02/24/2022 01:15:12 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 02/24/2022 01:15:13 - WARNING - huggingface_hub.repository - Revision `gallant-salad-7` does not exist. Created and checked out branch `gallant-salad-7`. 02/24/2022 01:15:13 - WARNING - huggingface_hub.repository - 02/24/2022 01:19:48 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 02/24/2022 01:19:49 - WARNING - huggingface_hub.repository - Revision `splendid-planet-8` does not exist. Created and checked out branch `splendid-planet-8`. 02/24/2022 01:19:49 - WARNING - huggingface_hub.repository - 02/24/2022 01:21:44 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 02/24/2022 01:21:44 - WARNING - huggingface_hub.repository - Revision `rich-lake-9` does not exist. Created and checked out branch `rich-lake-9`. 02/24/2022 01:21:44 - WARNING - huggingface_hub.repository - 02/24/2022 01:21:57 - WARNING - datasets.builder - Using custom data configuration lvwerra--codeparrot-clean-train-a1efdd1059bd841d 02/24/2022 01:21:59 - WARNING - datasets.builder - Using custom data configuration lvwerra--codeparrot-clean-valid-a800eb55c299abc0 02/24/2022 01:22:07 - INFO - codeparrot_training - Step 1: {'lr': 0.0, 'samples': 16, 'steps': 0, 'loss/train': 10.91077709197998} 02/24/2022 01:22:19 - INFO - codeparrot_training - Step 2: {'lr': 0.0, 'samples': 32, 'steps': 0, 'loss/train': 10.942980766296387} 02/24/2022 01:22:29 - INFO - codeparrot_training - Step 3: {'lr': 0.0, 'samples': 48, 'steps': 0, 'loss/train': 11.002518653869629} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 4: {'lr': 0.0, 'samples': 64, 'steps': 0, 'loss/train': 10.910642623901367} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 5: {'lr': 0.0, 'samples': 80, 'steps': 0, 'loss/train': 10.943111419677734} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 6: {'lr': 0.0, 'samples': 96, 'steps': 0, 'loss/train': 10.949981689453125} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 7: {'lr': 0.0, 'samples': 112, 'steps': 0, 'loss/train': 10.960782051086426} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 8: {'lr': 0.0, 'samples': 128, 'steps': 0, 'loss/train': 10.970269203186035} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 9: {'lr': 0.0, 'samples': 144, 'steps': 0, 'loss/train': 10.96359634399414} 02/24/2022 01:22:30 - INFO - codeparrot_training - Step 10: {'lr': 0.0, 'samples': 160, 'steps': 0, 'loss/train': 10.989349365234375} 02/24/2022 01:22:31 - INFO - codeparrot_training - Step 11: {'lr': 0.0, 'samples': 176, 'steps': 0, 'loss/train': 10.927996635437012} 02/24/2022 01:22:31 - INFO - codeparrot_training - Step 12: {'lr': 0.0, 'samples': 192, 'steps': 0, 'loss/train': 10.922099113464355} 02/24/2022 01:22:31 - INFO - codeparrot_training - Step 13: {'lr': 0.0, 'samples': 208, 'steps': 0, 'loss/train': 10.948519706726074} 02/24/2022 01:22:31 - INFO - codeparrot_training - Step 14: {'lr': 0.0, 'samples': 224, 'steps': 0, 'loss/train': 10.906710624694824} 02/24/2022 01:22:31 - INFO - codeparrot_training - Step 15: {'lr': 0.0, 'samples': 240, 'steps': 0, 'loss/train': 11.033504486083984} 02/24/2022 01:22:31 - INFO - codeparrot_training - Step 16: {'lr': 0.0, 'samples': 256, 'steps': 0, 'loss/train': 10.973214149475098} 02/24/2022 01:22:32 - INFO - codeparrot_training - Step 17: {'lr': 0.0, 'samples': 272, 'steps': 0, 'loss/train': 10.971339225769043} 02/24/2022 01:22:32 - INFO - codeparrot_training - Step 18: {'lr': 0.0, 'samples': 288, 'steps': 0, 'loss/train': 10.91922378540039} 02/24/2022 01:22:32 - INFO - codeparrot_training - Step 19: {'lr': 0.0, 'samples': 304, 'steps': 0, 'loss/train': 10.92506217956543} 02/24/2022 01:22:32 - INFO - codeparrot_training - Step 20: {'lr': 0.0, 'samples': 320, 'steps': 0, 'loss/train': 10.96170711517334} 02/24/2022 01:22:32 - INFO - codeparrot_training - Step 21: {'lr': 0.0, 'samples': 336, 'steps': 0, 'loss/train': 10.995943069458008} 02/24/2022 01:22:32 - INFO - codeparrot_training - Step 22: {'lr': 0.0, 'samples': 352, 'steps': 0, 'loss/train': 10.80260944366455} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 23: {'lr': 0.0, 'samples': 368, 'steps': 0, 'loss/train': 11.354506492614746} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 24: {'lr': 0.0, 'samples': 384, 'steps': 0, 'loss/train': 10.87732219696045} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 25: {'lr': 0.0, 'samples': 400, 'steps': 0, 'loss/train': 10.90416145324707} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 26: {'lr': 0.0, 'samples': 416, 'steps': 0, 'loss/train': 10.979859352111816} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 27: {'lr': 0.0, 'samples': 432, 'steps': 0, 'loss/train': 10.968029022216797} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 28: {'lr': 0.0, 'samples': 448, 'steps': 0, 'loss/train': 10.952177047729492} 02/24/2022 01:22:33 - INFO - codeparrot_training - Step 29: {'lr': 0.0, 'samples': 464, 'steps': 0, 'loss/train': 10.939998626708984} 02/24/2022 01:22:34 - INFO - codeparrot_training - Step 30: {'lr': 0.0, 'samples': 480, 'steps': 0, 'loss/train': 10.979020118713379} 02/24/2022 01:22:34 - INFO - codeparrot_training - Step 31: {'lr': 0.0, 'samples': 496, 'steps': 0, 'loss/train': 10.940269470214844} 02/24/2022 01:22:34 - INFO - codeparrot_training - Step 32: {'lr': 0.0, 'samples': 512, 'steps': 0, 'loss/train': 11.01601791381836} 02/24/2022 01:25:47 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 02/24/2022 01:25:48 - WARNING - huggingface_hub.repository - Revision `misty-fire-10` does not exist. Created and checked out branch `misty-fire-10`. 02/24/2022 01:25:48 - WARNING - huggingface_hub.repository - 02/24/2022 01:26:00 - WARNING - datasets.builder - Using custom data configuration lvwerra--codeparrot-clean-train-a1efdd1059bd841d 02/24/2022 01:26:02 - WARNING - datasets.builder - Using custom data configuration lvwerra--codeparrot-clean-valid-a800eb55c299abc0 02/24/2022 01:26:10 - INFO - codeparrot_training - Step 1: {'lr': 0.0, 'samples': 16, 'steps': 0, 'loss/train': 10.91077709197998} 02/24/2022 01:26:23 - INFO - codeparrot_training - Step 2: {'lr': 0.0, 'samples': 32, 'steps': 0, 'loss/train': 10.942980766296387} 02/24/2022 01:26:33 - INFO - codeparrot_training - Step 3: {'lr': 0.0, 'samples': 48, 'steps': 0, 'loss/train': 11.002518653869629} 02/24/2022 01:26:33 - INFO - codeparrot_training - Step 4: {'lr': 0.0, 'samples': 64, 'steps': 0, 'loss/train': 10.910642623901367} 02/24/2022 01:26:34 - INFO - codeparrot_training - Step 5: {'lr': 0.0, 'samples': 80, 'steps': 0, 'loss/train': 10.943111419677734} 02/24/2022 01:26:34 - INFO - codeparrot_training - Step 6: {'lr': 0.0, 'samples': 96, 'steps': 0, 'loss/train': 10.949981689453125} 02/24/2022 01:26:34 - INFO - codeparrot_training - Step 7: {'lr': 0.0, 'samples': 112, 'steps': 0, 'loss/train': 10.960782051086426} 02/24/2022 01:26:34 - INFO - codeparrot_training - Step 8: {'lr': 0.0, 'samples': 128, 'steps': 0, 'loss/train': 10.970269203186035} 02/24/2022 01:26:34 - INFO - codeparrot_training - Step 9: {'lr': 0.0, 'samples': 144, 'steps': 0, 'loss/train': 10.96359634399414} 02/24/2022 01:26:34 - INFO - codeparrot_training - Step 10: {'lr': 0.0, 'samples': 160, 'steps': 0, 'loss/train': 10.989349365234375} 02/24/2022 01:26:35 - INFO - codeparrot_training - Step 11: {'lr': 0.0, 'samples': 176, 'steps': 0, 'loss/train': 10.927996635437012} 02/24/2022 01:26:35 - INFO - codeparrot_training - Step 12: {'lr': 0.0, 'samples': 192, 'steps': 0, 'loss/train': 10.922099113464355} 02/24/2022 01:26:35 - INFO - codeparrot_training - Step 13: {'lr': 0.0, 'samples': 208, 'steps': 0, 'loss/train': 10.948519706726074} 02/24/2022 01:26:35 - INFO - codeparrot_training - Step 14: {'lr': 0.0, 'samples': 224, 'steps': 0, 'loss/train': 10.906710624694824} 02/24/2022 01:26:35 - INFO - codeparrot_training - Step 15: {'lr': 0.0, 'samples': 240, 'steps': 0, 'loss/train': 11.033504486083984} 02/24/2022 01:26:35 - INFO - codeparrot_training - Step 16: {'lr': 0.0, 'samples': 256, 'steps': 0, 'loss/train': 10.973214149475098} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 17: {'lr': 0.0, 'samples': 272, 'steps': 0, 'loss/train': 10.971339225769043} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 18: {'lr': 0.0, 'samples': 288, 'steps': 0, 'loss/train': 10.91922378540039} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 19: {'lr': 0.0, 'samples': 304, 'steps': 0, 'loss/train': 10.92506217956543} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 20: {'lr': 0.0, 'samples': 320, 'steps': 0, 'loss/train': 10.96170711517334} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 21: {'lr': 0.0, 'samples': 336, 'steps': 0, 'loss/train': 10.995943069458008} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 22: {'lr': 0.0, 'samples': 352, 'steps': 0, 'loss/train': 10.80260944366455} 02/24/2022 01:26:36 - INFO - codeparrot_training - Step 23: {'lr': 0.0, 'samples': 368, 'steps': 0, 'loss/train': 11.354506492614746} 02/24/2022 01:26:37 - INFO - codeparrot_training - Step 24: {'lr': 0.0, 'samples': 384, 'steps': 0, 'loss/train': 10.87732219696045} 02/24/2022 01:26:37 - INFO - codeparrot_training - Step 25: {'lr': 0.0, 'samples': 400, 'steps': 0, 'loss/train': 10.90416145324707} 02/24/2022 01:26:37 - INFO - codeparrot_training - Step 26: {'lr': 0.0, 'samples': 416, 'steps': 0, 'loss/train': 10.979859352111816} 02/24/2022 01:26:37 - INFO - codeparrot_training - Step 27: {'lr': 0.0, 'samples': 432, 'steps': 0, 'loss/train': 10.968029022216797} 02/24/2022 01:26:37 - INFO - codeparrot_training - Step 28: {'lr': 0.0, 'samples': 448, 'steps': 0, 'loss/train': 10.952177047729492} 02/24/2022 01:26:37 - INFO - codeparrot_training - Step 29: {'lr': 0.0, 'samples': 464, 'steps': 0, 'loss/train': 10.939998626708984} 02/24/2022 01:26:38 - INFO - codeparrot_training - Step 30: {'lr': 0.0, 'samples': 480, 'steps': 0, 'loss/train': 10.979020118713379} 02/24/2022 01:26:38 - INFO - codeparrot_training - Step 31: {'lr': 0.0, 'samples': 496, 'steps': 0, 'loss/train': 10.940269470214844} 02/24/2022 01:26:38 - INFO - codeparrot_training - Step 32: {'lr': 0.0, 'samples': 512, 'steps': 0, 'loss/train': 11.01601791381836} 02/24/2022 01:27:44 - INFO - codeparrot_training - Step 33: {'lr': 2.5e-07, 'samples': 528, 'steps': 1, 'loss/train': 10.974133491516113} 02/24/2022 01:27:44 - INFO - codeparrot_training - Step 34: {'lr': 2.5e-07, 'samples': 544, 'steps': 1, 'loss/train': 10.982958793640137} 02/24/2022 01:27:44 - INFO - codeparrot_training - Step 35: {'lr': 2.5e-07, 'samples': 560, 'steps': 1, 'loss/train': 10.95070743560791} 02/24/2022 01:27:45 - INFO - codeparrot_training - Step 36: {'lr': 2.5e-07, 'samples': 576, 'steps': 1, 'loss/train': 10.997182846069336} 02/24/2022 01:27:45 - INFO - codeparrot_training - Step 37: {'lr': 2.5e-07, 'samples': 592, 'steps': 1, 'loss/train': 10.971476554870605} 02/24/2022 01:27:45 - INFO - codeparrot_training - Step 38: {'lr': 2.5e-07, 'samples': 608, 'steps': 1, 'loss/train': 10.988201141357422} 02/24/2022 01:27:45 - INFO - codeparrot_training - Step 39: {'lr': 2.5e-07, 'samples': 624, 'steps': 1, 'loss/train': 10.913816452026367} 02/24/2022 01:27:45 - INFO - codeparrot_training - Step 40: {'lr': 2.5e-07, 'samples': 640, 'steps': 1, 'loss/train': 10.960057258605957} 02/24/2022 01:27:45 - INFO - codeparrot_training - Step 41: {'lr': 2.5e-07, 'samples': 656, 'steps': 1, 'loss/train': 10.945222854614258} 02/24/2022 01:27:46 - INFO - codeparrot_training - Step 42: {'lr': 2.5e-07, 'samples': 672, 'steps': 1, 'loss/train': 10.931300163269043} 02/24/2022 01:27:46 - INFO - codeparrot_training - Step 43: {'lr': 2.5e-07, 'samples': 688, 'steps': 1, 'loss/train': 10.93020248413086} 02/24/2022 01:27:46 - INFO - codeparrot_training - Step 44: {'lr': 2.5e-07, 'samples': 704, 'steps': 1, 'loss/train': 10.922496795654297} 02/24/2022 01:27:46 - INFO - codeparrot_training - Step 45: {'lr': 2.5e-07, 'samples': 720, 'steps': 1, 'loss/train': 10.934232711791992} 02/24/2022 01:27:46 - INFO - codeparrot_training - Step 46: {'lr': 2.5e-07, 'samples': 736, 'steps': 1, 'loss/train': 10.906913757324219} 02/24/2022 01:27:46 - INFO - codeparrot_training - Step 47: {'lr': 2.5e-07, 'samples': 752, 'steps': 1, 'loss/train': 10.988995552062988} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 48: {'lr': 2.5e-07, 'samples': 768, 'steps': 1, 'loss/train': 10.946273803710938} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 49: {'lr': 2.5e-07, 'samples': 784, 'steps': 1, 'loss/train': 10.94522476196289} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 50: {'lr': 2.5e-07, 'samples': 800, 'steps': 1, 'loss/train': 10.973355293273926} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 51: {'lr': 2.5e-07, 'samples': 816, 'steps': 1, 'loss/train': 10.965099334716797} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 52: {'lr': 2.5e-07, 'samples': 832, 'steps': 1, 'loss/train': 10.938562393188477} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 53: {'lr': 2.5e-07, 'samples': 848, 'steps': 1, 'loss/train': 10.90954303741455} 02/24/2022 01:27:47 - INFO - codeparrot_training - Step 54: {'lr': 2.5e-07, 'samples': 864, 'steps': 1, 'loss/train': 10.997761726379395} 02/24/2022 01:27:48 - INFO - codeparrot_training - Step 55: {'lr': 2.5e-07, 'samples': 880, 'steps': 1, 'loss/train': 10.994376182556152} 02/24/2022 01:27:48 - INFO - codeparrot_training - Step 56: {'lr': 2.5e-07, 'samples': 896, 'steps': 1, 'loss/train': 10.929353713989258} 02/24/2022 01:27:48 - INFO - codeparrot_training - Step 57: {'lr': 2.5e-07, 'samples': 912, 'steps': 1, 'loss/train': 10.877440452575684} 02/24/2022 01:27:48 - INFO - codeparrot_training - Step 58: {'lr': 2.5e-07, 'samples': 928, 'steps': 1, 'loss/train': 10.93001651763916} 02/24/2022 01:27:48 - INFO - codeparrot_training - Step 59: {'lr': 2.5e-07, 'samples': 944, 'steps': 1, 'loss/train': 10.937539100646973} 02/24/2022 01:27:48 - INFO - codeparrot_training - Step 60: {'lr': 2.5e-07, 'samples': 960, 'steps': 1, 'loss/train': 10.957945823669434} 02/24/2022 01:27:49 - INFO - codeparrot_training - Step 61: {'lr': 2.5e-07, 'samples': 976, 'steps': 1, 'loss/train': 11.001383781433105} 02/24/2022 01:27:49 - INFO - codeparrot_training - Step 62: {'lr': 2.5e-07, 'samples': 992, 'steps': 1, 'loss/train': 10.79994010925293} 02/24/2022 01:27:49 - INFO - codeparrot_training - Step 63: {'lr': 2.5e-07, 'samples': 1008, 'steps': 1, 'loss/train': 11.036500930786133} 02/24/2022 01:27:51 - INFO - codeparrot_training - Step 64: {'lr': 2.5e-07, 'samples': 1024, 'steps': 1, 'loss/train': 10.936944961547852} 02/24/2022 01:29:01 - INFO - codeparrot_training - Step 65: {'lr': 5e-07, 'samples': 1040, 'steps': 2, 'loss/train': 10.962884902954102} 02/24/2022 01:29:01 - INFO - codeparrot_training - Step 66: {'lr': 5e-07, 'samples': 1056, 'steps': 2, 'loss/train': 10.924057960510254} 02/24/2022 01:29:01 - INFO - codeparrot_training - Step 67: {'lr': 5e-07, 'samples': 1072, 'steps': 2, 'loss/train': 11.00478744506836} 02/24/2022 01:29:01 - INFO - codeparrot_training - Step 68: {'lr': 5e-07, 'samples': 1088, 'steps': 2, 'loss/train': 10.938624382019043} 02/24/2022 01:29:02 - INFO - codeparrot_training - Step 69: {'lr': 5e-07, 'samples': 1104, 'steps': 2, 'loss/train': 10.946737289428711} 02/24/2022 01:29:02 - INFO - codeparrot_training - Step 70: {'lr': 5e-07, 'samples': 1120, 'steps': 2, 'loss/train': 10.97043514251709} 02/24/2022 01:29:02 - INFO - codeparrot_training - Step 71: {'lr': 5e-07, 'samples': 1136, 'steps': 2, 'loss/train': 10.919570922851562} 02/24/2022 01:29:02 - INFO - codeparrot_training - Step 72: {'lr': 5e-07, 'samples': 1152, 'steps': 2, 'loss/train': 10.94793701171875} 02/24/2022 01:29:02 - INFO - codeparrot_training - Step 73: {'lr': 5e-07, 'samples': 1168, 'steps': 2, 'loss/train': 10.879217147827148} 02/24/2022 01:29:02 - INFO - codeparrot_training - Step 74: {'lr': 5e-07, 'samples': 1184, 'steps': 2, 'loss/train': 10.92875862121582} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 75: {'lr': 5e-07, 'samples': 1200, 'steps': 2, 'loss/train': 10.892864227294922} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 76: {'lr': 5e-07, 'samples': 1216, 'steps': 2, 'loss/train': 10.862300872802734} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 77: {'lr': 5e-07, 'samples': 1232, 'steps': 2, 'loss/train': 10.956947326660156} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 78: {'lr': 5e-07, 'samples': 1248, 'steps': 2, 'loss/train': 10.952834129333496} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 79: {'lr': 5e-07, 'samples': 1264, 'steps': 2, 'loss/train': 10.932134628295898} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 80: {'lr': 5e-07, 'samples': 1280, 'steps': 2, 'loss/train': 10.967159271240234} 02/24/2022 01:29:03 - INFO - codeparrot_training - Step 81: {'lr': 5e-07, 'samples': 1296, 'steps': 2, 'loss/train': 10.990038871765137} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 82: {'lr': 5e-07, 'samples': 1312, 'steps': 2, 'loss/train': 10.931655883789062} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 83: {'lr': 5e-07, 'samples': 1328, 'steps': 2, 'loss/train': 11.00280475616455} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 84: {'lr': 5e-07, 'samples': 1344, 'steps': 2, 'loss/train': 10.966012954711914} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 85: {'lr': 5e-07, 'samples': 1360, 'steps': 2, 'loss/train': 10.966472625732422} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 86: {'lr': 5e-07, 'samples': 1376, 'steps': 2, 'loss/train': 11.024860382080078} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 87: {'lr': 5e-07, 'samples': 1392, 'steps': 2, 'loss/train': 10.982053756713867} 02/24/2022 01:29:04 - INFO - codeparrot_training - Step 88: {'lr': 5e-07, 'samples': 1408, 'steps': 2, 'loss/train': 10.985541343688965} 02/24/2022 01:29:05 - INFO - codeparrot_training - Step 89: {'lr': 5e-07, 'samples': 1424, 'steps': 2, 'loss/train': 10.979313850402832} 02/24/2022 01:29:05 - INFO - codeparrot_training - Step 90: {'lr': 5e-07, 'samples': 1440, 'steps': 2, 'loss/train': 10.98116683959961} 02/24/2022 01:29:05 - INFO - codeparrot_training - Step 91: {'lr': 5e-07, 'samples': 1456, 'steps': 2, 'loss/train': 11.013162612915039} 02/24/2022 01:29:05 - INFO - codeparrot_training - Step 92: {'lr': 5e-07, 'samples': 1472, 'steps': 2, 'loss/train': 10.944647789001465} 02/24/2022 01:29:05 - INFO - codeparrot_training - Step 93: {'lr': 5e-07, 'samples': 1488, 'steps': 2, 'loss/train': 10.926380157470703} 02/24/2022 01:29:05 - INFO - codeparrot_training - Step 94: {'lr': 5e-07, 'samples': 1504, 'steps': 2, 'loss/train': 10.939397811889648} 02/24/2022 01:29:06 - INFO - codeparrot_training - Step 95: {'lr': 5e-07, 'samples': 1520, 'steps': 2, 'loss/train': 10.954940795898438} 02/24/2022 01:29:06 - INFO - codeparrot_training - Step 96: {'lr': 5e-07, 'samples': 1536, 'steps': 2, 'loss/train': 10.925644874572754} 02/24/2022 01:29:06 - INFO - codeparrot_training - Step 97: {'lr': 7.5e-07, 'samples': 1552, 'steps': 3, 'loss/train': 10.895140647888184} 02/24/2022 01:29:06 - INFO - codeparrot_training - Step 98: {'lr': 7.5e-07, 'samples': 1568, 'steps': 3, 'loss/train': 10.933130264282227} 02/24/2022 01:29:06 - INFO - codeparrot_training - Step 99: {'lr': 7.5e-07, 'samples': 1584, 'steps': 3, 'loss/train': 10.823505401611328} 02/24/2022 01:29:06 - INFO - codeparrot_training - Step 100: {'lr': 7.5e-07, 'samples': 1600, 'steps': 3, 'loss/train': 10.901811599731445} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 101: {'lr': 7.5e-07, 'samples': 1616, 'steps': 3, 'loss/train': 10.928942680358887} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 102: {'lr': 7.5e-07, 'samples': 1632, 'steps': 3, 'loss/train': 10.981282234191895} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 103: {'lr': 7.5e-07, 'samples': 1648, 'steps': 3, 'loss/train': 10.933391571044922} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 104: {'lr': 7.5e-07, 'samples': 1664, 'steps': 3, 'loss/train': 10.925240516662598} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 105: {'lr': 7.5e-07, 'samples': 1680, 'steps': 3, 'loss/train': 10.990243911743164} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 106: {'lr': 7.5e-07, 'samples': 1696, 'steps': 3, 'loss/train': 10.938335418701172} 02/24/2022 01:29:07 - INFO - codeparrot_training - Step 107: {'lr': 7.5e-07, 'samples': 1712, 'steps': 3, 'loss/train': 10.967667579650879} 02/24/2022 01:29:08 - INFO - codeparrot_training - Step 108: {'lr': 7.5e-07, 'samples': 1728, 'steps': 3, 'loss/train': 10.93986988067627} 02/24/2022 01:29:08 - INFO - codeparrot_training - Step 109: {'lr': 7.5e-07, 'samples': 1744, 'steps': 3, 'loss/train': 10.925867080688477} 02/24/2022 01:29:08 - INFO - codeparrot_training - Step 110: {'lr': 7.5e-07, 'samples': 1760, 'steps': 3, 'loss/train': 10.97072982788086} 02/24/2022 01:29:08 - INFO - codeparrot_training - Step 111: {'lr': 7.5e-07, 'samples': 1776, 'steps': 3, 'loss/train': 10.872678756713867} 02/24/2022 01:29:08 - INFO - codeparrot_training - Step 112: {'lr': 7.5e-07, 'samples': 1792, 'steps': 3, 'loss/train': 10.886422157287598} 02/24/2022 01:29:08 - INFO - codeparrot_training - Step 113: {'lr': 7.5e-07, 'samples': 1808, 'steps': 3, 'loss/train': 10.965313911437988} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 114: {'lr': 7.5e-07, 'samples': 1824, 'steps': 3, 'loss/train': 10.914895057678223} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 115: {'lr': 7.5e-07, 'samples': 1840, 'steps': 3, 'loss/train': 10.9727144241333} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 116: {'lr': 7.5e-07, 'samples': 1856, 'steps': 3, 'loss/train': 10.957098960876465} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 117: {'lr': 7.5e-07, 'samples': 1872, 'steps': 3, 'loss/train': 10.932665824890137} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 118: {'lr': 7.5e-07, 'samples': 1888, 'steps': 3, 'loss/train': 10.878507614135742} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 119: {'lr': 7.5e-07, 'samples': 1904, 'steps': 3, 'loss/train': 10.957688331604004} 02/24/2022 01:29:09 - INFO - codeparrot_training - Step 120: {'lr': 7.5e-07, 'samples': 1920, 'steps': 3, 'loss/train': 10.916014671325684} 02/24/2022 01:29:10 - INFO - codeparrot_training - Step 121: {'lr': 7.5e-07, 'samples': 1936, 'steps': 3, 'loss/train': 10.942499160766602} 02/24/2022 01:29:10 - INFO - codeparrot_training - Step 122: {'lr': 7.5e-07, 'samples': 1952, 'steps': 3, 'loss/train': 10.985833168029785} 02/24/2022 01:29:10 - INFO - codeparrot_training - Step 123: {'lr': 7.5e-07, 'samples': 1968, 'steps': 3, 'loss/train': 10.920166969299316} 02/24/2022 01:29:10 - INFO - codeparrot_training - Step 124: {'lr': 7.5e-07, 'samples': 1984, 'steps': 3, 'loss/train': 10.89915657043457} 02/24/2022 01:29:10 - INFO - codeparrot_training - Step 125: {'lr': 7.5e-07, 'samples': 2000, 'steps': 3, 'loss/train': 10.908203125} 02/24/2022 01:29:10 - INFO - codeparrot_training - Step 126: {'lr': 7.5e-07, 'samples': 2016, 'steps': 3, 'loss/train': 10.928319931030273} 02/24/2022 01:29:11 - INFO - codeparrot_training - Step 127: {'lr': 7.5e-07, 'samples': 2032, 'steps': 3, 'loss/train': 10.972715377807617} 02/24/2022 01:29:14 - INFO - codeparrot_training - Step 128: {'lr': 7.5e-07, 'samples': 2048, 'steps': 3, 'loss/train': 10.941082954406738} 02/24/2022 01:29:14 - INFO - codeparrot_training - Step 129: {'lr': 1e-06, 'samples': 2064, 'steps': 4, 'loss/train': 10.916863441467285} 02/24/2022 01:29:14 - INFO - codeparrot_training - Step 130: {'lr': 1e-06, 'samples': 2080, 'steps': 4, 'loss/train': 10.913991928100586} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 131: {'lr': 1e-06, 'samples': 2096, 'steps': 4, 'loss/train': 10.973291397094727} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 132: {'lr': 1e-06, 'samples': 2112, 'steps': 4, 'loss/train': 10.985342979431152} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 133: {'lr': 1e-06, 'samples': 2128, 'steps': 4, 'loss/train': 10.939002990722656} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 134: {'lr': 1e-06, 'samples': 2144, 'steps': 4, 'loss/train': 10.843615531921387} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 135: {'lr': 1e-06, 'samples': 2160, 'steps': 4, 'loss/train': 10.945176124572754} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 136: {'lr': 1e-06, 'samples': 2176, 'steps': 4, 'loss/train': 10.920622825622559} 02/24/2022 01:29:15 - INFO - codeparrot_training - Step 137: {'lr': 1e-06, 'samples': 2192, 'steps': 4, 'loss/train': 10.99799919128418} 02/24/2022 01:29:16 - INFO - codeparrot_training - Step 138: {'lr': 1e-06, 'samples': 2208, 'steps': 4, 'loss/train': 10.955973625183105} 02/24/2022 01:29:16 - INFO - codeparrot_training - Step 139: {'lr': 1e-06, 'samples': 2224, 'steps': 4, 'loss/train': 10.947653770446777} 02/24/2022 01:29:16 - INFO - codeparrot_training - Step 140: {'lr': 1e-06, 'samples': 2240, 'steps': 4, 'loss/train': 10.857400894165039} 02/24/2022 01:29:16 - INFO - codeparrot_training - Step 141: {'lr': 1e-06, 'samples': 2256, 'steps': 4, 'loss/train': 10.980046272277832} 02/24/2022 01:29:16 - INFO - codeparrot_training - Step 142: {'lr': 1e-06, 'samples': 2272, 'steps': 4, 'loss/train': 10.897544860839844} 02/24/2022 01:29:16 - INFO - codeparrot_training - Step 143: {'lr': 1e-06, 'samples': 2288, 'steps': 4, 'loss/train': 10.952252388000488} 02/24/2022 01:29:17 - INFO - codeparrot_training - Step 144: {'lr': 1e-06, 'samples': 2304, 'steps': 4, 'loss/train': 10.939106941223145} 02/24/2022 01:29:17 - INFO - codeparrot_training - Step 145: {'lr': 1e-06, 'samples': 2320, 'steps': 4, 'loss/train': 10.965592384338379} 02/24/2022 01:29:17 - INFO - codeparrot_training - Step 146: {'lr': 1e-06, 'samples': 2336, 'steps': 4, 'loss/train': 10.850253105163574} 02/24/2022 01:29:17 - INFO - codeparrot_training - Step 147: {'lr': 1e-06, 'samples': 2352, 'steps': 4, 'loss/train': 10.900897026062012} 02/24/2022 01:29:17 - INFO - codeparrot_training - Step 148: {'lr': 1e-06, 'samples': 2368, 'steps': 4, 'loss/train': 10.954386711120605} 02/24/2022 01:29:17 - INFO - codeparrot_training - Step 149: {'lr': 1e-06, 'samples': 2384, 'steps': 4, 'loss/train': 10.88646411895752} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 150: {'lr': 1e-06, 'samples': 2400, 'steps': 4, 'loss/train': 10.871075630187988} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 151: {'lr': 1e-06, 'samples': 2416, 'steps': 4, 'loss/train': 10.888216018676758} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 152: {'lr': 1e-06, 'samples': 2432, 'steps': 4, 'loss/train': 10.9573392868042} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 153: {'lr': 1e-06, 'samples': 2448, 'steps': 4, 'loss/train': 10.916056632995605} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 154: {'lr': 1e-06, 'samples': 2464, 'steps': 4, 'loss/train': 10.897372245788574} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 155: {'lr': 1e-06, 'samples': 2480, 'steps': 4, 'loss/train': 10.932211875915527} 02/24/2022 01:29:18 - INFO - codeparrot_training - Step 156: {'lr': 1e-06, 'samples': 2496, 'steps': 4, 'loss/train': 10.901747703552246} 02/24/2022 01:29:19 - INFO - codeparrot_training - Step 157: {'lr': 1e-06, 'samples': 2512, 'steps': 4, 'loss/train': 10.91716194152832} 02/24/2022 01:29:19 - INFO - codeparrot_training - Step 158: {'lr': 1e-06, 'samples': 2528, 'steps': 4, 'loss/train': 10.892692565917969} 02/24/2022 01:29:19 - INFO - codeparrot_training - Step 159: {'lr': 1e-06, 'samples': 2544, 'steps': 4, 'loss/train': 10.95022201538086} 02/24/2022 01:29:19 - INFO - codeparrot_training - Step 160: {'lr': 1e-06, 'samples': 2560, 'steps': 4, 'loss/train': 11.07193374633789} 02/24/2022 01:29:19 - INFO - codeparrot_training - Step 161: {'lr': 1.25e-06, 'samples': 2576, 'steps': 5, 'loss/train': 11.085668563842773} 02/24/2022 01:29:19 - INFO - codeparrot_training - Step 162: {'lr': 1.25e-06, 'samples': 2592, 'steps': 5, 'loss/train': 10.96953010559082} 02/24/2022 01:29:20 - INFO - codeparrot_training - Step 163: {'lr': 1.25e-06, 'samples': 2608, 'steps': 5, 'loss/train': 10.864081382751465} 02/24/2022 01:29:20 - INFO - codeparrot_training - Step 164: {'lr': 1.25e-06, 'samples': 2624, 'steps': 5, 'loss/train': 10.895583152770996} 02/24/2022 01:29:20 - INFO - codeparrot_training - Step 165: {'lr': 1.25e-06, 'samples': 2640, 'steps': 5, 'loss/train': 10.876590728759766} 02/24/2022 01:29:20 - INFO - codeparrot_training - Step 166: {'lr': 1.25e-06, 'samples': 2656, 'steps': 5, 'loss/train': 10.900389671325684} 02/24/2022 01:29:20 - INFO - codeparrot_training - Step 167: {'lr': 1.25e-06, 'samples': 2672, 'steps': 5, 'loss/train': 10.90485668182373} 02/24/2022 01:29:20 - INFO - codeparrot_training - Step 168: {'lr': 1.25e-06, 'samples': 2688, 'steps': 5, 'loss/train': 10.92477035522461} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 169: {'lr': 1.25e-06, 'samples': 2704, 'steps': 5, 'loss/train': 10.900215148925781} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 170: {'lr': 1.25e-06, 'samples': 2720, 'steps': 5, 'loss/train': 10.924201011657715} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 171: {'lr': 1.25e-06, 'samples': 2736, 'steps': 5, 'loss/train': 10.936107635498047} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 172: {'lr': 1.25e-06, 'samples': 2752, 'steps': 5, 'loss/train': 10.946706771850586} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 173: {'lr': 1.25e-06, 'samples': 2768, 'steps': 5, 'loss/train': 10.901002883911133} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 174: {'lr': 1.25e-06, 'samples': 2784, 'steps': 5, 'loss/train': 10.934091567993164} 02/24/2022 01:29:21 - INFO - codeparrot_training - Step 175: {'lr': 1.25e-06, 'samples': 2800, 'steps': 5, 'loss/train': 10.858243942260742} 02/24/2022 01:29:22 - INFO - codeparrot_training - Step 176: {'lr': 1.25e-06, 'samples': 2816, 'steps': 5, 'loss/train': 10.875936508178711} 02/24/2022 01:29:22 - INFO - codeparrot_training - Step 177: {'lr': 1.25e-06, 'samples': 2832, 'steps': 5, 'loss/train': 10.9532470703125} 02/24/2022 01:29:22 - INFO - codeparrot_training - Step 178: {'lr': 1.25e-06, 'samples': 2848, 'steps': 5, 'loss/train': 10.994404792785645} 02/24/2022 01:29:22 - INFO - codeparrot_training - Step 179: {'lr': 1.25e-06, 'samples': 2864, 'steps': 5, 'loss/train': 10.915818214416504} 02/24/2022 01:29:22 - INFO - codeparrot_training - Step 180: {'lr': 1.25e-06, 'samples': 2880, 'steps': 5, 'loss/train': 10.936393737792969} 02/24/2022 01:29:22 - INFO - codeparrot_training - Step 181: {'lr': 1.25e-06, 'samples': 2896, 'steps': 5, 'loss/train': 10.88879680633545} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 182: {'lr': 1.25e-06, 'samples': 2912, 'steps': 5, 'loss/train': 10.824249267578125} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 183: {'lr': 1.25e-06, 'samples': 2928, 'steps': 5, 'loss/train': 10.954201698303223} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 184: {'lr': 1.25e-06, 'samples': 2944, 'steps': 5, 'loss/train': 10.881426811218262} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 185: {'lr': 1.25e-06, 'samples': 2960, 'steps': 5, 'loss/train': 10.915998458862305} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 186: {'lr': 1.25e-06, 'samples': 2976, 'steps': 5, 'loss/train': 10.899885177612305} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 187: {'lr': 1.25e-06, 'samples': 2992, 'steps': 5, 'loss/train': 10.919334411621094} 02/24/2022 01:29:23 - INFO - codeparrot_training - Step 188: {'lr': 1.25e-06, 'samples': 3008, 'steps': 5, 'loss/train': 10.756229400634766} 02/24/2022 01:29:24 - INFO - codeparrot_training - Step 189: {'lr': 1.25e-06, 'samples': 3024, 'steps': 5, 'loss/train': 10.891862869262695} 02/24/2022 01:29:24 - INFO - codeparrot_training - Step 190: {'lr': 1.25e-06, 'samples': 3040, 'steps': 5, 'loss/train': 10.897555351257324} 02/24/2022 01:29:24 - INFO - codeparrot_training - Step 191: {'lr': 1.25e-06, 'samples': 3056, 'steps': 5, 'loss/train': 10.805274963378906} 02/24/2022 01:29:26 - INFO - codeparrot_training - Step 192: {'lr': 1.25e-06, 'samples': 3072, 'steps': 5, 'loss/train': 10.89697551727295} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 193: {'lr': 1.5e-06, 'samples': 3088, 'steps': 6, 'loss/train': 10.931663513183594} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 194: {'lr': 1.5e-06, 'samples': 3104, 'steps': 6, 'loss/train': 10.827430725097656} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 195: {'lr': 1.5e-06, 'samples': 3120, 'steps': 6, 'loss/train': 10.844098091125488} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 196: {'lr': 1.5e-06, 'samples': 3136, 'steps': 6, 'loss/train': 10.851478576660156} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 197: {'lr': 1.5e-06, 'samples': 3152, 'steps': 6, 'loss/train': 11.007500648498535} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 198: {'lr': 1.5e-06, 'samples': 3168, 'steps': 6, 'loss/train': 10.888507843017578} 02/24/2022 01:29:27 - INFO - codeparrot_training - Step 199: {'lr': 1.5e-06, 'samples': 3184, 'steps': 6, 'loss/train': 10.888191223144531} 02/24/2022 01:29:28 - INFO - codeparrot_training - Step 200: {'lr': 1.5e-06, 'samples': 3200, 'steps': 6, 'loss/train': 10.859925270080566} 02/24/2022 01:29:28 - INFO - codeparrot_training - Step 201: {'lr': 1.5e-06, 'samples': 3216, 'steps': 6, 'loss/train': 10.911829948425293} 02/24/2022 01:29:28 - INFO - codeparrot_training - Step 202: {'lr': 1.5e-06, 'samples': 3232, 'steps': 6, 'loss/train': 10.920969009399414} 02/24/2022 01:29:28 - INFO - codeparrot_training - Step 203: {'lr': 1.5e-06, 'samples': 3248, 'steps': 6, 'loss/train': 11.01217269897461} 02/24/2022 01:29:28 - INFO - codeparrot_training - Step 204: {'lr': 1.5e-06, 'samples': 3264, 'steps': 6, 'loss/train': 10.69102954864502} 02/24/2022 01:29:28 - INFO - codeparrot_training - Step 205: {'lr': 1.5e-06, 'samples': 3280, 'steps': 6, 'loss/train': 10.810357093811035} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 206: {'lr': 1.5e-06, 'samples': 3296, 'steps': 6, 'loss/train': 10.828718185424805} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 207: {'lr': 1.5e-06, 'samples': 3312, 'steps': 6, 'loss/train': 10.842058181762695} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 208: {'lr': 1.5e-06, 'samples': 3328, 'steps': 6, 'loss/train': 10.93617057800293} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 209: {'lr': 1.5e-06, 'samples': 3344, 'steps': 6, 'loss/train': 10.911834716796875} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 210: {'lr': 1.5e-06, 'samples': 3360, 'steps': 6, 'loss/train': 10.907511711120605} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 211: {'lr': 1.5e-06, 'samples': 3376, 'steps': 6, 'loss/train': 10.88161849975586} 02/24/2022 01:29:29 - INFO - codeparrot_training - Step 212: {'lr': 1.5e-06, 'samples': 3392, 'steps': 6, 'loss/train': 10.922626495361328} 02/24/2022 01:29:30 - INFO - codeparrot_training - Step 213: {'lr': 1.5e-06, 'samples': 3408, 'steps': 6, 'loss/train': 10.967133522033691} 02/24/2022 01:29:30 - INFO - codeparrot_training - Step 214: {'lr': 1.5e-06, 'samples': 3424, 'steps': 6, 'loss/train': 10.916214942932129} 02/24/2022 01:29:30 - INFO - codeparrot_training - Step 215: {'lr': 1.5e-06, 'samples': 3440, 'steps': 6, 'loss/train': 10.900676727294922} 02/24/2022 01:29:30 - INFO - codeparrot_training - Step 216: {'lr': 1.5e-06, 'samples': 3456, 'steps': 6, 'loss/train': 10.85409164428711} 02/24/2022 01:29:30 - INFO - codeparrot_training - Step 217: {'lr': 1.5e-06, 'samples': 3472, 'steps': 6, 'loss/train': 10.870858192443848} 02/24/2022 01:29:30 - INFO - codeparrot_training - Step 218: {'lr': 1.5e-06, 'samples': 3488, 'steps': 6, 'loss/train': 10.858830451965332} 02/24/2022 01:29:31 - INFO - codeparrot_training - Step 219: {'lr': 1.5e-06, 'samples': 3504, 'steps': 6, 'loss/train': 10.821305274963379} 02/24/2022 01:29:31 - INFO - codeparrot_training - Step 220: {'lr': 1.5e-06, 'samples': 3520, 'steps': 6, 'loss/train': 10.941291809082031} 02/24/2022 01:29:31 - INFO - codeparrot_training - Step 221: {'lr': 1.5e-06, 'samples': 3536, 'steps': 6, 'loss/train': 10.885320663452148} 02/24/2022 01:29:31 - INFO - codeparrot_training - Step 222: {'lr': 1.5e-06, 'samples': 3552, 'steps': 6, 'loss/train': 10.896737098693848} 02/24/2022 01:29:31 - INFO - codeparrot_training - Step 223: {'lr': 1.5e-06, 'samples': 3568, 'steps': 6, 'loss/train': 10.932000160217285} 02/24/2022 01:29:31 - INFO - codeparrot_training - Step 224: {'lr': 1.5e-06, 'samples': 3584, 'steps': 6, 'loss/train': 10.873857498168945} 02/24/2022 01:29:32 - INFO - codeparrot_training - Step 225: {'lr': 1.75e-06, 'samples': 3600, 'steps': 7, 'loss/train': 10.845444679260254} 02/24/2022 01:29:32 - INFO - codeparrot_training - Step 226: {'lr': 1.75e-06, 'samples': 3616, 'steps': 7, 'loss/train': 10.83541488647461} 02/24/2022 01:29:32 - INFO - codeparrot_training - Step 227: {'lr': 1.75e-06, 'samples': 3632, 'steps': 7, 'loss/train': 10.851716995239258} 02/24/2022 01:31:34 - INFO - codeparrot_training - Distributed environment: TPU Num processes: 8 Process index: 0 Local process index: 0 Device: xla:1 Use FP16 precision: False 02/24/2022 01:31:34 - WARNING - huggingface_hub.repository - Revision `floral-grass-11` does not exist. Created and checked out branch `floral-grass-11`. 02/24/2022 01:31:34 - WARNING - huggingface_hub.repository - 02/24/2022 01:31:47 - WARNING - datasets.builder - Using custom data configuration lvwerra--codeparrot-clean-train-a1efdd1059bd841d 02/24/2022 01:31:48 - WARNING - datasets.builder - Using custom data configuration lvwerra--codeparrot-clean-valid-a800eb55c299abc0 02/24/2022 01:32:22 - INFO - codeparrot_training - Step 0: {'lr': 0.0, 'samples': 512, 'steps': 0, 'loss/train': 11.01601791381836} 02/24/2022 01:33:30 - INFO - codeparrot_training - Step 1: {'lr': 2.5e-07, 'samples': 1024, 'steps': 1, 'loss/train': 10.936944961547852} 02/24/2022 01:34:46 - INFO - codeparrot_training - Step 2: {'lr': 5e-07, 'samples': 1536, 'steps': 2, 'loss/train': 10.925644874572754} 02/24/2022 01:34:52 - INFO - codeparrot_training - Step 3: {'lr': 7.5e-07, 'samples': 2048, 'steps': 3, 'loss/train': 10.941082954406738} 02/24/2022 01:34:56 - INFO - codeparrot_training - Step 4: {'lr': 1e-06, 'samples': 2560, 'steps': 4, 'loss/train': 11.07193374633789} 02/24/2022 01:35:01 - INFO - codeparrot_training - Step 5: {'lr': 1.25e-06, 'samples': 3072, 'steps': 5, 'loss/train': 10.89697551727295} 02/24/2022 01:35:05 - INFO - codeparrot_training - Step 6: {'lr': 1.5e-06, 'samples': 3584, 'steps': 6, 'loss/train': 10.873857498168945} 02/24/2022 01:35:10 - INFO - codeparrot_training - Step 7: {'lr': 1.75e-06, 'samples': 4096, 'steps': 7, 'loss/train': 10.825765609741211} 02/24/2022 01:35:13 - INFO - codeparrot_training - Step 8: {'lr': 2e-06, 'samples': 4608, 'steps': 8, 'loss/train': 10.752714157104492} 02/24/2022 01:35:19 - INFO - codeparrot_training - Step 9: {'lr': 2.25e-06, 'samples': 5120, 'steps': 9, 'loss/train': 10.731087684631348} 02/24/2022 01:35:22 - INFO - codeparrot_training - Step 10: {'lr': 2.5e-06, 'samples': 5632, 'steps': 10, 'loss/train': 10.78388786315918} 02/24/2022 01:35:28 - INFO - codeparrot_training - Step 11: {'lr': 2.75e-06, 'samples': 6144, 'steps': 11, 'loss/train': 10.686656951904297} 02/24/2022 01:35:31 - INFO - codeparrot_training - Step 12: {'lr': 3e-06, 'samples': 6656, 'steps': 12, 'loss/train': 10.692927360534668} 02/24/2022 01:35:38 - INFO - codeparrot_training - Step 13: {'lr': 3.25e-06, 'samples': 7168, 'steps': 13, 'loss/train': 10.630372047424316} 02/24/2022 01:35:41 - INFO - codeparrot_training - Step 14: {'lr': 3.5e-06, 'samples': 7680, 'steps': 14, 'loss/train': 10.618318557739258} 02/24/2022 01:35:46 - INFO - codeparrot_training - Step 15: {'lr': 3.75e-06, 'samples': 8192, 'steps': 15, 'loss/train': 10.532057762145996} 02/24/2022 01:35:50 - INFO - codeparrot_training - Step 16: {'lr': 4e-06, 'samples': 8704, 'steps': 16, 'loss/train': 10.441239356994629} 02/24/2022 01:35:55 - INFO - codeparrot_training - Step 17: {'lr': 4.250000000000001e-06, 'samples': 9216, 'steps': 17, 'loss/train': 10.438984870910645} 02/24/2022 01:35:59 - INFO - codeparrot_training - Step 18: {'lr': 4.5e-06, 'samples': 9728, 'steps': 18, 'loss/train': 10.231606483459473} 02/24/2022 01:36:04 - INFO - codeparrot_training - Step 19: {'lr': 4.75e-06, 'samples': 10240, 'steps': 19, 'loss/train': 10.179640769958496} 02/24/2022 01:36:07 - INFO - codeparrot_training - Step 20: {'lr': 5e-06, 'samples': 10752, 'steps': 20, 'loss/train': 9.973851203918457} 02/24/2022 01:36:13 - INFO - codeparrot_training - Step 21: {'lr': 5.2500000000000006e-06, 'samples': 11264, 'steps': 21, 'loss/train': 9.950591087341309} 02/24/2022 01:36:16 - INFO - codeparrot_training - Step 22: {'lr': 5.5e-06, 'samples': 11776, 'steps': 22, 'loss/train': 10.476153373718262} 02/24/2022 01:36:23 - INFO - codeparrot_training - Step 23: {'lr': 5.75e-06, 'samples': 12288, 'steps': 23, 'loss/train': 9.552414894104004} 02/24/2022 01:36:26 - INFO - codeparrot_training - Step 24: {'lr': 6e-06, 'samples': 12800, 'steps': 24, 'loss/train': 10.253140449523926} 02/24/2022 01:36:32 - INFO - codeparrot_training - Step 25: {'lr': 6.25e-06, 'samples': 13312, 'steps': 25, 'loss/train': 9.579047203063965} 02/24/2022 01:36:35 - INFO - codeparrot_training - Step 26: {'lr': 6.5e-06, 'samples': 13824, 'steps': 26, 'loss/train': 10.285812377929688} 02/24/2022 01:36:41 - INFO - codeparrot_training - Step 27: {'lr': 6.75e-06, 'samples': 14336, 'steps': 27, 'loss/train': 10.226362228393555} 02/24/2022 01:36:44 - INFO - codeparrot_training - Step 28: {'lr': 7e-06, 'samples': 14848, 'steps': 28, 'loss/train': 9.725104331970215} 02/24/2022 01:36:50 - INFO - codeparrot_training - Step 29: {'lr': 7.250000000000001e-06, 'samples': 15360, 'steps': 29, 'loss/train': 9.937402725219727} 02/24/2022 01:36:53 - INFO - codeparrot_training - Step 30: {'lr': 7.5e-06, 'samples': 15872, 'steps': 30, 'loss/train': 9.978718757629395} 02/24/2022 01:36:59 - INFO - codeparrot_training - Step 31: {'lr': 7.75e-06, 'samples': 16384, 'steps': 31, 'loss/train': 9.851637840270996} 02/24/2022 01:37:02 - INFO - codeparrot_training - Step 32: {'lr': 8e-06, 'samples': 16896, 'steps': 32, 'loss/train': 9.715484619140625} 02/24/2022 01:37:08 - INFO - codeparrot_training - Step 33: {'lr': 8.25e-06, 'samples': 17408, 'steps': 33, 'loss/train': 9.392934799194336} 02/24/2022 01:37:11 - INFO - codeparrot_training - Step 34: {'lr': 8.500000000000002e-06, 'samples': 17920, 'steps': 34, 'loss/train': 9.88122844696045} 02/24/2022 01:37:18 - INFO - codeparrot_training - Step 35: {'lr': 8.750000000000001e-06, 'samples': 18432, 'steps': 35, 'loss/train': 9.911639213562012} 02/24/2022 01:37:21 - INFO - codeparrot_training - Step 36: {'lr': 9e-06, 'samples': 18944, 'steps': 36, 'loss/train': 10.15811538696289} 02/24/2022 01:37:26 - INFO - codeparrot_training - Step 37: {'lr': 9.25e-06, 'samples': 19456, 'steps': 37, 'loss/train': 9.33040714263916} 02/24/2022 01:37:30 - INFO - codeparrot_training - Step 38: {'lr': 9.5e-06, 'samples': 19968, 'steps': 38, 'loss/train': 9.846328735351562} 02/24/2022 01:37:35 - INFO - codeparrot_training - Step 39: {'lr': 9.75e-06, 'samples': 20480, 'steps': 39, 'loss/train': 9.611072540283203} 02/24/2022 01:37:39 - INFO - codeparrot_training - Step 40: {'lr': 1e-05, 'samples': 20992, 'steps': 40, 'loss/train': 9.038996696472168} 02/24/2022 01:37:44 - INFO - codeparrot_training - Step 41: {'lr': 1.025e-05, 'samples': 21504, 'steps': 41, 'loss/train': 9.312322616577148} 02/24/2022 01:37:48 - INFO - codeparrot_training - Step 42: {'lr': 1.0500000000000001e-05, 'samples': 22016, 'steps': 42, 'loss/train': 9.412076950073242} 02/24/2022 01:37:53 - INFO - codeparrot_training - Step 43: {'lr': 1.0749999999999999e-05, 'samples': 22528, 'steps': 43, 'loss/train': 9.228338241577148} 02/24/2022 01:37:57 - INFO - codeparrot_training - Step 44: {'lr': 1.1e-05, 'samples': 23040, 'steps': 44, 'loss/train': 9.625134468078613} 02/24/2022 01:38:03 - INFO - codeparrot_training - Step 45: {'lr': 1.1249999999999999e-05, 'samples': 23552, 'steps': 45, 'loss/train': 9.419281005859375} 02/24/2022 01:38:06 - INFO - codeparrot_training - Step 46: {'lr': 1.15e-05, 'samples': 24064, 'steps': 46, 'loss/train': 9.415852546691895} 02/24/2022 01:38:11 - INFO - codeparrot_training - Step 47: {'lr': 1.1750000000000001e-05, 'samples': 24576, 'steps': 47, 'loss/train': 9.942625045776367} 02/24/2022 01:38:15 - INFO - codeparrot_training - Step 48: {'lr': 1.2e-05, 'samples': 25088, 'steps': 48, 'loss/train': 9.21979808807373} 02/24/2022 01:38:20 - INFO - codeparrot_training - Step 49: {'lr': 1.2250000000000001e-05, 'samples': 25600, 'steps': 49, 'loss/train': 9.481938362121582} 02/24/2022 01:38:26 - INFO - codeparrot_training - Step 50: {'lr': 1.25e-05, 'samples': 26112, 'steps': 50, 'loss/train': 9.65326976776123} 02/24/2022 01:38:29 - INFO - codeparrot_training - Step 51: {'lr': 1.275e-05, 'samples': 26624, 'steps': 51, 'loss/train': 9.695452690124512} 02/24/2022 01:38:35 - INFO - codeparrot_training - Step 52: {'lr': 1.3e-05, 'samples': 27136, 'steps': 52, 'loss/train': 9.77243709564209} 02/24/2022 01:38:38 - INFO - codeparrot_training - Step 53: {'lr': 1.325e-05, 'samples': 27648, 'steps': 53, 'loss/train': 9.193220138549805} 02/24/2022 01:38:44 - INFO - codeparrot_training - Step 54: {'lr': 1.35e-05, 'samples': 28160, 'steps': 54, 'loss/train': 9.582751274108887} 02/24/2022 01:38:47 - INFO - codeparrot_training - Step 55: {'lr': 1.375e-05, 'samples': 28672, 'steps': 55, 'loss/train': 10.538885116577148} 02/24/2022 01:38:53 - INFO - codeparrot_training - Step 56: {'lr': 1.4e-05, 'samples': 29184, 'steps': 56, 'loss/train': 9.603676795959473} 02/24/2022 01:38:56 - INFO - codeparrot_training - Step 57: {'lr': 1.425e-05, 'samples': 29696, 'steps': 57, 'loss/train': 9.933730125427246} 02/24/2022 01:39:02 - INFO - codeparrot_training - Step 58: {'lr': 1.4500000000000002e-05, 'samples': 30208, 'steps': 58, 'loss/train': 10.022768020629883} 02/24/2022 01:39:06 - INFO - codeparrot_training - Step 59: {'lr': 1.475e-05, 'samples': 30720, 'steps': 59, 'loss/train': 9.456661224365234} 02/24/2022 01:39:11 - INFO - codeparrot_training - Step 60: {'lr': 1.5e-05, 'samples': 31232, 'steps': 60, 'loss/train': 9.61857795715332} 02/24/2022 01:39:15 - INFO - codeparrot_training - Step 61: {'lr': 1.525e-05, 'samples': 31744, 'steps': 61, 'loss/train': 9.595654487609863} 02/24/2022 01:39:20 - INFO - codeparrot_training - Step 62: {'lr': 1.55e-05, 'samples': 32256, 'steps': 62, 'loss/train': 9.803630828857422} 02/24/2022 01:39:23 - INFO - codeparrot_training - Step 63: {'lr': 1.575e-05, 'samples': 32768, 'steps': 63, 'loss/train': 9.261384010314941} 02/24/2022 01:39:29 - INFO - codeparrot_training - Step 64: {'lr': 1.6e-05, 'samples': 33280, 'steps': 64, 'loss/train': 9.71921443939209} 02/24/2022 01:39:33 - INFO - codeparrot_training - Step 65: {'lr': 1.6250000000000002e-05, 'samples': 33792, 'steps': 65, 'loss/train': 9.623526573181152} 02/24/2022 01:39:38 - INFO - codeparrot_training - Step 66: {'lr': 1.65e-05, 'samples': 34304, 'steps': 66, 'loss/train': 9.844499588012695} 02/24/2022 01:39:42 - INFO - codeparrot_training - Step 67: {'lr': 1.675e-05, 'samples': 34816, 'steps': 67, 'loss/train': 9.011263847351074} 02/24/2022 01:39:47 - INFO - codeparrot_training - Step 68: {'lr': 1.7000000000000003e-05, 'samples': 35328, 'steps': 68, 'loss/train': 9.395312309265137} 02/24/2022 01:39:50 - INFO - codeparrot_training - Step 69: {'lr': 1.7250000000000003e-05, 'samples': 35840, 'steps': 69, 'loss/train': 9.74931812286377} 02/24/2022 01:39:57 - INFO - codeparrot_training - Step 70: {'lr': 1.7500000000000002e-05, 'samples': 36352, 'steps': 70, 'loss/train': 9.2435884475708} 02/24/2022 01:40:00 - INFO - codeparrot_training - Step 71: {'lr': 1.7749999999999998e-05, 'samples': 36864, 'steps': 71, 'loss/train': 9.436172485351562} 02/24/2022 01:40:06 - INFO - codeparrot_training - Step 72: {'lr': 1.8e-05, 'samples': 37376, 'steps': 72, 'loss/train': 9.279879570007324} 02/24/2022 01:40:09 - INFO - codeparrot_training - Step 73: {'lr': 1.825e-05, 'samples': 37888, 'steps': 73, 'loss/train': 9.521175384521484} 02/24/2022 01:40:15 - INFO - codeparrot_training - Step 74: {'lr': 1.85e-05, 'samples': 38400, 'steps': 74, 'loss/train': 9.995550155639648} 02/24/2022 01:40:18 - INFO - codeparrot_training - Step 75: {'lr': 1.875e-05, 'samples': 38912, 'steps': 75, 'loss/train': 9.35321044921875} 02/24/2022 01:40:23 - INFO - codeparrot_training - Step 76: {'lr': 1.9e-05, 'samples': 39424, 'steps': 76, 'loss/train': 9.540994644165039} 02/24/2022 01:40:27 - INFO - codeparrot_training - Step 77: {'lr': 1.925e-05, 'samples': 39936, 'steps': 77, 'loss/train': 9.684625625610352} 02/24/2022 01:40:32 - INFO - codeparrot_training - Step 78: {'lr': 1.95e-05, 'samples': 40448, 'steps': 78, 'loss/train': 9.534080505371094} 02/24/2022 01:40:36 - INFO - codeparrot_training - Step 79: {'lr': 1.975e-05, 'samples': 40960, 'steps': 79, 'loss/train': 9.40109920501709} 02/24/2022 01:40:42 - INFO - codeparrot_training - Step 80: {'lr': 2e-05, 'samples': 41472, 'steps': 80, 'loss/train': 9.779088020324707} 02/24/2022 01:40:46 - INFO - codeparrot_training - Step 81: {'lr': 2.025e-05, 'samples': 41984, 'steps': 81, 'loss/train': 9.618664741516113} 02/24/2022 01:40:51 - INFO - codeparrot_training - Step 82: {'lr': 2.05e-05, 'samples': 42496, 'steps': 82, 'loss/train': 9.514649391174316} 02/24/2022 01:40:55 - INFO - codeparrot_training - Step 83: {'lr': 2.0750000000000003e-05, 'samples': 43008, 'steps': 83, 'loss/train': 9.943718910217285} 02/24/2022 01:41:00 - INFO - codeparrot_training - Step 84: {'lr': 2.1000000000000002e-05, 'samples': 43520, 'steps': 84, 'loss/train': 9.961823463439941} 02/24/2022 01:41:04 - INFO - codeparrot_training - Step 85: {'lr': 2.125e-05, 'samples': 44032, 'steps': 85, 'loss/train': 9.23841667175293} 02/24/2022 01:41:09 - INFO - codeparrot_training - Step 86: {'lr': 2.1499999999999997e-05, 'samples': 44544, 'steps': 86, 'loss/train': 9.898496627807617} 02/24/2022 01:41:12 - INFO - codeparrot_training - Step 87: {'lr': 2.175e-05, 'samples': 45056, 'steps': 87, 'loss/train': 10.449203491210938} 02/24/2022 01:41:18 - INFO - codeparrot_training - Step 88: {'lr': 2.2e-05, 'samples': 45568, 'steps': 88, 'loss/train': 9.512351036071777} 02/24/2022 01:41:21 - INFO - codeparrot_training - Step 89: {'lr': 2.225e-05, 'samples': 46080, 'steps': 89, 'loss/train': 9.259896278381348} 02/24/2022 01:41:27 - INFO - codeparrot_training - Step 90: {'lr': 2.2499999999999998e-05, 'samples': 46592, 'steps': 90, 'loss/train': 9.768840789794922} 02/24/2022 01:41:31 - INFO - codeparrot_training - Step 91: {'lr': 2.275e-05, 'samples': 47104, 'steps': 91, 'loss/train': 9.244466781616211} 02/24/2022 01:41:36 - INFO - codeparrot_training - Step 92: {'lr': 2.3e-05, 'samples': 47616, 'steps': 92, 'loss/train': 9.151692390441895} 02/24/2022 01:41:40 - INFO - codeparrot_training - Step 93: {'lr': 2.325e-05, 'samples': 48128, 'steps': 93, 'loss/train': 9.288126945495605} 02/24/2022 01:41:45 - INFO - codeparrot_training - Step 94: {'lr': 2.3500000000000002e-05, 'samples': 48640, 'steps': 94, 'loss/train': 9.82015609741211} 02/24/2022 01:41:49 - INFO - codeparrot_training - Step 95: {'lr': 2.375e-05, 'samples': 49152, 'steps': 95, 'loss/train': 10.049545288085938} 02/24/2022 01:41:54 - INFO - codeparrot_training - Step 96: {'lr': 2.4e-05, 'samples': 49664, 'steps': 96, 'loss/train': 9.537553787231445} 02/24/2022 01:41:58 - INFO - codeparrot_training - Step 97: {'lr': 2.425e-05, 'samples': 50176, 'steps': 97, 'loss/train': 9.354862213134766} 02/24/2022 01:42:03 - INFO - codeparrot_training - Step 98: {'lr': 2.4500000000000003e-05, 'samples': 50688, 'steps': 98, 'loss/train': 9.080880165100098} 02/24/2022 01:42:07 - INFO - codeparrot_training - Step 99: {'lr': 2.4750000000000002e-05, 'samples': 51200, 'steps': 99, 'loss/train': 8.6892671585083} 02/24/2022 01:42:13 - INFO - codeparrot_training - Step 100: {'lr': 2.5e-05, 'samples': 51712, 'steps': 100, 'loss/train': 9.663463592529297} 02/24/2022 01:42:18 - INFO - codeparrot_training - Step 101: {'lr': 2.525e-05, 'samples': 52224, 'steps': 101, 'loss/train': 9.291194915771484} 02/24/2022 01:42:22 - INFO - codeparrot_training - Step 102: {'lr': 2.55e-05, 'samples': 52736, 'steps': 102, 'loss/train': 9.83191967010498} 02/24/2022 01:42:27 - INFO - codeparrot_training - Step 103: {'lr': 2.575e-05, 'samples': 53248, 'steps': 103, 'loss/train': 9.29751205444336} 02/24/2022 01:42:30 - INFO - codeparrot_training - Step 104: {'lr': 2.6e-05, 'samples': 53760, 'steps': 104, 'loss/train': 9.098075866699219} 02/24/2022 01:42:36 - INFO - codeparrot_training - Step 105: {'lr': 2.625e-05, 'samples': 54272, 'steps': 105, 'loss/train': 8.863370895385742} 02/24/2022 01:42:39 - INFO - codeparrot_training - Step 106: {'lr': 2.65e-05, 'samples': 54784, 'steps': 106, 'loss/train': 9.174436569213867} 02/24/2022 01:42:45 - INFO - codeparrot_training - Step 107: {'lr': 2.675e-05, 'samples': 55296, 'steps': 107, 'loss/train': 9.370267868041992} 02/24/2022 01:42:48 - INFO - codeparrot_training - Step 108: {'lr': 2.7e-05, 'samples': 55808, 'steps': 108, 'loss/train': 8.78602123260498} 02/24/2022 01:42:54 - INFO - codeparrot_training - Step 109: {'lr': 2.725e-05, 'samples': 56320, 'steps': 109, 'loss/train': 9.10840129852295} 02/24/2022 01:42:57 - INFO - codeparrot_training - Step 110: {'lr': 2.75e-05, 'samples': 56832, 'steps': 110, 'loss/train': 9.094649314880371} 02/24/2022 01:43:02 - INFO - codeparrot_training - Step 111: {'lr': 2.775e-05, 'samples': 57344, 'steps': 111, 'loss/train': 9.771917343139648} 02/24/2022 01:43:06 - INFO - codeparrot_training - Step 112: {'lr': 2.8e-05, 'samples': 57856, 'steps': 112, 'loss/train': 9.095439910888672} 02/24/2022 01:43:12 - INFO - codeparrot_training - Step 113: {'lr': 2.8250000000000002e-05, 'samples': 58368, 'steps': 113, 'loss/train': 9.394959449768066} 02/24/2022 01:43:15 - INFO - codeparrot_training - Step 114: {'lr': 2.85e-05, 'samples': 58880, 'steps': 114, 'loss/train': 9.721589088439941} 02/24/2022 01:43:21 - INFO - codeparrot_training - Step 115: {'lr': 2.875e-05, 'samples': 59392, 'steps': 115, 'loss/train': 7.896186828613281} 02/24/2022 01:43:24 - INFO - codeparrot_training - Step 116: {'lr': 2.9000000000000004e-05, 'samples': 59904, 'steps': 116, 'loss/train': 7.90242862701416} 02/24/2022 01:43:30 - INFO - codeparrot_training - Step 117: {'lr': 2.9250000000000003e-05, 'samples': 60416, 'steps': 117, 'loss/train': 9.375008583068848} 02/24/2022 01:43:33 - INFO - codeparrot_training - Step 118: {'lr': 2.95e-05, 'samples': 60928, 'steps': 118, 'loss/train': 8.5529203414917} 02/24/2022 01:43:39 - INFO - codeparrot_training - Step 119: {'lr': 2.9749999999999998e-05, 'samples': 61440, 'steps': 119, 'loss/train': 8.504928588867188} 02/24/2022 01:43:42 - INFO - codeparrot_training - Step 120: {'lr': 3e-05, 'samples': 61952, 'steps': 120, 'loss/train': 9.61538028717041} 02/24/2022 01:43:48 - INFO - codeparrot_training - Step 121: {'lr': 3.025e-05, 'samples': 62464, 'steps': 121, 'loss/train': 9.16252613067627} 02/24/2022 01:43:51 - INFO - codeparrot_training - Step 122: {'lr': 3.05e-05, 'samples': 62976, 'steps': 122, 'loss/train': 9.315605163574219} 02/24/2022 01:43:57 - INFO - codeparrot_training - Step 123: {'lr': 3.075e-05, 'samples': 63488, 'steps': 123, 'loss/train': 9.257500648498535} 02/24/2022 01:44:00 - INFO - codeparrot_training - Step 124: {'lr': 3.1e-05, 'samples': 64000, 'steps': 124, 'loss/train': 8.4464693069458} 02/24/2022 01:44:06 - INFO - codeparrot_training - Step 125: {'lr': 3.125e-05, 'samples': 64512, 'steps': 125, 'loss/train': 9.038679122924805} 02/24/2022 01:44:10 - INFO - codeparrot_training - Step 126: {'lr': 3.15e-05, 'samples': 65024, 'steps': 126, 'loss/train': 8.971589088439941} 02/24/2022 01:44:15 - INFO - codeparrot_training - Step 127: {'lr': 3.175e-05, 'samples': 65536, 'steps': 127, 'loss/train': 9.091878890991211} 02/24/2022 01:44:19 - INFO - codeparrot_training - Step 128: {'lr': 3.2e-05, 'samples': 66048, 'steps': 128, 'loss/train': 8.642971992492676} 02/24/2022 01:44:24 - INFO - codeparrot_training - Step 129: {'lr': 3.2250000000000005e-05, 'samples': 66560, 'steps': 129, 'loss/train': 9.095447540283203} 02/24/2022 01:44:27 - INFO - codeparrot_training - Step 130: {'lr': 3.2500000000000004e-05, 'samples': 67072, 'steps': 130, 'loss/train': 8.557703018188477} 02/24/2022 01:44:33 - INFO - codeparrot_training - Step 131: {'lr': 3.275e-05, 'samples': 67584, 'steps': 131, 'loss/train': 8.567814826965332} 02/24/2022 01:44:38 - INFO - codeparrot_training - Step 132: {'lr': 3.3e-05, 'samples': 68096, 'steps': 132, 'loss/train': 8.998651504516602} 02/24/2022 01:44:42 - INFO - codeparrot_training - Step 133: {'lr': 3.325e-05, 'samples': 68608, 'steps': 133, 'loss/train': 8.84678840637207} 02/24/2022 01:44:49 - INFO - codeparrot_training - Step 134: {'lr': 3.35e-05, 'samples': 69120, 'steps': 134, 'loss/train': 8.768152236938477} 02/24/2022 01:44:52 - INFO - codeparrot_training - Step 135: {'lr': 3.375e-05, 'samples': 69632, 'steps': 135, 'loss/train': 8.87886905670166} 02/24/2022 01:44:58 - INFO - codeparrot_training - Step 136: {'lr': 3.4000000000000007e-05, 'samples': 70144, 'steps': 136, 'loss/train': 9.644051551818848} 02/24/2022 01:45:01 - INFO - codeparrot_training - Step 137: {'lr': 3.4250000000000006e-05, 'samples': 70656, 'steps': 137, 'loss/train': 9.087210655212402} 02/24/2022 01:45:06 - INFO - codeparrot_training - Step 138: {'lr': 3.4500000000000005e-05, 'samples': 71168, 'steps': 138, 'loss/train': 9.06340503692627} 02/24/2022 01:45:10 - INFO - codeparrot_training - Step 139: {'lr': 3.4750000000000004e-05, 'samples': 71680, 'steps': 139, 'loss/train': 8.886314392089844} 02/24/2022 01:45:15 - INFO - codeparrot_training - Step 140: {'lr': 3.5000000000000004e-05, 'samples': 72192, 'steps': 140, 'loss/train': 9.091540336608887} 02/24/2022 01:45:19 - INFO - codeparrot_training - Step 141: {'lr': 3.5249999999999996e-05, 'samples': 72704, 'steps': 141, 'loss/train': 9.188919067382812} 02/24/2022 01:45:24 - INFO - codeparrot_training - Step 142: {'lr': 3.5499999999999996e-05, 'samples': 73216, 'steps': 142, 'loss/train': 9.140847206115723} 02/24/2022 01:45:28 - INFO - codeparrot_training - Step 143: {'lr': 3.5749999999999995e-05, 'samples': 73728, 'steps': 143, 'loss/train': 8.462065696716309} 02/24/2022 01:45:33 - INFO - codeparrot_training - Step 144: {'lr': 3.6e-05, 'samples': 74240, 'steps': 144, 'loss/train': 8.03787899017334} 02/24/2022 01:45:37 - INFO - codeparrot_training - Step 145: {'lr': 3.625e-05, 'samples': 74752, 'steps': 145, 'loss/train': 8.804536819458008} 02/24/2022 01:45:43 - INFO - codeparrot_training - Step 146: {'lr': 3.65e-05, 'samples': 75264, 'steps': 146, 'loss/train': 9.822867393493652} 02/24/2022 01:45:46 - INFO - codeparrot_training - Step 147: {'lr': 3.675e-05, 'samples': 75776, 'steps': 147, 'loss/train': 9.119757652282715} 02/24/2022 01:45:52 - INFO - codeparrot_training - Step 148: {'lr': 3.7e-05, 'samples': 76288, 'steps': 148, 'loss/train': 8.35252571105957} 02/24/2022 01:45:55 - INFO - codeparrot_training - Step 149: {'lr': 3.725e-05, 'samples': 76800, 'steps': 149, 'loss/train': 8.668418884277344} 02/24/2022 01:46:01 - INFO - codeparrot_training - Step 150: {'lr': 3.75e-05, 'samples': 77312, 'steps': 150, 'loss/train': 8.233305931091309} 02/24/2022 01:46:04 - INFO - codeparrot_training - Step 151: {'lr': 3.775e-05, 'samples': 77824, 'steps': 151, 'loss/train': 9.180137634277344} 02/24/2022 01:46:10 - INFO - codeparrot_training - Step 152: {'lr': 3.8e-05, 'samples': 78336, 'steps': 152, 'loss/train': 8.461322784423828} 02/24/2022 01:46:13 - INFO - codeparrot_training - Step 153: {'lr': 3.825e-05, 'samples': 78848, 'steps': 153, 'loss/train': 8.905062675476074} 02/24/2022 01:46:19 - INFO - codeparrot_training - Step 154: {'lr': 3.85e-05, 'samples': 79360, 'steps': 154, 'loss/train': 8.642675399780273} 02/24/2022 01:46:23 - INFO - codeparrot_training - Step 155: {'lr': 3.875e-05, 'samples': 79872, 'steps': 155, 'loss/train': 8.070643424987793} 02/24/2022 01:46:28 - INFO - codeparrot_training - Step 156: {'lr': 3.9e-05, 'samples': 80384, 'steps': 156, 'loss/train': 5.581755638122559} 02/24/2022 01:46:32 - INFO - codeparrot_training - Step 157: {'lr': 3.925e-05, 'samples': 80896, 'steps': 157, 'loss/train': 8.737156867980957} 02/24/2022 01:46:37 - INFO - codeparrot_training - Step 158: {'lr': 3.95e-05, 'samples': 81408, 'steps': 158, 'loss/train': 8.90294075012207} 02/24/2022 01:46:41 - INFO - codeparrot_training - Step 159: {'lr': 3.9750000000000004e-05, 'samples': 81920, 'steps': 159, 'loss/train': 8.663994789123535} 02/24/2022 01:46:46 - INFO - codeparrot_training - Step 160: {'lr': 4e-05, 'samples': 82432, 'steps': 160, 'loss/train': 8.823864936828613} 02/24/2022 01:46:50 - INFO - codeparrot_training - Step 161: {'lr': 4.025e-05, 'samples': 82944, 'steps': 161, 'loss/train': 8.195619583129883} 02/24/2022 01:46:55 - INFO - codeparrot_training - Step 162: {'lr': 4.05e-05, 'samples': 83456, 'steps': 162, 'loss/train': 8.817365646362305} 02/24/2022 01:46:58 - INFO - codeparrot_training - Step 163: {'lr': 4.075e-05, 'samples': 83968, 'steps': 163, 'loss/train': 9.630975723266602} 02/24/2022 01:47:04 - INFO - codeparrot_training - Step 164: {'lr': 4.1e-05, 'samples': 84480, 'steps': 164, 'loss/train': 8.188645362854004} 02/24/2022 01:47:07 - INFO - codeparrot_training - Step 165: {'lr': 4.125e-05, 'samples': 84992, 'steps': 165, 'loss/train': 10.159860610961914} 02/24/2022 01:47:13 - INFO - codeparrot_training - Step 166: {'lr': 4.1500000000000006e-05, 'samples': 85504, 'steps': 166, 'loss/train': 8.092951774597168} 02/24/2022 01:47:16 - INFO - codeparrot_training - Step 167: {'lr': 4.1750000000000005e-05, 'samples': 86016, 'steps': 167, 'loss/train': 9.026200294494629} 02/24/2022 01:47:22 - INFO - codeparrot_training - Step 168: {'lr': 4.2000000000000004e-05, 'samples': 86528, 'steps': 168, 'loss/train': 8.288607597351074} 02/24/2022 01:47:25 - INFO - codeparrot_training - Step 169: {'lr': 4.2250000000000004e-05, 'samples': 87040, 'steps': 169, 'loss/train': 7.285762786865234} 02/24/2022 01:47:31 - INFO - codeparrot_training - Step 170: {'lr': 4.25e-05, 'samples': 87552, 'steps': 170, 'loss/train': 8.039697647094727} 02/24/2022 01:47:35 - INFO - codeparrot_training - Step 171: {'lr': 4.275e-05, 'samples': 88064, 'steps': 171, 'loss/train': 8.41202163696289} 02/24/2022 01:47:40 - INFO - codeparrot_training - Step 172: {'lr': 4.2999999999999995e-05, 'samples': 88576, 'steps': 172, 'loss/train': 8.135007858276367} 02/24/2022 01:47:44 - INFO - codeparrot_training - Step 173: {'lr': 4.325e-05, 'samples': 89088, 'steps': 173, 'loss/train': 8.630109786987305} 02/24/2022 01:47:49 - INFO - codeparrot_training - Step 174: {'lr': 4.35e-05, 'samples': 89600, 'steps': 174, 'loss/train': 8.232574462890625} 02/24/2022 01:47:52 - INFO - codeparrot_training - Step 175: {'lr': 4.375e-05, 'samples': 90112, 'steps': 175, 'loss/train': 7.799593925476074} 02/24/2022 01:47:58 - INFO - codeparrot_training - Step 176: {'lr': 4.4e-05, 'samples': 90624, 'steps': 176, 'loss/train': 8.399349212646484} 02/24/2022 01:48:01 - INFO - codeparrot_training - Step 177: {'lr': 4.425e-05, 'samples': 91136, 'steps': 177, 'loss/train': 8.36412239074707} 02/24/2022 01:48:07 - INFO - codeparrot_training - Step 178: {'lr': 4.45e-05, 'samples': 91648, 'steps': 178, 'loss/train': 9.112398147583008} 02/24/2022 01:48:10 - INFO - codeparrot_training - Step 179: {'lr': 4.475e-05, 'samples': 92160, 'steps': 179, 'loss/train': 7.641449928283691} 02/24/2022 01:48:16 - INFO - codeparrot_training - Step 180: {'lr': 4.4999999999999996e-05, 'samples': 92672, 'steps': 180, 'loss/train': 9.119322776794434} 02/24/2022 01:48:19 - INFO - codeparrot_training - Step 181: {'lr': 4.525e-05, 'samples': 93184, 'steps': 181, 'loss/train': 8.35004711151123} 02/24/2022 01:48:25 - INFO - codeparrot_training - Step 182: {'lr': 4.55e-05, 'samples': 93696, 'steps': 182, 'loss/train': 8.301406860351562} 02/24/2022 01:48:29 - INFO - codeparrot_training - Step 183: {'lr': 4.575e-05, 'samples': 94208, 'steps': 183, 'loss/train': 8.395172119140625} 02/24/2022 01:48:34 - INFO - codeparrot_training - Step 184: {'lr': 4.6e-05, 'samples': 94720, 'steps': 184, 'loss/train': 10.358185768127441} 02/24/2022 01:48:38 - INFO - codeparrot_training - Step 185: {'lr': 4.625e-05, 'samples': 95232, 'steps': 185, 'loss/train': 7.754116535186768} 02/24/2022 01:48:43 - INFO - codeparrot_training - Step 186: {'lr': 4.65e-05, 'samples': 95744, 'steps': 186, 'loss/train': 8.634472846984863} 02/24/2022 01:48:46 - INFO - codeparrot_training - Step 187: {'lr': 4.675e-05, 'samples': 96256, 'steps': 187, 'loss/train': 9.413156509399414} 02/24/2022 01:48:52 - INFO - codeparrot_training - Step 188: {'lr': 4.7000000000000004e-05, 'samples': 96768, 'steps': 188, 'loss/train': 8.477133750915527} 02/24/2022 01:48:55 - INFO - codeparrot_training - Step 189: {'lr': 4.725e-05, 'samples': 97280, 'steps': 189, 'loss/train': 8.565733909606934} 02/24/2022 01:49:01 - INFO - codeparrot_training - Step 190: {'lr': 4.75e-05, 'samples': 97792, 'steps': 190, 'loss/train': 8.980598449707031} 02/24/2022 01:49:05 - INFO - codeparrot_training - Step 191: {'lr': 4.775e-05, 'samples': 98304, 'steps': 191, 'loss/train': 8.175108909606934} 02/24/2022 01:49:11 - INFO - codeparrot_training - Step 192: {'lr': 4.8e-05, 'samples': 98816, 'steps': 192, 'loss/train': 7.164796829223633} 02/24/2022 01:49:14 - INFO - codeparrot_training - Step 193: {'lr': 4.825e-05, 'samples': 99328, 'steps': 193, 'loss/train': 7.408548831939697} 02/24/2022 01:49:20 - INFO - codeparrot_training - Step 194: {'lr': 4.85e-05, 'samples': 99840, 'steps': 194, 'loss/train': 8.422624588012695} 02/24/2022 01:49:23 - INFO - codeparrot_training - Step 195: {'lr': 4.8750000000000006e-05, 'samples': 100352, 'steps': 195, 'loss/train': 7.857550621032715} 02/24/2022 01:49:29 - INFO - codeparrot_training - Step 196: {'lr': 4.9000000000000005e-05, 'samples': 100864, 'steps': 196, 'loss/train': 7.618260860443115} 02/24/2022 01:49:32 - INFO - codeparrot_training - Step 197: {'lr': 4.9250000000000004e-05, 'samples': 101376, 'steps': 197, 'loss/train': 8.605622291564941} 02/24/2022 01:49:37 - INFO - codeparrot_training - Step 198: {'lr': 4.9500000000000004e-05, 'samples': 101888, 'steps': 198, 'loss/train': 7.873624801635742} 02/24/2022 01:49:41 - INFO - codeparrot_training - Step 199: {'lr': 4.975e-05, 'samples': 102400, 'steps': 199, 'loss/train': 9.396079063415527} 02/24/2022 01:49:46 - INFO - codeparrot_training - Step 200: {'lr': 5e-05, 'samples': 102912, 'steps': 200, 'loss/train': 9.744832992553711} 02/24/2022 01:49:50 - INFO - codeparrot_training - Step 201: {'lr': 5.025e-05, 'samples': 103424, 'steps': 201, 'loss/train': 8.990636825561523} 02/24/2022 01:49:56 - INFO - codeparrot_training - Step 202: {'lr': 5.05e-05, 'samples': 103936, 'steps': 202, 'loss/train': 8.19791030883789} 02/24/2022 01:49:59 - INFO - codeparrot_training - Step 203: {'lr': 5.075000000000001e-05, 'samples': 104448, 'steps': 203, 'loss/train': 8.748708724975586} 02/24/2022 01:50:05 - INFO - codeparrot_training - Step 204: {'lr': 5.1e-05, 'samples': 104960, 'steps': 204, 'loss/train': 10.17846393585205} 02/24/2022 01:50:08 - INFO - codeparrot_training - Step 205: {'lr': 5.125e-05, 'samples': 105472, 'steps': 205, 'loss/train': 8.256689071655273} 02/24/2022 01:50:14 - INFO - codeparrot_training - Step 206: {'lr': 5.15e-05, 'samples': 105984, 'steps': 206, 'loss/train': 8.524040222167969} 02/24/2022 01:50:17 - INFO - codeparrot_training - Step 207: {'lr': 5.175e-05, 'samples': 106496, 'steps': 207, 'loss/train': 8.013733863830566} 02/24/2022 01:50:23 - INFO - codeparrot_training - Step 208: {'lr': 5.2e-05, 'samples': 107008, 'steps': 208, 'loss/train': 8.974383354187012} 02/24/2022 01:50:26 - INFO - codeparrot_training - Step 209: {'lr': 5.2249999999999996e-05, 'samples': 107520, 'steps': 209, 'loss/train': 7.138608932495117} 02/24/2022 01:50:32 - INFO - codeparrot_training - Step 210: {'lr': 5.25e-05, 'samples': 108032, 'steps': 210, 'loss/train': 8.761432647705078} 02/24/2022 01:50:35 - INFO - codeparrot_training - Step 211: {'lr': 5.275e-05, 'samples': 108544, 'steps': 211, 'loss/train': 8.35822868347168} 02/24/2022 01:50:41 - INFO - codeparrot_training - Step 212: {'lr': 5.3e-05, 'samples': 109056, 'steps': 212, 'loss/train': 7.769471168518066} 02/24/2022 01:50:44 - INFO - codeparrot_training - Step 213: {'lr': 5.325e-05, 'samples': 109568, 'steps': 213, 'loss/train': 7.3015055656433105} 02/24/2022 01:50:50 - INFO - codeparrot_training - Step 214: {'lr': 5.35e-05, 'samples': 110080, 'steps': 214, 'loss/train': 8.458576202392578} 02/24/2022 01:50:55 - INFO - codeparrot_training - Step 215: {'lr': 5.375e-05, 'samples': 110592, 'steps': 215, 'loss/train': 7.801426410675049} 02/24/2022 01:50:59 - INFO - codeparrot_training - Step 216: {'lr': 5.4e-05, 'samples': 111104, 'steps': 216, 'loss/train': 8.24003791809082} 02/24/2022 01:51:04 - INFO - codeparrot_training - Step 217: {'lr': 5.4250000000000004e-05, 'samples': 111616, 'steps': 217, 'loss/train': 8.818034172058105} 02/24/2022 01:51:08 - INFO - codeparrot_training - Step 218: {'lr': 5.45e-05, 'samples': 112128, 'steps': 218, 'loss/train': 7.830726146697998} 02/24/2022 01:51:13 - INFO - codeparrot_training - Step 219: {'lr': 5.475e-05, 'samples': 112640, 'steps': 219, 'loss/train': 8.32111644744873} 02/24/2022 01:51:17 - INFO - codeparrot_training - Step 220: {'lr': 5.5e-05, 'samples': 113152, 'steps': 220, 'loss/train': 7.4934258460998535} 02/24/2022 01:51:23 - INFO - codeparrot_training - Step 221: {'lr': 5.525e-05, 'samples': 113664, 'steps': 221, 'loss/train': 7.56221866607666} 02/24/2022 01:51:26 - INFO - codeparrot_training - Step 222: {'lr': 5.55e-05, 'samples': 114176, 'steps': 222, 'loss/train': 7.5190558433532715} 02/24/2022 01:51:31 - INFO - codeparrot_training - Step 223: {'lr': 5.575e-05, 'samples': 114688, 'steps': 223, 'loss/train': 7.6769304275512695} 02/24/2022 01:51:35 - INFO - codeparrot_training - Step 224: {'lr': 5.6e-05, 'samples': 115200, 'steps': 224, 'loss/train': 8.480721473693848} 02/24/2022 01:51:40 - INFO - codeparrot_training - Step 225: {'lr': 5.6250000000000005e-05, 'samples': 115712, 'steps': 225, 'loss/train': 7.005849838256836} 02/24/2022 01:51:44 - INFO - codeparrot_training - Step 226: {'lr': 5.6500000000000005e-05, 'samples': 116224, 'steps': 226, 'loss/train': 7.340949058532715} 02/24/2022 01:51:49 - INFO - codeparrot_training - Step 227: {'lr': 5.6750000000000004e-05, 'samples': 116736, 'steps': 227, 'loss/train': 8.233819961547852} 02/24/2022 01:51:53 - INFO - codeparrot_training - Step 228: {'lr': 5.7e-05, 'samples': 117248, 'steps': 228, 'loss/train': 9.251286506652832} 02/24/2022 01:51:58 - INFO - codeparrot_training - Step 229: {'lr': 5.725e-05, 'samples': 117760, 'steps': 229, 'loss/train': 7.45113468170166} 02/24/2022 01:52:01 - INFO - codeparrot_training - Step 230: {'lr': 5.75e-05, 'samples': 118272, 'steps': 230, 'loss/train': 7.624856472015381} 02/24/2022 01:52:07 - INFO - codeparrot_training - Step 231: {'lr': 5.775e-05, 'samples': 118784, 'steps': 231, 'loss/train': 8.095080375671387} 02/24/2022 01:52:11 - INFO - codeparrot_training - Step 232: {'lr': 5.800000000000001e-05, 'samples': 119296, 'steps': 232, 'loss/train': 6.90823221206665} 02/24/2022 01:52:16 - INFO - codeparrot_training - Step 233: {'lr': 5.8250000000000006e-05, 'samples': 119808, 'steps': 233, 'loss/train': 8.450687408447266} 02/24/2022 01:52:20 - INFO - codeparrot_training - Step 234: {'lr': 5.8500000000000006e-05, 'samples': 120320, 'steps': 234, 'loss/train': 7.578708648681641} 02/24/2022 01:52:25 - INFO - codeparrot_training - Step 235: {'lr': 5.875e-05, 'samples': 120832, 'steps': 235, 'loss/train': 8.322768211364746} 02/24/2022 01:52:29 - INFO - codeparrot_training - Step 236: {'lr': 5.9e-05, 'samples': 121344, 'steps': 236, 'loss/train': 7.880550861358643} 02/24/2022 01:52:35 - INFO - codeparrot_training - Step 237: {'lr': 5.925e-05, 'samples': 121856, 'steps': 237, 'loss/train': 8.012900352478027} 02/24/2022 01:52:38 - INFO - codeparrot_training - Step 238: {'lr': 5.9499999999999996e-05, 'samples': 122368, 'steps': 238, 'loss/train': 8.338845252990723} 02/24/2022 01:52:44 - INFO - codeparrot_training - Step 239: {'lr': 5.9749999999999995e-05, 'samples': 122880, 'steps': 239, 'loss/train': 8.256518363952637} 02/24/2022 01:52:47 - INFO - codeparrot_training - Step 240: {'lr': 6e-05, 'samples': 123392, 'steps': 240, 'loss/train': 8.110925674438477} 02/24/2022 01:52:53 - INFO - codeparrot_training - Step 241: {'lr': 6.025e-05, 'samples': 123904, 'steps': 241, 'loss/train': 8.951800346374512} 02/24/2022 01:52:56 - INFO - codeparrot_training - Step 242: {'lr': 6.05e-05, 'samples': 124416, 'steps': 242, 'loss/train': 8.23705768585205} 02/24/2022 01:53:02 - INFO - codeparrot_training - Step 243: {'lr': 6.075e-05, 'samples': 124928, 'steps': 243, 'loss/train': 7.9062042236328125} 02/24/2022 01:53:05 - INFO - codeparrot_training - Step 244: {'lr': 6.1e-05, 'samples': 125440, 'steps': 244, 'loss/train': 10.34733772277832} 02/24/2022 01:53:10 - INFO - codeparrot_training - Step 245: {'lr': 6.125e-05, 'samples': 125952, 'steps': 245, 'loss/train': 8.160472869873047} 02/24/2022 01:53:14 - INFO - codeparrot_training - Step 246: {'lr': 6.15e-05, 'samples': 126464, 'steps': 246, 'loss/train': 8.468631744384766} 02/24/2022 01:53:20 - INFO - codeparrot_training - Step 247: {'lr': 6.175e-05, 'samples': 126976, 'steps': 247, 'loss/train': 8.497027397155762} 02/24/2022 01:53:23 - INFO - codeparrot_training - Step 248: {'lr': 6.2e-05, 'samples': 127488, 'steps': 248, 'loss/train': 7.5156755447387695} 02/24/2022 01:53:29 - INFO - codeparrot_training - Step 249: {'lr': 6.225e-05, 'samples': 128000, 'steps': 249, 'loss/train': 8.726844787597656} 02/24/2022 01:53:32 - INFO - codeparrot_training - Step 250: {'lr': 6.25e-05, 'samples': 128512, 'steps': 250, 'loss/train': 8.230809211730957} 02/24/2022 01:53:38 - INFO - codeparrot_training - Step 251: {'lr': 6.275000000000001e-05, 'samples': 129024, 'steps': 251, 'loss/train': 8.839653015136719} 02/24/2022 01:53:41 - INFO - codeparrot_training - Step 252: {'lr': 6.3e-05, 'samples': 129536, 'steps': 252, 'loss/train': 7.990950584411621} 02/24/2022 01:53:47 - INFO - codeparrot_training - Step 253: {'lr': 6.325e-05, 'samples': 130048, 'steps': 253, 'loss/train': 7.565497398376465} 02/24/2022 01:53:50 - INFO - codeparrot_training - Step 254: {'lr': 6.35e-05, 'samples': 130560, 'steps': 254, 'loss/train': 7.676276206970215} 02/24/2022 01:53:55 - INFO - codeparrot_training - Step 255: {'lr': 6.375e-05, 'samples': 131072, 'steps': 255, 'loss/train': 8.033483505249023} 02/24/2022 01:53:59 - INFO - codeparrot_training - Step 256: {'lr': 6.4e-05, 'samples': 131584, 'steps': 256, 'loss/train': 8.045367240905762} 02/24/2022 01:54:05 - INFO - codeparrot_training - Step 257: {'lr': 6.425e-05, 'samples': 132096, 'steps': 257, 'loss/train': 8.210289001464844} 02/24/2022 01:54:08 - INFO - codeparrot_training - Step 258: {'lr': 6.450000000000001e-05, 'samples': 132608, 'steps': 258, 'loss/train': 7.078313827514648} 02/24/2022 01:54:14 - INFO - codeparrot_training - Step 259: {'lr': 6.475e-05, 'samples': 133120, 'steps': 259, 'loss/train': 9.826390266418457} 02/24/2022 01:54:18 - INFO - codeparrot_training - Step 260: {'lr': 6.500000000000001e-05, 'samples': 133632, 'steps': 260, 'loss/train': 7.573724746704102} 02/24/2022 01:54:23 - INFO - codeparrot_training - Step 261: {'lr': 6.525e-05, 'samples': 134144, 'steps': 261, 'loss/train': 8.298879623413086} 02/24/2022 01:54:26 - INFO - codeparrot_training - Step 262: {'lr': 6.55e-05, 'samples': 134656, 'steps': 262, 'loss/train': 8.041802406311035} 02/24/2022 01:54:32 - INFO - codeparrot_training - Step 263: {'lr': 6.575e-05, 'samples': 135168, 'steps': 263, 'loss/train': 8.012332916259766} 02/24/2022 01:54:35 - INFO - codeparrot_training - Step 264: {'lr': 6.6e-05, 'samples': 135680, 'steps': 264, 'loss/train': 7.702607154846191} 02/24/2022 01:54:41 - INFO - codeparrot_training - Step 265: {'lr': 6.625000000000001e-05, 'samples': 136192, 'steps': 265, 'loss/train': 8.194743156433105} 02/24/2022 01:54:44 - INFO - codeparrot_training - Step 266: {'lr': 6.65e-05, 'samples': 136704, 'steps': 266, 'loss/train': 8.723750114440918} 02/24/2022 01:54:50 - INFO - codeparrot_training - Step 267: {'lr': 6.675000000000001e-05, 'samples': 137216, 'steps': 267, 'loss/train': 7.491954326629639} 02/24/2022 01:54:54 - INFO - codeparrot_training - Step 268: {'lr': 6.7e-05, 'samples': 137728, 'steps': 268, 'loss/train': 8.32316780090332} 02/24/2022 01:54:59 - INFO - codeparrot_training - Step 269: {'lr': 6.725000000000001e-05, 'samples': 138240, 'steps': 269, 'loss/train': 7.601045608520508} 02/24/2022 01:55:03 - INFO - codeparrot_training - Step 270: {'lr': 6.75e-05, 'samples': 138752, 'steps': 270, 'loss/train': 8.582389831542969} 02/24/2022 01:55:08 - INFO - codeparrot_training - Step 271: {'lr': 6.775000000000001e-05, 'samples': 139264, 'steps': 271, 'loss/train': 7.930277347564697} 02/24/2022 01:55:12 - INFO - codeparrot_training - Step 272: {'lr': 6.800000000000001e-05, 'samples': 139776, 'steps': 272, 'loss/train': 7.568014621734619} 02/24/2022 01:55:18 - INFO - codeparrot_training - Step 273: {'lr': 6.825e-05, 'samples': 140288, 'steps': 273, 'loss/train': 7.254758358001709} 02/24/2022 01:55:21 - INFO - codeparrot_training - Step 274: {'lr': 6.850000000000001e-05, 'samples': 140800, 'steps': 274, 'loss/train': 7.01399564743042} 02/24/2022 01:55:26 - INFO - codeparrot_training - Step 275: {'lr': 6.875e-05, 'samples': 141312, 'steps': 275, 'loss/train': 7.368566989898682} 02/24/2022 01:55:30 - INFO - codeparrot_training - Step 276: {'lr': 6.900000000000001e-05, 'samples': 141824, 'steps': 276, 'loss/train': 7.3699870109558105} 02/24/2022 01:55:35 - INFO - codeparrot_training - Step 277: {'lr': 6.925e-05, 'samples': 142336, 'steps': 277, 'loss/train': 7.630679130554199} 02/24/2022 01:55:39 - INFO - codeparrot_training - Step 278: {'lr': 6.950000000000001e-05, 'samples': 142848, 'steps': 278, 'loss/train': 8.27759838104248} 02/24/2022 01:55:44 - INFO - codeparrot_training - Step 279: {'lr': 6.975e-05, 'samples': 143360, 'steps': 279, 'loss/train': 8.460628509521484} 02/24/2022 01:55:48 - INFO - codeparrot_training - Step 280: {'lr': 7.000000000000001e-05, 'samples': 143872, 'steps': 280, 'loss/train': 8.191814422607422} 02/24/2022 01:55:53 - INFO - codeparrot_training - Step 281: {'lr': 7.025000000000001e-05, 'samples': 144384, 'steps': 281, 'loss/train': 7.863226413726807} 02/24/2022 01:55:56 - INFO - codeparrot_training - Step 282: {'lr': 7.049999999999999e-05, 'samples': 144896, 'steps': 282, 'loss/train': 6.898287296295166} 02/24/2022 01:56:03 - INFO - codeparrot_training - Step 283: {'lr': 7.075e-05, 'samples': 145408, 'steps': 283, 'loss/train': 7.9318132400512695} 02/24/2022 01:56:06 - INFO - codeparrot_training - Step 284: {'lr': 7.099999999999999e-05, 'samples': 145920, 'steps': 284, 'loss/train': 8.058138847351074} 02/24/2022 01:56:12 - INFO - codeparrot_training - Step 285: {'lr': 7.125e-05, 'samples': 146432, 'steps': 285, 'loss/train': 7.8580546379089355} 02/24/2022 01:56:15 - INFO - codeparrot_training - Step 286: {'lr': 7.149999999999999e-05, 'samples': 146944, 'steps': 286, 'loss/train': 7.834128379821777} 02/24/2022 01:56:20 - INFO - codeparrot_training - Step 287: {'lr': 7.175e-05, 'samples': 147456, 'steps': 287, 'loss/train': 7.871674060821533} 02/24/2022 01:56:24 - INFO - codeparrot_training - Step 288: {'lr': 7.2e-05, 'samples': 147968, 'steps': 288, 'loss/train': 8.25328254699707} 02/24/2022 01:56:29 - INFO - codeparrot_training - Step 289: {'lr': 7.225e-05, 'samples': 148480, 'steps': 289, 'loss/train': 7.208605766296387} 02/24/2022 01:56:33 - INFO - codeparrot_training - Step 290: {'lr': 7.25e-05, 'samples': 148992, 'steps': 290, 'loss/train': 5.396283149719238} 02/24/2022 01:56:38 - INFO - codeparrot_training - Step 291: {'lr': 7.274999999999999e-05, 'samples': 149504, 'steps': 291, 'loss/train': 7.4003190994262695} 02/24/2022 01:56:42 - INFO - codeparrot_training - Step 292: {'lr': 7.3e-05, 'samples': 150016, 'steps': 292, 'loss/train': 8.524874687194824} 02/24/2022 01:56:48 - INFO - codeparrot_training - Step 293: {'lr': 7.324999999999999e-05, 'samples': 150528, 'steps': 293, 'loss/train': 5.260894775390625} 02/24/2022 01:56:52 - INFO - codeparrot_training - Step 294: {'lr': 7.35e-05, 'samples': 151040, 'steps': 294, 'loss/train': 7.498438835144043} 02/24/2022 01:56:57 - INFO - codeparrot_training - Step 295: {'lr': 7.375e-05, 'samples': 151552, 'steps': 295, 'loss/train': 7.932528972625732} 02/24/2022 01:57:00 - INFO - codeparrot_training - Step 296: {'lr': 7.4e-05, 'samples': 152064, 'steps': 296, 'loss/train': 7.992652416229248} 02/24/2022 01:57:06 - INFO - codeparrot_training - Step 297: {'lr': 7.425e-05, 'samples': 152576, 'steps': 297, 'loss/train': 7.58806037902832} 02/24/2022 01:57:09 - INFO - codeparrot_training - Step 298: {'lr': 7.45e-05, 'samples': 153088, 'steps': 298, 'loss/train': 8.329551696777344} 02/24/2022 01:57:15 - INFO - codeparrot_training - Step 299: {'lr': 7.475e-05, 'samples': 153600, 'steps': 299, 'loss/train': 7.876159191131592} 02/24/2022 01:57:19 - INFO - codeparrot_training - Step 300: {'lr': 7.5e-05, 'samples': 154112, 'steps': 300, 'loss/train': 8.205613136291504} 02/24/2022 01:57:24 - INFO - codeparrot_training - Step 301: {'lr': 7.525e-05, 'samples': 154624, 'steps': 301, 'loss/train': 8.59995174407959} 02/24/2022 01:57:28 - INFO - codeparrot_training - Step 302: {'lr': 7.55e-05, 'samples': 155136, 'steps': 302, 'loss/train': 8.156671524047852} 02/24/2022 01:57:34 - INFO - codeparrot_training - Step 303: {'lr': 7.575e-05, 'samples': 155648, 'steps': 303, 'loss/train': 7.888469696044922} 02/24/2022 01:57:37 - INFO - codeparrot_training - Step 304: {'lr': 7.6e-05, 'samples': 156160, 'steps': 304, 'loss/train': 7.206390857696533} 02/24/2022 01:57:42 - INFO - codeparrot_training - Step 305: {'lr': 7.625e-05, 'samples': 156672, 'steps': 305, 'loss/train': 8.198074340820312} 02/24/2022 01:57:46 - INFO - codeparrot_training - Step 306: {'lr': 7.65e-05, 'samples': 157184, 'steps': 306, 'loss/train': 7.702149868011475} 02/24/2022 01:57:51 - INFO - codeparrot_training - Step 307: {'lr': 7.675e-05, 'samples': 157696, 'steps': 307, 'loss/train': 8.235662460327148} 02/24/2022 01:57:55 - INFO - codeparrot_training - Step 308: {'lr': 7.7e-05, 'samples': 158208, 'steps': 308, 'loss/train': 8.55713939666748} 02/24/2022 01:58:00 - INFO - codeparrot_training - Step 309: {'lr': 7.725000000000001e-05, 'samples': 158720, 'steps': 309, 'loss/train': 7.453092098236084} 02/24/2022 01:58:04 - INFO - codeparrot_training - Step 310: {'lr': 7.75e-05, 'samples': 159232, 'steps': 310, 'loss/train': 7.088987350463867} 02/24/2022 01:58:09 - INFO - codeparrot_training - Step 311: {'lr': 7.775e-05, 'samples': 159744, 'steps': 311, 'loss/train': 5.286121368408203} 02/24/2022 01:58:13 - INFO - codeparrot_training - Step 312: {'lr': 7.8e-05, 'samples': 160256, 'steps': 312, 'loss/train': 7.3425517082214355} 02/24/2022 01:58:19 - INFO - codeparrot_training - Step 313: {'lr': 7.825e-05, 'samples': 160768, 'steps': 313, 'loss/train': 8.444873809814453} 02/24/2022 01:58:25 - INFO - codeparrot_training - Step 314: {'lr': 7.85e-05, 'samples': 161280, 'steps': 314, 'loss/train': 9.051511764526367} 02/24/2022 01:58:28 - INFO - codeparrot_training - Step 315: {'lr': 7.875e-05, 'samples': 161792, 'steps': 315, 'loss/train': 8.298068046569824} 02/24/2022 01:58:32 - INFO - codeparrot_training - Step 316: {'lr': 7.9e-05, 'samples': 162304, 'steps': 316, 'loss/train': 7.379024028778076} 02/24/2022 01:58:37 - INFO - codeparrot_training - Step 317: {'lr': 7.925e-05, 'samples': 162816, 'steps': 317, 'loss/train': 8.077301979064941} 02/24/2022 01:58:41 - INFO - codeparrot_training - Step 318: {'lr': 7.950000000000001e-05, 'samples': 163328, 'steps': 318, 'loss/train': 8.417896270751953} 02/24/2022 01:58:46 - INFO - codeparrot_training - Step 319: {'lr': 7.975e-05, 'samples': 163840, 'steps': 319, 'loss/train': 7.594058513641357} 02/24/2022 01:58:49 - INFO - codeparrot_training - Step 320: {'lr': 8e-05, 'samples': 164352, 'steps': 320, 'loss/train': 8.477936744689941} 02/24/2022 01:58:55 - INFO - codeparrot_training - Step 321: {'lr': 8.025e-05, 'samples': 164864, 'steps': 321, 'loss/train': 8.0873384475708} 02/24/2022 01:58:58 - INFO - codeparrot_training - Step 322: {'lr': 8.05e-05, 'samples': 165376, 'steps': 322, 'loss/train': 7.719216823577881} 02/24/2022 01:59:04 - INFO - codeparrot_training - Step 323: {'lr': 8.075e-05, 'samples': 165888, 'steps': 323, 'loss/train': 7.951709270477295} 02/24/2022 01:59:07 - INFO - codeparrot_training - Step 324: {'lr': 8.1e-05, 'samples': 166400, 'steps': 324, 'loss/train': 7.829857349395752} 02/24/2022 01:59:14 - INFO - codeparrot_training - Step 325: {'lr': 8.125000000000001e-05, 'samples': 166912, 'steps': 325, 'loss/train': 8.28011703491211} 02/24/2022 01:59:17 - INFO - codeparrot_training - Step 326: {'lr': 8.15e-05, 'samples': 167424, 'steps': 326, 'loss/train': 8.2847261428833} 02/24/2022 01:59:23 - INFO - codeparrot_training - Step 327: {'lr': 8.175000000000001e-05, 'samples': 167936, 'steps': 327, 'loss/train': 7.977908611297607} 02/24/2022 01:59:26 - INFO - codeparrot_training - Step 328: {'lr': 8.2e-05, 'samples': 168448, 'steps': 328, 'loss/train': 6.0809855461120605} 02/24/2022 01:59:31 - INFO - codeparrot_training - Step 329: {'lr': 8.225000000000001e-05, 'samples': 168960, 'steps': 329, 'loss/train': 7.6018290519714355} 02/24/2022 01:59:35 - INFO - codeparrot_training - Step 330: {'lr': 8.25e-05, 'samples': 169472, 'steps': 330, 'loss/train': 7.403947353363037} 02/24/2022 01:59:40 - INFO - codeparrot_training - Step 331: {'lr': 8.275e-05, 'samples': 169984, 'steps': 331, 'loss/train': 8.224590301513672} 02/24/2022 01:59:44 - INFO - codeparrot_training - Step 332: {'lr': 8.300000000000001e-05, 'samples': 170496, 'steps': 332, 'loss/train': 7.179081916809082} 02/24/2022 01:59:49 - INFO - codeparrot_training - Step 333: {'lr': 8.325e-05, 'samples': 171008, 'steps': 333, 'loss/train': 7.664137363433838} 02/24/2022 01:59:55 - INFO - codeparrot_training - Step 334: {'lr': 8.350000000000001e-05, 'samples': 171520, 'steps': 334, 'loss/train': 6.750515937805176} 02/24/2022 01:59:58 - INFO - codeparrot_training - Step 335: {'lr': 8.375e-05, 'samples': 172032, 'steps': 335, 'loss/train': 7.544463634490967} 02/24/2022 02:00:04 - INFO - codeparrot_training - Step 336: {'lr': 8.400000000000001e-05, 'samples': 172544, 'steps': 336, 'loss/train': 6.866092205047607} 02/24/2022 02:00:07 - INFO - codeparrot_training - Step 337: {'lr': 8.425e-05, 'samples': 173056, 'steps': 337, 'loss/train': 7.727400302886963} 02/24/2022 02:00:13 - INFO - codeparrot_training - Step 338: {'lr': 8.450000000000001e-05, 'samples': 173568, 'steps': 338, 'loss/train': 7.797516345977783} 02/24/2022 02:00:16 - INFO - codeparrot_training - Step 339: {'lr': 8.475000000000001e-05, 'samples': 174080, 'steps': 339, 'loss/train': 7.595932960510254} 02/24/2022 02:00:22 - INFO - codeparrot_training - Step 340: {'lr': 8.5e-05, 'samples': 174592, 'steps': 340, 'loss/train': 7.032388687133789} 02/24/2022 02:00:25 - INFO - codeparrot_training - Step 341: {'lr': 8.525000000000001e-05, 'samples': 175104, 'steps': 341, 'loss/train': 7.881145477294922} 02/24/2022 02:00:31 - INFO - codeparrot_training - Step 342: {'lr': 8.55e-05, 'samples': 175616, 'steps': 342, 'loss/train': 7.786242961883545} 02/24/2022 02:00:34 - INFO - codeparrot_training - Step 343: {'lr': 8.575000000000001e-05, 'samples': 176128, 'steps': 343, 'loss/train': 7.198973178863525} 02/24/2022 02:00:40 - INFO - codeparrot_training - Step 344: {'lr': 8.599999999999999e-05, 'samples': 176640, 'steps': 344, 'loss/train': 8.599320411682129} 02/24/2022 02:00:43 - INFO - codeparrot_training - Step 345: {'lr': 8.625e-05, 'samples': 177152, 'steps': 345, 'loss/train': 7.57554817199707} 02/24/2022 02:00:49 - INFO - codeparrot_training - Step 346: {'lr': 8.65e-05, 'samples': 177664, 'steps': 346, 'loss/train': 8.224437713623047} 02/24/2022 02:00:52 - INFO - codeparrot_training - Step 347: {'lr': 8.675e-05, 'samples': 178176, 'steps': 347, 'loss/train': 7.841677188873291} 02/24/2022 02:00:58 - INFO - codeparrot_training - Step 348: {'lr': 8.7e-05, 'samples': 178688, 'steps': 348, 'loss/train': 7.63005256652832} 02/24/2022 02:01:01 - INFO - codeparrot_training - Step 349: {'lr': 8.724999999999999e-05, 'samples': 179200, 'steps': 349, 'loss/train': 8.130577087402344} 02/24/2022 02:01:07 - INFO - codeparrot_training - Step 350: {'lr': 8.75e-05, 'samples': 179712, 'steps': 350, 'loss/train': 7.222663402557373} 02/24/2022 02:01:10 - INFO - codeparrot_training - Step 351: {'lr': 8.774999999999999e-05, 'samples': 180224, 'steps': 351, 'loss/train': 7.223997116088867} 02/24/2022 02:01:16 - INFO - codeparrot_training - Step 352: {'lr': 8.8e-05, 'samples': 180736, 'steps': 352, 'loss/train': 7.704534530639648} 02/24/2022 02:01:19 - INFO - codeparrot_training - Step 353: {'lr': 8.824999999999999e-05, 'samples': 181248, 'steps': 353, 'loss/train': 8.424806594848633} 02/24/2022 02:01:25 - INFO - codeparrot_training - Step 354: {'lr': 8.85e-05, 'samples': 181760, 'steps': 354, 'loss/train': 8.343356132507324} 02/24/2022 02:01:28 - INFO - codeparrot_training - Step 355: {'lr': 8.875e-05, 'samples': 182272, 'steps': 355, 'loss/train': 7.833737850189209} 02/24/2022 02:01:34 - INFO - codeparrot_training - Step 356: {'lr': 8.9e-05, 'samples': 182784, 'steps': 356, 'loss/train': 4.860916614532471} 02/24/2022 02:01:37 - INFO - codeparrot_training - Step 357: {'lr': 8.925e-05, 'samples': 183296, 'steps': 357, 'loss/train': 7.812404632568359} 02/24/2022 02:01:42 - INFO - codeparrot_training - Step 358: {'lr': 8.95e-05, 'samples': 183808, 'steps': 358, 'loss/train': 8.174966812133789} 02/24/2022 02:01:46 - INFO - codeparrot_training - Step 359: {'lr': 8.975e-05, 'samples': 184320, 'steps': 359, 'loss/train': 7.480123043060303} 02/24/2022 02:01:52 - INFO - codeparrot_training - Step 360: {'lr': 8.999999999999999e-05, 'samples': 184832, 'steps': 360, 'loss/train': 8.104669570922852} 02/24/2022 02:01:55 - INFO - codeparrot_training - Step 361: {'lr': 9.025e-05, 'samples': 185344, 'steps': 361, 'loss/train': 7.596373558044434} 02/24/2022 02:02:01 - INFO - codeparrot_training - Step 362: {'lr': 9.05e-05, 'samples': 185856, 'steps': 362, 'loss/train': 7.922679901123047} 02/24/2022 02:02:04 - INFO - codeparrot_training - Step 363: {'lr': 9.075e-05, 'samples': 186368, 'steps': 363, 'loss/train': 7.51069450378418} 02/24/2022 02:02:10 - INFO - codeparrot_training - Step 364: {'lr': 9.1e-05, 'samples': 186880, 'steps': 364, 'loss/train': 7.531533241271973} 02/24/2022 02:02:14 - INFO - codeparrot_training - Step 365: {'lr': 9.125e-05, 'samples': 187392, 'steps': 365, 'loss/train': 8.281204223632812} 02/24/2022 02:02:19 - INFO - codeparrot_training - Step 366: {'lr': 9.15e-05, 'samples': 187904, 'steps': 366, 'loss/train': 7.3093342781066895} 02/24/2022 02:02:23 - INFO - codeparrot_training - Step 367: {'lr': 9.175e-05, 'samples': 188416, 'steps': 367, 'loss/train': 8.077080726623535} 02/24/2022 02:02:28 - INFO - codeparrot_training - Step 368: {'lr': 9.2e-05, 'samples': 188928, 'steps': 368, 'loss/train': 7.250057697296143} 02/24/2022 02:02:31 - INFO - codeparrot_training - Step 369: {'lr': 9.225e-05, 'samples': 189440, 'steps': 369, 'loss/train': 7.866724967956543} 02/24/2022 02:02:37 - INFO - codeparrot_training - Step 370: {'lr': 9.25e-05, 'samples': 189952, 'steps': 370, 'loss/train': 6.862495422363281} 02/24/2022 02:02:41 - INFO - codeparrot_training - Step 371: {'lr': 9.275e-05, 'samples': 190464, 'steps': 371, 'loss/train': 8.284356117248535} 02/24/2022 02:02:46 - INFO - codeparrot_training - Step 372: {'lr': 9.3e-05, 'samples': 190976, 'steps': 372, 'loss/train': 7.869273662567139} 02/24/2022 02:02:50 - INFO - codeparrot_training - Step 373: {'lr': 9.325e-05, 'samples': 191488, 'steps': 373, 'loss/train': 8.418437957763672} 02/24/2022 02:02:55 - INFO - codeparrot_training - Step 374: {'lr': 9.35e-05, 'samples': 192000, 'steps': 374, 'loss/train': 10.001588821411133} 02/24/2022 02:02:59 - INFO - codeparrot_training - Step 375: {'lr': 9.375e-05, 'samples': 192512, 'steps': 375, 'loss/train': 7.9266862869262695} 02/24/2022 02:03:04 - INFO - codeparrot_training - Step 376: {'lr': 9.400000000000001e-05, 'samples': 193024, 'steps': 376, 'loss/train': 8.20534896850586} 02/24/2022 02:03:08 - INFO - codeparrot_training - Step 377: {'lr': 9.425e-05, 'samples': 193536, 'steps': 377, 'loss/train': 7.760704040527344} 02/24/2022 02:03:13 - INFO - codeparrot_training - Step 378: {'lr': 9.45e-05, 'samples': 194048, 'steps': 378, 'loss/train': 8.39869213104248} 02/24/2022 02:03:16 - INFO - codeparrot_training - Step 379: {'lr': 9.475e-05, 'samples': 194560, 'steps': 379, 'loss/train': 7.1428704261779785} 02/24/2022 02:03:23 - INFO - codeparrot_training - Step 380: {'lr': 9.5e-05, 'samples': 195072, 'steps': 380, 'loss/train': 7.34848690032959} 02/24/2022 02:03:26 - INFO - codeparrot_training - Step 381: {'lr': 9.525e-05, 'samples': 195584, 'steps': 381, 'loss/train': 7.505448341369629} 02/24/2022 02:03:32 - INFO - codeparrot_training - Step 382: {'lr': 9.55e-05, 'samples': 196096, 'steps': 382, 'loss/train': 7.698422908782959} 02/24/2022 02:03:35 - INFO - codeparrot_training - Step 383: {'lr': 9.575000000000001e-05, 'samples': 196608, 'steps': 383, 'loss/train': 8.264890670776367} 02/24/2022 02:03:40 - INFO - codeparrot_training - Step 384: {'lr': 9.6e-05, 'samples': 197120, 'steps': 384, 'loss/train': 7.688831329345703} 02/24/2022 02:03:44 - INFO - codeparrot_training - Step 385: {'lr': 9.625000000000001e-05, 'samples': 197632, 'steps': 385, 'loss/train': 7.142866134643555} 02/24/2022 02:03:50 - INFO - codeparrot_training - Step 386: {'lr': 9.65e-05, 'samples': 198144, 'steps': 386, 'loss/train': 7.939741611480713} 02/24/2022 02:03:53 - INFO - codeparrot_training - Step 387: {'lr': 9.675000000000001e-05, 'samples': 198656, 'steps': 387, 'loss/train': 7.401408672332764} 02/24/2022 02:03:59 - INFO - codeparrot_training - Step 388: {'lr': 9.7e-05, 'samples': 199168, 'steps': 388, 'loss/train': 7.98380708694458} 02/24/2022 02:04:02 - INFO - codeparrot_training - Step 389: {'lr': 9.725e-05, 'samples': 199680, 'steps': 389, 'loss/train': 5.516870498657227} 02/24/2022 02:04:08 - INFO - codeparrot_training - Step 390: {'lr': 9.750000000000001e-05, 'samples': 200192, 'steps': 390, 'loss/train': 8.51853084564209} 02/24/2022 02:04:13 - INFO - codeparrot_training - Step 391: {'lr': 9.775e-05, 'samples': 200704, 'steps': 391, 'loss/train': 7.015237331390381} 02/24/2022 02:04:17 - INFO - codeparrot_training - Step 392: {'lr': 9.800000000000001e-05, 'samples': 201216, 'steps': 392, 'loss/train': 9.577959060668945} 02/24/2022 02:04:22 - INFO - codeparrot_training - Step 393: {'lr': 9.825e-05, 'samples': 201728, 'steps': 393, 'loss/train': 7.536276817321777} 02/24/2022 02:04:25 - INFO - codeparrot_training - Step 394: {'lr': 9.850000000000001e-05, 'samples': 202240, 'steps': 394, 'loss/train': 7.226565837860107} 02/24/2022 02:04:32 - INFO - codeparrot_training - Step 395: {'lr': 9.875e-05, 'samples': 202752, 'steps': 395, 'loss/train': 8.628961563110352} 02/24/2022 02:04:35 - INFO - codeparrot_training - Step 396: {'lr': 9.900000000000001e-05, 'samples': 203264, 'steps': 396, 'loss/train': 7.343240737915039} 02/24/2022 02:04:40 - INFO - codeparrot_training - Step 397: {'lr': 9.925000000000001e-05, 'samples': 203776, 'steps': 397, 'loss/train': 7.60067081451416} 02/24/2022 02:04:44 - INFO - codeparrot_training - Step 398: {'lr': 9.95e-05, 'samples': 204288, 'steps': 398, 'loss/train': 7.258427143096924} 02/24/2022 02:04:49 - INFO - codeparrot_training - Step 399: {'lr': 9.975000000000001e-05, 'samples': 204800, 'steps': 399, 'loss/train': 6.988162517547607} 02/24/2022 02:04:53 - INFO - codeparrot_training - Step 400: {'lr': 0.0001, 'samples': 205312, 'steps': 400, 'loss/train': 7.182863235473633} 02/24/2022 02:04:58 - INFO - codeparrot_training - Step 401: {'lr': 0.00010025000000000001, 'samples': 205824, 'steps': 401, 'loss/train': 7.117374420166016} 02/24/2022 02:05:02 - INFO - codeparrot_training - Step 402: {'lr': 0.0001005, 'samples': 206336, 'steps': 402, 'loss/train': 7.725515842437744} 02/24/2022 02:05:07 - INFO - codeparrot_training - Step 403: {'lr': 0.00010075000000000001, 'samples': 206848, 'steps': 403, 'loss/train': 7.501499176025391} 02/24/2022 02:05:10 - INFO - codeparrot_training - Step 404: {'lr': 0.000101, 'samples': 207360, 'steps': 404, 'loss/train': 8.18635082244873} 02/24/2022 02:05:17 - INFO - codeparrot_training - Step 405: {'lr': 0.00010125000000000001, 'samples': 207872, 'steps': 405, 'loss/train': 8.187431335449219} 02/24/2022 02:05:20 - INFO - codeparrot_training - Step 406: {'lr': 0.00010150000000000001, 'samples': 208384, 'steps': 406, 'loss/train': 7.888332366943359} 02/24/2022 02:05:26 - INFO - codeparrot_training - Step 407: {'lr': 0.00010174999999999999, 'samples': 208896, 'steps': 407, 'loss/train': 7.384456157684326} 02/24/2022 02:05:29 - INFO - codeparrot_training - Step 408: {'lr': 0.000102, 'samples': 209408, 'steps': 408, 'loss/train': 7.803598403930664} 02/24/2022 02:05:35 - INFO - codeparrot_training - Step 409: {'lr': 0.00010224999999999999, 'samples': 209920, 'steps': 409, 'loss/train': 7.0473527908325195} 02/24/2022 02:05:38 - INFO - codeparrot_training - Step 410: {'lr': 0.0001025, 'samples': 210432, 'steps': 410, 'loss/train': 9.248546600341797} 02/24/2022 02:05:43 - INFO - codeparrot_training - Step 411: {'lr': 0.00010274999999999999, 'samples': 210944, 'steps': 411, 'loss/train': 7.117498397827148} 02/24/2022 02:05:47 - INFO - codeparrot_training - Step 412: {'lr': 0.000103, 'samples': 211456, 'steps': 412, 'loss/train': 7.365967273712158} 02/24/2022 02:05:52 - INFO - codeparrot_training - Step 413: {'lr': 0.00010325, 'samples': 211968, 'steps': 413, 'loss/train': 8.108454704284668} 02/24/2022 02:05:56 - INFO - codeparrot_training - Step 414: {'lr': 0.0001035, 'samples': 212480, 'steps': 414, 'loss/train': 7.5737385749816895} 02/24/2022 02:06:03 - INFO - codeparrot_training - Step 415: {'lr': 0.00010375, 'samples': 212992, 'steps': 415, 'loss/train': 8.08419132232666} 02/24/2022 02:06:06 - INFO - codeparrot_training - Step 416: {'lr': 0.000104, 'samples': 213504, 'steps': 416, 'loss/train': 7.4956183433532715} 02/24/2022 02:06:11 - INFO - codeparrot_training - Step 417: {'lr': 0.00010425, 'samples': 214016, 'steps': 417, 'loss/train': 7.642014503479004} 02/24/2022 02:06:15 - INFO - codeparrot_training - Step 418: {'lr': 0.00010449999999999999, 'samples': 214528, 'steps': 418, 'loss/train': 7.403052806854248} 02/24/2022 02:06:20 - INFO - codeparrot_training - Step 419: {'lr': 0.00010475, 'samples': 215040, 'steps': 419, 'loss/train': 7.374420166015625} 02/24/2022 02:06:24 - INFO - codeparrot_training - Step 420: {'lr': 0.000105, 'samples': 215552, 'steps': 420, 'loss/train': 7.816165924072266} 02/24/2022 02:06:29 - INFO - codeparrot_training - Step 421: {'lr': 0.00010525, 'samples': 216064, 'steps': 421, 'loss/train': 6.225070953369141} 02/24/2022 02:06:33 - INFO - codeparrot_training - Step 422: {'lr': 0.0001055, 'samples': 216576, 'steps': 422, 'loss/train': 7.872852325439453} 02/24/2022 02:06:38 - INFO - codeparrot_training - Step 423: {'lr': 0.00010575, 'samples': 217088, 'steps': 423, 'loss/train': 7.829193115234375} 02/24/2022 02:06:42 - INFO - codeparrot_training - Step 424: {'lr': 0.000106, 'samples': 217600, 'steps': 424, 'loss/train': 8.13473892211914} 02/24/2022 02:06:48 - INFO - codeparrot_training - Step 425: {'lr': 0.00010625, 'samples': 218112, 'steps': 425, 'loss/train': 3.2847564220428467} 02/24/2022 02:06:51 - INFO - codeparrot_training - Step 426: {'lr': 0.0001065, 'samples': 218624, 'steps': 426, 'loss/train': 6.639157295227051} 02/24/2022 02:06:57 - INFO - codeparrot_training - Step 427: {'lr': 0.00010675, 'samples': 219136, 'steps': 427, 'loss/train': 8.147637367248535} 02/24/2022 02:07:00 - INFO - codeparrot_training - Step 428: {'lr': 0.000107, 'samples': 219648, 'steps': 428, 'loss/train': 7.7012457847595215} 02/24/2022 02:07:06 - INFO - codeparrot_training - Step 429: {'lr': 0.00010725, 'samples': 220160, 'steps': 429, 'loss/train': 7.161632061004639} 02/24/2022 02:07:09 - INFO - codeparrot_training - Step 430: {'lr': 0.0001075, 'samples': 220672, 'steps': 430, 'loss/train': 7.4439873695373535} 02/24/2022 02:07:14 - INFO - codeparrot_training - Step 431: {'lr': 0.00010775, 'samples': 221184, 'steps': 431, 'loss/train': 8.414506912231445} 02/24/2022 02:07:18 - INFO - codeparrot_training - Step 432: {'lr': 0.000108, 'samples': 221696, 'steps': 432, 'loss/train': 7.436773777008057} 02/24/2022 02:07:23 - INFO - codeparrot_training - Step 433: {'lr': 0.00010825, 'samples': 222208, 'steps': 433, 'loss/train': 7.460289478302002} 02/24/2022 02:07:27 - INFO - codeparrot_training - Step 434: {'lr': 0.00010850000000000001, 'samples': 222720, 'steps': 434, 'loss/train': 8.266501426696777} 02/24/2022 02:07:32 - INFO - codeparrot_training - Step 435: {'lr': 0.00010875, 'samples': 223232, 'steps': 435, 'loss/train': 8.083974838256836} 02/24/2022 02:07:36 - INFO - codeparrot_training - Step 436: {'lr': 0.000109, 'samples': 223744, 'steps': 436, 'loss/train': 8.220951080322266} 02/24/2022 02:07:41 - INFO - codeparrot_training - Step 437: {'lr': 0.00010925, 'samples': 224256, 'steps': 437, 'loss/train': 6.722242832183838} 02/24/2022 02:07:45 - INFO - codeparrot_training - Step 438: {'lr': 0.0001095, 'samples': 224768, 'steps': 438, 'loss/train': 6.867252826690674} 02/24/2022 02:07:50 - INFO - codeparrot_training - Step 439: {'lr': 0.00010975, 'samples': 225280, 'steps': 439, 'loss/train': 6.986952304840088} 02/24/2022 02:07:54 - INFO - codeparrot_training - Step 440: {'lr': 0.00011, 'samples': 225792, 'steps': 440, 'loss/train': 7.2183146476745605} 02/24/2022 02:08:00 - INFO - codeparrot_training - Step 441: {'lr': 0.00011025, 'samples': 226304, 'steps': 441, 'loss/train': 7.320436954498291} 02/24/2022 02:08:03 - INFO - codeparrot_training - Step 442: {'lr': 0.0001105, 'samples': 226816, 'steps': 442, 'loss/train': 6.413392543792725} 02/24/2022 02:08:09 - INFO - codeparrot_training - Step 443: {'lr': 0.00011075000000000001, 'samples': 227328, 'steps': 443, 'loss/train': 7.322403430938721} 02/24/2022 02:08:12 - INFO - codeparrot_training - Step 444: {'lr': 0.000111, 'samples': 227840, 'steps': 444, 'loss/train': 7.285653114318848} 02/24/2022 02:08:18 - INFO - codeparrot_training - Step 445: {'lr': 0.00011125000000000001, 'samples': 228352, 'steps': 445, 'loss/train': 7.396255016326904} 02/24/2022 02:08:21 - INFO - codeparrot_training - Step 446: {'lr': 0.0001115, 'samples': 228864, 'steps': 446, 'loss/train': 7.2955827713012695} 02/24/2022 02:08:27 - INFO - codeparrot_training - Step 447: {'lr': 0.00011175, 'samples': 229376, 'steps': 447, 'loss/train': 4.9974751472473145} 02/24/2022 02:08:30 - INFO - codeparrot_training - Step 448: {'lr': 0.000112, 'samples': 229888, 'steps': 448, 'loss/train': 7.790682792663574} 02/24/2022 02:08:36 - INFO - codeparrot_training - Step 449: {'lr': 0.00011225, 'samples': 230400, 'steps': 449, 'loss/train': 7.6659979820251465} 02/24/2022 02:08:39 - INFO - codeparrot_training - Step 450: {'lr': 0.00011250000000000001, 'samples': 230912, 'steps': 450, 'loss/train': 7.553841590881348} 02/24/2022 02:08:45 - INFO - codeparrot_training - Step 451: {'lr': 0.00011275, 'samples': 231424, 'steps': 451, 'loss/train': 7.629961013793945} 02/24/2022 02:08:50 - INFO - codeparrot_training - Step 452: {'lr': 0.00011300000000000001, 'samples': 231936, 'steps': 452, 'loss/train': 4.840542316436768} 02/24/2022 02:08:54 - INFO - codeparrot_training - Step 453: {'lr': 0.00011325, 'samples': 232448, 'steps': 453, 'loss/train': 7.533297538757324} 02/24/2022 02:08:59 - INFO - codeparrot_training - Step 454: {'lr': 0.00011350000000000001, 'samples': 232960, 'steps': 454, 'loss/train': 7.914617538452148} 02/24/2022 02:09:03 - INFO - codeparrot_training - Step 455: {'lr': 0.00011375, 'samples': 233472, 'steps': 455, 'loss/train': 6.7937188148498535} 02/24/2022 02:09:08 - INFO - codeparrot_training - Step 456: {'lr': 0.000114, 'samples': 233984, 'steps': 456, 'loss/train': 6.527731418609619} 02/24/2022 02:09:12 - INFO - codeparrot_training - Step 457: {'lr': 0.00011425000000000001, 'samples': 234496, 'steps': 457, 'loss/train': 7.973141670227051} 02/24/2022 02:09:17 - INFO - codeparrot_training - Step 458: {'lr': 0.0001145, 'samples': 235008, 'steps': 458, 'loss/train': 7.240002155303955} 02/24/2022 02:09:21 - INFO - codeparrot_training - Step 459: {'lr': 0.00011475000000000001, 'samples': 235520, 'steps': 459, 'loss/train': 7.077408790588379} 02/24/2022 02:09:27 - INFO - codeparrot_training - Step 460: {'lr': 0.000115, 'samples': 236032, 'steps': 460, 'loss/train': 7.475757598876953} 02/24/2022 02:09:30 - INFO - codeparrot_training - Step 461: {'lr': 0.00011525000000000001, 'samples': 236544, 'steps': 461, 'loss/train': 7.6415300369262695} 02/24/2022 02:09:36 - INFO - codeparrot_training - Step 462: {'lr': 0.0001155, 'samples': 237056, 'steps': 462, 'loss/train': 8.935208320617676} 02/24/2022 02:09:39 - INFO - codeparrot_training - Step 463: {'lr': 0.00011575000000000001, 'samples': 237568, 'steps': 463, 'loss/train': 6.987576961517334} 02/24/2022 02:09:45 - INFO - codeparrot_training - Step 464: {'lr': 0.00011600000000000001, 'samples': 238080, 'steps': 464, 'loss/train': 7.757630825042725} 02/24/2022 02:09:48 - INFO - codeparrot_training - Step 465: {'lr': 0.00011625, 'samples': 238592, 'steps': 465, 'loss/train': 7.162644386291504} 02/24/2022 02:09:54 - INFO - codeparrot_training - Step 466: {'lr': 0.00011650000000000001, 'samples': 239104, 'steps': 466, 'loss/train': 8.270011901855469} 02/24/2022 02:09:57 - INFO - codeparrot_training - Step 467: {'lr': 0.00011675, 'samples': 239616, 'steps': 467, 'loss/train': 6.218019008636475} 02/24/2022 02:10:03 - INFO - codeparrot_training - Step 468: {'lr': 0.00011700000000000001, 'samples': 240128, 'steps': 468, 'loss/train': 6.125336647033691} 02/24/2022 02:10:06 - INFO - codeparrot_training - Step 469: {'lr': 0.00011724999999999999, 'samples': 240640, 'steps': 469, 'loss/train': 7.7407145500183105} 02/24/2022 02:10:12 - INFO - codeparrot_training - Step 470: {'lr': 0.0001175, 'samples': 241152, 'steps': 470, 'loss/train': 4.7091450691223145} 02/24/2022 02:10:15 - INFO - codeparrot_training - Step 471: {'lr': 0.00011775, 'samples': 241664, 'steps': 471, 'loss/train': 7.565977096557617} 02/24/2022 02:10:21 - INFO - codeparrot_training - Step 472: {'lr': 0.000118, 'samples': 242176, 'steps': 472, 'loss/train': 8.618671417236328} 02/24/2022 02:10:25 - INFO - codeparrot_training - Step 473: {'lr': 0.00011825, 'samples': 242688, 'steps': 473, 'loss/train': 7.5694475173950195} 02/24/2022 02:10:30 - INFO - codeparrot_training - Step 474: {'lr': 0.0001185, 'samples': 243200, 'steps': 474, 'loss/train': 6.634157657623291} 02/24/2022 02:10:34 - INFO - codeparrot_training - Step 475: {'lr': 0.00011875, 'samples': 243712, 'steps': 475, 'loss/train': 7.126220226287842} 02/24/2022 02:10:39 - INFO - codeparrot_training - Step 476: {'lr': 0.00011899999999999999, 'samples': 244224, 'steps': 476, 'loss/train': 7.417023181915283} 02/24/2022 02:10:43 - INFO - codeparrot_training - Step 477: {'lr': 0.00011925, 'samples': 244736, 'steps': 477, 'loss/train': 8.853533744812012} 02/24/2022 02:10:48 - INFO - codeparrot_training - Step 478: {'lr': 0.00011949999999999999, 'samples': 245248, 'steps': 478, 'loss/train': 6.4358439445495605} 02/24/2022 02:10:52 - INFO - codeparrot_training - Step 479: {'lr': 0.00011975, 'samples': 245760, 'steps': 479, 'loss/train': 8.242326736450195} 02/24/2022 02:10:57 - INFO - codeparrot_training - Step 480: {'lr': 0.00012, 'samples': 246272, 'steps': 480, 'loss/train': 7.858012676239014} 02/24/2022 02:11:01 - INFO - codeparrot_training - Step 481: {'lr': 0.00012025, 'samples': 246784, 'steps': 481, 'loss/train': 7.545612335205078} 02/24/2022 02:11:06 - INFO - codeparrot_training - Step 482: {'lr': 0.0001205, 'samples': 247296, 'steps': 482, 'loss/train': 7.130396842956543} 02/24/2022 02:11:10 - INFO - codeparrot_training - Step 483: {'lr': 0.00012075, 'samples': 247808, 'steps': 483, 'loss/train': 6.627236366271973} 02/24/2022 02:11:15 - INFO - codeparrot_training - Step 484: {'lr': 0.000121, 'samples': 248320, 'steps': 484, 'loss/train': 7.135700225830078} 02/24/2022 02:11:19 - INFO - codeparrot_training - Step 485: {'lr': 0.00012124999999999999, 'samples': 248832, 'steps': 485, 'loss/train': 7.578171253204346} 02/24/2022 02:11:24 - INFO - codeparrot_training - Step 486: {'lr': 0.0001215, 'samples': 249344, 'steps': 486, 'loss/train': 7.24437141418457} 02/24/2022 02:11:28 - INFO - codeparrot_training - Step 487: {'lr': 0.00012175, 'samples': 249856, 'steps': 487, 'loss/train': 4.430206775665283} 02/24/2022 02:11:34 - INFO - codeparrot_training - Step 488: {'lr': 0.000122, 'samples': 250368, 'steps': 488, 'loss/train': 7.954636096954346} 02/24/2022 02:11:37 - INFO - codeparrot_training - Step 489: {'lr': 0.00012225, 'samples': 250880, 'steps': 489, 'loss/train': 6.426181793212891} 02/24/2022 02:11:43 - INFO - codeparrot_training - Step 490: {'lr': 0.0001225, 'samples': 251392, 'steps': 490, 'loss/train': 7.321078300476074} 02/24/2022 02:11:46 - INFO - codeparrot_training - Step 491: {'lr': 0.00012275, 'samples': 251904, 'steps': 491, 'loss/train': 7.457513332366943} 02/24/2022 02:11:52 - INFO - codeparrot_training - Step 492: {'lr': 0.000123, 'samples': 252416, 'steps': 492, 'loss/train': 7.4673991203308105} 02/24/2022 02:11:55 - INFO - codeparrot_training - Step 493: {'lr': 0.00012325000000000001, 'samples': 252928, 'steps': 493, 'loss/train': 7.5666327476501465} 02/24/2022 02:12:01 - INFO - codeparrot_training - Step 494: {'lr': 0.0001235, 'samples': 253440, 'steps': 494, 'loss/train': 6.872106552124023} 02/24/2022 02:12:04 - INFO - codeparrot_training - Step 495: {'lr': 0.00012375, 'samples': 253952, 'steps': 495, 'loss/train': 7.914139270782471} 02/24/2022 02:12:10 - INFO - codeparrot_training - Step 496: {'lr': 0.000124, 'samples': 254464, 'steps': 496, 'loss/train': 7.269467830657959} 02/24/2022 02:12:13 - INFO - codeparrot_training - Step 497: {'lr': 0.00012425, 'samples': 254976, 'steps': 497, 'loss/train': 7.180480003356934} 02/24/2022 02:12:19 - INFO - codeparrot_training - Step 498: {'lr': 0.0001245, 'samples': 255488, 'steps': 498, 'loss/train': 7.44936466217041} 02/24/2022 02:12:22 - INFO - codeparrot_training - Step 499: {'lr': 0.00012475, 'samples': 256000, 'steps': 499, 'loss/train': 7.441803455352783} 02/24/2022 02:12:28 - INFO - codeparrot_training - Step 500: {'lr': 0.000125, 'samples': 256512, 'steps': 500, 'loss/train': 6.37514066696167} 02/24/2022 02:12:31 - INFO - codeparrot_training - Step 501: {'lr': 0.00012525, 'samples': 257024, 'steps': 501, 'loss/train': 7.1933794021606445} 02/24/2022 02:12:37 - INFO - codeparrot_training - Step 502: {'lr': 0.00012550000000000001, 'samples': 257536, 'steps': 502, 'loss/train': 7.345328330993652} 02/24/2022 02:12:40 - INFO - codeparrot_training - Step 503: {'lr': 0.00012575, 'samples': 258048, 'steps': 503, 'loss/train': 7.6105265617370605} 02/24/2022 02:12:46 - INFO - codeparrot_training - Step 504: {'lr': 0.000126, 'samples': 258560, 'steps': 504, 'loss/train': 8.217442512512207} 02/24/2022 02:12:49 - INFO - codeparrot_training - Step 505: {'lr': 0.00012625, 'samples': 259072, 'steps': 505, 'loss/train': 7.326420307159424} 02/24/2022 02:12:55 - INFO - codeparrot_training - Step 506: {'lr': 0.0001265, 'samples': 259584, 'steps': 506, 'loss/train': 7.832907199859619} 02/24/2022 02:12:58 - INFO - codeparrot_training - Step 507: {'lr': 0.00012675, 'samples': 260096, 'steps': 507, 'loss/train': 6.8345465660095215} 02/24/2022 02:13:05 - INFO - codeparrot_training - Step 508: {'lr': 0.000127, 'samples': 260608, 'steps': 508, 'loss/train': 7.008861064910889} 02/24/2022 02:13:08 - INFO - codeparrot_training - Step 509: {'lr': 0.00012725, 'samples': 261120, 'steps': 509, 'loss/train': 7.020104885101318} 02/24/2022 02:13:14 - INFO - codeparrot_training - Step 510: {'lr': 0.0001275, 'samples': 261632, 'steps': 510, 'loss/train': 7.388762474060059} 02/24/2022 02:13:17 - INFO - codeparrot_training - Step 511: {'lr': 0.00012775000000000002, 'samples': 262144, 'steps': 511, 'loss/train': 5.314887046813965} 02/24/2022 02:13:23 - INFO - codeparrot_training - Step 512: {'lr': 0.000128, 'samples': 262656, 'steps': 512, 'loss/train': 6.833120822906494} 02/24/2022 02:13:26 - INFO - codeparrot_training - Step 513: {'lr': 0.00012825, 'samples': 263168, 'steps': 513, 'loss/train': 4.8469953536987305} 02/24/2022 02:13:32 - INFO - codeparrot_training - Step 514: {'lr': 0.0001285, 'samples': 263680, 'steps': 514, 'loss/train': 7.254924774169922} 02/24/2022 02:13:35 - INFO - codeparrot_training - Step 515: {'lr': 0.00012875, 'samples': 264192, 'steps': 515, 'loss/train': 7.093168258666992} 02/24/2022 02:13:40 - INFO - codeparrot_training - Step 516: {'lr': 0.00012900000000000002, 'samples': 264704, 'steps': 516, 'loss/train': 6.9729108810424805} 02/24/2022 02:13:44 - INFO - codeparrot_training - Step 517: {'lr': 0.00012925, 'samples': 265216, 'steps': 517, 'loss/train': 7.3428215980529785} 02/24/2022 02:13:49 - INFO - codeparrot_training - Step 518: {'lr': 0.0001295, 'samples': 265728, 'steps': 518, 'loss/train': 7.924755096435547} 02/24/2022 02:13:53 - INFO - codeparrot_training - Step 519: {'lr': 0.00012975, 'samples': 266240, 'steps': 519, 'loss/train': 9.013562202453613} 02/24/2022 02:14:01 - INFO - codeparrot_training - Step 520: {'lr': 0.00013000000000000002, 'samples': 266752, 'steps': 520, 'loss/train': 6.795487403869629} 02/24/2022 02:14:04 - INFO - codeparrot_training - Step 521: {'lr': 0.00013025, 'samples': 267264, 'steps': 521, 'loss/train': 7.241908073425293} 02/24/2022 02:14:10 - INFO - codeparrot_training - Step 522: {'lr': 0.0001305, 'samples': 267776, 'steps': 522, 'loss/train': 6.705616474151611} 02/24/2022 02:14:13 - INFO - codeparrot_training - Step 523: {'lr': 0.00013075, 'samples': 268288, 'steps': 523, 'loss/train': 6.343685150146484} 02/24/2022 02:14:18 - INFO - codeparrot_training - Step 524: {'lr': 0.000131, 'samples': 268800, 'steps': 524, 'loss/train': 7.4530768394470215} 02/24/2022 02:14:22 - INFO - codeparrot_training - Step 525: {'lr': 0.00013125000000000002, 'samples': 269312, 'steps': 525, 'loss/train': 7.97170877456665} 02/24/2022 02:14:27 - INFO - codeparrot_training - Step 526: {'lr': 0.0001315, 'samples': 269824, 'steps': 526, 'loss/train': 6.723884582519531} 02/24/2022 02:14:31 - INFO - codeparrot_training - Step 527: {'lr': 0.00013175, 'samples': 270336, 'steps': 527, 'loss/train': 7.471153259277344} 02/24/2022 02:14:36 - INFO - codeparrot_training - Step 528: {'lr': 0.000132, 'samples': 270848, 'steps': 528, 'loss/train': 6.499663352966309} 02/24/2022 02:14:40 - INFO - codeparrot_training - Step 529: {'lr': 0.00013225000000000002, 'samples': 271360, 'steps': 529, 'loss/train': 7.3266730308532715} 02/24/2022 02:14:47 - INFO - codeparrot_training - Step 530: {'lr': 0.00013250000000000002, 'samples': 271872, 'steps': 530, 'loss/train': 6.079280853271484} 02/24/2022 02:14:50 - INFO - codeparrot_training - Step 531: {'lr': 0.00013275, 'samples': 272384, 'steps': 531, 'loss/train': 7.223377704620361} 02/24/2022 02:14:56 - INFO - codeparrot_training - Step 532: {'lr': 0.000133, 'samples': 272896, 'steps': 532, 'loss/train': 6.303267478942871} 02/24/2022 02:15:01 - INFO - codeparrot_training - Step 533: {'lr': 0.00013325, 'samples': 273408, 'steps': 533, 'loss/train': 6.488459587097168} 02/24/2022 02:15:05 - INFO - codeparrot_training - Step 534: {'lr': 0.00013350000000000002, 'samples': 273920, 'steps': 534, 'loss/train': 7.408422946929932} 02/24/2022 02:15:10 - INFO - codeparrot_training - Step 535: {'lr': 0.00013375, 'samples': 274432, 'steps': 535, 'loss/train': 7.3007588386535645} 02/24/2022 02:15:14 - INFO - codeparrot_training - Step 536: {'lr': 0.000134, 'samples': 274944, 'steps': 536, 'loss/train': 7.12288236618042} 02/24/2022 02:15:19 - INFO - codeparrot_training - Step 537: {'lr': 0.00013425, 'samples': 275456, 'steps': 537, 'loss/train': 6.852250099182129} 02/24/2022 02:15:23 - INFO - codeparrot_training - Step 538: {'lr': 0.00013450000000000002, 'samples': 275968, 'steps': 538, 'loss/train': 6.62864875793457} 02/24/2022 02:15:30 - INFO - codeparrot_training - Step 539: {'lr': 0.00013475000000000002, 'samples': 276480, 'steps': 539, 'loss/train': 7.144618034362793} 02/24/2022 02:15:33 - INFO - codeparrot_training - Step 540: {'lr': 0.000135, 'samples': 276992, 'steps': 540, 'loss/train': 6.56927490234375} 02/24/2022 02:15:39 - INFO - codeparrot_training - Step 541: {'lr': 0.00013525, 'samples': 277504, 'steps': 541, 'loss/train': 7.681875705718994} 02/24/2022 02:15:42 - INFO - codeparrot_training - Step 542: {'lr': 0.00013550000000000001, 'samples': 278016, 'steps': 542, 'loss/train': 6.560507297515869} 02/24/2022 02:15:48 - INFO - codeparrot_training - Step 543: {'lr': 0.00013575000000000002, 'samples': 278528, 'steps': 543, 'loss/train': 6.655404567718506} 02/24/2022 02:15:51 - INFO - codeparrot_training - Step 544: {'lr': 0.00013600000000000003, 'samples': 279040, 'steps': 544, 'loss/train': 6.519329071044922} 02/24/2022 02:15:57 - INFO - codeparrot_training - Step 545: {'lr': 0.00013625, 'samples': 279552, 'steps': 545, 'loss/train': 5.666187763214111} 02/24/2022 02:16:00 - INFO - codeparrot_training - Step 546: {'lr': 0.0001365, 'samples': 280064, 'steps': 546, 'loss/train': 7.2396979331970215} 02/24/2022 02:16:06 - INFO - codeparrot_training - Step 547: {'lr': 0.00013675000000000002, 'samples': 280576, 'steps': 547, 'loss/train': 6.303050994873047} 02/24/2022 02:16:09 - INFO - codeparrot_training - Step 548: {'lr': 0.00013700000000000002, 'samples': 281088, 'steps': 548, 'loss/train': 8.919416427612305} 02/24/2022 02:16:17 - INFO - codeparrot_training - Step 549: {'lr': 0.00013725, 'samples': 281600, 'steps': 549, 'loss/train': 4.016124248504639} 02/24/2022 02:16:20 - INFO - codeparrot_training - Step 550: {'lr': 0.0001375, 'samples': 282112, 'steps': 550, 'loss/train': 6.259299278259277} 02/24/2022 02:16:26 - INFO - codeparrot_training - Step 551: {'lr': 0.00013775000000000001, 'samples': 282624, 'steps': 551, 'loss/train': 6.790286064147949} 02/24/2022 02:16:29 - INFO - codeparrot_training - Step 552: {'lr': 0.00013800000000000002, 'samples': 283136, 'steps': 552, 'loss/train': 6.966577529907227} 02/24/2022 02:16:35 - INFO - codeparrot_training - Step 553: {'lr': 0.00013825000000000003, 'samples': 283648, 'steps': 553, 'loss/train': 7.551130771636963} 02/24/2022 02:16:38 - INFO - codeparrot_training - Step 554: {'lr': 0.0001385, 'samples': 284160, 'steps': 554, 'loss/train': 7.652850151062012} 02/24/2022 02:16:44 - INFO - codeparrot_training - Step 555: {'lr': 0.00013875, 'samples': 284672, 'steps': 555, 'loss/train': 7.365139007568359} 02/24/2022 02:16:47 - INFO - codeparrot_training - Step 556: {'lr': 0.00013900000000000002, 'samples': 285184, 'steps': 556, 'loss/train': 7.868185520172119} 02/24/2022 02:16:53 - INFO - codeparrot_training - Step 557: {'lr': 0.00013925000000000002, 'samples': 285696, 'steps': 557, 'loss/train': 7.059309959411621} 02/24/2022 02:16:56 - INFO - codeparrot_training - Step 558: {'lr': 0.0001395, 'samples': 286208, 'steps': 558, 'loss/train': 6.935769557952881} 02/24/2022 02:17:02 - INFO - codeparrot_training - Step 559: {'lr': 0.00013975, 'samples': 286720, 'steps': 559, 'loss/train': 6.963362693786621} 02/24/2022 02:17:05 - INFO - codeparrot_training - Step 560: {'lr': 0.00014000000000000001, 'samples': 287232, 'steps': 560, 'loss/train': 7.0291948318481445} 02/24/2022 02:17:11 - INFO - codeparrot_training - Step 561: {'lr': 0.00014025000000000002, 'samples': 287744, 'steps': 561, 'loss/train': 7.17509651184082} 02/24/2022 02:17:14 - INFO - codeparrot_training - Step 562: {'lr': 0.00014050000000000003, 'samples': 288256, 'steps': 562, 'loss/train': 6.840673446655273} 02/24/2022 02:17:20 - INFO - codeparrot_training - Step 563: {'lr': 0.00014074999999999998, 'samples': 288768, 'steps': 563, 'loss/train': 7.216966152191162} 02/24/2022 02:17:23 - INFO - codeparrot_training - Step 564: {'lr': 0.00014099999999999998, 'samples': 289280, 'steps': 564, 'loss/train': 7.15809440612793} 02/24/2022 02:17:29 - INFO - codeparrot_training - Step 565: {'lr': 0.00014125, 'samples': 289792, 'steps': 565, 'loss/train': 8.598647117614746} 02/24/2022 02:17:32 - INFO - codeparrot_training - Step 566: {'lr': 0.0001415, 'samples': 290304, 'steps': 566, 'loss/train': 6.900800704956055} 02/24/2022 02:17:39 - INFO - codeparrot_training - Step 567: {'lr': 0.00014175, 'samples': 290816, 'steps': 567, 'loss/train': 7.053728103637695} 02/24/2022 02:17:43 - INFO - codeparrot_training - Step 568: {'lr': 0.00014199999999999998, 'samples': 291328, 'steps': 568, 'loss/train': 7.268818378448486} 02/24/2022 02:17:48 - INFO - codeparrot_training - Step 569: {'lr': 0.00014225, 'samples': 291840, 'steps': 569, 'loss/train': 7.903963565826416} 02/24/2022 02:17:52 - INFO - codeparrot_training - Step 570: {'lr': 0.0001425, 'samples': 292352, 'steps': 570, 'loss/train': 7.795132160186768} 02/24/2022 02:17:57 - INFO - codeparrot_training - Step 571: {'lr': 0.00014275, 'samples': 292864, 'steps': 571, 'loss/train': 6.7889018058776855} 02/24/2022 02:18:03 - INFO - codeparrot_training - Step 572: {'lr': 0.00014299999999999998, 'samples': 293376, 'steps': 572, 'loss/train': 8.538932800292969} 02/24/2022 02:18:06 - INFO - codeparrot_training - Step 573: {'lr': 0.00014324999999999999, 'samples': 293888, 'steps': 573, 'loss/train': 6.632084369659424} 02/24/2022 02:18:11 - INFO - codeparrot_training - Step 574: {'lr': 0.0001435, 'samples': 294400, 'steps': 574, 'loss/train': 6.36543083190918} 02/24/2022 02:18:15 - INFO - codeparrot_training - Step 575: {'lr': 0.00014375, 'samples': 294912, 'steps': 575, 'loss/train': 6.891174793243408} 02/24/2022 02:18:22 - INFO - codeparrot_training - Step 576: {'lr': 0.000144, 'samples': 295424, 'steps': 576, 'loss/train': 4.885776519775391} 02/24/2022 02:18:26 - INFO - codeparrot_training - Step 577: {'lr': 0.00014424999999999998, 'samples': 295936, 'steps': 577, 'loss/train': 6.663343906402588} 02/24/2022 02:18:31 - INFO - codeparrot_training - Step 578: {'lr': 0.0001445, 'samples': 296448, 'steps': 578, 'loss/train': 5.8480987548828125} 02/24/2022 02:18:35 - INFO - codeparrot_training - Step 579: {'lr': 0.00014475, 'samples': 296960, 'steps': 579, 'loss/train': 7.730355739593506} 02/24/2022 02:18:40 - INFO - codeparrot_training - Step 580: {'lr': 0.000145, 'samples': 297472, 'steps': 580, 'loss/train': 7.83709716796875} 02/24/2022 02:18:44 - INFO - codeparrot_training - Step 581: {'lr': 0.00014524999999999998, 'samples': 297984, 'steps': 581, 'loss/train': 6.554624557495117} 02/24/2022 02:18:49 - INFO - codeparrot_training - Step 582: {'lr': 0.00014549999999999999, 'samples': 298496, 'steps': 582, 'loss/train': 7.251741886138916} 02/24/2022 02:18:52 - INFO - codeparrot_training - Step 583: {'lr': 0.00014575, 'samples': 299008, 'steps': 583, 'loss/train': 7.171562194824219} 02/24/2022 02:18:58 - INFO - codeparrot_training - Step 584: {'lr': 0.000146, 'samples': 299520, 'steps': 584, 'loss/train': 5.9258341789245605} 02/24/2022 02:19:01 - INFO - codeparrot_training - Step 585: {'lr': 0.00014625, 'samples': 300032, 'steps': 585, 'loss/train': 6.827935218811035} 02/24/2022 02:19:09 - INFO - codeparrot_training - Step 586: {'lr': 0.00014649999999999998, 'samples': 300544, 'steps': 586, 'loss/train': 6.947076797485352} 02/24/2022 02:19:12 - INFO - codeparrot_training - Step 587: {'lr': 0.00014675, 'samples': 301056, 'steps': 587, 'loss/train': 7.862649917602539} 02/24/2022 02:19:18 - INFO - codeparrot_training - Step 588: {'lr': 0.000147, 'samples': 301568, 'steps': 588, 'loss/train': 7.695960521697998} 02/24/2022 02:19:21 - INFO - codeparrot_training - Step 589: {'lr': 0.00014725, 'samples': 302080, 'steps': 589, 'loss/train': 3.70426344871521} 02/24/2022 02:19:26 - INFO - codeparrot_training - Step 590: {'lr': 0.0001475, 'samples': 302592, 'steps': 590, 'loss/train': 7.610629558563232} 02/24/2022 02:19:30 - INFO - codeparrot_training - Step 591: {'lr': 0.00014774999999999999, 'samples': 303104, 'steps': 591, 'loss/train': 6.706225872039795} 02/24/2022 02:19:35 - INFO - codeparrot_training - Step 592: {'lr': 0.000148, 'samples': 303616, 'steps': 592, 'loss/train': 7.33603572845459} 02/24/2022 02:19:39 - INFO - codeparrot_training - Step 593: {'lr': 0.00014825, 'samples': 304128, 'steps': 593, 'loss/train': 6.951836585998535} 02/24/2022 02:19:44 - INFO - codeparrot_training - Step 594: {'lr': 0.0001485, 'samples': 304640, 'steps': 594, 'loss/train': 6.484375953674316} 02/24/2022 02:19:48 - INFO - codeparrot_training - Step 595: {'lr': 0.00014874999999999998, 'samples': 305152, 'steps': 595, 'loss/train': 6.624990463256836} 02/24/2022 02:19:55 - INFO - codeparrot_training - Step 596: {'lr': 0.000149, 'samples': 305664, 'steps': 596, 'loss/train': 6.708791255950928} 02/24/2022 02:19:58 - INFO - codeparrot_training - Step 597: {'lr': 0.00014925, 'samples': 306176, 'steps': 597, 'loss/train': 5.992984294891357} 02/24/2022 02:20:04 - INFO - codeparrot_training - Step 598: {'lr': 0.0001495, 'samples': 306688, 'steps': 598, 'loss/train': 7.315062522888184} 02/24/2022 02:20:07 - INFO - codeparrot_training - Step 599: {'lr': 0.00014975, 'samples': 307200, 'steps': 599, 'loss/train': 7.080319881439209} 02/24/2022 02:20:13 - INFO - codeparrot_training - Step 600: {'lr': 0.00015, 'samples': 307712, 'steps': 600, 'loss/train': 6.57517671585083} 02/24/2022 02:20:16 - INFO - codeparrot_training - Step 601: {'lr': 0.00015025, 'samples': 308224, 'steps': 601, 'loss/train': 3.876352548599243} 02/24/2022 02:20:22 - INFO - codeparrot_training - Step 602: {'lr': 0.0001505, 'samples': 308736, 'steps': 602, 'loss/train': 6.989983558654785} 02/24/2022 02:20:25 - INFO - codeparrot_training - Step 603: {'lr': 0.00015075, 'samples': 309248, 'steps': 603, 'loss/train': 6.711729526519775} 02/24/2022 02:20:31 - INFO - codeparrot_training - Step 604: {'lr': 0.000151, 'samples': 309760, 'steps': 604, 'loss/train': 6.691339492797852} 02/24/2022 02:20:34 - INFO - codeparrot_training - Step 605: {'lr': 0.00015125, 'samples': 310272, 'steps': 605, 'loss/train': 7.045860767364502} 02/24/2022 02:20:39 - INFO - codeparrot_training - Step 606: {'lr': 0.0001515, 'samples': 310784, 'steps': 606, 'loss/train': 6.741722106933594} 02/24/2022 02:20:43 - INFO - codeparrot_training - Step 607: {'lr': 0.00015175, 'samples': 311296, 'steps': 607, 'loss/train': 7.056661128997803} 02/24/2022 02:20:49 - INFO - codeparrot_training - Step 608: {'lr': 0.000152, 'samples': 311808, 'steps': 608, 'loss/train': 6.854211807250977} 02/24/2022 02:20:52 - INFO - codeparrot_training - Step 609: {'lr': 0.00015225, 'samples': 312320, 'steps': 609, 'loss/train': 6.906655788421631} 02/24/2022 02:20:57 - INFO - codeparrot_training - Step 610: {'lr': 0.0001525, 'samples': 312832, 'steps': 610, 'loss/train': 5.857024192810059} 02/24/2022 02:21:01 - INFO - codeparrot_training - Step 611: {'lr': 0.00015275, 'samples': 313344, 'steps': 611, 'loss/train': 6.103697776794434} 02/24/2022 02:21:08 - INFO - codeparrot_training - Step 612: {'lr': 0.000153, 'samples': 313856, 'steps': 612, 'loss/train': 6.716464042663574} 02/24/2022 02:21:12 - INFO - codeparrot_training - Step 613: {'lr': 0.00015325, 'samples': 314368, 'steps': 613, 'loss/train': 6.938880920410156} 02/24/2022 02:21:17 - INFO - codeparrot_training - Step 614: {'lr': 0.0001535, 'samples': 314880, 'steps': 614, 'loss/train': 6.522587776184082} 02/24/2022 02:21:23 - INFO - codeparrot_training - Step 615: {'lr': 0.00015375, 'samples': 315392, 'steps': 615, 'loss/train': 6.33483362197876} 02/24/2022 02:21:26 - INFO - codeparrot_training - Step 616: {'lr': 0.000154, 'samples': 315904, 'steps': 616, 'loss/train': 6.6959333419799805} 02/24/2022 02:21:32 - INFO - codeparrot_training - Step 617: {'lr': 0.00015425, 'samples': 316416, 'steps': 617, 'loss/train': 6.425475120544434} 02/24/2022 02:21:35 - INFO - codeparrot_training - Step 618: {'lr': 0.00015450000000000001, 'samples': 316928, 'steps': 618, 'loss/train': 8.284980773925781} 02/24/2022 02:21:40 - INFO - codeparrot_training - Step 619: {'lr': 0.00015475, 'samples': 317440, 'steps': 619, 'loss/train': 4.899602890014648} 02/24/2022 02:21:44 - INFO - codeparrot_training - Step 620: {'lr': 0.000155, 'samples': 317952, 'steps': 620, 'loss/train': 7.085748195648193} 02/24/2022 02:21:51 - INFO - codeparrot_training - Step 621: {'lr': 0.00015525, 'samples': 318464, 'steps': 621, 'loss/train': 7.098223686218262} 02/24/2022 02:21:55 - INFO - codeparrot_training - Step 622: {'lr': 0.0001555, 'samples': 318976, 'steps': 622, 'loss/train': 6.562531471252441} 02/24/2022 02:22:00 - INFO - codeparrot_training - Step 623: {'lr': 0.00015575, 'samples': 319488, 'steps': 623, 'loss/train': 6.3596882820129395} 02/24/2022 02:22:03 - INFO - codeparrot_training - Step 624: {'lr': 0.000156, 'samples': 320000, 'steps': 624, 'loss/train': 7.403925895690918} 02/24/2022 02:22:09 - INFO - codeparrot_training - Step 625: {'lr': 0.00015625, 'samples': 320512, 'steps': 625, 'loss/train': 7.254275798797607} 02/24/2022 02:22:12 - INFO - codeparrot_training - Step 626: {'lr': 0.0001565, 'samples': 321024, 'steps': 626, 'loss/train': 7.162297248840332} 02/24/2022 02:22:18 - INFO - codeparrot_training - Step 627: {'lr': 0.00015675000000000002, 'samples': 321536, 'steps': 627, 'loss/train': 7.67238187789917} 02/24/2022 02:22:21 - INFO - codeparrot_training - Step 628: {'lr': 0.000157, 'samples': 322048, 'steps': 628, 'loss/train': 6.712908744812012} 02/24/2022 02:22:27 - INFO - codeparrot_training - Step 629: {'lr': 0.00015725, 'samples': 322560, 'steps': 629, 'loss/train': 6.164926052093506} 02/24/2022 02:22:30 - INFO - codeparrot_training - Step 630: {'lr': 0.0001575, 'samples': 323072, 'steps': 630, 'loss/train': 7.04307746887207} 02/24/2022 02:22:37 - INFO - codeparrot_training - Step 631: {'lr': 0.00015775, 'samples': 323584, 'steps': 631, 'loss/train': 7.326963901519775} 02/24/2022 02:22:41 - INFO - codeparrot_training - Step 632: {'lr': 0.000158, 'samples': 324096, 'steps': 632, 'loss/train': 7.378395080566406} 02/24/2022 02:22:46 - INFO - codeparrot_training - Step 633: {'lr': 0.00015825, 'samples': 324608, 'steps': 633, 'loss/train': 6.594248294830322} 02/24/2022 02:22:50 - INFO - codeparrot_training - Step 634: {'lr': 0.0001585, 'samples': 325120, 'steps': 634, 'loss/train': 6.691349983215332} 02/24/2022 02:22:55 - INFO - codeparrot_training - Step 635: {'lr': 0.00015875, 'samples': 325632, 'steps': 635, 'loss/train': 6.84854793548584} 02/24/2022 02:22:59 - INFO - codeparrot_training - Step 636: {'lr': 0.00015900000000000002, 'samples': 326144, 'steps': 636, 'loss/train': 7.24489164352417} 02/24/2022 02:23:04 - INFO - codeparrot_training - Step 637: {'lr': 0.00015925, 'samples': 326656, 'steps': 637, 'loss/train': 5.990461826324463} 02/24/2022 02:23:08 - INFO - codeparrot_training - Step 638: {'lr': 0.0001595, 'samples': 327168, 'steps': 638, 'loss/train': 6.8416428565979} 02/24/2022 02:23:13 - INFO - codeparrot_training - Step 639: {'lr': 0.00015975, 'samples': 327680, 'steps': 639, 'loss/train': 7.200850009918213} 02/24/2022 02:23:16 - INFO - codeparrot_training - Step 640: {'lr': 0.00016, 'samples': 328192, 'steps': 640, 'loss/train': 7.323842525482178} 02/24/2022 02:23:22 - INFO - codeparrot_training - Step 641: {'lr': 0.00016025000000000002, 'samples': 328704, 'steps': 641, 'loss/train': 6.524571895599365} 02/24/2022 02:23:25 - INFO - codeparrot_training - Step 642: {'lr': 0.0001605, 'samples': 329216, 'steps': 642, 'loss/train': 6.505560398101807} 02/24/2022 02:23:33 - INFO - codeparrot_training - Step 643: {'lr': 0.00016075, 'samples': 329728, 'steps': 643, 'loss/train': 7.442710876464844} 02/24/2022 02:23:37 - INFO - codeparrot_training - Step 644: {'lr': 0.000161, 'samples': 330240, 'steps': 644, 'loss/train': 7.752229690551758} 02/24/2022 02:23:42 - INFO - codeparrot_training - Step 645: {'lr': 0.00016125000000000002, 'samples': 330752, 'steps': 645, 'loss/train': 6.916399002075195} 02/24/2022 02:23:46 - INFO - codeparrot_training - Step 646: {'lr': 0.0001615, 'samples': 331264, 'steps': 646, 'loss/train': 6.4619011878967285} 02/24/2022 02:23:51 - INFO - codeparrot_training - Step 647: {'lr': 0.00016175, 'samples': 331776, 'steps': 647, 'loss/train': 6.152773380279541} 02/24/2022 02:23:54 - INFO - codeparrot_training - Step 648: {'lr': 0.000162, 'samples': 332288, 'steps': 648, 'loss/train': 6.222464084625244} 02/24/2022 02:24:00 - INFO - codeparrot_training - Step 649: {'lr': 0.00016225000000000001, 'samples': 332800, 'steps': 649, 'loss/train': 7.1147780418396} 02/24/2022 02:24:04 - INFO - codeparrot_training - Step 650: {'lr': 0.00016250000000000002, 'samples': 333312, 'steps': 650, 'loss/train': 4.591808795928955} 02/24/2022 02:24:09 - INFO - codeparrot_training - Step 651: {'lr': 0.00016275, 'samples': 333824, 'steps': 651, 'loss/train': 7.736910343170166} 02/24/2022 02:24:13 - INFO - codeparrot_training - Step 652: {'lr': 0.000163, 'samples': 334336, 'steps': 652, 'loss/train': 6.243583679199219} 02/24/2022 02:24:20 - INFO - codeparrot_training - Step 653: {'lr': 0.00016325, 'samples': 334848, 'steps': 653, 'loss/train': 6.474334239959717} 02/24/2022 02:24:23 - INFO - codeparrot_training - Step 654: {'lr': 0.00016350000000000002, 'samples': 335360, 'steps': 654, 'loss/train': 7.204182147979736} 02/24/2022 02:24:29 - INFO - codeparrot_training - Step 655: {'lr': 0.00016375000000000002, 'samples': 335872, 'steps': 655, 'loss/train': 6.291009902954102} 02/24/2022 02:24:32 - INFO - codeparrot_training - Step 656: {'lr': 0.000164, 'samples': 336384, 'steps': 656, 'loss/train': 5.735008239746094} 02/24/2022 02:24:38 - INFO - codeparrot_training - Step 657: {'lr': 0.00016425, 'samples': 336896, 'steps': 657, 'loss/train': 5.332834243774414} 02/24/2022 02:24:41 - INFO - codeparrot_training - Step 658: {'lr': 0.00016450000000000001, 'samples': 337408, 'steps': 658, 'loss/train': 6.378315448760986} 02/24/2022 02:24:47 - INFO - codeparrot_training - Step 659: {'lr': 0.00016475000000000002, 'samples': 337920, 'steps': 659, 'loss/train': 6.292074203491211} 02/24/2022 02:24:50 - INFO - codeparrot_training - Step 660: {'lr': 0.000165, 'samples': 338432, 'steps': 660, 'loss/train': 9.866281509399414} 02/24/2022 02:24:56 - INFO - codeparrot_training - Step 661: {'lr': 0.00016525, 'samples': 338944, 'steps': 661, 'loss/train': 6.443175792694092} 02/24/2022 02:24:59 - INFO - codeparrot_training - Step 662: {'lr': 0.0001655, 'samples': 339456, 'steps': 662, 'loss/train': 6.011322975158691} 02/24/2022 02:25:05 - INFO - codeparrot_training - Step 663: {'lr': 0.00016575000000000002, 'samples': 339968, 'steps': 663, 'loss/train': 7.997130393981934} 02/24/2022 02:25:08 - INFO - codeparrot_training - Step 664: {'lr': 0.00016600000000000002, 'samples': 340480, 'steps': 664, 'loss/train': 6.515625} 02/24/2022 02:25:14 - INFO - codeparrot_training - Step 665: {'lr': 0.00016625, 'samples': 340992, 'steps': 665, 'loss/train': 6.358397960662842} 02/24/2022 02:25:17 - INFO - codeparrot_training - Step 666: {'lr': 0.0001665, 'samples': 341504, 'steps': 666, 'loss/train': 7.1998515129089355} 02/24/2022 02:25:23 - INFO - codeparrot_training - Step 667: {'lr': 0.00016675000000000001, 'samples': 342016, 'steps': 667, 'loss/train': 6.34958028793335} 02/24/2022 02:25:26 - INFO - codeparrot_training - Step 668: {'lr': 0.00016700000000000002, 'samples': 342528, 'steps': 668, 'loss/train': 6.854135036468506} 02/24/2022 02:25:32 - INFO - codeparrot_training - Step 669: {'lr': 0.00016725000000000003, 'samples': 343040, 'steps': 669, 'loss/train': 5.932799339294434} 02/24/2022 02:25:36 - INFO - codeparrot_training - Step 670: {'lr': 0.0001675, 'samples': 343552, 'steps': 670, 'loss/train': 6.308033466339111} 02/24/2022 02:25:41 - INFO - codeparrot_training - Step 671: {'lr': 0.00016775, 'samples': 344064, 'steps': 671, 'loss/train': 7.460987567901611} 02/24/2022 02:25:45 - INFO - codeparrot_training - Step 672: {'lr': 0.00016800000000000002, 'samples': 344576, 'steps': 672, 'loss/train': 6.332848072052002} 02/24/2022 02:25:50 - INFO - codeparrot_training - Step 673: {'lr': 0.00016825000000000002, 'samples': 345088, 'steps': 673, 'loss/train': 6.770963668823242} 02/24/2022 02:25:54 - INFO - codeparrot_training - Step 674: {'lr': 0.0001685, 'samples': 345600, 'steps': 674, 'loss/train': 6.474719524383545} 02/24/2022 02:25:59 - INFO - codeparrot_training - Step 675: {'lr': 0.00016875, 'samples': 346112, 'steps': 675, 'loss/train': 6.697624206542969} 02/24/2022 02:26:03 - INFO - codeparrot_training - Step 676: {'lr': 0.00016900000000000002, 'samples': 346624, 'steps': 676, 'loss/train': 6.2510480880737305} 02/24/2022 02:26:09 - INFO - codeparrot_training - Step 677: {'lr': 0.00016925000000000002, 'samples': 347136, 'steps': 677, 'loss/train': 6.77142333984375} 02/24/2022 02:26:12 - INFO - codeparrot_training - Step 678: {'lr': 0.00016950000000000003, 'samples': 347648, 'steps': 678, 'loss/train': 6.4684247970581055} 02/24/2022 02:26:18 - INFO - codeparrot_training - Step 679: {'lr': 0.00016975, 'samples': 348160, 'steps': 679, 'loss/train': 6.551549434661865} 02/24/2022 02:26:21 - INFO - codeparrot_training - Step 680: {'lr': 0.00017, 'samples': 348672, 'steps': 680, 'loss/train': 6.417777061462402} 02/24/2022 02:26:27 - INFO - codeparrot_training - Step 681: {'lr': 0.00017025000000000002, 'samples': 349184, 'steps': 681, 'loss/train': 6.795881748199463} 02/24/2022 02:26:30 - INFO - codeparrot_training - Step 682: {'lr': 0.00017050000000000002, 'samples': 349696, 'steps': 682, 'loss/train': 6.6553568840026855} 02/24/2022 02:26:36 - INFO - codeparrot_training - Step 683: {'lr': 0.00017075, 'samples': 350208, 'steps': 683, 'loss/train': 6.663280010223389} 02/24/2022 02:26:39 - INFO - codeparrot_training - Step 684: {'lr': 0.000171, 'samples': 350720, 'steps': 684, 'loss/train': 6.603376865386963} 02/24/2022 02:26:45 - INFO - codeparrot_training - Step 685: {'lr': 0.00017125000000000002, 'samples': 351232, 'steps': 685, 'loss/train': 6.150605201721191} 02/24/2022 02:26:48 - INFO - codeparrot_training - Step 686: {'lr': 0.00017150000000000002, 'samples': 351744, 'steps': 686, 'loss/train': 6.475432395935059} 02/24/2022 02:26:54 - INFO - codeparrot_training - Step 687: {'lr': 0.00017175000000000003, 'samples': 352256, 'steps': 687, 'loss/train': 7.101222038269043} 02/24/2022 02:26:57 - INFO - codeparrot_training - Step 688: {'lr': 0.00017199999999999998, 'samples': 352768, 'steps': 688, 'loss/train': 6.253207206726074} 02/24/2022 02:27:03 - INFO - codeparrot_training - Step 689: {'lr': 0.00017224999999999999, 'samples': 353280, 'steps': 689, 'loss/train': 6.175331115722656} 02/24/2022 02:27:07 - INFO - codeparrot_training - Step 690: {'lr': 0.0001725, 'samples': 353792, 'steps': 690, 'loss/train': 5.864757061004639} 02/24/2022 02:27:12 - INFO - codeparrot_training - Step 691: {'lr': 0.00017275, 'samples': 354304, 'steps': 691, 'loss/train': 7.658307075500488} 02/24/2022 02:27:15 - INFO - codeparrot_training - Step 692: {'lr': 0.000173, 'samples': 354816, 'steps': 692, 'loss/train': 6.467141628265381} 02/24/2022 02:27:21 - INFO - codeparrot_training - Step 693: {'lr': 0.00017324999999999998, 'samples': 355328, 'steps': 693, 'loss/train': 6.138640403747559} 02/24/2022 02:27:26 - INFO - codeparrot_training - Step 694: {'lr': 0.0001735, 'samples': 355840, 'steps': 694, 'loss/train': 6.835862159729004} 02/24/2022 02:27:30 - INFO - codeparrot_training - Step 695: {'lr': 0.00017375, 'samples': 356352, 'steps': 695, 'loss/train': 6.1372199058532715} 02/24/2022 02:27:35 - INFO - codeparrot_training - Step 696: {'lr': 0.000174, 'samples': 356864, 'steps': 696, 'loss/train': 6.7231221199035645} 02/24/2022 02:27:39 - INFO - codeparrot_training - Step 697: {'lr': 0.00017424999999999998, 'samples': 357376, 'steps': 697, 'loss/train': 6.801931381225586} 02/24/2022 02:27:45 - INFO - codeparrot_training - Step 698: {'lr': 0.00017449999999999999, 'samples': 357888, 'steps': 698, 'loss/train': 6.4526753425598145} 02/24/2022 02:27:48 - INFO - codeparrot_training - Step 699: {'lr': 0.00017475, 'samples': 358400, 'steps': 699, 'loss/train': 5.960089206695557} 02/24/2022 02:27:54 - INFO - codeparrot_training - Step 700: {'lr': 0.000175, 'samples': 358912, 'steps': 700, 'loss/train': 6.215817928314209} 02/24/2022 02:27:57 - INFO - codeparrot_training - Step 701: {'lr': 0.00017525, 'samples': 359424, 'steps': 701, 'loss/train': 6.193975448608398} 02/24/2022 02:28:03 - INFO - codeparrot_training - Step 702: {'lr': 0.00017549999999999998, 'samples': 359936, 'steps': 702, 'loss/train': 6.0116167068481445} 02/24/2022 02:28:06 - INFO - codeparrot_training - Step 703: {'lr': 0.00017575, 'samples': 360448, 'steps': 703, 'loss/train': 6.346678256988525} 02/24/2022 02:28:12 - INFO - codeparrot_training - Step 704: {'lr': 0.000176, 'samples': 360960, 'steps': 704, 'loss/train': 6.734238147735596} 02/24/2022 02:28:15 - INFO - codeparrot_training - Step 705: {'lr': 0.00017625, 'samples': 361472, 'steps': 705, 'loss/train': 7.386608123779297} 02/24/2022 02:28:21 - INFO - codeparrot_training - Step 706: {'lr': 0.00017649999999999998, 'samples': 361984, 'steps': 706, 'loss/train': 7.438027381896973} 02/24/2022 02:28:24 - INFO - codeparrot_training - Step 707: {'lr': 0.00017675, 'samples': 362496, 'steps': 707, 'loss/train': 7.111026763916016} 02/24/2022 02:28:30 - INFO - codeparrot_training - Step 708: {'lr': 0.000177, 'samples': 363008, 'steps': 708, 'loss/train': 5.961710453033447} 02/24/2022 02:28:33 - INFO - codeparrot_training - Step 709: {'lr': 0.00017725, 'samples': 363520, 'steps': 709, 'loss/train': 6.055693626403809} 02/24/2022 02:28:39 - INFO - codeparrot_training - Step 710: {'lr': 0.0001775, 'samples': 364032, 'steps': 710, 'loss/train': 6.290581703186035} 02/24/2022 02:28:42 - INFO - codeparrot_training - Step 711: {'lr': 0.00017774999999999998, 'samples': 364544, 'steps': 711, 'loss/train': 6.256446838378906} 02/24/2022 02:28:47 - INFO - codeparrot_training - Step 712: {'lr': 0.000178, 'samples': 365056, 'steps': 712, 'loss/train': 5.427511692047119} 02/24/2022 02:28:51 - INFO - codeparrot_training - Step 713: {'lr': 0.00017825, 'samples': 365568, 'steps': 713, 'loss/train': 6.285609245300293} 02/24/2022 02:28:57 - INFO - codeparrot_training - Step 714: {'lr': 0.0001785, 'samples': 366080, 'steps': 714, 'loss/train': 6.4496331214904785} 02/24/2022 02:29:01 - INFO - codeparrot_training - Step 715: {'lr': 0.00017875, 'samples': 366592, 'steps': 715, 'loss/train': 4.660919666290283} 02/24/2022 02:29:06 - INFO - codeparrot_training - Step 716: {'lr': 0.000179, 'samples': 367104, 'steps': 716, 'loss/train': 5.834559440612793} 02/24/2022 02:29:10 - INFO - codeparrot_training - Step 717: {'lr': 0.00017925, 'samples': 367616, 'steps': 717, 'loss/train': 6.639416694641113} 02/24/2022 02:29:15 - INFO - codeparrot_training - Step 718: {'lr': 0.0001795, 'samples': 368128, 'steps': 718, 'loss/train': 6.017248153686523} 02/24/2022 02:29:19 - INFO - codeparrot_training - Step 719: {'lr': 0.00017975, 'samples': 368640, 'steps': 719, 'loss/train': 5.847497940063477} 02/24/2022 02:29:24 - INFO - codeparrot_training - Step 720: {'lr': 0.00017999999999999998, 'samples': 369152, 'steps': 720, 'loss/train': 7.500033378601074} 02/24/2022 02:29:28 - INFO - codeparrot_training - Step 721: {'lr': 0.00018025, 'samples': 369664, 'steps': 721, 'loss/train': 6.163767337799072} 02/24/2022 02:29:33 - INFO - codeparrot_training - Step 722: {'lr': 0.0001805, 'samples': 370176, 'steps': 722, 'loss/train': 5.948116779327393} 02/24/2022 02:29:37 - INFO - codeparrot_training - Step 723: {'lr': 0.00018075, 'samples': 370688, 'steps': 723, 'loss/train': 6.244548320770264} 02/24/2022 02:29:43 - INFO - codeparrot_training - Step 724: {'lr': 0.000181, 'samples': 371200, 'steps': 724, 'loss/train': 3.4994406700134277} 02/24/2022 02:29:46 - INFO - codeparrot_training - Step 725: {'lr': 0.00018125, 'samples': 371712, 'steps': 725, 'loss/train': 6.339785099029541} 02/24/2022 02:29:52 - INFO - codeparrot_training - Step 726: {'lr': 0.0001815, 'samples': 372224, 'steps': 726, 'loss/train': 6.456050395965576} 02/24/2022 02:29:55 - INFO - codeparrot_training - Step 727: {'lr': 0.00018175, 'samples': 372736, 'steps': 727, 'loss/train': 6.083957195281982} 02/24/2022 02:30:01 - INFO - codeparrot_training - Step 728: {'lr': 0.000182, 'samples': 373248, 'steps': 728, 'loss/train': 6.422770023345947} 02/24/2022 02:30:04 - INFO - codeparrot_training - Step 729: {'lr': 0.00018225, 'samples': 373760, 'steps': 729, 'loss/train': 6.4083757400512695} 02/24/2022 02:30:10 - INFO - codeparrot_training - Step 730: {'lr': 0.0001825, 'samples': 374272, 'steps': 730, 'loss/train': 6.041470527648926} 02/24/2022 02:30:13 - INFO - codeparrot_training - Step 731: {'lr': 0.00018275, 'samples': 374784, 'steps': 731, 'loss/train': 5.690192699432373} 02/24/2022 02:30:19 - INFO - codeparrot_training - Step 732: {'lr': 0.000183, 'samples': 375296, 'steps': 732, 'loss/train': 6.055638790130615} 02/24/2022 02:30:22 - INFO - codeparrot_training - Step 733: {'lr': 0.00018325, 'samples': 375808, 'steps': 733, 'loss/train': 6.33910608291626} 02/24/2022 02:30:28 - INFO - codeparrot_training - Step 734: {'lr': 0.0001835, 'samples': 376320, 'steps': 734, 'loss/train': 5.1370086669921875} 02/24/2022 02:30:32 - INFO - codeparrot_training - Step 735: {'lr': 0.00018375, 'samples': 376832, 'steps': 735, 'loss/train': 6.075075626373291} 02/24/2022 02:30:37 - INFO - codeparrot_training - Step 736: {'lr': 0.000184, 'samples': 377344, 'steps': 736, 'loss/train': 6.304084300994873} 02/24/2022 02:30:41 - INFO - codeparrot_training - Step 737: {'lr': 0.00018425, 'samples': 377856, 'steps': 737, 'loss/train': 6.102715969085693} 02/24/2022 02:30:46 - INFO - codeparrot_training - Step 738: {'lr': 0.0001845, 'samples': 378368, 'steps': 738, 'loss/train': 6.641720294952393} 02/24/2022 02:30:49 - INFO - codeparrot_training - Step 739: {'lr': 0.00018475, 'samples': 378880, 'steps': 739, 'loss/train': 6.921244144439697} 02/24/2022 02:30:55 - INFO - codeparrot_training - Step 740: {'lr': 0.000185, 'samples': 379392, 'steps': 740, 'loss/train': 6.271901607513428} 02/24/2022 02:30:58 - INFO - codeparrot_training - Step 741: {'lr': 0.00018525, 'samples': 379904, 'steps': 741, 'loss/train': 5.727572917938232} 02/24/2022 02:31:04 - INFO - codeparrot_training - Step 742: {'lr': 0.0001855, 'samples': 380416, 'steps': 742, 'loss/train': 6.31984281539917} 02/24/2022 02:31:07 - INFO - codeparrot_training - Step 743: {'lr': 0.00018575000000000002, 'samples': 380928, 'steps': 743, 'loss/train': 6.738889694213867} 02/24/2022 02:31:13 - INFO - codeparrot_training - Step 744: {'lr': 0.000186, 'samples': 381440, 'steps': 744, 'loss/train': 6.099252700805664} 02/24/2022 02:31:17 - INFO - codeparrot_training - Step 745: {'lr': 0.00018625, 'samples': 381952, 'steps': 745, 'loss/train': 6.482193470001221} 02/24/2022 02:31:22 - INFO - codeparrot_training - Step 746: {'lr': 0.0001865, 'samples': 382464, 'steps': 746, 'loss/train': 6.471707344055176} 02/24/2022 02:31:26 - INFO - codeparrot_training - Step 747: {'lr': 0.00018675, 'samples': 382976, 'steps': 747, 'loss/train': 7.385077476501465} 02/24/2022 02:31:31 - INFO - codeparrot_training - Step 748: {'lr': 0.000187, 'samples': 383488, 'steps': 748, 'loss/train': 10.252251625061035} 02/24/2022 02:31:35 - INFO - codeparrot_training - Step 749: {'lr': 0.00018725, 'samples': 384000, 'steps': 749, 'loss/train': 7.1071271896362305} 02/24/2022 02:31:40 - INFO - codeparrot_training - Step 750: {'lr': 0.0001875, 'samples': 384512, 'steps': 750, 'loss/train': 6.437521457672119} 02/24/2022 02:31:44 - INFO - codeparrot_training - Step 751: {'lr': 0.00018775, 'samples': 385024, 'steps': 751, 'loss/train': 6.625316143035889} 02/24/2022 02:31:49 - INFO - codeparrot_training - Step 752: {'lr': 0.00018800000000000002, 'samples': 385536, 'steps': 752, 'loss/train': 6.356978416442871} 02/24/2022 02:31:53 - INFO - codeparrot_training - Step 753: {'lr': 0.00018825, 'samples': 386048, 'steps': 753, 'loss/train': 6.442242622375488} 02/24/2022 02:31:58 - INFO - codeparrot_training - Step 754: {'lr': 0.0001885, 'samples': 386560, 'steps': 754, 'loss/train': 6.0052490234375} 02/24/2022 02:32:02 - INFO - codeparrot_training - Step 755: {'lr': 0.00018875, 'samples': 387072, 'steps': 755, 'loss/train': 6.370572566986084} 02/24/2022 02:32:07 - INFO - codeparrot_training - Step 756: {'lr': 0.000189, 'samples': 387584, 'steps': 756, 'loss/train': 6.61212682723999} 02/24/2022 02:32:10 - INFO - codeparrot_training - Step 757: {'lr': 0.00018925, 'samples': 388096, 'steps': 757, 'loss/train': 7.189175128936768} 02/24/2022 02:32:16 - INFO - codeparrot_training - Step 758: {'lr': 0.0001895, 'samples': 388608, 'steps': 758, 'loss/train': 6.407525539398193} 02/24/2022 02:32:19 - INFO - codeparrot_training - Step 759: {'lr': 0.00018975, 'samples': 389120, 'steps': 759, 'loss/train': 6.6135945320129395} 02/24/2022 02:32:26 - INFO - codeparrot_training - Step 760: {'lr': 0.00019, 'samples': 389632, 'steps': 760, 'loss/train': 5.9673566818237305} 02/24/2022 02:32:29 - INFO - codeparrot_training - Step 761: {'lr': 0.00019025000000000002, 'samples': 390144, 'steps': 761, 'loss/train': 5.967684268951416} 02/24/2022 02:32:34 - INFO - codeparrot_training - Step 762: {'lr': 0.0001905, 'samples': 390656, 'steps': 762, 'loss/train': 6.6024322509765625} 02/24/2022 02:32:38 - INFO - codeparrot_training - Step 763: {'lr': 0.00019075, 'samples': 391168, 'steps': 763, 'loss/train': 6.407318592071533} 02/24/2022 02:32:43 - INFO - codeparrot_training - Step 764: {'lr': 0.000191, 'samples': 391680, 'steps': 764, 'loss/train': 5.6461181640625} 02/24/2022 02:32:47 - INFO - codeparrot_training - Step 765: {'lr': 0.00019125000000000001, 'samples': 392192, 'steps': 765, 'loss/train': 6.097337245941162} 02/24/2022 02:32:52 - INFO - codeparrot_training - Step 766: {'lr': 0.00019150000000000002, 'samples': 392704, 'steps': 766, 'loss/train': 5.59611701965332} 02/24/2022 02:32:56 - INFO - codeparrot_training - Step 767: {'lr': 0.00019175, 'samples': 393216, 'steps': 767, 'loss/train': 5.815179347991943} 02/24/2022 02:33:01 - INFO - codeparrot_training - Step 768: {'lr': 0.000192, 'samples': 393728, 'steps': 768, 'loss/train': 5.541304111480713} 02/24/2022 02:33:05 - INFO - codeparrot_training - Step 769: {'lr': 0.00019225, 'samples': 394240, 'steps': 769, 'loss/train': 5.901695728302002} 02/24/2022 02:33:11 - INFO - codeparrot_training - Step 770: {'lr': 0.00019250000000000002, 'samples': 394752, 'steps': 770, 'loss/train': 6.1426472663879395} 02/24/2022 02:33:14 - INFO - codeparrot_training - Step 771: {'lr': 0.00019275, 'samples': 395264, 'steps': 771, 'loss/train': 6.149904727935791} 02/24/2022 02:33:20 - INFO - codeparrot_training - Step 772: {'lr': 0.000193, 'samples': 395776, 'steps': 772, 'loss/train': 5.976991176605225} 02/24/2022 02:33:25 - INFO - codeparrot_training - Step 773: {'lr': 0.00019325, 'samples': 396288, 'steps': 773, 'loss/train': 6.669251441955566} 02/24/2022 02:33:29 - INFO - codeparrot_training - Step 774: {'lr': 0.00019350000000000001, 'samples': 396800, 'steps': 774, 'loss/train': 6.111331462860107} 02/24/2022 02:33:34 - INFO - codeparrot_training - Step 775: {'lr': 0.00019375000000000002, 'samples': 397312, 'steps': 775, 'loss/train': 6.806671142578125} 02/24/2022 02:33:38 - INFO - codeparrot_training - Step 776: {'lr': 0.000194, 'samples': 397824, 'steps': 776, 'loss/train': 6.131837368011475} 02/24/2022 02:33:43 - INFO - codeparrot_training - Step 777: {'lr': 0.00019425, 'samples': 398336, 'steps': 777, 'loss/train': 5.915597915649414} 02/24/2022 02:33:47 - INFO - codeparrot_training - Step 778: {'lr': 0.0001945, 'samples': 398848, 'steps': 778, 'loss/train': 6.392073154449463} 02/24/2022 02:33:53 - INFO - codeparrot_training - Step 779: {'lr': 0.00019475000000000002, 'samples': 399360, 'steps': 779, 'loss/train': 7.551222324371338} 02/24/2022 02:33:56 - INFO - codeparrot_training - Step 780: {'lr': 0.00019500000000000002, 'samples': 399872, 'steps': 780, 'loss/train': 6.118439674377441} 02/24/2022 02:34:02 - INFO - codeparrot_training - Step 781: {'lr': 0.00019525, 'samples': 400384, 'steps': 781, 'loss/train': 6.616265296936035} 02/24/2022 02:34:05 - INFO - codeparrot_training - Step 782: {'lr': 0.0001955, 'samples': 400896, 'steps': 782, 'loss/train': 5.312283515930176} 02/24/2022 02:34:11 - INFO - codeparrot_training - Step 783: {'lr': 0.00019575000000000001, 'samples': 401408, 'steps': 783, 'loss/train': 3.7072436809539795} 02/24/2022 02:34:14 - INFO - codeparrot_training - Step 784: {'lr': 0.00019600000000000002, 'samples': 401920, 'steps': 784, 'loss/train': 6.311213493347168} 02/24/2022 02:34:20 - INFO - codeparrot_training - Step 785: {'lr': 0.00019625, 'samples': 402432, 'steps': 785, 'loss/train': 6.558773994445801} 02/24/2022 02:34:23 - INFO - codeparrot_training - Step 786: {'lr': 0.0001965, 'samples': 402944, 'steps': 786, 'loss/train': 5.938821792602539} 02/24/2022 02:34:29 - INFO - codeparrot_training - Step 787: {'lr': 0.00019675, 'samples': 403456, 'steps': 787, 'loss/train': 5.922603130340576} 02/24/2022 02:34:32 - INFO - codeparrot_training - Step 788: {'lr': 0.00019700000000000002, 'samples': 403968, 'steps': 788, 'loss/train': 8.191448211669922} 02/24/2022 02:34:38 - INFO - codeparrot_training - Step 789: {'lr': 0.00019725000000000002, 'samples': 404480, 'steps': 789, 'loss/train': 5.338706016540527} 02/24/2022 02:34:42 - INFO - codeparrot_training - Step 790: {'lr': 0.0001975, 'samples': 404992, 'steps': 790, 'loss/train': 7.5210161209106445} 02/24/2022 02:34:47 - INFO - codeparrot_training - Step 791: {'lr': 0.00019775, 'samples': 405504, 'steps': 791, 'loss/train': 7.093896865844727} 02/24/2022 02:34:51 - INFO - codeparrot_training - Step 792: {'lr': 0.00019800000000000002, 'samples': 406016, 'steps': 792, 'loss/train': 3.117628574371338} 02/24/2022 02:34:56 - INFO - codeparrot_training - Step 793: {'lr': 0.00019825000000000002, 'samples': 406528, 'steps': 793, 'loss/train': 5.884690284729004} 02/24/2022 02:34:59 - INFO - codeparrot_training - Step 794: {'lr': 0.00019850000000000003, 'samples': 407040, 'steps': 794, 'loss/train': 5.920525074005127} 02/24/2022 02:35:05 - INFO - codeparrot_training - Step 795: {'lr': 0.00019875, 'samples': 407552, 'steps': 795, 'loss/train': 6.231931209564209} 02/24/2022 02:35:08 - INFO - codeparrot_training - Step 796: {'lr': 0.000199, 'samples': 408064, 'steps': 796, 'loss/train': 5.558359146118164} 02/24/2022 02:35:14 - INFO - codeparrot_training - Step 797: {'lr': 0.00019925000000000002, 'samples': 408576, 'steps': 797, 'loss/train': 6.473911285400391} 02/24/2022 02:35:17 - INFO - codeparrot_training - Step 798: {'lr': 0.00019950000000000002, 'samples': 409088, 'steps': 798, 'loss/train': 6.409696578979492} 02/24/2022 02:35:23 - INFO - codeparrot_training - Step 799: {'lr': 0.00019975, 'samples': 409600, 'steps': 799, 'loss/train': 6.1711106300354} 02/24/2022 02:35:26 - INFO - codeparrot_training - Step 800: {'lr': 0.0002, 'samples': 410112, 'steps': 800, 'loss/train': 6.187984943389893} 02/24/2022 02:35:32 - INFO - codeparrot_training - Step 801: {'lr': 0.00020025000000000002, 'samples': 410624, 'steps': 801, 'loss/train': 7.001424312591553} 02/24/2022 02:35:35 - INFO - codeparrot_training - Step 802: {'lr': 0.00020050000000000002, 'samples': 411136, 'steps': 802, 'loss/train': 2.725290060043335} 02/24/2022 02:35:41 - INFO - codeparrot_training - Step 803: {'lr': 0.00020075000000000003, 'samples': 411648, 'steps': 803, 'loss/train': 6.731625556945801} 02/24/2022 02:35:44 - INFO - codeparrot_training - Step 804: {'lr': 0.000201, 'samples': 412160, 'steps': 804, 'loss/train': 6.009284019470215} 02/24/2022 02:35:50 - INFO - codeparrot_training - Step 805: {'lr': 0.00020125, 'samples': 412672, 'steps': 805, 'loss/train': 6.596358776092529} 02/24/2022 02:35:54 - INFO - codeparrot_training - Step 806: {'lr': 0.00020150000000000002, 'samples': 413184, 'steps': 806, 'loss/train': 6.380502700805664} 02/24/2022 02:35:59 - INFO - codeparrot_training - Step 807: {'lr': 0.00020175000000000003, 'samples': 413696, 'steps': 807, 'loss/train': 7.420418739318848} 02/24/2022 02:36:02 - INFO - codeparrot_training - Step 808: {'lr': 0.000202, 'samples': 414208, 'steps': 808, 'loss/train': 6.865699291229248} 02/24/2022 02:36:08 - INFO - codeparrot_training - Step 809: {'lr': 0.00020225, 'samples': 414720, 'steps': 809, 'loss/train': 6.789573669433594} 02/24/2022 02:36:11 - INFO - codeparrot_training - Step 810: {'lr': 0.00020250000000000002, 'samples': 415232, 'steps': 810, 'loss/train': 5.445998668670654} 02/24/2022 02:36:17 - INFO - codeparrot_training - Step 811: {'lr': 0.00020275000000000002, 'samples': 415744, 'steps': 811, 'loss/train': 9.829699516296387} 02/24/2022 02:36:21 - INFO - codeparrot_training - Step 812: {'lr': 0.00020300000000000003, 'samples': 416256, 'steps': 812, 'loss/train': 6.86912202835083} 02/24/2022 02:36:26 - INFO - codeparrot_training - Step 813: {'lr': 0.00020324999999999998, 'samples': 416768, 'steps': 813, 'loss/train': 6.221214771270752} 02/24/2022 02:36:29 - INFO - codeparrot_training - Step 814: {'lr': 0.00020349999999999999, 'samples': 417280, 'steps': 814, 'loss/train': 8.142473220825195} 02/24/2022 02:36:35 - INFO - codeparrot_training - Step 815: {'lr': 0.00020375, 'samples': 417792, 'steps': 815, 'loss/train': 5.884995937347412} 02/24/2022 02:36:38 - INFO - codeparrot_training - Step 816: {'lr': 0.000204, 'samples': 418304, 'steps': 816, 'loss/train': 5.674224853515625} 02/24/2022 02:36:45 - INFO - codeparrot_training - Step 817: {'lr': 0.00020425, 'samples': 418816, 'steps': 817, 'loss/train': 5.738276958465576} 02/24/2022 02:36:48 - INFO - codeparrot_training - Step 818: {'lr': 0.00020449999999999998, 'samples': 419328, 'steps': 818, 'loss/train': 2.4321036338806152} 02/24/2022 02:36:53 - INFO - codeparrot_training - Step 819: {'lr': 0.00020475, 'samples': 419840, 'steps': 819, 'loss/train': 6.171673774719238} 02/24/2022 02:36:57 - INFO - codeparrot_training - Step 820: {'lr': 0.000205, 'samples': 420352, 'steps': 820, 'loss/train': 6.443540096282959} 02/24/2022 02:37:02 - INFO - codeparrot_training - Step 821: {'lr': 0.00020525, 'samples': 420864, 'steps': 821, 'loss/train': 5.149608135223389} 02/24/2022 02:37:08 - INFO - codeparrot_training - Step 822: {'lr': 0.00020549999999999998, 'samples': 421376, 'steps': 822, 'loss/train': 5.614904880523682} 02/24/2022 02:37:11 - INFO - codeparrot_training - Step 823: {'lr': 0.00020575, 'samples': 421888, 'steps': 823, 'loss/train': 8.65012264251709} 02/24/2022 02:37:17 - INFO - codeparrot_training - Step 824: {'lr': 0.000206, 'samples': 422400, 'steps': 824, 'loss/train': 6.405791282653809} 02/24/2022 02:37:20 - INFO - codeparrot_training - Step 825: {'lr': 0.00020625, 'samples': 422912, 'steps': 825, 'loss/train': 5.550499439239502} 02/24/2022 02:37:27 - INFO - codeparrot_training - Step 826: {'lr': 0.0002065, 'samples': 423424, 'steps': 826, 'loss/train': 6.045300483703613} 02/24/2022 02:37:30 - INFO - codeparrot_training - Step 827: {'lr': 0.00020674999999999998, 'samples': 423936, 'steps': 827, 'loss/train': 6.330199241638184} 02/24/2022 02:37:35 - INFO - codeparrot_training - Step 828: {'lr': 0.000207, 'samples': 424448, 'steps': 828, 'loss/train': 4.9262375831604} 02/24/2022 02:37:39 - INFO - codeparrot_training - Step 829: {'lr': 0.00020725, 'samples': 424960, 'steps': 829, 'loss/train': 5.7279486656188965} 02/24/2022 02:37:44 - INFO - codeparrot_training - Step 830: {'lr': 0.0002075, 'samples': 425472, 'steps': 830, 'loss/train': 6.108678817749023} 02/24/2022 02:37:48 - INFO - codeparrot_training - Step 831: {'lr': 0.00020774999999999998, 'samples': 425984, 'steps': 831, 'loss/train': 5.691193580627441} 02/24/2022 02:37:53 - INFO - codeparrot_training - Step 832: {'lr': 0.000208, 'samples': 426496, 'steps': 832, 'loss/train': 6.178577899932861} 02/24/2022 02:37:57 - INFO - codeparrot_training - Step 833: {'lr': 0.00020825, 'samples': 427008, 'steps': 833, 'loss/train': 6.508179187774658} 02/24/2022 02:38:02 - INFO - codeparrot_training - Step 834: {'lr': 0.0002085, 'samples': 427520, 'steps': 834, 'loss/train': 5.967754364013672} 02/24/2022 02:38:06 - INFO - codeparrot_training - Step 835: {'lr': 0.00020875, 'samples': 428032, 'steps': 835, 'loss/train': 4.665596961975098} 02/24/2022 02:38:12 - INFO - codeparrot_training - Step 836: {'lr': 0.00020899999999999998, 'samples': 428544, 'steps': 836, 'loss/train': 6.464814186096191} 02/24/2022 02:38:15 - INFO - codeparrot_training - Step 837: {'lr': 0.00020925, 'samples': 429056, 'steps': 837, 'loss/train': 6.114542007446289} 02/24/2022 02:38:21 - INFO - codeparrot_training - Step 838: {'lr': 0.0002095, 'samples': 429568, 'steps': 838, 'loss/train': 7.669986724853516} 02/24/2022 02:38:24 - INFO - codeparrot_training - Step 839: {'lr': 0.00020975, 'samples': 430080, 'steps': 839, 'loss/train': 6.398980140686035} 02/24/2022 02:38:28 - INFO - codeparrot_training - Step 840: {'lr': 0.00021, 'samples': 430592, 'steps': 840, 'loss/train': 7.4006195068359375} 02/24/2022 02:38:33 - INFO - codeparrot_training - Step 841: {'lr': 0.00021025, 'samples': 431104, 'steps': 841, 'loss/train': 6.521213531494141} 02/24/2022 02:38:37 - INFO - codeparrot_training - Step 842: {'lr': 0.0002105, 'samples': 431616, 'steps': 842, 'loss/train': 6.6276726722717285} 02/24/2022 02:38:42 - INFO - codeparrot_training - Step 843: {'lr': 0.00021075, 'samples': 432128, 'steps': 843, 'loss/train': 6.0089006423950195} 02/24/2022 02:38:46 - INFO - codeparrot_training - Step 844: {'lr': 0.000211, 'samples': 432640, 'steps': 844, 'loss/train': 6.256646633148193} 02/24/2022 02:38:51 - INFO - codeparrot_training - Step 845: {'lr': 0.00021124999999999998, 'samples': 433152, 'steps': 845, 'loss/train': 6.229872226715088} 02/24/2022 02:38:57 - INFO - codeparrot_training - Step 846: {'lr': 0.0002115, 'samples': 433664, 'steps': 846, 'loss/train': 6.033138275146484} 02/24/2022 02:39:01 - INFO - codeparrot_training - Step 847: {'lr': 0.00021175, 'samples': 434176, 'steps': 847, 'loss/train': 5.928333282470703} 02/24/2022 02:39:06 - INFO - codeparrot_training - Step 848: {'lr': 0.000212, 'samples': 434688, 'steps': 848, 'loss/train': 2.4343342781066895} 02/24/2022 02:39:10 - INFO - codeparrot_training - Step 849: {'lr': 0.00021225, 'samples': 435200, 'steps': 849, 'loss/train': 5.912009239196777} 02/24/2022 02:39:13 - INFO - codeparrot_training - Step 850: {'lr': 0.0002125, 'samples': 435712, 'steps': 850, 'loss/train': 6.8696136474609375} 02/24/2022 02:39:19 - INFO - codeparrot_training - Step 851: {'lr': 0.00021275, 'samples': 436224, 'steps': 851, 'loss/train': 9.76145076751709} 02/24/2022 02:39:22 - INFO - codeparrot_training - Step 852: {'lr': 0.000213, 'samples': 436736, 'steps': 852, 'loss/train': 5.8401713371276855} 02/24/2022 02:39:28 - INFO - codeparrot_training - Step 853: {'lr': 0.00021325, 'samples': 437248, 'steps': 853, 'loss/train': 5.45580530166626} 02/24/2022 02:39:31 - INFO - codeparrot_training - Step 854: {'lr': 0.0002135, 'samples': 437760, 'steps': 854, 'loss/train': 5.753581523895264} 02/24/2022 02:39:37 - INFO - codeparrot_training - Step 855: {'lr': 0.00021375, 'samples': 438272, 'steps': 855, 'loss/train': 7.0582475662231445} 02/24/2022 02:39:40 - INFO - codeparrot_training - Step 856: {'lr': 0.000214, 'samples': 438784, 'steps': 856, 'loss/train': 5.507960796356201} 02/24/2022 02:39:46 - INFO - codeparrot_training - Step 857: {'lr': 0.00021425, 'samples': 439296, 'steps': 857, 'loss/train': 5.333662509918213} 02/24/2022 02:39:49 - INFO - codeparrot_training - Step 858: {'lr': 0.0002145, 'samples': 439808, 'steps': 858, 'loss/train': 5.536602020263672} 02/24/2022 02:39:55 - INFO - codeparrot_training - Step 859: {'lr': 0.00021475, 'samples': 440320, 'steps': 859, 'loss/train': 5.66477632522583} 02/24/2022 02:39:58 - INFO - codeparrot_training - Step 860: {'lr': 0.000215, 'samples': 440832, 'steps': 860, 'loss/train': 5.88394021987915} 02/24/2022 02:40:04 - INFO - codeparrot_training - Step 861: {'lr': 0.00021525, 'samples': 441344, 'steps': 861, 'loss/train': 7.011894226074219} 02/24/2022 02:40:10 - INFO - codeparrot_training - Step 862: {'lr': 0.0002155, 'samples': 441856, 'steps': 862, 'loss/train': 6.413607120513916} 02/24/2022 02:40:13 - INFO - codeparrot_training - Step 863: {'lr': 0.00021575, 'samples': 442368, 'steps': 863, 'loss/train': 6.333065986633301} 02/24/2022 02:40:19 - INFO - codeparrot_training - Step 864: {'lr': 0.000216, 'samples': 442880, 'steps': 864, 'loss/train': 7.09116268157959} 02/24/2022 02:40:22 - INFO - codeparrot_training - Step 865: {'lr': 0.00021625, 'samples': 443392, 'steps': 865, 'loss/train': 6.186008930206299} 02/24/2022 02:40:28 - INFO - codeparrot_training - Step 866: {'lr': 0.0002165, 'samples': 443904, 'steps': 866, 'loss/train': 6.317468643188477} 02/24/2022 02:40:31 - INFO - codeparrot_training - Step 867: {'lr': 0.00021675, 'samples': 444416, 'steps': 867, 'loss/train': 5.776604175567627} 02/24/2022 02:40:36 - INFO - codeparrot_training - Step 868: {'lr': 0.00021700000000000002, 'samples': 444928, 'steps': 868, 'loss/train': 5.373128890991211} 02/24/2022 02:40:40 - INFO - codeparrot_training - Step 869: {'lr': 0.00021725, 'samples': 445440, 'steps': 869, 'loss/train': 6.335923194885254} 02/24/2022 02:40:45 - INFO - codeparrot_training - Step 870: {'lr': 0.0002175, 'samples': 445952, 'steps': 870, 'loss/train': 6.347319602966309} 02/24/2022 02:40:49 - INFO - codeparrot_training - Step 871: {'lr': 0.00021775, 'samples': 446464, 'steps': 871, 'loss/train': 3.7130682468414307} 02/24/2022 02:40:55 - INFO - codeparrot_training - Step 872: {'lr': 0.000218, 'samples': 446976, 'steps': 872, 'loss/train': 8.389607429504395} 02/24/2022 02:40:58 - INFO - codeparrot_training - Step 873: {'lr': 0.00021825, 'samples': 447488, 'steps': 873, 'loss/train': 6.073687553405762} 02/24/2022 02:41:04 - INFO - codeparrot_training - Step 874: {'lr': 0.0002185, 'samples': 448000, 'steps': 874, 'loss/train': 6.788509845733643} 02/24/2022 02:41:07 - INFO - codeparrot_training - Step 875: {'lr': 0.00021875, 'samples': 448512, 'steps': 875, 'loss/train': 6.623488426208496} 02/24/2022 02:41:13 - INFO - codeparrot_training - Step 876: {'lr': 0.000219, 'samples': 449024, 'steps': 876, 'loss/train': 3.808100938796997} 02/24/2022 02:41:16 - INFO - codeparrot_training - Step 877: {'lr': 0.00021925000000000002, 'samples': 449536, 'steps': 877, 'loss/train': 5.787219047546387} 02/24/2022 02:41:22 - INFO - codeparrot_training - Step 878: {'lr': 0.0002195, 'samples': 450048, 'steps': 878, 'loss/train': 5.961319446563721} 02/24/2022 02:41:25 - INFO - codeparrot_training - Step 879: {'lr': 0.00021975, 'samples': 450560, 'steps': 879, 'loss/train': 6.06191873550415} 02/24/2022 02:41:31 - INFO - codeparrot_training - Step 880: {'lr': 0.00022, 'samples': 451072, 'steps': 880, 'loss/train': 5.730699062347412} 02/24/2022 02:41:34 - INFO - codeparrot_training - Step 881: {'lr': 0.00022025000000000001, 'samples': 451584, 'steps': 881, 'loss/train': 6.804256916046143} 02/24/2022 02:41:40 - INFO - codeparrot_training - Step 882: {'lr': 0.0002205, 'samples': 452096, 'steps': 882, 'loss/train': 5.483248233795166} 02/24/2022 02:41:43 - INFO - codeparrot_training - Step 883: {'lr': 0.00022075, 'samples': 452608, 'steps': 883, 'loss/train': 5.957112789154053} 02/24/2022 02:41:49 - INFO - codeparrot_training - Step 884: {'lr': 0.000221, 'samples': 453120, 'steps': 884, 'loss/train': 5.911479473114014} 02/24/2022 02:41:52 - INFO - codeparrot_training - Step 885: {'lr': 0.00022125, 'samples': 453632, 'steps': 885, 'loss/train': 5.84247350692749} 02/24/2022 02:41:58 - INFO - codeparrot_training - Step 886: {'lr': 0.00022150000000000002, 'samples': 454144, 'steps': 886, 'loss/train': 6.338929176330566} 02/24/2022 02:42:01 - INFO - codeparrot_training - Step 887: {'lr': 0.00022175, 'samples': 454656, 'steps': 887, 'loss/train': 7.415934085845947} 02/24/2022 02:42:07 - INFO - codeparrot_training - Step 888: {'lr': 0.000222, 'samples': 455168, 'steps': 888, 'loss/train': 6.88756799697876} 02/24/2022 02:42:10 - INFO - codeparrot_training - Step 889: {'lr': 0.00022225, 'samples': 455680, 'steps': 889, 'loss/train': 5.849971294403076} 02/24/2022 02:42:16 - INFO - codeparrot_training - Step 890: {'lr': 0.00022250000000000001, 'samples': 456192, 'steps': 890, 'loss/train': 6.605902671813965} 02/24/2022 02:42:19 - INFO - codeparrot_training - Step 891: {'lr': 0.00022275000000000002, 'samples': 456704, 'steps': 891, 'loss/train': 5.789719581604004} 02/24/2022 02:42:25 - INFO - codeparrot_training - Step 892: {'lr': 0.000223, 'samples': 457216, 'steps': 892, 'loss/train': 7.157408237457275} 02/24/2022 02:42:29 - INFO - codeparrot_training - Step 893: {'lr': 0.00022325, 'samples': 457728, 'steps': 893, 'loss/train': 5.905867099761963} 02/24/2022 02:42:34 - INFO - codeparrot_training - Step 894: {'lr': 0.0002235, 'samples': 458240, 'steps': 894, 'loss/train': 5.780603408813477} 02/24/2022 02:42:37 - INFO - codeparrot_training - Step 895: {'lr': 0.00022375000000000002, 'samples': 458752, 'steps': 895, 'loss/train': 5.804656028747559} 02/24/2022 02:42:43 - INFO - codeparrot_training - Step 896: {'lr': 0.000224, 'samples': 459264, 'steps': 896, 'loss/train': 5.820722579956055} 02/24/2022 02:42:48 - INFO - codeparrot_training - Step 897: {'lr': 0.00022425, 'samples': 459776, 'steps': 897, 'loss/train': 6.14693021774292} 02/24/2022 02:42:52 - INFO - codeparrot_training - Step 898: {'lr': 0.0002245, 'samples': 460288, 'steps': 898, 'loss/train': 6.593510627746582} 02/24/2022 02:42:57 - INFO - codeparrot_training - Step 899: {'lr': 0.00022475000000000001, 'samples': 460800, 'steps': 899, 'loss/train': 5.965980052947998} 02/24/2022 02:43:01 - INFO - codeparrot_training - Step 900: {'lr': 0.00022500000000000002, 'samples': 461312, 'steps': 900, 'loss/train': 5.33210563659668} 02/24/2022 02:43:06 - INFO - codeparrot_training - Step 901: {'lr': 0.00022525, 'samples': 461824, 'steps': 901, 'loss/train': 5.6558685302734375} 02/24/2022 02:43:10 - INFO - codeparrot_training - Step 902: {'lr': 0.0002255, 'samples': 462336, 'steps': 902, 'loss/train': 5.435738563537598} 02/24/2022 02:43:15 - INFO - codeparrot_training - Step 903: {'lr': 0.00022575, 'samples': 462848, 'steps': 903, 'loss/train': 6.804921627044678} 02/24/2022 02:43:19 - INFO - codeparrot_training - Step 904: {'lr': 0.00022600000000000002, 'samples': 463360, 'steps': 904, 'loss/train': 8.13878345489502} 02/24/2022 02:43:24 - INFO - codeparrot_training - Step 905: {'lr': 0.00022625000000000002, 'samples': 463872, 'steps': 905, 'loss/train': 6.534748077392578} 02/24/2022 02:43:28 - INFO - codeparrot_training - Step 906: {'lr': 0.0002265, 'samples': 464384, 'steps': 906, 'loss/train': 5.649563312530518} 02/24/2022 02:43:34 - INFO - codeparrot_training - Step 907: {'lr': 0.00022675, 'samples': 464896, 'steps': 907, 'loss/train': 6.934625625610352} 02/24/2022 02:43:37 - INFO - codeparrot_training - Step 908: {'lr': 0.00022700000000000002, 'samples': 465408, 'steps': 908, 'loss/train': 5.989643573760986} 02/24/2022 02:43:43 - INFO - codeparrot_training - Step 909: {'lr': 0.00022725000000000002, 'samples': 465920, 'steps': 909, 'loss/train': 6.277463436126709} 02/24/2022 02:43:46 - INFO - codeparrot_training - Step 910: {'lr': 0.0002275, 'samples': 466432, 'steps': 910, 'loss/train': 6.045421600341797} 02/24/2022 02:43:51 - INFO - codeparrot_training - Step 911: {'lr': 0.00022775, 'samples': 466944, 'steps': 911, 'loss/train': 5.83571720123291} 02/24/2022 02:43:55 - INFO - codeparrot_training - Step 912: {'lr': 0.000228, 'samples': 467456, 'steps': 912, 'loss/train': 4.9673309326171875} 02/24/2022 02:44:00 - INFO - codeparrot_training - Step 913: {'lr': 0.00022825000000000002, 'samples': 467968, 'steps': 913, 'loss/train': 5.016026020050049} 02/24/2022 02:44:04 - INFO - codeparrot_training - Step 914: {'lr': 0.00022850000000000002, 'samples': 468480, 'steps': 914, 'loss/train': 5.969403266906738} 02/24/2022 02:44:09 - INFO - codeparrot_training - Step 915: {'lr': 0.00022875, 'samples': 468992, 'steps': 915, 'loss/train': 5.889241695404053} 02/24/2022 02:44:13 - INFO - codeparrot_training - Step 916: {'lr': 0.000229, 'samples': 469504, 'steps': 916, 'loss/train': 5.977917671203613} 02/24/2022 02:44:19 - INFO - codeparrot_training - Step 917: {'lr': 0.00022925000000000002, 'samples': 470016, 'steps': 917, 'loss/train': 6.2948713302612305} 02/24/2022 02:44:22 - INFO - codeparrot_training - Step 918: {'lr': 0.00022950000000000002, 'samples': 470528, 'steps': 918, 'loss/train': 6.78879451751709} 02/24/2022 02:44:28 - INFO - codeparrot_training - Step 919: {'lr': 0.00022975000000000003, 'samples': 471040, 'steps': 919, 'loss/train': 6.277073860168457} 02/24/2022 02:44:31 - INFO - codeparrot_training - Step 920: {'lr': 0.00023, 'samples': 471552, 'steps': 920, 'loss/train': 6.629997730255127} 02/24/2022 02:44:37 - INFO - codeparrot_training - Step 921: {'lr': 0.00023025, 'samples': 472064, 'steps': 921, 'loss/train': 5.085743427276611} 02/24/2022 02:44:40 - INFO - codeparrot_training - Step 922: {'lr': 0.00023050000000000002, 'samples': 472576, 'steps': 922, 'loss/train': 5.860753536224365} 02/24/2022 02:44:46 - INFO - codeparrot_training - Step 923: {'lr': 0.00023075000000000003, 'samples': 473088, 'steps': 923, 'loss/train': 6.716079235076904} 02/24/2022 02:44:49 - INFO - codeparrot_training - Step 924: {'lr': 0.000231, 'samples': 473600, 'steps': 924, 'loss/train': 6.65992546081543} 02/24/2022 02:44:54 - INFO - codeparrot_training - Step 925: {'lr': 0.00023125, 'samples': 474112, 'steps': 925, 'loss/train': 6.191816329956055} 02/24/2022 02:44:58 - INFO - codeparrot_training - Step 926: {'lr': 0.00023150000000000002, 'samples': 474624, 'steps': 926, 'loss/train': 5.812952518463135} 02/24/2022 02:45:04 - INFO - codeparrot_training - Step 927: {'lr': 0.00023175000000000002, 'samples': 475136, 'steps': 927, 'loss/train': 4.18288516998291} 02/24/2022 02:45:08 - INFO - codeparrot_training - Step 928: {'lr': 0.00023200000000000003, 'samples': 475648, 'steps': 928, 'loss/train': 5.7861175537109375} 02/24/2022 02:45:13 - INFO - codeparrot_training - Step 929: {'lr': 0.00023225, 'samples': 476160, 'steps': 929, 'loss/train': 7.031640529632568} 02/24/2022 02:45:16 - INFO - codeparrot_training - Step 930: {'lr': 0.0002325, 'samples': 476672, 'steps': 930, 'loss/train': 6.2350239753723145} 02/24/2022 02:45:22 - INFO - codeparrot_training - Step 931: {'lr': 0.00023275000000000002, 'samples': 477184, 'steps': 931, 'loss/train': 5.944024562835693} 02/24/2022 02:45:25 - INFO - codeparrot_training - Step 932: {'lr': 0.00023300000000000003, 'samples': 477696, 'steps': 932, 'loss/train': 4.8677167892456055} 02/24/2022 02:45:31 - INFO - codeparrot_training - Step 933: {'lr': 0.00023325, 'samples': 478208, 'steps': 933, 'loss/train': 6.226642608642578} 02/24/2022 02:45:35 - INFO - codeparrot_training - Step 934: {'lr': 0.0002335, 'samples': 478720, 'steps': 934, 'loss/train': 5.019423484802246} 02/24/2022 02:45:40 - INFO - codeparrot_training - Step 935: {'lr': 0.00023375000000000002, 'samples': 479232, 'steps': 935, 'loss/train': 5.15052604675293} 02/24/2022 02:45:43 - INFO - codeparrot_training - Step 936: {'lr': 0.00023400000000000002, 'samples': 479744, 'steps': 936, 'loss/train': 6.223147392272949} 02/24/2022 02:45:49 - INFO - codeparrot_training - Step 937: {'lr': 0.00023425000000000003, 'samples': 480256, 'steps': 937, 'loss/train': 5.998800754547119} 02/24/2022 02:45:52 - INFO - codeparrot_training - Step 938: {'lr': 0.00023449999999999998, 'samples': 480768, 'steps': 938, 'loss/train': 5.864666938781738} 02/24/2022 02:45:58 - INFO - codeparrot_training - Step 939: {'lr': 0.00023475, 'samples': 481280, 'steps': 939, 'loss/train': 5.786550998687744} 02/24/2022 02:46:02 - INFO - codeparrot_training - Step 940: {'lr': 0.000235, 'samples': 481792, 'steps': 940, 'loss/train': 5.671061038970947} 02/24/2022 02:46:07 - INFO - codeparrot_training - Step 941: {'lr': 0.00023525, 'samples': 482304, 'steps': 941, 'loss/train': 5.6021623611450195} 02/24/2022 02:46:11 - INFO - codeparrot_training - Step 942: {'lr': 0.0002355, 'samples': 482816, 'steps': 942, 'loss/train': 5.248650550842285} 02/24/2022 02:46:16 - INFO - codeparrot_training - Step 943: {'lr': 0.00023574999999999998, 'samples': 483328, 'steps': 943, 'loss/train': 5.5844831466674805} 02/24/2022 02:46:20 - INFO - codeparrot_training - Step 944: {'lr': 0.000236, 'samples': 483840, 'steps': 944, 'loss/train': 6.027729034423828} 02/24/2022 02:46:25 - INFO - codeparrot_training - Step 945: {'lr': 0.00023625, 'samples': 484352, 'steps': 945, 'loss/train': 6.658106327056885} 02/24/2022 02:46:29 - INFO - codeparrot_training - Step 946: {'lr': 0.0002365, 'samples': 484864, 'steps': 946, 'loss/train': 6.287262916564941} 02/24/2022 02:46:34 - INFO - codeparrot_training - Step 947: {'lr': 0.00023674999999999998, 'samples': 485376, 'steps': 947, 'loss/train': 4.877590656280518} 02/24/2022 02:46:40 - INFO - codeparrot_training - Step 948: {'lr': 0.000237, 'samples': 485888, 'steps': 948, 'loss/train': 6.26812744140625} 02/24/2022 02:46:44 - INFO - codeparrot_training - Step 949: {'lr': 0.00023725, 'samples': 486400, 'steps': 949, 'loss/train': 6.118881702423096} 02/24/2022 02:46:49 - INFO - codeparrot_training - Step 950: {'lr': 0.0002375, 'samples': 486912, 'steps': 950, 'loss/train': 5.752974510192871} 02/24/2022 02:46:53 - INFO - codeparrot_training - Step 951: {'lr': 0.00023775, 'samples': 487424, 'steps': 951, 'loss/train': 6.376278400421143} 02/24/2022 02:46:58 - INFO - codeparrot_training - Step 952: {'lr': 0.00023799999999999998, 'samples': 487936, 'steps': 952, 'loss/train': 5.967774391174316} 02/24/2022 02:47:01 - INFO - codeparrot_training - Step 953: {'lr': 0.00023825, 'samples': 488448, 'steps': 953, 'loss/train': 6.946618556976318} 02/24/2022 02:47:07 - INFO - codeparrot_training - Step 954: {'lr': 0.0002385, 'samples': 488960, 'steps': 954, 'loss/train': 5.575984477996826} 02/24/2022 02:47:10 - INFO - codeparrot_training - Step 955: {'lr': 0.00023875, 'samples': 489472, 'steps': 955, 'loss/train': 5.178292274475098} 02/24/2022 02:47:16 - INFO - codeparrot_training - Step 956: {'lr': 0.00023899999999999998, 'samples': 489984, 'steps': 956, 'loss/train': 5.378644943237305} 02/24/2022 02:47:19 - INFO - codeparrot_training - Step 957: {'lr': 0.00023925, 'samples': 490496, 'steps': 957, 'loss/train': 5.6738996505737305} 02/24/2022 02:47:25 - INFO - codeparrot_training - Step 958: {'lr': 0.0002395, 'samples': 491008, 'steps': 958, 'loss/train': 5.746028423309326} 02/24/2022 02:47:28 - INFO - codeparrot_training - Step 959: {'lr': 0.00023975, 'samples': 491520, 'steps': 959, 'loss/train': 6.612490653991699} 02/24/2022 02:47:34 - INFO - codeparrot_training - Step 960: {'lr': 0.00024, 'samples': 492032, 'steps': 960, 'loss/train': 6.502299785614014} 02/24/2022 02:47:37 - INFO - codeparrot_training - Step 961: {'lr': 0.00024024999999999999, 'samples': 492544, 'steps': 961, 'loss/train': 9.734945297241211} 02/24/2022 02:47:43 - INFO - codeparrot_training - Step 962: {'lr': 0.0002405, 'samples': 493056, 'steps': 962, 'loss/train': 5.183950901031494} 02/24/2022 02:47:46 - INFO - codeparrot_training - Step 963: {'lr': 0.00024075, 'samples': 493568, 'steps': 963, 'loss/train': 5.328657150268555} 02/24/2022 02:47:52 - INFO - codeparrot_training - Step 964: {'lr': 0.000241, 'samples': 494080, 'steps': 964, 'loss/train': 6.440547466278076} 02/24/2022 02:47:56 - INFO - codeparrot_training - Step 965: {'lr': 0.00024125, 'samples': 494592, 'steps': 965, 'loss/train': 4.988674640655518} 02/24/2022 02:48:01 - INFO - codeparrot_training - Step 966: {'lr': 0.0002415, 'samples': 495104, 'steps': 966, 'loss/train': 5.711447715759277} 02/24/2022 02:48:05 - INFO - codeparrot_training - Step 967: {'lr': 0.00024175, 'samples': 495616, 'steps': 967, 'loss/train': 5.55291223526001} 02/24/2022 02:48:10 - INFO - codeparrot_training - Step 968: {'lr': 0.000242, 'samples': 496128, 'steps': 968, 'loss/train': 6.4952921867370605} 02/24/2022 02:48:13 - INFO - codeparrot_training - Step 969: {'lr': 0.00024225, 'samples': 496640, 'steps': 969, 'loss/train': 7.697079658508301} 02/24/2022 02:48:19 - INFO - codeparrot_training - Step 970: {'lr': 0.00024249999999999999, 'samples': 497152, 'steps': 970, 'loss/train': 6.5585784912109375} 02/24/2022 02:48:22 - INFO - codeparrot_training - Step 971: {'lr': 0.00024275, 'samples': 497664, 'steps': 971, 'loss/train': 5.085152626037598} 02/24/2022 02:48:28 - INFO - codeparrot_training - Step 972: {'lr': 0.000243, 'samples': 498176, 'steps': 972, 'loss/train': 5.955501556396484} 02/24/2022 02:48:31 - INFO - codeparrot_training - Step 973: {'lr': 0.00024325, 'samples': 498688, 'steps': 973, 'loss/train': 5.7349467277526855} 02/24/2022 02:48:37 - INFO - codeparrot_training - Step 974: {'lr': 0.0002435, 'samples': 499200, 'steps': 974, 'loss/train': 5.855245113372803} 02/24/2022 02:48:41 - INFO - codeparrot_training - Step 975: {'lr': 0.00024375, 'samples': 499712, 'steps': 975, 'loss/train': 6.428284645080566} 02/24/2022 02:48:46 - INFO - codeparrot_training - Step 976: {'lr': 0.000244, 'samples': 500224, 'steps': 976, 'loss/train': 4.7485737800598145} 02/24/2022 02:48:50 - INFO - codeparrot_training - Step 977: {'lr': 0.00024425, 'samples': 500736, 'steps': 977, 'loss/train': 5.669659614562988} 02/24/2022 02:48:55 - INFO - codeparrot_training - Step 978: {'lr': 0.0002445, 'samples': 501248, 'steps': 978, 'loss/train': 3.911973476409912} 02/24/2022 02:48:59 - INFO - codeparrot_training - Step 979: {'lr': 0.00024475, 'samples': 501760, 'steps': 979, 'loss/train': 6.481731414794922} 02/24/2022 02:49:04 - INFO - codeparrot_training - Step 980: {'lr': 0.000245, 'samples': 502272, 'steps': 980, 'loss/train': 6.454158782958984} 02/24/2022 02:49:08 - INFO - codeparrot_training - Step 981: {'lr': 0.00024525, 'samples': 502784, 'steps': 981, 'loss/train': 6.572815895080566} 02/24/2022 02:49:13 - INFO - codeparrot_training - Step 982: {'lr': 0.0002455, 'samples': 503296, 'steps': 982, 'loss/train': 5.37235164642334} 02/24/2022 02:49:16 - INFO - codeparrot_training - Step 983: {'lr': 0.00024575, 'samples': 503808, 'steps': 983, 'loss/train': 5.381284236907959} 02/24/2022 02:49:23 - INFO - codeparrot_training - Step 984: {'lr': 0.000246, 'samples': 504320, 'steps': 984, 'loss/train': 6.182966232299805} 02/24/2022 02:49:26 - INFO - codeparrot_training - Step 985: {'lr': 0.00024625, 'samples': 504832, 'steps': 985, 'loss/train': 4.443139553070068} 02/24/2022 02:49:32 - INFO - codeparrot_training - Step 986: {'lr': 0.00024650000000000003, 'samples': 505344, 'steps': 986, 'loss/train': 5.667211055755615} 02/24/2022 02:49:36 - INFO - codeparrot_training - Step 987: {'lr': 0.00024675, 'samples': 505856, 'steps': 987, 'loss/train': 4.547412872314453} 02/24/2022 02:49:41 - INFO - codeparrot_training - Step 988: {'lr': 0.000247, 'samples': 506368, 'steps': 988, 'loss/train': 5.689558982849121} 02/24/2022 02:49:45 - INFO - codeparrot_training - Step 989: {'lr': 0.00024725, 'samples': 506880, 'steps': 989, 'loss/train': 5.532820224761963} 02/24/2022 02:49:50 - INFO - codeparrot_training - Step 990: {'lr': 0.0002475, 'samples': 507392, 'steps': 990, 'loss/train': 6.047245979309082} 02/24/2022 02:49:54 - INFO - codeparrot_training - Step 991: {'lr': 0.00024775, 'samples': 507904, 'steps': 991, 'loss/train': 5.9628520011901855} 02/24/2022 02:49:59 - INFO - codeparrot_training - Step 992: {'lr': 0.000248, 'samples': 508416, 'steps': 992, 'loss/train': 5.37191104888916} 02/24/2022 02:50:03 - INFO - codeparrot_training - Step 993: {'lr': 0.00024825, 'samples': 508928, 'steps': 993, 'loss/train': 6.091874599456787} 02/24/2022 02:50:09 - INFO - codeparrot_training - Step 994: {'lr': 0.0002485, 'samples': 509440, 'steps': 994, 'loss/train': 5.176259994506836} 02/24/2022 02:50:12 - INFO - codeparrot_training - Step 995: {'lr': 0.00024875, 'samples': 509952, 'steps': 995, 'loss/train': 6.878044605255127} 02/24/2022 02:50:18 - INFO - codeparrot_training - Step 996: {'lr': 0.000249, 'samples': 510464, 'steps': 996, 'loss/train': 5.136017799377441} 02/24/2022 02:50:21 - INFO - codeparrot_training - Step 997: {'lr': 0.00024925, 'samples': 510976, 'steps': 997, 'loss/train': 6.837099075317383} 02/24/2022 02:50:27 - INFO - codeparrot_training - Step 998: {'lr': 0.0002495, 'samples': 511488, 'steps': 998, 'loss/train': 5.783529758453369} 02/24/2022 02:50:30 - INFO - codeparrot_training - Step 999: {'lr': 0.00024975, 'samples': 512000, 'steps': 999, 'loss/train': 6.577051639556885} 02/24/2022 02:50:30 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 02:52:17 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy * [new branch] floral-grass-11 -> floral-grass-11 02/24/2022 02:53:28 - INFO - codeparrot_training - Step 1000: {'lr': 0.00025, 'samples': 512512, 'steps': 1000, 'loss/train': 5.066281795501709} 02/24/2022 02:53:31 - INFO - codeparrot_training - Step 1001: {'lr': 0.00025025, 'samples': 513024, 'steps': 1001, 'loss/train': 4.80372953414917} 02/24/2022 02:53:37 - INFO - codeparrot_training - Step 1002: {'lr': 0.0002505, 'samples': 513536, 'steps': 1002, 'loss/train': 5.167027950286865} 02/24/2022 02:53:40 - INFO - codeparrot_training - Step 1003: {'lr': 0.00025075, 'samples': 514048, 'steps': 1003, 'loss/train': 6.6149187088012695} 02/24/2022 02:53:46 - INFO - codeparrot_training - Step 1004: {'lr': 0.00025100000000000003, 'samples': 514560, 'steps': 1004, 'loss/train': 5.51480770111084} 02/24/2022 02:53:49 - INFO - codeparrot_training - Step 1005: {'lr': 0.00025124999999999995, 'samples': 515072, 'steps': 1005, 'loss/train': 5.671285152435303} 02/24/2022 02:53:55 - INFO - codeparrot_training - Step 1006: {'lr': 0.0002515, 'samples': 515584, 'steps': 1006, 'loss/train': 5.495372772216797} 02/24/2022 02:53:58 - INFO - codeparrot_training - Step 1007: {'lr': 0.00025174999999999997, 'samples': 516096, 'steps': 1007, 'loss/train': 2.7495155334472656} 02/24/2022 02:54:03 - INFO - codeparrot_training - Step 1008: {'lr': 0.000252, 'samples': 516608, 'steps': 1008, 'loss/train': 4.522823810577393} 02/24/2022 02:54:07 - INFO - codeparrot_training - Step 1009: {'lr': 0.00025225, 'samples': 517120, 'steps': 1009, 'loss/train': 5.456524848937988} 02/24/2022 02:54:13 - INFO - codeparrot_training - Step 1010: {'lr': 0.0002525, 'samples': 517632, 'steps': 1010, 'loss/train': 5.721465110778809} 02/24/2022 02:54:16 - INFO - codeparrot_training - Step 1011: {'lr': 0.00025275, 'samples': 518144, 'steps': 1011, 'loss/train': 5.572442054748535} 02/24/2022 02:54:22 - INFO - codeparrot_training - Step 1012: {'lr': 0.000253, 'samples': 518656, 'steps': 1012, 'loss/train': 6.193021297454834} 02/24/2022 02:54:26 - INFO - codeparrot_training - Step 1013: {'lr': 0.00025325, 'samples': 519168, 'steps': 1013, 'loss/train': 5.142810344696045} 02/24/2022 02:54:31 - INFO - codeparrot_training - Step 1014: {'lr': 0.0002535, 'samples': 519680, 'steps': 1014, 'loss/train': 5.904702186584473} 02/24/2022 02:54:35 - INFO - codeparrot_training - Step 1015: {'lr': 0.00025374999999999996, 'samples': 520192, 'steps': 1015, 'loss/train': 5.99049186706543} 02/24/2022 02:54:40 - INFO - codeparrot_training - Step 1016: {'lr': 0.000254, 'samples': 520704, 'steps': 1016, 'loss/train': 5.846545696258545} 02/24/2022 02:54:44 - INFO - codeparrot_training - Step 1017: {'lr': 0.00025425, 'samples': 521216, 'steps': 1017, 'loss/train': 4.6801276206970215} 02/24/2022 02:54:49 - INFO - codeparrot_training - Step 1018: {'lr': 0.0002545, 'samples': 521728, 'steps': 1018, 'loss/train': 4.828365325927734} 02/24/2022 02:54:53 - INFO - codeparrot_training - Step 1019: {'lr': 0.00025475, 'samples': 522240, 'steps': 1019, 'loss/train': 5.723589897155762} 02/24/2022 02:54:59 - INFO - codeparrot_training - Step 1020: {'lr': 0.000255, 'samples': 522752, 'steps': 1020, 'loss/train': 5.405442237854004} 02/24/2022 02:55:02 - INFO - codeparrot_training - Step 1021: {'lr': 0.00025525, 'samples': 523264, 'steps': 1021, 'loss/train': 4.955881595611572} 02/24/2022 02:55:08 - INFO - codeparrot_training - Step 1022: {'lr': 0.00025550000000000003, 'samples': 523776, 'steps': 1022, 'loss/train': 5.056830406188965} 02/24/2022 02:55:11 - INFO - codeparrot_training - Step 1023: {'lr': 0.00025575, 'samples': 524288, 'steps': 1023, 'loss/train': 5.384207248687744} 02/24/2022 02:55:16 - INFO - codeparrot_training - Step 1024: {'lr': 0.000256, 'samples': 524800, 'steps': 1024, 'loss/train': 5.197412490844727} 02/24/2022 02:55:20 - INFO - codeparrot_training - Step 1025: {'lr': 0.00025624999999999997, 'samples': 525312, 'steps': 1025, 'loss/train': 5.356321334838867} 02/24/2022 02:55:25 - INFO - codeparrot_training - Step 1026: {'lr': 0.0002565, 'samples': 525824, 'steps': 1026, 'loss/train': 5.2317585945129395} 02/24/2022 02:55:29 - INFO - codeparrot_training - Step 1027: {'lr': 0.00025675, 'samples': 526336, 'steps': 1027, 'loss/train': 5.499627590179443} 02/24/2022 02:55:34 - INFO - codeparrot_training - Step 1028: {'lr': 0.000257, 'samples': 526848, 'steps': 1028, 'loss/train': 5.996160507202148} 02/24/2022 02:55:38 - INFO - codeparrot_training - Step 1029: {'lr': 0.00025725, 'samples': 527360, 'steps': 1029, 'loss/train': 5.855902671813965} 02/24/2022 02:55:44 - INFO - codeparrot_training - Step 1030: {'lr': 0.0002575, 'samples': 527872, 'steps': 1030, 'loss/train': 5.065313816070557} 02/24/2022 02:55:47 - INFO - codeparrot_training - Step 1031: {'lr': 0.00025775, 'samples': 528384, 'steps': 1031, 'loss/train': 6.560394287109375} 02/24/2022 02:55:53 - INFO - codeparrot_training - Step 1032: {'lr': 0.00025800000000000004, 'samples': 528896, 'steps': 1032, 'loss/train': 4.543503761291504} 02/24/2022 02:55:58 - INFO - codeparrot_training - Step 1033: {'lr': 0.00025824999999999996, 'samples': 529408, 'steps': 1033, 'loss/train': 5.719844341278076} 02/24/2022 02:56:02 - INFO - codeparrot_training - Step 1034: {'lr': 0.0002585, 'samples': 529920, 'steps': 1034, 'loss/train': 4.815674781799316} 02/24/2022 02:56:07 - INFO - codeparrot_training - Step 1035: {'lr': 0.00025875, 'samples': 530432, 'steps': 1035, 'loss/train': 4.833280086517334} 02/24/2022 02:56:11 - INFO - codeparrot_training - Step 1036: {'lr': 0.000259, 'samples': 530944, 'steps': 1036, 'loss/train': 5.193758487701416} 02/24/2022 02:56:16 - INFO - codeparrot_training - Step 1037: {'lr': 0.00025925, 'samples': 531456, 'steps': 1037, 'loss/train': 4.932142734527588} 02/24/2022 02:56:20 - INFO - codeparrot_training - Step 1038: {'lr': 0.0002595, 'samples': 531968, 'steps': 1038, 'loss/train': 5.557000160217285} 02/24/2022 02:56:26 - INFO - codeparrot_training - Step 1039: {'lr': 0.00025975, 'samples': 532480, 'steps': 1039, 'loss/train': 5.681889533996582} 02/24/2022 02:56:29 - INFO - codeparrot_training - Step 1040: {'lr': 0.00026000000000000003, 'samples': 532992, 'steps': 1040, 'loss/train': 4.866218566894531} 02/24/2022 02:56:35 - INFO - codeparrot_training - Step 1041: {'lr': 0.00026025, 'samples': 533504, 'steps': 1041, 'loss/train': 5.507498264312744} 02/24/2022 02:56:38 - INFO - codeparrot_training - Step 1042: {'lr': 0.0002605, 'samples': 534016, 'steps': 1042, 'loss/train': 5.162564754486084} 02/24/2022 02:56:44 - INFO - codeparrot_training - Step 1043: {'lr': 0.00026074999999999997, 'samples': 534528, 'steps': 1043, 'loss/train': 5.068989276885986} 02/24/2022 02:56:47 - INFO - codeparrot_training - Step 1044: {'lr': 0.000261, 'samples': 535040, 'steps': 1044, 'loss/train': 6.276032447814941} 02/24/2022 02:56:53 - INFO - codeparrot_training - Step 1045: {'lr': 0.00026125, 'samples': 535552, 'steps': 1045, 'loss/train': 8.02214527130127} 02/24/2022 02:56:56 - INFO - codeparrot_training - Step 1046: {'lr': 0.0002615, 'samples': 536064, 'steps': 1046, 'loss/train': 5.075617790222168} 02/24/2022 02:57:02 - INFO - codeparrot_training - Step 1047: {'lr': 0.00026175, 'samples': 536576, 'steps': 1047, 'loss/train': 5.763293743133545} 02/24/2022 02:57:05 - INFO - codeparrot_training - Step 1048: {'lr': 0.000262, 'samples': 537088, 'steps': 1048, 'loss/train': 1.702135682106018} 02/24/2022 02:57:11 - INFO - codeparrot_training - Step 1049: {'lr': 0.00026225, 'samples': 537600, 'steps': 1049, 'loss/train': 5.185730934143066} 02/24/2022 02:57:14 - INFO - codeparrot_training - Step 1050: {'lr': 0.00026250000000000004, 'samples': 538112, 'steps': 1050, 'loss/train': 3.9915056228637695} 02/24/2022 02:57:20 - INFO - codeparrot_training - Step 1051: {'lr': 0.00026274999999999996, 'samples': 538624, 'steps': 1051, 'loss/train': 4.6293792724609375} 02/24/2022 02:57:24 - INFO - codeparrot_training - Step 1052: {'lr': 0.000263, 'samples': 539136, 'steps': 1052, 'loss/train': 5.293740749359131} 02/24/2022 02:57:29 - INFO - codeparrot_training - Step 1053: {'lr': 0.00026325, 'samples': 539648, 'steps': 1053, 'loss/train': 5.670673370361328} 02/24/2022 02:57:33 - INFO - codeparrot_training - Step 1054: {'lr': 0.0002635, 'samples': 540160, 'steps': 1054, 'loss/train': 5.470986843109131} 02/24/2022 02:57:38 - INFO - codeparrot_training - Step 1055: {'lr': 0.00026375, 'samples': 540672, 'steps': 1055, 'loss/train': 5.483548164367676} 02/24/2022 02:57:42 - INFO - codeparrot_training - Step 1056: {'lr': 0.000264, 'samples': 541184, 'steps': 1056, 'loss/train': 4.262084007263184} 02/24/2022 02:57:50 - INFO - codeparrot_training - Step 1057: {'lr': 0.00026425, 'samples': 541696, 'steps': 1057, 'loss/train': 5.8530168533325195} 02/24/2022 02:57:53 - INFO - codeparrot_training - Step 1058: {'lr': 0.00026450000000000003, 'samples': 542208, 'steps': 1058, 'loss/train': 4.918515205383301} 02/24/2022 02:57:59 - INFO - codeparrot_training - Step 1059: {'lr': 0.00026475, 'samples': 542720, 'steps': 1059, 'loss/train': 4.118727684020996} 02/24/2022 02:58:02 - INFO - codeparrot_training - Step 1060: {'lr': 0.00026500000000000004, 'samples': 543232, 'steps': 1060, 'loss/train': 6.1179962158203125} 02/24/2022 02:58:08 - INFO - codeparrot_training - Step 1061: {'lr': 0.00026524999999999997, 'samples': 543744, 'steps': 1061, 'loss/train': 4.762015342712402} 02/24/2022 02:58:11 - INFO - codeparrot_training - Step 1062: {'lr': 0.0002655, 'samples': 544256, 'steps': 1062, 'loss/train': 4.535111904144287} 02/24/2022 02:58:17 - INFO - codeparrot_training - Step 1063: {'lr': 0.00026575, 'samples': 544768, 'steps': 1063, 'loss/train': 5.647745132446289} 02/24/2022 02:58:20 - INFO - codeparrot_training - Step 1064: {'lr': 0.000266, 'samples': 545280, 'steps': 1064, 'loss/train': 5.756256580352783} 02/24/2022 02:58:26 - INFO - codeparrot_training - Step 1065: {'lr': 0.00026625, 'samples': 545792, 'steps': 1065, 'loss/train': 5.518280982971191} 02/24/2022 02:58:29 - INFO - codeparrot_training - Step 1066: {'lr': 0.0002665, 'samples': 546304, 'steps': 1066, 'loss/train': 5.746287822723389} 02/24/2022 02:58:36 - INFO - codeparrot_training - Step 1067: {'lr': 0.00026675, 'samples': 546816, 'steps': 1067, 'loss/train': 5.017087936401367} 02/24/2022 02:58:40 - INFO - codeparrot_training - Step 1068: {'lr': 0.00026700000000000004, 'samples': 547328, 'steps': 1068, 'loss/train': 5.298943996429443} 02/24/2022 02:58:45 - INFO - codeparrot_training - Step 1069: {'lr': 0.00026725, 'samples': 547840, 'steps': 1069, 'loss/train': 5.6898298263549805} 02/24/2022 02:58:49 - INFO - codeparrot_training - Step 1070: {'lr': 0.0002675, 'samples': 548352, 'steps': 1070, 'loss/train': 3.6321702003479004} 02/24/2022 02:58:54 - INFO - codeparrot_training - Step 1071: {'lr': 0.00026775, 'samples': 548864, 'steps': 1071, 'loss/train': 5.318291187286377} 02/24/2022 02:58:58 - INFO - codeparrot_training - Step 1072: {'lr': 0.000268, 'samples': 549376, 'steps': 1072, 'loss/train': 5.162319183349609} 02/24/2022 02:59:03 - INFO - codeparrot_training - Step 1073: {'lr': 0.00026825, 'samples': 549888, 'steps': 1073, 'loss/train': 5.279951095581055} 02/24/2022 02:59:07 - INFO - codeparrot_training - Step 1074: {'lr': 0.0002685, 'samples': 550400, 'steps': 1074, 'loss/train': 5.386423110961914} 02/24/2022 02:59:12 - INFO - codeparrot_training - Step 1075: {'lr': 0.00026875, 'samples': 550912, 'steps': 1075, 'loss/train': 5.800024509429932} 02/24/2022 02:59:16 - INFO - codeparrot_training - Step 1076: {'lr': 0.00026900000000000003, 'samples': 551424, 'steps': 1076, 'loss/train': 4.981051445007324} 02/24/2022 02:59:23 - INFO - codeparrot_training - Step 1077: {'lr': 0.00026925, 'samples': 551936, 'steps': 1077, 'loss/train': 5.475964069366455} 02/24/2022 02:59:27 - INFO - codeparrot_training - Step 1078: {'lr': 0.00026950000000000005, 'samples': 552448, 'steps': 1078, 'loss/train': 6.140387535095215} 02/24/2022 02:59:32 - INFO - codeparrot_training - Step 1079: {'lr': 0.00026974999999999997, 'samples': 552960, 'steps': 1079, 'loss/train': 5.8200812339782715} 02/24/2022 02:59:35 - INFO - codeparrot_training - Step 1080: {'lr': 0.00027, 'samples': 553472, 'steps': 1080, 'loss/train': 2.874450206756592} 02/24/2022 02:59:41 - INFO - codeparrot_training - Step 1081: {'lr': 0.00027025, 'samples': 553984, 'steps': 1081, 'loss/train': 5.2649149894714355} 02/24/2022 02:59:44 - INFO - codeparrot_training - Step 1082: {'lr': 0.0002705, 'samples': 554496, 'steps': 1082, 'loss/train': 6.761886119842529} 02/24/2022 02:59:50 - INFO - codeparrot_training - Step 1083: {'lr': 0.00027075, 'samples': 555008, 'steps': 1083, 'loss/train': 4.219944000244141} 02/24/2022 02:59:53 - INFO - codeparrot_training - Step 1084: {'lr': 0.00027100000000000003, 'samples': 555520, 'steps': 1084, 'loss/train': 5.17498779296875} 02/24/2022 02:59:59 - INFO - codeparrot_training - Step 1085: {'lr': 0.00027125, 'samples': 556032, 'steps': 1085, 'loss/train': 6.553275108337402} 02/24/2022 03:00:03 - INFO - codeparrot_training - Step 1086: {'lr': 0.00027150000000000004, 'samples': 556544, 'steps': 1086, 'loss/train': 4.6969404220581055} 02/24/2022 03:00:08 - INFO - codeparrot_training - Step 1087: {'lr': 0.00027175, 'samples': 557056, 'steps': 1087, 'loss/train': 4.942883491516113} 02/24/2022 03:00:11 - INFO - codeparrot_training - Step 1088: {'lr': 0.00027200000000000005, 'samples': 557568, 'steps': 1088, 'loss/train': 4.379235744476318} 02/24/2022 03:00:19 - INFO - codeparrot_training - Step 1089: {'lr': 0.00027225, 'samples': 558080, 'steps': 1089, 'loss/train': 5.86265230178833} 02/24/2022 03:00:22 - INFO - codeparrot_training - Step 1090: {'lr': 0.0002725, 'samples': 558592, 'steps': 1090, 'loss/train': 4.575802803039551} 02/24/2022 03:00:28 - INFO - codeparrot_training - Step 1091: {'lr': 0.00027275, 'samples': 559104, 'steps': 1091, 'loss/train': 1.775154948234558} 02/24/2022 03:00:31 - INFO - codeparrot_training - Step 1092: {'lr': 0.000273, 'samples': 559616, 'steps': 1092, 'loss/train': 5.635185718536377} 02/24/2022 03:00:37 - INFO - codeparrot_training - Step 1093: {'lr': 0.00027325, 'samples': 560128, 'steps': 1093, 'loss/train': 4.302326679229736} 02/24/2022 03:00:40 - INFO - codeparrot_training - Step 1094: {'lr': 0.00027350000000000003, 'samples': 560640, 'steps': 1094, 'loss/train': 6.043479919433594} 02/24/2022 03:00:46 - INFO - codeparrot_training - Step 1095: {'lr': 0.00027375, 'samples': 561152, 'steps': 1095, 'loss/train': 6.240167140960693} 02/24/2022 03:00:49 - INFO - codeparrot_training - Step 1096: {'lr': 0.00027400000000000005, 'samples': 561664, 'steps': 1096, 'loss/train': 6.487496852874756} 02/24/2022 03:00:55 - INFO - codeparrot_training - Step 1097: {'lr': 0.00027425, 'samples': 562176, 'steps': 1097, 'loss/train': 4.986917018890381} 02/24/2022 03:00:58 - INFO - codeparrot_training - Step 1098: {'lr': 0.0002745, 'samples': 562688, 'steps': 1098, 'loss/train': 6.253907680511475} 02/24/2022 03:01:04 - INFO - codeparrot_training - Step 1099: {'lr': 0.00027475, 'samples': 563200, 'steps': 1099, 'loss/train': 5.558638095855713} 02/24/2022 03:01:09 - INFO - codeparrot_training - Step 1100: {'lr': 0.000275, 'samples': 563712, 'steps': 1100, 'loss/train': 6.074194431304932} 02/24/2022 03:01:13 - INFO - codeparrot_training - Step 1101: {'lr': 0.00027525, 'samples': 564224, 'steps': 1101, 'loss/train': 5.0031514167785645} 02/24/2022 03:01:20 - INFO - codeparrot_training - Step 1102: {'lr': 0.00027550000000000003, 'samples': 564736, 'steps': 1102, 'loss/train': 5.181690692901611} 02/24/2022 03:01:23 - INFO - codeparrot_training - Step 1103: {'lr': 0.00027575, 'samples': 565248, 'steps': 1103, 'loss/train': 6.462654113769531} 02/24/2022 03:01:29 - INFO - codeparrot_training - Step 1104: {'lr': 0.00027600000000000004, 'samples': 565760, 'steps': 1104, 'loss/train': 4.807124137878418} 02/24/2022 03:01:32 - INFO - codeparrot_training - Step 1105: {'lr': 0.00027625, 'samples': 566272, 'steps': 1105, 'loss/train': 5.1877007484436035} 02/24/2022 03:01:38 - INFO - codeparrot_training - Step 1106: {'lr': 0.00027650000000000005, 'samples': 566784, 'steps': 1106, 'loss/train': 5.409707069396973} 02/24/2022 03:01:41 - INFO - codeparrot_training - Step 1107: {'lr': 0.00027675, 'samples': 567296, 'steps': 1107, 'loss/train': 5.500791549682617} 02/24/2022 03:01:47 - INFO - codeparrot_training - Step 1108: {'lr': 0.000277, 'samples': 567808, 'steps': 1108, 'loss/train': 4.060062408447266} 02/24/2022 03:01:50 - INFO - codeparrot_training - Step 1109: {'lr': 0.00027725, 'samples': 568320, 'steps': 1109, 'loss/train': 4.874750137329102} 02/24/2022 03:01:56 - INFO - codeparrot_training - Step 1110: {'lr': 0.0002775, 'samples': 568832, 'steps': 1110, 'loss/train': 5.023270130157471} 02/24/2022 03:01:59 - INFO - codeparrot_training - Step 1111: {'lr': 0.00027775, 'samples': 569344, 'steps': 1111, 'loss/train': 5.3346076011657715} 02/24/2022 03:02:05 - INFO - codeparrot_training - Step 1112: {'lr': 0.00027800000000000004, 'samples': 569856, 'steps': 1112, 'loss/train': 4.889978885650635} 02/24/2022 03:02:08 - INFO - codeparrot_training - Step 1113: {'lr': 0.00027825, 'samples': 570368, 'steps': 1113, 'loss/train': 1.7322806119918823} 02/24/2022 03:02:15 - INFO - codeparrot_training - Step 1114: {'lr': 0.00027850000000000005, 'samples': 570880, 'steps': 1114, 'loss/train': 5.020740985870361} 02/24/2022 03:02:19 - INFO - codeparrot_training - Step 1115: {'lr': 0.00027875, 'samples': 571392, 'steps': 1115, 'loss/train': 4.7959160804748535} 02/24/2022 03:02:24 - INFO - codeparrot_training - Step 1116: {'lr': 0.000279, 'samples': 571904, 'steps': 1116, 'loss/train': 5.215575218200684} 02/24/2022 03:02:28 - INFO - codeparrot_training - Step 1117: {'lr': 0.00027925, 'samples': 572416, 'steps': 1117, 'loss/train': 3.991440773010254} 02/24/2022 03:02:33 - INFO - codeparrot_training - Step 1118: {'lr': 0.0002795, 'samples': 572928, 'steps': 1118, 'loss/train': 6.473158359527588} 02/24/2022 03:02:37 - INFO - codeparrot_training - Step 1119: {'lr': 0.00027975, 'samples': 573440, 'steps': 1119, 'loss/train': 4.704736709594727} 02/24/2022 03:02:42 - INFO - codeparrot_training - Step 1120: {'lr': 0.00028000000000000003, 'samples': 573952, 'steps': 1120, 'loss/train': 4.446264743804932} 02/24/2022 03:02:46 - INFO - codeparrot_training - Step 1121: {'lr': 0.00028025, 'samples': 574464, 'steps': 1121, 'loss/train': 4.785529613494873} 02/24/2022 03:02:53 - INFO - codeparrot_training - Step 1122: {'lr': 0.00028050000000000004, 'samples': 574976, 'steps': 1122, 'loss/train': 5.402191162109375} 02/24/2022 03:02:56 - INFO - codeparrot_training - Step 1123: {'lr': 0.00028075, 'samples': 575488, 'steps': 1123, 'loss/train': 5.555760860443115} 02/24/2022 03:03:02 - INFO - codeparrot_training - Step 1124: {'lr': 0.00028100000000000005, 'samples': 576000, 'steps': 1124, 'loss/train': 5.6269989013671875} 02/24/2022 03:03:05 - INFO - codeparrot_training - Step 1125: {'lr': 0.00028125000000000003, 'samples': 576512, 'steps': 1125, 'loss/train': 4.530860424041748} 02/24/2022 03:03:11 - INFO - codeparrot_training - Step 1126: {'lr': 0.00028149999999999996, 'samples': 577024, 'steps': 1126, 'loss/train': 5.3932785987854} 02/24/2022 03:03:14 - INFO - codeparrot_training - Step 1127: {'lr': 0.00028175, 'samples': 577536, 'steps': 1127, 'loss/train': 4.677549362182617} 02/24/2022 03:03:20 - INFO - codeparrot_training - Step 1128: {'lr': 0.00028199999999999997, 'samples': 578048, 'steps': 1128, 'loss/train': 5.662293434143066} 02/24/2022 03:03:24 - INFO - codeparrot_training - Step 1129: {'lr': 0.00028225, 'samples': 578560, 'steps': 1129, 'loss/train': 2.453404426574707} 02/24/2022 03:03:29 - INFO - codeparrot_training - Step 1130: {'lr': 0.0002825, 'samples': 579072, 'steps': 1130, 'loss/train': 4.762810707092285} 02/24/2022 03:03:33 - INFO - codeparrot_training - Step 1131: {'lr': 0.00028275, 'samples': 579584, 'steps': 1131, 'loss/train': 5.110528945922852} 02/24/2022 03:03:38 - INFO - codeparrot_training - Step 1132: {'lr': 0.000283, 'samples': 580096, 'steps': 1132, 'loss/train': 5.49754524230957} 02/24/2022 03:03:41 - INFO - codeparrot_training - Step 1133: {'lr': 0.00028325000000000003, 'samples': 580608, 'steps': 1133, 'loss/train': 5.466548442840576} 02/24/2022 03:03:47 - INFO - codeparrot_training - Step 1134: {'lr': 0.0002835, 'samples': 581120, 'steps': 1134, 'loss/train': 4.763579845428467} 02/24/2022 03:03:53 - INFO - codeparrot_training - Step 1135: {'lr': 0.00028375, 'samples': 581632, 'steps': 1135, 'loss/train': 4.981903076171875} 02/24/2022 03:03:56 - INFO - codeparrot_training - Step 1136: {'lr': 0.00028399999999999996, 'samples': 582144, 'steps': 1136, 'loss/train': 6.5357818603515625} 02/24/2022 03:04:03 - INFO - codeparrot_training - Step 1137: {'lr': 0.00028425, 'samples': 582656, 'steps': 1137, 'loss/train': 5.191011905670166} 02/24/2022 03:04:07 - INFO - codeparrot_training - Step 1138: {'lr': 0.0002845, 'samples': 583168, 'steps': 1138, 'loss/train': 4.785286903381348} 02/24/2022 03:04:12 - INFO - codeparrot_training - Step 1139: {'lr': 0.00028475, 'samples': 583680, 'steps': 1139, 'loss/train': 5.120326995849609} 02/24/2022 03:04:16 - INFO - codeparrot_training - Step 1140: {'lr': 0.000285, 'samples': 584192, 'steps': 1140, 'loss/train': 4.626264572143555} 02/24/2022 03:04:21 - INFO - codeparrot_training - Step 1141: {'lr': 0.00028525, 'samples': 584704, 'steps': 1141, 'loss/train': 6.172025203704834} 02/24/2022 03:04:25 - INFO - codeparrot_training - Step 1142: {'lr': 0.0002855, 'samples': 585216, 'steps': 1142, 'loss/train': 2.3676493167877197} 02/24/2022 03:04:30 - INFO - codeparrot_training - Step 1143: {'lr': 0.00028575000000000003, 'samples': 585728, 'steps': 1143, 'loss/train': 5.581705093383789} 02/24/2022 03:04:34 - INFO - codeparrot_training - Step 1144: {'lr': 0.00028599999999999996, 'samples': 586240, 'steps': 1144, 'loss/train': 3.964090347290039} 02/24/2022 03:04:39 - INFO - codeparrot_training - Step 1145: {'lr': 0.00028625, 'samples': 586752, 'steps': 1145, 'loss/train': 6.0772857666015625} 02/24/2022 03:04:43 - INFO - codeparrot_training - Step 1146: {'lr': 0.00028649999999999997, 'samples': 587264, 'steps': 1146, 'loss/train': 4.500453472137451} 02/24/2022 03:04:48 - INFO - codeparrot_training - Step 1147: {'lr': 0.00028675, 'samples': 587776, 'steps': 1147, 'loss/train': 4.486052989959717} 02/24/2022 03:04:52 - INFO - codeparrot_training - Step 1148: {'lr': 0.000287, 'samples': 588288, 'steps': 1148, 'loss/train': 5.942382335662842} 02/24/2022 03:04:59 - INFO - codeparrot_training - Step 1149: {'lr': 0.00028725, 'samples': 588800, 'steps': 1149, 'loss/train': 5.77067232131958} 02/24/2022 03:05:02 - INFO - codeparrot_training - Step 1150: {'lr': 0.0002875, 'samples': 589312, 'steps': 1150, 'loss/train': 5.469539165496826} 02/24/2022 03:05:08 - INFO - codeparrot_training - Step 1151: {'lr': 0.00028775000000000003, 'samples': 589824, 'steps': 1151, 'loss/train': 6.311186790466309} 02/24/2022 03:05:11 - INFO - codeparrot_training - Step 1152: {'lr': 0.000288, 'samples': 590336, 'steps': 1152, 'loss/train': 4.198432922363281} 02/24/2022 03:05:17 - INFO - codeparrot_training - Step 1153: {'lr': 0.00028825, 'samples': 590848, 'steps': 1153, 'loss/train': 3.5707523822784424} 02/24/2022 03:05:21 - INFO - codeparrot_training - Step 1154: {'lr': 0.00028849999999999997, 'samples': 591360, 'steps': 1154, 'loss/train': 4.79783296585083} 02/24/2022 03:05:26 - INFO - codeparrot_training - Step 1155: {'lr': 0.00028875, 'samples': 591872, 'steps': 1155, 'loss/train': 4.189629077911377} 02/24/2022 03:05:30 - INFO - codeparrot_training - Step 1156: {'lr': 0.000289, 'samples': 592384, 'steps': 1156, 'loss/train': 5.576852321624756} 02/24/2022 03:05:35 - INFO - codeparrot_training - Step 1157: {'lr': 0.00028925, 'samples': 592896, 'steps': 1157, 'loss/train': 6.0749921798706055} 02/24/2022 03:05:38 - INFO - codeparrot_training - Step 1158: {'lr': 0.0002895, 'samples': 593408, 'steps': 1158, 'loss/train': 5.3626556396484375} 02/24/2022 03:05:46 - INFO - codeparrot_training - Step 1159: {'lr': 0.00028975, 'samples': 593920, 'steps': 1159, 'loss/train': 4.907273769378662} 02/24/2022 03:05:49 - INFO - codeparrot_training - Step 1160: {'lr': 0.00029, 'samples': 594432, 'steps': 1160, 'loss/train': 6.00240421295166} 02/24/2022 03:05:53 - INFO - codeparrot_training - Step 1161: {'lr': 0.00029025000000000003, 'samples': 594944, 'steps': 1161, 'loss/train': 2.301327705383301} 02/24/2022 03:05:58 - INFO - codeparrot_training - Step 1162: {'lr': 0.00029049999999999996, 'samples': 595456, 'steps': 1162, 'loss/train': 5.6408281326293945} 02/24/2022 03:06:04 - INFO - codeparrot_training - Step 1163: {'lr': 0.00029075, 'samples': 595968, 'steps': 1163, 'loss/train': 4.322697162628174} 02/24/2022 03:06:07 - INFO - codeparrot_training - Step 1164: {'lr': 0.00029099999999999997, 'samples': 596480, 'steps': 1164, 'loss/train': 4.571609020233154} 02/24/2022 03:06:13 - INFO - codeparrot_training - Step 1165: {'lr': 0.00029125, 'samples': 596992, 'steps': 1165, 'loss/train': 5.720776557922363} 02/24/2022 03:06:16 - INFO - codeparrot_training - Step 1166: {'lr': 0.0002915, 'samples': 597504, 'steps': 1166, 'loss/train': 4.859494686126709} 02/24/2022 03:06:22 - INFO - codeparrot_training - Step 1167: {'lr': 0.00029175, 'samples': 598016, 'steps': 1167, 'loss/train': 5.917038917541504} 02/24/2022 03:06:25 - INFO - codeparrot_training - Step 1168: {'lr': 0.000292, 'samples': 598528, 'steps': 1168, 'loss/train': 5.487188816070557} 02/24/2022 03:06:32 - INFO - codeparrot_training - Step 1169: {'lr': 0.00029225000000000003, 'samples': 599040, 'steps': 1169, 'loss/train': 4.916807174682617} 02/24/2022 03:06:36 - INFO - codeparrot_training - Step 1170: {'lr': 0.0002925, 'samples': 599552, 'steps': 1170, 'loss/train': 4.521228313446045} 02/24/2022 03:06:41 - INFO - codeparrot_training - Step 1171: {'lr': 0.00029275000000000004, 'samples': 600064, 'steps': 1171, 'loss/train': 4.090498447418213} 02/24/2022 03:06:45 - INFO - codeparrot_training - Step 1172: {'lr': 0.00029299999999999997, 'samples': 600576, 'steps': 1172, 'loss/train': 4.494028091430664} 02/24/2022 03:06:50 - INFO - codeparrot_training - Step 1173: {'lr': 0.00029325, 'samples': 601088, 'steps': 1173, 'loss/train': 4.425089359283447} 02/24/2022 03:06:54 - INFO - codeparrot_training - Step 1174: {'lr': 0.0002935, 'samples': 601600, 'steps': 1174, 'loss/train': 5.415894508361816} 02/24/2022 03:06:59 - INFO - codeparrot_training - Step 1175: {'lr': 0.00029375, 'samples': 602112, 'steps': 1175, 'loss/train': 5.765039920806885} 02/24/2022 03:07:03 - INFO - codeparrot_training - Step 1176: {'lr': 0.000294, 'samples': 602624, 'steps': 1176, 'loss/train': 4.922140598297119} 02/24/2022 03:07:08 - INFO - codeparrot_training - Step 1177: {'lr': 0.00029425, 'samples': 603136, 'steps': 1177, 'loss/train': 5.992300987243652} 02/24/2022 03:07:12 - INFO - codeparrot_training - Step 1178: {'lr': 0.0002945, 'samples': 603648, 'steps': 1178, 'loss/train': 4.574462413787842} 02/24/2022 03:07:17 - INFO - codeparrot_training - Step 1179: {'lr': 0.00029475000000000004, 'samples': 604160, 'steps': 1179, 'loss/train': 4.694687366485596} 02/24/2022 03:07:21 - INFO - codeparrot_training - Step 1180: {'lr': 0.000295, 'samples': 604672, 'steps': 1180, 'loss/train': 5.516098499298096} 02/24/2022 03:07:26 - INFO - codeparrot_training - Step 1181: {'lr': 0.00029525, 'samples': 605184, 'steps': 1181, 'loss/train': 5.470821380615234} 02/24/2022 03:07:30 - INFO - codeparrot_training - Step 1182: {'lr': 0.00029549999999999997, 'samples': 605696, 'steps': 1182, 'loss/train': 6.128830432891846} 02/24/2022 03:07:35 - INFO - codeparrot_training - Step 1183: {'lr': 0.00029575, 'samples': 606208, 'steps': 1183, 'loss/train': 5.9133195877075195} 02/24/2022 03:07:38 - INFO - codeparrot_training - Step 1184: {'lr': 0.000296, 'samples': 606720, 'steps': 1184, 'loss/train': 4.757331848144531} 02/24/2022 03:07:46 - INFO - codeparrot_training - Step 1185: {'lr': 0.00029625, 'samples': 607232, 'steps': 1185, 'loss/train': 4.075989246368408} 02/24/2022 03:07:49 - INFO - codeparrot_training - Step 1186: {'lr': 0.0002965, 'samples': 607744, 'steps': 1186, 'loss/train': 4.66351842880249} 02/24/2022 03:07:55 - INFO - codeparrot_training - Step 1187: {'lr': 0.00029675000000000003, 'samples': 608256, 'steps': 1187, 'loss/train': 5.199337482452393} 02/24/2022 03:07:58 - INFO - codeparrot_training - Step 1188: {'lr': 0.000297, 'samples': 608768, 'steps': 1188, 'loss/train': 4.158010482788086} 02/24/2022 03:08:04 - INFO - codeparrot_training - Step 1189: {'lr': 0.00029725000000000004, 'samples': 609280, 'steps': 1189, 'loss/train': 5.813632965087891} 02/24/2022 03:08:07 - INFO - codeparrot_training - Step 1190: {'lr': 0.00029749999999999997, 'samples': 609792, 'steps': 1190, 'loss/train': 5.551846504211426} 02/24/2022 03:08:12 - INFO - codeparrot_training - Step 1191: {'lr': 0.00029775, 'samples': 610304, 'steps': 1191, 'loss/train': 5.164458751678467} 02/24/2022 03:08:16 - INFO - codeparrot_training - Step 1192: {'lr': 0.000298, 'samples': 610816, 'steps': 1192, 'loss/train': 4.805848598480225} 02/24/2022 03:08:22 - INFO - codeparrot_training - Step 1193: {'lr': 0.00029825, 'samples': 611328, 'steps': 1193, 'loss/train': 4.825739860534668} 02/24/2022 03:08:25 - INFO - codeparrot_training - Step 1194: {'lr': 0.0002985, 'samples': 611840, 'steps': 1194, 'loss/train': 4.308709144592285} 02/24/2022 03:08:32 - INFO - codeparrot_training - Step 1195: {'lr': 0.00029875, 'samples': 612352, 'steps': 1195, 'loss/train': 5.1788458824157715} 02/24/2022 03:08:36 - INFO - codeparrot_training - Step 1196: {'lr': 0.000299, 'samples': 612864, 'steps': 1196, 'loss/train': 4.943964004516602} 02/24/2022 03:08:41 - INFO - codeparrot_training - Step 1197: {'lr': 0.00029925000000000004, 'samples': 613376, 'steps': 1197, 'loss/train': 5.083621025085449} 02/24/2022 03:08:45 - INFO - codeparrot_training - Step 1198: {'lr': 0.0002995, 'samples': 613888, 'steps': 1198, 'loss/train': 1.606532335281372} 02/24/2022 03:08:50 - INFO - codeparrot_training - Step 1199: {'lr': 0.00029975000000000005, 'samples': 614400, 'steps': 1199, 'loss/train': 4.5193400382995605} 02/24/2022 03:08:54 - INFO - codeparrot_training - Step 1200: {'lr': 0.0003, 'samples': 614912, 'steps': 1200, 'loss/train': 4.09227991104126} 02/24/2022 03:08:59 - INFO - codeparrot_training - Step 1201: {'lr': 0.00030025, 'samples': 615424, 'steps': 1201, 'loss/train': 5.314067840576172} 02/24/2022 03:09:03 - INFO - codeparrot_training - Step 1202: {'lr': 0.0003005, 'samples': 615936, 'steps': 1202, 'loss/train': 6.291508674621582} 02/24/2022 03:09:08 - INFO - codeparrot_training - Step 1203: {'lr': 0.00030075, 'samples': 616448, 'steps': 1203, 'loss/train': 5.892355918884277} 02/24/2022 03:09:12 - INFO - codeparrot_training - Step 1204: {'lr': 0.000301, 'samples': 616960, 'steps': 1204, 'loss/train': 1.4669965505599976} 02/24/2022 03:09:19 - INFO - codeparrot_training - Step 1205: {'lr': 0.00030125000000000003, 'samples': 617472, 'steps': 1205, 'loss/train': 5.404385566711426} 02/24/2022 03:09:24 - INFO - codeparrot_training - Step 1206: {'lr': 0.0003015, 'samples': 617984, 'steps': 1206, 'loss/train': 4.801541805267334} 02/24/2022 03:09:28 - INFO - codeparrot_training - Step 1207: {'lr': 0.00030175000000000004, 'samples': 618496, 'steps': 1207, 'loss/train': 5.1355414390563965} 02/24/2022 03:09:33 - INFO - codeparrot_training - Step 1208: {'lr': 0.000302, 'samples': 619008, 'steps': 1208, 'loss/train': 4.8629279136657715} 02/24/2022 03:09:37 - INFO - codeparrot_training - Step 1209: {'lr': 0.00030225, 'samples': 619520, 'steps': 1209, 'loss/train': 5.296463966369629} 02/24/2022 03:09:42 - INFO - codeparrot_training - Step 1210: {'lr': 0.0003025, 'samples': 620032, 'steps': 1210, 'loss/train': 5.099849700927734} 02/24/2022 03:09:46 - INFO - codeparrot_training - Step 1211: {'lr': 0.00030275, 'samples': 620544, 'steps': 1211, 'loss/train': 4.357608318328857} 02/24/2022 03:09:51 - INFO - codeparrot_training - Step 1212: {'lr': 0.000303, 'samples': 621056, 'steps': 1212, 'loss/train': 4.461977005004883} 02/24/2022 03:09:54 - INFO - codeparrot_training - Step 1213: {'lr': 0.00030325, 'samples': 621568, 'steps': 1213, 'loss/train': 4.262706756591797} 02/24/2022 03:10:02 - INFO - codeparrot_training - Step 1214: {'lr': 0.0003035, 'samples': 622080, 'steps': 1214, 'loss/train': 4.767970561981201} 02/24/2022 03:10:05 - INFO - codeparrot_training - Step 1215: {'lr': 0.00030375000000000004, 'samples': 622592, 'steps': 1215, 'loss/train': 5.04935884475708} 02/24/2022 03:10:11 - INFO - codeparrot_training - Step 1216: {'lr': 0.000304, 'samples': 623104, 'steps': 1216, 'loss/train': 5.2728447914123535} 02/24/2022 03:10:14 - INFO - codeparrot_training - Step 1217: {'lr': 0.00030425000000000005, 'samples': 623616, 'steps': 1217, 'loss/train': 4.5625457763671875} 02/24/2022 03:10:20 - INFO - codeparrot_training - Step 1218: {'lr': 0.0003045, 'samples': 624128, 'steps': 1218, 'loss/train': 6.0289506912231445} 02/24/2022 03:10:23 - INFO - codeparrot_training - Step 1219: {'lr': 0.00030475, 'samples': 624640, 'steps': 1219, 'loss/train': 4.400296211242676} 02/24/2022 03:10:29 - INFO - codeparrot_training - Step 1220: {'lr': 0.000305, 'samples': 625152, 'steps': 1220, 'loss/train': 4.083799362182617} 02/24/2022 03:10:32 - INFO - codeparrot_training - Step 1221: {'lr': 0.00030525, 'samples': 625664, 'steps': 1221, 'loss/train': 4.796741485595703} 02/24/2022 03:10:38 - INFO - codeparrot_training - Step 1222: {'lr': 0.0003055, 'samples': 626176, 'steps': 1222, 'loss/train': 5.3789896965026855} 02/24/2022 03:10:41 - INFO - codeparrot_training - Step 1223: {'lr': 0.00030575000000000003, 'samples': 626688, 'steps': 1223, 'loss/train': 5.484334945678711} 02/24/2022 03:10:47 - INFO - codeparrot_training - Step 1224: {'lr': 0.000306, 'samples': 627200, 'steps': 1224, 'loss/train': 4.20302677154541} 02/24/2022 03:10:50 - INFO - codeparrot_training - Step 1225: {'lr': 0.00030625000000000004, 'samples': 627712, 'steps': 1225, 'loss/train': 7.1518049240112305} 02/24/2022 03:10:54 - INFO - codeparrot_training - Step 1226: {'lr': 0.0003065, 'samples': 628224, 'steps': 1226, 'loss/train': 4.745423793792725} 02/24/2022 03:11:00 - INFO - codeparrot_training - Step 1227: {'lr': 0.00030675, 'samples': 628736, 'steps': 1227, 'loss/train': 5.093043804168701} 02/24/2022 03:11:06 - INFO - codeparrot_training - Step 1228: {'lr': 0.000307, 'samples': 629248, 'steps': 1228, 'loss/train': 5.976129531860352} 02/24/2022 03:11:09 - INFO - codeparrot_training - Step 1229: {'lr': 0.00030725, 'samples': 629760, 'steps': 1229, 'loss/train': 5.111237049102783} 02/24/2022 03:11:15 - INFO - codeparrot_training - Step 1230: {'lr': 0.0003075, 'samples': 630272, 'steps': 1230, 'loss/train': 10.158740997314453} 02/24/2022 03:11:18 - INFO - codeparrot_training - Step 1231: {'lr': 0.00030775, 'samples': 630784, 'steps': 1231, 'loss/train': 4.465982437133789} 02/24/2022 03:11:24 - INFO - codeparrot_training - Step 1232: {'lr': 0.000308, 'samples': 631296, 'steps': 1232, 'loss/train': 5.0771260261535645} 02/24/2022 03:11:27 - INFO - codeparrot_training - Step 1233: {'lr': 0.00030825000000000004, 'samples': 631808, 'steps': 1233, 'loss/train': 4.414144039154053} 02/24/2022 03:11:33 - INFO - codeparrot_training - Step 1234: {'lr': 0.0003085, 'samples': 632320, 'steps': 1234, 'loss/train': 3.978367805480957} 02/24/2022 03:11:36 - INFO - codeparrot_training - Step 1235: {'lr': 0.00030875000000000005, 'samples': 632832, 'steps': 1235, 'loss/train': 4.054947376251221} 02/24/2022 03:11:42 - INFO - codeparrot_training - Step 1236: {'lr': 0.00030900000000000003, 'samples': 633344, 'steps': 1236, 'loss/train': 4.702358722686768} 02/24/2022 03:11:45 - INFO - codeparrot_training - Step 1237: {'lr': 0.00030925, 'samples': 633856, 'steps': 1237, 'loss/train': 5.610233306884766} 02/24/2022 03:11:51 - INFO - codeparrot_training - Step 1238: {'lr': 0.0003095, 'samples': 634368, 'steps': 1238, 'loss/train': 5.236934661865234} 02/24/2022 03:11:54 - INFO - codeparrot_training - Step 1239: {'lr': 0.00030975, 'samples': 634880, 'steps': 1239, 'loss/train': 5.027810573577881} 02/24/2022 03:12:00 - INFO - codeparrot_training - Step 1240: {'lr': 0.00031, 'samples': 635392, 'steps': 1240, 'loss/train': 4.405148506164551} 02/24/2022 03:12:04 - INFO - codeparrot_training - Step 1241: {'lr': 0.00031025000000000003, 'samples': 635904, 'steps': 1241, 'loss/train': 4.1442670822143555} 02/24/2022 03:12:09 - INFO - codeparrot_training - Step 1242: {'lr': 0.0003105, 'samples': 636416, 'steps': 1242, 'loss/train': 5.2206597328186035} 02/24/2022 03:12:12 - INFO - codeparrot_training - Step 1243: {'lr': 0.00031075000000000005, 'samples': 636928, 'steps': 1243, 'loss/train': 4.97514009475708} 02/24/2022 03:12:18 - INFO - codeparrot_training - Step 1244: {'lr': 0.000311, 'samples': 637440, 'steps': 1244, 'loss/train': 3.7692489624023438} 02/24/2022 03:12:21 - INFO - codeparrot_training - Step 1245: {'lr': 0.00031125000000000006, 'samples': 637952, 'steps': 1245, 'loss/train': 4.279240131378174} 02/24/2022 03:12:27 - INFO - codeparrot_training - Step 1246: {'lr': 0.0003115, 'samples': 638464, 'steps': 1246, 'loss/train': 4.816704750061035} 02/24/2022 03:12:31 - INFO - codeparrot_training - Step 1247: {'lr': 0.00031175, 'samples': 638976, 'steps': 1247, 'loss/train': 4.269006729125977} 02/24/2022 03:12:36 - INFO - codeparrot_training - Step 1248: {'lr': 0.000312, 'samples': 639488, 'steps': 1248, 'loss/train': 4.934327602386475} 02/24/2022 03:12:40 - INFO - codeparrot_training - Step 1249: {'lr': 0.00031225000000000003, 'samples': 640000, 'steps': 1249, 'loss/train': 4.7373552322387695} 02/24/2022 03:12:46 - INFO - codeparrot_training - Step 1250: {'lr': 0.0003125, 'samples': 640512, 'steps': 1250, 'loss/train': 5.402671813964844} 02/24/2022 03:12:49 - INFO - codeparrot_training - Step 1251: {'lr': 0.00031275, 'samples': 641024, 'steps': 1251, 'loss/train': 5.435028553009033} 02/24/2022 03:12:55 - INFO - codeparrot_training - Step 1252: {'lr': 0.000313, 'samples': 641536, 'steps': 1252, 'loss/train': 4.6016435623168945} 02/24/2022 03:12:58 - INFO - codeparrot_training - Step 1253: {'lr': 0.00031325, 'samples': 642048, 'steps': 1253, 'loss/train': 3.774256706237793} 02/24/2022 03:13:04 - INFO - codeparrot_training - Step 1254: {'lr': 0.00031350000000000003, 'samples': 642560, 'steps': 1254, 'loss/train': 3.7604446411132812} 02/24/2022 03:13:07 - INFO - codeparrot_training - Step 1255: {'lr': 0.00031374999999999996, 'samples': 643072, 'steps': 1255, 'loss/train': 6.243778705596924} 02/24/2022 03:13:12 - INFO - codeparrot_training - Step 1256: {'lr': 0.000314, 'samples': 643584, 'steps': 1256, 'loss/train': 5.965482711791992} 02/24/2022 03:13:16 - INFO - codeparrot_training - Step 1257: {'lr': 0.00031424999999999997, 'samples': 644096, 'steps': 1257, 'loss/train': 5.064489841461182} 02/24/2022 03:13:21 - INFO - codeparrot_training - Step 1258: {'lr': 0.0003145, 'samples': 644608, 'steps': 1258, 'loss/train': 4.7462158203125} 02/24/2022 03:13:25 - INFO - codeparrot_training - Step 1259: {'lr': 0.00031475, 'samples': 645120, 'steps': 1259, 'loss/train': 4.177390098571777} 02/24/2022 03:13:31 - INFO - codeparrot_training - Step 1260: {'lr': 0.000315, 'samples': 645632, 'steps': 1260, 'loss/train': 3.2847282886505127} 02/24/2022 03:13:34 - INFO - codeparrot_training - Step 1261: {'lr': 0.00031525, 'samples': 646144, 'steps': 1261, 'loss/train': 2.996202230453491} 02/24/2022 03:13:41 - INFO - codeparrot_training - Step 1262: {'lr': 0.0003155, 'samples': 646656, 'steps': 1262, 'loss/train': 5.227932453155518} 02/24/2022 03:13:44 - INFO - codeparrot_training - Step 1263: {'lr': 0.00031575, 'samples': 647168, 'steps': 1263, 'loss/train': 5.454081058502197} 02/24/2022 03:13:50 - INFO - codeparrot_training - Step 1264: {'lr': 0.000316, 'samples': 647680, 'steps': 1264, 'loss/train': 4.7259626388549805} 02/24/2022 03:13:53 - INFO - codeparrot_training - Step 1265: {'lr': 0.00031624999999999996, 'samples': 648192, 'steps': 1265, 'loss/train': 6.212442874908447} 02/24/2022 03:13:59 - INFO - codeparrot_training - Step 1266: {'lr': 0.0003165, 'samples': 648704, 'steps': 1266, 'loss/train': 4.92418098449707} 02/24/2022 03:14:02 - INFO - codeparrot_training - Step 1267: {'lr': 0.00031675, 'samples': 649216, 'steps': 1267, 'loss/train': 5.104456901550293} 02/24/2022 03:14:08 - INFO - codeparrot_training - Step 1268: {'lr': 0.000317, 'samples': 649728, 'steps': 1268, 'loss/train': 5.967926979064941} 02/24/2022 03:14:11 - INFO - codeparrot_training - Step 1269: {'lr': 0.00031725, 'samples': 650240, 'steps': 1269, 'loss/train': 4.988691806793213} 02/24/2022 03:14:17 - INFO - codeparrot_training - Step 1270: {'lr': 0.0003175, 'samples': 650752, 'steps': 1270, 'loss/train': 5.1615118980407715} 02/24/2022 03:14:20 - INFO - codeparrot_training - Step 1271: {'lr': 0.00031775, 'samples': 651264, 'steps': 1271, 'loss/train': 6.670093059539795} 02/24/2022 03:14:27 - INFO - codeparrot_training - Step 1272: {'lr': 0.00031800000000000003, 'samples': 651776, 'steps': 1272, 'loss/train': 6.492274761199951} 02/24/2022 03:14:30 - INFO - codeparrot_training - Step 1273: {'lr': 0.00031825, 'samples': 652288, 'steps': 1273, 'loss/train': 1.6502529382705688} 02/24/2022 03:14:35 - INFO - codeparrot_training - Step 1274: {'lr': 0.0003185, 'samples': 652800, 'steps': 1274, 'loss/train': 5.201275825500488} 02/24/2022 03:14:39 - INFO - codeparrot_training - Step 1275: {'lr': 0.00031874999999999997, 'samples': 653312, 'steps': 1275, 'loss/train': 2.714742422103882} 02/24/2022 03:14:44 - INFO - codeparrot_training - Step 1276: {'lr': 0.000319, 'samples': 653824, 'steps': 1276, 'loss/train': 1.6181998252868652} 02/24/2022 03:14:48 - INFO - codeparrot_training - Step 1277: {'lr': 0.00031925, 'samples': 654336, 'steps': 1277, 'loss/train': 4.431972026824951} 02/24/2022 03:14:53 - INFO - codeparrot_training - Step 1278: {'lr': 0.0003195, 'samples': 654848, 'steps': 1278, 'loss/train': 5.74703311920166} 02/24/2022 03:14:57 - INFO - codeparrot_training - Step 1279: {'lr': 0.00031975, 'samples': 655360, 'steps': 1279, 'loss/train': 3.972041606903076} 02/24/2022 03:15:02 - INFO - codeparrot_training - Step 1280: {'lr': 0.00032, 'samples': 655872, 'steps': 1280, 'loss/train': 4.932854175567627} 02/24/2022 03:15:06 - INFO - codeparrot_training - Step 1281: {'lr': 0.00032025, 'samples': 656384, 'steps': 1281, 'loss/train': 5.711575508117676} 02/24/2022 03:15:11 - INFO - codeparrot_training - Step 1282: {'lr': 0.00032050000000000004, 'samples': 656896, 'steps': 1282, 'loss/train': 4.543326377868652} 02/24/2022 03:15:15 - INFO - codeparrot_training - Step 1283: {'lr': 0.00032074999999999996, 'samples': 657408, 'steps': 1283, 'loss/train': 3.9023425579071045} 02/24/2022 03:15:20 - INFO - codeparrot_training - Step 1284: {'lr': 0.000321, 'samples': 657920, 'steps': 1284, 'loss/train': 4.438559055328369} 02/24/2022 03:15:24 - INFO - codeparrot_training - Step 1285: {'lr': 0.00032125, 'samples': 658432, 'steps': 1285, 'loss/train': 5.090837001800537} 02/24/2022 03:15:29 - INFO - codeparrot_training - Step 1286: {'lr': 0.0003215, 'samples': 658944, 'steps': 1286, 'loss/train': 3.9492552280426025} 02/24/2022 03:15:32 - INFO - codeparrot_training - Step 1287: {'lr': 0.00032175, 'samples': 659456, 'steps': 1287, 'loss/train': 4.6581549644470215} 02/24/2022 03:15:39 - INFO - codeparrot_training - Step 1288: {'lr': 0.000322, 'samples': 659968, 'steps': 1288, 'loss/train': 5.21675443649292} 02/24/2022 03:15:42 - INFO - codeparrot_training - Step 1289: {'lr': 0.00032225, 'samples': 660480, 'steps': 1289, 'loss/train': 5.406274318695068} 02/24/2022 03:15:48 - INFO - codeparrot_training - Step 1290: {'lr': 0.00032250000000000003, 'samples': 660992, 'steps': 1290, 'loss/train': 5.3217597007751465} 02/24/2022 03:15:51 - INFO - codeparrot_training - Step 1291: {'lr': 0.00032275, 'samples': 661504, 'steps': 1291, 'loss/train': 5.154272556304932} 02/24/2022 03:15:57 - INFO - codeparrot_training - Step 1292: {'lr': 0.000323, 'samples': 662016, 'steps': 1292, 'loss/train': 4.131954193115234} 02/24/2022 03:16:00 - INFO - codeparrot_training - Step 1293: {'lr': 0.00032324999999999997, 'samples': 662528, 'steps': 1293, 'loss/train': 4.235006332397461} 02/24/2022 03:16:06 - INFO - codeparrot_training - Step 1294: {'lr': 0.0003235, 'samples': 663040, 'steps': 1294, 'loss/train': 5.352925777435303} 02/24/2022 03:16:09 - INFO - codeparrot_training - Step 1295: {'lr': 0.00032375, 'samples': 663552, 'steps': 1295, 'loss/train': 4.743350505828857} 02/24/2022 03:16:15 - INFO - codeparrot_training - Step 1296: {'lr': 0.000324, 'samples': 664064, 'steps': 1296, 'loss/train': 5.727532863616943} 02/24/2022 03:16:21 - INFO - codeparrot_training - Step 1297: {'lr': 0.00032425, 'samples': 664576, 'steps': 1297, 'loss/train': 5.753353118896484} 02/24/2022 03:16:25 - INFO - codeparrot_training - Step 1298: {'lr': 0.00032450000000000003, 'samples': 665088, 'steps': 1298, 'loss/train': 4.1519455909729} 02/24/2022 03:16:31 - INFO - codeparrot_training - Step 1299: {'lr': 0.00032475, 'samples': 665600, 'steps': 1299, 'loss/train': 3.3136136531829834} 02/24/2022 03:16:34 - INFO - codeparrot_training - Step 1300: {'lr': 0.00032500000000000004, 'samples': 666112, 'steps': 1300, 'loss/train': 5.631933212280273} 02/24/2022 03:16:39 - INFO - codeparrot_training - Step 1301: {'lr': 0.00032524999999999996, 'samples': 666624, 'steps': 1301, 'loss/train': 3.7394909858703613} 02/24/2022 03:16:43 - INFO - codeparrot_training - Step 1302: {'lr': 0.0003255, 'samples': 667136, 'steps': 1302, 'loss/train': 4.254319190979004} 02/24/2022 03:16:49 - INFO - codeparrot_training - Step 1303: {'lr': 0.00032575, 'samples': 667648, 'steps': 1303, 'loss/train': 4.715116024017334} 02/24/2022 03:16:52 - INFO - codeparrot_training - Step 1304: {'lr': 0.000326, 'samples': 668160, 'steps': 1304, 'loss/train': 4.737923622131348} 02/24/2022 03:16:57 - INFO - codeparrot_training - Step 1305: {'lr': 0.00032625, 'samples': 668672, 'steps': 1305, 'loss/train': 5.72030782699585} 02/24/2022 03:17:01 - INFO - codeparrot_training - Step 1306: {'lr': 0.0003265, 'samples': 669184, 'steps': 1306, 'loss/train': 4.006705284118652} 02/24/2022 03:17:08 - INFO - codeparrot_training - Step 1307: {'lr': 0.00032675, 'samples': 669696, 'steps': 1307, 'loss/train': 5.079216003417969} 02/24/2022 03:17:11 - INFO - codeparrot_training - Step 1308: {'lr': 0.00032700000000000003, 'samples': 670208, 'steps': 1308, 'loss/train': 4.970621109008789} 02/24/2022 03:17:17 - INFO - codeparrot_training - Step 1309: {'lr': 0.00032725, 'samples': 670720, 'steps': 1309, 'loss/train': 3.7042927742004395} 02/24/2022 03:17:20 - INFO - codeparrot_training - Step 1310: {'lr': 0.00032750000000000005, 'samples': 671232, 'steps': 1310, 'loss/train': 1.5207202434539795} 02/24/2022 03:17:26 - INFO - codeparrot_training - Step 1311: {'lr': 0.00032774999999999997, 'samples': 671744, 'steps': 1311, 'loss/train': 4.482362270355225} 02/24/2022 03:17:29 - INFO - codeparrot_training - Step 1312: {'lr': 0.000328, 'samples': 672256, 'steps': 1312, 'loss/train': 4.10011100769043} 02/24/2022 03:17:35 - INFO - codeparrot_training - Step 1313: {'lr': 0.00032825, 'samples': 672768, 'steps': 1313, 'loss/train': 5.036800861358643} 02/24/2022 03:17:38 - INFO - codeparrot_training - Step 1314: {'lr': 0.0003285, 'samples': 673280, 'steps': 1314, 'loss/train': 4.329094886779785} 02/24/2022 03:17:43 - INFO - codeparrot_training - Step 1315: {'lr': 0.00032875, 'samples': 673792, 'steps': 1315, 'loss/train': 5.421314716339111} 02/24/2022 03:17:47 - INFO - codeparrot_training - Step 1316: {'lr': 0.00032900000000000003, 'samples': 674304, 'steps': 1316, 'loss/train': 3.2206647396087646} 02/24/2022 03:17:53 - INFO - codeparrot_training - Step 1317: {'lr': 0.00032925, 'samples': 674816, 'steps': 1317, 'loss/train': 4.149138450622559} 02/24/2022 03:17:57 - INFO - codeparrot_training - Step 1318: {'lr': 0.00032950000000000004, 'samples': 675328, 'steps': 1318, 'loss/train': 4.594919681549072} 02/24/2022 03:18:02 - INFO - codeparrot_training - Step 1319: {'lr': 0.00032975, 'samples': 675840, 'steps': 1319, 'loss/train': 2.8470685482025146} 02/24/2022 03:18:06 - INFO - codeparrot_training - Step 1320: {'lr': 0.00033, 'samples': 676352, 'steps': 1320, 'loss/train': 3.9833788871765137} 02/24/2022 03:18:09 - INFO - codeparrot_training - Step 1321: {'lr': 0.00033025, 'samples': 676864, 'steps': 1321, 'loss/train': 4.424703121185303} 02/24/2022 03:18:15 - INFO - codeparrot_training - Step 1322: {'lr': 0.0003305, 'samples': 677376, 'steps': 1322, 'loss/train': 4.049134254455566} 02/24/2022 03:18:18 - INFO - codeparrot_training - Step 1323: {'lr': 0.00033075, 'samples': 677888, 'steps': 1323, 'loss/train': 4.732639312744141} 02/24/2022 03:18:24 - INFO - codeparrot_training - Step 1324: {'lr': 0.000331, 'samples': 678400, 'steps': 1324, 'loss/train': 5.325099468231201} 02/24/2022 03:18:29 - INFO - codeparrot_training - Step 1325: {'lr': 0.00033125, 'samples': 678912, 'steps': 1325, 'loss/train': 4.985098838806152} 02/24/2022 03:18:32 - INFO - codeparrot_training - Step 1326: {'lr': 0.00033150000000000003, 'samples': 679424, 'steps': 1326, 'loss/train': 5.548741340637207} 02/24/2022 03:18:38 - INFO - codeparrot_training - Step 1327: {'lr': 0.00033175, 'samples': 679936, 'steps': 1327, 'loss/train': 4.608846664428711} 02/24/2022 03:18:41 - INFO - codeparrot_training - Step 1328: {'lr': 0.00033200000000000005, 'samples': 680448, 'steps': 1328, 'loss/train': 6.626372814178467} 02/24/2022 03:18:47 - INFO - codeparrot_training - Step 1329: {'lr': 0.00033224999999999997, 'samples': 680960, 'steps': 1329, 'loss/train': 4.52178955078125} 02/24/2022 03:18:50 - INFO - codeparrot_training - Step 1330: {'lr': 0.0003325, 'samples': 681472, 'steps': 1330, 'loss/train': 4.231669902801514} 02/24/2022 03:18:56 - INFO - codeparrot_training - Step 1331: {'lr': 0.00033275, 'samples': 681984, 'steps': 1331, 'loss/train': 5.275118350982666} 02/24/2022 03:18:59 - INFO - codeparrot_training - Step 1332: {'lr': 0.000333, 'samples': 682496, 'steps': 1332, 'loss/train': 1.3557944297790527} 02/24/2022 03:19:06 - INFO - codeparrot_training - Step 1333: {'lr': 0.00033325, 'samples': 683008, 'steps': 1333, 'loss/train': 4.469725608825684} 02/24/2022 03:19:09 - INFO - codeparrot_training - Step 1334: {'lr': 0.00033350000000000003, 'samples': 683520, 'steps': 1334, 'loss/train': 3.57256817817688} 02/24/2022 03:19:15 - INFO - codeparrot_training - Step 1335: {'lr': 0.00033375, 'samples': 684032, 'steps': 1335, 'loss/train': 4.704814910888672} 02/24/2022 03:19:19 - INFO - codeparrot_training - Step 1336: {'lr': 0.00033400000000000004, 'samples': 684544, 'steps': 1336, 'loss/train': 4.015722274780273} 02/24/2022 03:19:24 - INFO - codeparrot_training - Step 1337: {'lr': 0.00033425, 'samples': 685056, 'steps': 1337, 'loss/train': 4.798593521118164} 02/24/2022 03:19:27 - INFO - codeparrot_training - Step 1338: {'lr': 0.00033450000000000005, 'samples': 685568, 'steps': 1338, 'loss/train': 4.522674560546875} 02/24/2022 03:19:33 - INFO - codeparrot_training - Step 1339: {'lr': 0.00033475, 'samples': 686080, 'steps': 1339, 'loss/train': 4.385733127593994} 02/24/2022 03:19:36 - INFO - codeparrot_training - Step 1340: {'lr': 0.000335, 'samples': 686592, 'steps': 1340, 'loss/train': 5.581892967224121} 02/24/2022 03:19:42 - INFO - codeparrot_training - Step 1341: {'lr': 0.00033525, 'samples': 687104, 'steps': 1341, 'loss/train': 5.506782054901123} 02/24/2022 03:19:45 - INFO - codeparrot_training - Step 1342: {'lr': 0.0003355, 'samples': 687616, 'steps': 1342, 'loss/train': 4.064359664916992} 02/24/2022 03:19:52 - INFO - codeparrot_training - Step 1343: {'lr': 0.00033575, 'samples': 688128, 'steps': 1343, 'loss/train': 6.217005729675293} 02/24/2022 03:19:55 - INFO - codeparrot_training - Step 1344: {'lr': 0.00033600000000000004, 'samples': 688640, 'steps': 1344, 'loss/train': 3.9408841133117676} 02/24/2022 03:20:01 - INFO - codeparrot_training - Step 1345: {'lr': 0.00033625, 'samples': 689152, 'steps': 1345, 'loss/train': 4.430418014526367} 02/24/2022 03:20:04 - INFO - codeparrot_training - Step 1346: {'lr': 0.00033650000000000005, 'samples': 689664, 'steps': 1346, 'loss/train': 3.7917208671569824} 02/24/2022 03:20:10 - INFO - codeparrot_training - Step 1347: {'lr': 0.00033675, 'samples': 690176, 'steps': 1347, 'loss/train': 5.144470691680908} 02/24/2022 03:20:13 - INFO - codeparrot_training - Step 1348: {'lr': 0.000337, 'samples': 690688, 'steps': 1348, 'loss/train': 4.643494129180908} 02/24/2022 03:20:19 - INFO - codeparrot_training - Step 1349: {'lr': 0.00033725, 'samples': 691200, 'steps': 1349, 'loss/train': 5.293832302093506} 02/24/2022 03:20:22 - INFO - codeparrot_training - Step 1350: {'lr': 0.0003375, 'samples': 691712, 'steps': 1350, 'loss/train': 4.75413179397583} 02/24/2022 03:20:28 - INFO - codeparrot_training - Step 1351: {'lr': 0.00033775, 'samples': 692224, 'steps': 1351, 'loss/train': 4.326086044311523} 02/24/2022 03:20:31 - INFO - codeparrot_training - Step 1352: {'lr': 0.00033800000000000003, 'samples': 692736, 'steps': 1352, 'loss/train': 4.840494632720947} 02/24/2022 03:20:37 - INFO - codeparrot_training - Step 1353: {'lr': 0.00033825, 'samples': 693248, 'steps': 1353, 'loss/train': 4.518887996673584} 02/24/2022 03:20:41 - INFO - codeparrot_training - Step 1354: {'lr': 0.00033850000000000004, 'samples': 693760, 'steps': 1354, 'loss/train': 6.481963157653809} 02/24/2022 03:20:46 - INFO - codeparrot_training - Step 1355: {'lr': 0.00033875, 'samples': 694272, 'steps': 1355, 'loss/train': 6.150463104248047} 02/24/2022 03:20:50 - INFO - codeparrot_training - Step 1356: {'lr': 0.00033900000000000005, 'samples': 694784, 'steps': 1356, 'loss/train': 3.5370700359344482} 02/24/2022 03:20:55 - INFO - codeparrot_training - Step 1357: {'lr': 0.00033925, 'samples': 695296, 'steps': 1357, 'loss/train': 4.54744815826416} 02/24/2022 03:20:59 - INFO - codeparrot_training - Step 1358: {'lr': 0.0003395, 'samples': 695808, 'steps': 1358, 'loss/train': 3.866886615753174} 02/24/2022 03:21:04 - INFO - codeparrot_training - Step 1359: {'lr': 0.00033975, 'samples': 696320, 'steps': 1359, 'loss/train': 4.2036848068237305} 02/24/2022 03:21:08 - INFO - codeparrot_training - Step 1360: {'lr': 0.00034, 'samples': 696832, 'steps': 1360, 'loss/train': 4.554596424102783} 02/24/2022 03:21:13 - INFO - codeparrot_training - Step 1361: {'lr': 0.00034025, 'samples': 697344, 'steps': 1361, 'loss/train': 4.297909736633301} 02/24/2022 03:21:17 - INFO - codeparrot_training - Step 1362: {'lr': 0.00034050000000000004, 'samples': 697856, 'steps': 1362, 'loss/train': 4.0271687507629395} 02/24/2022 03:21:23 - INFO - codeparrot_training - Step 1363: {'lr': 0.00034075, 'samples': 698368, 'steps': 1363, 'loss/train': 3.2887182235717773} 02/24/2022 03:21:26 - INFO - codeparrot_training - Step 1364: {'lr': 0.00034100000000000005, 'samples': 698880, 'steps': 1364, 'loss/train': 4.91015625} 02/24/2022 03:21:32 - INFO - codeparrot_training - Step 1365: {'lr': 0.00034125000000000003, 'samples': 699392, 'steps': 1365, 'loss/train': 3.5127220153808594} 02/24/2022 03:21:35 - INFO - codeparrot_training - Step 1366: {'lr': 0.0003415, 'samples': 699904, 'steps': 1366, 'loss/train': 4.21429967880249} 02/24/2022 03:21:41 - INFO - codeparrot_training - Step 1367: {'lr': 0.00034175, 'samples': 700416, 'steps': 1367, 'loss/train': 4.592789173126221} 02/24/2022 03:21:44 - INFO - codeparrot_training - Step 1368: {'lr': 0.000342, 'samples': 700928, 'steps': 1368, 'loss/train': 4.821196556091309} 02/24/2022 03:21:50 - INFO - codeparrot_training - Step 1369: {'lr': 0.00034225, 'samples': 701440, 'steps': 1369, 'loss/train': 4.770508289337158} 02/24/2022 03:21:54 - INFO - codeparrot_training - Step 1370: {'lr': 0.00034250000000000003, 'samples': 701952, 'steps': 1370, 'loss/train': 2.9974818229675293} 02/24/2022 03:21:59 - INFO - codeparrot_training - Step 1371: {'lr': 0.00034275, 'samples': 702464, 'steps': 1371, 'loss/train': 4.845798969268799} 02/24/2022 03:22:02 - INFO - codeparrot_training - Step 1372: {'lr': 0.00034300000000000004, 'samples': 702976, 'steps': 1372, 'loss/train': 7.2618513107299805} 02/24/2022 03:22:08 - INFO - codeparrot_training - Step 1373: {'lr': 0.00034325, 'samples': 703488, 'steps': 1373, 'loss/train': 5.582989692687988} 02/24/2022 03:22:12 - INFO - codeparrot_training - Step 1374: {'lr': 0.00034350000000000006, 'samples': 704000, 'steps': 1374, 'loss/train': 4.1638641357421875} 02/24/2022 03:22:17 - INFO - codeparrot_training - Step 1375: {'lr': 0.00034375, 'samples': 704512, 'steps': 1375, 'loss/train': 4.44398832321167} 02/24/2022 03:22:21 - INFO - codeparrot_training - Step 1376: {'lr': 0.00034399999999999996, 'samples': 705024, 'steps': 1376, 'loss/train': 5.0551042556762695} 02/24/2022 03:22:26 - INFO - codeparrot_training - Step 1377: {'lr': 0.00034425, 'samples': 705536, 'steps': 1377, 'loss/train': 4.392722129821777} 02/24/2022 03:22:29 - INFO - codeparrot_training - Step 1378: {'lr': 0.00034449999999999997, 'samples': 706048, 'steps': 1378, 'loss/train': 3.820713758468628} 02/24/2022 03:22:35 - INFO - codeparrot_training - Step 1379: {'lr': 0.00034475, 'samples': 706560, 'steps': 1379, 'loss/train': 4.039594650268555} 02/24/2022 03:22:39 - INFO - codeparrot_training - Step 1380: {'lr': 0.000345, 'samples': 707072, 'steps': 1380, 'loss/train': 5.738271713256836} 02/24/2022 03:22:44 - INFO - codeparrot_training - Step 1381: {'lr': 0.00034525, 'samples': 707584, 'steps': 1381, 'loss/train': 4.054514408111572} 02/24/2022 03:22:48 - INFO - codeparrot_training - Step 1382: {'lr': 0.0003455, 'samples': 708096, 'steps': 1382, 'loss/train': 3.326011896133423} 02/24/2022 03:22:53 - INFO - codeparrot_training - Step 1383: {'lr': 0.00034575000000000003, 'samples': 708608, 'steps': 1383, 'loss/train': 4.194097995758057} 02/24/2022 03:22:57 - INFO - codeparrot_training - Step 1384: {'lr': 0.000346, 'samples': 709120, 'steps': 1384, 'loss/train': 3.6856579780578613} 02/24/2022 03:23:02 - INFO - codeparrot_training - Step 1385: {'lr': 0.00034625, 'samples': 709632, 'steps': 1385, 'loss/train': 4.916198253631592} 02/24/2022 03:23:06 - INFO - codeparrot_training - Step 1386: {'lr': 0.00034649999999999997, 'samples': 710144, 'steps': 1386, 'loss/train': 5.696788311004639} 02/24/2022 03:23:11 - INFO - codeparrot_training - Step 1387: {'lr': 0.00034675, 'samples': 710656, 'steps': 1387, 'loss/train': 3.7698769569396973} 02/24/2022 03:23:15 - INFO - codeparrot_training - Step 1388: {'lr': 0.000347, 'samples': 711168, 'steps': 1388, 'loss/train': 4.747680187225342} 02/24/2022 03:23:20 - INFO - codeparrot_training - Step 1389: {'lr': 0.00034725, 'samples': 711680, 'steps': 1389, 'loss/train': 3.7353427410125732} 02/24/2022 03:23:24 - INFO - codeparrot_training - Step 1390: {'lr': 0.0003475, 'samples': 712192, 'steps': 1390, 'loss/train': 4.51666259765625} 02/24/2022 03:23:30 - INFO - codeparrot_training - Step 1391: {'lr': 0.00034775, 'samples': 712704, 'steps': 1391, 'loss/train': 4.353278636932373} 02/24/2022 03:23:33 - INFO - codeparrot_training - Step 1392: {'lr': 0.000348, 'samples': 713216, 'steps': 1392, 'loss/train': 3.0993151664733887} 02/24/2022 03:23:39 - INFO - codeparrot_training - Step 1393: {'lr': 0.00034825000000000004, 'samples': 713728, 'steps': 1393, 'loss/train': 4.803194522857666} 02/24/2022 03:23:44 - INFO - codeparrot_training - Step 1394: {'lr': 0.00034849999999999996, 'samples': 714240, 'steps': 1394, 'loss/train': 3.762338876724243} 02/24/2022 03:23:48 - INFO - codeparrot_training - Step 1395: {'lr': 0.00034875, 'samples': 714752, 'steps': 1395, 'loss/train': 3.8587708473205566} 02/24/2022 03:23:53 - INFO - codeparrot_training - Step 1396: {'lr': 0.00034899999999999997, 'samples': 715264, 'steps': 1396, 'loss/train': 5.5884881019592285} 02/24/2022 03:23:57 - INFO - codeparrot_training - Step 1397: {'lr': 0.00034925, 'samples': 715776, 'steps': 1397, 'loss/train': 3.1794707775115967} 02/24/2022 03:24:02 - INFO - codeparrot_training - Step 1398: {'lr': 0.0003495, 'samples': 716288, 'steps': 1398, 'loss/train': 4.749282360076904} 02/24/2022 03:24:06 - INFO - codeparrot_training - Step 1399: {'lr': 0.00034975, 'samples': 716800, 'steps': 1399, 'loss/train': 4.140390396118164} 02/24/2022 03:24:12 - INFO - codeparrot_training - Step 1400: {'lr': 0.00035, 'samples': 717312, 'steps': 1400, 'loss/train': 5.647134304046631} 02/24/2022 03:24:15 - INFO - codeparrot_training - Step 1401: {'lr': 0.00035025000000000003, 'samples': 717824, 'steps': 1401, 'loss/train': 4.211458683013916} 02/24/2022 03:24:21 - INFO - codeparrot_training - Step 1402: {'lr': 0.0003505, 'samples': 718336, 'steps': 1402, 'loss/train': 4.266910552978516} 02/24/2022 03:24:24 - INFO - codeparrot_training - Step 1403: {'lr': 0.00035075, 'samples': 718848, 'steps': 1403, 'loss/train': 2.2526164054870605} 02/24/2022 03:24:30 - INFO - codeparrot_training - Step 1404: {'lr': 0.00035099999999999997, 'samples': 719360, 'steps': 1404, 'loss/train': 4.09318208694458} 02/24/2022 03:24:33 - INFO - codeparrot_training - Step 1405: {'lr': 0.00035125, 'samples': 719872, 'steps': 1405, 'loss/train': 5.502644062042236} 02/24/2022 03:24:39 - INFO - codeparrot_training - Step 1406: {'lr': 0.0003515, 'samples': 720384, 'steps': 1406, 'loss/train': 3.886813163757324} 02/24/2022 03:24:42 - INFO - codeparrot_training - Step 1407: {'lr': 0.00035175, 'samples': 720896, 'steps': 1407, 'loss/train': 4.046776294708252} 02/24/2022 03:24:48 - INFO - codeparrot_training - Step 1408: {'lr': 0.000352, 'samples': 721408, 'steps': 1408, 'loss/train': 4.324183940887451} 02/24/2022 03:24:51 - INFO - codeparrot_training - Step 1409: {'lr': 0.00035225, 'samples': 721920, 'steps': 1409, 'loss/train': 4.107088565826416} 02/24/2022 03:24:57 - INFO - codeparrot_training - Step 1410: {'lr': 0.0003525, 'samples': 722432, 'steps': 1410, 'loss/train': 4.725785732269287} 02/24/2022 03:25:01 - INFO - codeparrot_training - Step 1411: {'lr': 0.00035275000000000004, 'samples': 722944, 'steps': 1411, 'loss/train': 4.506245136260986} 02/24/2022 03:25:06 - INFO - codeparrot_training - Step 1412: {'lr': 0.00035299999999999996, 'samples': 723456, 'steps': 1412, 'loss/train': 4.756691932678223} 02/24/2022 03:25:10 - INFO - codeparrot_training - Step 1413: {'lr': 0.00035325, 'samples': 723968, 'steps': 1413, 'loss/train': 4.609408378601074} 02/24/2022 03:25:15 - INFO - codeparrot_training - Step 1414: {'lr': 0.0003535, 'samples': 724480, 'steps': 1414, 'loss/train': 4.31541109085083} 02/24/2022 03:25:19 - INFO - codeparrot_training - Step 1415: {'lr': 0.00035375, 'samples': 724992, 'steps': 1415, 'loss/train': 4.767493724822998} 02/24/2022 03:25:24 - INFO - codeparrot_training - Step 1416: {'lr': 0.000354, 'samples': 725504, 'steps': 1416, 'loss/train': 5.0047760009765625} 02/24/2022 03:25:28 - INFO - codeparrot_training - Step 1417: {'lr': 0.00035425, 'samples': 726016, 'steps': 1417, 'loss/train': 8.275381088256836} 02/24/2022 03:25:33 - INFO - codeparrot_training - Step 1418: {'lr': 0.0003545, 'samples': 726528, 'steps': 1418, 'loss/train': 4.996078968048096} 02/24/2022 03:25:36 - INFO - codeparrot_training - Step 1419: {'lr': 0.00035475000000000003, 'samples': 727040, 'steps': 1419, 'loss/train': 4.269710540771484} 02/24/2022 03:25:42 - INFO - codeparrot_training - Step 1420: {'lr': 0.000355, 'samples': 727552, 'steps': 1420, 'loss/train': 1.2767610549926758} 02/24/2022 03:25:45 - INFO - codeparrot_training - Step 1421: {'lr': 0.00035525000000000004, 'samples': 728064, 'steps': 1421, 'loss/train': 4.295624256134033} 02/24/2022 03:25:51 - INFO - codeparrot_training - Step 1422: {'lr': 0.00035549999999999997, 'samples': 728576, 'steps': 1422, 'loss/train': 5.73574686050415} 02/24/2022 03:25:54 - INFO - codeparrot_training - Step 1423: {'lr': 0.00035575, 'samples': 729088, 'steps': 1423, 'loss/train': 3.591066598892212} 02/24/2022 03:26:00 - INFO - codeparrot_training - Step 1424: {'lr': 0.000356, 'samples': 729600, 'steps': 1424, 'loss/train': 4.418255805969238} 02/24/2022 03:26:03 - INFO - codeparrot_training - Step 1425: {'lr': 0.00035625, 'samples': 730112, 'steps': 1425, 'loss/train': 1.3538577556610107} 02/24/2022 03:26:10 - INFO - codeparrot_training - Step 1426: {'lr': 0.0003565, 'samples': 730624, 'steps': 1426, 'loss/train': 4.249266147613525} 02/24/2022 03:26:13 - INFO - codeparrot_training - Step 1427: {'lr': 0.00035675, 'samples': 731136, 'steps': 1427, 'loss/train': 4.449369430541992} 02/24/2022 03:26:19 - INFO - codeparrot_training - Step 1428: {'lr': 0.000357, 'samples': 731648, 'steps': 1428, 'loss/train': 4.438675403594971} 02/24/2022 03:26:22 - INFO - codeparrot_training - Step 1429: {'lr': 0.00035725000000000004, 'samples': 732160, 'steps': 1429, 'loss/train': 4.605688095092773} 02/24/2022 03:26:28 - INFO - codeparrot_training - Step 1430: {'lr': 0.0003575, 'samples': 732672, 'steps': 1430, 'loss/train': 5.555391311645508} 02/24/2022 03:26:31 - INFO - codeparrot_training - Step 1431: {'lr': 0.00035775, 'samples': 733184, 'steps': 1431, 'loss/train': 2.1674280166625977} 02/24/2022 03:26:37 - INFO - codeparrot_training - Step 1432: {'lr': 0.000358, 'samples': 733696, 'steps': 1432, 'loss/train': 3.172797441482544} 02/24/2022 03:26:40 - INFO - codeparrot_training - Step 1433: {'lr': 0.00035825, 'samples': 734208, 'steps': 1433, 'loss/train': 4.789025783538818} 02/24/2022 03:26:46 - INFO - codeparrot_training - Step 1434: {'lr': 0.0003585, 'samples': 734720, 'steps': 1434, 'loss/train': 2.9427411556243896} 02/24/2022 03:26:49 - INFO - codeparrot_training - Step 1435: {'lr': 0.00035875, 'samples': 735232, 'steps': 1435, 'loss/train': 3.566638469696045} 02/24/2022 03:26:56 - INFO - codeparrot_training - Step 1436: {'lr': 0.000359, 'samples': 735744, 'steps': 1436, 'loss/train': 2.941498041152954} 02/24/2022 03:26:59 - INFO - codeparrot_training - Step 1437: {'lr': 0.00035925000000000003, 'samples': 736256, 'steps': 1437, 'loss/train': 4.232140064239502} 02/24/2022 03:27:05 - INFO - codeparrot_training - Step 1438: {'lr': 0.0003595, 'samples': 736768, 'steps': 1438, 'loss/train': 2.9871699810028076} 02/24/2022 03:27:08 - INFO - codeparrot_training - Step 1439: {'lr': 0.00035975000000000004, 'samples': 737280, 'steps': 1439, 'loss/train': 4.4560866355896} 02/24/2022 03:27:14 - INFO - codeparrot_training - Step 1440: {'lr': 0.00035999999999999997, 'samples': 737792, 'steps': 1440, 'loss/train': 4.410506725311279} 02/24/2022 03:27:17 - INFO - codeparrot_training - Step 1441: {'lr': 0.00036025, 'samples': 738304, 'steps': 1441, 'loss/train': 3.8036000728607178} 02/24/2022 03:27:23 - INFO - codeparrot_training - Step 1442: {'lr': 0.0003605, 'samples': 738816, 'steps': 1442, 'loss/train': 3.6715757846832275} 02/24/2022 03:27:26 - INFO - codeparrot_training - Step 1443: {'lr': 0.00036075, 'samples': 739328, 'steps': 1443, 'loss/train': 4.418936252593994} 02/24/2022 03:27:32 - INFO - codeparrot_training - Step 1444: {'lr': 0.000361, 'samples': 739840, 'steps': 1444, 'loss/train': 4.190974712371826} 02/24/2022 03:27:35 - INFO - codeparrot_training - Step 1445: {'lr': 0.00036125, 'samples': 740352, 'steps': 1445, 'loss/train': 4.757613182067871} 02/24/2022 03:27:41 - INFO - codeparrot_training - Step 1446: {'lr': 0.0003615, 'samples': 740864, 'steps': 1446, 'loss/train': 3.6293020248413086} 02/24/2022 03:27:45 - INFO - codeparrot_training - Step 1447: {'lr': 0.00036175000000000004, 'samples': 741376, 'steps': 1447, 'loss/train': 4.599608421325684} 02/24/2022 03:27:50 - INFO - codeparrot_training - Step 1448: {'lr': 0.000362, 'samples': 741888, 'steps': 1448, 'loss/train': 4.677462100982666} 02/24/2022 03:27:54 - INFO - codeparrot_training - Step 1449: {'lr': 0.00036225000000000005, 'samples': 742400, 'steps': 1449, 'loss/train': 3.832645893096924} 02/24/2022 03:27:59 - INFO - codeparrot_training - Step 1450: {'lr': 0.0003625, 'samples': 742912, 'steps': 1450, 'loss/train': 4.194993019104004} 02/24/2022 03:28:03 - INFO - codeparrot_training - Step 1451: {'lr': 0.00036275, 'samples': 743424, 'steps': 1451, 'loss/train': 4.173336982727051} 02/24/2022 03:28:08 - INFO - codeparrot_training - Step 1452: {'lr': 0.000363, 'samples': 743936, 'steps': 1452, 'loss/train': 5.092646598815918} 02/24/2022 03:28:14 - INFO - codeparrot_training - Step 1453: {'lr': 0.00036325, 'samples': 744448, 'steps': 1453, 'loss/train': 4.8541646003723145} 02/24/2022 03:28:17 - INFO - codeparrot_training - Step 1454: {'lr': 0.0003635, 'samples': 744960, 'steps': 1454, 'loss/train': 2.8798084259033203} 02/24/2022 03:28:24 - INFO - codeparrot_training - Step 1455: {'lr': 0.00036375000000000003, 'samples': 745472, 'steps': 1455, 'loss/train': 4.474486351013184} 02/24/2022 03:28:27 - INFO - codeparrot_training - Step 1456: {'lr': 0.000364, 'samples': 745984, 'steps': 1456, 'loss/train': 3.5735902786254883} 02/24/2022 03:28:33 - INFO - codeparrot_training - Step 1457: {'lr': 0.00036425000000000004, 'samples': 746496, 'steps': 1457, 'loss/train': 3.685084342956543} 02/24/2022 03:28:36 - INFO - codeparrot_training - Step 1458: {'lr': 0.0003645, 'samples': 747008, 'steps': 1458, 'loss/train': 3.5939478874206543} 02/24/2022 03:28:42 - INFO - codeparrot_training - Step 1459: {'lr': 0.00036475, 'samples': 747520, 'steps': 1459, 'loss/train': 4.002579689025879} 02/24/2022 03:28:45 - INFO - codeparrot_training - Step 1460: {'lr': 0.000365, 'samples': 748032, 'steps': 1460, 'loss/train': 4.364797115325928} 02/24/2022 03:28:50 - INFO - codeparrot_training - Step 1461: {'lr': 0.00036525, 'samples': 748544, 'steps': 1461, 'loss/train': 4.881330490112305} 02/24/2022 03:28:54 - INFO - codeparrot_training - Step 1462: {'lr': 0.0003655, 'samples': 749056, 'steps': 1462, 'loss/train': 4.901648998260498} 02/24/2022 03:28:59 - INFO - codeparrot_training - Step 1463: {'lr': 0.00036575, 'samples': 749568, 'steps': 1463, 'loss/train': 4.209507465362549} 02/24/2022 03:29:03 - INFO - codeparrot_training - Step 1464: {'lr': 0.000366, 'samples': 750080, 'steps': 1464, 'loss/train': 4.085138320922852} 02/24/2022 03:29:09 - INFO - codeparrot_training - Step 1465: {'lr': 0.00036625000000000004, 'samples': 750592, 'steps': 1465, 'loss/train': 5.039862155914307} 02/24/2022 03:29:12 - INFO - codeparrot_training - Step 1466: {'lr': 0.0003665, 'samples': 751104, 'steps': 1466, 'loss/train': 4.095442771911621} 02/24/2022 03:29:18 - INFO - codeparrot_training - Step 1467: {'lr': 0.00036675000000000005, 'samples': 751616, 'steps': 1467, 'loss/train': 9.734140396118164} 02/24/2022 03:29:21 - INFO - codeparrot_training - Step 1468: {'lr': 0.000367, 'samples': 752128, 'steps': 1468, 'loss/train': 3.7371835708618164} 02/24/2022 03:29:27 - INFO - codeparrot_training - Step 1469: {'lr': 0.00036725, 'samples': 752640, 'steps': 1469, 'loss/train': 4.368670463562012} 02/24/2022 03:29:30 - INFO - codeparrot_training - Step 1470: {'lr': 0.0003675, 'samples': 753152, 'steps': 1470, 'loss/train': 3.4546971321105957} 02/24/2022 03:29:36 - INFO - codeparrot_training - Step 1471: {'lr': 0.00036775, 'samples': 753664, 'steps': 1471, 'loss/train': 3.882535696029663} 02/24/2022 03:29:39 - INFO - codeparrot_training - Step 1472: {'lr': 0.000368, 'samples': 754176, 'steps': 1472, 'loss/train': 4.518320560455322} 02/24/2022 03:29:45 - INFO - codeparrot_training - Step 1473: {'lr': 0.00036825000000000003, 'samples': 754688, 'steps': 1473, 'loss/train': 4.773551940917969} 02/24/2022 03:29:48 - INFO - codeparrot_training - Step 1474: {'lr': 0.0003685, 'samples': 755200, 'steps': 1474, 'loss/train': 3.6483707427978516} 02/24/2022 03:29:53 - INFO - codeparrot_training - Step 1475: {'lr': 0.00036875000000000005, 'samples': 755712, 'steps': 1475, 'loss/train': 3.6492366790771484} 02/24/2022 03:29:57 - INFO - codeparrot_training - Step 1476: {'lr': 0.000369, 'samples': 756224, 'steps': 1476, 'loss/train': 4.988525390625} 02/24/2022 03:30:02 - INFO - codeparrot_training - Step 1477: {'lr': 0.00036925, 'samples': 756736, 'steps': 1477, 'loss/train': 4.443835735321045} 02/24/2022 03:30:06 - INFO - codeparrot_training - Step 1478: {'lr': 0.0003695, 'samples': 757248, 'steps': 1478, 'loss/train': 4.1710591316223145} 02/24/2022 03:30:11 - INFO - codeparrot_training - Step 1479: {'lr': 0.00036975, 'samples': 757760, 'steps': 1479, 'loss/train': 4.66081428527832} 02/24/2022 03:30:15 - INFO - codeparrot_training - Step 1480: {'lr': 0.00037, 'samples': 758272, 'steps': 1480, 'loss/train': 3.862652063369751} 02/24/2022 03:30:21 - INFO - codeparrot_training - Step 1481: {'lr': 0.00037025000000000003, 'samples': 758784, 'steps': 1481, 'loss/train': 6.220557689666748} 02/24/2022 03:30:26 - INFO - codeparrot_training - Step 1482: {'lr': 0.0003705, 'samples': 759296, 'steps': 1482, 'loss/train': 3.1184709072113037} 02/24/2022 03:30:30 - INFO - codeparrot_training - Step 1483: {'lr': 0.00037075000000000004, 'samples': 759808, 'steps': 1483, 'loss/train': 3.412602663040161} 02/24/2022 03:30:35 - INFO - codeparrot_training - Step 1484: {'lr': 0.000371, 'samples': 760320, 'steps': 1484, 'loss/train': 3.708744764328003} 02/24/2022 03:30:39 - INFO - codeparrot_training - Step 1485: {'lr': 0.00037125000000000005, 'samples': 760832, 'steps': 1485, 'loss/train': 4.213790416717529} 02/24/2022 03:30:44 - INFO - codeparrot_training - Step 1486: {'lr': 0.00037150000000000003, 'samples': 761344, 'steps': 1486, 'loss/train': 4.705002784729004} 02/24/2022 03:30:48 - INFO - codeparrot_training - Step 1487: {'lr': 0.00037175, 'samples': 761856, 'steps': 1487, 'loss/train': 5.278611660003662} 02/24/2022 03:30:53 - INFO - codeparrot_training - Step 1488: {'lr': 0.000372, 'samples': 762368, 'steps': 1488, 'loss/train': 3.840797185897827} 02/24/2022 03:30:57 - INFO - codeparrot_training - Step 1489: {'lr': 0.00037225, 'samples': 762880, 'steps': 1489, 'loss/train': 3.5598690509796143} 02/24/2022 03:31:03 - INFO - codeparrot_training - Step 1490: {'lr': 0.0003725, 'samples': 763392, 'steps': 1490, 'loss/train': 3.661531686782837} 02/24/2022 03:31:06 - INFO - codeparrot_training - Step 1491: {'lr': 0.00037275000000000003, 'samples': 763904, 'steps': 1491, 'loss/train': 4.435661792755127} 02/24/2022 03:31:12 - INFO - codeparrot_training - Step 1492: {'lr': 0.000373, 'samples': 764416, 'steps': 1492, 'loss/train': 4.603262901306152} 02/24/2022 03:31:15 - INFO - codeparrot_training - Step 1493: {'lr': 0.00037325000000000005, 'samples': 764928, 'steps': 1493, 'loss/train': 3.6793453693389893} 02/24/2022 03:31:21 - INFO - codeparrot_training - Step 1494: {'lr': 0.0003735, 'samples': 765440, 'steps': 1494, 'loss/train': 3.0818397998809814} 02/24/2022 03:31:24 - INFO - codeparrot_training - Step 1495: {'lr': 0.00037375000000000006, 'samples': 765952, 'steps': 1495, 'loss/train': 4.466464042663574} 02/24/2022 03:31:30 - INFO - codeparrot_training - Step 1496: {'lr': 0.000374, 'samples': 766464, 'steps': 1496, 'loss/train': 3.632730007171631} 02/24/2022 03:31:33 - INFO - codeparrot_training - Step 1497: {'lr': 0.00037425, 'samples': 766976, 'steps': 1497, 'loss/train': 1.4231302738189697} 02/24/2022 03:31:39 - INFO - codeparrot_training - Step 1498: {'lr': 0.0003745, 'samples': 767488, 'steps': 1498, 'loss/train': 4.891319751739502} 02/24/2022 03:31:42 - INFO - codeparrot_training - Step 1499: {'lr': 0.00037475000000000003, 'samples': 768000, 'steps': 1499, 'loss/train': 3.9893150329589844} 02/24/2022 03:31:48 - INFO - codeparrot_training - Step 1500: {'lr': 0.000375, 'samples': 768512, 'steps': 1500, 'loss/train': 2.9769890308380127} 02/24/2022 03:31:51 - INFO - codeparrot_training - Step 1501: {'lr': 0.00037525, 'samples': 769024, 'steps': 1501, 'loss/train': 3.0658252239227295} 02/24/2022 03:31:57 - INFO - codeparrot_training - Step 1502: {'lr': 0.0003755, 'samples': 769536, 'steps': 1502, 'loss/train': 4.065186500549316} 02/24/2022 03:32:01 - INFO - codeparrot_training - Step 1503: {'lr': 0.00037575, 'samples': 770048, 'steps': 1503, 'loss/train': 3.200500011444092} 02/24/2022 03:32:06 - INFO - codeparrot_training - Step 1504: {'lr': 0.00037600000000000003, 'samples': 770560, 'steps': 1504, 'loss/train': 4.314088821411133} 02/24/2022 03:32:10 - INFO - codeparrot_training - Step 1505: {'lr': 0.00037624999999999996, 'samples': 771072, 'steps': 1505, 'loss/train': 3.7412798404693604} 02/24/2022 03:32:15 - INFO - codeparrot_training - Step 1506: {'lr': 0.0003765, 'samples': 771584, 'steps': 1506, 'loss/train': 2.632077217102051} 02/24/2022 03:32:19 - INFO - codeparrot_training - Step 1507: {'lr': 0.00037674999999999997, 'samples': 772096, 'steps': 1507, 'loss/train': 4.561457633972168} 02/24/2022 03:32:24 - INFO - codeparrot_training - Step 1508: {'lr': 0.000377, 'samples': 772608, 'steps': 1508, 'loss/train': 2.8471598625183105} 02/24/2022 03:32:28 - INFO - codeparrot_training - Step 1509: {'lr': 0.00037725, 'samples': 773120, 'steps': 1509, 'loss/train': 3.7033565044403076} 02/24/2022 03:32:33 - INFO - codeparrot_training - Step 1510: {'lr': 0.0003775, 'samples': 773632, 'steps': 1510, 'loss/train': 4.7861433029174805} 02/24/2022 03:32:36 - INFO - codeparrot_training - Step 1511: {'lr': 0.00037775, 'samples': 774144, 'steps': 1511, 'loss/train': 4.188358306884766} 02/24/2022 03:32:43 - INFO - codeparrot_training - Step 1512: {'lr': 0.000378, 'samples': 774656, 'steps': 1512, 'loss/train': 3.815391778945923} 02/24/2022 03:32:47 - INFO - codeparrot_training - Step 1513: {'lr': 0.00037825, 'samples': 775168, 'steps': 1513, 'loss/train': 4.046994209289551} 02/24/2022 03:32:52 - INFO - codeparrot_training - Step 1514: {'lr': 0.0003785, 'samples': 775680, 'steps': 1514, 'loss/train': 2.8600573539733887} 02/24/2022 03:32:55 - INFO - codeparrot_training - Step 1515: {'lr': 0.00037874999999999996, 'samples': 776192, 'steps': 1515, 'loss/train': 4.728860855102539} 02/24/2022 03:33:01 - INFO - codeparrot_training - Step 1516: {'lr': 0.000379, 'samples': 776704, 'steps': 1516, 'loss/train': 4.52414083480835} 02/24/2022 03:33:04 - INFO - codeparrot_training - Step 1517: {'lr': 0.00037925, 'samples': 777216, 'steps': 1517, 'loss/train': 3.1684107780456543} 02/24/2022 03:33:10 - INFO - codeparrot_training - Step 1518: {'lr': 0.0003795, 'samples': 777728, 'steps': 1518, 'loss/train': 4.330784320831299} 02/24/2022 03:33:13 - INFO - codeparrot_training - Step 1519: {'lr': 0.00037975, 'samples': 778240, 'steps': 1519, 'loss/train': 4.734991550445557} 02/24/2022 03:33:19 - INFO - codeparrot_training - Step 1520: {'lr': 0.00038, 'samples': 778752, 'steps': 1520, 'loss/train': 4.218307971954346} 02/24/2022 03:33:22 - INFO - codeparrot_training - Step 1521: {'lr': 0.00038025, 'samples': 779264, 'steps': 1521, 'loss/train': 4.723255634307861} 02/24/2022 03:33:28 - INFO - codeparrot_training - Step 1522: {'lr': 0.00038050000000000003, 'samples': 779776, 'steps': 1522, 'loss/train': 5.335879802703857} 02/24/2022 03:33:31 - INFO - codeparrot_training - Step 1523: {'lr': 0.00038075, 'samples': 780288, 'steps': 1523, 'loss/train': 4.171871662139893} 02/24/2022 03:33:37 - INFO - codeparrot_training - Step 1524: {'lr': 0.000381, 'samples': 780800, 'steps': 1524, 'loss/train': 4.38768196105957} 02/24/2022 03:33:40 - INFO - codeparrot_training - Step 1525: {'lr': 0.00038124999999999997, 'samples': 781312, 'steps': 1525, 'loss/train': 3.8277640342712402} 02/24/2022 03:33:46 - INFO - codeparrot_training - Step 1526: {'lr': 0.0003815, 'samples': 781824, 'steps': 1526, 'loss/train': 4.827854633331299} 02/24/2022 03:33:50 - INFO - codeparrot_training - Step 1527: {'lr': 0.00038175, 'samples': 782336, 'steps': 1527, 'loss/train': 4.390958309173584} 02/24/2022 03:33:55 - INFO - codeparrot_training - Step 1528: {'lr': 0.000382, 'samples': 782848, 'steps': 1528, 'loss/train': 3.4712443351745605} 02/24/2022 03:33:59 - INFO - codeparrot_training - Step 1529: {'lr': 0.00038225, 'samples': 783360, 'steps': 1529, 'loss/train': 4.485414028167725} 02/24/2022 03:34:04 - INFO - codeparrot_training - Step 1530: {'lr': 0.00038250000000000003, 'samples': 783872, 'steps': 1530, 'loss/train': 4.224032402038574} 02/24/2022 03:34:08 - INFO - codeparrot_training - Step 1531: {'lr': 0.00038275, 'samples': 784384, 'steps': 1531, 'loss/train': 4.25079345703125} 02/24/2022 03:34:13 - INFO - codeparrot_training - Step 1532: {'lr': 0.00038300000000000004, 'samples': 784896, 'steps': 1532, 'loss/train': 3.665842056274414} 02/24/2022 03:34:17 - INFO - codeparrot_training - Step 1533: {'lr': 0.00038324999999999996, 'samples': 785408, 'steps': 1533, 'loss/train': 4.240571975708008} 02/24/2022 03:34:22 - INFO - codeparrot_training - Step 1534: {'lr': 0.0003835, 'samples': 785920, 'steps': 1534, 'loss/train': 3.8314082622528076} 02/24/2022 03:34:26 - INFO - codeparrot_training - Step 1535: {'lr': 0.00038375, 'samples': 786432, 'steps': 1535, 'loss/train': 3.0607237815856934} 02/24/2022 03:34:32 - INFO - codeparrot_training - Step 1536: {'lr': 0.000384, 'samples': 786944, 'steps': 1536, 'loss/train': 1.1733241081237793} 02/24/2022 03:34:35 - INFO - codeparrot_training - Step 1537: {'lr': 0.00038425, 'samples': 787456, 'steps': 1537, 'loss/train': 3.5049164295196533} 02/24/2022 03:34:41 - INFO - codeparrot_training - Step 1538: {'lr': 0.0003845, 'samples': 787968, 'steps': 1538, 'loss/train': 5.128952980041504} 02/24/2022 03:34:44 - INFO - codeparrot_training - Step 1539: {'lr': 0.00038475, 'samples': 788480, 'steps': 1539, 'loss/train': 3.32218599319458} 02/24/2022 03:34:50 - INFO - codeparrot_training - Step 1540: {'lr': 0.00038500000000000003, 'samples': 788992, 'steps': 1540, 'loss/train': 4.7126569747924805} 02/24/2022 03:34:53 - INFO - codeparrot_training - Step 1541: {'lr': 0.00038525, 'samples': 789504, 'steps': 1541, 'loss/train': 5.071201324462891} 02/24/2022 03:34:58 - INFO - codeparrot_training - Step 1542: {'lr': 0.0003855, 'samples': 790016, 'steps': 1542, 'loss/train': 4.748036861419678} 02/24/2022 03:35:02 - INFO - codeparrot_training - Step 1543: {'lr': 0.00038574999999999997, 'samples': 790528, 'steps': 1543, 'loss/train': 3.630767345428467} 02/24/2022 03:35:07 - INFO - codeparrot_training - Step 1544: {'lr': 0.000386, 'samples': 791040, 'steps': 1544, 'loss/train': 3.1564924716949463} 02/24/2022 03:35:11 - INFO - codeparrot_training - Step 1545: {'lr': 0.00038625, 'samples': 791552, 'steps': 1545, 'loss/train': 2.396965980529785} 02/24/2022 03:35:17 - INFO - codeparrot_training - Step 1546: {'lr': 0.0003865, 'samples': 792064, 'steps': 1546, 'loss/train': 3.0344560146331787} 02/24/2022 03:35:20 - INFO - codeparrot_training - Step 1547: {'lr': 0.00038675, 'samples': 792576, 'steps': 1547, 'loss/train': 4.657129764556885} 02/24/2022 03:35:26 - INFO - codeparrot_training - Step 1548: {'lr': 0.00038700000000000003, 'samples': 793088, 'steps': 1548, 'loss/train': 3.840728282928467} 02/24/2022 03:35:29 - INFO - codeparrot_training - Step 1549: {'lr': 0.00038725, 'samples': 793600, 'steps': 1549, 'loss/train': 5.157027721405029} 02/24/2022 03:35:35 - INFO - codeparrot_training - Step 1550: {'lr': 0.00038750000000000004, 'samples': 794112, 'steps': 1550, 'loss/train': 4.252553939819336} 02/24/2022 03:35:38 - INFO - codeparrot_training - Step 1551: {'lr': 0.00038774999999999997, 'samples': 794624, 'steps': 1551, 'loss/train': 3.6083273887634277} 02/24/2022 03:35:44 - INFO - codeparrot_training - Step 1552: {'lr': 0.000388, 'samples': 795136, 'steps': 1552, 'loss/train': 4.28788423538208} 02/24/2022 03:35:47 - INFO - codeparrot_training - Step 1553: {'lr': 0.00038825, 'samples': 795648, 'steps': 1553, 'loss/train': 5.174439430236816} 02/24/2022 03:35:53 - INFO - codeparrot_training - Step 1554: {'lr': 0.0003885, 'samples': 796160, 'steps': 1554, 'loss/train': 3.887331485748291} 02/24/2022 03:35:56 - INFO - codeparrot_training - Step 1555: {'lr': 0.00038875, 'samples': 796672, 'steps': 1555, 'loss/train': 3.7367477416992188} 02/24/2022 03:36:02 - INFO - codeparrot_training - Step 1556: {'lr': 0.000389, 'samples': 797184, 'steps': 1556, 'loss/train': 4.111754417419434} 02/24/2022 03:36:05 - INFO - codeparrot_training - Step 1557: {'lr': 0.00038925, 'samples': 797696, 'steps': 1557, 'loss/train': 3.9184670448303223} 02/24/2022 03:36:11 - INFO - codeparrot_training - Step 1558: {'lr': 0.00038950000000000003, 'samples': 798208, 'steps': 1558, 'loss/train': 3.515531063079834} 02/24/2022 03:36:15 - INFO - codeparrot_training - Step 1559: {'lr': 0.00038975, 'samples': 798720, 'steps': 1559, 'loss/train': 5.446542263031006} 02/24/2022 03:36:20 - INFO - codeparrot_training - Step 1560: {'lr': 0.00039000000000000005, 'samples': 799232, 'steps': 1560, 'loss/train': 3.9595837593078613} 02/24/2022 03:36:24 - INFO - codeparrot_training - Step 1561: {'lr': 0.00039024999999999997, 'samples': 799744, 'steps': 1561, 'loss/train': 3.99322509765625} 02/24/2022 03:36:29 - INFO - codeparrot_training - Step 1562: {'lr': 0.0003905, 'samples': 800256, 'steps': 1562, 'loss/train': 4.53653621673584} 02/24/2022 03:36:33 - INFO - codeparrot_training - Step 1563: {'lr': 0.00039075, 'samples': 800768, 'steps': 1563, 'loss/train': 4.1567606925964355} 02/24/2022 03:36:39 - INFO - codeparrot_training - Step 1564: {'lr': 0.000391, 'samples': 801280, 'steps': 1564, 'loss/train': 3.402520179748535} 02/24/2022 03:36:42 - INFO - codeparrot_training - Step 1565: {'lr': 0.00039125, 'samples': 801792, 'steps': 1565, 'loss/train': 3.962956428527832} 02/24/2022 03:36:48 - INFO - codeparrot_training - Step 1566: {'lr': 0.00039150000000000003, 'samples': 802304, 'steps': 1566, 'loss/train': 3.229032039642334} 02/24/2022 03:36:51 - INFO - codeparrot_training - Step 1567: {'lr': 0.00039175, 'samples': 802816, 'steps': 1567, 'loss/train': 5.345529556274414} 02/24/2022 03:36:57 - INFO - codeparrot_training - Step 1568: {'lr': 0.00039200000000000004, 'samples': 803328, 'steps': 1568, 'loss/train': 3.7535061836242676} 02/24/2022 03:37:00 - INFO - codeparrot_training - Step 1569: {'lr': 0.00039225, 'samples': 803840, 'steps': 1569, 'loss/train': 4.086646556854248} 02/24/2022 03:37:05 - INFO - codeparrot_training - Step 1570: {'lr': 0.0003925, 'samples': 804352, 'steps': 1570, 'loss/train': 3.606579542160034} 02/24/2022 03:37:09 - INFO - codeparrot_training - Step 1571: {'lr': 0.00039275, 'samples': 804864, 'steps': 1571, 'loss/train': 4.3846211433410645} 02/24/2022 03:37:15 - INFO - codeparrot_training - Step 1572: {'lr': 0.000393, 'samples': 805376, 'steps': 1572, 'loss/train': 5.237354755401611} 02/24/2022 03:37:19 - INFO - codeparrot_training - Step 1573: {'lr': 0.00039325, 'samples': 805888, 'steps': 1573, 'loss/train': 4.324764728546143} 02/24/2022 03:37:24 - INFO - codeparrot_training - Step 1574: {'lr': 0.0003935, 'samples': 806400, 'steps': 1574, 'loss/train': 3.2067623138427734} 02/24/2022 03:37:27 - INFO - codeparrot_training - Step 1575: {'lr': 0.00039375, 'samples': 806912, 'steps': 1575, 'loss/train': 4.206762313842773} 02/24/2022 03:37:33 - INFO - codeparrot_training - Step 1576: {'lr': 0.00039400000000000004, 'samples': 807424, 'steps': 1576, 'loss/train': 5.17689323425293} 02/24/2022 03:37:36 - INFO - codeparrot_training - Step 1577: {'lr': 0.00039425, 'samples': 807936, 'steps': 1577, 'loss/train': 3.9292101860046387} 02/24/2022 03:37:42 - INFO - codeparrot_training - Step 1578: {'lr': 0.00039450000000000005, 'samples': 808448, 'steps': 1578, 'loss/train': 3.6446094512939453} 02/24/2022 03:37:45 - INFO - codeparrot_training - Step 1579: {'lr': 0.00039474999999999997, 'samples': 808960, 'steps': 1579, 'loss/train': 5.564534664154053} 02/24/2022 03:37:51 - INFO - codeparrot_training - Step 1580: {'lr': 0.000395, 'samples': 809472, 'steps': 1580, 'loss/train': 4.369887351989746} 02/24/2022 03:37:54 - INFO - codeparrot_training - Step 1581: {'lr': 0.00039525, 'samples': 809984, 'steps': 1581, 'loss/train': 3.6876540184020996} 02/24/2022 03:38:01 - INFO - codeparrot_training - Step 1582: {'lr': 0.0003955, 'samples': 810496, 'steps': 1582, 'loss/train': 4.171565055847168} 02/24/2022 03:38:04 - INFO - codeparrot_training - Step 1583: {'lr': 0.00039575, 'samples': 811008, 'steps': 1583, 'loss/train': 4.222428321838379} 02/24/2022 03:38:10 - INFO - codeparrot_training - Step 1584: {'lr': 0.00039600000000000003, 'samples': 811520, 'steps': 1584, 'loss/train': 3.810997247695923} 02/24/2022 03:38:13 - INFO - codeparrot_training - Step 1585: {'lr': 0.00039625, 'samples': 812032, 'steps': 1585, 'loss/train': 4.887095928192139} 02/24/2022 03:38:19 - INFO - codeparrot_training - Step 1586: {'lr': 0.00039650000000000004, 'samples': 812544, 'steps': 1586, 'loss/train': 3.7610602378845215} 02/24/2022 03:38:22 - INFO - codeparrot_training - Step 1587: {'lr': 0.00039675, 'samples': 813056, 'steps': 1587, 'loss/train': 4.862823963165283} 02/24/2022 03:38:28 - INFO - codeparrot_training - Step 1588: {'lr': 0.00039700000000000005, 'samples': 813568, 'steps': 1588, 'loss/train': 6.103458881378174} 02/24/2022 03:38:31 - INFO - codeparrot_training - Step 1589: {'lr': 0.00039725, 'samples': 814080, 'steps': 1589, 'loss/train': 5.141780376434326} 02/24/2022 03:38:37 - INFO - codeparrot_training - Step 1590: {'lr': 0.0003975, 'samples': 814592, 'steps': 1590, 'loss/train': 4.060984134674072} 02/24/2022 03:38:40 - INFO - codeparrot_training - Step 1591: {'lr': 0.00039775, 'samples': 815104, 'steps': 1591, 'loss/train': 3.5712521076202393} 02/24/2022 03:38:46 - INFO - codeparrot_training - Step 1592: {'lr': 0.000398, 'samples': 815616, 'steps': 1592, 'loss/train': 4.213394641876221} 02/24/2022 03:38:49 - INFO - codeparrot_training - Step 1593: {'lr': 0.00039825, 'samples': 816128, 'steps': 1593, 'loss/train': 4.766323566436768} 02/24/2022 03:38:55 - INFO - codeparrot_training - Step 1594: {'lr': 0.00039850000000000004, 'samples': 816640, 'steps': 1594, 'loss/train': 3.537151336669922} 02/24/2022 03:38:59 - INFO - codeparrot_training - Step 1595: {'lr': 0.00039875, 'samples': 817152, 'steps': 1595, 'loss/train': 2.797670841217041} 02/24/2022 03:39:06 - INFO - codeparrot_training - Step 1596: {'lr': 0.00039900000000000005, 'samples': 817664, 'steps': 1596, 'loss/train': 4.90841007232666} 02/24/2022 03:39:09 - INFO - codeparrot_training - Step 1597: {'lr': 0.00039925000000000003, 'samples': 818176, 'steps': 1597, 'loss/train': 3.814776659011841} 02/24/2022 03:39:14 - INFO - codeparrot_training - Step 1598: {'lr': 0.0003995, 'samples': 818688, 'steps': 1598, 'loss/train': 4.990166664123535} 02/24/2022 03:39:18 - INFO - codeparrot_training - Step 1599: {'lr': 0.00039975, 'samples': 819200, 'steps': 1599, 'loss/train': 3.483522415161133} 02/24/2022 03:39:23 - INFO - codeparrot_training - Step 1600: {'lr': 0.0004, 'samples': 819712, 'steps': 1600, 'loss/train': 4.816373348236084} 02/24/2022 03:39:27 - INFO - codeparrot_training - Step 1601: {'lr': 0.00040025, 'samples': 820224, 'steps': 1601, 'loss/train': 2.801182270050049} 02/24/2022 03:39:32 - INFO - codeparrot_training - Step 1602: {'lr': 0.00040050000000000003, 'samples': 820736, 'steps': 1602, 'loss/train': 3.923861265182495} 02/24/2022 03:39:36 - INFO - codeparrot_training - Step 1603: {'lr': 0.00040075, 'samples': 821248, 'steps': 1603, 'loss/train': 3.9733026027679443} 02/24/2022 03:39:41 - INFO - codeparrot_training - Step 1604: {'lr': 0.00040100000000000004, 'samples': 821760, 'steps': 1604, 'loss/train': 5.037976264953613} 02/24/2022 03:39:45 - INFO - codeparrot_training - Step 1605: {'lr': 0.00040125, 'samples': 822272, 'steps': 1605, 'loss/train': 4.74677038192749} 02/24/2022 03:39:52 - INFO - codeparrot_training - Step 1606: {'lr': 0.00040150000000000006, 'samples': 822784, 'steps': 1606, 'loss/train': 5.230562210083008} 02/24/2022 03:39:55 - INFO - codeparrot_training - Step 1607: {'lr': 0.00040175, 'samples': 823296, 'steps': 1607, 'loss/train': 5.117575168609619} 02/24/2022 03:40:01 - INFO - codeparrot_training - Step 1608: {'lr': 0.000402, 'samples': 823808, 'steps': 1608, 'loss/train': 3.9092824459075928} 02/24/2022 03:40:04 - INFO - codeparrot_training - Step 1609: {'lr': 0.00040225, 'samples': 824320, 'steps': 1609, 'loss/train': 3.3457095623016357} 02/24/2022 03:40:10 - INFO - codeparrot_training - Step 1610: {'lr': 0.0004025, 'samples': 824832, 'steps': 1610, 'loss/train': 3.0912325382232666} 02/24/2022 03:40:13 - INFO - codeparrot_training - Step 1611: {'lr': 0.00040275, 'samples': 825344, 'steps': 1611, 'loss/train': 4.7266974449157715} 02/24/2022 03:40:19 - INFO - codeparrot_training - Step 1612: {'lr': 0.00040300000000000004, 'samples': 825856, 'steps': 1612, 'loss/train': 4.315446853637695} 02/24/2022 03:40:22 - INFO - codeparrot_training - Step 1613: {'lr': 0.00040325, 'samples': 826368, 'steps': 1613, 'loss/train': 1.2668206691741943} 02/24/2022 03:40:28 - INFO - codeparrot_training - Step 1614: {'lr': 0.00040350000000000005, 'samples': 826880, 'steps': 1614, 'loss/train': 4.686374187469482} 02/24/2022 03:40:31 - INFO - codeparrot_training - Step 1615: {'lr': 0.00040375000000000003, 'samples': 827392, 'steps': 1615, 'loss/train': 6.061168670654297} 02/24/2022 03:40:37 - INFO - codeparrot_training - Step 1616: {'lr': 0.000404, 'samples': 827904, 'steps': 1616, 'loss/train': 4.728420257568359} 02/24/2022 03:40:41 - INFO - codeparrot_training - Step 1617: {'lr': 0.00040425, 'samples': 828416, 'steps': 1617, 'loss/train': 4.256476402282715} 02/24/2022 03:40:46 - INFO - codeparrot_training - Step 1618: {'lr': 0.0004045, 'samples': 828928, 'steps': 1618, 'loss/train': 4.144393444061279} 02/24/2022 03:40:49 - INFO - codeparrot_training - Step 1619: {'lr': 0.00040475, 'samples': 829440, 'steps': 1619, 'loss/train': 3.679882764816284} 02/24/2022 03:40:55 - INFO - codeparrot_training - Step 1620: {'lr': 0.00040500000000000003, 'samples': 829952, 'steps': 1620, 'loss/train': 3.6705260276794434} 02/24/2022 03:40:58 - INFO - codeparrot_training - Step 1621: {'lr': 0.00040525, 'samples': 830464, 'steps': 1621, 'loss/train': 3.643399715423584} 02/24/2022 03:41:04 - INFO - codeparrot_training - Step 1622: {'lr': 0.00040550000000000004, 'samples': 830976, 'steps': 1622, 'loss/train': 6.2327680587768555} 02/24/2022 03:41:07 - INFO - codeparrot_training - Step 1623: {'lr': 0.00040575, 'samples': 831488, 'steps': 1623, 'loss/train': 5.0092692375183105} 02/24/2022 03:41:13 - INFO - codeparrot_training - Step 1624: {'lr': 0.00040600000000000006, 'samples': 832000, 'steps': 1624, 'loss/train': 3.2840332984924316} 02/24/2022 03:41:16 - INFO - codeparrot_training - Step 1625: {'lr': 0.00040625000000000004, 'samples': 832512, 'steps': 1625, 'loss/train': 4.8649091720581055} 02/24/2022 03:41:22 - INFO - codeparrot_training - Step 1626: {'lr': 0.00040649999999999996, 'samples': 833024, 'steps': 1626, 'loss/train': 3.7539560794830322} 02/24/2022 03:41:26 - INFO - codeparrot_training - Step 1627: {'lr': 0.00040675, 'samples': 833536, 'steps': 1627, 'loss/train': 3.8577213287353516} 02/24/2022 03:41:31 - INFO - codeparrot_training - Step 1628: {'lr': 0.00040699999999999997, 'samples': 834048, 'steps': 1628, 'loss/train': 3.5600366592407227} 02/24/2022 03:41:35 - INFO - codeparrot_training - Step 1629: {'lr': 0.00040725, 'samples': 834560, 'steps': 1629, 'loss/train': 0.8801425695419312} 02/24/2022 03:41:40 - INFO - codeparrot_training - Step 1630: {'lr': 0.0004075, 'samples': 835072, 'steps': 1630, 'loss/train': 3.8342957496643066} 02/24/2022 03:41:44 - INFO - codeparrot_training - Step 1631: {'lr': 0.00040775, 'samples': 835584, 'steps': 1631, 'loss/train': 4.777003765106201} 02/24/2022 03:41:49 - INFO - codeparrot_training - Step 1632: {'lr': 0.000408, 'samples': 836096, 'steps': 1632, 'loss/train': 5.079536437988281} 02/24/2022 03:41:53 - INFO - codeparrot_training - Step 1633: {'lr': 0.00040825000000000003, 'samples': 836608, 'steps': 1633, 'loss/train': 3.115975856781006} 02/24/2022 03:41:58 - INFO - codeparrot_training - Step 1634: {'lr': 0.0004085, 'samples': 837120, 'steps': 1634, 'loss/train': 3.5675368309020996} 02/24/2022 03:42:02 - INFO - codeparrot_training - Step 1635: {'lr': 0.00040875, 'samples': 837632, 'steps': 1635, 'loss/train': 3.040477752685547} 02/24/2022 03:42:07 - INFO - codeparrot_training - Step 1636: {'lr': 0.00040899999999999997, 'samples': 838144, 'steps': 1636, 'loss/train': 4.125766754150391} 02/24/2022 03:42:11 - INFO - codeparrot_training - Step 1637: {'lr': 0.00040925, 'samples': 838656, 'steps': 1637, 'loss/train': 2.4122812747955322} 02/24/2022 03:42:17 - INFO - codeparrot_training - Step 1638: {'lr': 0.0004095, 'samples': 839168, 'steps': 1638, 'loss/train': 2.6019678115844727} 02/24/2022 03:42:20 - INFO - codeparrot_training - Step 1639: {'lr': 0.00040975, 'samples': 839680, 'steps': 1639, 'loss/train': 4.266791343688965} 02/24/2022 03:42:26 - INFO - codeparrot_training - Step 1640: {'lr': 0.00041, 'samples': 840192, 'steps': 1640, 'loss/train': 3.685781240463257} 02/24/2022 03:42:29 - INFO - codeparrot_training - Step 1641: {'lr': 0.00041025, 'samples': 840704, 'steps': 1641, 'loss/train': 4.00113582611084} 02/24/2022 03:42:35 - INFO - codeparrot_training - Step 1642: {'lr': 0.0004105, 'samples': 841216, 'steps': 1642, 'loss/train': 1.5531821250915527} 02/24/2022 03:42:39 - INFO - codeparrot_training - Step 1643: {'lr': 0.00041075000000000004, 'samples': 841728, 'steps': 1643, 'loss/train': 3.4445090293884277} 02/24/2022 03:42:44 - INFO - codeparrot_training - Step 1644: {'lr': 0.00041099999999999996, 'samples': 842240, 'steps': 1644, 'loss/train': 3.668736696243286} 02/24/2022 03:42:48 - INFO - codeparrot_training - Step 1645: {'lr': 0.00041125, 'samples': 842752, 'steps': 1645, 'loss/train': 1.6072088479995728} 02/24/2022 03:42:53 - INFO - codeparrot_training - Step 1646: {'lr': 0.0004115, 'samples': 843264, 'steps': 1646, 'loss/train': 4.0394110679626465} 02/24/2022 03:42:56 - INFO - codeparrot_training - Step 1647: {'lr': 0.00041175, 'samples': 843776, 'steps': 1647, 'loss/train': 4.2862019538879395} 02/24/2022 03:43:02 - INFO - codeparrot_training - Step 1648: {'lr': 0.000412, 'samples': 844288, 'steps': 1648, 'loss/train': 4.242203235626221} 02/24/2022 03:43:06 - INFO - codeparrot_training - Step 1649: {'lr': 0.00041225, 'samples': 844800, 'steps': 1649, 'loss/train': 3.158301830291748} 02/24/2022 03:43:11 - INFO - codeparrot_training - Step 1650: {'lr': 0.0004125, 'samples': 845312, 'steps': 1650, 'loss/train': 4.465639114379883} 02/24/2022 03:43:15 - INFO - codeparrot_training - Step 1651: {'lr': 0.00041275000000000003, 'samples': 845824, 'steps': 1651, 'loss/train': 3.8523125648498535} 02/24/2022 03:43:20 - INFO - codeparrot_training - Step 1652: {'lr': 0.000413, 'samples': 846336, 'steps': 1652, 'loss/train': 4.363038539886475} 02/24/2022 03:43:24 - INFO - codeparrot_training - Step 1653: {'lr': 0.00041325, 'samples': 846848, 'steps': 1653, 'loss/train': 5.091957092285156} 02/24/2022 03:43:30 - INFO - codeparrot_training - Step 1654: {'lr': 0.00041349999999999997, 'samples': 847360, 'steps': 1654, 'loss/train': 4.1330485343933105} 02/24/2022 03:43:33 - INFO - codeparrot_training - Step 1655: {'lr': 0.00041375, 'samples': 847872, 'steps': 1655, 'loss/train': 4.860011577606201} 02/24/2022 03:43:39 - INFO - codeparrot_training - Step 1656: {'lr': 0.000414, 'samples': 848384, 'steps': 1656, 'loss/train': 3.5825374126434326} 02/24/2022 03:43:42 - INFO - codeparrot_training - Step 1657: {'lr': 0.00041425, 'samples': 848896, 'steps': 1657, 'loss/train': 3.5022332668304443} 02/24/2022 03:43:48 - INFO - codeparrot_training - Step 1658: {'lr': 0.0004145, 'samples': 849408, 'steps': 1658, 'loss/train': 3.397233486175537} 02/24/2022 03:43:51 - INFO - codeparrot_training - Step 1659: {'lr': 0.00041475, 'samples': 849920, 'steps': 1659, 'loss/train': 4.770598888397217} 02/24/2022 03:43:56 - INFO - codeparrot_training - Step 1660: {'lr': 0.000415, 'samples': 850432, 'steps': 1660, 'loss/train': 3.4581315517425537} 02/24/2022 03:44:00 - INFO - codeparrot_training - Step 1661: {'lr': 0.00041525000000000004, 'samples': 850944, 'steps': 1661, 'loss/train': 3.625771999359131} 02/24/2022 03:44:05 - INFO - codeparrot_training - Step 1662: {'lr': 0.00041549999999999996, 'samples': 851456, 'steps': 1662, 'loss/train': 3.6055209636688232} 02/24/2022 03:44:09 - INFO - codeparrot_training - Step 1663: {'lr': 0.00041575, 'samples': 851968, 'steps': 1663, 'loss/train': 3.7835278511047363} 02/24/2022 03:44:15 - INFO - codeparrot_training - Step 1664: {'lr': 0.000416, 'samples': 852480, 'steps': 1664, 'loss/train': 4.466989994049072} 02/24/2022 03:44:19 - INFO - codeparrot_training - Step 1665: {'lr': 0.00041625, 'samples': 852992, 'steps': 1665, 'loss/train': 4.221617698669434} 02/24/2022 03:44:24 - INFO - codeparrot_training - Step 1666: {'lr': 0.0004165, 'samples': 853504, 'steps': 1666, 'loss/train': 4.161709308624268} 02/24/2022 03:44:27 - INFO - codeparrot_training - Step 1667: {'lr': 0.00041675, 'samples': 854016, 'steps': 1667, 'loss/train': 3.2856905460357666} 02/24/2022 03:44:33 - INFO - codeparrot_training - Step 1668: {'lr': 0.000417, 'samples': 854528, 'steps': 1668, 'loss/train': 3.8417611122131348} 02/24/2022 03:44:36 - INFO - codeparrot_training - Step 1669: {'lr': 0.00041725000000000003, 'samples': 855040, 'steps': 1669, 'loss/train': 2.292628288269043} 02/24/2022 03:44:42 - INFO - codeparrot_training - Step 1670: {'lr': 0.0004175, 'samples': 855552, 'steps': 1670, 'loss/train': 3.4556281566619873} 02/24/2022 03:44:45 - INFO - codeparrot_training - Step 1671: {'lr': 0.00041775000000000004, 'samples': 856064, 'steps': 1671, 'loss/train': 4.098479747772217} 02/24/2022 03:44:51 - INFO - codeparrot_training - Step 1672: {'lr': 0.00041799999999999997, 'samples': 856576, 'steps': 1672, 'loss/train': 3.883469820022583} 02/24/2022 03:44:54 - INFO - codeparrot_training - Step 1673: {'lr': 0.00041825, 'samples': 857088, 'steps': 1673, 'loss/train': 3.452237367630005} 02/24/2022 03:45:00 - INFO - codeparrot_training - Step 1674: {'lr': 0.0004185, 'samples': 857600, 'steps': 1674, 'loss/train': 4.276821136474609} 02/24/2022 03:45:04 - INFO - codeparrot_training - Step 1675: {'lr': 0.00041875, 'samples': 858112, 'steps': 1675, 'loss/train': 3.78054141998291} 02/24/2022 03:45:09 - INFO - codeparrot_training - Step 1676: {'lr': 0.000419, 'samples': 858624, 'steps': 1676, 'loss/train': 3.6571803092956543} 02/24/2022 03:45:13 - INFO - codeparrot_training - Step 1677: {'lr': 0.00041925, 'samples': 859136, 'steps': 1677, 'loss/train': 0.9283031821250916} 02/24/2022 03:45:18 - INFO - codeparrot_training - Step 1678: {'lr': 0.0004195, 'samples': 859648, 'steps': 1678, 'loss/train': 4.2189040184021} 02/24/2022 03:45:22 - INFO - codeparrot_training - Step 1679: {'lr': 0.00041975000000000004, 'samples': 860160, 'steps': 1679, 'loss/train': 3.447009563446045} 02/24/2022 03:45:27 - INFO - codeparrot_training - Step 1680: {'lr': 0.00042, 'samples': 860672, 'steps': 1680, 'loss/train': 3.345370054244995} 02/24/2022 03:45:31 - INFO - codeparrot_training - Step 1681: {'lr': 0.00042025, 'samples': 861184, 'steps': 1681, 'loss/train': 4.379073619842529} 02/24/2022 03:45:36 - INFO - codeparrot_training - Step 1682: {'lr': 0.0004205, 'samples': 861696, 'steps': 1682, 'loss/train': 3.2736711502075195} 02/24/2022 03:45:42 - INFO - codeparrot_training - Step 1683: {'lr': 0.00042075, 'samples': 862208, 'steps': 1683, 'loss/train': 3.483628988265991} 02/24/2022 03:45:46 - INFO - codeparrot_training - Step 1684: {'lr': 0.000421, 'samples': 862720, 'steps': 1684, 'loss/train': 3.41766357421875} 02/24/2022 03:45:51 - INFO - codeparrot_training - Step 1685: {'lr': 0.00042125, 'samples': 863232, 'steps': 1685, 'loss/train': 2.9679110050201416} 02/24/2022 03:45:55 - INFO - codeparrot_training - Step 1686: {'lr': 0.0004215, 'samples': 863744, 'steps': 1686, 'loss/train': 5.177552700042725} 02/24/2022 03:46:00 - INFO - codeparrot_training - Step 1687: {'lr': 0.00042175000000000003, 'samples': 864256, 'steps': 1687, 'loss/train': 3.167175531387329} 02/24/2022 03:46:04 - INFO - codeparrot_training - Step 1688: {'lr': 0.000422, 'samples': 864768, 'steps': 1688, 'loss/train': 4.140285491943359} 02/24/2022 03:46:09 - INFO - codeparrot_training - Step 1689: {'lr': 0.00042225000000000005, 'samples': 865280, 'steps': 1689, 'loss/train': 4.430770397186279} 02/24/2022 03:46:12 - INFO - codeparrot_training - Step 1690: {'lr': 0.00042249999999999997, 'samples': 865792, 'steps': 1690, 'loss/train': 4.592250823974609} 02/24/2022 03:46:18 - INFO - codeparrot_training - Step 1691: {'lr': 0.00042275, 'samples': 866304, 'steps': 1691, 'loss/train': 3.8042314052581787} 02/24/2022 03:46:21 - INFO - codeparrot_training - Step 1692: {'lr': 0.000423, 'samples': 866816, 'steps': 1692, 'loss/train': 4.788865089416504} 02/24/2022 03:46:27 - INFO - codeparrot_training - Step 1693: {'lr': 0.00042325, 'samples': 867328, 'steps': 1693, 'loss/train': 4.038976192474365} 02/24/2022 03:46:30 - INFO - codeparrot_training - Step 1694: {'lr': 0.0004235, 'samples': 867840, 'steps': 1694, 'loss/train': 4.357811450958252} 02/24/2022 03:46:36 - INFO - codeparrot_training - Step 1695: {'lr': 0.00042375000000000003, 'samples': 868352, 'steps': 1695, 'loss/train': 3.128354072570801} 02/24/2022 03:46:39 - INFO - codeparrot_training - Step 1696: {'lr': 0.000424, 'samples': 868864, 'steps': 1696, 'loss/train': 4.368496417999268} 02/24/2022 03:46:45 - INFO - codeparrot_training - Step 1697: {'lr': 0.00042425000000000004, 'samples': 869376, 'steps': 1697, 'loss/train': 3.5256783962249756} 02/24/2022 03:46:48 - INFO - codeparrot_training - Step 1698: {'lr': 0.0004245, 'samples': 869888, 'steps': 1698, 'loss/train': 4.285900592803955} 02/24/2022 03:46:54 - INFO - codeparrot_training - Step 1699: {'lr': 0.00042475000000000005, 'samples': 870400, 'steps': 1699, 'loss/train': 3.7480287551879883} 02/24/2022 03:46:58 - INFO - codeparrot_training - Step 1700: {'lr': 0.000425, 'samples': 870912, 'steps': 1700, 'loss/train': 2.899427890777588} 02/24/2022 03:47:03 - INFO - codeparrot_training - Step 1701: {'lr': 0.00042525, 'samples': 871424, 'steps': 1701, 'loss/train': 3.787052869796753} 02/24/2022 03:47:07 - INFO - codeparrot_training - Step 1702: {'lr': 0.0004255, 'samples': 871936, 'steps': 1702, 'loss/train': 2.8308157920837402} 02/24/2022 03:47:12 - INFO - codeparrot_training - Step 1703: {'lr': 0.00042575, 'samples': 872448, 'steps': 1703, 'loss/train': 3.9945175647735596} 02/24/2022 03:47:16 - INFO - codeparrot_training - Step 1704: {'lr': 0.000426, 'samples': 872960, 'steps': 1704, 'loss/train': 4.1553425788879395} 02/24/2022 03:47:21 - INFO - codeparrot_training - Step 1705: {'lr': 0.00042625000000000003, 'samples': 873472, 'steps': 1705, 'loss/train': 3.1137735843658447} 02/24/2022 03:47:25 - INFO - codeparrot_training - Step 1706: {'lr': 0.0004265, 'samples': 873984, 'steps': 1706, 'loss/train': 3.8155298233032227} 02/24/2022 03:47:31 - INFO - codeparrot_training - Step 1707: {'lr': 0.00042675000000000005, 'samples': 874496, 'steps': 1707, 'loss/train': 4.7053632736206055} 02/24/2022 03:47:34 - INFO - codeparrot_training - Step 1708: {'lr': 0.000427, 'samples': 875008, 'steps': 1708, 'loss/train': 4.304908752441406} 02/24/2022 03:47:40 - INFO - codeparrot_training - Step 1709: {'lr': 0.00042725, 'samples': 875520, 'steps': 1709, 'loss/train': 1.082191824913025} 02/24/2022 03:47:43 - INFO - codeparrot_training - Step 1710: {'lr': 0.0004275, 'samples': 876032, 'steps': 1710, 'loss/train': 4.41619873046875} 02/24/2022 03:47:49 - INFO - codeparrot_training - Step 1711: {'lr': 0.00042775, 'samples': 876544, 'steps': 1711, 'loss/train': 2.172891855239868} 02/24/2022 03:47:52 - INFO - codeparrot_training - Step 1712: {'lr': 0.000428, 'samples': 877056, 'steps': 1712, 'loss/train': 4.226787567138672} 02/24/2022 03:47:58 - INFO - codeparrot_training - Step 1713: {'lr': 0.00042825000000000003, 'samples': 877568, 'steps': 1713, 'loss/train': 3.967400074005127} 02/24/2022 03:48:01 - INFO - codeparrot_training - Step 1714: {'lr': 0.0004285, 'samples': 878080, 'steps': 1714, 'loss/train': 4.285831928253174} 02/24/2022 03:48:07 - INFO - codeparrot_training - Step 1715: {'lr': 0.00042875000000000004, 'samples': 878592, 'steps': 1715, 'loss/train': 4.683849334716797} 02/24/2022 03:48:10 - INFO - codeparrot_training - Step 1716: {'lr': 0.000429, 'samples': 879104, 'steps': 1716, 'loss/train': 3.3186259269714355} 02/24/2022 03:48:16 - INFO - codeparrot_training - Step 1717: {'lr': 0.00042925000000000005, 'samples': 879616, 'steps': 1717, 'loss/train': 4.041749954223633} 02/24/2022 03:48:19 - INFO - codeparrot_training - Step 1718: {'lr': 0.0004295, 'samples': 880128, 'steps': 1718, 'loss/train': 3.5565340518951416} 02/24/2022 03:48:26 - INFO - codeparrot_training - Step 1719: {'lr': 0.00042975, 'samples': 880640, 'steps': 1719, 'loss/train': 3.8268656730651855} 02/24/2022 03:48:30 - INFO - codeparrot_training - Step 1720: {'lr': 0.00043, 'samples': 881152, 'steps': 1720, 'loss/train': 1.9114283323287964} 02/24/2022 03:48:35 - INFO - codeparrot_training - Step 1721: {'lr': 0.00043025, 'samples': 881664, 'steps': 1721, 'loss/train': 4.044600009918213} 02/24/2022 03:48:38 - INFO - codeparrot_training - Step 1722: {'lr': 0.0004305, 'samples': 882176, 'steps': 1722, 'loss/train': 3.9941177368164062} 02/24/2022 03:48:44 - INFO - codeparrot_training - Step 1723: {'lr': 0.00043075000000000003, 'samples': 882688, 'steps': 1723, 'loss/train': 3.646448850631714} 02/24/2022 03:48:47 - INFO - codeparrot_training - Step 1724: {'lr': 0.000431, 'samples': 883200, 'steps': 1724, 'loss/train': 3.9820351600646973} 02/24/2022 03:48:53 - INFO - codeparrot_training - Step 1725: {'lr': 0.00043125000000000005, 'samples': 883712, 'steps': 1725, 'loss/train': 3.4107930660247803} 02/24/2022 03:48:56 - INFO - codeparrot_training - Step 1726: {'lr': 0.0004315, 'samples': 884224, 'steps': 1726, 'loss/train': 4.530364513397217} 02/24/2022 03:49:02 - INFO - codeparrot_training - Step 1727: {'lr': 0.00043175, 'samples': 884736, 'steps': 1727, 'loss/train': 3.9251275062561035} 02/24/2022 03:49:05 - INFO - codeparrot_training - Step 1728: {'lr': 0.000432, 'samples': 885248, 'steps': 1728, 'loss/train': 4.252110004425049} 02/24/2022 03:49:12 - INFO - codeparrot_training - Step 1729: {'lr': 0.00043225, 'samples': 885760, 'steps': 1729, 'loss/train': 2.7589948177337646} 02/24/2022 03:49:15 - INFO - codeparrot_training - Step 1730: {'lr': 0.0004325, 'samples': 886272, 'steps': 1730, 'loss/train': 3.8178868293762207} 02/24/2022 03:49:21 - INFO - codeparrot_training - Step 1731: {'lr': 0.00043275000000000003, 'samples': 886784, 'steps': 1731, 'loss/train': 3.0542633533477783} 02/24/2022 03:49:24 - INFO - codeparrot_training - Step 1732: {'lr': 0.000433, 'samples': 887296, 'steps': 1732, 'loss/train': 3.4022269248962402} 02/24/2022 03:49:30 - INFO - codeparrot_training - Step 1733: {'lr': 0.00043325000000000004, 'samples': 887808, 'steps': 1733, 'loss/train': 6.1047844886779785} 02/24/2022 03:49:33 - INFO - codeparrot_training - Step 1734: {'lr': 0.0004335, 'samples': 888320, 'steps': 1734, 'loss/train': 3.8315672874450684} 02/24/2022 03:49:39 - INFO - codeparrot_training - Step 1735: {'lr': 0.00043375000000000005, 'samples': 888832, 'steps': 1735, 'loss/train': 4.061244487762451} 02/24/2022 03:49:42 - INFO - codeparrot_training - Step 1736: {'lr': 0.00043400000000000003, 'samples': 889344, 'steps': 1736, 'loss/train': 3.787344217300415} 02/24/2022 03:49:48 - INFO - codeparrot_training - Step 1737: {'lr': 0.00043425, 'samples': 889856, 'steps': 1737, 'loss/train': 4.335191249847412} 02/24/2022 03:49:51 - INFO - codeparrot_training - Step 1738: {'lr': 0.0004345, 'samples': 890368, 'steps': 1738, 'loss/train': 3.9164671897888184} 02/24/2022 03:49:57 - INFO - codeparrot_training - Step 1739: {'lr': 0.00043475, 'samples': 890880, 'steps': 1739, 'loss/train': 4.006381034851074} 02/24/2022 03:50:00 - INFO - codeparrot_training - Step 1740: {'lr': 0.000435, 'samples': 891392, 'steps': 1740, 'loss/train': 3.6912143230438232} 02/24/2022 03:50:06 - INFO - codeparrot_training - Step 1741: {'lr': 0.00043525000000000004, 'samples': 891904, 'steps': 1741, 'loss/train': 4.184384346008301} 02/24/2022 03:50:09 - INFO - codeparrot_training - Step 1742: {'lr': 0.0004355, 'samples': 892416, 'steps': 1742, 'loss/train': 4.126287460327148} 02/24/2022 03:50:15 - INFO - codeparrot_training - Step 1743: {'lr': 0.00043575000000000005, 'samples': 892928, 'steps': 1743, 'loss/train': 4.320517063140869} 02/24/2022 03:50:18 - INFO - codeparrot_training - Step 1744: {'lr': 0.000436, 'samples': 893440, 'steps': 1744, 'loss/train': 2.6963040828704834} 02/24/2022 03:50:25 - INFO - codeparrot_training - Step 1745: {'lr': 0.00043625000000000006, 'samples': 893952, 'steps': 1745, 'loss/train': 6.48025369644165} 02/24/2022 03:50:28 - INFO - codeparrot_training - Step 1746: {'lr': 0.0004365, 'samples': 894464, 'steps': 1746, 'loss/train': 4.3270039558410645} 02/24/2022 03:50:34 - INFO - codeparrot_training - Step 1747: {'lr': 0.00043675, 'samples': 894976, 'steps': 1747, 'loss/train': 4.204241752624512} 02/24/2022 03:50:37 - INFO - codeparrot_training - Step 1748: {'lr': 0.000437, 'samples': 895488, 'steps': 1748, 'loss/train': 5.211532115936279} 02/24/2022 03:50:43 - INFO - codeparrot_training - Step 1749: {'lr': 0.00043725000000000003, 'samples': 896000, 'steps': 1749, 'loss/train': 4.810113906860352} 02/24/2022 03:50:46 - INFO - codeparrot_training - Step 1750: {'lr': 0.0004375, 'samples': 896512, 'steps': 1750, 'loss/train': 3.8145179748535156} 02/24/2022 03:50:51 - INFO - codeparrot_training - Step 1751: {'lr': 0.00043775, 'samples': 897024, 'steps': 1751, 'loss/train': 3.480945348739624} 02/24/2022 03:50:55 - INFO - codeparrot_training - Step 1752: {'lr': 0.000438, 'samples': 897536, 'steps': 1752, 'loss/train': 4.3082194328308105} 02/24/2022 03:51:00 - INFO - codeparrot_training - Step 1753: {'lr': 0.00043825, 'samples': 898048, 'steps': 1753, 'loss/train': 3.9679627418518066} 02/24/2022 03:51:04 - INFO - codeparrot_training - Step 1754: {'lr': 0.00043850000000000003, 'samples': 898560, 'steps': 1754, 'loss/train': 3.5825259685516357} 02/24/2022 03:51:10 - INFO - codeparrot_training - Step 1755: {'lr': 0.00043874999999999996, 'samples': 899072, 'steps': 1755, 'loss/train': 2.855005979537964} 02/24/2022 03:51:13 - INFO - codeparrot_training - Step 1756: {'lr': 0.000439, 'samples': 899584, 'steps': 1756, 'loss/train': 3.516845703125} 02/24/2022 03:51:19 - INFO - codeparrot_training - Step 1757: {'lr': 0.00043924999999999997, 'samples': 900096, 'steps': 1757, 'loss/train': 4.522437572479248} 02/24/2022 03:51:22 - INFO - codeparrot_training - Step 1758: {'lr': 0.0004395, 'samples': 900608, 'steps': 1758, 'loss/train': 3.619581937789917} 02/24/2022 03:51:28 - INFO - codeparrot_training - Step 1759: {'lr': 0.00043975, 'samples': 901120, 'steps': 1759, 'loss/train': 3.721560001373291} 02/24/2022 03:51:31 - INFO - codeparrot_training - Step 1760: {'lr': 0.00044, 'samples': 901632, 'steps': 1760, 'loss/train': 4.21987771987915} 02/24/2022 03:51:37 - INFO - codeparrot_training - Step 1761: {'lr': 0.00044025, 'samples': 902144, 'steps': 1761, 'loss/train': 3.4904325008392334} 02/24/2022 03:51:40 - INFO - codeparrot_training - Step 1762: {'lr': 0.00044050000000000003, 'samples': 902656, 'steps': 1762, 'loss/train': 3.912849187850952} 02/24/2022 03:51:46 - INFO - codeparrot_training - Step 1763: {'lr': 0.00044075, 'samples': 903168, 'steps': 1763, 'loss/train': 1.1546976566314697} 02/24/2022 03:51:49 - INFO - codeparrot_training - Step 1764: {'lr': 0.000441, 'samples': 903680, 'steps': 1764, 'loss/train': 5.073992729187012} 02/24/2022 03:51:56 - INFO - codeparrot_training - Step 1765: {'lr': 0.00044124999999999996, 'samples': 904192, 'steps': 1765, 'loss/train': 3.295259475708008} 02/24/2022 03:51:59 - INFO - codeparrot_training - Step 1766: {'lr': 0.0004415, 'samples': 904704, 'steps': 1766, 'loss/train': 3.897064447402954} 02/24/2022 03:52:05 - INFO - codeparrot_training - Step 1767: {'lr': 0.00044175, 'samples': 905216, 'steps': 1767, 'loss/train': 3.720574140548706} 02/24/2022 03:52:08 - INFO - codeparrot_training - Step 1768: {'lr': 0.000442, 'samples': 905728, 'steps': 1768, 'loss/train': 2.6307895183563232} 02/24/2022 03:52:14 - INFO - codeparrot_training - Step 1769: {'lr': 0.00044225, 'samples': 906240, 'steps': 1769, 'loss/train': 3.19089412689209} 02/24/2022 03:52:17 - INFO - codeparrot_training - Step 1770: {'lr': 0.0004425, 'samples': 906752, 'steps': 1770, 'loss/train': 2.301189661026001} 02/24/2022 03:52:23 - INFO - codeparrot_training - Step 1771: {'lr': 0.00044275, 'samples': 907264, 'steps': 1771, 'loss/train': 3.2204225063323975} 02/24/2022 03:52:26 - INFO - codeparrot_training - Step 1772: {'lr': 0.00044300000000000003, 'samples': 907776, 'steps': 1772, 'loss/train': 3.2072110176086426} 02/24/2022 03:52:32 - INFO - codeparrot_training - Step 1773: {'lr': 0.00044325, 'samples': 908288, 'steps': 1773, 'loss/train': 4.46207332611084} 02/24/2022 03:52:35 - INFO - codeparrot_training - Step 1774: {'lr': 0.0004435, 'samples': 908800, 'steps': 1774, 'loss/train': 3.0252420902252197} 02/24/2022 03:52:42 - INFO - codeparrot_training - Step 1775: {'lr': 0.00044374999999999997, 'samples': 909312, 'steps': 1775, 'loss/train': 1.1468534469604492} 02/24/2022 03:52:46 - INFO - codeparrot_training - Step 1776: {'lr': 0.000444, 'samples': 909824, 'steps': 1776, 'loss/train': 3.9944698810577393} 02/24/2022 03:52:51 - INFO - codeparrot_training - Step 1777: {'lr': 0.00044425, 'samples': 910336, 'steps': 1777, 'loss/train': 6.24413537979126} 02/24/2022 03:52:55 - INFO - codeparrot_training - Step 1778: {'lr': 0.0004445, 'samples': 910848, 'steps': 1778, 'loss/train': 5.424709796905518} 02/24/2022 03:52:58 - INFO - codeparrot_training - Step 1779: {'lr': 0.00044475, 'samples': 911360, 'steps': 1779, 'loss/train': 4.296822547912598} 02/24/2022 03:53:04 - INFO - codeparrot_training - Step 1780: {'lr': 0.00044500000000000003, 'samples': 911872, 'steps': 1780, 'loss/train': 4.318593978881836} 02/24/2022 03:53:07 - INFO - codeparrot_training - Step 1781: {'lr': 0.00044525, 'samples': 912384, 'steps': 1781, 'loss/train': 4.570363521575928} 02/24/2022 03:53:13 - INFO - codeparrot_training - Step 1782: {'lr': 0.00044550000000000004, 'samples': 912896, 'steps': 1782, 'loss/train': 2.947409152984619} 02/24/2022 03:53:18 - INFO - codeparrot_training - Step 1783: {'lr': 0.00044574999999999997, 'samples': 913408, 'steps': 1783, 'loss/train': 3.573421001434326} 02/24/2022 03:53:22 - INFO - codeparrot_training - Step 1784: {'lr': 0.000446, 'samples': 913920, 'steps': 1784, 'loss/train': 4.229432582855225} 02/24/2022 03:53:27 - INFO - codeparrot_training - Step 1785: {'lr': 0.00044625, 'samples': 914432, 'steps': 1785, 'loss/train': 4.517875671386719} 02/24/2022 03:53:31 - INFO - codeparrot_training - Step 1786: {'lr': 0.0004465, 'samples': 914944, 'steps': 1786, 'loss/train': 4.342311382293701} 02/24/2022 03:53:37 - INFO - codeparrot_training - Step 1787: {'lr': 0.00044675, 'samples': 915456, 'steps': 1787, 'loss/train': 4.162026882171631} 02/24/2022 03:53:40 - INFO - codeparrot_training - Step 1788: {'lr': 0.000447, 'samples': 915968, 'steps': 1788, 'loss/train': 4.058498382568359} 02/24/2022 03:53:46 - INFO - codeparrot_training - Step 1789: {'lr': 0.00044725, 'samples': 916480, 'steps': 1789, 'loss/train': 5.35480260848999} 02/24/2022 03:53:49 - INFO - codeparrot_training - Step 1790: {'lr': 0.00044750000000000004, 'samples': 916992, 'steps': 1790, 'loss/train': 3.704540729522705} 02/24/2022 03:53:55 - INFO - codeparrot_training - Step 1791: {'lr': 0.00044775, 'samples': 917504, 'steps': 1791, 'loss/train': 3.169809341430664} 02/24/2022 03:53:58 - INFO - codeparrot_training - Step 1792: {'lr': 0.000448, 'samples': 918016, 'steps': 1792, 'loss/train': 5.216386795043945} 02/24/2022 03:54:03 - INFO - codeparrot_training - Step 1793: {'lr': 0.00044824999999999997, 'samples': 918528, 'steps': 1793, 'loss/train': 4.576674938201904} 02/24/2022 03:54:07 - INFO - codeparrot_training - Step 1794: {'lr': 0.0004485, 'samples': 919040, 'steps': 1794, 'loss/train': 4.403326511383057} 02/24/2022 03:54:12 - INFO - codeparrot_training - Step 1795: {'lr': 0.00044875, 'samples': 919552, 'steps': 1795, 'loss/train': 3.6770081520080566} 02/24/2022 03:54:16 - INFO - codeparrot_training - Step 1796: {'lr': 0.000449, 'samples': 920064, 'steps': 1796, 'loss/train': 4.592019557952881} 02/24/2022 03:54:21 - INFO - codeparrot_training - Step 1797: {'lr': 0.00044925, 'samples': 920576, 'steps': 1797, 'loss/train': 3.8015635013580322} 02/24/2022 03:54:25 - INFO - codeparrot_training - Step 1798: {'lr': 0.00044950000000000003, 'samples': 921088, 'steps': 1798, 'loss/train': 3.5735716819763184} 02/24/2022 03:54:30 - INFO - codeparrot_training - Step 1799: {'lr': 0.00044975, 'samples': 921600, 'steps': 1799, 'loss/train': 3.4813215732574463} 02/24/2022 03:54:34 - INFO - codeparrot_training - Step 1800: {'lr': 0.00045000000000000004, 'samples': 922112, 'steps': 1800, 'loss/train': 5.032011985778809} 02/24/2022 03:54:41 - INFO - codeparrot_training - Step 1801: {'lr': 0.00045024999999999997, 'samples': 922624, 'steps': 1801, 'loss/train': 4.501695156097412} 02/24/2022 03:54:44 - INFO - codeparrot_training - Step 1802: {'lr': 0.0004505, 'samples': 923136, 'steps': 1802, 'loss/train': 3.375959873199463} 02/24/2022 03:54:50 - INFO - codeparrot_training - Step 1803: {'lr': 0.00045075, 'samples': 923648, 'steps': 1803, 'loss/train': 3.2776548862457275} 02/24/2022 03:54:53 - INFO - codeparrot_training - Step 1804: {'lr': 0.000451, 'samples': 924160, 'steps': 1804, 'loss/train': 4.1599345207214355} 02/24/2022 03:54:59 - INFO - codeparrot_training - Step 1805: {'lr': 0.00045125, 'samples': 924672, 'steps': 1805, 'loss/train': 3.610893487930298} 02/24/2022 03:55:02 - INFO - codeparrot_training - Step 1806: {'lr': 0.0004515, 'samples': 925184, 'steps': 1806, 'loss/train': 4.139208793640137} 02/24/2022 03:55:08 - INFO - codeparrot_training - Step 1807: {'lr': 0.00045175, 'samples': 925696, 'steps': 1807, 'loss/train': 3.118744134902954} 02/24/2022 03:55:11 - INFO - codeparrot_training - Step 1808: {'lr': 0.00045200000000000004, 'samples': 926208, 'steps': 1808, 'loss/train': 3.631859302520752} 02/24/2022 03:55:17 - INFO - codeparrot_training - Step 1809: {'lr': 0.00045225, 'samples': 926720, 'steps': 1809, 'loss/train': 3.8989806175231934} 02/24/2022 03:55:20 - INFO - codeparrot_training - Step 1810: {'lr': 0.00045250000000000005, 'samples': 927232, 'steps': 1810, 'loss/train': 2.0652458667755127} 02/24/2022 03:55:27 - INFO - codeparrot_training - Step 1811: {'lr': 0.00045275, 'samples': 927744, 'steps': 1811, 'loss/train': 3.0327908992767334} 02/24/2022 03:55:31 - INFO - codeparrot_training - Step 1812: {'lr': 0.000453, 'samples': 928256, 'steps': 1812, 'loss/train': 4.495767593383789} 02/24/2022 03:55:36 - INFO - codeparrot_training - Step 1813: {'lr': 0.00045325, 'samples': 928768, 'steps': 1813, 'loss/train': 4.163337707519531} 02/24/2022 03:55:39 - INFO - codeparrot_training - Step 1814: {'lr': 0.0004535, 'samples': 929280, 'steps': 1814, 'loss/train': 3.809475898742676} 02/24/2022 03:55:45 - INFO - codeparrot_training - Step 1815: {'lr': 0.00045375, 'samples': 929792, 'steps': 1815, 'loss/train': 4.117321968078613} 02/24/2022 03:55:48 - INFO - codeparrot_training - Step 1816: {'lr': 0.00045400000000000003, 'samples': 930304, 'steps': 1816, 'loss/train': 3.2555124759674072} 02/24/2022 03:55:54 - INFO - codeparrot_training - Step 1817: {'lr': 0.00045425, 'samples': 930816, 'steps': 1817, 'loss/train': 3.4132704734802246} 02/24/2022 03:55:57 - INFO - codeparrot_training - Step 1818: {'lr': 0.00045450000000000004, 'samples': 931328, 'steps': 1818, 'loss/train': 3.0801494121551514} 02/24/2022 03:56:03 - INFO - codeparrot_training - Step 1819: {'lr': 0.00045475, 'samples': 931840, 'steps': 1819, 'loss/train': 3.2781450748443604} 02/24/2022 03:56:06 - INFO - codeparrot_training - Step 1820: {'lr': 0.000455, 'samples': 932352, 'steps': 1820, 'loss/train': 3.5847525596618652} 02/24/2022 03:56:12 - INFO - codeparrot_training - Step 1821: {'lr': 0.00045525, 'samples': 932864, 'steps': 1821, 'loss/train': 3.1453769207000732} 02/24/2022 03:56:16 - INFO - codeparrot_training - Step 1822: {'lr': 0.0004555, 'samples': 933376, 'steps': 1822, 'loss/train': 3.9529950618743896} 02/24/2022 03:56:21 - INFO - codeparrot_training - Step 1823: {'lr': 0.00045575, 'samples': 933888, 'steps': 1823, 'loss/train': 1.4370235204696655} 02/24/2022 03:56:25 - INFO - codeparrot_training - Step 1824: {'lr': 0.000456, 'samples': 934400, 'steps': 1824, 'loss/train': 5.172427177429199} 02/24/2022 03:56:30 - INFO - codeparrot_training - Step 1825: {'lr': 0.00045625, 'samples': 934912, 'steps': 1825, 'loss/train': 2.903763771057129} 02/24/2022 03:56:36 - INFO - codeparrot_training - Step 1826: {'lr': 0.00045650000000000004, 'samples': 935424, 'steps': 1826, 'loss/train': 4.424928188323975} 02/24/2022 03:56:39 - INFO - codeparrot_training - Step 1827: {'lr': 0.00045675, 'samples': 935936, 'steps': 1827, 'loss/train': 3.0317370891571045} 02/24/2022 03:56:45 - INFO - codeparrot_training - Step 1828: {'lr': 0.00045700000000000005, 'samples': 936448, 'steps': 1828, 'loss/train': 4.324011325836182} 02/24/2022 03:56:48 - INFO - codeparrot_training - Step 1829: {'lr': 0.00045725, 'samples': 936960, 'steps': 1829, 'loss/train': 3.033329725265503} 02/24/2022 03:56:54 - INFO - codeparrot_training - Step 1830: {'lr': 0.0004575, 'samples': 937472, 'steps': 1830, 'loss/train': 4.144485950469971} 02/24/2022 03:56:57 - INFO - codeparrot_training - Step 1831: {'lr': 0.00045775, 'samples': 937984, 'steps': 1831, 'loss/train': 4.162665843963623} 02/24/2022 03:57:03 - INFO - codeparrot_training - Step 1832: {'lr': 0.000458, 'samples': 938496, 'steps': 1832, 'loss/train': 1.6082497835159302} 02/24/2022 03:57:06 - INFO - codeparrot_training - Step 1833: {'lr': 0.00045825, 'samples': 939008, 'steps': 1833, 'loss/train': 3.437598705291748} 02/24/2022 03:57:12 - INFO - codeparrot_training - Step 1834: {'lr': 0.00045850000000000003, 'samples': 939520, 'steps': 1834, 'loss/train': 5.750239372253418} 02/24/2022 03:57:15 - INFO - codeparrot_training - Step 1835: {'lr': 0.00045875, 'samples': 940032, 'steps': 1835, 'loss/train': 4.496338367462158} 02/24/2022 03:57:20 - INFO - codeparrot_training - Step 1836: {'lr': 0.00045900000000000004, 'samples': 940544, 'steps': 1836, 'loss/train': 3.1345911026000977} 02/24/2022 03:57:24 - INFO - codeparrot_training - Step 1837: {'lr': 0.00045925, 'samples': 941056, 'steps': 1837, 'loss/train': 2.838383436203003} 02/24/2022 03:57:30 - INFO - codeparrot_training - Step 1838: {'lr': 0.00045950000000000006, 'samples': 941568, 'steps': 1838, 'loss/train': 4.904336929321289} 02/24/2022 03:57:33 - INFO - codeparrot_training - Step 1839: {'lr': 0.00045975, 'samples': 942080, 'steps': 1839, 'loss/train': 3.4146924018859863} 02/24/2022 03:57:39 - INFO - codeparrot_training - Step 1840: {'lr': 0.00046, 'samples': 942592, 'steps': 1840, 'loss/train': 4.4224748611450195} 02/24/2022 03:57:42 - INFO - codeparrot_training - Step 1841: {'lr': 0.00046025, 'samples': 943104, 'steps': 1841, 'loss/train': 2.8425915241241455} 02/24/2022 03:57:48 - INFO - codeparrot_training - Step 1842: {'lr': 0.0004605, 'samples': 943616, 'steps': 1842, 'loss/train': 4.913407802581787} 02/24/2022 03:57:52 - INFO - codeparrot_training - Step 1843: {'lr': 0.00046075, 'samples': 944128, 'steps': 1843, 'loss/train': 4.115130424499512} 02/24/2022 03:57:57 - INFO - codeparrot_training - Step 1844: {'lr': 0.00046100000000000004, 'samples': 944640, 'steps': 1844, 'loss/train': 3.745527744293213} 02/24/2022 03:58:01 - INFO - codeparrot_training - Step 1845: {'lr': 0.00046125, 'samples': 945152, 'steps': 1845, 'loss/train': 4.078271865844727} 02/24/2022 03:58:06 - INFO - codeparrot_training - Step 1846: {'lr': 0.00046150000000000005, 'samples': 945664, 'steps': 1846, 'loss/train': 4.371314525604248} 02/24/2022 03:58:09 - INFO - codeparrot_training - Step 1847: {'lr': 0.00046175000000000003, 'samples': 946176, 'steps': 1847, 'loss/train': 3.3703036308288574} 02/24/2022 03:58:16 - INFO - codeparrot_training - Step 1848: {'lr': 0.000462, 'samples': 946688, 'steps': 1848, 'loss/train': 5.860727787017822} 02/24/2022 03:58:19 - INFO - codeparrot_training - Step 1849: {'lr': 0.00046225, 'samples': 947200, 'steps': 1849, 'loss/train': 3.782621383666992} 02/24/2022 03:58:24 - INFO - codeparrot_training - Step 1850: {'lr': 0.0004625, 'samples': 947712, 'steps': 1850, 'loss/train': 3.4912502765655518} 02/24/2022 03:58:28 - INFO - codeparrot_training - Step 1851: {'lr': 0.00046275, 'samples': 948224, 'steps': 1851, 'loss/train': 3.6662824153900146} 02/24/2022 03:58:33 - INFO - codeparrot_training - Step 1852: {'lr': 0.00046300000000000003, 'samples': 948736, 'steps': 1852, 'loss/train': 3.2053329944610596} 02/24/2022 03:58:37 - INFO - codeparrot_training - Step 1853: {'lr': 0.00046325, 'samples': 949248, 'steps': 1853, 'loss/train': 3.854938268661499} 02/24/2022 03:58:42 - INFO - codeparrot_training - Step 1854: {'lr': 0.00046350000000000004, 'samples': 949760, 'steps': 1854, 'loss/train': 3.7356789112091064} 02/24/2022 03:58:46 - INFO - codeparrot_training - Step 1855: {'lr': 0.00046375, 'samples': 950272, 'steps': 1855, 'loss/train': 3.7279837131500244} 02/24/2022 03:58:51 - INFO - codeparrot_training - Step 1856: {'lr': 0.00046400000000000006, 'samples': 950784, 'steps': 1856, 'loss/train': 3.8213253021240234} 02/24/2022 03:58:55 - INFO - codeparrot_training - Step 1857: {'lr': 0.00046425, 'samples': 951296, 'steps': 1857, 'loss/train': 3.7587873935699463} 02/24/2022 03:59:01 - INFO - codeparrot_training - Step 1858: {'lr': 0.0004645, 'samples': 951808, 'steps': 1858, 'loss/train': 3.8782124519348145} 02/24/2022 03:59:05 - INFO - codeparrot_training - Step 1859: {'lr': 0.00046475, 'samples': 952320, 'steps': 1859, 'loss/train': 3.160205125808716} 02/24/2022 03:59:10 - INFO - codeparrot_training - Step 1860: {'lr': 0.000465, 'samples': 952832, 'steps': 1860, 'loss/train': 4.472071170806885} 02/24/2022 03:59:13 - INFO - codeparrot_training - Step 1861: {'lr': 0.00046525, 'samples': 953344, 'steps': 1861, 'loss/train': 4.162772178649902} 02/24/2022 03:59:19 - INFO - codeparrot_training - Step 1862: {'lr': 0.00046550000000000004, 'samples': 953856, 'steps': 1862, 'loss/train': 4.5032525062561035} 02/24/2022 03:59:23 - INFO - codeparrot_training - Step 1863: {'lr': 0.00046575, 'samples': 954368, 'steps': 1863, 'loss/train': 3.5368921756744385} 02/24/2022 03:59:28 - INFO - codeparrot_training - Step 1864: {'lr': 0.00046600000000000005, 'samples': 954880, 'steps': 1864, 'loss/train': 3.049611806869507} 02/24/2022 03:59:31 - INFO - codeparrot_training - Step 1865: {'lr': 0.00046625000000000003, 'samples': 955392, 'steps': 1865, 'loss/train': 2.394439697265625} 02/24/2022 03:59:37 - INFO - codeparrot_training - Step 1866: {'lr': 0.0004665, 'samples': 955904, 'steps': 1866, 'loss/train': 4.33760404586792} 02/24/2022 03:59:40 - INFO - codeparrot_training - Step 1867: {'lr': 0.00046675, 'samples': 956416, 'steps': 1867, 'loss/train': 3.278625011444092} 02/24/2022 03:59:47 - INFO - codeparrot_training - Step 1868: {'lr': 0.000467, 'samples': 956928, 'steps': 1868, 'loss/train': 4.229162693023682} 02/24/2022 03:59:51 - INFO - codeparrot_training - Step 1869: {'lr': 0.00046725, 'samples': 957440, 'steps': 1869, 'loss/train': 2.944370746612549} 02/24/2022 03:59:56 - INFO - codeparrot_training - Step 1870: {'lr': 0.00046750000000000003, 'samples': 957952, 'steps': 1870, 'loss/train': 3.1038155555725098} 02/24/2022 03:59:59 - INFO - codeparrot_training - Step 1871: {'lr': 0.00046775, 'samples': 958464, 'steps': 1871, 'loss/train': 3.9347496032714844} 02/24/2022 04:00:05 - INFO - codeparrot_training - Step 1872: {'lr': 0.00046800000000000005, 'samples': 958976, 'steps': 1872, 'loss/train': 6.10337495803833} 02/24/2022 04:00:08 - INFO - codeparrot_training - Step 1873: {'lr': 0.00046825, 'samples': 959488, 'steps': 1873, 'loss/train': 3.674490451812744} 02/24/2022 04:00:14 - INFO - codeparrot_training - Step 1874: {'lr': 0.00046850000000000006, 'samples': 960000, 'steps': 1874, 'loss/train': 2.7155466079711914} 02/24/2022 04:00:17 - INFO - codeparrot_training - Step 1875: {'lr': 0.00046875, 'samples': 960512, 'steps': 1875, 'loss/train': 3.057251453399658} 02/24/2022 04:00:23 - INFO - codeparrot_training - Step 1876: {'lr': 0.00046899999999999996, 'samples': 961024, 'steps': 1876, 'loss/train': 0.8594743609428406} 02/24/2022 04:00:26 - INFO - codeparrot_training - Step 1877: {'lr': 0.00046925, 'samples': 961536, 'steps': 1877, 'loss/train': 1.3336923122406006} 02/24/2022 04:00:32 - INFO - codeparrot_training - Step 1878: {'lr': 0.0004695, 'samples': 962048, 'steps': 1878, 'loss/train': 3.075562000274658} 02/24/2022 04:00:35 - INFO - codeparrot_training - Step 1879: {'lr': 0.00046975, 'samples': 962560, 'steps': 1879, 'loss/train': 4.606093883514404} 02/24/2022 04:00:41 - INFO - codeparrot_training - Step 1880: {'lr': 0.00047, 'samples': 963072, 'steps': 1880, 'loss/train': 3.365853786468506} 02/24/2022 04:00:44 - INFO - codeparrot_training - Step 1881: {'lr': 0.00047025, 'samples': 963584, 'steps': 1881, 'loss/train': 3.5819649696350098} 02/24/2022 04:00:50 - INFO - codeparrot_training - Step 1882: {'lr': 0.0004705, 'samples': 964096, 'steps': 1882, 'loss/train': 2.9109272956848145} 02/24/2022 04:00:53 - INFO - codeparrot_training - Step 1883: {'lr': 0.00047075000000000003, 'samples': 964608, 'steps': 1883, 'loss/train': 3.848289728164673} 02/24/2022 04:00:59 - INFO - codeparrot_training - Step 1884: {'lr': 0.000471, 'samples': 965120, 'steps': 1884, 'loss/train': 3.2787702083587646} 02/24/2022 04:01:03 - INFO - codeparrot_training - Step 1885: {'lr': 0.00047125, 'samples': 965632, 'steps': 1885, 'loss/train': 3.697462320327759} 02/24/2022 04:01:08 - INFO - codeparrot_training - Step 1886: {'lr': 0.00047149999999999997, 'samples': 966144, 'steps': 1886, 'loss/train': 4.535035610198975} 02/24/2022 04:01:12 - INFO - codeparrot_training - Step 1887: {'lr': 0.00047175, 'samples': 966656, 'steps': 1887, 'loss/train': 0.8160093426704407} 02/24/2022 04:01:17 - INFO - codeparrot_training - Step 1888: {'lr': 0.000472, 'samples': 967168, 'steps': 1888, 'loss/train': 4.502017974853516} 02/24/2022 04:01:21 - INFO - codeparrot_training - Step 1889: {'lr': 0.00047225, 'samples': 967680, 'steps': 1889, 'loss/train': 3.7890665531158447} 02/24/2022 04:01:26 - INFO - codeparrot_training - Step 1890: {'lr': 0.0004725, 'samples': 968192, 'steps': 1890, 'loss/train': 4.392154693603516} 02/24/2022 04:01:30 - INFO - codeparrot_training - Step 1891: {'lr': 0.00047275, 'samples': 968704, 'steps': 1891, 'loss/train': 2.513261318206787} 02/24/2022 04:01:35 - INFO - codeparrot_training - Step 1892: {'lr': 0.000473, 'samples': 969216, 'steps': 1892, 'loss/train': 3.9184131622314453} 02/24/2022 04:01:39 - INFO - codeparrot_training - Step 1893: {'lr': 0.00047325000000000004, 'samples': 969728, 'steps': 1893, 'loss/train': 4.705053329467773} 02/24/2022 04:01:45 - INFO - codeparrot_training - Step 1894: {'lr': 0.00047349999999999996, 'samples': 970240, 'steps': 1894, 'loss/train': 4.477334976196289} 02/24/2022 04:01:48 - INFO - codeparrot_training - Step 1895: {'lr': 0.00047375, 'samples': 970752, 'steps': 1895, 'loss/train': 4.102129936218262} 02/24/2022 04:01:54 - INFO - codeparrot_training - Step 1896: {'lr': 0.000474, 'samples': 971264, 'steps': 1896, 'loss/train': 3.6128153800964355} 02/24/2022 04:01:57 - INFO - codeparrot_training - Step 1897: {'lr': 0.00047425, 'samples': 971776, 'steps': 1897, 'loss/train': 1.9788767099380493} 02/24/2022 04:02:03 - INFO - codeparrot_training - Step 1898: {'lr': 0.0004745, 'samples': 972288, 'steps': 1898, 'loss/train': 5.206483364105225} 02/24/2022 04:02:07 - INFO - codeparrot_training - Step 1899: {'lr': 0.00047475, 'samples': 972800, 'steps': 1899, 'loss/train': 0.6392098665237427} 02/24/2022 04:02:12 - INFO - codeparrot_training - Step 1900: {'lr': 0.000475, 'samples': 973312, 'steps': 1900, 'loss/train': 3.522040367126465} 02/24/2022 04:02:17 - INFO - codeparrot_training - Step 1901: {'lr': 0.00047525000000000003, 'samples': 973824, 'steps': 1901, 'loss/train': 3.5939764976501465} 02/24/2022 04:02:21 - INFO - codeparrot_training - Step 1902: {'lr': 0.0004755, 'samples': 974336, 'steps': 1902, 'loss/train': 4.5955352783203125} 02/24/2022 04:02:27 - INFO - codeparrot_training - Step 1903: {'lr': 0.00047575, 'samples': 974848, 'steps': 1903, 'loss/train': 3.5274877548217773} 02/24/2022 04:02:31 - INFO - codeparrot_training - Step 1904: {'lr': 0.00047599999999999997, 'samples': 975360, 'steps': 1904, 'loss/train': 3.993286371231079} 02/24/2022 04:02:36 - INFO - codeparrot_training - Step 1905: {'lr': 0.00047625, 'samples': 975872, 'steps': 1905, 'loss/train': 3.3120734691619873} 02/24/2022 04:02:39 - INFO - codeparrot_training - Step 1906: {'lr': 0.0004765, 'samples': 976384, 'steps': 1906, 'loss/train': 3.9515020847320557} 02/24/2022 04:02:45 - INFO - codeparrot_training - Step 1907: {'lr': 0.00047675, 'samples': 976896, 'steps': 1907, 'loss/train': 4.023288249969482} 02/24/2022 04:02:48 - INFO - codeparrot_training - Step 1908: {'lr': 0.000477, 'samples': 977408, 'steps': 1908, 'loss/train': 3.07366681098938} 02/24/2022 04:02:54 - INFO - codeparrot_training - Step 1909: {'lr': 0.00047725, 'samples': 977920, 'steps': 1909, 'loss/train': 4.2689666748046875} 02/24/2022 04:02:57 - INFO - codeparrot_training - Step 1910: {'lr': 0.0004775, 'samples': 978432, 'steps': 1910, 'loss/train': 4.167940616607666} 02/24/2022 04:03:03 - INFO - codeparrot_training - Step 1911: {'lr': 0.00047775000000000004, 'samples': 978944, 'steps': 1911, 'loss/train': 4.071891784667969} 02/24/2022 04:03:06 - INFO - codeparrot_training - Step 1912: {'lr': 0.00047799999999999996, 'samples': 979456, 'steps': 1912, 'loss/train': 3.5444371700286865} 02/24/2022 04:03:13 - INFO - codeparrot_training - Step 1913: {'lr': 0.00047825, 'samples': 979968, 'steps': 1913, 'loss/train': 3.6563661098480225} 02/24/2022 04:03:17 - INFO - codeparrot_training - Step 1914: {'lr': 0.0004785, 'samples': 980480, 'steps': 1914, 'loss/train': 3.701944351196289} 02/24/2022 04:03:22 - INFO - codeparrot_training - Step 1915: {'lr': 0.00047875, 'samples': 980992, 'steps': 1915, 'loss/train': 2.2404544353485107} 02/24/2022 04:03:25 - INFO - codeparrot_training - Step 1916: {'lr': 0.000479, 'samples': 981504, 'steps': 1916, 'loss/train': 4.324679374694824} 02/24/2022 04:03:31 - INFO - codeparrot_training - Step 1917: {'lr': 0.00047925, 'samples': 982016, 'steps': 1917, 'loss/train': 3.3296546936035156} 02/24/2022 04:03:34 - INFO - codeparrot_training - Step 1918: {'lr': 0.0004795, 'samples': 982528, 'steps': 1918, 'loss/train': 4.014566898345947} 02/24/2022 04:03:40 - INFO - codeparrot_training - Step 1919: {'lr': 0.00047975000000000003, 'samples': 983040, 'steps': 1919, 'loss/train': 2.519977331161499} 02/24/2022 04:03:43 - INFO - codeparrot_training - Step 1920: {'lr': 0.00048, 'samples': 983552, 'steps': 1920, 'loss/train': 3.5396993160247803} 02/24/2022 04:03:49 - INFO - codeparrot_training - Step 1921: {'lr': 0.00048025000000000005, 'samples': 984064, 'steps': 1921, 'loss/train': 4.678669452667236} 02/24/2022 04:03:52 - INFO - codeparrot_training - Step 1922: {'lr': 0.00048049999999999997, 'samples': 984576, 'steps': 1922, 'loss/train': 3.478494882583618} 02/24/2022 04:03:58 - INFO - codeparrot_training - Step 1923: {'lr': 0.00048075, 'samples': 985088, 'steps': 1923, 'loss/train': 2.2176525592803955} 02/24/2022 04:04:01 - INFO - codeparrot_training - Step 1924: {'lr': 0.000481, 'samples': 985600, 'steps': 1924, 'loss/train': 4.056451797485352} 02/24/2022 04:04:07 - INFO - codeparrot_training - Step 1925: {'lr': 0.00048125, 'samples': 986112, 'steps': 1925, 'loss/train': 5.742905616760254} 02/24/2022 04:04:10 - INFO - codeparrot_training - Step 1926: {'lr': 0.0004815, 'samples': 986624, 'steps': 1926, 'loss/train': 4.071230888366699} 02/24/2022 04:04:16 - INFO - codeparrot_training - Step 1927: {'lr': 0.00048175000000000003, 'samples': 987136, 'steps': 1927, 'loss/train': 4.317747592926025} 02/24/2022 04:04:19 - INFO - codeparrot_training - Step 1928: {'lr': 0.000482, 'samples': 987648, 'steps': 1928, 'loss/train': 3.545318365097046} 02/24/2022 04:04:25 - INFO - codeparrot_training - Step 1929: {'lr': 0.00048225000000000004, 'samples': 988160, 'steps': 1929, 'loss/train': 4.278107166290283} 02/24/2022 04:04:29 - INFO - codeparrot_training - Step 1930: {'lr': 0.0004825, 'samples': 988672, 'steps': 1930, 'loss/train': 3.470717191696167} 02/24/2022 04:04:34 - INFO - codeparrot_training - Step 1931: {'lr': 0.00048275, 'samples': 989184, 'steps': 1931, 'loss/train': 4.630705833435059} 02/24/2022 04:04:38 - INFO - codeparrot_training - Step 1932: {'lr': 0.000483, 'samples': 989696, 'steps': 1932, 'loss/train': 4.433801651000977} 02/24/2022 04:04:43 - INFO - codeparrot_training - Step 1933: {'lr': 0.00048325, 'samples': 990208, 'steps': 1933, 'loss/train': 3.111100673675537} 02/24/2022 04:04:47 - INFO - codeparrot_training - Step 1934: {'lr': 0.0004835, 'samples': 990720, 'steps': 1934, 'loss/train': 4.2302565574646} 02/24/2022 04:04:52 - INFO - codeparrot_training - Step 1935: {'lr': 0.00048375, 'samples': 991232, 'steps': 1935, 'loss/train': 4.6753129959106445} 02/24/2022 04:04:56 - INFO - codeparrot_training - Step 1936: {'lr': 0.000484, 'samples': 991744, 'steps': 1936, 'loss/train': 4.574672698974609} 02/24/2022 04:05:01 - INFO - codeparrot_training - Step 1937: {'lr': 0.00048425000000000003, 'samples': 992256, 'steps': 1937, 'loss/train': 0.7327980399131775} 02/24/2022 04:05:07 - INFO - codeparrot_training - Step 1938: {'lr': 0.0004845, 'samples': 992768, 'steps': 1938, 'loss/train': 4.488827705383301} 02/24/2022 04:05:11 - INFO - codeparrot_training - Step 1939: {'lr': 0.00048475000000000005, 'samples': 993280, 'steps': 1939, 'loss/train': 3.215743064880371} 02/24/2022 04:05:16 - INFO - codeparrot_training - Step 1940: {'lr': 0.00048499999999999997, 'samples': 993792, 'steps': 1940, 'loss/train': 2.5476274490356445} 02/24/2022 04:05:19 - INFO - codeparrot_training - Step 1941: {'lr': 0.00048525, 'samples': 994304, 'steps': 1941, 'loss/train': 3.6032984256744385} 02/24/2022 04:05:25 - INFO - codeparrot_training - Step 1942: {'lr': 0.0004855, 'samples': 994816, 'steps': 1942, 'loss/train': 3.7020487785339355} 02/24/2022 04:05:28 - INFO - codeparrot_training - Step 1943: {'lr': 0.00048575, 'samples': 995328, 'steps': 1943, 'loss/train': 4.331693649291992} 02/24/2022 04:05:34 - INFO - codeparrot_training - Step 1944: {'lr': 0.000486, 'samples': 995840, 'steps': 1944, 'loss/train': 1.9530490636825562} 02/24/2022 04:05:37 - INFO - codeparrot_training - Step 1945: {'lr': 0.00048625000000000003, 'samples': 996352, 'steps': 1945, 'loss/train': 2.9386301040649414} 02/24/2022 04:05:43 - INFO - codeparrot_training - Step 1946: {'lr': 0.0004865, 'samples': 996864, 'steps': 1946, 'loss/train': 4.366998672485352} 02/24/2022 04:05:46 - INFO - codeparrot_training - Step 1947: {'lr': 0.00048675000000000004, 'samples': 997376, 'steps': 1947, 'loss/train': 3.3206379413604736} 02/24/2022 04:05:52 - INFO - codeparrot_training - Step 1948: {'lr': 0.000487, 'samples': 997888, 'steps': 1948, 'loss/train': 3.3380088806152344} 02/24/2022 04:05:55 - INFO - codeparrot_training - Step 1949: {'lr': 0.00048725000000000005, 'samples': 998400, 'steps': 1949, 'loss/train': 3.6555140018463135} 02/24/2022 04:06:01 - INFO - codeparrot_training - Step 1950: {'lr': 0.0004875, 'samples': 998912, 'steps': 1950, 'loss/train': 2.818737745285034} 02/24/2022 04:06:05 - INFO - codeparrot_training - Step 1951: {'lr': 0.00048775, 'samples': 999424, 'steps': 1951, 'loss/train': 2.6505634784698486} 02/24/2022 04:06:10 - INFO - codeparrot_training - Step 1952: {'lr': 0.000488, 'samples': 999936, 'steps': 1952, 'loss/train': 3.7129604816436768} 02/24/2022 04:06:14 - INFO - codeparrot_training - Step 1953: {'lr': 0.00048825, 'samples': 1000448, 'steps': 1953, 'loss/train': 4.589820861816406} 02/24/2022 04:06:19 - INFO - codeparrot_training - Step 1954: {'lr': 0.0004885, 'samples': 1000960, 'steps': 1954, 'loss/train': 3.3188226222991943} 02/24/2022 04:06:23 - INFO - codeparrot_training - Step 1955: {'lr': 0.00048875, 'samples': 1001472, 'steps': 1955, 'loss/train': 4.6001410484313965} 02/24/2022 04:06:28 - INFO - codeparrot_training - Step 1956: {'lr': 0.000489, 'samples': 1001984, 'steps': 1956, 'loss/train': 3.3966524600982666} 02/24/2022 04:06:32 - INFO - codeparrot_training - Step 1957: {'lr': 0.00048925, 'samples': 1002496, 'steps': 1957, 'loss/train': 3.7696356773376465} 02/24/2022 04:06:37 - INFO - codeparrot_training - Step 1958: {'lr': 0.0004895, 'samples': 1003008, 'steps': 1958, 'loss/train': 1.4602926969528198} 02/24/2022 04:06:41 - INFO - codeparrot_training - Step 1959: {'lr': 0.0004897500000000001, 'samples': 1003520, 'steps': 1959, 'loss/train': 2.9611456394195557} 02/24/2022 04:06:47 - INFO - codeparrot_training - Step 1960: {'lr': 0.00049, 'samples': 1004032, 'steps': 1960, 'loss/train': 4.06137228012085} 02/24/2022 04:06:50 - INFO - codeparrot_training - Step 1961: {'lr': 0.00049025, 'samples': 1004544, 'steps': 1961, 'loss/train': 3.5533761978149414} 02/24/2022 04:06:56 - INFO - codeparrot_training - Step 1962: {'lr': 0.0004905, 'samples': 1005056, 'steps': 1962, 'loss/train': 4.941059112548828} 02/24/2022 04:06:59 - INFO - codeparrot_training - Step 1963: {'lr': 0.0004907500000000001, 'samples': 1005568, 'steps': 1963, 'loss/train': 3.677499532699585} 02/24/2022 04:07:05 - INFO - codeparrot_training - Step 1964: {'lr': 0.000491, 'samples': 1006080, 'steps': 1964, 'loss/train': 2.110358238220215} 02/24/2022 04:07:08 - INFO - codeparrot_training - Step 1965: {'lr': 0.00049125, 'samples': 1006592, 'steps': 1965, 'loss/train': 3.6638340950012207} 02/24/2022 04:07:14 - INFO - codeparrot_training - Step 1966: {'lr': 0.0004915, 'samples': 1007104, 'steps': 1966, 'loss/train': 2.495577335357666} 02/24/2022 04:07:17 - INFO - codeparrot_training - Step 1967: {'lr': 0.00049175, 'samples': 1007616, 'steps': 1967, 'loss/train': 3.320880889892578} 02/24/2022 04:07:23 - INFO - codeparrot_training - Step 1968: {'lr': 0.000492, 'samples': 1008128, 'steps': 1968, 'loss/train': 3.538256883621216} 02/24/2022 04:07:26 - INFO - codeparrot_training - Step 1969: {'lr': 0.0004922500000000001, 'samples': 1008640, 'steps': 1969, 'loss/train': 3.474503517150879} 02/24/2022 04:07:31 - INFO - codeparrot_training - Step 1970: {'lr': 0.0004925, 'samples': 1009152, 'steps': 1970, 'loss/train': 3.4210190773010254} 02/24/2022 04:07:35 - INFO - codeparrot_training - Step 1971: {'lr': 0.00049275, 'samples': 1009664, 'steps': 1971, 'loss/train': 3.2094409465789795} 02/24/2022 04:07:40 - INFO - codeparrot_training - Step 1972: {'lr': 0.0004930000000000001, 'samples': 1010176, 'steps': 1972, 'loss/train': 3.4008500576019287} 02/24/2022 04:07:44 - INFO - codeparrot_training - Step 1973: {'lr': 0.00049325, 'samples': 1010688, 'steps': 1973, 'loss/train': 3.1762261390686035} 02/24/2022 04:07:49 - INFO - codeparrot_training - Step 1974: {'lr': 0.0004935, 'samples': 1011200, 'steps': 1974, 'loss/train': 3.091787338256836} 02/24/2022 04:07:53 - INFO - codeparrot_training - Step 1975: {'lr': 0.00049375, 'samples': 1011712, 'steps': 1975, 'loss/train': 1.7174049615859985} 02/24/2022 04:07:59 - INFO - codeparrot_training - Step 1976: {'lr': 0.000494, 'samples': 1012224, 'steps': 1976, 'loss/train': 4.197392463684082} 02/24/2022 04:08:03 - INFO - codeparrot_training - Step 1977: {'lr': 0.00049425, 'samples': 1012736, 'steps': 1977, 'loss/train': 2.6754541397094727} 02/24/2022 04:08:08 - INFO - codeparrot_training - Step 1978: {'lr': 0.0004945, 'samples': 1013248, 'steps': 1978, 'loss/train': 3.13645601272583} 02/24/2022 04:08:11 - INFO - codeparrot_training - Step 1979: {'lr': 0.0004947500000000001, 'samples': 1013760, 'steps': 1979, 'loss/train': 4.084655284881592} 02/24/2022 04:08:17 - INFO - codeparrot_training - Step 1980: {'lr': 0.000495, 'samples': 1014272, 'steps': 1980, 'loss/train': 4.445559501647949} 02/24/2022 04:08:20 - INFO - codeparrot_training - Step 1981: {'lr': 0.00049525, 'samples': 1014784, 'steps': 1981, 'loss/train': 2.8948557376861572} 02/24/2022 04:08:26 - INFO - codeparrot_training - Step 1982: {'lr': 0.0004955, 'samples': 1015296, 'steps': 1982, 'loss/train': 3.580780029296875} 02/24/2022 04:08:29 - INFO - codeparrot_training - Step 1983: {'lr': 0.00049575, 'samples': 1015808, 'steps': 1983, 'loss/train': 3.9996466636657715} 02/24/2022 04:08:35 - INFO - codeparrot_training - Step 1984: {'lr': 0.000496, 'samples': 1016320, 'steps': 1984, 'loss/train': 3.889043092727661} 02/24/2022 04:08:38 - INFO - codeparrot_training - Step 1985: {'lr': 0.0004962500000000001, 'samples': 1016832, 'steps': 1985, 'loss/train': 2.5638394355773926} 02/24/2022 04:08:45 - INFO - codeparrot_training - Step 1986: {'lr': 0.0004965, 'samples': 1017344, 'steps': 1986, 'loss/train': 4.096147537231445} 02/24/2022 04:08:48 - INFO - codeparrot_training - Step 1987: {'lr': 0.00049675, 'samples': 1017856, 'steps': 1987, 'loss/train': 3.9539005756378174} 02/24/2022 04:08:54 - INFO - codeparrot_training - Step 1988: {'lr': 0.000497, 'samples': 1018368, 'steps': 1988, 'loss/train': 1.1471844911575317} 02/24/2022 04:08:57 - INFO - codeparrot_training - Step 1989: {'lr': 0.0004972500000000001, 'samples': 1018880, 'steps': 1989, 'loss/train': 2.479738235473633} 02/24/2022 04:09:03 - INFO - codeparrot_training - Step 1990: {'lr': 0.0004975, 'samples': 1019392, 'steps': 1990, 'loss/train': 3.169276237487793} 02/24/2022 04:09:06 - INFO - codeparrot_training - Step 1991: {'lr': 0.00049775, 'samples': 1019904, 'steps': 1991, 'loss/train': 3.661322593688965} 02/24/2022 04:09:12 - INFO - codeparrot_training - Step 1992: {'lr': 0.000498, 'samples': 1020416, 'steps': 1992, 'loss/train': 3.7599399089813232} 02/24/2022 04:09:17 - INFO - codeparrot_training - Step 1993: {'lr': 0.00049825, 'samples': 1020928, 'steps': 1993, 'loss/train': 3.4758248329162598} 02/24/2022 04:09:20 - INFO - codeparrot_training - Step 1994: {'lr': 0.0004985, 'samples': 1021440, 'steps': 1994, 'loss/train': 6.070404052734375} 02/24/2022 04:09:27 - INFO - codeparrot_training - Step 1995: {'lr': 0.0004987500000000001, 'samples': 1021952, 'steps': 1995, 'loss/train': 4.1490583419799805} 02/24/2022 04:09:30 - INFO - codeparrot_training - Step 1996: {'lr': 0.000499, 'samples': 1022464, 'steps': 1996, 'loss/train': 3.0770957469940186} 02/24/2022 04:09:36 - INFO - codeparrot_training - Step 1997: {'lr': 0.00049925, 'samples': 1022976, 'steps': 1997, 'loss/train': 3.3806257247924805} 02/24/2022 04:09:39 - INFO - codeparrot_training - Step 1998: {'lr': 0.0004995, 'samples': 1023488, 'steps': 1998, 'loss/train': 2.887777805328369} 02/24/2022 04:09:45 - INFO - codeparrot_training - Step 1999: {'lr': 0.0004997500000000001, 'samples': 1024000, 'steps': 1999, 'loss/train': 9.467606544494629} 02/24/2022 04:09:45 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 04:10:02 - WARNING - huggingface_hub.repository - Several commits (2) will be pushed upstream. 02/24/2022 04:10:02 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 04:10:36 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy b1f5d23..0998bf2 floral-grass-11 -> floral-grass-11 02/24/2022 04:10:41 - INFO - codeparrot_training - Step 2000: {'lr': 0.0005, 'samples': 1024512, 'steps': 2000, 'loss/train': 2.800246477127075} 02/24/2022 04:10:44 - INFO - codeparrot_training - Step 2001: {'lr': 0.0004999999994645397, 'samples': 1025024, 'steps': 2001, 'loss/train': 0.9297313094139099} 02/24/2022 04:10:50 - INFO - codeparrot_training - Step 2002: {'lr': 0.0004999999978581587, 'samples': 1025536, 'steps': 2002, 'loss/train': 3.4139480590820312} 02/24/2022 04:10:53 - INFO - codeparrot_training - Step 2003: {'lr': 0.0004999999951808573, 'samples': 1026048, 'steps': 2003, 'loss/train': 1.5185656547546387} 02/24/2022 04:10:59 - INFO - codeparrot_training - Step 2004: {'lr': 0.0004999999914326351, 'samples': 1026560, 'steps': 2004, 'loss/train': 3.569647789001465} 02/24/2022 04:11:05 - INFO - codeparrot_training - Step 2005: {'lr': 0.0004999999866134924, 'samples': 1027072, 'steps': 2005, 'loss/train': 3.923659324645996} 02/24/2022 04:11:08 - INFO - codeparrot_training - Step 2006: {'lr': 0.0004999999807234292, 'samples': 1027584, 'steps': 2006, 'loss/train': 4.813257694244385} 02/24/2022 04:11:14 - INFO - codeparrot_training - Step 2007: {'lr': 0.0004999999737624453, 'samples': 1028096, 'steps': 2007, 'loss/train': 3.247771978378296} 02/24/2022 04:11:17 - INFO - codeparrot_training - Step 2008: {'lr': 0.0004999999657305411, 'samples': 1028608, 'steps': 2008, 'loss/train': 2.858778715133667} 02/24/2022 04:11:23 - INFO - codeparrot_training - Step 2009: {'lr': 0.0004999999566277163, 'samples': 1029120, 'steps': 2009, 'loss/train': 3.722787857055664} 02/24/2022 04:11:26 - INFO - codeparrot_training - Step 2010: {'lr': 0.0004999999464539711, 'samples': 1029632, 'steps': 2010, 'loss/train': 1.810246467590332} 02/24/2022 04:11:32 - INFO - codeparrot_training - Step 2011: {'lr': 0.0004999999352093055, 'samples': 1030144, 'steps': 2011, 'loss/train': 3.4813263416290283} 02/24/2022 04:11:35 - INFO - codeparrot_training - Step 2012: {'lr': 0.0004999999228937196, 'samples': 1030656, 'steps': 2012, 'loss/train': 3.003511667251587} 02/24/2022 04:11:41 - INFO - codeparrot_training - Step 2013: {'lr': 0.0004999999095072135, 'samples': 1031168, 'steps': 2013, 'loss/train': 3.13137149810791} 02/24/2022 04:11:44 - INFO - codeparrot_training - Step 2014: {'lr': 0.0004999998950497869, 'samples': 1031680, 'steps': 2014, 'loss/train': 2.7934157848358154} 02/24/2022 04:11:50 - INFO - codeparrot_training - Step 2015: {'lr': 0.0004999998795214404, 'samples': 1032192, 'steps': 2015, 'loss/train': 5.817224979400635} 02/24/2022 04:11:54 - INFO - codeparrot_training - Step 2016: {'lr': 0.0004999998629221736, 'samples': 1032704, 'steps': 2016, 'loss/train': 3.418705701828003} 02/24/2022 04:11:59 - INFO - codeparrot_training - Step 2017: {'lr': 0.0004999998452519869, 'samples': 1033216, 'steps': 2017, 'loss/train': 3.1369762420654297} 02/24/2022 04:12:03 - INFO - codeparrot_training - Step 2018: {'lr': 0.0004999998265108802, 'samples': 1033728, 'steps': 2018, 'loss/train': 3.6667239665985107} 02/24/2022 04:12:06 - INFO - codeparrot_training - Step 2019: {'lr': 0.0004999998066988537, 'samples': 1034240, 'steps': 2019, 'loss/train': 2.634143352508545} 02/24/2022 04:12:12 - INFO - codeparrot_training - Step 2020: {'lr': 0.0004999997858159073, 'samples': 1034752, 'steps': 2020, 'loss/train': 4.146056652069092} 02/24/2022 04:12:18 - INFO - codeparrot_training - Step 2021: {'lr': 0.0004999997638620412, 'samples': 1035264, 'steps': 2021, 'loss/train': 3.6164910793304443} 02/24/2022 04:12:21 - INFO - codeparrot_training - Step 2022: {'lr': 0.0004999997408372557, 'samples': 1035776, 'steps': 2022, 'loss/train': 4.4506402015686035} 02/24/2022 04:12:26 - INFO - codeparrot_training - Step 2023: {'lr': 0.0004999997167415504, 'samples': 1036288, 'steps': 2023, 'loss/train': 3.631222724914551} 02/24/2022 04:12:30 - INFO - codeparrot_training - Step 2024: {'lr': 0.0004999996915749259, 'samples': 1036800, 'steps': 2024, 'loss/train': 3.945448398590088} 02/24/2022 04:12:36 - INFO - codeparrot_training - Step 2025: {'lr': 0.0004999996653373821, 'samples': 1037312, 'steps': 2025, 'loss/train': 3.8391072750091553} 02/24/2022 04:12:39 - INFO - codeparrot_training - Step 2026: {'lr': 0.000499999638028919, 'samples': 1037824, 'steps': 2026, 'loss/train': 2.200242280960083} 02/24/2022 04:12:44 - INFO - codeparrot_training - Step 2027: {'lr': 0.0004999996096495369, 'samples': 1038336, 'steps': 2027, 'loss/train': 3.019376277923584} 02/24/2022 04:12:48 - INFO - codeparrot_training - Step 2028: {'lr': 0.0004999995801992359, 'samples': 1038848, 'steps': 2028, 'loss/train': 2.8726444244384766} 02/24/2022 04:12:53 - INFO - codeparrot_training - Step 2029: {'lr': 0.000499999549678016, 'samples': 1039360, 'steps': 2029, 'loss/train': 3.873802423477173} 02/24/2022 04:12:57 - INFO - codeparrot_training - Step 2030: {'lr': 0.0004999995180858774, 'samples': 1039872, 'steps': 2030, 'loss/train': 4.297247886657715} 02/24/2022 04:13:04 - INFO - codeparrot_training - Step 2031: {'lr': 0.0004999994854228203, 'samples': 1040384, 'steps': 2031, 'loss/train': 3.6109445095062256} 02/24/2022 04:13:07 - INFO - codeparrot_training - Step 2032: {'lr': 0.0004999994516888449, 'samples': 1040896, 'steps': 2032, 'loss/train': 5.221895694732666} 02/24/2022 04:13:13 - INFO - codeparrot_training - Step 2033: {'lr': 0.000499999416883951, 'samples': 1041408, 'steps': 2033, 'loss/train': 4.7753705978393555} 02/24/2022 04:13:16 - INFO - codeparrot_training - Step 2034: {'lr': 0.0004999993810081391, 'samples': 1041920, 'steps': 2034, 'loss/train': 3.140744686126709} 02/24/2022 04:13:22 - INFO - codeparrot_training - Step 2035: {'lr': 0.0004999993440614092, 'samples': 1042432, 'steps': 2035, 'loss/train': 3.2063350677490234} 02/24/2022 04:13:25 - INFO - codeparrot_training - Step 2036: {'lr': 0.0004999993060437616, 'samples': 1042944, 'steps': 2036, 'loss/train': 3.360121250152588} 02/24/2022 04:13:31 - INFO - codeparrot_training - Step 2037: {'lr': 0.0004999992669551962, 'samples': 1043456, 'steps': 2037, 'loss/train': 3.6778557300567627} 02/24/2022 04:13:34 - INFO - codeparrot_training - Step 2038: {'lr': 0.0004999992267957135, 'samples': 1043968, 'steps': 2038, 'loss/train': 3.849402666091919} 02/24/2022 04:13:40 - INFO - codeparrot_training - Step 2039: {'lr': 0.0004999991855653134, 'samples': 1044480, 'steps': 2039, 'loss/train': 3.1258656978607178} 02/24/2022 04:13:43 - INFO - codeparrot_training - Step 2040: {'lr': 0.0004999991432639963, 'samples': 1044992, 'steps': 2040, 'loss/train': 4.852616786956787} 02/24/2022 04:13:50 - INFO - codeparrot_training - Step 2041: {'lr': 0.0004999990998917621, 'samples': 1045504, 'steps': 2041, 'loss/train': 3.8511712551116943} 02/24/2022 04:13:53 - INFO - codeparrot_training - Step 2042: {'lr': 0.0004999990554486111, 'samples': 1046016, 'steps': 2042, 'loss/train': 4.233978748321533} 02/24/2022 04:13:59 - INFO - codeparrot_training - Step 2043: {'lr': 0.0004999990099345436, 'samples': 1046528, 'steps': 2043, 'loss/train': 3.5593159198760986} 02/24/2022 04:14:02 - INFO - codeparrot_training - Step 2044: {'lr': 0.0004999989633495597, 'samples': 1047040, 'steps': 2044, 'loss/train': 3.260753870010376} 02/24/2022 04:14:08 - INFO - codeparrot_training - Step 2045: {'lr': 0.0004999989156936597, 'samples': 1047552, 'steps': 2045, 'loss/train': 2.6524853706359863} 02/24/2022 04:14:12 - INFO - codeparrot_training - Step 2046: {'lr': 0.0004999988669668437, 'samples': 1048064, 'steps': 2046, 'loss/train': 3.4726810455322266} 02/24/2022 04:14:17 - INFO - codeparrot_training - Step 2047: {'lr': 0.0004999988171691119, 'samples': 1048576, 'steps': 2047, 'loss/train': 2.923299551010132} 02/24/2022 04:14:21 - INFO - codeparrot_training - Step 2048: {'lr': 0.0004999987663004646, 'samples': 1049088, 'steps': 2048, 'loss/train': 4.113481521606445} 02/24/2022 04:14:26 - INFO - codeparrot_training - Step 2049: {'lr': 0.0004999987143609019, 'samples': 1049600, 'steps': 2049, 'loss/train': 0.8320827484130859} 02/24/2022 04:14:30 - INFO - codeparrot_training - Step 2050: {'lr': 0.0004999986613504242, 'samples': 1050112, 'steps': 2050, 'loss/train': 4.12211275100708} 02/24/2022 04:14:36 - INFO - codeparrot_training - Step 2051: {'lr': 0.0004999986072690315, 'samples': 1050624, 'steps': 2051, 'loss/train': 4.0427093505859375} 02/24/2022 04:14:39 - INFO - codeparrot_training - Step 2052: {'lr': 0.0004999985521167242, 'samples': 1051136, 'steps': 2052, 'loss/train': 3.8508830070495605} 02/24/2022 04:14:45 - INFO - codeparrot_training - Step 2053: {'lr': 0.0004999984958935025, 'samples': 1051648, 'steps': 2053, 'loss/train': 3.516901969909668} 02/24/2022 04:14:48 - INFO - codeparrot_training - Step 2054: {'lr': 0.0004999984385993665, 'samples': 1052160, 'steps': 2054, 'loss/train': 3.1342296600341797} 02/24/2022 04:14:54 - INFO - codeparrot_training - Step 2055: {'lr': 0.0004999983802343168, 'samples': 1052672, 'steps': 2055, 'loss/train': 4.581167697906494} 02/24/2022 04:14:57 - INFO - codeparrot_training - Step 2056: {'lr': 0.0004999983207983532, 'samples': 1053184, 'steps': 2056, 'loss/train': 3.29864501953125} 02/24/2022 04:15:03 - INFO - codeparrot_training - Step 2057: {'lr': 0.0004999982602914763, 'samples': 1053696, 'steps': 2057, 'loss/train': 4.555438995361328} 02/24/2022 04:15:06 - INFO - codeparrot_training - Step 2058: {'lr': 0.0004999981987136862, 'samples': 1054208, 'steps': 2058, 'loss/train': 2.4252705574035645} 02/24/2022 04:15:12 - INFO - codeparrot_training - Step 2059: {'lr': 0.0004999981360649833, 'samples': 1054720, 'steps': 2059, 'loss/train': 4.049717426300049} 02/24/2022 04:15:15 - INFO - codeparrot_training - Step 2060: {'lr': 0.0004999980723453676, 'samples': 1055232, 'steps': 2060, 'loss/train': 3.7873756885528564} 02/24/2022 04:15:21 - INFO - codeparrot_training - Step 2061: {'lr': 0.0004999980075548397, 'samples': 1055744, 'steps': 2061, 'loss/train': 3.2403149604797363} 02/24/2022 04:15:24 - INFO - codeparrot_training - Step 2062: {'lr': 0.0004999979416933997, 'samples': 1056256, 'steps': 2062, 'loss/train': 4.091283321380615} 02/24/2022 04:15:31 - INFO - codeparrot_training - Step 2063: {'lr': 0.0004999978747610478, 'samples': 1056768, 'steps': 2063, 'loss/train': 3.3648386001586914} 02/24/2022 04:15:34 - INFO - codeparrot_training - Step 2064: {'lr': 0.0004999978067577843, 'samples': 1057280, 'steps': 2064, 'loss/train': 3.1449785232543945} 02/24/2022 04:15:40 - INFO - codeparrot_training - Step 2065: {'lr': 0.0004999977376836098, 'samples': 1057792, 'steps': 2065, 'loss/train': 2.325490951538086} 02/24/2022 04:15:43 - INFO - codeparrot_training - Step 2066: {'lr': 0.0004999976675385243, 'samples': 1058304, 'steps': 2066, 'loss/train': 3.4341890811920166} 02/24/2022 04:15:49 - INFO - codeparrot_training - Step 2067: {'lr': 0.0004999975963225282, 'samples': 1058816, 'steps': 2067, 'loss/train': 2.971423864364624} 02/24/2022 04:15:52 - INFO - codeparrot_training - Step 2068: {'lr': 0.0004999975240356217, 'samples': 1059328, 'steps': 2068, 'loss/train': 4.039858818054199} 02/24/2022 04:15:58 - INFO - codeparrot_training - Step 2069: {'lr': 0.0004999974506778053, 'samples': 1059840, 'steps': 2069, 'loss/train': 0.8792083859443665} 02/24/2022 04:16:01 - INFO - codeparrot_training - Step 2070: {'lr': 0.0004999973762490792, 'samples': 1060352, 'steps': 2070, 'loss/train': 2.9441537857055664} 02/24/2022 04:16:07 - INFO - codeparrot_training - Step 2071: {'lr': 0.0004999973007494436, 'samples': 1060864, 'steps': 2071, 'loss/train': 3.1561279296875} 02/24/2022 04:16:10 - INFO - codeparrot_training - Step 2072: {'lr': 0.000499997224178899, 'samples': 1061376, 'steps': 2072, 'loss/train': 5.323057651519775} 02/24/2022 04:16:16 - INFO - codeparrot_training - Step 2073: {'lr': 0.0004999971465374457, 'samples': 1061888, 'steps': 2073, 'loss/train': 3.414630651473999} 02/24/2022 04:16:19 - INFO - codeparrot_training - Step 2074: {'lr': 0.000499997067825084, 'samples': 1062400, 'steps': 2074, 'loss/train': 3.895528554916382} 02/24/2022 04:16:24 - INFO - codeparrot_training - Step 2075: {'lr': 0.0004999969880418142, 'samples': 1062912, 'steps': 2075, 'loss/train': 3.8276734352111816} 02/24/2022 04:16:28 - INFO - codeparrot_training - Step 2076: {'lr': 0.0004999969071876367, 'samples': 1063424, 'steps': 2076, 'loss/train': 3.8214621543884277} 02/24/2022 04:16:34 - INFO - codeparrot_training - Step 2077: {'lr': 0.0004999968252625519, 'samples': 1063936, 'steps': 2077, 'loss/train': 3.8029026985168457} 02/24/2022 04:16:38 - INFO - codeparrot_training - Step 2078: {'lr': 0.00049999674226656, 'samples': 1064448, 'steps': 2078, 'loss/train': 4.751585960388184} 02/24/2022 04:16:43 - INFO - codeparrot_training - Step 2079: {'lr': 0.0004999966581996616, 'samples': 1064960, 'steps': 2079, 'loss/train': 4.121493816375732} 02/24/2022 04:16:49 - INFO - codeparrot_training - Step 2080: {'lr': 0.0004999965730618567, 'samples': 1065472, 'steps': 2080, 'loss/train': 4.1601691246032715} 02/24/2022 04:16:52 - INFO - codeparrot_training - Step 2081: {'lr': 0.000499996486853146, 'samples': 1065984, 'steps': 2081, 'loss/train': 3.698397397994995} 02/24/2022 04:16:58 - INFO - codeparrot_training - Step 2082: {'lr': 0.0004999963995735296, 'samples': 1066496, 'steps': 2082, 'loss/train': 2.953812599182129} 02/24/2022 04:17:01 - INFO - codeparrot_training - Step 2083: {'lr': 0.0004999963112230081, 'samples': 1067008, 'steps': 2083, 'loss/train': 3.1523876190185547} 02/24/2022 04:17:05 - INFO - codeparrot_training - Step 2084: {'lr': 0.0004999962218015818, 'samples': 1067520, 'steps': 2084, 'loss/train': 1.5381730794906616} 02/24/2022 04:17:10 - INFO - codeparrot_training - Step 2085: {'lr': 0.0004999961313092511, 'samples': 1068032, 'steps': 2085, 'loss/train': 3.5067222118377686} 02/24/2022 04:17:16 - INFO - codeparrot_training - Step 2086: {'lr': 0.0004999960397460162, 'samples': 1068544, 'steps': 2086, 'loss/train': 1.5887590646743774} 02/24/2022 04:17:20 - INFO - codeparrot_training - Step 2087: {'lr': 0.0004999959471118778, 'samples': 1069056, 'steps': 2087, 'loss/train': 2.0652477741241455} 02/24/2022 04:17:25 - INFO - codeparrot_training - Step 2088: {'lr': 0.000499995853406836, 'samples': 1069568, 'steps': 2088, 'loss/train': 3.348379373550415} 02/24/2022 04:17:29 - INFO - codeparrot_training - Step 2089: {'lr': 0.0004999957586308914, 'samples': 1070080, 'steps': 2089, 'loss/train': 2.952920436859131} 02/24/2022 04:17:34 - INFO - codeparrot_training - Step 2090: {'lr': 0.0004999956627840445, 'samples': 1070592, 'steps': 2090, 'loss/train': 2.571131706237793} 02/24/2022 04:17:38 - INFO - codeparrot_training - Step 2091: {'lr': 0.0004999955658662954, 'samples': 1071104, 'steps': 2091, 'loss/train': 3.6060545444488525} 02/24/2022 04:17:43 - INFO - codeparrot_training - Step 2092: {'lr': 0.0004999954678776448, 'samples': 1071616, 'steps': 2092, 'loss/train': 3.0546963214874268} 02/24/2022 04:17:47 - INFO - codeparrot_training - Step 2093: {'lr': 0.0004999953688180929, 'samples': 1072128, 'steps': 2093, 'loss/train': 3.4562463760375977} 02/24/2022 04:17:52 - INFO - codeparrot_training - Step 2094: {'lr': 0.0004999952686876402, 'samples': 1072640, 'steps': 2094, 'loss/train': 3.625727653503418} 02/24/2022 04:17:56 - INFO - codeparrot_training - Step 2095: {'lr': 0.0004999951674862872, 'samples': 1073152, 'steps': 2095, 'loss/train': 1.858146071434021} 02/24/2022 04:18:02 - INFO - codeparrot_training - Step 2096: {'lr': 0.0004999950652140343, 'samples': 1073664, 'steps': 2096, 'loss/train': 4.236659526824951} 02/24/2022 04:18:05 - INFO - codeparrot_training - Step 2097: {'lr': 0.0004999949618708819, 'samples': 1074176, 'steps': 2097, 'loss/train': 3.1121630668640137} 02/24/2022 04:18:11 - INFO - codeparrot_training - Step 2098: {'lr': 0.0004999948574568305, 'samples': 1074688, 'steps': 2098, 'loss/train': 3.973454236984253} 02/24/2022 04:18:14 - INFO - codeparrot_training - Step 2099: {'lr': 0.0004999947519718805, 'samples': 1075200, 'steps': 2099, 'loss/train': 2.847280502319336} 02/24/2022 04:18:20 - INFO - codeparrot_training - Step 2100: {'lr': 0.0004999946454160324, 'samples': 1075712, 'steps': 2100, 'loss/train': 5.648211479187012} 02/24/2022 04:18:23 - INFO - codeparrot_training - Step 2101: {'lr': 0.0004999945377892865, 'samples': 1076224, 'steps': 2101, 'loss/train': 2.1025211811065674} 02/24/2022 04:18:29 - INFO - codeparrot_training - Step 2102: {'lr': 0.0004999944290916434, 'samples': 1076736, 'steps': 2102, 'loss/train': 3.2467496395111084} 02/24/2022 04:18:32 - INFO - codeparrot_training - Step 2103: {'lr': 0.0004999943193231037, 'samples': 1077248, 'steps': 2103, 'loss/train': 3.7901408672332764} 02/24/2022 04:18:38 - INFO - codeparrot_training - Step 2104: {'lr': 0.0004999942084836675, 'samples': 1077760, 'steps': 2104, 'loss/train': 1.9659900665283203} 02/24/2022 04:18:41 - INFO - codeparrot_training - Step 2105: {'lr': 0.0004999940965733356, 'samples': 1078272, 'steps': 2105, 'loss/train': 3.02752685546875} 02/24/2022 04:18:47 - INFO - codeparrot_training - Step 2106: {'lr': 0.0004999939835921085, 'samples': 1078784, 'steps': 2106, 'loss/train': 3.097346067428589} 02/24/2022 04:18:50 - INFO - codeparrot_training - Step 2107: {'lr': 0.0004999938695399864, 'samples': 1079296, 'steps': 2107, 'loss/train': 2.835118532180786} 02/24/2022 04:18:56 - INFO - codeparrot_training - Step 2108: {'lr': 0.00049999375441697, 'samples': 1079808, 'steps': 2108, 'loss/train': 3.388638496398926} 02/24/2022 04:18:59 - INFO - codeparrot_training - Step 2109: {'lr': 0.0004999936382230597, 'samples': 1080320, 'steps': 2109, 'loss/train': 4.3357648849487305} 02/24/2022 04:19:05 - INFO - codeparrot_training - Step 2110: {'lr': 0.000499993520958256, 'samples': 1080832, 'steps': 2110, 'loss/train': 3.3163487911224365} 02/24/2022 04:19:08 - INFO - codeparrot_training - Step 2111: {'lr': 0.0004999934026225595, 'samples': 1081344, 'steps': 2111, 'loss/train': 3.4119441509246826} 02/24/2022 04:19:15 - INFO - codeparrot_training - Step 2112: {'lr': 0.0004999932832159708, 'samples': 1081856, 'steps': 2112, 'loss/train': 2.135465621948242} 02/24/2022 04:19:18 - INFO - codeparrot_training - Step 2113: {'lr': 0.00049999316273849, 'samples': 1082368, 'steps': 2113, 'loss/train': 4.085206508636475} 02/24/2022 04:19:24 - INFO - codeparrot_training - Step 2114: {'lr': 0.0004999930411901181, 'samples': 1082880, 'steps': 2114, 'loss/train': 1.668155550956726} 02/24/2022 04:19:27 - INFO - codeparrot_training - Step 2115: {'lr': 0.0004999929185708551, 'samples': 1083392, 'steps': 2115, 'loss/train': 3.799992322921753} 02/24/2022 04:19:33 - INFO - codeparrot_training - Step 2116: {'lr': 0.000499992794880702, 'samples': 1083904, 'steps': 2116, 'loss/train': 4.510910511016846} 02/24/2022 04:19:36 - INFO - codeparrot_training - Step 2117: {'lr': 0.0004999926701196592, 'samples': 1084416, 'steps': 2117, 'loss/train': 2.6093909740448} 02/24/2022 04:19:42 - INFO - codeparrot_training - Step 2118: {'lr': 0.0004999925442877271, 'samples': 1084928, 'steps': 2118, 'loss/train': 5.3163018226623535} 02/24/2022 04:19:45 - INFO - codeparrot_training - Step 2119: {'lr': 0.0004999924173849063, 'samples': 1085440, 'steps': 2119, 'loss/train': 3.2637908458709717} 02/24/2022 04:19:51 - INFO - codeparrot_training - Step 2120: {'lr': 0.0004999922894111975, 'samples': 1085952, 'steps': 2120, 'loss/train': 0.513687014579773} 02/24/2022 04:19:54 - INFO - codeparrot_training - Step 2121: {'lr': 0.000499992160366601, 'samples': 1086464, 'steps': 2121, 'loss/train': 2.543294906616211} 02/24/2022 04:20:01 - INFO - codeparrot_training - Step 2122: {'lr': 0.0004999920302511175, 'samples': 1086976, 'steps': 2122, 'loss/train': 3.3987691402435303} 02/24/2022 04:20:04 - INFO - codeparrot_training - Step 2123: {'lr': 0.0004999918990647474, 'samples': 1087488, 'steps': 2123, 'loss/train': 4.51619291305542} 02/24/2022 04:20:09 - INFO - codeparrot_training - Step 2124: {'lr': 0.0004999917668074915, 'samples': 1088000, 'steps': 2124, 'loss/train': 1.8270645141601562} 02/24/2022 04:20:13 - INFO - codeparrot_training - Step 2125: {'lr': 0.0004999916334793503, 'samples': 1088512, 'steps': 2125, 'loss/train': 4.188164234161377} 02/24/2022 04:20:18 - INFO - codeparrot_training - Step 2126: {'lr': 0.0004999914990803242, 'samples': 1089024, 'steps': 2126, 'loss/train': 3.254211664199829} 02/24/2022 04:20:22 - INFO - codeparrot_training - Step 2127: {'lr': 0.000499991363610414, 'samples': 1089536, 'steps': 2127, 'loss/train': 3.083686351776123} 02/24/2022 04:20:27 - INFO - codeparrot_training - Step 2128: {'lr': 0.0004999912270696202, 'samples': 1090048, 'steps': 2128, 'loss/train': 4.00949239730835} 02/24/2022 04:20:31 - INFO - codeparrot_training - Step 2129: {'lr': 0.0004999910894579432, 'samples': 1090560, 'steps': 2129, 'loss/train': 2.9461610317230225} 02/24/2022 04:20:36 - INFO - codeparrot_training - Step 2130: {'lr': 0.000499990950775384, 'samples': 1091072, 'steps': 2130, 'loss/train': 3.2630045413970947} 02/24/2022 04:20:40 - INFO - codeparrot_training - Step 2131: {'lr': 0.0004999908110219428, 'samples': 1091584, 'steps': 2131, 'loss/train': 3.818962812423706} 02/24/2022 04:20:46 - INFO - codeparrot_training - Step 2132: {'lr': 0.0004999906701976203, 'samples': 1092096, 'steps': 2132, 'loss/train': 2.9355931282043457} 02/24/2022 04:20:49 - INFO - codeparrot_training - Step 2133: {'lr': 0.0004999905283024172, 'samples': 1092608, 'steps': 2133, 'loss/train': 4.199846267700195} 02/24/2022 04:20:55 - INFO - codeparrot_training - Step 2134: {'lr': 0.0004999903853363341, 'samples': 1093120, 'steps': 2134, 'loss/train': 4.1997551918029785} 02/24/2022 04:20:59 - INFO - codeparrot_training - Step 2135: {'lr': 0.0004999902412993715, 'samples': 1093632, 'steps': 2135, 'loss/train': 0.6705176830291748} 02/24/2022 04:21:04 - INFO - codeparrot_training - Step 2136: {'lr': 0.0004999900961915302, 'samples': 1094144, 'steps': 2136, 'loss/train': 3.620927572250366} 02/24/2022 04:21:08 - INFO - codeparrot_training - Step 2137: {'lr': 0.0004999899500128107, 'samples': 1094656, 'steps': 2137, 'loss/train': 5.4724955558776855} 02/24/2022 04:21:15 - INFO - codeparrot_training - Step 2138: {'lr': 0.0004999898027632135, 'samples': 1095168, 'steps': 2138, 'loss/train': 3.5491676330566406} 02/24/2022 04:21:19 - INFO - codeparrot_training - Step 2139: {'lr': 0.0004999896544427394, 'samples': 1095680, 'steps': 2139, 'loss/train': 2.685971736907959} 02/24/2022 04:21:24 - INFO - codeparrot_training - Step 2140: {'lr': 0.0004999895050513891, 'samples': 1096192, 'steps': 2140, 'loss/train': 3.04113507270813} 02/24/2022 04:21:28 - INFO - codeparrot_training - Step 2141: {'lr': 0.0004999893545891631, 'samples': 1096704, 'steps': 2141, 'loss/train': 3.4661216735839844} 02/24/2022 04:21:33 - INFO - codeparrot_training - Step 2142: {'lr': 0.000499989203056062, 'samples': 1097216, 'steps': 2142, 'loss/train': 4.2213544845581055} 02/24/2022 04:21:37 - INFO - codeparrot_training - Step 2143: {'lr': 0.0004999890504520866, 'samples': 1097728, 'steps': 2143, 'loss/train': 3.146622896194458} 02/24/2022 04:21:42 - INFO - codeparrot_training - Step 2144: {'lr': 0.0004999888967772375, 'samples': 1098240, 'steps': 2144, 'loss/train': 2.617624044418335} 02/24/2022 04:21:46 - INFO - codeparrot_training - Step 2145: {'lr': 0.0004999887420315153, 'samples': 1098752, 'steps': 2145, 'loss/train': 2.923344612121582} 02/24/2022 04:21:51 - INFO - codeparrot_training - Step 2146: {'lr': 0.0004999885862149207, 'samples': 1099264, 'steps': 2146, 'loss/train': 3.128354549407959} 02/24/2022 04:21:55 - INFO - codeparrot_training - Step 2147: {'lr': 0.0004999884293274545, 'samples': 1099776, 'steps': 2147, 'loss/train': 2.230713367462158} 02/24/2022 04:22:02 - INFO - codeparrot_training - Step 2148: {'lr': 0.0004999882713691171, 'samples': 1100288, 'steps': 2148, 'loss/train': 2.229236125946045} 02/24/2022 04:22:05 - INFO - codeparrot_training - Step 2149: {'lr': 0.0004999881123399093, 'samples': 1100800, 'steps': 2149, 'loss/train': 3.4961557388305664} 02/24/2022 04:22:11 - INFO - codeparrot_training - Step 2150: {'lr': 0.000499987952239832, 'samples': 1101312, 'steps': 2150, 'loss/train': 4.360259532928467} 02/24/2022 04:22:16 - INFO - codeparrot_training - Step 2151: {'lr': 0.0004999877910688856, 'samples': 1101824, 'steps': 2151, 'loss/train': 3.7019996643066406} 02/24/2022 04:22:20 - INFO - codeparrot_training - Step 2152: {'lr': 0.0004999876288270708, 'samples': 1102336, 'steps': 2152, 'loss/train': 1.0514161586761475} 02/24/2022 04:22:25 - INFO - codeparrot_training - Step 2153: {'lr': 0.0004999874655143886, 'samples': 1102848, 'steps': 2153, 'loss/train': 3.8178577423095703} 02/24/2022 04:22:29 - INFO - codeparrot_training - Step 2154: {'lr': 0.0004999873011308393, 'samples': 1103360, 'steps': 2154, 'loss/train': 4.491191387176514} 02/24/2022 04:22:34 - INFO - codeparrot_training - Step 2155: {'lr': 0.0004999871356764238, 'samples': 1103872, 'steps': 2155, 'loss/train': 3.949737787246704} 02/24/2022 04:22:38 - INFO - codeparrot_training - Step 2156: {'lr': 0.0004999869691511428, 'samples': 1104384, 'steps': 2156, 'loss/train': 4.077169418334961} 02/24/2022 04:22:45 - INFO - codeparrot_training - Step 2157: {'lr': 0.000499986801554997, 'samples': 1104896, 'steps': 2157, 'loss/train': 3.050219774246216} 02/24/2022 04:22:49 - INFO - codeparrot_training - Step 2158: {'lr': 0.0004999866328879871, 'samples': 1105408, 'steps': 2158, 'loss/train': 5.032126426696777} 02/24/2022 04:22:54 - INFO - codeparrot_training - Step 2159: {'lr': 0.0004999864631501139, 'samples': 1105920, 'steps': 2159, 'loss/train': 2.369687795639038} 02/24/2022 04:22:58 - INFO - codeparrot_training - Step 2160: {'lr': 0.000499986292341378, 'samples': 1106432, 'steps': 2160, 'loss/train': 3.0985145568847656} 02/24/2022 04:23:03 - INFO - codeparrot_training - Step 2161: {'lr': 0.0004999861204617803, 'samples': 1106944, 'steps': 2161, 'loss/train': 2.0317800045013428} 02/24/2022 04:23:07 - INFO - codeparrot_training - Step 2162: {'lr': 0.0004999859475113213, 'samples': 1107456, 'steps': 2162, 'loss/train': 1.1256765127182007} 02/24/2022 04:23:12 - INFO - codeparrot_training - Step 2163: {'lr': 0.0004999857734900021, 'samples': 1107968, 'steps': 2163, 'loss/train': 3.6377909183502197} 02/24/2022 04:23:16 - INFO - codeparrot_training - Step 2164: {'lr': 0.000499985598397823, 'samples': 1108480, 'steps': 2164, 'loss/train': 3.1779940128326416} 02/24/2022 04:23:21 - INFO - codeparrot_training - Step 2165: {'lr': 0.0004999854222347851, 'samples': 1108992, 'steps': 2165, 'loss/train': 4.294700622558594} 02/24/2022 04:23:25 - INFO - codeparrot_training - Step 2166: {'lr': 0.000499985245000889, 'samples': 1109504, 'steps': 2166, 'loss/train': 3.5069031715393066} 02/24/2022 04:23:30 - INFO - codeparrot_training - Step 2167: {'lr': 0.0004999850666961355, 'samples': 1110016, 'steps': 2167, 'loss/train': 2.9456210136413574} 02/24/2022 04:23:34 - INFO - codeparrot_training - Step 2168: {'lr': 0.0004999848873205254, 'samples': 1110528, 'steps': 2168, 'loss/train': 3.3173673152923584} 02/24/2022 04:23:37 - INFO - codeparrot_training - Step 2169: {'lr': 0.0004999847068740593, 'samples': 1111040, 'steps': 2169, 'loss/train': 1.4148213863372803} 02/24/2022 04:23:45 - INFO - codeparrot_training - Step 2170: {'lr': 0.0004999845253567382, 'samples': 1111552, 'steps': 2170, 'loss/train': 2.3650436401367188} 02/24/2022 04:23:48 - INFO - codeparrot_training - Step 2171: {'lr': 0.0004999843427685627, 'samples': 1112064, 'steps': 2171, 'loss/train': 3.2185676097869873} 02/24/2022 04:23:54 - INFO - codeparrot_training - Step 2172: {'lr': 0.0004999841591095337, 'samples': 1112576, 'steps': 2172, 'loss/train': 2.8058295249938965} 02/24/2022 04:23:57 - INFO - codeparrot_training - Step 2173: {'lr': 0.0004999839743796519, 'samples': 1113088, 'steps': 2173, 'loss/train': 1.9911302328109741} 02/24/2022 04:24:03 - INFO - codeparrot_training - Step 2174: {'lr': 0.0004999837885789182, 'samples': 1113600, 'steps': 2174, 'loss/train': 3.1840808391571045} 02/24/2022 04:24:06 - INFO - codeparrot_training - Step 2175: {'lr': 0.0004999836017073332, 'samples': 1114112, 'steps': 2175, 'loss/train': 3.3054006099700928} 02/24/2022 04:24:12 - INFO - codeparrot_training - Step 2176: {'lr': 0.000499983413764898, 'samples': 1114624, 'steps': 2176, 'loss/train': 3.170403480529785} 02/24/2022 04:24:17 - INFO - codeparrot_training - Step 2177: {'lr': 0.0004999832247516132, 'samples': 1115136, 'steps': 2177, 'loss/train': 3.2368595600128174} 02/24/2022 04:24:20 - INFO - codeparrot_training - Step 2178: {'lr': 0.0004999830346674796, 'samples': 1115648, 'steps': 2178, 'loss/train': 3.468134880065918} 02/24/2022 04:24:28 - INFO - codeparrot_training - Step 2179: {'lr': 0.000499982843512498, 'samples': 1116160, 'steps': 2179, 'loss/train': 3.5260887145996094} 02/24/2022 04:24:31 - INFO - codeparrot_training - Step 2180: {'lr': 0.0004999826512866693, 'samples': 1116672, 'steps': 2180, 'loss/train': 3.104518413543701} 02/24/2022 04:24:37 - INFO - codeparrot_training - Step 2181: {'lr': 0.0004999824579899944, 'samples': 1117184, 'steps': 2181, 'loss/train': 2.4731578826904297} 02/24/2022 04:24:40 - INFO - codeparrot_training - Step 2182: {'lr': 0.000499982263622474, 'samples': 1117696, 'steps': 2182, 'loss/train': 3.185030698776245} 02/24/2022 04:24:44 - INFO - codeparrot_training - Step 2183: {'lr': 0.0004999820681841088, 'samples': 1118208, 'steps': 2183, 'loss/train': 3.0085761547088623} 02/24/2022 04:24:49 - INFO - codeparrot_training - Step 2184: {'lr': 0.0004999818716748999, 'samples': 1118720, 'steps': 2184, 'loss/train': 3.471605062484741} 02/24/2022 04:24:55 - INFO - codeparrot_training - Step 2185: {'lr': 0.0004999816740948481, 'samples': 1119232, 'steps': 2185, 'loss/train': 3.210179328918457} 02/24/2022 04:24:58 - INFO - codeparrot_training - Step 2186: {'lr': 0.0004999814754439542, 'samples': 1119744, 'steps': 2186, 'loss/train': 4.003715515136719} 02/24/2022 04:25:04 - INFO - codeparrot_training - Step 2187: {'lr': 0.000499981275722219, 'samples': 1120256, 'steps': 2187, 'loss/train': 4.127617835998535} 02/24/2022 04:25:07 - INFO - codeparrot_training - Step 2188: {'lr': 0.0004999810749296434, 'samples': 1120768, 'steps': 2188, 'loss/train': 4.164938926696777} 02/24/2022 04:25:13 - INFO - codeparrot_training - Step 2189: {'lr': 0.0004999808730662282, 'samples': 1121280, 'steps': 2189, 'loss/train': 2.5026516914367676} 02/24/2022 04:25:16 - INFO - codeparrot_training - Step 2190: {'lr': 0.0004999806701319743, 'samples': 1121792, 'steps': 2190, 'loss/train': 2.7306406497955322} 02/24/2022 04:25:22 - INFO - codeparrot_training - Step 2191: {'lr': 0.0004999804661268827, 'samples': 1122304, 'steps': 2191, 'loss/train': 2.1373531818389893} 02/24/2022 04:25:25 - INFO - codeparrot_training - Step 2192: {'lr': 0.0004999802610509541, 'samples': 1122816, 'steps': 2192, 'loss/train': 3.4641823768615723} 02/24/2022 04:25:30 - INFO - codeparrot_training - Step 2193: {'lr': 0.0004999800549041894, 'samples': 1123328, 'steps': 2193, 'loss/train': 4.5540995597839355} 02/24/2022 04:25:34 - INFO - codeparrot_training - Step 2194: {'lr': 0.0004999798476865895, 'samples': 1123840, 'steps': 2194, 'loss/train': 3.797067165374756} 02/24/2022 04:25:41 - INFO - codeparrot_training - Step 2195: {'lr': 0.0004999796393981554, 'samples': 1124352, 'steps': 2195, 'loss/train': 3.8183090686798096} 02/24/2022 04:25:45 - INFO - codeparrot_training - Step 2196: {'lr': 0.0004999794300388879, 'samples': 1124864, 'steps': 2196, 'loss/train': 4.2558369636535645} 02/24/2022 04:25:50 - INFO - codeparrot_training - Step 2197: {'lr': 0.0004999792196087879, 'samples': 1125376, 'steps': 2197, 'loss/train': 4.3991289138793945} 02/24/2022 04:25:54 - INFO - codeparrot_training - Step 2198: {'lr': 0.0004999790081078562, 'samples': 1125888, 'steps': 2198, 'loss/train': 2.6684980392456055} 02/24/2022 04:26:00 - INFO - codeparrot_training - Step 2199: {'lr': 0.0004999787955360939, 'samples': 1126400, 'steps': 2199, 'loss/train': 6.624790668487549} 02/24/2022 04:26:03 - INFO - codeparrot_training - Step 2200: {'lr': 0.0004999785818935018, 'samples': 1126912, 'steps': 2200, 'loss/train': 2.3392221927642822} 02/24/2022 04:26:09 - INFO - codeparrot_training - Step 2201: {'lr': 0.0004999783671800808, 'samples': 1127424, 'steps': 2201, 'loss/train': 3.1122958660125732} 02/24/2022 04:26:12 - INFO - codeparrot_training - Step 2202: {'lr': 0.0004999781513958318, 'samples': 1127936, 'steps': 2202, 'loss/train': 3.565094470977783} 02/24/2022 04:26:17 - INFO - codeparrot_training - Step 2203: {'lr': 0.000499977934540756, 'samples': 1128448, 'steps': 2203, 'loss/train': 3.473228931427002} 02/24/2022 04:26:21 - INFO - codeparrot_training - Step 2204: {'lr': 0.0004999777166148539, 'samples': 1128960, 'steps': 2204, 'loss/train': 2.5281777381896973} 02/24/2022 04:26:28 - INFO - codeparrot_training - Step 2205: {'lr': 0.0004999774976181267, 'samples': 1129472, 'steps': 2205, 'loss/train': 3.73065185546875} 02/24/2022 04:26:32 - INFO - codeparrot_training - Step 2206: {'lr': 0.0004999772775505753, 'samples': 1129984, 'steps': 2206, 'loss/train': 3.899561882019043} 02/24/2022 04:26:37 - INFO - codeparrot_training - Step 2207: {'lr': 0.0004999770564122005, 'samples': 1130496, 'steps': 2207, 'loss/train': 3.804422616958618} 02/24/2022 04:26:41 - INFO - codeparrot_training - Step 2208: {'lr': 0.0004999768342030035, 'samples': 1131008, 'steps': 2208, 'loss/train': 4.123617172241211} 02/24/2022 04:26:46 - INFO - codeparrot_training - Step 2209: {'lr': 0.0004999766109229851, 'samples': 1131520, 'steps': 2209, 'loss/train': 3.929154634475708} 02/24/2022 04:26:49 - INFO - codeparrot_training - Step 2210: {'lr': 0.0004999763865721463, 'samples': 1132032, 'steps': 2210, 'loss/train': 4.022369861602783} 02/24/2022 04:26:55 - INFO - codeparrot_training - Step 2211: {'lr': 0.000499976161150488, 'samples': 1132544, 'steps': 2211, 'loss/train': 3.633319139480591} 02/24/2022 04:26:58 - INFO - codeparrot_training - Step 2212: {'lr': 0.0004999759346580111, 'samples': 1133056, 'steps': 2212, 'loss/train': 4.5179572105407715} 02/24/2022 04:27:04 - INFO - codeparrot_training - Step 2213: {'lr': 0.0004999757070947168, 'samples': 1133568, 'steps': 2213, 'loss/train': 3.564769744873047} 02/24/2022 04:27:07 - INFO - codeparrot_training - Step 2214: {'lr': 0.0004999754784606058, 'samples': 1134080, 'steps': 2214, 'loss/train': 3.7344629764556885} 02/24/2022 04:27:15 - INFO - codeparrot_training - Step 2215: {'lr': 0.0004999752487556794, 'samples': 1134592, 'steps': 2215, 'loss/train': 2.515857458114624} 02/24/2022 04:27:18 - INFO - codeparrot_training - Step 2216: {'lr': 0.0004999750179799383, 'samples': 1135104, 'steps': 2216, 'loss/train': 2.132930278778076} 02/24/2022 04:27:24 - INFO - codeparrot_training - Step 2217: {'lr': 0.0004999747861333838, 'samples': 1135616, 'steps': 2217, 'loss/train': 2.356292963027954} 02/24/2022 04:27:27 - INFO - codeparrot_training - Step 2218: {'lr': 0.0004999745532160164, 'samples': 1136128, 'steps': 2218, 'loss/train': 3.3841753005981445} 02/24/2022 04:27:33 - INFO - codeparrot_training - Step 2219: {'lr': 0.0004999743192278377, 'samples': 1136640, 'steps': 2219, 'loss/train': 1.976936936378479} 02/24/2022 04:27:36 - INFO - codeparrot_training - Step 2220: {'lr': 0.0004999740841688481, 'samples': 1137152, 'steps': 2220, 'loss/train': 2.2595696449279785} 02/24/2022 04:27:42 - INFO - codeparrot_training - Step 2221: {'lr': 0.000499973848039049, 'samples': 1137664, 'steps': 2221, 'loss/train': 3.168776750564575} 02/24/2022 04:27:45 - INFO - codeparrot_training - Step 2222: {'lr': 0.0004999736108384414, 'samples': 1138176, 'steps': 2222, 'loss/train': 3.64705491065979} 02/24/2022 04:27:51 - INFO - codeparrot_training - Step 2223: {'lr': 0.0004999733725670261, 'samples': 1138688, 'steps': 2223, 'loss/train': 1.1466491222381592} 02/24/2022 04:27:54 - INFO - codeparrot_training - Step 2224: {'lr': 0.0004999731332248044, 'samples': 1139200, 'steps': 2224, 'loss/train': 2.6170737743377686} 02/24/2022 04:28:01 - INFO - codeparrot_training - Step 2225: {'lr': 0.0004999728928117771, 'samples': 1139712, 'steps': 2225, 'loss/train': 2.871457576751709} 02/24/2022 04:28:05 - INFO - codeparrot_training - Step 2226: {'lr': 0.0004999726513279452, 'samples': 1140224, 'steps': 2226, 'loss/train': 4.05280065536499} 02/24/2022 04:28:10 - INFO - codeparrot_training - Step 2227: {'lr': 0.0004999724087733099, 'samples': 1140736, 'steps': 2227, 'loss/train': 3.5515193939208984} 02/24/2022 04:28:14 - INFO - codeparrot_training - Step 2228: {'lr': 0.0004999721651478723, 'samples': 1141248, 'steps': 2228, 'loss/train': 2.5241708755493164} 02/24/2022 04:28:19 - INFO - codeparrot_training - Step 2229: {'lr': 0.0004999719204516332, 'samples': 1141760, 'steps': 2229, 'loss/train': 3.4813103675842285} 02/24/2022 04:28:22 - INFO - codeparrot_training - Step 2230: {'lr': 0.0004999716746845937, 'samples': 1142272, 'steps': 2230, 'loss/train': 4.263186454772949} 02/24/2022 04:28:28 - INFO - codeparrot_training - Step 2231: {'lr': 0.0004999714278467551, 'samples': 1142784, 'steps': 2231, 'loss/train': 3.2247517108917236} 02/24/2022 04:28:33 - INFO - codeparrot_training - Step 2232: {'lr': 0.0004999711799381181, 'samples': 1143296, 'steps': 2232, 'loss/train': 4.176067352294922} 02/24/2022 04:28:37 - INFO - codeparrot_training - Step 2233: {'lr': 0.000499970930958684, 'samples': 1143808, 'steps': 2233, 'loss/train': 4.0900115966796875} 02/24/2022 04:28:44 - INFO - codeparrot_training - Step 2234: {'lr': 0.0004999706809084538, 'samples': 1144320, 'steps': 2234, 'loss/train': 2.9840047359466553} 02/24/2022 04:28:48 - INFO - codeparrot_training - Step 2235: {'lr': 0.0004999704297874287, 'samples': 1144832, 'steps': 2235, 'loss/train': 4.0390167236328125} 02/24/2022 04:28:53 - INFO - codeparrot_training - Step 2236: {'lr': 0.0004999701775956095, 'samples': 1145344, 'steps': 2236, 'loss/train': 2.902132749557495} 02/24/2022 04:28:57 - INFO - codeparrot_training - Step 2237: {'lr': 0.0004999699243329975, 'samples': 1145856, 'steps': 2237, 'loss/train': 3.204497814178467} 02/24/2022 04:29:02 - INFO - codeparrot_training - Step 2238: {'lr': 0.0004999696699995937, 'samples': 1146368, 'steps': 2238, 'loss/train': 0.8694908618927002} 02/24/2022 04:29:06 - INFO - codeparrot_training - Step 2239: {'lr': 0.0004999694145953992, 'samples': 1146880, 'steps': 2239, 'loss/train': 3.2490906715393066} 02/24/2022 04:29:11 - INFO - codeparrot_training - Step 2240: {'lr': 0.0004999691581204152, 'samples': 1147392, 'steps': 2240, 'loss/train': 3.2772016525268555} 02/24/2022 04:29:15 - INFO - codeparrot_training - Step 2241: {'lr': 0.0004999689005746426, 'samples': 1147904, 'steps': 2241, 'loss/train': 2.77553129196167} 02/24/2022 04:29:20 - INFO - codeparrot_training - Step 2242: {'lr': 0.0004999686419580827, 'samples': 1148416, 'steps': 2242, 'loss/train': 2.762347459793091} 02/24/2022 04:29:24 - INFO - codeparrot_training - Step 2243: {'lr': 0.0004999683822707364, 'samples': 1148928, 'steps': 2243, 'loss/train': 4.271193504333496} 02/24/2022 04:29:27 - INFO - codeparrot_training - Step 2244: {'lr': 0.0004999681215126049, 'samples': 1149440, 'steps': 2244, 'loss/train': 4.404847621917725} 02/24/2022 04:29:33 - INFO - codeparrot_training - Step 2245: {'lr': 0.0004999678596836894, 'samples': 1149952, 'steps': 2245, 'loss/train': 2.916435480117798} 02/24/2022 04:29:36 - INFO - codeparrot_training - Step 2246: {'lr': 0.000499967596783991, 'samples': 1150464, 'steps': 2246, 'loss/train': 4.588435649871826} 02/24/2022 04:29:42 - INFO - codeparrot_training - Step 2247: {'lr': 0.0004999673328135107, 'samples': 1150976, 'steps': 2247, 'loss/train': 4.4290642738342285} 02/24/2022 04:29:45 - INFO - codeparrot_training - Step 2248: {'lr': 0.0004999670677722498, 'samples': 1151488, 'steps': 2248, 'loss/train': 3.213757038116455} 02/24/2022 04:29:51 - INFO - codeparrot_training - Step 2249: {'lr': 0.0004999668016602094, 'samples': 1152000, 'steps': 2249, 'loss/train': 3.557638645172119} 02/24/2022 04:29:54 - INFO - codeparrot_training - Step 2250: {'lr': 0.0004999665344773905, 'samples': 1152512, 'steps': 2250, 'loss/train': 2.462035894393921} 02/24/2022 04:30:02 - INFO - codeparrot_training - Step 2251: {'lr': 0.0004999662662237943, 'samples': 1153024, 'steps': 2251, 'loss/train': 3.8249166011810303} 02/24/2022 04:30:05 - INFO - codeparrot_training - Step 2252: {'lr': 0.0004999659968994221, 'samples': 1153536, 'steps': 2252, 'loss/train': 3.099893569946289} 02/24/2022 04:30:11 - INFO - codeparrot_training - Step 2253: {'lr': 0.0004999657265042748, 'samples': 1154048, 'steps': 2253, 'loss/train': 3.291548013687134} 02/24/2022 04:30:14 - INFO - codeparrot_training - Step 2254: {'lr': 0.0004999654550383539, 'samples': 1154560, 'steps': 2254, 'loss/train': 2.822593927383423} 02/24/2022 04:30:20 - INFO - codeparrot_training - Step 2255: {'lr': 0.0004999651825016603, 'samples': 1155072, 'steps': 2255, 'loss/train': 4.550164222717285} 02/24/2022 04:30:25 - INFO - codeparrot_training - Step 2256: {'lr': 0.0004999649088941951, 'samples': 1155584, 'steps': 2256, 'loss/train': 3.122103452682495} 02/24/2022 04:30:29 - INFO - codeparrot_training - Step 2257: {'lr': 0.0004999646342159597, 'samples': 1156096, 'steps': 2257, 'loss/train': 2.8963000774383545} 02/24/2022 04:30:34 - INFO - codeparrot_training - Step 2258: {'lr': 0.0004999643584669552, 'samples': 1156608, 'steps': 2258, 'loss/train': 3.3787777423858643} 02/24/2022 04:30:38 - INFO - codeparrot_training - Step 2259: {'lr': 0.0004999640816471827, 'samples': 1157120, 'steps': 2259, 'loss/train': 3.468085289001465} 02/24/2022 04:30:43 - INFO - codeparrot_training - Step 2260: {'lr': 0.0004999638037566436, 'samples': 1157632, 'steps': 2260, 'loss/train': 3.627345323562622} 02/24/2022 04:30:46 - INFO - codeparrot_training - Step 2261: {'lr': 0.0004999635247953387, 'samples': 1158144, 'steps': 2261, 'loss/train': 4.028489112854004} 02/24/2022 04:30:54 - INFO - codeparrot_training - Step 2262: {'lr': 0.0004999632447632696, 'samples': 1158656, 'steps': 2262, 'loss/train': 2.716365337371826} 02/24/2022 04:30:57 - INFO - codeparrot_training - Step 2263: {'lr': 0.0004999629636604372, 'samples': 1159168, 'steps': 2263, 'loss/train': 3.508540630340576} 02/24/2022 04:31:03 - INFO - codeparrot_training - Step 2264: {'lr': 0.0004999626814868429, 'samples': 1159680, 'steps': 2264, 'loss/train': 3.9661898612976074} 02/24/2022 04:31:06 - INFO - codeparrot_training - Step 2265: {'lr': 0.0004999623982424879, 'samples': 1160192, 'steps': 2265, 'loss/train': 2.874356985092163} 02/24/2022 04:31:12 - INFO - codeparrot_training - Step 2266: {'lr': 0.0004999621139273733, 'samples': 1160704, 'steps': 2266, 'loss/train': 4.915088653564453} 02/24/2022 04:31:15 - INFO - codeparrot_training - Step 2267: {'lr': 0.0004999618285415004, 'samples': 1161216, 'steps': 2267, 'loss/train': 3.3333630561828613} 02/24/2022 04:31:21 - INFO - codeparrot_training - Step 2268: {'lr': 0.0004999615420848704, 'samples': 1161728, 'steps': 2268, 'loss/train': 3.9082980155944824} 02/24/2022 04:31:24 - INFO - codeparrot_training - Step 2269: {'lr': 0.0004999612545574845, 'samples': 1162240, 'steps': 2269, 'loss/train': 3.7582783699035645} 02/24/2022 04:31:31 - INFO - codeparrot_training - Step 2270: {'lr': 0.000499960965959344, 'samples': 1162752, 'steps': 2270, 'loss/train': 3.778946876525879} 02/24/2022 04:31:35 - INFO - codeparrot_training - Step 2271: {'lr': 0.0004999606762904501, 'samples': 1163264, 'steps': 2271, 'loss/train': 4.916256904602051} 02/24/2022 04:31:40 - INFO - codeparrot_training - Step 2272: {'lr': 0.000499960385550804, 'samples': 1163776, 'steps': 2272, 'loss/train': 2.7035017013549805} 02/24/2022 04:31:44 - INFO - codeparrot_training - Step 2273: {'lr': 0.000499960093740407, 'samples': 1164288, 'steps': 2273, 'loss/train': 2.951904535293579} 02/24/2022 04:31:49 - INFO - codeparrot_training - Step 2274: {'lr': 0.0004999598008592603, 'samples': 1164800, 'steps': 2274, 'loss/train': 3.863191843032837} 02/24/2022 04:31:53 - INFO - codeparrot_training - Step 2275: {'lr': 0.0004999595069073653, 'samples': 1165312, 'steps': 2275, 'loss/train': 3.578238010406494} 02/24/2022 04:31:58 - INFO - codeparrot_training - Step 2276: {'lr': 0.0004999592118847229, 'samples': 1165824, 'steps': 2276, 'loss/train': 3.3003287315368652} 02/24/2022 04:32:02 - INFO - codeparrot_training - Step 2277: {'lr': 0.0004999589157913348, 'samples': 1166336, 'steps': 2277, 'loss/train': 3.5974996089935303} 02/24/2022 04:32:07 - INFO - codeparrot_training - Step 2278: {'lr': 0.0004999586186272021, 'samples': 1166848, 'steps': 2278, 'loss/train': 3.835594654083252} 02/24/2022 04:32:11 - INFO - codeparrot_training - Step 2279: {'lr': 0.000499958320392326, 'samples': 1167360, 'steps': 2279, 'loss/train': 3.9866204261779785} 02/24/2022 04:32:16 - INFO - codeparrot_training - Step 2280: {'lr': 0.0004999580210867077, 'samples': 1167872, 'steps': 2280, 'loss/train': 2.536888599395752} 02/24/2022 04:32:20 - INFO - codeparrot_training - Step 2281: {'lr': 0.0004999577207103487, 'samples': 1168384, 'steps': 2281, 'loss/train': 4.099573135375977} 02/24/2022 04:32:25 - INFO - codeparrot_training - Step 2282: {'lr': 0.0004999574192632502, 'samples': 1168896, 'steps': 2282, 'loss/train': 2.6548068523406982} 02/24/2022 04:32:29 - INFO - codeparrot_training - Step 2283: {'lr': 0.0004999571167454135, 'samples': 1169408, 'steps': 2283, 'loss/train': 3.8463265895843506} 02/24/2022 04:32:34 - INFO - codeparrot_training - Step 2284: {'lr': 0.0004999568131568399, 'samples': 1169920, 'steps': 2284, 'loss/train': 0.6082117557525635} 02/24/2022 04:32:38 - INFO - codeparrot_training - Step 2285: {'lr': 0.0004999565084975306, 'samples': 1170432, 'steps': 2285, 'loss/train': 3.158193826675415} 02/24/2022 04:32:45 - INFO - codeparrot_training - Step 2286: {'lr': 0.0004999562027674871, 'samples': 1170944, 'steps': 2286, 'loss/train': 4.124403476715088} 02/24/2022 04:32:48 - INFO - codeparrot_training - Step 2287: {'lr': 0.0004999558959667105, 'samples': 1171456, 'steps': 2287, 'loss/train': 1.7018682956695557} 02/24/2022 04:32:54 - INFO - codeparrot_training - Step 2288: {'lr': 0.0004999555880952023, 'samples': 1171968, 'steps': 2288, 'loss/train': 4.156793117523193} 02/24/2022 04:32:57 - INFO - codeparrot_training - Step 2289: {'lr': 0.0004999552791529637, 'samples': 1172480, 'steps': 2289, 'loss/train': 2.108534812927246} 02/24/2022 04:33:03 - INFO - codeparrot_training - Step 2290: {'lr': 0.000499954969139996, 'samples': 1172992, 'steps': 2290, 'loss/train': 3.1388399600982666} 02/24/2022 04:33:06 - INFO - codeparrot_training - Step 2291: {'lr': 0.0004999546580563006, 'samples': 1173504, 'steps': 2291, 'loss/train': 4.162036418914795} 02/24/2022 04:33:12 - INFO - codeparrot_training - Step 2292: {'lr': 0.0004999543459018788, 'samples': 1174016, 'steps': 2292, 'loss/train': 3.205673933029175} 02/24/2022 04:33:15 - INFO - codeparrot_training - Step 2293: {'lr': 0.000499954032676732, 'samples': 1174528, 'steps': 2293, 'loss/train': 4.420656204223633} 02/24/2022 04:33:21 - INFO - codeparrot_training - Step 2294: {'lr': 0.0004999537183808614, 'samples': 1175040, 'steps': 2294, 'loss/train': 2.455386161804199} 02/24/2022 04:33:24 - INFO - codeparrot_training - Step 2295: {'lr': 0.0004999534030142686, 'samples': 1175552, 'steps': 2295, 'loss/train': 0.4072885811328888} 02/24/2022 04:33:32 - INFO - codeparrot_training - Step 2296: {'lr': 0.0004999530865769547, 'samples': 1176064, 'steps': 2296, 'loss/train': 2.254129409790039} 02/24/2022 04:33:35 - INFO - codeparrot_training - Step 2297: {'lr': 0.0004999527690689212, 'samples': 1176576, 'steps': 2297, 'loss/train': 3.1453521251678467} 02/24/2022 04:33:41 - INFO - codeparrot_training - Step 2298: {'lr': 0.0004999524504901694, 'samples': 1177088, 'steps': 2298, 'loss/train': 2.589794635772705} 02/24/2022 04:33:44 - INFO - codeparrot_training - Step 2299: {'lr': 0.0004999521308407006, 'samples': 1177600, 'steps': 2299, 'loss/train': 3.344181537628174} 02/24/2022 04:33:50 - INFO - codeparrot_training - Step 2300: {'lr': 0.0004999518101205162, 'samples': 1178112, 'steps': 2300, 'loss/train': 4.422452926635742} 02/24/2022 04:33:53 - INFO - codeparrot_training - Step 2301: {'lr': 0.0004999514883296176, 'samples': 1178624, 'steps': 2301, 'loss/train': 1.787958025932312} 02/24/2022 04:33:59 - INFO - codeparrot_training - Step 2302: {'lr': 0.0004999511654680064, 'samples': 1179136, 'steps': 2302, 'loss/train': 2.1183021068573} 02/24/2022 04:34:02 - INFO - codeparrot_training - Step 2303: {'lr': 0.0004999508415356836, 'samples': 1179648, 'steps': 2303, 'loss/train': 3.060662269592285} 02/24/2022 04:34:08 - INFO - codeparrot_training - Step 2304: {'lr': 0.0004999505165326509, 'samples': 1180160, 'steps': 2304, 'loss/train': 2.1927576065063477} 02/24/2022 04:34:11 - INFO - codeparrot_training - Step 2305: {'lr': 0.0004999501904589095, 'samples': 1180672, 'steps': 2305, 'loss/train': 2.572749137878418} 02/24/2022 04:34:18 - INFO - codeparrot_training - Step 2306: {'lr': 0.0004999498633144608, 'samples': 1181184, 'steps': 2306, 'loss/train': 2.5980634689331055} 02/24/2022 04:34:21 - INFO - codeparrot_training - Step 2307: {'lr': 0.0004999495350993062, 'samples': 1181696, 'steps': 2307, 'loss/train': 3.2687549591064453} 02/24/2022 04:34:27 - INFO - codeparrot_training - Step 2308: {'lr': 0.0004999492058134473, 'samples': 1182208, 'steps': 2308, 'loss/train': 5.166488170623779} 02/24/2022 04:34:30 - INFO - codeparrot_training - Step 2309: {'lr': 0.0004999488754568853, 'samples': 1182720, 'steps': 2309, 'loss/train': 3.4950520992279053} 02/24/2022 04:34:36 - INFO - codeparrot_training - Step 2310: {'lr': 0.0004999485440296216, 'samples': 1183232, 'steps': 2310, 'loss/train': 2.227299213409424} 02/24/2022 04:34:39 - INFO - codeparrot_training - Step 2311: {'lr': 0.0004999482115316579, 'samples': 1183744, 'steps': 2311, 'loss/train': 2.8919758796691895} 02/24/2022 04:34:45 - INFO - codeparrot_training - Step 2312: {'lr': 0.0004999478779629953, 'samples': 1184256, 'steps': 2312, 'loss/train': 2.2472426891326904} 02/24/2022 04:34:48 - INFO - codeparrot_training - Step 2313: {'lr': 0.0004999475433236354, 'samples': 1184768, 'steps': 2313, 'loss/train': 2.3866384029388428} 02/24/2022 04:34:54 - INFO - codeparrot_training - Step 2314: {'lr': 0.0004999472076135796, 'samples': 1185280, 'steps': 2314, 'loss/train': 3.3593130111694336} 02/24/2022 04:34:59 - INFO - codeparrot_training - Step 2315: {'lr': 0.0004999468708328293, 'samples': 1185792, 'steps': 2315, 'loss/train': 3.8776097297668457} 02/24/2022 04:35:02 - INFO - codeparrot_training - Step 2316: {'lr': 0.0004999465329813859, 'samples': 1186304, 'steps': 2316, 'loss/train': 3.015650749206543} 02/24/2022 04:35:09 - INFO - codeparrot_training - Step 2317: {'lr': 0.000499946194059251, 'samples': 1186816, 'steps': 2317, 'loss/train': 2.145082712173462} 02/24/2022 04:35:13 - INFO - codeparrot_training - Step 2318: {'lr': 0.000499945854066426, 'samples': 1187328, 'steps': 2318, 'loss/train': 2.344414472579956} 02/24/2022 04:35:18 - INFO - codeparrot_training - Step 2319: {'lr': 0.0004999455130029123, 'samples': 1187840, 'steps': 2319, 'loss/train': 3.8513545989990234} 02/24/2022 04:35:22 - INFO - codeparrot_training - Step 2320: {'lr': 0.0004999451708687113, 'samples': 1188352, 'steps': 2320, 'loss/train': 3.5533957481384277} 02/24/2022 04:35:27 - INFO - codeparrot_training - Step 2321: {'lr': 0.0004999448276638247, 'samples': 1188864, 'steps': 2321, 'loss/train': 3.5150415897369385} 02/24/2022 04:35:31 - INFO - codeparrot_training - Step 2322: {'lr': 0.0004999444833882538, 'samples': 1189376, 'steps': 2322, 'loss/train': 3.4148900508880615} 02/24/2022 04:35:36 - INFO - codeparrot_training - Step 2323: {'lr': 0.000499944138042, 'samples': 1189888, 'steps': 2323, 'loss/train': 3.8514246940612793} 02/24/2022 04:35:40 - INFO - codeparrot_training - Step 2324: {'lr': 0.000499943791625065, 'samples': 1190400, 'steps': 2324, 'loss/train': 2.228163480758667} 02/24/2022 04:35:46 - INFO - codeparrot_training - Step 2325: {'lr': 0.0004999434441374501, 'samples': 1190912, 'steps': 2325, 'loss/train': 1.3067525625228882} 02/24/2022 04:35:49 - INFO - codeparrot_training - Step 2326: {'lr': 0.0004999430955791569, 'samples': 1191424, 'steps': 2326, 'loss/train': 0.28261226415634155} 02/24/2022 04:35:55 - INFO - codeparrot_training - Step 2327: {'lr': 0.0004999427459501868, 'samples': 1191936, 'steps': 2327, 'loss/train': 6.6103057861328125} 02/24/2022 04:35:58 - INFO - codeparrot_training - Step 2328: {'lr': 0.0004999423952505414, 'samples': 1192448, 'steps': 2328, 'loss/train': 2.38521671295166} 02/24/2022 04:36:04 - INFO - codeparrot_training - Step 2329: {'lr': 0.000499942043480222, 'samples': 1192960, 'steps': 2329, 'loss/train': 4.625878810882568} 02/24/2022 04:36:07 - INFO - codeparrot_training - Step 2330: {'lr': 0.0004999416906392303, 'samples': 1193472, 'steps': 2330, 'loss/train': 1.787308931350708} 02/24/2022 04:36:13 - INFO - codeparrot_training - Step 2331: {'lr': 0.0004999413367275678, 'samples': 1193984, 'steps': 2331, 'loss/train': 3.660783529281616} 02/24/2022 04:36:17 - INFO - codeparrot_training - Step 2332: {'lr': 0.000499940981745236, 'samples': 1194496, 'steps': 2332, 'loss/train': 1.4107046127319336} 02/24/2022 04:36:23 - INFO - codeparrot_training - Step 2333: {'lr': 0.0004999406256922365, 'samples': 1195008, 'steps': 2333, 'loss/train': 3.45271635055542} 02/24/2022 04:36:27 - INFO - codeparrot_training - Step 2334: {'lr': 0.0004999402685685705, 'samples': 1195520, 'steps': 2334, 'loss/train': 3.881763458251953} 02/24/2022 04:36:30 - INFO - codeparrot_training - Step 2335: {'lr': 0.0004999399103742399, 'samples': 1196032, 'steps': 2335, 'loss/train': 2.9919068813323975} 02/24/2022 04:36:33 - INFO - codeparrot_training - Step 2336: {'lr': 0.000499939551109246, 'samples': 1196544, 'steps': 2336, 'loss/train': 4.510589599609375} 02/24/2022 04:36:39 - INFO - codeparrot_training - Step 2337: {'lr': 0.0004999391907735905, 'samples': 1197056, 'steps': 2337, 'loss/train': 3.028735637664795} 02/24/2022 04:36:42 - INFO - codeparrot_training - Step 2338: {'lr': 0.0004999388293672748, 'samples': 1197568, 'steps': 2338, 'loss/train': 3.2945544719696045} 02/24/2022 04:36:48 - INFO - codeparrot_training - Step 2339: {'lr': 0.0004999384668903006, 'samples': 1198080, 'steps': 2339, 'loss/train': 2.7194881439208984} 02/24/2022 04:36:52 - INFO - codeparrot_training - Step 2340: {'lr': 0.0004999381033426693, 'samples': 1198592, 'steps': 2340, 'loss/train': 2.4050533771514893} 02/24/2022 04:36:57 - INFO - codeparrot_training - Step 2341: {'lr': 0.0004999377387243827, 'samples': 1199104, 'steps': 2341, 'loss/train': 4.339027404785156} 02/24/2022 04:37:00 - INFO - codeparrot_training - Step 2342: {'lr': 0.0004999373730354419, 'samples': 1199616, 'steps': 2342, 'loss/train': 2.8714916706085205} 02/24/2022 04:37:07 - INFO - codeparrot_training - Step 2343: {'lr': 0.0004999370062758491, 'samples': 1200128, 'steps': 2343, 'loss/train': 3.7314019203186035} 02/24/2022 04:37:12 - INFO - codeparrot_training - Step 2344: {'lr': 0.0004999366384456052, 'samples': 1200640, 'steps': 2344, 'loss/train': 4.306217193603516} 02/24/2022 04:37:15 - INFO - codeparrot_training - Step 2345: {'lr': 0.0004999362695447123, 'samples': 1201152, 'steps': 2345, 'loss/train': 3.8925302028656006} 02/24/2022 04:37:21 - INFO - codeparrot_training - Step 2346: {'lr': 0.0004999358995731718, 'samples': 1201664, 'steps': 2346, 'loss/train': 3.634169816970825} 02/24/2022 04:37:24 - INFO - codeparrot_training - Step 2347: {'lr': 0.0004999355285309851, 'samples': 1202176, 'steps': 2347, 'loss/train': 3.4635636806488037} 02/24/2022 04:37:30 - INFO - codeparrot_training - Step 2348: {'lr': 0.0004999351564181541, 'samples': 1202688, 'steps': 2348, 'loss/train': 4.446084976196289} 02/24/2022 04:37:33 - INFO - codeparrot_training - Step 2349: {'lr': 0.0004999347832346802, 'samples': 1203200, 'steps': 2349, 'loss/train': 4.977412700653076} 02/24/2022 04:37:39 - INFO - codeparrot_training - Step 2350: {'lr': 0.0004999344089805651, 'samples': 1203712, 'steps': 2350, 'loss/train': 2.824918746948242} 02/24/2022 04:37:42 - INFO - codeparrot_training - Step 2351: {'lr': 0.0004999340336558104, 'samples': 1204224, 'steps': 2351, 'loss/train': 1.8706772327423096} 02/24/2022 04:37:48 - INFO - codeparrot_training - Step 2352: {'lr': 0.0004999336572604175, 'samples': 1204736, 'steps': 2352, 'loss/train': 3.604421854019165} 02/24/2022 04:37:52 - INFO - codeparrot_training - Step 2353: {'lr': 0.0004999332797943883, 'samples': 1205248, 'steps': 2353, 'loss/train': 3.882422924041748} 02/24/2022 04:37:57 - INFO - codeparrot_training - Step 2354: {'lr': 0.0004999329012577243, 'samples': 1205760, 'steps': 2354, 'loss/train': 3.8811521530151367} 02/24/2022 04:38:01 - INFO - codeparrot_training - Step 2355: {'lr': 0.000499932521650427, 'samples': 1206272, 'steps': 2355, 'loss/train': 3.0197741985321045} 02/24/2022 04:38:06 - INFO - codeparrot_training - Step 2356: {'lr': 0.0004999321409724982, 'samples': 1206784, 'steps': 2356, 'loss/train': 0.8928283452987671} 02/24/2022 04:38:10 - INFO - codeparrot_training - Step 2357: {'lr': 0.0004999317592239395, 'samples': 1207296, 'steps': 2357, 'loss/train': 3.751070261001587} 02/24/2022 04:38:15 - INFO - codeparrot_training - Step 2358: {'lr': 0.0004999313764047525, 'samples': 1207808, 'steps': 2358, 'loss/train': 4.281635761260986} 02/24/2022 04:38:19 - INFO - codeparrot_training - Step 2359: {'lr': 0.0004999309925149388, 'samples': 1208320, 'steps': 2359, 'loss/train': 3.909730911254883} 02/24/2022 04:38:24 - INFO - codeparrot_training - Step 2360: {'lr': 0.0004999306075545002, 'samples': 1208832, 'steps': 2360, 'loss/train': 5.49069881439209} 02/24/2022 04:38:28 - INFO - codeparrot_training - Step 2361: {'lr': 0.0004999302215234381, 'samples': 1209344, 'steps': 2361, 'loss/train': 3.3348541259765625} 02/24/2022 04:38:33 - INFO - codeparrot_training - Step 2362: {'lr': 0.0004999298344217543, 'samples': 1209856, 'steps': 2362, 'loss/train': 3.580683946609497} 02/24/2022 04:38:37 - INFO - codeparrot_training - Step 2363: {'lr': 0.0004999294462494506, 'samples': 1210368, 'steps': 2363, 'loss/train': 2.8717422485351562} 02/24/2022 04:38:43 - INFO - codeparrot_training - Step 2364: {'lr': 0.0004999290570065284, 'samples': 1210880, 'steps': 2364, 'loss/train': 0.5460531711578369} 02/24/2022 04:38:46 - INFO - codeparrot_training - Step 2365: {'lr': 0.0004999286666929895, 'samples': 1211392, 'steps': 2365, 'loss/train': 3.815709114074707} 02/24/2022 04:38:51 - INFO - codeparrot_training - Step 2366: {'lr': 0.0004999282753088356, 'samples': 1211904, 'steps': 2366, 'loss/train': 3.463447093963623} 02/24/2022 04:38:55 - INFO - codeparrot_training - Step 2367: {'lr': 0.0004999278828540682, 'samples': 1212416, 'steps': 2367, 'loss/train': 3.520075559616089} 02/24/2022 04:39:00 - INFO - codeparrot_training - Step 2368: {'lr': 0.0004999274893286893, 'samples': 1212928, 'steps': 2368, 'loss/train': 3.047358751296997} 02/24/2022 04:39:04 - INFO - codeparrot_training - Step 2369: {'lr': 0.0004999270947327003, 'samples': 1213440, 'steps': 2369, 'loss/train': 3.0929086208343506} 02/24/2022 04:39:09 - INFO - codeparrot_training - Step 2370: {'lr': 0.0004999266990661029, 'samples': 1213952, 'steps': 2370, 'loss/train': 3.5242362022399902} 02/24/2022 04:39:13 - INFO - codeparrot_training - Step 2371: {'lr': 0.0004999263023288989, 'samples': 1214464, 'steps': 2371, 'loss/train': 3.582456350326538} 02/24/2022 04:39:18 - INFO - codeparrot_training - Step 2372: {'lr': 0.0004999259045210901, 'samples': 1214976, 'steps': 2372, 'loss/train': 2.7310776710510254} 02/24/2022 04:39:22 - INFO - codeparrot_training - Step 2373: {'lr': 0.000499925505642678, 'samples': 1215488, 'steps': 2373, 'loss/train': 5.180909156799316} 02/24/2022 04:39:28 - INFO - codeparrot_training - Step 2374: {'lr': 0.0004999251056936645, 'samples': 1216000, 'steps': 2374, 'loss/train': 2.431812047958374} 02/24/2022 04:39:32 - INFO - codeparrot_training - Step 2375: {'lr': 0.000499924704674051, 'samples': 1216512, 'steps': 2375, 'loss/train': 4.3434953689575195} 02/24/2022 04:39:37 - INFO - codeparrot_training - Step 2376: {'lr': 0.0004999243025838396, 'samples': 1217024, 'steps': 2376, 'loss/train': 3.5757734775543213} 02/24/2022 04:39:40 - INFO - codeparrot_training - Step 2377: {'lr': 0.0004999238994230318, 'samples': 1217536, 'steps': 2377, 'loss/train': 0.8257812857627869} 02/24/2022 04:39:46 - INFO - codeparrot_training - Step 2378: {'lr': 0.0004999234951916293, 'samples': 1218048, 'steps': 2378, 'loss/train': 3.2002294063568115} 02/24/2022 04:39:49 - INFO - codeparrot_training - Step 2379: {'lr': 0.0004999230898896341, 'samples': 1218560, 'steps': 2379, 'loss/train': 2.6964566707611084} 02/24/2022 04:39:55 - INFO - codeparrot_training - Step 2380: {'lr': 0.0004999226835170476, 'samples': 1219072, 'steps': 2380, 'loss/train': 3.7121119499206543} 02/24/2022 04:39:58 - INFO - codeparrot_training - Step 2381: {'lr': 0.0004999222760738717, 'samples': 1219584, 'steps': 2381, 'loss/train': 2.8447868824005127} 02/24/2022 04:40:04 - INFO - codeparrot_training - Step 2382: {'lr': 0.0004999218675601081, 'samples': 1220096, 'steps': 2382, 'loss/train': 4.5413126945495605} 02/24/2022 04:40:07 - INFO - codeparrot_training - Step 2383: {'lr': 0.0004999214579757586, 'samples': 1220608, 'steps': 2383, 'loss/train': 2.644672155380249} 02/24/2022 04:40:13 - INFO - codeparrot_training - Step 2384: {'lr': 0.000499921047320825, 'samples': 1221120, 'steps': 2384, 'loss/train': 4.234299182891846} 02/24/2022 04:40:16 - INFO - codeparrot_training - Step 2385: {'lr': 0.000499920635595309, 'samples': 1221632, 'steps': 2385, 'loss/train': 4.1897759437561035} 02/24/2022 04:40:22 - INFO - codeparrot_training - Step 2386: {'lr': 0.0004999202227992122, 'samples': 1222144, 'steps': 2386, 'loss/train': 3.9284749031066895} 02/24/2022 04:40:25 - INFO - codeparrot_training - Step 2387: {'lr': 0.0004999198089325367, 'samples': 1222656, 'steps': 2387, 'loss/train': 3.6172170639038086} 02/24/2022 04:40:31 - INFO - codeparrot_training - Step 2388: {'lr': 0.0004999193939952839, 'samples': 1223168, 'steps': 2388, 'loss/train': 3.1974456310272217} 02/24/2022 04:40:34 - INFO - codeparrot_training - Step 2389: {'lr': 0.000499918977987456, 'samples': 1223680, 'steps': 2389, 'loss/train': 1.7424440383911133} 02/24/2022 04:40:40 - INFO - codeparrot_training - Step 2390: {'lr': 0.0004999185609090544, 'samples': 1224192, 'steps': 2390, 'loss/train': 3.1206183433532715} 02/24/2022 04:40:46 - INFO - codeparrot_training - Step 2391: {'lr': 0.0004999181427600811, 'samples': 1224704, 'steps': 2391, 'loss/train': 3.673250675201416} 02/24/2022 04:40:49 - INFO - codeparrot_training - Step 2392: {'lr': 0.0004999177235405378, 'samples': 1225216, 'steps': 2392, 'loss/train': 3.422680377960205} 02/24/2022 04:40:55 - INFO - codeparrot_training - Step 2393: {'lr': 0.0004999173032504264, 'samples': 1225728, 'steps': 2393, 'loss/train': 2.807474136352539} 02/24/2022 04:40:58 - INFO - codeparrot_training - Step 2394: {'lr': 0.0004999168818897486, 'samples': 1226240, 'steps': 2394, 'loss/train': 3.026815414428711} 02/24/2022 04:41:04 - INFO - codeparrot_training - Step 2395: {'lr': 0.0004999164594585062, 'samples': 1226752, 'steps': 2395, 'loss/train': 2.855661392211914} 02/24/2022 04:41:07 - INFO - codeparrot_training - Step 2396: {'lr': 0.0004999160359567011, 'samples': 1227264, 'steps': 2396, 'loss/train': 3.660255193710327} 02/24/2022 04:41:13 - INFO - codeparrot_training - Step 2397: {'lr': 0.000499915611384335, 'samples': 1227776, 'steps': 2397, 'loss/train': 2.143773317337036} 02/24/2022 04:41:16 - INFO - codeparrot_training - Step 2398: {'lr': 0.0004999151857414099, 'samples': 1228288, 'steps': 2398, 'loss/train': 4.035255432128906} 02/24/2022 04:41:22 - INFO - codeparrot_training - Step 2399: {'lr': 0.0004999147590279273, 'samples': 1228800, 'steps': 2399, 'loss/train': 0.5259669423103333} 02/24/2022 04:41:26 - INFO - codeparrot_training - Step 2400: {'lr': 0.0004999143312438893, 'samples': 1229312, 'steps': 2400, 'loss/train': 2.4696826934814453} 02/24/2022 04:41:31 - INFO - codeparrot_training - Step 2401: {'lr': 0.0004999139023892978, 'samples': 1229824, 'steps': 2401, 'loss/train': 3.858065605163574} 02/24/2022 04:41:35 - INFO - codeparrot_training - Step 2402: {'lr': 0.0004999134724641543, 'samples': 1230336, 'steps': 2402, 'loss/train': 3.43780517578125} 02/24/2022 04:41:40 - INFO - codeparrot_training - Step 2403: {'lr': 0.000499913041468461, 'samples': 1230848, 'steps': 2403, 'loss/train': 2.750088691711426} 02/24/2022 04:41:44 - INFO - codeparrot_training - Step 2404: {'lr': 0.0004999126094022195, 'samples': 1231360, 'steps': 2404, 'loss/train': 3.518949031829834} 02/24/2022 04:41:49 - INFO - codeparrot_training - Step 2405: {'lr': 0.0004999121762654318, 'samples': 1231872, 'steps': 2405, 'loss/train': 2.4769396781921387} 02/24/2022 04:41:53 - INFO - codeparrot_training - Step 2406: {'lr': 0.0004999117420580996, 'samples': 1232384, 'steps': 2406, 'loss/train': 4.040136814117432} 02/24/2022 04:41:58 - INFO - codeparrot_training - Step 2407: {'lr': 0.0004999113067802249, 'samples': 1232896, 'steps': 2407, 'loss/train': 3.6795735359191895} 02/24/2022 04:42:01 - INFO - codeparrot_training - Step 2408: {'lr': 0.0004999108704318095, 'samples': 1233408, 'steps': 2408, 'loss/train': 3.4916088581085205} 02/24/2022 04:42:08 - INFO - codeparrot_training - Step 2409: {'lr': 0.0004999104330128553, 'samples': 1233920, 'steps': 2409, 'loss/train': 3.254087209701538} 02/24/2022 04:42:12 - INFO - codeparrot_training - Step 2410: {'lr': 0.0004999099945233641, 'samples': 1234432, 'steps': 2410, 'loss/train': 3.413273572921753} 02/24/2022 04:42:17 - INFO - codeparrot_training - Step 2411: {'lr': 0.000499909554963338, 'samples': 1234944, 'steps': 2411, 'loss/train': 3.116635799407959} 02/24/2022 04:42:21 - INFO - codeparrot_training - Step 2412: {'lr': 0.0004999091143327786, 'samples': 1235456, 'steps': 2412, 'loss/train': 3.037827253341675} 02/24/2022 04:42:26 - INFO - codeparrot_training - Step 2413: {'lr': 0.000499908672631688, 'samples': 1235968, 'steps': 2413, 'loss/train': 4.417897701263428} 02/24/2022 04:42:30 - INFO - codeparrot_training - Step 2414: {'lr': 0.0004999082298600679, 'samples': 1236480, 'steps': 2414, 'loss/train': 3.452875852584839} 02/24/2022 04:42:35 - INFO - codeparrot_training - Step 2415: {'lr': 0.0004999077860179204, 'samples': 1236992, 'steps': 2415, 'loss/train': 2.8953804969787598} 02/24/2022 04:42:39 - INFO - codeparrot_training - Step 2416: {'lr': 0.0004999073411052472, 'samples': 1237504, 'steps': 2416, 'loss/train': 3.4752864837646484} 02/24/2022 04:42:44 - INFO - codeparrot_training - Step 2417: {'lr': 0.0004999068951220503, 'samples': 1238016, 'steps': 2417, 'loss/train': 3.343427896499634} 02/24/2022 04:42:47 - INFO - codeparrot_training - Step 2418: {'lr': 0.0004999064480683317, 'samples': 1238528, 'steps': 2418, 'loss/train': 2.558880567550659} 02/24/2022 04:42:54 - INFO - codeparrot_training - Step 2419: {'lr': 0.0004999059999440932, 'samples': 1239040, 'steps': 2419, 'loss/train': 4.296968936920166} 02/24/2022 04:42:57 - INFO - codeparrot_training - Step 2420: {'lr': 0.0004999055507493368, 'samples': 1239552, 'steps': 2420, 'loss/train': 3.0869860649108887} 02/24/2022 04:43:03 - INFO - codeparrot_training - Step 2421: {'lr': 0.0004999051004840642, 'samples': 1240064, 'steps': 2421, 'loss/train': 2.653878927230835} 02/24/2022 04:43:06 - INFO - codeparrot_training - Step 2422: {'lr': 0.0004999046491482777, 'samples': 1240576, 'steps': 2422, 'loss/train': 3.7434804439544678} 02/24/2022 04:43:12 - INFO - codeparrot_training - Step 2423: {'lr': 0.000499904196741979, 'samples': 1241088, 'steps': 2423, 'loss/train': 3.566549301147461} 02/24/2022 04:43:15 - INFO - codeparrot_training - Step 2424: {'lr': 0.00049990374326517, 'samples': 1241600, 'steps': 2424, 'loss/train': 3.7648866176605225} 02/24/2022 04:43:21 - INFO - codeparrot_training - Step 2425: {'lr': 0.0004999032887178527, 'samples': 1242112, 'steps': 2425, 'loss/train': 3.2500345706939697} 02/24/2022 04:43:24 - INFO - codeparrot_training - Step 2426: {'lr': 0.000499902833100029, 'samples': 1242624, 'steps': 2426, 'loss/train': 4.845691204071045} 02/24/2022 04:43:30 - INFO - codeparrot_training - Step 2427: {'lr': 0.0004999023764117011, 'samples': 1243136, 'steps': 2427, 'loss/train': 0.8710881471633911} 02/24/2022 04:43:33 - INFO - codeparrot_training - Step 2428: {'lr': 0.0004999019186528708, 'samples': 1243648, 'steps': 2428, 'loss/train': 4.0029497146606445} 02/24/2022 04:43:39 - INFO - codeparrot_training - Step 2429: {'lr': 0.0004999014598235399, 'samples': 1244160, 'steps': 2429, 'loss/train': 2.7371466159820557} 02/24/2022 04:43:42 - INFO - codeparrot_training - Step 2430: {'lr': 0.0004999009999237105, 'samples': 1244672, 'steps': 2430, 'loss/train': 4.078182220458984} 02/24/2022 04:43:48 - INFO - codeparrot_training - Step 2431: {'lr': 0.0004999005389533846, 'samples': 1245184, 'steps': 2431, 'loss/train': 3.9989120960235596} 02/24/2022 04:43:51 - INFO - codeparrot_training - Step 2432: {'lr': 0.0004999000769125642, 'samples': 1245696, 'steps': 2432, 'loss/train': 2.7581992149353027} 02/24/2022 04:43:57 - INFO - codeparrot_training - Step 2433: {'lr': 0.0004998996138012512, 'samples': 1246208, 'steps': 2433, 'loss/train': 1.4696320295333862} 02/24/2022 04:44:00 - INFO - codeparrot_training - Step 2434: {'lr': 0.0004998991496194475, 'samples': 1246720, 'steps': 2434, 'loss/train': 0.5218791365623474} 02/24/2022 04:44:06 - INFO - codeparrot_training - Step 2435: {'lr': 0.0004998986843671552, 'samples': 1247232, 'steps': 2435, 'loss/train': 1.8831707239151} 02/24/2022 04:44:10 - INFO - codeparrot_training - Step 2436: {'lr': 0.0004998982180443764, 'samples': 1247744, 'steps': 2436, 'loss/train': 3.620414972305298} 02/24/2022 04:44:15 - INFO - codeparrot_training - Step 2437: {'lr': 0.000499897750651113, 'samples': 1248256, 'steps': 2437, 'loss/train': 4.267504692077637} 02/24/2022 04:44:19 - INFO - codeparrot_training - Step 2438: {'lr': 0.0004998972821873668, 'samples': 1248768, 'steps': 2438, 'loss/train': 2.5739922523498535} 02/24/2022 04:44:24 - INFO - codeparrot_training - Step 2439: {'lr': 0.0004998968126531402, 'samples': 1249280, 'steps': 2439, 'loss/train': 3.807056188583374} 02/24/2022 04:44:28 - INFO - codeparrot_training - Step 2440: {'lr': 0.0004998963420484349, 'samples': 1249792, 'steps': 2440, 'loss/train': 3.8136444091796875} 02/24/2022 04:44:33 - INFO - codeparrot_training - Step 2441: {'lr': 0.0004998958703732532, 'samples': 1250304, 'steps': 2441, 'loss/train': 1.0674653053283691} 02/24/2022 04:44:37 - INFO - codeparrot_training - Step 2442: {'lr': 0.0004998953976275966, 'samples': 1250816, 'steps': 2442, 'loss/train': 3.5890421867370605} 02/24/2022 04:44:42 - INFO - codeparrot_training - Step 2443: {'lr': 0.0004998949238114677, 'samples': 1251328, 'steps': 2443, 'loss/train': 2.782356023788452} 02/24/2022 04:44:49 - INFO - codeparrot_training - Step 2444: {'lr': 0.0004998944489248683, 'samples': 1251840, 'steps': 2444, 'loss/train': 3.3370895385742188} 02/24/2022 04:44:52 - INFO - codeparrot_training - Step 2445: {'lr': 0.0004998939729678004, 'samples': 1252352, 'steps': 2445, 'loss/train': 2.900718927383423} 02/24/2022 04:44:57 - INFO - codeparrot_training - Step 2446: {'lr': 0.000499893495940266, 'samples': 1252864, 'steps': 2446, 'loss/train': 4.6819748878479} 02/24/2022 04:45:01 - INFO - codeparrot_training - Step 2447: {'lr': 0.0004998930178422673, 'samples': 1253376, 'steps': 2447, 'loss/train': 2.8547513484954834} 02/24/2022 04:45:06 - INFO - codeparrot_training - Step 2448: {'lr': 0.0004998925386738062, 'samples': 1253888, 'steps': 2448, 'loss/train': 4.188310146331787} 02/24/2022 04:45:10 - INFO - codeparrot_training - Step 2449: {'lr': 0.0004998920584348849, 'samples': 1254400, 'steps': 2449, 'loss/train': 3.1551120281219482} 02/24/2022 04:45:16 - INFO - codeparrot_training - Step 2450: {'lr': 0.0004998915771255053, 'samples': 1254912, 'steps': 2450, 'loss/train': 4.014820575714111} 02/24/2022 04:45:20 - INFO - codeparrot_training - Step 2451: {'lr': 0.0004998910947456696, 'samples': 1255424, 'steps': 2451, 'loss/train': 3.264758825302124} 02/24/2022 04:45:23 - INFO - codeparrot_training - Step 2452: {'lr': 0.0004998906112953797, 'samples': 1255936, 'steps': 2452, 'loss/train': 3.2256007194519043} 02/24/2022 04:45:29 - INFO - codeparrot_training - Step 2453: {'lr': 0.0004998901267746379, 'samples': 1256448, 'steps': 2453, 'loss/train': 2.6657519340515137} 02/24/2022 04:45:32 - INFO - codeparrot_training - Step 2454: {'lr': 0.0004998896411834461, 'samples': 1256960, 'steps': 2454, 'loss/train': 3.830369710922241} 02/24/2022 04:45:38 - INFO - codeparrot_training - Step 2455: {'lr': 0.0004998891545218063, 'samples': 1257472, 'steps': 2455, 'loss/train': 4.173727989196777} 02/24/2022 04:45:42 - INFO - codeparrot_training - Step 2456: {'lr': 0.0004998886667897209, 'samples': 1257984, 'steps': 2456, 'loss/train': 3.6768832206726074} 02/24/2022 04:45:47 - INFO - codeparrot_training - Step 2457: {'lr': 0.0004998881779871917, 'samples': 1258496, 'steps': 2457, 'loss/train': 2.9944214820861816} 02/24/2022 04:45:51 - INFO - codeparrot_training - Step 2458: {'lr': 0.0004998876881142208, 'samples': 1259008, 'steps': 2458, 'loss/train': 2.1139795780181885} 02/24/2022 04:45:56 - INFO - codeparrot_training - Step 2459: {'lr': 0.0004998871971708106, 'samples': 1259520, 'steps': 2459, 'loss/train': 3.3430514335632324} 02/24/2022 04:45:59 - INFO - codeparrot_training - Step 2460: {'lr': 0.0004998867051569627, 'samples': 1260032, 'steps': 2460, 'loss/train': 3.4140310287475586} 02/24/2022 04:46:05 - INFO - codeparrot_training - Step 2461: {'lr': 0.0004998862120726798, 'samples': 1260544, 'steps': 2461, 'loss/train': 2.1001102924346924} 02/24/2022 04:46:08 - INFO - codeparrot_training - Step 2462: {'lr': 0.0004998857179179636, 'samples': 1261056, 'steps': 2462, 'loss/train': 2.8193459510803223} 02/24/2022 04:46:14 - INFO - codeparrot_training - Step 2463: {'lr': 0.0004998852226928164, 'samples': 1261568, 'steps': 2463, 'loss/train': 1.900756597518921} 02/24/2022 04:46:17 - INFO - codeparrot_training - Step 2464: {'lr': 0.0004998847263972401, 'samples': 1262080, 'steps': 2464, 'loss/train': 3.9644336700439453} 02/24/2022 04:46:23 - INFO - codeparrot_training - Step 2465: {'lr': 0.0004998842290312371, 'samples': 1262592, 'steps': 2465, 'loss/train': 2.8959858417510986} 02/24/2022 04:46:26 - INFO - codeparrot_training - Step 2466: {'lr': 0.0004998837305948094, 'samples': 1263104, 'steps': 2466, 'loss/train': 2.9531450271606445} 02/24/2022 04:46:32 - INFO - codeparrot_training - Step 2467: {'lr': 0.0004998832310879591, 'samples': 1263616, 'steps': 2467, 'loss/train': 0.5945470929145813} 02/24/2022 04:46:35 - INFO - codeparrot_training - Step 2468: {'lr': 0.0004998827305106884, 'samples': 1264128, 'steps': 2468, 'loss/train': 3.9969992637634277} 02/24/2022 04:46:41 - INFO - codeparrot_training - Step 2469: {'lr': 0.0004998822288629995, 'samples': 1264640, 'steps': 2469, 'loss/train': 2.9935379028320312} 02/24/2022 04:46:44 - INFO - codeparrot_training - Step 2470: {'lr': 0.0004998817261448943, 'samples': 1265152, 'steps': 2470, 'loss/train': 3.9394543170928955} 02/24/2022 04:46:51 - INFO - codeparrot_training - Step 2471: {'lr': 0.0004998812223563754, 'samples': 1265664, 'steps': 2471, 'loss/train': 2.3813884258270264} 02/24/2022 04:46:54 - INFO - codeparrot_training - Step 2472: {'lr': 0.0004998807174974445, 'samples': 1266176, 'steps': 2472, 'loss/train': 2.9916625022888184} 02/24/2022 04:47:00 - INFO - codeparrot_training - Step 2473: {'lr': 0.0004998802115681039, 'samples': 1266688, 'steps': 2473, 'loss/train': 1.8967984914779663} 02/24/2022 04:47:03 - INFO - codeparrot_training - Step 2474: {'lr': 0.000499879704568356, 'samples': 1267200, 'steps': 2474, 'loss/train': 3.8154592514038086} 02/24/2022 04:47:09 - INFO - codeparrot_training - Step 2475: {'lr': 0.0004998791964982026, 'samples': 1267712, 'steps': 2475, 'loss/train': 0.8268159627914429} 02/24/2022 04:47:12 - INFO - codeparrot_training - Step 2476: {'lr': 0.0004998786873576462, 'samples': 1268224, 'steps': 2476, 'loss/train': 4.143904209136963} 02/24/2022 04:47:18 - INFO - codeparrot_training - Step 2477: {'lr': 0.0004998781771466889, 'samples': 1268736, 'steps': 2477, 'loss/train': 3.781113862991333} 02/24/2022 04:47:21 - INFO - codeparrot_training - Step 2478: {'lr': 0.0004998776658653327, 'samples': 1269248, 'steps': 2478, 'loss/train': 3.5197501182556152} 02/24/2022 04:47:27 - INFO - codeparrot_training - Step 2479: {'lr': 0.00049987715351358, 'samples': 1269760, 'steps': 2479, 'loss/train': 3.852661371231079} 02/24/2022 04:47:30 - INFO - codeparrot_training - Step 2480: {'lr': 0.0004998766400914329, 'samples': 1270272, 'steps': 2480, 'loss/train': 4.078242301940918} 02/24/2022 04:47:36 - INFO - codeparrot_training - Step 2481: {'lr': 0.0004998761255988936, 'samples': 1270784, 'steps': 2481, 'loss/train': 3.6569957733154297} 02/24/2022 04:47:40 - INFO - codeparrot_training - Step 2482: {'lr': 0.0004998756100359643, 'samples': 1271296, 'steps': 2482, 'loss/train': 4.2928595542907715} 02/24/2022 04:47:46 - INFO - codeparrot_training - Step 2483: {'lr': 0.0004998750934026474, 'samples': 1271808, 'steps': 2483, 'loss/train': 0.5672394037246704} 02/24/2022 04:47:49 - INFO - codeparrot_training - Step 2484: {'lr': 0.0004998745756989448, 'samples': 1272320, 'steps': 2484, 'loss/train': 0.7372508645057678} 02/24/2022 04:47:54 - INFO - codeparrot_training - Step 2485: {'lr': 0.0004998740569248588, 'samples': 1272832, 'steps': 2485, 'loss/train': 3.4258670806884766} 02/24/2022 04:47:58 - INFO - codeparrot_training - Step 2486: {'lr': 0.0004998735370803917, 'samples': 1273344, 'steps': 2486, 'loss/train': 3.582427978515625} 02/24/2022 04:48:03 - INFO - codeparrot_training - Step 2487: {'lr': 0.0004998730161655459, 'samples': 1273856, 'steps': 2487, 'loss/train': 2.98801851272583} 02/24/2022 04:48:07 - INFO - codeparrot_training - Step 2488: {'lr': 0.0004998724941803232, 'samples': 1274368, 'steps': 2488, 'loss/train': 2.776625633239746} 02/24/2022 04:48:12 - INFO - codeparrot_training - Step 2489: {'lr': 0.0004998719711247262, 'samples': 1274880, 'steps': 2489, 'loss/train': 2.3862781524658203} 02/24/2022 04:48:18 - INFO - codeparrot_training - Step 2490: {'lr': 0.0004998714469987571, 'samples': 1275392, 'steps': 2490, 'loss/train': 2.8626139163970947} 02/24/2022 04:48:22 - INFO - codeparrot_training - Step 2491: {'lr': 0.000499870921802418, 'samples': 1275904, 'steps': 2491, 'loss/train': 3.086052417755127} 02/24/2022 04:48:28 - INFO - codeparrot_training - Step 2492: {'lr': 0.0004998703955357111, 'samples': 1276416, 'steps': 2492, 'loss/train': 2.601979970932007} 02/24/2022 04:48:31 - INFO - codeparrot_training - Step 2493: {'lr': 0.0004998698681986389, 'samples': 1276928, 'steps': 2493, 'loss/train': 0.6036854982376099} 02/24/2022 04:48:37 - INFO - codeparrot_training - Step 2494: {'lr': 0.0004998693397912034, 'samples': 1277440, 'steps': 2494, 'loss/train': 3.421665668487549} 02/24/2022 04:48:40 - INFO - codeparrot_training - Step 2495: {'lr': 0.0004998688103134072, 'samples': 1277952, 'steps': 2495, 'loss/train': 3.5919039249420166} 02/24/2022 04:48:46 - INFO - codeparrot_training - Step 2496: {'lr': 0.0004998682797652522, 'samples': 1278464, 'steps': 2496, 'loss/train': 2.873520851135254} 02/24/2022 04:48:49 - INFO - codeparrot_training - Step 2497: {'lr': 0.0004998677481467408, 'samples': 1278976, 'steps': 2497, 'loss/train': 2.5793066024780273} 02/24/2022 04:48:55 - INFO - codeparrot_training - Step 2498: {'lr': 0.0004998672154578754, 'samples': 1279488, 'steps': 2498, 'loss/train': 3.4168808460235596} 02/24/2022 04:48:58 - INFO - codeparrot_training - Step 2499: {'lr': 0.0004998666816986582, 'samples': 1280000, 'steps': 2499, 'loss/train': 1.6697356700897217} 02/24/2022 04:49:04 - INFO - codeparrot_training - Step 2500: {'lr': 0.0004998661468690914, 'samples': 1280512, 'steps': 2500, 'loss/train': 4.405819892883301} 02/24/2022 04:49:07 - INFO - codeparrot_training - Step 2501: {'lr': 0.0004998656109691774, 'samples': 1281024, 'steps': 2501, 'loss/train': 3.3693747520446777} 02/24/2022 04:49:14 - INFO - codeparrot_training - Step 2502: {'lr': 0.0004998650739989185, 'samples': 1281536, 'steps': 2502, 'loss/train': 3.306903600692749} 02/24/2022 04:49:17 - INFO - codeparrot_training - Step 2503: {'lr': 0.0004998645359583169, 'samples': 1282048, 'steps': 2503, 'loss/train': 2.528913736343384} 02/24/2022 04:49:23 - INFO - codeparrot_training - Step 2504: {'lr': 0.0004998639968473751, 'samples': 1282560, 'steps': 2504, 'loss/train': 3.113189697265625} 02/24/2022 04:49:26 - INFO - codeparrot_training - Step 2505: {'lr': 0.0004998634566660952, 'samples': 1283072, 'steps': 2505, 'loss/train': 2.5438289642333984} 02/24/2022 04:49:32 - INFO - codeparrot_training - Step 2506: {'lr': 0.0004998629154144795, 'samples': 1283584, 'steps': 2506, 'loss/train': 3.474109649658203} 02/24/2022 04:49:35 - INFO - codeparrot_training - Step 2507: {'lr': 0.0004998623730925305, 'samples': 1284096, 'steps': 2507, 'loss/train': 4.123108863830566} 02/24/2022 04:49:41 - INFO - codeparrot_training - Step 2508: {'lr': 0.0004998618297002504, 'samples': 1284608, 'steps': 2508, 'loss/train': 2.341304063796997} 02/24/2022 04:49:44 - INFO - codeparrot_training - Step 2509: {'lr': 0.0004998612852376417, 'samples': 1285120, 'steps': 2509, 'loss/train': 2.595646858215332} 02/24/2022 04:49:50 - INFO - codeparrot_training - Step 2510: {'lr': 0.0004998607397047063, 'samples': 1285632, 'steps': 2510, 'loss/train': 2.518873691558838} 02/24/2022 04:49:53 - INFO - codeparrot_training - Step 2511: {'lr': 0.0004998601931014471, 'samples': 1286144, 'steps': 2511, 'loss/train': 3.6765215396881104} 02/24/2022 04:49:59 - INFO - codeparrot_training - Step 2512: {'lr': 0.0004998596454278661, 'samples': 1286656, 'steps': 2512, 'loss/train': 1.8189362287521362} 02/24/2022 04:50:03 - INFO - codeparrot_training - Step 2513: {'lr': 0.0004998590966839657, 'samples': 1287168, 'steps': 2513, 'loss/train': 5.485591888427734} 02/24/2022 04:50:08 - INFO - codeparrot_training - Step 2514: {'lr': 0.0004998585468697482, 'samples': 1287680, 'steps': 2514, 'loss/train': 3.153528928756714} 02/24/2022 04:50:12 - INFO - codeparrot_training - Step 2515: {'lr': 0.0004998579959852161, 'samples': 1288192, 'steps': 2515, 'loss/train': 3.1058945655822754} 02/24/2022 04:50:17 - INFO - codeparrot_training - Step 2516: {'lr': 0.0004998574440303718, 'samples': 1288704, 'steps': 2516, 'loss/train': 3.6950082778930664} 02/24/2022 04:50:21 - INFO - codeparrot_training - Step 2517: {'lr': 0.0004998568910052173, 'samples': 1289216, 'steps': 2517, 'loss/train': 2.8677847385406494} 02/24/2022 04:50:26 - INFO - codeparrot_training - Step 2518: {'lr': 0.0004998563369097554, 'samples': 1289728, 'steps': 2518, 'loss/train': 2.691347360610962} 02/24/2022 04:50:30 - INFO - codeparrot_training - Step 2519: {'lr': 0.0004998557817439882, 'samples': 1290240, 'steps': 2519, 'loss/train': 2.9570348262786865} 02/24/2022 04:50:35 - INFO - codeparrot_training - Step 2520: {'lr': 0.0004998552255079182, 'samples': 1290752, 'steps': 2520, 'loss/train': 3.1546807289123535} 02/24/2022 04:50:39 - INFO - codeparrot_training - Step 2521: {'lr': 0.0004998546682015478, 'samples': 1291264, 'steps': 2521, 'loss/train': 4.013919353485107} 02/24/2022 04:50:44 - INFO - codeparrot_training - Step 2522: {'lr': 0.0004998541098248793, 'samples': 1291776, 'steps': 2522, 'loss/train': 3.959960699081421} 02/24/2022 04:50:48 - INFO - codeparrot_training - Step 2523: {'lr': 0.0004998535503779151, 'samples': 1292288, 'steps': 2523, 'loss/train': 3.3066318035125732} 02/24/2022 04:50:53 - INFO - codeparrot_training - Step 2524: {'lr': 0.0004998529898606576, 'samples': 1292800, 'steps': 2524, 'loss/train': 0.5945372581481934} 02/24/2022 04:50:57 - INFO - codeparrot_training - Step 2525: {'lr': 0.0004998524282731093, 'samples': 1293312, 'steps': 2525, 'loss/train': 3.0410656929016113} 02/24/2022 04:51:02 - INFO - codeparrot_training - Step 2526: {'lr': 0.0004998518656152725, 'samples': 1293824, 'steps': 2526, 'loss/train': 3.017993927001953} 02/24/2022 04:51:06 - INFO - codeparrot_training - Step 2527: {'lr': 0.0004998513018871498, 'samples': 1294336, 'steps': 2527, 'loss/train': 3.1269800662994385} 02/24/2022 04:51:12 - INFO - codeparrot_training - Step 2528: {'lr': 0.0004998507370887433, 'samples': 1294848, 'steps': 2528, 'loss/train': 3.7678191661834717} 02/24/2022 04:51:15 - INFO - codeparrot_training - Step 2529: {'lr': 0.0004998501712200555, 'samples': 1295360, 'steps': 2529, 'loss/train': 0.6667803525924683} 02/24/2022 04:51:21 - INFO - codeparrot_training - Step 2530: {'lr': 0.000499849604281089, 'samples': 1295872, 'steps': 2530, 'loss/train': 3.203901529312134} 02/24/2022 04:51:24 - INFO - codeparrot_training - Step 2531: {'lr': 0.0004998490362718462, 'samples': 1296384, 'steps': 2531, 'loss/train': 3.729490041732788} 02/24/2022 04:51:30 - INFO - codeparrot_training - Step 2532: {'lr': 0.0004998484671923293, 'samples': 1296896, 'steps': 2532, 'loss/train': 4.017566680908203} 02/24/2022 04:51:34 - INFO - codeparrot_training - Step 2533: {'lr': 0.000499847897042541, 'samples': 1297408, 'steps': 2533, 'loss/train': 3.762479782104492} 02/24/2022 04:51:39 - INFO - codeparrot_training - Step 2534: {'lr': 0.0004998473258224837, 'samples': 1297920, 'steps': 2534, 'loss/train': 2.2608602046966553} 02/24/2022 04:51:42 - INFO - codeparrot_training - Step 2535: {'lr': 0.0004998467535321597, 'samples': 1298432, 'steps': 2535, 'loss/train': 3.631470203399658} 02/24/2022 04:51:48 - INFO - codeparrot_training - Step 2536: {'lr': 0.0004998461801715716, 'samples': 1298944, 'steps': 2536, 'loss/train': 3.0279645919799805} 02/24/2022 04:51:52 - INFO - codeparrot_training - Step 2537: {'lr': 0.0004998456057407218, 'samples': 1299456, 'steps': 2537, 'loss/train': 3.118901014328003} 02/24/2022 04:51:58 - INFO - codeparrot_training - Step 2538: {'lr': 0.0004998450302396127, 'samples': 1299968, 'steps': 2538, 'loss/train': 3.350200891494751} 02/24/2022 04:52:01 - INFO - codeparrot_training - Step 2539: {'lr': 0.0004998444536682469, 'samples': 1300480, 'steps': 2539, 'loss/train': 2.497511386871338} 02/24/2022 04:52:06 - INFO - codeparrot_training - Step 2540: {'lr': 0.0004998438760266267, 'samples': 1300992, 'steps': 2540, 'loss/train': 2.3810489177703857} 02/24/2022 04:52:10 - INFO - codeparrot_training - Step 2541: {'lr': 0.0004998432973147548, 'samples': 1301504, 'steps': 2541, 'loss/train': 2.9320499897003174} 02/24/2022 04:52:15 - INFO - codeparrot_training - Step 2542: {'lr': 0.0004998427175326335, 'samples': 1302016, 'steps': 2542, 'loss/train': 3.6699774265289307} 02/24/2022 04:52:19 - INFO - codeparrot_training - Step 2543: {'lr': 0.0004998421366802653, 'samples': 1302528, 'steps': 2543, 'loss/train': 3.641139030456543} 02/24/2022 04:52:24 - INFO - codeparrot_training - Step 2544: {'lr': 0.0004998415547576527, 'samples': 1303040, 'steps': 2544, 'loss/train': 3.030946731567383} 02/24/2022 04:52:28 - INFO - codeparrot_training - Step 2545: {'lr': 0.0004998409717647983, 'samples': 1303552, 'steps': 2545, 'loss/train': 3.1390645503997803} 02/24/2022 04:52:33 - INFO - codeparrot_training - Step 2546: {'lr': 0.0004998403877017044, 'samples': 1304064, 'steps': 2546, 'loss/train': 2.7460896968841553} 02/24/2022 04:52:39 - INFO - codeparrot_training - Step 2547: {'lr': 0.0004998398025683737, 'samples': 1304576, 'steps': 2547, 'loss/train': 2.8691656589508057} 02/24/2022 04:52:43 - INFO - codeparrot_training - Step 2548: {'lr': 0.0004998392163648085, 'samples': 1305088, 'steps': 2548, 'loss/train': 3.3751120567321777} 02/24/2022 04:52:48 - INFO - codeparrot_training - Step 2549: {'lr': 0.0004998386290910116, 'samples': 1305600, 'steps': 2549, 'loss/train': 3.710108995437622} 02/24/2022 04:52:52 - INFO - codeparrot_training - Step 2550: {'lr': 0.0004998380407469853, 'samples': 1306112, 'steps': 2550, 'loss/train': 3.1626763343811035} 02/24/2022 04:52:57 - INFO - codeparrot_training - Step 2551: {'lr': 0.0004998374513327321, 'samples': 1306624, 'steps': 2551, 'loss/train': 3.2368767261505127} 02/24/2022 04:53:01 - INFO - codeparrot_training - Step 2552: {'lr': 0.0004998368608482546, 'samples': 1307136, 'steps': 2552, 'loss/train': 2.595431327819824} 02/24/2022 04:53:06 - INFO - codeparrot_training - Step 2553: {'lr': 0.0004998362692935553, 'samples': 1307648, 'steps': 2553, 'loss/train': 4.266594886779785} 02/24/2022 04:53:10 - INFO - codeparrot_training - Step 2554: {'lr': 0.0004998356766686368, 'samples': 1308160, 'steps': 2554, 'loss/train': 3.5977730751037598} 02/24/2022 04:53:15 - INFO - codeparrot_training - Step 2555: {'lr': 0.0004998350829735016, 'samples': 1308672, 'steps': 2555, 'loss/train': 3.4415242671966553} 02/24/2022 04:53:19 - INFO - codeparrot_training - Step 2556: {'lr': 0.0004998344882081522, 'samples': 1309184, 'steps': 2556, 'loss/train': 1.8883965015411377} 02/24/2022 04:53:24 - INFO - codeparrot_training - Step 2557: {'lr': 0.0004998338923725913, 'samples': 1309696, 'steps': 2557, 'loss/train': 4.211737632751465} 02/24/2022 04:53:28 - INFO - codeparrot_training - Step 2558: {'lr': 0.0004998332954668211, 'samples': 1310208, 'steps': 2558, 'loss/train': 2.723522901535034} 02/24/2022 04:53:33 - INFO - codeparrot_training - Step 2559: {'lr': 0.0004998326974908446, 'samples': 1310720, 'steps': 2559, 'loss/train': 2.6041481494903564} 02/24/2022 04:53:37 - INFO - codeparrot_training - Step 2560: {'lr': 0.0004998320984446641, 'samples': 1311232, 'steps': 2560, 'loss/train': 3.5851409435272217} 02/24/2022 04:53:42 - INFO - codeparrot_training - Step 2561: {'lr': 0.0004998314983282821, 'samples': 1311744, 'steps': 2561, 'loss/train': 2.926137924194336} 02/24/2022 04:53:46 - INFO - codeparrot_training - Step 2562: {'lr': 0.0004998308971417015, 'samples': 1312256, 'steps': 2562, 'loss/train': 3.7162015438079834} 02/24/2022 04:53:52 - INFO - codeparrot_training - Step 2563: {'lr': 0.0004998302948849246, 'samples': 1312768, 'steps': 2563, 'loss/train': 3.186518669128418} 02/24/2022 04:53:55 - INFO - codeparrot_training - Step 2564: {'lr': 0.0004998296915579539, 'samples': 1313280, 'steps': 2564, 'loss/train': 3.566105365753174} 02/24/2022 04:54:01 - INFO - codeparrot_training - Step 2565: {'lr': 0.0004998290871607924, 'samples': 1313792, 'steps': 2565, 'loss/train': 3.6817195415496826} 02/24/2022 04:54:04 - INFO - codeparrot_training - Step 2566: {'lr': 0.0004998284816934422, 'samples': 1314304, 'steps': 2566, 'loss/train': 3.373403549194336} 02/24/2022 04:54:10 - INFO - codeparrot_training - Step 2567: {'lr': 0.0004998278751559062, 'samples': 1314816, 'steps': 2567, 'loss/train': 4.0760321617126465} 02/24/2022 04:54:13 - INFO - codeparrot_training - Step 2568: {'lr': 0.0004998272675481868, 'samples': 1315328, 'steps': 2568, 'loss/train': 4.096252918243408} 02/24/2022 04:54:19 - INFO - codeparrot_training - Step 2569: {'lr': 0.0004998266588702869, 'samples': 1315840, 'steps': 2569, 'loss/train': 2.4457826614379883} 02/24/2022 04:54:22 - INFO - codeparrot_training - Step 2570: {'lr': 0.0004998260491222088, 'samples': 1316352, 'steps': 2570, 'loss/train': 2.908329486846924} 02/24/2022 04:54:27 - INFO - codeparrot_training - Step 2571: {'lr': 0.0004998254383039552, 'samples': 1316864, 'steps': 2571, 'loss/train': 4.027587890625} 02/24/2022 04:54:31 - INFO - codeparrot_training - Step 2572: {'lr': 0.0004998248264155288, 'samples': 1317376, 'steps': 2572, 'loss/train': 3.6171653270721436} 02/24/2022 04:54:37 - INFO - codeparrot_training - Step 2573: {'lr': 0.0004998242134569322, 'samples': 1317888, 'steps': 2573, 'loss/train': 3.2733216285705566} 02/24/2022 04:54:41 - INFO - codeparrot_training - Step 2574: {'lr': 0.0004998235994281681, 'samples': 1318400, 'steps': 2574, 'loss/train': 4.285373687744141} 02/24/2022 04:54:46 - INFO - codeparrot_training - Step 2575: {'lr': 0.0004998229843292388, 'samples': 1318912, 'steps': 2575, 'loss/train': 2.5809576511383057} 02/24/2022 04:54:49 - INFO - codeparrot_training - Step 2576: {'lr': 0.0004998223681601474, 'samples': 1319424, 'steps': 2576, 'loss/train': 3.8315556049346924} 02/24/2022 04:54:55 - INFO - codeparrot_training - Step 2577: {'lr': 0.0004998217509208961, 'samples': 1319936, 'steps': 2577, 'loss/train': 3.1385366916656494} 02/24/2022 04:54:58 - INFO - codeparrot_training - Step 2578: {'lr': 0.0004998211326114878, 'samples': 1320448, 'steps': 2578, 'loss/train': 3.515986442565918} 02/24/2022 04:55:04 - INFO - codeparrot_training - Step 2579: {'lr': 0.0004998205132319252, 'samples': 1320960, 'steps': 2579, 'loss/train': 4.016100883483887} 02/24/2022 04:55:07 - INFO - codeparrot_training - Step 2580: {'lr': 0.0004998198927822108, 'samples': 1321472, 'steps': 2580, 'loss/train': 1.2040746212005615} 02/24/2022 04:55:13 - INFO - codeparrot_training - Step 2581: {'lr': 0.0004998192712623472, 'samples': 1321984, 'steps': 2581, 'loss/train': 2.934591770172119} 02/24/2022 04:55:16 - INFO - codeparrot_training - Step 2582: {'lr': 0.0004998186486723373, 'samples': 1322496, 'steps': 2582, 'loss/train': 2.857848644256592} 02/24/2022 04:55:23 - INFO - codeparrot_training - Step 2583: {'lr': 0.0004998180250121836, 'samples': 1323008, 'steps': 2583, 'loss/train': 3.1073317527770996} 02/24/2022 04:55:26 - INFO - codeparrot_training - Step 2584: {'lr': 0.0004998174002818887, 'samples': 1323520, 'steps': 2584, 'loss/train': 2.4528825283050537} 02/24/2022 04:55:32 - INFO - codeparrot_training - Step 2585: {'lr': 0.0004998167744814555, 'samples': 1324032, 'steps': 2585, 'loss/train': 3.731081962585449} 02/24/2022 04:55:35 - INFO - codeparrot_training - Step 2586: {'lr': 0.0004998161476108864, 'samples': 1324544, 'steps': 2586, 'loss/train': 6.180232048034668} 02/24/2022 04:55:40 - INFO - codeparrot_training - Step 2587: {'lr': 0.0004998155196701845, 'samples': 1325056, 'steps': 2587, 'loss/train': 3.3464818000793457} 02/24/2022 04:55:44 - INFO - codeparrot_training - Step 2588: {'lr': 0.000499814890659352, 'samples': 1325568, 'steps': 2588, 'loss/train': 3.4936163425445557} 02/24/2022 04:55:50 - INFO - codeparrot_training - Step 2589: {'lr': 0.000499814260578392, 'samples': 1326080, 'steps': 2589, 'loss/train': 3.774045944213867} 02/24/2022 04:55:53 - INFO - codeparrot_training - Step 2590: {'lr': 0.000499813629427307, 'samples': 1326592, 'steps': 2590, 'loss/train': 3.5812551975250244} 02/24/2022 04:55:59 - INFO - codeparrot_training - Step 2591: {'lr': 0.0004998129972060998, 'samples': 1327104, 'steps': 2591, 'loss/train': 2.8676490783691406} 02/24/2022 04:56:02 - INFO - codeparrot_training - Step 2592: {'lr': 0.000499812363914773, 'samples': 1327616, 'steps': 2592, 'loss/train': 2.083740472793579} 02/24/2022 04:56:09 - INFO - codeparrot_training - Step 2593: {'lr': 0.0004998117295533292, 'samples': 1328128, 'steps': 2593, 'loss/train': 3.420949935913086} 02/24/2022 04:56:12 - INFO - codeparrot_training - Step 2594: {'lr': 0.0004998110941217714, 'samples': 1328640, 'steps': 2594, 'loss/train': 3.7538506984710693} 02/24/2022 04:56:18 - INFO - codeparrot_training - Step 2595: {'lr': 0.0004998104576201022, 'samples': 1329152, 'steps': 2595, 'loss/train': 4.181107997894287} 02/24/2022 04:56:21 - INFO - codeparrot_training - Step 2596: {'lr': 0.0004998098200483243, 'samples': 1329664, 'steps': 2596, 'loss/train': 3.6615211963653564} 02/24/2022 04:56:27 - INFO - codeparrot_training - Step 2597: {'lr': 0.0004998091814064405, 'samples': 1330176, 'steps': 2597, 'loss/train': 4.06433629989624} 02/24/2022 04:56:30 - INFO - codeparrot_training - Step 2598: {'lr': 0.0004998085416944534, 'samples': 1330688, 'steps': 2598, 'loss/train': 3.4494833946228027} 02/24/2022 04:56:36 - INFO - codeparrot_training - Step 2599: {'lr': 0.000499807900912366, 'samples': 1331200, 'steps': 2599, 'loss/train': 3.337926149368286} 02/24/2022 04:56:39 - INFO - codeparrot_training - Step 2600: {'lr': 0.0004998072590601808, 'samples': 1331712, 'steps': 2600, 'loss/train': 3.3017685413360596} 02/24/2022 04:56:45 - INFO - codeparrot_training - Step 2601: {'lr': 0.0004998066161379006, 'samples': 1332224, 'steps': 2601, 'loss/train': 1.7428919076919556} 02/24/2022 04:56:48 - INFO - codeparrot_training - Step 2602: {'lr': 0.0004998059721455281, 'samples': 1332736, 'steps': 2602, 'loss/train': 3.700556755065918} 02/24/2022 04:56:54 - INFO - codeparrot_training - Step 2603: {'lr': 0.0004998053270830662, 'samples': 1333248, 'steps': 2603, 'loss/train': 3.682637929916382} 02/24/2022 04:56:57 - INFO - codeparrot_training - Step 2604: {'lr': 0.0004998046809505176, 'samples': 1333760, 'steps': 2604, 'loss/train': 2.8790128231048584} 02/24/2022 04:57:03 - INFO - codeparrot_training - Step 2605: {'lr': 0.0004998040337478851, 'samples': 1334272, 'steps': 2605, 'loss/train': 3.541307210922241} 02/24/2022 04:57:06 - INFO - codeparrot_training - Step 2606: {'lr': 0.0004998033854751715, 'samples': 1334784, 'steps': 2606, 'loss/train': 3.504415273666382} 02/24/2022 04:57:12 - INFO - codeparrot_training - Step 2607: {'lr': 0.0004998027361323794, 'samples': 1335296, 'steps': 2607, 'loss/train': 3.3908636569976807} 02/24/2022 04:57:15 - INFO - codeparrot_training - Step 2608: {'lr': 0.0004998020857195117, 'samples': 1335808, 'steps': 2608, 'loss/train': 2.3568642139434814} 02/24/2022 04:57:22 - INFO - codeparrot_training - Step 2609: {'lr': 0.0004998014342365712, 'samples': 1336320, 'steps': 2609, 'loss/train': 3.6828958988189697} 02/24/2022 04:57:25 - INFO - codeparrot_training - Step 2610: {'lr': 0.0004998007816835608, 'samples': 1336832, 'steps': 2610, 'loss/train': 3.263063669204712} 02/24/2022 04:57:31 - INFO - codeparrot_training - Step 2611: {'lr': 0.000499800128060483, 'samples': 1337344, 'steps': 2611, 'loss/train': 2.5680322647094727} 02/24/2022 04:57:34 - INFO - codeparrot_training - Step 2612: {'lr': 0.0004997994733673409, 'samples': 1337856, 'steps': 2612, 'loss/train': 3.1815738677978516} 02/24/2022 04:57:40 - INFO - codeparrot_training - Step 2613: {'lr': 0.000499798817604137, 'samples': 1338368, 'steps': 2613, 'loss/train': 1.6468865871429443} 02/24/2022 04:57:43 - INFO - codeparrot_training - Step 2614: {'lr': 0.0004997981607708745, 'samples': 1338880, 'steps': 2614, 'loss/train': 2.981415271759033} 02/24/2022 04:57:49 - INFO - codeparrot_training - Step 2615: {'lr': 0.0004997975028675558, 'samples': 1339392, 'steps': 2615, 'loss/train': 4.030995845794678} 02/24/2022 04:57:53 - INFO - codeparrot_training - Step 2616: {'lr': 0.0004997968438941841, 'samples': 1339904, 'steps': 2616, 'loss/train': 1.6613208055496216} 02/24/2022 04:57:58 - INFO - codeparrot_training - Step 2617: {'lr': 0.0004997961838507619, 'samples': 1340416, 'steps': 2617, 'loss/train': 3.0013906955718994} 02/24/2022 04:58:02 - INFO - codeparrot_training - Step 2618: {'lr': 0.0004997955227372923, 'samples': 1340928, 'steps': 2618, 'loss/train': 4.839288711547852} 02/24/2022 04:58:08 - INFO - codeparrot_training - Step 2619: {'lr': 0.000499794860553778, 'samples': 1341440, 'steps': 2619, 'loss/train': 3.9381189346313477} 02/24/2022 04:58:11 - INFO - codeparrot_training - Step 2620: {'lr': 0.0004997941973002216, 'samples': 1341952, 'steps': 2620, 'loss/train': 4.103443622589111} 02/24/2022 04:58:17 - INFO - codeparrot_training - Step 2621: {'lr': 0.0004997935329766265, 'samples': 1342464, 'steps': 2621, 'loss/train': 3.048365592956543} 02/24/2022 04:58:20 - INFO - codeparrot_training - Step 2622: {'lr': 0.000499792867582995, 'samples': 1342976, 'steps': 2622, 'loss/train': 3.6909701824188232} 02/24/2022 04:58:25 - INFO - codeparrot_training - Step 2623: {'lr': 0.0004997922011193303, 'samples': 1343488, 'steps': 2623, 'loss/train': 1.2625305652618408} 02/24/2022 04:58:29 - INFO - codeparrot_training - Step 2624: {'lr': 0.000499791533585635, 'samples': 1344000, 'steps': 2624, 'loss/train': 0.9226672649383545} 02/24/2022 04:58:34 - INFO - codeparrot_training - Step 2625: {'lr': 0.0004997908649819122, 'samples': 1344512, 'steps': 2625, 'loss/train': 3.233002185821533} 02/24/2022 04:58:38 - INFO - codeparrot_training - Step 2626: {'lr': 0.0004997901953081646, 'samples': 1345024, 'steps': 2626, 'loss/train': 1.8113363981246948} 02/24/2022 04:58:43 - INFO - codeparrot_training - Step 2627: {'lr': 0.0004997895245643951, 'samples': 1345536, 'steps': 2627, 'loss/train': 3.721475839614868} 02/24/2022 04:58:49 - INFO - codeparrot_training - Step 2628: {'lr': 0.0004997888527506067, 'samples': 1346048, 'steps': 2628, 'loss/train': 2.823287010192871} 02/24/2022 04:58:53 - INFO - codeparrot_training - Step 2629: {'lr': 0.000499788179866802, 'samples': 1346560, 'steps': 2629, 'loss/train': 3.3880221843719482} 02/24/2022 04:58:56 - INFO - codeparrot_training - Step 2630: {'lr': 0.0004997875059129843, 'samples': 1347072, 'steps': 2630, 'loss/train': 2.731090784072876} 02/24/2022 04:59:02 - INFO - codeparrot_training - Step 2631: {'lr': 0.000499786830889156, 'samples': 1347584, 'steps': 2631, 'loss/train': 3.708174228668213} 02/24/2022 04:59:06 - INFO - codeparrot_training - Step 2632: {'lr': 0.0004997861547953203, 'samples': 1348096, 'steps': 2632, 'loss/train': 3.243811845779419} 02/24/2022 04:59:11 - INFO - codeparrot_training - Step 2633: {'lr': 0.00049978547763148, 'samples': 1348608, 'steps': 2633, 'loss/train': 2.597032070159912} 02/24/2022 04:59:15 - INFO - codeparrot_training - Step 2634: {'lr': 0.0004997847993976381, 'samples': 1349120, 'steps': 2634, 'loss/train': 2.9924778938293457} 02/24/2022 04:59:20 - INFO - codeparrot_training - Step 2635: {'lr': 0.0004997841200937975, 'samples': 1349632, 'steps': 2635, 'loss/train': 4.333327770233154} 02/24/2022 04:59:24 - INFO - codeparrot_training - Step 2636: {'lr': 0.0004997834397199609, 'samples': 1350144, 'steps': 2636, 'loss/train': 3.035799741744995} 02/24/2022 04:59:29 - INFO - codeparrot_training - Step 2637: {'lr': 0.0004997827582761315, 'samples': 1350656, 'steps': 2637, 'loss/train': 4.465551853179932} 02/24/2022 04:59:33 - INFO - codeparrot_training - Step 2638: {'lr': 0.0004997820757623119, 'samples': 1351168, 'steps': 2638, 'loss/train': 3.823025703430176} 02/24/2022 04:59:38 - INFO - codeparrot_training - Step 2639: {'lr': 0.0004997813921785054, 'samples': 1351680, 'steps': 2639, 'loss/train': 4.360265254974365} 02/24/2022 04:59:42 - INFO - codeparrot_training - Step 2640: {'lr': 0.0004997807075247146, 'samples': 1352192, 'steps': 2640, 'loss/train': 2.638787269592285} 02/24/2022 04:59:48 - INFO - codeparrot_training - Step 2641: {'lr': 0.0004997800218009426, 'samples': 1352704, 'steps': 2641, 'loss/train': 2.7928121089935303} 02/24/2022 04:59:54 - INFO - codeparrot_training - Step 2642: {'lr': 0.0004997793350071923, 'samples': 1353216, 'steps': 2642, 'loss/train': 2.997230291366577} 02/24/2022 04:59:57 - INFO - codeparrot_training - Step 2643: {'lr': 0.0004997786471434666, 'samples': 1353728, 'steps': 2643, 'loss/train': 3.406938314437866} 02/24/2022 05:00:03 - INFO - codeparrot_training - Step 2644: {'lr': 0.0004997779582097686, 'samples': 1354240, 'steps': 2644, 'loss/train': 3.1807749271392822} 02/24/2022 05:00:06 - INFO - codeparrot_training - Step 2645: {'lr': 0.0004997772682061011, 'samples': 1354752, 'steps': 2645, 'loss/train': 3.620171070098877} 02/24/2022 05:00:12 - INFO - codeparrot_training - Step 2646: {'lr': 0.000499776577132467, 'samples': 1355264, 'steps': 2646, 'loss/train': 2.2439463138580322} 02/24/2022 05:00:15 - INFO - codeparrot_training - Step 2647: {'lr': 0.0004997758849888693, 'samples': 1355776, 'steps': 2647, 'loss/train': 3.4832112789154053} 02/24/2022 05:00:20 - INFO - codeparrot_training - Step 2648: {'lr': 0.0004997751917753113, 'samples': 1356288, 'steps': 2648, 'loss/train': 4.780531883239746} 02/24/2022 05:00:24 - INFO - codeparrot_training - Step 2649: {'lr': 0.0004997744974917955, 'samples': 1356800, 'steps': 2649, 'loss/train': 3.1612603664398193} 02/24/2022 05:00:29 - INFO - codeparrot_training - Step 2650: {'lr': 0.0004997738021383252, 'samples': 1357312, 'steps': 2650, 'loss/train': 3.9964981079101562} 02/24/2022 05:00:33 - INFO - codeparrot_training - Step 2651: {'lr': 0.000499773105714903, 'samples': 1357824, 'steps': 2651, 'loss/train': 2.747293472290039} 02/24/2022 05:00:38 - INFO - codeparrot_training - Step 2652: {'lr': 0.0004997724082215323, 'samples': 1358336, 'steps': 2652, 'loss/train': 2.8473522663116455} 02/24/2022 05:00:42 - INFO - codeparrot_training - Step 2653: {'lr': 0.0004997717096582159, 'samples': 1358848, 'steps': 2653, 'loss/train': 3.1151986122131348} 02/24/2022 05:00:48 - INFO - codeparrot_training - Step 2654: {'lr': 0.0004997710100249568, 'samples': 1359360, 'steps': 2654, 'loss/train': 0.9005175232887268} 02/24/2022 05:00:51 - INFO - codeparrot_training - Step 2655: {'lr': 0.000499770309321758, 'samples': 1359872, 'steps': 2655, 'loss/train': 3.2134785652160645} 02/24/2022 05:00:57 - INFO - codeparrot_training - Step 2656: {'lr': 0.0004997696075486225, 'samples': 1360384, 'steps': 2656, 'loss/train': 3.5971925258636475} 02/24/2022 05:01:00 - INFO - codeparrot_training - Step 2657: {'lr': 0.0004997689047055534, 'samples': 1360896, 'steps': 2657, 'loss/train': 3.3990397453308105} 02/24/2022 05:01:06 - INFO - codeparrot_training - Step 2658: {'lr': 0.0004997682007925535, 'samples': 1361408, 'steps': 2658, 'loss/train': 1.5455138683319092} 02/24/2022 05:01:09 - INFO - codeparrot_training - Step 2659: {'lr': 0.0004997674958096259, 'samples': 1361920, 'steps': 2659, 'loss/train': 4.938658714294434} 02/24/2022 05:01:15 - INFO - codeparrot_training - Step 2660: {'lr': 0.0004997667897567738, 'samples': 1362432, 'steps': 2660, 'loss/train': 1.8334852457046509} 02/24/2022 05:01:18 - INFO - codeparrot_training - Step 2661: {'lr': 0.000499766082634, 'samples': 1362944, 'steps': 2661, 'loss/train': 1.4658668041229248} 02/24/2022 05:01:24 - INFO - codeparrot_training - Step 2662: {'lr': 0.0004997653744413076, 'samples': 1363456, 'steps': 2662, 'loss/train': 3.3851983547210693} 02/24/2022 05:01:27 - INFO - codeparrot_training - Step 2663: {'lr': 0.0004997646651786996, 'samples': 1363968, 'steps': 2663, 'loss/train': 3.7666029930114746} 02/24/2022 05:01:34 - INFO - codeparrot_training - Step 2664: {'lr': 0.0004997639548461792, 'samples': 1364480, 'steps': 2664, 'loss/train': 1.4481109380722046} 02/24/2022 05:01:37 - INFO - codeparrot_training - Step 2665: {'lr': 0.0004997632434437493, 'samples': 1364992, 'steps': 2665, 'loss/train': 2.707970142364502} 02/24/2022 05:01:43 - INFO - codeparrot_training - Step 2666: {'lr': 0.0004997625309714129, 'samples': 1365504, 'steps': 2666, 'loss/train': 3.949827194213867} 02/24/2022 05:01:46 - INFO - codeparrot_training - Step 2667: {'lr': 0.0004997618174291732, 'samples': 1366016, 'steps': 2667, 'loss/train': 3.3576014041900635} 02/24/2022 05:01:51 - INFO - codeparrot_training - Step 2668: {'lr': 0.0004997611028170332, 'samples': 1366528, 'steps': 2668, 'loss/train': 3.399460554122925} 02/24/2022 05:01:55 - INFO - codeparrot_training - Step 2669: {'lr': 0.000499760387134996, 'samples': 1367040, 'steps': 2669, 'loss/train': 4.606263637542725} 02/24/2022 05:02:01 - INFO - codeparrot_training - Step 2670: {'lr': 0.0004997596703830645, 'samples': 1367552, 'steps': 2670, 'loss/train': 1.6423394680023193} 02/24/2022 05:02:04 - INFO - codeparrot_training - Step 2671: {'lr': 0.0004997589525612418, 'samples': 1368064, 'steps': 2671, 'loss/train': 3.123924493789673} 02/24/2022 05:02:09 - INFO - codeparrot_training - Step 2672: {'lr': 0.0004997582336695312, 'samples': 1368576, 'steps': 2672, 'loss/train': 3.1173181533813477} 02/24/2022 05:02:13 - INFO - codeparrot_training - Step 2673: {'lr': 0.0004997575137079355, 'samples': 1369088, 'steps': 2673, 'loss/train': 2.865905523300171} 02/24/2022 05:02:19 - INFO - codeparrot_training - Step 2674: {'lr': 0.0004997567926764581, 'samples': 1369600, 'steps': 2674, 'loss/train': 2.6781394481658936} 02/24/2022 05:02:22 - INFO - codeparrot_training - Step 2675: {'lr': 0.0004997560705751018, 'samples': 1370112, 'steps': 2675, 'loss/train': 2.5407214164733887} 02/24/2022 05:02:28 - INFO - codeparrot_training - Step 2676: {'lr': 0.0004997553474038698, 'samples': 1370624, 'steps': 2676, 'loss/train': 2.841752529144287} 02/24/2022 05:02:32 - INFO - codeparrot_training - Step 2677: {'lr': 0.0004997546231627652, 'samples': 1371136, 'steps': 2677, 'loss/train': 3.3782777786254883} 02/24/2022 05:02:37 - INFO - codeparrot_training - Step 2678: {'lr': 0.0004997538978517912, 'samples': 1371648, 'steps': 2678, 'loss/train': 4.038352012634277} 02/24/2022 05:02:40 - INFO - codeparrot_training - Step 2679: {'lr': 0.0004997531714709506, 'samples': 1372160, 'steps': 2679, 'loss/train': 3.6375372409820557} 02/24/2022 05:02:46 - INFO - codeparrot_training - Step 2680: {'lr': 0.0004997524440202469, 'samples': 1372672, 'steps': 2680, 'loss/train': 3.3535213470458984} 02/24/2022 05:02:49 - INFO - codeparrot_training - Step 2681: {'lr': 0.0004997517154996829, 'samples': 1373184, 'steps': 2681, 'loss/train': 3.3058862686157227} 02/24/2022 05:02:57 - INFO - codeparrot_training - Step 2682: {'lr': 0.000499750985909262, 'samples': 1373696, 'steps': 2682, 'loss/train': 2.548179864883423} 02/24/2022 05:03:01 - INFO - codeparrot_training - Step 2683: {'lr': 0.0004997502552489871, 'samples': 1374208, 'steps': 2683, 'loss/train': 3.678434371948242} 02/24/2022 05:03:06 - INFO - codeparrot_training - Step 2684: {'lr': 0.0004997495235188614, 'samples': 1374720, 'steps': 2684, 'loss/train': 0.5663205981254578} 02/24/2022 05:03:10 - INFO - codeparrot_training - Step 2685: {'lr': 0.0004997487907188881, 'samples': 1375232, 'steps': 2685, 'loss/train': 2.213472843170166} 02/24/2022 05:03:15 - INFO - codeparrot_training - Step 2686: {'lr': 0.0004997480568490702, 'samples': 1375744, 'steps': 2686, 'loss/train': 3.539717435836792} 02/24/2022 05:03:19 - INFO - codeparrot_training - Step 2687: {'lr': 0.0004997473219094111, 'samples': 1376256, 'steps': 2687, 'loss/train': 0.9934473037719727} 02/24/2022 05:03:24 - INFO - codeparrot_training - Step 2688: {'lr': 0.0004997465858999136, 'samples': 1376768, 'steps': 2688, 'loss/train': 2.6953916549682617} 02/24/2022 05:03:28 - INFO - codeparrot_training - Step 2689: {'lr': 0.0004997458488205811, 'samples': 1377280, 'steps': 2689, 'loss/train': 4.212241172790527} 02/24/2022 05:03:33 - INFO - codeparrot_training - Step 2690: {'lr': 0.0004997451106714166, 'samples': 1377792, 'steps': 2690, 'loss/train': 3.1946170330047607} 02/24/2022 05:03:37 - INFO - codeparrot_training - Step 2691: {'lr': 0.0004997443714524235, 'samples': 1378304, 'steps': 2691, 'loss/train': 1.975101351737976} 02/24/2022 05:03:44 - INFO - codeparrot_training - Step 2692: {'lr': 0.0004997436311636046, 'samples': 1378816, 'steps': 2692, 'loss/train': 2.9100875854492188} 02/24/2022 05:03:48 - INFO - codeparrot_training - Step 2693: {'lr': 0.0004997428898049635, 'samples': 1379328, 'steps': 2693, 'loss/train': 3.6551802158355713} 02/24/2022 05:03:53 - INFO - codeparrot_training - Step 2694: {'lr': 0.0004997421473765031, 'samples': 1379840, 'steps': 2694, 'loss/train': 2.9520838260650635} 02/24/2022 05:03:59 - INFO - codeparrot_training - Step 2695: {'lr': 0.0004997414038782266, 'samples': 1380352, 'steps': 2695, 'loss/train': 2.3361704349517822} 02/24/2022 05:04:02 - INFO - codeparrot_training - Step 2696: {'lr': 0.0004997406593101373, 'samples': 1380864, 'steps': 2696, 'loss/train': 3.740046501159668} 02/24/2022 05:04:08 - INFO - codeparrot_training - Step 2697: {'lr': 0.0004997399136722383, 'samples': 1381376, 'steps': 2697, 'loss/train': 1.9504197835922241} 02/24/2022 05:04:11 - INFO - codeparrot_training - Step 2698: {'lr': 0.0004997391669645327, 'samples': 1381888, 'steps': 2698, 'loss/train': 2.307004928588867} 02/24/2022 05:04:17 - INFO - codeparrot_training - Step 2699: {'lr': 0.0004997384191870239, 'samples': 1382400, 'steps': 2699, 'loss/train': 2.8918838500976562} 02/24/2022 05:04:20 - INFO - codeparrot_training - Step 2700: {'lr': 0.000499737670339715, 'samples': 1382912, 'steps': 2700, 'loss/train': 4.2642822265625} 02/24/2022 05:04:28 - INFO - codeparrot_training - Step 2701: {'lr': 0.0004997369204226093, 'samples': 1383424, 'steps': 2701, 'loss/train': 3.9701991081237793} 02/24/2022 05:04:31 - INFO - codeparrot_training - Step 2702: {'lr': 0.0004997361694357098, 'samples': 1383936, 'steps': 2702, 'loss/train': 1.7114585638046265} 02/24/2022 05:04:37 - INFO - codeparrot_training - Step 2703: {'lr': 0.00049973541737902, 'samples': 1384448, 'steps': 2703, 'loss/train': 3.1372931003570557} 02/24/2022 05:04:40 - INFO - codeparrot_training - Step 2704: {'lr': 0.0004997346642525428, 'samples': 1384960, 'steps': 2704, 'loss/train': 2.4286675453186035} 02/24/2022 05:04:46 - INFO - codeparrot_training - Step 2705: {'lr': 0.0004997339100562817, 'samples': 1385472, 'steps': 2705, 'loss/train': 3.4984283447265625} 02/24/2022 05:04:49 - INFO - codeparrot_training - Step 2706: {'lr': 0.0004997331547902398, 'samples': 1385984, 'steps': 2706, 'loss/train': 3.564594030380249} 02/24/2022 05:04:55 - INFO - codeparrot_training - Step 2707: {'lr': 0.0004997323984544204, 'samples': 1386496, 'steps': 2707, 'loss/train': 1.8964987993240356} 02/24/2022 05:04:58 - INFO - codeparrot_training - Step 2708: {'lr': 0.0004997316410488267, 'samples': 1387008, 'steps': 2708, 'loss/train': 3.7054755687713623} 02/24/2022 05:05:04 - INFO - codeparrot_training - Step 2709: {'lr': 0.0004997308825734619, 'samples': 1387520, 'steps': 2709, 'loss/train': 2.7727036476135254} 02/24/2022 05:05:07 - INFO - codeparrot_training - Step 2710: {'lr': 0.0004997301230283294, 'samples': 1388032, 'steps': 2710, 'loss/train': 4.350216865539551} 02/24/2022 05:05:15 - INFO - codeparrot_training - Step 2711: {'lr': 0.0004997293624134322, 'samples': 1388544, 'steps': 2711, 'loss/train': 3.0784571170806885} 02/24/2022 05:05:18 - INFO - codeparrot_training - Step 2712: {'lr': 0.0004997286007287738, 'samples': 1389056, 'steps': 2712, 'loss/train': 2.5586674213409424} 02/24/2022 05:05:22 - INFO - codeparrot_training - Step 2713: {'lr': 0.0004997278379743574, 'samples': 1389568, 'steps': 2713, 'loss/train': 0.2997039556503296} 02/24/2022 05:05:27 - INFO - codeparrot_training - Step 2714: {'lr': 0.0004997270741501861, 'samples': 1390080, 'steps': 2714, 'loss/train': 2.3912224769592285} 02/24/2022 05:05:33 - INFO - codeparrot_training - Step 2715: {'lr': 0.0004997263092562634, 'samples': 1390592, 'steps': 2715, 'loss/train': 2.553396463394165} 02/24/2022 05:05:36 - INFO - codeparrot_training - Step 2716: {'lr': 0.0004997255432925926, 'samples': 1391104, 'steps': 2716, 'loss/train': 2.2280917167663574} 02/24/2022 05:05:42 - INFO - codeparrot_training - Step 2717: {'lr': 0.0004997247762591766, 'samples': 1391616, 'steps': 2717, 'loss/train': 2.6942803859710693} 02/24/2022 05:05:45 - INFO - codeparrot_training - Step 2718: {'lr': 0.0004997240081560193, 'samples': 1392128, 'steps': 2718, 'loss/train': 7.449179172515869} 02/24/2022 05:05:51 - INFO - codeparrot_training - Step 2719: {'lr': 0.0004997232389831234, 'samples': 1392640, 'steps': 2719, 'loss/train': 2.6250391006469727} 02/24/2022 05:05:54 - INFO - codeparrot_training - Step 2720: {'lr': 0.0004997224687404926, 'samples': 1393152, 'steps': 2720, 'loss/train': 3.1600351333618164} 02/24/2022 05:06:00 - INFO - codeparrot_training - Step 2721: {'lr': 0.0004997216974281299, 'samples': 1393664, 'steps': 2721, 'loss/train': 3.197535276412964} 02/24/2022 05:06:03 - INFO - codeparrot_training - Step 2722: {'lr': 0.0004997209250460387, 'samples': 1394176, 'steps': 2722, 'loss/train': 2.6948647499084473} 02/24/2022 05:06:09 - INFO - codeparrot_training - Step 2723: {'lr': 0.0004997201515942225, 'samples': 1394688, 'steps': 2723, 'loss/train': 1.3640938997268677} 02/24/2022 05:06:12 - INFO - codeparrot_training - Step 2724: {'lr': 0.0004997193770726844, 'samples': 1395200, 'steps': 2724, 'loss/train': 4.225404739379883} 02/24/2022 05:06:17 - INFO - codeparrot_training - Step 2725: {'lr': 0.0004997186014814278, 'samples': 1395712, 'steps': 2725, 'loss/train': 2.874544382095337} 02/24/2022 05:06:21 - INFO - codeparrot_training - Step 2726: {'lr': 0.000499717824820456, 'samples': 1396224, 'steps': 2726, 'loss/train': 3.7581987380981445} 02/24/2022 05:06:28 - INFO - codeparrot_training - Step 2727: {'lr': 0.0004997170470897723, 'samples': 1396736, 'steps': 2727, 'loss/train': 3.2515029907226562} 02/24/2022 05:06:32 - INFO - codeparrot_training - Step 2728: {'lr': 0.0004997162682893801, 'samples': 1397248, 'steps': 2728, 'loss/train': 3.428917646408081} 02/24/2022 05:06:37 - INFO - codeparrot_training - Step 2729: {'lr': 0.0004997154884192827, 'samples': 1397760, 'steps': 2729, 'loss/train': 2.7987887859344482} 02/24/2022 05:06:41 - INFO - codeparrot_training - Step 2730: {'lr': 0.0004997147074794835, 'samples': 1398272, 'steps': 2730, 'loss/train': 2.8409271240234375} 02/24/2022 05:06:46 - INFO - codeparrot_training - Step 2731: {'lr': 0.0004997139254699856, 'samples': 1398784, 'steps': 2731, 'loss/train': 4.316927433013916} 02/24/2022 05:06:50 - INFO - codeparrot_training - Step 2732: {'lr': 0.0004997131423907927, 'samples': 1399296, 'steps': 2732, 'loss/train': 4.355013847351074} 02/24/2022 05:06:55 - INFO - codeparrot_training - Step 2733: {'lr': 0.000499712358241908, 'samples': 1399808, 'steps': 2733, 'loss/train': 3.3068785667419434} 02/24/2022 05:06:59 - INFO - codeparrot_training - Step 2734: {'lr': 0.0004997115730233349, 'samples': 1400320, 'steps': 2734, 'loss/train': 4.108318328857422} 02/24/2022 05:07:04 - INFO - codeparrot_training - Step 2735: {'lr': 0.0004997107867350765, 'samples': 1400832, 'steps': 2735, 'loss/train': 2.8441662788391113} 02/24/2022 05:07:08 - INFO - codeparrot_training - Step 2736: {'lr': 0.0004997099993771365, 'samples': 1401344, 'steps': 2736, 'loss/train': 2.534899950027466} 02/24/2022 05:07:15 - INFO - codeparrot_training - Step 2737: {'lr': 0.0004997092109495181, 'samples': 1401856, 'steps': 2737, 'loss/train': 3.0105106830596924} 02/24/2022 05:07:18 - INFO - codeparrot_training - Step 2738: {'lr': 0.0004997084214522249, 'samples': 1402368, 'steps': 2738, 'loss/train': 2.264688014984131} 02/24/2022 05:07:24 - INFO - codeparrot_training - Step 2739: {'lr': 0.0004997076308852599, 'samples': 1402880, 'steps': 2739, 'loss/train': 3.0359232425689697} 02/24/2022 05:07:27 - INFO - codeparrot_training - Step 2740: {'lr': 0.0004997068392486268, 'samples': 1403392, 'steps': 2740, 'loss/train': 1.9524973630905151} 02/24/2022 05:07:33 - INFO - codeparrot_training - Step 2741: {'lr': 0.0004997060465423288, 'samples': 1403904, 'steps': 2741, 'loss/train': 1.9212144613265991} 02/24/2022 05:07:36 - INFO - codeparrot_training - Step 2742: {'lr': 0.0004997052527663696, 'samples': 1404416, 'steps': 2742, 'loss/train': 0.3470495045185089} 02/24/2022 05:07:42 - INFO - codeparrot_training - Step 2743: {'lr': 0.0004997044579207522, 'samples': 1404928, 'steps': 2743, 'loss/train': 3.8068249225616455} 02/24/2022 05:07:46 - INFO - codeparrot_training - Step 2744: {'lr': 0.0004997036620054803, 'samples': 1405440, 'steps': 2744, 'loss/train': 3.403555393218994} 02/24/2022 05:07:51 - INFO - codeparrot_training - Step 2745: {'lr': 0.0004997028650205572, 'samples': 1405952, 'steps': 2745, 'loss/train': 9.212778091430664} 02/24/2022 05:07:54 - INFO - codeparrot_training - Step 2746: {'lr': 0.0004997020669659862, 'samples': 1406464, 'steps': 2746, 'loss/train': 2.6647017002105713} 02/24/2022 05:08:02 - INFO - codeparrot_training - Step 2747: {'lr': 0.000499701267841771, 'samples': 1406976, 'steps': 2747, 'loss/train': 3.697998046875} 02/24/2022 05:08:05 - INFO - codeparrot_training - Step 2748: {'lr': 0.0004997004676479147, 'samples': 1407488, 'steps': 2748, 'loss/train': 2.0773494243621826} 02/24/2022 05:08:11 - INFO - codeparrot_training - Step 2749: {'lr': 0.0004996996663844209, 'samples': 1408000, 'steps': 2749, 'loss/train': 1.2449995279312134} 02/24/2022 05:08:14 - INFO - codeparrot_training - Step 2750: {'lr': 0.0004996988640512931, 'samples': 1408512, 'steps': 2750, 'loss/train': 3.868187665939331} 02/24/2022 05:08:20 - INFO - codeparrot_training - Step 2751: {'lr': 0.0004996980606485346, 'samples': 1409024, 'steps': 2751, 'loss/train': 2.1135287284851074} 02/24/2022 05:08:23 - INFO - codeparrot_training - Step 2752: {'lr': 0.0004996972561761489, 'samples': 1409536, 'steps': 2752, 'loss/train': 3.0647342205047607} 02/24/2022 05:08:29 - INFO - codeparrot_training - Step 2753: {'lr': 0.0004996964506341395, 'samples': 1410048, 'steps': 2753, 'loss/train': 2.9392306804656982} 02/24/2022 05:08:32 - INFO - codeparrot_training - Step 2754: {'lr': 0.0004996956440225098, 'samples': 1410560, 'steps': 2754, 'loss/train': 4.008497714996338} 02/24/2022 05:08:38 - INFO - codeparrot_training - Step 2755: {'lr': 0.0004996948363412631, 'samples': 1411072, 'steps': 2755, 'loss/train': 2.9995925426483154} 02/24/2022 05:08:43 - INFO - codeparrot_training - Step 2756: {'lr': 0.0004996940275904031, 'samples': 1411584, 'steps': 2756, 'loss/train': 1.4026012420654297} 02/24/2022 05:08:46 - INFO - codeparrot_training - Step 2757: {'lr': 0.0004996932177699332, 'samples': 1412096, 'steps': 2757, 'loss/train': 3.455096483230591} 02/24/2022 05:08:52 - INFO - codeparrot_training - Step 2758: {'lr': 0.0004996924068798569, 'samples': 1412608, 'steps': 2758, 'loss/train': 2.144666910171509} 02/24/2022 05:08:56 - INFO - codeparrot_training - Step 2759: {'lr': 0.0004996915949201775, 'samples': 1413120, 'steps': 2759, 'loss/train': 3.213867664337158} 02/24/2022 05:09:01 - INFO - codeparrot_training - Step 2760: {'lr': 0.0004996907818908987, 'samples': 1413632, 'steps': 2760, 'loss/train': 3.3109800815582275} 02/24/2022 05:09:05 - INFO - codeparrot_training - Step 2761: {'lr': 0.0004996899677920238, 'samples': 1414144, 'steps': 2761, 'loss/train': 3.086364269256592} 02/24/2022 05:09:12 - INFO - codeparrot_training - Step 2762: {'lr': 0.0004996891526235564, 'samples': 1414656, 'steps': 2762, 'loss/train': 3.8656375408172607} 02/24/2022 05:09:15 - INFO - codeparrot_training - Step 2763: {'lr': 0.0004996883363854998, 'samples': 1415168, 'steps': 2763, 'loss/train': 2.529487133026123} 02/24/2022 05:09:21 - INFO - codeparrot_training - Step 2764: {'lr': 0.0004996875190778579, 'samples': 1415680, 'steps': 2764, 'loss/train': 1.996953010559082} 02/24/2022 05:09:24 - INFO - codeparrot_training - Step 2765: {'lr': 0.0004996867007006339, 'samples': 1416192, 'steps': 2765, 'loss/train': 3.3359246253967285} 02/24/2022 05:09:30 - INFO - codeparrot_training - Step 2766: {'lr': 0.0004996858812538312, 'samples': 1416704, 'steps': 2766, 'loss/train': 3.4091908931732178} 02/24/2022 05:09:33 - INFO - codeparrot_training - Step 2767: {'lr': 0.0004996850607374535, 'samples': 1417216, 'steps': 2767, 'loss/train': 3.3565618991851807} 02/24/2022 05:09:39 - INFO - codeparrot_training - Step 2768: {'lr': 0.0004996842391515044, 'samples': 1417728, 'steps': 2768, 'loss/train': 2.5190060138702393} 02/24/2022 05:09:42 - INFO - codeparrot_training - Step 2769: {'lr': 0.0004996834164959872, 'samples': 1418240, 'steps': 2769, 'loss/train': 3.107618570327759} 02/24/2022 05:09:48 - INFO - codeparrot_training - Step 2770: {'lr': 0.0004996825927709056, 'samples': 1418752, 'steps': 2770, 'loss/train': 3.328472852706909} 02/24/2022 05:09:51 - INFO - codeparrot_training - Step 2771: {'lr': 0.0004996817679762631, 'samples': 1419264, 'steps': 2771, 'loss/train': 3.195223808288574} 02/24/2022 05:09:58 - INFO - codeparrot_training - Step 2772: {'lr': 0.000499680942112063, 'samples': 1419776, 'steps': 2772, 'loss/train': 0.8028042316436768} 02/24/2022 05:10:02 - INFO - codeparrot_training - Step 2773: {'lr': 0.0004996801151783092, 'samples': 1420288, 'steps': 2773, 'loss/train': 3.554117202758789} 02/24/2022 05:10:07 - INFO - codeparrot_training - Step 2774: {'lr': 0.000499679287175005, 'samples': 1420800, 'steps': 2774, 'loss/train': 4.272123336791992} 02/24/2022 05:10:11 - INFO - codeparrot_training - Step 2775: {'lr': 0.000499678458102154, 'samples': 1421312, 'steps': 2775, 'loss/train': 1.5888338088989258} 02/24/2022 05:10:16 - INFO - codeparrot_training - Step 2776: {'lr': 0.0004996776279597598, 'samples': 1421824, 'steps': 2776, 'loss/train': 2.162257671356201} 02/24/2022 05:10:20 - INFO - codeparrot_training - Step 2777: {'lr': 0.0004996767967478259, 'samples': 1422336, 'steps': 2777, 'loss/train': 2.9495656490325928} 02/24/2022 05:10:25 - INFO - codeparrot_training - Step 2778: {'lr': 0.0004996759644663559, 'samples': 1422848, 'steps': 2778, 'loss/train': 2.884937286376953} 02/24/2022 05:10:29 - INFO - codeparrot_training - Step 2779: {'lr': 0.0004996751311153535, 'samples': 1423360, 'steps': 2779, 'loss/train': 3.068324565887451} 02/24/2022 05:10:34 - INFO - codeparrot_training - Step 2780: {'lr': 0.0004996742966948219, 'samples': 1423872, 'steps': 2780, 'loss/train': 2.154026985168457} 02/24/2022 05:10:38 - INFO - codeparrot_training - Step 2781: {'lr': 0.000499673461204765, 'samples': 1424384, 'steps': 2781, 'loss/train': 4.119757652282715} 02/24/2022 05:10:43 - INFO - codeparrot_training - Step 2782: {'lr': 0.0004996726246451862, 'samples': 1424896, 'steps': 2782, 'loss/train': 3.3859035968780518} 02/24/2022 05:10:47 - INFO - codeparrot_training - Step 2783: {'lr': 0.0004996717870160892, 'samples': 1425408, 'steps': 2783, 'loss/train': 2.421213388442993} 02/24/2022 05:10:54 - INFO - codeparrot_training - Step 2784: {'lr': 0.0004996709483174775, 'samples': 1425920, 'steps': 2784, 'loss/train': 2.6468234062194824} 02/24/2022 05:10:57 - INFO - codeparrot_training - Step 2785: {'lr': 0.0004996701085493547, 'samples': 1426432, 'steps': 2785, 'loss/train': 3.9776830673217773} 02/24/2022 05:11:03 - INFO - codeparrot_training - Step 2786: {'lr': 0.0004996692677117246, 'samples': 1426944, 'steps': 2786, 'loss/train': 2.395897150039673} 02/24/2022 05:11:06 - INFO - codeparrot_training - Step 2787: {'lr': 0.0004996684258045906, 'samples': 1427456, 'steps': 2787, 'loss/train': 2.4937233924865723} 02/24/2022 05:11:12 - INFO - codeparrot_training - Step 2788: {'lr': 0.0004996675828279562, 'samples': 1427968, 'steps': 2788, 'loss/train': 2.744579315185547} 02/24/2022 05:11:15 - INFO - codeparrot_training - Step 2789: {'lr': 0.0004996667387818254, 'samples': 1428480, 'steps': 2789, 'loss/train': 5.006727695465088} 02/24/2022 05:11:21 - INFO - codeparrot_training - Step 2790: {'lr': 0.0004996658936662013, 'samples': 1428992, 'steps': 2790, 'loss/train': 3.5648996829986572} 02/24/2022 05:11:24 - INFO - codeparrot_training - Step 2791: {'lr': 0.0004996650474810879, 'samples': 1429504, 'steps': 2791, 'loss/train': 3.327510356903076} 02/24/2022 05:11:30 - INFO - codeparrot_training - Step 2792: {'lr': 0.0004996642002264887, 'samples': 1430016, 'steps': 2792, 'loss/train': 3.139310836791992} 02/24/2022 05:11:33 - INFO - codeparrot_training - Step 2793: {'lr': 0.0004996633519024074, 'samples': 1430528, 'steps': 2793, 'loss/train': 3.672539710998535} 02/24/2022 05:11:41 - INFO - codeparrot_training - Step 2794: {'lr': 0.0004996625025088476, 'samples': 1431040, 'steps': 2794, 'loss/train': 3.735501766204834} 02/24/2022 05:11:44 - INFO - codeparrot_training - Step 2795: {'lr': 0.0004996616520458128, 'samples': 1431552, 'steps': 2795, 'loss/train': 3.2168426513671875} 02/24/2022 05:11:50 - INFO - codeparrot_training - Step 2796: {'lr': 0.0004996608005133068, 'samples': 1432064, 'steps': 2796, 'loss/train': 2.692532777786255} 02/24/2022 05:11:53 - INFO - codeparrot_training - Step 2797: {'lr': 0.0004996599479113333, 'samples': 1432576, 'steps': 2797, 'loss/train': 2.4207661151885986} 02/24/2022 05:11:59 - INFO - codeparrot_training - Step 2798: {'lr': 0.0004996590942398958, 'samples': 1433088, 'steps': 2798, 'loss/train': 3.100615978240967} 02/24/2022 05:12:02 - INFO - codeparrot_training - Step 2799: {'lr': 0.0004996582394989979, 'samples': 1433600, 'steps': 2799, 'loss/train': 1.690998911857605} 02/24/2022 05:12:08 - INFO - codeparrot_training - Step 2800: {'lr': 0.0004996573836886434, 'samples': 1434112, 'steps': 2800, 'loss/train': 2.2832417488098145} 02/24/2022 05:12:11 - INFO - codeparrot_training - Step 2801: {'lr': 0.0004996565268088362, 'samples': 1434624, 'steps': 2801, 'loss/train': 2.6704163551330566} 02/24/2022 05:12:17 - INFO - codeparrot_training - Step 2802: {'lr': 0.0004996556688595794, 'samples': 1435136, 'steps': 2802, 'loss/train': 1.6761455535888672} 02/24/2022 05:12:20 - INFO - codeparrot_training - Step 2803: {'lr': 0.0004996548098408772, 'samples': 1435648, 'steps': 2803, 'loss/train': 3.0169854164123535} 02/24/2022 05:12:26 - INFO - codeparrot_training - Step 2804: {'lr': 0.0004996539497527329, 'samples': 1436160, 'steps': 2804, 'loss/train': 3.3277251720428467} 02/24/2022 05:12:29 - INFO - codeparrot_training - Step 2805: {'lr': 0.0004996530885951505, 'samples': 1436672, 'steps': 2805, 'loss/train': 3.2999966144561768} 02/24/2022 05:12:35 - INFO - codeparrot_training - Step 2806: {'lr': 0.0004996522263681335, 'samples': 1437184, 'steps': 2806, 'loss/train': 3.297621965408325} 02/24/2022 05:12:38 - INFO - codeparrot_training - Step 2807: {'lr': 0.0004996513630716856, 'samples': 1437696, 'steps': 2807, 'loss/train': 3.0177981853485107} 02/24/2022 05:12:45 - INFO - codeparrot_training - Step 2808: {'lr': 0.0004996504987058105, 'samples': 1438208, 'steps': 2808, 'loss/train': 1.5601694583892822} 02/24/2022 05:12:51 - INFO - codeparrot_training - Step 2809: {'lr': 0.000499649633270512, 'samples': 1438720, 'steps': 2809, 'loss/train': 2.582554817199707} 02/24/2022 05:12:54 - INFO - codeparrot_training - Step 2810: {'lr': 0.0004996487667657938, 'samples': 1439232, 'steps': 2810, 'loss/train': 3.531783103942871} 02/24/2022 05:12:58 - INFO - codeparrot_training - Step 2811: {'lr': 0.0004996478991916595, 'samples': 1439744, 'steps': 2811, 'loss/train': 0.40621069073677063} 02/24/2022 05:13:03 - INFO - codeparrot_training - Step 2812: {'lr': 0.0004996470305481127, 'samples': 1440256, 'steps': 2812, 'loss/train': 1.916076421737671} 02/24/2022 05:13:07 - INFO - codeparrot_training - Step 2813: {'lr': 0.0004996461608351575, 'samples': 1440768, 'steps': 2813, 'loss/train': 2.6425673961639404} 02/24/2022 05:13:12 - INFO - codeparrot_training - Step 2814: {'lr': 0.0004996452900527974, 'samples': 1441280, 'steps': 2814, 'loss/train': 3.936494827270508} 02/24/2022 05:13:18 - INFO - codeparrot_training - Step 2815: {'lr': 0.0004996444182010361, 'samples': 1441792, 'steps': 2815, 'loss/train': 2.7605273723602295} 02/24/2022 05:13:21 - INFO - codeparrot_training - Step 2816: {'lr': 0.0004996435452798775, 'samples': 1442304, 'steps': 2816, 'loss/train': 2.6253724098205566} 02/24/2022 05:13:27 - INFO - codeparrot_training - Step 2817: {'lr': 0.000499642671289325, 'samples': 1442816, 'steps': 2817, 'loss/train': 1.991423487663269} 02/24/2022 05:13:30 - INFO - codeparrot_training - Step 2818: {'lr': 0.0004996417962293828, 'samples': 1443328, 'steps': 2818, 'loss/train': 1.931232213973999} 02/24/2022 05:13:38 - INFO - codeparrot_training - Step 2819: {'lr': 0.0004996409201000543, 'samples': 1443840, 'steps': 2819, 'loss/train': 1.2316335439682007} 02/24/2022 05:13:41 - INFO - codeparrot_training - Step 2820: {'lr': 0.0004996400429013434, 'samples': 1444352, 'steps': 2820, 'loss/train': 2.7481179237365723} 02/24/2022 05:13:47 - INFO - codeparrot_training - Step 2821: {'lr': 0.0004996391646332537, 'samples': 1444864, 'steps': 2821, 'loss/train': 3.186208963394165} 02/24/2022 05:13:50 - INFO - codeparrot_training - Step 2822: {'lr': 0.0004996382852957892, 'samples': 1445376, 'steps': 2822, 'loss/train': 3.020224094390869} 02/24/2022 05:13:56 - INFO - codeparrot_training - Step 2823: {'lr': 0.0004996374048889536, 'samples': 1445888, 'steps': 2823, 'loss/train': 0.28825661540031433} 02/24/2022 05:13:59 - INFO - codeparrot_training - Step 2824: {'lr': 0.0004996365234127506, 'samples': 1446400, 'steps': 2824, 'loss/train': 4.033950328826904} 02/24/2022 05:14:05 - INFO - codeparrot_training - Step 2825: {'lr': 0.000499635640867184, 'samples': 1446912, 'steps': 2825, 'loss/train': 2.2927799224853516} 02/24/2022 05:14:08 - INFO - codeparrot_training - Step 2826: {'lr': 0.0004996347572522575, 'samples': 1447424, 'steps': 2826, 'loss/train': 3.171036958694458} 02/24/2022 05:14:13 - INFO - codeparrot_training - Step 2827: {'lr': 0.000499633872567975, 'samples': 1447936, 'steps': 2827, 'loss/train': 2.999267578125} 02/24/2022 05:14:17 - INFO - codeparrot_training - Step 2828: {'lr': 0.0004996329868143404, 'samples': 1448448, 'steps': 2828, 'loss/train': 2.059232234954834} 02/24/2022 05:14:24 - INFO - codeparrot_training - Step 2829: {'lr': 0.0004996320999913572, 'samples': 1448960, 'steps': 2829, 'loss/train': 2.5450592041015625} 02/24/2022 05:14:28 - INFO - codeparrot_training - Step 2830: {'lr': 0.0004996312120990293, 'samples': 1449472, 'steps': 2830, 'loss/train': 3.0187501907348633} 02/24/2022 05:14:33 - INFO - codeparrot_training - Step 2831: {'lr': 0.0004996303231373607, 'samples': 1449984, 'steps': 2831, 'loss/train': 3.935559034347534} 02/24/2022 05:14:37 - INFO - codeparrot_training - Step 2832: {'lr': 0.000499629433106355, 'samples': 1450496, 'steps': 2832, 'loss/train': 3.5534451007843018} 02/24/2022 05:14:42 - INFO - codeparrot_training - Step 2833: {'lr': 0.000499628542006016, 'samples': 1451008, 'steps': 2833, 'loss/train': 3.1241888999938965} 02/24/2022 05:14:45 - INFO - codeparrot_training - Step 2834: {'lr': 0.0004996276498363477, 'samples': 1451520, 'steps': 2834, 'loss/train': 3.4773495197296143} 02/24/2022 05:14:51 - INFO - codeparrot_training - Step 2835: {'lr': 0.0004996267565973538, 'samples': 1452032, 'steps': 2835, 'loss/train': 3.3543875217437744} 02/24/2022 05:14:54 - INFO - codeparrot_training - Step 2836: {'lr': 0.0004996258622890381, 'samples': 1452544, 'steps': 2836, 'loss/train': 2.600266218185425} 02/24/2022 05:15:00 - INFO - codeparrot_training - Step 2837: {'lr': 0.0004996249669114045, 'samples': 1453056, 'steps': 2837, 'loss/train': 3.269315242767334} 02/24/2022 05:15:03 - INFO - codeparrot_training - Step 2838: {'lr': 0.0004996240704644568, 'samples': 1453568, 'steps': 2838, 'loss/train': 3.004892110824585} 02/24/2022 05:15:11 - INFO - codeparrot_training - Step 2839: {'lr': 0.0004996231729481989, 'samples': 1454080, 'steps': 2839, 'loss/train': 3.3889365196228027} 02/24/2022 05:15:14 - INFO - codeparrot_training - Step 2840: {'lr': 0.0004996222743626345, 'samples': 1454592, 'steps': 2840, 'loss/train': 3.0549867153167725} 02/24/2022 05:15:20 - INFO - codeparrot_training - Step 2841: {'lr': 0.0004996213747077675, 'samples': 1455104, 'steps': 2841, 'loss/train': 3.333037853240967} 02/24/2022 05:15:23 - INFO - codeparrot_training - Step 2842: {'lr': 0.0004996204739836019, 'samples': 1455616, 'steps': 2842, 'loss/train': 3.4401707649230957} 02/24/2022 05:15:29 - INFO - codeparrot_training - Step 2843: {'lr': 0.0004996195721901415, 'samples': 1456128, 'steps': 2843, 'loss/train': 2.8262827396392822} 02/24/2022 05:15:32 - INFO - codeparrot_training - Step 2844: {'lr': 0.00049961866932739, 'samples': 1456640, 'steps': 2844, 'loss/train': 4.0146684646606445} 02/24/2022 05:15:38 - INFO - codeparrot_training - Step 2845: {'lr': 0.0004996177653953514, 'samples': 1457152, 'steps': 2845, 'loss/train': 4.328563690185547} 02/24/2022 05:15:41 - INFO - codeparrot_training - Step 2846: {'lr': 0.0004996168603940296, 'samples': 1457664, 'steps': 2846, 'loss/train': 1.920601725578308} 02/24/2022 05:15:47 - INFO - codeparrot_training - Step 2847: {'lr': 0.0004996159543234285, 'samples': 1458176, 'steps': 2847, 'loss/train': 3.8002512454986572} 02/24/2022 05:15:50 - INFO - codeparrot_training - Step 2848: {'lr': 0.0004996150471835518, 'samples': 1458688, 'steps': 2848, 'loss/train': 2.7994513511657715} 02/24/2022 05:15:56 - INFO - codeparrot_training - Step 2849: {'lr': 0.0004996141389744035, 'samples': 1459200, 'steps': 2849, 'loss/train': 2.1357784271240234} 02/24/2022 05:15:59 - INFO - codeparrot_training - Step 2850: {'lr': 0.0004996132296959876, 'samples': 1459712, 'steps': 2850, 'loss/train': 2.8863883018493652} 02/24/2022 05:16:04 - INFO - codeparrot_training - Step 2851: {'lr': 0.0004996123193483076, 'samples': 1460224, 'steps': 2851, 'loss/train': 4.0175557136535645} 02/24/2022 05:16:08 - INFO - codeparrot_training - Step 2852: {'lr': 0.000499611407931368, 'samples': 1460736, 'steps': 2852, 'loss/train': 2.7112789154052734} 02/24/2022 05:16:14 - INFO - codeparrot_training - Step 2853: {'lr': 0.0004996104954451722, 'samples': 1461248, 'steps': 2853, 'loss/train': 3.1282994747161865} 02/24/2022 05:16:17 - INFO - codeparrot_training - Step 2854: {'lr': 0.0004996095818897245, 'samples': 1461760, 'steps': 2854, 'loss/train': 3.200387954711914} 02/24/2022 05:16:23 - INFO - codeparrot_training - Step 2855: {'lr': 0.0004996086672650284, 'samples': 1462272, 'steps': 2855, 'loss/train': 3.209383964538574} 02/24/2022 05:16:26 - INFO - codeparrot_training - Step 2856: {'lr': 0.0004996077515710881, 'samples': 1462784, 'steps': 2856, 'loss/train': 1.6988755464553833} 02/24/2022 05:16:32 - INFO - codeparrot_training - Step 2857: {'lr': 0.0004996068348079075, 'samples': 1463296, 'steps': 2857, 'loss/train': 5.365222454071045} 02/24/2022 05:16:35 - INFO - codeparrot_training - Step 2858: {'lr': 0.0004996059169754904, 'samples': 1463808, 'steps': 2858, 'loss/train': 4.547300338745117} 02/24/2022 05:16:41 - INFO - codeparrot_training - Step 2859: {'lr': 0.0004996049980738409, 'samples': 1464320, 'steps': 2859, 'loss/train': 2.6535141468048096} 02/24/2022 05:16:44 - INFO - codeparrot_training - Step 2860: {'lr': 0.0004996040781029629, 'samples': 1464832, 'steps': 2860, 'loss/train': 0.21506935358047485} 02/24/2022 05:16:50 - INFO - codeparrot_training - Step 2861: {'lr': 0.00049960315706286, 'samples': 1465344, 'steps': 2861, 'loss/train': 2.928163766860962} 02/24/2022 05:16:53 - INFO - codeparrot_training - Step 2862: {'lr': 0.0004996022349535367, 'samples': 1465856, 'steps': 2862, 'loss/train': 3.1556451320648193} 02/24/2022 05:16:59 - INFO - codeparrot_training - Step 2863: {'lr': 0.0004996013117749967, 'samples': 1466368, 'steps': 2863, 'loss/train': 2.578514575958252} 02/24/2022 05:17:02 - INFO - codeparrot_training - Step 2864: {'lr': 0.0004996003875272438, 'samples': 1466880, 'steps': 2864, 'loss/train': 2.881091356277466} 02/24/2022 05:17:08 - INFO - codeparrot_training - Step 2865: {'lr': 0.0004995994622102821, 'samples': 1467392, 'steps': 2865, 'loss/train': 2.6668548583984375} 02/24/2022 05:17:12 - INFO - codeparrot_training - Step 2866: {'lr': 0.0004995985358241156, 'samples': 1467904, 'steps': 2866, 'loss/train': 2.870293617248535} 02/24/2022 05:17:17 - INFO - codeparrot_training - Step 2867: {'lr': 0.0004995976083687482, 'samples': 1468416, 'steps': 2867, 'loss/train': 1.574471116065979} 02/24/2022 05:17:21 - INFO - codeparrot_training - Step 2868: {'lr': 0.000499596679844184, 'samples': 1468928, 'steps': 2868, 'loss/train': 2.1411514282226562} 02/24/2022 05:17:27 - INFO - codeparrot_training - Step 2869: {'lr': 0.0004995957502504268, 'samples': 1469440, 'steps': 2869, 'loss/train': 3.222968339920044} 02/24/2022 05:17:30 - INFO - codeparrot_training - Step 2870: {'lr': 0.0004995948195874807, 'samples': 1469952, 'steps': 2870, 'loss/train': 3.6101386547088623} 02/24/2022 05:17:36 - INFO - codeparrot_training - Step 2871: {'lr': 0.0004995938878553496, 'samples': 1470464, 'steps': 2871, 'loss/train': 4.255477428436279} 02/24/2022 05:17:39 - INFO - codeparrot_training - Step 2872: {'lr': 0.0004995929550540376, 'samples': 1470976, 'steps': 2872, 'loss/train': 2.392404556274414} 02/24/2022 05:17:45 - INFO - codeparrot_training - Step 2873: {'lr': 0.0004995920211835485, 'samples': 1471488, 'steps': 2873, 'loss/train': 2.3501245975494385} 02/24/2022 05:17:48 - INFO - codeparrot_training - Step 2874: {'lr': 0.0004995910862438866, 'samples': 1472000, 'steps': 2874, 'loss/train': 3.421520948410034} 02/24/2022 05:17:55 - INFO - codeparrot_training - Step 2875: {'lr': 0.0004995901502350556, 'samples': 1472512, 'steps': 2875, 'loss/train': 2.563455820083618} 02/24/2022 05:17:58 - INFO - codeparrot_training - Step 2876: {'lr': 0.0004995892131570598, 'samples': 1473024, 'steps': 2876, 'loss/train': 1.8721435070037842} 02/24/2022 05:18:04 - INFO - codeparrot_training - Step 2877: {'lr': 0.0004995882750099029, 'samples': 1473536, 'steps': 2877, 'loss/train': 2.436617136001587} 02/24/2022 05:18:07 - INFO - codeparrot_training - Step 2878: {'lr': 0.0004995873357935892, 'samples': 1474048, 'steps': 2878, 'loss/train': 4.313658237457275} 02/24/2022 05:18:13 - INFO - codeparrot_training - Step 2879: {'lr': 0.0004995863955081226, 'samples': 1474560, 'steps': 2879, 'loss/train': 2.886084794998169} 02/24/2022 05:18:16 - INFO - codeparrot_training - Step 2880: {'lr': 0.0004995854541535071, 'samples': 1475072, 'steps': 2880, 'loss/train': 2.8019585609436035} 02/24/2022 05:18:22 - INFO - codeparrot_training - Step 2881: {'lr': 0.0004995845117297468, 'samples': 1475584, 'steps': 2881, 'loss/train': 2.918586254119873} 02/24/2022 05:18:25 - INFO - codeparrot_training - Step 2882: {'lr': 0.0004995835682368457, 'samples': 1476096, 'steps': 2882, 'loss/train': 3.149017572402954} 02/24/2022 05:18:31 - INFO - codeparrot_training - Step 2883: {'lr': 0.0004995826236748078, 'samples': 1476608, 'steps': 2883, 'loss/train': 2.0591771602630615} 02/24/2022 05:18:34 - INFO - codeparrot_training - Step 2884: {'lr': 0.0004995816780436372, 'samples': 1477120, 'steps': 2884, 'loss/train': 3.566878318786621} 02/24/2022 05:18:40 - INFO - codeparrot_training - Step 2885: {'lr': 0.0004995807313433379, 'samples': 1477632, 'steps': 2885, 'loss/train': 4.372598171234131} 02/24/2022 05:18:44 - INFO - codeparrot_training - Step 2886: {'lr': 0.0004995797835739141, 'samples': 1478144, 'steps': 2886, 'loss/train': 1.5869191884994507} 02/24/2022 05:18:49 - INFO - codeparrot_training - Step 2887: {'lr': 0.0004995788347353697, 'samples': 1478656, 'steps': 2887, 'loss/train': 2.7778050899505615} 02/24/2022 05:18:55 - INFO - codeparrot_training - Step 2888: {'lr': 0.0004995778848277088, 'samples': 1479168, 'steps': 2888, 'loss/train': 3.9248712062835693} 02/24/2022 05:18:58 - INFO - codeparrot_training - Step 2889: {'lr': 0.0004995769338509357, 'samples': 1479680, 'steps': 2889, 'loss/train': 3.454069137573242} 02/24/2022 05:19:04 - INFO - codeparrot_training - Step 2890: {'lr': 0.000499575981805054, 'samples': 1480192, 'steps': 2890, 'loss/train': 2.604952812194824} 02/24/2022 05:19:07 - INFO - codeparrot_training - Step 2891: {'lr': 0.000499575028690068, 'samples': 1480704, 'steps': 2891, 'loss/train': 1.6018638610839844} 02/24/2022 05:19:13 - INFO - codeparrot_training - Step 2892: {'lr': 0.000499574074505982, 'samples': 1481216, 'steps': 2892, 'loss/train': 4.385003089904785} 02/24/2022 05:19:16 - INFO - codeparrot_training - Step 2893: {'lr': 0.0004995731192527999, 'samples': 1481728, 'steps': 2893, 'loss/train': 1.1551443338394165} 02/24/2022 05:19:22 - INFO - codeparrot_training - Step 2894: {'lr': 0.0004995721629305258, 'samples': 1482240, 'steps': 2894, 'loss/train': 2.4492008686065674} 02/24/2022 05:19:25 - INFO - codeparrot_training - Step 2895: {'lr': 0.0004995712055391638, 'samples': 1482752, 'steps': 2895, 'loss/train': 0.4130255877971649} 02/24/2022 05:19:30 - INFO - codeparrot_training - Step 2896: {'lr': 0.000499570247078718, 'samples': 1483264, 'steps': 2896, 'loss/train': 3.880509853363037} 02/24/2022 05:19:34 - INFO - codeparrot_training - Step 2897: {'lr': 0.0004995692875491925, 'samples': 1483776, 'steps': 2897, 'loss/train': 4.691442489624023} 02/24/2022 05:19:40 - INFO - codeparrot_training - Step 2898: {'lr': 0.0004995683269505914, 'samples': 1484288, 'steps': 2898, 'loss/train': 2.7149837017059326} 02/24/2022 05:19:43 - INFO - codeparrot_training - Step 2899: {'lr': 0.000499567365282919, 'samples': 1484800, 'steps': 2899, 'loss/train': 3.7517716884613037} 02/24/2022 05:19:49 - INFO - codeparrot_training - Step 2900: {'lr': 0.000499566402546179, 'samples': 1485312, 'steps': 2900, 'loss/train': 3.438077211380005} 02/24/2022 05:19:53 - INFO - codeparrot_training - Step 2901: {'lr': 0.0004995654387403758, 'samples': 1485824, 'steps': 2901, 'loss/train': 3.3144567012786865} 02/24/2022 05:19:58 - INFO - codeparrot_training - Step 2902: {'lr': 0.0004995644738655136, 'samples': 1486336, 'steps': 2902, 'loss/train': 2.5644874572753906} 02/24/2022 05:20:02 - INFO - codeparrot_training - Step 2903: {'lr': 0.0004995635079215965, 'samples': 1486848, 'steps': 2903, 'loss/train': 2.727362632751465} 02/24/2022 05:20:07 - INFO - codeparrot_training - Step 2904: {'lr': 0.0004995625409086285, 'samples': 1487360, 'steps': 2904, 'loss/train': 1.5900318622589111} 02/24/2022 05:20:11 - INFO - codeparrot_training - Step 2905: {'lr': 0.0004995615728266138, 'samples': 1487872, 'steps': 2905, 'loss/train': 3.285262107849121} 02/24/2022 05:20:16 - INFO - codeparrot_training - Step 2906: {'lr': 0.0004995606036755566, 'samples': 1488384, 'steps': 2906, 'loss/train': 3.105708122253418} 02/24/2022 05:20:20 - INFO - codeparrot_training - Step 2907: {'lr': 0.000499559633455461, 'samples': 1488896, 'steps': 2907, 'loss/train': 2.7195498943328857} 02/24/2022 05:20:25 - INFO - codeparrot_training - Step 2908: {'lr': 0.0004995586621663312, 'samples': 1489408, 'steps': 2908, 'loss/train': 2.7219977378845215} 02/24/2022 05:20:29 - INFO - codeparrot_training - Step 2909: {'lr': 0.0004995576898081713, 'samples': 1489920, 'steps': 2909, 'loss/train': 4.537677764892578} 02/24/2022 05:20:35 - INFO - codeparrot_training - Step 2910: {'lr': 0.0004995567163809855, 'samples': 1490432, 'steps': 2910, 'loss/train': 3.032747268676758} 02/24/2022 05:20:38 - INFO - codeparrot_training - Step 2911: {'lr': 0.000499555741884778, 'samples': 1490944, 'steps': 2911, 'loss/train': 3.2470128536224365} 02/24/2022 05:20:44 - INFO - codeparrot_training - Step 2912: {'lr': 0.000499554766319553, 'samples': 1491456, 'steps': 2912, 'loss/train': 2.6746344566345215} 02/24/2022 05:20:47 - INFO - codeparrot_training - Step 2913: {'lr': 0.0004995537896853146, 'samples': 1491968, 'steps': 2913, 'loss/train': 2.9834694862365723} 02/24/2022 05:20:53 - INFO - codeparrot_training - Step 2914: {'lr': 0.0004995528119820669, 'samples': 1492480, 'steps': 2914, 'loss/train': 2.5413784980773926} 02/24/2022 05:20:56 - INFO - codeparrot_training - Step 2915: {'lr': 0.0004995518332098143, 'samples': 1492992, 'steps': 2915, 'loss/train': 3.562969923019409} 02/24/2022 05:21:02 - INFO - codeparrot_training - Step 2916: {'lr': 0.0004995508533685608, 'samples': 1493504, 'steps': 2916, 'loss/train': 3.3154449462890625} 02/24/2022 05:21:05 - INFO - codeparrot_training - Step 2917: {'lr': 0.0004995498724583107, 'samples': 1494016, 'steps': 2917, 'loss/train': 3.7721385955810547} 02/24/2022 05:21:10 - INFO - codeparrot_training - Step 2918: {'lr': 0.0004995488904790682, 'samples': 1494528, 'steps': 2918, 'loss/train': 2.7396411895751953} 02/24/2022 05:21:14 - INFO - codeparrot_training - Step 2919: {'lr': 0.0004995479074308375, 'samples': 1495040, 'steps': 2919, 'loss/train': 3.9017858505249023} 02/24/2022 05:21:20 - INFO - codeparrot_training - Step 2920: {'lr': 0.0004995469233136228, 'samples': 1495552, 'steps': 2920, 'loss/train': 2.7082836627960205} 02/24/2022 05:21:23 - INFO - codeparrot_training - Step 2921: {'lr': 0.0004995459381274284, 'samples': 1496064, 'steps': 2921, 'loss/train': 3.10884952545166} 02/24/2022 05:21:29 - INFO - codeparrot_training - Step 2922: {'lr': 0.0004995449518722584, 'samples': 1496576, 'steps': 2922, 'loss/train': 2.602710485458374} 02/24/2022 05:21:32 - INFO - codeparrot_training - Step 2923: {'lr': 0.000499543964548117, 'samples': 1497088, 'steps': 2923, 'loss/train': 2.3507938385009766} 02/24/2022 05:21:38 - INFO - codeparrot_training - Step 2924: {'lr': 0.0004995429761550086, 'samples': 1497600, 'steps': 2924, 'loss/train': 2.5034258365631104} 02/24/2022 05:21:43 - INFO - codeparrot_training - Step 2925: {'lr': 0.0004995419866929373, 'samples': 1498112, 'steps': 2925, 'loss/train': 3.191732406616211} 02/24/2022 05:21:47 - INFO - codeparrot_training - Step 2926: {'lr': 0.0004995409961619073, 'samples': 1498624, 'steps': 2926, 'loss/train': 4.301843643188477} 02/24/2022 05:21:52 - INFO - codeparrot_training - Step 2927: {'lr': 0.0004995400045619229, 'samples': 1499136, 'steps': 2927, 'loss/train': 2.3361196517944336} 02/24/2022 05:21:56 - INFO - codeparrot_training - Step 2928: {'lr': 0.0004995390118929885, 'samples': 1499648, 'steps': 2928, 'loss/train': 2.6883368492126465} 02/24/2022 05:22:02 - INFO - codeparrot_training - Step 2929: {'lr': 0.0004995380181551081, 'samples': 1500160, 'steps': 2929, 'loss/train': 3.7777180671691895} 02/24/2022 05:22:05 - INFO - codeparrot_training - Step 2930: {'lr': 0.0004995370233482861, 'samples': 1500672, 'steps': 2930, 'loss/train': 3.621124267578125} 02/24/2022 05:22:11 - INFO - codeparrot_training - Step 2931: {'lr': 0.0004995360274725267, 'samples': 1501184, 'steps': 2931, 'loss/train': 2.3852555751800537} 02/24/2022 05:22:14 - INFO - codeparrot_training - Step 2932: {'lr': 0.0004995350305278342, 'samples': 1501696, 'steps': 2932, 'loss/train': 2.6886954307556152} 02/24/2022 05:22:20 - INFO - codeparrot_training - Step 2933: {'lr': 0.0004995340325142128, 'samples': 1502208, 'steps': 2933, 'loss/train': 3.690152883529663} 02/24/2022 05:22:23 - INFO - codeparrot_training - Step 2934: {'lr': 0.000499533033431667, 'samples': 1502720, 'steps': 2934, 'loss/train': 3.8104782104492188} 02/24/2022 05:22:29 - INFO - codeparrot_training - Step 2935: {'lr': 0.0004995320332802008, 'samples': 1503232, 'steps': 2935, 'loss/train': 2.6377949714660645} 02/24/2022 05:22:33 - INFO - codeparrot_training - Step 2936: {'lr': 0.0004995310320598187, 'samples': 1503744, 'steps': 2936, 'loss/train': 4.464433193206787} 02/24/2022 05:22:38 - INFO - codeparrot_training - Step 2937: {'lr': 0.0004995300297705248, 'samples': 1504256, 'steps': 2937, 'loss/train': 3.2768266201019287} 02/24/2022 05:22:42 - INFO - codeparrot_training - Step 2938: {'lr': 0.0004995290264123235, 'samples': 1504768, 'steps': 2938, 'loss/train': 4.1689324378967285} 02/24/2022 05:22:47 - INFO - codeparrot_training - Step 2939: {'lr': 0.0004995280219852192, 'samples': 1505280, 'steps': 2939, 'loss/train': 2.5472989082336426} 02/24/2022 05:22:51 - INFO - codeparrot_training - Step 2940: {'lr': 0.000499527016489216, 'samples': 1505792, 'steps': 2940, 'loss/train': 0.33765846490859985} 02/24/2022 05:22:57 - INFO - codeparrot_training - Step 2941: {'lr': 0.0004995260099243182, 'samples': 1506304, 'steps': 2941, 'loss/train': 2.321903944015503} 02/24/2022 05:23:00 - INFO - codeparrot_training - Step 2942: {'lr': 0.0004995250022905303, 'samples': 1506816, 'steps': 2942, 'loss/train': 4.7813334465026855} 02/24/2022 05:23:04 - INFO - codeparrot_training - Step 2943: {'lr': 0.0004995239935878565, 'samples': 1507328, 'steps': 2943, 'loss/train': 2.1477274894714355} 02/24/2022 05:23:09 - INFO - codeparrot_training - Step 2944: {'lr': 0.0004995229838163012, 'samples': 1507840, 'steps': 2944, 'loss/train': 3.494150400161743} 02/24/2022 05:23:12 - INFO - codeparrot_training - Step 2945: {'lr': 0.0004995219729758687, 'samples': 1508352, 'steps': 2945, 'loss/train': 2.7612357139587402} 02/24/2022 05:23:19 - INFO - codeparrot_training - Step 2946: {'lr': 0.0004995209610665632, 'samples': 1508864, 'steps': 2946, 'loss/train': 4.717722415924072} 02/24/2022 05:23:22 - INFO - codeparrot_training - Step 2947: {'lr': 0.0004995199480883892, 'samples': 1509376, 'steps': 2947, 'loss/train': 3.280971050262451} 02/24/2022 05:23:28 - INFO - codeparrot_training - Step 2948: {'lr': 0.0004995189340413509, 'samples': 1509888, 'steps': 2948, 'loss/train': 1.9670884609222412} 02/24/2022 05:23:31 - INFO - codeparrot_training - Step 2949: {'lr': 0.0004995179189254528, 'samples': 1510400, 'steps': 2949, 'loss/train': 3.457575798034668} 02/24/2022 05:23:37 - INFO - codeparrot_training - Step 2950: {'lr': 0.000499516902740699, 'samples': 1510912, 'steps': 2950, 'loss/train': 3.502692461013794} 02/24/2022 05:23:40 - INFO - codeparrot_training - Step 2951: {'lr': 0.0004995158854870942, 'samples': 1511424, 'steps': 2951, 'loss/train': 3.8437092304229736} 02/24/2022 05:23:46 - INFO - codeparrot_training - Step 2952: {'lr': 0.0004995148671646426, 'samples': 1511936, 'steps': 2952, 'loss/train': 1.654483675956726} 02/24/2022 05:23:49 - INFO - codeparrot_training - Step 2953: {'lr': 0.0004995138477733484, 'samples': 1512448, 'steps': 2953, 'loss/train': 4.670032024383545} 02/24/2022 05:23:55 - INFO - codeparrot_training - Step 2954: {'lr': 0.0004995128273132161, 'samples': 1512960, 'steps': 2954, 'loss/train': 3.2233850955963135} 02/24/2022 05:23:58 - INFO - codeparrot_training - Step 2955: {'lr': 0.0004995118057842502, 'samples': 1513472, 'steps': 2955, 'loss/train': 3.455089569091797} 02/24/2022 05:24:04 - INFO - codeparrot_training - Step 2956: {'lr': 0.0004995107831864549, 'samples': 1513984, 'steps': 2956, 'loss/train': 3.514132022857666} 02/24/2022 05:24:08 - INFO - codeparrot_training - Step 2957: {'lr': 0.0004995097595198346, 'samples': 1514496, 'steps': 2957, 'loss/train': 2.50236177444458} 02/24/2022 05:24:13 - INFO - codeparrot_training - Step 2958: {'lr': 0.0004995087347843938, 'samples': 1515008, 'steps': 2958, 'loss/train': 3.0795295238494873} 02/24/2022 05:24:17 - INFO - codeparrot_training - Step 2959: {'lr': 0.0004995077089801368, 'samples': 1515520, 'steps': 2959, 'loss/train': 2.86020565032959} 02/24/2022 05:24:22 - INFO - codeparrot_training - Step 2960: {'lr': 0.0004995066821070679, 'samples': 1516032, 'steps': 2960, 'loss/train': 2.777902126312256} 02/24/2022 05:24:26 - INFO - codeparrot_training - Step 2961: {'lr': 0.0004995056541651917, 'samples': 1516544, 'steps': 2961, 'loss/train': 3.6399078369140625} 02/24/2022 05:24:31 - INFO - codeparrot_training - Step 2962: {'lr': 0.0004995046251545125, 'samples': 1517056, 'steps': 2962, 'loss/train': 2.810255289077759} 02/24/2022 05:24:37 - INFO - codeparrot_training - Step 2963: {'lr': 0.0004995035950750346, 'samples': 1517568, 'steps': 2963, 'loss/train': 3.0590343475341797} 02/24/2022 05:24:40 - INFO - codeparrot_training - Step 2964: {'lr': 0.0004995025639267627, 'samples': 1518080, 'steps': 2964, 'loss/train': 4.054042339324951} 02/24/2022 05:24:46 - INFO - codeparrot_training - Step 2965: {'lr': 0.0004995015317097009, 'samples': 1518592, 'steps': 2965, 'loss/train': 2.602961778640747} 02/24/2022 05:24:50 - INFO - codeparrot_training - Step 2966: {'lr': 0.0004995004984238538, 'samples': 1519104, 'steps': 2966, 'loss/train': 2.1026241779327393} 02/24/2022 05:24:55 - INFO - codeparrot_training - Step 2967: {'lr': 0.0004994994640692258, 'samples': 1519616, 'steps': 2967, 'loss/train': 3.4214701652526855} 02/24/2022 05:24:59 - INFO - codeparrot_training - Step 2968: {'lr': 0.0004994984286458213, 'samples': 1520128, 'steps': 2968, 'loss/train': 1.6972887516021729} 02/24/2022 05:25:04 - INFO - codeparrot_training - Step 2969: {'lr': 0.0004994973921536447, 'samples': 1520640, 'steps': 2969, 'loss/train': 3.4412119388580322} 02/24/2022 05:25:08 - INFO - codeparrot_training - Step 2970: {'lr': 0.0004994963545927006, 'samples': 1521152, 'steps': 2970, 'loss/train': 2.8235559463500977} 02/24/2022 05:25:13 - INFO - codeparrot_training - Step 2971: {'lr': 0.0004994953159629934, 'samples': 1521664, 'steps': 2971, 'loss/train': 3.128722667694092} 02/24/2022 05:25:17 - INFO - codeparrot_training - Step 2972: {'lr': 0.0004994942762645274, 'samples': 1522176, 'steps': 2972, 'loss/train': 1.9544121026992798} 02/24/2022 05:25:22 - INFO - codeparrot_training - Step 2973: {'lr': 0.000499493235497307, 'samples': 1522688, 'steps': 2973, 'loss/train': 2.301903486251831} 02/24/2022 05:25:26 - INFO - codeparrot_training - Step 2974: {'lr': 0.000499492193661337, 'samples': 1523200, 'steps': 2974, 'loss/train': 2.6223721504211426} 02/24/2022 05:25:31 - INFO - codeparrot_training - Step 2975: {'lr': 0.0004994911507566216, 'samples': 1523712, 'steps': 2975, 'loss/train': 3.720458984375} 02/24/2022 05:25:35 - INFO - codeparrot_training - Step 2976: {'lr': 0.0004994901067831654, 'samples': 1524224, 'steps': 2976, 'loss/train': 3.7408065795898438} 02/24/2022 05:25:40 - INFO - codeparrot_training - Step 2977: {'lr': 0.0004994890617409728, 'samples': 1524736, 'steps': 2977, 'loss/train': 4.947940349578857} 02/24/2022 05:25:44 - INFO - codeparrot_training - Step 2978: {'lr': 0.0004994880156300482, 'samples': 1525248, 'steps': 2978, 'loss/train': 3.488401174545288} 02/24/2022 05:25:49 - INFO - codeparrot_training - Step 2979: {'lr': 0.0004994869684503962, 'samples': 1525760, 'steps': 2979, 'loss/train': 2.4275946617126465} 02/24/2022 05:25:53 - INFO - codeparrot_training - Step 2980: {'lr': 0.0004994859202020212, 'samples': 1526272, 'steps': 2980, 'loss/train': 1.637404203414917} 02/24/2022 05:26:00 - INFO - codeparrot_training - Step 2981: {'lr': 0.0004994848708849279, 'samples': 1526784, 'steps': 2981, 'loss/train': 1.961496114730835} 02/24/2022 05:26:03 - INFO - codeparrot_training - Step 2982: {'lr': 0.0004994838204991205, 'samples': 1527296, 'steps': 2982, 'loss/train': 2.5848701000213623} 02/24/2022 05:26:08 - INFO - codeparrot_training - Step 2983: {'lr': 0.0004994827690446036, 'samples': 1527808, 'steps': 2983, 'loss/train': 2.58817720413208} 02/24/2022 05:26:12 - INFO - codeparrot_training - Step 2984: {'lr': 0.0004994817165213817, 'samples': 1528320, 'steps': 2984, 'loss/train': 2.185640573501587} 02/24/2022 05:26:17 - INFO - codeparrot_training - Step 2985: {'lr': 0.0004994806629294594, 'samples': 1528832, 'steps': 2985, 'loss/train': 2.0475449562072754} 02/24/2022 05:26:21 - INFO - codeparrot_training - Step 2986: {'lr': 0.0004994796082688413, 'samples': 1529344, 'steps': 2986, 'loss/train': 2.093899965286255} 02/24/2022 05:26:26 - INFO - codeparrot_training - Step 2987: {'lr': 0.0004994785525395316, 'samples': 1529856, 'steps': 2987, 'loss/train': 2.775014877319336} 02/24/2022 05:26:30 - INFO - codeparrot_training - Step 2988: {'lr': 0.0004994774957415351, 'samples': 1530368, 'steps': 2988, 'loss/train': 4.215974807739258} 02/24/2022 05:26:35 - INFO - codeparrot_training - Step 2989: {'lr': 0.0004994764378748562, 'samples': 1530880, 'steps': 2989, 'loss/train': 3.640334129333496} 02/24/2022 05:26:39 - INFO - codeparrot_training - Step 2990: {'lr': 0.0004994753789394994, 'samples': 1531392, 'steps': 2990, 'loss/train': 1.996397614479065} 02/24/2022 05:26:45 - INFO - codeparrot_training - Step 2991: {'lr': 0.0004994743189354694, 'samples': 1531904, 'steps': 2991, 'loss/train': 2.74576735496521} 02/24/2022 05:26:49 - INFO - codeparrot_training - Step 2992: {'lr': 0.0004994732578627706, 'samples': 1532416, 'steps': 2992, 'loss/train': 2.9333362579345703} 02/24/2022 05:26:54 - INFO - codeparrot_training - Step 2993: {'lr': 0.0004994721957214076, 'samples': 1532928, 'steps': 2993, 'loss/train': 4.03203821182251} 02/24/2022 05:26:58 - INFO - codeparrot_training - Step 2994: {'lr': 0.0004994711325113849, 'samples': 1533440, 'steps': 2994, 'loss/train': 2.6782896518707275} 02/24/2022 05:27:03 - INFO - codeparrot_training - Step 2995: {'lr': 0.000499470068232707, 'samples': 1533952, 'steps': 2995, 'loss/train': 2.666483163833618} 02/24/2022 05:27:06 - INFO - codeparrot_training - Step 2996: {'lr': 0.0004994690028853787, 'samples': 1534464, 'steps': 2996, 'loss/train': 2.761693239212036} 02/24/2022 05:27:12 - INFO - codeparrot_training - Step 2997: {'lr': 0.0004994679364694043, 'samples': 1534976, 'steps': 2997, 'loss/train': 2.6543567180633545} 02/24/2022 05:27:15 - INFO - codeparrot_training - Step 2998: {'lr': 0.0004994668689847885, 'samples': 1535488, 'steps': 2998, 'loss/train': 3.5141983032226562} 02/24/2022 05:27:21 - INFO - codeparrot_training - Step 2999: {'lr': 0.0004994658004315358, 'samples': 1536000, 'steps': 2999, 'loss/train': 4.3175482749938965} 02/24/2022 05:27:21 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 05:27:38 - WARNING - huggingface_hub.repository - Several commits (3) will be pushed upstream. 02/24/2022 05:27:38 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 05:28:10 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 0998bf2..292ce9a floral-grass-11 -> floral-grass-11 02/24/2022 05:28:15 - INFO - codeparrot_training - Step 3000: {'lr': 0.0004994647308096509, 'samples': 1536512, 'steps': 3000, 'loss/train': 2.8926448822021484} 02/24/2022 05:28:21 - INFO - codeparrot_training - Step 3001: {'lr': 0.0004994636601191383, 'samples': 1537024, 'steps': 3001, 'loss/train': 2.7182507514953613} 02/24/2022 05:28:25 - INFO - codeparrot_training - Step 3002: {'lr': 0.0004994625883600025, 'samples': 1537536, 'steps': 3002, 'loss/train': 2.4522995948791504} 02/24/2022 05:28:30 - INFO - codeparrot_training - Step 3003: {'lr': 0.0004994615155322483, 'samples': 1538048, 'steps': 3003, 'loss/train': 2.787111282348633} 02/24/2022 05:28:34 - INFO - codeparrot_training - Step 3004: {'lr': 0.0004994604416358801, 'samples': 1538560, 'steps': 3004, 'loss/train': 3.58676815032959} 02/24/2022 05:28:39 - INFO - codeparrot_training - Step 3005: {'lr': 0.0004994593666709027, 'samples': 1539072, 'steps': 3005, 'loss/train': 3.154186725616455} 02/24/2022 05:28:43 - INFO - codeparrot_training - Step 3006: {'lr': 0.0004994582906373205, 'samples': 1539584, 'steps': 3006, 'loss/train': 2.457848310470581} 02/24/2022 05:28:48 - INFO - codeparrot_training - Step 3007: {'lr': 0.0004994572135351382, 'samples': 1540096, 'steps': 3007, 'loss/train': 3.2163219451904297} 02/24/2022 05:28:52 - INFO - codeparrot_training - Step 3008: {'lr': 0.0004994561353643604, 'samples': 1540608, 'steps': 3008, 'loss/train': 3.254566192626953} 02/24/2022 05:28:57 - INFO - codeparrot_training - Step 3009: {'lr': 0.0004994550561249917, 'samples': 1541120, 'steps': 3009, 'loss/train': 2.7579524517059326} 02/24/2022 05:29:01 - INFO - codeparrot_training - Step 3010: {'lr': 0.0004994539758170367, 'samples': 1541632, 'steps': 3010, 'loss/train': 3.395554542541504} 02/24/2022 05:29:06 - INFO - codeparrot_training - Step 3011: {'lr': 0.0004994528944405002, 'samples': 1542144, 'steps': 3011, 'loss/train': 3.2159061431884766} 02/24/2022 05:29:10 - INFO - codeparrot_training - Step 3012: {'lr': 0.0004994518119953867, 'samples': 1542656, 'steps': 3012, 'loss/train': 3.997093677520752} 02/24/2022 05:29:16 - INFO - codeparrot_training - Step 3013: {'lr': 0.0004994507284817009, 'samples': 1543168, 'steps': 3013, 'loss/train': 2.1546213626861572} 02/24/2022 05:29:19 - INFO - codeparrot_training - Step 3014: {'lr': 0.0004994496438994472, 'samples': 1543680, 'steps': 3014, 'loss/train': 3.6459944248199463} 02/24/2022 05:29:25 - INFO - codeparrot_training - Step 3015: {'lr': 0.0004994485582486306, 'samples': 1544192, 'steps': 3015, 'loss/train': 2.3420541286468506} 02/24/2022 05:29:30 - INFO - codeparrot_training - Step 3016: {'lr': 0.0004994474715292555, 'samples': 1544704, 'steps': 3016, 'loss/train': 3.2091407775878906} 02/24/2022 05:29:34 - INFO - codeparrot_training - Step 3017: {'lr': 0.0004994463837413268, 'samples': 1545216, 'steps': 3017, 'loss/train': 3.6281821727752686} 02/24/2022 05:29:40 - INFO - codeparrot_training - Step 3018: {'lr': 0.0004994452948848488, 'samples': 1545728, 'steps': 3018, 'loss/train': 2.7767629623413086} 02/24/2022 05:29:43 - INFO - codeparrot_training - Step 3019: {'lr': 0.0004994442049598265, 'samples': 1546240, 'steps': 3019, 'loss/train': 1.3820394277572632} 02/24/2022 05:29:47 - INFO - codeparrot_training - Step 3020: {'lr': 0.0004994431139662643, 'samples': 1546752, 'steps': 3020, 'loss/train': 2.668639898300171} 02/24/2022 05:29:52 - INFO - codeparrot_training - Step 3021: {'lr': 0.0004994420219041671, 'samples': 1547264, 'steps': 3021, 'loss/train': 4.131133079528809} 02/24/2022 05:29:56 - INFO - codeparrot_training - Step 3022: {'lr': 0.0004994409287735394, 'samples': 1547776, 'steps': 3022, 'loss/train': 3.1065754890441895} 02/24/2022 05:30:01 - INFO - codeparrot_training - Step 3023: {'lr': 0.0004994398345743861, 'samples': 1548288, 'steps': 3023, 'loss/train': 2.1964523792266846} 02/24/2022 05:30:05 - INFO - codeparrot_training - Step 3024: {'lr': 0.0004994387393067117, 'samples': 1548800, 'steps': 3024, 'loss/train': 2.6514766216278076} 02/24/2022 05:30:10 - INFO - codeparrot_training - Step 3025: {'lr': 0.0004994376429705208, 'samples': 1549312, 'steps': 3025, 'loss/train': 3.2547452449798584} 02/24/2022 05:30:14 - INFO - codeparrot_training - Step 3026: {'lr': 0.0004994365455658185, 'samples': 1549824, 'steps': 3026, 'loss/train': 3.4843621253967285} 02/24/2022 05:30:19 - INFO - codeparrot_training - Step 3027: {'lr': 0.000499435447092609, 'samples': 1550336, 'steps': 3027, 'loss/train': 2.9100887775421143} 02/24/2022 05:30:22 - INFO - codeparrot_training - Step 3028: {'lr': 0.0004994343475508974, 'samples': 1550848, 'steps': 3028, 'loss/train': 4.048941612243652} 02/24/2022 05:30:29 - INFO - codeparrot_training - Step 3029: {'lr': 0.0004994332469406882, 'samples': 1551360, 'steps': 3029, 'loss/train': 4.630547523498535} 02/24/2022 05:30:32 - INFO - codeparrot_training - Step 3030: {'lr': 0.0004994321452619863, 'samples': 1551872, 'steps': 3030, 'loss/train': 9.189669609069824} 02/24/2022 05:30:38 - INFO - codeparrot_training - Step 3031: {'lr': 0.0004994310425147962, 'samples': 1552384, 'steps': 3031, 'loss/train': 1.415884017944336} 02/24/2022 05:30:41 - INFO - codeparrot_training - Step 3032: {'lr': 0.0004994299386991227, 'samples': 1552896, 'steps': 3032, 'loss/train': 3.1991424560546875} 02/24/2022 05:30:47 - INFO - codeparrot_training - Step 3033: {'lr': 0.0004994288338149705, 'samples': 1553408, 'steps': 3033, 'loss/train': 2.432361125946045} 02/24/2022 05:30:50 - INFO - codeparrot_training - Step 3034: {'lr': 0.0004994277278623445, 'samples': 1553920, 'steps': 3034, 'loss/train': 3.262948989868164} 02/24/2022 05:30:56 - INFO - codeparrot_training - Step 3035: {'lr': 0.0004994266208412493, 'samples': 1554432, 'steps': 3035, 'loss/train': 2.9390087127685547} 02/24/2022 05:30:59 - INFO - codeparrot_training - Step 3036: {'lr': 0.0004994255127516895, 'samples': 1554944, 'steps': 3036, 'loss/train': 4.075638771057129} 02/24/2022 05:31:05 - INFO - codeparrot_training - Step 3037: {'lr': 0.0004994244035936701, 'samples': 1555456, 'steps': 3037, 'loss/train': 3.095963716506958} 02/24/2022 05:31:08 - INFO - codeparrot_training - Step 3038: {'lr': 0.0004994232933671958, 'samples': 1555968, 'steps': 3038, 'loss/train': 2.6832947731018066} 02/24/2022 05:31:15 - INFO - codeparrot_training - Step 3039: {'lr': 0.0004994221820722713, 'samples': 1556480, 'steps': 3039, 'loss/train': 0.6859299540519714} 02/24/2022 05:31:18 - INFO - codeparrot_training - Step 3040: {'lr': 0.0004994210697089013, 'samples': 1556992, 'steps': 3040, 'loss/train': 2.2006866931915283} 02/24/2022 05:31:23 - INFO - codeparrot_training - Step 3041: {'lr': 0.0004994199562770907, 'samples': 1557504, 'steps': 3041, 'loss/train': 1.9316508769989014} 02/24/2022 05:31:27 - INFO - codeparrot_training - Step 3042: {'lr': 0.0004994188417768443, 'samples': 1558016, 'steps': 3042, 'loss/train': 2.3153460025787354} 02/24/2022 05:31:32 - INFO - codeparrot_training - Step 3043: {'lr': 0.0004994177262081666, 'samples': 1558528, 'steps': 3043, 'loss/train': 2.8302969932556152} 02/24/2022 05:31:38 - INFO - codeparrot_training - Step 3044: {'lr': 0.0004994166095710626, 'samples': 1559040, 'steps': 3044, 'loss/train': 2.9031593799591064} 02/24/2022 05:31:42 - INFO - codeparrot_training - Step 3045: {'lr': 0.0004994154918655371, 'samples': 1559552, 'steps': 3045, 'loss/train': 3.1518476009368896} 02/24/2022 05:31:47 - INFO - codeparrot_training - Step 3046: {'lr': 0.0004994143730915948, 'samples': 1560064, 'steps': 3046, 'loss/train': 3.062596082687378} 02/24/2022 05:31:50 - INFO - codeparrot_training - Step 3047: {'lr': 0.0004994132532492406, 'samples': 1560576, 'steps': 3047, 'loss/train': 3.987868309020996} 02/24/2022 05:31:57 - INFO - codeparrot_training - Step 3048: {'lr': 0.0004994121323384791, 'samples': 1561088, 'steps': 3048, 'loss/train': 3.3002355098724365} 02/24/2022 05:32:01 - INFO - codeparrot_training - Step 3049: {'lr': 0.0004994110103593154, 'samples': 1561600, 'steps': 3049, 'loss/train': 2.5775294303894043} 02/24/2022 05:32:06 - INFO - codeparrot_training - Step 3050: {'lr': 0.0004994098873117539, 'samples': 1562112, 'steps': 3050, 'loss/train': 2.7640812397003174} 02/24/2022 05:32:09 - INFO - codeparrot_training - Step 3051: {'lr': 0.0004994087631957998, 'samples': 1562624, 'steps': 3051, 'loss/train': 2.525949001312256} 02/24/2022 05:32:15 - INFO - codeparrot_training - Step 3052: {'lr': 0.0004994076380114577, 'samples': 1563136, 'steps': 3052, 'loss/train': 2.823159694671631} 02/24/2022 05:32:18 - INFO - codeparrot_training - Step 3053: {'lr': 0.0004994065117587325, 'samples': 1563648, 'steps': 3053, 'loss/train': 2.5650107860565186} 02/24/2022 05:32:24 - INFO - codeparrot_training - Step 3054: {'lr': 0.0004994053844376289, 'samples': 1564160, 'steps': 3054, 'loss/train': 2.197035312652588} 02/24/2022 05:32:28 - INFO - codeparrot_training - Step 3055: {'lr': 0.000499404256048152, 'samples': 1564672, 'steps': 3055, 'loss/train': 3.0703396797180176} 02/24/2022 05:32:33 - INFO - codeparrot_training - Step 3056: {'lr': 0.0004994031265903063, 'samples': 1565184, 'steps': 3056, 'loss/train': 2.5021157264709473} 02/24/2022 05:32:36 - INFO - codeparrot_training - Step 3057: {'lr': 0.0004994019960640969, 'samples': 1565696, 'steps': 3057, 'loss/train': 3.484729528427124} 02/24/2022 05:32:43 - INFO - codeparrot_training - Step 3058: {'lr': 0.0004994008644695285, 'samples': 1566208, 'steps': 3058, 'loss/train': 3.1117043495178223} 02/24/2022 05:32:46 - INFO - codeparrot_training - Step 3059: {'lr': 0.0004993997318066061, 'samples': 1566720, 'steps': 3059, 'loss/train': 3.684542417526245} 02/24/2022 05:32:52 - INFO - codeparrot_training - Step 3060: {'lr': 0.0004993985980753342, 'samples': 1567232, 'steps': 3060, 'loss/train': 2.172954797744751} 02/24/2022 05:32:55 - INFO - codeparrot_training - Step 3061: {'lr': 0.0004993974632757181, 'samples': 1567744, 'steps': 3061, 'loss/train': 2.8374783992767334} 02/24/2022 05:33:01 - INFO - codeparrot_training - Step 3062: {'lr': 0.0004993963274077624, 'samples': 1568256, 'steps': 3062, 'loss/train': 2.33644962310791} 02/24/2022 05:33:04 - INFO - codeparrot_training - Step 3063: {'lr': 0.000499395190471472, 'samples': 1568768, 'steps': 3063, 'loss/train': 3.3544886112213135} 02/24/2022 05:33:10 - INFO - codeparrot_training - Step 3064: {'lr': 0.0004993940524668518, 'samples': 1569280, 'steps': 3064, 'loss/train': 2.6159868240356445} 02/24/2022 05:33:13 - INFO - codeparrot_training - Step 3065: {'lr': 0.0004993929133939067, 'samples': 1569792, 'steps': 3065, 'loss/train': 4.779129981994629} 02/24/2022 05:33:19 - INFO - codeparrot_training - Step 3066: {'lr': 0.0004993917732526416, 'samples': 1570304, 'steps': 3066, 'loss/train': 2.871546506881714} 02/24/2022 05:33:22 - INFO - codeparrot_training - Step 3067: {'lr': 0.0004993906320430613, 'samples': 1570816, 'steps': 3067, 'loss/train': 1.2023429870605469} 02/24/2022 05:33:26 - INFO - codeparrot_training - Step 3068: {'lr': 0.0004993894897651706, 'samples': 1571328, 'steps': 3068, 'loss/train': 3.7668938636779785} 02/24/2022 05:33:32 - INFO - codeparrot_training - Step 3069: {'lr': 0.0004993883464189747, 'samples': 1571840, 'steps': 3069, 'loss/train': 2.192412853240967} 02/24/2022 05:33:35 - INFO - codeparrot_training - Step 3070: {'lr': 0.0004993872020044781, 'samples': 1572352, 'steps': 3070, 'loss/train': 3.0252087116241455} 02/24/2022 05:33:41 - INFO - codeparrot_training - Step 3071: {'lr': 0.0004993860565216861, 'samples': 1572864, 'steps': 3071, 'loss/train': 2.8106606006622314} 02/24/2022 05:33:44 - INFO - codeparrot_training - Step 3072: {'lr': 0.0004993849099706034, 'samples': 1573376, 'steps': 3072, 'loss/train': 3.3133106231689453} 02/24/2022 05:33:50 - INFO - codeparrot_training - Step 3073: {'lr': 0.0004993837623512349, 'samples': 1573888, 'steps': 3073, 'loss/train': 2.569164276123047} 02/24/2022 05:33:53 - INFO - codeparrot_training - Step 3074: {'lr': 0.0004993826136635856, 'samples': 1574400, 'steps': 3074, 'loss/train': 1.7295336723327637} 02/24/2022 05:33:59 - INFO - codeparrot_training - Step 3075: {'lr': 0.0004993814639076602, 'samples': 1574912, 'steps': 3075, 'loss/train': 4.119635581970215} 02/24/2022 05:34:03 - INFO - codeparrot_training - Step 3076: {'lr': 0.000499380313083464, 'samples': 1575424, 'steps': 3076, 'loss/train': 2.8600194454193115} 02/24/2022 05:34:08 - INFO - codeparrot_training - Step 3077: {'lr': 0.0004993791611910017, 'samples': 1575936, 'steps': 3077, 'loss/train': 3.1440136432647705} 02/24/2022 05:34:12 - INFO - codeparrot_training - Step 3078: {'lr': 0.0004993780082302782, 'samples': 1576448, 'steps': 3078, 'loss/train': 3.172365665435791} 02/24/2022 05:34:17 - INFO - codeparrot_training - Step 3079: {'lr': 0.0004993768542012985, 'samples': 1576960, 'steps': 3079, 'loss/train': 3.4987597465515137} 02/24/2022 05:34:23 - INFO - codeparrot_training - Step 3080: {'lr': 0.0004993756991040675, 'samples': 1577472, 'steps': 3080, 'loss/train': 1.0102487802505493} 02/24/2022 05:34:26 - INFO - codeparrot_training - Step 3081: {'lr': 0.0004993745429385903, 'samples': 1577984, 'steps': 3081, 'loss/train': 3.1577799320220947} 02/24/2022 05:34:32 - INFO - codeparrot_training - Step 3082: {'lr': 0.0004993733857048717, 'samples': 1578496, 'steps': 3082, 'loss/train': 3.864769697189331} 02/24/2022 05:34:35 - INFO - codeparrot_training - Step 3083: {'lr': 0.0004993722274029167, 'samples': 1579008, 'steps': 3083, 'loss/train': 3.4096155166625977} 02/24/2022 05:34:41 - INFO - codeparrot_training - Step 3084: {'lr': 0.0004993710680327301, 'samples': 1579520, 'steps': 3084, 'loss/train': 2.480449676513672} 02/24/2022 05:34:45 - INFO - codeparrot_training - Step 3085: {'lr': 0.0004993699075943172, 'samples': 1580032, 'steps': 3085, 'loss/train': 2.4581048488616943} 02/24/2022 05:34:50 - INFO - codeparrot_training - Step 3086: {'lr': 0.0004993687460876829, 'samples': 1580544, 'steps': 3086, 'loss/train': 4.506467819213867} 02/24/2022 05:34:54 - INFO - codeparrot_training - Step 3087: {'lr': 0.0004993675835128319, 'samples': 1581056, 'steps': 3087, 'loss/train': 3.1030325889587402} 02/24/2022 05:34:59 - INFO - codeparrot_training - Step 3088: {'lr': 0.0004993664198697694, 'samples': 1581568, 'steps': 3088, 'loss/train': 2.63893461227417} 02/24/2022 05:35:03 - INFO - codeparrot_training - Step 3089: {'lr': 0.0004993652551585003, 'samples': 1582080, 'steps': 3089, 'loss/train': 2.6485581398010254} 02/24/2022 05:35:09 - INFO - codeparrot_training - Step 3090: {'lr': 0.0004993640893790298, 'samples': 1582592, 'steps': 3090, 'loss/train': 2.240339994430542} 02/24/2022 05:35:12 - INFO - codeparrot_training - Step 3091: {'lr': 0.0004993629225313625, 'samples': 1583104, 'steps': 3091, 'loss/train': 8.609245300292969} 02/24/2022 05:35:15 - INFO - codeparrot_training - Step 3092: {'lr': 0.0004993617546155037, 'samples': 1583616, 'steps': 3092, 'loss/train': 2.914644479751587} 02/24/2022 05:35:21 - INFO - codeparrot_training - Step 3093: {'lr': 0.0004993605856314584, 'samples': 1584128, 'steps': 3093, 'loss/train': 3.2939345836639404} 02/24/2022 05:35:24 - INFO - codeparrot_training - Step 3094: {'lr': 0.0004993594155792315, 'samples': 1584640, 'steps': 3094, 'loss/train': 3.5357043743133545} 02/24/2022 05:35:31 - INFO - codeparrot_training - Step 3095: {'lr': 0.000499358244458828, 'samples': 1585152, 'steps': 3095, 'loss/train': 4.278213977813721} 02/24/2022 05:35:34 - INFO - codeparrot_training - Step 3096: {'lr': 0.0004993570722702529, 'samples': 1585664, 'steps': 3096, 'loss/train': 4.091033935546875} 02/24/2022 05:35:39 - INFO - codeparrot_training - Step 3097: {'lr': 0.0004993558990135115, 'samples': 1586176, 'steps': 3097, 'loss/train': 2.089538812637329} 02/24/2022 05:35:43 - INFO - codeparrot_training - Step 3098: {'lr': 0.0004993547246886084, 'samples': 1586688, 'steps': 3098, 'loss/train': 3.41424298286438} 02/24/2022 05:35:49 - INFO - codeparrot_training - Step 3099: {'lr': 0.0004993535492955488, 'samples': 1587200, 'steps': 3099, 'loss/train': 2.994297742843628} 02/24/2022 05:35:54 - INFO - codeparrot_training - Step 3100: {'lr': 0.000499352372834338, 'samples': 1587712, 'steps': 3100, 'loss/train': 2.984280824661255} 02/24/2022 05:35:58 - INFO - codeparrot_training - Step 3101: {'lr': 0.0004993511953049807, 'samples': 1588224, 'steps': 3101, 'loss/train': 1.667233943939209} 02/24/2022 05:36:01 - INFO - codeparrot_training - Step 3102: {'lr': 0.000499350016707482, 'samples': 1588736, 'steps': 3102, 'loss/train': 1.7583197355270386} 02/24/2022 05:36:07 - INFO - codeparrot_training - Step 3103: {'lr': 0.0004993488370418471, 'samples': 1589248, 'steps': 3103, 'loss/train': 2.6337673664093018} 02/24/2022 05:36:13 - INFO - codeparrot_training - Step 3104: {'lr': 0.0004993476563080809, 'samples': 1589760, 'steps': 3104, 'loss/train': 3.095831871032715} 02/24/2022 05:36:16 - INFO - codeparrot_training - Step 3105: {'lr': 0.0004993464745061885, 'samples': 1590272, 'steps': 3105, 'loss/train': 3.2371175289154053} 02/24/2022 05:36:22 - INFO - codeparrot_training - Step 3106: {'lr': 0.0004993452916361751, 'samples': 1590784, 'steps': 3106, 'loss/train': 1.9171984195709229} 02/24/2022 05:36:25 - INFO - codeparrot_training - Step 3107: {'lr': 0.0004993441076980455, 'samples': 1591296, 'steps': 3107, 'loss/train': 2.9600372314453125} 02/24/2022 05:36:31 - INFO - codeparrot_training - Step 3108: {'lr': 0.0004993429226918051, 'samples': 1591808, 'steps': 3108, 'loss/train': 3.627802610397339} 02/24/2022 05:36:34 - INFO - codeparrot_training - Step 3109: {'lr': 0.0004993417366174586, 'samples': 1592320, 'steps': 3109, 'loss/train': 9.288830757141113} 02/24/2022 05:36:38 - INFO - codeparrot_training - Step 3110: {'lr': 0.0004993405494750113, 'samples': 1592832, 'steps': 3110, 'loss/train': 3.908906936645508} 02/24/2022 05:36:43 - INFO - codeparrot_training - Step 3111: {'lr': 0.0004993393612644683, 'samples': 1593344, 'steps': 3111, 'loss/train': 1.6921014785766602} 02/24/2022 05:36:49 - INFO - codeparrot_training - Step 3112: {'lr': 0.0004993381719858347, 'samples': 1593856, 'steps': 3112, 'loss/train': 3.225355863571167} 02/24/2022 05:36:52 - INFO - codeparrot_training - Step 3113: {'lr': 0.0004993369816391156, 'samples': 1594368, 'steps': 3113, 'loss/train': 3.1029157638549805} 02/24/2022 05:36:58 - INFO - codeparrot_training - Step 3114: {'lr': 0.0004993357902243158, 'samples': 1594880, 'steps': 3114, 'loss/train': 1.9244272708892822} 02/24/2022 05:37:01 - INFO - codeparrot_training - Step 3115: {'lr': 0.0004993345977414408, 'samples': 1595392, 'steps': 3115, 'loss/train': 1.8763879537582397} 02/24/2022 05:37:07 - INFO - codeparrot_training - Step 3116: {'lr': 0.0004993334041904957, 'samples': 1595904, 'steps': 3116, 'loss/train': 1.612082839012146} 02/24/2022 05:37:11 - INFO - codeparrot_training - Step 3117: {'lr': 0.0004993322095714853, 'samples': 1596416, 'steps': 3117, 'loss/train': 3.9887824058532715} 02/24/2022 05:37:16 - INFO - codeparrot_training - Step 3118: {'lr': 0.0004993310138844149, 'samples': 1596928, 'steps': 3118, 'loss/train': 2.430988311767578} 02/24/2022 05:37:20 - INFO - codeparrot_training - Step 3119: {'lr': 0.0004993298171292896, 'samples': 1597440, 'steps': 3119, 'loss/train': 3.412546396255493} 02/24/2022 05:37:25 - INFO - codeparrot_training - Step 3120: {'lr': 0.0004993286193061145, 'samples': 1597952, 'steps': 3120, 'loss/train': 2.7344772815704346} 02/24/2022 05:37:29 - INFO - codeparrot_training - Step 3121: {'lr': 0.0004993274204148949, 'samples': 1598464, 'steps': 3121, 'loss/train': 1.9452003240585327} 02/24/2022 05:37:34 - INFO - codeparrot_training - Step 3122: {'lr': 0.0004993262204556356, 'samples': 1598976, 'steps': 3122, 'loss/train': 4.483450412750244} 02/24/2022 05:37:38 - INFO - codeparrot_training - Step 3123: {'lr': 0.0004993250194283421, 'samples': 1599488, 'steps': 3123, 'loss/train': 2.142258882522583} 02/24/2022 05:37:43 - INFO - codeparrot_training - Step 3124: {'lr': 0.0004993238173330194, 'samples': 1600000, 'steps': 3124, 'loss/train': 3.2038655281066895} 02/24/2022 05:37:47 - INFO - codeparrot_training - Step 3125: {'lr': 0.0004993226141696725, 'samples': 1600512, 'steps': 3125, 'loss/train': 3.7841668128967285} 02/24/2022 05:37:52 - INFO - codeparrot_training - Step 3126: {'lr': 0.0004993214099383069, 'samples': 1601024, 'steps': 3126, 'loss/train': 0.7654164433479309} 02/24/2022 05:37:56 - INFO - codeparrot_training - Step 3127: {'lr': 0.0004993202046389274, 'samples': 1601536, 'steps': 3127, 'loss/train': 2.519116163253784} 02/24/2022 05:38:01 - INFO - codeparrot_training - Step 3128: {'lr': 0.0004993189982715392, 'samples': 1602048, 'steps': 3128, 'loss/train': 1.8190758228302002} 02/24/2022 05:38:05 - INFO - codeparrot_training - Step 3129: {'lr': 0.0004993177908361479, 'samples': 1602560, 'steps': 3129, 'loss/train': 0.8677091002464294} 02/24/2022 05:38:11 - INFO - codeparrot_training - Step 3130: {'lr': 0.000499316582332758, 'samples': 1603072, 'steps': 3130, 'loss/train': 3.7022597789764404} 02/24/2022 05:38:14 - INFO - codeparrot_training - Step 3131: {'lr': 0.0004993153727613753, 'samples': 1603584, 'steps': 3131, 'loss/train': 2.8413612842559814} 02/24/2022 05:38:20 - INFO - codeparrot_training - Step 3132: {'lr': 0.0004993141621220046, 'samples': 1604096, 'steps': 3132, 'loss/train': 2.6280641555786133} 02/24/2022 05:38:23 - INFO - codeparrot_training - Step 3133: {'lr': 0.0004993129504146512, 'samples': 1604608, 'steps': 3133, 'loss/train': 3.023607015609741} 02/24/2022 05:38:29 - INFO - codeparrot_training - Step 3134: {'lr': 0.0004993117376393203, 'samples': 1605120, 'steps': 3134, 'loss/train': 2.760035991668701} 02/24/2022 05:38:32 - INFO - codeparrot_training - Step 3135: {'lr': 0.000499310523796017, 'samples': 1605632, 'steps': 3135, 'loss/train': 3.1640913486480713} 02/24/2022 05:38:38 - INFO - codeparrot_training - Step 3136: {'lr': 0.0004993093088847466, 'samples': 1606144, 'steps': 3136, 'loss/train': 2.2486836910247803} 02/24/2022 05:38:41 - INFO - codeparrot_training - Step 3137: {'lr': 0.0004993080929055144, 'samples': 1606656, 'steps': 3137, 'loss/train': 3.125383138656616} 02/24/2022 05:38:47 - INFO - codeparrot_training - Step 3138: {'lr': 0.0004993068758583254, 'samples': 1607168, 'steps': 3138, 'loss/train': 1.7867794036865234} 02/24/2022 05:38:50 - INFO - codeparrot_training - Step 3139: {'lr': 0.0004993056577431849, 'samples': 1607680, 'steps': 3139, 'loss/train': 3.286060094833374} 02/24/2022 05:38:56 - INFO - codeparrot_training - Step 3140: {'lr': 0.0004993044385600982, 'samples': 1608192, 'steps': 3140, 'loss/train': 4.156434535980225} 02/24/2022 05:38:59 - INFO - codeparrot_training - Step 3141: {'lr': 0.0004993032183090704, 'samples': 1608704, 'steps': 3141, 'loss/train': 2.3331122398376465} 02/24/2022 05:39:05 - INFO - codeparrot_training - Step 3142: {'lr': 0.0004993019969901069, 'samples': 1609216, 'steps': 3142, 'loss/train': 3.361048698425293} 02/24/2022 05:39:09 - INFO - codeparrot_training - Step 3143: {'lr': 0.0004993007746032126, 'samples': 1609728, 'steps': 3143, 'loss/train': 1.409009575843811} 02/24/2022 05:39:14 - INFO - codeparrot_training - Step 3144: {'lr': 0.000499299551148393, 'samples': 1610240, 'steps': 3144, 'loss/train': 4.10382080078125} 02/24/2022 05:39:18 - INFO - codeparrot_training - Step 3145: {'lr': 0.0004992983266256533, 'samples': 1610752, 'steps': 3145, 'loss/train': 1.9379990100860596} 02/24/2022 05:39:24 - INFO - codeparrot_training - Step 3146: {'lr': 0.0004992971010349987, 'samples': 1611264, 'steps': 3146, 'loss/train': 2.8312385082244873} 02/24/2022 05:39:27 - INFO - codeparrot_training - Step 3147: {'lr': 0.0004992958743764346, 'samples': 1611776, 'steps': 3147, 'loss/train': 2.8191604614257812} 02/24/2022 05:39:33 - INFO - codeparrot_training - Step 3148: {'lr': 0.0004992946466499661, 'samples': 1612288, 'steps': 3148, 'loss/train': 3.0731639862060547} 02/24/2022 05:39:36 - INFO - codeparrot_training - Step 3149: {'lr': 0.0004992934178555984, 'samples': 1612800, 'steps': 3149, 'loss/train': 1.7696489095687866} 02/24/2022 05:39:42 - INFO - codeparrot_training - Step 3150: {'lr': 0.000499292187993337, 'samples': 1613312, 'steps': 3150, 'loss/train': 3.066833019256592} 02/24/2022 05:39:46 - INFO - codeparrot_training - Step 3151: {'lr': 0.0004992909570631868, 'samples': 1613824, 'steps': 3151, 'loss/train': 3.392374038696289} 02/24/2022 05:39:51 - INFO - codeparrot_training - Step 3152: {'lr': 0.0004992897250651535, 'samples': 1614336, 'steps': 3152, 'loss/train': 0.44543054699897766} 02/24/2022 05:39:55 - INFO - codeparrot_training - Step 3153: {'lr': 0.0004992884919992421, 'samples': 1614848, 'steps': 3153, 'loss/train': 3.2773361206054688} 02/24/2022 05:40:00 - INFO - codeparrot_training - Step 3154: {'lr': 0.000499287257865458, 'samples': 1615360, 'steps': 3154, 'loss/train': 3.7567567825317383} 02/24/2022 05:40:04 - INFO - codeparrot_training - Step 3155: {'lr': 0.0004992860226638064, 'samples': 1615872, 'steps': 3155, 'loss/train': 2.750204563140869} 02/24/2022 05:40:09 - INFO - codeparrot_training - Step 3156: {'lr': 0.0004992847863942927, 'samples': 1616384, 'steps': 3156, 'loss/train': 1.7656526565551758} 02/24/2022 05:40:13 - INFO - codeparrot_training - Step 3157: {'lr': 0.000499283549056922, 'samples': 1616896, 'steps': 3157, 'loss/train': 3.4308788776397705} 02/24/2022 05:40:18 - INFO - codeparrot_training - Step 3158: {'lr': 0.0004992823106516999, 'samples': 1617408, 'steps': 3158, 'loss/train': 2.693948745727539} 02/24/2022 05:40:22 - INFO - codeparrot_training - Step 3159: {'lr': 0.0004992810711786314, 'samples': 1617920, 'steps': 3159, 'loss/train': 0.32078981399536133} 02/24/2022 05:40:27 - INFO - codeparrot_training - Step 3160: {'lr': 0.000499279830637722, 'samples': 1618432, 'steps': 3160, 'loss/train': 3.464844226837158} 02/24/2022 05:40:31 - INFO - codeparrot_training - Step 3161: {'lr': 0.000499278589028977, 'samples': 1618944, 'steps': 3161, 'loss/train': 3.604499578475952} 02/24/2022 05:40:36 - INFO - codeparrot_training - Step 3162: {'lr': 0.0004992773463524016, 'samples': 1619456, 'steps': 3162, 'loss/train': 2.6285791397094727} 02/24/2022 05:40:40 - INFO - codeparrot_training - Step 3163: {'lr': 0.0004992761026080013, 'samples': 1619968, 'steps': 3163, 'loss/train': 2.7477588653564453} 02/24/2022 05:40:45 - INFO - codeparrot_training - Step 3164: {'lr': 0.0004992748577957812, 'samples': 1620480, 'steps': 3164, 'loss/train': 3.1901142597198486} 02/24/2022 05:40:49 - INFO - codeparrot_training - Step 3165: {'lr': 0.0004992736119157469, 'samples': 1620992, 'steps': 3165, 'loss/train': 3.7214250564575195} 02/24/2022 05:40:55 - INFO - codeparrot_training - Step 3166: {'lr': 0.0004992723649679035, 'samples': 1621504, 'steps': 3166, 'loss/train': 4.488134860992432} 02/24/2022 05:40:58 - INFO - codeparrot_training - Step 3167: {'lr': 0.0004992711169522565, 'samples': 1622016, 'steps': 3167, 'loss/train': 2.0717387199401855} 02/24/2022 05:41:04 - INFO - codeparrot_training - Step 3168: {'lr': 0.0004992698678688111, 'samples': 1622528, 'steps': 3168, 'loss/train': 3.143423557281494} 02/24/2022 05:41:08 - INFO - codeparrot_training - Step 3169: {'lr': 0.0004992686177175728, 'samples': 1623040, 'steps': 3169, 'loss/train': 7.139700412750244} 02/24/2022 05:41:13 - INFO - codeparrot_training - Step 3170: {'lr': 0.000499267366498547, 'samples': 1623552, 'steps': 3170, 'loss/train': 3.5356225967407227} 02/24/2022 05:41:16 - INFO - codeparrot_training - Step 3171: {'lr': 0.0004992661142117388, 'samples': 1624064, 'steps': 3171, 'loss/train': 2.750394105911255} 02/24/2022 05:41:22 - INFO - codeparrot_training - Step 3172: {'lr': 0.0004992648608571537, 'samples': 1624576, 'steps': 3172, 'loss/train': 1.6689268350601196} 02/24/2022 05:41:25 - INFO - codeparrot_training - Step 3173: {'lr': 0.0004992636064347971, 'samples': 1625088, 'steps': 3173, 'loss/train': 2.4697742462158203} 02/24/2022 05:41:31 - INFO - codeparrot_training - Step 3174: {'lr': 0.0004992623509446746, 'samples': 1625600, 'steps': 3174, 'loss/train': 2.9983716011047363} 02/24/2022 05:41:34 - INFO - codeparrot_training - Step 3175: {'lr': 0.0004992610943867911, 'samples': 1626112, 'steps': 3175, 'loss/train': 2.3993630409240723} 02/24/2022 05:41:40 - INFO - codeparrot_training - Step 3176: {'lr': 0.0004992598367611523, 'samples': 1626624, 'steps': 3176, 'loss/train': 2.1723434925079346} 02/24/2022 05:41:43 - INFO - codeparrot_training - Step 3177: {'lr': 0.0004992585780677634, 'samples': 1627136, 'steps': 3177, 'loss/train': 1.7839462757110596} 02/24/2022 05:41:50 - INFO - codeparrot_training - Step 3178: {'lr': 0.00049925731830663, 'samples': 1627648, 'steps': 3178, 'loss/train': 2.7818543910980225} 02/24/2022 05:41:53 - INFO - codeparrot_training - Step 3179: {'lr': 0.0004992560574777574, 'samples': 1628160, 'steps': 3179, 'loss/train': 3.399399757385254} 02/24/2022 05:41:59 - INFO - codeparrot_training - Step 3180: {'lr': 0.000499254795581151, 'samples': 1628672, 'steps': 3180, 'loss/train': 1.4273549318313599} 02/24/2022 05:42:02 - INFO - codeparrot_training - Step 3181: {'lr': 0.0004992535326168162, 'samples': 1629184, 'steps': 3181, 'loss/train': 2.8285038471221924} 02/24/2022 05:42:08 - INFO - codeparrot_training - Step 3182: {'lr': 0.0004992522685847583, 'samples': 1629696, 'steps': 3182, 'loss/train': 1.829829454421997} 02/24/2022 05:42:11 - INFO - codeparrot_training - Step 3183: {'lr': 0.000499251003484983, 'samples': 1630208, 'steps': 3183, 'loss/train': 3.3949851989746094} 02/24/2022 05:42:17 - INFO - codeparrot_training - Step 3184: {'lr': 0.0004992497373174955, 'samples': 1630720, 'steps': 3184, 'loss/train': 1.1417018175125122} 02/24/2022 05:42:20 - INFO - codeparrot_training - Step 3185: {'lr': 0.0004992484700823012, 'samples': 1631232, 'steps': 3185, 'loss/train': 3.5801939964294434} 02/24/2022 05:42:25 - INFO - codeparrot_training - Step 3186: {'lr': 0.0004992472017794057, 'samples': 1631744, 'steps': 3186, 'loss/train': 2.711984872817993} 02/24/2022 05:42:29 - INFO - codeparrot_training - Step 3187: {'lr': 0.0004992459324088143, 'samples': 1632256, 'steps': 3187, 'loss/train': 2.6519556045532227} 02/24/2022 05:42:35 - INFO - codeparrot_training - Step 3188: {'lr': 0.0004992446619705324, 'samples': 1632768, 'steps': 3188, 'loss/train': 3.6207399368286133} 02/24/2022 05:42:38 - INFO - codeparrot_training - Step 3189: {'lr': 0.0004992433904645654, 'samples': 1633280, 'steps': 3189, 'loss/train': 3.226090431213379} 02/24/2022 05:42:44 - INFO - codeparrot_training - Step 3190: {'lr': 0.0004992421178909191, 'samples': 1633792, 'steps': 3190, 'loss/train': 2.890526294708252} 02/24/2022 05:42:47 - INFO - codeparrot_training - Step 3191: {'lr': 0.0004992408442495986, 'samples': 1634304, 'steps': 3191, 'loss/train': 3.212261199951172} 02/24/2022 05:42:53 - INFO - codeparrot_training - Step 3192: {'lr': 0.0004992395695406095, 'samples': 1634816, 'steps': 3192, 'loss/train': 2.970785617828369} 02/24/2022 05:42:56 - INFO - codeparrot_training - Step 3193: {'lr': 0.0004992382937639572, 'samples': 1635328, 'steps': 3193, 'loss/train': 2.744068145751953} 02/24/2022 05:43:02 - INFO - codeparrot_training - Step 3194: {'lr': 0.0004992370169196472, 'samples': 1635840, 'steps': 3194, 'loss/train': 2.8357691764831543} 02/24/2022 05:43:07 - INFO - codeparrot_training - Step 3195: {'lr': 0.000499235739007685, 'samples': 1636352, 'steps': 3195, 'loss/train': 1.7496614456176758} 02/24/2022 05:43:11 - INFO - codeparrot_training - Step 3196: {'lr': 0.000499234460028076, 'samples': 1636864, 'steps': 3196, 'loss/train': 3.583536148071289} 02/24/2022 05:43:17 - INFO - codeparrot_training - Step 3197: {'lr': 0.0004992331799808258, 'samples': 1637376, 'steps': 3197, 'loss/train': 2.757659435272217} 02/24/2022 05:43:21 - INFO - codeparrot_training - Step 3198: {'lr': 0.0004992318988659396, 'samples': 1637888, 'steps': 3198, 'loss/train': 0.2308320850133896} 02/24/2022 05:43:24 - INFO - codeparrot_training - Step 3199: {'lr': 0.0004992306166834232, 'samples': 1638400, 'steps': 3199, 'loss/train': 3.5532004833221436} 02/24/2022 05:43:30 - INFO - codeparrot_training - Step 3200: {'lr': 0.000499229333433282, 'samples': 1638912, 'steps': 3200, 'loss/train': 3.832852363586426} 02/24/2022 05:43:33 - INFO - codeparrot_training - Step 3201: {'lr': 0.0004992280491155214, 'samples': 1639424, 'steps': 3201, 'loss/train': 3.0533409118652344} 02/24/2022 05:43:39 - INFO - codeparrot_training - Step 3202: {'lr': 0.0004992267637301471, 'samples': 1639936, 'steps': 3202, 'loss/train': 2.398283004760742} 02/24/2022 05:43:42 - INFO - codeparrot_training - Step 3203: {'lr': 0.0004992254772771644, 'samples': 1640448, 'steps': 3203, 'loss/train': 1.2057732343673706} 02/24/2022 05:43:48 - INFO - codeparrot_training - Step 3204: {'lr': 0.0004992241897565789, 'samples': 1640960, 'steps': 3204, 'loss/train': 2.909679889678955} 02/24/2022 05:43:51 - INFO - codeparrot_training - Step 3205: {'lr': 0.0004992229011683961, 'samples': 1641472, 'steps': 3205, 'loss/train': 3.1105668544769287} 02/24/2022 05:43:57 - INFO - codeparrot_training - Step 3206: {'lr': 0.0004992216115126216, 'samples': 1641984, 'steps': 3206, 'loss/train': 2.3970987796783447} 02/24/2022 05:44:00 - INFO - codeparrot_training - Step 3207: {'lr': 0.0004992203207892607, 'samples': 1642496, 'steps': 3207, 'loss/train': 2.6902811527252197} 02/24/2022 05:44:06 - INFO - codeparrot_training - Step 3208: {'lr': 0.0004992190289983192, 'samples': 1643008, 'steps': 3208, 'loss/train': 2.374669075012207} 02/24/2022 05:44:12 - INFO - codeparrot_training - Step 3209: {'lr': 0.0004992177361398026, 'samples': 1643520, 'steps': 3209, 'loss/train': 5.70389986038208} 02/24/2022 05:44:15 - INFO - codeparrot_training - Step 3210: {'lr': 0.0004992164422137162, 'samples': 1644032, 'steps': 3210, 'loss/train': 2.85617733001709} 02/24/2022 05:44:21 - INFO - codeparrot_training - Step 3211: {'lr': 0.0004992151472200657, 'samples': 1644544, 'steps': 3211, 'loss/train': 3.1170172691345215} 02/24/2022 05:44:24 - INFO - codeparrot_training - Step 3212: {'lr': 0.0004992138511588567, 'samples': 1645056, 'steps': 3212, 'loss/train': 3.403440475463867} 02/24/2022 05:44:30 - INFO - codeparrot_training - Step 3213: {'lr': 0.0004992125540300947, 'samples': 1645568, 'steps': 3213, 'loss/train': 2.5705549716949463} 02/24/2022 05:44:33 - INFO - codeparrot_training - Step 3214: {'lr': 0.0004992112558337852, 'samples': 1646080, 'steps': 3214, 'loss/train': 2.6941282749176025} 02/24/2022 05:44:39 - INFO - codeparrot_training - Step 3215: {'lr': 0.0004992099565699339, 'samples': 1646592, 'steps': 3215, 'loss/train': 4.322793960571289} 02/24/2022 05:44:42 - INFO - codeparrot_training - Step 3216: {'lr': 0.0004992086562385462, 'samples': 1647104, 'steps': 3216, 'loss/train': 3.3275146484375} 02/24/2022 05:44:48 - INFO - codeparrot_training - Step 3217: {'lr': 0.0004992073548396277, 'samples': 1647616, 'steps': 3217, 'loss/train': 3.8183724880218506} 02/24/2022 05:44:51 - INFO - codeparrot_training - Step 3218: {'lr': 0.0004992060523731842, 'samples': 1648128, 'steps': 3218, 'loss/train': 1.4780527353286743} 02/24/2022 05:44:56 - INFO - codeparrot_training - Step 3219: {'lr': 0.0004992047488392209, 'samples': 1648640, 'steps': 3219, 'loss/train': 2.7397172451019287} 02/24/2022 05:45:00 - INFO - codeparrot_training - Step 3220: {'lr': 0.0004992034442377437, 'samples': 1649152, 'steps': 3220, 'loss/train': 0.3047305941581726} 02/24/2022 05:45:07 - INFO - codeparrot_training - Step 3221: {'lr': 0.0004992021385687582, 'samples': 1649664, 'steps': 3221, 'loss/train': 3.753160238265991} 02/24/2022 05:45:10 - INFO - codeparrot_training - Step 3222: {'lr': 0.0004992008318322697, 'samples': 1650176, 'steps': 3222, 'loss/train': 5.040472030639648} 02/24/2022 05:45:16 - INFO - codeparrot_training - Step 3223: {'lr': 0.000499199524028284, 'samples': 1650688, 'steps': 3223, 'loss/train': 2.1115314960479736} 02/24/2022 05:45:19 - INFO - codeparrot_training - Step 3224: {'lr': 0.0004991982151568066, 'samples': 1651200, 'steps': 3224, 'loss/train': 2.7115297317504883} 02/24/2022 05:45:25 - INFO - codeparrot_training - Step 3225: {'lr': 0.0004991969052178433, 'samples': 1651712, 'steps': 3225, 'loss/train': 3.3633506298065186} 02/24/2022 05:45:28 - INFO - codeparrot_training - Step 3226: {'lr': 0.0004991955942113995, 'samples': 1652224, 'steps': 3226, 'loss/train': 2.8467767238616943} 02/24/2022 05:45:34 - INFO - codeparrot_training - Step 3227: {'lr': 0.0004991942821374809, 'samples': 1652736, 'steps': 3227, 'loss/train': 2.220717191696167} 02/24/2022 05:45:37 - INFO - codeparrot_training - Step 3228: {'lr': 0.0004991929689960932, 'samples': 1653248, 'steps': 3228, 'loss/train': 3.3960371017456055} 02/24/2022 05:45:43 - INFO - codeparrot_training - Step 3229: {'lr': 0.000499191654787242, 'samples': 1653760, 'steps': 3229, 'loss/train': 2.105057954788208} 02/24/2022 05:45:46 - INFO - codeparrot_training - Step 3230: {'lr': 0.0004991903395109328, 'samples': 1654272, 'steps': 3230, 'loss/train': 3.1729471683502197} 02/24/2022 05:45:53 - INFO - codeparrot_training - Step 3231: {'lr': 0.0004991890231671712, 'samples': 1654784, 'steps': 3231, 'loss/train': 1.2637964487075806} 02/24/2022 05:45:56 - INFO - codeparrot_training - Step 3232: {'lr': 0.0004991877057559631, 'samples': 1655296, 'steps': 3232, 'loss/train': 2.351189136505127} 02/24/2022 05:46:02 - INFO - codeparrot_training - Step 3233: {'lr': 0.0004991863872773139, 'samples': 1655808, 'steps': 3233, 'loss/train': 1.7066203355789185} 02/24/2022 05:46:05 - INFO - codeparrot_training - Step 3234: {'lr': 0.0004991850677312295, 'samples': 1656320, 'steps': 3234, 'loss/train': 1.730821967124939} 02/24/2022 05:46:11 - INFO - codeparrot_training - Step 3235: {'lr': 0.0004991837471177152, 'samples': 1656832, 'steps': 3235, 'loss/train': 3.146332263946533} 02/24/2022 05:46:14 - INFO - codeparrot_training - Step 3236: {'lr': 0.000499182425436777, 'samples': 1657344, 'steps': 3236, 'loss/train': 3.5814104080200195} 02/24/2022 05:46:20 - INFO - codeparrot_training - Step 3237: {'lr': 0.0004991811026884203, 'samples': 1657856, 'steps': 3237, 'loss/train': 2.492323398590088} 02/24/2022 05:46:23 - INFO - codeparrot_training - Step 3238: {'lr': 0.0004991797788726509, 'samples': 1658368, 'steps': 3238, 'loss/train': 2.071268081665039} 02/24/2022 05:46:29 - INFO - codeparrot_training - Step 3239: {'lr': 0.0004991784539894745, 'samples': 1658880, 'steps': 3239, 'loss/train': 2.5763838291168213} 02/24/2022 05:46:32 - INFO - codeparrot_training - Step 3240: {'lr': 0.0004991771280388967, 'samples': 1659392, 'steps': 3240, 'loss/train': 3.325284957885742} 02/24/2022 05:46:39 - INFO - codeparrot_training - Step 3241: {'lr': 0.0004991758010209232, 'samples': 1659904, 'steps': 3241, 'loss/train': 1.9424529075622559} 02/24/2022 05:46:42 - INFO - codeparrot_training - Step 3242: {'lr': 0.0004991744729355598, 'samples': 1660416, 'steps': 3242, 'loss/train': 3.8290340900421143} 02/24/2022 05:46:48 - INFO - codeparrot_training - Step 3243: {'lr': 0.0004991731437828119, 'samples': 1660928, 'steps': 3243, 'loss/train': 2.1688547134399414} 02/24/2022 05:46:51 - INFO - codeparrot_training - Step 3244: {'lr': 0.0004991718135626855, 'samples': 1661440, 'steps': 3244, 'loss/train': 2.0057246685028076} 02/24/2022 05:46:57 - INFO - codeparrot_training - Step 3245: {'lr': 0.0004991704822751861, 'samples': 1661952, 'steps': 3245, 'loss/train': 2.5122742652893066} 02/24/2022 05:47:00 - INFO - codeparrot_training - Step 3246: {'lr': 0.0004991691499203195, 'samples': 1662464, 'steps': 3246, 'loss/train': 3.7530465126037598} 02/24/2022 05:47:06 - INFO - codeparrot_training - Step 3247: {'lr': 0.0004991678164980914, 'samples': 1662976, 'steps': 3247, 'loss/train': 3.4744019508361816} 02/24/2022 05:47:09 - INFO - codeparrot_training - Step 3248: {'lr': 0.0004991664820085074, 'samples': 1663488, 'steps': 3248, 'loss/train': 2.2773642539978027} 02/24/2022 05:47:14 - INFO - codeparrot_training - Step 3249: {'lr': 0.0004991651464515735, 'samples': 1664000, 'steps': 3249, 'loss/train': 3.480146646499634} 02/24/2022 05:47:18 - INFO - codeparrot_training - Step 3250: {'lr': 0.0004991638098272951, 'samples': 1664512, 'steps': 3250, 'loss/train': 2.522777557373047} 02/24/2022 05:47:25 - INFO - codeparrot_training - Step 3251: {'lr': 0.000499162472135678, 'samples': 1665024, 'steps': 3251, 'loss/train': 3.02297043800354} 02/24/2022 05:47:28 - INFO - codeparrot_training - Step 3252: {'lr': 0.0004991611333767281, 'samples': 1665536, 'steps': 3252, 'loss/train': 3.784723997116089} 02/24/2022 05:47:34 - INFO - codeparrot_training - Step 3253: {'lr': 0.000499159793550451, 'samples': 1666048, 'steps': 3253, 'loss/train': 3.7307851314544678} 02/24/2022 05:47:39 - INFO - codeparrot_training - Step 3254: {'lr': 0.0004991584526568524, 'samples': 1666560, 'steps': 3254, 'loss/train': 2.665712833404541} 02/24/2022 05:47:43 - INFO - codeparrot_training - Step 3255: {'lr': 0.0004991571106959383, 'samples': 1667072, 'steps': 3255, 'loss/train': 2.9563913345336914} 02/24/2022 05:47:48 - INFO - codeparrot_training - Step 3256: {'lr': 0.000499155767667714, 'samples': 1667584, 'steps': 3256, 'loss/train': 2.1152777671813965} 02/24/2022 05:47:51 - INFO - codeparrot_training - Step 3257: {'lr': 0.0004991544235721857, 'samples': 1668096, 'steps': 3257, 'loss/train': 3.230518102645874} 02/24/2022 05:47:57 - INFO - codeparrot_training - Step 3258: {'lr': 0.0004991530784093589, 'samples': 1668608, 'steps': 3258, 'loss/train': 2.232226848602295} 02/24/2022 05:48:00 - INFO - codeparrot_training - Step 3259: {'lr': 0.0004991517321792394, 'samples': 1669120, 'steps': 3259, 'loss/train': 3.9473624229431152} 02/24/2022 05:48:06 - INFO - codeparrot_training - Step 3260: {'lr': 0.000499150384881833, 'samples': 1669632, 'steps': 3260, 'loss/train': 2.196275472640991} 02/24/2022 05:48:09 - INFO - codeparrot_training - Step 3261: {'lr': 0.0004991490365171454, 'samples': 1670144, 'steps': 3261, 'loss/train': 2.8099348545074463} 02/24/2022 05:48:16 - INFO - codeparrot_training - Step 3262: {'lr': 0.0004991476870851825, 'samples': 1670656, 'steps': 3262, 'loss/train': 3.0723612308502197} 02/24/2022 05:48:20 - INFO - codeparrot_training - Step 3263: {'lr': 0.0004991463365859501, 'samples': 1671168, 'steps': 3263, 'loss/train': 4.407003402709961} 02/24/2022 05:48:25 - INFO - codeparrot_training - Step 3264: {'lr': 0.0004991449850194538, 'samples': 1671680, 'steps': 3264, 'loss/train': 2.6636056900024414} 02/24/2022 05:48:28 - INFO - codeparrot_training - Step 3265: {'lr': 0.0004991436323856995, 'samples': 1672192, 'steps': 3265, 'loss/train': 2.2249515056610107} 02/24/2022 05:48:34 - INFO - codeparrot_training - Step 3266: {'lr': 0.0004991422786846931, 'samples': 1672704, 'steps': 3266, 'loss/train': 1.6259679794311523} 02/24/2022 05:48:37 - INFO - codeparrot_training - Step 3267: {'lr': 0.0004991409239164401, 'samples': 1673216, 'steps': 3267, 'loss/train': 2.8585119247436523} 02/24/2022 05:48:43 - INFO - codeparrot_training - Step 3268: {'lr': 0.0004991395680809467, 'samples': 1673728, 'steps': 3268, 'loss/train': 3.8965282440185547} 02/24/2022 05:48:46 - INFO - codeparrot_training - Step 3269: {'lr': 0.0004991382111782183, 'samples': 1674240, 'steps': 3269, 'loss/train': 3.693962812423706} 02/24/2022 05:48:52 - INFO - codeparrot_training - Step 3270: {'lr': 0.0004991368532082611, 'samples': 1674752, 'steps': 3270, 'loss/train': 1.0827059745788574} 02/24/2022 05:48:55 - INFO - codeparrot_training - Step 3271: {'lr': 0.0004991354941710806, 'samples': 1675264, 'steps': 3271, 'loss/train': 2.297900438308716} 02/24/2022 05:49:01 - INFO - codeparrot_training - Step 3272: {'lr': 0.0004991341340666828, 'samples': 1675776, 'steps': 3272, 'loss/train': 2.014286756515503} 02/24/2022 05:49:04 - INFO - codeparrot_training - Step 3273: {'lr': 0.0004991327728950736, 'samples': 1676288, 'steps': 3273, 'loss/train': 2.8018417358398438} 02/24/2022 05:49:10 - INFO - codeparrot_training - Step 3274: {'lr': 0.0004991314106562586, 'samples': 1676800, 'steps': 3274, 'loss/train': 2.6202468872070312} 02/24/2022 05:49:13 - INFO - codeparrot_training - Step 3275: {'lr': 0.0004991300473502437, 'samples': 1677312, 'steps': 3275, 'loss/train': 2.7593538761138916} 02/24/2022 05:49:19 - INFO - codeparrot_training - Step 3276: {'lr': 0.0004991286829770348, 'samples': 1677824, 'steps': 3276, 'loss/train': 3.337987184524536} 02/24/2022 05:49:23 - INFO - codeparrot_training - Step 3277: {'lr': 0.0004991273175366378, 'samples': 1678336, 'steps': 3277, 'loss/train': 2.0048723220825195} 02/24/2022 05:49:26 - INFO - codeparrot_training - Step 3278: {'lr': 0.0004991259510290584, 'samples': 1678848, 'steps': 3278, 'loss/train': 4.979170322418213} 02/24/2022 05:49:33 - INFO - codeparrot_training - Step 3279: {'lr': 0.0004991245834543025, 'samples': 1679360, 'steps': 3279, 'loss/train': 2.2831180095672607} 02/24/2022 05:49:36 - INFO - codeparrot_training - Step 3280: {'lr': 0.0004991232148123761, 'samples': 1679872, 'steps': 3280, 'loss/train': 2.6834723949432373} 02/24/2022 05:49:41 - INFO - codeparrot_training - Step 3281: {'lr': 0.0004991218451032849, 'samples': 1680384, 'steps': 3281, 'loss/train': 3.3644988536834717} 02/24/2022 05:49:45 - INFO - codeparrot_training - Step 3282: {'lr': 0.0004991204743270348, 'samples': 1680896, 'steps': 3282, 'loss/train': 3.373000383377075} 02/24/2022 05:49:50 - INFO - codeparrot_training - Step 3283: {'lr': 0.0004991191024836317, 'samples': 1681408, 'steps': 3283, 'loss/train': 3.6448915004730225} 02/24/2022 05:49:56 - INFO - codeparrot_training - Step 3284: {'lr': 0.0004991177295730815, 'samples': 1681920, 'steps': 3284, 'loss/train': 3.4570765495300293} 02/24/2022 05:49:59 - INFO - codeparrot_training - Step 3285: {'lr': 0.0004991163555953901, 'samples': 1682432, 'steps': 3285, 'loss/train': 3.0987887382507324} 02/24/2022 05:50:06 - INFO - codeparrot_training - Step 3286: {'lr': 0.0004991149805505632, 'samples': 1682944, 'steps': 3286, 'loss/train': 3.1108367443084717} 02/24/2022 05:50:10 - INFO - codeparrot_training - Step 3287: {'lr': 0.0004991136044386069, 'samples': 1683456, 'steps': 3287, 'loss/train': 1.6065137386322021} 02/24/2022 05:50:15 - INFO - codeparrot_training - Step 3288: {'lr': 0.0004991122272595271, 'samples': 1683968, 'steps': 3288, 'loss/train': 3.9230844974517822} 02/24/2022 05:50:19 - INFO - codeparrot_training - Step 3289: {'lr': 0.0004991108490133296, 'samples': 1684480, 'steps': 3289, 'loss/train': 3.022031545639038} 02/24/2022 05:50:24 - INFO - codeparrot_training - Step 3290: {'lr': 0.0004991094697000202, 'samples': 1684992, 'steps': 3290, 'loss/train': 2.7204723358154297} 02/24/2022 05:50:28 - INFO - codeparrot_training - Step 3291: {'lr': 0.000499108089319605, 'samples': 1685504, 'steps': 3291, 'loss/train': 1.8913851976394653} 02/24/2022 05:50:33 - INFO - codeparrot_training - Step 3292: {'lr': 0.0004991067078720899, 'samples': 1686016, 'steps': 3292, 'loss/train': 2.1258010864257812} 02/24/2022 05:50:37 - INFO - codeparrot_training - Step 3293: {'lr': 0.0004991053253574807, 'samples': 1686528, 'steps': 3293, 'loss/train': 3.767484188079834} 02/24/2022 05:50:42 - INFO - codeparrot_training - Step 3294: {'lr': 0.0004991039417757833, 'samples': 1687040, 'steps': 3294, 'loss/train': 2.5974926948547363} 02/24/2022 05:50:46 - INFO - codeparrot_training - Step 3295: {'lr': 0.0004991025571270039, 'samples': 1687552, 'steps': 3295, 'loss/train': 3.828691244125366} 02/24/2022 05:50:51 - INFO - codeparrot_training - Step 3296: {'lr': 0.000499101171411148, 'samples': 1688064, 'steps': 3296, 'loss/train': 2.7984731197357178} 02/24/2022 05:50:54 - INFO - codeparrot_training - Step 3297: {'lr': 0.000499099784628222, 'samples': 1688576, 'steps': 3297, 'loss/train': 2.1045572757720947} 02/24/2022 05:51:01 - INFO - codeparrot_training - Step 3298: {'lr': 0.0004990983967782316, 'samples': 1689088, 'steps': 3298, 'loss/train': 0.8454645276069641} 02/24/2022 05:51:04 - INFO - codeparrot_training - Step 3299: {'lr': 0.0004990970078611827, 'samples': 1689600, 'steps': 3299, 'loss/train': 3.2705276012420654} 02/24/2022 05:51:10 - INFO - codeparrot_training - Step 3300: {'lr': 0.0004990956178770814, 'samples': 1690112, 'steps': 3300, 'loss/train': 2.8591206073760986} 02/24/2022 05:51:13 - INFO - codeparrot_training - Step 3301: {'lr': 0.0004990942268259335, 'samples': 1690624, 'steps': 3301, 'loss/train': 4.083927154541016} 02/24/2022 05:51:19 - INFO - codeparrot_training - Step 3302: {'lr': 0.000499092834707745, 'samples': 1691136, 'steps': 3302, 'loss/train': 2.730400323867798} 02/24/2022 05:51:22 - INFO - codeparrot_training - Step 3303: {'lr': 0.000499091441522522, 'samples': 1691648, 'steps': 3303, 'loss/train': 2.869877815246582} 02/24/2022 05:51:28 - INFO - codeparrot_training - Step 3304: {'lr': 0.0004990900472702702, 'samples': 1692160, 'steps': 3304, 'loss/train': 2.0143260955810547} 02/24/2022 05:51:31 - INFO - codeparrot_training - Step 3305: {'lr': 0.0004990886519509959, 'samples': 1692672, 'steps': 3305, 'loss/train': 2.5578551292419434} 02/24/2022 05:51:37 - INFO - codeparrot_training - Step 3306: {'lr': 0.0004990872555647048, 'samples': 1693184, 'steps': 3306, 'loss/train': 2.9826536178588867} 02/24/2022 05:51:40 - INFO - codeparrot_training - Step 3307: {'lr': 0.0004990858581114029, 'samples': 1693696, 'steps': 3307, 'loss/train': 3.3972973823547363} 02/24/2022 05:51:46 - INFO - codeparrot_training - Step 3308: {'lr': 0.0004990844595910965, 'samples': 1694208, 'steps': 3308, 'loss/train': 2.664269208908081} 02/24/2022 05:51:49 - INFO - codeparrot_training - Step 3309: {'lr': 0.0004990830600037912, 'samples': 1694720, 'steps': 3309, 'loss/train': 2.6696882247924805} 02/24/2022 05:51:55 - INFO - codeparrot_training - Step 3310: {'lr': 0.0004990816593494933, 'samples': 1695232, 'steps': 3310, 'loss/train': 2.8404529094696045} 02/24/2022 05:51:58 - INFO - codeparrot_training - Step 3311: {'lr': 0.0004990802576282085, 'samples': 1695744, 'steps': 3311, 'loss/train': 3.3159096240997314} 02/24/2022 05:52:03 - INFO - codeparrot_training - Step 3312: {'lr': 0.0004990788548399431, 'samples': 1696256, 'steps': 3312, 'loss/train': 2.6696512699127197} 02/24/2022 05:52:07 - INFO - codeparrot_training - Step 3313: {'lr': 0.0004990774509847029, 'samples': 1696768, 'steps': 3313, 'loss/train': 3.5161855220794678} 02/24/2022 05:52:14 - INFO - codeparrot_training - Step 3314: {'lr': 0.0004990760460624941, 'samples': 1697280, 'steps': 3314, 'loss/train': 3.0681910514831543} 02/24/2022 05:52:17 - INFO - codeparrot_training - Step 3315: {'lr': 0.0004990746400733225, 'samples': 1697792, 'steps': 3315, 'loss/train': 3.209573268890381} 02/24/2022 05:52:23 - INFO - codeparrot_training - Step 3316: {'lr': 0.0004990732330171943, 'samples': 1698304, 'steps': 3316, 'loss/train': 1.746667742729187} 02/24/2022 05:52:28 - INFO - codeparrot_training - Step 3317: {'lr': 0.0004990718248941154, 'samples': 1698816, 'steps': 3317, 'loss/train': 3.3051421642303467} 02/24/2022 05:52:32 - INFO - codeparrot_training - Step 3318: {'lr': 0.0004990704157040919, 'samples': 1699328, 'steps': 3318, 'loss/train': 3.486098051071167} 02/24/2022 05:52:37 - INFO - codeparrot_training - Step 3319: {'lr': 0.0004990690054471299, 'samples': 1699840, 'steps': 3319, 'loss/train': 2.2478842735290527} 02/24/2022 05:52:41 - INFO - codeparrot_training - Step 3320: {'lr': 0.0004990675941232354, 'samples': 1700352, 'steps': 3320, 'loss/train': 1.7554570436477661} 02/24/2022 05:52:46 - INFO - codeparrot_training - Step 3321: {'lr': 0.0004990661817324142, 'samples': 1700864, 'steps': 3321, 'loss/train': 3.1901450157165527} 02/24/2022 05:52:49 - INFO - codeparrot_training - Step 3322: {'lr': 0.0004990647682746727, 'samples': 1701376, 'steps': 3322, 'loss/train': 3.148627996444702} 02/24/2022 05:52:56 - INFO - codeparrot_training - Step 3323: {'lr': 0.0004990633537500169, 'samples': 1701888, 'steps': 3323, 'loss/train': 4.160163879394531} 02/24/2022 05:52:59 - INFO - codeparrot_training - Step 3324: {'lr': 0.0004990619381584527, 'samples': 1702400, 'steps': 3324, 'loss/train': 2.2061586380004883} 02/24/2022 05:53:05 - INFO - codeparrot_training - Step 3325: {'lr': 0.0004990605214999862, 'samples': 1702912, 'steps': 3325, 'loss/train': 3.1617579460144043} 02/24/2022 05:53:08 - INFO - codeparrot_training - Step 3326: {'lr': 0.0004990591037746236, 'samples': 1703424, 'steps': 3326, 'loss/train': 4.343183994293213} 02/24/2022 05:53:13 - INFO - codeparrot_training - Step 3327: {'lr': 0.0004990576849823708, 'samples': 1703936, 'steps': 3327, 'loss/train': 1.9711787700653076} 02/24/2022 05:53:17 - INFO - codeparrot_training - Step 3328: {'lr': 0.000499056265123234, 'samples': 1704448, 'steps': 3328, 'loss/train': 2.978633403778076} 02/24/2022 05:53:23 - INFO - codeparrot_training - Step 3329: {'lr': 0.0004990548441972193, 'samples': 1704960, 'steps': 3329, 'loss/train': 2.6748435497283936} 02/24/2022 05:53:26 - INFO - codeparrot_training - Step 3330: {'lr': 0.0004990534222043325, 'samples': 1705472, 'steps': 3330, 'loss/train': 0.8951675295829773} 02/24/2022 05:53:31 - INFO - codeparrot_training - Step 3331: {'lr': 0.0004990519991445803, 'samples': 1705984, 'steps': 3331, 'loss/train': 3.2934584617614746} 02/24/2022 05:53:35 - INFO - codeparrot_training - Step 3332: {'lr': 0.0004990505750179682, 'samples': 1706496, 'steps': 3332, 'loss/train': 2.81172776222229} 02/24/2022 05:53:41 - INFO - codeparrot_training - Step 3333: {'lr': 0.0004990491498245024, 'samples': 1707008, 'steps': 3333, 'loss/train': 2.932238817214966} 02/24/2022 05:53:44 - INFO - codeparrot_training - Step 3334: {'lr': 0.0004990477235641893, 'samples': 1707520, 'steps': 3334, 'loss/train': 1.2628066539764404} 02/24/2022 05:53:50 - INFO - codeparrot_training - Step 3335: {'lr': 0.0004990462962370347, 'samples': 1708032, 'steps': 3335, 'loss/train': 3.303638219833374} 02/24/2022 05:53:54 - INFO - codeparrot_training - Step 3336: {'lr': 0.0004990448678430451, 'samples': 1708544, 'steps': 3336, 'loss/train': 2.033548355102539} 02/24/2022 05:53:59 - INFO - codeparrot_training - Step 3337: {'lr': 0.0004990434383822261, 'samples': 1709056, 'steps': 3337, 'loss/train': 3.910740852355957} 02/24/2022 05:54:02 - INFO - codeparrot_training - Step 3338: {'lr': 0.0004990420078545843, 'samples': 1709568, 'steps': 3338, 'loss/train': 0.6613812446594238} 02/24/2022 05:54:08 - INFO - codeparrot_training - Step 3339: {'lr': 0.0004990405762601254, 'samples': 1710080, 'steps': 3339, 'loss/train': 4.209691524505615} 02/24/2022 05:54:11 - INFO - codeparrot_training - Step 3340: {'lr': 0.000499039143598856, 'samples': 1710592, 'steps': 3340, 'loss/train': 4.026211261749268} 02/24/2022 05:54:17 - INFO - codeparrot_training - Step 3341: {'lr': 0.0004990377098707818, 'samples': 1711104, 'steps': 3341, 'loss/train': 3.085852861404419} 02/24/2022 05:54:20 - INFO - codeparrot_training - Step 3342: {'lr': 0.0004990362750759092, 'samples': 1711616, 'steps': 3342, 'loss/train': 2.8016932010650635} 02/24/2022 05:54:26 - INFO - codeparrot_training - Step 3343: {'lr': 0.0004990348392142443, 'samples': 1712128, 'steps': 3343, 'loss/train': 1.691659927368164} 02/24/2022 05:54:29 - INFO - codeparrot_training - Step 3344: {'lr': 0.0004990334022857932, 'samples': 1712640, 'steps': 3344, 'loss/train': 1.7974783182144165} 02/24/2022 05:54:36 - INFO - codeparrot_training - Step 3345: {'lr': 0.0004990319642905619, 'samples': 1713152, 'steps': 3345, 'loss/train': 3.4584097862243652} 02/24/2022 05:54:39 - INFO - codeparrot_training - Step 3346: {'lr': 0.000499030525228557, 'samples': 1713664, 'steps': 3346, 'loss/train': 2.7353296279907227} 02/24/2022 05:54:45 - INFO - codeparrot_training - Step 3347: {'lr': 0.0004990290850997843, 'samples': 1714176, 'steps': 3347, 'loss/train': 2.4467248916625977} 02/24/2022 05:54:48 - INFO - codeparrot_training - Step 3348: {'lr': 0.0004990276439042501, 'samples': 1714688, 'steps': 3348, 'loss/train': 1.4302935600280762} 02/24/2022 05:54:54 - INFO - codeparrot_training - Step 3349: {'lr': 0.0004990262016419606, 'samples': 1715200, 'steps': 3349, 'loss/train': 3.072995901107788} 02/24/2022 05:54:57 - INFO - codeparrot_training - Step 3350: {'lr': 0.0004990247583129218, 'samples': 1715712, 'steps': 3350, 'loss/train': 2.932849884033203} 02/24/2022 05:55:03 - INFO - codeparrot_training - Step 3351: {'lr': 0.00049902331391714, 'samples': 1716224, 'steps': 3351, 'loss/train': 2.1219727993011475} 02/24/2022 05:55:06 - INFO - codeparrot_training - Step 3352: {'lr': 0.0004990218684546216, 'samples': 1716736, 'steps': 3352, 'loss/train': 3.4314424991607666} 02/24/2022 05:55:12 - INFO - codeparrot_training - Step 3353: {'lr': 0.0004990204219253724, 'samples': 1717248, 'steps': 3353, 'loss/train': 3.2598094940185547} 02/24/2022 05:55:15 - INFO - codeparrot_training - Step 3354: {'lr': 0.0004990189743293989, 'samples': 1717760, 'steps': 3354, 'loss/train': 2.1458733081817627} 02/24/2022 05:55:21 - INFO - codeparrot_training - Step 3355: {'lr': 0.0004990175256667071, 'samples': 1718272, 'steps': 3355, 'loss/train': 3.1958394050598145} 02/24/2022 05:55:24 - INFO - codeparrot_training - Step 3356: {'lr': 0.0004990160759373033, 'samples': 1718784, 'steps': 3356, 'loss/train': 2.9634346961975098} 02/24/2022 05:55:30 - INFO - codeparrot_training - Step 3357: {'lr': 0.0004990146251411938, 'samples': 1719296, 'steps': 3357, 'loss/train': 3.195547103881836} 02/24/2022 05:55:33 - INFO - codeparrot_training - Step 3358: {'lr': 0.0004990131732783846, 'samples': 1719808, 'steps': 3358, 'loss/train': 3.348626136779785} 02/24/2022 05:55:40 - INFO - codeparrot_training - Step 3359: {'lr': 0.000499011720348882, 'samples': 1720320, 'steps': 3359, 'loss/train': 2.193225383758545} 02/24/2022 05:55:43 - INFO - codeparrot_training - Step 3360: {'lr': 0.0004990102663526924, 'samples': 1720832, 'steps': 3360, 'loss/train': 0.33492591977119446} 02/24/2022 05:55:49 - INFO - codeparrot_training - Step 3361: {'lr': 0.0004990088112898219, 'samples': 1721344, 'steps': 3361, 'loss/train': 1.6221529245376587} 02/24/2022 05:55:52 - INFO - codeparrot_training - Step 3362: {'lr': 0.0004990073551602766, 'samples': 1721856, 'steps': 3362, 'loss/train': 3.003966808319092} 02/24/2022 05:55:58 - INFO - codeparrot_training - Step 3363: {'lr': 0.000499005897964063, 'samples': 1722368, 'steps': 3363, 'loss/train': 3.1551339626312256} 02/24/2022 05:56:01 - INFO - codeparrot_training - Step 3364: {'lr': 0.0004990044397011871, 'samples': 1722880, 'steps': 3364, 'loss/train': 2.333383321762085} 02/24/2022 05:56:07 - INFO - codeparrot_training - Step 3365: {'lr': 0.0004990029803716552, 'samples': 1723392, 'steps': 3365, 'loss/train': 4.322406768798828} 02/24/2022 05:56:10 - INFO - codeparrot_training - Step 3366: {'lr': 0.0004990015199754736, 'samples': 1723904, 'steps': 3366, 'loss/train': 2.2222132682800293} 02/24/2022 05:56:16 - INFO - codeparrot_training - Step 3367: {'lr': 0.0004990000585126486, 'samples': 1724416, 'steps': 3367, 'loss/train': 2.321769952774048} 02/24/2022 05:56:20 - INFO - codeparrot_training - Step 3368: {'lr': 0.0004989985959831865, 'samples': 1724928, 'steps': 3368, 'loss/train': 3.484276294708252} 02/24/2022 05:56:23 - INFO - codeparrot_training - Step 3369: {'lr': 0.0004989971323870934, 'samples': 1725440, 'steps': 3369, 'loss/train': 2.8329782485961914} 02/24/2022 05:56:30 - INFO - codeparrot_training - Step 3370: {'lr': 0.0004989956677243757, 'samples': 1725952, 'steps': 3370, 'loss/train': 4.990061283111572} 02/24/2022 05:56:36 - INFO - codeparrot_training - Step 3371: {'lr': 0.0004989942019950395, 'samples': 1726464, 'steps': 3371, 'loss/train': 2.5743296146392822} 02/24/2022 05:56:39 - INFO - codeparrot_training - Step 3372: {'lr': 0.0004989927351990912, 'samples': 1726976, 'steps': 3372, 'loss/train': 2.976681709289551} 02/24/2022 05:56:45 - INFO - codeparrot_training - Step 3373: {'lr': 0.0004989912673365373, 'samples': 1727488, 'steps': 3373, 'loss/train': 3.4696974754333496} 02/24/2022 05:56:48 - INFO - codeparrot_training - Step 3374: {'lr': 0.0004989897984073837, 'samples': 1728000, 'steps': 3374, 'loss/train': 3.0720739364624023} 02/24/2022 05:56:54 - INFO - codeparrot_training - Step 3375: {'lr': 0.000498988328411637, 'samples': 1728512, 'steps': 3375, 'loss/train': 2.514559268951416} 02/24/2022 05:56:57 - INFO - codeparrot_training - Step 3376: {'lr': 0.0004989868573493032, 'samples': 1729024, 'steps': 3376, 'loss/train': 1.5759943723678589} 02/24/2022 05:57:02 - INFO - codeparrot_training - Step 3377: {'lr': 0.0004989853852203889, 'samples': 1729536, 'steps': 3377, 'loss/train': 2.459177017211914} 02/24/2022 05:57:06 - INFO - codeparrot_training - Step 3378: {'lr': 0.0004989839120249002, 'samples': 1730048, 'steps': 3378, 'loss/train': 3.1192870140075684} 02/24/2022 05:57:12 - INFO - codeparrot_training - Step 3379: {'lr': 0.0004989824377628435, 'samples': 1730560, 'steps': 3379, 'loss/train': 2.705767869949341} 02/24/2022 05:57:15 - INFO - codeparrot_training - Step 3380: {'lr': 0.0004989809624342251, 'samples': 1731072, 'steps': 3380, 'loss/train': 1.0068100690841675} 02/24/2022 05:57:21 - INFO - codeparrot_training - Step 3381: {'lr': 0.0004989794860390513, 'samples': 1731584, 'steps': 3381, 'loss/train': 4.0767927169799805} 02/24/2022 05:57:25 - INFO - codeparrot_training - Step 3382: {'lr': 0.0004989780085773285, 'samples': 1732096, 'steps': 3382, 'loss/train': 3.0383639335632324} 02/24/2022 05:57:30 - INFO - codeparrot_training - Step 3383: {'lr': 0.0004989765300490628, 'samples': 1732608, 'steps': 3383, 'loss/train': 3.3286867141723633} 02/24/2022 05:57:34 - INFO - codeparrot_training - Step 3384: {'lr': 0.0004989750504542609, 'samples': 1733120, 'steps': 3384, 'loss/train': 1.9907475709915161} 02/24/2022 05:57:39 - INFO - codeparrot_training - Step 3385: {'lr': 0.0004989735697929289, 'samples': 1733632, 'steps': 3385, 'loss/train': 2.196045160293579} 02/24/2022 05:57:43 - INFO - codeparrot_training - Step 3386: {'lr': 0.0004989720880650731, 'samples': 1734144, 'steps': 3386, 'loss/train': 3.305495262145996} 02/24/2022 05:57:48 - INFO - codeparrot_training - Step 3387: {'lr': 0.0004989706052707, 'samples': 1734656, 'steps': 3387, 'loss/train': 2.4244720935821533} 02/24/2022 05:57:52 - INFO - codeparrot_training - Step 3388: {'lr': 0.0004989691214098158, 'samples': 1735168, 'steps': 3388, 'loss/train': 7.428947448730469} 02/24/2022 05:57:57 - INFO - codeparrot_training - Step 3389: {'lr': 0.0004989676364824271, 'samples': 1735680, 'steps': 3389, 'loss/train': 3.755610942840576} 02/24/2022 05:58:01 - INFO - codeparrot_training - Step 3390: {'lr': 0.00049896615048854, 'samples': 1736192, 'steps': 3390, 'loss/train': 2.9101481437683105} 02/24/2022 05:58:06 - INFO - codeparrot_training - Step 3391: {'lr': 0.000498964663428161, 'samples': 1736704, 'steps': 3391, 'loss/train': 1.8431017398834229} 02/24/2022 05:58:10 - INFO - codeparrot_training - Step 3392: {'lr': 0.0004989631753012964, 'samples': 1737216, 'steps': 3392, 'loss/train': 1.9650423526763916} 02/24/2022 05:58:15 - INFO - codeparrot_training - Step 3393: {'lr': 0.0004989616861079527, 'samples': 1737728, 'steps': 3393, 'loss/train': 0.8594864010810852} 02/24/2022 05:58:19 - INFO - codeparrot_training - Step 3394: {'lr': 0.0004989601958481361, 'samples': 1738240, 'steps': 3394, 'loss/train': 2.525815486907959} 02/24/2022 05:58:25 - INFO - codeparrot_training - Step 3395: {'lr': 0.000498958704521853, 'samples': 1738752, 'steps': 3395, 'loss/train': 3.5477874279022217} 02/24/2022 05:58:28 - INFO - codeparrot_training - Step 3396: {'lr': 0.00049895721212911, 'samples': 1739264, 'steps': 3396, 'loss/train': 3.293307065963745} 02/24/2022 05:58:34 - INFO - codeparrot_training - Step 3397: {'lr': 0.0004989557186699133, 'samples': 1739776, 'steps': 3397, 'loss/train': 3.0308268070220947} 02/24/2022 05:58:37 - INFO - codeparrot_training - Step 3398: {'lr': 0.0004989542241442695, 'samples': 1740288, 'steps': 3398, 'loss/train': 3.868443012237549} 02/24/2022 05:58:43 - INFO - codeparrot_training - Step 3399: {'lr': 0.0004989527285521846, 'samples': 1740800, 'steps': 3399, 'loss/train': 1.2805525064468384} 02/24/2022 05:58:46 - INFO - codeparrot_training - Step 3400: {'lr': 0.0004989512318936654, 'samples': 1741312, 'steps': 3400, 'loss/train': 2.518308162689209} 02/24/2022 05:58:52 - INFO - codeparrot_training - Step 3401: {'lr': 0.0004989497341687182, 'samples': 1741824, 'steps': 3401, 'loss/train': 4.038780689239502} 02/24/2022 05:58:55 - INFO - codeparrot_training - Step 3402: {'lr': 0.0004989482353773494, 'samples': 1742336, 'steps': 3402, 'loss/train': 2.5255954265594482} 02/24/2022 05:59:01 - INFO - codeparrot_training - Step 3403: {'lr': 0.0004989467355195653, 'samples': 1742848, 'steps': 3403, 'loss/train': 2.20237398147583} 02/24/2022 05:59:04 - INFO - codeparrot_training - Step 3404: {'lr': 0.0004989452345953725, 'samples': 1743360, 'steps': 3404, 'loss/train': 3.1351206302642822} 02/24/2022 05:59:10 - INFO - codeparrot_training - Step 3405: {'lr': 0.0004989437326047774, 'samples': 1743872, 'steps': 3405, 'loss/train': 3.515432357788086} 02/24/2022 05:59:13 - INFO - codeparrot_training - Step 3406: {'lr': 0.0004989422295477863, 'samples': 1744384, 'steps': 3406, 'loss/train': 2.930018663406372} 02/24/2022 05:59:19 - INFO - codeparrot_training - Step 3407: {'lr': 0.0004989407254244058, 'samples': 1744896, 'steps': 3407, 'loss/train': 2.9464001655578613} 02/24/2022 05:59:23 - INFO - codeparrot_training - Step 3408: {'lr': 0.0004989392202346424, 'samples': 1745408, 'steps': 3408, 'loss/train': 4.28965425491333} 02/24/2022 05:59:29 - INFO - codeparrot_training - Step 3409: {'lr': 0.0004989377139785022, 'samples': 1745920, 'steps': 3409, 'loss/train': 2.82633376121521} 02/24/2022 05:59:32 - INFO - codeparrot_training - Step 3410: {'lr': 0.000498936206655992, 'samples': 1746432, 'steps': 3410, 'loss/train': 3.0110108852386475} 02/24/2022 05:59:38 - INFO - codeparrot_training - Step 3411: {'lr': 0.0004989346982671181, 'samples': 1746944, 'steps': 3411, 'loss/train': 2.8296079635620117} 02/24/2022 05:59:41 - INFO - codeparrot_training - Step 3412: {'lr': 0.0004989331888118869, 'samples': 1747456, 'steps': 3412, 'loss/train': 2.685868740081787} 02/24/2022 05:59:47 - INFO - codeparrot_training - Step 3413: {'lr': 0.0004989316782903052, 'samples': 1747968, 'steps': 3413, 'loss/train': 2.2202298641204834} 02/24/2022 05:59:50 - INFO - codeparrot_training - Step 3414: {'lr': 0.0004989301667023791, 'samples': 1748480, 'steps': 3414, 'loss/train': 3.193492889404297} 02/24/2022 05:59:56 - INFO - codeparrot_training - Step 3415: {'lr': 0.0004989286540481152, 'samples': 1748992, 'steps': 3415, 'loss/train': 2.7336885929107666} 02/24/2022 06:00:00 - INFO - codeparrot_training - Step 3416: {'lr': 0.00049892714032752, 'samples': 1749504, 'steps': 3416, 'loss/train': 1.3939684629440308} 02/24/2022 06:00:05 - INFO - codeparrot_training - Step 3417: {'lr': 0.0004989256255406001, 'samples': 1750016, 'steps': 3417, 'loss/train': 2.98162579536438} 02/24/2022 06:00:09 - INFO - codeparrot_training - Step 3418: {'lr': 0.0004989241096873617, 'samples': 1750528, 'steps': 3418, 'loss/train': 2.2810022830963135} 02/24/2022 06:00:14 - INFO - codeparrot_training - Step 3419: {'lr': 0.0004989225927678115, 'samples': 1751040, 'steps': 3419, 'loss/train': 2.1340138912200928} 02/24/2022 06:00:18 - INFO - codeparrot_training - Step 3420: {'lr': 0.000498921074781956, 'samples': 1751552, 'steps': 3420, 'loss/train': 2.4442453384399414} 02/24/2022 06:00:23 - INFO - codeparrot_training - Step 3421: {'lr': 0.0004989195557298016, 'samples': 1752064, 'steps': 3421, 'loss/train': 1.3915042877197266} 02/24/2022 06:00:27 - INFO - codeparrot_training - Step 3422: {'lr': 0.0004989180356113549, 'samples': 1752576, 'steps': 3422, 'loss/train': 3.096195697784424} 02/24/2022 06:00:33 - INFO - codeparrot_training - Step 3423: {'lr': 0.0004989165144266224, 'samples': 1753088, 'steps': 3423, 'loss/train': 3.191877841949463} 02/24/2022 06:00:36 - INFO - codeparrot_training - Step 3424: {'lr': 0.0004989149921756105, 'samples': 1753600, 'steps': 3424, 'loss/train': 3.200476884841919} 02/24/2022 06:00:42 - INFO - codeparrot_training - Step 3425: {'lr': 0.0004989134688583259, 'samples': 1754112, 'steps': 3425, 'loss/train': 3.352710485458374} 02/24/2022 06:00:46 - INFO - codeparrot_training - Step 3426: {'lr': 0.000498911944474775, 'samples': 1754624, 'steps': 3426, 'loss/train': 3.7004177570343018} 02/24/2022 06:00:52 - INFO - codeparrot_training - Step 3427: {'lr': 0.0004989104190249643, 'samples': 1755136, 'steps': 3427, 'loss/train': 0.3849867284297943} 02/24/2022 06:00:55 - INFO - codeparrot_training - Step 3428: {'lr': 0.0004989088925089005, 'samples': 1755648, 'steps': 3428, 'loss/train': 2.8095381259918213} 02/24/2022 06:01:00 - INFO - codeparrot_training - Step 3429: {'lr': 0.00049890736492659, 'samples': 1756160, 'steps': 3429, 'loss/train': 3.5617098808288574} 02/24/2022 06:01:04 - INFO - codeparrot_training - Step 3430: {'lr': 0.0004989058362780394, 'samples': 1756672, 'steps': 3430, 'loss/train': 3.3939056396484375} 02/24/2022 06:01:09 - INFO - codeparrot_training - Step 3431: {'lr': 0.0004989043065632552, 'samples': 1757184, 'steps': 3431, 'loss/train': 3.7800133228302} 02/24/2022 06:01:13 - INFO - codeparrot_training - Step 3432: {'lr': 0.0004989027757822441, 'samples': 1757696, 'steps': 3432, 'loss/train': 1.511250376701355} 02/24/2022 06:01:18 - INFO - codeparrot_training - Step 3433: {'lr': 0.0004989012439350124, 'samples': 1758208, 'steps': 3433, 'loss/train': 0.15233869850635529} 02/24/2022 06:01:22 - INFO - codeparrot_training - Step 3434: {'lr': 0.0004988997110215668, 'samples': 1758720, 'steps': 3434, 'loss/train': 3.2571072578430176} 02/24/2022 06:01:27 - INFO - codeparrot_training - Step 3435: {'lr': 0.0004988981770419141, 'samples': 1759232, 'steps': 3435, 'loss/train': 3.0362415313720703} 02/24/2022 06:01:31 - INFO - codeparrot_training - Step 3436: {'lr': 0.0004988966419960605, 'samples': 1759744, 'steps': 3436, 'loss/train': 2.339164972305298} 02/24/2022 06:01:36 - INFO - codeparrot_training - Step 3437: {'lr': 0.0004988951058840127, 'samples': 1760256, 'steps': 3437, 'loss/train': 2.506068706512451} 02/24/2022 06:01:40 - INFO - codeparrot_training - Step 3438: {'lr': 0.0004988935687057773, 'samples': 1760768, 'steps': 3438, 'loss/train': 3.698857307434082} 02/24/2022 06:01:45 - INFO - codeparrot_training - Step 3439: {'lr': 0.0004988920304613609, 'samples': 1761280, 'steps': 3439, 'loss/train': 3.0238447189331055} 02/24/2022 06:01:49 - INFO - codeparrot_training - Step 3440: {'lr': 0.00049889049115077, 'samples': 1761792, 'steps': 3440, 'loss/train': 1.8441886901855469} 02/24/2022 06:01:55 - INFO - codeparrot_training - Step 3441: {'lr': 0.0004988889507740113, 'samples': 1762304, 'steps': 3441, 'loss/train': 1.3593950271606445} 02/24/2022 06:01:59 - INFO - codeparrot_training - Step 3442: {'lr': 0.0004988874093310914, 'samples': 1762816, 'steps': 3442, 'loss/train': 3.2507877349853516} 02/24/2022 06:02:04 - INFO - codeparrot_training - Step 3443: {'lr': 0.000498885866822017, 'samples': 1763328, 'steps': 3443, 'loss/train': 2.106030225753784} 02/24/2022 06:02:08 - INFO - codeparrot_training - Step 3444: {'lr': 0.0004988843232467944, 'samples': 1763840, 'steps': 3444, 'loss/train': 2.3151698112487793} 02/24/2022 06:02:13 - INFO - codeparrot_training - Step 3445: {'lr': 0.0004988827786054304, 'samples': 1764352, 'steps': 3445, 'loss/train': 2.7319116592407227} 02/24/2022 06:02:16 - INFO - codeparrot_training - Step 3446: {'lr': 0.0004988812328979317, 'samples': 1764864, 'steps': 3446, 'loss/train': 3.381547451019287} 02/24/2022 06:02:22 - INFO - codeparrot_training - Step 3447: {'lr': 0.0004988796861243046, 'samples': 1765376, 'steps': 3447, 'loss/train': 2.0480639934539795} 02/24/2022 06:02:26 - INFO - codeparrot_training - Step 3448: {'lr': 0.0004988781382845562, 'samples': 1765888, 'steps': 3448, 'loss/train': 2.7525269985198975} 02/24/2022 06:02:31 - INFO - codeparrot_training - Step 3449: {'lr': 0.0004988765893786929, 'samples': 1766400, 'steps': 3449, 'loss/train': 3.872993230819702} 02/24/2022 06:02:35 - INFO - codeparrot_training - Step 3450: {'lr': 0.0004988750394067211, 'samples': 1766912, 'steps': 3450, 'loss/train': 2.2978477478027344} 02/24/2022 06:02:41 - INFO - codeparrot_training - Step 3451: {'lr': 0.0004988734883686479, 'samples': 1767424, 'steps': 3451, 'loss/train': 2.334455728530884} 02/24/2022 06:02:44 - INFO - codeparrot_training - Step 3452: {'lr': 0.0004988719362644795, 'samples': 1767936, 'steps': 3452, 'loss/train': 2.5293805599212646} 02/24/2022 06:02:50 - INFO - codeparrot_training - Step 3453: {'lr': 0.0004988703830942228, 'samples': 1768448, 'steps': 3453, 'loss/train': 2.462118625640869} 02/24/2022 06:02:53 - INFO - codeparrot_training - Step 3454: {'lr': 0.0004988688288578845, 'samples': 1768960, 'steps': 3454, 'loss/train': 2.1267988681793213} 02/24/2022 06:02:59 - INFO - codeparrot_training - Step 3455: {'lr': 0.0004988672735554711, 'samples': 1769472, 'steps': 3455, 'loss/train': 1.8279513120651245} 02/24/2022 06:03:02 - INFO - codeparrot_training - Step 3456: {'lr': 0.0004988657171869893, 'samples': 1769984, 'steps': 3456, 'loss/train': 1.1528892517089844} 02/24/2022 06:03:08 - INFO - codeparrot_training - Step 3457: {'lr': 0.0004988641597524458, 'samples': 1770496, 'steps': 3457, 'loss/train': 2.45343279838562} 02/24/2022 06:03:11 - INFO - codeparrot_training - Step 3458: {'lr': 0.0004988626012518473, 'samples': 1771008, 'steps': 3458, 'loss/train': 2.931540012359619} 02/24/2022 06:03:17 - INFO - codeparrot_training - Step 3459: {'lr': 0.0004988610416852004, 'samples': 1771520, 'steps': 3459, 'loss/train': 2.7208609580993652} 02/24/2022 06:03:20 - INFO - codeparrot_training - Step 3460: {'lr': 0.0004988594810525118, 'samples': 1772032, 'steps': 3460, 'loss/train': 3.9978678226470947} 02/24/2022 06:03:26 - INFO - codeparrot_training - Step 3461: {'lr': 0.0004988579193537883, 'samples': 1772544, 'steps': 3461, 'loss/train': 3.373157262802124} 02/24/2022 06:03:29 - INFO - codeparrot_training - Step 3462: {'lr': 0.0004988563565890364, 'samples': 1773056, 'steps': 3462, 'loss/train': 3.328639507293701} 02/24/2022 06:03:35 - INFO - codeparrot_training - Step 3463: {'lr': 0.000498854792758263, 'samples': 1773568, 'steps': 3463, 'loss/train': 3.3730952739715576} 02/24/2022 06:03:39 - INFO - codeparrot_training - Step 3464: {'lr': 0.0004988532278614745, 'samples': 1774080, 'steps': 3464, 'loss/train': 2.235300302505493} 02/24/2022 06:03:44 - INFO - codeparrot_training - Step 3465: {'lr': 0.0004988516618986779, 'samples': 1774592, 'steps': 3465, 'loss/train': 2.551022529602051} 02/24/2022 06:03:48 - INFO - codeparrot_training - Step 3466: {'lr': 0.0004988500948698799, 'samples': 1775104, 'steps': 3466, 'loss/train': 2.154386043548584} 02/24/2022 06:03:53 - INFO - codeparrot_training - Step 3467: {'lr': 0.000498848526775087, 'samples': 1775616, 'steps': 3467, 'loss/train': 2.2679810523986816} 02/24/2022 06:03:57 - INFO - codeparrot_training - Step 3468: {'lr': 0.0004988469576143059, 'samples': 1776128, 'steps': 3468, 'loss/train': 3.1192033290863037} 02/24/2022 06:04:02 - INFO - codeparrot_training - Step 3469: {'lr': 0.0004988453873875437, 'samples': 1776640, 'steps': 3469, 'loss/train': 1.99477219581604} 02/24/2022 06:04:06 - INFO - codeparrot_training - Step 3470: {'lr': 0.0004988438160948068, 'samples': 1777152, 'steps': 3470, 'loss/train': 3.6434619426727295} 02/24/2022 06:04:11 - INFO - codeparrot_training - Step 3471: {'lr': 0.000498842243736102, 'samples': 1777664, 'steps': 3471, 'loss/train': 2.7385129928588867} 02/24/2022 06:04:15 - INFO - codeparrot_training - Step 3472: {'lr': 0.000498840670311436, 'samples': 1778176, 'steps': 3472, 'loss/train': 2.3991098403930664} 02/24/2022 06:04:20 - INFO - codeparrot_training - Step 3473: {'lr': 0.0004988390958208156, 'samples': 1778688, 'steps': 3473, 'loss/train': 2.373166084289551} 02/24/2022 06:04:26 - INFO - codeparrot_training - Step 3474: {'lr': 0.0004988375202642475, 'samples': 1779200, 'steps': 3474, 'loss/train': 2.6730761528015137} 02/24/2022 06:04:29 - INFO - codeparrot_training - Step 3475: {'lr': 0.0004988359436417385, 'samples': 1779712, 'steps': 3475, 'loss/train': 2.573969602584839} 02/24/2022 06:04:36 - INFO - codeparrot_training - Step 3476: {'lr': 0.0004988343659532954, 'samples': 1780224, 'steps': 3476, 'loss/train': 2.7785582542419434} 02/24/2022 06:04:39 - INFO - codeparrot_training - Step 3477: {'lr': 0.0004988327871989249, 'samples': 1780736, 'steps': 3477, 'loss/train': 2.594221830368042} 02/24/2022 06:04:45 - INFO - codeparrot_training - Step 3478: {'lr': 0.0004988312073786336, 'samples': 1781248, 'steps': 3478, 'loss/train': 3.8832502365112305} 02/24/2022 06:04:48 - INFO - codeparrot_training - Step 3479: {'lr': 0.0004988296264924286, 'samples': 1781760, 'steps': 3479, 'loss/train': 3.013744592666626} 02/24/2022 06:04:54 - INFO - codeparrot_training - Step 3480: {'lr': 0.0004988280445403164, 'samples': 1782272, 'steps': 3480, 'loss/train': 2.1859991550445557} 02/24/2022 06:04:58 - INFO - codeparrot_training - Step 3481: {'lr': 0.0004988264615223038, 'samples': 1782784, 'steps': 3481, 'loss/train': 2.1419198513031006} 02/24/2022 06:05:01 - INFO - codeparrot_training - Step 3482: {'lr': 0.0004988248774383978, 'samples': 1783296, 'steps': 3482, 'loss/train': 0.30636340379714966} 02/24/2022 06:05:07 - INFO - codeparrot_training - Step 3483: {'lr': 0.0004988232922886049, 'samples': 1783808, 'steps': 3483, 'loss/train': 2.5050477981567383} 02/24/2022 06:05:10 - INFO - codeparrot_training - Step 3484: {'lr': 0.0004988217060729321, 'samples': 1784320, 'steps': 3484, 'loss/train': 2.3946900367736816} 02/24/2022 06:05:16 - INFO - codeparrot_training - Step 3485: {'lr': 0.0004988201187913861, 'samples': 1784832, 'steps': 3485, 'loss/train': 1.218027949333191} 02/24/2022 06:05:19 - INFO - codeparrot_training - Step 3486: {'lr': 0.0004988185304439737, 'samples': 1785344, 'steps': 3486, 'loss/train': 2.8386361598968506} 02/24/2022 06:05:25 - INFO - codeparrot_training - Step 3487: {'lr': 0.0004988169410307018, 'samples': 1785856, 'steps': 3487, 'loss/train': 3.3929803371429443} 02/24/2022 06:05:28 - INFO - codeparrot_training - Step 3488: {'lr': 0.0004988153505515771, 'samples': 1786368, 'steps': 3488, 'loss/train': 2.450044631958008} 02/24/2022 06:05:35 - INFO - codeparrot_training - Step 3489: {'lr': 0.0004988137590066064, 'samples': 1786880, 'steps': 3489, 'loss/train': 2.801856756210327} 02/24/2022 06:05:40 - INFO - codeparrot_training - Step 3490: {'lr': 0.0004988121663957966, 'samples': 1787392, 'steps': 3490, 'loss/train': 2.518889904022217} 02/24/2022 06:05:44 - INFO - codeparrot_training - Step 3491: {'lr': 0.0004988105727191546, 'samples': 1787904, 'steps': 3491, 'loss/train': 1.3204587697982788} 02/24/2022 06:05:49 - INFO - codeparrot_training - Step 3492: {'lr': 0.0004988089779766869, 'samples': 1788416, 'steps': 3492, 'loss/train': 2.1671030521392822} 02/24/2022 06:05:53 - INFO - codeparrot_training - Step 3493: {'lr': 0.0004988073821684006, 'samples': 1788928, 'steps': 3493, 'loss/train': 3.2876837253570557} 02/24/2022 06:05:58 - INFO - codeparrot_training - Step 3494: {'lr': 0.0004988057852943025, 'samples': 1789440, 'steps': 3494, 'loss/train': 2.6805965900421143} 02/24/2022 06:06:02 - INFO - codeparrot_training - Step 3495: {'lr': 0.0004988041873543995, 'samples': 1789952, 'steps': 3495, 'loss/train': 2.0916295051574707} 02/24/2022 06:06:07 - INFO - codeparrot_training - Step 3496: {'lr': 0.0004988025883486983, 'samples': 1790464, 'steps': 3496, 'loss/train': 2.741621494293213} 02/24/2022 06:06:11 - INFO - codeparrot_training - Step 3497: {'lr': 0.0004988009882772058, 'samples': 1790976, 'steps': 3497, 'loss/train': 3.0412755012512207} 02/24/2022 06:06:18 - INFO - codeparrot_training - Step 3498: {'lr': 0.0004987993871399289, 'samples': 1791488, 'steps': 3498, 'loss/train': 4.121082305908203} 02/24/2022 06:06:21 - INFO - codeparrot_training - Step 3499: {'lr': 0.0004987977849368744, 'samples': 1792000, 'steps': 3499, 'loss/train': 3.4514198303222656} 02/24/2022 06:06:27 - INFO - codeparrot_training - Step 3500: {'lr': 0.0004987961816680492, 'samples': 1792512, 'steps': 3500, 'loss/train': 3.965061664581299} 02/24/2022 06:06:30 - INFO - codeparrot_training - Step 3501: {'lr': 0.0004987945773334602, 'samples': 1793024, 'steps': 3501, 'loss/train': 2.789762496948242} 02/24/2022 06:06:36 - INFO - codeparrot_training - Step 3502: {'lr': 0.0004987929719331142, 'samples': 1793536, 'steps': 3502, 'loss/train': 2.0071256160736084} 02/24/2022 06:06:39 - INFO - codeparrot_training - Step 3503: {'lr': 0.0004987913654670181, 'samples': 1794048, 'steps': 3503, 'loss/train': 3.3820815086364746} 02/24/2022 06:06:44 - INFO - codeparrot_training - Step 3504: {'lr': 0.0004987897579351787, 'samples': 1794560, 'steps': 3504, 'loss/train': 2.395979881286621} 02/24/2022 06:06:48 - INFO - codeparrot_training - Step 3505: {'lr': 0.0004987881493376032, 'samples': 1795072, 'steps': 3505, 'loss/train': 3.3991336822509766} 02/24/2022 06:06:54 - INFO - codeparrot_training - Step 3506: {'lr': 0.0004987865396742981, 'samples': 1795584, 'steps': 3506, 'loss/train': 3.1336047649383545} 02/24/2022 06:06:57 - INFO - codeparrot_training - Step 3507: {'lr': 0.0004987849289452705, 'samples': 1796096, 'steps': 3507, 'loss/train': 0.3818409740924835} 02/24/2022 06:07:04 - INFO - codeparrot_training - Step 3508: {'lr': 0.0004987833171505272, 'samples': 1796608, 'steps': 3508, 'loss/train': 3.589928388595581} 02/24/2022 06:07:07 - INFO - codeparrot_training - Step 3509: {'lr': 0.0004987817042900753, 'samples': 1797120, 'steps': 3509, 'loss/train': 2.944444417953491} 02/24/2022 06:07:13 - INFO - codeparrot_training - Step 3510: {'lr': 0.0004987800903639216, 'samples': 1797632, 'steps': 3510, 'loss/train': 3.009467124938965} 02/24/2022 06:07:16 - INFO - codeparrot_training - Step 3511: {'lr': 0.0004987784753720728, 'samples': 1798144, 'steps': 3511, 'loss/train': 1.7856380939483643} 02/24/2022 06:07:22 - INFO - codeparrot_training - Step 3512: {'lr': 0.0004987768593145362, 'samples': 1798656, 'steps': 3512, 'loss/train': 0.23413746058940887} 02/24/2022 06:07:25 - INFO - codeparrot_training - Step 3513: {'lr': 0.0004987752421913185, 'samples': 1799168, 'steps': 3513, 'loss/train': 2.859423875808716} 02/24/2022 06:07:31 - INFO - codeparrot_training - Step 3514: {'lr': 0.0004987736240024264, 'samples': 1799680, 'steps': 3514, 'loss/train': 2.9557743072509766} 02/24/2022 06:07:34 - INFO - codeparrot_training - Step 3515: {'lr': 0.0004987720047478673, 'samples': 1800192, 'steps': 3515, 'loss/train': 2.4329938888549805} 02/24/2022 06:07:40 - INFO - codeparrot_training - Step 3516: {'lr': 0.000498770384427648, 'samples': 1800704, 'steps': 3516, 'loss/train': 2.0503909587860107} 02/24/2022 06:07:43 - INFO - codeparrot_training - Step 3517: {'lr': 0.0004987687630417753, 'samples': 1801216, 'steps': 3517, 'loss/train': 3.767716646194458} 02/24/2022 06:07:49 - INFO - codeparrot_training - Step 3518: {'lr': 0.0004987671405902562, 'samples': 1801728, 'steps': 3518, 'loss/train': 4.070925712585449} 02/24/2022 06:07:52 - INFO - codeparrot_training - Step 3519: {'lr': 0.0004987655170730976, 'samples': 1802240, 'steps': 3519, 'loss/train': 2.934185266494751} 02/24/2022 06:07:58 - INFO - codeparrot_training - Step 3520: {'lr': 0.0004987638924903066, 'samples': 1802752, 'steps': 3520, 'loss/train': 2.184549331665039} 02/24/2022 06:08:01 - INFO - codeparrot_training - Step 3521: {'lr': 0.00049876226684189, 'samples': 1803264, 'steps': 3521, 'loss/train': 2.0489423274993896} 02/24/2022 06:08:07 - INFO - codeparrot_training - Step 3522: {'lr': 0.0004987606401278549, 'samples': 1803776, 'steps': 3522, 'loss/train': 3.404755115509033} 02/24/2022 06:08:10 - INFO - codeparrot_training - Step 3523: {'lr': 0.0004987590123482082, 'samples': 1804288, 'steps': 3523, 'loss/train': 3.2636003494262695} 02/24/2022 06:08:16 - INFO - codeparrot_training - Step 3524: {'lr': 0.0004987573835029569, 'samples': 1804800, 'steps': 3524, 'loss/train': 2.906745433807373} 02/24/2022 06:08:20 - INFO - codeparrot_training - Step 3525: {'lr': 0.0004987557535921079, 'samples': 1805312, 'steps': 3525, 'loss/train': 3.216729164123535} 02/24/2022 06:08:25 - INFO - codeparrot_training - Step 3526: {'lr': 0.0004987541226156683, 'samples': 1805824, 'steps': 3526, 'loss/train': 3.9713149070739746} 02/24/2022 06:08:29 - INFO - codeparrot_training - Step 3527: {'lr': 0.0004987524905736451, 'samples': 1806336, 'steps': 3527, 'loss/train': 2.592583656311035} 02/24/2022 06:08:34 - INFO - codeparrot_training - Step 3528: {'lr': 0.000498750857466045, 'samples': 1806848, 'steps': 3528, 'loss/train': 2.5117557048797607} 02/24/2022 06:08:38 - INFO - codeparrot_training - Step 3529: {'lr': 0.0004987492232928753, 'samples': 1807360, 'steps': 3529, 'loss/train': 3.4634780883789062} 02/24/2022 06:08:43 - INFO - codeparrot_training - Step 3530: {'lr': 0.000498747588054143, 'samples': 1807872, 'steps': 3530, 'loss/train': 2.570119857788086} 02/24/2022 06:08:47 - INFO - codeparrot_training - Step 3531: {'lr': 0.0004987459517498549, 'samples': 1808384, 'steps': 3531, 'loss/train': 2.5747294425964355} 02/24/2022 06:08:52 - INFO - codeparrot_training - Step 3532: {'lr': 0.0004987443143800182, 'samples': 1808896, 'steps': 3532, 'loss/train': 2.667483329772949} 02/24/2022 06:08:56 - INFO - codeparrot_training - Step 3533: {'lr': 0.0004987426759446398, 'samples': 1809408, 'steps': 3533, 'loss/train': 1.96893310546875} 02/24/2022 06:09:02 - INFO - codeparrot_training - Step 3534: {'lr': 0.0004987410364437269, 'samples': 1809920, 'steps': 3534, 'loss/train': 3.6135401725769043} 02/24/2022 06:09:05 - INFO - codeparrot_training - Step 3535: {'lr': 0.0004987393958772862, 'samples': 1810432, 'steps': 3535, 'loss/train': 1.8765099048614502} 02/24/2022 06:09:11 - INFO - codeparrot_training - Step 3536: {'lr': 0.0004987377542453251, 'samples': 1810944, 'steps': 3536, 'loss/train': 2.8146049976348877} 02/24/2022 06:09:14 - INFO - codeparrot_training - Step 3537: {'lr': 0.0004987361115478502, 'samples': 1811456, 'steps': 3537, 'loss/train': 2.16068959236145} 02/24/2022 06:09:20 - INFO - codeparrot_training - Step 3538: {'lr': 0.000498734467784869, 'samples': 1811968, 'steps': 3538, 'loss/train': 2.9196808338165283} 02/24/2022 06:09:23 - INFO - codeparrot_training - Step 3539: {'lr': 0.0004987328229563883, 'samples': 1812480, 'steps': 3539, 'loss/train': 4.4412150382995605} 02/24/2022 06:09:29 - INFO - codeparrot_training - Step 3540: {'lr': 0.0004987311770624151, 'samples': 1812992, 'steps': 3540, 'loss/train': 2.660609483718872} 02/24/2022 06:09:33 - INFO - codeparrot_training - Step 3541: {'lr': 0.0004987295301029565, 'samples': 1813504, 'steps': 3541, 'loss/train': 4.113503456115723} 02/24/2022 06:09:38 - INFO - codeparrot_training - Step 3542: {'lr': 0.0004987278820780196, 'samples': 1814016, 'steps': 3542, 'loss/train': 2.2509279251098633} 02/24/2022 06:09:42 - INFO - codeparrot_training - Step 3543: {'lr': 0.0004987262329876114, 'samples': 1814528, 'steps': 3543, 'loss/train': 4.698426246643066} 02/24/2022 06:09:48 - INFO - codeparrot_training - Step 3544: {'lr': 0.000498724582831739, 'samples': 1815040, 'steps': 3544, 'loss/train': 2.624014139175415} 02/24/2022 06:09:51 - INFO - codeparrot_training - Step 3545: {'lr': 0.0004987229316104095, 'samples': 1815552, 'steps': 3545, 'loss/train': 2.6331729888916016} 02/24/2022 06:09:57 - INFO - codeparrot_training - Step 3546: {'lr': 0.00049872127932363, 'samples': 1816064, 'steps': 3546, 'loss/train': 2.3879826068878174} 02/24/2022 06:10:00 - INFO - codeparrot_training - Step 3547: {'lr': 0.0004987196259714074, 'samples': 1816576, 'steps': 3547, 'loss/train': 2.751999855041504} 02/24/2022 06:10:06 - INFO - codeparrot_training - Step 3548: {'lr': 0.000498717971553749, 'samples': 1817088, 'steps': 3548, 'loss/train': 2.8464198112487793} 02/24/2022 06:10:09 - INFO - codeparrot_training - Step 3549: {'lr': 0.0004987163160706617, 'samples': 1817600, 'steps': 3549, 'loss/train': 1.9754754304885864} 02/24/2022 06:10:15 - INFO - codeparrot_training - Step 3550: {'lr': 0.0004987146595221527, 'samples': 1818112, 'steps': 3550, 'loss/train': 1.407125473022461} 02/24/2022 06:10:18 - INFO - codeparrot_training - Step 3551: {'lr': 0.0004987130019082291, 'samples': 1818624, 'steps': 3551, 'loss/train': 2.565162420272827} 02/24/2022 06:10:24 - INFO - codeparrot_training - Step 3552: {'lr': 0.000498711343228898, 'samples': 1819136, 'steps': 3552, 'loss/train': 3.4164676666259766} 02/24/2022 06:10:27 - INFO - codeparrot_training - Step 3553: {'lr': 0.0004987096834841665, 'samples': 1819648, 'steps': 3553, 'loss/train': 3.1859078407287598} 02/24/2022 06:10:34 - INFO - codeparrot_training - Step 3554: {'lr': 0.0004987080226740416, 'samples': 1820160, 'steps': 3554, 'loss/train': 2.8668935298919678} 02/24/2022 06:10:37 - INFO - codeparrot_training - Step 3555: {'lr': 0.0004987063607985305, 'samples': 1820672, 'steps': 3555, 'loss/train': 2.7604565620422363} 02/24/2022 06:10:43 - INFO - codeparrot_training - Step 3556: {'lr': 0.0004987046978576404, 'samples': 1821184, 'steps': 3556, 'loss/train': 2.679964303970337} 02/24/2022 06:10:46 - INFO - codeparrot_training - Step 3557: {'lr': 0.0004987030338513783, 'samples': 1821696, 'steps': 3557, 'loss/train': 2.652012348175049} 02/24/2022 06:10:52 - INFO - codeparrot_training - Step 3558: {'lr': 0.0004987013687797514, 'samples': 1822208, 'steps': 3558, 'loss/train': 3.0175700187683105} 02/24/2022 06:10:55 - INFO - codeparrot_training - Step 3559: {'lr': 0.0004986997026427668, 'samples': 1822720, 'steps': 3559, 'loss/train': 3.9427781105041504} 02/24/2022 06:11:01 - INFO - codeparrot_training - Step 3560: {'lr': 0.0004986980354404316, 'samples': 1823232, 'steps': 3560, 'loss/train': 2.6489572525024414} 02/24/2022 06:11:04 - INFO - codeparrot_training - Step 3561: {'lr': 0.000498696367172753, 'samples': 1823744, 'steps': 3561, 'loss/train': 1.2118130922317505} 02/24/2022 06:11:10 - INFO - codeparrot_training - Step 3562: {'lr': 0.0004986946978397382, 'samples': 1824256, 'steps': 3562, 'loss/train': 3.7711474895477295} 02/24/2022 06:11:13 - INFO - codeparrot_training - Step 3563: {'lr': 0.0004986930274413942, 'samples': 1824768, 'steps': 3563, 'loss/train': 3.0194289684295654} 02/24/2022 06:11:19 - INFO - codeparrot_training - Step 3564: {'lr': 0.0004986913559777283, 'samples': 1825280, 'steps': 3564, 'loss/train': 2.7172155380249023} 02/24/2022 06:11:22 - INFO - codeparrot_training - Step 3565: {'lr': 0.0004986896834487477, 'samples': 1825792, 'steps': 3565, 'loss/train': 2.6076767444610596} 02/24/2022 06:11:28 - INFO - codeparrot_training - Step 3566: {'lr': 0.0004986880098544593, 'samples': 1826304, 'steps': 3566, 'loss/train': 3.3423800468444824} 02/24/2022 06:11:31 - INFO - codeparrot_training - Step 3567: {'lr': 0.0004986863351948705, 'samples': 1826816, 'steps': 3567, 'loss/train': 5.826650619506836} 02/24/2022 06:11:37 - INFO - codeparrot_training - Step 3568: {'lr': 0.0004986846594699883, 'samples': 1827328, 'steps': 3568, 'loss/train': 5.136133193969727} 02/24/2022 06:11:40 - INFO - codeparrot_training - Step 3569: {'lr': 0.0004986829826798202, 'samples': 1827840, 'steps': 3569, 'loss/train': 2.5577759742736816} 02/24/2022 06:11:47 - INFO - codeparrot_training - Step 3570: {'lr': 0.0004986813048243729, 'samples': 1828352, 'steps': 3570, 'loss/train': 3.0491981506347656} 02/24/2022 06:11:50 - INFO - codeparrot_training - Step 3571: {'lr': 0.000498679625903654, 'samples': 1828864, 'steps': 3571, 'loss/train': 2.9680490493774414} 02/24/2022 06:11:56 - INFO - codeparrot_training - Step 3572: {'lr': 0.0004986779459176706, 'samples': 1829376, 'steps': 3572, 'loss/train': 2.868858814239502} 02/24/2022 06:11:59 - INFO - codeparrot_training - Step 3573: {'lr': 0.0004986762648664298, 'samples': 1829888, 'steps': 3573, 'loss/train': 2.118520498275757} 02/24/2022 06:12:05 - INFO - codeparrot_training - Step 3574: {'lr': 0.0004986745827499389, 'samples': 1830400, 'steps': 3574, 'loss/train': 3.633444309234619} 02/24/2022 06:12:08 - INFO - codeparrot_training - Step 3575: {'lr': 0.0004986728995682049, 'samples': 1830912, 'steps': 3575, 'loss/train': 2.089087963104248} 02/24/2022 06:12:14 - INFO - codeparrot_training - Step 3576: {'lr': 0.0004986712153212352, 'samples': 1831424, 'steps': 3576, 'loss/train': 2.3902127742767334} 02/24/2022 06:12:17 - INFO - codeparrot_training - Step 3577: {'lr': 0.0004986695300090371, 'samples': 1831936, 'steps': 3577, 'loss/train': 2.7970895767211914} 02/24/2022 06:12:23 - INFO - codeparrot_training - Step 3578: {'lr': 0.0004986678436316175, 'samples': 1832448, 'steps': 3578, 'loss/train': 3.4691977500915527} 02/24/2022 06:12:26 - INFO - codeparrot_training - Step 3579: {'lr': 0.000498666156188984, 'samples': 1832960, 'steps': 3579, 'loss/train': 2.964557647705078} 02/24/2022 06:12:32 - INFO - codeparrot_training - Step 3580: {'lr': 0.0004986644676811436, 'samples': 1833472, 'steps': 3580, 'loss/train': 1.213300108909607} 02/24/2022 06:12:36 - INFO - codeparrot_training - Step 3581: {'lr': 0.0004986627781081035, 'samples': 1833984, 'steps': 3581, 'loss/train': 3.4480254650115967} 02/24/2022 06:12:41 - INFO - codeparrot_training - Step 3582: {'lr': 0.0004986610874698712, 'samples': 1834496, 'steps': 3582, 'loss/train': 2.341282844543457} 02/24/2022 06:12:45 - INFO - codeparrot_training - Step 3583: {'lr': 0.0004986593957664536, 'samples': 1835008, 'steps': 3583, 'loss/train': 2.6536827087402344} 02/24/2022 06:12:50 - INFO - codeparrot_training - Step 3584: {'lr': 0.0004986577029978581, 'samples': 1835520, 'steps': 3584, 'loss/train': 4.221004009246826} 02/24/2022 06:12:54 - INFO - codeparrot_training - Step 3585: {'lr': 0.000498656009164092, 'samples': 1836032, 'steps': 3585, 'loss/train': 0.5307567119598389} 02/24/2022 06:12:59 - INFO - codeparrot_training - Step 3586: {'lr': 0.0004986543142651625, 'samples': 1836544, 'steps': 3586, 'loss/train': 0.7687624096870422} 02/24/2022 06:13:03 - INFO - codeparrot_training - Step 3587: {'lr': 0.0004986526183010769, 'samples': 1837056, 'steps': 3587, 'loss/train': 1.857844591140747} 02/24/2022 06:13:08 - INFO - codeparrot_training - Step 3588: {'lr': 0.0004986509212718425, 'samples': 1837568, 'steps': 3588, 'loss/train': 4.479802131652832} 02/24/2022 06:13:12 - INFO - codeparrot_training - Step 3589: {'lr': 0.0004986492231774664, 'samples': 1838080, 'steps': 3589, 'loss/train': 3.233942747116089} 02/24/2022 06:13:18 - INFO - codeparrot_training - Step 3590: {'lr': 0.0004986475240179559, 'samples': 1838592, 'steps': 3590, 'loss/train': 3.5459752082824707} 02/24/2022 06:13:24 - INFO - codeparrot_training - Step 3591: {'lr': 0.0004986458237933185, 'samples': 1839104, 'steps': 3591, 'loss/train': 3.054870843887329} 02/24/2022 06:13:27 - INFO - codeparrot_training - Step 3592: {'lr': 0.0004986441225035614, 'samples': 1839616, 'steps': 3592, 'loss/train': 2.552980899810791} 02/24/2022 06:13:33 - INFO - codeparrot_training - Step 3593: {'lr': 0.0004986424201486918, 'samples': 1840128, 'steps': 3593, 'loss/train': 3.2321536540985107} 02/24/2022 06:13:36 - INFO - codeparrot_training - Step 3594: {'lr': 0.000498640716728717, 'samples': 1840640, 'steps': 3594, 'loss/train': 3.447997808456421} 02/24/2022 06:13:42 - INFO - codeparrot_training - Step 3595: {'lr': 0.0004986390122436443, 'samples': 1841152, 'steps': 3595, 'loss/train': 2.6887683868408203} 02/24/2022 06:13:45 - INFO - codeparrot_training - Step 3596: {'lr': 0.000498637306693481, 'samples': 1841664, 'steps': 3596, 'loss/train': 2.5113518238067627} 02/24/2022 06:13:51 - INFO - codeparrot_training - Step 3597: {'lr': 0.0004986356000782345, 'samples': 1842176, 'steps': 3597, 'loss/train': 3.2579522132873535} 02/24/2022 06:13:54 - INFO - codeparrot_training - Step 3598: {'lr': 0.0004986338923979119, 'samples': 1842688, 'steps': 3598, 'loss/train': 2.677898645401001} 02/24/2022 06:14:00 - INFO - codeparrot_training - Step 3599: {'lr': 0.0004986321836525209, 'samples': 1843200, 'steps': 3599, 'loss/train': 1.6726078987121582} 02/24/2022 06:14:04 - INFO - codeparrot_training - Step 3600: {'lr': 0.0004986304738420684, 'samples': 1843712, 'steps': 3600, 'loss/train': 2.5096471309661865} 02/24/2022 06:14:09 - INFO - codeparrot_training - Step 3601: {'lr': 0.0004986287629665619, 'samples': 1844224, 'steps': 3601, 'loss/train': 3.162659168243408} 02/24/2022 06:14:13 - INFO - codeparrot_training - Step 3602: {'lr': 0.0004986270510260087, 'samples': 1844736, 'steps': 3602, 'loss/train': 3.1090872287750244} 02/24/2022 06:14:18 - INFO - codeparrot_training - Step 3603: {'lr': 0.0004986253380204163, 'samples': 1845248, 'steps': 3603, 'loss/train': 2.805504322052002} 02/24/2022 06:14:22 - INFO - codeparrot_training - Step 3604: {'lr': 0.0004986236239497918, 'samples': 1845760, 'steps': 3604, 'loss/train': 3.178154230117798} 02/24/2022 06:14:27 - INFO - codeparrot_training - Step 3605: {'lr': 0.0004986219088141426, 'samples': 1846272, 'steps': 3605, 'loss/train': 2.967597246170044} 02/24/2022 06:14:31 - INFO - codeparrot_training - Step 3606: {'lr': 0.0004986201926134761, 'samples': 1846784, 'steps': 3606, 'loss/train': 2.679542064666748} 02/24/2022 06:14:37 - INFO - codeparrot_training - Step 3607: {'lr': 0.0004986184753477998, 'samples': 1847296, 'steps': 3607, 'loss/train': 3.167675256729126} 02/24/2022 06:14:40 - INFO - codeparrot_training - Step 3608: {'lr': 0.0004986167570171208, 'samples': 1847808, 'steps': 3608, 'loss/train': 2.4734816551208496} 02/24/2022 06:14:44 - INFO - codeparrot_training - Step 3609: {'lr': 0.0004986150376214465, 'samples': 1848320, 'steps': 3609, 'loss/train': 3.0577664375305176} 02/24/2022 06:14:49 - INFO - codeparrot_training - Step 3610: {'lr': 0.0004986133171607844, 'samples': 1848832, 'steps': 3610, 'loss/train': 3.364696502685547} 02/24/2022 06:14:53 - INFO - codeparrot_training - Step 3611: {'lr': 0.0004986115956351417, 'samples': 1849344, 'steps': 3611, 'loss/train': 2.9301700592041016} 02/24/2022 06:14:58 - INFO - codeparrot_training - Step 3612: {'lr': 0.000498609873044526, 'samples': 1849856, 'steps': 3612, 'loss/train': 2.5401082038879395} 02/24/2022 06:15:02 - INFO - codeparrot_training - Step 3613: {'lr': 0.0004986081493889444, 'samples': 1850368, 'steps': 3613, 'loss/train': 3.115487813949585} 02/24/2022 06:15:07 - INFO - codeparrot_training - Step 3614: {'lr': 0.0004986064246684046, 'samples': 1850880, 'steps': 3614, 'loss/train': 3.4398844242095947} 02/24/2022 06:15:11 - INFO - codeparrot_training - Step 3615: {'lr': 0.0004986046988829136, 'samples': 1851392, 'steps': 3615, 'loss/train': 3.2407708168029785} 02/24/2022 06:15:17 - INFO - codeparrot_training - Step 3616: {'lr': 0.0004986029720324791, 'samples': 1851904, 'steps': 3616, 'loss/train': 2.397080898284912} 02/24/2022 06:15:20 - INFO - codeparrot_training - Step 3617: {'lr': 0.0004986012441171085, 'samples': 1852416, 'steps': 3617, 'loss/train': 4.38129186630249} 02/24/2022 06:15:26 - INFO - codeparrot_training - Step 3618: {'lr': 0.000498599515136809, 'samples': 1852928, 'steps': 3618, 'loss/train': 3.37368106842041} 02/24/2022 06:15:29 - INFO - codeparrot_training - Step 3619: {'lr': 0.0004985977850915882, 'samples': 1853440, 'steps': 3619, 'loss/train': 1.7670608758926392} 02/24/2022 06:15:35 - INFO - codeparrot_training - Step 3620: {'lr': 0.0004985960539814534, 'samples': 1853952, 'steps': 3620, 'loss/train': 2.857828140258789} 02/24/2022 06:15:38 - INFO - codeparrot_training - Step 3621: {'lr': 0.000498594321806412, 'samples': 1854464, 'steps': 3621, 'loss/train': 2.5764071941375732} 02/24/2022 06:15:44 - INFO - codeparrot_training - Step 3622: {'lr': 0.0004985925885664716, 'samples': 1854976, 'steps': 3622, 'loss/train': 2.8645498752593994} 02/24/2022 06:15:49 - INFO - codeparrot_training - Step 3623: {'lr': 0.0004985908542616393, 'samples': 1855488, 'steps': 3623, 'loss/train': 0.915501594543457} 02/24/2022 06:15:53 - INFO - codeparrot_training - Step 3624: {'lr': 0.0004985891188919229, 'samples': 1856000, 'steps': 3624, 'loss/train': 3.1070609092712402} 02/24/2022 06:15:58 - INFO - codeparrot_training - Step 3625: {'lr': 0.0004985873824573296, 'samples': 1856512, 'steps': 3625, 'loss/train': 3.562913179397583} 02/24/2022 06:16:02 - INFO - codeparrot_training - Step 3626: {'lr': 0.0004985856449578667, 'samples': 1857024, 'steps': 3626, 'loss/train': 0.7748192548751831} 02/24/2022 06:16:08 - INFO - codeparrot_training - Step 3627: {'lr': 0.0004985839063935421, 'samples': 1857536, 'steps': 3627, 'loss/train': 3.190570116043091} 02/24/2022 06:16:12 - INFO - codeparrot_training - Step 3628: {'lr': 0.0004985821667643628, 'samples': 1858048, 'steps': 3628, 'loss/train': 2.9273016452789307} 02/24/2022 06:16:17 - INFO - codeparrot_training - Step 3629: {'lr': 0.0004985804260703364, 'samples': 1858560, 'steps': 3629, 'loss/train': 2.5061187744140625} 02/24/2022 06:16:21 - INFO - codeparrot_training - Step 3630: {'lr': 0.0004985786843114706, 'samples': 1859072, 'steps': 3630, 'loss/train': 1.3142304420471191} 02/24/2022 06:16:26 - INFO - codeparrot_training - Step 3631: {'lr': 0.0004985769414877725, 'samples': 1859584, 'steps': 3631, 'loss/train': 1.9471021890640259} 02/24/2022 06:16:30 - INFO - codeparrot_training - Step 3632: {'lr': 0.0004985751975992497, 'samples': 1860096, 'steps': 3632, 'loss/train': 1.9094825983047485} 02/24/2022 06:16:35 - INFO - codeparrot_training - Step 3633: {'lr': 0.0004985734526459098, 'samples': 1860608, 'steps': 3633, 'loss/train': 3.0633323192596436} 02/24/2022 06:16:39 - INFO - codeparrot_training - Step 3634: {'lr': 0.0004985717066277601, 'samples': 1861120, 'steps': 3634, 'loss/train': 2.8317251205444336} 02/24/2022 06:16:44 - INFO - codeparrot_training - Step 3635: {'lr': 0.0004985699595448081, 'samples': 1861632, 'steps': 3635, 'loss/train': 2.7256269454956055} 02/24/2022 06:16:48 - INFO - codeparrot_training - Step 3636: {'lr': 0.0004985682113970613, 'samples': 1862144, 'steps': 3636, 'loss/train': 4.393418312072754} 02/24/2022 06:16:54 - INFO - codeparrot_training - Step 3637: {'lr': 0.0004985664621845273, 'samples': 1862656, 'steps': 3637, 'loss/train': 2.0420732498168945} 02/24/2022 06:16:58 - INFO - codeparrot_training - Step 3638: {'lr': 0.0004985647119072135, 'samples': 1863168, 'steps': 3638, 'loss/train': 2.6048779487609863} 02/24/2022 06:17:03 - INFO - codeparrot_training - Step 3639: {'lr': 0.0004985629605651273, 'samples': 1863680, 'steps': 3639, 'loss/train': 3.3446872234344482} 02/24/2022 06:17:07 - INFO - codeparrot_training - Step 3640: {'lr': 0.0004985612081582763, 'samples': 1864192, 'steps': 3640, 'loss/train': 0.5013619065284729} 02/24/2022 06:17:12 - INFO - codeparrot_training - Step 3641: {'lr': 0.0004985594546866682, 'samples': 1864704, 'steps': 3641, 'loss/train': 3.417056083679199} 02/24/2022 06:17:16 - INFO - codeparrot_training - Step 3642: {'lr': 0.0004985577001503102, 'samples': 1865216, 'steps': 3642, 'loss/train': 4.4992995262146} 02/24/2022 06:17:21 - INFO - codeparrot_training - Step 3643: {'lr': 0.0004985559445492099, 'samples': 1865728, 'steps': 3643, 'loss/train': 2.3566882610321045} 02/24/2022 06:17:25 - INFO - codeparrot_training - Step 3644: {'lr': 0.0004985541878833749, 'samples': 1866240, 'steps': 3644, 'loss/train': 1.6865768432617188} 02/24/2022 06:17:30 - INFO - codeparrot_training - Step 3645: {'lr': 0.0004985524301528127, 'samples': 1866752, 'steps': 3645, 'loss/train': 3.049612522125244} 02/24/2022 06:17:34 - INFO - codeparrot_training - Step 3646: {'lr': 0.0004985506713575307, 'samples': 1867264, 'steps': 3646, 'loss/train': 3.0081145763397217} 02/24/2022 06:17:40 - INFO - codeparrot_training - Step 3647: {'lr': 0.0004985489114975368, 'samples': 1867776, 'steps': 3647, 'loss/train': 2.305170774459839} 02/24/2022 06:17:44 - INFO - codeparrot_training - Step 3648: {'lr': 0.0004985471505728381, 'samples': 1868288, 'steps': 3648, 'loss/train': 2.7749903202056885} 02/24/2022 06:17:49 - INFO - codeparrot_training - Step 3649: {'lr': 0.0004985453885834423, 'samples': 1868800, 'steps': 3649, 'loss/train': 3.4093470573425293} 02/24/2022 06:17:53 - INFO - codeparrot_training - Step 3650: {'lr': 0.0004985436255293571, 'samples': 1869312, 'steps': 3650, 'loss/train': 3.6422040462493896} 02/24/2022 06:17:58 - INFO - codeparrot_training - Step 3651: {'lr': 0.0004985418614105898, 'samples': 1869824, 'steps': 3651, 'loss/train': 3.808675527572632} 02/24/2022 06:18:02 - INFO - codeparrot_training - Step 3652: {'lr': 0.0004985400962271482, 'samples': 1870336, 'steps': 3652, 'loss/train': 5.389224529266357} 02/24/2022 06:18:07 - INFO - codeparrot_training - Step 3653: {'lr': 0.0004985383299790397, 'samples': 1870848, 'steps': 3653, 'loss/train': 2.6073951721191406} 02/24/2022 06:18:11 - INFO - codeparrot_training - Step 3654: {'lr': 0.0004985365626662719, 'samples': 1871360, 'steps': 3654, 'loss/train': 2.7892062664031982} 02/24/2022 06:18:16 - INFO - codeparrot_training - Step 3655: {'lr': 0.0004985347942888524, 'samples': 1871872, 'steps': 3655, 'loss/train': 1.9611306190490723} 02/24/2022 06:18:20 - INFO - codeparrot_training - Step 3656: {'lr': 0.0004985330248467888, 'samples': 1872384, 'steps': 3656, 'loss/train': 3.443881034851074} 02/24/2022 06:18:25 - INFO - codeparrot_training - Step 3657: {'lr': 0.0004985312543400886, 'samples': 1872896, 'steps': 3657, 'loss/train': 1.3975470066070557} 02/24/2022 06:18:29 - INFO - codeparrot_training - Step 3658: {'lr': 0.0004985294827687594, 'samples': 1873408, 'steps': 3658, 'loss/train': 2.50839900970459} 02/24/2022 06:18:34 - INFO - codeparrot_training - Step 3659: {'lr': 0.0004985277101328088, 'samples': 1873920, 'steps': 3659, 'loss/train': 1.0921086072921753} 02/24/2022 06:18:37 - INFO - codeparrot_training - Step 3660: {'lr': 0.0004985259364322445, 'samples': 1874432, 'steps': 3660, 'loss/train': 2.4239721298217773} 02/24/2022 06:18:43 - INFO - codeparrot_training - Step 3661: {'lr': 0.0004985241616670739, 'samples': 1874944, 'steps': 3661, 'loss/train': 3.2412805557250977} 02/24/2022 06:18:46 - INFO - codeparrot_training - Step 3662: {'lr': 0.0004985223858373048, 'samples': 1875456, 'steps': 3662, 'loss/train': 2.599910020828247} 02/24/2022 06:18:53 - INFO - codeparrot_training - Step 3663: {'lr': 0.0004985206089429447, 'samples': 1875968, 'steps': 3663, 'loss/train': 1.2451457977294922} 02/24/2022 06:18:56 - INFO - codeparrot_training - Step 3664: {'lr': 0.0004985188309840012, 'samples': 1876480, 'steps': 3664, 'loss/train': 2.9755656719207764} 02/24/2022 06:19:02 - INFO - codeparrot_training - Step 3665: {'lr': 0.0004985170519604819, 'samples': 1876992, 'steps': 3665, 'loss/train': 3.2344369888305664} 02/24/2022 06:19:05 - INFO - codeparrot_training - Step 3666: {'lr': 0.0004985152718723944, 'samples': 1877504, 'steps': 3666, 'loss/train': 2.8349993228912354} 02/24/2022 06:19:11 - INFO - codeparrot_training - Step 3667: {'lr': 0.0004985134907197466, 'samples': 1878016, 'steps': 3667, 'loss/train': 4.3324456214904785} 02/24/2022 06:19:14 - INFO - codeparrot_training - Step 3668: {'lr': 0.0004985117085025458, 'samples': 1878528, 'steps': 3668, 'loss/train': 3.3636906147003174} 02/24/2022 06:19:20 - INFO - codeparrot_training - Step 3669: {'lr': 0.0004985099252207998, 'samples': 1879040, 'steps': 3669, 'loss/train': 3.616454839706421} 02/24/2022 06:19:24 - INFO - codeparrot_training - Step 3670: {'lr': 0.0004985081408745161, 'samples': 1879552, 'steps': 3670, 'loss/train': 3.3376147747039795} 02/24/2022 06:19:29 - INFO - codeparrot_training - Step 3671: {'lr': 0.0004985063554637025, 'samples': 1880064, 'steps': 3671, 'loss/train': 2.973696708679199} 02/24/2022 06:19:33 - INFO - codeparrot_training - Step 3672: {'lr': 0.0004985045689883665, 'samples': 1880576, 'steps': 3672, 'loss/train': 4.10275936126709} 02/24/2022 06:19:39 - INFO - codeparrot_training - Step 3673: {'lr': 0.0004985027814485159, 'samples': 1881088, 'steps': 3673, 'loss/train': 0.3154342472553253} 02/24/2022 06:19:42 - INFO - codeparrot_training - Step 3674: {'lr': 0.0004985009928441584, 'samples': 1881600, 'steps': 3674, 'loss/train': 3.094592571258545} 02/24/2022 06:19:48 - INFO - codeparrot_training - Step 3675: {'lr': 0.0004984992031753014, 'samples': 1882112, 'steps': 3675, 'loss/train': 2.868638038635254} 02/24/2022 06:19:51 - INFO - codeparrot_training - Step 3676: {'lr': 0.0004984974124419528, 'samples': 1882624, 'steps': 3676, 'loss/train': 2.966783046722412} 02/24/2022 06:19:57 - INFO - codeparrot_training - Step 3677: {'lr': 0.0004984956206441201, 'samples': 1883136, 'steps': 3677, 'loss/train': 2.5661163330078125} 02/24/2022 06:20:01 - INFO - codeparrot_training - Step 3678: {'lr': 0.0004984938277818112, 'samples': 1883648, 'steps': 3678, 'loss/train': 2.9959144592285156} 02/24/2022 06:20:06 - INFO - codeparrot_training - Step 3679: {'lr': 0.0004984920338550335, 'samples': 1884160, 'steps': 3679, 'loss/train': 3.004345178604126} 02/24/2022 06:20:09 - INFO - codeparrot_training - Step 3680: {'lr': 0.0004984902388637949, 'samples': 1884672, 'steps': 3680, 'loss/train': 2.672820568084717} 02/24/2022 06:20:15 - INFO - codeparrot_training - Step 3681: {'lr': 0.0004984884428081031, 'samples': 1885184, 'steps': 3681, 'loss/train': 2.1061084270477295} 02/24/2022 06:20:18 - INFO - codeparrot_training - Step 3682: {'lr': 0.0004984866456879657, 'samples': 1885696, 'steps': 3682, 'loss/train': 2.685220956802368} 02/24/2022 06:20:25 - INFO - codeparrot_training - Step 3683: {'lr': 0.0004984848475033903, 'samples': 1886208, 'steps': 3683, 'loss/train': 2.81203031539917} 02/24/2022 06:20:28 - INFO - codeparrot_training - Step 3684: {'lr': 0.0004984830482543847, 'samples': 1886720, 'steps': 3684, 'loss/train': 3.422428607940674} 02/24/2022 06:20:34 - INFO - codeparrot_training - Step 3685: {'lr': 0.0004984812479409568, 'samples': 1887232, 'steps': 3685, 'loss/train': 2.7525634765625} 02/24/2022 06:20:37 - INFO - codeparrot_training - Step 3686: {'lr': 0.000498479446563114, 'samples': 1887744, 'steps': 3686, 'loss/train': 2.6657323837280273} 02/24/2022 06:20:43 - INFO - codeparrot_training - Step 3687: {'lr': 0.0004984776441208642, 'samples': 1888256, 'steps': 3687, 'loss/train': 3.3297717571258545} 02/24/2022 06:20:46 - INFO - codeparrot_training - Step 3688: {'lr': 0.000498475840614215, 'samples': 1888768, 'steps': 3688, 'loss/train': 1.8983154296875} 02/24/2022 06:20:52 - INFO - codeparrot_training - Step 3689: {'lr': 0.0004984740360431742, 'samples': 1889280, 'steps': 3689, 'loss/train': 3.2670509815216064} 02/24/2022 06:20:55 - INFO - codeparrot_training - Step 3690: {'lr': 0.0004984722304077496, 'samples': 1889792, 'steps': 3690, 'loss/train': 3.821131944656372} 02/24/2022 06:21:01 - INFO - codeparrot_training - Step 3691: {'lr': 0.0004984704237079489, 'samples': 1890304, 'steps': 3691, 'loss/train': 1.655848503112793} 02/24/2022 06:21:04 - INFO - codeparrot_training - Step 3692: {'lr': 0.0004984686159437798, 'samples': 1890816, 'steps': 3692, 'loss/train': 3.352936267852783} 02/24/2022 06:21:10 - INFO - codeparrot_training - Step 3693: {'lr': 0.00049846680711525, 'samples': 1891328, 'steps': 3693, 'loss/train': 1.884761095046997} 02/24/2022 06:21:14 - INFO - codeparrot_training - Step 3694: {'lr': 0.0004984649972223673, 'samples': 1891840, 'steps': 3694, 'loss/train': 3.1419994831085205} 02/24/2022 06:21:19 - INFO - codeparrot_training - Step 3695: {'lr': 0.0004984631862651395, 'samples': 1892352, 'steps': 3695, 'loss/train': 3.305544137954712} 02/24/2022 06:21:23 - INFO - codeparrot_training - Step 3696: {'lr': 0.0004984613742435742, 'samples': 1892864, 'steps': 3696, 'loss/train': 2.4752044677734375} 02/24/2022 06:21:28 - INFO - codeparrot_training - Step 3697: {'lr': 0.0004984595611576793, 'samples': 1893376, 'steps': 3697, 'loss/train': 2.936878204345703} 02/24/2022 06:21:32 - INFO - codeparrot_training - Step 3698: {'lr': 0.0004984577470074625, 'samples': 1893888, 'steps': 3698, 'loss/train': 1.8516157865524292} 02/24/2022 06:21:37 - INFO - codeparrot_training - Step 3699: {'lr': 0.0004984559317929317, 'samples': 1894400, 'steps': 3699, 'loss/train': 2.8204853534698486} 02/24/2022 06:21:43 - INFO - codeparrot_training - Step 3700: {'lr': 0.0004984541155140946, 'samples': 1894912, 'steps': 3700, 'loss/train': 5.0451788902282715} 02/24/2022 06:21:46 - INFO - codeparrot_training - Step 3701: {'lr': 0.0004984522981709589, 'samples': 1895424, 'steps': 3701, 'loss/train': 3.1432008743286133} 02/24/2022 06:21:52 - INFO - codeparrot_training - Step 3702: {'lr': 0.0004984504797635324, 'samples': 1895936, 'steps': 3702, 'loss/train': 2.3866403102874756} 02/24/2022 06:21:55 - INFO - codeparrot_training - Step 3703: {'lr': 0.000498448660291823, 'samples': 1896448, 'steps': 3703, 'loss/train': 3.1456377506256104} 02/24/2022 06:22:01 - INFO - codeparrot_training - Step 3704: {'lr': 0.0004984468397558384, 'samples': 1896960, 'steps': 3704, 'loss/train': 1.9401761293411255} 02/24/2022 06:22:04 - INFO - codeparrot_training - Step 3705: {'lr': 0.0004984450181555864, 'samples': 1897472, 'steps': 3705, 'loss/train': 2.999276876449585} 02/24/2022 06:22:10 - INFO - codeparrot_training - Step 3706: {'lr': 0.0004984431954910749, 'samples': 1897984, 'steps': 3706, 'loss/train': 2.5989420413970947} 02/24/2022 06:22:13 - INFO - codeparrot_training - Step 3707: {'lr': 0.0004984413717623117, 'samples': 1898496, 'steps': 3707, 'loss/train': 3.1828978061676025} 02/24/2022 06:22:19 - INFO - codeparrot_training - Step 3708: {'lr': 0.0004984395469693044, 'samples': 1899008, 'steps': 3708, 'loss/train': 2.2003378868103027} 02/24/2022 06:22:23 - INFO - codeparrot_training - Step 3709: {'lr': 0.000498437721112061, 'samples': 1899520, 'steps': 3709, 'loss/train': 2.798269510269165} 02/24/2022 06:22:28 - INFO - codeparrot_training - Step 3710: {'lr': 0.0004984358941905894, 'samples': 1900032, 'steps': 3710, 'loss/train': 3.3372676372528076} 02/24/2022 06:22:32 - INFO - codeparrot_training - Step 3711: {'lr': 0.0004984340662048972, 'samples': 1900544, 'steps': 3711, 'loss/train': 3.361177682876587} 02/24/2022 06:22:37 - INFO - codeparrot_training - Step 3712: {'lr': 0.0004984322371549924, 'samples': 1901056, 'steps': 3712, 'loss/train': 3.1103322505950928} 02/24/2022 06:22:41 - INFO - codeparrot_training - Step 3713: {'lr': 0.0004984304070408828, 'samples': 1901568, 'steps': 3713, 'loss/train': 1.899722695350647} 02/24/2022 06:22:47 - INFO - codeparrot_training - Step 3714: {'lr': 0.0004984285758625761, 'samples': 1902080, 'steps': 3714, 'loss/train': 2.3745577335357666} 02/24/2022 06:22:50 - INFO - codeparrot_training - Step 3715: {'lr': 0.0004984267436200805, 'samples': 1902592, 'steps': 3715, 'loss/train': 3.5624730587005615} 02/24/2022 06:22:56 - INFO - codeparrot_training - Step 3716: {'lr': 0.0004984249103134035, 'samples': 1903104, 'steps': 3716, 'loss/train': 2.43900990486145} 02/24/2022 06:22:59 - INFO - codeparrot_training - Step 3717: {'lr': 0.000498423075942553, 'samples': 1903616, 'steps': 3717, 'loss/train': 2.644005298614502} 02/24/2022 06:23:05 - INFO - codeparrot_training - Step 3718: {'lr': 0.0004984212405075369, 'samples': 1904128, 'steps': 3718, 'loss/train': 2.683960199356079} 02/24/2022 06:23:09 - INFO - codeparrot_training - Step 3719: {'lr': 0.0004984194040083632, 'samples': 1904640, 'steps': 3719, 'loss/train': 3.2765417098999023} 02/24/2022 06:23:14 - INFO - codeparrot_training - Step 3720: {'lr': 0.0004984175664450397, 'samples': 1905152, 'steps': 3720, 'loss/train': 3.5208442211151123} 02/24/2022 06:23:18 - INFO - codeparrot_training - Step 3721: {'lr': 0.0004984157278175741, 'samples': 1905664, 'steps': 3721, 'loss/train': 3.1742918491363525} 02/24/2022 06:23:23 - INFO - codeparrot_training - Step 3722: {'lr': 0.0004984138881259744, 'samples': 1906176, 'steps': 3722, 'loss/train': 2.0669407844543457} 02/24/2022 06:23:27 - INFO - codeparrot_training - Step 3723: {'lr': 0.0004984120473702486, 'samples': 1906688, 'steps': 3723, 'loss/train': 2.4079787731170654} 02/24/2022 06:23:32 - INFO - codeparrot_training - Step 3724: {'lr': 0.0004984102055504044, 'samples': 1907200, 'steps': 3724, 'loss/train': 2.955017328262329} 02/24/2022 06:23:36 - INFO - codeparrot_training - Step 3725: {'lr': 0.0004984083626664497, 'samples': 1907712, 'steps': 3725, 'loss/train': 3.057616949081421} 02/24/2022 06:23:41 - INFO - codeparrot_training - Step 3726: {'lr': 0.0004984065187183925, 'samples': 1908224, 'steps': 3726, 'loss/train': 2.7142789363861084} 02/24/2022 06:23:44 - INFO - codeparrot_training - Step 3727: {'lr': 0.0004984046737062407, 'samples': 1908736, 'steps': 3727, 'loss/train': 1.5987069606781006} 02/24/2022 06:23:51 - INFO - codeparrot_training - Step 3728: {'lr': 0.0004984028276300021, 'samples': 1909248, 'steps': 3728, 'loss/train': 2.8336312770843506} 02/24/2022 06:23:54 - INFO - codeparrot_training - Step 3729: {'lr': 0.0004984009804896846, 'samples': 1909760, 'steps': 3729, 'loss/train': 2.7531681060791016} 02/24/2022 06:24:00 - INFO - codeparrot_training - Step 3730: {'lr': 0.0004983991322852963, 'samples': 1910272, 'steps': 3730, 'loss/train': 9.213851928710938} 02/24/2022 06:24:03 - INFO - codeparrot_training - Step 3731: {'lr': 0.000498397283016845, 'samples': 1910784, 'steps': 3731, 'loss/train': 3.1641433238983154} 02/24/2022 06:24:10 - INFO - codeparrot_training - Step 3732: {'lr': 0.0004983954326843386, 'samples': 1911296, 'steps': 3732, 'loss/train': 2.9372360706329346} 02/24/2022 06:24:13 - INFO - codeparrot_training - Step 3733: {'lr': 0.000498393581287785, 'samples': 1911808, 'steps': 3733, 'loss/train': 2.870913505554199} 02/24/2022 06:24:16 - INFO - codeparrot_training - Step 3734: {'lr': 0.0004983917288271921, 'samples': 1912320, 'steps': 3734, 'loss/train': 2.4943997859954834} 02/24/2022 06:24:22 - INFO - codeparrot_training - Step 3735: {'lr': 0.0004983898753025681, 'samples': 1912832, 'steps': 3735, 'loss/train': 2.899331569671631} 02/24/2022 06:24:25 - INFO - codeparrot_training - Step 3736: {'lr': 0.0004983880207139205, 'samples': 1913344, 'steps': 3736, 'loss/train': 2.1709070205688477} 02/24/2022 06:24:31 - INFO - codeparrot_training - Step 3737: {'lr': 0.0004983861650612577, 'samples': 1913856, 'steps': 3737, 'loss/train': 2.6429572105407715} 02/24/2022 06:24:34 - INFO - codeparrot_training - Step 3738: {'lr': 0.0004983843083445873, 'samples': 1914368, 'steps': 3738, 'loss/train': 2.3169102668762207} 02/24/2022 06:24:40 - INFO - codeparrot_training - Step 3739: {'lr': 0.0004983824505639175, 'samples': 1914880, 'steps': 3739, 'loss/train': 2.5287582874298096} 02/24/2022 06:24:43 - INFO - codeparrot_training - Step 3740: {'lr': 0.000498380591719256, 'samples': 1915392, 'steps': 3740, 'loss/train': 3.4230027198791504} 02/24/2022 06:24:49 - INFO - codeparrot_training - Step 3741: {'lr': 0.0004983787318106111, 'samples': 1915904, 'steps': 3741, 'loss/train': 3.4316139221191406} 02/24/2022 06:24:53 - INFO - codeparrot_training - Step 3742: {'lr': 0.0004983768708379905, 'samples': 1916416, 'steps': 3742, 'loss/train': 2.9466638565063477} 02/24/2022 06:24:58 - INFO - codeparrot_training - Step 3743: {'lr': 0.0004983750088014023, 'samples': 1916928, 'steps': 3743, 'loss/train': 2.963587760925293} 02/24/2022 06:25:02 - INFO - codeparrot_training - Step 3744: {'lr': 0.0004983731457008544, 'samples': 1917440, 'steps': 3744, 'loss/train': 1.5645618438720703} 02/24/2022 06:25:07 - INFO - codeparrot_training - Step 3745: {'lr': 0.0004983712815363548, 'samples': 1917952, 'steps': 3745, 'loss/train': 2.88423490524292} 02/24/2022 06:25:11 - INFO - codeparrot_training - Step 3746: {'lr': 0.0004983694163079115, 'samples': 1918464, 'steps': 3746, 'loss/train': 2.7721304893493652} 02/24/2022 06:25:16 - INFO - codeparrot_training - Step 3747: {'lr': 0.0004983675500155325, 'samples': 1918976, 'steps': 3747, 'loss/train': 2.4178385734558105} 02/24/2022 06:25:20 - INFO - codeparrot_training - Step 3748: {'lr': 0.0004983656826592258, 'samples': 1919488, 'steps': 3748, 'loss/train': 7.331582546234131} 02/24/2022 06:25:25 - INFO - codeparrot_training - Step 3749: {'lr': 0.0004983638142389993, 'samples': 1920000, 'steps': 3749, 'loss/train': 2.3395955562591553} 02/24/2022 06:25:29 - INFO - codeparrot_training - Step 3750: {'lr': 0.000498361944754861, 'samples': 1920512, 'steps': 3750, 'loss/train': 3.1834793090820312} 02/24/2022 06:25:34 - INFO - codeparrot_training - Step 3751: {'lr': 0.0004983600742068192, 'samples': 1921024, 'steps': 3751, 'loss/train': 3.083505392074585} 02/24/2022 06:25:38 - INFO - codeparrot_training - Step 3752: {'lr': 0.0004983582025948816, 'samples': 1921536, 'steps': 3752, 'loss/train': 2.9713878631591797} 02/24/2022 06:25:43 - INFO - codeparrot_training - Step 3753: {'lr': 0.0004983563299190564, 'samples': 1922048, 'steps': 3753, 'loss/train': 3.736201763153076} 02/24/2022 06:25:47 - INFO - codeparrot_training - Step 3754: {'lr': 0.0004983544561793515, 'samples': 1922560, 'steps': 3754, 'loss/train': 1.7429077625274658} 02/24/2022 06:25:53 - INFO - codeparrot_training - Step 3755: {'lr': 0.000498352581375775, 'samples': 1923072, 'steps': 3755, 'loss/train': 2.9478325843811035} 02/24/2022 06:25:56 - INFO - codeparrot_training - Step 3756: {'lr': 0.0004983507055083349, 'samples': 1923584, 'steps': 3756, 'loss/train': 2.925215244293213} 02/24/2022 06:26:02 - INFO - codeparrot_training - Step 3757: {'lr': 0.0004983488285770391, 'samples': 1924096, 'steps': 3757, 'loss/train': 3.0634958744049072} 02/24/2022 06:26:05 - INFO - codeparrot_training - Step 3758: {'lr': 0.000498346950581896, 'samples': 1924608, 'steps': 3758, 'loss/train': 2.396843671798706} 02/24/2022 06:26:11 - INFO - codeparrot_training - Step 3759: {'lr': 0.0004983450715229132, 'samples': 1925120, 'steps': 3759, 'loss/train': 2.7595198154449463} 02/24/2022 06:26:14 - INFO - codeparrot_training - Step 3760: {'lr': 0.000498343191400099, 'samples': 1925632, 'steps': 3760, 'loss/train': 3.177227020263672} 02/24/2022 06:26:20 - INFO - codeparrot_training - Step 3761: {'lr': 0.0004983413102134616, 'samples': 1926144, 'steps': 3761, 'loss/train': 3.1766610145568848} 02/24/2022 06:26:23 - INFO - codeparrot_training - Step 3762: {'lr': 0.0004983394279630088, 'samples': 1926656, 'steps': 3762, 'loss/train': 2.3948452472686768} 02/24/2022 06:26:29 - INFO - codeparrot_training - Step 3763: {'lr': 0.0004983375446487488, 'samples': 1927168, 'steps': 3763, 'loss/train': 0.9162213206291199} 02/24/2022 06:26:32 - INFO - codeparrot_training - Step 3764: {'lr': 0.0004983356602706895, 'samples': 1927680, 'steps': 3764, 'loss/train': 2.3024191856384277} 02/24/2022 06:26:38 - INFO - codeparrot_training - Step 3765: {'lr': 0.0004983337748288391, 'samples': 1928192, 'steps': 3765, 'loss/train': 3.489478349685669} 02/24/2022 06:26:41 - INFO - codeparrot_training - Step 3766: {'lr': 0.0004983318883232058, 'samples': 1928704, 'steps': 3766, 'loss/train': 3.4569692611694336} 02/24/2022 06:26:47 - INFO - codeparrot_training - Step 3767: {'lr': 0.0004983300007537974, 'samples': 1929216, 'steps': 3767, 'loss/train': 3.213865041732788} 02/24/2022 06:26:51 - INFO - codeparrot_training - Step 3768: {'lr': 0.0004983281121206222, 'samples': 1929728, 'steps': 3768, 'loss/train': 3.08313250541687} 02/24/2022 06:26:56 - INFO - codeparrot_training - Step 3769: {'lr': 0.0004983262224236882, 'samples': 1930240, 'steps': 3769, 'loss/train': 2.926205635070801} 02/24/2022 06:27:00 - INFO - codeparrot_training - Step 3770: {'lr': 0.0004983243316630035, 'samples': 1930752, 'steps': 3770, 'loss/train': 3.129277229309082} 02/24/2022 06:27:05 - INFO - codeparrot_training - Step 3771: {'lr': 0.0004983224398385762, 'samples': 1931264, 'steps': 3771, 'loss/train': 2.537529230117798} 02/24/2022 06:27:09 - INFO - codeparrot_training - Step 3772: {'lr': 0.0004983205469504144, 'samples': 1931776, 'steps': 3772, 'loss/train': 1.9506564140319824} 02/24/2022 06:27:14 - INFO - codeparrot_training - Step 3773: {'lr': 0.0004983186529985263, 'samples': 1932288, 'steps': 3773, 'loss/train': 2.6856629848480225} 02/24/2022 06:27:18 - INFO - codeparrot_training - Step 3774: {'lr': 0.00049831675798292, 'samples': 1932800, 'steps': 3774, 'loss/train': 3.698329448699951} 02/24/2022 06:27:23 - INFO - codeparrot_training - Step 3775: {'lr': 0.0004983148619036034, 'samples': 1933312, 'steps': 3775, 'loss/train': 1.9796056747436523} 02/24/2022 06:27:27 - INFO - codeparrot_training - Step 3776: {'lr': 0.0004983129647605849, 'samples': 1933824, 'steps': 3776, 'loss/train': 3.0843496322631836} 02/24/2022 06:27:33 - INFO - codeparrot_training - Step 3777: {'lr': 0.0004983110665538724, 'samples': 1934336, 'steps': 3777, 'loss/train': 2.5527446269989014} 02/24/2022 06:27:36 - INFO - codeparrot_training - Step 3778: {'lr': 0.0004983091672834742, 'samples': 1934848, 'steps': 3778, 'loss/train': 3.087131977081299} 02/24/2022 06:27:42 - INFO - codeparrot_training - Step 3779: {'lr': 0.0004983072669493985, 'samples': 1935360, 'steps': 3779, 'loss/train': 3.097358226776123} 02/24/2022 06:27:45 - INFO - codeparrot_training - Step 3780: {'lr': 0.0004983053655516531, 'samples': 1935872, 'steps': 3780, 'loss/train': 2.8714654445648193} 02/24/2022 06:27:51 - INFO - codeparrot_training - Step 3781: {'lr': 0.0004983034630902465, 'samples': 1936384, 'steps': 3781, 'loss/train': 4.090614318847656} 02/24/2022 06:27:55 - INFO - codeparrot_training - Step 3782: {'lr': 0.0004983015595651867, 'samples': 1936896, 'steps': 3782, 'loss/train': 2.7963130474090576} 02/24/2022 06:28:00 - INFO - codeparrot_training - Step 3783: {'lr': 0.0004982996549764817, 'samples': 1937408, 'steps': 3783, 'loss/train': 2.333299398422241} 02/24/2022 06:28:03 - INFO - codeparrot_training - Step 3784: {'lr': 0.0004982977493241399, 'samples': 1937920, 'steps': 3784, 'loss/train': 2.328354835510254} 02/24/2022 06:28:09 - INFO - codeparrot_training - Step 3785: {'lr': 0.0004982958426081695, 'samples': 1938432, 'steps': 3785, 'loss/train': 2.2759294509887695} 02/24/2022 06:28:12 - INFO - codeparrot_training - Step 3786: {'lr': 0.0004982939348285784, 'samples': 1938944, 'steps': 3786, 'loss/train': 3.6218342781066895} 02/24/2022 06:28:19 - INFO - codeparrot_training - Step 3787: {'lr': 0.000498292025985375, 'samples': 1939456, 'steps': 3787, 'loss/train': 2.3638393878936768} 02/24/2022 06:28:22 - INFO - codeparrot_training - Step 3788: {'lr': 0.0004982901160785675, 'samples': 1939968, 'steps': 3788, 'loss/train': 3.03568959236145} 02/24/2022 06:28:28 - INFO - codeparrot_training - Step 3789: {'lr': 0.0004982882051081639, 'samples': 1940480, 'steps': 3789, 'loss/train': 4.6451802253723145} 02/24/2022 06:28:31 - INFO - codeparrot_training - Step 3790: {'lr': 0.0004982862930741725, 'samples': 1940992, 'steps': 3790, 'loss/train': 3.393296957015991} 02/24/2022 06:28:37 - INFO - codeparrot_training - Step 3791: {'lr': 0.0004982843799766014, 'samples': 1941504, 'steps': 3791, 'loss/train': 4.070493698120117} 02/24/2022 06:28:40 - INFO - codeparrot_training - Step 3792: {'lr': 0.0004982824658154589, 'samples': 1942016, 'steps': 3792, 'loss/train': 0.31920528411865234} 02/24/2022 06:28:46 - INFO - codeparrot_training - Step 3793: {'lr': 0.000498280550590753, 'samples': 1942528, 'steps': 3793, 'loss/train': 3.0541083812713623} 02/24/2022 06:28:49 - INFO - codeparrot_training - Step 3794: {'lr': 0.0004982786343024923, 'samples': 1943040, 'steps': 3794, 'loss/train': 1.9726697206497192} 02/24/2022 06:28:55 - INFO - codeparrot_training - Step 3795: {'lr': 0.0004982767169506847, 'samples': 1943552, 'steps': 3795, 'loss/train': 3.0337398052215576} 02/24/2022 06:28:58 - INFO - codeparrot_training - Step 3796: {'lr': 0.0004982747985353384, 'samples': 1944064, 'steps': 3796, 'loss/train': 2.9903831481933594} 02/24/2022 06:29:05 - INFO - codeparrot_training - Step 3797: {'lr': 0.0004982728790564616, 'samples': 1944576, 'steps': 3797, 'loss/train': 3.065185546875} 02/24/2022 06:29:08 - INFO - codeparrot_training - Step 3798: {'lr': 0.0004982709585140629, 'samples': 1945088, 'steps': 3798, 'loss/train': 3.254495620727539} 02/24/2022 06:29:14 - INFO - codeparrot_training - Step 3799: {'lr': 0.0004982690369081501, 'samples': 1945600, 'steps': 3799, 'loss/train': 2.4238195419311523} 02/24/2022 06:29:17 - INFO - codeparrot_training - Step 3800: {'lr': 0.0004982671142387316, 'samples': 1946112, 'steps': 3800, 'loss/train': 3.1700680255889893} 02/24/2022 06:29:23 - INFO - codeparrot_training - Step 3801: {'lr': 0.0004982651905058156, 'samples': 1946624, 'steps': 3801, 'loss/train': 2.4958302974700928} 02/24/2022 06:29:26 - INFO - codeparrot_training - Step 3802: {'lr': 0.0004982632657094104, 'samples': 1947136, 'steps': 3802, 'loss/train': 2.6902425289154053} 02/24/2022 06:29:32 - INFO - codeparrot_training - Step 3803: {'lr': 0.0004982613398495241, 'samples': 1947648, 'steps': 3803, 'loss/train': 3.3472297191619873} 02/24/2022 06:29:35 - INFO - codeparrot_training - Step 3804: {'lr': 0.0004982594129261652, 'samples': 1948160, 'steps': 3804, 'loss/train': 2.234116315841675} 02/24/2022 06:29:41 - INFO - codeparrot_training - Step 3805: {'lr': 0.0004982574849393416, 'samples': 1948672, 'steps': 3805, 'loss/train': 3.7738876342773438} 02/24/2022 06:29:44 - INFO - codeparrot_training - Step 3806: {'lr': 0.000498255555889062, 'samples': 1949184, 'steps': 3806, 'loss/train': 2.6245174407958984} 02/24/2022 06:29:51 - INFO - codeparrot_training - Step 3807: {'lr': 0.0004982536257753343, 'samples': 1949696, 'steps': 3807, 'loss/train': 2.472965955734253} 02/24/2022 06:29:56 - INFO - codeparrot_training - Step 3808: {'lr': 0.0004982516945981669, 'samples': 1950208, 'steps': 3808, 'loss/train': 2.5350589752197266} 02/24/2022 06:29:59 - INFO - codeparrot_training - Step 3809: {'lr': 0.0004982497623575681, 'samples': 1950720, 'steps': 3809, 'loss/train': 3.7140450477600098} 02/24/2022 06:30:05 - INFO - codeparrot_training - Step 3810: {'lr': 0.0004982478290535461, 'samples': 1951232, 'steps': 3810, 'loss/train': 2.433718681335449} 02/24/2022 06:30:09 - INFO - codeparrot_training - Step 3811: {'lr': 0.0004982458946861093, 'samples': 1951744, 'steps': 3811, 'loss/train': 2.9738035202026367} 02/24/2022 06:30:14 - INFO - codeparrot_training - Step 3812: {'lr': 0.0004982439592552658, 'samples': 1952256, 'steps': 3812, 'loss/train': 2.861863851547241} 02/24/2022 06:30:18 - INFO - codeparrot_training - Step 3813: {'lr': 0.0004982420227610242, 'samples': 1952768, 'steps': 3813, 'loss/train': 2.92391300201416} 02/24/2022 06:30:23 - INFO - codeparrot_training - Step 3814: {'lr': 0.0004982400852033924, 'samples': 1953280, 'steps': 3814, 'loss/train': 3.3727493286132812} 02/24/2022 06:30:27 - INFO - codeparrot_training - Step 3815: {'lr': 0.000498238146582379, 'samples': 1953792, 'steps': 3815, 'loss/train': 2.72053861618042} 02/24/2022 06:30:32 - INFO - codeparrot_training - Step 3816: {'lr': 0.0004982362068979921, 'samples': 1954304, 'steps': 3816, 'loss/train': 3.413799524307251} 02/24/2022 06:30:36 - INFO - codeparrot_training - Step 3817: {'lr': 0.0004982342661502403, 'samples': 1954816, 'steps': 3817, 'loss/train': 2.691828966140747} 02/24/2022 06:30:41 - INFO - codeparrot_training - Step 3818: {'lr': 0.0004982323243391315, 'samples': 1955328, 'steps': 3818, 'loss/train': 3.3134384155273438} 02/24/2022 06:30:45 - INFO - codeparrot_training - Step 3819: {'lr': 0.0004982303814646745, 'samples': 1955840, 'steps': 3819, 'loss/train': 2.3300282955169678} 02/24/2022 06:30:50 - INFO - codeparrot_training - Step 3820: {'lr': 0.0004982284375268772, 'samples': 1956352, 'steps': 3820, 'loss/train': 2.090731143951416} 02/24/2022 06:30:54 - INFO - codeparrot_training - Step 3821: {'lr': 0.0004982264925257481, 'samples': 1956864, 'steps': 3821, 'loss/train': 3.1457366943359375} 02/24/2022 06:31:00 - INFO - codeparrot_training - Step 3822: {'lr': 0.0004982245464612955, 'samples': 1957376, 'steps': 3822, 'loss/train': 2.9802441596984863} 02/24/2022 06:31:03 - INFO - codeparrot_training - Step 3823: {'lr': 0.0004982225993335279, 'samples': 1957888, 'steps': 3823, 'loss/train': 3.245405435562134} 02/24/2022 06:31:09 - INFO - codeparrot_training - Step 3824: {'lr': 0.0004982206511424534, 'samples': 1958400, 'steps': 3824, 'loss/train': 3.7082931995391846} 02/24/2022 06:31:12 - INFO - codeparrot_training - Step 3825: {'lr': 0.0004982187018880805, 'samples': 1958912, 'steps': 3825, 'loss/train': 3.3702168464660645} 02/24/2022 06:31:18 - INFO - codeparrot_training - Step 3826: {'lr': 0.0004982167515704174, 'samples': 1959424, 'steps': 3826, 'loss/train': 2.24308705329895} 02/24/2022 06:31:21 - INFO - codeparrot_training - Step 3827: {'lr': 0.0004982148001894727, 'samples': 1959936, 'steps': 3827, 'loss/train': 3.251370668411255} 02/24/2022 06:31:27 - INFO - codeparrot_training - Step 3828: {'lr': 0.0004982128477452546, 'samples': 1960448, 'steps': 3828, 'loss/train': 2.9376633167266846} 02/24/2022 06:31:30 - INFO - codeparrot_training - Step 3829: {'lr': 0.0004982108942377713, 'samples': 1960960, 'steps': 3829, 'loss/train': 2.1045656204223633} 02/24/2022 06:31:36 - INFO - codeparrot_training - Step 3830: {'lr': 0.0004982089396670316, 'samples': 1961472, 'steps': 3830, 'loss/train': 5.327694892883301} 02/24/2022 06:31:39 - INFO - codeparrot_training - Step 3831: {'lr': 0.0004982069840330435, 'samples': 1961984, 'steps': 3831, 'loss/train': 3.1833877563476562} 02/24/2022 06:31:46 - INFO - codeparrot_training - Step 3832: {'lr': 0.0004982050273358154, 'samples': 1962496, 'steps': 3832, 'loss/train': 3.3765788078308105} 02/24/2022 06:31:49 - INFO - codeparrot_training - Step 3833: {'lr': 0.0004982030695753558, 'samples': 1963008, 'steps': 3833, 'loss/train': 3.984766960144043} 02/24/2022 06:31:52 - INFO - codeparrot_training - Step 3834: {'lr': 0.0004982011107516732, 'samples': 1963520, 'steps': 3834, 'loss/train': 3.2367210388183594} 02/24/2022 06:31:58 - INFO - codeparrot_training - Step 3835: {'lr': 0.0004981991508647757, 'samples': 1964032, 'steps': 3835, 'loss/train': 2.9056060314178467} 02/24/2022 06:32:01 - INFO - codeparrot_training - Step 3836: {'lr': 0.0004981971899146719, 'samples': 1964544, 'steps': 3836, 'loss/train': 2.7002134323120117} 02/24/2022 06:32:07 - INFO - codeparrot_training - Step 3837: {'lr': 0.0004981952279013702, 'samples': 1965056, 'steps': 3837, 'loss/train': 1.8907198905944824} 02/24/2022 06:32:13 - INFO - codeparrot_training - Step 3838: {'lr': 0.0004981932648248789, 'samples': 1965568, 'steps': 3838, 'loss/train': 3.1737170219421387} 02/24/2022 06:32:16 - INFO - codeparrot_training - Step 3839: {'lr': 0.0004981913006852065, 'samples': 1966080, 'steps': 3839, 'loss/train': 2.749114751815796} 02/24/2022 06:32:21 - INFO - codeparrot_training - Step 3840: {'lr': 0.0004981893354823614, 'samples': 1966592, 'steps': 3840, 'loss/train': 2.7215335369110107} 02/24/2022 06:32:25 - INFO - codeparrot_training - Step 3841: {'lr': 0.000498187369216352, 'samples': 1967104, 'steps': 3841, 'loss/train': 2.360414505004883} 02/24/2022 06:32:32 - INFO - codeparrot_training - Step 3842: {'lr': 0.0004981854018871867, 'samples': 1967616, 'steps': 3842, 'loss/train': 1.6500130891799927} 02/24/2022 06:32:35 - INFO - codeparrot_training - Step 3843: {'lr': 0.0004981834334948738, 'samples': 1968128, 'steps': 3843, 'loss/train': 4.215713977813721} 02/24/2022 06:32:41 - INFO - codeparrot_training - Step 3844: {'lr': 0.0004981814640394221, 'samples': 1968640, 'steps': 3844, 'loss/train': 3.047409772872925} 02/24/2022 06:32:44 - INFO - codeparrot_training - Step 3845: {'lr': 0.0004981794935208397, 'samples': 1969152, 'steps': 3845, 'loss/train': 2.663679838180542} 02/24/2022 06:32:50 - INFO - codeparrot_training - Step 3846: {'lr': 0.0004981775219391352, 'samples': 1969664, 'steps': 3846, 'loss/train': 2.523682117462158} 02/24/2022 06:32:53 - INFO - codeparrot_training - Step 3847: {'lr': 0.000498175549294317, 'samples': 1970176, 'steps': 3847, 'loss/train': 3.081430196762085} 02/24/2022 06:32:59 - INFO - codeparrot_training - Step 3848: {'lr': 0.0004981735755863934, 'samples': 1970688, 'steps': 3848, 'loss/train': 5.446176528930664} 02/24/2022 06:33:02 - INFO - codeparrot_training - Step 3849: {'lr': 0.0004981716008153732, 'samples': 1971200, 'steps': 3849, 'loss/train': 2.8046562671661377} 02/24/2022 06:33:08 - INFO - codeparrot_training - Step 3850: {'lr': 0.0004981696249812646, 'samples': 1971712, 'steps': 3850, 'loss/train': 2.7847628593444824} 02/24/2022 06:33:11 - INFO - codeparrot_training - Step 3851: {'lr': 0.0004981676480840761, 'samples': 1972224, 'steps': 3851, 'loss/train': 2.9150774478912354} 02/24/2022 06:33:17 - INFO - codeparrot_training - Step 3852: {'lr': 0.0004981656701238162, 'samples': 1972736, 'steps': 3852, 'loss/train': 1.747209072113037} 02/24/2022 06:33:20 - INFO - codeparrot_training - Step 3853: {'lr': 0.0004981636911004934, 'samples': 1973248, 'steps': 3853, 'loss/train': 2.875384569168091} 02/24/2022 06:33:25 - INFO - codeparrot_training - Step 3854: {'lr': 0.0004981617110141162, 'samples': 1973760, 'steps': 3854, 'loss/train': 2.2483558654785156} 02/24/2022 06:33:29 - INFO - codeparrot_training - Step 3855: {'lr': 0.000498159729864693, 'samples': 1974272, 'steps': 3855, 'loss/train': 2.3501102924346924} 02/24/2022 06:33:34 - INFO - codeparrot_training - Step 3856: {'lr': 0.0004981577476522323, 'samples': 1974784, 'steps': 3856, 'loss/train': 2.564826488494873} 02/24/2022 06:33:38 - INFO - codeparrot_training - Step 3857: {'lr': 0.0004981557643767426, 'samples': 1975296, 'steps': 3857, 'loss/train': 2.1798336505889893} 02/24/2022 06:33:45 - INFO - codeparrot_training - Step 3858: {'lr': 0.0004981537800382323, 'samples': 1975808, 'steps': 3858, 'loss/train': 2.6843316555023193} 02/24/2022 06:33:48 - INFO - codeparrot_training - Step 3859: {'lr': 0.0004981517946367102, 'samples': 1976320, 'steps': 3859, 'loss/train': 2.897378921508789} 02/24/2022 06:33:54 - INFO - codeparrot_training - Step 3860: {'lr': 0.0004981498081721845, 'samples': 1976832, 'steps': 3860, 'loss/train': 2.6063075065612793} 02/24/2022 06:33:57 - INFO - codeparrot_training - Step 3861: {'lr': 0.0004981478206446638, 'samples': 1977344, 'steps': 3861, 'loss/train': 4.317244529724121} 02/24/2022 06:34:03 - INFO - codeparrot_training - Step 3862: {'lr': 0.0004981458320541567, 'samples': 1977856, 'steps': 3862, 'loss/train': 0.46900343894958496} 02/24/2022 06:34:06 - INFO - codeparrot_training - Step 3863: {'lr': 0.0004981438424006716, 'samples': 1978368, 'steps': 3863, 'loss/train': 2.3372273445129395} 02/24/2022 06:34:12 - INFO - codeparrot_training - Step 3864: {'lr': 0.0004981418516842171, 'samples': 1978880, 'steps': 3864, 'loss/train': 3.0440077781677246} 02/24/2022 06:34:15 - INFO - codeparrot_training - Step 3865: {'lr': 0.0004981398599048018, 'samples': 1979392, 'steps': 3865, 'loss/train': 2.0230464935302734} 02/24/2022 06:34:20 - INFO - codeparrot_training - Step 3866: {'lr': 0.000498137867062434, 'samples': 1979904, 'steps': 3866, 'loss/train': 3.9015941619873047} 02/24/2022 06:34:24 - INFO - codeparrot_training - Step 3867: {'lr': 0.0004981358731571223, 'samples': 1980416, 'steps': 3867, 'loss/train': 2.2192654609680176} 02/24/2022 06:34:31 - INFO - codeparrot_training - Step 3868: {'lr': 0.0004981338781888755, 'samples': 1980928, 'steps': 3868, 'loss/train': 3.102003812789917} 02/24/2022 06:34:34 - INFO - codeparrot_training - Step 3869: {'lr': 0.0004981318821577018, 'samples': 1981440, 'steps': 3869, 'loss/train': 3.187795877456665} 02/24/2022 06:34:40 - INFO - codeparrot_training - Step 3870: {'lr': 0.00049812988506361, 'samples': 1981952, 'steps': 3870, 'loss/train': 2.7443923950195312} 02/24/2022 06:34:43 - INFO - codeparrot_training - Step 3871: {'lr': 0.0004981278869066085, 'samples': 1982464, 'steps': 3871, 'loss/train': 2.060483932495117} 02/24/2022 06:34:48 - INFO - codeparrot_training - Step 3872: {'lr': 0.000498125887686706, 'samples': 1982976, 'steps': 3872, 'loss/train': 2.1937296390533447} 02/24/2022 06:34:52 - INFO - codeparrot_training - Step 3873: {'lr': 0.0004981238874039109, 'samples': 1983488, 'steps': 3873, 'loss/train': 2.440709352493286} 02/24/2022 06:34:58 - INFO - codeparrot_training - Step 3874: {'lr': 0.0004981218860582319, 'samples': 1984000, 'steps': 3874, 'loss/train': 3.473883867263794} 02/24/2022 06:35:01 - INFO - codeparrot_training - Step 3875: {'lr': 0.0004981198836496775, 'samples': 1984512, 'steps': 3875, 'loss/train': 3.1413447856903076} 02/24/2022 06:35:06 - INFO - codeparrot_training - Step 3876: {'lr': 0.0004981178801782563, 'samples': 1985024, 'steps': 3876, 'loss/train': 1.9248684644699097} 02/24/2022 06:35:10 - INFO - codeparrot_training - Step 3877: {'lr': 0.000498115875643977, 'samples': 1985536, 'steps': 3877, 'loss/train': 1.1404633522033691} 02/24/2022 06:35:16 - INFO - codeparrot_training - Step 3878: {'lr': 0.0004981138700468479, 'samples': 1986048, 'steps': 3878, 'loss/train': 4.5774126052856445} 02/24/2022 06:35:22 - INFO - codeparrot_training - Step 3879: {'lr': 0.0004981118633868779, 'samples': 1986560, 'steps': 3879, 'loss/train': 2.8465540409088135} 02/24/2022 06:35:25 - INFO - codeparrot_training - Step 3880: {'lr': 0.0004981098556640755, 'samples': 1987072, 'steps': 3880, 'loss/train': 1.1227093935012817} 02/24/2022 06:35:31 - INFO - codeparrot_training - Step 3881: {'lr': 0.0004981078468784491, 'samples': 1987584, 'steps': 3881, 'loss/train': 2.5524864196777344} 02/24/2022 06:35:34 - INFO - codeparrot_training - Step 3882: {'lr': 0.0004981058370300076, 'samples': 1988096, 'steps': 3882, 'loss/train': 2.31843638420105} 02/24/2022 06:35:40 - INFO - codeparrot_training - Step 3883: {'lr': 0.0004981038261187594, 'samples': 1988608, 'steps': 3883, 'loss/train': 1.8953057527542114} 02/24/2022 06:35:43 - INFO - codeparrot_training - Step 3884: {'lr': 0.0004981018141447133, 'samples': 1989120, 'steps': 3884, 'loss/train': 1.1048979759216309} 02/24/2022 06:35:49 - INFO - codeparrot_training - Step 3885: {'lr': 0.0004980998011078776, 'samples': 1989632, 'steps': 3885, 'loss/train': 4.669902801513672} 02/24/2022 06:35:52 - INFO - codeparrot_training - Step 3886: {'lr': 0.0004980977870082613, 'samples': 1990144, 'steps': 3886, 'loss/train': 1.7005499601364136} 02/24/2022 06:35:58 - INFO - codeparrot_training - Step 3887: {'lr': 0.0004980957718458729, 'samples': 1990656, 'steps': 3887, 'loss/train': 3.654224395751953} 02/24/2022 06:36:01 - INFO - codeparrot_training - Step 3888: {'lr': 0.0004980937556207207, 'samples': 1991168, 'steps': 3888, 'loss/train': 2.9529154300689697} 02/24/2022 06:36:07 - INFO - codeparrot_training - Step 3889: {'lr': 0.0004980917383328139, 'samples': 1991680, 'steps': 3889, 'loss/train': 2.5530948638916016} 02/24/2022 06:36:11 - INFO - codeparrot_training - Step 3890: {'lr': 0.0004980897199821609, 'samples': 1992192, 'steps': 3890, 'loss/train': 3.3051698207855225} 02/24/2022 06:36:16 - INFO - codeparrot_training - Step 3891: {'lr': 0.0004980877005687701, 'samples': 1992704, 'steps': 3891, 'loss/train': 2.874004602432251} 02/24/2022 06:36:20 - INFO - codeparrot_training - Step 3892: {'lr': 0.0004980856800926506, 'samples': 1993216, 'steps': 3892, 'loss/train': 3.540424346923828} 02/24/2022 06:36:25 - INFO - codeparrot_training - Step 3893: {'lr': 0.0004980836585538107, 'samples': 1993728, 'steps': 3893, 'loss/train': 3.150780200958252} 02/24/2022 06:36:29 - INFO - codeparrot_training - Step 3894: {'lr': 0.0004980816359522592, 'samples': 1994240, 'steps': 3894, 'loss/train': 2.766648292541504} 02/24/2022 06:36:34 - INFO - codeparrot_training - Step 3895: {'lr': 0.0004980796122880048, 'samples': 1994752, 'steps': 3895, 'loss/train': 2.0121874809265137} 02/24/2022 06:36:38 - INFO - codeparrot_training - Step 3896: {'lr': 0.000498077587561056, 'samples': 1995264, 'steps': 3896, 'loss/train': 2.6432979106903076} 02/24/2022 06:36:43 - INFO - codeparrot_training - Step 3897: {'lr': 0.0004980755617714216, 'samples': 1995776, 'steps': 3897, 'loss/train': 3.193638563156128} 02/24/2022 06:36:47 - INFO - codeparrot_training - Step 3898: {'lr': 0.0004980735349191104, 'samples': 1996288, 'steps': 3898, 'loss/train': 2.426842212677002} 02/24/2022 06:36:52 - INFO - codeparrot_training - Step 3899: {'lr': 0.0004980715070041308, 'samples': 1996800, 'steps': 3899, 'loss/train': 2.622724771499634} 02/24/2022 06:36:55 - INFO - codeparrot_training - Step 3900: {'lr': 0.0004980694780264917, 'samples': 1997312, 'steps': 3900, 'loss/train': 4.063630104064941} 02/24/2022 06:37:01 - INFO - codeparrot_training - Step 3901: {'lr': 0.0004980674479862018, 'samples': 1997824, 'steps': 3901, 'loss/train': 3.0027968883514404} 02/24/2022 06:37:05 - INFO - codeparrot_training - Step 3902: {'lr': 0.0004980654168832697, 'samples': 1998336, 'steps': 3902, 'loss/train': 3.2706844806671143} 02/24/2022 06:37:12 - INFO - codeparrot_training - Step 3903: {'lr': 0.0004980633847177041, 'samples': 1998848, 'steps': 3903, 'loss/train': 2.9794161319732666} 02/24/2022 06:37:15 - INFO - codeparrot_training - Step 3904: {'lr': 0.0004980613514895135, 'samples': 1999360, 'steps': 3904, 'loss/train': 1.8317804336547852} 02/24/2022 06:37:19 - INFO - codeparrot_training - Step 3905: {'lr': 0.0004980593171987072, 'samples': 1999872, 'steps': 3905, 'loss/train': 2.109004020690918} 02/24/2022 06:37:24 - INFO - codeparrot_training - Step 3906: {'lr': 0.0004980572818452934, 'samples': 2000384, 'steps': 3906, 'loss/train': 1.9249424934387207} 02/24/2022 06:37:28 - INFO - codeparrot_training - Step 3907: {'lr': 0.0004980552454292809, 'samples': 2000896, 'steps': 3907, 'loss/train': 3.3942930698394775} 02/24/2022 06:37:33 - INFO - codeparrot_training - Step 3908: {'lr': 0.0004980532079506786, 'samples': 2001408, 'steps': 3908, 'loss/train': 1.7629261016845703} 02/24/2022 06:37:36 - INFO - codeparrot_training - Step 3909: {'lr': 0.0004980511694094951, 'samples': 2001920, 'steps': 3909, 'loss/train': 1.864051342010498} 02/24/2022 06:37:42 - INFO - codeparrot_training - Step 3910: {'lr': 0.0004980491298057392, 'samples': 2002432, 'steps': 3910, 'loss/train': 2.3030428886413574} 02/24/2022 06:37:46 - INFO - codeparrot_training - Step 3911: {'lr': 0.0004980470891394194, 'samples': 2002944, 'steps': 3911, 'loss/train': 3.462634563446045} 02/24/2022 06:37:51 - INFO - codeparrot_training - Step 3912: {'lr': 0.0004980450474105448, 'samples': 2003456, 'steps': 3912, 'loss/train': 2.4956939220428467} 02/24/2022 06:37:55 - INFO - codeparrot_training - Step 3913: {'lr': 0.000498043004619124, 'samples': 2003968, 'steps': 3913, 'loss/train': 2.379676103591919} 02/24/2022 06:38:01 - INFO - codeparrot_training - Step 3914: {'lr': 0.0004980409607651656, 'samples': 2004480, 'steps': 3914, 'loss/train': 1.6961394548416138} 02/24/2022 06:38:04 - INFO - codeparrot_training - Step 3915: {'lr': 0.0004980389158486786, 'samples': 2004992, 'steps': 3915, 'loss/train': 3.039517879486084} 02/24/2022 06:38:10 - INFO - codeparrot_training - Step 3916: {'lr': 0.0004980368698696716, 'samples': 2005504, 'steps': 3916, 'loss/train': 3.665147304534912} 02/24/2022 06:38:13 - INFO - codeparrot_training - Step 3917: {'lr': 0.0004980348228281534, 'samples': 2006016, 'steps': 3917, 'loss/train': 2.4664700031280518} 02/24/2022 06:38:19 - INFO - codeparrot_training - Step 3918: {'lr': 0.0004980327747241329, 'samples': 2006528, 'steps': 3918, 'loss/train': 4.039323329925537} 02/24/2022 06:38:22 - INFO - codeparrot_training - Step 3919: {'lr': 0.0004980307255576185, 'samples': 2007040, 'steps': 3919, 'loss/train': 3.323735475540161} 02/24/2022 06:38:28 - INFO - codeparrot_training - Step 3920: {'lr': 0.0004980286753286195, 'samples': 2007552, 'steps': 3920, 'loss/train': 3.0495383739471436} 02/24/2022 06:38:31 - INFO - codeparrot_training - Step 3921: {'lr': 0.0004980266240371443, 'samples': 2008064, 'steps': 3921, 'loss/train': 4.615071773529053} 02/24/2022 06:38:37 - INFO - codeparrot_training - Step 3922: {'lr': 0.0004980245716832018, 'samples': 2008576, 'steps': 3922, 'loss/train': 2.739562511444092} 02/24/2022 06:38:40 - INFO - codeparrot_training - Step 3923: {'lr': 0.0004980225182668008, 'samples': 2009088, 'steps': 3923, 'loss/train': 2.093564748764038} 02/24/2022 06:38:46 - INFO - codeparrot_training - Step 3924: {'lr': 0.00049802046378795, 'samples': 2009600, 'steps': 3924, 'loss/train': 3.1308810710906982} 02/24/2022 06:38:49 - INFO - codeparrot_training - Step 3925: {'lr': 0.0004980184082466583, 'samples': 2010112, 'steps': 3925, 'loss/train': 4.367579936981201} 02/24/2022 06:38:56 - INFO - codeparrot_training - Step 3926: {'lr': 0.0004980163516429346, 'samples': 2010624, 'steps': 3926, 'loss/train': 2.273790121078491} 02/24/2022 06:38:59 - INFO - codeparrot_training - Step 3927: {'lr': 0.0004980142939767876, 'samples': 2011136, 'steps': 3927, 'loss/train': 3.024556875228882} 02/24/2022 06:39:05 - INFO - codeparrot_training - Step 3928: {'lr': 0.000498012235248226, 'samples': 2011648, 'steps': 3928, 'loss/train': 2.704887866973877} 02/24/2022 06:39:08 - INFO - codeparrot_training - Step 3929: {'lr': 0.0004980101754572589, 'samples': 2012160, 'steps': 3929, 'loss/train': 1.8712055683135986} 02/24/2022 06:39:13 - INFO - codeparrot_training - Step 3930: {'lr': 0.0004980081146038948, 'samples': 2012672, 'steps': 3930, 'loss/train': 3.405402421951294} 02/24/2022 06:39:17 - INFO - codeparrot_training - Step 3931: {'lr': 0.0004980060526881429, 'samples': 2013184, 'steps': 3931, 'loss/train': 3.6324095726013184} 02/24/2022 06:39:23 - INFO - codeparrot_training - Step 3932: {'lr': 0.0004980039897100115, 'samples': 2013696, 'steps': 3932, 'loss/train': 3.024447202682495} 02/24/2022 06:39:26 - INFO - codeparrot_training - Step 3933: {'lr': 0.0004980019256695101, 'samples': 2014208, 'steps': 3933, 'loss/train': 5.835325241088867} 02/24/2022 06:39:32 - INFO - codeparrot_training - Step 3934: {'lr': 0.000497999860566647, 'samples': 2014720, 'steps': 3934, 'loss/train': 2.5996549129486084} 02/24/2022 06:39:35 - INFO - codeparrot_training - Step 3935: {'lr': 0.0004979977944014313, 'samples': 2015232, 'steps': 3935, 'loss/train': 2.718518018722534} 02/24/2022 06:39:41 - INFO - codeparrot_training - Step 3936: {'lr': 0.0004979957271738718, 'samples': 2015744, 'steps': 3936, 'loss/train': 1.3581740856170654} 02/24/2022 06:39:44 - INFO - codeparrot_training - Step 3937: {'lr': 0.0004979936588839773, 'samples': 2016256, 'steps': 3937, 'loss/train': 2.5434367656707764} 02/24/2022 06:39:50 - INFO - codeparrot_training - Step 3938: {'lr': 0.0004979915895317567, 'samples': 2016768, 'steps': 3938, 'loss/train': 2.3870959281921387} 02/24/2022 06:39:53 - INFO - codeparrot_training - Step 3939: {'lr': 0.000497989519117219, 'samples': 2017280, 'steps': 3939, 'loss/train': 2.086479425430298} 02/24/2022 06:39:59 - INFO - codeparrot_training - Step 3940: {'lr': 0.0004979874476403729, 'samples': 2017792, 'steps': 3940, 'loss/train': 2.3216536045074463} 02/24/2022 06:40:02 - INFO - codeparrot_training - Step 3941: {'lr': 0.0004979853751012273, 'samples': 2018304, 'steps': 3941, 'loss/train': 2.391488790512085} 02/24/2022 06:40:08 - INFO - codeparrot_training - Step 3942: {'lr': 0.0004979833014997911, 'samples': 2018816, 'steps': 3942, 'loss/train': 2.9961180686950684} 02/24/2022 06:40:11 - INFO - codeparrot_training - Step 3943: {'lr': 0.0004979812268360731, 'samples': 2019328, 'steps': 3943, 'loss/train': 2.9539763927459717} 02/24/2022 06:40:17 - INFO - codeparrot_training - Step 3944: {'lr': 0.0004979791511100823, 'samples': 2019840, 'steps': 3944, 'loss/train': 3.0451676845550537} 02/24/2022 06:40:20 - INFO - codeparrot_training - Step 3945: {'lr': 0.0004979770743218276, 'samples': 2020352, 'steps': 3945, 'loss/train': 1.821937918663025} 02/24/2022 06:40:26 - INFO - codeparrot_training - Step 3946: {'lr': 0.0004979749964713179, 'samples': 2020864, 'steps': 3946, 'loss/train': 2.286525249481201} 02/24/2022 06:40:29 - INFO - codeparrot_training - Step 3947: {'lr': 0.000497972917558562, 'samples': 2021376, 'steps': 3947, 'loss/train': 3.20574951171875} 02/24/2022 06:40:35 - INFO - codeparrot_training - Step 3948: {'lr': 0.0004979708375835688, 'samples': 2021888, 'steps': 3948, 'loss/train': 2.0780210494995117} 02/24/2022 06:40:38 - INFO - codeparrot_training - Step 3949: {'lr': 0.0004979687565463475, 'samples': 2022400, 'steps': 3949, 'loss/train': 3.068037748336792} 02/24/2022 06:40:45 - INFO - codeparrot_training - Step 3950: {'lr': 0.0004979666744469065, 'samples': 2022912, 'steps': 3950, 'loss/train': 2.231921911239624} 02/24/2022 06:40:48 - INFO - codeparrot_training - Step 3951: {'lr': 0.0004979645912852552, 'samples': 2023424, 'steps': 3951, 'loss/train': 1.9839613437652588} 02/24/2022 06:40:54 - INFO - codeparrot_training - Step 3952: {'lr': 0.0004979625070614022, 'samples': 2023936, 'steps': 3952, 'loss/train': 3.6623475551605225} 02/24/2022 06:40:57 - INFO - codeparrot_training - Step 3953: {'lr': 0.0004979604217753566, 'samples': 2024448, 'steps': 3953, 'loss/train': 1.5192588567733765} 02/24/2022 06:41:03 - INFO - codeparrot_training - Step 3954: {'lr': 0.0004979583354271273, 'samples': 2024960, 'steps': 3954, 'loss/train': 3.713426113128662} 02/24/2022 06:41:06 - INFO - codeparrot_training - Step 3955: {'lr': 0.0004979562480167232, 'samples': 2025472, 'steps': 3955, 'loss/train': 1.8236944675445557} 02/24/2022 06:41:12 - INFO - codeparrot_training - Step 3956: {'lr': 0.0004979541595441534, 'samples': 2025984, 'steps': 3956, 'loss/train': 1.8752506971359253} 02/24/2022 06:41:17 - INFO - codeparrot_training - Step 3957: {'lr': 0.0004979520700094265, 'samples': 2026496, 'steps': 3957, 'loss/train': 2.3672893047332764} 02/24/2022 06:41:21 - INFO - codeparrot_training - Step 3958: {'lr': 0.0004979499794125518, 'samples': 2027008, 'steps': 3958, 'loss/train': 3.0423362255096436} 02/24/2022 06:41:27 - INFO - codeparrot_training - Step 3959: {'lr': 0.0004979478877535382, 'samples': 2027520, 'steps': 3959, 'loss/train': 0.24859654903411865} 02/24/2022 06:41:31 - INFO - codeparrot_training - Step 3960: {'lr': 0.0004979457950323945, 'samples': 2028032, 'steps': 3960, 'loss/train': 3.2089052200317383} 02/24/2022 06:41:36 - INFO - codeparrot_training - Step 3961: {'lr': 0.0004979437012491297, 'samples': 2028544, 'steps': 3961, 'loss/train': 2.6857504844665527} 02/24/2022 06:41:40 - INFO - codeparrot_training - Step 3962: {'lr': 0.0004979416064037528, 'samples': 2029056, 'steps': 3962, 'loss/train': 2.964184522628784} 02/24/2022 06:41:45 - INFO - codeparrot_training - Step 3963: {'lr': 0.0004979395104962728, 'samples': 2029568, 'steps': 3963, 'loss/train': 2.4108879566192627} 02/24/2022 06:41:49 - INFO - codeparrot_training - Step 3964: {'lr': 0.0004979374135266987, 'samples': 2030080, 'steps': 3964, 'loss/train': 2.143505096435547} 02/24/2022 06:41:54 - INFO - codeparrot_training - Step 3965: {'lr': 0.0004979353154950394, 'samples': 2030592, 'steps': 3965, 'loss/train': 0.5235998630523682} 02/24/2022 06:41:58 - INFO - codeparrot_training - Step 3966: {'lr': 0.0004979332164013041, 'samples': 2031104, 'steps': 3966, 'loss/train': 1.6597570180892944} 02/24/2022 06:42:03 - INFO - codeparrot_training - Step 3967: {'lr': 0.0004979311162455015, 'samples': 2031616, 'steps': 3967, 'loss/train': 3.270280361175537} 02/24/2022 06:42:07 - INFO - codeparrot_training - Step 3968: {'lr': 0.0004979290150276407, 'samples': 2032128, 'steps': 3968, 'loss/train': 4.413817882537842} 02/24/2022 06:42:12 - INFO - codeparrot_training - Step 3969: {'lr': 0.0004979269127477308, 'samples': 2032640, 'steps': 3969, 'loss/train': 3.0075902938842773} 02/24/2022 06:42:16 - INFO - codeparrot_training - Step 3970: {'lr': 0.0004979248094057806, 'samples': 2033152, 'steps': 3970, 'loss/train': 1.8251111507415771} 02/24/2022 06:42:22 - INFO - codeparrot_training - Step 3971: {'lr': 0.0004979227050017994, 'samples': 2033664, 'steps': 3971, 'loss/train': 1.5707528591156006} 02/24/2022 06:42:26 - INFO - codeparrot_training - Step 3972: {'lr': 0.000497920599535796, 'samples': 2034176, 'steps': 3972, 'loss/train': 2.2323122024536133} 02/24/2022 06:42:31 - INFO - codeparrot_training - Step 3973: {'lr': 0.0004979184930077794, 'samples': 2034688, 'steps': 3973, 'loss/train': 1.9746849536895752} 02/24/2022 06:42:35 - INFO - codeparrot_training - Step 3974: {'lr': 0.0004979163854177588, 'samples': 2035200, 'steps': 3974, 'loss/train': 2.7524733543395996} 02/24/2022 06:42:40 - INFO - codeparrot_training - Step 3975: {'lr': 0.0004979142767657432, 'samples': 2035712, 'steps': 3975, 'loss/train': 3.455688714981079} 02/24/2022 06:42:44 - INFO - codeparrot_training - Step 3976: {'lr': 0.0004979121670517413, 'samples': 2036224, 'steps': 3976, 'loss/train': 2.497265100479126} 02/24/2022 06:42:49 - INFO - codeparrot_training - Step 3977: {'lr': 0.0004979100562757626, 'samples': 2036736, 'steps': 3977, 'loss/train': 2.372575521469116} 02/24/2022 06:42:53 - INFO - codeparrot_training - Step 3978: {'lr': 0.0004979079444378159, 'samples': 2037248, 'steps': 3978, 'loss/train': 3.387920618057251} 02/24/2022 06:42:58 - INFO - codeparrot_training - Step 3979: {'lr': 0.0004979058315379103, 'samples': 2037760, 'steps': 3979, 'loss/train': 2.5600380897521973} 02/24/2022 06:43:02 - INFO - codeparrot_training - Step 3980: {'lr': 0.0004979037175760548, 'samples': 2038272, 'steps': 3980, 'loss/train': 1.1275111436843872} 02/24/2022 06:43:08 - INFO - codeparrot_training - Step 3981: {'lr': 0.0004979016025522586, 'samples': 2038784, 'steps': 3981, 'loss/train': 5.364108085632324} 02/24/2022 06:43:11 - INFO - codeparrot_training - Step 3982: {'lr': 0.0004978994864665305, 'samples': 2039296, 'steps': 3982, 'loss/train': 3.3783178329467773} 02/24/2022 06:43:17 - INFO - codeparrot_training - Step 3983: {'lr': 0.0004978973693188797, 'samples': 2039808, 'steps': 3983, 'loss/train': 3.17097806930542} 02/24/2022 06:43:20 - INFO - codeparrot_training - Step 3984: {'lr': 0.0004978952511093155, 'samples': 2040320, 'steps': 3984, 'loss/train': 3.197195529937744} 02/24/2022 06:43:26 - INFO - codeparrot_training - Step 3985: {'lr': 0.0004978931318378465, 'samples': 2040832, 'steps': 3985, 'loss/train': 2.1542203426361084} 02/24/2022 06:43:29 - INFO - codeparrot_training - Step 3986: {'lr': 0.0004978910115044822, 'samples': 2041344, 'steps': 3986, 'loss/train': 1.6286121606826782} 02/24/2022 06:43:35 - INFO - codeparrot_training - Step 3987: {'lr': 0.0004978888901092315, 'samples': 2041856, 'steps': 3987, 'loss/train': 3.493614673614502} 02/24/2022 06:43:38 - INFO - codeparrot_training - Step 3988: {'lr': 0.0004978867676521035, 'samples': 2042368, 'steps': 3988, 'loss/train': 3.4249188899993896} 02/24/2022 06:43:44 - INFO - codeparrot_training - Step 3989: {'lr': 0.0004978846441331073, 'samples': 2042880, 'steps': 3989, 'loss/train': 3.6243088245391846} 02/24/2022 06:43:47 - INFO - codeparrot_training - Step 3990: {'lr': 0.000497882519552252, 'samples': 2043392, 'steps': 3990, 'loss/train': 2.1013896465301514} 02/24/2022 06:43:53 - INFO - codeparrot_training - Step 3991: {'lr': 0.0004978803939095466, 'samples': 2043904, 'steps': 3991, 'loss/train': 2.831549644470215} 02/24/2022 06:43:56 - INFO - codeparrot_training - Step 3992: {'lr': 0.0004978782672050004, 'samples': 2044416, 'steps': 3992, 'loss/train': 3.109680414199829} 02/24/2022 06:44:02 - INFO - codeparrot_training - Step 3993: {'lr': 0.0004978761394386224, 'samples': 2044928, 'steps': 3993, 'loss/train': 1.1819266080856323} 02/24/2022 06:44:05 - INFO - codeparrot_training - Step 3994: {'lr': 0.0004978740106104218, 'samples': 2045440, 'steps': 3994, 'loss/train': 3.6159698963165283} 02/24/2022 06:44:11 - INFO - codeparrot_training - Step 3995: {'lr': 0.0004978718807204076, 'samples': 2045952, 'steps': 3995, 'loss/train': 2.9805691242218018} 02/24/2022 06:44:14 - INFO - codeparrot_training - Step 3996: {'lr': 0.0004978697497685889, 'samples': 2046464, 'steps': 3996, 'loss/train': 3.5271191596984863} 02/24/2022 06:44:20 - INFO - codeparrot_training - Step 3997: {'lr': 0.0004978676177549749, 'samples': 2046976, 'steps': 3997, 'loss/train': 2.2478177547454834} 02/24/2022 06:44:24 - INFO - codeparrot_training - Step 3998: {'lr': 0.0004978654846795748, 'samples': 2047488, 'steps': 3998, 'loss/train': 2.3059802055358887} 02/24/2022 06:44:29 - INFO - codeparrot_training - Step 3999: {'lr': 0.0004978633505423976, 'samples': 2048000, 'steps': 3999, 'loss/train': 2.555680751800537} 02/24/2022 06:44:29 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 06:44:46 - WARNING - huggingface_hub.repository - Several commits (4) will be pushed upstream. 02/24/2022 06:44:46 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 06:45:19 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 292ce9a..f6d96e1 floral-grass-11 -> floral-grass-11 02/24/2022 06:45:25 - INFO - codeparrot_training - Step 4000: {'lr': 0.0004978612153434526, 'samples': 2048512, 'steps': 4000, 'loss/train': 1.766768217086792} 02/24/2022 06:45:29 - INFO - codeparrot_training - Step 4001: {'lr': 0.0004978590790827488, 'samples': 2049024, 'steps': 4001, 'loss/train': 2.1696391105651855} 02/24/2022 06:45:34 - INFO - codeparrot_training - Step 4002: {'lr': 0.0004978569417602955, 'samples': 2049536, 'steps': 4002, 'loss/train': 3.4861037731170654} 02/24/2022 06:45:38 - INFO - codeparrot_training - Step 4003: {'lr': 0.0004978548033761017, 'samples': 2050048, 'steps': 4003, 'loss/train': 2.9585325717926025} 02/24/2022 06:45:43 - INFO - codeparrot_training - Step 4004: {'lr': 0.0004978526639301766, 'samples': 2050560, 'steps': 4004, 'loss/train': 2.1167385578155518} 02/24/2022 06:45:47 - INFO - codeparrot_training - Step 4005: {'lr': 0.0004978505234225294, 'samples': 2051072, 'steps': 4005, 'loss/train': 2.0740511417388916} 02/24/2022 06:45:53 - INFO - codeparrot_training - Step 4006: {'lr': 0.0004978483818531693, 'samples': 2051584, 'steps': 4006, 'loss/train': 2.959043025970459} 02/24/2022 06:45:56 - INFO - codeparrot_training - Step 4007: {'lr': 0.0004978462392221054, 'samples': 2052096, 'steps': 4007, 'loss/train': 2.3407022953033447} 02/24/2022 06:46:02 - INFO - codeparrot_training - Step 4008: {'lr': 0.0004978440955293468, 'samples': 2052608, 'steps': 4008, 'loss/train': 2.577820062637329} 02/24/2022 06:46:06 - INFO - codeparrot_training - Step 4009: {'lr': 0.000497841950774903, 'samples': 2053120, 'steps': 4009, 'loss/train': 2.4661107063293457} 02/24/2022 06:46:11 - INFO - codeparrot_training - Step 4010: {'lr': 0.0004978398049587828, 'samples': 2053632, 'steps': 4010, 'loss/train': 1.9514811038970947} 02/24/2022 06:46:15 - INFO - codeparrot_training - Step 4011: {'lr': 0.0004978376580809957, 'samples': 2054144, 'steps': 4011, 'loss/train': 2.8353562355041504} 02/24/2022 06:46:20 - INFO - codeparrot_training - Step 4012: {'lr': 0.0004978355101415507, 'samples': 2054656, 'steps': 4012, 'loss/train': 3.286431312561035} 02/24/2022 06:46:23 - INFO - codeparrot_training - Step 4013: {'lr': 0.0004978333611404571, 'samples': 2055168, 'steps': 4013, 'loss/train': 1.1409614086151123} 02/24/2022 06:46:29 - INFO - codeparrot_training - Step 4014: {'lr': 0.0004978312110777241, 'samples': 2055680, 'steps': 4014, 'loss/train': 0.5710865259170532} 02/24/2022 06:46:32 - INFO - codeparrot_training - Step 4015: {'lr': 0.0004978290599533609, 'samples': 2056192, 'steps': 4015, 'loss/train': 2.760190010070801} 02/24/2022 06:46:39 - INFO - codeparrot_training - Step 4016: {'lr': 0.0004978269077673766, 'samples': 2056704, 'steps': 4016, 'loss/train': 2.6205389499664307} 02/24/2022 06:46:42 - INFO - codeparrot_training - Step 4017: {'lr': 0.0004978247545197806, 'samples': 2057216, 'steps': 4017, 'loss/train': 1.9146541357040405} 02/24/2022 06:46:48 - INFO - codeparrot_training - Step 4018: {'lr': 0.0004978226002105821, 'samples': 2057728, 'steps': 4018, 'loss/train': 2.279057264328003} 02/24/2022 06:46:51 - INFO - codeparrot_training - Step 4019: {'lr': 0.0004978204448397902, 'samples': 2058240, 'steps': 4019, 'loss/train': 2.3974881172180176} 02/24/2022 06:46:57 - INFO - codeparrot_training - Step 4020: {'lr': 0.0004978182884074142, 'samples': 2058752, 'steps': 4020, 'loss/train': 3.1385862827301025} 02/24/2022 06:47:00 - INFO - codeparrot_training - Step 4021: {'lr': 0.0004978161309134633, 'samples': 2059264, 'steps': 4021, 'loss/train': 2.3737213611602783} 02/24/2022 06:47:06 - INFO - codeparrot_training - Step 4022: {'lr': 0.0004978139723579469, 'samples': 2059776, 'steps': 4022, 'loss/train': 4.163173198699951} 02/24/2022 06:47:09 - INFO - codeparrot_training - Step 4023: {'lr': 0.0004978118127408741, 'samples': 2060288, 'steps': 4023, 'loss/train': 1.2128463983535767} 02/24/2022 06:47:15 - INFO - codeparrot_training - Step 4024: {'lr': 0.0004978096520622541, 'samples': 2060800, 'steps': 4024, 'loss/train': 1.8585034608840942} 02/24/2022 06:47:18 - INFO - codeparrot_training - Step 4025: {'lr': 0.0004978074903220964, 'samples': 2061312, 'steps': 4025, 'loss/train': 2.923572063446045} 02/24/2022 06:47:24 - INFO - codeparrot_training - Step 4026: {'lr': 0.0004978053275204099, 'samples': 2061824, 'steps': 4026, 'loss/train': 1.7987560033798218} 02/24/2022 06:47:28 - INFO - codeparrot_training - Step 4027: {'lr': 0.0004978031636572042, 'samples': 2062336, 'steps': 4027, 'loss/train': 2.458836078643799} 02/24/2022 06:47:33 - INFO - codeparrot_training - Step 4028: {'lr': 0.0004978009987324884, 'samples': 2062848, 'steps': 4028, 'loss/train': 3.3272807598114014} 02/24/2022 06:47:37 - INFO - codeparrot_training - Step 4029: {'lr': 0.0004977988327462718, 'samples': 2063360, 'steps': 4029, 'loss/train': 2.1616814136505127} 02/24/2022 06:47:42 - INFO - codeparrot_training - Step 4030: {'lr': 0.0004977966656985637, 'samples': 2063872, 'steps': 4030, 'loss/train': 2.478296995162964} 02/24/2022 06:47:46 - INFO - codeparrot_training - Step 4031: {'lr': 0.0004977944975893733, 'samples': 2064384, 'steps': 4031, 'loss/train': 1.5593154430389404} 02/24/2022 06:47:51 - INFO - codeparrot_training - Step 4032: {'lr': 0.00049779232841871, 'samples': 2064896, 'steps': 4032, 'loss/train': 2.3311731815338135} 02/24/2022 06:47:55 - INFO - codeparrot_training - Step 4033: {'lr': 0.0004977901581865831, 'samples': 2065408, 'steps': 4033, 'loss/train': 2.844867467880249} 02/24/2022 06:48:00 - INFO - codeparrot_training - Step 4034: {'lr': 0.0004977879868930018, 'samples': 2065920, 'steps': 4034, 'loss/train': 2.878805637359619} 02/24/2022 06:48:04 - INFO - codeparrot_training - Step 4035: {'lr': 0.0004977858145379754, 'samples': 2066432, 'steps': 4035, 'loss/train': 2.0401692390441895} 02/24/2022 06:48:09 - INFO - codeparrot_training - Step 4036: {'lr': 0.0004977836411215133, 'samples': 2066944, 'steps': 4036, 'loss/train': 2.3467581272125244} 02/24/2022 06:48:13 - INFO - codeparrot_training - Step 4037: {'lr': 0.0004977814666436248, 'samples': 2067456, 'steps': 4037, 'loss/train': 1.9562848806381226} 02/24/2022 06:48:18 - INFO - codeparrot_training - Step 4038: {'lr': 0.0004977792911043191, 'samples': 2067968, 'steps': 4038, 'loss/train': 3.3585565090179443} 02/24/2022 06:48:21 - INFO - codeparrot_training - Step 4039: {'lr': 0.0004977771145036056, 'samples': 2068480, 'steps': 4039, 'loss/train': 2.3639299869537354} 02/24/2022 06:48:27 - INFO - codeparrot_training - Step 4040: {'lr': 0.0004977749368414937, 'samples': 2068992, 'steps': 4040, 'loss/train': 2.498365640640259} 02/24/2022 06:48:30 - INFO - codeparrot_training - Step 4041: {'lr': 0.0004977727581179926, 'samples': 2069504, 'steps': 4041, 'loss/train': 4.177948474884033} 02/24/2022 06:48:37 - INFO - codeparrot_training - Step 4042: {'lr': 0.0004977705783331117, 'samples': 2070016, 'steps': 4042, 'loss/train': 3.2485718727111816} 02/24/2022 06:48:40 - INFO - codeparrot_training - Step 4043: {'lr': 0.0004977683974868603, 'samples': 2070528, 'steps': 4043, 'loss/train': 3.354999303817749} 02/24/2022 06:48:45 - INFO - codeparrot_training - Step 4044: {'lr': 0.0004977662155792478, 'samples': 2071040, 'steps': 4044, 'loss/train': 3.3551292419433594} 02/24/2022 06:48:49 - INFO - codeparrot_training - Step 4045: {'lr': 0.0004977640326102834, 'samples': 2071552, 'steps': 4045, 'loss/train': 2.6851553916931152} 02/24/2022 06:48:54 - INFO - codeparrot_training - Step 4046: {'lr': 0.0004977618485799767, 'samples': 2072064, 'steps': 4046, 'loss/train': 3.9697463512420654} 02/24/2022 06:48:58 - INFO - codeparrot_training - Step 4047: {'lr': 0.0004977596634883368, 'samples': 2072576, 'steps': 4047, 'loss/train': 3.1436359882354736} 02/24/2022 06:49:03 - INFO - codeparrot_training - Step 4048: {'lr': 0.0004977574773353732, 'samples': 2073088, 'steps': 4048, 'loss/train': 3.32637357711792} 02/24/2022 06:49:07 - INFO - codeparrot_training - Step 4049: {'lr': 0.0004977552901210952, 'samples': 2073600, 'steps': 4049, 'loss/train': 1.3028780221939087} 02/24/2022 06:49:12 - INFO - codeparrot_training - Step 4050: {'lr': 0.0004977531018455124, 'samples': 2074112, 'steps': 4050, 'loss/train': 2.07358455657959} 02/24/2022 06:49:16 - INFO - codeparrot_training - Step 4051: {'lr': 0.0004977509125086338, 'samples': 2074624, 'steps': 4051, 'loss/train': 3.3087549209594727} 02/24/2022 06:49:22 - INFO - codeparrot_training - Step 4052: {'lr': 0.000497748722110469, 'samples': 2075136, 'steps': 4052, 'loss/train': 3.6010215282440186} 02/24/2022 06:49:26 - INFO - codeparrot_training - Step 4053: {'lr': 0.0004977465306510273, 'samples': 2075648, 'steps': 4053, 'loss/train': 3.058358907699585} 02/24/2022 06:49:31 - INFO - codeparrot_training - Step 4054: {'lr': 0.0004977443381303182, 'samples': 2076160, 'steps': 4054, 'loss/train': 2.0042192935943604} 02/24/2022 06:49:34 - INFO - codeparrot_training - Step 4055: {'lr': 0.000497742144548351, 'samples': 2076672, 'steps': 4055, 'loss/train': 2.343111276626587} 02/24/2022 06:49:40 - INFO - codeparrot_training - Step 4056: {'lr': 0.0004977399499051351, 'samples': 2077184, 'steps': 4056, 'loss/train': 2.196147918701172} 02/24/2022 06:49:43 - INFO - codeparrot_training - Step 4057: {'lr': 0.0004977377542006799, 'samples': 2077696, 'steps': 4057, 'loss/train': 1.5491563081741333} 02/24/2022 06:49:49 - INFO - codeparrot_training - Step 4058: {'lr': 0.0004977355574349949, 'samples': 2078208, 'steps': 4058, 'loss/train': 2.3003621101379395} 02/24/2022 06:49:53 - INFO - codeparrot_training - Step 4059: {'lr': 0.0004977333596080894, 'samples': 2078720, 'steps': 4059, 'loss/train': 1.8863720893859863} 02/24/2022 06:49:58 - INFO - codeparrot_training - Step 4060: {'lr': 0.0004977311607199729, 'samples': 2079232, 'steps': 4060, 'loss/train': 3.0794272422790527} 02/24/2022 06:50:01 - INFO - codeparrot_training - Step 4061: {'lr': 0.0004977289607706547, 'samples': 2079744, 'steps': 4061, 'loss/train': 3.6320080757141113} 02/24/2022 06:50:09 - INFO - codeparrot_training - Step 4062: {'lr': 0.0004977267597601443, 'samples': 2080256, 'steps': 4062, 'loss/train': 2.693032741546631} 02/24/2022 06:50:12 - INFO - codeparrot_training - Step 4063: {'lr': 0.0004977245576884511, 'samples': 2080768, 'steps': 4063, 'loss/train': 1.1405690908432007} 02/24/2022 06:50:17 - INFO - codeparrot_training - Step 4064: {'lr': 0.0004977223545555847, 'samples': 2081280, 'steps': 4064, 'loss/train': 2.3361263275146484} 02/24/2022 06:50:21 - INFO - codeparrot_training - Step 4065: {'lr': 0.0004977201503615543, 'samples': 2081792, 'steps': 4065, 'loss/train': 3.531859874725342} 02/24/2022 06:50:26 - INFO - codeparrot_training - Step 4066: {'lr': 0.0004977179451063694, 'samples': 2082304, 'steps': 4066, 'loss/train': 2.254673480987549} 02/24/2022 06:50:30 - INFO - codeparrot_training - Step 4067: {'lr': 0.0004977157387900395, 'samples': 2082816, 'steps': 4067, 'loss/train': 2.790304660797119} 02/24/2022 06:50:35 - INFO - codeparrot_training - Step 4068: {'lr': 0.0004977135314125741, 'samples': 2083328, 'steps': 4068, 'loss/train': 1.7385773658752441} 02/24/2022 06:50:39 - INFO - codeparrot_training - Step 4069: {'lr': 0.0004977113229739825, 'samples': 2083840, 'steps': 4069, 'loss/train': 2.9903883934020996} 02/24/2022 06:50:45 - INFO - codeparrot_training - Step 4070: {'lr': 0.0004977091134742743, 'samples': 2084352, 'steps': 4070, 'loss/train': 2.2825770378112793} 02/24/2022 06:50:48 - INFO - codeparrot_training - Step 4071: {'lr': 0.0004977069029134588, 'samples': 2084864, 'steps': 4071, 'loss/train': 1.849504828453064} 02/24/2022 06:50:54 - INFO - codeparrot_training - Step 4072: {'lr': 0.0004977046912915458, 'samples': 2085376, 'steps': 4072, 'loss/train': 2.6642861366271973} 02/24/2022 06:50:57 - INFO - codeparrot_training - Step 4073: {'lr': 0.0004977024786085444, 'samples': 2085888, 'steps': 4073, 'loss/train': 3.099130630493164} 02/24/2022 06:51:03 - INFO - codeparrot_training - Step 4074: {'lr': 0.0004977002648644642, 'samples': 2086400, 'steps': 4074, 'loss/train': 2.609877347946167} 02/24/2022 06:51:06 - INFO - codeparrot_training - Step 4075: {'lr': 0.0004976980500593149, 'samples': 2086912, 'steps': 4075, 'loss/train': 2.702996253967285} 02/24/2022 06:51:12 - INFO - codeparrot_training - Step 4076: {'lr': 0.0004976958341931057, 'samples': 2087424, 'steps': 4076, 'loss/train': 3.115689754486084} 02/24/2022 06:51:15 - INFO - codeparrot_training - Step 4077: {'lr': 0.0004976936172658462, 'samples': 2087936, 'steps': 4077, 'loss/train': 0.3984140455722809} 02/24/2022 06:51:21 - INFO - codeparrot_training - Step 4078: {'lr': 0.0004976913992775459, 'samples': 2088448, 'steps': 4078, 'loss/train': 2.721973419189453} 02/24/2022 06:51:25 - INFO - codeparrot_training - Step 4079: {'lr': 0.0004976891802282143, 'samples': 2088960, 'steps': 4079, 'loss/train': 3.176347494125366} 02/24/2022 06:51:30 - INFO - codeparrot_training - Step 4080: {'lr': 0.0004976869601178609, 'samples': 2089472, 'steps': 4080, 'loss/train': 3.010443687438965} 02/24/2022 06:51:34 - INFO - codeparrot_training - Step 4081: {'lr': 0.0004976847389464952, 'samples': 2089984, 'steps': 4081, 'loss/train': 2.6185731887817383} 02/24/2022 06:51:39 - INFO - codeparrot_training - Step 4082: {'lr': 0.0004976825167141268, 'samples': 2090496, 'steps': 4082, 'loss/train': 2.216217279434204} 02/24/2022 06:51:43 - INFO - codeparrot_training - Step 4083: {'lr': 0.000497680293420765, 'samples': 2091008, 'steps': 4083, 'loss/train': 1.74923574924469} 02/24/2022 06:51:48 - INFO - codeparrot_training - Step 4084: {'lr': 0.0004976780690664196, 'samples': 2091520, 'steps': 4084, 'loss/train': 2.7566394805908203} 02/24/2022 06:51:52 - INFO - codeparrot_training - Step 4085: {'lr': 0.0004976758436511, 'samples': 2092032, 'steps': 4085, 'loss/train': 0.7061077356338501} 02/24/2022 06:51:57 - INFO - codeparrot_training - Step 4086: {'lr': 0.0004976736171748156, 'samples': 2092544, 'steps': 4086, 'loss/train': 2.711764335632324} 02/24/2022 06:52:01 - INFO - codeparrot_training - Step 4087: {'lr': 0.0004976713896375762, 'samples': 2093056, 'steps': 4087, 'loss/train': 3.8529140949249268} 02/24/2022 06:52:07 - INFO - codeparrot_training - Step 4088: {'lr': 0.0004976691610393911, 'samples': 2093568, 'steps': 4088, 'loss/train': 4.321038246154785} 02/24/2022 06:52:11 - INFO - codeparrot_training - Step 4089: {'lr': 0.0004976669313802701, 'samples': 2094080, 'steps': 4089, 'loss/train': 2.645718812942505} 02/24/2022 06:52:16 - INFO - codeparrot_training - Step 4090: {'lr': 0.0004976647006602225, 'samples': 2094592, 'steps': 4090, 'loss/train': 3.2641632556915283} 02/24/2022 06:52:20 - INFO - codeparrot_training - Step 4091: {'lr': 0.0004976624688792581, 'samples': 2095104, 'steps': 4091, 'loss/train': 3.3384528160095215} 02/24/2022 06:52:25 - INFO - codeparrot_training - Step 4092: {'lr': 0.0004976602360373861, 'samples': 2095616, 'steps': 4092, 'loss/train': 3.3506109714508057} 02/24/2022 06:52:29 - INFO - codeparrot_training - Step 4093: {'lr': 0.0004976580021346164, 'samples': 2096128, 'steps': 4093, 'loss/train': 2.810047149658203} 02/24/2022 06:52:34 - INFO - codeparrot_training - Step 4094: {'lr': 0.0004976557671709585, 'samples': 2096640, 'steps': 4094, 'loss/train': 0.20184071362018585} 02/24/2022 06:52:40 - INFO - codeparrot_training - Step 4095: {'lr': 0.0004976535311464219, 'samples': 2097152, 'steps': 4095, 'loss/train': 3.3932225704193115} 02/24/2022 06:52:43 - INFO - codeparrot_training - Step 4096: {'lr': 0.0004976512940610162, 'samples': 2097664, 'steps': 4096, 'loss/train': 2.8875722885131836} 02/24/2022 06:52:50 - INFO - codeparrot_training - Step 4097: {'lr': 0.0004976490559147511, 'samples': 2098176, 'steps': 4097, 'loss/train': 1.573515772819519} 02/24/2022 06:52:53 - INFO - codeparrot_training - Step 4098: {'lr': 0.0004976468167076359, 'samples': 2098688, 'steps': 4098, 'loss/train': 1.7893885374069214} 02/24/2022 06:52:59 - INFO - codeparrot_training - Step 4099: {'lr': 0.0004976445764396805, 'samples': 2099200, 'steps': 4099, 'loss/train': 3.0921425819396973} 02/24/2022 06:53:02 - INFO - codeparrot_training - Step 4100: {'lr': 0.0004976423351108943, 'samples': 2099712, 'steps': 4100, 'loss/train': 3.3553755283355713} 02/24/2022 06:53:08 - INFO - codeparrot_training - Step 4101: {'lr': 0.0004976400927212871, 'samples': 2100224, 'steps': 4101, 'loss/train': 2.4786128997802734} 02/24/2022 06:53:11 - INFO - codeparrot_training - Step 4102: {'lr': 0.0004976378492708681, 'samples': 2100736, 'steps': 4102, 'loss/train': 3.5474250316619873} 02/24/2022 06:53:17 - INFO - codeparrot_training - Step 4103: {'lr': 0.0004976356047596475, 'samples': 2101248, 'steps': 4103, 'loss/train': 3.0893149375915527} 02/24/2022 06:53:20 - INFO - codeparrot_training - Step 4104: {'lr': 0.0004976333591876344, 'samples': 2101760, 'steps': 4104, 'loss/train': 1.6089959144592285} 02/24/2022 06:53:26 - INFO - codeparrot_training - Step 4105: {'lr': 0.0004976311125548387, 'samples': 2102272, 'steps': 4105, 'loss/train': 1.4295350313186646} 02/24/2022 06:53:29 - INFO - codeparrot_training - Step 4106: {'lr': 0.00049762886486127, 'samples': 2102784, 'steps': 4106, 'loss/train': 2.4978926181793213} 02/24/2022 06:53:35 - INFO - codeparrot_training - Step 4107: {'lr': 0.0004976266161069379, 'samples': 2103296, 'steps': 4107, 'loss/train': 2.4716033935546875} 02/24/2022 06:53:39 - INFO - codeparrot_training - Step 4108: {'lr': 0.0004976243662918518, 'samples': 2103808, 'steps': 4108, 'loss/train': 2.882061719894409} 02/24/2022 06:53:44 - INFO - codeparrot_training - Step 4109: {'lr': 0.0004976221154160217, 'samples': 2104320, 'steps': 4109, 'loss/train': 3.3347325325012207} 02/24/2022 06:53:48 - INFO - codeparrot_training - Step 4110: {'lr': 0.0004976198634794571, 'samples': 2104832, 'steps': 4110, 'loss/train': 2.1972203254699707} 02/24/2022 06:53:53 - INFO - codeparrot_training - Step 4111: {'lr': 0.0004976176104821675, 'samples': 2105344, 'steps': 4111, 'loss/train': 3.1521363258361816} 02/24/2022 06:53:57 - INFO - codeparrot_training - Step 4112: {'lr': 0.0004976153564241628, 'samples': 2105856, 'steps': 4112, 'loss/train': 2.865093231201172} 02/24/2022 06:54:02 - INFO - codeparrot_training - Step 4113: {'lr': 0.0004976131013054526, 'samples': 2106368, 'steps': 4113, 'loss/train': 2.075465202331543} 02/24/2022 06:54:06 - INFO - codeparrot_training - Step 4114: {'lr': 0.0004976108451260464, 'samples': 2106880, 'steps': 4114, 'loss/train': 1.5532543659210205} 02/24/2022 06:54:11 - INFO - codeparrot_training - Step 4115: {'lr': 0.000497608587885954, 'samples': 2107392, 'steps': 4115, 'loss/train': 3.418349027633667} 02/24/2022 06:54:15 - INFO - codeparrot_training - Step 4116: {'lr': 0.0004976063295851849, 'samples': 2107904, 'steps': 4116, 'loss/train': 0.24530726671218872} 02/24/2022 06:54:20 - INFO - codeparrot_training - Step 4117: {'lr': 0.000497604070223749, 'samples': 2108416, 'steps': 4117, 'loss/train': 3.3822782039642334} 02/24/2022 06:54:24 - INFO - codeparrot_training - Step 4118: {'lr': 0.0004976018098016559, 'samples': 2108928, 'steps': 4118, 'loss/train': 3.072525978088379} 02/24/2022 06:54:29 - INFO - codeparrot_training - Step 4119: {'lr': 0.0004975995483189153, 'samples': 2109440, 'steps': 4119, 'loss/train': 3.4508750438690186} 02/24/2022 06:54:33 - INFO - codeparrot_training - Step 4120: {'lr': 0.0004975972857755368, 'samples': 2109952, 'steps': 4120, 'loss/train': 3.0615954399108887} 02/24/2022 06:54:38 - INFO - codeparrot_training - Step 4121: {'lr': 0.0004975950221715302, 'samples': 2110464, 'steps': 4121, 'loss/train': 1.7298758029937744} 02/24/2022 06:54:42 - INFO - codeparrot_training - Step 4122: {'lr': 0.0004975927575069051, 'samples': 2110976, 'steps': 4122, 'loss/train': 2.7990801334381104} 02/24/2022 06:54:47 - INFO - codeparrot_training - Step 4123: {'lr': 0.0004975904917816713, 'samples': 2111488, 'steps': 4123, 'loss/train': 3.235246181488037} 02/24/2022 06:54:51 - INFO - codeparrot_training - Step 4124: {'lr': 0.0004975882249958385, 'samples': 2112000, 'steps': 4124, 'loss/train': 3.029106616973877} 02/24/2022 06:54:57 - INFO - codeparrot_training - Step 4125: {'lr': 0.0004975859571494162, 'samples': 2112512, 'steps': 4125, 'loss/train': 3.2794747352600098} 02/24/2022 06:55:01 - INFO - codeparrot_training - Step 4126: {'lr': 0.0004975836882424143, 'samples': 2113024, 'steps': 4126, 'loss/train': 3.6696321964263916} 02/24/2022 06:55:06 - INFO - codeparrot_training - Step 4127: {'lr': 0.0004975814182748426, 'samples': 2113536, 'steps': 4127, 'loss/train': 2.5619986057281494} 02/24/2022 06:55:10 - INFO - codeparrot_training - Step 4128: {'lr': 0.0004975791472467108, 'samples': 2114048, 'steps': 4128, 'loss/train': 3.685267686843872} 02/24/2022 06:55:15 - INFO - codeparrot_training - Step 4129: {'lr': 0.0004975768751580283, 'samples': 2114560, 'steps': 4129, 'loss/train': 1.5730431079864502} 02/24/2022 06:55:19 - INFO - codeparrot_training - Step 4130: {'lr': 0.0004975746020088052, 'samples': 2115072, 'steps': 4130, 'loss/train': 2.3806216716766357} 02/24/2022 06:55:24 - INFO - codeparrot_training - Step 4131: {'lr': 0.0004975723277990512, 'samples': 2115584, 'steps': 4131, 'loss/train': 2.7863998413085938} 02/24/2022 06:55:28 - INFO - codeparrot_training - Step 4132: {'lr': 0.0004975700525287758, 'samples': 2116096, 'steps': 4132, 'loss/train': 2.579864025115967} 02/24/2022 06:55:33 - INFO - codeparrot_training - Step 4133: {'lr': 0.0004975677761979891, 'samples': 2116608, 'steps': 4133, 'loss/train': 1.7505260705947876} 02/24/2022 06:55:37 - INFO - codeparrot_training - Step 4134: {'lr': 0.0004975654988067005, 'samples': 2117120, 'steps': 4134, 'loss/train': 3.4967141151428223} 02/24/2022 06:55:43 - INFO - codeparrot_training - Step 4135: {'lr': 0.00049756322035492, 'samples': 2117632, 'steps': 4135, 'loss/train': 2.701418161392212} 02/24/2022 06:55:47 - INFO - codeparrot_training - Step 4136: {'lr': 0.0004975609408426572, 'samples': 2118144, 'steps': 4136, 'loss/train': 2.120211601257324} 02/24/2022 06:55:52 - INFO - codeparrot_training - Step 4137: {'lr': 0.000497558660269922, 'samples': 2118656, 'steps': 4137, 'loss/train': 2.9140028953552246} 02/24/2022 06:55:56 - INFO - codeparrot_training - Step 4138: {'lr': 0.0004975563786367241, 'samples': 2119168, 'steps': 4138, 'loss/train': 2.5102291107177734} 02/24/2022 06:56:01 - INFO - codeparrot_training - Step 4139: {'lr': 0.0004975540959430732, 'samples': 2119680, 'steps': 4139, 'loss/train': 2.784776449203491} 02/24/2022 06:56:05 - INFO - codeparrot_training - Step 4140: {'lr': 0.0004975518121889793, 'samples': 2120192, 'steps': 4140, 'loss/train': 3.3934874534606934} 02/24/2022 06:56:10 - INFO - codeparrot_training - Step 4141: {'lr': 0.000497549527374452, 'samples': 2120704, 'steps': 4141, 'loss/train': 3.3145015239715576} 02/24/2022 06:56:14 - INFO - codeparrot_training - Step 4142: {'lr': 0.000497547241499501, 'samples': 2121216, 'steps': 4142, 'loss/train': 3.1953165531158447} 02/24/2022 06:56:19 - INFO - codeparrot_training - Step 4143: {'lr': 0.0004975449545641364, 'samples': 2121728, 'steps': 4143, 'loss/train': 2.2466766834259033} 02/24/2022 06:56:23 - INFO - codeparrot_training - Step 4144: {'lr': 0.0004975426665683678, 'samples': 2122240, 'steps': 4144, 'loss/train': 1.393843173980713} 02/24/2022 06:56:29 - INFO - codeparrot_training - Step 4145: {'lr': 0.000497540377512205, 'samples': 2122752, 'steps': 4145, 'loss/train': 2.6824259757995605} 02/24/2022 06:56:33 - INFO - codeparrot_training - Step 4146: {'lr': 0.0004975380873956577, 'samples': 2123264, 'steps': 4146, 'loss/train': 2.5177626609802246} 02/24/2022 06:56:38 - INFO - codeparrot_training - Step 4147: {'lr': 0.0004975357962187359, 'samples': 2123776, 'steps': 4147, 'loss/train': 2.309971570968628} 02/24/2022 06:56:42 - INFO - codeparrot_training - Step 4148: {'lr': 0.0004975335039814493, 'samples': 2124288, 'steps': 4148, 'loss/train': 0.4787849485874176} 02/24/2022 06:56:47 - INFO - codeparrot_training - Step 4149: {'lr': 0.0004975312106838079, 'samples': 2124800, 'steps': 4149, 'loss/train': 1.504604458808899} 02/24/2022 06:56:51 - INFO - codeparrot_training - Step 4150: {'lr': 0.0004975289163258214, 'samples': 2125312, 'steps': 4150, 'loss/train': 2.834517478942871} 02/24/2022 06:56:56 - INFO - codeparrot_training - Step 4151: {'lr': 0.0004975266209074995, 'samples': 2125824, 'steps': 4151, 'loss/train': 3.719064474105835} 02/24/2022 06:57:00 - INFO - codeparrot_training - Step 4152: {'lr': 0.0004975243244288522, 'samples': 2126336, 'steps': 4152, 'loss/train': 1.7685208320617676} 02/24/2022 06:57:05 - INFO - codeparrot_training - Step 4153: {'lr': 0.0004975220268898893, 'samples': 2126848, 'steps': 4153, 'loss/train': 0.5588983297348022} 02/24/2022 06:57:09 - INFO - codeparrot_training - Step 4154: {'lr': 0.0004975197282906207, 'samples': 2127360, 'steps': 4154, 'loss/train': 2.361112356185913} 02/24/2022 06:57:15 - INFO - codeparrot_training - Step 4155: {'lr': 0.0004975174286310562, 'samples': 2127872, 'steps': 4155, 'loss/train': 2.388824224472046} 02/24/2022 06:57:18 - INFO - codeparrot_training - Step 4156: {'lr': 0.0004975151279112054, 'samples': 2128384, 'steps': 4156, 'loss/train': 1.9848403930664062} 02/24/2022 06:57:24 - INFO - codeparrot_training - Step 4157: {'lr': 0.0004975128261310787, 'samples': 2128896, 'steps': 4157, 'loss/train': 2.1548051834106445} 02/24/2022 06:57:27 - INFO - codeparrot_training - Step 4158: {'lr': 0.0004975105232906854, 'samples': 2129408, 'steps': 4158, 'loss/train': 1.8559770584106445} 02/24/2022 06:57:33 - INFO - codeparrot_training - Step 4159: {'lr': 0.0004975082193900357, 'samples': 2129920, 'steps': 4159, 'loss/train': 3.228878974914551} 02/24/2022 06:57:36 - INFO - codeparrot_training - Step 4160: {'lr': 0.0004975059144291394, 'samples': 2130432, 'steps': 4160, 'loss/train': 2.7836742401123047} 02/24/2022 06:57:42 - INFO - codeparrot_training - Step 4161: {'lr': 0.0004975036084080063, 'samples': 2130944, 'steps': 4161, 'loss/train': 3.261890411376953} 02/24/2022 06:57:45 - INFO - codeparrot_training - Step 4162: {'lr': 0.0004975013013266464, 'samples': 2131456, 'steps': 4162, 'loss/train': 5.519766807556152} 02/24/2022 06:57:51 - INFO - codeparrot_training - Step 4163: {'lr': 0.0004974989931850695, 'samples': 2131968, 'steps': 4163, 'loss/train': 2.5995800495147705} 02/24/2022 06:57:54 - INFO - codeparrot_training - Step 4164: {'lr': 0.0004974966839832855, 'samples': 2132480, 'steps': 4164, 'loss/train': 1.9648898839950562} 02/24/2022 06:58:00 - INFO - codeparrot_training - Step 4165: {'lr': 0.0004974943737213042, 'samples': 2132992, 'steps': 4165, 'loss/train': 3.284193992614746} 02/24/2022 06:58:03 - INFO - codeparrot_training - Step 4166: {'lr': 0.0004974920623991356, 'samples': 2133504, 'steps': 4166, 'loss/train': 3.09946346282959} 02/24/2022 06:58:09 - INFO - codeparrot_training - Step 4167: {'lr': 0.0004974897500167898, 'samples': 2134016, 'steps': 4167, 'loss/train': 1.719670057296753} 02/24/2022 06:58:15 - INFO - codeparrot_training - Step 4168: {'lr': 0.0004974874365742763, 'samples': 2134528, 'steps': 4168, 'loss/train': 3.4264321327209473} 02/24/2022 06:58:19 - INFO - codeparrot_training - Step 4169: {'lr': 0.0004974851220716053, 'samples': 2135040, 'steps': 4169, 'loss/train': 3.7530100345611572} 02/24/2022 06:58:22 - INFO - codeparrot_training - Step 4170: {'lr': 0.0004974828065087867, 'samples': 2135552, 'steps': 4170, 'loss/train': 4.17598295211792} 02/24/2022 06:58:28 - INFO - codeparrot_training - Step 4171: {'lr': 0.0004974804898858302, 'samples': 2136064, 'steps': 4171, 'loss/train': 1.6517932415008545} 02/24/2022 06:58:31 - INFO - codeparrot_training - Step 4172: {'lr': 0.0004974781722027459, 'samples': 2136576, 'steps': 4172, 'loss/train': 2.7677624225616455} 02/24/2022 06:58:36 - INFO - codeparrot_training - Step 4173: {'lr': 0.0004974758534595436, 'samples': 2137088, 'steps': 4173, 'loss/train': 1.6782039403915405} 02/24/2022 06:58:42 - INFO - codeparrot_training - Step 4174: {'lr': 0.0004974735336562335, 'samples': 2137600, 'steps': 4174, 'loss/train': 2.7820169925689697} 02/24/2022 06:58:46 - INFO - codeparrot_training - Step 4175: {'lr': 0.0004974712127928252, 'samples': 2138112, 'steps': 4175, 'loss/train': 1.788802981376648} 02/24/2022 06:58:51 - INFO - codeparrot_training - Step 4176: {'lr': 0.000497468890869329, 'samples': 2138624, 'steps': 4176, 'loss/train': 3.1842432022094727} 02/24/2022 06:58:55 - INFO - codeparrot_training - Step 4177: {'lr': 0.0004974665678857545, 'samples': 2139136, 'steps': 4177, 'loss/train': 2.655348777770996} 02/24/2022 06:59:00 - INFO - codeparrot_training - Step 4178: {'lr': 0.0004974642438421118, 'samples': 2139648, 'steps': 4178, 'loss/train': 3.539243698120117} 02/24/2022 06:59:04 - INFO - codeparrot_training - Step 4179: {'lr': 0.0004974619187384109, 'samples': 2140160, 'steps': 4179, 'loss/train': 3.358081817626953} 02/24/2022 06:59:10 - INFO - codeparrot_training - Step 4180: {'lr': 0.0004974595925746618, 'samples': 2140672, 'steps': 4180, 'loss/train': 1.3600236177444458} 02/24/2022 06:59:13 - INFO - codeparrot_training - Step 4181: {'lr': 0.0004974572653508742, 'samples': 2141184, 'steps': 4181, 'loss/train': 1.4963819980621338} 02/24/2022 06:59:19 - INFO - codeparrot_training - Step 4182: {'lr': 0.0004974549370670584, 'samples': 2141696, 'steps': 4182, 'loss/train': 1.3458919525146484} 02/24/2022 06:59:22 - INFO - codeparrot_training - Step 4183: {'lr': 0.0004974526077232242, 'samples': 2142208, 'steps': 4183, 'loss/train': 3.2300713062286377} 02/24/2022 06:59:28 - INFO - codeparrot_training - Step 4184: {'lr': 0.0004974502773193815, 'samples': 2142720, 'steps': 4184, 'loss/train': 3.0200698375701904} 02/24/2022 06:59:31 - INFO - codeparrot_training - Step 4185: {'lr': 0.0004974479458555405, 'samples': 2143232, 'steps': 4185, 'loss/train': 1.4956889152526855} 02/24/2022 06:59:37 - INFO - codeparrot_training - Step 4186: {'lr': 0.000497445613331711, 'samples': 2143744, 'steps': 4186, 'loss/train': 2.869990825653076} 02/24/2022 06:59:40 - INFO - codeparrot_training - Step 4187: {'lr': 0.0004974432797479032, 'samples': 2144256, 'steps': 4187, 'loss/train': 1.9260053634643555} 02/24/2022 06:59:46 - INFO - codeparrot_training - Step 4188: {'lr': 0.0004974409451041268, 'samples': 2144768, 'steps': 4188, 'loss/train': 2.113548994064331} 02/24/2022 06:59:49 - INFO - codeparrot_training - Step 4189: {'lr': 0.0004974386094003921, 'samples': 2145280, 'steps': 4189, 'loss/train': 3.0446863174438477} 02/24/2022 06:59:56 - INFO - codeparrot_training - Step 4190: {'lr': 0.0004974362726367089, 'samples': 2145792, 'steps': 4190, 'loss/train': 1.829388976097107} 02/24/2022 07:00:00 - INFO - codeparrot_training - Step 4191: {'lr': 0.0004974339348130873, 'samples': 2146304, 'steps': 4191, 'loss/train': 3.40861177444458} 02/24/2022 07:00:03 - INFO - codeparrot_training - Step 4192: {'lr': 0.0004974315959295373, 'samples': 2146816, 'steps': 4192, 'loss/train': 2.2164175510406494} 02/24/2022 07:00:09 - INFO - codeparrot_training - Step 4193: {'lr': 0.0004974292559860688, 'samples': 2147328, 'steps': 4193, 'loss/train': 2.3503730297088623} 02/24/2022 07:00:12 - INFO - codeparrot_training - Step 4194: {'lr': 0.0004974269149826921, 'samples': 2147840, 'steps': 4194, 'loss/train': 3.8397929668426514} 02/24/2022 07:00:18 - INFO - codeparrot_training - Step 4195: {'lr': 0.0004974245729194169, 'samples': 2148352, 'steps': 4195, 'loss/train': 2.7743873596191406} 02/24/2022 07:00:21 - INFO - codeparrot_training - Step 4196: {'lr': 0.0004974222297962535, 'samples': 2148864, 'steps': 4196, 'loss/train': 0.25342845916748047} 02/24/2022 07:00:27 - INFO - codeparrot_training - Step 4197: {'lr': 0.0004974198856132118, 'samples': 2149376, 'steps': 4197, 'loss/train': 0.7477357983589172} 02/24/2022 07:00:30 - INFO - codeparrot_training - Step 4198: {'lr': 0.0004974175403703019, 'samples': 2149888, 'steps': 4198, 'loss/train': 3.1035780906677246} 02/24/2022 07:00:36 - INFO - codeparrot_training - Step 4199: {'lr': 0.0004974151940675338, 'samples': 2150400, 'steps': 4199, 'loss/train': 3.3665931224823} 02/24/2022 07:00:39 - INFO - codeparrot_training - Step 4200: {'lr': 0.0004974128467049176, 'samples': 2150912, 'steps': 4200, 'loss/train': 2.379422664642334} 02/24/2022 07:00:46 - INFO - codeparrot_training - Step 4201: {'lr': 0.0004974104982824632, 'samples': 2151424, 'steps': 4201, 'loss/train': 3.107706069946289} 02/24/2022 07:00:50 - INFO - codeparrot_training - Step 4202: {'lr': 0.0004974081488001809, 'samples': 2151936, 'steps': 4202, 'loss/train': 2.9332213401794434} 02/24/2022 07:00:55 - INFO - codeparrot_training - Step 4203: {'lr': 0.0004974057982580806, 'samples': 2152448, 'steps': 4203, 'loss/train': 3.1667959690093994} 02/24/2022 07:00:59 - INFO - codeparrot_training - Step 4204: {'lr': 0.0004974034466561725, 'samples': 2152960, 'steps': 4204, 'loss/train': 2.3205790519714355} 02/24/2022 07:01:04 - INFO - codeparrot_training - Step 4205: {'lr': 0.0004974010939944667, 'samples': 2153472, 'steps': 4205, 'loss/train': 2.7863240242004395} 02/24/2022 07:01:08 - INFO - codeparrot_training - Step 4206: {'lr': 0.0004973987402729729, 'samples': 2153984, 'steps': 4206, 'loss/train': 3.1548309326171875} 02/24/2022 07:01:13 - INFO - codeparrot_training - Step 4207: {'lr': 0.0004973963854917016, 'samples': 2154496, 'steps': 4207, 'loss/train': 2.5072789192199707} 02/24/2022 07:01:17 - INFO - codeparrot_training - Step 4208: {'lr': 0.0004973940296506627, 'samples': 2155008, 'steps': 4208, 'loss/train': 3.0419318675994873} 02/24/2022 07:01:22 - INFO - codeparrot_training - Step 4209: {'lr': 0.0004973916727498664, 'samples': 2155520, 'steps': 4209, 'loss/train': 2.2985165119171143} 02/24/2022 07:01:26 - INFO - codeparrot_training - Step 4210: {'lr': 0.0004973893147893227, 'samples': 2156032, 'steps': 4210, 'loss/train': 2.629042148590088} 02/24/2022 07:01:31 - INFO - codeparrot_training - Step 4211: {'lr': 0.0004973869557690417, 'samples': 2156544, 'steps': 4211, 'loss/train': 2.9459030628204346} 02/24/2022 07:01:35 - INFO - codeparrot_training - Step 4212: {'lr': 0.0004973845956890336, 'samples': 2157056, 'steps': 4212, 'loss/train': 1.8773908615112305} 02/24/2022 07:01:41 - INFO - codeparrot_training - Step 4213: {'lr': 0.0004973822345493084, 'samples': 2157568, 'steps': 4213, 'loss/train': 2.438155174255371} 02/24/2022 07:01:44 - INFO - codeparrot_training - Step 4214: {'lr': 0.0004973798723498762, 'samples': 2158080, 'steps': 4214, 'loss/train': 3.4575247764587402} 02/24/2022 07:01:50 - INFO - codeparrot_training - Step 4215: {'lr': 0.0004973775090907473, 'samples': 2158592, 'steps': 4215, 'loss/train': 2.744008779525757} 02/24/2022 07:01:53 - INFO - codeparrot_training - Step 4216: {'lr': 0.0004973751447719316, 'samples': 2159104, 'steps': 4216, 'loss/train': 2.998913526535034} 02/24/2022 07:01:59 - INFO - codeparrot_training - Step 4217: {'lr': 0.0004973727793934394, 'samples': 2159616, 'steps': 4217, 'loss/train': 3.5606839656829834} 02/24/2022 07:02:03 - INFO - codeparrot_training - Step 4218: {'lr': 0.0004973704129552808, 'samples': 2160128, 'steps': 4218, 'loss/train': 2.9075167179107666} 02/24/2022 07:02:08 - INFO - codeparrot_training - Step 4219: {'lr': 0.0004973680454574657, 'samples': 2160640, 'steps': 4219, 'loss/train': 2.903878688812256} 02/24/2022 07:02:12 - INFO - codeparrot_training - Step 4220: {'lr': 0.0004973656769000046, 'samples': 2161152, 'steps': 4220, 'loss/train': 3.636286973953247} 02/24/2022 07:02:17 - INFO - codeparrot_training - Step 4221: {'lr': 0.0004973633072829075, 'samples': 2161664, 'steps': 4221, 'loss/train': 2.547485828399658} 02/24/2022 07:02:21 - INFO - codeparrot_training - Step 4222: {'lr': 0.0004973609366061845, 'samples': 2162176, 'steps': 4222, 'loss/train': 1.3098418712615967} 02/24/2022 07:02:26 - INFO - codeparrot_training - Step 4223: {'lr': 0.0004973585648698457, 'samples': 2162688, 'steps': 4223, 'loss/train': 2.7077715396881104} 02/24/2022 07:02:30 - INFO - codeparrot_training - Step 4224: {'lr': 0.0004973561920739015, 'samples': 2163200, 'steps': 4224, 'loss/train': 1.7139753103256226} 02/24/2022 07:02:35 - INFO - codeparrot_training - Step 4225: {'lr': 0.0004973538182183618, 'samples': 2163712, 'steps': 4225, 'loss/train': 3.5012283325195312} 02/24/2022 07:02:39 - INFO - codeparrot_training - Step 4226: {'lr': 0.000497351443303237, 'samples': 2164224, 'steps': 4226, 'loss/train': 1.6470822095870972} 02/24/2022 07:02:45 - INFO - codeparrot_training - Step 4227: {'lr': 0.0004973490673285372, 'samples': 2164736, 'steps': 4227, 'loss/train': 1.578709602355957} 02/24/2022 07:02:48 - INFO - codeparrot_training - Step 4228: {'lr': 0.0004973466902942723, 'samples': 2165248, 'steps': 4228, 'loss/train': 2.0621304512023926} 02/24/2022 07:02:54 - INFO - codeparrot_training - Step 4229: {'lr': 0.0004973443122004529, 'samples': 2165760, 'steps': 4229, 'loss/train': 2.8437130451202393} 02/24/2022 07:02:57 - INFO - codeparrot_training - Step 4230: {'lr': 0.0004973419330470891, 'samples': 2166272, 'steps': 4230, 'loss/train': 3.1119749546051025} 02/24/2022 07:03:03 - INFO - codeparrot_training - Step 4231: {'lr': 0.0004973395528341908, 'samples': 2166784, 'steps': 4231, 'loss/train': 2.8233418464660645} 02/24/2022 07:03:06 - INFO - codeparrot_training - Step 4232: {'lr': 0.0004973371715617685, 'samples': 2167296, 'steps': 4232, 'loss/train': 3.0032131671905518} 02/24/2022 07:03:12 - INFO - codeparrot_training - Step 4233: {'lr': 0.0004973347892298322, 'samples': 2167808, 'steps': 4233, 'loss/train': 2.891402006149292} 02/24/2022 07:03:15 - INFO - codeparrot_training - Step 4234: {'lr': 0.0004973324058383924, 'samples': 2168320, 'steps': 4234, 'loss/train': 0.458845853805542} 02/24/2022 07:03:21 - INFO - codeparrot_training - Step 4235: {'lr': 0.0004973300213874589, 'samples': 2168832, 'steps': 4235, 'loss/train': 1.8262721300125122} 02/24/2022 07:03:24 - INFO - codeparrot_training - Step 4236: {'lr': 0.0004973276358770422, 'samples': 2169344, 'steps': 4236, 'loss/train': 2.673147201538086} 02/24/2022 07:03:30 - INFO - codeparrot_training - Step 4237: {'lr': 0.0004973252493071525, 'samples': 2169856, 'steps': 4237, 'loss/train': 2.2957446575164795} 02/24/2022 07:03:34 - INFO - codeparrot_training - Step 4238: {'lr': 0.0004973228616777999, 'samples': 2170368, 'steps': 4238, 'loss/train': 2.6661338806152344} 02/24/2022 07:03:40 - INFO - codeparrot_training - Step 4239: {'lr': 0.0004973204729889946, 'samples': 2170880, 'steps': 4239, 'loss/train': 2.6410841941833496} 02/24/2022 07:03:43 - INFO - codeparrot_training - Step 4240: {'lr': 0.0004973180832407472, 'samples': 2171392, 'steps': 4240, 'loss/train': 3.154447317123413} 02/24/2022 07:03:48 - INFO - codeparrot_training - Step 4241: {'lr': 0.0004973156924330674, 'samples': 2171904, 'steps': 4241, 'loss/train': 3.064802646636963} 02/24/2022 07:03:52 - INFO - codeparrot_training - Step 4242: {'lr': 0.0004973133005659658, 'samples': 2172416, 'steps': 4242, 'loss/train': 3.259582996368408} 02/24/2022 07:03:58 - INFO - codeparrot_training - Step 4243: {'lr': 0.0004973109076394526, 'samples': 2172928, 'steps': 4243, 'loss/train': 2.3444747924804688} 02/24/2022 07:04:01 - INFO - codeparrot_training - Step 4244: {'lr': 0.0004973085136535379, 'samples': 2173440, 'steps': 4244, 'loss/train': 2.6797614097595215} 02/24/2022 07:04:07 - INFO - codeparrot_training - Step 4245: {'lr': 0.000497306118608232, 'samples': 2173952, 'steps': 4245, 'loss/train': 1.1895850896835327} 02/24/2022 07:04:10 - INFO - codeparrot_training - Step 4246: {'lr': 0.0004973037225035454, 'samples': 2174464, 'steps': 4246, 'loss/train': 3.039938449859619} 02/24/2022 07:04:16 - INFO - codeparrot_training - Step 4247: {'lr': 0.0004973013253394881, 'samples': 2174976, 'steps': 4247, 'loss/train': 1.6467804908752441} 02/24/2022 07:04:19 - INFO - codeparrot_training - Step 4248: {'lr': 0.0004972989271160705, 'samples': 2175488, 'steps': 4248, 'loss/train': 3.7007312774658203} 02/24/2022 07:04:25 - INFO - codeparrot_training - Step 4249: {'lr': 0.0004972965278333028, 'samples': 2176000, 'steps': 4249, 'loss/train': 3.049842119216919} 02/24/2022 07:04:30 - INFO - codeparrot_training - Step 4250: {'lr': 0.0004972941274911952, 'samples': 2176512, 'steps': 4250, 'loss/train': 3.7673535346984863} 02/24/2022 07:04:34 - INFO - codeparrot_training - Step 4251: {'lr': 0.0004972917260897583, 'samples': 2177024, 'steps': 4251, 'loss/train': 2.8709936141967773} 02/24/2022 07:04:40 - INFO - codeparrot_training - Step 4252: {'lr': 0.0004972893236290019, 'samples': 2177536, 'steps': 4252, 'loss/train': 3.36670184135437} 02/24/2022 07:04:43 - INFO - codeparrot_training - Step 4253: {'lr': 0.0004972869201089367, 'samples': 2178048, 'steps': 4253, 'loss/train': 3.524763584136963} 02/24/2022 07:04:49 - INFO - codeparrot_training - Step 4254: {'lr': 0.0004972845155295729, 'samples': 2178560, 'steps': 4254, 'loss/train': 2.793093204498291} 02/24/2022 07:04:52 - INFO - codeparrot_training - Step 4255: {'lr': 0.0004972821098909207, 'samples': 2179072, 'steps': 4255, 'loss/train': 1.677321195602417} 02/24/2022 07:04:58 - INFO - codeparrot_training - Step 4256: {'lr': 0.0004972797031929904, 'samples': 2179584, 'steps': 4256, 'loss/train': 1.6498457193374634} 02/24/2022 07:05:01 - INFO - codeparrot_training - Step 4257: {'lr': 0.0004972772954357924, 'samples': 2180096, 'steps': 4257, 'loss/train': 2.9303619861602783} 02/24/2022 07:05:07 - INFO - codeparrot_training - Step 4258: {'lr': 0.0004972748866193371, 'samples': 2180608, 'steps': 4258, 'loss/train': 3.9064760208129883} 02/24/2022 07:05:10 - INFO - codeparrot_training - Step 4259: {'lr': 0.0004972724767436346, 'samples': 2181120, 'steps': 4259, 'loss/train': 2.702552556991577} 02/24/2022 07:05:16 - INFO - codeparrot_training - Step 4260: {'lr': 0.0004972700658086954, 'samples': 2181632, 'steps': 4260, 'loss/train': 2.3609423637390137} 02/24/2022 07:05:19 - INFO - codeparrot_training - Step 4261: {'lr': 0.0004972676538145298, 'samples': 2182144, 'steps': 4261, 'loss/train': 4.089694499969482} 02/24/2022 07:05:25 - INFO - codeparrot_training - Step 4262: {'lr': 0.0004972652407611479, 'samples': 2182656, 'steps': 4262, 'loss/train': 3.2853028774261475} 02/24/2022 07:05:29 - INFO - codeparrot_training - Step 4263: {'lr': 0.0004972628266485604, 'samples': 2183168, 'steps': 4263, 'loss/train': 0.9743463397026062} 02/24/2022 07:05:34 - INFO - codeparrot_training - Step 4264: {'lr': 0.0004972604114767774, 'samples': 2183680, 'steps': 4264, 'loss/train': 3.7340307235717773} 02/24/2022 07:05:38 - INFO - codeparrot_training - Step 4265: {'lr': 0.0004972579952458092, 'samples': 2184192, 'steps': 4265, 'loss/train': 2.7161076068878174} 02/24/2022 07:05:43 - INFO - codeparrot_training - Step 4266: {'lr': 0.0004972555779556664, 'samples': 2184704, 'steps': 4266, 'loss/train': 2.780529260635376} 02/24/2022 07:05:47 - INFO - codeparrot_training - Step 4267: {'lr': 0.0004972531596063592, 'samples': 2185216, 'steps': 4267, 'loss/train': 2.639686346054077} 02/24/2022 07:05:52 - INFO - codeparrot_training - Step 4268: {'lr': 0.000497250740197898, 'samples': 2185728, 'steps': 4268, 'loss/train': 1.4798073768615723} 02/24/2022 07:05:56 - INFO - codeparrot_training - Step 4269: {'lr': 0.0004972483197302931, 'samples': 2186240, 'steps': 4269, 'loss/train': 1.8299784660339355} 02/24/2022 07:06:01 - INFO - codeparrot_training - Step 4270: {'lr': 0.0004972458982035548, 'samples': 2186752, 'steps': 4270, 'loss/train': 2.852864980697632} 02/24/2022 07:06:05 - INFO - codeparrot_training - Step 4271: {'lr': 0.0004972434756176937, 'samples': 2187264, 'steps': 4271, 'loss/train': 8.189410209655762} 02/24/2022 07:06:11 - INFO - codeparrot_training - Step 4272: {'lr': 0.0004972410519727201, 'samples': 2187776, 'steps': 4272, 'loss/train': 2.3508660793304443} 02/24/2022 07:06:14 - INFO - codeparrot_training - Step 4273: {'lr': 0.0004972386272686443, 'samples': 2188288, 'steps': 4273, 'loss/train': 1.5492292642593384} 02/24/2022 07:06:20 - INFO - codeparrot_training - Step 4274: {'lr': 0.0004972362015054767, 'samples': 2188800, 'steps': 4274, 'loss/train': 3.1584479808807373} 02/24/2022 07:06:23 - INFO - codeparrot_training - Step 4275: {'lr': 0.0004972337746832278, 'samples': 2189312, 'steps': 4275, 'loss/train': 2.052060842514038} 02/24/2022 07:06:29 - INFO - codeparrot_training - Step 4276: {'lr': 0.0004972313468019077, 'samples': 2189824, 'steps': 4276, 'loss/train': 2.50105881690979} 02/24/2022 07:06:32 - INFO - codeparrot_training - Step 4277: {'lr': 0.0004972289178615273, 'samples': 2190336, 'steps': 4277, 'loss/train': 1.3975589275360107} 02/24/2022 07:06:38 - INFO - codeparrot_training - Step 4278: {'lr': 0.0004972264878620965, 'samples': 2190848, 'steps': 4278, 'loss/train': 2.5786895751953125} 02/24/2022 07:06:41 - INFO - codeparrot_training - Step 4279: {'lr': 0.000497224056803626, 'samples': 2191360, 'steps': 4279, 'loss/train': 2.410680055618286} 02/24/2022 07:06:47 - INFO - codeparrot_training - Step 4280: {'lr': 0.0004972216246861262, 'samples': 2191872, 'steps': 4280, 'loss/train': 2.154649019241333} 02/24/2022 07:06:50 - INFO - codeparrot_training - Step 4281: {'lr': 0.0004972191915096074, 'samples': 2192384, 'steps': 4281, 'loss/train': 2.531623363494873} 02/24/2022 07:06:56 - INFO - codeparrot_training - Step 4282: {'lr': 0.0004972167572740801, 'samples': 2192896, 'steps': 4282, 'loss/train': 2.173480749130249} 02/24/2022 07:07:00 - INFO - codeparrot_training - Step 4283: {'lr': 0.0004972143219795547, 'samples': 2193408, 'steps': 4283, 'loss/train': 0.3406789302825928} 02/24/2022 07:07:06 - INFO - codeparrot_training - Step 4284: {'lr': 0.0004972118856260416, 'samples': 2193920, 'steps': 4284, 'loss/train': 2.5170040130615234} 02/24/2022 07:07:09 - INFO - codeparrot_training - Step 4285: {'lr': 0.0004972094482135514, 'samples': 2194432, 'steps': 4285, 'loss/train': 3.0732195377349854} 02/24/2022 07:07:15 - INFO - codeparrot_training - Step 4286: {'lr': 0.0004972070097420943, 'samples': 2194944, 'steps': 4286, 'loss/train': 2.393036365509033} 02/24/2022 07:07:18 - INFO - codeparrot_training - Step 4287: {'lr': 0.0004972045702116809, 'samples': 2195456, 'steps': 4287, 'loss/train': 2.8728830814361572} 02/24/2022 07:07:24 - INFO - codeparrot_training - Step 4288: {'lr': 0.0004972021296223217, 'samples': 2195968, 'steps': 4288, 'loss/train': 2.898014545440674} 02/24/2022 07:07:27 - INFO - codeparrot_training - Step 4289: {'lr': 0.0004971996879740271, 'samples': 2196480, 'steps': 4289, 'loss/train': 2.6281983852386475} 02/24/2022 07:07:33 - INFO - codeparrot_training - Step 4290: {'lr': 0.0004971972452668074, 'samples': 2196992, 'steps': 4290, 'loss/train': 2.692924737930298} 02/24/2022 07:07:36 - INFO - codeparrot_training - Step 4291: {'lr': 0.0004971948015006732, 'samples': 2197504, 'steps': 4291, 'loss/train': 1.9726898670196533} 02/24/2022 07:07:42 - INFO - codeparrot_training - Step 4292: {'lr': 0.000497192356675635, 'samples': 2198016, 'steps': 4292, 'loss/train': 2.1184329986572266} 02/24/2022 07:07:45 - INFO - codeparrot_training - Step 4293: {'lr': 0.0004971899107917033, 'samples': 2198528, 'steps': 4293, 'loss/train': 1.4840930700302124} 02/24/2022 07:07:51 - INFO - codeparrot_training - Step 4294: {'lr': 0.0004971874638488884, 'samples': 2199040, 'steps': 4294, 'loss/train': 2.085710287094116} 02/24/2022 07:07:54 - INFO - codeparrot_training - Step 4295: {'lr': 0.000497185015847201, 'samples': 2199552, 'steps': 4295, 'loss/train': 2.7963337898254395} 02/24/2022 07:08:00 - INFO - codeparrot_training - Step 4296: {'lr': 0.0004971825667866515, 'samples': 2200064, 'steps': 4296, 'loss/train': 3.636133909225464} 02/24/2022 07:08:03 - INFO - codeparrot_training - Step 4297: {'lr': 0.0004971801166672502, 'samples': 2200576, 'steps': 4297, 'loss/train': 2.0215768814086914} 02/24/2022 07:08:09 - INFO - codeparrot_training - Step 4298: {'lr': 0.0004971776654890079, 'samples': 2201088, 'steps': 4298, 'loss/train': 2.9988853931427} 02/24/2022 07:08:13 - INFO - codeparrot_training - Step 4299: {'lr': 0.000497175213251935, 'samples': 2201600, 'steps': 4299, 'loss/train': 1.9716269969940186} 02/24/2022 07:08:18 - INFO - codeparrot_training - Step 4300: {'lr': 0.0004971727599560418, 'samples': 2202112, 'steps': 4300, 'loss/train': 1.5621925592422485} 02/24/2022 07:08:22 - INFO - codeparrot_training - Step 4301: {'lr': 0.0004971703056013392, 'samples': 2202624, 'steps': 4301, 'loss/train': 3.330235004425049} 02/24/2022 07:08:27 - INFO - codeparrot_training - Step 4302: {'lr': 0.0004971678501878374, 'samples': 2203136, 'steps': 4302, 'loss/train': 4.034710884094238} 02/24/2022 07:08:31 - INFO - codeparrot_training - Step 4303: {'lr': 0.000497165393715547, 'samples': 2203648, 'steps': 4303, 'loss/train': 1.7335206270217896} 02/24/2022 07:08:36 - INFO - codeparrot_training - Step 4304: {'lr': 0.0004971629361844785, 'samples': 2204160, 'steps': 4304, 'loss/train': 1.8462622165679932} 02/24/2022 07:08:40 - INFO - codeparrot_training - Step 4305: {'lr': 0.0004971604775946425, 'samples': 2204672, 'steps': 4305, 'loss/train': 3.579932928085327} 02/24/2022 07:08:45 - INFO - codeparrot_training - Step 4306: {'lr': 0.0004971580179460495, 'samples': 2205184, 'steps': 4306, 'loss/train': 1.6447175741195679} 02/24/2022 07:08:49 - INFO - codeparrot_training - Step 4307: {'lr': 0.0004971555572387101, 'samples': 2205696, 'steps': 4307, 'loss/train': 2.691662311553955} 02/24/2022 07:08:55 - INFO - codeparrot_training - Step 4308: {'lr': 0.0004971530954726346, 'samples': 2206208, 'steps': 4308, 'loss/train': 3.363874673843384} 02/24/2022 07:08:58 - INFO - codeparrot_training - Step 4309: {'lr': 0.0004971506326478339, 'samples': 2206720, 'steps': 4309, 'loss/train': 3.470662832260132} 02/24/2022 07:09:04 - INFO - codeparrot_training - Step 4310: {'lr': 0.0004971481687643184, 'samples': 2207232, 'steps': 4310, 'loss/train': 2.9335923194885254} 02/24/2022 07:09:07 - INFO - codeparrot_training - Step 4311: {'lr': 0.0004971457038220984, 'samples': 2207744, 'steps': 4311, 'loss/train': 2.650141477584839} 02/24/2022 07:09:13 - INFO - codeparrot_training - Step 4312: {'lr': 0.0004971432378211849, 'samples': 2208256, 'steps': 4312, 'loss/train': 3.1428043842315674} 02/24/2022 07:09:16 - INFO - codeparrot_training - Step 4313: {'lr': 0.0004971407707615881, 'samples': 2208768, 'steps': 4313, 'loss/train': 2.7456884384155273} 02/24/2022 07:09:24 - INFO - codeparrot_training - Step 4314: {'lr': 0.0004971383026433189, 'samples': 2209280, 'steps': 4314, 'loss/train': 2.857170343399048} 02/24/2022 07:09:28 - INFO - codeparrot_training - Step 4315: {'lr': 0.0004971358334663875, 'samples': 2209792, 'steps': 4315, 'loss/train': 1.715367317199707} 02/24/2022 07:09:33 - INFO - codeparrot_training - Step 4316: {'lr': 0.0004971333632308047, 'samples': 2210304, 'steps': 4316, 'loss/train': 3.0234556198120117} 02/24/2022 07:09:36 - INFO - codeparrot_training - Step 4317: {'lr': 0.000497130891936581, 'samples': 2210816, 'steps': 4317, 'loss/train': 3.300001859664917} 02/24/2022 07:09:42 - INFO - codeparrot_training - Step 4318: {'lr': 0.0004971284195837271, 'samples': 2211328, 'steps': 4318, 'loss/train': 2.5985090732574463} 02/24/2022 07:09:46 - INFO - codeparrot_training - Step 4319: {'lr': 0.0004971259461722536, 'samples': 2211840, 'steps': 4319, 'loss/train': 2.2258150577545166} 02/24/2022 07:09:51 - INFO - codeparrot_training - Step 4320: {'lr': 0.0004971234717021708, 'samples': 2212352, 'steps': 4320, 'loss/train': 1.5778388977050781} 02/24/2022 07:09:54 - INFO - codeparrot_training - Step 4321: {'lr': 0.0004971209961734897, 'samples': 2212864, 'steps': 4321, 'loss/train': 1.0925599336624146} 02/24/2022 07:10:00 - INFO - codeparrot_training - Step 4322: {'lr': 0.0004971185195862207, 'samples': 2213376, 'steps': 4322, 'loss/train': 2.530054807662964} 02/24/2022 07:10:03 - INFO - codeparrot_training - Step 4323: {'lr': 0.0004971160419403744, 'samples': 2213888, 'steps': 4323, 'loss/train': 0.9472009539604187} 02/24/2022 07:10:11 - INFO - codeparrot_training - Step 4324: {'lr': 0.0004971135632359614, 'samples': 2214400, 'steps': 4324, 'loss/train': 2.86696720123291} 02/24/2022 07:10:14 - INFO - codeparrot_training - Step 4325: {'lr': 0.0004971110834729925, 'samples': 2214912, 'steps': 4325, 'loss/train': 2.8923516273498535} 02/24/2022 07:10:20 - INFO - codeparrot_training - Step 4326: {'lr': 0.0004971086026514781, 'samples': 2215424, 'steps': 4326, 'loss/train': 2.862438201904297} 02/24/2022 07:10:23 - INFO - codeparrot_training - Step 4327: {'lr': 0.0004971061207714289, 'samples': 2215936, 'steps': 4327, 'loss/train': 2.9758591651916504} 02/24/2022 07:10:29 - INFO - codeparrot_training - Step 4328: {'lr': 0.0004971036378328556, 'samples': 2216448, 'steps': 4328, 'loss/train': 1.4875649213790894} 02/24/2022 07:10:32 - INFO - codeparrot_training - Step 4329: {'lr': 0.0004971011538357687, 'samples': 2216960, 'steps': 4329, 'loss/train': 2.493306875228882} 02/24/2022 07:10:38 - INFO - codeparrot_training - Step 4330: {'lr': 0.000497098668780179, 'samples': 2217472, 'steps': 4330, 'loss/train': 3.131075143814087} 02/24/2022 07:10:41 - INFO - codeparrot_training - Step 4331: {'lr': 0.000497096182666097, 'samples': 2217984, 'steps': 4331, 'loss/train': 3.5481016635894775} 02/24/2022 07:10:47 - INFO - codeparrot_training - Step 4332: {'lr': 0.0004970936954935334, 'samples': 2218496, 'steps': 4332, 'loss/train': 2.7496330738067627} 02/24/2022 07:10:50 - INFO - codeparrot_training - Step 4333: {'lr': 0.0004970912072624989, 'samples': 2219008, 'steps': 4333, 'loss/train': 2.2501373291015625} 02/24/2022 07:10:57 - INFO - codeparrot_training - Step 4334: {'lr': 0.0004970887179730041, 'samples': 2219520, 'steps': 4334, 'loss/train': 2.827500820159912} 02/24/2022 07:11:01 - INFO - codeparrot_training - Step 4335: {'lr': 0.0004970862276250599, 'samples': 2220032, 'steps': 4335, 'loss/train': 3.508098602294922} 02/24/2022 07:11:06 - INFO - codeparrot_training - Step 4336: {'lr': 0.0004970837362186766, 'samples': 2220544, 'steps': 4336, 'loss/train': 1.368098258972168} 02/24/2022 07:11:10 - INFO - codeparrot_training - Step 4337: {'lr': 0.0004970812437538649, 'samples': 2221056, 'steps': 4337, 'loss/train': 3.964770555496216} 02/24/2022 07:11:15 - INFO - codeparrot_training - Step 4338: {'lr': 0.0004970787502306357, 'samples': 2221568, 'steps': 4338, 'loss/train': 1.5995824337005615} 02/24/2022 07:11:19 - INFO - codeparrot_training - Step 4339: {'lr': 0.0004970762556489996, 'samples': 2222080, 'steps': 4339, 'loss/train': 2.1084766387939453} 02/24/2022 07:11:24 - INFO - codeparrot_training - Step 4340: {'lr': 0.0004970737600089673, 'samples': 2222592, 'steps': 4340, 'loss/train': 2.1236257553100586} 02/24/2022 07:11:27 - INFO - codeparrot_training - Step 4341: {'lr': 0.0004970712633105496, 'samples': 2223104, 'steps': 4341, 'loss/train': 3.0250909328460693} 02/24/2022 07:11:33 - INFO - codeparrot_training - Step 4342: {'lr': 0.0004970687655537568, 'samples': 2223616, 'steps': 4342, 'loss/train': 3.0731847286224365} 02/24/2022 07:11:37 - INFO - codeparrot_training - Step 4343: {'lr': 0.0004970662667386, 'samples': 2224128, 'steps': 4343, 'loss/train': 1.9997105598449707} 02/24/2022 07:11:44 - INFO - codeparrot_training - Step 4344: {'lr': 0.0004970637668650898, 'samples': 2224640, 'steps': 4344, 'loss/train': 0.218427374958992} 02/24/2022 07:11:47 - INFO - codeparrot_training - Step 4345: {'lr': 0.0004970612659332368, 'samples': 2225152, 'steps': 4345, 'loss/train': 2.3627676963806152} 02/24/2022 07:11:53 - INFO - codeparrot_training - Step 4346: {'lr': 0.0004970587639430518, 'samples': 2225664, 'steps': 4346, 'loss/train': 2.492006301879883} 02/24/2022 07:11:56 - INFO - codeparrot_training - Step 4347: {'lr': 0.0004970562608945455, 'samples': 2226176, 'steps': 4347, 'loss/train': 2.6539816856384277} 02/24/2022 07:12:02 - INFO - codeparrot_training - Step 4348: {'lr': 0.0004970537567877286, 'samples': 2226688, 'steps': 4348, 'loss/train': 3.71990966796875} 02/24/2022 07:12:05 - INFO - codeparrot_training - Step 4349: {'lr': 0.000497051251622612, 'samples': 2227200, 'steps': 4349, 'loss/train': 0.6602520942687988} 02/24/2022 07:12:11 - INFO - codeparrot_training - Step 4350: {'lr': 0.0004970487453992062, 'samples': 2227712, 'steps': 4350, 'loss/train': 2.111689329147339} 02/24/2022 07:12:16 - INFO - codeparrot_training - Step 4351: {'lr': 0.000497046238117522, 'samples': 2228224, 'steps': 4351, 'loss/train': 1.2905681133270264} 02/24/2022 07:12:20 - INFO - codeparrot_training - Step 4352: {'lr': 0.0004970437297775702, 'samples': 2228736, 'steps': 4352, 'loss/train': 3.3549752235412598} 02/24/2022 07:12:27 - INFO - codeparrot_training - Step 4353: {'lr': 0.0004970412203793614, 'samples': 2229248, 'steps': 4353, 'loss/train': 1.9374116659164429} 02/24/2022 07:12:31 - INFO - codeparrot_training - Step 4354: {'lr': 0.0004970387099229066, 'samples': 2229760, 'steps': 4354, 'loss/train': 3.28519606590271} 02/24/2022 07:12:36 - INFO - codeparrot_training - Step 4355: {'lr': 0.0004970361984082163, 'samples': 2230272, 'steps': 4355, 'loss/train': 2.5305819511413574} 02/24/2022 07:12:39 - INFO - codeparrot_training - Step 4356: {'lr': 0.0004970336858353014, 'samples': 2230784, 'steps': 4356, 'loss/train': 2.907233953475952} 02/24/2022 07:12:45 - INFO - codeparrot_training - Step 4357: {'lr': 0.0004970311722041727, 'samples': 2231296, 'steps': 4357, 'loss/train': 2.892895460128784} 02/24/2022 07:12:48 - INFO - codeparrot_training - Step 4358: {'lr': 0.0004970286575148408, 'samples': 2231808, 'steps': 4358, 'loss/train': 2.534895181655884} 02/24/2022 07:12:54 - INFO - codeparrot_training - Step 4359: {'lr': 0.0004970261417673165, 'samples': 2232320, 'steps': 4359, 'loss/train': 2.9031383991241455} 02/24/2022 07:12:57 - INFO - codeparrot_training - Step 4360: {'lr': 0.0004970236249616109, 'samples': 2232832, 'steps': 4360, 'loss/train': 2.6167120933532715} 02/24/2022 07:13:03 - INFO - codeparrot_training - Step 4361: {'lr': 0.0004970211070977344, 'samples': 2233344, 'steps': 4361, 'loss/train': 1.2957284450531006} 02/24/2022 07:13:06 - INFO - codeparrot_training - Step 4362: {'lr': 0.0004970185881756979, 'samples': 2233856, 'steps': 4362, 'loss/train': 2.167532205581665} 02/24/2022 07:13:12 - INFO - codeparrot_training - Step 4363: {'lr': 0.0004970160681955121, 'samples': 2234368, 'steps': 4363, 'loss/train': 3.269674062728882} 02/24/2022 07:13:15 - INFO - codeparrot_training - Step 4364: {'lr': 0.0004970135471571881, 'samples': 2234880, 'steps': 4364, 'loss/train': 1.3479008674621582} 02/24/2022 07:13:21 - INFO - codeparrot_training - Step 4365: {'lr': 0.0004970110250607364, 'samples': 2235392, 'steps': 4365, 'loss/train': 2.534153461456299} 02/24/2022 07:13:24 - INFO - codeparrot_training - Step 4366: {'lr': 0.000497008501906168, 'samples': 2235904, 'steps': 4366, 'loss/train': 3.1038362979888916} 02/24/2022 07:13:30 - INFO - codeparrot_training - Step 4367: {'lr': 0.0004970059776934935, 'samples': 2236416, 'steps': 4367, 'loss/train': 3.128077983856201} 02/24/2022 07:13:33 - INFO - codeparrot_training - Step 4368: {'lr': 0.0004970034524227238, 'samples': 2236928, 'steps': 4368, 'loss/train': 2.852567434310913} 02/24/2022 07:13:41 - INFO - codeparrot_training - Step 4369: {'lr': 0.0004970009260938698, 'samples': 2237440, 'steps': 4369, 'loss/train': 3.0271027088165283} 02/24/2022 07:13:44 - INFO - codeparrot_training - Step 4370: {'lr': 0.0004969983987069423, 'samples': 2237952, 'steps': 4370, 'loss/train': 2.3852086067199707} 02/24/2022 07:13:50 - INFO - codeparrot_training - Step 4371: {'lr': 0.000496995870261952, 'samples': 2238464, 'steps': 4371, 'loss/train': 3.10168719291687} 02/24/2022 07:13:53 - INFO - codeparrot_training - Step 4372: {'lr': 0.0004969933407589098, 'samples': 2238976, 'steps': 4372, 'loss/train': 2.5915274620056152} 02/24/2022 07:13:59 - INFO - codeparrot_training - Step 4373: {'lr': 0.0004969908101978267, 'samples': 2239488, 'steps': 4373, 'loss/train': 1.5831700563430786} 02/24/2022 07:14:02 - INFO - codeparrot_training - Step 4374: {'lr': 0.0004969882785787133, 'samples': 2240000, 'steps': 4374, 'loss/train': 2.789936065673828} 02/24/2022 07:14:08 - INFO - codeparrot_training - Step 4375: {'lr': 0.0004969857459015807, 'samples': 2240512, 'steps': 4375, 'loss/train': 1.6095125675201416} 02/24/2022 07:14:11 - INFO - codeparrot_training - Step 4376: {'lr': 0.0004969832121664394, 'samples': 2241024, 'steps': 4376, 'loss/train': 2.221705675125122} 02/24/2022 07:14:17 - INFO - codeparrot_training - Step 4377: {'lr': 0.0004969806773733004, 'samples': 2241536, 'steps': 4377, 'loss/train': 1.3647468090057373} 02/24/2022 07:14:20 - INFO - codeparrot_training - Step 4378: {'lr': 0.0004969781415221748, 'samples': 2242048, 'steps': 4378, 'loss/train': 2.6426949501037598} 02/24/2022 07:14:28 - INFO - codeparrot_training - Step 4379: {'lr': 0.0004969756046130731, 'samples': 2242560, 'steps': 4379, 'loss/train': 1.2999351024627686} 02/24/2022 07:14:31 - INFO - codeparrot_training - Step 4380: {'lr': 0.0004969730666460065, 'samples': 2243072, 'steps': 4380, 'loss/train': 3.769508123397827} 02/24/2022 07:14:37 - INFO - codeparrot_training - Step 4381: {'lr': 0.0004969705276209856, 'samples': 2243584, 'steps': 4381, 'loss/train': 3.0358543395996094} 02/24/2022 07:14:40 - INFO - codeparrot_training - Step 4382: {'lr': 0.0004969679875380214, 'samples': 2244096, 'steps': 4382, 'loss/train': 3.4092957973480225} 02/24/2022 07:14:46 - INFO - codeparrot_training - Step 4383: {'lr': 0.0004969654463971247, 'samples': 2244608, 'steps': 4383, 'loss/train': 2.6508755683898926} 02/24/2022 07:14:49 - INFO - codeparrot_training - Step 4384: {'lr': 0.0004969629041983065, 'samples': 2245120, 'steps': 4384, 'loss/train': 2.873786211013794} 02/24/2022 07:14:55 - INFO - codeparrot_training - Step 4385: {'lr': 0.0004969603609415777, 'samples': 2245632, 'steps': 4385, 'loss/train': 2.103040933609009} 02/24/2022 07:14:58 - INFO - codeparrot_training - Step 4386: {'lr': 0.000496957816626949, 'samples': 2246144, 'steps': 4386, 'loss/train': 3.001793146133423} 02/24/2022 07:15:04 - INFO - codeparrot_training - Step 4387: {'lr': 0.0004969552712544316, 'samples': 2246656, 'steps': 4387, 'loss/train': 2.502504825592041} 02/24/2022 07:15:07 - INFO - codeparrot_training - Step 4388: {'lr': 0.0004969527248240361, 'samples': 2247168, 'steps': 4388, 'loss/train': 2.7314093112945557} 02/24/2022 07:15:14 - INFO - codeparrot_training - Step 4389: {'lr': 0.0004969501773357736, 'samples': 2247680, 'steps': 4389, 'loss/train': 0.27108290791511536} 02/24/2022 07:15:18 - INFO - codeparrot_training - Step 4390: {'lr': 0.000496947628789655, 'samples': 2248192, 'steps': 4390, 'loss/train': 2.2951478958129883} 02/24/2022 07:15:23 - INFO - codeparrot_training - Step 4391: {'lr': 0.000496945079185691, 'samples': 2248704, 'steps': 4391, 'loss/train': 3.159607410430908} 02/24/2022 07:15:27 - INFO - codeparrot_training - Step 4392: {'lr': 0.0004969425285238928, 'samples': 2249216, 'steps': 4392, 'loss/train': 8.981085777282715} 02/24/2022 07:15:32 - INFO - codeparrot_training - Step 4393: {'lr': 0.0004969399768042713, 'samples': 2249728, 'steps': 4393, 'loss/train': 2.296780824661255} 02/24/2022 07:15:36 - INFO - codeparrot_training - Step 4394: {'lr': 0.0004969374240268373, 'samples': 2250240, 'steps': 4394, 'loss/train': 2.9942376613616943} 02/24/2022 07:15:41 - INFO - codeparrot_training - Step 4395: {'lr': 0.0004969348701916018, 'samples': 2250752, 'steps': 4395, 'loss/train': 3.0727157592773438} 02/24/2022 07:15:45 - INFO - codeparrot_training - Step 4396: {'lr': 0.0004969323152985756, 'samples': 2251264, 'steps': 4396, 'loss/train': 1.7974709272384644} 02/24/2022 07:15:50 - INFO - codeparrot_training - Step 4397: {'lr': 0.0004969297593477699, 'samples': 2251776, 'steps': 4397, 'loss/train': 3.8683183193206787} 02/24/2022 07:15:54 - INFO - codeparrot_training - Step 4398: {'lr': 0.0004969272023391955, 'samples': 2252288, 'steps': 4398, 'loss/train': 3.408257484436035} 02/24/2022 07:15:59 - INFO - codeparrot_training - Step 4399: {'lr': 0.0004969246442728633, 'samples': 2252800, 'steps': 4399, 'loss/train': 2.9289355278015137} 02/24/2022 07:16:03 - INFO - codeparrot_training - Step 4400: {'lr': 0.0004969220851487844, 'samples': 2253312, 'steps': 4400, 'loss/train': 2.622563123703003} 02/24/2022 07:16:10 - INFO - codeparrot_training - Step 4401: {'lr': 0.0004969195249669697, 'samples': 2253824, 'steps': 4401, 'loss/train': 3.680387496948242} 02/24/2022 07:16:13 - INFO - codeparrot_training - Step 4402: {'lr': 0.0004969169637274301, 'samples': 2254336, 'steps': 4402, 'loss/train': 1.7651646137237549} 02/24/2022 07:16:19 - INFO - codeparrot_training - Step 4403: {'lr': 0.0004969144014301767, 'samples': 2254848, 'steps': 4403, 'loss/train': 3.1497962474823} 02/24/2022 07:16:22 - INFO - codeparrot_training - Step 4404: {'lr': 0.0004969118380752205, 'samples': 2255360, 'steps': 4404, 'loss/train': 1.8213385343551636} 02/24/2022 07:16:28 - INFO - codeparrot_training - Step 4405: {'lr': 0.0004969092736625722, 'samples': 2255872, 'steps': 4405, 'loss/train': 1.1485289335250854} 02/24/2022 07:16:31 - INFO - codeparrot_training - Step 4406: {'lr': 0.000496906708192243, 'samples': 2256384, 'steps': 4406, 'loss/train': 2.1710715293884277} 02/24/2022 07:16:37 - INFO - codeparrot_training - Step 4407: {'lr': 0.000496904141664244, 'samples': 2256896, 'steps': 4407, 'loss/train': 4.6458210945129395} 02/24/2022 07:16:40 - INFO - codeparrot_training - Step 4408: {'lr': 0.0004969015740785859, 'samples': 2257408, 'steps': 4408, 'loss/train': 2.5125210285186768} 02/24/2022 07:16:46 - INFO - codeparrot_training - Step 4409: {'lr': 0.00049689900543528, 'samples': 2257920, 'steps': 4409, 'loss/train': 2.992716073989868} 02/24/2022 07:16:49 - INFO - codeparrot_training - Step 4410: {'lr': 0.0004968964357343371, 'samples': 2258432, 'steps': 4410, 'loss/train': 2.815979480743408} 02/24/2022 07:16:55 - INFO - codeparrot_training - Step 4411: {'lr': 0.0004968938649757682, 'samples': 2258944, 'steps': 4411, 'loss/train': 2.444709062576294} 02/24/2022 07:16:58 - INFO - codeparrot_training - Step 4412: {'lr': 0.0004968912931595845, 'samples': 2259456, 'steps': 4412, 'loss/train': 2.953953504562378} 02/24/2022 07:17:04 - INFO - codeparrot_training - Step 4413: {'lr': 0.0004968887202857968, 'samples': 2259968, 'steps': 4413, 'loss/train': 2.799375057220459} 02/24/2022 07:17:07 - INFO - codeparrot_training - Step 4414: {'lr': 0.0004968861463544163, 'samples': 2260480, 'steps': 4414, 'loss/train': 1.8723090887069702} 02/24/2022 07:17:15 - INFO - codeparrot_training - Step 4415: {'lr': 0.0004968835713654538, 'samples': 2260992, 'steps': 4415, 'loss/train': 2.4066226482391357} 02/24/2022 07:17:18 - INFO - codeparrot_training - Step 4416: {'lr': 0.0004968809953189206, 'samples': 2261504, 'steps': 4416, 'loss/train': 1.7056611776351929} 02/24/2022 07:17:24 - INFO - codeparrot_training - Step 4417: {'lr': 0.0004968784182148276, 'samples': 2262016, 'steps': 4417, 'loss/train': 2.899319648742676} 02/24/2022 07:17:27 - INFO - codeparrot_training - Step 4418: {'lr': 0.0004968758400531859, 'samples': 2262528, 'steps': 4418, 'loss/train': 1.9671186208724976} 02/24/2022 07:17:32 - INFO - codeparrot_training - Step 4419: {'lr': 0.0004968732608340064, 'samples': 2263040, 'steps': 4419, 'loss/train': 2.4352123737335205} 02/24/2022 07:17:36 - INFO - codeparrot_training - Step 4420: {'lr': 0.0004968706805573002, 'samples': 2263552, 'steps': 4420, 'loss/train': 3.3830208778381348} 02/24/2022 07:17:42 - INFO - codeparrot_training - Step 4421: {'lr': 0.0004968680992230785, 'samples': 2264064, 'steps': 4421, 'loss/train': 2.5741186141967773} 02/24/2022 07:17:45 - INFO - codeparrot_training - Step 4422: {'lr': 0.0004968655168313522, 'samples': 2264576, 'steps': 4422, 'loss/train': 2.9256064891815186} 02/24/2022 07:17:51 - INFO - codeparrot_training - Step 4423: {'lr': 0.0004968629333821324, 'samples': 2265088, 'steps': 4423, 'loss/train': 2.8239471912384033} 02/24/2022 07:17:54 - INFO - codeparrot_training - Step 4424: {'lr': 0.0004968603488754302, 'samples': 2265600, 'steps': 4424, 'loss/train': 3.7768757343292236} 02/24/2022 07:18:01 - INFO - codeparrot_training - Step 4425: {'lr': 0.0004968577633112566, 'samples': 2266112, 'steps': 4425, 'loss/train': 2.7473161220550537} 02/24/2022 07:18:05 - INFO - codeparrot_training - Step 4426: {'lr': 0.0004968551766896228, 'samples': 2266624, 'steps': 4426, 'loss/train': 3.1793689727783203} 02/24/2022 07:18:10 - INFO - codeparrot_training - Step 4427: {'lr': 0.0004968525890105399, 'samples': 2267136, 'steps': 4427, 'loss/train': 2.7369556427001953} 02/24/2022 07:18:16 - INFO - codeparrot_training - Step 4428: {'lr': 0.0004968500002740187, 'samples': 2267648, 'steps': 4428, 'loss/train': 3.915787696838379} 02/24/2022 07:18:19 - INFO - codeparrot_training - Step 4429: {'lr': 0.0004968474104800706, 'samples': 2268160, 'steps': 4429, 'loss/train': 2.2864506244659424} 02/24/2022 07:18:25 - INFO - codeparrot_training - Step 4430: {'lr': 0.0004968448196287066, 'samples': 2268672, 'steps': 4430, 'loss/train': 2.4751694202423096} 02/24/2022 07:18:28 - INFO - codeparrot_training - Step 4431: {'lr': 0.0004968422277199377, 'samples': 2269184, 'steps': 4431, 'loss/train': 0.6428107023239136} 02/24/2022 07:18:34 - INFO - codeparrot_training - Step 4432: {'lr': 0.000496839634753775, 'samples': 2269696, 'steps': 4432, 'loss/train': 3.907973051071167} 02/24/2022 07:18:37 - INFO - codeparrot_training - Step 4433: {'lr': 0.0004968370407302299, 'samples': 2270208, 'steps': 4433, 'loss/train': 2.3889451026916504} 02/24/2022 07:18:45 - INFO - codeparrot_training - Step 4434: {'lr': 0.0004968344456493132, 'samples': 2270720, 'steps': 4434, 'loss/train': 8.024490356445312} 02/24/2022 07:18:48 - INFO - codeparrot_training - Step 4435: {'lr': 0.000496831849511036, 'samples': 2271232, 'steps': 4435, 'loss/train': 3.2024617195129395} 02/24/2022 07:18:54 - INFO - codeparrot_training - Step 4436: {'lr': 0.0004968292523154096, 'samples': 2271744, 'steps': 4436, 'loss/train': 3.395526885986328} 02/24/2022 07:18:57 - INFO - codeparrot_training - Step 4437: {'lr': 0.0004968266540624452, 'samples': 2272256, 'steps': 4437, 'loss/train': 3.5550262928009033} 02/24/2022 07:19:01 - INFO - codeparrot_training - Step 4438: {'lr': 0.0004968240547521536, 'samples': 2272768, 'steps': 4438, 'loss/train': 4.299802780151367} 02/24/2022 07:19:06 - INFO - codeparrot_training - Step 4439: {'lr': 0.0004968214543845463, 'samples': 2273280, 'steps': 4439, 'loss/train': 2.9327893257141113} 02/24/2022 07:19:10 - INFO - codeparrot_training - Step 4440: {'lr': 0.0004968188529596341, 'samples': 2273792, 'steps': 4440, 'loss/train': 2.998858690261841} 02/24/2022 07:19:15 - INFO - codeparrot_training - Step 4441: {'lr': 0.0004968162504774284, 'samples': 2274304, 'steps': 4441, 'loss/train': 4.053661346435547} 02/24/2022 07:19:19 - INFO - codeparrot_training - Step 4442: {'lr': 0.0004968136469379403, 'samples': 2274816, 'steps': 4442, 'loss/train': 3.4228515625} 02/24/2022 07:19:24 - INFO - codeparrot_training - Step 4443: {'lr': 0.0004968110423411808, 'samples': 2275328, 'steps': 4443, 'loss/train': 2.4554667472839355} 02/24/2022 07:19:28 - INFO - codeparrot_training - Step 4444: {'lr': 0.0004968084366871612, 'samples': 2275840, 'steps': 4444, 'loss/train': 2.481663942337036} 02/24/2022 07:19:34 - INFO - codeparrot_training - Step 4445: {'lr': 0.0004968058299758926, 'samples': 2276352, 'steps': 4445, 'loss/train': 2.463561773300171} 02/24/2022 07:19:37 - INFO - codeparrot_training - Step 4446: {'lr': 0.0004968032222073863, 'samples': 2276864, 'steps': 4446, 'loss/train': 2.4199366569519043} 02/24/2022 07:19:44 - INFO - codeparrot_training - Step 4447: {'lr': 0.0004968006133816532, 'samples': 2277376, 'steps': 4447, 'loss/train': 2.901564359664917} 02/24/2022 07:19:48 - INFO - codeparrot_training - Step 4448: {'lr': 0.0004967980034987048, 'samples': 2277888, 'steps': 4448, 'loss/train': 2.1362946033477783} 02/24/2022 07:19:54 - INFO - codeparrot_training - Step 4449: {'lr': 0.0004967953925585521, 'samples': 2278400, 'steps': 4449, 'loss/train': 3.142230987548828} 02/24/2022 07:19:57 - INFO - codeparrot_training - Step 4450: {'lr': 0.0004967927805612063, 'samples': 2278912, 'steps': 4450, 'loss/train': 3.0722169876098633} 02/24/2022 07:20:03 - INFO - codeparrot_training - Step 4451: {'lr': 0.0004967901675066784, 'samples': 2279424, 'steps': 4451, 'loss/train': 3.438680410385132} 02/24/2022 07:20:06 - INFO - codeparrot_training - Step 4452: {'lr': 0.0004967875533949801, 'samples': 2279936, 'steps': 4452, 'loss/train': 1.9792863130569458} 02/24/2022 07:20:12 - INFO - codeparrot_training - Step 4453: {'lr': 0.000496784938226122, 'samples': 2280448, 'steps': 4453, 'loss/train': 1.7001590728759766} 02/24/2022 07:20:15 - INFO - codeparrot_training - Step 4454: {'lr': 0.0004967823220001158, 'samples': 2280960, 'steps': 4454, 'loss/train': 1.9118536710739136} 02/24/2022 07:20:20 - INFO - codeparrot_training - Step 4455: {'lr': 0.0004967797047169724, 'samples': 2281472, 'steps': 4455, 'loss/train': 0.26020562648773193} 02/24/2022 07:20:24 - INFO - codeparrot_training - Step 4456: {'lr': 0.0004967770863767031, 'samples': 2281984, 'steps': 4456, 'loss/train': 2.5394480228424072} 02/24/2022 07:20:29 - INFO - codeparrot_training - Step 4457: {'lr': 0.0004967744669793192, 'samples': 2282496, 'steps': 4457, 'loss/train': 3.2624049186706543} 02/24/2022 07:20:33 - INFO - codeparrot_training - Step 4458: {'lr': 0.0004967718465248317, 'samples': 2283008, 'steps': 4458, 'loss/train': 2.592275857925415} 02/24/2022 07:20:38 - INFO - codeparrot_training - Step 4459: {'lr': 0.000496769225013252, 'samples': 2283520, 'steps': 4459, 'loss/train': 2.834155797958374} 02/24/2022 07:20:42 - INFO - codeparrot_training - Step 4460: {'lr': 0.0004967666024445913, 'samples': 2284032, 'steps': 4460, 'loss/train': 3.2586326599121094} 02/24/2022 07:20:47 - INFO - codeparrot_training - Step 4461: {'lr': 0.000496763978818861, 'samples': 2284544, 'steps': 4461, 'loss/train': 2.9925897121429443} 02/24/2022 07:20:51 - INFO - codeparrot_training - Step 4462: {'lr': 0.000496761354136072, 'samples': 2285056, 'steps': 4462, 'loss/train': 2.9752209186553955} 02/24/2022 07:20:58 - INFO - codeparrot_training - Step 4463: {'lr': 0.0004967587283962358, 'samples': 2285568, 'steps': 4463, 'loss/train': 2.2733354568481445} 02/24/2022 07:21:02 - INFO - codeparrot_training - Step 4464: {'lr': 0.0004967561015993635, 'samples': 2286080, 'steps': 4464, 'loss/train': 3.6289825439453125} 02/24/2022 07:21:07 - INFO - codeparrot_training - Step 4465: {'lr': 0.0004967534737454665, 'samples': 2286592, 'steps': 4465, 'loss/train': 2.0765228271484375} 02/24/2022 07:21:11 - INFO - codeparrot_training - Step 4466: {'lr': 0.000496750844834556, 'samples': 2287104, 'steps': 4466, 'loss/train': 3.049776554107666} 02/24/2022 07:21:16 - INFO - codeparrot_training - Step 4467: {'lr': 0.000496748214866643, 'samples': 2287616, 'steps': 4467, 'loss/train': 0.45040270686149597} 02/24/2022 07:21:20 - INFO - codeparrot_training - Step 4468: {'lr': 0.0004967455838417392, 'samples': 2288128, 'steps': 4468, 'loss/train': 2.2237250804901123} 02/24/2022 07:21:25 - INFO - codeparrot_training - Step 4469: {'lr': 0.0004967429517598556, 'samples': 2288640, 'steps': 4469, 'loss/train': 2.066601037979126} 02/24/2022 07:21:29 - INFO - codeparrot_training - Step 4470: {'lr': 0.0004967403186210036, 'samples': 2289152, 'steps': 4470, 'loss/train': 3.095932960510254} 02/24/2022 07:21:34 - INFO - codeparrot_training - Step 4471: {'lr': 0.0004967376844251944, 'samples': 2289664, 'steps': 4471, 'loss/train': 5.598101615905762} 02/24/2022 07:21:38 - INFO - codeparrot_training - Step 4472: {'lr': 0.0004967350491724392, 'samples': 2290176, 'steps': 4472, 'loss/train': 2.4305169582366943} 02/24/2022 07:21:45 - INFO - codeparrot_training - Step 4473: {'lr': 0.0004967324128627495, 'samples': 2290688, 'steps': 4473, 'loss/train': 3.0877225399017334} 02/24/2022 07:21:48 - INFO - codeparrot_training - Step 4474: {'lr': 0.0004967297754961365, 'samples': 2291200, 'steps': 4474, 'loss/train': 1.8233672380447388} 02/24/2022 07:21:54 - INFO - codeparrot_training - Step 4475: {'lr': 0.0004967271370726115, 'samples': 2291712, 'steps': 4475, 'loss/train': 2.399275302886963} 02/24/2022 07:21:57 - INFO - codeparrot_training - Step 4476: {'lr': 0.0004967244975921857, 'samples': 2292224, 'steps': 4476, 'loss/train': 3.2833547592163086} 02/24/2022 07:22:03 - INFO - codeparrot_training - Step 4477: {'lr': 0.0004967218570548706, 'samples': 2292736, 'steps': 4477, 'loss/train': 2.1094300746917725} 02/24/2022 07:22:06 - INFO - codeparrot_training - Step 4478: {'lr': 0.0004967192154606774, 'samples': 2293248, 'steps': 4478, 'loss/train': 2.0430593490600586} 02/24/2022 07:22:12 - INFO - codeparrot_training - Step 4479: {'lr': 0.0004967165728096172, 'samples': 2293760, 'steps': 4479, 'loss/train': 3.1830222606658936} 02/24/2022 07:22:15 - INFO - codeparrot_training - Step 4480: {'lr': 0.0004967139291017018, 'samples': 2294272, 'steps': 4480, 'loss/train': 1.9834043979644775} 02/24/2022 07:22:21 - INFO - codeparrot_training - Step 4481: {'lr': 0.0004967112843369423, 'samples': 2294784, 'steps': 4481, 'loss/train': 3.3287105560302734} 02/24/2022 07:22:24 - INFO - codeparrot_training - Step 4482: {'lr': 0.0004967086385153499, 'samples': 2295296, 'steps': 4482, 'loss/train': 2.6500513553619385} 02/24/2022 07:22:30 - INFO - codeparrot_training - Step 4483: {'lr': 0.0004967059916369359, 'samples': 2295808, 'steps': 4483, 'loss/train': 3.129227876663208} 02/24/2022 07:22:34 - INFO - codeparrot_training - Step 4484: {'lr': 0.000496703343701712, 'samples': 2296320, 'steps': 4484, 'loss/train': 2.439460277557373} 02/24/2022 07:22:39 - INFO - codeparrot_training - Step 4485: {'lr': 0.0004967006947096892, 'samples': 2296832, 'steps': 4485, 'loss/train': 2.4997289180755615} 02/24/2022 07:22:43 - INFO - codeparrot_training - Step 4486: {'lr': 0.0004966980446608789, 'samples': 2297344, 'steps': 4486, 'loss/train': 3.0099687576293945} 02/24/2022 07:22:48 - INFO - codeparrot_training - Step 4487: {'lr': 0.0004966953935552925, 'samples': 2297856, 'steps': 4487, 'loss/train': 2.946878433227539} 02/24/2022 07:22:52 - INFO - codeparrot_training - Step 4488: {'lr': 0.0004966927413929415, 'samples': 2298368, 'steps': 4488, 'loss/train': 2.7749521732330322} 02/24/2022 07:22:57 - INFO - codeparrot_training - Step 4489: {'lr': 0.0004966900881738371, 'samples': 2298880, 'steps': 4489, 'loss/train': 2.306493043899536} 02/24/2022 07:23:01 - INFO - codeparrot_training - Step 4490: {'lr': 0.0004966874338979907, 'samples': 2299392, 'steps': 4490, 'loss/train': 3.477994441986084} 02/24/2022 07:23:06 - INFO - codeparrot_training - Step 4491: {'lr': 0.0004966847785654136, 'samples': 2299904, 'steps': 4491, 'loss/train': 3.2932941913604736} 02/24/2022 07:23:10 - INFO - codeparrot_training - Step 4492: {'lr': 0.0004966821221761173, 'samples': 2300416, 'steps': 4492, 'loss/train': 2.67145037651062} 02/24/2022 07:23:16 - INFO - codeparrot_training - Step 4493: {'lr': 0.0004966794647301131, 'samples': 2300928, 'steps': 4493, 'loss/train': 2.603548765182495} 02/24/2022 07:23:20 - INFO - codeparrot_training - Step 4494: {'lr': 0.0004966768062274125, 'samples': 2301440, 'steps': 4494, 'loss/train': 7.907590866088867} 02/24/2022 07:23:25 - INFO - codeparrot_training - Step 4495: {'lr': 0.0004966741466680266, 'samples': 2301952, 'steps': 4495, 'loss/train': 2.716430902481079} 02/24/2022 07:23:28 - INFO - codeparrot_training - Step 4496: {'lr': 0.000496671486051967, 'samples': 2302464, 'steps': 4496, 'loss/train': 3.171616315841675} 02/24/2022 07:23:34 - INFO - codeparrot_training - Step 4497: {'lr': 0.0004966688243792452, 'samples': 2302976, 'steps': 4497, 'loss/train': 2.0503103733062744} 02/24/2022 07:23:37 - INFO - codeparrot_training - Step 4498: {'lr': 0.0004966661616498724, 'samples': 2303488, 'steps': 4498, 'loss/train': 2.387420654296875} 02/24/2022 07:23:43 - INFO - codeparrot_training - Step 4499: {'lr': 0.0004966634978638601, 'samples': 2304000, 'steps': 4499, 'loss/train': 2.9485855102539062} 02/24/2022 07:23:46 - INFO - codeparrot_training - Step 4500: {'lr': 0.0004966608330212198, 'samples': 2304512, 'steps': 4500, 'loss/train': 2.670931816101074} 02/24/2022 07:23:52 - INFO - codeparrot_training - Step 4501: {'lr': 0.0004966581671219627, 'samples': 2305024, 'steps': 4501, 'loss/train': 3.4441380500793457} 02/24/2022 07:23:55 - INFO - codeparrot_training - Step 4502: {'lr': 0.0004966555001661004, 'samples': 2305536, 'steps': 4502, 'loss/train': 2.946645975112915} 02/24/2022 07:24:02 - INFO - codeparrot_training - Step 4503: {'lr': 0.0004966528321536442, 'samples': 2306048, 'steps': 4503, 'loss/train': 1.6140245199203491} 02/24/2022 07:24:05 - INFO - codeparrot_training - Step 4504: {'lr': 0.0004966501630846057, 'samples': 2306560, 'steps': 4504, 'loss/train': 7.584186553955078} 02/24/2022 07:24:10 - INFO - codeparrot_training - Step 4505: {'lr': 0.000496647492958996, 'samples': 2307072, 'steps': 4505, 'loss/train': 2.899822950363159} 02/24/2022 07:24:16 - INFO - codeparrot_training - Step 4506: {'lr': 0.000496644821776827, 'samples': 2307584, 'steps': 4506, 'loss/train': 1.99951171875} 02/24/2022 07:24:19 - INFO - codeparrot_training - Step 4507: {'lr': 0.0004966421495381098, 'samples': 2308096, 'steps': 4507, 'loss/train': 2.502471923828125} 02/24/2022 07:24:25 - INFO - codeparrot_training - Step 4508: {'lr': 0.0004966394762428559, 'samples': 2308608, 'steps': 4508, 'loss/train': 1.67020845413208} 02/24/2022 07:24:29 - INFO - codeparrot_training - Step 4509: {'lr': 0.0004966368018910768, 'samples': 2309120, 'steps': 4509, 'loss/train': 3.101428747177124} 02/24/2022 07:24:34 - INFO - codeparrot_training - Step 4510: {'lr': 0.000496634126482784, 'samples': 2309632, 'steps': 4510, 'loss/train': 1.0595159530639648} 02/24/2022 07:24:38 - INFO - codeparrot_training - Step 4511: {'lr': 0.000496631450017989, 'samples': 2310144, 'steps': 4511, 'loss/train': 0.08816852420568466} 02/24/2022 07:24:43 - INFO - codeparrot_training - Step 4512: {'lr': 0.0004966287724967032, 'samples': 2310656, 'steps': 4512, 'loss/train': 1.6504418849945068} 02/24/2022 07:24:47 - INFO - codeparrot_training - Step 4513: {'lr': 0.0004966260939189379, 'samples': 2311168, 'steps': 4513, 'loss/train': 1.6694890260696411} 02/24/2022 07:24:52 - INFO - codeparrot_training - Step 4514: {'lr': 0.0004966234142847048, 'samples': 2311680, 'steps': 4514, 'loss/train': 2.1192007064819336} 02/24/2022 07:24:56 - INFO - codeparrot_training - Step 4515: {'lr': 0.0004966207335940153, 'samples': 2312192, 'steps': 4515, 'loss/train': 3.1064274311065674} 02/24/2022 07:25:01 - INFO - codeparrot_training - Step 4516: {'lr': 0.0004966180518468808, 'samples': 2312704, 'steps': 4516, 'loss/train': 2.8479433059692383} 02/24/2022 07:25:05 - INFO - codeparrot_training - Step 4517: {'lr': 0.000496615369043313, 'samples': 2313216, 'steps': 4517, 'loss/train': 2.3023681640625} 02/24/2022 07:25:11 - INFO - codeparrot_training - Step 4518: {'lr': 0.0004966126851833233, 'samples': 2313728, 'steps': 4518, 'loss/train': 3.0059397220611572} 02/24/2022 07:25:14 - INFO - codeparrot_training - Step 4519: {'lr': 0.0004966100002669231, 'samples': 2314240, 'steps': 4519, 'loss/train': 1.8635908365249634} 02/24/2022 07:25:20 - INFO - codeparrot_training - Step 4520: {'lr': 0.0004966073142941239, 'samples': 2314752, 'steps': 4520, 'loss/train': 3.673736095428467} 02/24/2022 07:25:23 - INFO - codeparrot_training - Step 4521: {'lr': 0.0004966046272649372, 'samples': 2315264, 'steps': 4521, 'loss/train': 2.7314493656158447} 02/24/2022 07:25:29 - INFO - codeparrot_training - Step 4522: {'lr': 0.0004966019391793748, 'samples': 2315776, 'steps': 4522, 'loss/train': 1.319679856300354} 02/24/2022 07:25:32 - INFO - codeparrot_training - Step 4523: {'lr': 0.0004965992500374479, 'samples': 2316288, 'steps': 4523, 'loss/train': 3.444967031478882} 02/24/2022 07:25:38 - INFO - codeparrot_training - Step 4524: {'lr': 0.0004965965598391682, 'samples': 2316800, 'steps': 4524, 'loss/train': 3.1193411350250244} 02/24/2022 07:25:41 - INFO - codeparrot_training - Step 4525: {'lr': 0.000496593868584547, 'samples': 2317312, 'steps': 4525, 'loss/train': 2.2373745441436768} 02/24/2022 07:25:47 - INFO - codeparrot_training - Step 4526: {'lr': 0.0004965911762735961, 'samples': 2317824, 'steps': 4526, 'loss/train': 3.0222339630126953} 02/24/2022 07:25:51 - INFO - codeparrot_training - Step 4527: {'lr': 0.0004965884829063268, 'samples': 2318336, 'steps': 4527, 'loss/train': 2.2678277492523193} 02/24/2022 07:25:57 - INFO - codeparrot_training - Step 4528: {'lr': 0.0004965857884827508, 'samples': 2318848, 'steps': 4528, 'loss/train': 1.9348386526107788} 02/24/2022 07:26:00 - INFO - codeparrot_training - Step 4529: {'lr': 0.0004965830930028795, 'samples': 2319360, 'steps': 4529, 'loss/train': 3.5817720890045166} 02/24/2022 07:26:06 - INFO - codeparrot_training - Step 4530: {'lr': 0.0004965803964667246, 'samples': 2319872, 'steps': 4530, 'loss/train': 2.1504907608032227} 02/24/2022 07:26:09 - INFO - codeparrot_training - Step 4531: {'lr': 0.0004965776988742976, 'samples': 2320384, 'steps': 4531, 'loss/train': 2.881620168685913} 02/24/2022 07:26:15 - INFO - codeparrot_training - Step 4532: {'lr': 0.00049657500022561, 'samples': 2320896, 'steps': 4532, 'loss/train': 2.845905303955078} 02/24/2022 07:26:18 - INFO - codeparrot_training - Step 4533: {'lr': 0.0004965723005206734, 'samples': 2321408, 'steps': 4533, 'loss/train': 3.8281946182250977} 02/24/2022 07:26:24 - INFO - codeparrot_training - Step 4534: {'lr': 0.0004965695997594993, 'samples': 2321920, 'steps': 4534, 'loss/train': 3.05188250541687} 02/24/2022 07:26:27 - INFO - codeparrot_training - Step 4535: {'lr': 0.0004965668979420994, 'samples': 2322432, 'steps': 4535, 'loss/train': 1.9769772291183472} 02/24/2022 07:26:33 - INFO - codeparrot_training - Step 4536: {'lr': 0.0004965641950684852, 'samples': 2322944, 'steps': 4536, 'loss/train': 2.2932844161987305} 02/24/2022 07:26:36 - INFO - codeparrot_training - Step 4537: {'lr': 0.0004965614911386683, 'samples': 2323456, 'steps': 4537, 'loss/train': 2.8208165168762207} 02/24/2022 07:26:42 - INFO - codeparrot_training - Step 4538: {'lr': 0.0004965587861526602, 'samples': 2323968, 'steps': 4538, 'loss/train': 2.6197359561920166} 02/24/2022 07:26:45 - INFO - codeparrot_training - Step 4539: {'lr': 0.0004965560801104726, 'samples': 2324480, 'steps': 4539, 'loss/train': 2.682002067565918} 02/24/2022 07:26:51 - INFO - codeparrot_training - Step 4540: {'lr': 0.000496553373012117, 'samples': 2324992, 'steps': 4540, 'loss/train': 2.820742130279541} 02/24/2022 07:26:54 - INFO - codeparrot_training - Step 4541: {'lr': 0.0004965506648576052, 'samples': 2325504, 'steps': 4541, 'loss/train': 0.8161208629608154} 02/24/2022 07:27:00 - INFO - codeparrot_training - Step 4542: {'lr': 0.0004965479556469485, 'samples': 2326016, 'steps': 4542, 'loss/train': 2.9415500164031982} 02/24/2022 07:27:03 - INFO - codeparrot_training - Step 4543: {'lr': 0.0004965452453801586, 'samples': 2326528, 'steps': 4543, 'loss/train': 2.601675271987915} 02/24/2022 07:27:09 - INFO - codeparrot_training - Step 4544: {'lr': 0.0004965425340572472, 'samples': 2327040, 'steps': 4544, 'loss/train': 2.377925157546997} 02/24/2022 07:27:12 - INFO - codeparrot_training - Step 4545: {'lr': 0.0004965398216782258, 'samples': 2327552, 'steps': 4545, 'loss/train': 2.7842869758605957} 02/24/2022 07:27:18 - INFO - codeparrot_training - Step 4546: {'lr': 0.0004965371082431062, 'samples': 2328064, 'steps': 4546, 'loss/train': 3.3167741298675537} 02/24/2022 07:27:21 - INFO - codeparrot_training - Step 4547: {'lr': 0.0004965343937519, 'samples': 2328576, 'steps': 4547, 'loss/train': 2.836545705795288} 02/24/2022 07:27:28 - INFO - codeparrot_training - Step 4548: {'lr': 0.0004965316782046186, 'samples': 2329088, 'steps': 4548, 'loss/train': 2.41920804977417} 02/24/2022 07:27:31 - INFO - codeparrot_training - Step 4549: {'lr': 0.0004965289616012739, 'samples': 2329600, 'steps': 4549, 'loss/train': 2.9270546436309814} 02/24/2022 07:27:37 - INFO - codeparrot_training - Step 4550: {'lr': 0.0004965262439418772, 'samples': 2330112, 'steps': 4550, 'loss/train': 1.860337495803833} 02/24/2022 07:27:40 - INFO - codeparrot_training - Step 4551: {'lr': 0.0004965235252264405, 'samples': 2330624, 'steps': 4551, 'loss/train': 2.1710102558135986} 02/24/2022 07:27:46 - INFO - codeparrot_training - Step 4552: {'lr': 0.0004965208054549753, 'samples': 2331136, 'steps': 4552, 'loss/train': 3.1194827556610107} 02/24/2022 07:27:49 - INFO - codeparrot_training - Step 4553: {'lr': 0.0004965180846274931, 'samples': 2331648, 'steps': 4553, 'loss/train': 2.0802981853485107} 02/24/2022 07:27:55 - INFO - codeparrot_training - Step 4554: {'lr': 0.0004965153627440058, 'samples': 2332160, 'steps': 4554, 'loss/train': 1.9774519205093384} 02/24/2022 07:27:59 - INFO - codeparrot_training - Step 4555: {'lr': 0.000496512639804525, 'samples': 2332672, 'steps': 4555, 'loss/train': 8.99859619140625} 02/24/2022 07:28:02 - INFO - codeparrot_training - Step 4556: {'lr': 0.0004965099158090624, 'samples': 2333184, 'steps': 4556, 'loss/train': 1.8740646839141846} 02/24/2022 07:28:07 - INFO - codeparrot_training - Step 4557: {'lr': 0.0004965071907576294, 'samples': 2333696, 'steps': 4557, 'loss/train': 3.6951892375946045} 02/24/2022 07:28:13 - INFO - codeparrot_training - Step 4558: {'lr': 0.000496504464650238, 'samples': 2334208, 'steps': 4558, 'loss/train': 2.408580780029297} 02/24/2022 07:28:16 - INFO - codeparrot_training - Step 4559: {'lr': 0.0004965017374868997, 'samples': 2334720, 'steps': 4559, 'loss/train': 2.156168222427368} 02/24/2022 07:28:22 - INFO - codeparrot_training - Step 4560: {'lr': 0.0004964990092676262, 'samples': 2335232, 'steps': 4560, 'loss/train': 2.1013872623443604} 02/24/2022 07:28:25 - INFO - codeparrot_training - Step 4561: {'lr': 0.0004964962799924293, 'samples': 2335744, 'steps': 4561, 'loss/train': 3.039947748184204} 02/24/2022 07:28:31 - INFO - codeparrot_training - Step 4562: {'lr': 0.0004964935496613206, 'samples': 2336256, 'steps': 4562, 'loss/train': 2.92911696434021} 02/24/2022 07:28:34 - INFO - codeparrot_training - Step 4563: {'lr': 0.0004964908182743117, 'samples': 2336768, 'steps': 4563, 'loss/train': 2.880499839782715} 02/24/2022 07:28:40 - INFO - codeparrot_training - Step 4564: {'lr': 0.0004964880858314146, 'samples': 2337280, 'steps': 4564, 'loss/train': 1.8258914947509766} 02/24/2022 07:28:44 - INFO - codeparrot_training - Step 4565: {'lr': 0.0004964853523326406, 'samples': 2337792, 'steps': 4565, 'loss/train': 2.7664577960968018} 02/24/2022 07:28:50 - INFO - codeparrot_training - Step 4566: {'lr': 0.0004964826177780017, 'samples': 2338304, 'steps': 4566, 'loss/train': 1.474591612815857} 02/24/2022 07:28:53 - INFO - codeparrot_training - Step 4567: {'lr': 0.0004964798821675096, 'samples': 2338816, 'steps': 4567, 'loss/train': 2.8581604957580566} 02/24/2022 07:28:58 - INFO - codeparrot_training - Step 4568: {'lr': 0.0004964771455011758, 'samples': 2339328, 'steps': 4568, 'loss/train': 2.927184581756592} 02/24/2022 07:29:02 - INFO - codeparrot_training - Step 4569: {'lr': 0.0004964744077790123, 'samples': 2339840, 'steps': 4569, 'loss/train': 1.7048866748809814} 02/24/2022 07:29:08 - INFO - codeparrot_training - Step 4570: {'lr': 0.0004964716690010306, 'samples': 2340352, 'steps': 4570, 'loss/train': 2.103135108947754} 02/24/2022 07:29:11 - INFO - codeparrot_training - Step 4571: {'lr': 0.0004964689291672427, 'samples': 2340864, 'steps': 4571, 'loss/train': 2.819431781768799} 02/24/2022 07:29:16 - INFO - codeparrot_training - Step 4572: {'lr': 0.00049646618827766, 'samples': 2341376, 'steps': 4572, 'loss/train': 2.6605138778686523} 02/24/2022 07:29:20 - INFO - codeparrot_training - Step 4573: {'lr': 0.0004964634463322945, 'samples': 2341888, 'steps': 4573, 'loss/train': 1.7473955154418945} 02/24/2022 07:29:25 - INFO - codeparrot_training - Step 4574: {'lr': 0.0004964607033311579, 'samples': 2342400, 'steps': 4574, 'loss/train': 3.4722418785095215} 02/24/2022 07:29:29 - INFO - codeparrot_training - Step 4575: {'lr': 0.0004964579592742618, 'samples': 2342912, 'steps': 4575, 'loss/train': 3.3197989463806152} 02/24/2022 07:29:35 - INFO - codeparrot_training - Step 4576: {'lr': 0.000496455214161618, 'samples': 2343424, 'steps': 4576, 'loss/train': 1.1324106454849243} 02/24/2022 07:29:39 - INFO - codeparrot_training - Step 4577: {'lr': 0.0004964524679932385, 'samples': 2343936, 'steps': 4577, 'loss/train': 2.899233102798462} 02/24/2022 07:29:44 - INFO - codeparrot_training - Step 4578: {'lr': 0.0004964497207691349, 'samples': 2344448, 'steps': 4578, 'loss/train': 2.8270816802978516} 02/24/2022 07:29:48 - INFO - codeparrot_training - Step 4579: {'lr': 0.0004964469724893188, 'samples': 2344960, 'steps': 4579, 'loss/train': 2.535975217819214} 02/24/2022 07:29:53 - INFO - codeparrot_training - Step 4580: {'lr': 0.0004964442231538023, 'samples': 2345472, 'steps': 4580, 'loss/train': 2.6434524059295654} 02/24/2022 07:29:57 - INFO - codeparrot_training - Step 4581: {'lr': 0.0004964414727625968, 'samples': 2345984, 'steps': 4581, 'loss/train': 2.275078535079956} 02/24/2022 07:30:02 - INFO - codeparrot_training - Step 4582: {'lr': 0.0004964387213157143, 'samples': 2346496, 'steps': 4582, 'loss/train': 2.6823956966400146} 02/24/2022 07:30:06 - INFO - codeparrot_training - Step 4583: {'lr': 0.0004964359688131667, 'samples': 2347008, 'steps': 4583, 'loss/train': 1.7113174200057983} 02/24/2022 07:30:11 - INFO - codeparrot_training - Step 4584: {'lr': 0.0004964332152549657, 'samples': 2347520, 'steps': 4584, 'loss/train': 2.531528949737549} 02/24/2022 07:30:15 - INFO - codeparrot_training - Step 4585: {'lr': 0.0004964304606411229, 'samples': 2348032, 'steps': 4585, 'loss/train': 2.952152967453003} 02/24/2022 07:30:21 - INFO - codeparrot_training - Step 4586: {'lr': 0.0004964277049716503, 'samples': 2348544, 'steps': 4586, 'loss/train': 2.2497806549072266} 02/24/2022 07:30:24 - INFO - codeparrot_training - Step 4587: {'lr': 0.0004964249482465597, 'samples': 2349056, 'steps': 4587, 'loss/train': 3.607583522796631} 02/24/2022 07:30:30 - INFO - codeparrot_training - Step 4588: {'lr': 0.0004964221904658629, 'samples': 2349568, 'steps': 4588, 'loss/train': 3.379906177520752} 02/24/2022 07:30:33 - INFO - codeparrot_training - Step 4589: {'lr': 0.0004964194316295716, 'samples': 2350080, 'steps': 4589, 'loss/train': 2.5294156074523926} 02/24/2022 07:30:39 - INFO - codeparrot_training - Step 4590: {'lr': 0.0004964166717376978, 'samples': 2350592, 'steps': 4590, 'loss/train': 1.7820571660995483} 02/24/2022 07:30:42 - INFO - codeparrot_training - Step 4591: {'lr': 0.0004964139107902531, 'samples': 2351104, 'steps': 4591, 'loss/train': 3.2617321014404297} 02/24/2022 07:30:48 - INFO - codeparrot_training - Step 4592: {'lr': 0.0004964111487872495, 'samples': 2351616, 'steps': 4592, 'loss/train': 2.0228817462921143} 02/24/2022 07:30:51 - INFO - codeparrot_training - Step 4593: {'lr': 0.0004964083857286988, 'samples': 2352128, 'steps': 4593, 'loss/train': 1.648618221282959} 02/24/2022 07:30:56 - INFO - codeparrot_training - Step 4594: {'lr': 0.0004964056216146129, 'samples': 2352640, 'steps': 4594, 'loss/train': 3.0626327991485596} 02/24/2022 07:31:00 - INFO - codeparrot_training - Step 4595: {'lr': 0.0004964028564450034, 'samples': 2353152, 'steps': 4595, 'loss/train': 2.530766248703003} 02/24/2022 07:31:06 - INFO - codeparrot_training - Step 4596: {'lr': 0.0004964000902198824, 'samples': 2353664, 'steps': 4596, 'loss/train': 1.1859678030014038} 02/24/2022 07:31:09 - INFO - codeparrot_training - Step 4597: {'lr': 0.0004963973229392617, 'samples': 2354176, 'steps': 4597, 'loss/train': 2.338047504425049} 02/24/2022 07:31:14 - INFO - codeparrot_training - Step 4598: {'lr': 0.0004963945546031529, 'samples': 2354688, 'steps': 4598, 'loss/train': 3.690566301345825} 02/24/2022 07:31:18 - INFO - codeparrot_training - Step 4599: {'lr': 0.0004963917852115683, 'samples': 2355200, 'steps': 4599, 'loss/train': 2.787959337234497} 02/24/2022 07:31:24 - INFO - codeparrot_training - Step 4600: {'lr': 0.0004963890147645194, 'samples': 2355712, 'steps': 4600, 'loss/train': 0.9729641079902649} 02/24/2022 07:31:27 - INFO - codeparrot_training - Step 4601: {'lr': 0.0004963862432620183, 'samples': 2356224, 'steps': 4601, 'loss/train': 1.9808101654052734} 02/24/2022 07:31:33 - INFO - codeparrot_training - Step 4602: {'lr': 0.0004963834707040767, 'samples': 2356736, 'steps': 4602, 'loss/train': 2.3603668212890625} 02/24/2022 07:31:36 - INFO - codeparrot_training - Step 4603: {'lr': 0.0004963806970907066, 'samples': 2357248, 'steps': 4603, 'loss/train': 2.557182788848877} 02/24/2022 07:31:42 - INFO - codeparrot_training - Step 4604: {'lr': 0.0004963779224219197, 'samples': 2357760, 'steps': 4604, 'loss/train': 8.10139274597168} 02/24/2022 07:31:45 - INFO - codeparrot_training - Step 4605: {'lr': 0.0004963751466977281, 'samples': 2358272, 'steps': 4605, 'loss/train': 2.8341751098632812} 02/24/2022 07:31:51 - INFO - codeparrot_training - Step 4606: {'lr': 0.0004963723699181437, 'samples': 2358784, 'steps': 4606, 'loss/train': 4.430930137634277} 02/24/2022 07:31:54 - INFO - codeparrot_training - Step 4607: {'lr': 0.0004963695920831781, 'samples': 2359296, 'steps': 4607, 'loss/train': 2.1612918376922607} 02/24/2022 07:32:00 - INFO - codeparrot_training - Step 4608: {'lr': 0.0004963668131928436, 'samples': 2359808, 'steps': 4608, 'loss/train': 2.2669763565063477} 02/24/2022 07:32:03 - INFO - codeparrot_training - Step 4609: {'lr': 0.0004963640332471518, 'samples': 2360320, 'steps': 4609, 'loss/train': 2.771120548248291} 02/24/2022 07:32:09 - INFO - codeparrot_training - Step 4610: {'lr': 0.0004963612522461147, 'samples': 2360832, 'steps': 4610, 'loss/train': 3.9943385124206543} 02/24/2022 07:32:13 - INFO - codeparrot_training - Step 4611: {'lr': 0.0004963584701897443, 'samples': 2361344, 'steps': 4611, 'loss/train': 2.767150402069092} 02/24/2022 07:32:19 - INFO - codeparrot_training - Step 4612: {'lr': 0.0004963556870780523, 'samples': 2361856, 'steps': 4612, 'loss/train': 2.933321475982666} 02/24/2022 07:32:22 - INFO - codeparrot_training - Step 4613: {'lr': 0.0004963529029110509, 'samples': 2362368, 'steps': 4613, 'loss/train': 3.2616207599639893} 02/24/2022 07:32:28 - INFO - codeparrot_training - Step 4614: {'lr': 0.0004963501176887519, 'samples': 2362880, 'steps': 4614, 'loss/train': 3.9431509971618652} 02/24/2022 07:32:31 - INFO - codeparrot_training - Step 4615: {'lr': 0.000496347331411167, 'samples': 2363392, 'steps': 4615, 'loss/train': 1.420654535293579} 02/24/2022 07:32:37 - INFO - codeparrot_training - Step 4616: {'lr': 0.0004963445440783086, 'samples': 2363904, 'steps': 4616, 'loss/train': 3.3908069133758545} 02/24/2022 07:32:40 - INFO - codeparrot_training - Step 4617: {'lr': 0.0004963417556901882, 'samples': 2364416, 'steps': 4617, 'loss/train': 2.4702248573303223} 02/24/2022 07:32:46 - INFO - codeparrot_training - Step 4618: {'lr': 0.0004963389662468182, 'samples': 2364928, 'steps': 4618, 'loss/train': 2.7041637897491455} 02/24/2022 07:32:49 - INFO - codeparrot_training - Step 4619: {'lr': 0.0004963361757482101, 'samples': 2365440, 'steps': 4619, 'loss/train': 1.2446062564849854} 02/24/2022 07:32:55 - INFO - codeparrot_training - Step 4620: {'lr': 0.000496333384194376, 'samples': 2365952, 'steps': 4620, 'loss/train': 3.6648197174072266} 02/24/2022 07:32:58 - INFO - codeparrot_training - Step 4621: {'lr': 0.000496330591585328, 'samples': 2366464, 'steps': 4621, 'loss/train': 2.9328489303588867} 02/24/2022 07:33:05 - INFO - codeparrot_training - Step 4622: {'lr': 0.0004963277979210779, 'samples': 2366976, 'steps': 4622, 'loss/train': 2.2918753623962402} 02/24/2022 07:33:08 - INFO - codeparrot_training - Step 4623: {'lr': 0.0004963250032016379, 'samples': 2367488, 'steps': 4623, 'loss/train': 2.744584083557129} 02/24/2022 07:33:14 - INFO - codeparrot_training - Step 4624: {'lr': 0.0004963222074270197, 'samples': 2368000, 'steps': 4624, 'loss/train': 1.826094627380371} 02/24/2022 07:33:19 - INFO - codeparrot_training - Step 4625: {'lr': 0.0004963194105972353, 'samples': 2368512, 'steps': 4625, 'loss/train': 2.5796639919281006} 02/24/2022 07:33:22 - INFO - codeparrot_training - Step 4626: {'lr': 0.0004963166127122969, 'samples': 2369024, 'steps': 4626, 'loss/train': 2.089270830154419} 02/24/2022 07:33:28 - INFO - codeparrot_training - Step 4627: {'lr': 0.0004963138137722161, 'samples': 2369536, 'steps': 4627, 'loss/train': 2.129268169403076} 02/24/2022 07:33:31 - INFO - codeparrot_training - Step 4628: {'lr': 0.0004963110137770054, 'samples': 2370048, 'steps': 4628, 'loss/train': 2.4956653118133545} 02/24/2022 07:33:37 - INFO - codeparrot_training - Step 4629: {'lr': 0.0004963082127266764, 'samples': 2370560, 'steps': 4629, 'loss/train': 1.6712114810943604} 02/24/2022 07:33:40 - INFO - codeparrot_training - Step 4630: {'lr': 0.0004963054106212414, 'samples': 2371072, 'steps': 4630, 'loss/train': 3.452554225921631} 02/24/2022 07:33:47 - INFO - codeparrot_training - Step 4631: {'lr': 0.000496302607460712, 'samples': 2371584, 'steps': 4631, 'loss/train': 3.349071502685547} 02/24/2022 07:33:51 - INFO - codeparrot_training - Step 4632: {'lr': 0.0004962998032451005, 'samples': 2372096, 'steps': 4632, 'loss/train': 1.8830987215042114} 02/24/2022 07:33:56 - INFO - codeparrot_training - Step 4633: {'lr': 0.0004962969979744189, 'samples': 2372608, 'steps': 4633, 'loss/train': 1.2372952699661255} 02/24/2022 07:34:00 - INFO - codeparrot_training - Step 4634: {'lr': 0.0004962941916486791, 'samples': 2373120, 'steps': 4634, 'loss/train': 1.7662454843521118} 02/24/2022 07:34:05 - INFO - codeparrot_training - Step 4635: {'lr': 0.0004962913842678934, 'samples': 2373632, 'steps': 4635, 'loss/train': 2.36586332321167} 02/24/2022 07:34:09 - INFO - codeparrot_training - Step 4636: {'lr': 0.0004962885758320734, 'samples': 2374144, 'steps': 4636, 'loss/train': 6.441001892089844} 02/24/2022 07:34:14 - INFO - codeparrot_training - Step 4637: {'lr': 0.0004962857663412314, 'samples': 2374656, 'steps': 4637, 'loss/train': 3.0188519954681396} 02/24/2022 07:34:17 - INFO - codeparrot_training - Step 4638: {'lr': 0.0004962829557953794, 'samples': 2375168, 'steps': 4638, 'loss/train': 2.7057178020477295} 02/24/2022 07:34:23 - INFO - codeparrot_training - Step 4639: {'lr': 0.0004962801441945293, 'samples': 2375680, 'steps': 4639, 'loss/train': 2.828700065612793} 02/24/2022 07:34:27 - INFO - codeparrot_training - Step 4640: {'lr': 0.0004962773315386935, 'samples': 2376192, 'steps': 4640, 'loss/train': 2.5407299995422363} 02/24/2022 07:34:33 - INFO - codeparrot_training - Step 4641: {'lr': 0.0004962745178278837, 'samples': 2376704, 'steps': 4641, 'loss/train': 0.6332015991210938} 02/24/2022 07:34:36 - INFO - codeparrot_training - Step 4642: {'lr': 0.000496271703062112, 'samples': 2377216, 'steps': 4642, 'loss/train': 3.1000168323516846} 02/24/2022 07:34:42 - INFO - codeparrot_training - Step 4643: {'lr': 0.0004962688872413906, 'samples': 2377728, 'steps': 4643, 'loss/train': 2.8571813106536865} 02/24/2022 07:34:45 - INFO - codeparrot_training - Step 4644: {'lr': 0.0004962660703657315, 'samples': 2378240, 'steps': 4644, 'loss/train': 2.7817742824554443} 02/24/2022 07:34:51 - INFO - codeparrot_training - Step 4645: {'lr': 0.0004962632524351467, 'samples': 2378752, 'steps': 4645, 'loss/train': 2.790558099746704} 02/24/2022 07:34:54 - INFO - codeparrot_training - Step 4646: {'lr': 0.0004962604334496483, 'samples': 2379264, 'steps': 4646, 'loss/train': 2.701981782913208} 02/24/2022 07:35:00 - INFO - codeparrot_training - Step 4647: {'lr': 0.0004962576134092485, 'samples': 2379776, 'steps': 4647, 'loss/train': 3.595661163330078} 02/24/2022 07:35:03 - INFO - codeparrot_training - Step 4648: {'lr': 0.0004962547923139592, 'samples': 2380288, 'steps': 4648, 'loss/train': 2.838186025619507} 02/24/2022 07:35:09 - INFO - codeparrot_training - Step 4649: {'lr': 0.0004962519701637926, 'samples': 2380800, 'steps': 4649, 'loss/train': 3.0065674781799316} 02/24/2022 07:35:12 - INFO - codeparrot_training - Step 4650: {'lr': 0.0004962491469587607, 'samples': 2381312, 'steps': 4650, 'loss/train': 2.068279981613159} 02/24/2022 07:35:18 - INFO - codeparrot_training - Step 4651: {'lr': 0.0004962463226988758, 'samples': 2381824, 'steps': 4651, 'loss/train': 1.4434808492660522} 02/24/2022 07:35:21 - INFO - codeparrot_training - Step 4652: {'lr': 0.0004962434973841497, 'samples': 2382336, 'steps': 4652, 'loss/train': 2.6536037921905518} 02/24/2022 07:35:27 - INFO - codeparrot_training - Step 4653: {'lr': 0.0004962406710145946, 'samples': 2382848, 'steps': 4653, 'loss/train': 1.1333259344100952} 02/24/2022 07:35:30 - INFO - codeparrot_training - Step 4654: {'lr': 0.0004962378435902228, 'samples': 2383360, 'steps': 4654, 'loss/train': 2.8344967365264893} 02/24/2022 07:35:36 - INFO - codeparrot_training - Step 4655: {'lr': 0.0004962350151110461, 'samples': 2383872, 'steps': 4655, 'loss/train': 2.6559736728668213} 02/24/2022 07:35:40 - INFO - codeparrot_training - Step 4656: {'lr': 0.0004962321855770769, 'samples': 2384384, 'steps': 4656, 'loss/train': 3.159492254257202} 02/24/2022 07:35:46 - INFO - codeparrot_training - Step 4657: {'lr': 0.0004962293549883273, 'samples': 2384896, 'steps': 4657, 'loss/train': 2.220210552215576} 02/24/2022 07:35:49 - INFO - codeparrot_training - Step 4658: {'lr': 0.0004962265233448092, 'samples': 2385408, 'steps': 4658, 'loss/train': 1.3266245126724243} 02/24/2022 07:35:55 - INFO - codeparrot_training - Step 4659: {'lr': 0.0004962236906465349, 'samples': 2385920, 'steps': 4659, 'loss/train': 1.2635562419891357} 02/24/2022 07:35:58 - INFO - codeparrot_training - Step 4660: {'lr': 0.0004962208568935164, 'samples': 2386432, 'steps': 4660, 'loss/train': 1.07157564163208} 02/24/2022 07:36:04 - INFO - codeparrot_training - Step 4661: {'lr': 0.000496218022085766, 'samples': 2386944, 'steps': 4661, 'loss/train': 2.566544532775879} 02/24/2022 07:36:07 - INFO - codeparrot_training - Step 4662: {'lr': 0.0004962151862232958, 'samples': 2387456, 'steps': 4662, 'loss/train': 3.157766103744507} 02/24/2022 07:36:13 - INFO - codeparrot_training - Step 4663: {'lr': 0.000496212349306118, 'samples': 2387968, 'steps': 4663, 'loss/train': 1.444319725036621} 02/24/2022 07:36:16 - INFO - codeparrot_training - Step 4664: {'lr': 0.0004962095113342445, 'samples': 2388480, 'steps': 4664, 'loss/train': 3.160513401031494} 02/24/2022 07:36:22 - INFO - codeparrot_training - Step 4665: {'lr': 0.0004962066723076878, 'samples': 2388992, 'steps': 4665, 'loss/train': 3.587603807449341} 02/24/2022 07:36:25 - INFO - codeparrot_training - Step 4666: {'lr': 0.0004962038322264598, 'samples': 2389504, 'steps': 4666, 'loss/train': 2.4390740394592285} 02/24/2022 07:36:31 - INFO - codeparrot_training - Step 4667: {'lr': 0.0004962009910905728, 'samples': 2390016, 'steps': 4667, 'loss/train': 2.259981632232666} 02/24/2022 07:36:35 - INFO - codeparrot_training - Step 4668: {'lr': 0.0004961981489000389, 'samples': 2390528, 'steps': 4668, 'loss/train': 1.3510923385620117} 02/24/2022 07:36:40 - INFO - codeparrot_training - Step 4669: {'lr': 0.0004961953056548703, 'samples': 2391040, 'steps': 4669, 'loss/train': 2.9432783126831055} 02/24/2022 07:36:44 - INFO - codeparrot_training - Step 4670: {'lr': 0.0004961924613550793, 'samples': 2391552, 'steps': 4670, 'loss/train': 3.6587672233581543} 02/24/2022 07:36:49 - INFO - codeparrot_training - Step 4671: {'lr': 0.0004961896160006778, 'samples': 2392064, 'steps': 4671, 'loss/train': 2.511075973510742} 02/24/2022 07:36:53 - INFO - codeparrot_training - Step 4672: {'lr': 0.0004961867695916782, 'samples': 2392576, 'steps': 4672, 'loss/train': 1.8166834115982056} 02/24/2022 07:36:58 - INFO - codeparrot_training - Step 4673: {'lr': 0.0004961839221280927, 'samples': 2393088, 'steps': 4673, 'loss/train': 0.8847571611404419} 02/24/2022 07:37:02 - INFO - codeparrot_training - Step 4674: {'lr': 0.0004961810736099334, 'samples': 2393600, 'steps': 4674, 'loss/train': 0.10410743951797485} 02/24/2022 07:37:07 - INFO - codeparrot_training - Step 4675: {'lr': 0.0004961782240372126, 'samples': 2394112, 'steps': 4675, 'loss/train': 3.4463889598846436} 02/24/2022 07:37:11 - INFO - codeparrot_training - Step 4676: {'lr': 0.0004961753734099425, 'samples': 2394624, 'steps': 4676, 'loss/train': 1.7875908613204956} 02/24/2022 07:37:17 - INFO - codeparrot_training - Step 4677: {'lr': 0.0004961725217281352, 'samples': 2395136, 'steps': 4677, 'loss/train': 3.275717258453369} 02/24/2022 07:37:20 - INFO - codeparrot_training - Step 4678: {'lr': 0.0004961696689918029, 'samples': 2395648, 'steps': 4678, 'loss/train': 1.3053271770477295} 02/24/2022 07:37:26 - INFO - codeparrot_training - Step 4679: {'lr': 0.0004961668152009581, 'samples': 2396160, 'steps': 4679, 'loss/train': 1.3668562173843384} 02/24/2022 07:37:29 - INFO - codeparrot_training - Step 4680: {'lr': 0.0004961639603556127, 'samples': 2396672, 'steps': 4680, 'loss/train': 1.2597614526748657} 02/24/2022 07:37:35 - INFO - codeparrot_training - Step 4681: {'lr': 0.0004961611044557792, 'samples': 2397184, 'steps': 4681, 'loss/train': 4.308879852294922} 02/24/2022 07:37:38 - INFO - codeparrot_training - Step 4682: {'lr': 0.0004961582475014695, 'samples': 2397696, 'steps': 4682, 'loss/train': 1.345685362815857} 02/24/2022 07:37:44 - INFO - codeparrot_training - Step 4683: {'lr': 0.0004961553894926961, 'samples': 2398208, 'steps': 4683, 'loss/train': 2.627019166946411} 02/24/2022 07:37:47 - INFO - codeparrot_training - Step 4684: {'lr': 0.0004961525304294712, 'samples': 2398720, 'steps': 4684, 'loss/train': 3.0345568656921387} 02/24/2022 07:37:53 - INFO - codeparrot_training - Step 4685: {'lr': 0.000496149670311807, 'samples': 2399232, 'steps': 4685, 'loss/train': 1.9586538076400757} 02/24/2022 07:37:56 - INFO - codeparrot_training - Step 4686: {'lr': 0.0004961468091397158, 'samples': 2399744, 'steps': 4686, 'loss/train': 2.8094279766082764} 02/24/2022 07:38:02 - INFO - codeparrot_training - Step 4687: {'lr': 0.0004961439469132098, 'samples': 2400256, 'steps': 4687, 'loss/train': 1.7700942754745483} 02/24/2022 07:38:05 - INFO - codeparrot_training - Step 4688: {'lr': 0.0004961410836323014, 'samples': 2400768, 'steps': 4688, 'loss/train': 2.9086923599243164} 02/24/2022 07:38:11 - INFO - codeparrot_training - Step 4689: {'lr': 0.0004961382192970027, 'samples': 2401280, 'steps': 4689, 'loss/train': 3.96055006980896} 02/24/2022 07:38:15 - INFO - codeparrot_training - Step 4690: {'lr': 0.0004961353539073258, 'samples': 2401792, 'steps': 4690, 'loss/train': 2.3432586193084717} 02/24/2022 07:38:20 - INFO - codeparrot_training - Step 4691: {'lr': 0.0004961324874632835, 'samples': 2402304, 'steps': 4691, 'loss/train': 1.7943015098571777} 02/24/2022 07:38:24 - INFO - codeparrot_training - Step 4692: {'lr': 0.0004961296199648877, 'samples': 2402816, 'steps': 4692, 'loss/train': 2.798440456390381} 02/24/2022 07:38:29 - INFO - codeparrot_training - Step 4693: {'lr': 0.0004961267514121507, 'samples': 2403328, 'steps': 4693, 'loss/train': 0.7034985423088074} 02/24/2022 07:38:33 - INFO - codeparrot_training - Step 4694: {'lr': 0.0004961238818050849, 'samples': 2403840, 'steps': 4694, 'loss/train': 1.9754117727279663} 02/24/2022 07:38:38 - INFO - codeparrot_training - Step 4695: {'lr': 0.0004961210111437026, 'samples': 2404352, 'steps': 4695, 'loss/train': 1.690004587173462} 02/24/2022 07:38:42 - INFO - codeparrot_training - Step 4696: {'lr': 0.0004961181394280159, 'samples': 2404864, 'steps': 4696, 'loss/train': 0.06594810634851456} 02/24/2022 07:38:47 - INFO - codeparrot_training - Step 4697: {'lr': 0.0004961152666580373, 'samples': 2405376, 'steps': 4697, 'loss/train': 1.377718210220337} 02/24/2022 07:38:51 - INFO - codeparrot_training - Step 4698: {'lr': 0.0004961123928337791, 'samples': 2405888, 'steps': 4698, 'loss/train': 2.882286787033081} 02/24/2022 07:38:56 - INFO - codeparrot_training - Step 4699: {'lr': 0.0004961095179552535, 'samples': 2406400, 'steps': 4699, 'loss/train': 2.091817855834961} 02/24/2022 07:39:00 - INFO - codeparrot_training - Step 4700: {'lr': 0.0004961066420224729, 'samples': 2406912, 'steps': 4700, 'loss/train': 2.072143316268921} 02/24/2022 07:39:05 - INFO - codeparrot_training - Step 4701: {'lr': 0.0004961037650354496, 'samples': 2407424, 'steps': 4701, 'loss/train': 2.9821889400482178} 02/24/2022 07:39:08 - INFO - codeparrot_training - Step 4702: {'lr': 0.0004961008869941959, 'samples': 2407936, 'steps': 4702, 'loss/train': 2.035592555999756} 02/24/2022 07:39:15 - INFO - codeparrot_training - Step 4703: {'lr': 0.0004960980078987241, 'samples': 2408448, 'steps': 4703, 'loss/train': 9.838709831237793} 02/24/2022 07:39:18 - INFO - codeparrot_training - Step 4704: {'lr': 0.0004960951277490467, 'samples': 2408960, 'steps': 4704, 'loss/train': 2.8876585960388184} 02/24/2022 07:39:24 - INFO - codeparrot_training - Step 4705: {'lr': 0.0004960922465451758, 'samples': 2409472, 'steps': 4705, 'loss/train': 2.791590929031372} 02/24/2022 07:39:27 - INFO - codeparrot_training - Step 4706: {'lr': 0.0004960893642871239, 'samples': 2409984, 'steps': 4706, 'loss/train': 3.412191867828369} 02/24/2022 07:39:33 - INFO - codeparrot_training - Step 4707: {'lr': 0.0004960864809749034, 'samples': 2410496, 'steps': 4707, 'loss/train': 0.14033402502536774} 02/24/2022 07:39:36 - INFO - codeparrot_training - Step 4708: {'lr': 0.0004960835966085264, 'samples': 2411008, 'steps': 4708, 'loss/train': 3.120861053466797} 02/24/2022 07:39:42 - INFO - codeparrot_training - Step 4709: {'lr': 0.0004960807111880055, 'samples': 2411520, 'steps': 4709, 'loss/train': 1.9810534715652466} 02/24/2022 07:39:45 - INFO - codeparrot_training - Step 4710: {'lr': 0.000496077824713353, 'samples': 2412032, 'steps': 4710, 'loss/train': 2.2970259189605713} 02/24/2022 07:39:51 - INFO - codeparrot_training - Step 4711: {'lr': 0.0004960749371845812, 'samples': 2412544, 'steps': 4711, 'loss/train': 2.2321465015411377} 02/24/2022 07:39:54 - INFO - codeparrot_training - Step 4712: {'lr': 0.0004960720486017025, 'samples': 2413056, 'steps': 4712, 'loss/train': 3.103388786315918} 02/24/2022 07:40:00 - INFO - codeparrot_training - Step 4713: {'lr': 0.0004960691589647292, 'samples': 2413568, 'steps': 4713, 'loss/train': 2.979495048522949} 02/24/2022 07:40:04 - INFO - codeparrot_training - Step 4714: {'lr': 0.0004960662682736739, 'samples': 2414080, 'steps': 4714, 'loss/train': 3.1824088096618652} 02/24/2022 07:40:09 - INFO - codeparrot_training - Step 4715: {'lr': 0.0004960633765285487, 'samples': 2414592, 'steps': 4715, 'loss/train': 2.706354856491089} 02/24/2022 07:40:13 - INFO - codeparrot_training - Step 4716: {'lr': 0.0004960604837293663, 'samples': 2415104, 'steps': 4716, 'loss/train': 2.2891921997070312} 02/24/2022 07:40:18 - INFO - codeparrot_training - Step 4717: {'lr': 0.0004960575898761388, 'samples': 2415616, 'steps': 4717, 'loss/train': 4.0091447830200195} 02/24/2022 07:40:24 - INFO - codeparrot_training - Step 4718: {'lr': 0.0004960546949688788, 'samples': 2416128, 'steps': 4718, 'loss/train': 2.0257010459899902} 02/24/2022 07:40:27 - INFO - codeparrot_training - Step 4719: {'lr': 0.0004960517990075985, 'samples': 2416640, 'steps': 4719, 'loss/train': 2.9779868125915527} 02/24/2022 07:40:33 - INFO - codeparrot_training - Step 4720: {'lr': 0.0004960489019923105, 'samples': 2417152, 'steps': 4720, 'loss/train': 1.8966171741485596} 02/24/2022 07:40:36 - INFO - codeparrot_training - Step 4721: {'lr': 0.0004960460039230271, 'samples': 2417664, 'steps': 4721, 'loss/train': 0.7914944291114807} 02/24/2022 07:40:43 - INFO - codeparrot_training - Step 4722: {'lr': 0.0004960431047997608, 'samples': 2418176, 'steps': 4722, 'loss/train': 2.8662352561950684} 02/24/2022 07:40:46 - INFO - codeparrot_training - Step 4723: {'lr': 0.0004960402046225239, 'samples': 2418688, 'steps': 4723, 'loss/train': 2.5579631328582764} 02/24/2022 07:40:50 - INFO - codeparrot_training - Step 4724: {'lr': 0.0004960373033913289, 'samples': 2419200, 'steps': 4724, 'loss/train': 1.533189058303833} 02/24/2022 07:40:55 - INFO - codeparrot_training - Step 4725: {'lr': 0.0004960344011061882, 'samples': 2419712, 'steps': 4725, 'loss/train': 2.4321110248565674} 02/24/2022 07:41:01 - INFO - codeparrot_training - Step 4726: {'lr': 0.0004960314977671144, 'samples': 2420224, 'steps': 4726, 'loss/train': 2.3497602939605713} 02/24/2022 07:41:04 - INFO - codeparrot_training - Step 4727: {'lr': 0.0004960285933741196, 'samples': 2420736, 'steps': 4727, 'loss/train': 2.9004337787628174} 02/24/2022 07:41:10 - INFO - codeparrot_training - Step 4728: {'lr': 0.0004960256879272166, 'samples': 2421248, 'steps': 4728, 'loss/train': 3.0766336917877197} 02/24/2022 07:41:13 - INFO - codeparrot_training - Step 4729: {'lr': 0.0004960227814264175, 'samples': 2421760, 'steps': 4729, 'loss/train': 2.4607222080230713} 02/24/2022 07:41:19 - INFO - codeparrot_training - Step 4730: {'lr': 0.0004960198738717351, 'samples': 2422272, 'steps': 4730, 'loss/train': 2.8965721130371094} 02/24/2022 07:41:22 - INFO - codeparrot_training - Step 4731: {'lr': 0.0004960169652631815, 'samples': 2422784, 'steps': 4731, 'loss/train': 2.2890076637268066} 02/24/2022 07:41:28 - INFO - codeparrot_training - Step 4732: {'lr': 0.0004960140556007695, 'samples': 2423296, 'steps': 4732, 'loss/train': 1.5802545547485352} 02/24/2022 07:41:31 - INFO - codeparrot_training - Step 4733: {'lr': 0.0004960111448845114, 'samples': 2423808, 'steps': 4733, 'loss/train': 2.5724680423736572} 02/24/2022 07:41:37 - INFO - codeparrot_training - Step 4734: {'lr': 0.0004960082331144195, 'samples': 2424320, 'steps': 4734, 'loss/train': 2.1319644451141357} 02/24/2022 07:41:41 - INFO - codeparrot_training - Step 4735: {'lr': 0.0004960053202905066, 'samples': 2424832, 'steps': 4735, 'loss/train': 2.4284427165985107} 02/24/2022 07:41:46 - INFO - codeparrot_training - Step 4736: {'lr': 0.0004960024064127849, 'samples': 2425344, 'steps': 4736, 'loss/train': 2.9356603622436523} 02/24/2022 07:41:50 - INFO - codeparrot_training - Step 4737: {'lr': 0.0004959994914812671, 'samples': 2425856, 'steps': 4737, 'loss/train': 2.7840700149536133} 02/24/2022 07:41:55 - INFO - codeparrot_training - Step 4738: {'lr': 0.0004959965754959656, 'samples': 2426368, 'steps': 4738, 'loss/train': 2.49600887298584} 02/24/2022 07:41:59 - INFO - codeparrot_training - Step 4739: {'lr': 0.0004959936584568928, 'samples': 2426880, 'steps': 4739, 'loss/train': 2.3466830253601074} 02/24/2022 07:42:04 - INFO - codeparrot_training - Step 4740: {'lr': 0.0004959907403640614, 'samples': 2427392, 'steps': 4740, 'loss/train': 2.574820041656494} 02/24/2022 07:42:07 - INFO - codeparrot_training - Step 4741: {'lr': 0.0004959878212174837, 'samples': 2427904, 'steps': 4741, 'loss/train': 1.8421263694763184} 02/24/2022 07:42:13 - INFO - codeparrot_training - Step 4742: {'lr': 0.0004959849010171723, 'samples': 2428416, 'steps': 4742, 'loss/train': 1.4805868864059448} 02/24/2022 07:42:16 - INFO - codeparrot_training - Step 4743: {'lr': 0.0004959819797631397, 'samples': 2428928, 'steps': 4743, 'loss/train': 2.4147841930389404} 02/24/2022 07:42:22 - INFO - codeparrot_training - Step 4744: {'lr': 0.0004959790574553984, 'samples': 2429440, 'steps': 4744, 'loss/train': 3.531266689300537} 02/24/2022 07:42:26 - INFO - codeparrot_training - Step 4745: {'lr': 0.000495976134093961, 'samples': 2429952, 'steps': 4745, 'loss/train': 9.151768684387207} 02/24/2022 07:42:31 - INFO - codeparrot_training - Step 4746: {'lr': 0.0004959732096788398, 'samples': 2430464, 'steps': 4746, 'loss/train': 0.6532483696937561} 02/24/2022 07:42:35 - INFO - codeparrot_training - Step 4747: {'lr': 0.0004959702842100475, 'samples': 2430976, 'steps': 4747, 'loss/train': 2.721266984939575} 02/24/2022 07:42:41 - INFO - codeparrot_training - Step 4748: {'lr': 0.0004959673576875967, 'samples': 2431488, 'steps': 4748, 'loss/train': 3.1963508129119873} 02/24/2022 07:42:44 - INFO - codeparrot_training - Step 4749: {'lr': 0.0004959644301114998, 'samples': 2432000, 'steps': 4749, 'loss/train': 2.7351653575897217} 02/24/2022 07:42:50 - INFO - codeparrot_training - Step 4750: {'lr': 0.0004959615014817694, 'samples': 2432512, 'steps': 4750, 'loss/train': 3.450247049331665} 02/24/2022 07:42:53 - INFO - codeparrot_training - Step 4751: {'lr': 0.000495958571798418, 'samples': 2433024, 'steps': 4751, 'loss/train': 3.1717827320098877} 02/24/2022 07:42:57 - INFO - codeparrot_training - Step 4752: {'lr': 0.0004959556410614582, 'samples': 2433536, 'steps': 4752, 'loss/train': 3.368734836578369} 02/24/2022 07:43:02 - INFO - codeparrot_training - Step 4753: {'lr': 0.0004959527092709026, 'samples': 2434048, 'steps': 4753, 'loss/train': 2.6499545574188232} 02/24/2022 07:43:06 - INFO - codeparrot_training - Step 4754: {'lr': 0.0004959497764267636, 'samples': 2434560, 'steps': 4754, 'loss/train': 2.5016541481018066} 02/24/2022 07:43:12 - INFO - codeparrot_training - Step 4755: {'lr': 0.0004959468425290537, 'samples': 2435072, 'steps': 4755, 'loss/train': 2.6367087364196777} 02/24/2022 07:43:15 - INFO - codeparrot_training - Step 4756: {'lr': 0.0004959439075777858, 'samples': 2435584, 'steps': 4756, 'loss/train': 1.3696736097335815} 02/24/2022 07:43:20 - INFO - codeparrot_training - Step 4757: {'lr': 0.0004959409715729723, 'samples': 2436096, 'steps': 4757, 'loss/train': 3.7594516277313232} 02/24/2022 07:43:24 - INFO - codeparrot_training - Step 4758: {'lr': 0.0004959380345146258, 'samples': 2436608, 'steps': 4758, 'loss/train': 2.6778650283813477} 02/24/2022 07:43:29 - INFO - codeparrot_training - Step 4759: {'lr': 0.0004959350964027588, 'samples': 2437120, 'steps': 4759, 'loss/train': 2.489854335784912} 02/24/2022 07:43:33 - INFO - codeparrot_training - Step 4760: {'lr': 0.000495932157237384, 'samples': 2437632, 'steps': 4760, 'loss/train': 1.8308134078979492} 02/24/2022 07:43:39 - INFO - codeparrot_training - Step 4761: {'lr': 0.0004959292170185139, 'samples': 2438144, 'steps': 4761, 'loss/train': 2.295203685760498} 02/24/2022 07:43:42 - INFO - codeparrot_training - Step 4762: {'lr': 0.0004959262757461611, 'samples': 2438656, 'steps': 4762, 'loss/train': 1.9961360692977905} 02/24/2022 07:43:48 - INFO - codeparrot_training - Step 4763: {'lr': 0.0004959233334203382, 'samples': 2439168, 'steps': 4763, 'loss/train': 2.186973810195923} 02/24/2022 07:43:51 - INFO - codeparrot_training - Step 4764: {'lr': 0.0004959203900410579, 'samples': 2439680, 'steps': 4764, 'loss/train': 1.625566840171814} 02/24/2022 07:43:57 - INFO - codeparrot_training - Step 4765: {'lr': 0.0004959174456083327, 'samples': 2440192, 'steps': 4765, 'loss/train': 2.9025657176971436} 02/24/2022 07:44:00 - INFO - codeparrot_training - Step 4766: {'lr': 0.0004959145001221752, 'samples': 2440704, 'steps': 4766, 'loss/train': 2.9964284896850586} 02/24/2022 07:44:06 - INFO - codeparrot_training - Step 4767: {'lr': 0.0004959115535825982, 'samples': 2441216, 'steps': 4767, 'loss/train': 3.570746660232544} 02/24/2022 07:44:09 - INFO - codeparrot_training - Step 4768: {'lr': 0.000495908605989614, 'samples': 2441728, 'steps': 4768, 'loss/train': 3.0606961250305176} 02/24/2022 07:44:15 - INFO - codeparrot_training - Step 4769: {'lr': 0.0004959056573432357, 'samples': 2442240, 'steps': 4769, 'loss/train': 3.297722816467285} 02/24/2022 07:44:18 - INFO - codeparrot_training - Step 4770: {'lr': 0.0004959027076434754, 'samples': 2442752, 'steps': 4770, 'loss/train': 2.21748423576355} 02/24/2022 07:44:24 - INFO - codeparrot_training - Step 4771: {'lr': 0.000495899756890346, 'samples': 2443264, 'steps': 4771, 'loss/train': 2.501883029937744} 02/24/2022 07:44:28 - INFO - codeparrot_training - Step 4772: {'lr': 0.0004958968050838603, 'samples': 2443776, 'steps': 4772, 'loss/train': 9.372735023498535} 02/24/2022 07:44:33 - INFO - codeparrot_training - Step 4773: {'lr': 0.0004958938522240306, 'samples': 2444288, 'steps': 4773, 'loss/train': 3.2129766941070557} 02/24/2022 07:44:37 - INFO - codeparrot_training - Step 4774: {'lr': 0.0004958908983108697, 'samples': 2444800, 'steps': 4774, 'loss/train': 2.83373761177063} 02/24/2022 07:44:42 - INFO - codeparrot_training - Step 4775: {'lr': 0.0004958879433443903, 'samples': 2445312, 'steps': 4775, 'loss/train': 1.9512957334518433} 02/24/2022 07:44:48 - INFO - codeparrot_training - Step 4776: {'lr': 0.0004958849873246051, 'samples': 2445824, 'steps': 4776, 'loss/train': 3.2849318981170654} 02/24/2022 07:44:51 - INFO - codeparrot_training - Step 4777: {'lr': 0.0004958820302515268, 'samples': 2446336, 'steps': 4777, 'loss/train': 3.2047841548919678} 02/24/2022 07:44:57 - INFO - codeparrot_training - Step 4778: {'lr': 0.0004958790721251678, 'samples': 2446848, 'steps': 4778, 'loss/train': 3.4885566234588623} 02/24/2022 07:45:00 - INFO - codeparrot_training - Step 4779: {'lr': 0.000495876112945541, 'samples': 2447360, 'steps': 4779, 'loss/train': 2.5625088214874268} 02/24/2022 07:45:06 - INFO - codeparrot_training - Step 4780: {'lr': 0.0004958731527126589, 'samples': 2447872, 'steps': 4780, 'loss/train': 1.7746145725250244} 02/24/2022 07:45:10 - INFO - codeparrot_training - Step 4781: {'lr': 0.0004958701914265344, 'samples': 2448384, 'steps': 4781, 'loss/train': 2.9764153957366943} 02/24/2022 07:45:15 - INFO - codeparrot_training - Step 4782: {'lr': 0.0004958672290871799, 'samples': 2448896, 'steps': 4782, 'loss/train': 1.6381715536117554} 02/24/2022 07:45:19 - INFO - codeparrot_training - Step 4783: {'lr': 0.0004958642656946084, 'samples': 2449408, 'steps': 4783, 'loss/train': 2.3984715938568115} 02/24/2022 07:45:24 - INFO - codeparrot_training - Step 4784: {'lr': 0.0004958613012488324, 'samples': 2449920, 'steps': 4784, 'loss/train': 3.122859239578247} 02/24/2022 07:45:28 - INFO - codeparrot_training - Step 4785: {'lr': 0.0004958583357498647, 'samples': 2450432, 'steps': 4785, 'loss/train': 3.215186834335327} 02/24/2022 07:45:33 - INFO - codeparrot_training - Step 4786: {'lr': 0.000495855369197718, 'samples': 2450944, 'steps': 4786, 'loss/train': 2.622568130493164} 02/24/2022 07:45:37 - INFO - codeparrot_training - Step 4787: {'lr': 0.0004958524015924048, 'samples': 2451456, 'steps': 4787, 'loss/train': 2.390336751937866} 02/24/2022 07:45:42 - INFO - codeparrot_training - Step 4788: {'lr': 0.0004958494329339382, 'samples': 2451968, 'steps': 4788, 'loss/train': 2.356473445892334} 02/24/2022 07:45:46 - INFO - codeparrot_training - Step 4789: {'lr': 0.0004958464632223306, 'samples': 2452480, 'steps': 4789, 'loss/train': 2.064406394958496} 02/24/2022 07:45:51 - INFO - codeparrot_training - Step 4790: {'lr': 0.0004958434924575947, 'samples': 2452992, 'steps': 4790, 'loss/train': 2.5083656311035156} 02/24/2022 07:45:55 - INFO - codeparrot_training - Step 4791: {'lr': 0.0004958405206397434, 'samples': 2453504, 'steps': 4791, 'loss/train': 2.026231050491333} 02/24/2022 07:46:00 - INFO - codeparrot_training - Step 4792: {'lr': 0.0004958375477687896, 'samples': 2454016, 'steps': 4792, 'loss/train': 1.7246315479278564} 02/24/2022 07:46:04 - INFO - codeparrot_training - Step 4793: {'lr': 0.0004958345738447456, 'samples': 2454528, 'steps': 4793, 'loss/train': 1.6558524370193481} 02/24/2022 07:46:09 - INFO - codeparrot_training - Step 4794: {'lr': 0.0004958315988676244, 'samples': 2455040, 'steps': 4794, 'loss/train': 3.594749689102173} 02/24/2022 07:46:13 - INFO - codeparrot_training - Step 4795: {'lr': 0.0004958286228374387, 'samples': 2455552, 'steps': 4795, 'loss/train': 2.831083059310913} 02/24/2022 07:46:19 - INFO - codeparrot_training - Step 4796: {'lr': 0.0004958256457542011, 'samples': 2456064, 'steps': 4796, 'loss/train': 1.8108325004577637} 02/24/2022 07:46:23 - INFO - codeparrot_training - Step 4797: {'lr': 0.0004958226676179246, 'samples': 2456576, 'steps': 4797, 'loss/train': 3.0596425533294678} 02/24/2022 07:46:28 - INFO - codeparrot_training - Step 4798: {'lr': 0.0004958196884286218, 'samples': 2457088, 'steps': 4798, 'loss/train': 2.8492672443389893} 02/24/2022 07:46:32 - INFO - codeparrot_training - Step 4799: {'lr': 0.0004958167081863057, 'samples': 2457600, 'steps': 4799, 'loss/train': 2.4905967712402344} 02/24/2022 07:46:37 - INFO - codeparrot_training - Step 4800: {'lr': 0.0004958137268909887, 'samples': 2458112, 'steps': 4800, 'loss/train': 1.6208795309066772} 02/24/2022 07:46:41 - INFO - codeparrot_training - Step 4801: {'lr': 0.0004958107445426838, 'samples': 2458624, 'steps': 4801, 'loss/train': 2.433300256729126} 02/24/2022 07:46:46 - INFO - codeparrot_training - Step 4802: {'lr': 0.0004958077611414037, 'samples': 2459136, 'steps': 4802, 'loss/train': 2.9399757385253906} 02/24/2022 07:46:50 - INFO - codeparrot_training - Step 4803: {'lr': 0.0004958047766871612, 'samples': 2459648, 'steps': 4803, 'loss/train': 2.5095489025115967} 02/24/2022 07:46:55 - INFO - codeparrot_training - Step 4804: {'lr': 0.000495801791179969, 'samples': 2460160, 'steps': 4804, 'loss/train': 2.0438201427459717} 02/24/2022 07:46:59 - INFO - codeparrot_training - Step 4805: {'lr': 0.0004957988046198401, 'samples': 2460672, 'steps': 4805, 'loss/train': 2.9106061458587646} 02/24/2022 07:47:05 - INFO - codeparrot_training - Step 4806: {'lr': 0.0004957958170067872, 'samples': 2461184, 'steps': 4806, 'loss/train': 3.167832612991333} 02/24/2022 07:47:08 - INFO - codeparrot_training - Step 4807: {'lr': 0.000495792828340823, 'samples': 2461696, 'steps': 4807, 'loss/train': 3.0617074966430664} 02/24/2022 07:47:14 - INFO - codeparrot_training - Step 4808: {'lr': 0.0004957898386219603, 'samples': 2462208, 'steps': 4808, 'loss/train': 2.184767007827759} 02/24/2022 07:47:17 - INFO - codeparrot_training - Step 4809: {'lr': 0.0004957868478502121, 'samples': 2462720, 'steps': 4809, 'loss/train': 3.0218400955200195} 02/24/2022 07:47:23 - INFO - codeparrot_training - Step 4810: {'lr': 0.0004957838560255911, 'samples': 2463232, 'steps': 4810, 'loss/train': 3.329874277114868} 02/24/2022 07:47:26 - INFO - codeparrot_training - Step 4811: {'lr': 0.0004957808631481101, 'samples': 2463744, 'steps': 4811, 'loss/train': 4.094907283782959} 02/24/2022 07:47:32 - INFO - codeparrot_training - Step 4812: {'lr': 0.0004957778692177819, 'samples': 2464256, 'steps': 4812, 'loss/train': 1.6182080507278442} 02/24/2022 07:47:35 - INFO - codeparrot_training - Step 4813: {'lr': 0.0004957748742346193, 'samples': 2464768, 'steps': 4813, 'loss/train': 2.0050013065338135} 02/24/2022 07:47:41 - INFO - codeparrot_training - Step 4814: {'lr': 0.0004957718781986352, 'samples': 2465280, 'steps': 4814, 'loss/train': 3.025818109512329} 02/24/2022 07:47:44 - INFO - codeparrot_training - Step 4815: {'lr': 0.0004957688811098425, 'samples': 2465792, 'steps': 4815, 'loss/train': 3.4738645553588867} 02/24/2022 07:47:50 - INFO - codeparrot_training - Step 4816: {'lr': 0.0004957658829682539, 'samples': 2466304, 'steps': 4816, 'loss/train': 2.82922101020813} 02/24/2022 07:47:54 - INFO - codeparrot_training - Step 4817: {'lr': 0.0004957628837738823, 'samples': 2466816, 'steps': 4817, 'loss/train': 2.2350382804870605} 02/24/2022 07:47:59 - INFO - codeparrot_training - Step 4818: {'lr': 0.0004957598835267405, 'samples': 2467328, 'steps': 4818, 'loss/train': 1.9026799201965332} 02/24/2022 07:48:03 - INFO - codeparrot_training - Step 4819: {'lr': 0.0004957568822268415, 'samples': 2467840, 'steps': 4819, 'loss/train': 1.5194218158721924} 02/24/2022 07:48:08 - INFO - codeparrot_training - Step 4820: {'lr': 0.000495753879874198, 'samples': 2468352, 'steps': 4820, 'loss/train': 2.190692901611328} 02/24/2022 07:48:12 - INFO - codeparrot_training - Step 4821: {'lr': 0.0004957508764688227, 'samples': 2468864, 'steps': 4821, 'loss/train': 1.6686153411865234} 02/24/2022 07:48:17 - INFO - codeparrot_training - Step 4822: {'lr': 0.000495747872010729, 'samples': 2469376, 'steps': 4822, 'loss/train': 1.8834881782531738} 02/24/2022 07:48:21 - INFO - codeparrot_training - Step 4823: {'lr': 0.0004957448664999293, 'samples': 2469888, 'steps': 4823, 'loss/train': 2.2773869037628174} 02/24/2022 07:48:26 - INFO - codeparrot_training - Step 4824: {'lr': 0.0004957418599364367, 'samples': 2470400, 'steps': 4824, 'loss/train': 4.282899379730225} 02/24/2022 07:48:30 - INFO - codeparrot_training - Step 4825: {'lr': 0.000495738852320264, 'samples': 2470912, 'steps': 4825, 'loss/train': 2.3439242839813232} 02/24/2022 07:48:36 - INFO - codeparrot_training - Step 4826: {'lr': 0.000495735843651424, 'samples': 2471424, 'steps': 4826, 'loss/train': 2.6990227699279785} 02/24/2022 07:48:39 - INFO - codeparrot_training - Step 4827: {'lr': 0.0004957328339299297, 'samples': 2471936, 'steps': 4827, 'loss/train': 2.958448886871338} 02/24/2022 07:48:45 - INFO - codeparrot_training - Step 4828: {'lr': 0.0004957298231557939, 'samples': 2472448, 'steps': 4828, 'loss/train': 2.2011032104492188} 02/24/2022 07:48:48 - INFO - codeparrot_training - Step 4829: {'lr': 0.0004957268113290297, 'samples': 2472960, 'steps': 4829, 'loss/train': 2.080993175506592} 02/24/2022 07:48:54 - INFO - codeparrot_training - Step 4830: {'lr': 0.0004957237984496499, 'samples': 2473472, 'steps': 4830, 'loss/train': 2.498842239379883} 02/24/2022 07:48:59 - INFO - codeparrot_training - Step 4831: {'lr': 0.0004957207845176673, 'samples': 2473984, 'steps': 4831, 'loss/train': 1.2883267402648926} 02/24/2022 07:49:03 - INFO - codeparrot_training - Step 4832: {'lr': 0.0004957177695330948, 'samples': 2474496, 'steps': 4832, 'loss/train': 3.31538724899292} 02/24/2022 07:49:08 - INFO - codeparrot_training - Step 4833: {'lr': 0.0004957147534959455, 'samples': 2475008, 'steps': 4833, 'loss/train': 2.475583553314209} 02/24/2022 07:49:12 - INFO - codeparrot_training - Step 4834: {'lr': 0.0004957117364062321, 'samples': 2475520, 'steps': 4834, 'loss/train': 2.8139469623565674} 02/24/2022 07:49:17 - INFO - codeparrot_training - Step 4835: {'lr': 0.0004957087182639678, 'samples': 2476032, 'steps': 4835, 'loss/train': 1.918376088142395} 02/24/2022 07:49:21 - INFO - codeparrot_training - Step 4836: {'lr': 0.0004957056990691653, 'samples': 2476544, 'steps': 4836, 'loss/train': 2.501250982284546} 02/24/2022 07:49:26 - INFO - codeparrot_training - Step 4837: {'lr': 0.0004957026788218377, 'samples': 2477056, 'steps': 4837, 'loss/train': 3.309501886367798} 02/24/2022 07:49:30 - INFO - codeparrot_training - Step 4838: {'lr': 0.0004956996575219977, 'samples': 2477568, 'steps': 4838, 'loss/train': 1.1400737762451172} 02/24/2022 07:49:35 - INFO - codeparrot_training - Step 4839: {'lr': 0.0004956966351696584, 'samples': 2478080, 'steps': 4839, 'loss/train': 2.0966250896453857} 02/24/2022 07:49:39 - INFO - codeparrot_training - Step 4840: {'lr': 0.0004956936117648329, 'samples': 2478592, 'steps': 4840, 'loss/train': 1.8754032850265503} 02/24/2022 07:49:45 - INFO - codeparrot_training - Step 4841: {'lr': 0.0004956905873075338, 'samples': 2479104, 'steps': 4841, 'loss/train': 1.5526013374328613} 02/24/2022 07:49:48 - INFO - codeparrot_training - Step 4842: {'lr': 0.0004956875617977743, 'samples': 2479616, 'steps': 4842, 'loss/train': 2.265995979309082} 02/24/2022 07:49:54 - INFO - codeparrot_training - Step 4843: {'lr': 0.0004956845352355674, 'samples': 2480128, 'steps': 4843, 'loss/train': 2.154154062271118} 02/24/2022 07:49:57 - INFO - codeparrot_training - Step 4844: {'lr': 0.0004956815076209257, 'samples': 2480640, 'steps': 4844, 'loss/train': 2.8307952880859375} 02/24/2022 07:50:03 - INFO - codeparrot_training - Step 4845: {'lr': 0.0004956784789538626, 'samples': 2481152, 'steps': 4845, 'loss/train': 2.029360294342041} 02/24/2022 07:50:06 - INFO - codeparrot_training - Step 4846: {'lr': 0.000495675449234391, 'samples': 2481664, 'steps': 4846, 'loss/train': 2.8690202236175537} 02/24/2022 07:50:12 - INFO - codeparrot_training - Step 4847: {'lr': 0.0004956724184625237, 'samples': 2482176, 'steps': 4847, 'loss/train': 2.2508604526519775} 02/24/2022 07:50:15 - INFO - codeparrot_training - Step 4848: {'lr': 0.0004956693866382738, 'samples': 2482688, 'steps': 4848, 'loss/train': 1.745201826095581} 02/24/2022 07:50:21 - INFO - codeparrot_training - Step 4849: {'lr': 0.0004956663537616542, 'samples': 2483200, 'steps': 4849, 'loss/train': 2.035811185836792} 02/24/2022 07:50:24 - INFO - codeparrot_training - Step 4850: {'lr': 0.000495663319832678, 'samples': 2483712, 'steps': 4850, 'loss/train': 3.593346118927002} 02/24/2022 07:50:31 - INFO - codeparrot_training - Step 4851: {'lr': 0.0004956602848513581, 'samples': 2484224, 'steps': 4851, 'loss/train': 1.7882438898086548} 02/24/2022 07:50:34 - INFO - codeparrot_training - Step 4852: {'lr': 0.0004956572488177075, 'samples': 2484736, 'steps': 4852, 'loss/train': 2.32806658744812} 02/24/2022 07:50:40 - INFO - codeparrot_training - Step 4853: {'lr': 0.0004956542117317393, 'samples': 2485248, 'steps': 4853, 'loss/train': 2.117522954940796} 02/24/2022 07:50:43 - INFO - codeparrot_training - Step 4854: {'lr': 0.0004956511735934665, 'samples': 2485760, 'steps': 4854, 'loss/train': 2.0202419757843018} 02/24/2022 07:50:49 - INFO - codeparrot_training - Step 4855: {'lr': 0.000495648134402902, 'samples': 2486272, 'steps': 4855, 'loss/train': 2.7553694248199463} 02/24/2022 07:50:52 - INFO - codeparrot_training - Step 4856: {'lr': 0.0004956450941600589, 'samples': 2486784, 'steps': 4856, 'loss/train': 2.8186230659484863} 02/24/2022 07:50:58 - INFO - codeparrot_training - Step 4857: {'lr': 0.0004956420528649504, 'samples': 2487296, 'steps': 4857, 'loss/train': 2.7916064262390137} 02/24/2022 07:51:01 - INFO - codeparrot_training - Step 4858: {'lr': 0.0004956390105175892, 'samples': 2487808, 'steps': 4858, 'loss/train': 2.8463239669799805} 02/24/2022 07:51:07 - INFO - codeparrot_training - Step 4859: {'lr': 0.0004956359671179885, 'samples': 2488320, 'steps': 4859, 'loss/train': 3.3343734741210938} 02/24/2022 07:51:10 - INFO - codeparrot_training - Step 4860: {'lr': 0.0004956329226661612, 'samples': 2488832, 'steps': 4860, 'loss/train': 2.231766700744629} 02/24/2022 07:51:17 - INFO - codeparrot_training - Step 4861: {'lr': 0.0004956298771621206, 'samples': 2489344, 'steps': 4861, 'loss/train': 2.2583096027374268} 02/24/2022 07:51:20 - INFO - codeparrot_training - Step 4862: {'lr': 0.0004956268306058795, 'samples': 2489856, 'steps': 4862, 'loss/train': 3.411510944366455} 02/24/2022 07:51:26 - INFO - codeparrot_training - Step 4863: {'lr': 0.0004956237829974511, 'samples': 2490368, 'steps': 4863, 'loss/train': 1.1921054124832153} 02/24/2022 07:51:29 - INFO - codeparrot_training - Step 4864: {'lr': 0.0004956207343368485, 'samples': 2490880, 'steps': 4864, 'loss/train': 3.353442907333374} 02/24/2022 07:51:35 - INFO - codeparrot_training - Step 4865: {'lr': 0.0004956176846240845, 'samples': 2491392, 'steps': 4865, 'loss/train': 1.3059569597244263} 02/24/2022 07:51:38 - INFO - codeparrot_training - Step 4866: {'lr': 0.0004956146338591725, 'samples': 2491904, 'steps': 4866, 'loss/train': 2.7087414264678955} 02/24/2022 07:51:43 - INFO - codeparrot_training - Step 4867: {'lr': 0.0004956115820421253, 'samples': 2492416, 'steps': 4867, 'loss/train': 2.6254632472991943} 02/24/2022 07:51:47 - INFO - codeparrot_training - Step 4868: {'lr': 0.000495608529172956, 'samples': 2492928, 'steps': 4868, 'loss/train': 3.8123397827148438} 02/24/2022 07:51:52 - INFO - codeparrot_training - Step 4869: {'lr': 0.000495605475251678, 'samples': 2493440, 'steps': 4869, 'loss/train': 2.5492396354675293} 02/24/2022 07:51:56 - INFO - codeparrot_training - Step 4870: {'lr': 0.000495602420278304, 'samples': 2493952, 'steps': 4870, 'loss/train': 1.9818994998931885} 02/24/2022 07:52:02 - INFO - codeparrot_training - Step 4871: {'lr': 0.0004955993642528471, 'samples': 2494464, 'steps': 4871, 'loss/train': 2.781245231628418} 02/24/2022 07:52:05 - INFO - codeparrot_training - Step 4872: {'lr': 0.0004955963071753206, 'samples': 2494976, 'steps': 4872, 'loss/train': 2.337423801422119} 02/24/2022 07:52:11 - INFO - codeparrot_training - Step 4873: {'lr': 0.0004955932490457375, 'samples': 2495488, 'steps': 4873, 'loss/train': 2.89913010597229} 02/24/2022 07:52:14 - INFO - codeparrot_training - Step 4874: {'lr': 0.0004955901898641109, 'samples': 2496000, 'steps': 4874, 'loss/train': 3.992082357406616} 02/24/2022 07:52:20 - INFO - codeparrot_training - Step 4875: {'lr': 0.000495587129630454, 'samples': 2496512, 'steps': 4875, 'loss/train': 2.445347309112549} 02/24/2022 07:52:24 - INFO - codeparrot_training - Step 4876: {'lr': 0.0004955840683447797, 'samples': 2497024, 'steps': 4876, 'loss/train': 2.601616144180298} 02/24/2022 07:52:29 - INFO - codeparrot_training - Step 4877: {'lr': 0.0004955810060071012, 'samples': 2497536, 'steps': 4877, 'loss/train': 3.540217161178589} 02/24/2022 07:52:33 - INFO - codeparrot_training - Step 4878: {'lr': 0.0004955779426174318, 'samples': 2498048, 'steps': 4878, 'loss/train': 2.6730260848999023} 02/24/2022 07:52:38 - INFO - codeparrot_training - Step 4879: {'lr': 0.0004955748781757844, 'samples': 2498560, 'steps': 4879, 'loss/train': 3.049858808517456} 02/24/2022 07:52:42 - INFO - codeparrot_training - Step 4880: {'lr': 0.0004955718126821722, 'samples': 2499072, 'steps': 4880, 'loss/train': 2.644120216369629} 02/24/2022 07:52:49 - INFO - codeparrot_training - Step 4881: {'lr': 0.0004955687461366083, 'samples': 2499584, 'steps': 4881, 'loss/train': 2.384840250015259} 02/24/2022 07:52:52 - INFO - codeparrot_training - Step 4882: {'lr': 0.000495565678539106, 'samples': 2500096, 'steps': 4882, 'loss/train': 4.711916923522949} 02/24/2022 07:52:58 - INFO - codeparrot_training - Step 4883: {'lr': 0.0004955626098896782, 'samples': 2500608, 'steps': 4883, 'loss/train': 1.9414739608764648} 02/24/2022 07:53:01 - INFO - codeparrot_training - Step 4884: {'lr': 0.0004955595401883381, 'samples': 2501120, 'steps': 4884, 'loss/train': 2.6034696102142334} 02/24/2022 07:53:07 - INFO - codeparrot_training - Step 4885: {'lr': 0.0004955564694350989, 'samples': 2501632, 'steps': 4885, 'loss/train': 3.098513126373291} 02/24/2022 07:53:10 - INFO - codeparrot_training - Step 4886: {'lr': 0.0004955533976299739, 'samples': 2502144, 'steps': 4886, 'loss/train': 0.36670711636543274} 02/24/2022 07:53:16 - INFO - codeparrot_training - Step 4887: {'lr': 0.000495550324772976, 'samples': 2502656, 'steps': 4887, 'loss/train': 3.2154476642608643} 02/24/2022 07:53:19 - INFO - codeparrot_training - Step 4888: {'lr': 0.0004955472508641186, 'samples': 2503168, 'steps': 4888, 'loss/train': 3.24361515045166} 02/24/2022 07:53:25 - INFO - codeparrot_training - Step 4889: {'lr': 0.0004955441759034146, 'samples': 2503680, 'steps': 4889, 'loss/train': 0.37054336071014404} 02/24/2022 07:53:28 - INFO - codeparrot_training - Step 4890: {'lr': 0.0004955410998908774, 'samples': 2504192, 'steps': 4890, 'loss/train': 2.3825619220733643} 02/24/2022 07:53:34 - INFO - codeparrot_training - Step 4891: {'lr': 0.0004955380228265201, 'samples': 2504704, 'steps': 4891, 'loss/train': 0.7356753945350647} 02/24/2022 07:53:37 - INFO - codeparrot_training - Step 4892: {'lr': 0.0004955349447103559, 'samples': 2505216, 'steps': 4892, 'loss/train': 3.7161316871643066} 02/24/2022 07:53:44 - INFO - codeparrot_training - Step 4893: {'lr': 0.000495531865542398, 'samples': 2505728, 'steps': 4893, 'loss/train': 1.4669992923736572} 02/24/2022 07:53:47 - INFO - codeparrot_training - Step 4894: {'lr': 0.0004955287853226594, 'samples': 2506240, 'steps': 4894, 'loss/train': 2.5684609413146973} 02/24/2022 07:53:53 - INFO - codeparrot_training - Step 4895: {'lr': 0.0004955257040511534, 'samples': 2506752, 'steps': 4895, 'loss/train': 3.071998357772827} 02/24/2022 07:53:56 - INFO - codeparrot_training - Step 4896: {'lr': 0.0004955226217278934, 'samples': 2507264, 'steps': 4896, 'loss/train': 3.3817026615142822} 02/24/2022 07:54:01 - INFO - codeparrot_training - Step 4897: {'lr': 0.0004955195383528926, 'samples': 2507776, 'steps': 4897, 'loss/train': 2.566835403442383} 02/24/2022 07:54:05 - INFO - codeparrot_training - Step 4898: {'lr': 0.0004955164539261638, 'samples': 2508288, 'steps': 4898, 'loss/train': 3.8253231048583984} 02/24/2022 07:54:10 - INFO - codeparrot_training - Step 4899: {'lr': 0.0004955133684477205, 'samples': 2508800, 'steps': 4899, 'loss/train': 2.7843708992004395} 02/24/2022 07:54:14 - INFO - codeparrot_training - Step 4900: {'lr': 0.000495510281917576, 'samples': 2509312, 'steps': 4900, 'loss/train': 3.378871440887451} 02/24/2022 07:54:19 - INFO - codeparrot_training - Step 4901: {'lr': 0.0004955071943357433, 'samples': 2509824, 'steps': 4901, 'loss/train': 1.7449642419815063} 02/24/2022 07:54:23 - INFO - codeparrot_training - Step 4902: {'lr': 0.0004955041057022358, 'samples': 2510336, 'steps': 4902, 'loss/train': 3.5450022220611572} 02/24/2022 07:54:28 - INFO - codeparrot_training - Step 4903: {'lr': 0.0004955010160170667, 'samples': 2510848, 'steps': 4903, 'loss/train': 2.7884042263031006} 02/24/2022 07:54:32 - INFO - codeparrot_training - Step 4904: {'lr': 0.0004954979252802491, 'samples': 2511360, 'steps': 4904, 'loss/train': 1.781516194343567} 02/24/2022 07:54:37 - INFO - codeparrot_training - Step 4905: {'lr': 0.0004954948334917965, 'samples': 2511872, 'steps': 4905, 'loss/train': 2.7518675327301025} 02/24/2022 07:54:41 - INFO - codeparrot_training - Step 4906: {'lr': 0.0004954917406517218, 'samples': 2512384, 'steps': 4906, 'loss/train': 1.969922423362732} 02/24/2022 07:54:46 - INFO - codeparrot_training - Step 4907: {'lr': 0.0004954886467600386, 'samples': 2512896, 'steps': 4907, 'loss/train': 2.238471508026123} 02/24/2022 07:54:50 - INFO - codeparrot_training - Step 4908: {'lr': 0.0004954855518167599, 'samples': 2513408, 'steps': 4908, 'loss/train': 3.4221787452697754} 02/24/2022 07:54:57 - INFO - codeparrot_training - Step 4909: {'lr': 0.000495482455821899, 'samples': 2513920, 'steps': 4909, 'loss/train': 2.2988929748535156} 02/24/2022 07:55:00 - INFO - codeparrot_training - Step 4910: {'lr': 0.0004954793587754694, 'samples': 2514432, 'steps': 4910, 'loss/train': 2.313474416732788} 02/24/2022 07:55:06 - INFO - codeparrot_training - Step 4911: {'lr': 0.000495476260677484, 'samples': 2514944, 'steps': 4911, 'loss/train': 2.7729668617248535} 02/24/2022 07:55:09 - INFO - codeparrot_training - Step 4912: {'lr': 0.0004954731615279563, 'samples': 2515456, 'steps': 4912, 'loss/train': 3.1681737899780273} 02/24/2022 07:55:15 - INFO - codeparrot_training - Step 4913: {'lr': 0.0004954700613268995, 'samples': 2515968, 'steps': 4913, 'loss/train': 2.9729576110839844} 02/24/2022 07:55:18 - INFO - codeparrot_training - Step 4914: {'lr': 0.0004954669600743269, 'samples': 2516480, 'steps': 4914, 'loss/train': 1.7505345344543457} 02/24/2022 07:55:24 - INFO - codeparrot_training - Step 4915: {'lr': 0.0004954638577702519, 'samples': 2516992, 'steps': 4915, 'loss/train': 2.5069775581359863} 02/24/2022 07:55:27 - INFO - codeparrot_training - Step 4916: {'lr': 0.0004954607544146875, 'samples': 2517504, 'steps': 4916, 'loss/train': 2.729753017425537} 02/24/2022 07:55:34 - INFO - codeparrot_training - Step 4917: {'lr': 0.0004954576500076472, 'samples': 2518016, 'steps': 4917, 'loss/train': 2.9764349460601807} 02/24/2022 07:55:37 - INFO - codeparrot_training - Step 4918: {'lr': 0.0004954545445491444, 'samples': 2518528, 'steps': 4918, 'loss/train': 2.1443498134613037} 02/24/2022 07:55:43 - INFO - codeparrot_training - Step 4919: {'lr': 0.0004954514380391921, 'samples': 2519040, 'steps': 4919, 'loss/train': 1.6374441385269165} 02/24/2022 07:55:46 - INFO - codeparrot_training - Step 4920: {'lr': 0.0004954483304778039, 'samples': 2519552, 'steps': 4920, 'loss/train': 2.5993611812591553} 02/24/2022 07:55:52 - INFO - codeparrot_training - Step 4921: {'lr': 0.0004954452218649929, 'samples': 2520064, 'steps': 4921, 'loss/train': 1.3553427457809448} 02/24/2022 07:55:55 - INFO - codeparrot_training - Step 4922: {'lr': 0.0004954421122007727, 'samples': 2520576, 'steps': 4922, 'loss/train': 2.369081497192383} 02/24/2022 07:56:01 - INFO - codeparrot_training - Step 4923: {'lr': 0.0004954390014851563, 'samples': 2521088, 'steps': 4923, 'loss/train': 1.9360178709030151} 02/24/2022 07:56:04 - INFO - codeparrot_training - Step 4924: {'lr': 0.0004954358897181571, 'samples': 2521600, 'steps': 4924, 'loss/train': 3.213473081588745} 02/24/2022 07:56:10 - INFO - codeparrot_training - Step 4925: {'lr': 0.0004954327768997885, 'samples': 2522112, 'steps': 4925, 'loss/train': 1.8457989692687988} 02/24/2022 07:56:13 - INFO - codeparrot_training - Step 4926: {'lr': 0.0004954296630300638, 'samples': 2522624, 'steps': 4926, 'loss/train': 3.197357654571533} 02/24/2022 07:56:19 - INFO - codeparrot_training - Step 4927: {'lr': 0.0004954265481089965, 'samples': 2523136, 'steps': 4927, 'loss/train': 2.4606659412384033} 02/24/2022 07:56:22 - INFO - codeparrot_training - Step 4928: {'lr': 0.0004954234321365998, 'samples': 2523648, 'steps': 4928, 'loss/train': 2.2355072498321533} 02/24/2022 07:56:29 - INFO - codeparrot_training - Step 4929: {'lr': 0.0004954203151128868, 'samples': 2524160, 'steps': 4929, 'loss/train': 2.40077543258667} 02/24/2022 07:56:32 - INFO - codeparrot_training - Step 4930: {'lr': 0.0004954171970378713, 'samples': 2524672, 'steps': 4930, 'loss/train': 3.895455837249756} 02/24/2022 07:56:38 - INFO - codeparrot_training - Step 4931: {'lr': 0.0004954140779115664, 'samples': 2525184, 'steps': 4931, 'loss/train': 2.3514821529388428} 02/24/2022 07:56:43 - INFO - codeparrot_training - Step 4932: {'lr': 0.0004954109577339856, 'samples': 2525696, 'steps': 4932, 'loss/train': 2.851844310760498} 02/24/2022 07:56:47 - INFO - codeparrot_training - Step 4933: {'lr': 0.0004954078365051421, 'samples': 2526208, 'steps': 4933, 'loss/train': 1.7190126180648804} 02/24/2022 07:56:52 - INFO - codeparrot_training - Step 4934: {'lr': 0.0004954047142250494, 'samples': 2526720, 'steps': 4934, 'loss/train': 2.6115102767944336} 02/24/2022 07:56:56 - INFO - codeparrot_training - Step 4935: {'lr': 0.0004954015908937208, 'samples': 2527232, 'steps': 4935, 'loss/train': 3.137942314147949} 02/24/2022 07:57:01 - INFO - codeparrot_training - Step 4936: {'lr': 0.0004953984665111697, 'samples': 2527744, 'steps': 4936, 'loss/train': 1.0885988473892212} 02/24/2022 07:57:05 - INFO - codeparrot_training - Step 4937: {'lr': 0.0004953953410774095, 'samples': 2528256, 'steps': 4937, 'loss/train': 1.9508899450302124} 02/24/2022 07:57:10 - INFO - codeparrot_training - Step 4938: {'lr': 0.0004953922145924535, 'samples': 2528768, 'steps': 4938, 'loss/train': 2.9652576446533203} 02/24/2022 07:57:14 - INFO - codeparrot_training - Step 4939: {'lr': 0.0004953890870563153, 'samples': 2529280, 'steps': 4939, 'loss/train': 2.772034168243408} 02/24/2022 07:57:19 - INFO - codeparrot_training - Step 4940: {'lr': 0.0004953859584690081, 'samples': 2529792, 'steps': 4940, 'loss/train': 2.0161304473876953} 02/24/2022 07:57:23 - INFO - codeparrot_training - Step 4941: {'lr': 0.0004953828288305454, 'samples': 2530304, 'steps': 4941, 'loss/train': 1.348802089691162} 02/24/2022 07:57:29 - INFO - codeparrot_training - Step 4942: {'lr': 0.0004953796981409407, 'samples': 2530816, 'steps': 4942, 'loss/train': 3.405233383178711} 02/24/2022 07:57:32 - INFO - codeparrot_training - Step 4943: {'lr': 0.0004953765664002071, 'samples': 2531328, 'steps': 4943, 'loss/train': 2.201634407043457} 02/24/2022 07:57:38 - INFO - codeparrot_training - Step 4944: {'lr': 0.0004953734336083582, 'samples': 2531840, 'steps': 4944, 'loss/train': 3.315370798110962} 02/24/2022 07:57:41 - INFO - codeparrot_training - Step 4945: {'lr': 0.0004953702997654076, 'samples': 2532352, 'steps': 4945, 'loss/train': 2.597675085067749} 02/24/2022 07:57:47 - INFO - codeparrot_training - Step 4946: {'lr': 0.0004953671648713683, 'samples': 2532864, 'steps': 4946, 'loss/train': 2.51084566116333} 02/24/2022 07:57:50 - INFO - codeparrot_training - Step 4947: {'lr': 0.0004953640289262542, 'samples': 2533376, 'steps': 4947, 'loss/train': 3.502622127532959} 02/24/2022 07:57:56 - INFO - codeparrot_training - Step 4948: {'lr': 0.0004953608919300784, 'samples': 2533888, 'steps': 4948, 'loss/train': 3.09306263923645} 02/24/2022 07:57:59 - INFO - codeparrot_training - Step 4949: {'lr': 0.0004953577538828546, 'samples': 2534400, 'steps': 4949, 'loss/train': 2.943469285964966} 02/24/2022 07:58:05 - INFO - codeparrot_training - Step 4950: {'lr': 0.0004953546147845959, 'samples': 2534912, 'steps': 4950, 'loss/train': 2.031853675842285} 02/24/2022 07:58:08 - INFO - codeparrot_training - Step 4951: {'lr': 0.0004953514746353161, 'samples': 2535424, 'steps': 4951, 'loss/train': 1.1166609525680542} 02/24/2022 07:58:14 - INFO - codeparrot_training - Step 4952: {'lr': 0.0004953483334350283, 'samples': 2535936, 'steps': 4952, 'loss/train': 3.162663221359253} 02/24/2022 07:58:18 - INFO - codeparrot_training - Step 4953: {'lr': 0.0004953451911837463, 'samples': 2536448, 'steps': 4953, 'loss/train': 1.1098450422286987} 02/24/2022 07:58:23 - INFO - codeparrot_training - Step 4954: {'lr': 0.0004953420478814834, 'samples': 2536960, 'steps': 4954, 'loss/train': 2.0787851810455322} 02/24/2022 07:58:27 - INFO - codeparrot_training - Step 4955: {'lr': 0.000495338903528253, 'samples': 2537472, 'steps': 4955, 'loss/train': 1.8731580972671509} 02/24/2022 07:58:32 - INFO - codeparrot_training - Step 4956: {'lr': 0.0004953357581240686, 'samples': 2537984, 'steps': 4956, 'loss/train': 1.1025416851043701} 02/24/2022 07:58:36 - INFO - codeparrot_training - Step 4957: {'lr': 0.0004953326116689438, 'samples': 2538496, 'steps': 4957, 'loss/train': 2.0030360221862793} 02/24/2022 07:58:41 - INFO - codeparrot_training - Step 4958: {'lr': 0.000495329464162892, 'samples': 2539008, 'steps': 4958, 'loss/train': 3.1406311988830566} 02/24/2022 07:58:45 - INFO - codeparrot_training - Step 4959: {'lr': 0.0004953263156059266, 'samples': 2539520, 'steps': 4959, 'loss/train': 3.2624127864837646} 02/24/2022 07:58:50 - INFO - codeparrot_training - Step 4960: {'lr': 0.0004953231659980613, 'samples': 2540032, 'steps': 4960, 'loss/train': 3.0520901679992676} 02/24/2022 07:58:54 - INFO - codeparrot_training - Step 4961: {'lr': 0.0004953200153393094, 'samples': 2540544, 'steps': 4961, 'loss/train': 2.382667303085327} 02/24/2022 07:59:00 - INFO - codeparrot_training - Step 4962: {'lr': 0.0004953168636296845, 'samples': 2541056, 'steps': 4962, 'loss/train': 5.038918495178223} 02/24/2022 07:59:03 - INFO - codeparrot_training - Step 4963: {'lr': 0.0004953137108691999, 'samples': 2541568, 'steps': 4963, 'loss/train': 8.362198829650879} 02/24/2022 07:59:07 - INFO - codeparrot_training - Step 4964: {'lr': 0.0004953105570578693, 'samples': 2542080, 'steps': 4964, 'loss/train': 2.4258134365081787} 02/24/2022 07:59:13 - INFO - codeparrot_training - Step 4965: {'lr': 0.0004953074021957063, 'samples': 2542592, 'steps': 4965, 'loss/train': 3.187837600708008} 02/24/2022 07:59:16 - INFO - codeparrot_training - Step 4966: {'lr': 0.0004953042462827242, 'samples': 2543104, 'steps': 4966, 'loss/train': 1.609328031539917} 02/24/2022 07:59:22 - INFO - codeparrot_training - Step 4967: {'lr': 0.0004953010893189365, 'samples': 2543616, 'steps': 4967, 'loss/train': 2.3354032039642334} 02/24/2022 07:59:25 - INFO - codeparrot_training - Step 4968: {'lr': 0.000495297931304357, 'samples': 2544128, 'steps': 4968, 'loss/train': 1.3369238376617432} 02/24/2022 07:59:31 - INFO - codeparrot_training - Step 4969: {'lr': 0.000495294772238999, 'samples': 2544640, 'steps': 4969, 'loss/train': 2.266425132751465} 02/24/2022 07:59:34 - INFO - codeparrot_training - Step 4970: {'lr': 0.000495291612122876, 'samples': 2545152, 'steps': 4970, 'loss/train': 3.3470370769500732} 02/24/2022 07:59:40 - INFO - codeparrot_training - Step 4971: {'lr': 0.0004952884509560017, 'samples': 2545664, 'steps': 4971, 'loss/train': 3.1451051235198975} 02/24/2022 07:59:44 - INFO - codeparrot_training - Step 4972: {'lr': 0.0004952852887383895, 'samples': 2546176, 'steps': 4972, 'loss/train': 3.0164761543273926} 02/24/2022 07:59:49 - INFO - codeparrot_training - Step 4973: {'lr': 0.0004952821254700531, 'samples': 2546688, 'steps': 4973, 'loss/train': 2.5824825763702393} 02/24/2022 07:59:53 - INFO - codeparrot_training - Step 4974: {'lr': 0.0004952789611510059, 'samples': 2547200, 'steps': 4974, 'loss/train': 2.556976556777954} 02/24/2022 07:59:59 - INFO - codeparrot_training - Step 4975: {'lr': 0.0004952757957812615, 'samples': 2547712, 'steps': 4975, 'loss/train': 2.773397207260132} 02/24/2022 08:00:03 - INFO - codeparrot_training - Step 4976: {'lr': 0.0004952726293608335, 'samples': 2548224, 'steps': 4976, 'loss/train': 1.9987236261367798} 02/24/2022 08:00:08 - INFO - codeparrot_training - Step 4977: {'lr': 0.0004952694618897354, 'samples': 2548736, 'steps': 4977, 'loss/train': 1.3548074960708618} 02/24/2022 08:00:12 - INFO - codeparrot_training - Step 4978: {'lr': 0.0004952662933679809, 'samples': 2549248, 'steps': 4978, 'loss/train': 2.1706817150115967} 02/24/2022 08:00:17 - INFO - codeparrot_training - Step 4979: {'lr': 0.0004952631237955835, 'samples': 2549760, 'steps': 4979, 'loss/train': 2.3538026809692383} 02/24/2022 08:00:21 - INFO - codeparrot_training - Step 4980: {'lr': 0.0004952599531725567, 'samples': 2550272, 'steps': 4980, 'loss/train': 0.886324405670166} 02/24/2022 08:00:26 - INFO - codeparrot_training - Step 4981: {'lr': 0.0004952567814989141, 'samples': 2550784, 'steps': 4981, 'loss/train': 2.7025487422943115} 02/24/2022 08:00:30 - INFO - codeparrot_training - Step 4982: {'lr': 0.0004952536087746693, 'samples': 2551296, 'steps': 4982, 'loss/train': 2.810694694519043} 02/24/2022 08:00:35 - INFO - codeparrot_training - Step 4983: {'lr': 0.000495250434999836, 'samples': 2551808, 'steps': 4983, 'loss/train': 3.363544225692749} 02/24/2022 08:00:39 - INFO - codeparrot_training - Step 4984: {'lr': 0.0004952472601744277, 'samples': 2552320, 'steps': 4984, 'loss/train': 1.953047752380371} 02/24/2022 08:00:44 - INFO - codeparrot_training - Step 4985: {'lr': 0.000495244084298458, 'samples': 2552832, 'steps': 4985, 'loss/train': 2.4962847232818604} 02/24/2022 08:00:48 - INFO - codeparrot_training - Step 4986: {'lr': 0.0004952409073719405, 'samples': 2553344, 'steps': 4986, 'loss/train': 3.0354714393615723} 02/24/2022 08:00:53 - INFO - codeparrot_training - Step 4987: {'lr': 0.0004952377293948888, 'samples': 2553856, 'steps': 4987, 'loss/train': 2.5642433166503906} 02/24/2022 08:00:57 - INFO - codeparrot_training - Step 4988: {'lr': 0.0004952345503673166, 'samples': 2554368, 'steps': 4988, 'loss/train': 2.23897647857666} 02/24/2022 08:01:03 - INFO - codeparrot_training - Step 4989: {'lr': 0.0004952313702892375, 'samples': 2554880, 'steps': 4989, 'loss/train': 0.9788317680358887} 02/24/2022 08:01:06 - INFO - codeparrot_training - Step 4990: {'lr': 0.0004952281891606649, 'samples': 2555392, 'steps': 4990, 'loss/train': 1.6288464069366455} 02/24/2022 08:01:12 - INFO - codeparrot_training - Step 4991: {'lr': 0.0004952250069816127, 'samples': 2555904, 'steps': 4991, 'loss/train': 1.7622431516647339} 02/24/2022 08:01:15 - INFO - codeparrot_training - Step 4992: {'lr': 0.0004952218237520945, 'samples': 2556416, 'steps': 4992, 'loss/train': 0.8125408291816711} 02/24/2022 08:01:21 - INFO - codeparrot_training - Step 4993: {'lr': 0.0004952186394721239, 'samples': 2556928, 'steps': 4993, 'loss/train': 2.602688789367676} 02/24/2022 08:01:24 - INFO - codeparrot_training - Step 4994: {'lr': 0.0004952154541417144, 'samples': 2557440, 'steps': 4994, 'loss/train': 2.5203757286071777} 02/24/2022 08:01:30 - INFO - codeparrot_training - Step 4995: {'lr': 0.0004952122677608798, 'samples': 2557952, 'steps': 4995, 'loss/train': 2.7976837158203125} 02/24/2022 08:01:33 - INFO - codeparrot_training - Step 4996: {'lr': 0.0004952090803296337, 'samples': 2558464, 'steps': 4996, 'loss/train': 2.25091290473938} 02/24/2022 08:01:39 - INFO - codeparrot_training - Step 4997: {'lr': 0.0004952058918479899, 'samples': 2558976, 'steps': 4997, 'loss/train': 2.8913917541503906} 02/24/2022 08:01:42 - INFO - codeparrot_training - Step 4998: {'lr': 0.0004952027023159617, 'samples': 2559488, 'steps': 4998, 'loss/train': 3.246948719024658} 02/24/2022 08:01:48 - INFO - codeparrot_training - Step 4999: {'lr': 0.0004951995117335631, 'samples': 2560000, 'steps': 4999, 'loss/train': 2.412139415740967} 02/24/2022 08:01:48 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 08:02:04 - WARNING - huggingface_hub.repository - Several commits (5) will be pushed upstream. 02/24/2022 08:02:04 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 08:02:36 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy f6d96e1..8f2bfb2 floral-grass-11 -> floral-grass-11 02/24/2022 08:02:40 - INFO - codeparrot_training - Step 5000: {'lr': 0.0004951963201008077, 'samples': 2560512, 'steps': 5000, 'loss/train': 2.296778917312622} 02/24/2022 08:02:47 - INFO - codeparrot_training - Step 5001: {'lr': 0.000495193127417709, 'samples': 2561024, 'steps': 5001, 'loss/train': 2.5641069412231445} 02/24/2022 08:02:50 - INFO - codeparrot_training - Step 5002: {'lr': 0.0004951899336842809, 'samples': 2561536, 'steps': 5002, 'loss/train': 2.7109029293060303} 02/24/2022 08:02:56 - INFO - codeparrot_training - Step 5003: {'lr': 0.0004951867389005369, 'samples': 2562048, 'steps': 5003, 'loss/train': 2.5077733993530273} 02/24/2022 08:02:59 - INFO - codeparrot_training - Step 5004: {'lr': 0.0004951835430664908, 'samples': 2562560, 'steps': 5004, 'loss/train': 2.3653955459594727} 02/24/2022 08:03:05 - INFO - codeparrot_training - Step 5005: {'lr': 0.0004951803461821562, 'samples': 2563072, 'steps': 5005, 'loss/train': 2.838548183441162} 02/24/2022 08:03:09 - INFO - codeparrot_training - Step 5006: {'lr': 0.0004951771482475469, 'samples': 2563584, 'steps': 5006, 'loss/train': 2.699648141860962} 02/24/2022 08:03:14 - INFO - codeparrot_training - Step 5007: {'lr': 0.0004951739492626766, 'samples': 2564096, 'steps': 5007, 'loss/train': 2.0335335731506348} 02/24/2022 08:03:18 - INFO - codeparrot_training - Step 5008: {'lr': 0.0004951707492275589, 'samples': 2564608, 'steps': 5008, 'loss/train': 2.9375572204589844} 02/24/2022 08:03:23 - INFO - codeparrot_training - Step 5009: {'lr': 0.0004951675481422075, 'samples': 2565120, 'steps': 5009, 'loss/train': 2.2810325622558594} 02/24/2022 08:03:27 - INFO - codeparrot_training - Step 5010: {'lr': 0.0004951643460066363, 'samples': 2565632, 'steps': 5010, 'loss/train': 4.218005180358887} 02/24/2022 08:03:33 - INFO - codeparrot_training - Step 5011: {'lr': 0.0004951611428208589, 'samples': 2566144, 'steps': 5011, 'loss/train': 3.0080645084381104} 02/24/2022 08:03:36 - INFO - codeparrot_training - Step 5012: {'lr': 0.0004951579385848889, 'samples': 2566656, 'steps': 5012, 'loss/train': 2.662175416946411} 02/24/2022 08:03:42 - INFO - codeparrot_training - Step 5013: {'lr': 0.0004951547332987401, 'samples': 2567168, 'steps': 5013, 'loss/train': 1.3648183345794678} 02/24/2022 08:03:45 - INFO - codeparrot_training - Step 5014: {'lr': 0.0004951515269624265, 'samples': 2567680, 'steps': 5014, 'loss/train': 2.1121575832366943} 02/24/2022 08:03:51 - INFO - codeparrot_training - Step 5015: {'lr': 0.0004951483195759614, 'samples': 2568192, 'steps': 5015, 'loss/train': 3.2036707401275635} 02/24/2022 08:03:54 - INFO - codeparrot_training - Step 5016: {'lr': 0.0004951451111393588, 'samples': 2568704, 'steps': 5016, 'loss/train': 2.3789610862731934} 02/24/2022 08:04:00 - INFO - codeparrot_training - Step 5017: {'lr': 0.0004951419016526324, 'samples': 2569216, 'steps': 5017, 'loss/train': 2.780557155609131} 02/24/2022 08:04:03 - INFO - codeparrot_training - Step 5018: {'lr': 0.0004951386911157959, 'samples': 2569728, 'steps': 5018, 'loss/train': 2.697455883026123} 02/24/2022 08:04:09 - INFO - codeparrot_training - Step 5019: {'lr': 0.0004951354795288631, 'samples': 2570240, 'steps': 5019, 'loss/train': 3.0525665283203125} 02/24/2022 08:04:13 - INFO - codeparrot_training - Step 5020: {'lr': 0.0004951322668918477, 'samples': 2570752, 'steps': 5020, 'loss/train': 2.4270029067993164} 02/24/2022 08:04:16 - INFO - codeparrot_training - Step 5021: {'lr': 0.0004951290532047637, 'samples': 2571264, 'steps': 5021, 'loss/train': 0.8056251406669617} 02/24/2022 08:04:22 - INFO - codeparrot_training - Step 5022: {'lr': 0.0004951258384676244, 'samples': 2571776, 'steps': 5022, 'loss/train': 2.556169033050537} 02/24/2022 08:04:26 - INFO - codeparrot_training - Step 5023: {'lr': 0.0004951226226804441, 'samples': 2572288, 'steps': 5023, 'loss/train': 2.4295411109924316} 02/24/2022 08:04:31 - INFO - codeparrot_training - Step 5024: {'lr': 0.0004951194058432361, 'samples': 2572800, 'steps': 5024, 'loss/train': 2.9989123344421387} 02/24/2022 08:04:35 - INFO - codeparrot_training - Step 5025: {'lr': 0.0004951161879560146, 'samples': 2573312, 'steps': 5025, 'loss/train': 3.6421656608581543} 02/24/2022 08:04:40 - INFO - codeparrot_training - Step 5026: {'lr': 0.000495112969018793, 'samples': 2573824, 'steps': 5026, 'loss/train': 2.5617172718048096} 02/24/2022 08:04:44 - INFO - codeparrot_training - Step 5027: {'lr': 0.0004951097490315853, 'samples': 2574336, 'steps': 5027, 'loss/train': 2.223910331726074} 02/24/2022 08:04:49 - INFO - codeparrot_training - Step 5028: {'lr': 0.0004951065279944054, 'samples': 2574848, 'steps': 5028, 'loss/train': 3.514998197555542} 02/24/2022 08:04:53 - INFO - codeparrot_training - Step 5029: {'lr': 0.0004951033059072668, 'samples': 2575360, 'steps': 5029, 'loss/train': 2.5756258964538574} 02/24/2022 08:04:58 - INFO - codeparrot_training - Step 5030: {'lr': 0.0004951000827701836, 'samples': 2575872, 'steps': 5030, 'loss/train': 1.9040237665176392} 02/24/2022 08:05:02 - INFO - codeparrot_training - Step 5031: {'lr': 0.0004950968585831694, 'samples': 2576384, 'steps': 5031, 'loss/train': 1.8934390544891357} 02/24/2022 08:05:08 - INFO - codeparrot_training - Step 5032: {'lr': 0.0004950936333462381, 'samples': 2576896, 'steps': 5032, 'loss/train': 2.160876989364624} 02/24/2022 08:05:12 - INFO - codeparrot_training - Step 5033: {'lr': 0.0004950904070594036, 'samples': 2577408, 'steps': 5033, 'loss/train': 2.021130323410034} 02/24/2022 08:05:17 - INFO - codeparrot_training - Step 5034: {'lr': 0.0004950871797226795, 'samples': 2577920, 'steps': 5034, 'loss/train': 2.519498586654663} 02/24/2022 08:05:21 - INFO - codeparrot_training - Step 5035: {'lr': 0.0004950839513360798, 'samples': 2578432, 'steps': 5035, 'loss/train': 2.521667242050171} 02/24/2022 08:05:26 - INFO - codeparrot_training - Step 5036: {'lr': 0.0004950807218996182, 'samples': 2578944, 'steps': 5036, 'loss/train': 1.9231854677200317} 02/24/2022 08:05:30 - INFO - codeparrot_training - Step 5037: {'lr': 0.0004950774914133086, 'samples': 2579456, 'steps': 5037, 'loss/train': 1.0685920715332031} 02/24/2022 08:05:35 - INFO - codeparrot_training - Step 5038: {'lr': 0.0004950742598771649, 'samples': 2579968, 'steps': 5038, 'loss/train': 2.0484297275543213} 02/24/2022 08:05:41 - INFO - codeparrot_training - Step 5039: {'lr': 0.0004950710272912009, 'samples': 2580480, 'steps': 5039, 'loss/train': 2.969024658203125} 02/24/2022 08:05:44 - INFO - codeparrot_training - Step 5040: {'lr': 0.0004950677936554305, 'samples': 2580992, 'steps': 5040, 'loss/train': 2.8437583446502686} 02/24/2022 08:05:50 - INFO - codeparrot_training - Step 5041: {'lr': 0.0004950645589698674, 'samples': 2581504, 'steps': 5041, 'loss/train': 2.2863240242004395} 02/24/2022 08:05:53 - INFO - codeparrot_training - Step 5042: {'lr': 0.0004950613232345256, 'samples': 2582016, 'steps': 5042, 'loss/train': 2.68483304977417} 02/24/2022 08:05:59 - INFO - codeparrot_training - Step 5043: {'lr': 0.0004950580864494188, 'samples': 2582528, 'steps': 5043, 'loss/train': 3.0379183292388916} 02/24/2022 08:06:02 - INFO - codeparrot_training - Step 5044: {'lr': 0.0004950548486145611, 'samples': 2583040, 'steps': 5044, 'loss/train': 2.3883142471313477} 02/24/2022 08:06:08 - INFO - codeparrot_training - Step 5045: {'lr': 0.0004950516097299662, 'samples': 2583552, 'steps': 5045, 'loss/train': 2.279304027557373} 02/24/2022 08:06:11 - INFO - codeparrot_training - Step 5046: {'lr': 0.000495048369795648, 'samples': 2584064, 'steps': 5046, 'loss/train': 1.8159793615341187} 02/24/2022 08:06:18 - INFO - codeparrot_training - Step 5047: {'lr': 0.0004950451288116204, 'samples': 2584576, 'steps': 5047, 'loss/train': 2.7002248764038086} 02/24/2022 08:06:21 - INFO - codeparrot_training - Step 5048: {'lr': 0.0004950418867778973, 'samples': 2585088, 'steps': 5048, 'loss/train': 2.738889694213867} 02/24/2022 08:06:27 - INFO - codeparrot_training - Step 5049: {'lr': 0.0004950386436944925, 'samples': 2585600, 'steps': 5049, 'loss/train': 2.8263957500457764} 02/24/2022 08:06:30 - INFO - codeparrot_training - Step 5050: {'lr': 0.0004950353995614201, 'samples': 2586112, 'steps': 5050, 'loss/train': 2.439209461212158} 02/24/2022 08:06:36 - INFO - codeparrot_training - Step 5051: {'lr': 0.0004950321543786937, 'samples': 2586624, 'steps': 5051, 'loss/train': 2.6172542572021484} 02/24/2022 08:06:39 - INFO - codeparrot_training - Step 5052: {'lr': 0.0004950289081463273, 'samples': 2587136, 'steps': 5052, 'loss/train': 3.2669732570648193} 02/24/2022 08:06:45 - INFO - codeparrot_training - Step 5053: {'lr': 0.0004950256608643351, 'samples': 2587648, 'steps': 5053, 'loss/train': 1.2929632663726807} 02/24/2022 08:06:48 - INFO - codeparrot_training - Step 5054: {'lr': 0.0004950224125327307, 'samples': 2588160, 'steps': 5054, 'loss/train': 2.0164365768432617} 02/24/2022 08:06:54 - INFO - codeparrot_training - Step 5055: {'lr': 0.000495019163151528, 'samples': 2588672, 'steps': 5055, 'loss/train': 3.002439022064209} 02/24/2022 08:06:57 - INFO - codeparrot_training - Step 5056: {'lr': 0.0004950159127207411, 'samples': 2589184, 'steps': 5056, 'loss/train': 2.2478277683258057} 02/24/2022 08:07:04 - INFO - codeparrot_training - Step 5057: {'lr': 0.0004950126612403838, 'samples': 2589696, 'steps': 5057, 'loss/train': 3.1365468502044678} 02/24/2022 08:07:07 - INFO - codeparrot_training - Step 5058: {'lr': 0.00049500940871047, 'samples': 2590208, 'steps': 5058, 'loss/train': 3.099783420562744} 02/24/2022 08:07:12 - INFO - codeparrot_training - Step 5059: {'lr': 0.0004950061551310138, 'samples': 2590720, 'steps': 5059, 'loss/train': 2.2945163249969482} 02/24/2022 08:07:16 - INFO - codeparrot_training - Step 5060: {'lr': 0.0004950029005020289, 'samples': 2591232, 'steps': 5060, 'loss/train': 1.7013286352157593} 02/24/2022 08:07:21 - INFO - codeparrot_training - Step 5061: {'lr': 0.0004949996448235294, 'samples': 2591744, 'steps': 5061, 'loss/train': 1.8031805753707886} 02/24/2022 08:07:25 - INFO - codeparrot_training - Step 5062: {'lr': 0.0004949963880955293, 'samples': 2592256, 'steps': 5062, 'loss/train': 2.8380470275878906} 02/24/2022 08:07:30 - INFO - codeparrot_training - Step 5063: {'lr': 0.0004949931303180424, 'samples': 2592768, 'steps': 5063, 'loss/train': 2.81138277053833} 02/24/2022 08:07:34 - INFO - codeparrot_training - Step 5064: {'lr': 0.0004949898714910828, 'samples': 2593280, 'steps': 5064, 'loss/train': 2.8589231967926025} 02/24/2022 08:07:39 - INFO - codeparrot_training - Step 5065: {'lr': 0.0004949866116146643, 'samples': 2593792, 'steps': 5065, 'loss/train': 1.7443885803222656} 02/24/2022 08:07:43 - INFO - codeparrot_training - Step 5066: {'lr': 0.000494983350688801, 'samples': 2594304, 'steps': 5066, 'loss/train': 2.3759613037109375} 02/24/2022 08:07:49 - INFO - codeparrot_training - Step 5067: {'lr': 0.0004949800887135067, 'samples': 2594816, 'steps': 5067, 'loss/train': 1.7359416484832764} 02/24/2022 08:07:53 - INFO - codeparrot_training - Step 5068: {'lr': 0.0004949768256887956, 'samples': 2595328, 'steps': 5068, 'loss/train': 3.2318248748779297} 02/24/2022 08:07:58 - INFO - codeparrot_training - Step 5069: {'lr': 0.0004949735616146816, 'samples': 2595840, 'steps': 5069, 'loss/train': 3.1380720138549805} 02/24/2022 08:08:01 - INFO - codeparrot_training - Step 5070: {'lr': 0.0004949702964911787, 'samples': 2596352, 'steps': 5070, 'loss/train': 2.6741809844970703} 02/24/2022 08:08:07 - INFO - codeparrot_training - Step 5071: {'lr': 0.0004949670303183006, 'samples': 2596864, 'steps': 5071, 'loss/train': 3.599745512008667} 02/24/2022 08:08:10 - INFO - codeparrot_training - Step 5072: {'lr': 0.0004949637630960618, 'samples': 2597376, 'steps': 5072, 'loss/train': 5.534940719604492} 02/24/2022 08:08:16 - INFO - codeparrot_training - Step 5073: {'lr': 0.0004949604948244758, 'samples': 2597888, 'steps': 5073, 'loss/train': 2.1387131214141846} 02/24/2022 08:08:20 - INFO - codeparrot_training - Step 5074: {'lr': 0.0004949572255035569, 'samples': 2598400, 'steps': 5074, 'loss/train': 2.5684547424316406} 02/24/2022 08:08:25 - INFO - codeparrot_training - Step 5075: {'lr': 0.0004949539551333191, 'samples': 2598912, 'steps': 5075, 'loss/train': 1.5825589895248413} 02/24/2022 08:08:28 - INFO - codeparrot_training - Step 5076: {'lr': 0.0004949506837137763, 'samples': 2599424, 'steps': 5076, 'loss/train': 2.939244270324707} 02/24/2022 08:08:35 - INFO - codeparrot_training - Step 5077: {'lr': 0.0004949474112449424, 'samples': 2599936, 'steps': 5077, 'loss/train': 3.7290453910827637} 02/24/2022 08:08:38 - INFO - codeparrot_training - Step 5078: {'lr': 0.0004949441377268318, 'samples': 2600448, 'steps': 5078, 'loss/train': 2.5886638164520264} 02/24/2022 08:08:44 - INFO - codeparrot_training - Step 5079: {'lr': 0.0004949408631594582, 'samples': 2600960, 'steps': 5079, 'loss/train': 2.021549940109253} 02/24/2022 08:08:47 - INFO - codeparrot_training - Step 5080: {'lr': 0.0004949375875428357, 'samples': 2601472, 'steps': 5080, 'loss/train': 2.831801414489746} 02/24/2022 08:08:53 - INFO - codeparrot_training - Step 5081: {'lr': 0.0004949343108769784, 'samples': 2601984, 'steps': 5081, 'loss/train': 2.555068254470825} 02/24/2022 08:08:56 - INFO - codeparrot_training - Step 5082: {'lr': 0.0004949310331619002, 'samples': 2602496, 'steps': 5082, 'loss/train': 2.103515148162842} 02/24/2022 08:09:02 - INFO - codeparrot_training - Step 5083: {'lr': 0.0004949277543976153, 'samples': 2603008, 'steps': 5083, 'loss/train': 2.3618087768554688} 02/24/2022 08:09:05 - INFO - codeparrot_training - Step 5084: {'lr': 0.0004949244745841377, 'samples': 2603520, 'steps': 5084, 'loss/train': 3.752011775970459} 02/24/2022 08:09:11 - INFO - codeparrot_training - Step 5085: {'lr': 0.0004949211937214814, 'samples': 2604032, 'steps': 5085, 'loss/train': 1.4301141500473022} 02/24/2022 08:09:15 - INFO - codeparrot_training - Step 5086: {'lr': 0.0004949179118096604, 'samples': 2604544, 'steps': 5086, 'loss/train': 3.508284330368042} 02/24/2022 08:09:20 - INFO - codeparrot_training - Step 5087: {'lr': 0.0004949146288486889, 'samples': 2605056, 'steps': 5087, 'loss/train': 2.7847771644592285} 02/24/2022 08:09:24 - INFO - codeparrot_training - Step 5088: {'lr': 0.0004949113448385809, 'samples': 2605568, 'steps': 5088, 'loss/train': 1.6906352043151855} 02/24/2022 08:09:29 - INFO - codeparrot_training - Step 5089: {'lr': 0.0004949080597793505, 'samples': 2606080, 'steps': 5089, 'loss/train': 2.9376049041748047} 02/24/2022 08:09:33 - INFO - codeparrot_training - Step 5090: {'lr': 0.0004949047736710116, 'samples': 2606592, 'steps': 5090, 'loss/train': 2.1915385723114014} 02/24/2022 08:09:38 - INFO - codeparrot_training - Step 5091: {'lr': 0.0004949014865135786, 'samples': 2607104, 'steps': 5091, 'loss/train': 2.7079694271087646} 02/24/2022 08:09:41 - INFO - codeparrot_training - Step 5092: {'lr': 0.0004948981983070652, 'samples': 2607616, 'steps': 5092, 'loss/train': 1.000334620475769} 02/24/2022 08:09:48 - INFO - codeparrot_training - Step 5093: {'lr': 0.0004948949090514858, 'samples': 2608128, 'steps': 5093, 'loss/train': 1.335888385772705} 02/24/2022 08:09:51 - INFO - codeparrot_training - Step 5094: {'lr': 0.0004948916187468544, 'samples': 2608640, 'steps': 5094, 'loss/train': 2.8289735317230225} 02/24/2022 08:09:57 - INFO - codeparrot_training - Step 5095: {'lr': 0.000494888327393185, 'samples': 2609152, 'steps': 5095, 'loss/train': 2.2796788215637207} 02/24/2022 08:10:00 - INFO - codeparrot_training - Step 5096: {'lr': 0.0004948850349904919, 'samples': 2609664, 'steps': 5096, 'loss/train': 2.2301361560821533} 02/24/2022 08:10:06 - INFO - codeparrot_training - Step 5097: {'lr': 0.000494881741538789, 'samples': 2610176, 'steps': 5097, 'loss/train': 2.9230470657348633} 02/24/2022 08:10:09 - INFO - codeparrot_training - Step 5098: {'lr': 0.0004948784470380904, 'samples': 2610688, 'steps': 5098, 'loss/train': 2.715867757797241} 02/24/2022 08:10:15 - INFO - codeparrot_training - Step 5099: {'lr': 0.0004948751514884103, 'samples': 2611200, 'steps': 5099, 'loss/train': 2.704303503036499} 02/24/2022 08:10:18 - INFO - codeparrot_training - Step 5100: {'lr': 0.0004948718548897628, 'samples': 2611712, 'steps': 5100, 'loss/train': 3.723182201385498} 02/24/2022 08:10:24 - INFO - codeparrot_training - Step 5101: {'lr': 0.0004948685572421621, 'samples': 2612224, 'steps': 5101, 'loss/train': 2.602644443511963} 02/24/2022 08:10:27 - INFO - codeparrot_training - Step 5102: {'lr': 0.0004948652585456222, 'samples': 2612736, 'steps': 5102, 'loss/train': 2.9009594917297363} 02/24/2022 08:10:33 - INFO - codeparrot_training - Step 5103: {'lr': 0.0004948619588001574, 'samples': 2613248, 'steps': 5103, 'loss/train': 1.9833265542984009} 02/24/2022 08:10:37 - INFO - codeparrot_training - Step 5104: {'lr': 0.0004948586580057816, 'samples': 2613760, 'steps': 5104, 'loss/train': 1.3734633922576904} 02/24/2022 08:10:42 - INFO - codeparrot_training - Step 5105: {'lr': 0.0004948553561625091, 'samples': 2614272, 'steps': 5105, 'loss/train': 2.327197790145874} 02/24/2022 08:10:46 - INFO - codeparrot_training - Step 5106: {'lr': 0.000494852053270354, 'samples': 2614784, 'steps': 5106, 'loss/train': 1.6684679985046387} 02/24/2022 08:10:51 - INFO - codeparrot_training - Step 5107: {'lr': 0.0004948487493293305, 'samples': 2615296, 'steps': 5107, 'loss/train': 2.6789066791534424} 02/24/2022 08:10:55 - INFO - codeparrot_training - Step 5108: {'lr': 0.0004948454443394527, 'samples': 2615808, 'steps': 5108, 'loss/train': 3.3280553817749023} 02/24/2022 08:11:01 - INFO - codeparrot_training - Step 5109: {'lr': 0.0004948421383007347, 'samples': 2616320, 'steps': 5109, 'loss/train': 4.2329487800598145} 02/24/2022 08:11:04 - INFO - codeparrot_training - Step 5110: {'lr': 0.0004948388312131908, 'samples': 2616832, 'steps': 5110, 'loss/train': 1.8370198011398315} 02/24/2022 08:11:10 - INFO - codeparrot_training - Step 5111: {'lr': 0.0004948355230768349, 'samples': 2617344, 'steps': 5111, 'loss/train': 1.8667925596237183} 02/24/2022 08:11:13 - INFO - codeparrot_training - Step 5112: {'lr': 0.0004948322138916816, 'samples': 2617856, 'steps': 5112, 'loss/train': 2.9620187282562256} 02/24/2022 08:11:19 - INFO - codeparrot_training - Step 5113: {'lr': 0.0004948289036577447, 'samples': 2618368, 'steps': 5113, 'loss/train': 2.476203680038452} 02/24/2022 08:11:23 - INFO - codeparrot_training - Step 5114: {'lr': 0.0004948255923750385, 'samples': 2618880, 'steps': 5114, 'loss/train': 2.6586132049560547} 02/24/2022 08:11:28 - INFO - codeparrot_training - Step 5115: {'lr': 0.0004948222800435773, 'samples': 2619392, 'steps': 5115, 'loss/train': 4.017131805419922} 02/24/2022 08:11:32 - INFO - codeparrot_training - Step 5116: {'lr': 0.0004948189666633752, 'samples': 2619904, 'steps': 5116, 'loss/train': 1.6821726560592651} 02/24/2022 08:11:37 - INFO - codeparrot_training - Step 5117: {'lr': 0.0004948156522344463, 'samples': 2620416, 'steps': 5117, 'loss/train': 1.9291718006134033} 02/24/2022 08:11:41 - INFO - codeparrot_training - Step 5118: {'lr': 0.0004948123367568049, 'samples': 2620928, 'steps': 5118, 'loss/train': 3.2326972484588623} 02/24/2022 08:11:46 - INFO - codeparrot_training - Step 5119: {'lr': 0.0004948090202304652, 'samples': 2621440, 'steps': 5119, 'loss/train': 2.32305645942688} 02/24/2022 08:11:50 - INFO - codeparrot_training - Step 5120: {'lr': 0.0004948057026554415, 'samples': 2621952, 'steps': 5120, 'loss/train': 2.8522584438323975} 02/24/2022 08:11:55 - INFO - codeparrot_training - Step 5121: {'lr': 0.0004948023840317477, 'samples': 2622464, 'steps': 5121, 'loss/train': 2.4217073917388916} 02/24/2022 08:11:59 - INFO - codeparrot_training - Step 5122: {'lr': 0.0004947990643593983, 'samples': 2622976, 'steps': 5122, 'loss/train': 2.274604320526123} 02/24/2022 08:12:05 - INFO - codeparrot_training - Step 5123: {'lr': 0.0004947957436384076, 'samples': 2623488, 'steps': 5123, 'loss/train': 2.744251012802124} 02/24/2022 08:12:08 - INFO - codeparrot_training - Step 5124: {'lr': 0.0004947924218687894, 'samples': 2624000, 'steps': 5124, 'loss/train': 3.4140748977661133} 02/24/2022 08:12:14 - INFO - codeparrot_training - Step 5125: {'lr': 0.0004947890990505585, 'samples': 2624512, 'steps': 5125, 'loss/train': 2.790090799331665} 02/24/2022 08:12:17 - INFO - codeparrot_training - Step 5126: {'lr': 0.0004947857751837286, 'samples': 2625024, 'steps': 5126, 'loss/train': 2.408137798309326} 02/24/2022 08:12:23 - INFO - codeparrot_training - Step 5127: {'lr': 0.0004947824502683142, 'samples': 2625536, 'steps': 5127, 'loss/train': 2.5618083477020264} 02/24/2022 08:12:26 - INFO - codeparrot_training - Step 5128: {'lr': 0.0004947791243043296, 'samples': 2626048, 'steps': 5128, 'loss/train': 1.8575185537338257} 02/24/2022 08:12:32 - INFO - codeparrot_training - Step 5129: {'lr': 0.0004947757972917889, 'samples': 2626560, 'steps': 5129, 'loss/train': 2.426994800567627} 02/24/2022 08:12:35 - INFO - codeparrot_training - Step 5130: {'lr': 0.0004947724692307064, 'samples': 2627072, 'steps': 5130, 'loss/train': 2.640988826751709} 02/24/2022 08:12:41 - INFO - codeparrot_training - Step 5131: {'lr': 0.0004947691401210963, 'samples': 2627584, 'steps': 5131, 'loss/train': 2.1888606548309326} 02/24/2022 08:12:44 - INFO - codeparrot_training - Step 5132: {'lr': 0.0004947658099629731, 'samples': 2628096, 'steps': 5132, 'loss/train': 3.2527060508728027} 02/24/2022 08:12:50 - INFO - codeparrot_training - Step 5133: {'lr': 0.0004947624787563507, 'samples': 2628608, 'steps': 5133, 'loss/train': 2.5574615001678467} 02/24/2022 08:12:54 - INFO - codeparrot_training - Step 5134: {'lr': 0.0004947591465012436, 'samples': 2629120, 'steps': 5134, 'loss/train': 3.0218966007232666} 02/24/2022 08:12:59 - INFO - codeparrot_training - Step 5135: {'lr': 0.0004947558131976661, 'samples': 2629632, 'steps': 5135, 'loss/train': 2.7285172939300537} 02/24/2022 08:13:03 - INFO - codeparrot_training - Step 5136: {'lr': 0.0004947524788456324, 'samples': 2630144, 'steps': 5136, 'loss/train': 3.1026570796966553} 02/24/2022 08:13:08 - INFO - codeparrot_training - Step 5137: {'lr': 0.0004947491434451569, 'samples': 2630656, 'steps': 5137, 'loss/train': 2.9265973567962646} 02/24/2022 08:13:12 - INFO - codeparrot_training - Step 5138: {'lr': 0.0004947458069962537, 'samples': 2631168, 'steps': 5138, 'loss/train': 1.492129921913147} 02/24/2022 08:13:18 - INFO - codeparrot_training - Step 5139: {'lr': 0.0004947424694989371, 'samples': 2631680, 'steps': 5139, 'loss/train': 2.335777997970581} 02/24/2022 08:13:21 - INFO - codeparrot_training - Step 5140: {'lr': 0.0004947391309532216, 'samples': 2632192, 'steps': 5140, 'loss/train': 3.307926893234253} 02/24/2022 08:13:27 - INFO - codeparrot_training - Step 5141: {'lr': 0.0004947357913591213, 'samples': 2632704, 'steps': 5141, 'loss/train': 1.5015305280685425} 02/24/2022 08:13:30 - INFO - codeparrot_training - Step 5142: {'lr': 0.0004947324507166505, 'samples': 2633216, 'steps': 5142, 'loss/train': 1.8070907592773438} 02/24/2022 08:13:36 - INFO - codeparrot_training - Step 5143: {'lr': 0.0004947291090258238, 'samples': 2633728, 'steps': 5143, 'loss/train': 2.8901987075805664} 02/24/2022 08:13:39 - INFO - codeparrot_training - Step 5144: {'lr': 0.0004947257662866551, 'samples': 2634240, 'steps': 5144, 'loss/train': 3.076181650161743} 02/24/2022 08:13:45 - INFO - codeparrot_training - Step 5145: {'lr': 0.0004947224224991591, 'samples': 2634752, 'steps': 5145, 'loss/train': 2.332066774368286} 02/24/2022 08:13:48 - INFO - codeparrot_training - Step 5146: {'lr': 0.0004947190776633499, 'samples': 2635264, 'steps': 5146, 'loss/train': 2.256387710571289} 02/24/2022 08:13:54 - INFO - codeparrot_training - Step 5147: {'lr': 0.0004947157317792418, 'samples': 2635776, 'steps': 5147, 'loss/train': 2.681453227996826} 02/24/2022 08:13:57 - INFO - codeparrot_training - Step 5148: {'lr': 0.0004947123848468493, 'samples': 2636288, 'steps': 5148, 'loss/train': 2.7483694553375244} 02/24/2022 08:14:03 - INFO - codeparrot_training - Step 5149: {'lr': 0.0004947090368661866, 'samples': 2636800, 'steps': 5149, 'loss/train': 1.2899115085601807} 02/24/2022 08:14:07 - INFO - codeparrot_training - Step 5150: {'lr': 0.0004947056878372681, 'samples': 2637312, 'steps': 5150, 'loss/train': 1.7127106189727783} 02/24/2022 08:14:12 - INFO - codeparrot_training - Step 5151: {'lr': 0.0004947023377601082, 'samples': 2637824, 'steps': 5151, 'loss/train': 2.629730701446533} 02/24/2022 08:14:16 - INFO - codeparrot_training - Step 5152: {'lr': 0.0004946989866347211, 'samples': 2638336, 'steps': 5152, 'loss/train': 1.9000917673110962} 02/24/2022 08:14:21 - INFO - codeparrot_training - Step 5153: {'lr': 0.0004946956344611212, 'samples': 2638848, 'steps': 5153, 'loss/train': 3.3763539791107178} 02/24/2022 08:14:27 - INFO - codeparrot_training - Step 5154: {'lr': 0.000494692281239323, 'samples': 2639360, 'steps': 5154, 'loss/train': 2.3680930137634277} 02/24/2022 08:14:30 - INFO - codeparrot_training - Step 5155: {'lr': 0.0004946889269693408, 'samples': 2639872, 'steps': 5155, 'loss/train': 2.266359567642212} 02/24/2022 08:14:36 - INFO - codeparrot_training - Step 5156: {'lr': 0.0004946855716511888, 'samples': 2640384, 'steps': 5156, 'loss/train': 3.2081191539764404} 02/24/2022 08:14:39 - INFO - codeparrot_training - Step 5157: {'lr': 0.0004946822152848816, 'samples': 2640896, 'steps': 5157, 'loss/train': 2.1639814376831055} 02/24/2022 08:14:45 - INFO - codeparrot_training - Step 5158: {'lr': 0.0004946788578704335, 'samples': 2641408, 'steps': 5158, 'loss/train': 2.309159994125366} 02/24/2022 08:14:49 - INFO - codeparrot_training - Step 5159: {'lr': 0.0004946754994078588, 'samples': 2641920, 'steps': 5159, 'loss/train': 1.926787257194519} 02/24/2022 08:14:54 - INFO - codeparrot_training - Step 5160: {'lr': 0.000494672139897172, 'samples': 2642432, 'steps': 5160, 'loss/train': 3.462728261947632} 02/24/2022 08:14:58 - INFO - codeparrot_training - Step 5161: {'lr': 0.0004946687793383874, 'samples': 2642944, 'steps': 5161, 'loss/train': 2.4011433124542236} 02/24/2022 08:15:01 - INFO - codeparrot_training - Step 5162: {'lr': 0.0004946654177315194, 'samples': 2643456, 'steps': 5162, 'loss/train': 0.5060781240463257} 02/24/2022 08:15:07 - INFO - codeparrot_training - Step 5163: {'lr': 0.0004946620550765826, 'samples': 2643968, 'steps': 5163, 'loss/train': 2.1764190196990967} 02/24/2022 08:15:10 - INFO - codeparrot_training - Step 5164: {'lr': 0.0004946586913735911, 'samples': 2644480, 'steps': 5164, 'loss/train': 1.2177103757858276} 02/24/2022 08:15:16 - INFO - codeparrot_training - Step 5165: {'lr': 0.0004946553266225595, 'samples': 2644992, 'steps': 5165, 'loss/train': 2.6867284774780273} 02/24/2022 08:15:21 - INFO - codeparrot_training - Step 5166: {'lr': 0.0004946519608235022, 'samples': 2645504, 'steps': 5166, 'loss/train': 2.9138760566711426} 02/24/2022 08:15:25 - INFO - codeparrot_training - Step 5167: {'lr': 0.0004946485939764336, 'samples': 2646016, 'steps': 5167, 'loss/train': 2.5797221660614014} 02/24/2022 08:15:28 - INFO - codeparrot_training - Step 5168: {'lr': 0.000494645226081368, 'samples': 2646528, 'steps': 5168, 'loss/train': 3.369985342025757} 02/24/2022 08:15:35 - INFO - codeparrot_training - Step 5169: {'lr': 0.00049464185713832, 'samples': 2647040, 'steps': 5169, 'loss/train': 2.85191011428833} 02/24/2022 08:15:40 - INFO - codeparrot_training - Step 5170: {'lr': 0.000494638487147304, 'samples': 2647552, 'steps': 5170, 'loss/train': 2.664747953414917} 02/24/2022 08:15:44 - INFO - codeparrot_training - Step 5171: {'lr': 0.0004946351161083344, 'samples': 2648064, 'steps': 5171, 'loss/train': 2.5004470348358154} 02/24/2022 08:15:50 - INFO - codeparrot_training - Step 5172: {'lr': 0.0004946317440214257, 'samples': 2648576, 'steps': 5172, 'loss/train': 2.210859775543213} 02/24/2022 08:15:53 - INFO - codeparrot_training - Step 5173: {'lr': 0.000494628370886592, 'samples': 2649088, 'steps': 5173, 'loss/train': 2.713085412979126} 02/24/2022 08:15:57 - INFO - codeparrot_training - Step 5174: {'lr': 0.0004946249967038483, 'samples': 2649600, 'steps': 5174, 'loss/train': 2.681502342224121} 02/24/2022 08:16:02 - INFO - codeparrot_training - Step 5175: {'lr': 0.0004946216214732088, 'samples': 2650112, 'steps': 5175, 'loss/train': 2.96237850189209} 02/24/2022 08:16:06 - INFO - codeparrot_training - Step 5176: {'lr': 0.0004946182451946878, 'samples': 2650624, 'steps': 5176, 'loss/train': 1.3448158502578735} 02/24/2022 08:16:11 - INFO - codeparrot_training - Step 5177: {'lr': 0.0004946148678683001, 'samples': 2651136, 'steps': 5177, 'loss/train': 1.241605520248413} 02/24/2022 08:16:15 - INFO - codeparrot_training - Step 5178: {'lr': 0.0004946114894940599, 'samples': 2651648, 'steps': 5178, 'loss/train': 2.018709182739258} 02/24/2022 08:16:20 - INFO - codeparrot_training - Step 5179: {'lr': 0.0004946081100719817, 'samples': 2652160, 'steps': 5179, 'loss/train': 2.4574363231658936} 02/24/2022 08:16:24 - INFO - codeparrot_training - Step 5180: {'lr': 0.00049460472960208, 'samples': 2652672, 'steps': 5180, 'loss/train': 3.193479061126709} 02/24/2022 08:16:29 - INFO - codeparrot_training - Step 5181: {'lr': 0.0004946013480843694, 'samples': 2653184, 'steps': 5181, 'loss/train': 1.2273881435394287} 02/24/2022 08:16:33 - INFO - codeparrot_training - Step 5182: {'lr': 0.0004945979655188642, 'samples': 2653696, 'steps': 5182, 'loss/train': 1.6777675151824951} 02/24/2022 08:16:38 - INFO - codeparrot_training - Step 5183: {'lr': 0.0004945945819055791, 'samples': 2654208, 'steps': 5183, 'loss/train': 1.6399005651474} 02/24/2022 08:16:42 - INFO - codeparrot_training - Step 5184: {'lr': 0.0004945911972445284, 'samples': 2654720, 'steps': 5184, 'loss/train': 2.7347631454467773} 02/24/2022 08:16:48 - INFO - codeparrot_training - Step 5185: {'lr': 0.0004945878115357267, 'samples': 2655232, 'steps': 5185, 'loss/train': 2.2540037631988525} 02/24/2022 08:16:51 - INFO - codeparrot_training - Step 5186: {'lr': 0.0004945844247791886, 'samples': 2655744, 'steps': 5186, 'loss/train': 8.276962280273438} 02/24/2022 08:16:58 - INFO - codeparrot_training - Step 5187: {'lr': 0.0004945810369749283, 'samples': 2656256, 'steps': 5187, 'loss/train': 4.214089393615723} 02/24/2022 08:17:01 - INFO - codeparrot_training - Step 5188: {'lr': 0.0004945776481229605, 'samples': 2656768, 'steps': 5188, 'loss/train': 1.785870909690857} 02/24/2022 08:17:04 - INFO - codeparrot_training - Step 5189: {'lr': 0.0004945742582232999, 'samples': 2657280, 'steps': 5189, 'loss/train': 2.70741868019104} 02/24/2022 08:17:10 - INFO - codeparrot_training - Step 5190: {'lr': 0.0004945708672759606, 'samples': 2657792, 'steps': 5190, 'loss/train': 2.295666217803955} 02/24/2022 08:17:13 - INFO - codeparrot_training - Step 5191: {'lr': 0.0004945674752809575, 'samples': 2658304, 'steps': 5191, 'loss/train': 3.018218517303467} 02/24/2022 08:17:19 - INFO - codeparrot_training - Step 5192: {'lr': 0.000494564082238305, 'samples': 2658816, 'steps': 5192, 'loss/train': 1.5652316808700562} 02/24/2022 08:17:22 - INFO - codeparrot_training - Step 5193: {'lr': 0.0004945606881480176, 'samples': 2659328, 'steps': 5193, 'loss/train': 2.667050361633301} 02/24/2022 08:17:28 - INFO - codeparrot_training - Step 5194: {'lr': 0.0004945572930101098, 'samples': 2659840, 'steps': 5194, 'loss/train': 2.329467535018921} 02/24/2022 08:17:31 - INFO - codeparrot_training - Step 5195: {'lr': 0.0004945538968245964, 'samples': 2660352, 'steps': 5195, 'loss/train': 2.79240083694458} 02/24/2022 08:17:37 - INFO - codeparrot_training - Step 5196: {'lr': 0.0004945504995914917, 'samples': 2660864, 'steps': 5196, 'loss/train': 1.2632068395614624} 02/24/2022 08:17:40 - INFO - codeparrot_training - Step 5197: {'lr': 0.0004945471013108102, 'samples': 2661376, 'steps': 5197, 'loss/train': 2.8748514652252197} 02/24/2022 08:17:46 - INFO - codeparrot_training - Step 5198: {'lr': 0.0004945437019825668, 'samples': 2661888, 'steps': 5198, 'loss/train': 1.6277357339859009} 02/24/2022 08:17:50 - INFO - codeparrot_training - Step 5199: {'lr': 0.0004945403016067756, 'samples': 2662400, 'steps': 5199, 'loss/train': 2.3230233192443848} 02/24/2022 08:17:56 - INFO - codeparrot_training - Step 5200: {'lr': 0.0004945369001834514, 'samples': 2662912, 'steps': 5200, 'loss/train': 2.9772768020629883} 02/24/2022 08:17:59 - INFO - codeparrot_training - Step 5201: {'lr': 0.0004945334977126089, 'samples': 2663424, 'steps': 5201, 'loss/train': 1.9875446557998657} 02/24/2022 08:18:05 - INFO - codeparrot_training - Step 5202: {'lr': 0.0004945300941942624, 'samples': 2663936, 'steps': 5202, 'loss/train': 2.912196397781372} 02/24/2022 08:18:08 - INFO - codeparrot_training - Step 5203: {'lr': 0.0004945266896284268, 'samples': 2664448, 'steps': 5203, 'loss/train': 3.1419756412506104} 02/24/2022 08:18:14 - INFO - codeparrot_training - Step 5204: {'lr': 0.0004945232840151164, 'samples': 2664960, 'steps': 5204, 'loss/train': 2.4587762355804443} 02/24/2022 08:18:17 - INFO - codeparrot_training - Step 5205: {'lr': 0.0004945198773543459, 'samples': 2665472, 'steps': 5205, 'loss/train': 2.140803098678589} 02/24/2022 08:18:23 - INFO - codeparrot_training - Step 5206: {'lr': 0.0004945164696461299, 'samples': 2665984, 'steps': 5206, 'loss/train': 2.219256639480591} 02/24/2022 08:18:26 - INFO - codeparrot_training - Step 5207: {'lr': 0.000494513060890483, 'samples': 2666496, 'steps': 5207, 'loss/train': 2.2424747943878174} 02/24/2022 08:18:32 - INFO - codeparrot_training - Step 5208: {'lr': 0.0004945096510874197, 'samples': 2667008, 'steps': 5208, 'loss/train': 2.0000693798065186} 02/24/2022 08:18:36 - INFO - codeparrot_training - Step 5209: {'lr': 0.0004945062402369548, 'samples': 2667520, 'steps': 5209, 'loss/train': 3.330228805541992} 02/24/2022 08:18:42 - INFO - codeparrot_training - Step 5210: {'lr': 0.0004945028283391028, 'samples': 2668032, 'steps': 5210, 'loss/train': 0.6820445656776428} 02/24/2022 08:18:45 - INFO - codeparrot_training - Step 5211: {'lr': 0.0004944994153938783, 'samples': 2668544, 'steps': 5211, 'loss/train': 2.586907386779785} 02/24/2022 08:18:51 - INFO - codeparrot_training - Step 5212: {'lr': 0.0004944960014012959, 'samples': 2669056, 'steps': 5212, 'loss/train': 3.126750946044922} 02/24/2022 08:18:54 - INFO - codeparrot_training - Step 5213: {'lr': 0.0004944925863613704, 'samples': 2669568, 'steps': 5213, 'loss/train': 1.5274075269699097} 02/24/2022 08:19:00 - INFO - codeparrot_training - Step 5214: {'lr': 0.0004944891702741161, 'samples': 2670080, 'steps': 5214, 'loss/train': 3.2495155334472656} 02/24/2022 08:19:03 - INFO - codeparrot_training - Step 5215: {'lr': 0.0004944857531395479, 'samples': 2670592, 'steps': 5215, 'loss/train': 2.186718702316284} 02/24/2022 08:19:09 - INFO - codeparrot_training - Step 5216: {'lr': 0.0004944823349576805, 'samples': 2671104, 'steps': 5216, 'loss/train': 2.643941879272461} 02/24/2022 08:19:13 - INFO - codeparrot_training - Step 5217: {'lr': 0.0004944789157285283, 'samples': 2671616, 'steps': 5217, 'loss/train': 2.964428424835205} 02/24/2022 08:19:18 - INFO - codeparrot_training - Step 5218: {'lr': 0.0004944754954521061, 'samples': 2672128, 'steps': 5218, 'loss/train': 3.1117196083068848} 02/24/2022 08:19:22 - INFO - codeparrot_training - Step 5219: {'lr': 0.0004944720741284285, 'samples': 2672640, 'steps': 5219, 'loss/train': 3.036386251449585} 02/24/2022 08:19:27 - INFO - codeparrot_training - Step 5220: {'lr': 0.00049446865175751, 'samples': 2673152, 'steps': 5220, 'loss/train': 2.484792709350586} 02/24/2022 08:19:31 - INFO - codeparrot_training - Step 5221: {'lr': 0.0004944652283393656, 'samples': 2673664, 'steps': 5221, 'loss/train': 2.8877463340759277} 02/24/2022 08:19:36 - INFO - codeparrot_training - Step 5222: {'lr': 0.0004944618038740098, 'samples': 2674176, 'steps': 5222, 'loss/train': 2.831866502761841} 02/24/2022 08:19:40 - INFO - codeparrot_training - Step 5223: {'lr': 0.0004944583783614571, 'samples': 2674688, 'steps': 5223, 'loss/train': 2.8418331146240234} 02/24/2022 08:19:45 - INFO - codeparrot_training - Step 5224: {'lr': 0.0004944549518017225, 'samples': 2675200, 'steps': 5224, 'loss/train': 3.592921257019043} 02/24/2022 08:19:49 - INFO - codeparrot_training - Step 5225: {'lr': 0.0004944515241948204, 'samples': 2675712, 'steps': 5225, 'loss/train': 2.6743624210357666} 02/24/2022 08:19:54 - INFO - codeparrot_training - Step 5226: {'lr': 0.0004944480955407656, 'samples': 2676224, 'steps': 5226, 'loss/train': 1.7635077238082886} 02/24/2022 08:19:58 - INFO - codeparrot_training - Step 5227: {'lr': 0.0004944446658395728, 'samples': 2676736, 'steps': 5227, 'loss/train': 2.057612895965576} 02/24/2022 08:20:03 - INFO - codeparrot_training - Step 5228: {'lr': 0.0004944412350912567, 'samples': 2677248, 'steps': 5228, 'loss/train': 1.861214518547058} 02/24/2022 08:20:07 - INFO - codeparrot_training - Step 5229: {'lr': 0.000494437803295832, 'samples': 2677760, 'steps': 5229, 'loss/train': 2.5382578372955322} 02/24/2022 08:20:12 - INFO - codeparrot_training - Step 5230: {'lr': 0.0004944343704533133, 'samples': 2678272, 'steps': 5230, 'loss/train': 1.9002410173416138} 02/24/2022 08:20:16 - INFO - codeparrot_training - Step 5231: {'lr': 0.0004944309365637154, 'samples': 2678784, 'steps': 5231, 'loss/train': 2.0640037059783936} 02/24/2022 08:20:22 - INFO - codeparrot_training - Step 5232: {'lr': 0.000494427501627053, 'samples': 2679296, 'steps': 5232, 'loss/train': 5.441315174102783} 02/24/2022 08:20:25 - INFO - codeparrot_training - Step 5233: {'lr': 0.0004944240656433407, 'samples': 2679808, 'steps': 5233, 'loss/train': 1.594360589981079} 02/24/2022 08:20:31 - INFO - codeparrot_training - Step 5234: {'lr': 0.0004944206286125935, 'samples': 2680320, 'steps': 5234, 'loss/train': 1.448689579963684} 02/24/2022 08:20:34 - INFO - codeparrot_training - Step 5235: {'lr': 0.0004944171905348258, 'samples': 2680832, 'steps': 5235, 'loss/train': 8.80094051361084} 02/24/2022 08:20:40 - INFO - codeparrot_training - Step 5236: {'lr': 0.0004944137514100525, 'samples': 2681344, 'steps': 5236, 'loss/train': 5.9965009689331055} 02/24/2022 08:20:43 - INFO - codeparrot_training - Step 5237: {'lr': 0.0004944103112382883, 'samples': 2681856, 'steps': 5237, 'loss/train': 2.855015993118286} 02/24/2022 08:20:49 - INFO - codeparrot_training - Step 5238: {'lr': 0.0004944068700195479, 'samples': 2682368, 'steps': 5238, 'loss/train': 3.7001540660858154} 02/24/2022 08:20:52 - INFO - codeparrot_training - Step 5239: {'lr': 0.0004944034277538462, 'samples': 2682880, 'steps': 5239, 'loss/train': 1.8518186807632446} 02/24/2022 08:20:58 - INFO - codeparrot_training - Step 5240: {'lr': 0.0004943999844411977, 'samples': 2683392, 'steps': 5240, 'loss/train': 3.2561490535736084} 02/24/2022 08:21:01 - INFO - codeparrot_training - Step 5241: {'lr': 0.0004943965400816173, 'samples': 2683904, 'steps': 5241, 'loss/train': 2.6126956939697266} 02/24/2022 08:21:07 - INFO - codeparrot_training - Step 5242: {'lr': 0.0004943930946751197, 'samples': 2684416, 'steps': 5242, 'loss/train': 2.2587852478027344} 02/24/2022 08:21:11 - INFO - codeparrot_training - Step 5243: {'lr': 0.0004943896482217197, 'samples': 2684928, 'steps': 5243, 'loss/train': 2.234565496444702} 02/24/2022 08:21:17 - INFO - codeparrot_training - Step 5244: {'lr': 0.0004943862007214322, 'samples': 2685440, 'steps': 5244, 'loss/train': 2.7044122219085693} 02/24/2022 08:21:20 - INFO - codeparrot_training - Step 5245: {'lr': 0.0004943827521742716, 'samples': 2685952, 'steps': 5245, 'loss/train': 2.0423200130462646} 02/24/2022 08:21:26 - INFO - codeparrot_training - Step 5246: {'lr': 0.000494379302580253, 'samples': 2686464, 'steps': 5246, 'loss/train': 2.8277180194854736} 02/24/2022 08:21:29 - INFO - codeparrot_training - Step 5247: {'lr': 0.000494375851939391, 'samples': 2686976, 'steps': 5247, 'loss/train': 2.1107726097106934} 02/24/2022 08:21:35 - INFO - codeparrot_training - Step 5248: {'lr': 0.0004943724002517005, 'samples': 2687488, 'steps': 5248, 'loss/train': 1.5515531301498413} 02/24/2022 08:21:38 - INFO - codeparrot_training - Step 5249: {'lr': 0.0004943689475171962, 'samples': 2688000, 'steps': 5249, 'loss/train': 2.673150062561035} 02/24/2022 08:21:44 - INFO - codeparrot_training - Step 5250: {'lr': 0.000494365493735893, 'samples': 2688512, 'steps': 5250, 'loss/train': 4.120074272155762} 02/24/2022 08:21:47 - INFO - codeparrot_training - Step 5251: {'lr': 0.0004943620389078055, 'samples': 2689024, 'steps': 5251, 'loss/train': 2.5042974948883057} 02/24/2022 08:21:53 - INFO - codeparrot_training - Step 5252: {'lr': 0.0004943585830329487, 'samples': 2689536, 'steps': 5252, 'loss/train': 1.8689889907836914} 02/24/2022 08:21:56 - INFO - codeparrot_training - Step 5253: {'lr': 0.0004943551261113373, 'samples': 2690048, 'steps': 5253, 'loss/train': 3.1838936805725098} 02/24/2022 08:22:02 - INFO - codeparrot_training - Step 5254: {'lr': 0.0004943516681429861, 'samples': 2690560, 'steps': 5254, 'loss/train': 3.069115400314331} 02/24/2022 08:22:06 - INFO - codeparrot_training - Step 5255: {'lr': 0.0004943482091279101, 'samples': 2691072, 'steps': 5255, 'loss/train': 1.9096789360046387} 02/24/2022 08:22:11 - INFO - codeparrot_training - Step 5256: {'lr': 0.0004943447490661238, 'samples': 2691584, 'steps': 5256, 'loss/train': 2.5676307678222656} 02/24/2022 08:22:15 - INFO - codeparrot_training - Step 5257: {'lr': 0.0004943412879576422, 'samples': 2692096, 'steps': 5257, 'loss/train': 3.0552666187286377} 02/24/2022 08:22:20 - INFO - codeparrot_training - Step 5258: {'lr': 0.0004943378258024802, 'samples': 2692608, 'steps': 5258, 'loss/train': 2.801168918609619} 02/24/2022 08:22:24 - INFO - codeparrot_training - Step 5259: {'lr': 0.0004943343626006524, 'samples': 2693120, 'steps': 5259, 'loss/train': 2.228573799133301} 02/24/2022 08:22:29 - INFO - codeparrot_training - Step 5260: {'lr': 0.000494330898352174, 'samples': 2693632, 'steps': 5260, 'loss/train': 2.013396978378296} 02/24/2022 08:22:33 - INFO - codeparrot_training - Step 5261: {'lr': 0.0004943274330570594, 'samples': 2694144, 'steps': 5261, 'loss/train': 2.8861701488494873} 02/24/2022 08:22:38 - INFO - codeparrot_training - Step 5262: {'lr': 0.0004943239667153237, 'samples': 2694656, 'steps': 5262, 'loss/train': 2.649702548980713} 02/24/2022 08:22:42 - INFO - codeparrot_training - Step 5263: {'lr': 0.0004943204993269818, 'samples': 2695168, 'steps': 5263, 'loss/train': 2.612884283065796} 02/24/2022 08:22:48 - INFO - codeparrot_training - Step 5264: {'lr': 0.0004943170308920483, 'samples': 2695680, 'steps': 5264, 'loss/train': 2.069972038269043} 02/24/2022 08:22:51 - INFO - codeparrot_training - Step 5265: {'lr': 0.0004943135614105384, 'samples': 2696192, 'steps': 5265, 'loss/train': 3.1454107761383057} 02/24/2022 08:22:57 - INFO - codeparrot_training - Step 5266: {'lr': 0.0004943100908824667, 'samples': 2696704, 'steps': 5266, 'loss/train': 2.6592893600463867} 02/24/2022 08:23:00 - INFO - codeparrot_training - Step 5267: {'lr': 0.0004943066193078482, 'samples': 2697216, 'steps': 5267, 'loss/train': 2.364511489868164} 02/24/2022 08:23:06 - INFO - codeparrot_training - Step 5268: {'lr': 0.0004943031466866976, 'samples': 2697728, 'steps': 5268, 'loss/train': 2.4398560523986816} 02/24/2022 08:23:09 - INFO - codeparrot_training - Step 5269: {'lr': 0.00049429967301903, 'samples': 2698240, 'steps': 5269, 'loss/train': 2.7816364765167236} 02/24/2022 08:23:15 - INFO - codeparrot_training - Step 5270: {'lr': 0.0004942961983048601, 'samples': 2698752, 'steps': 5270, 'loss/train': 2.5790302753448486} 02/24/2022 08:23:20 - INFO - codeparrot_training - Step 5271: {'lr': 0.0004942927225442029, 'samples': 2699264, 'steps': 5271, 'loss/train': 2.625483989715576} 02/24/2022 08:23:24 - INFO - codeparrot_training - Step 5272: {'lr': 0.0004942892457370732, 'samples': 2699776, 'steps': 5272, 'loss/train': 2.663005828857422} 02/24/2022 08:23:29 - INFO - codeparrot_training - Step 5273: {'lr': 0.000494285767883486, 'samples': 2700288, 'steps': 5273, 'loss/train': 2.2452316284179688} 02/24/2022 08:23:33 - INFO - codeparrot_training - Step 5274: {'lr': 0.0004942822889834562, 'samples': 2700800, 'steps': 5274, 'loss/train': 1.6624966859817505} 02/24/2022 08:23:38 - INFO - codeparrot_training - Step 5275: {'lr': 0.0004942788090369985, 'samples': 2701312, 'steps': 5275, 'loss/train': 2.4746954441070557} 02/24/2022 08:23:42 - INFO - codeparrot_training - Step 5276: {'lr': 0.0004942753280441281, 'samples': 2701824, 'steps': 5276, 'loss/train': 1.6270679235458374} 02/24/2022 08:23:47 - INFO - codeparrot_training - Step 5277: {'lr': 0.0004942718460048596, 'samples': 2702336, 'steps': 5277, 'loss/train': 2.141404628753662} 02/24/2022 08:23:51 - INFO - codeparrot_training - Step 5278: {'lr': 0.0004942683629192082, 'samples': 2702848, 'steps': 5278, 'loss/train': 1.053605079650879} 02/24/2022 08:23:57 - INFO - codeparrot_training - Step 5279: {'lr': 0.0004942648787871886, 'samples': 2703360, 'steps': 5279, 'loss/train': 1.9970529079437256} 02/24/2022 08:24:00 - INFO - codeparrot_training - Step 5280: {'lr': 0.000494261393608816, 'samples': 2703872, 'steps': 5280, 'loss/train': 1.7869811058044434} 02/24/2022 08:24:06 - INFO - codeparrot_training - Step 5281: {'lr': 0.0004942579073841049, 'samples': 2704384, 'steps': 5281, 'loss/train': 1.0113152265548706} 02/24/2022 08:24:09 - INFO - codeparrot_training - Step 5282: {'lr': 0.0004942544201130706, 'samples': 2704896, 'steps': 5282, 'loss/train': 1.235727310180664} 02/24/2022 08:24:15 - INFO - codeparrot_training - Step 5283: {'lr': 0.000494250931795728, 'samples': 2705408, 'steps': 5283, 'loss/train': 0.6270483136177063} 02/24/2022 08:24:18 - INFO - codeparrot_training - Step 5284: {'lr': 0.0004942474424320919, 'samples': 2705920, 'steps': 5284, 'loss/train': 1.5817296504974365} 02/24/2022 08:24:24 - INFO - codeparrot_training - Step 5285: {'lr': 0.0004942439520221774, 'samples': 2706432, 'steps': 5285, 'loss/train': 2.4787728786468506} 02/24/2022 08:24:27 - INFO - codeparrot_training - Step 5286: {'lr': 0.0004942404605659991, 'samples': 2706944, 'steps': 5286, 'loss/train': 2.474702835083008} 02/24/2022 08:24:33 - INFO - codeparrot_training - Step 5287: {'lr': 0.0004942369680635724, 'samples': 2707456, 'steps': 5287, 'loss/train': 1.1714990139007568} 02/24/2022 08:24:36 - INFO - codeparrot_training - Step 5288: {'lr': 0.0004942334745149122, 'samples': 2707968, 'steps': 5288, 'loss/train': 2.6758017539978027} 02/24/2022 08:24:43 - INFO - codeparrot_training - Step 5289: {'lr': 0.0004942299799200332, 'samples': 2708480, 'steps': 5289, 'loss/train': 2.175381898880005} 02/24/2022 08:24:46 - INFO - codeparrot_training - Step 5290: {'lr': 0.0004942264842789506, 'samples': 2708992, 'steps': 5290, 'loss/train': 2.2731714248657227} 02/24/2022 08:24:52 - INFO - codeparrot_training - Step 5291: {'lr': 0.0004942229875916792, 'samples': 2709504, 'steps': 5291, 'loss/train': 1.9897421598434448} 02/24/2022 08:24:55 - INFO - codeparrot_training - Step 5292: {'lr': 0.0004942194898582341, 'samples': 2710016, 'steps': 5292, 'loss/train': 1.783277153968811} 02/24/2022 08:25:01 - INFO - codeparrot_training - Step 5293: {'lr': 0.0004942159910786303, 'samples': 2710528, 'steps': 5293, 'loss/train': 3.5113205909729004} 02/24/2022 08:25:04 - INFO - codeparrot_training - Step 5294: {'lr': 0.0004942124912528827, 'samples': 2711040, 'steps': 5294, 'loss/train': 2.5687026977539062} 02/24/2022 08:25:10 - INFO - codeparrot_training - Step 5295: {'lr': 0.0004942089903810064, 'samples': 2711552, 'steps': 5295, 'loss/train': 1.251427412033081} 02/24/2022 08:25:13 - INFO - codeparrot_training - Step 5296: {'lr': 0.0004942054884630162, 'samples': 2712064, 'steps': 5296, 'loss/train': 3.1526901721954346} 02/24/2022 08:25:19 - INFO - codeparrot_training - Step 5297: {'lr': 0.0004942019854989274, 'samples': 2712576, 'steps': 5297, 'loss/train': 2.863186836242676} 02/24/2022 08:25:22 - INFO - codeparrot_training - Step 5298: {'lr': 0.0004941984814887546, 'samples': 2713088, 'steps': 5298, 'loss/train': 1.3744895458221436} 02/24/2022 08:25:28 - INFO - codeparrot_training - Step 5299: {'lr': 0.0004941949764325133, 'samples': 2713600, 'steps': 5299, 'loss/train': 2.7279162406921387} 02/24/2022 08:25:32 - INFO - codeparrot_training - Step 5300: {'lr': 0.0004941914703302181, 'samples': 2714112, 'steps': 5300, 'loss/train': 4.265825271606445} 02/24/2022 08:25:37 - INFO - codeparrot_training - Step 5301: {'lr': 0.0004941879631818843, 'samples': 2714624, 'steps': 5301, 'loss/train': 3.0422871112823486} 02/24/2022 08:25:41 - INFO - codeparrot_training - Step 5302: {'lr': 0.0004941844549875267, 'samples': 2715136, 'steps': 5302, 'loss/train': 2.122356653213501} 02/24/2022 08:25:46 - INFO - codeparrot_training - Step 5303: {'lr': 0.0004941809457471605, 'samples': 2715648, 'steps': 5303, 'loss/train': 2.3341476917266846} 02/24/2022 08:25:50 - INFO - codeparrot_training - Step 5304: {'lr': 0.0004941774354608006, 'samples': 2716160, 'steps': 5304, 'loss/train': 2.035097360610962} 02/24/2022 08:25:55 - INFO - codeparrot_training - Step 5305: {'lr': 0.0004941739241284621, 'samples': 2716672, 'steps': 5305, 'loss/train': 2.286090612411499} 02/24/2022 08:25:59 - INFO - codeparrot_training - Step 5306: {'lr': 0.0004941704117501601, 'samples': 2717184, 'steps': 5306, 'loss/train': 1.87517511844635} 02/24/2022 08:26:04 - INFO - codeparrot_training - Step 5307: {'lr': 0.0004941668983259095, 'samples': 2717696, 'steps': 5307, 'loss/train': 1.714046597480774} 02/24/2022 08:26:08 - INFO - codeparrot_training - Step 5308: {'lr': 0.0004941633838557256, 'samples': 2718208, 'steps': 5308, 'loss/train': 2.1482901573181152} 02/24/2022 08:26:14 - INFO - codeparrot_training - Step 5309: {'lr': 0.0004941598683396232, 'samples': 2718720, 'steps': 5309, 'loss/train': 3.1135590076446533} 02/24/2022 08:26:17 - INFO - codeparrot_training - Step 5310: {'lr': 0.0004941563517776174, 'samples': 2719232, 'steps': 5310, 'loss/train': 2.7894322872161865} 02/24/2022 08:26:23 - INFO - codeparrot_training - Step 5311: {'lr': 0.0004941528341697234, 'samples': 2719744, 'steps': 5311, 'loss/train': 2.3022868633270264} 02/24/2022 08:26:26 - INFO - codeparrot_training - Step 5312: {'lr': 0.0004941493155159562, 'samples': 2720256, 'steps': 5312, 'loss/train': 0.8195092678070068} 02/24/2022 08:26:32 - INFO - codeparrot_training - Step 5313: {'lr': 0.0004941457958163308, 'samples': 2720768, 'steps': 5313, 'loss/train': 2.694387674331665} 02/24/2022 08:26:35 - INFO - codeparrot_training - Step 5314: {'lr': 0.0004941422750708623, 'samples': 2721280, 'steps': 5314, 'loss/train': 2.1537344455718994} 02/24/2022 08:26:41 - INFO - codeparrot_training - Step 5315: {'lr': 0.0004941387532795659, 'samples': 2721792, 'steps': 5315, 'loss/train': 3.722724437713623} 02/24/2022 08:26:45 - INFO - codeparrot_training - Step 5316: {'lr': 0.0004941352304424566, 'samples': 2722304, 'steps': 5316, 'loss/train': 2.6303935050964355} 02/24/2022 08:26:50 - INFO - codeparrot_training - Step 5317: {'lr': 0.0004941317065595495, 'samples': 2722816, 'steps': 5317, 'loss/train': 1.962284803390503} 02/24/2022 08:26:56 - INFO - codeparrot_training - Step 5318: {'lr': 0.0004941281816308596, 'samples': 2723328, 'steps': 5318, 'loss/train': 2.6302943229675293} 02/24/2022 08:26:59 - INFO - codeparrot_training - Step 5319: {'lr': 0.0004941246556564021, 'samples': 2723840, 'steps': 5319, 'loss/train': 1.9872844219207764} 02/24/2022 08:27:05 - INFO - codeparrot_training - Step 5320: {'lr': 0.0004941211286361922, 'samples': 2724352, 'steps': 5320, 'loss/train': 2.951214551925659} 02/24/2022 08:27:08 - INFO - codeparrot_training - Step 5321: {'lr': 0.0004941176005702448, 'samples': 2724864, 'steps': 5321, 'loss/train': 0.1776892989873886} 02/24/2022 08:27:14 - INFO - codeparrot_training - Step 5322: {'lr': 0.0004941140714585752, 'samples': 2725376, 'steps': 5322, 'loss/train': 3.4030232429504395} 02/24/2022 08:27:17 - INFO - codeparrot_training - Step 5323: {'lr': 0.0004941105413011984, 'samples': 2725888, 'steps': 5323, 'loss/train': 2.8220651149749756} 02/24/2022 08:27:24 - INFO - codeparrot_training - Step 5324: {'lr': 0.0004941070100981295, 'samples': 2726400, 'steps': 5324, 'loss/train': 2.2283096313476562} 02/24/2022 08:27:27 - INFO - codeparrot_training - Step 5325: {'lr': 0.0004941034778493837, 'samples': 2726912, 'steps': 5325, 'loss/train': 2.513944387435913} 02/24/2022 08:27:30 - INFO - codeparrot_training - Step 5326: {'lr': 0.0004940999445549762, 'samples': 2727424, 'steps': 5326, 'loss/train': 0.9540320038795471} 02/24/2022 08:27:36 - INFO - codeparrot_training - Step 5327: {'lr': 0.0004940964102149219, 'samples': 2727936, 'steps': 5327, 'loss/train': 2.9037890434265137} 02/24/2022 08:27:39 - INFO - codeparrot_training - Step 5328: {'lr': 0.0004940928748292363, 'samples': 2728448, 'steps': 5328, 'loss/train': 2.8778762817382812} 02/24/2022 08:27:45 - INFO - codeparrot_training - Step 5329: {'lr': 0.0004940893383979341, 'samples': 2728960, 'steps': 5329, 'loss/train': 1.5792394876480103} 02/24/2022 08:27:48 - INFO - codeparrot_training - Step 5330: {'lr': 0.0004940858009210308, 'samples': 2729472, 'steps': 5330, 'loss/train': 1.6234550476074219} 02/24/2022 08:27:54 - INFO - codeparrot_training - Step 5331: {'lr': 0.0004940822623985414, 'samples': 2729984, 'steps': 5331, 'loss/train': 3.0696349143981934} 02/24/2022 08:27:57 - INFO - codeparrot_training - Step 5332: {'lr': 0.0004940787228304811, 'samples': 2730496, 'steps': 5332, 'loss/train': 3.5931174755096436} 02/24/2022 08:28:03 - INFO - codeparrot_training - Step 5333: {'lr': 0.0004940751822168651, 'samples': 2731008, 'steps': 5333, 'loss/train': 2.6041266918182373} 02/24/2022 08:28:09 - INFO - codeparrot_training - Step 5334: {'lr': 0.0004940716405577086, 'samples': 2731520, 'steps': 5334, 'loss/train': 2.4724690914154053} 02/24/2022 08:28:13 - INFO - codeparrot_training - Step 5335: {'lr': 0.0004940680978530265, 'samples': 2732032, 'steps': 5335, 'loss/train': 2.6429800987243652} 02/24/2022 08:28:18 - INFO - codeparrot_training - Step 5336: {'lr': 0.0004940645541028343, 'samples': 2732544, 'steps': 5336, 'loss/train': 1.3455263376235962} 02/24/2022 08:28:22 - INFO - codeparrot_training - Step 5337: {'lr': 0.0004940610093071469, 'samples': 2733056, 'steps': 5337, 'loss/train': 4.012539386749268} 02/24/2022 08:28:27 - INFO - codeparrot_training - Step 5338: {'lr': 0.0004940574634659798, 'samples': 2733568, 'steps': 5338, 'loss/train': 2.4877099990844727} 02/24/2022 08:28:31 - INFO - codeparrot_training - Step 5339: {'lr': 0.000494053916579348, 'samples': 2734080, 'steps': 5339, 'loss/train': 2.3281702995300293} 02/24/2022 08:28:36 - INFO - codeparrot_training - Step 5340: {'lr': 0.0004940503686472667, 'samples': 2734592, 'steps': 5340, 'loss/train': 2.581996202468872} 02/24/2022 08:28:40 - INFO - codeparrot_training - Step 5341: {'lr': 0.0004940468196697511, 'samples': 2735104, 'steps': 5341, 'loss/train': 2.3619203567504883} 02/24/2022 08:28:43 - INFO - codeparrot_training - Step 5342: {'lr': 0.0004940432696468164, 'samples': 2735616, 'steps': 5342, 'loss/train': 3.4700541496276855} 02/24/2022 08:28:49 - INFO - codeparrot_training - Step 5343: {'lr': 0.0004940397185784778, 'samples': 2736128, 'steps': 5343, 'loss/train': 3.0689940452575684} 02/24/2022 08:28:55 - INFO - codeparrot_training - Step 5344: {'lr': 0.0004940361664647506, 'samples': 2736640, 'steps': 5344, 'loss/train': 2.58054780960083} 02/24/2022 08:28:58 - INFO - codeparrot_training - Step 5345: {'lr': 0.0004940326133056499, 'samples': 2737152, 'steps': 5345, 'loss/train': 0.9639036655426025} 02/24/2022 08:29:04 - INFO - codeparrot_training - Step 5346: {'lr': 0.000494029059101191, 'samples': 2737664, 'steps': 5346, 'loss/train': 1.3676584959030151} 02/24/2022 08:29:07 - INFO - codeparrot_training - Step 5347: {'lr': 0.0004940255038513891, 'samples': 2738176, 'steps': 5347, 'loss/train': 2.4100656509399414} 02/24/2022 08:29:11 - INFO - codeparrot_training - Step 5348: {'lr': 0.0004940219475562593, 'samples': 2738688, 'steps': 5348, 'loss/train': 0.26062944531440735} 02/24/2022 08:29:16 - INFO - codeparrot_training - Step 5349: {'lr': 0.0004940183902158172, 'samples': 2739200, 'steps': 5349, 'loss/train': 3.099381923675537} 02/24/2022 08:29:22 - INFO - codeparrot_training - Step 5350: {'lr': 0.0004940148318300777, 'samples': 2739712, 'steps': 5350, 'loss/train': 2.632383346557617} 02/24/2022 08:29:25 - INFO - codeparrot_training - Step 5351: {'lr': 0.0004940112723990561, 'samples': 2740224, 'steps': 5351, 'loss/train': 3.2728219032287598} 02/24/2022 08:29:31 - INFO - codeparrot_training - Step 5352: {'lr': 0.0004940077119227678, 'samples': 2740736, 'steps': 5352, 'loss/train': 2.341137409210205} 02/24/2022 08:29:34 - INFO - codeparrot_training - Step 5353: {'lr': 0.0004940041504012279, 'samples': 2741248, 'steps': 5353, 'loss/train': 3.1377363204956055} 02/24/2022 08:29:41 - INFO - codeparrot_training - Step 5354: {'lr': 0.0004940005878344517, 'samples': 2741760, 'steps': 5354, 'loss/train': 3.3415699005126953} 02/24/2022 08:29:44 - INFO - codeparrot_training - Step 5355: {'lr': 0.0004939970242224544, 'samples': 2742272, 'steps': 5355, 'loss/train': 2.1214072704315186} 02/24/2022 08:29:50 - INFO - codeparrot_training - Step 5356: {'lr': 0.0004939934595652513, 'samples': 2742784, 'steps': 5356, 'loss/train': 1.4875153303146362} 02/24/2022 08:29:53 - INFO - codeparrot_training - Step 5357: {'lr': 0.0004939898938628578, 'samples': 2743296, 'steps': 5357, 'loss/train': 2.943678379058838} 02/24/2022 08:29:59 - INFO - codeparrot_training - Step 5358: {'lr': 0.000493986327115289, 'samples': 2743808, 'steps': 5358, 'loss/train': 2.513535737991333} 02/24/2022 08:30:02 - INFO - codeparrot_training - Step 5359: {'lr': 0.0004939827593225602, 'samples': 2744320, 'steps': 5359, 'loss/train': 1.2540435791015625} 02/24/2022 08:30:08 - INFO - codeparrot_training - Step 5360: {'lr': 0.0004939791904846869, 'samples': 2744832, 'steps': 5360, 'loss/train': 0.37946397066116333} 02/24/2022 08:30:11 - INFO - codeparrot_training - Step 5361: {'lr': 0.0004939756206016841, 'samples': 2745344, 'steps': 5361, 'loss/train': 1.523438811302185} 02/24/2022 08:30:17 - INFO - codeparrot_training - Step 5362: {'lr': 0.0004939720496735672, 'samples': 2745856, 'steps': 5362, 'loss/train': 1.9133459329605103} 02/24/2022 08:30:20 - INFO - codeparrot_training - Step 5363: {'lr': 0.0004939684777003516, 'samples': 2746368, 'steps': 5363, 'loss/train': 1.180401086807251} 02/24/2022 08:30:26 - INFO - codeparrot_training - Step 5364: {'lr': 0.0004939649046820524, 'samples': 2746880, 'steps': 5364, 'loss/train': 2.737797498703003} 02/24/2022 08:30:29 - INFO - codeparrot_training - Step 5365: {'lr': 0.0004939613306186851, 'samples': 2747392, 'steps': 5365, 'loss/train': 1.868486762046814} 02/24/2022 08:30:35 - INFO - codeparrot_training - Step 5366: {'lr': 0.0004939577555102649, 'samples': 2747904, 'steps': 5366, 'loss/train': 2.5798048973083496} 02/24/2022 08:30:38 - INFO - codeparrot_training - Step 5367: {'lr': 0.0004939541793568072, 'samples': 2748416, 'steps': 5367, 'loss/train': 2.7518157958984375} 02/24/2022 08:30:44 - INFO - codeparrot_training - Step 5368: {'lr': 0.000493950602158327, 'samples': 2748928, 'steps': 5368, 'loss/train': 3.122177839279175} 02/24/2022 08:30:47 - INFO - codeparrot_training - Step 5369: {'lr': 0.0004939470239148403, 'samples': 2749440, 'steps': 5369, 'loss/train': 1.493192195892334} 02/24/2022 08:30:54 - INFO - codeparrot_training - Step 5370: {'lr': 0.0004939434446263617, 'samples': 2749952, 'steps': 5370, 'loss/train': 3.5367205142974854} 02/24/2022 08:30:57 - INFO - codeparrot_training - Step 5371: {'lr': 0.000493939864292907, 'samples': 2750464, 'steps': 5371, 'loss/train': 2.0677380561828613} 02/24/2022 08:31:03 - INFO - codeparrot_training - Step 5372: {'lr': 0.0004939362829144913, 'samples': 2750976, 'steps': 5372, 'loss/train': 3.0955259799957275} 02/24/2022 08:31:06 - INFO - codeparrot_training - Step 5373: {'lr': 0.00049393270049113, 'samples': 2751488, 'steps': 5373, 'loss/train': 1.9950581789016724} 02/24/2022 08:31:12 - INFO - codeparrot_training - Step 5374: {'lr': 0.0004939291170228385, 'samples': 2752000, 'steps': 5374, 'loss/train': 2.679351568222046} 02/24/2022 08:31:15 - INFO - codeparrot_training - Step 5375: {'lr': 0.0004939255325096321, 'samples': 2752512, 'steps': 5375, 'loss/train': 2.3657546043395996} 02/24/2022 08:31:21 - INFO - codeparrot_training - Step 5376: {'lr': 0.0004939219469515262, 'samples': 2753024, 'steps': 5376, 'loss/train': 2.3725273609161377} 02/24/2022 08:31:24 - INFO - codeparrot_training - Step 5377: {'lr': 0.0004939183603485363, 'samples': 2753536, 'steps': 5377, 'loss/train': 2.000995397567749} 02/24/2022 08:31:30 - INFO - codeparrot_training - Step 5378: {'lr': 0.0004939147727006773, 'samples': 2754048, 'steps': 5378, 'loss/train': 2.109294891357422} 02/24/2022 08:31:33 - INFO - codeparrot_training - Step 5379: {'lr': 0.000493911184007965, 'samples': 2754560, 'steps': 5379, 'loss/train': 3.860227584838867} 02/24/2022 08:31:39 - INFO - codeparrot_training - Step 5380: {'lr': 0.0004939075942704147, 'samples': 2755072, 'steps': 5380, 'loss/train': 2.5177245140075684} 02/24/2022 08:31:43 - INFO - codeparrot_training - Step 5381: {'lr': 0.0004939040034880416, 'samples': 2755584, 'steps': 5381, 'loss/train': 2.001504898071289} 02/24/2022 08:31:48 - INFO - codeparrot_training - Step 5382: {'lr': 0.0004939004116608612, 'samples': 2756096, 'steps': 5382, 'loss/train': 3.1134586334228516} 02/24/2022 08:31:52 - INFO - codeparrot_training - Step 5383: {'lr': 0.000493896818788889, 'samples': 2756608, 'steps': 5383, 'loss/train': 1.1286685466766357} 02/24/2022 08:31:57 - INFO - codeparrot_training - Step 5384: {'lr': 0.0004938932248721401, 'samples': 2757120, 'steps': 5384, 'loss/train': 2.4488136768341064} 02/24/2022 08:32:01 - INFO - codeparrot_training - Step 5385: {'lr': 0.0004938896299106302, 'samples': 2757632, 'steps': 5385, 'loss/train': 1.8871541023254395} 02/24/2022 08:32:06 - INFO - codeparrot_training - Step 5386: {'lr': 0.0004938860339043746, 'samples': 2758144, 'steps': 5386, 'loss/train': 0.17116351425647736} 02/24/2022 08:32:10 - INFO - codeparrot_training - Step 5387: {'lr': 0.0004938824368533886, 'samples': 2758656, 'steps': 5387, 'loss/train': 2.8661715984344482} 02/24/2022 08:32:15 - INFO - codeparrot_training - Step 5388: {'lr': 0.0004938788387576878, 'samples': 2759168, 'steps': 5388, 'loss/train': 1.751478672027588} 02/24/2022 08:32:19 - INFO - codeparrot_training - Step 5389: {'lr': 0.0004938752396172873, 'samples': 2759680, 'steps': 5389, 'loss/train': 1.6210495233535767} 02/24/2022 08:32:26 - INFO - codeparrot_training - Step 5390: {'lr': 0.0004938716394322028, 'samples': 2760192, 'steps': 5390, 'loss/train': 1.0106014013290405} 02/24/2022 08:32:30 - INFO - codeparrot_training - Step 5391: {'lr': 0.0004938680382024497, 'samples': 2760704, 'steps': 5391, 'loss/train': 2.8358685970306396} 02/24/2022 08:32:35 - INFO - codeparrot_training - Step 5392: {'lr': 0.0004938644359280433, 'samples': 2761216, 'steps': 5392, 'loss/train': 2.2548182010650635} 02/24/2022 08:32:39 - INFO - codeparrot_training - Step 5393: {'lr': 0.000493860832608999, 'samples': 2761728, 'steps': 5393, 'loss/train': 2.348198652267456} 02/24/2022 08:32:44 - INFO - codeparrot_training - Step 5394: {'lr': 0.0004938572282453326, 'samples': 2762240, 'steps': 5394, 'loss/train': 2.4740381240844727} 02/24/2022 08:32:48 - INFO - codeparrot_training - Step 5395: {'lr': 0.000493853622837059, 'samples': 2762752, 'steps': 5395, 'loss/train': 2.233853816986084} 02/24/2022 08:32:54 - INFO - codeparrot_training - Step 5396: {'lr': 0.000493850016384194, 'samples': 2763264, 'steps': 5396, 'loss/train': 0.5000482201576233} 02/24/2022 08:32:57 - INFO - codeparrot_training - Step 5397: {'lr': 0.000493846408886753, 'samples': 2763776, 'steps': 5397, 'loss/train': 2.442148208618164} 02/24/2022 08:33:03 - INFO - codeparrot_training - Step 5398: {'lr': 0.0004938428003447514, 'samples': 2764288, 'steps': 5398, 'loss/train': 2.3250579833984375} 02/24/2022 08:33:06 - INFO - codeparrot_training - Step 5399: {'lr': 0.0004938391907582046, 'samples': 2764800, 'steps': 5399, 'loss/train': 3.1738271713256836} 02/24/2022 08:33:14 - INFO - codeparrot_training - Step 5400: {'lr': 0.0004938355801271282, 'samples': 2765312, 'steps': 5400, 'loss/train': 2.0156095027923584} 02/24/2022 08:33:17 - INFO - codeparrot_training - Step 5401: {'lr': 0.0004938319684515375, 'samples': 2765824, 'steps': 5401, 'loss/train': 1.7207632064819336} 02/24/2022 08:33:23 - INFO - codeparrot_training - Step 5402: {'lr': 0.0004938283557314483, 'samples': 2766336, 'steps': 5402, 'loss/train': 1.1186944246292114} 02/24/2022 08:33:26 - INFO - codeparrot_training - Step 5403: {'lr': 0.0004938247419668757, 'samples': 2766848, 'steps': 5403, 'loss/train': 3.0246737003326416} 02/24/2022 08:33:32 - INFO - codeparrot_training - Step 5404: {'lr': 0.0004938211271578352, 'samples': 2767360, 'steps': 5404, 'loss/train': 3.333437204360962} 02/24/2022 08:33:35 - INFO - codeparrot_training - Step 5405: {'lr': 0.0004938175113043426, 'samples': 2767872, 'steps': 5405, 'loss/train': 3.0279159545898438} 02/24/2022 08:33:41 - INFO - codeparrot_training - Step 5406: {'lr': 0.0004938138944064131, 'samples': 2768384, 'steps': 5406, 'loss/train': 2.5381946563720703} 02/24/2022 08:33:44 - INFO - codeparrot_training - Step 5407: {'lr': 0.0004938102764640624, 'samples': 2768896, 'steps': 5407, 'loss/train': 1.55325186252594} 02/24/2022 08:33:50 - INFO - codeparrot_training - Step 5408: {'lr': 0.0004938066574773058, 'samples': 2769408, 'steps': 5408, 'loss/train': 2.9902753829956055} 02/24/2022 08:33:53 - INFO - codeparrot_training - Step 5409: {'lr': 0.000493803037446159, 'samples': 2769920, 'steps': 5409, 'loss/train': 2.5365686416625977} 02/24/2022 08:34:01 - INFO - codeparrot_training - Step 5410: {'lr': 0.0004937994163706374, 'samples': 2770432, 'steps': 5410, 'loss/train': 3.490772247314453} 02/24/2022 08:34:04 - INFO - codeparrot_training - Step 5411: {'lr': 0.0004937957942507564, 'samples': 2770944, 'steps': 5411, 'loss/train': 1.6447222232818604} 02/24/2022 08:34:10 - INFO - codeparrot_training - Step 5412: {'lr': 0.0004937921710865317, 'samples': 2771456, 'steps': 5412, 'loss/train': 2.1643147468566895} 02/24/2022 08:34:13 - INFO - codeparrot_training - Step 5413: {'lr': 0.0004937885468779787, 'samples': 2771968, 'steps': 5413, 'loss/train': 2.4104678630828857} 02/24/2022 08:34:19 - INFO - codeparrot_training - Step 5414: {'lr': 0.000493784921625113, 'samples': 2772480, 'steps': 5414, 'loss/train': 1.894930124282837} 02/24/2022 08:34:22 - INFO - codeparrot_training - Step 5415: {'lr': 0.0004937812953279502, 'samples': 2772992, 'steps': 5415, 'loss/train': 2.163015365600586} 02/24/2022 08:34:28 - INFO - codeparrot_training - Step 5416: {'lr': 0.0004937776679865057, 'samples': 2773504, 'steps': 5416, 'loss/train': 3.095672130584717} 02/24/2022 08:34:31 - INFO - codeparrot_training - Step 5417: {'lr': 0.000493774039600795, 'samples': 2774016, 'steps': 5417, 'loss/train': 2.0287392139434814} 02/24/2022 08:34:37 - INFO - codeparrot_training - Step 5418: {'lr': 0.0004937704101708338, 'samples': 2774528, 'steps': 5418, 'loss/train': 2.356771230697632} 02/24/2022 08:34:40 - INFO - codeparrot_training - Step 5419: {'lr': 0.0004937667796966374, 'samples': 2775040, 'steps': 5419, 'loss/train': 2.3700101375579834} 02/24/2022 08:34:46 - INFO - codeparrot_training - Step 5420: {'lr': 0.0004937631481782218, 'samples': 2775552, 'steps': 5420, 'loss/train': 2.759166717529297} 02/24/2022 08:34:49 - INFO - codeparrot_training - Step 5421: {'lr': 0.000493759515615602, 'samples': 2776064, 'steps': 5421, 'loss/train': 2.496807336807251} 02/24/2022 08:34:57 - INFO - codeparrot_training - Step 5422: {'lr': 0.000493755882008794, 'samples': 2776576, 'steps': 5422, 'loss/train': 2.534860372543335} 02/24/2022 08:35:00 - INFO - codeparrot_training - Step 5423: {'lr': 0.0004937522473578132, 'samples': 2777088, 'steps': 5423, 'loss/train': 1.8650087118148804} 02/24/2022 08:35:06 - INFO - codeparrot_training - Step 5424: {'lr': 0.0004937486116626752, 'samples': 2777600, 'steps': 5424, 'loss/train': 2.7128074169158936} 02/24/2022 08:35:09 - INFO - codeparrot_training - Step 5425: {'lr': 0.0004937449749233954, 'samples': 2778112, 'steps': 5425, 'loss/train': 1.823857307434082} 02/24/2022 08:35:15 - INFO - codeparrot_training - Step 5426: {'lr': 0.0004937413371399897, 'samples': 2778624, 'steps': 5426, 'loss/train': 2.4024977684020996} 02/24/2022 08:35:18 - INFO - codeparrot_training - Step 5427: {'lr': 0.0004937376983124734, 'samples': 2779136, 'steps': 5427, 'loss/train': 1.462932825088501} 02/24/2022 08:35:24 - INFO - codeparrot_training - Step 5428: {'lr': 0.0004937340584408622, 'samples': 2779648, 'steps': 5428, 'loss/train': 1.3764538764953613} 02/24/2022 08:35:27 - INFO - codeparrot_training - Step 5429: {'lr': 0.0004937304175251717, 'samples': 2780160, 'steps': 5429, 'loss/train': 2.4385201930999756} 02/24/2022 08:35:33 - INFO - codeparrot_training - Step 5430: {'lr': 0.0004937267755654174, 'samples': 2780672, 'steps': 5430, 'loss/train': 3.247731924057007} 02/24/2022 08:35:36 - INFO - codeparrot_training - Step 5431: {'lr': 0.0004937231325616152, 'samples': 2781184, 'steps': 5431, 'loss/train': 1.6430755853652954} 02/24/2022 08:35:42 - INFO - codeparrot_training - Step 5432: {'lr': 0.0004937194885137803, 'samples': 2781696, 'steps': 5432, 'loss/train': 2.4172794818878174} 02/24/2022 08:35:45 - INFO - codeparrot_training - Step 5433: {'lr': 0.0004937158434219286, 'samples': 2782208, 'steps': 5433, 'loss/train': 2.495945930480957} 02/24/2022 08:35:51 - INFO - codeparrot_training - Step 5434: {'lr': 0.0004937121972860755, 'samples': 2782720, 'steps': 5434, 'loss/train': 3.0654258728027344} 02/24/2022 08:35:54 - INFO - codeparrot_training - Step 5435: {'lr': 0.0004937085501062369, 'samples': 2783232, 'steps': 5435, 'loss/train': 1.7805092334747314} 02/24/2022 08:36:01 - INFO - codeparrot_training - Step 5436: {'lr': 0.0004937049018824282, 'samples': 2783744, 'steps': 5436, 'loss/train': 2.5049848556518555} 02/24/2022 08:36:07 - INFO - codeparrot_training - Step 5437: {'lr': 0.000493701252614665, 'samples': 2784256, 'steps': 5437, 'loss/train': 2.8795433044433594} 02/24/2022 08:36:11 - INFO - codeparrot_training - Step 5438: {'lr': 0.0004936976023029631, 'samples': 2784768, 'steps': 5438, 'loss/train': 2.858910083770752} 02/24/2022 08:36:16 - INFO - codeparrot_training - Step 5439: {'lr': 0.000493693950947338, 'samples': 2785280, 'steps': 5439, 'loss/train': 2.472738742828369} 02/24/2022 08:36:20 - INFO - codeparrot_training - Step 5440: {'lr': 0.0004936902985478055, 'samples': 2785792, 'steps': 5440, 'loss/train': 2.654956102371216} 02/24/2022 08:36:25 - INFO - codeparrot_training - Step 5441: {'lr': 0.000493686645104381, 'samples': 2786304, 'steps': 5441, 'loss/train': 3.063326358795166} 02/24/2022 08:36:29 - INFO - codeparrot_training - Step 5442: {'lr': 0.0004936829906170804, 'samples': 2786816, 'steps': 5442, 'loss/train': 1.4038671255111694} 02/24/2022 08:36:34 - INFO - codeparrot_training - Step 5443: {'lr': 0.0004936793350859192, 'samples': 2787328, 'steps': 5443, 'loss/train': 2.1081933975219727} 02/24/2022 08:36:38 - INFO - codeparrot_training - Step 5444: {'lr': 0.0004936756785109131, 'samples': 2787840, 'steps': 5444, 'loss/train': 2.282024383544922} 02/24/2022 08:36:45 - INFO - codeparrot_training - Step 5445: {'lr': 0.0004936720208920778, 'samples': 2788352, 'steps': 5445, 'loss/train': 2.9123265743255615} 02/24/2022 08:36:48 - INFO - codeparrot_training - Step 5446: {'lr': 0.0004936683622294289, 'samples': 2788864, 'steps': 5446, 'loss/train': 3.220599412918091} 02/24/2022 08:36:54 - INFO - codeparrot_training - Step 5447: {'lr': 0.0004936647025229822, 'samples': 2789376, 'steps': 5447, 'loss/train': 1.7034388780593872} 02/24/2022 08:36:57 - INFO - codeparrot_training - Step 5448: {'lr': 0.0004936610417727532, 'samples': 2789888, 'steps': 5448, 'loss/train': 0.8039000034332275} 02/24/2022 08:37:03 - INFO - codeparrot_training - Step 5449: {'lr': 0.0004936573799787575, 'samples': 2790400, 'steps': 5449, 'loss/train': 3.2148349285125732} 02/24/2022 08:37:06 - INFO - codeparrot_training - Step 5450: {'lr': 0.0004936537171410112, 'samples': 2790912, 'steps': 5450, 'loss/train': 2.899169683456421} 02/24/2022 08:37:12 - INFO - codeparrot_training - Step 5451: {'lr': 0.0004936500532595297, 'samples': 2791424, 'steps': 5451, 'loss/train': 3.7909696102142334} 02/24/2022 08:37:16 - INFO - codeparrot_training - Step 5452: {'lr': 0.0004936463883343287, 'samples': 2791936, 'steps': 5452, 'loss/train': 2.4795773029327393} 02/24/2022 08:37:19 - INFO - codeparrot_training - Step 5453: {'lr': 0.000493642722365424, 'samples': 2792448, 'steps': 5453, 'loss/train': 1.709165096282959} 02/24/2022 08:37:25 - INFO - codeparrot_training - Step 5454: {'lr': 0.0004936390553528313, 'samples': 2792960, 'steps': 5454, 'loss/train': 2.9883267879486084} 02/24/2022 08:37:32 - INFO - codeparrot_training - Step 5455: {'lr': 0.0004936353872965661, 'samples': 2793472, 'steps': 5455, 'loss/train': 2.1667511463165283} 02/24/2022 08:37:35 - INFO - codeparrot_training - Step 5456: {'lr': 0.0004936317181966443, 'samples': 2793984, 'steps': 5456, 'loss/train': 1.7340718507766724} 02/24/2022 08:37:41 - INFO - codeparrot_training - Step 5457: {'lr': 0.0004936280480530816, 'samples': 2794496, 'steps': 5457, 'loss/train': 1.9111151695251465} 02/24/2022 08:37:44 - INFO - codeparrot_training - Step 5458: {'lr': 0.0004936243768658937, 'samples': 2795008, 'steps': 5458, 'loss/train': 1.3520731925964355} 02/24/2022 08:37:50 - INFO - codeparrot_training - Step 5459: {'lr': 0.0004936207046350963, 'samples': 2795520, 'steps': 5459, 'loss/train': 2.566462516784668} 02/24/2022 08:37:53 - INFO - codeparrot_training - Step 5460: {'lr': 0.0004936170313607053, 'samples': 2796032, 'steps': 5460, 'loss/train': 2.4731314182281494} 02/24/2022 08:37:57 - INFO - codeparrot_training - Step 5461: {'lr': 0.0004936133570427361, 'samples': 2796544, 'steps': 5461, 'loss/train': 2.474146842956543} 02/24/2022 08:38:02 - INFO - codeparrot_training - Step 5462: {'lr': 0.0004936096816812046, 'samples': 2797056, 'steps': 5462, 'loss/train': 2.897639274597168} 02/24/2022 08:38:06 - INFO - codeparrot_training - Step 5463: {'lr': 0.0004936060052761268, 'samples': 2797568, 'steps': 5463, 'loss/train': 2.1374096870422363} 02/24/2022 08:38:11 - INFO - codeparrot_training - Step 5464: {'lr': 0.0004936023278275182, 'samples': 2798080, 'steps': 5464, 'loss/train': 2.5332071781158447} 02/24/2022 08:38:17 - INFO - codeparrot_training - Step 5465: {'lr': 0.0004935986493353944, 'samples': 2798592, 'steps': 5465, 'loss/train': 2.5430822372436523} 02/24/2022 08:38:20 - INFO - codeparrot_training - Step 5466: {'lr': 0.0004935949697997715, 'samples': 2799104, 'steps': 5466, 'loss/train': 2.320448637008667} 02/24/2022 08:38:28 - INFO - codeparrot_training - Step 5467: {'lr': 0.000493591289220665, 'samples': 2799616, 'steps': 5467, 'loss/train': 3.023228406906128} 02/24/2022 08:38:31 - INFO - codeparrot_training - Step 5468: {'lr': 0.0004935876075980908, 'samples': 2800128, 'steps': 5468, 'loss/train': 1.9866042137145996} 02/24/2022 08:38:37 - INFO - codeparrot_training - Step 5469: {'lr': 0.0004935839249320647, 'samples': 2800640, 'steps': 5469, 'loss/train': 2.737163543701172} 02/24/2022 08:38:40 - INFO - codeparrot_training - Step 5470: {'lr': 0.0004935802412226024, 'samples': 2801152, 'steps': 5470, 'loss/train': 1.5813744068145752} 02/24/2022 08:38:46 - INFO - codeparrot_training - Step 5471: {'lr': 0.0004935765564697195, 'samples': 2801664, 'steps': 5471, 'loss/train': 2.9298453330993652} 02/24/2022 08:38:49 - INFO - codeparrot_training - Step 5472: {'lr': 0.0004935728706734322, 'samples': 2802176, 'steps': 5472, 'loss/train': 2.430553674697876} 02/24/2022 08:38:55 - INFO - codeparrot_training - Step 5473: {'lr': 0.000493569183833756, 'samples': 2802688, 'steps': 5473, 'loss/train': 2.3587841987609863} 02/24/2022 08:38:58 - INFO - codeparrot_training - Step 5474: {'lr': 0.0004935654959507068, 'samples': 2803200, 'steps': 5474, 'loss/train': 2.5076725482940674} 02/24/2022 08:39:03 - INFO - codeparrot_training - Step 5475: {'lr': 0.0004935618070243003, 'samples': 2803712, 'steps': 5475, 'loss/train': 2.501732110977173} 02/24/2022 08:39:07 - INFO - codeparrot_training - Step 5476: {'lr': 0.0004935581170545523, 'samples': 2804224, 'steps': 5476, 'loss/train': 1.607458472251892} 02/24/2022 08:39:13 - INFO - codeparrot_training - Step 5477: {'lr': 0.0004935544260414787, 'samples': 2804736, 'steps': 5477, 'loss/train': 2.6318142414093018} 02/24/2022 08:39:16 - INFO - codeparrot_training - Step 5478: {'lr': 0.0004935507339850953, 'samples': 2805248, 'steps': 5478, 'loss/train': 2.7555673122406006} 02/24/2022 08:39:21 - INFO - codeparrot_training - Step 5479: {'lr': 0.0004935470408854179, 'samples': 2805760, 'steps': 5479, 'loss/train': 0.8822837471961975} 02/24/2022 08:39:25 - INFO - codeparrot_training - Step 5480: {'lr': 0.0004935433467424624, 'samples': 2806272, 'steps': 5480, 'loss/train': 2.194122552871704} 02/24/2022 08:39:32 - INFO - codeparrot_training - Step 5481: {'lr': 0.0004935396515562444, 'samples': 2806784, 'steps': 5481, 'loss/train': 2.570009708404541} 02/24/2022 08:39:36 - INFO - codeparrot_training - Step 5482: {'lr': 0.0004935359553267798, 'samples': 2807296, 'steps': 5482, 'loss/train': 2.700437307357788} 02/24/2022 08:39:41 - INFO - codeparrot_training - Step 5483: {'lr': 0.0004935322580540847, 'samples': 2807808, 'steps': 5483, 'loss/train': 2.5435824394226074} 02/24/2022 08:39:45 - INFO - codeparrot_training - Step 5484: {'lr': 0.0004935285597381747, 'samples': 2808320, 'steps': 5484, 'loss/train': 2.6107141971588135} 02/24/2022 08:39:50 - INFO - codeparrot_training - Step 5485: {'lr': 0.0004935248603790656, 'samples': 2808832, 'steps': 5485, 'loss/train': 1.681152582168579} 02/24/2022 08:39:54 - INFO - codeparrot_training - Step 5486: {'lr': 0.0004935211599767733, 'samples': 2809344, 'steps': 5486, 'loss/train': 2.324976921081543} 02/24/2022 08:39:59 - INFO - codeparrot_training - Step 5487: {'lr': 0.0004935174585313138, 'samples': 2809856, 'steps': 5487, 'loss/train': 3.132422685623169} 02/24/2022 08:40:03 - INFO - codeparrot_training - Step 5488: {'lr': 0.0004935137560427027, 'samples': 2810368, 'steps': 5488, 'loss/train': 2.7801859378814697} 02/24/2022 08:40:08 - INFO - codeparrot_training - Step 5489: {'lr': 0.000493510052510956, 'samples': 2810880, 'steps': 5489, 'loss/train': 2.628154754638672} 02/24/2022 08:40:12 - INFO - codeparrot_training - Step 5490: {'lr': 0.0004935063479360897, 'samples': 2811392, 'steps': 5490, 'loss/train': 5.426364421844482} 02/24/2022 08:40:19 - INFO - codeparrot_training - Step 5491: {'lr': 0.0004935026423181194, 'samples': 2811904, 'steps': 5491, 'loss/train': 1.8185917139053345} 02/24/2022 08:40:22 - INFO - codeparrot_training - Step 5492: {'lr': 0.0004934989356570611, 'samples': 2812416, 'steps': 5492, 'loss/train': 1.8836909532546997} 02/24/2022 08:40:28 - INFO - codeparrot_training - Step 5493: {'lr': 0.0004934952279529308, 'samples': 2812928, 'steps': 5493, 'loss/train': 2.1510279178619385} 02/24/2022 08:40:31 - INFO - codeparrot_training - Step 5494: {'lr': 0.0004934915192057441, 'samples': 2813440, 'steps': 5494, 'loss/train': 2.4905660152435303} 02/24/2022 08:40:37 - INFO - codeparrot_training - Step 5495: {'lr': 0.0004934878094155172, 'samples': 2813952, 'steps': 5495, 'loss/train': 2.4707770347595215} 02/24/2022 08:40:40 - INFO - codeparrot_training - Step 5496: {'lr': 0.0004934840985822657, 'samples': 2814464, 'steps': 5496, 'loss/train': 1.2118127346038818} 02/24/2022 08:40:46 - INFO - codeparrot_training - Step 5497: {'lr': 0.0004934803867060058, 'samples': 2814976, 'steps': 5497, 'loss/train': 3.060567855834961} 02/24/2022 08:40:49 - INFO - codeparrot_training - Step 5498: {'lr': 0.0004934766737867531, 'samples': 2815488, 'steps': 5498, 'loss/train': 2.6628003120422363} 02/24/2022 08:40:55 - INFO - codeparrot_training - Step 5499: {'lr': 0.0004934729598245237, 'samples': 2816000, 'steps': 5499, 'loss/train': 2.2565956115722656} 02/24/2022 08:40:58 - INFO - codeparrot_training - Step 5500: {'lr': 0.0004934692448193334, 'samples': 2816512, 'steps': 5500, 'loss/train': 2.064164161682129} 02/24/2022 08:41:04 - INFO - codeparrot_training - Step 5501: {'lr': 0.0004934655287711982, 'samples': 2817024, 'steps': 5501, 'loss/train': 2.3882791996002197} 02/24/2022 08:41:07 - INFO - codeparrot_training - Step 5502: {'lr': 0.0004934618116801341, 'samples': 2817536, 'steps': 5502, 'loss/train': 3.4337117671966553} 02/24/2022 08:41:15 - INFO - codeparrot_training - Step 5503: {'lr': 0.0004934580935461567, 'samples': 2818048, 'steps': 5503, 'loss/train': 3.1718528270721436} 02/24/2022 08:41:18 - INFO - codeparrot_training - Step 5504: {'lr': 0.0004934543743692822, 'samples': 2818560, 'steps': 5504, 'loss/train': 2.7611584663391113} 02/24/2022 08:41:24 - INFO - codeparrot_training - Step 5505: {'lr': 0.0004934506541495265, 'samples': 2819072, 'steps': 5505, 'loss/train': 1.4494328498840332} 02/24/2022 08:41:27 - INFO - codeparrot_training - Step 5506: {'lr': 0.0004934469328869056, 'samples': 2819584, 'steps': 5506, 'loss/train': 2.5046980381011963} 02/24/2022 08:41:33 - INFO - codeparrot_training - Step 5507: {'lr': 0.0004934432105814352, 'samples': 2820096, 'steps': 5507, 'loss/train': 3.164855718612671} 02/24/2022 08:41:36 - INFO - codeparrot_training - Step 5508: {'lr': 0.0004934394872331314, 'samples': 2820608, 'steps': 5508, 'loss/train': 3.4603705406188965} 02/24/2022 08:41:42 - INFO - codeparrot_training - Step 5509: {'lr': 0.0004934357628420101, 'samples': 2821120, 'steps': 5509, 'loss/train': 2.5381319522857666} 02/24/2022 08:41:45 - INFO - codeparrot_training - Step 5510: {'lr': 0.0004934320374080874, 'samples': 2821632, 'steps': 5510, 'loss/train': 3.3778696060180664} 02/24/2022 08:41:51 - INFO - codeparrot_training - Step 5511: {'lr': 0.000493428310931379, 'samples': 2822144, 'steps': 5511, 'loss/train': 2.120063066482544} 02/24/2022 08:41:54 - INFO - codeparrot_training - Step 5512: {'lr': 0.0004934245834119013, 'samples': 2822656, 'steps': 5512, 'loss/train': 2.2954940795898438} 02/24/2022 08:42:02 - INFO - codeparrot_training - Step 5513: {'lr': 0.0004934208548496697, 'samples': 2823168, 'steps': 5513, 'loss/train': 3.1563899517059326} 02/24/2022 08:42:05 - INFO - codeparrot_training - Step 5514: {'lr': 0.0004934171252447006, 'samples': 2823680, 'steps': 5514, 'loss/train': 1.0124396085739136} 02/24/2022 08:42:11 - INFO - codeparrot_training - Step 5515: {'lr': 0.0004934133945970097, 'samples': 2824192, 'steps': 5515, 'loss/train': 3.836822271347046} 02/24/2022 08:42:14 - INFO - codeparrot_training - Step 5516: {'lr': 0.0004934096629066133, 'samples': 2824704, 'steps': 5516, 'loss/train': 0.8433687090873718} 02/24/2022 08:42:20 - INFO - codeparrot_training - Step 5517: {'lr': 0.000493405930173527, 'samples': 2825216, 'steps': 5517, 'loss/train': 2.610063076019287} 02/24/2022 08:42:23 - INFO - codeparrot_training - Step 5518: {'lr': 0.0004934021963977671, 'samples': 2825728, 'steps': 5518, 'loss/train': 1.5580424070358276} 02/24/2022 08:42:29 - INFO - codeparrot_training - Step 5519: {'lr': 0.0004933984615793494, 'samples': 2826240, 'steps': 5519, 'loss/train': 1.7483354806900024} 02/24/2022 08:42:32 - INFO - codeparrot_training - Step 5520: {'lr': 0.0004933947257182901, 'samples': 2826752, 'steps': 5520, 'loss/train': 2.165771007537842} 02/24/2022 08:42:38 - INFO - codeparrot_training - Step 5521: {'lr': 0.000493390988814605, 'samples': 2827264, 'steps': 5521, 'loss/train': 3.6925840377807617} 02/24/2022 08:42:41 - INFO - codeparrot_training - Step 5522: {'lr': 0.0004933872508683101, 'samples': 2827776, 'steps': 5522, 'loss/train': 2.4756855964660645} 02/24/2022 08:42:47 - INFO - codeparrot_training - Step 5523: {'lr': 0.0004933835118794217, 'samples': 2828288, 'steps': 5523, 'loss/train': 0.5385767817497253} 02/24/2022 08:42:50 - INFO - codeparrot_training - Step 5524: {'lr': 0.0004933797718479555, 'samples': 2828800, 'steps': 5524, 'loss/train': 1.9447957277297974} 02/24/2022 08:42:56 - INFO - codeparrot_training - Step 5525: {'lr': 0.0004933760307739277, 'samples': 2829312, 'steps': 5525, 'loss/train': 4.624439239501953} 02/24/2022 08:43:00 - INFO - codeparrot_training - Step 5526: {'lr': 0.0004933722886573542, 'samples': 2829824, 'steps': 5526, 'loss/train': 1.0215610265731812} 02/24/2022 08:43:05 - INFO - codeparrot_training - Step 5527: {'lr': 0.0004933685454982511, 'samples': 2830336, 'steps': 5527, 'loss/train': 1.5295915603637695} 02/24/2022 08:43:08 - INFO - codeparrot_training - Step 5528: {'lr': 0.0004933648012966344, 'samples': 2830848, 'steps': 5528, 'loss/train': 2.2978732585906982} 02/24/2022 08:43:16 - INFO - codeparrot_training - Step 5529: {'lr': 0.0004933610560525203, 'samples': 2831360, 'steps': 5529, 'loss/train': 3.312633991241455} 02/24/2022 08:43:19 - INFO - codeparrot_training - Step 5530: {'lr': 0.0004933573097659246, 'samples': 2831872, 'steps': 5530, 'loss/train': 1.7465201616287231} 02/24/2022 08:43:25 - INFO - codeparrot_training - Step 5531: {'lr': 0.0004933535624368634, 'samples': 2832384, 'steps': 5531, 'loss/train': 2.419267177581787} 02/24/2022 08:43:28 - INFO - codeparrot_training - Step 5532: {'lr': 0.0004933498140653529, 'samples': 2832896, 'steps': 5532, 'loss/train': 3.0108892917633057} 02/24/2022 08:43:34 - INFO - codeparrot_training - Step 5533: {'lr': 0.0004933460646514092, 'samples': 2833408, 'steps': 5533, 'loss/train': 2.615257501602173} 02/24/2022 08:43:37 - INFO - codeparrot_training - Step 5534: {'lr': 0.000493342314195048, 'samples': 2833920, 'steps': 5534, 'loss/train': 3.0479815006256104} 02/24/2022 08:43:43 - INFO - codeparrot_training - Step 5535: {'lr': 0.0004933385626962858, 'samples': 2834432, 'steps': 5535, 'loss/train': 0.3741433322429657} 02/24/2022 08:43:46 - INFO - codeparrot_training - Step 5536: {'lr': 0.0004933348101551383, 'samples': 2834944, 'steps': 5536, 'loss/train': 1.148425579071045} 02/24/2022 08:43:52 - INFO - codeparrot_training - Step 5537: {'lr': 0.0004933310565716218, 'samples': 2835456, 'steps': 5537, 'loss/train': 3.763272523880005} 02/24/2022 08:43:55 - INFO - codeparrot_training - Step 5538: {'lr': 0.0004933273019457524, 'samples': 2835968, 'steps': 5538, 'loss/train': 2.1809885501861572} 02/24/2022 08:44:02 - INFO - codeparrot_training - Step 5539: {'lr': 0.0004933235462775459, 'samples': 2836480, 'steps': 5539, 'loss/train': 2.2543070316314697} 02/24/2022 08:44:06 - INFO - codeparrot_training - Step 5540: {'lr': 0.0004933197895670187, 'samples': 2836992, 'steps': 5540, 'loss/train': 3.752925157546997} 02/24/2022 08:44:11 - INFO - codeparrot_training - Step 5541: {'lr': 0.0004933160318141869, 'samples': 2837504, 'steps': 5541, 'loss/train': 2.4720349311828613} 02/24/2022 08:44:15 - INFO - codeparrot_training - Step 5542: {'lr': 0.0004933122730190663, 'samples': 2838016, 'steps': 5542, 'loss/train': 2.054579496383667} 02/24/2022 08:44:20 - INFO - codeparrot_training - Step 5543: {'lr': 0.0004933085131816733, 'samples': 2838528, 'steps': 5543, 'loss/train': 1.9743494987487793} 02/24/2022 08:44:24 - INFO - codeparrot_training - Step 5544: {'lr': 0.0004933047523020239, 'samples': 2839040, 'steps': 5544, 'loss/train': 3.0286362171173096} 02/24/2022 08:44:29 - INFO - codeparrot_training - Step 5545: {'lr': 0.0004933009903801341, 'samples': 2839552, 'steps': 5545, 'loss/train': 2.624907970428467} 02/24/2022 08:44:33 - INFO - codeparrot_training - Step 5546: {'lr': 0.0004932972274160202, 'samples': 2840064, 'steps': 5546, 'loss/train': 2.792562484741211} 02/24/2022 08:44:38 - INFO - codeparrot_training - Step 5547: {'lr': 0.0004932934634096982, 'samples': 2840576, 'steps': 5547, 'loss/train': 3.149885654449463} 02/24/2022 08:44:42 - INFO - codeparrot_training - Step 5548: {'lr': 0.0004932896983611843, 'samples': 2841088, 'steps': 5548, 'loss/train': 3.6530075073242188} 02/24/2022 08:44:49 - INFO - codeparrot_training - Step 5549: {'lr': 0.0004932859322704944, 'samples': 2841600, 'steps': 5549, 'loss/train': 1.6506565809249878} 02/24/2022 08:44:52 - INFO - codeparrot_training - Step 5550: {'lr': 0.000493282165137645, 'samples': 2842112, 'steps': 5550, 'loss/train': 2.8397183418273926} 02/24/2022 08:44:58 - INFO - codeparrot_training - Step 5551: {'lr': 0.0004932783969626521, 'samples': 2842624, 'steps': 5551, 'loss/train': 2.3374781608581543} 02/24/2022 08:45:01 - INFO - codeparrot_training - Step 5552: {'lr': 0.0004932746277455317, 'samples': 2843136, 'steps': 5552, 'loss/train': 2.3526647090911865} 02/24/2022 08:45:07 - INFO - codeparrot_training - Step 5553: {'lr': 0.0004932708574863, 'samples': 2843648, 'steps': 5553, 'loss/train': 2.5887060165405273} 02/24/2022 08:45:10 - INFO - codeparrot_training - Step 5554: {'lr': 0.0004932670861849733, 'samples': 2844160, 'steps': 5554, 'loss/train': 2.2699503898620605} 02/24/2022 08:45:16 - INFO - codeparrot_training - Step 5555: {'lr': 0.0004932633138415675, 'samples': 2844672, 'steps': 5555, 'loss/train': 2.6233675479888916} 02/24/2022 08:45:19 - INFO - codeparrot_training - Step 5556: {'lr': 0.000493259540456099, 'samples': 2845184, 'steps': 5556, 'loss/train': 2.745791435241699} 02/24/2022 08:45:25 - INFO - codeparrot_training - Step 5557: {'lr': 0.0004932557660285839, 'samples': 2845696, 'steps': 5557, 'loss/train': 2.101487398147583} 02/24/2022 08:45:28 - INFO - codeparrot_training - Step 5558: {'lr': 0.0004932519905590383, 'samples': 2846208, 'steps': 5558, 'loss/train': 2.3263118267059326} 02/24/2022 08:45:35 - INFO - codeparrot_training - Step 5559: {'lr': 0.0004932482140474785, 'samples': 2846720, 'steps': 5559, 'loss/train': 1.1069170236587524} 02/24/2022 08:45:38 - INFO - codeparrot_training - Step 5560: {'lr': 0.0004932444364939204, 'samples': 2847232, 'steps': 5560, 'loss/train': 2.67093563079834} 02/24/2022 08:45:44 - INFO - codeparrot_training - Step 5561: {'lr': 0.0004932406578983806, 'samples': 2847744, 'steps': 5561, 'loss/train': 2.045226573944092} 02/24/2022 08:45:47 - INFO - codeparrot_training - Step 5562: {'lr': 0.0004932368782608749, 'samples': 2848256, 'steps': 5562, 'loss/train': 3.291518449783325} 02/24/2022 08:45:53 - INFO - codeparrot_training - Step 5563: {'lr': 0.0004932330975814198, 'samples': 2848768, 'steps': 5563, 'loss/train': 2.0810651779174805} 02/24/2022 08:45:56 - INFO - codeparrot_training - Step 5564: {'lr': 0.0004932293158600312, 'samples': 2849280, 'steps': 5564, 'loss/train': 1.5393812656402588} 02/24/2022 08:46:02 - INFO - codeparrot_training - Step 5565: {'lr': 0.0004932255330967255, 'samples': 2849792, 'steps': 5565, 'loss/train': 2.9816529750823975} 02/24/2022 08:46:05 - INFO - codeparrot_training - Step 5566: {'lr': 0.0004932217492915189, 'samples': 2850304, 'steps': 5566, 'loss/train': 1.493029236793518} 02/24/2022 08:46:11 - INFO - codeparrot_training - Step 5567: {'lr': 0.0004932179644444274, 'samples': 2850816, 'steps': 5567, 'loss/train': 2.513721227645874} 02/24/2022 08:46:14 - INFO - codeparrot_training - Step 5568: {'lr': 0.0004932141785554676, 'samples': 2851328, 'steps': 5568, 'loss/train': 2.5890614986419678} 02/24/2022 08:46:20 - INFO - codeparrot_training - Step 5569: {'lr': 0.0004932103916246553, 'samples': 2851840, 'steps': 5569, 'loss/train': 1.7538347244262695} 02/24/2022 08:46:25 - INFO - codeparrot_training - Step 5570: {'lr': 0.000493206603652007, 'samples': 2852352, 'steps': 5570, 'loss/train': 1.822554588317871} 02/24/2022 08:46:28 - INFO - codeparrot_training - Step 5571: {'lr': 0.0004932028146375388, 'samples': 2852864, 'steps': 5571, 'loss/train': 1.128174901008606} 02/24/2022 08:46:34 - INFO - codeparrot_training - Step 5572: {'lr': 0.000493199024581267, 'samples': 2853376, 'steps': 5572, 'loss/train': 2.5253806114196777} 02/24/2022 08:46:38 - INFO - codeparrot_training - Step 5573: {'lr': 0.0004931952334832077, 'samples': 2853888, 'steps': 5573, 'loss/train': 0.4618869423866272} 02/24/2022 08:46:44 - INFO - codeparrot_training - Step 5574: {'lr': 0.0004931914413433773, 'samples': 2854400, 'steps': 5574, 'loss/train': 3.4767470359802246} 02/24/2022 08:46:47 - INFO - codeparrot_training - Step 5575: {'lr': 0.0004931876481617921, 'samples': 2854912, 'steps': 5575, 'loss/train': 2.4454307556152344} 02/24/2022 08:46:53 - INFO - codeparrot_training - Step 5576: {'lr': 0.0004931838539384681, 'samples': 2855424, 'steps': 5576, 'loss/train': 1.6719249486923218} 02/24/2022 08:46:56 - INFO - codeparrot_training - Step 5577: {'lr': 0.0004931800586734218, 'samples': 2855936, 'steps': 5577, 'loss/train': 2.5418787002563477} 02/24/2022 08:47:02 - INFO - codeparrot_training - Step 5578: {'lr': 0.0004931762623666692, 'samples': 2856448, 'steps': 5578, 'loss/train': 1.7851618528366089} 02/24/2022 08:47:05 - INFO - codeparrot_training - Step 5579: {'lr': 0.0004931724650182268, 'samples': 2856960, 'steps': 5579, 'loss/train': 0.44172510504722595} 02/24/2022 08:47:11 - INFO - codeparrot_training - Step 5580: {'lr': 0.0004931686666281108, 'samples': 2857472, 'steps': 5580, 'loss/train': 2.741643190383911} 02/24/2022 08:47:14 - INFO - codeparrot_training - Step 5581: {'lr': 0.0004931648671963373, 'samples': 2857984, 'steps': 5581, 'loss/train': 2.207061767578125} 02/24/2022 08:47:20 - INFO - codeparrot_training - Step 5582: {'lr': 0.000493161066722923, 'samples': 2858496, 'steps': 5582, 'loss/train': 1.6693904399871826} 02/24/2022 08:47:23 - INFO - codeparrot_training - Step 5583: {'lr': 0.0004931572652078837, 'samples': 2859008, 'steps': 5583, 'loss/train': 1.9169400930404663} 02/24/2022 08:47:29 - INFO - codeparrot_training - Step 5584: {'lr': 0.0004931534626512359, 'samples': 2859520, 'steps': 5584, 'loss/train': 3.8445916175842285} 02/24/2022 08:47:33 - INFO - codeparrot_training - Step 5585: {'lr': 0.0004931496590529959, 'samples': 2860032, 'steps': 5585, 'loss/train': 2.5335116386413574} 02/24/2022 08:47:38 - INFO - codeparrot_training - Step 5586: {'lr': 0.0004931458544131799, 'samples': 2860544, 'steps': 5586, 'loss/train': 3.1453921794891357} 02/24/2022 08:47:42 - INFO - codeparrot_training - Step 5587: {'lr': 0.0004931420487318044, 'samples': 2861056, 'steps': 5587, 'loss/train': 2.750887155532837} 02/24/2022 08:47:47 - INFO - codeparrot_training - Step 5588: {'lr': 0.0004931382420088855, 'samples': 2861568, 'steps': 5588, 'loss/train': 0.7489635348320007} 02/24/2022 08:47:51 - INFO - codeparrot_training - Step 5589: {'lr': 0.0004931344342444396, 'samples': 2862080, 'steps': 5589, 'loss/train': 2.8323230743408203} 02/24/2022 08:47:56 - INFO - codeparrot_training - Step 5590: {'lr': 0.000493130625438483, 'samples': 2862592, 'steps': 5590, 'loss/train': 1.8913551568984985} 02/24/2022 08:48:00 - INFO - codeparrot_training - Step 5591: {'lr': 0.000493126815591032, 'samples': 2863104, 'steps': 5591, 'loss/train': 3.3016164302825928} 02/24/2022 08:48:05 - INFO - codeparrot_training - Step 5592: {'lr': 0.0004931230047021028, 'samples': 2863616, 'steps': 5592, 'loss/train': 1.8535507917404175} 02/24/2022 08:48:09 - INFO - codeparrot_training - Step 5593: {'lr': 0.000493119192771712, 'samples': 2864128, 'steps': 5593, 'loss/train': 2.1000936031341553} 02/24/2022 08:48:15 - INFO - codeparrot_training - Step 5594: {'lr': 0.0004931153797998757, 'samples': 2864640, 'steps': 5594, 'loss/train': 1.4673081636428833} 02/24/2022 08:48:18 - INFO - codeparrot_training - Step 5595: {'lr': 0.0004931115657866103, 'samples': 2865152, 'steps': 5595, 'loss/train': 1.965951681137085} 02/24/2022 08:48:24 - INFO - codeparrot_training - Step 5596: {'lr': 0.0004931077507319322, 'samples': 2865664, 'steps': 5596, 'loss/train': 1.1762901544570923} 02/24/2022 08:48:27 - INFO - codeparrot_training - Step 5597: {'lr': 0.0004931039346358577, 'samples': 2866176, 'steps': 5597, 'loss/train': 2.6841182708740234} 02/24/2022 08:48:33 - INFO - codeparrot_training - Step 5598: {'lr': 0.0004931001174984032, 'samples': 2866688, 'steps': 5598, 'loss/train': 2.2216854095458984} 02/24/2022 08:48:36 - INFO - codeparrot_training - Step 5599: {'lr': 0.0004930962993195848, 'samples': 2867200, 'steps': 5599, 'loss/train': 2.5812947750091553} 02/24/2022 08:48:42 - INFO - codeparrot_training - Step 5600: {'lr': 0.0004930924800994192, 'samples': 2867712, 'steps': 5600, 'loss/train': 2.8477280139923096} 02/24/2022 08:48:45 - INFO - codeparrot_training - Step 5601: {'lr': 0.0004930886598379225, 'samples': 2868224, 'steps': 5601, 'loss/train': 1.5769703388214111} 02/24/2022 08:48:51 - INFO - codeparrot_training - Step 5602: {'lr': 0.0004930848385351112, 'samples': 2868736, 'steps': 5602, 'loss/train': 3.0094215869903564} 02/24/2022 08:48:54 - INFO - codeparrot_training - Step 5603: {'lr': 0.0004930810161910017, 'samples': 2869248, 'steps': 5603, 'loss/train': 2.762162208557129} 02/24/2022 08:49:00 - INFO - codeparrot_training - Step 5604: {'lr': 0.0004930771928056102, 'samples': 2869760, 'steps': 5604, 'loss/train': 2.9988224506378174} 02/24/2022 08:49:04 - INFO - codeparrot_training - Step 5605: {'lr': 0.0004930733683789533, 'samples': 2870272, 'steps': 5605, 'loss/train': 2.5743329524993896} 02/24/2022 08:49:09 - INFO - codeparrot_training - Step 5606: {'lr': 0.0004930695429110473, 'samples': 2870784, 'steps': 5606, 'loss/train': 2.397738218307495} 02/24/2022 08:49:13 - INFO - codeparrot_training - Step 5607: {'lr': 0.0004930657164019085, 'samples': 2871296, 'steps': 5607, 'loss/train': 3.0244300365448} 02/24/2022 08:49:18 - INFO - codeparrot_training - Step 5608: {'lr': 0.0004930618888515534, 'samples': 2871808, 'steps': 5608, 'loss/train': 2.8468844890594482} 02/24/2022 08:49:22 - INFO - codeparrot_training - Step 5609: {'lr': 0.0004930580602599983, 'samples': 2872320, 'steps': 5609, 'loss/train': 1.5977821350097656} 02/24/2022 08:49:28 - INFO - codeparrot_training - Step 5610: {'lr': 0.0004930542306272596, 'samples': 2872832, 'steps': 5610, 'loss/train': 2.6236846446990967} 02/24/2022 08:49:31 - INFO - codeparrot_training - Step 5611: {'lr': 0.0004930503999533538, 'samples': 2873344, 'steps': 5611, 'loss/train': 2.5708155632019043} 02/24/2022 08:49:37 - INFO - codeparrot_training - Step 5612: {'lr': 0.0004930465682382973, 'samples': 2873856, 'steps': 5612, 'loss/train': 1.275734305381775} 02/24/2022 08:49:40 - INFO - codeparrot_training - Step 5613: {'lr': 0.0004930427354821064, 'samples': 2874368, 'steps': 5613, 'loss/train': 1.1773325204849243} 02/24/2022 08:49:46 - INFO - codeparrot_training - Step 5614: {'lr': 0.0004930389016847977, 'samples': 2874880, 'steps': 5614, 'loss/train': 1.0492737293243408} 02/24/2022 08:49:49 - INFO - codeparrot_training - Step 5615: {'lr': 0.0004930350668463874, 'samples': 2875392, 'steps': 5615, 'loss/train': 2.9726195335388184} 02/24/2022 08:49:55 - INFO - codeparrot_training - Step 5616: {'lr': 0.0004930312309668922, 'samples': 2875904, 'steps': 5616, 'loss/train': 1.563391923904419} 02/24/2022 08:49:58 - INFO - codeparrot_training - Step 5617: {'lr': 0.0004930273940463283, 'samples': 2876416, 'steps': 5617, 'loss/train': 2.0094175338745117} 02/24/2022 08:50:04 - INFO - codeparrot_training - Step 5618: {'lr': 0.0004930235560847121, 'samples': 2876928, 'steps': 5618, 'loss/train': 2.9715640544891357} 02/24/2022 08:50:07 - INFO - codeparrot_training - Step 5619: {'lr': 0.0004930197170820603, 'samples': 2877440, 'steps': 5619, 'loss/train': 2.5607075691223145} 02/24/2022 08:50:13 - INFO - codeparrot_training - Step 5620: {'lr': 0.0004930158770383891, 'samples': 2877952, 'steps': 5620, 'loss/train': 0.5485643744468689} 02/24/2022 08:50:17 - INFO - codeparrot_training - Step 5621: {'lr': 0.0004930120359537153, 'samples': 2878464, 'steps': 5621, 'loss/train': 2.1816604137420654} 02/24/2022 08:50:22 - INFO - codeparrot_training - Step 5622: {'lr': 0.0004930081938280548, 'samples': 2878976, 'steps': 5622, 'loss/train': 1.880281686782837} 02/24/2022 08:50:26 - INFO - codeparrot_training - Step 5623: {'lr': 0.0004930043506614245, 'samples': 2879488, 'steps': 5623, 'loss/train': 2.9321084022521973} 02/24/2022 08:50:31 - INFO - codeparrot_training - Step 5624: {'lr': 0.0004930005064538406, 'samples': 2880000, 'steps': 5624, 'loss/train': 3.324977159500122} 02/24/2022 08:50:37 - INFO - codeparrot_training - Step 5625: {'lr': 0.0004929966612053199, 'samples': 2880512, 'steps': 5625, 'loss/train': 2.425710678100586} 02/24/2022 08:50:40 - INFO - codeparrot_training - Step 5626: {'lr': 0.0004929928149158785, 'samples': 2881024, 'steps': 5626, 'loss/train': 1.9112578630447388} 02/24/2022 08:50:46 - INFO - codeparrot_training - Step 5627: {'lr': 0.0004929889675855332, 'samples': 2881536, 'steps': 5627, 'loss/train': 2.19138240814209} 02/24/2022 08:50:49 - INFO - codeparrot_training - Step 5628: {'lr': 0.0004929851192143001, 'samples': 2882048, 'steps': 5628, 'loss/train': 2.4410400390625} 02/24/2022 08:50:56 - INFO - codeparrot_training - Step 5629: {'lr': 0.0004929812698021961, 'samples': 2882560, 'steps': 5629, 'loss/train': 3.1160783767700195} 02/24/2022 08:50:59 - INFO - codeparrot_training - Step 5630: {'lr': 0.0004929774193492373, 'samples': 2883072, 'steps': 5630, 'loss/train': 2.5817558765411377} 02/24/2022 08:51:05 - INFO - codeparrot_training - Step 5631: {'lr': 0.0004929735678554406, 'samples': 2883584, 'steps': 5631, 'loss/train': 2.8050694465637207} 02/24/2022 08:51:08 - INFO - codeparrot_training - Step 5632: {'lr': 0.0004929697153208221, 'samples': 2884096, 'steps': 5632, 'loss/train': 1.891499400138855} 02/24/2022 08:51:14 - INFO - codeparrot_training - Step 5633: {'lr': 0.0004929658617453986, 'samples': 2884608, 'steps': 5633, 'loss/train': 2.8051838874816895} 02/24/2022 08:51:17 - INFO - codeparrot_training - Step 5634: {'lr': 0.0004929620071291865, 'samples': 2885120, 'steps': 5634, 'loss/train': 2.0820560455322266} 02/24/2022 08:51:23 - INFO - codeparrot_training - Step 5635: {'lr': 0.0004929581514722023, 'samples': 2885632, 'steps': 5635, 'loss/train': 2.417781352996826} 02/24/2022 08:51:26 - INFO - codeparrot_training - Step 5636: {'lr': 0.0004929542947744625, 'samples': 2886144, 'steps': 5636, 'loss/train': 2.3993985652923584} 02/24/2022 08:51:32 - INFO - codeparrot_training - Step 5637: {'lr': 0.0004929504370359837, 'samples': 2886656, 'steps': 5637, 'loss/train': 2.83490252494812} 02/24/2022 08:51:35 - INFO - codeparrot_training - Step 5638: {'lr': 0.0004929465782567824, 'samples': 2887168, 'steps': 5638, 'loss/train': 3.0768420696258545} 02/24/2022 08:51:41 - INFO - codeparrot_training - Step 5639: {'lr': 0.000492942718436875, 'samples': 2887680, 'steps': 5639, 'loss/train': 2.4285888671875} 02/24/2022 08:51:44 - INFO - codeparrot_training - Step 5640: {'lr': 0.0004929388575762782, 'samples': 2888192, 'steps': 5640, 'loss/train': 1.8348264694213867} 02/24/2022 08:51:50 - INFO - codeparrot_training - Step 5641: {'lr': 0.0004929349956750085, 'samples': 2888704, 'steps': 5641, 'loss/train': 2.9415781497955322} 02/24/2022 08:51:54 - INFO - codeparrot_training - Step 5642: {'lr': 0.0004929311327330823, 'samples': 2889216, 'steps': 5642, 'loss/train': 1.738094687461853} 02/24/2022 08:51:59 - INFO - codeparrot_training - Step 5643: {'lr': 0.0004929272687505163, 'samples': 2889728, 'steps': 5643, 'loss/train': 3.554286003112793} 02/24/2022 08:52:03 - INFO - codeparrot_training - Step 5644: {'lr': 0.0004929234037273271, 'samples': 2890240, 'steps': 5644, 'loss/train': 2.46764874458313} 02/24/2022 08:52:08 - INFO - codeparrot_training - Step 5645: {'lr': 0.0004929195376635311, 'samples': 2890752, 'steps': 5645, 'loss/train': 2.29890513420105} 02/24/2022 08:52:12 - INFO - codeparrot_training - Step 5646: {'lr': 0.000492915670559145, 'samples': 2891264, 'steps': 5646, 'loss/train': 3.413724899291992} 02/24/2022 08:52:17 - INFO - codeparrot_training - Step 5647: {'lr': 0.0004929118024141853, 'samples': 2891776, 'steps': 5647, 'loss/train': 1.6830378770828247} 02/24/2022 08:52:21 - INFO - codeparrot_training - Step 5648: {'lr': 0.0004929079332286685, 'samples': 2892288, 'steps': 5648, 'loss/train': 2.7947282791137695} 02/24/2022 08:52:26 - INFO - codeparrot_training - Step 5649: {'lr': 0.0004929040630026112, 'samples': 2892800, 'steps': 5649, 'loss/train': 2.3367056846618652} 02/24/2022 08:52:30 - INFO - codeparrot_training - Step 5650: {'lr': 0.0004929001917360302, 'samples': 2893312, 'steps': 5650, 'loss/train': 2.1625170707702637} 02/24/2022 08:52:35 - INFO - codeparrot_training - Step 5651: {'lr': 0.0004928963194289419, 'samples': 2893824, 'steps': 5651, 'loss/train': 2.3080220222473145} 02/24/2022 08:52:39 - INFO - codeparrot_training - Step 5652: {'lr': 0.0004928924460813627, 'samples': 2894336, 'steps': 5652, 'loss/train': 2.702507257461548} 02/24/2022 08:52:44 - INFO - codeparrot_training - Step 5653: {'lr': 0.0004928885716933096, 'samples': 2894848, 'steps': 5653, 'loss/train': 3.0107593536376953} 02/24/2022 08:52:48 - INFO - codeparrot_training - Step 5654: {'lr': 0.0004928846962647988, 'samples': 2895360, 'steps': 5654, 'loss/train': 2.6054599285125732} 02/24/2022 08:52:54 - INFO - codeparrot_training - Step 5655: {'lr': 0.0004928808197958472, 'samples': 2895872, 'steps': 5655, 'loss/train': 1.2903093099594116} 02/24/2022 08:52:57 - INFO - codeparrot_training - Step 5656: {'lr': 0.0004928769422864712, 'samples': 2896384, 'steps': 5656, 'loss/train': 2.8213772773742676} 02/24/2022 08:53:03 - INFO - codeparrot_training - Step 5657: {'lr': 0.0004928730637366877, 'samples': 2896896, 'steps': 5657, 'loss/train': 1.6182501316070557} 02/24/2022 08:53:06 - INFO - codeparrot_training - Step 5658: {'lr': 0.000492869184146513, 'samples': 2897408, 'steps': 5658, 'loss/train': 1.7428618669509888} 02/24/2022 08:53:12 - INFO - codeparrot_training - Step 5659: {'lr': 0.0004928653035159638, 'samples': 2897920, 'steps': 5659, 'loss/train': 1.595118522644043} 02/24/2022 08:53:16 - INFO - codeparrot_training - Step 5660: {'lr': 0.0004928614218450568, 'samples': 2898432, 'steps': 5660, 'loss/train': 2.697014570236206} 02/24/2022 08:53:21 - INFO - codeparrot_training - Step 5661: {'lr': 0.0004928575391338085, 'samples': 2898944, 'steps': 5661, 'loss/train': 1.9353317022323608} 02/24/2022 08:53:25 - INFO - codeparrot_training - Step 5662: {'lr': 0.0004928536553822357, 'samples': 2899456, 'steps': 5662, 'loss/train': 3.699150800704956} 02/24/2022 08:53:30 - INFO - codeparrot_training - Step 5663: {'lr': 0.0004928497705903549, 'samples': 2899968, 'steps': 5663, 'loss/train': 3.097060441970825} 02/24/2022 08:53:34 - INFO - codeparrot_training - Step 5664: {'lr': 0.0004928458847581828, 'samples': 2900480, 'steps': 5664, 'loss/train': 2.043487548828125} 02/24/2022 08:53:39 - INFO - codeparrot_training - Step 5665: {'lr': 0.0004928419978857361, 'samples': 2900992, 'steps': 5665, 'loss/train': 2.7488622665405273} 02/24/2022 08:53:43 - INFO - codeparrot_training - Step 5666: {'lr': 0.0004928381099730314, 'samples': 2901504, 'steps': 5666, 'loss/train': 2.8547468185424805} 02/24/2022 08:53:49 - INFO - codeparrot_training - Step 5667: {'lr': 0.0004928342210200853, 'samples': 2902016, 'steps': 5667, 'loss/train': 1.4548543691635132} 02/24/2022 08:53:53 - INFO - codeparrot_training - Step 5668: {'lr': 0.0004928303310269145, 'samples': 2902528, 'steps': 5668, 'loss/train': 2.502410411834717} 02/24/2022 08:53:58 - INFO - codeparrot_training - Step 5669: {'lr': 0.0004928264399935357, 'samples': 2903040, 'steps': 5669, 'loss/train': 2.077807664871216} 02/24/2022 08:54:01 - INFO - codeparrot_training - Step 5670: {'lr': 0.0004928225479199655, 'samples': 2903552, 'steps': 5670, 'loss/train': 2.462615728378296} 02/24/2022 08:54:07 - INFO - codeparrot_training - Step 5671: {'lr': 0.0004928186548062206, 'samples': 2904064, 'steps': 5671, 'loss/train': 3.0940253734588623} 02/24/2022 08:54:10 - INFO - codeparrot_training - Step 5672: {'lr': 0.0004928147606523179, 'samples': 2904576, 'steps': 5672, 'loss/train': 2.4129199981689453} 02/24/2022 08:54:16 - INFO - codeparrot_training - Step 5673: {'lr': 0.0004928108654582736, 'samples': 2905088, 'steps': 5673, 'loss/train': 3.663492202758789} 02/24/2022 08:54:19 - INFO - codeparrot_training - Step 5674: {'lr': 0.0004928069692241048, 'samples': 2905600, 'steps': 5674, 'loss/train': 2.200190544128418} 02/24/2022 08:54:25 - INFO - codeparrot_training - Step 5675: {'lr': 0.000492803071949828, 'samples': 2906112, 'steps': 5675, 'loss/train': 3.141918659210205} 02/24/2022 08:54:28 - INFO - codeparrot_training - Step 5676: {'lr': 0.0004927991736354599, 'samples': 2906624, 'steps': 5676, 'loss/train': 2.6103997230529785} 02/24/2022 08:54:35 - INFO - codeparrot_training - Step 5677: {'lr': 0.0004927952742810173, 'samples': 2907136, 'steps': 5677, 'loss/train': 3.0645596981048584} 02/24/2022 08:54:38 - INFO - codeparrot_training - Step 5678: {'lr': 0.0004927913738865167, 'samples': 2907648, 'steps': 5678, 'loss/train': 2.324155569076538} 02/24/2022 08:54:43 - INFO - codeparrot_training - Step 5679: {'lr': 0.0004927874724519751, 'samples': 2908160, 'steps': 5679, 'loss/train': 2.973525047302246} 02/24/2022 08:54:47 - INFO - codeparrot_training - Step 5680: {'lr': 0.000492783569977409, 'samples': 2908672, 'steps': 5680, 'loss/train': 1.9238383769989014} 02/24/2022 08:54:52 - INFO - codeparrot_training - Step 5681: {'lr': 0.0004927796664628353, 'samples': 2909184, 'steps': 5681, 'loss/train': 2.038911819458008} 02/24/2022 08:54:56 - INFO - codeparrot_training - Step 5682: {'lr': 0.0004927757619082704, 'samples': 2909696, 'steps': 5682, 'loss/train': 1.9652752876281738} 02/24/2022 08:55:01 - INFO - codeparrot_training - Step 5683: {'lr': 0.0004927718563137313, 'samples': 2910208, 'steps': 5683, 'loss/train': 2.848663330078125} 02/24/2022 08:55:05 - INFO - codeparrot_training - Step 5684: {'lr': 0.0004927679496792347, 'samples': 2910720, 'steps': 5684, 'loss/train': 2.696091413497925} 02/24/2022 08:55:10 - INFO - codeparrot_training - Step 5685: {'lr': 0.0004927640420047973, 'samples': 2911232, 'steps': 5685, 'loss/train': 0.9404380917549133} 02/24/2022 08:55:14 - INFO - codeparrot_training - Step 5686: {'lr': 0.0004927601332904358, 'samples': 2911744, 'steps': 5686, 'loss/train': 2.1705567836761475} 02/24/2022 08:55:20 - INFO - codeparrot_training - Step 5687: {'lr': 0.0004927562235361669, 'samples': 2912256, 'steps': 5687, 'loss/train': 2.154585123062134} 02/24/2022 08:55:24 - INFO - codeparrot_training - Step 5688: {'lr': 0.0004927523127420076, 'samples': 2912768, 'steps': 5688, 'loss/train': 2.619330644607544} 02/24/2022 08:55:29 - INFO - codeparrot_training - Step 5689: {'lr': 0.0004927484009079743, 'samples': 2913280, 'steps': 5689, 'loss/train': 2.8479294776916504} 02/24/2022 08:55:33 - INFO - codeparrot_training - Step 5690: {'lr': 0.000492744488034084, 'samples': 2913792, 'steps': 5690, 'loss/train': 2.0539581775665283} 02/24/2022 08:55:38 - INFO - codeparrot_training - Step 5691: {'lr': 0.0004927405741203534, 'samples': 2914304, 'steps': 5691, 'loss/train': 3.403379201889038} 02/24/2022 08:55:42 - INFO - codeparrot_training - Step 5692: {'lr': 0.0004927366591667993, 'samples': 2914816, 'steps': 5692, 'loss/train': 1.4618353843688965} 02/24/2022 08:55:47 - INFO - codeparrot_training - Step 5693: {'lr': 0.0004927327431734383, 'samples': 2915328, 'steps': 5693, 'loss/train': 2.345198392868042} 02/24/2022 08:55:50 - INFO - codeparrot_training - Step 5694: {'lr': 0.0004927288261402875, 'samples': 2915840, 'steps': 5694, 'loss/train': 1.8501535654067993} 02/24/2022 08:55:56 - INFO - codeparrot_training - Step 5695: {'lr': 0.0004927249080673633, 'samples': 2916352, 'steps': 5695, 'loss/train': 2.8291263580322266} 02/24/2022 08:55:59 - INFO - codeparrot_training - Step 5696: {'lr': 0.0004927209889546828, 'samples': 2916864, 'steps': 5696, 'loss/train': 2.2671685218811035} 02/24/2022 08:56:05 - INFO - codeparrot_training - Step 5697: {'lr': 0.0004927170688022625, 'samples': 2917376, 'steps': 5697, 'loss/train': 1.8576054573059082} 02/24/2022 08:56:08 - INFO - codeparrot_training - Step 5698: {'lr': 0.0004927131476101195, 'samples': 2917888, 'steps': 5698, 'loss/train': 2.2511613368988037} 02/24/2022 08:56:14 - INFO - codeparrot_training - Step 5699: {'lr': 0.0004927092253782704, 'samples': 2918400, 'steps': 5699, 'loss/train': 3.254352331161499} 02/24/2022 08:56:17 - INFO - codeparrot_training - Step 5700: {'lr': 0.0004927053021067321, 'samples': 2918912, 'steps': 5700, 'loss/train': 2.630722999572754} 02/24/2022 08:56:23 - INFO - codeparrot_training - Step 5701: {'lr': 0.0004927013777955212, 'samples': 2919424, 'steps': 5701, 'loss/train': 2.9244372844696045} 02/24/2022 08:56:26 - INFO - codeparrot_training - Step 5702: {'lr': 0.0004926974524446548, 'samples': 2919936, 'steps': 5702, 'loss/train': 2.6032893657684326} 02/24/2022 08:56:33 - INFO - codeparrot_training - Step 5703: {'lr': 0.0004926935260541496, 'samples': 2920448, 'steps': 5703, 'loss/train': 2.5975043773651123} 02/24/2022 08:56:37 - INFO - codeparrot_training - Step 5704: {'lr': 0.0004926895986240222, 'samples': 2920960, 'steps': 5704, 'loss/train': 3.2763800621032715} 02/24/2022 08:56:42 - INFO - codeparrot_training - Step 5705: {'lr': 0.0004926856701542898, 'samples': 2921472, 'steps': 5705, 'loss/train': 1.6149892807006836} 02/24/2022 08:56:46 - INFO - codeparrot_training - Step 5706: {'lr': 0.000492681740644969, 'samples': 2921984, 'steps': 5706, 'loss/train': 3.00669264793396} 02/24/2022 08:56:51 - INFO - codeparrot_training - Step 5707: {'lr': 0.0004926778100960767, 'samples': 2922496, 'steps': 5707, 'loss/train': 2.7209267616271973} 02/24/2022 08:56:55 - INFO - codeparrot_training - Step 5708: {'lr': 0.0004926738785076297, 'samples': 2923008, 'steps': 5708, 'loss/train': 1.271486759185791} 02/24/2022 08:57:00 - INFO - codeparrot_training - Step 5709: {'lr': 0.0004926699458796448, 'samples': 2923520, 'steps': 5709, 'loss/train': 2.8055872917175293} 02/24/2022 08:57:04 - INFO - codeparrot_training - Step 5710: {'lr': 0.0004926660122121391, 'samples': 2924032, 'steps': 5710, 'loss/train': 1.2185041904449463} 02/24/2022 08:57:09 - INFO - codeparrot_training - Step 5711: {'lr': 0.0004926620775051291, 'samples': 2924544, 'steps': 5711, 'loss/train': 1.8776848316192627} 02/24/2022 08:57:13 - INFO - codeparrot_training - Step 5712: {'lr': 0.0004926581417586318, 'samples': 2925056, 'steps': 5712, 'loss/train': 1.6651860475540161} 02/24/2022 08:57:19 - INFO - codeparrot_training - Step 5713: {'lr': 0.0004926542049726642, 'samples': 2925568, 'steps': 5713, 'loss/train': 3.009711503982544} 02/24/2022 08:57:22 - INFO - codeparrot_training - Step 5714: {'lr': 0.0004926502671472429, 'samples': 2926080, 'steps': 5714, 'loss/train': 3.661102294921875} 02/24/2022 08:57:28 - INFO - codeparrot_training - Step 5715: {'lr': 0.000492646328282385, 'samples': 2926592, 'steps': 5715, 'loss/train': 3.2587244510650635} 02/24/2022 08:57:31 - INFO - codeparrot_training - Step 5716: {'lr': 0.0004926423883781073, 'samples': 2927104, 'steps': 5716, 'loss/train': 2.594268321990967} 02/24/2022 08:57:37 - INFO - codeparrot_training - Step 5717: {'lr': 0.0004926384474344265, 'samples': 2927616, 'steps': 5717, 'loss/train': 2.8316903114318848} 02/24/2022 08:57:40 - INFO - codeparrot_training - Step 5718: {'lr': 0.0004926345054513598, 'samples': 2928128, 'steps': 5718, 'loss/train': 2.4430084228515625} 02/24/2022 08:57:46 - INFO - codeparrot_training - Step 5719: {'lr': 0.0004926305624289238, 'samples': 2928640, 'steps': 5719, 'loss/train': 2.468344211578369} 02/24/2022 08:57:49 - INFO - codeparrot_training - Step 5720: {'lr': 0.0004926266183671356, 'samples': 2929152, 'steps': 5720, 'loss/train': 2.630418539047241} 02/24/2022 08:57:55 - INFO - codeparrot_training - Step 5721: {'lr': 0.000492622673266012, 'samples': 2929664, 'steps': 5721, 'loss/train': 0.7120659351348877} 02/24/2022 08:58:01 - INFO - codeparrot_training - Step 5722: {'lr': 0.0004926187271255698, 'samples': 2930176, 'steps': 5722, 'loss/train': 1.9260408878326416} 02/24/2022 08:58:04 - INFO - codeparrot_training - Step 5723: {'lr': 0.0004926147799458262, 'samples': 2930688, 'steps': 5723, 'loss/train': 1.9589388370513916} 02/24/2022 08:58:10 - INFO - codeparrot_training - Step 5724: {'lr': 0.0004926108317267979, 'samples': 2931200, 'steps': 5724, 'loss/train': 2.889728546142578} 02/24/2022 08:58:13 - INFO - codeparrot_training - Step 5725: {'lr': 0.0004926068824685017, 'samples': 2931712, 'steps': 5725, 'loss/train': 2.8991611003875732} 02/24/2022 08:58:19 - INFO - codeparrot_training - Step 5726: {'lr': 0.0004926029321709548, 'samples': 2932224, 'steps': 5726, 'loss/train': 1.6453745365142822} 02/24/2022 08:58:22 - INFO - codeparrot_training - Step 5727: {'lr': 0.0004925989808341738, 'samples': 2932736, 'steps': 5727, 'loss/train': 2.4324159622192383} 02/24/2022 08:58:28 - INFO - codeparrot_training - Step 5728: {'lr': 0.0004925950284581759, 'samples': 2933248, 'steps': 5728, 'loss/train': 2.519422769546509} 02/24/2022 08:58:31 - INFO - codeparrot_training - Step 5729: {'lr': 0.0004925910750429779, 'samples': 2933760, 'steps': 5729, 'loss/train': 3.4216084480285645} 02/24/2022 08:58:37 - INFO - codeparrot_training - Step 5730: {'lr': 0.0004925871205885968, 'samples': 2934272, 'steps': 5730, 'loss/train': 3.5515549182891846} 02/24/2022 08:58:40 - INFO - codeparrot_training - Step 5731: {'lr': 0.0004925831650950495, 'samples': 2934784, 'steps': 5731, 'loss/train': 2.1032602787017822} 02/24/2022 08:58:46 - INFO - codeparrot_training - Step 5732: {'lr': 0.000492579208562353, 'samples': 2935296, 'steps': 5732, 'loss/train': 1.753894329071045} 02/24/2022 08:58:49 - INFO - codeparrot_training - Step 5733: {'lr': 0.0004925752509905241, 'samples': 2935808, 'steps': 5733, 'loss/train': 3.2489020824432373} 02/24/2022 08:58:55 - INFO - codeparrot_training - Step 5734: {'lr': 0.0004925712923795799, 'samples': 2936320, 'steps': 5734, 'loss/train': 3.059749126434326} 02/24/2022 08:58:59 - INFO - codeparrot_training - Step 5735: {'lr': 0.0004925673327295374, 'samples': 2936832, 'steps': 5735, 'loss/train': 2.2254936695098877} 02/24/2022 08:59:04 - INFO - codeparrot_training - Step 5736: {'lr': 0.0004925633720404132, 'samples': 2937344, 'steps': 5736, 'loss/train': 2.7345802783966064} 02/24/2022 08:59:08 - INFO - codeparrot_training - Step 5737: {'lr': 0.0004925594103122248, 'samples': 2937856, 'steps': 5737, 'loss/train': 2.271477222442627} 02/24/2022 08:59:13 - INFO - codeparrot_training - Step 5738: {'lr': 0.0004925554475449888, 'samples': 2938368, 'steps': 5738, 'loss/train': 1.21036696434021} 02/24/2022 08:59:17 - INFO - codeparrot_training - Step 5739: {'lr': 0.0004925514837387223, 'samples': 2938880, 'steps': 5739, 'loss/train': 3.161942720413208} 02/24/2022 08:59:22 - INFO - codeparrot_training - Step 5740: {'lr': 0.0004925475188934423, 'samples': 2939392, 'steps': 5740, 'loss/train': 1.8844009637832642} 02/24/2022 08:59:26 - INFO - codeparrot_training - Step 5741: {'lr': 0.0004925435530091656, 'samples': 2939904, 'steps': 5741, 'loss/train': 2.592146635055542} 02/24/2022 08:59:31 - INFO - codeparrot_training - Step 5742: {'lr': 0.0004925395860859096, 'samples': 2940416, 'steps': 5742, 'loss/train': 2.1591148376464844} 02/24/2022 08:59:35 - INFO - codeparrot_training - Step 5743: {'lr': 0.0004925356181236908, 'samples': 2940928, 'steps': 5743, 'loss/train': 2.389815330505371} 02/24/2022 08:59:40 - INFO - codeparrot_training - Step 5744: {'lr': 0.0004925316491225265, 'samples': 2941440, 'steps': 5744, 'loss/train': 2.967702627182007} 02/24/2022 08:59:44 - INFO - codeparrot_training - Step 5745: {'lr': 0.0004925276790824336, 'samples': 2941952, 'steps': 5745, 'loss/train': 2.732590913772583} 02/24/2022 08:59:49 - INFO - codeparrot_training - Step 5746: {'lr': 0.0004925237080034291, 'samples': 2942464, 'steps': 5746, 'loss/train': 3.098123788833618} 02/24/2022 08:59:53 - INFO - codeparrot_training - Step 5747: {'lr': 0.0004925197358855301, 'samples': 2942976, 'steps': 5747, 'loss/train': 1.696014165878296} 02/24/2022 08:59:59 - INFO - codeparrot_training - Step 5748: {'lr': 0.0004925157627287536, 'samples': 2943488, 'steps': 5748, 'loss/train': 1.1901603937149048} 02/24/2022 09:00:02 - INFO - codeparrot_training - Step 5749: {'lr': 0.0004925117885331166, 'samples': 2944000, 'steps': 5749, 'loss/train': 2.0456440448760986} 02/24/2022 09:00:08 - INFO - codeparrot_training - Step 5750: {'lr': 0.000492507813298636, 'samples': 2944512, 'steps': 5750, 'loss/train': 1.910574197769165} 02/24/2022 09:00:11 - INFO - codeparrot_training - Step 5751: {'lr': 0.000492503837025329, 'samples': 2945024, 'steps': 5751, 'loss/train': 2.3957343101501465} 02/24/2022 09:00:17 - INFO - codeparrot_training - Step 5752: {'lr': 0.0004924998597132125, 'samples': 2945536, 'steps': 5752, 'loss/train': 2.528535842895508} 02/24/2022 09:00:21 - INFO - codeparrot_training - Step 5753: {'lr': 0.0004924958813623037, 'samples': 2946048, 'steps': 5753, 'loss/train': 2.163933753967285} 02/24/2022 09:00:26 - INFO - codeparrot_training - Step 5754: {'lr': 0.0004924919019726195, 'samples': 2946560, 'steps': 5754, 'loss/train': 2.6796388626098633} 02/24/2022 09:00:29 - INFO - codeparrot_training - Step 5755: {'lr': 0.000492487921544177, 'samples': 2947072, 'steps': 5755, 'loss/train': 2.167558193206787} 02/24/2022 09:00:35 - INFO - codeparrot_training - Step 5756: {'lr': 0.0004924839400769932, 'samples': 2947584, 'steps': 5756, 'loss/train': 2.3096165657043457} 02/24/2022 09:00:38 - INFO - codeparrot_training - Step 5757: {'lr': 0.0004924799575710852, 'samples': 2948096, 'steps': 5757, 'loss/train': 1.3764539957046509} 02/24/2022 09:00:45 - INFO - codeparrot_training - Step 5758: {'lr': 0.0004924759740264701, 'samples': 2948608, 'steps': 5758, 'loss/train': 2.897132635116577} 02/24/2022 09:00:48 - INFO - codeparrot_training - Step 5759: {'lr': 0.000492471989443165, 'samples': 2949120, 'steps': 5759, 'loss/train': 2.8032610416412354} 02/24/2022 09:00:54 - INFO - codeparrot_training - Step 5760: {'lr': 0.0004924680038211868, 'samples': 2949632, 'steps': 5760, 'loss/train': 1.8311803340911865} 02/24/2022 09:00:57 - INFO - codeparrot_training - Step 5761: {'lr': 0.0004924640171605526, 'samples': 2950144, 'steps': 5761, 'loss/train': 2.805656671524048} 02/24/2022 09:01:02 - INFO - codeparrot_training - Step 5762: {'lr': 0.0004924600294612796, 'samples': 2950656, 'steps': 5762, 'loss/train': 3.264230966567993} 02/24/2022 09:01:06 - INFO - codeparrot_training - Step 5763: {'lr': 0.0004924560407233848, 'samples': 2951168, 'steps': 5763, 'loss/train': 2.720665454864502} 02/24/2022 09:01:11 - INFO - codeparrot_training - Step 5764: {'lr': 0.0004924520509468854, 'samples': 2951680, 'steps': 5764, 'loss/train': 1.5945589542388916} 02/24/2022 09:01:15 - INFO - codeparrot_training - Step 5765: {'lr': 0.0004924480601317982, 'samples': 2952192, 'steps': 5765, 'loss/train': 3.415029525756836} 02/24/2022 09:01:20 - INFO - codeparrot_training - Step 5766: {'lr': 0.0004924440682781407, 'samples': 2952704, 'steps': 5766, 'loss/train': 2.8304691314697266} 02/24/2022 09:01:26 - INFO - codeparrot_training - Step 5767: {'lr': 0.0004924400753859297, 'samples': 2953216, 'steps': 5767, 'loss/train': 3.212505578994751} 02/24/2022 09:01:29 - INFO - codeparrot_training - Step 5768: {'lr': 0.0004924360814551825, 'samples': 2953728, 'steps': 5768, 'loss/train': 2.7774264812469482} 02/24/2022 09:01:33 - INFO - codeparrot_training - Step 5769: {'lr': 0.000492432086485916, 'samples': 2954240, 'steps': 5769, 'loss/train': 2.588395357131958} 02/24/2022 09:01:39 - INFO - codeparrot_training - Step 5770: {'lr': 0.0004924280904781475, 'samples': 2954752, 'steps': 5770, 'loss/train': 1.9354759454727173} 02/24/2022 09:01:45 - INFO - codeparrot_training - Step 5771: {'lr': 0.0004924240934318939, 'samples': 2955264, 'steps': 5771, 'loss/train': 2.613406181335449} 02/24/2022 09:01:48 - INFO - codeparrot_training - Step 5772: {'lr': 0.0004924200953471727, 'samples': 2955776, 'steps': 5772, 'loss/train': 2.8672473430633545} 02/24/2022 09:01:51 - INFO - codeparrot_training - Step 5773: {'lr': 0.0004924160962240005, 'samples': 2956288, 'steps': 5773, 'loss/train': 2.887533664703369} 02/24/2022 09:01:57 - INFO - codeparrot_training - Step 5774: {'lr': 0.0004924120960623949, 'samples': 2956800, 'steps': 5774, 'loss/train': 2.5469424724578857} 02/24/2022 09:02:00 - INFO - codeparrot_training - Step 5775: {'lr': 0.0004924080948623729, 'samples': 2957312, 'steps': 5775, 'loss/train': 2.3984427452087402} 02/24/2022 09:02:06 - INFO - codeparrot_training - Step 5776: {'lr': 0.0004924040926239515, 'samples': 2957824, 'steps': 5776, 'loss/train': 2.1007635593414307} 02/24/2022 09:02:09 - INFO - codeparrot_training - Step 5777: {'lr': 0.000492400089347148, 'samples': 2958336, 'steps': 5777, 'loss/train': 1.9850976467132568} 02/24/2022 09:02:15 - INFO - codeparrot_training - Step 5778: {'lr': 0.0004923960850319794, 'samples': 2958848, 'steps': 5778, 'loss/train': 2.062241315841675} 02/24/2022 09:02:18 - INFO - codeparrot_training - Step 5779: {'lr': 0.000492392079678463, 'samples': 2959360, 'steps': 5779, 'loss/train': 3.2259674072265625} 02/24/2022 09:02:24 - INFO - codeparrot_training - Step 5780: {'lr': 0.0004923880732866159, 'samples': 2959872, 'steps': 5780, 'loss/train': 2.1912457942962646} 02/24/2022 09:02:28 - INFO - codeparrot_training - Step 5781: {'lr': 0.0004923840658564553, 'samples': 2960384, 'steps': 5781, 'loss/train': 2.5156610012054443} 02/24/2022 09:02:33 - INFO - codeparrot_training - Step 5782: {'lr': 0.0004923800573879983, 'samples': 2960896, 'steps': 5782, 'loss/train': 1.65301513671875} 02/24/2022 09:02:39 - INFO - codeparrot_training - Step 5783: {'lr': 0.000492376047881262, 'samples': 2961408, 'steps': 5783, 'loss/train': 1.8982449769973755} 02/24/2022 09:02:42 - INFO - codeparrot_training - Step 5784: {'lr': 0.0004923720373362638, 'samples': 2961920, 'steps': 5784, 'loss/train': 2.152921199798584} 02/24/2022 09:02:48 - INFO - codeparrot_training - Step 5785: {'lr': 0.0004923680257530207, 'samples': 2962432, 'steps': 5785, 'loss/train': 2.8000853061676025} 02/24/2022 09:02:52 - INFO - codeparrot_training - Step 5786: {'lr': 0.0004923640131315499, 'samples': 2962944, 'steps': 5786, 'loss/train': 0.4488138258457184} 02/24/2022 09:02:55 - INFO - codeparrot_training - Step 5787: {'lr': 0.0004923599994718687, 'samples': 2963456, 'steps': 5787, 'loss/train': 0.4960688352584839} 02/24/2022 09:03:01 - INFO - codeparrot_training - Step 5788: {'lr': 0.0004923559847739941, 'samples': 2963968, 'steps': 5788, 'loss/train': 2.007481336593628} 02/24/2022 09:03:04 - INFO - codeparrot_training - Step 5789: {'lr': 0.0004923519690379436, 'samples': 2964480, 'steps': 5789, 'loss/train': 2.0681345462799072} 02/24/2022 09:03:10 - INFO - codeparrot_training - Step 5790: {'lr': 0.0004923479522637341, 'samples': 2964992, 'steps': 5790, 'loss/train': 2.125170946121216} 02/24/2022 09:03:14 - INFO - codeparrot_training - Step 5791: {'lr': 0.0004923439344513829, 'samples': 2965504, 'steps': 5791, 'loss/train': 1.9498525857925415} 02/24/2022 09:03:19 - INFO - codeparrot_training - Step 5792: {'lr': 0.0004923399156009073, 'samples': 2966016, 'steps': 5792, 'loss/train': 2.004350185394287} 02/24/2022 09:03:23 - INFO - codeparrot_training - Step 5793: {'lr': 0.0004923358957123245, 'samples': 2966528, 'steps': 5793, 'loss/train': 3.4744937419891357} 02/24/2022 09:03:29 - INFO - codeparrot_training - Step 5794: {'lr': 0.0004923318747856515, 'samples': 2967040, 'steps': 5794, 'loss/train': 2.7088301181793213} 02/24/2022 09:03:32 - INFO - codeparrot_training - Step 5795: {'lr': 0.0004923278528209059, 'samples': 2967552, 'steps': 5795, 'loss/train': 3.662992000579834} 02/24/2022 09:03:38 - INFO - codeparrot_training - Step 5796: {'lr': 0.0004923238298181047, 'samples': 2968064, 'steps': 5796, 'loss/train': 2.6054458618164062} 02/24/2022 09:03:41 - INFO - codeparrot_training - Step 5797: {'lr': 0.0004923198057772651, 'samples': 2968576, 'steps': 5797, 'loss/train': 0.2809189558029175} 02/24/2022 09:03:47 - INFO - codeparrot_training - Step 5798: {'lr': 0.0004923157806984044, 'samples': 2969088, 'steps': 5798, 'loss/train': 2.4262919425964355} 02/24/2022 09:03:50 - INFO - codeparrot_training - Step 5799: {'lr': 0.0004923117545815398, 'samples': 2969600, 'steps': 5799, 'loss/train': 1.0630701780319214} 02/24/2022 09:03:56 - INFO - codeparrot_training - Step 5800: {'lr': 0.0004923077274266886, 'samples': 2970112, 'steps': 5800, 'loss/train': 2.783456802368164} 02/24/2022 09:03:59 - INFO - codeparrot_training - Step 5801: {'lr': 0.0004923036992338681, 'samples': 2970624, 'steps': 5801, 'loss/train': 2.1558947563171387} 02/24/2022 09:04:05 - INFO - codeparrot_training - Step 5802: {'lr': 0.0004922996700030954, 'samples': 2971136, 'steps': 5802, 'loss/train': 2.67423939704895} 02/24/2022 09:04:09 - INFO - codeparrot_training - Step 5803: {'lr': 0.000492295639734388, 'samples': 2971648, 'steps': 5803, 'loss/train': 8.31851863861084} 02/24/2022 09:04:15 - INFO - codeparrot_training - Step 5804: {'lr': 0.0004922916084277629, 'samples': 2972160, 'steps': 5804, 'loss/train': 2.4561474323272705} 02/24/2022 09:04:18 - INFO - codeparrot_training - Step 5805: {'lr': 0.0004922875760832375, 'samples': 2972672, 'steps': 5805, 'loss/train': 2.176706075668335} 02/24/2022 09:04:24 - INFO - codeparrot_training - Step 5806: {'lr': 0.000492283542700829, 'samples': 2973184, 'steps': 5806, 'loss/train': 3.096372604370117} 02/24/2022 09:04:27 - INFO - codeparrot_training - Step 5807: {'lr': 0.0004922795082805549, 'samples': 2973696, 'steps': 5807, 'loss/train': 1.5772497653961182} 02/24/2022 09:04:33 - INFO - codeparrot_training - Step 5808: {'lr': 0.0004922754728224322, 'samples': 2974208, 'steps': 5808, 'loss/train': 2.438209056854248} 02/24/2022 09:04:36 - INFO - codeparrot_training - Step 5809: {'lr': 0.0004922714363264783, 'samples': 2974720, 'steps': 5809, 'loss/train': 0.3431764543056488} 02/24/2022 09:04:42 - INFO - codeparrot_training - Step 5810: {'lr': 0.0004922673987927106, 'samples': 2975232, 'steps': 5810, 'loss/train': 1.535483479499817} 02/24/2022 09:04:45 - INFO - codeparrot_training - Step 5811: {'lr': 0.0004922633602211462, 'samples': 2975744, 'steps': 5811, 'loss/train': 2.461210012435913} 02/24/2022 09:04:51 - INFO - codeparrot_training - Step 5812: {'lr': 0.0004922593206118025, 'samples': 2976256, 'steps': 5812, 'loss/train': 2.38917875289917} 02/24/2022 09:04:54 - INFO - codeparrot_training - Step 5813: {'lr': 0.0004922552799646968, 'samples': 2976768, 'steps': 5813, 'loss/train': 3.163112163543701} 02/24/2022 09:04:59 - INFO - codeparrot_training - Step 5814: {'lr': 0.0004922512382798463, 'samples': 2977280, 'steps': 5814, 'loss/train': 2.611781358718872} 02/24/2022 09:05:03 - INFO - codeparrot_training - Step 5815: {'lr': 0.0004922471955572686, 'samples': 2977792, 'steps': 5815, 'loss/train': 2.7317473888397217} 02/24/2022 09:05:09 - INFO - codeparrot_training - Step 5816: {'lr': 0.0004922431517969808, 'samples': 2978304, 'steps': 5816, 'loss/train': 1.324861764907837} 02/24/2022 09:05:13 - INFO - codeparrot_training - Step 5817: {'lr': 0.0004922391069990002, 'samples': 2978816, 'steps': 5817, 'loss/train': 3.162651777267456} 02/24/2022 09:05:18 - INFO - codeparrot_training - Step 5818: {'lr': 0.0004922350611633442, 'samples': 2979328, 'steps': 5818, 'loss/train': 1.4453734159469604} 02/24/2022 09:05:22 - INFO - codeparrot_training - Step 5819: {'lr': 0.0004922310142900302, 'samples': 2979840, 'steps': 5819, 'loss/train': 2.142991542816162} 02/24/2022 09:05:27 - INFO - codeparrot_training - Step 5820: {'lr': 0.0004922269663790753, 'samples': 2980352, 'steps': 5820, 'loss/train': 2.0249812602996826} 02/24/2022 09:05:31 - INFO - codeparrot_training - Step 5821: {'lr': 0.0004922229174304971, 'samples': 2980864, 'steps': 5821, 'loss/train': 2.174739360809326} 02/24/2022 09:05:36 - INFO - codeparrot_training - Step 5822: {'lr': 0.0004922188674443128, 'samples': 2981376, 'steps': 5822, 'loss/train': 2.2752199172973633} 02/24/2022 09:05:40 - INFO - codeparrot_training - Step 5823: {'lr': 0.0004922148164205398, 'samples': 2981888, 'steps': 5823, 'loss/train': 2.7619760036468506} 02/24/2022 09:05:45 - INFO - codeparrot_training - Step 5824: {'lr': 0.0004922107643591954, 'samples': 2982400, 'steps': 5824, 'loss/train': 2.9031381607055664} 02/24/2022 09:05:49 - INFO - codeparrot_training - Step 5825: {'lr': 0.000492206711260297, 'samples': 2982912, 'steps': 5825, 'loss/train': 1.8890568017959595} 02/24/2022 09:05:56 - INFO - codeparrot_training - Step 5826: {'lr': 0.000492202657123862, 'samples': 2983424, 'steps': 5826, 'loss/train': 3.002413511276245} 02/24/2022 09:05:59 - INFO - codeparrot_training - Step 5827: {'lr': 0.0004921986019499078, 'samples': 2983936, 'steps': 5827, 'loss/train': 3.159044027328491} 02/24/2022 09:06:05 - INFO - codeparrot_training - Step 5828: {'lr': 0.0004921945457384516, 'samples': 2984448, 'steps': 5828, 'loss/train': 2.5415422916412354} 02/24/2022 09:06:08 - INFO - codeparrot_training - Step 5829: {'lr': 0.0004921904884895108, 'samples': 2984960, 'steps': 5829, 'loss/train': 1.8086621761322021} 02/24/2022 09:06:14 - INFO - codeparrot_training - Step 5830: {'lr': 0.000492186430203103, 'samples': 2985472, 'steps': 5830, 'loss/train': 2.4877140522003174} 02/24/2022 09:06:17 - INFO - codeparrot_training - Step 5831: {'lr': 0.0004921823708792453, 'samples': 2985984, 'steps': 5831, 'loss/train': 0.10805436223745346} 02/24/2022 09:06:23 - INFO - codeparrot_training - Step 5832: {'lr': 0.0004921783105179552, 'samples': 2986496, 'steps': 5832, 'loss/train': 2.1557843685150146} 02/24/2022 09:06:26 - INFO - codeparrot_training - Step 5833: {'lr': 0.0004921742491192502, 'samples': 2987008, 'steps': 5833, 'loss/train': 1.9289510250091553} 02/24/2022 09:06:32 - INFO - codeparrot_training - Step 5834: {'lr': 0.0004921701866831477, 'samples': 2987520, 'steps': 5834, 'loss/train': 2.871187925338745} 02/24/2022 09:06:35 - INFO - codeparrot_training - Step 5835: {'lr': 0.000492166123209665, 'samples': 2988032, 'steps': 5835, 'loss/train': 3.156506299972534} 02/24/2022 09:06:41 - INFO - codeparrot_training - Step 5836: {'lr': 0.0004921620586988193, 'samples': 2988544, 'steps': 5836, 'loss/train': 2.2434442043304443} 02/24/2022 09:06:45 - INFO - codeparrot_training - Step 5837: {'lr': 0.0004921579931506285, 'samples': 2989056, 'steps': 5837, 'loss/train': 2.22975754737854} 02/24/2022 09:06:50 - INFO - codeparrot_training - Step 5838: {'lr': 0.0004921539265651096, 'samples': 2989568, 'steps': 5838, 'loss/train': 3.5447118282318115} 02/24/2022 09:06:54 - INFO - codeparrot_training - Step 5839: {'lr': 0.0004921498589422803, 'samples': 2990080, 'steps': 5839, 'loss/train': 2.172454595565796} 02/24/2022 09:06:59 - INFO - codeparrot_training - Step 5840: {'lr': 0.0004921457902821578, 'samples': 2990592, 'steps': 5840, 'loss/train': 2.519798994064331} 02/24/2022 09:07:03 - INFO - codeparrot_training - Step 5841: {'lr': 0.0004921417205847597, 'samples': 2991104, 'steps': 5841, 'loss/train': 3.081843852996826} 02/24/2022 09:07:08 - INFO - codeparrot_training - Step 5842: {'lr': 0.0004921376498501032, 'samples': 2991616, 'steps': 5842, 'loss/train': 2.781508684158325} 02/24/2022 09:07:12 - INFO - codeparrot_training - Step 5843: {'lr': 0.000492133578078206, 'samples': 2992128, 'steps': 5843, 'loss/train': 2.0008127689361572} 02/24/2022 09:07:17 - INFO - codeparrot_training - Step 5844: {'lr': 0.0004921295052690855, 'samples': 2992640, 'steps': 5844, 'loss/train': 2.194403886795044} 02/24/2022 09:07:22 - INFO - codeparrot_training - Step 5845: {'lr': 0.000492125431422759, 'samples': 2993152, 'steps': 5845, 'loss/train': 2.6696548461914062} 02/24/2022 09:07:26 - INFO - codeparrot_training - Step 5846: {'lr': 0.0004921213565392441, 'samples': 2993664, 'steps': 5846, 'loss/train': 1.3034484386444092} 02/24/2022 09:07:31 - INFO - codeparrot_training - Step 5847: {'lr': 0.000492117280618558, 'samples': 2994176, 'steps': 5847, 'loss/train': 2.9973793029785156} 02/24/2022 09:07:35 - INFO - codeparrot_training - Step 5848: {'lr': 0.0004921132036607186, 'samples': 2994688, 'steps': 5848, 'loss/train': 1.3244411945343018} 02/24/2022 09:07:41 - INFO - codeparrot_training - Step 5849: {'lr': 0.0004921091256657429, 'samples': 2995200, 'steps': 5849, 'loss/train': 3.166919469833374} 02/24/2022 09:07:44 - INFO - codeparrot_training - Step 5850: {'lr': 0.0004921050466336487, 'samples': 2995712, 'steps': 5850, 'loss/train': 2.2517757415771484} 02/24/2022 09:07:50 - INFO - codeparrot_training - Step 5851: {'lr': 0.0004921009665644535, 'samples': 2996224, 'steps': 5851, 'loss/train': 1.4505754709243774} 02/24/2022 09:07:54 - INFO - codeparrot_training - Step 5852: {'lr': 0.0004920968854581745, 'samples': 2996736, 'steps': 5852, 'loss/train': 2.563232898712158} 02/24/2022 09:07:59 - INFO - codeparrot_training - Step 5853: {'lr': 0.0004920928033148292, 'samples': 2997248, 'steps': 5853, 'loss/train': 0.4264124035835266} 02/24/2022 09:08:03 - INFO - codeparrot_training - Step 5854: {'lr': 0.0004920887201344353, 'samples': 2997760, 'steps': 5854, 'loss/train': 2.3094394207000732} 02/24/2022 09:08:08 - INFO - codeparrot_training - Step 5855: {'lr': 0.0004920846359170103, 'samples': 2998272, 'steps': 5855, 'loss/train': 2.546543598175049} 02/24/2022 09:08:12 - INFO - codeparrot_training - Step 5856: {'lr': 0.0004920805506625714, 'samples': 2998784, 'steps': 5856, 'loss/train': 1.9825351238250732} 02/24/2022 09:08:17 - INFO - codeparrot_training - Step 5857: {'lr': 0.0004920764643711364, 'samples': 2999296, 'steps': 5857, 'loss/train': 2.429476499557495} 02/24/2022 09:08:21 - INFO - codeparrot_training - Step 5858: {'lr': 0.0004920723770427226, 'samples': 2999808, 'steps': 5858, 'loss/train': 1.733453392982483} 02/24/2022 09:08:26 - INFO - codeparrot_training - Step 5859: {'lr': 0.0004920682886773478, 'samples': 3000320, 'steps': 5859, 'loss/train': 1.9622443914413452} 02/24/2022 09:08:30 - INFO - codeparrot_training - Step 5860: {'lr': 0.000492064199275029, 'samples': 3000832, 'steps': 5860, 'loss/train': 2.8610122203826904} 02/24/2022 09:08:36 - INFO - codeparrot_training - Step 5861: {'lr': 0.0004920601088357844, 'samples': 3001344, 'steps': 5861, 'loss/train': 2.8463919162750244} 02/24/2022 09:08:40 - INFO - codeparrot_training - Step 5862: {'lr': 0.0004920560173596309, 'samples': 3001856, 'steps': 5862, 'loss/train': 2.971566915512085} 02/24/2022 09:08:45 - INFO - codeparrot_training - Step 5863: {'lr': 0.0004920519248465864, 'samples': 3002368, 'steps': 5863, 'loss/train': 2.667762041091919} 02/24/2022 09:08:49 - INFO - codeparrot_training - Step 5864: {'lr': 0.0004920478312966683, 'samples': 3002880, 'steps': 5864, 'loss/train': 2.155449867248535} 02/24/2022 09:08:54 - INFO - codeparrot_training - Step 5865: {'lr': 0.0004920437367098941, 'samples': 3003392, 'steps': 5865, 'loss/train': 2.5240321159362793} 02/24/2022 09:08:58 - INFO - codeparrot_training - Step 5866: {'lr': 0.0004920396410862815, 'samples': 3003904, 'steps': 5866, 'loss/train': 3.0880231857299805} 02/24/2022 09:09:03 - INFO - codeparrot_training - Step 5867: {'lr': 0.0004920355444258479, 'samples': 3004416, 'steps': 5867, 'loss/train': 3.0169613361358643} 02/24/2022 09:09:07 - INFO - codeparrot_training - Step 5868: {'lr': 0.0004920314467286108, 'samples': 3004928, 'steps': 5868, 'loss/train': 1.5566153526306152} 02/24/2022 09:09:12 - INFO - codeparrot_training - Step 5869: {'lr': 0.0004920273479945878, 'samples': 3005440, 'steps': 5869, 'loss/train': 2.141596555709839} 02/24/2022 09:09:16 - INFO - codeparrot_training - Step 5870: {'lr': 0.0004920232482237966, 'samples': 3005952, 'steps': 5870, 'loss/train': 1.9738441705703735} 02/24/2022 09:09:23 - INFO - codeparrot_training - Step 5871: {'lr': 0.0004920191474162547, 'samples': 3006464, 'steps': 5871, 'loss/train': 2.539581298828125} 02/24/2022 09:09:26 - INFO - codeparrot_training - Step 5872: {'lr': 0.0004920150455719795, 'samples': 3006976, 'steps': 5872, 'loss/train': 3.6093921661376953} 02/24/2022 09:09:32 - INFO - codeparrot_training - Step 5873: {'lr': 0.0004920109426909887, 'samples': 3007488, 'steps': 5873, 'loss/train': 3.1514780521392822} 02/24/2022 09:09:35 - INFO - codeparrot_training - Step 5874: {'lr': 0.0004920068387733, 'samples': 3008000, 'steps': 5874, 'loss/train': 2.913638114929199} 02/24/2022 09:09:41 - INFO - codeparrot_training - Step 5875: {'lr': 0.0004920027338189307, 'samples': 3008512, 'steps': 5875, 'loss/train': 2.3762590885162354} 02/24/2022 09:09:44 - INFO - codeparrot_training - Step 5876: {'lr': 0.0004919986278278986, 'samples': 3009024, 'steps': 5876, 'loss/train': 2.9139466285705566} 02/24/2022 09:09:50 - INFO - codeparrot_training - Step 5877: {'lr': 0.0004919945208002212, 'samples': 3009536, 'steps': 5877, 'loss/train': 2.5311338901519775} 02/24/2022 09:09:53 - INFO - codeparrot_training - Step 5878: {'lr': 0.0004919904127359162, 'samples': 3010048, 'steps': 5878, 'loss/train': 2.225706100463867} 02/24/2022 09:09:59 - INFO - codeparrot_training - Step 5879: {'lr': 0.000491986303635001, 'samples': 3010560, 'steps': 5879, 'loss/train': 0.3612838387489319} 02/24/2022 09:10:02 - INFO - codeparrot_training - Step 5880: {'lr': 0.0004919821934974933, 'samples': 3011072, 'steps': 5880, 'loss/train': 2.4800500869750977} 02/24/2022 09:10:08 - INFO - codeparrot_training - Step 5881: {'lr': 0.0004919780823234108, 'samples': 3011584, 'steps': 5881, 'loss/train': 2.014688014984131} 02/24/2022 09:10:12 - INFO - codeparrot_training - Step 5882: {'lr': 0.000491973970112771, 'samples': 3012096, 'steps': 5882, 'loss/train': 2.718477725982666} 02/24/2022 09:10:17 - INFO - codeparrot_training - Step 5883: {'lr': 0.0004919698568655916, 'samples': 3012608, 'steps': 5883, 'loss/train': 2.08742356300354} 02/24/2022 09:10:21 - INFO - codeparrot_training - Step 5884: {'lr': 0.0004919657425818901, 'samples': 3013120, 'steps': 5884, 'loss/train': 3.067800760269165} 02/24/2022 09:10:26 - INFO - codeparrot_training - Step 5885: {'lr': 0.0004919616272616842, 'samples': 3013632, 'steps': 5885, 'loss/train': 2.602139949798584} 02/24/2022 09:10:30 - INFO - codeparrot_training - Step 5886: {'lr': 0.0004919575109049915, 'samples': 3014144, 'steps': 5886, 'loss/train': 2.448864459991455} 02/24/2022 09:10:35 - INFO - codeparrot_training - Step 5887: {'lr': 0.0004919533935118296, 'samples': 3014656, 'steps': 5887, 'loss/train': 2.2215683460235596} 02/24/2022 09:10:39 - INFO - codeparrot_training - Step 5888: {'lr': 0.0004919492750822163, 'samples': 3015168, 'steps': 5888, 'loss/train': 2.8769874572753906} 02/24/2022 09:10:44 - INFO - codeparrot_training - Step 5889: {'lr': 0.0004919451556161692, 'samples': 3015680, 'steps': 5889, 'loss/train': 2.2274057865142822} 02/24/2022 09:10:48 - INFO - codeparrot_training - Step 5890: {'lr': 0.0004919410351137058, 'samples': 3016192, 'steps': 5890, 'loss/train': 1.046937346458435} 02/24/2022 09:10:53 - INFO - codeparrot_training - Step 5891: {'lr': 0.0004919369135748438, 'samples': 3016704, 'steps': 5891, 'loss/train': 1.8074312210083008} 02/24/2022 09:10:57 - INFO - codeparrot_training - Step 5892: {'lr': 0.0004919327909996008, 'samples': 3017216, 'steps': 5892, 'loss/train': 2.5137054920196533} 02/24/2022 09:11:02 - INFO - codeparrot_training - Step 5893: {'lr': 0.0004919286673879948, 'samples': 3017728, 'steps': 5893, 'loss/train': 2.2542898654937744} 02/24/2022 09:11:06 - INFO - codeparrot_training - Step 5894: {'lr': 0.000491924542740043, 'samples': 3018240, 'steps': 5894, 'loss/train': 2.457650899887085} 02/24/2022 09:11:11 - INFO - codeparrot_training - Step 5895: {'lr': 0.0004919204170557634, 'samples': 3018752, 'steps': 5895, 'loss/train': 2.6880722045898438} 02/24/2022 09:11:14 - INFO - codeparrot_training - Step 5896: {'lr': 0.0004919162903351734, 'samples': 3019264, 'steps': 5896, 'loss/train': 2.5171055793762207} 02/24/2022 09:11:21 - INFO - codeparrot_training - Step 5897: {'lr': 0.000491912162578291, 'samples': 3019776, 'steps': 5897, 'loss/train': 2.186033010482788} 02/24/2022 09:11:24 - INFO - codeparrot_training - Step 5898: {'lr': 0.0004919080337851336, 'samples': 3020288, 'steps': 5898, 'loss/train': 1.3456863164901733} 02/24/2022 09:11:30 - INFO - codeparrot_training - Step 5899: {'lr': 0.000491903903955719, 'samples': 3020800, 'steps': 5899, 'loss/train': 1.9215354919433594} 02/24/2022 09:11:33 - INFO - codeparrot_training - Step 5900: {'lr': 0.0004918997730900649, 'samples': 3021312, 'steps': 5900, 'loss/train': 1.5159803628921509} 02/24/2022 09:11:39 - INFO - codeparrot_training - Step 5901: {'lr': 0.000491895641188189, 'samples': 3021824, 'steps': 5901, 'loss/train': 2.7701447010040283} 02/24/2022 09:11:42 - INFO - codeparrot_training - Step 5902: {'lr': 0.000491891508250109, 'samples': 3022336, 'steps': 5902, 'loss/train': 3.80869722366333} 02/24/2022 09:11:48 - INFO - codeparrot_training - Step 5903: {'lr': 0.0004918873742758426, 'samples': 3022848, 'steps': 5903, 'loss/train': 2.234260320663452} 02/24/2022 09:11:51 - INFO - codeparrot_training - Step 5904: {'lr': 0.0004918832392654074, 'samples': 3023360, 'steps': 5904, 'loss/train': 2.6218695640563965} 02/24/2022 09:11:57 - INFO - codeparrot_training - Step 5905: {'lr': 0.0004918791032188214, 'samples': 3023872, 'steps': 5905, 'loss/train': 3.7347638607025146} 02/24/2022 09:12:00 - INFO - codeparrot_training - Step 5906: {'lr': 0.0004918749661361019, 'samples': 3024384, 'steps': 5906, 'loss/train': 2.1661012172698975} 02/24/2022 09:12:06 - INFO - codeparrot_training - Step 5907: {'lr': 0.000491870828017267, 'samples': 3024896, 'steps': 5907, 'loss/train': 2.8488426208496094} 02/24/2022 09:12:10 - INFO - codeparrot_training - Step 5908: {'lr': 0.0004918666888623342, 'samples': 3025408, 'steps': 5908, 'loss/train': 2.373432159423828} 02/24/2022 09:12:15 - INFO - codeparrot_training - Step 5909: {'lr': 0.0004918625486713214, 'samples': 3025920, 'steps': 5909, 'loss/train': 1.8840291500091553} 02/24/2022 09:12:19 - INFO - codeparrot_training - Step 5910: {'lr': 0.0004918584074442462, 'samples': 3026432, 'steps': 5910, 'loss/train': 1.9829857349395752} 02/24/2022 09:12:24 - INFO - codeparrot_training - Step 5911: {'lr': 0.0004918542651811263, 'samples': 3026944, 'steps': 5911, 'loss/train': 2.4264886379241943} 02/24/2022 09:12:28 - INFO - codeparrot_training - Step 5912: {'lr': 0.0004918501218819796, 'samples': 3027456, 'steps': 5912, 'loss/train': 3.1741394996643066} 02/24/2022 09:12:33 - INFO - codeparrot_training - Step 5913: {'lr': 0.0004918459775468238, 'samples': 3027968, 'steps': 5913, 'loss/train': 2.344528913497925} 02/24/2022 09:12:39 - INFO - codeparrot_training - Step 5914: {'lr': 0.0004918418321756766, 'samples': 3028480, 'steps': 5914, 'loss/train': 2.5269615650177} 02/24/2022 09:12:42 - INFO - codeparrot_training - Step 5915: {'lr': 0.0004918376857685557, 'samples': 3028992, 'steps': 5915, 'loss/train': 1.707935094833374} 02/24/2022 09:12:48 - INFO - codeparrot_training - Step 5916: {'lr': 0.000491833538325479, 'samples': 3029504, 'steps': 5916, 'loss/train': 1.5108113288879395} 02/24/2022 09:12:52 - INFO - codeparrot_training - Step 5917: {'lr': 0.0004918293898464643, 'samples': 3030016, 'steps': 5917, 'loss/train': 1.932183861732483} 02/24/2022 09:12:57 - INFO - codeparrot_training - Step 5918: {'lr': 0.0004918252403315292, 'samples': 3030528, 'steps': 5918, 'loss/train': 2.0883026123046875} 02/24/2022 09:13:01 - INFO - codeparrot_training - Step 5919: {'lr': 0.0004918210897806916, 'samples': 3031040, 'steps': 5919, 'loss/train': 2.218614101409912} 02/24/2022 09:13:07 - INFO - codeparrot_training - Step 5920: {'lr': 0.0004918169381939692, 'samples': 3031552, 'steps': 5920, 'loss/train': 2.778390645980835} 02/24/2022 09:13:10 - INFO - codeparrot_training - Step 5921: {'lr': 0.0004918127855713799, 'samples': 3032064, 'steps': 5921, 'loss/train': 2.4599533081054688} 02/24/2022 09:13:16 - INFO - codeparrot_training - Step 5922: {'lr': 0.0004918086319129413, 'samples': 3032576, 'steps': 5922, 'loss/train': 2.545726776123047} 02/24/2022 09:13:19 - INFO - codeparrot_training - Step 5923: {'lr': 0.0004918044772186714, 'samples': 3033088, 'steps': 5923, 'loss/train': 2.906093120574951} 02/24/2022 09:13:22 - INFO - codeparrot_training - Step 5924: {'lr': 0.0004918003214885877, 'samples': 3033600, 'steps': 5924, 'loss/train': 3.477184295654297} 02/24/2022 09:13:28 - INFO - codeparrot_training - Step 5925: {'lr': 0.0004917961647227084, 'samples': 3034112, 'steps': 5925, 'loss/train': 2.5634021759033203} 02/24/2022 09:13:36 - INFO - codeparrot_training - Step 5926: {'lr': 0.0004917920069210511, 'samples': 3034624, 'steps': 5926, 'loss/train': 2.973698616027832} 02/24/2022 09:13:39 - INFO - codeparrot_training - Step 5927: {'lr': 0.0004917878480836336, 'samples': 3035136, 'steps': 5927, 'loss/train': 2.3524036407470703} 02/24/2022 09:13:45 - INFO - codeparrot_training - Step 5928: {'lr': 0.0004917836882104737, 'samples': 3035648, 'steps': 5928, 'loss/train': 2.9325084686279297} 02/24/2022 09:13:48 - INFO - codeparrot_training - Step 5929: {'lr': 0.0004917795273015892, 'samples': 3036160, 'steps': 5929, 'loss/train': 2.26198148727417} 02/24/2022 09:13:54 - INFO - codeparrot_training - Step 5930: {'lr': 0.0004917753653569981, 'samples': 3036672, 'steps': 5930, 'loss/train': 3.487244129180908} 02/24/2022 09:13:57 - INFO - codeparrot_training - Step 5931: {'lr': 0.000491771202376718, 'samples': 3037184, 'steps': 5931, 'loss/train': 3.7557060718536377} 02/24/2022 09:14:03 - INFO - codeparrot_training - Step 5932: {'lr': 0.000491767038360767, 'samples': 3037696, 'steps': 5932, 'loss/train': 2.0902822017669678} 02/24/2022 09:14:06 - INFO - codeparrot_training - Step 5933: {'lr': 0.0004917628733091626, 'samples': 3038208, 'steps': 5933, 'loss/train': 3.6177351474761963} 02/24/2022 09:14:12 - INFO - codeparrot_training - Step 5934: {'lr': 0.000491758707221923, 'samples': 3038720, 'steps': 5934, 'loss/train': 3.8787155151367188} 02/24/2022 09:14:15 - INFO - codeparrot_training - Step 5935: {'lr': 0.0004917545400990657, 'samples': 3039232, 'steps': 5935, 'loss/train': 1.9852008819580078} 02/24/2022 09:14:23 - INFO - codeparrot_training - Step 5936: {'lr': 0.0004917503719406087, 'samples': 3039744, 'steps': 5936, 'loss/train': 2.1472153663635254} 02/24/2022 09:14:26 - INFO - codeparrot_training - Step 5937: {'lr': 0.00049174620274657, 'samples': 3040256, 'steps': 5937, 'loss/train': 2.3118090629577637} 02/24/2022 09:14:32 - INFO - codeparrot_training - Step 5938: {'lr': 0.0004917420325169673, 'samples': 3040768, 'steps': 5938, 'loss/train': 2.198652505874634} 02/24/2022 09:14:35 - INFO - codeparrot_training - Step 5939: {'lr': 0.0004917378612518185, 'samples': 3041280, 'steps': 5939, 'loss/train': 1.9445152282714844} 02/24/2022 09:14:41 - INFO - codeparrot_training - Step 5940: {'lr': 0.0004917336889511414, 'samples': 3041792, 'steps': 5940, 'loss/train': 2.235792875289917} 02/24/2022 09:14:44 - INFO - codeparrot_training - Step 5941: {'lr': 0.0004917295156149539, 'samples': 3042304, 'steps': 5941, 'loss/train': 1.3855277299880981} 02/24/2022 09:14:50 - INFO - codeparrot_training - Step 5942: {'lr': 0.000491725341243274, 'samples': 3042816, 'steps': 5942, 'loss/train': 2.03230619430542} 02/24/2022 09:14:53 - INFO - codeparrot_training - Step 5943: {'lr': 0.0004917211658361196, 'samples': 3043328, 'steps': 5943, 'loss/train': 7.5376973152160645} 02/24/2022 09:14:59 - INFO - codeparrot_training - Step 5944: {'lr': 0.0004917169893935083, 'samples': 3043840, 'steps': 5944, 'loss/train': 1.9802172183990479} 02/24/2022 09:15:02 - INFO - codeparrot_training - Step 5945: {'lr': 0.0004917128119154582, 'samples': 3044352, 'steps': 5945, 'loss/train': 2.066563606262207} 02/24/2022 09:15:10 - INFO - codeparrot_training - Step 5946: {'lr': 0.0004917086334019872, 'samples': 3044864, 'steps': 5946, 'loss/train': 2.6638426780700684} 02/24/2022 09:15:13 - INFO - codeparrot_training - Step 5947: {'lr': 0.0004917044538531131, 'samples': 3045376, 'steps': 5947, 'loss/train': 2.811230421066284} 02/24/2022 09:15:16 - INFO - codeparrot_training - Step 5948: {'lr': 0.000491700273268854, 'samples': 3045888, 'steps': 5948, 'loss/train': 2.620330572128296} 02/24/2022 09:15:22 - INFO - codeparrot_training - Step 5949: {'lr': 0.0004916960916492276, 'samples': 3046400, 'steps': 5949, 'loss/train': 3.8698835372924805} 02/24/2022 09:15:25 - INFO - codeparrot_training - Step 5950: {'lr': 0.0004916919089942519, 'samples': 3046912, 'steps': 5950, 'loss/train': 2.5785202980041504} 02/24/2022 09:15:31 - INFO - codeparrot_training - Step 5951: {'lr': 0.0004916877253039448, 'samples': 3047424, 'steps': 5951, 'loss/train': 3.1925880908966064} 02/24/2022 09:15:37 - INFO - codeparrot_training - Step 5952: {'lr': 0.0004916835405783242, 'samples': 3047936, 'steps': 5952, 'loss/train': 1.9864548444747925} 02/24/2022 09:15:40 - INFO - codeparrot_training - Step 5953: {'lr': 0.0004916793548174081, 'samples': 3048448, 'steps': 5953, 'loss/train': 2.669029951095581} 02/24/2022 09:15:46 - INFO - codeparrot_training - Step 5954: {'lr': 0.0004916751680212145, 'samples': 3048960, 'steps': 5954, 'loss/train': 2.4319956302642822} 02/24/2022 09:15:49 - INFO - codeparrot_training - Step 5955: {'lr': 0.000491670980189761, 'samples': 3049472, 'steps': 5955, 'loss/train': 3.0009117126464844} 02/24/2022 09:15:55 - INFO - codeparrot_training - Step 5956: {'lr': 0.0004916667913230659, 'samples': 3049984, 'steps': 5956, 'loss/train': 2.445373296737671} 02/24/2022 09:15:58 - INFO - codeparrot_training - Step 5957: {'lr': 0.000491662601421147, 'samples': 3050496, 'steps': 5957, 'loss/train': 2.3539483547210693} 02/24/2022 09:16:05 - INFO - codeparrot_training - Step 5958: {'lr': 0.0004916584104840222, 'samples': 3051008, 'steps': 5958, 'loss/train': 3.514528512954712} 02/24/2022 09:16:09 - INFO - codeparrot_training - Step 5959: {'lr': 0.0004916542185117095, 'samples': 3051520, 'steps': 5959, 'loss/train': 2.8166308403015137} 02/24/2022 09:16:14 - INFO - codeparrot_training - Step 5960: {'lr': 0.0004916500255042268, 'samples': 3052032, 'steps': 5960, 'loss/train': 2.885915994644165} 02/24/2022 09:16:18 - INFO - codeparrot_training - Step 5961: {'lr': 0.0004916458314615923, 'samples': 3052544, 'steps': 5961, 'loss/train': 1.8288182020187378} 02/24/2022 09:16:23 - INFO - codeparrot_training - Step 5962: {'lr': 0.0004916416363838237, 'samples': 3053056, 'steps': 5962, 'loss/train': 3.064462661743164} 02/24/2022 09:16:27 - INFO - codeparrot_training - Step 5963: {'lr': 0.000491637440270939, 'samples': 3053568, 'steps': 5963, 'loss/train': 1.8593496084213257} 02/24/2022 09:16:32 - INFO - codeparrot_training - Step 5964: {'lr': 0.0004916332431229562, 'samples': 3054080, 'steps': 5964, 'loss/train': 2.246675968170166} 02/24/2022 09:16:36 - INFO - codeparrot_training - Step 5965: {'lr': 0.0004916290449398934, 'samples': 3054592, 'steps': 5965, 'loss/train': 2.6514439582824707} 02/24/2022 09:16:41 - INFO - codeparrot_training - Step 5966: {'lr': 0.0004916248457217686, 'samples': 3055104, 'steps': 5966, 'loss/train': 0.20919479429721832} 02/24/2022 09:16:45 - INFO - codeparrot_training - Step 5967: {'lr': 0.0004916206454685995, 'samples': 3055616, 'steps': 5967, 'loss/train': 2.311065196990967} 02/24/2022 09:16:50 - INFO - codeparrot_training - Step 5968: {'lr': 0.0004916164441804044, 'samples': 3056128, 'steps': 5968, 'loss/train': 1.8287731409072876} 02/24/2022 09:16:54 - INFO - codeparrot_training - Step 5969: {'lr': 0.0004916122418572011, 'samples': 3056640, 'steps': 5969, 'loss/train': 3.03713321685791} 02/24/2022 09:16:59 - INFO - codeparrot_training - Step 5970: {'lr': 0.0004916080384990077, 'samples': 3057152, 'steps': 5970, 'loss/train': 2.46460223197937} 02/24/2022 09:17:03 - INFO - codeparrot_training - Step 5971: {'lr': 0.0004916038341058423, 'samples': 3057664, 'steps': 5971, 'loss/train': 3.3667428493499756} 02/24/2022 09:17:09 - INFO - codeparrot_training - Step 5972: {'lr': 0.0004915996286777226, 'samples': 3058176, 'steps': 5972, 'loss/train': 2.676967144012451} 02/24/2022 09:17:12 - INFO - codeparrot_training - Step 5973: {'lr': 0.0004915954222146669, 'samples': 3058688, 'steps': 5973, 'loss/train': 2.5330419540405273} 02/24/2022 09:17:20 - INFO - codeparrot_training - Step 5974: {'lr': 0.0004915912147166932, 'samples': 3059200, 'steps': 5974, 'loss/train': 2.5565476417541504} 02/24/2022 09:17:23 - INFO - codeparrot_training - Step 5975: {'lr': 0.0004915870061838193, 'samples': 3059712, 'steps': 5975, 'loss/train': 2.807097911834717} 02/24/2022 09:17:29 - INFO - codeparrot_training - Step 5976: {'lr': 0.0004915827966160634, 'samples': 3060224, 'steps': 5976, 'loss/train': 2.2697389125823975} 02/24/2022 09:17:32 - INFO - codeparrot_training - Step 5977: {'lr': 0.0004915785860134436, 'samples': 3060736, 'steps': 5977, 'loss/train': 2.375128984451294} 02/24/2022 09:17:38 - INFO - codeparrot_training - Step 5978: {'lr': 0.0004915743743759779, 'samples': 3061248, 'steps': 5978, 'loss/train': 2.197355270385742} 02/24/2022 09:17:41 - INFO - codeparrot_training - Step 5979: {'lr': 0.0004915701617036842, 'samples': 3061760, 'steps': 5979, 'loss/train': 2.348787307739258} 02/24/2022 09:17:47 - INFO - codeparrot_training - Step 5980: {'lr': 0.0004915659479965806, 'samples': 3062272, 'steps': 5980, 'loss/train': 2.3333446979522705} 02/24/2022 09:17:50 - INFO - codeparrot_training - Step 5981: {'lr': 0.0004915617332546852, 'samples': 3062784, 'steps': 5981, 'loss/train': 2.1261661052703857} 02/24/2022 09:17:56 - INFO - codeparrot_training - Step 5982: {'lr': 0.0004915575174780161, 'samples': 3063296, 'steps': 5982, 'loss/train': 1.4645841121673584} 02/24/2022 09:17:59 - INFO - codeparrot_training - Step 5983: {'lr': 0.0004915533006665912, 'samples': 3063808, 'steps': 5983, 'loss/train': 2.9423398971557617} 02/24/2022 09:18:06 - INFO - codeparrot_training - Step 5984: {'lr': 0.0004915490828204287, 'samples': 3064320, 'steps': 5984, 'loss/train': 2.4179799556732178} 02/24/2022 09:18:10 - INFO - codeparrot_training - Step 5985: {'lr': 0.0004915448639395466, 'samples': 3064832, 'steps': 5985, 'loss/train': 2.138633966445923} 02/24/2022 09:18:15 - INFO - codeparrot_training - Step 5986: {'lr': 0.0004915406440239631, 'samples': 3065344, 'steps': 5986, 'loss/train': 3.034088611602783} 02/24/2022 09:18:19 - INFO - codeparrot_training - Step 5987: {'lr': 0.0004915364230736961, 'samples': 3065856, 'steps': 5987, 'loss/train': 8.288673400878906} 02/24/2022 09:18:24 - INFO - codeparrot_training - Step 5988: {'lr': 0.0004915322010887637, 'samples': 3066368, 'steps': 5988, 'loss/train': 2.684619665145874} 02/24/2022 09:18:28 - INFO - codeparrot_training - Step 5989: {'lr': 0.0004915279780691843, 'samples': 3066880, 'steps': 5989, 'loss/train': 2.3704662322998047} 02/24/2022 09:18:33 - INFO - codeparrot_training - Step 5990: {'lr': 0.0004915237540149755, 'samples': 3067392, 'steps': 5990, 'loss/train': 1.8893766403198242} 02/24/2022 09:18:37 - INFO - codeparrot_training - Step 5991: {'lr': 0.0004915195289261557, 'samples': 3067904, 'steps': 5991, 'loss/train': 2.4566521644592285} 02/24/2022 09:18:42 - INFO - codeparrot_training - Step 5992: {'lr': 0.0004915153028027429, 'samples': 3068416, 'steps': 5992, 'loss/train': 2.708566188812256} 02/24/2022 09:18:46 - INFO - codeparrot_training - Step 5993: {'lr': 0.0004915110756447552, 'samples': 3068928, 'steps': 5993, 'loss/train': 2.671201467514038} 02/24/2022 09:18:53 - INFO - codeparrot_training - Step 5994: {'lr': 0.0004915068474522109, 'samples': 3069440, 'steps': 5994, 'loss/train': 2.5488717555999756} 02/24/2022 09:18:56 - INFO - codeparrot_training - Step 5995: {'lr': 0.0004915026182251278, 'samples': 3069952, 'steps': 5995, 'loss/train': 1.259892225265503} 02/24/2022 09:19:02 - INFO - codeparrot_training - Step 5996: {'lr': 0.0004914983879635242, 'samples': 3070464, 'steps': 5996, 'loss/train': 2.0499367713928223} 02/24/2022 09:19:05 - INFO - codeparrot_training - Step 5997: {'lr': 0.0004914941566674183, 'samples': 3070976, 'steps': 5997, 'loss/train': 2.102783679962158} 02/24/2022 09:19:11 - INFO - codeparrot_training - Step 5998: {'lr': 0.0004914899243368279, 'samples': 3071488, 'steps': 5998, 'loss/train': 1.7195990085601807} 02/24/2022 09:19:14 - INFO - codeparrot_training - Step 5999: {'lr': 0.0004914856909717715, 'samples': 3072000, 'steps': 5999, 'loss/train': 2.396481513977051} 02/24/2022 09:19:14 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 09:19:31 - WARNING - huggingface_hub.repository - Several commits (6) will be pushed upstream. 02/24/2022 09:19:31 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 09:20:04 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 8f2bfb2..8b611e7 floral-grass-11 -> floral-grass-11 02/24/2022 09:20:11 - INFO - codeparrot_training - Step 6000: {'lr': 0.0004914814565722671, 'samples': 3072512, 'steps': 6000, 'loss/train': 1.604397177696228} 02/24/2022 09:20:14 - INFO - codeparrot_training - Step 6001: {'lr': 0.0004914772211383327, 'samples': 3073024, 'steps': 6001, 'loss/train': 3.018603801727295} 02/24/2022 09:20:20 - INFO - codeparrot_training - Step 6002: {'lr': 0.0004914729846699867, 'samples': 3073536, 'steps': 6002, 'loss/train': 1.4417643547058105} 02/24/2022 09:20:23 - INFO - codeparrot_training - Step 6003: {'lr': 0.000491468747167247, 'samples': 3074048, 'steps': 6003, 'loss/train': 1.4821195602416992} 02/24/2022 09:20:30 - INFO - codeparrot_training - Step 6004: {'lr': 0.0004914645086301319, 'samples': 3074560, 'steps': 6004, 'loss/train': 3.401012897491455} 02/24/2022 09:20:34 - INFO - codeparrot_training - Step 6005: {'lr': 0.0004914602690586596, 'samples': 3075072, 'steps': 6005, 'loss/train': 1.865563154220581} 02/24/2022 09:20:39 - INFO - codeparrot_training - Step 6006: {'lr': 0.0004914560284528481, 'samples': 3075584, 'steps': 6006, 'loss/train': 1.9332705736160278} 02/24/2022 09:20:43 - INFO - codeparrot_training - Step 6007: {'lr': 0.0004914517868127156, 'samples': 3076096, 'steps': 6007, 'loss/train': 2.0969417095184326} 02/24/2022 09:20:49 - INFO - codeparrot_training - Step 6008: {'lr': 0.0004914475441382804, 'samples': 3076608, 'steps': 6008, 'loss/train': 2.5750367641448975} 02/24/2022 09:20:52 - INFO - codeparrot_training - Step 6009: {'lr': 0.0004914433004295605, 'samples': 3077120, 'steps': 6009, 'loss/train': 2.3179712295532227} 02/24/2022 09:20:58 - INFO - codeparrot_training - Step 6010: {'lr': 0.0004914390556865743, 'samples': 3077632, 'steps': 6010, 'loss/train': 2.3691201210021973} 02/24/2022 09:21:01 - INFO - codeparrot_training - Step 6011: {'lr': 0.0004914348099093398, 'samples': 3078144, 'steps': 6011, 'loss/train': 1.8876723051071167} 02/24/2022 09:21:07 - INFO - codeparrot_training - Step 6012: {'lr': 0.0004914305630978751, 'samples': 3078656, 'steps': 6012, 'loss/train': 1.9158037900924683} 02/24/2022 09:21:10 - INFO - codeparrot_training - Step 6013: {'lr': 0.0004914263152521987, 'samples': 3079168, 'steps': 6013, 'loss/train': 2.1041481494903564} 02/24/2022 09:21:16 - INFO - codeparrot_training - Step 6014: {'lr': 0.0004914220663723286, 'samples': 3079680, 'steps': 6014, 'loss/train': 2.2101552486419678} 02/24/2022 09:21:19 - INFO - codeparrot_training - Step 6015: {'lr': 0.0004914178164582829, 'samples': 3080192, 'steps': 6015, 'loss/train': 1.28226900100708} 02/24/2022 09:21:25 - INFO - codeparrot_training - Step 6016: {'lr': 0.0004914135655100801, 'samples': 3080704, 'steps': 6016, 'loss/train': 1.7215476036071777} 02/24/2022 09:21:28 - INFO - codeparrot_training - Step 6017: {'lr': 0.0004914093135277381, 'samples': 3081216, 'steps': 6017, 'loss/train': 2.223314046859741} 02/24/2022 09:21:34 - INFO - codeparrot_training - Step 6018: {'lr': 0.0004914050605112753, 'samples': 3081728, 'steps': 6018, 'loss/train': 2.245091199874878} 02/24/2022 09:21:37 - INFO - codeparrot_training - Step 6019: {'lr': 0.00049140080646071, 'samples': 3082240, 'steps': 6019, 'loss/train': 1.5238486528396606} 02/24/2022 09:21:45 - INFO - codeparrot_training - Step 6020: {'lr': 0.0004913965513760601, 'samples': 3082752, 'steps': 6020, 'loss/train': 2.957625389099121} 02/24/2022 09:21:50 - INFO - codeparrot_training - Step 6021: {'lr': 0.0004913922952573442, 'samples': 3083264, 'steps': 6021, 'loss/train': 2.4479918479919434} 02/24/2022 09:21:54 - INFO - codeparrot_training - Step 6022: {'lr': 0.0004913880381045803, 'samples': 3083776, 'steps': 6022, 'loss/train': 3.338038682937622} 02/24/2022 09:21:57 - INFO - codeparrot_training - Step 6023: {'lr': 0.0004913837799177867, 'samples': 3084288, 'steps': 6023, 'loss/train': 1.1308441162109375} 02/24/2022 09:22:03 - INFO - codeparrot_training - Step 6024: {'lr': 0.0004913795206969815, 'samples': 3084800, 'steps': 6024, 'loss/train': 2.617490530014038} 02/24/2022 09:22:08 - INFO - codeparrot_training - Step 6025: {'lr': 0.0004913752604421833, 'samples': 3085312, 'steps': 6025, 'loss/train': 2.5893051624298096} 02/24/2022 09:22:12 - INFO - codeparrot_training - Step 6026: {'lr': 0.0004913709991534099, 'samples': 3085824, 'steps': 6026, 'loss/train': 2.3469371795654297} 02/24/2022 09:22:17 - INFO - codeparrot_training - Step 6027: {'lr': 0.00049136673683068, 'samples': 3086336, 'steps': 6027, 'loss/train': 2.451702117919922} 02/24/2022 09:22:21 - INFO - codeparrot_training - Step 6028: {'lr': 0.0004913624734740115, 'samples': 3086848, 'steps': 6028, 'loss/train': 2.1170425415039062} 02/24/2022 09:22:28 - INFO - codeparrot_training - Step 6029: {'lr': 0.0004913582090834229, 'samples': 3087360, 'steps': 6029, 'loss/train': 2.244554281234741} 02/24/2022 09:22:31 - INFO - codeparrot_training - Step 6030: {'lr': 0.0004913539436589323, 'samples': 3087872, 'steps': 6030, 'loss/train': 2.9589085578918457} 02/24/2022 09:22:35 - INFO - codeparrot_training - Step 6031: {'lr': 0.0004913496772005581, 'samples': 3088384, 'steps': 6031, 'loss/train': 3.629499673843384} 02/24/2022 09:22:40 - INFO - codeparrot_training - Step 6032: {'lr': 0.0004913454097083185, 'samples': 3088896, 'steps': 6032, 'loss/train': 2.1499903202056885} 02/24/2022 09:22:46 - INFO - codeparrot_training - Step 6033: {'lr': 0.0004913411411822318, 'samples': 3089408, 'steps': 6033, 'loss/train': 1.62830650806427} 02/24/2022 09:22:49 - INFO - codeparrot_training - Step 6034: {'lr': 0.0004913368716223162, 'samples': 3089920, 'steps': 6034, 'loss/train': 2.0829288959503174} 02/24/2022 09:22:55 - INFO - codeparrot_training - Step 6035: {'lr': 0.0004913326010285902, 'samples': 3090432, 'steps': 6035, 'loss/train': 1.9751555919647217} 02/24/2022 09:22:58 - INFO - codeparrot_training - Step 6036: {'lr': 0.0004913283294010719, 'samples': 3090944, 'steps': 6036, 'loss/train': 2.6310036182403564} 02/24/2022 09:23:04 - INFO - codeparrot_training - Step 6037: {'lr': 0.0004913240567397797, 'samples': 3091456, 'steps': 6037, 'loss/train': 2.592632532119751} 02/24/2022 09:23:07 - INFO - codeparrot_training - Step 6038: {'lr': 0.0004913197830447319, 'samples': 3091968, 'steps': 6038, 'loss/train': 2.8282647132873535} 02/24/2022 09:23:15 - INFO - codeparrot_training - Step 6039: {'lr': 0.0004913155083159467, 'samples': 3092480, 'steps': 6039, 'loss/train': 1.9306542873382568} 02/24/2022 09:23:18 - INFO - codeparrot_training - Step 6040: {'lr': 0.0004913112325534426, 'samples': 3092992, 'steps': 6040, 'loss/train': 3.1631524562835693} 02/24/2022 09:23:24 - INFO - codeparrot_training - Step 6041: {'lr': 0.0004913069557572376, 'samples': 3093504, 'steps': 6041, 'loss/train': 4.818521499633789} 02/24/2022 09:23:27 - INFO - codeparrot_training - Step 6042: {'lr': 0.0004913026779273504, 'samples': 3094016, 'steps': 6042, 'loss/train': 2.5206692218780518} 02/24/2022 09:23:33 - INFO - codeparrot_training - Step 6043: {'lr': 0.0004912983990637992, 'samples': 3094528, 'steps': 6043, 'loss/train': 2.197874069213867} 02/24/2022 09:23:36 - INFO - codeparrot_training - Step 6044: {'lr': 0.0004912941191666021, 'samples': 3095040, 'steps': 6044, 'loss/train': 1.9329408407211304} 02/24/2022 09:23:42 - INFO - codeparrot_training - Step 6045: {'lr': 0.0004912898382357777, 'samples': 3095552, 'steps': 6045, 'loss/train': 2.4365339279174805} 02/24/2022 09:23:45 - INFO - codeparrot_training - Step 6046: {'lr': 0.0004912855562713443, 'samples': 3096064, 'steps': 6046, 'loss/train': 2.190722703933716} 02/24/2022 09:23:51 - INFO - codeparrot_training - Step 6047: {'lr': 0.0004912812732733201, 'samples': 3096576, 'steps': 6047, 'loss/train': 1.9168121814727783} 02/24/2022 09:23:54 - INFO - codeparrot_training - Step 6048: {'lr': 0.0004912769892417236, 'samples': 3097088, 'steps': 6048, 'loss/train': 2.0972447395324707} 02/24/2022 09:24:02 - INFO - codeparrot_training - Step 6049: {'lr': 0.000491272704176573, 'samples': 3097600, 'steps': 6049, 'loss/train': 2.4642345905303955} 02/24/2022 09:24:05 - INFO - codeparrot_training - Step 6050: {'lr': 0.0004912684180778869, 'samples': 3098112, 'steps': 6050, 'loss/train': 1.5909693241119385} 02/24/2022 09:24:10 - INFO - codeparrot_training - Step 6051: {'lr': 0.0004912641309456834, 'samples': 3098624, 'steps': 6051, 'loss/train': 2.454633951187134} 02/24/2022 09:24:14 - INFO - codeparrot_training - Step 6052: {'lr': 0.000491259842779981, 'samples': 3099136, 'steps': 6052, 'loss/train': 1.9100749492645264} 02/24/2022 09:24:19 - INFO - codeparrot_training - Step 6053: {'lr': 0.0004912555535807981, 'samples': 3099648, 'steps': 6053, 'loss/train': 2.113274097442627} 02/24/2022 09:24:23 - INFO - codeparrot_training - Step 6054: {'lr': 0.0004912512633481529, 'samples': 3100160, 'steps': 6054, 'loss/train': 2.7591514587402344} 02/24/2022 09:24:28 - INFO - codeparrot_training - Step 6055: {'lr': 0.0004912469720820639, 'samples': 3100672, 'steps': 6055, 'loss/train': 3.426459312438965} 02/24/2022 09:24:32 - INFO - codeparrot_training - Step 6056: {'lr': 0.0004912426797825495, 'samples': 3101184, 'steps': 6056, 'loss/train': 1.6065919399261475} 02/24/2022 09:24:37 - INFO - codeparrot_training - Step 6057: {'lr': 0.0004912383864496281, 'samples': 3101696, 'steps': 6057, 'loss/train': 2.546449899673462} 02/24/2022 09:24:41 - INFO - codeparrot_training - Step 6058: {'lr': 0.0004912340920833182, 'samples': 3102208, 'steps': 6058, 'loss/train': 2.28745698928833} 02/24/2022 09:24:46 - INFO - codeparrot_training - Step 6059: {'lr': 0.0004912297966836378, 'samples': 3102720, 'steps': 6059, 'loss/train': 2.5935282707214355} 02/24/2022 09:24:50 - INFO - codeparrot_training - Step 6060: {'lr': 0.0004912255002506057, 'samples': 3103232, 'steps': 6060, 'loss/train': 2.5242879390716553} 02/24/2022 09:24:55 - INFO - codeparrot_training - Step 6061: {'lr': 0.00049122120278424, 'samples': 3103744, 'steps': 6061, 'loss/train': 3.297246217727661} 02/24/2022 09:24:59 - INFO - codeparrot_training - Step 6062: {'lr': 0.0004912169042845595, 'samples': 3104256, 'steps': 6062, 'loss/train': 2.0308449268341064} 02/24/2022 09:25:04 - INFO - codeparrot_training - Step 6063: {'lr': 0.0004912126047515821, 'samples': 3104768, 'steps': 6063, 'loss/train': 1.0052578449249268} 02/24/2022 09:25:08 - INFO - codeparrot_training - Step 6064: {'lr': 0.0004912083041853267, 'samples': 3105280, 'steps': 6064, 'loss/train': 5.7387518882751465} 02/24/2022 09:25:15 - INFO - codeparrot_training - Step 6065: {'lr': 0.0004912040025858114, 'samples': 3105792, 'steps': 6065, 'loss/train': 2.9070844650268555} 02/24/2022 09:25:19 - INFO - codeparrot_training - Step 6066: {'lr': 0.0004911996999530548, 'samples': 3106304, 'steps': 6066, 'loss/train': 3.1185414791107178} 02/24/2022 09:25:24 - INFO - codeparrot_training - Step 6067: {'lr': 0.0004911953962870754, 'samples': 3106816, 'steps': 6067, 'loss/train': 3.52547550201416} 02/24/2022 09:25:28 - INFO - codeparrot_training - Step 6068: {'lr': 0.0004911910915878913, 'samples': 3107328, 'steps': 6068, 'loss/train': 2.851006507873535} 02/24/2022 09:25:33 - INFO - codeparrot_training - Step 6069: {'lr': 0.0004911867858555212, 'samples': 3107840, 'steps': 6069, 'loss/train': 3.4294042587280273} 02/24/2022 09:25:37 - INFO - codeparrot_training - Step 6070: {'lr': 0.0004911824790899836, 'samples': 3108352, 'steps': 6070, 'loss/train': 1.884102702140808} 02/24/2022 09:25:42 - INFO - codeparrot_training - Step 6071: {'lr': 0.0004911781712912968, 'samples': 3108864, 'steps': 6071, 'loss/train': 1.6893190145492554} 02/24/2022 09:25:46 - INFO - codeparrot_training - Step 6072: {'lr': 0.0004911738624594793, 'samples': 3109376, 'steps': 6072, 'loss/train': 2.1315715312957764} 02/24/2022 09:25:51 - INFO - codeparrot_training - Step 6073: {'lr': 0.0004911695525945494, 'samples': 3109888, 'steps': 6073, 'loss/train': 2.465468168258667} 02/24/2022 09:25:58 - INFO - codeparrot_training - Step 6074: {'lr': 0.0004911652416965259, 'samples': 3110400, 'steps': 6074, 'loss/train': 2.6952435970306396} 02/24/2022 09:26:02 - INFO - codeparrot_training - Step 6075: {'lr': 0.000491160929765427, 'samples': 3110912, 'steps': 6075, 'loss/train': 3.1071462631225586} 02/24/2022 09:26:05 - INFO - codeparrot_training - Step 6076: {'lr': 0.0004911566168012714, 'samples': 3111424, 'steps': 6076, 'loss/train': 2.949596881866455} 02/24/2022 09:26:11 - INFO - codeparrot_training - Step 6077: {'lr': 0.0004911523028040772, 'samples': 3111936, 'steps': 6077, 'loss/train': 2.3221099376678467} 02/24/2022 09:26:16 - INFO - codeparrot_training - Step 6078: {'lr': 0.0004911479877738633, 'samples': 3112448, 'steps': 6078, 'loss/train': 2.6894407272338867} 02/24/2022 09:26:20 - INFO - codeparrot_training - Step 6079: {'lr': 0.0004911436717106478, 'samples': 3112960, 'steps': 6079, 'loss/train': 2.227316379547119} 02/24/2022 09:26:25 - INFO - codeparrot_training - Step 6080: {'lr': 0.0004911393546144495, 'samples': 3113472, 'steps': 6080, 'loss/train': 2.730419397354126} 02/24/2022 09:26:29 - INFO - codeparrot_training - Step 6081: {'lr': 0.0004911350364852868, 'samples': 3113984, 'steps': 6081, 'loss/train': 1.5856051445007324} 02/24/2022 09:26:34 - INFO - codeparrot_training - Step 6082: {'lr': 0.0004911307173231782, 'samples': 3114496, 'steps': 6082, 'loss/train': 2.1041507720947266} 02/24/2022 09:26:38 - INFO - codeparrot_training - Step 6083: {'lr': 0.000491126397128142, 'samples': 3115008, 'steps': 6083, 'loss/train': 1.594551682472229} 02/24/2022 09:26:45 - INFO - codeparrot_training - Step 6084: {'lr': 0.0004911220759001971, 'samples': 3115520, 'steps': 6084, 'loss/train': 2.4654273986816406} 02/24/2022 09:26:49 - INFO - codeparrot_training - Step 6085: {'lr': 0.0004911177536393616, 'samples': 3116032, 'steps': 6085, 'loss/train': 1.576617956161499} 02/24/2022 09:26:52 - INFO - codeparrot_training - Step 6086: {'lr': 0.0004911134303456543, 'samples': 3116544, 'steps': 6086, 'loss/train': 2.4746272563934326} 02/24/2022 09:26:58 - INFO - codeparrot_training - Step 6087: {'lr': 0.0004911091060190937, 'samples': 3117056, 'steps': 6087, 'loss/train': 2.7446303367614746} 02/24/2022 09:27:01 - INFO - codeparrot_training - Step 6088: {'lr': 0.0004911047806596981, 'samples': 3117568, 'steps': 6088, 'loss/train': 2.535982131958008} 02/24/2022 09:27:07 - INFO - codeparrot_training - Step 6089: {'lr': 0.0004911004542674863, 'samples': 3118080, 'steps': 6089, 'loss/train': 2.5550506114959717} 02/24/2022 09:27:10 - INFO - codeparrot_training - Step 6090: {'lr': 0.0004910961268424766, 'samples': 3118592, 'steps': 6090, 'loss/train': 2.3968963623046875} 02/24/2022 09:27:16 - INFO - codeparrot_training - Step 6091: {'lr': 0.0004910917983846877, 'samples': 3119104, 'steps': 6091, 'loss/train': 2.9215710163116455} 02/24/2022 09:27:19 - INFO - codeparrot_training - Step 6092: {'lr': 0.0004910874688941381, 'samples': 3119616, 'steps': 6092, 'loss/train': 3.4735233783721924} 02/24/2022 09:27:25 - INFO - codeparrot_training - Step 6093: {'lr': 0.0004910831383708464, 'samples': 3120128, 'steps': 6093, 'loss/train': 2.906996726989746} 02/24/2022 09:27:28 - INFO - codeparrot_training - Step 6094: {'lr': 0.000491078806814831, 'samples': 3120640, 'steps': 6094, 'loss/train': 1.6564717292785645} 02/24/2022 09:27:34 - INFO - codeparrot_training - Step 6095: {'lr': 0.0004910744742261106, 'samples': 3121152, 'steps': 6095, 'loss/train': 2.2512950897216797} 02/24/2022 09:27:37 - INFO - codeparrot_training - Step 6096: {'lr': 0.0004910701406047037, 'samples': 3121664, 'steps': 6096, 'loss/train': 2.8152337074279785} 02/24/2022 09:27:44 - INFO - codeparrot_training - Step 6097: {'lr': 0.0004910658059506289, 'samples': 3122176, 'steps': 6097, 'loss/train': 3.291065216064453} 02/24/2022 09:27:48 - INFO - codeparrot_training - Step 6098: {'lr': 0.0004910614702639045, 'samples': 3122688, 'steps': 6098, 'loss/train': 1.9583368301391602} 02/24/2022 09:27:53 - INFO - codeparrot_training - Step 6099: {'lr': 0.0004910571335445496, 'samples': 3123200, 'steps': 6099, 'loss/train': 2.0853025913238525} 02/24/2022 09:27:57 - INFO - codeparrot_training - Step 6100: {'lr': 0.0004910527957925823, 'samples': 3123712, 'steps': 6100, 'loss/train': 2.9351468086242676} 02/24/2022 09:28:03 - INFO - codeparrot_training - Step 6101: {'lr': 0.0004910484570080215, 'samples': 3124224, 'steps': 6101, 'loss/train': 2.6556482315063477} 02/24/2022 09:28:06 - INFO - codeparrot_training - Step 6102: {'lr': 0.0004910441171908855, 'samples': 3124736, 'steps': 6102, 'loss/train': 1.123103141784668} 02/24/2022 09:28:12 - INFO - codeparrot_training - Step 6103: {'lr': 0.0004910397763411931, 'samples': 3125248, 'steps': 6103, 'loss/train': 2.4701147079467773} 02/24/2022 09:28:15 - INFO - codeparrot_training - Step 6104: {'lr': 0.000491035434458963, 'samples': 3125760, 'steps': 6104, 'loss/train': 2.011813163757324} 02/24/2022 09:28:21 - INFO - codeparrot_training - Step 6105: {'lr': 0.0004910310915442135, 'samples': 3126272, 'steps': 6105, 'loss/train': 2.420161008834839} 02/24/2022 09:28:24 - INFO - codeparrot_training - Step 6106: {'lr': 0.0004910267475969633, 'samples': 3126784, 'steps': 6106, 'loss/train': 1.6771225929260254} 02/24/2022 09:28:30 - INFO - codeparrot_training - Step 6107: {'lr': 0.000491022402617231, 'samples': 3127296, 'steps': 6107, 'loss/train': 1.8822206258773804} 02/24/2022 09:28:33 - INFO - codeparrot_training - Step 6108: {'lr': 0.0004910180566050354, 'samples': 3127808, 'steps': 6108, 'loss/train': 2.1303060054779053} 02/24/2022 09:28:39 - INFO - codeparrot_training - Step 6109: {'lr': 0.0004910137095603949, 'samples': 3128320, 'steps': 6109, 'loss/train': 2.4091217517852783} 02/24/2022 09:28:42 - INFO - codeparrot_training - Step 6110: {'lr': 0.0004910093614833282, 'samples': 3128832, 'steps': 6110, 'loss/train': 2.64959454536438} 02/24/2022 09:28:49 - INFO - codeparrot_training - Step 6111: {'lr': 0.000491005012373854, 'samples': 3129344, 'steps': 6111, 'loss/train': 2.0367701053619385} 02/24/2022 09:28:53 - INFO - codeparrot_training - Step 6112: {'lr': 0.0004910006622319908, 'samples': 3129856, 'steps': 6112, 'loss/train': 1.6172839403152466} 02/24/2022 09:28:59 - INFO - codeparrot_training - Step 6113: {'lr': 0.0004909963110577573, 'samples': 3130368, 'steps': 6113, 'loss/train': 2.6684956550598145} 02/24/2022 09:29:02 - INFO - codeparrot_training - Step 6114: {'lr': 0.000490991958851172, 'samples': 3130880, 'steps': 6114, 'loss/train': 1.8965198993682861} 02/24/2022 09:29:08 - INFO - codeparrot_training - Step 6115: {'lr': 0.0004909876056122538, 'samples': 3131392, 'steps': 6115, 'loss/train': 2.818169116973877} 02/24/2022 09:29:11 - INFO - codeparrot_training - Step 6116: {'lr': 0.0004909832513410213, 'samples': 3131904, 'steps': 6116, 'loss/train': 2.6958489418029785} 02/24/2022 09:29:17 - INFO - codeparrot_training - Step 6117: {'lr': 0.000490978896037493, 'samples': 3132416, 'steps': 6117, 'loss/train': 3.0597338676452637} 02/24/2022 09:29:20 - INFO - codeparrot_training - Step 6118: {'lr': 0.0004909745397016876, 'samples': 3132928, 'steps': 6118, 'loss/train': 1.8273980617523193} 02/24/2022 09:29:25 - INFO - codeparrot_training - Step 6119: {'lr': 0.0004909701823336238, 'samples': 3133440, 'steps': 6119, 'loss/train': 2.6896169185638428} 02/24/2022 09:29:29 - INFO - codeparrot_training - Step 6120: {'lr': 0.0004909658239333202, 'samples': 3133952, 'steps': 6120, 'loss/train': 2.7352962493896484} 02/24/2022 09:29:36 - INFO - codeparrot_training - Step 6121: {'lr': 0.0004909614645007956, 'samples': 3134464, 'steps': 6121, 'loss/train': 2.0802161693573} 02/24/2022 09:29:40 - INFO - codeparrot_training - Step 6122: {'lr': 0.0004909571040360686, 'samples': 3134976, 'steps': 6122, 'loss/train': 3.7975966930389404} 02/24/2022 09:29:45 - INFO - codeparrot_training - Step 6123: {'lr': 0.0004909527425391579, 'samples': 3135488, 'steps': 6123, 'loss/train': 1.557277798652649} 02/24/2022 09:29:49 - INFO - codeparrot_training - Step 6124: {'lr': 0.0004909483800100822, 'samples': 3136000, 'steps': 6124, 'loss/train': 2.73663067817688} 02/24/2022 09:29:54 - INFO - codeparrot_training - Step 6125: {'lr': 0.00049094401644886, 'samples': 3136512, 'steps': 6125, 'loss/train': 1.576921820640564} 02/24/2022 09:29:58 - INFO - codeparrot_training - Step 6126: {'lr': 0.0004909396518555102, 'samples': 3137024, 'steps': 6126, 'loss/train': 2.7907321453094482} 02/24/2022 09:30:03 - INFO - codeparrot_training - Step 6127: {'lr': 0.0004909352862300514, 'samples': 3137536, 'steps': 6127, 'loss/train': 1.7028225660324097} 02/24/2022 09:30:07 - INFO - codeparrot_training - Step 6128: {'lr': 0.0004909309195725024, 'samples': 3138048, 'steps': 6128, 'loss/train': 2.3685123920440674} 02/24/2022 09:30:12 - INFO - codeparrot_training - Step 6129: {'lr': 0.0004909265518828819, 'samples': 3138560, 'steps': 6129, 'loss/train': 1.558257818222046} 02/24/2022 09:30:16 - INFO - codeparrot_training - Step 6130: {'lr': 0.0004909221831612085, 'samples': 3139072, 'steps': 6130, 'loss/train': 2.0935728549957275} 02/24/2022 09:30:23 - INFO - codeparrot_training - Step 6131: {'lr': 0.000490917813407501, 'samples': 3139584, 'steps': 6131, 'loss/train': 2.4071693420410156} 02/24/2022 09:30:26 - INFO - codeparrot_training - Step 6132: {'lr': 0.0004909134426217779, 'samples': 3140096, 'steps': 6132, 'loss/train': 3.3161418437957764} 02/24/2022 09:30:32 - INFO - codeparrot_training - Step 6133: {'lr': 0.0004909090708040583, 'samples': 3140608, 'steps': 6133, 'loss/train': 1.6046791076660156} 02/24/2022 09:30:37 - INFO - codeparrot_training - Step 6134: {'lr': 0.0004909046979543608, 'samples': 3141120, 'steps': 6134, 'loss/train': 1.3531415462493896} 02/24/2022 09:30:41 - INFO - codeparrot_training - Step 6135: {'lr': 0.000490900324072704, 'samples': 3141632, 'steps': 6135, 'loss/train': 1.4257930517196655} 02/24/2022 09:30:46 - INFO - codeparrot_training - Step 6136: {'lr': 0.0004908959491591065, 'samples': 3142144, 'steps': 6136, 'loss/train': 3.079061985015869} 02/24/2022 09:30:50 - INFO - codeparrot_training - Step 6137: {'lr': 0.0004908915732135874, 'samples': 3142656, 'steps': 6137, 'loss/train': 1.9684590101242065} 02/24/2022 09:30:53 - INFO - codeparrot_training - Step 6138: {'lr': 0.0004908871962361654, 'samples': 3143168, 'steps': 6138, 'loss/train': 0.4464505612850189} 02/24/2022 09:31:00 - INFO - codeparrot_training - Step 6139: {'lr': 0.0004908828182268591, 'samples': 3143680, 'steps': 6139, 'loss/train': 1.4488576650619507} 02/24/2022 09:31:03 - INFO - codeparrot_training - Step 6140: {'lr': 0.0004908784391856872, 'samples': 3144192, 'steps': 6140, 'loss/train': 2.665416955947876} 02/24/2022 09:31:10 - INFO - codeparrot_training - Step 6141: {'lr': 0.0004908740591126686, 'samples': 3144704, 'steps': 6141, 'loss/train': 1.779059648513794} 02/24/2022 09:31:14 - INFO - codeparrot_training - Step 6142: {'lr': 0.000490869678007822, 'samples': 3145216, 'steps': 6142, 'loss/train': 2.6022825241088867} 02/24/2022 09:31:19 - INFO - codeparrot_training - Step 6143: {'lr': 0.0004908652958711663, 'samples': 3145728, 'steps': 6143, 'loss/train': 2.301237106323242} 02/24/2022 09:31:23 - INFO - codeparrot_training - Step 6144: {'lr': 0.00049086091270272, 'samples': 3146240, 'steps': 6144, 'loss/train': 1.6943293809890747} 02/24/2022 09:31:28 - INFO - codeparrot_training - Step 6145: {'lr': 0.0004908565285025021, 'samples': 3146752, 'steps': 6145, 'loss/train': 1.8279458284378052} 02/24/2022 09:31:32 - INFO - codeparrot_training - Step 6146: {'lr': 0.0004908521432705312, 'samples': 3147264, 'steps': 6146, 'loss/train': 3.0824453830718994} 02/24/2022 09:31:37 - INFO - codeparrot_training - Step 6147: {'lr': 0.0004908477570068263, 'samples': 3147776, 'steps': 6147, 'loss/train': 2.9719717502593994} 02/24/2022 09:31:41 - INFO - codeparrot_training - Step 6148: {'lr': 0.0004908433697114062, 'samples': 3148288, 'steps': 6148, 'loss/train': 2.696833610534668} 02/24/2022 09:31:46 - INFO - codeparrot_training - Step 6149: {'lr': 0.0004908389813842894, 'samples': 3148800, 'steps': 6149, 'loss/train': 2.525071144104004} 02/24/2022 09:31:50 - INFO - codeparrot_training - Step 6150: {'lr': 0.0004908345920254949, 'samples': 3149312, 'steps': 6150, 'loss/train': 2.4214484691619873} 02/24/2022 09:31:55 - INFO - codeparrot_training - Step 6151: {'lr': 0.0004908302016350416, 'samples': 3149824, 'steps': 6151, 'loss/train': 2.583103656768799} 02/24/2022 09:31:59 - INFO - codeparrot_training - Step 6152: {'lr': 0.0004908258102129481, 'samples': 3150336, 'steps': 6152, 'loss/train': 1.8836390972137451} 02/24/2022 09:32:04 - INFO - codeparrot_training - Step 6153: {'lr': 0.0004908214177592334, 'samples': 3150848, 'steps': 6153, 'loss/train': 3.045994997024536} 02/24/2022 09:32:08 - INFO - codeparrot_training - Step 6154: {'lr': 0.000490817024273916, 'samples': 3151360, 'steps': 6154, 'loss/train': 0.5500620007514954} 02/24/2022 09:32:13 - INFO - codeparrot_training - Step 6155: {'lr': 0.0004908126297570152, 'samples': 3151872, 'steps': 6155, 'loss/train': 2.1244826316833496} 02/24/2022 09:32:17 - INFO - codeparrot_training - Step 6156: {'lr': 0.0004908082342085494, 'samples': 3152384, 'steps': 6156, 'loss/train': 2.3692545890808105} 02/24/2022 09:32:22 - INFO - codeparrot_training - Step 6157: {'lr': 0.0004908038376285375, 'samples': 3152896, 'steps': 6157, 'loss/train': 4.187961578369141} 02/24/2022 09:32:26 - INFO - codeparrot_training - Step 6158: {'lr': 0.0004907994400169986, 'samples': 3153408, 'steps': 6158, 'loss/train': 2.536123514175415} 02/24/2022 09:32:33 - INFO - codeparrot_training - Step 6159: {'lr': 0.0004907950413739514, 'samples': 3153920, 'steps': 6159, 'loss/train': 2.9844295978546143} 02/24/2022 09:32:37 - INFO - codeparrot_training - Step 6160: {'lr': 0.0004907906416994146, 'samples': 3154432, 'steps': 6160, 'loss/train': 2.656822443008423} 02/24/2022 09:32:42 - INFO - codeparrot_training - Step 6161: {'lr': 0.0004907862409934071, 'samples': 3154944, 'steps': 6161, 'loss/train': 3.5315799713134766} 02/24/2022 09:32:46 - INFO - codeparrot_training - Step 6162: {'lr': 0.0004907818392559479, 'samples': 3155456, 'steps': 6162, 'loss/train': 2.341357707977295} 02/24/2022 09:32:51 - INFO - codeparrot_training - Step 6163: {'lr': 0.0004907774364870557, 'samples': 3155968, 'steps': 6163, 'loss/train': 2.1793057918548584} 02/24/2022 09:32:55 - INFO - codeparrot_training - Step 6164: {'lr': 0.0004907730326867495, 'samples': 3156480, 'steps': 6164, 'loss/train': 2.491849422454834} 02/24/2022 09:33:00 - INFO - codeparrot_training - Step 6165: {'lr': 0.0004907686278550479, 'samples': 3156992, 'steps': 6165, 'loss/train': 1.4215545654296875} 02/24/2022 09:33:04 - INFO - codeparrot_training - Step 6166: {'lr': 0.0004907642219919701, 'samples': 3157504, 'steps': 6166, 'loss/train': 2.2843611240386963} 02/24/2022 09:33:11 - INFO - codeparrot_training - Step 6167: {'lr': 0.0004907598150975348, 'samples': 3158016, 'steps': 6167, 'loss/train': 3.4909417629241943} 02/24/2022 09:33:14 - INFO - codeparrot_training - Step 6168: {'lr': 0.0004907554071717609, 'samples': 3158528, 'steps': 6168, 'loss/train': 1.2315435409545898} 02/24/2022 09:33:20 - INFO - codeparrot_training - Step 6169: {'lr': 0.0004907509982146673, 'samples': 3159040, 'steps': 6169, 'loss/train': 3.2568936347961426} 02/24/2022 09:33:23 - INFO - codeparrot_training - Step 6170: {'lr': 0.0004907465882262728, 'samples': 3159552, 'steps': 6170, 'loss/train': 1.9939494132995605} 02/24/2022 09:33:29 - INFO - codeparrot_training - Step 6171: {'lr': 0.0004907421772065965, 'samples': 3160064, 'steps': 6171, 'loss/train': 3.2389276027679443} 02/24/2022 09:33:32 - INFO - codeparrot_training - Step 6172: {'lr': 0.000490737765155657, 'samples': 3160576, 'steps': 6172, 'loss/train': 2.159992218017578} 02/24/2022 09:33:38 - INFO - codeparrot_training - Step 6173: {'lr': 0.0004907333520734734, 'samples': 3161088, 'steps': 6173, 'loss/train': 1.7988319396972656} 02/24/2022 09:33:41 - INFO - codeparrot_training - Step 6174: {'lr': 0.0004907289379600646, 'samples': 3161600, 'steps': 6174, 'loss/train': 1.812206506729126} 02/24/2022 09:33:47 - INFO - codeparrot_training - Step 6175: {'lr': 0.0004907245228154495, 'samples': 3162112, 'steps': 6175, 'loss/train': 1.9122754335403442} 02/24/2022 09:33:50 - INFO - codeparrot_training - Step 6176: {'lr': 0.0004907201066396469, 'samples': 3162624, 'steps': 6176, 'loss/train': 0.5126199126243591} 02/24/2022 09:33:58 - INFO - codeparrot_training - Step 6177: {'lr': 0.0004907156894326758, 'samples': 3163136, 'steps': 6177, 'loss/train': 2.3578402996063232} 02/24/2022 09:34:01 - INFO - codeparrot_training - Step 6178: {'lr': 0.0004907112711945552, 'samples': 3163648, 'steps': 6178, 'loss/train': 1.8555666208267212} 02/24/2022 09:34:07 - INFO - codeparrot_training - Step 6179: {'lr': 0.000490706851925304, 'samples': 3164160, 'steps': 6179, 'loss/train': 2.283424139022827} 02/24/2022 09:34:10 - INFO - codeparrot_training - Step 6180: {'lr': 0.0004907024316249408, 'samples': 3164672, 'steps': 6180, 'loss/train': 2.6915483474731445} 02/24/2022 09:34:15 - INFO - codeparrot_training - Step 6181: {'lr': 0.0004906980102934852, 'samples': 3165184, 'steps': 6181, 'loss/train': 2.5719361305236816} 02/24/2022 09:34:21 - INFO - codeparrot_training - Step 6182: {'lr': 0.0004906935879309555, 'samples': 3165696, 'steps': 6182, 'loss/train': 1.5978578329086304} 02/24/2022 09:34:24 - INFO - codeparrot_training - Step 6183: {'lr': 0.0004906891645373709, 'samples': 3166208, 'steps': 6183, 'loss/train': 1.8750826120376587} 02/24/2022 09:34:30 - INFO - codeparrot_training - Step 6184: {'lr': 0.0004906847401127504, 'samples': 3166720, 'steps': 6184, 'loss/train': 1.9642512798309326} 02/24/2022 09:34:33 - INFO - codeparrot_training - Step 6185: {'lr': 0.0004906803146571129, 'samples': 3167232, 'steps': 6185, 'loss/train': 2.1332035064697266} 02/24/2022 09:34:39 - INFO - codeparrot_training - Step 6186: {'lr': 0.0004906758881704774, 'samples': 3167744, 'steps': 6186, 'loss/train': 2.0426807403564453} 02/24/2022 09:34:42 - INFO - codeparrot_training - Step 6187: {'lr': 0.0004906714606528628, 'samples': 3168256, 'steps': 6187, 'loss/train': 3.2241833209991455} 02/24/2022 09:34:50 - INFO - codeparrot_training - Step 6188: {'lr': 0.0004906670321042881, 'samples': 3168768, 'steps': 6188, 'loss/train': 3.909959316253662} 02/24/2022 09:34:53 - INFO - codeparrot_training - Step 6189: {'lr': 0.0004906626025247722, 'samples': 3169280, 'steps': 6189, 'loss/train': 4.036428451538086} 02/24/2022 09:34:59 - INFO - codeparrot_training - Step 6190: {'lr': 0.000490658171914334, 'samples': 3169792, 'steps': 6190, 'loss/train': 2.8523690700531006} 02/24/2022 09:35:02 - INFO - codeparrot_training - Step 6191: {'lr': 0.0004906537402729928, 'samples': 3170304, 'steps': 6191, 'loss/train': 2.8157458305358887} 02/24/2022 09:35:08 - INFO - codeparrot_training - Step 6192: {'lr': 0.0004906493076007675, 'samples': 3170816, 'steps': 6192, 'loss/train': 2.9344265460968018} 02/24/2022 09:35:12 - INFO - codeparrot_training - Step 6193: {'lr': 0.0004906448738976768, 'samples': 3171328, 'steps': 6193, 'loss/train': 2.2063586711883545} 02/24/2022 09:35:17 - INFO - codeparrot_training - Step 6194: {'lr': 0.0004906404391637397, 'samples': 3171840, 'steps': 6194, 'loss/train': 3.1388185024261475} 02/24/2022 09:35:21 - INFO - codeparrot_training - Step 6195: {'lr': 0.0004906360033989758, 'samples': 3172352, 'steps': 6195, 'loss/train': 3.407362461090088} 02/24/2022 09:35:26 - INFO - codeparrot_training - Step 6196: {'lr': 0.0004906315666034034, 'samples': 3172864, 'steps': 6196, 'loss/train': 2.4597115516662598} 02/24/2022 09:35:30 - INFO - codeparrot_training - Step 6197: {'lr': 0.0004906271287770418, 'samples': 3173376, 'steps': 6197, 'loss/train': 2.4107017517089844} 02/24/2022 09:35:35 - INFO - codeparrot_training - Step 6198: {'lr': 0.00049062268991991, 'samples': 3173888, 'steps': 6198, 'loss/train': 2.347597599029541} 02/24/2022 09:35:39 - INFO - codeparrot_training - Step 6199: {'lr': 0.0004906182500320269, 'samples': 3174400, 'steps': 6199, 'loss/train': 2.123532772064209} 02/24/2022 09:35:44 - INFO - codeparrot_training - Step 6200: {'lr': 0.0004906138091134118, 'samples': 3174912, 'steps': 6200, 'loss/train': 3.8120334148406982} 02/24/2022 09:35:48 - INFO - codeparrot_training - Step 6201: {'lr': 0.0004906093671640836, 'samples': 3175424, 'steps': 6201, 'loss/train': 3.379196882247925} 02/24/2022 09:35:51 - INFO - codeparrot_training - Step 6202: {'lr': 0.0004906049241840612, 'samples': 3175936, 'steps': 6202, 'loss/train': 2.8560476303100586} 02/24/2022 09:35:59 - INFO - codeparrot_training - Step 6203: {'lr': 0.0004906004801733635, 'samples': 3176448, 'steps': 6203, 'loss/train': 2.3828630447387695} 02/24/2022 09:36:04 - INFO - codeparrot_training - Step 6204: {'lr': 0.0004905960351320099, 'samples': 3176960, 'steps': 6204, 'loss/train': 1.625923752784729} 02/24/2022 09:36:07 - INFO - codeparrot_training - Step 6205: {'lr': 0.0004905915890600194, 'samples': 3177472, 'steps': 6205, 'loss/train': 0.898340106010437} 02/24/2022 09:36:13 - INFO - codeparrot_training - Step 6206: {'lr': 0.0004905871419574107, 'samples': 3177984, 'steps': 6206, 'loss/train': 3.0943093299865723} 02/24/2022 09:36:16 - INFO - codeparrot_training - Step 6207: {'lr': 0.0004905826938242032, 'samples': 3178496, 'steps': 6207, 'loss/train': 2.0490376949310303} 02/24/2022 09:36:22 - INFO - codeparrot_training - Step 6208: {'lr': 0.0004905782446604158, 'samples': 3179008, 'steps': 6208, 'loss/train': 1.7287081480026245} 02/24/2022 09:36:25 - INFO - codeparrot_training - Step 6209: {'lr': 0.0004905737944660676, 'samples': 3179520, 'steps': 6209, 'loss/train': 2.7915098667144775} 02/24/2022 09:36:31 - INFO - codeparrot_training - Step 6210: {'lr': 0.0004905693432411777, 'samples': 3180032, 'steps': 6210, 'loss/train': 3.150730609893799} 02/24/2022 09:36:35 - INFO - codeparrot_training - Step 6211: {'lr': 0.0004905648909857652, 'samples': 3180544, 'steps': 6211, 'loss/train': 3.344449520111084} 02/24/2022 09:36:40 - INFO - codeparrot_training - Step 6212: {'lr': 0.0004905604376998489, 'samples': 3181056, 'steps': 6212, 'loss/train': 2.2071542739868164} 02/24/2022 09:36:44 - INFO - codeparrot_training - Step 6213: {'lr': 0.0004905559833834482, 'samples': 3181568, 'steps': 6213, 'loss/train': 1.9531646966934204} 02/24/2022 09:36:51 - INFO - codeparrot_training - Step 6214: {'lr': 0.000490551528036582, 'samples': 3182080, 'steps': 6214, 'loss/train': 2.1269495487213135} 02/24/2022 09:36:54 - INFO - codeparrot_training - Step 6215: {'lr': 0.0004905470716592695, 'samples': 3182592, 'steps': 6215, 'loss/train': 2.228341579437256} 02/24/2022 09:37:00 - INFO - codeparrot_training - Step 6216: {'lr': 0.0004905426142515296, 'samples': 3183104, 'steps': 6216, 'loss/train': 0.1706215888261795} 02/24/2022 09:37:03 - INFO - codeparrot_training - Step 6217: {'lr': 0.0004905381558133817, 'samples': 3183616, 'steps': 6217, 'loss/train': 2.2180795669555664} 02/24/2022 09:37:09 - INFO - codeparrot_training - Step 6218: {'lr': 0.0004905336963448446, 'samples': 3184128, 'steps': 6218, 'loss/train': 2.185758352279663} 02/24/2022 09:37:12 - INFO - codeparrot_training - Step 6219: {'lr': 0.0004905292358459375, 'samples': 3184640, 'steps': 6219, 'loss/train': 3.3639838695526123} 02/24/2022 09:37:18 - INFO - codeparrot_training - Step 6220: {'lr': 0.0004905247743166796, 'samples': 3185152, 'steps': 6220, 'loss/train': 1.8612608909606934} 02/24/2022 09:37:21 - INFO - codeparrot_training - Step 6221: {'lr': 0.0004905203117570899, 'samples': 3185664, 'steps': 6221, 'loss/train': 2.313664436340332} 02/24/2022 09:37:27 - INFO - codeparrot_training - Step 6222: {'lr': 0.0004905158481671876, 'samples': 3186176, 'steps': 6222, 'loss/train': 1.283319354057312} 02/24/2022 09:37:30 - INFO - codeparrot_training - Step 6223: {'lr': 0.0004905113835469918, 'samples': 3186688, 'steps': 6223, 'loss/train': 0.3306739330291748} 02/24/2022 09:37:38 - INFO - codeparrot_training - Step 6224: {'lr': 0.0004905069178965214, 'samples': 3187200, 'steps': 6224, 'loss/train': 0.4955524504184723} 02/24/2022 09:37:41 - INFO - codeparrot_training - Step 6225: {'lr': 0.0004905024512157959, 'samples': 3187712, 'steps': 6225, 'loss/train': 1.918975591659546} 02/24/2022 09:37:47 - INFO - codeparrot_training - Step 6226: {'lr': 0.0004904979835048343, 'samples': 3188224, 'steps': 6226, 'loss/train': 2.428847074508667} 02/24/2022 09:37:50 - INFO - codeparrot_training - Step 6227: {'lr': 0.0004904935147636557, 'samples': 3188736, 'steps': 6227, 'loss/train': 1.7641596794128418} 02/24/2022 09:37:56 - INFO - codeparrot_training - Step 6228: {'lr': 0.0004904890449922792, 'samples': 3189248, 'steps': 6228, 'loss/train': 2.264202117919922} 02/24/2022 09:37:59 - INFO - codeparrot_training - Step 6229: {'lr': 0.0004904845741907241, 'samples': 3189760, 'steps': 6229, 'loss/train': 2.421705961227417} 02/24/2022 09:38:05 - INFO - codeparrot_training - Step 6230: {'lr': 0.0004904801023590094, 'samples': 3190272, 'steps': 6230, 'loss/train': 0.2653506398200989} 02/24/2022 09:38:08 - INFO - codeparrot_training - Step 6231: {'lr': 0.0004904756294971541, 'samples': 3190784, 'steps': 6231, 'loss/train': 2.5598371028900146} 02/24/2022 09:38:14 - INFO - codeparrot_training - Step 6232: {'lr': 0.0004904711556051778, 'samples': 3191296, 'steps': 6232, 'loss/train': 3.231922149658203} 02/24/2022 09:38:17 - INFO - codeparrot_training - Step 6233: {'lr': 0.0004904666806830992, 'samples': 3191808, 'steps': 6233, 'loss/train': 2.381088972091675} 02/24/2022 09:38:25 - INFO - codeparrot_training - Step 6234: {'lr': 0.0004904622047309379, 'samples': 3192320, 'steps': 6234, 'loss/train': 2.4190409183502197} 02/24/2022 09:38:28 - INFO - codeparrot_training - Step 6235: {'lr': 0.0004904577277487129, 'samples': 3192832, 'steps': 6235, 'loss/train': 2.964531898498535} 02/24/2022 09:38:34 - INFO - codeparrot_training - Step 6236: {'lr': 0.0004904532497364432, 'samples': 3193344, 'steps': 6236, 'loss/train': 2.159140110015869} 02/24/2022 09:38:37 - INFO - codeparrot_training - Step 6237: {'lr': 0.0004904487706941481, 'samples': 3193856, 'steps': 6237, 'loss/train': 1.2483268976211548} 02/24/2022 09:38:43 - INFO - codeparrot_training - Step 6238: {'lr': 0.000490444290621847, 'samples': 3194368, 'steps': 6238, 'loss/train': 3.2070467472076416} 02/24/2022 09:38:47 - INFO - codeparrot_training - Step 6239: {'lr': 0.0004904398095195588, 'samples': 3194880, 'steps': 6239, 'loss/train': 1.5273369550704956} 02/24/2022 09:38:50 - INFO - codeparrot_training - Step 6240: {'lr': 0.0004904353273873028, 'samples': 3195392, 'steps': 6240, 'loss/train': 3.5635406970977783} 02/24/2022 09:38:56 - INFO - codeparrot_training - Step 6241: {'lr': 0.0004904308442250983, 'samples': 3195904, 'steps': 6241, 'loss/train': 2.4514782428741455} 02/24/2022 09:38:59 - INFO - codeparrot_training - Step 6242: {'lr': 0.0004904263600329643, 'samples': 3196416, 'steps': 6242, 'loss/train': 2.590282917022705} 02/24/2022 09:39:05 - INFO - codeparrot_training - Step 6243: {'lr': 0.0004904218748109201, 'samples': 3196928, 'steps': 6243, 'loss/train': 1.0154826641082764} 02/24/2022 09:39:08 - INFO - codeparrot_training - Step 6244: {'lr': 0.000490417388558985, 'samples': 3197440, 'steps': 6244, 'loss/train': 2.410426378250122} 02/24/2022 09:39:14 - INFO - codeparrot_training - Step 6245: {'lr': 0.0004904129012771782, 'samples': 3197952, 'steps': 6245, 'loss/train': 2.0995585918426514} 02/24/2022 09:39:17 - INFO - codeparrot_training - Step 6246: {'lr': 0.0004904084129655188, 'samples': 3198464, 'steps': 6246, 'loss/train': 0.30579087138175964} 02/24/2022 09:39:23 - INFO - codeparrot_training - Step 6247: {'lr': 0.000490403923624026, 'samples': 3198976, 'steps': 6247, 'loss/train': 1.0417059659957886} 02/24/2022 09:39:26 - INFO - codeparrot_training - Step 6248: {'lr': 0.0004903994332527193, 'samples': 3199488, 'steps': 6248, 'loss/train': 2.3782055377960205} 02/24/2022 09:39:31 - INFO - codeparrot_training - Step 6249: {'lr': 0.0004903949418516178, 'samples': 3200000, 'steps': 6249, 'loss/train': 2.0115253925323486} 02/24/2022 09:39:35 - INFO - codeparrot_training - Step 6250: {'lr': 0.0004903904494207405, 'samples': 3200512, 'steps': 6250, 'loss/train': 2.2685794830322266} 02/24/2022 09:39:42 - INFO - codeparrot_training - Step 6251: {'lr': 0.000490385955960107, 'samples': 3201024, 'steps': 6251, 'loss/train': 2.6550660133361816} 02/24/2022 09:39:46 - INFO - codeparrot_training - Step 6252: {'lr': 0.0004903814614697363, 'samples': 3201536, 'steps': 6252, 'loss/train': 3.2062313556671143} 02/24/2022 09:39:51 - INFO - codeparrot_training - Step 6253: {'lr': 0.0004903769659496478, 'samples': 3202048, 'steps': 6253, 'loss/train': 3.9077563285827637} 02/24/2022 09:39:57 - INFO - codeparrot_training - Step 6254: {'lr': 0.0004903724693998607, 'samples': 3202560, 'steps': 6254, 'loss/train': 3.3275156021118164} 02/24/2022 09:40:00 - INFO - codeparrot_training - Step 6255: {'lr': 0.0004903679718203942, 'samples': 3203072, 'steps': 6255, 'loss/train': 1.2602351903915405} 02/24/2022 09:40:04 - INFO - codeparrot_training - Step 6256: {'lr': 0.0004903634732112678, 'samples': 3203584, 'steps': 6256, 'loss/train': 2.4089462757110596} 02/24/2022 09:40:09 - INFO - codeparrot_training - Step 6257: {'lr': 0.0004903589735725004, 'samples': 3204096, 'steps': 6257, 'loss/train': 1.7523692846298218} 02/24/2022 09:40:13 - INFO - codeparrot_training - Step 6258: {'lr': 0.0004903544729041116, 'samples': 3204608, 'steps': 6258, 'loss/train': 2.9210150241851807} 02/24/2022 09:40:18 - INFO - codeparrot_training - Step 6259: {'lr': 0.0004903499712061206, 'samples': 3205120, 'steps': 6259, 'loss/train': 3.1442933082580566} 02/24/2022 09:40:22 - INFO - codeparrot_training - Step 6260: {'lr': 0.0004903454684785465, 'samples': 3205632, 'steps': 6260, 'loss/train': 1.6726661920547485} 02/24/2022 09:40:29 - INFO - codeparrot_training - Step 6261: {'lr': 0.0004903409647214088, 'samples': 3206144, 'steps': 6261, 'loss/train': 2.656792402267456} 02/24/2022 09:40:32 - INFO - codeparrot_training - Step 6262: {'lr': 0.0004903364599347268, 'samples': 3206656, 'steps': 6262, 'loss/train': 2.310574769973755} 02/24/2022 09:40:38 - INFO - codeparrot_training - Step 6263: {'lr': 0.0004903319541185196, 'samples': 3207168, 'steps': 6263, 'loss/train': 2.427889823913574} 02/24/2022 09:40:41 - INFO - codeparrot_training - Step 6264: {'lr': 0.0004903274472728067, 'samples': 3207680, 'steps': 6264, 'loss/train': 2.4989237785339355} 02/24/2022 09:40:47 - INFO - codeparrot_training - Step 6265: {'lr': 0.0004903229393976073, 'samples': 3208192, 'steps': 6265, 'loss/train': 3.0231387615203857} 02/24/2022 09:40:50 - INFO - codeparrot_training - Step 6266: {'lr': 0.0004903184304929408, 'samples': 3208704, 'steps': 6266, 'loss/train': 2.4457943439483643} 02/24/2022 09:40:56 - INFO - codeparrot_training - Step 6267: {'lr': 0.0004903139205588264, 'samples': 3209216, 'steps': 6267, 'loss/train': 1.440609097480774} 02/24/2022 09:40:59 - INFO - codeparrot_training - Step 6268: {'lr': 0.0004903094095952834, 'samples': 3209728, 'steps': 6268, 'loss/train': 2.4216387271881104} 02/24/2022 09:41:07 - INFO - codeparrot_training - Step 6269: {'lr': 0.0004903048976023313, 'samples': 3210240, 'steps': 6269, 'loss/train': 1.8143019676208496} 02/24/2022 09:41:12 - INFO - codeparrot_training - Step 6270: {'lr': 0.0004903003845799893, 'samples': 3210752, 'steps': 6270, 'loss/train': 2.3480634689331055} 02/24/2022 09:41:16 - INFO - codeparrot_training - Step 6271: {'lr': 0.0004902958705282767, 'samples': 3211264, 'steps': 6271, 'loss/train': 0.1486256718635559} 02/24/2022 09:41:21 - INFO - codeparrot_training - Step 6272: {'lr': 0.000490291355447213, 'samples': 3211776, 'steps': 6272, 'loss/train': 1.9411009550094604} 02/24/2022 09:41:25 - INFO - codeparrot_training - Step 6273: {'lr': 0.0004902868393368174, 'samples': 3212288, 'steps': 6273, 'loss/train': 2.6457629203796387} 02/24/2022 09:41:30 - INFO - codeparrot_training - Step 6274: {'lr': 0.0004902823221971092, 'samples': 3212800, 'steps': 6274, 'loss/train': 3.7133326530456543} 02/24/2022 09:41:34 - INFO - codeparrot_training - Step 6275: {'lr': 0.000490277804028108, 'samples': 3213312, 'steps': 6275, 'loss/train': 2.910630226135254} 02/24/2022 09:41:39 - INFO - codeparrot_training - Step 6276: {'lr': 0.0004902732848298328, 'samples': 3213824, 'steps': 6276, 'loss/train': 2.3521180152893066} 02/24/2022 09:41:43 - INFO - codeparrot_training - Step 6277: {'lr': 0.0004902687646023032, 'samples': 3214336, 'steps': 6277, 'loss/train': 3.0850539207458496} 02/24/2022 09:41:49 - INFO - codeparrot_training - Step 6278: {'lr': 0.0004902642433455385, 'samples': 3214848, 'steps': 6278, 'loss/train': 1.6708606481552124} 02/24/2022 09:41:52 - INFO - codeparrot_training - Step 6279: {'lr': 0.0004902597210595581, 'samples': 3215360, 'steps': 6279, 'loss/train': 2.247377872467041} 02/24/2022 09:41:58 - INFO - codeparrot_training - Step 6280: {'lr': 0.0004902551977443813, 'samples': 3215872, 'steps': 6280, 'loss/train': 1.9600398540496826} 02/24/2022 09:42:01 - INFO - codeparrot_training - Step 6281: {'lr': 0.0004902506734000276, 'samples': 3216384, 'steps': 6281, 'loss/train': 1.8372225761413574} 02/24/2022 09:42:07 - INFO - codeparrot_training - Step 6282: {'lr': 0.0004902461480265163, 'samples': 3216896, 'steps': 6282, 'loss/train': 2.0583794116973877} 02/24/2022 09:42:10 - INFO - codeparrot_training - Step 6283: {'lr': 0.0004902416216238667, 'samples': 3217408, 'steps': 6283, 'loss/train': 2.1329660415649414} 02/24/2022 09:42:16 - INFO - codeparrot_training - Step 6284: {'lr': 0.0004902370941920984, 'samples': 3217920, 'steps': 6284, 'loss/train': 2.5528974533081055} 02/24/2022 09:42:19 - INFO - codeparrot_training - Step 6285: {'lr': 0.0004902325657312306, 'samples': 3218432, 'steps': 6285, 'loss/train': 3.1858463287353516} 02/24/2022 09:42:27 - INFO - codeparrot_training - Step 6286: {'lr': 0.0004902280362412828, 'samples': 3218944, 'steps': 6286, 'loss/train': 0.1197754368185997} 02/24/2022 09:42:30 - INFO - codeparrot_training - Step 6287: {'lr': 0.0004902235057222743, 'samples': 3219456, 'steps': 6287, 'loss/train': 2.2900893688201904} 02/24/2022 09:42:36 - INFO - codeparrot_training - Step 6288: {'lr': 0.0004902189741742246, 'samples': 3219968, 'steps': 6288, 'loss/train': 2.8684144020080566} 02/24/2022 09:42:39 - INFO - codeparrot_training - Step 6289: {'lr': 0.0004902144415971532, 'samples': 3220480, 'steps': 6289, 'loss/train': 2.5648696422576904} 02/24/2022 09:42:45 - INFO - codeparrot_training - Step 6290: {'lr': 0.0004902099079910794, 'samples': 3220992, 'steps': 6290, 'loss/train': 2.3180229663848877} 02/24/2022 09:42:48 - INFO - codeparrot_training - Step 6291: {'lr': 0.0004902053733560225, 'samples': 3221504, 'steps': 6291, 'loss/train': 1.8863275051116943} 02/24/2022 09:42:54 - INFO - codeparrot_training - Step 6292: {'lr': 0.0004902008376920021, 'samples': 3222016, 'steps': 6292, 'loss/train': 2.350266695022583} 02/24/2022 09:42:57 - INFO - codeparrot_training - Step 6293: {'lr': 0.0004901963009990376, 'samples': 3222528, 'steps': 6293, 'loss/train': 1.3136851787567139} 02/24/2022 09:43:02 - INFO - codeparrot_training - Step 6294: {'lr': 0.0004901917632771485, 'samples': 3223040, 'steps': 6294, 'loss/train': 4.365197658538818} 02/24/2022 09:43:06 - INFO - codeparrot_training - Step 6295: {'lr': 0.000490187224526354, 'samples': 3223552, 'steps': 6295, 'loss/train': 0.6939431428909302} 02/24/2022 09:43:13 - INFO - codeparrot_training - Step 6296: {'lr': 0.0004901826847466738, 'samples': 3224064, 'steps': 6296, 'loss/train': 1.7623052597045898} 02/24/2022 09:43:17 - INFO - codeparrot_training - Step 6297: {'lr': 0.0004901781439381272, 'samples': 3224576, 'steps': 6297, 'loss/train': 1.0246156454086304} 02/24/2022 09:43:22 - INFO - codeparrot_training - Step 6298: {'lr': 0.0004901736021007337, 'samples': 3225088, 'steps': 6298, 'loss/train': 3.066565752029419} 02/24/2022 09:43:26 - INFO - codeparrot_training - Step 6299: {'lr': 0.0004901690592345127, 'samples': 3225600, 'steps': 6299, 'loss/train': 2.188586711883545} 02/24/2022 09:43:31 - INFO - codeparrot_training - Step 6300: {'lr': 0.0004901645153394838, 'samples': 3226112, 'steps': 6300, 'loss/train': 1.838338851928711} 02/24/2022 09:43:35 - INFO - codeparrot_training - Step 6301: {'lr': 0.0004901599704156664, 'samples': 3226624, 'steps': 6301, 'loss/train': 2.1207802295684814} 02/24/2022 09:43:40 - INFO - codeparrot_training - Step 6302: {'lr': 0.00049015542446308, 'samples': 3227136, 'steps': 6302, 'loss/train': 3.0751938819885254} 02/24/2022 09:43:44 - INFO - codeparrot_training - Step 6303: {'lr': 0.0004901508774817438, 'samples': 3227648, 'steps': 6303, 'loss/train': 2.528035879135132} 02/24/2022 09:43:49 - INFO - codeparrot_training - Step 6304: {'lr': 0.0004901463294716776, 'samples': 3228160, 'steps': 6304, 'loss/train': 2.6392056941986084} 02/24/2022 09:43:53 - INFO - codeparrot_training - Step 6305: {'lr': 0.0004901417804329008, 'samples': 3228672, 'steps': 6305, 'loss/train': 2.7946455478668213} 02/24/2022 09:44:00 - INFO - codeparrot_training - Step 6306: {'lr': 0.0004901372303654329, 'samples': 3229184, 'steps': 6306, 'loss/train': 2.7061655521392822} 02/24/2022 09:44:03 - INFO - codeparrot_training - Step 6307: {'lr': 0.0004901326792692934, 'samples': 3229696, 'steps': 6307, 'loss/train': 2.344365358352661} 02/24/2022 09:44:09 - INFO - codeparrot_training - Step 6308: {'lr': 0.0004901281271445016, 'samples': 3230208, 'steps': 6308, 'loss/train': 2.365645170211792} 02/24/2022 09:44:13 - INFO - codeparrot_training - Step 6309: {'lr': 0.0004901235739910772, 'samples': 3230720, 'steps': 6309, 'loss/train': 2.601585626602173} 02/24/2022 09:44:18 - INFO - codeparrot_training - Step 6310: {'lr': 0.0004901190198090397, 'samples': 3231232, 'steps': 6310, 'loss/train': 3.2509396076202393} 02/24/2022 09:44:22 - INFO - codeparrot_training - Step 6311: {'lr': 0.0004901144645984086, 'samples': 3231744, 'steps': 6311, 'loss/train': 0.9778910279273987} 02/24/2022 09:44:27 - INFO - codeparrot_training - Step 6312: {'lr': 0.0004901099083592034, 'samples': 3232256, 'steps': 6312, 'loss/train': 2.777360677719116} 02/24/2022 09:44:31 - INFO - codeparrot_training - Step 6313: {'lr': 0.0004901053510914434, 'samples': 3232768, 'steps': 6313, 'loss/train': 2.0888760089874268} 02/24/2022 09:44:36 - INFO - codeparrot_training - Step 6314: {'lr': 0.0004901007927951485, 'samples': 3233280, 'steps': 6314, 'loss/train': 2.6442201137542725} 02/24/2022 09:44:43 - INFO - codeparrot_training - Step 6315: {'lr': 0.000490096233470338, 'samples': 3233792, 'steps': 6315, 'loss/train': 2.640913724899292} 02/24/2022 09:44:47 - INFO - codeparrot_training - Step 6316: {'lr': 0.0004900916731170314, 'samples': 3234304, 'steps': 6316, 'loss/train': 2.3063735961914062} 02/24/2022 09:44:50 - INFO - codeparrot_training - Step 6317: {'lr': 0.0004900871117352484, 'samples': 3234816, 'steps': 6317, 'loss/train': 2.2915847301483154} 02/24/2022 09:44:56 - INFO - codeparrot_training - Step 6318: {'lr': 0.0004900825493250084, 'samples': 3235328, 'steps': 6318, 'loss/train': 2.087425708770752} 02/24/2022 09:44:59 - INFO - codeparrot_training - Step 6319: {'lr': 0.000490077985886331, 'samples': 3235840, 'steps': 6319, 'loss/train': 2.266817569732666} 02/24/2022 09:45:05 - INFO - codeparrot_training - Step 6320: {'lr': 0.0004900734214192358, 'samples': 3236352, 'steps': 6320, 'loss/train': 0.4193190336227417} 02/24/2022 09:45:08 - INFO - codeparrot_training - Step 6321: {'lr': 0.0004900688559237422, 'samples': 3236864, 'steps': 6321, 'loss/train': 2.32072377204895} 02/24/2022 09:45:14 - INFO - codeparrot_training - Step 6322: {'lr': 0.0004900642893998699, 'samples': 3237376, 'steps': 6322, 'loss/train': 1.146812081336975} 02/24/2022 09:45:17 - INFO - codeparrot_training - Step 6323: {'lr': 0.0004900597218476385, 'samples': 3237888, 'steps': 6323, 'loss/train': 3.1454150676727295} 02/24/2022 09:45:23 - INFO - codeparrot_training - Step 6324: {'lr': 0.0004900551532670673, 'samples': 3238400, 'steps': 6324, 'loss/train': 2.032609701156616} 02/24/2022 09:45:26 - INFO - codeparrot_training - Step 6325: {'lr': 0.0004900505836581763, 'samples': 3238912, 'steps': 6325, 'loss/train': 2.3087079524993896} 02/24/2022 09:45:32 - INFO - codeparrot_training - Step 6326: {'lr': 0.0004900460130209845, 'samples': 3239424, 'steps': 6326, 'loss/train': 2.2144553661346436} 02/24/2022 09:45:35 - INFO - codeparrot_training - Step 6327: {'lr': 0.000490041441355512, 'samples': 3239936, 'steps': 6327, 'loss/train': 3.5378143787384033} 02/24/2022 09:45:41 - INFO - codeparrot_training - Step 6328: {'lr': 0.0004900368686617783, 'samples': 3240448, 'steps': 6328, 'loss/train': 2.095506191253662} 02/24/2022 09:45:44 - INFO - codeparrot_training - Step 6329: {'lr': 0.0004900322949398026, 'samples': 3240960, 'steps': 6329, 'loss/train': 2.1152613162994385} 02/24/2022 09:45:50 - INFO - codeparrot_training - Step 6330: {'lr': 0.000490027720189605, 'samples': 3241472, 'steps': 6330, 'loss/train': 2.2639265060424805} 02/24/2022 09:45:53 - INFO - codeparrot_training - Step 6331: {'lr': 0.0004900231444112047, 'samples': 3241984, 'steps': 6331, 'loss/train': 1.6293816566467285} 02/24/2022 09:45:59 - INFO - codeparrot_training - Step 6332: {'lr': 0.0004900185676046214, 'samples': 3242496, 'steps': 6332, 'loss/train': 2.488816022872925} 02/24/2022 09:46:06 - INFO - codeparrot_training - Step 6333: {'lr': 0.0004900139897698751, 'samples': 3243008, 'steps': 6333, 'loss/train': 2.4937851428985596} 02/24/2022 09:46:10 - INFO - codeparrot_training - Step 6334: {'lr': 0.0004900094109069848, 'samples': 3243520, 'steps': 6334, 'loss/train': 1.6818186044692993} 02/24/2022 09:46:15 - INFO - codeparrot_training - Step 6335: {'lr': 0.0004900048310159705, 'samples': 3244032, 'steps': 6335, 'loss/train': 2.6479344367980957} 02/24/2022 09:46:19 - INFO - codeparrot_training - Step 6336: {'lr': 0.0004900002500968516, 'samples': 3244544, 'steps': 6336, 'loss/train': 2.1148204803466797} 02/24/2022 09:46:24 - INFO - codeparrot_training - Step 6337: {'lr': 0.000489995668149648, 'samples': 3245056, 'steps': 6337, 'loss/train': 2.4322540760040283} 02/24/2022 09:46:27 - INFO - codeparrot_training - Step 6338: {'lr': 0.0004899910851743791, 'samples': 3245568, 'steps': 6338, 'loss/train': 1.680912971496582} 02/24/2022 09:46:33 - INFO - codeparrot_training - Step 6339: {'lr': 0.0004899865011710646, 'samples': 3246080, 'steps': 6339, 'loss/train': 1.9332932233810425} 02/24/2022 09:46:37 - INFO - codeparrot_training - Step 6340: {'lr': 0.0004899819161397241, 'samples': 3246592, 'steps': 6340, 'loss/train': 8.981110572814941} 02/24/2022 09:46:42 - INFO - codeparrot_training - Step 6341: {'lr': 0.0004899773300803774, 'samples': 3247104, 'steps': 6341, 'loss/train': 2.545008659362793} 02/24/2022 09:46:46 - INFO - codeparrot_training - Step 6342: {'lr': 0.0004899727429930438, 'samples': 3247616, 'steps': 6342, 'loss/train': 2.2221126556396484} 02/24/2022 09:46:53 - INFO - codeparrot_training - Step 6343: {'lr': 0.0004899681548777434, 'samples': 3248128, 'steps': 6343, 'loss/train': 2.2362680435180664} 02/24/2022 09:46:57 - INFO - codeparrot_training - Step 6344: {'lr': 0.0004899635657344954, 'samples': 3248640, 'steps': 6344, 'loss/train': 2.0811617374420166} 02/24/2022 09:47:02 - INFO - codeparrot_training - Step 6345: {'lr': 0.0004899589755633198, 'samples': 3249152, 'steps': 6345, 'loss/train': 2.934920310974121} 02/24/2022 09:47:06 - INFO - codeparrot_training - Step 6346: {'lr': 0.0004899543843642362, 'samples': 3249664, 'steps': 6346, 'loss/train': 3.1778531074523926} 02/24/2022 09:47:11 - INFO - codeparrot_training - Step 6347: {'lr': 0.0004899497921372641, 'samples': 3250176, 'steps': 6347, 'loss/train': 2.080864429473877} 02/24/2022 09:47:15 - INFO - codeparrot_training - Step 6348: {'lr': 0.0004899451988824233, 'samples': 3250688, 'steps': 6348, 'loss/train': 1.7839866876602173} 02/24/2022 09:47:20 - INFO - codeparrot_training - Step 6349: {'lr': 0.0004899406045997336, 'samples': 3251200, 'steps': 6349, 'loss/train': 2.041637420654297} 02/24/2022 09:47:23 - INFO - codeparrot_training - Step 6350: {'lr': 0.0004899360092892143, 'samples': 3251712, 'steps': 6350, 'loss/train': 2.313155174255371} 02/24/2022 09:47:29 - INFO - codeparrot_training - Step 6351: {'lr': 0.0004899314129508855, 'samples': 3252224, 'steps': 6351, 'loss/train': 1.9384204149246216} 02/24/2022 09:47:32 - INFO - codeparrot_training - Step 6352: {'lr': 0.0004899268155847667, 'samples': 3252736, 'steps': 6352, 'loss/train': 3.5875790119171143} 02/24/2022 09:47:40 - INFO - codeparrot_training - Step 6353: {'lr': 0.0004899222171908776, 'samples': 3253248, 'steps': 6353, 'loss/train': 1.1855950355529785} 02/24/2022 09:47:43 - INFO - codeparrot_training - Step 6354: {'lr': 0.0004899176177692379, 'samples': 3253760, 'steps': 6354, 'loss/train': 1.4508904218673706} 02/24/2022 09:47:49 - INFO - codeparrot_training - Step 6355: {'lr': 0.0004899130173198672, 'samples': 3254272, 'steps': 6355, 'loss/train': 2.6990723609924316} 02/24/2022 09:47:52 - INFO - codeparrot_training - Step 6356: {'lr': 0.0004899084158427855, 'samples': 3254784, 'steps': 6356, 'loss/train': 3.017900228500366} 02/24/2022 09:47:58 - INFO - codeparrot_training - Step 6357: {'lr': 0.0004899038133380121, 'samples': 3255296, 'steps': 6357, 'loss/train': 1.9416550397872925} 02/24/2022 09:48:01 - INFO - codeparrot_training - Step 6358: {'lr': 0.0004898992098055671, 'samples': 3255808, 'steps': 6358, 'loss/train': 1.2342463731765747} 02/24/2022 09:48:07 - INFO - codeparrot_training - Step 6359: {'lr': 0.00048989460524547, 'samples': 3256320, 'steps': 6359, 'loss/train': 2.060485601425171} 02/24/2022 09:48:10 - INFO - codeparrot_training - Step 6360: {'lr': 0.0004898899996577407, 'samples': 3256832, 'steps': 6360, 'loss/train': 2.177398920059204} 02/24/2022 09:48:16 - INFO - codeparrot_training - Step 6361: {'lr': 0.0004898853930423987, 'samples': 3257344, 'steps': 6361, 'loss/train': 2.6436080932617188} 02/24/2022 09:48:19 - INFO - codeparrot_training - Step 6362: {'lr': 0.0004898807853994639, 'samples': 3257856, 'steps': 6362, 'loss/train': 1.580167293548584} 02/24/2022 09:48:26 - INFO - codeparrot_training - Step 6363: {'lr': 0.000489876176728956, 'samples': 3258368, 'steps': 6363, 'loss/train': 3.411489963531494} 02/24/2022 09:48:30 - INFO - codeparrot_training - Step 6364: {'lr': 0.0004898715670308947, 'samples': 3258880, 'steps': 6364, 'loss/train': 2.096208333969116} 02/24/2022 09:48:35 - INFO - codeparrot_training - Step 6365: {'lr': 0.0004898669563052997, 'samples': 3259392, 'steps': 6365, 'loss/train': 2.6656792163848877} 02/24/2022 09:48:39 - INFO - codeparrot_training - Step 6366: {'lr': 0.0004898623445521909, 'samples': 3259904, 'steps': 6366, 'loss/train': 3.165503740310669} 02/24/2022 09:48:44 - INFO - codeparrot_training - Step 6367: {'lr': 0.000489857731771588, 'samples': 3260416, 'steps': 6367, 'loss/train': 2.6996541023254395} 02/24/2022 09:48:48 - INFO - codeparrot_training - Step 6368: {'lr': 0.0004898531179635108, 'samples': 3260928, 'steps': 6368, 'loss/train': 2.4503157138824463} 02/24/2022 09:48:53 - INFO - codeparrot_training - Step 6369: {'lr': 0.0004898485031279788, 'samples': 3261440, 'steps': 6369, 'loss/train': 4.438111782073975} 02/24/2022 09:48:57 - INFO - codeparrot_training - Step 6370: {'lr': 0.0004898438872650121, 'samples': 3261952, 'steps': 6370, 'loss/train': 2.9785115718841553} 02/24/2022 09:49:02 - INFO - codeparrot_training - Step 6371: {'lr': 0.0004898392703746304, 'samples': 3262464, 'steps': 6371, 'loss/train': 3.1848602294921875} 02/24/2022 09:49:06 - INFO - codeparrot_training - Step 6372: {'lr': 0.0004898346524568533, 'samples': 3262976, 'steps': 6372, 'loss/train': 3.0334019660949707} 02/24/2022 09:49:11 - INFO - codeparrot_training - Step 6373: {'lr': 0.0004898300335117008, 'samples': 3263488, 'steps': 6373, 'loss/train': 2.1745877265930176} 02/24/2022 09:49:15 - INFO - codeparrot_training - Step 6374: {'lr': 0.0004898254135391926, 'samples': 3264000, 'steps': 6374, 'loss/train': 2.3121368885040283} 02/24/2022 09:49:20 - INFO - codeparrot_training - Step 6375: {'lr': 0.0004898207925393485, 'samples': 3264512, 'steps': 6375, 'loss/train': 2.5151071548461914} 02/24/2022 09:49:24 - INFO - codeparrot_training - Step 6376: {'lr': 0.0004898161705121882, 'samples': 3265024, 'steps': 6376, 'loss/train': 1.7923381328582764} 02/24/2022 09:49:31 - INFO - codeparrot_training - Step 6377: {'lr': 0.0004898115474577315, 'samples': 3265536, 'steps': 6377, 'loss/train': 2.277191638946533} 02/24/2022 09:49:35 - INFO - codeparrot_training - Step 6378: {'lr': 0.0004898069233759985, 'samples': 3266048, 'steps': 6378, 'loss/train': 1.523551106452942} 02/24/2022 09:49:40 - INFO - codeparrot_training - Step 6379: {'lr': 0.0004898022982670085, 'samples': 3266560, 'steps': 6379, 'loss/train': 1.7236496210098267} 02/24/2022 09:49:44 - INFO - codeparrot_training - Step 6380: {'lr': 0.0004897976721307818, 'samples': 3267072, 'steps': 6380, 'loss/train': 2.342064619064331} 02/24/2022 09:49:49 - INFO - codeparrot_training - Step 6381: {'lr': 0.000489793044967338, 'samples': 3267584, 'steps': 6381, 'loss/train': 2.5133140087127686} 02/24/2022 09:49:53 - INFO - codeparrot_training - Step 6382: {'lr': 0.0004897884167766969, 'samples': 3268096, 'steps': 6382, 'loss/train': 2.8323864936828613} 02/24/2022 09:49:58 - INFO - codeparrot_training - Step 6383: {'lr': 0.0004897837875588784, 'samples': 3268608, 'steps': 6383, 'loss/train': 2.454118490219116} 02/24/2022 09:50:02 - INFO - codeparrot_training - Step 6384: {'lr': 0.0004897791573139022, 'samples': 3269120, 'steps': 6384, 'loss/train': 2.2801244258880615} 02/24/2022 09:50:08 - INFO - codeparrot_training - Step 6385: {'lr': 0.0004897745260417884, 'samples': 3269632, 'steps': 6385, 'loss/train': 1.8247528076171875} 02/24/2022 09:50:11 - INFO - codeparrot_training - Step 6386: {'lr': 0.0004897698937425566, 'samples': 3270144, 'steps': 6386, 'loss/train': 1.270604133605957} 02/24/2022 09:50:17 - INFO - codeparrot_training - Step 6387: {'lr': 0.0004897652604162266, 'samples': 3270656, 'steps': 6387, 'loss/train': 1.784490704536438} 02/24/2022 09:50:20 - INFO - codeparrot_training - Step 6388: {'lr': 0.0004897606260628184, 'samples': 3271168, 'steps': 6388, 'loss/train': 2.2898740768432617} 02/24/2022 09:50:28 - INFO - codeparrot_training - Step 6389: {'lr': 0.0004897559906823518, 'samples': 3271680, 'steps': 6389, 'loss/train': 1.523402214050293} 02/24/2022 09:50:31 - INFO - codeparrot_training - Step 6390: {'lr': 0.0004897513542748468, 'samples': 3272192, 'steps': 6390, 'loss/train': 3.0109260082244873} 02/24/2022 09:50:37 - INFO - codeparrot_training - Step 6391: {'lr': 0.0004897467168403231, 'samples': 3272704, 'steps': 6391, 'loss/train': 1.308800220489502} 02/24/2022 09:50:40 - INFO - codeparrot_training - Step 6392: {'lr': 0.0004897420783788006, 'samples': 3273216, 'steps': 6392, 'loss/train': 2.723564624786377} 02/24/2022 09:50:46 - INFO - codeparrot_training - Step 6393: {'lr': 0.0004897374388902991, 'samples': 3273728, 'steps': 6393, 'loss/train': 1.9558881521224976} 02/24/2022 09:50:49 - INFO - codeparrot_training - Step 6394: {'lr': 0.0004897327983748385, 'samples': 3274240, 'steps': 6394, 'loss/train': 0.46763065457344055} 02/24/2022 09:50:55 - INFO - codeparrot_training - Step 6395: {'lr': 0.0004897281568324387, 'samples': 3274752, 'steps': 6395, 'loss/train': 2.242149829864502} 02/24/2022 09:50:58 - INFO - codeparrot_training - Step 6396: {'lr': 0.0004897235142631197, 'samples': 3275264, 'steps': 6396, 'loss/train': 2.9120097160339355} 02/24/2022 09:51:03 - INFO - codeparrot_training - Step 6397: {'lr': 0.0004897188706669012, 'samples': 3275776, 'steps': 6397, 'loss/train': 3.2725422382354736} 02/24/2022 09:51:07 - INFO - codeparrot_training - Step 6398: {'lr': 0.0004897142260438032, 'samples': 3276288, 'steps': 6398, 'loss/train': 0.6915867924690247} 02/24/2022 09:51:14 - INFO - codeparrot_training - Step 6399: {'lr': 0.0004897095803938456, 'samples': 3276800, 'steps': 6399, 'loss/train': 2.606157064437866} 02/24/2022 09:51:18 - INFO - codeparrot_training - Step 6400: {'lr': 0.0004897049337170483, 'samples': 3277312, 'steps': 6400, 'loss/train': 3.485391855239868} 02/24/2022 09:51:23 - INFO - codeparrot_training - Step 6401: {'lr': 0.0004897002860134311, 'samples': 3277824, 'steps': 6401, 'loss/train': 2.1284947395324707} 02/24/2022 09:51:27 - INFO - codeparrot_training - Step 6402: {'lr': 0.0004896956372830141, 'samples': 3278336, 'steps': 6402, 'loss/train': 2.415383815765381} 02/24/2022 09:51:32 - INFO - codeparrot_training - Step 6403: {'lr': 0.000489690987525817, 'samples': 3278848, 'steps': 6403, 'loss/train': 2.4144012928009033} 02/24/2022 09:51:36 - INFO - codeparrot_training - Step 6404: {'lr': 0.0004896863367418598, 'samples': 3279360, 'steps': 6404, 'loss/train': 1.3919150829315186} 02/24/2022 09:51:41 - INFO - codeparrot_training - Step 6405: {'lr': 0.0004896816849311625, 'samples': 3279872, 'steps': 6405, 'loss/train': 3.055471420288086} 02/24/2022 09:51:45 - INFO - codeparrot_training - Step 6406: {'lr': 0.000489677032093745, 'samples': 3280384, 'steps': 6406, 'loss/train': 2.631194829940796} 02/24/2022 09:51:50 - INFO - codeparrot_training - Step 6407: {'lr': 0.0004896723782296272, 'samples': 3280896, 'steps': 6407, 'loss/train': 1.6270899772644043} 02/24/2022 09:51:54 - INFO - codeparrot_training - Step 6408: {'lr': 0.0004896677233388289, 'samples': 3281408, 'steps': 6408, 'loss/train': 1.408095121383667} 02/24/2022 09:52:01 - INFO - codeparrot_training - Step 6409: {'lr': 0.0004896630674213703, 'samples': 3281920, 'steps': 6409, 'loss/train': 3.0135111808776855} 02/24/2022 09:52:04 - INFO - codeparrot_training - Step 6410: {'lr': 0.0004896584104772712, 'samples': 3282432, 'steps': 6410, 'loss/train': 3.475926637649536} 02/24/2022 09:52:10 - INFO - codeparrot_training - Step 6411: {'lr': 0.0004896537525065516, 'samples': 3282944, 'steps': 6411, 'loss/train': 2.8265891075134277} 02/24/2022 09:52:13 - INFO - codeparrot_training - Step 6412: {'lr': 0.0004896490935092314, 'samples': 3283456, 'steps': 6412, 'loss/train': 3.4042012691497803} 02/24/2022 09:52:19 - INFO - codeparrot_training - Step 6413: {'lr': 0.0004896444334853305, 'samples': 3283968, 'steps': 6413, 'loss/train': 2.973315715789795} 02/24/2022 09:52:22 - INFO - codeparrot_training - Step 6414: {'lr': 0.000489639772434869, 'samples': 3284480, 'steps': 6414, 'loss/train': 2.2527859210968018} 02/24/2022 09:52:28 - INFO - codeparrot_training - Step 6415: {'lr': 0.0004896351103578669, 'samples': 3284992, 'steps': 6415, 'loss/train': 1.876029372215271} 02/24/2022 09:52:31 - INFO - codeparrot_training - Step 6416: {'lr': 0.0004896304472543439, 'samples': 3285504, 'steps': 6416, 'loss/train': 9.057655334472656} 02/24/2022 09:52:37 - INFO - codeparrot_training - Step 6417: {'lr': 0.0004896257831243204, 'samples': 3286016, 'steps': 6417, 'loss/train': 3.0918171405792236} 02/24/2022 09:52:40 - INFO - codeparrot_training - Step 6418: {'lr': 0.0004896211179678159, 'samples': 3286528, 'steps': 6418, 'loss/train': 3.134316921234131} 02/24/2022 09:52:48 - INFO - codeparrot_training - Step 6419: {'lr': 0.0004896164517848508, 'samples': 3287040, 'steps': 6419, 'loss/train': 2.4525632858276367} 02/24/2022 09:52:51 - INFO - codeparrot_training - Step 6420: {'lr': 0.0004896117845754448, 'samples': 3287552, 'steps': 6420, 'loss/train': 2.562161922454834} 02/24/2022 09:52:57 - INFO - codeparrot_training - Step 6421: {'lr': 0.0004896071163396179, 'samples': 3288064, 'steps': 6421, 'loss/train': 2.3492627143859863} 02/24/2022 09:53:00 - INFO - codeparrot_training - Step 6422: {'lr': 0.0004896024470773904, 'samples': 3288576, 'steps': 6422, 'loss/train': 1.8552896976470947} 02/24/2022 09:53:06 - INFO - codeparrot_training - Step 6423: {'lr': 0.000489597776788782, 'samples': 3289088, 'steps': 6423, 'loss/train': 1.3254491090774536} 02/24/2022 09:53:09 - INFO - codeparrot_training - Step 6424: {'lr': 0.0004895931054738128, 'samples': 3289600, 'steps': 6424, 'loss/train': 2.5252487659454346} 02/24/2022 09:53:15 - INFO - codeparrot_training - Step 6425: {'lr': 0.0004895884331325028, 'samples': 3290112, 'steps': 6425, 'loss/train': 2.3808586597442627} 02/24/2022 09:53:18 - INFO - codeparrot_training - Step 6426: {'lr': 0.0004895837597648721, 'samples': 3290624, 'steps': 6426, 'loss/train': 1.6469117403030396} 02/24/2022 09:53:24 - INFO - codeparrot_training - Step 6427: {'lr': 0.0004895790853709406, 'samples': 3291136, 'steps': 6427, 'loss/train': 2.1365747451782227} 02/24/2022 09:53:27 - INFO - codeparrot_training - Step 6428: {'lr': 0.0004895744099507284, 'samples': 3291648, 'steps': 6428, 'loss/train': 2.845057249069214} 02/24/2022 09:53:33 - INFO - codeparrot_training - Step 6429: {'lr': 0.0004895697335042555, 'samples': 3292160, 'steps': 6429, 'loss/train': 2.1584692001342773} 02/24/2022 09:53:38 - INFO - codeparrot_training - Step 6430: {'lr': 0.0004895650560315419, 'samples': 3292672, 'steps': 6430, 'loss/train': 2.8787264823913574} 02/24/2022 09:53:42 - INFO - codeparrot_training - Step 6431: {'lr': 0.0004895603775326077, 'samples': 3293184, 'steps': 6431, 'loss/train': 1.3936365842819214} 02/24/2022 09:53:47 - INFO - codeparrot_training - Step 6432: {'lr': 0.0004895556980074729, 'samples': 3293696, 'steps': 6432, 'loss/train': 1.6991254091262817} 02/24/2022 09:53:51 - INFO - codeparrot_training - Step 6433: {'lr': 0.0004895510174561576, 'samples': 3294208, 'steps': 6433, 'loss/train': 2.6044275760650635} 02/24/2022 09:53:58 - INFO - codeparrot_training - Step 6434: {'lr': 0.0004895463358786818, 'samples': 3294720, 'steps': 6434, 'loss/train': 1.8530462980270386} 02/24/2022 09:54:02 - INFO - codeparrot_training - Step 6435: {'lr': 0.0004895416532750655, 'samples': 3295232, 'steps': 6435, 'loss/train': 1.9207227230072021} 02/24/2022 09:54:07 - INFO - codeparrot_training - Step 6436: {'lr': 0.0004895369696453289, 'samples': 3295744, 'steps': 6436, 'loss/train': 2.309866428375244} 02/24/2022 09:54:11 - INFO - codeparrot_training - Step 6437: {'lr': 0.0004895322849894918, 'samples': 3296256, 'steps': 6437, 'loss/train': 2.4831600189208984} 02/24/2022 09:54:16 - INFO - codeparrot_training - Step 6438: {'lr': 0.0004895275993075747, 'samples': 3296768, 'steps': 6438, 'loss/train': 2.2389132976531982} 02/24/2022 09:54:20 - INFO - codeparrot_training - Step 6439: {'lr': 0.0004895229125995973, 'samples': 3297280, 'steps': 6439, 'loss/train': 1.4974126815795898} 02/24/2022 09:54:23 - INFO - codeparrot_training - Step 6440: {'lr': 0.0004895182248655798, 'samples': 3297792, 'steps': 6440, 'loss/train': 2.6597371101379395} 02/24/2022 09:54:29 - INFO - codeparrot_training - Step 6441: {'lr': 0.0004895135361055422, 'samples': 3298304, 'steps': 6441, 'loss/train': 2.5425000190734863} 02/24/2022 09:54:32 - INFO - codeparrot_training - Step 6442: {'lr': 0.0004895088463195049, 'samples': 3298816, 'steps': 6442, 'loss/train': 2.8660354614257812} 02/24/2022 09:54:38 - INFO - codeparrot_training - Step 6443: {'lr': 0.0004895041555074875, 'samples': 3299328, 'steps': 6443, 'loss/train': 2.6452040672302246} 02/24/2022 09:54:41 - INFO - codeparrot_training - Step 6444: {'lr': 0.0004894994636695105, 'samples': 3299840, 'steps': 6444, 'loss/train': 2.966923952102661} 02/24/2022 09:54:49 - INFO - codeparrot_training - Step 6445: {'lr': 0.0004894947708055938, 'samples': 3300352, 'steps': 6445, 'loss/train': 2.6917126178741455} 02/24/2022 09:54:54 - INFO - codeparrot_training - Step 6446: {'lr': 0.0004894900769157576, 'samples': 3300864, 'steps': 6446, 'loss/train': 1.5543131828308105} 02/24/2022 09:54:57 - INFO - codeparrot_training - Step 6447: {'lr': 0.0004894853820000219, 'samples': 3301376, 'steps': 6447, 'loss/train': 2.8028717041015625} 02/24/2022 09:55:03 - INFO - codeparrot_training - Step 6448: {'lr': 0.000489480686058407, 'samples': 3301888, 'steps': 6448, 'loss/train': 2.402170419692993} 02/24/2022 09:55:06 - INFO - codeparrot_training - Step 6449: {'lr': 0.0004894759890909326, 'samples': 3302400, 'steps': 6449, 'loss/train': 1.7828624248504639} 02/24/2022 09:55:12 - INFO - codeparrot_training - Step 6450: {'lr': 0.0004894712910976193, 'samples': 3302912, 'steps': 6450, 'loss/train': 2.1325206756591797} 02/24/2022 09:55:15 - INFO - codeparrot_training - Step 6451: {'lr': 0.000489466592078487, 'samples': 3303424, 'steps': 6451, 'loss/train': 0.8790604472160339} 02/24/2022 09:55:21 - INFO - codeparrot_training - Step 6452: {'lr': 0.0004894618920335558, 'samples': 3303936, 'steps': 6452, 'loss/train': 0.5870147347450256} 02/24/2022 09:55:24 - INFO - codeparrot_training - Step 6453: {'lr': 0.000489457190962846, 'samples': 3304448, 'steps': 6453, 'loss/train': 2.9822442531585693} 02/24/2022 09:55:31 - INFO - codeparrot_training - Step 6454: {'lr': 0.0004894524888663776, 'samples': 3304960, 'steps': 6454, 'loss/train': 2.5562589168548584} 02/24/2022 09:55:35 - INFO - codeparrot_training - Step 6455: {'lr': 0.0004894477857441707, 'samples': 3305472, 'steps': 6455, 'loss/train': 2.2890918254852295} 02/24/2022 09:55:40 - INFO - codeparrot_training - Step 6456: {'lr': 0.0004894430815962456, 'samples': 3305984, 'steps': 6456, 'loss/train': 2.3057167530059814} 02/24/2022 09:55:44 - INFO - codeparrot_training - Step 6457: {'lr': 0.0004894383764226224, 'samples': 3306496, 'steps': 6457, 'loss/train': 1.25165855884552} 02/24/2022 09:55:49 - INFO - codeparrot_training - Step 6458: {'lr': 0.0004894336702233212, 'samples': 3307008, 'steps': 6458, 'loss/train': 1.5881156921386719} 02/24/2022 09:55:53 - INFO - codeparrot_training - Step 6459: {'lr': 0.0004894289629983621, 'samples': 3307520, 'steps': 6459, 'loss/train': 2.387951135635376} 02/24/2022 09:55:58 - INFO - codeparrot_training - Step 6460: {'lr': 0.0004894242547477654, 'samples': 3308032, 'steps': 6460, 'loss/train': 3.081812858581543} 02/24/2022 09:56:02 - INFO - codeparrot_training - Step 6461: {'lr': 0.0004894195454715512, 'samples': 3308544, 'steps': 6461, 'loss/train': 2.0218260288238525} 02/24/2022 09:56:10 - INFO - codeparrot_training - Step 6462: {'lr': 0.0004894148351697398, 'samples': 3309056, 'steps': 6462, 'loss/train': 1.4597071409225464} 02/24/2022 09:56:13 - INFO - codeparrot_training - Step 6463: {'lr': 0.0004894101238423512, 'samples': 3309568, 'steps': 6463, 'loss/train': 1.536042332649231} 02/24/2022 09:56:19 - INFO - codeparrot_training - Step 6464: {'lr': 0.0004894054114894055, 'samples': 3310080, 'steps': 6464, 'loss/train': 1.6616467237472534} 02/24/2022 09:56:22 - INFO - codeparrot_training - Step 6465: {'lr': 0.0004894006981109232, 'samples': 3310592, 'steps': 6465, 'loss/train': 2.6865811347961426} 02/24/2022 09:56:28 - INFO - codeparrot_training - Step 6466: {'lr': 0.0004893959837069243, 'samples': 3311104, 'steps': 6466, 'loss/train': 3.2605366706848145} 02/24/2022 09:56:31 - INFO - codeparrot_training - Step 6467: {'lr': 0.0004893912682774291, 'samples': 3311616, 'steps': 6467, 'loss/train': 1.0845344066619873} 02/24/2022 09:56:37 - INFO - codeparrot_training - Step 6468: {'lr': 0.0004893865518224576, 'samples': 3312128, 'steps': 6468, 'loss/train': 2.5008442401885986} 02/24/2022 09:56:40 - INFO - codeparrot_training - Step 6469: {'lr': 0.0004893818343420302, 'samples': 3312640, 'steps': 6469, 'loss/train': 3.2619946002960205} 02/24/2022 09:56:46 - INFO - codeparrot_training - Step 6470: {'lr': 0.000489377115836167, 'samples': 3313152, 'steps': 6470, 'loss/train': 1.6051849126815796} 02/24/2022 09:56:49 - INFO - codeparrot_training - Step 6471: {'lr': 0.0004893723963048882, 'samples': 3313664, 'steps': 6471, 'loss/train': 2.2450814247131348} 02/24/2022 09:56:53 - INFO - codeparrot_training - Step 6472: {'lr': 0.0004893676757482142, 'samples': 3314176, 'steps': 6472, 'loss/train': 2.335479259490967} 02/24/2022 09:57:00 - INFO - codeparrot_training - Step 6473: {'lr': 0.0004893629541661649, 'samples': 3314688, 'steps': 6473, 'loss/train': 1.8403065204620361} 02/24/2022 09:57:04 - INFO - codeparrot_training - Step 6474: {'lr': 0.0004893582315587608, 'samples': 3315200, 'steps': 6474, 'loss/train': 1.6207207441329956} 02/24/2022 09:57:09 - INFO - codeparrot_training - Step 6475: {'lr': 0.0004893535079260221, 'samples': 3315712, 'steps': 6475, 'loss/train': 2.3650307655334473} 02/24/2022 09:57:13 - INFO - codeparrot_training - Step 6476: {'lr': 0.000489348783267969, 'samples': 3316224, 'steps': 6476, 'loss/train': 0.23510637879371643} 02/24/2022 09:57:18 - INFO - codeparrot_training - Step 6477: {'lr': 0.0004893440575846215, 'samples': 3316736, 'steps': 6477, 'loss/train': 1.888055443763733} 02/24/2022 09:57:22 - INFO - codeparrot_training - Step 6478: {'lr': 0.0004893393308760002, 'samples': 3317248, 'steps': 6478, 'loss/train': 2.07065749168396} 02/24/2022 09:57:27 - INFO - codeparrot_training - Step 6479: {'lr': 0.0004893346031421253, 'samples': 3317760, 'steps': 6479, 'loss/train': 2.236597776412964} 02/24/2022 09:57:31 - INFO - codeparrot_training - Step 6480: {'lr': 0.0004893298743830168, 'samples': 3318272, 'steps': 6480, 'loss/train': 2.4800174236297607} 02/24/2022 09:57:36 - INFO - codeparrot_training - Step 6481: {'lr': 0.0004893251445986952, 'samples': 3318784, 'steps': 6481, 'loss/train': 3.0782835483551025} 02/24/2022 09:57:40 - INFO - codeparrot_training - Step 6482: {'lr': 0.0004893204137891807, 'samples': 3319296, 'steps': 6482, 'loss/train': 2.900688409805298} 02/24/2022 09:57:47 - INFO - codeparrot_training - Step 6483: {'lr': 0.0004893156819544935, 'samples': 3319808, 'steps': 6483, 'loss/train': 2.7950785160064697} 02/24/2022 09:57:51 - INFO - codeparrot_training - Step 6484: {'lr': 0.0004893109490946539, 'samples': 3320320, 'steps': 6484, 'loss/train': 1.3772821426391602} 02/24/2022 09:57:56 - INFO - codeparrot_training - Step 6485: {'lr': 0.0004893062152096821, 'samples': 3320832, 'steps': 6485, 'loss/train': 2.1796483993530273} 02/24/2022 09:58:00 - INFO - codeparrot_training - Step 6486: {'lr': 0.0004893014802995985, 'samples': 3321344, 'steps': 6486, 'loss/train': 2.3409743309020996} 02/24/2022 09:58:05 - INFO - codeparrot_training - Step 6487: {'lr': 0.0004892967443644235, 'samples': 3321856, 'steps': 6487, 'loss/train': 2.5255603790283203} 02/24/2022 09:58:08 - INFO - codeparrot_training - Step 6488: {'lr': 0.0004892920074041771, 'samples': 3322368, 'steps': 6488, 'loss/train': 2.121716260910034} 02/24/2022 09:58:14 - INFO - codeparrot_training - Step 6489: {'lr': 0.0004892872694188797, 'samples': 3322880, 'steps': 6489, 'loss/train': 2.704684257507324} 02/24/2022 09:58:18 - INFO - codeparrot_training - Step 6490: {'lr': 0.0004892825304085517, 'samples': 3323392, 'steps': 6490, 'loss/train': 1.4401229619979858} 02/24/2022 09:58:23 - INFO - codeparrot_training - Step 6491: {'lr': 0.0004892777903732133, 'samples': 3323904, 'steps': 6491, 'loss/train': 1.776444911956787} 02/24/2022 09:58:26 - INFO - codeparrot_training - Step 6492: {'lr': 0.0004892730493128848, 'samples': 3324416, 'steps': 6492, 'loss/train': 3.5959036350250244} 02/24/2022 09:58:34 - INFO - codeparrot_training - Step 6493: {'lr': 0.0004892683072275865, 'samples': 3324928, 'steps': 6493, 'loss/train': 2.111992359161377} 02/24/2022 09:58:37 - INFO - codeparrot_training - Step 6494: {'lr': 0.0004892635641173389, 'samples': 3325440, 'steps': 6494, 'loss/train': 3.305833578109741} 02/24/2022 09:58:43 - INFO - codeparrot_training - Step 6495: {'lr': 0.0004892588199821619, 'samples': 3325952, 'steps': 6495, 'loss/train': 2.2491681575775146} 02/24/2022 09:58:46 - INFO - codeparrot_training - Step 6496: {'lr': 0.0004892540748220763, 'samples': 3326464, 'steps': 6496, 'loss/train': 2.2219161987304688} 02/24/2022 09:58:52 - INFO - codeparrot_training - Step 6497: {'lr': 0.0004892493286371022, 'samples': 3326976, 'steps': 6497, 'loss/train': 2.0182578563690186} 02/24/2022 09:58:55 - INFO - codeparrot_training - Step 6498: {'lr': 0.00048924458142726, 'samples': 3327488, 'steps': 6498, 'loss/train': 2.817366361618042} 02/24/2022 09:59:01 - INFO - codeparrot_training - Step 6499: {'lr': 0.0004892398331925698, 'samples': 3328000, 'steps': 6499, 'loss/train': 2.3096659183502197} 02/24/2022 09:59:04 - INFO - codeparrot_training - Step 6500: {'lr': 0.0004892350839330522, 'samples': 3328512, 'steps': 6500, 'loss/train': 2.039881944656372} 02/24/2022 09:59:10 - INFO - codeparrot_training - Step 6501: {'lr': 0.0004892303336487275, 'samples': 3329024, 'steps': 6501, 'loss/train': 1.4357212781906128} 02/24/2022 09:59:13 - INFO - codeparrot_training - Step 6502: {'lr': 0.000489225582339616, 'samples': 3329536, 'steps': 6502, 'loss/train': 2.9906527996063232} 02/24/2022 09:59:19 - INFO - codeparrot_training - Step 6503: {'lr': 0.000489220830005738, 'samples': 3330048, 'steps': 6503, 'loss/train': 1.452683925628662} 02/24/2022 09:59:22 - INFO - codeparrot_training - Step 6504: {'lr': 0.0004892160766471141, 'samples': 3330560, 'steps': 6504, 'loss/train': 1.4622899293899536} 02/24/2022 09:59:29 - INFO - codeparrot_training - Step 6505: {'lr': 0.0004892113222637643, 'samples': 3331072, 'steps': 6505, 'loss/train': 3.0667734146118164} 02/24/2022 09:59:35 - INFO - codeparrot_training - Step 6506: {'lr': 0.0004892065668557093, 'samples': 3331584, 'steps': 6506, 'loss/train': 2.964228630065918} 02/24/2022 09:59:39 - INFO - codeparrot_training - Step 6507: {'lr': 0.0004892018104229692, 'samples': 3332096, 'steps': 6507, 'loss/train': 3.244946241378784} 02/24/2022 09:59:42 - INFO - codeparrot_training - Step 6508: {'lr': 0.0004891970529655646, 'samples': 3332608, 'steps': 6508, 'loss/train': 1.858280897140503} 02/24/2022 09:59:48 - INFO - codeparrot_training - Step 6509: {'lr': 0.0004891922944835158, 'samples': 3333120, 'steps': 6509, 'loss/train': 3.8298871517181396} 02/24/2022 09:59:51 - INFO - codeparrot_training - Step 6510: {'lr': 0.000489187534976843, 'samples': 3333632, 'steps': 6510, 'loss/train': 2.3593363761901855} 02/24/2022 09:59:57 - INFO - codeparrot_training - Step 6511: {'lr': 0.0004891827744455668, 'samples': 3334144, 'steps': 6511, 'loss/train': 3.423246145248413} 02/24/2022 10:00:00 - INFO - codeparrot_training - Step 6512: {'lr': 0.0004891780128897077, 'samples': 3334656, 'steps': 6512, 'loss/train': 2.655895471572876} 02/24/2022 10:00:06 - INFO - codeparrot_training - Step 6513: {'lr': 0.0004891732503092858, 'samples': 3335168, 'steps': 6513, 'loss/train': 2.022547721862793} 02/24/2022 10:00:11 - INFO - codeparrot_training - Step 6514: {'lr': 0.0004891684867043216, 'samples': 3335680, 'steps': 6514, 'loss/train': 2.3471121788024902} 02/24/2022 10:00:15 - INFO - codeparrot_training - Step 6515: {'lr': 0.0004891637220748356, 'samples': 3336192, 'steps': 6515, 'loss/train': 1.8252805471420288} 02/24/2022 10:00:20 - INFO - codeparrot_training - Step 6516: {'lr': 0.0004891589564208482, 'samples': 3336704, 'steps': 6516, 'loss/train': 0.5026594400405884} 02/24/2022 10:00:23 - INFO - codeparrot_training - Step 6517: {'lr': 0.0004891541897423798, 'samples': 3337216, 'steps': 6517, 'loss/train': 1.8229773044586182} 02/24/2022 10:00:31 - INFO - codeparrot_training - Step 6518: {'lr': 0.0004891494220394507, 'samples': 3337728, 'steps': 6518, 'loss/train': 3.6086223125457764} 02/24/2022 10:00:34 - INFO - codeparrot_training - Step 6519: {'lr': 0.0004891446533120815, 'samples': 3338240, 'steps': 6519, 'loss/train': 2.2332448959350586} 02/24/2022 10:00:40 - INFO - codeparrot_training - Step 6520: {'lr': 0.0004891398835602925, 'samples': 3338752, 'steps': 6520, 'loss/train': 3.0605921745300293} 02/24/2022 10:00:44 - INFO - codeparrot_training - Step 6521: {'lr': 0.0004891351127841041, 'samples': 3339264, 'steps': 6521, 'loss/train': 5.062290668487549} 02/24/2022 10:00:47 - INFO - codeparrot_training - Step 6522: {'lr': 0.0004891303409835369, 'samples': 3339776, 'steps': 6522, 'loss/train': 2.3840460777282715} 02/24/2022 10:00:53 - INFO - codeparrot_training - Step 6523: {'lr': 0.0004891255681586113, 'samples': 3340288, 'steps': 6523, 'loss/train': 1.9411540031433105} 02/24/2022 10:00:56 - INFO - codeparrot_training - Step 6524: {'lr': 0.0004891207943093476, 'samples': 3340800, 'steps': 6524, 'loss/train': 2.0542826652526855} 02/24/2022 10:01:01 - INFO - codeparrot_training - Step 6525: {'lr': 0.0004891160194357663, 'samples': 3341312, 'steps': 6525, 'loss/train': 2.3490045070648193} 02/24/2022 10:01:05 - INFO - codeparrot_training - Step 6526: {'lr': 0.0004891112435378881, 'samples': 3341824, 'steps': 6526, 'loss/train': 2.155965566635132} 02/24/2022 10:01:11 - INFO - codeparrot_training - Step 6527: {'lr': 0.0004891064666157331, 'samples': 3342336, 'steps': 6527, 'loss/train': 2.277674674987793} 02/24/2022 10:01:14 - INFO - codeparrot_training - Step 6528: {'lr': 0.0004891016886693219, 'samples': 3342848, 'steps': 6528, 'loss/train': 1.6484726667404175} 02/24/2022 10:01:21 - INFO - codeparrot_training - Step 6529: {'lr': 0.0004890969096986751, 'samples': 3343360, 'steps': 6529, 'loss/train': 1.8307876586914062} 02/24/2022 10:01:27 - INFO - codeparrot_training - Step 6530: {'lr': 0.000489092129703813, 'samples': 3343872, 'steps': 6530, 'loss/train': 3.480984687805176} 02/24/2022 10:01:30 - INFO - codeparrot_training - Step 6531: {'lr': 0.0004890873486847561, 'samples': 3344384, 'steps': 6531, 'loss/train': 1.3929086923599243} 02/24/2022 10:01:36 - INFO - codeparrot_training - Step 6532: {'lr': 0.000489082566641525, 'samples': 3344896, 'steps': 6532, 'loss/train': 2.7073380947113037} 02/24/2022 10:01:39 - INFO - codeparrot_training - Step 6533: {'lr': 0.00048907778357414, 'samples': 3345408, 'steps': 6533, 'loss/train': 2.1377365589141846} 02/24/2022 10:01:45 - INFO - codeparrot_training - Step 6534: {'lr': 0.0004890729994826218, 'samples': 3345920, 'steps': 6534, 'loss/train': 3.1660683155059814} 02/24/2022 10:01:48 - INFO - codeparrot_training - Step 6535: {'lr': 0.0004890682143669908, 'samples': 3346432, 'steps': 6535, 'loss/train': 3.0280263423919678} 02/24/2022 10:01:52 - INFO - codeparrot_training - Step 6536: {'lr': 0.0004890634282272673, 'samples': 3346944, 'steps': 6536, 'loss/train': 0.28646907210350037} 02/24/2022 10:01:57 - INFO - codeparrot_training - Step 6537: {'lr': 0.0004890586410634722, 'samples': 3347456, 'steps': 6537, 'loss/train': 3.232842445373535} 02/24/2022 10:02:01 - INFO - codeparrot_training - Step 6538: {'lr': 0.0004890538528756256, 'samples': 3347968, 'steps': 6538, 'loss/train': 1.879063606262207} 02/24/2022 10:02:08 - INFO - codeparrot_training - Step 6539: {'lr': 0.0004890490636637484, 'samples': 3348480, 'steps': 6539, 'loss/train': 2.724001884460449} 02/24/2022 10:02:14 - INFO - codeparrot_training - Step 6540: {'lr': 0.0004890442734278608, 'samples': 3348992, 'steps': 6540, 'loss/train': 1.7801892757415771} 02/24/2022 10:02:18 - INFO - codeparrot_training - Step 6541: {'lr': 0.0004890394821679834, 'samples': 3349504, 'steps': 6541, 'loss/train': 1.900373935699463} 02/24/2022 10:02:23 - INFO - codeparrot_training - Step 6542: {'lr': 0.0004890346898841369, 'samples': 3350016, 'steps': 6542, 'loss/train': 2.83417010307312} 02/24/2022 10:02:27 - INFO - codeparrot_training - Step 6543: {'lr': 0.0004890298965763416, 'samples': 3350528, 'steps': 6543, 'loss/train': 0.7365725040435791} 02/24/2022 10:02:32 - INFO - codeparrot_training - Step 6544: {'lr': 0.0004890251022446181, 'samples': 3351040, 'steps': 6544, 'loss/train': 2.0251264572143555} 02/24/2022 10:02:36 - INFO - codeparrot_training - Step 6545: {'lr': 0.000489020306888987, 'samples': 3351552, 'steps': 6545, 'loss/train': 3.1508102416992188} 02/24/2022 10:02:41 - INFO - codeparrot_training - Step 6546: {'lr': 0.0004890155105094688, 'samples': 3352064, 'steps': 6546, 'loss/train': 1.8642537593841553} 02/24/2022 10:02:45 - INFO - codeparrot_training - Step 6547: {'lr': 0.0004890107131060841, 'samples': 3352576, 'steps': 6547, 'loss/train': 1.9416359663009644} 02/24/2022 10:02:52 - INFO - codeparrot_training - Step 6548: {'lr': 0.0004890059146788532, 'samples': 3353088, 'steps': 6548, 'loss/train': 2.778923749923706} 02/24/2022 10:02:55 - INFO - codeparrot_training - Step 6549: {'lr': 0.000489001115227797, 'samples': 3353600, 'steps': 6549, 'loss/train': 1.9057215452194214} 02/24/2022 10:03:01 - INFO - codeparrot_training - Step 6550: {'lr': 0.000488996314752936, 'samples': 3354112, 'steps': 6550, 'loss/train': 3.6777164936065674} 02/24/2022 10:03:04 - INFO - codeparrot_training - Step 6551: {'lr': 0.0004889915132542906, 'samples': 3354624, 'steps': 6551, 'loss/train': 0.23695413768291473} 02/24/2022 10:03:10 - INFO - codeparrot_training - Step 6552: {'lr': 0.0004889867107318814, 'samples': 3355136, 'steps': 6552, 'loss/train': 2.2470178604125977} 02/24/2022 10:03:13 - INFO - codeparrot_training - Step 6553: {'lr': 0.0004889819071857291, 'samples': 3355648, 'steps': 6553, 'loss/train': 1.051946997642517} 02/24/2022 10:03:19 - INFO - codeparrot_training - Step 6554: {'lr': 0.0004889771026158541, 'samples': 3356160, 'steps': 6554, 'loss/train': 2.677356243133545} 02/24/2022 10:03:22 - INFO - codeparrot_training - Step 6555: {'lr': 0.0004889722970222772, 'samples': 3356672, 'steps': 6555, 'loss/train': 2.6603970527648926} 02/24/2022 10:03:28 - INFO - codeparrot_training - Step 6556: {'lr': 0.0004889674904050188, 'samples': 3357184, 'steps': 6556, 'loss/train': 2.640587329864502} 02/24/2022 10:03:31 - INFO - codeparrot_training - Step 6557: {'lr': 0.0004889626827640994, 'samples': 3357696, 'steps': 6557, 'loss/train': 2.5008530616760254} 02/24/2022 10:03:37 - INFO - codeparrot_training - Step 6558: {'lr': 0.00048895787409954, 'samples': 3358208, 'steps': 6558, 'loss/train': 2.5547757148742676} 02/24/2022 10:03:41 - INFO - codeparrot_training - Step 6559: {'lr': 0.0004889530644113608, 'samples': 3358720, 'steps': 6559, 'loss/train': 2.7593259811401367} 02/24/2022 10:03:46 - INFO - codeparrot_training - Step 6560: {'lr': 0.0004889482536995825, 'samples': 3359232, 'steps': 6560, 'loss/train': 2.433361530303955} 02/24/2022 10:03:50 - INFO - codeparrot_training - Step 6561: {'lr': 0.0004889434419642259, 'samples': 3359744, 'steps': 6561, 'loss/train': 3.0989346504211426} 02/24/2022 10:03:55 - INFO - codeparrot_training - Step 6562: {'lr': 0.0004889386292053114, 'samples': 3360256, 'steps': 6562, 'loss/train': 0.5503135919570923} 02/24/2022 10:03:59 - INFO - codeparrot_training - Step 6563: {'lr': 0.0004889338154228596, 'samples': 3360768, 'steps': 6563, 'loss/train': 3.1529862880706787} 02/24/2022 10:04:06 - INFO - codeparrot_training - Step 6564: {'lr': 0.0004889290006168913, 'samples': 3361280, 'steps': 6564, 'loss/train': 2.479491710662842} 02/24/2022 10:04:09 - INFO - codeparrot_training - Step 6565: {'lr': 0.0004889241847874271, 'samples': 3361792, 'steps': 6565, 'loss/train': 1.8109992742538452} 02/24/2022 10:04:15 - INFO - codeparrot_training - Step 6566: {'lr': 0.0004889193679344874, 'samples': 3362304, 'steps': 6566, 'loss/train': 2.2763566970825195} 02/24/2022 10:04:18 - INFO - codeparrot_training - Step 6567: {'lr': 0.0004889145500580932, 'samples': 3362816, 'steps': 6567, 'loss/train': 2.791712522506714} 02/24/2022 10:04:22 - INFO - codeparrot_training - Step 6568: {'lr': 0.0004889097311582647, 'samples': 3363328, 'steps': 6568, 'loss/train': 1.8300089836120605} 02/24/2022 10:04:27 - INFO - codeparrot_training - Step 6569: {'lr': 0.000488904911235023, 'samples': 3363840, 'steps': 6569, 'loss/train': 2.486496686935425} 02/24/2022 10:04:33 - INFO - codeparrot_training - Step 6570: {'lr': 0.0004889000902883883, 'samples': 3364352, 'steps': 6570, 'loss/train': 2.468989133834839} 02/24/2022 10:04:36 - INFO - codeparrot_training - Step 6571: {'lr': 0.0004888952683183816, 'samples': 3364864, 'steps': 6571, 'loss/train': 2.3142385482788086} 02/24/2022 10:04:42 - INFO - codeparrot_training - Step 6572: {'lr': 0.0004888904453250233, 'samples': 3365376, 'steps': 6572, 'loss/train': 3.6092429161071777} 02/24/2022 10:04:45 - INFO - codeparrot_training - Step 6573: {'lr': 0.0004888856213083343, 'samples': 3365888, 'steps': 6573, 'loss/train': 2.0679149627685547} 02/24/2022 10:04:53 - INFO - codeparrot_training - Step 6574: {'lr': 0.0004888807962683353, 'samples': 3366400, 'steps': 6574, 'loss/train': 2.371832847595215} 02/24/2022 10:04:56 - INFO - codeparrot_training - Step 6575: {'lr': 0.0004888759702050466, 'samples': 3366912, 'steps': 6575, 'loss/train': 2.7449398040771484} 02/24/2022 10:05:02 - INFO - codeparrot_training - Step 6576: {'lr': 0.0004888711431184892, 'samples': 3367424, 'steps': 6576, 'loss/train': 1.9852746725082397} 02/24/2022 10:05:05 - INFO - codeparrot_training - Step 6577: {'lr': 0.0004888663150086835, 'samples': 3367936, 'steps': 6577, 'loss/train': 1.6808832883834839} 02/24/2022 10:05:09 - INFO - codeparrot_training - Step 6578: {'lr': 0.0004888614858756505, 'samples': 3368448, 'steps': 6578, 'loss/train': 3.339327096939087} 02/24/2022 10:05:15 - INFO - codeparrot_training - Step 6579: {'lr': 0.0004888566557194107, 'samples': 3368960, 'steps': 6579, 'loss/train': 2.0961108207702637} 02/24/2022 10:05:20 - INFO - codeparrot_training - Step 6580: {'lr': 0.0004888518245399849, 'samples': 3369472, 'steps': 6580, 'loss/train': 1.8694839477539062} 02/24/2022 10:05:24 - INFO - codeparrot_training - Step 6581: {'lr': 0.0004888469923373937, 'samples': 3369984, 'steps': 6581, 'loss/train': 2.3084683418273926} 02/24/2022 10:05:29 - INFO - codeparrot_training - Step 6582: {'lr': 0.0004888421591116578, 'samples': 3370496, 'steps': 6582, 'loss/train': 1.8113888502120972} 02/24/2022 10:05:33 - INFO - codeparrot_training - Step 6583: {'lr': 0.000488837324862798, 'samples': 3371008, 'steps': 6583, 'loss/train': 2.4665839672088623} 02/24/2022 10:05:40 - INFO - codeparrot_training - Step 6584: {'lr': 0.0004888324895908349, 'samples': 3371520, 'steps': 6584, 'loss/train': 3.1531171798706055} 02/24/2022 10:05:43 - INFO - codeparrot_training - Step 6585: {'lr': 0.0004888276532957892, 'samples': 3372032, 'steps': 6585, 'loss/train': 2.2655367851257324} 02/24/2022 10:05:49 - INFO - codeparrot_training - Step 6586: {'lr': 0.0004888228159776818, 'samples': 3372544, 'steps': 6586, 'loss/train': 1.1006475687026978} 02/24/2022 10:05:52 - INFO - codeparrot_training - Step 6587: {'lr': 0.0004888179776365331, 'samples': 3373056, 'steps': 6587, 'loss/train': 2.708115816116333} 02/24/2022 10:05:58 - INFO - codeparrot_training - Step 6588: {'lr': 0.0004888131382723641, 'samples': 3373568, 'steps': 6588, 'loss/train': 2.7441513538360596} 02/24/2022 10:06:01 - INFO - codeparrot_training - Step 6589: {'lr': 0.0004888082978851954, 'samples': 3374080, 'steps': 6589, 'loss/train': 7.269293785095215} 02/24/2022 10:06:07 - INFO - codeparrot_training - Step 6590: {'lr': 0.000488803456475048, 'samples': 3374592, 'steps': 6590, 'loss/train': 2.0342698097229004} 02/24/2022 10:06:10 - INFO - codeparrot_training - Step 6591: {'lr': 0.0004887986140419422, 'samples': 3375104, 'steps': 6591, 'loss/train': 3.4777088165283203} 02/24/2022 10:06:16 - INFO - codeparrot_training - Step 6592: {'lr': 0.000488793770585899, 'samples': 3375616, 'steps': 6592, 'loss/train': 2.5299012660980225} 02/24/2022 10:06:19 - INFO - codeparrot_training - Step 6593: {'lr': 0.0004887889261069392, 'samples': 3376128, 'steps': 6593, 'loss/train': 0.4990481436252594} 02/24/2022 10:06:25 - INFO - codeparrot_training - Step 6594: {'lr': 0.0004887840806050834, 'samples': 3376640, 'steps': 6594, 'loss/train': 2.1829071044921875} 02/24/2022 10:06:28 - INFO - codeparrot_training - Step 6595: {'lr': 0.0004887792340803524, 'samples': 3377152, 'steps': 6595, 'loss/train': 2.905155658721924} 02/24/2022 10:06:35 - INFO - codeparrot_training - Step 6596: {'lr': 0.000488774386532767, 'samples': 3377664, 'steps': 6596, 'loss/train': 1.8029149770736694} 02/24/2022 10:06:39 - INFO - codeparrot_training - Step 6597: {'lr': 0.0004887695379623481, 'samples': 3378176, 'steps': 6597, 'loss/train': 2.8679449558258057} 02/24/2022 10:06:45 - INFO - codeparrot_training - Step 6598: {'lr': 0.000488764688369116, 'samples': 3378688, 'steps': 6598, 'loss/train': 2.7318615913391113} 02/24/2022 10:06:48 - INFO - codeparrot_training - Step 6599: {'lr': 0.000488759837753092, 'samples': 3379200, 'steps': 6599, 'loss/train': 2.086902141571045} 02/24/2022 10:06:54 - INFO - codeparrot_training - Step 6600: {'lr': 0.0004887549861142967, 'samples': 3379712, 'steps': 6600, 'loss/train': 2.0606794357299805} 02/24/2022 10:06:57 - INFO - codeparrot_training - Step 6601: {'lr': 0.0004887501334527507, 'samples': 3380224, 'steps': 6601, 'loss/train': 2.4570682048797607} 02/24/2022 10:07:03 - INFO - codeparrot_training - Step 6602: {'lr': 0.000488745279768475, 'samples': 3380736, 'steps': 6602, 'loss/train': 2.378192186355591} 02/24/2022 10:07:06 - INFO - codeparrot_training - Step 6603: {'lr': 0.0004887404250614904, 'samples': 3381248, 'steps': 6603, 'loss/train': 2.676232099533081} 02/24/2022 10:07:10 - INFO - codeparrot_training - Step 6604: {'lr': 0.0004887355693318176, 'samples': 3381760, 'steps': 6604, 'loss/train': 1.8846713304519653} 02/24/2022 10:07:15 - INFO - codeparrot_training - Step 6605: {'lr': 0.0004887307125794775, 'samples': 3382272, 'steps': 6605, 'loss/train': 2.3327300548553467} 02/24/2022 10:07:21 - INFO - codeparrot_training - Step 6606: {'lr': 0.0004887258548044907, 'samples': 3382784, 'steps': 6606, 'loss/train': 2.2616655826568604} 02/24/2022 10:07:24 - INFO - codeparrot_training - Step 6607: {'lr': 0.0004887209960068782, 'samples': 3383296, 'steps': 6607, 'loss/train': 2.124143123626709} 02/24/2022 10:07:30 - INFO - codeparrot_training - Step 6608: {'lr': 0.0004887161361866607, 'samples': 3383808, 'steps': 6608, 'loss/train': 2.479696035385132} 02/24/2022 10:07:34 - INFO - codeparrot_training - Step 6609: {'lr': 0.0004887112753438592, 'samples': 3384320, 'steps': 6609, 'loss/train': 1.5773290395736694} 02/24/2022 10:07:41 - INFO - codeparrot_training - Step 6610: {'lr': 0.0004887064134784943, 'samples': 3384832, 'steps': 6610, 'loss/train': 1.6292513608932495} 02/24/2022 10:07:44 - INFO - codeparrot_training - Step 6611: {'lr': 0.0004887015505905869, 'samples': 3385344, 'steps': 6611, 'loss/train': 2.8065037727355957} 02/24/2022 10:07:50 - INFO - codeparrot_training - Step 6612: {'lr': 0.0004886966866801579, 'samples': 3385856, 'steps': 6612, 'loss/train': 1.255644679069519} 02/24/2022 10:07:53 - INFO - codeparrot_training - Step 6613: {'lr': 0.0004886918217472281, 'samples': 3386368, 'steps': 6613, 'loss/train': 2.8977606296539307} 02/24/2022 10:07:59 - INFO - codeparrot_training - Step 6614: {'lr': 0.0004886869557918183, 'samples': 3386880, 'steps': 6614, 'loss/train': 2.5531868934631348} 02/24/2022 10:08:02 - INFO - codeparrot_training - Step 6615: {'lr': 0.0004886820888139494, 'samples': 3387392, 'steps': 6615, 'loss/train': 2.9990012645721436} 02/24/2022 10:08:08 - INFO - codeparrot_training - Step 6616: {'lr': 0.0004886772208136422, 'samples': 3387904, 'steps': 6616, 'loss/train': 2.3838999271392822} 02/24/2022 10:08:11 - INFO - codeparrot_training - Step 6617: {'lr': 0.0004886723517909176, 'samples': 3388416, 'steps': 6617, 'loss/train': 1.844799518585205} 02/24/2022 10:08:17 - INFO - codeparrot_training - Step 6618: {'lr': 0.0004886674817457964, 'samples': 3388928, 'steps': 6618, 'loss/train': 2.758025884628296} 02/24/2022 10:08:20 - INFO - codeparrot_training - Step 6619: {'lr': 0.0004886626106782995, 'samples': 3389440, 'steps': 6619, 'loss/train': 3.207862377166748} 02/24/2022 10:08:27 - INFO - codeparrot_training - Step 6620: {'lr': 0.0004886577385884478, 'samples': 3389952, 'steps': 6620, 'loss/train': 2.013533592224121} 02/24/2022 10:08:31 - INFO - codeparrot_training - Step 6621: {'lr': 0.0004886528654762621, 'samples': 3390464, 'steps': 6621, 'loss/train': 2.4255099296569824} 02/24/2022 10:08:37 - INFO - codeparrot_training - Step 6622: {'lr': 0.0004886479913417633, 'samples': 3390976, 'steps': 6622, 'loss/train': 2.250338077545166} 02/24/2022 10:08:40 - INFO - codeparrot_training - Step 6623: {'lr': 0.0004886431161849722, 'samples': 3391488, 'steps': 6623, 'loss/train': 2.052408456802368} 02/24/2022 10:08:46 - INFO - codeparrot_training - Step 6624: {'lr': 0.0004886382400059099, 'samples': 3392000, 'steps': 6624, 'loss/train': 2.5310869216918945} 02/24/2022 10:08:49 - INFO - codeparrot_training - Step 6625: {'lr': 0.0004886333628045972, 'samples': 3392512, 'steps': 6625, 'loss/train': 2.2756600379943848} 02/24/2022 10:08:55 - INFO - codeparrot_training - Step 6626: {'lr': 0.0004886284845810548, 'samples': 3393024, 'steps': 6626, 'loss/train': 2.1461868286132812} 02/24/2022 10:08:58 - INFO - codeparrot_training - Step 6627: {'lr': 0.0004886236053353038, 'samples': 3393536, 'steps': 6627, 'loss/train': 1.1681115627288818} 02/24/2022 10:09:04 - INFO - codeparrot_training - Step 6628: {'lr': 0.000488618725067365, 'samples': 3394048, 'steps': 6628, 'loss/train': 3.136591672897339} 02/24/2022 10:09:07 - INFO - codeparrot_training - Step 6629: {'lr': 0.0004886138437772594, 'samples': 3394560, 'steps': 6629, 'loss/train': 1.9597866535186768} 02/24/2022 10:09:13 - INFO - codeparrot_training - Step 6630: {'lr': 0.0004886089614650078, 'samples': 3395072, 'steps': 6630, 'loss/train': 2.4114599227905273} 02/24/2022 10:09:16 - INFO - codeparrot_training - Step 6631: {'lr': 0.0004886040781306313, 'samples': 3395584, 'steps': 6631, 'loss/train': 5.021580696105957} 02/24/2022 10:09:23 - INFO - codeparrot_training - Step 6632: {'lr': 0.0004885991937741506, 'samples': 3396096, 'steps': 6632, 'loss/train': 3.04402756690979} 02/24/2022 10:09:26 - INFO - codeparrot_training - Step 6633: {'lr': 0.0004885943083955868, 'samples': 3396608, 'steps': 6633, 'loss/train': 2.7292466163635254} 02/24/2022 10:09:32 - INFO - codeparrot_training - Step 6634: {'lr': 0.0004885894219949607, 'samples': 3397120, 'steps': 6634, 'loss/train': 2.0121607780456543} 02/24/2022 10:09:35 - INFO - codeparrot_training - Step 6635: {'lr': 0.0004885845345722932, 'samples': 3397632, 'steps': 6635, 'loss/train': 1.775314450263977} 02/24/2022 10:09:41 - INFO - codeparrot_training - Step 6636: {'lr': 0.0004885796461276055, 'samples': 3398144, 'steps': 6636, 'loss/train': 2.724120855331421} 02/24/2022 10:09:44 - INFO - codeparrot_training - Step 6637: {'lr': 0.0004885747566609182, 'samples': 3398656, 'steps': 6637, 'loss/train': 2.1954236030578613} 02/24/2022 10:09:50 - INFO - codeparrot_training - Step 6638: {'lr': 0.0004885698661722524, 'samples': 3399168, 'steps': 6638, 'loss/train': 2.8365275859832764} 02/24/2022 10:09:53 - INFO - codeparrot_training - Step 6639: {'lr': 0.0004885649746616291, 'samples': 3399680, 'steps': 6639, 'loss/train': 3.1536190509796143} 02/24/2022 10:09:59 - INFO - codeparrot_training - Step 6640: {'lr': 0.0004885600821290692, 'samples': 3400192, 'steps': 6640, 'loss/train': 3.166158437728882} 02/24/2022 10:10:02 - INFO - codeparrot_training - Step 6641: {'lr': 0.0004885551885745937, 'samples': 3400704, 'steps': 6641, 'loss/train': 2.1950392723083496} 02/24/2022 10:10:09 - INFO - codeparrot_training - Step 6642: {'lr': 0.0004885502939982235, 'samples': 3401216, 'steps': 6642, 'loss/train': 2.4782874584198} 02/24/2022 10:10:13 - INFO - codeparrot_training - Step 6643: {'lr': 0.0004885453983999795, 'samples': 3401728, 'steps': 6643, 'loss/train': 1.5839953422546387} 02/24/2022 10:10:18 - INFO - codeparrot_training - Step 6644: {'lr': 0.0004885405017798828, 'samples': 3402240, 'steps': 6644, 'loss/train': 1.575046181678772} 02/24/2022 10:10:21 - INFO - codeparrot_training - Step 6645: {'lr': 0.0004885356041379544, 'samples': 3402752, 'steps': 6645, 'loss/train': 2.4168448448181152} 02/24/2022 10:10:27 - INFO - codeparrot_training - Step 6646: {'lr': 0.0004885307054742151, 'samples': 3403264, 'steps': 6646, 'loss/train': 1.755401611328125} 02/24/2022 10:10:30 - INFO - codeparrot_training - Step 6647: {'lr': 0.0004885258057886861, 'samples': 3403776, 'steps': 6647, 'loss/train': 1.2887043952941895} 02/24/2022 10:10:36 - INFO - codeparrot_training - Step 6648: {'lr': 0.0004885209050813882, 'samples': 3404288, 'steps': 6648, 'loss/train': 1.9484624862670898} 02/24/2022 10:10:39 - INFO - codeparrot_training - Step 6649: {'lr': 0.0004885160033523426, 'samples': 3404800, 'steps': 6649, 'loss/train': 1.7089738845825195} 02/24/2022 10:10:45 - INFO - codeparrot_training - Step 6650: {'lr': 0.0004885111006015701, 'samples': 3405312, 'steps': 6650, 'loss/train': 1.8647420406341553} 02/24/2022 10:10:48 - INFO - codeparrot_training - Step 6651: {'lr': 0.0004885061968290919, 'samples': 3405824, 'steps': 6651, 'loss/train': 1.6634795665740967} 02/24/2022 10:10:54 - INFO - codeparrot_training - Step 6652: {'lr': 0.0004885012920349287, 'samples': 3406336, 'steps': 6652, 'loss/train': 1.3363686800003052} 02/24/2022 10:10:58 - INFO - codeparrot_training - Step 6653: {'lr': 0.0004884963862191018, 'samples': 3406848, 'steps': 6653, 'loss/train': 0.18189160525798798} 02/24/2022 10:11:03 - INFO - codeparrot_training - Step 6654: {'lr': 0.0004884914793816321, 'samples': 3407360, 'steps': 6654, 'loss/train': 3.6349596977233887} 02/24/2022 10:11:07 - INFO - codeparrot_training - Step 6655: {'lr': 0.0004884865715225407, 'samples': 3407872, 'steps': 6655, 'loss/train': 1.5691801309585571} 02/24/2022 10:11:13 - INFO - codeparrot_training - Step 6656: {'lr': 0.0004884816626418484, 'samples': 3408384, 'steps': 6656, 'loss/train': 3.0810370445251465} 02/24/2022 10:11:17 - INFO - codeparrot_training - Step 6657: {'lr': 0.0004884767527395765, 'samples': 3408896, 'steps': 6657, 'loss/train': 3.0394320487976074} 02/24/2022 10:11:22 - INFO - codeparrot_training - Step 6658: {'lr': 0.0004884718418157459, 'samples': 3409408, 'steps': 6658, 'loss/train': 2.349281072616577} 02/24/2022 10:11:28 - INFO - codeparrot_training - Step 6659: {'lr': 0.0004884669298703775, 'samples': 3409920, 'steps': 6659, 'loss/train': 1.5560840368270874} 02/24/2022 10:11:32 - INFO - codeparrot_training - Step 6660: {'lr': 0.0004884620169034927, 'samples': 3410432, 'steps': 6660, 'loss/train': 2.6041526794433594} 02/24/2022 10:11:35 - INFO - codeparrot_training - Step 6661: {'lr': 0.0004884571029151123, 'samples': 3410944, 'steps': 6661, 'loss/train': 3.0832126140594482} 02/24/2022 10:11:41 - INFO - codeparrot_training - Step 6662: {'lr': 0.0004884521879052573, 'samples': 3411456, 'steps': 6662, 'loss/train': 2.4900426864624023} 02/24/2022 10:11:44 - INFO - codeparrot_training - Step 6663: {'lr': 0.000488447271873949, 'samples': 3411968, 'steps': 6663, 'loss/train': 3.26607084274292} 02/24/2022 10:11:50 - INFO - codeparrot_training - Step 6664: {'lr': 0.0004884423548212082, 'samples': 3412480, 'steps': 6664, 'loss/train': 3.2601382732391357} 02/24/2022 10:11:53 - INFO - codeparrot_training - Step 6665: {'lr': 0.000488437436747056, 'samples': 3412992, 'steps': 6665, 'loss/train': 0.47087666392326355} 02/24/2022 10:11:59 - INFO - codeparrot_training - Step 6666: {'lr': 0.0004884325176515137, 'samples': 3413504, 'steps': 6666, 'loss/train': 3.041734218597412} 02/24/2022 10:12:02 - INFO - codeparrot_training - Step 6667: {'lr': 0.000488427597534602, 'samples': 3414016, 'steps': 6667, 'loss/train': 1.6828534603118896} 02/24/2022 10:12:09 - INFO - codeparrot_training - Step 6668: {'lr': 0.0004884226763963423, 'samples': 3414528, 'steps': 6668, 'loss/train': 2.149714946746826} 02/24/2022 10:12:12 - INFO - codeparrot_training - Step 6669: {'lr': 0.0004884177542367556, 'samples': 3415040, 'steps': 6669, 'loss/train': 1.6940264701843262} 02/24/2022 10:12:18 - INFO - codeparrot_training - Step 6670: {'lr': 0.0004884128310558628, 'samples': 3415552, 'steps': 6670, 'loss/train': 1.0381897687911987} 02/24/2022 10:12:21 - INFO - codeparrot_training - Step 6671: {'lr': 0.0004884079068536853, 'samples': 3416064, 'steps': 6671, 'loss/train': 2.336825132369995} 02/24/2022 10:12:27 - INFO - codeparrot_training - Step 6672: {'lr': 0.000488402981630244, 'samples': 3416576, 'steps': 6672, 'loss/train': 2.4532577991485596} 02/24/2022 10:12:30 - INFO - codeparrot_training - Step 6673: {'lr': 0.00048839805538556, 'samples': 3417088, 'steps': 6673, 'loss/train': 2.9381401538848877} 02/24/2022 10:12:36 - INFO - codeparrot_training - Step 6674: {'lr': 0.0004883931281196544, 'samples': 3417600, 'steps': 6674, 'loss/train': 2.611172914505005} 02/24/2022 10:12:39 - INFO - codeparrot_training - Step 6675: {'lr': 0.0004883881998325484, 'samples': 3418112, 'steps': 6675, 'loss/train': 3.408902883529663} 02/24/2022 10:12:45 - INFO - codeparrot_training - Step 6676: {'lr': 0.000488383270524263, 'samples': 3418624, 'steps': 6676, 'loss/train': 1.6392360925674438} 02/24/2022 10:12:49 - INFO - codeparrot_training - Step 6677: {'lr': 0.0004883783401948194, 'samples': 3419136, 'steps': 6677, 'loss/train': 1.9712547063827515} 02/24/2022 10:12:54 - INFO - codeparrot_training - Step 6678: {'lr': 0.0004883734088442387, 'samples': 3419648, 'steps': 6678, 'loss/train': 2.7229816913604736} 02/24/2022 10:12:58 - INFO - codeparrot_training - Step 6679: {'lr': 0.0004883684764725419, 'samples': 3420160, 'steps': 6679, 'loss/train': 2.0310049057006836} 02/24/2022 10:13:04 - INFO - codeparrot_training - Step 6680: {'lr': 0.0004883635430797502, 'samples': 3420672, 'steps': 6680, 'loss/train': 2.5254099369049072} 02/24/2022 10:13:07 - INFO - codeparrot_training - Step 6681: {'lr': 0.000488358608665885, 'samples': 3421184, 'steps': 6681, 'loss/train': 2.7194700241088867} 02/24/2022 10:13:13 - INFO - codeparrot_training - Step 6682: {'lr': 0.000488353673230967, 'samples': 3421696, 'steps': 6682, 'loss/train': 2.399972677230835} 02/24/2022 10:13:16 - INFO - codeparrot_training - Step 6683: {'lr': 0.0004883487367750177, 'samples': 3422208, 'steps': 6683, 'loss/train': 2.159799098968506} 02/24/2022 10:13:22 - INFO - codeparrot_training - Step 6684: {'lr': 0.0004883437992980581, 'samples': 3422720, 'steps': 6684, 'loss/train': 1.2849992513656616} 02/24/2022 10:13:25 - INFO - codeparrot_training - Step 6685: {'lr': 0.0004883388608001093, 'samples': 3423232, 'steps': 6685, 'loss/train': 2.4541096687316895} 02/24/2022 10:13:31 - INFO - codeparrot_training - Step 6686: {'lr': 0.0004883339212811924, 'samples': 3423744, 'steps': 6686, 'loss/train': 1.4725836515426636} 02/24/2022 10:13:35 - INFO - codeparrot_training - Step 6687: {'lr': 0.0004883289807413288, 'samples': 3424256, 'steps': 6687, 'loss/train': 2.3076071739196777} 02/24/2022 10:13:40 - INFO - codeparrot_training - Step 6688: {'lr': 0.0004883240391805394, 'samples': 3424768, 'steps': 6688, 'loss/train': 3.1735002994537354} 02/24/2022 10:13:44 - INFO - codeparrot_training - Step 6689: {'lr': 0.0004883190965988455, 'samples': 3425280, 'steps': 6689, 'loss/train': 2.6082816123962402} 02/24/2022 10:13:49 - INFO - codeparrot_training - Step 6690: {'lr': 0.0004883141529962683, 'samples': 3425792, 'steps': 6690, 'loss/train': 2.8360228538513184} 02/24/2022 10:13:52 - INFO - codeparrot_training - Step 6691: {'lr': 0.000488309208372829, 'samples': 3426304, 'steps': 6691, 'loss/train': 2.1867799758911133} 02/24/2022 10:13:58 - INFO - codeparrot_training - Step 6692: {'lr': 0.0004883042627285488, 'samples': 3426816, 'steps': 6692, 'loss/train': 0.40237507224082947} 02/24/2022 10:14:02 - INFO - codeparrot_training - Step 6693: {'lr': 0.0004882993160634487, 'samples': 3427328, 'steps': 6693, 'loss/train': 1.9413496255874634} 02/24/2022 10:14:07 - INFO - codeparrot_training - Step 6694: {'lr': 0.0004882943683775499, 'samples': 3427840, 'steps': 6694, 'loss/train': 0.8966919183731079} 02/24/2022 10:14:11 - INFO - codeparrot_training - Step 6695: {'lr': 0.0004882894196708738, 'samples': 3428352, 'steps': 6695, 'loss/train': 3.11983585357666} 02/24/2022 10:14:16 - INFO - codeparrot_training - Step 6696: {'lr': 0.0004882844699434415, 'samples': 3428864, 'steps': 6696, 'loss/train': 3.249206066131592} 02/24/2022 10:14:20 - INFO - codeparrot_training - Step 6697: {'lr': 0.0004882795191952741, 'samples': 3429376, 'steps': 6697, 'loss/train': 1.4275602102279663} 02/24/2022 10:14:25 - INFO - codeparrot_training - Step 6698: {'lr': 0.0004882745674263931, 'samples': 3429888, 'steps': 6698, 'loss/train': 1.953503131866455} 02/24/2022 10:14:29 - INFO - codeparrot_training - Step 6699: {'lr': 0.00048826961463681936, 'samples': 3430400, 'steps': 6699, 'loss/train': 2.1883325576782227} 02/24/2022 10:14:34 - INFO - codeparrot_training - Step 6700: {'lr': 0.00048826466082657426, 'samples': 3430912, 'steps': 6700, 'loss/train': 2.350893020629883} 02/24/2022 10:14:38 - INFO - codeparrot_training - Step 6701: {'lr': 0.000488259705995679, 'samples': 3431424, 'steps': 6701, 'loss/train': 2.4833803176879883} 02/24/2022 10:14:44 - INFO - codeparrot_training - Step 6702: {'lr': 0.0004882547501441549, 'samples': 3431936, 'steps': 6702, 'loss/train': 2.158377170562744} 02/24/2022 10:14:50 - INFO - codeparrot_training - Step 6703: {'lr': 0.000488249793272023, 'samples': 3432448, 'steps': 6703, 'loss/train': 1.896710991859436} 02/24/2022 10:14:53 - INFO - codeparrot_training - Step 6704: {'lr': 0.0004882448353793048, 'samples': 3432960, 'steps': 6704, 'loss/train': 0.7881848812103271} 02/24/2022 10:14:59 - INFO - codeparrot_training - Step 6705: {'lr': 0.0004882398764660212, 'samples': 3433472, 'steps': 6705, 'loss/train': 2.688445568084717} 02/24/2022 10:15:02 - INFO - codeparrot_training - Step 6706: {'lr': 0.00048823491653219366, 'samples': 3433984, 'steps': 6706, 'loss/train': 2.4517712593078613} 02/24/2022 10:15:08 - INFO - codeparrot_training - Step 6707: {'lr': 0.00048822995557784343, 'samples': 3434496, 'steps': 6707, 'loss/train': 2.814756393432617} 02/24/2022 10:15:11 - INFO - codeparrot_training - Step 6708: {'lr': 0.00048822499360299165, 'samples': 3435008, 'steps': 6708, 'loss/train': 3.1927785873413086} 02/24/2022 10:15:17 - INFO - codeparrot_training - Step 6709: {'lr': 0.00048822003060765973, 'samples': 3435520, 'steps': 6709, 'loss/train': 1.3512688875198364} 02/24/2022 10:15:20 - INFO - codeparrot_training - Step 6710: {'lr': 0.00048821506659186875, 'samples': 3436032, 'steps': 6710, 'loss/train': 3.470193862915039} 02/24/2022 10:15:26 - INFO - codeparrot_training - Step 6711: {'lr': 0.0004882101015556402, 'samples': 3436544, 'steps': 6711, 'loss/train': 2.6784918308258057} 02/24/2022 10:15:29 - INFO - codeparrot_training - Step 6712: {'lr': 0.00048820513549899507, 'samples': 3437056, 'steps': 6712, 'loss/train': 3.514265537261963} 02/24/2022 10:15:33 - INFO - codeparrot_training - Step 6713: {'lr': 0.00048820016842195487, 'samples': 3437568, 'steps': 6713, 'loss/train': 2.6792969703674316} 02/24/2022 10:15:39 - INFO - codeparrot_training - Step 6714: {'lr': 0.0004881952003245408, 'samples': 3438080, 'steps': 6714, 'loss/train': 2.1285791397094727} 02/24/2022 10:15:42 - INFO - codeparrot_training - Step 6715: {'lr': 0.00048819023120677405, 'samples': 3438592, 'steps': 6715, 'loss/train': 1.4544459581375122} 02/24/2022 10:15:48 - INFO - codeparrot_training - Step 6716: {'lr': 0.000488185261068676, 'samples': 3439104, 'steps': 6716, 'loss/train': 2.244325637817383} 02/24/2022 10:15:54 - INFO - codeparrot_training - Step 6717: {'lr': 0.000488180289910268, 'samples': 3439616, 'steps': 6717, 'loss/train': 3.0336530208587646} 02/24/2022 10:15:57 - INFO - codeparrot_training - Step 6718: {'lr': 0.0004881753177315711, 'samples': 3440128, 'steps': 6718, 'loss/train': 2.2515945434570312} 02/24/2022 10:16:03 - INFO - codeparrot_training - Step 6719: {'lr': 0.0004881703445326069, 'samples': 3440640, 'steps': 6719, 'loss/train': 3.2806403636932373} 02/24/2022 10:16:06 - INFO - codeparrot_training - Step 6720: {'lr': 0.0004881653703133966, 'samples': 3441152, 'steps': 6720, 'loss/train': 3.607123851776123} 02/24/2022 10:16:12 - INFO - codeparrot_training - Step 6721: {'lr': 0.00048816039507396135, 'samples': 3441664, 'steps': 6721, 'loss/train': 2.0291852951049805} 02/24/2022 10:16:15 - INFO - codeparrot_training - Step 6722: {'lr': 0.00048815541881432273, 'samples': 3442176, 'steps': 6722, 'loss/train': 0.6921824812889099} 02/24/2022 10:16:22 - INFO - codeparrot_training - Step 6723: {'lr': 0.00048815044153450185, 'samples': 3442688, 'steps': 6723, 'loss/train': 2.089301586151123} 02/24/2022 10:16:25 - INFO - codeparrot_training - Step 6724: {'lr': 0.00048814546323452013, 'samples': 3443200, 'steps': 6724, 'loss/train': 3.177513360977173} 02/24/2022 10:16:31 - INFO - codeparrot_training - Step 6725: {'lr': 0.0004881404839143988, 'samples': 3443712, 'steps': 6725, 'loss/train': 2.178401231765747} 02/24/2022 10:16:34 - INFO - codeparrot_training - Step 6726: {'lr': 0.00048813550357415937, 'samples': 3444224, 'steps': 6726, 'loss/train': 1.9097955226898193} 02/24/2022 10:16:40 - INFO - codeparrot_training - Step 6727: {'lr': 0.00048813052221382294, 'samples': 3444736, 'steps': 6727, 'loss/train': 1.7666443586349487} 02/24/2022 10:16:43 - INFO - codeparrot_training - Step 6728: {'lr': 0.000488125539833411, 'samples': 3445248, 'steps': 6728, 'loss/train': 2.45284366607666} 02/24/2022 10:16:49 - INFO - codeparrot_training - Step 6729: {'lr': 0.0004881205564329449, 'samples': 3445760, 'steps': 6729, 'loss/train': 3.046703338623047} 02/24/2022 10:16:52 - INFO - codeparrot_training - Step 6730: {'lr': 0.00048811557201244594, 'samples': 3446272, 'steps': 6730, 'loss/train': 2.918105363845825} 02/24/2022 10:16:58 - INFO - codeparrot_training - Step 6731: {'lr': 0.0004881105865719355, 'samples': 3446784, 'steps': 6731, 'loss/train': 3.2985289096832275} 02/24/2022 10:17:01 - INFO - codeparrot_training - Step 6732: {'lr': 0.00048810560011143485, 'samples': 3447296, 'steps': 6732, 'loss/train': 2.6072769165039062} 02/24/2022 10:17:07 - INFO - codeparrot_training - Step 6733: {'lr': 0.0004881006126309654, 'samples': 3447808, 'steps': 6733, 'loss/train': 1.818380355834961} 02/24/2022 10:17:10 - INFO - codeparrot_training - Step 6734: {'lr': 0.00048809562413054864, 'samples': 3448320, 'steps': 6734, 'loss/train': 2.0489141941070557} 02/24/2022 10:17:16 - INFO - codeparrot_training - Step 6735: {'lr': 0.00048809063461020575, 'samples': 3448832, 'steps': 6735, 'loss/train': 3.461740016937256} 02/24/2022 10:17:19 - INFO - codeparrot_training - Step 6736: {'lr': 0.0004880856440699582, 'samples': 3449344, 'steps': 6736, 'loss/train': 2.1043686866760254} 02/24/2022 10:17:25 - INFO - codeparrot_training - Step 6737: {'lr': 0.00048808065250982737, 'samples': 3449856, 'steps': 6737, 'loss/train': 2.4198501110076904} 02/24/2022 10:17:28 - INFO - codeparrot_training - Step 6738: {'lr': 0.0004880756599298346, 'samples': 3450368, 'steps': 6738, 'loss/train': 0.9834362268447876} 02/24/2022 10:17:34 - INFO - codeparrot_training - Step 6739: {'lr': 0.0004880706663300013, 'samples': 3450880, 'steps': 6739, 'loss/train': 1.7150791883468628} 02/24/2022 10:17:38 - INFO - codeparrot_training - Step 6740: {'lr': 0.0004880656717103489, 'samples': 3451392, 'steps': 6740, 'loss/train': 3.4452342987060547} 02/24/2022 10:17:43 - INFO - codeparrot_training - Step 6741: {'lr': 0.00048806067607089866, 'samples': 3451904, 'steps': 6741, 'loss/train': 2.5446107387542725} 02/24/2022 10:17:47 - INFO - codeparrot_training - Step 6742: {'lr': 0.00048805567941167215, 'samples': 3452416, 'steps': 6742, 'loss/train': 1.396004319190979} 02/24/2022 10:17:52 - INFO - codeparrot_training - Step 6743: {'lr': 0.0004880506817326907, 'samples': 3452928, 'steps': 6743, 'loss/train': 1.0757395029067993} 02/24/2022 10:17:56 - INFO - codeparrot_training - Step 6744: {'lr': 0.0004880456830339757, 'samples': 3453440, 'steps': 6744, 'loss/train': 4.272819519042969} 02/24/2022 10:18:01 - INFO - codeparrot_training - Step 6745: {'lr': 0.00048804068331554864, 'samples': 3453952, 'steps': 6745, 'loss/train': 2.1425302028656006} 02/24/2022 10:18:05 - INFO - codeparrot_training - Step 6746: {'lr': 0.00048803568257743083, 'samples': 3454464, 'steps': 6746, 'loss/train': 2.6007919311523438} 02/24/2022 10:18:10 - INFO - codeparrot_training - Step 6747: {'lr': 0.00048803068081964375, 'samples': 3454976, 'steps': 6747, 'loss/train': 2.067784309387207} 02/24/2022 10:18:14 - INFO - codeparrot_training - Step 6748: {'lr': 0.00048802567804220875, 'samples': 3455488, 'steps': 6748, 'loss/train': 2.053147554397583} 02/24/2022 10:18:20 - INFO - codeparrot_training - Step 6749: {'lr': 0.0004880206742451474, 'samples': 3456000, 'steps': 6749, 'loss/train': 1.5851526260375977} 02/24/2022 10:18:23 - INFO - codeparrot_training - Step 6750: {'lr': 0.0004880156694284811, 'samples': 3456512, 'steps': 6750, 'loss/train': 2.108826160430908} 02/24/2022 10:18:29 - INFO - codeparrot_training - Step 6751: {'lr': 0.00048801066359223117, 'samples': 3457024, 'steps': 6751, 'loss/train': 2.5202291011810303} 02/24/2022 10:18:32 - INFO - codeparrot_training - Step 6752: {'lr': 0.00048800565673641917, 'samples': 3457536, 'steps': 6752, 'loss/train': 2.684048652648926} 02/24/2022 10:18:38 - INFO - codeparrot_training - Step 6753: {'lr': 0.00048800064886106654, 'samples': 3458048, 'steps': 6753, 'loss/train': 2.449248790740967} 02/24/2022 10:18:41 - INFO - codeparrot_training - Step 6754: {'lr': 0.0004879956399661947, 'samples': 3458560, 'steps': 6754, 'loss/train': 2.1509690284729004} 02/24/2022 10:18:47 - INFO - codeparrot_training - Step 6755: {'lr': 0.000487990630051825, 'samples': 3459072, 'steps': 6755, 'loss/train': 2.5240941047668457} 02/24/2022 10:18:50 - INFO - codeparrot_training - Step 6756: {'lr': 0.00048798561911797913, 'samples': 3459584, 'steps': 6756, 'loss/train': 1.8330283164978027} 02/24/2022 10:18:56 - INFO - codeparrot_training - Step 6757: {'lr': 0.0004879806071646784, 'samples': 3460096, 'steps': 6757, 'loss/train': 3.353050470352173} 02/24/2022 10:18:59 - INFO - codeparrot_training - Step 6758: {'lr': 0.00048797559419194427, 'samples': 3460608, 'steps': 6758, 'loss/train': 1.622212290763855} 02/24/2022 10:19:06 - INFO - codeparrot_training - Step 6759: {'lr': 0.00048797058019979837, 'samples': 3461120, 'steps': 6759, 'loss/train': 1.6564418077468872} 02/24/2022 10:19:09 - INFO - codeparrot_training - Step 6760: {'lr': 0.00048796556518826195, 'samples': 3461632, 'steps': 6760, 'loss/train': 2.486419439315796} 02/24/2022 10:19:15 - INFO - codeparrot_training - Step 6761: {'lr': 0.00048796054915735664, 'samples': 3462144, 'steps': 6761, 'loss/train': 2.234741449356079} 02/24/2022 10:19:18 - INFO - codeparrot_training - Step 6762: {'lr': 0.00048795553210710397, 'samples': 3462656, 'steps': 6762, 'loss/train': 2.0524232387542725} 02/24/2022 10:19:24 - INFO - codeparrot_training - Step 6763: {'lr': 0.00048795051403752534, 'samples': 3463168, 'steps': 6763, 'loss/train': 2.742936372756958} 02/24/2022 10:19:27 - INFO - codeparrot_training - Step 6764: {'lr': 0.0004879454949486422, 'samples': 3463680, 'steps': 6764, 'loss/train': 2.63751220703125} 02/24/2022 10:19:33 - INFO - codeparrot_training - Step 6765: {'lr': 0.00048794047484047615, 'samples': 3464192, 'steps': 6765, 'loss/train': 3.1249783039093018} 02/24/2022 10:19:36 - INFO - codeparrot_training - Step 6766: {'lr': 0.00048793545371304863, 'samples': 3464704, 'steps': 6766, 'loss/train': 2.2070140838623047} 02/24/2022 10:19:42 - INFO - codeparrot_training - Step 6767: {'lr': 0.0004879304315663813, 'samples': 3465216, 'steps': 6767, 'loss/train': 2.943758249282837} 02/24/2022 10:19:45 - INFO - codeparrot_training - Step 6768: {'lr': 0.00048792540840049544, 'samples': 3465728, 'steps': 6768, 'loss/train': 1.2936748266220093} 02/24/2022 10:19:51 - INFO - codeparrot_training - Step 6769: {'lr': 0.00048792038421541266, 'samples': 3466240, 'steps': 6769, 'loss/train': 1.001870036125183} 02/24/2022 10:19:55 - INFO - codeparrot_training - Step 6770: {'lr': 0.0004879153590111546, 'samples': 3466752, 'steps': 6770, 'loss/train': 2.885420083999634} 02/24/2022 10:20:00 - INFO - codeparrot_training - Step 6771: {'lr': 0.0004879103327877426, 'samples': 3467264, 'steps': 6771, 'loss/train': 2.096390962600708} 02/24/2022 10:20:04 - INFO - codeparrot_training - Step 6772: {'lr': 0.0004879053055451983, 'samples': 3467776, 'steps': 6772, 'loss/train': 2.433114528656006} 02/24/2022 10:20:09 - INFO - codeparrot_training - Step 6773: {'lr': 0.00048790027728354323, 'samples': 3468288, 'steps': 6773, 'loss/train': 2.698664903640747} 02/24/2022 10:20:12 - INFO - codeparrot_training - Step 6774: {'lr': 0.0004878952480027989, 'samples': 3468800, 'steps': 6774, 'loss/train': 2.100001573562622} 02/24/2022 10:20:18 - INFO - codeparrot_training - Step 6775: {'lr': 0.0004878902177029869, 'samples': 3469312, 'steps': 6775, 'loss/train': 2.992250919342041} 02/24/2022 10:20:21 - INFO - codeparrot_training - Step 6776: {'lr': 0.0004878851863841287, 'samples': 3469824, 'steps': 6776, 'loss/train': 1.8872079849243164} 02/24/2022 10:20:27 - INFO - codeparrot_training - Step 6777: {'lr': 0.00048788015404624597, 'samples': 3470336, 'steps': 6777, 'loss/train': 6.179102420806885} 02/24/2022 10:20:30 - INFO - codeparrot_training - Step 6778: {'lr': 0.0004878751206893601, 'samples': 3470848, 'steps': 6778, 'loss/train': 2.476759910583496} 02/24/2022 10:20:36 - INFO - codeparrot_training - Step 6779: {'lr': 0.0004878700863134928, 'samples': 3471360, 'steps': 6779, 'loss/train': 2.795654535293579} 02/24/2022 10:20:40 - INFO - codeparrot_training - Step 6780: {'lr': 0.00048786505091866564, 'samples': 3471872, 'steps': 6780, 'loss/train': 2.3990864753723145} 02/24/2022 10:20:45 - INFO - codeparrot_training - Step 6781: {'lr': 0.0004878600145049001, 'samples': 3472384, 'steps': 6781, 'loss/train': 1.354610562324524} 02/24/2022 10:20:51 - INFO - codeparrot_training - Step 6782: {'lr': 0.0004878549770722177, 'samples': 3472896, 'steps': 6782, 'loss/train': 2.642122983932495} 02/24/2022 10:20:54 - INFO - codeparrot_training - Step 6783: {'lr': 0.0004878499386206402, 'samples': 3473408, 'steps': 6783, 'loss/train': 2.210378885269165} 02/24/2022 10:20:58 - INFO - codeparrot_training - Step 6784: {'lr': 0.000487844899150189, 'samples': 3473920, 'steps': 6784, 'loss/train': 1.9647313356399536} 02/24/2022 10:21:04 - INFO - codeparrot_training - Step 6785: {'lr': 0.0004878398586608859, 'samples': 3474432, 'steps': 6785, 'loss/train': 2.1268482208251953} 02/24/2022 10:21:08 - INFO - codeparrot_training - Step 6786: {'lr': 0.0004878348171527523, 'samples': 3474944, 'steps': 6786, 'loss/train': 1.8752449750900269} 02/24/2022 10:21:13 - INFO - codeparrot_training - Step 6787: {'lr': 0.0004878297746258099, 'samples': 3475456, 'steps': 6787, 'loss/train': 4.299953460693359} 02/24/2022 10:21:19 - INFO - codeparrot_training - Step 6788: {'lr': 0.0004878247310800802, 'samples': 3475968, 'steps': 6788, 'loss/train': 2.0954201221466064} 02/24/2022 10:21:23 - INFO - codeparrot_training - Step 6789: {'lr': 0.0004878196865155849, 'samples': 3476480, 'steps': 6789, 'loss/train': 1.4868054389953613} 02/24/2022 10:21:26 - INFO - codeparrot_training - Step 6790: {'lr': 0.0004878146409323456, 'samples': 3476992, 'steps': 6790, 'loss/train': 2.5076186656951904} 02/24/2022 10:21:32 - INFO - codeparrot_training - Step 6791: {'lr': 0.00048780959433038386, 'samples': 3477504, 'steps': 6791, 'loss/train': 2.459852457046509} 02/24/2022 10:21:35 - INFO - codeparrot_training - Step 6792: {'lr': 0.00048780454670972127, 'samples': 3478016, 'steps': 6792, 'loss/train': 2.1706223487854004} 02/24/2022 10:21:41 - INFO - codeparrot_training - Step 6793: {'lr': 0.00048779949807037967, 'samples': 3478528, 'steps': 6793, 'loss/train': 2.4286062717437744} 02/24/2022 10:21:44 - INFO - codeparrot_training - Step 6794: {'lr': 0.0004877944484123804, 'samples': 3479040, 'steps': 6794, 'loss/train': 2.3741579055786133} 02/24/2022 10:21:51 - INFO - codeparrot_training - Step 6795: {'lr': 0.00048778939773574525, 'samples': 3479552, 'steps': 6795, 'loss/train': 3.1717588901519775} 02/24/2022 10:21:54 - INFO - codeparrot_training - Step 6796: {'lr': 0.0004877843460404959, 'samples': 3480064, 'steps': 6796, 'loss/train': 3.162328004837036} 02/24/2022 10:22:00 - INFO - codeparrot_training - Step 6797: {'lr': 0.00048777929332665385, 'samples': 3480576, 'steps': 6797, 'loss/train': 2.5856754779815674} 02/24/2022 10:22:03 - INFO - codeparrot_training - Step 6798: {'lr': 0.00048777423959424083, 'samples': 3481088, 'steps': 6798, 'loss/train': 3.2038838863372803} 02/24/2022 10:22:09 - INFO - codeparrot_training - Step 6799: {'lr': 0.00048776918484327847, 'samples': 3481600, 'steps': 6799, 'loss/train': 2.595067024230957} 02/24/2022 10:22:12 - INFO - codeparrot_training - Step 6800: {'lr': 0.0004877641290737884, 'samples': 3482112, 'steps': 6800, 'loss/train': 2.380418300628662} 02/24/2022 10:22:18 - INFO - codeparrot_training - Step 6801: {'lr': 0.0004877590722857923, 'samples': 3482624, 'steps': 6801, 'loss/train': 3.4098522663116455} 02/24/2022 10:22:21 - INFO - codeparrot_training - Step 6802: {'lr': 0.00048775401447931187, 'samples': 3483136, 'steps': 6802, 'loss/train': 1.665794014930725} 02/24/2022 10:22:27 - INFO - codeparrot_training - Step 6803: {'lr': 0.0004877489556543687, 'samples': 3483648, 'steps': 6803, 'loss/train': 1.0300036668777466} 02/24/2022 10:22:30 - INFO - codeparrot_training - Step 6804: {'lr': 0.00048774389581098454, 'samples': 3484160, 'steps': 6804, 'loss/train': 1.4774547815322876} 02/24/2022 10:22:36 - INFO - codeparrot_training - Step 6805: {'lr': 0.00048773883494918096, 'samples': 3484672, 'steps': 6805, 'loss/train': 0.5300992727279663} 02/24/2022 10:22:40 - INFO - codeparrot_training - Step 6806: {'lr': 0.0004877337730689797, 'samples': 3485184, 'steps': 6806, 'loss/train': 2.1091196537017822} 02/24/2022 10:22:45 - INFO - codeparrot_training - Step 6807: {'lr': 0.00048772871017040256, 'samples': 3485696, 'steps': 6807, 'loss/train': 0.8454818725585938} 02/24/2022 10:22:49 - INFO - codeparrot_training - Step 6808: {'lr': 0.000487723646253471, 'samples': 3486208, 'steps': 6808, 'loss/train': 2.7593982219696045} 02/24/2022 10:22:54 - INFO - codeparrot_training - Step 6809: {'lr': 0.00048771858131820684, 'samples': 3486720, 'steps': 6809, 'loss/train': 0.9625942707061768} 02/24/2022 10:22:58 - INFO - codeparrot_training - Step 6810: {'lr': 0.0004877135153646318, 'samples': 3487232, 'steps': 6810, 'loss/train': 2.464139938354492} 02/24/2022 10:23:03 - INFO - codeparrot_training - Step 6811: {'lr': 0.0004877084483927675, 'samples': 3487744, 'steps': 6811, 'loss/train': 2.0602457523345947} 02/24/2022 10:23:07 - INFO - codeparrot_training - Step 6812: {'lr': 0.00048770338040263574, 'samples': 3488256, 'steps': 6812, 'loss/train': 2.813868522644043} 02/24/2022 10:23:12 - INFO - codeparrot_training - Step 6813: {'lr': 0.00048769831139425815, 'samples': 3488768, 'steps': 6813, 'loss/train': 2.190279483795166} 02/24/2022 10:23:16 - INFO - codeparrot_training - Step 6814: {'lr': 0.0004876932413676565, 'samples': 3489280, 'steps': 6814, 'loss/train': 3.5951104164123535} 02/24/2022 10:23:22 - INFO - codeparrot_training - Step 6815: {'lr': 0.0004876881703228524, 'samples': 3489792, 'steps': 6815, 'loss/train': 1.5487090349197388} 02/24/2022 10:23:26 - INFO - codeparrot_training - Step 6816: {'lr': 0.0004876830982598677, 'samples': 3490304, 'steps': 6816, 'loss/train': 3.0427701473236084} 02/24/2022 10:23:31 - INFO - codeparrot_training - Step 6817: {'lr': 0.0004876780251787241, 'samples': 3490816, 'steps': 6817, 'loss/train': 2.7557144165039062} 02/24/2022 10:23:35 - INFO - codeparrot_training - Step 6818: {'lr': 0.0004876729510794433, 'samples': 3491328, 'steps': 6818, 'loss/train': 1.4410158395767212} 02/24/2022 10:23:40 - INFO - codeparrot_training - Step 6819: {'lr': 0.00048766787596204704, 'samples': 3491840, 'steps': 6819, 'loss/train': 2.8650803565979004} 02/24/2022 10:23:44 - INFO - codeparrot_training - Step 6820: {'lr': 0.000487662799826557, 'samples': 3492352, 'steps': 6820, 'loss/train': 3.0239808559417725} 02/24/2022 10:23:49 - INFO - codeparrot_training - Step 6821: {'lr': 0.00048765772267299513, 'samples': 3492864, 'steps': 6821, 'loss/train': 2.364173173904419} 02/24/2022 10:23:53 - INFO - codeparrot_training - Step 6822: {'lr': 0.00048765264450138297, 'samples': 3493376, 'steps': 6822, 'loss/train': 3.3968331813812256} 02/24/2022 10:23:58 - INFO - codeparrot_training - Step 6823: {'lr': 0.00048764756531174237, 'samples': 3493888, 'steps': 6823, 'loss/train': 2.6022391319274902} 02/24/2022 10:24:02 - INFO - codeparrot_training - Step 6824: {'lr': 0.000487642485104095, 'samples': 3494400, 'steps': 6824, 'loss/train': 1.6639845371246338} 02/24/2022 10:24:07 - INFO - codeparrot_training - Step 6825: {'lr': 0.0004876374038784627, 'samples': 3494912, 'steps': 6825, 'loss/train': 0.5440935492515564} 02/24/2022 10:24:11 - INFO - codeparrot_training - Step 6826: {'lr': 0.0004876323216348673, 'samples': 3495424, 'steps': 6826, 'loss/train': 1.8696203231811523} 02/24/2022 10:24:16 - INFO - codeparrot_training - Step 6827: {'lr': 0.0004876272383733304, 'samples': 3495936, 'steps': 6827, 'loss/train': 1.1684650182724} 02/24/2022 10:24:20 - INFO - codeparrot_training - Step 6828: {'lr': 0.0004876221540938739, 'samples': 3496448, 'steps': 6828, 'loss/train': 2.7044548988342285} 02/24/2022 10:24:26 - INFO - codeparrot_training - Step 6829: {'lr': 0.00048761706879651956, 'samples': 3496960, 'steps': 6829, 'loss/train': 2.303992748260498} 02/24/2022 10:24:31 - INFO - codeparrot_training - Step 6830: {'lr': 0.00048761198248128913, 'samples': 3497472, 'steps': 6830, 'loss/train': 2.1266443729400635} 02/24/2022 10:24:35 - INFO - codeparrot_training - Step 6831: {'lr': 0.00048760689514820444, 'samples': 3497984, 'steps': 6831, 'loss/train': 3.373574733734131} 02/24/2022 10:24:38 - INFO - codeparrot_training - Step 6832: {'lr': 0.0004876018067972872, 'samples': 3498496, 'steps': 6832, 'loss/train': 2.0352985858917236} 02/24/2022 10:24:44 - INFO - codeparrot_training - Step 6833: {'lr': 0.00048759671742855935, 'samples': 3499008, 'steps': 6833, 'loss/train': 2.7940714359283447} 02/24/2022 10:24:48 - INFO - codeparrot_training - Step 6834: {'lr': 0.00048759162704204253, 'samples': 3499520, 'steps': 6834, 'loss/train': 2.209381341934204} 02/24/2022 10:24:53 - INFO - codeparrot_training - Step 6835: {'lr': 0.0004875865356377587, 'samples': 3500032, 'steps': 6835, 'loss/train': 3.131401538848877} 02/24/2022 10:24:57 - INFO - codeparrot_training - Step 6836: {'lr': 0.0004875814432157295, 'samples': 3500544, 'steps': 6836, 'loss/train': 1.1384847164154053} 02/24/2022 10:25:02 - INFO - codeparrot_training - Step 6837: {'lr': 0.0004875763497759769, 'samples': 3501056, 'steps': 6837, 'loss/train': 2.843522071838379} 02/24/2022 10:25:06 - INFO - codeparrot_training - Step 6838: {'lr': 0.00048757125531852263, 'samples': 3501568, 'steps': 6838, 'loss/train': 1.1262342929840088} 02/24/2022 10:25:11 - INFO - codeparrot_training - Step 6839: {'lr': 0.00048756615984338857, 'samples': 3502080, 'steps': 6839, 'loss/train': 2.284484386444092} 02/24/2022 10:25:15 - INFO - codeparrot_training - Step 6840: {'lr': 0.0004875610633505965, 'samples': 3502592, 'steps': 6840, 'loss/train': 3.321861505508423} 02/24/2022 10:25:20 - INFO - codeparrot_training - Step 6841: {'lr': 0.00048755596584016824, 'samples': 3503104, 'steps': 6841, 'loss/train': 1.028484582901001} 02/24/2022 10:25:24 - INFO - codeparrot_training - Step 6842: {'lr': 0.0004875508673121257, 'samples': 3503616, 'steps': 6842, 'loss/train': 4.017027854919434} 02/24/2022 10:25:30 - INFO - codeparrot_training - Step 6843: {'lr': 0.00048754576776649066, 'samples': 3504128, 'steps': 6843, 'loss/train': 2.2528154850006104} 02/24/2022 10:25:34 - INFO - codeparrot_training - Step 6844: {'lr': 0.000487540667203285, 'samples': 3504640, 'steps': 6844, 'loss/train': 2.3947060108184814} 02/24/2022 10:25:39 - INFO - codeparrot_training - Step 6845: {'lr': 0.0004875355656225305, 'samples': 3505152, 'steps': 6845, 'loss/train': 3.461034059524536} 02/24/2022 10:25:43 - INFO - codeparrot_training - Step 6846: {'lr': 0.0004875304630242491, 'samples': 3505664, 'steps': 6846, 'loss/train': 1.7348238229751587} 02/24/2022 10:25:48 - INFO - codeparrot_training - Step 6847: {'lr': 0.00048752535940846267, 'samples': 3506176, 'steps': 6847, 'loss/train': 2.5499794483184814} 02/24/2022 10:25:54 - INFO - codeparrot_training - Step 6848: {'lr': 0.0004875202547751929, 'samples': 3506688, 'steps': 6848, 'loss/train': 0.7180147767066956} 02/24/2022 10:25:57 - INFO - codeparrot_training - Step 6849: {'lr': 0.00048751514912446185, 'samples': 3507200, 'steps': 6849, 'loss/train': 1.0598224401474} 02/24/2022 10:26:03 - INFO - codeparrot_training - Step 6850: {'lr': 0.0004875100424562914, 'samples': 3507712, 'steps': 6850, 'loss/train': 2.8479793071746826} 02/24/2022 10:26:06 - INFO - codeparrot_training - Step 6851: {'lr': 0.0004875049347707032, 'samples': 3508224, 'steps': 6851, 'loss/train': 1.7459993362426758} 02/24/2022 10:26:10 - INFO - codeparrot_training - Step 6852: {'lr': 0.00048749982606771934, 'samples': 3508736, 'steps': 6852, 'loss/train': 1.103472113609314} 02/24/2022 10:26:17 - INFO - codeparrot_training - Step 6853: {'lr': 0.00048749471634736163, 'samples': 3509248, 'steps': 6853, 'loss/train': 2.467160224914551} 02/24/2022 10:26:20 - INFO - codeparrot_training - Step 6854: {'lr': 0.0004874896056096521, 'samples': 3509760, 'steps': 6854, 'loss/train': 2.8990769386291504} 02/24/2022 10:26:26 - INFO - codeparrot_training - Step 6855: {'lr': 0.0004874844938546123, 'samples': 3510272, 'steps': 6855, 'loss/train': 1.8433427810668945} 02/24/2022 10:26:29 - INFO - codeparrot_training - Step 6856: {'lr': 0.0004874793810822644, 'samples': 3510784, 'steps': 6856, 'loss/train': 2.625093698501587} 02/24/2022 10:26:35 - INFO - codeparrot_training - Step 6857: {'lr': 0.00048747426729263036, 'samples': 3511296, 'steps': 6857, 'loss/train': 2.515077829360962} 02/24/2022 10:26:38 - INFO - codeparrot_training - Step 6858: {'lr': 0.0004874691524857318, 'samples': 3511808, 'steps': 6858, 'loss/train': 2.274050235748291} 02/24/2022 10:26:44 - INFO - codeparrot_training - Step 6859: {'lr': 0.00048746403666159087, 'samples': 3512320, 'steps': 6859, 'loss/train': 3.2021071910858154} 02/24/2022 10:26:47 - INFO - codeparrot_training - Step 6860: {'lr': 0.0004874589198202294, 'samples': 3512832, 'steps': 6860, 'loss/train': 2.418423652648926} 02/24/2022 10:26:52 - INFO - codeparrot_training - Step 6861: {'lr': 0.0004874538019616693, 'samples': 3513344, 'steps': 6861, 'loss/train': 2.798901081085205} 02/24/2022 10:26:56 - INFO - codeparrot_training - Step 6862: {'lr': 0.0004874486830859326, 'samples': 3513856, 'steps': 6862, 'loss/train': 2.0700607299804688} 02/24/2022 10:27:03 - INFO - codeparrot_training - Step 6863: {'lr': 0.0004874435631930411, 'samples': 3514368, 'steps': 6863, 'loss/train': 3.3550164699554443} 02/24/2022 10:27:08 - INFO - codeparrot_training - Step 6864: {'lr': 0.0004874384422830167, 'samples': 3514880, 'steps': 6864, 'loss/train': 1.916977047920227} 02/24/2022 10:27:12 - INFO - codeparrot_training - Step 6865: {'lr': 0.0004874333203558815, 'samples': 3515392, 'steps': 6865, 'loss/train': 2.3820817470550537} 02/24/2022 10:27:15 - INFO - codeparrot_training - Step 6866: {'lr': 0.0004874281974116573, 'samples': 3515904, 'steps': 6866, 'loss/train': 2.2846484184265137} 02/24/2022 10:27:21 - INFO - codeparrot_training - Step 6867: {'lr': 0.0004874230734503661, 'samples': 3516416, 'steps': 6867, 'loss/train': 1.8415343761444092} 02/24/2022 10:27:24 - INFO - codeparrot_training - Step 6868: {'lr': 0.00048741794847202984, 'samples': 3516928, 'steps': 6868, 'loss/train': 6.344202518463135} 02/24/2022 10:27:30 - INFO - codeparrot_training - Step 6869: {'lr': 0.00048741282247667054, 'samples': 3517440, 'steps': 6869, 'loss/train': 3.0049126148223877} 02/24/2022 10:27:33 - INFO - codeparrot_training - Step 6870: {'lr': 0.00048740769546431, 'samples': 3517952, 'steps': 6870, 'loss/train': 2.6367475986480713} 02/24/2022 10:27:39 - INFO - codeparrot_training - Step 6871: {'lr': 0.0004874025674349704, 'samples': 3518464, 'steps': 6871, 'loss/train': 1.9336625337600708} 02/24/2022 10:27:42 - INFO - codeparrot_training - Step 6872: {'lr': 0.00048739743838867344, 'samples': 3518976, 'steps': 6872, 'loss/train': 2.931097984313965} 02/24/2022 10:27:48 - INFO - codeparrot_training - Step 6873: {'lr': 0.0004873923083254413, 'samples': 3519488, 'steps': 6873, 'loss/train': 0.9362912774085999} 02/24/2022 10:27:54 - INFO - codeparrot_training - Step 6874: {'lr': 0.0004873871772452959, 'samples': 3520000, 'steps': 6874, 'loss/train': 0.2874855101108551} 02/24/2022 10:27:57 - INFO - codeparrot_training - Step 6875: {'lr': 0.00048738204514825917, 'samples': 3520512, 'steps': 6875, 'loss/train': 2.248539686203003} 02/24/2022 10:28:04 - INFO - codeparrot_training - Step 6876: {'lr': 0.0004873769120343532, 'samples': 3521024, 'steps': 6876, 'loss/train': 2.1301231384277344} 02/24/2022 10:28:07 - INFO - codeparrot_training - Step 6877: {'lr': 0.0004873717779035999, 'samples': 3521536, 'steps': 6877, 'loss/train': 1.4798696041107178} 02/24/2022 10:28:13 - INFO - codeparrot_training - Step 6878: {'lr': 0.00048736664275602124, 'samples': 3522048, 'steps': 6878, 'loss/train': 2.549281120300293} 02/24/2022 10:28:17 - INFO - codeparrot_training - Step 6879: {'lr': 0.00048736150659163925, 'samples': 3522560, 'steps': 6879, 'loss/train': 2.0744125843048096} 02/24/2022 10:28:20 - INFO - codeparrot_training - Step 6880: {'lr': 0.000487356369410476, 'samples': 3523072, 'steps': 6880, 'loss/train': 3.563528299331665} 02/24/2022 10:28:26 - INFO - codeparrot_training - Step 6881: {'lr': 0.00048735123121255335, 'samples': 3523584, 'steps': 6881, 'loss/train': 1.9544663429260254} 02/24/2022 10:28:29 - INFO - codeparrot_training - Step 6882: {'lr': 0.0004873460919978935, 'samples': 3524096, 'steps': 6882, 'loss/train': 2.161992311477661} 02/24/2022 10:28:35 - INFO - codeparrot_training - Step 6883: {'lr': 0.00048734095176651825, 'samples': 3524608, 'steps': 6883, 'loss/train': 2.713358163833618} 02/24/2022 10:28:38 - INFO - codeparrot_training - Step 6884: {'lr': 0.00048733581051844976, 'samples': 3525120, 'steps': 6884, 'loss/train': 3.582826614379883} 02/24/2022 10:28:44 - INFO - codeparrot_training - Step 6885: {'lr': 0.0004873306682537101, 'samples': 3525632, 'steps': 6885, 'loss/train': 2.634448766708374} 02/24/2022 10:28:47 - INFO - codeparrot_training - Step 6886: {'lr': 0.0004873255249723211, 'samples': 3526144, 'steps': 6886, 'loss/train': 2.45340895652771} 02/24/2022 10:28:53 - INFO - codeparrot_training - Step 6887: {'lr': 0.000487320380674305, 'samples': 3526656, 'steps': 6887, 'loss/train': 2.3462326526641846} 02/24/2022 10:28:56 - INFO - codeparrot_training - Step 6888: {'lr': 0.0004873152353596837, 'samples': 3527168, 'steps': 6888, 'loss/train': 1.6121147871017456} 02/24/2022 10:29:02 - INFO - codeparrot_training - Step 6889: {'lr': 0.00048731008902847927, 'samples': 3527680, 'steps': 6889, 'loss/train': 2.879335641860962} 02/24/2022 10:29:06 - INFO - codeparrot_training - Step 6890: {'lr': 0.0004873049416807138, 'samples': 3528192, 'steps': 6890, 'loss/train': 2.384963274002075} 02/24/2022 10:29:11 - INFO - codeparrot_training - Step 6891: {'lr': 0.00048729979331640927, 'samples': 3528704, 'steps': 6891, 'loss/train': 2.614638566970825} 02/24/2022 10:29:15 - INFO - codeparrot_training - Step 6892: {'lr': 0.0004872946439355879, 'samples': 3529216, 'steps': 6892, 'loss/train': 2.283127546310425} 02/24/2022 10:29:20 - INFO - codeparrot_training - Step 6893: {'lr': 0.0004872894935382715, 'samples': 3529728, 'steps': 6893, 'loss/train': 2.558030366897583} 02/24/2022 10:29:24 - INFO - codeparrot_training - Step 6894: {'lr': 0.00048728434212448233, 'samples': 3530240, 'steps': 6894, 'loss/train': 2.4514951705932617} 02/24/2022 10:29:29 - INFO - codeparrot_training - Step 6895: {'lr': 0.0004872791896942423, 'samples': 3530752, 'steps': 6895, 'loss/train': 1.9218103885650635} 02/24/2022 10:29:33 - INFO - codeparrot_training - Step 6896: {'lr': 0.0004872740362475737, 'samples': 3531264, 'steps': 6896, 'loss/train': 2.0093512535095215} 02/24/2022 10:29:38 - INFO - codeparrot_training - Step 6897: {'lr': 0.00048726888178449835, 'samples': 3531776, 'steps': 6897, 'loss/train': 1.5521132946014404} 02/24/2022 10:29:44 - INFO - codeparrot_training - Step 6898: {'lr': 0.00048726372630503845, 'samples': 3532288, 'steps': 6898, 'loss/train': 3.6632180213928223} 02/24/2022 10:29:48 - INFO - codeparrot_training - Step 6899: {'lr': 0.00048725856980921616, 'samples': 3532800, 'steps': 6899, 'loss/train': 1.849155306816101} 02/24/2022 10:29:53 - INFO - codeparrot_training - Step 6900: {'lr': 0.0004872534122970535, 'samples': 3533312, 'steps': 6900, 'loss/train': 0.6469098925590515} 02/24/2022 10:29:57 - INFO - codeparrot_training - Step 6901: {'lr': 0.00048724825376857253, 'samples': 3533824, 'steps': 6901, 'loss/train': 1.8186110258102417} 02/24/2022 10:30:03 - INFO - codeparrot_training - Step 6902: {'lr': 0.0004872430942237953, 'samples': 3534336, 'steps': 6902, 'loss/train': 1.0362597703933716} 02/24/2022 10:30:06 - INFO - codeparrot_training - Step 6903: {'lr': 0.0004872379336627441, 'samples': 3534848, 'steps': 6903, 'loss/train': 2.4244754314422607} 02/24/2022 10:30:11 - INFO - codeparrot_training - Step 6904: {'lr': 0.0004872327720854409, 'samples': 3535360, 'steps': 6904, 'loss/train': 1.561259388923645} 02/24/2022 10:30:15 - INFO - codeparrot_training - Step 6905: {'lr': 0.0004872276094919078, 'samples': 3535872, 'steps': 6905, 'loss/train': 2.42621111869812} 02/24/2022 10:30:20 - INFO - codeparrot_training - Step 6906: {'lr': 0.00048722244588216695, 'samples': 3536384, 'steps': 6906, 'loss/train': 2.059445381164551} 02/24/2022 10:30:24 - INFO - codeparrot_training - Step 6907: {'lr': 0.00048721728125624054, 'samples': 3536896, 'steps': 6907, 'loss/train': 2.377080202102661} 02/24/2022 10:30:30 - INFO - codeparrot_training - Step 6908: {'lr': 0.0004872121156141506, 'samples': 3537408, 'steps': 6908, 'loss/train': 0.697849452495575} 02/24/2022 10:30:33 - INFO - codeparrot_training - Step 6909: {'lr': 0.0004872069489559192, 'samples': 3537920, 'steps': 6909, 'loss/train': 2.2627804279327393} 02/24/2022 10:30:37 - INFO - codeparrot_training - Step 6910: {'lr': 0.00048720178128156856, 'samples': 3538432, 'steps': 6910, 'loss/train': 2.5727944374084473} 02/24/2022 10:30:42 - INFO - codeparrot_training - Step 6911: {'lr': 0.00048719661259112086, 'samples': 3538944, 'steps': 6911, 'loss/train': 1.2028506994247437} 02/24/2022 10:30:46 - INFO - codeparrot_training - Step 6912: {'lr': 0.0004871914428845982, 'samples': 3539456, 'steps': 6912, 'loss/train': 3.0690088272094727} 02/24/2022 10:30:51 - INFO - codeparrot_training - Step 6913: {'lr': 0.0004871862721620227, 'samples': 3539968, 'steps': 6913, 'loss/train': 3.2593812942504883} 02/24/2022 10:30:55 - INFO - codeparrot_training - Step 6914: {'lr': 0.0004871811004234165, 'samples': 3540480, 'steps': 6914, 'loss/train': 3.0025129318237305} 02/24/2022 10:31:01 - INFO - codeparrot_training - Step 6915: {'lr': 0.0004871759276688018, 'samples': 3540992, 'steps': 6915, 'loss/train': 3.003502607345581} 02/24/2022 10:31:05 - INFO - codeparrot_training - Step 6916: {'lr': 0.00048717075389820074, 'samples': 3541504, 'steps': 6916, 'loss/train': 2.9868600368499756} 02/24/2022 10:31:10 - INFO - codeparrot_training - Step 6917: {'lr': 0.0004871655791116355, 'samples': 3542016, 'steps': 6917, 'loss/train': 2.282160520553589} 02/24/2022 10:31:14 - INFO - codeparrot_training - Step 6918: {'lr': 0.00048716040330912816, 'samples': 3542528, 'steps': 6918, 'loss/train': 2.1020379066467285} 02/24/2022 10:31:19 - INFO - codeparrot_training - Step 6919: {'lr': 0.000487155226490701, 'samples': 3543040, 'steps': 6919, 'loss/train': 2.961427688598633} 02/24/2022 10:31:23 - INFO - codeparrot_training - Step 6920: {'lr': 0.0004871500486563761, 'samples': 3543552, 'steps': 6920, 'loss/train': 1.9515552520751953} 02/24/2022 10:31:29 - INFO - codeparrot_training - Step 6921: {'lr': 0.00048714486980617577, 'samples': 3544064, 'steps': 6921, 'loss/train': 2.1839396953582764} 02/24/2022 10:31:32 - INFO - codeparrot_training - Step 6922: {'lr': 0.00048713968994012216, 'samples': 3544576, 'steps': 6922, 'loss/train': 3.6249005794525146} 02/24/2022 10:31:38 - INFO - codeparrot_training - Step 6923: {'lr': 0.00048713450905823736, 'samples': 3545088, 'steps': 6923, 'loss/train': 1.58771550655365} 02/24/2022 10:31:41 - INFO - codeparrot_training - Step 6924: {'lr': 0.0004871293271605436, 'samples': 3545600, 'steps': 6924, 'loss/train': 3.40474534034729} 02/24/2022 10:31:47 - INFO - codeparrot_training - Step 6925: {'lr': 0.00048712414424706315, 'samples': 3546112, 'steps': 6925, 'loss/train': 1.908578634262085} 02/24/2022 10:31:51 - INFO - codeparrot_training - Step 6926: {'lr': 0.0004871189603178181, 'samples': 3546624, 'steps': 6926, 'loss/train': 2.4061009883880615} 02/24/2022 10:31:56 - INFO - codeparrot_training - Step 6927: {'lr': 0.00048711377537283073, 'samples': 3547136, 'steps': 6927, 'loss/train': 1.7508264780044556} 02/24/2022 10:32:00 - INFO - codeparrot_training - Step 6928: {'lr': 0.0004871085894121233, 'samples': 3547648, 'steps': 6928, 'loss/train': 2.1758227348327637} 02/24/2022 10:32:05 - INFO - codeparrot_training - Step 6929: {'lr': 0.00048710340243571796, 'samples': 3548160, 'steps': 6929, 'loss/train': 2.505138635635376} 02/24/2022 10:32:09 - INFO - codeparrot_training - Step 6930: {'lr': 0.0004870982144436369, 'samples': 3548672, 'steps': 6930, 'loss/train': 1.9964663982391357} 02/24/2022 10:32:14 - INFO - codeparrot_training - Step 6931: {'lr': 0.0004870930254359023, 'samples': 3549184, 'steps': 6931, 'loss/train': 2.8455898761749268} 02/24/2022 10:32:18 - INFO - codeparrot_training - Step 6932: {'lr': 0.00048708783541253655, 'samples': 3549696, 'steps': 6932, 'loss/train': 1.9579960107803345} 02/24/2022 10:32:23 - INFO - codeparrot_training - Step 6933: {'lr': 0.0004870826443735618, 'samples': 3550208, 'steps': 6933, 'loss/train': 3.1020195484161377} 02/24/2022 10:32:27 - INFO - codeparrot_training - Step 6934: {'lr': 0.0004870774523190003, 'samples': 3550720, 'steps': 6934, 'loss/train': 2.6237692832946777} 02/24/2022 10:32:33 - INFO - codeparrot_training - Step 6935: {'lr': 0.00048707225924887423, 'samples': 3551232, 'steps': 6935, 'loss/train': 1.7438384294509888} 02/24/2022 10:32:36 - INFO - codeparrot_training - Step 6936: {'lr': 0.0004870670651632059, 'samples': 3551744, 'steps': 6936, 'loss/train': 1.662680745124817} 02/24/2022 10:32:42 - INFO - codeparrot_training - Step 6937: {'lr': 0.0004870618700620175, 'samples': 3552256, 'steps': 6937, 'loss/train': 2.110260009765625} 02/24/2022 10:32:45 - INFO - codeparrot_training - Step 6938: {'lr': 0.0004870566739453314, 'samples': 3552768, 'steps': 6938, 'loss/train': 2.6720337867736816} 02/24/2022 10:32:51 - INFO - codeparrot_training - Step 6939: {'lr': 0.00048705147681316974, 'samples': 3553280, 'steps': 6939, 'loss/train': 2.551278591156006} 02/24/2022 10:32:54 - INFO - codeparrot_training - Step 6940: {'lr': 0.00048704627866555486, 'samples': 3553792, 'steps': 6940, 'loss/train': 1.92770254611969} 02/24/2022 10:33:00 - INFO - codeparrot_training - Step 6941: {'lr': 0.00048704107950250887, 'samples': 3554304, 'steps': 6941, 'loss/train': 1.7141563892364502} 02/24/2022 10:33:03 - INFO - codeparrot_training - Step 6942: {'lr': 0.0004870358793240543, 'samples': 3554816, 'steps': 6942, 'loss/train': 2.0228426456451416} 02/24/2022 10:33:09 - INFO - codeparrot_training - Step 6943: {'lr': 0.00048703067813021323, 'samples': 3555328, 'steps': 6943, 'loss/train': 2.6666617393493652} 02/24/2022 10:33:12 - INFO - codeparrot_training - Step 6944: {'lr': 0.000487025475921008, 'samples': 3555840, 'steps': 6944, 'loss/train': 1.5000585317611694} 02/24/2022 10:33:19 - INFO - codeparrot_training - Step 6945: {'lr': 0.0004870202726964609, 'samples': 3556352, 'steps': 6945, 'loss/train': 2.0828866958618164} 02/24/2022 10:33:22 - INFO - codeparrot_training - Step 6946: {'lr': 0.0004870150684565943, 'samples': 3556864, 'steps': 6946, 'loss/train': 0.9956005811691284} 02/24/2022 10:33:27 - INFO - codeparrot_training - Step 6947: {'lr': 0.00048700986320143026, 'samples': 3557376, 'steps': 6947, 'loss/train': 1.6129735708236694} 02/24/2022 10:33:31 - INFO - codeparrot_training - Step 6948: {'lr': 0.0004870046569309913, 'samples': 3557888, 'steps': 6948, 'loss/train': 2.7539186477661133} 02/24/2022 10:33:36 - INFO - codeparrot_training - Step 6949: {'lr': 0.0004869994496452996, 'samples': 3558400, 'steps': 6949, 'loss/train': 3.8030338287353516} 02/24/2022 10:33:40 - INFO - codeparrot_training - Step 6950: {'lr': 0.0004869942413443776, 'samples': 3558912, 'steps': 6950, 'loss/train': 3.0598342418670654} 02/24/2022 10:33:45 - INFO - codeparrot_training - Step 6951: {'lr': 0.0004869890320282475, 'samples': 3559424, 'steps': 6951, 'loss/train': 1.4042478799819946} 02/24/2022 10:33:49 - INFO - codeparrot_training - Step 6952: {'lr': 0.0004869838216969316, 'samples': 3559936, 'steps': 6952, 'loss/train': 1.1975939273834229} 02/24/2022 10:33:54 - INFO - codeparrot_training - Step 6953: {'lr': 0.0004869786103504523, 'samples': 3560448, 'steps': 6953, 'loss/train': 3.341160535812378} 02/24/2022 10:33:58 - INFO - codeparrot_training - Step 6954: {'lr': 0.0004869733979888319, 'samples': 3560960, 'steps': 6954, 'loss/train': 2.8100550174713135} 02/24/2022 10:34:03 - INFO - codeparrot_training - Step 6955: {'lr': 0.00048696818461209265, 'samples': 3561472, 'steps': 6955, 'loss/train': 1.7726314067840576} 02/24/2022 10:34:09 - INFO - codeparrot_training - Step 6956: {'lr': 0.0004869629702202569, 'samples': 3561984, 'steps': 6956, 'loss/train': 2.453686475753784} 02/24/2022 10:34:12 - INFO - codeparrot_training - Step 6957: {'lr': 0.0004869577548133471, 'samples': 3562496, 'steps': 6957, 'loss/train': 1.8270617723464966} 02/24/2022 10:34:18 - INFO - codeparrot_training - Step 6958: {'lr': 0.00048695253839138553, 'samples': 3563008, 'steps': 6958, 'loss/train': 2.497138023376465} 02/24/2022 10:34:21 - INFO - codeparrot_training - Step 6959: {'lr': 0.0004869473209543945, 'samples': 3563520, 'steps': 6959, 'loss/train': 3.538966417312622} 02/24/2022 10:34:28 - INFO - codeparrot_training - Step 6960: {'lr': 0.00048694210250239646, 'samples': 3564032, 'steps': 6960, 'loss/train': 4.415807723999023} 02/24/2022 10:34:31 - INFO - codeparrot_training - Step 6961: {'lr': 0.0004869368830354136, 'samples': 3564544, 'steps': 6961, 'loss/train': 3.716559410095215} 02/24/2022 10:34:37 - INFO - codeparrot_training - Step 6962: {'lr': 0.00048693166255346843, 'samples': 3565056, 'steps': 6962, 'loss/train': 2.985011100769043} 02/24/2022 10:34:40 - INFO - codeparrot_training - Step 6963: {'lr': 0.0004869264410565832, 'samples': 3565568, 'steps': 6963, 'loss/train': 3.226262331008911} 02/24/2022 10:34:46 - INFO - codeparrot_training - Step 6964: {'lr': 0.00048692121854478033, 'samples': 3566080, 'steps': 6964, 'loss/train': 2.240480899810791} 02/24/2022 10:34:49 - INFO - codeparrot_training - Step 6965: {'lr': 0.00048691599501808223, 'samples': 3566592, 'steps': 6965, 'loss/train': 1.6684958934783936} 02/24/2022 10:34:55 - INFO - codeparrot_training - Step 6966: {'lr': 0.0004869107704765112, 'samples': 3567104, 'steps': 6966, 'loss/train': 8.854053497314453} 02/24/2022 10:34:58 - INFO - codeparrot_training - Step 6967: {'lr': 0.00048690554492008967, 'samples': 3567616, 'steps': 6967, 'loss/train': 2.2074220180511475} 02/24/2022 10:35:04 - INFO - codeparrot_training - Step 6968: {'lr': 0.00048690031834884004, 'samples': 3568128, 'steps': 6968, 'loss/train': 3.1572303771972656} 02/24/2022 10:35:07 - INFO - codeparrot_training - Step 6969: {'lr': 0.0004868950907627846, 'samples': 3568640, 'steps': 6969, 'loss/train': 3.5579230785369873} 02/24/2022 10:35:13 - INFO - codeparrot_training - Step 6970: {'lr': 0.00048688986216194585, 'samples': 3569152, 'steps': 6970, 'loss/train': 3.30551815032959} 02/24/2022 10:35:17 - INFO - codeparrot_training - Step 6971: {'lr': 0.0004868846325463462, 'samples': 3569664, 'steps': 6971, 'loss/train': 3.2069735527038574} 02/24/2022 10:35:22 - INFO - codeparrot_training - Step 6972: {'lr': 0.000486879401916008, 'samples': 3570176, 'steps': 6972, 'loss/train': 1.6296110153198242} 02/24/2022 10:35:26 - INFO - codeparrot_training - Step 6973: {'lr': 0.0004868741702709536, 'samples': 3570688, 'steps': 6973, 'loss/train': 2.4016482830047607} 02/24/2022 10:35:29 - INFO - codeparrot_training - Step 6974: {'lr': 0.0004868689376112055, 'samples': 3571200, 'steps': 6974, 'loss/train': 3.2272472381591797} 02/24/2022 10:35:35 - INFO - codeparrot_training - Step 6975: {'lr': 0.000486863703936786, 'samples': 3571712, 'steps': 6975, 'loss/train': 3.6157171726226807} 02/24/2022 10:35:38 - INFO - codeparrot_training - Step 6976: {'lr': 0.0004868584692477178, 'samples': 3572224, 'steps': 6976, 'loss/train': 1.9117414951324463} 02/24/2022 10:35:44 - INFO - codeparrot_training - Step 6977: {'lr': 0.000486853233544023, 'samples': 3572736, 'steps': 6977, 'loss/train': 2.51379656791687} 02/24/2022 10:35:47 - INFO - codeparrot_training - Step 6978: {'lr': 0.0004868479968257241, 'samples': 3573248, 'steps': 6978, 'loss/train': 3.8637757301330566} 02/24/2022 10:35:53 - INFO - codeparrot_training - Step 6979: {'lr': 0.0004868427590928437, 'samples': 3573760, 'steps': 6979, 'loss/train': 2.8012845516204834} 02/24/2022 10:35:59 - INFO - codeparrot_training - Step 6980: {'lr': 0.0004868375203454041, 'samples': 3574272, 'steps': 6980, 'loss/train': 2.2839508056640625} 02/24/2022 10:36:02 - INFO - codeparrot_training - Step 6981: {'lr': 0.0004868322805834278, 'samples': 3574784, 'steps': 6981, 'loss/train': 2.9954354763031006} 02/24/2022 10:36:08 - INFO - codeparrot_training - Step 6982: {'lr': 0.0004868270398069371, 'samples': 3575296, 'steps': 6982, 'loss/train': 2.1910762786865234} 02/24/2022 10:36:11 - INFO - codeparrot_training - Step 6983: {'lr': 0.0004868217980159546, 'samples': 3575808, 'steps': 6983, 'loss/train': 2.2371723651885986} 02/24/2022 10:36:17 - INFO - codeparrot_training - Step 6984: {'lr': 0.0004868165552105028, 'samples': 3576320, 'steps': 6984, 'loss/train': 1.5835167169570923} 02/24/2022 10:36:20 - INFO - codeparrot_training - Step 6985: {'lr': 0.000486811311390604, 'samples': 3576832, 'steps': 6985, 'loss/train': 0.7449057698249817} 02/24/2022 10:36:26 - INFO - codeparrot_training - Step 6986: {'lr': 0.0004868060665562808, 'samples': 3577344, 'steps': 6986, 'loss/train': 0.6410086750984192} 02/24/2022 10:36:30 - INFO - codeparrot_training - Step 6987: {'lr': 0.0004868008207075555, 'samples': 3577856, 'steps': 6987, 'loss/train': 1.8432601690292358} 02/24/2022 10:36:35 - INFO - codeparrot_training - Step 6988: {'lr': 0.0004867955738444508, 'samples': 3578368, 'steps': 6988, 'loss/train': 2.0191097259521484} 02/24/2022 10:36:38 - INFO - codeparrot_training - Step 6989: {'lr': 0.000486790325966989, 'samples': 3578880, 'steps': 6989, 'loss/train': 1.547868013381958} 02/24/2022 10:36:44 - INFO - codeparrot_training - Step 6990: {'lr': 0.0004867850770751926, 'samples': 3579392, 'steps': 6990, 'loss/train': 3.213212251663208} 02/24/2022 10:36:47 - INFO - codeparrot_training - Step 6991: {'lr': 0.00048677982716908416, 'samples': 3579904, 'steps': 6991, 'loss/train': 2.3578999042510986} 02/24/2022 10:36:54 - INFO - codeparrot_training - Step 6992: {'lr': 0.0004867745762486861, 'samples': 3580416, 'steps': 6992, 'loss/train': 3.327143430709839} 02/24/2022 10:36:57 - INFO - codeparrot_training - Step 6993: {'lr': 0.0004867693243140209, 'samples': 3580928, 'steps': 6993, 'loss/train': 2.5125010013580322} 02/24/2022 10:37:01 - INFO - codeparrot_training - Step 6994: {'lr': 0.0004867640713651112, 'samples': 3581440, 'steps': 6994, 'loss/train': 3.126490592956543} 02/24/2022 10:37:06 - INFO - codeparrot_training - Step 6995: {'lr': 0.0004867588174019794, 'samples': 3581952, 'steps': 6995, 'loss/train': 2.614426374435425} 02/24/2022 10:37:10 - INFO - codeparrot_training - Step 6996: {'lr': 0.00048675356242464785, 'samples': 3582464, 'steps': 6996, 'loss/train': 2.09812593460083} 02/24/2022 10:37:15 - INFO - codeparrot_training - Step 6997: {'lr': 0.0004867483064331394, 'samples': 3582976, 'steps': 6997, 'loss/train': 2.2243432998657227} 02/24/2022 10:37:19 - INFO - codeparrot_training - Step 6998: {'lr': 0.00048674304942747626, 'samples': 3583488, 'steps': 6998, 'loss/train': 2.046218156814575} 02/24/2022 10:37:24 - INFO - codeparrot_training - Step 6999: {'lr': 0.0004867377914076811, 'samples': 3584000, 'steps': 6999, 'loss/train': 2.682992696762085} 02/24/2022 10:37:24 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 10:37:41 - WARNING - huggingface_hub.repository - Several commits (7) will be pushed upstream. 02/24/2022 10:37:41 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 10:38:15 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 8b611e7..1deb887 floral-grass-11 -> floral-grass-11 02/24/2022 10:38:20 - INFO - codeparrot_training - Step 7000: {'lr': 0.00048673253237377644, 'samples': 3584512, 'steps': 7000, 'loss/train': 1.8298699855804443} 02/24/2022 10:38:25 - INFO - codeparrot_training - Step 7001: {'lr': 0.00048672727232578476, 'samples': 3585024, 'steps': 7001, 'loss/train': 1.4196652173995972} 02/24/2022 10:38:29 - INFO - codeparrot_training - Step 7002: {'lr': 0.0004867220112637286, 'samples': 3585536, 'steps': 7002, 'loss/train': 1.1515252590179443} 02/24/2022 10:38:36 - INFO - codeparrot_training - Step 7003: {'lr': 0.00048671674918763055, 'samples': 3586048, 'steps': 7003, 'loss/train': 1.9506548643112183} 02/24/2022 10:38:40 - INFO - codeparrot_training - Step 7004: {'lr': 0.00048671148609751307, 'samples': 3586560, 'steps': 7004, 'loss/train': 2.682981491088867} 02/24/2022 10:38:45 - INFO - codeparrot_training - Step 7005: {'lr': 0.0004867062219933988, 'samples': 3587072, 'steps': 7005, 'loss/train': 4.058497905731201} 02/24/2022 10:38:49 - INFO - codeparrot_training - Step 7006: {'lr': 0.00048670095687531023, 'samples': 3587584, 'steps': 7006, 'loss/train': 1.0610548257827759} 02/24/2022 10:38:54 - INFO - codeparrot_training - Step 7007: {'lr': 0.0004866956907432699, 'samples': 3588096, 'steps': 7007, 'loss/train': 1.5664478540420532} 02/24/2022 10:38:58 - INFO - codeparrot_training - Step 7008: {'lr': 0.00048669042359730043, 'samples': 3588608, 'steps': 7008, 'loss/train': 2.1107544898986816} 02/24/2022 10:39:03 - INFO - codeparrot_training - Step 7009: {'lr': 0.00048668515543742426, 'samples': 3589120, 'steps': 7009, 'loss/train': 2.958948850631714} 02/24/2022 10:39:07 - INFO - codeparrot_training - Step 7010: {'lr': 0.0004866798862636641, 'samples': 3589632, 'steps': 7010, 'loss/train': 3.252990961074829} 02/24/2022 10:39:12 - INFO - codeparrot_training - Step 7011: {'lr': 0.0004866746160760425, 'samples': 3590144, 'steps': 7011, 'loss/train': 1.740478515625} 02/24/2022 10:39:16 - INFO - codeparrot_training - Step 7012: {'lr': 0.0004866693448745819, 'samples': 3590656, 'steps': 7012, 'loss/train': 2.313049077987671} 02/24/2022 10:39:23 - INFO - codeparrot_training - Step 7013: {'lr': 0.000486664072659305, 'samples': 3591168, 'steps': 7013, 'loss/train': 6.489420413970947} 02/24/2022 10:39:29 - INFO - codeparrot_training - Step 7014: {'lr': 0.0004866587994302344, 'samples': 3591680, 'steps': 7014, 'loss/train': 2.7838709354400635} 02/24/2022 10:39:32 - INFO - codeparrot_training - Step 7015: {'lr': 0.0004866535251873926, 'samples': 3592192, 'steps': 7015, 'loss/train': 0.38497358560562134} 02/24/2022 10:39:38 - INFO - codeparrot_training - Step 7016: {'lr': 0.0004866482499308023, 'samples': 3592704, 'steps': 7016, 'loss/train': 2.105242967605591} 02/24/2022 10:39:41 - INFO - codeparrot_training - Step 7017: {'lr': 0.000486642973660486, 'samples': 3593216, 'steps': 7017, 'loss/train': 1.8313387632369995} 02/24/2022 10:39:47 - INFO - codeparrot_training - Step 7018: {'lr': 0.00048663769637646636, 'samples': 3593728, 'steps': 7018, 'loss/train': 1.5400100946426392} 02/24/2022 10:39:50 - INFO - codeparrot_training - Step 7019: {'lr': 0.000486632418078766, 'samples': 3594240, 'steps': 7019, 'loss/train': 2.302502393722534} 02/24/2022 10:39:56 - INFO - codeparrot_training - Step 7020: {'lr': 0.0004866271387674075, 'samples': 3594752, 'steps': 7020, 'loss/train': 1.95855712890625} 02/24/2022 10:39:59 - INFO - codeparrot_training - Step 7021: {'lr': 0.00048662185844241347, 'samples': 3595264, 'steps': 7021, 'loss/train': 2.7917532920837402} 02/24/2022 10:40:05 - INFO - codeparrot_training - Step 7022: {'lr': 0.00048661657710380647, 'samples': 3595776, 'steps': 7022, 'loss/train': 2.4886507987976074} 02/24/2022 10:40:08 - INFO - codeparrot_training - Step 7023: {'lr': 0.00048661129475160926, 'samples': 3596288, 'steps': 7023, 'loss/train': 0.8547494411468506} 02/24/2022 10:40:16 - INFO - codeparrot_training - Step 7024: {'lr': 0.00048660601138584436, 'samples': 3596800, 'steps': 7024, 'loss/train': 1.7446014881134033} 02/24/2022 10:40:19 - INFO - codeparrot_training - Step 7025: {'lr': 0.00048660072700653446, 'samples': 3597312, 'steps': 7025, 'loss/train': 2.086608409881592} 02/24/2022 10:40:23 - INFO - codeparrot_training - Step 7026: {'lr': 0.0004865954416137022, 'samples': 3597824, 'steps': 7026, 'loss/train': 1.7990435361862183} 02/24/2022 10:40:28 - INFO - codeparrot_training - Step 7027: {'lr': 0.0004865901552073701, 'samples': 3598336, 'steps': 7027, 'loss/train': 0.3683944642543793} 02/24/2022 10:40:34 - INFO - codeparrot_training - Step 7028: {'lr': 0.00048658486778756097, 'samples': 3598848, 'steps': 7028, 'loss/train': 2.3204288482666016} 02/24/2022 10:40:37 - INFO - codeparrot_training - Step 7029: {'lr': 0.00048657957935429734, 'samples': 3599360, 'steps': 7029, 'loss/train': 2.0324978828430176} 02/24/2022 10:40:43 - INFO - codeparrot_training - Step 7030: {'lr': 0.000486574289907602, 'samples': 3599872, 'steps': 7030, 'loss/train': 2.7280023097991943} 02/24/2022 10:40:46 - INFO - codeparrot_training - Step 7031: {'lr': 0.0004865689994474974, 'samples': 3600384, 'steps': 7031, 'loss/train': 2.8704378604888916} 02/24/2022 10:40:54 - INFO - codeparrot_training - Step 7032: {'lr': 0.00048656370797400643, 'samples': 3600896, 'steps': 7032, 'loss/train': 2.6442599296569824} 02/24/2022 10:40:57 - INFO - codeparrot_training - Step 7033: {'lr': 0.00048655841548715163, 'samples': 3601408, 'steps': 7033, 'loss/train': 2.491255760192871} 02/24/2022 10:41:01 - INFO - codeparrot_training - Step 7034: {'lr': 0.00048655312198695567, 'samples': 3601920, 'steps': 7034, 'loss/train': 3.174478769302368} 02/24/2022 10:41:06 - INFO - codeparrot_training - Step 7035: {'lr': 0.00048654782747344126, 'samples': 3602432, 'steps': 7035, 'loss/train': 2.215595245361328} 02/24/2022 10:41:12 - INFO - codeparrot_training - Step 7036: {'lr': 0.00048654253194663113, 'samples': 3602944, 'steps': 7036, 'loss/train': 1.9380426406860352} 02/24/2022 10:41:15 - INFO - codeparrot_training - Step 7037: {'lr': 0.0004865372354065478, 'samples': 3603456, 'steps': 7037, 'loss/train': 2.1176156997680664} 02/24/2022 10:41:21 - INFO - codeparrot_training - Step 7038: {'lr': 0.00048653193785321415, 'samples': 3603968, 'steps': 7038, 'loss/train': 3.197876453399658} 02/24/2022 10:41:24 - INFO - codeparrot_training - Step 7039: {'lr': 0.00048652663928665273, 'samples': 3604480, 'steps': 7039, 'loss/train': 2.3217735290527344} 02/24/2022 10:41:30 - INFO - codeparrot_training - Step 7040: {'lr': 0.00048652133970688633, 'samples': 3604992, 'steps': 7040, 'loss/train': 2.416092872619629} 02/24/2022 10:41:34 - INFO - codeparrot_training - Step 7041: {'lr': 0.0004865160391139376, 'samples': 3605504, 'steps': 7041, 'loss/train': 2.928356409072876} 02/24/2022 10:41:37 - INFO - codeparrot_training - Step 7042: {'lr': 0.0004865107375078293, 'samples': 3606016, 'steps': 7042, 'loss/train': 2.22717022895813} 02/24/2022 10:41:43 - INFO - codeparrot_training - Step 7043: {'lr': 0.000486505434888584, 'samples': 3606528, 'steps': 7043, 'loss/train': 1.0353965759277344} 02/24/2022 10:41:46 - INFO - codeparrot_training - Step 7044: {'lr': 0.0004865001312562246, 'samples': 3607040, 'steps': 7044, 'loss/train': 3.5584471225738525} 02/24/2022 10:41:54 - INFO - codeparrot_training - Step 7045: {'lr': 0.0004864948266107737, 'samples': 3607552, 'steps': 7045, 'loss/train': 2.675661325454712} 02/24/2022 10:41:57 - INFO - codeparrot_training - Step 7046: {'lr': 0.0004864895209522541, 'samples': 3608064, 'steps': 7046, 'loss/train': 2.382627248764038} 02/24/2022 10:42:03 - INFO - codeparrot_training - Step 7047: {'lr': 0.00048648421428068843, 'samples': 3608576, 'steps': 7047, 'loss/train': 1.7274190187454224} 02/24/2022 10:42:06 - INFO - codeparrot_training - Step 7048: {'lr': 0.0004864789065960995, 'samples': 3609088, 'steps': 7048, 'loss/train': 2.419377326965332} 02/24/2022 10:42:12 - INFO - codeparrot_training - Step 7049: {'lr': 0.00048647359789851, 'samples': 3609600, 'steps': 7049, 'loss/train': 3.056361198425293} 02/24/2022 10:42:15 - INFO - codeparrot_training - Step 7050: {'lr': 0.00048646828818794274, 'samples': 3610112, 'steps': 7050, 'loss/train': 2.801461696624756} 02/24/2022 10:42:21 - INFO - codeparrot_training - Step 7051: {'lr': 0.00048646297746442044, 'samples': 3610624, 'steps': 7051, 'loss/train': 1.286455750465393} 02/24/2022 10:42:24 - INFO - codeparrot_training - Step 7052: {'lr': 0.0004864576657279658, 'samples': 3611136, 'steps': 7052, 'loss/train': 3.068563461303711} 02/24/2022 10:42:30 - INFO - codeparrot_training - Step 7053: {'lr': 0.0004864523529786016, 'samples': 3611648, 'steps': 7053, 'loss/train': 3.2697532176971436} 02/24/2022 10:42:33 - INFO - codeparrot_training - Step 7054: {'lr': 0.0004864470392163506, 'samples': 3612160, 'steps': 7054, 'loss/train': 2.8305177688598633} 02/24/2022 10:42:39 - INFO - codeparrot_training - Step 7055: {'lr': 0.0004864417244412355, 'samples': 3612672, 'steps': 7055, 'loss/train': 1.4956492185592651} 02/24/2022 10:42:42 - INFO - codeparrot_training - Step 7056: {'lr': 0.0004864364086532792, 'samples': 3613184, 'steps': 7056, 'loss/train': 3.037086248397827} 02/24/2022 10:42:48 - INFO - codeparrot_training - Step 7057: {'lr': 0.00048643109185250445, 'samples': 3613696, 'steps': 7057, 'loss/train': 2.29018497467041} 02/24/2022 10:42:51 - INFO - codeparrot_training - Step 7058: {'lr': 0.0004864257740389338, 'samples': 3614208, 'steps': 7058, 'loss/train': 2.9040513038635254} 02/24/2022 10:42:57 - INFO - codeparrot_training - Step 7059: {'lr': 0.00048642045521259044, 'samples': 3614720, 'steps': 7059, 'loss/train': 2.4331135749816895} 02/24/2022 10:43:00 - INFO - codeparrot_training - Step 7060: {'lr': 0.0004864151353734968, 'samples': 3615232, 'steps': 7060, 'loss/train': 2.06575345993042} 02/24/2022 10:43:08 - INFO - codeparrot_training - Step 7061: {'lr': 0.0004864098145216758, 'samples': 3615744, 'steps': 7061, 'loss/train': 1.384238362312317} 02/24/2022 10:43:11 - INFO - codeparrot_training - Step 7062: {'lr': 0.0004864044926571503, 'samples': 3616256, 'steps': 7062, 'loss/train': 1.0450429916381836} 02/24/2022 10:43:17 - INFO - codeparrot_training - Step 7063: {'lr': 0.00048639916977994286, 'samples': 3616768, 'steps': 7063, 'loss/train': 2.0263302326202393} 02/24/2022 10:43:20 - INFO - codeparrot_training - Step 7064: {'lr': 0.0004863938458900765, 'samples': 3617280, 'steps': 7064, 'loss/train': 3.104705333709717} 02/24/2022 10:43:26 - INFO - codeparrot_training - Step 7065: {'lr': 0.000486388520987574, 'samples': 3617792, 'steps': 7065, 'loss/train': 2.1649348735809326} 02/24/2022 10:43:29 - INFO - codeparrot_training - Step 7066: {'lr': 0.0004863831950724582, 'samples': 3618304, 'steps': 7066, 'loss/train': 2.2347323894500732} 02/24/2022 10:43:35 - INFO - codeparrot_training - Step 7067: {'lr': 0.00048637786814475175, 'samples': 3618816, 'steps': 7067, 'loss/train': 3.1933720111846924} 02/24/2022 10:43:38 - INFO - codeparrot_training - Step 7068: {'lr': 0.0004863725402044776, 'samples': 3619328, 'steps': 7068, 'loss/train': 2.077728748321533} 02/24/2022 10:43:45 - INFO - codeparrot_training - Step 7069: {'lr': 0.00048636721125165855, 'samples': 3619840, 'steps': 7069, 'loss/train': 2.586998701095581} 02/24/2022 10:43:49 - INFO - codeparrot_training - Step 7070: {'lr': 0.0004863618812863174, 'samples': 3620352, 'steps': 7070, 'loss/train': 2.167773485183716} 02/24/2022 10:43:54 - INFO - codeparrot_training - Step 7071: {'lr': 0.0004863565503084771, 'samples': 3620864, 'steps': 7071, 'loss/train': 2.40755558013916} 02/24/2022 10:43:58 - INFO - codeparrot_training - Step 7072: {'lr': 0.0004863512183181603, 'samples': 3621376, 'steps': 7072, 'loss/train': 3.4997470378875732} 02/24/2022 10:44:03 - INFO - codeparrot_training - Step 7073: {'lr': 0.0004863458853153899, 'samples': 3621888, 'steps': 7073, 'loss/train': 2.8361222743988037} 02/24/2022 10:44:07 - INFO - codeparrot_training - Step 7074: {'lr': 0.00048634055130018886, 'samples': 3622400, 'steps': 7074, 'loss/train': 2.2650959491729736} 02/24/2022 10:44:13 - INFO - codeparrot_training - Step 7075: {'lr': 0.00048633521627257993, 'samples': 3622912, 'steps': 7075, 'loss/train': 2.1696739196777344} 02/24/2022 10:44:16 - INFO - codeparrot_training - Step 7076: {'lr': 0.00048632988023258596, 'samples': 3623424, 'steps': 7076, 'loss/train': 1.913853406906128} 02/24/2022 10:44:22 - INFO - codeparrot_training - Step 7077: {'lr': 0.0004863245431802298, 'samples': 3623936, 'steps': 7077, 'loss/train': 1.8230143785476685} 02/24/2022 10:44:25 - INFO - codeparrot_training - Step 7078: {'lr': 0.0004863192051155344, 'samples': 3624448, 'steps': 7078, 'loss/train': 0.14730876684188843} 02/24/2022 10:44:31 - INFO - codeparrot_training - Step 7079: {'lr': 0.0004863138660385225, 'samples': 3624960, 'steps': 7079, 'loss/train': 0.562419593334198} 02/24/2022 10:44:34 - INFO - codeparrot_training - Step 7080: {'lr': 0.00048630852594921703, 'samples': 3625472, 'steps': 7080, 'loss/train': 4.191076755523682} 02/24/2022 10:44:41 - INFO - codeparrot_training - Step 7081: {'lr': 0.00048630318484764093, 'samples': 3625984, 'steps': 7081, 'loss/train': 3.7572033405303955} 02/24/2022 10:44:45 - INFO - codeparrot_training - Step 7082: {'lr': 0.000486297842733817, 'samples': 3626496, 'steps': 7082, 'loss/train': 2.141685724258423} 02/24/2022 10:44:50 - INFO - codeparrot_training - Step 7083: {'lr': 0.0004862924996077682, 'samples': 3627008, 'steps': 7083, 'loss/train': 1.2183120250701904} 02/24/2022 10:44:54 - INFO - codeparrot_training - Step 7084: {'lr': 0.0004862871554695173, 'samples': 3627520, 'steps': 7084, 'loss/train': 2.786996364593506} 02/24/2022 10:44:59 - INFO - codeparrot_training - Step 7085: {'lr': 0.00048628181031908725, 'samples': 3628032, 'steps': 7085, 'loss/train': 0.9586674571037292} 02/24/2022 10:45:03 - INFO - codeparrot_training - Step 7086: {'lr': 0.00048627646415650094, 'samples': 3628544, 'steps': 7086, 'loss/train': 3.023047924041748} 02/24/2022 10:45:08 - INFO - codeparrot_training - Step 7087: {'lr': 0.0004862711169817813, 'samples': 3629056, 'steps': 7087, 'loss/train': 1.8972781896591187} 02/24/2022 10:45:12 - INFO - codeparrot_training - Step 7088: {'lr': 0.0004862657687949512, 'samples': 3629568, 'steps': 7088, 'loss/train': 2.777175188064575} 02/24/2022 10:45:18 - INFO - codeparrot_training - Step 7089: {'lr': 0.0004862604195960336, 'samples': 3630080, 'steps': 7089, 'loss/train': 2.29752779006958} 02/24/2022 10:45:21 - INFO - codeparrot_training - Step 7090: {'lr': 0.00048625506938505136, 'samples': 3630592, 'steps': 7090, 'loss/train': 0.6763121485710144} 02/24/2022 10:45:27 - INFO - codeparrot_training - Step 7091: {'lr': 0.00048624971816202747, 'samples': 3631104, 'steps': 7091, 'loss/train': 2.6189115047454834} 02/24/2022 10:45:30 - INFO - codeparrot_training - Step 7092: {'lr': 0.0004862443659269848, 'samples': 3631616, 'steps': 7092, 'loss/train': 1.3140772581100464} 02/24/2022 10:45:36 - INFO - codeparrot_training - Step 7093: {'lr': 0.00048623901267994625, 'samples': 3632128, 'steps': 7093, 'loss/train': 2.539118766784668} 02/24/2022 10:45:39 - INFO - codeparrot_training - Step 7094: {'lr': 0.00048623365842093483, 'samples': 3632640, 'steps': 7094, 'loss/train': 1.5668152570724487} 02/24/2022 10:45:47 - INFO - codeparrot_training - Step 7095: {'lr': 0.00048622830314997334, 'samples': 3633152, 'steps': 7095, 'loss/train': 1.4457467794418335} 02/24/2022 10:45:50 - INFO - codeparrot_training - Step 7096: {'lr': 0.0004862229468670849, 'samples': 3633664, 'steps': 7096, 'loss/train': 2.4714720249176025} 02/24/2022 10:45:56 - INFO - codeparrot_training - Step 7097: {'lr': 0.0004862175895722923, 'samples': 3634176, 'steps': 7097, 'loss/train': 1.8877856731414795} 02/24/2022 10:45:59 - INFO - codeparrot_training - Step 7098: {'lr': 0.0004862122312656186, 'samples': 3634688, 'steps': 7098, 'loss/train': 2.70857834815979} 02/24/2022 10:46:05 - INFO - codeparrot_training - Step 7099: {'lr': 0.0004862068719470867, 'samples': 3635200, 'steps': 7099, 'loss/train': 1.7322553396224976} 02/24/2022 10:46:08 - INFO - codeparrot_training - Step 7100: {'lr': 0.00048620151161671955, 'samples': 3635712, 'steps': 7100, 'loss/train': 1.8875893354415894} 02/24/2022 10:46:14 - INFO - codeparrot_training - Step 7101: {'lr': 0.0004861961502745401, 'samples': 3636224, 'steps': 7101, 'loss/train': 2.606764078140259} 02/24/2022 10:46:17 - INFO - codeparrot_training - Step 7102: {'lr': 0.00048619078792057135, 'samples': 3636736, 'steps': 7102, 'loss/train': 3.5200204849243164} 02/24/2022 10:46:23 - INFO - codeparrot_training - Step 7103: {'lr': 0.00048618542455483625, 'samples': 3637248, 'steps': 7103, 'loss/train': 2.3999087810516357} 02/24/2022 10:46:26 - INFO - codeparrot_training - Step 7104: {'lr': 0.0004861800601773579, 'samples': 3637760, 'steps': 7104, 'loss/train': 2.8969504833221436} 02/24/2022 10:46:32 - INFO - codeparrot_training - Step 7105: {'lr': 0.00048617469478815905, 'samples': 3638272, 'steps': 7105, 'loss/train': 1.8044992685317993} 02/24/2022 10:46:35 - INFO - codeparrot_training - Step 7106: {'lr': 0.00048616932838726286, 'samples': 3638784, 'steps': 7106, 'loss/train': 3.151644706726074} 02/24/2022 10:46:43 - INFO - codeparrot_training - Step 7107: {'lr': 0.0004861639609746923, 'samples': 3639296, 'steps': 7107, 'loss/train': 2.1009297370910645} 02/24/2022 10:46:47 - INFO - codeparrot_training - Step 7108: {'lr': 0.0004861585925504702, 'samples': 3639808, 'steps': 7108, 'loss/train': 3.0692696571350098} 02/24/2022 10:46:52 - INFO - codeparrot_training - Step 7109: {'lr': 0.00048615322311461973, 'samples': 3640320, 'steps': 7109, 'loss/train': 3.0109736919403076} 02/24/2022 10:46:56 - INFO - codeparrot_training - Step 7110: {'lr': 0.0004861478526671639, 'samples': 3640832, 'steps': 7110, 'loss/train': 2.0940754413604736} 02/24/2022 10:47:01 - INFO - codeparrot_training - Step 7111: {'lr': 0.0004861424812081256, 'samples': 3641344, 'steps': 7111, 'loss/train': 1.7864515781402588} 02/24/2022 10:47:05 - INFO - codeparrot_training - Step 7112: {'lr': 0.0004861371087375279, 'samples': 3641856, 'steps': 7112, 'loss/train': 2.3304591178894043} 02/24/2022 10:47:10 - INFO - codeparrot_training - Step 7113: {'lr': 0.0004861317352553938, 'samples': 3642368, 'steps': 7113, 'loss/train': 3.2613091468811035} 02/24/2022 10:47:14 - INFO - codeparrot_training - Step 7114: {'lr': 0.0004861263607617463, 'samples': 3642880, 'steps': 7114, 'loss/train': 2.572568893432617} 02/24/2022 10:47:19 - INFO - codeparrot_training - Step 7115: {'lr': 0.00048612098525660855, 'samples': 3643392, 'steps': 7115, 'loss/train': 2.615438938140869} 02/24/2022 10:47:23 - INFO - codeparrot_training - Step 7116: {'lr': 0.00048611560874000335, 'samples': 3643904, 'steps': 7116, 'loss/train': 1.7394640445709229} 02/24/2022 10:47:30 - INFO - codeparrot_training - Step 7117: {'lr': 0.000486110231211954, 'samples': 3644416, 'steps': 7117, 'loss/train': 0.3638428747653961} 02/24/2022 10:47:33 - INFO - codeparrot_training - Step 7118: {'lr': 0.0004861048526724833, 'samples': 3644928, 'steps': 7118, 'loss/train': 2.4352030754089355} 02/24/2022 10:47:39 - INFO - codeparrot_training - Step 7119: {'lr': 0.00048609947312161435, 'samples': 3645440, 'steps': 7119, 'loss/train': 2.163443088531494} 02/24/2022 10:47:42 - INFO - codeparrot_training - Step 7120: {'lr': 0.0004860940925593703, 'samples': 3645952, 'steps': 7120, 'loss/train': 1.8580036163330078} 02/24/2022 10:47:48 - INFO - codeparrot_training - Step 7121: {'lr': 0.0004860887109857741, 'samples': 3646464, 'steps': 7121, 'loss/train': 1.9512933492660522} 02/24/2022 10:47:51 - INFO - codeparrot_training - Step 7122: {'lr': 0.0004860833284008488, 'samples': 3646976, 'steps': 7122, 'loss/train': 1.955841302871704} 02/24/2022 10:47:57 - INFO - codeparrot_training - Step 7123: {'lr': 0.00048607794480461753, 'samples': 3647488, 'steps': 7123, 'loss/train': 1.6247124671936035} 02/24/2022 10:48:00 - INFO - codeparrot_training - Step 7124: {'lr': 0.00048607256019710327, 'samples': 3648000, 'steps': 7124, 'loss/train': 2.3678576946258545} 02/24/2022 10:48:07 - INFO - codeparrot_training - Step 7125: {'lr': 0.0004860671745783292, 'samples': 3648512, 'steps': 7125, 'loss/train': 2.619330644607544} 02/24/2022 10:48:10 - INFO - codeparrot_training - Step 7126: {'lr': 0.0004860617879483182, 'samples': 3649024, 'steps': 7126, 'loss/train': 2.1089675426483154} 02/24/2022 10:48:14 - INFO - codeparrot_training - Step 7127: {'lr': 0.0004860564003070935, 'samples': 3649536, 'steps': 7127, 'loss/train': 2.939650774002075} 02/24/2022 10:48:20 - INFO - codeparrot_training - Step 7128: {'lr': 0.00048605101165467813, 'samples': 3650048, 'steps': 7128, 'loss/train': 3.040771484375} 02/24/2022 10:48:23 - INFO - codeparrot_training - Step 7129: {'lr': 0.00048604562199109524, 'samples': 3650560, 'steps': 7129, 'loss/train': 2.6924564838409424} 02/24/2022 10:48:29 - INFO - codeparrot_training - Step 7130: {'lr': 0.00048604023131636784, 'samples': 3651072, 'steps': 7130, 'loss/train': 2.4744179248809814} 02/24/2022 10:48:32 - INFO - codeparrot_training - Step 7131: {'lr': 0.00048603483963051896, 'samples': 3651584, 'steps': 7131, 'loss/train': 2.6079609394073486} 02/24/2022 10:48:38 - INFO - codeparrot_training - Step 7132: {'lr': 0.0004860294469335719, 'samples': 3652096, 'steps': 7132, 'loss/train': 1.791207194328308} 02/24/2022 10:48:41 - INFO - codeparrot_training - Step 7133: {'lr': 0.00048602405322554956, 'samples': 3652608, 'steps': 7133, 'loss/train': 2.0393855571746826} 02/24/2022 10:48:49 - INFO - codeparrot_training - Step 7134: {'lr': 0.00048601865850647516, 'samples': 3653120, 'steps': 7134, 'loss/train': 1.391977310180664} 02/24/2022 10:48:52 - INFO - codeparrot_training - Step 7135: {'lr': 0.0004860132627763717, 'samples': 3653632, 'steps': 7135, 'loss/train': 1.6162645816802979} 02/24/2022 10:48:58 - INFO - codeparrot_training - Step 7136: {'lr': 0.0004860078660352625, 'samples': 3654144, 'steps': 7136, 'loss/train': 2.4756364822387695} 02/24/2022 10:49:01 - INFO - codeparrot_training - Step 7137: {'lr': 0.0004860024682831704, 'samples': 3654656, 'steps': 7137, 'loss/train': 1.660615086555481} 02/24/2022 10:49:07 - INFO - codeparrot_training - Step 7138: {'lr': 0.0004859970695201187, 'samples': 3655168, 'steps': 7138, 'loss/train': 2.9241085052490234} 02/24/2022 10:49:11 - INFO - codeparrot_training - Step 7139: {'lr': 0.00048599166974613053, 'samples': 3655680, 'steps': 7139, 'loss/train': 1.275688886642456} 02/24/2022 10:49:16 - INFO - codeparrot_training - Step 7140: {'lr': 0.000485986268961229, 'samples': 3656192, 'steps': 7140, 'loss/train': 2.078481912612915} 02/24/2022 10:49:20 - INFO - codeparrot_training - Step 7141: {'lr': 0.0004859808671654372, 'samples': 3656704, 'steps': 7141, 'loss/train': 4.897994518280029} 02/24/2022 10:49:27 - INFO - codeparrot_training - Step 7142: {'lr': 0.00048597546435877824, 'samples': 3657216, 'steps': 7142, 'loss/train': 1.9437363147735596} 02/24/2022 10:49:30 - INFO - codeparrot_training - Step 7143: {'lr': 0.0004859700605412754, 'samples': 3657728, 'steps': 7143, 'loss/train': 2.2802278995513916} 02/24/2022 10:49:36 - INFO - codeparrot_training - Step 7144: {'lr': 0.0004859646557129517, 'samples': 3658240, 'steps': 7144, 'loss/train': 2.515331506729126} 02/24/2022 10:49:39 - INFO - codeparrot_training - Step 7145: {'lr': 0.0004859592498738304, 'samples': 3658752, 'steps': 7145, 'loss/train': 0.2366783618927002} 02/24/2022 10:49:45 - INFO - codeparrot_training - Step 7146: {'lr': 0.00048595384302393453, 'samples': 3659264, 'steps': 7146, 'loss/train': 3.8160417079925537} 02/24/2022 10:49:48 - INFO - codeparrot_training - Step 7147: {'lr': 0.00048594843516328734, 'samples': 3659776, 'steps': 7147, 'loss/train': 1.748029351234436} 02/24/2022 10:49:54 - INFO - codeparrot_training - Step 7148: {'lr': 0.000485943026291912, 'samples': 3660288, 'steps': 7148, 'loss/train': 1.8070690631866455} 02/24/2022 10:49:57 - INFO - codeparrot_training - Step 7149: {'lr': 0.0004859376164098317, 'samples': 3660800, 'steps': 7149, 'loss/train': 3.185234785079956} 02/24/2022 10:50:03 - INFO - codeparrot_training - Step 7150: {'lr': 0.0004859322055170695, 'samples': 3661312, 'steps': 7150, 'loss/train': 2.15921688079834} 02/24/2022 10:50:06 - INFO - codeparrot_training - Step 7151: {'lr': 0.00048592679361364867, 'samples': 3661824, 'steps': 7151, 'loss/train': 2.3122177124023438} 02/24/2022 10:50:12 - INFO - codeparrot_training - Step 7152: {'lr': 0.00048592138069959235, 'samples': 3662336, 'steps': 7152, 'loss/train': 2.613693952560425} 02/24/2022 10:50:16 - INFO - codeparrot_training - Step 7153: {'lr': 0.0004859159667749238, 'samples': 3662848, 'steps': 7153, 'loss/train': 2.06919002532959} 02/24/2022 10:50:23 - INFO - codeparrot_training - Step 7154: {'lr': 0.000485910551839666, 'samples': 3663360, 'steps': 7154, 'loss/train': 3.6004843711853027} 02/24/2022 10:50:26 - INFO - codeparrot_training - Step 7155: {'lr': 0.0004859051358938425, 'samples': 3663872, 'steps': 7155, 'loss/train': 1.2813680171966553} 02/24/2022 10:50:32 - INFO - codeparrot_training - Step 7156: {'lr': 0.00048589971893747626, 'samples': 3664384, 'steps': 7156, 'loss/train': 2.6149206161499023} 02/24/2022 10:50:35 - INFO - codeparrot_training - Step 7157: {'lr': 0.0004858943009705905, 'samples': 3664896, 'steps': 7157, 'loss/train': 3.144535541534424} 02/24/2022 10:50:41 - INFO - codeparrot_training - Step 7158: {'lr': 0.00048588888199320847, 'samples': 3665408, 'steps': 7158, 'loss/train': 1.6206649541854858} 02/24/2022 10:50:44 - INFO - codeparrot_training - Step 7159: {'lr': 0.0004858834620053534, 'samples': 3665920, 'steps': 7159, 'loss/train': 2.4694418907165527} 02/24/2022 10:50:50 - INFO - codeparrot_training - Step 7160: {'lr': 0.0004858780410070484, 'samples': 3666432, 'steps': 7160, 'loss/train': 1.972827434539795} 02/24/2022 10:50:53 - INFO - codeparrot_training - Step 7161: {'lr': 0.0004858726189983168, 'samples': 3666944, 'steps': 7161, 'loss/train': 1.6937215328216553} 02/24/2022 10:50:59 - INFO - codeparrot_training - Step 7162: {'lr': 0.00048586719597918185, 'samples': 3667456, 'steps': 7162, 'loss/train': 1.9544861316680908} 02/24/2022 10:51:02 - INFO - codeparrot_training - Step 7163: {'lr': 0.0004858617719496667, 'samples': 3667968, 'steps': 7163, 'loss/train': 2.36794376373291} 02/24/2022 10:51:10 - INFO - codeparrot_training - Step 7164: {'lr': 0.0004858563469097946, 'samples': 3668480, 'steps': 7164, 'loss/train': 2.1638131141662598} 02/24/2022 10:51:13 - INFO - codeparrot_training - Step 7165: {'lr': 0.0004858509208595888, 'samples': 3668992, 'steps': 7165, 'loss/train': 2.0760090351104736} 02/24/2022 10:51:19 - INFO - codeparrot_training - Step 7166: {'lr': 0.0004858454937990726, 'samples': 3669504, 'steps': 7166, 'loss/train': 2.2443134784698486} 02/24/2022 10:51:22 - INFO - codeparrot_training - Step 7167: {'lr': 0.0004858400657282691, 'samples': 3670016, 'steps': 7167, 'loss/train': 2.4196226596832275} 02/24/2022 10:51:28 - INFO - codeparrot_training - Step 7168: {'lr': 0.00048583463664720174, 'samples': 3670528, 'steps': 7168, 'loss/train': 3.0884318351745605} 02/24/2022 10:51:31 - INFO - codeparrot_training - Step 7169: {'lr': 0.00048582920655589366, 'samples': 3671040, 'steps': 7169, 'loss/train': 2.4953513145446777} 02/24/2022 10:51:37 - INFO - codeparrot_training - Step 7170: {'lr': 0.0004858237754543681, 'samples': 3671552, 'steps': 7170, 'loss/train': 1.272633671760559} 02/24/2022 10:51:40 - INFO - codeparrot_training - Step 7171: {'lr': 0.0004858183433426484, 'samples': 3672064, 'steps': 7171, 'loss/train': 2.6103909015655518} 02/24/2022 10:51:46 - INFO - codeparrot_training - Step 7172: {'lr': 0.0004858129102207578, 'samples': 3672576, 'steps': 7172, 'loss/train': 2.377828359603882} 02/24/2022 10:51:49 - INFO - codeparrot_training - Step 7173: {'lr': 0.00048580747608871955, 'samples': 3673088, 'steps': 7173, 'loss/train': 3.352987766265869} 02/24/2022 10:51:55 - INFO - codeparrot_training - Step 7174: {'lr': 0.000485802040946557, 'samples': 3673600, 'steps': 7174, 'loss/train': 2.291076898574829} 02/24/2022 10:51:58 - INFO - codeparrot_training - Step 7175: {'lr': 0.00048579660479429335, 'samples': 3674112, 'steps': 7175, 'loss/train': 2.5165326595306396} 02/24/2022 10:52:04 - INFO - codeparrot_training - Step 7176: {'lr': 0.00048579116763195184, 'samples': 3674624, 'steps': 7176, 'loss/train': 3.4540293216705322} 02/24/2022 10:52:07 - INFO - codeparrot_training - Step 7177: {'lr': 0.00048578572945955594, 'samples': 3675136, 'steps': 7177, 'loss/train': 2.3993983268737793} 02/24/2022 10:52:13 - INFO - codeparrot_training - Step 7178: {'lr': 0.00048578029027712883, 'samples': 3675648, 'steps': 7178, 'loss/train': 2.3527095317840576} 02/24/2022 10:52:16 - INFO - codeparrot_training - Step 7179: {'lr': 0.0004857748500846938, 'samples': 3676160, 'steps': 7179, 'loss/train': 1.432336449623108} 02/24/2022 10:52:23 - INFO - codeparrot_training - Step 7180: {'lr': 0.0004857694088822742, 'samples': 3676672, 'steps': 7180, 'loss/train': 1.1556800603866577} 02/24/2022 10:52:27 - INFO - codeparrot_training - Step 7181: {'lr': 0.00048576396666989333, 'samples': 3677184, 'steps': 7181, 'loss/train': 1.8337124586105347} 02/24/2022 10:52:32 - INFO - codeparrot_training - Step 7182: {'lr': 0.0004857585234475745, 'samples': 3677696, 'steps': 7182, 'loss/train': 3.0596120357513428} 02/24/2022 10:52:36 - INFO - codeparrot_training - Step 7183: {'lr': 0.00048575307921534095, 'samples': 3678208, 'steps': 7183, 'loss/train': 1.1250718832015991} 02/24/2022 10:52:41 - INFO - codeparrot_training - Step 7184: {'lr': 0.0004857476339732161, 'samples': 3678720, 'steps': 7184, 'loss/train': 1.85762619972229} 02/24/2022 10:52:45 - INFO - codeparrot_training - Step 7185: {'lr': 0.0004857421877212233, 'samples': 3679232, 'steps': 7185, 'loss/train': 2.8390424251556396} 02/24/2022 10:52:50 - INFO - codeparrot_training - Step 7186: {'lr': 0.00048573674045938577, 'samples': 3679744, 'steps': 7186, 'loss/train': 2.185783624649048} 02/24/2022 10:52:54 - INFO - codeparrot_training - Step 7187: {'lr': 0.00048573129218772686, 'samples': 3680256, 'steps': 7187, 'loss/train': 2.4759747982025146} 02/24/2022 10:52:59 - INFO - codeparrot_training - Step 7188: {'lr': 0.00048572584290627, 'samples': 3680768, 'steps': 7188, 'loss/train': 2.599696397781372} 02/24/2022 10:53:03 - INFO - codeparrot_training - Step 7189: {'lr': 0.00048572039261503855, 'samples': 3681280, 'steps': 7189, 'loss/train': 2.3892900943756104} 02/24/2022 10:53:10 - INFO - codeparrot_training - Step 7190: {'lr': 0.00048571494131405567, 'samples': 3681792, 'steps': 7190, 'loss/train': 2.8847944736480713} 02/24/2022 10:53:13 - INFO - codeparrot_training - Step 7191: {'lr': 0.0004857094890033449, 'samples': 3682304, 'steps': 7191, 'loss/train': 2.7730712890625} 02/24/2022 10:53:19 - INFO - codeparrot_training - Step 7192: {'lr': 0.0004857040356829295, 'samples': 3682816, 'steps': 7192, 'loss/train': 2.233396291732788} 02/24/2022 10:53:23 - INFO - codeparrot_training - Step 7193: {'lr': 0.00048569858135283285, 'samples': 3683328, 'steps': 7193, 'loss/train': 1.2707794904708862} 02/24/2022 10:53:28 - INFO - codeparrot_training - Step 7194: {'lr': 0.00048569312601307827, 'samples': 3683840, 'steps': 7194, 'loss/train': 3.1571671962738037} 02/24/2022 10:53:32 - INFO - codeparrot_training - Step 7195: {'lr': 0.00048568766966368925, 'samples': 3684352, 'steps': 7195, 'loss/train': 2.1208324432373047} 02/24/2022 10:53:37 - INFO - codeparrot_training - Step 7196: {'lr': 0.00048568221230468905, 'samples': 3684864, 'steps': 7196, 'loss/train': 2.512352705001831} 02/24/2022 10:53:41 - INFO - codeparrot_training - Step 7197: {'lr': 0.0004856767539361011, 'samples': 3685376, 'steps': 7197, 'loss/train': 1.8583602905273438} 02/24/2022 10:53:46 - INFO - codeparrot_training - Step 7198: {'lr': 0.0004856712945579488, 'samples': 3685888, 'steps': 7198, 'loss/train': 1.7931087017059326} 02/24/2022 10:53:49 - INFO - codeparrot_training - Step 7199: {'lr': 0.00048566583417025553, 'samples': 3686400, 'steps': 7199, 'loss/train': 2.4703731536865234} 02/24/2022 10:53:57 - INFO - codeparrot_training - Step 7200: {'lr': 0.00048566037277304465, 'samples': 3686912, 'steps': 7200, 'loss/train': 2.562755823135376} 02/24/2022 10:54:00 - INFO - codeparrot_training - Step 7201: {'lr': 0.00048565491036633946, 'samples': 3687424, 'steps': 7201, 'loss/train': 3.199000358581543} 02/24/2022 10:54:06 - INFO - codeparrot_training - Step 7202: {'lr': 0.00048564944695016356, 'samples': 3687936, 'steps': 7202, 'loss/train': 2.8276875019073486} 02/24/2022 10:54:11 - INFO - codeparrot_training - Step 7203: {'lr': 0.00048564398252454026, 'samples': 3688448, 'steps': 7203, 'loss/train': 2.888237953186035} 02/24/2022 10:54:15 - INFO - codeparrot_training - Step 7204: {'lr': 0.0004856385170894929, 'samples': 3688960, 'steps': 7204, 'loss/train': 2.2383034229278564} 02/24/2022 10:54:20 - INFO - codeparrot_training - Step 7205: {'lr': 0.00048563305064504503, 'samples': 3689472, 'steps': 7205, 'loss/train': 1.4651906490325928} 02/24/2022 10:54:24 - INFO - codeparrot_training - Step 7206: {'lr': 0.00048562758319121996, 'samples': 3689984, 'steps': 7206, 'loss/train': 1.9588793516159058} 02/24/2022 10:54:29 - INFO - codeparrot_training - Step 7207: {'lr': 0.00048562211472804115, 'samples': 3690496, 'steps': 7207, 'loss/train': 1.9141972064971924} 02/24/2022 10:54:33 - INFO - codeparrot_training - Step 7208: {'lr': 0.000485616645255532, 'samples': 3691008, 'steps': 7208, 'loss/train': 2.395052671432495} 02/24/2022 10:54:40 - INFO - codeparrot_training - Step 7209: {'lr': 0.00048561117477371595, 'samples': 3691520, 'steps': 7209, 'loss/train': 2.777357339859009} 02/24/2022 10:54:44 - INFO - codeparrot_training - Step 7210: {'lr': 0.0004856057032826165, 'samples': 3692032, 'steps': 7210, 'loss/train': 1.645520567893982} 02/24/2022 10:54:49 - INFO - codeparrot_training - Step 7211: {'lr': 0.000485600230782257, 'samples': 3692544, 'steps': 7211, 'loss/train': 2.1502439975738525} 02/24/2022 10:54:53 - INFO - codeparrot_training - Step 7212: {'lr': 0.00048559475727266086, 'samples': 3693056, 'steps': 7212, 'loss/train': 2.699481725692749} 02/24/2022 10:54:58 - INFO - codeparrot_training - Step 7213: {'lr': 0.00048558928275385167, 'samples': 3693568, 'steps': 7213, 'loss/train': 1.5347621440887451} 02/24/2022 10:55:02 - INFO - codeparrot_training - Step 7214: {'lr': 0.00048558380722585283, 'samples': 3694080, 'steps': 7214, 'loss/train': 3.3592445850372314} 02/24/2022 10:55:07 - INFO - codeparrot_training - Step 7215: {'lr': 0.00048557833068868766, 'samples': 3694592, 'steps': 7215, 'loss/train': 2.003713846206665} 02/24/2022 10:55:11 - INFO - codeparrot_training - Step 7216: {'lr': 0.00048557285314237975, 'samples': 3695104, 'steps': 7216, 'loss/train': 2.9488580226898193} 02/24/2022 10:55:16 - INFO - codeparrot_training - Step 7217: {'lr': 0.0004855673745869526, 'samples': 3695616, 'steps': 7217, 'loss/train': 2.2138144969940186} 02/24/2022 10:55:20 - INFO - codeparrot_training - Step 7218: {'lr': 0.00048556189502242956, 'samples': 3696128, 'steps': 7218, 'loss/train': 2.137820243835449} 02/24/2022 10:55:25 - INFO - codeparrot_training - Step 7219: {'lr': 0.00048555641444883424, 'samples': 3696640, 'steps': 7219, 'loss/train': 2.91664981842041} 02/24/2022 10:55:28 - INFO - codeparrot_training - Step 7220: {'lr': 0.00048555093286618996, 'samples': 3697152, 'steps': 7220, 'loss/train': 2.5578601360321045} 02/24/2022 10:55:34 - INFO - codeparrot_training - Step 7221: {'lr': 0.00048554545027452035, 'samples': 3697664, 'steps': 7221, 'loss/train': 2.4842004776000977} 02/24/2022 10:55:38 - INFO - codeparrot_training - Step 7222: {'lr': 0.00048553996667384877, 'samples': 3698176, 'steps': 7222, 'loss/train': 2.053813934326172} 02/24/2022 10:55:43 - INFO - codeparrot_training - Step 7223: {'lr': 0.00048553448206419876, 'samples': 3698688, 'steps': 7223, 'loss/train': 2.055654287338257} 02/24/2022 10:55:47 - INFO - codeparrot_training - Step 7224: {'lr': 0.0004855289964455938, 'samples': 3699200, 'steps': 7224, 'loss/train': 2.218465566635132} 02/24/2022 10:55:54 - INFO - codeparrot_training - Step 7225: {'lr': 0.0004855235098180575, 'samples': 3699712, 'steps': 7225, 'loss/train': 1.195311427116394} 02/24/2022 10:55:57 - INFO - codeparrot_training - Step 7226: {'lr': 0.00048551802218161315, 'samples': 3700224, 'steps': 7226, 'loss/train': 2.4262242317199707} 02/24/2022 10:56:04 - INFO - codeparrot_training - Step 7227: {'lr': 0.00048551253353628444, 'samples': 3700736, 'steps': 7227, 'loss/train': 3.3374390602111816} 02/24/2022 10:56:07 - INFO - codeparrot_training - Step 7228: {'lr': 0.0004855070438820949, 'samples': 3701248, 'steps': 7228, 'loss/train': 2.7477335929870605} 02/24/2022 10:56:11 - INFO - codeparrot_training - Step 7229: {'lr': 0.0004855015532190679, 'samples': 3701760, 'steps': 7229, 'loss/train': 2.0309486389160156} 02/24/2022 10:56:16 - INFO - codeparrot_training - Step 7230: {'lr': 0.0004854960615472269, 'samples': 3702272, 'steps': 7230, 'loss/train': 2.0270018577575684} 02/24/2022 10:56:20 - INFO - codeparrot_training - Step 7231: {'lr': 0.0004854905688665957, 'samples': 3702784, 'steps': 7231, 'loss/train': 1.3398030996322632} 02/24/2022 10:56:25 - INFO - codeparrot_training - Step 7232: {'lr': 0.00048548507517719766, 'samples': 3703296, 'steps': 7232, 'loss/train': 2.0843541622161865} 02/24/2022 10:56:29 - INFO - codeparrot_training - Step 7233: {'lr': 0.00048547958047905635, 'samples': 3703808, 'steps': 7233, 'loss/train': 2.3666794300079346} 02/24/2022 10:56:34 - INFO - codeparrot_training - Step 7234: {'lr': 0.00048547408477219524, 'samples': 3704320, 'steps': 7234, 'loss/train': 2.985812187194824} 02/24/2022 10:56:38 - INFO - codeparrot_training - Step 7235: {'lr': 0.00048546858805663797, 'samples': 3704832, 'steps': 7235, 'loss/train': 1.1768949031829834} 02/24/2022 10:56:43 - INFO - codeparrot_training - Step 7236: {'lr': 0.000485463090332408, 'samples': 3705344, 'steps': 7236, 'loss/train': 3.4861700534820557} 02/24/2022 10:56:47 - INFO - codeparrot_training - Step 7237: {'lr': 0.0004854575915995289, 'samples': 3705856, 'steps': 7237, 'loss/train': 1.7150379419326782} 02/24/2022 10:56:54 - INFO - codeparrot_training - Step 7238: {'lr': 0.0004854520918580243, 'samples': 3706368, 'steps': 7238, 'loss/train': 2.8902151584625244} 02/24/2022 10:56:57 - INFO - codeparrot_training - Step 7239: {'lr': 0.00048544659110791766, 'samples': 3706880, 'steps': 7239, 'loss/train': 2.8760859966278076} 02/24/2022 10:57:03 - INFO - codeparrot_training - Step 7240: {'lr': 0.0004854410893492326, 'samples': 3707392, 'steps': 7240, 'loss/train': 2.2347939014434814} 02/24/2022 10:57:06 - INFO - codeparrot_training - Step 7241: {'lr': 0.00048543558658199266, 'samples': 3707904, 'steps': 7241, 'loss/train': 2.0554895401000977} 02/24/2022 10:57:12 - INFO - codeparrot_training - Step 7242: {'lr': 0.0004854300828062215, 'samples': 3708416, 'steps': 7242, 'loss/train': 2.3495774269104004} 02/24/2022 10:57:15 - INFO - codeparrot_training - Step 7243: {'lr': 0.0004854245780219425, 'samples': 3708928, 'steps': 7243, 'loss/train': 1.511998176574707} 02/24/2022 10:57:21 - INFO - codeparrot_training - Step 7244: {'lr': 0.00048541907222917946, 'samples': 3709440, 'steps': 7244, 'loss/train': 0.863852322101593} 02/24/2022 10:57:24 - INFO - codeparrot_training - Step 7245: {'lr': 0.0004854135654279558, 'samples': 3709952, 'steps': 7245, 'loss/train': 2.0903382301330566} 02/24/2022 10:57:32 - INFO - codeparrot_training - Step 7246: {'lr': 0.0004854080576182952, 'samples': 3710464, 'steps': 7246, 'loss/train': 2.5613043308258057} 02/24/2022 10:57:35 - INFO - codeparrot_training - Step 7247: {'lr': 0.00048540254880022126, 'samples': 3710976, 'steps': 7247, 'loss/train': 2.1796469688415527} 02/24/2022 10:57:41 - INFO - codeparrot_training - Step 7248: {'lr': 0.00048539703897375753, 'samples': 3711488, 'steps': 7248, 'loss/train': 2.8391802310943604} 02/24/2022 10:57:44 - INFO - codeparrot_training - Step 7249: {'lr': 0.0004853915281389276, 'samples': 3712000, 'steps': 7249, 'loss/train': 2.178542137145996} 02/24/2022 10:57:50 - INFO - codeparrot_training - Step 7250: {'lr': 0.0004853860162957552, 'samples': 3712512, 'steps': 7250, 'loss/train': 2.441507339477539} 02/24/2022 10:57:53 - INFO - codeparrot_training - Step 7251: {'lr': 0.00048538050344426375, 'samples': 3713024, 'steps': 7251, 'loss/train': 1.7413239479064941} 02/24/2022 10:57:59 - INFO - codeparrot_training - Step 7252: {'lr': 0.0004853749895844771, 'samples': 3713536, 'steps': 7252, 'loss/train': 2.78940486907959} 02/24/2022 10:58:02 - INFO - codeparrot_training - Step 7253: {'lr': 0.00048536947471641855, 'samples': 3714048, 'steps': 7253, 'loss/train': 3.2126731872558594} 02/24/2022 10:58:08 - INFO - codeparrot_training - Step 7254: {'lr': 0.00048536395884011207, 'samples': 3714560, 'steps': 7254, 'loss/train': 8.389498710632324} 02/24/2022 10:58:11 - INFO - codeparrot_training - Step 7255: {'lr': 0.00048535844195558104, 'samples': 3715072, 'steps': 7255, 'loss/train': 1.8467371463775635} 02/24/2022 10:58:18 - INFO - codeparrot_training - Step 7256: {'lr': 0.0004853529240628493, 'samples': 3715584, 'steps': 7256, 'loss/train': 1.7006131410598755} 02/24/2022 10:58:22 - INFO - codeparrot_training - Step 7257: {'lr': 0.0004853474051619402, 'samples': 3716096, 'steps': 7257, 'loss/train': 2.6924703121185303} 02/24/2022 10:58:27 - INFO - codeparrot_training - Step 7258: {'lr': 0.0004853418852528776, 'samples': 3716608, 'steps': 7258, 'loss/train': 2.2776970863342285} 02/24/2022 10:58:31 - INFO - codeparrot_training - Step 7259: {'lr': 0.00048533636433568505, 'samples': 3717120, 'steps': 7259, 'loss/train': 1.6656603813171387} 02/24/2022 10:58:36 - INFO - codeparrot_training - Step 7260: {'lr': 0.00048533084241038637, 'samples': 3717632, 'steps': 7260, 'loss/train': 2.825521230697632} 02/24/2022 10:58:40 - INFO - codeparrot_training - Step 7261: {'lr': 0.00048532531947700496, 'samples': 3718144, 'steps': 7261, 'loss/train': 2.580369710922241} 02/24/2022 10:58:45 - INFO - codeparrot_training - Step 7262: {'lr': 0.00048531979553556473, 'samples': 3718656, 'steps': 7262, 'loss/train': 0.9673537015914917} 02/24/2022 10:58:49 - INFO - codeparrot_training - Step 7263: {'lr': 0.0004853142705860891, 'samples': 3719168, 'steps': 7263, 'loss/train': 1.1317963600158691} 02/24/2022 10:58:54 - INFO - codeparrot_training - Step 7264: {'lr': 0.00048530874462860194, 'samples': 3719680, 'steps': 7264, 'loss/train': 1.9410359859466553} 02/24/2022 10:59:00 - INFO - codeparrot_training - Step 7265: {'lr': 0.0004853032176631268, 'samples': 3720192, 'steps': 7265, 'loss/train': 2.244546413421631} 02/24/2022 10:59:03 - INFO - codeparrot_training - Step 7266: {'lr': 0.0004852976896896874, 'samples': 3720704, 'steps': 7266, 'loss/train': 1.9455451965332031} 02/24/2022 10:59:09 - INFO - codeparrot_training - Step 7267: {'lr': 0.0004852921607083074, 'samples': 3721216, 'steps': 7267, 'loss/train': 1.6454341411590576} 02/24/2022 10:59:12 - INFO - codeparrot_training - Step 7268: {'lr': 0.00048528663071901047, 'samples': 3721728, 'steps': 7268, 'loss/train': 2.1452298164367676} 02/24/2022 10:59:18 - INFO - codeparrot_training - Step 7269: {'lr': 0.00048528109972182043, 'samples': 3722240, 'steps': 7269, 'loss/train': 1.0160413980484009} 02/24/2022 10:59:21 - INFO - codeparrot_training - Step 7270: {'lr': 0.0004852755677167607, 'samples': 3722752, 'steps': 7270, 'loss/train': 2.9078474044799805} 02/24/2022 10:59:29 - INFO - codeparrot_training - Step 7271: {'lr': 0.00048527003470385534, 'samples': 3723264, 'steps': 7271, 'loss/train': 2.056849479675293} 02/24/2022 10:59:32 - INFO - codeparrot_training - Step 7272: {'lr': 0.0004852645006831278, 'samples': 3723776, 'steps': 7272, 'loss/train': 2.4244182109832764} 02/24/2022 10:59:38 - INFO - codeparrot_training - Step 7273: {'lr': 0.00048525896565460177, 'samples': 3724288, 'steps': 7273, 'loss/train': 2.0066919326782227} 02/24/2022 10:59:41 - INFO - codeparrot_training - Step 7274: {'lr': 0.00048525342961830106, 'samples': 3724800, 'steps': 7274, 'loss/train': 2.268441677093506} 02/24/2022 10:59:47 - INFO - codeparrot_training - Step 7275: {'lr': 0.0004852478925742494, 'samples': 3725312, 'steps': 7275, 'loss/train': 3.278693437576294} 02/24/2022 10:59:50 - INFO - codeparrot_training - Step 7276: {'lr': 0.0004852423545224704, 'samples': 3725824, 'steps': 7276, 'loss/train': 1.9232362508773804} 02/24/2022 10:59:56 - INFO - codeparrot_training - Step 7277: {'lr': 0.00048523681546298793, 'samples': 3726336, 'steps': 7277, 'loss/train': 1.712380051612854} 02/24/2022 10:59:59 - INFO - codeparrot_training - Step 7278: {'lr': 0.0004852312753958256, 'samples': 3726848, 'steps': 7278, 'loss/train': 3.691619873046875} 02/24/2022 11:00:05 - INFO - codeparrot_training - Step 7279: {'lr': 0.00048522573432100715, 'samples': 3727360, 'steps': 7279, 'loss/train': 1.2366740703582764} 02/24/2022 11:00:08 - INFO - codeparrot_training - Step 7280: {'lr': 0.0004852201922385564, 'samples': 3727872, 'steps': 7280, 'loss/train': 3.183555841445923} 02/24/2022 11:00:14 - INFO - codeparrot_training - Step 7281: {'lr': 0.000485214649148497, 'samples': 3728384, 'steps': 7281, 'loss/train': 2.197190523147583} 02/24/2022 11:00:17 - INFO - codeparrot_training - Step 7282: {'lr': 0.00048520910505085274, 'samples': 3728896, 'steps': 7282, 'loss/train': 1.9557613134384155} 02/24/2022 11:00:24 - INFO - codeparrot_training - Step 7283: {'lr': 0.0004852035599456474, 'samples': 3729408, 'steps': 7283, 'loss/train': 2.2826309204101562} 02/24/2022 11:00:28 - INFO - codeparrot_training - Step 7284: {'lr': 0.0004851980138329046, 'samples': 3729920, 'steps': 7284, 'loss/train': 2.773719549179077} 02/24/2022 11:00:33 - INFO - codeparrot_training - Step 7285: {'lr': 0.00048519246671264825, 'samples': 3730432, 'steps': 7285, 'loss/train': 3.537505626678467} 02/24/2022 11:00:37 - INFO - codeparrot_training - Step 7286: {'lr': 0.0004851869185849021, 'samples': 3730944, 'steps': 7286, 'loss/train': 2.2186882495880127} 02/24/2022 11:00:42 - INFO - codeparrot_training - Step 7287: {'lr': 0.0004851813694496898, 'samples': 3731456, 'steps': 7287, 'loss/train': 2.055197238922119} 02/24/2022 11:00:46 - INFO - codeparrot_training - Step 7288: {'lr': 0.00048517581930703526, 'samples': 3731968, 'steps': 7288, 'loss/train': 1.3483860492706299} 02/24/2022 11:00:51 - INFO - codeparrot_training - Step 7289: {'lr': 0.0004851702681569621, 'samples': 3732480, 'steps': 7289, 'loss/train': 3.1877801418304443} 02/24/2022 11:00:55 - INFO - codeparrot_training - Step 7290: {'lr': 0.0004851647159994943, 'samples': 3732992, 'steps': 7290, 'loss/train': 2.2647533416748047} 02/24/2022 11:01:00 - INFO - codeparrot_training - Step 7291: {'lr': 0.00048515916283465546, 'samples': 3733504, 'steps': 7291, 'loss/train': 2.734210729598999} 02/24/2022 11:01:04 - INFO - codeparrot_training - Step 7292: {'lr': 0.00048515360866246943, 'samples': 3734016, 'steps': 7292, 'loss/train': 2.4495162963867188} 02/24/2022 11:01:11 - INFO - codeparrot_training - Step 7293: {'lr': 0.00048514805348296, 'samples': 3734528, 'steps': 7293, 'loss/train': 1.844407320022583} 02/24/2022 11:01:14 - INFO - codeparrot_training - Step 7294: {'lr': 0.000485142497296151, 'samples': 3735040, 'steps': 7294, 'loss/train': 4.059875011444092} 02/24/2022 11:01:20 - INFO - codeparrot_training - Step 7295: {'lr': 0.00048513694010206623, 'samples': 3735552, 'steps': 7295, 'loss/train': 1.5836923122406006} 02/24/2022 11:01:23 - INFO - codeparrot_training - Step 7296: {'lr': 0.0004851313819007295, 'samples': 3736064, 'steps': 7296, 'loss/train': 2.7153892517089844} 02/24/2022 11:01:29 - INFO - codeparrot_training - Step 7297: {'lr': 0.0004851258226921645, 'samples': 3736576, 'steps': 7297, 'loss/train': 2.5042052268981934} 02/24/2022 11:01:32 - INFO - codeparrot_training - Step 7298: {'lr': 0.0004851202624763952, 'samples': 3737088, 'steps': 7298, 'loss/train': 2.5344438552856445} 02/24/2022 11:01:38 - INFO - codeparrot_training - Step 7299: {'lr': 0.0004851147012534453, 'samples': 3737600, 'steps': 7299, 'loss/train': 2.899864673614502} 02/24/2022 11:01:41 - INFO - codeparrot_training - Step 7300: {'lr': 0.00048510913902333875, 'samples': 3738112, 'steps': 7300, 'loss/train': 1.7969651222229004} 02/24/2022 11:01:47 - INFO - codeparrot_training - Step 7301: {'lr': 0.0004851035757860992, 'samples': 3738624, 'steps': 7301, 'loss/train': 3.2442026138305664} 02/24/2022 11:01:50 - INFO - codeparrot_training - Step 7302: {'lr': 0.0004850980115417507, 'samples': 3739136, 'steps': 7302, 'loss/train': 2.2781991958618164} 02/24/2022 11:01:58 - INFO - codeparrot_training - Step 7303: {'lr': 0.0004850924462903169, 'samples': 3739648, 'steps': 7303, 'loss/train': 2.26617169380188} 02/24/2022 11:02:01 - INFO - codeparrot_training - Step 7304: {'lr': 0.0004850868800318218, 'samples': 3740160, 'steps': 7304, 'loss/train': 2.0905508995056152} 02/24/2022 11:02:07 - INFO - codeparrot_training - Step 7305: {'lr': 0.00048508131276628905, 'samples': 3740672, 'steps': 7305, 'loss/train': 1.6095443964004517} 02/24/2022 11:02:10 - INFO - codeparrot_training - Step 7306: {'lr': 0.0004850757444937426, 'samples': 3741184, 'steps': 7306, 'loss/train': 2.637087821960449} 02/24/2022 11:02:16 - INFO - codeparrot_training - Step 7307: {'lr': 0.00048507017521420636, 'samples': 3741696, 'steps': 7307, 'loss/train': 1.8647308349609375} 02/24/2022 11:02:19 - INFO - codeparrot_training - Step 7308: {'lr': 0.0004850646049277041, 'samples': 3742208, 'steps': 7308, 'loss/train': 2.1614322662353516} 02/24/2022 11:02:25 - INFO - codeparrot_training - Step 7309: {'lr': 0.00048505903363425974, 'samples': 3742720, 'steps': 7309, 'loss/train': 2.0116732120513916} 02/24/2022 11:02:28 - INFO - codeparrot_training - Step 7310: {'lr': 0.0004850534613338972, 'samples': 3743232, 'steps': 7310, 'loss/train': 4.1565775871276855} 02/24/2022 11:02:34 - INFO - codeparrot_training - Step 7311: {'lr': 0.00048504788802664013, 'samples': 3743744, 'steps': 7311, 'loss/train': 2.4235846996307373} 02/24/2022 11:02:37 - INFO - codeparrot_training - Step 7312: {'lr': 0.00048504231371251255, 'samples': 3744256, 'steps': 7312, 'loss/train': 1.5881248712539673} 02/24/2022 11:02:43 - INFO - codeparrot_training - Step 7313: {'lr': 0.0004850367383915384, 'samples': 3744768, 'steps': 7313, 'loss/train': 1.9319747686386108} 02/24/2022 11:02:46 - INFO - codeparrot_training - Step 7314: {'lr': 0.00048503116206374147, 'samples': 3745280, 'steps': 7314, 'loss/train': 4.7429680824279785} 02/24/2022 11:02:52 - INFO - codeparrot_training - Step 7315: {'lr': 0.00048502558472914573, 'samples': 3745792, 'steps': 7315, 'loss/train': 0.5526663064956665} 02/24/2022 11:02:55 - INFO - codeparrot_training - Step 7316: {'lr': 0.00048502000638777487, 'samples': 3746304, 'steps': 7316, 'loss/train': 3.684763193130493} 02/24/2022 11:03:01 - INFO - codeparrot_training - Step 7317: {'lr': 0.000485014427039653, 'samples': 3746816, 'steps': 7317, 'loss/train': 2.744223117828369} 02/24/2022 11:03:04 - INFO - codeparrot_training - Step 7318: {'lr': 0.00048500884668480407, 'samples': 3747328, 'steps': 7318, 'loss/train': 2.4987661838531494} 02/24/2022 11:03:12 - INFO - codeparrot_training - Step 7319: {'lr': 0.00048500326532325167, 'samples': 3747840, 'steps': 7319, 'loss/train': 2.580307960510254} 02/24/2022 11:03:15 - INFO - codeparrot_training - Step 7320: {'lr': 0.00048499768295502, 'samples': 3748352, 'steps': 7320, 'loss/train': 3.0824406147003174} 02/24/2022 11:03:21 - INFO - codeparrot_training - Step 7321: {'lr': 0.0004849920995801329, 'samples': 3748864, 'steps': 7321, 'loss/train': 2.096445322036743} 02/24/2022 11:03:24 - INFO - codeparrot_training - Step 7322: {'lr': 0.00048498651519861426, 'samples': 3749376, 'steps': 7322, 'loss/train': 2.05833101272583} 02/24/2022 11:03:30 - INFO - codeparrot_training - Step 7323: {'lr': 0.00048498092981048797, 'samples': 3749888, 'steps': 7323, 'loss/train': 1.8068619966506958} 02/24/2022 11:03:33 - INFO - codeparrot_training - Step 7324: {'lr': 0.000484975343415778, 'samples': 3750400, 'steps': 7324, 'loss/train': 2.3918323516845703} 02/24/2022 11:03:39 - INFO - codeparrot_training - Step 7325: {'lr': 0.00048496975601450835, 'samples': 3750912, 'steps': 7325, 'loss/train': 2.4769656658172607} 02/24/2022 11:03:42 - INFO - codeparrot_training - Step 7326: {'lr': 0.0004849641676067027, 'samples': 3751424, 'steps': 7326, 'loss/train': 2.3107967376708984} 02/24/2022 11:03:48 - INFO - codeparrot_training - Step 7327: {'lr': 0.0004849585781923853, 'samples': 3751936, 'steps': 7327, 'loss/train': 2.417308807373047} 02/24/2022 11:03:51 - INFO - codeparrot_training - Step 7328: {'lr': 0.00048495298777157994, 'samples': 3752448, 'steps': 7328, 'loss/train': 0.8975260853767395} 02/24/2022 11:03:59 - INFO - codeparrot_training - Step 7329: {'lr': 0.00048494739634431057, 'samples': 3752960, 'steps': 7329, 'loss/train': 1.2878053188323975} 02/24/2022 11:04:04 - INFO - codeparrot_training - Step 7330: {'lr': 0.00048494180391060114, 'samples': 3753472, 'steps': 7330, 'loss/train': 2.769775390625} 02/24/2022 11:04:08 - INFO - codeparrot_training - Step 7331: {'lr': 0.0004849362104704756, 'samples': 3753984, 'steps': 7331, 'loss/train': 1.1494196653366089} 02/24/2022 11:04:13 - INFO - codeparrot_training - Step 7332: {'lr': 0.00048493061602395803, 'samples': 3754496, 'steps': 7332, 'loss/train': 2.467677354812622} 02/24/2022 11:04:17 - INFO - codeparrot_training - Step 7333: {'lr': 0.0004849250205710722, 'samples': 3755008, 'steps': 7333, 'loss/train': 1.2712596654891968} 02/24/2022 11:04:22 - INFO - codeparrot_training - Step 7334: {'lr': 0.0004849194241118423, 'samples': 3755520, 'steps': 7334, 'loss/train': 0.40487274527549744} 02/24/2022 11:04:26 - INFO - codeparrot_training - Step 7335: {'lr': 0.0004849138266462921, 'samples': 3756032, 'steps': 7335, 'loss/train': 1.7233330011367798} 02/24/2022 11:04:31 - INFO - codeparrot_training - Step 7336: {'lr': 0.0004849082281744457, 'samples': 3756544, 'steps': 7336, 'loss/train': 2.4098403453826904} 02/24/2022 11:04:35 - INFO - codeparrot_training - Step 7337: {'lr': 0.00048490262869632693, 'samples': 3757056, 'steps': 7337, 'loss/train': 2.3194680213928223} 02/24/2022 11:04:42 - INFO - codeparrot_training - Step 7338: {'lr': 0.00048489702821196003, 'samples': 3757568, 'steps': 7338, 'loss/train': 1.9878824949264526} 02/24/2022 11:04:46 - INFO - codeparrot_training - Step 7339: {'lr': 0.0004848914267213688, 'samples': 3758080, 'steps': 7339, 'loss/train': 2.571725845336914} 02/24/2022 11:04:51 - INFO - codeparrot_training - Step 7340: {'lr': 0.00048488582422457726, 'samples': 3758592, 'steps': 7340, 'loss/train': 2.17806077003479} 02/24/2022 11:04:54 - INFO - codeparrot_training - Step 7341: {'lr': 0.0004848802207216094, 'samples': 3759104, 'steps': 7341, 'loss/train': 2.1164746284484863} 02/24/2022 11:05:00 - INFO - codeparrot_training - Step 7342: {'lr': 0.0004848746162124894, 'samples': 3759616, 'steps': 7342, 'loss/train': 1.9124101400375366} 02/24/2022 11:05:03 - INFO - codeparrot_training - Step 7343: {'lr': 0.00048486901069724097, 'samples': 3760128, 'steps': 7343, 'loss/train': 2.7095611095428467} 02/24/2022 11:05:09 - INFO - codeparrot_training - Step 7344: {'lr': 0.0004848634041758884, 'samples': 3760640, 'steps': 7344, 'loss/train': 1.5187373161315918} 02/24/2022 11:05:12 - INFO - codeparrot_training - Step 7345: {'lr': 0.00048485779664845553, 'samples': 3761152, 'steps': 7345, 'loss/train': 2.075495481491089} 02/24/2022 11:05:18 - INFO - codeparrot_training - Step 7346: {'lr': 0.0004848521881149664, 'samples': 3761664, 'steps': 7346, 'loss/train': 2.744697332382202} 02/24/2022 11:05:21 - INFO - codeparrot_training - Step 7347: {'lr': 0.00048484657857544513, 'samples': 3762176, 'steps': 7347, 'loss/train': 1.3915081024169922} 02/24/2022 11:05:29 - INFO - codeparrot_training - Step 7348: {'lr': 0.0004848409680299156, 'samples': 3762688, 'steps': 7348, 'loss/train': 1.8964418172836304} 02/24/2022 11:05:32 - INFO - codeparrot_training - Step 7349: {'lr': 0.00048483535647840206, 'samples': 3763200, 'steps': 7349, 'loss/train': 2.719987154006958} 02/24/2022 11:05:38 - INFO - codeparrot_training - Step 7350: {'lr': 0.00048482974392092827, 'samples': 3763712, 'steps': 7350, 'loss/train': 1.4325661659240723} 02/24/2022 11:05:41 - INFO - codeparrot_training - Step 7351: {'lr': 0.0004848241303575185, 'samples': 3764224, 'steps': 7351, 'loss/train': 2.5114574432373047} 02/24/2022 11:05:47 - INFO - codeparrot_training - Step 7352: {'lr': 0.0004848185157881968, 'samples': 3764736, 'steps': 7352, 'loss/train': 2.7234508991241455} 02/24/2022 11:05:50 - INFO - codeparrot_training - Step 7353: {'lr': 0.0004848129002129871, 'samples': 3765248, 'steps': 7353, 'loss/train': 1.158555030822754} 02/24/2022 11:05:56 - INFO - codeparrot_training - Step 7354: {'lr': 0.0004848072836319134, 'samples': 3765760, 'steps': 7354, 'loss/train': 2.1142032146453857} 02/24/2022 11:05:59 - INFO - codeparrot_training - Step 7355: {'lr': 0.000484801666045, 'samples': 3766272, 'steps': 7355, 'loss/train': 1.7687695026397705} 02/24/2022 11:06:04 - INFO - codeparrot_training - Step 7356: {'lr': 0.0004847960474522707, 'samples': 3766784, 'steps': 7356, 'loss/train': 3.242324113845825} 02/24/2022 11:06:08 - INFO - codeparrot_training - Step 7357: {'lr': 0.00048479042785374974, 'samples': 3767296, 'steps': 7357, 'loss/train': 1.59627366065979} 02/24/2022 11:06:13 - INFO - codeparrot_training - Step 7358: {'lr': 0.0004847848072494611, 'samples': 3767808, 'steps': 7358, 'loss/train': 2.9463582038879395} 02/24/2022 11:06:17 - INFO - codeparrot_training - Step 7359: {'lr': 0.0004847791856394289, 'samples': 3768320, 'steps': 7359, 'loss/train': 1.2039793729782104} 02/24/2022 11:06:23 - INFO - codeparrot_training - Step 7360: {'lr': 0.00048477356302367724, 'samples': 3768832, 'steps': 7360, 'loss/train': 2.4084179401397705} 02/24/2022 11:06:26 - INFO - codeparrot_training - Step 7361: {'lr': 0.00048476793940223026, 'samples': 3769344, 'steps': 7361, 'loss/train': 2.2719168663024902} 02/24/2022 11:06:32 - INFO - codeparrot_training - Step 7362: {'lr': 0.0004847623147751119, 'samples': 3769856, 'steps': 7362, 'loss/train': 2.434413433074951} 02/24/2022 11:06:35 - INFO - codeparrot_training - Step 7363: {'lr': 0.00048475668914234636, 'samples': 3770368, 'steps': 7363, 'loss/train': 2.3244428634643555} 02/24/2022 11:06:42 - INFO - codeparrot_training - Step 7364: {'lr': 0.0004847510625039577, 'samples': 3770880, 'steps': 7364, 'loss/train': 1.4526770114898682} 02/24/2022 11:06:46 - INFO - codeparrot_training - Step 7365: {'lr': 0.00048474543485997005, 'samples': 3771392, 'steps': 7365, 'loss/train': 0.6961601376533508} 02/24/2022 11:06:51 - INFO - codeparrot_training - Step 7366: {'lr': 0.00048473980621040744, 'samples': 3771904, 'steps': 7366, 'loss/train': 3.000654697418213} 02/24/2022 11:06:55 - INFO - codeparrot_training - Step 7367: {'lr': 0.00048473417655529405, 'samples': 3772416, 'steps': 7367, 'loss/train': 1.0492222309112549} 02/24/2022 11:07:00 - INFO - codeparrot_training - Step 7368: {'lr': 0.000484728545894654, 'samples': 3772928, 'steps': 7368, 'loss/train': 2.8229119777679443} 02/24/2022 11:07:04 - INFO - codeparrot_training - Step 7369: {'lr': 0.00048472291422851135, 'samples': 3773440, 'steps': 7369, 'loss/train': 2.4911160469055176} 02/24/2022 11:07:09 - INFO - codeparrot_training - Step 7370: {'lr': 0.00048471728155689034, 'samples': 3773952, 'steps': 7370, 'loss/train': 2.662477970123291} 02/24/2022 11:07:13 - INFO - codeparrot_training - Step 7371: {'lr': 0.000484711647879815, 'samples': 3774464, 'steps': 7371, 'loss/train': 2.269019842147827} 02/24/2022 11:07:18 - INFO - codeparrot_training - Step 7372: {'lr': 0.00048470601319730946, 'samples': 3774976, 'steps': 7372, 'loss/train': 2.9981753826141357} 02/24/2022 11:07:22 - INFO - codeparrot_training - Step 7373: {'lr': 0.00048470037750939795, 'samples': 3775488, 'steps': 7373, 'loss/train': 1.798475742340088} 02/24/2022 11:07:29 - INFO - codeparrot_training - Step 7374: {'lr': 0.0004846947408161045, 'samples': 3776000, 'steps': 7374, 'loss/train': 2.5420162677764893} 02/24/2022 11:07:32 - INFO - codeparrot_training - Step 7375: {'lr': 0.0004846891031174533, 'samples': 3776512, 'steps': 7375, 'loss/train': 1.373158574104309} 02/24/2022 11:07:38 - INFO - codeparrot_training - Step 7376: {'lr': 0.00048468346441346853, 'samples': 3777024, 'steps': 7376, 'loss/train': 2.904468059539795} 02/24/2022 11:07:41 - INFO - codeparrot_training - Step 7377: {'lr': 0.00048467782470417434, 'samples': 3777536, 'steps': 7377, 'loss/train': 1.616076111793518} 02/24/2022 11:07:47 - INFO - codeparrot_training - Step 7378: {'lr': 0.0004846721839895948, 'samples': 3778048, 'steps': 7378, 'loss/train': 2.5133163928985596} 02/24/2022 11:07:50 - INFO - codeparrot_training - Step 7379: {'lr': 0.00048466654226975414, 'samples': 3778560, 'steps': 7379, 'loss/train': 1.7621402740478516} 02/24/2022 11:07:56 - INFO - codeparrot_training - Step 7380: {'lr': 0.00048466089954467663, 'samples': 3779072, 'steps': 7380, 'loss/train': 2.6987743377685547} 02/24/2022 11:07:59 - INFO - codeparrot_training - Step 7381: {'lr': 0.0004846552558143863, 'samples': 3779584, 'steps': 7381, 'loss/train': 2.5966460704803467} 02/24/2022 11:08:05 - INFO - codeparrot_training - Step 7382: {'lr': 0.00048464961107890734, 'samples': 3780096, 'steps': 7382, 'loss/train': 2.0526061058044434} 02/24/2022 11:08:08 - INFO - codeparrot_training - Step 7383: {'lr': 0.00048464396533826396, 'samples': 3780608, 'steps': 7383, 'loss/train': 2.589766263961792} 02/24/2022 11:08:16 - INFO - codeparrot_training - Step 7384: {'lr': 0.0004846383185924803, 'samples': 3781120, 'steps': 7384, 'loss/train': 2.858475685119629} 02/24/2022 11:08:19 - INFO - codeparrot_training - Step 7385: {'lr': 0.0004846326708415806, 'samples': 3781632, 'steps': 7385, 'loss/train': 0.6818851232528687} 02/24/2022 11:08:25 - INFO - codeparrot_training - Step 7386: {'lr': 0.00048462702208558906, 'samples': 3782144, 'steps': 7386, 'loss/train': 2.251932144165039} 02/24/2022 11:08:28 - INFO - codeparrot_training - Step 7387: {'lr': 0.0004846213723245299, 'samples': 3782656, 'steps': 7387, 'loss/train': 3.389828681945801} 02/24/2022 11:08:34 - INFO - codeparrot_training - Step 7388: {'lr': 0.00048461572155842725, 'samples': 3783168, 'steps': 7388, 'loss/train': 3.451742649078369} 02/24/2022 11:08:37 - INFO - codeparrot_training - Step 7389: {'lr': 0.0004846100697873054, 'samples': 3783680, 'steps': 7389, 'loss/train': 2.0010650157928467} 02/24/2022 11:08:43 - INFO - codeparrot_training - Step 7390: {'lr': 0.0004846044170111884, 'samples': 3784192, 'steps': 7390, 'loss/train': 2.340823173522949} 02/24/2022 11:08:46 - INFO - codeparrot_training - Step 7391: {'lr': 0.00048459876323010063, 'samples': 3784704, 'steps': 7391, 'loss/train': 3.3236348628997803} 02/24/2022 11:08:52 - INFO - codeparrot_training - Step 7392: {'lr': 0.00048459310844406624, 'samples': 3785216, 'steps': 7392, 'loss/train': 1.8938822746276855} 02/24/2022 11:08:55 - INFO - codeparrot_training - Step 7393: {'lr': 0.0004845874526531095, 'samples': 3785728, 'steps': 7393, 'loss/train': 2.624589204788208} 02/24/2022 11:09:03 - INFO - codeparrot_training - Step 7394: {'lr': 0.0004845817958572546, 'samples': 3786240, 'steps': 7394, 'loss/train': 1.487290620803833} 02/24/2022 11:09:06 - INFO - codeparrot_training - Step 7395: {'lr': 0.0004845761380565257, 'samples': 3786752, 'steps': 7395, 'loss/train': 2.2486672401428223} 02/24/2022 11:09:12 - INFO - codeparrot_training - Step 7396: {'lr': 0.0004845704792509472, 'samples': 3787264, 'steps': 7396, 'loss/train': 1.6457959413528442} 02/24/2022 11:09:15 - INFO - codeparrot_training - Step 7397: {'lr': 0.0004845648194405432, 'samples': 3787776, 'steps': 7397, 'loss/train': 2.4682440757751465} 02/24/2022 11:09:21 - INFO - codeparrot_training - Step 7398: {'lr': 0.00048455915862533804, 'samples': 3788288, 'steps': 7398, 'loss/train': 3.274388551712036} 02/24/2022 11:09:24 - INFO - codeparrot_training - Step 7399: {'lr': 0.0004845534968053559, 'samples': 3788800, 'steps': 7399, 'loss/train': 2.3076016902923584} 02/24/2022 11:09:30 - INFO - codeparrot_training - Step 7400: {'lr': 0.0004845478339806211, 'samples': 3789312, 'steps': 7400, 'loss/train': 1.9895148277282715} 02/24/2022 11:09:33 - INFO - codeparrot_training - Step 7401: {'lr': 0.0004845421701511578, 'samples': 3789824, 'steps': 7401, 'loss/train': 3.1556906700134277} 02/24/2022 11:09:39 - INFO - codeparrot_training - Step 7402: {'lr': 0.0004845365053169903, 'samples': 3790336, 'steps': 7402, 'loss/train': 1.161543846130371} 02/24/2022 11:09:42 - INFO - codeparrot_training - Step 7403: {'lr': 0.0004845308394781429, 'samples': 3790848, 'steps': 7403, 'loss/train': 1.674891471862793} 02/24/2022 11:09:48 - INFO - codeparrot_training - Step 7404: {'lr': 0.0004845251726346399, 'samples': 3791360, 'steps': 7404, 'loss/train': 1.5679104328155518} 02/24/2022 11:09:51 - INFO - codeparrot_training - Step 7405: {'lr': 0.0004845195047865055, 'samples': 3791872, 'steps': 7405, 'loss/train': 3.752072811126709} 02/24/2022 11:09:57 - INFO - codeparrot_training - Step 7406: {'lr': 0.00048451383593376394, 'samples': 3792384, 'steps': 7406, 'loss/train': 2.251842498779297} 02/24/2022 11:10:00 - INFO - codeparrot_training - Step 7407: {'lr': 0.0004845081660764397, 'samples': 3792896, 'steps': 7407, 'loss/train': 2.050541639328003} 02/24/2022 11:10:06 - INFO - codeparrot_training - Step 7408: {'lr': 0.0004845024952145569, 'samples': 3793408, 'steps': 7408, 'loss/train': 3.354140520095825} 02/24/2022 11:10:09 - INFO - codeparrot_training - Step 7409: {'lr': 0.00048449682334813983, 'samples': 3793920, 'steps': 7409, 'loss/train': 2.8606581687927246} 02/24/2022 11:10:17 - INFO - codeparrot_training - Step 7410: {'lr': 0.00048449115047721286, 'samples': 3794432, 'steps': 7410, 'loss/train': 1.5060209035873413} 02/24/2022 11:10:21 - INFO - codeparrot_training - Step 7411: {'lr': 0.00048448547660180034, 'samples': 3794944, 'steps': 7411, 'loss/train': 2.3049936294555664} 02/24/2022 11:10:26 - INFO - codeparrot_training - Step 7412: {'lr': 0.0004844798017219264, 'samples': 3795456, 'steps': 7412, 'loss/train': 2.2139499187469482} 02/24/2022 11:10:30 - INFO - codeparrot_training - Step 7413: {'lr': 0.00048447412583761543, 'samples': 3795968, 'steps': 7413, 'loss/train': 2.282728910446167} 02/24/2022 11:10:35 - INFO - codeparrot_training - Step 7414: {'lr': 0.00048446844894889173, 'samples': 3796480, 'steps': 7414, 'loss/train': 3.0450706481933594} 02/24/2022 11:10:39 - INFO - codeparrot_training - Step 7415: {'lr': 0.00048446277105577973, 'samples': 3796992, 'steps': 7415, 'loss/train': 0.5314150452613831} 02/24/2022 11:10:44 - INFO - codeparrot_training - Step 7416: {'lr': 0.0004844570921583037, 'samples': 3797504, 'steps': 7416, 'loss/train': 4.11193323135376} 02/24/2022 11:10:48 - INFO - codeparrot_training - Step 7417: {'lr': 0.00048445141225648785, 'samples': 3798016, 'steps': 7417, 'loss/train': 1.738006591796875} 02/24/2022 11:10:53 - INFO - codeparrot_training - Step 7418: {'lr': 0.00048444573135035665, 'samples': 3798528, 'steps': 7418, 'loss/train': 2.838282346725464} 02/24/2022 11:10:57 - INFO - codeparrot_training - Step 7419: {'lr': 0.00048444004943993434, 'samples': 3799040, 'steps': 7419, 'loss/train': 2.3490233421325684} 02/24/2022 11:11:05 - INFO - codeparrot_training - Step 7420: {'lr': 0.0004844343665252453, 'samples': 3799552, 'steps': 7420, 'loss/train': 2.526664972305298} 02/24/2022 11:11:08 - INFO - codeparrot_training - Step 7421: {'lr': 0.0004844286826063139, 'samples': 3800064, 'steps': 7421, 'loss/train': 2.28163480758667} 02/24/2022 11:11:12 - INFO - codeparrot_training - Step 7422: {'lr': 0.0004844229976831645, 'samples': 3800576, 'steps': 7422, 'loss/train': 2.114687919616699} 02/24/2022 11:11:17 - INFO - codeparrot_training - Step 7423: {'lr': 0.00048441731175582136, 'samples': 3801088, 'steps': 7423, 'loss/train': 2.468702554702759} 02/24/2022 11:11:21 - INFO - codeparrot_training - Step 7424: {'lr': 0.0004844116248243089, 'samples': 3801600, 'steps': 7424, 'loss/train': 2.777743101119995} 02/24/2022 11:11:26 - INFO - codeparrot_training - Step 7425: {'lr': 0.00048440593688865155, 'samples': 3802112, 'steps': 7425, 'loss/train': 2.7870445251464844} 02/24/2022 11:11:30 - INFO - codeparrot_training - Step 7426: {'lr': 0.0004844002479488735, 'samples': 3802624, 'steps': 7426, 'loss/train': 1.713482141494751} 02/24/2022 11:11:35 - INFO - codeparrot_training - Step 7427: {'lr': 0.0004843945580049992, 'samples': 3803136, 'steps': 7427, 'loss/train': 0.9529287219047546} 02/24/2022 11:11:39 - INFO - codeparrot_training - Step 7428: {'lr': 0.0004843888670570531, 'samples': 3803648, 'steps': 7428, 'loss/train': 0.5740416049957275} 02/24/2022 11:11:44 - INFO - codeparrot_training - Step 7429: {'lr': 0.00048438317510505954, 'samples': 3804160, 'steps': 7429, 'loss/train': 1.3869069814682007} 02/24/2022 11:11:48 - INFO - codeparrot_training - Step 7430: {'lr': 0.0004843774821490429, 'samples': 3804672, 'steps': 7430, 'loss/train': 2.052302837371826} 02/24/2022 11:11:55 - INFO - codeparrot_training - Step 7431: {'lr': 0.0004843717881890275, 'samples': 3805184, 'steps': 7431, 'loss/train': 3.2143967151641846} 02/24/2022 11:11:59 - INFO - codeparrot_training - Step 7432: {'lr': 0.0004843660932250378, 'samples': 3805696, 'steps': 7432, 'loss/train': 2.6610870361328125} 02/24/2022 11:12:04 - INFO - codeparrot_training - Step 7433: {'lr': 0.0004843603972570981, 'samples': 3806208, 'steps': 7433, 'loss/train': 0.7759596109390259} 02/24/2022 11:12:08 - INFO - codeparrot_training - Step 7434: {'lr': 0.00048435470028523295, 'samples': 3806720, 'steps': 7434, 'loss/train': 2.1584222316741943} 02/24/2022 11:12:13 - INFO - codeparrot_training - Step 7435: {'lr': 0.00048434900230946666, 'samples': 3807232, 'steps': 7435, 'loss/train': 2.5876851081848145} 02/24/2022 11:12:16 - INFO - codeparrot_training - Step 7436: {'lr': 0.0004843433033298237, 'samples': 3807744, 'steps': 7436, 'loss/train': 1.3958079814910889} 02/24/2022 11:12:22 - INFO - codeparrot_training - Step 7437: {'lr': 0.00048433760334632835, 'samples': 3808256, 'steps': 7437, 'loss/train': 2.6977572441101074} 02/24/2022 11:12:25 - INFO - codeparrot_training - Step 7438: {'lr': 0.0004843319023590052, 'samples': 3808768, 'steps': 7438, 'loss/train': 1.7582217454910278} 02/24/2022 11:12:31 - INFO - codeparrot_training - Step 7439: {'lr': 0.0004843262003678786, 'samples': 3809280, 'steps': 7439, 'loss/train': 1.7335035800933838} 02/24/2022 11:12:34 - INFO - codeparrot_training - Step 7440: {'lr': 0.0004843204973729729, 'samples': 3809792, 'steps': 7440, 'loss/train': 2.367358922958374} 02/24/2022 11:12:42 - INFO - codeparrot_training - Step 7441: {'lr': 0.0004843147933743126, 'samples': 3810304, 'steps': 7441, 'loss/train': 3.0420961380004883} 02/24/2022 11:12:45 - INFO - codeparrot_training - Step 7442: {'lr': 0.0004843090883719222, 'samples': 3810816, 'steps': 7442, 'loss/train': 1.9123307466506958} 02/24/2022 11:12:51 - INFO - codeparrot_training - Step 7443: {'lr': 0.00048430338236582596, 'samples': 3811328, 'steps': 7443, 'loss/train': 1.2915576696395874} 02/24/2022 11:12:57 - INFO - codeparrot_training - Step 7444: {'lr': 0.0004842976753560485, 'samples': 3811840, 'steps': 7444, 'loss/train': 2.6454339027404785} 02/24/2022 11:13:00 - INFO - codeparrot_training - Step 7445: {'lr': 0.00048429196734261413, 'samples': 3812352, 'steps': 7445, 'loss/train': 0.4806378483772278} 02/24/2022 11:13:03 - INFO - codeparrot_training - Step 7446: {'lr': 0.00048428625832554754, 'samples': 3812864, 'steps': 7446, 'loss/train': 4.000292778015137} 02/24/2022 11:13:09 - INFO - codeparrot_training - Step 7447: {'lr': 0.0004842805483048728, 'samples': 3813376, 'steps': 7447, 'loss/train': 2.2990851402282715} 02/24/2022 11:13:13 - INFO - codeparrot_training - Step 7448: {'lr': 0.0004842748372806147, 'samples': 3813888, 'steps': 7448, 'loss/train': 2.6723527908325195} 02/24/2022 11:13:18 - INFO - codeparrot_training - Step 7449: {'lr': 0.0004842691252527976, 'samples': 3814400, 'steps': 7449, 'loss/train': 2.4262754917144775} 02/24/2022 11:13:21 - INFO - codeparrot_training - Step 7450: {'lr': 0.00048426341222144586, 'samples': 3814912, 'steps': 7450, 'loss/train': 1.598927617073059} 02/24/2022 11:13:27 - INFO - codeparrot_training - Step 7451: {'lr': 0.00048425769818658416, 'samples': 3815424, 'steps': 7451, 'loss/train': 1.358352541923523} 02/24/2022 11:13:31 - INFO - codeparrot_training - Step 7452: {'lr': 0.0004842519831482368, 'samples': 3815936, 'steps': 7452, 'loss/train': 2.3954668045043945} 02/24/2022 11:13:36 - INFO - codeparrot_training - Step 7453: {'lr': 0.00048424626710642836, 'samples': 3816448, 'steps': 7453, 'loss/train': 1.1535263061523438} 02/24/2022 11:13:40 - INFO - codeparrot_training - Step 7454: {'lr': 0.0004842405500611833, 'samples': 3816960, 'steps': 7454, 'loss/train': 3.4303042888641357} 02/24/2022 11:13:45 - INFO - codeparrot_training - Step 7455: {'lr': 0.00048423483201252604, 'samples': 3817472, 'steps': 7455, 'loss/train': 2.8033041954040527} 02/24/2022 11:13:49 - INFO - codeparrot_training - Step 7456: {'lr': 0.0004842291129604812, 'samples': 3817984, 'steps': 7456, 'loss/train': 2.044224262237549} 02/24/2022 11:13:56 - INFO - codeparrot_training - Step 7457: {'lr': 0.0004842233929050732, 'samples': 3818496, 'steps': 7457, 'loss/train': 2.26042103767395} 02/24/2022 11:14:00 - INFO - codeparrot_training - Step 7458: {'lr': 0.00048421767184632657, 'samples': 3819008, 'steps': 7458, 'loss/train': 1.8025332689285278} 02/24/2022 11:14:05 - INFO - codeparrot_training - Step 7459: {'lr': 0.00048421194978426574, 'samples': 3819520, 'steps': 7459, 'loss/train': 2.354947328567505} 02/24/2022 11:14:09 - INFO - codeparrot_training - Step 7460: {'lr': 0.00048420622671891533, 'samples': 3820032, 'steps': 7460, 'loss/train': 1.2694612741470337} 02/24/2022 11:14:14 - INFO - codeparrot_training - Step 7461: {'lr': 0.0004842005026502999, 'samples': 3820544, 'steps': 7461, 'loss/train': 3.551039457321167} 02/24/2022 11:14:18 - INFO - codeparrot_training - Step 7462: {'lr': 0.00048419477757844376, 'samples': 3821056, 'steps': 7462, 'loss/train': 2.7074220180511475} 02/24/2022 11:14:23 - INFO - codeparrot_training - Step 7463: {'lr': 0.00048418905150337166, 'samples': 3821568, 'steps': 7463, 'loss/train': 2.501370906829834} 02/24/2022 11:14:27 - INFO - codeparrot_training - Step 7464: {'lr': 0.00048418332442510794, 'samples': 3822080, 'steps': 7464, 'loss/train': 1.9361931085586548} 02/24/2022 11:14:32 - INFO - codeparrot_training - Step 7465: {'lr': 0.00048417759634367726, 'samples': 3822592, 'steps': 7465, 'loss/train': 2.5320217609405518} 02/24/2022 11:14:36 - INFO - codeparrot_training - Step 7466: {'lr': 0.00048417186725910414, 'samples': 3823104, 'steps': 7466, 'loss/train': 2.402559280395508} 02/24/2022 11:14:43 - INFO - codeparrot_training - Step 7467: {'lr': 0.000484166137171413, 'samples': 3823616, 'steps': 7467, 'loss/train': 2.707710027694702} 02/24/2022 11:14:46 - INFO - codeparrot_training - Step 7468: {'lr': 0.0004841604060806286, 'samples': 3824128, 'steps': 7468, 'loss/train': 3.385481595993042} 02/24/2022 11:14:52 - INFO - codeparrot_training - Step 7469: {'lr': 0.00048415467398677534, 'samples': 3824640, 'steps': 7469, 'loss/train': 1.0140942335128784} 02/24/2022 11:14:55 - INFO - codeparrot_training - Step 7470: {'lr': 0.0004841489408898778, 'samples': 3825152, 'steps': 7470, 'loss/train': 2.7896058559417725} 02/24/2022 11:15:01 - INFO - codeparrot_training - Step 7471: {'lr': 0.0004841432067899605, 'samples': 3825664, 'steps': 7471, 'loss/train': 2.2625505924224854} 02/24/2022 11:15:04 - INFO - codeparrot_training - Step 7472: {'lr': 0.0004841374716870481, 'samples': 3826176, 'steps': 7472, 'loss/train': 2.1733639240264893} 02/24/2022 11:15:10 - INFO - codeparrot_training - Step 7473: {'lr': 0.0004841317355811651, 'samples': 3826688, 'steps': 7473, 'loss/train': 2.0763137340545654} 02/24/2022 11:15:15 - INFO - codeparrot_training - Step 7474: {'lr': 0.00048412599847233613, 'samples': 3827200, 'steps': 7474, 'loss/train': 2.003067970275879} 02/24/2022 11:15:19 - INFO - codeparrot_training - Step 7475: {'lr': 0.0004841202603605857, 'samples': 3827712, 'steps': 7475, 'loss/train': 1.6955280303955078} 02/24/2022 11:15:26 - INFO - codeparrot_training - Step 7476: {'lr': 0.0004841145212459384, 'samples': 3828224, 'steps': 7476, 'loss/train': 2.7507684230804443} 02/24/2022 11:15:30 - INFO - codeparrot_training - Step 7477: {'lr': 0.0004841087811284188, 'samples': 3828736, 'steps': 7477, 'loss/train': 2.3904590606689453} 02/24/2022 11:15:35 - INFO - codeparrot_training - Step 7478: {'lr': 0.0004841030400080516, 'samples': 3829248, 'steps': 7478, 'loss/train': 2.1411893367767334} 02/24/2022 11:15:39 - INFO - codeparrot_training - Step 7479: {'lr': 0.00048409729788486127, 'samples': 3829760, 'steps': 7479, 'loss/train': 2.5110843181610107} 02/24/2022 11:15:42 - INFO - codeparrot_training - Step 7480: {'lr': 0.00048409155475887244, 'samples': 3830272, 'steps': 7480, 'loss/train': 0.2696789503097534} 02/24/2022 11:15:48 - INFO - codeparrot_training - Step 7481: {'lr': 0.00048408581063010973, 'samples': 3830784, 'steps': 7481, 'loss/train': 1.666459083557129} 02/24/2022 11:15:51 - INFO - codeparrot_training - Step 7482: {'lr': 0.00048408006549859777, 'samples': 3831296, 'steps': 7482, 'loss/train': 2.4707143306732178} 02/24/2022 11:15:57 - INFO - codeparrot_training - Step 7483: {'lr': 0.00048407431936436116, 'samples': 3831808, 'steps': 7483, 'loss/train': 1.0383611917495728} 02/24/2022 11:16:00 - INFO - codeparrot_training - Step 7484: {'lr': 0.0004840685722274244, 'samples': 3832320, 'steps': 7484, 'loss/train': 0.20521792769432068} 02/24/2022 11:16:06 - INFO - codeparrot_training - Step 7485: {'lr': 0.00048406282408781226, 'samples': 3832832, 'steps': 7485, 'loss/train': 2.5948879718780518} 02/24/2022 11:16:09 - INFO - codeparrot_training - Step 7486: {'lr': 0.0004840570749455493, 'samples': 3833344, 'steps': 7486, 'loss/train': 1.5987448692321777} 02/24/2022 11:16:16 - INFO - codeparrot_training - Step 7487: {'lr': 0.00048405132480066015, 'samples': 3833856, 'steps': 7487, 'loss/train': 2.2542054653167725} 02/24/2022 11:16:20 - INFO - codeparrot_training - Step 7488: {'lr': 0.00048404557365316946, 'samples': 3834368, 'steps': 7488, 'loss/train': 2.9628989696502686} 02/24/2022 11:16:25 - INFO - codeparrot_training - Step 7489: {'lr': 0.00048403982150310184, 'samples': 3834880, 'steps': 7489, 'loss/train': 0.8160163760185242} 02/24/2022 11:16:29 - INFO - codeparrot_training - Step 7490: {'lr': 0.0004840340683504819, 'samples': 3835392, 'steps': 7490, 'loss/train': 0.9846836924552917} 02/24/2022 11:16:35 - INFO - codeparrot_training - Step 7491: {'lr': 0.0004840283141953343, 'samples': 3835904, 'steps': 7491, 'loss/train': 3.9652154445648193} 02/24/2022 11:16:38 - INFO - codeparrot_training - Step 7492: {'lr': 0.0004840225590376839, 'samples': 3836416, 'steps': 7492, 'loss/train': 2.300816297531128} 02/24/2022 11:16:44 - INFO - codeparrot_training - Step 7493: {'lr': 0.000484016802877555, 'samples': 3836928, 'steps': 7493, 'loss/train': 1.4028760194778442} 02/24/2022 11:16:47 - INFO - codeparrot_training - Step 7494: {'lr': 0.00048401104571497245, 'samples': 3837440, 'steps': 7494, 'loss/train': 2.2573134899139404} 02/24/2022 11:16:53 - INFO - codeparrot_training - Step 7495: {'lr': 0.00048400528754996086, 'samples': 3837952, 'steps': 7495, 'loss/train': 2.2583260536193848} 02/24/2022 11:16:56 - INFO - codeparrot_training - Step 7496: {'lr': 0.000483999528382545, 'samples': 3838464, 'steps': 7496, 'loss/train': 4.183949947357178} 02/24/2022 11:17:03 - INFO - codeparrot_training - Step 7497: {'lr': 0.00048399376821274943, 'samples': 3838976, 'steps': 7497, 'loss/train': 1.9686617851257324} 02/24/2022 11:17:07 - INFO - codeparrot_training - Step 7498: {'lr': 0.00048398800704059887, 'samples': 3839488, 'steps': 7498, 'loss/train': 2.413862466812134} 02/24/2022 11:17:13 - INFO - codeparrot_training - Step 7499: {'lr': 0.000483982244866118, 'samples': 3840000, 'steps': 7499, 'loss/train': 1.8729325532913208} 02/24/2022 11:17:16 - INFO - codeparrot_training - Step 7500: {'lr': 0.00048397648168933144, 'samples': 3840512, 'steps': 7500, 'loss/train': 1.2530790567398071} 02/24/2022 11:17:22 - INFO - codeparrot_training - Step 7501: {'lr': 0.00048397071751026395, 'samples': 3841024, 'steps': 7501, 'loss/train': 1.9784280061721802} 02/24/2022 11:17:25 - INFO - codeparrot_training - Step 7502: {'lr': 0.00048396495232894024, 'samples': 3841536, 'steps': 7502, 'loss/train': 2.662201166152954} 02/24/2022 11:17:29 - INFO - codeparrot_training - Step 7503: {'lr': 0.0004839591861453849, 'samples': 3842048, 'steps': 7503, 'loss/train': 8.562796592712402} 02/24/2022 11:17:34 - INFO - codeparrot_training - Step 7504: {'lr': 0.00048395341895962277, 'samples': 3842560, 'steps': 7504, 'loss/train': 2.5060017108917236} 02/24/2022 11:17:38 - INFO - codeparrot_training - Step 7505: {'lr': 0.0004839476507716784, 'samples': 3843072, 'steps': 7505, 'loss/train': 2.0978822708129883} 02/24/2022 11:17:43 - INFO - codeparrot_training - Step 7506: {'lr': 0.0004839418815815766, 'samples': 3843584, 'steps': 7506, 'loss/train': 3.61733078956604} 02/24/2022 11:17:47 - INFO - codeparrot_training - Step 7507: {'lr': 0.0004839361113893421, 'samples': 3844096, 'steps': 7507, 'loss/train': 2.0031683444976807} 02/24/2022 11:17:52 - INFO - codeparrot_training - Step 7508: {'lr': 0.0004839303401949996, 'samples': 3844608, 'steps': 7508, 'loss/train': 1.562530517578125} 02/24/2022 11:17:56 - INFO - codeparrot_training - Step 7509: {'lr': 0.00048392456799857374, 'samples': 3845120, 'steps': 7509, 'loss/train': 2.808486223220825} 02/24/2022 11:18:01 - INFO - codeparrot_training - Step 7510: {'lr': 0.0004839187948000893, 'samples': 3845632, 'steps': 7510, 'loss/train': 2.772757053375244} 02/24/2022 11:18:05 - INFO - codeparrot_training - Step 7511: {'lr': 0.0004839130205995711, 'samples': 3846144, 'steps': 7511, 'loss/train': 2.466223955154419} 02/24/2022 11:18:10 - INFO - codeparrot_training - Step 7512: {'lr': 0.0004839072453970438, 'samples': 3846656, 'steps': 7512, 'loss/train': 0.5400563478469849} 02/24/2022 11:18:14 - INFO - codeparrot_training - Step 7513: {'lr': 0.00048390146919253206, 'samples': 3847168, 'steps': 7513, 'loss/train': 2.3244948387145996} 02/24/2022 11:18:21 - INFO - codeparrot_training - Step 7514: {'lr': 0.0004838956919860607, 'samples': 3847680, 'steps': 7514, 'loss/train': 2.395681858062744} 02/24/2022 11:18:27 - INFO - codeparrot_training - Step 7515: {'lr': 0.0004838899137776545, 'samples': 3848192, 'steps': 7515, 'loss/train': 1.977725625038147} 02/24/2022 11:18:30 - INFO - codeparrot_training - Step 7516: {'lr': 0.00048388413456733814, 'samples': 3848704, 'steps': 7516, 'loss/train': 3.4100310802459717} 02/24/2022 11:18:34 - INFO - codeparrot_training - Step 7517: {'lr': 0.0004838783543551365, 'samples': 3849216, 'steps': 7517, 'loss/train': 2.434039354324341} 02/24/2022 11:18:39 - INFO - codeparrot_training - Step 7518: {'lr': 0.0004838725731410742, 'samples': 3849728, 'steps': 7518, 'loss/train': 0.9382692575454712} 02/24/2022 11:18:43 - INFO - codeparrot_training - Step 7519: {'lr': 0.00048386679092517605, 'samples': 3850240, 'steps': 7519, 'loss/train': 1.0077052116394043} 02/24/2022 11:18:48 - INFO - codeparrot_training - Step 7520: {'lr': 0.00048386100770746686, 'samples': 3850752, 'steps': 7520, 'loss/train': 1.7812910079956055} 02/24/2022 11:18:52 - INFO - codeparrot_training - Step 7521: {'lr': 0.00048385522348797134, 'samples': 3851264, 'steps': 7521, 'loss/train': 2.62276291847229} 02/24/2022 11:18:58 - INFO - codeparrot_training - Step 7522: {'lr': 0.0004838494382667143, 'samples': 3851776, 'steps': 7522, 'loss/train': 1.857709288597107} 02/24/2022 11:19:01 - INFO - codeparrot_training - Step 7523: {'lr': 0.0004838436520437205, 'samples': 3852288, 'steps': 7523, 'loss/train': 3.001187562942505} 02/24/2022 11:19:08 - INFO - codeparrot_training - Step 7524: {'lr': 0.00048383786481901483, 'samples': 3852800, 'steps': 7524, 'loss/train': 2.8849003314971924} 02/24/2022 11:19:12 - INFO - codeparrot_training - Step 7525: {'lr': 0.00048383207659262196, 'samples': 3853312, 'steps': 7525, 'loss/train': 3.820636034011841} 02/24/2022 11:19:17 - INFO - codeparrot_training - Step 7526: {'lr': 0.0004838262873645667, 'samples': 3853824, 'steps': 7526, 'loss/train': 1.8931299448013306} 02/24/2022 11:19:21 - INFO - codeparrot_training - Step 7527: {'lr': 0.00048382049713487383, 'samples': 3854336, 'steps': 7527, 'loss/train': 9.131762504577637} 02/24/2022 11:19:26 - INFO - codeparrot_training - Step 7528: {'lr': 0.00048381470590356835, 'samples': 3854848, 'steps': 7528, 'loss/train': 3.472043752670288} 02/24/2022 11:19:30 - INFO - codeparrot_training - Step 7529: {'lr': 0.00048380891367067483, 'samples': 3855360, 'steps': 7529, 'loss/train': 1.9659978151321411} 02/24/2022 11:19:35 - INFO - codeparrot_training - Step 7530: {'lr': 0.0004838031204362181, 'samples': 3855872, 'steps': 7530, 'loss/train': 2.6228482723236084} 02/24/2022 11:19:39 - INFO - codeparrot_training - Step 7531: {'lr': 0.0004837973262002231, 'samples': 3856384, 'steps': 7531, 'loss/train': 3.1839160919189453} 02/24/2022 11:19:44 - INFO - codeparrot_training - Step 7532: {'lr': 0.0004837915309627146, 'samples': 3856896, 'steps': 7532, 'loss/train': 2.2843639850616455} 02/24/2022 11:19:48 - INFO - codeparrot_training - Step 7533: {'lr': 0.00048378573472371744, 'samples': 3857408, 'steps': 7533, 'loss/train': 2.2604622840881348} 02/24/2022 11:19:53 - INFO - codeparrot_training - Step 7534: {'lr': 0.0004837799374832564, 'samples': 3857920, 'steps': 7534, 'loss/train': 1.4056577682495117} 02/24/2022 11:19:57 - INFO - codeparrot_training - Step 7535: {'lr': 0.0004837741392413563, 'samples': 3858432, 'steps': 7535, 'loss/train': 2.7577364444732666} 02/24/2022 11:20:03 - INFO - codeparrot_training - Step 7536: {'lr': 0.000483768339998042, 'samples': 3858944, 'steps': 7536, 'loss/train': 1.8355239629745483} 02/24/2022 11:20:06 - INFO - codeparrot_training - Step 7537: {'lr': 0.0004837625397533385, 'samples': 3859456, 'steps': 7537, 'loss/train': 1.8684333562850952} 02/24/2022 11:20:12 - INFO - codeparrot_training - Step 7538: {'lr': 0.00048375673850727043, 'samples': 3859968, 'steps': 7538, 'loss/train': 2.1198551654815674} 02/24/2022 11:20:15 - INFO - codeparrot_training - Step 7539: {'lr': 0.00048375093625986274, 'samples': 3860480, 'steps': 7539, 'loss/train': 2.906947135925293} 02/24/2022 11:20:21 - INFO - codeparrot_training - Step 7540: {'lr': 0.0004837451330111402, 'samples': 3860992, 'steps': 7540, 'loss/train': 1.2826128005981445} 02/24/2022 11:20:25 - INFO - codeparrot_training - Step 7541: {'lr': 0.0004837393287611278, 'samples': 3861504, 'steps': 7541, 'loss/train': 2.5487778186798096} 02/24/2022 11:20:30 - INFO - codeparrot_training - Step 7542: {'lr': 0.0004837335235098503, 'samples': 3862016, 'steps': 7542, 'loss/train': 1.4793578386306763} 02/24/2022 11:20:34 - INFO - codeparrot_training - Step 7543: {'lr': 0.0004837277172573326, 'samples': 3862528, 'steps': 7543, 'loss/train': 2.404090166091919} 02/24/2022 11:20:39 - INFO - codeparrot_training - Step 7544: {'lr': 0.00048372191000359955, 'samples': 3863040, 'steps': 7544, 'loss/train': 2.2764785289764404} 02/24/2022 11:20:43 - INFO - codeparrot_training - Step 7545: {'lr': 0.00048371610174867614, 'samples': 3863552, 'steps': 7545, 'loss/train': 2.0950586795806885} 02/24/2022 11:20:48 - INFO - codeparrot_training - Step 7546: {'lr': 0.00048371029249258716, 'samples': 3864064, 'steps': 7546, 'loss/train': 2.016496181488037} 02/24/2022 11:20:52 - INFO - codeparrot_training - Step 7547: {'lr': 0.0004837044822353574, 'samples': 3864576, 'steps': 7547, 'loss/train': 3.632533550262451} 02/24/2022 11:20:57 - INFO - codeparrot_training - Step 7548: {'lr': 0.0004836986709770119, 'samples': 3865088, 'steps': 7548, 'loss/train': 1.8181231021881104} 02/24/2022 11:21:01 - INFO - codeparrot_training - Step 7549: {'lr': 0.00048369285871757554, 'samples': 3865600, 'steps': 7549, 'loss/train': 1.7309651374816895} 02/24/2022 11:21:07 - INFO - codeparrot_training - Step 7550: {'lr': 0.0004836870454570731, 'samples': 3866112, 'steps': 7550, 'loss/train': 2.7915351390838623} 02/24/2022 11:21:10 - INFO - codeparrot_training - Step 7551: {'lr': 0.00048368123119552965, 'samples': 3866624, 'steps': 7551, 'loss/train': 2.872307300567627} 02/24/2022 11:21:16 - INFO - codeparrot_training - Step 7552: {'lr': 0.00048367541593296996, 'samples': 3867136, 'steps': 7552, 'loss/train': 2.6407103538513184} 02/24/2022 11:21:19 - INFO - codeparrot_training - Step 7553: {'lr': 0.00048366959966941893, 'samples': 3867648, 'steps': 7553, 'loss/train': 2.058753728866577} 02/24/2022 11:21:25 - INFO - codeparrot_training - Step 7554: {'lr': 0.0004836637824049016, 'samples': 3868160, 'steps': 7554, 'loss/train': 1.3336979150772095} 02/24/2022 11:21:28 - INFO - codeparrot_training - Step 7555: {'lr': 0.00048365796413944284, 'samples': 3868672, 'steps': 7555, 'loss/train': 1.9933503866195679} 02/24/2022 11:21:34 - INFO - codeparrot_training - Step 7556: {'lr': 0.00048365214487306753, 'samples': 3869184, 'steps': 7556, 'loss/train': 1.4395159482955933} 02/24/2022 11:21:37 - INFO - codeparrot_training - Step 7557: {'lr': 0.0004836463246058006, 'samples': 3869696, 'steps': 7557, 'loss/train': 1.4961059093475342} 02/24/2022 11:21:43 - INFO - codeparrot_training - Step 7558: {'lr': 0.0004836405033376671, 'samples': 3870208, 'steps': 7558, 'loss/train': 2.4053282737731934} 02/24/2022 11:21:49 - INFO - codeparrot_training - Step 7559: {'lr': 0.00048363468106869177, 'samples': 3870720, 'steps': 7559, 'loss/train': 3.01051664352417} 02/24/2022 11:21:52 - INFO - codeparrot_training - Step 7560: {'lr': 0.00048362885779889967, 'samples': 3871232, 'steps': 7560, 'loss/train': 1.41871178150177} 02/24/2022 11:21:58 - INFO - codeparrot_training - Step 7561: {'lr': 0.0004836230335283158, 'samples': 3871744, 'steps': 7561, 'loss/train': 0.9912055134773254} 02/24/2022 11:22:01 - INFO - codeparrot_training - Step 7562: {'lr': 0.00048361720825696494, 'samples': 3872256, 'steps': 7562, 'loss/train': 2.072070837020874} 02/24/2022 11:22:07 - INFO - codeparrot_training - Step 7563: {'lr': 0.0004836113819848722, 'samples': 3872768, 'steps': 7563, 'loss/train': 2.1366806030273438} 02/24/2022 11:22:10 - INFO - codeparrot_training - Step 7564: {'lr': 0.0004836055547120625, 'samples': 3873280, 'steps': 7564, 'loss/train': 1.8939872980117798} 02/24/2022 11:22:16 - INFO - codeparrot_training - Step 7565: {'lr': 0.0004835997264385607, 'samples': 3873792, 'steps': 7565, 'loss/train': 1.8421449661254883} 02/24/2022 11:22:19 - INFO - codeparrot_training - Step 7566: {'lr': 0.0004835938971643919, 'samples': 3874304, 'steps': 7566, 'loss/train': 2.350156545639038} 02/24/2022 11:22:25 - INFO - codeparrot_training - Step 7567: {'lr': 0.000483588066889581, 'samples': 3874816, 'steps': 7567, 'loss/train': 3.722226858139038} 02/24/2022 11:22:28 - INFO - codeparrot_training - Step 7568: {'lr': 0.00048358223561415306, 'samples': 3875328, 'steps': 7568, 'loss/train': 2.0755059719085693} 02/24/2022 11:22:35 - INFO - codeparrot_training - Step 7569: {'lr': 0.0004835764033381329, 'samples': 3875840, 'steps': 7569, 'loss/train': 0.9213374257087708} 02/24/2022 11:22:38 - INFO - codeparrot_training - Step 7570: {'lr': 0.00048357057006154566, 'samples': 3876352, 'steps': 7570, 'loss/train': 0.18913401663303375} 02/24/2022 11:22:44 - INFO - codeparrot_training - Step 7571: {'lr': 0.0004835647357844162, 'samples': 3876864, 'steps': 7571, 'loss/train': 2.3233468532562256} 02/24/2022 11:22:47 - INFO - codeparrot_training - Step 7572: {'lr': 0.00048355890050676966, 'samples': 3877376, 'steps': 7572, 'loss/train': 1.3482396602630615} 02/24/2022 11:22:53 - INFO - codeparrot_training - Step 7573: {'lr': 0.0004835530642286309, 'samples': 3877888, 'steps': 7573, 'loss/train': 1.8638635873794556} 02/24/2022 11:22:56 - INFO - codeparrot_training - Step 7574: {'lr': 0.000483547226950025, 'samples': 3878400, 'steps': 7574, 'loss/train': 2.173956871032715} 02/24/2022 11:23:02 - INFO - codeparrot_training - Step 7575: {'lr': 0.00048354138867097695, 'samples': 3878912, 'steps': 7575, 'loss/train': 1.251124620437622} 02/24/2022 11:23:05 - INFO - codeparrot_training - Step 7576: {'lr': 0.00048353554939151167, 'samples': 3879424, 'steps': 7576, 'loss/train': 1.7908861637115479} 02/24/2022 11:23:11 - INFO - codeparrot_training - Step 7577: {'lr': 0.00048352970911165434, 'samples': 3879936, 'steps': 7577, 'loss/train': 1.157549500465393} 02/24/2022 11:23:14 - INFO - codeparrot_training - Step 7578: {'lr': 0.0004835238678314299, 'samples': 3880448, 'steps': 7578, 'loss/train': 2.919440269470215} 02/24/2022 11:23:19 - INFO - codeparrot_training - Step 7579: {'lr': 0.00048351802555086335, 'samples': 3880960, 'steps': 7579, 'loss/train': 1.6097110509872437} 02/24/2022 11:23:23 - INFO - codeparrot_training - Step 7580: {'lr': 0.0004835121822699796, 'samples': 3881472, 'steps': 7580, 'loss/train': 2.9920217990875244} 02/24/2022 11:23:29 - INFO - codeparrot_training - Step 7581: {'lr': 0.00048350633798880397, 'samples': 3881984, 'steps': 7581, 'loss/train': 1.9408460855484009} 02/24/2022 11:23:32 - INFO - codeparrot_training - Step 7582: {'lr': 0.0004835004927073613, 'samples': 3882496, 'steps': 7582, 'loss/train': 1.9490169286727905} 02/24/2022 11:23:38 - INFO - codeparrot_training - Step 7583: {'lr': 0.0004834946464256766, 'samples': 3883008, 'steps': 7583, 'loss/train': 0.8694325685501099} 02/24/2022 11:23:42 - INFO - codeparrot_training - Step 7584: {'lr': 0.00048348879914377504, 'samples': 3883520, 'steps': 7584, 'loss/train': 1.966606616973877} 02/24/2022 11:23:45 - INFO - codeparrot_training - Step 7585: {'lr': 0.0004834829508616816, 'samples': 3884032, 'steps': 7585, 'loss/train': 2.655318260192871} 02/24/2022 11:23:52 - INFO - codeparrot_training - Step 7586: {'lr': 0.00048347710157942126, 'samples': 3884544, 'steps': 7586, 'loss/train': 2.219027042388916} 02/24/2022 11:23:55 - INFO - codeparrot_training - Step 7587: {'lr': 0.00048347125129701924, 'samples': 3885056, 'steps': 7587, 'loss/train': 1.47925865650177} 02/24/2022 11:24:00 - INFO - codeparrot_training - Step 7588: {'lr': 0.00048346540001450045, 'samples': 3885568, 'steps': 7588, 'loss/train': 0.6150884628295898} 02/24/2022 11:24:04 - INFO - codeparrot_training - Step 7589: {'lr': 0.0004834595477318901, 'samples': 3886080, 'steps': 7589, 'loss/train': 1.8040697574615479} 02/24/2022 11:24:09 - INFO - codeparrot_training - Step 7590: {'lr': 0.00048345369444921315, 'samples': 3886592, 'steps': 7590, 'loss/train': 1.8759030103683472} 02/24/2022 11:24:13 - INFO - codeparrot_training - Step 7591: {'lr': 0.00048344784016649467, 'samples': 3887104, 'steps': 7591, 'loss/train': 2.656978130340576} 02/24/2022 11:24:18 - INFO - codeparrot_training - Step 7592: {'lr': 0.0004834419848837598, 'samples': 3887616, 'steps': 7592, 'loss/train': 1.5592306852340698} 02/24/2022 11:24:22 - INFO - codeparrot_training - Step 7593: {'lr': 0.0004834361286010336, 'samples': 3888128, 'steps': 7593, 'loss/train': 2.9403769969940186} 02/24/2022 11:24:27 - INFO - codeparrot_training - Step 7594: {'lr': 0.0004834302713183411, 'samples': 3888640, 'steps': 7594, 'loss/train': 1.8358407020568848} 02/24/2022 11:24:31 - INFO - codeparrot_training - Step 7595: {'lr': 0.0004834244130357075, 'samples': 3889152, 'steps': 7595, 'loss/train': 2.9446747303009033} 02/24/2022 11:24:37 - INFO - codeparrot_training - Step 7596: {'lr': 0.0004834185537531578, 'samples': 3889664, 'steps': 7596, 'loss/train': 2.3520755767822266} 02/24/2022 11:24:40 - INFO - codeparrot_training - Step 7597: {'lr': 0.00048341269347071717, 'samples': 3890176, 'steps': 7597, 'loss/train': 2.0623204708099365} 02/24/2022 11:24:46 - INFO - codeparrot_training - Step 7598: {'lr': 0.00048340683218841066, 'samples': 3890688, 'steps': 7598, 'loss/train': 1.848036289215088} 02/24/2022 11:24:49 - INFO - codeparrot_training - Step 7599: {'lr': 0.00048340096990626336, 'samples': 3891200, 'steps': 7599, 'loss/train': 1.7646814584732056} 02/24/2022 11:24:55 - INFO - codeparrot_training - Step 7600: {'lr': 0.00048339510662430044, 'samples': 3891712, 'steps': 7600, 'loss/train': 2.6719794273376465} 02/24/2022 11:24:59 - INFO - codeparrot_training - Step 7601: {'lr': 0.000483389242342547, 'samples': 3892224, 'steps': 7601, 'loss/train': 2.8692915439605713} 02/24/2022 11:25:04 - INFO - codeparrot_training - Step 7602: {'lr': 0.00048338337706102817, 'samples': 3892736, 'steps': 7602, 'loss/train': 1.6060826778411865} 02/24/2022 11:25:08 - INFO - codeparrot_training - Step 7603: {'lr': 0.00048337751077976907, 'samples': 3893248, 'steps': 7603, 'loss/train': 3.6554031372070312} 02/24/2022 11:25:13 - INFO - codeparrot_training - Step 7604: {'lr': 0.0004833716434987948, 'samples': 3893760, 'steps': 7604, 'loss/train': 1.902369499206543} 02/24/2022 11:25:16 - INFO - codeparrot_training - Step 7605: {'lr': 0.0004833657752181305, 'samples': 3894272, 'steps': 7605, 'loss/train': 2.3780405521392822} 02/24/2022 11:25:23 - INFO - codeparrot_training - Step 7606: {'lr': 0.00048335990593780133, 'samples': 3894784, 'steps': 7606, 'loss/train': 2.47692608833313} 02/24/2022 11:25:27 - INFO - codeparrot_training - Step 7607: {'lr': 0.00048335403565783245, 'samples': 3895296, 'steps': 7607, 'loss/train': 2.3229684829711914} 02/24/2022 11:25:32 - INFO - codeparrot_training - Step 7608: {'lr': 0.0004833481643782489, 'samples': 3895808, 'steps': 7608, 'loss/train': 9.008742332458496} 02/24/2022 11:25:36 - INFO - codeparrot_training - Step 7609: {'lr': 0.000483342292099076, 'samples': 3896320, 'steps': 7609, 'loss/train': 1.9355552196502686} 02/24/2022 11:25:41 - INFO - codeparrot_training - Step 7610: {'lr': 0.0004833364188203387, 'samples': 3896832, 'steps': 7610, 'loss/train': 1.3129560947418213} 02/24/2022 11:25:45 - INFO - codeparrot_training - Step 7611: {'lr': 0.0004833305445420624, 'samples': 3897344, 'steps': 7611, 'loss/train': 0.2687065303325653} 02/24/2022 11:25:50 - INFO - codeparrot_training - Step 7612: {'lr': 0.0004833246692642721, 'samples': 3897856, 'steps': 7612, 'loss/train': 1.6502118110656738} 02/24/2022 11:25:54 - INFO - codeparrot_training - Step 7613: {'lr': 0.000483318792986993, 'samples': 3898368, 'steps': 7613, 'loss/train': 2.6306190490722656} 02/24/2022 11:25:59 - INFO - codeparrot_training - Step 7614: {'lr': 0.00048331291571025026, 'samples': 3898880, 'steps': 7614, 'loss/train': 2.7483718395233154} 02/24/2022 11:26:03 - INFO - codeparrot_training - Step 7615: {'lr': 0.0004833070374340691, 'samples': 3899392, 'steps': 7615, 'loss/train': 1.1649190187454224} 02/24/2022 11:26:10 - INFO - codeparrot_training - Step 7616: {'lr': 0.00048330115815847465, 'samples': 3899904, 'steps': 7616, 'loss/train': 1.7381356954574585} 02/24/2022 11:26:13 - INFO - codeparrot_training - Step 7617: {'lr': 0.00048329527788349216, 'samples': 3900416, 'steps': 7617, 'loss/train': 1.486620545387268} 02/24/2022 11:26:19 - INFO - codeparrot_training - Step 7618: {'lr': 0.0004832893966091467, 'samples': 3900928, 'steps': 7618, 'loss/train': 2.6989986896514893} 02/24/2022 11:26:22 - INFO - codeparrot_training - Step 7619: {'lr': 0.00048328351433546364, 'samples': 3901440, 'steps': 7619, 'loss/train': 1.774891972541809} 02/24/2022 11:26:27 - INFO - codeparrot_training - Step 7620: {'lr': 0.000483277631062468, 'samples': 3901952, 'steps': 7620, 'loss/train': 1.8501554727554321} 02/24/2022 11:26:31 - INFO - codeparrot_training - Step 7621: {'lr': 0.00048327174679018515, 'samples': 3902464, 'steps': 7621, 'loss/train': 3.5334436893463135} 02/24/2022 11:26:36 - INFO - codeparrot_training - Step 7622: {'lr': 0.00048326586151864015, 'samples': 3902976, 'steps': 7622, 'loss/train': 2.5045037269592285} 02/24/2022 11:26:40 - INFO - codeparrot_training - Step 7623: {'lr': 0.00048325997524785826, 'samples': 3903488, 'steps': 7623, 'loss/train': 2.4404690265655518} 02/24/2022 11:26:45 - INFO - codeparrot_training - Step 7624: {'lr': 0.00048325408797786476, 'samples': 3904000, 'steps': 7624, 'loss/train': 2.6604700088500977} 02/24/2022 11:26:49 - INFO - codeparrot_training - Step 7625: {'lr': 0.00048324819970868473, 'samples': 3904512, 'steps': 7625, 'loss/train': 1.8641486167907715} 02/24/2022 11:26:55 - INFO - codeparrot_training - Step 7626: {'lr': 0.0004832423104403435, 'samples': 3905024, 'steps': 7626, 'loss/train': 3.499929904937744} 02/24/2022 11:26:58 - INFO - codeparrot_training - Step 7627: {'lr': 0.0004832364201728663, 'samples': 3905536, 'steps': 7627, 'loss/train': 2.5727362632751465} 02/24/2022 11:27:04 - INFO - codeparrot_training - Step 7628: {'lr': 0.0004832305289062784, 'samples': 3906048, 'steps': 7628, 'loss/train': 2.061204671859741} 02/24/2022 11:27:07 - INFO - codeparrot_training - Step 7629: {'lr': 0.0004832246366406049, 'samples': 3906560, 'steps': 7629, 'loss/train': 1.4982973337173462} 02/24/2022 11:27:13 - INFO - codeparrot_training - Step 7630: {'lr': 0.00048321874337587105, 'samples': 3907072, 'steps': 7630, 'loss/train': 1.3451931476593018} 02/24/2022 11:27:16 - INFO - codeparrot_training - Step 7631: {'lr': 0.0004832128491121023, 'samples': 3907584, 'steps': 7631, 'loss/train': 1.123176097869873} 02/24/2022 11:27:23 - INFO - codeparrot_training - Step 7632: {'lr': 0.00048320695384932366, 'samples': 3908096, 'steps': 7632, 'loss/train': 2.087005853652954} 02/24/2022 11:27:26 - INFO - codeparrot_training - Step 7633: {'lr': 0.0004832010575875605, 'samples': 3908608, 'steps': 7633, 'loss/train': 1.8638484477996826} 02/24/2022 11:27:32 - INFO - codeparrot_training - Step 7634: {'lr': 0.0004831951603268381, 'samples': 3909120, 'steps': 7634, 'loss/train': 2.5032124519348145} 02/24/2022 11:27:35 - INFO - codeparrot_training - Step 7635: {'lr': 0.0004831892620671816, 'samples': 3909632, 'steps': 7635, 'loss/train': 1.503653645515442} 02/24/2022 11:27:41 - INFO - codeparrot_training - Step 7636: {'lr': 0.0004831833628086164, 'samples': 3910144, 'steps': 7636, 'loss/train': 1.7145801782608032} 02/24/2022 11:27:44 - INFO - codeparrot_training - Step 7637: {'lr': 0.0004831774625511677, 'samples': 3910656, 'steps': 7637, 'loss/train': 1.6310853958129883} 02/24/2022 11:27:50 - INFO - codeparrot_training - Step 7638: {'lr': 0.00048317156129486086, 'samples': 3911168, 'steps': 7638, 'loss/train': 1.7415744066238403} 02/24/2022 11:27:53 - INFO - codeparrot_training - Step 7639: {'lr': 0.000483165659039721, 'samples': 3911680, 'steps': 7639, 'loss/train': 2.416231393814087} 02/24/2022 11:27:59 - INFO - codeparrot_training - Step 7640: {'lr': 0.0004831597557857735, 'samples': 3912192, 'steps': 7640, 'loss/train': 2.7516090869903564} 02/24/2022 11:28:02 - INFO - codeparrot_training - Step 7641: {'lr': 0.0004831538515330437, 'samples': 3912704, 'steps': 7641, 'loss/train': 2.9884231090545654} 02/24/2022 11:28:09 - INFO - codeparrot_training - Step 7642: {'lr': 0.0004831479462815568, 'samples': 3913216, 'steps': 7642, 'loss/train': 2.332920551300049} 02/24/2022 11:28:12 - INFO - codeparrot_training - Step 7643: {'lr': 0.00048314204003133815, 'samples': 3913728, 'steps': 7643, 'loss/train': 3.0780816078186035} 02/24/2022 11:28:18 - INFO - codeparrot_training - Step 7644: {'lr': 0.00048313613278241305, 'samples': 3914240, 'steps': 7644, 'loss/train': 2.5479977130889893} 02/24/2022 11:28:21 - INFO - codeparrot_training - Step 7645: {'lr': 0.0004831302245348068, 'samples': 3914752, 'steps': 7645, 'loss/train': 1.8576456308364868} 02/24/2022 11:28:27 - INFO - codeparrot_training - Step 7646: {'lr': 0.0004831243152885446, 'samples': 3915264, 'steps': 7646, 'loss/train': 2.0708115100860596} 02/24/2022 11:28:32 - INFO - codeparrot_training - Step 7647: {'lr': 0.0004831184050436519, 'samples': 3915776, 'steps': 7647, 'loss/train': 0.713884174823761} 02/24/2022 11:28:36 - INFO - codeparrot_training - Step 7648: {'lr': 0.000483112493800154, 'samples': 3916288, 'steps': 7648, 'loss/train': 3.53208065032959} 02/24/2022 11:28:41 - INFO - codeparrot_training - Step 7649: {'lr': 0.0004831065815580762, 'samples': 3916800, 'steps': 7649, 'loss/train': 2.173996686935425} 02/24/2022 11:28:45 - INFO - codeparrot_training - Step 7650: {'lr': 0.0004831006683174438, 'samples': 3917312, 'steps': 7650, 'loss/train': 0.8287095427513123} 02/24/2022 11:28:51 - INFO - codeparrot_training - Step 7651: {'lr': 0.0004830947540782822, 'samples': 3917824, 'steps': 7651, 'loss/train': 2.5224125385284424} 02/24/2022 11:28:55 - INFO - codeparrot_training - Step 7652: {'lr': 0.0004830888388406166, 'samples': 3918336, 'steps': 7652, 'loss/train': 1.5298198461532593} 02/24/2022 11:29:00 - INFO - codeparrot_training - Step 7653: {'lr': 0.0004830829226044725, 'samples': 3918848, 'steps': 7653, 'loss/train': 1.9724185466766357} 02/24/2022 11:29:03 - INFO - codeparrot_training - Step 7654: {'lr': 0.0004830770053698752, 'samples': 3919360, 'steps': 7654, 'loss/train': 2.5069825649261475} 02/24/2022 11:29:09 - INFO - codeparrot_training - Step 7655: {'lr': 0.00048307108713684994, 'samples': 3919872, 'steps': 7655, 'loss/train': 2.3742592334747314} 02/24/2022 11:29:12 - INFO - codeparrot_training - Step 7656: {'lr': 0.00048306516790542223, 'samples': 3920384, 'steps': 7656, 'loss/train': 1.044836401939392} 02/24/2022 11:29:18 - INFO - codeparrot_training - Step 7657: {'lr': 0.00048305924767561725, 'samples': 3920896, 'steps': 7657, 'loss/train': 2.8518283367156982} 02/24/2022 11:29:21 - INFO - codeparrot_training - Step 7658: {'lr': 0.00048305332644746053, 'samples': 3921408, 'steps': 7658, 'loss/train': 2.051703929901123} 02/24/2022 11:29:27 - INFO - codeparrot_training - Step 7659: {'lr': 0.0004830474042209774, 'samples': 3921920, 'steps': 7659, 'loss/train': 1.8091627359390259} 02/24/2022 11:29:30 - INFO - codeparrot_training - Step 7660: {'lr': 0.00048304148099619304, 'samples': 3922432, 'steps': 7660, 'loss/train': 1.4749717712402344} 02/24/2022 11:29:36 - INFO - codeparrot_training - Step 7661: {'lr': 0.0004830355567731331, 'samples': 3922944, 'steps': 7661, 'loss/train': 2.0379819869995117} 02/24/2022 11:29:39 - INFO - codeparrot_training - Step 7662: {'lr': 0.0004830296315518228, 'samples': 3923456, 'steps': 7662, 'loss/train': 2.104915142059326} 02/24/2022 11:29:46 - INFO - codeparrot_training - Step 7663: {'lr': 0.00048302370533228754, 'samples': 3923968, 'steps': 7663, 'loss/train': 2.3708560466766357} 02/24/2022 11:29:49 - INFO - codeparrot_training - Step 7664: {'lr': 0.00048301777811455274, 'samples': 3924480, 'steps': 7664, 'loss/train': 2.7781424522399902} 02/24/2022 11:29:54 - INFO - codeparrot_training - Step 7665: {'lr': 0.0004830118498986438, 'samples': 3924992, 'steps': 7665, 'loss/train': 2.476651191711426} 02/24/2022 11:29:58 - INFO - codeparrot_training - Step 7666: {'lr': 0.000483005920684586, 'samples': 3925504, 'steps': 7666, 'loss/train': 2.8014779090881348} 02/24/2022 11:30:04 - INFO - codeparrot_training - Step 7667: {'lr': 0.0004829999904724049, 'samples': 3926016, 'steps': 7667, 'loss/train': 2.835610866546631} 02/24/2022 11:30:07 - INFO - codeparrot_training - Step 7668: {'lr': 0.0004829940592621258, 'samples': 3926528, 'steps': 7668, 'loss/train': 2.361767292022705} 02/24/2022 11:30:12 - INFO - codeparrot_training - Step 7669: {'lr': 0.00048298812705377414, 'samples': 3927040, 'steps': 7669, 'loss/train': 2.4721925258636475} 02/24/2022 11:30:16 - INFO - codeparrot_training - Step 7670: {'lr': 0.0004829821938473753, 'samples': 3927552, 'steps': 7670, 'loss/train': 2.1274666786193848} 02/24/2022 11:30:21 - INFO - codeparrot_training - Step 7671: {'lr': 0.0004829762596429548, 'samples': 3928064, 'steps': 7671, 'loss/train': 1.3183720111846924} 02/24/2022 11:30:25 - INFO - codeparrot_training - Step 7672: {'lr': 0.0004829703244405379, 'samples': 3928576, 'steps': 7672, 'loss/train': 2.008500576019287} 02/24/2022 11:30:31 - INFO - codeparrot_training - Step 7673: {'lr': 0.0004829643882401501, 'samples': 3929088, 'steps': 7673, 'loss/train': 2.1050355434417725} 02/24/2022 11:30:34 - INFO - codeparrot_training - Step 7674: {'lr': 0.0004829584510418169, 'samples': 3929600, 'steps': 7674, 'loss/train': 2.4840757846832275} 02/24/2022 11:30:40 - INFO - codeparrot_training - Step 7675: {'lr': 0.00048295251284556363, 'samples': 3930112, 'steps': 7675, 'loss/train': 2.3247227668762207} 02/24/2022 11:30:43 - INFO - codeparrot_training - Step 7676: {'lr': 0.0004829465736514157, 'samples': 3930624, 'steps': 7676, 'loss/train': 2.8374712467193604} 02/24/2022 11:30:49 - INFO - codeparrot_training - Step 7677: {'lr': 0.00048294063345939877, 'samples': 3931136, 'steps': 7677, 'loss/train': 1.3187438249588013} 02/24/2022 11:30:53 - INFO - codeparrot_training - Step 7678: {'lr': 0.000482934692269538, 'samples': 3931648, 'steps': 7678, 'loss/train': 2.2495052814483643} 02/24/2022 11:30:58 - INFO - codeparrot_training - Step 7679: {'lr': 0.00048292875008185896, 'samples': 3932160, 'steps': 7679, 'loss/train': 1.7550535202026367} 02/24/2022 11:31:02 - INFO - codeparrot_training - Step 7680: {'lr': 0.0004829228068963872, 'samples': 3932672, 'steps': 7680, 'loss/train': 2.868255615234375} 02/24/2022 11:31:07 - INFO - codeparrot_training - Step 7681: {'lr': 0.00048291686271314816, 'samples': 3933184, 'steps': 7681, 'loss/train': 1.4639145135879517} 02/24/2022 11:31:11 - INFO - codeparrot_training - Step 7682: {'lr': 0.0004829109175321671, 'samples': 3933696, 'steps': 7682, 'loss/train': 3.0324151515960693} 02/24/2022 11:31:16 - INFO - codeparrot_training - Step 7683: {'lr': 0.00048290497135346965, 'samples': 3934208, 'steps': 7683, 'loss/train': 2.0329792499542236} 02/24/2022 11:31:20 - INFO - codeparrot_training - Step 7684: {'lr': 0.0004828990241770813, 'samples': 3934720, 'steps': 7684, 'loss/train': 2.7160253524780273} 02/24/2022 11:31:25 - INFO - codeparrot_training - Step 7685: {'lr': 0.0004828930760030275, 'samples': 3935232, 'steps': 7685, 'loss/train': 3.1744160652160645} 02/24/2022 11:31:29 - INFO - codeparrot_training - Step 7686: {'lr': 0.0004828871268313337, 'samples': 3935744, 'steps': 7686, 'loss/train': 2.5418992042541504} 02/24/2022 11:31:34 - INFO - codeparrot_training - Step 7687: {'lr': 0.0004828811766620254, 'samples': 3936256, 'steps': 7687, 'loss/train': 4.689566612243652} 02/24/2022 11:31:38 - INFO - codeparrot_training - Step 7688: {'lr': 0.00048287522549512806, 'samples': 3936768, 'steps': 7688, 'loss/train': 1.0280145406723022} 02/24/2022 11:31:44 - INFO - codeparrot_training - Step 7689: {'lr': 0.0004828692733306672, 'samples': 3937280, 'steps': 7689, 'loss/train': 1.7278858423233032} 02/24/2022 11:31:47 - INFO - codeparrot_training - Step 7690: {'lr': 0.0004828633201686684, 'samples': 3937792, 'steps': 7690, 'loss/train': 2.4728946685791016} 02/24/2022 11:31:53 - INFO - codeparrot_training - Step 7691: {'lr': 0.00048285736600915696, 'samples': 3938304, 'steps': 7691, 'loss/train': 2.1119372844696045} 02/24/2022 11:31:56 - INFO - codeparrot_training - Step 7692: {'lr': 0.00048285141085215857, 'samples': 3938816, 'steps': 7692, 'loss/train': 0.23492492735385895} 02/24/2022 11:32:02 - INFO - codeparrot_training - Step 7693: {'lr': 0.0004828454546976987, 'samples': 3939328, 'steps': 7693, 'loss/train': 2.8611409664154053} 02/24/2022 11:32:05 - INFO - codeparrot_training - Step 7694: {'lr': 0.00048283949754580283, 'samples': 3939840, 'steps': 7694, 'loss/train': 1.9664379358291626} 02/24/2022 11:32:11 - INFO - codeparrot_training - Step 7695: {'lr': 0.00048283353939649644, 'samples': 3940352, 'steps': 7695, 'loss/train': 1.3437795639038086} 02/24/2022 11:32:14 - INFO - codeparrot_training - Step 7696: {'lr': 0.0004828275802498051, 'samples': 3940864, 'steps': 7696, 'loss/train': 2.141622543334961} 02/24/2022 11:32:20 - INFO - codeparrot_training - Step 7697: {'lr': 0.0004828216201057544, 'samples': 3941376, 'steps': 7697, 'loss/train': 0.8001936674118042} 02/24/2022 11:32:23 - INFO - codeparrot_training - Step 7698: {'lr': 0.00048281565896436966, 'samples': 3941888, 'steps': 7698, 'loss/train': 1.9529004096984863} 02/24/2022 11:32:29 - INFO - codeparrot_training - Step 7699: {'lr': 0.0004828096968256767, 'samples': 3942400, 'steps': 7699, 'loss/train': 2.188166856765747} 02/24/2022 11:32:33 - INFO - codeparrot_training - Step 7700: {'lr': 0.00048280373368970086, 'samples': 3942912, 'steps': 7700, 'loss/train': 2.3147594928741455} 02/24/2022 11:32:38 - INFO - codeparrot_training - Step 7701: {'lr': 0.0004827977695564678, 'samples': 3943424, 'steps': 7701, 'loss/train': 2.0119380950927734} 02/24/2022 11:32:42 - INFO - codeparrot_training - Step 7702: {'lr': 0.000482791804426003, 'samples': 3943936, 'steps': 7702, 'loss/train': 1.545360803604126} 02/24/2022 11:32:48 - INFO - codeparrot_training - Step 7703: {'lr': 0.00048278583829833207, 'samples': 3944448, 'steps': 7703, 'loss/train': 2.4977641105651855} 02/24/2022 11:32:51 - INFO - codeparrot_training - Step 7704: {'lr': 0.00048277987117348043, 'samples': 3944960, 'steps': 7704, 'loss/train': 8.53597354888916} 02/24/2022 11:32:57 - INFO - codeparrot_training - Step 7705: {'lr': 0.00048277390305147386, 'samples': 3945472, 'steps': 7705, 'loss/train': 0.19285978376865387} 02/24/2022 11:33:00 - INFO - codeparrot_training - Step 7706: {'lr': 0.0004827679339323377, 'samples': 3945984, 'steps': 7706, 'loss/train': 2.566706895828247} 02/24/2022 11:33:06 - INFO - codeparrot_training - Step 7707: {'lr': 0.0004827619638160977, 'samples': 3946496, 'steps': 7707, 'loss/train': 1.779285192489624} 02/24/2022 11:33:09 - INFO - codeparrot_training - Step 7708: {'lr': 0.00048275599270277927, 'samples': 3947008, 'steps': 7708, 'loss/train': 2.663087844848633} 02/24/2022 11:33:15 - INFO - codeparrot_training - Step 7709: {'lr': 0.00048275002059240815, 'samples': 3947520, 'steps': 7709, 'loss/train': 1.724759578704834} 02/24/2022 11:33:19 - INFO - codeparrot_training - Step 7710: {'lr': 0.00048274404748500975, 'samples': 3948032, 'steps': 7710, 'loss/train': 2.8975305557250977} 02/24/2022 11:33:24 - INFO - codeparrot_training - Step 7711: {'lr': 0.0004827380733806099, 'samples': 3948544, 'steps': 7711, 'loss/train': 2.604024648666382} 02/24/2022 11:33:28 - INFO - codeparrot_training - Step 7712: {'lr': 0.0004827320982792339, 'samples': 3949056, 'steps': 7712, 'loss/train': 2.541714668273926} 02/24/2022 11:33:33 - INFO - codeparrot_training - Step 7713: {'lr': 0.0004827261221809076, 'samples': 3949568, 'steps': 7713, 'loss/train': 2.7684316635131836} 02/24/2022 11:33:37 - INFO - codeparrot_training - Step 7714: {'lr': 0.00048272014508565645, 'samples': 3950080, 'steps': 7714, 'loss/train': 1.9567121267318726} 02/24/2022 11:33:43 - INFO - codeparrot_training - Step 7715: {'lr': 0.00048271416699350613, 'samples': 3950592, 'steps': 7715, 'loss/train': 1.8389208316802979} 02/24/2022 11:33:46 - INFO - codeparrot_training - Step 7716: {'lr': 0.0004827081879044821, 'samples': 3951104, 'steps': 7716, 'loss/train': 1.0541294813156128} 02/24/2022 11:33:52 - INFO - codeparrot_training - Step 7717: {'lr': 0.00048270220781861025, 'samples': 3951616, 'steps': 7717, 'loss/train': 0.4620235562324524} 02/24/2022 11:33:55 - INFO - codeparrot_training - Step 7718: {'lr': 0.000482696226735916, 'samples': 3952128, 'steps': 7718, 'loss/train': 1.644763469696045} 02/24/2022 11:34:01 - INFO - codeparrot_training - Step 7719: {'lr': 0.00048269024465642487, 'samples': 3952640, 'steps': 7719, 'loss/train': 2.339843988418579} 02/24/2022 11:34:04 - INFO - codeparrot_training - Step 7720: {'lr': 0.00048268426158016274, 'samples': 3953152, 'steps': 7720, 'loss/train': 1.6591130495071411} 02/24/2022 11:34:10 - INFO - codeparrot_training - Step 7721: {'lr': 0.0004826782775071551, 'samples': 3953664, 'steps': 7721, 'loss/train': 2.216982364654541} 02/24/2022 11:34:13 - INFO - codeparrot_training - Step 7722: {'lr': 0.00048267229243742753, 'samples': 3954176, 'steps': 7722, 'loss/train': 1.771715521812439} 02/24/2022 11:34:19 - INFO - codeparrot_training - Step 7723: {'lr': 0.00048266630637100585, 'samples': 3954688, 'steps': 7723, 'loss/train': 2.678609848022461} 02/24/2022 11:34:22 - INFO - codeparrot_training - Step 7724: {'lr': 0.00048266031930791555, 'samples': 3955200, 'steps': 7724, 'loss/train': 2.1432509422302246} 02/24/2022 11:34:28 - INFO - codeparrot_training - Step 7725: {'lr': 0.00048265433124818226, 'samples': 3955712, 'steps': 7725, 'loss/train': 0.9142553806304932} 02/24/2022 11:34:32 - INFO - codeparrot_training - Step 7726: {'lr': 0.00048264834219183175, 'samples': 3956224, 'steps': 7726, 'loss/train': 1.7260949611663818} 02/24/2022 11:34:37 - INFO - codeparrot_training - Step 7727: {'lr': 0.00048264235213888964, 'samples': 3956736, 'steps': 7727, 'loss/train': 1.4306950569152832} 02/24/2022 11:34:41 - INFO - codeparrot_training - Step 7728: {'lr': 0.00048263636108938153, 'samples': 3957248, 'steps': 7728, 'loss/train': 1.7211476564407349} 02/24/2022 11:34:46 - INFO - codeparrot_training - Step 7729: {'lr': 0.0004826303690433331, 'samples': 3957760, 'steps': 7729, 'loss/train': 2.3091349601745605} 02/24/2022 11:34:50 - INFO - codeparrot_training - Step 7730: {'lr': 0.0004826243760007701, 'samples': 3958272, 'steps': 7730, 'loss/train': 2.0332255363464355} 02/24/2022 11:34:56 - INFO - codeparrot_training - Step 7731: {'lr': 0.00048261838196171804, 'samples': 3958784, 'steps': 7731, 'loss/train': 2.8401684761047363} 02/24/2022 11:34:59 - INFO - codeparrot_training - Step 7732: {'lr': 0.0004826123869262028, 'samples': 3959296, 'steps': 7732, 'loss/train': 1.2814948558807373} 02/24/2022 11:35:05 - INFO - codeparrot_training - Step 7733: {'lr': 0.0004826063908942499, 'samples': 3959808, 'steps': 7733, 'loss/train': 2.7540884017944336} 02/24/2022 11:35:08 - INFO - codeparrot_training - Step 7734: {'lr': 0.00048260039386588513, 'samples': 3960320, 'steps': 7734, 'loss/train': 0.18227140605449677} 02/24/2022 11:35:15 - INFO - codeparrot_training - Step 7735: {'lr': 0.00048259439584113405, 'samples': 3960832, 'steps': 7735, 'loss/train': 0.3952483534812927} 02/24/2022 11:35:18 - INFO - codeparrot_training - Step 7736: {'lr': 0.00048258839682002253, 'samples': 3961344, 'steps': 7736, 'loss/train': 0.6021316051483154} 02/24/2022 11:35:24 - INFO - codeparrot_training - Step 7737: {'lr': 0.0004825823968025761, 'samples': 3961856, 'steps': 7737, 'loss/train': 0.1311848759651184} 02/24/2022 11:35:27 - INFO - codeparrot_training - Step 7738: {'lr': 0.0004825763957888206, 'samples': 3962368, 'steps': 7738, 'loss/train': 2.5508365631103516} 02/24/2022 11:35:33 - INFO - codeparrot_training - Step 7739: {'lr': 0.00048257039377878165, 'samples': 3962880, 'steps': 7739, 'loss/train': 2.9970643520355225} 02/24/2022 11:35:36 - INFO - codeparrot_training - Step 7740: {'lr': 0.00048256439077248495, 'samples': 3963392, 'steps': 7740, 'loss/train': 1.5972727537155151} 02/24/2022 11:35:42 - INFO - codeparrot_training - Step 7741: {'lr': 0.00048255838676995624, 'samples': 3963904, 'steps': 7741, 'loss/train': 1.511894941329956} 02/24/2022 11:35:45 - INFO - codeparrot_training - Step 7742: {'lr': 0.00048255238177122127, 'samples': 3964416, 'steps': 7742, 'loss/train': 3.0409815311431885} 02/24/2022 11:35:51 - INFO - codeparrot_training - Step 7743: {'lr': 0.0004825463757763058, 'samples': 3964928, 'steps': 7743, 'loss/train': 2.1414434909820557} 02/24/2022 11:35:54 - INFO - codeparrot_training - Step 7744: {'lr': 0.00048254036878523537, 'samples': 3965440, 'steps': 7744, 'loss/train': 1.647324562072754} 02/24/2022 11:36:00 - INFO - codeparrot_training - Step 7745: {'lr': 0.00048253436079803594, 'samples': 3965952, 'steps': 7745, 'loss/train': 2.089505910873413} 02/24/2022 11:36:04 - INFO - codeparrot_training - Step 7746: {'lr': 0.0004825283518147331, 'samples': 3966464, 'steps': 7746, 'loss/train': 1.6967997550964355} 02/24/2022 11:36:09 - INFO - codeparrot_training - Step 7747: {'lr': 0.00048252234183535265, 'samples': 3966976, 'steps': 7747, 'loss/train': 2.582773447036743} 02/24/2022 11:36:13 - INFO - codeparrot_training - Step 7748: {'lr': 0.0004825163308599203, 'samples': 3967488, 'steps': 7748, 'loss/train': 2.2498679161071777} 02/24/2022 11:36:18 - INFO - codeparrot_training - Step 7749: {'lr': 0.0004825103188884619, 'samples': 3968000, 'steps': 7749, 'loss/train': 1.5776172876358032} 02/24/2022 11:36:22 - INFO - codeparrot_training - Step 7750: {'lr': 0.000482504305921003, 'samples': 3968512, 'steps': 7750, 'loss/train': 2.4894113540649414} 02/24/2022 11:36:27 - INFO - codeparrot_training - Step 7751: {'lr': 0.00048249829195756954, 'samples': 3969024, 'steps': 7751, 'loss/train': 2.1953022480010986} 02/24/2022 11:36:30 - INFO - codeparrot_training - Step 7752: {'lr': 0.0004824922769981873, 'samples': 3969536, 'steps': 7752, 'loss/train': 2.151811122894287} 02/24/2022 11:36:36 - INFO - codeparrot_training - Step 7753: {'lr': 0.0004824862610428819, 'samples': 3970048, 'steps': 7753, 'loss/train': 3.196906566619873} 02/24/2022 11:36:39 - INFO - codeparrot_training - Step 7754: {'lr': 0.0004824802440916792, 'samples': 3970560, 'steps': 7754, 'loss/train': 3.315129041671753} 02/24/2022 11:36:46 - INFO - codeparrot_training - Step 7755: {'lr': 0.0004824742261446049, 'samples': 3971072, 'steps': 7755, 'loss/train': 0.5079135298728943} 02/24/2022 11:36:49 - INFO - codeparrot_training - Step 7756: {'lr': 0.0004824682072016849, 'samples': 3971584, 'steps': 7756, 'loss/train': 3.0118930339813232} 02/24/2022 11:36:55 - INFO - codeparrot_training - Step 7757: {'lr': 0.00048246218726294486, 'samples': 3972096, 'steps': 7757, 'loss/train': 2.474109411239624} 02/24/2022 11:36:58 - INFO - codeparrot_training - Step 7758: {'lr': 0.0004824561663284107, 'samples': 3972608, 'steps': 7758, 'loss/train': 0.28871428966522217} 02/24/2022 11:37:04 - INFO - codeparrot_training - Step 7759: {'lr': 0.0004824501443981081, 'samples': 3973120, 'steps': 7759, 'loss/train': 2.8747596740722656} 02/24/2022 11:37:07 - INFO - codeparrot_training - Step 7760: {'lr': 0.00048244412147206283, 'samples': 3973632, 'steps': 7760, 'loss/train': 0.5813701152801514} 02/24/2022 11:37:13 - INFO - codeparrot_training - Step 7761: {'lr': 0.00048243809755030086, 'samples': 3974144, 'steps': 7761, 'loss/train': 1.5651233196258545} 02/24/2022 11:37:16 - INFO - codeparrot_training - Step 7762: {'lr': 0.00048243207263284785, 'samples': 3974656, 'steps': 7762, 'loss/train': 2.242506742477417} 02/24/2022 11:37:22 - INFO - codeparrot_training - Step 7763: {'lr': 0.0004824260467197296, 'samples': 3975168, 'steps': 7763, 'loss/train': 1.8729777336120605} 02/24/2022 11:37:25 - INFO - codeparrot_training - Step 7764: {'lr': 0.000482420019810972, 'samples': 3975680, 'steps': 7764, 'loss/train': 1.818366527557373} 02/24/2022 11:37:31 - INFO - codeparrot_training - Step 7765: {'lr': 0.00048241399190660086, 'samples': 3976192, 'steps': 7765, 'loss/train': 2.531325578689575} 02/24/2022 11:37:35 - INFO - codeparrot_training - Step 7766: {'lr': 0.0004824079630066419, 'samples': 3976704, 'steps': 7766, 'loss/train': 2.398423910140991} 02/24/2022 11:37:40 - INFO - codeparrot_training - Step 7767: {'lr': 0.0004824019331111211, 'samples': 3977216, 'steps': 7767, 'loss/train': 3.0106937885284424} 02/24/2022 11:37:43 - INFO - codeparrot_training - Step 7768: {'lr': 0.0004823959022200642, 'samples': 3977728, 'steps': 7768, 'loss/train': 2.110288381576538} 02/24/2022 11:37:49 - INFO - codeparrot_training - Step 7769: {'lr': 0.00048238987033349706, 'samples': 3978240, 'steps': 7769, 'loss/train': 3.1188998222351074} 02/24/2022 11:37:52 - INFO - codeparrot_training - Step 7770: {'lr': 0.0004823838374514455, 'samples': 3978752, 'steps': 7770, 'loss/train': 1.878414511680603} 02/24/2022 11:37:59 - INFO - codeparrot_training - Step 7771: {'lr': 0.00048237780357393535, 'samples': 3979264, 'steps': 7771, 'loss/train': 1.8256481885910034} 02/24/2022 11:38:02 - INFO - codeparrot_training - Step 7772: {'lr': 0.00048237176870099256, 'samples': 3979776, 'steps': 7772, 'loss/train': 2.308722734451294} 02/24/2022 11:38:08 - INFO - codeparrot_training - Step 7773: {'lr': 0.0004823657328326427, 'samples': 3980288, 'steps': 7773, 'loss/train': 2.7769384384155273} 02/24/2022 11:38:11 - INFO - codeparrot_training - Step 7774: {'lr': 0.000482359695968912, 'samples': 3980800, 'steps': 7774, 'loss/train': 5.388213157653809} 02/24/2022 11:38:17 - INFO - codeparrot_training - Step 7775: {'lr': 0.0004823536581098261, 'samples': 3981312, 'steps': 7775, 'loss/train': 1.8067668676376343} 02/24/2022 11:38:20 - INFO - codeparrot_training - Step 7776: {'lr': 0.00048234761925541094, 'samples': 3981824, 'steps': 7776, 'loss/train': 3.3182616233825684} 02/24/2022 11:38:26 - INFO - codeparrot_training - Step 7777: {'lr': 0.0004823415794056923, 'samples': 3982336, 'steps': 7777, 'loss/train': 2.4434127807617188} 02/24/2022 11:38:29 - INFO - codeparrot_training - Step 7778: {'lr': 0.00048233553856069617, 'samples': 3982848, 'steps': 7778, 'loss/train': 2.5438945293426514} 02/24/2022 11:38:35 - INFO - codeparrot_training - Step 7779: {'lr': 0.00048232949672044834, 'samples': 3983360, 'steps': 7779, 'loss/train': 2.4991538524627686} 02/24/2022 11:38:38 - INFO - codeparrot_training - Step 7780: {'lr': 0.0004823234538849747, 'samples': 3983872, 'steps': 7780, 'loss/train': 2.984445333480835} 02/24/2022 11:38:44 - INFO - codeparrot_training - Step 7781: {'lr': 0.0004823174100543012, 'samples': 3984384, 'steps': 7781, 'loss/train': 2.254769802093506} 02/24/2022 11:38:48 - INFO - codeparrot_training - Step 7782: {'lr': 0.0004823113652284536, 'samples': 3984896, 'steps': 7782, 'loss/train': 1.9577986001968384} 02/24/2022 11:38:53 - INFO - codeparrot_training - Step 7783: {'lr': 0.00048230531940745793, 'samples': 3985408, 'steps': 7783, 'loss/train': 2.592783212661743} 02/24/2022 11:38:57 - INFO - codeparrot_training - Step 7784: {'lr': 0.0004822992725913401, 'samples': 3985920, 'steps': 7784, 'loss/train': 2.5873801708221436} 02/24/2022 11:39:02 - INFO - codeparrot_training - Step 7785: {'lr': 0.00048229322478012584, 'samples': 3986432, 'steps': 7785, 'loss/train': 2.0520689487457275} 02/24/2022 11:39:06 - INFO - codeparrot_training - Step 7786: {'lr': 0.0004822871759738412, 'samples': 3986944, 'steps': 7786, 'loss/train': 1.9086414575576782} 02/24/2022 11:39:11 - INFO - codeparrot_training - Step 7787: {'lr': 0.0004822811261725121, 'samples': 3987456, 'steps': 7787, 'loss/train': 2.414551258087158} 02/24/2022 11:39:15 - INFO - codeparrot_training - Step 7788: {'lr': 0.0004822750753761644, 'samples': 3987968, 'steps': 7788, 'loss/train': 2.2617695331573486} 02/24/2022 11:39:20 - INFO - codeparrot_training - Step 7789: {'lr': 0.00048226902358482405, 'samples': 3988480, 'steps': 7789, 'loss/train': 2.2469849586486816} 02/24/2022 11:39:24 - INFO - codeparrot_training - Step 7790: {'lr': 0.0004822629707985169, 'samples': 3988992, 'steps': 7790, 'loss/train': 3.169412851333618} 02/24/2022 11:39:30 - INFO - codeparrot_training - Step 7791: {'lr': 0.00048225691701726895, 'samples': 3989504, 'steps': 7791, 'loss/train': 2.17946457862854} 02/24/2022 11:39:34 - INFO - codeparrot_training - Step 7792: {'lr': 0.00048225086224110614, 'samples': 3990016, 'steps': 7792, 'loss/train': 0.47140467166900635} 02/24/2022 11:39:39 - INFO - codeparrot_training - Step 7793: {'lr': 0.00048224480647005437, 'samples': 3990528, 'steps': 7793, 'loss/train': 0.39862060546875} 02/24/2022 11:39:43 - INFO - codeparrot_training - Step 7794: {'lr': 0.0004822387497041396, 'samples': 3991040, 'steps': 7794, 'loss/train': 2.600405693054199} 02/24/2022 11:39:48 - INFO - codeparrot_training - Step 7795: {'lr': 0.00048223269194338776, 'samples': 3991552, 'steps': 7795, 'loss/train': 3.442826509475708} 02/24/2022 11:39:52 - INFO - codeparrot_training - Step 7796: {'lr': 0.0004822266331878248, 'samples': 3992064, 'steps': 7796, 'loss/train': 3.0081872940063477} 02/24/2022 11:39:57 - INFO - codeparrot_training - Step 7797: {'lr': 0.0004822205734374767, 'samples': 3992576, 'steps': 7797, 'loss/train': 1.3744837045669556} 02/24/2022 11:40:01 - INFO - codeparrot_training - Step 7798: {'lr': 0.00048221451269236937, 'samples': 3993088, 'steps': 7798, 'loss/train': 1.2777047157287598} 02/24/2022 11:40:06 - INFO - codeparrot_training - Step 7799: {'lr': 0.0004822084509525289, 'samples': 3993600, 'steps': 7799, 'loss/train': 2.253067970275879} 02/24/2022 11:40:10 - INFO - codeparrot_training - Step 7800: {'lr': 0.0004822023882179811, 'samples': 3994112, 'steps': 7800, 'loss/train': 2.231689929962158} 02/24/2022 11:40:15 - INFO - codeparrot_training - Step 7801: {'lr': 0.00048219632448875195, 'samples': 3994624, 'steps': 7801, 'loss/train': 2.465588331222534} 02/24/2022 11:40:21 - INFO - codeparrot_training - Step 7802: {'lr': 0.0004821902597648675, 'samples': 3995136, 'steps': 7802, 'loss/train': 2.680110454559326} 02/24/2022 11:40:24 - INFO - codeparrot_training - Step 7803: {'lr': 0.0004821841940463538, 'samples': 3995648, 'steps': 7803, 'loss/train': 1.8885477781295776} 02/24/2022 11:40:30 - INFO - codeparrot_training - Step 7804: {'lr': 0.0004821781273332366, 'samples': 3996160, 'steps': 7804, 'loss/train': 2.3127224445343018} 02/24/2022 11:40:33 - INFO - codeparrot_training - Step 7805: {'lr': 0.00048217205962554214, 'samples': 3996672, 'steps': 7805, 'loss/train': 2.5576257705688477} 02/24/2022 11:40:40 - INFO - codeparrot_training - Step 7806: {'lr': 0.0004821659909232963, 'samples': 3997184, 'steps': 7806, 'loss/train': 2.561901330947876} 02/24/2022 11:40:43 - INFO - codeparrot_training - Step 7807: {'lr': 0.000482159921226525, 'samples': 3997696, 'steps': 7807, 'loss/train': 2.2070515155792236} 02/24/2022 11:40:49 - INFO - codeparrot_training - Step 7808: {'lr': 0.00048215385053525434, 'samples': 3998208, 'steps': 7808, 'loss/train': 1.6095865964889526} 02/24/2022 11:40:52 - INFO - codeparrot_training - Step 7809: {'lr': 0.0004821477788495103, 'samples': 3998720, 'steps': 7809, 'loss/train': 2.2978527545928955} 02/24/2022 11:40:58 - INFO - codeparrot_training - Step 7810: {'lr': 0.0004821417061693189, 'samples': 3999232, 'steps': 7810, 'loss/train': 2.2504847049713135} 02/24/2022 11:41:01 - INFO - codeparrot_training - Step 7811: {'lr': 0.00048213563249470615, 'samples': 3999744, 'steps': 7811, 'loss/train': 2.2726497650146484} 02/24/2022 11:41:07 - INFO - codeparrot_training - Step 7812: {'lr': 0.00048212955782569805, 'samples': 4000256, 'steps': 7812, 'loss/train': 1.8694270849227905} 02/24/2022 11:41:10 - INFO - codeparrot_training - Step 7813: {'lr': 0.00048212348216232064, 'samples': 4000768, 'steps': 7813, 'loss/train': 1.7889246940612793} 02/24/2022 11:41:16 - INFO - codeparrot_training - Step 7814: {'lr': 0.0004821174055045999, 'samples': 4001280, 'steps': 7814, 'loss/train': 2.1357905864715576} 02/24/2022 11:41:19 - INFO - codeparrot_training - Step 7815: {'lr': 0.000482111327852562, 'samples': 4001792, 'steps': 7815, 'loss/train': 2.235445022583008} 02/24/2022 11:41:26 - INFO - codeparrot_training - Step 7816: {'lr': 0.0004821052492062328, 'samples': 4002304, 'steps': 7816, 'loss/train': 2.826972723007202} 02/24/2022 11:41:30 - INFO - codeparrot_training - Step 7817: {'lr': 0.0004820991695656385, 'samples': 4002816, 'steps': 7817, 'loss/train': 1.0054012537002563} 02/24/2022 11:41:33 - INFO - codeparrot_training - Step 7818: {'lr': 0.00048209308893080495, 'samples': 4003328, 'steps': 7818, 'loss/train': 2.198528528213501} 02/24/2022 11:41:39 - INFO - codeparrot_training - Step 7819: {'lr': 0.00048208700730175834, 'samples': 4003840, 'steps': 7819, 'loss/train': 2.274714708328247} 02/24/2022 11:41:42 - INFO - codeparrot_training - Step 7820: {'lr': 0.0004820809246785247, 'samples': 4004352, 'steps': 7820, 'loss/train': 3.313187599182129} 02/24/2022 11:41:48 - INFO - codeparrot_training - Step 7821: {'lr': 0.00048207484106113, 'samples': 4004864, 'steps': 7821, 'loss/train': 3.0961296558380127} 02/24/2022 11:41:51 - INFO - codeparrot_training - Step 7822: {'lr': 0.0004820687564496005, 'samples': 4005376, 'steps': 7822, 'loss/train': 0.856377363204956} 02/24/2022 11:41:57 - INFO - codeparrot_training - Step 7823: {'lr': 0.00048206267084396204, 'samples': 4005888, 'steps': 7823, 'loss/train': 2.9520039558410645} 02/24/2022 11:42:00 - INFO - codeparrot_training - Step 7824: {'lr': 0.0004820565842442408, 'samples': 4006400, 'steps': 7824, 'loss/train': 2.825042963027954} 02/24/2022 11:42:06 - INFO - codeparrot_training - Step 7825: {'lr': 0.00048205049665046287, 'samples': 4006912, 'steps': 7825, 'loss/train': 3.0491065979003906} 02/24/2022 11:42:09 - INFO - codeparrot_training - Step 7826: {'lr': 0.0004820444080626543, 'samples': 4007424, 'steps': 7826, 'loss/train': 2.5948143005371094} 02/24/2022 11:42:15 - INFO - codeparrot_training - Step 7827: {'lr': 0.00048203831848084115, 'samples': 4007936, 'steps': 7827, 'loss/train': 3.128384590148926} 02/24/2022 11:42:19 - INFO - codeparrot_training - Step 7828: {'lr': 0.0004820322279050495, 'samples': 4008448, 'steps': 7828, 'loss/train': 1.2974281311035156} 02/24/2022 11:42:24 - INFO - codeparrot_training - Step 7829: {'lr': 0.00048202613633530555, 'samples': 4008960, 'steps': 7829, 'loss/train': 2.6432459354400635} 02/24/2022 11:42:28 - INFO - codeparrot_training - Step 7830: {'lr': 0.00048202004377163524, 'samples': 4009472, 'steps': 7830, 'loss/train': 2.479123592376709} 02/24/2022 11:42:33 - INFO - codeparrot_training - Step 7831: {'lr': 0.00048201395021406476, 'samples': 4009984, 'steps': 7831, 'loss/train': 1.8705668449401855} 02/24/2022 11:42:37 - INFO - codeparrot_training - Step 7832: {'lr': 0.0004820078556626202, 'samples': 4010496, 'steps': 7832, 'loss/train': 1.2155200242996216} 02/24/2022 11:42:42 - INFO - codeparrot_training - Step 7833: {'lr': 0.0004820017601173276, 'samples': 4011008, 'steps': 7833, 'loss/train': 2.29663348197937} 02/24/2022 11:42:46 - INFO - codeparrot_training - Step 7834: {'lr': 0.00048199566357821314, 'samples': 4011520, 'steps': 7834, 'loss/train': 2.2540931701660156} 02/24/2022 11:42:51 - INFO - codeparrot_training - Step 7835: {'lr': 0.00048198956604530297, 'samples': 4012032, 'steps': 7835, 'loss/train': 2.5729522705078125} 02/24/2022 11:42:55 - INFO - codeparrot_training - Step 7836: {'lr': 0.0004819834675186231, 'samples': 4012544, 'steps': 7836, 'loss/train': 2.475855827331543} 02/24/2022 11:43:00 - INFO - codeparrot_training - Step 7837: {'lr': 0.0004819773679981998, 'samples': 4013056, 'steps': 7837, 'loss/train': 2.0142223834991455} 02/24/2022 11:43:04 - INFO - codeparrot_training - Step 7838: {'lr': 0.0004819712674840591, 'samples': 4013568, 'steps': 7838, 'loss/train': 1.6875241994857788} 02/24/2022 11:43:10 - INFO - codeparrot_training - Step 7839: {'lr': 0.00048196516597622706, 'samples': 4014080, 'steps': 7839, 'loss/train': 2.1893794536590576} 02/24/2022 11:43:13 - INFO - codeparrot_training - Step 7840: {'lr': 0.00048195906347473, 'samples': 4014592, 'steps': 7840, 'loss/train': 1.9948806762695312} 02/24/2022 11:43:19 - INFO - codeparrot_training - Step 7841: {'lr': 0.00048195295997959393, 'samples': 4015104, 'steps': 7841, 'loss/train': 2.6629714965820312} 02/24/2022 11:43:22 - INFO - codeparrot_training - Step 7842: {'lr': 0.00048194685549084507, 'samples': 4015616, 'steps': 7842, 'loss/train': 2.457061290740967} 02/24/2022 11:43:28 - INFO - codeparrot_training - Step 7843: {'lr': 0.00048194075000850944, 'samples': 4016128, 'steps': 7843, 'loss/train': 2.4506239891052246} 02/24/2022 11:43:31 - INFO - codeparrot_training - Step 7844: {'lr': 0.0004819346435326134, 'samples': 4016640, 'steps': 7844, 'loss/train': 2.20335054397583} 02/24/2022 11:43:37 - INFO - codeparrot_training - Step 7845: {'lr': 0.000481928536063183, 'samples': 4017152, 'steps': 7845, 'loss/train': 1.9625605344772339} 02/24/2022 11:43:40 - INFO - codeparrot_training - Step 7846: {'lr': 0.0004819224276002443, 'samples': 4017664, 'steps': 7846, 'loss/train': 2.8505752086639404} 02/24/2022 11:43:46 - INFO - codeparrot_training - Step 7847: {'lr': 0.0004819163181438236, 'samples': 4018176, 'steps': 7847, 'loss/train': 1.4816619157791138} 02/24/2022 11:43:49 - INFO - codeparrot_training - Step 7848: {'lr': 0.000481910207693947, 'samples': 4018688, 'steps': 7848, 'loss/train': 2.219057559967041} 02/24/2022 11:43:55 - INFO - codeparrot_training - Step 7849: {'lr': 0.0004819040962506408, 'samples': 4019200, 'steps': 7849, 'loss/train': 1.3804471492767334} 02/24/2022 11:43:58 - INFO - codeparrot_training - Step 7850: {'lr': 0.000481897983813931, 'samples': 4019712, 'steps': 7850, 'loss/train': 1.5484976768493652} 02/24/2022 11:44:04 - INFO - codeparrot_training - Step 7851: {'lr': 0.00048189187038384396, 'samples': 4020224, 'steps': 7851, 'loss/train': 3.0784573554992676} 02/24/2022 11:44:07 - INFO - codeparrot_training - Step 7852: {'lr': 0.00048188575596040575, 'samples': 4020736, 'steps': 7852, 'loss/train': 2.2429795265197754} 02/24/2022 11:44:14 - INFO - codeparrot_training - Step 7853: {'lr': 0.00048187964054364254, 'samples': 4021248, 'steps': 7853, 'loss/train': 2.7370266914367676} 02/24/2022 11:44:19 - INFO - codeparrot_training - Step 7854: {'lr': 0.0004818735241335807, 'samples': 4021760, 'steps': 7854, 'loss/train': 1.6324163675308228} 02/24/2022 11:44:23 - INFO - codeparrot_training - Step 7855: {'lr': 0.00048186740673024614, 'samples': 4022272, 'steps': 7855, 'loss/train': 1.7768681049346924} 02/24/2022 11:44:28 - INFO - codeparrot_training - Step 7856: {'lr': 0.00048186128833366536, 'samples': 4022784, 'steps': 7856, 'loss/train': 2.867032289505005} 02/24/2022 11:44:31 - INFO - codeparrot_training - Step 7857: {'lr': 0.0004818551689438644, 'samples': 4023296, 'steps': 7857, 'loss/train': 2.4767074584960938} 02/24/2022 11:44:37 - INFO - codeparrot_training - Step 7858: {'lr': 0.00048184904856086953, 'samples': 4023808, 'steps': 7858, 'loss/train': 2.5349812507629395} 02/24/2022 11:44:40 - INFO - codeparrot_training - Step 7859: {'lr': 0.0004818429271847069, 'samples': 4024320, 'steps': 7859, 'loss/train': 2.389404058456421} 02/24/2022 11:44:46 - INFO - codeparrot_training - Step 7860: {'lr': 0.00048183680481540293, 'samples': 4024832, 'steps': 7860, 'loss/train': 2.769411087036133} 02/24/2022 11:44:49 - INFO - codeparrot_training - Step 7861: {'lr': 0.0004818306814529836, 'samples': 4025344, 'steps': 7861, 'loss/train': 2.349100112915039} 02/24/2022 11:44:55 - INFO - codeparrot_training - Step 7862: {'lr': 0.00048182455709747525, 'samples': 4025856, 'steps': 7862, 'loss/train': 2.544666051864624} 02/24/2022 11:44:58 - INFO - codeparrot_training - Step 7863: {'lr': 0.0004818184317489041, 'samples': 4026368, 'steps': 7863, 'loss/train': 2.097475290298462} 02/24/2022 11:45:04 - INFO - codeparrot_training - Step 7864: {'lr': 0.00048181230540729643, 'samples': 4026880, 'steps': 7864, 'loss/train': 1.2829848527908325} 02/24/2022 11:45:08 - INFO - codeparrot_training - Step 7865: {'lr': 0.00048180617807267844, 'samples': 4027392, 'steps': 7865, 'loss/train': 3.1394948959350586} 02/24/2022 11:45:13 - INFO - codeparrot_training - Step 7866: {'lr': 0.0004818000497450764, 'samples': 4027904, 'steps': 7866, 'loss/train': 1.9057754278182983} 02/24/2022 11:45:17 - INFO - codeparrot_training - Step 7867: {'lr': 0.00048179392042451655, 'samples': 4028416, 'steps': 7867, 'loss/train': 1.6702289581298828} 02/24/2022 11:45:22 - INFO - codeparrot_training - Step 7868: {'lr': 0.0004817877901110251, 'samples': 4028928, 'steps': 7868, 'loss/train': 2.5258617401123047} 02/24/2022 11:45:26 - INFO - codeparrot_training - Step 7869: {'lr': 0.00048178165880462845, 'samples': 4029440, 'steps': 7869, 'loss/train': 2.1727590560913086} 02/24/2022 11:45:31 - INFO - codeparrot_training - Step 7870: {'lr': 0.0004817755265053527, 'samples': 4029952, 'steps': 7870, 'loss/train': 2.1347312927246094} 02/24/2022 11:45:35 - INFO - codeparrot_training - Step 7871: {'lr': 0.0004817693932132242, 'samples': 4030464, 'steps': 7871, 'loss/train': 1.851364254951477} 02/24/2022 11:45:40 - INFO - codeparrot_training - Step 7872: {'lr': 0.0004817632589282693, 'samples': 4030976, 'steps': 7872, 'loss/train': 1.3407553434371948} 02/24/2022 11:45:44 - INFO - codeparrot_training - Step 7873: {'lr': 0.00048175712365051407, 'samples': 4031488, 'steps': 7873, 'loss/train': 2.8045785427093506} 02/24/2022 11:45:50 - INFO - codeparrot_training - Step 7874: {'lr': 0.00048175098737998504, 'samples': 4032000, 'steps': 7874, 'loss/train': 2.6032423973083496} 02/24/2022 11:45:53 - INFO - codeparrot_training - Step 7875: {'lr': 0.0004817448501167082, 'samples': 4032512, 'steps': 7875, 'loss/train': 1.6382709741592407} 02/24/2022 11:45:59 - INFO - codeparrot_training - Step 7876: {'lr': 0.0004817387118607102, 'samples': 4033024, 'steps': 7876, 'loss/train': 1.7451155185699463} 02/24/2022 11:46:02 - INFO - codeparrot_training - Step 7877: {'lr': 0.00048173257261201695, 'samples': 4033536, 'steps': 7877, 'loss/train': 1.59663724899292} 02/24/2022 11:46:08 - INFO - codeparrot_training - Step 7878: {'lr': 0.00048172643237065504, 'samples': 4034048, 'steps': 7878, 'loss/train': 1.880545973777771} 02/24/2022 11:46:11 - INFO - codeparrot_training - Step 7879: {'lr': 0.00048172029113665075, 'samples': 4034560, 'steps': 7879, 'loss/train': 2.4292068481445312} 02/24/2022 11:46:17 - INFO - codeparrot_training - Step 7880: {'lr': 0.0004817141489100302, 'samples': 4035072, 'steps': 7880, 'loss/train': 2.0743541717529297} 02/24/2022 11:46:20 - INFO - codeparrot_training - Step 7881: {'lr': 0.00048170800569081985, 'samples': 4035584, 'steps': 7881, 'loss/train': 2.9938931465148926} 02/24/2022 11:46:26 - INFO - codeparrot_training - Step 7882: {'lr': 0.000481701861479046, 'samples': 4036096, 'steps': 7882, 'loss/train': 2.032970428466797} 02/24/2022 11:46:29 - INFO - codeparrot_training - Step 7883: {'lr': 0.000481695716274735, 'samples': 4036608, 'steps': 7883, 'loss/train': 1.8447604179382324} 02/24/2022 11:46:36 - INFO - codeparrot_training - Step 7884: {'lr': 0.000481689570077913, 'samples': 4037120, 'steps': 7884, 'loss/train': 2.6717259883880615} 02/24/2022 11:46:39 - INFO - codeparrot_training - Step 7885: {'lr': 0.00048168342288860646, 'samples': 4037632, 'steps': 7885, 'loss/train': 2.673819065093994} 02/24/2022 11:46:44 - INFO - codeparrot_training - Step 7886: {'lr': 0.00048167727470684176, 'samples': 4038144, 'steps': 7886, 'loss/train': 1.8334239721298218} 02/24/2022 11:46:48 - INFO - codeparrot_training - Step 7887: {'lr': 0.0004816711255326452, 'samples': 4038656, 'steps': 7887, 'loss/train': 2.5789427757263184} 02/24/2022 11:46:53 - INFO - codeparrot_training - Step 7888: {'lr': 0.00048166497536604306, 'samples': 4039168, 'steps': 7888, 'loss/train': 2.0291426181793213} 02/24/2022 11:46:57 - INFO - codeparrot_training - Step 7889: {'lr': 0.00048165882420706175, 'samples': 4039680, 'steps': 7889, 'loss/train': 2.59883975982666} 02/24/2022 11:47:02 - INFO - codeparrot_training - Step 7890: {'lr': 0.0004816526720557276, 'samples': 4040192, 'steps': 7890, 'loss/train': 0.48059460520744324} 02/24/2022 11:47:06 - INFO - codeparrot_training - Step 7891: {'lr': 0.0004816465189120669, 'samples': 4040704, 'steps': 7891, 'loss/train': 2.426546573638916} 02/24/2022 11:47:12 - INFO - codeparrot_training - Step 7892: {'lr': 0.00048164036477610616, 'samples': 4041216, 'steps': 7892, 'loss/train': 2.4815969467163086} 02/24/2022 11:47:15 - INFO - codeparrot_training - Step 7893: {'lr': 0.0004816342096478716, 'samples': 4041728, 'steps': 7893, 'loss/train': 2.3024590015411377} 02/24/2022 11:47:21 - INFO - codeparrot_training - Step 7894: {'lr': 0.00048162805352738966, 'samples': 4042240, 'steps': 7894, 'loss/train': 2.7481632232666016} 02/24/2022 11:47:24 - INFO - codeparrot_training - Step 7895: {'lr': 0.0004816218964146867, 'samples': 4042752, 'steps': 7895, 'loss/train': 2.0822837352752686} 02/24/2022 11:47:30 - INFO - codeparrot_training - Step 7896: {'lr': 0.000481615738309789, 'samples': 4043264, 'steps': 7896, 'loss/train': 2.5999977588653564} 02/24/2022 11:47:33 - INFO - codeparrot_training - Step 7897: {'lr': 0.00048160957921272306, 'samples': 4043776, 'steps': 7897, 'loss/train': 2.230018377304077} 02/24/2022 11:47:39 - INFO - codeparrot_training - Step 7898: {'lr': 0.00048160341912351523, 'samples': 4044288, 'steps': 7898, 'loss/train': 1.790880799293518} 02/24/2022 11:47:45 - INFO - codeparrot_training - Step 7899: {'lr': 0.00048159725804219195, 'samples': 4044800, 'steps': 7899, 'loss/train': 1.1448894739151} 02/24/2022 11:47:49 - INFO - codeparrot_training - Step 7900: {'lr': 0.00048159109596877954, 'samples': 4045312, 'steps': 7900, 'loss/train': 2.8563191890716553} 02/24/2022 11:47:54 - INFO - codeparrot_training - Step 7901: {'lr': 0.00048158493290330443, 'samples': 4045824, 'steps': 7901, 'loss/train': 2.86163592338562} 02/24/2022 11:47:58 - INFO - codeparrot_training - Step 7902: {'lr': 0.00048157876884579294, 'samples': 4046336, 'steps': 7902, 'loss/train': 1.7894618511199951} 02/24/2022 11:48:03 - INFO - codeparrot_training - Step 7903: {'lr': 0.00048157260379627154, 'samples': 4046848, 'steps': 7903, 'loss/train': 1.9722492694854736} 02/24/2022 11:48:07 - INFO - codeparrot_training - Step 7904: {'lr': 0.0004815664377547667, 'samples': 4047360, 'steps': 7904, 'loss/train': 1.7103582620620728} 02/24/2022 11:48:12 - INFO - codeparrot_training - Step 7905: {'lr': 0.0004815602707213047, 'samples': 4047872, 'steps': 7905, 'loss/train': 1.7719565629959106} 02/24/2022 11:48:16 - INFO - codeparrot_training - Step 7906: {'lr': 0.00048155410269591203, 'samples': 4048384, 'steps': 7906, 'loss/train': 2.8015847206115723} 02/24/2022 11:48:21 - INFO - codeparrot_training - Step 7907: {'lr': 0.00048154793367861514, 'samples': 4048896, 'steps': 7907, 'loss/train': 1.719612717628479} 02/24/2022 11:48:25 - INFO - codeparrot_training - Step 7908: {'lr': 0.00048154176366944045, 'samples': 4049408, 'steps': 7908, 'loss/train': 3.0167155265808105} 02/24/2022 11:48:31 - INFO - codeparrot_training - Step 7909: {'lr': 0.0004815355926684144, 'samples': 4049920, 'steps': 7909, 'loss/train': 2.863046169281006} 02/24/2022 11:48:35 - INFO - codeparrot_training - Step 7910: {'lr': 0.0004815294206755633, 'samples': 4050432, 'steps': 7910, 'loss/train': 1.4275684356689453} 02/24/2022 11:48:40 - INFO - codeparrot_training - Step 7911: {'lr': 0.0004815232476909137, 'samples': 4050944, 'steps': 7911, 'loss/train': 2.777683973312378} 02/24/2022 11:48:43 - INFO - codeparrot_training - Step 7912: {'lr': 0.00048151707371449213, 'samples': 4051456, 'steps': 7912, 'loss/train': 2.2816970348358154} 02/24/2022 11:48:49 - INFO - codeparrot_training - Step 7913: {'lr': 0.0004815108987463248, 'samples': 4051968, 'steps': 7913, 'loss/train': 2.0520927906036377} 02/24/2022 11:48:53 - INFO - codeparrot_training - Step 7914: {'lr': 0.00048150472278643834, 'samples': 4052480, 'steps': 7914, 'loss/train': 2.3371365070343018} 02/24/2022 11:48:58 - INFO - codeparrot_training - Step 7915: {'lr': 0.0004814985458348592, 'samples': 4052992, 'steps': 7915, 'loss/train': 1.3985146284103394} 02/24/2022 11:49:02 - INFO - codeparrot_training - Step 7916: {'lr': 0.00048149236789161374, 'samples': 4053504, 'steps': 7916, 'loss/train': 2.6492886543273926} 02/24/2022 11:49:08 - INFO - codeparrot_training - Step 7917: {'lr': 0.00048148618895672846, 'samples': 4054016, 'steps': 7917, 'loss/train': 1.949591040611267} 02/24/2022 11:49:11 - INFO - codeparrot_training - Step 7918: {'lr': 0.0004814800090302299, 'samples': 4054528, 'steps': 7918, 'loss/train': 1.3352735042572021} 02/24/2022 11:49:17 - INFO - codeparrot_training - Step 7919: {'lr': 0.00048147382811214445, 'samples': 4055040, 'steps': 7919, 'loss/train': 2.5388824939727783} 02/24/2022 11:49:20 - INFO - codeparrot_training - Step 7920: {'lr': 0.0004814676462024987, 'samples': 4055552, 'steps': 7920, 'loss/train': 0.9629352688789368} 02/24/2022 11:49:26 - INFO - codeparrot_training - Step 7921: {'lr': 0.000481461463301319, 'samples': 4056064, 'steps': 7921, 'loss/train': 2.6115283966064453} 02/24/2022 11:49:29 - INFO - codeparrot_training - Step 7922: {'lr': 0.00048145527940863186, 'samples': 4056576, 'steps': 7922, 'loss/train': 1.4665732383728027} 02/24/2022 11:49:35 - INFO - codeparrot_training - Step 7923: {'lr': 0.00048144909452446384, 'samples': 4057088, 'steps': 7923, 'loss/train': 2.37501859664917} 02/24/2022 11:49:38 - INFO - codeparrot_training - Step 7924: {'lr': 0.00048144290864884145, 'samples': 4057600, 'steps': 7924, 'loss/train': 2.3301010131835938} 02/24/2022 11:49:44 - INFO - codeparrot_training - Step 7925: {'lr': 0.000481436721781791, 'samples': 4058112, 'steps': 7925, 'loss/train': 1.3217674493789673} 02/24/2022 11:49:48 - INFO - codeparrot_training - Step 7926: {'lr': 0.00048143053392333917, 'samples': 4058624, 'steps': 7926, 'loss/train': 2.278921127319336} 02/24/2022 11:49:53 - INFO - codeparrot_training - Step 7927: {'lr': 0.00048142434507351245, 'samples': 4059136, 'steps': 7927, 'loss/train': 2.2163450717926025} 02/24/2022 11:49:57 - INFO - codeparrot_training - Step 7928: {'lr': 0.00048141815523233735, 'samples': 4059648, 'steps': 7928, 'loss/train': 2.4262280464172363} 02/24/2022 11:50:03 - INFO - codeparrot_training - Step 7929: {'lr': 0.00048141196439984026, 'samples': 4060160, 'steps': 7929, 'loss/train': 3.123971462249756} 02/24/2022 11:50:06 - INFO - codeparrot_training - Step 7930: {'lr': 0.0004814057725760479, 'samples': 4060672, 'steps': 7930, 'loss/train': 2.2646522521972656} 02/24/2022 11:50:12 - INFO - codeparrot_training - Step 7931: {'lr': 0.0004813995797609866, 'samples': 4061184, 'steps': 7931, 'loss/train': 3.0059664249420166} 02/24/2022 11:50:15 - INFO - codeparrot_training - Step 7932: {'lr': 0.000481393385954683, 'samples': 4061696, 'steps': 7932, 'loss/train': 0.17604969441890717} 02/24/2022 11:50:21 - INFO - codeparrot_training - Step 7933: {'lr': 0.00048138719115716367, 'samples': 4062208, 'steps': 7933, 'loss/train': 2.906609058380127} 02/24/2022 11:50:24 - INFO - codeparrot_training - Step 7934: {'lr': 0.00048138099536845503, 'samples': 4062720, 'steps': 7934, 'loss/train': 2.1302058696746826} 02/24/2022 11:50:30 - INFO - codeparrot_training - Step 7935: {'lr': 0.0004813747985885837, 'samples': 4063232, 'steps': 7935, 'loss/train': 2.5546796321868896} 02/24/2022 11:50:33 - INFO - codeparrot_training - Step 7936: {'lr': 0.00048136860081757617, 'samples': 4063744, 'steps': 7936, 'loss/train': 2.7443885803222656} 02/24/2022 11:50:39 - INFO - codeparrot_training - Step 7937: {'lr': 0.00048136240205545907, 'samples': 4064256, 'steps': 7937, 'loss/train': 2.2444028854370117} 02/24/2022 11:50:43 - INFO - codeparrot_training - Step 7938: {'lr': 0.0004813562023022588, 'samples': 4064768, 'steps': 7938, 'loss/train': 1.5303294658660889} 02/24/2022 11:50:48 - INFO - codeparrot_training - Step 7939: {'lr': 0.00048135000155800217, 'samples': 4065280, 'steps': 7939, 'loss/train': 2.5860888957977295} 02/24/2022 11:50:52 - INFO - codeparrot_training - Step 7940: {'lr': 0.0004813437998227155, 'samples': 4065792, 'steps': 7940, 'loss/train': 2.6325581073760986} 02/24/2022 11:50:57 - INFO - codeparrot_training - Step 7941: {'lr': 0.00048133759709642556, 'samples': 4066304, 'steps': 7941, 'loss/train': 2.1561849117279053} 02/24/2022 11:51:01 - INFO - codeparrot_training - Step 7942: {'lr': 0.00048133139337915866, 'samples': 4066816, 'steps': 7942, 'loss/train': 1.200860619544983} 02/24/2022 11:51:06 - INFO - codeparrot_training - Step 7943: {'lr': 0.00048132518867094167, 'samples': 4067328, 'steps': 7943, 'loss/train': 2.813310146331787} 02/24/2022 11:51:10 - INFO - codeparrot_training - Step 7944: {'lr': 0.00048131898297180085, 'samples': 4067840, 'steps': 7944, 'loss/train': 1.2833945751190186} 02/24/2022 11:51:16 - INFO - codeparrot_training - Step 7945: {'lr': 0.0004813127762817631, 'samples': 4068352, 'steps': 7945, 'loss/train': 2.419658899307251} 02/24/2022 11:51:20 - INFO - codeparrot_training - Step 7946: {'lr': 0.00048130656860085485, 'samples': 4068864, 'steps': 7946, 'loss/train': 1.3830996751785278} 02/24/2022 11:51:25 - INFO - codeparrot_training - Step 7947: {'lr': 0.0004813003599291027, 'samples': 4069376, 'steps': 7947, 'loss/train': 2.92842435836792} 02/24/2022 11:51:29 - INFO - codeparrot_training - Step 7948: {'lr': 0.0004812941502665332, 'samples': 4069888, 'steps': 7948, 'loss/train': 2.1732382774353027} 02/24/2022 11:51:34 - INFO - codeparrot_training - Step 7949: {'lr': 0.0004812879396131731, 'samples': 4070400, 'steps': 7949, 'loss/train': 1.84319269657135} 02/24/2022 11:51:38 - INFO - codeparrot_training - Step 7950: {'lr': 0.0004812817279690488, 'samples': 4070912, 'steps': 7950, 'loss/train': 2.645608425140381} 02/24/2022 11:51:43 - INFO - codeparrot_training - Step 7951: {'lr': 0.00048127551533418714, 'samples': 4071424, 'steps': 7951, 'loss/train': 2.622865676879883} 02/24/2022 11:51:47 - INFO - codeparrot_training - Step 7952: {'lr': 0.0004812693017086145, 'samples': 4071936, 'steps': 7952, 'loss/train': 2.6795732975006104} 02/24/2022 11:51:53 - INFO - codeparrot_training - Step 7953: {'lr': 0.0004812630870923577, 'samples': 4072448, 'steps': 7953, 'loss/train': 2.2243003845214844} 02/24/2022 11:51:56 - INFO - codeparrot_training - Step 7954: {'lr': 0.00048125687148544316, 'samples': 4072960, 'steps': 7954, 'loss/train': 2.7206146717071533} 02/24/2022 11:52:00 - INFO - codeparrot_training - Step 7955: {'lr': 0.0004812506548878977, 'samples': 4073472, 'steps': 7955, 'loss/train': 0.573676347732544} 02/24/2022 11:52:06 - INFO - codeparrot_training - Step 7956: {'lr': 0.0004812444372997479, 'samples': 4073984, 'steps': 7956, 'loss/train': 1.9819824695587158} 02/24/2022 11:52:09 - INFO - codeparrot_training - Step 7957: {'lr': 0.00048123821872102023, 'samples': 4074496, 'steps': 7957, 'loss/train': 1.3413965702056885} 02/24/2022 11:52:15 - INFO - codeparrot_training - Step 7958: {'lr': 0.00048123199915174153, 'samples': 4075008, 'steps': 7958, 'loss/train': 2.1608805656433105} 02/24/2022 11:52:21 - INFO - codeparrot_training - Step 7959: {'lr': 0.0004812257785919384, 'samples': 4075520, 'steps': 7959, 'loss/train': 1.45054292678833} 02/24/2022 11:52:24 - INFO - codeparrot_training - Step 7960: {'lr': 0.00048121955704163744, 'samples': 4076032, 'steps': 7960, 'loss/train': 1.766263484954834} 02/24/2022 11:52:30 - INFO - codeparrot_training - Step 7961: {'lr': 0.00048121333450086524, 'samples': 4076544, 'steps': 7961, 'loss/train': 0.1166217252612114} 02/24/2022 11:52:33 - INFO - codeparrot_training - Step 7962: {'lr': 0.00048120711096964866, 'samples': 4077056, 'steps': 7962, 'loss/train': 2.2934818267822266} 02/24/2022 11:52:39 - INFO - codeparrot_training - Step 7963: {'lr': 0.0004812008864480142, 'samples': 4077568, 'steps': 7963, 'loss/train': 3.1890931129455566} 02/24/2022 11:52:43 - INFO - codeparrot_training - Step 7964: {'lr': 0.0004811946609359885, 'samples': 4078080, 'steps': 7964, 'loss/train': 2.035816192626953} 02/24/2022 11:52:46 - INFO - codeparrot_training - Step 7965: {'lr': 0.00048118843443359827, 'samples': 4078592, 'steps': 7965, 'loss/train': 0.8988781571388245} 02/24/2022 11:52:52 - INFO - codeparrot_training - Step 7966: {'lr': 0.00048118220694087023, 'samples': 4079104, 'steps': 7966, 'loss/train': 2.727423906326294} 02/24/2022 11:52:56 - INFO - codeparrot_training - Step 7967: {'lr': 0.00048117597845783106, 'samples': 4079616, 'steps': 7967, 'loss/train': 2.156682014465332} 02/24/2022 11:53:01 - INFO - codeparrot_training - Step 7968: {'lr': 0.0004811697489845074, 'samples': 4080128, 'steps': 7968, 'loss/train': 1.8921197652816772} 02/24/2022 11:53:05 - INFO - codeparrot_training - Step 7969: {'lr': 0.0004811635185209259, 'samples': 4080640, 'steps': 7969, 'loss/train': 2.3413808345794678} 02/24/2022 11:53:10 - INFO - codeparrot_training - Step 7970: {'lr': 0.0004811572870671133, 'samples': 4081152, 'steps': 7970, 'loss/train': 2.5551137924194336} 02/24/2022 11:53:14 - INFO - codeparrot_training - Step 7971: {'lr': 0.0004811510546230963, 'samples': 4081664, 'steps': 7971, 'loss/train': 2.3741261959075928} 02/24/2022 11:53:19 - INFO - codeparrot_training - Step 7972: {'lr': 0.0004811448211889016, 'samples': 4082176, 'steps': 7972, 'loss/train': 2.4906249046325684} 02/24/2022 11:53:25 - INFO - codeparrot_training - Step 7973: {'lr': 0.0004811385867645558, 'samples': 4082688, 'steps': 7973, 'loss/train': 2.2428839206695557} 02/24/2022 11:53:28 - INFO - codeparrot_training - Step 7974: {'lr': 0.00048113235135008574, 'samples': 4083200, 'steps': 7974, 'loss/train': 2.148123025894165} 02/24/2022 11:53:34 - INFO - codeparrot_training - Step 7975: {'lr': 0.0004811261149455181, 'samples': 4083712, 'steps': 7975, 'loss/train': 2.495208501815796} 02/24/2022 11:53:38 - INFO - codeparrot_training - Step 7976: {'lr': 0.0004811198775508796, 'samples': 4084224, 'steps': 7976, 'loss/train': 2.1607701778411865} 02/24/2022 11:53:41 - INFO - codeparrot_training - Step 7977: {'lr': 0.0004811136391661969, 'samples': 4084736, 'steps': 7977, 'loss/train': 2.479353904724121} 02/24/2022 11:53:47 - INFO - codeparrot_training - Step 7978: {'lr': 0.0004811073997914967, 'samples': 4085248, 'steps': 7978, 'loss/train': 2.566638469696045} 02/24/2022 11:53:52 - INFO - codeparrot_training - Step 7979: {'lr': 0.00048110115942680585, 'samples': 4085760, 'steps': 7979, 'loss/train': 2.3170058727264404} 02/24/2022 11:53:56 - INFO - codeparrot_training - Step 7980: {'lr': 0.000481094918072151, 'samples': 4086272, 'steps': 7980, 'loss/train': 1.4209965467453003} 02/24/2022 11:54:01 - INFO - codeparrot_training - Step 7981: {'lr': 0.0004810886757275589, 'samples': 4086784, 'steps': 7981, 'loss/train': 2.3535349369049072} 02/24/2022 11:54:05 - INFO - codeparrot_training - Step 7982: {'lr': 0.0004810824323930563, 'samples': 4087296, 'steps': 7982, 'loss/train': 2.260396718978882} 02/24/2022 11:54:10 - INFO - codeparrot_training - Step 7983: {'lr': 0.00048107618806866994, 'samples': 4087808, 'steps': 7983, 'loss/train': 1.198960542678833} 02/24/2022 11:54:14 - INFO - codeparrot_training - Step 7984: {'lr': 0.0004810699427544265, 'samples': 4088320, 'steps': 7984, 'loss/train': 2.7593555450439453} 02/24/2022 11:54:19 - INFO - codeparrot_training - Step 7985: {'lr': 0.00048106369645035284, 'samples': 4088832, 'steps': 7985, 'loss/train': 1.948966383934021} 02/24/2022 11:54:23 - INFO - codeparrot_training - Step 7986: {'lr': 0.0004810574491564757, 'samples': 4089344, 'steps': 7986, 'loss/train': 3.0611252784729004} 02/24/2022 11:54:28 - INFO - codeparrot_training - Step 7987: {'lr': 0.0004810512008728218, 'samples': 4089856, 'steps': 7987, 'loss/train': 2.946183443069458} 02/24/2022 11:54:32 - INFO - codeparrot_training - Step 7988: {'lr': 0.00048104495159941794, 'samples': 4090368, 'steps': 7988, 'loss/train': 2.5046749114990234} 02/24/2022 11:54:37 - INFO - codeparrot_training - Step 7989: {'lr': 0.00048103870133629084, 'samples': 4090880, 'steps': 7989, 'loss/train': 1.8661895990371704} 02/24/2022 11:54:41 - INFO - codeparrot_training - Step 7990: {'lr': 0.00048103245008346735, 'samples': 4091392, 'steps': 7990, 'loss/train': 1.6969434022903442} 02/24/2022 11:54:47 - INFO - codeparrot_training - Step 7991: {'lr': 0.0004810261978409742, 'samples': 4091904, 'steps': 7991, 'loss/train': 2.6047844886779785} 02/24/2022 11:54:51 - INFO - codeparrot_training - Step 7992: {'lr': 0.00048101994460883815, 'samples': 4092416, 'steps': 7992, 'loss/train': 2.207082748413086} 02/24/2022 11:54:54 - INFO - codeparrot_training - Step 7993: {'lr': 0.00048101369038708596, 'samples': 4092928, 'steps': 7993, 'loss/train': 0.4118365943431854} 02/24/2022 11:55:00 - INFO - codeparrot_training - Step 7994: {'lr': 0.0004810074351757446, 'samples': 4093440, 'steps': 7994, 'loss/train': 2.19514536857605} 02/24/2022 11:55:03 - INFO - codeparrot_training - Step 7995: {'lr': 0.00048100117897484064, 'samples': 4093952, 'steps': 7995, 'loss/train': 2.972121238708496} 02/24/2022 11:55:09 - INFO - codeparrot_training - Step 7996: {'lr': 0.0004809949217844011, 'samples': 4094464, 'steps': 7996, 'loss/train': 2.363495349884033} 02/24/2022 11:55:12 - INFO - codeparrot_training - Step 7997: {'lr': 0.00048098866360445254, 'samples': 4094976, 'steps': 7997, 'loss/train': 1.8781957626342773} 02/24/2022 11:55:18 - INFO - codeparrot_training - Step 7998: {'lr': 0.00048098240443502195, 'samples': 4095488, 'steps': 7998, 'loss/train': 2.3064351081848145} 02/24/2022 11:55:21 - INFO - codeparrot_training - Step 7999: {'lr': 0.000480976144276136, 'samples': 4096000, 'steps': 7999, 'loss/train': 2.3386523723602295} 02/24/2022 11:55:21 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 11:55:38 - WARNING - huggingface_hub.repository - Several commits (8) will be pushed upstream. 02/24/2022 11:55:38 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 11:56:12 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 1deb887..869a374 floral-grass-11 -> floral-grass-11 02/24/2022 11:56:19 - INFO - codeparrot_training - Step 8000: {'lr': 0.0004809698831278217, 'samples': 4096512, 'steps': 8000, 'loss/train': 2.433603048324585} 02/24/2022 11:56:22 - INFO - codeparrot_training - Step 8001: {'lr': 0.0004809636209901057, 'samples': 4097024, 'steps': 8001, 'loss/train': 2.5256423950195312} 02/24/2022 11:56:28 - INFO - codeparrot_training - Step 8002: {'lr': 0.00048095735786301495, 'samples': 4097536, 'steps': 8002, 'loss/train': 2.812481641769409} 02/24/2022 11:56:32 - INFO - codeparrot_training - Step 8003: {'lr': 0.00048095109374657617, 'samples': 4098048, 'steps': 8003, 'loss/train': 1.7197555303573608} 02/24/2022 11:56:37 - INFO - codeparrot_training - Step 8004: {'lr': 0.00048094482864081625, 'samples': 4098560, 'steps': 8004, 'loss/train': 2.114607810974121} 02/24/2022 11:56:41 - INFO - codeparrot_training - Step 8005: {'lr': 0.00048093856254576196, 'samples': 4099072, 'steps': 8005, 'loss/train': 2.01350998878479} 02/24/2022 11:56:46 - INFO - codeparrot_training - Step 8006: {'lr': 0.0004809322954614403, 'samples': 4099584, 'steps': 8006, 'loss/train': 1.9032108783721924} 02/24/2022 11:56:50 - INFO - codeparrot_training - Step 8007: {'lr': 0.00048092602738787795, 'samples': 4100096, 'steps': 8007, 'loss/train': 2.137026071548462} 02/24/2022 11:56:55 - INFO - codeparrot_training - Step 8008: {'lr': 0.00048091975832510183, 'samples': 4100608, 'steps': 8008, 'loss/train': 2.5735089778900146} 02/24/2022 11:56:59 - INFO - codeparrot_training - Step 8009: {'lr': 0.00048091348827313885, 'samples': 4101120, 'steps': 8009, 'loss/train': 2.4302027225494385} 02/24/2022 11:57:05 - INFO - codeparrot_training - Step 8010: {'lr': 0.0004809072172320157, 'samples': 4101632, 'steps': 8010, 'loss/train': 1.8490601778030396} 02/24/2022 11:57:08 - INFO - codeparrot_training - Step 8011: {'lr': 0.0004809009452017594, 'samples': 4102144, 'steps': 8011, 'loss/train': 1.9899863004684448} 02/24/2022 11:57:14 - INFO - codeparrot_training - Step 8012: {'lr': 0.00048089467218239687, 'samples': 4102656, 'steps': 8012, 'loss/train': 1.6612939834594727} 02/24/2022 11:57:18 - INFO - codeparrot_training - Step 8013: {'lr': 0.0004808883981739548, 'samples': 4103168, 'steps': 8013, 'loss/train': 1.3895426988601685} 02/24/2022 11:57:23 - INFO - codeparrot_training - Step 8014: {'lr': 0.00048088212317646016, 'samples': 4103680, 'steps': 8014, 'loss/train': 1.6364991664886475} 02/24/2022 11:57:27 - INFO - codeparrot_training - Step 8015: {'lr': 0.00048087584718993975, 'samples': 4104192, 'steps': 8015, 'loss/train': 2.0440566539764404} 02/24/2022 11:57:32 - INFO - codeparrot_training - Step 8016: {'lr': 0.0004808695702144206, 'samples': 4104704, 'steps': 8016, 'loss/train': 2.1823010444641113} 02/24/2022 11:57:36 - INFO - codeparrot_training - Step 8017: {'lr': 0.0004808632922499295, 'samples': 4105216, 'steps': 8017, 'loss/train': 2.9521472454071045} 02/24/2022 11:57:41 - INFO - codeparrot_training - Step 8018: {'lr': 0.00048085701329649336, 'samples': 4105728, 'steps': 8018, 'loss/train': 1.6742266416549683} 02/24/2022 11:57:45 - INFO - codeparrot_training - Step 8019: {'lr': 0.0004808507333541391, 'samples': 4106240, 'steps': 8019, 'loss/train': 2.1094822883605957} 02/24/2022 11:57:50 - INFO - codeparrot_training - Step 8020: {'lr': 0.00048084445242289355, 'samples': 4106752, 'steps': 8020, 'loss/train': 2.3094542026519775} 02/24/2022 11:57:54 - INFO - codeparrot_training - Step 8021: {'lr': 0.0004808381705027837, 'samples': 4107264, 'steps': 8021, 'loss/train': 2.991875648498535} 02/24/2022 11:58:00 - INFO - codeparrot_training - Step 8022: {'lr': 0.00048083188759383646, 'samples': 4107776, 'steps': 8022, 'loss/train': 2.4072718620300293} 02/24/2022 11:58:03 - INFO - codeparrot_training - Step 8023: {'lr': 0.00048082560369607863, 'samples': 4108288, 'steps': 8023, 'loss/train': 1.6046478748321533} 02/24/2022 11:58:09 - INFO - codeparrot_training - Step 8024: {'lr': 0.0004808193188095372, 'samples': 4108800, 'steps': 8024, 'loss/train': 2.563946485519409} 02/24/2022 11:58:12 - INFO - codeparrot_training - Step 8025: {'lr': 0.00048081303293423923, 'samples': 4109312, 'steps': 8025, 'loss/train': 2.3632664680480957} 02/24/2022 11:58:18 - INFO - codeparrot_training - Step 8026: {'lr': 0.0004808067460702115, 'samples': 4109824, 'steps': 8026, 'loss/train': 2.011209487915039} 02/24/2022 11:58:21 - INFO - codeparrot_training - Step 8027: {'lr': 0.00048080045821748086, 'samples': 4110336, 'steps': 8027, 'loss/train': 3.8600800037384033} 02/24/2022 11:58:28 - INFO - codeparrot_training - Step 8028: {'lr': 0.00048079416937607436, 'samples': 4110848, 'steps': 8028, 'loss/train': 2.4174904823303223} 02/24/2022 11:58:32 - INFO - codeparrot_training - Step 8029: {'lr': 0.000480787879546019, 'samples': 4111360, 'steps': 8029, 'loss/train': 3.168501377105713} 02/24/2022 11:58:37 - INFO - codeparrot_training - Step 8030: {'lr': 0.00048078158872734157, 'samples': 4111872, 'steps': 8030, 'loss/train': 2.173494577407837} 02/24/2022 11:58:41 - INFO - codeparrot_training - Step 8031: {'lr': 0.0004807752969200691, 'samples': 4112384, 'steps': 8031, 'loss/train': 1.3554211854934692} 02/24/2022 11:58:46 - INFO - codeparrot_training - Step 8032: {'lr': 0.0004807690041242286, 'samples': 4112896, 'steps': 8032, 'loss/train': 2.6872105598449707} 02/24/2022 11:58:50 - INFO - codeparrot_training - Step 8033: {'lr': 0.00048076271033984687, 'samples': 4113408, 'steps': 8033, 'loss/train': 1.967897891998291} 02/24/2022 11:58:55 - INFO - codeparrot_training - Step 8034: {'lr': 0.00048075641556695107, 'samples': 4113920, 'steps': 8034, 'loss/train': 1.3190655708312988} 02/24/2022 11:58:59 - INFO - codeparrot_training - Step 8035: {'lr': 0.000480750119805568, 'samples': 4114432, 'steps': 8035, 'loss/train': 2.2726380825042725} 02/24/2022 11:59:04 - INFO - codeparrot_training - Step 8036: {'lr': 0.0004807438230557247, 'samples': 4114944, 'steps': 8036, 'loss/train': 1.9625569581985474} 02/24/2022 11:59:08 - INFO - codeparrot_training - Step 8037: {'lr': 0.00048073752531744814, 'samples': 4115456, 'steps': 8037, 'loss/train': 2.1068637371063232} 02/24/2022 11:59:14 - INFO - codeparrot_training - Step 8038: {'lr': 0.0004807312265907653, 'samples': 4115968, 'steps': 8038, 'loss/train': 1.6491332054138184} 02/24/2022 11:59:17 - INFO - codeparrot_training - Step 8039: {'lr': 0.0004807249268757031, 'samples': 4116480, 'steps': 8039, 'loss/train': 2.0935187339782715} 02/24/2022 11:59:23 - INFO - codeparrot_training - Step 8040: {'lr': 0.00048071862617228854, 'samples': 4116992, 'steps': 8040, 'loss/train': 2.45920729637146} 02/24/2022 11:59:26 - INFO - codeparrot_training - Step 8041: {'lr': 0.0004807123244805488, 'samples': 4117504, 'steps': 8041, 'loss/train': 3.753026247024536} 02/24/2022 11:59:32 - INFO - codeparrot_training - Step 8042: {'lr': 0.0004807060218005106, 'samples': 4118016, 'steps': 8042, 'loss/train': 2.635624647140503} 02/24/2022 11:59:36 - INFO - codeparrot_training - Step 8043: {'lr': 0.00048069971813220107, 'samples': 4118528, 'steps': 8043, 'loss/train': 2.515420913696289} 02/24/2022 11:59:41 - INFO - codeparrot_training - Step 8044: {'lr': 0.0004806934134756472, 'samples': 4119040, 'steps': 8044, 'loss/train': 4.101273536682129} 02/24/2022 11:59:44 - INFO - codeparrot_training - Step 8045: {'lr': 0.0004806871078308761, 'samples': 4119552, 'steps': 8045, 'loss/train': 4.552318096160889} 02/24/2022 11:59:50 - INFO - codeparrot_training - Step 8046: {'lr': 0.0004806808011979146, 'samples': 4120064, 'steps': 8046, 'loss/train': 2.0850627422332764} 02/24/2022 11:59:53 - INFO - codeparrot_training - Step 8047: {'lr': 0.00048067449357678984, 'samples': 4120576, 'steps': 8047, 'loss/train': 2.1984524726867676} 02/24/2022 12:00:00 - INFO - codeparrot_training - Step 8048: {'lr': 0.0004806681849675287, 'samples': 4121088, 'steps': 8048, 'loss/train': 2.4453721046447754} 02/24/2022 12:00:05 - INFO - codeparrot_training - Step 8049: {'lr': 0.00048066187537015837, 'samples': 4121600, 'steps': 8049, 'loss/train': 2.386759042739868} 02/24/2022 12:00:09 - INFO - codeparrot_training - Step 8050: {'lr': 0.00048065556478470584, 'samples': 4122112, 'steps': 8050, 'loss/train': 2.3366856575012207} 02/24/2022 12:00:14 - INFO - codeparrot_training - Step 8051: {'lr': 0.0004806492532111981, 'samples': 4122624, 'steps': 8051, 'loss/train': 2.038649320602417} 02/24/2022 12:00:18 - INFO - codeparrot_training - Step 8052: {'lr': 0.00048064294064966215, 'samples': 4123136, 'steps': 8052, 'loss/train': 2.676236629486084} 02/24/2022 12:00:23 - INFO - codeparrot_training - Step 8053: {'lr': 0.00048063662710012513, 'samples': 4123648, 'steps': 8053, 'loss/train': 2.9631597995758057} 02/24/2022 12:00:27 - INFO - codeparrot_training - Step 8054: {'lr': 0.000480630312562614, 'samples': 4124160, 'steps': 8054, 'loss/train': 2.490952730178833} 02/24/2022 12:00:32 - INFO - codeparrot_training - Step 8055: {'lr': 0.0004806239970371558, 'samples': 4124672, 'steps': 8055, 'loss/train': 1.6425117254257202} 02/24/2022 12:00:36 - INFO - codeparrot_training - Step 8056: {'lr': 0.0004806176805237777, 'samples': 4125184, 'steps': 8056, 'loss/train': 6.032262325286865} 02/24/2022 12:00:42 - INFO - codeparrot_training - Step 8057: {'lr': 0.0004806113630225066, 'samples': 4125696, 'steps': 8057, 'loss/train': 2.2534685134887695} 02/24/2022 12:00:45 - INFO - codeparrot_training - Step 8058: {'lr': 0.0004806050445333697, 'samples': 4126208, 'steps': 8058, 'loss/train': 1.3106892108917236} 02/24/2022 12:00:51 - INFO - codeparrot_training - Step 8059: {'lr': 0.00048059872505639415, 'samples': 4126720, 'steps': 8059, 'loss/train': 1.9426368474960327} 02/24/2022 12:00:54 - INFO - codeparrot_training - Step 8060: {'lr': 0.0004805924045916067, 'samples': 4127232, 'steps': 8060, 'loss/train': 1.356183648109436} 02/24/2022 12:01:00 - INFO - codeparrot_training - Step 8061: {'lr': 0.00048058608313903474, 'samples': 4127744, 'steps': 8061, 'loss/train': 2.119553565979004} 02/24/2022 12:01:03 - INFO - codeparrot_training - Step 8062: {'lr': 0.0004805797606987051, 'samples': 4128256, 'steps': 8062, 'loss/train': 2.763160467147827} 02/24/2022 12:01:09 - INFO - codeparrot_training - Step 8063: {'lr': 0.0004805734372706451, 'samples': 4128768, 'steps': 8063, 'loss/train': 1.8133933544158936} 02/24/2022 12:01:12 - INFO - codeparrot_training - Step 8064: {'lr': 0.0004805671128548816, 'samples': 4129280, 'steps': 8064, 'loss/train': 2.0286574363708496} 02/24/2022 12:01:18 - INFO - codeparrot_training - Step 8065: {'lr': 0.00048056078745144183, 'samples': 4129792, 'steps': 8065, 'loss/train': 3.311856985092163} 02/24/2022 12:01:21 - INFO - codeparrot_training - Step 8066: {'lr': 0.0004805544610603529, 'samples': 4130304, 'steps': 8066, 'loss/train': 3.5693204402923584} 02/24/2022 12:01:27 - INFO - codeparrot_training - Step 8067: {'lr': 0.00048054813368164184, 'samples': 4130816, 'steps': 8067, 'loss/train': 1.5602513551712036} 02/24/2022 12:01:30 - INFO - codeparrot_training - Step 8068: {'lr': 0.00048054180531533576, 'samples': 4131328, 'steps': 8068, 'loss/train': 3.062683582305908} 02/24/2022 12:01:36 - INFO - codeparrot_training - Step 8069: {'lr': 0.00048053547596146185, 'samples': 4131840, 'steps': 8069, 'loss/train': 1.6411796808242798} 02/24/2022 12:01:39 - INFO - codeparrot_training - Step 8070: {'lr': 0.0004805291456200471, 'samples': 4132352, 'steps': 8070, 'loss/train': 2.829799175262451} 02/24/2022 12:01:45 - INFO - codeparrot_training - Step 8071: {'lr': 0.0004805228142911188, 'samples': 4132864, 'steps': 8071, 'loss/train': 1.6796129941940308} 02/24/2022 12:01:48 - INFO - codeparrot_training - Step 8072: {'lr': 0.0004805164819747038, 'samples': 4133376, 'steps': 8072, 'loss/train': 2.3893637657165527} 02/24/2022 12:01:54 - INFO - codeparrot_training - Step 8073: {'lr': 0.0004805101486708295, 'samples': 4133888, 'steps': 8073, 'loss/train': 1.6499165296554565} 02/24/2022 12:01:58 - INFO - codeparrot_training - Step 8074: {'lr': 0.0004805038143795229, 'samples': 4134400, 'steps': 8074, 'loss/train': 2.346736431121826} 02/24/2022 12:02:03 - INFO - codeparrot_training - Step 8075: {'lr': 0.00048049747910081114, 'samples': 4134912, 'steps': 8075, 'loss/train': 2.0924017429351807} 02/24/2022 12:02:07 - INFO - codeparrot_training - Step 8076: {'lr': 0.0004804911428347214, 'samples': 4135424, 'steps': 8076, 'loss/train': 2.370426654815674} 02/24/2022 12:02:13 - INFO - codeparrot_training - Step 8077: {'lr': 0.0004804848055812807, 'samples': 4135936, 'steps': 8077, 'loss/train': 2.170945644378662} 02/24/2022 12:02:17 - INFO - codeparrot_training - Step 8078: {'lr': 0.0004804784673405164, 'samples': 4136448, 'steps': 8078, 'loss/train': 2.602884531021118} 02/24/2022 12:02:22 - INFO - codeparrot_training - Step 8079: {'lr': 0.00048047212811245545, 'samples': 4136960, 'steps': 8079, 'loss/train': 1.7192648649215698} 02/24/2022 12:02:26 - INFO - codeparrot_training - Step 8080: {'lr': 0.00048046578789712516, 'samples': 4137472, 'steps': 8080, 'loss/train': 2.4500865936279297} 02/24/2022 12:02:31 - INFO - codeparrot_training - Step 8081: {'lr': 0.0004804594466945525, 'samples': 4137984, 'steps': 8081, 'loss/train': 2.430422782897949} 02/24/2022 12:02:35 - INFO - codeparrot_training - Step 8082: {'lr': 0.00048045310450476486, 'samples': 4138496, 'steps': 8082, 'loss/train': 3.008835792541504} 02/24/2022 12:02:40 - INFO - codeparrot_training - Step 8083: {'lr': 0.0004804467613277893, 'samples': 4139008, 'steps': 8083, 'loss/train': 1.8665213584899902} 02/24/2022 12:02:44 - INFO - codeparrot_training - Step 8084: {'lr': 0.00048044041716365296, 'samples': 4139520, 'steps': 8084, 'loss/train': 2.3280811309814453} 02/24/2022 12:02:49 - INFO - codeparrot_training - Step 8085: {'lr': 0.000480434072012383, 'samples': 4140032, 'steps': 8085, 'loss/train': 1.5482207536697388} 02/24/2022 12:02:53 - INFO - codeparrot_training - Step 8086: {'lr': 0.0004804277258740067, 'samples': 4140544, 'steps': 8086, 'loss/train': 2.0478014945983887} 02/24/2022 12:02:59 - INFO - codeparrot_training - Step 8087: {'lr': 0.0004804213787485512, 'samples': 4141056, 'steps': 8087, 'loss/train': 2.316554069519043} 02/24/2022 12:03:03 - INFO - codeparrot_training - Step 8088: {'lr': 0.00048041503063604366, 'samples': 4141568, 'steps': 8088, 'loss/train': 2.178321123123169} 02/24/2022 12:03:08 - INFO - codeparrot_training - Step 8089: {'lr': 0.00048040868153651124, 'samples': 4142080, 'steps': 8089, 'loss/train': 2.8111791610717773} 02/24/2022 12:03:12 - INFO - codeparrot_training - Step 8090: {'lr': 0.00048040233144998123, 'samples': 4142592, 'steps': 8090, 'loss/train': 2.189178228378296} 02/24/2022 12:03:17 - INFO - codeparrot_training - Step 8091: {'lr': 0.0004803959803764808, 'samples': 4143104, 'steps': 8091, 'loss/train': 2.443894624710083} 02/24/2022 12:03:21 - INFO - codeparrot_training - Step 8092: {'lr': 0.0004803896283160372, 'samples': 4143616, 'steps': 8092, 'loss/train': 2.2112085819244385} 02/24/2022 12:03:26 - INFO - codeparrot_training - Step 8093: {'lr': 0.0004803832752686775, 'samples': 4144128, 'steps': 8093, 'loss/train': 3.663377046585083} 02/24/2022 12:03:30 - INFO - codeparrot_training - Step 8094: {'lr': 0.00048037692123442904, 'samples': 4144640, 'steps': 8094, 'loss/train': 2.401536464691162} 02/24/2022 12:03:35 - INFO - codeparrot_training - Step 8095: {'lr': 0.000480370566213319, 'samples': 4145152, 'steps': 8095, 'loss/train': 1.8455917835235596} 02/24/2022 12:03:39 - INFO - codeparrot_training - Step 8096: {'lr': 0.00048036421020537464, 'samples': 4145664, 'steps': 8096, 'loss/train': 3.3369877338409424} 02/24/2022 12:03:45 - INFO - codeparrot_training - Step 8097: {'lr': 0.0004803578532106231, 'samples': 4146176, 'steps': 8097, 'loss/train': 2.3343923091888428} 02/24/2022 12:03:49 - INFO - codeparrot_training - Step 8098: {'lr': 0.00048035149522909174, 'samples': 4146688, 'steps': 8098, 'loss/train': 3.571747303009033} 02/24/2022 12:03:54 - INFO - codeparrot_training - Step 8099: {'lr': 0.0004803451362608076, 'samples': 4147200, 'steps': 8099, 'loss/train': 2.8301188945770264} 02/24/2022 12:03:58 - INFO - codeparrot_training - Step 8100: {'lr': 0.00048033877630579815, 'samples': 4147712, 'steps': 8100, 'loss/train': 1.7285693883895874} 02/24/2022 12:04:03 - INFO - codeparrot_training - Step 8101: {'lr': 0.00048033241536409043, 'samples': 4148224, 'steps': 8101, 'loss/train': 2.335404872894287} 02/24/2022 12:04:09 - INFO - codeparrot_training - Step 8102: {'lr': 0.0004803260534357119, 'samples': 4148736, 'steps': 8102, 'loss/train': 2.0846755504608154} 02/24/2022 12:04:12 - INFO - codeparrot_training - Step 8103: {'lr': 0.00048031969052068956, 'samples': 4149248, 'steps': 8103, 'loss/train': 2.8624277114868164} 02/24/2022 12:04:18 - INFO - codeparrot_training - Step 8104: {'lr': 0.00048031332661905093, 'samples': 4149760, 'steps': 8104, 'loss/train': 2.3115487098693848} 02/24/2022 12:04:21 - INFO - codeparrot_training - Step 8105: {'lr': 0.000480306961730823, 'samples': 4150272, 'steps': 8105, 'loss/train': 0.7021949887275696} 02/24/2022 12:04:27 - INFO - codeparrot_training - Step 8106: {'lr': 0.00048030059585603326, 'samples': 4150784, 'steps': 8106, 'loss/train': 1.5503710508346558} 02/24/2022 12:04:31 - INFO - codeparrot_training - Step 8107: {'lr': 0.0004802942289947089, 'samples': 4151296, 'steps': 8107, 'loss/train': 3.0905518531799316} 02/24/2022 12:04:36 - INFO - codeparrot_training - Step 8108: {'lr': 0.00048028786114687715, 'samples': 4151808, 'steps': 8108, 'loss/train': 1.925814151763916} 02/24/2022 12:04:40 - INFO - codeparrot_training - Step 8109: {'lr': 0.0004802814923125654, 'samples': 4152320, 'steps': 8109, 'loss/train': 1.0505138635635376} 02/24/2022 12:04:45 - INFO - codeparrot_training - Step 8110: {'lr': 0.00048027512249180083, 'samples': 4152832, 'steps': 8110, 'loss/train': 3.64475417137146} 02/24/2022 12:04:49 - INFO - codeparrot_training - Step 8111: {'lr': 0.0004802687516846107, 'samples': 4153344, 'steps': 8111, 'loss/train': 1.3404834270477295} 02/24/2022 12:04:54 - INFO - codeparrot_training - Step 8112: {'lr': 0.0004802623798910224, 'samples': 4153856, 'steps': 8112, 'loss/train': 1.6767250299453735} 02/24/2022 12:04:58 - INFO - codeparrot_training - Step 8113: {'lr': 0.00048025600711106323, 'samples': 4154368, 'steps': 8113, 'loss/train': 0.15420502424240112} 02/24/2022 12:05:04 - INFO - codeparrot_training - Step 8114: {'lr': 0.00048024963334476035, 'samples': 4154880, 'steps': 8114, 'loss/train': 2.6755011081695557} 02/24/2022 12:05:07 - INFO - codeparrot_training - Step 8115: {'lr': 0.00048024325859214123, 'samples': 4155392, 'steps': 8115, 'loss/train': 1.9760112762451172} 02/24/2022 12:05:13 - INFO - codeparrot_training - Step 8116: {'lr': 0.00048023688285323305, 'samples': 4155904, 'steps': 8116, 'loss/train': 1.9275007247924805} 02/24/2022 12:05:16 - INFO - codeparrot_training - Step 8117: {'lr': 0.0004802305061280632, 'samples': 4156416, 'steps': 8117, 'loss/train': 2.2548446655273438} 02/24/2022 12:05:22 - INFO - codeparrot_training - Step 8118: {'lr': 0.0004802241284166589, 'samples': 4156928, 'steps': 8118, 'loss/train': 2.323350191116333} 02/24/2022 12:05:25 - INFO - codeparrot_training - Step 8119: {'lr': 0.00048021774971904765, 'samples': 4157440, 'steps': 8119, 'loss/train': 2.828622817993164} 02/24/2022 12:05:31 - INFO - codeparrot_training - Step 8120: {'lr': 0.0004802113700352566, 'samples': 4157952, 'steps': 8120, 'loss/train': 0.19782328605651855} 02/24/2022 12:05:34 - INFO - codeparrot_training - Step 8121: {'lr': 0.0004802049893653131, 'samples': 4158464, 'steps': 8121, 'loss/train': 1.5305522680282593} 02/24/2022 12:05:40 - INFO - codeparrot_training - Step 8122: {'lr': 0.0004801986077092446, 'samples': 4158976, 'steps': 8122, 'loss/train': 0.35005131363868713} 02/24/2022 12:05:44 - INFO - codeparrot_training - Step 8123: {'lr': 0.0004801922250670783, 'samples': 4159488, 'steps': 8123, 'loss/train': 2.779991865158081} 02/24/2022 12:05:49 - INFO - codeparrot_training - Step 8124: {'lr': 0.0004801858414388416, 'samples': 4160000, 'steps': 8124, 'loss/train': 1.8902994394302368} 02/24/2022 12:05:53 - INFO - codeparrot_training - Step 8125: {'lr': 0.0004801794568245619, 'samples': 4160512, 'steps': 8125, 'loss/train': 3.0531387329101562} 02/24/2022 12:05:58 - INFO - codeparrot_training - Step 8126: {'lr': 0.00048017307122426653, 'samples': 4161024, 'steps': 8126, 'loss/train': 1.2154006958007812} 02/24/2022 12:06:02 - INFO - codeparrot_training - Step 8127: {'lr': 0.0004801666846379827, 'samples': 4161536, 'steps': 8127, 'loss/train': 2.110605239868164} 02/24/2022 12:06:07 - INFO - codeparrot_training - Step 8128: {'lr': 0.00048016029706573793, 'samples': 4162048, 'steps': 8128, 'loss/train': 2.361940383911133} 02/24/2022 12:06:11 - INFO - codeparrot_training - Step 8129: {'lr': 0.0004801539085075596, 'samples': 4162560, 'steps': 8129, 'loss/train': 2.6462466716766357} 02/24/2022 12:06:16 - INFO - codeparrot_training - Step 8130: {'lr': 0.0004801475189634749, 'samples': 4163072, 'steps': 8130, 'loss/train': 2.303628444671631} 02/24/2022 12:06:20 - INFO - codeparrot_training - Step 8131: {'lr': 0.0004801411284335114, 'samples': 4163584, 'steps': 8131, 'loss/train': 2.897541046142578} 02/24/2022 12:06:26 - INFO - codeparrot_training - Step 8132: {'lr': 0.0004801347369176963, 'samples': 4164096, 'steps': 8132, 'loss/train': 2.826315402984619} 02/24/2022 12:06:30 - INFO - codeparrot_training - Step 8133: {'lr': 0.0004801283444160571, 'samples': 4164608, 'steps': 8133, 'loss/train': 1.4086928367614746} 02/24/2022 12:06:35 - INFO - codeparrot_training - Step 8134: {'lr': 0.0004801219509286212, 'samples': 4165120, 'steps': 8134, 'loss/train': 1.6862421035766602} 02/24/2022 12:06:39 - INFO - codeparrot_training - Step 8135: {'lr': 0.00048011555645541585, 'samples': 4165632, 'steps': 8135, 'loss/train': 0.4918065369129181} 02/24/2022 12:06:44 - INFO - codeparrot_training - Step 8136: {'lr': 0.00048010916099646854, 'samples': 4166144, 'steps': 8136, 'loss/train': 2.506182909011841} 02/24/2022 12:06:47 - INFO - codeparrot_training - Step 8137: {'lr': 0.0004801027645518067, 'samples': 4166656, 'steps': 8137, 'loss/train': 2.2698466777801514} 02/24/2022 12:06:53 - INFO - codeparrot_training - Step 8138: {'lr': 0.00048009636712145764, 'samples': 4167168, 'steps': 8138, 'loss/train': 2.1795644760131836} 02/24/2022 12:06:57 - INFO - codeparrot_training - Step 8139: {'lr': 0.00048008996870544887, 'samples': 4167680, 'steps': 8139, 'loss/train': 1.3044347763061523} 02/24/2022 12:07:02 - INFO - codeparrot_training - Step 8140: {'lr': 0.0004800835693038076, 'samples': 4168192, 'steps': 8140, 'loss/train': 2.018707513809204} 02/24/2022 12:07:06 - INFO - codeparrot_training - Step 8141: {'lr': 0.0004800771689165615, 'samples': 4168704, 'steps': 8141, 'loss/train': 1.9235914945602417} 02/24/2022 12:07:11 - INFO - codeparrot_training - Step 8142: {'lr': 0.00048007076754373785, 'samples': 4169216, 'steps': 8142, 'loss/train': 1.0019667148590088} 02/24/2022 12:07:15 - INFO - codeparrot_training - Step 8143: {'lr': 0.00048006436518536403, 'samples': 4169728, 'steps': 8143, 'loss/train': 2.8660309314727783} 02/24/2022 12:07:22 - INFO - codeparrot_training - Step 8144: {'lr': 0.0004800579618414676, 'samples': 4170240, 'steps': 8144, 'loss/train': 1.065003752708435} 02/24/2022 12:07:25 - INFO - codeparrot_training - Step 8145: {'lr': 0.00048005155751207584, 'samples': 4170752, 'steps': 8145, 'loss/train': 0.17463712394237518} 02/24/2022 12:07:31 - INFO - codeparrot_training - Step 8146: {'lr': 0.0004800451521972163, 'samples': 4171264, 'steps': 8146, 'loss/train': 2.980131149291992} 02/24/2022 12:07:34 - INFO - codeparrot_training - Step 8147: {'lr': 0.0004800387458969164, 'samples': 4171776, 'steps': 8147, 'loss/train': 2.8523271083831787} 02/24/2022 12:07:40 - INFO - codeparrot_training - Step 8148: {'lr': 0.00048003233861120356, 'samples': 4172288, 'steps': 8148, 'loss/train': 2.7215511798858643} 02/24/2022 12:07:43 - INFO - codeparrot_training - Step 8149: {'lr': 0.00048002593034010516, 'samples': 4172800, 'steps': 8149, 'loss/train': 2.1218414306640625} 02/24/2022 12:07:49 - INFO - codeparrot_training - Step 8150: {'lr': 0.00048001952108364876, 'samples': 4173312, 'steps': 8150, 'loss/train': 2.644815683364868} 02/24/2022 12:07:52 - INFO - codeparrot_training - Step 8151: {'lr': 0.00048001311084186173, 'samples': 4173824, 'steps': 8151, 'loss/train': 2.0464587211608887} 02/24/2022 12:07:58 - INFO - codeparrot_training - Step 8152: {'lr': 0.0004800066996147716, 'samples': 4174336, 'steps': 8152, 'loss/train': 2.6863176822662354} 02/24/2022 12:08:05 - INFO - codeparrot_training - Step 8153: {'lr': 0.0004800002874024058, 'samples': 4174848, 'steps': 8153, 'loss/train': 2.6502692699432373} 02/24/2022 12:08:08 - INFO - codeparrot_training - Step 8154: {'lr': 0.0004799938742047918, 'samples': 4175360, 'steps': 8154, 'loss/train': 1.6076663732528687} 02/24/2022 12:08:14 - INFO - codeparrot_training - Step 8155: {'lr': 0.0004799874600219571, 'samples': 4175872, 'steps': 8155, 'loss/train': 2.502009391784668} 02/24/2022 12:08:17 - INFO - codeparrot_training - Step 8156: {'lr': 0.00047998104485392915, 'samples': 4176384, 'steps': 8156, 'loss/train': 2.1580681800842285} 02/24/2022 12:08:23 - INFO - codeparrot_training - Step 8157: {'lr': 0.0004799746287007354, 'samples': 4176896, 'steps': 8157, 'loss/train': 1.8411126136779785} 02/24/2022 12:08:26 - INFO - codeparrot_training - Step 8158: {'lr': 0.00047996821156240333, 'samples': 4177408, 'steps': 8158, 'loss/train': 2.3313233852386475} 02/24/2022 12:08:30 - INFO - codeparrot_training - Step 8159: {'lr': 0.0004799617934389605, 'samples': 4177920, 'steps': 8159, 'loss/train': 2.6959192752838135} 02/24/2022 12:08:35 - INFO - codeparrot_training - Step 8160: {'lr': 0.00047995537433043444, 'samples': 4178432, 'steps': 8160, 'loss/train': 1.8789721727371216} 02/24/2022 12:08:39 - INFO - codeparrot_training - Step 8161: {'lr': 0.00047994895423685246, 'samples': 4178944, 'steps': 8161, 'loss/train': 1.0363401174545288} 02/24/2022 12:08:44 - INFO - codeparrot_training - Step 8162: {'lr': 0.0004799425331582423, 'samples': 4179456, 'steps': 8162, 'loss/train': 1.6722670793533325} 02/24/2022 12:08:48 - INFO - codeparrot_training - Step 8163: {'lr': 0.00047993611109463125, 'samples': 4179968, 'steps': 8163, 'loss/train': 2.320652723312378} 02/24/2022 12:08:54 - INFO - codeparrot_training - Step 8164: {'lr': 0.00047992968804604693, 'samples': 4180480, 'steps': 8164, 'loss/train': 2.011254072189331} 02/24/2022 12:08:57 - INFO - codeparrot_training - Step 8165: {'lr': 0.00047992326401251686, 'samples': 4180992, 'steps': 8165, 'loss/train': 2.0376811027526855} 02/24/2022 12:09:03 - INFO - codeparrot_training - Step 8166: {'lr': 0.0004799168389940685, 'samples': 4181504, 'steps': 8166, 'loss/train': 2.3124780654907227} 02/24/2022 12:09:06 - INFO - codeparrot_training - Step 8167: {'lr': 0.00047991041299072946, 'samples': 4182016, 'steps': 8167, 'loss/train': 2.154853105545044} 02/24/2022 12:09:12 - INFO - codeparrot_training - Step 8168: {'lr': 0.00047990398600252713, 'samples': 4182528, 'steps': 8168, 'loss/train': 2.822540521621704} 02/24/2022 12:09:15 - INFO - codeparrot_training - Step 8169: {'lr': 0.0004798975580294892, 'samples': 4183040, 'steps': 8169, 'loss/train': 2.153313159942627} 02/24/2022 12:09:21 - INFO - codeparrot_training - Step 8170: {'lr': 0.0004798911290716431, 'samples': 4183552, 'steps': 8170, 'loss/train': 1.9812062978744507} 02/24/2022 12:09:24 - INFO - codeparrot_training - Step 8171: {'lr': 0.0004798846991290164, 'samples': 4184064, 'steps': 8171, 'loss/train': 1.9204516410827637} 02/24/2022 12:09:30 - INFO - codeparrot_training - Step 8172: {'lr': 0.0004798782682016367, 'samples': 4184576, 'steps': 8172, 'loss/train': 1.9890244007110596} 02/24/2022 12:09:33 - INFO - codeparrot_training - Step 8173: {'lr': 0.0004798718362895315, 'samples': 4185088, 'steps': 8173, 'loss/train': 2.2296829223632812} 02/24/2022 12:09:39 - INFO - codeparrot_training - Step 8174: {'lr': 0.0004798654033927283, 'samples': 4185600, 'steps': 8174, 'loss/train': 1.7498815059661865} 02/24/2022 12:09:42 - INFO - codeparrot_training - Step 8175: {'lr': 0.00047985896951125464, 'samples': 4186112, 'steps': 8175, 'loss/train': 2.340967893600464} 02/24/2022 12:09:48 - INFO - codeparrot_training - Step 8176: {'lr': 0.00047985253464513823, 'samples': 4186624, 'steps': 8176, 'loss/train': 1.6529486179351807} 02/24/2022 12:09:51 - INFO - codeparrot_training - Step 8177: {'lr': 0.00047984609879440655, 'samples': 4187136, 'steps': 8177, 'loss/train': 2.792949914932251} 02/24/2022 12:09:57 - INFO - codeparrot_training - Step 8178: {'lr': 0.0004798396619590871, 'samples': 4187648, 'steps': 8178, 'loss/train': 2.7088193893432617} 02/24/2022 12:10:00 - INFO - codeparrot_training - Step 8179: {'lr': 0.0004798332241392076, 'samples': 4188160, 'steps': 8179, 'loss/train': 1.6455278396606445} 02/24/2022 12:10:07 - INFO - codeparrot_training - Step 8180: {'lr': 0.0004798267853347955, 'samples': 4188672, 'steps': 8180, 'loss/train': 2.6440117359161377} 02/24/2022 12:10:10 - INFO - codeparrot_training - Step 8181: {'lr': 0.00047982034554587837, 'samples': 4189184, 'steps': 8181, 'loss/train': 2.4521286487579346} 02/24/2022 12:10:16 - INFO - codeparrot_training - Step 8182: {'lr': 0.000479813904772484, 'samples': 4189696, 'steps': 8182, 'loss/train': 3.2595694065093994} 02/24/2022 12:10:19 - INFO - codeparrot_training - Step 8183: {'lr': 0.0004798074630146397, 'samples': 4190208, 'steps': 8183, 'loss/train': 1.3921838998794556} 02/24/2022 12:10:25 - INFO - codeparrot_training - Step 8184: {'lr': 0.0004798010202723733, 'samples': 4190720, 'steps': 8184, 'loss/train': 1.9243848323822021} 02/24/2022 12:10:28 - INFO - codeparrot_training - Step 8185: {'lr': 0.00047979457654571223, 'samples': 4191232, 'steps': 8185, 'loss/train': 2.5567169189453125} 02/24/2022 12:10:34 - INFO - codeparrot_training - Step 8186: {'lr': 0.0004797881318346842, 'samples': 4191744, 'steps': 8186, 'loss/train': 1.6026275157928467} 02/24/2022 12:10:37 - INFO - codeparrot_training - Step 8187: {'lr': 0.00047978168613931684, 'samples': 4192256, 'steps': 8187, 'loss/train': 2.5572195053100586} 02/24/2022 12:10:43 - INFO - codeparrot_training - Step 8188: {'lr': 0.0004797752394596376, 'samples': 4192768, 'steps': 8188, 'loss/train': 2.0416271686553955} 02/24/2022 12:10:49 - INFO - codeparrot_training - Step 8189: {'lr': 0.0004797687917956742, 'samples': 4193280, 'steps': 8189, 'loss/train': 2.3751988410949707} 02/24/2022 12:10:52 - INFO - codeparrot_training - Step 8190: {'lr': 0.0004797623431474543, 'samples': 4193792, 'steps': 8190, 'loss/train': 2.909899950027466} 02/24/2022 12:10:58 - INFO - codeparrot_training - Step 8191: {'lr': 0.0004797558935150055, 'samples': 4194304, 'steps': 8191, 'loss/train': 2.1544458866119385} 02/24/2022 12:11:01 - INFO - codeparrot_training - Step 8192: {'lr': 0.0004797494428983553, 'samples': 4194816, 'steps': 8192, 'loss/train': 2.2693026065826416} 02/24/2022 12:11:07 - INFO - codeparrot_training - Step 8193: {'lr': 0.0004797429912975316, 'samples': 4195328, 'steps': 8193, 'loss/train': 2.261376142501831} 02/24/2022 12:11:10 - INFO - codeparrot_training - Step 8194: {'lr': 0.00047973653871256173, 'samples': 4195840, 'steps': 8194, 'loss/train': 3.045262575149536} 02/24/2022 12:11:16 - INFO - codeparrot_training - Step 8195: {'lr': 0.00047973008514347353, 'samples': 4196352, 'steps': 8195, 'loss/train': 1.2056047916412354} 02/24/2022 12:11:19 - INFO - codeparrot_training - Step 8196: {'lr': 0.00047972363059029465, 'samples': 4196864, 'steps': 8196, 'loss/train': 1.6056427955627441} 02/24/2022 12:11:25 - INFO - codeparrot_training - Step 8197: {'lr': 0.0004797171750530526, 'samples': 4197376, 'steps': 8197, 'loss/train': 1.9963260889053345} 02/24/2022 12:11:28 - INFO - codeparrot_training - Step 8198: {'lr': 0.00047971071853177515, 'samples': 4197888, 'steps': 8198, 'loss/train': 2.772096872329712} 02/24/2022 12:11:35 - INFO - codeparrot_training - Step 8199: {'lr': 0.0004797042610264899, 'samples': 4198400, 'steps': 8199, 'loss/train': 2.4937896728515625} 02/24/2022 12:11:38 - INFO - codeparrot_training - Step 8200: {'lr': 0.0004796978025372246, 'samples': 4198912, 'steps': 8200, 'loss/train': 2.0932905673980713} 02/24/2022 12:11:44 - INFO - codeparrot_training - Step 8201: {'lr': 0.0004796913430640068, 'samples': 4199424, 'steps': 8201, 'loss/train': 2.6117615699768066} 02/24/2022 12:11:47 - INFO - codeparrot_training - Step 8202: {'lr': 0.0004796848826068642, 'samples': 4199936, 'steps': 8202, 'loss/train': 2.5743870735168457} 02/24/2022 12:11:53 - INFO - codeparrot_training - Step 8203: {'lr': 0.00047967842116582453, 'samples': 4200448, 'steps': 8203, 'loss/train': 2.954094409942627} 02/24/2022 12:11:56 - INFO - codeparrot_training - Step 8204: {'lr': 0.00047967195874091547, 'samples': 4200960, 'steps': 8204, 'loss/train': 1.1489640474319458} 02/24/2022 12:12:02 - INFO - codeparrot_training - Step 8205: {'lr': 0.00047966549533216466, 'samples': 4201472, 'steps': 8205, 'loss/train': 2.8725554943084717} 02/24/2022 12:12:05 - INFO - codeparrot_training - Step 8206: {'lr': 0.00047965903093959974, 'samples': 4201984, 'steps': 8206, 'loss/train': 3.312897205352783} 02/24/2022 12:12:11 - INFO - codeparrot_training - Step 8207: {'lr': 0.0004796525655632484, 'samples': 4202496, 'steps': 8207, 'loss/train': 1.7352651357650757} 02/24/2022 12:12:14 - INFO - codeparrot_training - Step 8208: {'lr': 0.0004796460992031385, 'samples': 4203008, 'steps': 8208, 'loss/train': 1.9571641683578491} 02/24/2022 12:12:21 - INFO - codeparrot_training - Step 8209: {'lr': 0.0004796396318592976, 'samples': 4203520, 'steps': 8209, 'loss/train': 2.4581120014190674} 02/24/2022 12:12:24 - INFO - codeparrot_training - Step 8210: {'lr': 0.00047963316353175344, 'samples': 4204032, 'steps': 8210, 'loss/train': 1.0137072801589966} 02/24/2022 12:12:30 - INFO - codeparrot_training - Step 8211: {'lr': 0.00047962669422053374, 'samples': 4204544, 'steps': 8211, 'loss/train': 1.7872384786605835} 02/24/2022 12:12:33 - INFO - codeparrot_training - Step 8212: {'lr': 0.0004796202239256662, 'samples': 4205056, 'steps': 8212, 'loss/train': 2.5947911739349365} 02/24/2022 12:12:39 - INFO - codeparrot_training - Step 8213: {'lr': 0.0004796137526471785, 'samples': 4205568, 'steps': 8213, 'loss/train': 1.6550679206848145} 02/24/2022 12:12:42 - INFO - codeparrot_training - Step 8214: {'lr': 0.0004796072803850984, 'samples': 4206080, 'steps': 8214, 'loss/train': 1.2377721071243286} 02/24/2022 12:12:48 - INFO - codeparrot_training - Step 8215: {'lr': 0.00047960080713945364, 'samples': 4206592, 'steps': 8215, 'loss/train': 4.210224628448486} 02/24/2022 12:12:51 - INFO - codeparrot_training - Step 8216: {'lr': 0.0004795943329102719, 'samples': 4207104, 'steps': 8216, 'loss/train': 2.362583875656128} 02/24/2022 12:12:57 - INFO - codeparrot_training - Step 8217: {'lr': 0.00047958785769758094, 'samples': 4207616, 'steps': 8217, 'loss/train': 1.7497318983078003} 02/24/2022 12:13:00 - INFO - codeparrot_training - Step 8218: {'lr': 0.0004795813815014085, 'samples': 4208128, 'steps': 8218, 'loss/train': 2.7496001720428467} 02/24/2022 12:13:06 - INFO - codeparrot_training - Step 8219: {'lr': 0.0004795749043217824, 'samples': 4208640, 'steps': 8219, 'loss/train': 2.711496591567993} 02/24/2022 12:13:09 - INFO - codeparrot_training - Step 8220: {'lr': 0.0004795684261587302, 'samples': 4209152, 'steps': 8220, 'loss/train': 2.292638063430786} 02/24/2022 12:13:15 - INFO - codeparrot_training - Step 8221: {'lr': 0.00047956194701227983, 'samples': 4209664, 'steps': 8221, 'loss/train': 2.5446643829345703} 02/24/2022 12:13:18 - INFO - codeparrot_training - Step 8222: {'lr': 0.000479555466882459, 'samples': 4210176, 'steps': 8222, 'loss/train': 1.4770768880844116} 02/24/2022 12:13:24 - INFO - codeparrot_training - Step 8223: {'lr': 0.00047954898576929534, 'samples': 4210688, 'steps': 8223, 'loss/train': 1.0413466691970825} 02/24/2022 12:13:27 - INFO - codeparrot_training - Step 8224: {'lr': 0.0004795425036728168, 'samples': 4211200, 'steps': 8224, 'loss/train': 2.090332269668579} 02/24/2022 12:13:33 - INFO - codeparrot_training - Step 8225: {'lr': 0.000479536020593051, 'samples': 4211712, 'steps': 8225, 'loss/train': 1.7069073915481567} 02/24/2022 12:13:37 - INFO - codeparrot_training - Step 8226: {'lr': 0.0004795295365300258, 'samples': 4212224, 'steps': 8226, 'loss/train': 1.806524395942688} 02/24/2022 12:13:42 - INFO - codeparrot_training - Step 8227: {'lr': 0.00047952305148376895, 'samples': 4212736, 'steps': 8227, 'loss/train': 2.024029493331909} 02/24/2022 12:13:46 - INFO - codeparrot_training - Step 8228: {'lr': 0.0004795165654543082, 'samples': 4213248, 'steps': 8228, 'loss/train': 2.5008585453033447} 02/24/2022 12:13:51 - INFO - codeparrot_training - Step 8229: {'lr': 0.0004795100784416714, 'samples': 4213760, 'steps': 8229, 'loss/train': 2.52839732170105} 02/24/2022 12:13:55 - INFO - codeparrot_training - Step 8230: {'lr': 0.0004795035904458863, 'samples': 4214272, 'steps': 8230, 'loss/train': 1.7009122371673584} 02/24/2022 12:14:00 - INFO - codeparrot_training - Step 8231: {'lr': 0.00047949710146698066, 'samples': 4214784, 'steps': 8231, 'loss/train': 2.181901693344116} 02/24/2022 12:14:04 - INFO - codeparrot_training - Step 8232: {'lr': 0.0004794906115049824, 'samples': 4215296, 'steps': 8232, 'loss/train': 1.8688772916793823} 02/24/2022 12:14:09 - INFO - codeparrot_training - Step 8233: {'lr': 0.00047948412055991916, 'samples': 4215808, 'steps': 8233, 'loss/train': 2.3490421772003174} 02/24/2022 12:14:13 - INFO - codeparrot_training - Step 8234: {'lr': 0.0004794776286318188, 'samples': 4216320, 'steps': 8234, 'loss/train': 2.4164299964904785} 02/24/2022 12:14:19 - INFO - codeparrot_training - Step 8235: {'lr': 0.0004794711357207092, 'samples': 4216832, 'steps': 8235, 'loss/train': 1.2209607362747192} 02/24/2022 12:14:23 - INFO - codeparrot_training - Step 8236: {'lr': 0.0004794646418266181, 'samples': 4217344, 'steps': 8236, 'loss/train': 1.6379138231277466} 02/24/2022 12:14:28 - INFO - codeparrot_training - Step 8237: {'lr': 0.0004794581469495733, 'samples': 4217856, 'steps': 8237, 'loss/train': 2.0143113136291504} 02/24/2022 12:14:31 - INFO - codeparrot_training - Step 8238: {'lr': 0.00047945165108960274, 'samples': 4218368, 'steps': 8238, 'loss/train': 2.0794875621795654} 02/24/2022 12:14:37 - INFO - codeparrot_training - Step 8239: {'lr': 0.0004794451542467341, 'samples': 4218880, 'steps': 8239, 'loss/train': 2.4172987937927246} 02/24/2022 12:14:40 - INFO - codeparrot_training - Step 8240: {'lr': 0.00047943865642099525, 'samples': 4219392, 'steps': 8240, 'loss/train': 2.9183919429779053} 02/24/2022 12:14:46 - INFO - codeparrot_training - Step 8241: {'lr': 0.0004794321576124141, 'samples': 4219904, 'steps': 8241, 'loss/train': 2.501436471939087} 02/24/2022 12:14:50 - INFO - codeparrot_training - Step 8242: {'lr': 0.0004794256578210184, 'samples': 4220416, 'steps': 8242, 'loss/train': 2.6892893314361572} 02/24/2022 12:14:55 - INFO - codeparrot_training - Step 8243: {'lr': 0.0004794191570468361, 'samples': 4220928, 'steps': 8243, 'loss/train': 2.8740122318267822} 02/24/2022 12:14:59 - INFO - codeparrot_training - Step 8244: {'lr': 0.00047941265528989496, 'samples': 4221440, 'steps': 8244, 'loss/train': 1.7507708072662354} 02/24/2022 12:15:04 - INFO - codeparrot_training - Step 8245: {'lr': 0.0004794061525502229, 'samples': 4221952, 'steps': 8245, 'loss/train': 2.9183146953582764} 02/24/2022 12:15:08 - INFO - codeparrot_training - Step 8246: {'lr': 0.00047939964882784766, 'samples': 4222464, 'steps': 8246, 'loss/train': 3.189979076385498} 02/24/2022 12:15:14 - INFO - codeparrot_training - Step 8247: {'lr': 0.0004793931441227972, 'samples': 4222976, 'steps': 8247, 'loss/train': 1.870833158493042} 02/24/2022 12:15:17 - INFO - codeparrot_training - Step 8248: {'lr': 0.00047938663843509927, 'samples': 4223488, 'steps': 8248, 'loss/train': 1.9074362516403198} 02/24/2022 12:15:23 - INFO - codeparrot_training - Step 8249: {'lr': 0.00047938013176478193, 'samples': 4224000, 'steps': 8249, 'loss/train': 2.826785087585449} 02/24/2022 12:15:26 - INFO - codeparrot_training - Step 8250: {'lr': 0.0004793736241118728, 'samples': 4224512, 'steps': 8250, 'loss/train': 2.700700521469116} 02/24/2022 12:15:32 - INFO - codeparrot_training - Step 8251: {'lr': 0.0004793671154764, 'samples': 4225024, 'steps': 8251, 'loss/train': 3.159172773361206} 02/24/2022 12:15:35 - INFO - codeparrot_training - Step 8252: {'lr': 0.0004793606058583913, 'samples': 4225536, 'steps': 8252, 'loss/train': 2.0935311317443848} 02/24/2022 12:15:41 - INFO - codeparrot_training - Step 8253: {'lr': 0.0004793540952578746, 'samples': 4226048, 'steps': 8253, 'loss/train': 2.501469612121582} 02/24/2022 12:15:44 - INFO - codeparrot_training - Step 8254: {'lr': 0.0004793475836748777, 'samples': 4226560, 'steps': 8254, 'loss/train': 1.8863047361373901} 02/24/2022 12:15:51 - INFO - codeparrot_training - Step 8255: {'lr': 0.0004793410711094287, 'samples': 4227072, 'steps': 8255, 'loss/train': 2.8314342498779297} 02/24/2022 12:15:54 - INFO - codeparrot_training - Step 8256: {'lr': 0.00047933455756155534, 'samples': 4227584, 'steps': 8256, 'loss/train': 1.3806148767471313} 02/24/2022 12:16:00 - INFO - codeparrot_training - Step 8257: {'lr': 0.00047932804303128557, 'samples': 4228096, 'steps': 8257, 'loss/train': 2.3710362911224365} 02/24/2022 12:16:05 - INFO - codeparrot_training - Step 8258: {'lr': 0.0004793215275186472, 'samples': 4228608, 'steps': 8258, 'loss/train': 2.5845117568969727} 02/24/2022 12:16:09 - INFO - codeparrot_training - Step 8259: {'lr': 0.0004793150110236684, 'samples': 4229120, 'steps': 8259, 'loss/train': 2.2541561126708984} 02/24/2022 12:16:14 - INFO - codeparrot_training - Step 8260: {'lr': 0.00047930849354637674, 'samples': 4229632, 'steps': 8260, 'loss/train': 1.9688395261764526} 02/24/2022 12:16:18 - INFO - codeparrot_training - Step 8261: {'lr': 0.00047930197508680027, 'samples': 4230144, 'steps': 8261, 'loss/train': 2.1843063831329346} 02/24/2022 12:16:23 - INFO - codeparrot_training - Step 8262: {'lr': 0.00047929545564496715, 'samples': 4230656, 'steps': 8262, 'loss/train': 2.6420583724975586} 02/24/2022 12:16:27 - INFO - codeparrot_training - Step 8263: {'lr': 0.0004792889352209049, 'samples': 4231168, 'steps': 8263, 'loss/train': 1.9495543241500854} 02/24/2022 12:16:32 - INFO - codeparrot_training - Step 8264: {'lr': 0.00047928241381464177, 'samples': 4231680, 'steps': 8264, 'loss/train': 1.4433046579360962} 02/24/2022 12:16:36 - INFO - codeparrot_training - Step 8265: {'lr': 0.00047927589142620556, 'samples': 4232192, 'steps': 8265, 'loss/train': 0.4476317763328552} 02/24/2022 12:16:41 - INFO - codeparrot_training - Step 8266: {'lr': 0.0004792693680556243, 'samples': 4232704, 'steps': 8266, 'loss/train': 0.9182693362236023} 02/24/2022 12:16:45 - INFO - codeparrot_training - Step 8267: {'lr': 0.0004792628437029258, 'samples': 4233216, 'steps': 8267, 'loss/train': 2.5585031509399414} 02/24/2022 12:16:50 - INFO - codeparrot_training - Step 8268: {'lr': 0.0004792563183681381, 'samples': 4233728, 'steps': 8268, 'loss/train': 1.8075768947601318} 02/24/2022 12:16:54 - INFO - codeparrot_training - Step 8269: {'lr': 0.0004792497920512891, 'samples': 4234240, 'steps': 8269, 'loss/train': 2.9600419998168945} 02/24/2022 12:17:00 - INFO - codeparrot_training - Step 8270: {'lr': 0.00047924326475240676, 'samples': 4234752, 'steps': 8270, 'loss/train': 2.1123013496398926} 02/24/2022 12:17:03 - INFO - codeparrot_training - Step 8271: {'lr': 0.00047923673647151915, 'samples': 4235264, 'steps': 8271, 'loss/train': 2.3942224979400635} 02/24/2022 12:17:09 - INFO - codeparrot_training - Step 8272: {'lr': 0.00047923020720865413, 'samples': 4235776, 'steps': 8272, 'loss/train': 2.0425853729248047} 02/24/2022 12:17:12 - INFO - codeparrot_training - Step 8273: {'lr': 0.0004792236769638396, 'samples': 4236288, 'steps': 8273, 'loss/train': 2.497483015060425} 02/24/2022 12:17:18 - INFO - codeparrot_training - Step 8274: {'lr': 0.00047921714573710374, 'samples': 4236800, 'steps': 8274, 'loss/train': 2.4915049076080322} 02/24/2022 12:17:21 - INFO - codeparrot_training - Step 8275: {'lr': 0.0004792106135284744, 'samples': 4237312, 'steps': 8275, 'loss/train': 2.1010518074035645} 02/24/2022 12:17:27 - INFO - codeparrot_training - Step 8276: {'lr': 0.00047920408033797954, 'samples': 4237824, 'steps': 8276, 'loss/train': 2.2919366359710693} 02/24/2022 12:17:30 - INFO - codeparrot_training - Step 8277: {'lr': 0.00047919754616564716, 'samples': 4238336, 'steps': 8277, 'loss/train': 2.2083323001861572} 02/24/2022 12:17:36 - INFO - codeparrot_training - Step 8278: {'lr': 0.0004791910110115053, 'samples': 4238848, 'steps': 8278, 'loss/train': 1.7486019134521484} 02/24/2022 12:17:39 - INFO - codeparrot_training - Step 8279: {'lr': 0.0004791844748755819, 'samples': 4239360, 'steps': 8279, 'loss/train': 3.083209753036499} 02/24/2022 12:17:45 - INFO - codeparrot_training - Step 8280: {'lr': 0.00047917793775790503, 'samples': 4239872, 'steps': 8280, 'loss/train': 2.515814781188965} 02/24/2022 12:17:48 - INFO - codeparrot_training - Step 8281: {'lr': 0.00047917139965850266, 'samples': 4240384, 'steps': 8281, 'loss/train': 2.1462748050689697} 02/24/2022 12:17:55 - INFO - codeparrot_training - Step 8282: {'lr': 0.0004791648605774027, 'samples': 4240896, 'steps': 8282, 'loss/train': 2.547936201095581} 02/24/2022 12:17:58 - INFO - codeparrot_training - Step 8283: {'lr': 0.00047915832051463326, 'samples': 4241408, 'steps': 8283, 'loss/train': 1.8159955739974976} 02/24/2022 12:18:04 - INFO - codeparrot_training - Step 8284: {'lr': 0.0004791517794702224, 'samples': 4241920, 'steps': 8284, 'loss/train': 1.4766459465026855} 02/24/2022 12:18:07 - INFO - codeparrot_training - Step 8285: {'lr': 0.00047914523744419803, 'samples': 4242432, 'steps': 8285, 'loss/train': 2.598806858062744} 02/24/2022 12:18:13 - INFO - codeparrot_training - Step 8286: {'lr': 0.00047913869443658825, 'samples': 4242944, 'steps': 8286, 'loss/train': 2.0815885066986084} 02/24/2022 12:18:16 - INFO - codeparrot_training - Step 8287: {'lr': 0.0004791321504474211, 'samples': 4243456, 'steps': 8287, 'loss/train': 1.115213394165039} 02/24/2022 12:18:22 - INFO - codeparrot_training - Step 8288: {'lr': 0.00047912560547672453, 'samples': 4243968, 'steps': 8288, 'loss/train': 1.9663575887680054} 02/24/2022 12:18:25 - INFO - codeparrot_training - Step 8289: {'lr': 0.0004791190595245266, 'samples': 4244480, 'steps': 8289, 'loss/train': 5.791067123413086} 02/24/2022 12:18:31 - INFO - codeparrot_training - Step 8290: {'lr': 0.0004791125125908554, 'samples': 4244992, 'steps': 8290, 'loss/train': 3.670483350753784} 02/24/2022 12:18:34 - INFO - codeparrot_training - Step 8291: {'lr': 0.000479105964675739, 'samples': 4245504, 'steps': 8291, 'loss/train': 1.0139940977096558} 02/24/2022 12:18:40 - INFO - codeparrot_training - Step 8292: {'lr': 0.0004790994157792053, 'samples': 4246016, 'steps': 8292, 'loss/train': 2.2026891708374023} 02/24/2022 12:18:44 - INFO - codeparrot_training - Step 8293: {'lr': 0.0004790928659012825, 'samples': 4246528, 'steps': 8293, 'loss/train': 2.047849416732788} 02/24/2022 12:18:49 - INFO - codeparrot_training - Step 8294: {'lr': 0.00047908631504199855, 'samples': 4247040, 'steps': 8294, 'loss/train': 2.145226001739502} 02/24/2022 12:18:53 - INFO - codeparrot_training - Step 8295: {'lr': 0.00047907976320138163, 'samples': 4247552, 'steps': 8295, 'loss/train': 2.48504376411438} 02/24/2022 12:18:58 - INFO - codeparrot_training - Step 8296: {'lr': 0.00047907321037945973, 'samples': 4248064, 'steps': 8296, 'loss/train': 2.725163221359253} 02/24/2022 12:19:02 - INFO - codeparrot_training - Step 8297: {'lr': 0.0004790666565762609, 'samples': 4248576, 'steps': 8297, 'loss/train': 0.6805282831192017} 02/24/2022 12:19:07 - INFO - codeparrot_training - Step 8298: {'lr': 0.0004790601017918134, 'samples': 4249088, 'steps': 8298, 'loss/train': 4.326928615570068} 02/24/2022 12:19:11 - INFO - codeparrot_training - Step 8299: {'lr': 0.00047905354602614504, 'samples': 4249600, 'steps': 8299, 'loss/train': 1.4267929792404175} 02/24/2022 12:19:16 - INFO - codeparrot_training - Step 8300: {'lr': 0.00047904698927928404, 'samples': 4250112, 'steps': 8300, 'loss/train': 2.0606985092163086} 02/24/2022 12:19:20 - INFO - codeparrot_training - Step 8301: {'lr': 0.0004790404315512584, 'samples': 4250624, 'steps': 8301, 'loss/train': 2.01861572265625} 02/24/2022 12:19:26 - INFO - codeparrot_training - Step 8302: {'lr': 0.0004790338728420963, 'samples': 4251136, 'steps': 8302, 'loss/train': 4.525434970855713} 02/24/2022 12:19:29 - INFO - codeparrot_training - Step 8303: {'lr': 0.0004790273131518259, 'samples': 4251648, 'steps': 8303, 'loss/train': 3.037768602371216} 02/24/2022 12:19:35 - INFO - codeparrot_training - Step 8304: {'lr': 0.00047902075248047515, 'samples': 4252160, 'steps': 8304, 'loss/train': 2.4598560333251953} 02/24/2022 12:19:39 - INFO - codeparrot_training - Step 8305: {'lr': 0.0004790141908280723, 'samples': 4252672, 'steps': 8305, 'loss/train': 1.3979452848434448} 02/24/2022 12:19:44 - INFO - codeparrot_training - Step 8306: {'lr': 0.00047900762819464527, 'samples': 4253184, 'steps': 8306, 'loss/train': 2.760977029800415} 02/24/2022 12:19:48 - INFO - codeparrot_training - Step 8307: {'lr': 0.0004790010645802223, 'samples': 4253696, 'steps': 8307, 'loss/train': 0.6402405500411987} 02/24/2022 12:19:53 - INFO - codeparrot_training - Step 8308: {'lr': 0.0004789944999848316, 'samples': 4254208, 'steps': 8308, 'loss/train': 2.015840530395508} 02/24/2022 12:19:57 - INFO - codeparrot_training - Step 8309: {'lr': 0.00047898793440850104, 'samples': 4254720, 'steps': 8309, 'loss/train': 2.688425064086914} 02/24/2022 12:20:03 - INFO - codeparrot_training - Step 8310: {'lr': 0.0004789813678512589, 'samples': 4255232, 'steps': 8310, 'loss/train': 1.593688726425171} 02/24/2022 12:20:06 - INFO - codeparrot_training - Step 8311: {'lr': 0.0004789748003131333, 'samples': 4255744, 'steps': 8311, 'loss/train': 2.0959279537200928} 02/24/2022 12:20:12 - INFO - codeparrot_training - Step 8312: {'lr': 0.00047896823179415237, 'samples': 4256256, 'steps': 8312, 'loss/train': 2.116199254989624} 02/24/2022 12:20:15 - INFO - codeparrot_training - Step 8313: {'lr': 0.00047896166229434423, 'samples': 4256768, 'steps': 8313, 'loss/train': 1.8071153163909912} 02/24/2022 12:20:21 - INFO - codeparrot_training - Step 8314: {'lr': 0.0004789550918137371, 'samples': 4257280, 'steps': 8314, 'loss/train': 2.653411865234375} 02/24/2022 12:20:24 - INFO - codeparrot_training - Step 8315: {'lr': 0.000478948520352359, 'samples': 4257792, 'steps': 8315, 'loss/train': 2.8427231311798096} 02/24/2022 12:20:30 - INFO - codeparrot_training - Step 8316: {'lr': 0.00047894194791023813, 'samples': 4258304, 'steps': 8316, 'loss/train': 2.527801275253296} 02/24/2022 12:20:33 - INFO - codeparrot_training - Step 8317: {'lr': 0.0004789353744874027, 'samples': 4258816, 'steps': 8317, 'loss/train': 1.9348825216293335} 02/24/2022 12:20:39 - INFO - codeparrot_training - Step 8318: {'lr': 0.0004789288000838808, 'samples': 4259328, 'steps': 8318, 'loss/train': 1.7931863069534302} 02/24/2022 12:20:43 - INFO - codeparrot_training - Step 8319: {'lr': 0.0004789222246997006, 'samples': 4259840, 'steps': 8319, 'loss/train': 1.3822815418243408} 02/24/2022 12:20:48 - INFO - codeparrot_training - Step 8320: {'lr': 0.00047891564833489034, 'samples': 4260352, 'steps': 8320, 'loss/train': 3.32655930519104} 02/24/2022 12:20:52 - INFO - codeparrot_training - Step 8321: {'lr': 0.000478909070989478, 'samples': 4260864, 'steps': 8321, 'loss/train': 2.1469578742980957} 02/24/2022 12:20:57 - INFO - codeparrot_training - Step 8322: {'lr': 0.00047890249266349194, 'samples': 4261376, 'steps': 8322, 'loss/train': 1.6860512495040894} 02/24/2022 12:21:01 - INFO - codeparrot_training - Step 8323: {'lr': 0.0004788959133569604, 'samples': 4261888, 'steps': 8323, 'loss/train': 0.27361732721328735} 02/24/2022 12:21:06 - INFO - codeparrot_training - Step 8324: {'lr': 0.00047888933306991136, 'samples': 4262400, 'steps': 8324, 'loss/train': 2.4335134029388428} 02/24/2022 12:21:10 - INFO - codeparrot_training - Step 8325: {'lr': 0.00047888275180237304, 'samples': 4262912, 'steps': 8325, 'loss/train': 1.720828652381897} 02/24/2022 12:21:15 - INFO - codeparrot_training - Step 8326: {'lr': 0.00047887616955437373, 'samples': 4263424, 'steps': 8326, 'loss/train': 2.173304557800293} 02/24/2022 12:21:19 - INFO - codeparrot_training - Step 8327: {'lr': 0.0004788695863259416, 'samples': 4263936, 'steps': 8327, 'loss/train': 3.2694969177246094} 02/24/2022 12:21:25 - INFO - codeparrot_training - Step 8328: {'lr': 0.0004788630021171049, 'samples': 4264448, 'steps': 8328, 'loss/train': 1.9894022941589355} 02/24/2022 12:21:28 - INFO - codeparrot_training - Step 8329: {'lr': 0.0004788564169278917, 'samples': 4264960, 'steps': 8329, 'loss/train': 1.9709280729293823} 02/24/2022 12:21:34 - INFO - codeparrot_training - Step 8330: {'lr': 0.00047884983075833023, 'samples': 4265472, 'steps': 8330, 'loss/train': 1.9120969772338867} 02/24/2022 12:21:37 - INFO - codeparrot_training - Step 8331: {'lr': 0.00047884324360844885, 'samples': 4265984, 'steps': 8331, 'loss/train': 3.297394275665283} 02/24/2022 12:21:43 - INFO - codeparrot_training - Step 8332: {'lr': 0.0004788366554782756, 'samples': 4266496, 'steps': 8332, 'loss/train': 1.3300307989120483} 02/24/2022 12:21:48 - INFO - codeparrot_training - Step 8333: {'lr': 0.00047883006636783887, 'samples': 4267008, 'steps': 8333, 'loss/train': 0.5932084918022156} 02/24/2022 12:21:52 - INFO - codeparrot_training - Step 8334: {'lr': 0.0004788234762771667, 'samples': 4267520, 'steps': 8334, 'loss/train': 1.8165948390960693} 02/24/2022 12:21:57 - INFO - codeparrot_training - Step 8335: {'lr': 0.0004788168852062875, 'samples': 4268032, 'steps': 8335, 'loss/train': 2.3402628898620605} 02/24/2022 12:22:01 - INFO - codeparrot_training - Step 8336: {'lr': 0.0004788102931552294, 'samples': 4268544, 'steps': 8336, 'loss/train': 2.7658097743988037} 02/24/2022 12:22:07 - INFO - codeparrot_training - Step 8337: {'lr': 0.00047880370012402064, 'samples': 4269056, 'steps': 8337, 'loss/train': 3.3505196571350098} 02/24/2022 12:22:10 - INFO - codeparrot_training - Step 8338: {'lr': 0.0004787971061126895, 'samples': 4269568, 'steps': 8338, 'loss/train': 2.4177515506744385} 02/24/2022 12:22:16 - INFO - codeparrot_training - Step 8339: {'lr': 0.0004787905111212642, 'samples': 4270080, 'steps': 8339, 'loss/train': 0.9822052717208862} 02/24/2022 12:22:19 - INFO - codeparrot_training - Step 8340: {'lr': 0.00047878391514977306, 'samples': 4270592, 'steps': 8340, 'loss/train': 2.2899115085601807} 02/24/2022 12:22:25 - INFO - codeparrot_training - Step 8341: {'lr': 0.0004787773181982442, 'samples': 4271104, 'steps': 8341, 'loss/train': 0.2807779312133789} 02/24/2022 12:22:28 - INFO - codeparrot_training - Step 8342: {'lr': 0.0004787707202667059, 'samples': 4271616, 'steps': 8342, 'loss/train': 2.1107070446014404} 02/24/2022 12:22:34 - INFO - codeparrot_training - Step 8343: {'lr': 0.00047876412135518655, 'samples': 4272128, 'steps': 8343, 'loss/train': 0.5293779969215393} 02/24/2022 12:22:37 - INFO - codeparrot_training - Step 8344: {'lr': 0.0004787575214637144, 'samples': 4272640, 'steps': 8344, 'loss/train': 2.0584027767181396} 02/24/2022 12:22:43 - INFO - codeparrot_training - Step 8345: {'lr': 0.00047875092059231756, 'samples': 4273152, 'steps': 8345, 'loss/train': 2.5660791397094727} 02/24/2022 12:22:46 - INFO - codeparrot_training - Step 8346: {'lr': 0.0004787443187410245, 'samples': 4273664, 'steps': 8346, 'loss/train': 2.752887725830078} 02/24/2022 12:22:52 - INFO - codeparrot_training - Step 8347: {'lr': 0.00047873771590986337, 'samples': 4274176, 'steps': 8347, 'loss/train': 2.2198922634124756} 02/24/2022 12:22:56 - INFO - codeparrot_training - Step 8348: {'lr': 0.00047873111209886245, 'samples': 4274688, 'steps': 8348, 'loss/train': 2.9151782989501953} 02/24/2022 12:23:01 - INFO - codeparrot_training - Step 8349: {'lr': 0.00047872450730805015, 'samples': 4275200, 'steps': 8349, 'loss/train': 0.9172837138175964} 02/24/2022 12:23:05 - INFO - codeparrot_training - Step 8350: {'lr': 0.00047871790153745464, 'samples': 4275712, 'steps': 8350, 'loss/train': 3.037362575531006} 02/24/2022 12:23:11 - INFO - codeparrot_training - Step 8351: {'lr': 0.0004787112947871043, 'samples': 4276224, 'steps': 8351, 'loss/train': 2.521303176879883} 02/24/2022 12:23:14 - INFO - codeparrot_training - Step 8352: {'lr': 0.0004787046870570274, 'samples': 4276736, 'steps': 8352, 'loss/train': 2.4298343658447266} 02/24/2022 12:23:18 - INFO - codeparrot_training - Step 8353: {'lr': 0.00047869807834725225, 'samples': 4277248, 'steps': 8353, 'loss/train': 1.12665593624115} 02/24/2022 12:23:23 - INFO - codeparrot_training - Step 8354: {'lr': 0.0004786914686578071, 'samples': 4277760, 'steps': 8354, 'loss/train': 2.123337745666504} 02/24/2022 12:23:27 - INFO - codeparrot_training - Step 8355: {'lr': 0.00047868485798872044, 'samples': 4278272, 'steps': 8355, 'loss/train': 4.201889514923096} 02/24/2022 12:23:32 - INFO - codeparrot_training - Step 8356: {'lr': 0.00047867824634002034, 'samples': 4278784, 'steps': 8356, 'loss/train': 0.5770955681800842} 02/24/2022 12:23:36 - INFO - codeparrot_training - Step 8357: {'lr': 0.0004786716337117353, 'samples': 4279296, 'steps': 8357, 'loss/train': 2.2288622856140137} 02/24/2022 12:23:41 - INFO - codeparrot_training - Step 8358: {'lr': 0.00047866502010389356, 'samples': 4279808, 'steps': 8358, 'loss/train': 2.534905433654785} 02/24/2022 12:23:45 - INFO - codeparrot_training - Step 8359: {'lr': 0.00047865840551652343, 'samples': 4280320, 'steps': 8359, 'loss/train': 1.0368577241897583} 02/24/2022 12:23:50 - INFO - codeparrot_training - Step 8360: {'lr': 0.0004786517899496534, 'samples': 4280832, 'steps': 8360, 'loss/train': 1.8622485399246216} 02/24/2022 12:23:54 - INFO - codeparrot_training - Step 8361: {'lr': 0.0004786451734033117, 'samples': 4281344, 'steps': 8361, 'loss/train': 1.3333227634429932} 02/24/2022 12:23:59 - INFO - codeparrot_training - Step 8362: {'lr': 0.00047863855587752666, 'samples': 4281856, 'steps': 8362, 'loss/train': 1.9483284950256348} 02/24/2022 12:24:03 - INFO - codeparrot_training - Step 8363: {'lr': 0.0004786319373723266, 'samples': 4282368, 'steps': 8363, 'loss/train': 2.3748581409454346} 02/24/2022 12:24:09 - INFO - codeparrot_training - Step 8364: {'lr': 0.00047862531788774, 'samples': 4282880, 'steps': 8364, 'loss/train': 2.4677064418792725} 02/24/2022 12:24:12 - INFO - codeparrot_training - Step 8365: {'lr': 0.00047861869742379503, 'samples': 4283392, 'steps': 8365, 'loss/train': 0.3311143219470978} 02/24/2022 12:24:18 - INFO - codeparrot_training - Step 8366: {'lr': 0.0004786120759805203, 'samples': 4283904, 'steps': 8366, 'loss/train': 1.3763526678085327} 02/24/2022 12:24:21 - INFO - codeparrot_training - Step 8367: {'lr': 0.0004786054535579439, 'samples': 4284416, 'steps': 8367, 'loss/train': 2.129305601119995} 02/24/2022 12:24:27 - INFO - codeparrot_training - Step 8368: {'lr': 0.0004785988301560944, 'samples': 4284928, 'steps': 8368, 'loss/train': 2.2499935626983643} 02/24/2022 12:24:31 - INFO - codeparrot_training - Step 8369: {'lr': 0.0004785922057750001, 'samples': 4285440, 'steps': 8369, 'loss/train': 1.6785212755203247} 02/24/2022 12:24:36 - INFO - codeparrot_training - Step 8370: {'lr': 0.00047858558041468925, 'samples': 4285952, 'steps': 8370, 'loss/train': 2.668321371078491} 02/24/2022 12:24:39 - INFO - codeparrot_training - Step 8371: {'lr': 0.0004785789540751905, 'samples': 4286464, 'steps': 8371, 'loss/train': 2.5227622985839844} 02/24/2022 12:24:45 - INFO - codeparrot_training - Step 8372: {'lr': 0.00047857232675653207, 'samples': 4286976, 'steps': 8372, 'loss/train': 1.676182508468628} 02/24/2022 12:24:48 - INFO - codeparrot_training - Step 8373: {'lr': 0.0004785656984587423, 'samples': 4287488, 'steps': 8373, 'loss/train': 2.2817189693450928} 02/24/2022 12:24:55 - INFO - codeparrot_training - Step 8374: {'lr': 0.0004785590691818498, 'samples': 4288000, 'steps': 8374, 'loss/train': 1.941512107849121} 02/24/2022 12:24:58 - INFO - codeparrot_training - Step 8375: {'lr': 0.0004785524389258827, 'samples': 4288512, 'steps': 8375, 'loss/train': 1.960147500038147} 02/24/2022 12:25:04 - INFO - codeparrot_training - Step 8376: {'lr': 0.0004785458076908695, 'samples': 4289024, 'steps': 8376, 'loss/train': 2.76289963722229} 02/24/2022 12:25:07 - INFO - codeparrot_training - Step 8377: {'lr': 0.00047853917547683873, 'samples': 4289536, 'steps': 8377, 'loss/train': 3.043684959411621} 02/24/2022 12:25:13 - INFO - codeparrot_training - Step 8378: {'lr': 0.00047853254228381864, 'samples': 4290048, 'steps': 8378, 'loss/train': 2.4906437397003174} 02/24/2022 12:25:16 - INFO - codeparrot_training - Step 8379: {'lr': 0.0004785259081118377, 'samples': 4290560, 'steps': 8379, 'loss/train': 2.2070980072021484} 02/24/2022 12:25:22 - INFO - codeparrot_training - Step 8380: {'lr': 0.0004785192729609244, 'samples': 4291072, 'steps': 8380, 'loss/train': 1.390824794769287} 02/24/2022 12:25:25 - INFO - codeparrot_training - Step 8381: {'lr': 0.00047851263683110706, 'samples': 4291584, 'steps': 8381, 'loss/train': 3.5089597702026367} 02/24/2022 12:25:31 - INFO - codeparrot_training - Step 8382: {'lr': 0.0004785059997224142, 'samples': 4292096, 'steps': 8382, 'loss/train': 1.415643334388733} 02/24/2022 12:25:34 - INFO - codeparrot_training - Step 8383: {'lr': 0.0004784993616348741, 'samples': 4292608, 'steps': 8383, 'loss/train': 2.391145944595337} 02/24/2022 12:25:41 - INFO - codeparrot_training - Step 8384: {'lr': 0.0004784927225685153, 'samples': 4293120, 'steps': 8384, 'loss/train': 1.3958176374435425} 02/24/2022 12:25:44 - INFO - codeparrot_training - Step 8385: {'lr': 0.0004784860825233662, 'samples': 4293632, 'steps': 8385, 'loss/train': 2.319824457168579} 02/24/2022 12:25:50 - INFO - codeparrot_training - Step 8386: {'lr': 0.00047847944149945545, 'samples': 4294144, 'steps': 8386, 'loss/train': 2.8788483142852783} 02/24/2022 12:25:53 - INFO - codeparrot_training - Step 8387: {'lr': 0.00047847279949681117, 'samples': 4294656, 'steps': 8387, 'loss/train': 2.1804423332214355} 02/24/2022 12:25:59 - INFO - codeparrot_training - Step 8388: {'lr': 0.000478466156515462, 'samples': 4295168, 'steps': 8388, 'loss/train': 2.3936657905578613} 02/24/2022 12:26:03 - INFO - codeparrot_training - Step 8389: {'lr': 0.0004784595125554364, 'samples': 4295680, 'steps': 8389, 'loss/train': 2.9301764965057373} 02/24/2022 12:26:08 - INFO - codeparrot_training - Step 8390: {'lr': 0.00047845286761676276, 'samples': 4296192, 'steps': 8390, 'loss/train': 2.4787824153900146} 02/24/2022 12:26:12 - INFO - codeparrot_training - Step 8391: {'lr': 0.00047844622169946954, 'samples': 4296704, 'steps': 8391, 'loss/train': 3.358727216720581} 02/24/2022 12:26:17 - INFO - codeparrot_training - Step 8392: {'lr': 0.0004784395748035853, 'samples': 4297216, 'steps': 8392, 'loss/train': 2.1815829277038574} 02/24/2022 12:26:21 - INFO - codeparrot_training - Step 8393: {'lr': 0.0004784329269291384, 'samples': 4297728, 'steps': 8393, 'loss/train': 0.7497676014900208} 02/24/2022 12:26:26 - INFO - codeparrot_training - Step 8394: {'lr': 0.0004784262780761575, 'samples': 4298240, 'steps': 8394, 'loss/train': 3.2470383644104004} 02/24/2022 12:26:30 - INFO - codeparrot_training - Step 8395: {'lr': 0.00047841962824467086, 'samples': 4298752, 'steps': 8395, 'loss/train': 1.9744614362716675} 02/24/2022 12:26:35 - INFO - codeparrot_training - Step 8396: {'lr': 0.000478412977434707, 'samples': 4299264, 'steps': 8396, 'loss/train': 2.033781051635742} 02/24/2022 12:26:38 - INFO - codeparrot_training - Step 8397: {'lr': 0.0004784063256462946, 'samples': 4299776, 'steps': 8397, 'loss/train': 1.9653384685516357} 02/24/2022 12:26:44 - INFO - codeparrot_training - Step 8398: {'lr': 0.00047839967287946196, 'samples': 4300288, 'steps': 8398, 'loss/train': 0.29436564445495605} 02/24/2022 12:26:47 - INFO - codeparrot_training - Step 8399: {'lr': 0.00047839301913423773, 'samples': 4300800, 'steps': 8399, 'loss/train': 2.555769205093384} 02/24/2022 12:26:54 - INFO - codeparrot_training - Step 8400: {'lr': 0.0004783863644106502, 'samples': 4301312, 'steps': 8400, 'loss/train': 2.3032169342041016} 02/24/2022 12:26:57 - INFO - codeparrot_training - Step 8401: {'lr': 0.0004783797087087281, 'samples': 4301824, 'steps': 8401, 'loss/train': 2.5960450172424316} 02/24/2022 12:27:03 - INFO - codeparrot_training - Step 8402: {'lr': 0.00047837305202849987, 'samples': 4302336, 'steps': 8402, 'loss/train': 1.1665267944335938} 02/24/2022 12:27:06 - INFO - codeparrot_training - Step 8403: {'lr': 0.0004783663943699939, 'samples': 4302848, 'steps': 8403, 'loss/train': 0.24033935368061066} 02/24/2022 12:27:12 - INFO - codeparrot_training - Step 8404: {'lr': 0.00047835973573323885, 'samples': 4303360, 'steps': 8404, 'loss/train': 1.2128452062606812} 02/24/2022 12:27:15 - INFO - codeparrot_training - Step 8405: {'lr': 0.00047835307611826327, 'samples': 4303872, 'steps': 8405, 'loss/train': 1.773014783859253} 02/24/2022 12:27:21 - INFO - codeparrot_training - Step 8406: {'lr': 0.0004783464155250955, 'samples': 4304384, 'steps': 8406, 'loss/train': 2.131950855255127} 02/24/2022 12:27:24 - INFO - codeparrot_training - Step 8407: {'lr': 0.00047833975395376426, 'samples': 4304896, 'steps': 8407, 'loss/train': 2.3662731647491455} 02/24/2022 12:27:30 - INFO - codeparrot_training - Step 8408: {'lr': 0.00047833309140429803, 'samples': 4305408, 'steps': 8408, 'loss/train': 2.124105215072632} 02/24/2022 12:27:33 - INFO - codeparrot_training - Step 8409: {'lr': 0.00047832642787672537, 'samples': 4305920, 'steps': 8409, 'loss/train': 0.120108462870121} 02/24/2022 12:27:40 - INFO - codeparrot_training - Step 8410: {'lr': 0.00047831976337107474, 'samples': 4306432, 'steps': 8410, 'loss/train': 1.3338755369186401} 02/24/2022 12:27:43 - INFO - codeparrot_training - Step 8411: {'lr': 0.00047831309788737476, 'samples': 4306944, 'steps': 8411, 'loss/train': 3.362307071685791} 02/24/2022 12:27:49 - INFO - codeparrot_training - Step 8412: {'lr': 0.000478306431425654, 'samples': 4307456, 'steps': 8412, 'loss/train': 1.5072788000106812} 02/24/2022 12:27:52 - INFO - codeparrot_training - Step 8413: {'lr': 0.0004782997639859409, 'samples': 4307968, 'steps': 8413, 'loss/train': 2.6732122898101807} 02/24/2022 12:27:58 - INFO - codeparrot_training - Step 8414: {'lr': 0.00047829309556826415, 'samples': 4308480, 'steps': 8414, 'loss/train': 2.0217456817626953} 02/24/2022 12:28:01 - INFO - codeparrot_training - Step 8415: {'lr': 0.0004782864261726523, 'samples': 4308992, 'steps': 8415, 'loss/train': 1.818701148033142} 02/24/2022 12:28:07 - INFO - codeparrot_training - Step 8416: {'lr': 0.0004782797557991339, 'samples': 4309504, 'steps': 8416, 'loss/train': 1.957191824913025} 02/24/2022 12:28:12 - INFO - codeparrot_training - Step 8417: {'lr': 0.00047827308444773746, 'samples': 4310016, 'steps': 8417, 'loss/train': 1.8469033241271973} 02/24/2022 12:28:16 - INFO - codeparrot_training - Step 8418: {'lr': 0.00047826641211849165, 'samples': 4310528, 'steps': 8418, 'loss/train': 0.8525946736335754} 02/24/2022 12:28:22 - INFO - codeparrot_training - Step 8419: {'lr': 0.000478259738811425, 'samples': 4311040, 'steps': 8419, 'loss/train': 1.5026800632476807} 02/24/2022 12:28:25 - INFO - codeparrot_training - Step 8420: {'lr': 0.0004782530645265661, 'samples': 4311552, 'steps': 8420, 'loss/train': 2.6665549278259277} 02/24/2022 12:28:31 - INFO - codeparrot_training - Step 8421: {'lr': 0.00047824638926394355, 'samples': 4312064, 'steps': 8421, 'loss/train': 2.0895164012908936} 02/24/2022 12:28:34 - INFO - codeparrot_training - Step 8422: {'lr': 0.0004782397130235859, 'samples': 4312576, 'steps': 8422, 'loss/train': 2.006305694580078} 02/24/2022 12:28:40 - INFO - codeparrot_training - Step 8423: {'lr': 0.0004782330358055219, 'samples': 4313088, 'steps': 8423, 'loss/train': 2.4006519317626953} 02/24/2022 12:28:43 - INFO - codeparrot_training - Step 8424: {'lr': 0.00047822635760977995, 'samples': 4313600, 'steps': 8424, 'loss/train': 2.131471633911133} 02/24/2022 12:28:49 - INFO - codeparrot_training - Step 8425: {'lr': 0.0004782196784363888, 'samples': 4314112, 'steps': 8425, 'loss/train': 1.9225215911865234} 02/24/2022 12:28:52 - INFO - codeparrot_training - Step 8426: {'lr': 0.000478212998285377, 'samples': 4314624, 'steps': 8426, 'loss/train': 1.5440889596939087} 02/24/2022 12:28:58 - INFO - codeparrot_training - Step 8427: {'lr': 0.0004782063171567732, 'samples': 4315136, 'steps': 8427, 'loss/train': 2.0610897541046143} 02/24/2022 12:29:01 - INFO - codeparrot_training - Step 8428: {'lr': 0.000478199635050606, 'samples': 4315648, 'steps': 8428, 'loss/train': 2.043809175491333} 02/24/2022 12:29:07 - INFO - codeparrot_training - Step 8429: {'lr': 0.000478192951966904, 'samples': 4316160, 'steps': 8429, 'loss/train': 0.31734123826026917} 02/24/2022 12:29:11 - INFO - codeparrot_training - Step 8430: {'lr': 0.00047818626790569586, 'samples': 4316672, 'steps': 8430, 'loss/train': 2.2577764987945557} 02/24/2022 12:29:16 - INFO - codeparrot_training - Step 8431: {'lr': 0.00047817958286701026, 'samples': 4317184, 'steps': 8431, 'loss/train': 2.492427349090576} 02/24/2022 12:29:20 - INFO - codeparrot_training - Step 8432: {'lr': 0.00047817289685087575, 'samples': 4317696, 'steps': 8432, 'loss/train': 2.4741947650909424} 02/24/2022 12:29:25 - INFO - codeparrot_training - Step 8433: {'lr': 0.00047816620985732095, 'samples': 4318208, 'steps': 8433, 'loss/train': 1.7228083610534668} 02/24/2022 12:29:29 - INFO - codeparrot_training - Step 8434: {'lr': 0.0004781595218863746, 'samples': 4318720, 'steps': 8434, 'loss/train': 2.0321130752563477} 02/24/2022 12:29:34 - INFO - codeparrot_training - Step 8435: {'lr': 0.00047815283293806533, 'samples': 4319232, 'steps': 8435, 'loss/train': 2.9467849731445312} 02/24/2022 12:29:38 - INFO - codeparrot_training - Step 8436: {'lr': 0.0004781461430124217, 'samples': 4319744, 'steps': 8436, 'loss/train': 1.9760843515396118} 02/24/2022 12:29:43 - INFO - codeparrot_training - Step 8437: {'lr': 0.0004781394521094725, 'samples': 4320256, 'steps': 8437, 'loss/train': 2.0097668170928955} 02/24/2022 12:29:47 - INFO - codeparrot_training - Step 8438: {'lr': 0.00047813276022924634, 'samples': 4320768, 'steps': 8438, 'loss/train': 3.0590879917144775} 02/24/2022 12:29:52 - INFO - codeparrot_training - Step 8439: {'lr': 0.0004781260673717718, 'samples': 4321280, 'steps': 8439, 'loss/train': 1.5833929777145386} 02/24/2022 12:29:56 - INFO - codeparrot_training - Step 8440: {'lr': 0.0004781193735370777, 'samples': 4321792, 'steps': 8440, 'loss/train': 3.2533435821533203} 02/24/2022 12:30:01 - INFO - codeparrot_training - Step 8441: {'lr': 0.0004781126787251926, 'samples': 4322304, 'steps': 8441, 'loss/train': 2.16023588180542} 02/24/2022 12:30:05 - INFO - codeparrot_training - Step 8442: {'lr': 0.0004781059829361453, 'samples': 4322816, 'steps': 8442, 'loss/train': 2.2788326740264893} 02/24/2022 12:30:10 - INFO - codeparrot_training - Step 8443: {'lr': 0.00047809928616996425, 'samples': 4323328, 'steps': 8443, 'loss/train': 3.5258076190948486} 02/24/2022 12:30:14 - INFO - codeparrot_training - Step 8444: {'lr': 0.00047809258842667837, 'samples': 4323840, 'steps': 8444, 'loss/train': 2.1777842044830322} 02/24/2022 12:30:20 - INFO - codeparrot_training - Step 8445: {'lr': 0.00047808588970631627, 'samples': 4324352, 'steps': 8445, 'loss/train': 2.0628180503845215} 02/24/2022 12:30:23 - INFO - codeparrot_training - Step 8446: {'lr': 0.0004780791900089066, 'samples': 4324864, 'steps': 8446, 'loss/train': 1.7411916255950928} 02/24/2022 12:30:29 - INFO - codeparrot_training - Step 8447: {'lr': 0.0004780724893344782, 'samples': 4325376, 'steps': 8447, 'loss/train': 2.3368139266967773} 02/24/2022 12:30:32 - INFO - codeparrot_training - Step 8448: {'lr': 0.00047806578768305963, 'samples': 4325888, 'steps': 8448, 'loss/train': 1.7770050764083862} 02/24/2022 12:30:38 - INFO - codeparrot_training - Step 8449: {'lr': 0.00047805908505467963, 'samples': 4326400, 'steps': 8449, 'loss/train': 2.472324848175049} 02/24/2022 12:30:41 - INFO - codeparrot_training - Step 8450: {'lr': 0.0004780523814493669, 'samples': 4326912, 'steps': 8450, 'loss/train': 2.0360825061798096} 02/24/2022 12:30:47 - INFO - codeparrot_training - Step 8451: {'lr': 0.0004780456768671503, 'samples': 4327424, 'steps': 8451, 'loss/train': 1.8343180418014526} 02/24/2022 12:30:50 - INFO - codeparrot_training - Step 8452: {'lr': 0.0004780389713080583, 'samples': 4327936, 'steps': 8452, 'loss/train': 2.3955178260803223} 02/24/2022 12:30:56 - INFO - codeparrot_training - Step 8453: {'lr': 0.0004780322647721198, 'samples': 4328448, 'steps': 8453, 'loss/train': 1.004384994506836} 02/24/2022 12:30:59 - INFO - codeparrot_training - Step 8454: {'lr': 0.00047802555725936347, 'samples': 4328960, 'steps': 8454, 'loss/train': 2.443615674972534} 02/24/2022 12:31:05 - INFO - codeparrot_training - Step 8455: {'lr': 0.00047801884876981813, 'samples': 4329472, 'steps': 8455, 'loss/train': 1.5048587322235107} 02/24/2022 12:31:09 - INFO - codeparrot_training - Step 8456: {'lr': 0.0004780121393035124, 'samples': 4329984, 'steps': 8456, 'loss/train': 2.0674431324005127} 02/24/2022 12:31:14 - INFO - codeparrot_training - Step 8457: {'lr': 0.00047800542886047506, 'samples': 4330496, 'steps': 8457, 'loss/train': 2.230578660964966} 02/24/2022 12:31:18 - INFO - codeparrot_training - Step 8458: {'lr': 0.00047799871744073485, 'samples': 4331008, 'steps': 8458, 'loss/train': 2.067634105682373} 02/24/2022 12:31:23 - INFO - codeparrot_training - Step 8459: {'lr': 0.00047799200504432054, 'samples': 4331520, 'steps': 8459, 'loss/train': 2.618751287460327} 02/24/2022 12:31:27 - INFO - codeparrot_training - Step 8460: {'lr': 0.0004779852916712609, 'samples': 4332032, 'steps': 8460, 'loss/train': 2.3800833225250244} 02/24/2022 12:31:32 - INFO - codeparrot_training - Step 8461: {'lr': 0.0004779785773215847, 'samples': 4332544, 'steps': 8461, 'loss/train': 2.295905828475952} 02/24/2022 12:31:36 - INFO - codeparrot_training - Step 8462: {'lr': 0.00047797186199532055, 'samples': 4333056, 'steps': 8462, 'loss/train': 4.059459209442139} 02/24/2022 12:31:41 - INFO - codeparrot_training - Step 8463: {'lr': 0.0004779651456924974, 'samples': 4333568, 'steps': 8463, 'loss/train': 2.6968955993652344} 02/24/2022 12:31:45 - INFO - codeparrot_training - Step 8464: {'lr': 0.00047795842841314394, 'samples': 4334080, 'steps': 8464, 'loss/train': 1.1903438568115234} 02/24/2022 12:31:51 - INFO - codeparrot_training - Step 8465: {'lr': 0.000477951710157289, 'samples': 4334592, 'steps': 8465, 'loss/train': 1.9200540781021118} 02/24/2022 12:31:54 - INFO - codeparrot_training - Step 8466: {'lr': 0.00047794499092496123, 'samples': 4335104, 'steps': 8466, 'loss/train': 2.4851508140563965} 02/24/2022 12:32:00 - INFO - codeparrot_training - Step 8467: {'lr': 0.00047793827071618955, 'samples': 4335616, 'steps': 8467, 'loss/train': 1.9378186464309692} 02/24/2022 12:32:03 - INFO - codeparrot_training - Step 8468: {'lr': 0.0004779315495310027, 'samples': 4336128, 'steps': 8468, 'loss/train': 1.8286139965057373} 02/24/2022 12:32:09 - INFO - codeparrot_training - Step 8469: {'lr': 0.00047792482736942955, 'samples': 4336640, 'steps': 8469, 'loss/train': 1.626105546951294} 02/24/2022 12:32:12 - INFO - codeparrot_training - Step 8470: {'lr': 0.00047791810423149873, 'samples': 4337152, 'steps': 8470, 'loss/train': 1.3139162063598633} 02/24/2022 12:32:18 - INFO - codeparrot_training - Step 8471: {'lr': 0.0004779113801172391, 'samples': 4337664, 'steps': 8471, 'loss/train': 2.4346654415130615} 02/24/2022 12:32:23 - INFO - codeparrot_training - Step 8472: {'lr': 0.0004779046550266795, 'samples': 4338176, 'steps': 8472, 'loss/train': 2.1775312423706055} 02/24/2022 12:32:27 - INFO - codeparrot_training - Step 8473: {'lr': 0.00047789792895984874, 'samples': 4338688, 'steps': 8473, 'loss/train': 1.46390962600708} 02/24/2022 12:32:30 - INFO - codeparrot_training - Step 8474: {'lr': 0.0004778912019167756, 'samples': 4339200, 'steps': 8474, 'loss/train': 2.397479295730591} 02/24/2022 12:32:36 - INFO - codeparrot_training - Step 8475: {'lr': 0.00047788447389748894, 'samples': 4339712, 'steps': 8475, 'loss/train': 2.3953919410705566} 02/24/2022 12:32:39 - INFO - codeparrot_training - Step 8476: {'lr': 0.0004778777449020176, 'samples': 4340224, 'steps': 8476, 'loss/train': 2.907946825027466} 02/24/2022 12:32:45 - INFO - codeparrot_training - Step 8477: {'lr': 0.0004778710149303903, 'samples': 4340736, 'steps': 8477, 'loss/train': 2.48589825630188} 02/24/2022 12:32:51 - INFO - codeparrot_training - Step 8478: {'lr': 0.00047786428398263595, 'samples': 4341248, 'steps': 8478, 'loss/train': 1.9270997047424316} 02/24/2022 12:32:54 - INFO - codeparrot_training - Step 8479: {'lr': 0.00047785755205878333, 'samples': 4341760, 'steps': 8479, 'loss/train': 1.5082348585128784} 02/24/2022 12:33:00 - INFO - codeparrot_training - Step 8480: {'lr': 0.0004778508191588613, 'samples': 4342272, 'steps': 8480, 'loss/train': 3.1460177898406982} 02/24/2022 12:33:03 - INFO - codeparrot_training - Step 8481: {'lr': 0.0004778440852828988, 'samples': 4342784, 'steps': 8481, 'loss/train': 1.150168776512146} 02/24/2022 12:33:09 - INFO - codeparrot_training - Step 8482: {'lr': 0.00047783735043092446, 'samples': 4343296, 'steps': 8482, 'loss/train': 1.5935105085372925} 02/24/2022 12:33:12 - INFO - codeparrot_training - Step 8483: {'lr': 0.0004778306146029674, 'samples': 4343808, 'steps': 8483, 'loss/train': 2.04718279838562} 02/24/2022 12:33:18 - INFO - codeparrot_training - Step 8484: {'lr': 0.0004778238777990562, 'samples': 4344320, 'steps': 8484, 'loss/train': 1.802907943725586} 02/24/2022 12:33:21 - INFO - codeparrot_training - Step 8485: {'lr': 0.00047781714001921997, 'samples': 4344832, 'steps': 8485, 'loss/train': 0.8413384556770325} 02/24/2022 12:33:27 - INFO - codeparrot_training - Step 8486: {'lr': 0.00047781040126348734, 'samples': 4345344, 'steps': 8486, 'loss/train': 2.87712025642395} 02/24/2022 12:33:30 - INFO - codeparrot_training - Step 8487: {'lr': 0.0004778036615318874, 'samples': 4345856, 'steps': 8487, 'loss/train': 1.5497286319732666} 02/24/2022 12:33:36 - INFO - codeparrot_training - Step 8488: {'lr': 0.0004777969208244488, 'samples': 4346368, 'steps': 8488, 'loss/train': 0.3731866180896759} 02/24/2022 12:33:39 - INFO - codeparrot_training - Step 8489: {'lr': 0.0004777901791412006, 'samples': 4346880, 'steps': 8489, 'loss/train': 2.040407657623291} 02/24/2022 12:33:45 - INFO - codeparrot_training - Step 8490: {'lr': 0.00047778343648217155, 'samples': 4347392, 'steps': 8490, 'loss/train': 1.8158525228500366} 02/24/2022 12:33:49 - INFO - codeparrot_training - Step 8491: {'lr': 0.00047777669284739064, 'samples': 4347904, 'steps': 8491, 'loss/train': 1.5633140802383423} 02/24/2022 12:33:54 - INFO - codeparrot_training - Step 8492: {'lr': 0.0004777699482368867, 'samples': 4348416, 'steps': 8492, 'loss/train': 2.3152101039886475} 02/24/2022 12:33:58 - INFO - codeparrot_training - Step 8493: {'lr': 0.0004777632026506886, 'samples': 4348928, 'steps': 8493, 'loss/train': 0.2024056315422058} 02/24/2022 12:34:03 - INFO - codeparrot_training - Step 8494: {'lr': 0.0004777564560888252, 'samples': 4349440, 'steps': 8494, 'loss/train': 2.1427817344665527} 02/24/2022 12:34:07 - INFO - codeparrot_training - Step 8495: {'lr': 0.0004777497085513256, 'samples': 4349952, 'steps': 8495, 'loss/train': 2.6349411010742188} 02/24/2022 12:34:12 - INFO - codeparrot_training - Step 8496: {'lr': 0.0004777429600382185, 'samples': 4350464, 'steps': 8496, 'loss/train': 1.7786177396774292} 02/24/2022 12:34:16 - INFO - codeparrot_training - Step 8497: {'lr': 0.00047773621054953287, 'samples': 4350976, 'steps': 8497, 'loss/train': 1.8175199031829834} 02/24/2022 12:34:21 - INFO - codeparrot_training - Step 8498: {'lr': 0.0004777294600852976, 'samples': 4351488, 'steps': 8498, 'loss/train': 0.35944968461990356} 02/24/2022 12:34:25 - INFO - codeparrot_training - Step 8499: {'lr': 0.0004777227086455417, 'samples': 4352000, 'steps': 8499, 'loss/train': 2.4719696044921875} 02/24/2022 12:34:31 - INFO - codeparrot_training - Step 8500: {'lr': 0.000477715956230294, 'samples': 4352512, 'steps': 8500, 'loss/train': 1.956946611404419} 02/24/2022 12:34:35 - INFO - codeparrot_training - Step 8501: {'lr': 0.0004777092028395834, 'samples': 4353024, 'steps': 8501, 'loss/train': 0.12164922058582306} 02/24/2022 12:34:40 - INFO - codeparrot_training - Step 8502: {'lr': 0.00047770244847343893, 'samples': 4353536, 'steps': 8502, 'loss/train': 1.2616498470306396} 02/24/2022 12:34:44 - INFO - codeparrot_training - Step 8503: {'lr': 0.0004776956931318895, 'samples': 4354048, 'steps': 8503, 'loss/train': 0.6837210059165955} 02/24/2022 12:34:49 - INFO - codeparrot_training - Step 8504: {'lr': 0.00047768893681496397, 'samples': 4354560, 'steps': 8504, 'loss/train': 0.9616917967796326} 02/24/2022 12:34:53 - INFO - codeparrot_training - Step 8505: {'lr': 0.0004776821795226913, 'samples': 4355072, 'steps': 8505, 'loss/train': 2.351278066635132} 02/24/2022 12:34:58 - INFO - codeparrot_training - Step 8506: {'lr': 0.0004776754212551006, 'samples': 4355584, 'steps': 8506, 'loss/train': 1.5699753761291504} 02/24/2022 12:35:02 - INFO - codeparrot_training - Step 8507: {'lr': 0.0004776686620122206, 'samples': 4356096, 'steps': 8507, 'loss/train': 5.027220726013184} 02/24/2022 12:35:08 - INFO - codeparrot_training - Step 8508: {'lr': 0.00047766190179408043, 'samples': 4356608, 'steps': 8508, 'loss/train': 2.2917065620422363} 02/24/2022 12:35:11 - INFO - codeparrot_training - Step 8509: {'lr': 0.00047765514060070887, 'samples': 4357120, 'steps': 8509, 'loss/train': 1.9001367092132568} 02/24/2022 12:35:17 - INFO - codeparrot_training - Step 8510: {'lr': 0.00047764837843213497, 'samples': 4357632, 'steps': 8510, 'loss/train': 2.093930721282959} 02/24/2022 12:35:20 - INFO - codeparrot_training - Step 8511: {'lr': 0.0004776416152883878, 'samples': 4358144, 'steps': 8511, 'loss/train': 2.780133008956909} 02/24/2022 12:35:26 - INFO - codeparrot_training - Step 8512: {'lr': 0.0004776348511694961, 'samples': 4358656, 'steps': 8512, 'loss/train': 2.6753807067871094} 02/24/2022 12:35:30 - INFO - codeparrot_training - Step 8513: {'lr': 0.0004776280860754891, 'samples': 4359168, 'steps': 8513, 'loss/train': 2.1889731884002686} 02/24/2022 12:35:35 - INFO - codeparrot_training - Step 8514: {'lr': 0.0004776213200063956, 'samples': 4359680, 'steps': 8514, 'loss/train': 3.200205087661743} 02/24/2022 12:35:39 - INFO - codeparrot_training - Step 8515: {'lr': 0.00047761455296224464, 'samples': 4360192, 'steps': 8515, 'loss/train': 2.1677839756011963} 02/24/2022 12:35:44 - INFO - codeparrot_training - Step 8516: {'lr': 0.0004776077849430652, 'samples': 4360704, 'steps': 8516, 'loss/train': 2.529313564300537} 02/24/2022 12:35:48 - INFO - codeparrot_training - Step 8517: {'lr': 0.00047760101594888633, 'samples': 4361216, 'steps': 8517, 'loss/train': 2.6973793506622314} 02/24/2022 12:35:53 - INFO - codeparrot_training - Step 8518: {'lr': 0.000477594245979737, 'samples': 4361728, 'steps': 8518, 'loss/train': 2.445666790008545} 02/24/2022 12:35:56 - INFO - codeparrot_training - Step 8519: {'lr': 0.0004775874750356461, 'samples': 4362240, 'steps': 8519, 'loss/train': 2.794433832168579} 02/24/2022 12:36:02 - INFO - codeparrot_training - Step 8520: {'lr': 0.00047758070311664283, 'samples': 4362752, 'steps': 8520, 'loss/train': 1.6525999307632446} 02/24/2022 12:36:05 - INFO - codeparrot_training - Step 8521: {'lr': 0.000477573930222756, 'samples': 4363264, 'steps': 8521, 'loss/train': 1.8866370916366577} 02/24/2022 12:36:12 - INFO - codeparrot_training - Step 8522: {'lr': 0.0004775671563540147, 'samples': 4363776, 'steps': 8522, 'loss/train': 2.0719964504241943} 02/24/2022 12:36:15 - INFO - codeparrot_training - Step 8523: {'lr': 0.000477560381510448, 'samples': 4364288, 'steps': 8523, 'loss/train': 2.1382999420166016} 02/24/2022 12:36:21 - INFO - codeparrot_training - Step 8524: {'lr': 0.00047755360569208495, 'samples': 4364800, 'steps': 8524, 'loss/train': 2.7998528480529785} 02/24/2022 12:36:24 - INFO - codeparrot_training - Step 8525: {'lr': 0.00047754682889895444, 'samples': 4365312, 'steps': 8525, 'loss/train': 1.8198583126068115} 02/24/2022 12:36:30 - INFO - codeparrot_training - Step 8526: {'lr': 0.00047754005113108557, 'samples': 4365824, 'steps': 8526, 'loss/train': 2.023512840270996} 02/24/2022 12:36:33 - INFO - codeparrot_training - Step 8527: {'lr': 0.0004775332723885074, 'samples': 4366336, 'steps': 8527, 'loss/train': 1.8731328248977661} 02/24/2022 12:36:39 - INFO - codeparrot_training - Step 8528: {'lr': 0.00047752649267124894, 'samples': 4366848, 'steps': 8528, 'loss/train': 1.8908895254135132} 02/24/2022 12:36:42 - INFO - codeparrot_training - Step 8529: {'lr': 0.0004775197119793392, 'samples': 4367360, 'steps': 8529, 'loss/train': 1.4178643226623535} 02/24/2022 12:36:48 - INFO - codeparrot_training - Step 8530: {'lr': 0.0004775129303128073, 'samples': 4367872, 'steps': 8530, 'loss/train': 2.051976203918457} 02/24/2022 12:36:51 - INFO - codeparrot_training - Step 8531: {'lr': 0.0004775061476716822, 'samples': 4368384, 'steps': 8531, 'loss/train': 2.5321860313415527} 02/24/2022 12:36:56 - INFO - codeparrot_training - Step 8532: {'lr': 0.000477499364055993, 'samples': 4368896, 'steps': 8532, 'loss/train': 0.30298274755477905} 02/24/2022 12:37:02 - INFO - codeparrot_training - Step 8533: {'lr': 0.00047749257946576887, 'samples': 4369408, 'steps': 8533, 'loss/train': 1.7422784566879272} 02/24/2022 12:37:05 - INFO - codeparrot_training - Step 8534: {'lr': 0.0004774857939010387, 'samples': 4369920, 'steps': 8534, 'loss/train': 2.6280179023742676} 02/24/2022 12:37:11 - INFO - codeparrot_training - Step 8535: {'lr': 0.0004774790073618316, 'samples': 4370432, 'steps': 8535, 'loss/train': 1.9418681859970093} 02/24/2022 12:37:14 - INFO - codeparrot_training - Step 8536: {'lr': 0.00047747221984817666, 'samples': 4370944, 'steps': 8536, 'loss/train': 0.8090088963508606} 02/24/2022 12:37:21 - INFO - codeparrot_training - Step 8537: {'lr': 0.000477465431360103, 'samples': 4371456, 'steps': 8537, 'loss/train': 1.2902745008468628} 02/24/2022 12:37:24 - INFO - codeparrot_training - Step 8538: {'lr': 0.00047745864189763964, 'samples': 4371968, 'steps': 8538, 'loss/train': 3.206209897994995} 02/24/2022 12:37:30 - INFO - codeparrot_training - Step 8539: {'lr': 0.0004774518514608157, 'samples': 4372480, 'steps': 8539, 'loss/train': 2.961583375930786} 02/24/2022 12:37:33 - INFO - codeparrot_training - Step 8540: {'lr': 0.00047744506004966024, 'samples': 4372992, 'steps': 8540, 'loss/train': 3.248595714569092} 02/24/2022 12:37:39 - INFO - codeparrot_training - Step 8541: {'lr': 0.0004774382676642024, 'samples': 4373504, 'steps': 8541, 'loss/train': 1.8727253675460815} 02/24/2022 12:37:42 - INFO - codeparrot_training - Step 8542: {'lr': 0.0004774314743044712, 'samples': 4374016, 'steps': 8542, 'loss/train': 2.283466100692749} 02/24/2022 12:37:47 - INFO - codeparrot_training - Step 8543: {'lr': 0.00047742467997049576, 'samples': 4374528, 'steps': 8543, 'loss/train': 2.719625949859619} 02/24/2022 12:37:51 - INFO - codeparrot_training - Step 8544: {'lr': 0.00047741788466230527, 'samples': 4375040, 'steps': 8544, 'loss/train': 2.3168108463287354} 02/24/2022 12:37:56 - INFO - codeparrot_training - Step 8545: {'lr': 0.00047741108837992877, 'samples': 4375552, 'steps': 8545, 'loss/train': 2.062335729598999} 02/24/2022 12:38:00 - INFO - codeparrot_training - Step 8546: {'lr': 0.0004774042911233953, 'samples': 4376064, 'steps': 8546, 'loss/train': 1.0751019716262817} 02/24/2022 12:38:06 - INFO - codeparrot_training - Step 8547: {'lr': 0.0004773974928927342, 'samples': 4376576, 'steps': 8547, 'loss/train': 2.533010482788086} 02/24/2022 12:38:09 - INFO - codeparrot_training - Step 8548: {'lr': 0.00047739069368797426, 'samples': 4377088, 'steps': 8548, 'loss/train': 2.0355021953582764} 02/24/2022 12:38:15 - INFO - codeparrot_training - Step 8549: {'lr': 0.0004773838935091449, 'samples': 4377600, 'steps': 8549, 'loss/train': 2.4210667610168457} 02/24/2022 12:38:18 - INFO - codeparrot_training - Step 8550: {'lr': 0.00047737709235627515, 'samples': 4378112, 'steps': 8550, 'loss/train': 1.7051926851272583} 02/24/2022 12:38:24 - INFO - codeparrot_training - Step 8551: {'lr': 0.00047737029022939414, 'samples': 4378624, 'steps': 8551, 'loss/train': 2.4141745567321777} 02/24/2022 12:38:27 - INFO - codeparrot_training - Step 8552: {'lr': 0.00047736348712853094, 'samples': 4379136, 'steps': 8552, 'loss/train': 1.9121068716049194} 02/24/2022 12:38:33 - INFO - codeparrot_training - Step 8553: {'lr': 0.00047735668305371484, 'samples': 4379648, 'steps': 8553, 'loss/train': 1.5458152294158936} 02/24/2022 12:38:36 - INFO - codeparrot_training - Step 8554: {'lr': 0.0004773498780049749, 'samples': 4380160, 'steps': 8554, 'loss/train': 2.028305768966675} 02/24/2022 12:38:42 - INFO - codeparrot_training - Step 8555: {'lr': 0.00047734307198234015, 'samples': 4380672, 'steps': 8555, 'loss/train': 2.7924082279205322} 02/24/2022 12:38:45 - INFO - codeparrot_training - Step 8556: {'lr': 0.00047733626498584, 'samples': 4381184, 'steps': 8556, 'loss/train': 2.3836426734924316} 02/24/2022 12:38:52 - INFO - codeparrot_training - Step 8557: {'lr': 0.0004773294570155035, 'samples': 4381696, 'steps': 8557, 'loss/train': 2.8018240928649902} 02/24/2022 12:38:55 - INFO - codeparrot_training - Step 8558: {'lr': 0.0004773226480713596, 'samples': 4382208, 'steps': 8558, 'loss/train': 2.0178024768829346} 02/24/2022 12:39:01 - INFO - codeparrot_training - Step 8559: {'lr': 0.00047731583815343784, 'samples': 4382720, 'steps': 8559, 'loss/train': 2.6657023429870605} 02/24/2022 12:39:04 - INFO - codeparrot_training - Step 8560: {'lr': 0.00047730902726176715, 'samples': 4383232, 'steps': 8560, 'loss/train': 2.0087077617645264} 02/24/2022 12:39:10 - INFO - codeparrot_training - Step 8561: {'lr': 0.00047730221539637677, 'samples': 4383744, 'steps': 8561, 'loss/train': 1.6984707117080688} 02/24/2022 12:39:13 - INFO - codeparrot_training - Step 8562: {'lr': 0.00047729540255729585, 'samples': 4384256, 'steps': 8562, 'loss/train': 2.4408490657806396} 02/24/2022 12:39:19 - INFO - codeparrot_training - Step 8563: {'lr': 0.0004772885887445536, 'samples': 4384768, 'steps': 8563, 'loss/train': 2.894321918487549} 02/24/2022 12:39:22 - INFO - codeparrot_training - Step 8564: {'lr': 0.0004772817739581793, 'samples': 4385280, 'steps': 8564, 'loss/train': 2.4830737113952637} 02/24/2022 12:39:28 - INFO - codeparrot_training - Step 8565: {'lr': 0.000477274958198202, 'samples': 4385792, 'steps': 8565, 'loss/train': 1.175816297531128} 02/24/2022 12:39:31 - INFO - codeparrot_training - Step 8566: {'lr': 0.0004772681414646509, 'samples': 4386304, 'steps': 8566, 'loss/train': 1.0876022577285767} 02/24/2022 12:39:37 - INFO - codeparrot_training - Step 8567: {'lr': 0.00047726132375755525, 'samples': 4386816, 'steps': 8567, 'loss/train': 1.7903074026107788} 02/24/2022 12:39:41 - INFO - codeparrot_training - Step 8568: {'lr': 0.00047725450507694433, 'samples': 4387328, 'steps': 8568, 'loss/train': 2.4472103118896484} 02/24/2022 12:39:46 - INFO - codeparrot_training - Step 8569: {'lr': 0.00047724768542284726, 'samples': 4387840, 'steps': 8569, 'loss/train': 2.4018044471740723} 02/24/2022 12:39:50 - INFO - codeparrot_training - Step 8570: {'lr': 0.0004772408647952932, 'samples': 4388352, 'steps': 8570, 'loss/train': 1.9022170305252075} 02/24/2022 12:39:56 - INFO - codeparrot_training - Step 8571: {'lr': 0.0004772340431943114, 'samples': 4388864, 'steps': 8571, 'loss/train': 1.9567608833312988} 02/24/2022 12:39:59 - INFO - codeparrot_training - Step 8572: {'lr': 0.0004772272206199312, 'samples': 4389376, 'steps': 8572, 'loss/train': 1.7057418823242188} 02/24/2022 12:40:05 - INFO - codeparrot_training - Step 8573: {'lr': 0.0004772203970721817, 'samples': 4389888, 'steps': 8573, 'loss/train': 0.4100549519062042} 02/24/2022 12:40:08 - INFO - codeparrot_training - Step 8574: {'lr': 0.0004772135725510922, 'samples': 4390400, 'steps': 8574, 'loss/train': 2.523175001144409} 02/24/2022 12:40:14 - INFO - codeparrot_training - Step 8575: {'lr': 0.0004772067470566919, 'samples': 4390912, 'steps': 8575, 'loss/train': 1.1470764875411987} 02/24/2022 12:40:18 - INFO - codeparrot_training - Step 8576: {'lr': 0.00047719992058901006, 'samples': 4391424, 'steps': 8576, 'loss/train': 2.4629693031311035} 02/24/2022 12:40:23 - INFO - codeparrot_training - Step 8577: {'lr': 0.00047719309314807584, 'samples': 4391936, 'steps': 8577, 'loss/train': 2.9566709995269775} 02/24/2022 12:40:27 - INFO - codeparrot_training - Step 8578: {'lr': 0.0004771862647339186, 'samples': 4392448, 'steps': 8578, 'loss/train': 1.8519346714019775} 02/24/2022 12:40:32 - INFO - codeparrot_training - Step 8579: {'lr': 0.0004771794353465675, 'samples': 4392960, 'steps': 8579, 'loss/train': 1.43887197971344} 02/24/2022 12:40:36 - INFO - codeparrot_training - Step 8580: {'lr': 0.00047717260498605186, 'samples': 4393472, 'steps': 8580, 'loss/train': 1.830541968345642} 02/24/2022 12:40:42 - INFO - codeparrot_training - Step 8581: {'lr': 0.0004771657736524009, 'samples': 4393984, 'steps': 8581, 'loss/train': 1.790372371673584} 02/24/2022 12:40:46 - INFO - codeparrot_training - Step 8582: {'lr': 0.00047715894134564395, 'samples': 4394496, 'steps': 8582, 'loss/train': 3.436487913131714} 02/24/2022 12:40:51 - INFO - codeparrot_training - Step 8583: {'lr': 0.0004771521080658102, 'samples': 4395008, 'steps': 8583, 'loss/train': 2.7922022342681885} 02/24/2022 12:40:54 - INFO - codeparrot_training - Step 8584: {'lr': 0.00047714527381292893, 'samples': 4395520, 'steps': 8584, 'loss/train': 3.391402006149292} 02/24/2022 12:41:00 - INFO - codeparrot_training - Step 8585: {'lr': 0.00047713843858702943, 'samples': 4396032, 'steps': 8585, 'loss/train': 2.1233153343200684} 02/24/2022 12:41:03 - INFO - codeparrot_training - Step 8586: {'lr': 0.000477131602388141, 'samples': 4396544, 'steps': 8586, 'loss/train': 2.618825912475586} 02/24/2022 12:41:09 - INFO - codeparrot_training - Step 8587: {'lr': 0.00047712476521629294, 'samples': 4397056, 'steps': 8587, 'loss/train': 1.6917307376861572} 02/24/2022 12:41:12 - INFO - codeparrot_training - Step 8588: {'lr': 0.0004771179270715145, 'samples': 4397568, 'steps': 8588, 'loss/train': 2.150625705718994} 02/24/2022 12:41:18 - INFO - codeparrot_training - Step 8589: {'lr': 0.000477111087953835, 'samples': 4398080, 'steps': 8589, 'loss/train': 2.749589681625366} 02/24/2022 12:41:21 - INFO - codeparrot_training - Step 8590: {'lr': 0.0004771042478632836, 'samples': 4398592, 'steps': 8590, 'loss/train': 2.3499083518981934} 02/24/2022 12:41:27 - INFO - codeparrot_training - Step 8591: {'lr': 0.0004770974067998898, 'samples': 4399104, 'steps': 8591, 'loss/train': 2.635939598083496} 02/24/2022 12:41:30 - INFO - codeparrot_training - Step 8592: {'lr': 0.0004770905647636828, 'samples': 4399616, 'steps': 8592, 'loss/train': 1.1321337223052979} 02/24/2022 12:41:36 - INFO - codeparrot_training - Step 8593: {'lr': 0.00047708372175469193, 'samples': 4400128, 'steps': 8593, 'loss/train': 2.657602071762085} 02/24/2022 12:41:40 - INFO - codeparrot_training - Step 8594: {'lr': 0.0004770768777729465, 'samples': 4400640, 'steps': 8594, 'loss/train': 2.3549373149871826} 02/24/2022 12:41:45 - INFO - codeparrot_training - Step 8595: {'lr': 0.0004770700328184758, 'samples': 4401152, 'steps': 8595, 'loss/train': 1.016804814338684} 02/24/2022 12:41:49 - INFO - codeparrot_training - Step 8596: {'lr': 0.00047706318689130924, 'samples': 4401664, 'steps': 8596, 'loss/train': 2.461090326309204} 02/24/2022 12:41:54 - INFO - codeparrot_training - Step 8597: {'lr': 0.0004770563399914761, 'samples': 4402176, 'steps': 8597, 'loss/train': 3.0149478912353516} 02/24/2022 12:41:58 - INFO - codeparrot_training - Step 8598: {'lr': 0.00047704949211900565, 'samples': 4402688, 'steps': 8598, 'loss/train': 2.4274518489837646} 02/24/2022 12:42:03 - INFO - codeparrot_training - Step 8599: {'lr': 0.0004770426432739273, 'samples': 4403200, 'steps': 8599, 'loss/train': 2.5701568126678467} 02/24/2022 12:42:07 - INFO - codeparrot_training - Step 8600: {'lr': 0.00047703579345627036, 'samples': 4403712, 'steps': 8600, 'loss/train': 1.3668849468231201} 02/24/2022 12:42:12 - INFO - codeparrot_training - Step 8601: {'lr': 0.00047702894266606413, 'samples': 4404224, 'steps': 8601, 'loss/train': 1.8855148553848267} 02/24/2022 12:42:16 - INFO - codeparrot_training - Step 8602: {'lr': 0.00047702209090333804, 'samples': 4404736, 'steps': 8602, 'loss/train': 3.4527690410614014} 02/24/2022 12:42:22 - INFO - codeparrot_training - Step 8603: {'lr': 0.0004770152381681214, 'samples': 4405248, 'steps': 8603, 'loss/train': 2.4510040283203125} 02/24/2022 12:42:26 - INFO - codeparrot_training - Step 8604: {'lr': 0.0004770083844604435, 'samples': 4405760, 'steps': 8604, 'loss/train': 2.505350351333618} 02/24/2022 12:42:32 - INFO - codeparrot_training - Step 8605: {'lr': 0.00047700152978033387, 'samples': 4406272, 'steps': 8605, 'loss/train': 2.3096370697021484} 02/24/2022 12:42:35 - INFO - codeparrot_training - Step 8606: {'lr': 0.0004769946741278217, 'samples': 4406784, 'steps': 8606, 'loss/train': 1.517177939414978} 02/24/2022 12:42:41 - INFO - codeparrot_training - Step 8607: {'lr': 0.00047698781750293644, 'samples': 4407296, 'steps': 8607, 'loss/train': 2.5887279510498047} 02/24/2022 12:42:44 - INFO - codeparrot_training - Step 8608: {'lr': 0.00047698095990570744, 'samples': 4407808, 'steps': 8608, 'loss/train': 2.6150879859924316} 02/24/2022 12:42:50 - INFO - codeparrot_training - Step 8609: {'lr': 0.00047697410133616414, 'samples': 4408320, 'steps': 8609, 'loss/train': 1.7853350639343262} 02/24/2022 12:42:53 - INFO - codeparrot_training - Step 8610: {'lr': 0.0004769672417943358, 'samples': 4408832, 'steps': 8610, 'loss/train': 1.4286537170410156} 02/24/2022 12:42:59 - INFO - codeparrot_training - Step 8611: {'lr': 0.00047696038128025185, 'samples': 4409344, 'steps': 8611, 'loss/train': 2.2056002616882324} 02/24/2022 12:43:02 - INFO - codeparrot_training - Step 8612: {'lr': 0.00047695351979394173, 'samples': 4409856, 'steps': 8612, 'loss/train': 1.807463526725769} 02/24/2022 12:43:08 - INFO - codeparrot_training - Step 8613: {'lr': 0.00047694665733543485, 'samples': 4410368, 'steps': 8613, 'loss/train': 3.0162248611450195} 02/24/2022 12:43:11 - INFO - codeparrot_training - Step 8614: {'lr': 0.00047693979390476046, 'samples': 4410880, 'steps': 8614, 'loss/train': 1.6820416450500488} 02/24/2022 12:43:17 - INFO - codeparrot_training - Step 8615: {'lr': 0.00047693292950194813, 'samples': 4411392, 'steps': 8615, 'loss/train': 2.2958154678344727} 02/24/2022 12:43:23 - INFO - codeparrot_training - Step 8616: {'lr': 0.0004769260641270271, 'samples': 4411904, 'steps': 8616, 'loss/train': 1.3787344694137573} 02/24/2022 12:43:26 - INFO - codeparrot_training - Step 8617: {'lr': 0.0004769191977800269, 'samples': 4412416, 'steps': 8617, 'loss/train': 2.0811045169830322} 02/24/2022 12:43:32 - INFO - codeparrot_training - Step 8618: {'lr': 0.0004769123304609769, 'samples': 4412928, 'steps': 8618, 'loss/train': 1.517041563987732} 02/24/2022 12:43:35 - INFO - codeparrot_training - Step 8619: {'lr': 0.0004769054621699066, 'samples': 4413440, 'steps': 8619, 'loss/train': 2.7214767932891846} 02/24/2022 12:43:41 - INFO - codeparrot_training - Step 8620: {'lr': 0.0004768985929068453, 'samples': 4413952, 'steps': 8620, 'loss/train': 1.9583196640014648} 02/24/2022 12:43:44 - INFO - codeparrot_training - Step 8621: {'lr': 0.0004768917226718225, 'samples': 4414464, 'steps': 8621, 'loss/train': 2.4468472003936768} 02/24/2022 12:43:50 - INFO - codeparrot_training - Step 8622: {'lr': 0.0004768848514648676, 'samples': 4414976, 'steps': 8622, 'loss/train': 2.4089601039886475} 02/24/2022 12:43:53 - INFO - codeparrot_training - Step 8623: {'lr': 0.0004768779792860101, 'samples': 4415488, 'steps': 8623, 'loss/train': 1.7274982929229736} 02/24/2022 12:43:59 - INFO - codeparrot_training - Step 8624: {'lr': 0.00047687110613527924, 'samples': 4416000, 'steps': 8624, 'loss/train': 1.3341413736343384} 02/24/2022 12:44:03 - INFO - codeparrot_training - Step 8625: {'lr': 0.0004768642320127047, 'samples': 4416512, 'steps': 8625, 'loss/train': 2.6436188220977783} 02/24/2022 12:44:08 - INFO - codeparrot_training - Step 8626: {'lr': 0.0004768573569183158, 'samples': 4417024, 'steps': 8626, 'loss/train': 1.0955262184143066} 02/24/2022 12:44:12 - INFO - codeparrot_training - Step 8627: {'lr': 0.000476850480852142, 'samples': 4417536, 'steps': 8627, 'loss/train': 2.3829903602600098} 02/24/2022 12:44:17 - INFO - codeparrot_training - Step 8628: {'lr': 0.0004768436038142128, 'samples': 4418048, 'steps': 8628, 'loss/train': 1.5524612665176392} 02/24/2022 12:44:21 - INFO - codeparrot_training - Step 8629: {'lr': 0.00047683672580455764, 'samples': 4418560, 'steps': 8629, 'loss/train': 3.0201761722564697} 02/24/2022 12:44:26 - INFO - codeparrot_training - Step 8630: {'lr': 0.00047682984682320597, 'samples': 4419072, 'steps': 8630, 'loss/train': 1.336091160774231} 02/24/2022 12:44:30 - INFO - codeparrot_training - Step 8631: {'lr': 0.0004768229668701872, 'samples': 4419584, 'steps': 8631, 'loss/train': 2.226219892501831} 02/24/2022 12:44:35 - INFO - codeparrot_training - Step 8632: {'lr': 0.00047681608594553093, 'samples': 4420096, 'steps': 8632, 'loss/train': 2.4161622524261475} 02/24/2022 12:44:39 - INFO - codeparrot_training - Step 8633: {'lr': 0.00047680920404926655, 'samples': 4420608, 'steps': 8633, 'loss/train': 2.278937578201294} 02/24/2022 12:44:45 - INFO - codeparrot_training - Step 8634: {'lr': 0.0004768023211814236, 'samples': 4421120, 'steps': 8634, 'loss/train': 0.4753687381744385} 02/24/2022 12:44:48 - INFO - codeparrot_training - Step 8635: {'lr': 0.0004767954373420315, 'samples': 4421632, 'steps': 8635, 'loss/train': 0.9312040209770203} 02/24/2022 12:44:54 - INFO - codeparrot_training - Step 8636: {'lr': 0.0004767885525311197, 'samples': 4422144, 'steps': 8636, 'loss/train': 0.989240288734436} 02/24/2022 12:44:57 - INFO - codeparrot_training - Step 8637: {'lr': 0.00047678166674871783, 'samples': 4422656, 'steps': 8637, 'loss/train': 2.154517889022827} 02/24/2022 12:45:03 - INFO - codeparrot_training - Step 8638: {'lr': 0.0004767747799948553, 'samples': 4423168, 'steps': 8638, 'loss/train': 1.625832438468933} 02/24/2022 12:45:06 - INFO - codeparrot_training - Step 8639: {'lr': 0.0004767678922695616, 'samples': 4423680, 'steps': 8639, 'loss/train': 2.3734169006347656} 02/24/2022 12:45:12 - INFO - codeparrot_training - Step 8640: {'lr': 0.0004767610035728662, 'samples': 4424192, 'steps': 8640, 'loss/train': 2.1616783142089844} 02/24/2022 12:45:16 - INFO - codeparrot_training - Step 8641: {'lr': 0.00047675411390479876, 'samples': 4424704, 'steps': 8641, 'loss/train': 2.3193087577819824} 02/24/2022 12:45:21 - INFO - codeparrot_training - Step 8642: {'lr': 0.0004767472232653887, 'samples': 4425216, 'steps': 8642, 'loss/train': 1.4009196758270264} 02/24/2022 12:45:25 - INFO - codeparrot_training - Step 8643: {'lr': 0.00047674033165466545, 'samples': 4425728, 'steps': 8643, 'loss/train': 1.9185664653778076} 02/24/2022 12:45:31 - INFO - codeparrot_training - Step 8644: {'lr': 0.0004767334390726588, 'samples': 4426240, 'steps': 8644, 'loss/train': 2.3010120391845703} 02/24/2022 12:45:34 - INFO - codeparrot_training - Step 8645: {'lr': 0.00047672654551939785, 'samples': 4426752, 'steps': 8645, 'loss/train': 2.000993251800537} 02/24/2022 12:45:40 - INFO - codeparrot_training - Step 8646: {'lr': 0.00047671965099491256, 'samples': 4427264, 'steps': 8646, 'loss/train': 2.680065631866455} 02/24/2022 12:45:43 - INFO - codeparrot_training - Step 8647: {'lr': 0.0004767127554992322, 'samples': 4427776, 'steps': 8647, 'loss/train': 2.972494602203369} 02/24/2022 12:45:49 - INFO - codeparrot_training - Step 8648: {'lr': 0.0004767058590323864, 'samples': 4428288, 'steps': 8648, 'loss/train': 1.4183294773101807} 02/24/2022 12:45:52 - INFO - codeparrot_training - Step 8649: {'lr': 0.00047669896159440464, 'samples': 4428800, 'steps': 8649, 'loss/train': 2.120751142501831} 02/24/2022 12:45:58 - INFO - codeparrot_training - Step 8650: {'lr': 0.00047669206318531654, 'samples': 4429312, 'steps': 8650, 'loss/train': 2.8385157585144043} 02/24/2022 12:46:01 - INFO - codeparrot_training - Step 8651: {'lr': 0.00047668516380515165, 'samples': 4429824, 'steps': 8651, 'loss/train': 2.1079602241516113} 02/24/2022 12:46:07 - INFO - codeparrot_training - Step 8652: {'lr': 0.0004766782634539395, 'samples': 4430336, 'steps': 8652, 'loss/train': 1.5508980751037598} 02/24/2022 12:46:10 - INFO - codeparrot_training - Step 8653: {'lr': 0.00047667136213170957, 'samples': 4430848, 'steps': 8653, 'loss/train': 3.2541048526763916} 02/24/2022 12:46:16 - INFO - codeparrot_training - Step 8654: {'lr': 0.00047666445983849163, 'samples': 4431360, 'steps': 8654, 'loss/train': 1.6311894655227661} 02/24/2022 12:46:19 - INFO - codeparrot_training - Step 8655: {'lr': 0.000476657556574315, 'samples': 4431872, 'steps': 8655, 'loss/train': 2.0137391090393066} 02/24/2022 12:46:25 - INFO - codeparrot_training - Step 8656: {'lr': 0.00047665065233920946, 'samples': 4432384, 'steps': 8656, 'loss/train': 1.5860389471054077} 02/24/2022 12:46:28 - INFO - codeparrot_training - Step 8657: {'lr': 0.0004766437471332045, 'samples': 4432896, 'steps': 8657, 'loss/train': 0.8020785450935364} 02/24/2022 12:46:34 - INFO - codeparrot_training - Step 8658: {'lr': 0.0004766368409563296, 'samples': 4433408, 'steps': 8658, 'loss/train': 2.4337875843048096} 02/24/2022 12:46:37 - INFO - codeparrot_training - Step 8659: {'lr': 0.0004766299338086145, 'samples': 4433920, 'steps': 8659, 'loss/train': 1.8769762516021729} 02/24/2022 12:46:43 - INFO - codeparrot_training - Step 8660: {'lr': 0.0004766230256900887, 'samples': 4434432, 'steps': 8660, 'loss/train': 2.092787504196167} 02/24/2022 12:46:47 - INFO - codeparrot_training - Step 8661: {'lr': 0.00047661611660078184, 'samples': 4434944, 'steps': 8661, 'loss/train': 2.9152870178222656} 02/24/2022 12:46:52 - INFO - codeparrot_training - Step 8662: {'lr': 0.0004766092065407235, 'samples': 4435456, 'steps': 8662, 'loss/train': 2.8709444999694824} 02/24/2022 12:46:56 - INFO - codeparrot_training - Step 8663: {'lr': 0.0004766022955099433, 'samples': 4435968, 'steps': 8663, 'loss/train': 2.575606346130371} 02/24/2022 12:47:01 - INFO - codeparrot_training - Step 8664: {'lr': 0.00047659538350847076, 'samples': 4436480, 'steps': 8664, 'loss/train': 1.45050847530365} 02/24/2022 12:47:05 - INFO - codeparrot_training - Step 8665: {'lr': 0.00047658847053633555, 'samples': 4436992, 'steps': 8665, 'loss/train': 1.629294753074646} 02/24/2022 12:47:10 - INFO - codeparrot_training - Step 8666: {'lr': 0.00047658155659356725, 'samples': 4437504, 'steps': 8666, 'loss/train': 2.524251937866211} 02/24/2022 12:47:14 - INFO - codeparrot_training - Step 8667: {'lr': 0.0004765746416801956, 'samples': 4438016, 'steps': 8667, 'loss/train': 2.461313247680664} 02/24/2022 12:47:19 - INFO - codeparrot_training - Step 8668: {'lr': 0.0004765677257962501, 'samples': 4438528, 'steps': 8668, 'loss/train': 2.9091362953186035} 02/24/2022 12:47:23 - INFO - codeparrot_training - Step 8669: {'lr': 0.0004765608089417604, 'samples': 4439040, 'steps': 8669, 'loss/train': 2.237517833709717} 02/24/2022 12:47:29 - INFO - codeparrot_training - Step 8670: {'lr': 0.0004765538911167562, 'samples': 4439552, 'steps': 8670, 'loss/train': 2.2874817848205566} 02/24/2022 12:47:32 - INFO - codeparrot_training - Step 8671: {'lr': 0.00047654697232126696, 'samples': 4440064, 'steps': 8671, 'loss/train': 1.8436689376831055} 02/24/2022 12:47:38 - INFO - codeparrot_training - Step 8672: {'lr': 0.00047654005255532247, 'samples': 4440576, 'steps': 8672, 'loss/train': 2.346289873123169} 02/24/2022 12:47:41 - INFO - codeparrot_training - Step 8673: {'lr': 0.0004765331318189523, 'samples': 4441088, 'steps': 8673, 'loss/train': 1.8630294799804688} 02/24/2022 12:47:47 - INFO - codeparrot_training - Step 8674: {'lr': 0.00047652621011218623, 'samples': 4441600, 'steps': 8674, 'loss/train': 3.237675428390503} 02/24/2022 12:47:50 - INFO - codeparrot_training - Step 8675: {'lr': 0.0004765192874350537, 'samples': 4442112, 'steps': 8675, 'loss/train': 1.8413033485412598} 02/24/2022 12:47:56 - INFO - codeparrot_training - Step 8676: {'lr': 0.0004765123637875845, 'samples': 4442624, 'steps': 8676, 'loss/train': 3.178135633468628} 02/24/2022 12:47:59 - INFO - codeparrot_training - Step 8677: {'lr': 0.00047650543916980827, 'samples': 4443136, 'steps': 8677, 'loss/train': 2.0916380882263184} 02/24/2022 12:48:05 - INFO - codeparrot_training - Step 8678: {'lr': 0.00047649851358175466, 'samples': 4443648, 'steps': 8678, 'loss/train': 2.175136089324951} 02/24/2022 12:48:08 - INFO - codeparrot_training - Step 8679: {'lr': 0.0004764915870234533, 'samples': 4444160, 'steps': 8679, 'loss/train': 2.0423433780670166} 02/24/2022 12:48:15 - INFO - codeparrot_training - Step 8680: {'lr': 0.000476484659494934, 'samples': 4444672, 'steps': 8680, 'loss/train': 3.073350429534912} 02/24/2022 12:48:18 - INFO - codeparrot_training - Step 8681: {'lr': 0.0004764777309962263, 'samples': 4445184, 'steps': 8681, 'loss/train': 2.4866981506347656} 02/24/2022 12:48:24 - INFO - codeparrot_training - Step 8682: {'lr': 0.0004764708015273599, 'samples': 4445696, 'steps': 8682, 'loss/train': 2.342972755432129} 02/24/2022 12:48:27 - INFO - codeparrot_training - Step 8683: {'lr': 0.0004764638710883644, 'samples': 4446208, 'steps': 8683, 'loss/train': 1.2774816751480103} 02/24/2022 12:48:33 - INFO - codeparrot_training - Step 8684: {'lr': 0.0004764569396792697, 'samples': 4446720, 'steps': 8684, 'loss/train': 1.7196332216262817} 02/24/2022 12:48:36 - INFO - codeparrot_training - Step 8685: {'lr': 0.00047645000730010535, 'samples': 4447232, 'steps': 8685, 'loss/train': 0.31277045607566833} 02/24/2022 12:48:42 - INFO - codeparrot_training - Step 8686: {'lr': 0.00047644307395090107, 'samples': 4447744, 'steps': 8686, 'loss/train': 1.9592556953430176} 02/24/2022 12:48:45 - INFO - codeparrot_training - Step 8687: {'lr': 0.0004764361396316866, 'samples': 4448256, 'steps': 8687, 'loss/train': 2.346013307571411} 02/24/2022 12:48:51 - INFO - codeparrot_training - Step 8688: {'lr': 0.0004764292043424916, 'samples': 4448768, 'steps': 8688, 'loss/train': 2.1401946544647217} 02/24/2022 12:48:54 - INFO - codeparrot_training - Step 8689: {'lr': 0.0004764222680833458, 'samples': 4449280, 'steps': 8689, 'loss/train': 2.2720906734466553} 02/24/2022 12:49:01 - INFO - codeparrot_training - Step 8690: {'lr': 0.0004764153308542788, 'samples': 4449792, 'steps': 8690, 'loss/train': 1.8749568462371826} 02/24/2022 12:49:04 - INFO - codeparrot_training - Step 8691: {'lr': 0.0004764083926553205, 'samples': 4450304, 'steps': 8691, 'loss/train': 2.8419225215911865} 02/24/2022 12:49:09 - INFO - codeparrot_training - Step 8692: {'lr': 0.00047640145348650057, 'samples': 4450816, 'steps': 8692, 'loss/train': 1.652675747871399} 02/24/2022 12:49:13 - INFO - codeparrot_training - Step 8693: {'lr': 0.0004763945133478486, 'samples': 4451328, 'steps': 8693, 'loss/train': 2.562063455581665} 02/24/2022 12:49:18 - INFO - codeparrot_training - Step 8694: {'lr': 0.0004763875722393945, 'samples': 4451840, 'steps': 8694, 'loss/train': 1.279549241065979} 02/24/2022 12:49:22 - INFO - codeparrot_training - Step 8695: {'lr': 0.000476380630161168, 'samples': 4452352, 'steps': 8695, 'loss/train': 8.204840660095215} 02/24/2022 12:49:27 - INFO - codeparrot_training - Step 8696: {'lr': 0.00047637368711319863, 'samples': 4452864, 'steps': 8696, 'loss/train': 1.8550820350646973} 02/24/2022 12:49:31 - INFO - codeparrot_training - Step 8697: {'lr': 0.00047636674309551626, 'samples': 4453376, 'steps': 8697, 'loss/train': 1.9099466800689697} 02/24/2022 12:49:36 - INFO - codeparrot_training - Step 8698: {'lr': 0.0004763597981081507, 'samples': 4453888, 'steps': 8698, 'loss/train': 1.665390968322754} 02/24/2022 12:49:40 - INFO - codeparrot_training - Step 8699: {'lr': 0.00047635285215113165, 'samples': 4454400, 'steps': 8699, 'loss/train': 2.509888172149658} 02/24/2022 12:49:46 - INFO - codeparrot_training - Step 8700: {'lr': 0.0004763459052244888, 'samples': 4454912, 'steps': 8700, 'loss/train': 0.6205786466598511} 02/24/2022 12:49:49 - INFO - codeparrot_training - Step 8701: {'lr': 0.0004763389573282521, 'samples': 4455424, 'steps': 8701, 'loss/train': 1.8865984678268433} 02/24/2022 12:49:55 - INFO - codeparrot_training - Step 8702: {'lr': 0.00047633200846245106, 'samples': 4455936, 'steps': 8702, 'loss/train': 2.541926383972168} 02/24/2022 12:49:58 - INFO - codeparrot_training - Step 8703: {'lr': 0.0004763250586271156, 'samples': 4456448, 'steps': 8703, 'loss/train': 2.171534776687622} 02/24/2022 12:50:04 - INFO - codeparrot_training - Step 8704: {'lr': 0.00047631810782227535, 'samples': 4456960, 'steps': 8704, 'loss/train': 1.604278802871704} 02/24/2022 12:50:07 - INFO - codeparrot_training - Step 8705: {'lr': 0.00047631115604796035, 'samples': 4457472, 'steps': 8705, 'loss/train': 1.72527277469635} 02/24/2022 12:50:14 - INFO - codeparrot_training - Step 8706: {'lr': 0.0004763042033042001, 'samples': 4457984, 'steps': 8706, 'loss/train': 0.8240301012992859} 02/24/2022 12:50:17 - INFO - codeparrot_training - Step 8707: {'lr': 0.0004762972495910246, 'samples': 4458496, 'steps': 8707, 'loss/train': 1.293099284172058} 02/24/2022 12:50:23 - INFO - codeparrot_training - Step 8708: {'lr': 0.00047629029490846346, 'samples': 4459008, 'steps': 8708, 'loss/train': 2.628904104232788} 02/24/2022 12:50:26 - INFO - codeparrot_training - Step 8709: {'lr': 0.0004762833392565466, 'samples': 4459520, 'steps': 8709, 'loss/train': 1.8404642343521118} 02/24/2022 12:50:32 - INFO - codeparrot_training - Step 8710: {'lr': 0.00047627638263530374, 'samples': 4460032, 'steps': 8710, 'loss/train': 1.2400083541870117} 02/24/2022 12:50:35 - INFO - codeparrot_training - Step 8711: {'lr': 0.00047626942504476477, 'samples': 4460544, 'steps': 8711, 'loss/train': 2.5367848873138428} 02/24/2022 12:50:41 - INFO - codeparrot_training - Step 8712: {'lr': 0.00047626246648495936, 'samples': 4461056, 'steps': 8712, 'loss/train': 2.0482499599456787} 02/24/2022 12:50:44 - INFO - codeparrot_training - Step 8713: {'lr': 0.0004762555069559175, 'samples': 4461568, 'steps': 8713, 'loss/train': 2.862839937210083} 02/24/2022 12:50:49 - INFO - codeparrot_training - Step 8714: {'lr': 0.00047624854645766875, 'samples': 4462080, 'steps': 8714, 'loss/train': 3.66365122795105} 02/24/2022 12:50:53 - INFO - codeparrot_training - Step 8715: {'lr': 0.0004762415849902431, 'samples': 4462592, 'steps': 8715, 'loss/train': 1.9151747226715088} 02/24/2022 12:50:59 - INFO - codeparrot_training - Step 8716: {'lr': 0.0004762346225536703, 'samples': 4463104, 'steps': 8716, 'loss/train': 2.9629907608032227} 02/24/2022 12:51:03 - INFO - codeparrot_training - Step 8717: {'lr': 0.0004762276591479804, 'samples': 4463616, 'steps': 8717, 'loss/train': 2.2446770668029785} 02/24/2022 12:51:08 - INFO - codeparrot_training - Step 8718: {'lr': 0.00047622069477320285, 'samples': 4464128, 'steps': 8718, 'loss/train': 1.707044243812561} 02/24/2022 12:51:12 - INFO - codeparrot_training - Step 8719: {'lr': 0.0004762137294293678, 'samples': 4464640, 'steps': 8719, 'loss/train': 1.3499451875686646} 02/24/2022 12:51:17 - INFO - codeparrot_training - Step 8720: {'lr': 0.0004762067631165049, 'samples': 4465152, 'steps': 8720, 'loss/train': 4.016249179840088} 02/24/2022 12:51:21 - INFO - codeparrot_training - Step 8721: {'lr': 0.0004761997958346441, 'samples': 4465664, 'steps': 8721, 'loss/train': 1.7870075702667236} 02/24/2022 12:51:26 - INFO - codeparrot_training - Step 8722: {'lr': 0.00047619282758381513, 'samples': 4466176, 'steps': 8722, 'loss/train': 2.3704593181610107} 02/24/2022 12:51:30 - INFO - codeparrot_training - Step 8723: {'lr': 0.0004761858583640479, 'samples': 4466688, 'steps': 8723, 'loss/train': 2.6884233951568604} 02/24/2022 12:51:35 - INFO - codeparrot_training - Step 8724: {'lr': 0.00047617888817537234, 'samples': 4467200, 'steps': 8724, 'loss/train': 1.7528960704803467} 02/24/2022 12:51:39 - INFO - codeparrot_training - Step 8725: {'lr': 0.00047617191701781824, 'samples': 4467712, 'steps': 8725, 'loss/train': 3.846125364303589} 02/24/2022 12:51:45 - INFO - codeparrot_training - Step 8726: {'lr': 0.0004761649448914155, 'samples': 4468224, 'steps': 8726, 'loss/train': 1.9477664232254028} 02/24/2022 12:51:48 - INFO - codeparrot_training - Step 8727: {'lr': 0.0004761579717961939, 'samples': 4468736, 'steps': 8727, 'loss/train': 2.341510057449341} 02/24/2022 12:51:54 - INFO - codeparrot_training - Step 8728: {'lr': 0.0004761509977321834, 'samples': 4469248, 'steps': 8728, 'loss/train': 2.20957088470459} 02/24/2022 12:51:57 - INFO - codeparrot_training - Step 8729: {'lr': 0.0004761440226994138, 'samples': 4469760, 'steps': 8729, 'loss/train': 2.3531336784362793} 02/24/2022 12:52:03 - INFO - codeparrot_training - Step 8730: {'lr': 0.000476137046697915, 'samples': 4470272, 'steps': 8730, 'loss/train': 1.7737395763397217} 02/24/2022 12:52:06 - INFO - codeparrot_training - Step 8731: {'lr': 0.0004761300697277169, 'samples': 4470784, 'steps': 8731, 'loss/train': 2.7883336544036865} 02/24/2022 12:52:12 - INFO - codeparrot_training - Step 8732: {'lr': 0.0004761230917888494, 'samples': 4471296, 'steps': 8732, 'loss/train': 2.4218623638153076} 02/24/2022 12:52:15 - INFO - codeparrot_training - Step 8733: {'lr': 0.00047611611288134236, 'samples': 4471808, 'steps': 8733, 'loss/train': 2.199613571166992} 02/24/2022 12:52:21 - INFO - codeparrot_training - Step 8734: {'lr': 0.00047610913300522576, 'samples': 4472320, 'steps': 8734, 'loss/train': 2.6561696529388428} 02/24/2022 12:52:24 - INFO - codeparrot_training - Step 8735: {'lr': 0.00047610215216052946, 'samples': 4472832, 'steps': 8735, 'loss/train': 2.19661808013916} 02/24/2022 12:52:30 - INFO - codeparrot_training - Step 8736: {'lr': 0.0004760951703472832, 'samples': 4473344, 'steps': 8736, 'loss/train': 3.079450845718384} 02/24/2022 12:52:33 - INFO - codeparrot_training - Step 8737: {'lr': 0.0004760881875655171, 'samples': 4473856, 'steps': 8737, 'loss/train': 1.9540948867797852} 02/24/2022 12:52:40 - INFO - codeparrot_training - Step 8738: {'lr': 0.000476081203815261, 'samples': 4474368, 'steps': 8738, 'loss/train': 2.0764429569244385} 02/24/2022 12:52:43 - INFO - codeparrot_training - Step 8739: {'lr': 0.0004760742190965447, 'samples': 4474880, 'steps': 8739, 'loss/train': 1.7902617454528809} 02/24/2022 12:52:49 - INFO - codeparrot_training - Step 8740: {'lr': 0.0004760672334093984, 'samples': 4475392, 'steps': 8740, 'loss/train': 2.467662811279297} 02/24/2022 12:52:52 - INFO - codeparrot_training - Step 8741: {'lr': 0.0004760602467538517, 'samples': 4475904, 'steps': 8741, 'loss/train': 8.840398788452148} 02/24/2022 12:52:58 - INFO - codeparrot_training - Step 8742: {'lr': 0.0004760532591299348, 'samples': 4476416, 'steps': 8742, 'loss/train': 2.396095037460327} 02/24/2022 12:53:01 - INFO - codeparrot_training - Step 8743: {'lr': 0.00047604627053767754, 'samples': 4476928, 'steps': 8743, 'loss/train': 1.8426545858383179} 02/24/2022 12:53:07 - INFO - codeparrot_training - Step 8744: {'lr': 0.0004760392809771098, 'samples': 4477440, 'steps': 8744, 'loss/train': 3.0879733562469482} 02/24/2022 12:53:10 - INFO - codeparrot_training - Step 8745: {'lr': 0.00047603229044826146, 'samples': 4477952, 'steps': 8745, 'loss/train': 2.3374834060668945} 02/24/2022 12:53:16 - INFO - codeparrot_training - Step 8746: {'lr': 0.00047602529895116264, 'samples': 4478464, 'steps': 8746, 'loss/train': 1.066630244255066} 02/24/2022 12:53:19 - INFO - codeparrot_training - Step 8747: {'lr': 0.0004760183064858432, 'samples': 4478976, 'steps': 8747, 'loss/train': 2.6336004734039307} 02/24/2022 12:53:25 - INFO - codeparrot_training - Step 8748: {'lr': 0.0004760113130523331, 'samples': 4479488, 'steps': 8748, 'loss/train': 2.3310763835906982} 02/24/2022 12:53:28 - INFO - codeparrot_training - Step 8749: {'lr': 0.0004760043186506624, 'samples': 4480000, 'steps': 8749, 'loss/train': 1.129913091659546} 02/24/2022 12:53:34 - INFO - codeparrot_training - Step 8750: {'lr': 0.0004759973232808609, 'samples': 4480512, 'steps': 8750, 'loss/train': 1.3763890266418457} 02/24/2022 12:53:37 - INFO - codeparrot_training - Step 8751: {'lr': 0.0004759903269429585, 'samples': 4481024, 'steps': 8751, 'loss/train': 2.6321349143981934} 02/24/2022 12:53:43 - INFO - codeparrot_training - Step 8752: {'lr': 0.00047598332963698543, 'samples': 4481536, 'steps': 8752, 'loss/train': 2.5127756595611572} 02/24/2022 12:53:47 - INFO - codeparrot_training - Step 8753: {'lr': 0.00047597633136297154, 'samples': 4482048, 'steps': 8753, 'loss/train': 2.2823503017425537} 02/24/2022 12:53:52 - INFO - codeparrot_training - Step 8754: {'lr': 0.0004759693321209467, 'samples': 4482560, 'steps': 8754, 'loss/train': 1.953088641166687} 02/24/2022 12:53:56 - INFO - codeparrot_training - Step 8755: {'lr': 0.00047596233191094114, 'samples': 4483072, 'steps': 8755, 'loss/train': 2.757328748703003} 02/24/2022 12:54:01 - INFO - codeparrot_training - Step 8756: {'lr': 0.0004759553307329846, 'samples': 4483584, 'steps': 8756, 'loss/train': 2.455230951309204} 02/24/2022 12:54:05 - INFO - codeparrot_training - Step 8757: {'lr': 0.00047594832858710725, 'samples': 4484096, 'steps': 8757, 'loss/train': 1.6920143365859985} 02/24/2022 12:54:10 - INFO - codeparrot_training - Step 8758: {'lr': 0.0004759413254733389, 'samples': 4484608, 'steps': 8758, 'loss/train': 1.721900224685669} 02/24/2022 12:54:14 - INFO - codeparrot_training - Step 8759: {'lr': 0.0004759343213917097, 'samples': 4485120, 'steps': 8759, 'loss/train': 2.8824124336242676} 02/24/2022 12:54:19 - INFO - codeparrot_training - Step 8760: {'lr': 0.0004759273163422496, 'samples': 4485632, 'steps': 8760, 'loss/train': 1.133579969406128} 02/24/2022 12:54:23 - INFO - codeparrot_training - Step 8761: {'lr': 0.00047592031032498875, 'samples': 4486144, 'steps': 8761, 'loss/train': 2.9116897583007812} 02/24/2022 12:54:30 - INFO - codeparrot_training - Step 8762: {'lr': 0.00047591330333995684, 'samples': 4486656, 'steps': 8762, 'loss/train': 2.5794615745544434} 02/24/2022 12:54:33 - INFO - codeparrot_training - Step 8763: {'lr': 0.0004759062953871842, 'samples': 4487168, 'steps': 8763, 'loss/train': 2.7960944175720215} 02/24/2022 12:54:39 - INFO - codeparrot_training - Step 8764: {'lr': 0.0004758992864667007, 'samples': 4487680, 'steps': 8764, 'loss/train': 2.0763607025146484} 02/24/2022 12:54:42 - INFO - codeparrot_training - Step 8765: {'lr': 0.0004758922765785363, 'samples': 4488192, 'steps': 8765, 'loss/train': 1.8580307960510254} 02/24/2022 12:54:47 - INFO - codeparrot_training - Step 8766: {'lr': 0.00047588526572272117, 'samples': 4488704, 'steps': 8766, 'loss/train': 1.993737816810608} 02/24/2022 12:54:51 - INFO - codeparrot_training - Step 8767: {'lr': 0.0004758782538992853, 'samples': 4489216, 'steps': 8767, 'loss/train': 2.3059587478637695} 02/24/2022 12:54:56 - INFO - codeparrot_training - Step 8768: {'lr': 0.00047587124110825874, 'samples': 4489728, 'steps': 8768, 'loss/train': 2.352890729904175} 02/24/2022 12:55:02 - INFO - codeparrot_training - Step 8769: {'lr': 0.0004758642273496714, 'samples': 4490240, 'steps': 8769, 'loss/train': 1.7700345516204834} 02/24/2022 12:55:05 - INFO - codeparrot_training - Step 8770: {'lr': 0.0004758572126235535, 'samples': 4490752, 'steps': 8770, 'loss/train': 2.174175977706909} 02/24/2022 12:55:11 - INFO - codeparrot_training - Step 8771: {'lr': 0.0004758501969299351, 'samples': 4491264, 'steps': 8771, 'loss/train': 2.276357650756836} 02/24/2022 12:55:14 - INFO - codeparrot_training - Step 8772: {'lr': 0.0004758431802688461, 'samples': 4491776, 'steps': 8772, 'loss/train': 0.6365671753883362} 02/24/2022 12:55:21 - INFO - codeparrot_training - Step 8773: {'lr': 0.00047583616264031657, 'samples': 4492288, 'steps': 8773, 'loss/train': 2.4573609828948975} 02/24/2022 12:55:25 - INFO - codeparrot_training - Step 8774: {'lr': 0.00047582914404437673, 'samples': 4492800, 'steps': 8774, 'loss/train': 2.514500617980957} 02/24/2022 12:55:30 - INFO - codeparrot_training - Step 8775: {'lr': 0.00047582212448105647, 'samples': 4493312, 'steps': 8775, 'loss/train': 1.933380126953125} 02/24/2022 12:55:34 - INFO - codeparrot_training - Step 8776: {'lr': 0.000475815103950386, 'samples': 4493824, 'steps': 8776, 'loss/train': 2.333723783493042} 02/24/2022 12:55:39 - INFO - codeparrot_training - Step 8777: {'lr': 0.00047580808245239526, 'samples': 4494336, 'steps': 8777, 'loss/train': 2.663935899734497} 02/24/2022 12:55:43 - INFO - codeparrot_training - Step 8778: {'lr': 0.0004758010599871145, 'samples': 4494848, 'steps': 8778, 'loss/train': 2.3152177333831787} 02/24/2022 12:55:48 - INFO - codeparrot_training - Step 8779: {'lr': 0.0004757940365545736, 'samples': 4495360, 'steps': 8779, 'loss/train': 1.993561029434204} 02/24/2022 12:55:52 - INFO - codeparrot_training - Step 8780: {'lr': 0.0004757870121548028, 'samples': 4495872, 'steps': 8780, 'loss/train': 2.6793200969696045} 02/24/2022 12:55:57 - INFO - codeparrot_training - Step 8781: {'lr': 0.00047577998678783207, 'samples': 4496384, 'steps': 8781, 'loss/train': 2.725834608078003} 02/24/2022 12:56:01 - INFO - codeparrot_training - Step 8782: {'lr': 0.0004757729604536917, 'samples': 4496896, 'steps': 8782, 'loss/train': 2.812669515609741} 02/24/2022 12:56:07 - INFO - codeparrot_training - Step 8783: {'lr': 0.0004757659331524115, 'samples': 4497408, 'steps': 8783, 'loss/train': 3.419722080230713} 02/24/2022 12:56:10 - INFO - codeparrot_training - Step 8784: {'lr': 0.00047575890488402183, 'samples': 4497920, 'steps': 8784, 'loss/train': 2.7267181873321533} 02/24/2022 12:56:16 - INFO - codeparrot_training - Step 8785: {'lr': 0.00047575187564855264, 'samples': 4498432, 'steps': 8785, 'loss/train': 1.5478872060775757} 02/24/2022 12:56:19 - INFO - codeparrot_training - Step 8786: {'lr': 0.00047574484544603415, 'samples': 4498944, 'steps': 8786, 'loss/train': 2.7265970706939697} 02/24/2022 12:56:25 - INFO - codeparrot_training - Step 8787: {'lr': 0.00047573781427649644, 'samples': 4499456, 'steps': 8787, 'loss/train': 2.3036603927612305} 02/24/2022 12:56:28 - INFO - codeparrot_training - Step 8788: {'lr': 0.00047573078213996954, 'samples': 4499968, 'steps': 8788, 'loss/train': 2.4199509620666504} 02/24/2022 12:56:34 - INFO - codeparrot_training - Step 8789: {'lr': 0.0004757237490364836, 'samples': 4500480, 'steps': 8789, 'loss/train': 2.5069308280944824} 02/24/2022 12:56:37 - INFO - codeparrot_training - Step 8790: {'lr': 0.00047571671496606893, 'samples': 4500992, 'steps': 8790, 'loss/train': 2.3331410884857178} 02/24/2022 12:56:43 - INFO - codeparrot_training - Step 8791: {'lr': 0.0004757096799287555, 'samples': 4501504, 'steps': 8791, 'loss/train': 1.9628428220748901} 02/24/2022 12:56:46 - INFO - codeparrot_training - Step 8792: {'lr': 0.0004757026439245735, 'samples': 4502016, 'steps': 8792, 'loss/train': 2.3894236087799072} 02/24/2022 12:56:52 - INFO - codeparrot_training - Step 8793: {'lr': 0.00047569560695355295, 'samples': 4502528, 'steps': 8793, 'loss/train': 3.1706998348236084} 02/24/2022 12:56:55 - INFO - codeparrot_training - Step 8794: {'lr': 0.0004756885690157241, 'samples': 4503040, 'steps': 8794, 'loss/train': 1.8535314798355103} 02/24/2022 12:57:01 - INFO - codeparrot_training - Step 8795: {'lr': 0.00047568153011111715, 'samples': 4503552, 'steps': 8795, 'loss/train': 1.6466879844665527} 02/24/2022 12:57:04 - INFO - codeparrot_training - Step 8796: {'lr': 0.00047567449023976213, 'samples': 4504064, 'steps': 8796, 'loss/train': 2.8835508823394775} 02/24/2022 12:57:11 - INFO - codeparrot_training - Step 8797: {'lr': 0.00047566744940168924, 'samples': 4504576, 'steps': 8797, 'loss/train': 2.6650779247283936} 02/24/2022 12:57:14 - INFO - codeparrot_training - Step 8798: {'lr': 0.0004756604075969287, 'samples': 4505088, 'steps': 8798, 'loss/train': 1.3685665130615234} 02/24/2022 12:57:20 - INFO - codeparrot_training - Step 8799: {'lr': 0.0004756533648255106, 'samples': 4505600, 'steps': 8799, 'loss/train': 2.1950764656066895} 02/24/2022 12:57:24 - INFO - codeparrot_training - Step 8800: {'lr': 0.0004756463210874652, 'samples': 4506112, 'steps': 8800, 'loss/train': 2.122185230255127} 02/24/2022 12:57:29 - INFO - codeparrot_training - Step 8801: {'lr': 0.0004756392763828226, 'samples': 4506624, 'steps': 8801, 'loss/train': 1.2308300733566284} 02/24/2022 12:57:33 - INFO - codeparrot_training - Step 8802: {'lr': 0.0004756322307116129, 'samples': 4507136, 'steps': 8802, 'loss/train': 2.173133134841919} 02/24/2022 12:57:38 - INFO - codeparrot_training - Step 8803: {'lr': 0.0004756251840738664, 'samples': 4507648, 'steps': 8803, 'loss/train': 2.749377489089966} 02/24/2022 12:57:42 - INFO - codeparrot_training - Step 8804: {'lr': 0.00047561813646961325, 'samples': 4508160, 'steps': 8804, 'loss/train': 1.2036086320877075} 02/24/2022 12:57:47 - INFO - codeparrot_training - Step 8805: {'lr': 0.00047561108789888367, 'samples': 4508672, 'steps': 8805, 'loss/train': 2.545456886291504} 02/24/2022 12:57:51 - INFO - codeparrot_training - Step 8806: {'lr': 0.0004756040383617078, 'samples': 4509184, 'steps': 8806, 'loss/train': 2.6579318046569824} 02/24/2022 12:57:56 - INFO - codeparrot_training - Step 8807: {'lr': 0.00047559698785811595, 'samples': 4509696, 'steps': 8807, 'loss/train': 2.8812873363494873} 02/24/2022 12:58:00 - INFO - codeparrot_training - Step 8808: {'lr': 0.0004755899363881382, 'samples': 4510208, 'steps': 8808, 'loss/train': 2.4146788120269775} 02/24/2022 12:58:06 - INFO - codeparrot_training - Step 8809: {'lr': 0.00047558288395180477, 'samples': 4510720, 'steps': 8809, 'loss/train': 1.9129815101623535} 02/24/2022 12:58:10 - INFO - codeparrot_training - Step 8810: {'lr': 0.0004755758305491459, 'samples': 4511232, 'steps': 8810, 'loss/train': 1.725149154663086} 02/24/2022 12:58:15 - INFO - codeparrot_training - Step 8811: {'lr': 0.0004755687761801918, 'samples': 4511744, 'steps': 8811, 'loss/train': 0.24333910644054413} 02/24/2022 12:58:19 - INFO - codeparrot_training - Step 8812: {'lr': 0.00047556172084497274, 'samples': 4512256, 'steps': 8812, 'loss/train': 2.304560899734497} 02/24/2022 12:58:24 - INFO - codeparrot_training - Step 8813: {'lr': 0.0004755546645435188, 'samples': 4512768, 'steps': 8813, 'loss/train': 2.0783276557922363} 02/24/2022 12:58:28 - INFO - codeparrot_training - Step 8814: {'lr': 0.0004755476072758604, 'samples': 4513280, 'steps': 8814, 'loss/train': 1.448231816291809} 02/24/2022 12:58:33 - INFO - codeparrot_training - Step 8815: {'lr': 0.0004755405490420276, 'samples': 4513792, 'steps': 8815, 'loss/train': 2.2205657958984375} 02/24/2022 12:58:37 - INFO - codeparrot_training - Step 8816: {'lr': 0.0004755334898420507, 'samples': 4514304, 'steps': 8816, 'loss/train': 2.519583225250244} 02/24/2022 12:58:42 - INFO - codeparrot_training - Step 8817: {'lr': 0.00047552642967596, 'samples': 4514816, 'steps': 8817, 'loss/train': 2.266977310180664} 02/24/2022 12:58:46 - INFO - codeparrot_training - Step 8818: {'lr': 0.00047551936854378564, 'samples': 4515328, 'steps': 8818, 'loss/train': 1.1768916845321655} 02/24/2022 12:58:52 - INFO - codeparrot_training - Step 8819: {'lr': 0.00047551230644555793, 'samples': 4515840, 'steps': 8819, 'loss/train': 2.3790745735168457} 02/24/2022 12:58:55 - INFO - codeparrot_training - Step 8820: {'lr': 0.00047550524338130706, 'samples': 4516352, 'steps': 8820, 'loss/train': 2.516787052154541} 02/24/2022 12:59:01 - INFO - codeparrot_training - Step 8821: {'lr': 0.00047549817935106344, 'samples': 4516864, 'steps': 8821, 'loss/train': 2.075530767440796} 02/24/2022 12:59:04 - INFO - codeparrot_training - Step 8822: {'lr': 0.00047549111435485716, 'samples': 4517376, 'steps': 8822, 'loss/train': 1.9331542253494263} 02/24/2022 12:59:10 - INFO - codeparrot_training - Step 8823: {'lr': 0.0004754840483927185, 'samples': 4517888, 'steps': 8823, 'loss/train': 2.2771711349487305} 02/24/2022 12:59:13 - INFO - codeparrot_training - Step 8824: {'lr': 0.0004754769814646779, 'samples': 4518400, 'steps': 8824, 'loss/train': 1.1682443618774414} 02/24/2022 12:59:19 - INFO - codeparrot_training - Step 8825: {'lr': 0.00047546991357076544, 'samples': 4518912, 'steps': 8825, 'loss/train': 2.4097707271575928} 02/24/2022 12:59:22 - INFO - codeparrot_training - Step 8826: {'lr': 0.00047546284471101143, 'samples': 4519424, 'steps': 8826, 'loss/train': 0.9507866501808167} 02/24/2022 12:59:28 - INFO - codeparrot_training - Step 8827: {'lr': 0.00047545577488544623, 'samples': 4519936, 'steps': 8827, 'loss/train': 2.25288987159729} 02/24/2022 12:59:31 - INFO - codeparrot_training - Step 8828: {'lr': 0.0004754487040941001, 'samples': 4520448, 'steps': 8828, 'loss/train': 1.9752769470214844} 02/24/2022 12:59:37 - INFO - codeparrot_training - Step 8829: {'lr': 0.00047544163233700324, 'samples': 4520960, 'steps': 8829, 'loss/train': 2.9149653911590576} 02/24/2022 12:59:41 - INFO - codeparrot_training - Step 8830: {'lr': 0.00047543455961418605, 'samples': 4521472, 'steps': 8830, 'loss/train': 1.2913155555725098} 02/24/2022 12:59:46 - INFO - codeparrot_training - Step 8831: {'lr': 0.0004754274859256788, 'samples': 4521984, 'steps': 8831, 'loss/train': 2.1174747943878174} 02/24/2022 12:59:50 - INFO - codeparrot_training - Step 8832: {'lr': 0.0004754204112715118, 'samples': 4522496, 'steps': 8832, 'loss/train': 2.7904491424560547} 02/24/2022 12:59:56 - INFO - codeparrot_training - Step 8833: {'lr': 0.0004754133356517153, 'samples': 4523008, 'steps': 8833, 'loss/train': 2.5591137409210205} 02/24/2022 12:59:59 - INFO - codeparrot_training - Step 8834: {'lr': 0.0004754062590663196, 'samples': 4523520, 'steps': 8834, 'loss/train': 2.704284429550171} 02/24/2022 13:00:05 - INFO - codeparrot_training - Step 8835: {'lr': 0.00047539918151535515, 'samples': 4524032, 'steps': 8835, 'loss/train': 0.2202148139476776} 02/24/2022 13:00:08 - INFO - codeparrot_training - Step 8836: {'lr': 0.00047539210299885217, 'samples': 4524544, 'steps': 8836, 'loss/train': 1.999778151512146} 02/24/2022 13:00:14 - INFO - codeparrot_training - Step 8837: {'lr': 0.00047538502351684097, 'samples': 4525056, 'steps': 8837, 'loss/train': 2.7448577880859375} 02/24/2022 13:00:19 - INFO - codeparrot_training - Step 8838: {'lr': 0.0004753779430693519, 'samples': 4525568, 'steps': 8838, 'loss/train': 1.8795011043548584} 02/24/2022 13:00:23 - INFO - codeparrot_training - Step 8839: {'lr': 0.0004753708616564153, 'samples': 4526080, 'steps': 8839, 'loss/train': 1.963841199874878} 02/24/2022 13:00:28 - INFO - codeparrot_training - Step 8840: {'lr': 0.00047536377927806143, 'samples': 4526592, 'steps': 8840, 'loss/train': 1.7777982950210571} 02/24/2022 13:00:32 - INFO - codeparrot_training - Step 8841: {'lr': 0.0004753566959343207, 'samples': 4527104, 'steps': 8841, 'loss/train': 1.4867552518844604} 02/24/2022 13:00:37 - INFO - codeparrot_training - Step 8842: {'lr': 0.0004753496116252235, 'samples': 4527616, 'steps': 8842, 'loss/train': 1.454268217086792} 02/24/2022 13:00:41 - INFO - codeparrot_training - Step 8843: {'lr': 0.0004753425263508001, 'samples': 4528128, 'steps': 8843, 'loss/train': 2.5899415016174316} 02/24/2022 13:00:47 - INFO - codeparrot_training - Step 8844: {'lr': 0.0004753354401110809, 'samples': 4528640, 'steps': 8844, 'loss/train': 1.742884635925293} 02/24/2022 13:00:51 - INFO - codeparrot_training - Step 8845: {'lr': 0.00047532835290609623, 'samples': 4529152, 'steps': 8845, 'loss/train': 1.3017199039459229} 02/24/2022 13:00:54 - INFO - codeparrot_training - Step 8846: {'lr': 0.00047532126473587635, 'samples': 4529664, 'steps': 8846, 'loss/train': 2.9739320278167725} 02/24/2022 13:01:00 - INFO - codeparrot_training - Step 8847: {'lr': 0.0004753141756004518, 'samples': 4530176, 'steps': 8847, 'loss/train': 0.5695855021476746} 02/24/2022 13:01:05 - INFO - codeparrot_training - Step 8848: {'lr': 0.00047530708549985287, 'samples': 4530688, 'steps': 8848, 'loss/train': 1.966842532157898} 02/24/2022 13:01:09 - INFO - codeparrot_training - Step 8849: {'lr': 0.00047529999443410986, 'samples': 4531200, 'steps': 8849, 'loss/train': 2.7894649505615234} 02/24/2022 13:01:14 - INFO - codeparrot_training - Step 8850: {'lr': 0.0004752929024032533, 'samples': 4531712, 'steps': 8850, 'loss/train': 2.735463857650757} 02/24/2022 13:01:18 - INFO - codeparrot_training - Step 8851: {'lr': 0.0004752858094073134, 'samples': 4532224, 'steps': 8851, 'loss/train': 3.1281850337982178} 02/24/2022 13:01:23 - INFO - codeparrot_training - Step 8852: {'lr': 0.0004752787154463207, 'samples': 4532736, 'steps': 8852, 'loss/train': 1.266960620880127} 02/24/2022 13:01:27 - INFO - codeparrot_training - Step 8853: {'lr': 0.0004752716205203055, 'samples': 4533248, 'steps': 8853, 'loss/train': 2.5410046577453613} 02/24/2022 13:01:33 - INFO - codeparrot_training - Step 8854: {'lr': 0.0004752645246292982, 'samples': 4533760, 'steps': 8854, 'loss/train': 2.573962450027466} 02/24/2022 13:01:37 - INFO - codeparrot_training - Step 8855: {'lr': 0.0004752574277733292, 'samples': 4534272, 'steps': 8855, 'loss/train': 2.521437406539917} 02/24/2022 13:01:40 - INFO - codeparrot_training - Step 8856: {'lr': 0.0004752503299524289, 'samples': 4534784, 'steps': 8856, 'loss/train': 2.4318201541900635} 02/24/2022 13:01:46 - INFO - codeparrot_training - Step 8857: {'lr': 0.0004752432311666277, 'samples': 4535296, 'steps': 8857, 'loss/train': 2.0192575454711914} 02/24/2022 13:01:49 - INFO - codeparrot_training - Step 8858: {'lr': 0.0004752361314159561, 'samples': 4535808, 'steps': 8858, 'loss/train': 3.0350465774536133} 02/24/2022 13:01:55 - INFO - codeparrot_training - Step 8859: {'lr': 0.0004752290307004444, 'samples': 4536320, 'steps': 8859, 'loss/train': 2.6406280994415283} 02/24/2022 13:01:58 - INFO - codeparrot_training - Step 8860: {'lr': 0.000475221929020123, 'samples': 4536832, 'steps': 8860, 'loss/train': 1.559017300605774} 02/24/2022 13:02:04 - INFO - codeparrot_training - Step 8861: {'lr': 0.00047521482637502246, 'samples': 4537344, 'steps': 8861, 'loss/train': 1.950171709060669} 02/24/2022 13:02:07 - INFO - codeparrot_training - Step 8862: {'lr': 0.00047520772276517297, 'samples': 4537856, 'steps': 8862, 'loss/train': 2.460857391357422} 02/24/2022 13:02:13 - INFO - codeparrot_training - Step 8863: {'lr': 0.0004752006181906052, 'samples': 4538368, 'steps': 8863, 'loss/train': 4.379079341888428} 02/24/2022 13:02:16 - INFO - codeparrot_training - Step 8864: {'lr': 0.00047519351265134954, 'samples': 4538880, 'steps': 8864, 'loss/train': 1.812307357788086} 02/24/2022 13:02:23 - INFO - codeparrot_training - Step 8865: {'lr': 0.0004751864061474364, 'samples': 4539392, 'steps': 8865, 'loss/train': 2.5313527584075928} 02/24/2022 13:02:26 - INFO - codeparrot_training - Step 8866: {'lr': 0.000475179298678896, 'samples': 4539904, 'steps': 8866, 'loss/train': 1.605868935585022} 02/24/2022 13:02:32 - INFO - codeparrot_training - Step 8867: {'lr': 0.0004751721902457592, 'samples': 4540416, 'steps': 8867, 'loss/train': 2.56888484954834} 02/24/2022 13:02:35 - INFO - codeparrot_training - Step 8868: {'lr': 0.0004751650808480561, 'samples': 4540928, 'steps': 8868, 'loss/train': 2.6334800720214844} 02/24/2022 13:02:41 - INFO - codeparrot_training - Step 8869: {'lr': 0.00047515797048581734, 'samples': 4541440, 'steps': 8869, 'loss/train': 2.85691499710083} 02/24/2022 13:02:44 - INFO - codeparrot_training - Step 8870: {'lr': 0.00047515085915907334, 'samples': 4541952, 'steps': 8870, 'loss/train': 1.6506918668746948} 02/24/2022 13:02:50 - INFO - codeparrot_training - Step 8871: {'lr': 0.00047514374686785454, 'samples': 4542464, 'steps': 8871, 'loss/train': 2.445449113845825} 02/24/2022 13:02:53 - INFO - codeparrot_training - Step 8872: {'lr': 0.00047513663361219144, 'samples': 4542976, 'steps': 8872, 'loss/train': 2.0163402557373047} 02/24/2022 13:02:59 - INFO - codeparrot_training - Step 8873: {'lr': 0.00047512951939211447, 'samples': 4543488, 'steps': 8873, 'loss/train': 2.0149426460266113} 02/24/2022 13:03:02 - INFO - codeparrot_training - Step 8874: {'lr': 0.0004751224042076542, 'samples': 4544000, 'steps': 8874, 'loss/train': 2.9443156719207764} 02/24/2022 13:03:08 - INFO - codeparrot_training - Step 8875: {'lr': 0.0004751152880588409, 'samples': 4544512, 'steps': 8875, 'loss/train': 2.747853994369507} 02/24/2022 13:03:12 - INFO - codeparrot_training - Step 8876: {'lr': 0.00047510817094570526, 'samples': 4545024, 'steps': 8876, 'loss/train': 1.9711774587631226} 02/24/2022 13:03:17 - INFO - codeparrot_training - Step 8877: {'lr': 0.0004751010528682777, 'samples': 4545536, 'steps': 8877, 'loss/train': 2.7659058570861816} 02/24/2022 13:03:21 - INFO - codeparrot_training - Step 8878: {'lr': 0.0004750939338265887, 'samples': 4546048, 'steps': 8878, 'loss/train': 2.2706782817840576} 02/24/2022 13:03:26 - INFO - codeparrot_training - Step 8879: {'lr': 0.0004750868138206688, 'samples': 4546560, 'steps': 8879, 'loss/train': 1.5981159210205078} 02/24/2022 13:03:30 - INFO - codeparrot_training - Step 8880: {'lr': 0.0004750796928505484, 'samples': 4547072, 'steps': 8880, 'loss/train': 2.0803921222686768} 02/24/2022 13:03:35 - INFO - codeparrot_training - Step 8881: {'lr': 0.0004750725709162581, 'samples': 4547584, 'steps': 8881, 'loss/train': 1.8010421991348267} 02/24/2022 13:03:39 - INFO - codeparrot_training - Step 8882: {'lr': 0.00047506544801782834, 'samples': 4548096, 'steps': 8882, 'loss/train': 2.5815365314483643} 02/24/2022 13:03:44 - INFO - codeparrot_training - Step 8883: {'lr': 0.00047505832415528973, 'samples': 4548608, 'steps': 8883, 'loss/train': 2.71773362159729} 02/24/2022 13:03:48 - INFO - codeparrot_training - Step 8884: {'lr': 0.0004750511993286727, 'samples': 4549120, 'steps': 8884, 'loss/train': 0.18191629648208618} 02/24/2022 13:03:54 - INFO - codeparrot_training - Step 8885: {'lr': 0.0004750440735380077, 'samples': 4549632, 'steps': 8885, 'loss/train': 2.0170774459838867} 02/24/2022 13:03:57 - INFO - codeparrot_training - Step 8886: {'lr': 0.00047503694678332543, 'samples': 4550144, 'steps': 8886, 'loss/train': 2.072664499282837} 02/24/2022 13:04:03 - INFO - codeparrot_training - Step 8887: {'lr': 0.00047502981906465634, 'samples': 4550656, 'steps': 8887, 'loss/train': 2.4431796073913574} 02/24/2022 13:04:06 - INFO - codeparrot_training - Step 8888: {'lr': 0.000475022690382031, 'samples': 4551168, 'steps': 8888, 'loss/train': 1.0116039514541626} 02/24/2022 13:04:12 - INFO - codeparrot_training - Step 8889: {'lr': 0.0004750155607354799, 'samples': 4551680, 'steps': 8889, 'loss/train': 2.342756986618042} 02/24/2022 13:04:15 - INFO - codeparrot_training - Step 8890: {'lr': 0.0004750084301250335, 'samples': 4552192, 'steps': 8890, 'loss/train': 2.442150831222534} 02/24/2022 13:04:22 - INFO - codeparrot_training - Step 8891: {'lr': 0.0004750012985507225, 'samples': 4552704, 'steps': 8891, 'loss/train': 2.6814463138580322} 02/24/2022 13:04:25 - INFO - codeparrot_training - Step 8892: {'lr': 0.0004749941660125774, 'samples': 4553216, 'steps': 8892, 'loss/train': 1.5645052194595337} 02/24/2022 13:04:31 - INFO - codeparrot_training - Step 8893: {'lr': 0.0004749870325106287, 'samples': 4553728, 'steps': 8893, 'loss/train': 2.6697940826416016} 02/24/2022 13:04:34 - INFO - codeparrot_training - Step 8894: {'lr': 0.00047497989804490693, 'samples': 4554240, 'steps': 8894, 'loss/train': 2.2933146953582764} 02/24/2022 13:04:40 - INFO - codeparrot_training - Step 8895: {'lr': 0.0004749727626154428, 'samples': 4554752, 'steps': 8895, 'loss/train': 2.0006628036499023} 02/24/2022 13:04:43 - INFO - codeparrot_training - Step 8896: {'lr': 0.0004749656262222668, 'samples': 4555264, 'steps': 8896, 'loss/train': 2.4212472438812256} 02/24/2022 13:04:49 - INFO - codeparrot_training - Step 8897: {'lr': 0.0004749584888654095, 'samples': 4555776, 'steps': 8897, 'loss/train': 2.1934757232666016} 02/24/2022 13:04:52 - INFO - codeparrot_training - Step 8898: {'lr': 0.0004749513505449014, 'samples': 4556288, 'steps': 8898, 'loss/train': 2.125549793243408} 02/24/2022 13:04:58 - INFO - codeparrot_training - Step 8899: {'lr': 0.00047494421126077313, 'samples': 4556800, 'steps': 8899, 'loss/train': 2.5988833904266357} 02/24/2022 13:05:01 - INFO - codeparrot_training - Step 8900: {'lr': 0.0004749370710130554, 'samples': 4557312, 'steps': 8900, 'loss/train': 2.2430381774902344} 02/24/2022 13:05:08 - INFO - codeparrot_training - Step 8901: {'lr': 0.0004749299298017786, 'samples': 4557824, 'steps': 8901, 'loss/train': 1.6011979579925537} 02/24/2022 13:05:11 - INFO - codeparrot_training - Step 8902: {'lr': 0.00047492278762697337, 'samples': 4558336, 'steps': 8902, 'loss/train': 2.3110179901123047} 02/24/2022 13:05:17 - INFO - codeparrot_training - Step 8903: {'lr': 0.0004749156444886704, 'samples': 4558848, 'steps': 8903, 'loss/train': 3.0487117767333984} 02/24/2022 13:05:20 - INFO - codeparrot_training - Step 8904: {'lr': 0.0004749085003869003, 'samples': 4559360, 'steps': 8904, 'loss/train': 0.6672154664993286} 02/24/2022 13:05:26 - INFO - codeparrot_training - Step 8905: {'lr': 0.00047490135532169347, 'samples': 4559872, 'steps': 8905, 'loss/train': 2.1986923217773438} 02/24/2022 13:05:29 - INFO - codeparrot_training - Step 8906: {'lr': 0.0004748942092930807, 'samples': 4560384, 'steps': 8906, 'loss/train': 2.151268720626831} 02/24/2022 13:05:35 - INFO - codeparrot_training - Step 8907: {'lr': 0.00047488706230109257, 'samples': 4560896, 'steps': 8907, 'loss/train': 2.0609793663024902} 02/24/2022 13:05:38 - INFO - codeparrot_training - Step 8908: {'lr': 0.00047487991434575963, 'samples': 4561408, 'steps': 8908, 'loss/train': 2.131075382232666} 02/24/2022 13:05:44 - INFO - codeparrot_training - Step 8909: {'lr': 0.0004748727654271126, 'samples': 4561920, 'steps': 8909, 'loss/train': 2.5970921516418457} 02/24/2022 13:05:47 - INFO - codeparrot_training - Step 8910: {'lr': 0.000474865615545182, 'samples': 4562432, 'steps': 8910, 'loss/train': 2.234992027282715} 02/24/2022 13:05:53 - INFO - codeparrot_training - Step 8911: {'lr': 0.0004748584646999985, 'samples': 4562944, 'steps': 8911, 'loss/train': 2.191291332244873} 02/24/2022 13:05:57 - INFO - codeparrot_training - Step 8912: {'lr': 0.0004748513128915928, 'samples': 4563456, 'steps': 8912, 'loss/train': 1.9498353004455566} 02/24/2022 13:06:02 - INFO - codeparrot_training - Step 8913: {'lr': 0.0004748441601199954, 'samples': 4563968, 'steps': 8913, 'loss/train': 2.004565715789795} 02/24/2022 13:06:06 - INFO - codeparrot_training - Step 8914: {'lr': 0.0004748370063852371, 'samples': 4564480, 'steps': 8914, 'loss/train': 3.04158353805542} 02/24/2022 13:06:11 - INFO - codeparrot_training - Step 8915: {'lr': 0.0004748298516873484, 'samples': 4564992, 'steps': 8915, 'loss/train': 2.4030370712280273} 02/24/2022 13:06:15 - INFO - codeparrot_training - Step 8916: {'lr': 0.00047482269602636, 'samples': 4565504, 'steps': 8916, 'loss/train': 2.060742139816284} 02/24/2022 13:06:20 - INFO - codeparrot_training - Step 8917: {'lr': 0.00047481553940230257, 'samples': 4566016, 'steps': 8917, 'loss/train': 2.0784223079681396} 02/24/2022 13:06:24 - INFO - codeparrot_training - Step 8918: {'lr': 0.0004748083818152067, 'samples': 4566528, 'steps': 8918, 'loss/train': 2.1121463775634766} 02/24/2022 13:06:29 - INFO - codeparrot_training - Step 8919: {'lr': 0.00047480122326510325, 'samples': 4567040, 'steps': 8919, 'loss/train': 2.3299055099487305} 02/24/2022 13:06:33 - INFO - codeparrot_training - Step 8920: {'lr': 0.0004747940637520226, 'samples': 4567552, 'steps': 8920, 'loss/train': 1.8456697463989258} 02/24/2022 13:06:39 - INFO - codeparrot_training - Step 8921: {'lr': 0.0004747869032759956, 'samples': 4568064, 'steps': 8921, 'loss/train': 3.1344287395477295} 02/24/2022 13:06:42 - INFO - codeparrot_training - Step 8922: {'lr': 0.00047477974183705293, 'samples': 4568576, 'steps': 8922, 'loss/train': 2.278343439102173} 02/24/2022 13:06:48 - INFO - codeparrot_training - Step 8923: {'lr': 0.0004747725794352252, 'samples': 4569088, 'steps': 8923, 'loss/train': 2.763631582260132} 02/24/2022 13:06:52 - INFO - codeparrot_training - Step 8924: {'lr': 0.00047476541607054313, 'samples': 4569600, 'steps': 8924, 'loss/train': 3.3006646633148193} 02/24/2022 13:06:57 - INFO - codeparrot_training - Step 8925: {'lr': 0.0004747582517430373, 'samples': 4570112, 'steps': 8925, 'loss/train': 1.9184359312057495} 02/24/2022 13:07:01 - INFO - codeparrot_training - Step 8926: {'lr': 0.00047475108645273856, 'samples': 4570624, 'steps': 8926, 'loss/train': 1.8225644826889038} 02/24/2022 13:07:06 - INFO - codeparrot_training - Step 8927: {'lr': 0.00047474392019967754, 'samples': 4571136, 'steps': 8927, 'loss/train': 2.553635358810425} 02/24/2022 13:07:10 - INFO - codeparrot_training - Step 8928: {'lr': 0.0004747367529838849, 'samples': 4571648, 'steps': 8928, 'loss/train': 2.4356164932250977} 02/24/2022 13:07:15 - INFO - codeparrot_training - Step 8929: {'lr': 0.0004747295848053914, 'samples': 4572160, 'steps': 8929, 'loss/train': 1.031925082206726} 02/24/2022 13:07:18 - INFO - codeparrot_training - Step 8930: {'lr': 0.0004747224156642277, 'samples': 4572672, 'steps': 8930, 'loss/train': 2.750985860824585} 02/24/2022 13:07:24 - INFO - codeparrot_training - Step 8931: {'lr': 0.00047471524556042454, 'samples': 4573184, 'steps': 8931, 'loss/train': 1.6292858123779297} 02/24/2022 13:07:27 - INFO - codeparrot_training - Step 8932: {'lr': 0.00047470807449401264, 'samples': 4573696, 'steps': 8932, 'loss/train': 2.339555263519287} 02/24/2022 13:07:34 - INFO - codeparrot_training - Step 8933: {'lr': 0.0004747009024650227, 'samples': 4574208, 'steps': 8933, 'loss/train': 2.4546921253204346} 02/24/2022 13:07:37 - INFO - codeparrot_training - Step 8934: {'lr': 0.00047469372947348546, 'samples': 4574720, 'steps': 8934, 'loss/train': 2.2428476810455322} 02/24/2022 13:07:43 - INFO - codeparrot_training - Step 8935: {'lr': 0.0004746865555194315, 'samples': 4575232, 'steps': 8935, 'loss/train': 1.9881370067596436} 02/24/2022 13:07:46 - INFO - codeparrot_training - Step 8936: {'lr': 0.00047467938060289185, 'samples': 4575744, 'steps': 8936, 'loss/train': 0.6411017179489136} 02/24/2022 13:07:52 - INFO - codeparrot_training - Step 8937: {'lr': 0.00047467220472389694, 'samples': 4576256, 'steps': 8937, 'loss/train': 2.5050201416015625} 02/24/2022 13:07:56 - INFO - codeparrot_training - Step 8938: {'lr': 0.0004746650278824777, 'samples': 4576768, 'steps': 8938, 'loss/train': 2.3526217937469482} 02/24/2022 13:08:01 - INFO - codeparrot_training - Step 8939: {'lr': 0.00047465785007866487, 'samples': 4577280, 'steps': 8939, 'loss/train': 2.139683246612549} 02/24/2022 13:08:05 - INFO - codeparrot_training - Step 8940: {'lr': 0.00047465067131248907, 'samples': 4577792, 'steps': 8940, 'loss/train': 3.1311049461364746} 02/24/2022 13:08:08 - INFO - codeparrot_training - Step 8941: {'lr': 0.0004746434915839812, 'samples': 4578304, 'steps': 8941, 'loss/train': 2.407386302947998} 02/24/2022 13:08:14 - INFO - codeparrot_training - Step 8942: {'lr': 0.00047463631089317195, 'samples': 4578816, 'steps': 8942, 'loss/train': 1.269031047821045} 02/24/2022 13:08:19 - INFO - codeparrot_training - Step 8943: {'lr': 0.000474629129240092, 'samples': 4579328, 'steps': 8943, 'loss/train': 1.9832195043563843} 02/24/2022 13:08:23 - INFO - codeparrot_training - Step 8944: {'lr': 0.0004746219466247722, 'samples': 4579840, 'steps': 8944, 'loss/train': 2.5644195079803467} 02/24/2022 13:08:28 - INFO - codeparrot_training - Step 8945: {'lr': 0.0004746147630472434, 'samples': 4580352, 'steps': 8945, 'loss/train': 4.009216785430908} 02/24/2022 13:08:32 - INFO - codeparrot_training - Step 8946: {'lr': 0.00047460757850753614, 'samples': 4580864, 'steps': 8946, 'loss/train': 2.1093924045562744} 02/24/2022 13:08:38 - INFO - codeparrot_training - Step 8947: {'lr': 0.00047460039300568143, 'samples': 4581376, 'steps': 8947, 'loss/train': 1.8118352890014648} 02/24/2022 13:08:41 - INFO - codeparrot_training - Step 8948: {'lr': 0.0004745932065417099, 'samples': 4581888, 'steps': 8948, 'loss/train': 5.964253902435303} 02/24/2022 13:08:45 - INFO - codeparrot_training - Step 8949: {'lr': 0.00047458601911565246, 'samples': 4582400, 'steps': 8949, 'loss/train': 2.5903995037078857} 02/24/2022 13:08:51 - INFO - codeparrot_training - Step 8950: {'lr': 0.0004745788307275398, 'samples': 4582912, 'steps': 8950, 'loss/train': 2.4588711261749268} 02/24/2022 13:08:54 - INFO - codeparrot_training - Step 8951: {'lr': 0.0004745716413774027, 'samples': 4583424, 'steps': 8951, 'loss/train': 1.8251460790634155} 02/24/2022 13:09:00 - INFO - codeparrot_training - Step 8952: {'lr': 0.000474564451065272, 'samples': 4583936, 'steps': 8952, 'loss/train': 1.827713966369629} 02/24/2022 13:09:04 - INFO - codeparrot_training - Step 8953: {'lr': 0.00047455725979117855, 'samples': 4584448, 'steps': 8953, 'loss/train': 1.045904517173767} 02/24/2022 13:09:09 - INFO - codeparrot_training - Step 8954: {'lr': 0.00047455006755515306, 'samples': 4584960, 'steps': 8954, 'loss/train': 2.504101276397705} 02/24/2022 13:09:13 - INFO - codeparrot_training - Step 8955: {'lr': 0.00047454287435722643, 'samples': 4585472, 'steps': 8955, 'loss/train': 2.354496955871582} 02/24/2022 13:09:18 - INFO - codeparrot_training - Step 8956: {'lr': 0.00047453568019742936, 'samples': 4585984, 'steps': 8956, 'loss/train': 3.8351099491119385} 02/24/2022 13:09:22 - INFO - codeparrot_training - Step 8957: {'lr': 0.0004745284850757928, 'samples': 4586496, 'steps': 8957, 'loss/train': 2.2186691761016846} 02/24/2022 13:09:28 - INFO - codeparrot_training - Step 8958: {'lr': 0.00047452128899234746, 'samples': 4587008, 'steps': 8958, 'loss/train': 2.7836594581604004} 02/24/2022 13:09:31 - INFO - codeparrot_training - Step 8959: {'lr': 0.0004745140919471243, 'samples': 4587520, 'steps': 8959, 'loss/train': 3.1200828552246094} 02/24/2022 13:09:37 - INFO - codeparrot_training - Step 8960: {'lr': 0.0004745068939401539, 'samples': 4588032, 'steps': 8960, 'loss/train': 0.8317255973815918} 02/24/2022 13:09:40 - INFO - codeparrot_training - Step 8961: {'lr': 0.0004744996949714674, 'samples': 4588544, 'steps': 8961, 'loss/train': 3.034991979598999} 02/24/2022 13:09:46 - INFO - codeparrot_training - Step 8962: {'lr': 0.0004744924950410954, 'samples': 4589056, 'steps': 8962, 'loss/train': 1.7224841117858887} 02/24/2022 13:09:49 - INFO - codeparrot_training - Step 8963: {'lr': 0.0004744852941490689, 'samples': 4589568, 'steps': 8963, 'loss/train': 2.6999459266662598} 02/24/2022 13:09:55 - INFO - codeparrot_training - Step 8964: {'lr': 0.0004744780922954186, 'samples': 4590080, 'steps': 8964, 'loss/train': 2.097709894180298} 02/24/2022 13:09:59 - INFO - codeparrot_training - Step 8965: {'lr': 0.00047447088948017555, 'samples': 4590592, 'steps': 8965, 'loss/train': 2.6860897541046143} 02/24/2022 13:10:04 - INFO - codeparrot_training - Step 8966: {'lr': 0.0004744636857033704, 'samples': 4591104, 'steps': 8966, 'loss/train': 3.8114237785339355} 02/24/2022 13:10:08 - INFO - codeparrot_training - Step 8967: {'lr': 0.00047445648096503413, 'samples': 4591616, 'steps': 8967, 'loss/train': 2.8054850101470947} 02/24/2022 13:10:11 - INFO - codeparrot_training - Step 8968: {'lr': 0.00047444927526519757, 'samples': 4592128, 'steps': 8968, 'loss/train': 2.804905414581299} 02/24/2022 13:10:17 - INFO - codeparrot_training - Step 8969: {'lr': 0.00047444206860389155, 'samples': 4592640, 'steps': 8969, 'loss/train': 2.3340671062469482} 02/24/2022 13:10:23 - INFO - codeparrot_training - Step 8970: {'lr': 0.00047443486098114703, 'samples': 4593152, 'steps': 8970, 'loss/train': 1.1994097232818604} 02/24/2022 13:10:26 - INFO - codeparrot_training - Step 8971: {'lr': 0.0004744276523969948, 'samples': 4593664, 'steps': 8971, 'loss/train': 2.310925006866455} 02/24/2022 13:10:32 - INFO - codeparrot_training - Step 8972: {'lr': 0.0004744204428514658, 'samples': 4594176, 'steps': 8972, 'loss/train': 1.7849228382110596} 02/24/2022 13:10:35 - INFO - codeparrot_training - Step 8973: {'lr': 0.0004744132323445908, 'samples': 4594688, 'steps': 8973, 'loss/train': 1.676094651222229} 02/24/2022 13:10:41 - INFO - codeparrot_training - Step 8974: {'lr': 0.00047440602087640084, 'samples': 4595200, 'steps': 8974, 'loss/train': 2.3776402473449707} 02/24/2022 13:10:44 - INFO - codeparrot_training - Step 8975: {'lr': 0.0004743988084469267, 'samples': 4595712, 'steps': 8975, 'loss/train': 2.1946041584014893} 02/24/2022 13:10:50 - INFO - codeparrot_training - Step 8976: {'lr': 0.00047439159505619936, 'samples': 4596224, 'steps': 8976, 'loss/train': 1.8234540224075317} 02/24/2022 13:10:53 - INFO - codeparrot_training - Step 8977: {'lr': 0.0004743843807042497, 'samples': 4596736, 'steps': 8977, 'loss/train': 2.3686439990997314} 02/24/2022 13:10:59 - INFO - codeparrot_training - Step 8978: {'lr': 0.0004743771653911086, 'samples': 4597248, 'steps': 8978, 'loss/train': 2.1785380840301514} 02/24/2022 13:11:02 - INFO - codeparrot_training - Step 8979: {'lr': 0.00047436994911680694, 'samples': 4597760, 'steps': 8979, 'loss/train': 2.5101404190063477} 02/24/2022 13:11:08 - INFO - codeparrot_training - Step 8980: {'lr': 0.0004743627318813757, 'samples': 4598272, 'steps': 8980, 'loss/train': 1.986647129058838} 02/24/2022 13:11:11 - INFO - codeparrot_training - Step 8981: {'lr': 0.00047435551368484567, 'samples': 4598784, 'steps': 8981, 'loss/train': 1.4231780767440796} 02/24/2022 13:11:17 - INFO - codeparrot_training - Step 8982: {'lr': 0.00047434829452724795, 'samples': 4599296, 'steps': 8982, 'loss/train': 3.5546858310699463} 02/24/2022 13:11:20 - INFO - codeparrot_training - Step 8983: {'lr': 0.00047434107440861336, 'samples': 4599808, 'steps': 8983, 'loss/train': 0.21147406101226807} 02/24/2022 13:11:26 - INFO - codeparrot_training - Step 8984: {'lr': 0.0004743338533289728, 'samples': 4600320, 'steps': 8984, 'loss/train': 2.249940872192383} 02/24/2022 13:11:30 - INFO - codeparrot_training - Step 8985: {'lr': 0.00047432663128835727, 'samples': 4600832, 'steps': 8985, 'loss/train': 1.261738896369934} 02/24/2022 13:11:35 - INFO - codeparrot_training - Step 8986: {'lr': 0.0004743194082867977, 'samples': 4601344, 'steps': 8986, 'loss/train': 2.7038543224334717} 02/24/2022 13:11:39 - INFO - codeparrot_training - Step 8987: {'lr': 0.000474312184324325, 'samples': 4601856, 'steps': 8987, 'loss/train': 2.480295419692993} 02/24/2022 13:11:44 - INFO - codeparrot_training - Step 8988: {'lr': 0.0004743049594009701, 'samples': 4602368, 'steps': 8988, 'loss/train': 1.866456151008606} 02/24/2022 13:11:48 - INFO - codeparrot_training - Step 8989: {'lr': 0.0004742977335167641, 'samples': 4602880, 'steps': 8989, 'loss/train': 2.4427878856658936} 02/24/2022 13:11:53 - INFO - codeparrot_training - Step 8990: {'lr': 0.0004742905066717377, 'samples': 4603392, 'steps': 8990, 'loss/train': 2.4546449184417725} 02/24/2022 13:11:57 - INFO - codeparrot_training - Step 8991: {'lr': 0.00047428327886592204, 'samples': 4603904, 'steps': 8991, 'loss/train': 2.667329788208008} 02/24/2022 13:12:02 - INFO - codeparrot_training - Step 8992: {'lr': 0.00047427605009934805, 'samples': 4604416, 'steps': 8992, 'loss/train': 2.1340625286102295} 02/24/2022 13:12:06 - INFO - codeparrot_training - Step 8993: {'lr': 0.00047426882037204663, 'samples': 4604928, 'steps': 8993, 'loss/train': 2.5201327800750732} 02/24/2022 13:12:12 - INFO - codeparrot_training - Step 8994: {'lr': 0.0004742615896840488, 'samples': 4605440, 'steps': 8994, 'loss/train': 2.1459014415740967} 02/24/2022 13:12:16 - INFO - codeparrot_training - Step 8995: {'lr': 0.00047425435803538554, 'samples': 4605952, 'steps': 8995, 'loss/train': 2.3032312393188477} 02/24/2022 13:12:21 - INFO - codeparrot_training - Step 8996: {'lr': 0.0004742471254260878, 'samples': 4606464, 'steps': 8996, 'loss/train': 1.379835844039917} 02/24/2022 13:12:24 - INFO - codeparrot_training - Step 8997: {'lr': 0.00047423989185618666, 'samples': 4606976, 'steps': 8997, 'loss/train': 2.6440694332122803} 02/24/2022 13:12:30 - INFO - codeparrot_training - Step 8998: {'lr': 0.00047423265732571295, 'samples': 4607488, 'steps': 8998, 'loss/train': 2.1104156970977783} 02/24/2022 13:12:34 - INFO - codeparrot_training - Step 8999: {'lr': 0.00047422542183469775, 'samples': 4608000, 'steps': 8999, 'loss/train': 3.3685567378997803} 02/24/2022 13:12:34 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 13:12:51 - WARNING - huggingface_hub.repository - Several commits (9) will be pushed upstream. 02/24/2022 13:12:51 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 13:13:25 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 869a374..8cd23e2 floral-grass-11 -> floral-grass-11 02/24/2022 13:13:31 - INFO - codeparrot_training - Step 9000: {'lr': 0.0004742181853831721, 'samples': 4608512, 'steps': 9000, 'loss/train': 2.7164244651794434} 02/24/2022 13:13:35 - INFO - codeparrot_training - Step 9001: {'lr': 0.00047421094797116687, 'samples': 4609024, 'steps': 9001, 'loss/train': 2.5897510051727295} 02/24/2022 13:13:40 - INFO - codeparrot_training - Step 9002: {'lr': 0.00047420370959871315, 'samples': 4609536, 'steps': 9002, 'loss/train': 2.2958545684814453} 02/24/2022 13:13:44 - INFO - codeparrot_training - Step 9003: {'lr': 0.000474196470265842, 'samples': 4610048, 'steps': 9003, 'loss/train': 2.6940038204193115} 02/24/2022 13:13:49 - INFO - codeparrot_training - Step 9004: {'lr': 0.0004741892299725843, 'samples': 4610560, 'steps': 9004, 'loss/train': 1.2317273616790771} 02/24/2022 13:13:53 - INFO - codeparrot_training - Step 9005: {'lr': 0.0004741819887189711, 'samples': 4611072, 'steps': 9005, 'loss/train': 2.348780632019043} 02/24/2022 13:13:59 - INFO - codeparrot_training - Step 9006: {'lr': 0.00047417474650503347, 'samples': 4611584, 'steps': 9006, 'loss/train': 1.9111452102661133} 02/24/2022 13:14:02 - INFO - codeparrot_training - Step 9007: {'lr': 0.00047416750333080244, 'samples': 4612096, 'steps': 9007, 'loss/train': 1.8345046043395996} 02/24/2022 13:14:08 - INFO - codeparrot_training - Step 9008: {'lr': 0.000474160259196309, 'samples': 4612608, 'steps': 9008, 'loss/train': 2.2105023860931396} 02/24/2022 13:14:11 - INFO - codeparrot_training - Step 9009: {'lr': 0.00047415301410158416, 'samples': 4613120, 'steps': 9009, 'loss/train': 2.2949531078338623} 02/24/2022 13:14:17 - INFO - codeparrot_training - Step 9010: {'lr': 0.00047414576804665897, 'samples': 4613632, 'steps': 9010, 'loss/train': 2.5287442207336426} 02/24/2022 13:14:20 - INFO - codeparrot_training - Step 9011: {'lr': 0.0004741385210315645, 'samples': 4614144, 'steps': 9011, 'loss/train': 3.376943826675415} 02/24/2022 13:14:26 - INFO - codeparrot_training - Step 9012: {'lr': 0.0004741312730563318, 'samples': 4614656, 'steps': 9012, 'loss/train': 2.552826404571533} 02/24/2022 13:14:29 - INFO - codeparrot_training - Step 9013: {'lr': 0.00047412402412099185, 'samples': 4615168, 'steps': 9013, 'loss/train': 1.4551596641540527} 02/24/2022 13:14:35 - INFO - codeparrot_training - Step 9014: {'lr': 0.00047411677422557586, 'samples': 4615680, 'steps': 9014, 'loss/train': 2.709442377090454} 02/24/2022 13:14:39 - INFO - codeparrot_training - Step 9015: {'lr': 0.0004741095233701147, 'samples': 4616192, 'steps': 9015, 'loss/train': 2.6846683025360107} 02/24/2022 13:14:44 - INFO - codeparrot_training - Step 9016: {'lr': 0.00047410227155463946, 'samples': 4616704, 'steps': 9016, 'loss/train': 1.409234642982483} 02/24/2022 13:14:48 - INFO - codeparrot_training - Step 9017: {'lr': 0.00047409501877918134, 'samples': 4617216, 'steps': 9017, 'loss/train': 2.198273181915283} 02/24/2022 13:14:53 - INFO - codeparrot_training - Step 9018: {'lr': 0.00047408776504377127, 'samples': 4617728, 'steps': 9018, 'loss/train': 3.9294626712799072} 02/24/2022 13:14:57 - INFO - codeparrot_training - Step 9019: {'lr': 0.00047408051034844036, 'samples': 4618240, 'steps': 9019, 'loss/train': 1.6127426624298096} 02/24/2022 13:15:03 - INFO - codeparrot_training - Step 9020: {'lr': 0.00047407325469321973, 'samples': 4618752, 'steps': 9020, 'loss/train': 1.723608374595642} 02/24/2022 13:15:06 - INFO - codeparrot_training - Step 9021: {'lr': 0.00047406599807814034, 'samples': 4619264, 'steps': 9021, 'loss/train': 1.8285869359970093} 02/24/2022 13:15:12 - INFO - codeparrot_training - Step 9022: {'lr': 0.00047405874050323346, 'samples': 4619776, 'steps': 9022, 'loss/train': 2.238917112350464} 02/24/2022 13:15:15 - INFO - codeparrot_training - Step 9023: {'lr': 0.00047405148196853005, 'samples': 4620288, 'steps': 9023, 'loss/train': 2.512418270111084} 02/24/2022 13:15:21 - INFO - codeparrot_training - Step 9024: {'lr': 0.0004740442224740612, 'samples': 4620800, 'steps': 9024, 'loss/train': 2.4134669303894043} 02/24/2022 13:15:24 - INFO - codeparrot_training - Step 9025: {'lr': 0.00047403696201985814, 'samples': 4621312, 'steps': 9025, 'loss/train': 2.0683298110961914} 02/24/2022 13:15:30 - INFO - codeparrot_training - Step 9026: {'lr': 0.0004740297006059517, 'samples': 4621824, 'steps': 9026, 'loss/train': 4.153469085693359} 02/24/2022 13:15:33 - INFO - codeparrot_training - Step 9027: {'lr': 0.00047402243823237335, 'samples': 4622336, 'steps': 9027, 'loss/train': 1.373235821723938} 02/24/2022 13:15:39 - INFO - codeparrot_training - Step 9028: {'lr': 0.0004740151748991539, 'samples': 4622848, 'steps': 9028, 'loss/train': 2.228574752807617} 02/24/2022 13:15:42 - INFO - codeparrot_training - Step 9029: {'lr': 0.00047400791060632464, 'samples': 4623360, 'steps': 9029, 'loss/train': 2.196526050567627} 02/24/2022 13:15:48 - INFO - codeparrot_training - Step 9030: {'lr': 0.0004740006453539166, 'samples': 4623872, 'steps': 9030, 'loss/train': 1.4205386638641357} 02/24/2022 13:15:55 - INFO - codeparrot_training - Step 9031: {'lr': 0.0004739933791419609, 'samples': 4624384, 'steps': 9031, 'loss/train': 2.9199771881103516} 02/24/2022 13:15:58 - INFO - codeparrot_training - Step 9032: {'lr': 0.0004739861119704887, 'samples': 4624896, 'steps': 9032, 'loss/train': 3.2379813194274902} 02/24/2022 13:16:04 - INFO - codeparrot_training - Step 9033: {'lr': 0.00047397884383953114, 'samples': 4625408, 'steps': 9033, 'loss/train': 2.1234307289123535} 02/24/2022 13:16:07 - INFO - codeparrot_training - Step 9034: {'lr': 0.0004739715747491193, 'samples': 4625920, 'steps': 9034, 'loss/train': 0.5818557739257812} 02/24/2022 13:16:13 - INFO - codeparrot_training - Step 9035: {'lr': 0.00047396430469928436, 'samples': 4626432, 'steps': 9035, 'loss/train': 1.9213566780090332} 02/24/2022 13:16:16 - INFO - codeparrot_training - Step 9036: {'lr': 0.0004739570336900575, 'samples': 4626944, 'steps': 9036, 'loss/train': 1.8206393718719482} 02/24/2022 13:16:22 - INFO - codeparrot_training - Step 9037: {'lr': 0.00047394976172146974, 'samples': 4627456, 'steps': 9037, 'loss/train': 2.1901352405548096} 02/24/2022 13:16:25 - INFO - codeparrot_training - Step 9038: {'lr': 0.0004739424887935524, 'samples': 4627968, 'steps': 9038, 'loss/train': 2.1107277870178223} 02/24/2022 13:16:31 - INFO - codeparrot_training - Step 9039: {'lr': 0.0004739352149063365, 'samples': 4628480, 'steps': 9039, 'loss/train': 2.2411656379699707} 02/24/2022 13:16:34 - INFO - codeparrot_training - Step 9040: {'lr': 0.0004739279400598532, 'samples': 4628992, 'steps': 9040, 'loss/train': 1.7671642303466797} 02/24/2022 13:16:40 - INFO - codeparrot_training - Step 9041: {'lr': 0.0004739206642541338, 'samples': 4629504, 'steps': 9041, 'loss/train': 1.9934426546096802} 02/24/2022 13:16:44 - INFO - codeparrot_training - Step 9042: {'lr': 0.0004739133874892093, 'samples': 4630016, 'steps': 9042, 'loss/train': 1.3417733907699585} 02/24/2022 13:16:50 - INFO - codeparrot_training - Step 9043: {'lr': 0.0004739061097651111, 'samples': 4630528, 'steps': 9043, 'loss/train': 0.42076340317726135} 02/24/2022 13:16:53 - INFO - codeparrot_training - Step 9044: {'lr': 0.00047389883108187004, 'samples': 4631040, 'steps': 9044, 'loss/train': 2.7384181022644043} 02/24/2022 13:16:58 - INFO - codeparrot_training - Step 9045: {'lr': 0.0004738915514395176, 'samples': 4631552, 'steps': 9045, 'loss/train': 1.4291678667068481} 02/24/2022 13:17:02 - INFO - codeparrot_training - Step 9046: {'lr': 0.0004738842708380847, 'samples': 4632064, 'steps': 9046, 'loss/train': 2.880378484725952} 02/24/2022 13:17:07 - INFO - codeparrot_training - Step 9047: {'lr': 0.0004738769892776028, 'samples': 4632576, 'steps': 9047, 'loss/train': 1.7809022665023804} 02/24/2022 13:17:11 - INFO - codeparrot_training - Step 9048: {'lr': 0.00047386970675810297, 'samples': 4633088, 'steps': 9048, 'loss/train': 1.3483664989471436} 02/24/2022 13:17:16 - INFO - codeparrot_training - Step 9049: {'lr': 0.00047386242327961635, 'samples': 4633600, 'steps': 9049, 'loss/train': 1.6377965211868286} 02/24/2022 13:17:20 - INFO - codeparrot_training - Step 9050: {'lr': 0.0004738551388421742, 'samples': 4634112, 'steps': 9050, 'loss/train': 1.6631475687026978} 02/24/2022 13:17:26 - INFO - codeparrot_training - Step 9051: {'lr': 0.00047384785344580784, 'samples': 4634624, 'steps': 9051, 'loss/train': 2.241671323776245} 02/24/2022 13:17:30 - INFO - codeparrot_training - Step 9052: {'lr': 0.00047384056709054824, 'samples': 4635136, 'steps': 9052, 'loss/train': 2.9415793418884277} 02/24/2022 13:17:35 - INFO - codeparrot_training - Step 9053: {'lr': 0.0004738332797764267, 'samples': 4635648, 'steps': 9053, 'loss/train': 2.354640007019043} 02/24/2022 13:17:39 - INFO - codeparrot_training - Step 9054: {'lr': 0.0004738259915034745, 'samples': 4636160, 'steps': 9054, 'loss/train': 1.0778690576553345} 02/24/2022 13:17:44 - INFO - codeparrot_training - Step 9055: {'lr': 0.00047381870227172285, 'samples': 4636672, 'steps': 9055, 'loss/train': 1.864437460899353} 02/24/2022 13:17:48 - INFO - codeparrot_training - Step 9056: {'lr': 0.0004738114120812029, 'samples': 4637184, 'steps': 9056, 'loss/train': 4.310451984405518} 02/24/2022 13:17:53 - INFO - codeparrot_training - Step 9057: {'lr': 0.000473804120931946, 'samples': 4637696, 'steps': 9057, 'loss/train': 1.8198201656341553} 02/24/2022 13:17:57 - INFO - codeparrot_training - Step 9058: {'lr': 0.0004737968288239832, 'samples': 4638208, 'steps': 9058, 'loss/train': 0.6944505572319031} 02/24/2022 13:18:02 - INFO - codeparrot_training - Step 9059: {'lr': 0.00047378953575734594, 'samples': 4638720, 'steps': 9059, 'loss/train': 2.7053723335266113} 02/24/2022 13:18:06 - INFO - codeparrot_training - Step 9060: {'lr': 0.0004737822417320654, 'samples': 4639232, 'steps': 9060, 'loss/train': 2.2772247791290283} 02/24/2022 13:18:11 - INFO - codeparrot_training - Step 9061: {'lr': 0.00047377494674817275, 'samples': 4639744, 'steps': 9061, 'loss/train': 1.6539815664291382} 02/24/2022 13:18:15 - INFO - codeparrot_training - Step 9062: {'lr': 0.00047376765080569925, 'samples': 4640256, 'steps': 9062, 'loss/train': 2.4359467029571533} 02/24/2022 13:18:21 - INFO - codeparrot_training - Step 9063: {'lr': 0.0004737603539046762, 'samples': 4640768, 'steps': 9063, 'loss/train': 2.046271800994873} 02/24/2022 13:18:24 - INFO - codeparrot_training - Step 9064: {'lr': 0.0004737530560451349, 'samples': 4641280, 'steps': 9064, 'loss/train': 2.2490618228912354} 02/24/2022 13:18:28 - INFO - codeparrot_training - Step 9065: {'lr': 0.00047374575722710656, 'samples': 4641792, 'steps': 9065, 'loss/train': 2.4609973430633545} 02/24/2022 13:18:33 - INFO - codeparrot_training - Step 9066: {'lr': 0.0004737384574506224, 'samples': 4642304, 'steps': 9066, 'loss/train': 1.7486547231674194} 02/24/2022 13:18:37 - INFO - codeparrot_training - Step 9067: {'lr': 0.0004737311567157137, 'samples': 4642816, 'steps': 9067, 'loss/train': 2.636962890625} 02/24/2022 13:18:43 - INFO - codeparrot_training - Step 9068: {'lr': 0.00047372385502241176, 'samples': 4643328, 'steps': 9068, 'loss/train': 2.1486315727233887} 02/24/2022 13:18:46 - INFO - codeparrot_training - Step 9069: {'lr': 0.00047371655237074794, 'samples': 4643840, 'steps': 9069, 'loss/train': 2.4533753395080566} 02/24/2022 13:18:52 - INFO - codeparrot_training - Step 9070: {'lr': 0.0004737092487607534, 'samples': 4644352, 'steps': 9070, 'loss/train': 2.2555012702941895} 02/24/2022 13:18:55 - INFO - codeparrot_training - Step 9071: {'lr': 0.00047370194419245955, 'samples': 4644864, 'steps': 9071, 'loss/train': 2.153978109359741} 02/24/2022 13:19:01 - INFO - codeparrot_training - Step 9072: {'lr': 0.00047369463866589755, 'samples': 4645376, 'steps': 9072, 'loss/train': 2.525999069213867} 02/24/2022 13:19:05 - INFO - codeparrot_training - Step 9073: {'lr': 0.00047368733218109874, 'samples': 4645888, 'steps': 9073, 'loss/train': 2.533388614654541} 02/24/2022 13:19:10 - INFO - codeparrot_training - Step 9074: {'lr': 0.00047368002473809447, 'samples': 4646400, 'steps': 9074, 'loss/train': 2.483051300048828} 02/24/2022 13:19:13 - INFO - codeparrot_training - Step 9075: {'lr': 0.0004736727163369159, 'samples': 4646912, 'steps': 9075, 'loss/train': 3.273833751678467} 02/24/2022 13:19:19 - INFO - codeparrot_training - Step 9076: {'lr': 0.00047366540697759454, 'samples': 4647424, 'steps': 9076, 'loss/train': 2.524005889892578} 02/24/2022 13:19:22 - INFO - codeparrot_training - Step 9077: {'lr': 0.00047365809666016155, 'samples': 4647936, 'steps': 9077, 'loss/train': 3.258361577987671} 02/24/2022 13:19:29 - INFO - codeparrot_training - Step 9078: {'lr': 0.00047365078538464826, 'samples': 4648448, 'steps': 9078, 'loss/train': 2.655883312225342} 02/24/2022 13:19:32 - INFO - codeparrot_training - Step 9079: {'lr': 0.0004736434731510861, 'samples': 4648960, 'steps': 9079, 'loss/train': 2.2690420150756836} 02/24/2022 13:19:38 - INFO - codeparrot_training - Step 9080: {'lr': 0.00047363615995950624, 'samples': 4649472, 'steps': 9080, 'loss/train': 2.4604878425598145} 02/24/2022 13:19:41 - INFO - codeparrot_training - Step 9081: {'lr': 0.0004736288458099401, 'samples': 4649984, 'steps': 9081, 'loss/train': 2.1250627040863037} 02/24/2022 13:19:47 - INFO - codeparrot_training - Step 9082: {'lr': 0.0004736215307024191, 'samples': 4650496, 'steps': 9082, 'loss/train': 2.5968241691589355} 02/24/2022 13:19:52 - INFO - codeparrot_training - Step 9083: {'lr': 0.0004736142146369744, 'samples': 4651008, 'steps': 9083, 'loss/train': 2.5302371978759766} 02/24/2022 13:19:56 - INFO - codeparrot_training - Step 9084: {'lr': 0.0004736068976136374, 'samples': 4651520, 'steps': 9084, 'loss/train': 2.263505220413208} 02/24/2022 13:20:01 - INFO - codeparrot_training - Step 9085: {'lr': 0.00047359957963243943, 'samples': 4652032, 'steps': 9085, 'loss/train': 1.1718926429748535} 02/24/2022 13:20:05 - INFO - codeparrot_training - Step 9086: {'lr': 0.0004735922606934119, 'samples': 4652544, 'steps': 9086, 'loss/train': 1.6612155437469482} 02/24/2022 13:20:11 - INFO - codeparrot_training - Step 9087: {'lr': 0.0004735849407965861, 'samples': 4653056, 'steps': 9087, 'loss/train': 1.924526333808899} 02/24/2022 13:20:15 - INFO - codeparrot_training - Step 9088: {'lr': 0.00047357761994199345, 'samples': 4653568, 'steps': 9088, 'loss/train': 1.666778564453125} 02/24/2022 13:20:20 - INFO - codeparrot_training - Step 9089: {'lr': 0.00047357029812966525, 'samples': 4654080, 'steps': 9089, 'loss/train': 2.3335912227630615} 02/24/2022 13:20:24 - INFO - codeparrot_training - Step 9090: {'lr': 0.0004735629753596328, 'samples': 4654592, 'steps': 9090, 'loss/train': 2.532266855239868} 02/24/2022 13:20:29 - INFO - codeparrot_training - Step 9091: {'lr': 0.00047355565163192763, 'samples': 4655104, 'steps': 9091, 'loss/train': 2.408735752105713} 02/24/2022 13:20:33 - INFO - codeparrot_training - Step 9092: {'lr': 0.00047354832694658104, 'samples': 4655616, 'steps': 9092, 'loss/train': 2.320756196975708} 02/24/2022 13:20:38 - INFO - codeparrot_training - Step 9093: {'lr': 0.00047354100130362443, 'samples': 4656128, 'steps': 9093, 'loss/train': 2.1825854778289795} 02/24/2022 13:20:42 - INFO - codeparrot_training - Step 9094: {'lr': 0.00047353367470308913, 'samples': 4656640, 'steps': 9094, 'loss/train': 1.8231948614120483} 02/24/2022 13:20:47 - INFO - codeparrot_training - Step 9095: {'lr': 0.0004735263471450065, 'samples': 4657152, 'steps': 9095, 'loss/train': 2.6250102519989014} 02/24/2022 13:20:51 - INFO - codeparrot_training - Step 9096: {'lr': 0.00047351901862940807, 'samples': 4657664, 'steps': 9096, 'loss/train': 2.6300764083862305} 02/24/2022 13:20:57 - INFO - codeparrot_training - Step 9097: {'lr': 0.000473511689156325, 'samples': 4658176, 'steps': 9097, 'loss/train': 1.944015622138977} 02/24/2022 13:21:01 - INFO - codeparrot_training - Step 9098: {'lr': 0.0004735043587257889, 'samples': 4658688, 'steps': 9098, 'loss/train': 1.8121219873428345} 02/24/2022 13:21:06 - INFO - codeparrot_training - Step 9099: {'lr': 0.00047349702733783113, 'samples': 4659200, 'steps': 9099, 'loss/train': 0.78632652759552} 02/24/2022 13:21:10 - INFO - codeparrot_training - Step 9100: {'lr': 0.00047348969499248306, 'samples': 4659712, 'steps': 9100, 'loss/train': 1.459047794342041} 02/24/2022 13:21:15 - INFO - codeparrot_training - Step 9101: {'lr': 0.0004734823616897761, 'samples': 4660224, 'steps': 9101, 'loss/train': 1.4093104600906372} 02/24/2022 13:21:19 - INFO - codeparrot_training - Step 9102: {'lr': 0.0004734750274297416, 'samples': 4660736, 'steps': 9102, 'loss/train': 1.9987393617630005} 02/24/2022 13:21:25 - INFO - codeparrot_training - Step 9103: {'lr': 0.0004734676922124111, 'samples': 4661248, 'steps': 9103, 'loss/train': 2.505825996398926} 02/24/2022 13:21:28 - INFO - codeparrot_training - Step 9104: {'lr': 0.00047346035603781597, 'samples': 4661760, 'steps': 9104, 'loss/train': 1.9600998163223267} 02/24/2022 13:21:34 - INFO - codeparrot_training - Step 9105: {'lr': 0.0004734530189059876, 'samples': 4662272, 'steps': 9105, 'loss/train': 2.631686210632324} 02/24/2022 13:21:37 - INFO - codeparrot_training - Step 9106: {'lr': 0.0004734456808169575, 'samples': 4662784, 'steps': 9106, 'loss/train': 1.004799485206604} 02/24/2022 13:21:43 - INFO - codeparrot_training - Step 9107: {'lr': 0.00047343834177075695, 'samples': 4663296, 'steps': 9107, 'loss/train': 2.3275928497314453} 02/24/2022 13:21:46 - INFO - codeparrot_training - Step 9108: {'lr': 0.0004734310017674176, 'samples': 4663808, 'steps': 9108, 'loss/train': 1.5682642459869385} 02/24/2022 13:21:52 - INFO - codeparrot_training - Step 9109: {'lr': 0.00047342366080697077, 'samples': 4664320, 'steps': 9109, 'loss/train': 2.061352014541626} 02/24/2022 13:21:55 - INFO - codeparrot_training - Step 9110: {'lr': 0.00047341631888944794, 'samples': 4664832, 'steps': 9110, 'loss/train': 1.8658485412597656} 02/24/2022 13:22:01 - INFO - codeparrot_training - Step 9111: {'lr': 0.0004734089760148805, 'samples': 4665344, 'steps': 9111, 'loss/train': 2.7326509952545166} 02/24/2022 13:22:04 - INFO - codeparrot_training - Step 9112: {'lr': 0.0004734016321832999, 'samples': 4665856, 'steps': 9112, 'loss/train': 2.248201608657837} 02/24/2022 13:22:10 - INFO - codeparrot_training - Step 9113: {'lr': 0.0004733942873947377, 'samples': 4666368, 'steps': 9113, 'loss/train': 1.2828941345214844} 02/24/2022 13:22:13 - INFO - codeparrot_training - Step 9114: {'lr': 0.00047338694164922535, 'samples': 4666880, 'steps': 9114, 'loss/train': 2.3066506385803223} 02/24/2022 13:22:19 - INFO - codeparrot_training - Step 9115: {'lr': 0.0004733795949467942, 'samples': 4667392, 'steps': 9115, 'loss/train': 2.111845016479492} 02/24/2022 13:22:23 - INFO - codeparrot_training - Step 9116: {'lr': 0.0004733722472874759, 'samples': 4667904, 'steps': 9116, 'loss/train': 2.4675910472869873} 02/24/2022 13:22:28 - INFO - codeparrot_training - Step 9117: {'lr': 0.0004733648986713017, 'samples': 4668416, 'steps': 9117, 'loss/train': 2.8457374572753906} 02/24/2022 13:22:32 - INFO - codeparrot_training - Step 9118: {'lr': 0.00047335754909830327, 'samples': 4668928, 'steps': 9118, 'loss/train': 1.828161358833313} 02/24/2022 13:22:37 - INFO - codeparrot_training - Step 9119: {'lr': 0.00047335019856851204, 'samples': 4669440, 'steps': 9119, 'loss/train': 1.8511936664581299} 02/24/2022 13:22:41 - INFO - codeparrot_training - Step 9120: {'lr': 0.0004733428470819594, 'samples': 4669952, 'steps': 9120, 'loss/train': 4.304659366607666} 02/24/2022 13:22:46 - INFO - codeparrot_training - Step 9121: {'lr': 0.000473335494638677, 'samples': 4670464, 'steps': 9121, 'loss/train': 1.58551025390625} 02/24/2022 13:22:50 - INFO - codeparrot_training - Step 9122: {'lr': 0.00047332814123869616, 'samples': 4670976, 'steps': 9122, 'loss/train': 1.7067691087722778} 02/24/2022 13:22:55 - INFO - codeparrot_training - Step 9123: {'lr': 0.0004733207868820486, 'samples': 4671488, 'steps': 9123, 'loss/train': 2.051353931427002} 02/24/2022 13:22:59 - INFO - codeparrot_training - Step 9124: {'lr': 0.0004733134315687656, 'samples': 4672000, 'steps': 9124, 'loss/train': 1.982812762260437} 02/24/2022 13:23:05 - INFO - codeparrot_training - Step 9125: {'lr': 0.00047330607529887884, 'samples': 4672512, 'steps': 9125, 'loss/train': 1.914789080619812} 02/24/2022 13:23:08 - INFO - codeparrot_training - Step 9126: {'lr': 0.00047329871807241976, 'samples': 4673024, 'steps': 9126, 'loss/train': 1.613128423690796} 02/24/2022 13:23:14 - INFO - codeparrot_training - Step 9127: {'lr': 0.00047329135988941984, 'samples': 4673536, 'steps': 9127, 'loss/train': 1.704093098640442} 02/24/2022 13:23:17 - INFO - codeparrot_training - Step 9128: {'lr': 0.00047328400074991064, 'samples': 4674048, 'steps': 9128, 'loss/train': 2.713627576828003} 02/24/2022 13:23:23 - INFO - codeparrot_training - Step 9129: {'lr': 0.00047327664065392375, 'samples': 4674560, 'steps': 9129, 'loss/train': 1.7202262878417969} 02/24/2022 13:23:26 - INFO - codeparrot_training - Step 9130: {'lr': 0.0004732692796014905, 'samples': 4675072, 'steps': 9130, 'loss/train': 1.0255662202835083} 02/24/2022 13:23:32 - INFO - codeparrot_training - Step 9131: {'lr': 0.00047326191759264265, 'samples': 4675584, 'steps': 9131, 'loss/train': 2.1260640621185303} 02/24/2022 13:23:35 - INFO - codeparrot_training - Step 9132: {'lr': 0.00047325455462741164, 'samples': 4676096, 'steps': 9132, 'loss/train': 1.1545182466506958} 02/24/2022 13:23:41 - INFO - codeparrot_training - Step 9133: {'lr': 0.00047324719070582894, 'samples': 4676608, 'steps': 9133, 'loss/train': 2.8290140628814697} 02/24/2022 13:23:44 - INFO - codeparrot_training - Step 9134: {'lr': 0.00047323982582792625, 'samples': 4677120, 'steps': 9134, 'loss/train': 1.455802083015442} 02/24/2022 13:23:50 - INFO - codeparrot_training - Step 9135: {'lr': 0.00047323245999373497, 'samples': 4677632, 'steps': 9135, 'loss/train': 2.5370726585388184} 02/24/2022 13:23:54 - INFO - codeparrot_training - Step 9136: {'lr': 0.0004732250932032867, 'samples': 4678144, 'steps': 9136, 'loss/train': 2.597081184387207} 02/24/2022 13:23:59 - INFO - codeparrot_training - Step 9137: {'lr': 0.0004732177254566131, 'samples': 4678656, 'steps': 9137, 'loss/train': 1.2114132642745972} 02/24/2022 13:24:02 - INFO - codeparrot_training - Step 9138: {'lr': 0.0004732103567537456, 'samples': 4679168, 'steps': 9138, 'loss/train': 2.1426267623901367} 02/24/2022 13:24:10 - INFO - codeparrot_training - Step 9139: {'lr': 0.00047320298709471574, 'samples': 4679680, 'steps': 9139, 'loss/train': 1.139854907989502} 02/24/2022 13:24:13 - INFO - codeparrot_training - Step 9140: {'lr': 0.0004731956164795552, 'samples': 4680192, 'steps': 9140, 'loss/train': 2.542071580886841} 02/24/2022 13:24:19 - INFO - codeparrot_training - Step 9141: {'lr': 0.0004731882449082956, 'samples': 4680704, 'steps': 9141, 'loss/train': 2.5392870903015137} 02/24/2022 13:24:22 - INFO - codeparrot_training - Step 9142: {'lr': 0.0004731808723809683, 'samples': 4681216, 'steps': 9142, 'loss/train': 1.6529693603515625} 02/24/2022 13:24:28 - INFO - codeparrot_training - Step 9143: {'lr': 0.0004731734988976051, 'samples': 4681728, 'steps': 9143, 'loss/train': 2.2855355739593506} 02/24/2022 13:24:32 - INFO - codeparrot_training - Step 9144: {'lr': 0.00047316612445823746, 'samples': 4682240, 'steps': 9144, 'loss/train': 2.11826491355896} 02/24/2022 13:24:37 - INFO - codeparrot_training - Step 9145: {'lr': 0.000473158749062897, 'samples': 4682752, 'steps': 9145, 'loss/train': 2.2974278926849365} 02/24/2022 13:24:40 - INFO - codeparrot_training - Step 9146: {'lr': 0.00047315137271161537, 'samples': 4683264, 'steps': 9146, 'loss/train': 2.2753851413726807} 02/24/2022 13:24:46 - INFO - codeparrot_training - Step 9147: {'lr': 0.00047314399540442407, 'samples': 4683776, 'steps': 9147, 'loss/train': 0.22249168157577515} 02/24/2022 13:24:50 - INFO - codeparrot_training - Step 9148: {'lr': 0.00047313661714135476, 'samples': 4684288, 'steps': 9148, 'loss/train': 2.55893874168396} 02/24/2022 13:24:56 - INFO - codeparrot_training - Step 9149: {'lr': 0.000473129237922439, 'samples': 4684800, 'steps': 9149, 'loss/train': 1.9013292789459229} 02/24/2022 13:25:00 - INFO - codeparrot_training - Step 9150: {'lr': 0.0004731218577477085, 'samples': 4685312, 'steps': 9150, 'loss/train': 2.48695969581604} 02/24/2022 13:25:05 - INFO - codeparrot_training - Step 9151: {'lr': 0.0004731144766171948, 'samples': 4685824, 'steps': 9151, 'loss/train': 1.3969038724899292} 02/24/2022 13:25:09 - INFO - codeparrot_training - Step 9152: {'lr': 0.0004731070945309295, 'samples': 4686336, 'steps': 9152, 'loss/train': 3.5837178230285645} 02/24/2022 13:25:14 - INFO - codeparrot_training - Step 9153: {'lr': 0.00047309971148894425, 'samples': 4686848, 'steps': 9153, 'loss/train': 2.50506854057312} 02/24/2022 13:25:18 - INFO - codeparrot_training - Step 9154: {'lr': 0.00047309232749127074, 'samples': 4687360, 'steps': 9154, 'loss/train': 0.8776212334632874} 02/24/2022 13:25:23 - INFO - codeparrot_training - Step 9155: {'lr': 0.0004730849425379404, 'samples': 4687872, 'steps': 9155, 'loss/train': 2.5917952060699463} 02/24/2022 13:25:27 - INFO - codeparrot_training - Step 9156: {'lr': 0.0004730775566289851, 'samples': 4688384, 'steps': 9156, 'loss/train': 2.4982125759124756} 02/24/2022 13:25:32 - INFO - codeparrot_training - Step 9157: {'lr': 0.0004730701697644364, 'samples': 4688896, 'steps': 9157, 'loss/train': 2.920562982559204} 02/24/2022 13:25:36 - INFO - codeparrot_training - Step 9158: {'lr': 0.00047306278194432597, 'samples': 4689408, 'steps': 9158, 'loss/train': 1.4460726976394653} 02/24/2022 13:25:43 - INFO - codeparrot_training - Step 9159: {'lr': 0.0004730553931686853, 'samples': 4689920, 'steps': 9159, 'loss/train': 0.9634991884231567} 02/24/2022 13:25:46 - INFO - codeparrot_training - Step 9160: {'lr': 0.00047304800343754615, 'samples': 4690432, 'steps': 9160, 'loss/train': 2.0099990367889404} 02/24/2022 13:25:52 - INFO - codeparrot_training - Step 9161: {'lr': 0.00047304061275094025, 'samples': 4690944, 'steps': 9161, 'loss/train': 2.5363290309906006} 02/24/2022 13:25:55 - INFO - codeparrot_training - Step 9162: {'lr': 0.0004730332211088992, 'samples': 4691456, 'steps': 9162, 'loss/train': 0.39027178287506104} 02/24/2022 13:26:01 - INFO - codeparrot_training - Step 9163: {'lr': 0.0004730258285114546, 'samples': 4691968, 'steps': 9163, 'loss/train': 2.634152412414551} 02/24/2022 13:26:04 - INFO - codeparrot_training - Step 9164: {'lr': 0.0004730184349586382, 'samples': 4692480, 'steps': 9164, 'loss/train': 2.871751070022583} 02/24/2022 13:26:10 - INFO - codeparrot_training - Step 9165: {'lr': 0.0004730110404504816, 'samples': 4692992, 'steps': 9165, 'loss/train': 1.150871753692627} 02/24/2022 13:26:13 - INFO - codeparrot_training - Step 9166: {'lr': 0.00047300364498701654, 'samples': 4693504, 'steps': 9166, 'loss/train': 1.7111626863479614} 02/24/2022 13:26:19 - INFO - codeparrot_training - Step 9167: {'lr': 0.00047299624856827474, 'samples': 4694016, 'steps': 9167, 'loss/train': 0.24416518211364746} 02/24/2022 13:26:22 - INFO - codeparrot_training - Step 9168: {'lr': 0.0004729888511942877, 'samples': 4694528, 'steps': 9168, 'loss/train': 1.2803181409835815} 02/24/2022 13:26:29 - INFO - codeparrot_training - Step 9169: {'lr': 0.0004729814528650873, 'samples': 4695040, 'steps': 9169, 'loss/train': 2.793639898300171} 02/24/2022 13:26:32 - INFO - codeparrot_training - Step 9170: {'lr': 0.00047297405358070517, 'samples': 4695552, 'steps': 9170, 'loss/train': 2.7124135494232178} 02/24/2022 13:26:38 - INFO - codeparrot_training - Step 9171: {'lr': 0.00047296665334117295, 'samples': 4696064, 'steps': 9171, 'loss/train': 1.1330863237380981} 02/24/2022 13:26:41 - INFO - codeparrot_training - Step 9172: {'lr': 0.0004729592521465224, 'samples': 4696576, 'steps': 9172, 'loss/train': 0.8860397338867188} 02/24/2022 13:26:47 - INFO - codeparrot_training - Step 9173: {'lr': 0.00047295184999678524, 'samples': 4697088, 'steps': 9173, 'loss/train': 2.290264844894409} 02/24/2022 13:26:52 - INFO - codeparrot_training - Step 9174: {'lr': 0.00047294444689199313, 'samples': 4697600, 'steps': 9174, 'loss/train': 1.532676100730896} 02/24/2022 13:26:56 - INFO - codeparrot_training - Step 9175: {'lr': 0.0004729370428321778, 'samples': 4698112, 'steps': 9175, 'loss/train': 2.995622396469116} 02/24/2022 13:27:01 - INFO - codeparrot_training - Step 9176: {'lr': 0.000472929637817371, 'samples': 4698624, 'steps': 9176, 'loss/train': 2.861117362976074} 02/24/2022 13:27:05 - INFO - codeparrot_training - Step 9177: {'lr': 0.0004729222318476044, 'samples': 4699136, 'steps': 9177, 'loss/train': 2.28058123588562} 02/24/2022 13:27:10 - INFO - codeparrot_training - Step 9178: {'lr': 0.0004729148249229097, 'samples': 4699648, 'steps': 9178, 'loss/train': 1.9214019775390625} 02/24/2022 13:27:13 - INFO - codeparrot_training - Step 9179: {'lr': 0.0004729074170433187, 'samples': 4700160, 'steps': 9179, 'loss/train': 1.9854960441589355} 02/24/2022 13:27:19 - INFO - codeparrot_training - Step 9180: {'lr': 0.0004729000082088631, 'samples': 4700672, 'steps': 9180, 'loss/train': 1.4668793678283691} 02/24/2022 13:27:22 - INFO - codeparrot_training - Step 9181: {'lr': 0.0004728925984195748, 'samples': 4701184, 'steps': 9181, 'loss/train': 1.8076584339141846} 02/24/2022 13:27:28 - INFO - codeparrot_training - Step 9182: {'lr': 0.00047288518767548516, 'samples': 4701696, 'steps': 9182, 'loss/train': 2.297683000564575} 02/24/2022 13:27:31 - INFO - codeparrot_training - Step 9183: {'lr': 0.0004728777759766263, 'samples': 4702208, 'steps': 9183, 'loss/train': 1.3928289413452148} 02/24/2022 13:27:38 - INFO - codeparrot_training - Step 9184: {'lr': 0.00047287036332302967, 'samples': 4702720, 'steps': 9184, 'loss/train': 3.4282398223876953} 02/24/2022 13:27:42 - INFO - codeparrot_training - Step 9185: {'lr': 0.0004728629497147273, 'samples': 4703232, 'steps': 9185, 'loss/train': 0.30095282196998596} 02/24/2022 13:27:47 - INFO - codeparrot_training - Step 9186: {'lr': 0.00047285553515175077, 'samples': 4703744, 'steps': 9186, 'loss/train': 2.145885705947876} 02/24/2022 13:27:51 - INFO - codeparrot_training - Step 9187: {'lr': 0.0004728481196341319, 'samples': 4704256, 'steps': 9187, 'loss/train': 0.6703117489814758} 02/24/2022 13:27:56 - INFO - codeparrot_training - Step 9188: {'lr': 0.0004728407031619025, 'samples': 4704768, 'steps': 9188, 'loss/train': 1.5894571542739868} 02/24/2022 13:28:00 - INFO - codeparrot_training - Step 9189: {'lr': 0.0004728332857350942, 'samples': 4705280, 'steps': 9189, 'loss/train': 2.1085352897644043} 02/24/2022 13:28:05 - INFO - codeparrot_training - Step 9190: {'lr': 0.00047282586735373887, 'samples': 4705792, 'steps': 9190, 'loss/train': 2.5246129035949707} 02/24/2022 13:28:09 - INFO - codeparrot_training - Step 9191: {'lr': 0.0004728184480178683, 'samples': 4706304, 'steps': 9191, 'loss/train': 2.775531530380249} 02/24/2022 13:28:14 - INFO - codeparrot_training - Step 9192: {'lr': 0.00047281102772751425, 'samples': 4706816, 'steps': 9192, 'loss/train': 1.6418699026107788} 02/24/2022 13:28:18 - INFO - codeparrot_training - Step 9193: {'lr': 0.0004728036064827086, 'samples': 4707328, 'steps': 9193, 'loss/train': 2.6444785594940186} 02/24/2022 13:28:25 - INFO - codeparrot_training - Step 9194: {'lr': 0.00047279618428348294, 'samples': 4707840, 'steps': 9194, 'loss/train': 1.7925957441329956} 02/24/2022 13:28:28 - INFO - codeparrot_training - Step 9195: {'lr': 0.00047278876112986923, 'samples': 4708352, 'steps': 9195, 'loss/train': 1.8629181385040283} 02/24/2022 13:28:34 - INFO - codeparrot_training - Step 9196: {'lr': 0.0004727813370218992, 'samples': 4708864, 'steps': 9196, 'loss/train': 2.582183599472046} 02/24/2022 13:28:37 - INFO - codeparrot_training - Step 9197: {'lr': 0.00047277391195960463, 'samples': 4709376, 'steps': 9197, 'loss/train': 4.056845188140869} 02/24/2022 13:28:43 - INFO - codeparrot_training - Step 9198: {'lr': 0.00047276648594301733, 'samples': 4709888, 'steps': 9198, 'loss/train': 2.2332284450531006} 02/24/2022 13:28:46 - INFO - codeparrot_training - Step 9199: {'lr': 0.0004727590589721692, 'samples': 4710400, 'steps': 9199, 'loss/train': 2.9682910442352295} 02/24/2022 13:28:52 - INFO - codeparrot_training - Step 9200: {'lr': 0.00047275163104709196, 'samples': 4710912, 'steps': 9200, 'loss/train': 2.4550609588623047} 02/24/2022 13:28:55 - INFO - codeparrot_training - Step 9201: {'lr': 0.0004727442021678175, 'samples': 4711424, 'steps': 9201, 'loss/train': 2.3716583251953125} 02/24/2022 13:29:01 - INFO - codeparrot_training - Step 9202: {'lr': 0.0004727367723343776, 'samples': 4711936, 'steps': 9202, 'loss/train': 1.4867007732391357} 02/24/2022 13:29:04 - INFO - codeparrot_training - Step 9203: {'lr': 0.0004727293415468041, 'samples': 4712448, 'steps': 9203, 'loss/train': 1.2493293285369873} 02/24/2022 13:29:11 - INFO - codeparrot_training - Step 9204: {'lr': 0.00047272190980512875, 'samples': 4712960, 'steps': 9204, 'loss/train': 2.5243260860443115} 02/24/2022 13:29:14 - INFO - codeparrot_training - Step 9205: {'lr': 0.0004727144771093835, 'samples': 4713472, 'steps': 9205, 'loss/train': 2.4383633136749268} 02/24/2022 13:29:20 - INFO - codeparrot_training - Step 9206: {'lr': 0.00047270704345960023, 'samples': 4713984, 'steps': 9206, 'loss/train': 1.977097511291504} 02/24/2022 13:29:23 - INFO - codeparrot_training - Step 9207: {'lr': 0.00047269960885581064, 'samples': 4714496, 'steps': 9207, 'loss/train': 2.731149196624756} 02/24/2022 13:29:29 - INFO - codeparrot_training - Step 9208: {'lr': 0.00047269217329804663, 'samples': 4715008, 'steps': 9208, 'loss/train': 2.0739552974700928} 02/24/2022 13:29:32 - INFO - codeparrot_training - Step 9209: {'lr': 0.00047268473678634007, 'samples': 4715520, 'steps': 9209, 'loss/train': 1.8604692220687866} 02/24/2022 13:29:38 - INFO - codeparrot_training - Step 9210: {'lr': 0.00047267729932072284, 'samples': 4716032, 'steps': 9210, 'loss/train': 2.654740571975708} 02/24/2022 13:29:42 - INFO - codeparrot_training - Step 9211: {'lr': 0.00047266986090122677, 'samples': 4716544, 'steps': 9211, 'loss/train': 2.5369644165039062} 02/24/2022 13:29:47 - INFO - codeparrot_training - Step 9212: {'lr': 0.0004726624215278836, 'samples': 4717056, 'steps': 9212, 'loss/train': 3.6587746143341064} 02/24/2022 13:29:51 - INFO - codeparrot_training - Step 9213: {'lr': 0.00047265498120072546, 'samples': 4717568, 'steps': 9213, 'loss/train': 2.129279136657715} 02/24/2022 13:29:54 - INFO - codeparrot_training - Step 9214: {'lr': 0.00047264753991978404, 'samples': 4718080, 'steps': 9214, 'loss/train': 3.0930986404418945} 02/24/2022 13:30:00 - INFO - codeparrot_training - Step 9215: {'lr': 0.00047264009768509127, 'samples': 4718592, 'steps': 9215, 'loss/train': 1.4839849472045898} 02/24/2022 13:30:03 - INFO - codeparrot_training - Step 9216: {'lr': 0.000472632654496679, 'samples': 4719104, 'steps': 9216, 'loss/train': 2.642756462097168} 02/24/2022 13:30:09 - INFO - codeparrot_training - Step 9217: {'lr': 0.00047262521035457914, 'samples': 4719616, 'steps': 9217, 'loss/train': 1.7101268768310547} 02/24/2022 13:30:15 - INFO - codeparrot_training - Step 9218: {'lr': 0.00047261776525882353, 'samples': 4720128, 'steps': 9218, 'loss/train': 3.559933662414551} 02/24/2022 13:30:18 - INFO - codeparrot_training - Step 9219: {'lr': 0.00047261031920944413, 'samples': 4720640, 'steps': 9219, 'loss/train': 2.0281102657318115} 02/24/2022 13:30:24 - INFO - codeparrot_training - Step 9220: {'lr': 0.0004726028722064728, 'samples': 4721152, 'steps': 9220, 'loss/train': 1.7223827838897705} 02/24/2022 13:30:27 - INFO - codeparrot_training - Step 9221: {'lr': 0.0004725954242499415, 'samples': 4721664, 'steps': 9221, 'loss/train': 3.576780080795288} 02/24/2022 13:30:33 - INFO - codeparrot_training - Step 9222: {'lr': 0.00047258797533988205, 'samples': 4722176, 'steps': 9222, 'loss/train': 1.093144178390503} 02/24/2022 13:30:37 - INFO - codeparrot_training - Step 9223: {'lr': 0.00047258052547632636, 'samples': 4722688, 'steps': 9223, 'loss/train': 2.557705879211426} 02/24/2022 13:30:42 - INFO - codeparrot_training - Step 9224: {'lr': 0.0004725730746593064, 'samples': 4723200, 'steps': 9224, 'loss/train': 1.4148427248001099} 02/24/2022 13:30:45 - INFO - codeparrot_training - Step 9225: {'lr': 0.0004725656228888541, 'samples': 4723712, 'steps': 9225, 'loss/train': 1.9137229919433594} 02/24/2022 13:30:51 - INFO - codeparrot_training - Step 9226: {'lr': 0.0004725581701650014, 'samples': 4724224, 'steps': 9226, 'loss/train': 1.8964622020721436} 02/24/2022 13:30:55 - INFO - codeparrot_training - Step 9227: {'lr': 0.00047255071648778004, 'samples': 4724736, 'steps': 9227, 'loss/train': 1.9602665901184082} 02/24/2022 13:31:00 - INFO - codeparrot_training - Step 9228: {'lr': 0.00047254326185722207, 'samples': 4725248, 'steps': 9228, 'loss/train': 2.1882286071777344} 02/24/2022 13:31:04 - INFO - codeparrot_training - Step 9229: {'lr': 0.00047253580627335944, 'samples': 4725760, 'steps': 9229, 'loss/train': 2.884420394897461} 02/24/2022 13:31:10 - INFO - codeparrot_training - Step 9230: {'lr': 0.00047252834973622414, 'samples': 4726272, 'steps': 9230, 'loss/train': 3.4539942741394043} 02/24/2022 13:31:13 - INFO - codeparrot_training - Step 9231: {'lr': 0.00047252089224584804, 'samples': 4726784, 'steps': 9231, 'loss/train': 2.046510934829712} 02/24/2022 13:31:19 - INFO - codeparrot_training - Step 9232: {'lr': 0.0004725134338022631, 'samples': 4727296, 'steps': 9232, 'loss/train': 2.272949695587158} 02/24/2022 13:31:22 - INFO - codeparrot_training - Step 9233: {'lr': 0.00047250597440550124, 'samples': 4727808, 'steps': 9233, 'loss/train': 2.060175657272339} 02/24/2022 13:31:28 - INFO - codeparrot_training - Step 9234: {'lr': 0.0004724985140555945, 'samples': 4728320, 'steps': 9234, 'loss/train': 1.7731454372406006} 02/24/2022 13:31:31 - INFO - codeparrot_training - Step 9235: {'lr': 0.0004724910527525748, 'samples': 4728832, 'steps': 9235, 'loss/train': 2.637456178665161} 02/24/2022 13:31:37 - INFO - codeparrot_training - Step 9236: {'lr': 0.0004724835904964739, 'samples': 4729344, 'steps': 9236, 'loss/train': 2.847226858139038} 02/24/2022 13:31:40 - INFO - codeparrot_training - Step 9237: {'lr': 0.00047247612728732407, 'samples': 4729856, 'steps': 9237, 'loss/train': 2.606318235397339} 02/24/2022 13:31:46 - INFO - codeparrot_training - Step 9238: {'lr': 0.0004724686631251572, 'samples': 4730368, 'steps': 9238, 'loss/train': 1.946047306060791} 02/24/2022 13:31:49 - INFO - codeparrot_training - Step 9239: {'lr': 0.00047246119801000507, 'samples': 4730880, 'steps': 9239, 'loss/train': 3.4506382942199707} 02/24/2022 13:31:56 - INFO - codeparrot_training - Step 9240: {'lr': 0.00047245373194189995, 'samples': 4731392, 'steps': 9240, 'loss/train': 1.6544857025146484} 02/24/2022 13:31:59 - INFO - codeparrot_training - Step 9241: {'lr': 0.0004724462649208736, 'samples': 4731904, 'steps': 9241, 'loss/train': 0.7858849763870239} 02/24/2022 13:32:04 - INFO - codeparrot_training - Step 9242: {'lr': 0.0004724387969469581, 'samples': 4732416, 'steps': 9242, 'loss/train': 2.2347378730773926} 02/24/2022 13:32:08 - INFO - codeparrot_training - Step 9243: {'lr': 0.00047243132802018544, 'samples': 4732928, 'steps': 9243, 'loss/train': 1.0245441198349} 02/24/2022 13:32:14 - INFO - codeparrot_training - Step 9244: {'lr': 0.00047242385814058764, 'samples': 4733440, 'steps': 9244, 'loss/train': 2.605638027191162} 02/24/2022 13:32:17 - INFO - codeparrot_training - Step 9245: {'lr': 0.0004724163873081966, 'samples': 4733952, 'steps': 9245, 'loss/train': 2.2589056491851807} 02/24/2022 13:32:23 - INFO - codeparrot_training - Step 9246: {'lr': 0.00047240891552304443, 'samples': 4734464, 'steps': 9246, 'loss/train': 1.3129860162734985} 02/24/2022 13:32:26 - INFO - codeparrot_training - Step 9247: {'lr': 0.0004724014427851631, 'samples': 4734976, 'steps': 9247, 'loss/train': 1.9890809059143066} 02/24/2022 13:32:32 - INFO - codeparrot_training - Step 9248: {'lr': 0.0004723939690945845, 'samples': 4735488, 'steps': 9248, 'loss/train': 2.149052858352661} 02/24/2022 13:32:35 - INFO - codeparrot_training - Step 9249: {'lr': 0.00047238649445134086, 'samples': 4736000, 'steps': 9249, 'loss/train': 2.1851701736450195} 02/24/2022 13:32:41 - INFO - codeparrot_training - Step 9250: {'lr': 0.00047237901885546405, 'samples': 4736512, 'steps': 9250, 'loss/train': 2.6466867923736572} 02/24/2022 13:32:44 - INFO - codeparrot_training - Step 9251: {'lr': 0.00047237154230698607, 'samples': 4737024, 'steps': 9251, 'loss/train': 1.60611891746521} 02/24/2022 13:32:50 - INFO - codeparrot_training - Step 9252: {'lr': 0.0004723640648059391, 'samples': 4737536, 'steps': 9252, 'loss/train': 2.268428325653076} 02/24/2022 13:32:54 - INFO - codeparrot_training - Step 9253: {'lr': 0.0004723565863523551, 'samples': 4738048, 'steps': 9253, 'loss/train': 2.487484931945801} 02/24/2022 13:32:59 - INFO - codeparrot_training - Step 9254: {'lr': 0.0004723491069462661, 'samples': 4738560, 'steps': 9254, 'loss/train': 2.7203564643859863} 02/24/2022 13:33:03 - INFO - codeparrot_training - Step 9255: {'lr': 0.00047234162658770407, 'samples': 4739072, 'steps': 9255, 'loss/train': 2.055199384689331} 02/24/2022 13:33:08 - INFO - codeparrot_training - Step 9256: {'lr': 0.00047233414527670113, 'samples': 4739584, 'steps': 9256, 'loss/train': 2.022843599319458} 02/24/2022 13:33:12 - INFO - codeparrot_training - Step 9257: {'lr': 0.0004723266630132893, 'samples': 4740096, 'steps': 9257, 'loss/train': 0.8023931980133057} 02/24/2022 13:33:17 - INFO - codeparrot_training - Step 9258: {'lr': 0.0004723191797975007, 'samples': 4740608, 'steps': 9258, 'loss/train': 2.3807342052459717} 02/24/2022 13:33:21 - INFO - codeparrot_training - Step 9259: {'lr': 0.00047231169562936726, 'samples': 4741120, 'steps': 9259, 'loss/train': 0.8737770915031433} 02/24/2022 13:33:26 - INFO - codeparrot_training - Step 9260: {'lr': 0.00047230421050892116, 'samples': 4741632, 'steps': 9260, 'loss/train': 3.2700276374816895} 02/24/2022 13:33:30 - INFO - codeparrot_training - Step 9261: {'lr': 0.00047229672443619433, 'samples': 4742144, 'steps': 9261, 'loss/train': 2.2991623878479004} 02/24/2022 13:33:36 - INFO - codeparrot_training - Step 9262: {'lr': 0.00047228923741121897, 'samples': 4742656, 'steps': 9262, 'loss/train': 2.6805050373077393} 02/24/2022 13:33:40 - INFO - codeparrot_training - Step 9263: {'lr': 0.0004722817494340271, 'samples': 4743168, 'steps': 9263, 'loss/train': 2.879227638244629} 02/24/2022 13:33:45 - INFO - codeparrot_training - Step 9264: {'lr': 0.00047227426050465085, 'samples': 4743680, 'steps': 9264, 'loss/train': 1.4752908945083618} 02/24/2022 13:33:49 - INFO - codeparrot_training - Step 9265: {'lr': 0.00047226677062312217, 'samples': 4744192, 'steps': 9265, 'loss/train': 2.207026243209839} 02/24/2022 13:33:54 - INFO - codeparrot_training - Step 9266: {'lr': 0.00047225927978947327, 'samples': 4744704, 'steps': 9266, 'loss/train': 1.518178939819336} 02/24/2022 13:33:58 - INFO - codeparrot_training - Step 9267: {'lr': 0.00047225178800373613, 'samples': 4745216, 'steps': 9267, 'loss/train': 1.1507806777954102} 02/24/2022 13:34:03 - INFO - codeparrot_training - Step 9268: {'lr': 0.00047224429526594296, 'samples': 4745728, 'steps': 9268, 'loss/train': 2.770920991897583} 02/24/2022 13:34:07 - INFO - codeparrot_training - Step 9269: {'lr': 0.0004722368015761258, 'samples': 4746240, 'steps': 9269, 'loss/train': 2.097870111465454} 02/24/2022 13:34:12 - INFO - codeparrot_training - Step 9270: {'lr': 0.0004722293069343168, 'samples': 4746752, 'steps': 9270, 'loss/train': 1.8893096446990967} 02/24/2022 13:34:16 - INFO - codeparrot_training - Step 9271: {'lr': 0.00047222181134054785, 'samples': 4747264, 'steps': 9271, 'loss/train': 2.4055800437927246} 02/24/2022 13:34:21 - INFO - codeparrot_training - Step 9272: {'lr': 0.0004722143147948513, 'samples': 4747776, 'steps': 9272, 'loss/train': 2.748605966567993} 02/24/2022 13:34:25 - INFO - codeparrot_training - Step 9273: {'lr': 0.0004722068172972593, 'samples': 4748288, 'steps': 9273, 'loss/train': 2.956089973449707} 02/24/2022 13:34:30 - INFO - codeparrot_training - Step 9274: {'lr': 0.00047219931884780376, 'samples': 4748800, 'steps': 9274, 'loss/train': 2.7256693840026855} 02/24/2022 13:34:34 - INFO - codeparrot_training - Step 9275: {'lr': 0.0004721918194465169, 'samples': 4749312, 'steps': 9275, 'loss/train': 2.108445167541504} 02/24/2022 13:34:39 - INFO - codeparrot_training - Step 9276: {'lr': 0.00047218431909343083, 'samples': 4749824, 'steps': 9276, 'loss/train': 2.1449239253997803} 02/24/2022 13:34:43 - INFO - codeparrot_training - Step 9277: {'lr': 0.0004721768177885777, 'samples': 4750336, 'steps': 9277, 'loss/train': 2.3894524574279785} 02/24/2022 13:34:49 - INFO - codeparrot_training - Step 9278: {'lr': 0.00047216931553198963, 'samples': 4750848, 'steps': 9278, 'loss/train': 1.8229236602783203} 02/24/2022 13:34:53 - INFO - codeparrot_training - Step 9279: {'lr': 0.0004721618123236987, 'samples': 4751360, 'steps': 9279, 'loss/train': 2.724698305130005} 02/24/2022 13:34:58 - INFO - codeparrot_training - Step 9280: {'lr': 0.0004721543081637372, 'samples': 4751872, 'steps': 9280, 'loss/train': 2.0684165954589844} 02/24/2022 13:35:02 - INFO - codeparrot_training - Step 9281: {'lr': 0.0004721468030521372, 'samples': 4752384, 'steps': 9281, 'loss/train': 2.5664479732513428} 02/24/2022 13:35:07 - INFO - codeparrot_training - Step 9282: {'lr': 0.0004721392969889308, 'samples': 4752896, 'steps': 9282, 'loss/train': 2.4262731075286865} 02/24/2022 13:35:11 - INFO - codeparrot_training - Step 9283: {'lr': 0.00047213178997415015, 'samples': 4753408, 'steps': 9283, 'loss/train': 2.0224199295043945} 02/24/2022 13:35:16 - INFO - codeparrot_training - Step 9284: {'lr': 0.00047212428200782744, 'samples': 4753920, 'steps': 9284, 'loss/train': 1.835253119468689} 02/24/2022 13:35:20 - INFO - codeparrot_training - Step 9285: {'lr': 0.0004721167730899949, 'samples': 4754432, 'steps': 9285, 'loss/train': 1.6648372411727905} 02/24/2022 13:35:25 - INFO - codeparrot_training - Step 9286: {'lr': 0.0004721092632206846, 'samples': 4754944, 'steps': 9286, 'loss/train': 2.646677017211914} 02/24/2022 13:35:29 - INFO - codeparrot_training - Step 9287: {'lr': 0.00047210175239992876, 'samples': 4755456, 'steps': 9287, 'loss/train': 2.926750659942627} 02/24/2022 13:35:35 - INFO - codeparrot_training - Step 9288: {'lr': 0.0004720942406277595, 'samples': 4755968, 'steps': 9288, 'loss/train': 1.5843933820724487} 02/24/2022 13:35:39 - INFO - codeparrot_training - Step 9289: {'lr': 0.0004720867279042091, 'samples': 4756480, 'steps': 9289, 'loss/train': 1.838757872581482} 02/24/2022 13:35:44 - INFO - codeparrot_training - Step 9290: {'lr': 0.00047207921422930967, 'samples': 4756992, 'steps': 9290, 'loss/train': 1.2665789127349854} 02/24/2022 13:35:48 - INFO - codeparrot_training - Step 9291: {'lr': 0.00047207169960309335, 'samples': 4757504, 'steps': 9291, 'loss/train': 1.840664267539978} 02/24/2022 13:35:53 - INFO - codeparrot_training - Step 9292: {'lr': 0.00047206418402559236, 'samples': 4758016, 'steps': 9292, 'loss/train': 2.8592965602874756} 02/24/2022 13:35:57 - INFO - codeparrot_training - Step 9293: {'lr': 0.000472056667496839, 'samples': 4758528, 'steps': 9293, 'loss/train': 2.743601083755493} 02/24/2022 13:36:02 - INFO - codeparrot_training - Step 9294: {'lr': 0.0004720491500168654, 'samples': 4759040, 'steps': 9294, 'loss/train': 1.9503521919250488} 02/24/2022 13:36:06 - INFO - codeparrot_training - Step 9295: {'lr': 0.0004720416315857037, 'samples': 4759552, 'steps': 9295, 'loss/train': 2.6600232124328613} 02/24/2022 13:36:11 - INFO - codeparrot_training - Step 9296: {'lr': 0.00047203411220338615, 'samples': 4760064, 'steps': 9296, 'loss/train': 2.670135498046875} 02/24/2022 13:36:17 - INFO - codeparrot_training - Step 9297: {'lr': 0.000472026591869945, 'samples': 4760576, 'steps': 9297, 'loss/train': 2.2512452602386475} 02/24/2022 13:36:21 - INFO - codeparrot_training - Step 9298: {'lr': 0.00047201907058541236, 'samples': 4761088, 'steps': 9298, 'loss/train': 2.2479238510131836} 02/24/2022 13:36:26 - INFO - codeparrot_training - Step 9299: {'lr': 0.0004720115483498206, 'samples': 4761600, 'steps': 9299, 'loss/train': 1.8936635255813599} 02/24/2022 13:36:30 - INFO - codeparrot_training - Step 9300: {'lr': 0.00047200402516320186, 'samples': 4762112, 'steps': 9300, 'loss/train': 1.2354745864868164} 02/24/2022 13:36:35 - INFO - codeparrot_training - Step 9301: {'lr': 0.00047199650102558834, 'samples': 4762624, 'steps': 9301, 'loss/train': 1.5507136583328247} 02/24/2022 13:36:39 - INFO - codeparrot_training - Step 9302: {'lr': 0.0004719889759370123, 'samples': 4763136, 'steps': 9302, 'loss/train': 2.3888351917266846} 02/24/2022 13:36:42 - INFO - codeparrot_training - Step 9303: {'lr': 0.00047198144989750603, 'samples': 4763648, 'steps': 9303, 'loss/train': 1.7949309349060059} 02/24/2022 13:36:48 - INFO - codeparrot_training - Step 9304: {'lr': 0.00047197392290710164, 'samples': 4764160, 'steps': 9304, 'loss/train': 2.4654555320739746} 02/24/2022 13:36:51 - INFO - codeparrot_training - Step 9305: {'lr': 0.0004719663949658315, 'samples': 4764672, 'steps': 9305, 'loss/train': 1.472735047340393} 02/24/2022 13:36:57 - INFO - codeparrot_training - Step 9306: {'lr': 0.00047195886607372773, 'samples': 4765184, 'steps': 9306, 'loss/train': 2.5258872509002686} 02/24/2022 13:37:00 - INFO - codeparrot_training - Step 9307: {'lr': 0.0004719513362308228, 'samples': 4765696, 'steps': 9307, 'loss/train': 1.6016918420791626} 02/24/2022 13:37:07 - INFO - codeparrot_training - Step 9308: {'lr': 0.0004719438054371487, 'samples': 4766208, 'steps': 9308, 'loss/train': 3.078833818435669} 02/24/2022 13:37:11 - INFO - codeparrot_training - Step 9309: {'lr': 0.00047193627369273786, 'samples': 4766720, 'steps': 9309, 'loss/train': 1.6228625774383545} 02/24/2022 13:37:16 - INFO - codeparrot_training - Step 9310: {'lr': 0.00047192874099762246, 'samples': 4767232, 'steps': 9310, 'loss/train': 2.5432825088500977} 02/24/2022 13:37:20 - INFO - codeparrot_training - Step 9311: {'lr': 0.00047192120735183485, 'samples': 4767744, 'steps': 9311, 'loss/train': 1.1503663063049316} 02/24/2022 13:37:25 - INFO - codeparrot_training - Step 9312: {'lr': 0.0004719136727554072, 'samples': 4768256, 'steps': 9312, 'loss/train': 2.543571949005127} 02/24/2022 13:37:28 - INFO - codeparrot_training - Step 9313: {'lr': 0.0004719061372083719, 'samples': 4768768, 'steps': 9313, 'loss/train': 2.7874090671539307} 02/24/2022 13:37:34 - INFO - codeparrot_training - Step 9314: {'lr': 0.00047189860071076114, 'samples': 4769280, 'steps': 9314, 'loss/train': 1.3268343210220337} 02/24/2022 13:37:40 - INFO - codeparrot_training - Step 9315: {'lr': 0.00047189106326260723, 'samples': 4769792, 'steps': 9315, 'loss/train': 3.752823829650879} 02/24/2022 13:37:43 - INFO - codeparrot_training - Step 9316: {'lr': 0.0004718835248639425, 'samples': 4770304, 'steps': 9316, 'loss/train': 2.344468116760254} 02/24/2022 13:37:46 - INFO - codeparrot_training - Step 9317: {'lr': 0.0004718759855147992, 'samples': 4770816, 'steps': 9317, 'loss/train': 2.220879554748535} 02/24/2022 13:37:52 - INFO - codeparrot_training - Step 9318: {'lr': 0.00047186844521520955, 'samples': 4771328, 'steps': 9318, 'loss/train': 2.433905839920044} 02/24/2022 13:37:58 - INFO - codeparrot_training - Step 9319: {'lr': 0.000471860903965206, 'samples': 4771840, 'steps': 9319, 'loss/train': 2.2181622982025146} 02/24/2022 13:38:01 - INFO - codeparrot_training - Step 9320: {'lr': 0.00047185336176482084, 'samples': 4772352, 'steps': 9320, 'loss/train': 1.553859829902649} 02/24/2022 13:38:04 - INFO - codeparrot_training - Step 9321: {'lr': 0.0004718458186140863, 'samples': 4772864, 'steps': 9321, 'loss/train': 1.4725899696350098} 02/24/2022 13:38:11 - INFO - codeparrot_training - Step 9322: {'lr': 0.0004718382745130346, 'samples': 4773376, 'steps': 9322, 'loss/train': 2.2867565155029297} 02/24/2022 13:38:14 - INFO - codeparrot_training - Step 9323: {'lr': 0.0004718307294616983, 'samples': 4773888, 'steps': 9323, 'loss/train': 1.739013433456421} 02/24/2022 13:38:20 - INFO - codeparrot_training - Step 9324: {'lr': 0.00047182318346010953, 'samples': 4774400, 'steps': 9324, 'loss/train': 1.8712568283081055} 02/24/2022 13:38:23 - INFO - codeparrot_training - Step 9325: {'lr': 0.0004718156365083007, 'samples': 4774912, 'steps': 9325, 'loss/train': 3.318185806274414} 02/24/2022 13:38:29 - INFO - codeparrot_training - Step 9326: {'lr': 0.0004718080886063041, 'samples': 4775424, 'steps': 9326, 'loss/train': 1.873559594154358} 02/24/2022 13:38:32 - INFO - codeparrot_training - Step 9327: {'lr': 0.00047180053975415216, 'samples': 4775936, 'steps': 9327, 'loss/train': 3.0185303688049316} 02/24/2022 13:38:38 - INFO - codeparrot_training - Step 9328: {'lr': 0.00047179298995187705, 'samples': 4776448, 'steps': 9328, 'loss/train': 1.2814826965332031} 02/24/2022 13:38:41 - INFO - codeparrot_training - Step 9329: {'lr': 0.00047178543919951124, 'samples': 4776960, 'steps': 9329, 'loss/train': 1.9439716339111328} 02/24/2022 13:38:47 - INFO - codeparrot_training - Step 9330: {'lr': 0.000471777887497087, 'samples': 4777472, 'steps': 9330, 'loss/train': 1.9582996368408203} 02/24/2022 13:38:50 - INFO - codeparrot_training - Step 9331: {'lr': 0.0004717703348446367, 'samples': 4777984, 'steps': 9331, 'loss/train': 1.719980239868164} 02/24/2022 13:38:56 - INFO - codeparrot_training - Step 9332: {'lr': 0.00047176278124219276, 'samples': 4778496, 'steps': 9332, 'loss/train': 2.2232604026794434} 02/24/2022 13:38:59 - INFO - codeparrot_training - Step 9333: {'lr': 0.0004717552266897874, 'samples': 4779008, 'steps': 9333, 'loss/train': 1.6723746061325073} 02/24/2022 13:39:05 - INFO - codeparrot_training - Step 9334: {'lr': 0.0004717476711874532, 'samples': 4779520, 'steps': 9334, 'loss/train': 3.317763328552246} 02/24/2022 13:39:09 - INFO - codeparrot_training - Step 9335: {'lr': 0.00047174011473522225, 'samples': 4780032, 'steps': 9335, 'loss/train': 2.113901376724243} 02/24/2022 13:39:14 - INFO - codeparrot_training - Step 9336: {'lr': 0.0004717325573331271, 'samples': 4780544, 'steps': 9336, 'loss/train': 4.578775405883789} 02/24/2022 13:39:18 - INFO - codeparrot_training - Step 9337: {'lr': 0.00047172499898120014, 'samples': 4781056, 'steps': 9337, 'loss/train': 1.9555407762527466} 02/24/2022 13:39:23 - INFO - codeparrot_training - Step 9338: {'lr': 0.0004717174396794737, 'samples': 4781568, 'steps': 9338, 'loss/train': 1.952053189277649} 02/24/2022 13:39:27 - INFO - codeparrot_training - Step 9339: {'lr': 0.00047170987942798004, 'samples': 4782080, 'steps': 9339, 'loss/train': 1.8725188970565796} 02/24/2022 13:39:32 - INFO - codeparrot_training - Step 9340: {'lr': 0.0004717023182267518, 'samples': 4782592, 'steps': 9340, 'loss/train': 1.8152172565460205} 02/24/2022 13:39:38 - INFO - codeparrot_training - Step 9341: {'lr': 0.00047169475607582113, 'samples': 4783104, 'steps': 9341, 'loss/train': 2.0683112144470215} 02/24/2022 13:39:41 - INFO - codeparrot_training - Step 9342: {'lr': 0.00047168719297522053, 'samples': 4783616, 'steps': 9342, 'loss/train': 1.3040565252304077} 02/24/2022 13:39:47 - INFO - codeparrot_training - Step 9343: {'lr': 0.0004716796289249824, 'samples': 4784128, 'steps': 9343, 'loss/train': 2.081977605819702} 02/24/2022 13:39:51 - INFO - codeparrot_training - Step 9344: {'lr': 0.0004716720639251392, 'samples': 4784640, 'steps': 9344, 'loss/train': 2.5326483249664307} 02/24/2022 13:39:56 - INFO - codeparrot_training - Step 9345: {'lr': 0.00047166449797572316, 'samples': 4785152, 'steps': 9345, 'loss/train': 2.550849676132202} 02/24/2022 13:40:00 - INFO - codeparrot_training - Step 9346: {'lr': 0.0004716569310767668, 'samples': 4785664, 'steps': 9346, 'loss/train': 1.447945237159729} 02/24/2022 13:40:06 - INFO - codeparrot_training - Step 9347: {'lr': 0.00047164936322830256, 'samples': 4786176, 'steps': 9347, 'loss/train': 2.047583818435669} 02/24/2022 13:40:09 - INFO - codeparrot_training - Step 9348: {'lr': 0.0004716417944303628, 'samples': 4786688, 'steps': 9348, 'loss/train': 2.749289035797119} 02/24/2022 13:40:14 - INFO - codeparrot_training - Step 9349: {'lr': 0.00047163422468298003, 'samples': 4787200, 'steps': 9349, 'loss/train': 2.25677227973938} 02/24/2022 13:40:18 - INFO - codeparrot_training - Step 9350: {'lr': 0.00047162665398618666, 'samples': 4787712, 'steps': 9350, 'loss/train': 1.3625988960266113} 02/24/2022 13:40:23 - INFO - codeparrot_training - Step 9351: {'lr': 0.00047161908234001496, 'samples': 4788224, 'steps': 9351, 'loss/train': 2.437257766723633} 02/24/2022 13:40:27 - INFO - codeparrot_training - Step 9352: {'lr': 0.0004716115097444975, 'samples': 4788736, 'steps': 9352, 'loss/train': 2.399075984954834} 02/24/2022 13:40:33 - INFO - codeparrot_training - Step 9353: {'lr': 0.0004716039361996668, 'samples': 4789248, 'steps': 9353, 'loss/train': 2.3500478267669678} 02/24/2022 13:40:37 - INFO - codeparrot_training - Step 9354: {'lr': 0.0004715963617055551, 'samples': 4789760, 'steps': 9354, 'loss/train': 2.6645054817199707} 02/24/2022 13:40:42 - INFO - codeparrot_training - Step 9355: {'lr': 0.00047158878626219505, 'samples': 4790272, 'steps': 9355, 'loss/train': 2.704805850982666} 02/24/2022 13:40:46 - INFO - codeparrot_training - Step 9356: {'lr': 0.00047158120986961897, 'samples': 4790784, 'steps': 9356, 'loss/train': 2.1819796562194824} 02/24/2022 13:40:51 - INFO - codeparrot_training - Step 9357: {'lr': 0.0004715736325278593, 'samples': 4791296, 'steps': 9357, 'loss/train': 1.9590816497802734} 02/24/2022 13:40:55 - INFO - codeparrot_training - Step 9358: {'lr': 0.0004715660542369485, 'samples': 4791808, 'steps': 9358, 'loss/train': 2.725348711013794} 02/24/2022 13:41:00 - INFO - codeparrot_training - Step 9359: {'lr': 0.0004715584749969192, 'samples': 4792320, 'steps': 9359, 'loss/train': 2.174389362335205} 02/24/2022 13:41:04 - INFO - codeparrot_training - Step 9360: {'lr': 0.00047155089480780364, 'samples': 4792832, 'steps': 9360, 'loss/train': 2.9488959312438965} 02/24/2022 13:41:09 - INFO - codeparrot_training - Step 9361: {'lr': 0.0004715433136696345, 'samples': 4793344, 'steps': 9361, 'loss/train': 1.9647334814071655} 02/24/2022 13:41:13 - INFO - codeparrot_training - Step 9362: {'lr': 0.0004715357315824441, 'samples': 4793856, 'steps': 9362, 'loss/train': 1.849334716796875} 02/24/2022 13:41:18 - INFO - codeparrot_training - Step 9363: {'lr': 0.00047152814854626494, 'samples': 4794368, 'steps': 9363, 'loss/train': 2.8237733840942383} 02/24/2022 13:41:22 - INFO - codeparrot_training - Step 9364: {'lr': 0.0004715205645611296, 'samples': 4794880, 'steps': 9364, 'loss/train': 2.231487274169922} 02/24/2022 13:41:27 - INFO - codeparrot_training - Step 9365: {'lr': 0.00047151297962707054, 'samples': 4795392, 'steps': 9365, 'loss/train': 2.5442285537719727} 02/24/2022 13:41:31 - INFO - codeparrot_training - Step 9366: {'lr': 0.00047150539374412004, 'samples': 4795904, 'steps': 9366, 'loss/train': 1.7987245321273804} 02/24/2022 13:41:36 - INFO - codeparrot_training - Step 9367: {'lr': 0.0004714978069123109, 'samples': 4796416, 'steps': 9367, 'loss/train': 1.5481021404266357} 02/24/2022 13:41:40 - INFO - codeparrot_training - Step 9368: {'lr': 0.00047149021913167545, 'samples': 4796928, 'steps': 9368, 'loss/train': 0.6987714767456055} 02/24/2022 13:41:46 - INFO - codeparrot_training - Step 9369: {'lr': 0.00047148263040224626, 'samples': 4797440, 'steps': 9369, 'loss/train': 2.306771755218506} 02/24/2022 13:41:49 - INFO - codeparrot_training - Step 9370: {'lr': 0.00047147504072405575, 'samples': 4797952, 'steps': 9370, 'loss/train': 2.630359649658203} 02/24/2022 13:41:55 - INFO - codeparrot_training - Step 9371: {'lr': 0.0004714674500971366, 'samples': 4798464, 'steps': 9371, 'loss/train': 1.6505128145217896} 02/24/2022 13:41:58 - INFO - codeparrot_training - Step 9372: {'lr': 0.00047145985852152115, 'samples': 4798976, 'steps': 9372, 'loss/train': 2.2431480884552} 02/24/2022 13:42:04 - INFO - codeparrot_training - Step 9373: {'lr': 0.000471452265997242, 'samples': 4799488, 'steps': 9373, 'loss/train': 2.020749092102051} 02/24/2022 13:42:08 - INFO - codeparrot_training - Step 9374: {'lr': 0.00047144467252433164, 'samples': 4800000, 'steps': 9374, 'loss/train': 2.1209564208984375} 02/24/2022 13:42:13 - INFO - codeparrot_training - Step 9375: {'lr': 0.00047143707810282266, 'samples': 4800512, 'steps': 9375, 'loss/train': 0.8257538080215454} 02/24/2022 13:42:16 - INFO - codeparrot_training - Step 9376: {'lr': 0.0004714294827327475, 'samples': 4801024, 'steps': 9376, 'loss/train': 1.7988255023956299} 02/24/2022 13:42:23 - INFO - codeparrot_training - Step 9377: {'lr': 0.00047142188641413873, 'samples': 4801536, 'steps': 9377, 'loss/train': 3.4448342323303223} 02/24/2022 13:42:26 - INFO - codeparrot_training - Step 9378: {'lr': 0.000471414289147029, 'samples': 4802048, 'steps': 9378, 'loss/train': 2.1589090824127197} 02/24/2022 13:42:30 - INFO - codeparrot_training - Step 9379: {'lr': 0.00047140669093145073, 'samples': 4802560, 'steps': 9379, 'loss/train': 2.4292073249816895} 02/24/2022 13:42:36 - INFO - codeparrot_training - Step 9380: {'lr': 0.00047139909176743643, 'samples': 4803072, 'steps': 9380, 'loss/train': 3.0854930877685547} 02/24/2022 13:42:40 - INFO - codeparrot_training - Step 9381: {'lr': 0.0004713914916550188, 'samples': 4803584, 'steps': 9381, 'loss/train': 1.1514456272125244} 02/24/2022 13:42:45 - INFO - codeparrot_training - Step 9382: {'lr': 0.00047138389059423033, 'samples': 4804096, 'steps': 9382, 'loss/train': 2.2677462100982666} 02/24/2022 13:42:49 - INFO - codeparrot_training - Step 9383: {'lr': 0.0004713762885851035, 'samples': 4804608, 'steps': 9383, 'loss/train': 2.2468223571777344} 02/24/2022 13:42:54 - INFO - codeparrot_training - Step 9384: {'lr': 0.000471368685627671, 'samples': 4805120, 'steps': 9384, 'loss/train': 2.3884036540985107} 02/24/2022 13:42:58 - INFO - codeparrot_training - Step 9385: {'lr': 0.00047136108172196535, 'samples': 4805632, 'steps': 9385, 'loss/train': 2.0233447551727295} 02/24/2022 13:43:04 - INFO - codeparrot_training - Step 9386: {'lr': 0.00047135347686801907, 'samples': 4806144, 'steps': 9386, 'loss/train': 2.1385862827301025} 02/24/2022 13:43:07 - INFO - codeparrot_training - Step 9387: {'lr': 0.0004713458710658648, 'samples': 4806656, 'steps': 9387, 'loss/train': 1.24008047580719} 02/24/2022 13:43:13 - INFO - codeparrot_training - Step 9388: {'lr': 0.0004713382643155351, 'samples': 4807168, 'steps': 9388, 'loss/train': 2.990281343460083} 02/24/2022 13:43:16 - INFO - codeparrot_training - Step 9389: {'lr': 0.00047133065661706254, 'samples': 4807680, 'steps': 9389, 'loss/train': 2.3468973636627197} 02/24/2022 13:43:22 - INFO - codeparrot_training - Step 9390: {'lr': 0.00047132304797047975, 'samples': 4808192, 'steps': 9390, 'loss/train': 2.8881564140319824} 02/24/2022 13:43:26 - INFO - codeparrot_training - Step 9391: {'lr': 0.00047131543837581935, 'samples': 4808704, 'steps': 9391, 'loss/train': 1.9530246257781982} 02/24/2022 13:43:31 - INFO - codeparrot_training - Step 9392: {'lr': 0.0004713078278331138, 'samples': 4809216, 'steps': 9392, 'loss/train': 2.6750190258026123} 02/24/2022 13:43:35 - INFO - codeparrot_training - Step 9393: {'lr': 0.00047130021634239584, 'samples': 4809728, 'steps': 9393, 'loss/train': 1.4577772617340088} 02/24/2022 13:43:40 - INFO - codeparrot_training - Step 9394: {'lr': 0.000471292603903698, 'samples': 4810240, 'steps': 9394, 'loss/train': 1.9654873609542847} 02/24/2022 13:43:44 - INFO - codeparrot_training - Step 9395: {'lr': 0.00047128499051705296, 'samples': 4810752, 'steps': 9395, 'loss/train': 3.8678324222564697} 02/24/2022 13:43:49 - INFO - codeparrot_training - Step 9396: {'lr': 0.00047127737618249323, 'samples': 4811264, 'steps': 9396, 'loss/train': 2.4186582565307617} 02/24/2022 13:43:53 - INFO - codeparrot_training - Step 9397: {'lr': 0.00047126976090005153, 'samples': 4811776, 'steps': 9397, 'loss/train': 3.3421342372894287} 02/24/2022 13:43:58 - INFO - codeparrot_training - Step 9398: {'lr': 0.00047126214466976034, 'samples': 4812288, 'steps': 9398, 'loss/train': 2.1782310009002686} 02/24/2022 13:44:02 - INFO - codeparrot_training - Step 9399: {'lr': 0.0004712545274916525, 'samples': 4812800, 'steps': 9399, 'loss/train': 2.024573802947998} 02/24/2022 13:44:08 - INFO - codeparrot_training - Step 9400: {'lr': 0.00047124690936576046, 'samples': 4813312, 'steps': 9400, 'loss/train': 2.6351754665374756} 02/24/2022 13:44:11 - INFO - codeparrot_training - Step 9401: {'lr': 0.000471239290292117, 'samples': 4813824, 'steps': 9401, 'loss/train': 2.5747451782226562} 02/24/2022 13:44:17 - INFO - codeparrot_training - Step 9402: {'lr': 0.00047123167027075455, 'samples': 4814336, 'steps': 9402, 'loss/train': 1.8039214611053467} 02/24/2022 13:44:20 - INFO - codeparrot_training - Step 9403: {'lr': 0.0004712240493017059, 'samples': 4814848, 'steps': 9403, 'loss/train': 2.802110433578491} 02/24/2022 13:44:26 - INFO - codeparrot_training - Step 9404: {'lr': 0.0004712164273850037, 'samples': 4815360, 'steps': 9404, 'loss/train': 2.291334629058838} 02/24/2022 13:44:29 - INFO - codeparrot_training - Step 9405: {'lr': 0.0004712088045206806, 'samples': 4815872, 'steps': 9405, 'loss/train': 2.223386764526367} 02/24/2022 13:44:35 - INFO - codeparrot_training - Step 9406: {'lr': 0.00047120118070876916, 'samples': 4816384, 'steps': 9406, 'loss/train': 2.235356330871582} 02/24/2022 13:44:38 - INFO - codeparrot_training - Step 9407: {'lr': 0.0004711935559493021, 'samples': 4816896, 'steps': 9407, 'loss/train': 1.396200180053711} 02/24/2022 13:44:44 - INFO - codeparrot_training - Step 9408: {'lr': 0.00047118593024231216, 'samples': 4817408, 'steps': 9408, 'loss/train': 2.0703699588775635} 02/24/2022 13:44:47 - INFO - codeparrot_training - Step 9409: {'lr': 0.00047117830358783184, 'samples': 4817920, 'steps': 9409, 'loss/train': 2.2124409675598145} 02/24/2022 13:44:53 - INFO - codeparrot_training - Step 9410: {'lr': 0.0004711706759858939, 'samples': 4818432, 'steps': 9410, 'loss/train': 2.479804039001465} 02/24/2022 13:44:56 - INFO - codeparrot_training - Step 9411: {'lr': 0.0004711630474365311, 'samples': 4818944, 'steps': 9411, 'loss/train': 2.5117905139923096} 02/24/2022 13:45:02 - INFO - codeparrot_training - Step 9412: {'lr': 0.000471155417939776, 'samples': 4819456, 'steps': 9412, 'loss/train': 1.632948875427246} 02/24/2022 13:45:05 - INFO - codeparrot_training - Step 9413: {'lr': 0.00047114778749566123, 'samples': 4819968, 'steps': 9413, 'loss/train': 1.818543553352356} 02/24/2022 13:45:11 - INFO - codeparrot_training - Step 9414: {'lr': 0.00047114015610421966, 'samples': 4820480, 'steps': 9414, 'loss/train': 2.543210744857788} 02/24/2022 13:45:14 - INFO - codeparrot_training - Step 9415: {'lr': 0.00047113252376548387, 'samples': 4820992, 'steps': 9415, 'loss/train': 2.46437931060791} 02/24/2022 13:45:21 - INFO - codeparrot_training - Step 9416: {'lr': 0.00047112489047948655, 'samples': 4821504, 'steps': 9416, 'loss/train': 2.1986892223358154} 02/24/2022 13:45:24 - INFO - codeparrot_training - Step 9417: {'lr': 0.0004711172562462604, 'samples': 4822016, 'steps': 9417, 'loss/train': 1.3839813470840454} 02/24/2022 13:45:30 - INFO - codeparrot_training - Step 9418: {'lr': 0.0004711096210658381, 'samples': 4822528, 'steps': 9418, 'loss/train': 1.9671071767807007} 02/24/2022 13:45:33 - INFO - codeparrot_training - Step 9419: {'lr': 0.0004711019849382525, 'samples': 4823040, 'steps': 9419, 'loss/train': 2.378391742706299} 02/24/2022 13:45:39 - INFO - codeparrot_training - Step 9420: {'lr': 0.0004710943478635361, 'samples': 4823552, 'steps': 9420, 'loss/train': 2.1312999725341797} 02/24/2022 13:45:42 - INFO - codeparrot_training - Step 9421: {'lr': 0.00047108670984172176, 'samples': 4824064, 'steps': 9421, 'loss/train': 0.6496623158454895} 02/24/2022 13:45:48 - INFO - codeparrot_training - Step 9422: {'lr': 0.00047107907087284216, 'samples': 4824576, 'steps': 9422, 'loss/train': 1.5175443887710571} 02/24/2022 13:45:51 - INFO - codeparrot_training - Step 9423: {'lr': 0.00047107143095693007, 'samples': 4825088, 'steps': 9423, 'loss/train': 1.9888180494308472} 02/24/2022 13:45:57 - INFO - codeparrot_training - Step 9424: {'lr': 0.0004710637900940181, 'samples': 4825600, 'steps': 9424, 'loss/train': 1.5517288446426392} 02/24/2022 13:46:00 - INFO - codeparrot_training - Step 9425: {'lr': 0.00047105614828413906, 'samples': 4826112, 'steps': 9425, 'loss/train': 2.545207977294922} 02/24/2022 13:46:07 - INFO - codeparrot_training - Step 9426: {'lr': 0.0004710485055273257, 'samples': 4826624, 'steps': 9426, 'loss/train': 2.6607866287231445} 02/24/2022 13:46:11 - INFO - codeparrot_training - Step 9427: {'lr': 0.00047104086182361073, 'samples': 4827136, 'steps': 9427, 'loss/train': 0.6701631546020508} 02/24/2022 13:46:16 - INFO - codeparrot_training - Step 9428: {'lr': 0.00047103321717302684, 'samples': 4827648, 'steps': 9428, 'loss/train': 1.7967125177383423} 02/24/2022 13:46:20 - INFO - codeparrot_training - Step 9429: {'lr': 0.00047102557157560686, 'samples': 4828160, 'steps': 9429, 'loss/train': 2.142204761505127} 02/24/2022 13:46:25 - INFO - codeparrot_training - Step 9430: {'lr': 0.00047101792503138353, 'samples': 4828672, 'steps': 9430, 'loss/train': 1.7491512298583984} 02/24/2022 13:46:29 - INFO - codeparrot_training - Step 9431: {'lr': 0.0004710102775403896, 'samples': 4829184, 'steps': 9431, 'loss/train': 2.050910472869873} 02/24/2022 13:46:34 - INFO - codeparrot_training - Step 9432: {'lr': 0.00047100262910265787, 'samples': 4829696, 'steps': 9432, 'loss/train': 1.6814892292022705} 02/24/2022 13:46:38 - INFO - codeparrot_training - Step 9433: {'lr': 0.00047099497971822096, 'samples': 4830208, 'steps': 9433, 'loss/train': 1.6766963005065918} 02/24/2022 13:46:43 - INFO - codeparrot_training - Step 9434: {'lr': 0.00047098732938711174, 'samples': 4830720, 'steps': 9434, 'loss/train': 3.0554845333099365} 02/24/2022 13:46:47 - INFO - codeparrot_training - Step 9435: {'lr': 0.00047097967810936305, 'samples': 4831232, 'steps': 9435, 'loss/train': 1.290590763092041} 02/24/2022 13:46:53 - INFO - codeparrot_training - Step 9436: {'lr': 0.00047097202588500747, 'samples': 4831744, 'steps': 9436, 'loss/train': 3.68501615524292} 02/24/2022 13:46:56 - INFO - codeparrot_training - Step 9437: {'lr': 0.000470964372714078, 'samples': 4832256, 'steps': 9437, 'loss/train': 2.2355496883392334} 02/24/2022 13:47:02 - INFO - codeparrot_training - Step 9438: {'lr': 0.00047095671859660726, 'samples': 4832768, 'steps': 9438, 'loss/train': 1.6161638498306274} 02/24/2022 13:47:05 - INFO - codeparrot_training - Step 9439: {'lr': 0.0004709490635326281, 'samples': 4833280, 'steps': 9439, 'loss/train': 1.6555230617523193} 02/24/2022 13:47:11 - INFO - codeparrot_training - Step 9440: {'lr': 0.0004709414075221734, 'samples': 4833792, 'steps': 9440, 'loss/train': 2.132615327835083} 02/24/2022 13:47:14 - INFO - codeparrot_training - Step 9441: {'lr': 0.00047093375056527577, 'samples': 4834304, 'steps': 9441, 'loss/train': 1.8542178869247437} 02/24/2022 13:47:20 - INFO - codeparrot_training - Step 9442: {'lr': 0.0004709260926619682, 'samples': 4834816, 'steps': 9442, 'loss/train': 2.3685977458953857} 02/24/2022 13:47:23 - INFO - codeparrot_training - Step 9443: {'lr': 0.00047091843381228326, 'samples': 4835328, 'steps': 9443, 'loss/train': 2.610386371612549} 02/24/2022 13:47:29 - INFO - codeparrot_training - Step 9444: {'lr': 0.000470910774016254, 'samples': 4835840, 'steps': 9444, 'loss/train': 3.0459628105163574} 02/24/2022 13:47:32 - INFO - codeparrot_training - Step 9445: {'lr': 0.0004709031132739131, 'samples': 4836352, 'steps': 9445, 'loss/train': 2.9118127822875977} 02/24/2022 13:47:38 - INFO - codeparrot_training - Step 9446: {'lr': 0.0004708954515852934, 'samples': 4836864, 'steps': 9446, 'loss/train': 2.26836895942688} 02/24/2022 13:47:42 - INFO - codeparrot_training - Step 9447: {'lr': 0.00047088778895042774, 'samples': 4837376, 'steps': 9447, 'loss/train': 2.414227247238159} 02/24/2022 13:47:48 - INFO - codeparrot_training - Step 9448: {'lr': 0.000470880125369349, 'samples': 4837888, 'steps': 9448, 'loss/train': 2.544114589691162} 02/24/2022 13:47:52 - INFO - codeparrot_training - Step 9449: {'lr': 0.0004708724608420898, 'samples': 4838400, 'steps': 9449, 'loss/train': 2.782829523086548} 02/24/2022 13:47:55 - INFO - codeparrot_training - Step 9450: {'lr': 0.0004708647953686832, 'samples': 4838912, 'steps': 9450, 'loss/train': 2.9846577644348145} 02/24/2022 13:48:01 - INFO - codeparrot_training - Step 9451: {'lr': 0.000470857128949162, 'samples': 4839424, 'steps': 9451, 'loss/train': 1.1818568706512451} 02/24/2022 13:48:04 - INFO - codeparrot_training - Step 9452: {'lr': 0.0004708494615835589, 'samples': 4839936, 'steps': 9452, 'loss/train': 2.4610939025878906} 02/24/2022 13:48:10 - INFO - codeparrot_training - Step 9453: {'lr': 0.0004708417932719068, 'samples': 4840448, 'steps': 9453, 'loss/train': 2.3494925498962402} 02/24/2022 13:48:13 - INFO - codeparrot_training - Step 9454: {'lr': 0.0004708341240142387, 'samples': 4840960, 'steps': 9454, 'loss/train': 2.653748035430908} 02/24/2022 13:48:19 - INFO - codeparrot_training - Step 9455: {'lr': 0.0004708264538105873, 'samples': 4841472, 'steps': 9455, 'loss/train': 2.767214059829712} 02/24/2022 13:48:22 - INFO - codeparrot_training - Step 9456: {'lr': 0.0004708187826609854, 'samples': 4841984, 'steps': 9456, 'loss/train': 2.1033360958099365} 02/24/2022 13:48:28 - INFO - codeparrot_training - Step 9457: {'lr': 0.0004708111105654661, 'samples': 4842496, 'steps': 9457, 'loss/train': 2.3968472480773926} 02/24/2022 13:48:31 - INFO - codeparrot_training - Step 9458: {'lr': 0.000470803437524062, 'samples': 4843008, 'steps': 9458, 'loss/train': 2.016186475753784} 02/24/2022 13:48:37 - INFO - codeparrot_training - Step 9459: {'lr': 0.00047079576353680614, 'samples': 4843520, 'steps': 9459, 'loss/train': 2.1598775386810303} 02/24/2022 13:48:40 - INFO - codeparrot_training - Step 9460: {'lr': 0.0004707880886037314, 'samples': 4844032, 'steps': 9460, 'loss/train': 2.0636279582977295} 02/24/2022 13:48:46 - INFO - codeparrot_training - Step 9461: {'lr': 0.00047078041272487046, 'samples': 4844544, 'steps': 9461, 'loss/train': 2.281644582748413} 02/24/2022 13:48:52 - INFO - codeparrot_training - Step 9462: {'lr': 0.00047077273590025637, 'samples': 4845056, 'steps': 9462, 'loss/train': 1.195705771446228} 02/24/2022 13:48:55 - INFO - codeparrot_training - Step 9463: {'lr': 0.00047076505812992204, 'samples': 4845568, 'steps': 9463, 'loss/train': 2.2360119819641113} 02/24/2022 13:48:59 - INFO - codeparrot_training - Step 9464: {'lr': 0.0004707573794139003, 'samples': 4846080, 'steps': 9464, 'loss/train': 2.893397331237793} 02/24/2022 13:49:05 - INFO - codeparrot_training - Step 9465: {'lr': 0.00047074969975222406, 'samples': 4846592, 'steps': 9465, 'loss/train': 2.2125468254089355} 02/24/2022 13:49:09 - INFO - codeparrot_training - Step 9466: {'lr': 0.0004707420191449261, 'samples': 4847104, 'steps': 9466, 'loss/train': 2.1745247840881348} 02/24/2022 13:49:14 - INFO - codeparrot_training - Step 9467: {'lr': 0.0004707343375920395, 'samples': 4847616, 'steps': 9467, 'loss/train': 2.440685272216797} 02/24/2022 13:49:18 - INFO - codeparrot_training - Step 9468: {'lr': 0.0004707266550935971, 'samples': 4848128, 'steps': 9468, 'loss/train': 2.183720827102661} 02/24/2022 13:49:23 - INFO - codeparrot_training - Step 9469: {'lr': 0.00047071897164963175, 'samples': 4848640, 'steps': 9469, 'loss/train': 0.6161959171295166} 02/24/2022 13:49:29 - INFO - codeparrot_training - Step 9470: {'lr': 0.00047071128726017643, 'samples': 4849152, 'steps': 9470, 'loss/train': 3.4612972736358643} 02/24/2022 13:49:32 - INFO - codeparrot_training - Step 9471: {'lr': 0.0004707036019252641, 'samples': 4849664, 'steps': 9471, 'loss/train': 2.284473419189453} 02/24/2022 13:49:35 - INFO - codeparrot_training - Step 9472: {'lr': 0.00047069591564492753, 'samples': 4850176, 'steps': 9472, 'loss/train': 2.2644498348236084} 02/24/2022 13:49:42 - INFO - codeparrot_training - Step 9473: {'lr': 0.00047068822841919976, 'samples': 4850688, 'steps': 9473, 'loss/train': 1.4850801229476929} 02/24/2022 13:49:46 - INFO - codeparrot_training - Step 9474: {'lr': 0.0004706805402481137, 'samples': 4851200, 'steps': 9474, 'loss/train': 2.9515137672424316} 02/24/2022 13:49:51 - INFO - codeparrot_training - Step 9475: {'lr': 0.00047067285113170233, 'samples': 4851712, 'steps': 9475, 'loss/train': 1.6336843967437744} 02/24/2022 13:49:55 - INFO - codeparrot_training - Step 9476: {'lr': 0.0004706651610699985, 'samples': 4852224, 'steps': 9476, 'loss/train': 2.651027202606201} 02/24/2022 13:50:00 - INFO - codeparrot_training - Step 9477: {'lr': 0.0004706574700630352, 'samples': 4852736, 'steps': 9477, 'loss/train': 2.1046602725982666} 02/24/2022 13:50:06 - INFO - codeparrot_training - Step 9478: {'lr': 0.0004706497781108453, 'samples': 4853248, 'steps': 9478, 'loss/train': 1.5995279550552368} 02/24/2022 13:50:09 - INFO - codeparrot_training - Step 9479: {'lr': 0.00047064208521346184, 'samples': 4853760, 'steps': 9479, 'loss/train': 3.694612979888916} 02/24/2022 13:50:15 - INFO - codeparrot_training - Step 9480: {'lr': 0.0004706343913709178, 'samples': 4854272, 'steps': 9480, 'loss/train': 2.999093532562256} 02/24/2022 13:50:18 - INFO - codeparrot_training - Step 9481: {'lr': 0.0004706266965832461, 'samples': 4854784, 'steps': 9481, 'loss/train': 2.6196208000183105} 02/24/2022 13:50:22 - INFO - codeparrot_training - Step 9482: {'lr': 0.0004706190008504796, 'samples': 4855296, 'steps': 9482, 'loss/train': 2.915841579437256} 02/24/2022 13:50:28 - INFO - codeparrot_training - Step 9483: {'lr': 0.00047061130417265143, 'samples': 4855808, 'steps': 9483, 'loss/train': 2.063542127609253} 02/24/2022 13:50:34 - INFO - codeparrot_training - Step 9484: {'lr': 0.0004706036065497944, 'samples': 4856320, 'steps': 9484, 'loss/train': 2.0744080543518066} 02/24/2022 13:50:37 - INFO - codeparrot_training - Step 9485: {'lr': 0.0004705959079819416, 'samples': 4856832, 'steps': 9485, 'loss/train': 0.7285880446434021} 02/24/2022 13:50:41 - INFO - codeparrot_training - Step 9486: {'lr': 0.0004705882084691261, 'samples': 4857344, 'steps': 9486, 'loss/train': 3.391324520111084} 02/24/2022 13:50:46 - INFO - codeparrot_training - Step 9487: {'lr': 0.00047058050801138064, 'samples': 4857856, 'steps': 9487, 'loss/train': 3.2877824306488037} 02/24/2022 13:50:50 - INFO - codeparrot_training - Step 9488: {'lr': 0.00047057280660873835, 'samples': 4858368, 'steps': 9488, 'loss/train': 2.6793205738067627} 02/24/2022 13:50:55 - INFO - codeparrot_training - Step 9489: {'lr': 0.0004705651042612322, 'samples': 4858880, 'steps': 9489, 'loss/train': 2.9948320388793945} 02/24/2022 13:50:59 - INFO - codeparrot_training - Step 9490: {'lr': 0.00047055740096889516, 'samples': 4859392, 'steps': 9490, 'loss/train': 2.651505470275879} 02/24/2022 13:51:04 - INFO - codeparrot_training - Step 9491: {'lr': 0.0004705496967317603, 'samples': 4859904, 'steps': 9491, 'loss/train': 1.615960717201233} 02/24/2022 13:51:08 - INFO - codeparrot_training - Step 9492: {'lr': 0.0004705419915498605, 'samples': 4860416, 'steps': 9492, 'loss/train': 2.577371835708618} 02/24/2022 13:51:14 - INFO - codeparrot_training - Step 9493: {'lr': 0.0004705342854232288, 'samples': 4860928, 'steps': 9493, 'loss/train': 2.549577474594116} 02/24/2022 13:51:17 - INFO - codeparrot_training - Step 9494: {'lr': 0.00047052657835189836, 'samples': 4861440, 'steps': 9494, 'loss/train': 0.5594264268875122} 02/24/2022 13:51:24 - INFO - codeparrot_training - Step 9495: {'lr': 0.00047051887033590205, 'samples': 4861952, 'steps': 9495, 'loss/train': 3.7060985565185547} 02/24/2022 13:51:27 - INFO - codeparrot_training - Step 9496: {'lr': 0.00047051116137527296, 'samples': 4862464, 'steps': 9496, 'loss/train': 2.390371561050415} 02/24/2022 13:51:33 - INFO - codeparrot_training - Step 9497: {'lr': 0.000470503451470044, 'samples': 4862976, 'steps': 9497, 'loss/train': 2.9564828872680664} 02/24/2022 13:51:36 - INFO - codeparrot_training - Step 9498: {'lr': 0.00047049574062024837, 'samples': 4863488, 'steps': 9498, 'loss/train': 2.4608426094055176} 02/24/2022 13:51:42 - INFO - codeparrot_training - Step 9499: {'lr': 0.0004704880288259189, 'samples': 4864000, 'steps': 9499, 'loss/train': 1.6270740032196045} 02/24/2022 13:51:45 - INFO - codeparrot_training - Step 9500: {'lr': 0.00047048031608708875, 'samples': 4864512, 'steps': 9500, 'loss/train': 0.9599283933639526} 02/24/2022 13:51:51 - INFO - codeparrot_training - Step 9501: {'lr': 0.00047047260240379096, 'samples': 4865024, 'steps': 9501, 'loss/train': 1.2787842750549316} 02/24/2022 13:51:54 - INFO - codeparrot_training - Step 9502: {'lr': 0.00047046488777605853, 'samples': 4865536, 'steps': 9502, 'loss/train': 1.7225559949874878} 02/24/2022 13:52:00 - INFO - codeparrot_training - Step 9503: {'lr': 0.0004704571722039246, 'samples': 4866048, 'steps': 9503, 'loss/train': 8.059077262878418} 02/24/2022 13:52:03 - INFO - codeparrot_training - Step 9504: {'lr': 0.00047044945568742205, 'samples': 4866560, 'steps': 9504, 'loss/train': 2.1754374504089355} 02/24/2022 13:52:09 - INFO - codeparrot_training - Step 9505: {'lr': 0.0004704417382265841, 'samples': 4867072, 'steps': 9505, 'loss/train': 2.2936434745788574} 02/24/2022 13:52:13 - INFO - codeparrot_training - Step 9506: {'lr': 0.0004704340198214437, 'samples': 4867584, 'steps': 9506, 'loss/train': 2.381317615509033} 02/24/2022 13:52:18 - INFO - codeparrot_training - Step 9507: {'lr': 0.00047042630047203394, 'samples': 4868096, 'steps': 9507, 'loss/train': 1.6862726211547852} 02/24/2022 13:52:22 - INFO - codeparrot_training - Step 9508: {'lr': 0.0004704185801783879, 'samples': 4868608, 'steps': 9508, 'loss/train': 1.9814400672912598} 02/24/2022 13:52:28 - INFO - codeparrot_training - Step 9509: {'lr': 0.0004704108589405387, 'samples': 4869120, 'steps': 9509, 'loss/train': 2.798516035079956} 02/24/2022 13:52:31 - INFO - codeparrot_training - Step 9510: {'lr': 0.0004704031367585193, 'samples': 4869632, 'steps': 9510, 'loss/train': 1.804969072341919} 02/24/2022 13:52:37 - INFO - codeparrot_training - Step 9511: {'lr': 0.0004703954136323629, 'samples': 4870144, 'steps': 9511, 'loss/train': 1.420964241027832} 02/24/2022 13:52:40 - INFO - codeparrot_training - Step 9512: {'lr': 0.0004703876895621025, 'samples': 4870656, 'steps': 9512, 'loss/train': 1.7509866952896118} 02/24/2022 13:52:46 - INFO - codeparrot_training - Step 9513: {'lr': 0.00047037996454777134, 'samples': 4871168, 'steps': 9513, 'loss/train': 2.491886854171753} 02/24/2022 13:52:49 - INFO - codeparrot_training - Step 9514: {'lr': 0.00047037223858940224, 'samples': 4871680, 'steps': 9514, 'loss/train': 1.3904900550842285} 02/24/2022 13:52:55 - INFO - codeparrot_training - Step 9515: {'lr': 0.00047036451168702855, 'samples': 4872192, 'steps': 9515, 'loss/train': 2.199233055114746} 02/24/2022 13:52:58 - INFO - codeparrot_training - Step 9516: {'lr': 0.0004703567838406832, 'samples': 4872704, 'steps': 9516, 'loss/train': 1.4684724807739258} 02/24/2022 13:53:04 - INFO - codeparrot_training - Step 9517: {'lr': 0.00047034905505039936, 'samples': 4873216, 'steps': 9517, 'loss/train': 2.099609375} 02/24/2022 13:53:07 - INFO - codeparrot_training - Step 9518: {'lr': 0.0004703413253162102, 'samples': 4873728, 'steps': 9518, 'loss/train': 2.147616386413574} 02/24/2022 13:53:13 - INFO - codeparrot_training - Step 9519: {'lr': 0.00047033359463814875, 'samples': 4874240, 'steps': 9519, 'loss/train': 2.0364086627960205} 02/24/2022 13:53:16 - INFO - codeparrot_training - Step 9520: {'lr': 0.00047032586301624804, 'samples': 4874752, 'steps': 9520, 'loss/train': 1.6985219717025757} 02/24/2022 13:53:22 - INFO - codeparrot_training - Step 9521: {'lr': 0.0004703181304505414, 'samples': 4875264, 'steps': 9521, 'loss/train': 1.8973575830459595} 02/24/2022 13:53:26 - INFO - codeparrot_training - Step 9522: {'lr': 0.0004703103969410618, 'samples': 4875776, 'steps': 9522, 'loss/train': 2.3894176483154297} 02/24/2022 13:53:31 - INFO - codeparrot_training - Step 9523: {'lr': 0.0004703026624878425, 'samples': 4876288, 'steps': 9523, 'loss/train': 2.59407639503479} 02/24/2022 13:53:34 - INFO - codeparrot_training - Step 9524: {'lr': 0.0004702949270909164, 'samples': 4876800, 'steps': 9524, 'loss/train': 3.0909056663513184} 02/24/2022 13:53:40 - INFO - codeparrot_training - Step 9525: {'lr': 0.0004702871907503169, 'samples': 4877312, 'steps': 9525, 'loss/train': 2.36826229095459} 02/24/2022 13:53:44 - INFO - codeparrot_training - Step 9526: {'lr': 0.000470279453466077, 'samples': 4877824, 'steps': 9526, 'loss/train': 2.572887897491455} 02/24/2022 13:53:49 - INFO - codeparrot_training - Step 9527: {'lr': 0.0004702717152382299, 'samples': 4878336, 'steps': 9527, 'loss/train': 2.394839286804199} 02/24/2022 13:53:53 - INFO - codeparrot_training - Step 9528: {'lr': 0.0004702639760668086, 'samples': 4878848, 'steps': 9528, 'loss/train': 2.151036500930786} 02/24/2022 13:53:58 - INFO - codeparrot_training - Step 9529: {'lr': 0.00047025623595184645, 'samples': 4879360, 'steps': 9529, 'loss/train': 2.4957728385925293} 02/24/2022 13:54:02 - INFO - codeparrot_training - Step 9530: {'lr': 0.0004702484948933765, 'samples': 4879872, 'steps': 9530, 'loss/train': 1.154759168624878} 02/24/2022 13:54:08 - INFO - codeparrot_training - Step 9531: {'lr': 0.000470240752891432, 'samples': 4880384, 'steps': 9531, 'loss/train': 2.056220769882202} 02/24/2022 13:54:11 - INFO - codeparrot_training - Step 9532: {'lr': 0.000470233009946046, 'samples': 4880896, 'steps': 9532, 'loss/train': 1.311429738998413} 02/24/2022 13:54:17 - INFO - codeparrot_training - Step 9533: {'lr': 0.0004702252660572517, 'samples': 4881408, 'steps': 9533, 'loss/train': 2.0231940746307373} 02/24/2022 13:54:20 - INFO - codeparrot_training - Step 9534: {'lr': 0.00047021752122508234, 'samples': 4881920, 'steps': 9534, 'loss/train': 2.327814817428589} 02/24/2022 13:54:26 - INFO - codeparrot_training - Step 9535: {'lr': 0.000470209775449571, 'samples': 4882432, 'steps': 9535, 'loss/train': 0.9399018883705139} 02/24/2022 13:54:29 - INFO - codeparrot_training - Step 9536: {'lr': 0.00047020202873075093, 'samples': 4882944, 'steps': 9536, 'loss/train': 1.3210853338241577} 02/24/2022 13:54:35 - INFO - codeparrot_training - Step 9537: {'lr': 0.0004701942810686552, 'samples': 4883456, 'steps': 9537, 'loss/train': 2.2215545177459717} 02/24/2022 13:54:40 - INFO - codeparrot_training - Step 9538: {'lr': 0.00047018653246331724, 'samples': 4883968, 'steps': 9538, 'loss/train': 2.9715187549591064} 02/24/2022 13:54:44 - INFO - codeparrot_training - Step 9539: {'lr': 0.00047017878291477, 'samples': 4884480, 'steps': 9539, 'loss/train': 2.44144344329834} 02/24/2022 13:54:50 - INFO - codeparrot_training - Step 9540: {'lr': 0.0004701710324230468, 'samples': 4884992, 'steps': 9540, 'loss/train': 2.4876208305358887} 02/24/2022 13:54:53 - INFO - codeparrot_training - Step 9541: {'lr': 0.00047016328098818086, 'samples': 4885504, 'steps': 9541, 'loss/train': 2.0679149627685547} 02/24/2022 13:54:59 - INFO - codeparrot_training - Step 9542: {'lr': 0.00047015552861020524, 'samples': 4886016, 'steps': 9542, 'loss/train': 1.988111138343811} 02/24/2022 13:55:02 - INFO - codeparrot_training - Step 9543: {'lr': 0.00047014777528915327, 'samples': 4886528, 'steps': 9543, 'loss/train': 1.9410239458084106} 02/24/2022 13:55:08 - INFO - codeparrot_training - Step 9544: {'lr': 0.0004701400210250581, 'samples': 4887040, 'steps': 9544, 'loss/train': 1.2016911506652832} 02/24/2022 13:55:11 - INFO - codeparrot_training - Step 9545: {'lr': 0.00047013226581795305, 'samples': 4887552, 'steps': 9545, 'loss/train': 2.673309087753296} 02/24/2022 13:55:17 - INFO - codeparrot_training - Step 9546: {'lr': 0.00047012450966787126, 'samples': 4888064, 'steps': 9546, 'loss/train': 2.13564133644104} 02/24/2022 13:55:20 - INFO - codeparrot_training - Step 9547: {'lr': 0.000470116752574846, 'samples': 4888576, 'steps': 9547, 'loss/train': 1.6354318857192993} 02/24/2022 13:55:26 - INFO - codeparrot_training - Step 9548: {'lr': 0.0004701089945389104, 'samples': 4889088, 'steps': 9548, 'loss/train': 2.253312349319458} 02/24/2022 13:55:29 - INFO - codeparrot_training - Step 9549: {'lr': 0.00047010123556009774, 'samples': 4889600, 'steps': 9549, 'loss/train': 2.0510005950927734} 02/24/2022 13:55:35 - INFO - codeparrot_training - Step 9550: {'lr': 0.0004700934756384413, 'samples': 4890112, 'steps': 9550, 'loss/train': 2.5258548259735107} 02/24/2022 13:55:39 - INFO - codeparrot_training - Step 9551: {'lr': 0.00047008571477397435, 'samples': 4890624, 'steps': 9551, 'loss/train': 2.133078098297119} 02/24/2022 13:55:44 - INFO - codeparrot_training - Step 9552: {'lr': 0.00047007795296673006, 'samples': 4891136, 'steps': 9552, 'loss/train': 1.5449477434158325} 02/24/2022 13:55:48 - INFO - codeparrot_training - Step 9553: {'lr': 0.00047007019021674167, 'samples': 4891648, 'steps': 9553, 'loss/train': 2.955099582672119} 02/24/2022 13:55:54 - INFO - codeparrot_training - Step 9554: {'lr': 0.0004700624265240425, 'samples': 4892160, 'steps': 9554, 'loss/train': 1.851338267326355} 02/24/2022 13:55:57 - INFO - codeparrot_training - Step 9555: {'lr': 0.00047005466188866575, 'samples': 4892672, 'steps': 9555, 'loss/train': 2.10430908203125} 02/24/2022 13:56:03 - INFO - codeparrot_training - Step 9556: {'lr': 0.00047004689631064474, 'samples': 4893184, 'steps': 9556, 'loss/train': 1.1847878694534302} 02/24/2022 13:56:06 - INFO - codeparrot_training - Step 9557: {'lr': 0.00047003912979001267, 'samples': 4893696, 'steps': 9557, 'loss/train': 0.5048022866249084} 02/24/2022 13:56:12 - INFO - codeparrot_training - Step 9558: {'lr': 0.0004700313623268028, 'samples': 4894208, 'steps': 9558, 'loss/train': 1.9121215343475342} 02/24/2022 13:56:15 - INFO - codeparrot_training - Step 9559: {'lr': 0.00047002359392104854, 'samples': 4894720, 'steps': 9559, 'loss/train': 1.4153485298156738} 02/24/2022 13:56:21 - INFO - codeparrot_training - Step 9560: {'lr': 0.000470015824572783, 'samples': 4895232, 'steps': 9560, 'loss/train': 2.3217670917510986} 02/24/2022 13:56:24 - INFO - codeparrot_training - Step 9561: {'lr': 0.00047000805428203953, 'samples': 4895744, 'steps': 9561, 'loss/train': 1.4297391176223755} 02/24/2022 13:56:30 - INFO - codeparrot_training - Step 9562: {'lr': 0.00047000028304885143, 'samples': 4896256, 'steps': 9562, 'loss/train': 2.0284175872802734} 02/24/2022 13:56:33 - INFO - codeparrot_training - Step 9563: {'lr': 0.00046999251087325204, 'samples': 4896768, 'steps': 9563, 'loss/train': 2.065173625946045} 02/24/2022 13:56:39 - INFO - codeparrot_training - Step 9564: {'lr': 0.0004699847377552745, 'samples': 4897280, 'steps': 9564, 'loss/train': 2.0511999130249023} 02/24/2022 13:56:42 - INFO - codeparrot_training - Step 9565: {'lr': 0.00046997696369495217, 'samples': 4897792, 'steps': 9565, 'loss/train': 2.5077998638153076} 02/24/2022 13:56:48 - INFO - codeparrot_training - Step 9566: {'lr': 0.00046996918869231843, 'samples': 4898304, 'steps': 9566, 'loss/train': 1.9838497638702393} 02/24/2022 13:56:52 - INFO - codeparrot_training - Step 9567: {'lr': 0.00046996141274740653, 'samples': 4898816, 'steps': 9567, 'loss/train': 2.8092308044433594} 02/24/2022 13:56:57 - INFO - codeparrot_training - Step 9568: {'lr': 0.00046995363586024977, 'samples': 4899328, 'steps': 9568, 'loss/train': 2.267569065093994} 02/24/2022 13:57:01 - INFO - codeparrot_training - Step 9569: {'lr': 0.0004699458580308815, 'samples': 4899840, 'steps': 9569, 'loss/train': 2.0510623455047607} 02/24/2022 13:57:06 - INFO - codeparrot_training - Step 9570: {'lr': 0.00046993807925933503, 'samples': 4900352, 'steps': 9570, 'loss/train': 0.7282016277313232} 02/24/2022 13:57:10 - INFO - codeparrot_training - Step 9571: {'lr': 0.00046993029954564363, 'samples': 4900864, 'steps': 9571, 'loss/train': 2.312633752822876} 02/24/2022 13:57:16 - INFO - codeparrot_training - Step 9572: {'lr': 0.0004699225188898407, 'samples': 4901376, 'steps': 9572, 'loss/train': 2.3100268840789795} 02/24/2022 13:57:19 - INFO - codeparrot_training - Step 9573: {'lr': 0.0004699147372919595, 'samples': 4901888, 'steps': 9573, 'loss/train': 1.6473491191864014} 02/24/2022 13:57:25 - INFO - codeparrot_training - Step 9574: {'lr': 0.00046990695475203337, 'samples': 4902400, 'steps': 9574, 'loss/train': 1.8104907274246216} 02/24/2022 13:57:28 - INFO - codeparrot_training - Step 9575: {'lr': 0.00046989917127009573, 'samples': 4902912, 'steps': 9575, 'loss/train': 2.1922624111175537} 02/24/2022 13:57:35 - INFO - codeparrot_training - Step 9576: {'lr': 0.0004698913868461798, 'samples': 4903424, 'steps': 9576, 'loss/train': 2.563850164413452} 02/24/2022 13:57:38 - INFO - codeparrot_training - Step 9577: {'lr': 0.00046988360148031904, 'samples': 4903936, 'steps': 9577, 'loss/train': 1.61285400390625} 02/24/2022 13:57:43 - INFO - codeparrot_training - Step 9578: {'lr': 0.0004698758151725468, 'samples': 4904448, 'steps': 9578, 'loss/train': 2.6799488067626953} 02/24/2022 13:57:47 - INFO - codeparrot_training - Step 9579: {'lr': 0.0004698680279228963, 'samples': 4904960, 'steps': 9579, 'loss/train': 1.6138412952423096} 02/24/2022 13:57:52 - INFO - codeparrot_training - Step 9580: {'lr': 0.000469860239731401, 'samples': 4905472, 'steps': 9580, 'loss/train': 3.2819690704345703} 02/24/2022 13:57:56 - INFO - codeparrot_training - Step 9581: {'lr': 0.00046985245059809436, 'samples': 4905984, 'steps': 9581, 'loss/train': 1.3251656293869019} 02/24/2022 13:58:01 - INFO - codeparrot_training - Step 9582: {'lr': 0.0004698446605230095, 'samples': 4906496, 'steps': 9582, 'loss/train': 1.789786696434021} 02/24/2022 13:58:05 - INFO - codeparrot_training - Step 9583: {'lr': 0.00046983686950618, 'samples': 4907008, 'steps': 9583, 'loss/train': 1.3839339017868042} 02/24/2022 13:58:10 - INFO - codeparrot_training - Step 9584: {'lr': 0.00046982907754763905, 'samples': 4907520, 'steps': 9584, 'loss/train': 1.4463045597076416} 02/24/2022 13:58:14 - INFO - codeparrot_training - Step 9585: {'lr': 0.00046982128464742026, 'samples': 4908032, 'steps': 9585, 'loss/train': 2.0827138423919678} 02/24/2022 13:58:20 - INFO - codeparrot_training - Step 9586: {'lr': 0.0004698134908055568, 'samples': 4908544, 'steps': 9586, 'loss/train': 0.997948169708252} 02/24/2022 13:58:23 - INFO - codeparrot_training - Step 9587: {'lr': 0.00046980569602208215, 'samples': 4909056, 'steps': 9587, 'loss/train': 2.402827501296997} 02/24/2022 13:58:29 - INFO - codeparrot_training - Step 9588: {'lr': 0.00046979790029702973, 'samples': 4909568, 'steps': 9588, 'loss/train': 1.7306206226348877} 02/24/2022 13:58:32 - INFO - codeparrot_training - Step 9589: {'lr': 0.0004697901036304329, 'samples': 4910080, 'steps': 9589, 'loss/train': 1.250905156135559} 02/24/2022 13:58:38 - INFO - codeparrot_training - Step 9590: {'lr': 0.00046978230602232507, 'samples': 4910592, 'steps': 9590, 'loss/train': 1.3780932426452637} 02/24/2022 13:58:41 - INFO - codeparrot_training - Step 9591: {'lr': 0.00046977450747273956, 'samples': 4911104, 'steps': 9591, 'loss/train': 1.7797541618347168} 02/24/2022 13:58:47 - INFO - codeparrot_training - Step 9592: {'lr': 0.00046976670798171, 'samples': 4911616, 'steps': 9592, 'loss/train': 2.5511999130249023} 02/24/2022 13:58:50 - INFO - codeparrot_training - Step 9593: {'lr': 0.00046975890754926943, 'samples': 4912128, 'steps': 9593, 'loss/train': 1.5138994455337524} 02/24/2022 13:58:56 - INFO - codeparrot_training - Step 9594: {'lr': 0.0004697511061754516, 'samples': 4912640, 'steps': 9594, 'loss/train': 2.4406418800354004} 02/24/2022 13:59:00 - INFO - codeparrot_training - Step 9595: {'lr': 0.00046974330386028985, 'samples': 4913152, 'steps': 9595, 'loss/train': 1.1683231592178345} 02/24/2022 13:59:05 - INFO - codeparrot_training - Step 9596: {'lr': 0.0004697355006038175, 'samples': 4913664, 'steps': 9596, 'loss/train': 2.3619003295898438} 02/24/2022 13:59:09 - INFO - codeparrot_training - Step 9597: {'lr': 0.00046972769640606804, 'samples': 4914176, 'steps': 9597, 'loss/train': 2.7334089279174805} 02/24/2022 13:59:14 - INFO - codeparrot_training - Step 9598: {'lr': 0.0004697198912670749, 'samples': 4914688, 'steps': 9598, 'loss/train': 2.702599048614502} 02/24/2022 13:59:18 - INFO - codeparrot_training - Step 9599: {'lr': 0.0004697120851868715, 'samples': 4915200, 'steps': 9599, 'loss/train': 0.9323081970214844} 02/24/2022 13:59:23 - INFO - codeparrot_training - Step 9600: {'lr': 0.00046970427816549133, 'samples': 4915712, 'steps': 9600, 'loss/train': 1.5623677968978882} 02/24/2022 13:59:27 - INFO - codeparrot_training - Step 9601: {'lr': 0.0004696964702029678, 'samples': 4916224, 'steps': 9601, 'loss/train': 1.9898372888565063} 02/24/2022 13:59:33 - INFO - codeparrot_training - Step 9602: {'lr': 0.00046968866129933436, 'samples': 4916736, 'steps': 9602, 'loss/train': 1.855385184288025} 02/24/2022 13:59:36 - INFO - codeparrot_training - Step 9603: {'lr': 0.0004696808514546244, 'samples': 4917248, 'steps': 9603, 'loss/train': 2.258364200592041} 02/24/2022 13:59:42 - INFO - codeparrot_training - Step 9604: {'lr': 0.0004696730406688715, 'samples': 4917760, 'steps': 9604, 'loss/train': 1.625893235206604} 02/24/2022 13:59:45 - INFO - codeparrot_training - Step 9605: {'lr': 0.000469665228942109, 'samples': 4918272, 'steps': 9605, 'loss/train': 8.917402267456055} 02/24/2022 13:59:51 - INFO - codeparrot_training - Step 9606: {'lr': 0.0004696574162743704, 'samples': 4918784, 'steps': 9606, 'loss/train': 2.3655173778533936} 02/24/2022 13:59:55 - INFO - codeparrot_training - Step 9607: {'lr': 0.00046964960266568926, 'samples': 4919296, 'steps': 9607, 'loss/train': 2.5318398475646973} 02/24/2022 14:00:00 - INFO - codeparrot_training - Step 9608: {'lr': 0.0004696417881160989, 'samples': 4919808, 'steps': 9608, 'loss/train': 1.6272956132888794} 02/24/2022 14:00:04 - INFO - codeparrot_training - Step 9609: {'lr': 0.0004696339726256328, 'samples': 4920320, 'steps': 9609, 'loss/train': 3.1039879322052} 02/24/2022 14:00:09 - INFO - codeparrot_training - Step 9610: {'lr': 0.00046962615619432457, 'samples': 4920832, 'steps': 9610, 'loss/train': 3.6105611324310303} 02/24/2022 14:00:13 - INFO - codeparrot_training - Step 9611: {'lr': 0.0004696183388222077, 'samples': 4921344, 'steps': 9611, 'loss/train': 2.0522360801696777} 02/24/2022 14:00:19 - INFO - codeparrot_training - Step 9612: {'lr': 0.0004696105205093155, 'samples': 4921856, 'steps': 9612, 'loss/train': 2.508798599243164} 02/24/2022 14:00:23 - INFO - codeparrot_training - Step 9613: {'lr': 0.0004696027012556816, 'samples': 4922368, 'steps': 9613, 'loss/train': 3.8638055324554443} 02/24/2022 14:00:28 - INFO - codeparrot_training - Step 9614: {'lr': 0.00046959488106133944, 'samples': 4922880, 'steps': 9614, 'loss/train': 2.9614624977111816} 02/24/2022 14:00:32 - INFO - codeparrot_training - Step 9615: {'lr': 0.0004695870599263226, 'samples': 4923392, 'steps': 9615, 'loss/train': 2.158031463623047} 02/24/2022 14:00:37 - INFO - codeparrot_training - Step 9616: {'lr': 0.0004695792378506645, 'samples': 4923904, 'steps': 9616, 'loss/train': 1.3664880990982056} 02/24/2022 14:00:41 - INFO - codeparrot_training - Step 9617: {'lr': 0.00046957141483439856, 'samples': 4924416, 'steps': 9617, 'loss/train': 1.960555911064148} 02/24/2022 14:00:46 - INFO - codeparrot_training - Step 9618: {'lr': 0.0004695635908775585, 'samples': 4924928, 'steps': 9618, 'loss/train': 1.426436424255371} 02/24/2022 14:00:50 - INFO - codeparrot_training - Step 9619: {'lr': 0.0004695557659801778, 'samples': 4925440, 'steps': 9619, 'loss/train': 1.223673939704895} 02/24/2022 14:00:55 - INFO - codeparrot_training - Step 9620: {'lr': 0.0004695479401422898, 'samples': 4925952, 'steps': 9620, 'loss/train': 2.2305169105529785} 02/24/2022 14:00:59 - INFO - codeparrot_training - Step 9621: {'lr': 0.0004695401133639282, 'samples': 4926464, 'steps': 9621, 'loss/train': 2.5820133686065674} 02/24/2022 14:01:05 - INFO - codeparrot_training - Step 9622: {'lr': 0.0004695322856451264, 'samples': 4926976, 'steps': 9622, 'loss/train': 2.4427130222320557} 02/24/2022 14:01:09 - INFO - codeparrot_training - Step 9623: {'lr': 0.00046952445698591805, 'samples': 4927488, 'steps': 9623, 'loss/train': 1.9843087196350098} 02/24/2022 14:01:14 - INFO - codeparrot_training - Step 9624: {'lr': 0.0004695166273863367, 'samples': 4928000, 'steps': 9624, 'loss/train': 1.5638004541397095} 02/24/2022 14:01:18 - INFO - codeparrot_training - Step 9625: {'lr': 0.00046950879684641567, 'samples': 4928512, 'steps': 9625, 'loss/train': 3.016756057739258} 02/24/2022 14:01:23 - INFO - codeparrot_training - Step 9626: {'lr': 0.00046950096536618876, 'samples': 4929024, 'steps': 9626, 'loss/train': 2.2594692707061768} 02/24/2022 14:01:27 - INFO - codeparrot_training - Step 9627: {'lr': 0.0004694931329456894, 'samples': 4929536, 'steps': 9627, 'loss/train': 3.36421537399292} 02/24/2022 14:01:32 - INFO - codeparrot_training - Step 9628: {'lr': 0.0004694852995849511, 'samples': 4930048, 'steps': 9628, 'loss/train': 0.8950254321098328} 02/24/2022 14:01:36 - INFO - codeparrot_training - Step 9629: {'lr': 0.00046947746528400755, 'samples': 4930560, 'steps': 9629, 'loss/train': 2.137103796005249} 02/24/2022 14:01:41 - INFO - codeparrot_training - Step 9630: {'lr': 0.00046946963004289223, 'samples': 4931072, 'steps': 9630, 'loss/train': 1.9981069564819336} 02/24/2022 14:01:44 - INFO - codeparrot_training - Step 9631: {'lr': 0.0004694617938616386, 'samples': 4931584, 'steps': 9631, 'loss/train': 1.5629870891571045} 02/24/2022 14:01:51 - INFO - codeparrot_training - Step 9632: {'lr': 0.00046945395674028047, 'samples': 4932096, 'steps': 9632, 'loss/train': 8.76471996307373} 02/24/2022 14:01:54 - INFO - codeparrot_training - Step 9633: {'lr': 0.0004694461186788512, 'samples': 4932608, 'steps': 9633, 'loss/train': 1.5739970207214355} 02/24/2022 14:02:00 - INFO - codeparrot_training - Step 9634: {'lr': 0.0004694382796773844, 'samples': 4933120, 'steps': 9634, 'loss/train': 1.3431178331375122} 02/24/2022 14:02:03 - INFO - codeparrot_training - Step 9635: {'lr': 0.0004694304397359137, 'samples': 4933632, 'steps': 9635, 'loss/train': 2.063695192337036} 02/24/2022 14:02:09 - INFO - codeparrot_training - Step 9636: {'lr': 0.00046942259885447273, 'samples': 4934144, 'steps': 9636, 'loss/train': 0.3628299832344055} 02/24/2022 14:02:12 - INFO - codeparrot_training - Step 9637: {'lr': 0.000469414757033095, 'samples': 4934656, 'steps': 9637, 'loss/train': 1.983220100402832} 02/24/2022 14:02:18 - INFO - codeparrot_training - Step 9638: {'lr': 0.00046940691427181414, 'samples': 4935168, 'steps': 9638, 'loss/train': 1.1638362407684326} 02/24/2022 14:02:21 - INFO - codeparrot_training - Step 9639: {'lr': 0.00046939907057066374, 'samples': 4935680, 'steps': 9639, 'loss/train': 2.3545024394989014} 02/24/2022 14:02:27 - INFO - codeparrot_training - Step 9640: {'lr': 0.0004693912259296773, 'samples': 4936192, 'steps': 9640, 'loss/train': 1.83735990524292} 02/24/2022 14:02:30 - INFO - codeparrot_training - Step 9641: {'lr': 0.0004693833803488886, 'samples': 4936704, 'steps': 9641, 'loss/train': 3.072589635848999} 02/24/2022 14:02:35 - INFO - codeparrot_training - Step 9642: {'lr': 0.00046937553382833116, 'samples': 4937216, 'steps': 9642, 'loss/train': 2.5270586013793945} 02/24/2022 14:02:39 - INFO - codeparrot_training - Step 9643: {'lr': 0.00046936768636803857, 'samples': 4937728, 'steps': 9643, 'loss/train': 2.449934959411621} 02/24/2022 14:02:44 - INFO - codeparrot_training - Step 9644: {'lr': 0.00046935983796804443, 'samples': 4938240, 'steps': 9644, 'loss/train': 1.4152569770812988} 02/24/2022 14:02:48 - INFO - codeparrot_training - Step 9645: {'lr': 0.00046935198862838246, 'samples': 4938752, 'steps': 9645, 'loss/train': 2.1745424270629883} 02/24/2022 14:02:53 - INFO - codeparrot_training - Step 9646: {'lr': 0.00046934413834908616, 'samples': 4939264, 'steps': 9646, 'loss/train': 2.4889042377471924} 02/24/2022 14:02:59 - INFO - codeparrot_training - Step 9647: {'lr': 0.0004693362871301893, 'samples': 4939776, 'steps': 9647, 'loss/train': 2.2839608192443848} 02/24/2022 14:03:02 - INFO - codeparrot_training - Step 9648: {'lr': 0.0004693284349717254, 'samples': 4940288, 'steps': 9648, 'loss/train': 3.5323565006256104} 02/24/2022 14:03:09 - INFO - codeparrot_training - Step 9649: {'lr': 0.00046932058187372803, 'samples': 4940800, 'steps': 9649, 'loss/train': 1.9513170719146729} 02/24/2022 14:03:12 - INFO - codeparrot_training - Step 9650: {'lr': 0.00046931272783623106, 'samples': 4941312, 'steps': 9650, 'loss/train': 2.5716538429260254} 02/24/2022 14:03:17 - INFO - codeparrot_training - Step 9651: {'lr': 0.00046930487285926797, 'samples': 4941824, 'steps': 9651, 'loss/train': 1.8524171113967896} 02/24/2022 14:03:21 - INFO - codeparrot_training - Step 9652: {'lr': 0.00046929701694287243, 'samples': 4942336, 'steps': 9652, 'loss/train': 2.0290989875793457} 02/24/2022 14:03:26 - INFO - codeparrot_training - Step 9653: {'lr': 0.0004692891600870781, 'samples': 4942848, 'steps': 9653, 'loss/train': 2.1385068893432617} 02/24/2022 14:03:30 - INFO - codeparrot_training - Step 9654: {'lr': 0.00046928130229191865, 'samples': 4943360, 'steps': 9654, 'loss/train': 1.0425946712493896} 02/24/2022 14:03:35 - INFO - codeparrot_training - Step 9655: {'lr': 0.00046927344355742774, 'samples': 4943872, 'steps': 9655, 'loss/train': 1.6587049961090088} 02/24/2022 14:03:39 - INFO - codeparrot_training - Step 9656: {'lr': 0.00046926558388363904, 'samples': 4944384, 'steps': 9656, 'loss/train': 2.826145648956299} 02/24/2022 14:03:44 - INFO - codeparrot_training - Step 9657: {'lr': 0.00046925772327058616, 'samples': 4944896, 'steps': 9657, 'loss/train': 2.489088296890259} 02/24/2022 14:03:48 - INFO - codeparrot_training - Step 9658: {'lr': 0.0004692498617183028, 'samples': 4945408, 'steps': 9658, 'loss/train': 2.4383544921875} 02/24/2022 14:03:54 - INFO - codeparrot_training - Step 9659: {'lr': 0.0004692419992268227, 'samples': 4945920, 'steps': 9659, 'loss/train': 2.3415846824645996} 02/24/2022 14:03:57 - INFO - codeparrot_training - Step 9660: {'lr': 0.00046923413579617944, 'samples': 4946432, 'steps': 9660, 'loss/train': 1.6319698095321655} 02/24/2022 14:04:03 - INFO - codeparrot_training - Step 9661: {'lr': 0.00046922627142640685, 'samples': 4946944, 'steps': 9661, 'loss/train': 2.659302234649658} 02/24/2022 14:04:06 - INFO - codeparrot_training - Step 9662: {'lr': 0.00046921840611753845, 'samples': 4947456, 'steps': 9662, 'loss/train': 1.9705803394317627} 02/24/2022 14:04:12 - INFO - codeparrot_training - Step 9663: {'lr': 0.000469210539869608, 'samples': 4947968, 'steps': 9663, 'loss/train': 3.4343323707580566} 02/24/2022 14:04:16 - INFO - codeparrot_training - Step 9664: {'lr': 0.0004692026726826493, 'samples': 4948480, 'steps': 9664, 'loss/train': 3.209012269973755} 02/24/2022 14:04:21 - INFO - codeparrot_training - Step 9665: {'lr': 0.0004691948045566958, 'samples': 4948992, 'steps': 9665, 'loss/train': 1.8484220504760742} 02/24/2022 14:04:24 - INFO - codeparrot_training - Step 9666: {'lr': 0.0004691869354917815, 'samples': 4949504, 'steps': 9666, 'loss/train': 2.241541624069214} 02/24/2022 14:04:30 - INFO - codeparrot_training - Step 9667: {'lr': 0.0004691790654879399, 'samples': 4950016, 'steps': 9667, 'loss/train': 1.3001562356948853} 02/24/2022 14:04:33 - INFO - codeparrot_training - Step 9668: {'lr': 0.00046917119454520487, 'samples': 4950528, 'steps': 9668, 'loss/train': 2.233103036880493} 02/24/2022 14:04:39 - INFO - codeparrot_training - Step 9669: {'lr': 0.0004691633226636099, 'samples': 4951040, 'steps': 9669, 'loss/train': 2.446730136871338} 02/24/2022 14:04:43 - INFO - codeparrot_training - Step 9670: {'lr': 0.0004691554498431889, 'samples': 4951552, 'steps': 9670, 'loss/train': 2.168839693069458} 02/24/2022 14:04:51 - INFO - codeparrot_training - Step 9671: {'lr': 0.00046914757608397555, 'samples': 4952064, 'steps': 9671, 'loss/train': 0.19441667199134827} 02/24/2022 14:04:54 - INFO - codeparrot_training - Step 9672: {'lr': 0.00046913970138600357, 'samples': 4952576, 'steps': 9672, 'loss/train': 2.23901104927063} 02/24/2022 14:05:00 - INFO - codeparrot_training - Step 9673: {'lr': 0.0004691318257493067, 'samples': 4953088, 'steps': 9673, 'loss/train': 2.6474897861480713} 02/24/2022 14:05:03 - INFO - codeparrot_training - Step 9674: {'lr': 0.00046912394917391866, 'samples': 4953600, 'steps': 9674, 'loss/train': 2.262817621231079} 02/24/2022 14:05:09 - INFO - codeparrot_training - Step 9675: {'lr': 0.00046911607165987324, 'samples': 4954112, 'steps': 9675, 'loss/train': 1.7118260860443115} 02/24/2022 14:05:12 - INFO - codeparrot_training - Step 9676: {'lr': 0.0004691081932072041, 'samples': 4954624, 'steps': 9676, 'loss/train': 2.1820528507232666} 02/24/2022 14:05:18 - INFO - codeparrot_training - Step 9677: {'lr': 0.0004691003138159451, 'samples': 4955136, 'steps': 9677, 'loss/train': 2.8769967555999756} 02/24/2022 14:05:21 - INFO - codeparrot_training - Step 9678: {'lr': 0.00046909243348612986, 'samples': 4955648, 'steps': 9678, 'loss/train': 1.680458903312683} 02/24/2022 14:05:27 - INFO - codeparrot_training - Step 9679: {'lr': 0.0004690845522177922, 'samples': 4956160, 'steps': 9679, 'loss/train': 1.559897780418396} 02/24/2022 14:05:30 - INFO - codeparrot_training - Step 9680: {'lr': 0.0004690766700109659, 'samples': 4956672, 'steps': 9680, 'loss/train': 1.483007550239563} 02/24/2022 14:05:37 - INFO - codeparrot_training - Step 9681: {'lr': 0.0004690687868656847, 'samples': 4957184, 'steps': 9681, 'loss/train': 1.9457534551620483} 02/24/2022 14:05:41 - INFO - codeparrot_training - Step 9682: {'lr': 0.00046906090278198246, 'samples': 4957696, 'steps': 9682, 'loss/train': 1.7990607023239136} 02/24/2022 14:05:46 - INFO - codeparrot_training - Step 9683: {'lr': 0.00046905301775989277, 'samples': 4958208, 'steps': 9683, 'loss/train': 1.5773755311965942} 02/24/2022 14:05:50 - INFO - codeparrot_training - Step 9684: {'lr': 0.0004690451317994495, 'samples': 4958720, 'steps': 9684, 'loss/train': 2.0951998233795166} 02/24/2022 14:05:55 - INFO - codeparrot_training - Step 9685: {'lr': 0.00046903724490068654, 'samples': 4959232, 'steps': 9685, 'loss/train': 1.9504728317260742} 02/24/2022 14:05:59 - INFO - codeparrot_training - Step 9686: {'lr': 0.00046902935706363754, 'samples': 4959744, 'steps': 9686, 'loss/train': 1.9299465417861938} 02/24/2022 14:06:05 - INFO - codeparrot_training - Step 9687: {'lr': 0.0004690214682883363, 'samples': 4960256, 'steps': 9687, 'loss/train': 2.704273223876953} 02/24/2022 14:06:08 - INFO - codeparrot_training - Step 9688: {'lr': 0.00046901357857481664, 'samples': 4960768, 'steps': 9688, 'loss/train': 2.1331183910369873} 02/24/2022 14:06:14 - INFO - codeparrot_training - Step 9689: {'lr': 0.0004690056879231124, 'samples': 4961280, 'steps': 9689, 'loss/train': 1.3145971298217773} 02/24/2022 14:06:17 - INFO - codeparrot_training - Step 9690: {'lr': 0.0004689977963332572, 'samples': 4961792, 'steps': 9690, 'loss/train': 2.0762250423431396} 02/24/2022 14:06:23 - INFO - codeparrot_training - Step 9691: {'lr': 0.0004689899038052852, 'samples': 4962304, 'steps': 9691, 'loss/train': 1.8556798696517944} 02/24/2022 14:06:26 - INFO - codeparrot_training - Step 9692: {'lr': 0.0004689820103392298, 'samples': 4962816, 'steps': 9692, 'loss/train': 2.8912041187286377} 02/24/2022 14:06:33 - INFO - codeparrot_training - Step 9693: {'lr': 0.0004689741159351251, 'samples': 4963328, 'steps': 9693, 'loss/train': 1.6728296279907227} 02/24/2022 14:06:37 - INFO - codeparrot_training - Step 9694: {'lr': 0.00046896622059300477, 'samples': 4963840, 'steps': 9694, 'loss/train': 1.652101993560791} 02/24/2022 14:06:42 - INFO - codeparrot_training - Step 9695: {'lr': 0.00046895832431290266, 'samples': 4964352, 'steps': 9695, 'loss/train': 3.0453104972839355} 02/24/2022 14:06:46 - INFO - codeparrot_training - Step 9696: {'lr': 0.0004689504270948527, 'samples': 4964864, 'steps': 9696, 'loss/train': 2.844454050064087} 02/24/2022 14:06:51 - INFO - codeparrot_training - Step 9697: {'lr': 0.00046894252893888854, 'samples': 4965376, 'steps': 9697, 'loss/train': 1.9529099464416504} 02/24/2022 14:06:55 - INFO - codeparrot_training - Step 9698: {'lr': 0.0004689346298450442, 'samples': 4965888, 'steps': 9698, 'loss/train': 1.6889833211898804} 02/24/2022 14:07:00 - INFO - codeparrot_training - Step 9699: {'lr': 0.0004689267298133534, 'samples': 4966400, 'steps': 9699, 'loss/train': 1.4936554431915283} 02/24/2022 14:07:04 - INFO - codeparrot_training - Step 9700: {'lr': 0.00046891882884384997, 'samples': 4966912, 'steps': 9700, 'loss/train': 1.9704091548919678} 02/24/2022 14:07:09 - INFO - codeparrot_training - Step 9701: {'lr': 0.00046891092693656777, 'samples': 4967424, 'steps': 9701, 'loss/train': 2.710467576980591} 02/24/2022 14:07:13 - INFO - codeparrot_training - Step 9702: {'lr': 0.0004689030240915407, 'samples': 4967936, 'steps': 9702, 'loss/train': 2.1302390098571777} 02/24/2022 14:07:20 - INFO - codeparrot_training - Step 9703: {'lr': 0.0004688951203088026, 'samples': 4968448, 'steps': 9703, 'loss/train': 2.377704381942749} 02/24/2022 14:07:24 - INFO - codeparrot_training - Step 9704: {'lr': 0.00046888721558838734, 'samples': 4968960, 'steps': 9704, 'loss/train': 0.8488210439682007} 02/24/2022 14:07:29 - INFO - codeparrot_training - Step 9705: {'lr': 0.0004688793099303287, 'samples': 4969472, 'steps': 9705, 'loss/train': 1.5677474737167358} 02/24/2022 14:07:33 - INFO - codeparrot_training - Step 9706: {'lr': 0.0004688714033346606, 'samples': 4969984, 'steps': 9706, 'loss/train': 1.6840240955352783} 02/24/2022 14:07:38 - INFO - codeparrot_training - Step 9707: {'lr': 0.000468863495801417, 'samples': 4970496, 'steps': 9707, 'loss/train': 1.791087031364441} 02/24/2022 14:07:42 - INFO - codeparrot_training - Step 9708: {'lr': 0.00046885558733063157, 'samples': 4971008, 'steps': 9708, 'loss/train': 0.8697268962860107} 02/24/2022 14:07:47 - INFO - codeparrot_training - Step 9709: {'lr': 0.00046884767792233827, 'samples': 4971520, 'steps': 9709, 'loss/train': 3.0986404418945312} 02/24/2022 14:07:53 - INFO - codeparrot_training - Step 9710: {'lr': 0.00046883976757657107, 'samples': 4972032, 'steps': 9710, 'loss/train': 2.8138427734375} 02/24/2022 14:07:56 - INFO - codeparrot_training - Step 9711: {'lr': 0.00046883185629336386, 'samples': 4972544, 'steps': 9711, 'loss/train': 1.4844030141830444} 02/24/2022 14:08:02 - INFO - codeparrot_training - Step 9712: {'lr': 0.0004688239440727504, 'samples': 4973056, 'steps': 9712, 'loss/train': 1.3667101860046387} 02/24/2022 14:08:05 - INFO - codeparrot_training - Step 9713: {'lr': 0.00046881603091476466, 'samples': 4973568, 'steps': 9713, 'loss/train': 1.4271619319915771} 02/24/2022 14:08:11 - INFO - codeparrot_training - Step 9714: {'lr': 0.0004688081168194405, 'samples': 4974080, 'steps': 9714, 'loss/train': 1.250153660774231} 02/24/2022 14:08:14 - INFO - codeparrot_training - Step 9715: {'lr': 0.0004688002017868119, 'samples': 4974592, 'steps': 9715, 'loss/train': 0.762831449508667} 02/24/2022 14:08:22 - INFO - codeparrot_training - Step 9716: {'lr': 0.0004687922858169126, 'samples': 4975104, 'steps': 9716, 'loss/train': 1.4025824069976807} 02/24/2022 14:08:25 - INFO - codeparrot_training - Step 9717: {'lr': 0.0004687843689097767, 'samples': 4975616, 'steps': 9717, 'loss/train': 2.28603196144104} 02/24/2022 14:08:31 - INFO - codeparrot_training - Step 9718: {'lr': 0.0004687764510654381, 'samples': 4976128, 'steps': 9718, 'loss/train': 2.9806973934173584} 02/24/2022 14:08:35 - INFO - codeparrot_training - Step 9719: {'lr': 0.0004687685322839306, 'samples': 4976640, 'steps': 9719, 'loss/train': 2.082766056060791} 02/24/2022 14:08:38 - INFO - codeparrot_training - Step 9720: {'lr': 0.00046876061256528813, 'samples': 4977152, 'steps': 9720, 'loss/train': 1.6067500114440918} 02/24/2022 14:08:44 - INFO - codeparrot_training - Step 9721: {'lr': 0.00046875269190954465, 'samples': 4977664, 'steps': 9721, 'loss/train': 2.112417697906494} 02/24/2022 14:08:47 - INFO - codeparrot_training - Step 9722: {'lr': 0.00046874477031673417, 'samples': 4978176, 'steps': 9722, 'loss/train': 2.8875627517700195} 02/24/2022 14:08:53 - INFO - codeparrot_training - Step 9723: {'lr': 0.00046873684778689053, 'samples': 4978688, 'steps': 9723, 'loss/train': 1.9508256912231445} 02/24/2022 14:08:56 - INFO - codeparrot_training - Step 9724: {'lr': 0.00046872892432004765, 'samples': 4979200, 'steps': 9724, 'loss/train': 2.1635444164276123} 02/24/2022 14:09:02 - INFO - codeparrot_training - Step 9725: {'lr': 0.00046872099991623954, 'samples': 4979712, 'steps': 9725, 'loss/train': 2.386948347091675} 02/24/2022 14:09:05 - INFO - codeparrot_training - Step 9726: {'lr': 0.0004687130745755002, 'samples': 4980224, 'steps': 9726, 'loss/train': 1.574517846107483} 02/24/2022 14:09:13 - INFO - codeparrot_training - Step 9727: {'lr': 0.0004687051482978634, 'samples': 4980736, 'steps': 9727, 'loss/train': 2.435608148574829} 02/24/2022 14:09:16 - INFO - codeparrot_training - Step 9728: {'lr': 0.0004686972210833632, 'samples': 4981248, 'steps': 9728, 'loss/train': 2.3228836059570312} 02/24/2022 14:09:22 - INFO - codeparrot_training - Step 9729: {'lr': 0.00046868929293203355, 'samples': 4981760, 'steps': 9729, 'loss/train': 1.6026933193206787} 02/24/2022 14:09:25 - INFO - codeparrot_training - Step 9730: {'lr': 0.0004686813638439085, 'samples': 4982272, 'steps': 9730, 'loss/train': 2.3634378910064697} 02/24/2022 14:09:30 - INFO - codeparrot_training - Step 9731: {'lr': 0.00046867343381902185, 'samples': 4982784, 'steps': 9731, 'loss/train': 1.9287230968475342} 02/24/2022 14:09:34 - INFO - codeparrot_training - Step 9732: {'lr': 0.0004686655028574076, 'samples': 4983296, 'steps': 9732, 'loss/train': 2.705789804458618} 02/24/2022 14:09:39 - INFO - codeparrot_training - Step 9733: {'lr': 0.0004686575709590998, 'samples': 4983808, 'steps': 9733, 'loss/train': 2.2005279064178467} 02/24/2022 14:09:43 - INFO - codeparrot_training - Step 9734: {'lr': 0.00046864963812413244, 'samples': 4984320, 'steps': 9734, 'loss/train': 0.14529983699321747} 02/24/2022 14:09:48 - INFO - codeparrot_training - Step 9735: {'lr': 0.00046864170435253946, 'samples': 4984832, 'steps': 9735, 'loss/train': 1.6733207702636719} 02/24/2022 14:09:52 - INFO - codeparrot_training - Step 9736: {'lr': 0.0004686337696443548, 'samples': 4985344, 'steps': 9736, 'loss/train': 2.8610265254974365} 02/24/2022 14:09:57 - INFO - codeparrot_training - Step 9737: {'lr': 0.0004686258339996125, 'samples': 4985856, 'steps': 9737, 'loss/train': 1.4688588380813599} 02/24/2022 14:10:01 - INFO - codeparrot_training - Step 9738: {'lr': 0.0004686178974183466, 'samples': 4986368, 'steps': 9738, 'loss/train': 2.0714304447174072} 02/24/2022 14:10:08 - INFO - codeparrot_training - Step 9739: {'lr': 0.00046860995990059096, 'samples': 4986880, 'steps': 9739, 'loss/train': 2.5236568450927734} 02/24/2022 14:10:12 - INFO - codeparrot_training - Step 9740: {'lr': 0.00046860202144637976, 'samples': 4987392, 'steps': 9740, 'loss/train': 2.6240153312683105} 02/24/2022 14:10:17 - INFO - codeparrot_training - Step 9741: {'lr': 0.0004685940820557468, 'samples': 4987904, 'steps': 9741, 'loss/train': 1.732853651046753} 02/24/2022 14:10:21 - INFO - codeparrot_training - Step 9742: {'lr': 0.0004685861417287263, 'samples': 4988416, 'steps': 9742, 'loss/train': 2.13775897026062} 02/24/2022 14:10:26 - INFO - codeparrot_training - Step 9743: {'lr': 0.00046857820046535215, 'samples': 4988928, 'steps': 9743, 'loss/train': 1.5943260192871094} 02/24/2022 14:10:30 - INFO - codeparrot_training - Step 9744: {'lr': 0.0004685702582656584, 'samples': 4989440, 'steps': 9744, 'loss/train': 1.6716082096099854} 02/24/2022 14:10:35 - INFO - codeparrot_training - Step 9745: {'lr': 0.0004685623151296791, 'samples': 4989952, 'steps': 9745, 'loss/train': 2.218304395675659} 02/24/2022 14:10:39 - INFO - codeparrot_training - Step 9746: {'lr': 0.0004685543710574482, 'samples': 4990464, 'steps': 9746, 'loss/train': 2.802156448364258} 02/24/2022 14:10:44 - INFO - codeparrot_training - Step 9747: {'lr': 0.00046854642604899976, 'samples': 4990976, 'steps': 9747, 'loss/train': 2.275113344192505} 02/24/2022 14:10:48 - INFO - codeparrot_training - Step 9748: {'lr': 0.00046853848010436783, 'samples': 4991488, 'steps': 9748, 'loss/train': 1.5402599573135376} 02/24/2022 14:10:55 - INFO - codeparrot_training - Step 9749: {'lr': 0.00046853053322358653, 'samples': 4992000, 'steps': 9749, 'loss/train': 2.3010404109954834} 02/24/2022 14:10:58 - INFO - codeparrot_training - Step 9750: {'lr': 0.00046852258540668973, 'samples': 4992512, 'steps': 9750, 'loss/train': 2.3874311447143555} 02/24/2022 14:11:04 - INFO - codeparrot_training - Step 9751: {'lr': 0.0004685146366537116, 'samples': 4993024, 'steps': 9751, 'loss/train': 2.006460189819336} 02/24/2022 14:11:08 - INFO - codeparrot_training - Step 9752: {'lr': 0.00046850668696468614, 'samples': 4993536, 'steps': 9752, 'loss/train': 2.3719584941864014} 02/24/2022 14:11:13 - INFO - codeparrot_training - Step 9753: {'lr': 0.0004684987363396474, 'samples': 4994048, 'steps': 9753, 'loss/train': 0.9575660824775696} 02/24/2022 14:11:16 - INFO - codeparrot_training - Step 9754: {'lr': 0.0004684907847786295, 'samples': 4994560, 'steps': 9754, 'loss/train': 2.432325839996338} 02/24/2022 14:11:22 - INFO - codeparrot_training - Step 9755: {'lr': 0.0004684828322816664, 'samples': 4995072, 'steps': 9755, 'loss/train': 0.5235753655433655} 02/24/2022 14:11:26 - INFO - codeparrot_training - Step 9756: {'lr': 0.00046847487884879227, 'samples': 4995584, 'steps': 9756, 'loss/train': 1.5608141422271729} 02/24/2022 14:11:31 - INFO - codeparrot_training - Step 9757: {'lr': 0.0004684669244800411, 'samples': 4996096, 'steps': 9757, 'loss/train': 2.2094027996063232} 02/24/2022 14:11:35 - INFO - codeparrot_training - Step 9758: {'lr': 0.00046845896917544703, 'samples': 4996608, 'steps': 9758, 'loss/train': 2.599088668823242} 02/24/2022 14:11:40 - INFO - codeparrot_training - Step 9759: {'lr': 0.00046845101293504403, 'samples': 4997120, 'steps': 9759, 'loss/train': 0.33489975333213806} 02/24/2022 14:11:44 - INFO - codeparrot_training - Step 9760: {'lr': 0.00046844305575886636, 'samples': 4997632, 'steps': 9760, 'loss/train': 2.0347495079040527} 02/24/2022 14:11:49 - INFO - codeparrot_training - Step 9761: {'lr': 0.00046843509764694794, 'samples': 4998144, 'steps': 9761, 'loss/train': 2.290970802307129} 02/24/2022 14:11:53 - INFO - codeparrot_training - Step 9762: {'lr': 0.0004684271385993229, 'samples': 4998656, 'steps': 9762, 'loss/train': 2.3579905033111572} 02/24/2022 14:11:58 - INFO - codeparrot_training - Step 9763: {'lr': 0.0004684191786160254, 'samples': 4999168, 'steps': 9763, 'loss/train': 2.1428382396698} 02/24/2022 14:12:02 - INFO - codeparrot_training - Step 9764: {'lr': 0.0004684112176970895, 'samples': 4999680, 'steps': 9764, 'loss/train': 2.238593816757202} 02/24/2022 14:12:09 - INFO - codeparrot_training - Step 9765: {'lr': 0.0004684032558425493, 'samples': 5000192, 'steps': 9765, 'loss/train': 1.7709112167358398} 02/24/2022 14:12:13 - INFO - codeparrot_training - Step 9766: {'lr': 0.00046839529305243885, 'samples': 5000704, 'steps': 9766, 'loss/train': 2.64963960647583} 02/24/2022 14:12:18 - INFO - codeparrot_training - Step 9767: {'lr': 0.00046838732932679236, 'samples': 5001216, 'steps': 9767, 'loss/train': 2.3232438564300537} 02/24/2022 14:12:22 - INFO - codeparrot_training - Step 9768: {'lr': 0.0004683793646656439, 'samples': 5001728, 'steps': 9768, 'loss/train': 1.7110859155654907} 02/24/2022 14:12:27 - INFO - codeparrot_training - Step 9769: {'lr': 0.00046837139906902753, 'samples': 5002240, 'steps': 9769, 'loss/train': 1.7164878845214844} 02/24/2022 14:12:31 - INFO - codeparrot_training - Step 9770: {'lr': 0.00046836343253697744, 'samples': 5002752, 'steps': 9770, 'loss/train': 2.112417697906494} 02/24/2022 14:12:36 - INFO - codeparrot_training - Step 9771: {'lr': 0.0004683554650695278, 'samples': 5003264, 'steps': 9771, 'loss/train': 2.0658204555511475} 02/24/2022 14:12:40 - INFO - codeparrot_training - Step 9772: {'lr': 0.0004683474966667127, 'samples': 5003776, 'steps': 9772, 'loss/train': 2.604013204574585} 02/24/2022 14:12:45 - INFO - codeparrot_training - Step 9773: {'lr': 0.00046833952732856614, 'samples': 5004288, 'steps': 9773, 'loss/train': 2.144122362136841} 02/24/2022 14:12:49 - INFO - codeparrot_training - Step 9774: {'lr': 0.00046833155705512246, 'samples': 5004800, 'steps': 9774, 'loss/train': 1.988416075706482} 02/24/2022 14:12:56 - INFO - codeparrot_training - Step 9775: {'lr': 0.0004683235858464157, 'samples': 5005312, 'steps': 9775, 'loss/train': 2.118572950363159} 02/24/2022 14:13:00 - INFO - codeparrot_training - Step 9776: {'lr': 0.0004683156137024801, 'samples': 5005824, 'steps': 9776, 'loss/train': 3.1579926013946533} 02/24/2022 14:13:05 - INFO - codeparrot_training - Step 9777: {'lr': 0.0004683076406233496, 'samples': 5006336, 'steps': 9777, 'loss/train': 2.162621259689331} 02/24/2022 14:13:08 - INFO - codeparrot_training - Step 9778: {'lr': 0.0004682996666090585, 'samples': 5006848, 'steps': 9778, 'loss/train': 2.5025382041931152} 02/24/2022 14:13:14 - INFO - codeparrot_training - Step 9779: {'lr': 0.00046829169165964104, 'samples': 5007360, 'steps': 9779, 'loss/train': 3.010511636734009} 02/24/2022 14:13:18 - INFO - codeparrot_training - Step 9780: {'lr': 0.0004682837157751313, 'samples': 5007872, 'steps': 9780, 'loss/train': 0.49823588132858276} 02/24/2022 14:13:23 - INFO - codeparrot_training - Step 9781: {'lr': 0.00046827573895556334, 'samples': 5008384, 'steps': 9781, 'loss/train': 2.351402997970581} 02/24/2022 14:13:27 - INFO - codeparrot_training - Step 9782: {'lr': 0.00046826776120097147, 'samples': 5008896, 'steps': 9782, 'loss/train': 2.7306134700775146} 02/24/2022 14:13:32 - INFO - codeparrot_training - Step 9783: {'lr': 0.0004682597825113898, 'samples': 5009408, 'steps': 9783, 'loss/train': 2.374734878540039} 02/24/2022 14:13:36 - INFO - codeparrot_training - Step 9784: {'lr': 0.00046825180288685253, 'samples': 5009920, 'steps': 9784, 'loss/train': 2.564307451248169} 02/24/2022 14:13:43 - INFO - codeparrot_training - Step 9785: {'lr': 0.00046824382232739386, 'samples': 5010432, 'steps': 9785, 'loss/train': 2.068251371383667} 02/24/2022 14:13:47 - INFO - codeparrot_training - Step 9786: {'lr': 0.00046823584083304794, 'samples': 5010944, 'steps': 9786, 'loss/train': 1.333812952041626} 02/24/2022 14:13:52 - INFO - codeparrot_training - Step 9787: {'lr': 0.00046822785840384897, 'samples': 5011456, 'steps': 9787, 'loss/train': 2.36928653717041} 02/24/2022 14:13:56 - INFO - codeparrot_training - Step 9788: {'lr': 0.0004682198750398312, 'samples': 5011968, 'steps': 9788, 'loss/train': 0.7014451026916504} 02/24/2022 14:14:01 - INFO - codeparrot_training - Step 9789: {'lr': 0.0004682118907410287, 'samples': 5012480, 'steps': 9789, 'loss/train': 2.3775744438171387} 02/24/2022 14:14:05 - INFO - codeparrot_training - Step 9790: {'lr': 0.00046820390550747585, 'samples': 5012992, 'steps': 9790, 'loss/train': 2.2849907875061035} 02/24/2022 14:14:10 - INFO - codeparrot_training - Step 9791: {'lr': 0.0004681959193392067, 'samples': 5013504, 'steps': 9791, 'loss/train': 2.189612627029419} 02/24/2022 14:14:14 - INFO - codeparrot_training - Step 9792: {'lr': 0.00046818793223625543, 'samples': 5014016, 'steps': 9792, 'loss/train': 1.6522659063339233} 02/24/2022 14:14:20 - INFO - codeparrot_training - Step 9793: {'lr': 0.0004681799441986564, 'samples': 5014528, 'steps': 9793, 'loss/train': 2.472900390625} 02/24/2022 14:14:23 - INFO - codeparrot_training - Step 9794: {'lr': 0.00046817195522644387, 'samples': 5015040, 'steps': 9794, 'loss/train': 3.4724793434143066} 02/24/2022 14:14:29 - INFO - codeparrot_training - Step 9795: {'lr': 0.00046816396531965186, 'samples': 5015552, 'steps': 9795, 'loss/train': 1.3836543560028076} 02/24/2022 14:14:32 - INFO - codeparrot_training - Step 9796: {'lr': 0.0004681559744783147, 'samples': 5016064, 'steps': 9796, 'loss/train': 1.037685751914978} 02/24/2022 14:14:38 - INFO - codeparrot_training - Step 9797: {'lr': 0.00046814798270246663, 'samples': 5016576, 'steps': 9797, 'loss/train': 2.375507116317749} 02/24/2022 14:14:45 - INFO - codeparrot_training - Step 9798: {'lr': 0.00046813998999214193, 'samples': 5017088, 'steps': 9798, 'loss/train': 1.9001340866088867} 02/24/2022 14:14:48 - INFO - codeparrot_training - Step 9799: {'lr': 0.0004681319963473747, 'samples': 5017600, 'steps': 9799, 'loss/train': 2.62568998336792} 02/24/2022 14:14:54 - INFO - codeparrot_training - Step 9800: {'lr': 0.0004681240017681993, 'samples': 5018112, 'steps': 9800, 'loss/train': 2.387746572494507} 02/24/2022 14:14:57 - INFO - codeparrot_training - Step 9801: {'lr': 0.0004681160062546499, 'samples': 5018624, 'steps': 9801, 'loss/train': 2.5254976749420166} 02/24/2022 14:15:03 - INFO - codeparrot_training - Step 9802: {'lr': 0.00046810800980676083, 'samples': 5019136, 'steps': 9802, 'loss/train': 1.4963001012802124} 02/24/2022 14:15:06 - INFO - codeparrot_training - Step 9803: {'lr': 0.0004681000124245663, 'samples': 5019648, 'steps': 9803, 'loss/train': 3.0191409587860107} 02/24/2022 14:15:12 - INFO - codeparrot_training - Step 9804: {'lr': 0.0004680920141081005, 'samples': 5020160, 'steps': 9804, 'loss/train': 1.0616954565048218} 02/24/2022 14:15:15 - INFO - codeparrot_training - Step 9805: {'lr': 0.00046808401485739793, 'samples': 5020672, 'steps': 9805, 'loss/train': 1.6365487575531006} 02/24/2022 14:15:21 - INFO - codeparrot_training - Step 9806: {'lr': 0.00046807601467249255, 'samples': 5021184, 'steps': 9806, 'loss/train': 1.661150574684143} 02/24/2022 14:15:24 - INFO - codeparrot_training - Step 9807: {'lr': 0.0004680680135534188, 'samples': 5021696, 'steps': 9807, 'loss/train': 1.475576400756836} 02/24/2022 14:15:30 - INFO - codeparrot_training - Step 9808: {'lr': 0.00046806001150021095, 'samples': 5022208, 'steps': 9808, 'loss/train': 3.4492859840393066} 02/24/2022 14:15:34 - INFO - codeparrot_training - Step 9809: {'lr': 0.0004680520085129032, 'samples': 5022720, 'steps': 9809, 'loss/train': 1.5156227350234985} 02/24/2022 14:15:37 - INFO - codeparrot_training - Step 9810: {'lr': 0.00046804400459152994, 'samples': 5023232, 'steps': 9810, 'loss/train': 3.0102250576019287} 02/24/2022 14:15:44 - INFO - codeparrot_training - Step 9811: {'lr': 0.0004680359997361254, 'samples': 5023744, 'steps': 9811, 'loss/train': 0.6598783731460571} 02/24/2022 14:15:48 - INFO - codeparrot_training - Step 9812: {'lr': 0.0004680279939467238, 'samples': 5024256, 'steps': 9812, 'loss/train': 1.3877547979354858} 02/24/2022 14:15:53 - INFO - codeparrot_training - Step 9813: {'lr': 0.0004680199872233596, 'samples': 5024768, 'steps': 9813, 'loss/train': 1.8924779891967773} 02/24/2022 14:15:57 - INFO - codeparrot_training - Step 9814: {'lr': 0.00046801197956606693, 'samples': 5025280, 'steps': 9814, 'loss/train': 2.622920036315918} 02/24/2022 14:16:02 - INFO - codeparrot_training - Step 9815: {'lr': 0.00046800397097488024, 'samples': 5025792, 'steps': 9815, 'loss/train': 1.2758949995040894} 02/24/2022 14:16:06 - INFO - codeparrot_training - Step 9816: {'lr': 0.0004679959614498337, 'samples': 5026304, 'steps': 9816, 'loss/train': 1.6469138860702515} 02/24/2022 14:16:11 - INFO - codeparrot_training - Step 9817: {'lr': 0.0004679879509909617, 'samples': 5026816, 'steps': 9817, 'loss/train': 1.9287909269332886} 02/24/2022 14:16:15 - INFO - codeparrot_training - Step 9818: {'lr': 0.00046797993959829857, 'samples': 5027328, 'steps': 9818, 'loss/train': 2.843106985092163} 02/24/2022 14:16:20 - INFO - codeparrot_training - Step 9819: {'lr': 0.00046797192727187855, 'samples': 5027840, 'steps': 9819, 'loss/train': 1.385992169380188} 02/24/2022 14:16:24 - INFO - codeparrot_training - Step 9820: {'lr': 0.000467963914011736, 'samples': 5028352, 'steps': 9820, 'loss/train': 2.058119297027588} 02/24/2022 14:16:31 - INFO - codeparrot_training - Step 9821: {'lr': 0.0004679558998179053, 'samples': 5028864, 'steps': 9821, 'loss/train': 1.1119884252548218} 02/24/2022 14:16:34 - INFO - codeparrot_training - Step 9822: {'lr': 0.0004679478846904207, 'samples': 5029376, 'steps': 9822, 'loss/train': 1.6504911184310913} 02/24/2022 14:16:40 - INFO - codeparrot_training - Step 9823: {'lr': 0.00046793986862931654, 'samples': 5029888, 'steps': 9823, 'loss/train': 2.2548770904541016} 02/24/2022 14:16:45 - INFO - codeparrot_training - Step 9824: {'lr': 0.0004679318516346273, 'samples': 5030400, 'steps': 9824, 'loss/train': 2.2708287239074707} 02/24/2022 14:16:49 - INFO - codeparrot_training - Step 9825: {'lr': 0.00046792383370638705, 'samples': 5030912, 'steps': 9825, 'loss/train': 3.161379814147949} 02/24/2022 14:16:54 - INFO - codeparrot_training - Step 9826: {'lr': 0.0004679158148446304, 'samples': 5031424, 'steps': 9826, 'loss/train': 1.0474443435668945} 02/24/2022 14:16:58 - INFO - codeparrot_training - Step 9827: {'lr': 0.00046790779504939155, 'samples': 5031936, 'steps': 9827, 'loss/train': 2.720407485961914} 02/24/2022 14:17:03 - INFO - codeparrot_training - Step 9828: {'lr': 0.00046789977432070497, 'samples': 5032448, 'steps': 9828, 'loss/train': 2.0093250274658203} 02/24/2022 14:17:07 - INFO - codeparrot_training - Step 9829: {'lr': 0.00046789175265860483, 'samples': 5032960, 'steps': 9829, 'loss/train': 1.6026208400726318} 02/24/2022 14:17:14 - INFO - codeparrot_training - Step 9830: {'lr': 0.00046788373006312567, 'samples': 5033472, 'steps': 9830, 'loss/train': 2.3966925144195557} 02/24/2022 14:17:17 - INFO - codeparrot_training - Step 9831: {'lr': 0.0004678757065343019, 'samples': 5033984, 'steps': 9831, 'loss/train': 2.5034494400024414} 02/24/2022 14:17:23 - INFO - codeparrot_training - Step 9832: {'lr': 0.0004678676820721677, 'samples': 5034496, 'steps': 9832, 'loss/train': 2.020383358001709} 02/24/2022 14:17:26 - INFO - codeparrot_training - Step 9833: {'lr': 0.00046785965667675745, 'samples': 5035008, 'steps': 9833, 'loss/train': 1.6288156509399414} 02/24/2022 14:17:32 - INFO - codeparrot_training - Step 9834: {'lr': 0.00046785163034810567, 'samples': 5035520, 'steps': 9834, 'loss/train': 2.30257248878479} 02/24/2022 14:17:35 - INFO - codeparrot_training - Step 9835: {'lr': 0.00046784360308624675, 'samples': 5036032, 'steps': 9835, 'loss/train': 1.4314764738082886} 02/24/2022 14:17:41 - INFO - codeparrot_training - Step 9836: {'lr': 0.0004678355748912149, 'samples': 5036544, 'steps': 9836, 'loss/train': 2.7425527572631836} 02/24/2022 14:17:44 - INFO - codeparrot_training - Step 9837: {'lr': 0.0004678275457630447, 'samples': 5037056, 'steps': 9837, 'loss/train': 1.969691514968872} 02/24/2022 14:17:50 - INFO - codeparrot_training - Step 9838: {'lr': 0.0004678195157017704, 'samples': 5037568, 'steps': 9838, 'loss/train': 2.5779781341552734} 02/24/2022 14:17:53 - INFO - codeparrot_training - Step 9839: {'lr': 0.00046781148470742654, 'samples': 5038080, 'steps': 9839, 'loss/train': 2.4346694946289062} 02/24/2022 14:17:59 - INFO - codeparrot_training - Step 9840: {'lr': 0.0004678034527800474, 'samples': 5038592, 'steps': 9840, 'loss/train': 1.6345288753509521} 02/24/2022 14:18:02 - INFO - codeparrot_training - Step 9841: {'lr': 0.0004677954199196674, 'samples': 5039104, 'steps': 9841, 'loss/train': 2.04012393951416} 02/24/2022 14:18:08 - INFO - codeparrot_training - Step 9842: {'lr': 0.00046778738612632097, 'samples': 5039616, 'steps': 9842, 'loss/train': 0.6263843774795532} 02/24/2022 14:18:11 - INFO - codeparrot_training - Step 9843: {'lr': 0.00046777935140004256, 'samples': 5040128, 'steps': 9843, 'loss/train': 1.6836732625961304} 02/24/2022 14:18:17 - INFO - codeparrot_training - Step 9844: {'lr': 0.00046777131574086663, 'samples': 5040640, 'steps': 9844, 'loss/train': 2.8924083709716797} 02/24/2022 14:18:20 - INFO - codeparrot_training - Step 9845: {'lr': 0.0004677632791488274, 'samples': 5041152, 'steps': 9845, 'loss/train': 1.0286898612976074} 02/24/2022 14:18:28 - INFO - codeparrot_training - Step 9846: {'lr': 0.00046775524162395954, 'samples': 5041664, 'steps': 9846, 'loss/train': 1.0133346319198608} 02/24/2022 14:18:31 - INFO - codeparrot_training - Step 9847: {'lr': 0.00046774720316629734, 'samples': 5042176, 'steps': 9847, 'loss/train': 1.944876790046692} 02/24/2022 14:18:37 - INFO - codeparrot_training - Step 9848: {'lr': 0.00046773916377587524, 'samples': 5042688, 'steps': 9848, 'loss/train': 1.5478363037109375} 02/24/2022 14:18:40 - INFO - codeparrot_training - Step 9849: {'lr': 0.00046773112345272773, 'samples': 5043200, 'steps': 9849, 'loss/train': 2.0946617126464844} 02/24/2022 14:18:46 - INFO - codeparrot_training - Step 9850: {'lr': 0.0004677230821968892, 'samples': 5043712, 'steps': 9850, 'loss/train': 1.566332459449768} 02/24/2022 14:18:49 - INFO - codeparrot_training - Step 9851: {'lr': 0.00046771504000839417, 'samples': 5044224, 'steps': 9851, 'loss/train': 1.9542189836502075} 02/24/2022 14:18:55 - INFO - codeparrot_training - Step 9852: {'lr': 0.0004677069968872769, 'samples': 5044736, 'steps': 9852, 'loss/train': 2.1629180908203125} 02/24/2022 14:18:58 - INFO - codeparrot_training - Step 9853: {'lr': 0.0004676989528335721, 'samples': 5045248, 'steps': 9853, 'loss/train': 2.5073561668395996} 02/24/2022 14:19:04 - INFO - codeparrot_training - Step 9854: {'lr': 0.0004676909078473142, 'samples': 5045760, 'steps': 9854, 'loss/train': 2.185811758041382} 02/24/2022 14:19:07 - INFO - codeparrot_training - Step 9855: {'lr': 0.00046768286192853736, 'samples': 5046272, 'steps': 9855, 'loss/train': 1.9570590257644653} 02/24/2022 14:19:14 - INFO - codeparrot_training - Step 9856: {'lr': 0.00046767481507727646, 'samples': 5046784, 'steps': 9856, 'loss/train': 2.4177072048187256} 02/24/2022 14:19:18 - INFO - codeparrot_training - Step 9857: {'lr': 0.00046766676729356564, 'samples': 5047296, 'steps': 9857, 'loss/train': 3.0085244178771973} 02/24/2022 14:19:23 - INFO - codeparrot_training - Step 9858: {'lr': 0.0004676587185774396, 'samples': 5047808, 'steps': 9858, 'loss/train': 2.3172130584716797} 02/24/2022 14:19:27 - INFO - codeparrot_training - Step 9859: {'lr': 0.00046765066892893266, 'samples': 5048320, 'steps': 9859, 'loss/train': 2.5215938091278076} 02/24/2022 14:19:32 - INFO - codeparrot_training - Step 9860: {'lr': 0.00046764261834807944, 'samples': 5048832, 'steps': 9860, 'loss/train': 1.548329472541809} 02/24/2022 14:19:36 - INFO - codeparrot_training - Step 9861: {'lr': 0.0004676345668349142, 'samples': 5049344, 'steps': 9861, 'loss/train': 2.180375337600708} 02/24/2022 14:19:42 - INFO - codeparrot_training - Step 9862: {'lr': 0.0004676265143894717, 'samples': 5049856, 'steps': 9862, 'loss/train': 1.731307864189148} 02/24/2022 14:19:45 - INFO - codeparrot_training - Step 9863: {'lr': 0.0004676184610117863, 'samples': 5050368, 'steps': 9863, 'loss/train': 3.3570234775543213} 02/24/2022 14:19:51 - INFO - codeparrot_training - Step 9864: {'lr': 0.0004676104067018925, 'samples': 5050880, 'steps': 9864, 'loss/train': 2.9153382778167725} 02/24/2022 14:19:54 - INFO - codeparrot_training - Step 9865: {'lr': 0.0004676023514598249, 'samples': 5051392, 'steps': 9865, 'loss/train': 2.4567854404449463} 02/24/2022 14:20:01 - INFO - codeparrot_training - Step 9866: {'lr': 0.0004675942952856178, 'samples': 5051904, 'steps': 9866, 'loss/train': 1.0560890436172485} 02/24/2022 14:20:05 - INFO - codeparrot_training - Step 9867: {'lr': 0.0004675862381793059, 'samples': 5052416, 'steps': 9867, 'loss/train': 3.4936273097991943} 02/24/2022 14:20:10 - INFO - codeparrot_training - Step 9868: {'lr': 0.0004675781801409236, 'samples': 5052928, 'steps': 9868, 'loss/train': 0.6611115336418152} 02/24/2022 14:20:14 - INFO - codeparrot_training - Step 9869: {'lr': 0.00046757012117050554, 'samples': 5053440, 'steps': 9869, 'loss/train': 2.094862699508667} 02/24/2022 14:20:19 - INFO - codeparrot_training - Step 9870: {'lr': 0.00046756206126808607, 'samples': 5053952, 'steps': 9870, 'loss/train': 1.258583664894104} 02/24/2022 14:20:23 - INFO - codeparrot_training - Step 9871: {'lr': 0.0004675540004336999, 'samples': 5054464, 'steps': 9871, 'loss/train': 1.1366286277770996} 02/24/2022 14:20:28 - INFO - codeparrot_training - Step 9872: {'lr': 0.00046754593866738144, 'samples': 5054976, 'steps': 9872, 'loss/train': 2.1207761764526367} 02/24/2022 14:20:32 - INFO - codeparrot_training - Step 9873: {'lr': 0.0004675378759691652, 'samples': 5055488, 'steps': 9873, 'loss/train': 2.7102530002593994} 02/24/2022 14:20:37 - INFO - codeparrot_training - Step 9874: {'lr': 0.00046752981233908587, 'samples': 5056000, 'steps': 9874, 'loss/train': 5.332662582397461} 02/24/2022 14:20:41 - INFO - codeparrot_training - Step 9875: {'lr': 0.0004675217477771779, 'samples': 5056512, 'steps': 9875, 'loss/train': 1.0341824293136597} 02/24/2022 14:20:46 - INFO - codeparrot_training - Step 9876: {'lr': 0.0004675136822834758, 'samples': 5057024, 'steps': 9876, 'loss/train': 2.6010963916778564} 02/24/2022 14:20:50 - INFO - codeparrot_training - Step 9877: {'lr': 0.0004675056158580141, 'samples': 5057536, 'steps': 9877, 'loss/train': 2.239691972732544} 02/24/2022 14:20:57 - INFO - codeparrot_training - Step 9878: {'lr': 0.0004674975485008275, 'samples': 5058048, 'steps': 9878, 'loss/train': 1.7435237169265747} 02/24/2022 14:21:01 - INFO - codeparrot_training - Step 9879: {'lr': 0.00046748948021195036, 'samples': 5058560, 'steps': 9879, 'loss/train': 2.204639434814453} 02/24/2022 14:21:06 - INFO - codeparrot_training - Step 9880: {'lr': 0.0004674814109914174, 'samples': 5059072, 'steps': 9880, 'loss/train': 1.5364415645599365} 02/24/2022 14:21:10 - INFO - codeparrot_training - Step 9881: {'lr': 0.00046747334083926316, 'samples': 5059584, 'steps': 9881, 'loss/train': 2.10906720161438} 02/24/2022 14:21:15 - INFO - codeparrot_training - Step 9882: {'lr': 0.0004674652697555222, 'samples': 5060096, 'steps': 9882, 'loss/train': 1.8647351264953613} 02/24/2022 14:21:19 - INFO - codeparrot_training - Step 9883: {'lr': 0.000467457197740229, 'samples': 5060608, 'steps': 9883, 'loss/train': 2.3197133541107178} 02/24/2022 14:21:24 - INFO - codeparrot_training - Step 9884: {'lr': 0.00046744912479341826, 'samples': 5061120, 'steps': 9884, 'loss/train': 2.5268044471740723} 02/24/2022 14:21:28 - INFO - codeparrot_training - Step 9885: {'lr': 0.0004674410509151246, 'samples': 5061632, 'steps': 9885, 'loss/train': 2.8732500076293945} 02/24/2022 14:21:33 - INFO - codeparrot_training - Step 9886: {'lr': 0.0004674329761053824, 'samples': 5062144, 'steps': 9886, 'loss/train': 2.8711466789245605} 02/24/2022 14:21:37 - INFO - codeparrot_training - Step 9887: {'lr': 0.00046742490036422635, 'samples': 5062656, 'steps': 9887, 'loss/train': 2.5218918323516846} 02/24/2022 14:21:42 - INFO - codeparrot_training - Step 9888: {'lr': 0.00046741682369169115, 'samples': 5063168, 'steps': 9888, 'loss/train': 2.0408434867858887} 02/24/2022 14:21:46 - INFO - codeparrot_training - Step 9889: {'lr': 0.00046740874608781126, 'samples': 5063680, 'steps': 9889, 'loss/train': 1.8912031650543213} 02/24/2022 14:21:51 - INFO - codeparrot_training - Step 9890: {'lr': 0.0004674006675526214, 'samples': 5064192, 'steps': 9890, 'loss/train': 2.1615066528320312} 02/24/2022 14:21:55 - INFO - codeparrot_training - Step 9891: {'lr': 0.00046739258808615607, 'samples': 5064704, 'steps': 9891, 'loss/train': 1.8317413330078125} 02/24/2022 14:22:02 - INFO - codeparrot_training - Step 9892: {'lr': 0.00046738450768845, 'samples': 5065216, 'steps': 9892, 'loss/train': 0.7311815023422241} 02/24/2022 14:22:06 - INFO - codeparrot_training - Step 9893: {'lr': 0.0004673764263595376, 'samples': 5065728, 'steps': 9893, 'loss/train': 0.9409798383712769} 02/24/2022 14:22:11 - INFO - codeparrot_training - Step 9894: {'lr': 0.00046736834409945364, 'samples': 5066240, 'steps': 9894, 'loss/train': 0.8589231371879578} 02/24/2022 14:22:15 - INFO - codeparrot_training - Step 9895: {'lr': 0.0004673602609082328, 'samples': 5066752, 'steps': 9895, 'loss/train': 0.2276872843503952} 02/24/2022 14:22:20 - INFO - codeparrot_training - Step 9896: {'lr': 0.00046735217678590957, 'samples': 5067264, 'steps': 9896, 'loss/train': 1.5629442930221558} 02/24/2022 14:22:24 - INFO - codeparrot_training - Step 9897: {'lr': 0.0004673440917325186, 'samples': 5067776, 'steps': 9897, 'loss/train': 2.7554752826690674} 02/24/2022 14:22:29 - INFO - codeparrot_training - Step 9898: {'lr': 0.00046733600574809465, 'samples': 5068288, 'steps': 9898, 'loss/train': 1.9369906187057495} 02/24/2022 14:22:33 - INFO - codeparrot_training - Step 9899: {'lr': 0.0004673279188326722, 'samples': 5068800, 'steps': 9899, 'loss/train': 2.524409055709839} 02/24/2022 14:22:38 - INFO - codeparrot_training - Step 9900: {'lr': 0.00046731983098628597, 'samples': 5069312, 'steps': 9900, 'loss/train': 1.9200384616851807} 02/24/2022 14:22:42 - INFO - codeparrot_training - Step 9901: {'lr': 0.00046731174220897054, 'samples': 5069824, 'steps': 9901, 'loss/train': 8.869290351867676} 02/24/2022 14:22:50 - INFO - codeparrot_training - Step 9902: {'lr': 0.0004673036525007607, 'samples': 5070336, 'steps': 9902, 'loss/train': 2.715398073196411} 02/24/2022 14:22:53 - INFO - codeparrot_training - Step 9903: {'lr': 0.0004672955618616909, 'samples': 5070848, 'steps': 9903, 'loss/train': 1.3328404426574707} 02/24/2022 14:22:59 - INFO - codeparrot_training - Step 9904: {'lr': 0.00046728747029179594, 'samples': 5071360, 'steps': 9904, 'loss/train': 2.451209783554077} 02/24/2022 14:23:02 - INFO - codeparrot_training - Step 9905: {'lr': 0.00046727937779111054, 'samples': 5071872, 'steps': 9905, 'loss/train': 1.2708052396774292} 02/24/2022 14:23:08 - INFO - codeparrot_training - Step 9906: {'lr': 0.0004672712843596693, 'samples': 5072384, 'steps': 9906, 'loss/train': 2.001110076904297} 02/24/2022 14:23:11 - INFO - codeparrot_training - Step 9907: {'lr': 0.0004672631899975067, 'samples': 5072896, 'steps': 9907, 'loss/train': 1.4222224950790405} 02/24/2022 14:23:17 - INFO - codeparrot_training - Step 9908: {'lr': 0.0004672550947046577, 'samples': 5073408, 'steps': 9908, 'loss/train': 2.0502870082855225} 02/24/2022 14:23:20 - INFO - codeparrot_training - Step 9909: {'lr': 0.0004672469984811568, 'samples': 5073920, 'steps': 9909, 'loss/train': 0.7946783304214478} 02/24/2022 14:23:26 - INFO - codeparrot_training - Step 9910: {'lr': 0.00046723890132703886, 'samples': 5074432, 'steps': 9910, 'loss/train': 3.2919411659240723} 02/24/2022 14:23:29 - INFO - codeparrot_training - Step 9911: {'lr': 0.0004672308032423384, 'samples': 5074944, 'steps': 9911, 'loss/train': 0.11130419373512268} 02/24/2022 14:23:35 - INFO - codeparrot_training - Step 9912: {'lr': 0.0004672227042270901, 'samples': 5075456, 'steps': 9912, 'loss/train': 1.9338780641555786} 02/24/2022 14:23:38 - INFO - codeparrot_training - Step 9913: {'lr': 0.00046721460428132873, 'samples': 5075968, 'steps': 9913, 'loss/train': 0.8318203091621399} 02/24/2022 14:23:46 - INFO - codeparrot_training - Step 9914: {'lr': 0.00046720650340508895, 'samples': 5076480, 'steps': 9914, 'loss/train': 3.330626964569092} 02/24/2022 14:23:49 - INFO - codeparrot_training - Step 9915: {'lr': 0.00046719840159840557, 'samples': 5076992, 'steps': 9915, 'loss/train': 2.098567485809326} 02/24/2022 14:23:55 - INFO - codeparrot_training - Step 9916: {'lr': 0.00046719029886131317, 'samples': 5077504, 'steps': 9916, 'loss/train': 1.9815986156463623} 02/24/2022 14:23:58 - INFO - codeparrot_training - Step 9917: {'lr': 0.0004671821951938464, 'samples': 5078016, 'steps': 9917, 'loss/train': 0.5972822904586792} 02/24/2022 14:24:04 - INFO - codeparrot_training - Step 9918: {'lr': 0.0004671740905960401, 'samples': 5078528, 'steps': 9918, 'loss/train': 0.6264111399650574} 02/24/2022 14:24:07 - INFO - codeparrot_training - Step 9919: {'lr': 0.00046716598506792905, 'samples': 5079040, 'steps': 9919, 'loss/train': 1.2439059019088745} 02/24/2022 14:24:13 - INFO - codeparrot_training - Step 9920: {'lr': 0.00046715787860954785, 'samples': 5079552, 'steps': 9920, 'loss/train': 1.1996192932128906} 02/24/2022 14:24:16 - INFO - codeparrot_training - Step 9921: {'lr': 0.0004671497712209312, 'samples': 5080064, 'steps': 9921, 'loss/train': 3.2870521545410156} 02/24/2022 14:24:22 - INFO - codeparrot_training - Step 9922: {'lr': 0.0004671416629021139, 'samples': 5080576, 'steps': 9922, 'loss/train': 1.2741830348968506} 02/24/2022 14:24:25 - INFO - codeparrot_training - Step 9923: {'lr': 0.0004671335536531307, 'samples': 5081088, 'steps': 9923, 'loss/train': 2.3450915813446045} 02/24/2022 14:24:32 - INFO - codeparrot_training - Step 9924: {'lr': 0.00046712544347401623, 'samples': 5081600, 'steps': 9924, 'loss/train': 2.386737108230591} 02/24/2022 14:24:36 - INFO - codeparrot_training - Step 9925: {'lr': 0.0004671173323648054, 'samples': 5082112, 'steps': 9925, 'loss/train': 2.5172245502471924} 02/24/2022 14:24:41 - INFO - codeparrot_training - Step 9926: {'lr': 0.00046710922032553283, 'samples': 5082624, 'steps': 9926, 'loss/train': 2.627974510192871} 02/24/2022 14:24:45 - INFO - codeparrot_training - Step 9927: {'lr': 0.00046710110735623326, 'samples': 5083136, 'steps': 9927, 'loss/train': 2.4732723236083984} 02/24/2022 14:24:50 - INFO - codeparrot_training - Step 9928: {'lr': 0.00046709299345694156, 'samples': 5083648, 'steps': 9928, 'loss/train': 2.4833905696868896} 02/24/2022 14:24:54 - INFO - codeparrot_training - Step 9929: {'lr': 0.00046708487862769235, 'samples': 5084160, 'steps': 9929, 'loss/train': 1.388100028038025} 02/24/2022 14:24:59 - INFO - codeparrot_training - Step 9930: {'lr': 0.0004670767628685204, 'samples': 5084672, 'steps': 9930, 'loss/train': 2.0667126178741455} 02/24/2022 14:25:03 - INFO - codeparrot_training - Step 9931: {'lr': 0.00046706864617946064, 'samples': 5085184, 'steps': 9931, 'loss/train': 2.4349706172943115} 02/24/2022 14:25:08 - INFO - codeparrot_training - Step 9932: {'lr': 0.0004670605285605477, 'samples': 5085696, 'steps': 9932, 'loss/train': 3.1051206588745117} 02/24/2022 14:25:12 - INFO - codeparrot_training - Step 9933: {'lr': 0.0004670524100118163, 'samples': 5086208, 'steps': 9933, 'loss/train': 1.7971051931381226} 02/24/2022 14:25:17 - INFO - codeparrot_training - Step 9934: {'lr': 0.00046704429053330137, 'samples': 5086720, 'steps': 9934, 'loss/train': 2.1329686641693115} 02/24/2022 14:25:21 - INFO - codeparrot_training - Step 9935: {'lr': 0.00046703617012503764, 'samples': 5087232, 'steps': 9935, 'loss/train': 1.5983407497406006} 02/24/2022 14:25:26 - INFO - codeparrot_training - Step 9936: {'lr': 0.00046702804878705987, 'samples': 5087744, 'steps': 9936, 'loss/train': 2.7194325923919678} 02/24/2022 14:25:29 - INFO - codeparrot_training - Step 9937: {'lr': 0.00046701992651940275, 'samples': 5088256, 'steps': 9937, 'loss/train': 2.697232723236084} 02/24/2022 14:25:37 - INFO - codeparrot_training - Step 9938: {'lr': 0.00046701180332210125, 'samples': 5088768, 'steps': 9938, 'loss/train': 2.7391722202301025} 02/24/2022 14:25:40 - INFO - codeparrot_training - Step 9939: {'lr': 0.0004670036791951901, 'samples': 5089280, 'steps': 9939, 'loss/train': 2.522125005722046} 02/24/2022 14:25:46 - INFO - codeparrot_training - Step 9940: {'lr': 0.0004669955541387041, 'samples': 5089792, 'steps': 9940, 'loss/train': 2.5859646797180176} 02/24/2022 14:25:49 - INFO - codeparrot_training - Step 9941: {'lr': 0.000466987428152678, 'samples': 5090304, 'steps': 9941, 'loss/train': 3.9463722705841064} 02/24/2022 14:25:55 - INFO - codeparrot_training - Step 9942: {'lr': 0.00046697930123714673, 'samples': 5090816, 'steps': 9942, 'loss/train': 0.7443364858627319} 02/24/2022 14:25:58 - INFO - codeparrot_training - Step 9943: {'lr': 0.000466971173392145, 'samples': 5091328, 'steps': 9943, 'loss/train': 1.603397011756897} 02/24/2022 14:26:04 - INFO - codeparrot_training - Step 9944: {'lr': 0.0004669630446177077, 'samples': 5091840, 'steps': 9944, 'loss/train': 1.6332961320877075} 02/24/2022 14:26:07 - INFO - codeparrot_training - Step 9945: {'lr': 0.00046695491491386955, 'samples': 5092352, 'steps': 9945, 'loss/train': 1.6669552326202393} 02/24/2022 14:26:13 - INFO - codeparrot_training - Step 9946: {'lr': 0.0004669467842806654, 'samples': 5092864, 'steps': 9946, 'loss/train': 2.0874228477478027} 02/24/2022 14:26:16 - INFO - codeparrot_training - Step 9947: {'lr': 0.00046693865271813016, 'samples': 5093376, 'steps': 9947, 'loss/train': 2.525726079940796} 02/24/2022 14:26:23 - INFO - codeparrot_training - Step 9948: {'lr': 0.0004669305202262987, 'samples': 5093888, 'steps': 9948, 'loss/train': 1.0922309160232544} 02/24/2022 14:26:27 - INFO - codeparrot_training - Step 9949: {'lr': 0.00046692238680520564, 'samples': 5094400, 'steps': 9949, 'loss/train': 1.541135311126709} 02/24/2022 14:26:32 - INFO - codeparrot_training - Step 9950: {'lr': 0.00046691425245488607, 'samples': 5094912, 'steps': 9950, 'loss/train': 2.081545114517212} 02/24/2022 14:26:36 - INFO - codeparrot_training - Step 9951: {'lr': 0.0004669061171753746, 'samples': 5095424, 'steps': 9951, 'loss/train': 2.0083301067352295} 02/24/2022 14:26:42 - INFO - codeparrot_training - Step 9952: {'lr': 0.0004668979809667063, 'samples': 5095936, 'steps': 9952, 'loss/train': 2.4489872455596924} 02/24/2022 14:26:45 - INFO - codeparrot_training - Step 9953: {'lr': 0.0004668898438289159, 'samples': 5096448, 'steps': 9953, 'loss/train': 0.8532716035842896} 02/24/2022 14:26:51 - INFO - codeparrot_training - Step 9954: {'lr': 0.00046688170576203827, 'samples': 5096960, 'steps': 9954, 'loss/train': 4.200775623321533} 02/24/2022 14:26:54 - INFO - codeparrot_training - Step 9955: {'lr': 0.00046687356676610825, 'samples': 5097472, 'steps': 9955, 'loss/train': 2.063951015472412} 02/24/2022 14:26:59 - INFO - codeparrot_training - Step 9956: {'lr': 0.00046686542684116073, 'samples': 5097984, 'steps': 9956, 'loss/train': 2.317816972732544} 02/24/2022 14:27:05 - INFO - codeparrot_training - Step 9957: {'lr': 0.00046685728598723063, 'samples': 5098496, 'steps': 9957, 'loss/train': 2.47550368309021} 02/24/2022 14:27:08 - INFO - codeparrot_training - Step 9958: {'lr': 0.00046684914420435275, 'samples': 5099008, 'steps': 9958, 'loss/train': 1.1509557962417603} 02/24/2022 14:27:16 - INFO - codeparrot_training - Step 9959: {'lr': 0.00046684100149256205, 'samples': 5099520, 'steps': 9959, 'loss/train': 2.005819320678711} 02/24/2022 14:27:19 - INFO - codeparrot_training - Step 9960: {'lr': 0.0004668328578518933, 'samples': 5100032, 'steps': 9960, 'loss/train': 2.0511999130249023} 02/24/2022 14:27:24 - INFO - codeparrot_training - Step 9961: {'lr': 0.0004668247132823814, 'samples': 5100544, 'steps': 9961, 'loss/train': 1.570127010345459} 02/24/2022 14:27:28 - INFO - codeparrot_training - Step 9962: {'lr': 0.00046681656778406136, 'samples': 5101056, 'steps': 9962, 'loss/train': 2.2478654384613037} 02/24/2022 14:27:34 - INFO - codeparrot_training - Step 9963: {'lr': 0.000466808421356968, 'samples': 5101568, 'steps': 9963, 'loss/train': 1.4018045663833618} 02/24/2022 14:27:37 - INFO - codeparrot_training - Step 9964: {'lr': 0.00046680027400113614, 'samples': 5102080, 'steps': 9964, 'loss/train': 2.3996355533599854} 02/24/2022 14:27:43 - INFO - codeparrot_training - Step 9965: {'lr': 0.0004667921257166008, 'samples': 5102592, 'steps': 9965, 'loss/train': 2.374163866043091} 02/24/2022 14:27:46 - INFO - codeparrot_training - Step 9966: {'lr': 0.00046678397650339677, 'samples': 5103104, 'steps': 9966, 'loss/train': 1.4455193281173706} 02/24/2022 14:27:51 - INFO - codeparrot_training - Step 9967: {'lr': 0.00046677582636155904, 'samples': 5103616, 'steps': 9967, 'loss/train': 2.928968667984009} 02/24/2022 14:27:55 - INFO - codeparrot_training - Step 9968: {'lr': 0.00046676767529112254, 'samples': 5104128, 'steps': 9968, 'loss/train': 2.4172563552856445} 02/24/2022 14:28:02 - INFO - codeparrot_training - Step 9969: {'lr': 0.0004667595232921221, 'samples': 5104640, 'steps': 9969, 'loss/train': 2.744659185409546} 02/24/2022 14:28:06 - INFO - codeparrot_training - Step 9970: {'lr': 0.00046675137036459273, 'samples': 5105152, 'steps': 9970, 'loss/train': 2.0865478515625} 02/24/2022 14:28:11 - INFO - codeparrot_training - Step 9971: {'lr': 0.0004667432165085693, 'samples': 5105664, 'steps': 9971, 'loss/train': 1.994526743888855} 02/24/2022 14:28:15 - INFO - codeparrot_training - Step 9972: {'lr': 0.00046673506172408675, 'samples': 5106176, 'steps': 9972, 'loss/train': 1.5064557790756226} 02/24/2022 14:28:20 - INFO - codeparrot_training - Step 9973: {'lr': 0.0004667269060111801, 'samples': 5106688, 'steps': 9973, 'loss/train': 1.2326725721359253} 02/24/2022 14:28:24 - INFO - codeparrot_training - Step 9974: {'lr': 0.0004667187493698841, 'samples': 5107200, 'steps': 9974, 'loss/train': 2.0082168579101562} 02/24/2022 14:28:30 - INFO - codeparrot_training - Step 9975: {'lr': 0.00046671059180023377, 'samples': 5107712, 'steps': 9975, 'loss/train': 2.4531643390655518} 02/24/2022 14:28:33 - INFO - codeparrot_training - Step 9976: {'lr': 0.0004667024333022642, 'samples': 5108224, 'steps': 9976, 'loss/train': 2.482114553451538} 02/24/2022 14:28:39 - INFO - codeparrot_training - Step 9977: {'lr': 0.00046669427387601017, 'samples': 5108736, 'steps': 9977, 'loss/train': 2.3870275020599365} 02/24/2022 14:28:42 - INFO - codeparrot_training - Step 9978: {'lr': 0.0004666861135215066, 'samples': 5109248, 'steps': 9978, 'loss/train': 2.2954249382019043} 02/24/2022 14:28:48 - INFO - codeparrot_training - Step 9979: {'lr': 0.0004666779522387886, 'samples': 5109760, 'steps': 9979, 'loss/train': 2.2171740531921387} 02/24/2022 14:28:51 - INFO - codeparrot_training - Step 9980: {'lr': 0.000466669790027891, 'samples': 5110272, 'steps': 9980, 'loss/train': 1.9995105266571045} 02/24/2022 14:28:57 - INFO - codeparrot_training - Step 9981: {'lr': 0.00046666162688884893, 'samples': 5110784, 'steps': 9981, 'loss/train': 2.267690420150757} 02/24/2022 14:29:00 - INFO - codeparrot_training - Step 9982: {'lr': 0.0004666534628216972, 'samples': 5111296, 'steps': 9982, 'loss/train': 2.74352765083313} 02/24/2022 14:29:06 - INFO - codeparrot_training - Step 9983: {'lr': 0.0004666452978264708, 'samples': 5111808, 'steps': 9983, 'loss/train': 2.284552812576294} 02/24/2022 14:29:09 - INFO - codeparrot_training - Step 9984: {'lr': 0.0004666371319032047, 'samples': 5112320, 'steps': 9984, 'loss/train': 2.207453489303589} 02/24/2022 14:29:17 - INFO - codeparrot_training - Step 9985: {'lr': 0.00046662896505193395, 'samples': 5112832, 'steps': 9985, 'loss/train': 2.010751962661743} 02/24/2022 14:29:20 - INFO - codeparrot_training - Step 9986: {'lr': 0.00046662079727269356, 'samples': 5113344, 'steps': 9986, 'loss/train': 1.626746654510498} 02/24/2022 14:29:26 - INFO - codeparrot_training - Step 9987: {'lr': 0.0004666126285655184, 'samples': 5113856, 'steps': 9987, 'loss/train': 2.2685327529907227} 02/24/2022 14:29:29 - INFO - codeparrot_training - Step 9988: {'lr': 0.0004666044589304436, 'samples': 5114368, 'steps': 9988, 'loss/train': 1.9239364862442017} 02/24/2022 14:29:33 - INFO - codeparrot_training - Step 9989: {'lr': 0.000466596288367504, 'samples': 5114880, 'steps': 9989, 'loss/train': 2.3868231773376465} 02/24/2022 14:29:38 - INFO - codeparrot_training - Step 9990: {'lr': 0.0004665881168767346, 'samples': 5115392, 'steps': 9990, 'loss/train': 1.8015762567520142} 02/24/2022 14:29:42 - INFO - codeparrot_training - Step 9991: {'lr': 0.00046657994445817064, 'samples': 5115904, 'steps': 9991, 'loss/train': 1.8796157836914062} 02/24/2022 14:29:48 - INFO - codeparrot_training - Step 9992: {'lr': 0.0004665717711118469, 'samples': 5116416, 'steps': 9992, 'loss/train': 2.834174633026123} 02/24/2022 14:29:51 - INFO - codeparrot_training - Step 9993: {'lr': 0.00046656359683779845, 'samples': 5116928, 'steps': 9993, 'loss/train': 2.1030397415161133} 02/24/2022 14:29:57 - INFO - codeparrot_training - Step 9994: {'lr': 0.00046655542163606033, 'samples': 5117440, 'steps': 9994, 'loss/train': 1.9597599506378174} 02/24/2022 14:30:04 - INFO - codeparrot_training - Step 9995: {'lr': 0.0004665472455066675, 'samples': 5117952, 'steps': 9995, 'loss/train': 2.9539687633514404} 02/24/2022 14:30:08 - INFO - codeparrot_training - Step 9996: {'lr': 0.0004665390684496551, 'samples': 5118464, 'steps': 9996, 'loss/train': 2.8192200660705566} 02/24/2022 14:30:13 - INFO - codeparrot_training - Step 9997: {'lr': 0.0004665308904650581, 'samples': 5118976, 'steps': 9997, 'loss/train': 1.3336280584335327} 02/24/2022 14:30:17 - INFO - codeparrot_training - Step 9998: {'lr': 0.00046652271155291146, 'samples': 5119488, 'steps': 9998, 'loss/train': 2.2988455295562744} 02/24/2022 14:30:22 - INFO - codeparrot_training - Step 9999: {'lr': 0.0004665145317132503, 'samples': 5120000, 'steps': 9999, 'loss/train': 1.8172866106033325} 02/24/2022 14:30:22 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 14:30:39 - WARNING - huggingface_hub.repository - Several commits (10) will be pushed upstream. 02/24/2022 14:30:39 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 14:31:13 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 8cd23e2..c2659cd floral-grass-11 -> floral-grass-11 02/24/2022 14:31:18 - INFO - codeparrot_training - Step 10000: {'lr': 0.00046650635094610973, 'samples': 5120512, 'steps': 10000, 'loss/train': 1.3753174543380737} 02/24/2022 14:31:23 - INFO - codeparrot_training - Step 10001: {'lr': 0.00046649816925152456, 'samples': 5121024, 'steps': 10001, 'loss/train': 2.441896438598633} 02/24/2022 14:31:27 - INFO - codeparrot_training - Step 10002: {'lr': 0.00046648998662953003, 'samples': 5121536, 'steps': 10002, 'loss/train': 1.613208293914795} 02/24/2022 14:31:30 - INFO - codeparrot_training - Step 10003: {'lr': 0.00046648180308016116, 'samples': 5122048, 'steps': 10003, 'loss/train': 3.103625774383545} 02/24/2022 14:31:36 - INFO - codeparrot_training - Step 10004: {'lr': 0.00046647361860345293, 'samples': 5122560, 'steps': 10004, 'loss/train': 2.8003487586975098} 02/24/2022 14:31:39 - INFO - codeparrot_training - Step 10005: {'lr': 0.00046646543319944057, 'samples': 5123072, 'steps': 10005, 'loss/train': 3.8220736980438232} 02/24/2022 14:31:46 - INFO - codeparrot_training - Step 10006: {'lr': 0.00046645724686815893, 'samples': 5123584, 'steps': 10006, 'loss/train': 2.9445724487304688} 02/24/2022 14:31:50 - INFO - codeparrot_training - Step 10007: {'lr': 0.00046644905960964325, 'samples': 5124096, 'steps': 10007, 'loss/train': 2.2180960178375244} 02/24/2022 14:31:55 - INFO - codeparrot_training - Step 10008: {'lr': 0.00046644087142392845, 'samples': 5124608, 'steps': 10008, 'loss/train': 2.4489901065826416} 02/24/2022 14:32:01 - INFO - codeparrot_training - Step 10009: {'lr': 0.00046643268231104975, 'samples': 5125120, 'steps': 10009, 'loss/train': 1.7591034173965454} 02/24/2022 14:32:04 - INFO - codeparrot_training - Step 10010: {'lr': 0.00046642449227104213, 'samples': 5125632, 'steps': 10010, 'loss/train': 2.5898399353027344} 02/24/2022 14:32:10 - INFO - codeparrot_training - Step 10011: {'lr': 0.00046641630130394066, 'samples': 5126144, 'steps': 10011, 'loss/train': 1.4844536781311035} 02/24/2022 14:32:13 - INFO - codeparrot_training - Step 10012: {'lr': 0.0004664081094097805, 'samples': 5126656, 'steps': 10012, 'loss/train': 2.326171398162842} 02/24/2022 14:32:19 - INFO - codeparrot_training - Step 10013: {'lr': 0.00046639991658859684, 'samples': 5127168, 'steps': 10013, 'loss/train': 1.8968192338943481} 02/24/2022 14:32:22 - INFO - codeparrot_training - Step 10014: {'lr': 0.00046639172284042453, 'samples': 5127680, 'steps': 10014, 'loss/train': 1.6029521226882935} 02/24/2022 14:32:29 - INFO - codeparrot_training - Step 10015: {'lr': 0.00046638352816529883, 'samples': 5128192, 'steps': 10015, 'loss/train': 1.8007721900939941} 02/24/2022 14:32:33 - INFO - codeparrot_training - Step 10016: {'lr': 0.00046637533256325476, 'samples': 5128704, 'steps': 10016, 'loss/train': 2.874091625213623} 02/24/2022 14:32:38 - INFO - codeparrot_training - Step 10017: {'lr': 0.0004663671360343275, 'samples': 5129216, 'steps': 10017, 'loss/train': 1.657929539680481} 02/24/2022 14:32:42 - INFO - codeparrot_training - Step 10018: {'lr': 0.00046635893857855217, 'samples': 5129728, 'steps': 10018, 'loss/train': 1.4173426628112793} 02/24/2022 14:32:47 - INFO - codeparrot_training - Step 10019: {'lr': 0.0004663507401959638, 'samples': 5130240, 'steps': 10019, 'loss/train': 2.41587233543396} 02/24/2022 14:32:51 - INFO - codeparrot_training - Step 10020: {'lr': 0.00046634254088659757, 'samples': 5130752, 'steps': 10020, 'loss/train': 2.928043842315674} 02/24/2022 14:32:56 - INFO - codeparrot_training - Step 10021: {'lr': 0.00046633434065048855, 'samples': 5131264, 'steps': 10021, 'loss/train': 2.5403943061828613} 02/24/2022 14:33:00 - INFO - codeparrot_training - Step 10022: {'lr': 0.000466326139487672, 'samples': 5131776, 'steps': 10022, 'loss/train': 2.6487104892730713} 02/24/2022 14:33:05 - INFO - codeparrot_training - Step 10023: {'lr': 0.0004663179373981829, 'samples': 5132288, 'steps': 10023, 'loss/train': 1.5769028663635254} 02/24/2022 14:33:09 - INFO - codeparrot_training - Step 10024: {'lr': 0.0004663097343820565, 'samples': 5132800, 'steps': 10024, 'loss/train': 1.0461530685424805} 02/24/2022 14:33:14 - INFO - codeparrot_training - Step 10025: {'lr': 0.00046630153043932784, 'samples': 5133312, 'steps': 10025, 'loss/train': 2.093100070953369} 02/24/2022 14:33:18 - INFO - codeparrot_training - Step 10026: {'lr': 0.00046629332557003215, 'samples': 5133824, 'steps': 10026, 'loss/train': 1.5797640085220337} 02/24/2022 14:33:23 - INFO - codeparrot_training - Step 10027: {'lr': 0.00046628511977420443, 'samples': 5134336, 'steps': 10027, 'loss/train': 2.1969242095947266} 02/24/2022 14:33:27 - INFO - codeparrot_training - Step 10028: {'lr': 0.00046627691305188004, 'samples': 5134848, 'steps': 10028, 'loss/train': 2.601287841796875} 02/24/2022 14:33:34 - INFO - codeparrot_training - Step 10029: {'lr': 0.00046626870540309394, 'samples': 5135360, 'steps': 10029, 'loss/train': 2.6741883754730225} 02/24/2022 14:33:38 - INFO - codeparrot_training - Step 10030: {'lr': 0.00046626049682788143, 'samples': 5135872, 'steps': 10030, 'loss/train': 1.5147557258605957} 02/24/2022 14:33:44 - INFO - codeparrot_training - Step 10031: {'lr': 0.00046625228732627763, 'samples': 5136384, 'steps': 10031, 'loss/train': 2.425189733505249} 02/24/2022 14:33:47 - INFO - codeparrot_training - Step 10032: {'lr': 0.00046624407689831773, 'samples': 5136896, 'steps': 10032, 'loss/train': 2.1255462169647217} 02/24/2022 14:33:53 - INFO - codeparrot_training - Step 10033: {'lr': 0.0004662358655440368, 'samples': 5137408, 'steps': 10033, 'loss/train': 2.444613456726074} 02/24/2022 14:33:56 - INFO - codeparrot_training - Step 10034: {'lr': 0.0004662276532634701, 'samples': 5137920, 'steps': 10034, 'loss/train': 2.3182966709136963} 02/24/2022 14:34:02 - INFO - codeparrot_training - Step 10035: {'lr': 0.0004662194400566528, 'samples': 5138432, 'steps': 10035, 'loss/train': 2.421915054321289} 02/24/2022 14:34:05 - INFO - codeparrot_training - Step 10036: {'lr': 0.0004662112259236201, 'samples': 5138944, 'steps': 10036, 'loss/train': 1.6606255769729614} 02/24/2022 14:34:11 - INFO - codeparrot_training - Step 10037: {'lr': 0.00046620301086440713, 'samples': 5139456, 'steps': 10037, 'loss/train': 2.000070810317993} 02/24/2022 14:34:14 - INFO - codeparrot_training - Step 10038: {'lr': 0.00046619479487904915, 'samples': 5139968, 'steps': 10038, 'loss/train': 1.8782694339752197} 02/24/2022 14:34:20 - INFO - codeparrot_training - Step 10039: {'lr': 0.0004661865779675813, 'samples': 5140480, 'steps': 10039, 'loss/train': 1.970636010169983} 02/24/2022 14:34:24 - INFO - codeparrot_training - Step 10040: {'lr': 0.0004661783601300388, 'samples': 5140992, 'steps': 10040, 'loss/train': 1.6922369003295898} 02/24/2022 14:34:31 - INFO - codeparrot_training - Step 10041: {'lr': 0.00046617014136645686, 'samples': 5141504, 'steps': 10041, 'loss/train': 2.4848549365997314} 02/24/2022 14:34:34 - INFO - codeparrot_training - Step 10042: {'lr': 0.00046616192167687066, 'samples': 5142016, 'steps': 10042, 'loss/train': 2.2349853515625} 02/24/2022 14:34:40 - INFO - codeparrot_training - Step 10043: {'lr': 0.00046615370106131536, 'samples': 5142528, 'steps': 10043, 'loss/train': 2.6010119915008545} 02/24/2022 14:34:43 - INFO - codeparrot_training - Step 10044: {'lr': 0.00046614547951982636, 'samples': 5143040, 'steps': 10044, 'loss/train': 2.7375640869140625} 02/24/2022 14:34:49 - INFO - codeparrot_training - Step 10045: {'lr': 0.00046613725705243873, 'samples': 5143552, 'steps': 10045, 'loss/train': 2.314051628112793} 02/24/2022 14:34:52 - INFO - codeparrot_training - Step 10046: {'lr': 0.0004661290336591877, 'samples': 5144064, 'steps': 10046, 'loss/train': 1.0697977542877197} 02/24/2022 14:34:58 - INFO - codeparrot_training - Step 10047: {'lr': 0.0004661208093401085, 'samples': 5144576, 'steps': 10047, 'loss/train': 2.632368564605713} 02/24/2022 14:35:01 - INFO - codeparrot_training - Step 10048: {'lr': 0.0004661125840952364, 'samples': 5145088, 'steps': 10048, 'loss/train': 3.0041842460632324} 02/24/2022 14:35:07 - INFO - codeparrot_training - Step 10049: {'lr': 0.0004661043579246066, 'samples': 5145600, 'steps': 10049, 'loss/train': 1.8701436519622803} 02/24/2022 14:35:10 - INFO - codeparrot_training - Step 10050: {'lr': 0.00046609613082825436, 'samples': 5146112, 'steps': 10050, 'loss/train': 3.2705018520355225} 02/24/2022 14:35:18 - INFO - codeparrot_training - Step 10051: {'lr': 0.00046608790280621494, 'samples': 5146624, 'steps': 10051, 'loss/train': 2.601288318634033} 02/24/2022 14:35:21 - INFO - codeparrot_training - Step 10052: {'lr': 0.0004660796738585235, 'samples': 5147136, 'steps': 10052, 'loss/train': 0.8518520593643188} 02/24/2022 14:35:26 - INFO - codeparrot_training - Step 10053: {'lr': 0.0004660714439852154, 'samples': 5147648, 'steps': 10053, 'loss/train': 1.528947114944458} 02/24/2022 14:35:30 - INFO - codeparrot_training - Step 10054: {'lr': 0.0004660632131863258, 'samples': 5148160, 'steps': 10054, 'loss/train': 1.620219111442566} 02/24/2022 14:35:35 - INFO - codeparrot_training - Step 10055: {'lr': 0.0004660549814618901, 'samples': 5148672, 'steps': 10055, 'loss/train': 2.8213744163513184} 02/24/2022 14:35:39 - INFO - codeparrot_training - Step 10056: {'lr': 0.00046604674881194335, 'samples': 5149184, 'steps': 10056, 'loss/train': 2.5623490810394287} 02/24/2022 14:35:44 - INFO - codeparrot_training - Step 10057: {'lr': 0.000466038515236521, 'samples': 5149696, 'steps': 10057, 'loss/train': 2.445793390274048} 02/24/2022 14:35:48 - INFO - codeparrot_training - Step 10058: {'lr': 0.0004660302807356582, 'samples': 5150208, 'steps': 10058, 'loss/train': 1.1359987258911133} 02/24/2022 14:35:53 - INFO - codeparrot_training - Step 10059: {'lr': 0.0004660220453093903, 'samples': 5150720, 'steps': 10059, 'loss/train': 0.461545467376709} 02/24/2022 14:35:57 - INFO - codeparrot_training - Step 10060: {'lr': 0.0004660138089577526, 'samples': 5151232, 'steps': 10060, 'loss/train': 2.7554819583892822} 02/24/2022 14:36:04 - INFO - codeparrot_training - Step 10061: {'lr': 0.00046600557168078026, 'samples': 5151744, 'steps': 10061, 'loss/train': 1.467819333076477} 02/24/2022 14:36:08 - INFO - codeparrot_training - Step 10062: {'lr': 0.0004659973334785087, 'samples': 5152256, 'steps': 10062, 'loss/train': 2.1805968284606934} 02/24/2022 14:36:13 - INFO - codeparrot_training - Step 10063: {'lr': 0.00046598909435097315, 'samples': 5152768, 'steps': 10063, 'loss/train': 2.4827444553375244} 02/24/2022 14:36:17 - INFO - codeparrot_training - Step 10064: {'lr': 0.0004659808542982088, 'samples': 5153280, 'steps': 10064, 'loss/train': 2.6353976726531982} 02/24/2022 14:36:22 - INFO - codeparrot_training - Step 10065: {'lr': 0.0004659726133202512, 'samples': 5153792, 'steps': 10065, 'loss/train': 1.5538828372955322} 02/24/2022 14:36:26 - INFO - codeparrot_training - Step 10066: {'lr': 0.0004659643714171354, 'samples': 5154304, 'steps': 10066, 'loss/train': 3.414149045944214} 02/24/2022 14:36:31 - INFO - codeparrot_training - Step 10067: {'lr': 0.00046595612858889686, 'samples': 5154816, 'steps': 10067, 'loss/train': 2.5281076431274414} 02/24/2022 14:36:35 - INFO - codeparrot_training - Step 10068: {'lr': 0.00046594788483557084, 'samples': 5155328, 'steps': 10068, 'loss/train': 2.7684803009033203} 02/24/2022 14:36:40 - INFO - codeparrot_training - Step 10069: {'lr': 0.00046593964015719257, 'samples': 5155840, 'steps': 10069, 'loss/train': 2.6864817142486572} 02/24/2022 14:36:44 - INFO - codeparrot_training - Step 10070: {'lr': 0.0004659313945537975, 'samples': 5156352, 'steps': 10070, 'loss/train': 2.4148175716400146} 02/24/2022 14:36:49 - INFO - codeparrot_training - Step 10071: {'lr': 0.00046592314802542095, 'samples': 5156864, 'steps': 10071, 'loss/train': 1.5033036470413208} 02/24/2022 14:36:53 - INFO - codeparrot_training - Step 10072: {'lr': 0.0004659149005720982, 'samples': 5157376, 'steps': 10072, 'loss/train': 1.4138095378875732} 02/24/2022 14:36:58 - INFO - codeparrot_training - Step 10073: {'lr': 0.00046590665219386454, 'samples': 5157888, 'steps': 10073, 'loss/train': 1.4330164194107056} 02/24/2022 14:37:02 - INFO - codeparrot_training - Step 10074: {'lr': 0.0004658984028907553, 'samples': 5158400, 'steps': 10074, 'loss/train': 1.749715805053711} 02/24/2022 14:37:09 - INFO - codeparrot_training - Step 10075: {'lr': 0.0004658901526628059, 'samples': 5158912, 'steps': 10075, 'loss/train': 2.8505806922912598} 02/24/2022 14:37:13 - INFO - codeparrot_training - Step 10076: {'lr': 0.00046588190151005163, 'samples': 5159424, 'steps': 10076, 'loss/train': 2.787666082382202} 02/24/2022 14:37:18 - INFO - codeparrot_training - Step 10077: {'lr': 0.00046587364943252783, 'samples': 5159936, 'steps': 10077, 'loss/train': 1.2507398128509521} 02/24/2022 14:37:22 - INFO - codeparrot_training - Step 10078: {'lr': 0.00046586539643026994, 'samples': 5160448, 'steps': 10078, 'loss/train': 3.095327854156494} 02/24/2022 14:37:27 - INFO - codeparrot_training - Step 10079: {'lr': 0.0004658571425033131, 'samples': 5160960, 'steps': 10079, 'loss/train': 2.252150774002075} 02/24/2022 14:37:31 - INFO - codeparrot_training - Step 10080: {'lr': 0.0004658488876516929, 'samples': 5161472, 'steps': 10080, 'loss/train': 0.7183483242988586} 02/24/2022 14:37:36 - INFO - codeparrot_training - Step 10081: {'lr': 0.0004658406318754446, 'samples': 5161984, 'steps': 10081, 'loss/train': 1.6840614080429077} 02/24/2022 14:37:40 - INFO - codeparrot_training - Step 10082: {'lr': 0.0004658323751746036, 'samples': 5162496, 'steps': 10082, 'loss/train': 1.894370675086975} 02/24/2022 14:37:45 - INFO - codeparrot_training - Step 10083: {'lr': 0.00046582411754920517, 'samples': 5163008, 'steps': 10083, 'loss/train': 2.1872527599334717} 02/24/2022 14:37:49 - INFO - codeparrot_training - Step 10084: {'lr': 0.0004658158589992848, 'samples': 5163520, 'steps': 10084, 'loss/train': 3.0928308963775635} 02/24/2022 14:37:54 - INFO - codeparrot_training - Step 10085: {'lr': 0.00046580759952487776, 'samples': 5164032, 'steps': 10085, 'loss/train': 1.4812055826187134} 02/24/2022 14:37:58 - INFO - codeparrot_training - Step 10086: {'lr': 0.00046579933912601956, 'samples': 5164544, 'steps': 10086, 'loss/train': 1.1863903999328613} 02/24/2022 14:38:05 - INFO - codeparrot_training - Step 10087: {'lr': 0.00046579107780274543, 'samples': 5165056, 'steps': 10087, 'loss/train': 1.3037912845611572} 02/24/2022 14:38:09 - INFO - codeparrot_training - Step 10088: {'lr': 0.00046578281555509094, 'samples': 5165568, 'steps': 10088, 'loss/train': 2.437044143676758} 02/24/2022 14:38:14 - INFO - codeparrot_training - Step 10089: {'lr': 0.0004657745523830914, 'samples': 5166080, 'steps': 10089, 'loss/train': 1.872549295425415} 02/24/2022 14:38:18 - INFO - codeparrot_training - Step 10090: {'lr': 0.0004657662882867821, 'samples': 5166592, 'steps': 10090, 'loss/train': 0.8082488775253296} 02/24/2022 14:38:23 - INFO - codeparrot_training - Step 10091: {'lr': 0.0004657580232661985, 'samples': 5167104, 'steps': 10091, 'loss/train': 1.8585642576217651} 02/24/2022 14:38:27 - INFO - codeparrot_training - Step 10092: {'lr': 0.00046574975732137613, 'samples': 5167616, 'steps': 10092, 'loss/train': 2.1034302711486816} 02/24/2022 14:38:32 - INFO - codeparrot_training - Step 10093: {'lr': 0.0004657414904523504, 'samples': 5168128, 'steps': 10093, 'loss/train': 2.5881094932556152} 02/24/2022 14:38:36 - INFO - codeparrot_training - Step 10094: {'lr': 0.0004657332226591565, 'samples': 5168640, 'steps': 10094, 'loss/train': 2.7554502487182617} 02/24/2022 14:38:41 - INFO - codeparrot_training - Step 10095: {'lr': 0.00046572495394183, 'samples': 5169152, 'steps': 10095, 'loss/train': 2.156113386154175} 02/24/2022 14:38:45 - INFO - codeparrot_training - Step 10096: {'lr': 0.00046571668430040624, 'samples': 5169664, 'steps': 10096, 'loss/train': 0.9854620099067688} 02/24/2022 14:38:52 - INFO - codeparrot_training - Step 10097: {'lr': 0.0004657084137349208, 'samples': 5170176, 'steps': 10097, 'loss/train': 1.3659499883651733} 02/24/2022 14:38:56 - INFO - codeparrot_training - Step 10098: {'lr': 0.0004657001422454089, 'samples': 5170688, 'steps': 10098, 'loss/train': 3.0307724475860596} 02/24/2022 14:39:01 - INFO - codeparrot_training - Step 10099: {'lr': 0.0004656918698319062, 'samples': 5171200, 'steps': 10099, 'loss/train': 1.0281710624694824} 02/24/2022 14:39:05 - INFO - codeparrot_training - Step 10100: {'lr': 0.00046568359649444796, 'samples': 5171712, 'steps': 10100, 'loss/train': 2.967120885848999} 02/24/2022 14:39:10 - INFO - codeparrot_training - Step 10101: {'lr': 0.0004656753222330697, 'samples': 5172224, 'steps': 10101, 'loss/train': 1.1926450729370117} 02/24/2022 14:39:14 - INFO - codeparrot_training - Step 10102: {'lr': 0.0004656670470478068, 'samples': 5172736, 'steps': 10102, 'loss/train': 2.2363133430480957} 02/24/2022 14:39:19 - INFO - codeparrot_training - Step 10103: {'lr': 0.0004656587709386948, 'samples': 5173248, 'steps': 10103, 'loss/train': 2.533914566040039} 02/24/2022 14:39:22 - INFO - codeparrot_training - Step 10104: {'lr': 0.00046565049390576906, 'samples': 5173760, 'steps': 10104, 'loss/train': 1.83980131149292} 02/24/2022 14:39:28 - INFO - codeparrot_training - Step 10105: {'lr': 0.0004656422159490652, 'samples': 5174272, 'steps': 10105, 'loss/train': 2.584303140640259} 02/24/2022 14:39:31 - INFO - codeparrot_training - Step 10106: {'lr': 0.00046563393706861847, 'samples': 5174784, 'steps': 10106, 'loss/train': 1.3872103691101074} 02/24/2022 14:39:37 - INFO - codeparrot_training - Step 10107: {'lr': 0.00046562565726446437, 'samples': 5175296, 'steps': 10107, 'loss/train': 2.090421676635742} 02/24/2022 14:39:40 - INFO - codeparrot_training - Step 10108: {'lr': 0.0004656173765366385, 'samples': 5175808, 'steps': 10108, 'loss/train': 2.775364637374878} 02/24/2022 14:39:46 - INFO - codeparrot_training - Step 10109: {'lr': 0.00046560909488517623, 'samples': 5176320, 'steps': 10109, 'loss/train': 3.0249388217926025} 02/24/2022 14:39:50 - INFO - codeparrot_training - Step 10110: {'lr': 0.0004656008123101131, 'samples': 5176832, 'steps': 10110, 'loss/train': 0.6997179985046387} 02/24/2022 14:39:56 - INFO - codeparrot_training - Step 10111: {'lr': 0.0004655925288114845, 'samples': 5177344, 'steps': 10111, 'loss/train': 1.3380553722381592} 02/24/2022 14:39:59 - INFO - codeparrot_training - Step 10112: {'lr': 0.000465584244389326, 'samples': 5177856, 'steps': 10112, 'loss/train': 2.2695019245147705} 02/24/2022 14:40:06 - INFO - codeparrot_training - Step 10113: {'lr': 0.000465575959043673, 'samples': 5178368, 'steps': 10113, 'loss/train': 0.9244392514228821} 02/24/2022 14:40:10 - INFO - codeparrot_training - Step 10114: {'lr': 0.0004655676727745611, 'samples': 5178880, 'steps': 10114, 'loss/train': 1.605212926864624} 02/24/2022 14:40:15 - INFO - codeparrot_training - Step 10115: {'lr': 0.0004655593855820257, 'samples': 5179392, 'steps': 10115, 'loss/train': 1.8432780504226685} 02/24/2022 14:40:19 - INFO - codeparrot_training - Step 10116: {'lr': 0.00046555109746610244, 'samples': 5179904, 'steps': 10116, 'loss/train': 1.3577485084533691} 02/24/2022 14:40:24 - INFO - codeparrot_training - Step 10117: {'lr': 0.0004655428084268266, 'samples': 5180416, 'steps': 10117, 'loss/train': 1.7021634578704834} 02/24/2022 14:40:28 - INFO - codeparrot_training - Step 10118: {'lr': 0.00046553451846423387, 'samples': 5180928, 'steps': 10118, 'loss/train': 2.499723434448242} 02/24/2022 14:40:33 - INFO - codeparrot_training - Step 10119: {'lr': 0.0004655262275783597, 'samples': 5181440, 'steps': 10119, 'loss/train': 0.8837375640869141} 02/24/2022 14:40:37 - INFO - codeparrot_training - Step 10120: {'lr': 0.00046551793576923964, 'samples': 5181952, 'steps': 10120, 'loss/train': 1.524060606956482} 02/24/2022 14:40:42 - INFO - codeparrot_training - Step 10121: {'lr': 0.0004655096430369091, 'samples': 5182464, 'steps': 10121, 'loss/train': 1.3317874670028687} 02/24/2022 14:40:46 - INFO - codeparrot_training - Step 10122: {'lr': 0.00046550134938140375, 'samples': 5182976, 'steps': 10122, 'loss/train': 1.8425887823104858} 02/24/2022 14:40:53 - INFO - codeparrot_training - Step 10123: {'lr': 0.00046549305480275894, 'samples': 5183488, 'steps': 10123, 'loss/train': 1.166218876838684} 02/24/2022 14:40:56 - INFO - codeparrot_training - Step 10124: {'lr': 0.0004654847593010104, 'samples': 5184000, 'steps': 10124, 'loss/train': 1.6765146255493164} 02/24/2022 14:41:02 - INFO - codeparrot_training - Step 10125: {'lr': 0.00046547646287619363, 'samples': 5184512, 'steps': 10125, 'loss/train': 4.011613368988037} 02/24/2022 14:41:05 - INFO - codeparrot_training - Step 10126: {'lr': 0.00046546816552834404, 'samples': 5185024, 'steps': 10126, 'loss/train': 2.2532076835632324} 02/24/2022 14:41:11 - INFO - codeparrot_training - Step 10127: {'lr': 0.00046545986725749725, 'samples': 5185536, 'steps': 10127, 'loss/train': 2.7270307540893555} 02/24/2022 14:41:14 - INFO - codeparrot_training - Step 10128: {'lr': 0.0004654515680636888, 'samples': 5186048, 'steps': 10128, 'loss/train': 2.2106196880340576} 02/24/2022 14:41:20 - INFO - codeparrot_training - Step 10129: {'lr': 0.00046544326794695424, 'samples': 5186560, 'steps': 10129, 'loss/train': 3.012795925140381} 02/24/2022 14:41:24 - INFO - codeparrot_training - Step 10130: {'lr': 0.00046543496690732914, 'samples': 5187072, 'steps': 10130, 'loss/train': 2.5087478160858154} 02/24/2022 14:41:29 - INFO - codeparrot_training - Step 10131: {'lr': 0.0004654266649448491, 'samples': 5187584, 'steps': 10131, 'loss/train': 2.468700885772705} 02/24/2022 14:41:33 - INFO - codeparrot_training - Step 10132: {'lr': 0.00046541836205954955, 'samples': 5188096, 'steps': 10132, 'loss/train': 1.8703206777572632} 02/24/2022 14:41:40 - INFO - codeparrot_training - Step 10133: {'lr': 0.0004654100582514662, 'samples': 5188608, 'steps': 10133, 'loss/train': 2.7259156703948975} 02/24/2022 14:41:43 - INFO - codeparrot_training - Step 10134: {'lr': 0.0004654017535206345, 'samples': 5189120, 'steps': 10134, 'loss/train': 1.6057307720184326} 02/24/2022 14:41:49 - INFO - codeparrot_training - Step 10135: {'lr': 0.00046539344786709013, 'samples': 5189632, 'steps': 10135, 'loss/train': 2.871105670928955} 02/24/2022 14:41:52 - INFO - codeparrot_training - Step 10136: {'lr': 0.0004653851412908686, 'samples': 5190144, 'steps': 10136, 'loss/train': 1.8268897533416748} 02/24/2022 14:41:58 - INFO - codeparrot_training - Step 10137: {'lr': 0.0004653768337920056, 'samples': 5190656, 'steps': 10137, 'loss/train': 2.5554840564727783} 02/24/2022 14:42:02 - INFO - codeparrot_training - Step 10138: {'lr': 0.00046536852537053654, 'samples': 5191168, 'steps': 10138, 'loss/train': 1.1800227165222168} 02/24/2022 14:42:07 - INFO - codeparrot_training - Step 10139: {'lr': 0.00046536021602649715, 'samples': 5191680, 'steps': 10139, 'loss/train': 2.837959051132202} 02/24/2022 14:42:10 - INFO - codeparrot_training - Step 10140: {'lr': 0.0004653519057599229, 'samples': 5192192, 'steps': 10140, 'loss/train': 2.6552610397338867} 02/24/2022 14:42:16 - INFO - codeparrot_training - Step 10141: {'lr': 0.0004653435945708496, 'samples': 5192704, 'steps': 10141, 'loss/train': 2.022085189819336} 02/24/2022 14:42:19 - INFO - codeparrot_training - Step 10142: {'lr': 0.00046533528245931266, 'samples': 5193216, 'steps': 10142, 'loss/train': 1.8562556505203247} 02/24/2022 14:42:25 - INFO - codeparrot_training - Step 10143: {'lr': 0.0004653269694253477, 'samples': 5193728, 'steps': 10143, 'loss/train': 0.3280055522918701} 02/24/2022 14:42:28 - INFO - codeparrot_training - Step 10144: {'lr': 0.00046531865546899044, 'samples': 5194240, 'steps': 10144, 'loss/train': 1.831819772720337} 02/24/2022 14:42:36 - INFO - codeparrot_training - Step 10145: {'lr': 0.00046531034059027644, 'samples': 5194752, 'steps': 10145, 'loss/train': 1.441911220550537} 02/24/2022 14:42:39 - INFO - codeparrot_training - Step 10146: {'lr': 0.0004653020247892412, 'samples': 5195264, 'steps': 10146, 'loss/train': 2.196563243865967} 02/24/2022 14:42:45 - INFO - codeparrot_training - Step 10147: {'lr': 0.0004652937080659206, 'samples': 5195776, 'steps': 10147, 'loss/train': 2.6439316272735596} 02/24/2022 14:42:48 - INFO - codeparrot_training - Step 10148: {'lr': 0.00046528539042035, 'samples': 5196288, 'steps': 10148, 'loss/train': 2.147430658340454} 02/24/2022 14:42:54 - INFO - codeparrot_training - Step 10149: {'lr': 0.0004652770718525652, 'samples': 5196800, 'steps': 10149, 'loss/train': 1.8699212074279785} 02/24/2022 14:42:57 - INFO - codeparrot_training - Step 10150: {'lr': 0.0004652687523626018, 'samples': 5197312, 'steps': 10150, 'loss/train': 2.5532984733581543} 02/24/2022 14:43:03 - INFO - codeparrot_training - Step 10151: {'lr': 0.0004652604319504954, 'samples': 5197824, 'steps': 10151, 'loss/train': 2.511273145675659} 02/24/2022 14:43:06 - INFO - codeparrot_training - Step 10152: {'lr': 0.00046525211061628163, 'samples': 5198336, 'steps': 10152, 'loss/train': 1.8393397331237793} 02/24/2022 14:43:12 - INFO - codeparrot_training - Step 10153: {'lr': 0.0004652437883599962, 'samples': 5198848, 'steps': 10153, 'loss/train': 1.9094717502593994} 02/24/2022 14:43:15 - INFO - codeparrot_training - Step 10154: {'lr': 0.0004652354651816747, 'samples': 5199360, 'steps': 10154, 'loss/train': 1.2962555885314941} 02/24/2022 14:43:21 - INFO - codeparrot_training - Step 10155: {'lr': 0.0004652271410813529, 'samples': 5199872, 'steps': 10155, 'loss/train': 1.813214898109436} 02/24/2022 14:43:24 - INFO - codeparrot_training - Step 10156: {'lr': 0.0004652188160590663, 'samples': 5200384, 'steps': 10156, 'loss/train': 1.9466413259506226} 02/24/2022 14:43:30 - INFO - codeparrot_training - Step 10157: {'lr': 0.00046521049011485064, 'samples': 5200896, 'steps': 10157, 'loss/train': 2.275838851928711} 02/24/2022 14:43:33 - INFO - codeparrot_training - Step 10158: {'lr': 0.0004652021632487415, 'samples': 5201408, 'steps': 10158, 'loss/train': 1.5771344900131226} 02/24/2022 14:43:40 - INFO - codeparrot_training - Step 10159: {'lr': 0.00046519383546077476, 'samples': 5201920, 'steps': 10159, 'loss/train': 0.6537622213363647} 02/24/2022 14:43:44 - INFO - codeparrot_training - Step 10160: {'lr': 0.0004651855067509859, 'samples': 5202432, 'steps': 10160, 'loss/train': 2.155850410461426} 02/24/2022 14:43:49 - INFO - codeparrot_training - Step 10161: {'lr': 0.00046517717711941066, 'samples': 5202944, 'steps': 10161, 'loss/train': 1.7206408977508545} 02/24/2022 14:43:53 - INFO - codeparrot_training - Step 10162: {'lr': 0.0004651688465660847, 'samples': 5203456, 'steps': 10162, 'loss/train': 2.2713823318481445} 02/24/2022 14:43:58 - INFO - codeparrot_training - Step 10163: {'lr': 0.00046516051509104376, 'samples': 5203968, 'steps': 10163, 'loss/train': 0.8373558521270752} 02/24/2022 14:44:02 - INFO - codeparrot_training - Step 10164: {'lr': 0.0004651521826943235, 'samples': 5204480, 'steps': 10164, 'loss/train': 2.706667423248291} 02/24/2022 14:44:07 - INFO - codeparrot_training - Step 10165: {'lr': 0.00046514384937595965, 'samples': 5204992, 'steps': 10165, 'loss/train': 2.2613797187805176} 02/24/2022 14:44:11 - INFO - codeparrot_training - Step 10166: {'lr': 0.0004651355151359878, 'samples': 5205504, 'steps': 10166, 'loss/train': 3.5909008979797363} 02/24/2022 14:44:16 - INFO - codeparrot_training - Step 10167: {'lr': 0.0004651271799744437, 'samples': 5206016, 'steps': 10167, 'loss/train': 2.2181339263916016} 02/24/2022 14:44:20 - INFO - codeparrot_training - Step 10168: {'lr': 0.0004651188438913631, 'samples': 5206528, 'steps': 10168, 'loss/train': 2.4214959144592285} 02/24/2022 14:44:25 - INFO - codeparrot_training - Step 10169: {'lr': 0.0004651105068867817, 'samples': 5207040, 'steps': 10169, 'loss/train': 2.7240946292877197} 02/24/2022 14:44:29 - INFO - codeparrot_training - Step 10170: {'lr': 0.00046510216896073517, 'samples': 5207552, 'steps': 10170, 'loss/train': 1.931358814239502} 02/24/2022 14:44:36 - INFO - codeparrot_training - Step 10171: {'lr': 0.00046509383011325925, 'samples': 5208064, 'steps': 10171, 'loss/train': 2.3203563690185547} 02/24/2022 14:44:40 - INFO - codeparrot_training - Step 10172: {'lr': 0.0004650854903443896, 'samples': 5208576, 'steps': 10172, 'loss/train': 2.6517884731292725} 02/24/2022 14:44:45 - INFO - codeparrot_training - Step 10173: {'lr': 0.0004650771496541621, 'samples': 5209088, 'steps': 10173, 'loss/train': 2.381789445877075} 02/24/2022 14:44:49 - INFO - codeparrot_training - Step 10174: {'lr': 0.0004650688080426123, 'samples': 5209600, 'steps': 10174, 'loss/train': 9.420103073120117} 02/24/2022 14:44:54 - INFO - codeparrot_training - Step 10175: {'lr': 0.0004650604655097761, 'samples': 5210112, 'steps': 10175, 'loss/train': 1.1089231967926025} 02/24/2022 14:45:00 - INFO - codeparrot_training - Step 10176: {'lr': 0.00046505212205568916, 'samples': 5210624, 'steps': 10176, 'loss/train': 1.0034704208374023} 02/24/2022 14:45:03 - INFO - codeparrot_training - Step 10177: {'lr': 0.0004650437776803872, 'samples': 5211136, 'steps': 10177, 'loss/train': 2.005133867263794} 02/24/2022 14:45:09 - INFO - codeparrot_training - Step 10178: {'lr': 0.00046503543238390595, 'samples': 5211648, 'steps': 10178, 'loss/train': 2.20976185798645} 02/24/2022 14:45:12 - INFO - codeparrot_training - Step 10179: {'lr': 0.0004650270861662812, 'samples': 5212160, 'steps': 10179, 'loss/train': 2.0330116748809814} 02/24/2022 14:45:20 - INFO - codeparrot_training - Step 10180: {'lr': 0.00046501873902754867, 'samples': 5212672, 'steps': 10180, 'loss/train': 1.8069636821746826} 02/24/2022 14:45:23 - INFO - codeparrot_training - Step 10181: {'lr': 0.00046501039096774415, 'samples': 5213184, 'steps': 10181, 'loss/train': 2.5174460411071777} 02/24/2022 14:45:29 - INFO - codeparrot_training - Step 10182: {'lr': 0.00046500204198690343, 'samples': 5213696, 'steps': 10182, 'loss/train': 2.5565450191497803} 02/24/2022 14:45:32 - INFO - codeparrot_training - Step 10183: {'lr': 0.0004649936920850622, 'samples': 5214208, 'steps': 10183, 'loss/train': 3.108539581298828} 02/24/2022 14:45:38 - INFO - codeparrot_training - Step 10184: {'lr': 0.00046498534126225625, 'samples': 5214720, 'steps': 10184, 'loss/train': 1.167891263961792} 02/24/2022 14:45:41 - INFO - codeparrot_training - Step 10185: {'lr': 0.0004649769895185214, 'samples': 5215232, 'steps': 10185, 'loss/train': 2.0358080863952637} 02/24/2022 14:45:46 - INFO - codeparrot_training - Step 10186: {'lr': 0.00046496863685389336, 'samples': 5215744, 'steps': 10186, 'loss/train': 3.45823335647583} 02/24/2022 14:45:50 - INFO - codeparrot_training - Step 10187: {'lr': 0.00046496028326840796, 'samples': 5216256, 'steps': 10187, 'loss/train': 2.4293813705444336} 02/24/2022 14:45:55 - INFO - codeparrot_training - Step 10188: {'lr': 0.000464951928762101, 'samples': 5216768, 'steps': 10188, 'loss/train': 1.1065857410430908} 02/24/2022 14:45:59 - INFO - codeparrot_training - Step 10189: {'lr': 0.00046494357333500816, 'samples': 5217280, 'steps': 10189, 'loss/train': 1.3676937818527222} 02/24/2022 14:46:07 - INFO - codeparrot_training - Step 10190: {'lr': 0.00046493521698716536, 'samples': 5217792, 'steps': 10190, 'loss/train': 1.7447195053100586} 02/24/2022 14:46:10 - INFO - codeparrot_training - Step 10191: {'lr': 0.00046492685971860826, 'samples': 5218304, 'steps': 10191, 'loss/train': 2.390455722808838} 02/24/2022 14:46:16 - INFO - codeparrot_training - Step 10192: {'lr': 0.00046491850152937276, 'samples': 5218816, 'steps': 10192, 'loss/train': 2.055619478225708} 02/24/2022 14:46:19 - INFO - codeparrot_training - Step 10193: {'lr': 0.0004649101424194947, 'samples': 5219328, 'steps': 10193, 'loss/train': 2.118504524230957} 02/24/2022 14:46:26 - INFO - codeparrot_training - Step 10194: {'lr': 0.0004649017823890098, 'samples': 5219840, 'steps': 10194, 'loss/train': 2.7681474685668945} 02/24/2022 14:46:29 - INFO - codeparrot_training - Step 10195: {'lr': 0.0004648934214379539, 'samples': 5220352, 'steps': 10195, 'loss/train': 2.295779228210449} 02/24/2022 14:46:33 - INFO - codeparrot_training - Step 10196: {'lr': 0.00046488505956636286, 'samples': 5220864, 'steps': 10196, 'loss/train': 2.5109503269195557} 02/24/2022 14:46:38 - INFO - codeparrot_training - Step 10197: {'lr': 0.00046487669677427237, 'samples': 5221376, 'steps': 10197, 'loss/train': 0.6117112040519714} 02/24/2022 14:46:42 - INFO - codeparrot_training - Step 10198: {'lr': 0.0004648683330617184, 'samples': 5221888, 'steps': 10198, 'loss/train': 2.5975136756896973} 02/24/2022 14:46:47 - INFO - codeparrot_training - Step 10199: {'lr': 0.00046485996842873676, 'samples': 5222400, 'steps': 10199, 'loss/train': 1.9540332555770874} 02/24/2022 14:46:51 - INFO - codeparrot_training - Step 10200: {'lr': 0.0004648516028753632, 'samples': 5222912, 'steps': 10200, 'loss/train': 2.2206778526306152} 02/24/2022 14:46:57 - INFO - codeparrot_training - Step 10201: {'lr': 0.00046484323640163356, 'samples': 5223424, 'steps': 10201, 'loss/train': 2.2316691875457764} 02/24/2022 14:47:00 - INFO - codeparrot_training - Step 10202: {'lr': 0.00046483486900758374, 'samples': 5223936, 'steps': 10202, 'loss/train': 1.1625144481658936} 02/24/2022 14:47:06 - INFO - codeparrot_training - Step 10203: {'lr': 0.0004648265006932496, 'samples': 5224448, 'steps': 10203, 'loss/train': 2.249019145965576} 02/24/2022 14:47:09 - INFO - codeparrot_training - Step 10204: {'lr': 0.0004648181314586669, 'samples': 5224960, 'steps': 10204, 'loss/train': 2.6625969409942627} 02/24/2022 14:47:17 - INFO - codeparrot_training - Step 10205: {'lr': 0.00046480976130387156, 'samples': 5225472, 'steps': 10205, 'loss/train': 2.237865686416626} 02/24/2022 14:47:20 - INFO - codeparrot_training - Step 10206: {'lr': 0.0004648013902288994, 'samples': 5225984, 'steps': 10206, 'loss/train': 3.1730339527130127} 02/24/2022 14:47:26 - INFO - codeparrot_training - Step 10207: {'lr': 0.0004647930182337863, 'samples': 5226496, 'steps': 10207, 'loss/train': 2.2114531993865967} 02/24/2022 14:47:29 - INFO - codeparrot_training - Step 10208: {'lr': 0.0004647846453185681, 'samples': 5227008, 'steps': 10208, 'loss/train': 3.4303669929504395} 02/24/2022 14:47:35 - INFO - codeparrot_training - Step 10209: {'lr': 0.0004647762714832807, 'samples': 5227520, 'steps': 10209, 'loss/train': 2.6461150646209717} 02/24/2022 14:47:38 - INFO - codeparrot_training - Step 10210: {'lr': 0.00046476789672795994, 'samples': 5228032, 'steps': 10210, 'loss/train': 1.9035426378250122} 02/24/2022 14:47:44 - INFO - codeparrot_training - Step 10211: {'lr': 0.00046475952105264176, 'samples': 5228544, 'steps': 10211, 'loss/train': 1.794641137123108} 02/24/2022 14:47:47 - INFO - codeparrot_training - Step 10212: {'lr': 0.0004647511444573619, 'samples': 5229056, 'steps': 10212, 'loss/train': 3.3274474143981934} 02/24/2022 14:47:53 - INFO - codeparrot_training - Step 10213: {'lr': 0.00046474276694215635, 'samples': 5229568, 'steps': 10213, 'loss/train': 1.993664026260376} 02/24/2022 14:47:56 - INFO - codeparrot_training - Step 10214: {'lr': 0.000464734388507061, 'samples': 5230080, 'steps': 10214, 'loss/train': 2.331465005874634} 02/24/2022 14:48:03 - INFO - codeparrot_training - Step 10215: {'lr': 0.00046472600915211174, 'samples': 5230592, 'steps': 10215, 'loss/train': 2.0613880157470703} 02/24/2022 14:48:06 - INFO - codeparrot_training - Step 10216: {'lr': 0.00046471762887734437, 'samples': 5231104, 'steps': 10216, 'loss/train': 3.046858072280884} 02/24/2022 14:48:12 - INFO - codeparrot_training - Step 10217: {'lr': 0.0004647092476827949, 'samples': 5231616, 'steps': 10217, 'loss/train': 1.8909368515014648} 02/24/2022 14:48:15 - INFO - codeparrot_training - Step 10218: {'lr': 0.0004647008655684992, 'samples': 5232128, 'steps': 10218, 'loss/train': 2.1916615962982178} 02/24/2022 14:48:21 - INFO - codeparrot_training - Step 10219: {'lr': 0.00046469248253449316, 'samples': 5232640, 'steps': 10219, 'loss/train': 2.1193439960479736} 02/24/2022 14:48:24 - INFO - codeparrot_training - Step 10220: {'lr': 0.0004646840985808126, 'samples': 5233152, 'steps': 10220, 'loss/train': 1.7376508712768555} 02/24/2022 14:48:30 - INFO - codeparrot_training - Step 10221: {'lr': 0.00046467571370749366, 'samples': 5233664, 'steps': 10221, 'loss/train': 1.3292498588562012} 02/24/2022 14:48:33 - INFO - codeparrot_training - Step 10222: {'lr': 0.0004646673279145721, 'samples': 5234176, 'steps': 10222, 'loss/train': 2.7708213329315186} 02/24/2022 14:48:39 - INFO - codeparrot_training - Step 10223: {'lr': 0.00046465894120208384, 'samples': 5234688, 'steps': 10223, 'loss/train': 1.8738151788711548} 02/24/2022 14:48:42 - INFO - codeparrot_training - Step 10224: {'lr': 0.00046465055357006494, 'samples': 5235200, 'steps': 10224, 'loss/train': 1.7299456596374512} 02/24/2022 14:48:48 - INFO - codeparrot_training - Step 10225: {'lr': 0.00046464216501855104, 'samples': 5235712, 'steps': 10225, 'loss/train': 3.709348440170288} 02/24/2022 14:48:52 - INFO - codeparrot_training - Step 10226: {'lr': 0.0004646337755475784, 'samples': 5236224, 'steps': 10226, 'loss/train': 1.841853380203247} 02/24/2022 14:48:58 - INFO - codeparrot_training - Step 10227: {'lr': 0.00046462538515718276, 'samples': 5236736, 'steps': 10227, 'loss/train': 2.1630375385284424} 02/24/2022 14:49:01 - INFO - codeparrot_training - Step 10228: {'lr': 0.0004646169938474002, 'samples': 5237248, 'steps': 10228, 'loss/train': 6.674635887145996} 02/24/2022 14:49:07 - INFO - codeparrot_training - Step 10229: {'lr': 0.0004646086016182666, 'samples': 5237760, 'steps': 10229, 'loss/train': 2.4645204544067383} 02/24/2022 14:49:10 - INFO - codeparrot_training - Step 10230: {'lr': 0.00046460020846981776, 'samples': 5238272, 'steps': 10230, 'loss/train': 2.987922191619873} 02/24/2022 14:49:16 - INFO - codeparrot_training - Step 10231: {'lr': 0.00046459181440208986, 'samples': 5238784, 'steps': 10231, 'loss/train': 5.388568878173828} 02/24/2022 14:49:19 - INFO - codeparrot_training - Step 10232: {'lr': 0.0004645834194151187, 'samples': 5239296, 'steps': 10232, 'loss/train': 1.9100579023361206} 02/24/2022 14:49:25 - INFO - codeparrot_training - Step 10233: {'lr': 0.00046457502350894046, 'samples': 5239808, 'steps': 10233, 'loss/train': 2.3590855598449707} 02/24/2022 14:49:28 - INFO - codeparrot_training - Step 10234: {'lr': 0.0004645666266835908, 'samples': 5240320, 'steps': 10234, 'loss/train': 1.1606158018112183} 02/24/2022 14:49:34 - INFO - codeparrot_training - Step 10235: {'lr': 0.0004645582289391059, 'samples': 5240832, 'steps': 10235, 'loss/train': 2.213454484939575} 02/24/2022 14:49:37 - INFO - codeparrot_training - Step 10236: {'lr': 0.00046454983027552165, 'samples': 5241344, 'steps': 10236, 'loss/train': 0.4401305317878723} 02/24/2022 14:49:44 - INFO - codeparrot_training - Step 10237: {'lr': 0.0004645414306928741, 'samples': 5241856, 'steps': 10237, 'loss/train': 1.0115946531295776} 02/24/2022 14:49:47 - INFO - codeparrot_training - Step 10238: {'lr': 0.0004645330301911992, 'samples': 5242368, 'steps': 10238, 'loss/train': 1.8077300786972046} 02/24/2022 14:49:53 - INFO - codeparrot_training - Step 10239: {'lr': 0.0004645246287705329, 'samples': 5242880, 'steps': 10239, 'loss/train': 1.8756779432296753} 02/24/2022 14:49:56 - INFO - codeparrot_training - Step 10240: {'lr': 0.0004645162264309112, 'samples': 5243392, 'steps': 10240, 'loss/train': 1.9027059078216553} 02/24/2022 14:50:02 - INFO - codeparrot_training - Step 10241: {'lr': 0.0004645078231723701, 'samples': 5243904, 'steps': 10241, 'loss/train': 1.5016604661941528} 02/24/2022 14:50:05 - INFO - codeparrot_training - Step 10242: {'lr': 0.0004644994189949455, 'samples': 5244416, 'steps': 10242, 'loss/train': 1.804394006729126} 02/24/2022 14:50:11 - INFO - codeparrot_training - Step 10243: {'lr': 0.00046449101389867364, 'samples': 5244928, 'steps': 10243, 'loss/train': 2.33488392829895} 02/24/2022 14:50:14 - INFO - codeparrot_training - Step 10244: {'lr': 0.0004644826078835903, 'samples': 5245440, 'steps': 10244, 'loss/train': 1.6047513484954834} 02/24/2022 14:50:20 - INFO - codeparrot_training - Step 10245: {'lr': 0.00046447420094973167, 'samples': 5245952, 'steps': 10245, 'loss/train': 1.8958096504211426} 02/24/2022 14:50:23 - INFO - codeparrot_training - Step 10246: {'lr': 0.0004644657930971336, 'samples': 5246464, 'steps': 10246, 'loss/train': 2.2004952430725098} 02/24/2022 14:50:29 - INFO - codeparrot_training - Step 10247: {'lr': 0.00046445738432583216, 'samples': 5246976, 'steps': 10247, 'loss/train': 1.1674318313598633} 02/24/2022 14:50:32 - INFO - codeparrot_training - Step 10248: {'lr': 0.00046444897463586345, 'samples': 5247488, 'steps': 10248, 'loss/train': 2.0538482666015625} 02/24/2022 14:50:38 - INFO - codeparrot_training - Step 10249: {'lr': 0.00046444056402726336, 'samples': 5248000, 'steps': 10249, 'loss/train': 1.1609246730804443} 02/24/2022 14:50:41 - INFO - codeparrot_training - Step 10250: {'lr': 0.00046443215250006805, 'samples': 5248512, 'steps': 10250, 'loss/train': 2.9854023456573486} 02/24/2022 14:50:48 - INFO - codeparrot_training - Step 10251: {'lr': 0.00046442374005431345, 'samples': 5249024, 'steps': 10251, 'loss/train': 2.560157060623169} 02/24/2022 14:50:51 - INFO - codeparrot_training - Step 10252: {'lr': 0.0004644153266900356, 'samples': 5249536, 'steps': 10252, 'loss/train': 2.1460869312286377} 02/24/2022 14:50:55 - INFO - codeparrot_training - Step 10253: {'lr': 0.0004644069124072706, 'samples': 5250048, 'steps': 10253, 'loss/train': 2.3790555000305176} 02/24/2022 14:51:00 - INFO - codeparrot_training - Step 10254: {'lr': 0.0004643984972060545, 'samples': 5250560, 'steps': 10254, 'loss/train': 1.6774470806121826} 02/24/2022 14:51:04 - INFO - codeparrot_training - Step 10255: {'lr': 0.00046439008108642335, 'samples': 5251072, 'steps': 10255, 'loss/train': 1.579734444618225} 02/24/2022 14:51:09 - INFO - codeparrot_training - Step 10256: {'lr': 0.0004643816640484131, 'samples': 5251584, 'steps': 10256, 'loss/train': 2.8604841232299805} 02/24/2022 14:51:13 - INFO - codeparrot_training - Step 10257: {'lr': 0.0004643732460920599, 'samples': 5252096, 'steps': 10257, 'loss/train': 1.7653487920761108} 02/24/2022 14:51:18 - INFO - codeparrot_training - Step 10258: {'lr': 0.00046436482721739976, 'samples': 5252608, 'steps': 10258, 'loss/train': 2.517033815383911} 02/24/2022 14:51:24 - INFO - codeparrot_training - Step 10259: {'lr': 0.00046435640742446875, 'samples': 5253120, 'steps': 10259, 'loss/train': 3.2435221672058105} 02/24/2022 14:51:27 - INFO - codeparrot_training - Step 10260: {'lr': 0.000464347986713303, 'samples': 5253632, 'steps': 10260, 'loss/train': 1.3066688776016235} 02/24/2022 14:51:32 - INFO - codeparrot_training - Step 10261: {'lr': 0.00046433956508393855, 'samples': 5254144, 'steps': 10261, 'loss/train': 2.3456404209136963} 02/24/2022 14:51:36 - INFO - codeparrot_training - Step 10262: {'lr': 0.0004643311425364114, 'samples': 5254656, 'steps': 10262, 'loss/train': 2.8051798343658447} 02/24/2022 14:51:42 - INFO - codeparrot_training - Step 10263: {'lr': 0.0004643227190707577, 'samples': 5255168, 'steps': 10263, 'loss/train': 3.029179334640503} 02/24/2022 14:51:46 - INFO - codeparrot_training - Step 10264: {'lr': 0.00046431429468701363, 'samples': 5255680, 'steps': 10264, 'loss/train': 2.406796932220459} 02/24/2022 14:51:51 - INFO - codeparrot_training - Step 10265: {'lr': 0.0004643058693852151, 'samples': 5256192, 'steps': 10265, 'loss/train': 0.2629365921020508} 02/24/2022 14:51:55 - INFO - codeparrot_training - Step 10266: {'lr': 0.0004642974431653983, 'samples': 5256704, 'steps': 10266, 'loss/train': 1.8382028341293335} 02/24/2022 14:52:00 - INFO - codeparrot_training - Step 10267: {'lr': 0.00046428901602759933, 'samples': 5257216, 'steps': 10267, 'loss/train': 2.460885763168335} 02/24/2022 14:52:04 - INFO - codeparrot_training - Step 10268: {'lr': 0.00046428058797185417, 'samples': 5257728, 'steps': 10268, 'loss/train': 2.5617311000823975} 02/24/2022 14:52:09 - INFO - codeparrot_training - Step 10269: {'lr': 0.0004642721589981991, 'samples': 5258240, 'steps': 10269, 'loss/train': 4.107093334197998} 02/24/2022 14:52:13 - INFO - codeparrot_training - Step 10270: {'lr': 0.00046426372910667003, 'samples': 5258752, 'steps': 10270, 'loss/train': 0.7126080393791199} 02/24/2022 14:52:18 - INFO - codeparrot_training - Step 10271: {'lr': 0.00046425529829730326, 'samples': 5259264, 'steps': 10271, 'loss/train': 2.419494152069092} 02/24/2022 14:52:22 - INFO - codeparrot_training - Step 10272: {'lr': 0.0004642468665701348, 'samples': 5259776, 'steps': 10272, 'loss/train': 2.052345037460327} 02/24/2022 14:52:28 - INFO - codeparrot_training - Step 10273: {'lr': 0.0004642384339252008, 'samples': 5260288, 'steps': 10273, 'loss/train': 4.275566577911377} 02/24/2022 14:52:32 - INFO - codeparrot_training - Step 10274: {'lr': 0.0004642300003625374, 'samples': 5260800, 'steps': 10274, 'loss/train': 2.1266286373138428} 02/24/2022 14:52:38 - INFO - codeparrot_training - Step 10275: {'lr': 0.0004642215658821807, 'samples': 5261312, 'steps': 10275, 'loss/train': 2.6137502193450928} 02/24/2022 14:52:41 - INFO - codeparrot_training - Step 10276: {'lr': 0.0004642131304841668, 'samples': 5261824, 'steps': 10276, 'loss/train': 3.646739959716797} 02/24/2022 14:52:45 - INFO - codeparrot_training - Step 10277: {'lr': 0.00046420469416853197, 'samples': 5262336, 'steps': 10277, 'loss/train': 1.6540005207061768} 02/24/2022 14:52:50 - INFO - codeparrot_training - Step 10278: {'lr': 0.0004641962569353121, 'samples': 5262848, 'steps': 10278, 'loss/train': 1.4199060201644897} 02/24/2022 14:52:54 - INFO - codeparrot_training - Step 10279: {'lr': 0.0004641878187845436, 'samples': 5263360, 'steps': 10279, 'loss/train': 2.1314585208892822} 02/24/2022 14:52:59 - INFO - codeparrot_training - Step 10280: {'lr': 0.00046417937971626245, 'samples': 5263872, 'steps': 10280, 'loss/train': 1.3049476146697998} 02/24/2022 14:53:03 - INFO - codeparrot_training - Step 10281: {'lr': 0.00046417093973050486, 'samples': 5264384, 'steps': 10281, 'loss/train': 3.138810157775879} 02/24/2022 14:53:08 - INFO - codeparrot_training - Step 10282: {'lr': 0.0004641624988273069, 'samples': 5264896, 'steps': 10282, 'loss/train': 2.1349916458129883} 02/24/2022 14:53:12 - INFO - codeparrot_training - Step 10283: {'lr': 0.0004641540570067049, 'samples': 5265408, 'steps': 10283, 'loss/train': 1.3850854635238647} 02/24/2022 14:53:18 - INFO - codeparrot_training - Step 10284: {'lr': 0.0004641456142687348, 'samples': 5265920, 'steps': 10284, 'loss/train': 2.4753170013427734} 02/24/2022 14:53:22 - INFO - codeparrot_training - Step 10285: {'lr': 0.000464137170613433, 'samples': 5266432, 'steps': 10285, 'loss/train': 1.1113992929458618} 02/24/2022 14:53:27 - INFO - codeparrot_training - Step 10286: {'lr': 0.00046412872604083554, 'samples': 5266944, 'steps': 10286, 'loss/train': 3.43532395362854} 02/24/2022 14:53:31 - INFO - codeparrot_training - Step 10287: {'lr': 0.00046412028055097855, 'samples': 5267456, 'steps': 10287, 'loss/train': 2.749605417251587} 02/24/2022 14:53:36 - INFO - codeparrot_training - Step 10288: {'lr': 0.00046411183414389834, 'samples': 5267968, 'steps': 10288, 'loss/train': 1.774010419845581} 02/24/2022 14:53:40 - INFO - codeparrot_training - Step 10289: {'lr': 0.000464103386819631, 'samples': 5268480, 'steps': 10289, 'loss/train': 2.439481019973755} 02/24/2022 14:53:45 - INFO - codeparrot_training - Step 10290: {'lr': 0.00046409493857821273, 'samples': 5268992, 'steps': 10290, 'loss/train': 1.5167802572250366} 02/24/2022 14:53:49 - INFO - codeparrot_training - Step 10291: {'lr': 0.00046408648941967975, 'samples': 5269504, 'steps': 10291, 'loss/train': 0.35078924894332886} 02/24/2022 14:53:54 - INFO - codeparrot_training - Step 10292: {'lr': 0.0004640780393440682, 'samples': 5270016, 'steps': 10292, 'loss/train': 1.8131710290908813} 02/24/2022 14:53:58 - INFO - codeparrot_training - Step 10293: {'lr': 0.0004640695883514143, 'samples': 5270528, 'steps': 10293, 'loss/train': 2.1447877883911133} 02/24/2022 14:54:04 - INFO - codeparrot_training - Step 10294: {'lr': 0.0004640611364417543, 'samples': 5271040, 'steps': 10294, 'loss/train': 2.1572015285491943} 02/24/2022 14:54:07 - INFO - codeparrot_training - Step 10295: {'lr': 0.0004640526836151243, 'samples': 5271552, 'steps': 10295, 'loss/train': 2.3479580879211426} 02/24/2022 14:54:13 - INFO - codeparrot_training - Step 10296: {'lr': 0.0004640442298715606, 'samples': 5272064, 'steps': 10296, 'loss/train': 2.0927531719207764} 02/24/2022 14:54:17 - INFO - codeparrot_training - Step 10297: {'lr': 0.0004640357752110994, 'samples': 5272576, 'steps': 10297, 'loss/train': 2.046898365020752} 02/24/2022 14:54:22 - INFO - codeparrot_training - Step 10298: {'lr': 0.00046402731963377685, 'samples': 5273088, 'steps': 10298, 'loss/train': 1.562732458114624} 02/24/2022 14:54:26 - INFO - codeparrot_training - Step 10299: {'lr': 0.0004640188631396293, 'samples': 5273600, 'steps': 10299, 'loss/train': 2.679598569869995} 02/24/2022 14:54:31 - INFO - codeparrot_training - Step 10300: {'lr': 0.0004640104057286929, 'samples': 5274112, 'steps': 10300, 'loss/train': 1.7256455421447754} 02/24/2022 14:54:35 - INFO - codeparrot_training - Step 10301: {'lr': 0.0004640019474010038, 'samples': 5274624, 'steps': 10301, 'loss/train': 1.2026939392089844} 02/24/2022 14:54:40 - INFO - codeparrot_training - Step 10302: {'lr': 0.00046399348815659837, 'samples': 5275136, 'steps': 10302, 'loss/train': 2.291311264038086} 02/24/2022 14:54:44 - INFO - codeparrot_training - Step 10303: {'lr': 0.0004639850279955128, 'samples': 5275648, 'steps': 10303, 'loss/train': 1.6144028902053833} 02/24/2022 14:54:49 - INFO - codeparrot_training - Step 10304: {'lr': 0.0004639765669177833, 'samples': 5276160, 'steps': 10304, 'loss/train': 2.092270851135254} 02/24/2022 14:54:53 - INFO - codeparrot_training - Step 10305: {'lr': 0.0004639681049234461, 'samples': 5276672, 'steps': 10305, 'loss/train': 1.8585162162780762} 02/24/2022 14:54:58 - INFO - codeparrot_training - Step 10306: {'lr': 0.0004639596420125375, 'samples': 5277184, 'steps': 10306, 'loss/train': 2.8110735416412354} 02/24/2022 14:55:02 - INFO - codeparrot_training - Step 10307: {'lr': 0.0004639511781850937, 'samples': 5277696, 'steps': 10307, 'loss/train': 2.093008518218994} 02/24/2022 14:55:07 - INFO - codeparrot_training - Step 10308: {'lr': 0.000463942713441151, 'samples': 5278208, 'steps': 10308, 'loss/train': 1.4693944454193115} 02/24/2022 14:55:11 - INFO - codeparrot_training - Step 10309: {'lr': 0.00046393424778074573, 'samples': 5278720, 'steps': 10309, 'loss/train': 2.2482521533966064} 02/24/2022 14:55:17 - INFO - codeparrot_training - Step 10310: {'lr': 0.000463925781203914, 'samples': 5279232, 'steps': 10310, 'loss/train': 1.3743386268615723} 02/24/2022 14:55:20 - INFO - codeparrot_training - Step 10311: {'lr': 0.00046391731371069224, 'samples': 5279744, 'steps': 10311, 'loss/train': 2.4796366691589355} 02/24/2022 14:55:26 - INFO - codeparrot_training - Step 10312: {'lr': 0.00046390884530111656, 'samples': 5280256, 'steps': 10312, 'loss/train': 2.6968576908111572} 02/24/2022 14:55:29 - INFO - codeparrot_training - Step 10313: {'lr': 0.0004639003759752233, 'samples': 5280768, 'steps': 10313, 'loss/train': 1.3039103746414185} 02/24/2022 14:55:35 - INFO - codeparrot_training - Step 10314: {'lr': 0.00046389190573304875, 'samples': 5281280, 'steps': 10314, 'loss/train': 2.093705654144287} 02/24/2022 14:55:38 - INFO - codeparrot_training - Step 10315: {'lr': 0.0004638834345746292, 'samples': 5281792, 'steps': 10315, 'loss/train': 0.8147538900375366} 02/24/2022 14:55:44 - INFO - codeparrot_training - Step 10316: {'lr': 0.00046387496250000095, 'samples': 5282304, 'steps': 10316, 'loss/train': 2.4193246364593506} 02/24/2022 14:55:49 - INFO - codeparrot_training - Step 10317: {'lr': 0.00046386648950920027, 'samples': 5282816, 'steps': 10317, 'loss/train': 1.7856088876724243} 02/24/2022 14:55:53 - INFO - codeparrot_training - Step 10318: {'lr': 0.0004638580156022635, 'samples': 5283328, 'steps': 10318, 'loss/train': 3.212846279144287} 02/24/2022 14:55:59 - INFO - codeparrot_training - Step 10319: {'lr': 0.0004638495407792268, 'samples': 5283840, 'steps': 10319, 'loss/train': 1.845513939857483} 02/24/2022 14:56:02 - INFO - codeparrot_training - Step 10320: {'lr': 0.0004638410650401267, 'samples': 5284352, 'steps': 10320, 'loss/train': 1.9201545715332031} 02/24/2022 14:56:08 - INFO - codeparrot_training - Step 10321: {'lr': 0.0004638325883849993, 'samples': 5284864, 'steps': 10321, 'loss/train': 3.156074047088623} 02/24/2022 14:56:11 - INFO - codeparrot_training - Step 10322: {'lr': 0.00046382411081388096, 'samples': 5285376, 'steps': 10322, 'loss/train': 0.6073839068412781} 02/24/2022 14:56:17 - INFO - codeparrot_training - Step 10323: {'lr': 0.0004638156323268081, 'samples': 5285888, 'steps': 10323, 'loss/train': 2.121464252471924} 02/24/2022 14:56:20 - INFO - codeparrot_training - Step 10324: {'lr': 0.00046380715292381695, 'samples': 5286400, 'steps': 10324, 'loss/train': 1.2324066162109375} 02/24/2022 14:56:26 - INFO - codeparrot_training - Step 10325: {'lr': 0.0004637986726049438, 'samples': 5286912, 'steps': 10325, 'loss/train': 3.2524149417877197} 02/24/2022 14:56:29 - INFO - codeparrot_training - Step 10326: {'lr': 0.00046379019137022506, 'samples': 5287424, 'steps': 10326, 'loss/train': 2.1994616985321045} 02/24/2022 14:56:35 - INFO - codeparrot_training - Step 10327: {'lr': 0.000463781709219697, 'samples': 5287936, 'steps': 10327, 'loss/train': 2.3758130073547363} 02/24/2022 14:56:38 - INFO - codeparrot_training - Step 10328: {'lr': 0.000463773226153396, 'samples': 5288448, 'steps': 10328, 'loss/train': 0.9522534012794495} 02/24/2022 14:56:44 - INFO - codeparrot_training - Step 10329: {'lr': 0.0004637647421713584, 'samples': 5288960, 'steps': 10329, 'loss/train': 2.19032883644104} 02/24/2022 14:56:48 - INFO - codeparrot_training - Step 10330: {'lr': 0.0004637562572736205, 'samples': 5289472, 'steps': 10330, 'loss/train': 1.8007209300994873} 02/24/2022 14:56:53 - INFO - codeparrot_training - Step 10331: {'lr': 0.00046374777146021865, 'samples': 5289984, 'steps': 10331, 'loss/train': 2.995356798171997} 02/24/2022 14:56:57 - INFO - codeparrot_training - Step 10332: {'lr': 0.00046373928473118927, 'samples': 5290496, 'steps': 10332, 'loss/train': 1.9197447299957275} 02/24/2022 14:57:02 - INFO - codeparrot_training - Step 10333: {'lr': 0.0004637307970865686, 'samples': 5291008, 'steps': 10333, 'loss/train': 3.180255174636841} 02/24/2022 14:57:06 - INFO - codeparrot_training - Step 10334: {'lr': 0.00046372230852639314, 'samples': 5291520, 'steps': 10334, 'loss/train': 2.269125461578369} 02/24/2022 14:57:11 - INFO - codeparrot_training - Step 10335: {'lr': 0.0004637138190506991, 'samples': 5292032, 'steps': 10335, 'loss/train': 2.0725109577178955} 02/24/2022 14:57:15 - INFO - codeparrot_training - Step 10336: {'lr': 0.00046370532865952296, 'samples': 5292544, 'steps': 10336, 'loss/train': 2.2971909046173096} 02/24/2022 14:57:20 - INFO - codeparrot_training - Step 10337: {'lr': 0.0004636968373529011, 'samples': 5293056, 'steps': 10337, 'loss/train': 2.3532931804656982} 02/24/2022 14:57:24 - INFO - codeparrot_training - Step 10338: {'lr': 0.00046368834513086976, 'samples': 5293568, 'steps': 10338, 'loss/train': 2.426110029220581} 02/24/2022 14:57:30 - INFO - codeparrot_training - Step 10339: {'lr': 0.00046367985199346546, 'samples': 5294080, 'steps': 10339, 'loss/train': 1.3799611330032349} 02/24/2022 14:57:33 - INFO - codeparrot_training - Step 10340: {'lr': 0.00046367135794072445, 'samples': 5294592, 'steps': 10340, 'loss/train': 1.8972915410995483} 02/24/2022 14:57:39 - INFO - codeparrot_training - Step 10341: {'lr': 0.00046366286297268327, 'samples': 5295104, 'steps': 10341, 'loss/train': 2.365474224090576} 02/24/2022 14:57:42 - INFO - codeparrot_training - Step 10342: {'lr': 0.0004636543670893782, 'samples': 5295616, 'steps': 10342, 'loss/train': 1.4184718132019043} 02/24/2022 14:57:48 - INFO - codeparrot_training - Step 10343: {'lr': 0.0004636458702908457, 'samples': 5296128, 'steps': 10343, 'loss/train': 1.5894575119018555} 02/24/2022 14:57:51 - INFO - codeparrot_training - Step 10344: {'lr': 0.0004636373725771221, 'samples': 5296640, 'steps': 10344, 'loss/train': 1.9423401355743408} 02/24/2022 14:57:57 - INFO - codeparrot_training - Step 10345: {'lr': 0.0004636288739482438, 'samples': 5297152, 'steps': 10345, 'loss/train': 1.7704288959503174} 02/24/2022 14:58:00 - INFO - codeparrot_training - Step 10346: {'lr': 0.0004636203744042473, 'samples': 5297664, 'steps': 10346, 'loss/train': 2.4374425411224365} 02/24/2022 14:58:06 - INFO - codeparrot_training - Step 10347: {'lr': 0.0004636118739451689, 'samples': 5298176, 'steps': 10347, 'loss/train': 3.2590057849884033} 02/24/2022 14:58:09 - INFO - codeparrot_training - Step 10348: {'lr': 0.0004636033725710451, 'samples': 5298688, 'steps': 10348, 'loss/train': 2.1918511390686035} 02/24/2022 14:58:15 - INFO - codeparrot_training - Step 10349: {'lr': 0.00046359487028191224, 'samples': 5299200, 'steps': 10349, 'loss/train': 3.533186912536621} 02/24/2022 14:58:18 - INFO - codeparrot_training - Step 10350: {'lr': 0.0004635863670778068, 'samples': 5299712, 'steps': 10350, 'loss/train': 1.1902270317077637} 02/24/2022 14:58:24 - INFO - codeparrot_training - Step 10351: {'lr': 0.00046357786295876517, 'samples': 5300224, 'steps': 10351, 'loss/train': 1.4413707256317139} 02/24/2022 14:58:28 - INFO - codeparrot_training - Step 10352: {'lr': 0.0004635693579248238, 'samples': 5300736, 'steps': 10352, 'loss/train': 2.7832536697387695} 02/24/2022 14:58:33 - INFO - codeparrot_training - Step 10353: {'lr': 0.0004635608519760191, 'samples': 5301248, 'steps': 10353, 'loss/train': 1.747622013092041} 02/24/2022 14:58:37 - INFO - codeparrot_training - Step 10354: {'lr': 0.00046355234511238756, 'samples': 5301760, 'steps': 10354, 'loss/train': 2.0732836723327637} 02/24/2022 14:58:42 - INFO - codeparrot_training - Step 10355: {'lr': 0.00046354383733396553, 'samples': 5302272, 'steps': 10355, 'loss/train': 2.2942445278167725} 02/24/2022 14:58:45 - INFO - codeparrot_training - Step 10356: {'lr': 0.0004635353286407896, 'samples': 5302784, 'steps': 10356, 'loss/train': 2.1034181118011475} 02/24/2022 14:58:52 - INFO - codeparrot_training - Step 10357: {'lr': 0.00046352681903289605, 'samples': 5303296, 'steps': 10357, 'loss/train': 1.9662588834762573} 02/24/2022 14:58:55 - INFO - codeparrot_training - Step 10358: {'lr': 0.00046351830851032146, 'samples': 5303808, 'steps': 10358, 'loss/train': 1.8031994104385376} 02/24/2022 14:59:01 - INFO - codeparrot_training - Step 10359: {'lr': 0.00046350979707310226, 'samples': 5304320, 'steps': 10359, 'loss/train': 2.544552803039551} 02/24/2022 14:59:04 - INFO - codeparrot_training - Step 10360: {'lr': 0.00046350128472127483, 'samples': 5304832, 'steps': 10360, 'loss/train': 1.9541394710540771} 02/24/2022 14:59:10 - INFO - codeparrot_training - Step 10361: {'lr': 0.00046349277145487565, 'samples': 5305344, 'steps': 10361, 'loss/train': 1.484208106994629} 02/24/2022 14:59:13 - INFO - codeparrot_training - Step 10362: {'lr': 0.00046348425727394126, 'samples': 5305856, 'steps': 10362, 'loss/train': 0.7156141400337219} 02/24/2022 14:59:19 - INFO - codeparrot_training - Step 10363: {'lr': 0.0004634757421785082, 'samples': 5306368, 'steps': 10363, 'loss/train': 2.1460771560668945} 02/24/2022 14:59:22 - INFO - codeparrot_training - Step 10364: {'lr': 0.0004634672261686127, 'samples': 5306880, 'steps': 10364, 'loss/train': 3.2442281246185303} 02/24/2022 14:59:27 - INFO - codeparrot_training - Step 10365: {'lr': 0.0004634587092442915, 'samples': 5307392, 'steps': 10365, 'loss/train': 1.8552302122116089} 02/24/2022 14:59:34 - INFO - codeparrot_training - Step 10366: {'lr': 0.00046345019140558085, 'samples': 5307904, 'steps': 10366, 'loss/train': 2.884028434753418} 02/24/2022 14:59:37 - INFO - codeparrot_training - Step 10367: {'lr': 0.0004634416726525175, 'samples': 5308416, 'steps': 10367, 'loss/train': 2.494704484939575} 02/24/2022 14:59:43 - INFO - codeparrot_training - Step 10368: {'lr': 0.00046343315298513765, 'samples': 5308928, 'steps': 10368, 'loss/train': 4.402544021606445} 02/24/2022 14:59:46 - INFO - codeparrot_training - Step 10369: {'lr': 0.0004634246324034781, 'samples': 5309440, 'steps': 10369, 'loss/train': 2.1542961597442627} 02/24/2022 14:59:52 - INFO - codeparrot_training - Step 10370: {'lr': 0.0004634161109075751, 'samples': 5309952, 'steps': 10370, 'loss/train': 2.0879409313201904} 02/24/2022 14:59:55 - INFO - codeparrot_training - Step 10371: {'lr': 0.0004634075884974652, 'samples': 5310464, 'steps': 10371, 'loss/train': 1.6788160800933838} 02/24/2022 14:59:59 - INFO - codeparrot_training - Step 10372: {'lr': 0.00046339906517318507, 'samples': 5310976, 'steps': 10372, 'loss/train': 1.8942443132400513} 02/24/2022 15:00:04 - INFO - codeparrot_training - Step 10373: {'lr': 0.0004633905409347711, 'samples': 5311488, 'steps': 10373, 'loss/train': 2.1707465648651123} 02/24/2022 15:00:11 - INFO - codeparrot_training - Step 10374: {'lr': 0.00046338201578225975, 'samples': 5312000, 'steps': 10374, 'loss/train': 2.4812490940093994} 02/24/2022 15:00:15 - INFO - codeparrot_training - Step 10375: {'lr': 0.0004633734897156876, 'samples': 5312512, 'steps': 10375, 'loss/train': 2.2935268878936768} 02/24/2022 15:00:20 - INFO - codeparrot_training - Step 10376: {'lr': 0.0004633649627350912, 'samples': 5313024, 'steps': 10376, 'loss/train': 1.2265043258666992} 02/24/2022 15:00:24 - INFO - codeparrot_training - Step 10377: {'lr': 0.000463356434840507, 'samples': 5313536, 'steps': 10377, 'loss/train': 2.8925282955169678} 02/24/2022 15:00:29 - INFO - codeparrot_training - Step 10378: {'lr': 0.0004633479060319717, 'samples': 5314048, 'steps': 10378, 'loss/train': 2.223686695098877} 02/24/2022 15:00:33 - INFO - codeparrot_training - Step 10379: {'lr': 0.00046333937630952163, 'samples': 5314560, 'steps': 10379, 'loss/train': 2.1045138835906982} 02/24/2022 15:00:38 - INFO - codeparrot_training - Step 10380: {'lr': 0.00046333084567319344, 'samples': 5315072, 'steps': 10380, 'loss/train': 1.5444692373275757} 02/24/2022 15:00:42 - INFO - codeparrot_training - Step 10381: {'lr': 0.0004633223141230236, 'samples': 5315584, 'steps': 10381, 'loss/train': 2.1285784244537354} 02/24/2022 15:00:47 - INFO - codeparrot_training - Step 10382: {'lr': 0.0004633137816590488, 'samples': 5316096, 'steps': 10382, 'loss/train': 1.839762568473816} 02/24/2022 15:00:51 - INFO - codeparrot_training - Step 10383: {'lr': 0.00046330524828130536, 'samples': 5316608, 'steps': 10383, 'loss/train': 1.6819720268249512} 02/24/2022 15:00:56 - INFO - codeparrot_training - Step 10384: {'lr': 0.00046329671398983007, 'samples': 5317120, 'steps': 10384, 'loss/train': 2.060786485671997} 02/24/2022 15:01:00 - INFO - codeparrot_training - Step 10385: {'lr': 0.0004632881787846594, 'samples': 5317632, 'steps': 10385, 'loss/train': 0.47012028098106384} 02/24/2022 15:01:06 - INFO - codeparrot_training - Step 10386: {'lr': 0.0004632796426658298, 'samples': 5318144, 'steps': 10386, 'loss/train': 2.015021800994873} 02/24/2022 15:01:09 - INFO - codeparrot_training - Step 10387: {'lr': 0.00046327110563337804, 'samples': 5318656, 'steps': 10387, 'loss/train': 2.4878122806549072} 02/24/2022 15:01:15 - INFO - codeparrot_training - Step 10388: {'lr': 0.00046326256768734053, 'samples': 5319168, 'steps': 10388, 'loss/train': 1.9941998720169067} 02/24/2022 15:01:18 - INFO - codeparrot_training - Step 10389: {'lr': 0.0004632540288277539, 'samples': 5319680, 'steps': 10389, 'loss/train': 0.9611613154411316} 02/24/2022 15:01:24 - INFO - codeparrot_training - Step 10390: {'lr': 0.0004632454890546547, 'samples': 5320192, 'steps': 10390, 'loss/train': 2.2257487773895264} 02/24/2022 15:01:28 - INFO - codeparrot_training - Step 10391: {'lr': 0.0004632369483680796, 'samples': 5320704, 'steps': 10391, 'loss/train': 3.8595004081726074} 02/24/2022 15:01:33 - INFO - codeparrot_training - Step 10392: {'lr': 0.0004632284067680651, 'samples': 5321216, 'steps': 10392, 'loss/train': 1.0521191358566284} 02/24/2022 15:01:37 - INFO - codeparrot_training - Step 10393: {'lr': 0.0004632198642546478, 'samples': 5321728, 'steps': 10393, 'loss/train': 0.564673662185669} 02/24/2022 15:01:42 - INFO - codeparrot_training - Step 10394: {'lr': 0.0004632113208278643, 'samples': 5322240, 'steps': 10394, 'loss/train': 2.7666842937469482} 02/24/2022 15:01:46 - INFO - codeparrot_training - Step 10395: {'lr': 0.00046320277648775123, 'samples': 5322752, 'steps': 10395, 'loss/train': 3.366778612136841} 02/24/2022 15:01:52 - INFO - codeparrot_training - Step 10396: {'lr': 0.0004631942312343452, 'samples': 5323264, 'steps': 10396, 'loss/train': 1.51768159866333} 02/24/2022 15:01:55 - INFO - codeparrot_training - Step 10397: {'lr': 0.00046318568506768267, 'samples': 5323776, 'steps': 10397, 'loss/train': 3.0890817642211914} 02/24/2022 15:02:01 - INFO - codeparrot_training - Step 10398: {'lr': 0.0004631771379878005, 'samples': 5324288, 'steps': 10398, 'loss/train': 1.256666660308838} 02/24/2022 15:02:04 - INFO - codeparrot_training - Step 10399: {'lr': 0.00046316858999473506, 'samples': 5324800, 'steps': 10399, 'loss/train': 1.6612883806228638} 02/24/2022 15:02:10 - INFO - codeparrot_training - Step 10400: {'lr': 0.00046316004108852305, 'samples': 5325312, 'steps': 10400, 'loss/train': 2.061941385269165} 02/24/2022 15:02:13 - INFO - codeparrot_training - Step 10401: {'lr': 0.0004631514912692012, 'samples': 5325824, 'steps': 10401, 'loss/train': 3.0965874195098877} 02/24/2022 15:02:20 - INFO - codeparrot_training - Step 10402: {'lr': 0.00046314294053680593, 'samples': 5326336, 'steps': 10402, 'loss/train': 2.3951609134674072} 02/24/2022 15:02:23 - INFO - codeparrot_training - Step 10403: {'lr': 0.0004631343888913741, 'samples': 5326848, 'steps': 10403, 'loss/train': 1.980438232421875} 02/24/2022 15:02:27 - INFO - codeparrot_training - Step 10404: {'lr': 0.00046312583633294213, 'samples': 5327360, 'steps': 10404, 'loss/train': 1.0517843961715698} 02/24/2022 15:02:32 - INFO - codeparrot_training - Step 10405: {'lr': 0.0004631172828615469, 'samples': 5327872, 'steps': 10405, 'loss/train': 2.7647383213043213} 02/24/2022 15:02:36 - INFO - codeparrot_training - Step 10406: {'lr': 0.0004631087284772247, 'samples': 5328384, 'steps': 10406, 'loss/train': 2.3980698585510254} 02/24/2022 15:02:41 - INFO - codeparrot_training - Step 10407: {'lr': 0.0004631001731800125, 'samples': 5328896, 'steps': 10407, 'loss/train': 3.1227638721466064} 02/24/2022 15:02:45 - INFO - codeparrot_training - Step 10408: {'lr': 0.0004630916169699468, 'samples': 5329408, 'steps': 10408, 'loss/train': 1.1033917665481567} 02/24/2022 15:02:50 - INFO - codeparrot_training - Step 10409: {'lr': 0.00046308305984706435, 'samples': 5329920, 'steps': 10409, 'loss/train': 1.6527396440505981} 02/24/2022 15:02:54 - INFO - codeparrot_training - Step 10410: {'lr': 0.00046307450181140163, 'samples': 5330432, 'steps': 10410, 'loss/train': 2.8946332931518555} 02/24/2022 15:02:59 - INFO - codeparrot_training - Step 10411: {'lr': 0.00046306594286299544, 'samples': 5330944, 'steps': 10411, 'loss/train': 1.771353006362915} 02/24/2022 15:03:03 - INFO - codeparrot_training - Step 10412: {'lr': 0.0004630573830018824, 'samples': 5331456, 'steps': 10412, 'loss/train': 2.3282008171081543} 02/24/2022 15:03:09 - INFO - codeparrot_training - Step 10413: {'lr': 0.00046304882222809917, 'samples': 5331968, 'steps': 10413, 'loss/train': 1.580085039138794} 02/24/2022 15:03:12 - INFO - codeparrot_training - Step 10414: {'lr': 0.0004630402605416825, 'samples': 5332480, 'steps': 10414, 'loss/train': 3.142756700515747} 02/24/2022 15:03:18 - INFO - codeparrot_training - Step 10415: {'lr': 0.0004630316979426689, 'samples': 5332992, 'steps': 10415, 'loss/train': 2.4721357822418213} 02/24/2022 15:03:21 - INFO - codeparrot_training - Step 10416: {'lr': 0.00046302313443109523, 'samples': 5333504, 'steps': 10416, 'loss/train': 2.2335314750671387} 02/24/2022 15:03:27 - INFO - codeparrot_training - Step 10417: {'lr': 0.00046301457000699807, 'samples': 5334016, 'steps': 10417, 'loss/train': 3.5762007236480713} 02/24/2022 15:03:30 - INFO - codeparrot_training - Step 10418: {'lr': 0.0004630060046704141, 'samples': 5334528, 'steps': 10418, 'loss/train': 2.231334924697876} 02/24/2022 15:03:36 - INFO - codeparrot_training - Step 10419: {'lr': 0.0004629974384213801, 'samples': 5335040, 'steps': 10419, 'loss/train': 2.758619546890259} 02/24/2022 15:03:39 - INFO - codeparrot_training - Step 10420: {'lr': 0.0004629888712599327, 'samples': 5335552, 'steps': 10420, 'loss/train': 1.9766424894332886} 02/24/2022 15:03:45 - INFO - codeparrot_training - Step 10421: {'lr': 0.0004629803031861086, 'samples': 5336064, 'steps': 10421, 'loss/train': 3.0400912761688232} 02/24/2022 15:03:48 - INFO - codeparrot_training - Step 10422: {'lr': 0.0004629717341999445, 'samples': 5336576, 'steps': 10422, 'loss/train': 2.448423385620117} 02/24/2022 15:03:55 - INFO - codeparrot_training - Step 10423: {'lr': 0.0004629631643014771, 'samples': 5337088, 'steps': 10423, 'loss/train': 1.903863787651062} 02/24/2022 15:03:58 - INFO - codeparrot_training - Step 10424: {'lr': 0.00046295459349074316, 'samples': 5337600, 'steps': 10424, 'loss/train': 2.0889322757720947} 02/24/2022 15:04:04 - INFO - codeparrot_training - Step 10425: {'lr': 0.00046294602176777936, 'samples': 5338112, 'steps': 10425, 'loss/train': 1.5387037992477417} 02/24/2022 15:04:07 - INFO - codeparrot_training - Step 10426: {'lr': 0.0004629374491326224, 'samples': 5338624, 'steps': 10426, 'loss/train': 1.9643384218215942} 02/24/2022 15:04:13 - INFO - codeparrot_training - Step 10427: {'lr': 0.00046292887558530905, 'samples': 5339136, 'steps': 10427, 'loss/train': 1.6737027168273926} 02/24/2022 15:04:16 - INFO - codeparrot_training - Step 10428: {'lr': 0.000462920301125876, 'samples': 5339648, 'steps': 10428, 'loss/train': 2.879610776901245} 02/24/2022 15:04:22 - INFO - codeparrot_training - Step 10429: {'lr': 0.0004629117257543599, 'samples': 5340160, 'steps': 10429, 'loss/train': 2.397786855697632} 02/24/2022 15:04:25 - INFO - codeparrot_training - Step 10430: {'lr': 0.0004629031494707977, 'samples': 5340672, 'steps': 10430, 'loss/train': 2.504424810409546} 02/24/2022 15:04:31 - INFO - codeparrot_training - Step 10431: {'lr': 0.00046289457227522595, 'samples': 5341184, 'steps': 10431, 'loss/train': 2.152984142303467} 02/24/2022 15:04:34 - INFO - codeparrot_training - Step 10432: {'lr': 0.0004628859941676815, 'samples': 5341696, 'steps': 10432, 'loss/train': 2.405923843383789} 02/24/2022 15:04:40 - INFO - codeparrot_training - Step 10433: {'lr': 0.000462877415148201, 'samples': 5342208, 'steps': 10433, 'loss/train': 1.4465056657791138} 02/24/2022 15:04:44 - INFO - codeparrot_training - Step 10434: {'lr': 0.0004628688352168213, 'samples': 5342720, 'steps': 10434, 'loss/train': 1.8572214841842651} 02/24/2022 15:04:49 - INFO - codeparrot_training - Step 10435: {'lr': 0.00046286025437357905, 'samples': 5343232, 'steps': 10435, 'loss/train': 0.44998013973236084} 02/24/2022 15:04:53 - INFO - codeparrot_training - Step 10436: {'lr': 0.00046285167261851114, 'samples': 5343744, 'steps': 10436, 'loss/train': 0.5047389268875122} 02/24/2022 15:04:58 - INFO - codeparrot_training - Step 10437: {'lr': 0.00046284308995165414, 'samples': 5344256, 'steps': 10437, 'loss/train': 1.4467048645019531} 02/24/2022 15:05:02 - INFO - codeparrot_training - Step 10438: {'lr': 0.00046283450637304497, 'samples': 5344768, 'steps': 10438, 'loss/train': 2.112070083618164} 02/24/2022 15:05:07 - INFO - codeparrot_training - Step 10439: {'lr': 0.0004628259218827204, 'samples': 5345280, 'steps': 10439, 'loss/train': 1.5674123764038086} 02/24/2022 15:05:11 - INFO - codeparrot_training - Step 10440: {'lr': 0.0004628173364807171, 'samples': 5345792, 'steps': 10440, 'loss/train': 2.8122034072875977} 02/24/2022 15:05:16 - INFO - codeparrot_training - Step 10441: {'lr': 0.00046280875016707195, 'samples': 5346304, 'steps': 10441, 'loss/train': 1.816400170326233} 02/24/2022 15:05:20 - INFO - codeparrot_training - Step 10442: {'lr': 0.0004628001629418217, 'samples': 5346816, 'steps': 10442, 'loss/train': 2.770390033721924} 02/24/2022 15:05:26 - INFO - codeparrot_training - Step 10443: {'lr': 0.0004627915748050031, 'samples': 5347328, 'steps': 10443, 'loss/train': 2.79180908203125} 02/24/2022 15:05:29 - INFO - codeparrot_training - Step 10444: {'lr': 0.000462782985756653, 'samples': 5347840, 'steps': 10444, 'loss/train': 1.995596170425415} 02/24/2022 15:05:35 - INFO - codeparrot_training - Step 10445: {'lr': 0.0004627743957968081, 'samples': 5348352, 'steps': 10445, 'loss/train': 1.672791838645935} 02/24/2022 15:05:39 - INFO - codeparrot_training - Step 10446: {'lr': 0.00046276580492550523, 'samples': 5348864, 'steps': 10446, 'loss/train': 2.561729907989502} 02/24/2022 15:05:44 - INFO - codeparrot_training - Step 10447: {'lr': 0.0004627572131427813, 'samples': 5349376, 'steps': 10447, 'loss/train': 1.3663554191589355} 02/24/2022 15:05:50 - INFO - codeparrot_training - Step 10448: {'lr': 0.000462748620448673, 'samples': 5349888, 'steps': 10448, 'loss/train': 2.5932722091674805} 02/24/2022 15:05:53 - INFO - codeparrot_training - Step 10449: {'lr': 0.00046274002684321716, 'samples': 5350400, 'steps': 10449, 'loss/train': 2.050666570663452} 02/24/2022 15:05:59 - INFO - codeparrot_training - Step 10450: {'lr': 0.00046273143232645054, 'samples': 5350912, 'steps': 10450, 'loss/train': 2.084097146987915} 02/24/2022 15:06:02 - INFO - codeparrot_training - Step 10451: {'lr': 0.0004627228368984101, 'samples': 5351424, 'steps': 10451, 'loss/train': 1.6187797784805298} 02/24/2022 15:06:08 - INFO - codeparrot_training - Step 10452: {'lr': 0.0004627142405591325, 'samples': 5351936, 'steps': 10452, 'loss/train': 1.9603959321975708} 02/24/2022 15:06:11 - INFO - codeparrot_training - Step 10453: {'lr': 0.00046270564330865466, 'samples': 5352448, 'steps': 10453, 'loss/train': 0.7993656992912292} 02/24/2022 15:06:17 - INFO - codeparrot_training - Step 10454: {'lr': 0.0004626970451470134, 'samples': 5352960, 'steps': 10454, 'loss/train': 1.8133491277694702} 02/24/2022 15:06:21 - INFO - codeparrot_training - Step 10455: {'lr': 0.0004626884460742455, 'samples': 5353472, 'steps': 10455, 'loss/train': 1.907637596130371} 02/24/2022 15:06:24 - INFO - codeparrot_training - Step 10456: {'lr': 0.00046267984609038793, 'samples': 5353984, 'steps': 10456, 'loss/train': 8.732415199279785} 02/24/2022 15:06:30 - INFO - codeparrot_training - Step 10457: {'lr': 0.0004626712451954773, 'samples': 5354496, 'steps': 10457, 'loss/train': 2.1762301921844482} 02/24/2022 15:06:33 - INFO - codeparrot_training - Step 10458: {'lr': 0.0004626626433895507, 'samples': 5355008, 'steps': 10458, 'loss/train': 0.9044533967971802} 02/24/2022 15:06:40 - INFO - codeparrot_training - Step 10459: {'lr': 0.00046265404067264484, 'samples': 5355520, 'steps': 10459, 'loss/train': 2.3630051612854004} 02/24/2022 15:06:43 - INFO - codeparrot_training - Step 10460: {'lr': 0.00046264543704479654, 'samples': 5356032, 'steps': 10460, 'loss/train': 2.51969051361084} 02/24/2022 15:06:49 - INFO - codeparrot_training - Step 10461: {'lr': 0.0004626368325060428, 'samples': 5356544, 'steps': 10461, 'loss/train': 1.684999704360962} 02/24/2022 15:06:52 - INFO - codeparrot_training - Step 10462: {'lr': 0.00046262822705642025, 'samples': 5357056, 'steps': 10462, 'loss/train': 2.6834254264831543} 02/24/2022 15:06:58 - INFO - codeparrot_training - Step 10463: {'lr': 0.00046261962069596603, 'samples': 5357568, 'steps': 10463, 'loss/train': 2.5085251331329346} 02/24/2022 15:07:01 - INFO - codeparrot_training - Step 10464: {'lr': 0.0004626110134247168, 'samples': 5358080, 'steps': 10464, 'loss/train': 2.391413688659668} 02/24/2022 15:07:07 - INFO - codeparrot_training - Step 10465: {'lr': 0.0004626024052427095, 'samples': 5358592, 'steps': 10465, 'loss/train': 1.5142011642456055} 02/24/2022 15:07:10 - INFO - codeparrot_training - Step 10466: {'lr': 0.00046259379614998103, 'samples': 5359104, 'steps': 10466, 'loss/train': 2.719632387161255} 02/24/2022 15:07:16 - INFO - codeparrot_training - Step 10467: {'lr': 0.00046258518614656827, 'samples': 5359616, 'steps': 10467, 'loss/train': 1.3169454336166382} 02/24/2022 15:07:19 - INFO - codeparrot_training - Step 10468: {'lr': 0.0004625765752325081, 'samples': 5360128, 'steps': 10468, 'loss/train': 2.7741124629974365} 02/24/2022 15:07:25 - INFO - codeparrot_training - Step 10469: {'lr': 0.0004625679634078372, 'samples': 5360640, 'steps': 10469, 'loss/train': 3.4212605953216553} 02/24/2022 15:07:29 - INFO - codeparrot_training - Step 10470: {'lr': 0.0004625593506725928, 'samples': 5361152, 'steps': 10470, 'loss/train': 2.8670907020568848} 02/24/2022 15:07:34 - INFO - codeparrot_training - Step 10471: {'lr': 0.0004625507370268116, 'samples': 5361664, 'steps': 10471, 'loss/train': 2.548942804336548} 02/24/2022 15:07:38 - INFO - codeparrot_training - Step 10472: {'lr': 0.00046254212247053055, 'samples': 5362176, 'steps': 10472, 'loss/train': 2.207308053970337} 02/24/2022 15:07:43 - INFO - codeparrot_training - Step 10473: {'lr': 0.00046253350700378655, 'samples': 5362688, 'steps': 10473, 'loss/train': 1.3895708322525024} 02/24/2022 15:07:47 - INFO - codeparrot_training - Step 10474: {'lr': 0.0004625248906266165, 'samples': 5363200, 'steps': 10474, 'loss/train': 2.06227970123291} 02/24/2022 15:07:52 - INFO - codeparrot_training - Step 10475: {'lr': 0.00046251627333905723, 'samples': 5363712, 'steps': 10475, 'loss/train': 2.119565486907959} 02/24/2022 15:07:56 - INFO - codeparrot_training - Step 10476: {'lr': 0.0004625076551411458, 'samples': 5364224, 'steps': 10476, 'loss/train': 1.8095057010650635} 02/24/2022 15:08:01 - INFO - codeparrot_training - Step 10477: {'lr': 0.000462499036032919, 'samples': 5364736, 'steps': 10477, 'loss/train': 2.1605141162872314} 02/24/2022 15:08:05 - INFO - codeparrot_training - Step 10478: {'lr': 0.0004624904160144138, 'samples': 5365248, 'steps': 10478, 'loss/train': 2.337312936782837} 02/24/2022 15:08:11 - INFO - codeparrot_training - Step 10479: {'lr': 0.00046248179508566716, 'samples': 5365760, 'steps': 10479, 'loss/train': 1.1478919982910156} 02/24/2022 15:08:15 - INFO - codeparrot_training - Step 10480: {'lr': 0.000462473173246716, 'samples': 5366272, 'steps': 10480, 'loss/train': 2.8255481719970703} 02/24/2022 15:08:20 - INFO - codeparrot_training - Step 10481: {'lr': 0.00046246455049759716, 'samples': 5366784, 'steps': 10481, 'loss/train': 1.9251611232757568} 02/24/2022 15:08:24 - INFO - codeparrot_training - Step 10482: {'lr': 0.00046245592683834773, 'samples': 5367296, 'steps': 10482, 'loss/train': 1.8433847427368164} 02/24/2022 15:08:29 - INFO - codeparrot_training - Step 10483: {'lr': 0.00046244730226900453, 'samples': 5367808, 'steps': 10483, 'loss/train': 1.8261759281158447} 02/24/2022 15:08:33 - INFO - codeparrot_training - Step 10484: {'lr': 0.00046243867678960463, 'samples': 5368320, 'steps': 10484, 'loss/train': 0.6374279260635376} 02/24/2022 15:08:38 - INFO - codeparrot_training - Step 10485: {'lr': 0.00046243005040018484, 'samples': 5368832, 'steps': 10485, 'loss/train': 2.717811107635498} 02/24/2022 15:08:42 - INFO - codeparrot_training - Step 10486: {'lr': 0.0004624214231007821, 'samples': 5369344, 'steps': 10486, 'loss/train': 1.7226450443267822} 02/24/2022 15:08:47 - INFO - codeparrot_training - Step 10487: {'lr': 0.0004624127948914335, 'samples': 5369856, 'steps': 10487, 'loss/train': 1.8952337503433228} 02/24/2022 15:08:51 - INFO - codeparrot_training - Step 10488: {'lr': 0.0004624041657721759, 'samples': 5370368, 'steps': 10488, 'loss/train': 3.1727347373962402} 02/24/2022 15:08:56 - INFO - codeparrot_training - Step 10489: {'lr': 0.0004623955357430464, 'samples': 5370880, 'steps': 10489, 'loss/train': 1.7089414596557617} 02/24/2022 15:09:00 - INFO - codeparrot_training - Step 10490: {'lr': 0.0004623869048040817, 'samples': 5371392, 'steps': 10490, 'loss/train': 2.485283374786377} 02/24/2022 15:09:06 - INFO - codeparrot_training - Step 10491: {'lr': 0.0004623782729553191, 'samples': 5371904, 'steps': 10491, 'loss/train': 3.259850263595581} 02/24/2022 15:09:09 - INFO - codeparrot_training - Step 10492: {'lr': 0.00046236964019679533, 'samples': 5372416, 'steps': 10492, 'loss/train': 2.32462739944458} 02/24/2022 15:09:15 - INFO - codeparrot_training - Step 10493: {'lr': 0.0004623610065285475, 'samples': 5372928, 'steps': 10493, 'loss/train': 1.2415096759796143} 02/24/2022 15:09:18 - INFO - codeparrot_training - Step 10494: {'lr': 0.00046235237195061253, 'samples': 5373440, 'steps': 10494, 'loss/train': 2.2489330768585205} 02/24/2022 15:09:24 - INFO - codeparrot_training - Step 10495: {'lr': 0.00046234373646302743, 'samples': 5373952, 'steps': 10495, 'loss/train': 1.502928614616394} 02/24/2022 15:09:28 - INFO - codeparrot_training - Step 10496: {'lr': 0.00046233510006582913, 'samples': 5374464, 'steps': 10496, 'loss/train': 2.110473871231079} 02/24/2022 15:09:33 - INFO - codeparrot_training - Step 10497: {'lr': 0.00046232646275905475, 'samples': 5374976, 'steps': 10497, 'loss/train': 2.732248306274414} 02/24/2022 15:09:36 - INFO - codeparrot_training - Step 10498: {'lr': 0.00046231782454274117, 'samples': 5375488, 'steps': 10498, 'loss/train': 0.805681049823761} 02/24/2022 15:09:42 - INFO - codeparrot_training - Step 10499: {'lr': 0.00046230918541692557, 'samples': 5376000, 'steps': 10499, 'loss/train': 1.8752543926239014} 02/24/2022 15:09:45 - INFO - codeparrot_training - Step 10500: {'lr': 0.00046230054538164475, 'samples': 5376512, 'steps': 10500, 'loss/train': 2.2577922344207764} 02/24/2022 15:09:52 - INFO - codeparrot_training - Step 10501: {'lr': 0.0004622919044369358, 'samples': 5377024, 'steps': 10501, 'loss/train': 1.9790390729904175} 02/24/2022 15:09:55 - INFO - codeparrot_training - Step 10502: {'lr': 0.00046228326258283576, 'samples': 5377536, 'steps': 10502, 'loss/train': 2.1342992782592773} 02/24/2022 15:10:01 - INFO - codeparrot_training - Step 10503: {'lr': 0.0004622746198193816, 'samples': 5378048, 'steps': 10503, 'loss/train': 1.1480612754821777} 02/24/2022 15:10:04 - INFO - codeparrot_training - Step 10504: {'lr': 0.00046226597614661044, 'samples': 5378560, 'steps': 10504, 'loss/train': 1.5505146980285645} 02/24/2022 15:10:10 - INFO - codeparrot_training - Step 10505: {'lr': 0.00046225733156455916, 'samples': 5379072, 'steps': 10505, 'loss/train': 1.1708292961120605} 02/24/2022 15:10:13 - INFO - codeparrot_training - Step 10506: {'lr': 0.00046224868607326494, 'samples': 5379584, 'steps': 10506, 'loss/train': 1.649767279624939} 02/24/2022 15:10:19 - INFO - codeparrot_training - Step 10507: {'lr': 0.00046224003967276474, 'samples': 5380096, 'steps': 10507, 'loss/train': 2.3073229789733887} 02/24/2022 15:10:22 - INFO - codeparrot_training - Step 10508: {'lr': 0.00046223139236309553, 'samples': 5380608, 'steps': 10508, 'loss/train': 1.979972004890442} 02/24/2022 15:10:28 - INFO - codeparrot_training - Step 10509: {'lr': 0.0004622227441442945, 'samples': 5381120, 'steps': 10509, 'loss/train': 1.5092341899871826} 02/24/2022 15:10:31 - INFO - codeparrot_training - Step 10510: {'lr': 0.00046221409501639863, 'samples': 5381632, 'steps': 10510, 'loss/train': 2.424701452255249} 02/24/2022 15:10:37 - INFO - codeparrot_training - Step 10511: {'lr': 0.0004622054449794449, 'samples': 5382144, 'steps': 10511, 'loss/train': 2.3061470985412598} 02/24/2022 15:10:40 - INFO - codeparrot_training - Step 10512: {'lr': 0.0004621967940334705, 'samples': 5382656, 'steps': 10512, 'loss/train': 2.4230778217315674} 02/24/2022 15:10:46 - INFO - codeparrot_training - Step 10513: {'lr': 0.00046218814217851233, 'samples': 5383168, 'steps': 10513, 'loss/train': 1.9350486993789673} 02/24/2022 15:10:49 - INFO - codeparrot_training - Step 10514: {'lr': 0.0004621794894146076, 'samples': 5383680, 'steps': 10514, 'loss/train': 1.7866779565811157} 02/24/2022 15:10:55 - INFO - codeparrot_training - Step 10515: {'lr': 0.0004621708357417933, 'samples': 5384192, 'steps': 10515, 'loss/train': 3.268122673034668} 02/24/2022 15:11:01 - INFO - codeparrot_training - Step 10516: {'lr': 0.00046216218116010646, 'samples': 5384704, 'steps': 10516, 'loss/train': 1.798248052597046} 02/24/2022 15:11:04 - INFO - codeparrot_training - Step 10517: {'lr': 0.00046215352566958423, 'samples': 5385216, 'steps': 10517, 'loss/train': 2.2035131454467773} 02/24/2022 15:11:10 - INFO - codeparrot_training - Step 10518: {'lr': 0.00046214486927026373, 'samples': 5385728, 'steps': 10518, 'loss/train': 1.996370792388916} 02/24/2022 15:11:13 - INFO - codeparrot_training - Step 10519: {'lr': 0.0004621362119621819, 'samples': 5386240, 'steps': 10519, 'loss/train': 2.58803653717041} 02/24/2022 15:11:19 - INFO - codeparrot_training - Step 10520: {'lr': 0.00046212755374537594, 'samples': 5386752, 'steps': 10520, 'loss/train': 0.9014103412628174} 02/24/2022 15:11:22 - INFO - codeparrot_training - Step 10521: {'lr': 0.00046211889461988286, 'samples': 5387264, 'steps': 10521, 'loss/train': 2.401191473007202} 02/24/2022 15:11:28 - INFO - codeparrot_training - Step 10522: {'lr': 0.0004621102345857399, 'samples': 5387776, 'steps': 10522, 'loss/train': 2.2674074172973633} 02/24/2022 15:11:31 - INFO - codeparrot_training - Step 10523: {'lr': 0.0004621015736429839, 'samples': 5388288, 'steps': 10523, 'loss/train': 1.804938554763794} 02/24/2022 15:11:36 - INFO - codeparrot_training - Step 10524: {'lr': 0.00046209291179165216, 'samples': 5388800, 'steps': 10524, 'loss/train': 2.3942768573760986} 02/24/2022 15:11:40 - INFO - codeparrot_training - Step 10525: {'lr': 0.0004620842490317817, 'samples': 5389312, 'steps': 10525, 'loss/train': 1.725327730178833} 02/24/2022 15:11:46 - INFO - codeparrot_training - Step 10526: {'lr': 0.0004620755853634097, 'samples': 5389824, 'steps': 10526, 'loss/train': 2.086566686630249} 02/24/2022 15:11:50 - INFO - codeparrot_training - Step 10527: {'lr': 0.00046206692078657325, 'samples': 5390336, 'steps': 10527, 'loss/train': 2.959632158279419} 02/24/2022 15:11:55 - INFO - codeparrot_training - Step 10528: {'lr': 0.0004620582553013094, 'samples': 5390848, 'steps': 10528, 'loss/train': 1.8163808584213257} 02/24/2022 15:11:59 - INFO - codeparrot_training - Step 10529: {'lr': 0.00046204958890765536, 'samples': 5391360, 'steps': 10529, 'loss/train': 1.567827582359314} 02/24/2022 15:12:04 - INFO - codeparrot_training - Step 10530: {'lr': 0.0004620409216056483, 'samples': 5391872, 'steps': 10530, 'loss/train': 1.542049765586853} 02/24/2022 15:12:08 - INFO - codeparrot_training - Step 10531: {'lr': 0.00046203225339532515, 'samples': 5392384, 'steps': 10531, 'loss/train': 2.0172841548919678} 02/24/2022 15:12:13 - INFO - codeparrot_training - Step 10532: {'lr': 0.00046202358427672313, 'samples': 5392896, 'steps': 10532, 'loss/train': 2.221881151199341} 02/24/2022 15:12:17 - INFO - codeparrot_training - Step 10533: {'lr': 0.0004620149142498795, 'samples': 5393408, 'steps': 10533, 'loss/train': 2.393617630004883} 02/24/2022 15:12:22 - INFO - codeparrot_training - Step 10534: {'lr': 0.0004620062433148312, 'samples': 5393920, 'steps': 10534, 'loss/train': 2.6493568420410156} 02/24/2022 15:12:26 - INFO - codeparrot_training - Step 10535: {'lr': 0.00046199757147161554, 'samples': 5394432, 'steps': 10535, 'loss/train': 0.931510865688324} 02/24/2022 15:12:32 - INFO - codeparrot_training - Step 10536: {'lr': 0.00046198889872026963, 'samples': 5394944, 'steps': 10536, 'loss/train': 2.0688412189483643} 02/24/2022 15:12:36 - INFO - codeparrot_training - Step 10537: {'lr': 0.0004619802250608305, 'samples': 5395456, 'steps': 10537, 'loss/train': 1.9365615844726562} 02/24/2022 15:12:41 - INFO - codeparrot_training - Step 10538: {'lr': 0.0004619715504933354, 'samples': 5395968, 'steps': 10538, 'loss/train': 1.5142697095870972} 02/24/2022 15:12:45 - INFO - codeparrot_training - Step 10539: {'lr': 0.00046196287501782155, 'samples': 5396480, 'steps': 10539, 'loss/train': 2.4453554153442383} 02/24/2022 15:12:50 - INFO - codeparrot_training - Step 10540: {'lr': 0.00046195419863432604, 'samples': 5396992, 'steps': 10540, 'loss/train': 1.559774398803711} 02/24/2022 15:12:54 - INFO - codeparrot_training - Step 10541: {'lr': 0.000461945521342886, 'samples': 5397504, 'steps': 10541, 'loss/train': 1.8514355421066284} 02/24/2022 15:12:59 - INFO - codeparrot_training - Step 10542: {'lr': 0.0004619368431435387, 'samples': 5398016, 'steps': 10542, 'loss/train': 0.19608449935913086} 02/24/2022 15:13:03 - INFO - codeparrot_training - Step 10543: {'lr': 0.0004619281640363212, 'samples': 5398528, 'steps': 10543, 'loss/train': 2.3130812644958496} 02/24/2022 15:13:08 - INFO - codeparrot_training - Step 10544: {'lr': 0.0004619194840212708, 'samples': 5399040, 'steps': 10544, 'loss/train': 1.5492312908172607} 02/24/2022 15:13:12 - INFO - codeparrot_training - Step 10545: {'lr': 0.00046191080309842457, 'samples': 5399552, 'steps': 10545, 'loss/train': 2.719177722930908} 02/24/2022 15:13:18 - INFO - codeparrot_training - Step 10546: {'lr': 0.0004619021212678198, 'samples': 5400064, 'steps': 10546, 'loss/train': 1.6152968406677246} 02/24/2022 15:13:22 - INFO - codeparrot_training - Step 10547: {'lr': 0.0004618934385294936, 'samples': 5400576, 'steps': 10547, 'loss/train': 2.703350782394409} 02/24/2022 15:13:27 - INFO - codeparrot_training - Step 10548: {'lr': 0.0004618847548834833, 'samples': 5401088, 'steps': 10548, 'loss/train': 2.941051721572876} 02/24/2022 15:13:30 - INFO - codeparrot_training - Step 10549: {'lr': 0.0004618760703298258, 'samples': 5401600, 'steps': 10549, 'loss/train': 2.6009654998779297} 02/24/2022 15:13:36 - INFO - codeparrot_training - Step 10550: {'lr': 0.0004618673848685586, 'samples': 5402112, 'steps': 10550, 'loss/train': 0.8980075716972351} 02/24/2022 15:13:40 - INFO - codeparrot_training - Step 10551: {'lr': 0.00046185869849971884, 'samples': 5402624, 'steps': 10551, 'loss/train': 2.0634307861328125} 02/24/2022 15:13:45 - INFO - codeparrot_training - Step 10552: {'lr': 0.0004618500112233436, 'samples': 5403136, 'steps': 10552, 'loss/train': 0.6996252536773682} 02/24/2022 15:13:49 - INFO - codeparrot_training - Step 10553: {'lr': 0.0004618413230394702, 'samples': 5403648, 'steps': 10553, 'loss/train': 2.558077573776245} 02/24/2022 15:13:54 - INFO - codeparrot_training - Step 10554: {'lr': 0.0004618326339481359, 'samples': 5404160, 'steps': 10554, 'loss/train': 2.5209367275238037} 02/24/2022 15:13:58 - INFO - codeparrot_training - Step 10555: {'lr': 0.00046182394394937774, 'samples': 5404672, 'steps': 10555, 'loss/train': 1.4326109886169434} 02/24/2022 15:14:03 - INFO - codeparrot_training - Step 10556: {'lr': 0.00046181525304323325, 'samples': 5405184, 'steps': 10556, 'loss/train': 2.0825324058532715} 02/24/2022 15:14:07 - INFO - codeparrot_training - Step 10557: {'lr': 0.0004618065612297393, 'samples': 5405696, 'steps': 10557, 'loss/train': 1.4221198558807373} 02/24/2022 15:14:12 - INFO - codeparrot_training - Step 10558: {'lr': 0.00046179786850893335, 'samples': 5406208, 'steps': 10558, 'loss/train': 2.5532338619232178} 02/24/2022 15:14:16 - INFO - codeparrot_training - Step 10559: {'lr': 0.0004617891748808526, 'samples': 5406720, 'steps': 10559, 'loss/train': 1.6183686256408691} 02/24/2022 15:14:22 - INFO - codeparrot_training - Step 10560: {'lr': 0.0004617804803455343, 'samples': 5407232, 'steps': 10560, 'loss/train': 2.1259236335754395} 02/24/2022 15:14:25 - INFO - codeparrot_training - Step 10561: {'lr': 0.0004617717849030156, 'samples': 5407744, 'steps': 10561, 'loss/train': 1.8071929216384888} 02/24/2022 15:14:31 - INFO - codeparrot_training - Step 10562: {'lr': 0.00046176308855333395, 'samples': 5408256, 'steps': 10562, 'loss/train': 2.0819756984710693} 02/24/2022 15:14:35 - INFO - codeparrot_training - Step 10563: {'lr': 0.00046175439129652636, 'samples': 5408768, 'steps': 10563, 'loss/train': 2.60795259475708} 02/24/2022 15:14:38 - INFO - codeparrot_training - Step 10564: {'lr': 0.0004617456931326302, 'samples': 5409280, 'steps': 10564, 'loss/train': 1.0081745386123657} 02/24/2022 15:14:44 - INFO - codeparrot_training - Step 10565: {'lr': 0.00046173699406168277, 'samples': 5409792, 'steps': 10565, 'loss/train': 2.4007821083068848} 02/24/2022 15:14:47 - INFO - codeparrot_training - Step 10566: {'lr': 0.00046172829408372125, 'samples': 5410304, 'steps': 10566, 'loss/train': 3.254162549972534} 02/24/2022 15:14:52 - INFO - codeparrot_training - Step 10567: {'lr': 0.000461719593198783, 'samples': 5410816, 'steps': 10567, 'loss/train': 1.541430115699768} 02/24/2022 15:14:58 - INFO - codeparrot_training - Step 10568: {'lr': 0.0004617108914069052, 'samples': 5411328, 'steps': 10568, 'loss/train': 1.8166934251785278} 02/24/2022 15:15:01 - INFO - codeparrot_training - Step 10569: {'lr': 0.00046170218870812517, 'samples': 5411840, 'steps': 10569, 'loss/train': 2.7922027111053467} 02/24/2022 15:15:07 - INFO - codeparrot_training - Step 10570: {'lr': 0.0004616934851024802, 'samples': 5412352, 'steps': 10570, 'loss/train': 1.125774621963501} 02/24/2022 15:15:10 - INFO - codeparrot_training - Step 10571: {'lr': 0.00046168478059000753, 'samples': 5412864, 'steps': 10571, 'loss/train': 1.5489860773086548} 02/24/2022 15:15:17 - INFO - codeparrot_training - Step 10572: {'lr': 0.0004616760751707445, 'samples': 5413376, 'steps': 10572, 'loss/train': 1.6841152906417847} 02/24/2022 15:15:20 - INFO - codeparrot_training - Step 10573: {'lr': 0.0004616673688447284, 'samples': 5413888, 'steps': 10573, 'loss/train': 2.14851450920105} 02/24/2022 15:15:26 - INFO - codeparrot_training - Step 10574: {'lr': 0.0004616586616119964, 'samples': 5414400, 'steps': 10574, 'loss/train': 1.840282917022705} 02/24/2022 15:15:29 - INFO - codeparrot_training - Step 10575: {'lr': 0.0004616499534725861, 'samples': 5414912, 'steps': 10575, 'loss/train': 1.9462885856628418} 02/24/2022 15:15:35 - INFO - codeparrot_training - Step 10576: {'lr': 0.00046164124442653445, 'samples': 5415424, 'steps': 10576, 'loss/train': 1.85312020778656} 02/24/2022 15:15:38 - INFO - codeparrot_training - Step 10577: {'lr': 0.00046163253447387896, 'samples': 5415936, 'steps': 10577, 'loss/train': 2.4246721267700195} 02/24/2022 15:15:44 - INFO - codeparrot_training - Step 10578: {'lr': 0.0004616238236146569, 'samples': 5416448, 'steps': 10578, 'loss/train': 2.5775511264801025} 02/24/2022 15:15:47 - INFO - codeparrot_training - Step 10579: {'lr': 0.0004616151118489056, 'samples': 5416960, 'steps': 10579, 'loss/train': 2.937816858291626} 02/24/2022 15:15:53 - INFO - codeparrot_training - Step 10580: {'lr': 0.0004616063991766623, 'samples': 5417472, 'steps': 10580, 'loss/train': 0.9874318242073059} 02/24/2022 15:15:56 - INFO - codeparrot_training - Step 10581: {'lr': 0.00046159768559796437, 'samples': 5417984, 'steps': 10581, 'loss/train': 1.7239164113998413} 02/24/2022 15:16:02 - INFO - codeparrot_training - Step 10582: {'lr': 0.0004615889711128492, 'samples': 5418496, 'steps': 10582, 'loss/train': 2.2855613231658936} 02/24/2022 15:16:06 - INFO - codeparrot_training - Step 10583: {'lr': 0.00046158025572135404, 'samples': 5419008, 'steps': 10583, 'loss/train': 3.1169204711914062} 02/24/2022 15:16:11 - INFO - codeparrot_training - Step 10584: {'lr': 0.00046157153942351625, 'samples': 5419520, 'steps': 10584, 'loss/train': 1.6316964626312256} 02/24/2022 15:16:15 - INFO - codeparrot_training - Step 10585: {'lr': 0.0004615628222193732, 'samples': 5420032, 'steps': 10585, 'loss/train': 4.91056489944458} 02/24/2022 15:16:20 - INFO - codeparrot_training - Step 10586: {'lr': 0.00046155410410896215, 'samples': 5420544, 'steps': 10586, 'loss/train': 2.4259636402130127} 02/24/2022 15:16:24 - INFO - codeparrot_training - Step 10587: {'lr': 0.00046154538509232044, 'samples': 5421056, 'steps': 10587, 'loss/train': 2.300079107284546} 02/24/2022 15:16:29 - INFO - codeparrot_training - Step 10588: {'lr': 0.00046153666516948554, 'samples': 5421568, 'steps': 10588, 'loss/train': 1.9895038604736328} 02/24/2022 15:16:33 - INFO - codeparrot_training - Step 10589: {'lr': 0.0004615279443404948, 'samples': 5422080, 'steps': 10589, 'loss/train': 1.9219489097595215} 02/24/2022 15:16:38 - INFO - codeparrot_training - Step 10590: {'lr': 0.0004615192226053855, 'samples': 5422592, 'steps': 10590, 'loss/train': 1.7846968173980713} 02/24/2022 15:16:42 - INFO - codeparrot_training - Step 10591: {'lr': 0.0004615104999641949, 'samples': 5423104, 'steps': 10591, 'loss/train': 2.1698429584503174} 02/24/2022 15:16:48 - INFO - codeparrot_training - Step 10592: {'lr': 0.0004615017764169606, 'samples': 5423616, 'steps': 10592, 'loss/train': 0.9332539439201355} 02/24/2022 15:16:52 - INFO - codeparrot_training - Step 10593: {'lr': 0.0004614930519637198, 'samples': 5424128, 'steps': 10593, 'loss/train': 2.5743725299835205} 02/24/2022 15:16:57 - INFO - codeparrot_training - Step 10594: {'lr': 0.0004614843266045099, 'samples': 5424640, 'steps': 10594, 'loss/train': 2.1349785327911377} 02/24/2022 15:17:01 - INFO - codeparrot_training - Step 10595: {'lr': 0.0004614756003393683, 'samples': 5425152, 'steps': 10595, 'loss/train': 1.0838006734848022} 02/24/2022 15:17:06 - INFO - codeparrot_training - Step 10596: {'lr': 0.00046146687316833235, 'samples': 5425664, 'steps': 10596, 'loss/train': 3.6415281295776367} 02/24/2022 15:17:10 - INFO - codeparrot_training - Step 10597: {'lr': 0.00046145814509143955, 'samples': 5426176, 'steps': 10597, 'loss/train': 1.4740712642669678} 02/24/2022 15:17:15 - INFO - codeparrot_training - Step 10598: {'lr': 0.0004614494161087271, 'samples': 5426688, 'steps': 10598, 'loss/train': 2.9835093021392822} 02/24/2022 15:17:19 - INFO - codeparrot_training - Step 10599: {'lr': 0.00046144068622023263, 'samples': 5427200, 'steps': 10599, 'loss/train': 1.950249195098877} 02/24/2022 15:17:24 - INFO - codeparrot_training - Step 10600: {'lr': 0.00046143195542599336, 'samples': 5427712, 'steps': 10600, 'loss/train': 1.9784170389175415} 02/24/2022 15:17:28 - INFO - codeparrot_training - Step 10601: {'lr': 0.00046142322372604667, 'samples': 5428224, 'steps': 10601, 'loss/train': 2.5392589569091797} 02/24/2022 15:17:33 - INFO - codeparrot_training - Step 10602: {'lr': 0.00046141449112043, 'samples': 5428736, 'steps': 10602, 'loss/train': 3.438002586364746} 02/24/2022 15:17:37 - INFO - codeparrot_training - Step 10603: {'lr': 0.0004614057576091809, 'samples': 5429248, 'steps': 10603, 'loss/train': 1.6824004650115967} 02/24/2022 15:17:42 - INFO - codeparrot_training - Step 10604: {'lr': 0.00046139702319233656, 'samples': 5429760, 'steps': 10604, 'loss/train': 3.164616823196411} 02/24/2022 15:17:46 - INFO - codeparrot_training - Step 10605: {'lr': 0.00046138828786993456, 'samples': 5430272, 'steps': 10605, 'loss/train': 2.3131275177001953} 02/24/2022 15:17:51 - INFO - codeparrot_training - Step 10606: {'lr': 0.0004613795516420122, 'samples': 5430784, 'steps': 10606, 'loss/train': 1.7368401288986206} 02/24/2022 15:17:55 - INFO - codeparrot_training - Step 10607: {'lr': 0.000461370814508607, 'samples': 5431296, 'steps': 10607, 'loss/train': 2.6561665534973145} 02/24/2022 15:18:01 - INFO - codeparrot_training - Step 10608: {'lr': 0.00046136207646975635, 'samples': 5431808, 'steps': 10608, 'loss/train': 2.068272352218628} 02/24/2022 15:18:05 - INFO - codeparrot_training - Step 10609: {'lr': 0.0004613533375254977, 'samples': 5432320, 'steps': 10609, 'loss/train': 2.232407331466675} 02/24/2022 15:18:10 - INFO - codeparrot_training - Step 10610: {'lr': 0.00046134459767586847, 'samples': 5432832, 'steps': 10610, 'loss/train': 2.3803439140319824} 02/24/2022 15:18:14 - INFO - codeparrot_training - Step 10611: {'lr': 0.00046133585692090603, 'samples': 5433344, 'steps': 10611, 'loss/train': 1.705320119857788} 02/24/2022 15:18:19 - INFO - codeparrot_training - Step 10612: {'lr': 0.0004613271152606479, 'samples': 5433856, 'steps': 10612, 'loss/train': 1.8641812801361084} 02/24/2022 15:18:23 - INFO - codeparrot_training - Step 10613: {'lr': 0.00046131837269513154, 'samples': 5434368, 'steps': 10613, 'loss/train': 1.3345494270324707} 02/24/2022 15:18:28 - INFO - codeparrot_training - Step 10614: {'lr': 0.00046130962922439435, 'samples': 5434880, 'steps': 10614, 'loss/train': 3.0980565547943115} 02/24/2022 15:18:32 - INFO - codeparrot_training - Step 10615: {'lr': 0.00046130088484847383, 'samples': 5435392, 'steps': 10615, 'loss/train': 1.9705994129180908} 02/24/2022 15:18:37 - INFO - codeparrot_training - Step 10616: {'lr': 0.0004612921395674074, 'samples': 5435904, 'steps': 10616, 'loss/train': 2.215280055999756} 02/24/2022 15:18:41 - INFO - codeparrot_training - Step 10617: {'lr': 0.00046128339338123253, 'samples': 5436416, 'steps': 10617, 'loss/train': 2.13520884513855} 02/24/2022 15:18:47 - INFO - codeparrot_training - Step 10618: {'lr': 0.0004612746462899867, 'samples': 5436928, 'steps': 10618, 'loss/train': 2.596813440322876} 02/24/2022 15:18:50 - INFO - codeparrot_training - Step 10619: {'lr': 0.00046126589829370736, 'samples': 5437440, 'steps': 10619, 'loss/train': 2.2310431003570557} 02/24/2022 15:18:56 - INFO - codeparrot_training - Step 10620: {'lr': 0.00046125714939243204, 'samples': 5437952, 'steps': 10620, 'loss/train': 2.1403725147247314} 02/24/2022 15:18:59 - INFO - codeparrot_training - Step 10621: {'lr': 0.00046124839958619815, 'samples': 5438464, 'steps': 10621, 'loss/train': 2.138474464416504} 02/24/2022 15:19:05 - INFO - codeparrot_training - Step 10622: {'lr': 0.0004612396488750432, 'samples': 5438976, 'steps': 10622, 'loss/train': 2.5780928134918213} 02/24/2022 15:19:08 - INFO - codeparrot_training - Step 10623: {'lr': 0.00046123089725900464, 'samples': 5439488, 'steps': 10623, 'loss/train': 2.0884387493133545} 02/24/2022 15:19:14 - INFO - codeparrot_training - Step 10624: {'lr': 0.00046122214473812005, 'samples': 5440000, 'steps': 10624, 'loss/train': 1.0501275062561035} 02/24/2022 15:19:17 - INFO - codeparrot_training - Step 10625: {'lr': 0.0004612133913124268, 'samples': 5440512, 'steps': 10625, 'loss/train': 2.3679120540618896} 02/24/2022 15:19:23 - INFO - codeparrot_training - Step 10626: {'lr': 0.00046120463698196245, 'samples': 5441024, 'steps': 10626, 'loss/train': 2.4974405765533447} 02/24/2022 15:19:26 - INFO - codeparrot_training - Step 10627: {'lr': 0.00046119588174676454, 'samples': 5441536, 'steps': 10627, 'loss/train': 2.1923253536224365} 02/24/2022 15:19:32 - INFO - codeparrot_training - Step 10628: {'lr': 0.0004611871256068705, 'samples': 5442048, 'steps': 10628, 'loss/train': 2.741480827331543} 02/24/2022 15:19:36 - INFO - codeparrot_training - Step 10629: {'lr': 0.0004611783685623179, 'samples': 5442560, 'steps': 10629, 'loss/train': 3.049236536026001} 02/24/2022 15:19:41 - INFO - codeparrot_training - Step 10630: {'lr': 0.00046116961061314424, 'samples': 5443072, 'steps': 10630, 'loss/train': 2.523798704147339} 02/24/2022 15:19:45 - INFO - codeparrot_training - Step 10631: {'lr': 0.00046116085175938694, 'samples': 5443584, 'steps': 10631, 'loss/train': 4.917596340179443} 02/24/2022 15:19:50 - INFO - codeparrot_training - Step 10632: {'lr': 0.00046115209200108366, 'samples': 5444096, 'steps': 10632, 'loss/train': 1.781620740890503} 02/24/2022 15:19:54 - INFO - codeparrot_training - Step 10633: {'lr': 0.00046114333133827194, 'samples': 5444608, 'steps': 10633, 'loss/train': 2.1847450733184814} 02/24/2022 15:19:59 - INFO - codeparrot_training - Step 10634: {'lr': 0.0004611345697709891, 'samples': 5445120, 'steps': 10634, 'loss/train': 2.132807731628418} 02/24/2022 15:20:03 - INFO - codeparrot_training - Step 10635: {'lr': 0.0004611258072992729, 'samples': 5445632, 'steps': 10635, 'loss/train': 1.9459401369094849} 02/24/2022 15:20:08 - INFO - codeparrot_training - Step 10636: {'lr': 0.0004611170439231607, 'samples': 5446144, 'steps': 10636, 'loss/train': 2.5081591606140137} 02/24/2022 15:20:12 - INFO - codeparrot_training - Step 10637: {'lr': 0.0004611082796426902, 'samples': 5446656, 'steps': 10637, 'loss/train': 2.3935770988464355} 02/24/2022 15:20:18 - INFO - codeparrot_training - Step 10638: {'lr': 0.00046109951445789883, 'samples': 5447168, 'steps': 10638, 'loss/train': 1.6926934719085693} 02/24/2022 15:20:22 - INFO - codeparrot_training - Step 10639: {'lr': 0.00046109074836882415, 'samples': 5447680, 'steps': 10639, 'loss/train': 1.7920057773590088} 02/24/2022 15:20:27 - INFO - codeparrot_training - Step 10640: {'lr': 0.00046108198137550377, 'samples': 5448192, 'steps': 10640, 'loss/train': 1.946357011795044} 02/24/2022 15:20:33 - INFO - codeparrot_training - Step 10641: {'lr': 0.0004610732134779752, 'samples': 5448704, 'steps': 10641, 'loss/train': 2.303535223007202} 02/24/2022 15:20:36 - INFO - codeparrot_training - Step 10642: {'lr': 0.000461064444676276, 'samples': 5449216, 'steps': 10642, 'loss/train': 2.0571703910827637} 02/24/2022 15:20:40 - INFO - codeparrot_training - Step 10643: {'lr': 0.0004610556749704438, 'samples': 5449728, 'steps': 10643, 'loss/train': 1.1858084201812744} 02/24/2022 15:20:45 - INFO - codeparrot_training - Step 10644: {'lr': 0.000461046904360516, 'samples': 5450240, 'steps': 10644, 'loss/train': 3.731553316116333} 02/24/2022 15:20:51 - INFO - codeparrot_training - Step 10645: {'lr': 0.0004610381328465303, 'samples': 5450752, 'steps': 10645, 'loss/train': 2.4287173748016357} 02/24/2022 15:20:54 - INFO - codeparrot_training - Step 10646: {'lr': 0.0004610293604285243, 'samples': 5451264, 'steps': 10646, 'loss/train': 1.9319217205047607} 02/24/2022 15:21:00 - INFO - codeparrot_training - Step 10647: {'lr': 0.0004610205871065355, 'samples': 5451776, 'steps': 10647, 'loss/train': 0.9761449098587036} 02/24/2022 15:21:03 - INFO - codeparrot_training - Step 10648: {'lr': 0.0004610118128806016, 'samples': 5452288, 'steps': 10648, 'loss/train': 2.363436698913574} 02/24/2022 15:21:09 - INFO - codeparrot_training - Step 10649: {'lr': 0.0004610030377507599, 'samples': 5452800, 'steps': 10649, 'loss/train': 2.598527669906616} 02/24/2022 15:21:12 - INFO - codeparrot_training - Step 10650: {'lr': 0.0004609942617170483, 'samples': 5453312, 'steps': 10650, 'loss/train': 1.288580060005188} 02/24/2022 15:21:18 - INFO - codeparrot_training - Step 10651: {'lr': 0.0004609854847795043, 'samples': 5453824, 'steps': 10651, 'loss/train': 1.7000689506530762} 02/24/2022 15:21:21 - INFO - codeparrot_training - Step 10652: {'lr': 0.0004609767069381655, 'samples': 5454336, 'steps': 10652, 'loss/train': 2.576692819595337} 02/24/2022 15:21:27 - INFO - codeparrot_training - Step 10653: {'lr': 0.00046096792819306945, 'samples': 5454848, 'steps': 10653, 'loss/train': 1.9039483070373535} 02/24/2022 15:21:31 - INFO - codeparrot_training - Step 10654: {'lr': 0.00046095914854425376, 'samples': 5455360, 'steps': 10654, 'loss/train': 2.774906635284424} 02/24/2022 15:21:36 - INFO - codeparrot_training - Step 10655: {'lr': 0.00046095036799175606, 'samples': 5455872, 'steps': 10655, 'loss/train': 1.4152904748916626} 02/24/2022 15:21:40 - INFO - codeparrot_training - Step 10656: {'lr': 0.000460941586535614, 'samples': 5456384, 'steps': 10656, 'loss/train': 3.429565668106079} 02/24/2022 15:21:45 - INFO - codeparrot_training - Step 10657: {'lr': 0.00046093280417586517, 'samples': 5456896, 'steps': 10657, 'loss/train': 0.16530273854732513} 02/24/2022 15:21:49 - INFO - codeparrot_training - Step 10658: {'lr': 0.0004609240209125472, 'samples': 5457408, 'steps': 10658, 'loss/train': 1.9924647808074951} 02/24/2022 15:21:54 - INFO - codeparrot_training - Step 10659: {'lr': 0.00046091523674569765, 'samples': 5457920, 'steps': 10659, 'loss/train': 1.7584319114685059} 02/24/2022 15:21:58 - INFO - codeparrot_training - Step 10660: {'lr': 0.00046090645167535415, 'samples': 5458432, 'steps': 10660, 'loss/train': 2.420118570327759} 02/24/2022 15:22:03 - INFO - codeparrot_training - Step 10661: {'lr': 0.00046089766570155447, 'samples': 5458944, 'steps': 10661, 'loss/train': 0.7556182146072388} 02/24/2022 15:22:07 - INFO - codeparrot_training - Step 10662: {'lr': 0.0004608888788243362, 'samples': 5459456, 'steps': 10662, 'loss/train': 2.5232532024383545} 02/24/2022 15:22:13 - INFO - codeparrot_training - Step 10663: {'lr': 0.00046088009104373683, 'samples': 5459968, 'steps': 10663, 'loss/train': 1.2358235120773315} 02/24/2022 15:22:17 - INFO - codeparrot_training - Step 10664: {'lr': 0.0004608713023597941, 'samples': 5460480, 'steps': 10664, 'loss/train': 2.8571650981903076} 02/24/2022 15:22:22 - INFO - codeparrot_training - Step 10665: {'lr': 0.0004608625127725458, 'samples': 5460992, 'steps': 10665, 'loss/train': 2.98763108253479} 02/24/2022 15:22:26 - INFO - codeparrot_training - Step 10666: {'lr': 0.0004608537222820294, 'samples': 5461504, 'steps': 10666, 'loss/train': 2.2872185707092285} 02/24/2022 15:22:31 - INFO - codeparrot_training - Step 10667: {'lr': 0.0004608449308882826, 'samples': 5462016, 'steps': 10667, 'loss/train': 2.736595869064331} 02/24/2022 15:22:35 - INFO - codeparrot_training - Step 10668: {'lr': 0.000460836138591343, 'samples': 5462528, 'steps': 10668, 'loss/train': 1.7485947608947754} 02/24/2022 15:22:40 - INFO - codeparrot_training - Step 10669: {'lr': 0.0004608273453912484, 'samples': 5463040, 'steps': 10669, 'loss/train': 1.9313406944274902} 02/24/2022 15:22:44 - INFO - codeparrot_training - Step 10670: {'lr': 0.0004608185512880364, 'samples': 5463552, 'steps': 10670, 'loss/train': 3.0228350162506104} 02/24/2022 15:22:49 - INFO - codeparrot_training - Step 10671: {'lr': 0.0004608097562817446, 'samples': 5464064, 'steps': 10671, 'loss/train': 3.2250454425811768} 02/24/2022 15:22:52 - INFO - codeparrot_training - Step 10672: {'lr': 0.0004608009603724108, 'samples': 5464576, 'steps': 10672, 'loss/train': 2.2080538272857666} 02/24/2022 15:22:58 - INFO - codeparrot_training - Step 10673: {'lr': 0.0004607921635600726, 'samples': 5465088, 'steps': 10673, 'loss/train': 1.8597644567489624} 02/24/2022 15:23:01 - INFO - codeparrot_training - Step 10674: {'lr': 0.00046078336584476777, 'samples': 5465600, 'steps': 10674, 'loss/train': 1.579274296760559} 02/24/2022 15:23:08 - INFO - codeparrot_training - Step 10675: {'lr': 0.00046077456722653387, 'samples': 5466112, 'steps': 10675, 'loss/train': 1.0815410614013672} 02/24/2022 15:23:11 - INFO - codeparrot_training - Step 10676: {'lr': 0.00046076576770540865, 'samples': 5466624, 'steps': 10676, 'loss/train': 2.622030019760132} 02/24/2022 15:23:17 - INFO - codeparrot_training - Step 10677: {'lr': 0.00046075696728142986, 'samples': 5467136, 'steps': 10677, 'loss/train': 1.6888563632965088} 02/24/2022 15:23:20 - INFO - codeparrot_training - Step 10678: {'lr': 0.0004607481659546351, 'samples': 5467648, 'steps': 10678, 'loss/train': 1.7208749055862427} 02/24/2022 15:23:26 - INFO - codeparrot_training - Step 10679: {'lr': 0.0004607393637250621, 'samples': 5468160, 'steps': 10679, 'loss/train': 3.6741042137145996} 02/24/2022 15:23:29 - INFO - codeparrot_training - Step 10680: {'lr': 0.00046073056059274867, 'samples': 5468672, 'steps': 10680, 'loss/train': 1.9011914730072021} 02/24/2022 15:23:35 - INFO - codeparrot_training - Step 10681: {'lr': 0.0004607217565577323, 'samples': 5469184, 'steps': 10681, 'loss/train': 2.0433027744293213} 02/24/2022 15:23:38 - INFO - codeparrot_training - Step 10682: {'lr': 0.0004607129516200509, 'samples': 5469696, 'steps': 10682, 'loss/train': 3.231966495513916} 02/24/2022 15:23:44 - INFO - codeparrot_training - Step 10683: {'lr': 0.00046070414577974216, 'samples': 5470208, 'steps': 10683, 'loss/train': 1.6576157808303833} 02/24/2022 15:23:47 - INFO - codeparrot_training - Step 10684: {'lr': 0.00046069533903684374, 'samples': 5470720, 'steps': 10684, 'loss/train': 1.943198800086975} 02/24/2022 15:23:53 - INFO - codeparrot_training - Step 10685: {'lr': 0.00046068653139139337, 'samples': 5471232, 'steps': 10685, 'loss/train': 2.388160467147827} 02/24/2022 15:23:57 - INFO - codeparrot_training - Step 10686: {'lr': 0.0004606777228434288, 'samples': 5471744, 'steps': 10686, 'loss/train': 2.79186749458313} 02/24/2022 15:24:02 - INFO - codeparrot_training - Step 10687: {'lr': 0.00046066891339298783, 'samples': 5472256, 'steps': 10687, 'loss/train': 2.1405656337738037} 02/24/2022 15:24:06 - INFO - codeparrot_training - Step 10688: {'lr': 0.0004606601030401081, 'samples': 5472768, 'steps': 10688, 'loss/train': 2.591496706008911} 02/24/2022 15:24:12 - INFO - codeparrot_training - Step 10689: {'lr': 0.00046065129178482733, 'samples': 5473280, 'steps': 10689, 'loss/train': 1.8618096113204956} 02/24/2022 15:24:15 - INFO - codeparrot_training - Step 10690: {'lr': 0.0004606424796271834, 'samples': 5473792, 'steps': 10690, 'loss/train': 3.25329852104187} 02/24/2022 15:24:21 - INFO - codeparrot_training - Step 10691: {'lr': 0.0004606336665672139, 'samples': 5474304, 'steps': 10691, 'loss/train': 1.3476084470748901} 02/24/2022 15:24:24 - INFO - codeparrot_training - Step 10692: {'lr': 0.00046062485260495666, 'samples': 5474816, 'steps': 10692, 'loss/train': 2.275230646133423} 02/24/2022 15:24:30 - INFO - codeparrot_training - Step 10693: {'lr': 0.00046061603774044945, 'samples': 5475328, 'steps': 10693, 'loss/train': 1.5449998378753662} 02/24/2022 15:24:33 - INFO - codeparrot_training - Step 10694: {'lr': 0.00046060722197373, 'samples': 5475840, 'steps': 10694, 'loss/train': 1.3837188482284546} 02/24/2022 15:24:39 - INFO - codeparrot_training - Step 10695: {'lr': 0.0004605984053048361, 'samples': 5476352, 'steps': 10695, 'loss/train': 0.9562422037124634} 02/24/2022 15:24:42 - INFO - codeparrot_training - Step 10696: {'lr': 0.0004605895877338055, 'samples': 5476864, 'steps': 10696, 'loss/train': 3.118607521057129} 02/24/2022 15:24:48 - INFO - codeparrot_training - Step 10697: {'lr': 0.000460580769260676, 'samples': 5477376, 'steps': 10697, 'loss/train': 2.567796230316162} 02/24/2022 15:24:51 - INFO - codeparrot_training - Step 10698: {'lr': 0.0004605719498854853, 'samples': 5477888, 'steps': 10698, 'loss/train': 1.7771592140197754} 02/24/2022 15:24:57 - INFO - codeparrot_training - Step 10699: {'lr': 0.0004605631296082713, 'samples': 5478400, 'steps': 10699, 'loss/train': 2.086648941040039} 02/24/2022 15:25:00 - INFO - codeparrot_training - Step 10700: {'lr': 0.0004605543084290716, 'samples': 5478912, 'steps': 10700, 'loss/train': 2.4179673194885254} 02/24/2022 15:25:06 - INFO - codeparrot_training - Step 10701: {'lr': 0.00046054548634792426, 'samples': 5479424, 'steps': 10701, 'loss/train': 3.4130523204803467} 02/24/2022 15:25:10 - INFO - codeparrot_training - Step 10702: {'lr': 0.0004605366633648668, 'samples': 5479936, 'steps': 10702, 'loss/train': 2.9978597164154053} 02/24/2022 15:25:15 - INFO - codeparrot_training - Step 10703: {'lr': 0.00046052783947993713, 'samples': 5480448, 'steps': 10703, 'loss/train': 1.5903500318527222} 02/24/2022 15:25:19 - INFO - codeparrot_training - Step 10704: {'lr': 0.0004605190146931731, 'samples': 5480960, 'steps': 10704, 'loss/train': 1.268998384475708} 02/24/2022 15:25:24 - INFO - codeparrot_training - Step 10705: {'lr': 0.0004605101890046124, 'samples': 5481472, 'steps': 10705, 'loss/train': 2.714695930480957} 02/24/2022 15:25:28 - INFO - codeparrot_training - Step 10706: {'lr': 0.00046050136241429295, 'samples': 5481984, 'steps': 10706, 'loss/train': 2.2369601726531982} 02/24/2022 15:25:33 - INFO - codeparrot_training - Step 10707: {'lr': 0.0004604925349222525, 'samples': 5482496, 'steps': 10707, 'loss/train': 2.965559959411621} 02/24/2022 15:25:37 - INFO - codeparrot_training - Step 10708: {'lr': 0.00046048370652852885, 'samples': 5483008, 'steps': 10708, 'loss/train': 0.43953466415405273} 02/24/2022 15:25:42 - INFO - codeparrot_training - Step 10709: {'lr': 0.00046047487723315986, 'samples': 5483520, 'steps': 10709, 'loss/train': 0.8210500478744507} 02/24/2022 15:25:46 - INFO - codeparrot_training - Step 10710: {'lr': 0.0004604660470361832, 'samples': 5484032, 'steps': 10710, 'loss/train': 2.3636443614959717} 02/24/2022 15:25:52 - INFO - codeparrot_training - Step 10711: {'lr': 0.000460457215937637, 'samples': 5484544, 'steps': 10711, 'loss/train': 1.7563261985778809} 02/24/2022 15:25:56 - INFO - codeparrot_training - Step 10712: {'lr': 0.00046044838393755885, 'samples': 5485056, 'steps': 10712, 'loss/train': 1.3773894309997559} 02/24/2022 15:26:01 - INFO - codeparrot_training - Step 10713: {'lr': 0.0004604395510359867, 'samples': 5485568, 'steps': 10713, 'loss/train': 1.2420666217803955} 02/24/2022 15:26:05 - INFO - codeparrot_training - Step 10714: {'lr': 0.0004604307172329582, 'samples': 5486080, 'steps': 10714, 'loss/train': 1.2597712278366089} 02/24/2022 15:26:10 - INFO - codeparrot_training - Step 10715: {'lr': 0.0004604218825285114, 'samples': 5486592, 'steps': 10715, 'loss/train': 2.457947015762329} 02/24/2022 15:26:14 - INFO - codeparrot_training - Step 10716: {'lr': 0.00046041304692268407, 'samples': 5487104, 'steps': 10716, 'loss/train': 1.957581877708435} 02/24/2022 15:26:19 - INFO - codeparrot_training - Step 10717: {'lr': 0.00046040421041551404, 'samples': 5487616, 'steps': 10717, 'loss/train': 1.2665934562683105} 02/24/2022 15:26:23 - INFO - codeparrot_training - Step 10718: {'lr': 0.00046039537300703926, 'samples': 5488128, 'steps': 10718, 'loss/train': 2.0186431407928467} 02/24/2022 15:26:28 - INFO - codeparrot_training - Step 10719: {'lr': 0.00046038653469729747, 'samples': 5488640, 'steps': 10719, 'loss/train': 1.9662803411483765} 02/24/2022 15:26:32 - INFO - codeparrot_training - Step 10720: {'lr': 0.00046037769548632656, 'samples': 5489152, 'steps': 10720, 'loss/train': 0.5897735357284546} 02/24/2022 15:26:38 - INFO - codeparrot_training - Step 10721: {'lr': 0.0004603688553741644, 'samples': 5489664, 'steps': 10721, 'loss/train': 2.4701669216156006} 02/24/2022 15:26:42 - INFO - codeparrot_training - Step 10722: {'lr': 0.0004603600143608488, 'samples': 5490176, 'steps': 10722, 'loss/train': 2.212117910385132} 02/24/2022 15:26:47 - INFO - codeparrot_training - Step 10723: {'lr': 0.00046035117244641783, 'samples': 5490688, 'steps': 10723, 'loss/train': 2.6752448081970215} 02/24/2022 15:26:51 - INFO - codeparrot_training - Step 10724: {'lr': 0.0004603423296309092, 'samples': 5491200, 'steps': 10724, 'loss/train': 1.9107458591461182} 02/24/2022 15:26:56 - INFO - codeparrot_training - Step 10725: {'lr': 0.0004603334859143608, 'samples': 5491712, 'steps': 10725, 'loss/train': 1.4822959899902344} 02/24/2022 15:27:00 - INFO - codeparrot_training - Step 10726: {'lr': 0.0004603246412968105, 'samples': 5492224, 'steps': 10726, 'loss/train': 2.253957748413086} 02/24/2022 15:27:05 - INFO - codeparrot_training - Step 10727: {'lr': 0.00046031579577829616, 'samples': 5492736, 'steps': 10727, 'loss/train': 1.159075379371643} 02/24/2022 15:27:09 - INFO - codeparrot_training - Step 10728: {'lr': 0.00046030694935885586, 'samples': 5493248, 'steps': 10728, 'loss/train': 1.7162890434265137} 02/24/2022 15:27:14 - INFO - codeparrot_training - Step 10729: {'lr': 0.00046029810203852736, 'samples': 5493760, 'steps': 10729, 'loss/train': 1.3259321451187134} 02/24/2022 15:27:18 - INFO - codeparrot_training - Step 10730: {'lr': 0.00046028925381734855, 'samples': 5494272, 'steps': 10730, 'loss/train': 2.19993257522583} 02/24/2022 15:27:24 - INFO - codeparrot_training - Step 10731: {'lr': 0.00046028040469535734, 'samples': 5494784, 'steps': 10731, 'loss/train': 2.3265724182128906} 02/24/2022 15:27:27 - INFO - codeparrot_training - Step 10732: {'lr': 0.00046027155467259166, 'samples': 5495296, 'steps': 10732, 'loss/train': 1.720679759979248} 02/24/2022 15:27:33 - INFO - codeparrot_training - Step 10733: {'lr': 0.00046026270374908935, 'samples': 5495808, 'steps': 10733, 'loss/train': 2.1227848529815674} 02/24/2022 15:27:36 - INFO - codeparrot_training - Step 10734: {'lr': 0.0004602538519248884, 'samples': 5496320, 'steps': 10734, 'loss/train': 2.029430389404297} 02/24/2022 15:27:42 - INFO - codeparrot_training - Step 10735: {'lr': 0.00046024499920002676, 'samples': 5496832, 'steps': 10735, 'loss/train': 2.8200113773345947} 02/24/2022 15:27:45 - INFO - codeparrot_training - Step 10736: {'lr': 0.0004602361455745423, 'samples': 5497344, 'steps': 10736, 'loss/train': 2.5483901500701904} 02/24/2022 15:27:52 - INFO - codeparrot_training - Step 10737: {'lr': 0.00046022729104847293, 'samples': 5497856, 'steps': 10737, 'loss/train': 2.080850124359131} 02/24/2022 15:27:56 - INFO - codeparrot_training - Step 10738: {'lr': 0.0004602184356218566, 'samples': 5498368, 'steps': 10738, 'loss/train': 4.888075351715088} 02/24/2022 15:28:01 - INFO - codeparrot_training - Step 10739: {'lr': 0.0004602095792947312, 'samples': 5498880, 'steps': 10739, 'loss/train': 2.311472177505493} 02/24/2022 15:28:05 - INFO - codeparrot_training - Step 10740: {'lr': 0.00046020072206713484, 'samples': 5499392, 'steps': 10740, 'loss/train': 2.3736765384674072} 02/24/2022 15:28:10 - INFO - codeparrot_training - Step 10741: {'lr': 0.0004601918639391052, 'samples': 5499904, 'steps': 10741, 'loss/train': 1.785421371459961} 02/24/2022 15:28:14 - INFO - codeparrot_training - Step 10742: {'lr': 0.0004601830049106804, 'samples': 5500416, 'steps': 10742, 'loss/train': 2.111750841140747} 02/24/2022 15:28:19 - INFO - codeparrot_training - Step 10743: {'lr': 0.0004601741449818984, 'samples': 5500928, 'steps': 10743, 'loss/train': 2.075619697570801} 02/24/2022 15:28:23 - INFO - codeparrot_training - Step 10744: {'lr': 0.000460165284152797, 'samples': 5501440, 'steps': 10744, 'loss/train': 2.629167318344116} 02/24/2022 15:28:28 - INFO - codeparrot_training - Step 10745: {'lr': 0.0004601564224234143, 'samples': 5501952, 'steps': 10745, 'loss/train': 2.1144609451293945} 02/24/2022 15:28:32 - INFO - codeparrot_training - Step 10746: {'lr': 0.00046014755979378825, 'samples': 5502464, 'steps': 10746, 'loss/train': 2.298802614212036} 02/24/2022 15:28:38 - INFO - codeparrot_training - Step 10747: {'lr': 0.0004601386962639568, 'samples': 5502976, 'steps': 10747, 'loss/train': 1.6870520114898682} 02/24/2022 15:28:42 - INFO - codeparrot_training - Step 10748: {'lr': 0.0004601298318339578, 'samples': 5503488, 'steps': 10748, 'loss/train': 0.5674211382865906} 02/24/2022 15:28:47 - INFO - codeparrot_training - Step 10749: {'lr': 0.0004601209665038294, 'samples': 5504000, 'steps': 10749, 'loss/train': 1.2293298244476318} 02/24/2022 15:28:51 - INFO - codeparrot_training - Step 10750: {'lr': 0.0004601121002736095, 'samples': 5504512, 'steps': 10750, 'loss/train': 3.513444662094116} 02/24/2022 15:28:56 - INFO - codeparrot_training - Step 10751: {'lr': 0.0004601032331433361, 'samples': 5505024, 'steps': 10751, 'loss/train': 2.7403571605682373} 02/24/2022 15:29:00 - INFO - codeparrot_training - Step 10752: {'lr': 0.00046009436511304714, 'samples': 5505536, 'steps': 10752, 'loss/train': 2.4913318157196045} 02/24/2022 15:29:05 - INFO - codeparrot_training - Step 10753: {'lr': 0.0004600854961827806, 'samples': 5506048, 'steps': 10753, 'loss/train': 2.5256471633911133} 02/24/2022 15:29:09 - INFO - codeparrot_training - Step 10754: {'lr': 0.00046007662635257453, 'samples': 5506560, 'steps': 10754, 'loss/train': 2.621028423309326} 02/24/2022 15:29:14 - INFO - codeparrot_training - Step 10755: {'lr': 0.0004600677556224669, 'samples': 5507072, 'steps': 10755, 'loss/train': 2.0305933952331543} 02/24/2022 15:29:21 - INFO - codeparrot_training - Step 10756: {'lr': 0.00046005888399249575, 'samples': 5507584, 'steps': 10756, 'loss/train': 2.801264762878418} 02/24/2022 15:29:24 - INFO - codeparrot_training - Step 10757: {'lr': 0.000460050011462699, 'samples': 5508096, 'steps': 10757, 'loss/train': 2.604865074157715} 02/24/2022 15:29:30 - INFO - codeparrot_training - Step 10758: {'lr': 0.0004600411380331146, 'samples': 5508608, 'steps': 10758, 'loss/train': 1.9681214094161987} 02/24/2022 15:29:33 - INFO - codeparrot_training - Step 10759: {'lr': 0.0004600322637037808, 'samples': 5509120, 'steps': 10759, 'loss/train': 2.2678351402282715} 02/24/2022 15:29:37 - INFO - codeparrot_training - Step 10760: {'lr': 0.00046002338847473545, 'samples': 5509632, 'steps': 10760, 'loss/train': 1.1324044466018677} 02/24/2022 15:29:43 - INFO - codeparrot_training - Step 10761: {'lr': 0.00046001451234601665, 'samples': 5510144, 'steps': 10761, 'loss/train': 5.01030158996582} 02/24/2022 15:29:46 - INFO - codeparrot_training - Step 10762: {'lr': 0.0004600056353176623, 'samples': 5510656, 'steps': 10762, 'loss/train': 3.141014337539673} 02/24/2022 15:29:52 - INFO - codeparrot_training - Step 10763: {'lr': 0.00045999675738971047, 'samples': 5511168, 'steps': 10763, 'loss/train': 1.300363302230835} 02/24/2022 15:29:55 - INFO - codeparrot_training - Step 10764: {'lr': 0.00045998787856219925, 'samples': 5511680, 'steps': 10764, 'loss/train': 2.2624716758728027} 02/24/2022 15:30:01 - INFO - codeparrot_training - Step 10765: {'lr': 0.0004599789988351666, 'samples': 5512192, 'steps': 10765, 'loss/train': 3.016848564147949} 02/24/2022 15:30:04 - INFO - codeparrot_training - Step 10766: {'lr': 0.0004599701182086506, 'samples': 5512704, 'steps': 10766, 'loss/train': 2.2997207641601562} 02/24/2022 15:30:10 - INFO - codeparrot_training - Step 10767: {'lr': 0.0004599612366826893, 'samples': 5513216, 'steps': 10767, 'loss/train': 1.757472276687622} 02/24/2022 15:30:16 - INFO - codeparrot_training - Step 10768: {'lr': 0.00045995235425732076, 'samples': 5513728, 'steps': 10768, 'loss/train': 1.4633433818817139} 02/24/2022 15:30:19 - INFO - codeparrot_training - Step 10769: {'lr': 0.00045994347093258295, 'samples': 5514240, 'steps': 10769, 'loss/train': 2.670070171356201} 02/24/2022 15:30:25 - INFO - codeparrot_training - Step 10770: {'lr': 0.00045993458670851397, 'samples': 5514752, 'steps': 10770, 'loss/train': 2.204164743423462} 02/24/2022 15:30:28 - INFO - codeparrot_training - Step 10771: {'lr': 0.0004599257015851519, 'samples': 5515264, 'steps': 10771, 'loss/train': 2.4114902019500732} 02/24/2022 15:30:34 - INFO - codeparrot_training - Step 10772: {'lr': 0.0004599168155625348, 'samples': 5515776, 'steps': 10772, 'loss/train': 1.9435793161392212} 02/24/2022 15:30:37 - INFO - codeparrot_training - Step 10773: {'lr': 0.00045990792864070075, 'samples': 5516288, 'steps': 10773, 'loss/train': 3.077359199523926} 02/24/2022 15:30:43 - INFO - codeparrot_training - Step 10774: {'lr': 0.0004598990408196878, 'samples': 5516800, 'steps': 10774, 'loss/train': 3.066190481185913} 02/24/2022 15:30:46 - INFO - codeparrot_training - Step 10775: {'lr': 0.00045989015209953394, 'samples': 5517312, 'steps': 10775, 'loss/train': 0.8637327551841736} 02/24/2022 15:30:52 - INFO - codeparrot_training - Step 10776: {'lr': 0.00045988126248027735, 'samples': 5517824, 'steps': 10776, 'loss/train': 0.5748113393783569} 02/24/2022 15:30:55 - INFO - codeparrot_training - Step 10777: {'lr': 0.00045987237196195603, 'samples': 5518336, 'steps': 10777, 'loss/train': 2.196979522705078} 02/24/2022 15:31:01 - INFO - codeparrot_training - Step 10778: {'lr': 0.00045986348054460815, 'samples': 5518848, 'steps': 10778, 'loss/train': 2.402722120285034} 02/24/2022 15:31:04 - INFO - codeparrot_training - Step 10779: {'lr': 0.00045985458822827175, 'samples': 5519360, 'steps': 10779, 'loss/train': 1.4091347455978394} 02/24/2022 15:31:10 - INFO - codeparrot_training - Step 10780: {'lr': 0.0004598456950129849, 'samples': 5519872, 'steps': 10780, 'loss/train': 2.7950053215026855} 02/24/2022 15:31:13 - INFO - codeparrot_training - Step 10781: {'lr': 0.00045983680089878575, 'samples': 5520384, 'steps': 10781, 'loss/train': 2.7543787956237793} 02/24/2022 15:31:20 - INFO - codeparrot_training - Step 10782: {'lr': 0.0004598279058857124, 'samples': 5520896, 'steps': 10782, 'loss/train': 1.9277764558792114} 02/24/2022 15:31:23 - INFO - codeparrot_training - Step 10783: {'lr': 0.00045981900997380296, 'samples': 5521408, 'steps': 10783, 'loss/train': 2.1256773471832275} 02/24/2022 15:31:29 - INFO - codeparrot_training - Step 10784: {'lr': 0.0004598101131630954, 'samples': 5521920, 'steps': 10784, 'loss/train': 1.8175573348999023} 02/24/2022 15:31:32 - INFO - codeparrot_training - Step 10785: {'lr': 0.00045980121545362805, 'samples': 5522432, 'steps': 10785, 'loss/train': 2.7589738368988037} 02/24/2022 15:31:38 - INFO - codeparrot_training - Step 10786: {'lr': 0.0004597923168454389, 'samples': 5522944, 'steps': 10786, 'loss/train': 2.586043119430542} 02/24/2022 15:31:41 - INFO - codeparrot_training - Step 10787: {'lr': 0.000459783417338566, 'samples': 5523456, 'steps': 10787, 'loss/train': 1.959873914718628} 02/24/2022 15:31:47 - INFO - codeparrot_training - Step 10788: {'lr': 0.0004597745169330476, 'samples': 5523968, 'steps': 10788, 'loss/train': 2.5134518146514893} 02/24/2022 15:31:50 - INFO - codeparrot_training - Step 10789: {'lr': 0.0004597656156289217, 'samples': 5524480, 'steps': 10789, 'loss/train': 2.8788723945617676} 02/24/2022 15:31:56 - INFO - codeparrot_training - Step 10790: {'lr': 0.0004597567134262266, 'samples': 5524992, 'steps': 10790, 'loss/train': 1.55586576461792} 02/24/2022 15:31:59 - INFO - codeparrot_training - Step 10791: {'lr': 0.00045974781032500034, 'samples': 5525504, 'steps': 10791, 'loss/train': 1.7677415609359741} 02/24/2022 15:32:06 - INFO - codeparrot_training - Step 10792: {'lr': 0.00045973890632528106, 'samples': 5526016, 'steps': 10792, 'loss/train': 1.8218841552734375} 02/24/2022 15:32:09 - INFO - codeparrot_training - Step 10793: {'lr': 0.00045973000142710696, 'samples': 5526528, 'steps': 10793, 'loss/train': 2.461500644683838} 02/24/2022 15:32:13 - INFO - codeparrot_training - Step 10794: {'lr': 0.000459721095630516, 'samples': 5527040, 'steps': 10794, 'loss/train': 3.791186809539795} 02/24/2022 15:32:18 - INFO - codeparrot_training - Step 10795: {'lr': 0.00045971218893554655, 'samples': 5527552, 'steps': 10795, 'loss/train': 1.9880272150039673} 02/24/2022 15:32:22 - INFO - codeparrot_training - Step 10796: {'lr': 0.0004597032813422367, 'samples': 5528064, 'steps': 10796, 'loss/train': 2.9160258769989014} 02/24/2022 15:32:27 - INFO - codeparrot_training - Step 10797: {'lr': 0.00045969437285062453, 'samples': 5528576, 'steps': 10797, 'loss/train': 2.4157333374023438} 02/24/2022 15:32:31 - INFO - codeparrot_training - Step 10798: {'lr': 0.00045968546346074823, 'samples': 5529088, 'steps': 10798, 'loss/train': 3.254408359527588} 02/24/2022 15:32:37 - INFO - codeparrot_training - Step 10799: {'lr': 0.000459676553172646, 'samples': 5529600, 'steps': 10799, 'loss/train': 2.1163313388824463} 02/24/2022 15:32:40 - INFO - codeparrot_training - Step 10800: {'lr': 0.00045966764198635603, 'samples': 5530112, 'steps': 10800, 'loss/train': 2.4462223052978516} 02/24/2022 15:32:46 - INFO - codeparrot_training - Step 10801: {'lr': 0.0004596587299019164, 'samples': 5530624, 'steps': 10801, 'loss/train': 2.0351014137268066} 02/24/2022 15:32:49 - INFO - codeparrot_training - Step 10802: {'lr': 0.0004596498169193654, 'samples': 5531136, 'steps': 10802, 'loss/train': 1.1944851875305176} 02/24/2022 15:32:55 - INFO - codeparrot_training - Step 10803: {'lr': 0.00045964090303874115, 'samples': 5531648, 'steps': 10803, 'loss/train': 2.605156421661377} 02/24/2022 15:32:59 - INFO - codeparrot_training - Step 10804: {'lr': 0.0004596319882600818, 'samples': 5532160, 'steps': 10804, 'loss/train': 2.180009603500366} 02/24/2022 15:33:04 - INFO - codeparrot_training - Step 10805: {'lr': 0.00045962307258342564, 'samples': 5532672, 'steps': 10805, 'loss/train': 2.020578622817993} 02/24/2022 15:33:08 - INFO - codeparrot_training - Step 10806: {'lr': 0.00045961415600881075, 'samples': 5533184, 'steps': 10806, 'loss/train': 2.613784074783325} 02/24/2022 15:33:13 - INFO - codeparrot_training - Step 10807: {'lr': 0.0004596052385362754, 'samples': 5533696, 'steps': 10807, 'loss/train': 2.621772050857544} 02/24/2022 15:33:17 - INFO - codeparrot_training - Step 10808: {'lr': 0.00045959632016585774, 'samples': 5534208, 'steps': 10808, 'loss/train': 2.413472890853882} 02/24/2022 15:33:22 - INFO - codeparrot_training - Step 10809: {'lr': 0.00045958740089759606, 'samples': 5534720, 'steps': 10809, 'loss/train': 1.3405696153640747} 02/24/2022 15:33:26 - INFO - codeparrot_training - Step 10810: {'lr': 0.0004595784807315284, 'samples': 5535232, 'steps': 10810, 'loss/train': 2.480562210083008} 02/24/2022 15:33:31 - INFO - codeparrot_training - Step 10811: {'lr': 0.0004595695596676932, 'samples': 5535744, 'steps': 10811, 'loss/train': 1.643795371055603} 02/24/2022 15:33:35 - INFO - codeparrot_training - Step 10812: {'lr': 0.00045956063770612843, 'samples': 5536256, 'steps': 10812, 'loss/train': 2.5044960975646973} 02/24/2022 15:33:41 - INFO - codeparrot_training - Step 10813: {'lr': 0.00045955171484687255, 'samples': 5536768, 'steps': 10813, 'loss/train': 3.2657511234283447} 02/24/2022 15:33:44 - INFO - codeparrot_training - Step 10814: {'lr': 0.0004595427910899636, 'samples': 5537280, 'steps': 10814, 'loss/train': 1.3925988674163818} 02/24/2022 15:33:50 - INFO - codeparrot_training - Step 10815: {'lr': 0.00045953386643543987, 'samples': 5537792, 'steps': 10815, 'loss/train': 1.8427921533584595} 02/24/2022 15:33:53 - INFO - codeparrot_training - Step 10816: {'lr': 0.0004595249408833396, 'samples': 5538304, 'steps': 10816, 'loss/train': 2.8641726970672607} 02/24/2022 15:33:59 - INFO - codeparrot_training - Step 10817: {'lr': 0.00045951601443370107, 'samples': 5538816, 'steps': 10817, 'loss/train': 2.1586177349090576} 02/24/2022 15:34:02 - INFO - codeparrot_training - Step 10818: {'lr': 0.00045950708708656236, 'samples': 5539328, 'steps': 10818, 'loss/train': 2.1059703826904297} 02/24/2022 15:34:08 - INFO - codeparrot_training - Step 10819: {'lr': 0.0004594981588419619, 'samples': 5539840, 'steps': 10819, 'loss/train': 2.009880781173706} 02/24/2022 15:34:11 - INFO - codeparrot_training - Step 10820: {'lr': 0.00045948922969993777, 'samples': 5540352, 'steps': 10820, 'loss/train': 3.571247100830078} 02/24/2022 15:34:17 - INFO - codeparrot_training - Step 10821: {'lr': 0.00045948029966052834, 'samples': 5540864, 'steps': 10821, 'loss/train': 0.4738519489765167} 02/24/2022 15:34:20 - INFO - codeparrot_training - Step 10822: {'lr': 0.0004594713687237718, 'samples': 5541376, 'steps': 10822, 'loss/train': 2.604982614517212} 02/24/2022 15:34:26 - INFO - codeparrot_training - Step 10823: {'lr': 0.00045946243688970643, 'samples': 5541888, 'steps': 10823, 'loss/train': 2.6622674465179443} 02/24/2022 15:34:29 - INFO - codeparrot_training - Step 10824: {'lr': 0.00045945350415837056, 'samples': 5542400, 'steps': 10824, 'loss/train': 2.3063488006591797} 02/24/2022 15:34:36 - INFO - codeparrot_training - Step 10825: {'lr': 0.00045944457052980237, 'samples': 5542912, 'steps': 10825, 'loss/train': 1.1940544843673706} 02/24/2022 15:34:39 - INFO - codeparrot_training - Step 10826: {'lr': 0.0004594356360040401, 'samples': 5543424, 'steps': 10826, 'loss/train': 1.5733952522277832} 02/24/2022 15:34:44 - INFO - codeparrot_training - Step 10827: {'lr': 0.0004594267005811221, 'samples': 5543936, 'steps': 10827, 'loss/train': 1.4370841979980469} 02/24/2022 15:34:48 - INFO - codeparrot_training - Step 10828: {'lr': 0.0004594177642610866, 'samples': 5544448, 'steps': 10828, 'loss/train': 2.525839328765869} 02/24/2022 15:34:54 - INFO - codeparrot_training - Step 10829: {'lr': 0.0004594088270439719, 'samples': 5544960, 'steps': 10829, 'loss/train': 1.8740156888961792} 02/24/2022 15:34:57 - INFO - codeparrot_training - Step 10830: {'lr': 0.00045939988892981624, 'samples': 5545472, 'steps': 10830, 'loss/train': 1.3015533685684204} 02/24/2022 15:35:03 - INFO - codeparrot_training - Step 10831: {'lr': 0.00045939094991865806, 'samples': 5545984, 'steps': 10831, 'loss/train': 2.91900634765625} 02/24/2022 15:35:06 - INFO - codeparrot_training - Step 10832: {'lr': 0.00045938201001053546, 'samples': 5546496, 'steps': 10832, 'loss/train': 2.7854232788085938} 02/24/2022 15:35:12 - INFO - codeparrot_training - Step 10833: {'lr': 0.00045937306920548684, 'samples': 5547008, 'steps': 10833, 'loss/train': 2.633469581604004} 02/24/2022 15:35:15 - INFO - codeparrot_training - Step 10834: {'lr': 0.0004593641275035504, 'samples': 5547520, 'steps': 10834, 'loss/train': 2.394116163253784} 02/24/2022 15:35:21 - INFO - codeparrot_training - Step 10835: {'lr': 0.00045935518490476456, 'samples': 5548032, 'steps': 10835, 'loss/train': 2.002525806427002} 02/24/2022 15:35:24 - INFO - codeparrot_training - Step 10836: {'lr': 0.00045934624140916763, 'samples': 5548544, 'steps': 10836, 'loss/train': 2.4536209106445312} 02/24/2022 15:35:30 - INFO - codeparrot_training - Step 10837: {'lr': 0.0004593372970167978, 'samples': 5549056, 'steps': 10837, 'loss/train': 1.2019845247268677} 02/24/2022 15:35:36 - INFO - codeparrot_training - Step 10838: {'lr': 0.0004593283517276936, 'samples': 5549568, 'steps': 10838, 'loss/train': 2.4683220386505127} 02/24/2022 15:35:39 - INFO - codeparrot_training - Step 10839: {'lr': 0.0004593194055418931, 'samples': 5550080, 'steps': 10839, 'loss/train': 1.516682744026184} 02/24/2022 15:35:45 - INFO - codeparrot_training - Step 10840: {'lr': 0.00045931045845943474, 'samples': 5550592, 'steps': 10840, 'loss/train': 2.4613759517669678} 02/24/2022 15:35:48 - INFO - codeparrot_training - Step 10841: {'lr': 0.00045930151048035684, 'samples': 5551104, 'steps': 10841, 'loss/train': 1.887755036354065} 02/24/2022 15:35:54 - INFO - codeparrot_training - Step 10842: {'lr': 0.0004592925616046978, 'samples': 5551616, 'steps': 10842, 'loss/train': 1.085592269897461} 02/24/2022 15:35:57 - INFO - codeparrot_training - Step 10843: {'lr': 0.0004592836118324958, 'samples': 5552128, 'steps': 10843, 'loss/train': 1.5014142990112305} 02/24/2022 15:36:01 - INFO - codeparrot_training - Step 10844: {'lr': 0.0004592746611637893, 'samples': 5552640, 'steps': 10844, 'loss/train': 2.393275022506714} 02/24/2022 15:36:07 - INFO - codeparrot_training - Step 10845: {'lr': 0.00045926570959861656, 'samples': 5553152, 'steps': 10845, 'loss/train': 1.8171045780181885} 02/24/2022 15:36:10 - INFO - codeparrot_training - Step 10846: {'lr': 0.000459256757137016, 'samples': 5553664, 'steps': 10846, 'loss/train': 1.567799687385559} 02/24/2022 15:36:16 - INFO - codeparrot_training - Step 10847: {'lr': 0.00045924780377902595, 'samples': 5554176, 'steps': 10847, 'loss/train': 1.839835286140442} 02/24/2022 15:36:19 - INFO - codeparrot_training - Step 10848: {'lr': 0.00045923884952468475, 'samples': 5554688, 'steps': 10848, 'loss/train': 3.0737197399139404} 02/24/2022 15:36:25 - INFO - codeparrot_training - Step 10849: {'lr': 0.00045922989437403074, 'samples': 5555200, 'steps': 10849, 'loss/train': 2.6786131858825684} 02/24/2022 15:36:29 - INFO - codeparrot_training - Step 10850: {'lr': 0.0004592209383271023, 'samples': 5555712, 'steps': 10850, 'loss/train': 2.7745351791381836} 02/24/2022 15:36:34 - INFO - codeparrot_training - Step 10851: {'lr': 0.0004592119813839378, 'samples': 5556224, 'steps': 10851, 'loss/train': 1.8272593021392822} 02/24/2022 15:36:38 - INFO - codeparrot_training - Step 10852: {'lr': 0.0004592030235445757, 'samples': 5556736, 'steps': 10852, 'loss/train': 2.176370859146118} 02/24/2022 15:36:43 - INFO - codeparrot_training - Step 10853: {'lr': 0.00045919406480905413, 'samples': 5557248, 'steps': 10853, 'loss/train': 9.989912033081055} 02/24/2022 15:36:47 - INFO - codeparrot_training - Step 10854: {'lr': 0.0004591851051774117, 'samples': 5557760, 'steps': 10854, 'loss/train': 2.11321759223938} 02/24/2022 15:36:52 - INFO - codeparrot_training - Step 10855: {'lr': 0.00045917614464968665, 'samples': 5558272, 'steps': 10855, 'loss/train': 2.581662893295288} 02/24/2022 15:36:56 - INFO - codeparrot_training - Step 10856: {'lr': 0.0004591671832259174, 'samples': 5558784, 'steps': 10856, 'loss/train': 2.2340810298919678} 02/24/2022 15:37:01 - INFO - codeparrot_training - Step 10857: {'lr': 0.00045915822090614243, 'samples': 5559296, 'steps': 10857, 'loss/train': 1.391809105873108} 02/24/2022 15:37:05 - INFO - codeparrot_training - Step 10858: {'lr': 0.00045914925769040006, 'samples': 5559808, 'steps': 10858, 'loss/train': 1.831588625907898} 02/24/2022 15:37:11 - INFO - codeparrot_training - Step 10859: {'lr': 0.0004591402935787287, 'samples': 5560320, 'steps': 10859, 'loss/train': 2.4328787326812744} 02/24/2022 15:37:15 - INFO - codeparrot_training - Step 10860: {'lr': 0.00045913132857116663, 'samples': 5560832, 'steps': 10860, 'loss/train': 1.9180657863616943} 02/24/2022 15:37:20 - INFO - codeparrot_training - Step 10861: {'lr': 0.00045912236266775245, 'samples': 5561344, 'steps': 10861, 'loss/train': 2.2141122817993164} 02/24/2022 15:37:24 - INFO - codeparrot_training - Step 10862: {'lr': 0.0004591133958685244, 'samples': 5561856, 'steps': 10862, 'loss/train': 2.738919734954834} 02/24/2022 15:37:29 - INFO - codeparrot_training - Step 10863: {'lr': 0.00045910442817352095, 'samples': 5562368, 'steps': 10863, 'loss/train': 2.1409969329833984} 02/24/2022 15:37:33 - INFO - codeparrot_training - Step 10864: {'lr': 0.0004590954595827806, 'samples': 5562880, 'steps': 10864, 'loss/train': 1.8137991428375244} 02/24/2022 15:37:38 - INFO - codeparrot_training - Step 10865: {'lr': 0.00045908649009634165, 'samples': 5563392, 'steps': 10865, 'loss/train': 1.4388374090194702} 02/24/2022 15:37:42 - INFO - codeparrot_training - Step 10866: {'lr': 0.0004590775197142426, 'samples': 5563904, 'steps': 10866, 'loss/train': 1.8825113773345947} 02/24/2022 15:37:47 - INFO - codeparrot_training - Step 10867: {'lr': 0.0004590685484365218, 'samples': 5564416, 'steps': 10867, 'loss/train': 2.259903907775879} 02/24/2022 15:37:51 - INFO - codeparrot_training - Step 10868: {'lr': 0.00045905957626321775, 'samples': 5564928, 'steps': 10868, 'loss/train': 1.8798149824142456} 02/24/2022 15:37:56 - INFO - codeparrot_training - Step 10869: {'lr': 0.0004590506031943689, 'samples': 5565440, 'steps': 10869, 'loss/train': 1.683966875076294} 02/24/2022 15:38:00 - INFO - codeparrot_training - Step 10870: {'lr': 0.00045904162923001356, 'samples': 5565952, 'steps': 10870, 'loss/train': 2.2319300174713135} 02/24/2022 15:38:06 - INFO - codeparrot_training - Step 10871: {'lr': 0.00045903265437019036, 'samples': 5566464, 'steps': 10871, 'loss/train': 1.6371709108352661} 02/24/2022 15:38:09 - INFO - codeparrot_training - Step 10872: {'lr': 0.00045902367861493754, 'samples': 5566976, 'steps': 10872, 'loss/train': 2.279087781906128} 02/24/2022 15:38:15 - INFO - codeparrot_training - Step 10873: {'lr': 0.00045901470196429376, 'samples': 5567488, 'steps': 10873, 'loss/train': 1.701271891593933} 02/24/2022 15:38:18 - INFO - codeparrot_training - Step 10874: {'lr': 0.0004590057244182972, 'samples': 5568000, 'steps': 10874, 'loss/train': 1.8044042587280273} 02/24/2022 15:38:24 - INFO - codeparrot_training - Step 10875: {'lr': 0.0004589967459769867, 'samples': 5568512, 'steps': 10875, 'loss/train': 1.7072850465774536} 02/24/2022 15:38:27 - INFO - codeparrot_training - Step 10876: {'lr': 0.00045898776664040036, 'samples': 5569024, 'steps': 10876, 'loss/train': 1.484846830368042} 02/24/2022 15:38:33 - INFO - codeparrot_training - Step 10877: {'lr': 0.00045897878640857684, 'samples': 5569536, 'steps': 10877, 'loss/train': 2.2672948837280273} 02/24/2022 15:38:36 - INFO - codeparrot_training - Step 10878: {'lr': 0.00045896980528155454, 'samples': 5570048, 'steps': 10878, 'loss/train': 1.7351562976837158} 02/24/2022 15:38:42 - INFO - codeparrot_training - Step 10879: {'lr': 0.0004589608232593719, 'samples': 5570560, 'steps': 10879, 'loss/train': 1.8515613079071045} 02/24/2022 15:38:45 - INFO - codeparrot_training - Step 10880: {'lr': 0.0004589518403420676, 'samples': 5571072, 'steps': 10880, 'loss/train': 2.7688052654266357} 02/24/2022 15:38:52 - INFO - codeparrot_training - Step 10881: {'lr': 0.0004589428565296798, 'samples': 5571584, 'steps': 10881, 'loss/train': 1.9160237312316895} 02/24/2022 15:38:55 - INFO - codeparrot_training - Step 10882: {'lr': 0.0004589338718222473, 'samples': 5572096, 'steps': 10882, 'loss/train': 2.819997549057007} 02/24/2022 15:39:01 - INFO - codeparrot_training - Step 10883: {'lr': 0.0004589248862198083, 'samples': 5572608, 'steps': 10883, 'loss/train': 1.2545146942138672} 02/24/2022 15:39:04 - INFO - codeparrot_training - Step 10884: {'lr': 0.0004589158997224015, 'samples': 5573120, 'steps': 10884, 'loss/train': 2.1836278438568115} 02/24/2022 15:39:10 - INFO - codeparrot_training - Step 10885: {'lr': 0.0004589069123300653, 'samples': 5573632, 'steps': 10885, 'loss/train': 2.161229372024536} 02/24/2022 15:39:13 - INFO - codeparrot_training - Step 10886: {'lr': 0.0004588979240428383, 'samples': 5574144, 'steps': 10886, 'loss/train': 1.6314976215362549} 02/24/2022 15:39:19 - INFO - codeparrot_training - Step 10887: {'lr': 0.00045888893486075875, 'samples': 5574656, 'steps': 10887, 'loss/train': 1.9694976806640625} 02/24/2022 15:39:22 - INFO - codeparrot_training - Step 10888: {'lr': 0.0004588799447838655, 'samples': 5575168, 'steps': 10888, 'loss/train': 2.6182644367218018} 02/24/2022 15:39:28 - INFO - codeparrot_training - Step 10889: {'lr': 0.0004588709538121968, 'samples': 5575680, 'steps': 10889, 'loss/train': 3.0257925987243652} 02/24/2022 15:39:31 - INFO - codeparrot_training - Step 10890: {'lr': 0.00045886196194579133, 'samples': 5576192, 'steps': 10890, 'loss/train': 2.9008426666259766} 02/24/2022 15:39:37 - INFO - codeparrot_training - Step 10891: {'lr': 0.00045885296918468746, 'samples': 5576704, 'steps': 10891, 'loss/train': 2.428581714630127} 02/24/2022 15:39:40 - INFO - codeparrot_training - Step 10892: {'lr': 0.0004588439755289238, 'samples': 5577216, 'steps': 10892, 'loss/train': 2.0430517196655273} 02/24/2022 15:39:46 - INFO - codeparrot_training - Step 10893: {'lr': 0.00045883498097853894, 'samples': 5577728, 'steps': 10893, 'loss/train': 1.7443327903747559} 02/24/2022 15:39:49 - INFO - codeparrot_training - Step 10894: {'lr': 0.00045882598553357125, 'samples': 5578240, 'steps': 10894, 'loss/train': 2.746281862258911} 02/24/2022 15:39:55 - INFO - codeparrot_training - Step 10895: {'lr': 0.00045881698919405937, 'samples': 5578752, 'steps': 10895, 'loss/train': 2.3394124507904053} 02/24/2022 15:39:58 - INFO - codeparrot_training - Step 10896: {'lr': 0.00045880799196004187, 'samples': 5579264, 'steps': 10896, 'loss/train': 1.704870581626892} 02/24/2022 15:40:05 - INFO - codeparrot_training - Step 10897: {'lr': 0.00045879899383155715, 'samples': 5579776, 'steps': 10897, 'loss/train': 1.8913606405258179} 02/24/2022 15:40:08 - INFO - codeparrot_training - Step 10898: {'lr': 0.00045878999480864386, 'samples': 5580288, 'steps': 10898, 'loss/train': 2.157862663269043} 02/24/2022 15:40:14 - INFO - codeparrot_training - Step 10899: {'lr': 0.0004587809948913406, 'samples': 5580800, 'steps': 10899, 'loss/train': 1.7100781202316284} 02/24/2022 15:40:17 - INFO - codeparrot_training - Step 10900: {'lr': 0.00045877199407968577, 'samples': 5581312, 'steps': 10900, 'loss/train': 4.836899757385254} 02/24/2022 15:40:23 - INFO - codeparrot_training - Step 10901: {'lr': 0.00045876299237371807, 'samples': 5581824, 'steps': 10901, 'loss/train': 1.926213264465332} 02/24/2022 15:40:26 - INFO - codeparrot_training - Step 10902: {'lr': 0.00045875398977347596, 'samples': 5582336, 'steps': 10902, 'loss/train': 2.1030848026275635} 02/24/2022 15:40:32 - INFO - codeparrot_training - Step 10903: {'lr': 0.00045874498627899806, 'samples': 5582848, 'steps': 10903, 'loss/train': 2.2585361003875732} 02/24/2022 15:40:36 - INFO - codeparrot_training - Step 10904: {'lr': 0.00045873598189032295, 'samples': 5583360, 'steps': 10904, 'loss/train': 2.3475728034973145} 02/24/2022 15:40:41 - INFO - codeparrot_training - Step 10905: {'lr': 0.0004587269766074891, 'samples': 5583872, 'steps': 10905, 'loss/train': 1.4917577505111694} 02/24/2022 15:40:45 - INFO - codeparrot_training - Step 10906: {'lr': 0.0004587179704305353, 'samples': 5584384, 'steps': 10906, 'loss/train': 2.9434027671813965} 02/24/2022 15:40:52 - INFO - codeparrot_training - Step 10907: {'lr': 0.00045870896335949987, 'samples': 5584896, 'steps': 10907, 'loss/train': 1.5502668619155884} 02/24/2022 15:40:55 - INFO - codeparrot_training - Step 10908: {'lr': 0.00045869995539442153, 'samples': 5585408, 'steps': 10908, 'loss/train': 1.712061882019043} 02/24/2022 15:41:00 - INFO - codeparrot_training - Step 10909: {'lr': 0.0004586909465353388, 'samples': 5585920, 'steps': 10909, 'loss/train': 2.5004844665527344} 02/24/2022 15:41:06 - INFO - codeparrot_training - Step 10910: {'lr': 0.0004586819367822904, 'samples': 5586432, 'steps': 10910, 'loss/train': 2.4756505489349365} 02/24/2022 15:41:09 - INFO - codeparrot_training - Step 10911: {'lr': 0.00045867292613531484, 'samples': 5586944, 'steps': 10911, 'loss/train': 2.9871819019317627} 02/24/2022 15:41:15 - INFO - codeparrot_training - Step 10912: {'lr': 0.0004586639145944508, 'samples': 5587456, 'steps': 10912, 'loss/train': 1.3270901441574097} 02/24/2022 15:41:18 - INFO - codeparrot_training - Step 10913: {'lr': 0.0004586549021597367, 'samples': 5587968, 'steps': 10913, 'loss/train': 1.992263674736023} 02/24/2022 15:41:24 - INFO - codeparrot_training - Step 10914: {'lr': 0.00045864588883121125, 'samples': 5588480, 'steps': 10914, 'loss/train': 1.609466791152954} 02/24/2022 15:41:27 - INFO - codeparrot_training - Step 10915: {'lr': 0.00045863687460891313, 'samples': 5588992, 'steps': 10915, 'loss/train': 2.2666447162628174} 02/24/2022 15:41:34 - INFO - codeparrot_training - Step 10916: {'lr': 0.0004586278594928808, 'samples': 5589504, 'steps': 10916, 'loss/train': 2.4565460681915283} 02/24/2022 15:41:37 - INFO - codeparrot_training - Step 10917: {'lr': 0.0004586188434831531, 'samples': 5590016, 'steps': 10917, 'loss/train': 2.018866539001465} 02/24/2022 15:41:43 - INFO - codeparrot_training - Step 10918: {'lr': 0.00045860982657976835, 'samples': 5590528, 'steps': 10918, 'loss/train': 1.8322100639343262} 02/24/2022 15:41:46 - INFO - codeparrot_training - Step 10919: {'lr': 0.00045860080878276546, 'samples': 5591040, 'steps': 10919, 'loss/train': 2.1766133308410645} 02/24/2022 15:41:52 - INFO - codeparrot_training - Step 10920: {'lr': 0.0004585917900921829, 'samples': 5591552, 'steps': 10920, 'loss/train': 1.6431580781936646} 02/24/2022 15:41:55 - INFO - codeparrot_training - Step 10921: {'lr': 0.0004585827705080594, 'samples': 5592064, 'steps': 10921, 'loss/train': 0.9396345615386963} 02/24/2022 15:42:01 - INFO - codeparrot_training - Step 10922: {'lr': 0.0004585737500304335, 'samples': 5592576, 'steps': 10922, 'loss/train': 1.7167716026306152} 02/24/2022 15:42:04 - INFO - codeparrot_training - Step 10923: {'lr': 0.0004585647286593439, 'samples': 5593088, 'steps': 10923, 'loss/train': 1.7322036027908325} 02/24/2022 15:42:10 - INFO - codeparrot_training - Step 10924: {'lr': 0.0004585557063948292, 'samples': 5593600, 'steps': 10924, 'loss/train': 1.560473084449768} 02/24/2022 15:42:13 - INFO - codeparrot_training - Step 10925: {'lr': 0.00045854668323692813, 'samples': 5594112, 'steps': 10925, 'loss/train': 3.1254656314849854} 02/24/2022 15:42:19 - INFO - codeparrot_training - Step 10926: {'lr': 0.00045853765918567926, 'samples': 5594624, 'steps': 10926, 'loss/train': 2.171631336212158} 02/24/2022 15:42:22 - INFO - codeparrot_training - Step 10927: {'lr': 0.00045852863424112125, 'samples': 5595136, 'steps': 10927, 'loss/train': 2.027039051055908} 02/24/2022 15:42:28 - INFO - codeparrot_training - Step 10928: {'lr': 0.0004585196084032928, 'samples': 5595648, 'steps': 10928, 'loss/train': 2.348907470703125} 02/24/2022 15:42:31 - INFO - codeparrot_training - Step 10929: {'lr': 0.0004585105816722326, 'samples': 5596160, 'steps': 10929, 'loss/train': 1.5306440591812134} 02/24/2022 15:42:37 - INFO - codeparrot_training - Step 10930: {'lr': 0.0004585015540479792, 'samples': 5596672, 'steps': 10930, 'loss/train': 2.6781058311462402} 02/24/2022 15:42:40 - INFO - codeparrot_training - Step 10931: {'lr': 0.00045849252553057144, 'samples': 5597184, 'steps': 10931, 'loss/train': 2.7256364822387695} 02/24/2022 15:42:46 - INFO - codeparrot_training - Step 10932: {'lr': 0.00045848349612004786, 'samples': 5597696, 'steps': 10932, 'loss/train': 2.1196982860565186} 02/24/2022 15:42:50 - INFO - codeparrot_training - Step 10933: {'lr': 0.0004584744658164472, 'samples': 5598208, 'steps': 10933, 'loss/train': 1.3694515228271484} 02/24/2022 15:42:55 - INFO - codeparrot_training - Step 10934: {'lr': 0.00045846543461980805, 'samples': 5598720, 'steps': 10934, 'loss/train': 1.7228940725326538} 02/24/2022 15:42:59 - INFO - codeparrot_training - Step 10935: {'lr': 0.0004584564025301693, 'samples': 5599232, 'steps': 10935, 'loss/train': 2.2873504161834717} 02/24/2022 15:43:04 - INFO - codeparrot_training - Step 10936: {'lr': 0.00045844736954756937, 'samples': 5599744, 'steps': 10936, 'loss/train': 1.5120030641555786} 02/24/2022 15:43:08 - INFO - codeparrot_training - Step 10937: {'lr': 0.0004584383356720472, 'samples': 5600256, 'steps': 10937, 'loss/train': 1.1111712455749512} 02/24/2022 15:43:13 - INFO - codeparrot_training - Step 10938: {'lr': 0.0004584293009036414, 'samples': 5600768, 'steps': 10938, 'loss/train': 1.3232792615890503} 02/24/2022 15:43:17 - INFO - codeparrot_training - Step 10939: {'lr': 0.0004584202652423906, 'samples': 5601280, 'steps': 10939, 'loss/train': 1.0158010721206665} 02/24/2022 15:43:22 - INFO - codeparrot_training - Step 10940: {'lr': 0.0004584112286883336, 'samples': 5601792, 'steps': 10940, 'loss/train': 0.9630935788154602} 02/24/2022 15:43:26 - INFO - codeparrot_training - Step 10941: {'lr': 0.00045840219124150907, 'samples': 5602304, 'steps': 10941, 'loss/train': 2.6655266284942627} 02/24/2022 15:43:32 - INFO - codeparrot_training - Step 10942: {'lr': 0.0004583931529019557, 'samples': 5602816, 'steps': 10942, 'loss/train': 4.958943843841553} 02/24/2022 15:43:36 - INFO - codeparrot_training - Step 10943: {'lr': 0.00045838411366971225, 'samples': 5603328, 'steps': 10943, 'loss/train': 1.0271648168563843} 02/24/2022 15:43:41 - INFO - codeparrot_training - Step 10944: {'lr': 0.00045837507354481744, 'samples': 5603840, 'steps': 10944, 'loss/train': 1.8632652759552002} 02/24/2022 15:43:44 - INFO - codeparrot_training - Step 10945: {'lr': 0.00045836603252731004, 'samples': 5604352, 'steps': 10945, 'loss/train': 1.5784664154052734} 02/24/2022 15:43:50 - INFO - codeparrot_training - Step 10946: {'lr': 0.0004583569906172286, 'samples': 5604864, 'steps': 10946, 'loss/train': 2.515704393386841} 02/24/2022 15:43:53 - INFO - codeparrot_training - Step 10947: {'lr': 0.000458347947814612, 'samples': 5605376, 'steps': 10947, 'loss/train': 1.323952317237854} 02/24/2022 15:43:59 - INFO - codeparrot_training - Step 10948: {'lr': 0.00045833890411949897, 'samples': 5605888, 'steps': 10948, 'loss/train': 1.5894198417663574} 02/24/2022 15:44:03 - INFO - codeparrot_training - Step 10949: {'lr': 0.0004583298595319283, 'samples': 5606400, 'steps': 10949, 'loss/train': 1.3341354131698608} 02/24/2022 15:44:08 - INFO - codeparrot_training - Step 10950: {'lr': 0.0004583208140519386, 'samples': 5606912, 'steps': 10950, 'loss/train': 2.3943071365356445} 02/24/2022 15:44:12 - INFO - codeparrot_training - Step 10951: {'lr': 0.00045831176767956866, 'samples': 5607424, 'steps': 10951, 'loss/train': 1.355905294418335} 02/24/2022 15:44:18 - INFO - codeparrot_training - Step 10952: {'lr': 0.0004583027204148573, 'samples': 5607936, 'steps': 10952, 'loss/train': 2.6468935012817383} 02/24/2022 15:44:21 - INFO - codeparrot_training - Step 10953: {'lr': 0.00045829367225784317, 'samples': 5608448, 'steps': 10953, 'loss/train': 0.3694375455379486} 02/24/2022 15:44:27 - INFO - codeparrot_training - Step 10954: {'lr': 0.0004582846232085651, 'samples': 5608960, 'steps': 10954, 'loss/train': 3.0442774295806885} 02/24/2022 15:44:30 - INFO - codeparrot_training - Step 10955: {'lr': 0.0004582755732670619, 'samples': 5609472, 'steps': 10955, 'loss/train': 1.7182607650756836} 02/24/2022 15:44:36 - INFO - codeparrot_training - Step 10956: {'lr': 0.00045826652243337226, 'samples': 5609984, 'steps': 10956, 'loss/train': 1.9290764331817627} 02/24/2022 15:44:39 - INFO - codeparrot_training - Step 10957: {'lr': 0.0004582574707075349, 'samples': 5610496, 'steps': 10957, 'loss/train': 3.0623626708984375} 02/24/2022 15:44:45 - INFO - codeparrot_training - Step 10958: {'lr': 0.00045824841808958874, 'samples': 5611008, 'steps': 10958, 'loss/train': 2.4898617267608643} 02/24/2022 15:44:50 - INFO - codeparrot_training - Step 10959: {'lr': 0.0004582393645795725, 'samples': 5611520, 'steps': 10959, 'loss/train': 1.9221982955932617} 02/24/2022 15:44:54 - INFO - codeparrot_training - Step 10960: {'lr': 0.00045823031017752484, 'samples': 5612032, 'steps': 10960, 'loss/train': 2.3630614280700684} 02/24/2022 15:45:00 - INFO - codeparrot_training - Step 10961: {'lr': 0.00045822125488348474, 'samples': 5612544, 'steps': 10961, 'loss/train': 1.5384422540664673} 02/24/2022 15:45:03 - INFO - codeparrot_training - Step 10962: {'lr': 0.00045821219869749086, 'samples': 5613056, 'steps': 10962, 'loss/train': 2.370610237121582} 02/24/2022 15:45:09 - INFO - codeparrot_training - Step 10963: {'lr': 0.00045820314161958207, 'samples': 5613568, 'steps': 10963, 'loss/train': 1.781028389930725} 02/24/2022 15:45:12 - INFO - codeparrot_training - Step 10964: {'lr': 0.00045819408364979714, 'samples': 5614080, 'steps': 10964, 'loss/train': 0.5244473814964294} 02/24/2022 15:45:18 - INFO - codeparrot_training - Step 10965: {'lr': 0.0004581850247881749, 'samples': 5614592, 'steps': 10965, 'loss/train': 2.107774496078491} 02/24/2022 15:45:21 - INFO - codeparrot_training - Step 10966: {'lr': 0.000458175965034754, 'samples': 5615104, 'steps': 10966, 'loss/train': 3.912019729614258} 02/24/2022 15:45:27 - INFO - codeparrot_training - Step 10967: {'lr': 0.0004581669043895734, 'samples': 5615616, 'steps': 10967, 'loss/train': 2.6622848510742188} 02/24/2022 15:45:30 - INFO - codeparrot_training - Step 10968: {'lr': 0.000458157842852672, 'samples': 5616128, 'steps': 10968, 'loss/train': 2.020294427871704} 02/24/2022 15:45:36 - INFO - codeparrot_training - Step 10969: {'lr': 0.0004581487804240884, 'samples': 5616640, 'steps': 10969, 'loss/train': 1.6713902950286865} 02/24/2022 15:45:39 - INFO - codeparrot_training - Step 10970: {'lr': 0.00045813971710386147, 'samples': 5617152, 'steps': 10970, 'loss/train': 1.7464865446090698} 02/24/2022 15:45:45 - INFO - codeparrot_training - Step 10971: {'lr': 0.0004581306528920302, 'samples': 5617664, 'steps': 10971, 'loss/train': 2.3608994483947754} 02/24/2022 15:45:48 - INFO - codeparrot_training - Step 10972: {'lr': 0.0004581215877886332, 'samples': 5618176, 'steps': 10972, 'loss/train': 3.615652561187744} 02/24/2022 15:45:54 - INFO - codeparrot_training - Step 10973: {'lr': 0.0004581125217937095, 'samples': 5618688, 'steps': 10973, 'loss/train': 2.6661667823791504} 02/24/2022 15:45:57 - INFO - codeparrot_training - Step 10974: {'lr': 0.00045810345490729777, 'samples': 5619200, 'steps': 10974, 'loss/train': 1.650611400604248} 02/24/2022 15:46:03 - INFO - codeparrot_training - Step 10975: {'lr': 0.00045809438712943694, 'samples': 5619712, 'steps': 10975, 'loss/train': 1.4975214004516602} 02/24/2022 15:46:06 - INFO - codeparrot_training - Step 10976: {'lr': 0.0004580853184601659, 'samples': 5620224, 'steps': 10976, 'loss/train': 1.5695899724960327} 02/24/2022 15:46:13 - INFO - codeparrot_training - Step 10977: {'lr': 0.00045807624889952336, 'samples': 5620736, 'steps': 10977, 'loss/train': 0.6024316549301147} 02/24/2022 15:46:16 - INFO - codeparrot_training - Step 10978: {'lr': 0.0004580671784475482, 'samples': 5621248, 'steps': 10978, 'loss/train': 1.947343349456787} 02/24/2022 15:46:21 - INFO - codeparrot_training - Step 10979: {'lr': 0.0004580581071042794, 'samples': 5621760, 'steps': 10979, 'loss/train': 2.1722631454467773} 02/24/2022 15:46:25 - INFO - codeparrot_training - Step 10980: {'lr': 0.00045804903486975566, 'samples': 5622272, 'steps': 10980, 'loss/train': 2.6500258445739746} 02/24/2022 15:46:31 - INFO - codeparrot_training - Step 10981: {'lr': 0.00045803996174401595, 'samples': 5622784, 'steps': 10981, 'loss/train': 2.291271209716797} 02/24/2022 15:46:34 - INFO - codeparrot_training - Step 10982: {'lr': 0.00045803088772709914, 'samples': 5623296, 'steps': 10982, 'loss/train': 3.9501991271972656} 02/24/2022 15:46:40 - INFO - codeparrot_training - Step 10983: {'lr': 0.00045802181281904403, 'samples': 5623808, 'steps': 10983, 'loss/train': 2.3780431747436523} 02/24/2022 15:46:43 - INFO - codeparrot_training - Step 10984: {'lr': 0.00045801273701988955, 'samples': 5624320, 'steps': 10984, 'loss/train': 2.2774314880371094} 02/24/2022 15:46:49 - INFO - codeparrot_training - Step 10985: {'lr': 0.0004580036603296746, 'samples': 5624832, 'steps': 10985, 'loss/train': 2.2266483306884766} 02/24/2022 15:46:52 - INFO - codeparrot_training - Step 10986: {'lr': 0.00045799458274843786, 'samples': 5625344, 'steps': 10986, 'loss/train': 6.575760364532471} 02/24/2022 15:46:58 - INFO - codeparrot_training - Step 10987: {'lr': 0.0004579855042762185, 'samples': 5625856, 'steps': 10987, 'loss/train': 2.204773187637329} 02/24/2022 15:47:02 - INFO - codeparrot_training - Step 10988: {'lr': 0.00045797642491305523, 'samples': 5626368, 'steps': 10988, 'loss/train': 1.8324697017669678} 02/24/2022 15:47:07 - INFO - codeparrot_training - Step 10989: {'lr': 0.00045796734465898705, 'samples': 5626880, 'steps': 10989, 'loss/train': 2.0985260009765625} 02/24/2022 15:47:11 - INFO - codeparrot_training - Step 10990: {'lr': 0.00045795826351405276, 'samples': 5627392, 'steps': 10990, 'loss/train': 2.278046131134033} 02/24/2022 15:47:16 - INFO - codeparrot_training - Step 10991: {'lr': 0.00045794918147829135, 'samples': 5627904, 'steps': 10991, 'loss/train': 2.584219217300415} 02/24/2022 15:47:20 - INFO - codeparrot_training - Step 10992: {'lr': 0.00045794009855174163, 'samples': 5628416, 'steps': 10992, 'loss/train': 2.37514591217041} 02/24/2022 15:47:25 - INFO - codeparrot_training - Step 10993: {'lr': 0.0004579310147344425, 'samples': 5628928, 'steps': 10993, 'loss/train': 2.830026626586914} 02/24/2022 15:47:29 - INFO - codeparrot_training - Step 10994: {'lr': 0.000457921930026433, 'samples': 5629440, 'steps': 10994, 'loss/train': 1.923923373222351} 02/24/2022 15:47:34 - INFO - codeparrot_training - Step 10995: {'lr': 0.00045791284442775205, 'samples': 5629952, 'steps': 10995, 'loss/train': 2.3469104766845703} 02/24/2022 15:47:38 - INFO - codeparrot_training - Step 10996: {'lr': 0.0004579037579384384, 'samples': 5630464, 'steps': 10996, 'loss/train': 1.6095459461212158} 02/24/2022 15:47:45 - INFO - codeparrot_training - Step 10997: {'lr': 0.00045789467055853104, 'samples': 5630976, 'steps': 10997, 'loss/train': 2.2742464542388916} 02/24/2022 15:47:48 - INFO - codeparrot_training - Step 10998: {'lr': 0.000457885582288069, 'samples': 5631488, 'steps': 10998, 'loss/train': 1.8267676830291748} 02/24/2022 15:47:54 - INFO - codeparrot_training - Step 10999: {'lr': 0.0004578764931270911, 'samples': 5632000, 'steps': 10999, 'loss/train': 1.6772180795669556} 02/24/2022 15:47:54 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 15:48:11 - WARNING - huggingface_hub.repository - Several commits (11) will be pushed upstream. 02/24/2022 15:48:11 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 15:48:44 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy c2659cd..3d9f7da floral-grass-11 -> floral-grass-11 02/24/2022 15:48:49 - INFO - codeparrot_training - Step 11000: {'lr': 0.00045786740307563633, 'samples': 5632512, 'steps': 11000, 'loss/train': 2.46366024017334} 02/24/2022 15:48:54 - INFO - codeparrot_training - Step 11001: {'lr': 0.0004578583121337436, 'samples': 5633024, 'steps': 11001, 'loss/train': 2.297736644744873} 02/24/2022 15:48:58 - INFO - codeparrot_training - Step 11002: {'lr': 0.0004578492203014518, 'samples': 5633536, 'steps': 11002, 'loss/train': 2.6836161613464355} 02/24/2022 15:49:03 - INFO - codeparrot_training - Step 11003: {'lr': 0.00045784012757880006, 'samples': 5634048, 'steps': 11003, 'loss/train': 3.5049614906311035} 02/24/2022 15:49:07 - INFO - codeparrot_training - Step 11004: {'lr': 0.00045783103396582713, 'samples': 5634560, 'steps': 11004, 'loss/train': 2.30288028717041} 02/24/2022 15:49:13 - INFO - codeparrot_training - Step 11005: {'lr': 0.0004578219394625721, 'samples': 5635072, 'steps': 11005, 'loss/train': 2.979309558868408} 02/24/2022 15:49:16 - INFO - codeparrot_training - Step 11006: {'lr': 0.0004578128440690738, 'samples': 5635584, 'steps': 11006, 'loss/train': 1.7592554092407227} 02/24/2022 15:49:22 - INFO - codeparrot_training - Step 11007: {'lr': 0.00045780374778537134, 'samples': 5636096, 'steps': 11007, 'loss/train': 1.4693125486373901} 02/24/2022 15:49:26 - INFO - codeparrot_training - Step 11008: {'lr': 0.00045779465061150356, 'samples': 5636608, 'steps': 11008, 'loss/train': 1.9468352794647217} 02/24/2022 15:49:31 - INFO - codeparrot_training - Step 11009: {'lr': 0.0004577855525475095, 'samples': 5637120, 'steps': 11009, 'loss/train': 2.276190996170044} 02/24/2022 15:49:35 - INFO - codeparrot_training - Step 11010: {'lr': 0.0004577764535934281, 'samples': 5637632, 'steps': 11010, 'loss/train': 1.9736546277999878} 02/24/2022 15:49:40 - INFO - codeparrot_training - Step 11011: {'lr': 0.00045776735374929834, 'samples': 5638144, 'steps': 11011, 'loss/train': 2.3329107761383057} 02/24/2022 15:49:44 - INFO - codeparrot_training - Step 11012: {'lr': 0.00045775825301515923, 'samples': 5638656, 'steps': 11012, 'loss/train': 1.7059834003448486} 02/24/2022 15:49:50 - INFO - codeparrot_training - Step 11013: {'lr': 0.00045774915139104973, 'samples': 5639168, 'steps': 11013, 'loss/train': 1.292406678199768} 02/24/2022 15:49:54 - INFO - codeparrot_training - Step 11014: {'lr': 0.0004577400488770088, 'samples': 5639680, 'steps': 11014, 'loss/train': 1.446625828742981} 02/24/2022 15:49:57 - INFO - codeparrot_training - Step 11015: {'lr': 0.0004577309454730755, 'samples': 5640192, 'steps': 11015, 'loss/train': 3.102872371673584} 02/24/2022 15:50:02 - INFO - codeparrot_training - Step 11016: {'lr': 0.00045772184117928884, 'samples': 5640704, 'steps': 11016, 'loss/train': 2.154526710510254} 02/24/2022 15:50:06 - INFO - codeparrot_training - Step 11017: {'lr': 0.00045771273599568767, 'samples': 5641216, 'steps': 11017, 'loss/train': 2.4339072704315186} 02/24/2022 15:50:11 - INFO - codeparrot_training - Step 11018: {'lr': 0.0004577036299223112, 'samples': 5641728, 'steps': 11018, 'loss/train': 2.4836628437042236} 02/24/2022 15:50:15 - INFO - codeparrot_training - Step 11019: {'lr': 0.0004576945229591982, 'samples': 5642240, 'steps': 11019, 'loss/train': 2.18385648727417} 02/24/2022 15:50:21 - INFO - codeparrot_training - Step 11020: {'lr': 0.0004576854151063879, 'samples': 5642752, 'steps': 11020, 'loss/train': 1.7703357934951782} 02/24/2022 15:50:25 - INFO - codeparrot_training - Step 11021: {'lr': 0.0004576763063639192, 'samples': 5643264, 'steps': 11021, 'loss/train': 1.8063669204711914} 02/24/2022 15:50:30 - INFO - codeparrot_training - Step 11022: {'lr': 0.0004576671967318312, 'samples': 5643776, 'steps': 11022, 'loss/train': 1.3862773180007935} 02/24/2022 15:50:33 - INFO - codeparrot_training - Step 11023: {'lr': 0.0004576580862101628, 'samples': 5644288, 'steps': 11023, 'loss/train': 1.5508840084075928} 02/24/2022 15:50:39 - INFO - codeparrot_training - Step 11024: {'lr': 0.00045764897479895315, 'samples': 5644800, 'steps': 11024, 'loss/train': 2.3896539211273193} 02/24/2022 15:50:43 - INFO - codeparrot_training - Step 11025: {'lr': 0.00045763986249824126, 'samples': 5645312, 'steps': 11025, 'loss/train': 1.9844008684158325} 02/24/2022 15:50:48 - INFO - codeparrot_training - Step 11026: {'lr': 0.00045763074930806606, 'samples': 5645824, 'steps': 11026, 'loss/train': 1.9455288648605347} 02/24/2022 15:50:52 - INFO - codeparrot_training - Step 11027: {'lr': 0.0004576216352284667, 'samples': 5646336, 'steps': 11027, 'loss/train': 1.5345526933670044} 02/24/2022 15:50:57 - INFO - codeparrot_training - Step 11028: {'lr': 0.0004576125202594822, 'samples': 5646848, 'steps': 11028, 'loss/train': 4.6256422996521} 02/24/2022 15:51:01 - INFO - codeparrot_training - Step 11029: {'lr': 0.0004576034044011515, 'samples': 5647360, 'steps': 11029, 'loss/train': 1.9145762920379639} 02/24/2022 15:51:06 - INFO - codeparrot_training - Step 11030: {'lr': 0.00045759428765351377, 'samples': 5647872, 'steps': 11030, 'loss/train': 2.9236321449279785} 02/24/2022 15:51:10 - INFO - codeparrot_training - Step 11031: {'lr': 0.0004575851700166081, 'samples': 5648384, 'steps': 11031, 'loss/train': 3.1974756717681885} 02/24/2022 15:51:15 - INFO - codeparrot_training - Step 11032: {'lr': 0.0004575760514904734, 'samples': 5648896, 'steps': 11032, 'loss/train': 2.6109697818756104} 02/24/2022 15:51:19 - INFO - codeparrot_training - Step 11033: {'lr': 0.0004575669320751489, 'samples': 5649408, 'steps': 11033, 'loss/train': 1.1202114820480347} 02/24/2022 15:51:25 - INFO - codeparrot_training - Step 11034: {'lr': 0.00045755781177067345, 'samples': 5649920, 'steps': 11034, 'loss/train': 2.89687180519104} 02/24/2022 15:51:29 - INFO - codeparrot_training - Step 11035: {'lr': 0.00045754869057708635, 'samples': 5650432, 'steps': 11035, 'loss/train': 2.5690252780914307} 02/24/2022 15:51:34 - INFO - codeparrot_training - Step 11036: {'lr': 0.00045753956849442647, 'samples': 5650944, 'steps': 11036, 'loss/train': 2.1626386642456055} 02/24/2022 15:51:38 - INFO - codeparrot_training - Step 11037: {'lr': 0.00045753044552273306, 'samples': 5651456, 'steps': 11037, 'loss/train': 1.834442377090454} 02/24/2022 15:51:43 - INFO - codeparrot_training - Step 11038: {'lr': 0.0004575213216620451, 'samples': 5651968, 'steps': 11038, 'loss/train': 1.7041386365890503} 02/24/2022 15:51:47 - INFO - codeparrot_training - Step 11039: {'lr': 0.0004575121969124016, 'samples': 5652480, 'steps': 11039, 'loss/train': 2.4016833305358887} 02/24/2022 15:51:52 - INFO - codeparrot_training - Step 11040: {'lr': 0.00045750307127384186, 'samples': 5652992, 'steps': 11040, 'loss/train': 2.4795610904693604} 02/24/2022 15:51:56 - INFO - codeparrot_training - Step 11041: {'lr': 0.0004574939447464048, 'samples': 5653504, 'steps': 11041, 'loss/train': 2.1513404846191406} 02/24/2022 15:52:01 - INFO - codeparrot_training - Step 11042: {'lr': 0.0004574848173301296, 'samples': 5654016, 'steps': 11042, 'loss/train': 2.031777858734131} 02/24/2022 15:52:05 - INFO - codeparrot_training - Step 11043: {'lr': 0.0004574756890250553, 'samples': 5654528, 'steps': 11043, 'loss/train': 1.6828848123550415} 02/24/2022 15:52:11 - INFO - codeparrot_training - Step 11044: {'lr': 0.00045746655983122105, 'samples': 5655040, 'steps': 11044, 'loss/train': 2.246591567993164} 02/24/2022 15:52:14 - INFO - codeparrot_training - Step 11045: {'lr': 0.0004574574297486659, 'samples': 5655552, 'steps': 11045, 'loss/train': 3.035916805267334} 02/24/2022 15:52:20 - INFO - codeparrot_training - Step 11046: {'lr': 0.00045744829877742907, 'samples': 5656064, 'steps': 11046, 'loss/train': 1.0365140438079834} 02/24/2022 15:52:23 - INFO - codeparrot_training - Step 11047: {'lr': 0.0004574391669175495, 'samples': 5656576, 'steps': 11047, 'loss/train': 1.1585502624511719} 02/24/2022 15:52:29 - INFO - codeparrot_training - Step 11048: {'lr': 0.0004574300341690665, 'samples': 5657088, 'steps': 11048, 'loss/train': 0.404594749212265} 02/24/2022 15:52:33 - INFO - codeparrot_training - Step 11049: {'lr': 0.000457420900532019, 'samples': 5657600, 'steps': 11049, 'loss/train': 2.6901657581329346} 02/24/2022 15:52:38 - INFO - codeparrot_training - Step 11050: {'lr': 0.0004574117660064463, 'samples': 5658112, 'steps': 11050, 'loss/train': 1.9391030073165894} 02/24/2022 15:52:41 - INFO - codeparrot_training - Step 11051: {'lr': 0.0004574026305923875, 'samples': 5658624, 'steps': 11051, 'loss/train': 1.7978076934814453} 02/24/2022 15:52:47 - INFO - codeparrot_training - Step 11052: {'lr': 0.0004573934942898816, 'samples': 5659136, 'steps': 11052, 'loss/train': 1.687556505203247} 02/24/2022 15:52:50 - INFO - codeparrot_training - Step 11053: {'lr': 0.0004573843570989679, 'samples': 5659648, 'steps': 11053, 'loss/train': 2.8811450004577637} 02/24/2022 15:52:56 - INFO - codeparrot_training - Step 11054: {'lr': 0.00045737521901968535, 'samples': 5660160, 'steps': 11054, 'loss/train': 3.078970432281494} 02/24/2022 15:53:00 - INFO - codeparrot_training - Step 11055: {'lr': 0.00045736608005207327, 'samples': 5660672, 'steps': 11055, 'loss/train': 2.828577756881714} 02/24/2022 15:53:06 - INFO - codeparrot_training - Step 11056: {'lr': 0.0004573569401961708, 'samples': 5661184, 'steps': 11056, 'loss/train': 2.040673017501831} 02/24/2022 15:53:10 - INFO - codeparrot_training - Step 11057: {'lr': 0.000457347799452017, 'samples': 5661696, 'steps': 11057, 'loss/train': 1.9416203498840332} 02/24/2022 15:53:16 - INFO - codeparrot_training - Step 11058: {'lr': 0.000457338657819651, 'samples': 5662208, 'steps': 11058, 'loss/train': 2.4838335514068604} 02/24/2022 15:53:19 - INFO - codeparrot_training - Step 11059: {'lr': 0.00045732951529911216, 'samples': 5662720, 'steps': 11059, 'loss/train': 2.3747220039367676} 02/24/2022 15:53:25 - INFO - codeparrot_training - Step 11060: {'lr': 0.0004573203718904394, 'samples': 5663232, 'steps': 11060, 'loss/train': 2.607297658920288} 02/24/2022 15:53:28 - INFO - codeparrot_training - Step 11061: {'lr': 0.00045731122759367206, 'samples': 5663744, 'steps': 11061, 'loss/train': 3.3181095123291016} 02/24/2022 15:53:34 - INFO - codeparrot_training - Step 11062: {'lr': 0.00045730208240884926, 'samples': 5664256, 'steps': 11062, 'loss/train': 2.0406525135040283} 02/24/2022 15:53:37 - INFO - codeparrot_training - Step 11063: {'lr': 0.0004572929363360101, 'samples': 5664768, 'steps': 11063, 'loss/train': 2.2629809379577637} 02/24/2022 15:53:43 - INFO - codeparrot_training - Step 11064: {'lr': 0.0004572837893751939, 'samples': 5665280, 'steps': 11064, 'loss/train': 3.5304272174835205} 02/24/2022 15:53:46 - INFO - codeparrot_training - Step 11065: {'lr': 0.0004572746415264397, 'samples': 5665792, 'steps': 11065, 'loss/train': 1.5774226188659668} 02/24/2022 15:53:53 - INFO - codeparrot_training - Step 11066: {'lr': 0.0004572654927897868, 'samples': 5666304, 'steps': 11066, 'loss/train': 1.3631993532180786} 02/24/2022 15:53:56 - INFO - codeparrot_training - Step 11067: {'lr': 0.0004572563431652743, 'samples': 5666816, 'steps': 11067, 'loss/train': 2.773375988006592} 02/24/2022 15:54:02 - INFO - codeparrot_training - Step 11068: {'lr': 0.00045724719265294143, 'samples': 5667328, 'steps': 11068, 'loss/train': 2.2670059204101562} 02/24/2022 15:54:05 - INFO - codeparrot_training - Step 11069: {'lr': 0.00045723804125282744, 'samples': 5667840, 'steps': 11069, 'loss/train': 2.206634759902954} 02/24/2022 15:54:11 - INFO - codeparrot_training - Step 11070: {'lr': 0.0004572288889649715, 'samples': 5668352, 'steps': 11070, 'loss/train': 2.9405438899993896} 02/24/2022 15:54:14 - INFO - codeparrot_training - Step 11071: {'lr': 0.00045721973578941277, 'samples': 5668864, 'steps': 11071, 'loss/train': 2.217150926589966} 02/24/2022 15:54:20 - INFO - codeparrot_training - Step 11072: {'lr': 0.00045721058172619043, 'samples': 5669376, 'steps': 11072, 'loss/train': 2.799722671508789} 02/24/2022 15:54:23 - INFO - codeparrot_training - Step 11073: {'lr': 0.00045720142677534387, 'samples': 5669888, 'steps': 11073, 'loss/train': 1.9608020782470703} 02/24/2022 15:54:29 - INFO - codeparrot_training - Step 11074: {'lr': 0.00045719227093691216, 'samples': 5670400, 'steps': 11074, 'loss/train': 2.630922317504883} 02/24/2022 15:54:32 - INFO - codeparrot_training - Step 11075: {'lr': 0.0004571831142109345, 'samples': 5670912, 'steps': 11075, 'loss/train': 2.085696220397949} 02/24/2022 15:54:38 - INFO - codeparrot_training - Step 11076: {'lr': 0.0004571739565974502, 'samples': 5671424, 'steps': 11076, 'loss/train': 2.1132121086120605} 02/24/2022 15:54:41 - INFO - codeparrot_training - Step 11077: {'lr': 0.0004571647980964985, 'samples': 5671936, 'steps': 11077, 'loss/train': 3.8341281414031982} 02/24/2022 15:54:47 - INFO - codeparrot_training - Step 11078: {'lr': 0.0004571556387081185, 'samples': 5672448, 'steps': 11078, 'loss/train': 1.9625836610794067} 02/24/2022 15:54:50 - INFO - codeparrot_training - Step 11079: {'lr': 0.0004571464784323496, 'samples': 5672960, 'steps': 11079, 'loss/train': 2.8605496883392334} 02/24/2022 15:54:56 - INFO - codeparrot_training - Step 11080: {'lr': 0.0004571373172692309, 'samples': 5673472, 'steps': 11080, 'loss/train': 2.9483017921447754} 02/24/2022 15:55:00 - INFO - codeparrot_training - Step 11081: {'lr': 0.0004571281552188018, 'samples': 5673984, 'steps': 11081, 'loss/train': 1.8165431022644043} 02/24/2022 15:55:06 - INFO - codeparrot_training - Step 11082: {'lr': 0.0004571189922811013, 'samples': 5674496, 'steps': 11082, 'loss/train': 1.9531326293945312} 02/24/2022 15:55:09 - INFO - codeparrot_training - Step 11083: {'lr': 0.00045710982845616893, 'samples': 5675008, 'steps': 11083, 'loss/train': 2.7474465370178223} 02/24/2022 15:55:15 - INFO - codeparrot_training - Step 11084: {'lr': 0.0004571006637440438, 'samples': 5675520, 'steps': 11084, 'loss/train': 3.15207576751709} 02/24/2022 15:55:18 - INFO - codeparrot_training - Step 11085: {'lr': 0.00045709149814476515, 'samples': 5676032, 'steps': 11085, 'loss/train': 2.6176068782806396} 02/24/2022 15:55:24 - INFO - codeparrot_training - Step 11086: {'lr': 0.0004570823316583723, 'samples': 5676544, 'steps': 11086, 'loss/train': 1.690643548965454} 02/24/2022 15:55:27 - INFO - codeparrot_training - Step 11087: {'lr': 0.00045707316428490453, 'samples': 5677056, 'steps': 11087, 'loss/train': 1.8370248079299927} 02/24/2022 15:55:33 - INFO - codeparrot_training - Step 11088: {'lr': 0.0004570639960244011, 'samples': 5677568, 'steps': 11088, 'loss/train': 2.0845835208892822} 02/24/2022 15:55:36 - INFO - codeparrot_training - Step 11089: {'lr': 0.00045705482687690113, 'samples': 5678080, 'steps': 11089, 'loss/train': 2.016822099685669} 02/24/2022 15:55:42 - INFO - codeparrot_training - Step 11090: {'lr': 0.00045704565684244415, 'samples': 5678592, 'steps': 11090, 'loss/train': 1.3957535028457642} 02/24/2022 15:55:45 - INFO - codeparrot_training - Step 11091: {'lr': 0.0004570364859210693, 'samples': 5679104, 'steps': 11091, 'loss/train': 0.9114518761634827} 02/24/2022 15:55:51 - INFO - codeparrot_training - Step 11092: {'lr': 0.0004570273141128158, 'samples': 5679616, 'steps': 11092, 'loss/train': 1.5152301788330078} 02/24/2022 15:55:55 - INFO - codeparrot_training - Step 11093: {'lr': 0.00045701814141772313, 'samples': 5680128, 'steps': 11093, 'loss/train': 1.3387616872787476} 02/24/2022 15:56:00 - INFO - codeparrot_training - Step 11094: {'lr': 0.0004570089678358305, 'samples': 5680640, 'steps': 11094, 'loss/train': 1.7889165878295898} 02/24/2022 15:56:04 - INFO - codeparrot_training - Step 11095: {'lr': 0.000456999793367177, 'samples': 5681152, 'steps': 11095, 'loss/train': 2.937065601348877} 02/24/2022 15:56:09 - INFO - codeparrot_training - Step 11096: {'lr': 0.0004569906180118023, 'samples': 5681664, 'steps': 11096, 'loss/train': 2.175995111465454} 02/24/2022 15:56:13 - INFO - codeparrot_training - Step 11097: {'lr': 0.0004569814417697454, 'samples': 5682176, 'steps': 11097, 'loss/train': 1.1215555667877197} 02/24/2022 15:56:18 - INFO - codeparrot_training - Step 11098: {'lr': 0.0004569722646410458, 'samples': 5682688, 'steps': 11098, 'loss/train': 1.065351128578186} 02/24/2022 15:56:22 - INFO - codeparrot_training - Step 11099: {'lr': 0.0004569630866257428, 'samples': 5683200, 'steps': 11099, 'loss/train': 0.2542405128479004} 02/24/2022 15:56:28 - INFO - codeparrot_training - Step 11100: {'lr': 0.00045695390772387557, 'samples': 5683712, 'steps': 11100, 'loss/train': 1.764005184173584} 02/24/2022 15:56:31 - INFO - codeparrot_training - Step 11101: {'lr': 0.00045694472793548346, 'samples': 5684224, 'steps': 11101, 'loss/train': 2.918515682220459} 02/24/2022 15:56:37 - INFO - codeparrot_training - Step 11102: {'lr': 0.0004569355472606059, 'samples': 5684736, 'steps': 11102, 'loss/train': 2.4069817066192627} 02/24/2022 15:56:41 - INFO - codeparrot_training - Step 11103: {'lr': 0.0004569263656992822, 'samples': 5685248, 'steps': 11103, 'loss/train': 2.837353229522705} 02/24/2022 15:56:46 - INFO - codeparrot_training - Step 11104: {'lr': 0.0004569171832515516, 'samples': 5685760, 'steps': 11104, 'loss/train': 2.0978844165802} 02/24/2022 15:56:50 - INFO - codeparrot_training - Step 11105: {'lr': 0.0004569079999174536, 'samples': 5686272, 'steps': 11105, 'loss/train': 3.774681568145752} 02/24/2022 15:56:55 - INFO - codeparrot_training - Step 11106: {'lr': 0.0004568988156970273, 'samples': 5686784, 'steps': 11106, 'loss/train': 2.0071756839752197} 02/24/2022 15:56:59 - INFO - codeparrot_training - Step 11107: {'lr': 0.00045688963059031226, 'samples': 5687296, 'steps': 11107, 'loss/train': 2.8168768882751465} 02/24/2022 15:57:04 - INFO - codeparrot_training - Step 11108: {'lr': 0.00045688044459734766, 'samples': 5687808, 'steps': 11108, 'loss/train': 1.557365894317627} 02/24/2022 15:57:08 - INFO - codeparrot_training - Step 11109: {'lr': 0.00045687125771817294, 'samples': 5688320, 'steps': 11109, 'loss/train': 1.7244561910629272} 02/24/2022 15:57:13 - INFO - codeparrot_training - Step 11110: {'lr': 0.00045686206995282754, 'samples': 5688832, 'steps': 11110, 'loss/train': 2.27374529838562} 02/24/2022 15:57:19 - INFO - codeparrot_training - Step 11111: {'lr': 0.00045685288130135063, 'samples': 5689344, 'steps': 11111, 'loss/train': 2.531306743621826} 02/24/2022 15:57:23 - INFO - codeparrot_training - Step 11112: {'lr': 0.00045684369176378164, 'samples': 5689856, 'steps': 11112, 'loss/train': 0.8942400217056274} 02/24/2022 15:57:28 - INFO - codeparrot_training - Step 11113: {'lr': 0.00045683450134016, 'samples': 5690368, 'steps': 11113, 'loss/train': 1.580580472946167} 02/24/2022 15:57:32 - INFO - codeparrot_training - Step 11114: {'lr': 0.0004568253100305251, 'samples': 5690880, 'steps': 11114, 'loss/train': 2.280998706817627} 02/24/2022 15:57:37 - INFO - codeparrot_training - Step 11115: {'lr': 0.0004568161178349161, 'samples': 5691392, 'steps': 11115, 'loss/train': 1.9796950817108154} 02/24/2022 15:57:41 - INFO - codeparrot_training - Step 11116: {'lr': 0.0004568069247533726, 'samples': 5691904, 'steps': 11116, 'loss/train': 1.0214215517044067} 02/24/2022 15:57:46 - INFO - codeparrot_training - Step 11117: {'lr': 0.0004567977307859339, 'samples': 5692416, 'steps': 11117, 'loss/train': 2.094801664352417} 02/24/2022 15:57:50 - INFO - codeparrot_training - Step 11118: {'lr': 0.0004567885359326394, 'samples': 5692928, 'steps': 11118, 'loss/train': 1.8895941972732544} 02/24/2022 15:57:55 - INFO - codeparrot_training - Step 11119: {'lr': 0.00045677934019352844, 'samples': 5693440, 'steps': 11119, 'loss/train': 2.691798210144043} 02/24/2022 15:57:59 - INFO - codeparrot_training - Step 11120: {'lr': 0.00045677014356864043, 'samples': 5693952, 'steps': 11120, 'loss/train': 1.1120326519012451} 02/24/2022 15:58:04 - INFO - codeparrot_training - Step 11121: {'lr': 0.00045676094605801487, 'samples': 5694464, 'steps': 11121, 'loss/train': 1.6420798301696777} 02/24/2022 15:58:08 - INFO - codeparrot_training - Step 11122: {'lr': 0.00045675174766169105, 'samples': 5694976, 'steps': 11122, 'loss/train': 2.191995143890381} 02/24/2022 15:58:14 - INFO - codeparrot_training - Step 11123: {'lr': 0.0004567425483797083, 'samples': 5695488, 'steps': 11123, 'loss/train': 2.5102522373199463} 02/24/2022 15:58:17 - INFO - codeparrot_training - Step 11124: {'lr': 0.0004567333482121062, 'samples': 5696000, 'steps': 11124, 'loss/train': 4.087818622589111} 02/24/2022 15:58:23 - INFO - codeparrot_training - Step 11125: {'lr': 0.0004567241471589241, 'samples': 5696512, 'steps': 11125, 'loss/train': 2.909599542617798} 02/24/2022 15:58:26 - INFO - codeparrot_training - Step 11126: {'lr': 0.0004567149452202013, 'samples': 5697024, 'steps': 11126, 'loss/train': 2.11537504196167} 02/24/2022 15:58:33 - INFO - codeparrot_training - Step 11127: {'lr': 0.0004567057423959774, 'samples': 5697536, 'steps': 11127, 'loss/train': 2.41999888420105} 02/24/2022 15:58:36 - INFO - codeparrot_training - Step 11128: {'lr': 0.0004566965386862917, 'samples': 5698048, 'steps': 11128, 'loss/train': 2.5000498294830322} 02/24/2022 15:58:42 - INFO - codeparrot_training - Step 11129: {'lr': 0.0004566873340911837, 'samples': 5698560, 'steps': 11129, 'loss/train': 2.079416275024414} 02/24/2022 15:58:45 - INFO - codeparrot_training - Step 11130: {'lr': 0.00045667812861069275, 'samples': 5699072, 'steps': 11130, 'loss/train': 1.9394596815109253} 02/24/2022 15:58:51 - INFO - codeparrot_training - Step 11131: {'lr': 0.00045666892224485836, 'samples': 5699584, 'steps': 11131, 'loss/train': 3.117314338684082} 02/24/2022 15:58:54 - INFO - codeparrot_training - Step 11132: {'lr': 0.0004566597149937199, 'samples': 5700096, 'steps': 11132, 'loss/train': 4.116604328155518} 02/24/2022 15:59:00 - INFO - codeparrot_training - Step 11133: {'lr': 0.0004566505068573168, 'samples': 5700608, 'steps': 11133, 'loss/train': 3.082699775695801} 02/24/2022 15:59:03 - INFO - codeparrot_training - Step 11134: {'lr': 0.00045664129783568866, 'samples': 5701120, 'steps': 11134, 'loss/train': 1.719556212425232} 02/24/2022 15:59:09 - INFO - codeparrot_training - Step 11135: {'lr': 0.00045663208792887474, 'samples': 5701632, 'steps': 11135, 'loss/train': 0.7279012799263} 02/24/2022 15:59:12 - INFO - codeparrot_training - Step 11136: {'lr': 0.0004566228771369146, 'samples': 5702144, 'steps': 11136, 'loss/train': 2.2131645679473877} 02/24/2022 15:59:18 - INFO - codeparrot_training - Step 11137: {'lr': 0.00045661366545984763, 'samples': 5702656, 'steps': 11137, 'loss/train': 1.8204002380371094} 02/24/2022 15:59:21 - INFO - codeparrot_training - Step 11138: {'lr': 0.00045660445289771336, 'samples': 5703168, 'steps': 11138, 'loss/train': 1.9106638431549072} 02/24/2022 15:59:28 - INFO - codeparrot_training - Step 11139: {'lr': 0.00045659523945055114, 'samples': 5703680, 'steps': 11139, 'loss/train': 2.272371292114258} 02/24/2022 15:59:31 - INFO - codeparrot_training - Step 11140: {'lr': 0.0004565860251184006, 'samples': 5704192, 'steps': 11140, 'loss/train': 2.1323587894439697} 02/24/2022 15:59:36 - INFO - codeparrot_training - Step 11141: {'lr': 0.0004565768099013011, 'samples': 5704704, 'steps': 11141, 'loss/train': 1.0485512018203735} 02/24/2022 15:59:40 - INFO - codeparrot_training - Step 11142: {'lr': 0.00045656759379929213, 'samples': 5705216, 'steps': 11142, 'loss/train': 1.9648613929748535} 02/24/2022 15:59:46 - INFO - codeparrot_training - Step 11143: {'lr': 0.0004565583768124132, 'samples': 5705728, 'steps': 11143, 'loss/train': 2.54841947555542} 02/24/2022 15:59:49 - INFO - codeparrot_training - Step 11144: {'lr': 0.0004565491589407038, 'samples': 5706240, 'steps': 11144, 'loss/train': 2.3678460121154785} 02/24/2022 15:59:54 - INFO - codeparrot_training - Step 11145: {'lr': 0.0004565399401842034, 'samples': 5706752, 'steps': 11145, 'loss/train': 2.0219571590423584} 02/24/2022 15:59:58 - INFO - codeparrot_training - Step 11146: {'lr': 0.0004565307205429514, 'samples': 5707264, 'steps': 11146, 'loss/train': 1.5630725622177124} 02/24/2022 16:00:03 - INFO - codeparrot_training - Step 11147: {'lr': 0.00045652150001698744, 'samples': 5707776, 'steps': 11147, 'loss/train': 2.420351982116699} 02/24/2022 16:00:07 - INFO - codeparrot_training - Step 11148: {'lr': 0.00045651227860635094, 'samples': 5708288, 'steps': 11148, 'loss/train': 2.3975882530212402} 02/24/2022 16:00:13 - INFO - codeparrot_training - Step 11149: {'lr': 0.00045650305631108137, 'samples': 5708800, 'steps': 11149, 'loss/train': 2.3108203411102295} 02/24/2022 16:00:17 - INFO - codeparrot_training - Step 11150: {'lr': 0.0004564938331312183, 'samples': 5709312, 'steps': 11150, 'loss/train': 2.3248162269592285} 02/24/2022 16:00:22 - INFO - codeparrot_training - Step 11151: {'lr': 0.00045648460906680123, 'samples': 5709824, 'steps': 11151, 'loss/train': 2.3360297679901123} 02/24/2022 16:00:26 - INFO - codeparrot_training - Step 11152: {'lr': 0.00045647538411786964, 'samples': 5710336, 'steps': 11152, 'loss/train': 2.2024340629577637} 02/24/2022 16:00:31 - INFO - codeparrot_training - Step 11153: {'lr': 0.00045646615828446316, 'samples': 5710848, 'steps': 11153, 'loss/train': 1.7577011585235596} 02/24/2022 16:00:35 - INFO - codeparrot_training - Step 11154: {'lr': 0.00045645693156662104, 'samples': 5711360, 'steps': 11154, 'loss/train': 3.6065192222595215} 02/24/2022 16:00:40 - INFO - codeparrot_training - Step 11155: {'lr': 0.0004564477039643831, 'samples': 5711872, 'steps': 11155, 'loss/train': 1.989689826965332} 02/24/2022 16:00:44 - INFO - codeparrot_training - Step 11156: {'lr': 0.0004564384754777888, 'samples': 5712384, 'steps': 11156, 'loss/train': 0.5113958120346069} 02/24/2022 16:00:49 - INFO - codeparrot_training - Step 11157: {'lr': 0.0004564292461068775, 'samples': 5712896, 'steps': 11157, 'loss/train': 1.392565369606018} 02/24/2022 16:00:53 - INFO - codeparrot_training - Step 11158: {'lr': 0.00045642001585168896, 'samples': 5713408, 'steps': 11158, 'loss/train': 2.381349802017212} 02/24/2022 16:00:59 - INFO - codeparrot_training - Step 11159: {'lr': 0.0004564107847122626, 'samples': 5713920, 'steps': 11159, 'loss/train': 1.497516393661499} 02/24/2022 16:01:03 - INFO - codeparrot_training - Step 11160: {'lr': 0.0004564015526886379, 'samples': 5714432, 'steps': 11160, 'loss/train': 1.8312599658966064} 02/24/2022 16:01:08 - INFO - codeparrot_training - Step 11161: {'lr': 0.0004563923197808546, 'samples': 5714944, 'steps': 11161, 'loss/train': 1.6338485479354858} 02/24/2022 16:01:12 - INFO - codeparrot_training - Step 11162: {'lr': 0.00045638308598895205, 'samples': 5715456, 'steps': 11162, 'loss/train': 1.7676514387130737} 02/24/2022 16:01:17 - INFO - codeparrot_training - Step 11163: {'lr': 0.0004563738513129699, 'samples': 5715968, 'steps': 11163, 'loss/train': 2.410141706466675} 02/24/2022 16:01:21 - INFO - codeparrot_training - Step 11164: {'lr': 0.0004563646157529477, 'samples': 5716480, 'steps': 11164, 'loss/train': 2.0964887142181396} 02/24/2022 16:01:26 - INFO - codeparrot_training - Step 11165: {'lr': 0.0004563553793089251, 'samples': 5716992, 'steps': 11165, 'loss/train': 2.0609984397888184} 02/24/2022 16:01:29 - INFO - codeparrot_training - Step 11166: {'lr': 0.00045634614198094154, 'samples': 5717504, 'steps': 11166, 'loss/train': 0.6208281517028809} 02/24/2022 16:01:35 - INFO - codeparrot_training - Step 11167: {'lr': 0.0004563369037690366, 'samples': 5718016, 'steps': 11167, 'loss/train': 3.829484224319458} 02/24/2022 16:01:41 - INFO - codeparrot_training - Step 11168: {'lr': 0.00045632766467324995, 'samples': 5718528, 'steps': 11168, 'loss/train': 2.1517574787139893} 02/24/2022 16:01:44 - INFO - codeparrot_training - Step 11169: {'lr': 0.00045631842469362103, 'samples': 5719040, 'steps': 11169, 'loss/train': 2.7359609603881836} 02/24/2022 16:01:50 - INFO - codeparrot_training - Step 11170: {'lr': 0.00045630918383018947, 'samples': 5719552, 'steps': 11170, 'loss/train': 1.998534917831421} 02/24/2022 16:01:53 - INFO - codeparrot_training - Step 11171: {'lr': 0.00045629994208299496, 'samples': 5720064, 'steps': 11171, 'loss/train': 2.9150965213775635} 02/24/2022 16:01:59 - INFO - codeparrot_training - Step 11172: {'lr': 0.0004562906994520769, 'samples': 5720576, 'steps': 11172, 'loss/train': 1.0025087594985962} 02/24/2022 16:02:03 - INFO - codeparrot_training - Step 11173: {'lr': 0.0004562814559374751, 'samples': 5721088, 'steps': 11173, 'loss/train': 1.4408923387527466} 02/24/2022 16:02:08 - INFO - codeparrot_training - Step 11174: {'lr': 0.000456272211539229, 'samples': 5721600, 'steps': 11174, 'loss/train': 0.9095917344093323} 02/24/2022 16:02:12 - INFO - codeparrot_training - Step 11175: {'lr': 0.00045626296625737823, 'samples': 5722112, 'steps': 11175, 'loss/train': 1.359091877937317} 02/24/2022 16:02:17 - INFO - codeparrot_training - Step 11176: {'lr': 0.0004562537200919624, 'samples': 5722624, 'steps': 11176, 'loss/train': 1.0864464044570923} 02/24/2022 16:02:21 - INFO - codeparrot_training - Step 11177: {'lr': 0.00045624447304302117, 'samples': 5723136, 'steps': 11177, 'loss/train': 2.413727283477783} 02/24/2022 16:02:27 - INFO - codeparrot_training - Step 11178: {'lr': 0.00045623522511059405, 'samples': 5723648, 'steps': 11178, 'loss/train': 2.8752694129943848} 02/24/2022 16:02:30 - INFO - codeparrot_training - Step 11179: {'lr': 0.00045622597629472073, 'samples': 5724160, 'steps': 11179, 'loss/train': 3.9342124462127686} 02/24/2022 16:02:36 - INFO - codeparrot_training - Step 11180: {'lr': 0.0004562167265954409, 'samples': 5724672, 'steps': 11180, 'loss/train': 1.9125604629516602} 02/24/2022 16:02:39 - INFO - codeparrot_training - Step 11181: {'lr': 0.000456207476012794, 'samples': 5725184, 'steps': 11181, 'loss/train': 1.9933083057403564} 02/24/2022 16:02:45 - INFO - codeparrot_training - Step 11182: {'lr': 0.0004561982245468198, 'samples': 5725696, 'steps': 11182, 'loss/train': 2.228924512863159} 02/24/2022 16:02:48 - INFO - codeparrot_training - Step 11183: {'lr': 0.0004561889721975578, 'samples': 5726208, 'steps': 11183, 'loss/train': 0.5284889936447144} 02/24/2022 16:02:54 - INFO - codeparrot_training - Step 11184: {'lr': 0.0004561797189650478, 'samples': 5726720, 'steps': 11184, 'loss/train': 1.4588679075241089} 02/24/2022 16:02:58 - INFO - codeparrot_training - Step 11185: {'lr': 0.0004561704648493293, 'samples': 5727232, 'steps': 11185, 'loss/train': 1.699294924736023} 02/24/2022 16:03:03 - INFO - codeparrot_training - Step 11186: {'lr': 0.00045616120985044205, 'samples': 5727744, 'steps': 11186, 'loss/train': 1.9070420265197754} 02/24/2022 16:03:07 - INFO - codeparrot_training - Step 11187: {'lr': 0.0004561519539684256, 'samples': 5728256, 'steps': 11187, 'loss/train': 1.4898838996887207} 02/24/2022 16:03:12 - INFO - codeparrot_training - Step 11188: {'lr': 0.00045614269720331964, 'samples': 5728768, 'steps': 11188, 'loss/train': 1.5310871601104736} 02/24/2022 16:03:16 - INFO - codeparrot_training - Step 11189: {'lr': 0.00045613343955516386, 'samples': 5729280, 'steps': 11189, 'loss/train': 2.431586265563965} 02/24/2022 16:03:21 - INFO - codeparrot_training - Step 11190: {'lr': 0.00045612418102399785, 'samples': 5729792, 'steps': 11190, 'loss/train': 1.268957257270813} 02/24/2022 16:03:25 - INFO - codeparrot_training - Step 11191: {'lr': 0.00045611492160986127, 'samples': 5730304, 'steps': 11191, 'loss/train': 1.8090838193893433} 02/24/2022 16:03:30 - INFO - codeparrot_training - Step 11192: {'lr': 0.00045610566131279386, 'samples': 5730816, 'steps': 11192, 'loss/train': 2.358632802963257} 02/24/2022 16:03:34 - INFO - codeparrot_training - Step 11193: {'lr': 0.00045609640013283525, 'samples': 5731328, 'steps': 11193, 'loss/train': 2.8368661403656006} 02/24/2022 16:03:40 - INFO - codeparrot_training - Step 11194: {'lr': 0.00045608713807002507, 'samples': 5731840, 'steps': 11194, 'loss/train': 2.0189576148986816} 02/24/2022 16:03:44 - INFO - codeparrot_training - Step 11195: {'lr': 0.000456077875124403, 'samples': 5732352, 'steps': 11195, 'loss/train': 1.6095008850097656} 02/24/2022 16:03:49 - INFO - codeparrot_training - Step 11196: {'lr': 0.00045606861129600883, 'samples': 5732864, 'steps': 11196, 'loss/train': 1.506212592124939} 02/24/2022 16:03:53 - INFO - codeparrot_training - Step 11197: {'lr': 0.00045605934658488214, 'samples': 5733376, 'steps': 11197, 'loss/train': 1.74277925491333} 02/24/2022 16:03:58 - INFO - codeparrot_training - Step 11198: {'lr': 0.0004560500809910626, 'samples': 5733888, 'steps': 11198, 'loss/train': 2.550379753112793} 02/24/2022 16:04:02 - INFO - codeparrot_training - Step 11199: {'lr': 0.00045604081451459, 'samples': 5734400, 'steps': 11199, 'loss/train': 0.9492202997207642} 02/24/2022 16:04:07 - INFO - codeparrot_training - Step 11200: {'lr': 0.0004560315471555039, 'samples': 5734912, 'steps': 11200, 'loss/train': 2.018465280532837} 02/24/2022 16:04:11 - INFO - codeparrot_training - Step 11201: {'lr': 0.00045602227891384416, 'samples': 5735424, 'steps': 11201, 'loss/train': 1.4421931505203247} 02/24/2022 16:04:16 - INFO - codeparrot_training - Step 11202: {'lr': 0.00045601300978965033, 'samples': 5735936, 'steps': 11202, 'loss/train': 1.4803434610366821} 02/24/2022 16:04:20 - INFO - codeparrot_training - Step 11203: {'lr': 0.00045600373978296223, 'samples': 5736448, 'steps': 11203, 'loss/train': 1.64203679561615} 02/24/2022 16:04:26 - INFO - codeparrot_training - Step 11204: {'lr': 0.0004559944688938195, 'samples': 5736960, 'steps': 11204, 'loss/train': 2.1750948429107666} 02/24/2022 16:04:29 - INFO - codeparrot_training - Step 11205: {'lr': 0.0004559851971222618, 'samples': 5737472, 'steps': 11205, 'loss/train': 2.479008674621582} 02/24/2022 16:04:35 - INFO - codeparrot_training - Step 11206: {'lr': 0.00045597592446832905, 'samples': 5737984, 'steps': 11206, 'loss/train': 2.1021623611450195} 02/24/2022 16:04:39 - INFO - codeparrot_training - Step 11207: {'lr': 0.0004559666509320608, 'samples': 5738496, 'steps': 11207, 'loss/train': 2.791691780090332} 02/24/2022 16:04:44 - INFO - codeparrot_training - Step 11208: {'lr': 0.0004559573765134968, 'samples': 5739008, 'steps': 11208, 'loss/train': 2.5238149166107178} 02/24/2022 16:04:48 - INFO - codeparrot_training - Step 11209: {'lr': 0.0004559481012126768, 'samples': 5739520, 'steps': 11209, 'loss/train': 2.6910507678985596} 02/24/2022 16:04:53 - INFO - codeparrot_training - Step 11210: {'lr': 0.00045593882502964055, 'samples': 5740032, 'steps': 11210, 'loss/train': 2.5532312393188477} 02/24/2022 16:04:57 - INFO - codeparrot_training - Step 11211: {'lr': 0.00045592954796442784, 'samples': 5740544, 'steps': 11211, 'loss/train': 2.537386178970337} 02/24/2022 16:05:02 - INFO - codeparrot_training - Step 11212: {'lr': 0.0004559202700170782, 'samples': 5741056, 'steps': 11212, 'loss/train': 2.9790687561035156} 02/24/2022 16:05:06 - INFO - codeparrot_training - Step 11213: {'lr': 0.00045591099118763156, 'samples': 5741568, 'steps': 11213, 'loss/train': 2.4957096576690674} 02/24/2022 16:05:11 - INFO - codeparrot_training - Step 11214: {'lr': 0.0004559017114761276, 'samples': 5742080, 'steps': 11214, 'loss/train': 1.2813564538955688} 02/24/2022 16:05:15 - INFO - codeparrot_training - Step 11215: {'lr': 0.00045589243088260613, 'samples': 5742592, 'steps': 11215, 'loss/train': 2.0508625507354736} 02/24/2022 16:05:20 - INFO - codeparrot_training - Step 11216: {'lr': 0.00045588314940710683, 'samples': 5743104, 'steps': 11216, 'loss/train': 2.667091131210327} 02/24/2022 16:05:24 - INFO - codeparrot_training - Step 11217: {'lr': 0.00045587386704966956, 'samples': 5743616, 'steps': 11217, 'loss/train': 2.8601770401000977} 02/24/2022 16:05:29 - INFO - codeparrot_training - Step 11218: {'lr': 0.00045586458381033395, 'samples': 5744128, 'steps': 11218, 'loss/train': 1.357225775718689} 02/24/2022 16:05:33 - INFO - codeparrot_training - Step 11219: {'lr': 0.00045585529968913984, 'samples': 5744640, 'steps': 11219, 'loss/train': 2.775374174118042} 02/24/2022 16:05:39 - INFO - codeparrot_training - Step 11220: {'lr': 0.00045584601468612703, 'samples': 5745152, 'steps': 11220, 'loss/train': 2.3387062549591064} 02/24/2022 16:05:42 - INFO - codeparrot_training - Step 11221: {'lr': 0.0004558367288013352, 'samples': 5745664, 'steps': 11221, 'loss/train': 3.0921833515167236} 02/24/2022 16:05:48 - INFO - codeparrot_training - Step 11222: {'lr': 0.00045582744203480417, 'samples': 5746176, 'steps': 11222, 'loss/train': 2.494013547897339} 02/24/2022 16:05:51 - INFO - codeparrot_training - Step 11223: {'lr': 0.0004558181543865738, 'samples': 5746688, 'steps': 11223, 'loss/train': 1.687036395072937} 02/24/2022 16:05:57 - INFO - codeparrot_training - Step 11224: {'lr': 0.0004558088658566838, 'samples': 5747200, 'steps': 11224, 'loss/train': 1.4024417400360107} 02/24/2022 16:06:00 - INFO - codeparrot_training - Step 11225: {'lr': 0.000455799576445174, 'samples': 5747712, 'steps': 11225, 'loss/train': 2.1459367275238037} 02/24/2022 16:06:06 - INFO - codeparrot_training - Step 11226: {'lr': 0.00045579028615208404, 'samples': 5748224, 'steps': 11226, 'loss/train': 2.2687666416168213} 02/24/2022 16:06:09 - INFO - codeparrot_training - Step 11227: {'lr': 0.000455780994977454, 'samples': 5748736, 'steps': 11227, 'loss/train': 1.6789630651474} 02/24/2022 16:06:15 - INFO - codeparrot_training - Step 11228: {'lr': 0.0004557717029213234, 'samples': 5749248, 'steps': 11228, 'loss/train': 1.9144201278686523} 02/24/2022 16:06:18 - INFO - codeparrot_training - Step 11229: {'lr': 0.00045576240998373226, 'samples': 5749760, 'steps': 11229, 'loss/train': 2.738412618637085} 02/24/2022 16:06:25 - INFO - codeparrot_training - Step 11230: {'lr': 0.00045575311616472024, 'samples': 5750272, 'steps': 11230, 'loss/train': 1.6283036470413208} 02/24/2022 16:06:28 - INFO - codeparrot_training - Step 11231: {'lr': 0.0004557438214643272, 'samples': 5750784, 'steps': 11231, 'loss/train': 2.378791332244873} 02/24/2022 16:06:33 - INFO - codeparrot_training - Step 11232: {'lr': 0.00045573452588259296, 'samples': 5751296, 'steps': 11232, 'loss/train': 2.9709901809692383} 02/24/2022 16:06:37 - INFO - codeparrot_training - Step 11233: {'lr': 0.0004557252294195573, 'samples': 5751808, 'steps': 11233, 'loss/train': 2.266359806060791} 02/24/2022 16:06:42 - INFO - codeparrot_training - Step 11234: {'lr': 0.00045571593207526016, 'samples': 5752320, 'steps': 11234, 'loss/train': 1.540785312652588} 02/24/2022 16:06:46 - INFO - codeparrot_training - Step 11235: {'lr': 0.00045570663384974125, 'samples': 5752832, 'steps': 11235, 'loss/train': 2.662827491760254} 02/24/2022 16:06:52 - INFO - codeparrot_training - Step 11236: {'lr': 0.00045569733474304044, 'samples': 5753344, 'steps': 11236, 'loss/train': 2.371854066848755} 02/24/2022 16:06:55 - INFO - codeparrot_training - Step 11237: {'lr': 0.0004556880347551976, 'samples': 5753856, 'steps': 11237, 'loss/train': 1.1751444339752197} 02/24/2022 16:07:00 - INFO - codeparrot_training - Step 11238: {'lr': 0.0004556787338862525, 'samples': 5754368, 'steps': 11238, 'loss/train': 1.4755576848983765} 02/24/2022 16:07:04 - INFO - codeparrot_training - Step 11239: {'lr': 0.000455669432136245, 'samples': 5754880, 'steps': 11239, 'loss/train': 2.0230255126953125} 02/24/2022 16:07:11 - INFO - codeparrot_training - Step 11240: {'lr': 0.00045566012950521497, 'samples': 5755392, 'steps': 11240, 'loss/train': 2.75071120262146} 02/24/2022 16:07:14 - INFO - codeparrot_training - Step 11241: {'lr': 0.0004556508259932022, 'samples': 5755904, 'steps': 11241, 'loss/train': 3.0111143589019775} 02/24/2022 16:07:20 - INFO - codeparrot_training - Step 11242: {'lr': 0.0004556415216002467, 'samples': 5756416, 'steps': 11242, 'loss/train': 1.4160038232803345} 02/24/2022 16:07:23 - INFO - codeparrot_training - Step 11243: {'lr': 0.0004556322163263882, 'samples': 5756928, 'steps': 11243, 'loss/train': 1.7679202556610107} 02/24/2022 16:07:29 - INFO - codeparrot_training - Step 11244: {'lr': 0.00045562291017166653, 'samples': 5757440, 'steps': 11244, 'loss/train': 2.0016725063323975} 02/24/2022 16:07:32 - INFO - codeparrot_training - Step 11245: {'lr': 0.0004556136031361216, 'samples': 5757952, 'steps': 11245, 'loss/train': 2.387760877609253} 02/24/2022 16:07:38 - INFO - codeparrot_training - Step 11246: {'lr': 0.0004556042952197933, 'samples': 5758464, 'steps': 11246, 'loss/train': 1.3299553394317627} 02/24/2022 16:07:41 - INFO - codeparrot_training - Step 11247: {'lr': 0.00045559498642272153, 'samples': 5758976, 'steps': 11247, 'loss/train': 2.1121790409088135} 02/24/2022 16:07:47 - INFO - codeparrot_training - Step 11248: {'lr': 0.0004555856767449461, 'samples': 5759488, 'steps': 11248, 'loss/train': 0.45214852690696716} 02/24/2022 16:07:50 - INFO - codeparrot_training - Step 11249: {'lr': 0.00045557636618650686, 'samples': 5760000, 'steps': 11249, 'loss/train': 1.3540096282958984} 02/24/2022 16:07:56 - INFO - codeparrot_training - Step 11250: {'lr': 0.00045556705474744376, 'samples': 5760512, 'steps': 11250, 'loss/train': 1.697840929031372} 02/24/2022 16:08:00 - INFO - codeparrot_training - Step 11251: {'lr': 0.00045555774242779675, 'samples': 5761024, 'steps': 11251, 'loss/train': 3.0322370529174805} 02/24/2022 16:08:06 - INFO - codeparrot_training - Step 11252: {'lr': 0.0004555484292276055, 'samples': 5761536, 'steps': 11252, 'loss/train': 2.3741860389709473} 02/24/2022 16:08:10 - INFO - codeparrot_training - Step 11253: {'lr': 0.0004555391151469102, 'samples': 5762048, 'steps': 11253, 'loss/train': 1.6380577087402344} 02/24/2022 16:08:13 - INFO - codeparrot_training - Step 11254: {'lr': 0.00045552980018575054, 'samples': 5762560, 'steps': 11254, 'loss/train': 1.7942290306091309} 02/24/2022 16:08:21 - INFO - codeparrot_training - Step 11255: {'lr': 0.0004555204843441665, 'samples': 5763072, 'steps': 11255, 'loss/train': 1.5617128610610962} 02/24/2022 16:08:26 - INFO - codeparrot_training - Step 11256: {'lr': 0.0004555111676221979, 'samples': 5763584, 'steps': 11256, 'loss/train': 2.575984477996826} 02/24/2022 16:08:30 - INFO - codeparrot_training - Step 11257: {'lr': 0.00045550185001988475, 'samples': 5764096, 'steps': 11257, 'loss/train': 1.7319233417510986} 02/24/2022 16:08:36 - INFO - codeparrot_training - Step 11258: {'lr': 0.00045549253153726694, 'samples': 5764608, 'steps': 11258, 'loss/train': 2.1164321899414062} 02/24/2022 16:08:39 - INFO - codeparrot_training - Step 11259: {'lr': 0.00045548321217438436, 'samples': 5765120, 'steps': 11259, 'loss/train': 2.0037665367126465} 02/24/2022 16:08:45 - INFO - codeparrot_training - Step 11260: {'lr': 0.00045547389193127696, 'samples': 5765632, 'steps': 11260, 'loss/train': 2.569166898727417} 02/24/2022 16:08:48 - INFO - codeparrot_training - Step 11261: {'lr': 0.00045546457080798463, 'samples': 5766144, 'steps': 11261, 'loss/train': 8.904472351074219} 02/24/2022 16:08:54 - INFO - codeparrot_training - Step 11262: {'lr': 0.00045545524880454734, 'samples': 5766656, 'steps': 11262, 'loss/train': 1.5161077976226807} 02/24/2022 16:08:57 - INFO - codeparrot_training - Step 11263: {'lr': 0.000455445925921005, 'samples': 5767168, 'steps': 11263, 'loss/train': 2.1050312519073486} 02/24/2022 16:09:05 - INFO - codeparrot_training - Step 11264: {'lr': 0.00045543660215739755, 'samples': 5767680, 'steps': 11264, 'loss/train': 1.7605210542678833} 02/24/2022 16:09:08 - INFO - codeparrot_training - Step 11265: {'lr': 0.00045542727751376495, 'samples': 5768192, 'steps': 11265, 'loss/train': 2.0572397708892822} 02/24/2022 16:09:14 - INFO - codeparrot_training - Step 11266: {'lr': 0.00045541795199014715, 'samples': 5768704, 'steps': 11266, 'loss/train': 2.817103624343872} 02/24/2022 16:09:17 - INFO - codeparrot_training - Step 11267: {'lr': 0.00045540862558658403, 'samples': 5769216, 'steps': 11267, 'loss/train': 2.798259735107422} 02/24/2022 16:09:23 - INFO - codeparrot_training - Step 11268: {'lr': 0.00045539929830311555, 'samples': 5769728, 'steps': 11268, 'loss/train': 2.311845302581787} 02/24/2022 16:09:26 - INFO - codeparrot_training - Step 11269: {'lr': 0.00045538997013978166, 'samples': 5770240, 'steps': 11269, 'loss/train': 0.5399313569068909} 02/24/2022 16:09:32 - INFO - codeparrot_training - Step 11270: {'lr': 0.0004553806410966225, 'samples': 5770752, 'steps': 11270, 'loss/train': 3.216301202774048} 02/24/2022 16:09:35 - INFO - codeparrot_training - Step 11271: {'lr': 0.0004553713111736778, 'samples': 5771264, 'steps': 11271, 'loss/train': 1.0413765907287598} 02/24/2022 16:09:41 - INFO - codeparrot_training - Step 11272: {'lr': 0.0004553619803709876, 'samples': 5771776, 'steps': 11272, 'loss/train': 1.5031672716140747} 02/24/2022 16:09:44 - INFO - codeparrot_training - Step 11273: {'lr': 0.00045535264868859195, 'samples': 5772288, 'steps': 11273, 'loss/train': 1.1409170627593994} 02/24/2022 16:09:50 - INFO - codeparrot_training - Step 11274: {'lr': 0.0004553433161265307, 'samples': 5772800, 'steps': 11274, 'loss/train': 0.9867144823074341} 02/24/2022 16:09:53 - INFO - codeparrot_training - Step 11275: {'lr': 0.00045533398268484396, 'samples': 5773312, 'steps': 11275, 'loss/train': 2.376530408859253} 02/24/2022 16:10:00 - INFO - codeparrot_training - Step 11276: {'lr': 0.00045532464836357155, 'samples': 5773824, 'steps': 11276, 'loss/train': 1.2782824039459229} 02/24/2022 16:10:04 - INFO - codeparrot_training - Step 11277: {'lr': 0.0004553153131627536, 'samples': 5774336, 'steps': 11277, 'loss/train': 0.9035613536834717} 02/24/2022 16:10:09 - INFO - codeparrot_training - Step 11278: {'lr': 0.00045530597708243, 'samples': 5774848, 'steps': 11278, 'loss/train': 2.114393711090088} 02/24/2022 16:10:13 - INFO - codeparrot_training - Step 11279: {'lr': 0.0004552966401226408, 'samples': 5775360, 'steps': 11279, 'loss/train': 1.7024188041687012} 02/24/2022 16:10:19 - INFO - codeparrot_training - Step 11280: {'lr': 0.000455287302283426, 'samples': 5775872, 'steps': 11280, 'loss/train': 2.1052308082580566} 02/24/2022 16:10:22 - INFO - codeparrot_training - Step 11281: {'lr': 0.00045527796356482566, 'samples': 5776384, 'steps': 11281, 'loss/train': 0.893500566482544} 02/24/2022 16:10:28 - INFO - codeparrot_training - Step 11282: {'lr': 0.00045526862396687957, 'samples': 5776896, 'steps': 11282, 'loss/train': 2.396361827850342} 02/24/2022 16:10:31 - INFO - codeparrot_training - Step 11283: {'lr': 0.000455259283489628, 'samples': 5777408, 'steps': 11283, 'loss/train': 1.5983911752700806} 02/24/2022 16:10:37 - INFO - codeparrot_training - Step 11284: {'lr': 0.0004552499421331107, 'samples': 5777920, 'steps': 11284, 'loss/train': 1.8370423316955566} 02/24/2022 16:10:40 - INFO - codeparrot_training - Step 11285: {'lr': 0.0004552405998973679, 'samples': 5778432, 'steps': 11285, 'loss/train': 1.6687424182891846} 02/24/2022 16:10:48 - INFO - codeparrot_training - Step 11286: {'lr': 0.0004552312567824395, 'samples': 5778944, 'steps': 11286, 'loss/train': 2.222606658935547} 02/24/2022 16:10:51 - INFO - codeparrot_training - Step 11287: {'lr': 0.00045522191278836563, 'samples': 5779456, 'steps': 11287, 'loss/train': 1.79940664768219} 02/24/2022 16:10:57 - INFO - codeparrot_training - Step 11288: {'lr': 0.00045521256791518616, 'samples': 5779968, 'steps': 11288, 'loss/train': 2.376446485519409} 02/24/2022 16:11:00 - INFO - codeparrot_training - Step 11289: {'lr': 0.0004552032221629413, 'samples': 5780480, 'steps': 11289, 'loss/train': 1.4775240421295166} 02/24/2022 16:11:06 - INFO - codeparrot_training - Step 11290: {'lr': 0.000455193875531671, 'samples': 5780992, 'steps': 11290, 'loss/train': 2.178053855895996} 02/24/2022 16:11:09 - INFO - codeparrot_training - Step 11291: {'lr': 0.00045518452802141524, 'samples': 5781504, 'steps': 11291, 'loss/train': 1.578875184059143} 02/24/2022 16:11:15 - INFO - codeparrot_training - Step 11292: {'lr': 0.0004551751796322141, 'samples': 5782016, 'steps': 11292, 'loss/train': 2.116464614868164} 02/24/2022 16:11:18 - INFO - codeparrot_training - Step 11293: {'lr': 0.00045516583036410777, 'samples': 5782528, 'steps': 11293, 'loss/train': 2.084806442260742} 02/24/2022 16:11:24 - INFO - codeparrot_training - Step 11294: {'lr': 0.00045515648021713604, 'samples': 5783040, 'steps': 11294, 'loss/train': 1.847497582435608} 02/24/2022 16:11:27 - INFO - codeparrot_training - Step 11295: {'lr': 0.0004551471291913391, 'samples': 5783552, 'steps': 11295, 'loss/train': 2.00285267829895} 02/24/2022 16:11:34 - INFO - codeparrot_training - Step 11296: {'lr': 0.00045513777728675703, 'samples': 5784064, 'steps': 11296, 'loss/train': 1.6667442321777344} 02/24/2022 16:11:38 - INFO - codeparrot_training - Step 11297: {'lr': 0.0004551284245034298, 'samples': 5784576, 'steps': 11297, 'loss/train': 1.6538283824920654} 02/24/2022 16:11:43 - INFO - codeparrot_training - Step 11298: {'lr': 0.00045511907084139767, 'samples': 5785088, 'steps': 11298, 'loss/train': 0.9538973569869995} 02/24/2022 16:11:47 - INFO - codeparrot_training - Step 11299: {'lr': 0.0004551097163007005, 'samples': 5785600, 'steps': 11299, 'loss/train': 3.656559705734253} 02/24/2022 16:11:52 - INFO - codeparrot_training - Step 11300: {'lr': 0.0004551003608813784, 'samples': 5786112, 'steps': 11300, 'loss/train': 2.665013074874878} 02/24/2022 16:11:56 - INFO - codeparrot_training - Step 11301: {'lr': 0.00045509100458347154, 'samples': 5786624, 'steps': 11301, 'loss/train': 1.9199810028076172} 02/24/2022 16:12:01 - INFO - codeparrot_training - Step 11302: {'lr': 0.0004550816474070199, 'samples': 5787136, 'steps': 11302, 'loss/train': 2.007091760635376} 02/24/2022 16:12:05 - INFO - codeparrot_training - Step 11303: {'lr': 0.0004550722893520636, 'samples': 5787648, 'steps': 11303, 'loss/train': 1.9990875720977783} 02/24/2022 16:12:10 - INFO - codeparrot_training - Step 11304: {'lr': 0.0004550629304186428, 'samples': 5788160, 'steps': 11304, 'loss/train': 1.3781377077102661} 02/24/2022 16:12:14 - INFO - codeparrot_training - Step 11305: {'lr': 0.0004550535706067974, 'samples': 5788672, 'steps': 11305, 'loss/train': 2.8855373859405518} 02/24/2022 16:12:19 - INFO - codeparrot_training - Step 11306: {'lr': 0.0004550442099165677, 'samples': 5789184, 'steps': 11306, 'loss/train': 2.104621410369873} 02/24/2022 16:12:23 - INFO - codeparrot_training - Step 11307: {'lr': 0.0004550348483479937, 'samples': 5789696, 'steps': 11307, 'loss/train': 1.7566425800323486} 02/24/2022 16:12:28 - INFO - codeparrot_training - Step 11308: {'lr': 0.00045502548590111553, 'samples': 5790208, 'steps': 11308, 'loss/train': 2.3883728981018066} 02/24/2022 16:12:32 - INFO - codeparrot_training - Step 11309: {'lr': 0.0004550161225759732, 'samples': 5790720, 'steps': 11309, 'loss/train': 2.064016342163086} 02/24/2022 16:12:38 - INFO - codeparrot_training - Step 11310: {'lr': 0.000455006758372607, 'samples': 5791232, 'steps': 11310, 'loss/train': 2.9384281635284424} 02/24/2022 16:12:41 - INFO - codeparrot_training - Step 11311: {'lr': 0.00045499739329105696, 'samples': 5791744, 'steps': 11311, 'loss/train': 1.8259005546569824} 02/24/2022 16:12:49 - INFO - codeparrot_training - Step 11312: {'lr': 0.00045498802733136306, 'samples': 5792256, 'steps': 11312, 'loss/train': 2.3329851627349854} 02/24/2022 16:12:52 - INFO - codeparrot_training - Step 11313: {'lr': 0.00045497866049356564, 'samples': 5792768, 'steps': 11313, 'loss/train': 1.7216702699661255} 02/24/2022 16:12:56 - INFO - codeparrot_training - Step 11314: {'lr': 0.0004549692927777047, 'samples': 5793280, 'steps': 11314, 'loss/train': 2.485945224761963} 02/24/2022 16:13:01 - INFO - codeparrot_training - Step 11315: {'lr': 0.00045495992418382035, 'samples': 5793792, 'steps': 11315, 'loss/train': 2.65579891204834} 02/24/2022 16:13:04 - INFO - codeparrot_training - Step 11316: {'lr': 0.0004549505547119529, 'samples': 5794304, 'steps': 11316, 'loss/train': 2.4479236602783203} 02/24/2022 16:13:10 - INFO - codeparrot_training - Step 11317: {'lr': 0.00045494118436214225, 'samples': 5794816, 'steps': 11317, 'loss/train': 2.8074517250061035} 02/24/2022 16:13:13 - INFO - codeparrot_training - Step 11318: {'lr': 0.00045493181313442866, 'samples': 5795328, 'steps': 11318, 'loss/train': 2.9917047023773193} 02/24/2022 16:13:19 - INFO - codeparrot_training - Step 11319: {'lr': 0.00045492244102885224, 'samples': 5795840, 'steps': 11319, 'loss/train': 3.807037353515625} 02/24/2022 16:13:22 - INFO - codeparrot_training - Step 11320: {'lr': 0.00045491306804545316, 'samples': 5796352, 'steps': 11320, 'loss/train': 1.3030238151550293} 02/24/2022 16:13:28 - INFO - codeparrot_training - Step 11321: {'lr': 0.0004549036941842716, 'samples': 5796864, 'steps': 11321, 'loss/train': 2.7639851570129395} 02/24/2022 16:13:31 - INFO - codeparrot_training - Step 11322: {'lr': 0.0004548943194453476, 'samples': 5797376, 'steps': 11322, 'loss/train': 2.7611591815948486} 02/24/2022 16:13:39 - INFO - codeparrot_training - Step 11323: {'lr': 0.0004548849438287214, 'samples': 5797888, 'steps': 11323, 'loss/train': 1.9540956020355225} 02/24/2022 16:13:44 - INFO - codeparrot_training - Step 11324: {'lr': 0.00045487556733443327, 'samples': 5798400, 'steps': 11324, 'loss/train': 2.3607444763183594} 02/24/2022 16:13:48 - INFO - codeparrot_training - Step 11325: {'lr': 0.00045486618996252315, 'samples': 5798912, 'steps': 11325, 'loss/train': 2.5105535984039307} 02/24/2022 16:13:53 - INFO - codeparrot_training - Step 11326: {'lr': 0.0004548568117130314, 'samples': 5799424, 'steps': 11326, 'loss/train': 1.458544373512268} 02/24/2022 16:13:57 - INFO - codeparrot_training - Step 11327: {'lr': 0.00045484743258599803, 'samples': 5799936, 'steps': 11327, 'loss/train': 2.256856918334961} 02/24/2022 16:14:02 - INFO - codeparrot_training - Step 11328: {'lr': 0.0004548380525814634, 'samples': 5800448, 'steps': 11328, 'loss/train': 1.4750778675079346} 02/24/2022 16:14:06 - INFO - codeparrot_training - Step 11329: {'lr': 0.0004548286716994676, 'samples': 5800960, 'steps': 11329, 'loss/train': 1.6293151378631592} 02/24/2022 16:14:11 - INFO - codeparrot_training - Step 11330: {'lr': 0.0004548192899400507, 'samples': 5801472, 'steps': 11330, 'loss/train': 1.9873050451278687} 02/24/2022 16:14:15 - INFO - codeparrot_training - Step 11331: {'lr': 0.0004548099073032531, 'samples': 5801984, 'steps': 11331, 'loss/train': 2.73116397857666} 02/24/2022 16:14:22 - INFO - codeparrot_training - Step 11332: {'lr': 0.00045480052378911483, 'samples': 5802496, 'steps': 11332, 'loss/train': 1.5129947662353516} 02/24/2022 16:14:25 - INFO - codeparrot_training - Step 11333: {'lr': 0.0004547911393976762, 'samples': 5803008, 'steps': 11333, 'loss/train': 1.450227975845337} 02/24/2022 16:14:31 - INFO - codeparrot_training - Step 11334: {'lr': 0.00045478175412897733, 'samples': 5803520, 'steps': 11334, 'loss/train': 2.6027538776397705} 02/24/2022 16:14:34 - INFO - codeparrot_training - Step 11335: {'lr': 0.00045477236798305846, 'samples': 5804032, 'steps': 11335, 'loss/train': 2.1613070964813232} 02/24/2022 16:14:40 - INFO - codeparrot_training - Step 11336: {'lr': 0.00045476298095995985, 'samples': 5804544, 'steps': 11336, 'loss/train': 2.1947970390319824} 02/24/2022 16:14:43 - INFO - codeparrot_training - Step 11337: {'lr': 0.0004547535930597215, 'samples': 5805056, 'steps': 11337, 'loss/train': 1.9178993701934814} 02/24/2022 16:14:49 - INFO - codeparrot_training - Step 11338: {'lr': 0.0004547442042823839, 'samples': 5805568, 'steps': 11338, 'loss/train': 2.148998975753784} 02/24/2022 16:14:52 - INFO - codeparrot_training - Step 11339: {'lr': 0.0004547348146279871, 'samples': 5806080, 'steps': 11339, 'loss/train': 2.377134084701538} 02/24/2022 16:14:58 - INFO - codeparrot_training - Step 11340: {'lr': 0.00045472542409657135, 'samples': 5806592, 'steps': 11340, 'loss/train': 2.880746603012085} 02/24/2022 16:15:01 - INFO - codeparrot_training - Step 11341: {'lr': 0.00045471603268817696, 'samples': 5807104, 'steps': 11341, 'loss/train': 2.8682913780212402} 02/24/2022 16:15:07 - INFO - codeparrot_training - Step 11342: {'lr': 0.000454706640402844, 'samples': 5807616, 'steps': 11342, 'loss/train': 2.337826728820801} 02/24/2022 16:15:10 - INFO - codeparrot_training - Step 11343: {'lr': 0.00045469724724061286, 'samples': 5808128, 'steps': 11343, 'loss/train': 2.1685545444488525} 02/24/2022 16:15:16 - INFO - codeparrot_training - Step 11344: {'lr': 0.0004546878532015236, 'samples': 5808640, 'steps': 11344, 'loss/train': 2.416700601577759} 02/24/2022 16:15:19 - INFO - codeparrot_training - Step 11345: {'lr': 0.00045467845828561673, 'samples': 5809152, 'steps': 11345, 'loss/train': 1.496660590171814} 02/24/2022 16:15:27 - INFO - codeparrot_training - Step 11346: {'lr': 0.0004546690624929322, 'samples': 5809664, 'steps': 11346, 'loss/train': 1.6992676258087158} 02/24/2022 16:15:30 - INFO - codeparrot_training - Step 11347: {'lr': 0.0004546596658235105, 'samples': 5810176, 'steps': 11347, 'loss/train': 2.159672737121582} 02/24/2022 16:15:36 - INFO - codeparrot_training - Step 11348: {'lr': 0.00045465026827739175, 'samples': 5810688, 'steps': 11348, 'loss/train': 1.746140956878662} 02/24/2022 16:15:39 - INFO - codeparrot_training - Step 11349: {'lr': 0.00045464086985461615, 'samples': 5811200, 'steps': 11349, 'loss/train': 1.85089910030365} 02/24/2022 16:15:44 - INFO - codeparrot_training - Step 11350: {'lr': 0.0004546314705552241, 'samples': 5811712, 'steps': 11350, 'loss/train': 2.0434978008270264} 02/24/2022 16:15:48 - INFO - codeparrot_training - Step 11351: {'lr': 0.00045462207037925593, 'samples': 5812224, 'steps': 11351, 'loss/train': 2.6612701416015625} 02/24/2022 16:15:54 - INFO - codeparrot_training - Step 11352: {'lr': 0.0004546126693267516, 'samples': 5812736, 'steps': 11352, 'loss/train': 0.9742980003356934} 02/24/2022 16:15:57 - INFO - codeparrot_training - Step 11353: {'lr': 0.0004546032673977517, 'samples': 5813248, 'steps': 11353, 'loss/train': 7.829508304595947} 02/24/2022 16:16:03 - INFO - codeparrot_training - Step 11354: {'lr': 0.0004545938645922963, 'samples': 5813760, 'steps': 11354, 'loss/train': 2.8821663856506348} 02/24/2022 16:16:06 - INFO - codeparrot_training - Step 11355: {'lr': 0.0004545844609104258, 'samples': 5814272, 'steps': 11355, 'loss/train': 1.7871776819229126} 02/24/2022 16:16:10 - INFO - codeparrot_training - Step 11356: {'lr': 0.0004545750563521804, 'samples': 5814784, 'steps': 11356, 'loss/train': 2.7154881954193115} 02/24/2022 16:16:16 - INFO - codeparrot_training - Step 11357: {'lr': 0.0004545656509176004, 'samples': 5815296, 'steps': 11357, 'loss/train': 2.296501636505127} 02/24/2022 16:16:23 - INFO - codeparrot_training - Step 11358: {'lr': 0.0004545562446067261, 'samples': 5815808, 'steps': 11358, 'loss/train': 0.6772823929786682} 02/24/2022 16:16:26 - INFO - codeparrot_training - Step 11359: {'lr': 0.00045454683741959787, 'samples': 5816320, 'steps': 11359, 'loss/train': 2.5950610637664795} 02/24/2022 16:16:32 - INFO - codeparrot_training - Step 11360: {'lr': 0.0004545374293562559, 'samples': 5816832, 'steps': 11360, 'loss/train': 1.4573802947998047} 02/24/2022 16:16:35 - INFO - codeparrot_training - Step 11361: {'lr': 0.00045452802041674045, 'samples': 5817344, 'steps': 11361, 'loss/train': 0.7797079682350159} 02/24/2022 16:16:41 - INFO - codeparrot_training - Step 11362: {'lr': 0.000454518610601092, 'samples': 5817856, 'steps': 11362, 'loss/train': 0.794467031955719} 02/24/2022 16:16:44 - INFO - codeparrot_training - Step 11363: {'lr': 0.0004545091999093508, 'samples': 5818368, 'steps': 11363, 'loss/train': 2.0298173427581787} 02/24/2022 16:16:50 - INFO - codeparrot_training - Step 11364: {'lr': 0.00045449978834155705, 'samples': 5818880, 'steps': 11364, 'loss/train': 1.0094642639160156} 02/24/2022 16:16:53 - INFO - codeparrot_training - Step 11365: {'lr': 0.00045449037589775123, 'samples': 5819392, 'steps': 11365, 'loss/train': 1.0958517789840698} 02/24/2022 16:16:59 - INFO - codeparrot_training - Step 11366: {'lr': 0.00045448096257797344, 'samples': 5819904, 'steps': 11366, 'loss/train': 2.5343425273895264} 02/24/2022 16:17:02 - INFO - codeparrot_training - Step 11367: {'lr': 0.0004544715483822642, 'samples': 5820416, 'steps': 11367, 'loss/train': 1.2284151315689087} 02/24/2022 16:17:09 - INFO - codeparrot_training - Step 11368: {'lr': 0.00045446213331066376, 'samples': 5820928, 'steps': 11368, 'loss/train': 1.1153687238693237} 02/24/2022 16:17:13 - INFO - codeparrot_training - Step 11369: {'lr': 0.0004544527173632125, 'samples': 5821440, 'steps': 11369, 'loss/train': 2.504488468170166} 02/24/2022 16:17:18 - INFO - codeparrot_training - Step 11370: {'lr': 0.00045444330053995074, 'samples': 5821952, 'steps': 11370, 'loss/train': 3.0068347454071045} 02/24/2022 16:17:22 - INFO - codeparrot_training - Step 11371: {'lr': 0.00045443388284091877, 'samples': 5822464, 'steps': 11371, 'loss/train': 1.8943393230438232} 02/24/2022 16:17:27 - INFO - codeparrot_training - Step 11372: {'lr': 0.0004544244642661569, 'samples': 5822976, 'steps': 11372, 'loss/train': 2.1895809173583984} 02/24/2022 16:17:31 - INFO - codeparrot_training - Step 11373: {'lr': 0.0004544150448157056, 'samples': 5823488, 'steps': 11373, 'loss/train': 1.8967657089233398} 02/24/2022 16:17:36 - INFO - codeparrot_training - Step 11374: {'lr': 0.0004544056244896052, 'samples': 5824000, 'steps': 11374, 'loss/train': 2.8636271953582764} 02/24/2022 16:17:40 - INFO - codeparrot_training - Step 11375: {'lr': 0.00045439620328789593, 'samples': 5824512, 'steps': 11375, 'loss/train': 1.1131142377853394} 02/24/2022 16:17:45 - INFO - codeparrot_training - Step 11376: {'lr': 0.00045438678121061826, 'samples': 5825024, 'steps': 11376, 'loss/train': 1.6215119361877441} 02/24/2022 16:17:49 - INFO - codeparrot_training - Step 11377: {'lr': 0.0004543773582578125, 'samples': 5825536, 'steps': 11377, 'loss/train': 1.6648873090744019} 02/24/2022 16:17:56 - INFO - codeparrot_training - Step 11378: {'lr': 0.00045436793442951907, 'samples': 5826048, 'steps': 11378, 'loss/train': 1.9659498929977417} 02/24/2022 16:18:00 - INFO - codeparrot_training - Step 11379: {'lr': 0.0004543585097257783, 'samples': 5826560, 'steps': 11379, 'loss/train': 2.6243858337402344} 02/24/2022 16:18:05 - INFO - codeparrot_training - Step 11380: {'lr': 0.0004543490841466306, 'samples': 5827072, 'steps': 11380, 'loss/train': 2.2534351348876953} 02/24/2022 16:18:09 - INFO - codeparrot_training - Step 11381: {'lr': 0.00045433965769211616, 'samples': 5827584, 'steps': 11381, 'loss/train': 3.113907814025879} 02/24/2022 16:18:14 - INFO - codeparrot_training - Step 11382: {'lr': 0.00045433023036227566, 'samples': 5828096, 'steps': 11382, 'loss/train': 2.194993257522583} 02/24/2022 16:18:18 - INFO - codeparrot_training - Step 11383: {'lr': 0.00045432080215714927, 'samples': 5828608, 'steps': 11383, 'loss/train': 2.3546202182769775} 02/24/2022 16:18:23 - INFO - codeparrot_training - Step 11384: {'lr': 0.00045431137307677753, 'samples': 5829120, 'steps': 11384, 'loss/train': 1.6404659748077393} 02/24/2022 16:18:27 - INFO - codeparrot_training - Step 11385: {'lr': 0.00045430194312120066, 'samples': 5829632, 'steps': 11385, 'loss/train': 2.023556709289551} 02/24/2022 16:18:32 - INFO - codeparrot_training - Step 11386: {'lr': 0.0004542925122904591, 'samples': 5830144, 'steps': 11386, 'loss/train': 2.4634549617767334} 02/24/2022 16:18:36 - INFO - codeparrot_training - Step 11387: {'lr': 0.00045428308058459335, 'samples': 5830656, 'steps': 11387, 'loss/train': 2.151942014694214} 02/24/2022 16:18:43 - INFO - codeparrot_training - Step 11388: {'lr': 0.00045427364800364374, 'samples': 5831168, 'steps': 11388, 'loss/train': 2.8610587120056152} 02/24/2022 16:18:46 - INFO - codeparrot_training - Step 11389: {'lr': 0.00045426421454765065, 'samples': 5831680, 'steps': 11389, 'loss/train': 0.571302592754364} 02/24/2022 16:18:52 - INFO - codeparrot_training - Step 11390: {'lr': 0.0004542547802166546, 'samples': 5832192, 'steps': 11390, 'loss/train': 1.377107858657837} 02/24/2022 16:18:55 - INFO - codeparrot_training - Step 11391: {'lr': 0.00045424534501069594, 'samples': 5832704, 'steps': 11391, 'loss/train': 2.3089232444763184} 02/24/2022 16:19:01 - INFO - codeparrot_training - Step 11392: {'lr': 0.00045423590892981503, 'samples': 5833216, 'steps': 11392, 'loss/train': 2.594899892807007} 02/24/2022 16:19:04 - INFO - codeparrot_training - Step 11393: {'lr': 0.0004542264719740523, 'samples': 5833728, 'steps': 11393, 'loss/train': 0.6809685826301575} 02/24/2022 16:19:10 - INFO - codeparrot_training - Step 11394: {'lr': 0.0004542170341434483, 'samples': 5834240, 'steps': 11394, 'loss/train': 1.1828583478927612} 02/24/2022 16:19:13 - INFO - codeparrot_training - Step 11395: {'lr': 0.00045420759543804326, 'samples': 5834752, 'steps': 11395, 'loss/train': 1.4844894409179688} 02/24/2022 16:19:19 - INFO - codeparrot_training - Step 11396: {'lr': 0.0004541981558578778, 'samples': 5835264, 'steps': 11396, 'loss/train': 2.7818479537963867} 02/24/2022 16:19:22 - INFO - codeparrot_training - Step 11397: {'lr': 0.0004541887154029922, 'samples': 5835776, 'steps': 11397, 'loss/train': 1.0714787244796753} 02/24/2022 16:19:28 - INFO - codeparrot_training - Step 11398: {'lr': 0.0004541792740734271, 'samples': 5836288, 'steps': 11398, 'loss/train': 1.311425805091858} 02/24/2022 16:19:32 - INFO - codeparrot_training - Step 11399: {'lr': 0.0004541698318692228, 'samples': 5836800, 'steps': 11399, 'loss/train': 2.1383049488067627} 02/24/2022 16:19:37 - INFO - codeparrot_training - Step 11400: {'lr': 0.0004541603887904198, 'samples': 5837312, 'steps': 11400, 'loss/train': 1.4966486692428589} 02/24/2022 16:19:41 - INFO - codeparrot_training - Step 11401: {'lr': 0.0004541509448370584, 'samples': 5837824, 'steps': 11401, 'loss/train': 1.64617121219635} 02/24/2022 16:19:46 - INFO - codeparrot_training - Step 11402: {'lr': 0.00045414150000917927, 'samples': 5838336, 'steps': 11402, 'loss/train': 2.2392823696136475} 02/24/2022 16:19:50 - INFO - codeparrot_training - Step 11403: {'lr': 0.0004541320543068227, 'samples': 5838848, 'steps': 11403, 'loss/train': 2.1169583797454834} 02/24/2022 16:19:55 - INFO - codeparrot_training - Step 11404: {'lr': 0.00045412260773002933, 'samples': 5839360, 'steps': 11404, 'loss/train': 2.5532174110412598} 02/24/2022 16:19:58 - INFO - codeparrot_training - Step 11405: {'lr': 0.0004541131602788395, 'samples': 5839872, 'steps': 11405, 'loss/train': 1.873200535774231} 02/24/2022 16:20:06 - INFO - codeparrot_training - Step 11406: {'lr': 0.00045410371195329365, 'samples': 5840384, 'steps': 11406, 'loss/train': 2.221608877182007} 02/24/2022 16:20:09 - INFO - codeparrot_training - Step 11407: {'lr': 0.00045409426275343234, 'samples': 5840896, 'steps': 11407, 'loss/train': 3.276916027069092} 02/24/2022 16:20:15 - INFO - codeparrot_training - Step 11408: {'lr': 0.00045408481267929604, 'samples': 5841408, 'steps': 11408, 'loss/train': 1.904630184173584} 02/24/2022 16:20:20 - INFO - codeparrot_training - Step 11409: {'lr': 0.0004540753617309251, 'samples': 5841920, 'steps': 11409, 'loss/train': 2.037550449371338} 02/24/2022 16:20:24 - INFO - codeparrot_training - Step 11410: {'lr': 0.0004540659099083602, 'samples': 5842432, 'steps': 11410, 'loss/train': 0.9711087942123413} 02/24/2022 16:20:29 - INFO - codeparrot_training - Step 11411: {'lr': 0.0004540564572116418, 'samples': 5842944, 'steps': 11411, 'loss/train': 1.0147775411605835} 02/24/2022 16:20:33 - INFO - codeparrot_training - Step 11412: {'lr': 0.0004540470036408102, 'samples': 5843456, 'steps': 11412, 'loss/train': 2.869175672531128} 02/24/2022 16:20:40 - INFO - codeparrot_training - Step 11413: {'lr': 0.0004540375491959061, 'samples': 5843968, 'steps': 11413, 'loss/train': 3.204055070877075} 02/24/2022 16:20:44 - INFO - codeparrot_training - Step 11414: {'lr': 0.00045402809387697, 'samples': 5844480, 'steps': 11414, 'loss/train': 2.127800941467285} 02/24/2022 16:20:49 - INFO - codeparrot_training - Step 11415: {'lr': 0.00045401863768404217, 'samples': 5844992, 'steps': 11415, 'loss/train': 1.7463570833206177} 02/24/2022 16:20:53 - INFO - codeparrot_training - Step 11416: {'lr': 0.0004540091806171634, 'samples': 5845504, 'steps': 11416, 'loss/train': 3.21770977973938} 02/24/2022 16:20:58 - INFO - codeparrot_training - Step 11417: {'lr': 0.000453999722676374, 'samples': 5846016, 'steps': 11417, 'loss/train': 2.4473228454589844} 02/24/2022 16:21:02 - INFO - codeparrot_training - Step 11418: {'lr': 0.0004539902638617146, 'samples': 5846528, 'steps': 11418, 'loss/train': 2.504650831222534} 02/24/2022 16:21:07 - INFO - codeparrot_training - Step 11419: {'lr': 0.0004539808041732257, 'samples': 5847040, 'steps': 11419, 'loss/train': 0.8776189684867859} 02/24/2022 16:21:11 - INFO - codeparrot_training - Step 11420: {'lr': 0.0004539713436109478, 'samples': 5847552, 'steps': 11420, 'loss/train': 2.253885269165039} 02/24/2022 16:21:16 - INFO - codeparrot_training - Step 11421: {'lr': 0.00045396188217492145, 'samples': 5848064, 'steps': 11421, 'loss/train': 2.1496660709381104} 02/24/2022 16:21:20 - INFO - codeparrot_training - Step 11422: {'lr': 0.00045395241986518714, 'samples': 5848576, 'steps': 11422, 'loss/train': 1.4651926755905151} 02/24/2022 16:21:27 - INFO - codeparrot_training - Step 11423: {'lr': 0.0004539429566817854, 'samples': 5849088, 'steps': 11423, 'loss/train': 2.836231231689453} 02/24/2022 16:21:30 - INFO - codeparrot_training - Step 11424: {'lr': 0.00045393349262475686, 'samples': 5849600, 'steps': 11424, 'loss/train': 1.9236829280853271} 02/24/2022 16:21:36 - INFO - codeparrot_training - Step 11425: {'lr': 0.000453924027694142, 'samples': 5850112, 'steps': 11425, 'loss/train': 1.5504837036132812} 02/24/2022 16:21:40 - INFO - codeparrot_training - Step 11426: {'lr': 0.00045391456188998124, 'samples': 5850624, 'steps': 11426, 'loss/train': 2.539696455001831} 02/24/2022 16:21:45 - INFO - codeparrot_training - Step 11427: {'lr': 0.00045390509521231535, 'samples': 5851136, 'steps': 11427, 'loss/train': 3.041769027709961} 02/24/2022 16:21:49 - INFO - codeparrot_training - Step 11428: {'lr': 0.00045389562766118475, 'samples': 5851648, 'steps': 11428, 'loss/train': 2.6716833114624023} 02/24/2022 16:21:54 - INFO - codeparrot_training - Step 11429: {'lr': 0.00045388615923663004, 'samples': 5852160, 'steps': 11429, 'loss/train': 1.4363834857940674} 02/24/2022 16:21:58 - INFO - codeparrot_training - Step 11430: {'lr': 0.0004538766899386917, 'samples': 5852672, 'steps': 11430, 'loss/train': 2.0188353061676025} 02/24/2022 16:22:03 - INFO - codeparrot_training - Step 11431: {'lr': 0.00045386721976741043, 'samples': 5853184, 'steps': 11431, 'loss/train': 2.653979778289795} 02/24/2022 16:22:07 - INFO - codeparrot_training - Step 11432: {'lr': 0.0004538577487228267, 'samples': 5853696, 'steps': 11432, 'loss/train': 2.52278470993042} 02/24/2022 16:22:12 - INFO - codeparrot_training - Step 11433: {'lr': 0.0004538482768049811, 'samples': 5854208, 'steps': 11433, 'loss/train': 1.777289628982544} 02/24/2022 16:22:16 - INFO - codeparrot_training - Step 11434: {'lr': 0.00045383880401391423, 'samples': 5854720, 'steps': 11434, 'loss/train': 2.0804903507232666} 02/24/2022 16:22:23 - INFO - codeparrot_training - Step 11435: {'lr': 0.00045382933034966667, 'samples': 5855232, 'steps': 11435, 'loss/train': 2.8244993686676025} 02/24/2022 16:22:26 - INFO - codeparrot_training - Step 11436: {'lr': 0.0004538198558122789, 'samples': 5855744, 'steps': 11436, 'loss/train': 2.139791250228882} 02/24/2022 16:22:32 - INFO - codeparrot_training - Step 11437: {'lr': 0.0004538103804017917, 'samples': 5856256, 'steps': 11437, 'loss/train': 1.7734744548797607} 02/24/2022 16:22:35 - INFO - codeparrot_training - Step 11438: {'lr': 0.00045380090411824547, 'samples': 5856768, 'steps': 11438, 'loss/train': 1.9485567808151245} 02/24/2022 16:22:41 - INFO - codeparrot_training - Step 11439: {'lr': 0.0004537914269616809, 'samples': 5857280, 'steps': 11439, 'loss/train': 2.719700574874878} 02/24/2022 16:22:44 - INFO - codeparrot_training - Step 11440: {'lr': 0.00045378194893213854, 'samples': 5857792, 'steps': 11440, 'loss/train': 1.1889444589614868} 02/24/2022 16:22:50 - INFO - codeparrot_training - Step 11441: {'lr': 0.00045377247002965904, 'samples': 5858304, 'steps': 11441, 'loss/train': 1.3447911739349365} 02/24/2022 16:22:53 - INFO - codeparrot_training - Step 11442: {'lr': 0.000453762990254283, 'samples': 5858816, 'steps': 11442, 'loss/train': 1.1948394775390625} 02/24/2022 16:22:59 - INFO - codeparrot_training - Step 11443: {'lr': 0.000453753509606051, 'samples': 5859328, 'steps': 11443, 'loss/train': 0.699555516242981} 02/24/2022 16:23:02 - INFO - codeparrot_training - Step 11444: {'lr': 0.0004537440280850037, 'samples': 5859840, 'steps': 11444, 'loss/train': 1.7442504167556763} 02/24/2022 16:23:10 - INFO - codeparrot_training - Step 11445: {'lr': 0.00045373454569118166, 'samples': 5860352, 'steps': 11445, 'loss/train': 3.653651475906372} 02/24/2022 16:23:13 - INFO - codeparrot_training - Step 11446: {'lr': 0.0004537250624246255, 'samples': 5860864, 'steps': 11446, 'loss/train': 1.8762763738632202} 02/24/2022 16:23:19 - INFO - codeparrot_training - Step 11447: {'lr': 0.00045371557828537585, 'samples': 5861376, 'steps': 11447, 'loss/train': 2.6544928550720215} 02/24/2022 16:23:22 - INFO - codeparrot_training - Step 11448: {'lr': 0.0004537060932734733, 'samples': 5861888, 'steps': 11448, 'loss/train': 2.730875015258789} 02/24/2022 16:23:28 - INFO - codeparrot_training - Step 11449: {'lr': 0.0004536966073889587, 'samples': 5862400, 'steps': 11449, 'loss/train': 2.141685724258423} 02/24/2022 16:23:31 - INFO - codeparrot_training - Step 11450: {'lr': 0.00045368712063187237, 'samples': 5862912, 'steps': 11450, 'loss/train': 1.514075756072998} 02/24/2022 16:23:37 - INFO - codeparrot_training - Step 11451: {'lr': 0.0004536776330022552, 'samples': 5863424, 'steps': 11451, 'loss/train': 2.724309206008911} 02/24/2022 16:23:40 - INFO - codeparrot_training - Step 11452: {'lr': 0.0004536681445001476, 'samples': 5863936, 'steps': 11452, 'loss/train': 1.3461116552352905} 02/24/2022 16:23:45 - INFO - codeparrot_training - Step 11453: {'lr': 0.0004536586551255904, 'samples': 5864448, 'steps': 11453, 'loss/train': 0.9570274353027344} 02/24/2022 16:23:49 - INFO - codeparrot_training - Step 11454: {'lr': 0.0004536491648786242, 'samples': 5864960, 'steps': 11454, 'loss/train': 2.104924201965332} 02/24/2022 16:23:55 - INFO - codeparrot_training - Step 11455: {'lr': 0.0004536396737592896, 'samples': 5865472, 'steps': 11455, 'loss/train': 1.296418309211731} 02/24/2022 16:23:58 - INFO - codeparrot_training - Step 11456: {'lr': 0.0004536301817676274, 'samples': 5865984, 'steps': 11456, 'loss/train': 1.55818510055542} 02/24/2022 16:24:04 - INFO - codeparrot_training - Step 11457: {'lr': 0.00045362068890367804, 'samples': 5866496, 'steps': 11457, 'loss/train': 0.9206820726394653} 02/24/2022 16:24:07 - INFO - codeparrot_training - Step 11458: {'lr': 0.0004536111951674824, 'samples': 5867008, 'steps': 11458, 'loss/train': 2.4314370155334473} 02/24/2022 16:24:12 - INFO - codeparrot_training - Step 11459: {'lr': 0.000453601700559081, 'samples': 5867520, 'steps': 11459, 'loss/train': 2.436211347579956} 02/24/2022 16:24:16 - INFO - codeparrot_training - Step 11460: {'lr': 0.00045359220507851456, 'samples': 5868032, 'steps': 11460, 'loss/train': 1.680085301399231} 02/24/2022 16:24:23 - INFO - codeparrot_training - Step 11461: {'lr': 0.0004535827087258238, 'samples': 5868544, 'steps': 11461, 'loss/train': 3.255981922149658} 02/24/2022 16:24:29 - INFO - codeparrot_training - Step 11462: {'lr': 0.00045357321150104934, 'samples': 5869056, 'steps': 11462, 'loss/train': 1.8159868717193604} 02/24/2022 16:24:32 - INFO - codeparrot_training - Step 11463: {'lr': 0.0004535637134042319, 'samples': 5869568, 'steps': 11463, 'loss/train': 1.8969529867172241} 02/24/2022 16:24:38 - INFO - codeparrot_training - Step 11464: {'lr': 0.00045355421443541214, 'samples': 5870080, 'steps': 11464, 'loss/train': 1.7035434246063232} 02/24/2022 16:24:41 - INFO - codeparrot_training - Step 11465: {'lr': 0.00045354471459463076, 'samples': 5870592, 'steps': 11465, 'loss/train': 2.244917869567871} 02/24/2022 16:24:47 - INFO - codeparrot_training - Step 11466: {'lr': 0.0004535352138819284, 'samples': 5871104, 'steps': 11466, 'loss/train': 1.6521811485290527} 02/24/2022 16:24:50 - INFO - codeparrot_training - Step 11467: {'lr': 0.0004535257122973459, 'samples': 5871616, 'steps': 11467, 'loss/train': 1.382369875907898} 02/24/2022 16:24:56 - INFO - codeparrot_training - Step 11468: {'lr': 0.0004535162098409238, 'samples': 5872128, 'steps': 11468, 'loss/train': 1.3611055612564087} 02/24/2022 16:24:59 - INFO - codeparrot_training - Step 11469: {'lr': 0.000453506706512703, 'samples': 5872640, 'steps': 11469, 'loss/train': 2.5269525051116943} 02/24/2022 16:25:07 - INFO - codeparrot_training - Step 11470: {'lr': 0.00045349720231272395, 'samples': 5873152, 'steps': 11470, 'loss/train': 1.9398821592330933} 02/24/2022 16:25:10 - INFO - codeparrot_training - Step 11471: {'lr': 0.0004534876972410276, 'samples': 5873664, 'steps': 11471, 'loss/train': 2.157519817352295} 02/24/2022 16:25:16 - INFO - codeparrot_training - Step 11472: {'lr': 0.0004534781912976545, 'samples': 5874176, 'steps': 11472, 'loss/train': 2.940617084503174} 02/24/2022 16:25:19 - INFO - codeparrot_training - Step 11473: {'lr': 0.00045346868448264553, 'samples': 5874688, 'steps': 11473, 'loss/train': 2.435476779937744} 02/24/2022 16:25:25 - INFO - codeparrot_training - Step 11474: {'lr': 0.00045345917679604126, 'samples': 5875200, 'steps': 11474, 'loss/train': 2.280003309249878} 02/24/2022 16:25:28 - INFO - codeparrot_training - Step 11475: {'lr': 0.0004534496682378825, 'samples': 5875712, 'steps': 11475, 'loss/train': 1.8867299556732178} 02/24/2022 16:25:34 - INFO - codeparrot_training - Step 11476: {'lr': 0.00045344015880821, 'samples': 5876224, 'steps': 11476, 'loss/train': 4.021996974945068} 02/24/2022 16:25:37 - INFO - codeparrot_training - Step 11477: {'lr': 0.0004534306485070644, 'samples': 5876736, 'steps': 11477, 'loss/train': 2.0361852645874023} 02/24/2022 16:25:43 - INFO - codeparrot_training - Step 11478: {'lr': 0.0004534211373344864, 'samples': 5877248, 'steps': 11478, 'loss/train': 2.842474937438965} 02/24/2022 16:25:46 - INFO - codeparrot_training - Step 11479: {'lr': 0.00045341162529051704, 'samples': 5877760, 'steps': 11479, 'loss/train': 0.6766850352287292} 02/24/2022 16:25:53 - INFO - codeparrot_training - Step 11480: {'lr': 0.0004534021123751968, 'samples': 5878272, 'steps': 11480, 'loss/train': 1.167156457901001} 02/24/2022 16:25:57 - INFO - codeparrot_training - Step 11481: {'lr': 0.0004533925985885664, 'samples': 5878784, 'steps': 11481, 'loss/train': 3.0065672397613525} 02/24/2022 16:26:02 - INFO - codeparrot_training - Step 11482: {'lr': 0.00045338308393066685, 'samples': 5879296, 'steps': 11482, 'loss/train': 1.7414146661758423} 02/24/2022 16:26:06 - INFO - codeparrot_training - Step 11483: {'lr': 0.00045337356840153864, 'samples': 5879808, 'steps': 11483, 'loss/train': 1.34998619556427} 02/24/2022 16:26:11 - INFO - codeparrot_training - Step 11484: {'lr': 0.00045336405200122266, 'samples': 5880320, 'steps': 11484, 'loss/train': 0.5113691687583923} 02/24/2022 16:26:15 - INFO - codeparrot_training - Step 11485: {'lr': 0.0004533545347297597, 'samples': 5880832, 'steps': 11485, 'loss/train': 1.5709724426269531} 02/24/2022 16:26:20 - INFO - codeparrot_training - Step 11486: {'lr': 0.0004533450165871904, 'samples': 5881344, 'steps': 11486, 'loss/train': 3.4390742778778076} 02/24/2022 16:26:24 - INFO - codeparrot_training - Step 11487: {'lr': 0.00045333549757355573, 'samples': 5881856, 'steps': 11487, 'loss/train': 1.904098391532898} 02/24/2022 16:26:30 - INFO - codeparrot_training - Step 11488: {'lr': 0.0004533259776888963, 'samples': 5882368, 'steps': 11488, 'loss/train': 0.5446158647537231} 02/24/2022 16:26:33 - INFO - codeparrot_training - Step 11489: {'lr': 0.00045331645693325295, 'samples': 5882880, 'steps': 11489, 'loss/train': 3.0803136825561523} 02/24/2022 16:26:41 - INFO - codeparrot_training - Step 11490: {'lr': 0.0004533069353066664, 'samples': 5883392, 'steps': 11490, 'loss/train': 1.295883059501648} 02/24/2022 16:26:44 - INFO - codeparrot_training - Step 11491: {'lr': 0.0004532974128091776, 'samples': 5883904, 'steps': 11491, 'loss/train': 1.9756821393966675} 02/24/2022 16:26:49 - INFO - codeparrot_training - Step 11492: {'lr': 0.00045328788944082717, 'samples': 5884416, 'steps': 11492, 'loss/train': 1.6426712274551392} 02/24/2022 16:26:53 - INFO - codeparrot_training - Step 11493: {'lr': 0.000453278365201656, 'samples': 5884928, 'steps': 11493, 'loss/train': 0.24516497552394867} 02/24/2022 16:26:58 - INFO - codeparrot_training - Step 11494: {'lr': 0.00045326884009170486, 'samples': 5885440, 'steps': 11494, 'loss/train': 2.5774052143096924} 02/24/2022 16:27:02 - INFO - codeparrot_training - Step 11495: {'lr': 0.0004532593141110145, 'samples': 5885952, 'steps': 11495, 'loss/train': 1.4146664142608643} 02/24/2022 16:27:07 - INFO - codeparrot_training - Step 11496: {'lr': 0.00045324978725962584, 'samples': 5886464, 'steps': 11496, 'loss/train': 2.5068442821502686} 02/24/2022 16:27:11 - INFO - codeparrot_training - Step 11497: {'lr': 0.0004532402595375796, 'samples': 5886976, 'steps': 11497, 'loss/train': 2.1259090900421143} 02/24/2022 16:27:16 - INFO - codeparrot_training - Step 11498: {'lr': 0.0004532307309449167, 'samples': 5887488, 'steps': 11498, 'loss/train': 2.485146999359131} 02/24/2022 16:27:20 - INFO - codeparrot_training - Step 11499: {'lr': 0.00045322120148167777, 'samples': 5888000, 'steps': 11499, 'loss/train': 2.150620222091675} 02/24/2022 16:27:25 - INFO - codeparrot_training - Step 11500: {'lr': 0.0004532116711479038, 'samples': 5888512, 'steps': 11500, 'loss/train': 1.8521806001663208} 02/24/2022 16:27:29 - INFO - codeparrot_training - Step 11501: {'lr': 0.00045320213994363555, 'samples': 5889024, 'steps': 11501, 'loss/train': 2.786912679672241} 02/24/2022 16:27:34 - INFO - codeparrot_training - Step 11502: {'lr': 0.00045319260786891394, 'samples': 5889536, 'steps': 11502, 'loss/train': 0.6149995923042297} 02/24/2022 16:27:38 - INFO - codeparrot_training - Step 11503: {'lr': 0.0004531830749237796, 'samples': 5890048, 'steps': 11503, 'loss/train': 0.7495447993278503} 02/24/2022 16:27:44 - INFO - codeparrot_training - Step 11504: {'lr': 0.00045317354110827344, 'samples': 5890560, 'steps': 11504, 'loss/train': 1.61348557472229} 02/24/2022 16:27:47 - INFO - codeparrot_training - Step 11505: {'lr': 0.0004531640064224365, 'samples': 5891072, 'steps': 11505, 'loss/train': 1.9044476747512817} 02/24/2022 16:27:54 - INFO - codeparrot_training - Step 11506: {'lr': 0.00045315447086630937, 'samples': 5891584, 'steps': 11506, 'loss/train': 2.6038100719451904} 02/24/2022 16:27:58 - INFO - codeparrot_training - Step 11507: {'lr': 0.000453144934439933, 'samples': 5892096, 'steps': 11507, 'loss/train': 2.211012601852417} 02/24/2022 16:28:03 - INFO - codeparrot_training - Step 11508: {'lr': 0.0004531353971433483, 'samples': 5892608, 'steps': 11508, 'loss/train': 3.507301092147827} 02/24/2022 16:28:07 - INFO - codeparrot_training - Step 11509: {'lr': 0.000453125858976596, 'samples': 5893120, 'steps': 11509, 'loss/train': 1.801110029220581} 02/24/2022 16:28:13 - INFO - codeparrot_training - Step 11510: {'lr': 0.000453116319939717, 'samples': 5893632, 'steps': 11510, 'loss/train': 2.172649383544922} 02/24/2022 16:28:16 - INFO - codeparrot_training - Step 11511: {'lr': 0.0004531067800327523, 'samples': 5894144, 'steps': 11511, 'loss/train': 2.3457870483398438} 02/24/2022 16:28:22 - INFO - codeparrot_training - Step 11512: {'lr': 0.0004530972392557425, 'samples': 5894656, 'steps': 11512, 'loss/train': 1.9584993124008179} 02/24/2022 16:28:25 - INFO - codeparrot_training - Step 11513: {'lr': 0.0004530876976087288, 'samples': 5895168, 'steps': 11513, 'loss/train': 2.7640602588653564} 02/24/2022 16:28:31 - INFO - codeparrot_training - Step 11514: {'lr': 0.00045307815509175177, 'samples': 5895680, 'steps': 11514, 'loss/train': 3.1797080039978027} 02/24/2022 16:28:34 - INFO - codeparrot_training - Step 11515: {'lr': 0.00045306861170485235, 'samples': 5896192, 'steps': 11515, 'loss/train': 1.0782145261764526} 02/24/2022 16:28:41 - INFO - codeparrot_training - Step 11516: {'lr': 0.00045305906744807156, 'samples': 5896704, 'steps': 11516, 'loss/train': 1.984292984008789} 02/24/2022 16:28:45 - INFO - codeparrot_training - Step 11517: {'lr': 0.0004530495223214502, 'samples': 5897216, 'steps': 11517, 'loss/train': 2.7023088932037354} 02/24/2022 16:28:51 - INFO - codeparrot_training - Step 11518: {'lr': 0.00045303997632502915, 'samples': 5897728, 'steps': 11518, 'loss/train': 2.081519603729248} 02/24/2022 16:28:54 - INFO - codeparrot_training - Step 11519: {'lr': 0.00045303042945884933, 'samples': 5898240, 'steps': 11519, 'loss/train': 9.124129295349121} 02/24/2022 16:29:00 - INFO - codeparrot_training - Step 11520: {'lr': 0.0004530208817229516, 'samples': 5898752, 'steps': 11520, 'loss/train': 3.5649006366729736} 02/24/2022 16:29:03 - INFO - codeparrot_training - Step 11521: {'lr': 0.00045301133311737685, 'samples': 5899264, 'steps': 11521, 'loss/train': 1.4435570240020752} 02/24/2022 16:29:09 - INFO - codeparrot_training - Step 11522: {'lr': 0.00045300178364216605, 'samples': 5899776, 'steps': 11522, 'loss/train': 2.2180709838867188} 02/24/2022 16:29:12 - INFO - codeparrot_training - Step 11523: {'lr': 0.00045299223329736004, 'samples': 5900288, 'steps': 11523, 'loss/train': 2.1526942253112793} 02/24/2022 16:29:18 - INFO - codeparrot_training - Step 11524: {'lr': 0.00045298268208299983, 'samples': 5900800, 'steps': 11524, 'loss/train': 3.007046937942505} 02/24/2022 16:29:21 - INFO - codeparrot_training - Step 11525: {'lr': 0.0004529731299991262, 'samples': 5901312, 'steps': 11525, 'loss/train': 2.2633237838745117} 02/24/2022 16:29:28 - INFO - codeparrot_training - Step 11526: {'lr': 0.00045296357704578016, 'samples': 5901824, 'steps': 11526, 'loss/train': 0.4225294291973114} 02/24/2022 16:29:32 - INFO - codeparrot_training - Step 11527: {'lr': 0.0004529540232230026, 'samples': 5902336, 'steps': 11527, 'loss/train': 2.5396106243133545} 02/24/2022 16:29:37 - INFO - codeparrot_training - Step 11528: {'lr': 0.00045294446853083446, 'samples': 5902848, 'steps': 11528, 'loss/train': 1.1456432342529297} 02/24/2022 16:29:41 - INFO - codeparrot_training - Step 11529: {'lr': 0.0004529349129693166, 'samples': 5903360, 'steps': 11529, 'loss/train': 2.259439468383789} 02/24/2022 16:29:46 - INFO - codeparrot_training - Step 11530: {'lr': 0.0004529253565384901, 'samples': 5903872, 'steps': 11530, 'loss/train': 2.5124430656433105} 02/24/2022 16:29:50 - INFO - codeparrot_training - Step 11531: {'lr': 0.00045291579923839576, 'samples': 5904384, 'steps': 11531, 'loss/train': 2.6305928230285645} 02/24/2022 16:29:55 - INFO - codeparrot_training - Step 11532: {'lr': 0.0004529062410690745, 'samples': 5904896, 'steps': 11532, 'loss/train': 1.2240854501724243} 02/24/2022 16:29:59 - INFO - codeparrot_training - Step 11533: {'lr': 0.00045289668203056743, 'samples': 5905408, 'steps': 11533, 'loss/train': 2.1779584884643555} 02/24/2022 16:30:04 - INFO - codeparrot_training - Step 11534: {'lr': 0.00045288712212291537, 'samples': 5905920, 'steps': 11534, 'loss/train': 1.7150356769561768} 02/24/2022 16:30:08 - INFO - codeparrot_training - Step 11535: {'lr': 0.0004528775613461593, 'samples': 5906432, 'steps': 11535, 'loss/train': 1.0887821912765503} 02/24/2022 16:30:13 - INFO - codeparrot_training - Step 11536: {'lr': 0.0004528679997003403, 'samples': 5906944, 'steps': 11536, 'loss/train': 2.1326656341552734} 02/24/2022 16:30:17 - INFO - codeparrot_training - Step 11537: {'lr': 0.000452858437185499, 'samples': 5907456, 'steps': 11537, 'loss/train': 1.573224663734436} 02/24/2022 16:30:22 - INFO - codeparrot_training - Step 11538: {'lr': 0.00045284887380167674, 'samples': 5907968, 'steps': 11538, 'loss/train': 2.3158607482910156} 02/24/2022 16:30:26 - INFO - codeparrot_training - Step 11539: {'lr': 0.0004528393095489142, 'samples': 5908480, 'steps': 11539, 'loss/train': 2.030322313308716} 02/24/2022 16:30:32 - INFO - codeparrot_training - Step 11540: {'lr': 0.0004528297444272525, 'samples': 5908992, 'steps': 11540, 'loss/train': 2.8145506381988525} 02/24/2022 16:30:35 - INFO - codeparrot_training - Step 11541: {'lr': 0.0004528201784367326, 'samples': 5909504, 'steps': 11541, 'loss/train': 2.621253252029419} 02/24/2022 16:30:42 - INFO - codeparrot_training - Step 11542: {'lr': 0.00045281061157739544, 'samples': 5910016, 'steps': 11542, 'loss/train': 2.1283960342407227} 02/24/2022 16:30:46 - INFO - codeparrot_training - Step 11543: {'lr': 0.000452801043849282, 'samples': 5910528, 'steps': 11543, 'loss/train': 2.3949320316314697} 02/24/2022 16:30:51 - INFO - codeparrot_training - Step 11544: {'lr': 0.00045279147525243335, 'samples': 5911040, 'steps': 11544, 'loss/train': 2.003279209136963} 02/24/2022 16:30:55 - INFO - codeparrot_training - Step 11545: {'lr': 0.0004527819057868904, 'samples': 5911552, 'steps': 11545, 'loss/train': 2.2588186264038086} 02/24/2022 16:31:00 - INFO - codeparrot_training - Step 11546: {'lr': 0.00045277233545269415, 'samples': 5912064, 'steps': 11546, 'loss/train': 1.4942573308944702} 02/24/2022 16:31:04 - INFO - codeparrot_training - Step 11547: {'lr': 0.00045276276424988554, 'samples': 5912576, 'steps': 11547, 'loss/train': 2.608991861343384} 02/24/2022 16:31:10 - INFO - codeparrot_training - Step 11548: {'lr': 0.0004527531921785057, 'samples': 5913088, 'steps': 11548, 'loss/train': 2.479332208633423} 02/24/2022 16:31:13 - INFO - codeparrot_training - Step 11549: {'lr': 0.00045274361923859554, 'samples': 5913600, 'steps': 11549, 'loss/train': 3.089203357696533} 02/24/2022 16:31:19 - INFO - codeparrot_training - Step 11550: {'lr': 0.0004527340454301961, 'samples': 5914112, 'steps': 11550, 'loss/train': 2.0475707054138184} 02/24/2022 16:31:22 - INFO - codeparrot_training - Step 11551: {'lr': 0.0004527244707533483, 'samples': 5914624, 'steps': 11551, 'loss/train': 2.8862383365631104} 02/24/2022 16:31:29 - INFO - codeparrot_training - Step 11552: {'lr': 0.00045271489520809337, 'samples': 5915136, 'steps': 11552, 'loss/train': 3.0720741748809814} 02/24/2022 16:31:33 - INFO - codeparrot_training - Step 11553: {'lr': 0.0004527053187944722, 'samples': 5915648, 'steps': 11553, 'loss/train': 2.62570858001709} 02/24/2022 16:31:38 - INFO - codeparrot_training - Step 11554: {'lr': 0.00045269574151252567, 'samples': 5916160, 'steps': 11554, 'loss/train': 2.2534070014953613} 02/24/2022 16:31:42 - INFO - codeparrot_training - Step 11555: {'lr': 0.00045268616336229504, 'samples': 5916672, 'steps': 11555, 'loss/train': 2.829556941986084} 02/24/2022 16:31:47 - INFO - codeparrot_training - Step 11556: {'lr': 0.0004526765843438213, 'samples': 5917184, 'steps': 11556, 'loss/train': 3.0566415786743164} 02/24/2022 16:31:51 - INFO - codeparrot_training - Step 11557: {'lr': 0.0004526670044571454, 'samples': 5917696, 'steps': 11557, 'loss/train': 1.6900790929794312} 02/24/2022 16:31:56 - INFO - codeparrot_training - Step 11558: {'lr': 0.00045265742370230835, 'samples': 5918208, 'steps': 11558, 'loss/train': 2.8533871173858643} 02/24/2022 16:32:00 - INFO - codeparrot_training - Step 11559: {'lr': 0.00045264784207935127, 'samples': 5918720, 'steps': 11559, 'loss/train': 2.1948347091674805} 02/24/2022 16:32:05 - INFO - codeparrot_training - Step 11560: {'lr': 0.0004526382595883152, 'samples': 5919232, 'steps': 11560, 'loss/train': 2.42537522315979} 02/24/2022 16:32:09 - INFO - codeparrot_training - Step 11561: {'lr': 0.0004526286762292411, 'samples': 5919744, 'steps': 11561, 'loss/train': 2.9881153106689453} 02/24/2022 16:32:16 - INFO - codeparrot_training - Step 11562: {'lr': 0.00045261909200217023, 'samples': 5920256, 'steps': 11562, 'loss/train': 2.443436622619629} 02/24/2022 16:32:20 - INFO - codeparrot_training - Step 11563: {'lr': 0.0004526095069071434, 'samples': 5920768, 'steps': 11563, 'loss/train': 0.5942236185073853} 02/24/2022 16:32:25 - INFO - codeparrot_training - Step 11564: {'lr': 0.0004525999209442018, 'samples': 5921280, 'steps': 11564, 'loss/train': 9.764789581298828} 02/24/2022 16:32:29 - INFO - codeparrot_training - Step 11565: {'lr': 0.0004525903341133865, 'samples': 5921792, 'steps': 11565, 'loss/train': 1.7377638816833496} 02/24/2022 16:32:34 - INFO - codeparrot_training - Step 11566: {'lr': 0.0004525807464147385, 'samples': 5922304, 'steps': 11566, 'loss/train': 2.552253484725952} 02/24/2022 16:32:38 - INFO - codeparrot_training - Step 11567: {'lr': 0.00045257115784829897, 'samples': 5922816, 'steps': 11567, 'loss/train': 1.2742775678634644} 02/24/2022 16:32:43 - INFO - codeparrot_training - Step 11568: {'lr': 0.00045256156841410884, 'samples': 5923328, 'steps': 11568, 'loss/train': 2.5633201599121094} 02/24/2022 16:32:47 - INFO - codeparrot_training - Step 11569: {'lr': 0.0004525519781122093, 'samples': 5923840, 'steps': 11569, 'loss/train': 2.5515196323394775} 02/24/2022 16:32:52 - INFO - codeparrot_training - Step 11570: {'lr': 0.00045254238694264145, 'samples': 5924352, 'steps': 11570, 'loss/train': 0.513909637928009} 02/24/2022 16:32:56 - INFO - codeparrot_training - Step 11571: {'lr': 0.00045253279490544627, 'samples': 5924864, 'steps': 11571, 'loss/train': 2.1761553287506104} 02/24/2022 16:33:03 - INFO - codeparrot_training - Step 11572: {'lr': 0.0004525232020006649, 'samples': 5925376, 'steps': 11572, 'loss/train': 2.4223105907440186} 02/24/2022 16:33:07 - INFO - codeparrot_training - Step 11573: {'lr': 0.00045251360822833855, 'samples': 5925888, 'steps': 11573, 'loss/train': 2.2352139949798584} 02/24/2022 16:33:12 - INFO - codeparrot_training - Step 11574: {'lr': 0.00045250401358850814, 'samples': 5926400, 'steps': 11574, 'loss/train': 2.3021950721740723} 02/24/2022 16:33:16 - INFO - codeparrot_training - Step 11575: {'lr': 0.00045249441808121484, 'samples': 5926912, 'steps': 11575, 'loss/train': 2.0894877910614014} 02/24/2022 16:33:21 - INFO - codeparrot_training - Step 11576: {'lr': 0.0004524848217064997, 'samples': 5927424, 'steps': 11576, 'loss/train': 1.4455124139785767} 02/24/2022 16:33:27 - INFO - codeparrot_training - Step 11577: {'lr': 0.0004524752244644039, 'samples': 5927936, 'steps': 11577, 'loss/train': 2.033019542694092} 02/24/2022 16:33:30 - INFO - codeparrot_training - Step 11578: {'lr': 0.0004524656263549686, 'samples': 5928448, 'steps': 11578, 'loss/train': 2.201353073120117} 02/24/2022 16:33:36 - INFO - codeparrot_training - Step 11579: {'lr': 0.0004524560273782348, 'samples': 5928960, 'steps': 11579, 'loss/train': 1.791818618774414} 02/24/2022 16:33:39 - INFO - codeparrot_training - Step 11580: {'lr': 0.00045244642753424364, 'samples': 5929472, 'steps': 11580, 'loss/train': 1.9452608823776245} 02/24/2022 16:33:47 - INFO - codeparrot_training - Step 11581: {'lr': 0.0004524368268230363, 'samples': 5929984, 'steps': 11581, 'loss/train': 2.6028406620025635} 02/24/2022 16:33:50 - INFO - codeparrot_training - Step 11582: {'lr': 0.00045242722524465386, 'samples': 5930496, 'steps': 11582, 'loss/train': 8.990501403808594} 02/24/2022 16:33:56 - INFO - codeparrot_training - Step 11583: {'lr': 0.00045241762279913745, 'samples': 5931008, 'steps': 11583, 'loss/train': 1.931323766708374} 02/24/2022 16:33:59 - INFO - codeparrot_training - Step 11584: {'lr': 0.0004524080194865283, 'samples': 5931520, 'steps': 11584, 'loss/train': 3.437455177307129} 02/24/2022 16:34:05 - INFO - codeparrot_training - Step 11585: {'lr': 0.00045239841530686736, 'samples': 5932032, 'steps': 11585, 'loss/train': 2.2022390365600586} 02/24/2022 16:34:08 - INFO - codeparrot_training - Step 11586: {'lr': 0.000452388810260196, 'samples': 5932544, 'steps': 11586, 'loss/train': 4.043230056762695} 02/24/2022 16:34:14 - INFO - codeparrot_training - Step 11587: {'lr': 0.0004523792043465551, 'samples': 5933056, 'steps': 11587, 'loss/train': 2.509674310684204} 02/24/2022 16:34:17 - INFO - codeparrot_training - Step 11588: {'lr': 0.00045236959756598605, 'samples': 5933568, 'steps': 11588, 'loss/train': 2.6086578369140625} 02/24/2022 16:34:23 - INFO - codeparrot_training - Step 11589: {'lr': 0.0004523599899185299, 'samples': 5934080, 'steps': 11589, 'loss/train': 2.2144455909729004} 02/24/2022 16:34:26 - INFO - codeparrot_training - Step 11590: {'lr': 0.0004523503814042277, 'samples': 5934592, 'steps': 11590, 'loss/train': 2.5732262134552} 02/24/2022 16:34:32 - INFO - codeparrot_training - Step 11591: {'lr': 0.00045234077202312086, 'samples': 5935104, 'steps': 11591, 'loss/train': 1.3404759168624878} 02/24/2022 16:34:35 - INFO - codeparrot_training - Step 11592: {'lr': 0.00045233116177525036, 'samples': 5935616, 'steps': 11592, 'loss/train': 1.8101704120635986} 02/24/2022 16:34:41 - INFO - codeparrot_training - Step 11593: {'lr': 0.00045232155066065737, 'samples': 5936128, 'steps': 11593, 'loss/train': 4.502933025360107} 02/24/2022 16:34:44 - INFO - codeparrot_training - Step 11594: {'lr': 0.00045231193867938314, 'samples': 5936640, 'steps': 11594, 'loss/train': 2.710329294204712} 02/24/2022 16:34:50 - INFO - codeparrot_training - Step 11595: {'lr': 0.0004523023258314688, 'samples': 5937152, 'steps': 11595, 'loss/train': 3.074796199798584} 02/24/2022 16:34:53 - INFO - codeparrot_training - Step 11596: {'lr': 0.00045229271211695554, 'samples': 5937664, 'steps': 11596, 'loss/train': 3.2067344188690186} 02/24/2022 16:35:01 - INFO - codeparrot_training - Step 11597: {'lr': 0.00045228309753588447, 'samples': 5938176, 'steps': 11597, 'loss/train': 2.750671625137329} 02/24/2022 16:35:04 - INFO - codeparrot_training - Step 11598: {'lr': 0.0004522734820882969, 'samples': 5938688, 'steps': 11598, 'loss/train': 1.4830482006072998} 02/24/2022 16:35:10 - INFO - codeparrot_training - Step 11599: {'lr': 0.00045226386577423394, 'samples': 5939200, 'steps': 11599, 'loss/train': 0.29062801599502563} 02/24/2022 16:35:13 - INFO - codeparrot_training - Step 11600: {'lr': 0.0004522542485937369, 'samples': 5939712, 'steps': 11600, 'loss/train': 2.089154005050659} 02/24/2022 16:35:19 - INFO - codeparrot_training - Step 11601: {'lr': 0.0004522446305468468, 'samples': 5940224, 'steps': 11601, 'loss/train': 2.1879985332489014} 02/24/2022 16:35:22 - INFO - codeparrot_training - Step 11602: {'lr': 0.00045223501163360494, 'samples': 5940736, 'steps': 11602, 'loss/train': 1.8698228597640991} 02/24/2022 16:35:28 - INFO - codeparrot_training - Step 11603: {'lr': 0.0004522253918540524, 'samples': 5941248, 'steps': 11603, 'loss/train': 2.562459945678711} 02/24/2022 16:35:31 - INFO - codeparrot_training - Step 11604: {'lr': 0.00045221577120823064, 'samples': 5941760, 'steps': 11604, 'loss/train': 0.799372673034668} 02/24/2022 16:35:37 - INFO - codeparrot_training - Step 11605: {'lr': 0.00045220614969618066, 'samples': 5942272, 'steps': 11605, 'loss/train': 1.231338620185852} 02/24/2022 16:35:40 - INFO - codeparrot_training - Step 11606: {'lr': 0.0004521965273179438, 'samples': 5942784, 'steps': 11606, 'loss/train': 1.9436644315719604} 02/24/2022 16:35:50 - INFO - codeparrot_training - Step 11607: {'lr': 0.00045218690407356117, 'samples': 5943296, 'steps': 11607, 'loss/train': 2.1516036987304688} 02/24/2022 16:35:54 - INFO - codeparrot_training - Step 11608: {'lr': 0.00045217727996307405, 'samples': 5943808, 'steps': 11608, 'loss/train': 1.9412357807159424} 02/24/2022 16:35:59 - INFO - codeparrot_training - Step 11609: {'lr': 0.0004521676549865237, 'samples': 5944320, 'steps': 11609, 'loss/train': 1.489535927772522} 02/24/2022 16:36:02 - INFO - codeparrot_training - Step 11610: {'lr': 0.0004521580291439513, 'samples': 5944832, 'steps': 11610, 'loss/train': 2.8663041591644287} 02/24/2022 16:36:08 - INFO - codeparrot_training - Step 11611: {'lr': 0.00045214840243539803, 'samples': 5945344, 'steps': 11611, 'loss/train': 2.324397563934326} 02/24/2022 16:36:11 - INFO - codeparrot_training - Step 11612: {'lr': 0.00045213877486090524, 'samples': 5945856, 'steps': 11612, 'loss/train': 2.4788997173309326} 02/24/2022 16:36:17 - INFO - codeparrot_training - Step 11613: {'lr': 0.0004521291464205141, 'samples': 5946368, 'steps': 11613, 'loss/train': 1.1995437145233154} 02/24/2022 16:36:20 - INFO - codeparrot_training - Step 11614: {'lr': 0.0004521195171142659, 'samples': 5946880, 'steps': 11614, 'loss/train': 2.288405656814575} 02/24/2022 16:36:26 - INFO - codeparrot_training - Step 11615: {'lr': 0.0004521098869422019, 'samples': 5947392, 'steps': 11615, 'loss/train': 2.9078524112701416} 02/24/2022 16:36:29 - INFO - codeparrot_training - Step 11616: {'lr': 0.00045210025590436333, 'samples': 5947904, 'steps': 11616, 'loss/train': 2.2069733142852783} 02/24/2022 16:36:37 - INFO - codeparrot_training - Step 11617: {'lr': 0.00045209062400079135, 'samples': 5948416, 'steps': 11617, 'loss/train': 1.4807133674621582} 02/24/2022 16:36:40 - INFO - codeparrot_training - Step 11618: {'lr': 0.00045208099123152735, 'samples': 5948928, 'steps': 11618, 'loss/train': 2.1155927181243896} 02/24/2022 16:36:46 - INFO - codeparrot_training - Step 11619: {'lr': 0.00045207135759661255, 'samples': 5949440, 'steps': 11619, 'loss/train': 2.448025941848755} 02/24/2022 16:36:49 - INFO - codeparrot_training - Step 11620: {'lr': 0.0004520617230960883, 'samples': 5949952, 'steps': 11620, 'loss/train': 2.3883309364318848} 02/24/2022 16:36:55 - INFO - codeparrot_training - Step 11621: {'lr': 0.0004520520877299957, 'samples': 5950464, 'steps': 11621, 'loss/train': 2.565316915512085} 02/24/2022 16:36:58 - INFO - codeparrot_training - Step 11622: {'lr': 0.00045204245149837606, 'samples': 5950976, 'steps': 11622, 'loss/train': 3.1621944904327393} 02/24/2022 16:37:04 - INFO - codeparrot_training - Step 11623: {'lr': 0.00045203281440127087, 'samples': 5951488, 'steps': 11623, 'loss/train': 0.9470974206924438} 02/24/2022 16:37:07 - INFO - codeparrot_training - Step 11624: {'lr': 0.00045202317643872113, 'samples': 5952000, 'steps': 11624, 'loss/train': 2.404804229736328} 02/24/2022 16:37:13 - INFO - codeparrot_training - Step 11625: {'lr': 0.0004520135376107683, 'samples': 5952512, 'steps': 11625, 'loss/train': 2.227313995361328} 02/24/2022 16:37:16 - INFO - codeparrot_training - Step 11626: {'lr': 0.00045200389791745364, 'samples': 5953024, 'steps': 11626, 'loss/train': 2.097395181655884} 02/24/2022 16:37:22 - INFO - codeparrot_training - Step 11627: {'lr': 0.0004519942573588184, 'samples': 5953536, 'steps': 11627, 'loss/train': 2.2431325912475586} 02/24/2022 16:37:26 - INFO - codeparrot_training - Step 11628: {'lr': 0.00045198461593490394, 'samples': 5954048, 'steps': 11628, 'loss/train': 6.506643772125244} 02/24/2022 16:37:31 - INFO - codeparrot_training - Step 11629: {'lr': 0.0004519749736457515, 'samples': 5954560, 'steps': 11629, 'loss/train': 1.7646405696868896} 02/24/2022 16:37:35 - INFO - codeparrot_training - Step 11630: {'lr': 0.00045196533049140234, 'samples': 5955072, 'steps': 11630, 'loss/train': 0.810029149055481} 02/24/2022 16:37:40 - INFO - codeparrot_training - Step 11631: {'lr': 0.0004519556864718979, 'samples': 5955584, 'steps': 11631, 'loss/train': 2.208754777908325} 02/24/2022 16:37:43 - INFO - codeparrot_training - Step 11632: {'lr': 0.00045194604158727936, 'samples': 5956096, 'steps': 11632, 'loss/train': 2.148669481277466} 02/24/2022 16:37:51 - INFO - codeparrot_training - Step 11633: {'lr': 0.0004519363958375882, 'samples': 5956608, 'steps': 11633, 'loss/train': 1.9196035861968994} 02/24/2022 16:37:54 - INFO - codeparrot_training - Step 11634: {'lr': 0.00045192674922286556, 'samples': 5957120, 'steps': 11634, 'loss/train': 2.1501853466033936} 02/24/2022 16:38:00 - INFO - codeparrot_training - Step 11635: {'lr': 0.00045191710174315294, 'samples': 5957632, 'steps': 11635, 'loss/train': 2.6451566219329834} 02/24/2022 16:38:03 - INFO - codeparrot_training - Step 11636: {'lr': 0.0004519074533984915, 'samples': 5958144, 'steps': 11636, 'loss/train': 1.8730019330978394} 02/24/2022 16:38:09 - INFO - codeparrot_training - Step 11637: {'lr': 0.0004518978041889227, 'samples': 5958656, 'steps': 11637, 'loss/train': 1.8419873714447021} 02/24/2022 16:38:12 - INFO - codeparrot_training - Step 11638: {'lr': 0.00045188815411448767, 'samples': 5959168, 'steps': 11638, 'loss/train': 3.911593437194824} 02/24/2022 16:38:18 - INFO - codeparrot_training - Step 11639: {'lr': 0.00045187850317522806, 'samples': 5959680, 'steps': 11639, 'loss/train': 2.027243137359619} 02/24/2022 16:38:21 - INFO - codeparrot_training - Step 11640: {'lr': 0.00045186885137118494, 'samples': 5960192, 'steps': 11640, 'loss/train': 2.275195360183716} 02/24/2022 16:38:27 - INFO - codeparrot_training - Step 11641: {'lr': 0.0004518591987023999, 'samples': 5960704, 'steps': 11641, 'loss/train': 2.997321605682373} 02/24/2022 16:38:30 - INFO - codeparrot_training - Step 11642: {'lr': 0.000451849545168914, 'samples': 5961216, 'steps': 11642, 'loss/train': 2.1607887744903564} 02/24/2022 16:38:38 - INFO - codeparrot_training - Step 11643: {'lr': 0.00045183989077076883, 'samples': 5961728, 'steps': 11643, 'loss/train': 1.8438206911087036} 02/24/2022 16:38:41 - INFO - codeparrot_training - Step 11644: {'lr': 0.00045183023550800564, 'samples': 5962240, 'steps': 11644, 'loss/train': 2.06140398979187} 02/24/2022 16:38:47 - INFO - codeparrot_training - Step 11645: {'lr': 0.0004518205793806658, 'samples': 5962752, 'steps': 11645, 'loss/train': 2.292933225631714} 02/24/2022 16:38:50 - INFO - codeparrot_training - Step 11646: {'lr': 0.0004518109223887907, 'samples': 5963264, 'steps': 11646, 'loss/train': 2.528597593307495} 02/24/2022 16:38:56 - INFO - codeparrot_training - Step 11647: {'lr': 0.0004518012645324217, 'samples': 5963776, 'steps': 11647, 'loss/train': 2.747081756591797} 02/24/2022 16:38:59 - INFO - codeparrot_training - Step 11648: {'lr': 0.00045179160581160005, 'samples': 5964288, 'steps': 11648, 'loss/train': 2.287997007369995} 02/24/2022 16:39:05 - INFO - codeparrot_training - Step 11649: {'lr': 0.0004517819462263674, 'samples': 5964800, 'steps': 11649, 'loss/train': 2.0138418674468994} 02/24/2022 16:39:08 - INFO - codeparrot_training - Step 11650: {'lr': 0.0004517722857767649, 'samples': 5965312, 'steps': 11650, 'loss/train': 1.583605408668518} 02/24/2022 16:39:14 - INFO - codeparrot_training - Step 11651: {'lr': 0.0004517626244628339, 'samples': 5965824, 'steps': 11651, 'loss/train': 0.8320340514183044} 02/24/2022 16:39:17 - INFO - codeparrot_training - Step 11652: {'lr': 0.000451752962284616, 'samples': 5966336, 'steps': 11652, 'loss/train': 2.384977340698242} 02/24/2022 16:39:23 - INFO - codeparrot_training - Step 11653: {'lr': 0.0004517432992421524, 'samples': 5966848, 'steps': 11653, 'loss/train': 1.457533836364746} 02/24/2022 16:39:26 - INFO - codeparrot_training - Step 11654: {'lr': 0.00045173363533548464, 'samples': 5967360, 'steps': 11654, 'loss/train': 1.8976963758468628} 02/24/2022 16:39:34 - INFO - codeparrot_training - Step 11655: {'lr': 0.00045172397056465405, 'samples': 5967872, 'steps': 11655, 'loss/train': 2.159296989440918} 02/24/2022 16:39:37 - INFO - codeparrot_training - Step 11656: {'lr': 0.000451714304929702, 'samples': 5968384, 'steps': 11656, 'loss/train': 1.856048822402954} 02/24/2022 16:39:43 - INFO - codeparrot_training - Step 11657: {'lr': 0.0004517046384306699, 'samples': 5968896, 'steps': 11657, 'loss/train': 1.6358002424240112} 02/24/2022 16:39:46 - INFO - codeparrot_training - Step 11658: {'lr': 0.00045169497106759915, 'samples': 5969408, 'steps': 11658, 'loss/train': 1.6984803676605225} 02/24/2022 16:39:52 - INFO - codeparrot_training - Step 11659: {'lr': 0.0004516853028405312, 'samples': 5969920, 'steps': 11659, 'loss/train': 0.12451069056987762} 02/24/2022 16:39:55 - INFO - codeparrot_training - Step 11660: {'lr': 0.0004516756337495075, 'samples': 5970432, 'steps': 11660, 'loss/train': 2.5406110286712646} 02/24/2022 16:40:01 - INFO - codeparrot_training - Step 11661: {'lr': 0.00045166596379456935, 'samples': 5970944, 'steps': 11661, 'loss/train': 2.4823601245880127} 02/24/2022 16:40:04 - INFO - codeparrot_training - Step 11662: {'lr': 0.0004516562929757584, 'samples': 5971456, 'steps': 11662, 'loss/train': 2.3144519329071045} 02/24/2022 16:40:10 - INFO - codeparrot_training - Step 11663: {'lr': 0.0004516466212931158, 'samples': 5971968, 'steps': 11663, 'loss/train': 1.219640851020813} 02/24/2022 16:40:13 - INFO - codeparrot_training - Step 11664: {'lr': 0.00045163694874668316, 'samples': 5972480, 'steps': 11664, 'loss/train': 2.5993688106536865} 02/24/2022 16:40:21 - INFO - codeparrot_training - Step 11665: {'lr': 0.0004516272753365018, 'samples': 5972992, 'steps': 11665, 'loss/train': 2.4267616271972656} 02/24/2022 16:40:24 - INFO - codeparrot_training - Step 11666: {'lr': 0.0004516176010626132, 'samples': 5973504, 'steps': 11666, 'loss/train': 0.460162878036499} 02/24/2022 16:40:30 - INFO - codeparrot_training - Step 11667: {'lr': 0.00045160792592505893, 'samples': 5974016, 'steps': 11667, 'loss/train': 1.9066812992095947} 02/24/2022 16:40:33 - INFO - codeparrot_training - Step 11668: {'lr': 0.0004515982499238802, 'samples': 5974528, 'steps': 11668, 'loss/train': 2.975303888320923} 02/24/2022 16:40:39 - INFO - codeparrot_training - Step 11669: {'lr': 0.0004515885730591187, 'samples': 5975040, 'steps': 11669, 'loss/train': 1.0526678562164307} 02/24/2022 16:40:42 - INFO - codeparrot_training - Step 11670: {'lr': 0.0004515788953308156, 'samples': 5975552, 'steps': 11670, 'loss/train': 2.3513824939727783} 02/24/2022 16:40:48 - INFO - codeparrot_training - Step 11671: {'lr': 0.00045156921673901267, 'samples': 5976064, 'steps': 11671, 'loss/train': 2.775592565536499} 02/24/2022 16:40:51 - INFO - codeparrot_training - Step 11672: {'lr': 0.0004515595372837512, 'samples': 5976576, 'steps': 11672, 'loss/train': 2.224604606628418} 02/24/2022 16:40:57 - INFO - codeparrot_training - Step 11673: {'lr': 0.00045154985696507267, 'samples': 5977088, 'steps': 11673, 'loss/train': 1.5069808959960938} 02/24/2022 16:41:00 - INFO - codeparrot_training - Step 11674: {'lr': 0.0004515401757830185, 'samples': 5977600, 'steps': 11674, 'loss/train': 3.1457831859588623} 02/24/2022 16:41:08 - INFO - codeparrot_training - Step 11675: {'lr': 0.0004515304937376302, 'samples': 5978112, 'steps': 11675, 'loss/train': 1.969599723815918} 02/24/2022 16:41:11 - INFO - codeparrot_training - Step 11676: {'lr': 0.00045152081082894935, 'samples': 5978624, 'steps': 11676, 'loss/train': 1.2151014804840088} 02/24/2022 16:41:16 - INFO - codeparrot_training - Step 11677: {'lr': 0.00045151112705701723, 'samples': 5979136, 'steps': 11677, 'loss/train': 1.9878487586975098} 02/24/2022 16:41:20 - INFO - codeparrot_training - Step 11678: {'lr': 0.00045150144242187554, 'samples': 5979648, 'steps': 11678, 'loss/train': 2.6872684955596924} 02/24/2022 16:41:25 - INFO - codeparrot_training - Step 11679: {'lr': 0.0004514917569235656, 'samples': 5980160, 'steps': 11679, 'loss/train': 3.4789350032806396} 02/24/2022 16:41:29 - INFO - codeparrot_training - Step 11680: {'lr': 0.00045148207056212896, 'samples': 5980672, 'steps': 11680, 'loss/train': 1.7429273128509521} 02/24/2022 16:41:35 - INFO - codeparrot_training - Step 11681: {'lr': 0.0004514723833376071, 'samples': 5981184, 'steps': 11681, 'loss/train': 1.8754760026931763} 02/24/2022 16:41:38 - INFO - codeparrot_training - Step 11682: {'lr': 0.00045146269525004153, 'samples': 5981696, 'steps': 11682, 'loss/train': 1.0769193172454834} 02/24/2022 16:41:44 - INFO - codeparrot_training - Step 11683: {'lr': 0.00045145300629947374, 'samples': 5982208, 'steps': 11683, 'loss/train': 2.6670479774475098} 02/24/2022 16:41:47 - INFO - codeparrot_training - Step 11684: {'lr': 0.0004514433164859453, 'samples': 5982720, 'steps': 11684, 'loss/train': 2.9119553565979004} 02/24/2022 16:41:53 - INFO - codeparrot_training - Step 11685: {'lr': 0.00045143362580949754, 'samples': 5983232, 'steps': 11685, 'loss/train': 1.419554591178894} 02/24/2022 16:41:56 - INFO - codeparrot_training - Step 11686: {'lr': 0.00045142393427017214, 'samples': 5983744, 'steps': 11686, 'loss/train': 1.2289128303527832} 02/24/2022 16:42:02 - INFO - codeparrot_training - Step 11687: {'lr': 0.0004514142418680106, 'samples': 5984256, 'steps': 11687, 'loss/train': 1.5331218242645264} 02/24/2022 16:42:05 - INFO - codeparrot_training - Step 11688: {'lr': 0.00045140454860305435, 'samples': 5984768, 'steps': 11688, 'loss/train': 2.273690700531006} 02/24/2022 16:42:11 - INFO - codeparrot_training - Step 11689: {'lr': 0.000451394854475345, 'samples': 5985280, 'steps': 11689, 'loss/train': 1.7599692344665527} 02/24/2022 16:42:14 - INFO - codeparrot_training - Step 11690: {'lr': 0.0004513851594849241, 'samples': 5985792, 'steps': 11690, 'loss/train': 1.0157694816589355} 02/24/2022 16:42:21 - INFO - codeparrot_training - Step 11691: {'lr': 0.000451375463631833, 'samples': 5986304, 'steps': 11691, 'loss/train': 2.3648629188537598} 02/24/2022 16:42:25 - INFO - codeparrot_training - Step 11692: {'lr': 0.0004513657669161134, 'samples': 5986816, 'steps': 11692, 'loss/train': 1.7742458581924438} 02/24/2022 16:42:30 - INFO - codeparrot_training - Step 11693: {'lr': 0.0004513560693378068, 'samples': 5987328, 'steps': 11693, 'loss/train': 2.3537003993988037} 02/24/2022 16:42:36 - INFO - codeparrot_training - Step 11694: {'lr': 0.00045134637089695484, 'samples': 5987840, 'steps': 11694, 'loss/train': 1.635169506072998} 02/24/2022 16:42:39 - INFO - codeparrot_training - Step 11695: {'lr': 0.0004513366715935988, 'samples': 5988352, 'steps': 11695, 'loss/train': 2.943849563598633} 02/24/2022 16:42:45 - INFO - codeparrot_training - Step 11696: {'lr': 0.00045132697142778044, 'samples': 5988864, 'steps': 11696, 'loss/train': 2.0444397926330566} 02/24/2022 16:42:48 - INFO - codeparrot_training - Step 11697: {'lr': 0.00045131727039954137, 'samples': 5989376, 'steps': 11697, 'loss/train': 1.8476845026016235} 02/24/2022 16:42:54 - INFO - codeparrot_training - Step 11698: {'lr': 0.00045130756850892296, 'samples': 5989888, 'steps': 11698, 'loss/train': 1.385201096534729} 02/24/2022 16:42:57 - INFO - codeparrot_training - Step 11699: {'lr': 0.00045129786575596683, 'samples': 5990400, 'steps': 11699, 'loss/train': 2.6225171089172363} 02/24/2022 16:43:04 - INFO - codeparrot_training - Step 11700: {'lr': 0.00045128816214071453, 'samples': 5990912, 'steps': 11700, 'loss/train': 1.614296793937683} 02/24/2022 16:43:08 - INFO - codeparrot_training - Step 11701: {'lr': 0.00045127845766320773, 'samples': 5991424, 'steps': 11701, 'loss/train': 0.1792672872543335} 02/24/2022 16:43:13 - INFO - codeparrot_training - Step 11702: {'lr': 0.0004512687523234879, 'samples': 5991936, 'steps': 11702, 'loss/train': 2.3859634399414062} 02/24/2022 16:43:17 - INFO - codeparrot_training - Step 11703: {'lr': 0.0004512590461215967, 'samples': 5992448, 'steps': 11703, 'loss/train': 2.563322067260742} 02/24/2022 16:43:22 - INFO - codeparrot_training - Step 11704: {'lr': 0.0004512493390575756, 'samples': 5992960, 'steps': 11704, 'loss/train': 1.340606927871704} 02/24/2022 16:43:26 - INFO - codeparrot_training - Step 11705: {'lr': 0.0004512396311314662, 'samples': 5993472, 'steps': 11705, 'loss/train': 2.1104423999786377} 02/24/2022 16:43:31 - INFO - codeparrot_training - Step 11706: {'lr': 0.00045122992234331017, 'samples': 5993984, 'steps': 11706, 'loss/train': 2.493543863296509} 02/24/2022 16:43:35 - INFO - codeparrot_training - Step 11707: {'lr': 0.00045122021269314907, 'samples': 5994496, 'steps': 11707, 'loss/train': 2.080507278442383} 02/24/2022 16:43:40 - INFO - codeparrot_training - Step 11708: {'lr': 0.0004512105021810244, 'samples': 5995008, 'steps': 11708, 'loss/train': 2.3038547039031982} 02/24/2022 16:43:44 - INFO - codeparrot_training - Step 11709: {'lr': 0.0004512007908069779, 'samples': 5995520, 'steps': 11709, 'loss/train': 2.1975417137145996} 02/24/2022 16:43:51 - INFO - codeparrot_training - Step 11710: {'lr': 0.0004511910785710511, 'samples': 5996032, 'steps': 11710, 'loss/train': 1.5053011178970337} 02/24/2022 16:43:55 - INFO - codeparrot_training - Step 11711: {'lr': 0.0004511813654732856, 'samples': 5996544, 'steps': 11711, 'loss/train': 1.2798415422439575} 02/24/2022 16:44:00 - INFO - codeparrot_training - Step 11712: {'lr': 0.00045117165151372296, 'samples': 5997056, 'steps': 11712, 'loss/train': 2.2211251258850098} 02/24/2022 16:44:04 - INFO - codeparrot_training - Step 11713: {'lr': 0.0004511619366924049, 'samples': 5997568, 'steps': 11713, 'loss/train': 1.8240395784378052} 02/24/2022 16:44:09 - INFO - codeparrot_training - Step 11714: {'lr': 0.00045115222100937293, 'samples': 5998080, 'steps': 11714, 'loss/train': 1.8571007251739502} 02/24/2022 16:44:13 - INFO - codeparrot_training - Step 11715: {'lr': 0.00045114250446466874, 'samples': 5998592, 'steps': 11715, 'loss/train': 2.3949577808380127} 02/24/2022 16:44:18 - INFO - codeparrot_training - Step 11716: {'lr': 0.00045113278705833396, 'samples': 5999104, 'steps': 11716, 'loss/train': 2.4111225605010986} 02/24/2022 16:44:22 - INFO - codeparrot_training - Step 11717: {'lr': 0.00045112306879041016, 'samples': 5999616, 'steps': 11717, 'loss/train': 2.1929476261138916} 02/24/2022 16:44:27 - INFO - codeparrot_training - Step 11718: {'lr': 0.000451113349660939, 'samples': 6000128, 'steps': 11718, 'loss/train': 2.0388059616088867} 02/24/2022 16:44:31 - INFO - codeparrot_training - Step 11719: {'lr': 0.0004511036296699621, 'samples': 6000640, 'steps': 11719, 'loss/train': 1.0479477643966675} 02/24/2022 16:44:38 - INFO - codeparrot_training - Step 11720: {'lr': 0.0004510939088175211, 'samples': 6001152, 'steps': 11720, 'loss/train': 1.0591456890106201} 02/24/2022 16:44:41 - INFO - codeparrot_training - Step 11721: {'lr': 0.00045108418710365774, 'samples': 6001664, 'steps': 11721, 'loss/train': 1.9625349044799805} 02/24/2022 16:44:47 - INFO - codeparrot_training - Step 11722: {'lr': 0.0004510744645284135, 'samples': 6002176, 'steps': 11722, 'loss/train': 1.909144401550293} 02/24/2022 16:44:50 - INFO - codeparrot_training - Step 11723: {'lr': 0.00045106474109183004, 'samples': 6002688, 'steps': 11723, 'loss/train': 1.3695712089538574} 02/24/2022 16:44:56 - INFO - codeparrot_training - Step 11724: {'lr': 0.00045105501679394916, 'samples': 6003200, 'steps': 11724, 'loss/train': 1.601331353187561} 02/24/2022 16:45:00 - INFO - codeparrot_training - Step 11725: {'lr': 0.00045104529163481245, 'samples': 6003712, 'steps': 11725, 'loss/train': 1.130944848060608} 02/24/2022 16:45:05 - INFO - codeparrot_training - Step 11726: {'lr': 0.0004510355656144615, 'samples': 6004224, 'steps': 11726, 'loss/train': 1.0114940404891968} 02/24/2022 16:45:09 - INFO - codeparrot_training - Step 11727: {'lr': 0.000451025838732938, 'samples': 6004736, 'steps': 11727, 'loss/train': 0.564415454864502} 02/24/2022 16:45:14 - INFO - codeparrot_training - Step 11728: {'lr': 0.0004510161109902837, 'samples': 6005248, 'steps': 11728, 'loss/train': 2.2881507873535156} 02/24/2022 16:45:18 - INFO - codeparrot_training - Step 11729: {'lr': 0.00045100638238654013, 'samples': 6005760, 'steps': 11729, 'loss/train': 1.274651288986206} 02/24/2022 16:45:23 - INFO - codeparrot_training - Step 11730: {'lr': 0.00045099665292174917, 'samples': 6006272, 'steps': 11730, 'loss/train': 1.4472706317901611} 02/24/2022 16:45:27 - INFO - codeparrot_training - Step 11731: {'lr': 0.00045098692259595233, 'samples': 6006784, 'steps': 11731, 'loss/train': 0.3856137990951538} 02/24/2022 16:45:32 - INFO - codeparrot_training - Step 11732: {'lr': 0.00045097719140919126, 'samples': 6007296, 'steps': 11732, 'loss/train': 2.1976161003112793} 02/24/2022 16:45:36 - INFO - codeparrot_training - Step 11733: {'lr': 0.00045096745936150774, 'samples': 6007808, 'steps': 11733, 'loss/train': 0.322746604681015} 02/24/2022 16:45:41 - INFO - codeparrot_training - Step 11734: {'lr': 0.00045095772645294347, 'samples': 6008320, 'steps': 11734, 'loss/train': 0.7797043919563293} 02/24/2022 16:45:45 - INFO - codeparrot_training - Step 11735: {'lr': 0.00045094799268354007, 'samples': 6008832, 'steps': 11735, 'loss/train': 2.494370222091675} 02/24/2022 16:45:52 - INFO - codeparrot_training - Step 11736: {'lr': 0.00045093825805333934, 'samples': 6009344, 'steps': 11736, 'loss/train': 2.7397022247314453} 02/24/2022 16:45:56 - INFO - codeparrot_training - Step 11737: {'lr': 0.0004509285225623829, 'samples': 6009856, 'steps': 11737, 'loss/train': 1.7188767194747925} 02/24/2022 16:46:01 - INFO - codeparrot_training - Step 11738: {'lr': 0.0004509187862107125, 'samples': 6010368, 'steps': 11738, 'loss/train': 3.339857578277588} 02/24/2022 16:46:05 - INFO - codeparrot_training - Step 11739: {'lr': 0.0004509090489983697, 'samples': 6010880, 'steps': 11739, 'loss/train': 1.6962321996688843} 02/24/2022 16:46:10 - INFO - codeparrot_training - Step 11740: {'lr': 0.0004508993109253964, 'samples': 6011392, 'steps': 11740, 'loss/train': 1.9202880859375} 02/24/2022 16:46:14 - INFO - codeparrot_training - Step 11741: {'lr': 0.00045088957199183427, 'samples': 6011904, 'steps': 11741, 'loss/train': 1.823652982711792} 02/24/2022 16:46:19 - INFO - codeparrot_training - Step 11742: {'lr': 0.000450879832197725, 'samples': 6012416, 'steps': 11742, 'loss/train': 1.4651334285736084} 02/24/2022 16:46:23 - INFO - codeparrot_training - Step 11743: {'lr': 0.0004508700915431103, 'samples': 6012928, 'steps': 11743, 'loss/train': 1.7653381824493408} 02/24/2022 16:46:28 - INFO - codeparrot_training - Step 11744: {'lr': 0.0004508603500280319, 'samples': 6013440, 'steps': 11744, 'loss/train': 0.8583138585090637} 02/24/2022 16:46:32 - INFO - codeparrot_training - Step 11745: {'lr': 0.00045085060765253157, 'samples': 6013952, 'steps': 11745, 'loss/train': 2.611677646636963} 02/24/2022 16:46:39 - INFO - codeparrot_training - Step 11746: {'lr': 0.00045084086441665093, 'samples': 6014464, 'steps': 11746, 'loss/train': 1.703393578529358} 02/24/2022 16:46:43 - INFO - codeparrot_training - Step 11747: {'lr': 0.00045083112032043196, 'samples': 6014976, 'steps': 11747, 'loss/train': 0.5210812091827393} 02/24/2022 16:46:48 - INFO - codeparrot_training - Step 11748: {'lr': 0.0004508213753639161, 'samples': 6015488, 'steps': 11748, 'loss/train': 2.5765676498413086} 02/24/2022 16:46:52 - INFO - codeparrot_training - Step 11749: {'lr': 0.0004508116295471453, 'samples': 6016000, 'steps': 11749, 'loss/train': 1.7753199338912964} 02/24/2022 16:46:57 - INFO - codeparrot_training - Step 11750: {'lr': 0.0004508018828701612, 'samples': 6016512, 'steps': 11750, 'loss/train': 1.9259554147720337} 02/24/2022 16:47:01 - INFO - codeparrot_training - Step 11751: {'lr': 0.0004507921353330057, 'samples': 6017024, 'steps': 11751, 'loss/train': 2.5465378761291504} 02/24/2022 16:47:06 - INFO - codeparrot_training - Step 11752: {'lr': 0.0004507823869357204, 'samples': 6017536, 'steps': 11752, 'loss/train': 1.3102563619613647} 02/24/2022 16:47:10 - INFO - codeparrot_training - Step 11753: {'lr': 0.00045077263767834703, 'samples': 6018048, 'steps': 11753, 'loss/train': 1.4255242347717285} 02/24/2022 16:47:15 - INFO - codeparrot_training - Step 11754: {'lr': 0.00045076288756092754, 'samples': 6018560, 'steps': 11754, 'loss/train': 1.6492993831634521} 02/24/2022 16:47:19 - INFO - codeparrot_training - Step 11755: {'lr': 0.0004507531365835035, 'samples': 6019072, 'steps': 11755, 'loss/train': 2.3588688373565674} 02/24/2022 16:47:26 - INFO - codeparrot_training - Step 11756: {'lr': 0.00045074338474611683, 'samples': 6019584, 'steps': 11756, 'loss/train': 2.6841821670532227} 02/24/2022 16:47:30 - INFO - codeparrot_training - Step 11757: {'lr': 0.00045073363204880916, 'samples': 6020096, 'steps': 11757, 'loss/train': 2.7651474475860596} 02/24/2022 16:47:35 - INFO - codeparrot_training - Step 11758: {'lr': 0.0004507238784916224, 'samples': 6020608, 'steps': 11758, 'loss/train': 2.136057138442993} 02/24/2022 16:47:39 - INFO - codeparrot_training - Step 11759: {'lr': 0.0004507141240745983, 'samples': 6021120, 'steps': 11759, 'loss/train': 2.7878289222717285} 02/24/2022 16:47:44 - INFO - codeparrot_training - Step 11760: {'lr': 0.0004507043687977787, 'samples': 6021632, 'steps': 11760, 'loss/train': 2.0140068531036377} 02/24/2022 16:47:48 - INFO - codeparrot_training - Step 11761: {'lr': 0.00045069461266120515, 'samples': 6022144, 'steps': 11761, 'loss/train': 1.6322925090789795} 02/24/2022 16:47:53 - INFO - codeparrot_training - Step 11762: {'lr': 0.0004506848556649197, 'samples': 6022656, 'steps': 11762, 'loss/train': 2.4961063861846924} 02/24/2022 16:47:57 - INFO - codeparrot_training - Step 11763: {'lr': 0.0004506750978089641, 'samples': 6023168, 'steps': 11763, 'loss/train': 1.8917371034622192} 02/24/2022 16:48:02 - INFO - codeparrot_training - Step 11764: {'lr': 0.00045066533909338005, 'samples': 6023680, 'steps': 11764, 'loss/train': 1.9698625802993774} 02/24/2022 16:48:05 - INFO - codeparrot_training - Step 11765: {'lr': 0.00045065557951820935, 'samples': 6024192, 'steps': 11765, 'loss/train': 2.764829397201538} 02/24/2022 16:48:13 - INFO - codeparrot_training - Step 11766: {'lr': 0.0004506458190834939, 'samples': 6024704, 'steps': 11766, 'loss/train': 2.0404257774353027} 02/24/2022 16:48:16 - INFO - codeparrot_training - Step 11767: {'lr': 0.0004506360577892755, 'samples': 6025216, 'steps': 11767, 'loss/train': 0.7921394109725952} 02/24/2022 16:48:22 - INFO - codeparrot_training - Step 11768: {'lr': 0.00045062629563559595, 'samples': 6025728, 'steps': 11768, 'loss/train': 1.5200183391571045} 02/24/2022 16:48:25 - INFO - codeparrot_training - Step 11769: {'lr': 0.00045061653262249703, 'samples': 6026240, 'steps': 11769, 'loss/train': 1.6946516036987305} 02/24/2022 16:48:31 - INFO - codeparrot_training - Step 11770: {'lr': 0.0004506067687500206, 'samples': 6026752, 'steps': 11770, 'loss/train': 1.5203983783721924} 02/24/2022 16:48:34 - INFO - codeparrot_training - Step 11771: {'lr': 0.00045059700401820846, 'samples': 6027264, 'steps': 11771, 'loss/train': 2.541382074356079} 02/24/2022 16:48:40 - INFO - codeparrot_training - Step 11772: {'lr': 0.00045058723842710246, 'samples': 6027776, 'steps': 11772, 'loss/train': 2.290221929550171} 02/24/2022 16:48:43 - INFO - codeparrot_training - Step 11773: {'lr': 0.0004505774719767444, 'samples': 6028288, 'steps': 11773, 'loss/train': 1.734066367149353} 02/24/2022 16:48:49 - INFO - codeparrot_training - Step 11774: {'lr': 0.0004505677046671761, 'samples': 6028800, 'steps': 11774, 'loss/train': 2.921297311782837} 02/24/2022 16:48:52 - INFO - codeparrot_training - Step 11775: {'lr': 0.0004505579364984396, 'samples': 6029312, 'steps': 11775, 'loss/train': 2.769432544708252} 02/24/2022 16:49:00 - INFO - codeparrot_training - Step 11776: {'lr': 0.0004505481674705764, 'samples': 6029824, 'steps': 11776, 'loss/train': 2.7122838497161865} 02/24/2022 16:49:03 - INFO - codeparrot_training - Step 11777: {'lr': 0.0004505383975836286, 'samples': 6030336, 'steps': 11777, 'loss/train': 2.342007875442505} 02/24/2022 16:49:09 - INFO - codeparrot_training - Step 11778: {'lr': 0.00045052862683763806, 'samples': 6030848, 'steps': 11778, 'loss/train': 2.0701496601104736} 02/24/2022 16:49:12 - INFO - codeparrot_training - Step 11779: {'lr': 0.0004505188552326465, 'samples': 6031360, 'steps': 11779, 'loss/train': 1.4011969566345215} 02/24/2022 16:49:17 - INFO - codeparrot_training - Step 11780: {'lr': 0.00045050908276869585, 'samples': 6031872, 'steps': 11780, 'loss/train': 2.5028140544891357} 02/24/2022 16:49:21 - INFO - codeparrot_training - Step 11781: {'lr': 0.00045049930944582783, 'samples': 6032384, 'steps': 11781, 'loss/train': 2.1235408782958984} 02/24/2022 16:49:27 - INFO - codeparrot_training - Step 11782: {'lr': 0.0004504895352640846, 'samples': 6032896, 'steps': 11782, 'loss/train': 2.391179084777832} 02/24/2022 16:49:30 - INFO - codeparrot_training - Step 11783: {'lr': 0.0004504797602235078, 'samples': 6033408, 'steps': 11783, 'loss/train': 1.413794994354248} 02/24/2022 16:49:35 - INFO - codeparrot_training - Step 11784: {'lr': 0.0004504699843241394, 'samples': 6033920, 'steps': 11784, 'loss/train': 2.193089246749878} 02/24/2022 16:49:39 - INFO - codeparrot_training - Step 11785: {'lr': 0.0004504602075660212, 'samples': 6034432, 'steps': 11785, 'loss/train': 1.088712453842163} 02/24/2022 16:49:44 - INFO - codeparrot_training - Step 11786: {'lr': 0.00045045042994919514, 'samples': 6034944, 'steps': 11786, 'loss/train': 1.206525206565857} 02/24/2022 16:49:48 - INFO - codeparrot_training - Step 11787: {'lr': 0.00045044065147370303, 'samples': 6035456, 'steps': 11787, 'loss/train': 2.6903977394104004} 02/24/2022 16:49:54 - INFO - codeparrot_training - Step 11788: {'lr': 0.0004504308721395869, 'samples': 6035968, 'steps': 11788, 'loss/train': 2.8004629611968994} 02/24/2022 16:49:58 - INFO - codeparrot_training - Step 11789: {'lr': 0.0004504210919468886, 'samples': 6036480, 'steps': 11789, 'loss/train': 1.1341214179992676} 02/24/2022 16:50:03 - INFO - codeparrot_training - Step 11790: {'lr': 0.0004504113108956499, 'samples': 6036992, 'steps': 11790, 'loss/train': 1.790855050086975} 02/24/2022 16:50:07 - INFO - codeparrot_training - Step 11791: {'lr': 0.0004504015289859128, 'samples': 6037504, 'steps': 11791, 'loss/train': 2.1810033321380615} 02/24/2022 16:50:12 - INFO - codeparrot_training - Step 11792: {'lr': 0.00045039174621771915, 'samples': 6038016, 'steps': 11792, 'loss/train': 2.7970197200775146} 02/24/2022 16:50:16 - INFO - codeparrot_training - Step 11793: {'lr': 0.0004503819625911109, 'samples': 6038528, 'steps': 11793, 'loss/train': 1.2452048063278198} 02/24/2022 16:50:21 - INFO - codeparrot_training - Step 11794: {'lr': 0.00045037217810613004, 'samples': 6039040, 'steps': 11794, 'loss/train': 3.1736698150634766} 02/24/2022 16:50:25 - INFO - codeparrot_training - Step 11795: {'lr': 0.0004503623927628183, 'samples': 6039552, 'steps': 11795, 'loss/train': 1.923999309539795} 02/24/2022 16:50:30 - INFO - codeparrot_training - Step 11796: {'lr': 0.0004503526065612177, 'samples': 6040064, 'steps': 11796, 'loss/train': 1.8024576902389526} 02/24/2022 16:50:34 - INFO - codeparrot_training - Step 11797: {'lr': 0.0004503428195013702, 'samples': 6040576, 'steps': 11797, 'loss/train': 1.5015267133712769} 02/24/2022 16:50:40 - INFO - codeparrot_training - Step 11798: {'lr': 0.00045033303158331764, 'samples': 6041088, 'steps': 11798, 'loss/train': 2.173752546310425} 02/24/2022 16:50:43 - INFO - codeparrot_training - Step 11799: {'lr': 0.00045032324280710204, 'samples': 6041600, 'steps': 11799, 'loss/train': 2.2663638591766357} 02/24/2022 16:50:49 - INFO - codeparrot_training - Step 11800: {'lr': 0.0004503134531727652, 'samples': 6042112, 'steps': 11800, 'loss/train': 2.9252853393554688} 02/24/2022 16:50:52 - INFO - codeparrot_training - Step 11801: {'lr': 0.00045030366268034917, 'samples': 6042624, 'steps': 11801, 'loss/train': 1.0678006410598755} 02/24/2022 16:50:58 - INFO - codeparrot_training - Step 11802: {'lr': 0.00045029387132989587, 'samples': 6043136, 'steps': 11802, 'loss/train': 2.732165575027466} 02/24/2022 16:51:01 - INFO - codeparrot_training - Step 11803: {'lr': 0.0004502840791214472, 'samples': 6043648, 'steps': 11803, 'loss/train': 2.073514461517334} 02/24/2022 16:51:07 - INFO - codeparrot_training - Step 11804: {'lr': 0.00045027428605504507, 'samples': 6044160, 'steps': 11804, 'loss/train': 2.5620059967041016} 02/24/2022 16:51:10 - INFO - codeparrot_training - Step 11805: {'lr': 0.00045026449213073154, 'samples': 6044672, 'steps': 11805, 'loss/train': 1.0992964506149292} 02/24/2022 16:51:16 - INFO - codeparrot_training - Step 11806: {'lr': 0.00045025469734854856, 'samples': 6045184, 'steps': 11806, 'loss/train': 1.7111836671829224} 02/24/2022 16:51:19 - INFO - codeparrot_training - Step 11807: {'lr': 0.00045024490170853806, 'samples': 6045696, 'steps': 11807, 'loss/train': 3.2235193252563477} 02/24/2022 16:51:26 - INFO - codeparrot_training - Step 11808: {'lr': 0.000450235105210742, 'samples': 6046208, 'steps': 11808, 'loss/train': 2.294288158416748} 02/24/2022 16:51:30 - INFO - codeparrot_training - Step 11809: {'lr': 0.0004502253078552022, 'samples': 6046720, 'steps': 11809, 'loss/train': 2.357346534729004} 02/24/2022 16:51:35 - INFO - codeparrot_training - Step 11810: {'lr': 0.00045021550964196086, 'samples': 6047232, 'steps': 11810, 'loss/train': 2.650636911392212} 02/24/2022 16:51:39 - INFO - codeparrot_training - Step 11811: {'lr': 0.0004502057105710598, 'samples': 6047744, 'steps': 11811, 'loss/train': 2.0565428733825684} 02/24/2022 16:51:44 - INFO - codeparrot_training - Step 11812: {'lr': 0.00045019591064254105, 'samples': 6048256, 'steps': 11812, 'loss/train': 1.7709914445877075} 02/24/2022 16:51:48 - INFO - codeparrot_training - Step 11813: {'lr': 0.00045018610985644663, 'samples': 6048768, 'steps': 11813, 'loss/train': 2.0230753421783447} 02/24/2022 16:51:53 - INFO - codeparrot_training - Step 11814: {'lr': 0.00045017630821281854, 'samples': 6049280, 'steps': 11814, 'loss/train': 2.387709379196167} 02/24/2022 16:51:57 - INFO - codeparrot_training - Step 11815: {'lr': 0.0004501665057116986, 'samples': 6049792, 'steps': 11815, 'loss/train': 2.8525302410125732} 02/24/2022 16:52:02 - INFO - codeparrot_training - Step 11816: {'lr': 0.00045015670235312895, 'samples': 6050304, 'steps': 11816, 'loss/train': 3.3121023178100586} 02/24/2022 16:52:06 - INFO - codeparrot_training - Step 11817: {'lr': 0.00045014689813715147, 'samples': 6050816, 'steps': 11817, 'loss/train': 2.3521053791046143} 02/24/2022 16:52:12 - INFO - codeparrot_training - Step 11818: {'lr': 0.00045013709306380837, 'samples': 6051328, 'steps': 11818, 'loss/train': 2.2122199535369873} 02/24/2022 16:52:15 - INFO - codeparrot_training - Step 11819: {'lr': 0.00045012728713314146, 'samples': 6051840, 'steps': 11819, 'loss/train': 1.578689455986023} 02/24/2022 16:52:21 - INFO - codeparrot_training - Step 11820: {'lr': 0.00045011748034519275, 'samples': 6052352, 'steps': 11820, 'loss/train': 1.8257839679718018} 02/24/2022 16:52:24 - INFO - codeparrot_training - Step 11821: {'lr': 0.00045010767270000436, 'samples': 6052864, 'steps': 11821, 'loss/train': 1.7844138145446777} 02/24/2022 16:52:30 - INFO - codeparrot_training - Step 11822: {'lr': 0.00045009786419761825, 'samples': 6053376, 'steps': 11822, 'loss/train': 2.06961727142334} 02/24/2022 16:52:33 - INFO - codeparrot_training - Step 11823: {'lr': 0.00045008805483807637, 'samples': 6053888, 'steps': 11823, 'loss/train': 2.0493924617767334} 02/24/2022 16:52:39 - INFO - codeparrot_training - Step 11824: {'lr': 0.0004500782446214208, 'samples': 6054400, 'steps': 11824, 'loss/train': 1.7303874492645264} 02/24/2022 16:52:42 - INFO - codeparrot_training - Step 11825: {'lr': 0.00045006843354769354, 'samples': 6054912, 'steps': 11825, 'loss/train': 1.4131265878677368} 02/24/2022 16:52:48 - INFO - codeparrot_training - Step 11826: {'lr': 0.0004500586216169367, 'samples': 6055424, 'steps': 11826, 'loss/train': 2.13592267036438} 02/24/2022 16:52:51 - INFO - codeparrot_training - Step 11827: {'lr': 0.0004500488088291923, 'samples': 6055936, 'steps': 11827, 'loss/train': 2.2689785957336426} 02/24/2022 16:52:57 - INFO - codeparrot_training - Step 11828: {'lr': 0.0004500389951845022, 'samples': 6056448, 'steps': 11828, 'loss/train': 2.451796293258667} 02/24/2022 16:53:00 - INFO - codeparrot_training - Step 11829: {'lr': 0.00045002918068290864, 'samples': 6056960, 'steps': 11829, 'loss/train': 2.7076683044433594} 02/24/2022 16:53:06 - INFO - codeparrot_training - Step 11830: {'lr': 0.00045001936532445354, 'samples': 6057472, 'steps': 11830, 'loss/train': 2.344712495803833} 02/24/2022 16:53:09 - INFO - codeparrot_training - Step 11831: {'lr': 0.000450009549109179, 'samples': 6057984, 'steps': 11831, 'loss/train': 0.39552760124206543} 02/24/2022 16:53:15 - INFO - codeparrot_training - Step 11832: {'lr': 0.0004499997320371271, 'samples': 6058496, 'steps': 11832, 'loss/train': 0.9307349324226379} 02/24/2022 16:53:18 - INFO - codeparrot_training - Step 11833: {'lr': 0.0004499899141083399, 'samples': 6059008, 'steps': 11833, 'loss/train': 1.6942479610443115} 02/24/2022 16:53:25 - INFO - codeparrot_training - Step 11834: {'lr': 0.0004499800953228593, 'samples': 6059520, 'steps': 11834, 'loss/train': 1.9799652099609375} 02/24/2022 16:53:28 - INFO - codeparrot_training - Step 11835: {'lr': 0.00044997027568072754, 'samples': 6060032, 'steps': 11835, 'loss/train': 3.412726640701294} 02/24/2022 16:53:34 - INFO - codeparrot_training - Step 11836: {'lr': 0.00044996045518198657, 'samples': 6060544, 'steps': 11836, 'loss/train': 1.8916040658950806} 02/24/2022 16:53:37 - INFO - codeparrot_training - Step 11837: {'lr': 0.00044995063382667855, 'samples': 6061056, 'steps': 11837, 'loss/train': 3.093398332595825} 02/24/2022 16:53:43 - INFO - codeparrot_training - Step 11838: {'lr': 0.0004499408116148455, 'samples': 6061568, 'steps': 11838, 'loss/train': 1.9078476428985596} 02/24/2022 16:53:48 - INFO - codeparrot_training - Step 11839: {'lr': 0.00044993098854652954, 'samples': 6062080, 'steps': 11839, 'loss/train': 2.662196636199951} 02/24/2022 16:53:52 - INFO - codeparrot_training - Step 11840: {'lr': 0.0004499211646217727, 'samples': 6062592, 'steps': 11840, 'loss/train': 2.1113712787628174} 02/24/2022 16:53:57 - INFO - codeparrot_training - Step 11841: {'lr': 0.000449911339840617, 'samples': 6063104, 'steps': 11841, 'loss/train': 2.349804162979126} 02/24/2022 16:54:01 - INFO - codeparrot_training - Step 11842: {'lr': 0.00044990151420310463, 'samples': 6063616, 'steps': 11842, 'loss/train': 1.6631113290786743} 02/24/2022 16:54:08 - INFO - codeparrot_training - Step 11843: {'lr': 0.0004498916877092776, 'samples': 6064128, 'steps': 11843, 'loss/train': 2.216902256011963} 02/24/2022 16:54:11 - INFO - codeparrot_training - Step 11844: {'lr': 0.00044988186035917817, 'samples': 6064640, 'steps': 11844, 'loss/train': 2.565882444381714} 02/24/2022 16:54:16 - INFO - codeparrot_training - Step 11845: {'lr': 0.00044987203215284823, 'samples': 6065152, 'steps': 11845, 'loss/train': 3.2680232524871826} 02/24/2022 16:54:20 - INFO - codeparrot_training - Step 11846: {'lr': 0.00044986220309033, 'samples': 6065664, 'steps': 11846, 'loss/train': 1.5452680587768555} 02/24/2022 16:54:25 - INFO - codeparrot_training - Step 11847: {'lr': 0.00044985237317166554, 'samples': 6066176, 'steps': 11847, 'loss/train': 2.617960214614868} 02/24/2022 16:54:29 - INFO - codeparrot_training - Step 11848: {'lr': 0.00044984254239689703, 'samples': 6066688, 'steps': 11848, 'loss/train': 2.240201950073242} 02/24/2022 16:54:34 - INFO - codeparrot_training - Step 11849: {'lr': 0.00044983271076606644, 'samples': 6067200, 'steps': 11849, 'loss/train': 1.4940757751464844} 02/24/2022 16:54:38 - INFO - codeparrot_training - Step 11850: {'lr': 0.000449822878279216, 'samples': 6067712, 'steps': 11850, 'loss/train': 2.268136501312256} 02/24/2022 16:54:43 - INFO - codeparrot_training - Step 11851: {'lr': 0.00044981304493638786, 'samples': 6068224, 'steps': 11851, 'loss/train': 0.21787181496620178} 02/24/2022 16:54:47 - INFO - codeparrot_training - Step 11852: {'lr': 0.00044980321073762405, 'samples': 6068736, 'steps': 11852, 'loss/train': 1.5555155277252197} 02/24/2022 16:54:52 - INFO - codeparrot_training - Step 11853: {'lr': 0.0004497933756829667, 'samples': 6069248, 'steps': 11853, 'loss/train': 0.5362102389335632} 02/24/2022 16:54:56 - INFO - codeparrot_training - Step 11854: {'lr': 0.000449783539772458, 'samples': 6069760, 'steps': 11854, 'loss/train': 8.661615371704102} 02/24/2022 16:55:02 - INFO - codeparrot_training - Step 11855: {'lr': 0.00044977370300614, 'samples': 6070272, 'steps': 11855, 'loss/train': 2.184626340866089} 02/24/2022 16:55:06 - INFO - codeparrot_training - Step 11856: {'lr': 0.00044976386538405494, 'samples': 6070784, 'steps': 11856, 'loss/train': 1.9198890924453735} 02/24/2022 16:55:11 - INFO - codeparrot_training - Step 11857: {'lr': 0.0004497540269062449, 'samples': 6071296, 'steps': 11857, 'loss/train': 1.743076205253601} 02/24/2022 16:55:15 - INFO - codeparrot_training - Step 11858: {'lr': 0.00044974418757275206, 'samples': 6071808, 'steps': 11858, 'loss/train': 1.8038883209228516} 02/24/2022 16:55:20 - INFO - codeparrot_training - Step 11859: {'lr': 0.00044973434738361853, 'samples': 6072320, 'steps': 11859, 'loss/train': 1.8507227897644043} 02/24/2022 16:55:24 - INFO - codeparrot_training - Step 11860: {'lr': 0.0004497245063388865, 'samples': 6072832, 'steps': 11860, 'loss/train': 1.4610706567764282} 02/24/2022 16:55:29 - INFO - codeparrot_training - Step 11861: {'lr': 0.0004497146644385981, 'samples': 6073344, 'steps': 11861, 'loss/train': 1.0989198684692383} 02/24/2022 16:55:33 - INFO - codeparrot_training - Step 11862: {'lr': 0.00044970482168279547, 'samples': 6073856, 'steps': 11862, 'loss/train': 1.9074040651321411} 02/24/2022 16:55:38 - INFO - codeparrot_training - Step 11863: {'lr': 0.0004496949780715208, 'samples': 6074368, 'steps': 11863, 'loss/train': 1.4850995540618896} 02/24/2022 16:55:41 - INFO - codeparrot_training - Step 11864: {'lr': 0.00044968513360481624, 'samples': 6074880, 'steps': 11864, 'loss/train': 2.1204400062561035} 02/24/2022 16:55:48 - INFO - codeparrot_training - Step 11865: {'lr': 0.000449675288282724, 'samples': 6075392, 'steps': 11865, 'loss/train': 2.1871776580810547} 02/24/2022 16:55:51 - INFO - codeparrot_training - Step 11866: {'lr': 0.0004496654421052862, 'samples': 6075904, 'steps': 11866, 'loss/train': 2.3699865341186523} 02/24/2022 16:55:57 - INFO - codeparrot_training - Step 11867: {'lr': 0.00044965559507254504, 'samples': 6076416, 'steps': 11867, 'loss/train': 2.8026418685913086} 02/24/2022 16:56:00 - INFO - codeparrot_training - Step 11868: {'lr': 0.0004496457471845428, 'samples': 6076928, 'steps': 11868, 'loss/train': 1.7902960777282715} 02/24/2022 16:56:06 - INFO - codeparrot_training - Step 11869: {'lr': 0.0004496358984413215, 'samples': 6077440, 'steps': 11869, 'loss/train': 2.4285483360290527} 02/24/2022 16:56:09 - INFO - codeparrot_training - Step 11870: {'lr': 0.0004496260488429234, 'samples': 6077952, 'steps': 11870, 'loss/train': 1.9883067607879639} 02/24/2022 16:56:15 - INFO - codeparrot_training - Step 11871: {'lr': 0.0004496161983893907, 'samples': 6078464, 'steps': 11871, 'loss/train': 2.715402364730835} 02/24/2022 16:56:18 - INFO - codeparrot_training - Step 11872: {'lr': 0.0004496063470807656, 'samples': 6078976, 'steps': 11872, 'loss/train': 2.8464128971099854} 02/24/2022 16:56:24 - INFO - codeparrot_training - Step 11873: {'lr': 0.0004495964949170903, 'samples': 6079488, 'steps': 11873, 'loss/train': 2.8089919090270996} 02/24/2022 16:56:27 - INFO - codeparrot_training - Step 11874: {'lr': 0.000449586641898407, 'samples': 6080000, 'steps': 11874, 'loss/train': 1.1081875562667847} 02/24/2022 16:56:34 - INFO - codeparrot_training - Step 11875: {'lr': 0.0004495767880247579, 'samples': 6080512, 'steps': 11875, 'loss/train': 3.8264901638031006} 02/24/2022 16:56:37 - INFO - codeparrot_training - Step 11876: {'lr': 0.0004495669332961852, 'samples': 6081024, 'steps': 11876, 'loss/train': 2.336467981338501} 02/24/2022 16:56:43 - INFO - codeparrot_training - Step 11877: {'lr': 0.0004495570777127311, 'samples': 6081536, 'steps': 11877, 'loss/train': 1.2555092573165894} 02/24/2022 16:56:46 - INFO - codeparrot_training - Step 11878: {'lr': 0.00044954722127443786, 'samples': 6082048, 'steps': 11878, 'loss/train': 2.372148036956787} 02/24/2022 16:56:52 - INFO - codeparrot_training - Step 11879: {'lr': 0.0004495373639813477, 'samples': 6082560, 'steps': 11879, 'loss/train': 2.5146970748901367} 02/24/2022 16:56:55 - INFO - codeparrot_training - Step 11880: {'lr': 0.00044952750583350287, 'samples': 6083072, 'steps': 11880, 'loss/train': 2.2158963680267334} 02/24/2022 16:57:01 - INFO - codeparrot_training - Step 11881: {'lr': 0.00044951764683094555, 'samples': 6083584, 'steps': 11881, 'loss/train': 2.6528873443603516} 02/24/2022 16:57:04 - INFO - codeparrot_training - Step 11882: {'lr': 0.000449507786973718, 'samples': 6084096, 'steps': 11882, 'loss/train': 2.0729784965515137} 02/24/2022 16:57:10 - INFO - codeparrot_training - Step 11883: {'lr': 0.0004494979262618624, 'samples': 6084608, 'steps': 11883, 'loss/train': 1.4883427619934082} 02/24/2022 16:57:13 - INFO - codeparrot_training - Step 11884: {'lr': 0.00044948806469542095, 'samples': 6085120, 'steps': 11884, 'loss/train': 1.772378921508789} 02/24/2022 16:57:19 - INFO - codeparrot_training - Step 11885: {'lr': 0.0004494782022744361, 'samples': 6085632, 'steps': 11885, 'loss/train': 1.4385184049606323} 02/24/2022 16:57:22 - INFO - codeparrot_training - Step 11886: {'lr': 0.0004494683389989499, 'samples': 6086144, 'steps': 11886, 'loss/train': 1.2901983261108398} 02/24/2022 16:57:28 - INFO - codeparrot_training - Step 11887: {'lr': 0.0004494584748690047, 'samples': 6086656, 'steps': 11887, 'loss/train': 2.284496784210205} 02/24/2022 16:57:31 - INFO - codeparrot_training - Step 11888: {'lr': 0.00044944860988464276, 'samples': 6087168, 'steps': 11888, 'loss/train': 1.466820240020752} 02/24/2022 16:57:38 - INFO - codeparrot_training - Step 11889: {'lr': 0.0004494387440459063, 'samples': 6087680, 'steps': 11889, 'loss/train': 1.2328132390975952} 02/24/2022 16:57:41 - INFO - codeparrot_training - Step 11890: {'lr': 0.00044942887735283755, 'samples': 6088192, 'steps': 11890, 'loss/train': 3.2682878971099854} 02/24/2022 16:57:47 - INFO - codeparrot_training - Step 11891: {'lr': 0.00044941900980547886, 'samples': 6088704, 'steps': 11891, 'loss/train': 2.589184284210205} 02/24/2022 16:57:50 - INFO - codeparrot_training - Step 11892: {'lr': 0.00044940914140387245, 'samples': 6089216, 'steps': 11892, 'loss/train': 1.2583006620407104} 02/24/2022 16:57:56 - INFO - codeparrot_training - Step 11893: {'lr': 0.00044939927214806055, 'samples': 6089728, 'steps': 11893, 'loss/train': 1.39590322971344} 02/24/2022 16:57:59 - INFO - codeparrot_training - Step 11894: {'lr': 0.0004493894020380855, 'samples': 6090240, 'steps': 11894, 'loss/train': 1.5713070631027222} 02/24/2022 16:58:03 - INFO - codeparrot_training - Step 11895: {'lr': 0.0004493795310739896, 'samples': 6090752, 'steps': 11895, 'loss/train': 1.9005963802337646} 02/24/2022 16:58:08 - INFO - codeparrot_training - Step 11896: {'lr': 0.00044936965925581506, 'samples': 6091264, 'steps': 11896, 'loss/train': 1.2607002258300781} 02/24/2022 16:58:12 - INFO - codeparrot_training - Step 11897: {'lr': 0.0004493597865836042, 'samples': 6091776, 'steps': 11897, 'loss/train': 1.736268401145935} 02/24/2022 16:58:17 - INFO - codeparrot_training - Step 11898: {'lr': 0.00044934991305739936, 'samples': 6092288, 'steps': 11898, 'loss/train': 2.399876832962036} 02/24/2022 16:58:21 - INFO - codeparrot_training - Step 11899: {'lr': 0.00044934003867724284, 'samples': 6092800, 'steps': 11899, 'loss/train': 2.5512542724609375} 02/24/2022 16:58:26 - INFO - codeparrot_training - Step 11900: {'lr': 0.0004493301634431768, 'samples': 6093312, 'steps': 11900, 'loss/train': 0.8661245703697205} 02/24/2022 16:58:30 - INFO - codeparrot_training - Step 11901: {'lr': 0.00044932028735524367, 'samples': 6093824, 'steps': 11901, 'loss/train': 1.3039920330047607} 02/24/2022 16:58:36 - INFO - codeparrot_training - Step 11902: {'lr': 0.0004493104104134857, 'samples': 6094336, 'steps': 11902, 'loss/train': 2.297304391860962} 02/24/2022 16:58:39 - INFO - codeparrot_training - Step 11903: {'lr': 0.0004493005326179452, 'samples': 6094848, 'steps': 11903, 'loss/train': 2.4173314571380615} 02/24/2022 16:58:45 - INFO - codeparrot_training - Step 11904: {'lr': 0.00044929065396866457, 'samples': 6095360, 'steps': 11904, 'loss/train': 2.2927234172821045} 02/24/2022 16:58:48 - INFO - codeparrot_training - Step 11905: {'lr': 0.00044928077446568606, 'samples': 6095872, 'steps': 11905, 'loss/train': 2.6560232639312744} 02/24/2022 16:58:54 - INFO - codeparrot_training - Step 11906: {'lr': 0.000449270894109052, 'samples': 6096384, 'steps': 11906, 'loss/train': 2.1722636222839355} 02/24/2022 16:58:57 - INFO - codeparrot_training - Step 11907: {'lr': 0.0004492610128988046, 'samples': 6096896, 'steps': 11907, 'loss/train': 1.3685529232025146} 02/24/2022 16:59:03 - INFO - codeparrot_training - Step 11908: {'lr': 0.00044925113083498636, 'samples': 6097408, 'steps': 11908, 'loss/train': 1.2973147630691528} 02/24/2022 16:59:08 - INFO - codeparrot_training - Step 11909: {'lr': 0.00044924124791763956, 'samples': 6097920, 'steps': 11909, 'loss/train': 3.2404911518096924} 02/24/2022 16:59:12 - INFO - codeparrot_training - Step 11910: {'lr': 0.0004492313641468065, 'samples': 6098432, 'steps': 11910, 'loss/train': 1.7464429140090942} 02/24/2022 16:59:18 - INFO - codeparrot_training - Step 11911: {'lr': 0.00044922147952252957, 'samples': 6098944, 'steps': 11911, 'loss/train': 0.7987387180328369} 02/24/2022 16:59:22 - INFO - codeparrot_training - Step 11912: {'lr': 0.000449211594044851, 'samples': 6099456, 'steps': 11912, 'loss/train': 2.8895716667175293} 02/24/2022 16:59:27 - INFO - codeparrot_training - Step 11913: {'lr': 0.0004492017077138133, 'samples': 6099968, 'steps': 11913, 'loss/train': 1.7488881349563599} 02/24/2022 16:59:31 - INFO - codeparrot_training - Step 11914: {'lr': 0.00044919182052945866, 'samples': 6100480, 'steps': 11914, 'loss/train': 1.0501044988632202} 02/24/2022 16:59:36 - INFO - codeparrot_training - Step 11915: {'lr': 0.00044918193249182957, 'samples': 6100992, 'steps': 11915, 'loss/train': 1.9240261316299438} 02/24/2022 16:59:40 - INFO - codeparrot_training - Step 11916: {'lr': 0.0004491720436009683, 'samples': 6101504, 'steps': 11916, 'loss/train': 1.406111478805542} 02/24/2022 16:59:45 - INFO - codeparrot_training - Step 11917: {'lr': 0.0004491621538569173, 'samples': 6102016, 'steps': 11917, 'loss/train': 0.8316056728363037} 02/24/2022 16:59:49 - INFO - codeparrot_training - Step 11918: {'lr': 0.0004491522632597188, 'samples': 6102528, 'steps': 11918, 'loss/train': 1.0837879180908203} 02/24/2022 16:59:54 - INFO - codeparrot_training - Step 11919: {'lr': 0.0004491423718094153, 'samples': 6103040, 'steps': 11919, 'loss/train': 1.7377420663833618} 02/24/2022 16:59:58 - INFO - codeparrot_training - Step 11920: {'lr': 0.00044913247950604905, 'samples': 6103552, 'steps': 11920, 'loss/train': 0.8596081137657166} 02/24/2022 17:00:03 - INFO - codeparrot_training - Step 11921: {'lr': 0.0004491225863496625, 'samples': 6104064, 'steps': 11921, 'loss/train': 1.740216851234436} 02/24/2022 17:00:07 - INFO - codeparrot_training - Step 11922: {'lr': 0.0004491126923402981, 'samples': 6104576, 'steps': 11922, 'loss/train': 1.1672195196151733} 02/24/2022 17:00:12 - INFO - codeparrot_training - Step 11923: {'lr': 0.0004491027974779981, 'samples': 6105088, 'steps': 11923, 'loss/train': 1.5681185722351074} 02/24/2022 17:00:16 - INFO - codeparrot_training - Step 11924: {'lr': 0.00044909290176280495, 'samples': 6105600, 'steps': 11924, 'loss/train': 2.9294259548187256} 02/24/2022 17:00:21 - INFO - codeparrot_training - Step 11925: {'lr': 0.000449083005194761, 'samples': 6106112, 'steps': 11925, 'loss/train': 1.2542585134506226} 02/24/2022 17:00:25 - INFO - codeparrot_training - Step 11926: {'lr': 0.0004490731077739087, 'samples': 6106624, 'steps': 11926, 'loss/train': 2.0052225589752197} 02/24/2022 17:00:31 - INFO - codeparrot_training - Step 11927: {'lr': 0.0004490632095002904, 'samples': 6107136, 'steps': 11927, 'loss/train': 2.3806068897247314} 02/24/2022 17:00:35 - INFO - codeparrot_training - Step 11928: {'lr': 0.00044905331037394853, 'samples': 6107648, 'steps': 11928, 'loss/train': 2.7983317375183105} 02/24/2022 17:00:40 - INFO - codeparrot_training - Step 11929: {'lr': 0.00044904341039492544, 'samples': 6108160, 'steps': 11929, 'loss/train': 0.13732248544692993} 02/24/2022 17:00:44 - INFO - codeparrot_training - Step 11930: {'lr': 0.00044903350956326365, 'samples': 6108672, 'steps': 11930, 'loss/train': 2.2320008277893066} 02/24/2022 17:00:49 - INFO - codeparrot_training - Step 11931: {'lr': 0.0004490236078790055, 'samples': 6109184, 'steps': 11931, 'loss/train': 2.022023916244507} 02/24/2022 17:00:53 - INFO - codeparrot_training - Step 11932: {'lr': 0.0004490137053421934, 'samples': 6109696, 'steps': 11932, 'loss/train': 1.7930134534835815} 02/24/2022 17:00:58 - INFO - codeparrot_training - Step 11933: {'lr': 0.00044900380195286974, 'samples': 6110208, 'steps': 11933, 'loss/train': 1.9943904876708984} 02/24/2022 17:01:02 - INFO - codeparrot_training - Step 11934: {'lr': 0.00044899389771107704, 'samples': 6110720, 'steps': 11934, 'loss/train': 1.4732900857925415} 02/24/2022 17:01:07 - INFO - codeparrot_training - Step 11935: {'lr': 0.00044898399261685765, 'samples': 6111232, 'steps': 11935, 'loss/train': 1.4811468124389648} 02/24/2022 17:01:10 - INFO - codeparrot_training - Step 11936: {'lr': 0.00044897408667025397, 'samples': 6111744, 'steps': 11936, 'loss/train': 1.9996992349624634} 02/24/2022 17:01:17 - INFO - codeparrot_training - Step 11937: {'lr': 0.00044896417987130854, 'samples': 6112256, 'steps': 11937, 'loss/train': 1.7018797397613525} 02/24/2022 17:01:20 - INFO - codeparrot_training - Step 11938: {'lr': 0.0004489542722200637, 'samples': 6112768, 'steps': 11938, 'loss/train': 2.303635358810425} 02/24/2022 17:01:26 - INFO - codeparrot_training - Step 11939: {'lr': 0.000448944363716562, 'samples': 6113280, 'steps': 11939, 'loss/train': 2.693312644958496} 02/24/2022 17:01:29 - INFO - codeparrot_training - Step 11940: {'lr': 0.0004489344543608458, 'samples': 6113792, 'steps': 11940, 'loss/train': 1.2922580242156982} 02/24/2022 17:01:35 - INFO - codeparrot_training - Step 11941: {'lr': 0.00044892454415295746, 'samples': 6114304, 'steps': 11941, 'loss/train': 2.4177403450012207} 02/24/2022 17:01:39 - INFO - codeparrot_training - Step 11942: {'lr': 0.0004489146330929397, 'samples': 6114816, 'steps': 11942, 'loss/train': 2.755218267440796} 02/24/2022 17:01:42 - INFO - codeparrot_training - Step 11943: {'lr': 0.0004489047211808347, 'samples': 6115328, 'steps': 11943, 'loss/train': 1.957213282585144} 02/24/2022 17:01:48 - INFO - codeparrot_training - Step 11944: {'lr': 0.0004488948084166851, 'samples': 6115840, 'steps': 11944, 'loss/train': 2.2804155349731445} 02/24/2022 17:01:51 - INFO - codeparrot_training - Step 11945: {'lr': 0.00044888489480053324, 'samples': 6116352, 'steps': 11945, 'loss/train': 2.0375568866729736} 02/24/2022 17:01:57 - INFO - codeparrot_training - Step 11946: {'lr': 0.00044887498033242167, 'samples': 6116864, 'steps': 11946, 'loss/train': 1.9059584140777588} 02/24/2022 17:02:00 - INFO - codeparrot_training - Step 11947: {'lr': 0.0004488650650123929, 'samples': 6117376, 'steps': 11947, 'loss/train': 1.1402106285095215} 02/24/2022 17:02:06 - INFO - codeparrot_training - Step 11948: {'lr': 0.00044885514884048926, 'samples': 6117888, 'steps': 11948, 'loss/train': 2.292330026626587} 02/24/2022 17:02:10 - INFO - codeparrot_training - Step 11949: {'lr': 0.0004488452318167533, 'samples': 6118400, 'steps': 11949, 'loss/train': 2.6049935817718506} 02/24/2022 17:02:15 - INFO - codeparrot_training - Step 11950: {'lr': 0.00044883531394122753, 'samples': 6118912, 'steps': 11950, 'loss/train': 0.7761608958244324} 02/24/2022 17:02:19 - INFO - codeparrot_training - Step 11951: {'lr': 0.00044882539521395436, 'samples': 6119424, 'steps': 11951, 'loss/train': 2.873300075531006} 02/24/2022 17:02:24 - INFO - codeparrot_training - Step 11952: {'lr': 0.0004488154756349764, 'samples': 6119936, 'steps': 11952, 'loss/train': 4.269147872924805} 02/24/2022 17:02:28 - INFO - codeparrot_training - Step 11953: {'lr': 0.0004488055552043361, 'samples': 6120448, 'steps': 11953, 'loss/train': 2.1729979515075684} 02/24/2022 17:02:33 - INFO - codeparrot_training - Step 11954: {'lr': 0.0004487956339220759, 'samples': 6120960, 'steps': 11954, 'loss/train': 0.8693559765815735} 02/24/2022 17:02:37 - INFO - codeparrot_training - Step 11955: {'lr': 0.00044878571178823826, 'samples': 6121472, 'steps': 11955, 'loss/train': 1.254451870918274} 02/24/2022 17:02:42 - INFO - codeparrot_training - Step 11956: {'lr': 0.00044877578880286585, 'samples': 6121984, 'steps': 11956, 'loss/train': 0.19196289777755737} 02/24/2022 17:02:46 - INFO - codeparrot_training - Step 11957: {'lr': 0.000448765864966001, 'samples': 6122496, 'steps': 11957, 'loss/train': 2.527200698852539} 02/24/2022 17:02:52 - INFO - codeparrot_training - Step 11958: {'lr': 0.00044875594027768634, 'samples': 6123008, 'steps': 11958, 'loss/train': 1.9800655841827393} 02/24/2022 17:02:56 - INFO - codeparrot_training - Step 11959: {'lr': 0.00044874601473796435, 'samples': 6123520, 'steps': 11959, 'loss/train': 0.3376052975654602} 02/24/2022 17:03:01 - INFO - codeparrot_training - Step 11960: {'lr': 0.00044873608834687754, 'samples': 6124032, 'steps': 11960, 'loss/train': 2.2521958351135254} 02/24/2022 17:03:04 - INFO - codeparrot_training - Step 11961: {'lr': 0.0004487261611044684, 'samples': 6124544, 'steps': 11961, 'loss/train': 1.5861146450042725} 02/24/2022 17:03:10 - INFO - codeparrot_training - Step 11962: {'lr': 0.0004487162330107795, 'samples': 6125056, 'steps': 11962, 'loss/train': 2.6860108375549316} 02/24/2022 17:03:14 - INFO - codeparrot_training - Step 11963: {'lr': 0.0004487063040658534, 'samples': 6125568, 'steps': 11963, 'loss/train': 2.2602968215942383} 02/24/2022 17:03:19 - INFO - codeparrot_training - Step 11964: {'lr': 0.00044869637426973256, 'samples': 6126080, 'steps': 11964, 'loss/train': 2.0055363178253174} 02/24/2022 17:03:23 - INFO - codeparrot_training - Step 11965: {'lr': 0.0004486864436224595, 'samples': 6126592, 'steps': 11965, 'loss/train': 1.4304295778274536} 02/24/2022 17:03:29 - INFO - codeparrot_training - Step 11966: {'lr': 0.0004486765121240769, 'samples': 6127104, 'steps': 11966, 'loss/train': 2.6333274841308594} 02/24/2022 17:03:32 - INFO - codeparrot_training - Step 11967: {'lr': 0.0004486665797746271, 'samples': 6127616, 'steps': 11967, 'loss/train': 0.22088871896266937} 02/24/2022 17:03:38 - INFO - codeparrot_training - Step 11968: {'lr': 0.00044865664657415286, 'samples': 6128128, 'steps': 11968, 'loss/train': 2.322220802307129} 02/24/2022 17:03:42 - INFO - codeparrot_training - Step 11969: {'lr': 0.00044864671252269663, 'samples': 6128640, 'steps': 11969, 'loss/train': 2.567811965942383} 02/24/2022 17:03:47 - INFO - codeparrot_training - Step 11970: {'lr': 0.00044863677762030087, 'samples': 6129152, 'steps': 11970, 'loss/train': 1.8674544095993042} 02/24/2022 17:03:51 - INFO - codeparrot_training - Step 11971: {'lr': 0.0004486268418670083, 'samples': 6129664, 'steps': 11971, 'loss/train': 1.2539787292480469} 02/24/2022 17:03:56 - INFO - codeparrot_training - Step 11972: {'lr': 0.00044861690526286135, 'samples': 6130176, 'steps': 11972, 'loss/train': 2.8972015380859375} 02/24/2022 17:04:00 - INFO - codeparrot_training - Step 11973: {'lr': 0.00044860696780790266, 'samples': 6130688, 'steps': 11973, 'loss/train': 1.7852134704589844} 02/24/2022 17:04:06 - INFO - codeparrot_training - Step 11974: {'lr': 0.00044859702950217486, 'samples': 6131200, 'steps': 11974, 'loss/train': 3.0962789058685303} 02/24/2022 17:04:09 - INFO - codeparrot_training - Step 11975: {'lr': 0.00044858709034572035, 'samples': 6131712, 'steps': 11975, 'loss/train': 2.5287649631500244} 02/24/2022 17:04:15 - INFO - codeparrot_training - Step 11976: {'lr': 0.00044857715033858183, 'samples': 6132224, 'steps': 11976, 'loss/train': 2.1950368881225586} 02/24/2022 17:04:18 - INFO - codeparrot_training - Step 11977: {'lr': 0.0004485672094808019, 'samples': 6132736, 'steps': 11977, 'loss/train': 1.9684149026870728} 02/24/2022 17:04:24 - INFO - codeparrot_training - Step 11978: {'lr': 0.0004485572677724231, 'samples': 6133248, 'steps': 11978, 'loss/train': 1.745532512664795} 02/24/2022 17:04:27 - INFO - codeparrot_training - Step 11979: {'lr': 0.00044854732521348796, 'samples': 6133760, 'steps': 11979, 'loss/train': 1.880918025970459} 02/24/2022 17:04:33 - INFO - codeparrot_training - Step 11980: {'lr': 0.0004485373818040391, 'samples': 6134272, 'steps': 11980, 'loss/train': 1.9257597923278809} 02/24/2022 17:04:36 - INFO - codeparrot_training - Step 11981: {'lr': 0.00044852743754411915, 'samples': 6134784, 'steps': 11981, 'loss/train': 1.7359751462936401} 02/24/2022 17:04:42 - INFO - codeparrot_training - Step 11982: {'lr': 0.00044851749243377085, 'samples': 6135296, 'steps': 11982, 'loss/train': 1.5540597438812256} 02/24/2022 17:04:45 - INFO - codeparrot_training - Step 11983: {'lr': 0.0004485075464730365, 'samples': 6135808, 'steps': 11983, 'loss/train': 1.6910945177078247} 02/24/2022 17:04:51 - INFO - codeparrot_training - Step 11984: {'lr': 0.0004484975996619589, 'samples': 6136320, 'steps': 11984, 'loss/train': 0.38395506143569946} 02/24/2022 17:04:55 - INFO - codeparrot_training - Step 11985: {'lr': 0.0004484876520005805, 'samples': 6136832, 'steps': 11985, 'loss/train': 1.7679054737091064} 02/24/2022 17:05:01 - INFO - codeparrot_training - Step 11986: {'lr': 0.0004484777034889441, 'samples': 6137344, 'steps': 11986, 'loss/train': 2.9316844940185547} 02/24/2022 17:05:04 - INFO - codeparrot_training - Step 11987: {'lr': 0.0004484677541270923, 'samples': 6137856, 'steps': 11987, 'loss/train': 1.8936922550201416} 02/24/2022 17:05:10 - INFO - codeparrot_training - Step 11988: {'lr': 0.00044845780391506763, 'samples': 6138368, 'steps': 11988, 'loss/train': 3.8764872550964355} 02/24/2022 17:05:13 - INFO - codeparrot_training - Step 11989: {'lr': 0.0004484478528529128, 'samples': 6138880, 'steps': 11989, 'loss/train': 2.5315418243408203} 02/24/2022 17:05:19 - INFO - codeparrot_training - Step 11990: {'lr': 0.00044843790094067026, 'samples': 6139392, 'steps': 11990, 'loss/train': 1.575141191482544} 02/24/2022 17:05:23 - INFO - codeparrot_training - Step 11991: {'lr': 0.00044842794817838286, 'samples': 6139904, 'steps': 11991, 'loss/train': 2.0063159465789795} 02/24/2022 17:05:26 - INFO - codeparrot_training - Step 11992: {'lr': 0.0004484179945660931, 'samples': 6140416, 'steps': 11992, 'loss/train': 1.4903126955032349} 02/24/2022 17:05:32 - INFO - codeparrot_training - Step 11993: {'lr': 0.00044840804010384366, 'samples': 6140928, 'steps': 11993, 'loss/train': 2.07969069480896} 02/24/2022 17:05:35 - INFO - codeparrot_training - Step 11994: {'lr': 0.00044839808479167723, 'samples': 6141440, 'steps': 11994, 'loss/train': 2.9866795539855957} 02/24/2022 17:05:41 - INFO - codeparrot_training - Step 11995: {'lr': 0.00044838812862963627, 'samples': 6141952, 'steps': 11995, 'loss/train': 2.889941453933716} 02/24/2022 17:05:47 - INFO - codeparrot_training - Step 11996: {'lr': 0.00044837817161776366, 'samples': 6142464, 'steps': 11996, 'loss/train': 1.9506334066390991} 02/24/2022 17:05:51 - INFO - codeparrot_training - Step 11997: {'lr': 0.00044836821375610194, 'samples': 6142976, 'steps': 11997, 'loss/train': 1.876620888710022} 02/24/2022 17:05:56 - INFO - codeparrot_training - Step 11998: {'lr': 0.0004483582550446938, 'samples': 6143488, 'steps': 11998, 'loss/train': 1.6544023752212524} 02/24/2022 17:06:00 - INFO - codeparrot_training - Step 11999: {'lr': 0.0004483482954835819, 'samples': 6144000, 'steps': 11999, 'loss/train': 1.3312416076660156} 02/24/2022 17:06:00 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 17:06:17 - WARNING - huggingface_hub.repository - Several commits (12) will be pushed upstream. 02/24/2022 17:06:17 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 17:06:51 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 3d9f7da..e455b12 floral-grass-11 -> floral-grass-11 02/24/2022 17:06:57 - INFO - codeparrot_training - Step 12000: {'lr': 0.0004483383350728088, 'samples': 6144512, 'steps': 12000, 'loss/train': 1.8612650632858276} 02/24/2022 17:07:00 - INFO - codeparrot_training - Step 12001: {'lr': 0.00044832837381241733, 'samples': 6145024, 'steps': 12001, 'loss/train': 2.081125020980835} 02/24/2022 17:07:06 - INFO - codeparrot_training - Step 12002: {'lr': 0.00044831841170245003, 'samples': 6145536, 'steps': 12002, 'loss/train': 3.2302181720733643} 02/24/2022 17:07:09 - INFO - codeparrot_training - Step 12003: {'lr': 0.0004483084487429496, 'samples': 6146048, 'steps': 12003, 'loss/train': 0.415947288274765} 02/24/2022 17:07:16 - INFO - codeparrot_training - Step 12004: {'lr': 0.00044829848493395884, 'samples': 6146560, 'steps': 12004, 'loss/train': 1.6646977663040161} 02/24/2022 17:07:19 - INFO - codeparrot_training - Step 12005: {'lr': 0.00044828852027552023, 'samples': 6147072, 'steps': 12005, 'loss/train': 2.7473230361938477} 02/24/2022 17:07:25 - INFO - codeparrot_training - Step 12006: {'lr': 0.00044827855476767665, 'samples': 6147584, 'steps': 12006, 'loss/train': 2.021182060241699} 02/24/2022 17:07:28 - INFO - codeparrot_training - Step 12007: {'lr': 0.00044826858841047067, 'samples': 6148096, 'steps': 12007, 'loss/train': 2.6072206497192383} 02/24/2022 17:07:34 - INFO - codeparrot_training - Step 12008: {'lr': 0.00044825862120394504, 'samples': 6148608, 'steps': 12008, 'loss/train': 1.318609356880188} 02/24/2022 17:07:37 - INFO - codeparrot_training - Step 12009: {'lr': 0.00044824865314814245, 'samples': 6149120, 'steps': 12009, 'loss/train': 1.8420289754867554} 02/24/2022 17:07:43 - INFO - codeparrot_training - Step 12010: {'lr': 0.00044823868424310553, 'samples': 6149632, 'steps': 12010, 'loss/train': 2.6879255771636963} 02/24/2022 17:07:46 - INFO - codeparrot_training - Step 12011: {'lr': 0.00044822871448887703, 'samples': 6150144, 'steps': 12011, 'loss/train': 1.9901670217514038} 02/24/2022 17:07:52 - INFO - codeparrot_training - Step 12012: {'lr': 0.0004482187438854997, 'samples': 6150656, 'steps': 12012, 'loss/train': 2.8354454040527344} 02/24/2022 17:07:55 - INFO - codeparrot_training - Step 12013: {'lr': 0.00044820877243301617, 'samples': 6151168, 'steps': 12013, 'loss/train': 2.2613120079040527} 02/24/2022 17:07:59 - INFO - codeparrot_training - Step 12014: {'lr': 0.00044819880013146924, 'samples': 6151680, 'steps': 12014, 'loss/train': 1.8379782438278198} 02/24/2022 17:08:05 - INFO - codeparrot_training - Step 12015: {'lr': 0.0004481888269809016, 'samples': 6152192, 'steps': 12015, 'loss/train': 2.1393845081329346} 02/24/2022 17:08:08 - INFO - codeparrot_training - Step 12016: {'lr': 0.0004481788529813559, 'samples': 6152704, 'steps': 12016, 'loss/train': 2.1670098304748535} 02/24/2022 17:08:14 - INFO - codeparrot_training - Step 12017: {'lr': 0.00044816887813287494, 'samples': 6153216, 'steps': 12017, 'loss/train': 2.503826856613159} 02/24/2022 17:08:19 - INFO - codeparrot_training - Step 12018: {'lr': 0.0004481589024355014, 'samples': 6153728, 'steps': 12018, 'loss/train': 1.9746654033660889} 02/24/2022 17:08:23 - INFO - codeparrot_training - Step 12019: {'lr': 0.00044814892588927816, 'samples': 6154240, 'steps': 12019, 'loss/train': 3.7525477409362793} 02/24/2022 17:08:29 - INFO - codeparrot_training - Step 12020: {'lr': 0.00044813894849424777, 'samples': 6154752, 'steps': 12020, 'loss/train': 2.1124234199523926} 02/24/2022 17:08:32 - INFO - codeparrot_training - Step 12021: {'lr': 0.00044812897025045295, 'samples': 6155264, 'steps': 12021, 'loss/train': 3.585268974304199} 02/24/2022 17:08:38 - INFO - codeparrot_training - Step 12022: {'lr': 0.00044811899115793666, 'samples': 6155776, 'steps': 12022, 'loss/train': 1.7901548147201538} 02/24/2022 17:08:41 - INFO - codeparrot_training - Step 12023: {'lr': 0.0004481090112167415, 'samples': 6156288, 'steps': 12023, 'loss/train': 1.735459566116333} 02/24/2022 17:08:47 - INFO - codeparrot_training - Step 12024: {'lr': 0.0004480990304269102, 'samples': 6156800, 'steps': 12024, 'loss/train': 1.8389549255371094} 02/24/2022 17:08:50 - INFO - codeparrot_training - Step 12025: {'lr': 0.00044808904878848555, 'samples': 6157312, 'steps': 12025, 'loss/train': 2.4541585445404053} 02/24/2022 17:08:56 - INFO - codeparrot_training - Step 12026: {'lr': 0.00044807906630151033, 'samples': 6157824, 'steps': 12026, 'loss/train': 1.9353373050689697} 02/24/2022 17:08:59 - INFO - codeparrot_training - Step 12027: {'lr': 0.00044806908296602733, 'samples': 6158336, 'steps': 12027, 'loss/train': 1.9195120334625244} 02/24/2022 17:09:05 - INFO - codeparrot_training - Step 12028: {'lr': 0.0004480590987820793, 'samples': 6158848, 'steps': 12028, 'loss/train': 2.1406238079071045} 02/24/2022 17:09:08 - INFO - codeparrot_training - Step 12029: {'lr': 0.00044804911374970893, 'samples': 6159360, 'steps': 12029, 'loss/train': 2.5416066646575928} 02/24/2022 17:09:15 - INFO - codeparrot_training - Step 12030: {'lr': 0.000448039127868959, 'samples': 6159872, 'steps': 12030, 'loss/train': 6.636228561401367} 02/24/2022 17:09:18 - INFO - codeparrot_training - Step 12031: {'lr': 0.0004480291411398724, 'samples': 6160384, 'steps': 12031, 'loss/train': 2.637648344039917} 02/24/2022 17:09:24 - INFO - codeparrot_training - Step 12032: {'lr': 0.0004480191535624918, 'samples': 6160896, 'steps': 12032, 'loss/train': 1.6071592569351196} 02/24/2022 17:09:27 - INFO - codeparrot_training - Step 12033: {'lr': 0.00044800916513686, 'samples': 6161408, 'steps': 12033, 'loss/train': 1.6605761051177979} 02/24/2022 17:09:33 - INFO - codeparrot_training - Step 12034: {'lr': 0.00044799917586301987, 'samples': 6161920, 'steps': 12034, 'loss/train': 2.0364322662353516} 02/24/2022 17:09:36 - INFO - codeparrot_training - Step 12035: {'lr': 0.00044798918574101413, 'samples': 6162432, 'steps': 12035, 'loss/train': 2.449662446975708} 02/24/2022 17:09:42 - INFO - codeparrot_training - Step 12036: {'lr': 0.00044797919477088555, 'samples': 6162944, 'steps': 12036, 'loss/train': 2.2326977252960205} 02/24/2022 17:09:45 - INFO - codeparrot_training - Step 12037: {'lr': 0.00044796920295267696, 'samples': 6163456, 'steps': 12037, 'loss/train': 1.6525148153305054} 02/24/2022 17:09:50 - INFO - codeparrot_training - Step 12038: {'lr': 0.0004479592102864313, 'samples': 6163968, 'steps': 12038, 'loss/train': 2.1086108684539795} 02/24/2022 17:09:54 - INFO - codeparrot_training - Step 12039: {'lr': 0.0004479492167721911, 'samples': 6164480, 'steps': 12039, 'loss/train': 1.9792639017105103} 02/24/2022 17:10:00 - INFO - codeparrot_training - Step 12040: {'lr': 0.0004479392224099993, 'samples': 6164992, 'steps': 12040, 'loss/train': 2.3902857303619385} 02/24/2022 17:10:04 - INFO - codeparrot_training - Step 12041: {'lr': 0.00044792922719989883, 'samples': 6165504, 'steps': 12041, 'loss/train': 2.0644795894622803} 02/24/2022 17:10:09 - INFO - codeparrot_training - Step 12042: {'lr': 0.00044791923114193233, 'samples': 6166016, 'steps': 12042, 'loss/train': 2.6241061687469482} 02/24/2022 17:10:13 - INFO - codeparrot_training - Step 12043: {'lr': 0.0004479092342361427, 'samples': 6166528, 'steps': 12043, 'loss/train': 1.2237296104431152} 02/24/2022 17:10:18 - INFO - codeparrot_training - Step 12044: {'lr': 0.0004478992364825728, 'samples': 6167040, 'steps': 12044, 'loss/train': 2.6538925170898438} 02/24/2022 17:10:22 - INFO - codeparrot_training - Step 12045: {'lr': 0.00044788923788126534, 'samples': 6167552, 'steps': 12045, 'loss/train': 2.507875442504883} 02/24/2022 17:10:27 - INFO - codeparrot_training - Step 12046: {'lr': 0.00044787923843226323, 'samples': 6168064, 'steps': 12046, 'loss/train': 2.967284917831421} 02/24/2022 17:10:31 - INFO - codeparrot_training - Step 12047: {'lr': 0.0004478692381356093, 'samples': 6168576, 'steps': 12047, 'loss/train': 2.0878546237945557} 02/24/2022 17:10:36 - INFO - codeparrot_training - Step 12048: {'lr': 0.00044785923699134646, 'samples': 6169088, 'steps': 12048, 'loss/train': 1.0714280605316162} 02/24/2022 17:10:40 - INFO - codeparrot_training - Step 12049: {'lr': 0.0004478492349995174, 'samples': 6169600, 'steps': 12049, 'loss/train': 3.110184907913208} 02/24/2022 17:10:45 - INFO - codeparrot_training - Step 12050: {'lr': 0.00044783923216016507, 'samples': 6170112, 'steps': 12050, 'loss/train': 2.5074799060821533} 02/24/2022 17:10:49 - INFO - codeparrot_training - Step 12051: {'lr': 0.0004478292284733323, 'samples': 6170624, 'steps': 12051, 'loss/train': 1.9328501224517822} 02/24/2022 17:10:54 - INFO - codeparrot_training - Step 12052: {'lr': 0.00044781922393906186, 'samples': 6171136, 'steps': 12052, 'loss/train': 0.18410548567771912} 02/24/2022 17:10:58 - INFO - codeparrot_training - Step 12053: {'lr': 0.00044780921855739676, 'samples': 6171648, 'steps': 12053, 'loss/train': 2.585103988647461} 02/24/2022 17:11:04 - INFO - codeparrot_training - Step 12054: {'lr': 0.00044779921232837973, 'samples': 6172160, 'steps': 12054, 'loss/train': 2.492974281311035} 02/24/2022 17:11:07 - INFO - codeparrot_training - Step 12055: {'lr': 0.0004477892052520537, 'samples': 6172672, 'steps': 12055, 'loss/train': 1.8947025537490845} 02/24/2022 17:11:13 - INFO - codeparrot_training - Step 12056: {'lr': 0.0004477791973284616, 'samples': 6173184, 'steps': 12056, 'loss/train': 2.6554815769195557} 02/24/2022 17:11:17 - INFO - codeparrot_training - Step 12057: {'lr': 0.00044776918855764616, 'samples': 6173696, 'steps': 12057, 'loss/train': 0.9762059450149536} 02/24/2022 17:11:22 - INFO - codeparrot_training - Step 12058: {'lr': 0.00044775917893965025, 'samples': 6174208, 'steps': 12058, 'loss/train': 1.3967198133468628} 02/24/2022 17:11:26 - INFO - codeparrot_training - Step 12059: {'lr': 0.00044774916847451683, 'samples': 6174720, 'steps': 12059, 'loss/train': 2.5747334957122803} 02/24/2022 17:11:31 - INFO - codeparrot_training - Step 12060: {'lr': 0.0004477391571622889, 'samples': 6175232, 'steps': 12060, 'loss/train': 2.2955400943756104} 02/24/2022 17:11:35 - INFO - codeparrot_training - Step 12061: {'lr': 0.00044772914500300907, 'samples': 6175744, 'steps': 12061, 'loss/train': 1.6977120637893677} 02/24/2022 17:11:41 - INFO - codeparrot_training - Step 12062: {'lr': 0.0004477191319967204, 'samples': 6176256, 'steps': 12062, 'loss/train': 1.8222932815551758} 02/24/2022 17:11:44 - INFO - codeparrot_training - Step 12063: {'lr': 0.0004477091181434658, 'samples': 6176768, 'steps': 12063, 'loss/train': 1.5017566680908203} 02/24/2022 17:11:50 - INFO - codeparrot_training - Step 12064: {'lr': 0.00044769910344328803, 'samples': 6177280, 'steps': 12064, 'loss/train': 2.077258825302124} 02/24/2022 17:11:53 - INFO - codeparrot_training - Step 12065: {'lr': 0.00044768908789623015, 'samples': 6177792, 'steps': 12065, 'loss/train': 2.035762071609497} 02/24/2022 17:12:00 - INFO - codeparrot_training - Step 12066: {'lr': 0.00044767907150233496, 'samples': 6178304, 'steps': 12066, 'loss/train': 1.5400667190551758} 02/24/2022 17:12:03 - INFO - codeparrot_training - Step 12067: {'lr': 0.0004476690542616454, 'samples': 6178816, 'steps': 12067, 'loss/train': 3.015357255935669} 02/24/2022 17:12:09 - INFO - codeparrot_training - Step 12068: {'lr': 0.00044765903617420436, 'samples': 6179328, 'steps': 12068, 'loss/train': 1.8078761100769043} 02/24/2022 17:12:12 - INFO - codeparrot_training - Step 12069: {'lr': 0.0004476490172400548, 'samples': 6179840, 'steps': 12069, 'loss/train': 2.3099827766418457} 02/24/2022 17:12:17 - INFO - codeparrot_training - Step 12070: {'lr': 0.00044763899745923965, 'samples': 6180352, 'steps': 12070, 'loss/train': 2.4852898120880127} 02/24/2022 17:12:21 - INFO - codeparrot_training - Step 12071: {'lr': 0.0004476289768318017, 'samples': 6180864, 'steps': 12071, 'loss/train': 3.00781512260437} 02/24/2022 17:12:27 - INFO - codeparrot_training - Step 12072: {'lr': 0.00044761895535778404, 'samples': 6181376, 'steps': 12072, 'loss/train': 1.6103401184082031} 02/24/2022 17:12:30 - INFO - codeparrot_training - Step 12073: {'lr': 0.0004476089330372295, 'samples': 6181888, 'steps': 12073, 'loss/train': 2.1514294147491455} 02/24/2022 17:12:35 - INFO - codeparrot_training - Step 12074: {'lr': 0.00044759890987018105, 'samples': 6182400, 'steps': 12074, 'loss/train': 3.306278705596924} 02/24/2022 17:12:39 - INFO - codeparrot_training - Step 12075: {'lr': 0.0004475888858566816, 'samples': 6182912, 'steps': 12075, 'loss/train': 1.6765960454940796} 02/24/2022 17:12:45 - INFO - codeparrot_training - Step 12076: {'lr': 0.00044757886099677416, 'samples': 6183424, 'steps': 12076, 'loss/train': 1.4008985757827759} 02/24/2022 17:12:49 - INFO - codeparrot_training - Step 12077: {'lr': 0.0004475688352905015, 'samples': 6183936, 'steps': 12077, 'loss/train': 2.1884684562683105} 02/24/2022 17:12:54 - INFO - codeparrot_training - Step 12078: {'lr': 0.00044755880873790675, 'samples': 6184448, 'steps': 12078, 'loss/train': 1.8656758069992065} 02/24/2022 17:12:58 - INFO - codeparrot_training - Step 12079: {'lr': 0.00044754878133903284, 'samples': 6184960, 'steps': 12079, 'loss/train': 1.9137187004089355} 02/24/2022 17:13:03 - INFO - codeparrot_training - Step 12080: {'lr': 0.0004475387530939226, 'samples': 6185472, 'steps': 12080, 'loss/train': 2.4631569385528564} 02/24/2022 17:13:07 - INFO - codeparrot_training - Step 12081: {'lr': 0.00044752872400261913, 'samples': 6185984, 'steps': 12081, 'loss/train': 1.814349889755249} 02/24/2022 17:13:12 - INFO - codeparrot_training - Step 12082: {'lr': 0.0004475186940651653, 'samples': 6186496, 'steps': 12082, 'loss/train': 1.066323161125183} 02/24/2022 17:13:16 - INFO - codeparrot_training - Step 12083: {'lr': 0.0004475086632816041, 'samples': 6187008, 'steps': 12083, 'loss/train': 1.6021504402160645} 02/24/2022 17:13:21 - INFO - codeparrot_training - Step 12084: {'lr': 0.00044749863165197845, 'samples': 6187520, 'steps': 12084, 'loss/train': 1.9285125732421875} 02/24/2022 17:13:25 - INFO - codeparrot_training - Step 12085: {'lr': 0.00044748859917633144, 'samples': 6188032, 'steps': 12085, 'loss/train': 2.1447486877441406} 02/24/2022 17:13:31 - INFO - codeparrot_training - Step 12086: {'lr': 0.00044747856585470604, 'samples': 6188544, 'steps': 12086, 'loss/train': 1.8635483980178833} 02/24/2022 17:13:34 - INFO - codeparrot_training - Step 12087: {'lr': 0.00044746853168714507, 'samples': 6189056, 'steps': 12087, 'loss/train': 2.179267168045044} 02/24/2022 17:13:40 - INFO - codeparrot_training - Step 12088: {'lr': 0.0004474584966736917, 'samples': 6189568, 'steps': 12088, 'loss/train': 2.1392414569854736} 02/24/2022 17:13:43 - INFO - codeparrot_training - Step 12089: {'lr': 0.00044744846081438874, 'samples': 6190080, 'steps': 12089, 'loss/train': 1.8742955923080444} 02/24/2022 17:13:49 - INFO - codeparrot_training - Step 12090: {'lr': 0.0004474384241092793, 'samples': 6190592, 'steps': 12090, 'loss/train': 2.2031991481781006} 02/24/2022 17:13:52 - INFO - codeparrot_training - Step 12091: {'lr': 0.00044742838655840636, 'samples': 6191104, 'steps': 12091, 'loss/train': 2.4598612785339355} 02/24/2022 17:13:58 - INFO - codeparrot_training - Step 12092: {'lr': 0.0004474183481618129, 'samples': 6191616, 'steps': 12092, 'loss/train': 2.594998359680176} 02/24/2022 17:14:02 - INFO - codeparrot_training - Step 12093: {'lr': 0.00044740830891954196, 'samples': 6192128, 'steps': 12093, 'loss/train': 2.505722761154175} 02/24/2022 17:14:07 - INFO - codeparrot_training - Step 12094: {'lr': 0.0004473982688316365, 'samples': 6192640, 'steps': 12094, 'loss/train': 0.6227912902832031} 02/24/2022 17:14:11 - INFO - codeparrot_training - Step 12095: {'lr': 0.0004473882278981395, 'samples': 6193152, 'steps': 12095, 'loss/train': 2.19868540763855} 02/24/2022 17:14:17 - INFO - codeparrot_training - Step 12096: {'lr': 0.000447378186119094, 'samples': 6193664, 'steps': 12096, 'loss/train': 0.2705650329589844} 02/24/2022 17:14:21 - INFO - codeparrot_training - Step 12097: {'lr': 0.00044736814349454303, 'samples': 6194176, 'steps': 12097, 'loss/train': 1.6578989028930664} 02/24/2022 17:14:26 - INFO - codeparrot_training - Step 12098: {'lr': 0.0004473581000245296, 'samples': 6194688, 'steps': 12098, 'loss/train': 1.0228785276412964} 02/24/2022 17:14:30 - INFO - codeparrot_training - Step 12099: {'lr': 0.00044734805570909676, 'samples': 6195200, 'steps': 12099, 'loss/train': 2.7728471755981445} 02/24/2022 17:14:35 - INFO - codeparrot_training - Step 12100: {'lr': 0.0004473380105482875, 'samples': 6195712, 'steps': 12100, 'loss/train': 0.7916803956031799} 02/24/2022 17:14:39 - INFO - codeparrot_training - Step 12101: {'lr': 0.0004473279645421449, 'samples': 6196224, 'steps': 12101, 'loss/train': 3.0350024700164795} 02/24/2022 17:14:44 - INFO - codeparrot_training - Step 12102: {'lr': 0.00044731791769071197, 'samples': 6196736, 'steps': 12102, 'loss/train': 1.2722065448760986} 02/24/2022 17:14:48 - INFO - codeparrot_training - Step 12103: {'lr': 0.00044730786999403166, 'samples': 6197248, 'steps': 12103, 'loss/train': 0.8307028412818909} 02/24/2022 17:14:53 - INFO - codeparrot_training - Step 12104: {'lr': 0.00044729782145214717, 'samples': 6197760, 'steps': 12104, 'loss/train': 2.607999324798584} 02/24/2022 17:14:56 - INFO - codeparrot_training - Step 12105: {'lr': 0.0004472877720651014, 'samples': 6198272, 'steps': 12105, 'loss/train': 2.4205119609832764} 02/24/2022 17:15:02 - INFO - codeparrot_training - Step 12106: {'lr': 0.0004472777218329375, 'samples': 6198784, 'steps': 12106, 'loss/train': 2.9563732147216797} 02/24/2022 17:15:06 - INFO - codeparrot_training - Step 12107: {'lr': 0.00044726767075569843, 'samples': 6199296, 'steps': 12107, 'loss/train': 0.15740205347537994} 02/24/2022 17:15:11 - INFO - codeparrot_training - Step 12108: {'lr': 0.0004472576188334273, 'samples': 6199808, 'steps': 12108, 'loss/train': 2.1738076210021973} 02/24/2022 17:15:15 - INFO - codeparrot_training - Step 12109: {'lr': 0.00044724756606616726, 'samples': 6200320, 'steps': 12109, 'loss/train': 2.113445520401001} 02/24/2022 17:15:20 - INFO - codeparrot_training - Step 12110: {'lr': 0.00044723751245396117, 'samples': 6200832, 'steps': 12110, 'loss/train': 1.9405531883239746} 02/24/2022 17:15:24 - INFO - codeparrot_training - Step 12111: {'lr': 0.00044722745799685227, 'samples': 6201344, 'steps': 12111, 'loss/train': 1.2801152467727661} 02/24/2022 17:15:30 - INFO - codeparrot_training - Step 12112: {'lr': 0.00044721740269488354, 'samples': 6201856, 'steps': 12112, 'loss/train': 2.6905722618103027} 02/24/2022 17:15:33 - INFO - codeparrot_training - Step 12113: {'lr': 0.0004472073465480981, 'samples': 6202368, 'steps': 12113, 'loss/train': 1.1298489570617676} 02/24/2022 17:15:39 - INFO - codeparrot_training - Step 12114: {'lr': 0.000447197289556539, 'samples': 6202880, 'steps': 12114, 'loss/train': 2.1920905113220215} 02/24/2022 17:15:42 - INFO - codeparrot_training - Step 12115: {'lr': 0.0004471872317202493, 'samples': 6203392, 'steps': 12115, 'loss/train': 1.6994199752807617} 02/24/2022 17:15:48 - INFO - codeparrot_training - Step 12116: {'lr': 0.0004471771730392722, 'samples': 6203904, 'steps': 12116, 'loss/train': 1.4567934274673462} 02/24/2022 17:15:51 - INFO - codeparrot_training - Step 12117: {'lr': 0.00044716711351365057, 'samples': 6204416, 'steps': 12117, 'loss/train': 2.0766303539276123} 02/24/2022 17:15:57 - INFO - codeparrot_training - Step 12118: {'lr': 0.00044715705314342776, 'samples': 6204928, 'steps': 12118, 'loss/train': 1.7800449132919312} 02/24/2022 17:16:00 - INFO - codeparrot_training - Step 12119: {'lr': 0.0004471469919286467, 'samples': 6205440, 'steps': 12119, 'loss/train': 2.729935646057129} 02/24/2022 17:16:06 - INFO - codeparrot_training - Step 12120: {'lr': 0.0004471369298693505, 'samples': 6205952, 'steps': 12120, 'loss/train': 2.7959799766540527} 02/24/2022 17:16:09 - INFO - codeparrot_training - Step 12121: {'lr': 0.0004471268669655822, 'samples': 6206464, 'steps': 12121, 'loss/train': 1.8632409572601318} 02/24/2022 17:16:16 - INFO - codeparrot_training - Step 12122: {'lr': 0.0004471168032173852, 'samples': 6206976, 'steps': 12122, 'loss/train': 1.7287883758544922} 02/24/2022 17:16:19 - INFO - codeparrot_training - Step 12123: {'lr': 0.0004471067386248023, 'samples': 6207488, 'steps': 12123, 'loss/train': 1.9525649547576904} 02/24/2022 17:16:25 - INFO - codeparrot_training - Step 12124: {'lr': 0.0004470966731878767, 'samples': 6208000, 'steps': 12124, 'loss/train': 2.1880342960357666} 02/24/2022 17:16:28 - INFO - codeparrot_training - Step 12125: {'lr': 0.0004470866069066516, 'samples': 6208512, 'steps': 12125, 'loss/train': 1.3273388147354126} 02/24/2022 17:16:34 - INFO - codeparrot_training - Step 12126: {'lr': 0.00044707653978117004, 'samples': 6209024, 'steps': 12126, 'loss/train': 2.819064140319824} 02/24/2022 17:16:37 - INFO - codeparrot_training - Step 12127: {'lr': 0.00044706647181147507, 'samples': 6209536, 'steps': 12127, 'loss/train': 3.3440170288085938} 02/24/2022 17:16:43 - INFO - codeparrot_training - Step 12128: {'lr': 0.00044705640299761004, 'samples': 6210048, 'steps': 12128, 'loss/train': 2.1373674869537354} 02/24/2022 17:16:46 - INFO - codeparrot_training - Step 12129: {'lr': 0.0004470463333396179, 'samples': 6210560, 'steps': 12129, 'loss/train': 1.064770221710205} 02/24/2022 17:16:52 - INFO - codeparrot_training - Step 12130: {'lr': 0.0004470362628375418, 'samples': 6211072, 'steps': 12130, 'loss/train': 1.7912788391113281} 02/24/2022 17:16:55 - INFO - codeparrot_training - Step 12131: {'lr': 0.000447026191491425, 'samples': 6211584, 'steps': 12131, 'loss/train': 2.478670597076416} 02/24/2022 17:17:01 - INFO - codeparrot_training - Step 12132: {'lr': 0.0004470161193013105, 'samples': 6212096, 'steps': 12132, 'loss/train': 2.443627119064331} 02/24/2022 17:17:04 - INFO - codeparrot_training - Step 12133: {'lr': 0.0004470060462672415, 'samples': 6212608, 'steps': 12133, 'loss/train': 2.4356517791748047} 02/24/2022 17:17:10 - INFO - codeparrot_training - Step 12134: {'lr': 0.0004469959723892612, 'samples': 6213120, 'steps': 12134, 'loss/train': 2.313127279281616} 02/24/2022 17:17:16 - INFO - codeparrot_training - Step 12135: {'lr': 0.0004469858976674126, 'samples': 6213632, 'steps': 12135, 'loss/train': 1.4179664850234985} 02/24/2022 17:17:19 - INFO - codeparrot_training - Step 12136: {'lr': 0.000446975822101739, 'samples': 6214144, 'steps': 12136, 'loss/train': 2.2050106525421143} 02/24/2022 17:17:25 - INFO - codeparrot_training - Step 12137: {'lr': 0.00044696574569228365, 'samples': 6214656, 'steps': 12137, 'loss/train': 2.052475690841675} 02/24/2022 17:17:28 - INFO - codeparrot_training - Step 12138: {'lr': 0.00044695566843908947, 'samples': 6215168, 'steps': 12138, 'loss/train': 2.0426907539367676} 02/24/2022 17:17:34 - INFO - codeparrot_training - Step 12139: {'lr': 0.0004469455903421998, 'samples': 6215680, 'steps': 12139, 'loss/train': 1.3605375289916992} 02/24/2022 17:17:37 - INFO - codeparrot_training - Step 12140: {'lr': 0.0004469355114016577, 'samples': 6216192, 'steps': 12140, 'loss/train': 0.4752911329269409} 02/24/2022 17:17:43 - INFO - codeparrot_training - Step 12141: {'lr': 0.0004469254316175065, 'samples': 6216704, 'steps': 12141, 'loss/train': 1.0319311618804932} 02/24/2022 17:17:46 - INFO - codeparrot_training - Step 12142: {'lr': 0.0004469153509897892, 'samples': 6217216, 'steps': 12142, 'loss/train': 2.084182024002075} 02/24/2022 17:17:52 - INFO - codeparrot_training - Step 12143: {'lr': 0.00044690526951854907, 'samples': 6217728, 'steps': 12143, 'loss/train': 1.1665127277374268} 02/24/2022 17:17:56 - INFO - codeparrot_training - Step 12144: {'lr': 0.0004468951872038293, 'samples': 6218240, 'steps': 12144, 'loss/train': 2.1122260093688965} 02/24/2022 17:18:01 - INFO - codeparrot_training - Step 12145: {'lr': 0.00044688510404567307, 'samples': 6218752, 'steps': 12145, 'loss/train': 2.1666266918182373} 02/24/2022 17:18:05 - INFO - codeparrot_training - Step 12146: {'lr': 0.0004468750200441236, 'samples': 6219264, 'steps': 12146, 'loss/train': 0.9853890538215637} 02/24/2022 17:18:10 - INFO - codeparrot_training - Step 12147: {'lr': 0.00044686493519922405, 'samples': 6219776, 'steps': 12147, 'loss/train': 2.296105146408081} 02/24/2022 17:18:14 - INFO - codeparrot_training - Step 12148: {'lr': 0.00044685484951101763, 'samples': 6220288, 'steps': 12148, 'loss/train': 1.2640005350112915} 02/24/2022 17:18:19 - INFO - codeparrot_training - Step 12149: {'lr': 0.0004468447629795475, 'samples': 6220800, 'steps': 12149, 'loss/train': 1.3010812997817993} 02/24/2022 17:18:23 - INFO - codeparrot_training - Step 12150: {'lr': 0.00044683467560485696, 'samples': 6221312, 'steps': 12150, 'loss/train': 0.910775363445282} 02/24/2022 17:18:28 - INFO - codeparrot_training - Step 12151: {'lr': 0.00044682458738698916, 'samples': 6221824, 'steps': 12151, 'loss/train': 2.668936252593994} 02/24/2022 17:18:32 - INFO - codeparrot_training - Step 12152: {'lr': 0.0004468144983259873, 'samples': 6222336, 'steps': 12152, 'loss/train': 2.7772552967071533} 02/24/2022 17:18:38 - INFO - codeparrot_training - Step 12153: {'lr': 0.00044680440842189464, 'samples': 6222848, 'steps': 12153, 'loss/train': 1.2003204822540283} 02/24/2022 17:18:41 - INFO - codeparrot_training - Step 12154: {'lr': 0.0004467943176747544, 'samples': 6223360, 'steps': 12154, 'loss/train': 2.720294713973999} 02/24/2022 17:18:47 - INFO - codeparrot_training - Step 12155: {'lr': 0.0004467842260846098, 'samples': 6223872, 'steps': 12155, 'loss/train': 1.2617138624191284} 02/24/2022 17:18:50 - INFO - codeparrot_training - Step 12156: {'lr': 0.00044677413365150397, 'samples': 6224384, 'steps': 12156, 'loss/train': 6.09367036819458} 02/24/2022 17:18:56 - INFO - codeparrot_training - Step 12157: {'lr': 0.00044676404037548035, 'samples': 6224896, 'steps': 12157, 'loss/train': 0.2070409655570984} 02/24/2022 17:18:59 - INFO - codeparrot_training - Step 12158: {'lr': 0.0004467539462565821, 'samples': 6225408, 'steps': 12158, 'loss/train': 2.41174578666687} 02/24/2022 17:19:05 - INFO - codeparrot_training - Step 12159: {'lr': 0.0004467438512948523, 'samples': 6225920, 'steps': 12159, 'loss/train': 2.365654230117798} 02/24/2022 17:19:09 - INFO - codeparrot_training - Step 12160: {'lr': 0.00044673375549033435, 'samples': 6226432, 'steps': 12160, 'loss/train': 2.6146440505981445} 02/24/2022 17:19:14 - INFO - codeparrot_training - Step 12161: {'lr': 0.0004467236588430714, 'samples': 6226944, 'steps': 12161, 'loss/train': 2.173468828201294} 02/24/2022 17:19:18 - INFO - codeparrot_training - Step 12162: {'lr': 0.00044671356135310685, 'samples': 6227456, 'steps': 12162, 'loss/train': 2.3044090270996094} 02/24/2022 17:19:23 - INFO - codeparrot_training - Step 12163: {'lr': 0.0004467034630204839, 'samples': 6227968, 'steps': 12163, 'loss/train': 2.7269012928009033} 02/24/2022 17:19:27 - INFO - codeparrot_training - Step 12164: {'lr': 0.0004466933638452457, 'samples': 6228480, 'steps': 12164, 'loss/train': 1.8663148880004883} 02/24/2022 17:19:33 - INFO - codeparrot_training - Step 12165: {'lr': 0.0004466832638274356, 'samples': 6228992, 'steps': 12165, 'loss/train': 2.520116090774536} 02/24/2022 17:19:36 - INFO - codeparrot_training - Step 12166: {'lr': 0.0004466731629670969, 'samples': 6229504, 'steps': 12166, 'loss/train': 0.5261411070823669} 02/24/2022 17:19:42 - INFO - codeparrot_training - Step 12167: {'lr': 0.00044666306126427276, 'samples': 6230016, 'steps': 12167, 'loss/train': 1.8091787099838257} 02/24/2022 17:19:45 - INFO - codeparrot_training - Step 12168: {'lr': 0.00044665295871900655, 'samples': 6230528, 'steps': 12168, 'loss/train': 1.778935432434082} 02/24/2022 17:19:51 - INFO - codeparrot_training - Step 12169: {'lr': 0.0004466428553313415, 'samples': 6231040, 'steps': 12169, 'loss/train': 1.3250904083251953} 02/24/2022 17:19:55 - INFO - codeparrot_training - Step 12170: {'lr': 0.0004466327511013208, 'samples': 6231552, 'steps': 12170, 'loss/train': 2.5258848667144775} 02/24/2022 17:20:00 - INFO - codeparrot_training - Step 12171: {'lr': 0.00044662264602898794, 'samples': 6232064, 'steps': 12171, 'loss/train': 2.155921220779419} 02/24/2022 17:20:04 - INFO - codeparrot_training - Step 12172: {'lr': 0.00044661254011438614, 'samples': 6232576, 'steps': 12172, 'loss/train': 2.612653970718384} 02/24/2022 17:20:09 - INFO - codeparrot_training - Step 12173: {'lr': 0.00044660243335755854, 'samples': 6233088, 'steps': 12173, 'loss/train': 2.268120765686035} 02/24/2022 17:20:13 - INFO - codeparrot_training - Step 12174: {'lr': 0.00044659232575854866, 'samples': 6233600, 'steps': 12174, 'loss/train': 1.769982933998108} 02/24/2022 17:20:18 - INFO - codeparrot_training - Step 12175: {'lr': 0.00044658221731739954, 'samples': 6234112, 'steps': 12175, 'loss/train': 2.595690965652466} 02/24/2022 17:20:22 - INFO - codeparrot_training - Step 12176: {'lr': 0.0004465721080341547, 'samples': 6234624, 'steps': 12176, 'loss/train': 1.7044028043746948} 02/24/2022 17:20:27 - INFO - codeparrot_training - Step 12177: {'lr': 0.00044656199790885743, 'samples': 6235136, 'steps': 12177, 'loss/train': 2.585681915283203} 02/24/2022 17:20:31 - INFO - codeparrot_training - Step 12178: {'lr': 0.0004465518869415509, 'samples': 6235648, 'steps': 12178, 'loss/train': 2.7242603302001953} 02/24/2022 17:20:37 - INFO - codeparrot_training - Step 12179: {'lr': 0.0004465417751322785, 'samples': 6236160, 'steps': 12179, 'loss/train': 2.0500900745391846} 02/24/2022 17:20:41 - INFO - codeparrot_training - Step 12180: {'lr': 0.00044653166248108357, 'samples': 6236672, 'steps': 12180, 'loss/train': 2.794024705886841} 02/24/2022 17:20:46 - INFO - codeparrot_training - Step 12181: {'lr': 0.00044652154898800937, 'samples': 6237184, 'steps': 12181, 'loss/train': 1.8298736810684204} 02/24/2022 17:20:49 - INFO - codeparrot_training - Step 12182: {'lr': 0.0004465114346530993, 'samples': 6237696, 'steps': 12182, 'loss/train': 2.546334981918335} 02/24/2022 17:20:55 - INFO - codeparrot_training - Step 12183: {'lr': 0.0004465013194763966, 'samples': 6238208, 'steps': 12183, 'loss/train': 2.531102418899536} 02/24/2022 17:20:58 - INFO - codeparrot_training - Step 12184: {'lr': 0.0004464912034579447, 'samples': 6238720, 'steps': 12184, 'loss/train': 2.207943916320801} 02/24/2022 17:21:04 - INFO - codeparrot_training - Step 12185: {'lr': 0.00044648108659778687, 'samples': 6239232, 'steps': 12185, 'loss/train': 1.4557642936706543} 02/24/2022 17:21:07 - INFO - codeparrot_training - Step 12186: {'lr': 0.0004464709688959664, 'samples': 6239744, 'steps': 12186, 'loss/train': 2.5606086254119873} 02/24/2022 17:21:13 - INFO - codeparrot_training - Step 12187: {'lr': 0.0004464608503525267, 'samples': 6240256, 'steps': 12187, 'loss/train': 2.242934226989746} 02/24/2022 17:21:16 - INFO - codeparrot_training - Step 12188: {'lr': 0.0004464507309675111, 'samples': 6240768, 'steps': 12188, 'loss/train': 0.28196999430656433} 02/24/2022 17:21:23 - INFO - codeparrot_training - Step 12189: {'lr': 0.000446440610740963, 'samples': 6241280, 'steps': 12189, 'loss/train': 2.3940367698669434} 02/24/2022 17:21:26 - INFO - codeparrot_training - Step 12190: {'lr': 0.0004464304896729257, 'samples': 6241792, 'steps': 12190, 'loss/train': 2.5269486904144287} 02/24/2022 17:21:32 - INFO - codeparrot_training - Step 12191: {'lr': 0.0004464203677634424, 'samples': 6242304, 'steps': 12191, 'loss/train': 2.1737985610961914} 02/24/2022 17:21:35 - INFO - codeparrot_training - Step 12192: {'lr': 0.0004464102450125568, 'samples': 6242816, 'steps': 12192, 'loss/train': 1.9607230424880981} 02/24/2022 17:21:41 - INFO - codeparrot_training - Step 12193: {'lr': 0.00044640012142031196, 'samples': 6243328, 'steps': 12193, 'loss/train': 2.8272314071655273} 02/24/2022 17:21:44 - INFO - codeparrot_training - Step 12194: {'lr': 0.0004463899969867514, 'samples': 6243840, 'steps': 12194, 'loss/train': 1.5703818798065186} 02/24/2022 17:21:50 - INFO - codeparrot_training - Step 12195: {'lr': 0.0004463798717119185, 'samples': 6244352, 'steps': 12195, 'loss/train': 2.894139289855957} 02/24/2022 17:21:53 - INFO - codeparrot_training - Step 12196: {'lr': 0.00044636974559585655, 'samples': 6244864, 'steps': 12196, 'loss/train': 2.4204652309417725} 02/24/2022 17:21:59 - INFO - codeparrot_training - Step 12197: {'lr': 0.00044635961863860894, 'samples': 6245376, 'steps': 12197, 'loss/train': 0.8477448225021362} 02/24/2022 17:22:02 - INFO - codeparrot_training - Step 12198: {'lr': 0.00044634949084021913, 'samples': 6245888, 'steps': 12198, 'loss/train': 1.9968920946121216} 02/24/2022 17:22:09 - INFO - codeparrot_training - Step 12199: {'lr': 0.0004463393622007305, 'samples': 6246400, 'steps': 12199, 'loss/train': 1.155855655670166} 02/24/2022 17:22:13 - INFO - codeparrot_training - Step 12200: {'lr': 0.0004463292327201862, 'samples': 6246912, 'steps': 12200, 'loss/train': 2.209712505340576} 02/24/2022 17:22:18 - INFO - codeparrot_training - Step 12201: {'lr': 0.0004463191023986299, 'samples': 6247424, 'steps': 12201, 'loss/train': 1.8642038106918335} 02/24/2022 17:22:21 - INFO - codeparrot_training - Step 12202: {'lr': 0.00044630897123610497, 'samples': 6247936, 'steps': 12202, 'loss/train': 2.290565013885498} 02/24/2022 17:22:27 - INFO - codeparrot_training - Step 12203: {'lr': 0.0004462988392326547, 'samples': 6248448, 'steps': 12203, 'loss/train': 1.4512784481048584} 02/24/2022 17:22:30 - INFO - codeparrot_training - Step 12204: {'lr': 0.00044628870638832254, 'samples': 6248960, 'steps': 12204, 'loss/train': 0.11816435307264328} 02/24/2022 17:22:36 - INFO - codeparrot_training - Step 12205: {'lr': 0.00044627857270315187, 'samples': 6249472, 'steps': 12205, 'loss/train': 1.8517296314239502} 02/24/2022 17:22:40 - INFO - codeparrot_training - Step 12206: {'lr': 0.00044626843817718615, 'samples': 6249984, 'steps': 12206, 'loss/train': 2.1940414905548096} 02/24/2022 17:22:45 - INFO - codeparrot_training - Step 12207: {'lr': 0.00044625830281046875, 'samples': 6250496, 'steps': 12207, 'loss/train': 2.197786808013916} 02/24/2022 17:22:48 - INFO - codeparrot_training - Step 12208: {'lr': 0.0004462481666030431, 'samples': 6251008, 'steps': 12208, 'loss/train': 1.139675259590149} 02/24/2022 17:22:54 - INFO - codeparrot_training - Step 12209: {'lr': 0.0004462380295549526, 'samples': 6251520, 'steps': 12209, 'loss/train': 0.5278382897377014} 02/24/2022 17:22:58 - INFO - codeparrot_training - Step 12210: {'lr': 0.0004462278916662407, 'samples': 6252032, 'steps': 12210, 'loss/train': 0.5970871448516846} 02/24/2022 17:23:03 - INFO - codeparrot_training - Step 12211: {'lr': 0.00044621775293695085, 'samples': 6252544, 'steps': 12211, 'loss/train': 2.36564040184021} 02/24/2022 17:23:06 - INFO - codeparrot_training - Step 12212: {'lr': 0.00044620761336712646, 'samples': 6253056, 'steps': 12212, 'loss/train': 1.9601918458938599} 02/24/2022 17:23:12 - INFO - codeparrot_training - Step 12213: {'lr': 0.0004461974729568109, 'samples': 6253568, 'steps': 12213, 'loss/train': 2.555851459503174} 02/24/2022 17:23:15 - INFO - codeparrot_training - Step 12214: {'lr': 0.0004461873317060477, 'samples': 6254080, 'steps': 12214, 'loss/train': 2.2025561332702637} 02/24/2022 17:23:22 - INFO - codeparrot_training - Step 12215: {'lr': 0.00044617718961488024, 'samples': 6254592, 'steps': 12215, 'loss/train': 1.5442249774932861} 02/24/2022 17:23:25 - INFO - codeparrot_training - Step 12216: {'lr': 0.000446167046683352, 'samples': 6255104, 'steps': 12216, 'loss/train': 2.2014803886413574} 02/24/2022 17:23:31 - INFO - codeparrot_training - Step 12217: {'lr': 0.0004461569029115065, 'samples': 6255616, 'steps': 12217, 'loss/train': 2.4430060386657715} 02/24/2022 17:23:34 - INFO - codeparrot_training - Step 12218: {'lr': 0.000446146758299387, 'samples': 6256128, 'steps': 12218, 'loss/train': 2.3343770503997803} 02/24/2022 17:23:40 - INFO - codeparrot_training - Step 12219: {'lr': 0.0004461366128470371, 'samples': 6256640, 'steps': 12219, 'loss/train': 1.514944314956665} 02/24/2022 17:23:43 - INFO - codeparrot_training - Step 12220: {'lr': 0.0004461264665545003, 'samples': 6257152, 'steps': 12220, 'loss/train': 2.2627553939819336} 02/24/2022 17:23:49 - INFO - codeparrot_training - Step 12221: {'lr': 0.00044611631942182, 'samples': 6257664, 'steps': 12221, 'loss/train': 2.365278482437134} 02/24/2022 17:23:52 - INFO - codeparrot_training - Step 12222: {'lr': 0.0004461061714490395, 'samples': 6258176, 'steps': 12222, 'loss/train': 1.005839467048645} 02/24/2022 17:23:58 - INFO - codeparrot_training - Step 12223: {'lr': 0.0004460960226362026, 'samples': 6258688, 'steps': 12223, 'loss/train': 2.1605939865112305} 02/24/2022 17:24:01 - INFO - codeparrot_training - Step 12224: {'lr': 0.0004460858729833525, 'samples': 6259200, 'steps': 12224, 'loss/train': 1.6568334102630615} 02/24/2022 17:24:08 - INFO - codeparrot_training - Step 12225: {'lr': 0.00044607572249053283, 'samples': 6259712, 'steps': 12225, 'loss/train': 0.4331347346305847} 02/24/2022 17:24:11 - INFO - codeparrot_training - Step 12226: {'lr': 0.0004460655711577871, 'samples': 6260224, 'steps': 12226, 'loss/train': 1.968904972076416} 02/24/2022 17:24:17 - INFO - codeparrot_training - Step 12227: {'lr': 0.00044605541898515863, 'samples': 6260736, 'steps': 12227, 'loss/train': 1.7360906600952148} 02/24/2022 17:24:20 - INFO - codeparrot_training - Step 12228: {'lr': 0.00044604526597269103, 'samples': 6261248, 'steps': 12228, 'loss/train': 1.6190905570983887} 02/24/2022 17:24:25 - INFO - codeparrot_training - Step 12229: {'lr': 0.0004460351121204277, 'samples': 6261760, 'steps': 12229, 'loss/train': 1.6429935693740845} 02/24/2022 17:24:29 - INFO - codeparrot_training - Step 12230: {'lr': 0.00044602495742841226, 'samples': 6262272, 'steps': 12230, 'loss/train': 2.3980512619018555} 02/24/2022 17:24:35 - INFO - codeparrot_training - Step 12231: {'lr': 0.00044601480189668816, 'samples': 6262784, 'steps': 12231, 'loss/train': 1.9587669372558594} 02/24/2022 17:24:38 - INFO - codeparrot_training - Step 12232: {'lr': 0.00044600464552529886, 'samples': 6263296, 'steps': 12232, 'loss/train': 2.0523691177368164} 02/24/2022 17:24:44 - INFO - codeparrot_training - Step 12233: {'lr': 0.0004459944883142879, 'samples': 6263808, 'steps': 12233, 'loss/train': 2.524299383163452} 02/24/2022 17:24:49 - INFO - codeparrot_training - Step 12234: {'lr': 0.0004459843302636988, 'samples': 6264320, 'steps': 12234, 'loss/train': 2.87491774559021} 02/24/2022 17:24:53 - INFO - codeparrot_training - Step 12235: {'lr': 0.000445974171373575, 'samples': 6264832, 'steps': 12235, 'loss/train': 1.8807936906814575} 02/24/2022 17:24:59 - INFO - codeparrot_training - Step 12236: {'lr': 0.0004459640116439602, 'samples': 6265344, 'steps': 12236, 'loss/train': 2.117929458618164} 02/24/2022 17:25:03 - INFO - codeparrot_training - Step 12237: {'lr': 0.0004459538510748977, 'samples': 6265856, 'steps': 12237, 'loss/train': 2.6591475009918213} 02/24/2022 17:25:08 - INFO - codeparrot_training - Step 12238: {'lr': 0.0004459436896664312, 'samples': 6266368, 'steps': 12238, 'loss/train': 2.887000799179077} 02/24/2022 17:25:11 - INFO - codeparrot_training - Step 12239: {'lr': 0.00044593352741860404, 'samples': 6266880, 'steps': 12239, 'loss/train': 2.488243579864502} 02/24/2022 17:25:17 - INFO - codeparrot_training - Step 12240: {'lr': 0.00044592336433145995, 'samples': 6267392, 'steps': 12240, 'loss/train': 2.0779874324798584} 02/24/2022 17:25:21 - INFO - codeparrot_training - Step 12241: {'lr': 0.00044591320040504237, 'samples': 6267904, 'steps': 12241, 'loss/train': 2.8562235832214355} 02/24/2022 17:25:26 - INFO - codeparrot_training - Step 12242: {'lr': 0.00044590303563939485, 'samples': 6268416, 'steps': 12242, 'loss/train': 1.9997836351394653} 02/24/2022 17:25:30 - INFO - codeparrot_training - Step 12243: {'lr': 0.0004458928700345609, 'samples': 6268928, 'steps': 12243, 'loss/train': 2.0879430770874023} 02/24/2022 17:25:35 - INFO - codeparrot_training - Step 12244: {'lr': 0.00044588270359058416, 'samples': 6269440, 'steps': 12244, 'loss/train': 1.1297696828842163} 02/24/2022 17:25:39 - INFO - codeparrot_training - Step 12245: {'lr': 0.000445872536307508, 'samples': 6269952, 'steps': 12245, 'loss/train': 1.9869401454925537} 02/24/2022 17:25:45 - INFO - codeparrot_training - Step 12246: {'lr': 0.0004458623681853762, 'samples': 6270464, 'steps': 12246, 'loss/train': 0.7562997937202454} 02/24/2022 17:25:48 - INFO - codeparrot_training - Step 12247: {'lr': 0.0004458521992242322, 'samples': 6270976, 'steps': 12247, 'loss/train': 1.6579252481460571} 02/24/2022 17:25:54 - INFO - codeparrot_training - Step 12248: {'lr': 0.00044584202942411956, 'samples': 6271488, 'steps': 12248, 'loss/train': 1.962679147720337} 02/24/2022 17:25:57 - INFO - codeparrot_training - Step 12249: {'lr': 0.00044583185878508183, 'samples': 6272000, 'steps': 12249, 'loss/train': 1.5631684064865112} 02/24/2022 17:26:03 - INFO - codeparrot_training - Step 12250: {'lr': 0.0004458216873071626, 'samples': 6272512, 'steps': 12250, 'loss/train': 2.2190117835998535} 02/24/2022 17:26:06 - INFO - codeparrot_training - Step 12251: {'lr': 0.00044581151499040547, 'samples': 6273024, 'steps': 12251, 'loss/train': 0.9030312895774841} 02/24/2022 17:26:12 - INFO - codeparrot_training - Step 12252: {'lr': 0.000445801341834854, 'samples': 6273536, 'steps': 12252, 'loss/train': 2.4146976470947266} 02/24/2022 17:26:15 - INFO - codeparrot_training - Step 12253: {'lr': 0.0004457911678405517, 'samples': 6274048, 'steps': 12253, 'loss/train': 0.5449259281158447} 02/24/2022 17:26:21 - INFO - codeparrot_training - Step 12254: {'lr': 0.0004457809930075422, 'samples': 6274560, 'steps': 12254, 'loss/train': 1.6004161834716797} 02/24/2022 17:26:24 - INFO - codeparrot_training - Step 12255: {'lr': 0.0004457708173358691, 'samples': 6275072, 'steps': 12255, 'loss/train': 2.691077947616577} 02/24/2022 17:26:30 - INFO - codeparrot_training - Step 12256: {'lr': 0.00044576064082557605, 'samples': 6275584, 'steps': 12256, 'loss/train': 2.0691773891448975} 02/24/2022 17:26:33 - INFO - codeparrot_training - Step 12257: {'lr': 0.0004457504634767066, 'samples': 6276096, 'steps': 12257, 'loss/train': 2.9726624488830566} 02/24/2022 17:26:39 - INFO - codeparrot_training - Step 12258: {'lr': 0.0004457402852893042, 'samples': 6276608, 'steps': 12258, 'loss/train': 2.6471009254455566} 02/24/2022 17:26:43 - INFO - codeparrot_training - Step 12259: {'lr': 0.0004457301062634126, 'samples': 6277120, 'steps': 12259, 'loss/train': 2.3239684104919434} 02/24/2022 17:26:48 - INFO - codeparrot_training - Step 12260: {'lr': 0.0004457199263990754, 'samples': 6277632, 'steps': 12260, 'loss/train': 2.2256107330322266} 02/24/2022 17:26:52 - INFO - codeparrot_training - Step 12261: {'lr': 0.0004457097456963362, 'samples': 6278144, 'steps': 12261, 'loss/train': 1.5917807817459106} 02/24/2022 17:26:58 - INFO - codeparrot_training - Step 12262: {'lr': 0.0004456995641552386, 'samples': 6278656, 'steps': 12262, 'loss/train': 1.9505926370620728} 02/24/2022 17:27:01 - INFO - codeparrot_training - Step 12263: {'lr': 0.0004456893817758262, 'samples': 6279168, 'steps': 12263, 'loss/train': 1.3096234798431396} 02/24/2022 17:27:07 - INFO - codeparrot_training - Step 12264: {'lr': 0.00044567919855814257, 'samples': 6279680, 'steps': 12264, 'loss/train': 2.4166879653930664} 02/24/2022 17:27:10 - INFO - codeparrot_training - Step 12265: {'lr': 0.0004456690145022314, 'samples': 6280192, 'steps': 12265, 'loss/train': 2.4174020290374756} 02/24/2022 17:27:16 - INFO - codeparrot_training - Step 12266: {'lr': 0.0004456588296081364, 'samples': 6280704, 'steps': 12266, 'loss/train': 2.4440877437591553} 02/24/2022 17:27:19 - INFO - codeparrot_training - Step 12267: {'lr': 0.000445648643875901, 'samples': 6281216, 'steps': 12267, 'loss/train': 2.1894125938415527} 02/24/2022 17:27:25 - INFO - codeparrot_training - Step 12268: {'lr': 0.000445638457305569, 'samples': 6281728, 'steps': 12268, 'loss/train': 2.184979200363159} 02/24/2022 17:27:28 - INFO - codeparrot_training - Step 12269: {'lr': 0.00044562826989718397, 'samples': 6282240, 'steps': 12269, 'loss/train': 2.278891086578369} 02/24/2022 17:27:34 - INFO - codeparrot_training - Step 12270: {'lr': 0.00044561808165078954, 'samples': 6282752, 'steps': 12270, 'loss/train': 1.8342580795288086} 02/24/2022 17:27:37 - INFO - codeparrot_training - Step 12271: {'lr': 0.0004456078925664293, 'samples': 6283264, 'steps': 12271, 'loss/train': 3.0874226093292236} 02/24/2022 17:27:44 - INFO - codeparrot_training - Step 12272: {'lr': 0.000445597702644147, 'samples': 6283776, 'steps': 12272, 'loss/train': 2.4969401359558105} 02/24/2022 17:27:47 - INFO - codeparrot_training - Step 12273: {'lr': 0.0004455875118839863, 'samples': 6284288, 'steps': 12273, 'loss/train': 2.95229172706604} 02/24/2022 17:27:53 - INFO - codeparrot_training - Step 12274: {'lr': 0.00044557732028599077, 'samples': 6284800, 'steps': 12274, 'loss/train': 1.8821544647216797} 02/24/2022 17:27:56 - INFO - codeparrot_training - Step 12275: {'lr': 0.0004455671278502041, 'samples': 6285312, 'steps': 12275, 'loss/train': 1.5584900379180908} 02/24/2022 17:28:02 - INFO - codeparrot_training - Step 12276: {'lr': 0.00044555693457667, 'samples': 6285824, 'steps': 12276, 'loss/train': 2.4532015323638916} 02/24/2022 17:28:05 - INFO - codeparrot_training - Step 12277: {'lr': 0.000445546740465432, 'samples': 6286336, 'steps': 12277, 'loss/train': 1.792319893836975} 02/24/2022 17:28:11 - INFO - codeparrot_training - Step 12278: {'lr': 0.00044553654551653387, 'samples': 6286848, 'steps': 12278, 'loss/train': 4.381943702697754} 02/24/2022 17:28:14 - INFO - codeparrot_training - Step 12279: {'lr': 0.0004455263497300194, 'samples': 6287360, 'steps': 12279, 'loss/train': 0.990138053894043} 02/24/2022 17:28:20 - INFO - codeparrot_training - Step 12280: {'lr': 0.000445516153105932, 'samples': 6287872, 'steps': 12280, 'loss/train': 1.7454262971878052} 02/24/2022 17:28:24 - INFO - codeparrot_training - Step 12281: {'lr': 0.0004455059556443155, 'samples': 6288384, 'steps': 12281, 'loss/train': 2.2448575496673584} 02/24/2022 17:28:30 - INFO - codeparrot_training - Step 12282: {'lr': 0.0004454957573452136, 'samples': 6288896, 'steps': 12282, 'loss/train': 1.0142549276351929} 02/24/2022 17:28:33 - INFO - codeparrot_training - Step 12283: {'lr': 0.0004454855582086699, 'samples': 6289408, 'steps': 12283, 'loss/train': 2.5627574920654297} 02/24/2022 17:28:39 - INFO - codeparrot_training - Step 12284: {'lr': 0.0004454753582347282, 'samples': 6289920, 'steps': 12284, 'loss/train': 2.0157546997070312} 02/24/2022 17:28:42 - INFO - codeparrot_training - Step 12285: {'lr': 0.00044546515742343207, 'samples': 6290432, 'steps': 12285, 'loss/train': 1.9214407205581665} 02/24/2022 17:28:48 - INFO - codeparrot_training - Step 12286: {'lr': 0.00044545495577482535, 'samples': 6290944, 'steps': 12286, 'loss/train': 1.8595911264419556} 02/24/2022 17:28:51 - INFO - codeparrot_training - Step 12287: {'lr': 0.00044544475328895164, 'samples': 6291456, 'steps': 12287, 'loss/train': 1.9937409162521362} 02/24/2022 17:28:57 - INFO - codeparrot_training - Step 12288: {'lr': 0.00044543454996585463, 'samples': 6291968, 'steps': 12288, 'loss/train': 1.5022523403167725} 02/24/2022 17:29:00 - INFO - codeparrot_training - Step 12289: {'lr': 0.0004454243458055781, 'samples': 6292480, 'steps': 12289, 'loss/train': 2.310431718826294} 02/24/2022 17:29:06 - INFO - codeparrot_training - Step 12290: {'lr': 0.00044541414080816573, 'samples': 6292992, 'steps': 12290, 'loss/train': 1.9270498752593994} 02/24/2022 17:29:09 - INFO - codeparrot_training - Step 12291: {'lr': 0.00044540393497366124, 'samples': 6293504, 'steps': 12291, 'loss/train': 2.380450963973999} 02/24/2022 17:29:16 - INFO - codeparrot_training - Step 12292: {'lr': 0.00044539372830210833, 'samples': 6294016, 'steps': 12292, 'loss/train': 1.314091682434082} 02/24/2022 17:29:19 - INFO - codeparrot_training - Step 12293: {'lr': 0.0004453835207935507, 'samples': 6294528, 'steps': 12293, 'loss/train': 1.027348518371582} 02/24/2022 17:29:25 - INFO - codeparrot_training - Step 12294: {'lr': 0.0004453733124480321, 'samples': 6295040, 'steps': 12294, 'loss/train': 2.0489554405212402} 02/24/2022 17:29:28 - INFO - codeparrot_training - Step 12295: {'lr': 0.0004453631032655964, 'samples': 6295552, 'steps': 12295, 'loss/train': 1.8813971281051636} 02/24/2022 17:29:34 - INFO - codeparrot_training - Step 12296: {'lr': 0.00044535289324628704, 'samples': 6296064, 'steps': 12296, 'loss/train': 2.194880485534668} 02/24/2022 17:29:37 - INFO - codeparrot_training - Step 12297: {'lr': 0.00044534268239014796, 'samples': 6296576, 'steps': 12297, 'loss/train': 2.218430995941162} 02/24/2022 17:29:43 - INFO - codeparrot_training - Step 12298: {'lr': 0.00044533247069722295, 'samples': 6297088, 'steps': 12298, 'loss/train': 1.4462721347808838} 02/24/2022 17:29:46 - INFO - codeparrot_training - Step 12299: {'lr': 0.0004453222581675556, 'samples': 6297600, 'steps': 12299, 'loss/train': 1.7703735828399658} 02/24/2022 17:29:52 - INFO - codeparrot_training - Step 12300: {'lr': 0.0004453120448011897, 'samples': 6298112, 'steps': 12300, 'loss/train': 1.7118444442749023} 02/24/2022 17:29:55 - INFO - codeparrot_training - Step 12301: {'lr': 0.00044530183059816896, 'samples': 6298624, 'steps': 12301, 'loss/train': 1.4803307056427002} 02/24/2022 17:30:01 - INFO - codeparrot_training - Step 12302: {'lr': 0.00044529161555853725, 'samples': 6299136, 'steps': 12302, 'loss/train': 1.002977728843689} 02/24/2022 17:30:04 - INFO - codeparrot_training - Step 12303: {'lr': 0.0004452813996823383, 'samples': 6299648, 'steps': 12303, 'loss/train': 1.6941735744476318} 02/24/2022 17:30:10 - INFO - codeparrot_training - Step 12304: {'lr': 0.00044527118296961576, 'samples': 6300160, 'steps': 12304, 'loss/train': 0.8799401521682739} 02/24/2022 17:30:13 - INFO - codeparrot_training - Step 12305: {'lr': 0.0004452609654204136, 'samples': 6300672, 'steps': 12305, 'loss/train': 2.210076093673706} 02/24/2022 17:30:20 - INFO - codeparrot_training - Step 12306: {'lr': 0.0004452507470347754, 'samples': 6301184, 'steps': 12306, 'loss/train': 2.1028568744659424} 02/24/2022 17:30:24 - INFO - codeparrot_training - Step 12307: {'lr': 0.00044524052781274497, 'samples': 6301696, 'steps': 12307, 'loss/train': 2.4408137798309326} 02/24/2022 17:30:29 - INFO - codeparrot_training - Step 12308: {'lr': 0.00044523030775436617, 'samples': 6302208, 'steps': 12308, 'loss/train': 1.1121882200241089} 02/24/2022 17:30:33 - INFO - codeparrot_training - Step 12309: {'lr': 0.0004452200868596827, 'samples': 6302720, 'steps': 12309, 'loss/train': 2.3073625564575195} 02/24/2022 17:30:38 - INFO - codeparrot_training - Step 12310: {'lr': 0.0004452098651287384, 'samples': 6303232, 'steps': 12310, 'loss/train': 1.9814388751983643} 02/24/2022 17:30:42 - INFO - codeparrot_training - Step 12311: {'lr': 0.000445199642561577, 'samples': 6303744, 'steps': 12311, 'loss/train': 1.386033535003662} 02/24/2022 17:30:47 - INFO - codeparrot_training - Step 12312: {'lr': 0.0004451894191582423, 'samples': 6304256, 'steps': 12312, 'loss/train': 2.078249216079712} 02/24/2022 17:30:51 - INFO - codeparrot_training - Step 12313: {'lr': 0.0004451791949187781, 'samples': 6304768, 'steps': 12313, 'loss/train': 2.4705617427825928} 02/24/2022 17:30:56 - INFO - codeparrot_training - Step 12314: {'lr': 0.0004451689698432282, 'samples': 6305280, 'steps': 12314, 'loss/train': 1.6156132221221924} 02/24/2022 17:31:00 - INFO - codeparrot_training - Step 12315: {'lr': 0.0004451587439316365, 'samples': 6305792, 'steps': 12315, 'loss/train': 0.9699227809906006} 02/24/2022 17:31:06 - INFO - codeparrot_training - Step 12316: {'lr': 0.0004451485171840466, 'samples': 6306304, 'steps': 12316, 'loss/train': 1.938400387763977} 02/24/2022 17:31:12 - INFO - codeparrot_training - Step 12317: {'lr': 0.0004451382896005024, 'samples': 6306816, 'steps': 12317, 'loss/train': 2.1531031131744385} 02/24/2022 17:31:15 - INFO - codeparrot_training - Step 12318: {'lr': 0.00044512806118104784, 'samples': 6307328, 'steps': 12318, 'loss/train': 1.8844478130340576} 02/24/2022 17:31:21 - INFO - codeparrot_training - Step 12319: {'lr': 0.0004451178319257265, 'samples': 6307840, 'steps': 12319, 'loss/train': 2.1367862224578857} 02/24/2022 17:31:24 - INFO - codeparrot_training - Step 12320: {'lr': 0.0004451076018345824, 'samples': 6308352, 'steps': 12320, 'loss/train': 1.9447153806686401} 02/24/2022 17:31:28 - INFO - codeparrot_training - Step 12321: {'lr': 0.00044509737090765933, 'samples': 6308864, 'steps': 12321, 'loss/train': 2.543639659881592} 02/24/2022 17:31:34 - INFO - codeparrot_training - Step 12322: {'lr': 0.00044508713914500107, 'samples': 6309376, 'steps': 12322, 'loss/train': 2.049834966659546} 02/24/2022 17:31:37 - INFO - codeparrot_training - Step 12323: {'lr': 0.0004450769065466514, 'samples': 6309888, 'steps': 12323, 'loss/train': 2.6177823543548584} 02/24/2022 17:31:43 - INFO - codeparrot_training - Step 12324: {'lr': 0.0004450666731126542, 'samples': 6310400, 'steps': 12324, 'loss/train': 2.3640940189361572} 02/24/2022 17:31:46 - INFO - codeparrot_training - Step 12325: {'lr': 0.0004450564388430533, 'samples': 6310912, 'steps': 12325, 'loss/train': 1.936601996421814} 02/24/2022 17:31:53 - INFO - codeparrot_training - Step 12326: {'lr': 0.0004450462037378926, 'samples': 6311424, 'steps': 12326, 'loss/train': 2.0119669437408447} 02/24/2022 17:31:56 - INFO - codeparrot_training - Step 12327: {'lr': 0.0004450359677972159, 'samples': 6311936, 'steps': 12327, 'loss/train': 1.828921914100647} 02/24/2022 17:32:02 - INFO - codeparrot_training - Step 12328: {'lr': 0.000445025731021067, 'samples': 6312448, 'steps': 12328, 'loss/train': 1.2264736890792847} 02/24/2022 17:32:05 - INFO - codeparrot_training - Step 12329: {'lr': 0.0004450154934094898, 'samples': 6312960, 'steps': 12329, 'loss/train': 1.1868711709976196} 02/24/2022 17:32:11 - INFO - codeparrot_training - Step 12330: {'lr': 0.0004450052549625282, 'samples': 6313472, 'steps': 12330, 'loss/train': 1.7256996631622314} 02/24/2022 17:32:14 - INFO - codeparrot_training - Step 12331: {'lr': 0.000444995015680226, 'samples': 6313984, 'steps': 12331, 'loss/train': 5.192365646362305} 02/24/2022 17:32:20 - INFO - codeparrot_training - Step 12332: {'lr': 0.0004449847755626271, 'samples': 6314496, 'steps': 12332, 'loss/train': 2.1573946475982666} 02/24/2022 17:32:23 - INFO - codeparrot_training - Step 12333: {'lr': 0.00044497453460977523, 'samples': 6315008, 'steps': 12333, 'loss/train': 2.4214375019073486} 02/24/2022 17:32:29 - INFO - codeparrot_training - Step 12334: {'lr': 0.0004449642928217144, 'samples': 6315520, 'steps': 12334, 'loss/train': 2.800600051879883} 02/24/2022 17:32:32 - INFO - codeparrot_training - Step 12335: {'lr': 0.0004449540501984885, 'samples': 6316032, 'steps': 12335, 'loss/train': 2.9670016765594482} 02/24/2022 17:32:38 - INFO - codeparrot_training - Step 12336: {'lr': 0.0004449438067401413, 'samples': 6316544, 'steps': 12336, 'loss/train': 2.426023483276367} 02/24/2022 17:32:42 - INFO - codeparrot_training - Step 12337: {'lr': 0.0004449335624467168, 'samples': 6317056, 'steps': 12337, 'loss/train': 1.6832247972488403} 02/24/2022 17:32:48 - INFO - codeparrot_training - Step 12338: {'lr': 0.00044492331731825875, 'samples': 6317568, 'steps': 12338, 'loss/train': 1.774661898612976} 02/24/2022 17:32:52 - INFO - codeparrot_training - Step 12339: {'lr': 0.0004449130713548111, 'samples': 6318080, 'steps': 12339, 'loss/train': 2.1241750717163086} 02/24/2022 17:32:57 - INFO - codeparrot_training - Step 12340: {'lr': 0.00044490282455641783, 'samples': 6318592, 'steps': 12340, 'loss/train': 2.314002275466919} 02/24/2022 17:33:01 - INFO - codeparrot_training - Step 12341: {'lr': 0.0004448925769231227, 'samples': 6319104, 'steps': 12341, 'loss/train': 1.9071742296218872} 02/24/2022 17:33:06 - INFO - codeparrot_training - Step 12342: {'lr': 0.0004448823284549696, 'samples': 6319616, 'steps': 12342, 'loss/train': 2.9264605045318604} 02/24/2022 17:33:10 - INFO - codeparrot_training - Step 12343: {'lr': 0.00044487207915200257, 'samples': 6320128, 'steps': 12343, 'loss/train': 0.30424219369888306} 02/24/2022 17:33:15 - INFO - codeparrot_training - Step 12344: {'lr': 0.0004448618290142654, 'samples': 6320640, 'steps': 12344, 'loss/train': 0.4632541537284851} 02/24/2022 17:33:19 - INFO - codeparrot_training - Step 12345: {'lr': 0.000444851578041802, 'samples': 6321152, 'steps': 12345, 'loss/train': 1.989852786064148} 02/24/2022 17:33:24 - INFO - codeparrot_training - Step 12346: {'lr': 0.00044484132623465633, 'samples': 6321664, 'steps': 12346, 'loss/train': 2.248664379119873} 02/24/2022 17:33:28 - INFO - codeparrot_training - Step 12347: {'lr': 0.0004448310735928723, 'samples': 6322176, 'steps': 12347, 'loss/train': 1.5307692289352417} 02/24/2022 17:33:33 - INFO - codeparrot_training - Step 12348: {'lr': 0.0004448208201164938, 'samples': 6322688, 'steps': 12348, 'loss/train': 3.1500887870788574} 02/24/2022 17:33:37 - INFO - codeparrot_training - Step 12349: {'lr': 0.0004448105658055648, 'samples': 6323200, 'steps': 12349, 'loss/train': 0.2151261866092682} 02/24/2022 17:33:42 - INFO - codeparrot_training - Step 12350: {'lr': 0.00044480031066012916, 'samples': 6323712, 'steps': 12350, 'loss/train': 1.7735192775726318} 02/24/2022 17:33:46 - INFO - codeparrot_training - Step 12351: {'lr': 0.00044479005468023086, 'samples': 6324224, 'steps': 12351, 'loss/train': 3.1518568992614746} 02/24/2022 17:33:52 - INFO - codeparrot_training - Step 12352: {'lr': 0.0004447797978659138, 'samples': 6324736, 'steps': 12352, 'loss/train': 2.223313093185425} 02/24/2022 17:33:55 - INFO - codeparrot_training - Step 12353: {'lr': 0.000444769540217222, 'samples': 6325248, 'steps': 12353, 'loss/train': 1.8045709133148193} 02/24/2022 17:34:01 - INFO - codeparrot_training - Step 12354: {'lr': 0.0004447592817341993, 'samples': 6325760, 'steps': 12354, 'loss/train': 0.15722961723804474} 02/24/2022 17:34:04 - INFO - codeparrot_training - Step 12355: {'lr': 0.0004447490224168896, 'samples': 6326272, 'steps': 12355, 'loss/train': 2.2276923656463623} 02/24/2022 17:34:10 - INFO - codeparrot_training - Step 12356: {'lr': 0.00044473876226533703, 'samples': 6326784, 'steps': 12356, 'loss/train': 1.558146595954895} 02/24/2022 17:34:13 - INFO - codeparrot_training - Step 12357: {'lr': 0.0004447285012795854, 'samples': 6327296, 'steps': 12357, 'loss/train': 1.7817211151123047} 02/24/2022 17:34:19 - INFO - codeparrot_training - Step 12358: {'lr': 0.0004447182394596788, 'samples': 6327808, 'steps': 12358, 'loss/train': 1.5735222101211548} 02/24/2022 17:34:22 - INFO - codeparrot_training - Step 12359: {'lr': 0.000444707976805661, 'samples': 6328320, 'steps': 12359, 'loss/train': 2.140296220779419} 02/24/2022 17:34:28 - INFO - codeparrot_training - Step 12360: {'lr': 0.00044469771331757604, 'samples': 6328832, 'steps': 12360, 'loss/train': 1.716566801071167} 02/24/2022 17:34:31 - INFO - codeparrot_training - Step 12361: {'lr': 0.00044468744899546785, 'samples': 6329344, 'steps': 12361, 'loss/train': 1.831556797027588} 02/24/2022 17:34:38 - INFO - codeparrot_training - Step 12362: {'lr': 0.0004446771838393806, 'samples': 6329856, 'steps': 12362, 'loss/train': 2.313117742538452} 02/24/2022 17:34:42 - INFO - codeparrot_training - Step 12363: {'lr': 0.00044466691784935796, 'samples': 6330368, 'steps': 12363, 'loss/train': 2.639930486679077} 02/24/2022 17:34:47 - INFO - codeparrot_training - Step 12364: {'lr': 0.00044465665102544415, 'samples': 6330880, 'steps': 12364, 'loss/train': 1.7952097654342651} 02/24/2022 17:34:53 - INFO - codeparrot_training - Step 12365: {'lr': 0.000444646383367683, 'samples': 6331392, 'steps': 12365, 'loss/train': 1.296103835105896} 02/24/2022 17:34:56 - INFO - codeparrot_training - Step 12366: {'lr': 0.00044463611487611864, 'samples': 6331904, 'steps': 12366, 'loss/train': 0.6917550563812256} 02/24/2022 17:35:02 - INFO - codeparrot_training - Step 12367: {'lr': 0.0004446258455507949, 'samples': 6332416, 'steps': 12367, 'loss/train': 2.458496570587158} 02/24/2022 17:35:05 - INFO - codeparrot_training - Step 12368: {'lr': 0.00044461557539175587, 'samples': 6332928, 'steps': 12368, 'loss/train': 2.906895875930786} 02/24/2022 17:35:11 - INFO - codeparrot_training - Step 12369: {'lr': 0.0004446053043990455, 'samples': 6333440, 'steps': 12369, 'loss/train': 2.5737407207489014} 02/24/2022 17:35:14 - INFO - codeparrot_training - Step 12370: {'lr': 0.00044459503257270776, 'samples': 6333952, 'steps': 12370, 'loss/train': 2.17844557762146} 02/24/2022 17:35:21 - INFO - codeparrot_training - Step 12371: {'lr': 0.0004445847599127868, 'samples': 6334464, 'steps': 12371, 'loss/train': 2.2148101329803467} 02/24/2022 17:35:24 - INFO - codeparrot_training - Step 12372: {'lr': 0.0004445744864193264, 'samples': 6334976, 'steps': 12372, 'loss/train': 3.0531060695648193} 02/24/2022 17:35:30 - INFO - codeparrot_training - Step 12373: {'lr': 0.00044456421209237073, 'samples': 6335488, 'steps': 12373, 'loss/train': 1.7490960359573364} 02/24/2022 17:35:33 - INFO - codeparrot_training - Step 12374: {'lr': 0.00044455393693196375, 'samples': 6336000, 'steps': 12374, 'loss/train': 2.6144533157348633} 02/24/2022 17:35:39 - INFO - codeparrot_training - Step 12375: {'lr': 0.00044454366093814947, 'samples': 6336512, 'steps': 12375, 'loss/train': 2.1560230255126953} 02/24/2022 17:35:43 - INFO - codeparrot_training - Step 12376: {'lr': 0.0004445333841109719, 'samples': 6337024, 'steps': 12376, 'loss/train': 4.896445274353027} 02/24/2022 17:35:48 - INFO - codeparrot_training - Step 12377: {'lr': 0.0004445231064504751, 'samples': 6337536, 'steps': 12377, 'loss/train': 1.9981356859207153} 02/24/2022 17:35:52 - INFO - codeparrot_training - Step 12378: {'lr': 0.00044451282795670313, 'samples': 6338048, 'steps': 12378, 'loss/train': 1.5984537601470947} 02/24/2022 17:35:57 - INFO - codeparrot_training - Step 12379: {'lr': 0.0004445025486297, 'samples': 6338560, 'steps': 12379, 'loss/train': 2.0769050121307373} 02/24/2022 17:36:01 - INFO - codeparrot_training - Step 12380: {'lr': 0.00044449226846950964, 'samples': 6339072, 'steps': 12380, 'loss/train': 2.7042148113250732} 02/24/2022 17:36:07 - INFO - codeparrot_training - Step 12381: {'lr': 0.0004444819874761762, 'samples': 6339584, 'steps': 12381, 'loss/train': 2.5061631202697754} 02/24/2022 17:36:10 - INFO - codeparrot_training - Step 12382: {'lr': 0.0004444717056497436, 'samples': 6340096, 'steps': 12382, 'loss/train': 2.3966879844665527} 02/24/2022 17:36:16 - INFO - codeparrot_training - Step 12383: {'lr': 0.00044446142299025605, 'samples': 6340608, 'steps': 12383, 'loss/train': 1.4159033298492432} 02/24/2022 17:36:19 - INFO - codeparrot_training - Step 12384: {'lr': 0.0004444511394977575, 'samples': 6341120, 'steps': 12384, 'loss/train': 1.7560759782791138} 02/24/2022 17:36:25 - INFO - codeparrot_training - Step 12385: {'lr': 0.0004444408551722919, 'samples': 6341632, 'steps': 12385, 'loss/train': 3.5217506885528564} 02/24/2022 17:36:28 - INFO - codeparrot_training - Step 12386: {'lr': 0.00044443057001390354, 'samples': 6342144, 'steps': 12386, 'loss/train': 1.8106368780136108} 02/24/2022 17:36:34 - INFO - codeparrot_training - Step 12387: {'lr': 0.00044442028402263636, 'samples': 6342656, 'steps': 12387, 'loss/train': 1.9505242109298706} 02/24/2022 17:36:37 - INFO - codeparrot_training - Step 12388: {'lr': 0.00044440999719853435, 'samples': 6343168, 'steps': 12388, 'loss/train': 2.5981204509735107} 02/24/2022 17:36:43 - INFO - codeparrot_training - Step 12389: {'lr': 0.0004443997095416417, 'samples': 6343680, 'steps': 12389, 'loss/train': 2.685946464538574} 02/24/2022 17:36:46 - INFO - codeparrot_training - Step 12390: {'lr': 0.0004443894210520024, 'samples': 6344192, 'steps': 12390, 'loss/train': 0.8968327641487122} 02/24/2022 17:36:52 - INFO - codeparrot_training - Step 12391: {'lr': 0.0004443791317296606, 'samples': 6344704, 'steps': 12391, 'loss/train': 2.7141690254211426} 02/24/2022 17:36:55 - INFO - codeparrot_training - Step 12392: {'lr': 0.0004443688415746602, 'samples': 6345216, 'steps': 12392, 'loss/train': 1.8027390241622925} 02/24/2022 17:37:01 - INFO - codeparrot_training - Step 12393: {'lr': 0.0004443585505870456, 'samples': 6345728, 'steps': 12393, 'loss/train': 2.7816147804260254} 02/24/2022 17:37:04 - INFO - codeparrot_training - Step 12394: {'lr': 0.0004443482587668605, 'samples': 6346240, 'steps': 12394, 'loss/train': 2.1532182693481445} 02/24/2022 17:37:10 - INFO - codeparrot_training - Step 12395: {'lr': 0.00044433796611414924, 'samples': 6346752, 'steps': 12395, 'loss/train': 2.3728513717651367} 02/24/2022 17:37:16 - INFO - codeparrot_training - Step 12396: {'lr': 0.0004443276726289558, 'samples': 6347264, 'steps': 12396, 'loss/train': 2.032423734664917} 02/24/2022 17:37:19 - INFO - codeparrot_training - Step 12397: {'lr': 0.00044431737831132433, 'samples': 6347776, 'steps': 12397, 'loss/train': 2.170013189315796} 02/24/2022 17:37:25 - INFO - codeparrot_training - Step 12398: {'lr': 0.000444307083161299, 'samples': 6348288, 'steps': 12398, 'loss/train': 2.83723521232605} 02/24/2022 17:37:28 - INFO - codeparrot_training - Step 12399: {'lr': 0.00044429678717892366, 'samples': 6348800, 'steps': 12399, 'loss/train': 2.6319732666015625} 02/24/2022 17:37:34 - INFO - codeparrot_training - Step 12400: {'lr': 0.0004442864903642427, 'samples': 6349312, 'steps': 12400, 'loss/train': 2.7992143630981445} 02/24/2022 17:37:37 - INFO - codeparrot_training - Step 12401: {'lr': 0.00044427619271730014, 'samples': 6349824, 'steps': 12401, 'loss/train': 2.6513051986694336} 02/24/2022 17:37:43 - INFO - codeparrot_training - Step 12402: {'lr': 0.00044426589423814003, 'samples': 6350336, 'steps': 12402, 'loss/train': 1.4257817268371582} 02/24/2022 17:37:46 - INFO - codeparrot_training - Step 12403: {'lr': 0.00044425559492680645, 'samples': 6350848, 'steps': 12403, 'loss/train': 2.921034574508667} 02/24/2022 17:37:52 - INFO - codeparrot_training - Step 12404: {'lr': 0.00044424529478334364, 'samples': 6351360, 'steps': 12404, 'loss/train': 2.418518543243408} 02/24/2022 17:37:55 - INFO - codeparrot_training - Step 12405: {'lr': 0.00044423499380779566, 'samples': 6351872, 'steps': 12405, 'loss/train': 1.8839558362960815} 02/24/2022 17:38:01 - INFO - codeparrot_training - Step 12406: {'lr': 0.00044422469200020666, 'samples': 6352384, 'steps': 12406, 'loss/train': 1.413009762763977} 02/24/2022 17:38:04 - INFO - codeparrot_training - Step 12407: {'lr': 0.0004442143893606207, 'samples': 6352896, 'steps': 12407, 'loss/train': 2.4514241218566895} 02/24/2022 17:38:11 - INFO - codeparrot_training - Step 12408: {'lr': 0.000444204085889082, 'samples': 6353408, 'steps': 12408, 'loss/train': 2.5671350955963135} 02/24/2022 17:38:14 - INFO - codeparrot_training - Step 12409: {'lr': 0.00044419378158563465, 'samples': 6353920, 'steps': 12409, 'loss/train': 2.8062143325805664} 02/24/2022 17:38:20 - INFO - codeparrot_training - Step 12410: {'lr': 0.0004441834764503228, 'samples': 6354432, 'steps': 12410, 'loss/train': 2.195612907409668} 02/24/2022 17:38:23 - INFO - codeparrot_training - Step 12411: {'lr': 0.0004441731704831906, 'samples': 6354944, 'steps': 12411, 'loss/train': 2.5319154262542725} 02/24/2022 17:38:29 - INFO - codeparrot_training - Step 12412: {'lr': 0.0004441628636842822, 'samples': 6355456, 'steps': 12412, 'loss/train': 2.2401556968688965} 02/24/2022 17:38:32 - INFO - codeparrot_training - Step 12413: {'lr': 0.0004441525560536418, 'samples': 6355968, 'steps': 12413, 'loss/train': 1.3314924240112305} 02/24/2022 17:38:37 - INFO - codeparrot_training - Step 12414: {'lr': 0.0004441422475913134, 'samples': 6356480, 'steps': 12414, 'loss/train': 2.3003504276275635} 02/24/2022 17:38:41 - INFO - codeparrot_training - Step 12415: {'lr': 0.0004441319382973413, 'samples': 6356992, 'steps': 12415, 'loss/train': 2.1423094272613525} 02/24/2022 17:38:46 - INFO - codeparrot_training - Step 12416: {'lr': 0.00044412162817176966, 'samples': 6357504, 'steps': 12416, 'loss/train': 2.11195707321167} 02/24/2022 17:38:50 - INFO - codeparrot_training - Step 12417: {'lr': 0.0004441113172146426, 'samples': 6358016, 'steps': 12417, 'loss/train': 1.632970929145813} 02/24/2022 17:38:56 - INFO - codeparrot_training - Step 12418: {'lr': 0.00044410100542600423, 'samples': 6358528, 'steps': 12418, 'loss/train': 2.1068115234375} 02/24/2022 17:39:00 - INFO - codeparrot_training - Step 12419: {'lr': 0.00044409069280589887, 'samples': 6359040, 'steps': 12419, 'loss/train': 2.087383985519409} 02/24/2022 17:39:05 - INFO - codeparrot_training - Step 12420: {'lr': 0.0004440803793543705, 'samples': 6359552, 'steps': 12420, 'loss/train': 2.1730427742004395} 02/24/2022 17:39:09 - INFO - codeparrot_training - Step 12421: {'lr': 0.00044407006507146354, 'samples': 6360064, 'steps': 12421, 'loss/train': 1.5413535833358765} 02/24/2022 17:39:15 - INFO - codeparrot_training - Step 12422: {'lr': 0.000444059749957222, 'samples': 6360576, 'steps': 12422, 'loss/train': 5.75645637512207} 02/24/2022 17:39:18 - INFO - codeparrot_training - Step 12423: {'lr': 0.00044404943401169005, 'samples': 6361088, 'steps': 12423, 'loss/train': 2.319354295730591} 02/24/2022 17:39:21 - INFO - codeparrot_training - Step 12424: {'lr': 0.00044403911723491196, 'samples': 6361600, 'steps': 12424, 'loss/train': 2.6930625438690186} 02/24/2022 17:39:27 - INFO - codeparrot_training - Step 12425: {'lr': 0.000444028799626932, 'samples': 6362112, 'steps': 12425, 'loss/train': 2.0183658599853516} 02/24/2022 17:39:30 - INFO - codeparrot_training - Step 12426: {'lr': 0.0004440184811877942, 'samples': 6362624, 'steps': 12426, 'loss/train': 1.2730671167373657} 02/24/2022 17:39:36 - INFO - codeparrot_training - Step 12427: {'lr': 0.0004440081619175428, 'samples': 6363136, 'steps': 12427, 'loss/train': 2.4213807582855225} 02/24/2022 17:39:39 - INFO - codeparrot_training - Step 12428: {'lr': 0.00044399784181622216, 'samples': 6363648, 'steps': 12428, 'loss/train': 2.496400833129883} 02/24/2022 17:39:46 - INFO - codeparrot_training - Step 12429: {'lr': 0.0004439875208838763, 'samples': 6364160, 'steps': 12429, 'loss/train': 2.9761877059936523} 02/24/2022 17:39:49 - INFO - codeparrot_training - Step 12430: {'lr': 0.00044397719912054944, 'samples': 6364672, 'steps': 12430, 'loss/train': 2.6443240642547607} 02/24/2022 17:39:55 - INFO - codeparrot_training - Step 12431: {'lr': 0.00044396687652628586, 'samples': 6365184, 'steps': 12431, 'loss/train': 1.5773134231567383} 02/24/2022 17:39:58 - INFO - codeparrot_training - Step 12432: {'lr': 0.00044395655310112985, 'samples': 6365696, 'steps': 12432, 'loss/train': 0.9131450653076172} 02/24/2022 17:40:04 - INFO - codeparrot_training - Step 12433: {'lr': 0.00044394622884512554, 'samples': 6366208, 'steps': 12433, 'loss/train': 2.3642284870147705} 02/24/2022 17:40:09 - INFO - codeparrot_training - Step 12434: {'lr': 0.00044393590375831716, 'samples': 6366720, 'steps': 12434, 'loss/train': 2.6101467609405518} 02/24/2022 17:40:12 - INFO - codeparrot_training - Step 12435: {'lr': 0.00044392557784074895, 'samples': 6367232, 'steps': 12435, 'loss/train': 2.382425308227539} 02/24/2022 17:40:18 - INFO - codeparrot_training - Step 12436: {'lr': 0.0004439152510924651, 'samples': 6367744, 'steps': 12436, 'loss/train': 2.946174144744873} 02/24/2022 17:40:22 - INFO - codeparrot_training - Step 12437: {'lr': 0.0004439049235135099, 'samples': 6368256, 'steps': 12437, 'loss/train': 1.6053688526153564} 02/24/2022 17:40:28 - INFO - codeparrot_training - Step 12438: {'lr': 0.0004438945951039276, 'samples': 6368768, 'steps': 12438, 'loss/train': 2.4451630115509033} 02/24/2022 17:40:31 - INFO - codeparrot_training - Step 12439: {'lr': 0.0004438842658637624, 'samples': 6369280, 'steps': 12439, 'loss/train': 2.1276416778564453} 02/24/2022 17:40:37 - INFO - codeparrot_training - Step 12440: {'lr': 0.0004438739357930586, 'samples': 6369792, 'steps': 12440, 'loss/train': 2.3103365898132324} 02/24/2022 17:40:40 - INFO - codeparrot_training - Step 12441: {'lr': 0.00044386360489186047, 'samples': 6370304, 'steps': 12441, 'loss/train': 2.244352340698242} 02/24/2022 17:40:46 - INFO - codeparrot_training - Step 12442: {'lr': 0.00044385327316021214, 'samples': 6370816, 'steps': 12442, 'loss/train': 3.0721912384033203} 02/24/2022 17:40:49 - INFO - codeparrot_training - Step 12443: {'lr': 0.000443842940598158, 'samples': 6371328, 'steps': 12443, 'loss/train': 1.8053581714630127} 02/24/2022 17:40:55 - INFO - codeparrot_training - Step 12444: {'lr': 0.00044383260720574214, 'samples': 6371840, 'steps': 12444, 'loss/train': 1.652478575706482} 02/24/2022 17:40:58 - INFO - codeparrot_training - Step 12445: {'lr': 0.00044382227298300905, 'samples': 6372352, 'steps': 12445, 'loss/train': 2.2976057529449463} 02/24/2022 17:41:04 - INFO - codeparrot_training - Step 12446: {'lr': 0.0004438119379300028, 'samples': 6372864, 'steps': 12446, 'loss/train': 1.7779102325439453} 02/24/2022 17:41:07 - INFO - codeparrot_training - Step 12447: {'lr': 0.00044380160204676787, 'samples': 6373376, 'steps': 12447, 'loss/train': 2.2807915210723877} 02/24/2022 17:41:13 - INFO - codeparrot_training - Step 12448: {'lr': 0.00044379126533334836, 'samples': 6373888, 'steps': 12448, 'loss/train': 2.2795181274414062} 02/24/2022 17:41:16 - INFO - codeparrot_training - Step 12449: {'lr': 0.00044378092778978864, 'samples': 6374400, 'steps': 12449, 'loss/train': 1.6085028648376465} 02/24/2022 17:41:22 - INFO - codeparrot_training - Step 12450: {'lr': 0.00044377058941613283, 'samples': 6374912, 'steps': 12450, 'loss/train': 1.0612393617630005} 02/24/2022 17:41:25 - INFO - codeparrot_training - Step 12451: {'lr': 0.0004437602502124255, 'samples': 6375424, 'steps': 12451, 'loss/train': 1.4914125204086304} 02/24/2022 17:41:31 - INFO - codeparrot_training - Step 12452: {'lr': 0.0004437499101787107, 'samples': 6375936, 'steps': 12452, 'loss/train': 0.9224465489387512} 02/24/2022 17:41:34 - INFO - codeparrot_training - Step 12453: {'lr': 0.0004437395693150328, 'samples': 6376448, 'steps': 12453, 'loss/train': 2.0059292316436768} 02/24/2022 17:41:41 - INFO - codeparrot_training - Step 12454: {'lr': 0.0004437292276214361, 'samples': 6376960, 'steps': 12454, 'loss/train': 3.6045403480529785} 02/24/2022 17:41:44 - INFO - codeparrot_training - Step 12455: {'lr': 0.000443718885097965, 'samples': 6377472, 'steps': 12455, 'loss/train': 1.1657453775405884} 02/24/2022 17:41:50 - INFO - codeparrot_training - Step 12456: {'lr': 0.0004437085417446636, 'samples': 6377984, 'steps': 12456, 'loss/train': 3.50653338432312} 02/24/2022 17:41:53 - INFO - codeparrot_training - Step 12457: {'lr': 0.0004436981975615764, 'samples': 6378496, 'steps': 12457, 'loss/train': 1.7108455896377563} 02/24/2022 17:41:59 - INFO - codeparrot_training - Step 12458: {'lr': 0.00044368785254874754, 'samples': 6379008, 'steps': 12458, 'loss/train': 3.026381492614746} 02/24/2022 17:42:02 - INFO - codeparrot_training - Step 12459: {'lr': 0.00044367750670622143, 'samples': 6379520, 'steps': 12459, 'loss/train': 1.752224326133728} 02/24/2022 17:42:08 - INFO - codeparrot_training - Step 12460: {'lr': 0.0004436671600340424, 'samples': 6380032, 'steps': 12460, 'loss/train': 2.599832773208618} 02/24/2022 17:42:11 - INFO - codeparrot_training - Step 12461: {'lr': 0.00044365681253225476, 'samples': 6380544, 'steps': 12461, 'loss/train': 2.228529930114746} 02/24/2022 17:42:17 - INFO - codeparrot_training - Step 12462: {'lr': 0.0004436464642009029, 'samples': 6381056, 'steps': 12462, 'loss/train': 2.433157444000244} 02/24/2022 17:42:20 - INFO - codeparrot_training - Step 12463: {'lr': 0.00044363611504003096, 'samples': 6381568, 'steps': 12463, 'loss/train': 1.9093401432037354} 02/24/2022 17:42:26 - INFO - codeparrot_training - Step 12464: {'lr': 0.00044362576504968344, 'samples': 6382080, 'steps': 12464, 'loss/train': 0.8747289180755615} 02/24/2022 17:42:30 - INFO - codeparrot_training - Step 12465: {'lr': 0.0004436154142299046, 'samples': 6382592, 'steps': 12465, 'loss/train': 1.8154183626174927} 02/24/2022 17:42:35 - INFO - codeparrot_training - Step 12466: {'lr': 0.00044360506258073884, 'samples': 6383104, 'steps': 12466, 'loss/train': 2.9957594871520996} 02/24/2022 17:42:39 - INFO - codeparrot_training - Step 12467: {'lr': 0.0004435947101022305, 'samples': 6383616, 'steps': 12467, 'loss/train': 1.6825523376464844} 02/24/2022 17:42:44 - INFO - codeparrot_training - Step 12468: {'lr': 0.0004435843567944239, 'samples': 6384128, 'steps': 12468, 'loss/train': 1.962235450744629} 02/24/2022 17:42:48 - INFO - codeparrot_training - Step 12469: {'lr': 0.0004435740026573633, 'samples': 6384640, 'steps': 12469, 'loss/train': 2.2906713485717773} 02/24/2022 17:42:53 - INFO - codeparrot_training - Step 12470: {'lr': 0.0004435636476910932, 'samples': 6385152, 'steps': 12470, 'loss/train': 1.7940757274627686} 02/24/2022 17:42:57 - INFO - codeparrot_training - Step 12471: {'lr': 0.00044355329189565783, 'samples': 6385664, 'steps': 12471, 'loss/train': 2.1791868209838867} 02/24/2022 17:43:02 - INFO - codeparrot_training - Step 12472: {'lr': 0.00044354293527110167, 'samples': 6386176, 'steps': 12472, 'loss/train': 1.994907021522522} 02/24/2022 17:43:06 - INFO - codeparrot_training - Step 12473: {'lr': 0.000443532577817469, 'samples': 6386688, 'steps': 12473, 'loss/train': 2.4598143100738525} 02/24/2022 17:43:12 - INFO - codeparrot_training - Step 12474: {'lr': 0.0004435222195348043, 'samples': 6387200, 'steps': 12474, 'loss/train': 2.6731014251708984} 02/24/2022 17:43:15 - INFO - codeparrot_training - Step 12475: {'lr': 0.00044351186042315184, 'samples': 6387712, 'steps': 12475, 'loss/train': 2.357680082321167} 02/24/2022 17:43:21 - INFO - codeparrot_training - Step 12476: {'lr': 0.000443501500482556, 'samples': 6388224, 'steps': 12476, 'loss/train': 2.2906453609466553} 02/24/2022 17:43:24 - INFO - codeparrot_training - Step 12477: {'lr': 0.0004434911397130612, 'samples': 6388736, 'steps': 12477, 'loss/train': 2.5177316665649414} 02/24/2022 17:43:30 - INFO - codeparrot_training - Step 12478: {'lr': 0.0004434807781147117, 'samples': 6389248, 'steps': 12478, 'loss/train': 1.243457555770874} 02/24/2022 17:43:33 - INFO - codeparrot_training - Step 12479: {'lr': 0.0004434704156875521, 'samples': 6389760, 'steps': 12479, 'loss/train': 2.449394941329956} 02/24/2022 17:43:39 - INFO - codeparrot_training - Step 12480: {'lr': 0.00044346005243162654, 'samples': 6390272, 'steps': 12480, 'loss/train': 1.990755558013916} 02/24/2022 17:43:43 - INFO - codeparrot_training - Step 12481: {'lr': 0.0004434496883469796, 'samples': 6390784, 'steps': 12481, 'loss/train': 2.4250757694244385} 02/24/2022 17:43:49 - INFO - codeparrot_training - Step 12482: {'lr': 0.0004434393234336557, 'samples': 6391296, 'steps': 12482, 'loss/train': 2.508273124694824} 02/24/2022 17:43:52 - INFO - codeparrot_training - Step 12483: {'lr': 0.0004434289576916991, 'samples': 6391808, 'steps': 12483, 'loss/train': 2.3634986877441406} 02/24/2022 17:43:56 - INFO - codeparrot_training - Step 12484: {'lr': 0.00044341859112115425, 'samples': 6392320, 'steps': 12484, 'loss/train': 4.695265769958496} 02/24/2022 17:44:01 - INFO - codeparrot_training - Step 12485: {'lr': 0.00044340822372206557, 'samples': 6392832, 'steps': 12485, 'loss/train': 0.9202659130096436} 02/24/2022 17:44:05 - INFO - codeparrot_training - Step 12486: {'lr': 0.00044339785549447756, 'samples': 6393344, 'steps': 12486, 'loss/train': 2.216762065887451} 02/24/2022 17:44:11 - INFO - codeparrot_training - Step 12487: {'lr': 0.00044338748643843446, 'samples': 6393856, 'steps': 12487, 'loss/train': 3.6331098079681396} 02/24/2022 17:44:14 - INFO - codeparrot_training - Step 12488: {'lr': 0.00044337711655398083, 'samples': 6394368, 'steps': 12488, 'loss/train': 1.5643867254257202} 02/24/2022 17:44:20 - INFO - codeparrot_training - Step 12489: {'lr': 0.00044336674584116096, 'samples': 6394880, 'steps': 12489, 'loss/train': 2.4457035064697266} 02/24/2022 17:44:23 - INFO - codeparrot_training - Step 12490: {'lr': 0.0004433563743000195, 'samples': 6395392, 'steps': 12490, 'loss/train': 2.6636414527893066} 02/24/2022 17:44:29 - INFO - codeparrot_training - Step 12491: {'lr': 0.0004433460019306006, 'samples': 6395904, 'steps': 12491, 'loss/train': 0.632786750793457} 02/24/2022 17:44:32 - INFO - codeparrot_training - Step 12492: {'lr': 0.00044333562873294884, 'samples': 6396416, 'steps': 12492, 'loss/train': 1.246692180633545} 02/24/2022 17:44:38 - INFO - codeparrot_training - Step 12493: {'lr': 0.00044332525470710865, 'samples': 6396928, 'steps': 12493, 'loss/train': 2.4409217834472656} 02/24/2022 17:44:41 - INFO - codeparrot_training - Step 12494: {'lr': 0.0004433148798531245, 'samples': 6397440, 'steps': 12494, 'loss/train': 2.7363319396972656} 02/24/2022 17:44:47 - INFO - codeparrot_training - Step 12495: {'lr': 0.0004433045041710407, 'samples': 6397952, 'steps': 12495, 'loss/train': 1.8529976606369019} 02/24/2022 17:44:50 - INFO - codeparrot_training - Step 12496: {'lr': 0.0004432941276609018, 'samples': 6398464, 'steps': 12496, 'loss/train': 2.633256196975708} 02/24/2022 17:44:56 - INFO - codeparrot_training - Step 12497: {'lr': 0.00044328375032275227, 'samples': 6398976, 'steps': 12497, 'loss/train': 2.59234619140625} 02/24/2022 17:44:59 - INFO - codeparrot_training - Step 12498: {'lr': 0.00044327337215663656, 'samples': 6399488, 'steps': 12498, 'loss/train': 1.4717844724655151} 02/24/2022 17:45:05 - INFO - codeparrot_training - Step 12499: {'lr': 0.000443262993162599, 'samples': 6400000, 'steps': 12499, 'loss/train': 1.216798186302185} 02/24/2022 17:45:08 - INFO - codeparrot_training - Step 12500: {'lr': 0.0004432526133406842, 'samples': 6400512, 'steps': 12500, 'loss/train': 2.088089942932129} 02/24/2022 17:45:14 - INFO - codeparrot_training - Step 12501: {'lr': 0.00044324223269093666, 'samples': 6401024, 'steps': 12501, 'loss/train': 1.7118074893951416} 02/24/2022 17:45:18 - INFO - codeparrot_training - Step 12502: {'lr': 0.00044323185121340064, 'samples': 6401536, 'steps': 12502, 'loss/train': 2.1711461544036865} 02/24/2022 17:45:23 - INFO - codeparrot_training - Step 12503: {'lr': 0.00044322146890812076, 'samples': 6402048, 'steps': 12503, 'loss/train': 1.2165560722351074} 02/24/2022 17:45:27 - INFO - codeparrot_training - Step 12504: {'lr': 0.0004432110857751415, 'samples': 6402560, 'steps': 12504, 'loss/train': 1.8111298084259033} 02/24/2022 17:45:32 - INFO - codeparrot_training - Step 12505: {'lr': 0.0004432007018145072, 'samples': 6403072, 'steps': 12505, 'loss/train': 2.0570878982543945} 02/24/2022 17:45:36 - INFO - codeparrot_training - Step 12506: {'lr': 0.00044319031702626255, 'samples': 6403584, 'steps': 12506, 'loss/train': 1.601704478263855} 02/24/2022 17:45:41 - INFO - codeparrot_training - Step 12507: {'lr': 0.0004431799314104519, 'samples': 6404096, 'steps': 12507, 'loss/train': 1.5467791557312012} 02/24/2022 17:45:45 - INFO - codeparrot_training - Step 12508: {'lr': 0.0004431695449671197, 'samples': 6404608, 'steps': 12508, 'loss/train': 1.7713934183120728} 02/24/2022 17:45:50 - INFO - codeparrot_training - Step 12509: {'lr': 0.00044315915769631054, 'samples': 6405120, 'steps': 12509, 'loss/train': 1.8101553916931152} 02/24/2022 17:45:54 - INFO - codeparrot_training - Step 12510: {'lr': 0.0004431487695980689, 'samples': 6405632, 'steps': 12510, 'loss/train': 2.5609304904937744} 02/24/2022 17:45:59 - INFO - codeparrot_training - Step 12511: {'lr': 0.0004431383806724393, 'samples': 6406144, 'steps': 12511, 'loss/train': 1.944899082183838} 02/24/2022 17:46:05 - INFO - codeparrot_training - Step 12512: {'lr': 0.0004431279909194661, 'samples': 6406656, 'steps': 12512, 'loss/train': 1.7914735078811646} 02/24/2022 17:46:09 - INFO - codeparrot_training - Step 12513: {'lr': 0.000443117600339194, 'samples': 6407168, 'steps': 12513, 'loss/train': 1.561868667602539} 02/24/2022 17:46:14 - INFO - codeparrot_training - Step 12514: {'lr': 0.0004431072089316674, 'samples': 6407680, 'steps': 12514, 'loss/train': 2.2434191703796387} 02/24/2022 17:46:18 - INFO - codeparrot_training - Step 12515: {'lr': 0.0004430968166969308, 'samples': 6408192, 'steps': 12515, 'loss/train': 1.392062783241272} 02/24/2022 17:46:23 - INFO - codeparrot_training - Step 12516: {'lr': 0.00044308642363502884, 'samples': 6408704, 'steps': 12516, 'loss/train': 1.9453023672103882} 02/24/2022 17:46:27 - INFO - codeparrot_training - Step 12517: {'lr': 0.00044307602974600594, 'samples': 6409216, 'steps': 12517, 'loss/train': 2.2774412631988525} 02/24/2022 17:46:32 - INFO - codeparrot_training - Step 12518: {'lr': 0.00044306563502990656, 'samples': 6409728, 'steps': 12518, 'loss/train': 1.5544239282608032} 02/24/2022 17:46:36 - INFO - codeparrot_training - Step 12519: {'lr': 0.0004430552394867753, 'samples': 6410240, 'steps': 12519, 'loss/train': 0.8071713447570801} 02/24/2022 17:46:41 - INFO - codeparrot_training - Step 12520: {'lr': 0.0004430448431166567, 'samples': 6410752, 'steps': 12520, 'loss/train': 2.0709426403045654} 02/24/2022 17:46:45 - INFO - codeparrot_training - Step 12521: {'lr': 0.00044303444591959533, 'samples': 6411264, 'steps': 12521, 'loss/train': 2.409284830093384} 02/24/2022 17:46:51 - INFO - codeparrot_training - Step 12522: {'lr': 0.00044302404789563573, 'samples': 6411776, 'steps': 12522, 'loss/train': 1.6684871912002563} 02/24/2022 17:46:54 - INFO - codeparrot_training - Step 12523: {'lr': 0.0004430136490448223, 'samples': 6412288, 'steps': 12523, 'loss/train': 1.6257758140563965} 02/24/2022 17:47:00 - INFO - codeparrot_training - Step 12524: {'lr': 0.0004430032493671998, 'samples': 6412800, 'steps': 12524, 'loss/train': 2.116621255874634} 02/24/2022 17:47:03 - INFO - codeparrot_training - Step 12525: {'lr': 0.0004429928488628126, 'samples': 6413312, 'steps': 12525, 'loss/train': 2.120816707611084} 02/24/2022 17:47:09 - INFO - codeparrot_training - Step 12526: {'lr': 0.00044298244753170535, 'samples': 6413824, 'steps': 12526, 'loss/train': 2.6973717212677} 02/24/2022 17:47:12 - INFO - codeparrot_training - Step 12527: {'lr': 0.00044297204537392253, 'samples': 6414336, 'steps': 12527, 'loss/train': 2.123155117034912} 02/24/2022 17:47:18 - INFO - codeparrot_training - Step 12528: {'lr': 0.00044296164238950874, 'samples': 6414848, 'steps': 12528, 'loss/train': 2.4255053997039795} 02/24/2022 17:47:21 - INFO - codeparrot_training - Step 12529: {'lr': 0.0004429512385785086, 'samples': 6415360, 'steps': 12529, 'loss/train': 2.0674798488616943} 02/24/2022 17:47:27 - INFO - codeparrot_training - Step 12530: {'lr': 0.0004429408339409666, 'samples': 6415872, 'steps': 12530, 'loss/train': 2.0020055770874023} 02/24/2022 17:47:31 - INFO - codeparrot_training - Step 12531: {'lr': 0.00044293042847692735, 'samples': 6416384, 'steps': 12531, 'loss/train': 2.289703845977783} 02/24/2022 17:47:37 - INFO - codeparrot_training - Step 12532: {'lr': 0.00044292002218643533, 'samples': 6416896, 'steps': 12532, 'loss/train': 2.27740216255188} 02/24/2022 17:47:40 - INFO - codeparrot_training - Step 12533: {'lr': 0.00044290961506953525, 'samples': 6417408, 'steps': 12533, 'loss/train': 1.8238201141357422} 02/24/2022 17:47:46 - INFO - codeparrot_training - Step 12534: {'lr': 0.0004428992071262716, 'samples': 6417920, 'steps': 12534, 'loss/train': 1.3410807847976685} 02/24/2022 17:47:49 - INFO - codeparrot_training - Step 12535: {'lr': 0.00044288879835668903, 'samples': 6418432, 'steps': 12535, 'loss/train': 2.37658429145813} 02/24/2022 17:47:55 - INFO - codeparrot_training - Step 12536: {'lr': 0.0004428783887608321, 'samples': 6418944, 'steps': 12536, 'loss/train': 2.2396929264068604} 02/24/2022 17:47:58 - INFO - codeparrot_training - Step 12537: {'lr': 0.0004428679783387454, 'samples': 6419456, 'steps': 12537, 'loss/train': 2.3694121837615967} 02/24/2022 17:48:04 - INFO - codeparrot_training - Step 12538: {'lr': 0.00044285756709047354, 'samples': 6419968, 'steps': 12538, 'loss/train': 2.369859218597412} 02/24/2022 17:48:08 - INFO - codeparrot_training - Step 12539: {'lr': 0.0004428471550160611, 'samples': 6420480, 'steps': 12539, 'loss/train': 0.912290096282959} 02/24/2022 17:48:11 - INFO - codeparrot_training - Step 12540: {'lr': 0.00044283674211555266, 'samples': 6420992, 'steps': 12540, 'loss/train': 0.6784152984619141} 02/24/2022 17:48:17 - INFO - codeparrot_training - Step 12541: {'lr': 0.0004428263283889928, 'samples': 6421504, 'steps': 12541, 'loss/train': 1.8253849744796753} 02/24/2022 17:48:20 - INFO - codeparrot_training - Step 12542: {'lr': 0.0004428159138364263, 'samples': 6422016, 'steps': 12542, 'loss/train': 2.7968780994415283} 02/24/2022 17:48:26 - INFO - codeparrot_training - Step 12543: {'lr': 0.0004428054984578975, 'samples': 6422528, 'steps': 12543, 'loss/train': 1.9343206882476807} 02/24/2022 17:48:29 - INFO - codeparrot_training - Step 12544: {'lr': 0.0004427950822534513, 'samples': 6423040, 'steps': 12544, 'loss/train': 2.3636536598205566} 02/24/2022 17:48:35 - INFO - codeparrot_training - Step 12545: {'lr': 0.0004427846652231321, 'samples': 6423552, 'steps': 12545, 'loss/train': 0.32711923122406006} 02/24/2022 17:48:38 - INFO - codeparrot_training - Step 12546: {'lr': 0.0004427742473669847, 'samples': 6424064, 'steps': 12546, 'loss/train': 1.8101640939712524} 02/24/2022 17:48:45 - INFO - codeparrot_training - Step 12547: {'lr': 0.00044276382868505356, 'samples': 6424576, 'steps': 12547, 'loss/train': 2.392697334289551} 02/24/2022 17:48:48 - INFO - codeparrot_training - Step 12548: {'lr': 0.0004427534091773834, 'samples': 6425088, 'steps': 12548, 'loss/train': 1.978995680809021} 02/24/2022 17:48:54 - INFO - codeparrot_training - Step 12549: {'lr': 0.00044274298884401886, 'samples': 6425600, 'steps': 12549, 'loss/train': 2.957902431488037} 02/24/2022 17:48:57 - INFO - codeparrot_training - Step 12550: {'lr': 0.0004427325676850045, 'samples': 6426112, 'steps': 12550, 'loss/train': 1.2002750635147095} 02/24/2022 17:49:03 - INFO - codeparrot_training - Step 12551: {'lr': 0.00044272214570038513, 'samples': 6426624, 'steps': 12551, 'loss/train': 2.8818886280059814} 02/24/2022 17:49:06 - INFO - codeparrot_training - Step 12552: {'lr': 0.00044271172289020525, 'samples': 6427136, 'steps': 12552, 'loss/train': 2.1332151889801025} 02/24/2022 17:49:12 - INFO - codeparrot_training - Step 12553: {'lr': 0.00044270129925450945, 'samples': 6427648, 'steps': 12553, 'loss/train': 1.5723611116409302} 02/24/2022 17:49:15 - INFO - codeparrot_training - Step 12554: {'lr': 0.00044269087479334256, 'samples': 6428160, 'steps': 12554, 'loss/train': 2.017009735107422} 02/24/2022 17:49:21 - INFO - codeparrot_training - Step 12555: {'lr': 0.00044268044950674913, 'samples': 6428672, 'steps': 12555, 'loss/train': 1.7089089155197144} 02/24/2022 17:49:24 - INFO - codeparrot_training - Step 12556: {'lr': 0.0004426700233947738, 'samples': 6429184, 'steps': 12556, 'loss/train': 2.3912200927734375} 02/24/2022 17:49:31 - INFO - codeparrot_training - Step 12557: {'lr': 0.00044265959645746136, 'samples': 6429696, 'steps': 12557, 'loss/train': 1.775288701057434} 02/24/2022 17:49:34 - INFO - codeparrot_training - Step 12558: {'lr': 0.0004426491686948563, 'samples': 6430208, 'steps': 12558, 'loss/train': 2.274704933166504} 02/24/2022 17:49:40 - INFO - codeparrot_training - Step 12559: {'lr': 0.00044263874010700343, 'samples': 6430720, 'steps': 12559, 'loss/train': 1.691240906715393} 02/24/2022 17:49:43 - INFO - codeparrot_training - Step 12560: {'lr': 0.0004426283106939473, 'samples': 6431232, 'steps': 12560, 'loss/train': 2.5085697174072266} 02/24/2022 17:49:48 - INFO - codeparrot_training - Step 12561: {'lr': 0.0004426178804557327, 'samples': 6431744, 'steps': 12561, 'loss/train': 1.1207683086395264} 02/24/2022 17:49:52 - INFO - codeparrot_training - Step 12562: {'lr': 0.0004426074493924043, 'samples': 6432256, 'steps': 12562, 'loss/train': 2.3670997619628906} 02/24/2022 17:49:57 - INFO - codeparrot_training - Step 12563: {'lr': 0.00044259701750400674, 'samples': 6432768, 'steps': 12563, 'loss/train': 2.062629461288452} 02/24/2022 17:50:01 - INFO - codeparrot_training - Step 12564: {'lr': 0.00044258658479058463, 'samples': 6433280, 'steps': 12564, 'loss/train': 2.1066765785217285} 02/24/2022 17:50:06 - INFO - codeparrot_training - Step 12565: {'lr': 0.00044257615125218273, 'samples': 6433792, 'steps': 12565, 'loss/train': 1.4604438543319702} 02/24/2022 17:50:10 - INFO - codeparrot_training - Step 12566: {'lr': 0.00044256571688884583, 'samples': 6434304, 'steps': 12566, 'loss/train': 1.3201088905334473} 02/24/2022 17:50:15 - INFO - codeparrot_training - Step 12567: {'lr': 0.00044255528170061853, 'samples': 6434816, 'steps': 12567, 'loss/train': 2.093223810195923} 02/24/2022 17:50:19 - INFO - codeparrot_training - Step 12568: {'lr': 0.00044254484568754556, 'samples': 6435328, 'steps': 12568, 'loss/train': 1.8225075006484985} 02/24/2022 17:50:25 - INFO - codeparrot_training - Step 12569: {'lr': 0.0004425344088496716, 'samples': 6435840, 'steps': 12569, 'loss/train': 1.8812451362609863} 02/24/2022 17:50:29 - INFO - codeparrot_training - Step 12570: {'lr': 0.00044252397118704133, 'samples': 6436352, 'steps': 12570, 'loss/train': 1.8809235095977783} 02/24/2022 17:50:34 - INFO - codeparrot_training - Step 12571: {'lr': 0.0004425135326996995, 'samples': 6436864, 'steps': 12571, 'loss/train': 2.5634331703186035} 02/24/2022 17:50:38 - INFO - codeparrot_training - Step 12572: {'lr': 0.0004425030933876909, 'samples': 6437376, 'steps': 12572, 'loss/train': 7.52122163772583} 02/24/2022 17:50:43 - INFO - codeparrot_training - Step 12573: {'lr': 0.00044249265325106013, 'samples': 6437888, 'steps': 12573, 'loss/train': 2.3944826126098633} 02/24/2022 17:50:47 - INFO - codeparrot_training - Step 12574: {'lr': 0.000442482212289852, 'samples': 6438400, 'steps': 12574, 'loss/train': 1.7205818891525269} 02/24/2022 17:50:52 - INFO - codeparrot_training - Step 12575: {'lr': 0.00044247177050411114, 'samples': 6438912, 'steps': 12575, 'loss/train': 2.094712018966675} 02/24/2022 17:50:56 - INFO - codeparrot_training - Step 12576: {'lr': 0.00044246132789388235, 'samples': 6439424, 'steps': 12576, 'loss/train': 2.1907360553741455} 02/24/2022 17:51:01 - INFO - codeparrot_training - Step 12577: {'lr': 0.00044245088445921035, 'samples': 6439936, 'steps': 12577, 'loss/train': 1.7764767408370972} 02/24/2022 17:51:05 - INFO - codeparrot_training - Step 12578: {'lr': 0.00044244044020013985, 'samples': 6440448, 'steps': 12578, 'loss/train': 1.8151516914367676} 02/24/2022 17:51:11 - INFO - codeparrot_training - Step 12579: {'lr': 0.0004424299951167156, 'samples': 6440960, 'steps': 12579, 'loss/train': 2.9270567893981934} 02/24/2022 17:51:14 - INFO - codeparrot_training - Step 12580: {'lr': 0.0004424195492089824, 'samples': 6441472, 'steps': 12580, 'loss/train': 2.571474552154541} 02/24/2022 17:51:20 - INFO - codeparrot_training - Step 12581: {'lr': 0.0004424091024769849, 'samples': 6441984, 'steps': 12581, 'loss/train': 1.6912142038345337} 02/24/2022 17:51:23 - INFO - codeparrot_training - Step 12582: {'lr': 0.00044239865492076794, 'samples': 6442496, 'steps': 12582, 'loss/train': 1.6894973516464233} 02/24/2022 17:51:29 - INFO - codeparrot_training - Step 12583: {'lr': 0.0004423882065403762, 'samples': 6443008, 'steps': 12583, 'loss/train': 2.1041579246520996} 02/24/2022 17:51:32 - INFO - codeparrot_training - Step 12584: {'lr': 0.0004423777573358545, 'samples': 6443520, 'steps': 12584, 'loss/train': 1.6610794067382812} 02/24/2022 17:51:38 - INFO - codeparrot_training - Step 12585: {'lr': 0.0004423673073072476, 'samples': 6444032, 'steps': 12585, 'loss/train': 1.4454855918884277} 02/24/2022 17:51:41 - INFO - codeparrot_training - Step 12586: {'lr': 0.0004423568564546002, 'samples': 6444544, 'steps': 12586, 'loss/train': 2.52445650100708} 02/24/2022 17:51:47 - INFO - codeparrot_training - Step 12587: {'lr': 0.00044234640477795707, 'samples': 6445056, 'steps': 12587, 'loss/train': 2.1523823738098145} 02/24/2022 17:51:50 - INFO - codeparrot_training - Step 12588: {'lr': 0.0004423359522773631, 'samples': 6445568, 'steps': 12588, 'loss/train': 2.0760316848754883} 02/24/2022 17:51:56 - INFO - codeparrot_training - Step 12589: {'lr': 0.00044232549895286294, 'samples': 6446080, 'steps': 12589, 'loss/train': 1.85196852684021} 02/24/2022 17:51:59 - INFO - codeparrot_training - Step 12590: {'lr': 0.00044231504480450145, 'samples': 6446592, 'steps': 12590, 'loss/train': 2.3095033168792725} 02/24/2022 17:52:06 - INFO - codeparrot_training - Step 12591: {'lr': 0.0004423045898323233, 'samples': 6447104, 'steps': 12591, 'loss/train': 2.6988797187805176} 02/24/2022 17:52:09 - INFO - codeparrot_training - Step 12592: {'lr': 0.0004422941340363734, 'samples': 6447616, 'steps': 12592, 'loss/train': 3.069305896759033} 02/24/2022 17:52:14 - INFO - codeparrot_training - Step 12593: {'lr': 0.0004422836774166965, 'samples': 6448128, 'steps': 12593, 'loss/train': 2.4666504859924316} 02/24/2022 17:52:18 - INFO - codeparrot_training - Step 12594: {'lr': 0.00044227321997333737, 'samples': 6448640, 'steps': 12594, 'loss/train': 2.4254045486450195} 02/24/2022 17:52:24 - INFO - codeparrot_training - Step 12595: {'lr': 0.0004422627617063408, 'samples': 6449152, 'steps': 12595, 'loss/train': 2.7385494709014893} 02/24/2022 17:52:27 - INFO - codeparrot_training - Step 12596: {'lr': 0.00044225230261575165, 'samples': 6449664, 'steps': 12596, 'loss/train': 0.8888677358627319} 02/24/2022 17:52:33 - INFO - codeparrot_training - Step 12597: {'lr': 0.00044224184270161466, 'samples': 6450176, 'steps': 12597, 'loss/train': 2.4945390224456787} 02/24/2022 17:52:38 - INFO - codeparrot_training - Step 12598: {'lr': 0.0004422313819639747, 'samples': 6450688, 'steps': 12598, 'loss/train': 1.8744642734527588} 02/24/2022 17:52:41 - INFO - codeparrot_training - Step 12599: {'lr': 0.0004422209204028765, 'samples': 6451200, 'steps': 12599, 'loss/train': 1.5629653930664062} 02/24/2022 17:52:47 - INFO - codeparrot_training - Step 12600: {'lr': 0.0004422104580183649, 'samples': 6451712, 'steps': 12600, 'loss/train': 2.2322680950164795} 02/24/2022 17:52:50 - INFO - codeparrot_training - Step 12601: {'lr': 0.0004421999948104848, 'samples': 6452224, 'steps': 12601, 'loss/train': 2.2347562313079834} 02/24/2022 17:52:56 - INFO - codeparrot_training - Step 12602: {'lr': 0.00044218953077928083, 'samples': 6452736, 'steps': 12602, 'loss/train': 2.259594440460205} 02/24/2022 17:52:59 - INFO - codeparrot_training - Step 12603: {'lr': 0.000442179065924798, 'samples': 6453248, 'steps': 12603, 'loss/train': 2.3209760189056396} 02/24/2022 17:53:06 - INFO - codeparrot_training - Step 12604: {'lr': 0.0004421686002470811, 'samples': 6453760, 'steps': 12604, 'loss/train': 2.8240911960601807} 02/24/2022 17:53:10 - INFO - codeparrot_training - Step 12605: {'lr': 0.0004421581337461749, 'samples': 6454272, 'steps': 12605, 'loss/train': 2.103368043899536} 02/24/2022 17:53:15 - INFO - codeparrot_training - Step 12606: {'lr': 0.00044214766642212435, 'samples': 6454784, 'steps': 12606, 'loss/train': 2.179652690887451} 02/24/2022 17:53:19 - INFO - codeparrot_training - Step 12607: {'lr': 0.00044213719827497413, 'samples': 6455296, 'steps': 12607, 'loss/train': 2.760122537612915} 02/24/2022 17:53:24 - INFO - codeparrot_training - Step 12608: {'lr': 0.0004421267293047692, 'samples': 6455808, 'steps': 12608, 'loss/train': 2.891826629638672} 02/24/2022 17:53:28 - INFO - codeparrot_training - Step 12609: {'lr': 0.00044211625951155433, 'samples': 6456320, 'steps': 12609, 'loss/train': 2.0851290225982666} 02/24/2022 17:53:33 - INFO - codeparrot_training - Step 12610: {'lr': 0.00044210578889537446, 'samples': 6456832, 'steps': 12610, 'loss/train': 2.0992162227630615} 02/24/2022 17:53:37 - INFO - codeparrot_training - Step 12611: {'lr': 0.0004420953174562743, 'samples': 6457344, 'steps': 12611, 'loss/train': 1.9473077058792114} 02/24/2022 17:53:42 - INFO - codeparrot_training - Step 12612: {'lr': 0.0004420848451942989, 'samples': 6457856, 'steps': 12612, 'loss/train': 2.036018133163452} 02/24/2022 17:53:46 - INFO - codeparrot_training - Step 12613: {'lr': 0.000442074372109493, 'samples': 6458368, 'steps': 12613, 'loss/train': 1.6547542810440063} 02/24/2022 17:53:52 - INFO - codeparrot_training - Step 12614: {'lr': 0.0004420638982019014, 'samples': 6458880, 'steps': 12614, 'loss/train': 3.098187208175659} 02/24/2022 17:53:56 - INFO - codeparrot_training - Step 12615: {'lr': 0.0004420534234715691, 'samples': 6459392, 'steps': 12615, 'loss/train': 2.1233577728271484} 02/24/2022 17:54:01 - INFO - codeparrot_training - Step 12616: {'lr': 0.00044204294791854094, 'samples': 6459904, 'steps': 12616, 'loss/train': 0.32390645146369934} 02/24/2022 17:54:05 - INFO - codeparrot_training - Step 12617: {'lr': 0.00044203247154286175, 'samples': 6460416, 'steps': 12617, 'loss/train': 0.4583117961883545} 02/24/2022 17:54:10 - INFO - codeparrot_training - Step 12618: {'lr': 0.0004420219943445765, 'samples': 6460928, 'steps': 12618, 'loss/train': 1.7064653635025024} 02/24/2022 17:54:14 - INFO - codeparrot_training - Step 12619: {'lr': 0.0004420115163237299, 'samples': 6461440, 'steps': 12619, 'loss/train': 3.1020309925079346} 02/24/2022 17:54:19 - INFO - codeparrot_training - Step 12620: {'lr': 0.000442001037480367, 'samples': 6461952, 'steps': 12620, 'loss/train': 1.4251264333724976} 02/24/2022 17:54:23 - INFO - codeparrot_training - Step 12621: {'lr': 0.0004419905578145326, 'samples': 6462464, 'steps': 12621, 'loss/train': 2.4654786586761475} 02/24/2022 17:54:28 - INFO - codeparrot_training - Step 12622: {'lr': 0.00044198007732627155, 'samples': 6462976, 'steps': 12622, 'loss/train': 1.6898380517959595} 02/24/2022 17:54:32 - INFO - codeparrot_training - Step 12623: {'lr': 0.00044196959601562884, 'samples': 6463488, 'steps': 12623, 'loss/train': 2.1188395023345947} 02/24/2022 17:54:37 - INFO - codeparrot_training - Step 12624: {'lr': 0.0004419591138826494, 'samples': 6464000, 'steps': 12624, 'loss/train': 2.0805227756500244} 02/24/2022 17:54:41 - INFO - codeparrot_training - Step 12625: {'lr': 0.000441948630927378, 'samples': 6464512, 'steps': 12625, 'loss/train': 1.6058732271194458} 02/24/2022 17:54:47 - INFO - codeparrot_training - Step 12626: {'lr': 0.0004419381471498597, 'samples': 6465024, 'steps': 12626, 'loss/train': 2.070448160171509} 02/24/2022 17:54:51 - INFO - codeparrot_training - Step 12627: {'lr': 0.00044192766255013926, 'samples': 6465536, 'steps': 12627, 'loss/train': 1.9055109024047852} 02/24/2022 17:54:56 - INFO - codeparrot_training - Step 12628: {'lr': 0.0004419171771282616, 'samples': 6466048, 'steps': 12628, 'loss/train': 1.4694446325302124} 02/24/2022 17:55:00 - INFO - codeparrot_training - Step 12629: {'lr': 0.0004419066908842718, 'samples': 6466560, 'steps': 12629, 'loss/train': 1.4293460845947266} 02/24/2022 17:55:05 - INFO - codeparrot_training - Step 12630: {'lr': 0.0004418962038182146, 'samples': 6467072, 'steps': 12630, 'loss/train': 1.2345119714736938} 02/24/2022 17:55:09 - INFO - codeparrot_training - Step 12631: {'lr': 0.00044188571593013504, 'samples': 6467584, 'steps': 12631, 'loss/train': 1.4546709060668945} 02/24/2022 17:55:14 - INFO - codeparrot_training - Step 12632: {'lr': 0.000441875227220078, 'samples': 6468096, 'steps': 12632, 'loss/train': 3.092456340789795} 02/24/2022 17:55:18 - INFO - codeparrot_training - Step 12633: {'lr': 0.00044186473768808844, 'samples': 6468608, 'steps': 12633, 'loss/train': 1.1801881790161133} 02/24/2022 17:55:23 - INFO - codeparrot_training - Step 12634: {'lr': 0.0004418542473342112, 'samples': 6469120, 'steps': 12634, 'loss/train': 2.223172903060913} 02/24/2022 17:55:27 - INFO - codeparrot_training - Step 12635: {'lr': 0.0004418437561584914, 'samples': 6469632, 'steps': 12635, 'loss/train': 2.7130913734436035} 02/24/2022 17:55:33 - INFO - codeparrot_training - Step 12636: {'lr': 0.00044183326416097373, 'samples': 6470144, 'steps': 12636, 'loss/train': 1.7788987159729004} 02/24/2022 17:55:36 - INFO - codeparrot_training - Step 12637: {'lr': 0.0004418227713417033, 'samples': 6470656, 'steps': 12637, 'loss/train': 2.1338772773742676} 02/24/2022 17:55:42 - INFO - codeparrot_training - Step 12638: {'lr': 0.0004418122777007251, 'samples': 6471168, 'steps': 12638, 'loss/train': 1.1886552572250366} 02/24/2022 17:55:45 - INFO - codeparrot_training - Step 12639: {'lr': 0.00044180178323808395, 'samples': 6471680, 'steps': 12639, 'loss/train': 2.2855968475341797} 02/24/2022 17:55:51 - INFO - codeparrot_training - Step 12640: {'lr': 0.00044179128795382493, 'samples': 6472192, 'steps': 12640, 'loss/train': 2.012110710144043} 02/24/2022 17:55:54 - INFO - codeparrot_training - Step 12641: {'lr': 0.00044178079184799284, 'samples': 6472704, 'steps': 12641, 'loss/train': 3.346752882003784} 02/24/2022 17:56:00 - INFO - codeparrot_training - Step 12642: {'lr': 0.0004417702949206328, 'samples': 6473216, 'steps': 12642, 'loss/train': 2.072218418121338} 02/24/2022 17:56:03 - INFO - codeparrot_training - Step 12643: {'lr': 0.0004417597971717897, 'samples': 6473728, 'steps': 12643, 'loss/train': 2.219122886657715} 02/24/2022 17:56:09 - INFO - codeparrot_training - Step 12644: {'lr': 0.0004417492986015085, 'samples': 6474240, 'steps': 12644, 'loss/train': 1.8840543031692505} 02/24/2022 17:56:12 - INFO - codeparrot_training - Step 12645: {'lr': 0.00044173879920983417, 'samples': 6474752, 'steps': 12645, 'loss/train': 2.312716484069824} 02/24/2022 17:56:18 - INFO - codeparrot_training - Step 12646: {'lr': 0.00044172829899681175, 'samples': 6475264, 'steps': 12646, 'loss/train': 1.9317117929458618} 02/24/2022 17:56:21 - INFO - codeparrot_training - Step 12647: {'lr': 0.00044171779796248623, 'samples': 6475776, 'steps': 12647, 'loss/train': 1.3755879402160645} 02/24/2022 17:56:27 - INFO - codeparrot_training - Step 12648: {'lr': 0.0004417072961069024, 'samples': 6476288, 'steps': 12648, 'loss/train': 2.609785318374634} 02/24/2022 17:56:30 - INFO - codeparrot_training - Step 12649: {'lr': 0.0004416967934301055, 'samples': 6476800, 'steps': 12649, 'loss/train': 1.818744421005249} 02/24/2022 17:56:36 - INFO - codeparrot_training - Step 12650: {'lr': 0.00044168628993214036, 'samples': 6477312, 'steps': 12650, 'loss/train': 2.233635663986206} 02/24/2022 17:56:40 - INFO - codeparrot_training - Step 12651: {'lr': 0.0004416757856130521, 'samples': 6477824, 'steps': 12651, 'loss/train': 0.1283031702041626} 02/24/2022 17:56:45 - INFO - codeparrot_training - Step 12652: {'lr': 0.0004416652804728855, 'samples': 6478336, 'steps': 12652, 'loss/train': 1.9103279113769531} 02/24/2022 17:56:49 - INFO - codeparrot_training - Step 12653: {'lr': 0.0004416547745116858, 'samples': 6478848, 'steps': 12653, 'loss/train': 0.5882845520973206} 02/24/2022 17:56:55 - INFO - codeparrot_training - Step 12654: {'lr': 0.00044164426772949785, 'samples': 6479360, 'steps': 12654, 'loss/train': 0.4565441310405731} 02/24/2022 17:56:58 - INFO - codeparrot_training - Step 12655: {'lr': 0.0004416337601263667, 'samples': 6479872, 'steps': 12655, 'loss/train': 1.2241612672805786} 02/24/2022 17:57:04 - INFO - codeparrot_training - Step 12656: {'lr': 0.00044162325170233745, 'samples': 6480384, 'steps': 12656, 'loss/train': 1.6748687028884888} 02/24/2022 17:57:07 - INFO - codeparrot_training - Step 12657: {'lr': 0.00044161274245745497, 'samples': 6480896, 'steps': 12657, 'loss/train': 0.9990090727806091} 02/24/2022 17:57:13 - INFO - codeparrot_training - Step 12658: {'lr': 0.00044160223239176445, 'samples': 6481408, 'steps': 12658, 'loss/train': 2.565922498703003} 02/24/2022 17:57:16 - INFO - codeparrot_training - Step 12659: {'lr': 0.0004415917215053107, 'samples': 6481920, 'steps': 12659, 'loss/train': 1.7146501541137695} 02/24/2022 17:57:22 - INFO - codeparrot_training - Step 12660: {'lr': 0.00044158120979813885, 'samples': 6482432, 'steps': 12660, 'loss/train': 1.6578271389007568} 02/24/2022 17:57:25 - INFO - codeparrot_training - Step 12661: {'lr': 0.000441570697270294, 'samples': 6482944, 'steps': 12661, 'loss/train': 2.0979325771331787} 02/24/2022 17:57:31 - INFO - codeparrot_training - Step 12662: {'lr': 0.00044156018392182105, 'samples': 6483456, 'steps': 12662, 'loss/train': 1.3775895833969116} 02/24/2022 17:57:35 - INFO - codeparrot_training - Step 12663: {'lr': 0.00044154966975276514, 'samples': 6483968, 'steps': 12663, 'loss/train': 1.668207049369812} 02/24/2022 17:57:40 - INFO - codeparrot_training - Step 12664: {'lr': 0.00044153915476317126, 'samples': 6484480, 'steps': 12664, 'loss/train': 2.0565552711486816} 02/24/2022 17:57:44 - INFO - codeparrot_training - Step 12665: {'lr': 0.00044152863895308446, 'samples': 6484992, 'steps': 12665, 'loss/train': 1.4628971815109253} 02/24/2022 17:57:50 - INFO - codeparrot_training - Step 12666: {'lr': 0.0004415181223225497, 'samples': 6485504, 'steps': 12666, 'loss/train': 2.9463400840759277} 02/24/2022 17:57:53 - INFO - codeparrot_training - Step 12667: {'lr': 0.0004415076048716122, 'samples': 6486016, 'steps': 12667, 'loss/train': 1.8107576370239258} 02/24/2022 17:57:59 - INFO - codeparrot_training - Step 12668: {'lr': 0.00044149708660031704, 'samples': 6486528, 'steps': 12668, 'loss/train': 0.6945056915283203} 02/24/2022 17:58:02 - INFO - codeparrot_training - Step 12669: {'lr': 0.000441486567508709, 'samples': 6487040, 'steps': 12669, 'loss/train': 2.2073943614959717} 02/24/2022 17:58:08 - INFO - codeparrot_training - Step 12670: {'lr': 0.0004414760475968334, 'samples': 6487552, 'steps': 12670, 'loss/train': 2.9699811935424805} 02/24/2022 17:58:11 - INFO - codeparrot_training - Step 12671: {'lr': 0.0004414655268647352, 'samples': 6488064, 'steps': 12671, 'loss/train': 1.8050999641418457} 02/24/2022 17:58:17 - INFO - codeparrot_training - Step 12672: {'lr': 0.0004414550053124594, 'samples': 6488576, 'steps': 12672, 'loss/train': 1.6099040508270264} 02/24/2022 17:58:21 - INFO - codeparrot_training - Step 12673: {'lr': 0.0004414444829400512, 'samples': 6489088, 'steps': 12673, 'loss/train': 2.013878107070923} 02/24/2022 17:58:26 - INFO - codeparrot_training - Step 12674: {'lr': 0.00044143395974755565, 'samples': 6489600, 'steps': 12674, 'loss/train': 2.288484811782837} 02/24/2022 17:58:30 - INFO - codeparrot_training - Step 12675: {'lr': 0.00044142343573501787, 'samples': 6490112, 'steps': 12675, 'loss/train': 1.8778387308120728} 02/24/2022 17:58:35 - INFO - codeparrot_training - Step 12676: {'lr': 0.0004414129109024827, 'samples': 6490624, 'steps': 12676, 'loss/train': 1.9495888948440552} 02/24/2022 17:58:39 - INFO - codeparrot_training - Step 12677: {'lr': 0.00044140238524999556, 'samples': 6491136, 'steps': 12677, 'loss/train': 2.473463535308838} 02/24/2022 17:58:44 - INFO - codeparrot_training - Step 12678: {'lr': 0.0004413918587776013, 'samples': 6491648, 'steps': 12678, 'loss/train': 2.351705551147461} 02/24/2022 17:58:48 - INFO - codeparrot_training - Step 12679: {'lr': 0.0004413813314853451, 'samples': 6492160, 'steps': 12679, 'loss/train': 1.420594334602356} 02/24/2022 17:58:53 - INFO - codeparrot_training - Step 12680: {'lr': 0.00044137080337327205, 'samples': 6492672, 'steps': 12680, 'loss/train': 2.236908435821533} 02/24/2022 17:58:57 - INFO - codeparrot_training - Step 12681: {'lr': 0.00044136027444142723, 'samples': 6493184, 'steps': 12681, 'loss/train': 0.9558871984481812} 02/24/2022 17:59:02 - INFO - codeparrot_training - Step 12682: {'lr': 0.0004413497446898558, 'samples': 6493696, 'steps': 12682, 'loss/train': 1.4634085893630981} 02/24/2022 17:59:06 - INFO - codeparrot_training - Step 12683: {'lr': 0.0004413392141186028, 'samples': 6494208, 'steps': 12683, 'loss/train': 2.386815309524536} 02/24/2022 17:59:11 - INFO - codeparrot_training - Step 12684: {'lr': 0.00044132868272771334, 'samples': 6494720, 'steps': 12684, 'loss/train': 2.317451000213623} 02/24/2022 17:59:15 - INFO - codeparrot_training - Step 12685: {'lr': 0.0004413181505172326, 'samples': 6495232, 'steps': 12685, 'loss/train': 2.307893753051758} 02/24/2022 17:59:20 - INFO - codeparrot_training - Step 12686: {'lr': 0.0004413076174872056, 'samples': 6495744, 'steps': 12686, 'loss/train': 2.906240701675415} 02/24/2022 17:59:24 - INFO - codeparrot_training - Step 12687: {'lr': 0.0004412970836376776, 'samples': 6496256, 'steps': 12687, 'loss/train': 1.8739978075027466} 02/24/2022 17:59:30 - INFO - codeparrot_training - Step 12688: {'lr': 0.00044128654896869357, 'samples': 6496768, 'steps': 12688, 'loss/train': 2.8445639610290527} 02/24/2022 17:59:34 - INFO - codeparrot_training - Step 12689: {'lr': 0.00044127601348029874, 'samples': 6497280, 'steps': 12689, 'loss/train': 1.6321231126785278} 02/24/2022 17:59:39 - INFO - codeparrot_training - Step 12690: {'lr': 0.0004412654771725382, 'samples': 6497792, 'steps': 12690, 'loss/train': 2.0426223278045654} 02/24/2022 17:59:43 - INFO - codeparrot_training - Step 12691: {'lr': 0.00044125494004545703, 'samples': 6498304, 'steps': 12691, 'loss/train': 2.3062593936920166} 02/24/2022 17:59:48 - INFO - codeparrot_training - Step 12692: {'lr': 0.0004412444020991004, 'samples': 6498816, 'steps': 12692, 'loss/train': 1.3057618141174316} 02/24/2022 17:59:52 - INFO - codeparrot_training - Step 12693: {'lr': 0.00044123386333351364, 'samples': 6499328, 'steps': 12693, 'loss/train': 2.1326727867126465} 02/24/2022 17:59:57 - INFO - codeparrot_training - Step 12694: {'lr': 0.00044122332374874166, 'samples': 6499840, 'steps': 12694, 'loss/train': 0.8666550517082214} 02/24/2022 18:00:01 - INFO - codeparrot_training - Step 12695: {'lr': 0.0004412127833448296, 'samples': 6500352, 'steps': 12695, 'loss/train': 1.8085170984268188} 02/24/2022 18:00:06 - INFO - codeparrot_training - Step 12696: {'lr': 0.00044120224212182283, 'samples': 6500864, 'steps': 12696, 'loss/train': 2.3644838333129883} 02/24/2022 18:00:10 - INFO - codeparrot_training - Step 12697: {'lr': 0.0004411917000797663, 'samples': 6501376, 'steps': 12697, 'loss/train': 1.6678123474121094} 02/24/2022 18:00:16 - INFO - codeparrot_training - Step 12698: {'lr': 0.0004411811572187052, 'samples': 6501888, 'steps': 12698, 'loss/train': 2.0448853969573975} 02/24/2022 18:00:20 - INFO - codeparrot_training - Step 12699: {'lr': 0.0004411706135386847, 'samples': 6502400, 'steps': 12699, 'loss/train': 1.9106106758117676} 02/24/2022 18:00:25 - INFO - codeparrot_training - Step 12700: {'lr': 0.0004411600690397501, 'samples': 6502912, 'steps': 12700, 'loss/train': 1.6755871772766113} 02/24/2022 18:00:29 - INFO - codeparrot_training - Step 12701: {'lr': 0.0004411495237219464, 'samples': 6503424, 'steps': 12701, 'loss/train': 1.4629929065704346} 02/24/2022 18:00:34 - INFO - codeparrot_training - Step 12702: {'lr': 0.00044113897758531884, 'samples': 6503936, 'steps': 12702, 'loss/train': 2.561931610107422} 02/24/2022 18:00:38 - INFO - codeparrot_training - Step 12703: {'lr': 0.00044112843062991264, 'samples': 6504448, 'steps': 12703, 'loss/train': 1.908326268196106} 02/24/2022 18:00:43 - INFO - codeparrot_training - Step 12704: {'lr': 0.0004411178828557729, 'samples': 6504960, 'steps': 12704, 'loss/train': 2.5652146339416504} 02/24/2022 18:00:47 - INFO - codeparrot_training - Step 12705: {'lr': 0.00044110733426294484, 'samples': 6505472, 'steps': 12705, 'loss/train': 3.0432803630828857} 02/24/2022 18:00:52 - INFO - codeparrot_training - Step 12706: {'lr': 0.00044109678485147367, 'samples': 6505984, 'steps': 12706, 'loss/train': 2.6836273670196533} 02/24/2022 18:00:56 - INFO - codeparrot_training - Step 12707: {'lr': 0.00044108623462140454, 'samples': 6506496, 'steps': 12707, 'loss/train': 2.0609307289123535} 02/24/2022 18:01:02 - INFO - codeparrot_training - Step 12708: {'lr': 0.0004410756835727826, 'samples': 6507008, 'steps': 12708, 'loss/train': 1.7371406555175781} 02/24/2022 18:01:06 - INFO - codeparrot_training - Step 12709: {'lr': 0.0004410651317056532, 'samples': 6507520, 'steps': 12709, 'loss/train': 4.048654079437256} 02/24/2022 18:01:11 - INFO - codeparrot_training - Step 12710: {'lr': 0.0004410545790200614, 'samples': 6508032, 'steps': 12710, 'loss/train': 2.5537118911743164} 02/24/2022 18:01:15 - INFO - codeparrot_training - Step 12711: {'lr': 0.00044104402551605246, 'samples': 6508544, 'steps': 12711, 'loss/train': 1.905525803565979} 02/24/2022 18:01:20 - INFO - codeparrot_training - Step 12712: {'lr': 0.00044103347119367155, 'samples': 6509056, 'steps': 12712, 'loss/train': 2.7290377616882324} 02/24/2022 18:01:23 - INFO - codeparrot_training - Step 12713: {'lr': 0.0004410229160529639, 'samples': 6509568, 'steps': 12713, 'loss/train': 1.5874825716018677} 02/24/2022 18:01:29 - INFO - codeparrot_training - Step 12714: {'lr': 0.0004410123600939747, 'samples': 6510080, 'steps': 12714, 'loss/train': 2.000169277191162} 02/24/2022 18:01:32 - INFO - codeparrot_training - Step 12715: {'lr': 0.00044100180331674933, 'samples': 6510592, 'steps': 12715, 'loss/train': 2.084585666656494} 02/24/2022 18:01:38 - INFO - codeparrot_training - Step 12716: {'lr': 0.00044099124572133283, 'samples': 6511104, 'steps': 12716, 'loss/train': 2.405771255493164} 02/24/2022 18:01:41 - INFO - codeparrot_training - Step 12717: {'lr': 0.0004409806873077704, 'samples': 6511616, 'steps': 12717, 'loss/train': 1.4571832418441772} 02/24/2022 18:01:48 - INFO - codeparrot_training - Step 12718: {'lr': 0.0004409701280761075, 'samples': 6512128, 'steps': 12718, 'loss/train': 1.939368486404419} 02/24/2022 18:01:51 - INFO - codeparrot_training - Step 12719: {'lr': 0.0004409595680263891, 'samples': 6512640, 'steps': 12719, 'loss/train': 1.5244413614273071} 02/24/2022 18:01:57 - INFO - codeparrot_training - Step 12720: {'lr': 0.0004409490071586606, 'samples': 6513152, 'steps': 12720, 'loss/train': 2.952091932296753} 02/24/2022 18:02:01 - INFO - codeparrot_training - Step 12721: {'lr': 0.00044093844547296715, 'samples': 6513664, 'steps': 12721, 'loss/train': 1.806796908378601} 02/24/2022 18:02:06 - INFO - codeparrot_training - Step 12722: {'lr': 0.000440927882969354, 'samples': 6514176, 'steps': 12722, 'loss/train': 2.065204620361328} 02/24/2022 18:02:10 - INFO - codeparrot_training - Step 12723: {'lr': 0.0004409173196478665, 'samples': 6514688, 'steps': 12723, 'loss/train': 1.4060415029525757} 02/24/2022 18:02:15 - INFO - codeparrot_training - Step 12724: {'lr': 0.00044090675550854973, 'samples': 6515200, 'steps': 12724, 'loss/train': 1.4661648273468018} 02/24/2022 18:02:19 - INFO - codeparrot_training - Step 12725: {'lr': 0.00044089619055144916, 'samples': 6515712, 'steps': 12725, 'loss/train': 2.989802122116089} 02/24/2022 18:02:25 - INFO - codeparrot_training - Step 12726: {'lr': 0.0004408856247766098, 'samples': 6516224, 'steps': 12726, 'loss/train': 2.071134567260742} 02/24/2022 18:02:28 - INFO - codeparrot_training - Step 12727: {'lr': 0.00044087505818407715, 'samples': 6516736, 'steps': 12727, 'loss/train': 1.6357592344284058} 02/24/2022 18:02:34 - INFO - codeparrot_training - Step 12728: {'lr': 0.00044086449077389636, 'samples': 6517248, 'steps': 12728, 'loss/train': 2.09582781791687} 02/24/2022 18:02:37 - INFO - codeparrot_training - Step 12729: {'lr': 0.0004408539225461126, 'samples': 6517760, 'steps': 12729, 'loss/train': 2.021911382675171} 02/24/2022 18:02:43 - INFO - codeparrot_training - Step 12730: {'lr': 0.0004408433535007713, 'samples': 6518272, 'steps': 12730, 'loss/train': 1.679516315460205} 02/24/2022 18:02:46 - INFO - codeparrot_training - Step 12731: {'lr': 0.0004408327836379177, 'samples': 6518784, 'steps': 12731, 'loss/train': 2.283358097076416} 02/24/2022 18:02:52 - INFO - codeparrot_training - Step 12732: {'lr': 0.0004408222129575969, 'samples': 6519296, 'steps': 12732, 'loss/train': 2.89127516746521} 02/24/2022 18:02:55 - INFO - codeparrot_training - Step 12733: {'lr': 0.0004408116414598545, 'samples': 6519808, 'steps': 12733, 'loss/train': 2.235253095626831} 02/24/2022 18:03:01 - INFO - codeparrot_training - Step 12734: {'lr': 0.0004408010691447356, 'samples': 6520320, 'steps': 12734, 'loss/train': 3.0562398433685303} 02/24/2022 18:03:05 - INFO - codeparrot_training - Step 12735: {'lr': 0.00044079049601228543, 'samples': 6520832, 'steps': 12735, 'loss/train': 1.7523605823516846} 02/24/2022 18:03:10 - INFO - codeparrot_training - Step 12736: {'lr': 0.00044077992206254934, 'samples': 6521344, 'steps': 12736, 'loss/train': 1.1580567359924316} 02/24/2022 18:03:14 - INFO - codeparrot_training - Step 12737: {'lr': 0.0004407693472955727, 'samples': 6521856, 'steps': 12737, 'loss/train': 2.4555981159210205} 02/24/2022 18:03:19 - INFO - codeparrot_training - Step 12738: {'lr': 0.00044075877171140075, 'samples': 6522368, 'steps': 12738, 'loss/train': 2.8675730228424072} 02/24/2022 18:03:23 - INFO - codeparrot_training - Step 12739: {'lr': 0.00044074819531007885, 'samples': 6522880, 'steps': 12739, 'loss/train': 0.9625195264816284} 02/24/2022 18:03:28 - INFO - codeparrot_training - Step 12740: {'lr': 0.0004407376180916522, 'samples': 6523392, 'steps': 12740, 'loss/train': 2.4256608486175537} 02/24/2022 18:03:32 - INFO - codeparrot_training - Step 12741: {'lr': 0.00044072704005616614, 'samples': 6523904, 'steps': 12741, 'loss/train': 2.437060832977295} 02/24/2022 18:03:37 - INFO - codeparrot_training - Step 12742: {'lr': 0.00044071646120366604, 'samples': 6524416, 'steps': 12742, 'loss/train': 0.27724093198776245} 02/24/2022 18:03:41 - INFO - codeparrot_training - Step 12743: {'lr': 0.00044070588153419715, 'samples': 6524928, 'steps': 12743, 'loss/train': 2.35150408744812} 02/24/2022 18:03:47 - INFO - codeparrot_training - Step 12744: {'lr': 0.00044069530104780486, 'samples': 6525440, 'steps': 12744, 'loss/train': 1.5001403093338013} 02/24/2022 18:03:51 - INFO - codeparrot_training - Step 12745: {'lr': 0.00044068471974453437, 'samples': 6525952, 'steps': 12745, 'loss/train': 2.5214011669158936} 02/24/2022 18:03:56 - INFO - codeparrot_training - Step 12746: {'lr': 0.0004406741376244312, 'samples': 6526464, 'steps': 12746, 'loss/train': 1.9468610286712646} 02/24/2022 18:04:00 - INFO - codeparrot_training - Step 12747: {'lr': 0.00044066355468754047, 'samples': 6526976, 'steps': 12747, 'loss/train': 2.306725263595581} 02/24/2022 18:04:05 - INFO - codeparrot_training - Step 12748: {'lr': 0.00044065297093390764, 'samples': 6527488, 'steps': 12748, 'loss/train': 2.6221346855163574} 02/24/2022 18:04:09 - INFO - codeparrot_training - Step 12749: {'lr': 0.0004406423863635781, 'samples': 6528000, 'steps': 12749, 'loss/train': 2.135566234588623} 02/24/2022 18:04:14 - INFO - codeparrot_training - Step 12750: {'lr': 0.00044063180097659704, 'samples': 6528512, 'steps': 12750, 'loss/train': 1.3207799196243286} 02/24/2022 18:04:18 - INFO - codeparrot_training - Step 12751: {'lr': 0.00044062121477300985, 'samples': 6529024, 'steps': 12751, 'loss/train': 1.6424869298934937} 02/24/2022 18:04:23 - INFO - codeparrot_training - Step 12752: {'lr': 0.000440610627752862, 'samples': 6529536, 'steps': 12752, 'loss/train': 2.7339706420898438} 02/24/2022 18:04:27 - INFO - codeparrot_training - Step 12753: {'lr': 0.0004406000399161987, 'samples': 6530048, 'steps': 12753, 'loss/train': 2.766383647918701} 02/24/2022 18:04:33 - INFO - codeparrot_training - Step 12754: {'lr': 0.00044058945126306535, 'samples': 6530560, 'steps': 12754, 'loss/train': 2.0135092735290527} 02/24/2022 18:04:36 - INFO - codeparrot_training - Step 12755: {'lr': 0.0004405788617935073, 'samples': 6531072, 'steps': 12755, 'loss/train': 1.3996175527572632} 02/24/2022 18:04:42 - INFO - codeparrot_training - Step 12756: {'lr': 0.0004405682715075699, 'samples': 6531584, 'steps': 12756, 'loss/train': 2.252906084060669} 02/24/2022 18:04:45 - INFO - codeparrot_training - Step 12757: {'lr': 0.0004405576804052985, 'samples': 6532096, 'steps': 12757, 'loss/train': 1.8308268785476685} 02/24/2022 18:04:51 - INFO - codeparrot_training - Step 12758: {'lr': 0.0004405470884867386, 'samples': 6532608, 'steps': 12758, 'loss/train': 1.2382327318191528} 02/24/2022 18:04:55 - INFO - codeparrot_training - Step 12759: {'lr': 0.00044053649575193543, 'samples': 6533120, 'steps': 12759, 'loss/train': 3.1187937259674072} 02/24/2022 18:05:00 - INFO - codeparrot_training - Step 12760: {'lr': 0.00044052590220093445, 'samples': 6533632, 'steps': 12760, 'loss/train': 0.9883297681808472} 02/24/2022 18:05:04 - INFO - codeparrot_training - Step 12761: {'lr': 0.00044051530783378103, 'samples': 6534144, 'steps': 12761, 'loss/train': 1.8920166492462158} 02/24/2022 18:05:09 - INFO - codeparrot_training - Step 12762: {'lr': 0.0004405047126505204, 'samples': 6534656, 'steps': 12762, 'loss/train': 1.4813631772994995} 02/24/2022 18:05:13 - INFO - codeparrot_training - Step 12763: {'lr': 0.0004404941166511982, 'samples': 6535168, 'steps': 12763, 'loss/train': 2.164565086364746} 02/24/2022 18:05:19 - INFO - codeparrot_training - Step 12764: {'lr': 0.00044048351983585966, 'samples': 6535680, 'steps': 12764, 'loss/train': 0.6382386088371277} 02/24/2022 18:05:25 - INFO - codeparrot_training - Step 12765: {'lr': 0.00044047292220455016, 'samples': 6536192, 'steps': 12765, 'loss/train': 2.824276924133301} 02/24/2022 18:05:28 - INFO - codeparrot_training - Step 12766: {'lr': 0.0004404623237573152, 'samples': 6536704, 'steps': 12766, 'loss/train': 2.3192460536956787} 02/24/2022 18:05:31 - INFO - codeparrot_training - Step 12767: {'lr': 0.00044045172449420005, 'samples': 6537216, 'steps': 12767, 'loss/train': 1.1514025926589966} 02/24/2022 18:05:37 - INFO - codeparrot_training - Step 12768: {'lr': 0.00044044112441525026, 'samples': 6537728, 'steps': 12768, 'loss/train': 1.3377079963684082} 02/24/2022 18:05:42 - INFO - codeparrot_training - Step 12769: {'lr': 0.0004404305235205112, 'samples': 6538240, 'steps': 12769, 'loss/train': 2.5847256183624268} 02/24/2022 18:05:46 - INFO - codeparrot_training - Step 12770: {'lr': 0.0004404199218100281, 'samples': 6538752, 'steps': 12770, 'loss/train': 2.267228841781616} 02/24/2022 18:05:51 - INFO - codeparrot_training - Step 12771: {'lr': 0.00044040931928384665, 'samples': 6539264, 'steps': 12771, 'loss/train': 1.6417549848556519} 02/24/2022 18:05:55 - INFO - codeparrot_training - Step 12772: {'lr': 0.0004403987159420121, 'samples': 6539776, 'steps': 12772, 'loss/train': 1.6507021188735962} 02/24/2022 18:06:01 - INFO - codeparrot_training - Step 12773: {'lr': 0.0004403881117845699, 'samples': 6540288, 'steps': 12773, 'loss/train': 1.3791691064834595} 02/24/2022 18:06:04 - INFO - codeparrot_training - Step 12774: {'lr': 0.00044037750681156547, 'samples': 6540800, 'steps': 12774, 'loss/train': 2.366238832473755} 02/24/2022 18:06:10 - INFO - codeparrot_training - Step 12775: {'lr': 0.0004403669010230443, 'samples': 6541312, 'steps': 12775, 'loss/train': 1.6738210916519165} 02/24/2022 18:06:13 - INFO - codeparrot_training - Step 12776: {'lr': 0.00044035629441905173, 'samples': 6541824, 'steps': 12776, 'loss/train': 2.1039857864379883} 02/24/2022 18:06:19 - INFO - codeparrot_training - Step 12777: {'lr': 0.0004403456869996333, 'samples': 6542336, 'steps': 12777, 'loss/train': 1.8654409646987915} 02/24/2022 18:06:22 - INFO - codeparrot_training - Step 12778: {'lr': 0.0004403350787648343, 'samples': 6542848, 'steps': 12778, 'loss/train': 1.9655073881149292} 02/24/2022 18:06:29 - INFO - codeparrot_training - Step 12779: {'lr': 0.0004403244697147003, 'samples': 6543360, 'steps': 12779, 'loss/train': 2.61484956741333} 02/24/2022 18:06:32 - INFO - codeparrot_training - Step 12780: {'lr': 0.00044031385984927675, 'samples': 6543872, 'steps': 12780, 'loss/train': 5.149417877197266} 02/24/2022 18:06:38 - INFO - codeparrot_training - Step 12781: {'lr': 0.000440303249168609, 'samples': 6544384, 'steps': 12781, 'loss/train': 1.9456908702850342} 02/24/2022 18:06:41 - INFO - codeparrot_training - Step 12782: {'lr': 0.0004402926376727425, 'samples': 6544896, 'steps': 12782, 'loss/train': 1.6736698150634766} 02/24/2022 18:06:45 - INFO - codeparrot_training - Step 12783: {'lr': 0.0004402820253617229, 'samples': 6545408, 'steps': 12783, 'loss/train': 1.9663290977478027} 02/24/2022 18:06:51 - INFO - codeparrot_training - Step 12784: {'lr': 0.0004402714122355955, 'samples': 6545920, 'steps': 12784, 'loss/train': 1.8992382287979126} 02/24/2022 18:06:54 - INFO - codeparrot_training - Step 12785: {'lr': 0.00044026079829440567, 'samples': 6546432, 'steps': 12785, 'loss/train': 2.6155500411987305} 02/24/2022 18:07:00 - INFO - codeparrot_training - Step 12786: {'lr': 0.0004402501835381991, 'samples': 6546944, 'steps': 12786, 'loss/train': 3.9356064796447754} 02/24/2022 18:07:03 - INFO - codeparrot_training - Step 12787: {'lr': 0.00044023956796702116, 'samples': 6547456, 'steps': 12787, 'loss/train': 1.4840117692947388} 02/24/2022 18:07:09 - INFO - codeparrot_training - Step 12788: {'lr': 0.0004402289515809172, 'samples': 6547968, 'steps': 12788, 'loss/train': 1.8973579406738281} 02/24/2022 18:07:12 - INFO - codeparrot_training - Step 12789: {'lr': 0.00044021833437993296, 'samples': 6548480, 'steps': 12789, 'loss/train': 1.8810702562332153} 02/24/2022 18:07:19 - INFO - codeparrot_training - Step 12790: {'lr': 0.0004402077163641137, 'samples': 6548992, 'steps': 12790, 'loss/train': 2.2041003704071045} 02/24/2022 18:07:22 - INFO - codeparrot_training - Step 12791: {'lr': 0.000440197097533505, 'samples': 6549504, 'steps': 12791, 'loss/train': 0.6744171977043152} 02/24/2022 18:07:27 - INFO - codeparrot_training - Step 12792: {'lr': 0.00044018647788815235, 'samples': 6550016, 'steps': 12792, 'loss/train': 1.9453179836273193} 02/24/2022 18:07:33 - INFO - codeparrot_training - Step 12793: {'lr': 0.00044017585742810124, 'samples': 6550528, 'steps': 12793, 'loss/train': 2.635545015335083} 02/24/2022 18:07:36 - INFO - codeparrot_training - Step 12794: {'lr': 0.0004401652361533971, 'samples': 6551040, 'steps': 12794, 'loss/train': 1.8903262615203857} 02/24/2022 18:07:42 - INFO - codeparrot_training - Step 12795: {'lr': 0.00044015461406408544, 'samples': 6551552, 'steps': 12795, 'loss/train': 2.7911055088043213} 02/24/2022 18:07:45 - INFO - codeparrot_training - Step 12796: {'lr': 0.00044014399116021184, 'samples': 6552064, 'steps': 12796, 'loss/train': 1.8545019626617432} 02/24/2022 18:07:51 - INFO - codeparrot_training - Step 12797: {'lr': 0.00044013336744182176, 'samples': 6552576, 'steps': 12797, 'loss/train': 2.298102855682373} 02/24/2022 18:07:55 - INFO - codeparrot_training - Step 12798: {'lr': 0.0004401227429089607, 'samples': 6553088, 'steps': 12798, 'loss/train': 1.5750998258590698} 02/24/2022 18:08:01 - INFO - codeparrot_training - Step 12799: {'lr': 0.00044011211756167425, 'samples': 6553600, 'steps': 12799, 'loss/train': 2.0372374057769775} 02/24/2022 18:08:04 - INFO - codeparrot_training - Step 12800: {'lr': 0.0004401014914000078, 'samples': 6554112, 'steps': 12800, 'loss/train': 2.5791444778442383} 02/24/2022 18:08:10 - INFO - codeparrot_training - Step 12801: {'lr': 0.00044009086442400684, 'samples': 6554624, 'steps': 12801, 'loss/train': 2.1451714038848877} 02/24/2022 18:08:13 - INFO - codeparrot_training - Step 12802: {'lr': 0.0004400802366337171, 'samples': 6555136, 'steps': 12802, 'loss/train': 1.864613652229309} 02/24/2022 18:08:19 - INFO - codeparrot_training - Step 12803: {'lr': 0.00044006960802918393, 'samples': 6555648, 'steps': 12803, 'loss/train': 1.5845685005187988} 02/24/2022 18:08:22 - INFO - codeparrot_training - Step 12804: {'lr': 0.0004400589786104529, 'samples': 6556160, 'steps': 12804, 'loss/train': 2.4343771934509277} 02/24/2022 18:08:28 - INFO - codeparrot_training - Step 12805: {'lr': 0.0004400483483775696, 'samples': 6556672, 'steps': 12805, 'loss/train': 2.7292592525482178} 02/24/2022 18:08:31 - INFO - codeparrot_training - Step 12806: {'lr': 0.00044003771733057943, 'samples': 6557184, 'steps': 12806, 'loss/train': 2.767972230911255} 02/24/2022 18:08:37 - INFO - codeparrot_training - Step 12807: {'lr': 0.0004400270854695281, 'samples': 6557696, 'steps': 12807, 'loss/train': 1.6104705333709717} 02/24/2022 18:08:40 - INFO - codeparrot_training - Step 12808: {'lr': 0.0004400164527944611, 'samples': 6558208, 'steps': 12808, 'loss/train': 2.1192734241485596} 02/24/2022 18:08:44 - INFO - codeparrot_training - Step 12809: {'lr': 0.0004400058193054239, 'samples': 6558720, 'steps': 12809, 'loss/train': 2.5320072174072266} 02/24/2022 18:08:50 - INFO - codeparrot_training - Step 12810: {'lr': 0.0004399951850024621, 'samples': 6559232, 'steps': 12810, 'loss/train': 2.262166738510132} 02/24/2022 18:08:54 - INFO - codeparrot_training - Step 12811: {'lr': 0.0004399845498856213, 'samples': 6559744, 'steps': 12811, 'loss/train': 0.6867892742156982} 02/24/2022 18:08:59 - INFO - codeparrot_training - Step 12812: {'lr': 0.000439973913954947, 'samples': 6560256, 'steps': 12812, 'loss/train': 2.066507577896118} 02/24/2022 18:09:03 - INFO - codeparrot_training - Step 12813: {'lr': 0.0004399632772104848, 'samples': 6560768, 'steps': 12813, 'loss/train': 1.0127909183502197} 02/24/2022 18:09:08 - INFO - codeparrot_training - Step 12814: {'lr': 0.00043995263965228016, 'samples': 6561280, 'steps': 12814, 'loss/train': 1.0926337242126465} 02/24/2022 18:09:12 - INFO - codeparrot_training - Step 12815: {'lr': 0.00043994200128037877, 'samples': 6561792, 'steps': 12815, 'loss/train': 2.4272892475128174} 02/24/2022 18:09:17 - INFO - codeparrot_training - Step 12816: {'lr': 0.0004399313620948262, 'samples': 6562304, 'steps': 12816, 'loss/train': 2.5267903804779053} 02/24/2022 18:09:23 - INFO - codeparrot_training - Step 12817: {'lr': 0.00043992072209566793, 'samples': 6562816, 'steps': 12817, 'loss/train': 2.595418691635132} 02/24/2022 18:09:26 - INFO - codeparrot_training - Step 12818: {'lr': 0.0004399100812829496, 'samples': 6563328, 'steps': 12818, 'loss/train': 2.586636781692505} 02/24/2022 18:09:30 - INFO - codeparrot_training - Step 12819: {'lr': 0.00043989943965671685, 'samples': 6563840, 'steps': 12819, 'loss/train': 2.1191513538360596} 02/24/2022 18:09:35 - INFO - codeparrot_training - Step 12820: {'lr': 0.00043988879721701515, 'samples': 6564352, 'steps': 12820, 'loss/train': 1.8866201639175415} 02/24/2022 18:09:41 - INFO - codeparrot_training - Step 12821: {'lr': 0.0004398781539638901, 'samples': 6564864, 'steps': 12821, 'loss/train': 3.1851587295532227} 02/24/2022 18:09:44 - INFO - codeparrot_training - Step 12822: {'lr': 0.00043986750989738737, 'samples': 6565376, 'steps': 12822, 'loss/train': 2.0842721462249756} 02/24/2022 18:09:50 - INFO - codeparrot_training - Step 12823: {'lr': 0.0004398568650175525, 'samples': 6565888, 'steps': 12823, 'loss/train': 2.483586549758911} 02/24/2022 18:09:53 - INFO - codeparrot_training - Step 12824: {'lr': 0.00043984621932443115, 'samples': 6566400, 'steps': 12824, 'loss/train': 2.3778891563415527} 02/24/2022 18:09:59 - INFO - codeparrot_training - Step 12825: {'lr': 0.0004398355728180689, 'samples': 6566912, 'steps': 12825, 'loss/train': 2.1336655616760254} 02/24/2022 18:10:03 - INFO - codeparrot_training - Step 12826: {'lr': 0.0004398249254985113, 'samples': 6567424, 'steps': 12826, 'loss/train': 2.381819725036621} 02/24/2022 18:10:09 - INFO - codeparrot_training - Step 12827: {'lr': 0.00043981427736580395, 'samples': 6567936, 'steps': 12827, 'loss/train': 2.261563539505005} 02/24/2022 18:10:12 - INFO - codeparrot_training - Step 12828: {'lr': 0.00043980362841999253, 'samples': 6568448, 'steps': 12828, 'loss/train': 2.7353408336639404} 02/24/2022 18:10:16 - INFO - codeparrot_training - Step 12829: {'lr': 0.0004397929786611227, 'samples': 6568960, 'steps': 12829, 'loss/train': 1.0050386190414429} 02/24/2022 18:10:23 - INFO - codeparrot_training - Step 12830: {'lr': 0.00043978232808923996, 'samples': 6569472, 'steps': 12830, 'loss/train': 0.19775129854679108} 02/24/2022 18:10:26 - INFO - codeparrot_training - Step 12831: {'lr': 0.00043977167670439, 'samples': 6569984, 'steps': 12831, 'loss/train': 1.5964388847351074} 02/24/2022 18:10:32 - INFO - codeparrot_training - Step 12832: {'lr': 0.0004397610245066184, 'samples': 6570496, 'steps': 12832, 'loss/train': 1.4163870811462402} 02/24/2022 18:10:35 - INFO - codeparrot_training - Step 12833: {'lr': 0.00043975037149597085, 'samples': 6571008, 'steps': 12833, 'loss/train': 1.9741743803024292} 02/24/2022 18:10:41 - INFO - codeparrot_training - Step 12834: {'lr': 0.00043973971767249297, 'samples': 6571520, 'steps': 12834, 'loss/train': 2.2019591331481934} 02/24/2022 18:10:44 - INFO - codeparrot_training - Step 12835: {'lr': 0.0004397290630362304, 'samples': 6572032, 'steps': 12835, 'loss/train': 2.1687161922454834} 02/24/2022 18:10:50 - INFO - codeparrot_training - Step 12836: {'lr': 0.0004397184075872288, 'samples': 6572544, 'steps': 12836, 'loss/train': 2.214421510696411} 02/24/2022 18:10:53 - INFO - codeparrot_training - Step 12837: {'lr': 0.00043970775132553375, 'samples': 6573056, 'steps': 12837, 'loss/train': 0.9080241322517395} 02/24/2022 18:10:59 - INFO - codeparrot_training - Step 12838: {'lr': 0.00043969709425119085, 'samples': 6573568, 'steps': 12838, 'loss/train': 1.4517842531204224} 02/24/2022 18:11:02 - INFO - codeparrot_training - Step 12839: {'lr': 0.000439686436364246, 'samples': 6574080, 'steps': 12839, 'loss/train': 1.7576817274093628} 02/24/2022 18:11:09 - INFO - codeparrot_training - Step 12840: {'lr': 0.00043967577766474455, 'samples': 6574592, 'steps': 12840, 'loss/train': 1.9305963516235352} 02/24/2022 18:11:13 - INFO - codeparrot_training - Step 12841: {'lr': 0.00043966511815273233, 'samples': 6575104, 'steps': 12841, 'loss/train': 1.845685601234436} 02/24/2022 18:11:19 - INFO - codeparrot_training - Step 12842: {'lr': 0.00043965445782825495, 'samples': 6575616, 'steps': 12842, 'loss/train': 2.229151964187622} 02/24/2022 18:11:22 - INFO - codeparrot_training - Step 12843: {'lr': 0.00043964379669135815, 'samples': 6576128, 'steps': 12843, 'loss/train': 1.5379745960235596} 02/24/2022 18:11:28 - INFO - codeparrot_training - Step 12844: {'lr': 0.00043963313474208753, 'samples': 6576640, 'steps': 12844, 'loss/train': 0.8637090921401978} 02/24/2022 18:11:31 - INFO - codeparrot_training - Step 12845: {'lr': 0.0004396224719804888, 'samples': 6577152, 'steps': 12845, 'loss/train': 2.0980193614959717} 02/24/2022 18:11:37 - INFO - codeparrot_training - Step 12846: {'lr': 0.0004396118084066075, 'samples': 6577664, 'steps': 12846, 'loss/train': 1.9521963596343994} 02/24/2022 18:11:40 - INFO - codeparrot_training - Step 12847: {'lr': 0.00043960114402048957, 'samples': 6578176, 'steps': 12847, 'loss/train': 2.294322967529297} 02/24/2022 18:11:46 - INFO - codeparrot_training - Step 12848: {'lr': 0.0004395904788221805, 'samples': 6578688, 'steps': 12848, 'loss/train': 2.5258712768554688} 02/24/2022 18:11:49 - INFO - codeparrot_training - Step 12849: {'lr': 0.00043957981281172597, 'samples': 6579200, 'steps': 12849, 'loss/train': 1.0141184329986572} 02/24/2022 18:11:56 - INFO - codeparrot_training - Step 12850: {'lr': 0.00043956914598917177, 'samples': 6579712, 'steps': 12850, 'loss/train': 1.8903825283050537} 02/24/2022 18:11:59 - INFO - codeparrot_training - Step 12851: {'lr': 0.00043955847835456353, 'samples': 6580224, 'steps': 12851, 'loss/train': 2.1995418071746826} 02/24/2022 18:12:05 - INFO - codeparrot_training - Step 12852: {'lr': 0.00043954780990794695, 'samples': 6580736, 'steps': 12852, 'loss/train': 0.5545242428779602} 02/24/2022 18:12:08 - INFO - codeparrot_training - Step 12853: {'lr': 0.0004395371406493677, 'samples': 6581248, 'steps': 12853, 'loss/train': 1.7464067935943604} 02/24/2022 18:12:14 - INFO - codeparrot_training - Step 12854: {'lr': 0.0004395264705788716, 'samples': 6581760, 'steps': 12854, 'loss/train': 2.0774736404418945} 02/24/2022 18:12:17 - INFO - codeparrot_training - Step 12855: {'lr': 0.00043951579969650424, 'samples': 6582272, 'steps': 12855, 'loss/train': 3.024357795715332} 02/24/2022 18:12:23 - INFO - codeparrot_training - Step 12856: {'lr': 0.00043950512800231136, 'samples': 6582784, 'steps': 12856, 'loss/train': 1.100299596786499} 02/24/2022 18:12:26 - INFO - codeparrot_training - Step 12857: {'lr': 0.0004394944554963387, 'samples': 6583296, 'steps': 12857, 'loss/train': 1.196905255317688} 02/24/2022 18:12:32 - INFO - codeparrot_training - Step 12858: {'lr': 0.000439483782178632, 'samples': 6583808, 'steps': 12858, 'loss/train': 2.6736104488372803} 02/24/2022 18:12:35 - INFO - codeparrot_training - Step 12859: {'lr': 0.0004394731080492369, 'samples': 6584320, 'steps': 12859, 'loss/train': 1.8345617055892944} 02/24/2022 18:12:41 - INFO - codeparrot_training - Step 12860: {'lr': 0.0004394624331081992, 'samples': 6584832, 'steps': 12860, 'loss/train': 2.343151569366455} 02/24/2022 18:12:45 - INFO - codeparrot_training - Step 12861: {'lr': 0.00043945175735556454, 'samples': 6585344, 'steps': 12861, 'loss/train': 0.6751885414123535} 02/24/2022 18:12:50 - INFO - codeparrot_training - Step 12862: {'lr': 0.0004394410807913788, 'samples': 6585856, 'steps': 12862, 'loss/train': 1.2433993816375732} 02/24/2022 18:12:54 - INFO - codeparrot_training - Step 12863: {'lr': 0.0004394304034156875, 'samples': 6586368, 'steps': 12863, 'loss/train': 0.16667544841766357} 02/24/2022 18:13:00 - INFO - codeparrot_training - Step 12864: {'lr': 0.00043941972522853665, 'samples': 6586880, 'steps': 12864, 'loss/train': 2.7264668941497803} 02/24/2022 18:13:03 - INFO - codeparrot_training - Step 12865: {'lr': 0.00043940904622997176, 'samples': 6587392, 'steps': 12865, 'loss/train': 1.565804123878479} 02/24/2022 18:13:08 - INFO - codeparrot_training - Step 12866: {'lr': 0.00043939836642003865, 'samples': 6587904, 'steps': 12866, 'loss/train': 3.104874610900879} 02/24/2022 18:13:12 - INFO - codeparrot_training - Step 12867: {'lr': 0.0004393876857987831, 'samples': 6588416, 'steps': 12867, 'loss/train': 1.7425435781478882} 02/24/2022 18:13:18 - INFO - codeparrot_training - Step 12868: {'lr': 0.0004393770043662508, 'samples': 6588928, 'steps': 12868, 'loss/train': 0.7137417197227478} 02/24/2022 18:13:21 - INFO - codeparrot_training - Step 12869: {'lr': 0.0004393663221224876, 'samples': 6589440, 'steps': 12869, 'loss/train': 2.7801997661590576} 02/24/2022 18:13:27 - INFO - codeparrot_training - Step 12870: {'lr': 0.00043935563906753923, 'samples': 6589952, 'steps': 12870, 'loss/train': 1.2647124528884888} 02/24/2022 18:13:31 - INFO - codeparrot_training - Step 12871: {'lr': 0.0004393449552014514, 'samples': 6590464, 'steps': 12871, 'loss/train': 2.162167549133301} 02/24/2022 18:13:36 - INFO - codeparrot_training - Step 12872: {'lr': 0.00043933427052426986, 'samples': 6590976, 'steps': 12872, 'loss/train': 2.060293436050415} 02/24/2022 18:13:40 - INFO - codeparrot_training - Step 12873: {'lr': 0.00043932358503604054, 'samples': 6591488, 'steps': 12873, 'loss/train': 1.4157055616378784} 02/24/2022 18:13:45 - INFO - codeparrot_training - Step 12874: {'lr': 0.000439312898736809, 'samples': 6592000, 'steps': 12874, 'loss/train': 2.2592687606811523} 02/24/2022 18:13:49 - INFO - codeparrot_training - Step 12875: {'lr': 0.00043930221162662115, 'samples': 6592512, 'steps': 12875, 'loss/train': 1.7817293405532837} 02/24/2022 18:13:54 - INFO - codeparrot_training - Step 12876: {'lr': 0.0004392915237055227, 'samples': 6593024, 'steps': 12876, 'loss/train': 2.011043071746826} 02/24/2022 18:13:58 - INFO - codeparrot_training - Step 12877: {'lr': 0.00043928083497355954, 'samples': 6593536, 'steps': 12877, 'loss/train': 1.628848910331726} 02/24/2022 18:14:03 - INFO - codeparrot_training - Step 12878: {'lr': 0.0004392701454307773, 'samples': 6594048, 'steps': 12878, 'loss/train': 2.68397855758667} 02/24/2022 18:14:06 - INFO - codeparrot_training - Step 12879: {'lr': 0.00043925945507722195, 'samples': 6594560, 'steps': 12879, 'loss/train': 1.3266445398330688} 02/24/2022 18:14:12 - INFO - codeparrot_training - Step 12880: {'lr': 0.0004392487639129391, 'samples': 6595072, 'steps': 12880, 'loss/train': 1.838348388671875} 02/24/2022 18:14:16 - INFO - codeparrot_training - Step 12881: {'lr': 0.0004392380719379747, 'samples': 6595584, 'steps': 12881, 'loss/train': 1.6737326383590698} 02/24/2022 18:14:21 - INFO - codeparrot_training - Step 12882: {'lr': 0.0004392273791523744, 'samples': 6596096, 'steps': 12882, 'loss/train': 2.3976011276245117} 02/24/2022 18:14:25 - INFO - codeparrot_training - Step 12883: {'lr': 0.0004392166855561842, 'samples': 6596608, 'steps': 12883, 'loss/train': 1.4859952926635742} 02/24/2022 18:14:30 - INFO - codeparrot_training - Step 12884: {'lr': 0.0004392059911494498, 'samples': 6597120, 'steps': 12884, 'loss/train': 1.7109577655792236} 02/24/2022 18:14:34 - INFO - codeparrot_training - Step 12885: {'lr': 0.00043919529593221696, 'samples': 6597632, 'steps': 12885, 'loss/train': 1.3060704469680786} 02/24/2022 18:14:41 - INFO - codeparrot_training - Step 12886: {'lr': 0.00043918459990453156, 'samples': 6598144, 'steps': 12886, 'loss/train': 2.6402320861816406} 02/24/2022 18:14:44 - INFO - codeparrot_training - Step 12887: {'lr': 0.00043917390306643945, 'samples': 6598656, 'steps': 12887, 'loss/train': 3.1464781761169434} 02/24/2022 18:14:50 - INFO - codeparrot_training - Step 12888: {'lr': 0.0004391632054179864, 'samples': 6599168, 'steps': 12888, 'loss/train': 1.0135523080825806} 02/24/2022 18:14:53 - INFO - codeparrot_training - Step 12889: {'lr': 0.00043915250695921815, 'samples': 6599680, 'steps': 12889, 'loss/train': 1.9060496091842651} 02/24/2022 18:14:59 - INFO - codeparrot_training - Step 12890: {'lr': 0.00043914180769018073, 'samples': 6600192, 'steps': 12890, 'loss/train': 0.6762253046035767} 02/24/2022 18:15:02 - INFO - codeparrot_training - Step 12891: {'lr': 0.0004391311076109198, 'samples': 6600704, 'steps': 12891, 'loss/train': 1.163953423500061} 02/24/2022 18:15:08 - INFO - codeparrot_training - Step 12892: {'lr': 0.00043912040672148135, 'samples': 6601216, 'steps': 12892, 'loss/train': 1.510191798210144} 02/24/2022 18:15:11 - INFO - codeparrot_training - Step 12893: {'lr': 0.00043910970502191105, 'samples': 6601728, 'steps': 12893, 'loss/train': 2.106886148452759} 02/24/2022 18:15:17 - INFO - codeparrot_training - Step 12894: {'lr': 0.00043909900251225476, 'samples': 6602240, 'steps': 12894, 'loss/train': 1.1448532342910767} 02/24/2022 18:15:20 - INFO - codeparrot_training - Step 12895: {'lr': 0.00043908829919255855, 'samples': 6602752, 'steps': 12895, 'loss/train': 1.8956762552261353} 02/24/2022 18:15:27 - INFO - codeparrot_training - Step 12896: {'lr': 0.00043907759506286797, 'samples': 6603264, 'steps': 12896, 'loss/train': 1.6714473962783813} 02/24/2022 18:15:31 - INFO - codeparrot_training - Step 12897: {'lr': 0.0004390668901232291, 'samples': 6603776, 'steps': 12897, 'loss/train': 1.9626426696777344} 02/24/2022 18:15:36 - INFO - codeparrot_training - Step 12898: {'lr': 0.00043905618437368766, 'samples': 6604288, 'steps': 12898, 'loss/train': 2.3590214252471924} 02/24/2022 18:15:40 - INFO - codeparrot_training - Step 12899: {'lr': 0.0004390454778142896, 'samples': 6604800, 'steps': 12899, 'loss/train': 2.0362513065338135} 02/24/2022 18:15:45 - INFO - codeparrot_training - Step 12900: {'lr': 0.00043903477044508066, 'samples': 6605312, 'steps': 12900, 'loss/train': 1.9012597799301147} 02/24/2022 18:15:49 - INFO - codeparrot_training - Step 12901: {'lr': 0.0004390240622661069, 'samples': 6605824, 'steps': 12901, 'loss/train': 0.5600261688232422} 02/24/2022 18:15:54 - INFO - codeparrot_training - Step 12902: {'lr': 0.000439013353277414, 'samples': 6606336, 'steps': 12902, 'loss/train': 1.669532299041748} 02/24/2022 18:15:58 - INFO - codeparrot_training - Step 12903: {'lr': 0.00043900264347904796, 'samples': 6606848, 'steps': 12903, 'loss/train': 2.456749439239502} 02/24/2022 18:16:03 - INFO - codeparrot_training - Step 12904: {'lr': 0.00043899193287105456, 'samples': 6607360, 'steps': 12904, 'loss/train': 1.5611181259155273} 02/24/2022 18:16:07 - INFO - codeparrot_training - Step 12905: {'lr': 0.0004389812214534798, 'samples': 6607872, 'steps': 12905, 'loss/train': 1.9544636011123657} 02/24/2022 18:16:12 - INFO - codeparrot_training - Step 12906: {'lr': 0.00043897050922636947, 'samples': 6608384, 'steps': 12906, 'loss/train': 2.197460412979126} 02/24/2022 18:16:16 - INFO - codeparrot_training - Step 12907: {'lr': 0.00043895979618976944, 'samples': 6608896, 'steps': 12907, 'loss/train': 2.3713438510894775} 02/24/2022 18:16:22 - INFO - codeparrot_training - Step 12908: {'lr': 0.00043894908234372564, 'samples': 6609408, 'steps': 12908, 'loss/train': 2.0503082275390625} 02/24/2022 18:16:26 - INFO - codeparrot_training - Step 12909: {'lr': 0.00043893836768828405, 'samples': 6609920, 'steps': 12909, 'loss/train': 2.2581403255462646} 02/24/2022 18:16:31 - INFO - codeparrot_training - Step 12910: {'lr': 0.0004389276522234904, 'samples': 6610432, 'steps': 12910, 'loss/train': 2.018444538116455} 02/24/2022 18:16:35 - INFO - codeparrot_training - Step 12911: {'lr': 0.00043891693594939077, 'samples': 6610944, 'steps': 12911, 'loss/train': 1.256234884262085} 02/24/2022 18:16:40 - INFO - codeparrot_training - Step 12912: {'lr': 0.0004389062188660309, 'samples': 6611456, 'steps': 12912, 'loss/train': 2.297994613647461} 02/24/2022 18:16:44 - INFO - codeparrot_training - Step 12913: {'lr': 0.00043889550097345675, 'samples': 6611968, 'steps': 12913, 'loss/train': 3.1912038326263428} 02/24/2022 18:16:49 - INFO - codeparrot_training - Step 12914: {'lr': 0.0004388847822717144, 'samples': 6612480, 'steps': 12914, 'loss/train': 0.8842859268188477} 02/24/2022 18:16:53 - INFO - codeparrot_training - Step 12915: {'lr': 0.0004388740627608495, 'samples': 6612992, 'steps': 12915, 'loss/train': 2.4785799980163574} 02/24/2022 18:16:58 - INFO - codeparrot_training - Step 12916: {'lr': 0.0004388633424409081, 'samples': 6613504, 'steps': 12916, 'loss/train': 0.728019654750824} 02/24/2022 18:17:02 - INFO - codeparrot_training - Step 12917: {'lr': 0.0004388526213119361, 'samples': 6614016, 'steps': 12917, 'loss/train': 1.5862822532653809} 02/24/2022 18:17:07 - INFO - codeparrot_training - Step 12918: {'lr': 0.00043884189937397946, 'samples': 6614528, 'steps': 12918, 'loss/train': 1.834671974182129} 02/24/2022 18:17:11 - INFO - codeparrot_training - Step 12919: {'lr': 0.00043883117662708404, 'samples': 6615040, 'steps': 12919, 'loss/train': 1.723826289176941} 02/24/2022 18:17:16 - INFO - codeparrot_training - Step 12920: {'lr': 0.0004388204530712959, 'samples': 6615552, 'steps': 12920, 'loss/train': 1.9067957401275635} 02/24/2022 18:17:20 - INFO - codeparrot_training - Step 12921: {'lr': 0.00043880972870666084, 'samples': 6616064, 'steps': 12921, 'loss/train': 1.8042864799499512} 02/24/2022 18:17:26 - INFO - codeparrot_training - Step 12922: {'lr': 0.0004387990035332249, 'samples': 6616576, 'steps': 12922, 'loss/train': 3.0477044582366943} 02/24/2022 18:17:30 - INFO - codeparrot_training - Step 12923: {'lr': 0.00043878827755103404, 'samples': 6617088, 'steps': 12923, 'loss/train': 2.090799331665039} 02/24/2022 18:17:35 - INFO - codeparrot_training - Step 12924: {'lr': 0.00043877755076013406, 'samples': 6617600, 'steps': 12924, 'loss/train': 1.9760048389434814} 02/24/2022 18:17:39 - INFO - codeparrot_training - Step 12925: {'lr': 0.00043876682316057095, 'samples': 6618112, 'steps': 12925, 'loss/train': 2.9608991146087646} 02/24/2022 18:17:44 - INFO - codeparrot_training - Step 12926: {'lr': 0.0004387560947523908, 'samples': 6618624, 'steps': 12926, 'loss/train': 2.5437731742858887} 02/24/2022 18:17:48 - INFO - codeparrot_training - Step 12927: {'lr': 0.0004387453655356394, 'samples': 6619136, 'steps': 12927, 'loss/train': 2.234454870223999} 02/24/2022 18:17:53 - INFO - codeparrot_training - Step 12928: {'lr': 0.00043873463551036284, 'samples': 6619648, 'steps': 12928, 'loss/train': 0.7627431750297546} 02/24/2022 18:17:57 - INFO - codeparrot_training - Step 12929: {'lr': 0.000438723904676607, 'samples': 6620160, 'steps': 12929, 'loss/train': 1.6176753044128418} 02/24/2022 18:18:02 - INFO - codeparrot_training - Step 12930: {'lr': 0.0004387131730344179, 'samples': 6620672, 'steps': 12930, 'loss/train': 2.3379111289978027} 02/24/2022 18:18:06 - INFO - codeparrot_training - Step 12931: {'lr': 0.00043870244058384145, 'samples': 6621184, 'steps': 12931, 'loss/train': 1.8111900091171265} 02/24/2022 18:18:12 - INFO - codeparrot_training - Step 12932: {'lr': 0.0004386917073249237, 'samples': 6621696, 'steps': 12932, 'loss/train': 1.5553728342056274} 02/24/2022 18:18:15 - INFO - codeparrot_training - Step 12933: {'lr': 0.00043868097325771064, 'samples': 6622208, 'steps': 12933, 'loss/train': 2.6121296882629395} 02/24/2022 18:18:21 - INFO - codeparrot_training - Step 12934: {'lr': 0.0004386702383822482, 'samples': 6622720, 'steps': 12934, 'loss/train': 2.1837029457092285} 02/24/2022 18:18:24 - INFO - codeparrot_training - Step 12935: {'lr': 0.00043865950269858224, 'samples': 6623232, 'steps': 12935, 'loss/train': 0.8351127505302429} 02/24/2022 18:18:30 - INFO - codeparrot_training - Step 12936: {'lr': 0.000438648766206759, 'samples': 6623744, 'steps': 12936, 'loss/train': 2.213078022003174} 02/24/2022 18:18:33 - INFO - codeparrot_training - Step 12937: {'lr': 0.0004386380289068243, 'samples': 6624256, 'steps': 12937, 'loss/train': 1.1169012784957886} 02/24/2022 18:18:39 - INFO - codeparrot_training - Step 12938: {'lr': 0.0004386272907988242, 'samples': 6624768, 'steps': 12938, 'loss/train': 1.2164520025253296} 02/24/2022 18:18:42 - INFO - codeparrot_training - Step 12939: {'lr': 0.0004386165518828047, 'samples': 6625280, 'steps': 12939, 'loss/train': 1.025045394897461} 02/24/2022 18:18:48 - INFO - codeparrot_training - Step 12940: {'lr': 0.0004386058121588117, 'samples': 6625792, 'steps': 12940, 'loss/train': 1.1823631525039673} 02/24/2022 18:18:51 - INFO - codeparrot_training - Step 12941: {'lr': 0.0004385950716268914, 'samples': 6626304, 'steps': 12941, 'loss/train': 1.8581124544143677} 02/24/2022 18:18:58 - INFO - codeparrot_training - Step 12942: {'lr': 0.0004385843302870896, 'samples': 6626816, 'steps': 12942, 'loss/train': 1.457020878791809} 02/24/2022 18:19:01 - INFO - codeparrot_training - Step 12943: {'lr': 0.0004385735881394525, 'samples': 6627328, 'steps': 12943, 'loss/train': 1.3479233980178833} 02/24/2022 18:19:07 - INFO - codeparrot_training - Step 12944: {'lr': 0.00043856284518402594, 'samples': 6627840, 'steps': 12944, 'loss/train': 2.614276885986328} 02/24/2022 18:19:10 - INFO - codeparrot_training - Step 12945: {'lr': 0.00043855210142085613, 'samples': 6628352, 'steps': 12945, 'loss/train': 1.8227146863937378} 02/24/2022 18:19:16 - INFO - codeparrot_training - Step 12946: {'lr': 0.00043854135684998893, 'samples': 6628864, 'steps': 12946, 'loss/train': 2.0301523208618164} 02/24/2022 18:19:21 - INFO - codeparrot_training - Step 12947: {'lr': 0.0004385306114714704, 'samples': 6629376, 'steps': 12947, 'loss/train': 1.6931921243667603} 02/24/2022 18:19:25 - INFO - codeparrot_training - Step 12948: {'lr': 0.0004385198652853466, 'samples': 6629888, 'steps': 12948, 'loss/train': 1.820383071899414} 02/24/2022 18:19:28 - INFO - codeparrot_training - Step 12949: {'lr': 0.00043850911829166364, 'samples': 6630400, 'steps': 12949, 'loss/train': 1.475953459739685} 02/24/2022 18:19:34 - INFO - codeparrot_training - Step 12950: {'lr': 0.00043849837049046735, 'samples': 6630912, 'steps': 12950, 'loss/train': 2.13820743560791} 02/24/2022 18:19:37 - INFO - codeparrot_training - Step 12951: {'lr': 0.000438487621881804, 'samples': 6631424, 'steps': 12951, 'loss/train': 2.1887857913970947} 02/24/2022 18:19:44 - INFO - codeparrot_training - Step 12952: {'lr': 0.00043847687246571955, 'samples': 6631936, 'steps': 12952, 'loss/train': 1.4776455163955688} 02/24/2022 18:19:48 - INFO - codeparrot_training - Step 12953: {'lr': 0.0004384661222422599, 'samples': 6632448, 'steps': 12953, 'loss/train': 0.7721335291862488} 02/24/2022 18:19:53 - INFO - codeparrot_training - Step 12954: {'lr': 0.00043845537121147126, 'samples': 6632960, 'steps': 12954, 'loss/train': 0.9670889973640442} 02/24/2022 18:19:59 - INFO - codeparrot_training - Step 12955: {'lr': 0.00043844461937339976, 'samples': 6633472, 'steps': 12955, 'loss/train': 2.7996771335601807} 02/24/2022 18:20:02 - INFO - codeparrot_training - Step 12956: {'lr': 0.00043843386672809127, 'samples': 6633984, 'steps': 12956, 'loss/train': 0.17765621840953827} 02/24/2022 18:20:08 - INFO - codeparrot_training - Step 12957: {'lr': 0.00043842311327559194, 'samples': 6634496, 'steps': 12957, 'loss/train': 2.5870721340179443} 02/24/2022 18:20:11 - INFO - codeparrot_training - Step 12958: {'lr': 0.0004384123590159478, 'samples': 6635008, 'steps': 12958, 'loss/train': 2.4279298782348633} 02/24/2022 18:20:15 - INFO - codeparrot_training - Step 12959: {'lr': 0.000438401603949205, 'samples': 6635520, 'steps': 12959, 'loss/train': 1.7032833099365234} 02/24/2022 18:20:20 - INFO - codeparrot_training - Step 12960: {'lr': 0.0004383908480754095, 'samples': 6636032, 'steps': 12960, 'loss/train': 2.8009588718414307} 02/24/2022 18:20:24 - INFO - codeparrot_training - Step 12961: {'lr': 0.0004383800913946074, 'samples': 6636544, 'steps': 12961, 'loss/train': 2.1121156215667725} 02/24/2022 18:20:29 - INFO - codeparrot_training - Step 12962: {'lr': 0.00043836933390684486, 'samples': 6637056, 'steps': 12962, 'loss/train': 1.0790863037109375} 02/24/2022 18:20:33 - INFO - codeparrot_training - Step 12963: {'lr': 0.0004383585756121679, 'samples': 6637568, 'steps': 12963, 'loss/train': 2.1460609436035156} 02/24/2022 18:20:38 - INFO - codeparrot_training - Step 12964: {'lr': 0.00043834781651062263, 'samples': 6638080, 'steps': 12964, 'loss/train': 2.0029067993164062} 02/24/2022 18:20:42 - INFO - codeparrot_training - Step 12965: {'lr': 0.00043833705660225507, 'samples': 6638592, 'steps': 12965, 'loss/train': 1.537346363067627} 02/24/2022 18:20:47 - INFO - codeparrot_training - Step 12966: {'lr': 0.0004383262958871114, 'samples': 6639104, 'steps': 12966, 'loss/train': 1.7548096179962158} 02/24/2022 18:20:53 - INFO - codeparrot_training - Step 12967: {'lr': 0.0004383155343652377, 'samples': 6639616, 'steps': 12967, 'loss/train': 2.1672027111053467} 02/24/2022 18:20:56 - INFO - codeparrot_training - Step 12968: {'lr': 0.00043830477203668, 'samples': 6640128, 'steps': 12968, 'loss/train': 1.9574977159500122} 02/24/2022 18:21:03 - INFO - codeparrot_training - Step 12969: {'lr': 0.00043829400890148446, 'samples': 6640640, 'steps': 12969, 'loss/train': 1.378501296043396} 02/24/2022 18:21:06 - INFO - codeparrot_training - Step 12970: {'lr': 0.0004382832449596972, 'samples': 6641152, 'steps': 12970, 'loss/train': 1.8503460884094238} 02/24/2022 18:21:12 - INFO - codeparrot_training - Step 12971: {'lr': 0.0004382724802113643, 'samples': 6641664, 'steps': 12971, 'loss/train': 1.3277899026870728} 02/24/2022 18:21:15 - INFO - codeparrot_training - Step 12972: {'lr': 0.0004382617146565319, 'samples': 6642176, 'steps': 12972, 'loss/train': 2.1431236267089844} 02/24/2022 18:21:21 - INFO - codeparrot_training - Step 12973: {'lr': 0.00043825094829524604, 'samples': 6642688, 'steps': 12973, 'loss/train': 2.597724676132202} 02/24/2022 18:21:24 - INFO - codeparrot_training - Step 12974: {'lr': 0.0004382401811275529, 'samples': 6643200, 'steps': 12974, 'loss/train': 0.6234491467475891} 02/24/2022 18:21:29 - INFO - codeparrot_training - Step 12975: {'lr': 0.0004382294131534986, 'samples': 6643712, 'steps': 12975, 'loss/train': 1.8240588903427124} 02/24/2022 18:21:33 - INFO - codeparrot_training - Step 12976: {'lr': 0.00043821864437312933, 'samples': 6644224, 'steps': 12976, 'loss/train': 2.4910523891448975} 02/24/2022 18:21:38 - INFO - codeparrot_training - Step 12977: {'lr': 0.00043820787478649105, 'samples': 6644736, 'steps': 12977, 'loss/train': 1.7990282773971558} 02/24/2022 18:21:42 - INFO - codeparrot_training - Step 12978: {'lr': 0.00043819710439363, 'samples': 6645248, 'steps': 12978, 'loss/train': 2.0579190254211426} 02/24/2022 18:21:48 - INFO - codeparrot_training - Step 12979: {'lr': 0.00043818633319459244, 'samples': 6645760, 'steps': 12979, 'loss/train': 2.8401296138763428} 02/24/2022 18:21:51 - INFO - codeparrot_training - Step 12980: {'lr': 0.00043817556118942426, 'samples': 6646272, 'steps': 12980, 'loss/train': 2.048617124557495} 02/24/2022 18:21:57 - INFO - codeparrot_training - Step 12981: {'lr': 0.00043816478837817183, 'samples': 6646784, 'steps': 12981, 'loss/train': 1.5831965208053589} 02/24/2022 18:22:00 - INFO - codeparrot_training - Step 12982: {'lr': 0.0004381540147608811, 'samples': 6647296, 'steps': 12982, 'loss/train': 1.0544679164886475} 02/24/2022 18:22:06 - INFO - codeparrot_training - Step 12983: {'lr': 0.00043814324033759834, 'samples': 6647808, 'steps': 12983, 'loss/train': 1.7173771858215332} 02/24/2022 18:22:09 - INFO - codeparrot_training - Step 12984: {'lr': 0.0004381324651083697, 'samples': 6648320, 'steps': 12984, 'loss/train': 1.8546584844589233} 02/24/2022 18:22:15 - INFO - codeparrot_training - Step 12985: {'lr': 0.00043812168907324137, 'samples': 6648832, 'steps': 12985, 'loss/train': 2.1783199310302734} 02/24/2022 18:22:18 - INFO - codeparrot_training - Step 12986: {'lr': 0.0004381109122322594, 'samples': 6649344, 'steps': 12986, 'loss/train': 0.3237724006175995} 02/24/2022 18:22:25 - INFO - codeparrot_training - Step 12987: {'lr': 0.00043810013458547007, 'samples': 6649856, 'steps': 12987, 'loss/train': 2.743412971496582} 02/24/2022 18:22:28 - INFO - codeparrot_training - Step 12988: {'lr': 0.00043808935613291934, 'samples': 6650368, 'steps': 12988, 'loss/train': 1.7829055786132812} 02/24/2022 18:22:34 - INFO - codeparrot_training - Step 12989: {'lr': 0.0004380785768746537, 'samples': 6650880, 'steps': 12989, 'loss/train': 1.761879563331604} 02/24/2022 18:22:37 - INFO - codeparrot_training - Step 12990: {'lr': 0.00043806779681071907, 'samples': 6651392, 'steps': 12990, 'loss/train': 1.6081982851028442} 02/24/2022 18:22:43 - INFO - codeparrot_training - Step 12991: {'lr': 0.00043805701594116175, 'samples': 6651904, 'steps': 12991, 'loss/train': 1.6534076929092407} 02/24/2022 18:22:46 - INFO - codeparrot_training - Step 12992: {'lr': 0.00043804623426602784, 'samples': 6652416, 'steps': 12992, 'loss/train': 1.8382694721221924} 02/24/2022 18:22:52 - INFO - codeparrot_training - Step 12993: {'lr': 0.00043803545178536365, 'samples': 6652928, 'steps': 12993, 'loss/train': 1.7728208303451538} 02/24/2022 18:22:55 - INFO - codeparrot_training - Step 12994: {'lr': 0.00043802466849921526, 'samples': 6653440, 'steps': 12994, 'loss/train': 1.5112247467041016} 02/24/2022 18:23:01 - INFO - codeparrot_training - Step 12995: {'lr': 0.0004380138844076289, 'samples': 6653952, 'steps': 12995, 'loss/train': 2.430936098098755} 02/24/2022 18:23:04 - INFO - codeparrot_training - Step 12996: {'lr': 0.00043800309951065076, 'samples': 6654464, 'steps': 12996, 'loss/train': 1.249114751815796} 02/24/2022 18:23:10 - INFO - codeparrot_training - Step 12997: {'lr': 0.000437992313808327, 'samples': 6654976, 'steps': 12997, 'loss/train': 0.6308334469795227} 02/24/2022 18:23:13 - INFO - codeparrot_training - Step 12998: {'lr': 0.0004379815273007039, 'samples': 6655488, 'steps': 12998, 'loss/train': 2.557718276977539} 02/24/2022 18:23:19 - INFO - codeparrot_training - Step 12999: {'lr': 0.0004379707399878276, 'samples': 6656000, 'steps': 12999, 'loss/train': 1.6013245582580566} 02/24/2022 18:23:19 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 18:23:36 - WARNING - huggingface_hub.repository - Several commits (13) will be pushed upstream. 02/24/2022 18:23:36 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 18:24:23 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy e455b12..7d01f0b floral-grass-11 -> floral-grass-11 02/24/2022 18:24:28 - INFO - codeparrot_training - Step 13000: {'lr': 0.00043795995186974435, 'samples': 6656512, 'steps': 13000, 'loss/train': 2.1346521377563477} 02/24/2022 18:24:33 - INFO - codeparrot_training - Step 13001: {'lr': 0.0004379491629465004, 'samples': 6657024, 'steps': 13001, 'loss/train': 1.4408684968948364} 02/24/2022 18:24:37 - INFO - codeparrot_training - Step 13002: {'lr': 0.00043793837321814185, 'samples': 6657536, 'steps': 13002, 'loss/train': 1.5857553482055664} 02/24/2022 18:24:42 - INFO - codeparrot_training - Step 13003: {'lr': 0.000437927582684715, 'samples': 6658048, 'steps': 13003, 'loss/train': 1.689317226409912} 02/24/2022 18:24:46 - INFO - codeparrot_training - Step 13004: {'lr': 0.0004379167913462661, 'samples': 6658560, 'steps': 13004, 'loss/train': 1.8870562314987183} 02/24/2022 18:24:51 - INFO - codeparrot_training - Step 13005: {'lr': 0.0004379059992028412, 'samples': 6659072, 'steps': 13005, 'loss/train': 2.039428234100342} 02/24/2022 18:24:55 - INFO - codeparrot_training - Step 13006: {'lr': 0.00043789520625448685, 'samples': 6659584, 'steps': 13006, 'loss/train': 2.1801905632019043} 02/24/2022 18:25:01 - INFO - codeparrot_training - Step 13007: {'lr': 0.000437884412501249, 'samples': 6660096, 'steps': 13007, 'loss/train': 2.1378047466278076} 02/24/2022 18:25:04 - INFO - codeparrot_training - Step 13008: {'lr': 0.00043787361794317403, 'samples': 6660608, 'steps': 13008, 'loss/train': 2.4507038593292236} 02/24/2022 18:25:10 - INFO - codeparrot_training - Step 13009: {'lr': 0.0004378628225803081, 'samples': 6661120, 'steps': 13009, 'loss/train': 1.9727485179901123} 02/24/2022 18:25:13 - INFO - codeparrot_training - Step 13010: {'lr': 0.0004378520264126975, 'samples': 6661632, 'steps': 13010, 'loss/train': 1.06065833568573} 02/24/2022 18:25:19 - INFO - codeparrot_training - Step 13011: {'lr': 0.0004378412294403885, 'samples': 6662144, 'steps': 13011, 'loss/train': 1.6586016416549683} 02/24/2022 18:25:22 - INFO - codeparrot_training - Step 13012: {'lr': 0.0004378304316634273, 'samples': 6662656, 'steps': 13012, 'loss/train': 2.9551174640655518} 02/24/2022 18:25:29 - INFO - codeparrot_training - Step 13013: {'lr': 0.0004378196330818602, 'samples': 6663168, 'steps': 13013, 'loss/train': 2.4696385860443115} 02/24/2022 18:25:32 - INFO - codeparrot_training - Step 13014: {'lr': 0.00043780883369573336, 'samples': 6663680, 'steps': 13014, 'loss/train': 2.9421322345733643} 02/24/2022 18:25:38 - INFO - codeparrot_training - Step 13015: {'lr': 0.00043779803350509316, 'samples': 6664192, 'steps': 13015, 'loss/train': 1.4251253604888916} 02/24/2022 18:25:41 - INFO - codeparrot_training - Step 13016: {'lr': 0.0004377872325099858, 'samples': 6664704, 'steps': 13016, 'loss/train': 0.14430415630340576} 02/24/2022 18:25:47 - INFO - codeparrot_training - Step 13017: {'lr': 0.0004377764307104576, 'samples': 6665216, 'steps': 13017, 'loss/train': 3.7814066410064697} 02/24/2022 18:25:50 - INFO - codeparrot_training - Step 13018: {'lr': 0.00043776562810655473, 'samples': 6665728, 'steps': 13018, 'loss/train': 1.9222338199615479} 02/24/2022 18:25:56 - INFO - codeparrot_training - Step 13019: {'lr': 0.0004377548246983236, 'samples': 6666240, 'steps': 13019, 'loss/train': 1.3617608547210693} 02/24/2022 18:26:00 - INFO - codeparrot_training - Step 13020: {'lr': 0.0004377440204858104, 'samples': 6666752, 'steps': 13020, 'loss/train': 1.7688344717025757} 02/24/2022 18:26:05 - INFO - codeparrot_training - Step 13021: {'lr': 0.0004377332154690614, 'samples': 6667264, 'steps': 13021, 'loss/train': 1.9365235567092896} 02/24/2022 18:26:09 - INFO - codeparrot_training - Step 13022: {'lr': 0.0004377224096481229, 'samples': 6667776, 'steps': 13022, 'loss/train': 2.4146294593811035} 02/24/2022 18:26:14 - INFO - codeparrot_training - Step 13023: {'lr': 0.0004377116030230413, 'samples': 6668288, 'steps': 13023, 'loss/train': 2.081615686416626} 02/24/2022 18:26:18 - INFO - codeparrot_training - Step 13024: {'lr': 0.0004377007955938628, 'samples': 6668800, 'steps': 13024, 'loss/train': 1.7210839986801147} 02/24/2022 18:26:24 - INFO - codeparrot_training - Step 13025: {'lr': 0.0004376899873606336, 'samples': 6669312, 'steps': 13025, 'loss/train': 1.253495693206787} 02/24/2022 18:26:27 - INFO - codeparrot_training - Step 13026: {'lr': 0.0004376791783234001, 'samples': 6669824, 'steps': 13026, 'loss/train': 2.019789934158325} 02/24/2022 18:26:33 - INFO - codeparrot_training - Step 13027: {'lr': 0.0004376683684822086, 'samples': 6670336, 'steps': 13027, 'loss/train': 0.6084689497947693} 02/24/2022 18:26:36 - INFO - codeparrot_training - Step 13028: {'lr': 0.0004376575578371055, 'samples': 6670848, 'steps': 13028, 'loss/train': 1.7511945962905884} 02/24/2022 18:26:42 - INFO - codeparrot_training - Step 13029: {'lr': 0.0004376467463881369, 'samples': 6671360, 'steps': 13029, 'loss/train': 1.479759931564331} 02/24/2022 18:26:45 - INFO - codeparrot_training - Step 13030: {'lr': 0.0004376359341353492, 'samples': 6671872, 'steps': 13030, 'loss/train': 3.672703266143799} 02/24/2022 18:26:51 - INFO - codeparrot_training - Step 13031: {'lr': 0.00043762512107878884, 'samples': 6672384, 'steps': 13031, 'loss/train': 2.759281873703003} 02/24/2022 18:26:54 - INFO - codeparrot_training - Step 13032: {'lr': 0.00043761430721850206, 'samples': 6672896, 'steps': 13032, 'loss/train': 1.9744319915771484} 02/24/2022 18:27:00 - INFO - codeparrot_training - Step 13033: {'lr': 0.0004376034925545351, 'samples': 6673408, 'steps': 13033, 'loss/train': 2.404545783996582} 02/24/2022 18:27:03 - INFO - codeparrot_training - Step 13034: {'lr': 0.0004375926770869343, 'samples': 6673920, 'steps': 13034, 'loss/train': 2.4459218978881836} 02/24/2022 18:27:09 - INFO - codeparrot_training - Step 13035: {'lr': 0.00043758186081574614, 'samples': 6674432, 'steps': 13035, 'loss/train': 2.392439126968384} 02/24/2022 18:27:13 - INFO - codeparrot_training - Step 13036: {'lr': 0.00043757104374101677, 'samples': 6674944, 'steps': 13036, 'loss/train': 2.167207956314087} 02/24/2022 18:27:18 - INFO - codeparrot_training - Step 13037: {'lr': 0.00043756022586279264, 'samples': 6675456, 'steps': 13037, 'loss/train': 1.6302781105041504} 02/24/2022 18:27:22 - INFO - codeparrot_training - Step 13038: {'lr': 0.00043754940718112, 'samples': 6675968, 'steps': 13038, 'loss/train': 2.507948637008667} 02/24/2022 18:27:28 - INFO - codeparrot_training - Step 13039: {'lr': 0.0004375385876960454, 'samples': 6676480, 'steps': 13039, 'loss/train': 0.4598539173603058} 02/24/2022 18:27:31 - INFO - codeparrot_training - Step 13040: {'lr': 0.0004375277674076149, 'samples': 6676992, 'steps': 13040, 'loss/train': 2.299009084701538} 02/24/2022 18:27:37 - INFO - codeparrot_training - Step 13041: {'lr': 0.00043751694631587504, 'samples': 6677504, 'steps': 13041, 'loss/train': 1.8661608695983887} 02/24/2022 18:27:40 - INFO - codeparrot_training - Step 13042: {'lr': 0.00043750612442087215, 'samples': 6678016, 'steps': 13042, 'loss/train': 1.9540120363235474} 02/24/2022 18:27:46 - INFO - codeparrot_training - Step 13043: {'lr': 0.0004374953017226525, 'samples': 6678528, 'steps': 13043, 'loss/train': 2.130275249481201} 02/24/2022 18:27:49 - INFO - codeparrot_training - Step 13044: {'lr': 0.0004374844782212626, 'samples': 6679040, 'steps': 13044, 'loss/train': 2.635179042816162} 02/24/2022 18:27:55 - INFO - codeparrot_training - Step 13045: {'lr': 0.0004374736539167487, 'samples': 6679552, 'steps': 13045, 'loss/train': 2.7220656871795654} 02/24/2022 18:27:58 - INFO - codeparrot_training - Step 13046: {'lr': 0.0004374628288091571, 'samples': 6680064, 'steps': 13046, 'loss/train': 2.7231671810150146} 02/24/2022 18:28:04 - INFO - codeparrot_training - Step 13047: {'lr': 0.0004374520028985344, 'samples': 6680576, 'steps': 13047, 'loss/train': 0.4303608536720276} 02/24/2022 18:28:07 - INFO - codeparrot_training - Step 13048: {'lr': 0.0004374411761849268, 'samples': 6681088, 'steps': 13048, 'loss/train': 1.8444992303848267} 02/24/2022 18:28:13 - INFO - codeparrot_training - Step 13049: {'lr': 0.0004374303486683807, 'samples': 6681600, 'steps': 13049, 'loss/train': 0.9657476544380188} 02/24/2022 18:28:17 - INFO - codeparrot_training - Step 13050: {'lr': 0.0004374195203489425, 'samples': 6682112, 'steps': 13050, 'loss/train': 2.81044864654541} 02/24/2022 18:28:23 - INFO - codeparrot_training - Step 13051: {'lr': 0.0004374086912266586, 'samples': 6682624, 'steps': 13051, 'loss/train': 3.1938674449920654} 02/24/2022 18:28:26 - INFO - codeparrot_training - Step 13052: {'lr': 0.0004373978613015753, 'samples': 6683136, 'steps': 13052, 'loss/train': 2.0069711208343506} 02/24/2022 18:28:32 - INFO - codeparrot_training - Step 13053: {'lr': 0.0004373870305737392, 'samples': 6683648, 'steps': 13053, 'loss/train': 2.423370122909546} 02/24/2022 18:28:35 - INFO - codeparrot_training - Step 13054: {'lr': 0.00043737619904319654, 'samples': 6684160, 'steps': 13054, 'loss/train': 1.4854636192321777} 02/24/2022 18:28:41 - INFO - codeparrot_training - Step 13055: {'lr': 0.0004373653667099937, 'samples': 6684672, 'steps': 13055, 'loss/train': 1.288827896118164} 02/24/2022 18:28:44 - INFO - codeparrot_training - Step 13056: {'lr': 0.00043735453357417707, 'samples': 6685184, 'steps': 13056, 'loss/train': 2.0451629161834717} 02/24/2022 18:28:50 - INFO - codeparrot_training - Step 13057: {'lr': 0.00043734369963579323, 'samples': 6685696, 'steps': 13057, 'loss/train': 1.8356995582580566} 02/24/2022 18:28:53 - INFO - codeparrot_training - Step 13058: {'lr': 0.0004373328648948884, 'samples': 6686208, 'steps': 13058, 'loss/train': 1.7394015789031982} 02/24/2022 18:28:59 - INFO - codeparrot_training - Step 13059: {'lr': 0.0004373220293515091, 'samples': 6686720, 'steps': 13059, 'loss/train': 1.4280154705047607} 02/24/2022 18:29:02 - INFO - codeparrot_training - Step 13060: {'lr': 0.00043731119300570166, 'samples': 6687232, 'steps': 13060, 'loss/train': 1.967885971069336} 02/24/2022 18:29:08 - INFO - codeparrot_training - Step 13061: {'lr': 0.0004373003558575126, 'samples': 6687744, 'steps': 13061, 'loss/train': 2.5358312129974365} 02/24/2022 18:29:12 - INFO - codeparrot_training - Step 13062: {'lr': 0.00043728951790698823, 'samples': 6688256, 'steps': 13062, 'loss/train': 2.411454439163208} 02/24/2022 18:29:17 - INFO - codeparrot_training - Step 13063: {'lr': 0.00043727867915417505, 'samples': 6688768, 'steps': 13063, 'loss/train': 1.7838704586029053} 02/24/2022 18:29:21 - INFO - codeparrot_training - Step 13064: {'lr': 0.00043726783959911953, 'samples': 6689280, 'steps': 13064, 'loss/train': 1.6130849123001099} 02/24/2022 18:29:27 - INFO - codeparrot_training - Step 13065: {'lr': 0.00043725699924186803, 'samples': 6689792, 'steps': 13065, 'loss/train': 0.8311630487442017} 02/24/2022 18:29:30 - INFO - codeparrot_training - Step 13066: {'lr': 0.00043724615808246695, 'samples': 6690304, 'steps': 13066, 'loss/train': 2.0785346031188965} 02/24/2022 18:29:35 - INFO - codeparrot_training - Step 13067: {'lr': 0.0004372353161209628, 'samples': 6690816, 'steps': 13067, 'loss/train': 0.2719285190105438} 02/24/2022 18:29:39 - INFO - codeparrot_training - Step 13068: {'lr': 0.000437224473357402, 'samples': 6691328, 'steps': 13068, 'loss/train': 1.5799168348312378} 02/24/2022 18:29:44 - INFO - codeparrot_training - Step 13069: {'lr': 0.0004372136297918311, 'samples': 6691840, 'steps': 13069, 'loss/train': 1.7601932287216187} 02/24/2022 18:29:50 - INFO - codeparrot_training - Step 13070: {'lr': 0.0004372027854242964, 'samples': 6692352, 'steps': 13070, 'loss/train': 1.5112253427505493} 02/24/2022 18:29:54 - INFO - codeparrot_training - Step 13071: {'lr': 0.0004371919402548444, 'samples': 6692864, 'steps': 13071, 'loss/train': 0.9533903002738953} 02/24/2022 18:30:00 - INFO - codeparrot_training - Step 13072: {'lr': 0.00043718109428352156, 'samples': 6693376, 'steps': 13072, 'loss/train': 2.0507044792175293} 02/24/2022 18:30:03 - INFO - codeparrot_training - Step 13073: {'lr': 0.00043717024751037436, 'samples': 6693888, 'steps': 13073, 'loss/train': 1.9658995866775513} 02/24/2022 18:30:09 - INFO - codeparrot_training - Step 13074: {'lr': 0.0004371593999354493, 'samples': 6694400, 'steps': 13074, 'loss/train': 1.318845272064209} 02/24/2022 18:30:12 - INFO - codeparrot_training - Step 13075: {'lr': 0.0004371485515587927, 'samples': 6694912, 'steps': 13075, 'loss/train': 1.7926405668258667} 02/24/2022 18:30:18 - INFO - codeparrot_training - Step 13076: {'lr': 0.0004371377023804512, 'samples': 6695424, 'steps': 13076, 'loss/train': 2.1227822303771973} 02/24/2022 18:30:21 - INFO - codeparrot_training - Step 13077: {'lr': 0.00043712685240047125, 'samples': 6695936, 'steps': 13077, 'loss/train': 2.1113078594207764} 02/24/2022 18:30:27 - INFO - codeparrot_training - Step 13078: {'lr': 0.00043711600161889917, 'samples': 6696448, 'steps': 13078, 'loss/train': 2.54327654838562} 02/24/2022 18:30:30 - INFO - codeparrot_training - Step 13079: {'lr': 0.0004371051500357816, 'samples': 6696960, 'steps': 13079, 'loss/train': 3.1392409801483154} 02/24/2022 18:30:36 - INFO - codeparrot_training - Step 13080: {'lr': 0.000437094297651165, 'samples': 6697472, 'steps': 13080, 'loss/train': 0.3492695689201355} 02/24/2022 18:30:40 - INFO - codeparrot_training - Step 13081: {'lr': 0.00043708344446509586, 'samples': 6697984, 'steps': 13081, 'loss/train': 2.4953410625457764} 02/24/2022 18:30:45 - INFO - codeparrot_training - Step 13082: {'lr': 0.0004370725904776206, 'samples': 6698496, 'steps': 13082, 'loss/train': 2.3644306659698486} 02/24/2022 18:30:49 - INFO - codeparrot_training - Step 13083: {'lr': 0.0004370617356887858, 'samples': 6699008, 'steps': 13083, 'loss/train': 2.571532726287842} 02/24/2022 18:30:54 - INFO - codeparrot_training - Step 13084: {'lr': 0.00043705088009863793, 'samples': 6699520, 'steps': 13084, 'loss/train': 2.2510273456573486} 02/24/2022 18:30:58 - INFO - codeparrot_training - Step 13085: {'lr': 0.0004370400237072234, 'samples': 6700032, 'steps': 13085, 'loss/train': 2.9907498359680176} 02/24/2022 18:31:03 - INFO - codeparrot_training - Step 13086: {'lr': 0.0004370291665145889, 'samples': 6700544, 'steps': 13086, 'loss/train': 1.3774452209472656} 02/24/2022 18:31:07 - INFO - codeparrot_training - Step 13087: {'lr': 0.00043701830852078076, 'samples': 6701056, 'steps': 13087, 'loss/train': 2.6954946517944336} 02/24/2022 18:31:12 - INFO - codeparrot_training - Step 13088: {'lr': 0.0004370074497258456, 'samples': 6701568, 'steps': 13088, 'loss/train': 1.8511327505111694} 02/24/2022 18:31:16 - INFO - codeparrot_training - Step 13089: {'lr': 0.00043699659012983, 'samples': 6702080, 'steps': 13089, 'loss/train': 2.4330554008483887} 02/24/2022 18:31:21 - INFO - codeparrot_training - Step 13090: {'lr': 0.00043698572973278026, 'samples': 6702592, 'steps': 13090, 'loss/train': 0.9355196952819824} 02/24/2022 18:31:25 - INFO - codeparrot_training - Step 13091: {'lr': 0.0004369748685347431, 'samples': 6703104, 'steps': 13091, 'loss/train': 2.0947518348693848} 02/24/2022 18:31:30 - INFO - codeparrot_training - Step 13092: {'lr': 0.00043696400653576496, 'samples': 6703616, 'steps': 13092, 'loss/train': 2.21494722366333} 02/24/2022 18:31:34 - INFO - codeparrot_training - Step 13093: {'lr': 0.00043695314373589234, 'samples': 6704128, 'steps': 13093, 'loss/train': 1.9236592054367065} 02/24/2022 18:31:39 - INFO - codeparrot_training - Step 13094: {'lr': 0.00043694228013517185, 'samples': 6704640, 'steps': 13094, 'loss/train': 1.3976666927337646} 02/24/2022 18:31:43 - INFO - codeparrot_training - Step 13095: {'lr': 0.00043693141573365003, 'samples': 6705152, 'steps': 13095, 'loss/train': 2.096853494644165} 02/24/2022 18:31:49 - INFO - codeparrot_training - Step 13096: {'lr': 0.0004369205505313733, 'samples': 6705664, 'steps': 13096, 'loss/train': 2.0255119800567627} 02/24/2022 18:31:52 - INFO - codeparrot_training - Step 13097: {'lr': 0.0004369096845283883, 'samples': 6706176, 'steps': 13097, 'loss/train': 1.203255534172058} 02/24/2022 18:31:58 - INFO - codeparrot_training - Step 13098: {'lr': 0.0004368988177247416, 'samples': 6706688, 'steps': 13098, 'loss/train': 1.4004253149032593} 02/24/2022 18:32:01 - INFO - codeparrot_training - Step 13099: {'lr': 0.00043688795012047975, 'samples': 6707200, 'steps': 13099, 'loss/train': 3.4179847240448} 02/24/2022 18:32:07 - INFO - codeparrot_training - Step 13100: {'lr': 0.00043687708171564923, 'samples': 6707712, 'steps': 13100, 'loss/train': 1.6716430187225342} 02/24/2022 18:32:11 - INFO - codeparrot_training - Step 13101: {'lr': 0.0004368662125102966, 'samples': 6708224, 'steps': 13101, 'loss/train': 0.7230225205421448} 02/24/2022 18:32:16 - INFO - codeparrot_training - Step 13102: {'lr': 0.00043685534250446846, 'samples': 6708736, 'steps': 13102, 'loss/train': 2.184340238571167} 02/24/2022 18:32:20 - INFO - codeparrot_training - Step 13103: {'lr': 0.0004368444716982114, 'samples': 6709248, 'steps': 13103, 'loss/train': 1.14542818069458} 02/24/2022 18:32:25 - INFO - codeparrot_training - Step 13104: {'lr': 0.0004368336000915719, 'samples': 6709760, 'steps': 13104, 'loss/train': 2.194945812225342} 02/24/2022 18:32:29 - INFO - codeparrot_training - Step 13105: {'lr': 0.0004368227276845966, 'samples': 6710272, 'steps': 13105, 'loss/train': 1.4404478073120117} 02/24/2022 18:32:35 - INFO - codeparrot_training - Step 13106: {'lr': 0.0004368118544773321, 'samples': 6710784, 'steps': 13106, 'loss/train': 2.1111202239990234} 02/24/2022 18:32:39 - INFO - codeparrot_training - Step 13107: {'lr': 0.00043680098046982495, 'samples': 6711296, 'steps': 13107, 'loss/train': 3.037461042404175} 02/24/2022 18:32:44 - INFO - codeparrot_training - Step 13108: {'lr': 0.00043679010566212163, 'samples': 6711808, 'steps': 13108, 'loss/train': 1.7464441061019897} 02/24/2022 18:32:48 - INFO - codeparrot_training - Step 13109: {'lr': 0.0004367792300542689, 'samples': 6712320, 'steps': 13109, 'loss/train': 1.7309730052947998} 02/24/2022 18:32:53 - INFO - codeparrot_training - Step 13110: {'lr': 0.00043676835364631316, 'samples': 6712832, 'steps': 13110, 'loss/train': 1.6087268590927124} 02/24/2022 18:32:57 - INFO - codeparrot_training - Step 13111: {'lr': 0.00043675747643830116, 'samples': 6713344, 'steps': 13111, 'loss/train': 3.231724977493286} 02/24/2022 18:33:02 - INFO - codeparrot_training - Step 13112: {'lr': 0.0004367465984302794, 'samples': 6713856, 'steps': 13112, 'loss/train': 1.6795626878738403} 02/24/2022 18:33:06 - INFO - codeparrot_training - Step 13113: {'lr': 0.0004367357196222946, 'samples': 6714368, 'steps': 13113, 'loss/train': 1.807559847831726} 02/24/2022 18:33:12 - INFO - codeparrot_training - Step 13114: {'lr': 0.00043672484001439316, 'samples': 6714880, 'steps': 13114, 'loss/train': 2.354203462600708} 02/24/2022 18:33:15 - INFO - codeparrot_training - Step 13115: {'lr': 0.00043671395960662184, 'samples': 6715392, 'steps': 13115, 'loss/train': 2.1525044441223145} 02/24/2022 18:33:21 - INFO - codeparrot_training - Step 13116: {'lr': 0.0004367030783990272, 'samples': 6715904, 'steps': 13116, 'loss/train': 1.8892061710357666} 02/24/2022 18:33:25 - INFO - codeparrot_training - Step 13117: {'lr': 0.0004366921963916559, 'samples': 6716416, 'steps': 13117, 'loss/train': 1.7278432846069336} 02/24/2022 18:33:30 - INFO - codeparrot_training - Step 13118: {'lr': 0.0004366813135845545, 'samples': 6716928, 'steps': 13118, 'loss/train': 3.2987256050109863} 02/24/2022 18:33:34 - INFO - codeparrot_training - Step 13119: {'lr': 0.00043667042997776965, 'samples': 6717440, 'steps': 13119, 'loss/train': 2.8281095027923584} 02/24/2022 18:33:39 - INFO - codeparrot_training - Step 13120: {'lr': 0.00043665954557134786, 'samples': 6717952, 'steps': 13120, 'loss/train': 2.366917133331299} 02/24/2022 18:33:43 - INFO - codeparrot_training - Step 13121: {'lr': 0.0004366486603653359, 'samples': 6718464, 'steps': 13121, 'loss/train': 1.9473650455474854} 02/24/2022 18:33:48 - INFO - codeparrot_training - Step 13122: {'lr': 0.00043663777435978037, 'samples': 6718976, 'steps': 13122, 'loss/train': 2.808049440383911} 02/24/2022 18:33:52 - INFO - codeparrot_training - Step 13123: {'lr': 0.0004366268875547278, 'samples': 6719488, 'steps': 13123, 'loss/train': 2.011204957962036} 02/24/2022 18:33:57 - INFO - codeparrot_training - Step 13124: {'lr': 0.000436615999950225, 'samples': 6720000, 'steps': 13124, 'loss/train': 2.8301994800567627} 02/24/2022 18:34:01 - INFO - codeparrot_training - Step 13125: {'lr': 0.0004366051115463184, 'samples': 6720512, 'steps': 13125, 'loss/train': 2.068774700164795} 02/24/2022 18:34:07 - INFO - codeparrot_training - Step 13126: {'lr': 0.0004365942223430549, 'samples': 6721024, 'steps': 13126, 'loss/train': 1.955896258354187} 02/24/2022 18:34:10 - INFO - codeparrot_training - Step 13127: {'lr': 0.0004365833323404809, 'samples': 6721536, 'steps': 13127, 'loss/train': 2.6178290843963623} 02/24/2022 18:34:16 - INFO - codeparrot_training - Step 13128: {'lr': 0.0004365724415386432, 'samples': 6722048, 'steps': 13128, 'loss/train': 1.4386388063430786} 02/24/2022 18:34:19 - INFO - codeparrot_training - Step 13129: {'lr': 0.0004365615499375884, 'samples': 6722560, 'steps': 13129, 'loss/train': 2.6756575107574463} 02/24/2022 18:34:25 - INFO - codeparrot_training - Step 13130: {'lr': 0.0004365506575373631, 'samples': 6723072, 'steps': 13130, 'loss/train': 1.0711939334869385} 02/24/2022 18:34:28 - INFO - codeparrot_training - Step 13131: {'lr': 0.0004365397643380141, 'samples': 6723584, 'steps': 13131, 'loss/train': 2.709085464477539} 02/24/2022 18:34:34 - INFO - codeparrot_training - Step 13132: {'lr': 0.000436528870339588, 'samples': 6724096, 'steps': 13132, 'loss/train': 1.7947182655334473} 02/24/2022 18:34:37 - INFO - codeparrot_training - Step 13133: {'lr': 0.0004365179755421314, 'samples': 6724608, 'steps': 13133, 'loss/train': 2.4967987537384033} 02/24/2022 18:34:43 - INFO - codeparrot_training - Step 13134: {'lr': 0.00043650707994569095, 'samples': 6725120, 'steps': 13134, 'loss/train': 1.885949730873108} 02/24/2022 18:34:46 - INFO - codeparrot_training - Step 13135: {'lr': 0.0004364961835503135, 'samples': 6725632, 'steps': 13135, 'loss/train': 2.3373868465423584} 02/24/2022 18:34:52 - INFO - codeparrot_training - Step 13136: {'lr': 0.00043648528635604556, 'samples': 6726144, 'steps': 13136, 'loss/train': 3.3740181922912598} 02/24/2022 18:34:55 - INFO - codeparrot_training - Step 13137: {'lr': 0.00043647438836293383, 'samples': 6726656, 'steps': 13137, 'loss/train': 1.6684520244598389} 02/24/2022 18:35:01 - INFO - codeparrot_training - Step 13138: {'lr': 0.0004364634895710251, 'samples': 6727168, 'steps': 13138, 'loss/train': 2.4966161251068115} 02/24/2022 18:35:05 - INFO - codeparrot_training - Step 13139: {'lr': 0.000436452589980366, 'samples': 6727680, 'steps': 13139, 'loss/train': 2.312422037124634} 02/24/2022 18:35:10 - INFO - codeparrot_training - Step 13140: {'lr': 0.00043644168959100315, 'samples': 6728192, 'steps': 13140, 'loss/train': 1.892502784729004} 02/24/2022 18:35:13 - INFO - codeparrot_training - Step 13141: {'lr': 0.0004364307884029834, 'samples': 6728704, 'steps': 13141, 'loss/train': 2.480815887451172} 02/24/2022 18:35:20 - INFO - codeparrot_training - Step 13142: {'lr': 0.0004364198864163533, 'samples': 6729216, 'steps': 13142, 'loss/train': 2.2424378395080566} 02/24/2022 18:35:23 - INFO - codeparrot_training - Step 13143: {'lr': 0.00043640898363115954, 'samples': 6729728, 'steps': 13143, 'loss/train': 2.198420286178589} 02/24/2022 18:35:29 - INFO - codeparrot_training - Step 13144: {'lr': 0.000436398080047449, 'samples': 6730240, 'steps': 13144, 'loss/train': 0.4985601305961609} 02/24/2022 18:35:32 - INFO - codeparrot_training - Step 13145: {'lr': 0.0004363871756652682, 'samples': 6730752, 'steps': 13145, 'loss/train': 2.4132590293884277} 02/24/2022 18:35:38 - INFO - codeparrot_training - Step 13146: {'lr': 0.00043637627048466395, 'samples': 6731264, 'steps': 13146, 'loss/train': 1.7860790491104126} 02/24/2022 18:35:41 - INFO - codeparrot_training - Step 13147: {'lr': 0.00043636536450568293, 'samples': 6731776, 'steps': 13147, 'loss/train': 1.7798588275909424} 02/24/2022 18:35:47 - INFO - codeparrot_training - Step 13148: {'lr': 0.0004363544577283718, 'samples': 6732288, 'steps': 13148, 'loss/train': 2.481811285018921} 02/24/2022 18:35:50 - INFO - codeparrot_training - Step 13149: {'lr': 0.00043634355015277745, 'samples': 6732800, 'steps': 13149, 'loss/train': 2.4156012535095215} 02/24/2022 18:35:56 - INFO - codeparrot_training - Step 13150: {'lr': 0.0004363326417789465, 'samples': 6733312, 'steps': 13150, 'loss/train': 2.1469268798828125} 02/24/2022 18:35:59 - INFO - codeparrot_training - Step 13151: {'lr': 0.0004363217326069256, 'samples': 6733824, 'steps': 13151, 'loss/train': 2.174891948699951} 02/24/2022 18:36:06 - INFO - codeparrot_training - Step 13152: {'lr': 0.0004363108226367616, 'samples': 6734336, 'steps': 13152, 'loss/train': 1.8401378393173218} 02/24/2022 18:36:09 - INFO - codeparrot_training - Step 13153: {'lr': 0.0004362999118685012, 'samples': 6734848, 'steps': 13153, 'loss/train': 2.1276886463165283} 02/24/2022 18:36:15 - INFO - codeparrot_training - Step 13154: {'lr': 0.0004362890003021911, 'samples': 6735360, 'steps': 13154, 'loss/train': 3.3623147010803223} 02/24/2022 18:36:18 - INFO - codeparrot_training - Step 13155: {'lr': 0.00043627808793787813, 'samples': 6735872, 'steps': 13155, 'loss/train': 1.6755790710449219} 02/24/2022 18:36:24 - INFO - codeparrot_training - Step 13156: {'lr': 0.00043626717477560897, 'samples': 6736384, 'steps': 13156, 'loss/train': 2.3421478271484375} 02/24/2022 18:36:27 - INFO - codeparrot_training - Step 13157: {'lr': 0.00043625626081543033, 'samples': 6736896, 'steps': 13157, 'loss/train': 0.820044219493866} 02/24/2022 18:36:33 - INFO - codeparrot_training - Step 13158: {'lr': 0.0004362453460573891, 'samples': 6737408, 'steps': 13158, 'loss/train': 1.2742438316345215} 02/24/2022 18:36:36 - INFO - codeparrot_training - Step 13159: {'lr': 0.0004362344305015319, 'samples': 6737920, 'steps': 13159, 'loss/train': 2.3827643394470215} 02/24/2022 18:36:42 - INFO - codeparrot_training - Step 13160: {'lr': 0.0004362235141479055, 'samples': 6738432, 'steps': 13160, 'loss/train': 1.021011471748352} 02/24/2022 18:36:45 - INFO - codeparrot_training - Step 13161: {'lr': 0.00043621259699655674, 'samples': 6738944, 'steps': 13161, 'loss/train': 1.639274001121521} 02/24/2022 18:36:52 - INFO - codeparrot_training - Step 13162: {'lr': 0.0004362016790475324, 'samples': 6739456, 'steps': 13162, 'loss/train': 1.8226795196533203} 02/24/2022 18:36:55 - INFO - codeparrot_training - Step 13163: {'lr': 0.0004361907603008791, 'samples': 6739968, 'steps': 13163, 'loss/train': 1.8377959728240967} 02/24/2022 18:37:01 - INFO - codeparrot_training - Step 13164: {'lr': 0.00043617984075664375, 'samples': 6740480, 'steps': 13164, 'loss/train': 2.4114608764648438} 02/24/2022 18:37:04 - INFO - codeparrot_training - Step 13165: {'lr': 0.000436168920414873, 'samples': 6740992, 'steps': 13165, 'loss/train': 5.044597625732422} 02/24/2022 18:37:10 - INFO - codeparrot_training - Step 13166: {'lr': 0.0004361579992756138, 'samples': 6741504, 'steps': 13166, 'loss/train': 2.1394155025482178} 02/24/2022 18:37:13 - INFO - codeparrot_training - Step 13167: {'lr': 0.00043614707733891285, 'samples': 6742016, 'steps': 13167, 'loss/train': 2.0595955848693848} 02/24/2022 18:37:19 - INFO - codeparrot_training - Step 13168: {'lr': 0.00043613615460481685, 'samples': 6742528, 'steps': 13168, 'loss/train': 1.9713562726974487} 02/24/2022 18:37:22 - INFO - codeparrot_training - Step 13169: {'lr': 0.0004361252310733728, 'samples': 6743040, 'steps': 13169, 'loss/train': 2.32989764213562} 02/24/2022 18:37:28 - INFO - codeparrot_training - Step 13170: {'lr': 0.0004361143067446273, 'samples': 6743552, 'steps': 13170, 'loss/train': 2.230102777481079} 02/24/2022 18:37:31 - INFO - codeparrot_training - Step 13171: {'lr': 0.00043610338161862713, 'samples': 6744064, 'steps': 13171, 'loss/train': 2.675751209259033} 02/24/2022 18:37:38 - INFO - codeparrot_training - Step 13172: {'lr': 0.00043609245569541924, 'samples': 6744576, 'steps': 13172, 'loss/train': 2.435805082321167} 02/24/2022 18:37:41 - INFO - codeparrot_training - Step 13173: {'lr': 0.0004360815289750503, 'samples': 6745088, 'steps': 13173, 'loss/train': 1.8633240461349487} 02/24/2022 18:37:47 - INFO - codeparrot_training - Step 13174: {'lr': 0.0004360706014575672, 'samples': 6745600, 'steps': 13174, 'loss/train': 2.5471673011779785} 02/24/2022 18:37:50 - INFO - codeparrot_training - Step 13175: {'lr': 0.00043605967314301673, 'samples': 6746112, 'steps': 13175, 'loss/train': 0.826678454875946} 02/24/2022 18:37:56 - INFO - codeparrot_training - Step 13176: {'lr': 0.0004360487440314458, 'samples': 6746624, 'steps': 13176, 'loss/train': 2.2414722442626953} 02/24/2022 18:38:00 - INFO - codeparrot_training - Step 13177: {'lr': 0.000436037814122901, 'samples': 6747136, 'steps': 13177, 'loss/train': 4.344521522521973} 02/24/2022 18:38:05 - INFO - codeparrot_training - Step 13178: {'lr': 0.0004360268834174294, 'samples': 6747648, 'steps': 13178, 'loss/train': 0.8745490312576294} 02/24/2022 18:38:09 - INFO - codeparrot_training - Step 13179: {'lr': 0.00043601595191507757, 'samples': 6748160, 'steps': 13179, 'loss/train': 1.3443816900253296} 02/24/2022 18:38:15 - INFO - codeparrot_training - Step 13180: {'lr': 0.0004360050196158925, 'samples': 6748672, 'steps': 13180, 'loss/train': 1.440228819847107} 02/24/2022 18:38:19 - INFO - codeparrot_training - Step 13181: {'lr': 0.000435994086519921, 'samples': 6749184, 'steps': 13181, 'loss/train': 2.425391435623169} 02/24/2022 18:38:22 - INFO - codeparrot_training - Step 13182: {'lr': 0.00043598315262720995, 'samples': 6749696, 'steps': 13182, 'loss/train': 3.68023681640625} 02/24/2022 18:38:28 - INFO - codeparrot_training - Step 13183: {'lr': 0.00043597221793780606, 'samples': 6750208, 'steps': 13183, 'loss/train': 1.579888939857483} 02/24/2022 18:38:31 - INFO - codeparrot_training - Step 13184: {'lr': 0.0004359612824517563, 'samples': 6750720, 'steps': 13184, 'loss/train': 1.7140872478485107} 02/24/2022 18:38:37 - INFO - codeparrot_training - Step 13185: {'lr': 0.0004359503461691074, 'samples': 6751232, 'steps': 13185, 'loss/train': 2.5259652137756348} 02/24/2022 18:38:40 - INFO - codeparrot_training - Step 13186: {'lr': 0.00043593940908990625, 'samples': 6751744, 'steps': 13186, 'loss/train': 1.8807111978530884} 02/24/2022 18:38:46 - INFO - codeparrot_training - Step 13187: {'lr': 0.00043592847121419974, 'samples': 6752256, 'steps': 13187, 'loss/train': 2.509537696838379} 02/24/2022 18:38:49 - INFO - codeparrot_training - Step 13188: {'lr': 0.00043591753254203474, 'samples': 6752768, 'steps': 13188, 'loss/train': 2.5340421199798584} 02/24/2022 18:38:56 - INFO - codeparrot_training - Step 13189: {'lr': 0.00043590659307345803, 'samples': 6753280, 'steps': 13189, 'loss/train': 2.152726888656616} 02/24/2022 18:39:00 - INFO - codeparrot_training - Step 13190: {'lr': 0.0004358956528085165, 'samples': 6753792, 'steps': 13190, 'loss/train': 2.254812479019165} 02/24/2022 18:39:05 - INFO - codeparrot_training - Step 13191: {'lr': 0.0004358847117472571, 'samples': 6754304, 'steps': 13191, 'loss/train': 2.7281293869018555} 02/24/2022 18:39:09 - INFO - codeparrot_training - Step 13192: {'lr': 0.00043587376988972655, 'samples': 6754816, 'steps': 13192, 'loss/train': 2.2432076930999756} 02/24/2022 18:39:14 - INFO - codeparrot_training - Step 13193: {'lr': 0.0004358628272359718, 'samples': 6755328, 'steps': 13193, 'loss/train': 4.2456955909729} 02/24/2022 18:39:18 - INFO - codeparrot_training - Step 13194: {'lr': 0.0004358518837860397, 'samples': 6755840, 'steps': 13194, 'loss/train': 2.813513994216919} 02/24/2022 18:39:23 - INFO - codeparrot_training - Step 13195: {'lr': 0.0004358409395399772, 'samples': 6756352, 'steps': 13195, 'loss/train': 1.5284693241119385} 02/24/2022 18:39:27 - INFO - codeparrot_training - Step 13196: {'lr': 0.00043582999449783103, 'samples': 6756864, 'steps': 13196, 'loss/train': 2.2048540115356445} 02/24/2022 18:39:32 - INFO - codeparrot_training - Step 13197: {'lr': 0.00043581904865964825, 'samples': 6757376, 'steps': 13197, 'loss/train': 2.481902837753296} 02/24/2022 18:39:36 - INFO - codeparrot_training - Step 13198: {'lr': 0.0004358081020254756, 'samples': 6757888, 'steps': 13198, 'loss/train': 2.173698902130127} 02/24/2022 18:39:42 - INFO - codeparrot_training - Step 13199: {'lr': 0.0004357971545953601, 'samples': 6758400, 'steps': 13199, 'loss/train': 2.6389975547790527} 02/24/2022 18:39:46 - INFO - codeparrot_training - Step 13200: {'lr': 0.00043578620636934855, 'samples': 6758912, 'steps': 13200, 'loss/train': 2.111168384552002} 02/24/2022 18:39:51 - INFO - codeparrot_training - Step 13201: {'lr': 0.0004357752573474879, 'samples': 6759424, 'steps': 13201, 'loss/train': 2.589674234390259} 02/24/2022 18:39:55 - INFO - codeparrot_training - Step 13202: {'lr': 0.0004357643075298251, 'samples': 6759936, 'steps': 13202, 'loss/train': 1.3990951776504517} 02/24/2022 18:40:00 - INFO - codeparrot_training - Step 13203: {'lr': 0.00043575335691640695, 'samples': 6760448, 'steps': 13203, 'loss/train': 1.8983817100524902} 02/24/2022 18:40:04 - INFO - codeparrot_training - Step 13204: {'lr': 0.0004357424055072804, 'samples': 6760960, 'steps': 13204, 'loss/train': 1.6345925331115723} 02/24/2022 18:40:09 - INFO - codeparrot_training - Step 13205: {'lr': 0.0004357314533024923, 'samples': 6761472, 'steps': 13205, 'loss/train': 2.351658582687378} 02/24/2022 18:40:13 - INFO - codeparrot_training - Step 13206: {'lr': 0.0004357205003020897, 'samples': 6761984, 'steps': 13206, 'loss/train': 1.238976001739502} 02/24/2022 18:40:18 - INFO - codeparrot_training - Step 13207: {'lr': 0.00043570954650611944, 'samples': 6762496, 'steps': 13207, 'loss/train': 2.736942768096924} 02/24/2022 18:40:22 - INFO - codeparrot_training - Step 13208: {'lr': 0.00043569859191462847, 'samples': 6763008, 'steps': 13208, 'loss/train': 1.632069706916809} 02/24/2022 18:40:28 - INFO - codeparrot_training - Step 13209: {'lr': 0.0004356876365276636, 'samples': 6763520, 'steps': 13209, 'loss/train': 0.4474453926086426} 02/24/2022 18:40:31 - INFO - codeparrot_training - Step 13210: {'lr': 0.00043567668034527195, 'samples': 6764032, 'steps': 13210, 'loss/train': 1.9706964492797852} 02/24/2022 18:40:37 - INFO - codeparrot_training - Step 13211: {'lr': 0.0004356657233675004, 'samples': 6764544, 'steps': 13211, 'loss/train': 1.9530082941055298} 02/24/2022 18:40:40 - INFO - codeparrot_training - Step 13212: {'lr': 0.00043565476559439577, 'samples': 6765056, 'steps': 13212, 'loss/train': 2.1657114028930664} 02/24/2022 18:40:46 - INFO - codeparrot_training - Step 13213: {'lr': 0.0004356438070260051, 'samples': 6765568, 'steps': 13213, 'loss/train': 2.159106492996216} 02/24/2022 18:40:49 - INFO - codeparrot_training - Step 13214: {'lr': 0.00043563284766237533, 'samples': 6766080, 'steps': 13214, 'loss/train': 4.27814245223999} 02/24/2022 18:40:55 - INFO - codeparrot_training - Step 13215: {'lr': 0.00043562188750355336, 'samples': 6766592, 'steps': 13215, 'loss/train': 2.5424985885620117} 02/24/2022 18:40:58 - INFO - codeparrot_training - Step 13216: {'lr': 0.0004356109265495861, 'samples': 6767104, 'steps': 13216, 'loss/train': 2.1559038162231445} 02/24/2022 18:41:04 - INFO - codeparrot_training - Step 13217: {'lr': 0.00043559996480052067, 'samples': 6767616, 'steps': 13217, 'loss/train': 1.9655873775482178} 02/24/2022 18:41:07 - INFO - codeparrot_training - Step 13218: {'lr': 0.0004355890022564039, 'samples': 6768128, 'steps': 13218, 'loss/train': 2.3823423385620117} 02/24/2022 18:41:13 - INFO - codeparrot_training - Step 13219: {'lr': 0.00043557803891728275, 'samples': 6768640, 'steps': 13219, 'loss/train': 1.8842852115631104} 02/24/2022 18:41:17 - INFO - codeparrot_training - Step 13220: {'lr': 0.00043556707478320425, 'samples': 6769152, 'steps': 13220, 'loss/train': 1.9322139024734497} 02/24/2022 18:41:22 - INFO - codeparrot_training - Step 13221: {'lr': 0.00043555610985421527, 'samples': 6769664, 'steps': 13221, 'loss/train': 2.273167610168457} 02/24/2022 18:41:26 - INFO - codeparrot_training - Step 13222: {'lr': 0.0004355451441303629, 'samples': 6770176, 'steps': 13222, 'loss/train': 2.4323787689208984} 02/24/2022 18:41:31 - INFO - codeparrot_training - Step 13223: {'lr': 0.000435534177611694, 'samples': 6770688, 'steps': 13223, 'loss/train': 1.3483549356460571} 02/24/2022 18:41:35 - INFO - codeparrot_training - Step 13224: {'lr': 0.0004355232102982556, 'samples': 6771200, 'steps': 13224, 'loss/train': 2.819002389907837} 02/24/2022 18:41:40 - INFO - codeparrot_training - Step 13225: {'lr': 0.00043551224219009473, 'samples': 6771712, 'steps': 13225, 'loss/train': 1.9850126504898071} 02/24/2022 18:41:44 - INFO - codeparrot_training - Step 13226: {'lr': 0.0004355012732872583, 'samples': 6772224, 'steps': 13226, 'loss/train': 1.8604122400283813} 02/24/2022 18:41:50 - INFO - codeparrot_training - Step 13227: {'lr': 0.00043549030358979324, 'samples': 6772736, 'steps': 13227, 'loss/train': 1.710444688796997} 02/24/2022 18:41:53 - INFO - codeparrot_training - Step 13228: {'lr': 0.0004354793330977467, 'samples': 6773248, 'steps': 13228, 'loss/train': 2.4048092365264893} 02/24/2022 18:41:59 - INFO - codeparrot_training - Step 13229: {'lr': 0.00043546836181116555, 'samples': 6773760, 'steps': 13229, 'loss/train': 2.7660040855407715} 02/24/2022 18:42:03 - INFO - codeparrot_training - Step 13230: {'lr': 0.0004354573897300969, 'samples': 6774272, 'steps': 13230, 'loss/train': 1.83676278591156} 02/24/2022 18:42:08 - INFO - codeparrot_training - Step 13231: {'lr': 0.0004354464168545876, 'samples': 6774784, 'steps': 13231, 'loss/train': 2.3647661209106445} 02/24/2022 18:42:12 - INFO - codeparrot_training - Step 13232: {'lr': 0.0004354354431846848, 'samples': 6775296, 'steps': 13232, 'loss/train': 0.3141121566295624} 02/24/2022 18:42:17 - INFO - codeparrot_training - Step 13233: {'lr': 0.0004354244687204354, 'samples': 6775808, 'steps': 13233, 'loss/train': 1.9678032398223877} 02/24/2022 18:42:21 - INFO - codeparrot_training - Step 13234: {'lr': 0.00043541349346188653, 'samples': 6776320, 'steps': 13234, 'loss/train': 1.8628443479537964} 02/24/2022 18:42:26 - INFO - codeparrot_training - Step 13235: {'lr': 0.000435402517409085, 'samples': 6776832, 'steps': 13235, 'loss/train': 1.5543805360794067} 02/24/2022 18:42:32 - INFO - codeparrot_training - Step 13236: {'lr': 0.0004353915405620781, 'samples': 6777344, 'steps': 13236, 'loss/train': 3.8039443492889404} 02/24/2022 18:42:35 - INFO - codeparrot_training - Step 13237: {'lr': 0.0004353805629209126, 'samples': 6777856, 'steps': 13237, 'loss/train': 2.3002946376800537} 02/24/2022 18:42:39 - INFO - codeparrot_training - Step 13238: {'lr': 0.0004353695844856357, 'samples': 6778368, 'steps': 13238, 'loss/train': 1.9879086017608643} 02/24/2022 18:42:44 - INFO - codeparrot_training - Step 13239: {'lr': 0.00043535860525629436, 'samples': 6778880, 'steps': 13239, 'loss/train': 1.1805089712142944} 02/24/2022 18:42:50 - INFO - codeparrot_training - Step 13240: {'lr': 0.00043534762523293557, 'samples': 6779392, 'steps': 13240, 'loss/train': 2.4244401454925537} 02/24/2022 18:42:53 - INFO - codeparrot_training - Step 13241: {'lr': 0.00043533664441560636, 'samples': 6779904, 'steps': 13241, 'loss/train': 1.9856505393981934} 02/24/2022 18:42:59 - INFO - codeparrot_training - Step 13242: {'lr': 0.0004353256628043539, 'samples': 6780416, 'steps': 13242, 'loss/train': 2.9495370388031006} 02/24/2022 18:43:02 - INFO - codeparrot_training - Step 13243: {'lr': 0.00043531468039922515, 'samples': 6780928, 'steps': 13243, 'loss/train': 2.4557976722717285} 02/24/2022 18:43:09 - INFO - codeparrot_training - Step 13244: {'lr': 0.0004353036972002671, 'samples': 6781440, 'steps': 13244, 'loss/train': 1.9739067554473877} 02/24/2022 18:43:12 - INFO - codeparrot_training - Step 13245: {'lr': 0.0004352927132075269, 'samples': 6781952, 'steps': 13245, 'loss/train': 0.5142982602119446} 02/24/2022 18:43:18 - INFO - codeparrot_training - Step 13246: {'lr': 0.00043528172842105154, 'samples': 6782464, 'steps': 13246, 'loss/train': 2.5212056636810303} 02/24/2022 18:43:21 - INFO - codeparrot_training - Step 13247: {'lr': 0.00043527074284088806, 'samples': 6782976, 'steps': 13247, 'loss/train': 1.782245397567749} 02/24/2022 18:43:27 - INFO - codeparrot_training - Step 13248: {'lr': 0.0004352597564670836, 'samples': 6783488, 'steps': 13248, 'loss/train': 1.6626681089401245} 02/24/2022 18:43:30 - INFO - codeparrot_training - Step 13249: {'lr': 0.00043524876929968516, 'samples': 6784000, 'steps': 13249, 'loss/train': 1.9425214529037476} 02/24/2022 18:43:36 - INFO - codeparrot_training - Step 13250: {'lr': 0.0004352377813387398, 'samples': 6784512, 'steps': 13250, 'loss/train': 0.9346581697463989} 02/24/2022 18:43:39 - INFO - codeparrot_training - Step 13251: {'lr': 0.0004352267925842946, 'samples': 6785024, 'steps': 13251, 'loss/train': 2.778203248977661} 02/24/2022 18:43:45 - INFO - codeparrot_training - Step 13252: {'lr': 0.00043521580303639663, 'samples': 6785536, 'steps': 13252, 'loss/train': 1.0520458221435547} 02/24/2022 18:43:48 - INFO - codeparrot_training - Step 13253: {'lr': 0.000435204812695093, 'samples': 6786048, 'steps': 13253, 'loss/train': 3.075870990753174} 02/24/2022 18:43:54 - INFO - codeparrot_training - Step 13254: {'lr': 0.00043519382156043075, 'samples': 6786560, 'steps': 13254, 'loss/train': 2.560676097869873} 02/24/2022 18:43:58 - INFO - codeparrot_training - Step 13255: {'lr': 0.0004351828296324569, 'samples': 6787072, 'steps': 13255, 'loss/train': 2.248198986053467} 02/24/2022 18:44:03 - INFO - codeparrot_training - Step 13256: {'lr': 0.00043517183691121875, 'samples': 6787584, 'steps': 13256, 'loss/train': 2.514848470687866} 02/24/2022 18:44:07 - INFO - codeparrot_training - Step 13257: {'lr': 0.00043516084339676316, 'samples': 6788096, 'steps': 13257, 'loss/train': 2.6968235969543457} 02/24/2022 18:44:12 - INFO - codeparrot_training - Step 13258: {'lr': 0.00043514984908913734, 'samples': 6788608, 'steps': 13258, 'loss/train': 2.6198337078094482} 02/24/2022 18:44:16 - INFO - codeparrot_training - Step 13259: {'lr': 0.0004351388539883883, 'samples': 6789120, 'steps': 13259, 'loss/train': 1.7377967834472656} 02/24/2022 18:44:21 - INFO - codeparrot_training - Step 13260: {'lr': 0.00043512785809456323, 'samples': 6789632, 'steps': 13260, 'loss/train': 1.7979042530059814} 02/24/2022 18:44:25 - INFO - codeparrot_training - Step 13261: {'lr': 0.00043511686140770925, 'samples': 6790144, 'steps': 13261, 'loss/train': 3.0527443885803223} 02/24/2022 18:44:30 - INFO - codeparrot_training - Step 13262: {'lr': 0.0004351058639278734, 'samples': 6790656, 'steps': 13262, 'loss/train': 0.6716596484184265} 02/24/2022 18:44:34 - INFO - codeparrot_training - Step 13263: {'lr': 0.0004350948656551028, 'samples': 6791168, 'steps': 13263, 'loss/train': 3.097402334213257} 02/24/2022 18:44:39 - INFO - codeparrot_training - Step 13264: {'lr': 0.0004350838665894445, 'samples': 6791680, 'steps': 13264, 'loss/train': 1.6201318502426147} 02/24/2022 18:44:43 - INFO - codeparrot_training - Step 13265: {'lr': 0.0004350728667309458, 'samples': 6792192, 'steps': 13265, 'loss/train': 2.168053388595581} 02/24/2022 18:44:49 - INFO - codeparrot_training - Step 13266: {'lr': 0.0004350618660796536, 'samples': 6792704, 'steps': 13266, 'loss/train': 1.9852432012557983} 02/24/2022 18:44:52 - INFO - codeparrot_training - Step 13267: {'lr': 0.0004350508646356152, 'samples': 6793216, 'steps': 13267, 'loss/train': 2.677222967147827} 02/24/2022 18:44:58 - INFO - codeparrot_training - Step 13268: {'lr': 0.00043503986239887765, 'samples': 6793728, 'steps': 13268, 'loss/train': 1.8020939826965332} 02/24/2022 18:45:01 - INFO - codeparrot_training - Step 13269: {'lr': 0.0004350288593694881, 'samples': 6794240, 'steps': 13269, 'loss/train': 2.4397799968719482} 02/24/2022 18:45:07 - INFO - codeparrot_training - Step 13270: {'lr': 0.00043501785554749363, 'samples': 6794752, 'steps': 13270, 'loss/train': 2.659473180770874} 02/24/2022 18:45:10 - INFO - codeparrot_training - Step 13271: {'lr': 0.00043500685093294145, 'samples': 6795264, 'steps': 13271, 'loss/train': 2.226707935333252} 02/24/2022 18:45:16 - INFO - codeparrot_training - Step 13272: {'lr': 0.0004349958455258786, 'samples': 6795776, 'steps': 13272, 'loss/train': 1.7915911674499512} 02/24/2022 18:45:19 - INFO - codeparrot_training - Step 13273: {'lr': 0.00043498483932635237, 'samples': 6796288, 'steps': 13273, 'loss/train': 1.2942596673965454} 02/24/2022 18:45:25 - INFO - codeparrot_training - Step 13274: {'lr': 0.0004349738323344098, 'samples': 6796800, 'steps': 13274, 'loss/train': 2.5414822101593018} 02/24/2022 18:45:28 - INFO - codeparrot_training - Step 13275: {'lr': 0.00043496282455009807, 'samples': 6797312, 'steps': 13275, 'loss/train': 2.5828800201416016} 02/24/2022 18:45:34 - INFO - codeparrot_training - Step 13276: {'lr': 0.00043495181597346435, 'samples': 6797824, 'steps': 13276, 'loss/train': 2.121838092803955} 02/24/2022 18:45:37 - INFO - codeparrot_training - Step 13277: {'lr': 0.0004349408066045557, 'samples': 6798336, 'steps': 13277, 'loss/train': 1.9473315477371216} 02/24/2022 18:45:43 - INFO - codeparrot_training - Step 13278: {'lr': 0.00043492979644341943, 'samples': 6798848, 'steps': 13278, 'loss/train': 2.4124510288238525} 02/24/2022 18:45:46 - INFO - codeparrot_training - Step 13279: {'lr': 0.0004349187854901026, 'samples': 6799360, 'steps': 13279, 'loss/train': 2.3382465839385986} 02/24/2022 18:45:53 - INFO - codeparrot_training - Step 13280: {'lr': 0.00043490777374465244, 'samples': 6799872, 'steps': 13280, 'loss/train': 1.5391281843185425} 02/24/2022 18:45:57 - INFO - codeparrot_training - Step 13281: {'lr': 0.0004348967612071161, 'samples': 6800384, 'steps': 13281, 'loss/train': 2.163550615310669} 02/24/2022 18:46:02 - INFO - codeparrot_training - Step 13282: {'lr': 0.0004348857478775407, 'samples': 6800896, 'steps': 13282, 'loss/train': 2.435809850692749} 02/24/2022 18:46:06 - INFO - codeparrot_training - Step 13283: {'lr': 0.00043487473375597354, 'samples': 6801408, 'steps': 13283, 'loss/train': 1.8940573930740356} 02/24/2022 18:46:11 - INFO - codeparrot_training - Step 13284: {'lr': 0.00043486371884246164, 'samples': 6801920, 'steps': 13284, 'loss/train': 1.351772665977478} 02/24/2022 18:46:15 - INFO - codeparrot_training - Step 13285: {'lr': 0.0004348527031370523, 'samples': 6802432, 'steps': 13285, 'loss/train': 2.4928598403930664} 02/24/2022 18:46:20 - INFO - codeparrot_training - Step 13286: {'lr': 0.00043484168663979265, 'samples': 6802944, 'steps': 13286, 'loss/train': 2.2198801040649414} 02/24/2022 18:46:24 - INFO - codeparrot_training - Step 13287: {'lr': 0.00043483066935073, 'samples': 6803456, 'steps': 13287, 'loss/train': 2.9466030597686768} 02/24/2022 18:46:29 - INFO - codeparrot_training - Step 13288: {'lr': 0.0004348196512699114, 'samples': 6803968, 'steps': 13288, 'loss/train': 2.033850908279419} 02/24/2022 18:46:33 - INFO - codeparrot_training - Step 13289: {'lr': 0.00043480863239738404, 'samples': 6804480, 'steps': 13289, 'loss/train': 3.5231668949127197} 02/24/2022 18:46:39 - INFO - codeparrot_training - Step 13290: {'lr': 0.0004347976127331953, 'samples': 6804992, 'steps': 13290, 'loss/train': 0.2825157344341278} 02/24/2022 18:46:43 - INFO - codeparrot_training - Step 13291: {'lr': 0.00043478659227739216, 'samples': 6805504, 'steps': 13291, 'loss/train': 2.493319034576416} 02/24/2022 18:46:48 - INFO - codeparrot_training - Step 13292: {'lr': 0.00043477557103002197, 'samples': 6806016, 'steps': 13292, 'loss/train': 2.368492603302002} 02/24/2022 18:46:52 - INFO - codeparrot_training - Step 13293: {'lr': 0.00043476454899113193, 'samples': 6806528, 'steps': 13293, 'loss/train': 2.5041770935058594} 02/24/2022 18:46:57 - INFO - codeparrot_training - Step 13294: {'lr': 0.00043475352616076927, 'samples': 6807040, 'steps': 13294, 'loss/train': 1.8983755111694336} 02/24/2022 18:47:01 - INFO - codeparrot_training - Step 13295: {'lr': 0.0004347425025389811, 'samples': 6807552, 'steps': 13295, 'loss/train': 2.371962308883667} 02/24/2022 18:47:07 - INFO - codeparrot_training - Step 13296: {'lr': 0.0004347314781258147, 'samples': 6808064, 'steps': 13296, 'loss/train': 1.1902852058410645} 02/24/2022 18:47:10 - INFO - codeparrot_training - Step 13297: {'lr': 0.00043472045292131735, 'samples': 6808576, 'steps': 13297, 'loss/train': 1.7911893129348755} 02/24/2022 18:47:16 - INFO - codeparrot_training - Step 13298: {'lr': 0.0004347094269255362, 'samples': 6809088, 'steps': 13298, 'loss/train': 2.4793660640716553} 02/24/2022 18:47:19 - INFO - codeparrot_training - Step 13299: {'lr': 0.0004346984001385186, 'samples': 6809600, 'steps': 13299, 'loss/train': 1.0510674715042114} 02/24/2022 18:47:25 - INFO - codeparrot_training - Step 13300: {'lr': 0.00043468737256031155, 'samples': 6810112, 'steps': 13300, 'loss/train': 1.7443426847457886} 02/24/2022 18:47:29 - INFO - codeparrot_training - Step 13301: {'lr': 0.00043467634419096257, 'samples': 6810624, 'steps': 13301, 'loss/train': 2.1588470935821533} 02/24/2022 18:47:34 - INFO - codeparrot_training - Step 13302: {'lr': 0.00043466531503051875, 'samples': 6811136, 'steps': 13302, 'loss/train': 2.269651174545288} 02/24/2022 18:47:38 - INFO - codeparrot_training - Step 13303: {'lr': 0.0004346542850790273, 'samples': 6811648, 'steps': 13303, 'loss/train': 0.8095269799232483} 02/24/2022 18:47:44 - INFO - codeparrot_training - Step 13304: {'lr': 0.00043464325433653563, 'samples': 6812160, 'steps': 13304, 'loss/train': 2.1540334224700928} 02/24/2022 18:47:47 - INFO - codeparrot_training - Step 13305: {'lr': 0.00043463222280309076, 'samples': 6812672, 'steps': 13305, 'loss/train': 2.9058022499084473} 02/24/2022 18:47:51 - INFO - codeparrot_training - Step 13306: {'lr': 0.00043462119047874015, 'samples': 6813184, 'steps': 13306, 'loss/train': 2.6206648349761963} 02/24/2022 18:47:56 - INFO - codeparrot_training - Step 13307: {'lr': 0.000434610157363531, 'samples': 6813696, 'steps': 13307, 'loss/train': 2.822443962097168} 02/24/2022 18:48:00 - INFO - codeparrot_training - Step 13308: {'lr': 0.0004345991234575105, 'samples': 6814208, 'steps': 13308, 'loss/train': 2.512821912765503} 02/24/2022 18:48:05 - INFO - codeparrot_training - Step 13309: {'lr': 0.00043458808876072595, 'samples': 6814720, 'steps': 13309, 'loss/train': 2.2585558891296387} 02/24/2022 18:48:09 - INFO - codeparrot_training - Step 13310: {'lr': 0.0004345770532732247, 'samples': 6815232, 'steps': 13310, 'loss/train': 2.420112133026123} 02/24/2022 18:48:15 - INFO - codeparrot_training - Step 13311: {'lr': 0.00043456601699505407, 'samples': 6815744, 'steps': 13311, 'loss/train': 0.8712084293365479} 02/24/2022 18:48:18 - INFO - codeparrot_training - Step 13312: {'lr': 0.00043455497992626104, 'samples': 6816256, 'steps': 13312, 'loss/train': 1.9645135402679443} 02/24/2022 18:48:24 - INFO - codeparrot_training - Step 13313: {'lr': 0.0004345439420668932, 'samples': 6816768, 'steps': 13313, 'loss/train': 1.6409363746643066} 02/24/2022 18:48:28 - INFO - codeparrot_training - Step 13314: {'lr': 0.0004345329034169977, 'samples': 6817280, 'steps': 13314, 'loss/train': 2.681473731994629} 02/24/2022 18:48:33 - INFO - codeparrot_training - Step 13315: {'lr': 0.00043452186397662174, 'samples': 6817792, 'steps': 13315, 'loss/train': 2.7246925830841064} 02/24/2022 18:48:37 - INFO - codeparrot_training - Step 13316: {'lr': 0.0004345108237458128, 'samples': 6818304, 'steps': 13316, 'loss/train': 1.6731756925582886} 02/24/2022 18:48:42 - INFO - codeparrot_training - Step 13317: {'lr': 0.00043449978272461806, 'samples': 6818816, 'steps': 13317, 'loss/train': 2.075159788131714} 02/24/2022 18:48:46 - INFO - codeparrot_training - Step 13318: {'lr': 0.0004344887409130848, 'samples': 6819328, 'steps': 13318, 'loss/train': 2.192598342895508} 02/24/2022 18:48:51 - INFO - codeparrot_training - Step 13319: {'lr': 0.0004344776983112604, 'samples': 6819840, 'steps': 13319, 'loss/train': 0.6538500189781189} 02/24/2022 18:48:55 - INFO - codeparrot_training - Step 13320: {'lr': 0.0004344666549191921, 'samples': 6820352, 'steps': 13320, 'loss/train': 2.219744920730591} 02/24/2022 18:49:00 - INFO - codeparrot_training - Step 13321: {'lr': 0.0004344556107369272, 'samples': 6820864, 'steps': 13321, 'loss/train': 2.6723968982696533} 02/24/2022 18:49:04 - INFO - codeparrot_training - Step 13322: {'lr': 0.00043444456576451307, 'samples': 6821376, 'steps': 13322, 'loss/train': 1.642527461051941} 02/24/2022 18:49:09 - INFO - codeparrot_training - Step 13323: {'lr': 0.000434433520001997, 'samples': 6821888, 'steps': 13323, 'loss/train': 2.5387818813323975} 02/24/2022 18:49:13 - INFO - codeparrot_training - Step 13324: {'lr': 0.0004344224734494263, 'samples': 6822400, 'steps': 13324, 'loss/train': 1.1320644617080688} 02/24/2022 18:49:18 - INFO - codeparrot_training - Step 13325: {'lr': 0.00043441142610684826, 'samples': 6822912, 'steps': 13325, 'loss/train': 0.46869421005249023} 02/24/2022 18:49:22 - INFO - codeparrot_training - Step 13326: {'lr': 0.0004344003779743102, 'samples': 6823424, 'steps': 13326, 'loss/train': 3.0613038539886475} 02/24/2022 18:49:27 - INFO - codeparrot_training - Step 13327: {'lr': 0.0004343893290518595, 'samples': 6823936, 'steps': 13327, 'loss/train': 2.912856340408325} 02/24/2022 18:49:31 - INFO - codeparrot_training - Step 13328: {'lr': 0.0004343782793395435, 'samples': 6824448, 'steps': 13328, 'loss/train': 1.0939773321151733} 02/24/2022 18:49:38 - INFO - codeparrot_training - Step 13329: {'lr': 0.00043436722883740943, 'samples': 6824960, 'steps': 13329, 'loss/train': 2.795685291290283} 02/24/2022 18:49:41 - INFO - codeparrot_training - Step 13330: {'lr': 0.0004343561775455047, 'samples': 6825472, 'steps': 13330, 'loss/train': 2.1160271167755127} 02/24/2022 18:49:47 - INFO - codeparrot_training - Step 13331: {'lr': 0.00043434512546387674, 'samples': 6825984, 'steps': 13331, 'loss/train': 2.545982599258423} 02/24/2022 18:49:50 - INFO - codeparrot_training - Step 13332: {'lr': 0.0004343340725925727, 'samples': 6826496, 'steps': 13332, 'loss/train': 2.631455898284912} 02/24/2022 18:49:56 - INFO - codeparrot_training - Step 13333: {'lr': 0.0004343230189316401, 'samples': 6827008, 'steps': 13333, 'loss/train': 1.516426682472229} 02/24/2022 18:49:59 - INFO - codeparrot_training - Step 13334: {'lr': 0.00043431196448112615, 'samples': 6827520, 'steps': 13334, 'loss/train': 2.9467992782592773} 02/24/2022 18:50:05 - INFO - codeparrot_training - Step 13335: {'lr': 0.0004343009092410783, 'samples': 6828032, 'steps': 13335, 'loss/train': 2.1306159496307373} 02/24/2022 18:50:10 - INFO - codeparrot_training - Step 13336: {'lr': 0.0004342898532115439, 'samples': 6828544, 'steps': 13336, 'loss/train': 1.820534348487854} 02/24/2022 18:50:14 - INFO - codeparrot_training - Step 13337: {'lr': 0.00043427879639257024, 'samples': 6829056, 'steps': 13337, 'loss/train': 1.6871719360351562} 02/24/2022 18:50:20 - INFO - codeparrot_training - Step 13338: {'lr': 0.0004342677387842048, 'samples': 6829568, 'steps': 13338, 'loss/train': 2.1549885272979736} 02/24/2022 18:50:24 - INFO - codeparrot_training - Step 13339: {'lr': 0.0004342566803864948, 'samples': 6830080, 'steps': 13339, 'loss/train': 1.6332987546920776} 02/24/2022 18:50:29 - INFO - codeparrot_training - Step 13340: {'lr': 0.0004342456211994877, 'samples': 6830592, 'steps': 13340, 'loss/train': 0.8337832689285278} 02/24/2022 18:50:33 - INFO - codeparrot_training - Step 13341: {'lr': 0.0004342345612232309, 'samples': 6831104, 'steps': 13341, 'loss/train': 1.2280763387680054} 02/24/2022 18:50:39 - INFO - codeparrot_training - Step 13342: {'lr': 0.0004342235004577717, 'samples': 6831616, 'steps': 13342, 'loss/train': 2.5863442420959473} 02/24/2022 18:50:42 - INFO - codeparrot_training - Step 13343: {'lr': 0.00043421243890315753, 'samples': 6832128, 'steps': 13343, 'loss/train': 0.7359707951545715} 02/24/2022 18:50:45 - INFO - codeparrot_training - Step 13344: {'lr': 0.0004342013765594358, 'samples': 6832640, 'steps': 13344, 'loss/train': 3.0660793781280518} 02/24/2022 18:50:52 - INFO - codeparrot_training - Step 13345: {'lr': 0.0004341903134266538, 'samples': 6833152, 'steps': 13345, 'loss/train': 2.3280622959136963} 02/24/2022 18:50:55 - INFO - codeparrot_training - Step 13346: {'lr': 0.0004341792495048591, 'samples': 6833664, 'steps': 13346, 'loss/train': 2.480565071105957} 02/24/2022 18:51:01 - INFO - codeparrot_training - Step 13347: {'lr': 0.00043416818479409894, 'samples': 6834176, 'steps': 13347, 'loss/train': 2.3103058338165283} 02/24/2022 18:51:04 - INFO - codeparrot_training - Step 13348: {'lr': 0.0004341571192944207, 'samples': 6834688, 'steps': 13348, 'loss/train': 0.11379817873239517} 02/24/2022 18:51:08 - INFO - codeparrot_training - Step 13349: {'lr': 0.00043414605300587183, 'samples': 6835200, 'steps': 13349, 'loss/train': 2.5075957775115967} 02/24/2022 18:51:14 - INFO - codeparrot_training - Step 13350: {'lr': 0.0004341349859284998, 'samples': 6835712, 'steps': 13350, 'loss/train': 2.269487142562866} 02/24/2022 18:51:17 - INFO - codeparrot_training - Step 13351: {'lr': 0.0004341239180623519, 'samples': 6836224, 'steps': 13351, 'loss/train': 1.6705011129379272} 02/24/2022 18:51:23 - INFO - codeparrot_training - Step 13352: {'lr': 0.0004341128494074756, 'samples': 6836736, 'steps': 13352, 'loss/train': 1.96543288230896} 02/24/2022 18:51:28 - INFO - codeparrot_training - Step 13353: {'lr': 0.00043410177996391837, 'samples': 6837248, 'steps': 13353, 'loss/train': 2.173870086669922} 02/24/2022 18:51:32 - INFO - codeparrot_training - Step 13354: {'lr': 0.00043409070973172753, 'samples': 6837760, 'steps': 13354, 'loss/train': 1.942000389099121} 02/24/2022 18:51:35 - INFO - codeparrot_training - Step 13355: {'lr': 0.0004340796387109506, 'samples': 6838272, 'steps': 13355, 'loss/train': 2.818225383758545} 02/24/2022 18:51:41 - INFO - codeparrot_training - Step 13356: {'lr': 0.00043406856690163487, 'samples': 6838784, 'steps': 13356, 'loss/train': 1.9128395318984985} 02/24/2022 18:51:48 - INFO - codeparrot_training - Step 13357: {'lr': 0.0004340574943038279, 'samples': 6839296, 'steps': 13357, 'loss/train': 1.6974326372146606} 02/24/2022 18:51:52 - INFO - codeparrot_training - Step 13358: {'lr': 0.00043404642091757705, 'samples': 6839808, 'steps': 13358, 'loss/train': 0.8889933228492737} 02/24/2022 18:51:57 - INFO - codeparrot_training - Step 13359: {'lr': 0.0004340353467429299, 'samples': 6840320, 'steps': 13359, 'loss/train': 2.474644899368286} 02/24/2022 18:52:01 - INFO - codeparrot_training - Step 13360: {'lr': 0.00043402427177993366, 'samples': 6840832, 'steps': 13360, 'loss/train': 2.1150295734405518} 02/24/2022 18:52:06 - INFO - codeparrot_training - Step 13361: {'lr': 0.00043401319602863584, 'samples': 6841344, 'steps': 13361, 'loss/train': 1.3615708351135254} 02/24/2022 18:52:10 - INFO - codeparrot_training - Step 13362: {'lr': 0.0004340021194890839, 'samples': 6841856, 'steps': 13362, 'loss/train': 1.4819518327713013} 02/24/2022 18:52:15 - INFO - codeparrot_training - Step 13363: {'lr': 0.0004339910421613253, 'samples': 6842368, 'steps': 13363, 'loss/train': 2.0121209621429443} 02/24/2022 18:52:19 - INFO - codeparrot_training - Step 13364: {'lr': 0.0004339799640454076, 'samples': 6842880, 'steps': 13364, 'loss/train': 1.5883493423461914} 02/24/2022 18:52:24 - INFO - codeparrot_training - Step 13365: {'lr': 0.0004339688851413781, 'samples': 6843392, 'steps': 13365, 'loss/train': 1.5042500495910645} 02/24/2022 18:52:28 - INFO - codeparrot_training - Step 13366: {'lr': 0.0004339578054492843, 'samples': 6843904, 'steps': 13366, 'loss/train': 2.0957536697387695} 02/24/2022 18:52:34 - INFO - codeparrot_training - Step 13367: {'lr': 0.0004339467249691737, 'samples': 6844416, 'steps': 13367, 'loss/train': 2.4097070693969727} 02/24/2022 18:52:37 - INFO - codeparrot_training - Step 13368: {'lr': 0.0004339356437010937, 'samples': 6844928, 'steps': 13368, 'loss/train': 2.522966146469116} 02/24/2022 18:52:43 - INFO - codeparrot_training - Step 13369: {'lr': 0.00043392456164509185, 'samples': 6845440, 'steps': 13369, 'loss/train': 2.0167455673217773} 02/24/2022 18:52:46 - INFO - codeparrot_training - Step 13370: {'lr': 0.00043391347880121554, 'samples': 6845952, 'steps': 13370, 'loss/train': 2.096789598464966} 02/24/2022 18:52:52 - INFO - codeparrot_training - Step 13371: {'lr': 0.00043390239516951235, 'samples': 6846464, 'steps': 13371, 'loss/train': 2.0977468490600586} 02/24/2022 18:52:56 - INFO - codeparrot_training - Step 13372: {'lr': 0.0004338913107500297, 'samples': 6846976, 'steps': 13372, 'loss/train': 3.343346357345581} 02/24/2022 18:53:01 - INFO - codeparrot_training - Step 13373: {'lr': 0.00043388022554281504, 'samples': 6847488, 'steps': 13373, 'loss/train': 2.03631854057312} 02/24/2022 18:53:05 - INFO - codeparrot_training - Step 13374: {'lr': 0.00043386913954791584, 'samples': 6848000, 'steps': 13374, 'loss/train': 2.520256519317627} 02/24/2022 18:53:10 - INFO - codeparrot_training - Step 13375: {'lr': 0.0004338580527653797, 'samples': 6848512, 'steps': 13375, 'loss/train': 2.7449727058410645} 02/24/2022 18:53:14 - INFO - codeparrot_training - Step 13376: {'lr': 0.000433846965195254, 'samples': 6849024, 'steps': 13376, 'loss/train': 1.5822027921676636} 02/24/2022 18:53:20 - INFO - codeparrot_training - Step 13377: {'lr': 0.0004338358768375863, 'samples': 6849536, 'steps': 13377, 'loss/train': 1.7537716627120972} 02/24/2022 18:53:23 - INFO - codeparrot_training - Step 13378: {'lr': 0.000433824787692424, 'samples': 6850048, 'steps': 13378, 'loss/train': 2.6050710678100586} 02/24/2022 18:53:29 - INFO - codeparrot_training - Step 13379: {'lr': 0.0004338136977598148, 'samples': 6850560, 'steps': 13379, 'loss/train': 2.313286542892456} 02/24/2022 18:53:32 - INFO - codeparrot_training - Step 13380: {'lr': 0.000433802607039806, 'samples': 6851072, 'steps': 13380, 'loss/train': 1.917705774307251} 02/24/2022 18:53:38 - INFO - codeparrot_training - Step 13381: {'lr': 0.00043379151553244523, 'samples': 6851584, 'steps': 13381, 'loss/train': 2.555523157119751} 02/24/2022 18:53:41 - INFO - codeparrot_training - Step 13382: {'lr': 0.00043378042323778, 'samples': 6852096, 'steps': 13382, 'loss/train': 1.6866884231567383} 02/24/2022 18:53:47 - INFO - codeparrot_training - Step 13383: {'lr': 0.00043376933015585776, 'samples': 6852608, 'steps': 13383, 'loss/train': 0.8979901671409607} 02/24/2022 18:53:50 - INFO - codeparrot_training - Step 13384: {'lr': 0.000433758236286726, 'samples': 6853120, 'steps': 13384, 'loss/train': 2.1274333000183105} 02/24/2022 18:53:56 - INFO - codeparrot_training - Step 13385: {'lr': 0.0004337471416304324, 'samples': 6853632, 'steps': 13385, 'loss/train': 2.098496913909912} 02/24/2022 18:53:59 - INFO - codeparrot_training - Step 13386: {'lr': 0.00043373604618702436, 'samples': 6854144, 'steps': 13386, 'loss/train': 1.6126383543014526} 02/24/2022 18:54:06 - INFO - codeparrot_training - Step 13387: {'lr': 0.00043372494995654943, 'samples': 6854656, 'steps': 13387, 'loss/train': 1.942667007446289} 02/24/2022 18:54:09 - INFO - codeparrot_training - Step 13388: {'lr': 0.00043371385293905517, 'samples': 6855168, 'steps': 13388, 'loss/train': 2.040477752685547} 02/24/2022 18:54:15 - INFO - codeparrot_training - Step 13389: {'lr': 0.0004337027551345891, 'samples': 6855680, 'steps': 13389, 'loss/train': 0.6316954493522644} 02/24/2022 18:54:18 - INFO - codeparrot_training - Step 13390: {'lr': 0.0004336916565431987, 'samples': 6856192, 'steps': 13390, 'loss/train': 1.7285830974578857} 02/24/2022 18:54:24 - INFO - codeparrot_training - Step 13391: {'lr': 0.0004336805571649316, 'samples': 6856704, 'steps': 13391, 'loss/train': 0.21150599420070648} 02/24/2022 18:54:27 - INFO - codeparrot_training - Step 13392: {'lr': 0.0004336694569998354, 'samples': 6857216, 'steps': 13392, 'loss/train': 1.7394059896469116} 02/24/2022 18:54:33 - INFO - codeparrot_training - Step 13393: {'lr': 0.00043365835604795746, 'samples': 6857728, 'steps': 13393, 'loss/train': 2.860260248184204} 02/24/2022 18:54:36 - INFO - codeparrot_training - Step 13394: {'lr': 0.0004336472543093455, 'samples': 6858240, 'steps': 13394, 'loss/train': 2.5539393424987793} 02/24/2022 18:54:42 - INFO - codeparrot_training - Step 13395: {'lr': 0.000433636151784047, 'samples': 6858752, 'steps': 13395, 'loss/train': 1.6724191904067993} 02/24/2022 18:54:45 - INFO - codeparrot_training - Step 13396: {'lr': 0.00043362504847210956, 'samples': 6859264, 'steps': 13396, 'loss/train': 1.963356614112854} 02/24/2022 18:54:51 - INFO - codeparrot_training - Step 13397: {'lr': 0.0004336139443735807, 'samples': 6859776, 'steps': 13397, 'loss/train': 2.2137465476989746} 02/24/2022 18:54:54 - INFO - codeparrot_training - Step 13398: {'lr': 0.000433602839488508, 'samples': 6860288, 'steps': 13398, 'loss/train': 2.2427098751068115} 02/24/2022 18:55:00 - INFO - codeparrot_training - Step 13399: {'lr': 0.00043359173381693906, 'samples': 6860800, 'steps': 13399, 'loss/train': 1.8561426401138306} 02/24/2022 18:55:03 - INFO - codeparrot_training - Step 13400: {'lr': 0.0004335806273589214, 'samples': 6861312, 'steps': 13400, 'loss/train': 0.9804264903068542} 02/24/2022 18:55:09 - INFO - codeparrot_training - Step 13401: {'lr': 0.00043356952011450265, 'samples': 6861824, 'steps': 13401, 'loss/train': 2.3413875102996826} 02/24/2022 18:55:13 - INFO - codeparrot_training - Step 13402: {'lr': 0.0004335584120837304, 'samples': 6862336, 'steps': 13402, 'loss/train': 1.9771844148635864} 02/24/2022 18:55:19 - INFO - codeparrot_training - Step 13403: {'lr': 0.0004335473032666521, 'samples': 6862848, 'steps': 13403, 'loss/train': 2.8203139305114746} 02/24/2022 18:55:22 - INFO - codeparrot_training - Step 13404: {'lr': 0.00043353619366331546, 'samples': 6863360, 'steps': 13404, 'loss/train': 1.2091772556304932} 02/24/2022 18:55:28 - INFO - codeparrot_training - Step 13405: {'lr': 0.0004335250832737681, 'samples': 6863872, 'steps': 13405, 'loss/train': 1.1968530416488647} 02/24/2022 18:55:31 - INFO - codeparrot_training - Step 13406: {'lr': 0.00043351397209805755, 'samples': 6864384, 'steps': 13406, 'loss/train': 1.3702644109725952} 02/24/2022 18:55:37 - INFO - codeparrot_training - Step 13407: {'lr': 0.0004335028601362314, 'samples': 6864896, 'steps': 13407, 'loss/train': 1.7492936849594116} 02/24/2022 18:55:40 - INFO - codeparrot_training - Step 13408: {'lr': 0.0004334917473883373, 'samples': 6865408, 'steps': 13408, 'loss/train': 2.7847790718078613} 02/24/2022 18:55:46 - INFO - codeparrot_training - Step 13409: {'lr': 0.0004334806338544227, 'samples': 6865920, 'steps': 13409, 'loss/train': 2.966379165649414} 02/24/2022 18:55:49 - INFO - codeparrot_training - Step 13410: {'lr': 0.0004334695195345355, 'samples': 6866432, 'steps': 13410, 'loss/train': 1.479927659034729} 02/24/2022 18:55:55 - INFO - codeparrot_training - Step 13411: {'lr': 0.000433458404428723, 'samples': 6866944, 'steps': 13411, 'loss/train': 2.035550355911255} 02/24/2022 18:55:58 - INFO - codeparrot_training - Step 13412: {'lr': 0.00043344728853703297, 'samples': 6867456, 'steps': 13412, 'loss/train': 2.2750933170318604} 02/24/2022 18:56:04 - INFO - codeparrot_training - Step 13413: {'lr': 0.00043343617185951305, 'samples': 6867968, 'steps': 13413, 'loss/train': 2.1236705780029297} 02/24/2022 18:56:08 - INFO - codeparrot_training - Step 13414: {'lr': 0.0004334250543962108, 'samples': 6868480, 'steps': 13414, 'loss/train': 2.128988027572632} 02/24/2022 18:56:13 - INFO - codeparrot_training - Step 13415: {'lr': 0.00043341393614717384, 'samples': 6868992, 'steps': 13415, 'loss/train': 1.898743987083435} 02/24/2022 18:56:17 - INFO - codeparrot_training - Step 13416: {'lr': 0.0004334028171124499, 'samples': 6869504, 'steps': 13416, 'loss/train': 2.1568007469177246} 02/24/2022 18:56:22 - INFO - codeparrot_training - Step 13417: {'lr': 0.0004333916972920864, 'samples': 6870016, 'steps': 13417, 'loss/train': 2.27823543548584} 02/24/2022 18:56:26 - INFO - codeparrot_training - Step 13418: {'lr': 0.00043338057668613117, 'samples': 6870528, 'steps': 13418, 'loss/train': 1.9554824829101562} 02/24/2022 18:56:31 - INFO - codeparrot_training - Step 13419: {'lr': 0.00043336945529463177, 'samples': 6871040, 'steps': 13419, 'loss/train': 1.2947747707366943} 02/24/2022 18:56:35 - INFO - codeparrot_training - Step 13420: {'lr': 0.00043335833311763597, 'samples': 6871552, 'steps': 13420, 'loss/train': 1.5561705827713013} 02/24/2022 18:56:40 - INFO - codeparrot_training - Step 13421: {'lr': 0.00043334721015519115, 'samples': 6872064, 'steps': 13421, 'loss/train': 2.5657753944396973} 02/24/2022 18:56:44 - INFO - codeparrot_training - Step 13422: {'lr': 0.00043333608640734513, 'samples': 6872576, 'steps': 13422, 'loss/train': 2.1511964797973633} 02/24/2022 18:56:51 - INFO - codeparrot_training - Step 13423: {'lr': 0.0004333249618741455, 'samples': 6873088, 'steps': 13423, 'loss/train': 1.474923014640808} 02/24/2022 18:56:54 - INFO - codeparrot_training - Step 13424: {'lr': 0.00043331383655564003, 'samples': 6873600, 'steps': 13424, 'loss/train': 1.8163540363311768} 02/24/2022 18:56:59 - INFO - codeparrot_training - Step 13425: {'lr': 0.0004333027104518762, 'samples': 6874112, 'steps': 13425, 'loss/train': 2.0045571327209473} 02/24/2022 18:57:03 - INFO - codeparrot_training - Step 13426: {'lr': 0.00043329158356290187, 'samples': 6874624, 'steps': 13426, 'loss/train': 1.8382295370101929} 02/24/2022 18:57:08 - INFO - codeparrot_training - Step 13427: {'lr': 0.00043328045588876454, 'samples': 6875136, 'steps': 13427, 'loss/train': 1.6040394306182861} 02/24/2022 18:57:12 - INFO - codeparrot_training - Step 13428: {'lr': 0.0004332693274295119, 'samples': 6875648, 'steps': 13428, 'loss/train': 1.1823590993881226} 02/24/2022 18:57:17 - INFO - codeparrot_training - Step 13429: {'lr': 0.0004332581981851917, 'samples': 6876160, 'steps': 13429, 'loss/train': 1.4518547058105469} 02/24/2022 18:57:21 - INFO - codeparrot_training - Step 13430: {'lr': 0.00043324706815585156, 'samples': 6876672, 'steps': 13430, 'loss/train': 1.9056583642959595} 02/24/2022 18:57:26 - INFO - codeparrot_training - Step 13431: {'lr': 0.00043323593734153915, 'samples': 6877184, 'steps': 13431, 'loss/train': 1.6445872783660889} 02/24/2022 18:57:30 - INFO - codeparrot_training - Step 13432: {'lr': 0.00043322480574230215, 'samples': 6877696, 'steps': 13432, 'loss/train': 1.6481211185455322} 02/24/2022 18:57:35 - INFO - codeparrot_training - Step 13433: {'lr': 0.00043321367335818833, 'samples': 6878208, 'steps': 13433, 'loss/train': 2.1485066413879395} 02/24/2022 18:57:39 - INFO - codeparrot_training - Step 13434: {'lr': 0.0004332025401892453, 'samples': 6878720, 'steps': 13434, 'loss/train': 2.0059142112731934} 02/24/2022 18:57:45 - INFO - codeparrot_training - Step 13435: {'lr': 0.00043319140623552073, 'samples': 6879232, 'steps': 13435, 'loss/train': 2.371858835220337} 02/24/2022 18:57:49 - INFO - codeparrot_training - Step 13436: {'lr': 0.0004331802714970624, 'samples': 6879744, 'steps': 13436, 'loss/train': 3.014699697494507} 02/24/2022 18:57:54 - INFO - codeparrot_training - Step 13437: {'lr': 0.00043316913597391785, 'samples': 6880256, 'steps': 13437, 'loss/train': 2.0820155143737793} 02/24/2022 18:57:58 - INFO - codeparrot_training - Step 13438: {'lr': 0.00043315799966613496, 'samples': 6880768, 'steps': 13438, 'loss/train': 1.8914895057678223} 02/24/2022 18:58:03 - INFO - codeparrot_training - Step 13439: {'lr': 0.00043314686257376136, 'samples': 6881280, 'steps': 13439, 'loss/train': 1.5246726274490356} 02/24/2022 18:58:07 - INFO - codeparrot_training - Step 13440: {'lr': 0.0004331357246968447, 'samples': 6881792, 'steps': 13440, 'loss/train': 0.6227343082427979} 02/24/2022 18:58:12 - INFO - codeparrot_training - Step 13441: {'lr': 0.0004331245860354328, 'samples': 6882304, 'steps': 13441, 'loss/train': 2.1620190143585205} 02/24/2022 18:58:16 - INFO - codeparrot_training - Step 13442: {'lr': 0.0004331134465895733, 'samples': 6882816, 'steps': 13442, 'loss/train': 0.47370097041130066} 02/24/2022 18:58:21 - INFO - codeparrot_training - Step 13443: {'lr': 0.00043310230635931394, 'samples': 6883328, 'steps': 13443, 'loss/train': 1.625957727432251} 02/24/2022 18:58:25 - INFO - codeparrot_training - Step 13444: {'lr': 0.0004330911653447024, 'samples': 6883840, 'steps': 13444, 'loss/train': 1.1147783994674683} 02/24/2022 18:58:30 - INFO - codeparrot_training - Step 13445: {'lr': 0.0004330800235457866, 'samples': 6884352, 'steps': 13445, 'loss/train': 2.186103582382202} 02/24/2022 18:58:34 - INFO - codeparrot_training - Step 13446: {'lr': 0.00043306888096261394, 'samples': 6884864, 'steps': 13446, 'loss/train': 3.074098825454712} 02/24/2022 18:58:39 - INFO - codeparrot_training - Step 13447: {'lr': 0.0004330577375952324, 'samples': 6885376, 'steps': 13447, 'loss/train': 1.6368879079818726} 02/24/2022 18:58:43 - INFO - codeparrot_training - Step 13448: {'lr': 0.0004330465934436896, 'samples': 6885888, 'steps': 13448, 'loss/train': 0.1662389039993286} 02/24/2022 18:58:50 - INFO - codeparrot_training - Step 13449: {'lr': 0.0004330354485080334, 'samples': 6886400, 'steps': 13449, 'loss/train': 0.9146532416343689} 02/24/2022 18:58:53 - INFO - codeparrot_training - Step 13450: {'lr': 0.0004330243027883114, 'samples': 6886912, 'steps': 13450, 'loss/train': 2.3914132118225098} 02/24/2022 18:58:58 - INFO - codeparrot_training - Step 13451: {'lr': 0.0004330131562845714, 'samples': 6887424, 'steps': 13451, 'loss/train': 2.218575954437256} 02/24/2022 18:59:02 - INFO - codeparrot_training - Step 13452: {'lr': 0.00043300200899686113, 'samples': 6887936, 'steps': 13452, 'loss/train': 1.9424468278884888} 02/24/2022 18:59:07 - INFO - codeparrot_training - Step 13453: {'lr': 0.0004329908609252284, 'samples': 6888448, 'steps': 13453, 'loss/train': 2.5715761184692383} 02/24/2022 18:59:11 - INFO - codeparrot_training - Step 13454: {'lr': 0.00043297971206972095, 'samples': 6888960, 'steps': 13454, 'loss/train': 1.081486463546753} 02/24/2022 18:59:17 - INFO - codeparrot_training - Step 13455: {'lr': 0.0004329685624303865, 'samples': 6889472, 'steps': 13455, 'loss/train': 2.1514930725097656} 02/24/2022 18:59:20 - INFO - codeparrot_training - Step 13456: {'lr': 0.0004329574120072728, 'samples': 6889984, 'steps': 13456, 'loss/train': 2.117000102996826} 02/24/2022 18:59:26 - INFO - codeparrot_training - Step 13457: {'lr': 0.00043294626080042767, 'samples': 6890496, 'steps': 13457, 'loss/train': 2.0635335445404053} 02/24/2022 18:59:29 - INFO - codeparrot_training - Step 13458: {'lr': 0.0004329351088098988, 'samples': 6891008, 'steps': 13458, 'loss/train': 1.7340364456176758} 02/24/2022 18:59:36 - INFO - codeparrot_training - Step 13459: {'lr': 0.0004329239560357341, 'samples': 6891520, 'steps': 13459, 'loss/train': 1.3921512365341187} 02/24/2022 18:59:39 - INFO - codeparrot_training - Step 13460: {'lr': 0.0004329128024779812, 'samples': 6892032, 'steps': 13460, 'loss/train': 1.8784161806106567} 02/24/2022 18:59:45 - INFO - codeparrot_training - Step 13461: {'lr': 0.00043290164813668795, 'samples': 6892544, 'steps': 13461, 'loss/train': 2.0356743335723877} 02/24/2022 18:59:48 - INFO - codeparrot_training - Step 13462: {'lr': 0.0004328904930119021, 'samples': 6893056, 'steps': 13462, 'loss/train': 2.024158477783203} 02/24/2022 18:59:54 - INFO - codeparrot_training - Step 13463: {'lr': 0.0004328793371036714, 'samples': 6893568, 'steps': 13463, 'loss/train': 1.0979338884353638} 02/24/2022 18:59:57 - INFO - codeparrot_training - Step 13464: {'lr': 0.0004328681804120438, 'samples': 6894080, 'steps': 13464, 'loss/train': 1.8582379817962646} 02/24/2022 19:00:03 - INFO - codeparrot_training - Step 13465: {'lr': 0.000432857022937067, 'samples': 6894592, 'steps': 13465, 'loss/train': 1.650192379951477} 02/24/2022 19:00:06 - INFO - codeparrot_training - Step 13466: {'lr': 0.00043284586467878865, 'samples': 6895104, 'steps': 13466, 'loss/train': 2.0719406604766846} 02/24/2022 19:00:12 - INFO - codeparrot_training - Step 13467: {'lr': 0.0004328347056372568, 'samples': 6895616, 'steps': 13467, 'loss/train': 2.1869418621063232} 02/24/2022 19:00:15 - INFO - codeparrot_training - Step 13468: {'lr': 0.00043282354581251903, 'samples': 6896128, 'steps': 13468, 'loss/train': 1.8174875974655151} 02/24/2022 19:00:22 - INFO - codeparrot_training - Step 13469: {'lr': 0.0004328123852046233, 'samples': 6896640, 'steps': 13469, 'loss/train': 1.8327851295471191} 02/24/2022 19:00:25 - INFO - codeparrot_training - Step 13470: {'lr': 0.0004328012238136173, 'samples': 6897152, 'steps': 13470, 'loss/train': 2.7620625495910645} 02/24/2022 19:00:31 - INFO - codeparrot_training - Step 13471: {'lr': 0.000432790061639549, 'samples': 6897664, 'steps': 13471, 'loss/train': 1.7467918395996094} 02/24/2022 19:00:34 - INFO - codeparrot_training - Step 13472: {'lr': 0.00043277889868246605, 'samples': 6898176, 'steps': 13472, 'loss/train': 3.432149887084961} 02/24/2022 19:00:40 - INFO - codeparrot_training - Step 13473: {'lr': 0.0004327677349424164, 'samples': 6898688, 'steps': 13473, 'loss/train': 1.9549671411514282} 02/24/2022 19:00:43 - INFO - codeparrot_training - Step 13474: {'lr': 0.0004327565704194477, 'samples': 6899200, 'steps': 13474, 'loss/train': 1.76982843875885} 02/24/2022 19:00:49 - INFO - codeparrot_training - Step 13475: {'lr': 0.0004327454051136079, 'samples': 6899712, 'steps': 13475, 'loss/train': 2.368537425994873} 02/24/2022 19:00:53 - INFO - codeparrot_training - Step 13476: {'lr': 0.0004327342390249449, 'samples': 6900224, 'steps': 13476, 'loss/train': 2.650076389312744} 02/24/2022 19:00:58 - INFO - codeparrot_training - Step 13477: {'lr': 0.00043272307215350635, 'samples': 6900736, 'steps': 13477, 'loss/train': 0.5039814114570618} 02/24/2022 19:01:02 - INFO - codeparrot_training - Step 13478: {'lr': 0.0004327119044993403, 'samples': 6901248, 'steps': 13478, 'loss/train': 2.7400102615356445} 02/24/2022 19:01:07 - INFO - codeparrot_training - Step 13479: {'lr': 0.0004327007360624944, 'samples': 6901760, 'steps': 13479, 'loss/train': 2.3195197582244873} 02/24/2022 19:01:10 - INFO - codeparrot_training - Step 13480: {'lr': 0.0004326895668430165, 'samples': 6902272, 'steps': 13480, 'loss/train': 3.7669460773468018} 02/24/2022 19:01:17 - INFO - codeparrot_training - Step 13481: {'lr': 0.0004326783968409546, 'samples': 6902784, 'steps': 13481, 'loss/train': 2.371272087097168} 02/24/2022 19:01:20 - INFO - codeparrot_training - Step 13482: {'lr': 0.00043266722605635644, 'samples': 6903296, 'steps': 13482, 'loss/train': 2.170497417449951} 02/24/2022 19:01:26 - INFO - codeparrot_training - Step 13483: {'lr': 0.0004326560544892699, 'samples': 6903808, 'steps': 13483, 'loss/train': 1.9850348234176636} 02/24/2022 19:01:29 - INFO - codeparrot_training - Step 13484: {'lr': 0.00043264488213974275, 'samples': 6904320, 'steps': 13484, 'loss/train': 1.9562952518463135} 02/24/2022 19:01:35 - INFO - codeparrot_training - Step 13485: {'lr': 0.00043263370900782297, 'samples': 6904832, 'steps': 13485, 'loss/train': 2.2027151584625244} 02/24/2022 19:01:38 - INFO - codeparrot_training - Step 13486: {'lr': 0.0004326225350935583, 'samples': 6905344, 'steps': 13486, 'loss/train': 1.579418659210205} 02/24/2022 19:01:44 - INFO - codeparrot_training - Step 13487: {'lr': 0.00043261136039699676, 'samples': 6905856, 'steps': 13487, 'loss/train': 1.213860273361206} 02/24/2022 19:01:47 - INFO - codeparrot_training - Step 13488: {'lr': 0.0004326001849181862, 'samples': 6906368, 'steps': 13488, 'loss/train': 2.603863000869751} 02/24/2022 19:01:53 - INFO - codeparrot_training - Step 13489: {'lr': 0.0004325890086571743, 'samples': 6906880, 'steps': 13489, 'loss/train': 1.5113205909729004} 02/24/2022 19:01:56 - INFO - codeparrot_training - Step 13490: {'lr': 0.00043257783161400917, 'samples': 6907392, 'steps': 13490, 'loss/train': 2.793163776397705} 02/24/2022 19:02:02 - INFO - codeparrot_training - Step 13491: {'lr': 0.0004325666537887385, 'samples': 6907904, 'steps': 13491, 'loss/train': 0.5322578549385071} 02/24/2022 19:02:05 - INFO - codeparrot_training - Step 13492: {'lr': 0.00043255547518141033, 'samples': 6908416, 'steps': 13492, 'loss/train': 2.004286050796509} 02/24/2022 19:02:11 - INFO - codeparrot_training - Step 13493: {'lr': 0.0004325442957920724, 'samples': 6908928, 'steps': 13493, 'loss/train': 0.12340640276670456} 02/24/2022 19:02:14 - INFO - codeparrot_training - Step 13494: {'lr': 0.0004325331156207727, 'samples': 6909440, 'steps': 13494, 'loss/train': 0.9747650623321533} 02/24/2022 19:02:21 - INFO - codeparrot_training - Step 13495: {'lr': 0.00043252193466755906, 'samples': 6909952, 'steps': 13495, 'loss/train': 1.4896043539047241} 02/24/2022 19:02:24 - INFO - codeparrot_training - Step 13496: {'lr': 0.0004325107529324795, 'samples': 6910464, 'steps': 13496, 'loss/train': 2.659832715988159} 02/24/2022 19:02:30 - INFO - codeparrot_training - Step 13497: {'lr': 0.0004324995704155817, 'samples': 6910976, 'steps': 13497, 'loss/train': 1.9089082479476929} 02/24/2022 19:02:33 - INFO - codeparrot_training - Step 13498: {'lr': 0.0004324883871169138, 'samples': 6911488, 'steps': 13498, 'loss/train': 2.02074933052063} 02/24/2022 19:02:39 - INFO - codeparrot_training - Step 13499: {'lr': 0.00043247720303652353, 'samples': 6912000, 'steps': 13499, 'loss/train': 0.494108110666275} 02/24/2022 19:02:42 - INFO - codeparrot_training - Step 13500: {'lr': 0.0004324660181744589, 'samples': 6912512, 'steps': 13500, 'loss/train': 2.467935085296631} 02/24/2022 19:02:48 - INFO - codeparrot_training - Step 13501: {'lr': 0.00043245483253076777, 'samples': 6913024, 'steps': 13501, 'loss/train': 1.9143275022506714} 02/24/2022 19:02:51 - INFO - codeparrot_training - Step 13502: {'lr': 0.0004324436461054981, 'samples': 6913536, 'steps': 13502, 'loss/train': 1.2303646802902222} 02/24/2022 19:02:57 - INFO - codeparrot_training - Step 13503: {'lr': 0.00043243245889869775, 'samples': 6914048, 'steps': 13503, 'loss/train': 3.229280948638916} 02/24/2022 19:03:00 - INFO - codeparrot_training - Step 13504: {'lr': 0.0004324212709104147, 'samples': 6914560, 'steps': 13504, 'loss/train': 1.7756668329238892} 02/24/2022 19:03:06 - INFO - codeparrot_training - Step 13505: {'lr': 0.0004324100821406969, 'samples': 6915072, 'steps': 13505, 'loss/train': 0.8699193596839905} 02/24/2022 19:03:09 - INFO - codeparrot_training - Step 13506: {'lr': 0.00043239889258959215, 'samples': 6915584, 'steps': 13506, 'loss/train': 1.8626275062561035} 02/24/2022 19:03:16 - INFO - codeparrot_training - Step 13507: {'lr': 0.00043238770225714854, 'samples': 6916096, 'steps': 13507, 'loss/train': 0.3798440992832184} 02/24/2022 19:03:20 - INFO - codeparrot_training - Step 13508: {'lr': 0.00043237651114341383, 'samples': 6916608, 'steps': 13508, 'loss/train': 2.3320095539093018} 02/24/2022 19:03:25 - INFO - codeparrot_training - Step 13509: {'lr': 0.0004323653192484361, 'samples': 6917120, 'steps': 13509, 'loss/train': 2.872229814529419} 02/24/2022 19:03:29 - INFO - codeparrot_training - Step 13510: {'lr': 0.0004323541265722633, 'samples': 6917632, 'steps': 13510, 'loss/train': 1.3713935613632202} 02/24/2022 19:03:34 - INFO - codeparrot_training - Step 13511: {'lr': 0.0004323429331149432, 'samples': 6918144, 'steps': 13511, 'loss/train': 1.5069878101348877} 02/24/2022 19:03:38 - INFO - codeparrot_training - Step 13512: {'lr': 0.000432331738876524, 'samples': 6918656, 'steps': 13512, 'loss/train': 3.0355465412139893} 02/24/2022 19:03:43 - INFO - codeparrot_training - Step 13513: {'lr': 0.00043232054385705345, 'samples': 6919168, 'steps': 13513, 'loss/train': 1.9760501384735107} 02/24/2022 19:03:47 - INFO - codeparrot_training - Step 13514: {'lr': 0.0004323093480565796, 'samples': 6919680, 'steps': 13514, 'loss/train': 1.7026857137680054} 02/24/2022 19:03:52 - INFO - codeparrot_training - Step 13515: {'lr': 0.0004322981514751504, 'samples': 6920192, 'steps': 13515, 'loss/train': 1.505321741104126} 02/24/2022 19:03:56 - INFO - codeparrot_training - Step 13516: {'lr': 0.0004322869541128138, 'samples': 6920704, 'steps': 13516, 'loss/train': 1.5018874406814575} 02/24/2022 19:04:02 - INFO - codeparrot_training - Step 13517: {'lr': 0.00043227575596961783, 'samples': 6921216, 'steps': 13517, 'loss/train': 2.7122464179992676} 02/24/2022 19:04:06 - INFO - codeparrot_training - Step 13518: {'lr': 0.00043226455704561034, 'samples': 6921728, 'steps': 13518, 'loss/train': 2.9626035690307617} 02/24/2022 19:04:11 - INFO - codeparrot_training - Step 13519: {'lr': 0.0004322533573408394, 'samples': 6922240, 'steps': 13519, 'loss/train': 1.6918985843658447} 02/24/2022 19:04:15 - INFO - codeparrot_training - Step 13520: {'lr': 0.00043224215685535287, 'samples': 6922752, 'steps': 13520, 'loss/train': 2.0553183555603027} 02/24/2022 19:04:20 - INFO - codeparrot_training - Step 13521: {'lr': 0.0004322309555891989, 'samples': 6923264, 'steps': 13521, 'loss/train': 2.2857887744903564} 02/24/2022 19:04:24 - INFO - codeparrot_training - Step 13522: {'lr': 0.00043221975354242536, 'samples': 6923776, 'steps': 13522, 'loss/train': 2.6552069187164307} 02/24/2022 19:04:29 - INFO - codeparrot_training - Step 13523: {'lr': 0.0004322085507150802, 'samples': 6924288, 'steps': 13523, 'loss/train': 0.5546449422836304} 02/24/2022 19:04:33 - INFO - codeparrot_training - Step 13524: {'lr': 0.00043219734710721146, 'samples': 6924800, 'steps': 13524, 'loss/train': 1.7232391834259033} 02/24/2022 19:04:38 - INFO - codeparrot_training - Step 13525: {'lr': 0.00043218614271886725, 'samples': 6925312, 'steps': 13525, 'loss/train': 1.8191306591033936} 02/24/2022 19:04:42 - INFO - codeparrot_training - Step 13526: {'lr': 0.0004321749375500954, 'samples': 6925824, 'steps': 13526, 'loss/train': 2.7485036849975586} 02/24/2022 19:04:48 - INFO - codeparrot_training - Step 13527: {'lr': 0.0004321637316009439, 'samples': 6926336, 'steps': 13527, 'loss/train': 2.03973650932312} 02/24/2022 19:04:51 - INFO - codeparrot_training - Step 13528: {'lr': 0.00043215252487146096, 'samples': 6926848, 'steps': 13528, 'loss/train': 1.5198118686676025} 02/24/2022 19:04:57 - INFO - codeparrot_training - Step 13529: {'lr': 0.0004321413173616943, 'samples': 6927360, 'steps': 13529, 'loss/train': 1.5036728382110596} 02/24/2022 19:05:00 - INFO - codeparrot_training - Step 13530: {'lr': 0.00043213010907169213, 'samples': 6927872, 'steps': 13530, 'loss/train': 2.8774921894073486} 02/24/2022 19:05:06 - INFO - codeparrot_training - Step 13531: {'lr': 0.00043211890000150247, 'samples': 6928384, 'steps': 13531, 'loss/train': 1.2315572500228882} 02/24/2022 19:05:09 - INFO - codeparrot_training - Step 13532: {'lr': 0.0004321076901511731, 'samples': 6928896, 'steps': 13532, 'loss/train': 1.9138476848602295} 02/24/2022 19:05:15 - INFO - codeparrot_training - Step 13533: {'lr': 0.00043209647952075235, 'samples': 6929408, 'steps': 13533, 'loss/train': 1.8021141290664673} 02/24/2022 19:05:18 - INFO - codeparrot_training - Step 13534: {'lr': 0.00043208526811028806, 'samples': 6929920, 'steps': 13534, 'loss/train': 1.293871283531189} 02/24/2022 19:05:24 - INFO - codeparrot_training - Step 13535: {'lr': 0.00043207405591982835, 'samples': 6930432, 'steps': 13535, 'loss/train': 1.8187755346298218} 02/24/2022 19:05:27 - INFO - codeparrot_training - Step 13536: {'lr': 0.0004320628429494212, 'samples': 6930944, 'steps': 13536, 'loss/train': 2.0494000911712646} 02/24/2022 19:05:33 - INFO - codeparrot_training - Step 13537: {'lr': 0.00043205162919911455, 'samples': 6931456, 'steps': 13537, 'loss/train': 2.9351913928985596} 02/24/2022 19:05:36 - INFO - codeparrot_training - Step 13538: {'lr': 0.0004320404146689566, 'samples': 6931968, 'steps': 13538, 'loss/train': 1.784485936164856} 02/24/2022 19:05:42 - INFO - codeparrot_training - Step 13539: {'lr': 0.0004320291993589953, 'samples': 6932480, 'steps': 13539, 'loss/train': 1.4332714080810547} 02/24/2022 19:05:45 - INFO - codeparrot_training - Step 13540: {'lr': 0.0004320179832692787, 'samples': 6932992, 'steps': 13540, 'loss/train': 2.744737148284912} 02/24/2022 19:05:51 - INFO - codeparrot_training - Step 13541: {'lr': 0.0004320067663998549, 'samples': 6933504, 'steps': 13541, 'loss/train': 2.7020082473754883} 02/24/2022 19:05:57 - INFO - codeparrot_training - Step 13542: {'lr': 0.00043199554875077183, 'samples': 6934016, 'steps': 13542, 'loss/train': 2.465879201889038} 02/24/2022 19:06:01 - INFO - codeparrot_training - Step 13543: {'lr': 0.00043198433032207774, 'samples': 6934528, 'steps': 13543, 'loss/train': 1.4109089374542236} 02/24/2022 19:06:06 - INFO - codeparrot_training - Step 13544: {'lr': 0.00043197311111382045, 'samples': 6935040, 'steps': 13544, 'loss/train': 1.873335599899292} 02/24/2022 19:06:10 - INFO - codeparrot_training - Step 13545: {'lr': 0.0004319618911260482, 'samples': 6935552, 'steps': 13545, 'loss/train': 2.226513147354126} 02/24/2022 19:06:13 - INFO - codeparrot_training - Step 13546: {'lr': 0.0004319506703588089, 'samples': 6936064, 'steps': 13546, 'loss/train': 1.7113468647003174} 02/24/2022 19:06:19 - INFO - codeparrot_training - Step 13547: {'lr': 0.00043193944881215075, 'samples': 6936576, 'steps': 13547, 'loss/train': 2.309114456176758} 02/24/2022 19:06:22 - INFO - codeparrot_training - Step 13548: {'lr': 0.00043192822648612184, 'samples': 6937088, 'steps': 13548, 'loss/train': 0.13835661113262177} 02/24/2022 19:06:28 - INFO - codeparrot_training - Step 13549: {'lr': 0.0004319170033807701, 'samples': 6937600, 'steps': 13549, 'loss/train': 1.448536992073059} 02/24/2022 19:06:34 - INFO - codeparrot_training - Step 13550: {'lr': 0.00043190577949614375, 'samples': 6938112, 'steps': 13550, 'loss/train': 1.6425881385803223} 02/24/2022 19:06:38 - INFO - codeparrot_training - Step 13551: {'lr': 0.00043189455483229073, 'samples': 6938624, 'steps': 13551, 'loss/train': 2.04683518409729} 02/24/2022 19:06:41 - INFO - codeparrot_training - Step 13552: {'lr': 0.00043188332938925923, 'samples': 6939136, 'steps': 13552, 'loss/train': 2.404266357421875} 02/24/2022 19:06:45 - INFO - codeparrot_training - Step 13553: {'lr': 0.0004318721031670973, 'samples': 6939648, 'steps': 13553, 'loss/train': 2.4058384895324707} 02/24/2022 19:06:51 - INFO - codeparrot_training - Step 13554: {'lr': 0.00043186087616585303, 'samples': 6940160, 'steps': 13554, 'loss/train': 2.257333755493164} 02/24/2022 19:06:54 - INFO - codeparrot_training - Step 13555: {'lr': 0.0004318496483855745, 'samples': 6940672, 'steps': 13555, 'loss/train': 2.4622716903686523} 02/24/2022 19:07:00 - INFO - codeparrot_training - Step 13556: {'lr': 0.0004318384198263099, 'samples': 6941184, 'steps': 13556, 'loss/train': 1.5970168113708496} 02/24/2022 19:07:03 - INFO - codeparrot_training - Step 13557: {'lr': 0.00043182719048810714, 'samples': 6941696, 'steps': 13557, 'loss/train': 2.133849620819092} 02/24/2022 19:07:09 - INFO - codeparrot_training - Step 13558: {'lr': 0.00043181596037101443, 'samples': 6942208, 'steps': 13558, 'loss/train': 2.686640501022339} 02/24/2022 19:07:12 - INFO - codeparrot_training - Step 13559: {'lr': 0.00043180472947508, 'samples': 6942720, 'steps': 13559, 'loss/train': 1.2532762289047241} 02/24/2022 19:07:18 - INFO - codeparrot_training - Step 13560: {'lr': 0.0004317934978003517, 'samples': 6943232, 'steps': 13560, 'loss/train': 3.6173346042633057} 02/24/2022 19:07:21 - INFO - codeparrot_training - Step 13561: {'lr': 0.0004317822653468778, 'samples': 6943744, 'steps': 13561, 'loss/train': 2.8468985557556152} 02/24/2022 19:07:27 - INFO - codeparrot_training - Step 13562: {'lr': 0.00043177103211470647, 'samples': 6944256, 'steps': 13562, 'loss/train': 1.8747682571411133} 02/24/2022 19:07:30 - INFO - codeparrot_training - Step 13563: {'lr': 0.00043175979810388575, 'samples': 6944768, 'steps': 13563, 'loss/train': 2.063985824584961} 02/24/2022 19:07:37 - INFO - codeparrot_training - Step 13564: {'lr': 0.0004317485633144638, 'samples': 6945280, 'steps': 13564, 'loss/train': 3.4184486865997314} 02/24/2022 19:07:41 - INFO - codeparrot_training - Step 13565: {'lr': 0.0004317373277464886, 'samples': 6945792, 'steps': 13565, 'loss/train': 1.3864854574203491} 02/24/2022 19:07:46 - INFO - codeparrot_training - Step 13566: {'lr': 0.0004317260914000085, 'samples': 6946304, 'steps': 13566, 'loss/train': 1.22243332862854} 02/24/2022 19:07:50 - INFO - codeparrot_training - Step 13567: {'lr': 0.00043171485427507145, 'samples': 6946816, 'steps': 13567, 'loss/train': 1.9268128871917725} 02/24/2022 19:07:55 - INFO - codeparrot_training - Step 13568: {'lr': 0.0004317036163717257, 'samples': 6947328, 'steps': 13568, 'loss/train': 4.796426773071289} 02/24/2022 19:07:59 - INFO - codeparrot_training - Step 13569: {'lr': 0.00043169237769001936, 'samples': 6947840, 'steps': 13569, 'loss/train': 2.196671724319458} 02/24/2022 19:08:04 - INFO - codeparrot_training - Step 13570: {'lr': 0.0004316811382300006, 'samples': 6948352, 'steps': 13570, 'loss/train': 2.194404125213623} 02/24/2022 19:08:08 - INFO - codeparrot_training - Step 13571: {'lr': 0.0004316698979917175, 'samples': 6948864, 'steps': 13571, 'loss/train': 1.0558006763458252} 02/24/2022 19:08:13 - INFO - codeparrot_training - Step 13572: {'lr': 0.0004316586569752182, 'samples': 6949376, 'steps': 13572, 'loss/train': 2.7078187465667725} 02/24/2022 19:08:17 - INFO - codeparrot_training - Step 13573: {'lr': 0.00043164741518055097, 'samples': 6949888, 'steps': 13573, 'loss/train': 2.9460179805755615} 02/24/2022 19:08:23 - INFO - codeparrot_training - Step 13574: {'lr': 0.0004316361726077639, 'samples': 6950400, 'steps': 13574, 'loss/train': 1.184583067893982} 02/24/2022 19:08:26 - INFO - codeparrot_training - Step 13575: {'lr': 0.0004316249292569051, 'samples': 6950912, 'steps': 13575, 'loss/train': 1.6814242601394653} 02/24/2022 19:08:32 - INFO - codeparrot_training - Step 13576: {'lr': 0.0004316136851280228, 'samples': 6951424, 'steps': 13576, 'loss/train': 0.5795541405677795} 02/24/2022 19:08:36 - INFO - codeparrot_training - Step 13577: {'lr': 0.00043160244022116514, 'samples': 6951936, 'steps': 13577, 'loss/train': 1.4604476690292358} 02/24/2022 19:08:41 - INFO - codeparrot_training - Step 13578: {'lr': 0.0004315911945363802, 'samples': 6952448, 'steps': 13578, 'loss/train': 1.8934495449066162} 02/24/2022 19:08:45 - INFO - codeparrot_training - Step 13579: {'lr': 0.00043157994807371634, 'samples': 6952960, 'steps': 13579, 'loss/train': 1.6934781074523926} 02/24/2022 19:08:50 - INFO - codeparrot_training - Step 13580: {'lr': 0.00043156870083322166, 'samples': 6953472, 'steps': 13580, 'loss/train': 1.825491189956665} 02/24/2022 19:08:54 - INFO - codeparrot_training - Step 13581: {'lr': 0.0004315574528149443, 'samples': 6953984, 'steps': 13581, 'loss/train': 1.064252495765686} 02/24/2022 19:08:59 - INFO - codeparrot_training - Step 13582: {'lr': 0.00043154620401893244, 'samples': 6954496, 'steps': 13582, 'loss/train': 2.3033533096313477} 02/24/2022 19:09:03 - INFO - codeparrot_training - Step 13583: {'lr': 0.0004315349544452343, 'samples': 6955008, 'steps': 13583, 'loss/train': 1.6521732807159424} 02/24/2022 19:09:08 - INFO - codeparrot_training - Step 13584: {'lr': 0.00043152370409389794, 'samples': 6955520, 'steps': 13584, 'loss/train': 3.037415027618408} 02/24/2022 19:09:12 - INFO - codeparrot_training - Step 13585: {'lr': 0.00043151245296497184, 'samples': 6956032, 'steps': 13585, 'loss/train': 1.4827784299850464} 02/24/2022 19:09:17 - INFO - codeparrot_training - Step 13586: {'lr': 0.000431501201058504, 'samples': 6956544, 'steps': 13586, 'loss/train': 2.637298583984375} 02/24/2022 19:09:21 - INFO - codeparrot_training - Step 13587: {'lr': 0.0004314899483745426, 'samples': 6957056, 'steps': 13587, 'loss/train': 2.552401065826416} 02/24/2022 19:09:26 - INFO - codeparrot_training - Step 13588: {'lr': 0.0004314786949131359, 'samples': 6957568, 'steps': 13588, 'loss/train': 2.3268959522247314} 02/24/2022 19:09:30 - INFO - codeparrot_training - Step 13589: {'lr': 0.0004314674406743321, 'samples': 6958080, 'steps': 13589, 'loss/train': 2.118502378463745} 02/24/2022 19:09:36 - INFO - codeparrot_training - Step 13590: {'lr': 0.00043145618565817946, 'samples': 6958592, 'steps': 13590, 'loss/train': 2.1962571144104004} 02/24/2022 19:09:41 - INFO - codeparrot_training - Step 13591: {'lr': 0.00043144492986472603, 'samples': 6959104, 'steps': 13591, 'loss/train': 1.1924455165863037} 02/24/2022 19:09:45 - INFO - codeparrot_training - Step 13592: {'lr': 0.0004314336732940202, 'samples': 6959616, 'steps': 13592, 'loss/train': 2.9314956665039062} 02/24/2022 19:09:50 - INFO - codeparrot_training - Step 13593: {'lr': 0.0004314224159461102, 'samples': 6960128, 'steps': 13593, 'loss/train': 2.478891134262085} 02/24/2022 19:09:54 - INFO - codeparrot_training - Step 13594: {'lr': 0.0004314111578210441, 'samples': 6960640, 'steps': 13594, 'loss/train': 1.8915724754333496} 02/24/2022 19:09:59 - INFO - codeparrot_training - Step 13595: {'lr': 0.0004313998989188702, 'samples': 6961152, 'steps': 13595, 'loss/train': 1.4155244827270508} 02/24/2022 19:10:03 - INFO - codeparrot_training - Step 13596: {'lr': 0.00043138863923963664, 'samples': 6961664, 'steps': 13596, 'loss/train': 2.005995750427246} 02/24/2022 19:10:08 - INFO - codeparrot_training - Step 13597: {'lr': 0.0004313773787833919, 'samples': 6962176, 'steps': 13597, 'loss/train': 1.8183064460754395} 02/24/2022 19:10:12 - INFO - codeparrot_training - Step 13598: {'lr': 0.0004313661175501841, 'samples': 6962688, 'steps': 13598, 'loss/train': 2.2222249507904053} 02/24/2022 19:10:18 - INFO - codeparrot_training - Step 13599: {'lr': 0.00043135485554006127, 'samples': 6963200, 'steps': 13599, 'loss/train': 1.8670982122421265} 02/24/2022 19:10:21 - INFO - codeparrot_training - Step 13600: {'lr': 0.0004313435927530719, 'samples': 6963712, 'steps': 13600, 'loss/train': 2.011847496032715} 02/24/2022 19:10:27 - INFO - codeparrot_training - Step 13601: {'lr': 0.00043133232918926426, 'samples': 6964224, 'steps': 13601, 'loss/train': 2.1976442337036133} 02/24/2022 19:10:30 - INFO - codeparrot_training - Step 13602: {'lr': 0.0004313210648486864, 'samples': 6964736, 'steps': 13602, 'loss/train': 1.497009515762329} 02/24/2022 19:10:36 - INFO - codeparrot_training - Step 13603: {'lr': 0.00043130979973138664, 'samples': 6965248, 'steps': 13603, 'loss/train': 1.391070008277893} 02/24/2022 19:10:39 - INFO - codeparrot_training - Step 13604: {'lr': 0.00043129853383741334, 'samples': 6965760, 'steps': 13604, 'loss/train': 3.091587781906128} 02/24/2022 19:10:45 - INFO - codeparrot_training - Step 13605: {'lr': 0.00043128726716681464, 'samples': 6966272, 'steps': 13605, 'loss/train': 1.5055516958236694} 02/24/2022 19:10:48 - INFO - codeparrot_training - Step 13606: {'lr': 0.0004312759997196389, 'samples': 6966784, 'steps': 13606, 'loss/train': 1.4908604621887207} 02/24/2022 19:10:54 - INFO - codeparrot_training - Step 13607: {'lr': 0.00043126473149593424, 'samples': 6967296, 'steps': 13607, 'loss/train': 1.7723292112350464} 02/24/2022 19:10:57 - INFO - codeparrot_training - Step 13608: {'lr': 0.00043125346249574915, 'samples': 6967808, 'steps': 13608, 'loss/train': 2.4015259742736816} 02/24/2022 19:11:03 - INFO - codeparrot_training - Step 13609: {'lr': 0.0004312421927191318, 'samples': 6968320, 'steps': 13609, 'loss/train': 2.184799909591675} 02/24/2022 19:11:07 - INFO - codeparrot_training - Step 13610: {'lr': 0.00043123092216613035, 'samples': 6968832, 'steps': 13610, 'loss/train': 2.1721858978271484} 02/24/2022 19:11:12 - INFO - codeparrot_training - Step 13611: {'lr': 0.0004312196508367932, 'samples': 6969344, 'steps': 13611, 'loss/train': 2.9096438884735107} 02/24/2022 19:11:16 - INFO - codeparrot_training - Step 13612: {'lr': 0.0004312083787311686, 'samples': 6969856, 'steps': 13612, 'loss/train': 2.17563533782959} 02/24/2022 19:11:21 - INFO - codeparrot_training - Step 13613: {'lr': 0.0004311971058493049, 'samples': 6970368, 'steps': 13613, 'loss/train': 1.903287649154663} 02/24/2022 19:11:25 - INFO - codeparrot_training - Step 13614: {'lr': 0.0004311858321912503, 'samples': 6970880, 'steps': 13614, 'loss/train': 1.232810616493225} 02/24/2022 19:11:31 - INFO - codeparrot_training - Step 13615: {'lr': 0.0004311745577570531, 'samples': 6971392, 'steps': 13615, 'loss/train': 0.4791630208492279} 02/24/2022 19:11:34 - INFO - codeparrot_training - Step 13616: {'lr': 0.0004311632825467617, 'samples': 6971904, 'steps': 13616, 'loss/train': 2.1557745933532715} 02/24/2022 19:11:39 - INFO - codeparrot_training - Step 13617: {'lr': 0.00043115200656042426, 'samples': 6972416, 'steps': 13617, 'loss/train': 1.3522979021072388} 02/24/2022 19:11:43 - INFO - codeparrot_training - Step 13618: {'lr': 0.00043114072979808914, 'samples': 6972928, 'steps': 13618, 'loss/train': 1.9285956621170044} 02/24/2022 19:11:48 - INFO - codeparrot_training - Step 13619: {'lr': 0.00043112945225980473, 'samples': 6973440, 'steps': 13619, 'loss/train': 1.6338764429092407} 02/24/2022 19:11:52 - INFO - codeparrot_training - Step 13620: {'lr': 0.00043111817394561917, 'samples': 6973952, 'steps': 13620, 'loss/train': 2.6576662063598633} 02/24/2022 19:11:58 - INFO - codeparrot_training - Step 13621: {'lr': 0.0004311068948555809, 'samples': 6974464, 'steps': 13621, 'loss/train': 1.338390827178955} 02/24/2022 19:12:01 - INFO - codeparrot_training - Step 13622: {'lr': 0.0004310956149897382, 'samples': 6974976, 'steps': 13622, 'loss/train': 1.3402029275894165} 02/24/2022 19:12:06 - INFO - codeparrot_training - Step 13623: {'lr': 0.00043108433434813943, 'samples': 6975488, 'steps': 13623, 'loss/train': 1.94597327709198} 02/24/2022 19:12:10 - INFO - codeparrot_training - Step 13624: {'lr': 0.00043107305293083276, 'samples': 6976000, 'steps': 13624, 'loss/train': 2.299471855163574} 02/24/2022 19:12:16 - INFO - codeparrot_training - Step 13625: {'lr': 0.0004310617707378668, 'samples': 6976512, 'steps': 13625, 'loss/train': 2.528211832046509} 02/24/2022 19:12:20 - INFO - codeparrot_training - Step 13626: {'lr': 0.0004310504877692896, 'samples': 6977024, 'steps': 13626, 'loss/train': 3.587096929550171} 02/24/2022 19:12:25 - INFO - codeparrot_training - Step 13627: {'lr': 0.00043103920402514956, 'samples': 6977536, 'steps': 13627, 'loss/train': 1.9988280534744263} 02/24/2022 19:12:29 - INFO - codeparrot_training - Step 13628: {'lr': 0.00043102791950549513, 'samples': 6978048, 'steps': 13628, 'loss/train': 2.581573963165283} 02/24/2022 19:12:34 - INFO - codeparrot_training - Step 13629: {'lr': 0.00043101663421037453, 'samples': 6978560, 'steps': 13629, 'loss/train': 2.308349132537842} 02/24/2022 19:12:38 - INFO - codeparrot_training - Step 13630: {'lr': 0.00043100534813983617, 'samples': 6979072, 'steps': 13630, 'loss/train': 1.44800865650177} 02/24/2022 19:12:43 - INFO - codeparrot_training - Step 13631: {'lr': 0.00043099406129392835, 'samples': 6979584, 'steps': 13631, 'loss/train': 2.951619863510132} 02/24/2022 19:12:47 - INFO - codeparrot_training - Step 13632: {'lr': 0.00043098277367269953, 'samples': 6980096, 'steps': 13632, 'loss/train': 2.598477363586426} 02/24/2022 19:12:52 - INFO - codeparrot_training - Step 13633: {'lr': 0.0004309714852761979, 'samples': 6980608, 'steps': 13633, 'loss/train': 1.5477696657180786} 02/24/2022 19:12:56 - INFO - codeparrot_training - Step 13634: {'lr': 0.0004309601961044719, 'samples': 6981120, 'steps': 13634, 'loss/train': 2.181486129760742} 02/24/2022 19:13:02 - INFO - codeparrot_training - Step 13635: {'lr': 0.0004309489061575699, 'samples': 6981632, 'steps': 13635, 'loss/train': 1.551332712173462} 02/24/2022 19:13:05 - INFO - codeparrot_training - Step 13636: {'lr': 0.0004309376154355402, 'samples': 6982144, 'steps': 13636, 'loss/train': 2.204392910003662} 02/24/2022 19:13:11 - INFO - codeparrot_training - Step 13637: {'lr': 0.00043092632393843124, 'samples': 6982656, 'steps': 13637, 'loss/train': 1.9445143938064575} 02/24/2022 19:13:14 - INFO - codeparrot_training - Step 13638: {'lr': 0.00043091503166629136, 'samples': 6983168, 'steps': 13638, 'loss/train': 2.391655921936035} 02/24/2022 19:13:20 - INFO - codeparrot_training - Step 13639: {'lr': 0.000430903738619169, 'samples': 6983680, 'steps': 13639, 'loss/train': 1.3919063806533813} 02/24/2022 19:13:23 - INFO - codeparrot_training - Step 13640: {'lr': 0.00043089244479711233, 'samples': 6984192, 'steps': 13640, 'loss/train': 2.4541327953338623} 02/24/2022 19:13:29 - INFO - codeparrot_training - Step 13641: {'lr': 0.00043088115020016994, 'samples': 6984704, 'steps': 13641, 'loss/train': 1.7544317245483398} 02/24/2022 19:13:34 - INFO - codeparrot_training - Step 13642: {'lr': 0.00043086985482839016, 'samples': 6985216, 'steps': 13642, 'loss/train': 1.6198045015335083} 02/24/2022 19:13:38 - INFO - codeparrot_training - Step 13643: {'lr': 0.00043085855868182135, 'samples': 6985728, 'steps': 13643, 'loss/train': 2.3710861206054688} 02/24/2022 19:13:45 - INFO - codeparrot_training - Step 13644: {'lr': 0.0004308472617605118, 'samples': 6986240, 'steps': 13644, 'loss/train': 1.6309189796447754} 02/24/2022 19:13:48 - INFO - codeparrot_training - Step 13645: {'lr': 0.00043083596406451015, 'samples': 6986752, 'steps': 13645, 'loss/train': 0.1144043579697609} 02/24/2022 19:13:54 - INFO - codeparrot_training - Step 13646: {'lr': 0.0004308246655938646, 'samples': 6987264, 'steps': 13646, 'loss/train': 1.7623697519302368} 02/24/2022 19:13:57 - INFO - codeparrot_training - Step 13647: {'lr': 0.0004308133663486236, 'samples': 6987776, 'steps': 13647, 'loss/train': 2.623563289642334} 02/24/2022 19:14:03 - INFO - codeparrot_training - Step 13648: {'lr': 0.00043080206632883553, 'samples': 6988288, 'steps': 13648, 'loss/train': 1.596634864807129} 02/24/2022 19:14:06 - INFO - codeparrot_training - Step 13649: {'lr': 0.0004307907655345488, 'samples': 6988800, 'steps': 13649, 'loss/train': 2.036360502243042} 02/24/2022 19:14:12 - INFO - codeparrot_training - Step 13650: {'lr': 0.0004307794639658119, 'samples': 6989312, 'steps': 13650, 'loss/train': 1.4130501747131348} 02/24/2022 19:14:15 - INFO - codeparrot_training - Step 13651: {'lr': 0.0004307681616226732, 'samples': 6989824, 'steps': 13651, 'loss/train': 1.5663180351257324} 02/24/2022 19:14:21 - INFO - codeparrot_training - Step 13652: {'lr': 0.000430756858505181, 'samples': 6990336, 'steps': 13652, 'loss/train': 2.8578941822052} 02/24/2022 19:14:24 - INFO - codeparrot_training - Step 13653: {'lr': 0.0004307455546133838, 'samples': 6990848, 'steps': 13653, 'loss/train': 0.7819316387176514} 02/24/2022 19:14:30 - INFO - codeparrot_training - Step 13654: {'lr': 0.00043073424994733014, 'samples': 6991360, 'steps': 13654, 'loss/train': 1.5253733396530151} 02/24/2022 19:14:33 - INFO - codeparrot_training - Step 13655: {'lr': 0.0004307229445070683, 'samples': 6991872, 'steps': 13655, 'loss/train': 2.6275267601013184} 02/24/2022 19:14:40 - INFO - codeparrot_training - Step 13656: {'lr': 0.0004307116382926468, 'samples': 6992384, 'steps': 13656, 'loss/train': 2.8324756622314453} 02/24/2022 19:14:44 - INFO - codeparrot_training - Step 13657: {'lr': 0.0004307003313041139, 'samples': 6992896, 'steps': 13657, 'loss/train': 1.418109655380249} 02/24/2022 19:14:49 - INFO - codeparrot_training - Step 13658: {'lr': 0.0004306890235415183, 'samples': 6993408, 'steps': 13658, 'loss/train': 2.501063346862793} 02/24/2022 19:14:53 - INFO - codeparrot_training - Step 13659: {'lr': 0.0004306777150049082, 'samples': 6993920, 'steps': 13659, 'loss/train': 1.6798590421676636} 02/24/2022 19:14:58 - INFO - codeparrot_training - Step 13660: {'lr': 0.0004306664056943322, 'samples': 6994432, 'steps': 13660, 'loss/train': 8.496380805969238} 02/24/2022 19:15:02 - INFO - codeparrot_training - Step 13661: {'lr': 0.0004306550956098386, 'samples': 6994944, 'steps': 13661, 'loss/train': 1.7745139598846436} 02/24/2022 19:15:07 - INFO - codeparrot_training - Step 13662: {'lr': 0.000430643784751476, 'samples': 6995456, 'steps': 13662, 'loss/train': 2.3376801013946533} 02/24/2022 19:15:11 - INFO - codeparrot_training - Step 13663: {'lr': 0.0004306324731192929, 'samples': 6995968, 'steps': 13663, 'loss/train': 1.2211185693740845} 02/24/2022 19:15:16 - INFO - codeparrot_training - Step 13664: {'lr': 0.00043062116071333745, 'samples': 6996480, 'steps': 13664, 'loss/train': 3.09124755859375} 02/24/2022 19:15:20 - INFO - codeparrot_training - Step 13665: {'lr': 0.0004306098475336584, 'samples': 6996992, 'steps': 13665, 'loss/train': 2.5851945877075195} 02/24/2022 19:15:26 - INFO - codeparrot_training - Step 13666: {'lr': 0.0004305985335803041, 'samples': 6997504, 'steps': 13666, 'loss/train': 1.8962674140930176} 02/24/2022 19:15:29 - INFO - codeparrot_training - Step 13667: {'lr': 0.000430587218853323, 'samples': 6998016, 'steps': 13667, 'loss/train': 2.4098546504974365} 02/24/2022 19:15:35 - INFO - codeparrot_training - Step 13668: {'lr': 0.0004305759033527636, 'samples': 6998528, 'steps': 13668, 'loss/train': 1.8245320320129395} 02/24/2022 19:15:38 - INFO - codeparrot_training - Step 13669: {'lr': 0.0004305645870786744, 'samples': 6999040, 'steps': 13669, 'loss/train': 2.325986862182617} 02/24/2022 19:15:44 - INFO - codeparrot_training - Step 13670: {'lr': 0.00043055327003110384, 'samples': 6999552, 'steps': 13670, 'loss/train': 1.0751984119415283} 02/24/2022 19:15:47 - INFO - codeparrot_training - Step 13671: {'lr': 0.00043054195221010037, 'samples': 7000064, 'steps': 13671, 'loss/train': 2.629345655441284} 02/24/2022 19:15:53 - INFO - codeparrot_training - Step 13672: {'lr': 0.00043053063361571256, 'samples': 7000576, 'steps': 13672, 'loss/train': 2.271709680557251} 02/24/2022 19:15:56 - INFO - codeparrot_training - Step 13673: {'lr': 0.0004305193142479888, 'samples': 7001088, 'steps': 13673, 'loss/train': 1.5242174863815308} 02/24/2022 19:16:02 - INFO - codeparrot_training - Step 13674: {'lr': 0.0004305079941069776, 'samples': 7001600, 'steps': 13674, 'loss/train': 2.424975633621216} 02/24/2022 19:16:05 - INFO - codeparrot_training - Step 13675: {'lr': 0.0004304966731927276, 'samples': 7002112, 'steps': 13675, 'loss/train': 2.5726494789123535} 02/24/2022 19:16:11 - INFO - codeparrot_training - Step 13676: {'lr': 0.000430485351505287, 'samples': 7002624, 'steps': 13676, 'loss/train': 1.4505350589752197} 02/24/2022 19:16:14 - INFO - codeparrot_training - Step 13677: {'lr': 0.00043047402904470455, 'samples': 7003136, 'steps': 13677, 'loss/train': 1.4280500411987305} 02/24/2022 19:16:19 - INFO - codeparrot_training - Step 13678: {'lr': 0.00043046270581102865, 'samples': 7003648, 'steps': 13678, 'loss/train': 1.0448979139328003} 02/24/2022 19:16:23 - INFO - codeparrot_training - Step 13679: {'lr': 0.00043045138180430783, 'samples': 7004160, 'steps': 13679, 'loss/train': 0.19420510530471802} 02/24/2022 19:16:28 - INFO - codeparrot_training - Step 13680: {'lr': 0.00043044005702459054, 'samples': 7004672, 'steps': 13680, 'loss/train': 2.1825525760650635} 02/24/2022 19:16:32 - INFO - codeparrot_training - Step 13681: {'lr': 0.0004304287314719254, 'samples': 7005184, 'steps': 13681, 'loss/train': 1.7918729782104492} 02/24/2022 19:16:38 - INFO - codeparrot_training - Step 13682: {'lr': 0.00043041740514636085, 'samples': 7005696, 'steps': 13682, 'loss/train': 3.2685329914093018} 02/24/2022 19:16:42 - INFO - codeparrot_training - Step 13683: {'lr': 0.0004304060780479454, 'samples': 7006208, 'steps': 13683, 'loss/train': 2.373638153076172} 02/24/2022 19:16:47 - INFO - codeparrot_training - Step 13684: {'lr': 0.0004303947501767276, 'samples': 7006720, 'steps': 13684, 'loss/train': 1.5358431339263916} 02/24/2022 19:16:51 - INFO - codeparrot_training - Step 13685: {'lr': 0.0004303834215327561, 'samples': 7007232, 'steps': 13685, 'loss/train': 1.933910608291626} 02/24/2022 19:16:56 - INFO - codeparrot_training - Step 13686: {'lr': 0.00043037209211607913, 'samples': 7007744, 'steps': 13686, 'loss/train': 1.9050073623657227} 02/24/2022 19:17:00 - INFO - codeparrot_training - Step 13687: {'lr': 0.00043036076192674546, 'samples': 7008256, 'steps': 13687, 'loss/train': 2.118858575820923} 02/24/2022 19:17:06 - INFO - codeparrot_training - Step 13688: {'lr': 0.00043034943096480357, 'samples': 7008768, 'steps': 13688, 'loss/train': 1.9457725286483765} 02/24/2022 19:17:09 - INFO - codeparrot_training - Step 13689: {'lr': 0.000430338099230302, 'samples': 7009280, 'steps': 13689, 'loss/train': 2.4389522075653076} 02/24/2022 19:17:15 - INFO - codeparrot_training - Step 13690: {'lr': 0.00043032676672328916, 'samples': 7009792, 'steps': 13690, 'loss/train': 1.885743260383606} 02/24/2022 19:17:18 - INFO - codeparrot_training - Step 13691: {'lr': 0.00043031543344381384, 'samples': 7010304, 'steps': 13691, 'loss/train': 1.9437928199768066} 02/24/2022 19:17:25 - INFO - codeparrot_training - Step 13692: {'lr': 0.0004303040993919244, 'samples': 7010816, 'steps': 13692, 'loss/train': 2.450385808944702} 02/24/2022 19:17:28 - INFO - codeparrot_training - Step 13693: {'lr': 0.00043029276456766946, 'samples': 7011328, 'steps': 13693, 'loss/train': 1.8689649105072021} 02/24/2022 19:17:34 - INFO - codeparrot_training - Step 13694: {'lr': 0.00043028142897109754, 'samples': 7011840, 'steps': 13694, 'loss/train': 1.6598165035247803} 02/24/2022 19:17:37 - INFO - codeparrot_training - Step 13695: {'lr': 0.0004302700926022573, 'samples': 7012352, 'steps': 13695, 'loss/train': 1.3444586992263794} 02/24/2022 19:17:43 - INFO - codeparrot_training - Step 13696: {'lr': 0.0004302587554611972, 'samples': 7012864, 'steps': 13696, 'loss/train': 2.2317521572113037} 02/24/2022 19:17:46 - INFO - codeparrot_training - Step 13697: {'lr': 0.0004302474175479658, 'samples': 7013376, 'steps': 13697, 'loss/train': 2.0625648498535156} 02/24/2022 19:17:52 - INFO - codeparrot_training - Step 13698: {'lr': 0.0004302360788626117, 'samples': 7013888, 'steps': 13698, 'loss/train': 1.4547176361083984} 02/24/2022 19:17:57 - INFO - codeparrot_training - Step 13699: {'lr': 0.00043022473940518345, 'samples': 7014400, 'steps': 13699, 'loss/train': 1.1368025541305542} 02/24/2022 19:18:01 - INFO - codeparrot_training - Step 13700: {'lr': 0.0004302133991757297, 'samples': 7014912, 'steps': 13700, 'loss/train': 2.170872449874878} 02/24/2022 19:18:07 - INFO - codeparrot_training - Step 13701: {'lr': 0.00043020205817429895, 'samples': 7015424, 'steps': 13701, 'loss/train': 1.2136664390563965} 02/24/2022 19:18:10 - INFO - codeparrot_training - Step 13702: {'lr': 0.0004301907164009398, 'samples': 7015936, 'steps': 13702, 'loss/train': 1.1966272592544556} 02/24/2022 19:18:14 - INFO - codeparrot_training - Step 13703: {'lr': 0.00043017937385570083, 'samples': 7016448, 'steps': 13703, 'loss/train': 2.951408624649048} 02/24/2022 19:18:20 - INFO - codeparrot_training - Step 13704: {'lr': 0.00043016803053863063, 'samples': 7016960, 'steps': 13704, 'loss/train': 1.6956672668457031} 02/24/2022 19:18:23 - INFO - codeparrot_training - Step 13705: {'lr': 0.00043015668644977783, 'samples': 7017472, 'steps': 13705, 'loss/train': 2.078306198120117} 02/24/2022 19:18:29 - INFO - codeparrot_training - Step 13706: {'lr': 0.000430145341589191, 'samples': 7017984, 'steps': 13706, 'loss/train': 1.5965362787246704} 02/24/2022 19:18:32 - INFO - codeparrot_training - Step 13707: {'lr': 0.0004301339959569187, 'samples': 7018496, 'steps': 13707, 'loss/train': 1.8603307008743286} 02/24/2022 19:18:38 - INFO - codeparrot_training - Step 13708: {'lr': 0.00043012264955300954, 'samples': 7019008, 'steps': 13708, 'loss/train': 1.7873365879058838} 02/24/2022 19:18:41 - INFO - codeparrot_training - Step 13709: {'lr': 0.0004301113023775122, 'samples': 7019520, 'steps': 13709, 'loss/train': 1.5965356826782227} 02/24/2022 19:18:47 - INFO - codeparrot_training - Step 13710: {'lr': 0.00043009995443047517, 'samples': 7020032, 'steps': 13710, 'loss/train': 1.4226073026657104} 02/24/2022 19:18:51 - INFO - codeparrot_training - Step 13711: {'lr': 0.0004300886057119472, 'samples': 7020544, 'steps': 13711, 'loss/train': 2.959944725036621} 02/24/2022 19:18:56 - INFO - codeparrot_training - Step 13712: {'lr': 0.00043007725622197675, 'samples': 7021056, 'steps': 13712, 'loss/train': 1.8905389308929443} 02/24/2022 19:19:00 - INFO - codeparrot_training - Step 13713: {'lr': 0.00043006590596061256, 'samples': 7021568, 'steps': 13713, 'loss/train': 1.7996097803115845} 02/24/2022 19:19:05 - INFO - codeparrot_training - Step 13714: {'lr': 0.0004300545549279032, 'samples': 7022080, 'steps': 13714, 'loss/train': 2.627183437347412} 02/24/2022 19:19:09 - INFO - codeparrot_training - Step 13715: {'lr': 0.0004300432031238973, 'samples': 7022592, 'steps': 13715, 'loss/train': 1.592428207397461} 02/24/2022 19:19:14 - INFO - codeparrot_training - Step 13716: {'lr': 0.00043003185054864344, 'samples': 7023104, 'steps': 13716, 'loss/train': 1.9637314081192017} 02/24/2022 19:19:18 - INFO - codeparrot_training - Step 13717: {'lr': 0.0004300204972021903, 'samples': 7023616, 'steps': 13717, 'loss/train': 2.055371046066284} 02/24/2022 19:19:24 - INFO - codeparrot_training - Step 13718: {'lr': 0.00043000914308458663, 'samples': 7024128, 'steps': 13718, 'loss/train': 2.8856101036071777} 02/24/2022 19:19:27 - INFO - codeparrot_training - Step 13719: {'lr': 0.0004299977881958808, 'samples': 7024640, 'steps': 13719, 'loss/train': 2.062641143798828} 02/24/2022 19:19:33 - INFO - codeparrot_training - Step 13720: {'lr': 0.0004299864325361217, 'samples': 7025152, 'steps': 13720, 'loss/train': 2.971893548965454} 02/24/2022 19:19:36 - INFO - codeparrot_training - Step 13721: {'lr': 0.00042997507610535783, 'samples': 7025664, 'steps': 13721, 'loss/train': 2.027390480041504} 02/24/2022 19:19:42 - INFO - codeparrot_training - Step 13722: {'lr': 0.00042996371890363796, 'samples': 7026176, 'steps': 13722, 'loss/train': 0.9466265439987183} 02/24/2022 19:19:46 - INFO - codeparrot_training - Step 13723: {'lr': 0.00042995236093101055, 'samples': 7026688, 'steps': 13723, 'loss/train': 2.238375663757324} 02/24/2022 19:19:51 - INFO - codeparrot_training - Step 13724: {'lr': 0.0004299410021875244, 'samples': 7027200, 'steps': 13724, 'loss/train': 1.4154530763626099} 02/24/2022 19:19:55 - INFO - codeparrot_training - Step 13725: {'lr': 0.00042992964267322823, 'samples': 7027712, 'steps': 13725, 'loss/train': 1.372147560119629} 02/24/2022 19:20:01 - INFO - codeparrot_training - Step 13726: {'lr': 0.00042991828238817046, 'samples': 7028224, 'steps': 13726, 'loss/train': 2.961617946624756} 02/24/2022 19:20:04 - INFO - codeparrot_training - Step 13727: {'lr': 0.0004299069213324, 'samples': 7028736, 'steps': 13727, 'loss/train': 1.9149107933044434} 02/24/2022 19:20:08 - INFO - codeparrot_training - Step 13728: {'lr': 0.0004298955595059654, 'samples': 7029248, 'steps': 13728, 'loss/train': 1.019468903541565} 02/24/2022 19:20:15 - INFO - codeparrot_training - Step 13729: {'lr': 0.00042988419690891534, 'samples': 7029760, 'steps': 13729, 'loss/train': 2.0500237941741943} 02/24/2022 19:20:18 - INFO - codeparrot_training - Step 13730: {'lr': 0.00042987283354129846, 'samples': 7030272, 'steps': 13730, 'loss/train': 1.713131070137024} 02/24/2022 19:20:24 - INFO - codeparrot_training - Step 13731: {'lr': 0.0004298614694031635, 'samples': 7030784, 'steps': 13731, 'loss/train': 3.7322075366973877} 02/24/2022 19:20:29 - INFO - codeparrot_training - Step 13732: {'lr': 0.0004298501044945591, 'samples': 7031296, 'steps': 13732, 'loss/train': 1.6995245218276978} 02/24/2022 19:20:33 - INFO - codeparrot_training - Step 13733: {'lr': 0.000429838738815534, 'samples': 7031808, 'steps': 13733, 'loss/train': 2.088724374771118} 02/24/2022 19:20:38 - INFO - codeparrot_training - Step 13734: {'lr': 0.00042982737236613687, 'samples': 7032320, 'steps': 13734, 'loss/train': 2.341543674468994} 02/24/2022 19:20:42 - INFO - codeparrot_training - Step 13735: {'lr': 0.00042981600514641635, 'samples': 7032832, 'steps': 13735, 'loss/train': 2.0815110206604004} 02/24/2022 19:20:45 - INFO - codeparrot_training - Step 13736: {'lr': 0.00042980463715642115, 'samples': 7033344, 'steps': 13736, 'loss/train': 2.4765756130218506} 02/24/2022 19:20:51 - INFO - codeparrot_training - Step 13737: {'lr': 0.0004297932683962, 'samples': 7033856, 'steps': 13737, 'loss/train': 2.6112093925476074} 02/24/2022 19:20:54 - INFO - codeparrot_training - Step 13738: {'lr': 0.00042978189886580157, 'samples': 7034368, 'steps': 13738, 'loss/train': 1.3575396537780762} 02/24/2022 19:21:00 - INFO - codeparrot_training - Step 13739: {'lr': 0.00042977052856527456, 'samples': 7034880, 'steps': 13739, 'loss/train': 1.7285126447677612} 02/24/2022 19:21:04 - INFO - codeparrot_training - Step 13740: {'lr': 0.00042975915749466763, 'samples': 7035392, 'steps': 13740, 'loss/train': 2.420454740524292} 02/24/2022 19:21:09 - INFO - codeparrot_training - Step 13741: {'lr': 0.0004297477856540296, 'samples': 7035904, 'steps': 13741, 'loss/train': 2.9618031978607178} 02/24/2022 19:21:13 - INFO - codeparrot_training - Step 13742: {'lr': 0.00042973641304340916, 'samples': 7036416, 'steps': 13742, 'loss/train': 2.3210268020629883} 02/24/2022 19:21:18 - INFO - codeparrot_training - Step 13743: {'lr': 0.00042972503966285503, 'samples': 7036928, 'steps': 13743, 'loss/train': 2.0137109756469727} 02/24/2022 19:21:24 - INFO - codeparrot_training - Step 13744: {'lr': 0.00042971366551241587, 'samples': 7037440, 'steps': 13744, 'loss/train': 1.216625452041626} 02/24/2022 19:21:27 - INFO - codeparrot_training - Step 13745: {'lr': 0.00042970229059214037, 'samples': 7037952, 'steps': 13745, 'loss/train': 1.7579693794250488} 02/24/2022 19:21:33 - INFO - codeparrot_training - Step 13746: {'lr': 0.0004296909149020774, 'samples': 7038464, 'steps': 13746, 'loss/train': 2.274695634841919} 02/24/2022 19:21:36 - INFO - codeparrot_training - Step 13747: {'lr': 0.0004296795384422756, 'samples': 7038976, 'steps': 13747, 'loss/train': 2.0781469345092773} 02/24/2022 19:21:42 - INFO - codeparrot_training - Step 13748: {'lr': 0.00042966816121278365, 'samples': 7039488, 'steps': 13748, 'loss/train': 2.7990827560424805} 02/24/2022 19:21:46 - INFO - codeparrot_training - Step 13749: {'lr': 0.00042965678321365045, 'samples': 7040000, 'steps': 13749, 'loss/train': 2.105743885040283} 02/24/2022 19:21:51 - INFO - codeparrot_training - Step 13750: {'lr': 0.00042964540444492453, 'samples': 7040512, 'steps': 13750, 'loss/train': 1.5459169149398804} 02/24/2022 19:21:55 - INFO - codeparrot_training - Step 13751: {'lr': 0.00042963402490665484, 'samples': 7041024, 'steps': 13751, 'loss/train': 2.268190383911133} 02/24/2022 19:22:01 - INFO - codeparrot_training - Step 13752: {'lr': 0.0004296226445988899, 'samples': 7041536, 'steps': 13752, 'loss/train': 2.209700345993042} 02/24/2022 19:22:04 - INFO - codeparrot_training - Step 13753: {'lr': 0.0004296112635216787, 'samples': 7042048, 'steps': 13753, 'loss/train': 1.8361694812774658} 02/24/2022 19:22:10 - INFO - codeparrot_training - Step 13754: {'lr': 0.00042959988167506983, 'samples': 7042560, 'steps': 13754, 'loss/train': 1.5439350605010986} 02/24/2022 19:22:13 - INFO - codeparrot_training - Step 13755: {'lr': 0.00042958849905911213, 'samples': 7043072, 'steps': 13755, 'loss/train': 1.547868013381958} 02/24/2022 19:22:19 - INFO - codeparrot_training - Step 13756: {'lr': 0.0004295771156738543, 'samples': 7043584, 'steps': 13756, 'loss/train': 0.36390218138694763} 02/24/2022 19:22:22 - INFO - codeparrot_training - Step 13757: {'lr': 0.00042956573151934507, 'samples': 7044096, 'steps': 13757, 'loss/train': 0.7468468546867371} 02/24/2022 19:22:28 - INFO - codeparrot_training - Step 13758: {'lr': 0.00042955434659563334, 'samples': 7044608, 'steps': 13758, 'loss/train': 1.6567022800445557} 02/24/2022 19:22:31 - INFO - codeparrot_training - Step 13759: {'lr': 0.00042954296090276777, 'samples': 7045120, 'steps': 13759, 'loss/train': 0.9269330501556396} 02/24/2022 19:22:36 - INFO - codeparrot_training - Step 13760: {'lr': 0.0004295315744407972, 'samples': 7045632, 'steps': 13760, 'loss/train': 1.100716471672058} 02/24/2022 19:22:40 - INFO - codeparrot_training - Step 13761: {'lr': 0.0004295201872097704, 'samples': 7046144, 'steps': 13761, 'loss/train': 1.6133983135223389} 02/24/2022 19:22:45 - INFO - codeparrot_training - Step 13762: {'lr': 0.0004295087992097361, 'samples': 7046656, 'steps': 13762, 'loss/train': 2.823902130126953} 02/24/2022 19:22:49 - INFO - codeparrot_training - Step 13763: {'lr': 0.00042949741044074306, 'samples': 7047168, 'steps': 13763, 'loss/train': 1.0988500118255615} 02/24/2022 19:22:55 - INFO - codeparrot_training - Step 13764: {'lr': 0.00042948602090284014, 'samples': 7047680, 'steps': 13764, 'loss/train': 1.7745718955993652} 02/24/2022 19:22:59 - INFO - codeparrot_training - Step 13765: {'lr': 0.00042947463059607606, 'samples': 7048192, 'steps': 13765, 'loss/train': 1.2690480947494507} 02/24/2022 19:23:04 - INFO - codeparrot_training - Step 13766: {'lr': 0.0004294632395204997, 'samples': 7048704, 'steps': 13766, 'loss/train': 2.875951051712036} 02/24/2022 19:23:08 - INFO - codeparrot_training - Step 13767: {'lr': 0.0004294518476761598, 'samples': 7049216, 'steps': 13767, 'loss/train': 0.4162357747554779} 02/24/2022 19:23:13 - INFO - codeparrot_training - Step 13768: {'lr': 0.00042944045506310515, 'samples': 7049728, 'steps': 13768, 'loss/train': 2.7400615215301514} 02/24/2022 19:23:17 - INFO - codeparrot_training - Step 13769: {'lr': 0.0004294290616813846, 'samples': 7050240, 'steps': 13769, 'loss/train': 1.1161319017410278} 02/24/2022 19:23:22 - INFO - codeparrot_training - Step 13770: {'lr': 0.00042941766753104696, 'samples': 7050752, 'steps': 13770, 'loss/train': 2.0202841758728027} 02/24/2022 19:23:26 - INFO - codeparrot_training - Step 13771: {'lr': 0.00042940627261214094, 'samples': 7051264, 'steps': 13771, 'loss/train': 2.79789137840271} 02/24/2022 19:23:31 - INFO - codeparrot_training - Step 13772: {'lr': 0.00042939487692471534, 'samples': 7051776, 'steps': 13772, 'loss/train': 2.5752716064453125} 02/24/2022 19:23:35 - INFO - codeparrot_training - Step 13773: {'lr': 0.0004293834804688192, 'samples': 7052288, 'steps': 13773, 'loss/train': 2.0483415126800537} 02/24/2022 19:23:41 - INFO - codeparrot_training - Step 13774: {'lr': 0.00042937208324450116, 'samples': 7052800, 'steps': 13774, 'loss/train': 9.001176834106445} 02/24/2022 19:23:44 - INFO - codeparrot_training - Step 13775: {'lr': 0.00042936068525181004, 'samples': 7053312, 'steps': 13775, 'loss/train': 1.8988256454467773} 02/24/2022 19:23:50 - INFO - codeparrot_training - Step 13776: {'lr': 0.00042934928649079467, 'samples': 7053824, 'steps': 13776, 'loss/train': 0.8681538105010986} 02/24/2022 19:23:53 - INFO - codeparrot_training - Step 13777: {'lr': 0.0004293378869615039, 'samples': 7054336, 'steps': 13777, 'loss/train': 1.0394240617752075} 02/24/2022 19:23:59 - INFO - codeparrot_training - Step 13778: {'lr': 0.00042932648666398667, 'samples': 7054848, 'steps': 13778, 'loss/train': 0.8579873442649841} 02/24/2022 19:24:02 - INFO - codeparrot_training - Step 13779: {'lr': 0.0004293150855982916, 'samples': 7055360, 'steps': 13779, 'loss/train': 1.411608099937439} 02/24/2022 19:24:08 - INFO - codeparrot_training - Step 13780: {'lr': 0.0004293036837644677, 'samples': 7055872, 'steps': 13780, 'loss/train': 1.1304799318313599} 02/24/2022 19:24:11 - INFO - codeparrot_training - Step 13781: {'lr': 0.0004292922811625637, 'samples': 7056384, 'steps': 13781, 'loss/train': 1.9228967428207397} 02/24/2022 19:24:17 - INFO - codeparrot_training - Step 13782: {'lr': 0.0004292808777926286, 'samples': 7056896, 'steps': 13782, 'loss/train': 1.3900909423828125} 02/24/2022 19:24:20 - INFO - codeparrot_training - Step 13783: {'lr': 0.0004292694736547111, 'samples': 7057408, 'steps': 13783, 'loss/train': 2.519258737564087} 02/24/2022 19:24:26 - INFO - codeparrot_training - Step 13784: {'lr': 0.0004292580687488601, 'samples': 7057920, 'steps': 13784, 'loss/train': 2.419724225997925} 02/24/2022 19:24:30 - INFO - codeparrot_training - Step 13785: {'lr': 0.00042924666307512437, 'samples': 7058432, 'steps': 13785, 'loss/train': 2.781606674194336} 02/24/2022 19:24:36 - INFO - codeparrot_training - Step 13786: {'lr': 0.000429235256633553, 'samples': 7058944, 'steps': 13786, 'loss/train': 2.4013800621032715} 02/24/2022 19:24:39 - INFO - codeparrot_training - Step 13787: {'lr': 0.0004292238494241946, 'samples': 7059456, 'steps': 13787, 'loss/train': 2.3291451930999756} 02/24/2022 19:24:45 - INFO - codeparrot_training - Step 13788: {'lr': 0.00042921244144709817, 'samples': 7059968, 'steps': 13788, 'loss/train': 2.263153314590454} 02/24/2022 19:24:48 - INFO - codeparrot_training - Step 13789: {'lr': 0.0004292010327023125, 'samples': 7060480, 'steps': 13789, 'loss/train': 1.587754487991333} 02/24/2022 19:24:54 - INFO - codeparrot_training - Step 13790: {'lr': 0.00042918962318988664, 'samples': 7060992, 'steps': 13790, 'loss/train': 1.3965115547180176} 02/24/2022 19:24:57 - INFO - codeparrot_training - Step 13791: {'lr': 0.00042917821290986926, 'samples': 7061504, 'steps': 13791, 'loss/train': 2.5567522048950195} 02/24/2022 19:25:03 - INFO - codeparrot_training - Step 13792: {'lr': 0.0004291668018623093, 'samples': 7062016, 'steps': 13792, 'loss/train': 2.453965902328491} 02/24/2022 19:25:06 - INFO - codeparrot_training - Step 13793: {'lr': 0.00042915539004725564, 'samples': 7062528, 'steps': 13793, 'loss/train': 1.6210194826126099} 02/24/2022 19:25:12 - INFO - codeparrot_training - Step 13794: {'lr': 0.0004291439774647572, 'samples': 7063040, 'steps': 13794, 'loss/train': 2.0600123405456543} 02/24/2022 19:25:16 - INFO - codeparrot_training - Step 13795: {'lr': 0.00042913256411486277, 'samples': 7063552, 'steps': 13795, 'loss/train': 2.1193196773529053} 02/24/2022 19:25:21 - INFO - codeparrot_training - Step 13796: {'lr': 0.0004291211499976214, 'samples': 7064064, 'steps': 13796, 'loss/train': 2.3493311405181885} 02/24/2022 19:25:25 - INFO - codeparrot_training - Step 13797: {'lr': 0.00042910973511308195, 'samples': 7064576, 'steps': 13797, 'loss/train': 1.6887013912200928} 02/24/2022 19:25:30 - INFO - codeparrot_training - Step 13798: {'lr': 0.0004290983194612932, 'samples': 7065088, 'steps': 13798, 'loss/train': 1.3335912227630615} 02/24/2022 19:25:34 - INFO - codeparrot_training - Step 13799: {'lr': 0.00042908690304230415, 'samples': 7065600, 'steps': 13799, 'loss/train': 1.6617976427078247} 02/24/2022 19:25:40 - INFO - codeparrot_training - Step 13800: {'lr': 0.00042907548585616363, 'samples': 7066112, 'steps': 13800, 'loss/train': 2.6258771419525146} 02/24/2022 19:25:43 - INFO - codeparrot_training - Step 13801: {'lr': 0.00042906406790292053, 'samples': 7066624, 'steps': 13801, 'loss/train': 2.3570327758789062} 02/24/2022 19:25:49 - INFO - codeparrot_training - Step 13802: {'lr': 0.00042905264918262386, 'samples': 7067136, 'steps': 13802, 'loss/train': 1.6215236186981201} 02/24/2022 19:25:52 - INFO - codeparrot_training - Step 13803: {'lr': 0.00042904122969532256, 'samples': 7067648, 'steps': 13803, 'loss/train': 1.531136393547058} 02/24/2022 19:25:58 - INFO - codeparrot_training - Step 13804: {'lr': 0.0004290298094410655, 'samples': 7068160, 'steps': 13804, 'loss/train': 1.1906249523162842} 02/24/2022 19:26:01 - INFO - codeparrot_training - Step 13805: {'lr': 0.0004290183884199015, 'samples': 7068672, 'steps': 13805, 'loss/train': 1.763102412223816} 02/24/2022 19:26:07 - INFO - codeparrot_training - Step 13806: {'lr': 0.00042900696663187963, 'samples': 7069184, 'steps': 13806, 'loss/train': 1.5061942338943481} 02/24/2022 19:26:10 - INFO - codeparrot_training - Step 13807: {'lr': 0.00042899554407704876, 'samples': 7069696, 'steps': 13807, 'loss/train': 2.170198440551758} 02/24/2022 19:26:16 - INFO - codeparrot_training - Step 13808: {'lr': 0.0004289841207554578, 'samples': 7070208, 'steps': 13808, 'loss/train': 2.469888210296631} 02/24/2022 19:26:19 - INFO - codeparrot_training - Step 13809: {'lr': 0.0004289726966671557, 'samples': 7070720, 'steps': 13809, 'loss/train': 2.30928897857666} 02/24/2022 19:26:26 - INFO - codeparrot_training - Step 13810: {'lr': 0.00042896127181219135, 'samples': 7071232, 'steps': 13810, 'loss/train': 2.186579704284668} 02/24/2022 19:26:29 - INFO - codeparrot_training - Step 13811: {'lr': 0.0004289498461906138, 'samples': 7071744, 'steps': 13811, 'loss/train': 1.1831737756729126} 02/24/2022 19:26:34 - INFO - codeparrot_training - Step 13812: {'lr': 0.00042893841980247194, 'samples': 7072256, 'steps': 13812, 'loss/train': 1.0793462991714478} 02/24/2022 19:26:38 - INFO - codeparrot_training - Step 13813: {'lr': 0.00042892699264781463, 'samples': 7072768, 'steps': 13813, 'loss/train': 2.049481153488159} 02/24/2022 19:26:43 - INFO - codeparrot_training - Step 13814: {'lr': 0.000428915564726691, 'samples': 7073280, 'steps': 13814, 'loss/train': 2.2335526943206787} 02/24/2022 19:26:47 - INFO - codeparrot_training - Step 13815: {'lr': 0.0004289041360391499, 'samples': 7073792, 'steps': 13815, 'loss/train': 2.0923101902008057} 02/24/2022 19:26:52 - INFO - codeparrot_training - Step 13816: {'lr': 0.0004288927065852402, 'samples': 7074304, 'steps': 13816, 'loss/train': 2.344900369644165} 02/24/2022 19:26:56 - INFO - codeparrot_training - Step 13817: {'lr': 0.000428881276365011, 'samples': 7074816, 'steps': 13817, 'loss/train': 0.7411916255950928} 02/24/2022 19:27:02 - INFO - codeparrot_training - Step 13818: {'lr': 0.00042886984537851124, 'samples': 7075328, 'steps': 13818, 'loss/train': 0.9422500133514404} 02/24/2022 19:27:05 - INFO - codeparrot_training - Step 13819: {'lr': 0.0004288584136257898, 'samples': 7075840, 'steps': 13819, 'loss/train': 2.4227230548858643} 02/24/2022 19:27:12 - INFO - codeparrot_training - Step 13820: {'lr': 0.00042884698110689574, 'samples': 7076352, 'steps': 13820, 'loss/train': 2.399198293685913} 02/24/2022 19:27:15 - INFO - codeparrot_training - Step 13821: {'lr': 0.000428835547821878, 'samples': 7076864, 'steps': 13821, 'loss/train': 2.7497496604919434} 02/24/2022 19:27:21 - INFO - codeparrot_training - Step 13822: {'lr': 0.00042882411377078556, 'samples': 7077376, 'steps': 13822, 'loss/train': 2.3676512241363525} 02/24/2022 19:27:24 - INFO - codeparrot_training - Step 13823: {'lr': 0.00042881267895366736, 'samples': 7077888, 'steps': 13823, 'loss/train': 1.8745065927505493} 02/24/2022 19:27:30 - INFO - codeparrot_training - Step 13824: {'lr': 0.00042880124337057253, 'samples': 7078400, 'steps': 13824, 'loss/train': 3.031111240386963} 02/24/2022 19:27:33 - INFO - codeparrot_training - Step 13825: {'lr': 0.00042878980702154985, 'samples': 7078912, 'steps': 13825, 'loss/train': 2.54958438873291} 02/24/2022 19:27:39 - INFO - codeparrot_training - Step 13826: {'lr': 0.00042877836990664844, 'samples': 7079424, 'steps': 13826, 'loss/train': 2.217719554901123} 02/24/2022 19:27:42 - INFO - codeparrot_training - Step 13827: {'lr': 0.00042876693202591724, 'samples': 7079936, 'steps': 13827, 'loss/train': 2.8435380458831787} 02/24/2022 19:27:48 - INFO - codeparrot_training - Step 13828: {'lr': 0.0004287554933794053, 'samples': 7080448, 'steps': 13828, 'loss/train': 1.2885924577713013} 02/24/2022 19:27:54 - INFO - codeparrot_training - Step 13829: {'lr': 0.0004287440539671616, 'samples': 7080960, 'steps': 13829, 'loss/train': 2.5434088706970215} 02/24/2022 19:27:58 - INFO - codeparrot_training - Step 13830: {'lr': 0.0004287326137892351, 'samples': 7081472, 'steps': 13830, 'loss/train': 1.6488548517227173} 02/24/2022 19:28:03 - INFO - codeparrot_training - Step 13831: {'lr': 0.00042872117284567486, 'samples': 7081984, 'steps': 13831, 'loss/train': 1.8399345874786377} 02/24/2022 19:28:07 - INFO - codeparrot_training - Step 13832: {'lr': 0.0004287097311365299, 'samples': 7082496, 'steps': 13832, 'loss/train': 1.0356048345565796} 02/24/2022 19:28:12 - INFO - codeparrot_training - Step 13833: {'lr': 0.0004286982886618491, 'samples': 7083008, 'steps': 13833, 'loss/train': 1.7882654666900635} 02/24/2022 19:28:16 - INFO - codeparrot_training - Step 13834: {'lr': 0.0004286868454216816, 'samples': 7083520, 'steps': 13834, 'loss/train': 1.4949910640716553} 02/24/2022 19:28:21 - INFO - codeparrot_training - Step 13835: {'lr': 0.00042867540141607643, 'samples': 7084032, 'steps': 13835, 'loss/train': 0.19167253375053406} 02/24/2022 19:28:25 - INFO - codeparrot_training - Step 13836: {'lr': 0.0004286639566450826, 'samples': 7084544, 'steps': 13836, 'loss/train': 1.452298641204834} 02/24/2022 19:28:31 - INFO - codeparrot_training - Step 13837: {'lr': 0.00042865251110874903, 'samples': 7085056, 'steps': 13837, 'loss/train': 0.1963367760181427} 02/24/2022 19:28:34 - INFO - codeparrot_training - Step 13838: {'lr': 0.00042864106480712495, 'samples': 7085568, 'steps': 13838, 'loss/train': 2.1012227535247803} 02/24/2022 19:28:40 - INFO - codeparrot_training - Step 13839: {'lr': 0.00042862961774025915, 'samples': 7086080, 'steps': 13839, 'loss/train': 1.7118444442749023} 02/24/2022 19:28:43 - INFO - codeparrot_training - Step 13840: {'lr': 0.00042861816990820087, 'samples': 7086592, 'steps': 13840, 'loss/train': 1.9184969663619995} 02/24/2022 19:28:49 - INFO - codeparrot_training - Step 13841: {'lr': 0.00042860672131099904, 'samples': 7087104, 'steps': 13841, 'loss/train': 2.407456874847412} 02/24/2022 19:28:53 - INFO - codeparrot_training - Step 13842: {'lr': 0.00042859527194870275, 'samples': 7087616, 'steps': 13842, 'loss/train': 1.7508593797683716} 02/24/2022 19:28:58 - INFO - codeparrot_training - Step 13843: {'lr': 0.000428583821821361, 'samples': 7088128, 'steps': 13843, 'loss/train': 2.053947925567627} 02/24/2022 19:29:02 - INFO - codeparrot_training - Step 13844: {'lr': 0.00042857237092902285, 'samples': 7088640, 'steps': 13844, 'loss/train': 2.5115182399749756} 02/24/2022 19:29:05 - INFO - codeparrot_training - Step 13845: {'lr': 0.0004285609192717374, 'samples': 7089152, 'steps': 13845, 'loss/train': 2.3884294033050537} 02/24/2022 19:29:11 - INFO - codeparrot_training - Step 13846: {'lr': 0.00042854946684955366, 'samples': 7089664, 'steps': 13846, 'loss/train': 2.2477335929870605} 02/24/2022 19:29:14 - INFO - codeparrot_training - Step 13847: {'lr': 0.00042853801366252067, 'samples': 7090176, 'steps': 13847, 'loss/train': 2.7596707344055176} 02/24/2022 19:29:20 - INFO - codeparrot_training - Step 13848: {'lr': 0.00042852655971068756, 'samples': 7090688, 'steps': 13848, 'loss/train': 0.15768425166606903} 02/24/2022 19:29:24 - INFO - codeparrot_training - Step 13849: {'lr': 0.0004285151049941033, 'samples': 7091200, 'steps': 13849, 'loss/train': 2.1418614387512207} 02/24/2022 19:29:29 - INFO - codeparrot_training - Step 13850: {'lr': 0.00042850364951281707, 'samples': 7091712, 'steps': 13850, 'loss/train': 1.3765848875045776} 02/24/2022 19:29:33 - INFO - codeparrot_training - Step 13851: {'lr': 0.00042849219326687786, 'samples': 7092224, 'steps': 13851, 'loss/train': 1.527712106704712} 02/24/2022 19:29:38 - INFO - codeparrot_training - Step 13852: {'lr': 0.0004284807362563348, 'samples': 7092736, 'steps': 13852, 'loss/train': 2.095750570297241} 02/24/2022 19:29:42 - INFO - codeparrot_training - Step 13853: {'lr': 0.00042846927848123694, 'samples': 7093248, 'steps': 13853, 'loss/train': 2.2116305828094482} 02/24/2022 19:29:47 - INFO - codeparrot_training - Step 13854: {'lr': 0.00042845781994163334, 'samples': 7093760, 'steps': 13854, 'loss/train': 1.7456697225570679} 02/24/2022 19:29:51 - INFO - codeparrot_training - Step 13855: {'lr': 0.00042844636063757316, 'samples': 7094272, 'steps': 13855, 'loss/train': 2.178934335708618} 02/24/2022 19:29:57 - INFO - codeparrot_training - Step 13856: {'lr': 0.00042843490056910534, 'samples': 7094784, 'steps': 13856, 'loss/train': 1.930143117904663} 02/24/2022 19:30:00 - INFO - codeparrot_training - Step 13857: {'lr': 0.0004284234397362791, 'samples': 7095296, 'steps': 13857, 'loss/train': 2.493631362915039} 02/24/2022 19:30:06 - INFO - codeparrot_training - Step 13858: {'lr': 0.0004284119781391436, 'samples': 7095808, 'steps': 13858, 'loss/train': 2.002756357192993} 02/24/2022 19:30:09 - INFO - codeparrot_training - Step 13859: {'lr': 0.00042840051577774766, 'samples': 7096320, 'steps': 13859, 'loss/train': 1.677224040031433} 02/24/2022 19:30:15 - INFO - codeparrot_training - Step 13860: {'lr': 0.00042838905265214067, 'samples': 7096832, 'steps': 13860, 'loss/train': 2.9802050590515137} 02/24/2022 19:30:18 - INFO - codeparrot_training - Step 13861: {'lr': 0.0004283775887623716, 'samples': 7097344, 'steps': 13861, 'loss/train': 2.2648096084594727} 02/24/2022 19:30:24 - INFO - codeparrot_training - Step 13862: {'lr': 0.0004283661241084896, 'samples': 7097856, 'steps': 13862, 'loss/train': 2.162268877029419} 02/24/2022 19:30:27 - INFO - codeparrot_training - Step 13863: {'lr': 0.0004283546586905437, 'samples': 7098368, 'steps': 13863, 'loss/train': 1.8374004364013672} 02/24/2022 19:30:33 - INFO - codeparrot_training - Step 13864: {'lr': 0.00042834319250858316, 'samples': 7098880, 'steps': 13864, 'loss/train': 1.72138512134552} 02/24/2022 19:30:36 - INFO - codeparrot_training - Step 13865: {'lr': 0.000428331725562657, 'samples': 7099392, 'steps': 13865, 'loss/train': 1.6780811548233032} 02/24/2022 19:30:42 - INFO - codeparrot_training - Step 13866: {'lr': 0.0004283202578528143, 'samples': 7099904, 'steps': 13866, 'loss/train': 2.028473138809204} 02/24/2022 19:30:45 - INFO - codeparrot_training - Step 13867: {'lr': 0.00042830878937910426, 'samples': 7100416, 'steps': 13867, 'loss/train': 2.8656997680664062} 02/24/2022 19:30:51 - INFO - codeparrot_training - Step 13868: {'lr': 0.000428297320141576, 'samples': 7100928, 'steps': 13868, 'loss/train': 0.9876556992530823} 02/24/2022 19:30:57 - INFO - codeparrot_training - Step 13869: {'lr': 0.00042828585014027863, 'samples': 7101440, 'steps': 13869, 'loss/train': 2.00734281539917} 02/24/2022 19:31:00 - INFO - codeparrot_training - Step 13870: {'lr': 0.0004282743793752613, 'samples': 7101952, 'steps': 13870, 'loss/train': 1.3925220966339111} 02/24/2022 19:31:06 - INFO - codeparrot_training - Step 13871: {'lr': 0.0004282629078465732, 'samples': 7102464, 'steps': 13871, 'loss/train': 1.8243123292922974} 02/24/2022 19:31:10 - INFO - codeparrot_training - Step 13872: {'lr': 0.0004282514355542633, 'samples': 7102976, 'steps': 13872, 'loss/train': 2.651815414428711} 02/24/2022 19:31:15 - INFO - codeparrot_training - Step 13873: {'lr': 0.0004282399624983808, 'samples': 7103488, 'steps': 13873, 'loss/train': 2.9097788333892822} 02/24/2022 19:31:19 - INFO - codeparrot_training - Step 13874: {'lr': 0.000428228488678975, 'samples': 7104000, 'steps': 13874, 'loss/train': 0.12431791424751282} 02/24/2022 19:31:24 - INFO - codeparrot_training - Step 13875: {'lr': 0.000428217014096095, 'samples': 7104512, 'steps': 13875, 'loss/train': 3.4056780338287354} 02/24/2022 19:31:27 - INFO - codeparrot_training - Step 13876: {'lr': 0.00042820553874978987, 'samples': 7105024, 'steps': 13876, 'loss/train': 1.8900150060653687} 02/24/2022 19:31:33 - INFO - codeparrot_training - Step 13877: {'lr': 0.0004281940626401087, 'samples': 7105536, 'steps': 13877, 'loss/train': 2.707639217376709} 02/24/2022 19:31:37 - INFO - codeparrot_training - Step 13878: {'lr': 0.0004281825857671008, 'samples': 7106048, 'steps': 13878, 'loss/train': 1.4920686483383179} 02/24/2022 19:31:42 - INFO - codeparrot_training - Step 13879: {'lr': 0.00042817110813081526, 'samples': 7106560, 'steps': 13879, 'loss/train': 2.484635829925537} 02/24/2022 19:31:46 - INFO - codeparrot_training - Step 13880: {'lr': 0.00042815962973130134, 'samples': 7107072, 'steps': 13880, 'loss/train': 2.5091655254364014} 02/24/2022 19:31:51 - INFO - codeparrot_training - Step 13881: {'lr': 0.00042814815056860814, 'samples': 7107584, 'steps': 13881, 'loss/train': 2.1803884506225586} 02/24/2022 19:31:55 - INFO - codeparrot_training - Step 13882: {'lr': 0.0004281366706427848, 'samples': 7108096, 'steps': 13882, 'loss/train': 1.9432035684585571} 02/24/2022 19:32:03 - INFO - codeparrot_training - Step 13883: {'lr': 0.0004281251899538805, 'samples': 7108608, 'steps': 13883, 'loss/train': 2.4190433025360107} 02/24/2022 19:32:06 - INFO - codeparrot_training - Step 13884: {'lr': 0.0004281137085019445, 'samples': 7109120, 'steps': 13884, 'loss/train': 2.6437957286834717} 02/24/2022 19:32:12 - INFO - codeparrot_training - Step 13885: {'lr': 0.0004281022262870259, 'samples': 7109632, 'steps': 13885, 'loss/train': 2.5035533905029297} 02/24/2022 19:32:15 - INFO - codeparrot_training - Step 13886: {'lr': 0.00042809074330917387, 'samples': 7110144, 'steps': 13886, 'loss/train': 2.959852695465088} 02/24/2022 19:32:21 - INFO - codeparrot_training - Step 13887: {'lr': 0.00042807925956843775, 'samples': 7110656, 'steps': 13887, 'loss/train': 2.476499319076538} 02/24/2022 19:32:24 - INFO - codeparrot_training - Step 13888: {'lr': 0.0004280677750648665, 'samples': 7111168, 'steps': 13888, 'loss/train': 2.1213388442993164} 02/24/2022 19:32:30 - INFO - codeparrot_training - Step 13889: {'lr': 0.0004280562897985095, 'samples': 7111680, 'steps': 13889, 'loss/train': 2.1540985107421875} 02/24/2022 19:32:33 - INFO - codeparrot_training - Step 13890: {'lr': 0.00042804480376941597, 'samples': 7112192, 'steps': 13890, 'loss/train': 2.0236587524414062} 02/24/2022 19:32:39 - INFO - codeparrot_training - Step 13891: {'lr': 0.0004280333169776349, 'samples': 7112704, 'steps': 13891, 'loss/train': 2.021449089050293} 02/24/2022 19:32:42 - INFO - codeparrot_training - Step 13892: {'lr': 0.00042802182942321576, 'samples': 7113216, 'steps': 13892, 'loss/train': 1.9408913850784302} 02/24/2022 19:32:50 - INFO - codeparrot_training - Step 13893: {'lr': 0.00042801034110620756, 'samples': 7113728, 'steps': 13893, 'loss/train': 2.097508668899536} 02/24/2022 19:32:54 - INFO - codeparrot_training - Step 13894: {'lr': 0.00042799885202665964, 'samples': 7114240, 'steps': 13894, 'loss/train': 1.7103838920593262} 02/24/2022 19:32:59 - INFO - codeparrot_training - Step 13895: {'lr': 0.0004279873621846211, 'samples': 7114752, 'steps': 13895, 'loss/train': 1.814162254333496} 02/24/2022 19:33:03 - INFO - codeparrot_training - Step 13896: {'lr': 0.0004279758715801412, 'samples': 7115264, 'steps': 13896, 'loss/train': 3.1128132343292236} 02/24/2022 19:33:08 - INFO - codeparrot_training - Step 13897: {'lr': 0.0004279643802132692, 'samples': 7115776, 'steps': 13897, 'loss/train': 2.096381664276123} 02/24/2022 19:33:12 - INFO - codeparrot_training - Step 13898: {'lr': 0.0004279528880840544, 'samples': 7116288, 'steps': 13898, 'loss/train': 1.4319543838500977} 02/24/2022 19:33:18 - INFO - codeparrot_training - Step 13899: {'lr': 0.00042794139519254583, 'samples': 7116800, 'steps': 13899, 'loss/train': 1.8660544157028198} 02/24/2022 19:33:21 - INFO - codeparrot_training - Step 13900: {'lr': 0.00042792990153879285, 'samples': 7117312, 'steps': 13900, 'loss/train': 2.2695634365081787} 02/24/2022 19:33:27 - INFO - codeparrot_training - Step 13901: {'lr': 0.00042791840712284466, 'samples': 7117824, 'steps': 13901, 'loss/train': 2.7423362731933594} 02/24/2022 19:33:30 - INFO - codeparrot_training - Step 13902: {'lr': 0.0004279069119447505, 'samples': 7118336, 'steps': 13902, 'loss/train': 0.5829432010650635} 02/24/2022 19:33:37 - INFO - codeparrot_training - Step 13903: {'lr': 0.0004278954160045597, 'samples': 7118848, 'steps': 13903, 'loss/train': 1.1926600933074951} 02/24/2022 19:33:41 - INFO - codeparrot_training - Step 13904: {'lr': 0.0004278839193023214, 'samples': 7119360, 'steps': 13904, 'loss/train': 2.1163101196289062} 02/24/2022 19:33:47 - INFO - codeparrot_training - Step 13905: {'lr': 0.00042787242183808485, 'samples': 7119872, 'steps': 13905, 'loss/train': 2.2231385707855225} 02/24/2022 19:33:50 - INFO - codeparrot_training - Step 13906: {'lr': 0.00042786092361189927, 'samples': 7120384, 'steps': 13906, 'loss/train': 2.0194473266601562} 02/24/2022 19:33:56 - INFO - codeparrot_training - Step 13907: {'lr': 0.00042784942462381403, 'samples': 7120896, 'steps': 13907, 'loss/train': 1.9014095067977905} 02/24/2022 19:33:59 - INFO - codeparrot_training - Step 13908: {'lr': 0.0004278379248738783, 'samples': 7121408, 'steps': 13908, 'loss/train': 2.3189709186553955} 02/24/2022 19:34:03 - INFO - codeparrot_training - Step 13909: {'lr': 0.00042782642436214137, 'samples': 7121920, 'steps': 13909, 'loss/train': 2.4209084510803223} 02/24/2022 19:34:08 - INFO - codeparrot_training - Step 13910: {'lr': 0.00042781492308865255, 'samples': 7122432, 'steps': 13910, 'loss/train': 0.5683369040489197} 02/24/2022 19:34:12 - INFO - codeparrot_training - Step 13911: {'lr': 0.000427803421053461, 'samples': 7122944, 'steps': 13911, 'loss/train': 2.0672590732574463} 02/24/2022 19:34:17 - INFO - codeparrot_training - Step 13912: {'lr': 0.0004277919182566161, 'samples': 7123456, 'steps': 13912, 'loss/train': 1.339026927947998} 02/24/2022 19:34:20 - INFO - codeparrot_training - Step 13913: {'lr': 0.0004277804146981671, 'samples': 7123968, 'steps': 13913, 'loss/train': 1.1579540967941284} 02/24/2022 19:34:28 - INFO - codeparrot_training - Step 13914: {'lr': 0.00042776891037816324, 'samples': 7124480, 'steps': 13914, 'loss/train': 2.402513265609741} 02/24/2022 19:34:34 - INFO - codeparrot_training - Step 13915: {'lr': 0.00042775740529665373, 'samples': 7124992, 'steps': 13915, 'loss/train': 1.5199159383773804} 02/24/2022 19:34:37 - INFO - codeparrot_training - Step 13916: {'lr': 0.000427745899453688, 'samples': 7125504, 'steps': 13916, 'loss/train': 2.0406506061553955} 02/24/2022 19:34:43 - INFO - codeparrot_training - Step 13917: {'lr': 0.0004277343928493153, 'samples': 7126016, 'steps': 13917, 'loss/train': 2.941728353500366} 02/24/2022 19:34:46 - INFO - codeparrot_training - Step 13918: {'lr': 0.0004277228854835849, 'samples': 7126528, 'steps': 13918, 'loss/train': 2.309279441833496} 02/24/2022 19:34:52 - INFO - codeparrot_training - Step 13919: {'lr': 0.0004277113773565461, 'samples': 7127040, 'steps': 13919, 'loss/train': 2.1819634437561035} 02/24/2022 19:34:55 - INFO - codeparrot_training - Step 13920: {'lr': 0.00042769986846824813, 'samples': 7127552, 'steps': 13920, 'loss/train': 2.8887381553649902} 02/24/2022 19:35:01 - INFO - codeparrot_training - Step 13921: {'lr': 0.00042768835881874036, 'samples': 7128064, 'steps': 13921, 'loss/train': 1.737459421157837} 02/24/2022 19:35:04 - INFO - codeparrot_training - Step 13922: {'lr': 0.00042767684840807214, 'samples': 7128576, 'steps': 13922, 'loss/train': 1.5378804206848145} 02/24/2022 19:35:11 - INFO - codeparrot_training - Step 13923: {'lr': 0.00042766533723629264, 'samples': 7129088, 'steps': 13923, 'loss/train': 2.1596288681030273} 02/24/2022 19:35:15 - INFO - codeparrot_training - Step 13924: {'lr': 0.0004276538253034513, 'samples': 7129600, 'steps': 13924, 'loss/train': 2.5023815631866455} 02/24/2022 19:35:20 - INFO - codeparrot_training - Step 13925: {'lr': 0.0004276423126095974, 'samples': 7130112, 'steps': 13925, 'loss/train': 2.2584335803985596} 02/24/2022 19:35:24 - INFO - codeparrot_training - Step 13926: {'lr': 0.0004276307991547802, 'samples': 7130624, 'steps': 13926, 'loss/train': 2.548325777053833} 02/24/2022 19:35:29 - INFO - codeparrot_training - Step 13927: {'lr': 0.0004276192849390491, 'samples': 7131136, 'steps': 13927, 'loss/train': 3.0007622241973877} 02/24/2022 19:35:33 - INFO - codeparrot_training - Step 13928: {'lr': 0.0004276077699624534, 'samples': 7131648, 'steps': 13928, 'loss/train': 1.0408521890640259} 02/24/2022 19:35:36 - INFO - codeparrot_training - Step 13929: {'lr': 0.00042759625422504236, 'samples': 7132160, 'steps': 13929, 'loss/train': 1.2168833017349243} 02/24/2022 19:35:43 - INFO - codeparrot_training - Step 13930: {'lr': 0.00042758473772686533, 'samples': 7132672, 'steps': 13930, 'loss/train': 1.8865442276000977} 02/24/2022 19:35:46 - INFO - codeparrot_training - Step 13931: {'lr': 0.0004275732204679718, 'samples': 7133184, 'steps': 13931, 'loss/train': 2.5318686962127686} 02/24/2022 19:35:51 - INFO - codeparrot_training - Step 13932: {'lr': 0.0004275617024484109, 'samples': 7133696, 'steps': 13932, 'loss/train': 1.77742338180542} 02/24/2022 19:35:55 - INFO - codeparrot_training - Step 13933: {'lr': 0.000427550183668232, 'samples': 7134208, 'steps': 13933, 'loss/train': 3.4306695461273193} 02/24/2022 19:36:00 - INFO - codeparrot_training - Step 13934: {'lr': 0.00042753866412748455, 'samples': 7134720, 'steps': 13934, 'loss/train': 0.13265100121498108} 02/24/2022 19:36:04 - INFO - codeparrot_training - Step 13935: {'lr': 0.00042752714382621784, 'samples': 7135232, 'steps': 13935, 'loss/train': 2.451072931289673} 02/24/2022 19:36:09 - INFO - codeparrot_training - Step 13936: {'lr': 0.0004275156227644812, 'samples': 7135744, 'steps': 13936, 'loss/train': 1.64609694480896} 02/24/2022 19:36:13 - INFO - codeparrot_training - Step 13937: {'lr': 0.00042750410094232394, 'samples': 7136256, 'steps': 13937, 'loss/train': 1.148285150527954} 02/24/2022 19:36:18 - INFO - codeparrot_training - Step 13938: {'lr': 0.0004274925783597956, 'samples': 7136768, 'steps': 13938, 'loss/train': 2.6608564853668213} 02/24/2022 19:36:22 - INFO - codeparrot_training - Step 13939: {'lr': 0.0004274810550169453, 'samples': 7137280, 'steps': 13939, 'loss/train': 1.4991357326507568} 02/24/2022 19:36:29 - INFO - codeparrot_training - Step 13940: {'lr': 0.00042746953091382254, 'samples': 7137792, 'steps': 13940, 'loss/train': 1.3732562065124512} 02/24/2022 19:36:33 - INFO - codeparrot_training - Step 13941: {'lr': 0.00042745800605047677, 'samples': 7138304, 'steps': 13941, 'loss/train': 2.1584925651550293} 02/24/2022 19:36:38 - INFO - codeparrot_training - Step 13942: {'lr': 0.00042744648042695717, 'samples': 7138816, 'steps': 13942, 'loss/train': 1.8618122339248657} 02/24/2022 19:36:42 - INFO - codeparrot_training - Step 13943: {'lr': 0.0004274349540433132, 'samples': 7139328, 'steps': 13943, 'loss/train': 1.9625802040100098} 02/24/2022 19:36:47 - INFO - codeparrot_training - Step 13944: {'lr': 0.00042742342689959425, 'samples': 7139840, 'steps': 13944, 'loss/train': 2.0464932918548584} 02/24/2022 19:36:51 - INFO - codeparrot_training - Step 13945: {'lr': 0.00042741189899584965, 'samples': 7140352, 'steps': 13945, 'loss/train': 1.3461633920669556} 02/24/2022 19:36:56 - INFO - codeparrot_training - Step 13946: {'lr': 0.00042740037033212877, 'samples': 7140864, 'steps': 13946, 'loss/train': 2.097648859024048} 02/24/2022 19:37:00 - INFO - codeparrot_training - Step 13947: {'lr': 0.0004273888409084811, 'samples': 7141376, 'steps': 13947, 'loss/train': 2.3671865463256836} 02/24/2022 19:37:05 - INFO - codeparrot_training - Step 13948: {'lr': 0.0004273773107249559, 'samples': 7141888, 'steps': 13948, 'loss/train': 2.186274766921997} 02/24/2022 19:37:09 - INFO - codeparrot_training - Step 13949: {'lr': 0.0004273657797816027, 'samples': 7142400, 'steps': 13949, 'loss/train': 2.4363222122192383} 02/24/2022 19:37:14 - INFO - codeparrot_training - Step 13950: {'lr': 0.0004273542480784708, 'samples': 7142912, 'steps': 13950, 'loss/train': 1.1649454832077026} 02/24/2022 19:37:18 - INFO - codeparrot_training - Step 13951: {'lr': 0.00042734271561560956, 'samples': 7143424, 'steps': 13951, 'loss/train': 1.4500069618225098} 02/24/2022 19:37:25 - INFO - codeparrot_training - Step 13952: {'lr': 0.00042733118239306845, 'samples': 7143936, 'steps': 13952, 'loss/train': 1.8847614526748657} 02/24/2022 19:37:28 - INFO - codeparrot_training - Step 13953: {'lr': 0.0004273196484108969, 'samples': 7144448, 'steps': 13953, 'loss/train': 1.9720548391342163} 02/24/2022 19:37:34 - INFO - codeparrot_training - Step 13954: {'lr': 0.00042730811366914435, 'samples': 7144960, 'steps': 13954, 'loss/train': 2.0264317989349365} 02/24/2022 19:37:37 - INFO - codeparrot_training - Step 13955: {'lr': 0.0004272965781678601, 'samples': 7145472, 'steps': 13955, 'loss/train': 2.0273795127868652} 02/24/2022 19:37:43 - INFO - codeparrot_training - Step 13956: {'lr': 0.0004272850419070935, 'samples': 7145984, 'steps': 13956, 'loss/train': 2.9655637741088867} 02/24/2022 19:37:46 - INFO - codeparrot_training - Step 13957: {'lr': 0.00042727350488689416, 'samples': 7146496, 'steps': 13957, 'loss/train': 3.0586349964141846} 02/24/2022 19:37:52 - INFO - codeparrot_training - Step 13958: {'lr': 0.00042726196710731135, 'samples': 7147008, 'steps': 13958, 'loss/train': 2.34102463722229} 02/24/2022 19:37:55 - INFO - codeparrot_training - Step 13959: {'lr': 0.0004272504285683947, 'samples': 7147520, 'steps': 13959, 'loss/train': 1.2469335794448853} 02/24/2022 19:38:03 - INFO - codeparrot_training - Step 13960: {'lr': 0.0004272388892701934, 'samples': 7148032, 'steps': 13960, 'loss/train': 2.715869903564453} 02/24/2022 19:38:06 - INFO - codeparrot_training - Step 13961: {'lr': 0.000427227349212757, 'samples': 7148544, 'steps': 13961, 'loss/train': 2.4502198696136475} 02/24/2022 19:38:11 - INFO - codeparrot_training - Step 13962: {'lr': 0.0004272158083961348, 'samples': 7149056, 'steps': 13962, 'loss/train': 2.77317214012146} 02/24/2022 19:38:15 - INFO - codeparrot_training - Step 13963: {'lr': 0.0004272042668203765, 'samples': 7149568, 'steps': 13963, 'loss/train': 2.952699899673462} 02/24/2022 19:38:20 - INFO - codeparrot_training - Step 13964: {'lr': 0.00042719272448553137, 'samples': 7150080, 'steps': 13964, 'loss/train': 0.8886147737503052} 02/24/2022 19:38:24 - INFO - codeparrot_training - Step 13965: {'lr': 0.00042718118139164883, 'samples': 7150592, 'steps': 13965, 'loss/train': 1.395202398300171} 02/24/2022 19:38:30 - INFO - codeparrot_training - Step 13966: {'lr': 0.00042716963753877836, 'samples': 7151104, 'steps': 13966, 'loss/train': 2.372715711593628} 02/24/2022 19:38:33 - INFO - codeparrot_training - Step 13967: {'lr': 0.0004271580929269695, 'samples': 7151616, 'steps': 13967, 'loss/train': 1.9061944484710693} 02/24/2022 19:38:39 - INFO - codeparrot_training - Step 13968: {'lr': 0.0004271465475562716, 'samples': 7152128, 'steps': 13968, 'loss/train': 1.1169462203979492} 02/24/2022 19:38:42 - INFO - codeparrot_training - Step 13969: {'lr': 0.00042713500142673404, 'samples': 7152640, 'steps': 13969, 'loss/train': 1.716810703277588} 02/24/2022 19:38:48 - INFO - codeparrot_training - Step 13970: {'lr': 0.00042712345453840644, 'samples': 7153152, 'steps': 13970, 'loss/train': 1.4124257564544678} 02/24/2022 19:38:51 - INFO - codeparrot_training - Step 13971: {'lr': 0.00042711190689133827, 'samples': 7153664, 'steps': 13971, 'loss/train': 2.0908589363098145} 02/24/2022 19:38:58 - INFO - codeparrot_training - Step 13972: {'lr': 0.0004271003584855788, 'samples': 7154176, 'steps': 13972, 'loss/train': 2.243739604949951} 02/24/2022 19:39:02 - INFO - codeparrot_training - Step 13973: {'lr': 0.0004270888093211778, 'samples': 7154688, 'steps': 13973, 'loss/train': 2.46085786819458} 02/24/2022 19:39:07 - INFO - codeparrot_training - Step 13974: {'lr': 0.0004270772593981844, 'samples': 7155200, 'steps': 13974, 'loss/train': 3.0705995559692383} 02/24/2022 19:39:11 - INFO - codeparrot_training - Step 13975: {'lr': 0.0004270657087166484, 'samples': 7155712, 'steps': 13975, 'loss/train': 2.056504011154175} 02/24/2022 19:39:16 - INFO - codeparrot_training - Step 13976: {'lr': 0.000427054157276619, 'samples': 7156224, 'steps': 13976, 'loss/train': 1.7271323204040527} 02/24/2022 19:39:20 - INFO - codeparrot_training - Step 13977: {'lr': 0.0004270426050781458, 'samples': 7156736, 'steps': 13977, 'loss/train': 2.601325511932373} 02/24/2022 19:39:25 - INFO - codeparrot_training - Step 13978: {'lr': 0.00042703105212127846, 'samples': 7157248, 'steps': 13978, 'loss/train': 1.536615014076233} 02/24/2022 19:39:29 - INFO - codeparrot_training - Step 13979: {'lr': 0.0004270194984060662, 'samples': 7157760, 'steps': 13979, 'loss/train': 2.1201298236846924} 02/24/2022 19:39:35 - INFO - codeparrot_training - Step 13980: {'lr': 0.0004270079439325586, 'samples': 7158272, 'steps': 13980, 'loss/train': 2.6276681423187256} 02/24/2022 19:39:38 - INFO - codeparrot_training - Step 13981: {'lr': 0.0004269963887008053, 'samples': 7158784, 'steps': 13981, 'loss/train': 1.9070311784744263} 02/24/2022 19:39:44 - INFO - codeparrot_training - Step 13982: {'lr': 0.00042698483271085555, 'samples': 7159296, 'steps': 13982, 'loss/train': 1.881844401359558} 02/24/2022 19:39:47 - INFO - codeparrot_training - Step 13983: {'lr': 0.0004269732759627589, 'samples': 7159808, 'steps': 13983, 'loss/train': 1.634737491607666} 02/24/2022 19:39:53 - INFO - codeparrot_training - Step 13984: {'lr': 0.0004269617184565651, 'samples': 7160320, 'steps': 13984, 'loss/train': 1.7941128015518188} 02/24/2022 19:39:56 - INFO - codeparrot_training - Step 13985: {'lr': 0.00042695016019232343, 'samples': 7160832, 'steps': 13985, 'loss/train': 1.1788517236709595} 02/24/2022 19:40:03 - INFO - codeparrot_training - Step 13986: {'lr': 0.0004269386011700834, 'samples': 7161344, 'steps': 13986, 'loss/train': 1.9005893468856812} 02/24/2022 19:40:07 - INFO - codeparrot_training - Step 13987: {'lr': 0.00042692704138989467, 'samples': 7161856, 'steps': 13987, 'loss/train': 1.90274178981781} 02/24/2022 19:40:12 - INFO - codeparrot_training - Step 13988: {'lr': 0.00042691548085180666, 'samples': 7162368, 'steps': 13988, 'loss/train': 1.7172231674194336} 02/24/2022 19:40:16 - INFO - codeparrot_training - Step 13989: {'lr': 0.00042690391955586886, 'samples': 7162880, 'steps': 13989, 'loss/train': 1.4420868158340454} 02/24/2022 19:40:21 - INFO - codeparrot_training - Step 13990: {'lr': 0.00042689235750213093, 'samples': 7163392, 'steps': 13990, 'loss/train': 2.505735397338867} 02/24/2022 19:40:25 - INFO - codeparrot_training - Step 13991: {'lr': 0.0004268807946906422, 'samples': 7163904, 'steps': 13991, 'loss/train': 2.922675848007202} 02/24/2022 19:40:31 - INFO - codeparrot_training - Step 13992: {'lr': 0.0004268692311214524, 'samples': 7164416, 'steps': 13992, 'loss/train': 2.881211757659912} 02/24/2022 19:40:34 - INFO - codeparrot_training - Step 13993: {'lr': 0.00042685766679461095, 'samples': 7164928, 'steps': 13993, 'loss/train': 2.036750078201294} 02/24/2022 19:40:39 - INFO - codeparrot_training - Step 13994: {'lr': 0.0004268461017101674, 'samples': 7165440, 'steps': 13994, 'loss/train': 1.924837350845337} 02/24/2022 19:40:43 - INFO - codeparrot_training - Step 13995: {'lr': 0.00042683453586817136, 'samples': 7165952, 'steps': 13995, 'loss/train': 1.8309330940246582} 02/24/2022 19:40:50 - INFO - codeparrot_training - Step 13996: {'lr': 0.00042682296926867226, 'samples': 7166464, 'steps': 13996, 'loss/train': 2.459993839263916} 02/24/2022 19:40:54 - INFO - codeparrot_training - Step 13997: {'lr': 0.0004268114019117197, 'samples': 7166976, 'steps': 13997, 'loss/train': 1.8924742937088013} 02/24/2022 19:40:59 - INFO - codeparrot_training - Step 13998: {'lr': 0.00042679983379736324, 'samples': 7167488, 'steps': 13998, 'loss/train': 1.7144615650177002} 02/24/2022 19:41:03 - INFO - codeparrot_training - Step 13999: {'lr': 0.0004267882649256525, 'samples': 7168000, 'steps': 13999, 'loss/train': 1.6556921005249023} 02/24/2022 19:41:03 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 19:41:20 - WARNING - huggingface_hub.repository - Several commits (14) will be pushed upstream. 02/24/2022 19:41:20 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 19:41:58 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 7d01f0b..3a61b16 floral-grass-11 -> floral-grass-11 02/24/2022 19:42:05 - INFO - codeparrot_training - Step 14000: {'lr': 0.00042677669529663686, 'samples': 7168512, 'steps': 14000, 'loss/train': 2.7728171348571777} 02/24/2022 19:42:08 - INFO - codeparrot_training - Step 14001: {'lr': 0.0004267651249103661, 'samples': 7169024, 'steps': 14001, 'loss/train': 2.645996332168579} 02/24/2022 19:42:14 - INFO - codeparrot_training - Step 14002: {'lr': 0.00042675355376688964, 'samples': 7169536, 'steps': 14002, 'loss/train': 2.4357352256774902} 02/24/2022 19:42:17 - INFO - codeparrot_training - Step 14003: {'lr': 0.000426741981866257, 'samples': 7170048, 'steps': 14003, 'loss/train': 2.1884565353393555} 02/24/2022 19:42:23 - INFO - codeparrot_training - Step 14004: {'lr': 0.00042673040920851793, 'samples': 7170560, 'steps': 14004, 'loss/train': 1.3800054788589478} 02/24/2022 19:42:26 - INFO - codeparrot_training - Step 14005: {'lr': 0.00042671883579372186, 'samples': 7171072, 'steps': 14005, 'loss/train': 2.5828280448913574} 02/24/2022 19:42:32 - INFO - codeparrot_training - Step 14006: {'lr': 0.00042670726162191843, 'samples': 7171584, 'steps': 14006, 'loss/train': 1.8915454149246216} 02/24/2022 19:42:35 - INFO - codeparrot_training - Step 14007: {'lr': 0.0004266956866931572, 'samples': 7172096, 'steps': 14007, 'loss/train': 1.3992984294891357} 02/24/2022 19:42:43 - INFO - codeparrot_training - Step 14008: {'lr': 0.0004266841110074878, 'samples': 7172608, 'steps': 14008, 'loss/train': 1.6871919631958008} 02/24/2022 19:42:46 - INFO - codeparrot_training - Step 14009: {'lr': 0.0004266725345649597, 'samples': 7173120, 'steps': 14009, 'loss/train': 2.5861899852752686} 02/24/2022 19:42:52 - INFO - codeparrot_training - Step 14010: {'lr': 0.0004266609573656226, 'samples': 7173632, 'steps': 14010, 'loss/train': 2.141249656677246} 02/24/2022 19:42:56 - INFO - codeparrot_training - Step 14011: {'lr': 0.000426649379409526, 'samples': 7174144, 'steps': 14011, 'loss/train': 3.1508538722991943} 02/24/2022 19:43:01 - INFO - codeparrot_training - Step 14012: {'lr': 0.00042663780069671965, 'samples': 7174656, 'steps': 14012, 'loss/train': 1.845304250717163} 02/24/2022 19:43:05 - INFO - codeparrot_training - Step 14013: {'lr': 0.000426626221227253, 'samples': 7175168, 'steps': 14013, 'loss/train': 1.7898507118225098} 02/24/2022 19:43:10 - INFO - codeparrot_training - Step 14014: {'lr': 0.00042661464100117566, 'samples': 7175680, 'steps': 14014, 'loss/train': 0.6427545547485352} 02/24/2022 19:43:14 - INFO - codeparrot_training - Step 14015: {'lr': 0.00042660306001853735, 'samples': 7176192, 'steps': 14015, 'loss/train': 1.653831958770752} 02/24/2022 19:43:19 - INFO - codeparrot_training - Step 14016: {'lr': 0.0004265914782793875, 'samples': 7176704, 'steps': 14016, 'loss/train': 0.8801273107528687} 02/24/2022 19:43:23 - INFO - codeparrot_training - Step 14017: {'lr': 0.000426579895783776, 'samples': 7177216, 'steps': 14017, 'loss/train': 2.425933837890625} 02/24/2022 19:43:30 - INFO - codeparrot_training - Step 14018: {'lr': 0.0004265683125317521, 'samples': 7177728, 'steps': 14018, 'loss/train': 1.94931960105896} 02/24/2022 19:43:33 - INFO - codeparrot_training - Step 14019: {'lr': 0.0004265567285233658, 'samples': 7178240, 'steps': 14019, 'loss/train': 2.7595584392547607} 02/24/2022 19:43:39 - INFO - codeparrot_training - Step 14020: {'lr': 0.0004265451437586664, 'samples': 7178752, 'steps': 14020, 'loss/train': 1.0570260286331177} 02/24/2022 19:43:42 - INFO - codeparrot_training - Step 14021: {'lr': 0.0004265335582377038, 'samples': 7179264, 'steps': 14021, 'loss/train': 1.486286997795105} 02/24/2022 19:43:48 - INFO - codeparrot_training - Step 14022: {'lr': 0.0004265219719605273, 'samples': 7179776, 'steps': 14022, 'loss/train': 2.3152973651885986} 02/24/2022 19:43:51 - INFO - codeparrot_training - Step 14023: {'lr': 0.0004265103849271869, 'samples': 7180288, 'steps': 14023, 'loss/train': 2.016923427581787} 02/24/2022 19:43:57 - INFO - codeparrot_training - Step 14024: {'lr': 0.000426498797137732, 'samples': 7180800, 'steps': 14024, 'loss/train': 1.445349097251892} 02/24/2022 19:44:00 - INFO - codeparrot_training - Step 14025: {'lr': 0.0004264872085922122, 'samples': 7181312, 'steps': 14025, 'loss/train': 0.40025123953819275} 02/24/2022 19:44:06 - INFO - codeparrot_training - Step 14026: {'lr': 0.0004264756192906774, 'samples': 7181824, 'steps': 14026, 'loss/train': 2.2487075328826904} 02/24/2022 19:44:09 - INFO - codeparrot_training - Step 14027: {'lr': 0.000426464029233177, 'samples': 7182336, 'steps': 14027, 'loss/train': 1.7559268474578857} 02/24/2022 19:44:17 - INFO - codeparrot_training - Step 14028: {'lr': 0.0004264524384197608, 'samples': 7182848, 'steps': 14028, 'loss/train': 2.0156359672546387} 02/24/2022 19:44:20 - INFO - codeparrot_training - Step 14029: {'lr': 0.0004264408468504783, 'samples': 7183360, 'steps': 14029, 'loss/train': 1.5344573259353638} 02/24/2022 19:44:26 - INFO - codeparrot_training - Step 14030: {'lr': 0.00042642925452537927, 'samples': 7183872, 'steps': 14030, 'loss/train': 2.616739511489868} 02/24/2022 19:44:29 - INFO - codeparrot_training - Step 14031: {'lr': 0.0004264176614445133, 'samples': 7184384, 'steps': 14031, 'loss/train': 2.3235089778900146} 02/24/2022 19:44:35 - INFO - codeparrot_training - Step 14032: {'lr': 0.0004264060676079302, 'samples': 7184896, 'steps': 14032, 'loss/train': 0.9734417200088501} 02/24/2022 19:44:38 - INFO - codeparrot_training - Step 14033: {'lr': 0.00042639447301567944, 'samples': 7185408, 'steps': 14033, 'loss/train': 1.7799558639526367} 02/24/2022 19:44:44 - INFO - codeparrot_training - Step 14034: {'lr': 0.0004263828776678108, 'samples': 7185920, 'steps': 14034, 'loss/train': 3.505218029022217} 02/24/2022 19:44:47 - INFO - codeparrot_training - Step 14035: {'lr': 0.00042637128156437385, 'samples': 7186432, 'steps': 14035, 'loss/train': 1.9882664680480957} 02/24/2022 19:44:53 - INFO - codeparrot_training - Step 14036: {'lr': 0.0004263596847054184, 'samples': 7186944, 'steps': 14036, 'loss/train': 2.313277244567871} 02/24/2022 19:44:56 - INFO - codeparrot_training - Step 14037: {'lr': 0.00042634808709099403, 'samples': 7187456, 'steps': 14037, 'loss/train': 1.884895920753479} 02/24/2022 19:45:02 - INFO - codeparrot_training - Step 14038: {'lr': 0.0004263364887211505, 'samples': 7187968, 'steps': 14038, 'loss/train': 1.0802332162857056} 02/24/2022 19:45:05 - INFO - codeparrot_training - Step 14039: {'lr': 0.0004263248895959374, 'samples': 7188480, 'steps': 14039, 'loss/train': 0.7658008337020874} 02/24/2022 19:45:11 - INFO - codeparrot_training - Step 14040: {'lr': 0.0004263132897154044, 'samples': 7188992, 'steps': 14040, 'loss/train': 0.4843223989009857} 02/24/2022 19:45:14 - INFO - codeparrot_training - Step 14041: {'lr': 0.0004263016890796014, 'samples': 7189504, 'steps': 14041, 'loss/train': 1.9767647981643677} 02/24/2022 19:45:20 - INFO - codeparrot_training - Step 14042: {'lr': 0.0004262900876885778, 'samples': 7190016, 'steps': 14042, 'loss/train': 2.517286777496338} 02/24/2022 19:45:23 - INFO - codeparrot_training - Step 14043: {'lr': 0.0004262784855423836, 'samples': 7190528, 'steps': 14043, 'loss/train': 2.145695924758911} 02/24/2022 19:45:31 - INFO - codeparrot_training - Step 14044: {'lr': 0.00042626688264106816, 'samples': 7191040, 'steps': 14044, 'loss/train': 1.0589336156845093} 02/24/2022 19:45:35 - INFO - codeparrot_training - Step 14045: {'lr': 0.00042625527898468155, 'samples': 7191552, 'steps': 14045, 'loss/train': 2.410911798477173} 02/24/2022 19:45:40 - INFO - codeparrot_training - Step 14046: {'lr': 0.0004262436745732732, 'samples': 7192064, 'steps': 14046, 'loss/train': 2.678050994873047} 02/24/2022 19:45:44 - INFO - codeparrot_training - Step 14047: {'lr': 0.00042623206940689285, 'samples': 7192576, 'steps': 14047, 'loss/train': 2.223356246948242} 02/24/2022 19:45:49 - INFO - codeparrot_training - Step 14048: {'lr': 0.00042622046348559034, 'samples': 7193088, 'steps': 14048, 'loss/train': 1.682914137840271} 02/24/2022 19:45:53 - INFO - codeparrot_training - Step 14049: {'lr': 0.0004262088568094153, 'samples': 7193600, 'steps': 14049, 'loss/train': 1.5549159049987793} 02/24/2022 19:45:58 - INFO - codeparrot_training - Step 14050: {'lr': 0.0004261972493784175, 'samples': 7194112, 'steps': 14050, 'loss/train': 2.370227575302124} 02/24/2022 19:46:02 - INFO - codeparrot_training - Step 14051: {'lr': 0.0004261856411926467, 'samples': 7194624, 'steps': 14051, 'loss/train': 2.1616947650909424} 02/24/2022 19:46:07 - INFO - codeparrot_training - Step 14052: {'lr': 0.0004261740322521525, 'samples': 7195136, 'steps': 14052, 'loss/train': 2.090914011001587} 02/24/2022 19:46:11 - INFO - codeparrot_training - Step 14053: {'lr': 0.00042616242255698463, 'samples': 7195648, 'steps': 14053, 'loss/train': 2.1879055500030518} 02/24/2022 19:46:18 - INFO - codeparrot_training - Step 14054: {'lr': 0.0004261508121071929, 'samples': 7196160, 'steps': 14054, 'loss/train': 1.4653798341751099} 02/24/2022 19:46:22 - INFO - codeparrot_training - Step 14055: {'lr': 0.00042613920090282706, 'samples': 7196672, 'steps': 14055, 'loss/train': 3.5175087451934814} 02/24/2022 19:46:27 - INFO - codeparrot_training - Step 14056: {'lr': 0.0004261275889439368, 'samples': 7197184, 'steps': 14056, 'loss/train': 1.1598094701766968} 02/24/2022 19:46:31 - INFO - codeparrot_training - Step 14057: {'lr': 0.0004261159762305719, 'samples': 7197696, 'steps': 14057, 'loss/train': 2.758617639541626} 02/24/2022 19:46:36 - INFO - codeparrot_training - Step 14058: {'lr': 0.00042610436276278196, 'samples': 7198208, 'steps': 14058, 'loss/train': 2.1537179946899414} 02/24/2022 19:46:40 - INFO - codeparrot_training - Step 14059: {'lr': 0.00042609274854061695, 'samples': 7198720, 'steps': 14059, 'loss/train': 6.597321510314941} 02/24/2022 19:46:45 - INFO - codeparrot_training - Step 14060: {'lr': 0.0004260811335641266, 'samples': 7199232, 'steps': 14060, 'loss/train': 2.0042102336883545} 02/24/2022 19:46:49 - INFO - codeparrot_training - Step 14061: {'lr': 0.00042606951783336045, 'samples': 7199744, 'steps': 14061, 'loss/train': 2.3713490962982178} 02/24/2022 19:46:54 - INFO - codeparrot_training - Step 14062: {'lr': 0.0004260579013483684, 'samples': 7200256, 'steps': 14062, 'loss/train': 0.5766487717628479} 02/24/2022 19:46:58 - INFO - codeparrot_training - Step 14063: {'lr': 0.0004260462841092003, 'samples': 7200768, 'steps': 14063, 'loss/train': 2.5429704189300537} 02/24/2022 19:47:05 - INFO - codeparrot_training - Step 14064: {'lr': 0.00042603466611590575, 'samples': 7201280, 'steps': 14064, 'loss/train': 1.9629775285720825} 02/24/2022 19:47:09 - INFO - codeparrot_training - Step 14065: {'lr': 0.00042602304736853464, 'samples': 7201792, 'steps': 14065, 'loss/train': 2.5161023139953613} 02/24/2022 19:47:14 - INFO - codeparrot_training - Step 14066: {'lr': 0.00042601142786713664, 'samples': 7202304, 'steps': 14066, 'loss/train': 0.9914633631706238} 02/24/2022 19:47:18 - INFO - codeparrot_training - Step 14067: {'lr': 0.0004259998076117616, 'samples': 7202816, 'steps': 14067, 'loss/train': 0.9146500825881958} 02/24/2022 19:47:23 - INFO - codeparrot_training - Step 14068: {'lr': 0.00042598818660245926, 'samples': 7203328, 'steps': 14068, 'loss/train': 2.3664193153381348} 02/24/2022 19:47:27 - INFO - codeparrot_training - Step 14069: {'lr': 0.00042597656483927936, 'samples': 7203840, 'steps': 14069, 'loss/train': 1.3827407360076904} 02/24/2022 19:47:32 - INFO - codeparrot_training - Step 14070: {'lr': 0.0004259649423222718, 'samples': 7204352, 'steps': 14070, 'loss/train': 2.288076639175415} 02/24/2022 19:47:36 - INFO - codeparrot_training - Step 14071: {'lr': 0.0004259533190514863, 'samples': 7204864, 'steps': 14071, 'loss/train': 2.987483263015747} 02/24/2022 19:47:41 - INFO - codeparrot_training - Step 14072: {'lr': 0.00042594169502697265, 'samples': 7205376, 'steps': 14072, 'loss/train': 2.6726582050323486} 02/24/2022 19:47:45 - INFO - codeparrot_training - Step 14073: {'lr': 0.0004259300702487806, 'samples': 7205888, 'steps': 14073, 'loss/train': 1.5380879640579224} 02/24/2022 19:47:52 - INFO - codeparrot_training - Step 14074: {'lr': 0.00042591844471696005, 'samples': 7206400, 'steps': 14074, 'loss/train': 1.6843796968460083} 02/24/2022 19:47:56 - INFO - codeparrot_training - Step 14075: {'lr': 0.00042590681843156073, 'samples': 7206912, 'steps': 14075, 'loss/train': 1.7758569717407227} 02/24/2022 19:48:01 - INFO - codeparrot_training - Step 14076: {'lr': 0.00042589519139263246, 'samples': 7207424, 'steps': 14076, 'loss/train': 1.8768200874328613} 02/24/2022 19:48:05 - INFO - codeparrot_training - Step 14077: {'lr': 0.0004258835636002251, 'samples': 7207936, 'steps': 14077, 'loss/train': 2.116335391998291} 02/24/2022 19:48:10 - INFO - codeparrot_training - Step 14078: {'lr': 0.0004258719350543883, 'samples': 7208448, 'steps': 14078, 'loss/train': 2.2044074535369873} 02/24/2022 19:48:14 - INFO - codeparrot_training - Step 14079: {'lr': 0.00042586030575517196, 'samples': 7208960, 'steps': 14079, 'loss/train': 1.9954184293746948} 02/24/2022 19:48:19 - INFO - codeparrot_training - Step 14080: {'lr': 0.00042584867570262595, 'samples': 7209472, 'steps': 14080, 'loss/train': 2.493478298187256} 02/24/2022 19:48:23 - INFO - codeparrot_training - Step 14081: {'lr': 0.00042583704489680007, 'samples': 7209984, 'steps': 14081, 'loss/train': 1.6662728786468506} 02/24/2022 19:48:29 - INFO - codeparrot_training - Step 14082: {'lr': 0.00042582541333774414, 'samples': 7210496, 'steps': 14082, 'loss/train': 1.7109266519546509} 02/24/2022 19:48:32 - INFO - codeparrot_training - Step 14083: {'lr': 0.0004258137810255079, 'samples': 7211008, 'steps': 14083, 'loss/train': 2.7853987216949463} 02/24/2022 19:48:38 - INFO - codeparrot_training - Step 14084: {'lr': 0.0004258021479601414, 'samples': 7211520, 'steps': 14084, 'loss/train': 1.8407634496688843} 02/24/2022 19:48:41 - INFO - codeparrot_training - Step 14085: {'lr': 0.00042579051414169417, 'samples': 7212032, 'steps': 14085, 'loss/train': 1.635709524154663} 02/24/2022 19:48:47 - INFO - codeparrot_training - Step 14086: {'lr': 0.0004257788795702162, 'samples': 7212544, 'steps': 14086, 'loss/train': 2.503282308578491} 02/24/2022 19:48:50 - INFO - codeparrot_training - Step 14087: {'lr': 0.0004257672442457574, 'samples': 7213056, 'steps': 14087, 'loss/train': 2.4399311542510986} 02/24/2022 19:48:56 - INFO - codeparrot_training - Step 14088: {'lr': 0.00042575560816836755, 'samples': 7213568, 'steps': 14088, 'loss/train': 2.224973201751709} 02/24/2022 19:48:59 - INFO - codeparrot_training - Step 14089: {'lr': 0.00042574397133809646, 'samples': 7214080, 'steps': 14089, 'loss/train': 2.33028244972229} 02/24/2022 19:49:07 - INFO - codeparrot_training - Step 14090: {'lr': 0.000425732333754994, 'samples': 7214592, 'steps': 14090, 'loss/train': 1.267633080482483} 02/24/2022 19:49:10 - INFO - codeparrot_training - Step 14091: {'lr': 0.00042572069541911, 'samples': 7215104, 'steps': 14091, 'loss/train': 1.4393713474273682} 02/24/2022 19:49:16 - INFO - codeparrot_training - Step 14092: {'lr': 0.0004257090563304943, 'samples': 7215616, 'steps': 14092, 'loss/train': 1.8586256504058838} 02/24/2022 19:49:19 - INFO - codeparrot_training - Step 14093: {'lr': 0.0004256974164891969, 'samples': 7216128, 'steps': 14093, 'loss/train': 2.453240156173706} 02/24/2022 19:49:25 - INFO - codeparrot_training - Step 14094: {'lr': 0.00042568577589526744, 'samples': 7216640, 'steps': 14094, 'loss/train': 3.355185031890869} 02/24/2022 19:49:28 - INFO - codeparrot_training - Step 14095: {'lr': 0.00042567413454875605, 'samples': 7217152, 'steps': 14095, 'loss/train': 2.238398790359497} 02/24/2022 19:49:35 - INFO - codeparrot_training - Step 14096: {'lr': 0.00042566249244971235, 'samples': 7217664, 'steps': 14096, 'loss/train': 1.4047503471374512} 02/24/2022 19:49:38 - INFO - codeparrot_training - Step 14097: {'lr': 0.0004256508495981863, 'samples': 7218176, 'steps': 14097, 'loss/train': 1.0151554346084595} 02/24/2022 19:49:41 - INFO - codeparrot_training - Step 14098: {'lr': 0.00042563920599422776, 'samples': 7218688, 'steps': 14098, 'loss/train': 1.2435219287872314} 02/24/2022 19:49:47 - INFO - codeparrot_training - Step 14099: {'lr': 0.00042562756163788673, 'samples': 7219200, 'steps': 14099, 'loss/train': 1.9987332820892334} 02/24/2022 19:49:50 - INFO - codeparrot_training - Step 14100: {'lr': 0.00042561591652921294, 'samples': 7219712, 'steps': 14100, 'loss/train': 1.8893072605133057} 02/24/2022 19:49:58 - INFO - codeparrot_training - Step 14101: {'lr': 0.00042560427066825636, 'samples': 7220224, 'steps': 14101, 'loss/train': 2.0572547912597656} 02/24/2022 19:50:01 - INFO - codeparrot_training - Step 14102: {'lr': 0.0004255926240550668, 'samples': 7220736, 'steps': 14102, 'loss/train': 1.5312286615371704} 02/24/2022 19:50:07 - INFO - codeparrot_training - Step 14103: {'lr': 0.0004255809766896942, 'samples': 7221248, 'steps': 14103, 'loss/train': 2.3437764644622803} 02/24/2022 19:50:10 - INFO - codeparrot_training - Step 14104: {'lr': 0.00042556932857218855, 'samples': 7221760, 'steps': 14104, 'loss/train': 1.6085742712020874} 02/24/2022 19:50:16 - INFO - codeparrot_training - Step 14105: {'lr': 0.0004255576797025995, 'samples': 7222272, 'steps': 14105, 'loss/train': 2.0601792335510254} 02/24/2022 19:50:19 - INFO - codeparrot_training - Step 14106: {'lr': 0.0004255460300809772, 'samples': 7222784, 'steps': 14106, 'loss/train': 2.894132375717163} 02/24/2022 19:50:25 - INFO - codeparrot_training - Step 14107: {'lr': 0.00042553437970737143, 'samples': 7223296, 'steps': 14107, 'loss/train': 1.829621434211731} 02/24/2022 19:50:28 - INFO - codeparrot_training - Step 14108: {'lr': 0.00042552272858183203, 'samples': 7223808, 'steps': 14108, 'loss/train': 1.725304365158081} 02/24/2022 19:50:34 - INFO - codeparrot_training - Step 14109: {'lr': 0.0004255110767044091, 'samples': 7224320, 'steps': 14109, 'loss/train': 2.5196967124938965} 02/24/2022 19:50:37 - INFO - codeparrot_training - Step 14110: {'lr': 0.0004254994240751524, 'samples': 7224832, 'steps': 14110, 'loss/train': 2.227799654006958} 02/24/2022 19:50:45 - INFO - codeparrot_training - Step 14111: {'lr': 0.00042548777069411194, 'samples': 7225344, 'steps': 14111, 'loss/train': 2.149305582046509} 02/24/2022 19:50:48 - INFO - codeparrot_training - Step 14112: {'lr': 0.0004254761165613375, 'samples': 7225856, 'steps': 14112, 'loss/train': 2.230198860168457} 02/24/2022 19:50:54 - INFO - codeparrot_training - Step 14113: {'lr': 0.00042546446167687914, 'samples': 7226368, 'steps': 14113, 'loss/train': 0.24831725656986237} 02/24/2022 19:50:57 - INFO - codeparrot_training - Step 14114: {'lr': 0.00042545280604078673, 'samples': 7226880, 'steps': 14114, 'loss/train': 1.4795970916748047} 02/24/2022 19:51:03 - INFO - codeparrot_training - Step 14115: {'lr': 0.0004254411496531103, 'samples': 7227392, 'steps': 14115, 'loss/train': 2.3489580154418945} 02/24/2022 19:51:08 - INFO - codeparrot_training - Step 14116: {'lr': 0.0004254294925138996, 'samples': 7227904, 'steps': 14116, 'loss/train': 2.704899549484253} 02/24/2022 19:51:12 - INFO - codeparrot_training - Step 14117: {'lr': 0.00042541783462320473, 'samples': 7228416, 'steps': 14117, 'loss/train': 0.9145331978797913} 02/24/2022 19:51:17 - INFO - codeparrot_training - Step 14118: {'lr': 0.00042540617598107544, 'samples': 7228928, 'steps': 14118, 'loss/train': 1.9691981077194214} 02/24/2022 19:51:21 - INFO - codeparrot_training - Step 14119: {'lr': 0.00042539451658756195, 'samples': 7229440, 'steps': 14119, 'loss/train': 1.7203046083450317} 02/24/2022 19:51:26 - INFO - codeparrot_training - Step 14120: {'lr': 0.000425382856442714, 'samples': 7229952, 'steps': 14120, 'loss/train': 1.2800098657608032} 02/24/2022 19:51:30 - INFO - codeparrot_training - Step 14121: {'lr': 0.0004253711955465815, 'samples': 7230464, 'steps': 14121, 'loss/train': 2.357043504714966} 02/24/2022 19:51:37 - INFO - codeparrot_training - Step 14122: {'lr': 0.00042535953389921454, 'samples': 7230976, 'steps': 14122, 'loss/train': 2.3500945568084717} 02/24/2022 19:51:40 - INFO - codeparrot_training - Step 14123: {'lr': 0.000425347871500663, 'samples': 7231488, 'steps': 14123, 'loss/train': 1.2270649671554565} 02/24/2022 19:51:46 - INFO - codeparrot_training - Step 14124: {'lr': 0.0004253362083509769, 'samples': 7232000, 'steps': 14124, 'loss/train': 2.0851705074310303} 02/24/2022 19:51:49 - INFO - codeparrot_training - Step 14125: {'lr': 0.0004253245444502061, 'samples': 7232512, 'steps': 14125, 'loss/train': 1.5507687330245972} 02/24/2022 19:51:55 - INFO - codeparrot_training - Step 14126: {'lr': 0.00042531287979840065, 'samples': 7233024, 'steps': 14126, 'loss/train': 2.677147626876831} 02/24/2022 19:51:58 - INFO - codeparrot_training - Step 14127: {'lr': 0.0004253012143956105, 'samples': 7233536, 'steps': 14127, 'loss/train': 2.2290821075439453} 02/24/2022 19:52:04 - INFO - codeparrot_training - Step 14128: {'lr': 0.0004252895482418856, 'samples': 7234048, 'steps': 14128, 'loss/train': 1.985878348350525} 02/24/2022 19:52:07 - INFO - codeparrot_training - Step 14129: {'lr': 0.00042527788133727595, 'samples': 7234560, 'steps': 14129, 'loss/train': 1.995582938194275} 02/24/2022 19:52:13 - INFO - codeparrot_training - Step 14130: {'lr': 0.0004252662136818315, 'samples': 7235072, 'steps': 14130, 'loss/train': 1.4644620418548584} 02/24/2022 19:52:17 - INFO - codeparrot_training - Step 14131: {'lr': 0.00042525454527560225, 'samples': 7235584, 'steps': 14131, 'loss/train': 1.3274897336959839} 02/24/2022 19:52:20 - INFO - codeparrot_training - Step 14132: {'lr': 0.0004252428761186382, 'samples': 7236096, 'steps': 14132, 'loss/train': 0.8161274194717407} 02/24/2022 19:52:26 - INFO - codeparrot_training - Step 14133: {'lr': 0.00042523120621098924, 'samples': 7236608, 'steps': 14133, 'loss/train': 2.4763269424438477} 02/24/2022 19:52:31 - INFO - codeparrot_training - Step 14134: {'lr': 0.0004252195355527055, 'samples': 7237120, 'steps': 14134, 'loss/train': 2.014841079711914} 02/24/2022 19:52:35 - INFO - codeparrot_training - Step 14135: {'lr': 0.0004252078641438369, 'samples': 7237632, 'steps': 14135, 'loss/train': 1.8089255094528198} 02/24/2022 19:52:42 - INFO - codeparrot_training - Step 14136: {'lr': 0.00042519619198443337, 'samples': 7238144, 'steps': 14136, 'loss/train': 2.2746827602386475} 02/24/2022 19:52:45 - INFO - codeparrot_training - Step 14137: {'lr': 0.0004251845190745451, 'samples': 7238656, 'steps': 14137, 'loss/train': 3.2400434017181396} 02/24/2022 19:52:51 - INFO - codeparrot_training - Step 14138: {'lr': 0.00042517284541422195, 'samples': 7239168, 'steps': 14138, 'loss/train': 2.6276497840881348} 02/24/2022 19:52:54 - INFO - codeparrot_training - Step 14139: {'lr': 0.00042516117100351394, 'samples': 7239680, 'steps': 14139, 'loss/train': 1.9705607891082764} 02/24/2022 19:53:00 - INFO - codeparrot_training - Step 14140: {'lr': 0.0004251494958424711, 'samples': 7240192, 'steps': 14140, 'loss/train': 1.336334466934204} 02/24/2022 19:53:03 - INFO - codeparrot_training - Step 14141: {'lr': 0.0004251378199311434, 'samples': 7240704, 'steps': 14141, 'loss/train': 2.1847548484802246} 02/24/2022 19:53:09 - INFO - codeparrot_training - Step 14142: {'lr': 0.0004251261432695809, 'samples': 7241216, 'steps': 14142, 'loss/train': 2.5474154949188232} 02/24/2022 19:53:12 - INFO - codeparrot_training - Step 14143: {'lr': 0.00042511446585783363, 'samples': 7241728, 'steps': 14143, 'loss/train': 2.114548921585083} 02/24/2022 19:53:18 - INFO - codeparrot_training - Step 14144: {'lr': 0.0004251027876959516, 'samples': 7242240, 'steps': 14144, 'loss/train': 3.132277727127075} 02/24/2022 19:53:21 - INFO - codeparrot_training - Step 14145: {'lr': 0.0004250911087839848, 'samples': 7242752, 'steps': 14145, 'loss/train': 2.365048408508301} 02/24/2022 19:53:29 - INFO - codeparrot_training - Step 14146: {'lr': 0.0004250794291219833, 'samples': 7243264, 'steps': 14146, 'loss/train': 2.6687538623809814} 02/24/2022 19:53:32 - INFO - codeparrot_training - Step 14147: {'lr': 0.00042506774870999716, 'samples': 7243776, 'steps': 14147, 'loss/train': 1.9452260732650757} 02/24/2022 19:53:38 - INFO - codeparrot_training - Step 14148: {'lr': 0.00042505606754807634, 'samples': 7244288, 'steps': 14148, 'loss/train': 2.4273617267608643} 02/24/2022 19:53:41 - INFO - codeparrot_training - Step 14149: {'lr': 0.00042504438563627093, 'samples': 7244800, 'steps': 14149, 'loss/train': 1.8435360193252563} 02/24/2022 19:53:47 - INFO - codeparrot_training - Step 14150: {'lr': 0.0004250327029746309, 'samples': 7245312, 'steps': 14150, 'loss/train': 2.2571325302124023} 02/24/2022 19:53:50 - INFO - codeparrot_training - Step 14151: {'lr': 0.0004250210195632064, 'samples': 7245824, 'steps': 14151, 'loss/train': 1.7857688665390015} 02/24/2022 19:53:56 - INFO - codeparrot_training - Step 14152: {'lr': 0.00042500933540204745, 'samples': 7246336, 'steps': 14152, 'loss/train': 2.543046474456787} 02/24/2022 19:53:59 - INFO - codeparrot_training - Step 14153: {'lr': 0.00042499765049120396, 'samples': 7246848, 'steps': 14153, 'loss/train': 1.796617865562439} 02/24/2022 19:54:05 - INFO - codeparrot_training - Step 14154: {'lr': 0.0004249859648307263, 'samples': 7247360, 'steps': 14154, 'loss/train': 2.202575445175171} 02/24/2022 19:54:08 - INFO - codeparrot_training - Step 14155: {'lr': 0.0004249742784206642, 'samples': 7247872, 'steps': 14155, 'loss/train': 1.7143428325653076} 02/24/2022 19:54:14 - INFO - codeparrot_training - Step 14156: {'lr': 0.00042496259126106786, 'samples': 7248384, 'steps': 14156, 'loss/train': 0.9983693957328796} 02/24/2022 19:54:18 - INFO - codeparrot_training - Step 14157: {'lr': 0.00042495090335198735, 'samples': 7248896, 'steps': 14157, 'loss/train': 3.1260788440704346} 02/24/2022 19:54:25 - INFO - codeparrot_training - Step 14158: {'lr': 0.0004249392146934726, 'samples': 7249408, 'steps': 14158, 'loss/train': 2.291740894317627} 02/24/2022 19:54:28 - INFO - codeparrot_training - Step 14159: {'lr': 0.000424927525285574, 'samples': 7249920, 'steps': 14159, 'loss/train': 2.944366931915283} 02/24/2022 19:54:34 - INFO - codeparrot_training - Step 14160: {'lr': 0.00042491583512834137, 'samples': 7250432, 'steps': 14160, 'loss/train': 2.893218994140625} 02/24/2022 19:54:37 - INFO - codeparrot_training - Step 14161: {'lr': 0.00042490414422182484, 'samples': 7250944, 'steps': 14161, 'loss/train': 2.6081125736236572} 02/24/2022 19:54:43 - INFO - codeparrot_training - Step 14162: {'lr': 0.00042489245256607447, 'samples': 7251456, 'steps': 14162, 'loss/train': 2.4035706520080566} 02/24/2022 19:54:46 - INFO - codeparrot_training - Step 14163: {'lr': 0.0004248807601611404, 'samples': 7251968, 'steps': 14163, 'loss/train': 2.323502779006958} 02/24/2022 19:54:52 - INFO - codeparrot_training - Step 14164: {'lr': 0.0004248690670070726, 'samples': 7252480, 'steps': 14164, 'loss/train': 0.11929647624492645} 02/24/2022 19:54:55 - INFO - codeparrot_training - Step 14165: {'lr': 0.00042485737310392135, 'samples': 7252992, 'steps': 14165, 'loss/train': 1.1046714782714844} 02/24/2022 19:55:01 - INFO - codeparrot_training - Step 14166: {'lr': 0.0004248456784517366, 'samples': 7253504, 'steps': 14166, 'loss/train': 1.9417747259140015} 02/24/2022 19:55:04 - INFO - codeparrot_training - Step 14167: {'lr': 0.00042483398305056847, 'samples': 7254016, 'steps': 14167, 'loss/train': 1.3783197402954102} 02/24/2022 19:55:12 - INFO - codeparrot_training - Step 14168: {'lr': 0.0004248222869004671, 'samples': 7254528, 'steps': 14168, 'loss/train': 2.4155406951904297} 02/24/2022 19:55:15 - INFO - codeparrot_training - Step 14169: {'lr': 0.00042481059000148253, 'samples': 7255040, 'steps': 14169, 'loss/train': 1.2719244956970215} 02/24/2022 19:55:21 - INFO - codeparrot_training - Step 14170: {'lr': 0.00042479889235366486, 'samples': 7255552, 'steps': 14170, 'loss/train': 1.524439811706543} 02/24/2022 19:55:24 - INFO - codeparrot_training - Step 14171: {'lr': 0.0004247871939570643, 'samples': 7256064, 'steps': 14171, 'loss/train': 1.6970051527023315} 02/24/2022 19:55:30 - INFO - codeparrot_training - Step 14172: {'lr': 0.00042477549481173093, 'samples': 7256576, 'steps': 14172, 'loss/train': 0.6546095609664917} 02/24/2022 19:55:34 - INFO - codeparrot_training - Step 14173: {'lr': 0.00042476379491771475, 'samples': 7257088, 'steps': 14173, 'loss/train': 2.554852247238159} 02/24/2022 19:55:37 - INFO - codeparrot_training - Step 14174: {'lr': 0.00042475209427506614, 'samples': 7257600, 'steps': 14174, 'loss/train': 2.718362331390381} 02/24/2022 19:55:43 - INFO - codeparrot_training - Step 14175: {'lr': 0.00042474039288383484, 'samples': 7258112, 'steps': 14175, 'loss/train': 1.319974660873413} 02/24/2022 19:55:46 - INFO - codeparrot_training - Step 14176: {'lr': 0.0004247286907440713, 'samples': 7258624, 'steps': 14176, 'loss/train': 1.9846550226211548} 02/24/2022 19:55:52 - INFO - codeparrot_training - Step 14177: {'lr': 0.00042471698785582546, 'samples': 7259136, 'steps': 14177, 'loss/train': 1.2104663848876953} 02/24/2022 19:55:57 - INFO - codeparrot_training - Step 14178: {'lr': 0.00042470528421914767, 'samples': 7259648, 'steps': 14178, 'loss/train': 1.5844451189041138} 02/24/2022 19:56:01 - INFO - codeparrot_training - Step 14179: {'lr': 0.0004246935798340877, 'samples': 7260160, 'steps': 14179, 'loss/train': 1.9999110698699951} 02/24/2022 19:56:06 - INFO - codeparrot_training - Step 14180: {'lr': 0.0004246818747006961, 'samples': 7260672, 'steps': 14180, 'loss/train': 0.9666487574577332} 02/24/2022 19:56:10 - INFO - codeparrot_training - Step 14181: {'lr': 0.0004246701688190227, 'samples': 7261184, 'steps': 14181, 'loss/train': 2.428150177001953} 02/24/2022 19:56:17 - INFO - codeparrot_training - Step 14182: {'lr': 0.0004246584621891179, 'samples': 7261696, 'steps': 14182, 'loss/train': 2.5712311267852783} 02/24/2022 19:56:21 - INFO - codeparrot_training - Step 14183: {'lr': 0.00042464675481103154, 'samples': 7262208, 'steps': 14183, 'loss/train': 0.8073933720588684} 02/24/2022 19:56:26 - INFO - codeparrot_training - Step 14184: {'lr': 0.00042463504668481403, 'samples': 7262720, 'steps': 14184, 'loss/train': 1.8930400609970093} 02/24/2022 19:56:30 - INFO - codeparrot_training - Step 14185: {'lr': 0.00042462333781051535, 'samples': 7263232, 'steps': 14185, 'loss/train': 1.7692681550979614} 02/24/2022 19:56:33 - INFO - codeparrot_training - Step 14186: {'lr': 0.00042461162818818585, 'samples': 7263744, 'steps': 14186, 'loss/train': 0.5340458750724792} 02/24/2022 19:56:39 - INFO - codeparrot_training - Step 14187: {'lr': 0.0004245999178178755, 'samples': 7264256, 'steps': 14187, 'loss/train': 1.3101933002471924} 02/24/2022 19:56:44 - INFO - codeparrot_training - Step 14188: {'lr': 0.0004245882066996346, 'samples': 7264768, 'steps': 14188, 'loss/train': 2.090367317199707} 02/24/2022 19:56:48 - INFO - codeparrot_training - Step 14189: {'lr': 0.0004245764948335132, 'samples': 7265280, 'steps': 14189, 'loss/train': 7.02206563949585} 02/24/2022 19:56:54 - INFO - codeparrot_training - Step 14190: {'lr': 0.0004245647822195616, 'samples': 7265792, 'steps': 14190, 'loss/train': 2.1099836826324463} 02/24/2022 19:56:57 - INFO - codeparrot_training - Step 14191: {'lr': 0.00042455306885782985, 'samples': 7266304, 'steps': 14191, 'loss/train': 2.362506866455078} 02/24/2022 19:57:02 - INFO - codeparrot_training - Step 14192: {'lr': 0.00042454135474836817, 'samples': 7266816, 'steps': 14192, 'loss/train': 1.3173161745071411} 02/24/2022 19:57:06 - INFO - codeparrot_training - Step 14193: {'lr': 0.00042452963989122685, 'samples': 7267328, 'steps': 14193, 'loss/train': 1.62032151222229} 02/24/2022 19:57:13 - INFO - codeparrot_training - Step 14194: {'lr': 0.00042451792428645587, 'samples': 7267840, 'steps': 14194, 'loss/train': 2.348407030105591} 02/24/2022 19:57:17 - INFO - codeparrot_training - Step 14195: {'lr': 0.0004245062079341055, 'samples': 7268352, 'steps': 14195, 'loss/train': 1.8568233251571655} 02/24/2022 19:57:22 - INFO - codeparrot_training - Step 14196: {'lr': 0.000424494490834226, 'samples': 7268864, 'steps': 14196, 'loss/train': 1.1280829906463623} 02/24/2022 19:57:26 - INFO - codeparrot_training - Step 14197: {'lr': 0.0004244827729868675, 'samples': 7269376, 'steps': 14197, 'loss/train': 0.7390331625938416} 02/24/2022 19:57:31 - INFO - codeparrot_training - Step 14198: {'lr': 0.00042447105439208024, 'samples': 7269888, 'steps': 14198, 'loss/train': 1.3491523265838623} 02/24/2022 19:57:35 - INFO - codeparrot_training - Step 14199: {'lr': 0.0004244593350499143, 'samples': 7270400, 'steps': 14199, 'loss/train': 1.9855657815933228} 02/24/2022 19:57:40 - INFO - codeparrot_training - Step 14200: {'lr': 0.00042444761496042004, 'samples': 7270912, 'steps': 14200, 'loss/train': 2.601789951324463} 02/24/2022 19:57:44 - INFO - codeparrot_training - Step 14201: {'lr': 0.0004244358941236476, 'samples': 7271424, 'steps': 14201, 'loss/train': 1.7302453517913818} 02/24/2022 19:57:49 - INFO - codeparrot_training - Step 14202: {'lr': 0.00042442417253964713, 'samples': 7271936, 'steps': 14202, 'loss/train': 1.6395988464355469} 02/24/2022 19:57:53 - INFO - codeparrot_training - Step 14203: {'lr': 0.00042441245020846885, 'samples': 7272448, 'steps': 14203, 'loss/train': 1.1476298570632935} 02/24/2022 19:58:00 - INFO - codeparrot_training - Step 14204: {'lr': 0.00042440072713016317, 'samples': 7272960, 'steps': 14204, 'loss/train': 1.3235162496566772} 02/24/2022 19:58:04 - INFO - codeparrot_training - Step 14205: {'lr': 0.00042438900330478, 'samples': 7273472, 'steps': 14205, 'loss/train': 0.17766611278057098} 02/24/2022 19:58:09 - INFO - codeparrot_training - Step 14206: {'lr': 0.00042437727873236974, 'samples': 7273984, 'steps': 14206, 'loss/train': 1.7074435949325562} 02/24/2022 19:58:13 - INFO - codeparrot_training - Step 14207: {'lr': 0.00042436555341298266, 'samples': 7274496, 'steps': 14207, 'loss/train': 2.560800790786743} 02/24/2022 19:58:18 - INFO - codeparrot_training - Step 14208: {'lr': 0.0004243538273466689, 'samples': 7275008, 'steps': 14208, 'loss/train': 2.2613449096679688} 02/24/2022 19:58:22 - INFO - codeparrot_training - Step 14209: {'lr': 0.00042434210053347865, 'samples': 7275520, 'steps': 14209, 'loss/train': 1.077513337135315} 02/24/2022 19:58:27 - INFO - codeparrot_training - Step 14210: {'lr': 0.0004243303729734622, 'samples': 7276032, 'steps': 14210, 'loss/train': 0.9633165001869202} 02/24/2022 19:58:31 - INFO - codeparrot_training - Step 14211: {'lr': 0.0004243186446666699, 'samples': 7276544, 'steps': 14211, 'loss/train': 1.5966123342514038} 02/24/2022 19:58:36 - INFO - codeparrot_training - Step 14212: {'lr': 0.00042430691561315176, 'samples': 7277056, 'steps': 14212, 'loss/train': 1.075243592262268} 02/24/2022 19:58:40 - INFO - codeparrot_training - Step 14213: {'lr': 0.0004242951858129582, 'samples': 7277568, 'steps': 14213, 'loss/train': 2.289094924926758} 02/24/2022 19:58:47 - INFO - codeparrot_training - Step 14214: {'lr': 0.0004242834552661394, 'samples': 7278080, 'steps': 14214, 'loss/train': 1.7578402757644653} 02/24/2022 19:58:50 - INFO - codeparrot_training - Step 14215: {'lr': 0.0004242717239727456, 'samples': 7278592, 'steps': 14215, 'loss/train': 1.8033943176269531} 02/24/2022 19:58:56 - INFO - codeparrot_training - Step 14216: {'lr': 0.00042425999193282713, 'samples': 7279104, 'steps': 14216, 'loss/train': 1.9938198328018188} 02/24/2022 19:58:59 - INFO - codeparrot_training - Step 14217: {'lr': 0.0004242482591464342, 'samples': 7279616, 'steps': 14217, 'loss/train': 1.4476081132888794} 02/24/2022 19:59:05 - INFO - codeparrot_training - Step 14218: {'lr': 0.0004242365256136169, 'samples': 7280128, 'steps': 14218, 'loss/train': 1.5051683187484741} 02/24/2022 19:59:09 - INFO - codeparrot_training - Step 14219: {'lr': 0.00042422479133442573, 'samples': 7280640, 'steps': 14219, 'loss/train': 2.4768805503845215} 02/24/2022 19:59:14 - INFO - codeparrot_training - Step 14220: {'lr': 0.00042421305630891093, 'samples': 7281152, 'steps': 14220, 'loss/train': 2.524547576904297} 02/24/2022 19:59:17 - INFO - codeparrot_training - Step 14221: {'lr': 0.0004242013205371227, 'samples': 7281664, 'steps': 14221, 'loss/train': 2.484830856323242} 02/24/2022 19:59:23 - INFO - codeparrot_training - Step 14222: {'lr': 0.00042418958401911134, 'samples': 7282176, 'steps': 14222, 'loss/train': 2.2180001735687256} 02/24/2022 19:59:27 - INFO - codeparrot_training - Step 14223: {'lr': 0.000424177846754927, 'samples': 7282688, 'steps': 14223, 'loss/train': 2.3699991703033447} 02/24/2022 19:59:32 - INFO - codeparrot_training - Step 14224: {'lr': 0.0004241661087446202, 'samples': 7283200, 'steps': 14224, 'loss/train': 1.6239110231399536} 02/24/2022 19:59:36 - INFO - codeparrot_training - Step 14225: {'lr': 0.00042415436998824105, 'samples': 7283712, 'steps': 14225, 'loss/train': 2.4087109565734863} 02/24/2022 19:59:41 - INFO - codeparrot_training - Step 14226: {'lr': 0.0004241426304858399, 'samples': 7284224, 'steps': 14226, 'loss/train': 2.291109561920166} 02/24/2022 19:59:45 - INFO - codeparrot_training - Step 14227: {'lr': 0.00042413089023746696, 'samples': 7284736, 'steps': 14227, 'loss/train': 2.7332041263580322} 02/24/2022 19:59:50 - INFO - codeparrot_training - Step 14228: {'lr': 0.00042411914924317265, 'samples': 7285248, 'steps': 14228, 'loss/train': 0.664466381072998} 02/24/2022 19:59:54 - INFO - codeparrot_training - Step 14229: {'lr': 0.00042410740750300715, 'samples': 7285760, 'steps': 14229, 'loss/train': 1.952970266342163} 02/24/2022 20:00:01 - INFO - codeparrot_training - Step 14230: {'lr': 0.0004240956650170208, 'samples': 7286272, 'steps': 14230, 'loss/train': 2.496864080429077} 02/24/2022 20:00:04 - INFO - codeparrot_training - Step 14231: {'lr': 0.00042408392178526396, 'samples': 7286784, 'steps': 14231, 'loss/train': 2.1443912982940674} 02/24/2022 20:00:10 - INFO - codeparrot_training - Step 14232: {'lr': 0.0004240721778077868, 'samples': 7287296, 'steps': 14232, 'loss/train': 2.285691022872925} 02/24/2022 20:00:13 - INFO - codeparrot_training - Step 14233: {'lr': 0.0004240604330846397, 'samples': 7287808, 'steps': 14233, 'loss/train': 2.256314277648926} 02/24/2022 20:00:19 - INFO - codeparrot_training - Step 14234: {'lr': 0.000424048687615873, 'samples': 7288320, 'steps': 14234, 'loss/train': 2.262995958328247} 02/24/2022 20:00:22 - INFO - codeparrot_training - Step 14235: {'lr': 0.00042403694140153705, 'samples': 7288832, 'steps': 14235, 'loss/train': 1.2854812145233154} 02/24/2022 20:00:28 - INFO - codeparrot_training - Step 14236: {'lr': 0.00042402519444168207, 'samples': 7289344, 'steps': 14236, 'loss/train': 1.9728502035140991} 02/24/2022 20:00:31 - INFO - codeparrot_training - Step 14237: {'lr': 0.00042401344673635846, 'samples': 7289856, 'steps': 14237, 'loss/train': 2.1000592708587646} 02/24/2022 20:00:37 - INFO - codeparrot_training - Step 14238: {'lr': 0.00042400169828561636, 'samples': 7290368, 'steps': 14238, 'loss/train': 0.6835358738899231} 02/24/2022 20:00:40 - INFO - codeparrot_training - Step 14239: {'lr': 0.0004239899490895063, 'samples': 7290880, 'steps': 14239, 'loss/train': 2.3722028732299805} 02/24/2022 20:00:48 - INFO - codeparrot_training - Step 14240: {'lr': 0.00042397819914807855, 'samples': 7291392, 'steps': 14240, 'loss/train': 2.6693899631500244} 02/24/2022 20:00:51 - INFO - codeparrot_training - Step 14241: {'lr': 0.00042396644846138355, 'samples': 7291904, 'steps': 14241, 'loss/train': 1.8931578397750854} 02/24/2022 20:00:57 - INFO - codeparrot_training - Step 14242: {'lr': 0.00042395469702947135, 'samples': 7292416, 'steps': 14242, 'loss/train': 0.32943904399871826} 02/24/2022 20:01:00 - INFO - codeparrot_training - Step 14243: {'lr': 0.0004239429448523925, 'samples': 7292928, 'steps': 14243, 'loss/train': 1.087774634361267} 02/24/2022 20:01:06 - INFO - codeparrot_training - Step 14244: {'lr': 0.00042393119193019743, 'samples': 7293440, 'steps': 14244, 'loss/train': 1.0334482192993164} 02/24/2022 20:01:09 - INFO - codeparrot_training - Step 14245: {'lr': 0.00042391943826293623, 'samples': 7293952, 'steps': 14245, 'loss/train': 0.9405443072319031} 02/24/2022 20:01:15 - INFO - codeparrot_training - Step 14246: {'lr': 0.0004239076838506595, 'samples': 7294464, 'steps': 14246, 'loss/train': 0.5584782958030701} 02/24/2022 20:01:18 - INFO - codeparrot_training - Step 14247: {'lr': 0.0004238959286934174, 'samples': 7294976, 'steps': 14247, 'loss/train': 2.3326168060302734} 02/24/2022 20:01:23 - INFO - codeparrot_training - Step 14248: {'lr': 0.0004238841727912603, 'samples': 7295488, 'steps': 14248, 'loss/train': 1.9053739309310913} 02/24/2022 20:01:27 - INFO - codeparrot_training - Step 14249: {'lr': 0.00042387241614423875, 'samples': 7296000, 'steps': 14249, 'loss/train': 2.0522050857543945} 02/24/2022 20:01:34 - INFO - codeparrot_training - Step 14250: {'lr': 0.0004238606587524029, 'samples': 7296512, 'steps': 14250, 'loss/train': 2.627589464187622} 02/24/2022 20:01:38 - INFO - codeparrot_training - Step 14251: {'lr': 0.0004238489006158033, 'samples': 7297024, 'steps': 14251, 'loss/train': 2.3849539756774902} 02/24/2022 20:01:43 - INFO - codeparrot_training - Step 14252: {'lr': 0.00042383714173449007, 'samples': 7297536, 'steps': 14252, 'loss/train': 1.99263334274292} 02/24/2022 20:01:47 - INFO - codeparrot_training - Step 14253: {'lr': 0.0004238253821085138, 'samples': 7298048, 'steps': 14253, 'loss/train': 1.7893986701965332} 02/24/2022 20:01:52 - INFO - codeparrot_training - Step 14254: {'lr': 0.00042381362173792475, 'samples': 7298560, 'steps': 14254, 'loss/train': 1.9920262098312378} 02/24/2022 20:01:56 - INFO - codeparrot_training - Step 14255: {'lr': 0.00042380186062277337, 'samples': 7299072, 'steps': 14255, 'loss/train': 1.7049756050109863} 02/24/2022 20:02:02 - INFO - codeparrot_training - Step 14256: {'lr': 0.00042379009876311, 'samples': 7299584, 'steps': 14256, 'loss/train': 2.2791190147399902} 02/24/2022 20:02:06 - INFO - codeparrot_training - Step 14257: {'lr': 0.00042377833615898496, 'samples': 7300096, 'steps': 14257, 'loss/train': 0.9907761216163635} 02/24/2022 20:02:09 - INFO - codeparrot_training - Step 14258: {'lr': 0.0004237665728104488, 'samples': 7300608, 'steps': 14258, 'loss/train': 1.5103892087936401} 02/24/2022 20:02:14 - INFO - codeparrot_training - Step 14259: {'lr': 0.0004237548087175518, 'samples': 7301120, 'steps': 14259, 'loss/train': 1.8606855869293213} 02/24/2022 20:02:18 - INFO - codeparrot_training - Step 14260: {'lr': 0.00042374304388034437, 'samples': 7301632, 'steps': 14260, 'loss/train': 1.6345964670181274} 02/24/2022 20:02:25 - INFO - codeparrot_training - Step 14261: {'lr': 0.00042373127829887694, 'samples': 7302144, 'steps': 14261, 'loss/train': 1.6888474225997925} 02/24/2022 20:02:29 - INFO - codeparrot_training - Step 14262: {'lr': 0.0004237195119731998, 'samples': 7302656, 'steps': 14262, 'loss/train': 3.3613123893737793} 02/24/2022 20:02:34 - INFO - codeparrot_training - Step 14263: {'lr': 0.0004237077449033635, 'samples': 7303168, 'steps': 14263, 'loss/train': 1.1451348066329956} 02/24/2022 20:02:38 - INFO - codeparrot_training - Step 14264: {'lr': 0.0004236959770894183, 'samples': 7303680, 'steps': 14264, 'loss/train': 2.098689317703247} 02/24/2022 20:02:43 - INFO - codeparrot_training - Step 14265: {'lr': 0.0004236842085314148, 'samples': 7304192, 'steps': 14265, 'loss/train': 1.7305004596710205} 02/24/2022 20:02:47 - INFO - codeparrot_training - Step 14266: {'lr': 0.0004236724392294032, 'samples': 7304704, 'steps': 14266, 'loss/train': 2.318592071533203} 02/24/2022 20:02:52 - INFO - codeparrot_training - Step 14267: {'lr': 0.0004236606691834341, 'samples': 7305216, 'steps': 14267, 'loss/train': 1.944257378578186} 02/24/2022 20:02:56 - INFO - codeparrot_training - Step 14268: {'lr': 0.0004236488983935578, 'samples': 7305728, 'steps': 14268, 'loss/train': 2.0166215896606445} 02/24/2022 20:03:01 - INFO - codeparrot_training - Step 14269: {'lr': 0.0004236371268598248, 'samples': 7306240, 'steps': 14269, 'loss/train': 2.130969524383545} 02/24/2022 20:03:05 - INFO - codeparrot_training - Step 14270: {'lr': 0.0004236253545822855, 'samples': 7306752, 'steps': 14270, 'loss/train': 1.1413593292236328} 02/24/2022 20:03:10 - INFO - codeparrot_training - Step 14271: {'lr': 0.00042361358156099016, 'samples': 7307264, 'steps': 14271, 'loss/train': 2.35860538482666} 02/24/2022 20:03:14 - INFO - codeparrot_training - Step 14272: {'lr': 0.0004236018077959895, 'samples': 7307776, 'steps': 14272, 'loss/train': 1.9288330078125} 02/24/2022 20:03:19 - INFO - codeparrot_training - Step 14273: {'lr': 0.00042359003328733383, 'samples': 7308288, 'steps': 14273, 'loss/train': 1.3647397756576538} 02/24/2022 20:03:23 - INFO - codeparrot_training - Step 14274: {'lr': 0.0004235782580350734, 'samples': 7308800, 'steps': 14274, 'loss/train': 2.0449132919311523} 02/24/2022 20:03:28 - INFO - codeparrot_training - Step 14275: {'lr': 0.0004235664820392591, 'samples': 7309312, 'steps': 14275, 'loss/train': 1.743349552154541} 02/24/2022 20:03:36 - INFO - codeparrot_training - Step 14276: {'lr': 0.0004235547052999409, 'samples': 7309824, 'steps': 14276, 'loss/train': 1.7291473150253296} 02/24/2022 20:03:39 - INFO - codeparrot_training - Step 14277: {'lr': 0.0004235429278171695, 'samples': 7310336, 'steps': 14277, 'loss/train': 2.043515682220459} 02/24/2022 20:03:43 - INFO - codeparrot_training - Step 14278: {'lr': 0.00042353114959099535, 'samples': 7310848, 'steps': 14278, 'loss/train': 1.9309711456298828} 02/24/2022 20:03:48 - INFO - codeparrot_training - Step 14279: {'lr': 0.0004235193706214688, 'samples': 7311360, 'steps': 14279, 'loss/train': 2.1979947090148926} 02/24/2022 20:03:52 - INFO - codeparrot_training - Step 14280: {'lr': 0.00042350759090864043, 'samples': 7311872, 'steps': 14280, 'loss/train': 1.9417647123336792} 02/24/2022 20:03:57 - INFO - codeparrot_training - Step 14281: {'lr': 0.00042349581045256055, 'samples': 7312384, 'steps': 14281, 'loss/train': 2.2637672424316406} 02/24/2022 20:04:01 - INFO - codeparrot_training - Step 14282: {'lr': 0.00042348402925327977, 'samples': 7312896, 'steps': 14282, 'loss/train': 2.0324196815490723} 02/24/2022 20:04:06 - INFO - codeparrot_training - Step 14283: {'lr': 0.00042347224731084854, 'samples': 7313408, 'steps': 14283, 'loss/train': 2.368499279022217} 02/24/2022 20:04:10 - INFO - codeparrot_training - Step 14284: {'lr': 0.0004234604646253172, 'samples': 7313920, 'steps': 14284, 'loss/train': 2.141674041748047} 02/24/2022 20:04:15 - INFO - codeparrot_training - Step 14285: {'lr': 0.0004234486811967364, 'samples': 7314432, 'steps': 14285, 'loss/train': 1.910029411315918} 02/24/2022 20:04:19 - INFO - codeparrot_training - Step 14286: {'lr': 0.00042343689702515643, 'samples': 7314944, 'steps': 14286, 'loss/train': 2.3125417232513428} 02/24/2022 20:04:26 - INFO - codeparrot_training - Step 14287: {'lr': 0.0004234251121106279, 'samples': 7315456, 'steps': 14287, 'loss/train': 1.4323766231536865} 02/24/2022 20:04:29 - INFO - codeparrot_training - Step 14288: {'lr': 0.00042341332645320126, 'samples': 7315968, 'steps': 14288, 'loss/train': 3.8594212532043457} 02/24/2022 20:04:35 - INFO - codeparrot_training - Step 14289: {'lr': 0.000423401540052927, 'samples': 7316480, 'steps': 14289, 'loss/train': 1.9006260633468628} 02/24/2022 20:04:38 - INFO - codeparrot_training - Step 14290: {'lr': 0.0004233897529098556, 'samples': 7316992, 'steps': 14290, 'loss/train': 1.530927300453186} 02/24/2022 20:04:44 - INFO - codeparrot_training - Step 14291: {'lr': 0.0004233779650240376, 'samples': 7317504, 'steps': 14291, 'loss/train': 0.7660928964614868} 02/24/2022 20:04:47 - INFO - codeparrot_training - Step 14292: {'lr': 0.00042336617639552335, 'samples': 7318016, 'steps': 14292, 'loss/train': 2.562587022781372} 02/24/2022 20:04:53 - INFO - codeparrot_training - Step 14293: {'lr': 0.00042335438702436354, 'samples': 7318528, 'steps': 14293, 'loss/train': 2.1339643001556396} 02/24/2022 20:04:56 - INFO - codeparrot_training - Step 14294: {'lr': 0.0004233425969106085, 'samples': 7319040, 'steps': 14294, 'loss/train': 0.5550722479820251} 02/24/2022 20:05:02 - INFO - codeparrot_training - Step 14295: {'lr': 0.00042333080605430883, 'samples': 7319552, 'steps': 14295, 'loss/train': 2.4108386039733887} 02/24/2022 20:05:10 - INFO - codeparrot_training - Step 14296: {'lr': 0.00042331901445551514, 'samples': 7320064, 'steps': 14296, 'loss/train': 1.6498088836669922} 02/24/2022 20:05:13 - INFO - codeparrot_training - Step 14297: {'lr': 0.00042330722211427775, 'samples': 7320576, 'steps': 14297, 'loss/train': 1.024876594543457} 02/24/2022 20:05:19 - INFO - codeparrot_training - Step 14298: {'lr': 0.00042329542903064724, 'samples': 7321088, 'steps': 14298, 'loss/train': 1.3504054546356201} 02/24/2022 20:05:22 - INFO - codeparrot_training - Step 14299: {'lr': 0.00042328363520467417, 'samples': 7321600, 'steps': 14299, 'loss/train': 1.8318637609481812} 02/24/2022 20:05:28 - INFO - codeparrot_training - Step 14300: {'lr': 0.000423271840636409, 'samples': 7322112, 'steps': 14300, 'loss/train': 1.520402193069458} 02/24/2022 20:05:31 - INFO - codeparrot_training - Step 14301: {'lr': 0.0004232600453259023, 'samples': 7322624, 'steps': 14301, 'loss/train': 2.715498208999634} 02/24/2022 20:05:37 - INFO - codeparrot_training - Step 14302: {'lr': 0.0004232482492732046, 'samples': 7323136, 'steps': 14302, 'loss/train': 2.625418186187744} 02/24/2022 20:05:40 - INFO - codeparrot_training - Step 14303: {'lr': 0.00042323645247836636, 'samples': 7323648, 'steps': 14303, 'loss/train': 1.2860190868377686} 02/24/2022 20:05:46 - INFO - codeparrot_training - Step 14304: {'lr': 0.00042322465494143814, 'samples': 7324160, 'steps': 14304, 'loss/train': 0.4741033613681793} 02/24/2022 20:05:49 - INFO - codeparrot_training - Step 14305: {'lr': 0.00042321285666247063, 'samples': 7324672, 'steps': 14305, 'loss/train': 2.146327257156372} 02/24/2022 20:05:57 - INFO - codeparrot_training - Step 14306: {'lr': 0.0004232010576415141, 'samples': 7325184, 'steps': 14306, 'loss/train': 1.637241005897522} 02/24/2022 20:06:00 - INFO - codeparrot_training - Step 14307: {'lr': 0.00042318925787861937, 'samples': 7325696, 'steps': 14307, 'loss/train': 2.294499158859253} 02/24/2022 20:06:06 - INFO - codeparrot_training - Step 14308: {'lr': 0.0004231774573738367, 'samples': 7326208, 'steps': 14308, 'loss/train': 2.5879604816436768} 02/24/2022 20:06:09 - INFO - codeparrot_training - Step 14309: {'lr': 0.000423165656127217, 'samples': 7326720, 'steps': 14309, 'loss/train': 1.6816283464431763} 02/24/2022 20:06:15 - INFO - codeparrot_training - Step 14310: {'lr': 0.00042315385413881047, 'samples': 7327232, 'steps': 14310, 'loss/train': 1.2102160453796387} 02/24/2022 20:06:18 - INFO - codeparrot_training - Step 14311: {'lr': 0.00042314205140866785, 'samples': 7327744, 'steps': 14311, 'loss/train': 2.049474000930786} 02/24/2022 20:06:24 - INFO - codeparrot_training - Step 14312: {'lr': 0.00042313024793683965, 'samples': 7328256, 'steps': 14312, 'loss/train': 1.9472191333770752} 02/24/2022 20:06:27 - INFO - codeparrot_training - Step 14313: {'lr': 0.0004231184437233765, 'samples': 7328768, 'steps': 14313, 'loss/train': 1.5723084211349487} 02/24/2022 20:06:33 - INFO - codeparrot_training - Step 14314: {'lr': 0.0004231066387683288, 'samples': 7329280, 'steps': 14314, 'loss/train': 1.6288281679153442} 02/24/2022 20:06:36 - INFO - codeparrot_training - Step 14315: {'lr': 0.0004230948330717472, 'samples': 7329792, 'steps': 14315, 'loss/train': 0.5535238981246948} 02/24/2022 20:06:42 - INFO - codeparrot_training - Step 14316: {'lr': 0.0004230830266336825, 'samples': 7330304, 'steps': 14316, 'loss/train': 0.5326436758041382} 02/24/2022 20:06:45 - INFO - codeparrot_training - Step 14317: {'lr': 0.00042307121945418493, 'samples': 7330816, 'steps': 14317, 'loss/train': 2.1080830097198486} 02/24/2022 20:06:51 - INFO - codeparrot_training - Step 14318: {'lr': 0.00042305941153330525, 'samples': 7331328, 'steps': 14318, 'loss/train': 2.036498785018921} 02/24/2022 20:06:54 - INFO - codeparrot_training - Step 14319: {'lr': 0.00042304760287109394, 'samples': 7331840, 'steps': 14319, 'loss/train': 2.572666645050049} 02/24/2022 20:06:59 - INFO - codeparrot_training - Step 14320: {'lr': 0.0004230357934676017, 'samples': 7332352, 'steps': 14320, 'loss/train': 2.471630334854126} 02/24/2022 20:07:03 - INFO - codeparrot_training - Step 14321: {'lr': 0.00042302398332287903, 'samples': 7332864, 'steps': 14321, 'loss/train': 1.4810775518417358} 02/24/2022 20:07:10 - INFO - codeparrot_training - Step 14322: {'lr': 0.00042301217243697665, 'samples': 7333376, 'steps': 14322, 'loss/train': 1.4826350212097168} 02/24/2022 20:07:14 - INFO - codeparrot_training - Step 14323: {'lr': 0.00042300036080994495, 'samples': 7333888, 'steps': 14323, 'loss/train': 2.3870391845703125} 02/24/2022 20:07:19 - INFO - codeparrot_training - Step 14324: {'lr': 0.00042298854844183476, 'samples': 7334400, 'steps': 14324, 'loss/train': 2.327512741088867} 02/24/2022 20:07:23 - INFO - codeparrot_training - Step 14325: {'lr': 0.0004229767353326964, 'samples': 7334912, 'steps': 14325, 'loss/train': 2.4142353534698486} 02/24/2022 20:07:28 - INFO - codeparrot_training - Step 14326: {'lr': 0.0004229649214825808, 'samples': 7335424, 'steps': 14326, 'loss/train': 1.994194746017456} 02/24/2022 20:07:32 - INFO - codeparrot_training - Step 14327: {'lr': 0.0004229531068915383, 'samples': 7335936, 'steps': 14327, 'loss/train': 1.6865442991256714} 02/24/2022 20:07:37 - INFO - codeparrot_training - Step 14328: {'lr': 0.0004229412915596196, 'samples': 7336448, 'steps': 14328, 'loss/train': 1.3904539346694946} 02/24/2022 20:07:41 - INFO - codeparrot_training - Step 14329: {'lr': 0.0004229294754868754, 'samples': 7336960, 'steps': 14329, 'loss/train': 1.4125627279281616} 02/24/2022 20:07:46 - INFO - codeparrot_training - Step 14330: {'lr': 0.0004229176586733562, 'samples': 7337472, 'steps': 14330, 'loss/train': 2.1270740032196045} 02/24/2022 20:07:50 - INFO - codeparrot_training - Step 14331: {'lr': 0.0004229058411191126, 'samples': 7337984, 'steps': 14331, 'loss/train': 2.4757940769195557} 02/24/2022 20:07:58 - INFO - codeparrot_training - Step 14332: {'lr': 0.0004228940228241953, 'samples': 7338496, 'steps': 14332, 'loss/train': 2.683319568634033} 02/24/2022 20:08:01 - INFO - codeparrot_training - Step 14333: {'lr': 0.0004228822037886549, 'samples': 7339008, 'steps': 14333, 'loss/train': 2.4039764404296875} 02/24/2022 20:08:05 - INFO - codeparrot_training - Step 14334: {'lr': 0.00042287038401254214, 'samples': 7339520, 'steps': 14334, 'loss/train': 2.4975473880767822} 02/24/2022 20:08:10 - INFO - codeparrot_training - Step 14335: {'lr': 0.00042285856349590746, 'samples': 7340032, 'steps': 14335, 'loss/train': 1.7661595344543457} 02/24/2022 20:08:14 - INFO - codeparrot_training - Step 14336: {'lr': 0.0004228467422388016, 'samples': 7340544, 'steps': 14336, 'loss/train': 2.0884792804718018} 02/24/2022 20:08:19 - INFO - codeparrot_training - Step 14337: {'lr': 0.00042283492024127524, 'samples': 7341056, 'steps': 14337, 'loss/train': 1.9482609033584595} 02/24/2022 20:08:23 - INFO - codeparrot_training - Step 14338: {'lr': 0.00042282309750337887, 'samples': 7341568, 'steps': 14338, 'loss/train': 2.5940208435058594} 02/24/2022 20:08:28 - INFO - codeparrot_training - Step 14339: {'lr': 0.0004228112740251632, 'samples': 7342080, 'steps': 14339, 'loss/train': 0.18865668773651123} 02/24/2022 20:08:32 - INFO - codeparrot_training - Step 14340: {'lr': 0.00042279944980667906, 'samples': 7342592, 'steps': 14340, 'loss/train': 0.9635124206542969} 02/24/2022 20:08:37 - INFO - codeparrot_training - Step 14341: {'lr': 0.00042278762484797684, 'samples': 7343104, 'steps': 14341, 'loss/train': 1.851822853088379} 02/24/2022 20:08:41 - INFO - codeparrot_training - Step 14342: {'lr': 0.0004227757991491073, 'samples': 7343616, 'steps': 14342, 'loss/train': 2.107936143875122} 02/24/2022 20:08:48 - INFO - codeparrot_training - Step 14343: {'lr': 0.0004227639727101211, 'samples': 7344128, 'steps': 14343, 'loss/train': 2.010317087173462} 02/24/2022 20:08:52 - INFO - codeparrot_training - Step 14344: {'lr': 0.0004227521455310689, 'samples': 7344640, 'steps': 14344, 'loss/train': 2.015282154083252} 02/24/2022 20:08:57 - INFO - codeparrot_training - Step 14345: {'lr': 0.0004227403176120014, 'samples': 7345152, 'steps': 14345, 'loss/train': 1.9161802530288696} 02/24/2022 20:09:01 - INFO - codeparrot_training - Step 14346: {'lr': 0.00042272848895296924, 'samples': 7345664, 'steps': 14346, 'loss/train': 1.7873947620391846} 02/24/2022 20:09:06 - INFO - codeparrot_training - Step 14347: {'lr': 0.000422716659554023, 'samples': 7346176, 'steps': 14347, 'loss/train': 2.7401559352874756} 02/24/2022 20:09:09 - INFO - codeparrot_training - Step 14348: {'lr': 0.00042270482941521347, 'samples': 7346688, 'steps': 14348, 'loss/train': 1.6684261560440063} 02/24/2022 20:09:15 - INFO - codeparrot_training - Step 14349: {'lr': 0.0004226929985365913, 'samples': 7347200, 'steps': 14349, 'loss/train': 2.3258559703826904} 02/24/2022 20:09:21 - INFO - codeparrot_training - Step 14350: {'lr': 0.00042268116691820723, 'samples': 7347712, 'steps': 14350, 'loss/train': 1.7296404838562012} 02/24/2022 20:09:24 - INFO - codeparrot_training - Step 14351: {'lr': 0.00042266933456011174, 'samples': 7348224, 'steps': 14351, 'loss/train': 2.2408125400543213} 02/24/2022 20:09:31 - INFO - codeparrot_training - Step 14352: {'lr': 0.0004226575014623557, 'samples': 7348736, 'steps': 14352, 'loss/train': 2.1970057487487793} 02/24/2022 20:09:35 - INFO - codeparrot_training - Step 14353: {'lr': 0.0004226456676249898, 'samples': 7349248, 'steps': 14353, 'loss/train': 1.3432856798171997} 02/24/2022 20:09:40 - INFO - codeparrot_training - Step 14354: {'lr': 0.0004226338330480646, 'samples': 7349760, 'steps': 14354, 'loss/train': 1.8615167140960693} 02/24/2022 20:09:44 - INFO - codeparrot_training - Step 14355: {'lr': 0.00042262199773163096, 'samples': 7350272, 'steps': 14355, 'loss/train': 0.9743638634681702} 02/24/2022 20:09:49 - INFO - codeparrot_training - Step 14356: {'lr': 0.00042261016167573944, 'samples': 7350784, 'steps': 14356, 'loss/train': 2.3044936656951904} 02/24/2022 20:09:53 - INFO - codeparrot_training - Step 14357: {'lr': 0.0004225983248804408, 'samples': 7351296, 'steps': 14357, 'loss/train': 2.7075257301330566} 02/24/2022 20:09:58 - INFO - codeparrot_training - Step 14358: {'lr': 0.0004225864873457858, 'samples': 7351808, 'steps': 14358, 'loss/train': 2.5473790168762207} 02/24/2022 20:10:02 - INFO - codeparrot_training - Step 14359: {'lr': 0.0004225746490718251, 'samples': 7352320, 'steps': 14359, 'loss/train': 2.723165988922119} 02/24/2022 20:10:07 - INFO - codeparrot_training - Step 14360: {'lr': 0.0004225628100586093, 'samples': 7352832, 'steps': 14360, 'loss/train': 2.216243267059326} 02/24/2022 20:10:11 - INFO - codeparrot_training - Step 14361: {'lr': 0.0004225509703061893, 'samples': 7353344, 'steps': 14361, 'loss/train': 1.7508578300476074} 02/24/2022 20:10:16 - INFO - codeparrot_training - Step 14362: {'lr': 0.0004225391298146157, 'samples': 7353856, 'steps': 14362, 'loss/train': 1.519669532775879} 02/24/2022 20:10:20 - INFO - codeparrot_training - Step 14363: {'lr': 0.0004225272885839392, 'samples': 7354368, 'steps': 14363, 'loss/train': 2.3524415493011475} 02/24/2022 20:10:25 - INFO - codeparrot_training - Step 14364: {'lr': 0.0004225154466142107, 'samples': 7354880, 'steps': 14364, 'loss/train': 1.6151716709136963} 02/24/2022 20:10:29 - INFO - codeparrot_training - Step 14365: {'lr': 0.0004225036039054807, 'samples': 7355392, 'steps': 14365, 'loss/train': 2.6133036613464355} 02/24/2022 20:10:34 - INFO - codeparrot_training - Step 14366: {'lr': 0.00042249176045780013, 'samples': 7355904, 'steps': 14366, 'loss/train': 1.6383662223815918} 02/24/2022 20:10:38 - INFO - codeparrot_training - Step 14367: {'lr': 0.0004224799162712195, 'samples': 7356416, 'steps': 14367, 'loss/train': 2.3020594120025635} 02/24/2022 20:10:46 - INFO - codeparrot_training - Step 14368: {'lr': 0.0004224680713457898, 'samples': 7356928, 'steps': 14368, 'loss/train': 1.793048620223999} 02/24/2022 20:10:49 - INFO - codeparrot_training - Step 14369: {'lr': 0.00042245622568156164, 'samples': 7357440, 'steps': 14369, 'loss/train': 0.45681455731391907} 02/24/2022 20:10:55 - INFO - codeparrot_training - Step 14370: {'lr': 0.0004224443792785857, 'samples': 7357952, 'steps': 14370, 'loss/train': 1.9955453872680664} 02/24/2022 20:10:58 - INFO - codeparrot_training - Step 14371: {'lr': 0.0004224325321369128, 'samples': 7358464, 'steps': 14371, 'loss/train': 1.1860418319702148} 02/24/2022 20:11:04 - INFO - codeparrot_training - Step 14372: {'lr': 0.0004224206842565937, 'samples': 7358976, 'steps': 14372, 'loss/train': 1.9157577753067017} 02/24/2022 20:11:07 - INFO - codeparrot_training - Step 14373: {'lr': 0.00042240883563767916, 'samples': 7359488, 'steps': 14373, 'loss/train': 1.9489727020263672} 02/24/2022 20:11:13 - INFO - codeparrot_training - Step 14374: {'lr': 0.00042239698628021994, 'samples': 7360000, 'steps': 14374, 'loss/train': 2.3002262115478516} 02/24/2022 20:11:16 - INFO - codeparrot_training - Step 14375: {'lr': 0.0004223851361842668, 'samples': 7360512, 'steps': 14375, 'loss/train': 1.7936546802520752} 02/24/2022 20:11:22 - INFO - codeparrot_training - Step 14376: {'lr': 0.00042237328534987034, 'samples': 7361024, 'steps': 14376, 'loss/train': 2.042577028274536} 02/24/2022 20:11:25 - INFO - codeparrot_training - Step 14377: {'lr': 0.0004223614337770816, 'samples': 7361536, 'steps': 14377, 'loss/train': 2.534580707550049} 02/24/2022 20:11:31 - INFO - codeparrot_training - Step 14378: {'lr': 0.0004223495814659511, 'samples': 7362048, 'steps': 14378, 'loss/train': 1.127556562423706} 02/24/2022 20:11:34 - INFO - codeparrot_training - Step 14379: {'lr': 0.00042233772841652974, 'samples': 7362560, 'steps': 14379, 'loss/train': 1.4802942276000977} 02/24/2022 20:11:41 - INFO - codeparrot_training - Step 14380: {'lr': 0.00042232587462886833, 'samples': 7363072, 'steps': 14380, 'loss/train': 2.082639217376709} 02/24/2022 20:11:45 - INFO - codeparrot_training - Step 14381: {'lr': 0.0004223140201030176, 'samples': 7363584, 'steps': 14381, 'loss/train': 1.6701768636703491} 02/24/2022 20:11:50 - INFO - codeparrot_training - Step 14382: {'lr': 0.0004223021648390283, 'samples': 7364096, 'steps': 14382, 'loss/train': 3.672809362411499} 02/24/2022 20:11:54 - INFO - codeparrot_training - Step 14383: {'lr': 0.0004222903088369512, 'samples': 7364608, 'steps': 14383, 'loss/train': 1.7200851440429688} 02/24/2022 20:11:59 - INFO - codeparrot_training - Step 14384: {'lr': 0.0004222784520968371, 'samples': 7365120, 'steps': 14384, 'loss/train': 0.5897279977798462} 02/24/2022 20:12:03 - INFO - codeparrot_training - Step 14385: {'lr': 0.000422266594618737, 'samples': 7365632, 'steps': 14385, 'loss/train': 2.610156536102295} 02/24/2022 20:12:08 - INFO - codeparrot_training - Step 14386: {'lr': 0.0004222547364027013, 'samples': 7366144, 'steps': 14386, 'loss/train': 1.5327962636947632} 02/24/2022 20:12:12 - INFO - codeparrot_training - Step 14387: {'lr': 0.0004222428774487811, 'samples': 7366656, 'steps': 14387, 'loss/train': 2.266096830368042} 02/24/2022 20:12:19 - INFO - codeparrot_training - Step 14388: {'lr': 0.00042223101775702704, 'samples': 7367168, 'steps': 14388, 'loss/train': 1.8165912628173828} 02/24/2022 20:12:23 - INFO - codeparrot_training - Step 14389: {'lr': 0.00042221915732749006, 'samples': 7367680, 'steps': 14389, 'loss/train': 2.7712199687957764} 02/24/2022 20:12:29 - INFO - codeparrot_training - Step 14390: {'lr': 0.0004222072961602209, 'samples': 7368192, 'steps': 14390, 'loss/train': 2.4479289054870605} 02/24/2022 20:12:32 - INFO - codeparrot_training - Step 14391: {'lr': 0.0004221954342552703, 'samples': 7368704, 'steps': 14391, 'loss/train': 1.5435402393341064} 02/24/2022 20:12:35 - INFO - codeparrot_training - Step 14392: {'lr': 0.00042218357161268917, 'samples': 7369216, 'steps': 14392, 'loss/train': 0.8719975352287292} 02/24/2022 20:12:41 - INFO - codeparrot_training - Step 14393: {'lr': 0.0004221717082325283, 'samples': 7369728, 'steps': 14393, 'loss/train': 0.4197543263435364} 02/24/2022 20:12:45 - INFO - codeparrot_training - Step 14394: {'lr': 0.00042215984411483854, 'samples': 7370240, 'steps': 14394, 'loss/train': 1.9302916526794434} 02/24/2022 20:12:50 - INFO - codeparrot_training - Step 14395: {'lr': 0.00042214797925967064, 'samples': 7370752, 'steps': 14395, 'loss/train': 1.0474448204040527} 02/24/2022 20:12:56 - INFO - codeparrot_training - Step 14396: {'lr': 0.00042213611366707547, 'samples': 7371264, 'steps': 14396, 'loss/train': 1.5193946361541748} 02/24/2022 20:12:59 - INFO - codeparrot_training - Step 14397: {'lr': 0.0004221242473371038, 'samples': 7371776, 'steps': 14397, 'loss/train': 2.213353395462036} 02/24/2022 20:13:05 - INFO - codeparrot_training - Step 14398: {'lr': 0.00042211238026980657, 'samples': 7372288, 'steps': 14398, 'loss/train': 2.2594025135040283} 02/24/2022 20:13:08 - INFO - codeparrot_training - Step 14399: {'lr': 0.0004221005124652345, 'samples': 7372800, 'steps': 14399, 'loss/train': 1.2874884605407715} 02/24/2022 20:13:14 - INFO - codeparrot_training - Step 14400: {'lr': 0.0004220886439234385, 'samples': 7373312, 'steps': 14400, 'loss/train': 1.7433128356933594} 02/24/2022 20:13:17 - INFO - codeparrot_training - Step 14401: {'lr': 0.0004220767746444694, 'samples': 7373824, 'steps': 14401, 'loss/train': 1.84574556350708} 02/24/2022 20:13:23 - INFO - codeparrot_training - Step 14402: {'lr': 0.0004220649046283781, 'samples': 7374336, 'steps': 14402, 'loss/train': 2.269367218017578} 02/24/2022 20:13:26 - INFO - codeparrot_training - Step 14403: {'lr': 0.00042205303387521533, 'samples': 7374848, 'steps': 14403, 'loss/train': 2.172515630722046} 02/24/2022 20:13:33 - INFO - codeparrot_training - Step 14404: {'lr': 0.00042204116238503197, 'samples': 7375360, 'steps': 14404, 'loss/train': 2.2918453216552734} 02/24/2022 20:13:37 - INFO - codeparrot_training - Step 14405: {'lr': 0.00042202929015787893, 'samples': 7375872, 'steps': 14405, 'loss/train': 2.335841655731201} 02/24/2022 20:13:42 - INFO - codeparrot_training - Step 14406: {'lr': 0.000422017417193807, 'samples': 7376384, 'steps': 14406, 'loss/train': 2.2437055110931396} 02/24/2022 20:13:45 - INFO - codeparrot_training - Step 14407: {'lr': 0.0004220055434928671, 'samples': 7376896, 'steps': 14407, 'loss/train': 2.7684497833251953} 02/24/2022 20:13:52 - INFO - codeparrot_training - Step 14408: {'lr': 0.0004219936690551101, 'samples': 7377408, 'steps': 14408, 'loss/train': 2.7071282863616943} 02/24/2022 20:13:56 - INFO - codeparrot_training - Step 14409: {'lr': 0.0004219817938805869, 'samples': 7377920, 'steps': 14409, 'loss/train': 1.9697179794311523} 02/24/2022 20:14:01 - INFO - codeparrot_training - Step 14410: {'lr': 0.0004219699179693481, 'samples': 7378432, 'steps': 14410, 'loss/train': 2.549269914627075} 02/24/2022 20:14:05 - INFO - codeparrot_training - Step 14411: {'lr': 0.000421958041321445, 'samples': 7378944, 'steps': 14411, 'loss/train': 0.2498295158147812} 02/24/2022 20:14:10 - INFO - codeparrot_training - Step 14412: {'lr': 0.0004219461639369281, 'samples': 7379456, 'steps': 14412, 'loss/train': 0.8178668022155762} 02/24/2022 20:14:14 - INFO - codeparrot_training - Step 14413: {'lr': 0.0004219342858158485, 'samples': 7379968, 'steps': 14413, 'loss/train': 2.323140859603882} 02/24/2022 20:14:19 - INFO - codeparrot_training - Step 14414: {'lr': 0.000421922406958257, 'samples': 7380480, 'steps': 14414, 'loss/train': 1.9101383686065674} 02/24/2022 20:14:23 - INFO - codeparrot_training - Step 14415: {'lr': 0.00042191052736420445, 'samples': 7380992, 'steps': 14415, 'loss/train': 1.9701310396194458} 02/24/2022 20:14:29 - INFO - codeparrot_training - Step 14416: {'lr': 0.0004218986470337419, 'samples': 7381504, 'steps': 14416, 'loss/train': 2.2721662521362305} 02/24/2022 20:14:32 - INFO - codeparrot_training - Step 14417: {'lr': 0.00042188676596692, 'samples': 7382016, 'steps': 14417, 'loss/train': 2.5398759841918945} 02/24/2022 20:14:40 - INFO - codeparrot_training - Step 14418: {'lr': 0.0004218748841637899, 'samples': 7382528, 'steps': 14418, 'loss/train': 2.5376124382019043} 02/24/2022 20:14:43 - INFO - codeparrot_training - Step 14419: {'lr': 0.0004218630016244023, 'samples': 7383040, 'steps': 14419, 'loss/train': 2.3700544834136963} 02/24/2022 20:14:47 - INFO - codeparrot_training - Step 14420: {'lr': 0.0004218511183488082, 'samples': 7383552, 'steps': 14420, 'loss/train': 1.721614122390747} 02/24/2022 20:14:52 - INFO - codeparrot_training - Step 14421: {'lr': 0.0004218392343370584, 'samples': 7384064, 'steps': 14421, 'loss/train': 1.2554845809936523} 02/24/2022 20:14:55 - INFO - codeparrot_training - Step 14422: {'lr': 0.000421827349589204, 'samples': 7384576, 'steps': 14422, 'loss/train': 1.450324296951294} 02/24/2022 20:15:01 - INFO - codeparrot_training - Step 14423: {'lr': 0.0004218154641052957, 'samples': 7385088, 'steps': 14423, 'loss/train': 1.9200878143310547} 02/24/2022 20:15:04 - INFO - codeparrot_training - Step 14424: {'lr': 0.0004218035778853846, 'samples': 7385600, 'steps': 14424, 'loss/train': 2.5163228511810303} 02/24/2022 20:15:10 - INFO - codeparrot_training - Step 14425: {'lr': 0.0004217916909295215, 'samples': 7386112, 'steps': 14425, 'loss/train': 3.1031835079193115} 02/24/2022 20:15:14 - INFO - codeparrot_training - Step 14426: {'lr': 0.00042177980323775734, 'samples': 7386624, 'steps': 14426, 'loss/train': 2.2497856616973877} 02/24/2022 20:15:19 - INFO - codeparrot_training - Step 14427: {'lr': 0.00042176791481014303, 'samples': 7387136, 'steps': 14427, 'loss/train': 2.626216411590576} 02/24/2022 20:15:23 - INFO - codeparrot_training - Step 14428: {'lr': 0.0004217560256467295, 'samples': 7387648, 'steps': 14428, 'loss/train': 0.8603918552398682} 02/24/2022 20:15:29 - INFO - codeparrot_training - Step 14429: {'lr': 0.00042174413574756775, 'samples': 7388160, 'steps': 14429, 'loss/train': 2.1737658977508545} 02/24/2022 20:15:33 - INFO - codeparrot_training - Step 14430: {'lr': 0.0004217322451127086, 'samples': 7388672, 'steps': 14430, 'loss/train': 2.1812446117401123} 02/24/2022 20:15:38 - INFO - codeparrot_training - Step 14431: {'lr': 0.00042172035374220306, 'samples': 7389184, 'steps': 14431, 'loss/train': 1.7511701583862305} 02/24/2022 20:15:42 - INFO - codeparrot_training - Step 14432: {'lr': 0.0004217084616361021, 'samples': 7389696, 'steps': 14432, 'loss/train': 2.7966084480285645} 02/24/2022 20:15:47 - INFO - codeparrot_training - Step 14433: {'lr': 0.00042169656879445657, 'samples': 7390208, 'steps': 14433, 'loss/train': 2.0498228073120117} 02/24/2022 20:15:51 - INFO - codeparrot_training - Step 14434: {'lr': 0.00042168467521731747, 'samples': 7390720, 'steps': 14434, 'loss/train': 2.652679443359375} 02/24/2022 20:15:56 - INFO - codeparrot_training - Step 14435: {'lr': 0.00042167278090473573, 'samples': 7391232, 'steps': 14435, 'loss/train': 1.6542110443115234} 02/24/2022 20:16:00 - INFO - codeparrot_training - Step 14436: {'lr': 0.0004216608858567623, 'samples': 7391744, 'steps': 14436, 'loss/train': 1.975625991821289} 02/24/2022 20:16:05 - INFO - codeparrot_training - Step 14437: {'lr': 0.00042164899007344814, 'samples': 7392256, 'steps': 14437, 'loss/train': 2.3574142456054688} 02/24/2022 20:16:09 - INFO - codeparrot_training - Step 14438: {'lr': 0.00042163709355484425, 'samples': 7392768, 'steps': 14438, 'loss/train': 2.790147066116333} 02/24/2022 20:16:15 - INFO - codeparrot_training - Step 14439: {'lr': 0.0004216251963010015, 'samples': 7393280, 'steps': 14439, 'loss/train': 2.61586332321167} 02/24/2022 20:16:18 - INFO - codeparrot_training - Step 14440: {'lr': 0.0004216132983119709, 'samples': 7393792, 'steps': 14440, 'loss/train': 2.4092423915863037} 02/24/2022 20:16:24 - INFO - codeparrot_training - Step 14441: {'lr': 0.00042160139958780346, 'samples': 7394304, 'steps': 14441, 'loss/train': 1.951647162437439} 02/24/2022 20:16:27 - INFO - codeparrot_training - Step 14442: {'lr': 0.0004215895001285501, 'samples': 7394816, 'steps': 14442, 'loss/train': 1.4745805263519287} 02/24/2022 20:16:33 - INFO - codeparrot_training - Step 14443: {'lr': 0.0004215775999342618, 'samples': 7395328, 'steps': 14443, 'loss/train': 2.3024535179138184} 02/24/2022 20:16:36 - INFO - codeparrot_training - Step 14444: {'lr': 0.0004215656990049896, 'samples': 7395840, 'steps': 14444, 'loss/train': 2.0168559551239014} 02/24/2022 20:16:42 - INFO - codeparrot_training - Step 14445: {'lr': 0.0004215537973407844, 'samples': 7396352, 'steps': 14445, 'loss/train': 2.563513994216919} 02/24/2022 20:16:45 - INFO - codeparrot_training - Step 14446: {'lr': 0.0004215418949416972, 'samples': 7396864, 'steps': 14446, 'loss/train': 2.3443524837493896} 02/24/2022 20:16:51 - INFO - codeparrot_training - Step 14447: {'lr': 0.00042152999180777894, 'samples': 7397376, 'steps': 14447, 'loss/train': 1.9824228286743164} 02/24/2022 20:16:54 - INFO - codeparrot_training - Step 14448: {'lr': 0.0004215180879390807, 'samples': 7397888, 'steps': 14448, 'loss/train': 2.6413373947143555} 02/24/2022 20:17:01 - INFO - codeparrot_training - Step 14449: {'lr': 0.0004215061833356535, 'samples': 7398400, 'steps': 14449, 'loss/train': 3.082911252975464} 02/24/2022 20:17:04 - INFO - codeparrot_training - Step 14450: {'lr': 0.00042149427799754817, 'samples': 7398912, 'steps': 14450, 'loss/train': 2.255244255065918} 02/24/2022 20:17:10 - INFO - codeparrot_training - Step 14451: {'lr': 0.00042148237192481586, 'samples': 7399424, 'steps': 14451, 'loss/train': 2.464221715927124} 02/24/2022 20:17:13 - INFO - codeparrot_training - Step 14452: {'lr': 0.0004214704651175075, 'samples': 7399936, 'steps': 14452, 'loss/train': 4.910278797149658} 02/24/2022 20:17:19 - INFO - codeparrot_training - Step 14453: {'lr': 0.0004214585575756742, 'samples': 7400448, 'steps': 14453, 'loss/train': 0.29434093832969666} 02/24/2022 20:17:22 - INFO - codeparrot_training - Step 14454: {'lr': 0.0004214466492993668, 'samples': 7400960, 'steps': 14454, 'loss/train': 2.264237880706787} 02/24/2022 20:17:28 - INFO - codeparrot_training - Step 14455: {'lr': 0.00042143474028863637, 'samples': 7401472, 'steps': 14455, 'loss/train': 2.308858633041382} 02/24/2022 20:17:31 - INFO - codeparrot_training - Step 14456: {'lr': 0.000421422830543534, 'samples': 7401984, 'steps': 14456, 'loss/train': 8.902766227722168} 02/24/2022 20:17:37 - INFO - codeparrot_training - Step 14457: {'lr': 0.0004214109200641106, 'samples': 7402496, 'steps': 14457, 'loss/train': 1.2082109451293945} 02/24/2022 20:17:40 - INFO - codeparrot_training - Step 14458: {'lr': 0.00042139900885041734, 'samples': 7403008, 'steps': 14458, 'loss/train': 1.3440091609954834} 02/24/2022 20:17:47 - INFO - codeparrot_training - Step 14459: {'lr': 0.00042138709690250507, 'samples': 7403520, 'steps': 14459, 'loss/train': 0.8292269706726074} 02/24/2022 20:17:50 - INFO - codeparrot_training - Step 14460: {'lr': 0.0004213751842204249, 'samples': 7404032, 'steps': 14460, 'loss/train': 2.31581711769104} 02/24/2022 20:17:56 - INFO - codeparrot_training - Step 14461: {'lr': 0.00042136327080422785, 'samples': 7404544, 'steps': 14461, 'loss/train': 2.3676931858062744} 02/24/2022 20:17:59 - INFO - codeparrot_training - Step 14462: {'lr': 0.0004213513566539651, 'samples': 7405056, 'steps': 14462, 'loss/train': 1.2420704364776611} 02/24/2022 20:18:05 - INFO - codeparrot_training - Step 14463: {'lr': 0.0004213394417696874, 'samples': 7405568, 'steps': 14463, 'loss/train': 2.387732982635498} 02/24/2022 20:18:08 - INFO - codeparrot_training - Step 14464: {'lr': 0.00042132752615144597, 'samples': 7406080, 'steps': 14464, 'loss/train': 2.154527187347412} 02/24/2022 20:18:14 - INFO - codeparrot_training - Step 14465: {'lr': 0.00042131560979929186, 'samples': 7406592, 'steps': 14465, 'loss/train': 1.8868833780288696} 02/24/2022 20:18:17 - INFO - codeparrot_training - Step 14466: {'lr': 0.00042130369271327605, 'samples': 7407104, 'steps': 14466, 'loss/train': 1.738230586051941} 02/24/2022 20:18:23 - INFO - codeparrot_training - Step 14467: {'lr': 0.0004212917748934496, 'samples': 7407616, 'steps': 14467, 'loss/train': 2.4486939907073975} 02/24/2022 20:18:26 - INFO - codeparrot_training - Step 14468: {'lr': 0.00042127985633986365, 'samples': 7408128, 'steps': 14468, 'loss/train': 1.370111346244812} 02/24/2022 20:18:32 - INFO - codeparrot_training - Step 14469: {'lr': 0.00042126793705256913, 'samples': 7408640, 'steps': 14469, 'loss/train': 2.3062503337860107} 02/24/2022 20:18:35 - INFO - codeparrot_training - Step 14470: {'lr': 0.00042125601703161706, 'samples': 7409152, 'steps': 14470, 'loss/train': 1.824141502380371} 02/24/2022 20:18:41 - INFO - codeparrot_training - Step 14471: {'lr': 0.00042124409627705873, 'samples': 7409664, 'steps': 14471, 'loss/train': 3.3604238033294678} 02/24/2022 20:18:44 - INFO - codeparrot_training - Step 14472: {'lr': 0.00042123217478894504, 'samples': 7410176, 'steps': 14472, 'loss/train': 1.6430108547210693} 02/24/2022 20:18:50 - INFO - codeparrot_training - Step 14473: {'lr': 0.0004212202525673271, 'samples': 7410688, 'steps': 14473, 'loss/train': 0.5625841021537781} 02/24/2022 20:18:53 - INFO - codeparrot_training - Step 14474: {'lr': 0.00042120832961225585, 'samples': 7411200, 'steps': 14474, 'loss/train': 1.4384527206420898} 02/24/2022 20:18:59 - INFO - codeparrot_training - Step 14475: {'lr': 0.00042119640592378263, 'samples': 7411712, 'steps': 14475, 'loss/train': 0.25691255927085876} 02/24/2022 20:19:03 - INFO - codeparrot_training - Step 14476: {'lr': 0.00042118448150195827, 'samples': 7412224, 'steps': 14476, 'loss/train': 2.408867359161377} 02/24/2022 20:19:09 - INFO - codeparrot_training - Step 14477: {'lr': 0.000421172556346834, 'samples': 7412736, 'steps': 14477, 'loss/train': 3.917689085006714} 02/24/2022 20:19:12 - INFO - codeparrot_training - Step 14478: {'lr': 0.00042116063045846073, 'samples': 7413248, 'steps': 14478, 'loss/train': 2.466057538986206} 02/24/2022 20:19:16 - INFO - codeparrot_training - Step 14479: {'lr': 0.00042114870383688985, 'samples': 7413760, 'steps': 14479, 'loss/train': 1.545590877532959} 02/24/2022 20:19:21 - INFO - codeparrot_training - Step 14480: {'lr': 0.0004211367764821722, 'samples': 7414272, 'steps': 14480, 'loss/train': 1.8706343173980713} 02/24/2022 20:19:25 - INFO - codeparrot_training - Step 14481: {'lr': 0.00042112484839435893, 'samples': 7414784, 'steps': 14481, 'loss/train': 2.249089241027832} 02/24/2022 20:19:30 - INFO - codeparrot_training - Step 14482: {'lr': 0.00042111291957350113, 'samples': 7415296, 'steps': 14482, 'loss/train': 2.7552359104156494} 02/24/2022 20:19:34 - INFO - codeparrot_training - Step 14483: {'lr': 0.00042110099001964996, 'samples': 7415808, 'steps': 14483, 'loss/train': 1.8637077808380127} 02/24/2022 20:19:40 - INFO - codeparrot_training - Step 14484: {'lr': 0.0004210890597328564, 'samples': 7416320, 'steps': 14484, 'loss/train': 2.5025877952575684} 02/24/2022 20:19:43 - INFO - codeparrot_training - Step 14485: {'lr': 0.0004210771287131717, 'samples': 7416832, 'steps': 14485, 'loss/train': 2.380509614944458} 02/24/2022 20:19:48 - INFO - codeparrot_training - Step 14486: {'lr': 0.00042106519696064694, 'samples': 7417344, 'steps': 14486, 'loss/train': 2.055649518966675} 02/24/2022 20:19:52 - INFO - codeparrot_training - Step 14487: {'lr': 0.0004210532644753331, 'samples': 7417856, 'steps': 14487, 'loss/train': 1.5627260208129883} 02/24/2022 20:19:58 - INFO - codeparrot_training - Step 14488: {'lr': 0.00042104133125728146, 'samples': 7418368, 'steps': 14488, 'loss/train': 2.8318421840667725} 02/24/2022 20:20:02 - INFO - codeparrot_training - Step 14489: {'lr': 0.00042102939730654304, 'samples': 7418880, 'steps': 14489, 'loss/train': 2.5699374675750732} 02/24/2022 20:20:07 - INFO - codeparrot_training - Step 14490: {'lr': 0.000421017462623169, 'samples': 7419392, 'steps': 14490, 'loss/train': 1.6295500993728638} 02/24/2022 20:20:10 - INFO - codeparrot_training - Step 14491: {'lr': 0.0004210055272072104, 'samples': 7419904, 'steps': 14491, 'loss/train': 1.347360610961914} 02/24/2022 20:20:16 - INFO - codeparrot_training - Step 14492: {'lr': 0.00042099359105871856, 'samples': 7420416, 'steps': 14492, 'loss/train': 1.6652920246124268} 02/24/2022 20:20:19 - INFO - codeparrot_training - Step 14493: {'lr': 0.0004209816541777444, 'samples': 7420928, 'steps': 14493, 'loss/train': 2.105761766433716} 02/24/2022 20:20:25 - INFO - codeparrot_training - Step 14494: {'lr': 0.0004209697165643391, 'samples': 7421440, 'steps': 14494, 'loss/train': 2.452751636505127} 02/24/2022 20:20:29 - INFO - codeparrot_training - Step 14495: {'lr': 0.0004209577782185538, 'samples': 7421952, 'steps': 14495, 'loss/train': 2.643202066421509} 02/24/2022 20:20:34 - INFO - codeparrot_training - Step 14496: {'lr': 0.0004209458391404397, 'samples': 7422464, 'steps': 14496, 'loss/train': 1.2692300081253052} 02/24/2022 20:20:38 - INFO - codeparrot_training - Step 14497: {'lr': 0.0004209338993300479, 'samples': 7422976, 'steps': 14497, 'loss/train': 1.755126953125} 02/24/2022 20:20:45 - INFO - codeparrot_training - Step 14498: {'lr': 0.00042092195878742954, 'samples': 7423488, 'steps': 14498, 'loss/train': 1.7072489261627197} 02/24/2022 20:20:48 - INFO - codeparrot_training - Step 14499: {'lr': 0.0004209100175126358, 'samples': 7424000, 'steps': 14499, 'loss/train': 1.4411416053771973} 02/24/2022 20:20:54 - INFO - codeparrot_training - Step 14500: {'lr': 0.0004208980755057178, 'samples': 7424512, 'steps': 14500, 'loss/train': 2.59185528755188} 02/24/2022 20:20:57 - INFO - codeparrot_training - Step 14501: {'lr': 0.0004208861327667268, 'samples': 7425024, 'steps': 14501, 'loss/train': 1.7651935815811157} 02/24/2022 20:21:03 - INFO - codeparrot_training - Step 14502: {'lr': 0.00042087418929571377, 'samples': 7425536, 'steps': 14502, 'loss/train': 1.57146418094635} 02/24/2022 20:21:06 - INFO - codeparrot_training - Step 14503: {'lr': 0.00042086224509272995, 'samples': 7426048, 'steps': 14503, 'loss/train': 1.0619361400604248} 02/24/2022 20:21:11 - INFO - codeparrot_training - Step 14504: {'lr': 0.0004208503001578266, 'samples': 7426560, 'steps': 14504, 'loss/train': 2.572526693344116} 02/24/2022 20:21:15 - INFO - codeparrot_training - Step 14505: {'lr': 0.00042083835449105477, 'samples': 7427072, 'steps': 14505, 'loss/train': 1.7435468435287476} 02/24/2022 20:21:20 - INFO - codeparrot_training - Step 14506: {'lr': 0.00042082640809246576, 'samples': 7427584, 'steps': 14506, 'loss/train': 0.6964829564094543} 02/24/2022 20:21:24 - INFO - codeparrot_training - Step 14507: {'lr': 0.0004208144609621106, 'samples': 7428096, 'steps': 14507, 'loss/train': 3.0581626892089844} 02/24/2022 20:21:31 - INFO - codeparrot_training - Step 14508: {'lr': 0.0004208025131000405, 'samples': 7428608, 'steps': 14508, 'loss/train': 0.207466721534729} 02/24/2022 20:21:36 - INFO - codeparrot_training - Step 14509: {'lr': 0.0004207905645063067, 'samples': 7429120, 'steps': 14509, 'loss/train': 1.0727956295013428} 02/24/2022 20:21:40 - INFO - codeparrot_training - Step 14510: {'lr': 0.00042077861518096033, 'samples': 7429632, 'steps': 14510, 'loss/train': 2.526712656021118} 02/24/2022 20:21:45 - INFO - codeparrot_training - Step 14511: {'lr': 0.0004207666651240526, 'samples': 7430144, 'steps': 14511, 'loss/train': 1.8233929872512817} 02/24/2022 20:21:49 - INFO - codeparrot_training - Step 14512: {'lr': 0.0004207547143356347, 'samples': 7430656, 'steps': 14512, 'loss/train': 2.5123543739318848} 02/24/2022 20:21:54 - INFO - codeparrot_training - Step 14513: {'lr': 0.00042074276281575787, 'samples': 7431168, 'steps': 14513, 'loss/train': 2.1859428882598877} 02/24/2022 20:21:58 - INFO - codeparrot_training - Step 14514: {'lr': 0.00042073081056447325, 'samples': 7431680, 'steps': 14514, 'loss/train': 2.1641855239868164} 02/24/2022 20:22:03 - INFO - codeparrot_training - Step 14515: {'lr': 0.00042071885758183204, 'samples': 7432192, 'steps': 14515, 'loss/train': 1.7970335483551025} 02/24/2022 20:22:07 - INFO - codeparrot_training - Step 14516: {'lr': 0.00042070690386788545, 'samples': 7432704, 'steps': 14516, 'loss/train': 2.465810537338257} 02/24/2022 20:22:12 - INFO - codeparrot_training - Step 14517: {'lr': 0.0004206949494226847, 'samples': 7433216, 'steps': 14517, 'loss/train': 1.938108205795288} 02/24/2022 20:22:16 - INFO - codeparrot_training - Step 14518: {'lr': 0.000420682994246281, 'samples': 7433728, 'steps': 14518, 'loss/train': 2.027783155441284} 02/24/2022 20:22:21 - INFO - codeparrot_training - Step 14519: {'lr': 0.00042067103833872554, 'samples': 7434240, 'steps': 14519, 'loss/train': 1.2425369024276733} 02/24/2022 20:22:25 - INFO - codeparrot_training - Step 14520: {'lr': 0.0004206590817000695, 'samples': 7434752, 'steps': 14520, 'loss/train': 1.8573615550994873} 02/24/2022 20:22:30 - INFO - codeparrot_training - Step 14521: {'lr': 0.0004206471243303642, 'samples': 7435264, 'steps': 14521, 'loss/train': 0.4201807975769043} 02/24/2022 20:22:34 - INFO - codeparrot_training - Step 14522: {'lr': 0.0004206351662296608, 'samples': 7435776, 'steps': 14522, 'loss/train': 1.7703566551208496} 02/24/2022 20:22:41 - INFO - codeparrot_training - Step 14523: {'lr': 0.0004206232073980105, 'samples': 7436288, 'steps': 14523, 'loss/train': 1.8556761741638184} 02/24/2022 20:22:44 - INFO - codeparrot_training - Step 14524: {'lr': 0.00042061124783546454, 'samples': 7436800, 'steps': 14524, 'loss/train': 2.1972217559814453} 02/24/2022 20:22:50 - INFO - codeparrot_training - Step 14525: {'lr': 0.0004205992875420742, 'samples': 7437312, 'steps': 14525, 'loss/train': 1.4722493886947632} 02/24/2022 20:22:53 - INFO - codeparrot_training - Step 14526: {'lr': 0.0004205873265178907, 'samples': 7437824, 'steps': 14526, 'loss/train': 2.1752734184265137} 02/24/2022 20:22:59 - INFO - codeparrot_training - Step 14527: {'lr': 0.0004205753647629653, 'samples': 7438336, 'steps': 14527, 'loss/train': 2.447154998779297} 02/24/2022 20:23:02 - INFO - codeparrot_training - Step 14528: {'lr': 0.0004205634022773491, 'samples': 7438848, 'steps': 14528, 'loss/train': 2.4360098838806152} 02/24/2022 20:23:08 - INFO - codeparrot_training - Step 14529: {'lr': 0.0004205514390610935, 'samples': 7439360, 'steps': 14529, 'loss/train': 2.466944456100464} 02/24/2022 20:23:11 - INFO - codeparrot_training - Step 14530: {'lr': 0.00042053947511424975, 'samples': 7439872, 'steps': 14530, 'loss/train': 1.8803566694259644} 02/24/2022 20:23:17 - INFO - codeparrot_training - Step 14531: {'lr': 0.00042052751043686895, 'samples': 7440384, 'steps': 14531, 'loss/train': 2.230368137359619} 02/24/2022 20:23:20 - INFO - codeparrot_training - Step 14532: {'lr': 0.00042051554502900245, 'samples': 7440896, 'steps': 14532, 'loss/train': 2.1919503211975098} 02/24/2022 20:23:27 - INFO - codeparrot_training - Step 14533: {'lr': 0.0004205035788907015, 'samples': 7441408, 'steps': 14533, 'loss/train': 1.9169282913208008} 02/24/2022 20:23:30 - INFO - codeparrot_training - Step 14534: {'lr': 0.0004204916120220174, 'samples': 7441920, 'steps': 14534, 'loss/train': 0.8139283657073975} 02/24/2022 20:23:36 - INFO - codeparrot_training - Step 14535: {'lr': 0.00042047964442300137, 'samples': 7442432, 'steps': 14535, 'loss/train': 1.701749563217163} 02/24/2022 20:23:39 - INFO - codeparrot_training - Step 14536: {'lr': 0.0004204676760937046, 'samples': 7442944, 'steps': 14536, 'loss/train': 1.6795092821121216} 02/24/2022 20:23:45 - INFO - codeparrot_training - Step 14537: {'lr': 0.00042045570703417857, 'samples': 7443456, 'steps': 14537, 'loss/train': 1.0003081560134888} 02/24/2022 20:23:48 - INFO - codeparrot_training - Step 14538: {'lr': 0.00042044373724447434, 'samples': 7443968, 'steps': 14538, 'loss/train': 1.6481555700302124} 02/24/2022 20:23:54 - INFO - codeparrot_training - Step 14539: {'lr': 0.0004204317667246432, 'samples': 7444480, 'steps': 14539, 'loss/train': 1.1439129114151} 02/24/2022 20:23:57 - INFO - codeparrot_training - Step 14540: {'lr': 0.00042041979547473665, 'samples': 7444992, 'steps': 14540, 'loss/train': 1.3678139448165894} 02/24/2022 20:24:03 - INFO - codeparrot_training - Step 14541: {'lr': 0.0004204078234948057, 'samples': 7445504, 'steps': 14541, 'loss/train': 0.9234797954559326} 02/24/2022 20:24:06 - INFO - codeparrot_training - Step 14542: {'lr': 0.00042039585078490173, 'samples': 7446016, 'steps': 14542, 'loss/train': 2.22986102104187} 02/24/2022 20:24:13 - INFO - codeparrot_training - Step 14543: {'lr': 0.000420383877345076, 'samples': 7446528, 'steps': 14543, 'loss/train': 2.308837890625} 02/24/2022 20:24:16 - INFO - codeparrot_training - Step 14544: {'lr': 0.00042037190317538, 'samples': 7447040, 'steps': 14544, 'loss/train': 2.48694109916687} 02/24/2022 20:24:22 - INFO - codeparrot_training - Step 14545: {'lr': 0.00042035992827586474, 'samples': 7447552, 'steps': 14545, 'loss/train': 1.7496896982192993} 02/24/2022 20:24:25 - INFO - codeparrot_training - Step 14546: {'lr': 0.00042034795264658163, 'samples': 7448064, 'steps': 14546, 'loss/train': 8.975650787353516} 02/24/2022 20:24:31 - INFO - codeparrot_training - Step 14547: {'lr': 0.00042033597628758206, 'samples': 7448576, 'steps': 14547, 'loss/train': 2.4917502403259277} 02/24/2022 20:24:34 - INFO - codeparrot_training - Step 14548: {'lr': 0.00042032399919891724, 'samples': 7449088, 'steps': 14548, 'loss/train': 1.7235925197601318} 02/24/2022 20:24:40 - INFO - codeparrot_training - Step 14549: {'lr': 0.0004203120213806385, 'samples': 7449600, 'steps': 14549, 'loss/train': 1.7175612449645996} 02/24/2022 20:24:43 - INFO - codeparrot_training - Step 14550: {'lr': 0.0004203000428327971, 'samples': 7450112, 'steps': 14550, 'loss/train': 0.7223354578018188} 02/24/2022 20:24:49 - INFO - codeparrot_training - Step 14551: {'lr': 0.00042028806355544443, 'samples': 7450624, 'steps': 14551, 'loss/train': 1.2693114280700684} 02/24/2022 20:24:52 - INFO - codeparrot_training - Step 14552: {'lr': 0.0004202760835486317, 'samples': 7451136, 'steps': 14552, 'loss/train': 3.00299334526062} 02/24/2022 20:24:58 - INFO - codeparrot_training - Step 14553: {'lr': 0.00042026410281241033, 'samples': 7451648, 'steps': 14553, 'loss/train': 2.212014675140381} 02/24/2022 20:25:02 - INFO - codeparrot_training - Step 14554: {'lr': 0.00042025212134683165, 'samples': 7452160, 'steps': 14554, 'loss/train': 1.0328600406646729} 02/24/2022 20:25:07 - INFO - codeparrot_training - Step 14555: {'lr': 0.0004202401391519469, 'samples': 7452672, 'steps': 14555, 'loss/train': 2.4593493938446045} 02/24/2022 20:25:11 - INFO - codeparrot_training - Step 14556: {'lr': 0.0004202281562278075, 'samples': 7453184, 'steps': 14556, 'loss/train': 2.254045009613037} 02/24/2022 20:25:16 - INFO - codeparrot_training - Step 14557: {'lr': 0.0004202161725744647, 'samples': 7453696, 'steps': 14557, 'loss/train': 0.7042149901390076} 02/24/2022 20:25:20 - INFO - codeparrot_training - Step 14558: {'lr': 0.0004202041881919699, 'samples': 7454208, 'steps': 14558, 'loss/train': 1.346352219581604} 02/24/2022 20:25:25 - INFO - codeparrot_training - Step 14559: {'lr': 0.0004201922030803743, 'samples': 7454720, 'steps': 14559, 'loss/train': 1.8507301807403564} 02/24/2022 20:25:29 - INFO - codeparrot_training - Step 14560: {'lr': 0.0004201802172397295, 'samples': 7455232, 'steps': 14560, 'loss/train': 1.4599277973175049} 02/24/2022 20:25:34 - INFO - codeparrot_training - Step 14561: {'lr': 0.0004201682306700866, 'samples': 7455744, 'steps': 14561, 'loss/train': 3.0319459438323975} 02/24/2022 20:25:38 - INFO - codeparrot_training - Step 14562: {'lr': 0.00042015624337149703, 'samples': 7456256, 'steps': 14562, 'loss/train': 2.387321949005127} 02/24/2022 20:25:44 - INFO - codeparrot_training - Step 14563: {'lr': 0.0004201442553440121, 'samples': 7456768, 'steps': 14563, 'loss/train': 1.8804051876068115} 02/24/2022 20:25:47 - INFO - codeparrot_training - Step 14564: {'lr': 0.00042013226658768333, 'samples': 7457280, 'steps': 14564, 'loss/train': 0.9310999512672424} 02/24/2022 20:25:53 - INFO - codeparrot_training - Step 14565: {'lr': 0.0004201202771025618, 'samples': 7457792, 'steps': 14565, 'loss/train': 1.5589817762374878} 02/24/2022 20:25:56 - INFO - codeparrot_training - Step 14566: {'lr': 0.0004201082868886992, 'samples': 7458304, 'steps': 14566, 'loss/train': 2.2237401008605957} 02/24/2022 20:26:02 - INFO - codeparrot_training - Step 14567: {'lr': 0.00042009629594614656, 'samples': 7458816, 'steps': 14567, 'loss/train': 2.1755847930908203} 02/24/2022 20:26:05 - INFO - codeparrot_training - Step 14568: {'lr': 0.0004200843042749555, 'samples': 7459328, 'steps': 14568, 'loss/train': 1.5562310218811035} 02/24/2022 20:26:12 - INFO - codeparrot_training - Step 14569: {'lr': 0.0004200723118751772, 'samples': 7459840, 'steps': 14569, 'loss/train': 2.437448501586914} 02/24/2022 20:26:17 - INFO - codeparrot_training - Step 14570: {'lr': 0.00042006031874686315, 'samples': 7460352, 'steps': 14570, 'loss/train': 1.4461536407470703} 02/24/2022 20:26:21 - INFO - codeparrot_training - Step 14571: {'lr': 0.00042004832489006474, 'samples': 7460864, 'steps': 14571, 'loss/train': 2.2945306301116943} 02/24/2022 20:26:24 - INFO - codeparrot_training - Step 14572: {'lr': 0.0004200363303048332, 'samples': 7461376, 'steps': 14572, 'loss/train': 1.3468681573867798} 02/24/2022 20:26:30 - INFO - codeparrot_training - Step 14573: {'lr': 0.00042002433499122016, 'samples': 7461888, 'steps': 14573, 'loss/train': 0.2137172967195511} 02/24/2022 20:26:33 - INFO - codeparrot_training - Step 14574: {'lr': 0.00042001233894927684, 'samples': 7462400, 'steps': 14574, 'loss/train': 2.0592262744903564} 02/24/2022 20:26:39 - INFO - codeparrot_training - Step 14575: {'lr': 0.0004200003421790546, 'samples': 7462912, 'steps': 14575, 'loss/train': 2.0852370262145996} 02/24/2022 20:26:42 - INFO - codeparrot_training - Step 14576: {'lr': 0.0004199883446806048, 'samples': 7463424, 'steps': 14576, 'loss/train': 2.2405941486358643} 02/24/2022 20:26:48 - INFO - codeparrot_training - Step 14577: {'lr': 0.00041997634645397897, 'samples': 7463936, 'steps': 14577, 'loss/train': 2.425971508026123} 02/24/2022 20:26:52 - INFO - codeparrot_training - Step 14578: {'lr': 0.0004199643474992285, 'samples': 7464448, 'steps': 14578, 'loss/train': 3.7889187335968018} 02/24/2022 20:26:58 - INFO - codeparrot_training - Step 14579: {'lr': 0.00041995234781640466, 'samples': 7464960, 'steps': 14579, 'loss/train': 2.0633435249328613} 02/24/2022 20:27:01 - INFO - codeparrot_training - Step 14580: {'lr': 0.00041994034740555896, 'samples': 7465472, 'steps': 14580, 'loss/train': 2.2397541999816895} 02/24/2022 20:27:07 - INFO - codeparrot_training - Step 14581: {'lr': 0.00041992834626674273, 'samples': 7465984, 'steps': 14581, 'loss/train': 2.570878744125366} 02/24/2022 20:27:10 - INFO - codeparrot_training - Step 14582: {'lr': 0.0004199163444000075, 'samples': 7466496, 'steps': 14582, 'loss/train': 2.2161812782287598} 02/24/2022 20:27:16 - INFO - codeparrot_training - Step 14583: {'lr': 0.00041990434180540453, 'samples': 7467008, 'steps': 14583, 'loss/train': 1.61948823928833} 02/24/2022 20:27:19 - INFO - codeparrot_training - Step 14584: {'lr': 0.00041989233848298534, 'samples': 7467520, 'steps': 14584, 'loss/train': 2.2366573810577393} 02/24/2022 20:27:25 - INFO - codeparrot_training - Step 14585: {'lr': 0.00041988033443280136, 'samples': 7468032, 'steps': 14585, 'loss/train': 1.9982514381408691} 02/24/2022 20:27:28 - INFO - codeparrot_training - Step 14586: {'lr': 0.00041986832965490396, 'samples': 7468544, 'steps': 14586, 'loss/train': 1.779753565788269} 02/24/2022 20:27:34 - INFO - codeparrot_training - Step 14587: {'lr': 0.0004198563241493445, 'samples': 7469056, 'steps': 14587, 'loss/train': 3.2324678897857666} 02/24/2022 20:27:37 - INFO - codeparrot_training - Step 14588: {'lr': 0.00041984431791617456, 'samples': 7469568, 'steps': 14588, 'loss/train': 2.060723066329956} 02/24/2022 20:27:44 - INFO - codeparrot_training - Step 14589: {'lr': 0.00041983231095544545, 'samples': 7470080, 'steps': 14589, 'loss/train': 0.5975165367126465} 02/24/2022 20:27:48 - INFO - codeparrot_training - Step 14590: {'lr': 0.00041982030326720866, 'samples': 7470592, 'steps': 14590, 'loss/train': 2.745650053024292} 02/24/2022 20:27:53 - INFO - codeparrot_training - Step 14591: {'lr': 0.00041980829485151563, 'samples': 7471104, 'steps': 14591, 'loss/train': 1.505582571029663} 02/24/2022 20:27:57 - INFO - codeparrot_training - Step 14592: {'lr': 0.00041979628570841776, 'samples': 7471616, 'steps': 14592, 'loss/train': 2.2861671447753906} 02/24/2022 20:28:02 - INFO - codeparrot_training - Step 14593: {'lr': 0.00041978427583796654, 'samples': 7472128, 'steps': 14593, 'loss/train': 1.7498937845230103} 02/24/2022 20:28:06 - INFO - codeparrot_training - Step 14594: {'lr': 0.00041977226524021337, 'samples': 7472640, 'steps': 14594, 'loss/train': 3.7761788368225098} 02/24/2022 20:28:11 - INFO - codeparrot_training - Step 14595: {'lr': 0.0004197602539152098, 'samples': 7473152, 'steps': 14595, 'loss/train': 1.72860586643219} 02/24/2022 20:28:15 - INFO - codeparrot_training - Step 14596: {'lr': 0.00041974824186300706, 'samples': 7473664, 'steps': 14596, 'loss/train': 1.5018503665924072} 02/24/2022 20:28:20 - INFO - codeparrot_training - Step 14597: {'lr': 0.0004197362290836569, 'samples': 7474176, 'steps': 14597, 'loss/train': 2.6293368339538574} 02/24/2022 20:28:24 - INFO - codeparrot_training - Step 14598: {'lr': 0.00041972421557721055, 'samples': 7474688, 'steps': 14598, 'loss/train': 1.987839698791504} 02/24/2022 20:28:30 - INFO - codeparrot_training - Step 14599: {'lr': 0.00041971220134371957, 'samples': 7475200, 'steps': 14599, 'loss/train': 3.2312142848968506} 02/24/2022 20:28:34 - INFO - codeparrot_training - Step 14600: {'lr': 0.00041970018638323546, 'samples': 7475712, 'steps': 14600, 'loss/train': 0.6105589270591736} 02/24/2022 20:28:39 - INFO - codeparrot_training - Step 14601: {'lr': 0.0004196881706958096, 'samples': 7476224, 'steps': 14601, 'loss/train': 1.999027967453003} 02/24/2022 20:28:43 - INFO - codeparrot_training - Step 14602: {'lr': 0.00041967615428149346, 'samples': 7476736, 'steps': 14602, 'loss/train': 1.6984055042266846} 02/24/2022 20:28:48 - INFO - codeparrot_training - Step 14603: {'lr': 0.0004196641371403386, 'samples': 7477248, 'steps': 14603, 'loss/train': 2.222597360610962} 02/24/2022 20:28:52 - INFO - codeparrot_training - Step 14604: {'lr': 0.00041965211927239644, 'samples': 7477760, 'steps': 14604, 'loss/train': 2.177447557449341} 02/24/2022 20:28:57 - INFO - codeparrot_training - Step 14605: {'lr': 0.0004196401006777185, 'samples': 7478272, 'steps': 14605, 'loss/train': 2.0778374671936035} 02/24/2022 20:29:01 - INFO - codeparrot_training - Step 14606: {'lr': 0.00041962808135635624, 'samples': 7478784, 'steps': 14606, 'loss/train': 1.7459238767623901} 02/24/2022 20:29:06 - INFO - codeparrot_training - Step 14607: {'lr': 0.00041961606130836105, 'samples': 7479296, 'steps': 14607, 'loss/train': 1.6056931018829346} 02/24/2022 20:29:10 - INFO - codeparrot_training - Step 14608: {'lr': 0.0004196040405337845, 'samples': 7479808, 'steps': 14608, 'loss/train': 2.2504143714904785} 02/24/2022 20:29:15 - INFO - codeparrot_training - Step 14609: {'lr': 0.0004195920190326782, 'samples': 7480320, 'steps': 14609, 'loss/train': 2.0761241912841797} 02/24/2022 20:29:19 - INFO - codeparrot_training - Step 14610: {'lr': 0.0004195799968050935, 'samples': 7480832, 'steps': 14610, 'loss/train': 3.2137739658355713} 02/24/2022 20:29:24 - INFO - codeparrot_training - Step 14611: {'lr': 0.000419567973851082, 'samples': 7481344, 'steps': 14611, 'loss/train': 2.1327872276306152} 02/24/2022 20:29:28 - INFO - codeparrot_training - Step 14612: {'lr': 0.0004195559501706951, 'samples': 7481856, 'steps': 14612, 'loss/train': 2.407708168029785} 02/24/2022 20:29:33 - INFO - codeparrot_training - Step 14613: {'lr': 0.00041954392576398433, 'samples': 7482368, 'steps': 14613, 'loss/train': 2.487175703048706} 02/24/2022 20:29:37 - INFO - codeparrot_training - Step 14614: {'lr': 0.0004195319006310012, 'samples': 7482880, 'steps': 14614, 'loss/train': 2.634536027908325} 02/24/2022 20:29:43 - INFO - codeparrot_training - Step 14615: {'lr': 0.0004195198747717973, 'samples': 7483392, 'steps': 14615, 'loss/train': 1.9193918704986572} 02/24/2022 20:29:47 - INFO - codeparrot_training - Step 14616: {'lr': 0.00041950784818642404, 'samples': 7483904, 'steps': 14616, 'loss/train': 2.3051199913024902} 02/24/2022 20:29:52 - INFO - codeparrot_training - Step 14617: {'lr': 0.000419495820874933, 'samples': 7484416, 'steps': 14617, 'loss/train': 2.5814459323883057} 02/24/2022 20:29:56 - INFO - codeparrot_training - Step 14618: {'lr': 0.0004194837928373757, 'samples': 7484928, 'steps': 14618, 'loss/train': 1.7086278200149536} 02/24/2022 20:30:01 - INFO - codeparrot_training - Step 14619: {'lr': 0.0004194717640738036, 'samples': 7485440, 'steps': 14619, 'loss/train': 2.177126884460449} 02/24/2022 20:30:05 - INFO - codeparrot_training - Step 14620: {'lr': 0.0004194597345842683, 'samples': 7485952, 'steps': 14620, 'loss/train': 2.3541409969329834} 02/24/2022 20:30:10 - INFO - codeparrot_training - Step 14621: {'lr': 0.00041944770436882134, 'samples': 7486464, 'steps': 14621, 'loss/train': 1.2967948913574219} 02/24/2022 20:30:14 - INFO - codeparrot_training - Step 14622: {'lr': 0.00041943567342751423, 'samples': 7486976, 'steps': 14622, 'loss/train': 1.6909550428390503} 02/24/2022 20:30:19 - INFO - codeparrot_training - Step 14623: {'lr': 0.0004194236417603985, 'samples': 7487488, 'steps': 14623, 'loss/train': 0.7737172245979309} 02/24/2022 20:30:23 - INFO - codeparrot_training - Step 14624: {'lr': 0.0004194116093675256, 'samples': 7488000, 'steps': 14624, 'loss/train': 0.6357579231262207} 02/24/2022 20:30:29 - INFO - codeparrot_training - Step 14625: {'lr': 0.0004193995762489472, 'samples': 7488512, 'steps': 14625, 'loss/train': 1.3895153999328613} 02/24/2022 20:30:33 - INFO - codeparrot_training - Step 14626: {'lr': 0.0004193875424047148, 'samples': 7489024, 'steps': 14626, 'loss/train': 2.646043062210083} 02/24/2022 20:30:38 - INFO - codeparrot_training - Step 14627: {'lr': 0.00041937550783488, 'samples': 7489536, 'steps': 14627, 'loss/train': 2.2098827362060547} 02/24/2022 20:30:42 - INFO - codeparrot_training - Step 14628: {'lr': 0.00041936347253949426, 'samples': 7490048, 'steps': 14628, 'loss/train': 1.5238178968429565} 02/24/2022 20:30:47 - INFO - codeparrot_training - Step 14629: {'lr': 0.00041935143651860917, 'samples': 7490560, 'steps': 14629, 'loss/train': 2.241729736328125} 02/24/2022 20:30:51 - INFO - codeparrot_training - Step 14630: {'lr': 0.0004193393997722764, 'samples': 7491072, 'steps': 14630, 'loss/train': 2.1297357082366943} 02/24/2022 20:30:56 - INFO - codeparrot_training - Step 14631: {'lr': 0.00041932736230054725, 'samples': 7491584, 'steps': 14631, 'loss/train': 2.4730610847473145} 02/24/2022 20:31:00 - INFO - codeparrot_training - Step 14632: {'lr': 0.0004193153241034736, 'samples': 7492096, 'steps': 14632, 'loss/train': 2.0761828422546387} 02/24/2022 20:31:06 - INFO - codeparrot_training - Step 14633: {'lr': 0.00041930328518110675, 'samples': 7492608, 'steps': 14633, 'loss/train': 2.089456796646118} 02/24/2022 20:31:09 - INFO - codeparrot_training - Step 14634: {'lr': 0.0004192912455334985, 'samples': 7493120, 'steps': 14634, 'loss/train': 2.841273069381714} 02/24/2022 20:31:15 - INFO - codeparrot_training - Step 14635: {'lr': 0.0004192792051607002, 'samples': 7493632, 'steps': 14635, 'loss/train': 2.551028251647949} 02/24/2022 20:31:19 - INFO - codeparrot_training - Step 14636: {'lr': 0.00041926716406276367, 'samples': 7494144, 'steps': 14636, 'loss/train': 1.873466968536377} 02/24/2022 20:31:24 - INFO - codeparrot_training - Step 14637: {'lr': 0.0004192551222397402, 'samples': 7494656, 'steps': 14637, 'loss/train': 1.8465358018875122} 02/24/2022 20:31:28 - INFO - codeparrot_training - Step 14638: {'lr': 0.0004192430796916816, 'samples': 7495168, 'steps': 14638, 'loss/train': 2.47739315032959} 02/24/2022 20:31:33 - INFO - codeparrot_training - Step 14639: {'lr': 0.0004192310364186394, 'samples': 7495680, 'steps': 14639, 'loss/train': 1.9412683248519897} 02/24/2022 20:31:37 - INFO - codeparrot_training - Step 14640: {'lr': 0.0004192189924206652, 'samples': 7496192, 'steps': 14640, 'loss/train': 1.650928258895874} 02/24/2022 20:31:42 - INFO - codeparrot_training - Step 14641: {'lr': 0.0004192069476978105, 'samples': 7496704, 'steps': 14641, 'loss/train': 1.1390290260314941} 02/24/2022 20:31:46 - INFO - codeparrot_training - Step 14642: {'lr': 0.000419194902250127, 'samples': 7497216, 'steps': 14642, 'loss/train': 1.85894775390625} 02/24/2022 20:31:51 - INFO - codeparrot_training - Step 14643: {'lr': 0.0004191828560776663, 'samples': 7497728, 'steps': 14643, 'loss/train': 0.8702483773231506} 02/24/2022 20:31:55 - INFO - codeparrot_training - Step 14644: {'lr': 0.00041917080918047996, 'samples': 7498240, 'steps': 14644, 'loss/train': 2.2353756427764893} 02/24/2022 20:32:00 - INFO - codeparrot_training - Step 14645: {'lr': 0.00041915876155861954, 'samples': 7498752, 'steps': 14645, 'loss/train': 1.6367141008377075} 02/24/2022 20:32:04 - INFO - codeparrot_training - Step 14646: {'lr': 0.0004191467132121367, 'samples': 7499264, 'steps': 14646, 'loss/train': 1.5821138620376587} 02/24/2022 20:32:09 - INFO - codeparrot_training - Step 14647: {'lr': 0.00041913466414108315, 'samples': 7499776, 'steps': 14647, 'loss/train': 2.1365838050842285} 02/24/2022 20:32:13 - INFO - codeparrot_training - Step 14648: {'lr': 0.0004191226143455103, 'samples': 7500288, 'steps': 14648, 'loss/train': 1.8549156188964844} 02/24/2022 20:32:19 - INFO - codeparrot_training - Step 14649: {'lr': 0.00041911056382546997, 'samples': 7500800, 'steps': 14649, 'loss/train': 0.9458596706390381} 02/24/2022 20:32:22 - INFO - codeparrot_training - Step 14650: {'lr': 0.00041909851258101357, 'samples': 7501312, 'steps': 14650, 'loss/train': 1.5656195878982544} 02/24/2022 20:32:26 - INFO - codeparrot_training - Step 14651: {'lr': 0.0004190864606121929, 'samples': 7501824, 'steps': 14651, 'loss/train': 3.424837112426758} 02/24/2022 20:32:33 - INFO - codeparrot_training - Step 14652: {'lr': 0.0004190744079190595, 'samples': 7502336, 'steps': 14652, 'loss/train': 1.837128758430481} 02/24/2022 20:32:38 - INFO - codeparrot_training - Step 14653: {'lr': 0.0004190623545016651, 'samples': 7502848, 'steps': 14653, 'loss/train': 2.3164870738983154} 02/24/2022 20:32:42 - INFO - codeparrot_training - Step 14654: {'lr': 0.00041905030036006106, 'samples': 7503360, 'steps': 14654, 'loss/train': 1.6023658514022827} 02/24/2022 20:32:47 - INFO - codeparrot_training - Step 14655: {'lr': 0.00041903824549429936, 'samples': 7503872, 'steps': 14655, 'loss/train': 0.9231467247009277} 02/24/2022 20:32:51 - INFO - codeparrot_training - Step 14656: {'lr': 0.00041902618990443156, 'samples': 7504384, 'steps': 14656, 'loss/train': 2.1635947227478027} 02/24/2022 20:32:57 - INFO - codeparrot_training - Step 14657: {'lr': 0.0004190141335905091, 'samples': 7504896, 'steps': 14657, 'loss/train': 1.8210846185684204} 02/24/2022 20:33:00 - INFO - codeparrot_training - Step 14658: {'lr': 0.0004190020765525838, 'samples': 7505408, 'steps': 14658, 'loss/train': 2.2828598022460938} 02/24/2022 20:33:06 - INFO - codeparrot_training - Step 14659: {'lr': 0.0004189900187907073, 'samples': 7505920, 'steps': 14659, 'loss/train': 1.8248789310455322} 02/24/2022 20:33:09 - INFO - codeparrot_training - Step 14660: {'lr': 0.0004189779603049312, 'samples': 7506432, 'steps': 14660, 'loss/train': 1.1439238786697388} 02/24/2022 20:33:15 - INFO - codeparrot_training - Step 14661: {'lr': 0.00041896590109530713, 'samples': 7506944, 'steps': 14661, 'loss/train': 1.7079750299453735} 02/24/2022 20:33:19 - INFO - codeparrot_training - Step 14662: {'lr': 0.00041895384116188685, 'samples': 7507456, 'steps': 14662, 'loss/train': 1.5416775941848755} 02/24/2022 20:33:24 - INFO - codeparrot_training - Step 14663: {'lr': 0.000418941780504722, 'samples': 7507968, 'steps': 14663, 'loss/train': 1.5117465257644653} 02/24/2022 20:33:28 - INFO - codeparrot_training - Step 14664: {'lr': 0.00041892971912386415, 'samples': 7508480, 'steps': 14664, 'loss/train': 2.0657784938812256} 02/24/2022 20:33:33 - INFO - codeparrot_training - Step 14665: {'lr': 0.000418917657019365, 'samples': 7508992, 'steps': 14665, 'loss/train': 2.7559168338775635} 02/24/2022 20:33:37 - INFO - codeparrot_training - Step 14666: {'lr': 0.0004189055941912763, 'samples': 7509504, 'steps': 14666, 'loss/train': 2.1266298294067383} 02/24/2022 20:33:40 - INFO - codeparrot_training - Step 14667: {'lr': 0.0004188935306396496, 'samples': 7510016, 'steps': 14667, 'loss/train': 2.2662570476531982} 02/24/2022 20:33:46 - INFO - codeparrot_training - Step 14668: {'lr': 0.00041888146636453674, 'samples': 7510528, 'steps': 14668, 'loss/train': 2.0040929317474365} 02/24/2022 20:33:50 - INFO - codeparrot_training - Step 14669: {'lr': 0.0004188694013659892, 'samples': 7511040, 'steps': 14669, 'loss/train': 2.4362144470214844} 02/24/2022 20:33:55 - INFO - codeparrot_training - Step 14670: {'lr': 0.0004188573356440588, 'samples': 7511552, 'steps': 14670, 'loss/train': 1.0447674989700317} 02/24/2022 20:33:59 - INFO - codeparrot_training - Step 14671: {'lr': 0.0004188452691987973, 'samples': 7512064, 'steps': 14671, 'loss/train': 2.4095516204833984} 02/24/2022 20:34:04 - INFO - codeparrot_training - Step 14672: {'lr': 0.0004188332020302561, 'samples': 7512576, 'steps': 14672, 'loss/train': 2.437413215637207} 02/24/2022 20:34:08 - INFO - codeparrot_training - Step 14673: {'lr': 0.0004188211341384872, 'samples': 7513088, 'steps': 14673, 'loss/train': 1.5840586423873901} 02/24/2022 20:34:14 - INFO - codeparrot_training - Step 14674: {'lr': 0.0004188090655235421, 'samples': 7513600, 'steps': 14674, 'loss/train': 3.0262913703918457} 02/24/2022 20:34:19 - INFO - codeparrot_training - Step 14675: {'lr': 0.00041879699618547263, 'samples': 7514112, 'steps': 14675, 'loss/train': 2.727795362472534} 02/24/2022 20:34:23 - INFO - codeparrot_training - Step 14676: {'lr': 0.0004187849261243304, 'samples': 7514624, 'steps': 14676, 'loss/train': 0.6858757734298706} 02/24/2022 20:34:28 - INFO - codeparrot_training - Step 14677: {'lr': 0.0004187728553401671, 'samples': 7515136, 'steps': 14677, 'loss/train': 2.3487091064453125} 02/24/2022 20:34:32 - INFO - codeparrot_training - Step 14678: {'lr': 0.0004187607838330345, 'samples': 7515648, 'steps': 14678, 'loss/train': 2.055164098739624} 02/24/2022 20:34:37 - INFO - codeparrot_training - Step 14679: {'lr': 0.0004187487116029843, 'samples': 7516160, 'steps': 14679, 'loss/train': 2.552825927734375} 02/24/2022 20:34:41 - INFO - codeparrot_training - Step 14680: {'lr': 0.0004187366386500683, 'samples': 7516672, 'steps': 14680, 'loss/train': 2.208420515060425} 02/24/2022 20:34:46 - INFO - codeparrot_training - Step 14681: {'lr': 0.00041872456497433797, 'samples': 7517184, 'steps': 14681, 'loss/train': 2.609623432159424} 02/24/2022 20:34:50 - INFO - codeparrot_training - Step 14682: {'lr': 0.00041871249057584526, 'samples': 7517696, 'steps': 14682, 'loss/train': 2.8586361408233643} 02/24/2022 20:34:57 - INFO - codeparrot_training - Step 14683: {'lr': 0.00041870041545464176, 'samples': 7518208, 'steps': 14683, 'loss/train': 2.1922357082366943} 02/24/2022 20:35:00 - INFO - codeparrot_training - Step 14684: {'lr': 0.00041868833961077935, 'samples': 7518720, 'steps': 14684, 'loss/train': 2.1248395442962646} 02/24/2022 20:35:06 - INFO - codeparrot_training - Step 14685: {'lr': 0.0004186762630443096, 'samples': 7519232, 'steps': 14685, 'loss/train': 1.9496482610702515} 02/24/2022 20:35:09 - INFO - codeparrot_training - Step 14686: {'lr': 0.0004186641857552842, 'samples': 7519744, 'steps': 14686, 'loss/train': 2.205439805984497} 02/24/2022 20:35:15 - INFO - codeparrot_training - Step 14687: {'lr': 0.0004186521077437551, 'samples': 7520256, 'steps': 14687, 'loss/train': 1.977354884147644} 02/24/2022 20:35:18 - INFO - codeparrot_training - Step 14688: {'lr': 0.00041864002900977393, 'samples': 7520768, 'steps': 14688, 'loss/train': 2.500113010406494} 02/24/2022 20:35:24 - INFO - codeparrot_training - Step 14689: {'lr': 0.0004186279495533923, 'samples': 7521280, 'steps': 14689, 'loss/train': 1.9217275381088257} 02/24/2022 20:35:27 - INFO - codeparrot_training - Step 14690: {'lr': 0.0004186158693746622, 'samples': 7521792, 'steps': 14690, 'loss/train': 2.2815613746643066} 02/24/2022 20:35:33 - INFO - codeparrot_training - Step 14691: {'lr': 0.0004186037884736352, 'samples': 7522304, 'steps': 14691, 'loss/train': 2.5176374912261963} 02/24/2022 20:35:37 - INFO - codeparrot_training - Step 14692: {'lr': 0.0004185917068503632, 'samples': 7522816, 'steps': 14692, 'loss/train': 1.900394320487976} 02/24/2022 20:35:40 - INFO - codeparrot_training - Step 14693: {'lr': 0.00041857962450489786, 'samples': 7523328, 'steps': 14693, 'loss/train': 0.40461859107017517} 02/24/2022 20:35:46 - INFO - codeparrot_training - Step 14694: {'lr': 0.0004185675414372908, 'samples': 7523840, 'steps': 14694, 'loss/train': 2.1908669471740723} 02/24/2022 20:35:52 - INFO - codeparrot_training - Step 14695: {'lr': 0.000418555457647594, 'samples': 7524352, 'steps': 14695, 'loss/train': 2.508284330368042} 02/24/2022 20:35:55 - INFO - codeparrot_training - Step 14696: {'lr': 0.00041854337313585913, 'samples': 7524864, 'steps': 14696, 'loss/train': 0.7164973616600037} 02/24/2022 20:36:01 - INFO - codeparrot_training - Step 14697: {'lr': 0.00041853128790213804, 'samples': 7525376, 'steps': 14697, 'loss/train': 1.2111140489578247} 02/24/2022 20:36:05 - INFO - codeparrot_training - Step 14698: {'lr': 0.0004185192019464823, 'samples': 7525888, 'steps': 14698, 'loss/train': 1.7549132108688354} 02/24/2022 20:36:10 - INFO - codeparrot_training - Step 14699: {'lr': 0.0004185071152689439, 'samples': 7526400, 'steps': 14699, 'loss/train': 1.9711965322494507} 02/24/2022 20:36:14 - INFO - codeparrot_training - Step 14700: {'lr': 0.0004184950278695745, 'samples': 7526912, 'steps': 14700, 'loss/train': 1.5679363012313843} 02/24/2022 20:36:19 - INFO - codeparrot_training - Step 14701: {'lr': 0.0004184829397484259, 'samples': 7527424, 'steps': 14701, 'loss/train': 1.7270411252975464} 02/24/2022 20:36:23 - INFO - codeparrot_training - Step 14702: {'lr': 0.00041847085090554985, 'samples': 7527936, 'steps': 14702, 'loss/train': 2.149169921875} 02/24/2022 20:36:28 - INFO - codeparrot_training - Step 14703: {'lr': 0.00041845876134099825, 'samples': 7528448, 'steps': 14703, 'loss/train': 2.5090620517730713} 02/24/2022 20:36:32 - INFO - codeparrot_training - Step 14704: {'lr': 0.0004184466710548227, 'samples': 7528960, 'steps': 14704, 'loss/train': 2.6662349700927734} 02/24/2022 20:36:37 - INFO - codeparrot_training - Step 14705: {'lr': 0.0004184345800470752, 'samples': 7529472, 'steps': 14705, 'loss/train': 2.8832342624664307} 02/24/2022 20:36:41 - INFO - codeparrot_training - Step 14706: {'lr': 0.00041842248831780736, 'samples': 7529984, 'steps': 14706, 'loss/train': 2.292116165161133} 02/24/2022 20:36:47 - INFO - codeparrot_training - Step 14707: {'lr': 0.0004184103958670712, 'samples': 7530496, 'steps': 14707, 'loss/train': 1.8573864698410034} 02/24/2022 20:36:50 - INFO - codeparrot_training - Step 14708: {'lr': 0.00041839830269491823, 'samples': 7531008, 'steps': 14708, 'loss/train': 2.167811870574951} 02/24/2022 20:36:56 - INFO - codeparrot_training - Step 14709: {'lr': 0.00041838620880140046, 'samples': 7531520, 'steps': 14709, 'loss/train': 1.3617523908615112} 02/24/2022 20:36:59 - INFO - codeparrot_training - Step 14710: {'lr': 0.00041837411418656965, 'samples': 7532032, 'steps': 14710, 'loss/train': 0.7553223371505737} 02/24/2022 20:37:05 - INFO - codeparrot_training - Step 14711: {'lr': 0.0004183620188504776, 'samples': 7532544, 'steps': 14711, 'loss/train': 1.7091599702835083} 02/24/2022 20:37:08 - INFO - codeparrot_training - Step 14712: {'lr': 0.0004183499227931761, 'samples': 7533056, 'steps': 14712, 'loss/train': 1.4914801120758057} 02/24/2022 20:37:14 - INFO - codeparrot_training - Step 14713: {'lr': 0.00041833782601471704, 'samples': 7533568, 'steps': 14713, 'loss/train': 1.6754792928695679} 02/24/2022 20:37:17 - INFO - codeparrot_training - Step 14714: {'lr': 0.0004183257285151521, 'samples': 7534080, 'steps': 14714, 'loss/train': 1.8323981761932373} 02/24/2022 20:37:23 - INFO - codeparrot_training - Step 14715: {'lr': 0.00041831363029453327, 'samples': 7534592, 'steps': 14715, 'loss/train': 2.3271398544311523} 02/24/2022 20:37:26 - INFO - codeparrot_training - Step 14716: {'lr': 0.0004183015313529123, 'samples': 7535104, 'steps': 14716, 'loss/train': 2.863957405090332} 02/24/2022 20:37:33 - INFO - codeparrot_training - Step 14717: {'lr': 0.00041828943169034094, 'samples': 7535616, 'steps': 14717, 'loss/train': 1.5282034873962402} 02/24/2022 20:37:36 - INFO - codeparrot_training - Step 14718: {'lr': 0.0004182773313068711, 'samples': 7536128, 'steps': 14718, 'loss/train': 2.154193878173828} 02/24/2022 20:37:42 - INFO - codeparrot_training - Step 14719: {'lr': 0.00041826523020255463, 'samples': 7536640, 'steps': 14719, 'loss/train': 1.8944965600967407} 02/24/2022 20:37:45 - INFO - codeparrot_training - Step 14720: {'lr': 0.00041825312837744333, 'samples': 7537152, 'steps': 14720, 'loss/train': 1.7460154294967651} 02/24/2022 20:37:51 - INFO - codeparrot_training - Step 14721: {'lr': 0.00041824102583158906, 'samples': 7537664, 'steps': 14721, 'loss/train': 3.624465227127075} 02/24/2022 20:37:54 - INFO - codeparrot_training - Step 14722: {'lr': 0.0004182289225650437, 'samples': 7538176, 'steps': 14722, 'loss/train': 2.5229992866516113} 02/24/2022 20:38:00 - INFO - codeparrot_training - Step 14723: {'lr': 0.00041821681857785904, 'samples': 7538688, 'steps': 14723, 'loss/train': 7.863982200622559} 02/24/2022 20:38:03 - INFO - codeparrot_training - Step 14724: {'lr': 0.0004182047138700869, 'samples': 7539200, 'steps': 14724, 'loss/train': 3.5256266593933105} 02/24/2022 20:38:09 - INFO - codeparrot_training - Step 14725: {'lr': 0.0004181926084417792, 'samples': 7539712, 'steps': 14725, 'loss/train': 1.2098037004470825} 02/24/2022 20:38:12 - INFO - codeparrot_training - Step 14726: {'lr': 0.0004181805022929878, 'samples': 7540224, 'steps': 14726, 'loss/train': 1.508366584777832} 02/24/2022 20:38:18 - INFO - codeparrot_training - Step 14727: {'lr': 0.0004181683954237645, 'samples': 7540736, 'steps': 14727, 'loss/train': 2.637098550796509} 02/24/2022 20:38:22 - INFO - codeparrot_training - Step 14728: {'lr': 0.00041815628783416117, 'samples': 7541248, 'steps': 14728, 'loss/train': 2.6529407501220703} 02/24/2022 20:38:27 - INFO - codeparrot_training - Step 14729: {'lr': 0.00041814417952422975, 'samples': 7541760, 'steps': 14729, 'loss/train': 1.285556674003601} 02/24/2022 20:38:31 - INFO - codeparrot_training - Step 14730: {'lr': 0.000418132070494022, 'samples': 7542272, 'steps': 14730, 'loss/train': 0.8838563561439514} 02/24/2022 20:38:37 - INFO - codeparrot_training - Step 14731: {'lr': 0.00041811996074358993, 'samples': 7542784, 'steps': 14731, 'loss/train': 2.138936996459961} 02/24/2022 20:38:40 - INFO - codeparrot_training - Step 14732: {'lr': 0.00041810785027298524, 'samples': 7543296, 'steps': 14732, 'loss/train': 2.455162763595581} 02/24/2022 20:38:46 - INFO - codeparrot_training - Step 14733: {'lr': 0.00041809573908225997, 'samples': 7543808, 'steps': 14733, 'loss/train': 2.405764579772949} 02/24/2022 20:38:49 - INFO - codeparrot_training - Step 14734: {'lr': 0.00041808362717146594, 'samples': 7544320, 'steps': 14734, 'loss/train': 1.3611860275268555} 02/24/2022 20:38:55 - INFO - codeparrot_training - Step 14735: {'lr': 0.00041807151454065493, 'samples': 7544832, 'steps': 14735, 'loss/train': 1.5316896438598633} 02/24/2022 20:38:58 - INFO - codeparrot_training - Step 14736: {'lr': 0.00041805940118987904, 'samples': 7545344, 'steps': 14736, 'loss/train': 1.4424585103988647} 02/24/2022 20:39:04 - INFO - codeparrot_training - Step 14737: {'lr': 0.0004180472871191899, 'samples': 7545856, 'steps': 14737, 'loss/train': 2.603076219558716} 02/24/2022 20:39:07 - INFO - codeparrot_training - Step 14738: {'lr': 0.0004180351723286396, 'samples': 7546368, 'steps': 14738, 'loss/train': 2.431905508041382} 02/24/2022 20:39:13 - INFO - codeparrot_training - Step 14739: {'lr': 0.00041802305681828007, 'samples': 7546880, 'steps': 14739, 'loss/train': 2.0364010334014893} 02/24/2022 20:39:16 - INFO - codeparrot_training - Step 14740: {'lr': 0.00041801094058816304, 'samples': 7547392, 'steps': 14740, 'loss/train': 2.643197774887085} 02/24/2022 20:39:22 - INFO - codeparrot_training - Step 14741: {'lr': 0.0004179988236383405, 'samples': 7547904, 'steps': 14741, 'loss/train': 1.226442575454712} 02/24/2022 20:39:25 - INFO - codeparrot_training - Step 14742: {'lr': 0.00041798670596886433, 'samples': 7548416, 'steps': 14742, 'loss/train': 2.3600006103515625} 02/24/2022 20:39:31 - INFO - codeparrot_training - Step 14743: {'lr': 0.00041797458757978647, 'samples': 7548928, 'steps': 14743, 'loss/train': 1.5221009254455566} 02/24/2022 20:39:34 - INFO - codeparrot_training - Step 14744: {'lr': 0.0004179624684711588, 'samples': 7549440, 'steps': 14744, 'loss/train': 2.237016439437866} 02/24/2022 20:39:40 - INFO - codeparrot_training - Step 14745: {'lr': 0.0004179503486430333, 'samples': 7549952, 'steps': 14745, 'loss/train': 1.6865696907043457} 02/24/2022 20:39:44 - INFO - codeparrot_training - Step 14746: {'lr': 0.00041793822809546176, 'samples': 7550464, 'steps': 14746, 'loss/train': 1.893465518951416} 02/24/2022 20:39:50 - INFO - codeparrot_training - Step 14747: {'lr': 0.0004179261068284963, 'samples': 7550976, 'steps': 14747, 'loss/train': 1.814063310623169} 02/24/2022 20:39:53 - INFO - codeparrot_training - Step 14748: {'lr': 0.00041791398484218855, 'samples': 7551488, 'steps': 14748, 'loss/train': 1.319667100906372} 02/24/2022 20:39:59 - INFO - codeparrot_training - Step 14749: {'lr': 0.0004179018621365908, 'samples': 7552000, 'steps': 14749, 'loss/train': 0.8811789155006409} 02/24/2022 20:40:02 - INFO - codeparrot_training - Step 14750: {'lr': 0.00041788973871175465, 'samples': 7552512, 'steps': 14750, 'loss/train': 2.1467764377593994} 02/24/2022 20:40:08 - INFO - codeparrot_training - Step 14751: {'lr': 0.00041787761456773214, 'samples': 7553024, 'steps': 14751, 'loss/train': 2.9769864082336426} 02/24/2022 20:40:11 - INFO - codeparrot_training - Step 14752: {'lr': 0.00041786548970457535, 'samples': 7553536, 'steps': 14752, 'loss/train': 2.068859815597534} 02/24/2022 20:40:17 - INFO - codeparrot_training - Step 14753: {'lr': 0.000417853364122336, 'samples': 7554048, 'steps': 14753, 'loss/train': 2.2789199352264404} 02/24/2022 20:40:20 - INFO - codeparrot_training - Step 14754: {'lr': 0.0004178412378210662, 'samples': 7554560, 'steps': 14754, 'loss/train': 3.3096373081207275} 02/24/2022 20:40:26 - INFO - codeparrot_training - Step 14755: {'lr': 0.0004178291108008179, 'samples': 7555072, 'steps': 14755, 'loss/train': 0.914291501045227} 02/24/2022 20:40:30 - INFO - codeparrot_training - Step 14756: {'lr': 0.00041781698306164283, 'samples': 7555584, 'steps': 14756, 'loss/train': 1.2013415098190308} 02/24/2022 20:40:35 - INFO - codeparrot_training - Step 14757: {'lr': 0.0004178048546035932, 'samples': 7556096, 'steps': 14757, 'loss/train': 1.8637973070144653} 02/24/2022 20:40:39 - INFO - codeparrot_training - Step 14758: {'lr': 0.00041779272542672086, 'samples': 7556608, 'steps': 14758, 'loss/train': 1.29334557056427} 02/24/2022 20:40:44 - INFO - codeparrot_training - Step 14759: {'lr': 0.00041778059553107766, 'samples': 7557120, 'steps': 14759, 'loss/train': 2.095696449279785} 02/24/2022 20:40:48 - INFO - codeparrot_training - Step 14760: {'lr': 0.00041776846491671575, 'samples': 7557632, 'steps': 14760, 'loss/train': 1.901415228843689} 02/24/2022 20:40:53 - INFO - codeparrot_training - Step 14761: {'lr': 0.000417756333583687, 'samples': 7558144, 'steps': 14761, 'loss/train': 2.9715511798858643} 02/24/2022 20:40:57 - INFO - codeparrot_training - Step 14762: {'lr': 0.0004177442015320434, 'samples': 7558656, 'steps': 14762, 'loss/train': 1.8797039985656738} 02/24/2022 20:41:02 - INFO - codeparrot_training - Step 14763: {'lr': 0.0004177320687618369, 'samples': 7559168, 'steps': 14763, 'loss/train': 2.5225167274475098} 02/24/2022 20:41:06 - INFO - codeparrot_training - Step 14764: {'lr': 0.0004177199352731194, 'samples': 7559680, 'steps': 14764, 'loss/train': 2.6675233840942383} 02/24/2022 20:41:12 - INFO - codeparrot_training - Step 14765: {'lr': 0.0004177078010659431, 'samples': 7560192, 'steps': 14765, 'loss/train': 7.817892074584961} 02/24/2022 20:41:16 - INFO - codeparrot_training - Step 14766: {'lr': 0.0004176956661403597, 'samples': 7560704, 'steps': 14766, 'loss/train': 2.0508601665496826} 02/24/2022 20:41:19 - INFO - codeparrot_training - Step 14767: {'lr': 0.0004176835304964214, 'samples': 7561216, 'steps': 14767, 'loss/train': 1.1477147340774536} 02/24/2022 20:41:25 - INFO - codeparrot_training - Step 14768: {'lr': 0.00041767139413418, 'samples': 7561728, 'steps': 14768, 'loss/train': 2.1556835174560547} 02/24/2022 20:41:28 - INFO - codeparrot_training - Step 14769: {'lr': 0.00041765925705368766, 'samples': 7562240, 'steps': 14769, 'loss/train': 1.9438164234161377} 02/24/2022 20:41:34 - INFO - codeparrot_training - Step 14770: {'lr': 0.00041764711925499633, 'samples': 7562752, 'steps': 14770, 'loss/train': 2.140820026397705} 02/24/2022 20:41:37 - INFO - codeparrot_training - Step 14771: {'lr': 0.0004176349807381579, 'samples': 7563264, 'steps': 14771, 'loss/train': 2.330643892288208} 02/24/2022 20:41:43 - INFO - codeparrot_training - Step 14772: {'lr': 0.0004176228415032245, 'samples': 7563776, 'steps': 14772, 'loss/train': 1.9728665351867676} 02/24/2022 20:41:46 - INFO - codeparrot_training - Step 14773: {'lr': 0.000417610701550248, 'samples': 7564288, 'steps': 14773, 'loss/train': 2.3594679832458496} 02/24/2022 20:41:52 - INFO - codeparrot_training - Step 14774: {'lr': 0.0004175985608792806, 'samples': 7564800, 'steps': 14774, 'loss/train': 2.7506489753723145} 02/24/2022 20:41:55 - INFO - codeparrot_training - Step 14775: {'lr': 0.00041758641949037414, 'samples': 7565312, 'steps': 14775, 'loss/train': 1.4936597347259521} 02/24/2022 20:42:02 - INFO - codeparrot_training - Step 14776: {'lr': 0.00041757427738358066, 'samples': 7565824, 'steps': 14776, 'loss/train': 2.4827866554260254} 02/24/2022 20:42:05 - INFO - codeparrot_training - Step 14777: {'lr': 0.00041756213455895215, 'samples': 7566336, 'steps': 14777, 'loss/train': 1.9582109451293945} 02/24/2022 20:42:11 - INFO - codeparrot_training - Step 14778: {'lr': 0.00041754999101654066, 'samples': 7566848, 'steps': 14778, 'loss/train': 2.7278366088867188} 02/24/2022 20:42:14 - INFO - codeparrot_training - Step 14779: {'lr': 0.0004175378467563983, 'samples': 7567360, 'steps': 14779, 'loss/train': 1.2823923826217651} 02/24/2022 20:42:20 - INFO - codeparrot_training - Step 14780: {'lr': 0.00041752570177857695, 'samples': 7567872, 'steps': 14780, 'loss/train': 2.1559641361236572} 02/24/2022 20:42:23 - INFO - codeparrot_training - Step 14781: {'lr': 0.0004175135560831287, 'samples': 7568384, 'steps': 14781, 'loss/train': 1.3038510084152222} 02/24/2022 20:42:29 - INFO - codeparrot_training - Step 14782: {'lr': 0.00041750140967010554, 'samples': 7568896, 'steps': 14782, 'loss/train': 1.337658166885376} 02/24/2022 20:42:32 - INFO - codeparrot_training - Step 14783: {'lr': 0.00041748926253955954, 'samples': 7569408, 'steps': 14783, 'loss/train': 1.157994031906128} 02/24/2022 20:42:38 - INFO - codeparrot_training - Step 14784: {'lr': 0.0004174771146915427, 'samples': 7569920, 'steps': 14784, 'loss/train': 1.7247217893600464} 02/24/2022 20:42:41 - INFO - codeparrot_training - Step 14785: {'lr': 0.00041746496612610705, 'samples': 7570432, 'steps': 14785, 'loss/train': 2.2335851192474365} 02/24/2022 20:42:47 - INFO - codeparrot_training - Step 14786: {'lr': 0.00041745281684330476, 'samples': 7570944, 'steps': 14786, 'loss/train': 1.3352926969528198} 02/24/2022 20:42:50 - INFO - codeparrot_training - Step 14787: {'lr': 0.0004174406668431877, 'samples': 7571456, 'steps': 14787, 'loss/train': 2.600802421569824} 02/24/2022 20:42:56 - INFO - codeparrot_training - Step 14788: {'lr': 0.000417428516125808, 'samples': 7571968, 'steps': 14788, 'loss/train': 1.2635252475738525} 02/24/2022 20:43:02 - INFO - codeparrot_training - Step 14789: {'lr': 0.0004174163646912178, 'samples': 7572480, 'steps': 14789, 'loss/train': 1.9817677736282349} 02/24/2022 20:43:06 - INFO - codeparrot_training - Step 14790: {'lr': 0.0004174042125394689, 'samples': 7572992, 'steps': 14790, 'loss/train': 1.7664518356323242} 02/24/2022 20:43:11 - INFO - codeparrot_training - Step 14791: {'lr': 0.00041739205967061366, 'samples': 7573504, 'steps': 14791, 'loss/train': 1.8326669931411743} 02/24/2022 20:43:15 - INFO - codeparrot_training - Step 14792: {'lr': 0.0004173799060847039, 'samples': 7574016, 'steps': 14792, 'loss/train': 1.4710299968719482} 02/24/2022 20:43:18 - INFO - codeparrot_training - Step 14793: {'lr': 0.00041736775178179174, 'samples': 7574528, 'steps': 14793, 'loss/train': 2.086339235305786} 02/24/2022 20:43:24 - INFO - codeparrot_training - Step 14794: {'lr': 0.0004173555967619294, 'samples': 7575040, 'steps': 14794, 'loss/train': 1.2483069896697998} 02/24/2022 20:43:27 - INFO - codeparrot_training - Step 14795: {'lr': 0.00041734344102516873, 'samples': 7575552, 'steps': 14795, 'loss/train': 1.9300806522369385} 02/24/2022 20:43:33 - INFO - codeparrot_training - Step 14796: {'lr': 0.0004173312845715619, 'samples': 7576064, 'steps': 14796, 'loss/train': 2.8084537982940674} 02/24/2022 20:43:37 - INFO - codeparrot_training - Step 14797: {'lr': 0.000417319127401161, 'samples': 7576576, 'steps': 14797, 'loss/train': 2.187028169631958} 02/24/2022 20:43:42 - INFO - codeparrot_training - Step 14798: {'lr': 0.00041730696951401816, 'samples': 7577088, 'steps': 14798, 'loss/train': 1.3990521430969238} 02/24/2022 20:43:46 - INFO - codeparrot_training - Step 14799: {'lr': 0.00041729481091018527, 'samples': 7577600, 'steps': 14799, 'loss/train': 2.188141107559204} 02/24/2022 20:43:51 - INFO - codeparrot_training - Step 14800: {'lr': 0.0004172826515897146, 'samples': 7578112, 'steps': 14800, 'loss/train': 3.5938236713409424} 02/24/2022 20:43:54 - INFO - codeparrot_training - Step 14801: {'lr': 0.0004172704915526581, 'samples': 7578624, 'steps': 14801, 'loss/train': 2.7196619510650635} 02/24/2022 20:44:01 - INFO - codeparrot_training - Step 14802: {'lr': 0.000417258330799068, 'samples': 7579136, 'steps': 14802, 'loss/train': 2.588388442993164} 02/24/2022 20:44:06 - INFO - codeparrot_training - Step 14803: {'lr': 0.00041724616932899627, 'samples': 7579648, 'steps': 14803, 'loss/train': 2.099151611328125} 02/24/2022 20:44:10 - INFO - codeparrot_training - Step 14804: {'lr': 0.0004172340071424951, 'samples': 7580160, 'steps': 14804, 'loss/train': 0.7498244643211365} 02/24/2022 20:44:15 - INFO - codeparrot_training - Step 14805: {'lr': 0.0004172218442396165, 'samples': 7580672, 'steps': 14805, 'loss/train': 1.2891371250152588} 02/24/2022 20:44:19 - INFO - codeparrot_training - Step 14806: {'lr': 0.00041720968062041266, 'samples': 7581184, 'steps': 14806, 'loss/train': 2.330594062805176} 02/24/2022 20:44:24 - INFO - codeparrot_training - Step 14807: {'lr': 0.0004171975162849356, 'samples': 7581696, 'steps': 14807, 'loss/train': 1.589237928390503} 02/24/2022 20:44:28 - INFO - codeparrot_training - Step 14808: {'lr': 0.0004171853512332375, 'samples': 7582208, 'steps': 14808, 'loss/train': 1.9926923513412476} 02/24/2022 20:44:33 - INFO - codeparrot_training - Step 14809: {'lr': 0.00041717318546537045, 'samples': 7582720, 'steps': 14809, 'loss/train': 2.358306407928467} 02/24/2022 20:44:37 - INFO - codeparrot_training - Step 14810: {'lr': 0.0004171610189813866, 'samples': 7583232, 'steps': 14810, 'loss/train': 2.5473005771636963} 02/24/2022 20:44:43 - INFO - codeparrot_training - Step 14811: {'lr': 0.000417148851781338, 'samples': 7583744, 'steps': 14811, 'loss/train': 2.297606945037842} 02/24/2022 20:44:46 - INFO - codeparrot_training - Step 14812: {'lr': 0.0004171366838652767, 'samples': 7584256, 'steps': 14812, 'loss/train': 2.0433497428894043} 02/24/2022 20:44:52 - INFO - codeparrot_training - Step 14813: {'lr': 0.000417124515233255, 'samples': 7584768, 'steps': 14813, 'loss/train': 1.1630451679229736} 02/24/2022 20:44:55 - INFO - codeparrot_training - Step 14814: {'lr': 0.00041711234588532497, 'samples': 7585280, 'steps': 14814, 'loss/train': 2.403754472732544} 02/24/2022 20:45:01 - INFO - codeparrot_training - Step 14815: {'lr': 0.0004171001758215387, 'samples': 7585792, 'steps': 14815, 'loss/train': 1.3927111625671387} 02/24/2022 20:45:04 - INFO - codeparrot_training - Step 14816: {'lr': 0.0004170880050419483, 'samples': 7586304, 'steps': 14816, 'loss/train': 1.7989267110824585} 02/24/2022 20:45:10 - INFO - codeparrot_training - Step 14817: {'lr': 0.00041707583354660597, 'samples': 7586816, 'steps': 14817, 'loss/train': 2.126197099685669} 02/24/2022 20:45:13 - INFO - codeparrot_training - Step 14818: {'lr': 0.0004170636613355638, 'samples': 7587328, 'steps': 14818, 'loss/train': 1.4171420335769653} 02/24/2022 20:45:19 - INFO - codeparrot_training - Step 14819: {'lr': 0.000417051488408874, 'samples': 7587840, 'steps': 14819, 'loss/train': 1.7377910614013672} 02/24/2022 20:45:22 - INFO - codeparrot_training - Step 14820: {'lr': 0.00041703931476658857, 'samples': 7588352, 'steps': 14820, 'loss/train': 1.913621187210083} 02/24/2022 20:45:29 - INFO - codeparrot_training - Step 14821: {'lr': 0.0004170271404087598, 'samples': 7588864, 'steps': 14821, 'loss/train': 1.0714280605316162} 02/24/2022 20:45:33 - INFO - codeparrot_training - Step 14822: {'lr': 0.0004170149653354398, 'samples': 7589376, 'steps': 14822, 'loss/train': 2.133225679397583} 02/24/2022 20:45:38 - INFO - codeparrot_training - Step 14823: {'lr': 0.0004170027895466807, 'samples': 7589888, 'steps': 14823, 'loss/train': 1.0843403339385986} 02/24/2022 20:45:42 - INFO - codeparrot_training - Step 14824: {'lr': 0.00041699061304253476, 'samples': 7590400, 'steps': 14824, 'loss/train': 2.45759916305542} 02/24/2022 20:45:47 - INFO - codeparrot_training - Step 14825: {'lr': 0.00041697843582305406, 'samples': 7590912, 'steps': 14825, 'loss/train': 0.9926760792732239} 02/24/2022 20:45:51 - INFO - codeparrot_training - Step 14826: {'lr': 0.0004169662578882907, 'samples': 7591424, 'steps': 14826, 'loss/train': 2.2052457332611084} 02/24/2022 20:45:56 - INFO - codeparrot_training - Step 14827: {'lr': 0.0004169540792382969, 'samples': 7591936, 'steps': 14827, 'loss/train': 2.6973533630371094} 02/24/2022 20:46:00 - INFO - codeparrot_training - Step 14828: {'lr': 0.0004169418998731249, 'samples': 7592448, 'steps': 14828, 'loss/train': 2.4513397216796875} 02/24/2022 20:46:05 - INFO - codeparrot_training - Step 14829: {'lr': 0.0004169297197928268, 'samples': 7592960, 'steps': 14829, 'loss/train': 0.5002651214599609} 02/24/2022 20:46:09 - INFO - codeparrot_training - Step 14830: {'lr': 0.0004169175389974548, 'samples': 7593472, 'steps': 14830, 'loss/train': 1.257468581199646} 02/24/2022 20:46:14 - INFO - codeparrot_training - Step 14831: {'lr': 0.0004169053574870609, 'samples': 7593984, 'steps': 14831, 'loss/train': 1.9362249374389648} 02/24/2022 20:46:18 - INFO - codeparrot_training - Step 14832: {'lr': 0.0004168931752616977, 'samples': 7594496, 'steps': 14832, 'loss/train': 2.6308228969573975} 02/24/2022 20:46:23 - INFO - codeparrot_training - Step 14833: {'lr': 0.00041688099232141694, 'samples': 7595008, 'steps': 14833, 'loss/train': 2.483644485473633} 02/24/2022 20:46:27 - INFO - codeparrot_training - Step 14834: {'lr': 0.0004168688086662711, 'samples': 7595520, 'steps': 14834, 'loss/train': 2.298835039138794} 02/24/2022 20:46:32 - INFO - codeparrot_training - Step 14835: {'lr': 0.0004168566242963122, 'samples': 7596032, 'steps': 14835, 'loss/train': 2.0751938819885254} 02/24/2022 20:46:35 - INFO - codeparrot_training - Step 14836: {'lr': 0.00041684443921159253, 'samples': 7596544, 'steps': 14836, 'loss/train': 2.780609607696533} 02/24/2022 20:46:42 - INFO - codeparrot_training - Step 14837: {'lr': 0.00041683225341216426, 'samples': 7597056, 'steps': 14837, 'loss/train': 2.2594830989837646} 02/24/2022 20:46:45 - INFO - codeparrot_training - Step 14838: {'lr': 0.0004168200668980796, 'samples': 7597568, 'steps': 14838, 'loss/train': 1.9914158582687378} 02/24/2022 20:46:51 - INFO - codeparrot_training - Step 14839: {'lr': 0.0004168078796693908, 'samples': 7598080, 'steps': 14839, 'loss/train': 1.541532278060913} 02/24/2022 20:46:54 - INFO - codeparrot_training - Step 14840: {'lr': 0.00041679569172614996, 'samples': 7598592, 'steps': 14840, 'loss/train': 2.2549946308135986} 02/24/2022 20:47:00 - INFO - codeparrot_training - Step 14841: {'lr': 0.0004167835030684093, 'samples': 7599104, 'steps': 14841, 'loss/train': 1.0146172046661377} 02/24/2022 20:47:03 - INFO - codeparrot_training - Step 14842: {'lr': 0.0004167713136962211, 'samples': 7599616, 'steps': 14842, 'loss/train': 1.6750144958496094} 02/24/2022 20:47:09 - INFO - codeparrot_training - Step 14843: {'lr': 0.00041675912360963766, 'samples': 7600128, 'steps': 14843, 'loss/train': 2.6504602432250977} 02/24/2022 20:47:12 - INFO - codeparrot_training - Step 14844: {'lr': 0.0004167469328087109, 'samples': 7600640, 'steps': 14844, 'loss/train': 2.073831796646118} 02/24/2022 20:47:18 - INFO - codeparrot_training - Step 14845: {'lr': 0.0004167347412934933, 'samples': 7601152, 'steps': 14845, 'loss/train': 2.2274749279022217} 02/24/2022 20:47:21 - INFO - codeparrot_training - Step 14846: {'lr': 0.00041672254906403703, 'samples': 7601664, 'steps': 14846, 'loss/train': 1.2622736692428589} 02/24/2022 20:47:27 - INFO - codeparrot_training - Step 14847: {'lr': 0.00041671035612039434, 'samples': 7602176, 'steps': 14847, 'loss/train': 3.036304235458374} 02/24/2022 20:47:31 - INFO - codeparrot_training - Step 14848: {'lr': 0.0004166981624626174, 'samples': 7602688, 'steps': 14848, 'loss/train': 2.4200854301452637} 02/24/2022 20:47:36 - INFO - codeparrot_training - Step 14849: {'lr': 0.00041668596809075835, 'samples': 7603200, 'steps': 14849, 'loss/train': 2.3739476203918457} 02/24/2022 20:47:40 - INFO - codeparrot_training - Step 14850: {'lr': 0.0004166737730048697, 'samples': 7603712, 'steps': 14850, 'loss/train': 1.7012600898742676} 02/24/2022 20:47:45 - INFO - codeparrot_training - Step 14851: {'lr': 0.00041666157720500344, 'samples': 7604224, 'steps': 14851, 'loss/train': 2.3902628421783447} 02/24/2022 20:47:49 - INFO - codeparrot_training - Step 14852: {'lr': 0.00041664938069121195, 'samples': 7604736, 'steps': 14852, 'loss/train': 1.6821991205215454} 02/24/2022 20:47:54 - INFO - codeparrot_training - Step 14853: {'lr': 0.0004166371834635474, 'samples': 7605248, 'steps': 14853, 'loss/train': 2.3953053951263428} 02/24/2022 20:47:58 - INFO - codeparrot_training - Step 14854: {'lr': 0.00041662498552206206, 'samples': 7605760, 'steps': 14854, 'loss/train': 1.5478614568710327} 02/24/2022 20:48:03 - INFO - codeparrot_training - Step 14855: {'lr': 0.00041661278686680827, 'samples': 7606272, 'steps': 14855, 'loss/train': 2.187920570373535} 02/24/2022 20:48:07 - INFO - codeparrot_training - Step 14856: {'lr': 0.00041660058749783813, 'samples': 7606784, 'steps': 14856, 'loss/train': 1.7837194204330444} 02/24/2022 20:48:13 - INFO - codeparrot_training - Step 14857: {'lr': 0.000416588387415204, 'samples': 7607296, 'steps': 14857, 'loss/train': 1.7008607387542725} 02/24/2022 20:48:17 - INFO - codeparrot_training - Step 14858: {'lr': 0.0004165761866189581, 'samples': 7607808, 'steps': 14858, 'loss/train': 1.649808406829834} 02/24/2022 20:48:22 - INFO - codeparrot_training - Step 14859: {'lr': 0.00041656398510915273, 'samples': 7608320, 'steps': 14859, 'loss/train': 1.8994311094284058} 02/24/2022 20:48:26 - INFO - codeparrot_training - Step 14860: {'lr': 0.00041655178288584006, 'samples': 7608832, 'steps': 14860, 'loss/train': 1.8425670862197876} 02/24/2022 20:48:31 - INFO - codeparrot_training - Step 14861: {'lr': 0.00041653957994907255, 'samples': 7609344, 'steps': 14861, 'loss/train': 1.5617280006408691} 02/24/2022 20:48:35 - INFO - codeparrot_training - Step 14862: {'lr': 0.0004165273762989023, 'samples': 7609856, 'steps': 14862, 'loss/train': 1.6858184337615967} 02/24/2022 20:48:40 - INFO - codeparrot_training - Step 14863: {'lr': 0.0004165151719353817, 'samples': 7610368, 'steps': 14863, 'loss/train': 1.5079296827316284} 02/24/2022 20:48:44 - INFO - codeparrot_training - Step 14864: {'lr': 0.0004165029668585629, 'samples': 7610880, 'steps': 14864, 'loss/train': 2.8091301918029785} 02/24/2022 20:48:49 - INFO - codeparrot_training - Step 14865: {'lr': 0.00041649076106849836, 'samples': 7611392, 'steps': 14865, 'loss/train': 1.4334156513214111} 02/24/2022 20:48:55 - INFO - codeparrot_training - Step 14866: {'lr': 0.0004164785545652402, 'samples': 7611904, 'steps': 14866, 'loss/train': 2.2660720348358154} 02/24/2022 20:48:58 - INFO - codeparrot_training - Step 14867: {'lr': 0.0004164663473488408, 'samples': 7612416, 'steps': 14867, 'loss/train': 2.3031933307647705} 02/24/2022 20:49:04 - INFO - codeparrot_training - Step 14868: {'lr': 0.0004164541394193524, 'samples': 7612928, 'steps': 14868, 'loss/train': 1.8336914777755737} 02/24/2022 20:49:08 - INFO - codeparrot_training - Step 14869: {'lr': 0.00041644193077682734, 'samples': 7613440, 'steps': 14869, 'loss/train': 0.2681225836277008} 02/24/2022 20:49:13 - INFO - codeparrot_training - Step 14870: {'lr': 0.0004164297214213179, 'samples': 7613952, 'steps': 14870, 'loss/train': 1.74448823928833} 02/24/2022 20:49:17 - INFO - codeparrot_training - Step 14871: {'lr': 0.0004164175113528763, 'samples': 7614464, 'steps': 14871, 'loss/train': 1.7467695474624634} 02/24/2022 20:49:22 - INFO - codeparrot_training - Step 14872: {'lr': 0.000416405300571555, 'samples': 7614976, 'steps': 14872, 'loss/train': 2.1374170780181885} 02/24/2022 20:49:26 - INFO - codeparrot_training - Step 14873: {'lr': 0.00041639308907740624, 'samples': 7615488, 'steps': 14873, 'loss/train': 2.4194884300231934} 02/24/2022 20:49:31 - INFO - codeparrot_training - Step 14874: {'lr': 0.0004163808768704823, 'samples': 7616000, 'steps': 14874, 'loss/train': 2.1897895336151123} 02/24/2022 20:49:35 - INFO - codeparrot_training - Step 14875: {'lr': 0.0004163686639508356, 'samples': 7616512, 'steps': 14875, 'loss/train': 1.8933578729629517} 02/24/2022 20:49:40 - INFO - codeparrot_training - Step 14876: {'lr': 0.00041635645031851826, 'samples': 7617024, 'steps': 14876, 'loss/train': 1.790216326713562} 02/24/2022 20:49:44 - INFO - codeparrot_training - Step 14877: {'lr': 0.0004163442359735827, 'samples': 7617536, 'steps': 14877, 'loss/train': 1.200398564338684} 02/24/2022 20:49:50 - INFO - codeparrot_training - Step 14878: {'lr': 0.00041633202091608136, 'samples': 7618048, 'steps': 14878, 'loss/train': 2.0175182819366455} 02/24/2022 20:49:53 - INFO - codeparrot_training - Step 14879: {'lr': 0.00041631980514606636, 'samples': 7618560, 'steps': 14879, 'loss/train': 0.13233886659145355} 02/24/2022 20:49:58 - INFO - codeparrot_training - Step 14880: {'lr': 0.0004163075886635902, 'samples': 7619072, 'steps': 14880, 'loss/train': 2.0524795055389404} 02/24/2022 20:50:02 - INFO - codeparrot_training - Step 14881: {'lr': 0.0004162953714687051, 'samples': 7619584, 'steps': 14881, 'loss/train': 1.709363341331482} 02/24/2022 20:50:08 - INFO - codeparrot_training - Step 14882: {'lr': 0.0004162831535614635, 'samples': 7620096, 'steps': 14882, 'loss/train': 3.060720205307007} 02/24/2022 20:50:12 - INFO - codeparrot_training - Step 14883: {'lr': 0.0004162709349419176, 'samples': 7620608, 'steps': 14883, 'loss/train': 1.8922135829925537} 02/24/2022 20:50:17 - INFO - codeparrot_training - Step 14884: {'lr': 0.0004162587156101198, 'samples': 7621120, 'steps': 14884, 'loss/train': 2.6358642578125} 02/24/2022 20:50:21 - INFO - codeparrot_training - Step 14885: {'lr': 0.0004162464955661225, 'samples': 7621632, 'steps': 14885, 'loss/train': 1.724726915359497} 02/24/2022 20:50:24 - INFO - codeparrot_training - Step 14886: {'lr': 0.000416234274809978, 'samples': 7622144, 'steps': 14886, 'loss/train': 0.09583525359630585} 02/24/2022 20:50:30 - INFO - codeparrot_training - Step 14887: {'lr': 0.00041622205334173863, 'samples': 7622656, 'steps': 14887, 'loss/train': 2.0287089347839355} 02/24/2022 20:50:33 - INFO - codeparrot_training - Step 14888: {'lr': 0.00041620983116145673, 'samples': 7623168, 'steps': 14888, 'loss/train': 1.413889765739441} 02/24/2022 20:50:39 - INFO - codeparrot_training - Step 14889: {'lr': 0.00041619760826918474, 'samples': 7623680, 'steps': 14889, 'loss/train': 0.8066677451133728} 02/24/2022 20:50:42 - INFO - codeparrot_training - Step 14890: {'lr': 0.00041618538466497496, 'samples': 7624192, 'steps': 14890, 'loss/train': 1.8946795463562012} 02/24/2022 20:50:48 - INFO - codeparrot_training - Step 14891: {'lr': 0.00041617316034887983, 'samples': 7624704, 'steps': 14891, 'loss/train': 0.7714298367500305} 02/24/2022 20:50:51 - INFO - codeparrot_training - Step 14892: {'lr': 0.00041616093532095155, 'samples': 7625216, 'steps': 14892, 'loss/train': 2.578319787979126} 02/24/2022 20:50:58 - INFO - codeparrot_training - Step 14893: {'lr': 0.00041614870958124264, 'samples': 7625728, 'steps': 14893, 'loss/train': 2.4218244552612305} 02/24/2022 20:51:01 - INFO - codeparrot_training - Step 14894: {'lr': 0.00041613648312980537, 'samples': 7626240, 'steps': 14894, 'loss/train': 0.8149571418762207} 02/24/2022 20:51:07 - INFO - codeparrot_training - Step 14895: {'lr': 0.00041612425596669215, 'samples': 7626752, 'steps': 14895, 'loss/train': 2.4576964378356934} 02/24/2022 20:51:12 - INFO - codeparrot_training - Step 14896: {'lr': 0.0004161120280919554, 'samples': 7627264, 'steps': 14896, 'loss/train': 2.0430126190185547} 02/24/2022 20:51:16 - INFO - codeparrot_training - Step 14897: {'lr': 0.00041609979950564747, 'samples': 7627776, 'steps': 14897, 'loss/train': 1.5276203155517578} 02/24/2022 20:51:21 - INFO - codeparrot_training - Step 14898: {'lr': 0.00041608757020782073, 'samples': 7628288, 'steps': 14898, 'loss/train': 2.11814546585083} 02/24/2022 20:51:25 - INFO - codeparrot_training - Step 14899: {'lr': 0.0004160753401985276, 'samples': 7628800, 'steps': 14899, 'loss/train': 1.3218504190444946} 02/24/2022 20:51:30 - INFO - codeparrot_training - Step 14900: {'lr': 0.00041606310947782046, 'samples': 7629312, 'steps': 14900, 'loss/train': 1.7159701585769653} 02/24/2022 20:51:34 - INFO - codeparrot_training - Step 14901: {'lr': 0.00041605087804575167, 'samples': 7629824, 'steps': 14901, 'loss/train': 2.6790263652801514} 02/24/2022 20:51:39 - INFO - codeparrot_training - Step 14902: {'lr': 0.0004160386459023736, 'samples': 7630336, 'steps': 14902, 'loss/train': 2.3310153484344482} 02/24/2022 20:51:43 - INFO - codeparrot_training - Step 14903: {'lr': 0.00041602641304773876, 'samples': 7630848, 'steps': 14903, 'loss/train': 1.4768877029418945} 02/24/2022 20:51:49 - INFO - codeparrot_training - Step 14904: {'lr': 0.0004160141794818995, 'samples': 7631360, 'steps': 14904, 'loss/train': 1.1713708639144897} 02/24/2022 20:51:53 - INFO - codeparrot_training - Step 14905: {'lr': 0.00041600194520490815, 'samples': 7631872, 'steps': 14905, 'loss/train': 2.23829984664917} 02/24/2022 20:51:58 - INFO - codeparrot_training - Step 14906: {'lr': 0.0004159897102168172, 'samples': 7632384, 'steps': 14906, 'loss/train': 3.2544000148773193} 02/24/2022 20:52:02 - INFO - codeparrot_training - Step 14907: {'lr': 0.00041597747451767905, 'samples': 7632896, 'steps': 14907, 'loss/train': 2.157799243927002} 02/24/2022 20:52:07 - INFO - codeparrot_training - Step 14908: {'lr': 0.00041596523810754607, 'samples': 7633408, 'steps': 14908, 'loss/train': 2.5491750240325928} 02/24/2022 20:52:11 - INFO - codeparrot_training - Step 14909: {'lr': 0.0004159530009864707, 'samples': 7633920, 'steps': 14909, 'loss/train': 0.29711586236953735} 02/24/2022 20:52:16 - INFO - codeparrot_training - Step 14910: {'lr': 0.0004159407631545054, 'samples': 7634432, 'steps': 14910, 'loss/train': 2.1845693588256836} 02/24/2022 20:52:20 - INFO - codeparrot_training - Step 14911: {'lr': 0.0004159285246117026, 'samples': 7634944, 'steps': 14911, 'loss/train': 2.1220433712005615} 02/24/2022 20:52:25 - INFO - codeparrot_training - Step 14912: {'lr': 0.00041591628535811464, 'samples': 7635456, 'steps': 14912, 'loss/train': 1.818053126335144} 02/24/2022 20:52:29 - INFO - codeparrot_training - Step 14913: {'lr': 0.000415904045393794, 'samples': 7635968, 'steps': 14913, 'loss/train': 1.3587634563446045} 02/24/2022 20:52:35 - INFO - codeparrot_training - Step 14914: {'lr': 0.0004158918047187931, 'samples': 7636480, 'steps': 14914, 'loss/train': 2.447115182876587} 02/24/2022 20:52:38 - INFO - codeparrot_training - Step 14915: {'lr': 0.0004158795633331645, 'samples': 7636992, 'steps': 14915, 'loss/train': 2.1404457092285156} 02/24/2022 20:52:44 - INFO - codeparrot_training - Step 14916: {'lr': 0.00041586732123696037, 'samples': 7637504, 'steps': 14916, 'loss/train': 1.956944465637207} 02/24/2022 20:52:47 - INFO - codeparrot_training - Step 14917: {'lr': 0.0004158550784302334, 'samples': 7638016, 'steps': 14917, 'loss/train': 2.7461376190185547} 02/24/2022 20:52:53 - INFO - codeparrot_training - Step 14918: {'lr': 0.0004158428349130359, 'samples': 7638528, 'steps': 14918, 'loss/train': 2.5754482746124268} 02/24/2022 20:52:56 - INFO - codeparrot_training - Step 14919: {'lr': 0.00041583059068542034, 'samples': 7639040, 'steps': 14919, 'loss/train': 1.840137004852295} 02/24/2022 20:53:02 - INFO - codeparrot_training - Step 14920: {'lr': 0.0004158183457474392, 'samples': 7639552, 'steps': 14920, 'loss/train': 0.7040673494338989} 02/24/2022 20:53:05 - INFO - codeparrot_training - Step 14921: {'lr': 0.00041580610009914486, 'samples': 7640064, 'steps': 14921, 'loss/train': 1.3049136400222778} 02/24/2022 20:53:11 - INFO - codeparrot_training - Step 14922: {'lr': 0.00041579385374058996, 'samples': 7640576, 'steps': 14922, 'loss/train': 2.4330427646636963} 02/24/2022 20:53:14 - INFO - codeparrot_training - Step 14923: {'lr': 0.00041578160667182676, 'samples': 7641088, 'steps': 14923, 'loss/train': 1.8912090063095093} 02/24/2022 20:53:20 - INFO - codeparrot_training - Step 14924: {'lr': 0.00041576935889290777, 'samples': 7641600, 'steps': 14924, 'loss/train': 1.8363546133041382} 02/24/2022 20:53:24 - INFO - codeparrot_training - Step 14925: {'lr': 0.0004157571104038856, 'samples': 7642112, 'steps': 14925, 'loss/train': 1.9110095500946045} 02/24/2022 20:53:27 - INFO - codeparrot_training - Step 14926: {'lr': 0.00041574486120481255, 'samples': 7642624, 'steps': 14926, 'loss/train': 1.2370458841323853} 02/24/2022 20:53:33 - INFO - codeparrot_training - Step 14927: {'lr': 0.0004157326112957411, 'samples': 7643136, 'steps': 14927, 'loss/train': 2.8494725227355957} 02/24/2022 20:53:36 - INFO - codeparrot_training - Step 14928: {'lr': 0.0004157203606767238, 'samples': 7643648, 'steps': 14928, 'loss/train': 1.1753263473510742} 02/24/2022 20:53:42 - INFO - codeparrot_training - Step 14929: {'lr': 0.0004157081093478131, 'samples': 7644160, 'steps': 14929, 'loss/train': 1.904313564300537} 02/24/2022 20:53:45 - INFO - codeparrot_training - Step 14930: {'lr': 0.00041569585730906147, 'samples': 7644672, 'steps': 14930, 'loss/train': 1.8031725883483887} 02/24/2022 20:53:51 - INFO - codeparrot_training - Step 14931: {'lr': 0.0004156836045605214, 'samples': 7645184, 'steps': 14931, 'loss/train': 2.3395514488220215} 02/24/2022 20:53:55 - INFO - codeparrot_training - Step 14932: {'lr': 0.0004156713511022454, 'samples': 7645696, 'steps': 14932, 'loss/train': 2.0929887294769287} 02/24/2022 20:54:00 - INFO - codeparrot_training - Step 14933: {'lr': 0.00041565909693428593, 'samples': 7646208, 'steps': 14933, 'loss/train': 1.9599897861480713} 02/24/2022 20:54:04 - INFO - codeparrot_training - Step 14934: {'lr': 0.00041564684205669546, 'samples': 7646720, 'steps': 14934, 'loss/train': 1.6113312244415283} 02/24/2022 20:54:11 - INFO - codeparrot_training - Step 14935: {'lr': 0.00041563458646952655, 'samples': 7647232, 'steps': 14935, 'loss/train': 1.6996887922286987} 02/24/2022 20:54:14 - INFO - codeparrot_training - Step 14936: {'lr': 0.0004156223301728316, 'samples': 7647744, 'steps': 14936, 'loss/train': 2.5070126056671143} 02/24/2022 20:54:20 - INFO - codeparrot_training - Step 14937: {'lr': 0.00041561007316666333, 'samples': 7648256, 'steps': 14937, 'loss/train': 2.176846504211426} 02/24/2022 20:54:23 - INFO - codeparrot_training - Step 14938: {'lr': 0.00041559781545107393, 'samples': 7648768, 'steps': 14938, 'loss/train': 2.6045141220092773} 02/24/2022 20:54:29 - INFO - codeparrot_training - Step 14939: {'lr': 0.00041558555702611615, 'samples': 7649280, 'steps': 14939, 'loss/train': 2.515040874481201} 02/24/2022 20:54:34 - INFO - codeparrot_training - Step 14940: {'lr': 0.0004155732978918424, 'samples': 7649792, 'steps': 14940, 'loss/train': 1.1698956489562988} 02/24/2022 20:54:38 - INFO - codeparrot_training - Step 14941: {'lr': 0.00041556103804830523, 'samples': 7650304, 'steps': 14941, 'loss/train': 1.5823616981506348} 02/24/2022 20:54:43 - INFO - codeparrot_training - Step 14942: {'lr': 0.0004155487774955572, 'samples': 7650816, 'steps': 14942, 'loss/train': 2.547513484954834} 02/24/2022 20:54:47 - INFO - codeparrot_training - Step 14943: {'lr': 0.00041553651623365076, 'samples': 7651328, 'steps': 14943, 'loss/train': 2.0599911212921143} 02/24/2022 20:54:52 - INFO - codeparrot_training - Step 14944: {'lr': 0.00041552425426263836, 'samples': 7651840, 'steps': 14944, 'loss/train': 2.4216253757476807} 02/24/2022 20:54:56 - INFO - codeparrot_training - Step 14945: {'lr': 0.00041551199158257264, 'samples': 7652352, 'steps': 14945, 'loss/train': 1.619296669960022} 02/24/2022 20:55:02 - INFO - codeparrot_training - Step 14946: {'lr': 0.00041549972819350615, 'samples': 7652864, 'steps': 14946, 'loss/train': 2.132974863052368} 02/24/2022 20:55:06 - INFO - codeparrot_training - Step 14947: {'lr': 0.00041548746409549134, 'samples': 7653376, 'steps': 14947, 'loss/train': 1.4858053922653198} 02/24/2022 20:55:11 - INFO - codeparrot_training - Step 14948: {'lr': 0.0004154751992885808, 'samples': 7653888, 'steps': 14948, 'loss/train': 1.736487627029419} 02/24/2022 20:55:15 - INFO - codeparrot_training - Step 14949: {'lr': 0.0004154629337728271, 'samples': 7654400, 'steps': 14949, 'loss/train': 2.457109212875366} 02/24/2022 20:55:20 - INFO - codeparrot_training - Step 14950: {'lr': 0.00041545066754828264, 'samples': 7654912, 'steps': 14950, 'loss/train': 1.6297321319580078} 02/24/2022 20:55:24 - INFO - codeparrot_training - Step 14951: {'lr': 0.00041543840061500007, 'samples': 7655424, 'steps': 14951, 'loss/train': 2.04052996635437} 02/24/2022 20:55:29 - INFO - codeparrot_training - Step 14952: {'lr': 0.000415426132973032, 'samples': 7655936, 'steps': 14952, 'loss/train': 2.7318904399871826} 02/24/2022 20:55:33 - INFO - codeparrot_training - Step 14953: {'lr': 0.0004154138646224308, 'samples': 7656448, 'steps': 14953, 'loss/train': 2.008251428604126} 02/24/2022 20:55:38 - INFO - codeparrot_training - Step 14954: {'lr': 0.0004154015955632492, 'samples': 7656960, 'steps': 14954, 'loss/train': 2.7928543090820312} 02/24/2022 20:55:42 - INFO - codeparrot_training - Step 14955: {'lr': 0.0004153893257955397, 'samples': 7657472, 'steps': 14955, 'loss/train': 1.9538930654525757} 02/24/2022 20:55:48 - INFO - codeparrot_training - Step 14956: {'lr': 0.00041537705531935476, 'samples': 7657984, 'steps': 14956, 'loss/train': 2.53762149810791} 02/24/2022 20:55:51 - INFO - codeparrot_training - Step 14957: {'lr': 0.0004153647841347471, 'samples': 7658496, 'steps': 14957, 'loss/train': 2.2134952545166016} 02/24/2022 20:55:57 - INFO - codeparrot_training - Step 14958: {'lr': 0.0004153525122417692, 'samples': 7659008, 'steps': 14958, 'loss/train': 2.2954773902893066} 02/24/2022 20:56:00 - INFO - codeparrot_training - Step 14959: {'lr': 0.00041534023964047363, 'samples': 7659520, 'steps': 14959, 'loss/train': 2.664625883102417} 02/24/2022 20:56:06 - INFO - codeparrot_training - Step 14960: {'lr': 0.00041532796633091297, 'samples': 7660032, 'steps': 14960, 'loss/train': 2.5796895027160645} 02/24/2022 20:56:09 - INFO - codeparrot_training - Step 14961: {'lr': 0.0004153156923131398, 'samples': 7660544, 'steps': 14961, 'loss/train': 2.0717506408691406} 02/24/2022 20:56:15 - INFO - codeparrot_training - Step 14962: {'lr': 0.0004153034175872067, 'samples': 7661056, 'steps': 14962, 'loss/train': 2.521807909011841} 02/24/2022 20:56:18 - INFO - codeparrot_training - Step 14963: {'lr': 0.00041529114215316633, 'samples': 7661568, 'steps': 14963, 'loss/train': 1.5495537519454956} 02/24/2022 20:56:24 - INFO - codeparrot_training - Step 14964: {'lr': 0.0004152788660110711, 'samples': 7662080, 'steps': 14964, 'loss/train': 1.9410985708236694} 02/24/2022 20:56:27 - INFO - codeparrot_training - Step 14965: {'lr': 0.0004152665891609737, 'samples': 7662592, 'steps': 14965, 'loss/train': 0.3258343040943146} 02/24/2022 20:56:33 - INFO - codeparrot_training - Step 14966: {'lr': 0.0004152543116029267, 'samples': 7663104, 'steps': 14966, 'loss/train': 2.791116952896118} 02/24/2022 20:56:37 - INFO - codeparrot_training - Step 14967: {'lr': 0.0004152420333369827, 'samples': 7663616, 'steps': 14967, 'loss/train': 1.664000153541565} 02/24/2022 20:56:42 - INFO - codeparrot_training - Step 14968: {'lr': 0.00041522975436319445, 'samples': 7664128, 'steps': 14968, 'loss/train': 2.229444980621338} 02/24/2022 20:56:46 - INFO - codeparrot_training - Step 14969: {'lr': 0.00041521747468161417, 'samples': 7664640, 'steps': 14969, 'loss/train': 1.7846778631210327} 02/24/2022 20:56:51 - INFO - codeparrot_training - Step 14970: {'lr': 0.00041520519429229485, 'samples': 7665152, 'steps': 14970, 'loss/train': 0.5800321102142334} 02/24/2022 20:56:55 - INFO - codeparrot_training - Step 14971: {'lr': 0.00041519291319528886, 'samples': 7665664, 'steps': 14971, 'loss/train': 1.9767426252365112} 02/24/2022 20:57:00 - INFO - codeparrot_training - Step 14972: {'lr': 0.00041518063139064893, 'samples': 7666176, 'steps': 14972, 'loss/train': 1.750622272491455} 02/24/2022 20:57:04 - INFO - codeparrot_training - Step 14973: {'lr': 0.0004151683488784276, 'samples': 7666688, 'steps': 14973, 'loss/train': 1.5277704000473022} 02/24/2022 20:57:09 - INFO - codeparrot_training - Step 14974: {'lr': 0.00041515606565867746, 'samples': 7667200, 'steps': 14974, 'loss/train': 1.2477033138275146} 02/24/2022 20:57:13 - INFO - codeparrot_training - Step 14975: {'lr': 0.0004151437817314513, 'samples': 7667712, 'steps': 14975, 'loss/train': 2.1250593662261963} 02/24/2022 20:57:19 - INFO - codeparrot_training - Step 14976: {'lr': 0.00041513149709680155, 'samples': 7668224, 'steps': 14976, 'loss/train': 1.9217205047607422} 02/24/2022 20:57:22 - INFO - codeparrot_training - Step 14977: {'lr': 0.00041511921175478085, 'samples': 7668736, 'steps': 14977, 'loss/train': 1.381117582321167} 02/24/2022 20:57:28 - INFO - codeparrot_training - Step 14978: {'lr': 0.0004151069257054419, 'samples': 7669248, 'steps': 14978, 'loss/train': 2.457695960998535} 02/24/2022 20:57:31 - INFO - codeparrot_training - Step 14979: {'lr': 0.0004150946389488374, 'samples': 7669760, 'steps': 14979, 'loss/train': 0.7559409141540527} 02/24/2022 20:57:37 - INFO - codeparrot_training - Step 14980: {'lr': 0.0004150823514850198, 'samples': 7670272, 'steps': 14980, 'loss/train': 2.514883518218994} 02/24/2022 20:57:41 - INFO - codeparrot_training - Step 14981: {'lr': 0.00041507006331404186, 'samples': 7670784, 'steps': 14981, 'loss/train': 2.1716983318328857} 02/24/2022 20:57:46 - INFO - codeparrot_training - Step 14982: {'lr': 0.00041505777443595615, 'samples': 7671296, 'steps': 14982, 'loss/train': 1.7020424604415894} 02/24/2022 20:57:50 - INFO - codeparrot_training - Step 14983: {'lr': 0.0004150454848508154, 'samples': 7671808, 'steps': 14983, 'loss/train': 2.3099896907806396} 02/24/2022 20:57:55 - INFO - codeparrot_training - Step 14984: {'lr': 0.00041503319455867216, 'samples': 7672320, 'steps': 14984, 'loss/train': 1.9879225492477417} 02/24/2022 20:57:59 - INFO - codeparrot_training - Step 14985: {'lr': 0.0004150209035595791, 'samples': 7672832, 'steps': 14985, 'loss/train': 2.8142597675323486} 02/24/2022 20:58:05 - INFO - codeparrot_training - Step 14986: {'lr': 0.000415008611853589, 'samples': 7673344, 'steps': 14986, 'loss/train': 2.348301649093628} 02/24/2022 20:58:08 - INFO - codeparrot_training - Step 14987: {'lr': 0.0004149963194407543, 'samples': 7673856, 'steps': 14987, 'loss/train': 1.9153321981430054} 02/24/2022 20:58:14 - INFO - codeparrot_training - Step 14988: {'lr': 0.00041498402632112776, 'samples': 7674368, 'steps': 14988, 'loss/train': 1.4021347761154175} 02/24/2022 20:58:17 - INFO - codeparrot_training - Step 14989: {'lr': 0.00041497173249476204, 'samples': 7674880, 'steps': 14989, 'loss/train': 1.3956748247146606} 02/24/2022 20:58:23 - INFO - codeparrot_training - Step 14990: {'lr': 0.0004149594379617099, 'samples': 7675392, 'steps': 14990, 'loss/train': 2.416032075881958} 02/24/2022 20:58:26 - INFO - codeparrot_training - Step 14991: {'lr': 0.00041494714272202385, 'samples': 7675904, 'steps': 14991, 'loss/train': 1.4470218420028687} 02/24/2022 20:58:33 - INFO - codeparrot_training - Step 14992: {'lr': 0.00041493484677575655, 'samples': 7676416, 'steps': 14992, 'loss/train': 3.7259156703948975} 02/24/2022 20:58:36 - INFO - codeparrot_training - Step 14993: {'lr': 0.00041492255012296077, 'samples': 7676928, 'steps': 14993, 'loss/train': 1.637926459312439} 02/24/2022 20:58:42 - INFO - codeparrot_training - Step 14994: {'lr': 0.0004149102527636892, 'samples': 7677440, 'steps': 14994, 'loss/train': 2.3843772411346436} 02/24/2022 20:58:45 - INFO - codeparrot_training - Step 14995: {'lr': 0.0004148979546979944, 'samples': 7677952, 'steps': 14995, 'loss/train': 1.8539739847183228} 02/24/2022 20:58:51 - INFO - codeparrot_training - Step 14996: {'lr': 0.00041488565592592917, 'samples': 7678464, 'steps': 14996, 'loss/train': 1.4838346242904663} 02/24/2022 20:58:54 - INFO - codeparrot_training - Step 14997: {'lr': 0.0004148733564475462, 'samples': 7678976, 'steps': 14997, 'loss/train': 2.3169679641723633} 02/24/2022 20:59:00 - INFO - codeparrot_training - Step 14998: {'lr': 0.000414861056262898, 'samples': 7679488, 'steps': 14998, 'loss/train': 1.7019398212432861} 02/24/2022 20:59:03 - INFO - codeparrot_training - Step 14999: {'lr': 0.0004148487553720375, 'samples': 7680000, 'steps': 14999, 'loss/train': 2.059992551803589} 02/24/2022 20:59:03 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 20:59:20 - WARNING - huggingface_hub.repository - Several commits (15) will be pushed upstream. 02/24/2022 20:59:20 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 20:59:54 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 3a61b16..751fbfb floral-grass-11 -> floral-grass-11 02/24/2022 21:00:01 - INFO - codeparrot_training - Step 15000: {'lr': 0.0004148364537750172, 'samples': 7680512, 'steps': 15000, 'loss/train': 1.8882983922958374} 02/24/2022 21:00:04 - INFO - codeparrot_training - Step 15001: {'lr': 0.0004148241514718899, 'samples': 7681024, 'steps': 15001, 'loss/train': 3.1234290599823} 02/24/2022 21:00:11 - INFO - codeparrot_training - Step 15002: {'lr': 0.00041481184846270836, 'samples': 7681536, 'steps': 15002, 'loss/train': 2.020622968673706} 02/24/2022 21:00:14 - INFO - codeparrot_training - Step 15003: {'lr': 0.00041479954474752507, 'samples': 7682048, 'steps': 15003, 'loss/train': 2.7225804328918457} 02/24/2022 21:00:20 - INFO - codeparrot_training - Step 15004: {'lr': 0.0004147872403263929, 'samples': 7682560, 'steps': 15004, 'loss/train': 1.2915410995483398} 02/24/2022 21:00:23 - INFO - codeparrot_training - Step 15005: {'lr': 0.0004147749351993645, 'samples': 7683072, 'steps': 15005, 'loss/train': 3.145185947418213} 02/24/2022 21:00:29 - INFO - codeparrot_training - Step 15006: {'lr': 0.0004147626293664926, 'samples': 7683584, 'steps': 15006, 'loss/train': 2.1505861282348633} 02/24/2022 21:00:32 - INFO - codeparrot_training - Step 15007: {'lr': 0.00041475032282783, 'samples': 7684096, 'steps': 15007, 'loss/train': 1.9363479614257812} 02/24/2022 21:00:38 - INFO - codeparrot_training - Step 15008: {'lr': 0.0004147380155834293, 'samples': 7684608, 'steps': 15008, 'loss/train': 0.9024702906608582} 02/24/2022 21:00:41 - INFO - codeparrot_training - Step 15009: {'lr': 0.00041472570763334316, 'samples': 7685120, 'steps': 15009, 'loss/train': 2.2938032150268555} 02/24/2022 21:00:47 - INFO - codeparrot_training - Step 15010: {'lr': 0.00041471339897762447, 'samples': 7685632, 'steps': 15010, 'loss/train': 2.2601678371429443} 02/24/2022 21:00:50 - INFO - codeparrot_training - Step 15011: {'lr': 0.0004147010896163259, 'samples': 7686144, 'steps': 15011, 'loss/train': 2.4319632053375244} 02/24/2022 21:00:56 - INFO - codeparrot_training - Step 15012: {'lr': 0.00041468877954950006, 'samples': 7686656, 'steps': 15012, 'loss/train': 1.8616952896118164} 02/24/2022 21:00:59 - INFO - codeparrot_training - Step 15013: {'lr': 0.0004146764687771999, 'samples': 7687168, 'steps': 15013, 'loss/train': 2.0138206481933594} 02/24/2022 21:01:06 - INFO - codeparrot_training - Step 15014: {'lr': 0.00041466415729947794, 'samples': 7687680, 'steps': 15014, 'loss/train': 1.6361883878707886} 02/24/2022 21:01:09 - INFO - codeparrot_training - Step 15015: {'lr': 0.0004146518451163871, 'samples': 7688192, 'steps': 15015, 'loss/train': 1.0572295188903809} 02/24/2022 21:01:15 - INFO - codeparrot_training - Step 15016: {'lr': 0.00041463953222798, 'samples': 7688704, 'steps': 15016, 'loss/train': 3.8093860149383545} 02/24/2022 21:01:18 - INFO - codeparrot_training - Step 15017: {'lr': 0.00041462721863430943, 'samples': 7689216, 'steps': 15017, 'loss/train': 1.4273110628128052} 02/24/2022 21:01:24 - INFO - codeparrot_training - Step 15018: {'lr': 0.0004146149043354281, 'samples': 7689728, 'steps': 15018, 'loss/train': 1.517775535583496} 02/24/2022 21:01:27 - INFO - codeparrot_training - Step 15019: {'lr': 0.0004146025893313888, 'samples': 7690240, 'steps': 15019, 'loss/train': 1.854677438735962} 02/24/2022 21:01:33 - INFO - codeparrot_training - Step 15020: {'lr': 0.00041459027362224433, 'samples': 7690752, 'steps': 15020, 'loss/train': 2.0060839653015137} 02/24/2022 21:01:36 - INFO - codeparrot_training - Step 15021: {'lr': 0.0004145779572080473, 'samples': 7691264, 'steps': 15021, 'loss/train': 2.086730718612671} 02/24/2022 21:01:42 - INFO - codeparrot_training - Step 15022: {'lr': 0.0004145656400888506, 'samples': 7691776, 'steps': 15022, 'loss/train': 2.4009227752685547} 02/24/2022 21:01:45 - INFO - codeparrot_training - Step 15023: {'lr': 0.000414553322264707, 'samples': 7692288, 'steps': 15023, 'loss/train': 1.1341274976730347} 02/24/2022 21:01:52 - INFO - codeparrot_training - Step 15024: {'lr': 0.00041454100373566915, 'samples': 7692800, 'steps': 15024, 'loss/train': 1.7949512004852295} 02/24/2022 21:01:55 - INFO - codeparrot_training - Step 15025: {'lr': 0.00041452868450178994, 'samples': 7693312, 'steps': 15025, 'loss/train': 1.8109724521636963} 02/24/2022 21:02:01 - INFO - codeparrot_training - Step 15026: {'lr': 0.00041451636456312207, 'samples': 7693824, 'steps': 15026, 'loss/train': 0.7747088670730591} 02/24/2022 21:02:04 - INFO - codeparrot_training - Step 15027: {'lr': 0.0004145040439197183, 'samples': 7694336, 'steps': 15027, 'loss/train': 2.311444044113159} 02/24/2022 21:02:10 - INFO - codeparrot_training - Step 15028: {'lr': 0.00041449172257163156, 'samples': 7694848, 'steps': 15028, 'loss/train': 2.1977365016937256} 02/24/2022 21:02:13 - INFO - codeparrot_training - Step 15029: {'lr': 0.00041447940051891435, 'samples': 7695360, 'steps': 15029, 'loss/train': 2.079603433609009} 02/24/2022 21:02:19 - INFO - codeparrot_training - Step 15030: {'lr': 0.00041446707776161975, 'samples': 7695872, 'steps': 15030, 'loss/train': 1.5588449239730835} 02/24/2022 21:02:24 - INFO - codeparrot_training - Step 15031: {'lr': 0.00041445475429980033, 'samples': 7696384, 'steps': 15031, 'loss/train': 0.8621600270271301} 02/24/2022 21:02:28 - INFO - codeparrot_training - Step 15032: {'lr': 0.000414442430133509, 'samples': 7696896, 'steps': 15032, 'loss/train': 3.114267110824585} 02/24/2022 21:02:33 - INFO - codeparrot_training - Step 15033: {'lr': 0.0004144301052627985, 'samples': 7697408, 'steps': 15033, 'loss/train': 1.6453732252120972} 02/24/2022 21:02:37 - INFO - codeparrot_training - Step 15034: {'lr': 0.00041441777968772165, 'samples': 7697920, 'steps': 15034, 'loss/train': 2.32963228225708} 02/24/2022 21:02:42 - INFO - codeparrot_training - Step 15035: {'lr': 0.00041440545340833124, 'samples': 7698432, 'steps': 15035, 'loss/train': 2.181037425994873} 02/24/2022 21:02:46 - INFO - codeparrot_training - Step 15036: {'lr': 0.00041439312642468007, 'samples': 7698944, 'steps': 15036, 'loss/train': 2.291114091873169} 02/24/2022 21:02:51 - INFO - codeparrot_training - Step 15037: {'lr': 0.000414380798736821, 'samples': 7699456, 'steps': 15037, 'loss/train': 2.567660093307495} 02/24/2022 21:02:55 - INFO - codeparrot_training - Step 15038: {'lr': 0.0004143684703448067, 'samples': 7699968, 'steps': 15038, 'loss/train': 2.418159246444702} 02/24/2022 21:03:01 - INFO - codeparrot_training - Step 15039: {'lr': 0.0004143561412486901, 'samples': 7700480, 'steps': 15039, 'loss/train': 1.9684832096099854} 02/24/2022 21:03:04 - INFO - codeparrot_training - Step 15040: {'lr': 0.00041434381144852395, 'samples': 7700992, 'steps': 15040, 'loss/train': 1.1164835691452026} 02/24/2022 21:03:10 - INFO - codeparrot_training - Step 15041: {'lr': 0.00041433148094436115, 'samples': 7701504, 'steps': 15041, 'loss/train': 2.4944045543670654} 02/24/2022 21:03:13 - INFO - codeparrot_training - Step 15042: {'lr': 0.0004143191497362544, 'samples': 7702016, 'steps': 15042, 'loss/train': 1.8259201049804688} 02/24/2022 21:03:19 - INFO - codeparrot_training - Step 15043: {'lr': 0.0004143068178242566, 'samples': 7702528, 'steps': 15043, 'loss/train': 5.458270072937012} 02/24/2022 21:03:23 - INFO - codeparrot_training - Step 15044: {'lr': 0.00041429448520842064, 'samples': 7703040, 'steps': 15044, 'loss/train': 1.0957911014556885} 02/24/2022 21:03:28 - INFO - codeparrot_training - Step 15045: {'lr': 0.00041428215188879926, 'samples': 7703552, 'steps': 15045, 'loss/train': 1.8701893091201782} 02/24/2022 21:03:32 - INFO - codeparrot_training - Step 15046: {'lr': 0.0004142698178654453, 'samples': 7704064, 'steps': 15046, 'loss/train': 1.3205199241638184} 02/24/2022 21:03:37 - INFO - codeparrot_training - Step 15047: {'lr': 0.0004142574831384115, 'samples': 7704576, 'steps': 15047, 'loss/train': 1.8461235761642456} 02/24/2022 21:03:41 - INFO - codeparrot_training - Step 15048: {'lr': 0.0004142451477077509, 'samples': 7705088, 'steps': 15048, 'loss/train': 2.698197841644287} 02/24/2022 21:03:47 - INFO - codeparrot_training - Step 15049: {'lr': 0.00041423281157351624, 'samples': 7705600, 'steps': 15049, 'loss/train': 2.1005477905273438} 02/24/2022 21:03:50 - INFO - codeparrot_training - Step 15050: {'lr': 0.00041422047473576033, 'samples': 7706112, 'steps': 15050, 'loss/train': 1.5033564567565918} 02/24/2022 21:03:56 - INFO - codeparrot_training - Step 15051: {'lr': 0.0004142081371945361, 'samples': 7706624, 'steps': 15051, 'loss/train': 2.7540948390960693} 02/24/2022 21:03:59 - INFO - codeparrot_training - Step 15052: {'lr': 0.00041419579894989633, 'samples': 7707136, 'steps': 15052, 'loss/train': 2.084900140762329} 02/24/2022 21:04:05 - INFO - codeparrot_training - Step 15053: {'lr': 0.0004141834600018939, 'samples': 7707648, 'steps': 15053, 'loss/train': 2.363438129425049} 02/24/2022 21:04:08 - INFO - codeparrot_training - Step 15054: {'lr': 0.00041417112035058157, 'samples': 7708160, 'steps': 15054, 'loss/train': 2.3568363189697266} 02/24/2022 21:04:14 - INFO - codeparrot_training - Step 15055: {'lr': 0.00041415877999601236, 'samples': 7708672, 'steps': 15055, 'loss/train': 1.926734447479248} 02/24/2022 21:04:17 - INFO - codeparrot_training - Step 15056: {'lr': 0.0004141464389382391, 'samples': 7709184, 'steps': 15056, 'loss/train': 2.082723379135132} 02/24/2022 21:04:24 - INFO - codeparrot_training - Step 15057: {'lr': 0.0004141340971773147, 'samples': 7709696, 'steps': 15057, 'loss/train': 2.5265214443206787} 02/24/2022 21:04:28 - INFO - codeparrot_training - Step 15058: {'lr': 0.00041412175471329174, 'samples': 7710208, 'steps': 15058, 'loss/train': 2.0780766010284424} 02/24/2022 21:04:33 - INFO - codeparrot_training - Step 15059: {'lr': 0.0004141094115462234, 'samples': 7710720, 'steps': 15059, 'loss/train': 0.3992597162723541} 02/24/2022 21:04:37 - INFO - codeparrot_training - Step 15060: {'lr': 0.00041409706767616246, 'samples': 7711232, 'steps': 15060, 'loss/train': 1.4950170516967773} 02/24/2022 21:04:42 - INFO - codeparrot_training - Step 15061: {'lr': 0.0004140847231031618, 'samples': 7711744, 'steps': 15061, 'loss/train': 1.5474239587783813} 02/24/2022 21:04:46 - INFO - codeparrot_training - Step 15062: {'lr': 0.00041407237782727427, 'samples': 7712256, 'steps': 15062, 'loss/train': 1.1488640308380127} 02/24/2022 21:04:51 - INFO - codeparrot_training - Step 15063: {'lr': 0.0004140600318485527, 'samples': 7712768, 'steps': 15063, 'loss/train': 1.7876847982406616} 02/24/2022 21:04:55 - INFO - codeparrot_training - Step 15064: {'lr': 0.0004140476851670502, 'samples': 7713280, 'steps': 15064, 'loss/train': 1.9944267272949219} 02/24/2022 21:05:00 - INFO - codeparrot_training - Step 15065: {'lr': 0.00041403533778281934, 'samples': 7713792, 'steps': 15065, 'loss/train': 1.2161877155303955} 02/24/2022 21:05:04 - INFO - codeparrot_training - Step 15066: {'lr': 0.0004140229896959132, 'samples': 7714304, 'steps': 15066, 'loss/train': 0.2673460841178894} 02/24/2022 21:05:07 - INFO - codeparrot_training - Step 15067: {'lr': 0.00041401064090638474, 'samples': 7714816, 'steps': 15067, 'loss/train': 2.6707763671875} 02/24/2022 21:05:13 - INFO - codeparrot_training - Step 15068: {'lr': 0.0004139982914142868, 'samples': 7715328, 'steps': 15068, 'loss/train': 2.6089210510253906} 02/24/2022 21:05:19 - INFO - codeparrot_training - Step 15069: {'lr': 0.00041398594121967215, 'samples': 7715840, 'steps': 15069, 'loss/train': 2.262225389480591} 02/24/2022 21:05:23 - INFO - codeparrot_training - Step 15070: {'lr': 0.0004139735903225939, 'samples': 7716352, 'steps': 15070, 'loss/train': 1.618273377418518} 02/24/2022 21:05:28 - INFO - codeparrot_training - Step 15071: {'lr': 0.0004139612387231048, 'samples': 7716864, 'steps': 15071, 'loss/train': 1.9760456085205078} 02/24/2022 21:05:32 - INFO - codeparrot_training - Step 15072: {'lr': 0.0004139488864212578, 'samples': 7717376, 'steps': 15072, 'loss/train': 1.9000672101974487} 02/24/2022 21:05:37 - INFO - codeparrot_training - Step 15073: {'lr': 0.0004139365334171059, 'samples': 7717888, 'steps': 15073, 'loss/train': 2.652770757675171} 02/24/2022 21:05:41 - INFO - codeparrot_training - Step 15074: {'lr': 0.0004139241797107019, 'samples': 7718400, 'steps': 15074, 'loss/train': 2.3371622562408447} 02/24/2022 21:05:46 - INFO - codeparrot_training - Step 15075: {'lr': 0.00041391182530209873, 'samples': 7718912, 'steps': 15075, 'loss/train': 1.8400214910507202} 02/24/2022 21:05:50 - INFO - codeparrot_training - Step 15076: {'lr': 0.0004138994701913494, 'samples': 7719424, 'steps': 15076, 'loss/train': 1.8306312561035156} 02/24/2022 21:05:56 - INFO - codeparrot_training - Step 15077: {'lr': 0.00041388711437850676, 'samples': 7719936, 'steps': 15077, 'loss/train': 1.7445478439331055} 02/24/2022 21:05:59 - INFO - codeparrot_training - Step 15078: {'lr': 0.00041387475786362386, 'samples': 7720448, 'steps': 15078, 'loss/train': 1.8092299699783325} 02/24/2022 21:06:05 - INFO - codeparrot_training - Step 15079: {'lr': 0.0004138624006467534, 'samples': 7720960, 'steps': 15079, 'loss/train': 2.0429253578186035} 02/24/2022 21:06:08 - INFO - codeparrot_training - Step 15080: {'lr': 0.00041385004272794846, 'samples': 7721472, 'steps': 15080, 'loss/train': 1.4502273797988892} 02/24/2022 21:06:14 - INFO - codeparrot_training - Step 15081: {'lr': 0.00041383768410726207, 'samples': 7721984, 'steps': 15081, 'loss/train': 1.8575890064239502} 02/24/2022 21:06:17 - INFO - codeparrot_training - Step 15082: {'lr': 0.000413825324784747, 'samples': 7722496, 'steps': 15082, 'loss/train': 2.518663167953491} 02/24/2022 21:06:23 - INFO - codeparrot_training - Step 15083: {'lr': 0.00041381296476045626, 'samples': 7723008, 'steps': 15083, 'loss/train': 1.9789882898330688} 02/24/2022 21:06:26 - INFO - codeparrot_training - Step 15084: {'lr': 0.0004138006040344428, 'samples': 7723520, 'steps': 15084, 'loss/train': 1.4686317443847656} 02/24/2022 21:06:33 - INFO - codeparrot_training - Step 15085: {'lr': 0.0004137882426067595, 'samples': 7724032, 'steps': 15085, 'loss/train': 1.7022567987442017} 02/24/2022 21:06:36 - INFO - codeparrot_training - Step 15086: {'lr': 0.0004137758804774594, 'samples': 7724544, 'steps': 15086, 'loss/train': 1.742958664894104} 02/24/2022 21:06:42 - INFO - codeparrot_training - Step 15087: {'lr': 0.0004137635176465955, 'samples': 7725056, 'steps': 15087, 'loss/train': 1.699378490447998} 02/24/2022 21:06:45 - INFO - codeparrot_training - Step 15088: {'lr': 0.00041375115411422064, 'samples': 7725568, 'steps': 15088, 'loss/train': 1.7454075813293457} 02/24/2022 21:06:51 - INFO - codeparrot_training - Step 15089: {'lr': 0.0004137387898803878, 'samples': 7726080, 'steps': 15089, 'loss/train': 1.2566828727722168} 02/24/2022 21:06:54 - INFO - codeparrot_training - Step 15090: {'lr': 0.0004137264249451501, 'samples': 7726592, 'steps': 15090, 'loss/train': 2.319958448410034} 02/24/2022 21:07:00 - INFO - codeparrot_training - Step 15091: {'lr': 0.00041371405930856026, 'samples': 7727104, 'steps': 15091, 'loss/train': 2.1931228637695312} 02/24/2022 21:07:03 - INFO - codeparrot_training - Step 15092: {'lr': 0.00041370169297067145, 'samples': 7727616, 'steps': 15092, 'loss/train': 3.3477466106414795} 02/24/2022 21:07:09 - INFO - codeparrot_training - Step 15093: {'lr': 0.0004136893259315365, 'samples': 7728128, 'steps': 15093, 'loss/train': 2.4416987895965576} 02/24/2022 21:07:12 - INFO - codeparrot_training - Step 15094: {'lr': 0.00041367695819120854, 'samples': 7728640, 'steps': 15094, 'loss/train': 1.09297776222229} 02/24/2022 21:07:19 - INFO - codeparrot_training - Step 15095: {'lr': 0.0004136645897497404, 'samples': 7729152, 'steps': 15095, 'loss/train': 0.9568027257919312} 02/24/2022 21:07:23 - INFO - codeparrot_training - Step 15096: {'lr': 0.0004136522206071852, 'samples': 7729664, 'steps': 15096, 'loss/train': 1.05661940574646} 02/24/2022 21:07:28 - INFO - codeparrot_training - Step 15097: {'lr': 0.0004136398507635958, 'samples': 7730176, 'steps': 15097, 'loss/train': 2.524338483810425} 02/24/2022 21:07:31 - INFO - codeparrot_training - Step 15098: {'lr': 0.00041362748021902526, 'samples': 7730688, 'steps': 15098, 'loss/train': 1.827979326248169} 02/24/2022 21:07:37 - INFO - codeparrot_training - Step 15099: {'lr': 0.0004136151089735265, 'samples': 7731200, 'steps': 15099, 'loss/train': 1.1120802164077759} 02/24/2022 21:07:41 - INFO - codeparrot_training - Step 15100: {'lr': 0.00041360273702715263, 'samples': 7731712, 'steps': 15100, 'loss/train': 1.7640358209609985} 02/24/2022 21:07:46 - INFO - codeparrot_training - Step 15101: {'lr': 0.0004135903643799566, 'samples': 7732224, 'steps': 15101, 'loss/train': 1.9734355211257935} 02/24/2022 21:07:50 - INFO - codeparrot_training - Step 15102: {'lr': 0.00041357799103199127, 'samples': 7732736, 'steps': 15102, 'loss/train': 2.2872936725616455} 02/24/2022 21:07:55 - INFO - codeparrot_training - Step 15103: {'lr': 0.00041356561698330984, 'samples': 7733248, 'steps': 15103, 'loss/train': 2.3059182167053223} 02/24/2022 21:07:59 - INFO - codeparrot_training - Step 15104: {'lr': 0.0004135532422339653, 'samples': 7733760, 'steps': 15104, 'loss/train': 1.5279898643493652} 02/24/2022 21:08:05 - INFO - codeparrot_training - Step 15105: {'lr': 0.00041354086678401056, 'samples': 7734272, 'steps': 15105, 'loss/train': 2.68605375289917} 02/24/2022 21:08:09 - INFO - codeparrot_training - Step 15106: {'lr': 0.00041352849063349865, 'samples': 7734784, 'steps': 15106, 'loss/train': 3.3742055892944336} 02/24/2022 21:08:14 - INFO - codeparrot_training - Step 15107: {'lr': 0.0004135161137824827, 'samples': 7735296, 'steps': 15107, 'loss/train': 2.2265923023223877} 02/24/2022 21:08:18 - INFO - codeparrot_training - Step 15108: {'lr': 0.0004135037362310155, 'samples': 7735808, 'steps': 15108, 'loss/train': 0.9467727541923523} 02/24/2022 21:08:23 - INFO - codeparrot_training - Step 15109: {'lr': 0.0004134913579791503, 'samples': 7736320, 'steps': 15109, 'loss/train': 1.1220277547836304} 02/24/2022 21:08:27 - INFO - codeparrot_training - Step 15110: {'lr': 0.00041347897902694003, 'samples': 7736832, 'steps': 15110, 'loss/train': 1.8450486660003662} 02/24/2022 21:08:32 - INFO - codeparrot_training - Step 15111: {'lr': 0.00041346659937443775, 'samples': 7737344, 'steps': 15111, 'loss/train': 2.6501808166503906} 02/24/2022 21:08:36 - INFO - codeparrot_training - Step 15112: {'lr': 0.00041345421902169645, 'samples': 7737856, 'steps': 15112, 'loss/train': 0.6919535994529724} 02/24/2022 21:08:42 - INFO - codeparrot_training - Step 15113: {'lr': 0.0004134418379687691, 'samples': 7738368, 'steps': 15113, 'loss/train': 2.2594401836395264} 02/24/2022 21:08:45 - INFO - codeparrot_training - Step 15114: {'lr': 0.0004134294562157089, 'samples': 7738880, 'steps': 15114, 'loss/train': 1.914897084236145} 02/24/2022 21:08:51 - INFO - codeparrot_training - Step 15115: {'lr': 0.00041341707376256877, 'samples': 7739392, 'steps': 15115, 'loss/train': 2.244562864303589} 02/24/2022 21:08:55 - INFO - codeparrot_training - Step 15116: {'lr': 0.00041340469060940183, 'samples': 7739904, 'steps': 15116, 'loss/train': 3.3928701877593994} 02/24/2022 21:09:00 - INFO - codeparrot_training - Step 15117: {'lr': 0.0004133923067562611, 'samples': 7740416, 'steps': 15117, 'loss/train': 1.800842523574829} 02/24/2022 21:09:04 - INFO - codeparrot_training - Step 15118: {'lr': 0.0004133799222031995, 'samples': 7740928, 'steps': 15118, 'loss/train': 2.7789714336395264} 02/24/2022 21:09:10 - INFO - codeparrot_training - Step 15119: {'lr': 0.0004133675369502703, 'samples': 7741440, 'steps': 15119, 'loss/train': 1.9856935739517212} 02/24/2022 21:09:13 - INFO - codeparrot_training - Step 15120: {'lr': 0.0004133551509975264, 'samples': 7741952, 'steps': 15120, 'loss/train': 0.8502724170684814} 02/24/2022 21:09:19 - INFO - codeparrot_training - Step 15121: {'lr': 0.0004133427643450209, 'samples': 7742464, 'steps': 15121, 'loss/train': 0.7499862909317017} 02/24/2022 21:09:22 - INFO - codeparrot_training - Step 15122: {'lr': 0.0004133303769928068, 'samples': 7742976, 'steps': 15122, 'loss/train': 2.680826425552368} 02/24/2022 21:09:28 - INFO - codeparrot_training - Step 15123: {'lr': 0.00041331798894093735, 'samples': 7743488, 'steps': 15123, 'loss/train': 1.6448231935501099} 02/24/2022 21:09:31 - INFO - codeparrot_training - Step 15124: {'lr': 0.0004133056001894655, 'samples': 7744000, 'steps': 15124, 'loss/train': 2.183262586593628} 02/24/2022 21:09:37 - INFO - codeparrot_training - Step 15125: {'lr': 0.0004132932107384442, 'samples': 7744512, 'steps': 15125, 'loss/train': 2.57751727104187} 02/24/2022 21:09:40 - INFO - codeparrot_training - Step 15126: {'lr': 0.0004132808205879267, 'samples': 7745024, 'steps': 15126, 'loss/train': 2.0060176849365234} 02/24/2022 21:09:46 - INFO - codeparrot_training - Step 15127: {'lr': 0.000413268429737966, 'samples': 7745536, 'steps': 15127, 'loss/train': 1.2702980041503906} 02/24/2022 21:09:49 - INFO - codeparrot_training - Step 15128: {'lr': 0.00041325603818861517, 'samples': 7746048, 'steps': 15128, 'loss/train': 1.5013447999954224} 02/24/2022 21:09:55 - INFO - codeparrot_training - Step 15129: {'lr': 0.00041324364593992735, 'samples': 7746560, 'steps': 15129, 'loss/train': 2.2891414165496826} 02/24/2022 21:09:58 - INFO - codeparrot_training - Step 15130: {'lr': 0.00041323125299195563, 'samples': 7747072, 'steps': 15130, 'loss/train': 0.6653809547424316} 02/24/2022 21:10:04 - INFO - codeparrot_training - Step 15131: {'lr': 0.000413218859344753, 'samples': 7747584, 'steps': 15131, 'loss/train': 0.8936456441879272} 02/24/2022 21:10:08 - INFO - codeparrot_training - Step 15132: {'lr': 0.00041320646499837254, 'samples': 7748096, 'steps': 15132, 'loss/train': 2.230276584625244} 02/24/2022 21:10:13 - INFO - codeparrot_training - Step 15133: {'lr': 0.00041319406995286753, 'samples': 7748608, 'steps': 15133, 'loss/train': 2.176027536392212} 02/24/2022 21:10:17 - INFO - codeparrot_training - Step 15134: {'lr': 0.0004131816742082909, 'samples': 7749120, 'steps': 15134, 'loss/train': 1.561685562133789} 02/24/2022 21:10:22 - INFO - codeparrot_training - Step 15135: {'lr': 0.00041316927776469575, 'samples': 7749632, 'steps': 15135, 'loss/train': 3.352414846420288} 02/24/2022 21:10:26 - INFO - codeparrot_training - Step 15136: {'lr': 0.00041315688062213524, 'samples': 7750144, 'steps': 15136, 'loss/train': 2.368332624435425} 02/24/2022 21:10:31 - INFO - codeparrot_training - Step 15137: {'lr': 0.0004131444827806625, 'samples': 7750656, 'steps': 15137, 'loss/train': 1.6380267143249512} 02/24/2022 21:10:35 - INFO - codeparrot_training - Step 15138: {'lr': 0.00041313208424033056, 'samples': 7751168, 'steps': 15138, 'loss/train': 0.9953458309173584} 02/24/2022 21:10:41 - INFO - codeparrot_training - Step 15139: {'lr': 0.0004131196850011926, 'samples': 7751680, 'steps': 15139, 'loss/train': 1.9391264915466309} 02/24/2022 21:10:44 - INFO - codeparrot_training - Step 15140: {'lr': 0.0004131072850633017, 'samples': 7752192, 'steps': 15140, 'loss/train': 2.290933132171631} 02/24/2022 21:10:51 - INFO - codeparrot_training - Step 15141: {'lr': 0.00041309488442671093, 'samples': 7752704, 'steps': 15141, 'loss/train': 2.3020575046539307} 02/24/2022 21:10:54 - INFO - codeparrot_training - Step 15142: {'lr': 0.00041308248309147356, 'samples': 7753216, 'steps': 15142, 'loss/train': 2.1128060817718506} 02/24/2022 21:10:59 - INFO - codeparrot_training - Step 15143: {'lr': 0.00041307008105764256, 'samples': 7753728, 'steps': 15143, 'loss/train': 2.1988351345062256} 02/24/2022 21:11:03 - INFO - codeparrot_training - Step 15144: {'lr': 0.0004130576783252712, 'samples': 7754240, 'steps': 15144, 'loss/train': 2.371443510055542} 02/24/2022 21:11:09 - INFO - codeparrot_training - Step 15145: {'lr': 0.00041304527489441237, 'samples': 7754752, 'steps': 15145, 'loss/train': 1.2331807613372803} 02/24/2022 21:11:12 - INFO - codeparrot_training - Step 15146: {'lr': 0.0004130328707651195, 'samples': 7755264, 'steps': 15146, 'loss/train': 0.9676528573036194} 02/24/2022 21:11:18 - INFO - codeparrot_training - Step 15147: {'lr': 0.00041302046593744547, 'samples': 7755776, 'steps': 15147, 'loss/train': 1.0516717433929443} 02/24/2022 21:11:21 - INFO - codeparrot_training - Step 15148: {'lr': 0.00041300806041144356, 'samples': 7756288, 'steps': 15148, 'loss/train': 1.3651198148727417} 02/24/2022 21:11:27 - INFO - codeparrot_training - Step 15149: {'lr': 0.0004129956541871669, 'samples': 7756800, 'steps': 15149, 'loss/train': 1.8809645175933838} 02/24/2022 21:11:30 - INFO - codeparrot_training - Step 15150: {'lr': 0.00041298324726466855, 'samples': 7757312, 'steps': 15150, 'loss/train': 1.4244515895843506} 02/24/2022 21:11:36 - INFO - codeparrot_training - Step 15151: {'lr': 0.0004129708396440018, 'samples': 7757824, 'steps': 15151, 'loss/train': 1.4894310235977173} 02/24/2022 21:11:40 - INFO - codeparrot_training - Step 15152: {'lr': 0.00041295843132521973, 'samples': 7758336, 'steps': 15152, 'loss/train': 2.0084636211395264} 02/24/2022 21:11:45 - INFO - codeparrot_training - Step 15153: {'lr': 0.0004129460223083754, 'samples': 7758848, 'steps': 15153, 'loss/train': 1.8689044713974} 02/24/2022 21:11:49 - INFO - codeparrot_training - Step 15154: {'lr': 0.0004129336125935221, 'samples': 7759360, 'steps': 15154, 'loss/train': 0.3137350380420685} 02/24/2022 21:11:54 - INFO - codeparrot_training - Step 15155: {'lr': 0.000412921202180713, 'samples': 7759872, 'steps': 15155, 'loss/train': 1.7021143436431885} 02/24/2022 21:11:58 - INFO - codeparrot_training - Step 15156: {'lr': 0.00041290879107000114, 'samples': 7760384, 'steps': 15156, 'loss/train': 1.8608148097991943} 02/24/2022 21:12:03 - INFO - codeparrot_training - Step 15157: {'lr': 0.00041289637926143974, 'samples': 7760896, 'steps': 15157, 'loss/train': 2.5681891441345215} 02/24/2022 21:12:07 - INFO - codeparrot_training - Step 15158: {'lr': 0.000412883966755082, 'samples': 7761408, 'steps': 15158, 'loss/train': 2.3331186771392822} 02/24/2022 21:12:12 - INFO - codeparrot_training - Step 15159: {'lr': 0.000412871553550981, 'samples': 7761920, 'steps': 15159, 'loss/train': 1.8314324617385864} 02/24/2022 21:12:16 - INFO - codeparrot_training - Step 15160: {'lr': 0.00041285913964919006, 'samples': 7762432, 'steps': 15160, 'loss/train': 2.451927661895752} 02/24/2022 21:12:22 - INFO - codeparrot_training - Step 15161: {'lr': 0.0004128467250497623, 'samples': 7762944, 'steps': 15161, 'loss/train': 8.79805850982666} 02/24/2022 21:12:26 - INFO - codeparrot_training - Step 15162: {'lr': 0.00041283430975275085, 'samples': 7763456, 'steps': 15162, 'loss/train': 1.1373577117919922} 02/24/2022 21:12:31 - INFO - codeparrot_training - Step 15163: {'lr': 0.0004128218937582089, 'samples': 7763968, 'steps': 15163, 'loss/train': 1.5468223094940186} 02/24/2022 21:12:35 - INFO - codeparrot_training - Step 15164: {'lr': 0.00041280947706618965, 'samples': 7764480, 'steps': 15164, 'loss/train': 2.591919183731079} 02/24/2022 21:12:40 - INFO - codeparrot_training - Step 15165: {'lr': 0.00041279705967674636, 'samples': 7764992, 'steps': 15165, 'loss/train': 2.084432363510132} 02/24/2022 21:12:44 - INFO - codeparrot_training - Step 15166: {'lr': 0.00041278464158993214, 'samples': 7765504, 'steps': 15166, 'loss/train': 2.1277060508728027} 02/24/2022 21:12:49 - INFO - codeparrot_training - Step 15167: {'lr': 0.0004127722228058002, 'samples': 7766016, 'steps': 15167, 'loss/train': 1.946176528930664} 02/24/2022 21:12:53 - INFO - codeparrot_training - Step 15168: {'lr': 0.0004127598033244037, 'samples': 7766528, 'steps': 15168, 'loss/train': 1.6518677473068237} 02/24/2022 21:12:58 - INFO - codeparrot_training - Step 15169: {'lr': 0.0004127473831457959, 'samples': 7767040, 'steps': 15169, 'loss/train': 2.1408612728118896} 02/24/2022 21:13:02 - INFO - codeparrot_training - Step 15170: {'lr': 0.00041273496227003004, 'samples': 7767552, 'steps': 15170, 'loss/train': 1.7052603960037231} 02/24/2022 21:13:08 - INFO - codeparrot_training - Step 15171: {'lr': 0.0004127225406971592, 'samples': 7768064, 'steps': 15171, 'loss/train': 2.443181276321411} 02/24/2022 21:13:11 - INFO - codeparrot_training - Step 15172: {'lr': 0.00041271011842723676, 'samples': 7768576, 'steps': 15172, 'loss/train': 0.3734479546546936} 02/24/2022 21:13:17 - INFO - codeparrot_training - Step 15173: {'lr': 0.00041269769546031576, 'samples': 7769088, 'steps': 15173, 'loss/train': 1.757319688796997} 02/24/2022 21:13:20 - INFO - codeparrot_training - Step 15174: {'lr': 0.0004126852717964495, 'samples': 7769600, 'steps': 15174, 'loss/train': 2.8355581760406494} 02/24/2022 21:13:26 - INFO - codeparrot_training - Step 15175: {'lr': 0.0004126728474356912, 'samples': 7770112, 'steps': 15175, 'loss/train': 3.0845136642456055} 02/24/2022 21:13:29 - INFO - codeparrot_training - Step 15176: {'lr': 0.0004126604223780941, 'samples': 7770624, 'steps': 15176, 'loss/train': 1.1025235652923584} 02/24/2022 21:13:35 - INFO - codeparrot_training - Step 15177: {'lr': 0.00041264799662371144, 'samples': 7771136, 'steps': 15177, 'loss/train': 2.0873119831085205} 02/24/2022 21:13:38 - INFO - codeparrot_training - Step 15178: {'lr': 0.0004126355701725963, 'samples': 7771648, 'steps': 15178, 'loss/train': 1.4091343879699707} 02/24/2022 21:13:44 - INFO - codeparrot_training - Step 15179: {'lr': 0.00041262314302480216, 'samples': 7772160, 'steps': 15179, 'loss/train': 2.303143262863159} 02/24/2022 21:13:47 - INFO - codeparrot_training - Step 15180: {'lr': 0.000412610715180382, 'samples': 7772672, 'steps': 15180, 'loss/train': 1.3482143878936768} 02/24/2022 21:13:53 - INFO - codeparrot_training - Step 15181: {'lr': 0.0004125982866393892, 'samples': 7773184, 'steps': 15181, 'loss/train': 0.5258373022079468} 02/24/2022 21:13:56 - INFO - codeparrot_training - Step 15182: {'lr': 0.0004125858574018769, 'samples': 7773696, 'steps': 15182, 'loss/train': 0.338752806186676} 02/24/2022 21:14:02 - INFO - codeparrot_training - Step 15183: {'lr': 0.0004125734274678986, 'samples': 7774208, 'steps': 15183, 'loss/train': 3.331299066543579} 02/24/2022 21:14:06 - INFO - codeparrot_training - Step 15184: {'lr': 0.0004125609968375072, 'samples': 7774720, 'steps': 15184, 'loss/train': 1.6494851112365723} 02/24/2022 21:14:11 - INFO - codeparrot_training - Step 15185: {'lr': 0.00041254856551075616, 'samples': 7775232, 'steps': 15185, 'loss/train': 2.7761707305908203} 02/24/2022 21:14:15 - INFO - codeparrot_training - Step 15186: {'lr': 0.0004125361334876987, 'samples': 7775744, 'steps': 15186, 'loss/train': 2.21018648147583} 02/24/2022 21:14:21 - INFO - codeparrot_training - Step 15187: {'lr': 0.000412523700768388, 'samples': 7776256, 'steps': 15187, 'loss/train': 0.8293166160583496} 02/24/2022 21:14:24 - INFO - codeparrot_training - Step 15188: {'lr': 0.0004125112673528775, 'samples': 7776768, 'steps': 15188, 'loss/train': 0.7854377031326294} 02/24/2022 21:14:30 - INFO - codeparrot_training - Step 15189: {'lr': 0.0004124988332412202, 'samples': 7777280, 'steps': 15189, 'loss/train': 1.8319604396820068} 02/24/2022 21:14:33 - INFO - codeparrot_training - Step 15190: {'lr': 0.00041248639843346953, 'samples': 7777792, 'steps': 15190, 'loss/train': 2.1161608695983887} 02/24/2022 21:14:39 - INFO - codeparrot_training - Step 15191: {'lr': 0.0004124739629296787, 'samples': 7778304, 'steps': 15191, 'loss/train': 2.0222506523132324} 02/24/2022 21:14:42 - INFO - codeparrot_training - Step 15192: {'lr': 0.00041246152672990105, 'samples': 7778816, 'steps': 15192, 'loss/train': 1.7895108461380005} 02/24/2022 21:14:48 - INFO - codeparrot_training - Step 15193: {'lr': 0.00041244908983418985, 'samples': 7779328, 'steps': 15193, 'loss/train': 1.4329543113708496} 02/24/2022 21:14:51 - INFO - codeparrot_training - Step 15194: {'lr': 0.0004124366522425982, 'samples': 7779840, 'steps': 15194, 'loss/train': 2.520211696624756} 02/24/2022 21:14:57 - INFO - codeparrot_training - Step 15195: {'lr': 0.0004124242139551796, 'samples': 7780352, 'steps': 15195, 'loss/train': 1.280103087425232} 02/24/2022 21:15:00 - INFO - codeparrot_training - Step 15196: {'lr': 0.00041241177497198725, 'samples': 7780864, 'steps': 15196, 'loss/train': 1.65938401222229} 02/24/2022 21:15:07 - INFO - codeparrot_training - Step 15197: {'lr': 0.00041239933529307437, 'samples': 7781376, 'steps': 15197, 'loss/train': 1.9791457653045654} 02/24/2022 21:15:10 - INFO - codeparrot_training - Step 15198: {'lr': 0.00041238689491849434, 'samples': 7781888, 'steps': 15198, 'loss/train': 1.8962838649749756} 02/24/2022 21:15:16 - INFO - codeparrot_training - Step 15199: {'lr': 0.00041237445384830043, 'samples': 7782400, 'steps': 15199, 'loss/train': 1.5832501649856567} 02/24/2022 21:15:19 - INFO - codeparrot_training - Step 15200: {'lr': 0.0004123620120825459, 'samples': 7782912, 'steps': 15200, 'loss/train': 2.6603612899780273} 02/24/2022 21:15:25 - INFO - codeparrot_training - Step 15201: {'lr': 0.0004123495696212841, 'samples': 7783424, 'steps': 15201, 'loss/train': 1.180666208267212} 02/24/2022 21:15:28 - INFO - codeparrot_training - Step 15202: {'lr': 0.00041233712646456823, 'samples': 7783936, 'steps': 15202, 'loss/train': 2.764810085296631} 02/24/2022 21:15:34 - INFO - codeparrot_training - Step 15203: {'lr': 0.0004123246826124517, 'samples': 7784448, 'steps': 15203, 'loss/train': 1.659140706062317} 02/24/2022 21:15:37 - INFO - codeparrot_training - Step 15204: {'lr': 0.00041231223806498777, 'samples': 7784960, 'steps': 15204, 'loss/train': 2.4636311531066895} 02/24/2022 21:15:43 - INFO - codeparrot_training - Step 15205: {'lr': 0.0004122997928222298, 'samples': 7785472, 'steps': 15205, 'loss/train': 1.4005694389343262} 02/24/2022 21:15:46 - INFO - codeparrot_training - Step 15206: {'lr': 0.000412287346884231, 'samples': 7785984, 'steps': 15206, 'loss/train': 3.509793758392334} 02/24/2022 21:15:52 - INFO - codeparrot_training - Step 15207: {'lr': 0.00041227490025104474, 'samples': 7786496, 'steps': 15207, 'loss/train': 2.6965408325195312} 02/24/2022 21:15:56 - INFO - codeparrot_training - Step 15208: {'lr': 0.00041226245292272433, 'samples': 7787008, 'steps': 15208, 'loss/train': 2.2454934120178223} 02/24/2022 21:16:02 - INFO - codeparrot_training - Step 15209: {'lr': 0.00041225000489932315, 'samples': 7787520, 'steps': 15209, 'loss/train': 2.4381909370422363} 02/24/2022 21:16:05 - INFO - codeparrot_training - Step 15210: {'lr': 0.00041223755618089445, 'samples': 7788032, 'steps': 15210, 'loss/train': 0.518037736415863} 02/24/2022 21:16:11 - INFO - codeparrot_training - Step 15211: {'lr': 0.0004122251067674915, 'samples': 7788544, 'steps': 15211, 'loss/train': 1.7804791927337646} 02/24/2022 21:16:14 - INFO - codeparrot_training - Step 15212: {'lr': 0.00041221265665916776, 'samples': 7789056, 'steps': 15212, 'loss/train': 3.487273931503296} 02/24/2022 21:16:20 - INFO - codeparrot_training - Step 15213: {'lr': 0.0004122002058559765, 'samples': 7789568, 'steps': 15213, 'loss/train': 1.683398723602295} 02/24/2022 21:16:23 - INFO - codeparrot_training - Step 15214: {'lr': 0.00041218775435797106, 'samples': 7790080, 'steps': 15214, 'loss/train': 2.7925148010253906} 02/24/2022 21:16:29 - INFO - codeparrot_training - Step 15215: {'lr': 0.0004121753021652048, 'samples': 7790592, 'steps': 15215, 'loss/train': 3.0459232330322266} 02/24/2022 21:16:32 - INFO - codeparrot_training - Step 15216: {'lr': 0.0004121628492777311, 'samples': 7791104, 'steps': 15216, 'loss/train': 2.6547303199768066} 02/24/2022 21:16:38 - INFO - codeparrot_training - Step 15217: {'lr': 0.0004121503956956031, 'samples': 7791616, 'steps': 15217, 'loss/train': 2.400583267211914} 02/24/2022 21:16:41 - INFO - codeparrot_training - Step 15218: {'lr': 0.0004121379414188744, 'samples': 7792128, 'steps': 15218, 'loss/train': 2.671135902404785} 02/24/2022 21:16:47 - INFO - codeparrot_training - Step 15219: {'lr': 0.0004121254864475982, 'samples': 7792640, 'steps': 15219, 'loss/train': 2.5322554111480713} 02/24/2022 21:16:51 - INFO - codeparrot_training - Step 15220: {'lr': 0.0004121130307818279, 'samples': 7793152, 'steps': 15220, 'loss/train': 3.069416046142578} 02/24/2022 21:16:57 - INFO - codeparrot_training - Step 15221: {'lr': 0.00041210057442161687, 'samples': 7793664, 'steps': 15221, 'loss/train': 2.3191070556640625} 02/24/2022 21:17:00 - INFO - codeparrot_training - Step 15222: {'lr': 0.0004120881173670184, 'samples': 7794176, 'steps': 15222, 'loss/train': 1.047019600868225} 02/24/2022 21:17:06 - INFO - codeparrot_training - Step 15223: {'lr': 0.000412075659618086, 'samples': 7794688, 'steps': 15223, 'loss/train': 2.5615899562835693} 02/24/2022 21:17:09 - INFO - codeparrot_training - Step 15224: {'lr': 0.0004120632011748728, 'samples': 7795200, 'steps': 15224, 'loss/train': 1.7023859024047852} 02/24/2022 21:17:15 - INFO - codeparrot_training - Step 15225: {'lr': 0.00041205074203743244, 'samples': 7795712, 'steps': 15225, 'loss/train': 0.7751212120056152} 02/24/2022 21:17:18 - INFO - codeparrot_training - Step 15226: {'lr': 0.00041203828220581805, 'samples': 7796224, 'steps': 15226, 'loss/train': 2.151682138442993} 02/24/2022 21:17:24 - INFO - codeparrot_training - Step 15227: {'lr': 0.00041202582168008324, 'samples': 7796736, 'steps': 15227, 'loss/train': 2.54848051071167} 02/24/2022 21:17:27 - INFO - codeparrot_training - Step 15228: {'lr': 0.00041201336046028117, 'samples': 7797248, 'steps': 15228, 'loss/train': 2.7814154624938965} 02/24/2022 21:17:33 - INFO - codeparrot_training - Step 15229: {'lr': 0.0004120008985464654, 'samples': 7797760, 'steps': 15229, 'loss/train': 0.9572920799255371} 02/24/2022 21:17:36 - INFO - codeparrot_training - Step 15230: {'lr': 0.0004119884359386891, 'samples': 7798272, 'steps': 15230, 'loss/train': 2.5589332580566406} 02/24/2022 21:17:42 - INFO - codeparrot_training - Step 15231: {'lr': 0.0004119759726370058, 'samples': 7798784, 'steps': 15231, 'loss/train': 2.0106046199798584} 02/24/2022 21:17:45 - INFO - codeparrot_training - Step 15232: {'lr': 0.0004119635086414689, 'samples': 7799296, 'steps': 15232, 'loss/train': 1.7771512269973755} 02/24/2022 21:17:52 - INFO - codeparrot_training - Step 15233: {'lr': 0.0004119510439521318, 'samples': 7799808, 'steps': 15233, 'loss/train': 1.5864630937576294} 02/24/2022 21:17:55 - INFO - codeparrot_training - Step 15234: {'lr': 0.0004119385785690478, 'samples': 7800320, 'steps': 15234, 'loss/train': 1.471492052078247} 02/24/2022 21:18:01 - INFO - codeparrot_training - Step 15235: {'lr': 0.0004119261124922703, 'samples': 7800832, 'steps': 15235, 'loss/train': 1.107556700706482} 02/24/2022 21:18:04 - INFO - codeparrot_training - Step 15236: {'lr': 0.00041191364572185286, 'samples': 7801344, 'steps': 15236, 'loss/train': 1.073188066482544} 02/24/2022 21:18:10 - INFO - codeparrot_training - Step 15237: {'lr': 0.0004119011782578487, 'samples': 7801856, 'steps': 15237, 'loss/train': 2.5408918857574463} 02/24/2022 21:18:13 - INFO - codeparrot_training - Step 15238: {'lr': 0.00041188871010031135, 'samples': 7802368, 'steps': 15238, 'loss/train': 2.212597370147705} 02/24/2022 21:18:19 - INFO - codeparrot_training - Step 15239: {'lr': 0.0004118762412492941, 'samples': 7802880, 'steps': 15239, 'loss/train': 1.766115665435791} 02/24/2022 21:18:22 - INFO - codeparrot_training - Step 15240: {'lr': 0.00041186377170485057, 'samples': 7803392, 'steps': 15240, 'loss/train': 1.3296129703521729} 02/24/2022 21:18:28 - INFO - codeparrot_training - Step 15241: {'lr': 0.00041185130146703387, 'samples': 7803904, 'steps': 15241, 'loss/train': 2.002227306365967} 02/24/2022 21:18:31 - INFO - codeparrot_training - Step 15242: {'lr': 0.0004118388305358977, 'samples': 7804416, 'steps': 15242, 'loss/train': 2.3698153495788574} 02/24/2022 21:18:37 - INFO - codeparrot_training - Step 15243: {'lr': 0.0004118263589114953, 'samples': 7804928, 'steps': 15243, 'loss/train': 1.9070606231689453} 02/24/2022 21:18:41 - INFO - codeparrot_training - Step 15244: {'lr': 0.00041181388659388026, 'samples': 7805440, 'steps': 15244, 'loss/train': 1.0141035318374634} 02/24/2022 21:18:46 - INFO - codeparrot_training - Step 15245: {'lr': 0.00041180141358310586, 'samples': 7805952, 'steps': 15245, 'loss/train': 2.0231447219848633} 02/24/2022 21:18:50 - INFO - codeparrot_training - Step 15246: {'lr': 0.00041178893987922556, 'samples': 7806464, 'steps': 15246, 'loss/train': 2.5819156169891357} 02/24/2022 21:18:55 - INFO - codeparrot_training - Step 15247: {'lr': 0.0004117764654822929, 'samples': 7806976, 'steps': 15247, 'loss/train': 1.5870237350463867} 02/24/2022 21:18:59 - INFO - codeparrot_training - Step 15248: {'lr': 0.0004117639903923611, 'samples': 7807488, 'steps': 15248, 'loss/train': 2.968144416809082} 02/24/2022 21:19:04 - INFO - codeparrot_training - Step 15249: {'lr': 0.0004117515146094838, 'samples': 7808000, 'steps': 15249, 'loss/train': 2.2506957054138184} 02/24/2022 21:19:08 - INFO - codeparrot_training - Step 15250: {'lr': 0.0004117390381337144, 'samples': 7808512, 'steps': 15250, 'loss/train': 2.5934667587280273} 02/24/2022 21:19:14 - INFO - codeparrot_training - Step 15251: {'lr': 0.00041172656096510624, 'samples': 7809024, 'steps': 15251, 'loss/train': 1.9710956811904907} 02/24/2022 21:19:17 - INFO - codeparrot_training - Step 15252: {'lr': 0.0004117140831037129, 'samples': 7809536, 'steps': 15252, 'loss/train': 3.5833899974823} 02/24/2022 21:19:23 - INFO - codeparrot_training - Step 15253: {'lr': 0.00041170160454958785, 'samples': 7810048, 'steps': 15253, 'loss/train': 1.8081755638122559} 02/24/2022 21:19:26 - INFO - codeparrot_training - Step 15254: {'lr': 0.00041168912530278434, 'samples': 7810560, 'steps': 15254, 'loss/train': 0.6070297360420227} 02/24/2022 21:19:32 - INFO - codeparrot_training - Step 15255: {'lr': 0.00041167664536335605, 'samples': 7811072, 'steps': 15255, 'loss/train': 2.3863115310668945} 02/24/2022 21:19:36 - INFO - codeparrot_training - Step 15256: {'lr': 0.0004116641647313563, 'samples': 7811584, 'steps': 15256, 'loss/train': 1.7708550691604614} 02/24/2022 21:19:41 - INFO - codeparrot_training - Step 15257: {'lr': 0.00041165168340683857, 'samples': 7812096, 'steps': 15257, 'loss/train': 1.7683138847351074} 02/24/2022 21:19:45 - INFO - codeparrot_training - Step 15258: {'lr': 0.0004116392013898564, 'samples': 7812608, 'steps': 15258, 'loss/train': 1.6458642482757568} 02/24/2022 21:19:50 - INFO - codeparrot_training - Step 15259: {'lr': 0.0004116267186804632, 'samples': 7813120, 'steps': 15259, 'loss/train': 1.3484008312225342} 02/24/2022 21:19:54 - INFO - codeparrot_training - Step 15260: {'lr': 0.0004116142352787125, 'samples': 7813632, 'steps': 15260, 'loss/train': 2.22202467918396} 02/24/2022 21:19:59 - INFO - codeparrot_training - Step 15261: {'lr': 0.0004116017511846577, 'samples': 7814144, 'steps': 15261, 'loss/train': 0.2391817569732666} 02/24/2022 21:20:03 - INFO - codeparrot_training - Step 15262: {'lr': 0.00041158926639835234, 'samples': 7814656, 'steps': 15262, 'loss/train': 2.558009386062622} 02/24/2022 21:20:08 - INFO - codeparrot_training - Step 15263: {'lr': 0.00041157678091984987, 'samples': 7815168, 'steps': 15263, 'loss/train': 2.524418354034424} 02/24/2022 21:20:12 - INFO - codeparrot_training - Step 15264: {'lr': 0.0004115642947492038, 'samples': 7815680, 'steps': 15264, 'loss/train': 1.3695374727249146} 02/24/2022 21:20:17 - INFO - codeparrot_training - Step 15265: {'lr': 0.0004115518078864675, 'samples': 7816192, 'steps': 15265, 'loss/train': 2.512526512145996} 02/24/2022 21:20:21 - INFO - codeparrot_training - Step 15266: {'lr': 0.0004115393203316946, 'samples': 7816704, 'steps': 15266, 'loss/train': 1.702041745185852} 02/24/2022 21:20:26 - INFO - codeparrot_training - Step 15267: {'lr': 0.00041152683208493855, 'samples': 7817216, 'steps': 15267, 'loss/train': 2.142240524291992} 02/24/2022 21:20:30 - INFO - codeparrot_training - Step 15268: {'lr': 0.0004115143431462529, 'samples': 7817728, 'steps': 15268, 'loss/train': 1.0065242052078247} 02/24/2022 21:20:36 - INFO - codeparrot_training - Step 15269: {'lr': 0.000411501853515691, 'samples': 7818240, 'steps': 15269, 'loss/train': 1.8284953832626343} 02/24/2022 21:20:39 - INFO - codeparrot_training - Step 15270: {'lr': 0.00041148936319330656, 'samples': 7818752, 'steps': 15270, 'loss/train': 1.7723549604415894} 02/24/2022 21:20:45 - INFO - codeparrot_training - Step 15271: {'lr': 0.0004114768721791529, 'samples': 7819264, 'steps': 15271, 'loss/train': 1.4771944284439087} 02/24/2022 21:20:48 - INFO - codeparrot_training - Step 15272: {'lr': 0.00041146438047328347, 'samples': 7819776, 'steps': 15272, 'loss/train': 2.116975784301758} 02/24/2022 21:20:54 - INFO - codeparrot_training - Step 15273: {'lr': 0.00041145188807575206, 'samples': 7820288, 'steps': 15273, 'loss/train': 2.1834287643432617} 02/24/2022 21:20:57 - INFO - codeparrot_training - Step 15274: {'lr': 0.000411439394986612, 'samples': 7820800, 'steps': 15274, 'loss/train': 0.97948157787323} 02/24/2022 21:21:03 - INFO - codeparrot_training - Step 15275: {'lr': 0.00041142690120591686, 'samples': 7821312, 'steps': 15275, 'loss/train': 3.0842692852020264} 02/24/2022 21:21:06 - INFO - codeparrot_training - Step 15276: {'lr': 0.0004114144067337201, 'samples': 7821824, 'steps': 15276, 'loss/train': 1.8898016214370728} 02/24/2022 21:21:12 - INFO - codeparrot_training - Step 15277: {'lr': 0.0004114019115700752, 'samples': 7822336, 'steps': 15277, 'loss/train': 1.829763650894165} 02/24/2022 21:21:15 - INFO - codeparrot_training - Step 15278: {'lr': 0.00041138941571503587, 'samples': 7822848, 'steps': 15278, 'loss/train': 1.9355000257492065} 02/24/2022 21:21:21 - INFO - codeparrot_training - Step 15279: {'lr': 0.0004113769191686555, 'samples': 7823360, 'steps': 15279, 'loss/train': 1.2846788167953491} 02/24/2022 21:21:24 - INFO - codeparrot_training - Step 15280: {'lr': 0.00041136442193098765, 'samples': 7823872, 'steps': 15280, 'loss/train': 2.2391481399536133} 02/24/2022 21:21:30 - INFO - codeparrot_training - Step 15281: {'lr': 0.00041135192400208585, 'samples': 7824384, 'steps': 15281, 'loss/train': 2.0184123516082764} 02/24/2022 21:21:34 - INFO - codeparrot_training - Step 15282: {'lr': 0.00041133942538200364, 'samples': 7824896, 'steps': 15282, 'loss/train': 2.0375027656555176} 02/24/2022 21:21:39 - INFO - codeparrot_training - Step 15283: {'lr': 0.0004113269260707946, 'samples': 7825408, 'steps': 15283, 'loss/train': 2.2702300548553467} 02/24/2022 21:21:43 - INFO - codeparrot_training - Step 15284: {'lr': 0.0004113144260685122, 'samples': 7825920, 'steps': 15284, 'loss/train': 2.0824499130249023} 02/24/2022 21:21:48 - INFO - codeparrot_training - Step 15285: {'lr': 0.00041130192537521, 'samples': 7826432, 'steps': 15285, 'loss/train': 1.392338752746582} 02/24/2022 21:21:52 - INFO - codeparrot_training - Step 15286: {'lr': 0.0004112894239909416, 'samples': 7826944, 'steps': 15286, 'loss/train': 1.3191852569580078} 02/24/2022 21:21:57 - INFO - codeparrot_training - Step 15287: {'lr': 0.0004112769219157605, 'samples': 7827456, 'steps': 15287, 'loss/train': 2.404175281524658} 02/24/2022 21:22:01 - INFO - codeparrot_training - Step 15288: {'lr': 0.00041126441914972036, 'samples': 7827968, 'steps': 15288, 'loss/train': 2.292914867401123} 02/24/2022 21:22:07 - INFO - codeparrot_training - Step 15289: {'lr': 0.00041125191569287456, 'samples': 7828480, 'steps': 15289, 'loss/train': 1.7711459398269653} 02/24/2022 21:22:10 - INFO - codeparrot_training - Step 15290: {'lr': 0.0004112394115452768, 'samples': 7828992, 'steps': 15290, 'loss/train': 2.0460455417633057} 02/24/2022 21:22:16 - INFO - codeparrot_training - Step 15291: {'lr': 0.00041122690670698054, 'samples': 7829504, 'steps': 15291, 'loss/train': 2.0301671028137207} 02/24/2022 21:22:19 - INFO - codeparrot_training - Step 15292: {'lr': 0.0004112144011780395, 'samples': 7830016, 'steps': 15292, 'loss/train': 1.9984729290008545} 02/24/2022 21:22:25 - INFO - codeparrot_training - Step 15293: {'lr': 0.00041120189495850713, 'samples': 7830528, 'steps': 15293, 'loss/train': 1.0612894296646118} 02/24/2022 21:22:29 - INFO - codeparrot_training - Step 15294: {'lr': 0.000411189388048437, 'samples': 7831040, 'steps': 15294, 'loss/train': 2.338893413543701} 02/24/2022 21:22:34 - INFO - codeparrot_training - Step 15295: {'lr': 0.0004111768804478827, 'samples': 7831552, 'steps': 15295, 'loss/train': 1.4726080894470215} 02/24/2022 21:22:38 - INFO - codeparrot_training - Step 15296: {'lr': 0.00041116437215689785, 'samples': 7832064, 'steps': 15296, 'loss/train': 1.815688133239746} 02/24/2022 21:22:43 - INFO - codeparrot_training - Step 15297: {'lr': 0.000411151863175536, 'samples': 7832576, 'steps': 15297, 'loss/train': 2.408508062362671} 02/24/2022 21:22:47 - INFO - codeparrot_training - Step 15298: {'lr': 0.00041113935350385074, 'samples': 7833088, 'steps': 15298, 'loss/train': 2.378605365753174} 02/24/2022 21:22:52 - INFO - codeparrot_training - Step 15299: {'lr': 0.0004111268431418957, 'samples': 7833600, 'steps': 15299, 'loss/train': 1.9653651714324951} 02/24/2022 21:22:56 - INFO - codeparrot_training - Step 15300: {'lr': 0.0004111143320897244, 'samples': 7834112, 'steps': 15300, 'loss/train': 2.33986496925354} 02/24/2022 21:23:03 - INFO - codeparrot_training - Step 15301: {'lr': 0.0004111018203473904, 'samples': 7834624, 'steps': 15301, 'loss/train': 3.817391872406006} 02/24/2022 21:23:06 - INFO - codeparrot_training - Step 15302: {'lr': 0.0004110893079149474, 'samples': 7835136, 'steps': 15302, 'loss/train': 2.3886022567749023} 02/24/2022 21:23:12 - INFO - codeparrot_training - Step 15303: {'lr': 0.000411076794792449, 'samples': 7835648, 'steps': 15303, 'loss/train': 2.1376419067382812} 02/24/2022 21:23:15 - INFO - codeparrot_training - Step 15304: {'lr': 0.0004110642809799487, 'samples': 7836160, 'steps': 15304, 'loss/train': 1.9003889560699463} 02/24/2022 21:23:21 - INFO - codeparrot_training - Step 15305: {'lr': 0.0004110517664775002, 'samples': 7836672, 'steps': 15305, 'loss/train': 2.674116611480713} 02/24/2022 21:23:24 - INFO - codeparrot_training - Step 15306: {'lr': 0.00041103925128515705, 'samples': 7837184, 'steps': 15306, 'loss/train': 2.4968910217285156} 02/24/2022 21:23:30 - INFO - codeparrot_training - Step 15307: {'lr': 0.0004110267354029729, 'samples': 7837696, 'steps': 15307, 'loss/train': 1.8054479360580444} 02/24/2022 21:23:33 - INFO - codeparrot_training - Step 15308: {'lr': 0.0004110142188310013, 'samples': 7838208, 'steps': 15308, 'loss/train': 0.8322159647941589} 02/24/2022 21:23:39 - INFO - codeparrot_training - Step 15309: {'lr': 0.00041100170156929596, 'samples': 7838720, 'steps': 15309, 'loss/train': 2.217796802520752} 02/24/2022 21:23:42 - INFO - codeparrot_training - Step 15310: {'lr': 0.0004109891836179105, 'samples': 7839232, 'steps': 15310, 'loss/train': 1.836825966835022} 02/24/2022 21:23:48 - INFO - codeparrot_training - Step 15311: {'lr': 0.0004109766649768984, 'samples': 7839744, 'steps': 15311, 'loss/train': 1.6202685832977295} 02/24/2022 21:23:51 - INFO - codeparrot_training - Step 15312: {'lr': 0.00041096414564631347, 'samples': 7840256, 'steps': 15312, 'loss/train': 2.732639789581299} 02/24/2022 21:23:57 - INFO - codeparrot_training - Step 15313: {'lr': 0.00041095162562620915, 'samples': 7840768, 'steps': 15313, 'loss/train': 1.9145981073379517} 02/24/2022 21:24:01 - INFO - codeparrot_training - Step 15314: {'lr': 0.00041093910491663926, 'samples': 7841280, 'steps': 15314, 'loss/train': 1.0388280153274536} 02/24/2022 21:24:06 - INFO - codeparrot_training - Step 15315: {'lr': 0.0004109265835176573, 'samples': 7841792, 'steps': 15315, 'loss/train': 2.090623140335083} 02/24/2022 21:24:09 - INFO - codeparrot_training - Step 15316: {'lr': 0.00041091406142931705, 'samples': 7842304, 'steps': 15316, 'loss/train': 0.30290287733078003} 02/24/2022 21:24:16 - INFO - codeparrot_training - Step 15317: {'lr': 0.00041090153865167196, 'samples': 7842816, 'steps': 15317, 'loss/train': 1.705505609512329} 02/24/2022 21:24:19 - INFO - codeparrot_training - Step 15318: {'lr': 0.0004108890151847758, 'samples': 7843328, 'steps': 15318, 'loss/train': 1.8073900938034058} 02/24/2022 21:24:25 - INFO - codeparrot_training - Step 15319: {'lr': 0.0004108764910286822, 'samples': 7843840, 'steps': 15319, 'loss/train': 2.8852314949035645} 02/24/2022 21:24:28 - INFO - codeparrot_training - Step 15320: {'lr': 0.00041086396618344475, 'samples': 7844352, 'steps': 15320, 'loss/train': 2.709652900695801} 02/24/2022 21:24:34 - INFO - codeparrot_training - Step 15321: {'lr': 0.0004108514406491172, 'samples': 7844864, 'steps': 15321, 'loss/train': 2.0468101501464844} 02/24/2022 21:24:37 - INFO - codeparrot_training - Step 15322: {'lr': 0.0004108389144257531, 'samples': 7845376, 'steps': 15322, 'loss/train': 1.4501640796661377} 02/24/2022 21:24:43 - INFO - codeparrot_training - Step 15323: {'lr': 0.0004108263875134062, 'samples': 7845888, 'steps': 15323, 'loss/train': 1.9559978246688843} 02/24/2022 21:24:46 - INFO - codeparrot_training - Step 15324: {'lr': 0.0004108138599121301, 'samples': 7846400, 'steps': 15324, 'loss/train': 2.1975150108337402} 02/24/2022 21:24:52 - INFO - codeparrot_training - Step 15325: {'lr': 0.00041080133162197855, 'samples': 7846912, 'steps': 15325, 'loss/train': 2.0719447135925293} 02/24/2022 21:24:55 - INFO - codeparrot_training - Step 15326: {'lr': 0.0004107888026430051, 'samples': 7847424, 'steps': 15326, 'loss/train': 2.1070544719696045} 02/24/2022 21:25:01 - INFO - codeparrot_training - Step 15327: {'lr': 0.0004107762729752635, 'samples': 7847936, 'steps': 15327, 'loss/train': 2.381730556488037} 02/24/2022 21:25:05 - INFO - codeparrot_training - Step 15328: {'lr': 0.00041076374261880735, 'samples': 7848448, 'steps': 15328, 'loss/train': 0.9070643782615662} 02/24/2022 21:25:10 - INFO - codeparrot_training - Step 15329: {'lr': 0.0004107512115736904, 'samples': 7848960, 'steps': 15329, 'loss/train': 1.4146337509155273} 02/24/2022 21:25:14 - INFO - codeparrot_training - Step 15330: {'lr': 0.0004107386798399664, 'samples': 7849472, 'steps': 15330, 'loss/train': 1.7019490003585815} 02/24/2022 21:25:19 - INFO - codeparrot_training - Step 15331: {'lr': 0.00041072614741768877, 'samples': 7849984, 'steps': 15331, 'loss/train': 1.81195867061615} 02/24/2022 21:25:23 - INFO - codeparrot_training - Step 15332: {'lr': 0.00041071361430691143, 'samples': 7850496, 'steps': 15332, 'loss/train': 1.6398224830627441} 02/24/2022 21:25:28 - INFO - codeparrot_training - Step 15333: {'lr': 0.00041070108050768805, 'samples': 7851008, 'steps': 15333, 'loss/train': 2.1147563457489014} 02/24/2022 21:25:32 - INFO - codeparrot_training - Step 15334: {'lr': 0.00041068854602007224, 'samples': 7851520, 'steps': 15334, 'loss/train': 1.8375039100646973} 02/24/2022 21:25:37 - INFO - codeparrot_training - Step 15335: {'lr': 0.0004106760108441177, 'samples': 7852032, 'steps': 15335, 'loss/train': 2.073981761932373} 02/24/2022 21:25:41 - INFO - codeparrot_training - Step 15336: {'lr': 0.0004106634749798782, 'samples': 7852544, 'steps': 15336, 'loss/train': 2.662733554840088} 02/24/2022 21:25:48 - INFO - codeparrot_training - Step 15337: {'lr': 0.0004106509384274073, 'samples': 7853056, 'steps': 15337, 'loss/train': 0.6032879948616028} 02/24/2022 21:25:51 - INFO - codeparrot_training - Step 15338: {'lr': 0.0004106384011867589, 'samples': 7853568, 'steps': 15338, 'loss/train': 1.042371392250061} 02/24/2022 21:25:57 - INFO - codeparrot_training - Step 15339: {'lr': 0.00041062586325798654, 'samples': 7854080, 'steps': 15339, 'loss/train': 1.7701929807662964} 02/24/2022 21:26:00 - INFO - codeparrot_training - Step 15340: {'lr': 0.000410613324641144, 'samples': 7854592, 'steps': 15340, 'loss/train': 1.5835610628128052} 02/24/2022 21:26:06 - INFO - codeparrot_training - Step 15341: {'lr': 0.000410600785336285, 'samples': 7855104, 'steps': 15341, 'loss/train': 1.518117070198059} 02/24/2022 21:26:09 - INFO - codeparrot_training - Step 15342: {'lr': 0.0004105882453434632, 'samples': 7855616, 'steps': 15342, 'loss/train': 1.9308040142059326} 02/24/2022 21:26:15 - INFO - codeparrot_training - Step 15343: {'lr': 0.0004105757046627323, 'samples': 7856128, 'steps': 15343, 'loss/train': 2.4660017490386963} 02/24/2022 21:26:18 - INFO - codeparrot_training - Step 15344: {'lr': 0.00041056316329414613, 'samples': 7856640, 'steps': 15344, 'loss/train': 2.105105400085449} 02/24/2022 21:26:24 - INFO - codeparrot_training - Step 15345: {'lr': 0.0004105506212377583, 'samples': 7857152, 'steps': 15345, 'loss/train': 2.8279001712799072} 02/24/2022 21:26:27 - INFO - codeparrot_training - Step 15346: {'lr': 0.0004105380784936227, 'samples': 7857664, 'steps': 15346, 'loss/train': 2.1776533126831055} 02/24/2022 21:26:33 - INFO - codeparrot_training - Step 15347: {'lr': 0.0004105255350617928, 'samples': 7858176, 'steps': 15347, 'loss/train': 2.864000082015991} 02/24/2022 21:26:37 - INFO - codeparrot_training - Step 15348: {'lr': 0.0004105129909423226, 'samples': 7858688, 'steps': 15348, 'loss/train': 1.0162395238876343} 02/24/2022 21:26:42 - INFO - codeparrot_training - Step 15349: {'lr': 0.0004105004461352657, 'samples': 7859200, 'steps': 15349, 'loss/train': 2.198106527328491} 02/24/2022 21:26:46 - INFO - codeparrot_training - Step 15350: {'lr': 0.00041048790064067577, 'samples': 7859712, 'steps': 15350, 'loss/train': 2.8883795738220215} 02/24/2022 21:26:51 - INFO - codeparrot_training - Step 15351: {'lr': 0.0004104753544586067, 'samples': 7860224, 'steps': 15351, 'loss/train': 2.16083025932312} 02/24/2022 21:26:55 - INFO - codeparrot_training - Step 15352: {'lr': 0.0004104628075891121, 'samples': 7860736, 'steps': 15352, 'loss/train': 1.572336196899414} 02/24/2022 21:27:01 - INFO - codeparrot_training - Step 15353: {'lr': 0.00041045026003224593, 'samples': 7861248, 'steps': 15353, 'loss/train': 1.6916844844818115} 02/24/2022 21:27:04 - INFO - codeparrot_training - Step 15354: {'lr': 0.00041043771178806164, 'samples': 7861760, 'steps': 15354, 'loss/train': 3.4172563552856445} 02/24/2022 21:27:10 - INFO - codeparrot_training - Step 15355: {'lr': 0.00041042516285661325, 'samples': 7862272, 'steps': 15355, 'loss/train': 2.7301809787750244} 02/24/2022 21:27:13 - INFO - codeparrot_training - Step 15356: {'lr': 0.00041041261323795437, 'samples': 7862784, 'steps': 15356, 'loss/train': 2.331615447998047} 02/24/2022 21:27:19 - INFO - codeparrot_training - Step 15357: {'lr': 0.00041040006293213883, 'samples': 7863296, 'steps': 15357, 'loss/train': 1.5986061096191406} 02/24/2022 21:27:22 - INFO - codeparrot_training - Step 15358: {'lr': 0.0004103875119392203, 'samples': 7863808, 'steps': 15358, 'loss/train': 1.8355193138122559} 02/24/2022 21:27:28 - INFO - codeparrot_training - Step 15359: {'lr': 0.00041037496025925256, 'samples': 7864320, 'steps': 15359, 'loss/train': 0.35641559958457947} 02/24/2022 21:27:31 - INFO - codeparrot_training - Step 15360: {'lr': 0.0004103624078922895, 'samples': 7864832, 'steps': 15360, 'loss/train': 2.09504771232605} 02/24/2022 21:27:37 - INFO - codeparrot_training - Step 15361: {'lr': 0.0004103498548383847, 'samples': 7865344, 'steps': 15361, 'loss/train': 2.1360926628112793} 02/24/2022 21:27:40 - INFO - codeparrot_training - Step 15362: {'lr': 0.00041033730109759216, 'samples': 7865856, 'steps': 15362, 'loss/train': 2.20161771774292} 02/24/2022 21:27:47 - INFO - codeparrot_training - Step 15363: {'lr': 0.00041032474666996544, 'samples': 7866368, 'steps': 15363, 'loss/train': 1.1102571487426758} 02/24/2022 21:27:50 - INFO - codeparrot_training - Step 15364: {'lr': 0.0004103121915555585, 'samples': 7866880, 'steps': 15364, 'loss/train': 1.4902901649475098} 02/24/2022 21:27:56 - INFO - codeparrot_training - Step 15365: {'lr': 0.00041029963575442494, 'samples': 7867392, 'steps': 15365, 'loss/train': 1.7745418548583984} 02/24/2022 21:27:59 - INFO - codeparrot_training - Step 15366: {'lr': 0.0004102870792666187, 'samples': 7867904, 'steps': 15366, 'loss/train': 0.8546589016914368} 02/24/2022 21:28:05 - INFO - codeparrot_training - Step 15367: {'lr': 0.0004102745220921935, 'samples': 7868416, 'steps': 15367, 'loss/train': 3.093583822250366} 02/24/2022 21:28:08 - INFO - codeparrot_training - Step 15368: {'lr': 0.0004102619642312031, 'samples': 7868928, 'steps': 15368, 'loss/train': 2.066190242767334} 02/24/2022 21:28:14 - INFO - codeparrot_training - Step 15369: {'lr': 0.0004102494056837014, 'samples': 7869440, 'steps': 15369, 'loss/train': 4.045064449310303} 02/24/2022 21:28:17 - INFO - codeparrot_training - Step 15370: {'lr': 0.00041023684644974213, 'samples': 7869952, 'steps': 15370, 'loss/train': 2.705396890640259} 02/24/2022 21:28:23 - INFO - codeparrot_training - Step 15371: {'lr': 0.00041022428652937905, 'samples': 7870464, 'steps': 15371, 'loss/train': 2.214348793029785} 02/24/2022 21:28:26 - INFO - codeparrot_training - Step 15372: {'lr': 0.000410211725922666, 'samples': 7870976, 'steps': 15372, 'loss/train': 1.8127859830856323} 02/24/2022 21:28:33 - INFO - codeparrot_training - Step 15373: {'lr': 0.00041019916462965684, 'samples': 7871488, 'steps': 15373, 'loss/train': 1.5097529888153076} 02/24/2022 21:28:36 - INFO - codeparrot_training - Step 15374: {'lr': 0.0004101866026504053, 'samples': 7872000, 'steps': 15374, 'loss/train': 2.180556058883667} 02/24/2022 21:28:42 - INFO - codeparrot_training - Step 15375: {'lr': 0.00041017403998496523, 'samples': 7872512, 'steps': 15375, 'loss/train': 1.9452483654022217} 02/24/2022 21:28:45 - INFO - codeparrot_training - Step 15376: {'lr': 0.0004101614766333904, 'samples': 7873024, 'steps': 15376, 'loss/train': 1.2768136262893677} 02/24/2022 21:28:51 - INFO - codeparrot_training - Step 15377: {'lr': 0.0004101489125957347, 'samples': 7873536, 'steps': 15377, 'loss/train': 1.8025894165039062} 02/24/2022 21:28:54 - INFO - codeparrot_training - Step 15378: {'lr': 0.0004101363478720519, 'samples': 7874048, 'steps': 15378, 'loss/train': 2.392350673675537} 02/24/2022 21:29:00 - INFO - codeparrot_training - Step 15379: {'lr': 0.0004101237824623958, 'samples': 7874560, 'steps': 15379, 'loss/train': 1.0656155347824097} 02/24/2022 21:29:04 - INFO - codeparrot_training - Step 15380: {'lr': 0.00041011121636682024, 'samples': 7875072, 'steps': 15380, 'loss/train': 2.7382652759552} 02/24/2022 21:29:09 - INFO - codeparrot_training - Step 15381: {'lr': 0.0004100986495853791, 'samples': 7875584, 'steps': 15381, 'loss/train': 1.5157382488250732} 02/24/2022 21:29:12 - INFO - codeparrot_training - Step 15382: {'lr': 0.00041008608211812625, 'samples': 7876096, 'steps': 15382, 'loss/train': 2.0826926231384277} 02/24/2022 21:29:19 - INFO - codeparrot_training - Step 15383: {'lr': 0.00041007351396511537, 'samples': 7876608, 'steps': 15383, 'loss/train': 1.5692298412322998} 02/24/2022 21:29:22 - INFO - codeparrot_training - Step 15384: {'lr': 0.00041006094512640044, 'samples': 7877120, 'steps': 15384, 'loss/train': 1.4416319131851196} 02/24/2022 21:29:28 - INFO - codeparrot_training - Step 15385: {'lr': 0.00041004837560203525, 'samples': 7877632, 'steps': 15385, 'loss/train': 9.425049781799316} 02/24/2022 21:29:31 - INFO - codeparrot_training - Step 15386: {'lr': 0.0004100358053920736, 'samples': 7878144, 'steps': 15386, 'loss/train': 1.5387446880340576} 02/24/2022 21:29:37 - INFO - codeparrot_training - Step 15387: {'lr': 0.00041002323449656943, 'samples': 7878656, 'steps': 15387, 'loss/train': 2.125946283340454} 02/24/2022 21:29:41 - INFO - codeparrot_training - Step 15388: {'lr': 0.00041001066291557653, 'samples': 7879168, 'steps': 15388, 'loss/train': 2.213087797164917} 02/24/2022 21:29:46 - INFO - codeparrot_training - Step 15389: {'lr': 0.0004099980906491487, 'samples': 7879680, 'steps': 15389, 'loss/train': 9.124073028564453} 02/24/2022 21:29:50 - INFO - codeparrot_training - Step 15390: {'lr': 0.0004099855176973399, 'samples': 7880192, 'steps': 15390, 'loss/train': 1.6886916160583496} 02/24/2022 21:29:55 - INFO - codeparrot_training - Step 15391: {'lr': 0.0004099729440602039, 'samples': 7880704, 'steps': 15391, 'loss/train': 2.2797696590423584} 02/24/2022 21:29:59 - INFO - codeparrot_training - Step 15392: {'lr': 0.0004099603697377946, 'samples': 7881216, 'steps': 15392, 'loss/train': 2.404973268508911} 02/24/2022 21:30:05 - INFO - codeparrot_training - Step 15393: {'lr': 0.000409947794730166, 'samples': 7881728, 'steps': 15393, 'loss/train': 2.6005899906158447} 02/24/2022 21:30:09 - INFO - codeparrot_training - Step 15394: {'lr': 0.0004099352190373716, 'samples': 7882240, 'steps': 15394, 'loss/train': 2.1604795455932617} 02/24/2022 21:30:14 - INFO - codeparrot_training - Step 15395: {'lr': 0.0004099226426594657, 'samples': 7882752, 'steps': 15395, 'loss/train': 1.7626302242279053} 02/24/2022 21:30:17 - INFO - codeparrot_training - Step 15396: {'lr': 0.0004099100655965019, 'samples': 7883264, 'steps': 15396, 'loss/train': 2.0931098461151123} 02/24/2022 21:30:23 - INFO - codeparrot_training - Step 15397: {'lr': 0.0004098974878485342, 'samples': 7883776, 'steps': 15397, 'loss/train': 1.9197818040847778} 02/24/2022 21:30:26 - INFO - codeparrot_training - Step 15398: {'lr': 0.0004098849094156164, 'samples': 7884288, 'steps': 15398, 'loss/train': 1.7820371389389038} 02/24/2022 21:30:32 - INFO - codeparrot_training - Step 15399: {'lr': 0.0004098723302978025, 'samples': 7884800, 'steps': 15399, 'loss/train': 1.904728651046753} 02/24/2022 21:30:36 - INFO - codeparrot_training - Step 15400: {'lr': 0.00040985975049514617, 'samples': 7885312, 'steps': 15400, 'loss/train': 1.2827519178390503} 02/24/2022 21:30:41 - INFO - codeparrot_training - Step 15401: {'lr': 0.00040984717000770157, 'samples': 7885824, 'steps': 15401, 'loss/train': 1.6896698474884033} 02/24/2022 21:30:45 - INFO - codeparrot_training - Step 15402: {'lr': 0.00040983458883552237, 'samples': 7886336, 'steps': 15402, 'loss/train': 1.2777215242385864} 02/24/2022 21:30:51 - INFO - codeparrot_training - Step 15403: {'lr': 0.00040982200697866256, 'samples': 7886848, 'steps': 15403, 'loss/train': 2.300790548324585} 02/24/2022 21:30:55 - INFO - codeparrot_training - Step 15404: {'lr': 0.00040980942443717596, 'samples': 7887360, 'steps': 15404, 'loss/train': 2.621476888656616} 02/24/2022 21:31:00 - INFO - codeparrot_training - Step 15405: {'lr': 0.0004097968412111166, 'samples': 7887872, 'steps': 15405, 'loss/train': 1.7218024730682373} 02/24/2022 21:31:04 - INFO - codeparrot_training - Step 15406: {'lr': 0.0004097842573005383, 'samples': 7888384, 'steps': 15406, 'loss/train': 2.013035297393799} 02/24/2022 21:31:10 - INFO - codeparrot_training - Step 15407: {'lr': 0.000409771672705495, 'samples': 7888896, 'steps': 15407, 'loss/train': 2.40028977394104} 02/24/2022 21:31:13 - INFO - codeparrot_training - Step 15408: {'lr': 0.0004097590874260405, 'samples': 7889408, 'steps': 15408, 'loss/train': 2.3176109790802} 02/24/2022 21:31:17 - INFO - codeparrot_training - Step 15409: {'lr': 0.0004097465014622289, 'samples': 7889920, 'steps': 15409, 'loss/train': 2.6842987537384033} 02/24/2022 21:31:22 - INFO - codeparrot_training - Step 15410: {'lr': 0.00040973391481411396, 'samples': 7890432, 'steps': 15410, 'loss/train': 2.919426918029785} 02/24/2022 21:31:26 - INFO - codeparrot_training - Step 15411: {'lr': 0.00040972132748174966, 'samples': 7890944, 'steps': 15411, 'loss/train': 2.180241346359253} 02/24/2022 21:31:31 - INFO - codeparrot_training - Step 15412: {'lr': 0.00040970873946518993, 'samples': 7891456, 'steps': 15412, 'loss/train': 2.3224427700042725} 02/24/2022 21:31:35 - INFO - codeparrot_training - Step 15413: {'lr': 0.00040969615076448865, 'samples': 7891968, 'steps': 15413, 'loss/train': 1.9502239227294922} 02/24/2022 21:31:40 - INFO - codeparrot_training - Step 15414: {'lr': 0.0004096835613796998, 'samples': 7892480, 'steps': 15414, 'loss/train': 1.4000823497772217} 02/24/2022 21:31:44 - INFO - codeparrot_training - Step 15415: {'lr': 0.00040967097131087727, 'samples': 7892992, 'steps': 15415, 'loss/train': 1.8175748586654663} 02/24/2022 21:31:49 - INFO - codeparrot_training - Step 15416: {'lr': 0.00040965838055807493, 'samples': 7893504, 'steps': 15416, 'loss/train': 2.2393364906311035} 02/24/2022 21:31:53 - INFO - codeparrot_training - Step 15417: {'lr': 0.00040964578912134687, 'samples': 7894016, 'steps': 15417, 'loss/train': 1.8740324974060059} 02/24/2022 21:31:58 - INFO - codeparrot_training - Step 15418: {'lr': 0.00040963319700074684, 'samples': 7894528, 'steps': 15418, 'loss/train': 1.4016344547271729} 02/24/2022 21:32:02 - INFO - codeparrot_training - Step 15419: {'lr': 0.00040962060419632906, 'samples': 7895040, 'steps': 15419, 'loss/train': 2.2686307430267334} 02/24/2022 21:32:08 - INFO - codeparrot_training - Step 15420: {'lr': 0.00040960801070814715, 'samples': 7895552, 'steps': 15420, 'loss/train': 1.1163126230239868} 02/24/2022 21:32:11 - INFO - codeparrot_training - Step 15421: {'lr': 0.00040959541653625526, 'samples': 7896064, 'steps': 15421, 'loss/train': 2.352973699569702} 02/24/2022 21:32:17 - INFO - codeparrot_training - Step 15422: {'lr': 0.0004095828216807073, 'samples': 7896576, 'steps': 15422, 'loss/train': 1.7794021368026733} 02/24/2022 21:32:20 - INFO - codeparrot_training - Step 15423: {'lr': 0.00040957022614155714, 'samples': 7897088, 'steps': 15423, 'loss/train': 2.371039628982544} 02/24/2022 21:32:26 - INFO - codeparrot_training - Step 15424: {'lr': 0.0004095576299188589, 'samples': 7897600, 'steps': 15424, 'loss/train': 2.479933738708496} 02/24/2022 21:32:29 - INFO - codeparrot_training - Step 15425: {'lr': 0.0004095450330126663, 'samples': 7898112, 'steps': 15425, 'loss/train': 1.5423216819763184} 02/24/2022 21:32:35 - INFO - codeparrot_training - Step 15426: {'lr': 0.0004095324354230335, 'samples': 7898624, 'steps': 15426, 'loss/train': 1.5639524459838867} 02/24/2022 21:32:38 - INFO - codeparrot_training - Step 15427: {'lr': 0.0004095198371500145, 'samples': 7899136, 'steps': 15427, 'loss/train': 2.206367015838623} 02/24/2022 21:32:44 - INFO - codeparrot_training - Step 15428: {'lr': 0.00040950723819366307, 'samples': 7899648, 'steps': 15428, 'loss/train': 1.9861165285110474} 02/24/2022 21:32:48 - INFO - codeparrot_training - Step 15429: {'lr': 0.00040949463855403326, 'samples': 7900160, 'steps': 15429, 'loss/train': 2.516212224960327} 02/24/2022 21:32:54 - INFO - codeparrot_training - Step 15430: {'lr': 0.00040948203823117915, 'samples': 7900672, 'steps': 15430, 'loss/train': 1.336027979850769} 02/24/2022 21:32:57 - INFO - codeparrot_training - Step 15431: {'lr': 0.00040946943722515455, 'samples': 7901184, 'steps': 15431, 'loss/train': 1.6456513404846191} 02/24/2022 21:33:03 - INFO - codeparrot_training - Step 15432: {'lr': 0.0004094568355360135, 'samples': 7901696, 'steps': 15432, 'loss/train': 3.6719956398010254} 02/24/2022 21:33:06 - INFO - codeparrot_training - Step 15433: {'lr': 0.00040944423316381006, 'samples': 7902208, 'steps': 15433, 'loss/train': 1.9775714874267578} 02/24/2022 21:33:12 - INFO - codeparrot_training - Step 15434: {'lr': 0.0004094316301085982, 'samples': 7902720, 'steps': 15434, 'loss/train': 0.892765998840332} 02/24/2022 21:33:15 - INFO - codeparrot_training - Step 15435: {'lr': 0.00040941902637043183, 'samples': 7903232, 'steps': 15435, 'loss/train': 2.8887033462524414} 02/24/2022 21:33:21 - INFO - codeparrot_training - Step 15436: {'lr': 0.00040940642194936495, 'samples': 7903744, 'steps': 15436, 'loss/train': 2.3517863750457764} 02/24/2022 21:33:25 - INFO - codeparrot_training - Step 15437: {'lr': 0.0004093938168454515, 'samples': 7904256, 'steps': 15437, 'loss/train': 1.8546351194381714} 02/24/2022 21:33:30 - INFO - codeparrot_training - Step 15438: {'lr': 0.00040938121105874573, 'samples': 7904768, 'steps': 15438, 'loss/train': 2.90852952003479} 02/24/2022 21:33:33 - INFO - codeparrot_training - Step 15439: {'lr': 0.0004093686045893013, 'samples': 7905280, 'steps': 15439, 'loss/train': 2.670231342315674} 02/24/2022 21:33:40 - INFO - codeparrot_training - Step 15440: {'lr': 0.00040935599743717243, 'samples': 7905792, 'steps': 15440, 'loss/train': 2.0736353397369385} 02/24/2022 21:33:43 - INFO - codeparrot_training - Step 15441: {'lr': 0.00040934338960241305, 'samples': 7906304, 'steps': 15441, 'loss/train': 2.002852439880371} 02/24/2022 21:33:49 - INFO - codeparrot_training - Step 15442: {'lr': 0.00040933078108507727, 'samples': 7906816, 'steps': 15442, 'loss/train': 2.8125252723693848} 02/24/2022 21:33:54 - INFO - codeparrot_training - Step 15443: {'lr': 0.00040931817188521894, 'samples': 7907328, 'steps': 15443, 'loss/train': 2.2092647552490234} 02/24/2022 21:33:58 - INFO - codeparrot_training - Step 15444: {'lr': 0.00040930556200289214, 'samples': 7907840, 'steps': 15444, 'loss/train': 0.7220437526702881} 02/24/2022 21:34:03 - INFO - codeparrot_training - Step 15445: {'lr': 0.00040929295143815093, 'samples': 7908352, 'steps': 15445, 'loss/train': 1.5290563106536865} 02/24/2022 21:34:07 - INFO - codeparrot_training - Step 15446: {'lr': 0.0004092803401910493, 'samples': 7908864, 'steps': 15446, 'loss/train': 1.9819129705429077} 02/24/2022 21:34:12 - INFO - codeparrot_training - Step 15447: {'lr': 0.00040926772826164126, 'samples': 7909376, 'steps': 15447, 'loss/train': 2.1988632678985596} 02/24/2022 21:34:16 - INFO - codeparrot_training - Step 15448: {'lr': 0.0004092551156499809, 'samples': 7909888, 'steps': 15448, 'loss/train': 2.420659065246582} 02/24/2022 21:34:21 - INFO - codeparrot_training - Step 15449: {'lr': 0.000409242502356122, 'samples': 7910400, 'steps': 15449, 'loss/train': 2.5587782859802246} 02/24/2022 21:34:25 - INFO - codeparrot_training - Step 15450: {'lr': 0.000409229888380119, 'samples': 7910912, 'steps': 15450, 'loss/train': 2.089151382446289} 02/24/2022 21:34:31 - INFO - codeparrot_training - Step 15451: {'lr': 0.00040921727372202565, 'samples': 7911424, 'steps': 15451, 'loss/train': 1.9240504503250122} 02/24/2022 21:34:34 - INFO - codeparrot_training - Step 15452: {'lr': 0.000409204658381896, 'samples': 7911936, 'steps': 15452, 'loss/train': 2.5743579864501953} 02/24/2022 21:34:40 - INFO - codeparrot_training - Step 15453: {'lr': 0.00040919204235978425, 'samples': 7912448, 'steps': 15453, 'loss/train': 2.241792678833008} 02/24/2022 21:34:43 - INFO - codeparrot_training - Step 15454: {'lr': 0.0004091794256557443, 'samples': 7912960, 'steps': 15454, 'loss/train': 2.547938585281372} 02/24/2022 21:34:47 - INFO - codeparrot_training - Step 15455: {'lr': 0.00040916680826983017, 'samples': 7913472, 'steps': 15455, 'loss/train': 2.1530346870422363} 02/24/2022 21:34:53 - INFO - codeparrot_training - Step 15456: {'lr': 0.00040915419020209605, 'samples': 7913984, 'steps': 15456, 'loss/train': 2.095562696456909} 02/24/2022 21:34:56 - INFO - codeparrot_training - Step 15457: {'lr': 0.0004091415714525959, 'samples': 7914496, 'steps': 15457, 'loss/train': 1.792824149131775} 02/24/2022 21:35:02 - INFO - codeparrot_training - Step 15458: {'lr': 0.0004091289520213838, 'samples': 7915008, 'steps': 15458, 'loss/train': 2.257463216781616} 02/24/2022 21:35:05 - INFO - codeparrot_training - Step 15459: {'lr': 0.0004091163319085137, 'samples': 7915520, 'steps': 15459, 'loss/train': 0.46837449073791504} 02/24/2022 21:35:13 - INFO - codeparrot_training - Step 15460: {'lr': 0.0004091037111140399, 'samples': 7916032, 'steps': 15460, 'loss/train': 2.1919593811035156} 02/24/2022 21:35:17 - INFO - codeparrot_training - Step 15461: {'lr': 0.00040909108963801624, 'samples': 7916544, 'steps': 15461, 'loss/train': 1.4890090227127075} 02/24/2022 21:35:22 - INFO - codeparrot_training - Step 15462: {'lr': 0.0004090784674804969, 'samples': 7917056, 'steps': 15462, 'loss/train': 1.857080340385437} 02/24/2022 21:35:26 - INFO - codeparrot_training - Step 15463: {'lr': 0.0004090658446415359, 'samples': 7917568, 'steps': 15463, 'loss/train': 1.6178237199783325} 02/24/2022 21:35:31 - INFO - codeparrot_training - Step 15464: {'lr': 0.0004090532211211874, 'samples': 7918080, 'steps': 15464, 'loss/train': 2.5663673877716064} 02/24/2022 21:35:37 - INFO - codeparrot_training - Step 15465: {'lr': 0.0004090405969195053, 'samples': 7918592, 'steps': 15465, 'loss/train': 1.9339476823806763} 02/24/2022 21:35:40 - INFO - codeparrot_training - Step 15466: {'lr': 0.0004090279720365438, 'samples': 7919104, 'steps': 15466, 'loss/train': 1.789537787437439} 02/24/2022 21:35:46 - INFO - codeparrot_training - Step 15467: {'lr': 0.00040901534647235703, 'samples': 7919616, 'steps': 15467, 'loss/train': 1.8607381582260132} 02/24/2022 21:35:50 - INFO - codeparrot_training - Step 15468: {'lr': 0.00040900272022699897, 'samples': 7920128, 'steps': 15468, 'loss/train': 1.4169374704360962} 02/24/2022 21:35:53 - INFO - codeparrot_training - Step 15469: {'lr': 0.00040899009330052375, 'samples': 7920640, 'steps': 15469, 'loss/train': 2.213528633117676} 02/24/2022 21:36:00 - INFO - codeparrot_training - Step 15470: {'lr': 0.00040897746569298546, 'samples': 7921152, 'steps': 15470, 'loss/train': 1.7157049179077148} 02/24/2022 21:36:04 - INFO - codeparrot_training - Step 15471: {'lr': 0.0004089648374044382, 'samples': 7921664, 'steps': 15471, 'loss/train': 1.051945686340332} 02/24/2022 21:36:09 - INFO - codeparrot_training - Step 15472: {'lr': 0.000408952208434936, 'samples': 7922176, 'steps': 15472, 'loss/train': 1.2381268739700317} 02/24/2022 21:36:13 - INFO - codeparrot_training - Step 15473: {'lr': 0.00040893957878453314, 'samples': 7922688, 'steps': 15473, 'loss/train': 2.5180225372314453} 02/24/2022 21:36:18 - INFO - codeparrot_training - Step 15474: {'lr': 0.0004089269484532834, 'samples': 7923200, 'steps': 15474, 'loss/train': 2.0026073455810547} 02/24/2022 21:36:22 - INFO - codeparrot_training - Step 15475: {'lr': 0.00040891431744124123, 'samples': 7923712, 'steps': 15475, 'loss/train': 2.1056971549987793} 02/24/2022 21:36:27 - INFO - codeparrot_training - Step 15476: {'lr': 0.00040890168574846055, 'samples': 7924224, 'steps': 15476, 'loss/train': 1.2067630290985107} 02/24/2022 21:36:31 - INFO - codeparrot_training - Step 15477: {'lr': 0.0004088890533749955, 'samples': 7924736, 'steps': 15477, 'loss/train': 2.467292308807373} 02/24/2022 21:36:36 - INFO - codeparrot_training - Step 15478: {'lr': 0.0004088764203209002, 'samples': 7925248, 'steps': 15478, 'loss/train': 2.263564348220825} 02/24/2022 21:36:40 - INFO - codeparrot_training - Step 15479: {'lr': 0.0004088637865862287, 'samples': 7925760, 'steps': 15479, 'loss/train': 2.318178653717041} 02/24/2022 21:36:47 - INFO - codeparrot_training - Step 15480: {'lr': 0.0004088511521710352, 'samples': 7926272, 'steps': 15480, 'loss/train': 2.8448283672332764} 02/24/2022 21:36:51 - INFO - codeparrot_training - Step 15481: {'lr': 0.0004088385170753739, 'samples': 7926784, 'steps': 15481, 'loss/train': 1.6736156940460205} 02/24/2022 21:36:56 - INFO - codeparrot_training - Step 15482: {'lr': 0.00040882588129929876, 'samples': 7927296, 'steps': 15482, 'loss/train': 1.9599905014038086} 02/24/2022 21:37:00 - INFO - codeparrot_training - Step 15483: {'lr': 0.000408813244842864, 'samples': 7927808, 'steps': 15483, 'loss/train': 2.5051231384277344} 02/24/2022 21:37:05 - INFO - codeparrot_training - Step 15484: {'lr': 0.0004088006077061237, 'samples': 7928320, 'steps': 15484, 'loss/train': 2.259176254272461} 02/24/2022 21:37:09 - INFO - codeparrot_training - Step 15485: {'lr': 0.00040878796988913204, 'samples': 7928832, 'steps': 15485, 'loss/train': 1.6392414569854736} 02/24/2022 21:37:14 - INFO - codeparrot_training - Step 15486: {'lr': 0.00040877533139194313, 'samples': 7929344, 'steps': 15486, 'loss/train': 2.424464225769043} 02/24/2022 21:37:18 - INFO - codeparrot_training - Step 15487: {'lr': 0.00040876269221461117, 'samples': 7929856, 'steps': 15487, 'loss/train': 2.7809624671936035} 02/24/2022 21:37:23 - INFO - codeparrot_training - Step 15488: {'lr': 0.0004087500523571902, 'samples': 7930368, 'steps': 15488, 'loss/train': 0.3770766854286194} 02/24/2022 21:37:27 - INFO - codeparrot_training - Step 15489: {'lr': 0.0004087374118197344, 'samples': 7930880, 'steps': 15489, 'loss/train': 1.3468271493911743} 02/24/2022 21:37:34 - INFO - codeparrot_training - Step 15490: {'lr': 0.00040872477060229797, 'samples': 7931392, 'steps': 15490, 'loss/train': 1.9579812288284302} 02/24/2022 21:37:37 - INFO - codeparrot_training - Step 15491: {'lr': 0.00040871212870493504, 'samples': 7931904, 'steps': 15491, 'loss/train': 2.244948387145996} 02/24/2022 21:37:43 - INFO - codeparrot_training - Step 15492: {'lr': 0.0004086994861276996, 'samples': 7932416, 'steps': 15492, 'loss/train': 1.9149377346038818} 02/24/2022 21:37:48 - INFO - codeparrot_training - Step 15493: {'lr': 0.00040868684287064617, 'samples': 7932928, 'steps': 15493, 'loss/train': 2.311769962310791} 02/24/2022 21:37:52 - INFO - codeparrot_training - Step 15494: {'lr': 0.0004086741989338285, 'samples': 7933440, 'steps': 15494, 'loss/train': 2.153244733810425} 02/24/2022 21:37:57 - INFO - codeparrot_training - Step 15495: {'lr': 0.0004086615543173011, 'samples': 7933952, 'steps': 15495, 'loss/train': 0.1000339463353157} 02/24/2022 21:38:01 - INFO - codeparrot_training - Step 15496: {'lr': 0.0004086489090211178, 'samples': 7934464, 'steps': 15496, 'loss/train': 2.5734522342681885} 02/24/2022 21:38:06 - INFO - codeparrot_training - Step 15497: {'lr': 0.00040863626304533316, 'samples': 7934976, 'steps': 15497, 'loss/train': 1.5227742195129395} 02/24/2022 21:38:10 - INFO - codeparrot_training - Step 15498: {'lr': 0.000408623616390001, 'samples': 7935488, 'steps': 15498, 'loss/train': 2.099524736404419} 02/24/2022 21:38:16 - INFO - codeparrot_training - Step 15499: {'lr': 0.00040861096905517574, 'samples': 7936000, 'steps': 15499, 'loss/train': 1.5269643068313599} 02/24/2022 21:38:19 - INFO - codeparrot_training - Step 15500: {'lr': 0.0004085983210409114, 'samples': 7936512, 'steps': 15500, 'loss/train': 3.403447389602661} 02/24/2022 21:38:22 - INFO - codeparrot_training - Step 15501: {'lr': 0.00040858567234726217, 'samples': 7937024, 'steps': 15501, 'loss/train': 1.84128999710083} 02/24/2022 21:38:28 - INFO - codeparrot_training - Step 15502: {'lr': 0.00040857302297428233, 'samples': 7937536, 'steps': 15502, 'loss/train': 1.6496556997299194} 02/24/2022 21:38:33 - INFO - codeparrot_training - Step 15503: {'lr': 0.000408560372922026, 'samples': 7938048, 'steps': 15503, 'loss/train': 2.2017412185668945} 02/24/2022 21:38:37 - INFO - codeparrot_training - Step 15504: {'lr': 0.00040854772219054737, 'samples': 7938560, 'steps': 15504, 'loss/train': 1.757104516029358} 02/24/2022 21:38:44 - INFO - codeparrot_training - Step 15505: {'lr': 0.00040853507077990073, 'samples': 7939072, 'steps': 15505, 'loss/train': 1.9060847759246826} 02/24/2022 21:38:48 - INFO - codeparrot_training - Step 15506: {'lr': 0.00040852241869014004, 'samples': 7939584, 'steps': 15506, 'loss/train': 1.3947877883911133} 02/24/2022 21:38:53 - INFO - codeparrot_training - Step 15507: {'lr': 0.00040850976592131974, 'samples': 7940096, 'steps': 15507, 'loss/train': 3.2720694541931152} 02/24/2022 21:38:57 - INFO - codeparrot_training - Step 15508: {'lr': 0.0004084971124734939, 'samples': 7940608, 'steps': 15508, 'loss/train': 2.748504877090454} 02/24/2022 21:39:02 - INFO - codeparrot_training - Step 15509: {'lr': 0.0004084844583467168, 'samples': 7941120, 'steps': 15509, 'loss/train': 2.437459707260132} 02/24/2022 21:39:06 - INFO - codeparrot_training - Step 15510: {'lr': 0.00040847180354104256, 'samples': 7941632, 'steps': 15510, 'loss/train': 2.965855360031128} 02/24/2022 21:39:11 - INFO - codeparrot_training - Step 15511: {'lr': 0.00040845914805652544, 'samples': 7942144, 'steps': 15511, 'loss/train': 1.7638673782348633} 02/24/2022 21:39:15 - INFO - codeparrot_training - Step 15512: {'lr': 0.0004084464918932197, 'samples': 7942656, 'steps': 15512, 'loss/train': 2.1305906772613525} 02/24/2022 21:39:20 - INFO - codeparrot_training - Step 15513: {'lr': 0.0004084338350511795, 'samples': 7943168, 'steps': 15513, 'loss/train': 2.331066846847534} 02/24/2022 21:39:24 - INFO - codeparrot_training - Step 15514: {'lr': 0.00040842117753045893, 'samples': 7943680, 'steps': 15514, 'loss/train': 1.5385148525238037} 02/24/2022 21:39:31 - INFO - codeparrot_training - Step 15515: {'lr': 0.0004084085193311124, 'samples': 7944192, 'steps': 15515, 'loss/train': 2.8428070545196533} 02/24/2022 21:39:34 - INFO - codeparrot_training - Step 15516: {'lr': 0.0004083958604531941, 'samples': 7944704, 'steps': 15516, 'loss/train': 1.7632235288619995} 02/24/2022 21:39:40 - INFO - codeparrot_training - Step 15517: {'lr': 0.0004083832008967583, 'samples': 7945216, 'steps': 15517, 'loss/train': 1.6807687282562256} 02/24/2022 21:39:43 - INFO - codeparrot_training - Step 15518: {'lr': 0.00040837054066185906, 'samples': 7945728, 'steps': 15518, 'loss/train': 1.4972398281097412} 02/24/2022 21:39:49 - INFO - codeparrot_training - Step 15519: {'lr': 0.0004083578797485508, 'samples': 7946240, 'steps': 15519, 'loss/train': 1.4638450145721436} 02/24/2022 21:39:52 - INFO - codeparrot_training - Step 15520: {'lr': 0.00040834521815688753, 'samples': 7946752, 'steps': 15520, 'loss/train': 2.972620964050293} 02/24/2022 21:39:58 - INFO - codeparrot_training - Step 15521: {'lr': 0.00040833255588692375, 'samples': 7947264, 'steps': 15521, 'loss/train': 2.0566892623901367} 02/24/2022 21:40:01 - INFO - codeparrot_training - Step 15522: {'lr': 0.0004083198929387135, 'samples': 7947776, 'steps': 15522, 'loss/train': 3.4358222484588623} 02/24/2022 21:40:07 - INFO - codeparrot_training - Step 15523: {'lr': 0.0004083072293123111, 'samples': 7948288, 'steps': 15523, 'loss/train': 1.9729145765304565} 02/24/2022 21:40:10 - INFO - codeparrot_training - Step 15524: {'lr': 0.00040829456500777084, 'samples': 7948800, 'steps': 15524, 'loss/train': 2.6530888080596924} 02/24/2022 21:40:16 - INFO - codeparrot_training - Step 15525: {'lr': 0.00040828190002514694, 'samples': 7949312, 'steps': 15525, 'loss/train': 1.9839402437210083} 02/24/2022 21:40:19 - INFO - codeparrot_training - Step 15526: {'lr': 0.0004082692343644936, 'samples': 7949824, 'steps': 15526, 'loss/train': 2.1632790565490723} 02/24/2022 21:40:27 - INFO - codeparrot_training - Step 15527: {'lr': 0.00040825656802586513, 'samples': 7950336, 'steps': 15527, 'loss/train': 2.0929946899414062} 02/24/2022 21:40:30 - INFO - codeparrot_training - Step 15528: {'lr': 0.00040824390100931585, 'samples': 7950848, 'steps': 15528, 'loss/train': 1.8413575887680054} 02/24/2022 21:40:36 - INFO - codeparrot_training - Step 15529: {'lr': 0.00040823123331489985, 'samples': 7951360, 'steps': 15529, 'loss/train': 2.7376160621643066} 02/24/2022 21:40:39 - INFO - codeparrot_training - Step 15530: {'lr': 0.0004082185649426715, 'samples': 7951872, 'steps': 15530, 'loss/train': 2.369424819946289} 02/24/2022 21:40:45 - INFO - codeparrot_training - Step 15531: {'lr': 0.0004082058958926851, 'samples': 7952384, 'steps': 15531, 'loss/train': 1.0001769065856934} 02/24/2022 21:40:48 - INFO - codeparrot_training - Step 15532: {'lr': 0.0004081932261649949, 'samples': 7952896, 'steps': 15532, 'loss/train': 2.2717673778533936} 02/24/2022 21:40:54 - INFO - codeparrot_training - Step 15533: {'lr': 0.00040818055575965505, 'samples': 7953408, 'steps': 15533, 'loss/train': 1.7797858715057373} 02/24/2022 21:40:57 - INFO - codeparrot_training - Step 15534: {'lr': 0.0004081678846767199, 'samples': 7953920, 'steps': 15534, 'loss/train': 3.093263626098633} 02/24/2022 21:41:03 - INFO - codeparrot_training - Step 15535: {'lr': 0.00040815521291624393, 'samples': 7954432, 'steps': 15535, 'loss/train': 1.6688371896743774} 02/24/2022 21:41:06 - INFO - codeparrot_training - Step 15536: {'lr': 0.0004081425404782811, 'samples': 7954944, 'steps': 15536, 'loss/train': 1.3239980936050415} 02/24/2022 21:41:14 - INFO - codeparrot_training - Step 15537: {'lr': 0.0004081298673628859, 'samples': 7955456, 'steps': 15537, 'loss/train': 2.282769203186035} 02/24/2022 21:41:17 - INFO - codeparrot_training - Step 15538: {'lr': 0.00040811719357011257, 'samples': 7955968, 'steps': 15538, 'loss/train': 1.4894928932189941} 02/24/2022 21:41:23 - INFO - codeparrot_training - Step 15539: {'lr': 0.00040810451910001537, 'samples': 7956480, 'steps': 15539, 'loss/train': 0.8797029852867126} 02/24/2022 21:41:26 - INFO - codeparrot_training - Step 15540: {'lr': 0.00040809184395264867, 'samples': 7956992, 'steps': 15540, 'loss/train': 2.388523578643799} 02/24/2022 21:41:32 - INFO - codeparrot_training - Step 15541: {'lr': 0.0004080791681280667, 'samples': 7957504, 'steps': 15541, 'loss/train': 1.1261372566223145} 02/24/2022 21:41:35 - INFO - codeparrot_training - Step 15542: {'lr': 0.00040806649162632364, 'samples': 7958016, 'steps': 15542, 'loss/train': 1.809226632118225} 02/24/2022 21:41:41 - INFO - codeparrot_training - Step 15543: {'lr': 0.000408053814447474, 'samples': 7958528, 'steps': 15543, 'loss/train': 1.465032935142517} 02/24/2022 21:41:44 - INFO - codeparrot_training - Step 15544: {'lr': 0.00040804113659157203, 'samples': 7959040, 'steps': 15544, 'loss/train': 2.1486222743988037} 02/24/2022 21:41:50 - INFO - codeparrot_training - Step 15545: {'lr': 0.00040802845805867205, 'samples': 7959552, 'steps': 15545, 'loss/train': 2.0902352333068848} 02/24/2022 21:41:53 - INFO - codeparrot_training - Step 15546: {'lr': 0.0004080157788488282, 'samples': 7960064, 'steps': 15546, 'loss/train': 2.96724271774292} 02/24/2022 21:41:59 - INFO - codeparrot_training - Step 15547: {'lr': 0.0004080030989620951, 'samples': 7960576, 'steps': 15547, 'loss/train': 2.1055681705474854} 02/24/2022 21:42:02 - INFO - codeparrot_training - Step 15548: {'lr': 0.0004079904183985268, 'samples': 7961088, 'steps': 15548, 'loss/train': 1.768223762512207} 02/24/2022 21:42:08 - INFO - codeparrot_training - Step 15549: {'lr': 0.0004079777371581777, 'samples': 7961600, 'steps': 15549, 'loss/train': 1.4073724746704102} 02/24/2022 21:42:11 - INFO - codeparrot_training - Step 15550: {'lr': 0.00040796505524110215, 'samples': 7962112, 'steps': 15550, 'loss/train': 2.4263346195220947} 02/24/2022 21:42:19 - INFO - codeparrot_training - Step 15551: {'lr': 0.00040795237264735454, 'samples': 7962624, 'steps': 15551, 'loss/train': 2.130971670150757} 02/24/2022 21:42:22 - INFO - codeparrot_training - Step 15552: {'lr': 0.00040793968937698905, 'samples': 7963136, 'steps': 15552, 'loss/train': 1.4335674047470093} 02/24/2022 21:42:28 - INFO - codeparrot_training - Step 15553: {'lr': 0.00040792700543006014, 'samples': 7963648, 'steps': 15553, 'loss/train': 2.335895538330078} 02/24/2022 21:42:31 - INFO - codeparrot_training - Step 15554: {'lr': 0.000407914320806622, 'samples': 7964160, 'steps': 15554, 'loss/train': 1.5521836280822754} 02/24/2022 21:42:37 - INFO - codeparrot_training - Step 15555: {'lr': 0.0004079016355067291, 'samples': 7964672, 'steps': 15555, 'loss/train': 2.253251552581787} 02/24/2022 21:42:40 - INFO - codeparrot_training - Step 15556: {'lr': 0.0004078889495304357, 'samples': 7965184, 'steps': 15556, 'loss/train': 1.2411023378372192} 02/24/2022 21:42:46 - INFO - codeparrot_training - Step 15557: {'lr': 0.00040787626287779624, 'samples': 7965696, 'steps': 15557, 'loss/train': 1.9537452459335327} 02/24/2022 21:42:49 - INFO - codeparrot_training - Step 15558: {'lr': 0.0004078635755488649, 'samples': 7966208, 'steps': 15558, 'loss/train': 2.891079902648926} 02/24/2022 21:42:55 - INFO - codeparrot_training - Step 15559: {'lr': 0.00040785088754369627, 'samples': 7966720, 'steps': 15559, 'loss/train': 2.240905523300171} 02/24/2022 21:42:58 - INFO - codeparrot_training - Step 15560: {'lr': 0.00040783819886234445, 'samples': 7967232, 'steps': 15560, 'loss/train': 2.200089693069458} 02/24/2022 21:43:06 - INFO - codeparrot_training - Step 15561: {'lr': 0.000407825509504864, 'samples': 7967744, 'steps': 15561, 'loss/train': 0.8187501430511475} 02/24/2022 21:43:09 - INFO - codeparrot_training - Step 15562: {'lr': 0.00040781281947130897, 'samples': 7968256, 'steps': 15562, 'loss/train': 2.1469154357910156} 02/24/2022 21:43:15 - INFO - codeparrot_training - Step 15563: {'lr': 0.0004078001287617342, 'samples': 7968768, 'steps': 15563, 'loss/train': 2.6158902645111084} 02/24/2022 21:43:19 - INFO - codeparrot_training - Step 15564: {'lr': 0.0004077874373761936, 'samples': 7969280, 'steps': 15564, 'loss/train': 2.2594118118286133} 02/24/2022 21:43:24 - INFO - codeparrot_training - Step 15565: {'lr': 0.0004077747453147418, 'samples': 7969792, 'steps': 15565, 'loss/train': 3.2596192359924316} 02/24/2022 21:43:28 - INFO - codeparrot_training - Step 15566: {'lr': 0.0004077620525774331, 'samples': 7970304, 'steps': 15566, 'loss/train': 2.3571548461914062} 02/24/2022 21:43:31 - INFO - codeparrot_training - Step 15567: {'lr': 0.0004077493591643219, 'samples': 7970816, 'steps': 15567, 'loss/train': 1.845572590827942} 02/24/2022 21:43:37 - INFO - codeparrot_training - Step 15568: {'lr': 0.00040773666507546244, 'samples': 7971328, 'steps': 15568, 'loss/train': 2.420595169067383} 02/24/2022 21:43:40 - INFO - codeparrot_training - Step 15569: {'lr': 0.00040772397031090923, 'samples': 7971840, 'steps': 15569, 'loss/train': 2.204481840133667} 02/24/2022 21:43:46 - INFO - codeparrot_training - Step 15570: {'lr': 0.0004077112748707166, 'samples': 7972352, 'steps': 15570, 'loss/train': 1.5307304859161377} 02/24/2022 21:43:49 - INFO - codeparrot_training - Step 15571: {'lr': 0.000407698578754939, 'samples': 7972864, 'steps': 15571, 'loss/train': 1.8535393476486206} 02/24/2022 21:43:57 - INFO - codeparrot_training - Step 15572: {'lr': 0.0004076858819636307, 'samples': 7973376, 'steps': 15572, 'loss/train': 1.7659270763397217} 02/24/2022 21:44:02 - INFO - codeparrot_training - Step 15573: {'lr': 0.0004076731844968462, 'samples': 7973888, 'steps': 15573, 'loss/train': 2.3016152381896973} 02/24/2022 21:44:06 - INFO - codeparrot_training - Step 15574: {'lr': 0.00040766048635463984, 'samples': 7974400, 'steps': 15574, 'loss/train': 2.24153995513916} 02/24/2022 21:44:09 - INFO - codeparrot_training - Step 15575: {'lr': 0.000407647787537066, 'samples': 7974912, 'steps': 15575, 'loss/train': 2.6576950550079346} 02/24/2022 21:44:14 - INFO - codeparrot_training - Step 15576: {'lr': 0.00040763508804417904, 'samples': 7975424, 'steps': 15576, 'loss/train': 1.8348135948181152} 02/24/2022 21:44:20 - INFO - codeparrot_training - Step 15577: {'lr': 0.0004076223878760335, 'samples': 7975936, 'steps': 15577, 'loss/train': 1.6207209825515747} 02/24/2022 21:44:24 - INFO - codeparrot_training - Step 15578: {'lr': 0.0004076096870326837, 'samples': 7976448, 'steps': 15578, 'loss/train': 2.6876683235168457} 02/24/2022 21:44:29 - INFO - codeparrot_training - Step 15579: {'lr': 0.000407596985514184, 'samples': 7976960, 'steps': 15579, 'loss/train': 1.0152466297149658} 02/24/2022 21:44:33 - INFO - codeparrot_training - Step 15580: {'lr': 0.00040758428332058895, 'samples': 7977472, 'steps': 15580, 'loss/train': 2.602926731109619} 02/24/2022 21:44:38 - INFO - codeparrot_training - Step 15581: {'lr': 0.00040757158045195274, 'samples': 7977984, 'steps': 15581, 'loss/train': 2.8576900959014893} 02/24/2022 21:44:42 - INFO - codeparrot_training - Step 15582: {'lr': 0.00040755887690833005, 'samples': 7978496, 'steps': 15582, 'loss/train': 2.028103828430176} 02/24/2022 21:44:47 - INFO - codeparrot_training - Step 15583: {'lr': 0.00040754617268977503, 'samples': 7979008, 'steps': 15583, 'loss/train': 1.3518954515457153} 02/24/2022 21:44:51 - INFO - codeparrot_training - Step 15584: {'lr': 0.0004075334677963423, 'samples': 7979520, 'steps': 15584, 'loss/train': 2.0328445434570312} 02/24/2022 21:44:56 - INFO - codeparrot_training - Step 15585: {'lr': 0.00040752076222808623, 'samples': 7980032, 'steps': 15585, 'loss/train': 0.06505563855171204} 02/24/2022 21:45:00 - INFO - codeparrot_training - Step 15586: {'lr': 0.00040750805598506115, 'samples': 7980544, 'steps': 15586, 'loss/train': 2.275618076324463} 02/24/2022 21:45:07 - INFO - codeparrot_training - Step 15587: {'lr': 0.00040749534906732167, 'samples': 7981056, 'steps': 15587, 'loss/train': 1.9834320545196533} 02/24/2022 21:45:11 - INFO - codeparrot_training - Step 15588: {'lr': 0.0004074826414749221, 'samples': 7981568, 'steps': 15588, 'loss/train': 2.576552152633667} 02/24/2022 21:45:16 - INFO - codeparrot_training - Step 15589: {'lr': 0.00040746993320791685, 'samples': 7982080, 'steps': 15589, 'loss/train': 0.2182197868824005} 02/24/2022 21:45:20 - INFO - codeparrot_training - Step 15590: {'lr': 0.00040745722426636043, 'samples': 7982592, 'steps': 15590, 'loss/train': 1.162914514541626} 02/24/2022 21:45:23 - INFO - codeparrot_training - Step 15591: {'lr': 0.0004074445146503073, 'samples': 7983104, 'steps': 15591, 'loss/train': 2.2130420207977295} 02/24/2022 21:45:29 - INFO - codeparrot_training - Step 15592: {'lr': 0.00040743180435981187, 'samples': 7983616, 'steps': 15592, 'loss/train': 2.697943687438965} 02/24/2022 21:45:33 - INFO - codeparrot_training - Step 15593: {'lr': 0.0004074190933949286, 'samples': 7984128, 'steps': 15593, 'loss/train': 2.4169728755950928} 02/24/2022 21:45:38 - INFO - codeparrot_training - Step 15594: {'lr': 0.00040740638175571175, 'samples': 7984640, 'steps': 15594, 'loss/train': 1.820448875427246} 02/24/2022 21:45:41 - INFO - codeparrot_training - Step 15595: {'lr': 0.0004073936694422161, 'samples': 7985152, 'steps': 15595, 'loss/train': 0.09863793104887009} 02/24/2022 21:45:47 - INFO - codeparrot_training - Step 15596: {'lr': 0.0004073809564544959, 'samples': 7985664, 'steps': 15596, 'loss/train': 2.099989891052246} 02/24/2022 21:45:51 - INFO - codeparrot_training - Step 15597: {'lr': 0.0004073682427926057, 'samples': 7986176, 'steps': 15597, 'loss/train': 2.8123772144317627} 02/24/2022 21:45:58 - INFO - codeparrot_training - Step 15598: {'lr': 0.00040735552845659986, 'samples': 7986688, 'steps': 15598, 'loss/train': 1.7853034734725952} 02/24/2022 21:46:03 - INFO - codeparrot_training - Step 15599: {'lr': 0.00040734281344653294, 'samples': 7987200, 'steps': 15599, 'loss/train': 1.9252949953079224} 02/24/2022 21:46:07 - INFO - codeparrot_training - Step 15600: {'lr': 0.0004073300977624594, 'samples': 7987712, 'steps': 15600, 'loss/train': 1.0375285148620605} 02/24/2022 21:46:12 - INFO - codeparrot_training - Step 15601: {'lr': 0.0004073173814044336, 'samples': 7988224, 'steps': 15601, 'loss/train': 2.023422956466675} 02/24/2022 21:46:16 - INFO - codeparrot_training - Step 15602: {'lr': 0.0004073046643725101, 'samples': 7988736, 'steps': 15602, 'loss/train': 1.3903703689575195} 02/24/2022 21:46:21 - INFO - codeparrot_training - Step 15603: {'lr': 0.0004072919466667434, 'samples': 7989248, 'steps': 15603, 'loss/train': 3.008213758468628} 02/24/2022 21:46:25 - INFO - codeparrot_training - Step 15604: {'lr': 0.000407279228287188, 'samples': 7989760, 'steps': 15604, 'loss/train': 2.3658792972564697} 02/24/2022 21:46:30 - INFO - codeparrot_training - Step 15605: {'lr': 0.00040726650923389825, 'samples': 7990272, 'steps': 15605, 'loss/train': 3.367514133453369} 02/24/2022 21:46:34 - INFO - codeparrot_training - Step 15606: {'lr': 0.00040725378950692874, 'samples': 7990784, 'steps': 15606, 'loss/train': 2.2920644283294678} 02/24/2022 21:46:41 - INFO - codeparrot_training - Step 15607: {'lr': 0.0004072410691063339, 'samples': 7991296, 'steps': 15607, 'loss/train': 2.5761048793792725} 02/24/2022 21:46:44 - INFO - codeparrot_training - Step 15608: {'lr': 0.00040722834803216834, 'samples': 7991808, 'steps': 15608, 'loss/train': 2.1576735973358154} 02/24/2022 21:46:50 - INFO - codeparrot_training - Step 15609: {'lr': 0.0004072156262844864, 'samples': 7992320, 'steps': 15609, 'loss/train': 2.205754518508911} 02/24/2022 21:46:53 - INFO - codeparrot_training - Step 15610: {'lr': 0.0004072029038633426, 'samples': 7992832, 'steps': 15610, 'loss/train': 2.0891125202178955} 02/24/2022 21:46:59 - INFO - codeparrot_training - Step 15611: {'lr': 0.0004071901807687915, 'samples': 7993344, 'steps': 15611, 'loss/train': 1.7078381776809692} 02/24/2022 21:47:02 - INFO - codeparrot_training - Step 15612: {'lr': 0.0004071774570008876, 'samples': 7993856, 'steps': 15612, 'loss/train': 1.4559953212738037} 02/24/2022 21:47:08 - INFO - codeparrot_training - Step 15613: {'lr': 0.00040716473255968534, 'samples': 7994368, 'steps': 15613, 'loss/train': 2.5761075019836426} 02/24/2022 21:47:11 - INFO - codeparrot_training - Step 15614: {'lr': 0.0004071520074452393, 'samples': 7994880, 'steps': 15614, 'loss/train': 1.5095183849334717} 02/24/2022 21:47:17 - INFO - codeparrot_training - Step 15615: {'lr': 0.000407139281657604, 'samples': 7995392, 'steps': 15615, 'loss/train': 2.1408703327178955} 02/24/2022 21:47:20 - INFO - codeparrot_training - Step 15616: {'lr': 0.0004071265551968338, 'samples': 7995904, 'steps': 15616, 'loss/train': 2.1084792613983154} 02/24/2022 21:47:28 - INFO - codeparrot_training - Step 15617: {'lr': 0.0004071138280629835, 'samples': 7996416, 'steps': 15617, 'loss/train': 2.583139657974243} 02/24/2022 21:47:31 - INFO - codeparrot_training - Step 15618: {'lr': 0.00040710110025610733, 'samples': 7996928, 'steps': 15618, 'loss/train': 1.5119836330413818} 02/24/2022 21:47:37 - INFO - codeparrot_training - Step 15619: {'lr': 0.00040708837177626, 'samples': 7997440, 'steps': 15619, 'loss/train': 2.2536675930023193} 02/24/2022 21:47:40 - INFO - codeparrot_training - Step 15620: {'lr': 0.00040707564262349594, 'samples': 7997952, 'steps': 15620, 'loss/train': 2.496870756149292} 02/24/2022 21:47:46 - INFO - codeparrot_training - Step 15621: {'lr': 0.00040706291279786965, 'samples': 7998464, 'steps': 15621, 'loss/train': 1.6723992824554443} 02/24/2022 21:47:49 - INFO - codeparrot_training - Step 15622: {'lr': 0.0004070501822994358, 'samples': 7998976, 'steps': 15622, 'loss/train': 2.4864156246185303} 02/24/2022 21:47:55 - INFO - codeparrot_training - Step 15623: {'lr': 0.00040703745112824876, 'samples': 7999488, 'steps': 15623, 'loss/train': 2.650505781173706} 02/24/2022 21:47:58 - INFO - codeparrot_training - Step 15624: {'lr': 0.00040702471928436316, 'samples': 8000000, 'steps': 15624, 'loss/train': 2.7441656589508057} 02/24/2022 21:48:04 - INFO - codeparrot_training - Step 15625: {'lr': 0.00040701198676783355, 'samples': 8000512, 'steps': 15625, 'loss/train': 1.6590642929077148} 02/24/2022 21:48:07 - INFO - codeparrot_training - Step 15626: {'lr': 0.00040699925357871446, 'samples': 8001024, 'steps': 15626, 'loss/train': 2.945499897003174} 02/24/2022 21:48:13 - INFO - codeparrot_training - Step 15627: {'lr': 0.00040698651971706037, 'samples': 8001536, 'steps': 15627, 'loss/train': 0.9308986067771912} 02/24/2022 21:48:16 - INFO - codeparrot_training - Step 15628: {'lr': 0.00040697378518292593, 'samples': 8002048, 'steps': 15628, 'loss/train': 2.228546619415283} 02/24/2022 21:48:22 - INFO - codeparrot_training - Step 15629: {'lr': 0.0004069610499763656, 'samples': 8002560, 'steps': 15629, 'loss/train': 1.2791314125061035} 02/24/2022 21:48:26 - INFO - codeparrot_training - Step 15630: {'lr': 0.00040694831409743406, 'samples': 8003072, 'steps': 15630, 'loss/train': 2.3972809314727783} 02/24/2022 21:48:31 - INFO - codeparrot_training - Step 15631: {'lr': 0.00040693557754618566, 'samples': 8003584, 'steps': 15631, 'loss/train': 2.058864116668701} 02/24/2022 21:48:35 - INFO - codeparrot_training - Step 15632: {'lr': 0.00040692284032267515, 'samples': 8004096, 'steps': 15632, 'loss/train': 2.842947244644165} 02/24/2022 21:48:41 - INFO - codeparrot_training - Step 15633: {'lr': 0.00040691010242695696, 'samples': 8004608, 'steps': 15633, 'loss/train': 2.114063024520874} 02/24/2022 21:48:45 - INFO - codeparrot_training - Step 15634: {'lr': 0.00040689736385908574, 'samples': 8005120, 'steps': 15634, 'loss/train': 2.1035783290863037} 02/24/2022 21:48:50 - INFO - codeparrot_training - Step 15635: {'lr': 0.0004068846246191161, 'samples': 8005632, 'steps': 15635, 'loss/train': 1.6431081295013428} 02/24/2022 21:48:54 - INFO - codeparrot_training - Step 15636: {'lr': 0.00040687188470710245, 'samples': 8006144, 'steps': 15636, 'loss/train': 1.4904038906097412} 02/24/2022 21:49:00 - INFO - codeparrot_training - Step 15637: {'lr': 0.00040685914412309955, 'samples': 8006656, 'steps': 15637, 'loss/train': 2.2176451683044434} 02/24/2022 21:49:04 - INFO - codeparrot_training - Step 15638: {'lr': 0.0004068464028671618, 'samples': 8007168, 'steps': 15638, 'loss/train': 2.1982195377349854} 02/24/2022 21:49:07 - INFO - codeparrot_training - Step 15639: {'lr': 0.00040683366093934394, 'samples': 8007680, 'steps': 15639, 'loss/train': 2.6619622707366943} 02/24/2022 21:49:13 - INFO - codeparrot_training - Step 15640: {'lr': 0.0004068209183397004, 'samples': 8008192, 'steps': 15640, 'loss/train': 1.3103344440460205} 02/24/2022 21:49:16 - INFO - codeparrot_training - Step 15641: {'lr': 0.0004068081750682859, 'samples': 8008704, 'steps': 15641, 'loss/train': 1.9242271184921265} 02/24/2022 21:49:22 - INFO - codeparrot_training - Step 15642: {'lr': 0.00040679543112515494, 'samples': 8009216, 'steps': 15642, 'loss/train': 1.9095739126205444} 02/24/2022 21:49:25 - INFO - codeparrot_training - Step 15643: {'lr': 0.00040678268651036213, 'samples': 8009728, 'steps': 15643, 'loss/train': 2.1943652629852295} 02/24/2022 21:49:31 - INFO - codeparrot_training - Step 15644: {'lr': 0.0004067699412239622, 'samples': 8010240, 'steps': 15644, 'loss/train': 0.7311398983001709} 02/24/2022 21:49:35 - INFO - codeparrot_training - Step 15645: {'lr': 0.00040675719526600947, 'samples': 8010752, 'steps': 15645, 'loss/train': 1.8090782165527344} 02/24/2022 21:49:40 - INFO - codeparrot_training - Step 15646: {'lr': 0.0004067444486365587, 'samples': 8011264, 'steps': 15646, 'loss/train': 2.4807727336883545} 02/24/2022 21:49:44 - INFO - codeparrot_training - Step 15647: {'lr': 0.00040673170133566453, 'samples': 8011776, 'steps': 15647, 'loss/train': 1.0575315952301025} 02/24/2022 21:49:49 - INFO - codeparrot_training - Step 15648: {'lr': 0.0004067189533633815, 'samples': 8012288, 'steps': 15648, 'loss/train': 1.96336829662323} 02/24/2022 21:49:53 - INFO - codeparrot_training - Step 15649: {'lr': 0.00040670620471976426, 'samples': 8012800, 'steps': 15649, 'loss/train': 1.8861486911773682} 02/24/2022 21:49:58 - INFO - codeparrot_training - Step 15650: {'lr': 0.0004066934554048674, 'samples': 8013312, 'steps': 15650, 'loss/train': 3.094194173812866} 02/24/2022 21:50:02 - INFO - codeparrot_training - Step 15651: {'lr': 0.00040668070541874553, 'samples': 8013824, 'steps': 15651, 'loss/train': 2.193488836288452} 02/24/2022 21:50:07 - INFO - codeparrot_training - Step 15652: {'lr': 0.00040666795476145326, 'samples': 8014336, 'steps': 15652, 'loss/train': 2.2909913063049316} 02/24/2022 21:50:11 - INFO - codeparrot_training - Step 15653: {'lr': 0.00040665520343304516, 'samples': 8014848, 'steps': 15653, 'loss/train': 0.6748460531234741} 02/24/2022 21:50:16 - INFO - codeparrot_training - Step 15654: {'lr': 0.00040664245143357604, 'samples': 8015360, 'steps': 15654, 'loss/train': 2.3541500568389893} 02/24/2022 21:50:20 - INFO - codeparrot_training - Step 15655: {'lr': 0.0004066296987631003, 'samples': 8015872, 'steps': 15655, 'loss/train': 2.6186037063598633} 02/24/2022 21:50:26 - INFO - codeparrot_training - Step 15656: {'lr': 0.0004066169454216727, 'samples': 8016384, 'steps': 15656, 'loss/train': 2.1786561012268066} 02/24/2022 21:50:29 - INFO - codeparrot_training - Step 15657: {'lr': 0.00040660419140934787, 'samples': 8016896, 'steps': 15657, 'loss/train': 2.1989779472351074} 02/24/2022 21:50:35 - INFO - codeparrot_training - Step 15658: {'lr': 0.0004065914367261804, 'samples': 8017408, 'steps': 15658, 'loss/train': 2.0224032402038574} 02/24/2022 21:50:38 - INFO - codeparrot_training - Step 15659: {'lr': 0.00040657868137222486, 'samples': 8017920, 'steps': 15659, 'loss/train': 2.074057102203369} 02/24/2022 21:50:44 - INFO - codeparrot_training - Step 15660: {'lr': 0.000406565925347536, 'samples': 8018432, 'steps': 15660, 'loss/train': 2.2125167846679688} 02/24/2022 21:50:47 - INFO - codeparrot_training - Step 15661: {'lr': 0.0004065531686521685, 'samples': 8018944, 'steps': 15661, 'loss/train': 1.2297435998916626} 02/24/2022 21:50:53 - INFO - codeparrot_training - Step 15662: {'lr': 0.00040654041128617693, 'samples': 8019456, 'steps': 15662, 'loss/train': 0.8650669455528259} 02/24/2022 21:50:56 - INFO - codeparrot_training - Step 15663: {'lr': 0.0004065276532496158, 'samples': 8019968, 'steps': 15663, 'loss/train': 2.5333142280578613} 02/24/2022 21:51:03 - INFO - codeparrot_training - Step 15664: {'lr': 0.0004065148945425401, 'samples': 8020480, 'steps': 15664, 'loss/train': 1.7127292156219482} 02/24/2022 21:51:06 - INFO - codeparrot_training - Step 15665: {'lr': 0.0004065021351650042, 'samples': 8020992, 'steps': 15665, 'loss/train': 1.736020565032959} 02/24/2022 21:51:12 - INFO - codeparrot_training - Step 15666: {'lr': 0.00040648937511706285, 'samples': 8021504, 'steps': 15666, 'loss/train': 2.829667329788208} 02/24/2022 21:51:15 - INFO - codeparrot_training - Step 15667: {'lr': 0.0004064766143987707, 'samples': 8022016, 'steps': 15667, 'loss/train': 0.9548608660697937} 02/24/2022 21:51:21 - INFO - codeparrot_training - Step 15668: {'lr': 0.00040646385301018243, 'samples': 8022528, 'steps': 15668, 'loss/train': 3.3888297080993652} 02/24/2022 21:51:24 - INFO - codeparrot_training - Step 15669: {'lr': 0.0004064510909513527, 'samples': 8023040, 'steps': 15669, 'loss/train': 2.4436657428741455} 02/24/2022 21:51:30 - INFO - codeparrot_training - Step 15670: {'lr': 0.00040643832822233615, 'samples': 8023552, 'steps': 15670, 'loss/train': 2.4292538166046143} 02/24/2022 21:51:33 - INFO - codeparrot_training - Step 15671: {'lr': 0.0004064255648231875, 'samples': 8024064, 'steps': 15671, 'loss/train': 2.216278076171875} 02/24/2022 21:51:39 - INFO - codeparrot_training - Step 15672: {'lr': 0.00040641280075396144, 'samples': 8024576, 'steps': 15672, 'loss/train': 1.429274559020996} 02/24/2022 21:51:42 - INFO - codeparrot_training - Step 15673: {'lr': 0.00040640003601471255, 'samples': 8025088, 'steps': 15673, 'loss/train': 3.0734481811523438} 02/24/2022 21:51:48 - INFO - codeparrot_training - Step 15674: {'lr': 0.00040638727060549556, 'samples': 8025600, 'steps': 15674, 'loss/train': 1.9044876098632812} 02/24/2022 21:51:52 - INFO - codeparrot_training - Step 15675: {'lr': 0.00040637450452636517, 'samples': 8026112, 'steps': 15675, 'loss/train': 1.3907132148742676} 02/24/2022 21:51:57 - INFO - codeparrot_training - Step 15676: {'lr': 0.00040636173777737613, 'samples': 8026624, 'steps': 15676, 'loss/train': 0.8431764245033264} 02/24/2022 21:52:01 - INFO - codeparrot_training - Step 15677: {'lr': 0.000406348970358583, 'samples': 8027136, 'steps': 15677, 'loss/train': 1.6959254741668701} 02/24/2022 21:52:06 - INFO - codeparrot_training - Step 15678: {'lr': 0.00040633620227004054, 'samples': 8027648, 'steps': 15678, 'loss/train': 2.103938579559326} 02/24/2022 21:52:10 - INFO - codeparrot_training - Step 15679: {'lr': 0.0004063234335118033, 'samples': 8028160, 'steps': 15679, 'loss/train': 1.601721167564392} 02/24/2022 21:52:15 - INFO - codeparrot_training - Step 15680: {'lr': 0.00040631066408392636, 'samples': 8028672, 'steps': 15680, 'loss/train': 0.8082705736160278} 02/24/2022 21:52:19 - INFO - codeparrot_training - Step 15681: {'lr': 0.000406297893986464, 'samples': 8029184, 'steps': 15681, 'loss/train': 1.3983267545700073} 02/24/2022 21:52:25 - INFO - codeparrot_training - Step 15682: {'lr': 0.0004062851232194711, 'samples': 8029696, 'steps': 15682, 'loss/train': 2.9970362186431885} 02/24/2022 21:52:29 - INFO - codeparrot_training - Step 15683: {'lr': 0.00040627235178300236, 'samples': 8030208, 'steps': 15683, 'loss/train': 2.2244744300842285} 02/24/2022 21:52:35 - INFO - codeparrot_training - Step 15684: {'lr': 0.0004062595796771126, 'samples': 8030720, 'steps': 15684, 'loss/train': 1.7679158449172974} 02/24/2022 21:52:38 - INFO - codeparrot_training - Step 15685: {'lr': 0.0004062468069018563, 'samples': 8031232, 'steps': 15685, 'loss/train': 1.208449363708496} 02/24/2022 21:52:42 - INFO - codeparrot_training - Step 15686: {'lr': 0.0004062340334572883, 'samples': 8031744, 'steps': 15686, 'loss/train': 1.7850629091262817} 02/24/2022 21:52:47 - INFO - codeparrot_training - Step 15687: {'lr': 0.0004062212593434634, 'samples': 8032256, 'steps': 15687, 'loss/train': 2.580045700073242} 02/24/2022 21:52:51 - INFO - codeparrot_training - Step 15688: {'lr': 0.0004062084845604361, 'samples': 8032768, 'steps': 15688, 'loss/train': 1.4563696384429932} 02/24/2022 21:52:56 - INFO - codeparrot_training - Step 15689: {'lr': 0.00040619570910826135, 'samples': 8033280, 'steps': 15689, 'loss/train': 2.5587480068206787} 02/24/2022 21:53:00 - INFO - codeparrot_training - Step 15690: {'lr': 0.0004061829329869937, 'samples': 8033792, 'steps': 15690, 'loss/train': 1.687915563583374} 02/24/2022 21:53:06 - INFO - codeparrot_training - Step 15691: {'lr': 0.0004061701561966881, 'samples': 8034304, 'steps': 15691, 'loss/train': 2.7886128425598145} 02/24/2022 21:53:09 - INFO - codeparrot_training - Step 15692: {'lr': 0.000406157378737399, 'samples': 8034816, 'steps': 15692, 'loss/train': 2.4578754901885986} 02/24/2022 21:53:16 - INFO - codeparrot_training - Step 15693: {'lr': 0.00040614460060918136, 'samples': 8035328, 'steps': 15693, 'loss/train': 1.6393136978149414} 02/24/2022 21:53:20 - INFO - codeparrot_training - Step 15694: {'lr': 0.0004061318218120898, 'samples': 8035840, 'steps': 15694, 'loss/train': 2.4925897121429443} 02/24/2022 21:53:25 - INFO - codeparrot_training - Step 15695: {'lr': 0.000406119042346179, 'samples': 8036352, 'steps': 15695, 'loss/train': 2.0249102115631104} 02/24/2022 21:53:29 - INFO - codeparrot_training - Step 15696: {'lr': 0.0004061062622115039, 'samples': 8036864, 'steps': 15696, 'loss/train': 2.170201539993286} 02/24/2022 21:53:34 - INFO - codeparrot_training - Step 15697: {'lr': 0.0004060934814081192, 'samples': 8037376, 'steps': 15697, 'loss/train': 2.5346755981445312} 02/24/2022 21:53:38 - INFO - codeparrot_training - Step 15698: {'lr': 0.00040608069993607954, 'samples': 8037888, 'steps': 15698, 'loss/train': 1.5353726148605347} 02/24/2022 21:53:43 - INFO - codeparrot_training - Step 15699: {'lr': 0.00040606791779543966, 'samples': 8038400, 'steps': 15699, 'loss/train': 2.2519142627716064} 02/24/2022 21:53:47 - INFO - codeparrot_training - Step 15700: {'lr': 0.00040605513498625443, 'samples': 8038912, 'steps': 15700, 'loss/train': 2.1619937419891357} 02/24/2022 21:53:52 - INFO - codeparrot_training - Step 15701: {'lr': 0.00040604235150857855, 'samples': 8039424, 'steps': 15701, 'loss/train': 2.3589928150177} 02/24/2022 21:53:56 - INFO - codeparrot_training - Step 15702: {'lr': 0.00040602956736246677, 'samples': 8039936, 'steps': 15702, 'loss/train': 2.3140981197357178} 02/24/2022 21:54:02 - INFO - codeparrot_training - Step 15703: {'lr': 0.00040601678254797394, 'samples': 8040448, 'steps': 15703, 'loss/train': 2.0368294715881348} 02/24/2022 21:54:05 - INFO - codeparrot_training - Step 15704: {'lr': 0.00040600399706515466, 'samples': 8040960, 'steps': 15704, 'loss/train': 2.1878161430358887} 02/24/2022 21:54:11 - INFO - codeparrot_training - Step 15705: {'lr': 0.0004059912109140638, 'samples': 8041472, 'steps': 15705, 'loss/train': 2.316575288772583} 02/24/2022 21:54:14 - INFO - codeparrot_training - Step 15706: {'lr': 0.00040597842409475615, 'samples': 8041984, 'steps': 15706, 'loss/train': 1.708521842956543} 02/24/2022 21:54:20 - INFO - codeparrot_training - Step 15707: {'lr': 0.00040596563660728646, 'samples': 8042496, 'steps': 15707, 'loss/train': 1.4838895797729492} 02/24/2022 21:54:23 - INFO - codeparrot_training - Step 15708: {'lr': 0.00040595284845170956, 'samples': 8043008, 'steps': 15708, 'loss/train': 1.8247803449630737} 02/24/2022 21:54:29 - INFO - codeparrot_training - Step 15709: {'lr': 0.0004059400596280801, 'samples': 8043520, 'steps': 15709, 'loss/train': 3.237764596939087} 02/24/2022 21:54:32 - INFO - codeparrot_training - Step 15710: {'lr': 0.00040592727013645297, 'samples': 8044032, 'steps': 15710, 'loss/train': 1.5520082712173462} 02/24/2022 21:54:38 - INFO - codeparrot_training - Step 15711: {'lr': 0.0004059144799768829, 'samples': 8044544, 'steps': 15711, 'loss/train': 1.7852360010147095} 02/24/2022 21:54:41 - INFO - codeparrot_training - Step 15712: {'lr': 0.00040590168914942477, 'samples': 8045056, 'steps': 15712, 'loss/train': 1.7346484661102295} 02/24/2022 21:54:48 - INFO - codeparrot_training - Step 15713: {'lr': 0.0004058888976541333, 'samples': 8045568, 'steps': 15713, 'loss/train': 2.3986096382141113} 02/24/2022 21:54:51 - INFO - codeparrot_training - Step 15714: {'lr': 0.00040587610549106326, 'samples': 8046080, 'steps': 15714, 'loss/train': 1.5621291399002075} 02/24/2022 21:54:57 - INFO - codeparrot_training - Step 15715: {'lr': 0.00040586331266026943, 'samples': 8046592, 'steps': 15715, 'loss/train': 2.487247943878174} 02/24/2022 21:55:00 - INFO - codeparrot_training - Step 15716: {'lr': 0.0004058505191618067, 'samples': 8047104, 'steps': 15716, 'loss/train': 3.226485252380371} 02/24/2022 21:55:06 - INFO - codeparrot_training - Step 15717: {'lr': 0.0004058377249957299, 'samples': 8047616, 'steps': 15717, 'loss/train': 9.076168060302734} 02/24/2022 21:55:09 - INFO - codeparrot_training - Step 15718: {'lr': 0.0004058249301620937, 'samples': 8048128, 'steps': 15718, 'loss/train': 1.2968710660934448} 02/24/2022 21:55:15 - INFO - codeparrot_training - Step 15719: {'lr': 0.00040581213466095304, 'samples': 8048640, 'steps': 15719, 'loss/train': 1.729250192642212} 02/24/2022 21:55:18 - INFO - codeparrot_training - Step 15720: {'lr': 0.0004057993384923626, 'samples': 8049152, 'steps': 15720, 'loss/train': 2.3715617656707764} 02/24/2022 21:55:24 - INFO - codeparrot_training - Step 15721: {'lr': 0.0004057865416563773, 'samples': 8049664, 'steps': 15721, 'loss/train': 2.581035852432251} 02/24/2022 21:55:27 - INFO - codeparrot_training - Step 15722: {'lr': 0.0004057737441530519, 'samples': 8050176, 'steps': 15722, 'loss/train': 1.50064218044281} 02/24/2022 21:55:33 - INFO - codeparrot_training - Step 15723: {'lr': 0.0004057609459824412, 'samples': 8050688, 'steps': 15723, 'loss/train': 3.061279058456421} 02/24/2022 21:55:36 - INFO - codeparrot_training - Step 15724: {'lr': 0.00040574814714460015, 'samples': 8051200, 'steps': 15724, 'loss/train': 2.3681302070617676} 02/24/2022 21:55:42 - INFO - codeparrot_training - Step 15725: {'lr': 0.0004057353476395835, 'samples': 8051712, 'steps': 15725, 'loss/train': 2.540001153945923} 02/24/2022 21:55:45 - INFO - codeparrot_training - Step 15726: {'lr': 0.00040572254746744607, 'samples': 8052224, 'steps': 15726, 'loss/train': 2.6839778423309326} 02/24/2022 21:55:52 - INFO - codeparrot_training - Step 15727: {'lr': 0.00040570974662824266, 'samples': 8052736, 'steps': 15727, 'loss/train': 1.0916556119918823} 02/24/2022 21:55:55 - INFO - codeparrot_training - Step 15728: {'lr': 0.00040569694512202815, 'samples': 8053248, 'steps': 15728, 'loss/train': 2.3467535972595215} 02/24/2022 21:56:01 - INFO - codeparrot_training - Step 15729: {'lr': 0.00040568414294885736, 'samples': 8053760, 'steps': 15729, 'loss/train': 1.5626912117004395} 02/24/2022 21:56:04 - INFO - codeparrot_training - Step 15730: {'lr': 0.00040567134010878513, 'samples': 8054272, 'steps': 15730, 'loss/train': 2.3129680156707764} 02/24/2022 21:56:10 - INFO - codeparrot_training - Step 15731: {'lr': 0.00040565853660186633, 'samples': 8054784, 'steps': 15731, 'loss/train': 2.3960771560668945} 02/24/2022 21:56:13 - INFO - codeparrot_training - Step 15732: {'lr': 0.0004056457324281557, 'samples': 8055296, 'steps': 15732, 'loss/train': 1.1214845180511475} 02/24/2022 21:56:19 - INFO - codeparrot_training - Step 15733: {'lr': 0.0004056329275877083, 'samples': 8055808, 'steps': 15733, 'loss/train': 2.5366082191467285} 02/24/2022 21:56:22 - INFO - codeparrot_training - Step 15734: {'lr': 0.00040562012208057886, 'samples': 8056320, 'steps': 15734, 'loss/train': 2.354208469390869} 02/24/2022 21:56:28 - INFO - codeparrot_training - Step 15735: {'lr': 0.0004056073159068222, 'samples': 8056832, 'steps': 15735, 'loss/train': 1.8322428464889526} 02/24/2022 21:56:31 - INFO - codeparrot_training - Step 15736: {'lr': 0.0004055945090664931, 'samples': 8057344, 'steps': 15736, 'loss/train': 1.8507310152053833} 02/24/2022 21:56:37 - INFO - codeparrot_training - Step 15737: {'lr': 0.0004055817015596467, 'samples': 8057856, 'steps': 15737, 'loss/train': 0.676263153553009} 02/24/2022 21:56:40 - INFO - codeparrot_training - Step 15738: {'lr': 0.00040556889338633754, 'samples': 8058368, 'steps': 15738, 'loss/train': 2.2110085487365723} 02/24/2022 21:56:47 - INFO - codeparrot_training - Step 15739: {'lr': 0.00040555608454662074, 'samples': 8058880, 'steps': 15739, 'loss/train': 2.2826342582702637} 02/24/2022 21:56:50 - INFO - codeparrot_training - Step 15740: {'lr': 0.00040554327504055106, 'samples': 8059392, 'steps': 15740, 'loss/train': 4.769762992858887} 02/24/2022 21:56:55 - INFO - codeparrot_training - Step 15741: {'lr': 0.00040553046486818336, 'samples': 8059904, 'steps': 15741, 'loss/train': 0.9186894297599792} 02/24/2022 21:56:59 - INFO - codeparrot_training - Step 15742: {'lr': 0.0004055176540295725, 'samples': 8060416, 'steps': 15742, 'loss/train': 2.7471678256988525} 02/24/2022 21:57:04 - INFO - codeparrot_training - Step 15743: {'lr': 0.00040550484252477347, 'samples': 8060928, 'steps': 15743, 'loss/train': 0.5740983486175537} 02/24/2022 21:57:08 - INFO - codeparrot_training - Step 15744: {'lr': 0.00040549203035384105, 'samples': 8061440, 'steps': 15744, 'loss/train': 1.8564348220825195} 02/24/2022 21:57:13 - INFO - codeparrot_training - Step 15745: {'lr': 0.0004054792175168301, 'samples': 8061952, 'steps': 15745, 'loss/train': 2.8452157974243164} 02/24/2022 21:57:17 - INFO - codeparrot_training - Step 15746: {'lr': 0.00040546640401379556, 'samples': 8062464, 'steps': 15746, 'loss/train': 1.6108318567276} 02/24/2022 21:57:22 - INFO - codeparrot_training - Step 15747: {'lr': 0.0004054535898447924, 'samples': 8062976, 'steps': 15747, 'loss/train': 1.845558524131775} 02/24/2022 21:57:26 - INFO - codeparrot_training - Step 15748: {'lr': 0.0004054407750098753, 'samples': 8063488, 'steps': 15748, 'loss/train': 1.8081797361373901} 02/24/2022 21:57:33 - INFO - codeparrot_training - Step 15749: {'lr': 0.0004054279595090994, 'samples': 8064000, 'steps': 15749, 'loss/train': 2.148766279220581} 02/24/2022 21:57:36 - INFO - codeparrot_training - Step 15750: {'lr': 0.0004054151433425194, 'samples': 8064512, 'steps': 15750, 'loss/train': 2.715588331222534} 02/24/2022 21:57:41 - INFO - codeparrot_training - Step 15751: {'lr': 0.00040540232651019027, 'samples': 8065024, 'steps': 15751, 'loss/train': 2.917924404144287} 02/24/2022 21:57:45 - INFO - codeparrot_training - Step 15752: {'lr': 0.0004053895090121669, 'samples': 8065536, 'steps': 15752, 'loss/train': 1.883616328239441} 02/24/2022 21:57:51 - INFO - codeparrot_training - Step 15753: {'lr': 0.00040537669084850426, 'samples': 8066048, 'steps': 15753, 'loss/train': 1.8007880449295044} 02/24/2022 21:57:56 - INFO - codeparrot_training - Step 15754: {'lr': 0.0004053638720192572, 'samples': 8066560, 'steps': 15754, 'loss/train': 2.740985155105591} 02/24/2022 21:58:00 - INFO - codeparrot_training - Step 15755: {'lr': 0.00040535105252448067, 'samples': 8067072, 'steps': 15755, 'loss/train': 1.8929885625839233} 02/24/2022 21:58:05 - INFO - codeparrot_training - Step 15756: {'lr': 0.0004053382323642295, 'samples': 8067584, 'steps': 15756, 'loss/train': 0.9658442735671997} 02/24/2022 21:58:09 - INFO - codeparrot_training - Step 15757: {'lr': 0.0004053254115385587, 'samples': 8068096, 'steps': 15757, 'loss/train': 1.908900260925293} 02/24/2022 21:58:15 - INFO - codeparrot_training - Step 15758: {'lr': 0.00040531259004752317, 'samples': 8068608, 'steps': 15758, 'loss/train': 2.4591195583343506} 02/24/2022 21:58:19 - INFO - codeparrot_training - Step 15759: {'lr': 0.00040529976789117786, 'samples': 8069120, 'steps': 15759, 'loss/train': 1.210711121559143} 02/24/2022 21:58:24 - INFO - codeparrot_training - Step 15760: {'lr': 0.0004052869450695776, 'samples': 8069632, 'steps': 15760, 'loss/train': 1.6907556056976318} 02/24/2022 21:58:28 - INFO - codeparrot_training - Step 15761: {'lr': 0.00040527412158277744, 'samples': 8070144, 'steps': 15761, 'loss/train': 1.694442868232727} 02/24/2022 21:58:34 - INFO - codeparrot_training - Step 15762: {'lr': 0.00040526129743083216, 'samples': 8070656, 'steps': 15762, 'loss/train': 1.5524027347564697} 02/24/2022 21:58:37 - INFO - codeparrot_training - Step 15763: {'lr': 0.0004052484726137968, 'samples': 8071168, 'steps': 15763, 'loss/train': 2.5131072998046875} 02/24/2022 21:58:43 - INFO - codeparrot_training - Step 15764: {'lr': 0.00040523564713172634, 'samples': 8071680, 'steps': 15764, 'loss/train': 1.8341569900512695} 02/24/2022 21:58:46 - INFO - codeparrot_training - Step 15765: {'lr': 0.0004052228209846756, 'samples': 8072192, 'steps': 15765, 'loss/train': 1.048017144203186} 02/24/2022 21:58:52 - INFO - codeparrot_training - Step 15766: {'lr': 0.0004052099941726996, 'samples': 8072704, 'steps': 15766, 'loss/train': 2.483851194381714} 02/24/2022 21:58:55 - INFO - codeparrot_training - Step 15767: {'lr': 0.0004051971666958533, 'samples': 8073216, 'steps': 15767, 'loss/train': 2.489689588546753} 02/24/2022 21:59:01 - INFO - codeparrot_training - Step 15768: {'lr': 0.0004051843385541916, 'samples': 8073728, 'steps': 15768, 'loss/train': 2.7603371143341064} 02/24/2022 21:59:04 - INFO - codeparrot_training - Step 15769: {'lr': 0.00040517150974776945, 'samples': 8074240, 'steps': 15769, 'loss/train': 1.8423763513565063} 02/24/2022 21:59:10 - INFO - codeparrot_training - Step 15770: {'lr': 0.00040515868027664185, 'samples': 8074752, 'steps': 15770, 'loss/train': 3.674001693725586} 02/24/2022 21:59:13 - INFO - codeparrot_training - Step 15771: {'lr': 0.00040514585014086367, 'samples': 8075264, 'steps': 15771, 'loss/train': 1.7027298212051392} 02/24/2022 21:59:19 - INFO - codeparrot_training - Step 15772: {'lr': 0.00040513301934049005, 'samples': 8075776, 'steps': 15772, 'loss/train': 2.7007603645324707} 02/24/2022 21:59:22 - INFO - codeparrot_training - Step 15773: {'lr': 0.00040512018787557574, 'samples': 8076288, 'steps': 15773, 'loss/train': 1.7654494047164917} 02/24/2022 21:59:29 - INFO - codeparrot_training - Step 15774: {'lr': 0.0004051073557461759, 'samples': 8076800, 'steps': 15774, 'loss/train': 0.9119476675987244} 02/24/2022 21:59:32 - INFO - codeparrot_training - Step 15775: {'lr': 0.00040509452295234527, 'samples': 8077312, 'steps': 15775, 'loss/train': 0.22977015376091003} 02/24/2022 21:59:39 - INFO - codeparrot_training - Step 15776: {'lr': 0.00040508168949413904, 'samples': 8077824, 'steps': 15776, 'loss/train': 2.417126417160034} 02/24/2022 21:59:42 - INFO - codeparrot_training - Step 15777: {'lr': 0.0004050688553716121, 'samples': 8078336, 'steps': 15777, 'loss/train': 2.347332715988159} 02/24/2022 21:59:46 - INFO - codeparrot_training - Step 15778: {'lr': 0.0004050560205848194, 'samples': 8078848, 'steps': 15778, 'loss/train': 1.4732513427734375} 02/24/2022 21:59:51 - INFO - codeparrot_training - Step 15779: {'lr': 0.0004050431851338159, 'samples': 8079360, 'steps': 15779, 'loss/train': 1.280692219734192} 02/24/2022 21:59:55 - INFO - codeparrot_training - Step 15780: {'lr': 0.00040503034901865666, 'samples': 8079872, 'steps': 15780, 'loss/train': 2.910001277923584} 02/24/2022 22:00:00 - INFO - codeparrot_training - Step 15781: {'lr': 0.00040501751223939665, 'samples': 8080384, 'steps': 15781, 'loss/train': 1.4696500301361084} 02/24/2022 22:00:04 - INFO - codeparrot_training - Step 15782: {'lr': 0.00040500467479609084, 'samples': 8080896, 'steps': 15782, 'loss/train': 1.9289052486419678} 02/24/2022 22:00:09 - INFO - codeparrot_training - Step 15783: {'lr': 0.00040499183668879415, 'samples': 8081408, 'steps': 15783, 'loss/train': 2.3125689029693604} 02/24/2022 22:00:13 - INFO - codeparrot_training - Step 15784: {'lr': 0.0004049789979175617, 'samples': 8081920, 'steps': 15784, 'loss/train': 2.5349185466766357} 02/24/2022 22:00:19 - INFO - codeparrot_training - Step 15785: {'lr': 0.00040496615848244845, 'samples': 8082432, 'steps': 15785, 'loss/train': 2.0821986198425293} 02/24/2022 22:00:22 - INFO - codeparrot_training - Step 15786: {'lr': 0.00040495331838350933, 'samples': 8082944, 'steps': 15786, 'loss/train': 2.451448917388916} 02/24/2022 22:00:28 - INFO - codeparrot_training - Step 15787: {'lr': 0.00040494047762079953, 'samples': 8083456, 'steps': 15787, 'loss/train': 2.157874822616577} 02/24/2022 22:00:31 - INFO - codeparrot_training - Step 15788: {'lr': 0.0004049276361943738, 'samples': 8083968, 'steps': 15788, 'loss/train': 4.5446038246154785} 02/24/2022 22:00:37 - INFO - codeparrot_training - Step 15789: {'lr': 0.00040491479410428735, 'samples': 8084480, 'steps': 15789, 'loss/train': 1.0855789184570312} 02/24/2022 22:00:40 - INFO - codeparrot_training - Step 15790: {'lr': 0.00040490195135059503, 'samples': 8084992, 'steps': 15790, 'loss/train': 2.2109575271606445} 02/24/2022 22:00:46 - INFO - codeparrot_training - Step 15791: {'lr': 0.000404889107933352, 'samples': 8085504, 'steps': 15791, 'loss/train': 1.4307068586349487} 02/24/2022 22:00:49 - INFO - codeparrot_training - Step 15792: {'lr': 0.0004048762638526132, 'samples': 8086016, 'steps': 15792, 'loss/train': 3.499948740005493} 02/24/2022 22:00:55 - INFO - codeparrot_training - Step 15793: {'lr': 0.0004048634191084336, 'samples': 8086528, 'steps': 15793, 'loss/train': 2.250875234603882} 02/24/2022 22:00:58 - INFO - codeparrot_training - Step 15794: {'lr': 0.0004048505737008684, 'samples': 8087040, 'steps': 15794, 'loss/train': 1.9795410633087158} 02/24/2022 22:01:04 - INFO - codeparrot_training - Step 15795: {'lr': 0.0004048377276299724, 'samples': 8087552, 'steps': 15795, 'loss/train': 2.4537510871887207} 02/24/2022 22:01:08 - INFO - codeparrot_training - Step 15796: {'lr': 0.00040482488089580083, 'samples': 8088064, 'steps': 15796, 'loss/train': 1.2694814205169678} 02/24/2022 22:01:13 - INFO - codeparrot_training - Step 15797: {'lr': 0.00040481203349840864, 'samples': 8088576, 'steps': 15797, 'loss/train': 2.8642430305480957} 02/24/2022 22:01:17 - INFO - codeparrot_training - Step 15798: {'lr': 0.0004047991854378508, 'samples': 8089088, 'steps': 15798, 'loss/train': 3.310959577560425} 02/24/2022 22:01:23 - INFO - codeparrot_training - Step 15799: {'lr': 0.00040478633671418244, 'samples': 8089600, 'steps': 15799, 'loss/train': 1.7746672630310059} 02/24/2022 22:01:26 - INFO - codeparrot_training - Step 15800: {'lr': 0.00040477348732745853, 'samples': 8090112, 'steps': 15800, 'loss/train': 2.1471316814422607} 02/24/2022 22:01:32 - INFO - codeparrot_training - Step 15801: {'lr': 0.00040476063727773416, 'samples': 8090624, 'steps': 15801, 'loss/train': 1.0932163000106812} 02/24/2022 22:01:35 - INFO - codeparrot_training - Step 15802: {'lr': 0.0004047477865650644, 'samples': 8091136, 'steps': 15802, 'loss/train': 2.674537181854248} 02/24/2022 22:01:40 - INFO - codeparrot_training - Step 15803: {'lr': 0.00040473493518950414, 'samples': 8091648, 'steps': 15803, 'loss/train': 1.7999407052993774} 02/24/2022 22:01:44 - INFO - codeparrot_training - Step 15804: {'lr': 0.00040472208315110866, 'samples': 8092160, 'steps': 15804, 'loss/train': 2.390216827392578} 02/24/2022 22:02:51 - INFO - codeparrot_training - Step 15805: {'lr': 0.0004047092304499329, 'samples': 8092672, 'steps': 15805, 'loss/train': 2.115192174911499} 02/24/2022 22:02:55 - INFO - codeparrot_training - Step 15806: {'lr': 0.0004046963770860319, 'samples': 8093184, 'steps': 15806, 'loss/train': 0.5324437022209167} 02/24/2022 22:03:00 - INFO - codeparrot_training - Step 15807: {'lr': 0.0004046835230594608, 'samples': 8093696, 'steps': 15807, 'loss/train': 1.6067605018615723} 02/24/2022 22:03:04 - INFO - codeparrot_training - Step 15808: {'lr': 0.0004046706683702744, 'samples': 8094208, 'steps': 15808, 'loss/train': 2.509697914123535} 02/24/2022 22:03:09 - INFO - codeparrot_training - Step 15809: {'lr': 0.0004046578130185282, 'samples': 8094720, 'steps': 15809, 'loss/train': 0.5408334136009216} 02/24/2022 22:03:13 - INFO - codeparrot_training - Step 15810: {'lr': 0.00040464495700427694, 'samples': 8095232, 'steps': 15810, 'loss/train': 2.224310874938965} 02/24/2022 22:03:18 - INFO - codeparrot_training - Step 15811: {'lr': 0.0004046321003275759, 'samples': 8095744, 'steps': 15811, 'loss/train': 2.0357437133789062} 02/24/2022 22:03:22 - INFO - codeparrot_training - Step 15812: {'lr': 0.00040461924298847987, 'samples': 8096256, 'steps': 15812, 'loss/train': 2.466475009918213} 02/24/2022 22:03:27 - INFO - codeparrot_training - Step 15813: {'lr': 0.0004046063849870442, 'samples': 8096768, 'steps': 15813, 'loss/train': 1.7881413698196411} 02/24/2022 22:03:31 - INFO - codeparrot_training - Step 15814: {'lr': 0.00040459352632332387, 'samples': 8097280, 'steps': 15814, 'loss/train': 0.9035694003105164} 02/24/2022 22:03:36 - INFO - codeparrot_training - Step 15815: {'lr': 0.0004045806669973739, 'samples': 8097792, 'steps': 15815, 'loss/train': 1.3460819721221924} 02/24/2022 22:03:42 - INFO - codeparrot_training - Step 15816: {'lr': 0.00040456780700924956, 'samples': 8098304, 'steps': 15816, 'loss/train': 2.337918996810913} 02/24/2022 22:03:45 - INFO - codeparrot_training - Step 15817: {'lr': 0.0004045549463590057, 'samples': 8098816, 'steps': 15817, 'loss/train': 0.1956883817911148} 02/24/2022 22:03:51 - INFO - codeparrot_training - Step 15818: {'lr': 0.0004045420850466975, 'samples': 8099328, 'steps': 15818, 'loss/train': 3.039705991744995} 02/24/2022 22:03:54 - INFO - codeparrot_training - Step 15819: {'lr': 0.00040452922307238016, 'samples': 8099840, 'steps': 15819, 'loss/train': 2.5365610122680664} 02/24/2022 22:04:01 - INFO - codeparrot_training - Step 15820: {'lr': 0.00040451636043610875, 'samples': 8100352, 'steps': 15820, 'loss/train': 1.4724889993667603} 02/24/2022 22:04:04 - INFO - codeparrot_training - Step 15821: {'lr': 0.0004045034971379382, 'samples': 8100864, 'steps': 15821, 'loss/train': 1.7322403192520142} 02/24/2022 22:04:10 - INFO - codeparrot_training - Step 15822: {'lr': 0.0004044906331779238, 'samples': 8101376, 'steps': 15822, 'loss/train': 3.248962640762329} 02/24/2022 22:04:13 - INFO - codeparrot_training - Step 15823: {'lr': 0.00040447776855612053, 'samples': 8101888, 'steps': 15823, 'loss/train': 1.9005476236343384} 02/24/2022 22:04:19 - INFO - codeparrot_training - Step 15824: {'lr': 0.0004044649032725836, 'samples': 8102400, 'steps': 15824, 'loss/train': 2.2104201316833496} 02/24/2022 22:04:22 - INFO - codeparrot_training - Step 15825: {'lr': 0.000404452037327368, 'samples': 8102912, 'steps': 15825, 'loss/train': 1.8906991481781006} 02/24/2022 22:04:28 - INFO - codeparrot_training - Step 15826: {'lr': 0.00040443917072052906, 'samples': 8103424, 'steps': 15826, 'loss/train': 1.7152348756790161} 02/24/2022 22:04:31 - INFO - codeparrot_training - Step 15827: {'lr': 0.0004044263034521216, 'samples': 8103936, 'steps': 15827, 'loss/train': 1.5793169736862183} 02/24/2022 22:04:37 - INFO - codeparrot_training - Step 15828: {'lr': 0.000404413435522201, 'samples': 8104448, 'steps': 15828, 'loss/train': 1.8360563516616821} 02/24/2022 22:04:40 - INFO - codeparrot_training - Step 15829: {'lr': 0.00040440056693082224, 'samples': 8104960, 'steps': 15829, 'loss/train': 2.155177354812622} 02/24/2022 22:04:47 - INFO - codeparrot_training - Step 15830: {'lr': 0.0004043876976780404, 'samples': 8105472, 'steps': 15830, 'loss/train': 1.060920238494873} 02/24/2022 22:04:50 - INFO - codeparrot_training - Step 15831: {'lr': 0.0004043748277639108, 'samples': 8105984, 'steps': 15831, 'loss/train': 2.553969144821167} 02/24/2022 22:04:56 - INFO - codeparrot_training - Step 15832: {'lr': 0.0004043619571884884, 'samples': 8106496, 'steps': 15832, 'loss/train': 2.750020980834961} 02/24/2022 22:04:59 - INFO - codeparrot_training - Step 15833: {'lr': 0.0004043490859518284, 'samples': 8107008, 'steps': 15833, 'loss/train': 1.9830418825149536} 02/24/2022 22:05:05 - INFO - codeparrot_training - Step 15834: {'lr': 0.0004043362140539859, 'samples': 8107520, 'steps': 15834, 'loss/train': 0.7290531396865845} 02/24/2022 22:05:08 - INFO - codeparrot_training - Step 15835: {'lr': 0.00040432334149501613, 'samples': 8108032, 'steps': 15835, 'loss/train': 1.781973123550415} 02/24/2022 22:05:14 - INFO - codeparrot_training - Step 15836: {'lr': 0.00040431046827497415, 'samples': 8108544, 'steps': 15836, 'loss/train': 2.0285396575927734} 02/24/2022 22:05:18 - INFO - codeparrot_training - Step 15837: {'lr': 0.00040429759439391513, 'samples': 8109056, 'steps': 15837, 'loss/train': 1.9817787408828735} 02/24/2022 22:05:23 - INFO - codeparrot_training - Step 15838: {'lr': 0.00040428471985189416, 'samples': 8109568, 'steps': 15838, 'loss/train': 2.752528190612793} 02/24/2022 22:05:27 - INFO - codeparrot_training - Step 15839: {'lr': 0.0004042718446489665, 'samples': 8110080, 'steps': 15839, 'loss/train': 1.9874250888824463} 02/24/2022 22:05:33 - INFO - codeparrot_training - Step 15840: {'lr': 0.0004042589687851872, 'samples': 8110592, 'steps': 15840, 'loss/train': 2.269784927368164} 02/24/2022 22:05:36 - INFO - codeparrot_training - Step 15841: {'lr': 0.00040424609226061146, 'samples': 8111104, 'steps': 15841, 'loss/train': 2.017077684402466} 02/24/2022 22:05:42 - INFO - codeparrot_training - Step 15842: {'lr': 0.0004042332150752944, 'samples': 8111616, 'steps': 15842, 'loss/train': 2.5844528675079346} 02/24/2022 22:05:45 - INFO - codeparrot_training - Step 15843: {'lr': 0.0004042203372292913, 'samples': 8112128, 'steps': 15843, 'loss/train': 1.9932184219360352} 02/24/2022 22:05:51 - INFO - codeparrot_training - Step 15844: {'lr': 0.00040420745872265726, 'samples': 8112640, 'steps': 15844, 'loss/train': 2.2793402671813965} 02/24/2022 22:05:54 - INFO - codeparrot_training - Step 15845: {'lr': 0.0004041945795554474, 'samples': 8113152, 'steps': 15845, 'loss/train': 0.9564520716667175} 02/24/2022 22:06:00 - INFO - codeparrot_training - Step 15846: {'lr': 0.0004041816997277169, 'samples': 8113664, 'steps': 15846, 'loss/train': 2.2356438636779785} 02/24/2022 22:06:03 - INFO - codeparrot_training - Step 15847: {'lr': 0.000404168819239521, 'samples': 8114176, 'steps': 15847, 'loss/train': 1.691184639930725} 02/24/2022 22:06:09 - INFO - codeparrot_training - Step 15848: {'lr': 0.0004041559380909148, 'samples': 8114688, 'steps': 15848, 'loss/train': 2.523899793624878} 02/24/2022 22:06:12 - INFO - codeparrot_training - Step 15849: {'lr': 0.00040414305628195347, 'samples': 8115200, 'steps': 15849, 'loss/train': 0.7413461208343506} 02/24/2022 22:06:18 - INFO - codeparrot_training - Step 15850: {'lr': 0.00040413017381269237, 'samples': 8115712, 'steps': 15850, 'loss/train': 2.901477336883545} 02/24/2022 22:06:21 - INFO - codeparrot_training - Step 15851: {'lr': 0.00040411729068318635, 'samples': 8116224, 'steps': 15851, 'loss/train': 1.7757055759429932} 02/24/2022 22:06:27 - INFO - codeparrot_training - Step 15852: {'lr': 0.0004041044068934909, 'samples': 8116736, 'steps': 15852, 'loss/train': 2.344928741455078} 02/24/2022 22:06:30 - INFO - codeparrot_training - Step 15853: {'lr': 0.00040409152244366117, 'samples': 8117248, 'steps': 15853, 'loss/train': 0.5482961535453796} 02/24/2022 22:06:36 - INFO - codeparrot_training - Step 15854: {'lr': 0.00040407863733375217, 'samples': 8117760, 'steps': 15854, 'loss/train': 0.8105165958404541} 02/24/2022 22:06:39 - INFO - codeparrot_training - Step 15855: {'lr': 0.0004040657515638193, 'samples': 8118272, 'steps': 15855, 'loss/train': 1.8301023244857788} 02/24/2022 22:06:46 - INFO - codeparrot_training - Step 15856: {'lr': 0.0004040528651339176, 'samples': 8118784, 'steps': 15856, 'loss/train': 2.081894636154175} 02/24/2022 22:06:49 - INFO - codeparrot_training - Step 15857: {'lr': 0.00040403997804410244, 'samples': 8119296, 'steps': 15857, 'loss/train': 1.9402638673782349} 02/24/2022 22:06:55 - INFO - codeparrot_training - Step 15858: {'lr': 0.00040402709029442883, 'samples': 8119808, 'steps': 15858, 'loss/train': 1.6192325353622437} 02/24/2022 22:06:58 - INFO - codeparrot_training - Step 15859: {'lr': 0.0004040142018849521, 'samples': 8120320, 'steps': 15859, 'loss/train': 2.533816337585449} 02/24/2022 22:07:03 - INFO - codeparrot_training - Step 15860: {'lr': 0.0004040013128157275, 'samples': 8120832, 'steps': 15860, 'loss/train': 2.0810694694519043} 02/24/2022 22:07:07 - INFO - codeparrot_training - Step 15861: {'lr': 0.0004039884230868101, 'samples': 8121344, 'steps': 15861, 'loss/train': 2.0571820735931396} 02/24/2022 22:07:13 - INFO - codeparrot_training - Step 15862: {'lr': 0.0004039755326982552, 'samples': 8121856, 'steps': 15862, 'loss/train': 1.2313518524169922} 02/24/2022 22:07:16 - INFO - codeparrot_training - Step 15863: {'lr': 0.000403962641650118, 'samples': 8122368, 'steps': 15863, 'loss/train': 2.065330743789673} 02/24/2022 22:07:22 - INFO - codeparrot_training - Step 15864: {'lr': 0.0004039497499424538, 'samples': 8122880, 'steps': 15864, 'loss/train': 1.1708543300628662} 02/24/2022 22:07:25 - INFO - codeparrot_training - Step 15865: {'lr': 0.00040393685757531776, 'samples': 8123392, 'steps': 15865, 'loss/train': 2.123807907104492} 02/24/2022 22:07:32 - INFO - codeparrot_training - Step 15866: {'lr': 0.000403923964548765, 'samples': 8123904, 'steps': 15866, 'loss/train': 2.3617241382598877} 02/24/2022 22:07:35 - INFO - codeparrot_training - Step 15867: {'lr': 0.0004039110708628509, 'samples': 8124416, 'steps': 15867, 'loss/train': 1.7418897151947021} 02/24/2022 22:07:41 - INFO - codeparrot_training - Step 15868: {'lr': 0.00040389817651763073, 'samples': 8124928, 'steps': 15868, 'loss/train': 1.7907893657684326} 02/24/2022 22:07:44 - INFO - codeparrot_training - Step 15869: {'lr': 0.0004038852815131595, 'samples': 8125440, 'steps': 15869, 'loss/train': 2.2889366149902344} 02/24/2022 22:07:49 - INFO - codeparrot_training - Step 15870: {'lr': 0.0004038723858494927, 'samples': 8125952, 'steps': 15870, 'loss/train': 2.0409507751464844} 02/24/2022 22:07:53 - INFO - codeparrot_training - Step 15871: {'lr': 0.00040385948952668537, 'samples': 8126464, 'steps': 15871, 'loss/train': 2.307403802871704} 02/24/2022 22:07:59 - INFO - codeparrot_training - Step 15872: {'lr': 0.0004038465925447929, 'samples': 8126976, 'steps': 15872, 'loss/train': 2.6729981899261475} 02/24/2022 22:08:02 - INFO - codeparrot_training - Step 15873: {'lr': 0.00040383369490387043, 'samples': 8127488, 'steps': 15873, 'loss/train': 3.0144176483154297} 02/24/2022 22:08:08 - INFO - codeparrot_training - Step 15874: {'lr': 0.0004038207966039733, 'samples': 8128000, 'steps': 15874, 'loss/train': 1.056424856185913} 02/24/2022 22:08:11 - INFO - codeparrot_training - Step 15875: {'lr': 0.00040380789764515667, 'samples': 8128512, 'steps': 15875, 'loss/train': 2.3025918006896973} 02/24/2022 22:08:17 - INFO - codeparrot_training - Step 15876: {'lr': 0.0004037949980274759, 'samples': 8129024, 'steps': 15876, 'loss/train': 2.5147294998168945} 02/24/2022 22:08:20 - INFO - codeparrot_training - Step 15877: {'lr': 0.0004037820977509862, 'samples': 8129536, 'steps': 15877, 'loss/train': 2.1982905864715576} 02/24/2022 22:08:26 - INFO - codeparrot_training - Step 15878: {'lr': 0.00040376919681574285, 'samples': 8130048, 'steps': 15878, 'loss/train': 2.2790567874908447} 02/24/2022 22:08:30 - INFO - codeparrot_training - Step 15879: {'lr': 0.000403756295221801, 'samples': 8130560, 'steps': 15879, 'loss/train': 1.8554037809371948} 02/24/2022 22:08:35 - INFO - codeparrot_training - Step 15880: {'lr': 0.00040374339296921606, 'samples': 8131072, 'steps': 15880, 'loss/train': 2.983825206756592} 02/24/2022 22:08:39 - INFO - codeparrot_training - Step 15881: {'lr': 0.00040373049005804323, 'samples': 8131584, 'steps': 15881, 'loss/train': 1.3516827821731567} 02/24/2022 22:08:44 - INFO - codeparrot_training - Step 15882: {'lr': 0.00040371758648833776, 'samples': 8132096, 'steps': 15882, 'loss/train': 1.6576025485992432} 02/24/2022 22:08:48 - INFO - codeparrot_training - Step 15883: {'lr': 0.00040370468226015507, 'samples': 8132608, 'steps': 15883, 'loss/train': 2.253601551055908} 02/24/2022 22:08:53 - INFO - codeparrot_training - Step 15884: {'lr': 0.0004036917773735502, 'samples': 8133120, 'steps': 15884, 'loss/train': 2.3390848636627197} 02/24/2022 22:08:57 - INFO - codeparrot_training - Step 15885: {'lr': 0.00040367887182857866, 'samples': 8133632, 'steps': 15885, 'loss/train': 1.6015355587005615} 02/24/2022 22:09:02 - INFO - codeparrot_training - Step 15886: {'lr': 0.00040366596562529554, 'samples': 8134144, 'steps': 15886, 'loss/train': 2.5205025672912598} 02/24/2022 22:09:06 - INFO - codeparrot_training - Step 15887: {'lr': 0.00040365305876375636, 'samples': 8134656, 'steps': 15887, 'loss/train': 1.9925780296325684} 02/24/2022 22:09:12 - INFO - codeparrot_training - Step 15888: {'lr': 0.0004036401512440161, 'samples': 8135168, 'steps': 15888, 'loss/train': 2.685431480407715} 02/24/2022 22:09:16 - INFO - codeparrot_training - Step 15889: {'lr': 0.0004036272430661303, 'samples': 8135680, 'steps': 15889, 'loss/train': 1.1180616617202759} 02/24/2022 22:09:21 - INFO - codeparrot_training - Step 15890: {'lr': 0.0004036143342301542, 'samples': 8136192, 'steps': 15890, 'loss/train': 1.2874945402145386} 02/24/2022 22:09:25 - INFO - codeparrot_training - Step 15891: {'lr': 0.000403601424736143, 'samples': 8136704, 'steps': 15891, 'loss/train': 2.018214464187622} 02/24/2022 22:09:30 - INFO - codeparrot_training - Step 15892: {'lr': 0.0004035885145841521, 'samples': 8137216, 'steps': 15892, 'loss/train': 1.9177649021148682} 02/24/2022 22:09:34 - INFO - codeparrot_training - Step 15893: {'lr': 0.00040357560377423675, 'samples': 8137728, 'steps': 15893, 'loss/train': 1.4360467195510864} 02/24/2022 22:09:39 - INFO - codeparrot_training - Step 15894: {'lr': 0.0004035626923064524, 'samples': 8138240, 'steps': 15894, 'loss/train': 2.937727451324463} 02/24/2022 22:09:43 - INFO - codeparrot_training - Step 15895: {'lr': 0.00040354978018085407, 'samples': 8138752, 'steps': 15895, 'loss/train': 1.301561713218689} 02/24/2022 22:09:48 - INFO - codeparrot_training - Step 15896: {'lr': 0.00040353686739749733, 'samples': 8139264, 'steps': 15896, 'loss/train': 2.101635694503784} 02/24/2022 22:09:52 - INFO - codeparrot_training - Step 15897: {'lr': 0.00040352395395643737, 'samples': 8139776, 'steps': 15897, 'loss/train': 2.8872873783111572} 02/24/2022 22:09:57 - INFO - codeparrot_training - Step 15898: {'lr': 0.00040351103985772964, 'samples': 8140288, 'steps': 15898, 'loss/train': 1.7483012676239014} 02/24/2022 22:10:01 - INFO - codeparrot_training - Step 15899: {'lr': 0.00040349812510142923, 'samples': 8140800, 'steps': 15899, 'loss/train': 1.4598948955535889} 02/24/2022 22:10:06 - INFO - codeparrot_training - Step 15900: {'lr': 0.0004034852096875916, 'samples': 8141312, 'steps': 15900, 'loss/train': 0.5955232381820679} 02/24/2022 22:10:10 - INFO - codeparrot_training - Step 15901: {'lr': 0.0004034722936162721, 'samples': 8141824, 'steps': 15901, 'loss/train': 2.769268035888672} 02/24/2022 22:10:16 - INFO - codeparrot_training - Step 15902: {'lr': 0.00040345937688752607, 'samples': 8142336, 'steps': 15902, 'loss/train': 0.09566738456487656} 02/24/2022 22:10:19 - INFO - codeparrot_training - Step 15903: {'lr': 0.0004034464595014088, 'samples': 8142848, 'steps': 15903, 'loss/train': 1.6336512565612793} 02/24/2022 22:10:25 - INFO - codeparrot_training - Step 15904: {'lr': 0.00040343354145797554, 'samples': 8143360, 'steps': 15904, 'loss/train': 1.7764639854431152} 02/24/2022 22:10:28 - INFO - codeparrot_training - Step 15905: {'lr': 0.0004034206227572818, 'samples': 8143872, 'steps': 15905, 'loss/train': 2.9088022708892822} 02/24/2022 22:10:34 - INFO - codeparrot_training - Step 15906: {'lr': 0.0004034077033993828, 'samples': 8144384, 'steps': 15906, 'loss/train': 2.1309866905212402} 02/24/2022 22:10:37 - INFO - codeparrot_training - Step 15907: {'lr': 0.00040339478338433386, 'samples': 8144896, 'steps': 15907, 'loss/train': 2.089580774307251} 02/24/2022 22:10:43 - INFO - codeparrot_training - Step 15908: {'lr': 0.0004033818627121904, 'samples': 8145408, 'steps': 15908, 'loss/train': 2.587886095046997} 02/24/2022 22:10:46 - INFO - codeparrot_training - Step 15909: {'lr': 0.00040336894138300777, 'samples': 8145920, 'steps': 15909, 'loss/train': 1.3302668333053589} 02/24/2022 22:10:52 - INFO - codeparrot_training - Step 15910: {'lr': 0.0004033560193968413, 'samples': 8146432, 'steps': 15910, 'loss/train': 1.1433871984481812} 02/24/2022 22:10:55 - INFO - codeparrot_training - Step 15911: {'lr': 0.00040334309675374636, 'samples': 8146944, 'steps': 15911, 'loss/train': 2.8243627548217773} 02/24/2022 22:11:01 - INFO - codeparrot_training - Step 15912: {'lr': 0.0004033301734537782, 'samples': 8147456, 'steps': 15912, 'loss/train': 1.226119875907898} 02/24/2022 22:11:04 - INFO - codeparrot_training - Step 15913: {'lr': 0.0004033172494969923, 'samples': 8147968, 'steps': 15913, 'loss/train': 1.8127819299697876} 02/24/2022 22:11:10 - INFO - codeparrot_training - Step 15914: {'lr': 0.000403304324883444, 'samples': 8148480, 'steps': 15914, 'loss/train': 1.9139840602874756} 02/24/2022 22:11:16 - INFO - codeparrot_training - Step 15915: {'lr': 0.00040329139961318863, 'samples': 8148992, 'steps': 15915, 'loss/train': 1.569076657295227} 02/24/2022 22:11:19 - INFO - codeparrot_training - Step 15916: {'lr': 0.00040327847368628163, 'samples': 8149504, 'steps': 15916, 'loss/train': 0.9138757586479187} 02/24/2022 22:11:25 - INFO - codeparrot_training - Step 15917: {'lr': 0.0004032655471027783, 'samples': 8150016, 'steps': 15917, 'loss/train': 1.0135868787765503} 02/24/2022 22:11:28 - INFO - codeparrot_training - Step 15918: {'lr': 0.000403252619862734, 'samples': 8150528, 'steps': 15918, 'loss/train': 2.693162441253662} 02/24/2022 22:11:34 - INFO - codeparrot_training - Step 15919: {'lr': 0.0004032396919662041, 'samples': 8151040, 'steps': 15919, 'loss/train': 2.888972043991089} 02/24/2022 22:11:37 - INFO - codeparrot_training - Step 15920: {'lr': 0.00040322676341324415, 'samples': 8151552, 'steps': 15920, 'loss/train': 1.0917152166366577} 02/24/2022 22:11:43 - INFO - codeparrot_training - Step 15921: {'lr': 0.0004032138342039093, 'samples': 8152064, 'steps': 15921, 'loss/train': 1.5159212350845337} 02/24/2022 22:11:46 - INFO - codeparrot_training - Step 15922: {'lr': 0.0004032009043382551, 'samples': 8152576, 'steps': 15922, 'loss/train': 2.320605754852295} 02/24/2022 22:11:53 - INFO - codeparrot_training - Step 15923: {'lr': 0.0004031879738163368, 'samples': 8153088, 'steps': 15923, 'loss/train': 2.4609010219573975} 02/24/2022 22:11:56 - INFO - codeparrot_training - Step 15924: {'lr': 0.00040317504263820994, 'samples': 8153600, 'steps': 15924, 'loss/train': 1.660744309425354} 02/24/2022 22:12:02 - INFO - codeparrot_training - Step 15925: {'lr': 0.0004031621108039298, 'samples': 8154112, 'steps': 15925, 'loss/train': 2.0532703399658203} 02/24/2022 22:12:05 - INFO - codeparrot_training - Step 15926: {'lr': 0.0004031491783135518, 'samples': 8154624, 'steps': 15926, 'loss/train': 2.271287679672241} 02/24/2022 22:12:11 - INFO - codeparrot_training - Step 15927: {'lr': 0.0004031362451671314, 'samples': 8155136, 'steps': 15927, 'loss/train': 1.370354413986206} 02/24/2022 22:12:14 - INFO - codeparrot_training - Step 15928: {'lr': 0.00040312331136472385, 'samples': 8155648, 'steps': 15928, 'loss/train': 3.042844772338867} 02/24/2022 22:12:20 - INFO - codeparrot_training - Step 15929: {'lr': 0.00040311037690638477, 'samples': 8156160, 'steps': 15929, 'loss/train': 1.244518756866455} 02/24/2022 22:12:23 - INFO - codeparrot_training - Step 15930: {'lr': 0.00040309744179216936, 'samples': 8156672, 'steps': 15930, 'loss/train': 2.078324794769287} 02/24/2022 22:12:29 - INFO - codeparrot_training - Step 15931: {'lr': 0.0004030845060221332, 'samples': 8157184, 'steps': 15931, 'loss/train': 1.725034475326538} 02/24/2022 22:12:32 - INFO - codeparrot_training - Step 15932: {'lr': 0.00040307156959633154, 'samples': 8157696, 'steps': 15932, 'loss/train': 2.2365212440490723} 02/24/2022 22:12:38 - INFO - codeparrot_training - Step 15933: {'lr': 0.00040305863251482, 'samples': 8158208, 'steps': 15933, 'loss/train': 1.8629781007766724} 02/24/2022 22:12:42 - INFO - codeparrot_training - Step 15934: {'lr': 0.00040304569477765375, 'samples': 8158720, 'steps': 15934, 'loss/train': 1.8882840871810913} 02/24/2022 22:12:47 - INFO - codeparrot_training - Step 15935: {'lr': 0.0004030327563848885, 'samples': 8159232, 'steps': 15935, 'loss/train': 1.6045397520065308} 02/24/2022 22:12:51 - INFO - codeparrot_training - Step 15936: {'lr': 0.00040301981733657934, 'samples': 8159744, 'steps': 15936, 'loss/train': 2.470729351043701} 02/24/2022 22:12:56 - INFO - codeparrot_training - Step 15937: {'lr': 0.00040300687763278196, 'samples': 8160256, 'steps': 15937, 'loss/train': 1.779634714126587} 02/24/2022 22:12:59 - INFO - codeparrot_training - Step 15938: {'lr': 0.0004029939372735517, 'samples': 8160768, 'steps': 15938, 'loss/train': 2.12949275970459} 02/24/2022 22:13:05 - INFO - codeparrot_training - Step 15939: {'lr': 0.000402980996258944, 'samples': 8161280, 'steps': 15939, 'loss/train': 2.9251275062561035} 02/24/2022 22:13:08 - INFO - codeparrot_training - Step 15940: {'lr': 0.00040296805458901427, 'samples': 8161792, 'steps': 15940, 'loss/train': 0.7486175894737244} 02/24/2022 22:13:14 - INFO - codeparrot_training - Step 15941: {'lr': 0.0004029551122638179, 'samples': 8162304, 'steps': 15941, 'loss/train': 2.53194522857666} 02/24/2022 22:13:17 - INFO - codeparrot_training - Step 15942: {'lr': 0.0004029421692834105, 'samples': 8162816, 'steps': 15942, 'loss/train': 1.3122670650482178} 02/24/2022 22:13:23 - INFO - codeparrot_training - Step 15943: {'lr': 0.0004029292256478474, 'samples': 8163328, 'steps': 15943, 'loss/train': 2.271367311477661} 02/24/2022 22:13:26 - INFO - codeparrot_training - Step 15944: {'lr': 0.00040291628135718404, 'samples': 8163840, 'steps': 15944, 'loss/train': 1.7397887706756592} 02/24/2022 22:13:32 - INFO - codeparrot_training - Step 15945: {'lr': 0.0004029033364114759, 'samples': 8164352, 'steps': 15945, 'loss/train': 2.0750691890716553} 02/24/2022 22:13:36 - INFO - codeparrot_training - Step 15946: {'lr': 0.00040289039081077837, 'samples': 8164864, 'steps': 15946, 'loss/train': 2.336592674255371} 02/24/2022 22:13:42 - INFO - codeparrot_training - Step 15947: {'lr': 0.00040287744455514703, 'samples': 8165376, 'steps': 15947, 'loss/train': 1.6677109003067017} 02/24/2022 22:13:46 - INFO - codeparrot_training - Step 15948: {'lr': 0.00040286449764463715, 'samples': 8165888, 'steps': 15948, 'loss/train': 1.7711118459701538} 02/24/2022 22:13:51 - INFO - codeparrot_training - Step 15949: {'lr': 0.0004028515500793044, 'samples': 8166400, 'steps': 15949, 'loss/train': 2.3861563205718994} 02/24/2022 22:13:55 - INFO - codeparrot_training - Step 15950: {'lr': 0.0004028386018592041, 'samples': 8166912, 'steps': 15950, 'loss/train': 2.4803884029388428} 02/24/2022 22:14:00 - INFO - codeparrot_training - Step 15951: {'lr': 0.0004028256529843918, 'samples': 8167424, 'steps': 15951, 'loss/train': 1.5475083589553833} 02/24/2022 22:14:04 - INFO - codeparrot_training - Step 15952: {'lr': 0.00040281270345492295, 'samples': 8167936, 'steps': 15952, 'loss/train': 3.054276704788208} 02/24/2022 22:14:09 - INFO - codeparrot_training - Step 15953: {'lr': 0.00040279975327085294, 'samples': 8168448, 'steps': 15953, 'loss/train': 0.556505024433136} 02/24/2022 22:14:13 - INFO - codeparrot_training - Step 15954: {'lr': 0.00040278680243223733, 'samples': 8168960, 'steps': 15954, 'loss/train': 1.144440770149231} 02/24/2022 22:14:18 - INFO - codeparrot_training - Step 15955: {'lr': 0.00040277385093913154, 'samples': 8169472, 'steps': 15955, 'loss/train': 2.0256495475769043} 02/24/2022 22:14:22 - INFO - codeparrot_training - Step 15956: {'lr': 0.0004027608987915912, 'samples': 8169984, 'steps': 15956, 'loss/train': 2.6622085571289062} 02/24/2022 22:14:27 - INFO - codeparrot_training - Step 15957: {'lr': 0.0004027479459896716, 'samples': 8170496, 'steps': 15957, 'loss/train': 1.8714369535446167} 02/24/2022 22:14:31 - INFO - codeparrot_training - Step 15958: {'lr': 0.0004027349925334282, 'samples': 8171008, 'steps': 15958, 'loss/train': 2.069997549057007} 02/24/2022 22:14:38 - INFO - codeparrot_training - Step 15959: {'lr': 0.00040272203842291676, 'samples': 8171520, 'steps': 15959, 'loss/train': 2.0072083473205566} 02/24/2022 22:14:41 - INFO - codeparrot_training - Step 15960: {'lr': 0.00040270908365819247, 'samples': 8172032, 'steps': 15960, 'loss/train': 1.9171329736709595} 02/24/2022 22:14:47 - INFO - codeparrot_training - Step 15961: {'lr': 0.000402696128239311, 'samples': 8172544, 'steps': 15961, 'loss/train': 1.3928046226501465} 02/24/2022 22:14:50 - INFO - codeparrot_training - Step 15962: {'lr': 0.00040268317216632783, 'samples': 8173056, 'steps': 15962, 'loss/train': 2.8390164375305176} 02/24/2022 22:14:56 - INFO - codeparrot_training - Step 15963: {'lr': 0.0004026702154392984, 'samples': 8173568, 'steps': 15963, 'loss/train': 1.405389428138733} 02/24/2022 22:14:59 - INFO - codeparrot_training - Step 15964: {'lr': 0.0004026572580582783, 'samples': 8174080, 'steps': 15964, 'loss/train': 1.5241801738739014} 02/24/2022 22:15:05 - INFO - codeparrot_training - Step 15965: {'lr': 0.000402644300023323, 'samples': 8174592, 'steps': 15965, 'loss/train': 2.094996929168701} 02/24/2022 22:15:08 - INFO - codeparrot_training - Step 15966: {'lr': 0.0004026313413344879, 'samples': 8175104, 'steps': 15966, 'loss/train': 1.3782401084899902} 02/24/2022 22:15:14 - INFO - codeparrot_training - Step 15967: {'lr': 0.0004026183819918286, 'samples': 8175616, 'steps': 15967, 'loss/train': 2.631969690322876} 02/24/2022 22:15:17 - INFO - codeparrot_training - Step 15968: {'lr': 0.00040260542199540064, 'samples': 8176128, 'steps': 15968, 'loss/train': 2.5334632396698} 02/24/2022 22:15:24 - INFO - codeparrot_training - Step 15969: {'lr': 0.00040259246134525953, 'samples': 8176640, 'steps': 15969, 'loss/train': 1.8797965049743652} 02/24/2022 22:15:28 - INFO - codeparrot_training - Step 15970: {'lr': 0.0004025795000414608, 'samples': 8177152, 'steps': 15970, 'loss/train': 2.752537250518799} 02/24/2022 22:15:33 - INFO - codeparrot_training - Step 15971: {'lr': 0.0004025665380840599, 'samples': 8177664, 'steps': 15971, 'loss/train': 2.20194411277771} 02/24/2022 22:15:37 - INFO - codeparrot_training - Step 15972: {'lr': 0.00040255357547311235, 'samples': 8178176, 'steps': 15972, 'loss/train': 2.222989082336426} 02/24/2022 22:15:42 - INFO - codeparrot_training - Step 15973: {'lr': 0.0004025406122086738, 'samples': 8178688, 'steps': 15973, 'loss/train': 3.159116268157959} 02/24/2022 22:15:46 - INFO - codeparrot_training - Step 15974: {'lr': 0.0004025276482907996, 'samples': 8179200, 'steps': 15974, 'loss/train': 2.6125540733337402} 02/24/2022 22:15:51 - INFO - codeparrot_training - Step 15975: {'lr': 0.0004025146837195455, 'samples': 8179712, 'steps': 15975, 'loss/train': 8.890109062194824} 02/24/2022 22:15:55 - INFO - codeparrot_training - Step 15976: {'lr': 0.00040250171849496685, 'samples': 8180224, 'steps': 15976, 'loss/train': 3.0437135696411133} 02/24/2022 22:16:00 - INFO - codeparrot_training - Step 15977: {'lr': 0.0004024887526171193, 'samples': 8180736, 'steps': 15977, 'loss/train': 1.925284743309021} 02/24/2022 22:16:04 - INFO - codeparrot_training - Step 15978: {'lr': 0.0004024757860860584, 'samples': 8181248, 'steps': 15978, 'loss/train': 1.3974815607070923} 02/24/2022 22:16:10 - INFO - codeparrot_training - Step 15979: {'lr': 0.00040246281890183954, 'samples': 8181760, 'steps': 15979, 'loss/train': 0.5565561652183533} 02/24/2022 22:16:13 - INFO - codeparrot_training - Step 15980: {'lr': 0.0004024498510645185, 'samples': 8182272, 'steps': 15980, 'loss/train': 1.920519232749939} 02/24/2022 22:16:19 - INFO - codeparrot_training - Step 15981: {'lr': 0.00040243688257415064, 'samples': 8182784, 'steps': 15981, 'loss/train': 1.2565110921859741} 02/24/2022 22:16:22 - INFO - codeparrot_training - Step 15982: {'lr': 0.00040242391343079157, 'samples': 8183296, 'steps': 15982, 'loss/train': 1.9350614547729492} 02/24/2022 22:16:28 - INFO - codeparrot_training - Step 15983: {'lr': 0.00040241094363449684, 'samples': 8183808, 'steps': 15983, 'loss/train': 1.8248515129089355} 02/24/2022 22:16:31 - INFO - codeparrot_training - Step 15984: {'lr': 0.000402397973185322, 'samples': 8184320, 'steps': 15984, 'loss/train': 2.6002328395843506} 02/24/2022 22:16:37 - INFO - codeparrot_training - Step 15985: {'lr': 0.0004023850020833227, 'samples': 8184832, 'steps': 15985, 'loss/train': 2.369809627532959} 02/24/2022 22:16:40 - INFO - codeparrot_training - Step 15986: {'lr': 0.00040237203032855446, 'samples': 8185344, 'steps': 15986, 'loss/train': 1.9571726322174072} 02/24/2022 22:16:46 - INFO - codeparrot_training - Step 15987: {'lr': 0.00040235905792107275, 'samples': 8185856, 'steps': 15987, 'loss/train': 2.0061137676239014} 02/24/2022 22:16:49 - INFO - codeparrot_training - Step 15988: {'lr': 0.00040234608486093326, 'samples': 8186368, 'steps': 15988, 'loss/train': 2.181933641433716} 02/24/2022 22:16:57 - INFO - codeparrot_training - Step 15989: {'lr': 0.00040233311114819156, 'samples': 8186880, 'steps': 15989, 'loss/train': 2.0181827545166016} 02/24/2022 22:17:00 - INFO - codeparrot_training - Step 15990: {'lr': 0.00040232013678290316, 'samples': 8187392, 'steps': 15990, 'loss/train': 2.413337469100952} 02/24/2022 22:17:04 - INFO - codeparrot_training - Step 15991: {'lr': 0.0004023071617651236, 'samples': 8187904, 'steps': 15991, 'loss/train': 2.2617931365966797} 02/24/2022 22:17:09 - INFO - codeparrot_training - Step 15992: {'lr': 0.0004022941860949085, 'samples': 8188416, 'steps': 15992, 'loss/train': 2.7247140407562256} 02/24/2022 22:17:13 - INFO - codeparrot_training - Step 15993: {'lr': 0.00040228120977231355, 'samples': 8188928, 'steps': 15993, 'loss/train': 2.6344995498657227} 02/24/2022 22:17:18 - INFO - codeparrot_training - Step 15994: {'lr': 0.00040226823279739427, 'samples': 8189440, 'steps': 15994, 'loss/train': 1.873080849647522} 02/24/2022 22:17:22 - INFO - codeparrot_training - Step 15995: {'lr': 0.00040225525517020616, 'samples': 8189952, 'steps': 15995, 'loss/train': 3.998002290725708} 02/24/2022 22:17:27 - INFO - codeparrot_training - Step 15996: {'lr': 0.0004022422768908049, 'samples': 8190464, 'steps': 15996, 'loss/train': 3.5603339672088623} 02/24/2022 22:17:31 - INFO - codeparrot_training - Step 15997: {'lr': 0.00040222929795924613, 'samples': 8190976, 'steps': 15997, 'loss/train': 0.8349915146827698} 02/24/2022 22:17:36 - INFO - codeparrot_training - Step 15998: {'lr': 0.0004022163183755853, 'samples': 8191488, 'steps': 15998, 'loss/train': 1.9355260133743286} 02/24/2022 22:17:40 - INFO - codeparrot_training - Step 15999: {'lr': 0.0004022033381398781, 'samples': 8192000, 'steps': 15999, 'loss/train': 2.3557024002075195} 02/24/2022 22:17:40 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 22:17:57 - WARNING - huggingface_hub.repository - Several commits (16) will be pushed upstream. 02/24/2022 22:17:57 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 22:18:30 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 751fbfb..704f602 floral-grass-11 -> floral-grass-11 02/24/2022 22:18:38 - INFO - codeparrot_training - Step 16000: {'lr': 0.0004021903572521802, 'samples': 8192512, 'steps': 16000, 'loss/train': 1.7201647758483887} 02/24/2022 22:18:41 - INFO - codeparrot_training - Step 16001: {'lr': 0.0004021773757125471, 'samples': 8193024, 'steps': 16001, 'loss/train': 0.508444607257843} 02/24/2022 22:18:47 - INFO - codeparrot_training - Step 16002: {'lr': 0.0004021643935210344, 'samples': 8193536, 'steps': 16002, 'loss/train': 1.3964498043060303} 02/24/2022 22:18:53 - INFO - codeparrot_training - Step 16003: {'lr': 0.0004021514106776978, 'samples': 8194048, 'steps': 16003, 'loss/train': 0.9580065011978149} 02/24/2022 22:18:56 - INFO - codeparrot_training - Step 16004: {'lr': 0.00040213842718259287, 'samples': 8194560, 'steps': 16004, 'loss/train': 1.0566712617874146} 02/24/2022 22:19:02 - INFO - codeparrot_training - Step 16005: {'lr': 0.00040212544303577525, 'samples': 8195072, 'steps': 16005, 'loss/train': 2.0069048404693604} 02/24/2022 22:19:05 - INFO - codeparrot_training - Step 16006: {'lr': 0.00040211245823730047, 'samples': 8195584, 'steps': 16006, 'loss/train': 2.917506217956543} 02/24/2022 22:19:11 - INFO - codeparrot_training - Step 16007: {'lr': 0.00040209947278722425, 'samples': 8196096, 'steps': 16007, 'loss/train': 1.8989200592041016} 02/24/2022 22:19:14 - INFO - codeparrot_training - Step 16008: {'lr': 0.0004020864866856022, 'samples': 8196608, 'steps': 16008, 'loss/train': 1.8675284385681152} 02/24/2022 22:19:20 - INFO - codeparrot_training - Step 16009: {'lr': 0.0004020734999324899, 'samples': 8197120, 'steps': 16009, 'loss/train': 1.8313961029052734} 02/24/2022 22:19:24 - INFO - codeparrot_training - Step 16010: {'lr': 0.0004020605125279431, 'samples': 8197632, 'steps': 16010, 'loss/train': 2.8082687854766846} 02/24/2022 22:19:29 - INFO - codeparrot_training - Step 16011: {'lr': 0.0004020475244720173, 'samples': 8198144, 'steps': 16011, 'loss/train': 1.8834885358810425} 02/24/2022 22:19:33 - INFO - codeparrot_training - Step 16012: {'lr': 0.0004020345357647681, 'samples': 8198656, 'steps': 16012, 'loss/train': 0.8034518361091614} 02/24/2022 22:19:38 - INFO - codeparrot_training - Step 16013: {'lr': 0.0004020215464062513, 'samples': 8199168, 'steps': 16013, 'loss/train': 2.044654607772827} 02/24/2022 22:19:42 - INFO - codeparrot_training - Step 16014: {'lr': 0.0004020085563965226, 'samples': 8199680, 'steps': 16014, 'loss/train': 0.5688008666038513} 02/24/2022 22:19:47 - INFO - codeparrot_training - Step 16015: {'lr': 0.00040199556573563736, 'samples': 8200192, 'steps': 16015, 'loss/train': 2.2517144680023193} 02/24/2022 22:19:51 - INFO - codeparrot_training - Step 16016: {'lr': 0.0004019825744236514, 'samples': 8200704, 'steps': 16016, 'loss/train': 1.289293885231018} 02/24/2022 22:19:56 - INFO - codeparrot_training - Step 16017: {'lr': 0.00040196958246062033, 'samples': 8201216, 'steps': 16017, 'loss/train': 1.4459197521209717} 02/24/2022 22:20:00 - INFO - codeparrot_training - Step 16018: {'lr': 0.00040195658984659987, 'samples': 8201728, 'steps': 16018, 'loss/train': 2.841352701187134} 02/24/2022 22:20:05 - INFO - codeparrot_training - Step 16019: {'lr': 0.0004019435965816456, 'samples': 8202240, 'steps': 16019, 'loss/train': 2.743588924407959} 02/24/2022 22:20:09 - INFO - codeparrot_training - Step 16020: {'lr': 0.0004019306026658132, 'samples': 8202752, 'steps': 16020, 'loss/train': 2.453803777694702} 02/24/2022 22:20:15 - INFO - codeparrot_training - Step 16021: {'lr': 0.00040191760809915833, 'samples': 8203264, 'steps': 16021, 'loss/train': 1.5826425552368164} 02/24/2022 22:20:18 - INFO - codeparrot_training - Step 16022: {'lr': 0.00040190461288173675, 'samples': 8203776, 'steps': 16022, 'loss/train': 2.028092622756958} 02/24/2022 22:20:24 - INFO - codeparrot_training - Step 16023: {'lr': 0.000401891617013604, 'samples': 8204288, 'steps': 16023, 'loss/train': 1.5239996910095215} 02/24/2022 22:20:27 - INFO - codeparrot_training - Step 16024: {'lr': 0.00040187862049481573, 'samples': 8204800, 'steps': 16024, 'loss/train': 2.1686224937438965} 02/24/2022 22:20:33 - INFO - codeparrot_training - Step 16025: {'lr': 0.00040186562332542773, 'samples': 8205312, 'steps': 16025, 'loss/train': 2.1555397510528564} 02/24/2022 22:20:36 - INFO - codeparrot_training - Step 16026: {'lr': 0.0004018526255054956, 'samples': 8205824, 'steps': 16026, 'loss/train': 1.6400172710418701} 02/24/2022 22:20:42 - INFO - codeparrot_training - Step 16027: {'lr': 0.00040183962703507515, 'samples': 8206336, 'steps': 16027, 'loss/train': 1.959862232208252} 02/24/2022 22:20:45 - INFO - codeparrot_training - Step 16028: {'lr': 0.00040182662791422185, 'samples': 8206848, 'steps': 16028, 'loss/train': 1.8384616374969482} 02/24/2022 22:20:51 - INFO - codeparrot_training - Step 16029: {'lr': 0.0004018136281429915, 'samples': 8207360, 'steps': 16029, 'loss/train': 1.9850516319274902} 02/24/2022 22:20:55 - INFO - codeparrot_training - Step 16030: {'lr': 0.0004018006277214398, 'samples': 8207872, 'steps': 16030, 'loss/train': 2.3242154121398926} 02/24/2022 22:21:01 - INFO - codeparrot_training - Step 16031: {'lr': 0.00040178762664962235, 'samples': 8208384, 'steps': 16031, 'loss/train': 1.7208975553512573} 02/24/2022 22:21:04 - INFO - codeparrot_training - Step 16032: {'lr': 0.000401774624927595, 'samples': 8208896, 'steps': 16032, 'loss/train': 1.891169786453247} 02/24/2022 22:21:10 - INFO - codeparrot_training - Step 16033: {'lr': 0.00040176162255541325, 'samples': 8209408, 'steps': 16033, 'loss/train': 3.9602861404418945} 02/24/2022 22:21:13 - INFO - codeparrot_training - Step 16034: {'lr': 0.00040174861953313297, 'samples': 8209920, 'steps': 16034, 'loss/train': 2.547063112258911} 02/24/2022 22:21:19 - INFO - codeparrot_training - Step 16035: {'lr': 0.00040173561586080974, 'samples': 8210432, 'steps': 16035, 'loss/train': 1.5564717054367065} 02/24/2022 22:21:22 - INFO - codeparrot_training - Step 16036: {'lr': 0.0004017226115384994, 'samples': 8210944, 'steps': 16036, 'loss/train': 2.483438014984131} 02/24/2022 22:21:28 - INFO - codeparrot_training - Step 16037: {'lr': 0.00040170960656625744, 'samples': 8211456, 'steps': 16037, 'loss/train': 3.138737916946411} 02/24/2022 22:21:31 - INFO - codeparrot_training - Step 16038: {'lr': 0.00040169660094413977, 'samples': 8211968, 'steps': 16038, 'loss/train': 2.3410651683807373} 02/24/2022 22:21:37 - INFO - codeparrot_training - Step 16039: {'lr': 0.00040168359467220206, 'samples': 8212480, 'steps': 16039, 'loss/train': 2.397145986557007} 02/24/2022 22:21:40 - INFO - codeparrot_training - Step 16040: {'lr': 0.00040167058775049993, 'samples': 8212992, 'steps': 16040, 'loss/train': 2.042656898498535} 02/24/2022 22:21:46 - INFO - codeparrot_training - Step 16041: {'lr': 0.0004016575801790892, 'samples': 8213504, 'steps': 16041, 'loss/train': 1.7949374914169312} 02/24/2022 22:21:49 - INFO - codeparrot_training - Step 16042: {'lr': 0.0004016445719580256, 'samples': 8214016, 'steps': 16042, 'loss/train': 1.7485071420669556} 02/24/2022 22:21:55 - INFO - codeparrot_training - Step 16043: {'lr': 0.0004016315630873647, 'samples': 8214528, 'steps': 16043, 'loss/train': 2.3833954334259033} 02/24/2022 22:21:58 - INFO - codeparrot_training - Step 16044: {'lr': 0.00040161855356716245, 'samples': 8215040, 'steps': 16044, 'loss/train': 1.5439420938491821} 02/24/2022 22:22:05 - INFO - codeparrot_training - Step 16045: {'lr': 0.00040160554339747434, 'samples': 8215552, 'steps': 16045, 'loss/train': 1.802834153175354} 02/24/2022 22:22:08 - INFO - codeparrot_training - Step 16046: {'lr': 0.00040159253257835624, 'samples': 8216064, 'steps': 16046, 'loss/train': 2.036452531814575} 02/24/2022 22:22:14 - INFO - codeparrot_training - Step 16047: {'lr': 0.00040157952110986397, 'samples': 8216576, 'steps': 16047, 'loss/train': 2.3060808181762695} 02/24/2022 22:22:17 - INFO - codeparrot_training - Step 16048: {'lr': 0.00040156650899205305, 'samples': 8217088, 'steps': 16048, 'loss/train': 2.501120090484619} 02/24/2022 22:22:23 - INFO - codeparrot_training - Step 16049: {'lr': 0.00040155349622497937, 'samples': 8217600, 'steps': 16049, 'loss/train': 2.1082980632781982} 02/24/2022 22:22:26 - INFO - codeparrot_training - Step 16050: {'lr': 0.0004015404828086987, 'samples': 8218112, 'steps': 16050, 'loss/train': 2.3823835849761963} 02/24/2022 22:22:32 - INFO - codeparrot_training - Step 16051: {'lr': 0.0004015274687432667, 'samples': 8218624, 'steps': 16051, 'loss/train': 2.436448097229004} 02/24/2022 22:22:35 - INFO - codeparrot_training - Step 16052: {'lr': 0.0004015144540287391, 'samples': 8219136, 'steps': 16052, 'loss/train': 2.071312665939331} 02/24/2022 22:22:41 - INFO - codeparrot_training - Step 16053: {'lr': 0.00040150143866517164, 'samples': 8219648, 'steps': 16053, 'loss/train': 2.40167498588562} 02/24/2022 22:22:47 - INFO - codeparrot_training - Step 16054: {'lr': 0.0004014884226526202, 'samples': 8220160, 'steps': 16054, 'loss/train': 2.357085943222046} 02/24/2022 22:22:50 - INFO - codeparrot_training - Step 16055: {'lr': 0.0004014754059911405, 'samples': 8220672, 'steps': 16055, 'loss/train': 2.284782886505127} 02/24/2022 22:22:56 - INFO - codeparrot_training - Step 16056: {'lr': 0.0004014623886807882, 'samples': 8221184, 'steps': 16056, 'loss/train': 1.9049789905548096} 02/24/2022 22:22:59 - INFO - codeparrot_training - Step 16057: {'lr': 0.0004014493707216191, 'samples': 8221696, 'steps': 16057, 'loss/train': 2.8246161937713623} 02/24/2022 22:23:05 - INFO - codeparrot_training - Step 16058: {'lr': 0.00040143635211368903, 'samples': 8222208, 'steps': 16058, 'loss/train': 3.0233166217803955} 02/24/2022 22:23:08 - INFO - codeparrot_training - Step 16059: {'lr': 0.0004014233328570537, 'samples': 8222720, 'steps': 16059, 'loss/train': 2.371473550796509} 02/24/2022 22:23:12 - INFO - codeparrot_training - Step 16060: {'lr': 0.0004014103129517689, 'samples': 8223232, 'steps': 16060, 'loss/train': 2.972662925720215} 02/24/2022 22:23:18 - INFO - codeparrot_training - Step 16061: {'lr': 0.00040139729239789036, 'samples': 8223744, 'steps': 16061, 'loss/train': 1.79690420627594} 02/24/2022 22:23:21 - INFO - codeparrot_training - Step 16062: {'lr': 0.0004013842711954739, 'samples': 8224256, 'steps': 16062, 'loss/train': 1.8570235967636108} 02/24/2022 22:23:27 - INFO - codeparrot_training - Step 16063: {'lr': 0.0004013712493445753, 'samples': 8224768, 'steps': 16063, 'loss/train': 2.9462969303131104} 02/24/2022 22:23:30 - INFO - codeparrot_training - Step 16064: {'lr': 0.00040135822684525036, 'samples': 8225280, 'steps': 16064, 'loss/train': 1.238988995552063} 02/24/2022 22:23:36 - INFO - codeparrot_training - Step 16065: {'lr': 0.0004013452036975548, 'samples': 8225792, 'steps': 16065, 'loss/train': 2.1266133785247803} 02/24/2022 22:23:39 - INFO - codeparrot_training - Step 16066: {'lr': 0.0004013321799015445, 'samples': 8226304, 'steps': 16066, 'loss/train': 1.759783387184143} 02/24/2022 22:23:46 - INFO - codeparrot_training - Step 16067: {'lr': 0.00040131915545727517, 'samples': 8226816, 'steps': 16067, 'loss/train': 2.3940839767456055} 02/24/2022 22:23:49 - INFO - codeparrot_training - Step 16068: {'lr': 0.00040130613036480265, 'samples': 8227328, 'steps': 16068, 'loss/train': 1.8836443424224854} 02/24/2022 22:23:55 - INFO - codeparrot_training - Step 16069: {'lr': 0.0004012931046241827, 'samples': 8227840, 'steps': 16069, 'loss/train': 0.15590330958366394} 02/24/2022 22:23:58 - INFO - codeparrot_training - Step 16070: {'lr': 0.00040128007823547106, 'samples': 8228352, 'steps': 16070, 'loss/train': 2.9946024417877197} 02/24/2022 22:24:04 - INFO - codeparrot_training - Step 16071: {'lr': 0.00040126705119872367, 'samples': 8228864, 'steps': 16071, 'loss/train': 3.3204965591430664} 02/24/2022 22:24:07 - INFO - codeparrot_training - Step 16072: {'lr': 0.00040125402351399623, 'samples': 8229376, 'steps': 16072, 'loss/train': 2.1243174076080322} 02/24/2022 22:24:13 - INFO - codeparrot_training - Step 16073: {'lr': 0.0004012409951813446, 'samples': 8229888, 'steps': 16073, 'loss/train': 2.5254733562469482} 02/24/2022 22:24:18 - INFO - codeparrot_training - Step 16074: {'lr': 0.0004012279662008246, 'samples': 8230400, 'steps': 16074, 'loss/train': 2.2832231521606445} 02/24/2022 22:24:22 - INFO - codeparrot_training - Step 16075: {'lr': 0.000401214936572492, 'samples': 8230912, 'steps': 16075, 'loss/train': 1.5007426738739014} 02/24/2022 22:24:28 - INFO - codeparrot_training - Step 16076: {'lr': 0.0004012019062964026, 'samples': 8231424, 'steps': 16076, 'loss/train': 1.786483645439148} 02/24/2022 22:24:32 - INFO - codeparrot_training - Step 16077: {'lr': 0.0004011888753726123, 'samples': 8231936, 'steps': 16077, 'loss/train': 2.955906867980957} 02/24/2022 22:24:38 - INFO - codeparrot_training - Step 16078: {'lr': 0.00040117584380117675, 'samples': 8232448, 'steps': 16078, 'loss/train': 3.329847574234009} 02/24/2022 22:24:41 - INFO - codeparrot_training - Step 16079: {'lr': 0.000401162811582152, 'samples': 8232960, 'steps': 16079, 'loss/train': 1.9173682928085327} 02/24/2022 22:24:47 - INFO - codeparrot_training - Step 16080: {'lr': 0.00040114977871559375, 'samples': 8233472, 'steps': 16080, 'loss/train': 1.4451696872711182} 02/24/2022 22:24:50 - INFO - codeparrot_training - Step 16081: {'lr': 0.0004011367452015578, 'samples': 8233984, 'steps': 16081, 'loss/train': 0.7013158202171326} 02/24/2022 22:24:56 - INFO - codeparrot_training - Step 16082: {'lr': 0.00040112371104010004, 'samples': 8234496, 'steps': 16082, 'loss/train': 2.2582297325134277} 02/24/2022 22:24:59 - INFO - codeparrot_training - Step 16083: {'lr': 0.00040111067623127626, 'samples': 8235008, 'steps': 16083, 'loss/train': 1.5379586219787598} 02/24/2022 22:25:05 - INFO - codeparrot_training - Step 16084: {'lr': 0.0004010976407751424, 'samples': 8235520, 'steps': 16084, 'loss/train': 2.447960138320923} 02/24/2022 22:25:08 - INFO - codeparrot_training - Step 16085: {'lr': 0.00040108460467175425, 'samples': 8236032, 'steps': 16085, 'loss/train': 1.7054228782653809} 02/24/2022 22:25:14 - INFO - codeparrot_training - Step 16086: {'lr': 0.00040107156792116753, 'samples': 8236544, 'steps': 16086, 'loss/train': 1.2864183187484741} 02/24/2022 22:25:18 - INFO - codeparrot_training - Step 16087: {'lr': 0.0004010585305234382, 'samples': 8237056, 'steps': 16087, 'loss/train': 1.4124513864517212} 02/24/2022 22:25:21 - INFO - codeparrot_training - Step 16088: {'lr': 0.00040104549247862217, 'samples': 8237568, 'steps': 16088, 'loss/train': 2.1191415786743164} 02/24/2022 22:25:27 - INFO - codeparrot_training - Step 16089: {'lr': 0.0004010324537867751, 'samples': 8238080, 'steps': 16089, 'loss/train': 1.8204383850097656} 02/24/2022 22:25:30 - INFO - codeparrot_training - Step 16090: {'lr': 0.000401019414447953, 'samples': 8238592, 'steps': 16090, 'loss/train': 0.9256582856178284} 02/24/2022 22:25:36 - INFO - codeparrot_training - Step 16091: {'lr': 0.0004010063744622117, 'samples': 8239104, 'steps': 16091, 'loss/train': 1.932460069656372} 02/24/2022 22:25:39 - INFO - codeparrot_training - Step 16092: {'lr': 0.00040099333382960707, 'samples': 8239616, 'steps': 16092, 'loss/train': 1.5542957782745361} 02/24/2022 22:25:45 - INFO - codeparrot_training - Step 16093: {'lr': 0.00040098029255019484, 'samples': 8240128, 'steps': 16093, 'loss/train': 2.8286492824554443} 02/24/2022 22:25:49 - INFO - codeparrot_training - Step 16094: {'lr': 0.0004009672506240311, 'samples': 8240640, 'steps': 16094, 'loss/train': 1.290589451789856} 02/24/2022 22:25:54 - INFO - codeparrot_training - Step 16095: {'lr': 0.00040095420805117153, 'samples': 8241152, 'steps': 16095, 'loss/train': 1.6316267251968384} 02/24/2022 22:26:00 - INFO - codeparrot_training - Step 16096: {'lr': 0.0004009411648316721, 'samples': 8241664, 'steps': 16096, 'loss/train': 2.666778564453125} 02/24/2022 22:26:03 - INFO - codeparrot_training - Step 16097: {'lr': 0.0004009281209655886, 'samples': 8242176, 'steps': 16097, 'loss/train': 3.057589292526245} 02/24/2022 22:26:09 - INFO - codeparrot_training - Step 16098: {'lr': 0.000400915076452977, 'samples': 8242688, 'steps': 16098, 'loss/train': 2.0071325302124023} 02/24/2022 22:26:12 - INFO - codeparrot_training - Step 16099: {'lr': 0.0004009020312938931, 'samples': 8243200, 'steps': 16099, 'loss/train': 1.0506532192230225} 02/24/2022 22:26:16 - INFO - codeparrot_training - Step 16100: {'lr': 0.0004008889854883929, 'samples': 8243712, 'steps': 16100, 'loss/train': 2.901155471801758} 02/24/2022 22:26:22 - INFO - codeparrot_training - Step 16101: {'lr': 0.0004008759390365321, 'samples': 8244224, 'steps': 16101, 'loss/train': 2.186505079269409} 02/24/2022 22:26:25 - INFO - codeparrot_training - Step 16102: {'lr': 0.00040086289193836674, 'samples': 8244736, 'steps': 16102, 'loss/train': 2.6797337532043457} 02/24/2022 22:26:33 - INFO - codeparrot_training - Step 16103: {'lr': 0.00040084984419395264, 'samples': 8245248, 'steps': 16103, 'loss/train': 2.0044150352478027} 02/24/2022 22:26:36 - INFO - codeparrot_training - Step 16104: {'lr': 0.00040083679580334565, 'samples': 8245760, 'steps': 16104, 'loss/train': 2.3971686363220215} 02/24/2022 22:26:42 - INFO - codeparrot_training - Step 16105: {'lr': 0.00040082374676660176, 'samples': 8246272, 'steps': 16105, 'loss/train': 2.8351309299468994} 02/24/2022 22:26:45 - INFO - codeparrot_training - Step 16106: {'lr': 0.00040081069708377686, 'samples': 8246784, 'steps': 16106, 'loss/train': 2.7092063426971436} 02/24/2022 22:26:51 - INFO - codeparrot_training - Step 16107: {'lr': 0.0004007976467549268, 'samples': 8247296, 'steps': 16107, 'loss/train': 3.106229066848755} 02/24/2022 22:26:54 - INFO - codeparrot_training - Step 16108: {'lr': 0.0004007845957801075, 'samples': 8247808, 'steps': 16108, 'loss/train': 2.0391921997070312} 02/24/2022 22:27:00 - INFO - codeparrot_training - Step 16109: {'lr': 0.0004007715441593749, 'samples': 8248320, 'steps': 16109, 'loss/train': 2.3040809631347656} 02/24/2022 22:27:03 - INFO - codeparrot_training - Step 16110: {'lr': 0.0004007584918927849, 'samples': 8248832, 'steps': 16110, 'loss/train': 1.5427764654159546} 02/24/2022 22:27:10 - INFO - codeparrot_training - Step 16111: {'lr': 0.0004007454389803933, 'samples': 8249344, 'steps': 16111, 'loss/train': 2.039019823074341} 02/24/2022 22:27:13 - INFO - codeparrot_training - Step 16112: {'lr': 0.00040073238542225623, 'samples': 8249856, 'steps': 16112, 'loss/train': 1.693688988685608} 02/24/2022 22:27:19 - INFO - codeparrot_training - Step 16113: {'lr': 0.00040071933121842943, 'samples': 8250368, 'steps': 16113, 'loss/train': 2.0680792331695557} 02/24/2022 22:27:23 - INFO - codeparrot_training - Step 16114: {'lr': 0.00040070627636896886, 'samples': 8250880, 'steps': 16114, 'loss/train': 3.236002206802368} 02/24/2022 22:27:26 - INFO - codeparrot_training - Step 16115: {'lr': 0.0004006932208739304, 'samples': 8251392, 'steps': 16115, 'loss/train': 1.1472545862197876} 02/24/2022 22:27:32 - INFO - codeparrot_training - Step 16116: {'lr': 0.0004006801647333701, 'samples': 8251904, 'steps': 16116, 'loss/train': 1.2056994438171387} 02/24/2022 22:27:35 - INFO - codeparrot_training - Step 16117: {'lr': 0.0004006671079473438, 'samples': 8252416, 'steps': 16117, 'loss/train': 0.5380105972290039} 02/24/2022 22:27:41 - INFO - codeparrot_training - Step 16118: {'lr': 0.00040065405051590745, 'samples': 8252928, 'steps': 16118, 'loss/train': 1.1405670642852783} 02/24/2022 22:27:44 - INFO - codeparrot_training - Step 16119: {'lr': 0.000400640992439117, 'samples': 8253440, 'steps': 16119, 'loss/train': 1.7501696348190308} 02/24/2022 22:27:50 - INFO - codeparrot_training - Step 16120: {'lr': 0.0004006279337170283, 'samples': 8253952, 'steps': 16120, 'loss/train': 1.8721909523010254} 02/24/2022 22:27:53 - INFO - codeparrot_training - Step 16121: {'lr': 0.00040061487434969744, 'samples': 8254464, 'steps': 16121, 'loss/train': 0.3439841568470001} 02/24/2022 22:27:59 - INFO - codeparrot_training - Step 16122: {'lr': 0.00040060181433718037, 'samples': 8254976, 'steps': 16122, 'loss/train': 2.461881160736084} 02/24/2022 22:28:02 - INFO - codeparrot_training - Step 16123: {'lr': 0.00040058875367953285, 'samples': 8255488, 'steps': 16123, 'loss/train': 2.0427472591400146} 02/24/2022 22:28:09 - INFO - codeparrot_training - Step 16124: {'lr': 0.0004005756923768109, 'samples': 8256000, 'steps': 16124, 'loss/train': 3.9292702674865723} 02/24/2022 22:28:12 - INFO - codeparrot_training - Step 16125: {'lr': 0.0004005626304290705, 'samples': 8256512, 'steps': 16125, 'loss/train': 2.3789103031158447} 02/24/2022 22:28:18 - INFO - codeparrot_training - Step 16126: {'lr': 0.00040054956783636765, 'samples': 8257024, 'steps': 16126, 'loss/train': 1.660765528678894} 02/24/2022 22:28:21 - INFO - codeparrot_training - Step 16127: {'lr': 0.00040053650459875823, 'samples': 8257536, 'steps': 16127, 'loss/train': 1.7264443635940552} 02/24/2022 22:28:27 - INFO - codeparrot_training - Step 16128: {'lr': 0.0004005234407162982, 'samples': 8258048, 'steps': 16128, 'loss/train': 1.4429055452346802} 02/24/2022 22:28:30 - INFO - codeparrot_training - Step 16129: {'lr': 0.00040051037618904365, 'samples': 8258560, 'steps': 16129, 'loss/train': 0.1482071429491043} 02/24/2022 22:28:36 - INFO - codeparrot_training - Step 16130: {'lr': 0.0004004973110170503, 'samples': 8259072, 'steps': 16130, 'loss/train': 2.3455629348754883} 02/24/2022 22:28:39 - INFO - codeparrot_training - Step 16131: {'lr': 0.0004004842452003743, 'samples': 8259584, 'steps': 16131, 'loss/train': 2.281982660293579} 02/24/2022 22:28:45 - INFO - codeparrot_training - Step 16132: {'lr': 0.0004004711787390716, 'samples': 8260096, 'steps': 16132, 'loss/train': 2.7971956729888916} 02/24/2022 22:28:48 - INFO - codeparrot_training - Step 16133: {'lr': 0.0004004581116331981, 'samples': 8260608, 'steps': 16133, 'loss/train': 1.4828740358352661} 02/24/2022 22:28:54 - INFO - codeparrot_training - Step 16134: {'lr': 0.00040044504388280996, 'samples': 8261120, 'steps': 16134, 'loss/train': 2.1721882820129395} 02/24/2022 22:28:57 - INFO - codeparrot_training - Step 16135: {'lr': 0.00040043197548796295, 'samples': 8261632, 'steps': 16135, 'loss/train': 1.853606104850769} 02/24/2022 22:29:03 - INFO - codeparrot_training - Step 16136: {'lr': 0.0004004189064487131, 'samples': 8262144, 'steps': 16136, 'loss/train': 1.6603351831436157} 02/24/2022 22:29:06 - INFO - codeparrot_training - Step 16137: {'lr': 0.00040040583676511645, 'samples': 8262656, 'steps': 16137, 'loss/train': 2.022278070449829} 02/24/2022 22:29:12 - INFO - codeparrot_training - Step 16138: {'lr': 0.0004003927664372289, 'samples': 8263168, 'steps': 16138, 'loss/train': 2.8447370529174805} 02/24/2022 22:29:15 - INFO - codeparrot_training - Step 16139: {'lr': 0.00040037969546510653, 'samples': 8263680, 'steps': 16139, 'loss/train': 1.3555209636688232} 02/24/2022 22:29:21 - INFO - codeparrot_training - Step 16140: {'lr': 0.0004003666238488053, 'samples': 8264192, 'steps': 16140, 'loss/train': 0.3209831118583679} 02/24/2022 22:29:25 - INFO - codeparrot_training - Step 16141: {'lr': 0.00040035355158838114, 'samples': 8264704, 'steps': 16141, 'loss/train': 1.6687036752700806} 02/24/2022 22:29:30 - INFO - codeparrot_training - Step 16142: {'lr': 0.0004003404786838902, 'samples': 8265216, 'steps': 16142, 'loss/train': 1.3506438732147217} 02/24/2022 22:29:34 - INFO - codeparrot_training - Step 16143: {'lr': 0.0004003274051353884, 'samples': 8265728, 'steps': 16143, 'loss/train': 1.289398431777954} 02/24/2022 22:29:40 - INFO - codeparrot_training - Step 16144: {'lr': 0.00040031433094293167, 'samples': 8266240, 'steps': 16144, 'loss/train': 1.6556278467178345} 02/24/2022 22:29:43 - INFO - codeparrot_training - Step 16145: {'lr': 0.0004003012561065761, 'samples': 8266752, 'steps': 16145, 'loss/train': 2.8677425384521484} 02/24/2022 22:29:49 - INFO - codeparrot_training - Step 16146: {'lr': 0.0004002881806263776, 'samples': 8267264, 'steps': 16146, 'loss/train': 3.008291006088257} 02/24/2022 22:29:52 - INFO - codeparrot_training - Step 16147: {'lr': 0.0004002751045023924, 'samples': 8267776, 'steps': 16147, 'loss/train': 2.081298828125} 02/24/2022 22:29:58 - INFO - codeparrot_training - Step 16148: {'lr': 0.00040026202773467623, 'samples': 8268288, 'steps': 16148, 'loss/train': 3.3766801357269287} 02/24/2022 22:30:01 - INFO - codeparrot_training - Step 16149: {'lr': 0.00040024895032328536, 'samples': 8268800, 'steps': 16149, 'loss/train': 0.7873048186302185} 02/24/2022 22:30:07 - INFO - codeparrot_training - Step 16150: {'lr': 0.0004002358722682756, 'samples': 8269312, 'steps': 16150, 'loss/train': 2.3426740169525146} 02/24/2022 22:30:11 - INFO - codeparrot_training - Step 16151: {'lr': 0.00040022279356970316, 'samples': 8269824, 'steps': 16151, 'loss/train': 1.2303619384765625} 02/24/2022 22:30:16 - INFO - codeparrot_training - Step 16152: {'lr': 0.0004002097142276239, 'samples': 8270336, 'steps': 16152, 'loss/train': 4.20693302154541} 02/24/2022 22:30:20 - INFO - codeparrot_training - Step 16153: {'lr': 0.00040019663424209397, 'samples': 8270848, 'steps': 16153, 'loss/train': 1.566499948501587} 02/24/2022 22:30:25 - INFO - codeparrot_training - Step 16154: {'lr': 0.0004001835536131693, 'samples': 8271360, 'steps': 16154, 'loss/train': 3.4570751190185547} 02/24/2022 22:30:29 - INFO - codeparrot_training - Step 16155: {'lr': 0.00040017047234090596, 'samples': 8271872, 'steps': 16155, 'loss/train': 2.058189630508423} 02/24/2022 22:30:34 - INFO - codeparrot_training - Step 16156: {'lr': 0.00040015739042536, 'samples': 8272384, 'steps': 16156, 'loss/train': 2.1488616466522217} 02/24/2022 22:30:38 - INFO - codeparrot_training - Step 16157: {'lr': 0.00040014430786658754, 'samples': 8272896, 'steps': 16157, 'loss/train': 1.2444789409637451} 02/24/2022 22:30:43 - INFO - codeparrot_training - Step 16158: {'lr': 0.0004001312246646446, 'samples': 8273408, 'steps': 16158, 'loss/train': 0.3768030107021332} 02/24/2022 22:30:47 - INFO - codeparrot_training - Step 16159: {'lr': 0.000400118140819587, 'samples': 8273920, 'steps': 16159, 'loss/train': 2.8664608001708984} 02/24/2022 22:30:53 - INFO - codeparrot_training - Step 16160: {'lr': 0.00040010505633147106, 'samples': 8274432, 'steps': 16160, 'loss/train': 2.529768228530884} 02/24/2022 22:30:57 - INFO - codeparrot_training - Step 16161: {'lr': 0.0004000919712003526, 'samples': 8274944, 'steps': 16161, 'loss/train': 2.0496413707733154} 02/24/2022 22:31:02 - INFO - codeparrot_training - Step 16162: {'lr': 0.0004000788854262879, 'samples': 8275456, 'steps': 16162, 'loss/train': 2.005619764328003} 02/24/2022 22:31:06 - INFO - codeparrot_training - Step 16163: {'lr': 0.00040006579900933294, 'samples': 8275968, 'steps': 16163, 'loss/train': 3.9936861991882324} 02/24/2022 22:31:11 - INFO - codeparrot_training - Step 16164: {'lr': 0.00040005271194954367, 'samples': 8276480, 'steps': 16164, 'loss/train': 2.508627414703369} 02/24/2022 22:31:15 - INFO - codeparrot_training - Step 16165: {'lr': 0.00040003962424697625, 'samples': 8276992, 'steps': 16165, 'loss/train': 2.586475133895874} 02/24/2022 22:31:20 - INFO - codeparrot_training - Step 16166: {'lr': 0.0004000265359016867, 'samples': 8277504, 'steps': 16166, 'loss/train': 1.5947999954223633} 02/24/2022 22:31:24 - INFO - codeparrot_training - Step 16167: {'lr': 0.0004000134469137312, 'samples': 8278016, 'steps': 16167, 'loss/train': 2.631283760070801} 02/24/2022 22:31:30 - INFO - codeparrot_training - Step 16168: {'lr': 0.00040000035728316564, 'samples': 8278528, 'steps': 16168, 'loss/train': 1.6275321245193481} 02/24/2022 22:31:33 - INFO - codeparrot_training - Step 16169: {'lr': 0.0003999872670100462, 'samples': 8279040, 'steps': 16169, 'loss/train': 2.1954967975616455} 02/24/2022 22:31:40 - INFO - codeparrot_training - Step 16170: {'lr': 0.000399974176094429, 'samples': 8279552, 'steps': 16170, 'loss/train': 2.0370078086853027} 02/24/2022 22:31:43 - INFO - codeparrot_training - Step 16171: {'lr': 0.00039996108453637, 'samples': 8280064, 'steps': 16171, 'loss/train': 2.3156166076660156} 02/24/2022 22:31:48 - INFO - codeparrot_training - Step 16172: {'lr': 0.0003999479923359253, 'samples': 8280576, 'steps': 16172, 'loss/train': 1.9278333187103271} 02/24/2022 22:31:52 - INFO - codeparrot_training - Step 16173: {'lr': 0.00039993489949315103, 'samples': 8281088, 'steps': 16173, 'loss/train': 1.853058099746704} 02/24/2022 22:31:57 - INFO - codeparrot_training - Step 16174: {'lr': 0.0003999218060081032, 'samples': 8281600, 'steps': 16174, 'loss/train': 2.7687489986419678} 02/24/2022 22:32:01 - INFO - codeparrot_training - Step 16175: {'lr': 0.0003999087118808381, 'samples': 8282112, 'steps': 16175, 'loss/train': 2.3673040866851807} 02/24/2022 22:32:07 - INFO - codeparrot_training - Step 16176: {'lr': 0.0003998956171114116, 'samples': 8282624, 'steps': 16176, 'loss/train': 1.9622160196304321} 02/24/2022 22:32:10 - INFO - codeparrot_training - Step 16177: {'lr': 0.0003998825216998799, 'samples': 8283136, 'steps': 16177, 'loss/train': 1.7406855821609497} 02/24/2022 22:32:15 - INFO - codeparrot_training - Step 16178: {'lr': 0.00039986942564629904, 'samples': 8283648, 'steps': 16178, 'loss/train': 1.0941768884658813} 02/24/2022 22:32:19 - INFO - codeparrot_training - Step 16179: {'lr': 0.0003998563289507251, 'samples': 8284160, 'steps': 16179, 'loss/train': 1.4929808378219604} 02/24/2022 22:32:25 - INFO - codeparrot_training - Step 16180: {'lr': 0.0003998432316132143, 'samples': 8284672, 'steps': 16180, 'loss/train': 2.450066089630127} 02/24/2022 22:32:28 - INFO - codeparrot_training - Step 16181: {'lr': 0.0003998301336338227, 'samples': 8285184, 'steps': 16181, 'loss/train': 1.1138701438903809} 02/24/2022 22:32:34 - INFO - codeparrot_training - Step 16182: {'lr': 0.0003998170350126064, 'samples': 8285696, 'steps': 16182, 'loss/train': 1.7946586608886719} 02/24/2022 22:32:37 - INFO - codeparrot_training - Step 16183: {'lr': 0.0003998039357496214, 'samples': 8286208, 'steps': 16183, 'loss/train': 1.4287916421890259} 02/24/2022 22:32:43 - INFO - codeparrot_training - Step 16184: {'lr': 0.000399790835844924, 'samples': 8286720, 'steps': 16184, 'loss/train': 1.0363694429397583} 02/24/2022 22:32:46 - INFO - codeparrot_training - Step 16185: {'lr': 0.00039977773529857016, 'samples': 8287232, 'steps': 16185, 'loss/train': 4.024621963500977} 02/24/2022 22:32:53 - INFO - codeparrot_training - Step 16186: {'lr': 0.00039976463411061606, 'samples': 8287744, 'steps': 16186, 'loss/train': 1.800007939338684} 02/24/2022 22:32:56 - INFO - codeparrot_training - Step 16187: {'lr': 0.00039975153228111784, 'samples': 8288256, 'steps': 16187, 'loss/train': 1.5483638048171997} 02/24/2022 22:33:02 - INFO - codeparrot_training - Step 16188: {'lr': 0.0003997384298101316, 'samples': 8288768, 'steps': 16188, 'loss/train': 1.9315979480743408} 02/24/2022 22:33:05 - INFO - codeparrot_training - Step 16189: {'lr': 0.0003997253266977135, 'samples': 8289280, 'steps': 16189, 'loss/train': 2.8159971237182617} 02/24/2022 22:33:11 - INFO - codeparrot_training - Step 16190: {'lr': 0.0003997122229439196, 'samples': 8289792, 'steps': 16190, 'loss/train': 1.053288459777832} 02/24/2022 22:33:14 - INFO - codeparrot_training - Step 16191: {'lr': 0.00039969911854880613, 'samples': 8290304, 'steps': 16191, 'loss/train': 2.3093559741973877} 02/24/2022 22:33:20 - INFO - codeparrot_training - Step 16192: {'lr': 0.0003996860135124292, 'samples': 8290816, 'steps': 16192, 'loss/train': 3.026240825653076} 02/24/2022 22:33:23 - INFO - codeparrot_training - Step 16193: {'lr': 0.00039967290783484485, 'samples': 8291328, 'steps': 16193, 'loss/train': 1.3618230819702148} 02/24/2022 22:33:29 - INFO - codeparrot_training - Step 16194: {'lr': 0.00039965980151610925, 'samples': 8291840, 'steps': 16194, 'loss/train': 0.6006201505661011} 02/24/2022 22:33:32 - INFO - codeparrot_training - Step 16195: {'lr': 0.0003996466945562787, 'samples': 8292352, 'steps': 16195, 'loss/train': 2.187267303466797} 02/24/2022 22:33:38 - INFO - codeparrot_training - Step 16196: {'lr': 0.00039963358695540907, 'samples': 8292864, 'steps': 16196, 'loss/train': 1.5534999370574951} 02/24/2022 22:33:42 - INFO - codeparrot_training - Step 16197: {'lr': 0.00039962047871355686, 'samples': 8293376, 'steps': 16197, 'loss/train': 1.945918083190918} 02/24/2022 22:33:47 - INFO - codeparrot_training - Step 16198: {'lr': 0.00039960736983077783, 'samples': 8293888, 'steps': 16198, 'loss/train': 2.2027645111083984} 02/24/2022 22:33:51 - INFO - codeparrot_training - Step 16199: {'lr': 0.0003995942603071285, 'samples': 8294400, 'steps': 16199, 'loss/train': 2.0365982055664062} 02/24/2022 22:33:56 - INFO - codeparrot_training - Step 16200: {'lr': 0.0003995811501426648, 'samples': 8294912, 'steps': 16200, 'loss/train': 1.5766929388046265} 02/24/2022 22:34:00 - INFO - codeparrot_training - Step 16201: {'lr': 0.0003995680393374429, 'samples': 8295424, 'steps': 16201, 'loss/train': 2.29494309425354} 02/24/2022 22:34:05 - INFO - codeparrot_training - Step 16202: {'lr': 0.00039955492789151904, 'samples': 8295936, 'steps': 16202, 'loss/train': 1.6093108654022217} 02/24/2022 22:34:09 - INFO - codeparrot_training - Step 16203: {'lr': 0.0003995418158049494, 'samples': 8296448, 'steps': 16203, 'loss/train': 1.9988603591918945} 02/24/2022 22:34:14 - INFO - codeparrot_training - Step 16204: {'lr': 0.0003995287030777901, 'samples': 8296960, 'steps': 16204, 'loss/train': 2.9860129356384277} 02/24/2022 22:34:18 - INFO - codeparrot_training - Step 16205: {'lr': 0.0003995155897100973, 'samples': 8297472, 'steps': 16205, 'loss/train': 1.9104325771331787} 02/24/2022 22:34:24 - INFO - codeparrot_training - Step 16206: {'lr': 0.0003995024757019272, 'samples': 8297984, 'steps': 16206, 'loss/train': 2.1127588748931885} 02/24/2022 22:34:27 - INFO - codeparrot_training - Step 16207: {'lr': 0.00039948936105333593, 'samples': 8298496, 'steps': 16207, 'loss/train': 0.9390119314193726} 02/24/2022 22:34:33 - INFO - codeparrot_training - Step 16208: {'lr': 0.0003994762457643797, 'samples': 8299008, 'steps': 16208, 'loss/train': 2.228416681289673} 02/24/2022 22:34:36 - INFO - codeparrot_training - Step 16209: {'lr': 0.0003994631298351148, 'samples': 8299520, 'steps': 16209, 'loss/train': 2.1857070922851562} 02/24/2022 22:34:42 - INFO - codeparrot_training - Step 16210: {'lr': 0.0003994500132655972, 'samples': 8300032, 'steps': 16210, 'loss/train': 1.95086669921875} 02/24/2022 22:34:45 - INFO - codeparrot_training - Step 16211: {'lr': 0.0003994368960558832, 'samples': 8300544, 'steps': 16211, 'loss/train': 2.3637008666992188} 02/24/2022 22:34:51 - INFO - codeparrot_training - Step 16212: {'lr': 0.0003994237782060291, 'samples': 8301056, 'steps': 16212, 'loss/train': 2.4839963912963867} 02/24/2022 22:34:54 - INFO - codeparrot_training - Step 16213: {'lr': 0.00039941065971609084, 'samples': 8301568, 'steps': 16213, 'loss/train': 0.7934821844100952} 02/24/2022 22:35:00 - INFO - codeparrot_training - Step 16214: {'lr': 0.00039939754058612487, 'samples': 8302080, 'steps': 16214, 'loss/train': 2.159456253051758} 02/24/2022 22:35:06 - INFO - codeparrot_training - Step 16215: {'lr': 0.0003993844208161872, 'samples': 8302592, 'steps': 16215, 'loss/train': 2.200657367706299} 02/24/2022 22:35:09 - INFO - codeparrot_training - Step 16216: {'lr': 0.0003993713004063341, 'samples': 8303104, 'steps': 16216, 'loss/train': 0.35117772221565247} 02/24/2022 22:35:15 - INFO - codeparrot_training - Step 16217: {'lr': 0.0003993581793566219, 'samples': 8303616, 'steps': 16217, 'loss/train': 1.3549160957336426} 02/24/2022 22:35:19 - INFO - codeparrot_training - Step 16218: {'lr': 0.00039934505766710656, 'samples': 8304128, 'steps': 16218, 'loss/train': 1.5180988311767578} 02/24/2022 22:35:24 - INFO - codeparrot_training - Step 16219: {'lr': 0.0003993319353378445, 'samples': 8304640, 'steps': 16219, 'loss/train': 2.35263991355896} 02/24/2022 22:35:28 - INFO - codeparrot_training - Step 16220: {'lr': 0.0003993188123688918, 'samples': 8305152, 'steps': 16220, 'loss/train': 1.7665669918060303} 02/24/2022 22:35:33 - INFO - codeparrot_training - Step 16221: {'lr': 0.00039930568876030473, 'samples': 8305664, 'steps': 16221, 'loss/train': 0.6001395583152771} 02/24/2022 22:35:37 - INFO - codeparrot_training - Step 16222: {'lr': 0.0003992925645121395, 'samples': 8306176, 'steps': 16222, 'loss/train': 2.160346269607544} 02/24/2022 22:35:42 - INFO - codeparrot_training - Step 16223: {'lr': 0.00039927943962445234, 'samples': 8306688, 'steps': 16223, 'loss/train': 2.4018616676330566} 02/24/2022 22:35:46 - INFO - codeparrot_training - Step 16224: {'lr': 0.0003992663140972994, 'samples': 8307200, 'steps': 16224, 'loss/train': 2.0112392902374268} 02/24/2022 22:35:51 - INFO - codeparrot_training - Step 16225: {'lr': 0.0003992531879307371, 'samples': 8307712, 'steps': 16225, 'loss/train': 2.5191287994384766} 02/24/2022 22:35:55 - INFO - codeparrot_training - Step 16226: {'lr': 0.0003992400611248214, 'samples': 8308224, 'steps': 16226, 'loss/train': 2.7199790477752686} 02/24/2022 22:36:00 - INFO - codeparrot_training - Step 16227: {'lr': 0.0003992269336796087, 'samples': 8308736, 'steps': 16227, 'loss/train': 1.5472825765609741} 02/24/2022 22:36:04 - INFO - codeparrot_training - Step 16228: {'lr': 0.0003992138055951552, 'samples': 8309248, 'steps': 16228, 'loss/train': 1.000227928161621} 02/24/2022 22:36:09 - INFO - codeparrot_training - Step 16229: {'lr': 0.00039920067687151717, 'samples': 8309760, 'steps': 16229, 'loss/train': 1.5652979612350464} 02/24/2022 22:36:13 - INFO - codeparrot_training - Step 16230: {'lr': 0.0003991875475087508, 'samples': 8310272, 'steps': 16230, 'loss/train': 1.9046932458877563} 02/24/2022 22:36:19 - INFO - codeparrot_training - Step 16231: {'lr': 0.00039917441750691237, 'samples': 8310784, 'steps': 16231, 'loss/train': 2.010688066482544} 02/24/2022 22:36:23 - INFO - codeparrot_training - Step 16232: {'lr': 0.0003991612868660581, 'samples': 8311296, 'steps': 16232, 'loss/train': 1.3565386533737183} 02/24/2022 22:36:29 - INFO - codeparrot_training - Step 16233: {'lr': 0.0003991481555862442, 'samples': 8311808, 'steps': 16233, 'loss/train': 2.3106303215026855} 02/24/2022 22:36:32 - INFO - codeparrot_training - Step 16234: {'lr': 0.00039913502366752704, 'samples': 8312320, 'steps': 16234, 'loss/train': 0.16075341403484344} 02/24/2022 22:36:38 - INFO - codeparrot_training - Step 16235: {'lr': 0.0003991218911099627, 'samples': 8312832, 'steps': 16235, 'loss/train': 1.3604589700698853} 02/24/2022 22:36:41 - INFO - codeparrot_training - Step 16236: {'lr': 0.0003991087579136076, 'samples': 8313344, 'steps': 16236, 'loss/train': 2.4689462184906006} 02/24/2022 22:36:47 - INFO - codeparrot_training - Step 16237: {'lr': 0.00039909562407851784, 'samples': 8313856, 'steps': 16237, 'loss/train': 1.8585337400436401} 02/24/2022 22:36:50 - INFO - codeparrot_training - Step 16238: {'lr': 0.0003990824896047498, 'samples': 8314368, 'steps': 16238, 'loss/train': 2.2633919715881348} 02/24/2022 22:36:56 - INFO - codeparrot_training - Step 16239: {'lr': 0.00039906935449235983, 'samples': 8314880, 'steps': 16239, 'loss/train': 2.590285062789917} 02/24/2022 22:36:59 - INFO - codeparrot_training - Step 16240: {'lr': 0.00039905621874140396, 'samples': 8315392, 'steps': 16240, 'loss/train': 0.5288260579109192} 02/24/2022 22:37:06 - INFO - codeparrot_training - Step 16241: {'lr': 0.00039904308235193866, 'samples': 8315904, 'steps': 16241, 'loss/train': 2.1751840114593506} 02/24/2022 22:37:09 - INFO - codeparrot_training - Step 16242: {'lr': 0.00039902994532402004, 'samples': 8316416, 'steps': 16242, 'loss/train': 2.2145566940307617} 02/24/2022 22:37:15 - INFO - codeparrot_training - Step 16243: {'lr': 0.0003990168076577045, 'samples': 8316928, 'steps': 16243, 'loss/train': 0.974034309387207} 02/24/2022 22:37:18 - INFO - codeparrot_training - Step 16244: {'lr': 0.00039900366935304824, 'samples': 8317440, 'steps': 16244, 'loss/train': 1.8895635604858398} 02/24/2022 22:37:24 - INFO - codeparrot_training - Step 16245: {'lr': 0.00039899053041010765, 'samples': 8317952, 'steps': 16245, 'loss/train': 1.9377793073654175} 02/24/2022 22:37:27 - INFO - codeparrot_training - Step 16246: {'lr': 0.00039897739082893883, 'samples': 8318464, 'steps': 16246, 'loss/train': 2.3315436840057373} 02/24/2022 22:37:33 - INFO - codeparrot_training - Step 16247: {'lr': 0.0003989642506095983, 'samples': 8318976, 'steps': 16247, 'loss/train': 1.5239439010620117} 02/24/2022 22:37:36 - INFO - codeparrot_training - Step 16248: {'lr': 0.0003989511097521421, 'samples': 8319488, 'steps': 16248, 'loss/train': 1.4171808958053589} 02/24/2022 22:37:42 - INFO - codeparrot_training - Step 16249: {'lr': 0.00039893796825662676, 'samples': 8320000, 'steps': 16249, 'loss/train': 3.3668322563171387} 02/24/2022 22:37:45 - INFO - codeparrot_training - Step 16250: {'lr': 0.0003989248261231084, 'samples': 8320512, 'steps': 16250, 'loss/train': 2.0585360527038574} 02/24/2022 22:37:52 - INFO - codeparrot_training - Step 16251: {'lr': 0.0003989116833516433, 'samples': 8321024, 'steps': 16251, 'loss/train': 1.05678391456604} 02/24/2022 22:37:55 - INFO - codeparrot_training - Step 16252: {'lr': 0.000398898539942288, 'samples': 8321536, 'steps': 16252, 'loss/train': 1.8924137353897095} 02/24/2022 22:38:00 - INFO - codeparrot_training - Step 16253: {'lr': 0.0003988853958950984, 'samples': 8322048, 'steps': 16253, 'loss/train': 1.5327723026275635} 02/24/2022 22:38:04 - INFO - codeparrot_training - Step 16254: {'lr': 0.00039887225121013124, 'samples': 8322560, 'steps': 16254, 'loss/train': 2.595731735229492} 02/24/2022 22:38:09 - INFO - codeparrot_training - Step 16255: {'lr': 0.0003988591058874426, 'samples': 8323072, 'steps': 16255, 'loss/train': 1.7835301160812378} 02/24/2022 22:38:13 - INFO - codeparrot_training - Step 16256: {'lr': 0.00039884595992708877, 'samples': 8323584, 'steps': 16256, 'loss/train': 2.2080979347229004} 02/24/2022 22:38:19 - INFO - codeparrot_training - Step 16257: {'lr': 0.0003988328133291261, 'samples': 8324096, 'steps': 16257, 'loss/train': 1.7378042936325073} 02/24/2022 22:38:22 - INFO - codeparrot_training - Step 16258: {'lr': 0.000398819666093611, 'samples': 8324608, 'steps': 16258, 'loss/train': 0.9707382917404175} 02/24/2022 22:38:28 - INFO - codeparrot_training - Step 16259: {'lr': 0.0003988065182205996, 'samples': 8325120, 'steps': 16259, 'loss/train': 2.9446816444396973} 02/24/2022 22:38:31 - INFO - codeparrot_training - Step 16260: {'lr': 0.0003987933697101484, 'samples': 8325632, 'steps': 16260, 'loss/train': 1.6062431335449219} 02/24/2022 22:38:37 - INFO - codeparrot_training - Step 16261: {'lr': 0.0003987802205623136, 'samples': 8326144, 'steps': 16261, 'loss/train': 2.2242660522460938} 02/24/2022 22:38:41 - INFO - codeparrot_training - Step 16262: {'lr': 0.0003987670707771516, 'samples': 8326656, 'steps': 16262, 'loss/train': 3.5205233097076416} 02/24/2022 22:38:46 - INFO - codeparrot_training - Step 16263: {'lr': 0.0003987539203547187, 'samples': 8327168, 'steps': 16263, 'loss/train': 0.9441400766372681} 02/24/2022 22:38:50 - INFO - codeparrot_training - Step 16264: {'lr': 0.00039874076929507124, 'samples': 8327680, 'steps': 16264, 'loss/train': 3.093445301055908} 02/24/2022 22:38:55 - INFO - codeparrot_training - Step 16265: {'lr': 0.0003987276175982656, 'samples': 8328192, 'steps': 16265, 'loss/train': 2.159552812576294} 02/24/2022 22:38:59 - INFO - codeparrot_training - Step 16266: {'lr': 0.00039871446526435806, 'samples': 8328704, 'steps': 16266, 'loss/train': 1.5061544179916382} 02/24/2022 22:39:04 - INFO - codeparrot_training - Step 16267: {'lr': 0.00039870131229340495, 'samples': 8329216, 'steps': 16267, 'loss/train': 1.717469334602356} 02/24/2022 22:39:08 - INFO - codeparrot_training - Step 16268: {'lr': 0.00039868815868546257, 'samples': 8329728, 'steps': 16268, 'loss/train': 1.8577830791473389} 02/24/2022 22:39:13 - INFO - codeparrot_training - Step 16269: {'lr': 0.00039867500444058747, 'samples': 8330240, 'steps': 16269, 'loss/train': 1.3528252840042114} 02/24/2022 22:39:17 - INFO - codeparrot_training - Step 16270: {'lr': 0.0003986618495588358, 'samples': 8330752, 'steps': 16270, 'loss/train': 0.8437669277191162} 02/24/2022 22:39:23 - INFO - codeparrot_training - Step 16271: {'lr': 0.00039864869404026394, 'samples': 8331264, 'steps': 16271, 'loss/train': 0.5787619948387146} 02/24/2022 22:39:26 - INFO - codeparrot_training - Step 16272: {'lr': 0.0003986355378849283, 'samples': 8331776, 'steps': 16272, 'loss/train': 0.3689199388027191} 02/24/2022 22:39:32 - INFO - codeparrot_training - Step 16273: {'lr': 0.00039862238109288523, 'samples': 8332288, 'steps': 16273, 'loss/train': 2.3050448894500732} 02/24/2022 22:39:35 - INFO - codeparrot_training - Step 16274: {'lr': 0.0003986092236641911, 'samples': 8332800, 'steps': 16274, 'loss/train': 2.035551071166992} 02/24/2022 22:39:41 - INFO - codeparrot_training - Step 16275: {'lr': 0.00039859606559890215, 'samples': 8333312, 'steps': 16275, 'loss/train': 1.84000563621521} 02/24/2022 22:39:44 - INFO - codeparrot_training - Step 16276: {'lr': 0.0003985829068970749, 'samples': 8333824, 'steps': 16276, 'loss/train': 1.827824592590332} 02/24/2022 22:39:50 - INFO - codeparrot_training - Step 16277: {'lr': 0.00039856974755876563, 'samples': 8334336, 'steps': 16277, 'loss/train': 0.7145095467567444} 02/24/2022 22:39:54 - INFO - codeparrot_training - Step 16278: {'lr': 0.0003985565875840308, 'samples': 8334848, 'steps': 16278, 'loss/train': 0.6022953987121582} 02/24/2022 22:39:59 - INFO - codeparrot_training - Step 16279: {'lr': 0.0003985434269729267, 'samples': 8335360, 'steps': 16279, 'loss/train': 2.246063470840454} 02/24/2022 22:40:03 - INFO - codeparrot_training - Step 16280: {'lr': 0.00039853026572550965, 'samples': 8335872, 'steps': 16280, 'loss/train': 3.149411916732788} 02/24/2022 22:40:08 - INFO - codeparrot_training - Step 16281: {'lr': 0.00039851710384183615, 'samples': 8336384, 'steps': 16281, 'loss/train': 1.8245059251785278} 02/24/2022 22:40:12 - INFO - codeparrot_training - Step 16282: {'lr': 0.0003985039413219626, 'samples': 8336896, 'steps': 16282, 'loss/train': 1.8205175399780273} 02/24/2022 22:40:17 - INFO - codeparrot_training - Step 16283: {'lr': 0.0003984907781659452, 'samples': 8337408, 'steps': 16283, 'loss/train': 1.714830994606018} 02/24/2022 22:40:21 - INFO - codeparrot_training - Step 16284: {'lr': 0.00039847761437384054, 'samples': 8337920, 'steps': 16284, 'loss/train': 2.2593464851379395} 02/24/2022 22:40:26 - INFO - codeparrot_training - Step 16285: {'lr': 0.0003984644499457049, 'samples': 8338432, 'steps': 16285, 'loss/train': 1.9405473470687866} 02/24/2022 22:40:30 - INFO - codeparrot_training - Step 16286: {'lr': 0.0003984512848815948, 'samples': 8338944, 'steps': 16286, 'loss/train': 0.8835693001747131} 02/24/2022 22:40:36 - INFO - codeparrot_training - Step 16287: {'lr': 0.00039843811918156635, 'samples': 8339456, 'steps': 16287, 'loss/train': 1.8923888206481934} 02/24/2022 22:40:39 - INFO - codeparrot_training - Step 16288: {'lr': 0.0003984249528456762, 'samples': 8339968, 'steps': 16288, 'loss/train': 1.098946213722229} 02/24/2022 22:40:45 - INFO - codeparrot_training - Step 16289: {'lr': 0.00039841178587398074, 'samples': 8340480, 'steps': 16289, 'loss/train': 2.3656797409057617} 02/24/2022 22:40:48 - INFO - codeparrot_training - Step 16290: {'lr': 0.0003983986182665362, 'samples': 8340992, 'steps': 16290, 'loss/train': 1.1832685470581055} 02/24/2022 22:40:54 - INFO - codeparrot_training - Step 16291: {'lr': 0.00039838545002339926, 'samples': 8341504, 'steps': 16291, 'loss/train': 2.0258514881134033} 02/24/2022 22:40:57 - INFO - codeparrot_training - Step 16292: {'lr': 0.0003983722811446261, 'samples': 8342016, 'steps': 16292, 'loss/train': 1.7101701498031616} 02/24/2022 22:41:03 - INFO - codeparrot_training - Step 16293: {'lr': 0.00039835911163027315, 'samples': 8342528, 'steps': 16293, 'loss/train': 2.7284963130950928} 02/24/2022 22:41:06 - INFO - codeparrot_training - Step 16294: {'lr': 0.00039834594148039693, 'samples': 8343040, 'steps': 16294, 'loss/train': 1.4164475202560425} 02/24/2022 22:41:12 - INFO - codeparrot_training - Step 16295: {'lr': 0.0003983327706950538, 'samples': 8343552, 'steps': 16295, 'loss/train': 1.9087039232254028} 02/24/2022 22:41:15 - INFO - codeparrot_training - Step 16296: {'lr': 0.00039831959927430017, 'samples': 8344064, 'steps': 16296, 'loss/train': 2.2220218181610107} 02/24/2022 22:41:21 - INFO - codeparrot_training - Step 16297: {'lr': 0.00039830642721819254, 'samples': 8344576, 'steps': 16297, 'loss/train': 3.1141719818115234} 02/24/2022 22:41:24 - INFO - codeparrot_training - Step 16298: {'lr': 0.0003982932545267872, 'samples': 8345088, 'steps': 16298, 'loss/train': 2.0572309494018555} 02/24/2022 22:41:31 - INFO - codeparrot_training - Step 16299: {'lr': 0.00039828008120014057, 'samples': 8345600, 'steps': 16299, 'loss/train': 2.0216176509857178} 02/24/2022 22:41:34 - INFO - codeparrot_training - Step 16300: {'lr': 0.00039826690723830926, 'samples': 8346112, 'steps': 16300, 'loss/train': 1.9893757104873657} 02/24/2022 22:41:40 - INFO - codeparrot_training - Step 16301: {'lr': 0.00039825373264134955, 'samples': 8346624, 'steps': 16301, 'loss/train': 1.7790496349334717} 02/24/2022 22:41:43 - INFO - codeparrot_training - Step 16302: {'lr': 0.00039824055740931804, 'samples': 8347136, 'steps': 16302, 'loss/train': 2.3769853115081787} 02/24/2022 22:41:49 - INFO - codeparrot_training - Step 16303: {'lr': 0.0003982273815422709, 'samples': 8347648, 'steps': 16303, 'loss/train': 1.9558449983596802} 02/24/2022 22:41:52 - INFO - codeparrot_training - Step 16304: {'lr': 0.00039821420504026486, 'samples': 8348160, 'steps': 16304, 'loss/train': 1.4359934329986572} 02/24/2022 22:41:58 - INFO - codeparrot_training - Step 16305: {'lr': 0.0003982010279033561, 'samples': 8348672, 'steps': 16305, 'loss/train': 2.1049954891204834} 02/24/2022 22:42:01 - INFO - codeparrot_training - Step 16306: {'lr': 0.0003981878501316013, 'samples': 8349184, 'steps': 16306, 'loss/train': 2.4715776443481445} 02/24/2022 22:42:07 - INFO - codeparrot_training - Step 16307: {'lr': 0.0003981746717250567, 'samples': 8349696, 'steps': 16307, 'loss/train': 2.257695436477661} 02/24/2022 22:42:10 - INFO - codeparrot_training - Step 16308: {'lr': 0.000398161492683779, 'samples': 8350208, 'steps': 16308, 'loss/train': 1.9007292985916138} 02/24/2022 22:42:17 - INFO - codeparrot_training - Step 16309: {'lr': 0.0003981483130078244, 'samples': 8350720, 'steps': 16309, 'loss/train': 2.6354665756225586} 02/24/2022 22:42:20 - INFO - codeparrot_training - Step 16310: {'lr': 0.0003981351326972495, 'samples': 8351232, 'steps': 16310, 'loss/train': 1.637117862701416} 02/24/2022 22:42:26 - INFO - codeparrot_training - Step 16311: {'lr': 0.00039812195175211075, 'samples': 8351744, 'steps': 16311, 'loss/train': 2.451132297515869} 02/24/2022 22:42:30 - INFO - codeparrot_training - Step 16312: {'lr': 0.0003981087701724645, 'samples': 8352256, 'steps': 16312, 'loss/train': 3.009580135345459} 02/24/2022 22:42:35 - INFO - codeparrot_training - Step 16313: {'lr': 0.00039809558795836743, 'samples': 8352768, 'steps': 16313, 'loss/train': 1.534165859222412} 02/24/2022 22:42:39 - INFO - codeparrot_training - Step 16314: {'lr': 0.00039808240510987584, 'samples': 8353280, 'steps': 16314, 'loss/train': 1.796891450881958} 02/24/2022 22:42:44 - INFO - codeparrot_training - Step 16315: {'lr': 0.0003980692216270462, 'samples': 8353792, 'steps': 16315, 'loss/train': 1.7552646398544312} 02/24/2022 22:42:47 - INFO - codeparrot_training - Step 16316: {'lr': 0.00039805603750993514, 'samples': 8354304, 'steps': 16316, 'loss/train': 1.9753177165985107} 02/24/2022 22:42:53 - INFO - codeparrot_training - Step 16317: {'lr': 0.0003980428527585989, 'samples': 8354816, 'steps': 16317, 'loss/train': 1.973376989364624} 02/24/2022 22:42:57 - INFO - codeparrot_training - Step 16318: {'lr': 0.0003980296673730942, 'samples': 8355328, 'steps': 16318, 'loss/train': 1.7562350034713745} 02/24/2022 22:43:02 - INFO - codeparrot_training - Step 16319: {'lr': 0.0003980164813534773, 'samples': 8355840, 'steps': 16319, 'loss/train': 2.663815975189209} 02/24/2022 22:43:06 - INFO - codeparrot_training - Step 16320: {'lr': 0.0003980032946998049, 'samples': 8356352, 'steps': 16320, 'loss/train': 2.2288382053375244} 02/24/2022 22:43:11 - INFO - codeparrot_training - Step 16321: {'lr': 0.00039799010741213336, 'samples': 8356864, 'steps': 16321, 'loss/train': 1.9420993328094482} 02/24/2022 22:43:15 - INFO - codeparrot_training - Step 16322: {'lr': 0.0003979769194905192, 'samples': 8357376, 'steps': 16322, 'loss/train': 2.2489728927612305} 02/24/2022 22:43:20 - INFO - codeparrot_training - Step 16323: {'lr': 0.0003979637309350188, 'samples': 8357888, 'steps': 16323, 'loss/train': 0.9923384189605713} 02/24/2022 22:43:24 - INFO - codeparrot_training - Step 16324: {'lr': 0.0003979505417456889, 'samples': 8358400, 'steps': 16324, 'loss/train': 2.8191654682159424} 02/24/2022 22:43:30 - INFO - codeparrot_training - Step 16325: {'lr': 0.00039793735192258575, 'samples': 8358912, 'steps': 16325, 'loss/train': 2.2891783714294434} 02/24/2022 22:43:33 - INFO - codeparrot_training - Step 16326: {'lr': 0.000397924161465766, 'samples': 8359424, 'steps': 16326, 'loss/train': 2.5036966800689697} 02/24/2022 22:43:39 - INFO - codeparrot_training - Step 16327: {'lr': 0.0003979109703752861, 'samples': 8359936, 'steps': 16327, 'loss/train': 1.8108277320861816} 02/24/2022 22:43:42 - INFO - codeparrot_training - Step 16328: {'lr': 0.00039789777865120257, 'samples': 8360448, 'steps': 16328, 'loss/train': 2.2767579555511475} 02/24/2022 22:43:48 - INFO - codeparrot_training - Step 16329: {'lr': 0.00039788458629357195, 'samples': 8360960, 'steps': 16329, 'loss/train': 1.2777466773986816} 02/24/2022 22:43:51 - INFO - codeparrot_training - Step 16330: {'lr': 0.0003978713933024507, 'samples': 8361472, 'steps': 16330, 'loss/train': 1.5918738842010498} 02/24/2022 22:43:57 - INFO - codeparrot_training - Step 16331: {'lr': 0.0003978581996778954, 'samples': 8361984, 'steps': 16331, 'loss/train': 1.9526430368423462} 02/24/2022 22:44:00 - INFO - codeparrot_training - Step 16332: {'lr': 0.0003978450054199625, 'samples': 8362496, 'steps': 16332, 'loss/train': 1.2859208583831787} 02/24/2022 22:44:07 - INFO - codeparrot_training - Step 16333: {'lr': 0.0003978318105287085, 'samples': 8363008, 'steps': 16333, 'loss/train': 2.0269393920898438} 02/24/2022 22:44:10 - INFO - codeparrot_training - Step 16334: {'lr': 0.00039781861500419, 'samples': 8363520, 'steps': 16334, 'loss/train': 2.8162662982940674} 02/24/2022 22:44:16 - INFO - codeparrot_training - Step 16335: {'lr': 0.00039780541884646347, 'samples': 8364032, 'steps': 16335, 'loss/train': 2.0910496711730957} 02/24/2022 22:44:20 - INFO - codeparrot_training - Step 16336: {'lr': 0.0003977922220555855, 'samples': 8364544, 'steps': 16336, 'loss/train': 2.6627039909362793} 02/24/2022 22:44:25 - INFO - codeparrot_training - Step 16337: {'lr': 0.0003977790246316125, 'samples': 8365056, 'steps': 16337, 'loss/train': 1.933947205543518} 02/24/2022 22:44:28 - INFO - codeparrot_training - Step 16338: {'lr': 0.00039776582657460115, 'samples': 8365568, 'steps': 16338, 'loss/train': 2.351294755935669} 02/24/2022 22:44:34 - INFO - codeparrot_training - Step 16339: {'lr': 0.000397752627884608, 'samples': 8366080, 'steps': 16339, 'loss/train': 1.0967384576797485} 02/24/2022 22:44:38 - INFO - codeparrot_training - Step 16340: {'lr': 0.0003977394285616893, 'samples': 8366592, 'steps': 16340, 'loss/train': 2.1124072074890137} 02/24/2022 22:44:43 - INFO - codeparrot_training - Step 16341: {'lr': 0.000397726228605902, 'samples': 8367104, 'steps': 16341, 'loss/train': 2.3344578742980957} 02/24/2022 22:44:47 - INFO - codeparrot_training - Step 16342: {'lr': 0.00039771302801730235, 'samples': 8367616, 'steps': 16342, 'loss/train': 2.6118335723876953} 02/24/2022 22:44:52 - INFO - codeparrot_training - Step 16343: {'lr': 0.00039769982679594703, 'samples': 8368128, 'steps': 16343, 'loss/train': 1.5699336528778076} 02/24/2022 22:44:55 - INFO - codeparrot_training - Step 16344: {'lr': 0.0003976866249418925, 'samples': 8368640, 'steps': 16344, 'loss/train': 2.2197580337524414} 02/24/2022 22:45:02 - INFO - codeparrot_training - Step 16345: {'lr': 0.0003976734224551954, 'samples': 8369152, 'steps': 16345, 'loss/train': 2.216350793838501} 02/24/2022 22:45:05 - INFO - codeparrot_training - Step 16346: {'lr': 0.0003976602193359122, 'samples': 8369664, 'steps': 16346, 'loss/train': 1.852243423461914} 02/24/2022 22:45:11 - INFO - codeparrot_training - Step 16347: {'lr': 0.00039764701558409955, 'samples': 8370176, 'steps': 16347, 'loss/train': 1.8337823152542114} 02/24/2022 22:45:14 - INFO - codeparrot_training - Step 16348: {'lr': 0.000397633811199814, 'samples': 8370688, 'steps': 16348, 'loss/train': 1.6812801361083984} 02/24/2022 22:45:20 - INFO - codeparrot_training - Step 16349: {'lr': 0.000397620606183112, 'samples': 8371200, 'steps': 16349, 'loss/train': 2.2033019065856934} 02/24/2022 22:45:23 - INFO - codeparrot_training - Step 16350: {'lr': 0.00039760740053405033, 'samples': 8371712, 'steps': 16350, 'loss/train': 2.432379961013794} 02/24/2022 22:45:29 - INFO - codeparrot_training - Step 16351: {'lr': 0.00039759419425268526, 'samples': 8372224, 'steps': 16351, 'loss/train': 2.5131478309631348} 02/24/2022 22:45:32 - INFO - codeparrot_training - Step 16352: {'lr': 0.00039758098733907364, 'samples': 8372736, 'steps': 16352, 'loss/train': 1.3854668140411377} 02/24/2022 22:45:38 - INFO - codeparrot_training - Step 16353: {'lr': 0.00039756777979327193, 'samples': 8373248, 'steps': 16353, 'loss/train': 0.9563204050064087} 02/24/2022 22:45:41 - INFO - codeparrot_training - Step 16354: {'lr': 0.0003975545716153367, 'samples': 8373760, 'steps': 16354, 'loss/train': 2.311582565307617} 02/24/2022 22:45:48 - INFO - codeparrot_training - Step 16355: {'lr': 0.0003975413628053245, 'samples': 8374272, 'steps': 16355, 'loss/train': 2.278615951538086} 02/24/2022 22:45:51 - INFO - codeparrot_training - Step 16356: {'lr': 0.000397528153363292, 'samples': 8374784, 'steps': 16356, 'loss/train': 2.7309703826904297} 02/24/2022 22:45:57 - INFO - codeparrot_training - Step 16357: {'lr': 0.00039751494328929565, 'samples': 8375296, 'steps': 16357, 'loss/train': 0.7731642127037048} 02/24/2022 22:46:00 - INFO - codeparrot_training - Step 16358: {'lr': 0.00039750173258339225, 'samples': 8375808, 'steps': 16358, 'loss/train': 1.1144764423370361} 02/24/2022 22:46:04 - INFO - codeparrot_training - Step 16359: {'lr': 0.00039748852124563816, 'samples': 8376320, 'steps': 16359, 'loss/train': 1.6835076808929443} 02/24/2022 22:46:10 - INFO - codeparrot_training - Step 16360: {'lr': 0.0003974753092760901, 'samples': 8376832, 'steps': 16360, 'loss/train': 3.235959053039551} 02/24/2022 22:46:15 - INFO - codeparrot_training - Step 16361: {'lr': 0.00039746209667480473, 'samples': 8377344, 'steps': 16361, 'loss/train': 1.918245553970337} 02/24/2022 22:46:19 - INFO - codeparrot_training - Step 16362: {'lr': 0.00039744888344183846, 'samples': 8377856, 'steps': 16362, 'loss/train': 1.8856500387191772} 02/24/2022 22:46:24 - INFO - codeparrot_training - Step 16363: {'lr': 0.00039743566957724805, 'samples': 8378368, 'steps': 16363, 'loss/train': 1.7472566366195679} 02/24/2022 22:46:27 - INFO - codeparrot_training - Step 16364: {'lr': 0.00039742245508109, 'samples': 8378880, 'steps': 16364, 'loss/train': 1.9533767700195312} 02/24/2022 22:46:33 - INFO - codeparrot_training - Step 16365: {'lr': 0.000397409239953421, 'samples': 8379392, 'steps': 16365, 'loss/train': 1.1026155948638916} 02/24/2022 22:46:37 - INFO - codeparrot_training - Step 16366: {'lr': 0.00039739602419429755, 'samples': 8379904, 'steps': 16366, 'loss/train': 3.1734139919281006} 02/24/2022 22:46:42 - INFO - codeparrot_training - Step 16367: {'lr': 0.00039738280780377645, 'samples': 8380416, 'steps': 16367, 'loss/train': 2.220407724380493} 02/24/2022 22:46:46 - INFO - codeparrot_training - Step 16368: {'lr': 0.0003973695907819141, 'samples': 8380928, 'steps': 16368, 'loss/train': 1.9063369035720825} 02/24/2022 22:46:51 - INFO - codeparrot_training - Step 16369: {'lr': 0.0003973563731287673, 'samples': 8381440, 'steps': 16369, 'loss/train': 2.800663948059082} 02/24/2022 22:46:55 - INFO - codeparrot_training - Step 16370: {'lr': 0.00039734315484439255, 'samples': 8381952, 'steps': 16370, 'loss/train': 2.1165618896484375} 02/24/2022 22:47:01 - INFO - codeparrot_training - Step 16371: {'lr': 0.0003973299359288465, 'samples': 8382464, 'steps': 16371, 'loss/train': 2.0720815658569336} 02/24/2022 22:47:04 - INFO - codeparrot_training - Step 16372: {'lr': 0.0003973167163821858, 'samples': 8382976, 'steps': 16372, 'loss/train': 2.5066123008728027} 02/24/2022 22:47:10 - INFO - codeparrot_training - Step 16373: {'lr': 0.0003973034962044671, 'samples': 8383488, 'steps': 16373, 'loss/train': 0.39379721879959106} 02/24/2022 22:47:13 - INFO - codeparrot_training - Step 16374: {'lr': 0.00039729027539574696, 'samples': 8384000, 'steps': 16374, 'loss/train': 1.7836464643478394} 02/24/2022 22:47:19 - INFO - codeparrot_training - Step 16375: {'lr': 0.00039727705395608203, 'samples': 8384512, 'steps': 16375, 'loss/train': 1.5476856231689453} 02/24/2022 22:47:22 - INFO - codeparrot_training - Step 16376: {'lr': 0.00039726383188552907, 'samples': 8385024, 'steps': 16376, 'loss/train': 1.8719788789749146} 02/24/2022 22:47:28 - INFO - codeparrot_training - Step 16377: {'lr': 0.00039725060918414446, 'samples': 8385536, 'steps': 16377, 'loss/train': 2.55049991607666} 02/24/2022 22:47:31 - INFO - codeparrot_training - Step 16378: {'lr': 0.0003972373858519851, 'samples': 8386048, 'steps': 16378, 'loss/train': 2.5014894008636475} 02/24/2022 22:47:37 - INFO - codeparrot_training - Step 16379: {'lr': 0.00039722416188910754, 'samples': 8386560, 'steps': 16379, 'loss/train': 1.4758388996124268} 02/24/2022 22:47:40 - INFO - codeparrot_training - Step 16380: {'lr': 0.00039721093729556836, 'samples': 8387072, 'steps': 16380, 'loss/train': 1.4191232919692993} 02/24/2022 22:47:46 - INFO - codeparrot_training - Step 16381: {'lr': 0.0003971977120714243, 'samples': 8387584, 'steps': 16381, 'loss/train': 1.8881176710128784} 02/24/2022 22:47:50 - INFO - codeparrot_training - Step 16382: {'lr': 0.000397184486216732, 'samples': 8388096, 'steps': 16382, 'loss/train': 1.9483017921447754} 02/24/2022 22:47:55 - INFO - codeparrot_training - Step 16383: {'lr': 0.0003971712597315481, 'samples': 8388608, 'steps': 16383, 'loss/train': 1.5617362260818481} 02/24/2022 22:47:59 - INFO - codeparrot_training - Step 16384: {'lr': 0.0003971580326159292, 'samples': 8389120, 'steps': 16384, 'loss/train': 2.1248550415039062} 02/24/2022 22:48:04 - INFO - codeparrot_training - Step 16385: {'lr': 0.0003971448048699321, 'samples': 8389632, 'steps': 16385, 'loss/train': 2.250239133834839} 02/24/2022 22:48:08 - INFO - codeparrot_training - Step 16386: {'lr': 0.00039713157649361327, 'samples': 8390144, 'steps': 16386, 'loss/train': 2.240642786026001} 02/24/2022 22:48:13 - INFO - codeparrot_training - Step 16387: {'lr': 0.00039711834748702956, 'samples': 8390656, 'steps': 16387, 'loss/train': 2.2107701301574707} 02/24/2022 22:48:17 - INFO - codeparrot_training - Step 16388: {'lr': 0.0003971051178502375, 'samples': 8391168, 'steps': 16388, 'loss/train': 4.451059818267822} 02/24/2022 22:48:22 - INFO - codeparrot_training - Step 16389: {'lr': 0.00039709188758329394, 'samples': 8391680, 'steps': 16389, 'loss/train': 2.0136067867279053} 02/24/2022 22:48:26 - INFO - codeparrot_training - Step 16390: {'lr': 0.0003970786566862553, 'samples': 8392192, 'steps': 16390, 'loss/train': 1.6645402908325195} 02/24/2022 22:48:32 - INFO - codeparrot_training - Step 16391: {'lr': 0.00039706542515917853, 'samples': 8392704, 'steps': 16391, 'loss/train': 1.0072708129882812} 02/24/2022 22:48:35 - INFO - codeparrot_training - Step 16392: {'lr': 0.00039705219300212015, 'samples': 8393216, 'steps': 16392, 'loss/train': 1.4125769138336182} 02/24/2022 22:48:41 - INFO - codeparrot_training - Step 16393: {'lr': 0.00039703896021513684, 'samples': 8393728, 'steps': 16393, 'loss/train': 0.5865223407745361} 02/24/2022 22:48:44 - INFO - codeparrot_training - Step 16394: {'lr': 0.0003970257267982853, 'samples': 8394240, 'steps': 16394, 'loss/train': 1.7379567623138428} 02/24/2022 22:48:50 - INFO - codeparrot_training - Step 16395: {'lr': 0.0003970124927516222, 'samples': 8394752, 'steps': 16395, 'loss/train': 1.4561570882797241} 02/24/2022 22:48:53 - INFO - codeparrot_training - Step 16396: {'lr': 0.0003969992580752043, 'samples': 8395264, 'steps': 16396, 'loss/train': 1.6364994049072266} 02/24/2022 22:48:59 - INFO - codeparrot_training - Step 16397: {'lr': 0.00039698602276908826, 'samples': 8395776, 'steps': 16397, 'loss/train': 2.2902824878692627} 02/24/2022 22:49:02 - INFO - codeparrot_training - Step 16398: {'lr': 0.0003969727868333308, 'samples': 8396288, 'steps': 16398, 'loss/train': 3.540570020675659} 02/24/2022 22:49:08 - INFO - codeparrot_training - Step 16399: {'lr': 0.00039695955026798857, 'samples': 8396800, 'steps': 16399, 'loss/train': 2.3943612575531006} 02/24/2022 22:49:11 - INFO - codeparrot_training - Step 16400: {'lr': 0.0003969463130731183, 'samples': 8397312, 'steps': 16400, 'loss/train': 0.6846969127655029} 02/24/2022 22:49:18 - INFO - codeparrot_training - Step 16401: {'lr': 0.00039693307524877664, 'samples': 8397824, 'steps': 16401, 'loss/train': 2.4312009811401367} 02/24/2022 22:49:21 - INFO - codeparrot_training - Step 16402: {'lr': 0.0003969198367950204, 'samples': 8398336, 'steps': 16402, 'loss/train': 1.9659732580184937} 02/24/2022 22:49:26 - INFO - codeparrot_training - Step 16403: {'lr': 0.00039690659771190616, 'samples': 8398848, 'steps': 16403, 'loss/train': 2.5383570194244385} 02/24/2022 22:49:30 - INFO - codeparrot_training - Step 16404: {'lr': 0.0003968933579994908, 'samples': 8399360, 'steps': 16404, 'loss/train': 1.7474948167800903} 02/24/2022 22:49:35 - INFO - codeparrot_training - Step 16405: {'lr': 0.0003968801176578309, 'samples': 8399872, 'steps': 16405, 'loss/train': 4.142742156982422} 02/24/2022 22:49:39 - INFO - codeparrot_training - Step 16406: {'lr': 0.00039686687668698316, 'samples': 8400384, 'steps': 16406, 'loss/train': 0.7344520688056946} 02/24/2022 22:49:44 - INFO - codeparrot_training - Step 16407: {'lr': 0.00039685363508700443, 'samples': 8400896, 'steps': 16407, 'loss/train': 3.1351852416992188} 02/24/2022 22:49:48 - INFO - codeparrot_training - Step 16408: {'lr': 0.00039684039285795133, 'samples': 8401408, 'steps': 16408, 'loss/train': 2.0528266429901123} 02/24/2022 22:49:54 - INFO - codeparrot_training - Step 16409: {'lr': 0.0003968271499998806, 'samples': 8401920, 'steps': 16409, 'loss/train': 1.8270974159240723} 02/24/2022 22:49:57 - INFO - codeparrot_training - Step 16410: {'lr': 0.000396813906512849, 'samples': 8402432, 'steps': 16410, 'loss/train': 2.603074312210083} 02/24/2022 22:50:03 - INFO - codeparrot_training - Step 16411: {'lr': 0.00039680066239691325, 'samples': 8402944, 'steps': 16411, 'loss/train': 2.0200417041778564} 02/24/2022 22:50:06 - INFO - codeparrot_training - Step 16412: {'lr': 0.00039678741765213006, 'samples': 8403456, 'steps': 16412, 'loss/train': 2.427785873413086} 02/24/2022 22:50:12 - INFO - codeparrot_training - Step 16413: {'lr': 0.00039677417227855624, 'samples': 8403968, 'steps': 16413, 'loss/train': 0.5865729451179504} 02/24/2022 22:50:15 - INFO - codeparrot_training - Step 16414: {'lr': 0.0003967609262762484, 'samples': 8404480, 'steps': 16414, 'loss/train': 1.9257038831710815} 02/24/2022 22:50:21 - INFO - codeparrot_training - Step 16415: {'lr': 0.0003967476796452634, 'samples': 8404992, 'steps': 16415, 'loss/train': 2.423224925994873} 02/24/2022 22:50:24 - INFO - codeparrot_training - Step 16416: {'lr': 0.00039673443238565786, 'samples': 8405504, 'steps': 16416, 'loss/train': 1.0601638555526733} 02/24/2022 22:50:31 - INFO - codeparrot_training - Step 16417: {'lr': 0.0003967211844974887, 'samples': 8406016, 'steps': 16417, 'loss/train': 0.7265795469284058} 02/24/2022 22:50:35 - INFO - codeparrot_training - Step 16418: {'lr': 0.0003967079359808125, 'samples': 8406528, 'steps': 16418, 'loss/train': 1.439018964767456} 02/24/2022 22:50:40 - INFO - codeparrot_training - Step 16419: {'lr': 0.0003966946868356861, 'samples': 8407040, 'steps': 16419, 'loss/train': 2.1818454265594482} 02/24/2022 22:50:44 - INFO - codeparrot_training - Step 16420: {'lr': 0.0003966814370621663, 'samples': 8407552, 'steps': 16420, 'loss/train': 1.3772715330123901} 02/24/2022 22:50:49 - INFO - codeparrot_training - Step 16421: {'lr': 0.00039666818666030974, 'samples': 8408064, 'steps': 16421, 'loss/train': 1.5109524726867676} 02/24/2022 22:50:53 - INFO - codeparrot_training - Step 16422: {'lr': 0.0003966549356301733, 'samples': 8408576, 'steps': 16422, 'loss/train': 1.0947314500808716} 02/24/2022 22:50:58 - INFO - codeparrot_training - Step 16423: {'lr': 0.0003966416839718136, 'samples': 8409088, 'steps': 16423, 'loss/train': 1.9209057092666626} 02/24/2022 22:51:02 - INFO - codeparrot_training - Step 16424: {'lr': 0.00039662843168528756, 'samples': 8409600, 'steps': 16424, 'loss/train': 1.0642305612564087} 02/24/2022 22:51:07 - INFO - codeparrot_training - Step 16425: {'lr': 0.00039661517877065183, 'samples': 8410112, 'steps': 16425, 'loss/train': 1.898046612739563} 02/24/2022 22:51:11 - INFO - codeparrot_training - Step 16426: {'lr': 0.0003966019252279633, 'samples': 8410624, 'steps': 16426, 'loss/train': 2.0211141109466553} 02/24/2022 22:51:18 - INFO - codeparrot_training - Step 16427: {'lr': 0.00039658867105727856, 'samples': 8411136, 'steps': 16427, 'loss/train': 2.5976674556732178} 02/24/2022 22:51:23 - INFO - codeparrot_training - Step 16428: {'lr': 0.0003965754162586547, 'samples': 8411648, 'steps': 16428, 'loss/train': 1.5950156450271606} 02/24/2022 22:51:27 - INFO - codeparrot_training - Step 16429: {'lr': 0.0003965621608321481, 'samples': 8412160, 'steps': 16429, 'loss/train': 1.2250365018844604} 02/24/2022 22:51:32 - INFO - codeparrot_training - Step 16430: {'lr': 0.0003965489047778158, 'samples': 8412672, 'steps': 16430, 'loss/train': 1.175410509109497} 02/24/2022 22:51:36 - INFO - codeparrot_training - Step 16431: {'lr': 0.0003965356480957145, 'samples': 8413184, 'steps': 16431, 'loss/train': 1.8192424774169922} 02/24/2022 22:51:41 - INFO - codeparrot_training - Step 16432: {'lr': 0.0003965223907859011, 'samples': 8413696, 'steps': 16432, 'loss/train': 2.0306432247161865} 02/24/2022 22:51:45 - INFO - codeparrot_training - Step 16433: {'lr': 0.00039650913284843225, 'samples': 8414208, 'steps': 16433, 'loss/train': 2.4991097450256348} 02/24/2022 22:51:50 - INFO - codeparrot_training - Step 16434: {'lr': 0.00039649587428336474, 'samples': 8414720, 'steps': 16434, 'loss/train': 3.0235767364501953} 02/24/2022 22:51:54 - INFO - codeparrot_training - Step 16435: {'lr': 0.00039648261509075554, 'samples': 8415232, 'steps': 16435, 'loss/train': 1.6170494556427002} 02/24/2022 22:51:59 - INFO - codeparrot_training - Step 16436: {'lr': 0.00039646935527066124, 'samples': 8415744, 'steps': 16436, 'loss/train': 2.5134470462799072} 02/24/2022 22:52:02 - INFO - codeparrot_training - Step 16437: {'lr': 0.0003964560948231388, 'samples': 8416256, 'steps': 16437, 'loss/train': 1.8894404172897339} 02/24/2022 22:52:09 - INFO - codeparrot_training - Step 16438: {'lr': 0.0003964428337482449, 'samples': 8416768, 'steps': 16438, 'loss/train': 1.789795160293579} 02/24/2022 22:52:12 - INFO - codeparrot_training - Step 16439: {'lr': 0.00039642957204603647, 'samples': 8417280, 'steps': 16439, 'loss/train': 2.028066873550415} 02/24/2022 22:52:18 - INFO - codeparrot_training - Step 16440: {'lr': 0.0003964163097165702, 'samples': 8417792, 'steps': 16440, 'loss/train': 2.473353624343872} 02/24/2022 22:52:21 - INFO - codeparrot_training - Step 16441: {'lr': 0.0003964030467599029, 'samples': 8418304, 'steps': 16441, 'loss/train': 0.18668408691883087} 02/24/2022 22:52:27 - INFO - codeparrot_training - Step 16442: {'lr': 0.00039638978317609155, 'samples': 8418816, 'steps': 16442, 'loss/train': 1.9909363985061646} 02/24/2022 22:52:30 - INFO - codeparrot_training - Step 16443: {'lr': 0.0003963765189651928, 'samples': 8419328, 'steps': 16443, 'loss/train': 2.2685911655426025} 02/24/2022 22:52:36 - INFO - codeparrot_training - Step 16444: {'lr': 0.0003963632541272635, 'samples': 8419840, 'steps': 16444, 'loss/train': 2.473036289215088} 02/24/2022 22:52:39 - INFO - codeparrot_training - Step 16445: {'lr': 0.00039634998866236047, 'samples': 8420352, 'steps': 16445, 'loss/train': 1.807316780090332} 02/24/2022 22:52:45 - INFO - codeparrot_training - Step 16446: {'lr': 0.0003963367225705406, 'samples': 8420864, 'steps': 16446, 'loss/train': 2.210996150970459} 02/24/2022 22:52:48 - INFO - codeparrot_training - Step 16447: {'lr': 0.0003963234558518607, 'samples': 8421376, 'steps': 16447, 'loss/train': 2.2673728466033936} 02/24/2022 22:52:54 - INFO - codeparrot_training - Step 16448: {'lr': 0.0003963101885063776, 'samples': 8421888, 'steps': 16448, 'loss/train': 2.0479724407196045} 02/24/2022 22:52:58 - INFO - codeparrot_training - Step 16449: {'lr': 0.000396296920534148, 'samples': 8422400, 'steps': 16449, 'loss/train': 3.3648974895477295} 02/24/2022 22:53:04 - INFO - codeparrot_training - Step 16450: {'lr': 0.000396283651935229, 'samples': 8422912, 'steps': 16450, 'loss/train': 1.3978142738342285} 02/24/2022 22:53:07 - INFO - codeparrot_training - Step 16451: {'lr': 0.0003962703827096771, 'samples': 8423424, 'steps': 16451, 'loss/train': 1.3956341743469238} 02/24/2022 22:53:13 - INFO - codeparrot_training - Step 16452: {'lr': 0.00039625711285754943, 'samples': 8423936, 'steps': 16452, 'loss/train': 1.7806352376937866} 02/24/2022 22:53:16 - INFO - codeparrot_training - Step 16453: {'lr': 0.00039624384237890275, 'samples': 8424448, 'steps': 16453, 'loss/train': 1.9342843294143677} 02/24/2022 22:53:22 - INFO - codeparrot_training - Step 16454: {'lr': 0.00039623057127379386, 'samples': 8424960, 'steps': 16454, 'loss/train': 1.8664958477020264} 02/24/2022 22:53:25 - INFO - codeparrot_training - Step 16455: {'lr': 0.0003962172995422796, 'samples': 8425472, 'steps': 16455, 'loss/train': 2.6650211811065674} 02/24/2022 22:53:30 - INFO - codeparrot_training - Step 16456: {'lr': 0.00039620402718441687, 'samples': 8425984, 'steps': 16456, 'loss/train': 3.2424442768096924} 02/24/2022 22:53:34 - INFO - codeparrot_training - Step 16457: {'lr': 0.0003961907542002626, 'samples': 8426496, 'steps': 16457, 'loss/train': 2.916908025741577} 02/24/2022 22:53:39 - INFO - codeparrot_training - Step 16458: {'lr': 0.00039617748058987345, 'samples': 8427008, 'steps': 16458, 'loss/train': 1.9213870763778687} 02/24/2022 22:53:43 - INFO - codeparrot_training - Step 16459: {'lr': 0.0003961642063533065, 'samples': 8427520, 'steps': 16459, 'loss/train': 1.4892544746398926} 02/24/2022 22:53:49 - INFO - codeparrot_training - Step 16460: {'lr': 0.0003961509314906184, 'samples': 8428032, 'steps': 16460, 'loss/train': 1.9083231687545776} 02/24/2022 22:53:52 - INFO - codeparrot_training - Step 16461: {'lr': 0.0003961376560018662, 'samples': 8428544, 'steps': 16461, 'loss/train': 2.6866848468780518} 02/24/2022 22:53:56 - INFO - codeparrot_training - Step 16462: {'lr': 0.0003961243798871066, 'samples': 8429056, 'steps': 16462, 'loss/train': 0.6899096965789795} 02/24/2022 22:54:02 - INFO - codeparrot_training - Step 16463: {'lr': 0.00039611110314639663, 'samples': 8429568, 'steps': 16463, 'loss/train': 2.3329179286956787} 02/24/2022 22:54:07 - INFO - codeparrot_training - Step 16464: {'lr': 0.00039609782577979306, 'samples': 8430080, 'steps': 16464, 'loss/train': 1.103137731552124} 02/24/2022 22:54:11 - INFO - codeparrot_training - Step 16465: {'lr': 0.0003960845477873528, 'samples': 8430592, 'steps': 16465, 'loss/train': 2.858816385269165} 02/24/2022 22:54:16 - INFO - codeparrot_training - Step 16466: {'lr': 0.00039607126916913274, 'samples': 8431104, 'steps': 16466, 'loss/train': 1.550572156906128} 02/24/2022 22:54:20 - INFO - codeparrot_training - Step 16467: {'lr': 0.00039605798992518973, 'samples': 8431616, 'steps': 16467, 'loss/train': 0.09146829694509506} 02/24/2022 22:54:26 - INFO - codeparrot_training - Step 16468: {'lr': 0.00039604471005558065, 'samples': 8432128, 'steps': 16468, 'loss/train': 2.452885627746582} 02/24/2022 22:54:29 - INFO - codeparrot_training - Step 16469: {'lr': 0.0003960314295603624, 'samples': 8432640, 'steps': 16469, 'loss/train': 0.927303671836853} 02/24/2022 22:54:33 - INFO - codeparrot_training - Step 16470: {'lr': 0.00039601814843959193, 'samples': 8433152, 'steps': 16470, 'loss/train': 2.736640214920044} 02/24/2022 22:54:38 - INFO - codeparrot_training - Step 16471: {'lr': 0.00039600486669332603, 'samples': 8433664, 'steps': 16471, 'loss/train': 2.7113940715789795} 02/24/2022 22:54:41 - INFO - codeparrot_training - Step 16472: {'lr': 0.00039599158432162163, 'samples': 8434176, 'steps': 16472, 'loss/train': 1.2780364751815796} 02/24/2022 22:54:48 - INFO - codeparrot_training - Step 16473: {'lr': 0.0003959783013245357, 'samples': 8434688, 'steps': 16473, 'loss/train': 2.6382501125335693} 02/24/2022 22:54:53 - INFO - codeparrot_training - Step 16474: {'lr': 0.000395965017702125, 'samples': 8435200, 'steps': 16474, 'loss/train': 2.0374083518981934} 02/24/2022 22:54:57 - INFO - codeparrot_training - Step 16475: {'lr': 0.00039595173345444656, 'samples': 8435712, 'steps': 16475, 'loss/train': 1.105425477027893} 02/24/2022 22:55:00 - INFO - codeparrot_training - Step 16476: {'lr': 0.0003959384485815573, 'samples': 8436224, 'steps': 16476, 'loss/train': 1.7489532232284546} 02/24/2022 22:55:06 - INFO - codeparrot_training - Step 16477: {'lr': 0.000395925163083514, 'samples': 8436736, 'steps': 16477, 'loss/train': 1.8258665800094604} 02/24/2022 22:55:11 - INFO - codeparrot_training - Step 16478: {'lr': 0.00039591187696037366, 'samples': 8437248, 'steps': 16478, 'loss/train': 2.2233920097351074} 02/24/2022 22:55:15 - INFO - codeparrot_training - Step 16479: {'lr': 0.0003958985902121931, 'samples': 8437760, 'steps': 16479, 'loss/train': 3.151305913925171} 02/24/2022 22:55:21 - INFO - codeparrot_training - Step 16480: {'lr': 0.00039588530283902936, 'samples': 8438272, 'steps': 16480, 'loss/train': 1.291198968887329} 02/24/2022 22:55:24 - INFO - codeparrot_training - Step 16481: {'lr': 0.00039587201484093937, 'samples': 8438784, 'steps': 16481, 'loss/train': 2.1003127098083496} 02/24/2022 22:55:27 - INFO - codeparrot_training - Step 16482: {'lr': 0.0003958587262179799, 'samples': 8439296, 'steps': 16482, 'loss/train': 1.629981279373169} 02/24/2022 22:55:34 - INFO - codeparrot_training - Step 16483: {'lr': 0.00039584543697020804, 'samples': 8439808, 'steps': 16483, 'loss/train': 2.0426506996154785} 02/24/2022 22:55:37 - INFO - codeparrot_training - Step 16484: {'lr': 0.00039583214709768054, 'samples': 8440320, 'steps': 16484, 'loss/train': 1.9853456020355225} 02/24/2022 22:55:43 - INFO - codeparrot_training - Step 16485: {'lr': 0.00039581885660045445, 'samples': 8440832, 'steps': 16485, 'loss/train': 1.8960797786712646} 02/24/2022 22:55:47 - INFO - codeparrot_training - Step 16486: {'lr': 0.0003958055654785867, 'samples': 8441344, 'steps': 16486, 'loss/train': 2.179119825363159} 02/24/2022 22:55:52 - INFO - codeparrot_training - Step 16487: {'lr': 0.0003957922737321343, 'samples': 8441856, 'steps': 16487, 'loss/train': 2.332688331604004} 02/24/2022 22:55:56 - INFO - codeparrot_training - Step 16488: {'lr': 0.00039577898136115397, 'samples': 8442368, 'steps': 16488, 'loss/train': 2.169513463973999} 02/24/2022 22:56:01 - INFO - codeparrot_training - Step 16489: {'lr': 0.00039576568836570283, 'samples': 8442880, 'steps': 16489, 'loss/train': 2.0098929405212402} 02/24/2022 22:56:05 - INFO - codeparrot_training - Step 16490: {'lr': 0.0003957523947458377, 'samples': 8443392, 'steps': 16490, 'loss/train': 1.5059328079223633} 02/24/2022 22:56:10 - INFO - codeparrot_training - Step 16491: {'lr': 0.00039573910050161564, 'samples': 8443904, 'steps': 16491, 'loss/train': 1.3284860849380493} 02/24/2022 22:56:14 - INFO - codeparrot_training - Step 16492: {'lr': 0.0003957258056330936, 'samples': 8444416, 'steps': 16492, 'loss/train': 2.356466054916382} 02/24/2022 22:56:19 - INFO - codeparrot_training - Step 16493: {'lr': 0.00039571251014032847, 'samples': 8444928, 'steps': 16493, 'loss/train': 4.82999849319458} 02/24/2022 22:56:23 - INFO - codeparrot_training - Step 16494: {'lr': 0.00039569921402337715, 'samples': 8445440, 'steps': 16494, 'loss/train': 2.3106446266174316} 02/24/2022 22:56:28 - INFO - codeparrot_training - Step 16495: {'lr': 0.00039568591728229667, 'samples': 8445952, 'steps': 16495, 'loss/train': 0.6705731749534607} 02/24/2022 22:56:32 - INFO - codeparrot_training - Step 16496: {'lr': 0.00039567261991714406, 'samples': 8446464, 'steps': 16496, 'loss/train': 1.957373023033142} 02/24/2022 22:56:37 - INFO - codeparrot_training - Step 16497: {'lr': 0.0003956593219279761, 'samples': 8446976, 'steps': 16497, 'loss/train': 1.888087272644043} 02/24/2022 22:56:41 - INFO - codeparrot_training - Step 16498: {'lr': 0.00039564602331484993, 'samples': 8447488, 'steps': 16498, 'loss/train': 1.583459496498108} 02/24/2022 22:56:46 - INFO - codeparrot_training - Step 16499: {'lr': 0.0003956327240778224, 'samples': 8448000, 'steps': 16499, 'loss/train': 2.1672847270965576} 02/24/2022 22:56:50 - INFO - codeparrot_training - Step 16500: {'lr': 0.00039561942421695057, 'samples': 8448512, 'steps': 16500, 'loss/train': 1.9693784713745117} 02/24/2022 22:56:56 - INFO - codeparrot_training - Step 16501: {'lr': 0.00039560612373229135, 'samples': 8449024, 'steps': 16501, 'loss/train': 1.5708659887313843} 02/24/2022 22:57:00 - INFO - codeparrot_training - Step 16502: {'lr': 0.0003955928226239017, 'samples': 8449536, 'steps': 16502, 'loss/train': 0.6870248317718506} 02/24/2022 22:57:05 - INFO - codeparrot_training - Step 16503: {'lr': 0.00039557952089183863, 'samples': 8450048, 'steps': 16503, 'loss/train': 2.149991035461426} 02/24/2022 22:57:09 - INFO - codeparrot_training - Step 16504: {'lr': 0.00039556621853615914, 'samples': 8450560, 'steps': 16504, 'loss/train': 2.26275897026062} 02/24/2022 22:57:14 - INFO - codeparrot_training - Step 16505: {'lr': 0.0003955529155569202, 'samples': 8451072, 'steps': 16505, 'loss/train': 2.676116466522217} 02/24/2022 22:57:18 - INFO - codeparrot_training - Step 16506: {'lr': 0.0003955396119541788, 'samples': 8451584, 'steps': 16506, 'loss/train': 0.8906749486923218} 02/24/2022 22:57:23 - INFO - codeparrot_training - Step 16507: {'lr': 0.00039552630772799185, 'samples': 8452096, 'steps': 16507, 'loss/train': 1.9210397005081177} 02/24/2022 22:57:27 - INFO - codeparrot_training - Step 16508: {'lr': 0.0003955130028784165, 'samples': 8452608, 'steps': 16508, 'loss/train': 2.290168523788452} 02/24/2022 22:57:33 - INFO - codeparrot_training - Step 16509: {'lr': 0.00039549969740550954, 'samples': 8453120, 'steps': 16509, 'loss/train': 2.4497992992401123} 02/24/2022 22:57:36 - INFO - codeparrot_training - Step 16510: {'lr': 0.00039548639130932816, 'samples': 8453632, 'steps': 16510, 'loss/train': 2.114773988723755} 02/24/2022 22:57:42 - INFO - codeparrot_training - Step 16511: {'lr': 0.00039547308458992927, 'samples': 8454144, 'steps': 16511, 'loss/train': 1.4165219068527222} 02/24/2022 22:57:45 - INFO - codeparrot_training - Step 16512: {'lr': 0.00039545977724736984, 'samples': 8454656, 'steps': 16512, 'loss/train': 2.0097784996032715} 02/24/2022 22:57:53 - INFO - codeparrot_training - Step 16513: {'lr': 0.00039544646928170695, 'samples': 8455168, 'steps': 16513, 'loss/train': 2.3383729457855225} 02/24/2022 22:57:56 - INFO - codeparrot_training - Step 16514: {'lr': 0.0003954331606929976, 'samples': 8455680, 'steps': 16514, 'loss/train': 0.8184428811073303} 02/24/2022 22:58:02 - INFO - codeparrot_training - Step 16515: {'lr': 0.00039541985148129865, 'samples': 8456192, 'steps': 16515, 'loss/train': 1.6029225587844849} 02/24/2022 22:58:05 - INFO - codeparrot_training - Step 16516: {'lr': 0.00039540654164666735, 'samples': 8456704, 'steps': 16516, 'loss/train': 2.4327926635742188} 02/24/2022 22:58:11 - INFO - codeparrot_training - Step 16517: {'lr': 0.00039539323118916055, 'samples': 8457216, 'steps': 16517, 'loss/train': 2.031630516052246} 02/24/2022 22:58:14 - INFO - codeparrot_training - Step 16518: {'lr': 0.0003953799201088353, 'samples': 8457728, 'steps': 16518, 'loss/train': 1.4740921258926392} 02/24/2022 22:58:20 - INFO - codeparrot_training - Step 16519: {'lr': 0.00039536660840574866, 'samples': 8458240, 'steps': 16519, 'loss/train': 2.8197085857391357} 02/24/2022 22:58:23 - INFO - codeparrot_training - Step 16520: {'lr': 0.0003953532960799577, 'samples': 8458752, 'steps': 16520, 'loss/train': 2.4500954151153564} 02/24/2022 22:58:29 - INFO - codeparrot_training - Step 16521: {'lr': 0.00039533998313151926, 'samples': 8459264, 'steps': 16521, 'loss/train': 2.0887598991394043} 02/24/2022 22:58:32 - INFO - codeparrot_training - Step 16522: {'lr': 0.0003953266695604906, 'samples': 8459776, 'steps': 16522, 'loss/train': 2.133399724960327} 02/24/2022 22:58:38 - INFO - codeparrot_training - Step 16523: {'lr': 0.0003953133553669285, 'samples': 8460288, 'steps': 16523, 'loss/train': 1.4083654880523682} 02/24/2022 22:58:42 - INFO - codeparrot_training - Step 16524: {'lr': 0.0003953000405508902, 'samples': 8460800, 'steps': 16524, 'loss/train': 1.5588939189910889} 02/24/2022 22:58:47 - INFO - codeparrot_training - Step 16525: {'lr': 0.00039528672511243256, 'samples': 8461312, 'steps': 16525, 'loss/train': 1.4627403020858765} 02/24/2022 22:58:51 - INFO - codeparrot_training - Step 16526: {'lr': 0.0003952734090516129, 'samples': 8461824, 'steps': 16526, 'loss/train': 2.596207618713379} 02/24/2022 22:58:56 - INFO - codeparrot_training - Step 16527: {'lr': 0.000395260092368488, 'samples': 8462336, 'steps': 16527, 'loss/train': 2.116630792617798} 02/24/2022 22:59:00 - INFO - codeparrot_training - Step 16528: {'lr': 0.000395246775063115, 'samples': 8462848, 'steps': 16528, 'loss/train': 1.810567021369934} 02/24/2022 22:59:05 - INFO - codeparrot_training - Step 16529: {'lr': 0.0003952334571355509, 'samples': 8463360, 'steps': 16529, 'loss/train': 2.4026710987091064} 02/24/2022 22:59:09 - INFO - codeparrot_training - Step 16530: {'lr': 0.0003952201385858528, 'samples': 8463872, 'steps': 16530, 'loss/train': 1.9620431661605835} 02/24/2022 22:59:14 - INFO - codeparrot_training - Step 16531: {'lr': 0.00039520681941407777, 'samples': 8464384, 'steps': 16531, 'loss/train': 2.0947883129119873} 02/24/2022 22:59:21 - INFO - codeparrot_training - Step 16532: {'lr': 0.00039519349962028276, 'samples': 8464896, 'steps': 16532, 'loss/train': 2.0926616191864014} 02/24/2022 22:59:24 - INFO - codeparrot_training - Step 16533: {'lr': 0.000395180179204525, 'samples': 8465408, 'steps': 16533, 'loss/train': 1.0830365419387817} 02/24/2022 22:59:30 - INFO - codeparrot_training - Step 16534: {'lr': 0.0003951668581668614, 'samples': 8465920, 'steps': 16534, 'loss/train': 2.152033805847168} 02/24/2022 22:59:33 - INFO - codeparrot_training - Step 16535: {'lr': 0.0003951535365073491, 'samples': 8466432, 'steps': 16535, 'loss/train': 1.3243255615234375} 02/24/2022 22:59:37 - INFO - codeparrot_training - Step 16536: {'lr': 0.00039514021422604515, 'samples': 8466944, 'steps': 16536, 'loss/train': 2.636953592300415} 02/24/2022 22:59:42 - INFO - codeparrot_training - Step 16537: {'lr': 0.0003951268913230066, 'samples': 8467456, 'steps': 16537, 'loss/train': 2.2711310386657715} 02/24/2022 22:59:45 - INFO - codeparrot_training - Step 16538: {'lr': 0.0003951135677982904, 'samples': 8467968, 'steps': 16538, 'loss/train': 2.9004974365234375} 02/24/2022 22:59:52 - INFO - codeparrot_training - Step 16539: {'lr': 0.000395100243651954, 'samples': 8468480, 'steps': 16539, 'loss/train': 2.2376620769500732} 02/24/2022 22:59:55 - INFO - codeparrot_training - Step 16540: {'lr': 0.00039508691888405403, 'samples': 8468992, 'steps': 16540, 'loss/train': 8.386177062988281} 02/24/2022 23:00:01 - INFO - codeparrot_training - Step 16541: {'lr': 0.0003950735934946478, 'samples': 8469504, 'steps': 16541, 'loss/train': 2.437643051147461} 02/24/2022 23:00:04 - INFO - codeparrot_training - Step 16542: {'lr': 0.0003950602674837924, 'samples': 8470016, 'steps': 16542, 'loss/train': 1.8111413717269897} 02/24/2022 23:00:09 - INFO - codeparrot_training - Step 16543: {'lr': 0.0003950469408515449, 'samples': 8470528, 'steps': 16543, 'loss/train': 1.3587454557418823} 02/24/2022 23:00:13 - INFO - codeparrot_training - Step 16544: {'lr': 0.00039503361359796235, 'samples': 8471040, 'steps': 16544, 'loss/train': 1.3132683038711548} 02/24/2022 23:00:19 - INFO - codeparrot_training - Step 16545: {'lr': 0.00039502028572310186, 'samples': 8471552, 'steps': 16545, 'loss/train': 1.629294514656067} 02/24/2022 23:00:23 - INFO - codeparrot_training - Step 16546: {'lr': 0.0003950069572270205, 'samples': 8472064, 'steps': 16546, 'loss/train': 2.0858993530273438} 02/24/2022 23:00:28 - INFO - codeparrot_training - Step 16547: {'lr': 0.00039499362810977535, 'samples': 8472576, 'steps': 16547, 'loss/train': 2.7927310466766357} 02/24/2022 23:00:32 - INFO - codeparrot_training - Step 16548: {'lr': 0.00039498029837142356, 'samples': 8473088, 'steps': 16548, 'loss/train': 1.405111312866211} 02/24/2022 23:00:37 - INFO - codeparrot_training - Step 16549: {'lr': 0.0003949669680120223, 'samples': 8473600, 'steps': 16549, 'loss/train': 1.546103596687317} 02/24/2022 23:00:41 - INFO - codeparrot_training - Step 16550: {'lr': 0.00039495363703162843, 'samples': 8474112, 'steps': 16550, 'loss/train': 2.0481925010681152} 02/24/2022 23:00:46 - INFO - codeparrot_training - Step 16551: {'lr': 0.00039494030543029925, 'samples': 8474624, 'steps': 16551, 'loss/train': 1.6093497276306152} 02/24/2022 23:00:50 - INFO - codeparrot_training - Step 16552: {'lr': 0.0003949269732080919, 'samples': 8475136, 'steps': 16552, 'loss/train': 2.665602684020996} 02/24/2022 23:00:55 - INFO - codeparrot_training - Step 16553: {'lr': 0.0003949136403650633, 'samples': 8475648, 'steps': 16553, 'loss/train': 2.1096978187561035} 02/24/2022 23:00:59 - INFO - codeparrot_training - Step 16554: {'lr': 0.0003949003069012708, 'samples': 8476160, 'steps': 16554, 'loss/train': 1.4082001447677612} 02/24/2022 23:01:05 - INFO - codeparrot_training - Step 16555: {'lr': 0.0003948869728167713, 'samples': 8476672, 'steps': 16555, 'loss/train': 2.1503357887268066} 02/24/2022 23:01:09 - INFO - codeparrot_training - Step 16556: {'lr': 0.0003948736381116221, 'samples': 8477184, 'steps': 16556, 'loss/train': 0.7984724640846252} 02/24/2022 23:01:14 - INFO - codeparrot_training - Step 16557: {'lr': 0.0003948603027858802, 'samples': 8477696, 'steps': 16557, 'loss/train': 1.2415456771850586} 02/24/2022 23:01:18 - INFO - codeparrot_training - Step 16558: {'lr': 0.00039484696683960276, 'samples': 8478208, 'steps': 16558, 'loss/train': 1.808021068572998} 02/24/2022 23:01:23 - INFO - codeparrot_training - Step 16559: {'lr': 0.0003948336302728469, 'samples': 8478720, 'steps': 16559, 'loss/train': 1.9360491037368774} 02/24/2022 23:01:27 - INFO - codeparrot_training - Step 16560: {'lr': 0.0003948202930856697, 'samples': 8479232, 'steps': 16560, 'loss/train': 1.6901130676269531} 02/24/2022 23:01:33 - INFO - codeparrot_training - Step 16561: {'lr': 0.0003948069552781285, 'samples': 8479744, 'steps': 16561, 'loss/train': 2.4037675857543945} 02/24/2022 23:01:36 - INFO - codeparrot_training - Step 16562: {'lr': 0.00039479361685028016, 'samples': 8480256, 'steps': 16562, 'loss/train': 2.626187562942505} 02/24/2022 23:01:42 - INFO - codeparrot_training - Step 16563: {'lr': 0.00039478027780218193, 'samples': 8480768, 'steps': 16563, 'loss/train': 2.0008575916290283} 02/24/2022 23:01:45 - INFO - codeparrot_training - Step 16564: {'lr': 0.00039476693813389105, 'samples': 8481280, 'steps': 16564, 'loss/train': 2.60748028755188} 02/24/2022 23:01:51 - INFO - codeparrot_training - Step 16565: {'lr': 0.0003947535978454645, 'samples': 8481792, 'steps': 16565, 'loss/train': 1.9211615324020386} 02/24/2022 23:01:54 - INFO - codeparrot_training - Step 16566: {'lr': 0.0003947402569369596, 'samples': 8482304, 'steps': 16566, 'loss/train': 1.8298922777175903} 02/24/2022 23:02:00 - INFO - codeparrot_training - Step 16567: {'lr': 0.0003947269154084333, 'samples': 8482816, 'steps': 16567, 'loss/train': 2.71044921875} 02/24/2022 23:02:03 - INFO - codeparrot_training - Step 16568: {'lr': 0.0003947135732599428, 'samples': 8483328, 'steps': 16568, 'loss/train': 1.496972918510437} 02/24/2022 23:02:09 - INFO - codeparrot_training - Step 16569: {'lr': 0.00039470023049154544, 'samples': 8483840, 'steps': 16569, 'loss/train': 1.642454743385315} 02/24/2022 23:02:13 - INFO - codeparrot_training - Step 16570: {'lr': 0.00039468688710329826, 'samples': 8484352, 'steps': 16570, 'loss/train': 2.7987024784088135} 02/24/2022 23:02:18 - INFO - codeparrot_training - Step 16571: {'lr': 0.0003946735430952583, 'samples': 8484864, 'steps': 16571, 'loss/train': 1.9255377054214478} 02/24/2022 23:02:22 - INFO - codeparrot_training - Step 16572: {'lr': 0.0003946601984674828, 'samples': 8485376, 'steps': 16572, 'loss/train': 1.349665880203247} 02/24/2022 23:02:27 - INFO - codeparrot_training - Step 16573: {'lr': 0.00039464685322002904, 'samples': 8485888, 'steps': 16573, 'loss/train': 1.7452784776687622} 02/24/2022 23:02:31 - INFO - codeparrot_training - Step 16574: {'lr': 0.000394633507352954, 'samples': 8486400, 'steps': 16574, 'loss/train': 2.2632179260253906} 02/24/2022 23:02:36 - INFO - codeparrot_training - Step 16575: {'lr': 0.00039462016086631505, 'samples': 8486912, 'steps': 16575, 'loss/train': 1.8841798305511475} 02/24/2022 23:02:40 - INFO - codeparrot_training - Step 16576: {'lr': 0.00039460681376016915, 'samples': 8487424, 'steps': 16576, 'loss/train': 1.313531517982483} 02/24/2022 23:02:45 - INFO - codeparrot_training - Step 16577: {'lr': 0.0003945934660345736, 'samples': 8487936, 'steps': 16577, 'loss/train': 2.0716559886932373} 02/24/2022 23:02:49 - INFO - codeparrot_training - Step 16578: {'lr': 0.00039458011768958557, 'samples': 8488448, 'steps': 16578, 'loss/train': 1.9649165868759155} 02/24/2022 23:02:54 - INFO - codeparrot_training - Step 16579: {'lr': 0.00039456676872526227, 'samples': 8488960, 'steps': 16579, 'loss/train': 2.3955788612365723} 02/24/2022 23:02:58 - INFO - codeparrot_training - Step 16580: {'lr': 0.00039455341914166074, 'samples': 8489472, 'steps': 16580, 'loss/train': 0.5866998434066772} 02/24/2022 23:03:04 - INFO - codeparrot_training - Step 16581: {'lr': 0.0003945400689388384, 'samples': 8489984, 'steps': 16581, 'loss/train': 1.8699214458465576} 02/24/2022 23:03:07 - INFO - codeparrot_training - Step 16582: {'lr': 0.00039452671811685214, 'samples': 8490496, 'steps': 16582, 'loss/train': 1.2274113893508911} 02/24/2022 23:03:13 - INFO - codeparrot_training - Step 16583: {'lr': 0.00039451336667575945, 'samples': 8491008, 'steps': 16583, 'loss/train': 1.9552297592163086} 02/24/2022 23:03:16 - INFO - codeparrot_training - Step 16584: {'lr': 0.0003945000146156173, 'samples': 8491520, 'steps': 16584, 'loss/train': 2.1564347743988037} 02/24/2022 23:03:22 - INFO - codeparrot_training - Step 16585: {'lr': 0.00039448666193648305, 'samples': 8492032, 'steps': 16585, 'loss/train': 2.390486717224121} 02/24/2022 23:03:26 - INFO - codeparrot_training - Step 16586: {'lr': 0.0003944733086384137, 'samples': 8492544, 'steps': 16586, 'loss/train': 2.964930772781372} 02/24/2022 23:03:31 - INFO - codeparrot_training - Step 16587: {'lr': 0.00039445995472146665, 'samples': 8493056, 'steps': 16587, 'loss/train': 2.803877592086792} 02/24/2022 23:03:34 - INFO - codeparrot_training - Step 16588: {'lr': 0.000394446600185699, 'samples': 8493568, 'steps': 16588, 'loss/train': 1.9944179058074951} 02/24/2022 23:03:40 - INFO - codeparrot_training - Step 16589: {'lr': 0.000394433245031168, 'samples': 8494080, 'steps': 16589, 'loss/train': 3.2032272815704346} 02/24/2022 23:03:43 - INFO - codeparrot_training - Step 16590: {'lr': 0.0003944198892579309, 'samples': 8494592, 'steps': 16590, 'loss/train': 1.1729189157485962} 02/24/2022 23:03:50 - INFO - codeparrot_training - Step 16591: {'lr': 0.0003944065328660447, 'samples': 8495104, 'steps': 16591, 'loss/train': 2.2717106342315674} 02/24/2022 23:03:53 - INFO - codeparrot_training - Step 16592: {'lr': 0.0003943931758555669, 'samples': 8495616, 'steps': 16592, 'loss/train': 1.2033859491348267} 02/24/2022 23:03:59 - INFO - codeparrot_training - Step 16593: {'lr': 0.00039437981822655453, 'samples': 8496128, 'steps': 16593, 'loss/train': 1.4606666564941406} 02/24/2022 23:04:02 - INFO - codeparrot_training - Step 16594: {'lr': 0.00039436645997906487, 'samples': 8496640, 'steps': 16594, 'loss/train': 3.0710957050323486} 02/24/2022 23:04:08 - INFO - codeparrot_training - Step 16595: {'lr': 0.00039435310111315513, 'samples': 8497152, 'steps': 16595, 'loss/train': 1.5867692232131958} 02/24/2022 23:04:11 - INFO - codeparrot_training - Step 16596: {'lr': 0.00039433974162888266, 'samples': 8497664, 'steps': 16596, 'loss/train': 1.9993149042129517} 02/24/2022 23:04:17 - INFO - codeparrot_training - Step 16597: {'lr': 0.0003943263815263044, 'samples': 8498176, 'steps': 16597, 'loss/train': 0.9557214379310608} 02/24/2022 23:04:20 - INFO - codeparrot_training - Step 16598: {'lr': 0.0003943130208054778, 'samples': 8498688, 'steps': 16598, 'loss/train': 2.130176544189453} 02/24/2022 23:04:26 - INFO - codeparrot_training - Step 16599: {'lr': 0.0003942996594664601, 'samples': 8499200, 'steps': 16599, 'loss/train': 1.9649536609649658} 02/24/2022 23:04:29 - INFO - codeparrot_training - Step 16600: {'lr': 0.00039428629750930846, 'samples': 8499712, 'steps': 16600, 'loss/train': 2.0388505458831787} 02/24/2022 23:04:35 - INFO - codeparrot_training - Step 16601: {'lr': 0.0003942729349340801, 'samples': 8500224, 'steps': 16601, 'loss/train': 1.426533579826355} 02/24/2022 23:04:39 - INFO - codeparrot_training - Step 16602: {'lr': 0.00039425957174083224, 'samples': 8500736, 'steps': 16602, 'loss/train': 2.302284002304077} 02/24/2022 23:04:44 - INFO - codeparrot_training - Step 16603: {'lr': 0.0003942462079296223, 'samples': 8501248, 'steps': 16603, 'loss/train': 2.7168502807617188} 02/24/2022 23:04:48 - INFO - codeparrot_training - Step 16604: {'lr': 0.00039423284350050735, 'samples': 8501760, 'steps': 16604, 'loss/train': 2.181910514831543} 02/24/2022 23:04:54 - INFO - codeparrot_training - Step 16605: {'lr': 0.00039421947845354476, 'samples': 8502272, 'steps': 16605, 'loss/train': 1.5540157556533813} 02/24/2022 23:04:59 - INFO - codeparrot_training - Step 16606: {'lr': 0.0003942061127887916, 'samples': 8502784, 'steps': 16606, 'loss/train': 1.2098674774169922} 02/24/2022 23:05:03 - INFO - codeparrot_training - Step 16607: {'lr': 0.00039419274650630536, 'samples': 8503296, 'steps': 16607, 'loss/train': 1.5457173585891724} 02/24/2022 23:05:08 - INFO - codeparrot_training - Step 16608: {'lr': 0.00039417937960614316, 'samples': 8503808, 'steps': 16608, 'loss/train': 2.610319137573242} 02/24/2022 23:05:12 - INFO - codeparrot_training - Step 16609: {'lr': 0.0003941660120883622, 'samples': 8504320, 'steps': 16609, 'loss/train': 1.9802271127700806} 02/24/2022 23:05:18 - INFO - codeparrot_training - Step 16610: {'lr': 0.0003941526439530199, 'samples': 8504832, 'steps': 16610, 'loss/train': 1.3905479907989502} 02/24/2022 23:05:21 - INFO - codeparrot_training - Step 16611: {'lr': 0.00039413927520017347, 'samples': 8505344, 'steps': 16611, 'loss/train': 2.3664402961730957} 02/24/2022 23:05:25 - INFO - codeparrot_training - Step 16612: {'lr': 0.00039412590582988007, 'samples': 8505856, 'steps': 16612, 'loss/train': 3.08532452583313} 02/24/2022 23:05:30 - INFO - codeparrot_training - Step 16613: {'lr': 0.00039411253584219707, 'samples': 8506368, 'steps': 16613, 'loss/train': 1.6768769025802612} 02/24/2022 23:05:34 - INFO - codeparrot_training - Step 16614: {'lr': 0.0003940991652371818, 'samples': 8506880, 'steps': 16614, 'loss/train': 2.472137451171875} 02/24/2022 23:05:40 - INFO - codeparrot_training - Step 16615: {'lr': 0.0003940857940148914, 'samples': 8507392, 'steps': 16615, 'loss/train': 2.0025970935821533} 02/24/2022 23:05:44 - INFO - codeparrot_training - Step 16616: {'lr': 0.00039407242217538317, 'samples': 8507904, 'steps': 16616, 'loss/train': 0.3123548924922943} 02/24/2022 23:05:49 - INFO - codeparrot_training - Step 16617: {'lr': 0.00039405904971871454, 'samples': 8508416, 'steps': 16617, 'loss/train': 1.2360575199127197} 02/24/2022 23:05:53 - INFO - codeparrot_training - Step 16618: {'lr': 0.00039404567664494264, 'samples': 8508928, 'steps': 16618, 'loss/train': 2.026491641998291} 02/24/2022 23:05:58 - INFO - codeparrot_training - Step 16619: {'lr': 0.0003940323029541248, 'samples': 8509440, 'steps': 16619, 'loss/train': 2.540133476257324} 02/24/2022 23:06:02 - INFO - codeparrot_training - Step 16620: {'lr': 0.00039401892864631826, 'samples': 8509952, 'steps': 16620, 'loss/train': 1.6243343353271484} 02/24/2022 23:06:07 - INFO - codeparrot_training - Step 16621: {'lr': 0.0003940055537215804, 'samples': 8510464, 'steps': 16621, 'loss/train': 1.334533929824829} 02/24/2022 23:06:11 - INFO - codeparrot_training - Step 16622: {'lr': 0.0003939921781799685, 'samples': 8510976, 'steps': 16622, 'loss/train': 2.263381242752075} 02/24/2022 23:06:16 - INFO - codeparrot_training - Step 16623: {'lr': 0.0003939788020215398, 'samples': 8511488, 'steps': 16623, 'loss/train': 1.5070136785507202} 02/24/2022 23:06:20 - INFO - codeparrot_training - Step 16624: {'lr': 0.0003939654252463517, 'samples': 8512000, 'steps': 16624, 'loss/train': 1.554828405380249} 02/24/2022 23:06:26 - INFO - codeparrot_training - Step 16625: {'lr': 0.00039395204785446137, 'samples': 8512512, 'steps': 16625, 'loss/train': 2.508723497390747} 02/24/2022 23:06:29 - INFO - codeparrot_training - Step 16626: {'lr': 0.00039393866984592616, 'samples': 8513024, 'steps': 16626, 'loss/train': 1.5081697702407837} 02/24/2022 23:06:35 - INFO - codeparrot_training - Step 16627: {'lr': 0.00039392529122080343, 'samples': 8513536, 'steps': 16627, 'loss/train': 2.061289072036743} 02/24/2022 23:06:39 - INFO - codeparrot_training - Step 16628: {'lr': 0.0003939119119791504, 'samples': 8514048, 'steps': 16628, 'loss/train': 2.5437963008880615} 02/24/2022 23:06:44 - INFO - codeparrot_training - Step 16629: {'lr': 0.0003938985321210245, 'samples': 8514560, 'steps': 16629, 'loss/train': 1.99045729637146} 02/24/2022 23:06:48 - INFO - codeparrot_training - Step 16630: {'lr': 0.00039388515164648293, 'samples': 8515072, 'steps': 16630, 'loss/train': 2.375284194946289} 02/24/2022 23:06:53 - INFO - codeparrot_training - Step 16631: {'lr': 0.0003938717705555831, 'samples': 8515584, 'steps': 16631, 'loss/train': 2.666696071624756} 02/24/2022 23:06:57 - INFO - codeparrot_training - Step 16632: {'lr': 0.0003938583888483823, 'samples': 8516096, 'steps': 16632, 'loss/train': 1.9341421127319336} 02/24/2022 23:07:02 - INFO - codeparrot_training - Step 16633: {'lr': 0.0003938450065249378, 'samples': 8516608, 'steps': 16633, 'loss/train': 1.1969630718231201} 02/24/2022 23:07:06 - INFO - codeparrot_training - Step 16634: {'lr': 0.00039383162358530696, 'samples': 8517120, 'steps': 16634, 'loss/train': 4.823916912078857} 02/24/2022 23:07:11 - INFO - codeparrot_training - Step 16635: {'lr': 0.0003938182400295471, 'samples': 8517632, 'steps': 16635, 'loss/train': 1.5374633073806763} 02/24/2022 23:07:15 - INFO - codeparrot_training - Step 16636: {'lr': 0.00039380485585771563, 'samples': 8518144, 'steps': 16636, 'loss/train': 2.1145589351654053} 02/24/2022 23:07:22 - INFO - codeparrot_training - Step 16637: {'lr': 0.00039379147106986985, 'samples': 8518656, 'steps': 16637, 'loss/train': 1.8385825157165527} 02/24/2022 23:07:25 - INFO - codeparrot_training - Step 16638: {'lr': 0.00039377808566606697, 'samples': 8519168, 'steps': 16638, 'loss/train': 3.803720235824585} 02/24/2022 23:07:31 - INFO - codeparrot_training - Step 16639: {'lr': 0.00039376469964636445, 'samples': 8519680, 'steps': 16639, 'loss/train': 0.5922945737838745} 02/24/2022 23:07:34 - INFO - codeparrot_training - Step 16640: {'lr': 0.0003937513130108197, 'samples': 8520192, 'steps': 16640, 'loss/train': 1.630444049835205} 02/24/2022 23:07:40 - INFO - codeparrot_training - Step 16641: {'lr': 0.00039373792575948986, 'samples': 8520704, 'steps': 16641, 'loss/train': 0.5477926731109619} 02/24/2022 23:07:43 - INFO - codeparrot_training - Step 16642: {'lr': 0.00039372453789243245, 'samples': 8521216, 'steps': 16642, 'loss/train': 1.720116138458252} 02/24/2022 23:07:49 - INFO - codeparrot_training - Step 16643: {'lr': 0.0003937111494097047, 'samples': 8521728, 'steps': 16643, 'loss/train': 0.5263066291809082} 02/24/2022 23:07:52 - INFO - codeparrot_training - Step 16644: {'lr': 0.0003936977603113641, 'samples': 8522240, 'steps': 16644, 'loss/train': 2.607430934906006} 02/24/2022 23:07:58 - INFO - codeparrot_training - Step 16645: {'lr': 0.00039368437059746785, 'samples': 8522752, 'steps': 16645, 'loss/train': 1.9224461317062378} 02/24/2022 23:08:01 - INFO - codeparrot_training - Step 16646: {'lr': 0.0003936709802680734, 'samples': 8523264, 'steps': 16646, 'loss/train': 2.295910596847534} 02/24/2022 23:08:08 - INFO - codeparrot_training - Step 16647: {'lr': 0.0003936575893232381, 'samples': 8523776, 'steps': 16647, 'loss/train': 1.4181220531463623} 02/24/2022 23:08:11 - INFO - codeparrot_training - Step 16648: {'lr': 0.0003936441977630193, 'samples': 8524288, 'steps': 16648, 'loss/train': 1.8561968803405762} 02/24/2022 23:08:17 - INFO - codeparrot_training - Step 16649: {'lr': 0.0003936308055874744, 'samples': 8524800, 'steps': 16649, 'loss/train': 2.004794120788574} 02/24/2022 23:08:20 - INFO - codeparrot_training - Step 16650: {'lr': 0.00039361741279666065, 'samples': 8525312, 'steps': 16650, 'loss/train': 1.9476447105407715} 02/24/2022 23:08:26 - INFO - codeparrot_training - Step 16651: {'lr': 0.0003936040193906356, 'samples': 8525824, 'steps': 16651, 'loss/train': 1.379144549369812} 02/24/2022 23:08:30 - INFO - codeparrot_training - Step 16652: {'lr': 0.00039359062536945645, 'samples': 8526336, 'steps': 16652, 'loss/train': 2.1489012241363525} 02/24/2022 23:08:33 - INFO - codeparrot_training - Step 16653: {'lr': 0.00039357723073318076, 'samples': 8526848, 'steps': 16653, 'loss/train': 2.451748847961426} 02/24/2022 23:08:39 - INFO - codeparrot_training - Step 16654: {'lr': 0.0003935638354818657, 'samples': 8527360, 'steps': 16654, 'loss/train': 1.9240946769714355} 02/24/2022 23:08:42 - INFO - codeparrot_training - Step 16655: {'lr': 0.0003935504396155688, 'samples': 8527872, 'steps': 16655, 'loss/train': 3.0455963611602783} 02/24/2022 23:08:48 - INFO - codeparrot_training - Step 16656: {'lr': 0.00039353704313434745, 'samples': 8528384, 'steps': 16656, 'loss/train': 2.579360246658325} 02/24/2022 23:08:51 - INFO - codeparrot_training - Step 16657: {'lr': 0.000393523646038259, 'samples': 8528896, 'steps': 16657, 'loss/train': 1.9909262657165527} 02/24/2022 23:08:57 - INFO - codeparrot_training - Step 16658: {'lr': 0.0003935102483273607, 'samples': 8529408, 'steps': 16658, 'loss/train': 1.5683646202087402} 02/24/2022 23:09:01 - INFO - codeparrot_training - Step 16659: {'lr': 0.0003934968500017101, 'samples': 8529920, 'steps': 16659, 'loss/train': 2.350641965866089} 02/24/2022 23:09:07 - INFO - codeparrot_training - Step 16660: {'lr': 0.0003934834510613646, 'samples': 8530432, 'steps': 16660, 'loss/train': 2.930534601211548} 02/24/2022 23:09:10 - INFO - codeparrot_training - Step 16661: {'lr': 0.00039347005150638156, 'samples': 8530944, 'steps': 16661, 'loss/train': 0.17369619011878967} 02/24/2022 23:09:16 - INFO - codeparrot_training - Step 16662: {'lr': 0.0003934566513368183, 'samples': 8531456, 'steps': 16662, 'loss/train': 1.937572717666626} 02/24/2022 23:09:19 - INFO - codeparrot_training - Step 16663: {'lr': 0.00039344325055273236, 'samples': 8531968, 'steps': 16663, 'loss/train': 1.3789302110671997} 02/24/2022 23:09:25 - INFO - codeparrot_training - Step 16664: {'lr': 0.0003934298491541811, 'samples': 8532480, 'steps': 16664, 'loss/train': 1.9850187301635742} 02/24/2022 23:09:29 - INFO - codeparrot_training - Step 16665: {'lr': 0.00039341644714122195, 'samples': 8532992, 'steps': 16665, 'loss/train': 1.5947589874267578} 02/24/2022 23:09:34 - INFO - codeparrot_training - Step 16666: {'lr': 0.00039340304451391216, 'samples': 8533504, 'steps': 16666, 'loss/train': 0.872013509273529} 02/24/2022 23:09:38 - INFO - codeparrot_training - Step 16667: {'lr': 0.00039338964127230935, 'samples': 8534016, 'steps': 16667, 'loss/train': 1.8149328231811523} 02/24/2022 23:09:43 - INFO - codeparrot_training - Step 16668: {'lr': 0.00039337623741647084, 'samples': 8534528, 'steps': 16668, 'loss/train': 2.2334938049316406} 02/24/2022 23:09:47 - INFO - codeparrot_training - Step 16669: {'lr': 0.000393362832946454, 'samples': 8535040, 'steps': 16669, 'loss/train': 2.6417410373687744} 02/24/2022 23:09:52 - INFO - codeparrot_training - Step 16670: {'lr': 0.0003933494278623164, 'samples': 8535552, 'steps': 16670, 'loss/train': 1.5269017219543457} 02/24/2022 23:09:56 - INFO - codeparrot_training - Step 16671: {'lr': 0.0003933360221641153, 'samples': 8536064, 'steps': 16671, 'loss/train': 1.4327607154846191} 02/24/2022 23:10:01 - INFO - codeparrot_training - Step 16672: {'lr': 0.0003933226158519082, 'samples': 8536576, 'steps': 16672, 'loss/train': 1.799267053604126} 02/24/2022 23:10:05 - INFO - codeparrot_training - Step 16673: {'lr': 0.0003933092089257525, 'samples': 8537088, 'steps': 16673, 'loss/train': 2.336268901824951} 02/24/2022 23:10:11 - INFO - codeparrot_training - Step 16674: {'lr': 0.0003932958013857057, 'samples': 8537600, 'steps': 16674, 'loss/train': 1.5832922458648682} 02/24/2022 23:10:14 - INFO - codeparrot_training - Step 16675: {'lr': 0.0003932823932318252, 'samples': 8538112, 'steps': 16675, 'loss/train': 2.7384278774261475} 02/24/2022 23:10:20 - INFO - codeparrot_training - Step 16676: {'lr': 0.0003932689844641684, 'samples': 8538624, 'steps': 16676, 'loss/train': 2.256833553314209} 02/24/2022 23:10:23 - INFO - codeparrot_training - Step 16677: {'lr': 0.00039325557508279276, 'samples': 8539136, 'steps': 16677, 'loss/train': 1.615507960319519} 02/24/2022 23:10:29 - INFO - codeparrot_training - Step 16678: {'lr': 0.00039324216508775567, 'samples': 8539648, 'steps': 16678, 'loss/train': 1.879104733467102} 02/24/2022 23:10:32 - INFO - codeparrot_training - Step 16679: {'lr': 0.0003932287544791148, 'samples': 8540160, 'steps': 16679, 'loss/train': 2.7534518241882324} 02/24/2022 23:10:38 - INFO - codeparrot_training - Step 16680: {'lr': 0.00039321534325692726, 'samples': 8540672, 'steps': 16680, 'loss/train': 2.967466115951538} 02/24/2022 23:10:41 - INFO - codeparrot_training - Step 16681: {'lr': 0.0003932019314212507, 'samples': 8541184, 'steps': 16681, 'loss/train': 2.105475664138794} 02/24/2022 23:10:47 - INFO - codeparrot_training - Step 16682: {'lr': 0.0003931885189721426, 'samples': 8541696, 'steps': 16682, 'loss/train': 1.5834599733352661} 02/24/2022 23:10:50 - INFO - codeparrot_training - Step 16683: {'lr': 0.00039317510590966033, 'samples': 8542208, 'steps': 16683, 'loss/train': 0.7416520714759827} 02/24/2022 23:10:57 - INFO - codeparrot_training - Step 16684: {'lr': 0.0003931616922338613, 'samples': 8542720, 'steps': 16684, 'loss/train': 1.5058794021606445} 02/24/2022 23:11:00 - INFO - codeparrot_training - Step 16685: {'lr': 0.00039314827794480314, 'samples': 8543232, 'steps': 16685, 'loss/train': 2.1089189052581787} 02/24/2022 23:11:06 - INFO - codeparrot_training - Step 16686: {'lr': 0.00039313486304254315, 'samples': 8543744, 'steps': 16686, 'loss/train': 2.7367444038391113} 02/24/2022 23:11:09 - INFO - codeparrot_training - Step 16687: {'lr': 0.00039312144752713885, 'samples': 8544256, 'steps': 16687, 'loss/train': 1.6765453815460205} 02/24/2022 23:11:15 - INFO - codeparrot_training - Step 16688: {'lr': 0.00039310803139864777, 'samples': 8544768, 'steps': 16688, 'loss/train': 1.931235432624817} 02/24/2022 23:11:18 - INFO - codeparrot_training - Step 16689: {'lr': 0.00039309461465712725, 'samples': 8545280, 'steps': 16689, 'loss/train': 1.2274912595748901} 02/24/2022 23:11:24 - INFO - codeparrot_training - Step 16690: {'lr': 0.00039308119730263494, 'samples': 8545792, 'steps': 16690, 'loss/train': 2.0927178859710693} 02/24/2022 23:11:27 - INFO - codeparrot_training - Step 16691: {'lr': 0.00039306777933522806, 'samples': 8546304, 'steps': 16691, 'loss/train': 2.453559398651123} 02/24/2022 23:11:33 - INFO - codeparrot_training - Step 16692: {'lr': 0.00039305436075496436, 'samples': 8546816, 'steps': 16692, 'loss/train': 1.3919786214828491} 02/24/2022 23:11:36 - INFO - codeparrot_training - Step 16693: {'lr': 0.0003930409415619012, 'samples': 8547328, 'steps': 16693, 'loss/train': 1.6564058065414429} 02/24/2022 23:11:42 - INFO - codeparrot_training - Step 16694: {'lr': 0.000393027521756096, 'samples': 8547840, 'steps': 16694, 'loss/train': 1.0032819509506226} 02/24/2022 23:11:46 - INFO - codeparrot_training - Step 16695: {'lr': 0.0003930141013376064, 'samples': 8548352, 'steps': 16695, 'loss/train': 2.062098979949951} 02/24/2022 23:11:51 - INFO - codeparrot_training - Step 16696: {'lr': 0.00039300068030648976, 'samples': 8548864, 'steps': 16696, 'loss/train': 1.3452891111373901} 02/24/2022 23:11:55 - INFO - codeparrot_training - Step 16697: {'lr': 0.0003929872586628036, 'samples': 8549376, 'steps': 16697, 'loss/train': 2.5055952072143555} 02/24/2022 23:12:00 - INFO - codeparrot_training - Step 16698: {'lr': 0.00039297383640660545, 'samples': 8549888, 'steps': 16698, 'loss/train': 1.6175209283828735} 02/24/2022 23:12:04 - INFO - codeparrot_training - Step 16699: {'lr': 0.0003929604135379528, 'samples': 8550400, 'steps': 16699, 'loss/train': 1.2649298906326294} 02/24/2022 23:12:09 - INFO - codeparrot_training - Step 16700: {'lr': 0.000392946990056903, 'samples': 8550912, 'steps': 16700, 'loss/train': 1.211811900138855} 02/24/2022 23:12:13 - INFO - codeparrot_training - Step 16701: {'lr': 0.0003929335659635139, 'samples': 8551424, 'steps': 16701, 'loss/train': 1.6003845930099487} 02/24/2022 23:12:18 - INFO - codeparrot_training - Step 16702: {'lr': 0.00039292014125784266, 'samples': 8551936, 'steps': 16702, 'loss/train': 1.6087394952774048} 02/24/2022 23:12:22 - INFO - codeparrot_training - Step 16703: {'lr': 0.00039290671593994697, 'samples': 8552448, 'steps': 16703, 'loss/train': 1.4104448556900024} 02/24/2022 23:12:28 - INFO - codeparrot_training - Step 16704: {'lr': 0.0003928932900098842, 'samples': 8552960, 'steps': 16704, 'loss/train': 3.354512929916382} 02/24/2022 23:12:31 - INFO - codeparrot_training - Step 16705: {'lr': 0.00039287986346771205, 'samples': 8553472, 'steps': 16705, 'loss/train': 1.54456627368927} 02/24/2022 23:12:37 - INFO - codeparrot_training - Step 16706: {'lr': 0.0003928664363134879, 'samples': 8553984, 'steps': 16706, 'loss/train': 1.5391348600387573} 02/24/2022 23:12:41 - INFO - codeparrot_training - Step 16707: {'lr': 0.00039285300854726926, 'samples': 8554496, 'steps': 16707, 'loss/train': 9.982333183288574} 02/24/2022 23:12:46 - INFO - codeparrot_training - Step 16708: {'lr': 0.00039283958016911373, 'samples': 8555008, 'steps': 16708, 'loss/train': 1.522818922996521} 02/24/2022 23:12:50 - INFO - codeparrot_training - Step 16709: {'lr': 0.00039282615117907884, 'samples': 8555520, 'steps': 16709, 'loss/train': 1.8078899383544922} 02/24/2022 23:12:55 - INFO - codeparrot_training - Step 16710: {'lr': 0.00039281272157722205, 'samples': 8556032, 'steps': 16710, 'loss/train': 2.7611310482025146} 02/24/2022 23:12:59 - INFO - codeparrot_training - Step 16711: {'lr': 0.0003927992913636008, 'samples': 8556544, 'steps': 16711, 'loss/train': 2.0730113983154297} 02/24/2022 23:13:04 - INFO - codeparrot_training - Step 16712: {'lr': 0.0003927858605382728, 'samples': 8557056, 'steps': 16712, 'loss/train': 3.581165075302124} 02/24/2022 23:13:08 - INFO - codeparrot_training - Step 16713: {'lr': 0.0003927724291012955, 'samples': 8557568, 'steps': 16713, 'loss/train': 2.925849199295044} 02/24/2022 23:13:13 - INFO - codeparrot_training - Step 16714: {'lr': 0.00039275899705272656, 'samples': 8558080, 'steps': 16714, 'loss/train': 3.4336538314819336} 02/24/2022 23:13:17 - INFO - codeparrot_training - Step 16715: {'lr': 0.00039274556439262325, 'samples': 8558592, 'steps': 16715, 'loss/train': 2.148698568344116} 02/24/2022 23:13:22 - INFO - codeparrot_training - Step 16716: {'lr': 0.0003927321311210434, 'samples': 8559104, 'steps': 16716, 'loss/train': 2.845339059829712} 02/24/2022 23:13:26 - INFO - codeparrot_training - Step 16717: {'lr': 0.00039271869723804434, 'samples': 8559616, 'steps': 16717, 'loss/train': 0.6309954524040222} 02/24/2022 23:13:31 - INFO - codeparrot_training - Step 16718: {'lr': 0.0003927052627436837, 'samples': 8560128, 'steps': 16718, 'loss/train': 2.1624178886413574} 02/24/2022 23:13:35 - INFO - codeparrot_training - Step 16719: {'lr': 0.000392691827638019, 'samples': 8560640, 'steps': 16719, 'loss/train': 2.5281877517700195} 02/24/2022 23:13:41 - INFO - codeparrot_training - Step 16720: {'lr': 0.000392678391921108, 'samples': 8561152, 'steps': 16720, 'loss/train': 2.620974540710449} 02/24/2022 23:13:45 - INFO - codeparrot_training - Step 16721: {'lr': 0.00039266495559300786, 'samples': 8561664, 'steps': 16721, 'loss/train': 2.650250196456909} 02/24/2022 23:13:50 - INFO - codeparrot_training - Step 16722: {'lr': 0.00039265151865377644, 'samples': 8562176, 'steps': 16722, 'loss/train': 2.0051259994506836} 02/24/2022 23:13:54 - INFO - codeparrot_training - Step 16723: {'lr': 0.0003926380811034712, 'samples': 8562688, 'steps': 16723, 'loss/train': 1.9259918928146362} 02/24/2022 23:13:59 - INFO - codeparrot_training - Step 16724: {'lr': 0.0003926246429421497, 'samples': 8563200, 'steps': 16724, 'loss/train': 2.457704544067383} 02/24/2022 23:14:03 - INFO - codeparrot_training - Step 16725: {'lr': 0.0003926112041698696, 'samples': 8563712, 'steps': 16725, 'loss/train': 3.1315267086029053} 02/24/2022 23:14:08 - INFO - codeparrot_training - Step 16726: {'lr': 0.0003925977647866883, 'samples': 8564224, 'steps': 16726, 'loss/train': 2.3196165561676025} 02/24/2022 23:14:12 - INFO - codeparrot_training - Step 16727: {'lr': 0.0003925843247926635, 'samples': 8564736, 'steps': 16727, 'loss/train': 2.237645387649536} 02/24/2022 23:14:17 - INFO - codeparrot_training - Step 16728: {'lr': 0.00039257088418785267, 'samples': 8565248, 'steps': 16728, 'loss/train': 1.8360320329666138} 02/24/2022 23:14:21 - INFO - codeparrot_training - Step 16729: {'lr': 0.00039255744297231354, 'samples': 8565760, 'steps': 16729, 'loss/train': 2.4263439178466797} 02/24/2022 23:14:27 - INFO - codeparrot_training - Step 16730: {'lr': 0.0003925440011461035, 'samples': 8566272, 'steps': 16730, 'loss/train': 2.2055110931396484} 02/24/2022 23:14:31 - INFO - codeparrot_training - Step 16731: {'lr': 0.0003925305587092802, 'samples': 8566784, 'steps': 16731, 'loss/train': 1.0306373834609985} 02/24/2022 23:14:36 - INFO - codeparrot_training - Step 16732: {'lr': 0.00039251711566190133, 'samples': 8567296, 'steps': 16732, 'loss/train': 1.720516324043274} 02/24/2022 23:14:40 - INFO - codeparrot_training - Step 16733: {'lr': 0.0003925036720040244, 'samples': 8567808, 'steps': 16733, 'loss/train': 2.2145564556121826} 02/24/2022 23:14:45 - INFO - codeparrot_training - Step 16734: {'lr': 0.000392490227735707, 'samples': 8568320, 'steps': 16734, 'loss/train': 1.176186203956604} 02/24/2022 23:14:49 - INFO - codeparrot_training - Step 16735: {'lr': 0.0003924767828570066, 'samples': 8568832, 'steps': 16735, 'loss/train': 0.7137700915336609} 02/24/2022 23:14:54 - INFO - codeparrot_training - Step 16736: {'lr': 0.00039246333736798095, 'samples': 8569344, 'steps': 16736, 'loss/train': 2.636673927307129} 02/24/2022 23:14:57 - INFO - codeparrot_training - Step 16737: {'lr': 0.00039244989126868755, 'samples': 8569856, 'steps': 16737, 'loss/train': 0.8950685858726501} 02/24/2022 23:15:03 - INFO - codeparrot_training - Step 16738: {'lr': 0.0003924364445591842, 'samples': 8570368, 'steps': 16738, 'loss/train': 2.578615188598633} 02/24/2022 23:15:06 - INFO - codeparrot_training - Step 16739: {'lr': 0.0003924229972395282, 'samples': 8570880, 'steps': 16739, 'loss/train': 1.9543473720550537} 02/24/2022 23:15:13 - INFO - codeparrot_training - Step 16740: {'lr': 0.00039240954930977744, 'samples': 8571392, 'steps': 16740, 'loss/train': 1.864445447921753} 02/24/2022 23:15:16 - INFO - codeparrot_training - Step 16741: {'lr': 0.0003923961007699893, 'samples': 8571904, 'steps': 16741, 'loss/train': 2.237811326980591} 02/24/2022 23:15:22 - INFO - codeparrot_training - Step 16742: {'lr': 0.00039238265162022147, 'samples': 8572416, 'steps': 16742, 'loss/train': 1.9068998098373413} 02/24/2022 23:15:25 - INFO - codeparrot_training - Step 16743: {'lr': 0.0003923692018605316, 'samples': 8572928, 'steps': 16743, 'loss/train': 2.144944429397583} 02/24/2022 23:15:31 - INFO - codeparrot_training - Step 16744: {'lr': 0.0003923557514909773, 'samples': 8573440, 'steps': 16744, 'loss/train': 3.1549019813537598} 02/24/2022 23:15:34 - INFO - codeparrot_training - Step 16745: {'lr': 0.00039234230051161614, 'samples': 8573952, 'steps': 16745, 'loss/train': 0.9322651624679565} 02/24/2022 23:15:40 - INFO - codeparrot_training - Step 16746: {'lr': 0.00039232884892250575, 'samples': 8574464, 'steps': 16746, 'loss/train': 2.5181801319122314} 02/24/2022 23:15:43 - INFO - codeparrot_training - Step 16747: {'lr': 0.00039231539672370376, 'samples': 8574976, 'steps': 16747, 'loss/train': 2.181405544281006} 02/24/2022 23:15:49 - INFO - codeparrot_training - Step 16748: {'lr': 0.00039230194391526784, 'samples': 8575488, 'steps': 16748, 'loss/train': 2.0367963314056396} 02/24/2022 23:15:52 - INFO - codeparrot_training - Step 16749: {'lr': 0.0003922884904972556, 'samples': 8576000, 'steps': 16749, 'loss/train': 2.0779237747192383} 02/24/2022 23:15:58 - INFO - codeparrot_training - Step 16750: {'lr': 0.0003922750364697246, 'samples': 8576512, 'steps': 16750, 'loss/train': 3.7160446643829346} 02/24/2022 23:16:02 - INFO - codeparrot_training - Step 16751: {'lr': 0.0003922615818327325, 'samples': 8577024, 'steps': 16751, 'loss/train': 2.6534552574157715} 02/24/2022 23:16:08 - INFO - codeparrot_training - Step 16752: {'lr': 0.000392248126586337, 'samples': 8577536, 'steps': 16752, 'loss/train': 2.5692691802978516} 02/24/2022 23:16:11 - INFO - codeparrot_training - Step 16753: {'lr': 0.0003922346707305957, 'samples': 8578048, 'steps': 16753, 'loss/train': 0.07850679010152817} 02/24/2022 23:16:17 - INFO - codeparrot_training - Step 16754: {'lr': 0.00039222121426556617, 'samples': 8578560, 'steps': 16754, 'loss/train': 2.6701011657714844} 02/24/2022 23:16:21 - INFO - codeparrot_training - Step 16755: {'lr': 0.0003922077571913062, 'samples': 8579072, 'steps': 16755, 'loss/train': 1.1906167268753052} 02/24/2022 23:16:26 - INFO - codeparrot_training - Step 16756: {'lr': 0.00039219429950787326, 'samples': 8579584, 'steps': 16756, 'loss/train': 2.0138866901397705} 02/24/2022 23:16:30 - INFO - codeparrot_training - Step 16757: {'lr': 0.0003921808412153252, 'samples': 8580096, 'steps': 16757, 'loss/train': 2.505519151687622} 02/24/2022 23:16:35 - INFO - codeparrot_training - Step 16758: {'lr': 0.0003921673823137195, 'samples': 8580608, 'steps': 16758, 'loss/train': 2.7859370708465576} 02/24/2022 23:16:39 - INFO - codeparrot_training - Step 16759: {'lr': 0.00039215392280311383, 'samples': 8581120, 'steps': 16759, 'loss/train': 4.237369060516357} 02/24/2022 23:16:44 - INFO - codeparrot_training - Step 16760: {'lr': 0.000392140462683566, 'samples': 8581632, 'steps': 16760, 'loss/train': 1.6873570680618286} 02/24/2022 23:16:48 - INFO - codeparrot_training - Step 16761: {'lr': 0.0003921270019551335, 'samples': 8582144, 'steps': 16761, 'loss/train': 2.5333385467529297} 02/24/2022 23:16:54 - INFO - codeparrot_training - Step 16762: {'lr': 0.00039211354061787407, 'samples': 8582656, 'steps': 16762, 'loss/train': 2.4803080558776855} 02/24/2022 23:16:57 - INFO - codeparrot_training - Step 16763: {'lr': 0.0003921000786718454, 'samples': 8583168, 'steps': 16763, 'loss/train': 2.0843112468719482} 02/24/2022 23:17:03 - INFO - codeparrot_training - Step 16764: {'lr': 0.0003920866161171051, 'samples': 8583680, 'steps': 16764, 'loss/train': 2.050842761993408} 02/24/2022 23:17:06 - INFO - codeparrot_training - Step 16765: {'lr': 0.0003920731529537108, 'samples': 8584192, 'steps': 16765, 'loss/train': 2.262209892272949} 02/24/2022 23:17:12 - INFO - codeparrot_training - Step 16766: {'lr': 0.00039205968918172026, 'samples': 8584704, 'steps': 16766, 'loss/train': 1.8851791620254517} 02/24/2022 23:17:16 - INFO - codeparrot_training - Step 16767: {'lr': 0.00039204622480119107, 'samples': 8585216, 'steps': 16767, 'loss/train': 1.0261403322219849} 02/24/2022 23:17:22 - INFO - codeparrot_training - Step 16768: {'lr': 0.000392032759812181, 'samples': 8585728, 'steps': 16768, 'loss/train': 2.142001152038574} 02/24/2022 23:17:25 - INFO - codeparrot_training - Step 16769: {'lr': 0.0003920192942147477, 'samples': 8586240, 'steps': 16769, 'loss/train': 1.6848701238632202} 02/24/2022 23:17:28 - INFO - codeparrot_training - Step 16770: {'lr': 0.00039200582800894885, 'samples': 8586752, 'steps': 16770, 'loss/train': 1.4346401691436768} 02/24/2022 23:17:34 - INFO - codeparrot_training - Step 16771: {'lr': 0.00039199236119484207, 'samples': 8587264, 'steps': 16771, 'loss/train': 2.4493155479431152} 02/24/2022 23:17:40 - INFO - codeparrot_training - Step 16772: {'lr': 0.0003919788937724852, 'samples': 8587776, 'steps': 16772, 'loss/train': 2.5252976417541504} 02/24/2022 23:17:43 - INFO - codeparrot_training - Step 16773: {'lr': 0.0003919654257419357, 'samples': 8588288, 'steps': 16773, 'loss/train': 2.1317074298858643} 02/24/2022 23:17:49 - INFO - codeparrot_training - Step 16774: {'lr': 0.0003919519571032515, 'samples': 8588800, 'steps': 16774, 'loss/train': 2.12827205657959} 02/24/2022 23:17:52 - INFO - codeparrot_training - Step 16775: {'lr': 0.00039193848785649016, 'samples': 8589312, 'steps': 16775, 'loss/train': 1.1113438606262207} 02/24/2022 23:17:56 - INFO - codeparrot_training - Step 16776: {'lr': 0.0003919250180017094, 'samples': 8589824, 'steps': 16776, 'loss/train': 2.9014785289764404} 02/24/2022 23:18:02 - INFO - codeparrot_training - Step 16777: {'lr': 0.00039191154753896696, 'samples': 8590336, 'steps': 16777, 'loss/train': 1.968258261680603} 02/24/2022 23:18:07 - INFO - codeparrot_training - Step 16778: {'lr': 0.00039189807646832045, 'samples': 8590848, 'steps': 16778, 'loss/train': 2.1374475955963135} 02/24/2022 23:18:11 - INFO - codeparrot_training - Step 16779: {'lr': 0.0003918846047898277, 'samples': 8591360, 'steps': 16779, 'loss/train': 3.5726158618927} 02/24/2022 23:18:14 - INFO - codeparrot_training - Step 16780: {'lr': 0.00039187113250354635, 'samples': 8591872, 'steps': 16780, 'loss/train': 1.8912001848220825} 02/24/2022 23:18:20 - INFO - codeparrot_training - Step 16781: {'lr': 0.00039185765960953405, 'samples': 8592384, 'steps': 16781, 'loss/train': 3.314514398574829} 02/24/2022 23:18:24 - INFO - codeparrot_training - Step 16782: {'lr': 0.0003918441861078486, 'samples': 8592896, 'steps': 16782, 'loss/train': 1.6754136085510254} 02/24/2022 23:18:29 - INFO - codeparrot_training - Step 16783: {'lr': 0.0003918307119985477, 'samples': 8593408, 'steps': 16783, 'loss/train': 2.063513994216919} 02/24/2022 23:18:33 - INFO - codeparrot_training - Step 16784: {'lr': 0.0003918172372816892, 'samples': 8593920, 'steps': 16784, 'loss/train': 1.3096811771392822} 02/24/2022 23:18:38 - INFO - codeparrot_training - Step 16785: {'lr': 0.0003918037619573305, 'samples': 8594432, 'steps': 16785, 'loss/train': 1.8276004791259766} 02/24/2022 23:18:42 - INFO - codeparrot_training - Step 16786: {'lr': 0.0003917902860255296, 'samples': 8594944, 'steps': 16786, 'loss/train': 1.6900595426559448} 02/24/2022 23:18:49 - INFO - codeparrot_training - Step 16787: {'lr': 0.0003917768094863441, 'samples': 8595456, 'steps': 16787, 'loss/train': 1.3743575811386108} 02/24/2022 23:18:52 - INFO - codeparrot_training - Step 16788: {'lr': 0.00039176333233983187, 'samples': 8595968, 'steps': 16788, 'loss/train': 2.1047770977020264} 02/24/2022 23:18:58 - INFO - codeparrot_training - Step 16789: {'lr': 0.0003917498545860504, 'samples': 8596480, 'steps': 16789, 'loss/train': 1.0771292448043823} 02/24/2022 23:19:02 - INFO - codeparrot_training - Step 16790: {'lr': 0.0003917363762250576, 'samples': 8596992, 'steps': 16790, 'loss/train': 3.275146961212158} 02/24/2022 23:19:08 - INFO - codeparrot_training - Step 16791: {'lr': 0.00039172289725691124, 'samples': 8597504, 'steps': 16791, 'loss/train': 1.6128149032592773} 02/24/2022 23:19:11 - INFO - codeparrot_training - Step 16792: {'lr': 0.000391709417681669, 'samples': 8598016, 'steps': 16792, 'loss/train': 1.7847570180892944} 02/24/2022 23:19:14 - INFO - codeparrot_training - Step 16793: {'lr': 0.0003916959374993885, 'samples': 8598528, 'steps': 16793, 'loss/train': 2.2029881477355957} 02/24/2022 23:19:20 - INFO - codeparrot_training - Step 16794: {'lr': 0.0003916824567101277, 'samples': 8599040, 'steps': 16794, 'loss/train': 1.8639984130859375} 02/24/2022 23:19:23 - INFO - codeparrot_training - Step 16795: {'lr': 0.0003916689753139442, 'samples': 8599552, 'steps': 16795, 'loss/train': 2.024332284927368} 02/24/2022 23:19:29 - INFO - codeparrot_training - Step 16796: {'lr': 0.0003916554933108958, 'samples': 8600064, 'steps': 16796, 'loss/train': 2.4026010036468506} 02/24/2022 23:19:32 - INFO - codeparrot_training - Step 16797: {'lr': 0.0003916420107010402, 'samples': 8600576, 'steps': 16797, 'loss/train': 2.0735392570495605} 02/24/2022 23:19:39 - INFO - codeparrot_training - Step 16798: {'lr': 0.0003916285274844353, 'samples': 8601088, 'steps': 16798, 'loss/train': 1.0994086265563965} 02/24/2022 23:19:42 - INFO - codeparrot_training - Step 16799: {'lr': 0.0003916150436611387, 'samples': 8601600, 'steps': 16799, 'loss/train': 2.4959099292755127} 02/24/2022 23:19:48 - INFO - codeparrot_training - Step 16800: {'lr': 0.0003916015592312082, 'samples': 8602112, 'steps': 16800, 'loss/train': 0.2565116286277771} 02/24/2022 23:19:51 - INFO - codeparrot_training - Step 16801: {'lr': 0.00039158807419470166, 'samples': 8602624, 'steps': 16801, 'loss/train': 2.1922967433929443} 02/24/2022 23:19:57 - INFO - codeparrot_training - Step 16802: {'lr': 0.0003915745885516767, 'samples': 8603136, 'steps': 16802, 'loss/train': 1.4452614784240723} 02/24/2022 23:20:00 - INFO - codeparrot_training - Step 16803: {'lr': 0.0003915611023021912, 'samples': 8603648, 'steps': 16803, 'loss/train': 1.3495919704437256} 02/24/2022 23:20:06 - INFO - codeparrot_training - Step 16804: {'lr': 0.00039154761544630287, 'samples': 8604160, 'steps': 16804, 'loss/train': 2.431001901626587} 02/24/2022 23:20:09 - INFO - codeparrot_training - Step 16805: {'lr': 0.0003915341279840695, 'samples': 8604672, 'steps': 16805, 'loss/train': 3.085236072540283} 02/24/2022 23:20:15 - INFO - codeparrot_training - Step 16806: {'lr': 0.00039152063991554885, 'samples': 8605184, 'steps': 16806, 'loss/train': 2.1057345867156982} 02/24/2022 23:20:18 - INFO - codeparrot_training - Step 16807: {'lr': 0.0003915071512407987, 'samples': 8605696, 'steps': 16807, 'loss/train': 1.8922282457351685} 02/24/2022 23:20:25 - INFO - codeparrot_training - Step 16808: {'lr': 0.0003914936619598769, 'samples': 8606208, 'steps': 16808, 'loss/train': 1.7595235109329224} 02/24/2022 23:20:28 - INFO - codeparrot_training - Step 16809: {'lr': 0.00039148017207284115, 'samples': 8606720, 'steps': 16809, 'loss/train': 1.1769860982894897} 02/24/2022 23:20:34 - INFO - codeparrot_training - Step 16810: {'lr': 0.0003914666815797493, 'samples': 8607232, 'steps': 16810, 'loss/train': 0.7896278500556946} 02/24/2022 23:20:37 - INFO - codeparrot_training - Step 16811: {'lr': 0.00039145319048065907, 'samples': 8607744, 'steps': 16811, 'loss/train': 1.8864352703094482} 02/24/2022 23:20:43 - INFO - codeparrot_training - Step 16812: {'lr': 0.00039143969877562833, 'samples': 8608256, 'steps': 16812, 'loss/train': 0.4437282383441925} 02/24/2022 23:20:46 - INFO - codeparrot_training - Step 16813: {'lr': 0.00039142620646471485, 'samples': 8608768, 'steps': 16813, 'loss/train': 1.2218278646469116} 02/24/2022 23:20:52 - INFO - codeparrot_training - Step 16814: {'lr': 0.00039141271354797635, 'samples': 8609280, 'steps': 16814, 'loss/train': 2.371201276779175} 02/24/2022 23:20:55 - INFO - codeparrot_training - Step 16815: {'lr': 0.0003913992200254707, 'samples': 8609792, 'steps': 16815, 'loss/train': 2.3276758193969727} 02/24/2022 23:21:01 - INFO - codeparrot_training - Step 16816: {'lr': 0.0003913857258972557, 'samples': 8610304, 'steps': 16816, 'loss/train': 1.5418903827667236} 02/24/2022 23:21:04 - INFO - codeparrot_training - Step 16817: {'lr': 0.0003913722311633892, 'samples': 8610816, 'steps': 16817, 'loss/train': 2.5481350421905518} 02/24/2022 23:21:10 - INFO - codeparrot_training - Step 16818: {'lr': 0.0003913587358239288, 'samples': 8611328, 'steps': 16818, 'loss/train': 1.9053610563278198} 02/24/2022 23:21:13 - INFO - codeparrot_training - Step 16819: {'lr': 0.0003913452398789326, 'samples': 8611840, 'steps': 16819, 'loss/train': 2.0319366455078125} 02/24/2022 23:21:19 - INFO - codeparrot_training - Step 16820: {'lr': 0.0003913317433284582, 'samples': 8612352, 'steps': 16820, 'loss/train': 1.5430309772491455} 02/24/2022 23:21:22 - INFO - codeparrot_training - Step 16821: {'lr': 0.00039131824617256354, 'samples': 8612864, 'steps': 16821, 'loss/train': 2.3261728286743164} 02/24/2022 23:21:28 - INFO - codeparrot_training - Step 16822: {'lr': 0.0003913047484113064, 'samples': 8613376, 'steps': 16822, 'loss/train': 2.5518836975097656} 02/24/2022 23:21:31 - INFO - codeparrot_training - Step 16823: {'lr': 0.0003912912500447445, 'samples': 8613888, 'steps': 16823, 'loss/train': 1.9733940362930298} 02/24/2022 23:21:37 - INFO - codeparrot_training - Step 16824: {'lr': 0.0003912777510729358, 'samples': 8614400, 'steps': 16824, 'loss/train': 1.2581181526184082} 02/24/2022 23:21:40 - INFO - codeparrot_training - Step 16825: {'lr': 0.0003912642514959381, 'samples': 8614912, 'steps': 16825, 'loss/train': 1.8309385776519775} 02/24/2022 23:21:47 - INFO - codeparrot_training - Step 16826: {'lr': 0.00039125075131380923, 'samples': 8615424, 'steps': 16826, 'loss/train': 1.498382329940796} 02/24/2022 23:21:50 - INFO - codeparrot_training - Step 16827: {'lr': 0.00039123725052660696, 'samples': 8615936, 'steps': 16827, 'loss/train': 1.355548620223999} 02/24/2022 23:21:56 - INFO - codeparrot_training - Step 16828: {'lr': 0.00039122374913438913, 'samples': 8616448, 'steps': 16828, 'loss/train': 1.8002629280090332} 02/24/2022 23:21:59 - INFO - codeparrot_training - Step 16829: {'lr': 0.00039121024713721365, 'samples': 8616960, 'steps': 16829, 'loss/train': 2.014267683029175} 02/24/2022 23:22:05 - INFO - codeparrot_training - Step 16830: {'lr': 0.0003911967445351382, 'samples': 8617472, 'steps': 16830, 'loss/train': 1.4848769903182983} 02/24/2022 23:22:08 - INFO - codeparrot_training - Step 16831: {'lr': 0.00039118324132822083, 'samples': 8617984, 'steps': 16831, 'loss/train': 1.7573505640029907} 02/24/2022 23:22:14 - INFO - codeparrot_training - Step 16832: {'lr': 0.0003911697375165193, 'samples': 8618496, 'steps': 16832, 'loss/train': 2.4055910110473633} 02/24/2022 23:22:17 - INFO - codeparrot_training - Step 16833: {'lr': 0.00039115623310009135, 'samples': 8619008, 'steps': 16833, 'loss/train': 1.5562704801559448} 02/24/2022 23:22:23 - INFO - codeparrot_training - Step 16834: {'lr': 0.00039114272807899496, 'samples': 8619520, 'steps': 16834, 'loss/train': 2.246941566467285} 02/24/2022 23:22:27 - INFO - codeparrot_training - Step 16835: {'lr': 0.000391129222453288, 'samples': 8620032, 'steps': 16835, 'loss/train': 0.9668616056442261} 02/24/2022 23:22:32 - INFO - codeparrot_training - Step 16836: {'lr': 0.00039111571622302824, 'samples': 8620544, 'steps': 16836, 'loss/train': 1.7109888792037964} 02/24/2022 23:22:36 - INFO - codeparrot_training - Step 16837: {'lr': 0.0003911022093882736, 'samples': 8621056, 'steps': 16837, 'loss/train': 0.9387457370758057} 02/24/2022 23:22:41 - INFO - codeparrot_training - Step 16838: {'lr': 0.00039108870194908175, 'samples': 8621568, 'steps': 16838, 'loss/train': 1.5231893062591553} 02/24/2022 23:22:45 - INFO - codeparrot_training - Step 16839: {'lr': 0.00039107519390551085, 'samples': 8622080, 'steps': 16839, 'loss/train': 1.5574750900268555} 02/24/2022 23:22:50 - INFO - codeparrot_training - Step 16840: {'lr': 0.00039106168525761855, 'samples': 8622592, 'steps': 16840, 'loss/train': 2.036545991897583} 02/24/2022 23:22:54 - INFO - codeparrot_training - Step 16841: {'lr': 0.00039104817600546277, 'samples': 8623104, 'steps': 16841, 'loss/train': 2.179760217666626} 02/24/2022 23:22:59 - INFO - codeparrot_training - Step 16842: {'lr': 0.00039103466614910144, 'samples': 8623616, 'steps': 16842, 'loss/train': 2.4726362228393555} 02/24/2022 23:23:03 - INFO - codeparrot_training - Step 16843: {'lr': 0.0003910211556885923, 'samples': 8624128, 'steps': 16843, 'loss/train': 2.223010778427124} 02/24/2022 23:23:10 - INFO - codeparrot_training - Step 16844: {'lr': 0.0003910076446239934, 'samples': 8624640, 'steps': 16844, 'loss/train': 2.6325414180755615} 02/24/2022 23:23:13 - INFO - codeparrot_training - Step 16845: {'lr': 0.00039099413295536246, 'samples': 8625152, 'steps': 16845, 'loss/train': 6.680877208709717} 02/24/2022 23:23:19 - INFO - codeparrot_training - Step 16846: {'lr': 0.0003909806206827575, 'samples': 8625664, 'steps': 16846, 'loss/train': 2.528313636779785} 02/24/2022 23:23:22 - INFO - codeparrot_training - Step 16847: {'lr': 0.00039096710780623625, 'samples': 8626176, 'steps': 16847, 'loss/train': 2.1600160598754883} 02/24/2022 23:23:28 - INFO - codeparrot_training - Step 16848: {'lr': 0.0003909535943258567, 'samples': 8626688, 'steps': 16848, 'loss/train': 2.4478447437286377} 02/24/2022 23:23:31 - INFO - codeparrot_training - Step 16849: {'lr': 0.0003909400802416767, 'samples': 8627200, 'steps': 16849, 'loss/train': 2.6089699268341064} 02/24/2022 23:23:37 - INFO - codeparrot_training - Step 16850: {'lr': 0.00039092656555375416, 'samples': 8627712, 'steps': 16850, 'loss/train': 1.8667513132095337} 02/24/2022 23:23:40 - INFO - codeparrot_training - Step 16851: {'lr': 0.00039091305026214704, 'samples': 8628224, 'steps': 16851, 'loss/train': 2.1120376586914062} 02/24/2022 23:23:46 - INFO - codeparrot_training - Step 16852: {'lr': 0.0003908995343669131, 'samples': 8628736, 'steps': 16852, 'loss/train': 2.670600652694702} 02/24/2022 23:23:49 - INFO - codeparrot_training - Step 16853: {'lr': 0.0003908860178681102, 'samples': 8629248, 'steps': 16853, 'loss/train': 3.1181223392486572} 02/24/2022 23:23:56 - INFO - codeparrot_training - Step 16854: {'lr': 0.0003908725007657964, 'samples': 8629760, 'steps': 16854, 'loss/train': 3.1714067459106445} 02/24/2022 23:23:59 - INFO - codeparrot_training - Step 16855: {'lr': 0.0003908589830600296, 'samples': 8630272, 'steps': 16855, 'loss/train': 0.7268884778022766} 02/24/2022 23:24:05 - INFO - codeparrot_training - Step 16856: {'lr': 0.0003908454647508676, 'samples': 8630784, 'steps': 16856, 'loss/train': 2.485808849334717} 02/24/2022 23:24:09 - INFO - codeparrot_training - Step 16857: {'lr': 0.00039083194583836836, 'samples': 8631296, 'steps': 16857, 'loss/train': 3.775114059448242} 02/24/2022 23:24:12 - INFO - codeparrot_training - Step 16858: {'lr': 0.0003908184263225898, 'samples': 8631808, 'steps': 16858, 'loss/train': 1.8956369161605835} 02/24/2022 23:24:19 - INFO - codeparrot_training - Step 16859: {'lr': 0.0003908049062035898, 'samples': 8632320, 'steps': 16859, 'loss/train': 1.420224905014038} 02/24/2022 23:24:22 - INFO - codeparrot_training - Step 16860: {'lr': 0.0003907913854814262, 'samples': 8632832, 'steps': 16860, 'loss/train': 2.164921998977661} 02/24/2022 23:24:26 - INFO - codeparrot_training - Step 16861: {'lr': 0.00039077786415615714, 'samples': 8633344, 'steps': 16861, 'loss/train': 1.5208487510681152} 02/24/2022 23:24:31 - INFO - codeparrot_training - Step 16862: {'lr': 0.0003907643422278404, 'samples': 8633856, 'steps': 16862, 'loss/train': 2.4336390495300293} 02/24/2022 23:24:35 - INFO - codeparrot_training - Step 16863: {'lr': 0.00039075081969653383, 'samples': 8634368, 'steps': 16863, 'loss/train': 1.6878808736801147} 02/24/2022 23:24:40 - INFO - codeparrot_training - Step 16864: {'lr': 0.0003907372965622955, 'samples': 8634880, 'steps': 16864, 'loss/train': 0.24565070867538452} 02/24/2022 23:24:44 - INFO - codeparrot_training - Step 16865: {'lr': 0.0003907237728251833, 'samples': 8635392, 'steps': 16865, 'loss/train': 2.3726556301116943} 02/24/2022 23:24:49 - INFO - codeparrot_training - Step 16866: {'lr': 0.0003907102484852551, 'samples': 8635904, 'steps': 16866, 'loss/train': 1.6078999042510986} 02/24/2022 23:24:53 - INFO - codeparrot_training - Step 16867: {'lr': 0.0003906967235425689, 'samples': 8636416, 'steps': 16867, 'loss/train': 1.6059931516647339} 02/24/2022 23:24:58 - INFO - codeparrot_training - Step 16868: {'lr': 0.0003906831979971826, 'samples': 8636928, 'steps': 16868, 'loss/train': 4.194704055786133} 02/24/2022 23:25:02 - INFO - codeparrot_training - Step 16869: {'lr': 0.0003906696718491541, 'samples': 8637440, 'steps': 16869, 'loss/train': 2.009641170501709} 02/24/2022 23:25:08 - INFO - codeparrot_training - Step 16870: {'lr': 0.0003906561450985415, 'samples': 8637952, 'steps': 16870, 'loss/train': 1.9054350852966309} 02/24/2022 23:25:11 - INFO - codeparrot_training - Step 16871: {'lr': 0.00039064261774540254, 'samples': 8638464, 'steps': 16871, 'loss/train': 3.1805734634399414} 02/24/2022 23:25:17 - INFO - codeparrot_training - Step 16872: {'lr': 0.0003906290897897953, 'samples': 8638976, 'steps': 16872, 'loss/train': 2.461371660232544} 02/24/2022 23:25:21 - INFO - codeparrot_training - Step 16873: {'lr': 0.00039061556123177777, 'samples': 8639488, 'steps': 16873, 'loss/train': 1.5450149774551392} 02/24/2022 23:25:26 - INFO - codeparrot_training - Step 16874: {'lr': 0.00039060203207140774, 'samples': 8640000, 'steps': 16874, 'loss/train': 2.7042887210845947} 02/24/2022 23:25:30 - INFO - codeparrot_training - Step 16875: {'lr': 0.0003905885023087433, 'samples': 8640512, 'steps': 16875, 'loss/train': 1.750491738319397} 02/24/2022 23:25:35 - INFO - codeparrot_training - Step 16876: {'lr': 0.0003905749719438423, 'samples': 8641024, 'steps': 16876, 'loss/train': 0.6731309294700623} 02/24/2022 23:25:39 - INFO - codeparrot_training - Step 16877: {'lr': 0.00039056144097676285, 'samples': 8641536, 'steps': 16877, 'loss/train': 1.1712673902511597} 02/24/2022 23:25:44 - INFO - codeparrot_training - Step 16878: {'lr': 0.0003905479094075627, 'samples': 8642048, 'steps': 16878, 'loss/train': 2.443943977355957} 02/24/2022 23:25:48 - INFO - codeparrot_training - Step 16879: {'lr': 0.00039053437723630003, 'samples': 8642560, 'steps': 16879, 'loss/train': 1.6826013326644897} 02/24/2022 23:25:53 - INFO - codeparrot_training - Step 16880: {'lr': 0.00039052084446303264, 'samples': 8643072, 'steps': 16880, 'loss/train': 2.573608160018921} 02/24/2022 23:25:57 - INFO - codeparrot_training - Step 16881: {'lr': 0.0003905073110878186, 'samples': 8643584, 'steps': 16881, 'loss/train': 0.9606328010559082} 02/24/2022 23:26:04 - INFO - codeparrot_training - Step 16882: {'lr': 0.00039049377711071595, 'samples': 8644096, 'steps': 16882, 'loss/train': 2.4206340312957764} 02/24/2022 23:26:07 - INFO - codeparrot_training - Step 16883: {'lr': 0.00039048024253178243, 'samples': 8644608, 'steps': 16883, 'loss/train': 1.2557203769683838} 02/24/2022 23:26:13 - INFO - codeparrot_training - Step 16884: {'lr': 0.00039046670735107627, 'samples': 8645120, 'steps': 16884, 'loss/train': 2.5004947185516357} 02/24/2022 23:26:16 - INFO - codeparrot_training - Step 16885: {'lr': 0.00039045317156865525, 'samples': 8645632, 'steps': 16885, 'loss/train': 1.8755720853805542} 02/24/2022 23:26:22 - INFO - codeparrot_training - Step 16886: {'lr': 0.0003904396351845775, 'samples': 8646144, 'steps': 16886, 'loss/train': 1.6499665975570679} 02/24/2022 23:26:25 - INFO - codeparrot_training - Step 16887: {'lr': 0.00039042609819890087, 'samples': 8646656, 'steps': 16887, 'loss/train': 1.7086681127548218} 02/24/2022 23:26:31 - INFO - codeparrot_training - Step 16888: {'lr': 0.0003904125606116835, 'samples': 8647168, 'steps': 16888, 'loss/train': 0.6751857399940491} 02/24/2022 23:26:34 - INFO - codeparrot_training - Step 16889: {'lr': 0.0003903990224229833, 'samples': 8647680, 'steps': 16889, 'loss/train': 1.9928969144821167} 02/24/2022 23:26:40 - INFO - codeparrot_training - Step 16890: {'lr': 0.00039038548363285825, 'samples': 8648192, 'steps': 16890, 'loss/train': 1.6685214042663574} 02/24/2022 23:26:43 - INFO - codeparrot_training - Step 16891: {'lr': 0.00039037194424136634, 'samples': 8648704, 'steps': 16891, 'loss/train': 2.2034332752227783} 02/24/2022 23:26:50 - INFO - codeparrot_training - Step 16892: {'lr': 0.0003903584042485656, 'samples': 8649216, 'steps': 16892, 'loss/train': 1.961581826210022} 02/24/2022 23:26:53 - INFO - codeparrot_training - Step 16893: {'lr': 0.00039034486365451405, 'samples': 8649728, 'steps': 16893, 'loss/train': 2.6328325271606445} 02/24/2022 23:26:59 - INFO - codeparrot_training - Step 16894: {'lr': 0.00039033132245926974, 'samples': 8650240, 'steps': 16894, 'loss/train': 2.49061918258667} 02/24/2022 23:27:02 - INFO - codeparrot_training - Step 16895: {'lr': 0.0003903177806628905, 'samples': 8650752, 'steps': 16895, 'loss/train': 2.2321810722351074} 02/24/2022 23:27:08 - INFO - codeparrot_training - Step 16896: {'lr': 0.00039030423826543446, 'samples': 8651264, 'steps': 16896, 'loss/train': 2.102652072906494} 02/24/2022 23:27:11 - INFO - codeparrot_training - Step 16897: {'lr': 0.0003902906952669596, 'samples': 8651776, 'steps': 16897, 'loss/train': 0.927318274974823} 02/24/2022 23:27:17 - INFO - codeparrot_training - Step 16898: {'lr': 0.000390277151667524, 'samples': 8652288, 'steps': 16898, 'loss/train': 2.801112174987793} 02/24/2022 23:27:20 - INFO - codeparrot_training - Step 16899: {'lr': 0.0003902636074671856, 'samples': 8652800, 'steps': 16899, 'loss/train': 1.952659010887146} 02/24/2022 23:27:26 - INFO - codeparrot_training - Step 16900: {'lr': 0.0003902500626660025, 'samples': 8653312, 'steps': 16900, 'loss/train': 0.9871919751167297} 02/24/2022 23:27:29 - INFO - codeparrot_training - Step 16901: {'lr': 0.00039023651726403263, 'samples': 8653824, 'steps': 16901, 'loss/train': 2.5099470615386963} 02/24/2022 23:27:35 - INFO - codeparrot_training - Step 16902: {'lr': 0.00039022297126133397, 'samples': 8654336, 'steps': 16902, 'loss/train': 2.026329517364502} 02/24/2022 23:27:39 - INFO - codeparrot_training - Step 16903: {'lr': 0.0003902094246579647, 'samples': 8654848, 'steps': 16903, 'loss/train': 1.2148100137710571} 02/24/2022 23:27:44 - INFO - codeparrot_training - Step 16904: {'lr': 0.00039019587745398276, 'samples': 8655360, 'steps': 16904, 'loss/train': 2.444507122039795} 02/24/2022 23:27:48 - INFO - codeparrot_training - Step 16905: {'lr': 0.00039018232964944623, 'samples': 8655872, 'steps': 16905, 'loss/train': 2.254626989364624} 02/24/2022 23:27:53 - INFO - codeparrot_training - Step 16906: {'lr': 0.0003901687812444131, 'samples': 8656384, 'steps': 16906, 'loss/train': 8.72937297821045} 02/24/2022 23:27:57 - INFO - codeparrot_training - Step 16907: {'lr': 0.0003901552322389414, 'samples': 8656896, 'steps': 16907, 'loss/train': 2.806398868560791} 02/24/2022 23:28:02 - INFO - codeparrot_training - Step 16908: {'lr': 0.00039014168263308926, 'samples': 8657408, 'steps': 16908, 'loss/train': 0.8207800984382629} 02/24/2022 23:28:06 - INFO - codeparrot_training - Step 16909: {'lr': 0.00039012813242691454, 'samples': 8657920, 'steps': 16909, 'loss/train': 1.2087421417236328} 02/24/2022 23:28:11 - INFO - codeparrot_training - Step 16910: {'lr': 0.00039011458162047547, 'samples': 8658432, 'steps': 16910, 'loss/train': 2.4208099842071533} 02/24/2022 23:28:15 - INFO - codeparrot_training - Step 16911: {'lr': 0.00039010103021383, 'samples': 8658944, 'steps': 16911, 'loss/train': 1.7641187906265259} 02/24/2022 23:28:20 - INFO - codeparrot_training - Step 16912: {'lr': 0.00039008747820703615, 'samples': 8659456, 'steps': 16912, 'loss/train': 1.685917854309082} 02/24/2022 23:28:24 - INFO - codeparrot_training - Step 16913: {'lr': 0.0003900739256001521, 'samples': 8659968, 'steps': 16913, 'loss/train': 2.2563421726226807} 02/24/2022 23:28:29 - INFO - codeparrot_training - Step 16914: {'lr': 0.00039006037239323584, 'samples': 8660480, 'steps': 16914, 'loss/train': 1.4826017618179321} 02/24/2022 23:28:33 - INFO - codeparrot_training - Step 16915: {'lr': 0.00039004681858634537, 'samples': 8660992, 'steps': 16915, 'loss/train': 0.29030516743659973} 02/24/2022 23:28:38 - INFO - codeparrot_training - Step 16916: {'lr': 0.0003900332641795388, 'samples': 8661504, 'steps': 16916, 'loss/train': 2.1742749214172363} 02/24/2022 23:28:42 - INFO - codeparrot_training - Step 16917: {'lr': 0.0003900197091728742, 'samples': 8662016, 'steps': 16917, 'loss/train': 2.3685264587402344} 02/24/2022 23:28:48 - INFO - codeparrot_training - Step 16918: {'lr': 0.0003900061535664097, 'samples': 8662528, 'steps': 16918, 'loss/train': 1.8964872360229492} 02/24/2022 23:28:51 - INFO - codeparrot_training - Step 16919: {'lr': 0.0003899925973602032, 'samples': 8663040, 'steps': 16919, 'loss/train': 1.3185784816741943} 02/24/2022 23:28:57 - INFO - codeparrot_training - Step 16920: {'lr': 0.0003899790405543129, 'samples': 8663552, 'steps': 16920, 'loss/train': 1.9329650402069092} 02/24/2022 23:29:00 - INFO - codeparrot_training - Step 16921: {'lr': 0.0003899654831487969, 'samples': 8664064, 'steps': 16921, 'loss/train': 2.5319745540618896} 02/24/2022 23:29:06 - INFO - codeparrot_training - Step 16922: {'lr': 0.0003899519251437131, 'samples': 8664576, 'steps': 16922, 'loss/train': 2.0593087673187256} 02/24/2022 23:29:09 - INFO - codeparrot_training - Step 16923: {'lr': 0.00038993836653911974, 'samples': 8665088, 'steps': 16923, 'loss/train': 1.5930911302566528} 02/24/2022 23:29:15 - INFO - codeparrot_training - Step 16924: {'lr': 0.00038992480733507487, 'samples': 8665600, 'steps': 16924, 'loss/train': 2.0232458114624023} 02/24/2022 23:29:18 - INFO - codeparrot_training - Step 16925: {'lr': 0.0003899112475316365, 'samples': 8666112, 'steps': 16925, 'loss/train': 2.715932846069336} 02/24/2022 23:29:24 - INFO - codeparrot_training - Step 16926: {'lr': 0.00038989768712886287, 'samples': 8666624, 'steps': 16926, 'loss/train': 1.8036901950836182} 02/24/2022 23:29:27 - INFO - codeparrot_training - Step 16927: {'lr': 0.0003898841261268119, 'samples': 8667136, 'steps': 16927, 'loss/train': 2.5591657161712646} 02/24/2022 23:29:33 - INFO - codeparrot_training - Step 16928: {'lr': 0.00038987056452554177, 'samples': 8667648, 'steps': 16928, 'loss/train': 1.5738908052444458} 02/24/2022 23:29:37 - INFO - codeparrot_training - Step 16929: {'lr': 0.00038985700232511055, 'samples': 8668160, 'steps': 16929, 'loss/train': 1.8292303085327148} 02/24/2022 23:29:43 - INFO - codeparrot_training - Step 16930: {'lr': 0.0003898434395255763, 'samples': 8668672, 'steps': 16930, 'loss/train': 2.2562856674194336} 02/24/2022 23:29:46 - INFO - codeparrot_training - Step 16931: {'lr': 0.0003898298761269973, 'samples': 8669184, 'steps': 16931, 'loss/train': 1.6968156099319458} 02/24/2022 23:29:52 - INFO - codeparrot_training - Step 16932: {'lr': 0.0003898163121294314, 'samples': 8669696, 'steps': 16932, 'loss/train': 2.213306427001953} 02/24/2022 23:29:55 - INFO - codeparrot_training - Step 16933: {'lr': 0.0003898027475329368, 'samples': 8670208, 'steps': 16933, 'loss/train': 1.7215467691421509} 02/24/2022 23:30:01 - INFO - codeparrot_training - Step 16934: {'lr': 0.00038978918233757167, 'samples': 8670720, 'steps': 16934, 'loss/train': 1.9487850666046143} 02/24/2022 23:30:04 - INFO - codeparrot_training - Step 16935: {'lr': 0.000389775616543394, 'samples': 8671232, 'steps': 16935, 'loss/train': 2.101301431655884} 02/24/2022 23:30:10 - INFO - codeparrot_training - Step 16936: {'lr': 0.00038976205015046206, 'samples': 8671744, 'steps': 16936, 'loss/train': 1.3145732879638672} 02/24/2022 23:30:13 - INFO - codeparrot_training - Step 16937: {'lr': 0.00038974848315883383, 'samples': 8672256, 'steps': 16937, 'loss/train': 2.593108892440796} 02/24/2022 23:30:19 - INFO - codeparrot_training - Step 16938: {'lr': 0.00038973491556856755, 'samples': 8672768, 'steps': 16938, 'loss/train': 2.0943918228149414} 02/24/2022 23:30:23 - INFO - codeparrot_training - Step 16939: {'lr': 0.0003897213473797212, 'samples': 8673280, 'steps': 16939, 'loss/train': 1.4675676822662354} 02/24/2022 23:30:28 - INFO - codeparrot_training - Step 16940: {'lr': 0.0003897077785923529, 'samples': 8673792, 'steps': 16940, 'loss/train': 2.2290542125701904} 02/24/2022 23:30:32 - INFO - codeparrot_training - Step 16941: {'lr': 0.0003896942092065209, 'samples': 8674304, 'steps': 16941, 'loss/train': 1.8714227676391602} 02/24/2022 23:30:37 - INFO - codeparrot_training - Step 16942: {'lr': 0.0003896806392222833, 'samples': 8674816, 'steps': 16942, 'loss/train': 1.1439050436019897} 02/24/2022 23:30:41 - INFO - codeparrot_training - Step 16943: {'lr': 0.00038966706863969815, 'samples': 8675328, 'steps': 16943, 'loss/train': 1.649509310722351} 02/24/2022 23:30:46 - INFO - codeparrot_training - Step 16944: {'lr': 0.00038965349745882365, 'samples': 8675840, 'steps': 16944, 'loss/train': 2.2249338626861572} 02/24/2022 23:30:50 - INFO - codeparrot_training - Step 16945: {'lr': 0.00038963992567971794, 'samples': 8676352, 'steps': 16945, 'loss/train': 1.5210684537887573} 02/24/2022 23:30:55 - INFO - codeparrot_training - Step 16946: {'lr': 0.0003896263533024391, 'samples': 8676864, 'steps': 16946, 'loss/train': 1.8658944368362427} 02/24/2022 23:30:59 - INFO - codeparrot_training - Step 16947: {'lr': 0.0003896127803270453, 'samples': 8677376, 'steps': 16947, 'loss/train': 1.8243488073349} 02/24/2022 23:31:04 - INFO - codeparrot_training - Step 16948: {'lr': 0.0003895992067535946, 'samples': 8677888, 'steps': 16948, 'loss/train': 2.0701653957366943} 02/24/2022 23:31:08 - INFO - codeparrot_training - Step 16949: {'lr': 0.0003895856325821454, 'samples': 8678400, 'steps': 16949, 'loss/train': 1.8434815406799316} 02/24/2022 23:31:14 - INFO - codeparrot_training - Step 16950: {'lr': 0.00038957205781275554, 'samples': 8678912, 'steps': 16950, 'loss/train': 2.7598280906677246} 02/24/2022 23:31:17 - INFO - codeparrot_training - Step 16951: {'lr': 0.00038955848244548333, 'samples': 8679424, 'steps': 16951, 'loss/train': 0.7306210398674011} 02/24/2022 23:31:23 - INFO - codeparrot_training - Step 16952: {'lr': 0.00038954490648038687, 'samples': 8679936, 'steps': 16952, 'loss/train': 2.261073112487793} 02/24/2022 23:31:26 - INFO - codeparrot_training - Step 16953: {'lr': 0.0003895313299175244, 'samples': 8680448, 'steps': 16953, 'loss/train': 2.1445531845092773} 02/24/2022 23:31:32 - INFO - codeparrot_training - Step 16954: {'lr': 0.000389517752756954, 'samples': 8680960, 'steps': 16954, 'loss/train': 1.6433805227279663} 02/24/2022 23:31:36 - INFO - codeparrot_training - Step 16955: {'lr': 0.0003895041749987338, 'samples': 8681472, 'steps': 16955, 'loss/train': 2.915520668029785} 02/24/2022 23:31:41 - INFO - codeparrot_training - Step 16956: {'lr': 0.00038949059664292207, 'samples': 8681984, 'steps': 16956, 'loss/train': 1.8010644912719727} 02/24/2022 23:31:45 - INFO - codeparrot_training - Step 16957: {'lr': 0.0003894770176895769, 'samples': 8682496, 'steps': 16957, 'loss/train': 2.249030590057373} 02/24/2022 23:31:50 - INFO - codeparrot_training - Step 16958: {'lr': 0.0003894634381387565, 'samples': 8683008, 'steps': 16958, 'loss/train': 1.8238641023635864} 02/24/2022 23:31:53 - INFO - codeparrot_training - Step 16959: {'lr': 0.00038944985799051896, 'samples': 8683520, 'steps': 16959, 'loss/train': 1.926099419593811} 02/24/2022 23:31:59 - INFO - codeparrot_training - Step 16960: {'lr': 0.0003894362772449226, 'samples': 8684032, 'steps': 16960, 'loss/train': 1.838199257850647} 02/24/2022 23:32:03 - INFO - codeparrot_training - Step 16961: {'lr': 0.0003894226959020254, 'samples': 8684544, 'steps': 16961, 'loss/train': 1.834027647972107} 02/24/2022 23:32:08 - INFO - codeparrot_training - Step 16962: {'lr': 0.00038940911396188573, 'samples': 8685056, 'steps': 16962, 'loss/train': 2.2430877685546875} 02/24/2022 23:32:15 - INFO - codeparrot_training - Step 16963: {'lr': 0.0003893955314245616, 'samples': 8685568, 'steps': 16963, 'loss/train': 1.9298113584518433} 02/24/2022 23:32:18 - INFO - codeparrot_training - Step 16964: {'lr': 0.0003893819482901113, 'samples': 8686080, 'steps': 16964, 'loss/train': 1.9258403778076172} 02/24/2022 23:32:24 - INFO - codeparrot_training - Step 16965: {'lr': 0.000389368364558593, 'samples': 8686592, 'steps': 16965, 'loss/train': 1.7546473741531372} 02/24/2022 23:32:27 - INFO - codeparrot_training - Step 16966: {'lr': 0.00038935478023006487, 'samples': 8687104, 'steps': 16966, 'loss/train': 2.231564521789551} 02/24/2022 23:32:33 - INFO - codeparrot_training - Step 16967: {'lr': 0.0003893411953045852, 'samples': 8687616, 'steps': 16967, 'loss/train': 1.3445745706558228} 02/24/2022 23:32:36 - INFO - codeparrot_training - Step 16968: {'lr': 0.000389327609782212, 'samples': 8688128, 'steps': 16968, 'loss/train': 1.9698094129562378} 02/24/2022 23:32:42 - INFO - codeparrot_training - Step 16969: {'lr': 0.0003893140236630036, 'samples': 8688640, 'steps': 16969, 'loss/train': 2.612501621246338} 02/24/2022 23:32:45 - INFO - codeparrot_training - Step 16970: {'lr': 0.0003893004369470181, 'samples': 8689152, 'steps': 16970, 'loss/train': 2.2755954265594482} 02/24/2022 23:32:51 - INFO - codeparrot_training - Step 16971: {'lr': 0.00038928684963431383, 'samples': 8689664, 'steps': 16971, 'loss/train': 2.605947256088257} 02/24/2022 23:32:54 - INFO - codeparrot_training - Step 16972: {'lr': 0.00038927326172494894, 'samples': 8690176, 'steps': 16972, 'loss/train': 2.1987826824188232} 02/24/2022 23:33:01 - INFO - codeparrot_training - Step 16973: {'lr': 0.0003892596732189816, 'samples': 8690688, 'steps': 16973, 'loss/train': 1.293381929397583} 02/24/2022 23:33:04 - INFO - codeparrot_training - Step 16974: {'lr': 0.00038924608411647, 'samples': 8691200, 'steps': 16974, 'loss/train': 1.3081204891204834} 02/24/2022 23:33:10 - INFO - codeparrot_training - Step 16975: {'lr': 0.00038923249441747245, 'samples': 8691712, 'steps': 16975, 'loss/train': 2.2672359943389893} 02/24/2022 23:33:13 - INFO - codeparrot_training - Step 16976: {'lr': 0.000389218904122047, 'samples': 8692224, 'steps': 16976, 'loss/train': 1.9623104333877563} 02/24/2022 23:33:19 - INFO - codeparrot_training - Step 16977: {'lr': 0.00038920531323025206, 'samples': 8692736, 'steps': 16977, 'loss/train': 3.1134965419769287} 02/24/2022 23:33:22 - INFO - codeparrot_training - Step 16978: {'lr': 0.0003891917217421458, 'samples': 8693248, 'steps': 16978, 'loss/train': 1.8949763774871826} 02/24/2022 23:33:28 - INFO - codeparrot_training - Step 16979: {'lr': 0.00038917812965778625, 'samples': 8693760, 'steps': 16979, 'loss/train': 1.8277864456176758} 02/24/2022 23:33:31 - INFO - codeparrot_training - Step 16980: {'lr': 0.00038916453697723194, 'samples': 8694272, 'steps': 16980, 'loss/train': 1.6719551086425781} 02/24/2022 23:33:37 - INFO - codeparrot_training - Step 16981: {'lr': 0.00038915094370054083, 'samples': 8694784, 'steps': 16981, 'loss/train': 2.123537302017212} 02/24/2022 23:33:40 - INFO - codeparrot_training - Step 16982: {'lr': 0.00038913734982777136, 'samples': 8695296, 'steps': 16982, 'loss/train': 1.8104923963546753} 02/24/2022 23:33:46 - INFO - codeparrot_training - Step 16983: {'lr': 0.0003891237553589816, 'samples': 8695808, 'steps': 16983, 'loss/train': 1.3323637247085571} 02/24/2022 23:33:49 - INFO - codeparrot_training - Step 16984: {'lr': 0.00038911016029422984, 'samples': 8696320, 'steps': 16984, 'loss/train': 2.28753924369812} 02/24/2022 23:33:55 - INFO - codeparrot_training - Step 16985: {'lr': 0.0003890965646335744, 'samples': 8696832, 'steps': 16985, 'loss/train': 1.9426664113998413} 02/24/2022 23:33:59 - INFO - codeparrot_training - Step 16986: {'lr': 0.0003890829683770734, 'samples': 8697344, 'steps': 16986, 'loss/train': 2.0136423110961914} 02/24/2022 23:34:04 - INFO - codeparrot_training - Step 16987: {'lr': 0.0003890693715247851, 'samples': 8697856, 'steps': 16987, 'loss/train': 2.2473838329315186} 02/24/2022 23:34:08 - INFO - codeparrot_training - Step 16988: {'lr': 0.0003890557740767678, 'samples': 8698368, 'steps': 16988, 'loss/train': 5.835962772369385} 02/24/2022 23:34:13 - INFO - codeparrot_training - Step 16989: {'lr': 0.0003890421760330798, 'samples': 8698880, 'steps': 16989, 'loss/train': 1.5661112070083618} 02/24/2022 23:34:17 - INFO - codeparrot_training - Step 16990: {'lr': 0.0003890285773937792, 'samples': 8699392, 'steps': 16990, 'loss/train': 2.22489595413208} 02/24/2022 23:34:22 - INFO - codeparrot_training - Step 16991: {'lr': 0.0003890149781589243, 'samples': 8699904, 'steps': 16991, 'loss/train': 1.6395831108093262} 02/24/2022 23:34:26 - INFO - codeparrot_training - Step 16992: {'lr': 0.0003890013783285733, 'samples': 8700416, 'steps': 16992, 'loss/train': 2.580901861190796} 02/24/2022 23:34:31 - INFO - codeparrot_training - Step 16993: {'lr': 0.00038898777790278465, 'samples': 8700928, 'steps': 16993, 'loss/train': 2.2957491874694824} 02/24/2022 23:34:35 - INFO - codeparrot_training - Step 16994: {'lr': 0.00038897417688161644, 'samples': 8701440, 'steps': 16994, 'loss/train': 2.2015116214752197} 02/24/2022 23:34:41 - INFO - codeparrot_training - Step 16995: {'lr': 0.0003889605752651271, 'samples': 8701952, 'steps': 16995, 'loss/train': 1.9655791521072388} 02/24/2022 23:34:44 - INFO - codeparrot_training - Step 16996: {'lr': 0.0003889469730533746, 'samples': 8702464, 'steps': 16996, 'loss/train': 1.8086227178573608} 02/24/2022 23:34:50 - INFO - codeparrot_training - Step 16997: {'lr': 0.0003889333702464175, 'samples': 8702976, 'steps': 16997, 'loss/train': 1.8407796621322632} 02/24/2022 23:34:53 - INFO - codeparrot_training - Step 16998: {'lr': 0.00038891976684431395, 'samples': 8703488, 'steps': 16998, 'loss/train': 2.391925096511841} 02/24/2022 23:34:59 - INFO - codeparrot_training - Step 16999: {'lr': 0.0003889061628471222, 'samples': 8704000, 'steps': 16999, 'loss/train': 3.4161200523376465} 02/24/2022 23:34:59 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/24/2022 23:35:16 - WARNING - huggingface_hub.repository - Several commits (17) will be pushed upstream. 02/24/2022 23:35:16 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/24/2022 23:35:49 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 704f602..4fad2a5 floral-grass-11 -> floral-grass-11 02/24/2022 23:35:54 - INFO - codeparrot_training - Step 17000: {'lr': 0.00038889255825490053, 'samples': 8704512, 'steps': 17000, 'loss/train': 1.384848713874817} 02/24/2022 23:35:59 - INFO - codeparrot_training - Step 17001: {'lr': 0.0003888789530677073, 'samples': 8705024, 'steps': 17001, 'loss/train': 4.113182067871094} 02/24/2022 23:36:03 - INFO - codeparrot_training - Step 17002: {'lr': 0.00038886534728560073, 'samples': 8705536, 'steps': 17002, 'loss/train': 2.253518581390381} 02/24/2022 23:36:08 - INFO - codeparrot_training - Step 17003: {'lr': 0.0003888517409086391, 'samples': 8706048, 'steps': 17003, 'loss/train': 1.5202544927597046} 02/24/2022 23:36:12 - INFO - codeparrot_training - Step 17004: {'lr': 0.0003888381339368807, 'samples': 8706560, 'steps': 17004, 'loss/train': 1.6984606981277466} 02/24/2022 23:36:17 - INFO - codeparrot_training - Step 17005: {'lr': 0.00038882452637038377, 'samples': 8707072, 'steps': 17005, 'loss/train': 2.1926732063293457} 02/24/2022 23:36:21 - INFO - codeparrot_training - Step 17006: {'lr': 0.00038881091820920676, 'samples': 8707584, 'steps': 17006, 'loss/train': 1.8134039640426636} 02/24/2022 23:36:26 - INFO - codeparrot_training - Step 17007: {'lr': 0.00038879730945340775, 'samples': 8708096, 'steps': 17007, 'loss/train': 2.5462334156036377} 02/24/2022 23:36:30 - INFO - codeparrot_training - Step 17008: {'lr': 0.0003887837001030452, 'samples': 8708608, 'steps': 17008, 'loss/train': 2.75063157081604} 02/24/2022 23:36:35 - INFO - codeparrot_training - Step 17009: {'lr': 0.00038877009015817734, 'samples': 8709120, 'steps': 17009, 'loss/train': 1.5253517627716064} 02/24/2022 23:36:39 - INFO - codeparrot_training - Step 17010: {'lr': 0.0003887564796188625, 'samples': 8709632, 'steps': 17010, 'loss/train': 2.22629451751709} 02/24/2022 23:36:45 - INFO - codeparrot_training - Step 17011: {'lr': 0.0003887428684851589, 'samples': 8710144, 'steps': 17011, 'loss/train': 2.340531826019287} 02/24/2022 23:36:49 - INFO - codeparrot_training - Step 17012: {'lr': 0.00038872925675712493, 'samples': 8710656, 'steps': 17012, 'loss/train': 1.8950774669647217} 02/24/2022 23:36:54 - INFO - codeparrot_training - Step 17013: {'lr': 0.00038871564443481886, 'samples': 8711168, 'steps': 17013, 'loss/train': 2.8091981410980225} 02/24/2022 23:36:58 - INFO - codeparrot_training - Step 17014: {'lr': 0.0003887020315182991, 'samples': 8711680, 'steps': 17014, 'loss/train': 1.9716668128967285} 02/24/2022 23:37:03 - INFO - codeparrot_training - Step 17015: {'lr': 0.0003886884180076238, 'samples': 8712192, 'steps': 17015, 'loss/train': 1.541662573814392} 02/24/2022 23:37:07 - INFO - codeparrot_training - Step 17016: {'lr': 0.0003886748039028514, 'samples': 8712704, 'steps': 17016, 'loss/train': 2.596200942993164} 02/24/2022 23:37:12 - INFO - codeparrot_training - Step 17017: {'lr': 0.00038866118920404013, 'samples': 8713216, 'steps': 17017, 'loss/train': 2.6591274738311768} 02/24/2022 23:37:16 - INFO - codeparrot_training - Step 17018: {'lr': 0.0003886475739112484, 'samples': 8713728, 'steps': 17018, 'loss/train': 1.9131158590316772} 02/24/2022 23:37:21 - INFO - codeparrot_training - Step 17019: {'lr': 0.0003886339580245344, 'samples': 8714240, 'steps': 17019, 'loss/train': 1.7564442157745361} 02/24/2022 23:37:25 - INFO - codeparrot_training - Step 17020: {'lr': 0.00038862034154395664, 'samples': 8714752, 'steps': 17020, 'loss/train': 1.9179941415786743} 02/24/2022 23:37:31 - INFO - codeparrot_training - Step 17021: {'lr': 0.00038860672446957336, 'samples': 8715264, 'steps': 17021, 'loss/train': 2.935488700866699} 02/24/2022 23:37:35 - INFO - codeparrot_training - Step 17022: {'lr': 0.00038859310680144276, 'samples': 8715776, 'steps': 17022, 'loss/train': 1.9156081676483154} 02/24/2022 23:37:40 - INFO - codeparrot_training - Step 17023: {'lr': 0.0003885794885396234, 'samples': 8716288, 'steps': 17023, 'loss/train': 2.3114750385284424} 02/24/2022 23:37:44 - INFO - codeparrot_training - Step 17024: {'lr': 0.00038856586968417353, 'samples': 8716800, 'steps': 17024, 'loss/train': 2.5909228324890137} 02/24/2022 23:37:49 - INFO - codeparrot_training - Step 17025: {'lr': 0.0003885522502351514, 'samples': 8717312, 'steps': 17025, 'loss/train': 5.457999229431152} 02/24/2022 23:37:53 - INFO - codeparrot_training - Step 17026: {'lr': 0.0003885386301926155, 'samples': 8717824, 'steps': 17026, 'loss/train': 2.549170970916748} 02/24/2022 23:37:58 - INFO - codeparrot_training - Step 17027: {'lr': 0.00038852500955662407, 'samples': 8718336, 'steps': 17027, 'loss/train': 1.4599891901016235} 02/24/2022 23:38:02 - INFO - codeparrot_training - Step 17028: {'lr': 0.0003885113883272355, 'samples': 8718848, 'steps': 17028, 'loss/train': 2.0304415225982666} 02/24/2022 23:38:07 - INFO - codeparrot_training - Step 17029: {'lr': 0.0003884977665045081, 'samples': 8719360, 'steps': 17029, 'loss/train': 1.807279109954834} 02/24/2022 23:38:11 - INFO - codeparrot_training - Step 17030: {'lr': 0.0003884841440885003, 'samples': 8719872, 'steps': 17030, 'loss/train': 0.4575875699520111} 02/24/2022 23:38:17 - INFO - codeparrot_training - Step 17031: {'lr': 0.0003884705210792703, 'samples': 8720384, 'steps': 17031, 'loss/train': 1.2923752069473267} 02/24/2022 23:38:20 - INFO - codeparrot_training - Step 17032: {'lr': 0.00038845689747687664, 'samples': 8720896, 'steps': 17032, 'loss/train': 2.4116322994232178} 02/24/2022 23:38:26 - INFO - codeparrot_training - Step 17033: {'lr': 0.0003884432732813776, 'samples': 8721408, 'steps': 17033, 'loss/train': 1.3514584302902222} 02/24/2022 23:38:29 - INFO - codeparrot_training - Step 17034: {'lr': 0.00038842964849283146, 'samples': 8721920, 'steps': 17034, 'loss/train': 2.2514805793762207} 02/24/2022 23:38:35 - INFO - codeparrot_training - Step 17035: {'lr': 0.0003884160231112968, 'samples': 8722432, 'steps': 17035, 'loss/train': 1.9595551490783691} 02/24/2022 23:38:38 - INFO - codeparrot_training - Step 17036: {'lr': 0.00038840239713683165, 'samples': 8722944, 'steps': 17036, 'loss/train': 0.9615404009819031} 02/24/2022 23:38:44 - INFO - codeparrot_training - Step 17037: {'lr': 0.00038838877056949475, 'samples': 8723456, 'steps': 17037, 'loss/train': 1.247662901878357} 02/24/2022 23:38:47 - INFO - codeparrot_training - Step 17038: {'lr': 0.00038837514340934424, 'samples': 8723968, 'steps': 17038, 'loss/train': 1.8019756078720093} 02/24/2022 23:38:53 - INFO - codeparrot_training - Step 17039: {'lr': 0.0003883615156564385, 'samples': 8724480, 'steps': 17039, 'loss/train': 0.6995099186897278} 02/24/2022 23:38:56 - INFO - codeparrot_training - Step 17040: {'lr': 0.000388347887310836, 'samples': 8724992, 'steps': 17040, 'loss/train': 2.377166986465454} 02/24/2022 23:39:04 - INFO - codeparrot_training - Step 17041: {'lr': 0.0003883342583725952, 'samples': 8725504, 'steps': 17041, 'loss/train': 2.044576406478882} 02/24/2022 23:39:07 - INFO - codeparrot_training - Step 17042: {'lr': 0.0003883206288417742, 'samples': 8726016, 'steps': 17042, 'loss/train': 1.3499064445495605} 02/24/2022 23:39:13 - INFO - codeparrot_training - Step 17043: {'lr': 0.0003883069987184316, 'samples': 8726528, 'steps': 17043, 'loss/train': 1.1370898485183716} 02/24/2022 23:39:16 - INFO - codeparrot_training - Step 17044: {'lr': 0.0003882933680026257, 'samples': 8727040, 'steps': 17044, 'loss/train': 1.7531712055206299} 02/24/2022 23:39:22 - INFO - codeparrot_training - Step 17045: {'lr': 0.000388279736694415, 'samples': 8727552, 'steps': 17045, 'loss/train': 1.5097105503082275} 02/24/2022 23:39:25 - INFO - codeparrot_training - Step 17046: {'lr': 0.00038826610479385774, 'samples': 8728064, 'steps': 17046, 'loss/train': 1.8655685186386108} 02/24/2022 23:39:31 - INFO - codeparrot_training - Step 17047: {'lr': 0.00038825247230101244, 'samples': 8728576, 'steps': 17047, 'loss/train': 1.5133527517318726} 02/24/2022 23:39:37 - INFO - codeparrot_training - Step 17048: {'lr': 0.0003882388392159375, 'samples': 8729088, 'steps': 17048, 'loss/train': 2.1465346813201904} 02/24/2022 23:39:40 - INFO - codeparrot_training - Step 17049: {'lr': 0.0003882252055386912, 'samples': 8729600, 'steps': 17049, 'loss/train': 2.474968433380127} 02/24/2022 23:39:43 - INFO - codeparrot_training - Step 17050: {'lr': 0.00038821157126933204, 'samples': 8730112, 'steps': 17050, 'loss/train': 2.2572433948516846} 02/24/2022 23:39:51 - INFO - codeparrot_training - Step 17051: {'lr': 0.00038819793640791834, 'samples': 8730624, 'steps': 17051, 'loss/train': 1.7960113286972046} 02/24/2022 23:39:54 - INFO - codeparrot_training - Step 17052: {'lr': 0.0003881843009545086, 'samples': 8731136, 'steps': 17052, 'loss/train': 1.1132287979125977} 02/24/2022 23:40:00 - INFO - codeparrot_training - Step 17053: {'lr': 0.0003881706649091612, 'samples': 8731648, 'steps': 17053, 'loss/train': 0.7072966694831848} 02/24/2022 23:40:06 - INFO - codeparrot_training - Step 17054: {'lr': 0.0003881570282719346, 'samples': 8732160, 'steps': 17054, 'loss/train': 2.306298017501831} 02/24/2022 23:40:09 - INFO - codeparrot_training - Step 17055: {'lr': 0.00038814339104288706, 'samples': 8732672, 'steps': 17055, 'loss/train': 2.1409566402435303} 02/24/2022 23:40:15 - INFO - codeparrot_training - Step 17056: {'lr': 0.00038812975322207713, 'samples': 8733184, 'steps': 17056, 'loss/train': 1.8545455932617188} 02/24/2022 23:40:18 - INFO - codeparrot_training - Step 17057: {'lr': 0.0003881161148095632, 'samples': 8733696, 'steps': 17057, 'loss/train': 1.762295126914978} 02/24/2022 23:40:24 - INFO - codeparrot_training - Step 17058: {'lr': 0.0003881024758054037, 'samples': 8734208, 'steps': 17058, 'loss/train': 2.726966381072998} 02/24/2022 23:40:27 - INFO - codeparrot_training - Step 17059: {'lr': 0.00038808883620965705, 'samples': 8734720, 'steps': 17059, 'loss/train': 2.3146941661834717} 02/24/2022 23:40:33 - INFO - codeparrot_training - Step 17060: {'lr': 0.00038807519602238174, 'samples': 8735232, 'steps': 17060, 'loss/train': 1.8632471561431885} 02/24/2022 23:40:36 - INFO - codeparrot_training - Step 17061: {'lr': 0.00038806155524363594, 'samples': 8735744, 'steps': 17061, 'loss/train': 1.590744972229004} 02/24/2022 23:40:43 - INFO - codeparrot_training - Step 17062: {'lr': 0.00038804791387347844, 'samples': 8736256, 'steps': 17062, 'loss/train': 0.5118932723999023} 02/24/2022 23:40:47 - INFO - codeparrot_training - Step 17063: {'lr': 0.0003880342719119675, 'samples': 8736768, 'steps': 17063, 'loss/train': 2.611489772796631} 02/24/2022 23:40:52 - INFO - codeparrot_training - Step 17064: {'lr': 0.0003880206293591615, 'samples': 8737280, 'steps': 17064, 'loss/train': 1.9074097871780396} 02/24/2022 23:40:56 - INFO - codeparrot_training - Step 17065: {'lr': 0.000388006986215119, 'samples': 8737792, 'steps': 17065, 'loss/train': 1.3274916410446167} 02/24/2022 23:41:01 - INFO - codeparrot_training - Step 17066: {'lr': 0.0003879933424798984, 'samples': 8738304, 'steps': 17066, 'loss/train': 0.15926577150821686} 02/24/2022 23:41:05 - INFO - codeparrot_training - Step 17067: {'lr': 0.0003879796981535582, 'samples': 8738816, 'steps': 17067, 'loss/train': 2.1513419151306152} 02/24/2022 23:41:10 - INFO - codeparrot_training - Step 17068: {'lr': 0.00038796605323615664, 'samples': 8739328, 'steps': 17068, 'loss/train': 2.6258888244628906} 02/24/2022 23:41:14 - INFO - codeparrot_training - Step 17069: {'lr': 0.00038795240772775244, 'samples': 8739840, 'steps': 17069, 'loss/train': 1.524785041809082} 02/24/2022 23:41:19 - INFO - codeparrot_training - Step 17070: {'lr': 0.0003879387616284038, 'samples': 8740352, 'steps': 17070, 'loss/train': 2.2197866439819336} 02/24/2022 23:41:23 - INFO - codeparrot_training - Step 17071: {'lr': 0.0003879251149381694, 'samples': 8740864, 'steps': 17071, 'loss/train': 2.068557024002075} 02/24/2022 23:41:30 - INFO - codeparrot_training - Step 17072: {'lr': 0.0003879114676571076, 'samples': 8741376, 'steps': 17072, 'loss/train': 2.4184749126434326} 02/24/2022 23:41:34 - INFO - codeparrot_training - Step 17073: {'lr': 0.00038789781978527683, 'samples': 8741888, 'steps': 17073, 'loss/train': 2.5602409839630127} 02/24/2022 23:41:39 - INFO - codeparrot_training - Step 17074: {'lr': 0.0003878841713227356, 'samples': 8742400, 'steps': 17074, 'loss/train': 2.6567413806915283} 02/24/2022 23:41:43 - INFO - codeparrot_training - Step 17075: {'lr': 0.00038787052226954235, 'samples': 8742912, 'steps': 17075, 'loss/train': 2.5525169372558594} 02/24/2022 23:41:48 - INFO - codeparrot_training - Step 17076: {'lr': 0.0003878568726257556, 'samples': 8743424, 'steps': 17076, 'loss/train': 4.4807233810424805} 02/24/2022 23:41:52 - INFO - codeparrot_training - Step 17077: {'lr': 0.0003878432223914338, 'samples': 8743936, 'steps': 17077, 'loss/train': 2.2350618839263916} 02/24/2022 23:41:57 - INFO - codeparrot_training - Step 17078: {'lr': 0.00038782957156663535, 'samples': 8744448, 'steps': 17078, 'loss/train': 2.6714046001434326} 02/24/2022 23:42:01 - INFO - codeparrot_training - Step 17079: {'lr': 0.0003878159201514188, 'samples': 8744960, 'steps': 17079, 'loss/train': 2.422719717025757} 02/24/2022 23:42:07 - INFO - codeparrot_training - Step 17080: {'lr': 0.00038780226814584263, 'samples': 8745472, 'steps': 17080, 'loss/train': 1.8320813179016113} 02/24/2022 23:42:10 - INFO - codeparrot_training - Step 17081: {'lr': 0.00038778861554996524, 'samples': 8745984, 'steps': 17081, 'loss/train': 1.4904323816299438} 02/24/2022 23:42:16 - INFO - codeparrot_training - Step 17082: {'lr': 0.00038777496236384526, 'samples': 8746496, 'steps': 17082, 'loss/train': 1.8217389583587646} 02/24/2022 23:42:19 - INFO - codeparrot_training - Step 17083: {'lr': 0.000387761308587541, 'samples': 8747008, 'steps': 17083, 'loss/train': 2.243881940841675} 02/24/2022 23:42:25 - INFO - codeparrot_training - Step 17084: {'lr': 0.0003877476542211111, 'samples': 8747520, 'steps': 17084, 'loss/train': 0.8803138136863708} 02/24/2022 23:42:28 - INFO - codeparrot_training - Step 17085: {'lr': 0.00038773399926461395, 'samples': 8748032, 'steps': 17085, 'loss/train': 2.692265510559082} 02/24/2022 23:42:34 - INFO - codeparrot_training - Step 17086: {'lr': 0.0003877203437181081, 'samples': 8748544, 'steps': 17086, 'loss/train': 2.32560133934021} 02/24/2022 23:42:37 - INFO - codeparrot_training - Step 17087: {'lr': 0.0003877066875816521, 'samples': 8749056, 'steps': 17087, 'loss/train': 1.8611980676651} 02/24/2022 23:42:45 - INFO - codeparrot_training - Step 17088: {'lr': 0.00038769303085530425, 'samples': 8749568, 'steps': 17088, 'loss/train': 1.200783610343933} 02/24/2022 23:42:48 - INFO - codeparrot_training - Step 17089: {'lr': 0.0003876793735391233, 'samples': 8750080, 'steps': 17089, 'loss/train': 2.1668903827667236} 02/24/2022 23:42:54 - INFO - codeparrot_training - Step 17090: {'lr': 0.00038766571563316756, 'samples': 8750592, 'steps': 17090, 'loss/train': 1.193395733833313} 02/24/2022 23:42:57 - INFO - codeparrot_training - Step 17091: {'lr': 0.00038765205713749563, 'samples': 8751104, 'steps': 17091, 'loss/train': 1.4541212320327759} 02/24/2022 23:43:03 - INFO - codeparrot_training - Step 17092: {'lr': 0.0003876383980521659, 'samples': 8751616, 'steps': 17092, 'loss/train': 1.65485680103302} 02/24/2022 23:43:06 - INFO - codeparrot_training - Step 17093: {'lr': 0.0003876247383772371, 'samples': 8752128, 'steps': 17093, 'loss/train': 0.4849706292152405} 02/24/2022 23:43:12 - INFO - codeparrot_training - Step 17094: {'lr': 0.00038761107811276756, 'samples': 8752640, 'steps': 17094, 'loss/train': 2.5173490047454834} 02/24/2022 23:43:15 - INFO - codeparrot_training - Step 17095: {'lr': 0.00038759741725881593, 'samples': 8753152, 'steps': 17095, 'loss/train': 3.143068552017212} 02/24/2022 23:43:21 - INFO - codeparrot_training - Step 17096: {'lr': 0.0003875837558154406, 'samples': 8753664, 'steps': 17096, 'loss/train': 2.195749044418335} 02/24/2022 23:43:24 - INFO - codeparrot_training - Step 17097: {'lr': 0.00038757009378270014, 'samples': 8754176, 'steps': 17097, 'loss/train': 1.4201284646987915} 02/24/2022 23:43:32 - INFO - codeparrot_training - Step 17098: {'lr': 0.0003875564311606531, 'samples': 8754688, 'steps': 17098, 'loss/train': 0.7694094181060791} 02/24/2022 23:43:35 - INFO - codeparrot_training - Step 17099: {'lr': 0.000387542767949358, 'samples': 8755200, 'steps': 17099, 'loss/train': 1.7876856327056885} 02/24/2022 23:43:41 - INFO - codeparrot_training - Step 17100: {'lr': 0.0003875291041488734, 'samples': 8755712, 'steps': 17100, 'loss/train': 1.1753270626068115} 02/24/2022 23:43:44 - INFO - codeparrot_training - Step 17101: {'lr': 0.00038751543975925766, 'samples': 8756224, 'steps': 17101, 'loss/train': 2.3657734394073486} 02/24/2022 23:43:50 - INFO - codeparrot_training - Step 17102: {'lr': 0.00038750177478056956, 'samples': 8756736, 'steps': 17102, 'loss/train': 1.3064559698104858} 02/24/2022 23:43:53 - INFO - codeparrot_training - Step 17103: {'lr': 0.0003874881092128675, 'samples': 8757248, 'steps': 17103, 'loss/train': 2.9817073345184326} 02/24/2022 23:43:59 - INFO - codeparrot_training - Step 17104: {'lr': 0.00038747444305621, 'samples': 8757760, 'steps': 17104, 'loss/train': 0.7996066808700562} 02/24/2022 23:44:02 - INFO - codeparrot_training - Step 17105: {'lr': 0.0003874607763106556, 'samples': 8758272, 'steps': 17105, 'loss/train': 2.700169801712036} 02/24/2022 23:44:08 - INFO - codeparrot_training - Step 17106: {'lr': 0.00038744710897626293, 'samples': 8758784, 'steps': 17106, 'loss/train': 2.0943641662597656} 02/24/2022 23:44:11 - INFO - codeparrot_training - Step 17107: {'lr': 0.00038743344105309055, 'samples': 8759296, 'steps': 17107, 'loss/train': 2.6027300357818604} 02/24/2022 23:44:19 - INFO - codeparrot_training - Step 17108: {'lr': 0.0003874197725411969, 'samples': 8759808, 'steps': 17108, 'loss/train': 1.9064092636108398} 02/24/2022 23:44:22 - INFO - codeparrot_training - Step 17109: {'lr': 0.0003874061034406405, 'samples': 8760320, 'steps': 17109, 'loss/train': 1.881370186805725} 02/24/2022 23:44:28 - INFO - codeparrot_training - Step 17110: {'lr': 0.00038739243375148, 'samples': 8760832, 'steps': 17110, 'loss/train': 2.559566020965576} 02/24/2022 23:44:31 - INFO - codeparrot_training - Step 17111: {'lr': 0.0003873787634737741, 'samples': 8761344, 'steps': 17111, 'loss/train': 2.025768995285034} 02/24/2022 23:44:37 - INFO - codeparrot_training - Step 17112: {'lr': 0.00038736509260758103, 'samples': 8761856, 'steps': 17112, 'loss/train': 1.2163435220718384} 02/24/2022 23:44:40 - INFO - codeparrot_training - Step 17113: {'lr': 0.00038735142115295965, 'samples': 8762368, 'steps': 17113, 'loss/train': 1.903591513633728} 02/24/2022 23:44:46 - INFO - codeparrot_training - Step 17114: {'lr': 0.00038733774910996825, 'samples': 8762880, 'steps': 17114, 'loss/train': 1.4680233001708984} 02/24/2022 23:44:49 - INFO - codeparrot_training - Step 17115: {'lr': 0.00038732407647866567, 'samples': 8763392, 'steps': 17115, 'loss/train': 2.2477405071258545} 02/24/2022 23:44:55 - INFO - codeparrot_training - Step 17116: {'lr': 0.00038731040325911027, 'samples': 8763904, 'steps': 17116, 'loss/train': 1.5594722032546997} 02/24/2022 23:44:58 - INFO - codeparrot_training - Step 17117: {'lr': 0.0003872967294513608, 'samples': 8764416, 'steps': 17117, 'loss/train': 1.8007839918136597} 02/24/2022 23:45:05 - INFO - codeparrot_training - Step 17118: {'lr': 0.0003872830550554757, 'samples': 8764928, 'steps': 17118, 'loss/train': 2.6001248359680176} 02/24/2022 23:45:09 - INFO - codeparrot_training - Step 17119: {'lr': 0.0003872693800715135, 'samples': 8765440, 'steps': 17119, 'loss/train': 2.1997592449188232} 02/24/2022 23:45:14 - INFO - codeparrot_training - Step 17120: {'lr': 0.00038725570449953296, 'samples': 8765952, 'steps': 17120, 'loss/train': 2.052170991897583} 02/24/2022 23:45:18 - INFO - codeparrot_training - Step 17121: {'lr': 0.00038724202833959254, 'samples': 8766464, 'steps': 17121, 'loss/train': 2.8242053985595703} 02/24/2022 23:45:23 - INFO - codeparrot_training - Step 17122: {'lr': 0.00038722835159175087, 'samples': 8766976, 'steps': 17122, 'loss/train': 2.0248873233795166} 02/24/2022 23:45:27 - INFO - codeparrot_training - Step 17123: {'lr': 0.00038721467425606644, 'samples': 8767488, 'steps': 17123, 'loss/train': 1.7550607919692993} 02/24/2022 23:45:32 - INFO - codeparrot_training - Step 17124: {'lr': 0.000387200996332598, 'samples': 8768000, 'steps': 17124, 'loss/train': 1.693726897239685} 02/24/2022 23:45:36 - INFO - codeparrot_training - Step 17125: {'lr': 0.000387187317821404, 'samples': 8768512, 'steps': 17125, 'loss/train': 1.760815143585205} 02/24/2022 23:45:41 - INFO - codeparrot_training - Step 17126: {'lr': 0.0003871736387225431, 'samples': 8769024, 'steps': 17126, 'loss/train': 3.0384926795959473} 02/24/2022 23:45:45 - INFO - codeparrot_training - Step 17127: {'lr': 0.0003871599590360739, 'samples': 8769536, 'steps': 17127, 'loss/train': 2.5729176998138428} 02/24/2022 23:45:52 - INFO - codeparrot_training - Step 17128: {'lr': 0.000387146278762055, 'samples': 8770048, 'steps': 17128, 'loss/train': 1.2496367692947388} 02/24/2022 23:45:56 - INFO - codeparrot_training - Step 17129: {'lr': 0.000387132597900545, 'samples': 8770560, 'steps': 17129, 'loss/train': 2.1538803577423096} 02/24/2022 23:46:01 - INFO - codeparrot_training - Step 17130: {'lr': 0.0003871189164516025, 'samples': 8771072, 'steps': 17130, 'loss/train': 2.379394769668579} 02/24/2022 23:46:05 - INFO - codeparrot_training - Step 17131: {'lr': 0.000387105234415286, 'samples': 8771584, 'steps': 17131, 'loss/train': 1.665564775466919} 02/24/2022 23:46:10 - INFO - codeparrot_training - Step 17132: {'lr': 0.00038709155179165436, 'samples': 8772096, 'steps': 17132, 'loss/train': 2.272392749786377} 02/24/2022 23:46:14 - INFO - codeparrot_training - Step 17133: {'lr': 0.000387077868580766, 'samples': 8772608, 'steps': 17133, 'loss/train': 0.39942196011543274} 02/24/2022 23:46:19 - INFO - codeparrot_training - Step 17134: {'lr': 0.00038706418478267945, 'samples': 8773120, 'steps': 17134, 'loss/train': 1.7531652450561523} 02/24/2022 23:46:23 - INFO - codeparrot_training - Step 17135: {'lr': 0.0003870505003974536, 'samples': 8773632, 'steps': 17135, 'loss/train': 1.9936140775680542} 02/24/2022 23:46:28 - INFO - codeparrot_training - Step 17136: {'lr': 0.0003870368154251469, 'samples': 8774144, 'steps': 17136, 'loss/train': 1.8988162279129028} 02/24/2022 23:46:32 - INFO - codeparrot_training - Step 17137: {'lr': 0.000387023129865818, 'samples': 8774656, 'steps': 17137, 'loss/train': 2.8043172359466553} 02/24/2022 23:46:37 - INFO - codeparrot_training - Step 17138: {'lr': 0.00038700944371952543, 'samples': 8775168, 'steps': 17138, 'loss/train': 1.4487859010696411} 02/24/2022 23:46:41 - INFO - codeparrot_training - Step 17139: {'lr': 0.00038699575698632806, 'samples': 8775680, 'steps': 17139, 'loss/train': 3.4096124172210693} 02/24/2022 23:46:46 - INFO - codeparrot_training - Step 17140: {'lr': 0.00038698206966628426, 'samples': 8776192, 'steps': 17140, 'loss/train': 2.037593364715576} 02/24/2022 23:46:50 - INFO - codeparrot_training - Step 17141: {'lr': 0.00038696838175945284, 'samples': 8776704, 'steps': 17141, 'loss/train': 2.4670138359069824} 02/24/2022 23:46:55 - INFO - codeparrot_training - Step 17142: {'lr': 0.0003869546932658923, 'samples': 8777216, 'steps': 17142, 'loss/train': 1.7054975032806396} 02/24/2022 23:46:59 - INFO - codeparrot_training - Step 17143: {'lr': 0.0003869410041856614, 'samples': 8777728, 'steps': 17143, 'loss/train': 2.2802484035491943} 02/24/2022 23:47:04 - INFO - codeparrot_training - Step 17144: {'lr': 0.0003869273145188186, 'samples': 8778240, 'steps': 17144, 'loss/train': 2.6949515342712402} 02/24/2022 23:47:08 - INFO - codeparrot_training - Step 17145: {'lr': 0.00038691362426542273, 'samples': 8778752, 'steps': 17145, 'loss/train': 3.853553056716919} 02/24/2022 23:47:15 - INFO - codeparrot_training - Step 17146: {'lr': 0.0003868999334255324, 'samples': 8779264, 'steps': 17146, 'loss/train': 1.9217404127120972} 02/24/2022 23:47:19 - INFO - codeparrot_training - Step 17147: {'lr': 0.00038688624199920623, 'samples': 8779776, 'steps': 17147, 'loss/train': 2.070007085800171} 02/24/2022 23:47:24 - INFO - codeparrot_training - Step 17148: {'lr': 0.0003868725499865029, 'samples': 8780288, 'steps': 17148, 'loss/train': 2.3876490592956543} 02/24/2022 23:47:27 - INFO - codeparrot_training - Step 17149: {'lr': 0.00038685885738748096, 'samples': 8780800, 'steps': 17149, 'loss/train': 2.0504651069641113} 02/24/2022 23:47:33 - INFO - codeparrot_training - Step 17150: {'lr': 0.0003868451642021992, 'samples': 8781312, 'steps': 17150, 'loss/train': 2.0499277114868164} 02/24/2022 23:47:37 - INFO - codeparrot_training - Step 17151: {'lr': 0.0003868314704307161, 'samples': 8781824, 'steps': 17151, 'loss/train': 2.0177595615386963} 02/24/2022 23:47:43 - INFO - codeparrot_training - Step 17152: {'lr': 0.0003868177760730905, 'samples': 8782336, 'steps': 17152, 'loss/train': 0.8068851232528687} 02/24/2022 23:47:46 - INFO - codeparrot_training - Step 17153: {'lr': 0.00038680408112938097, 'samples': 8782848, 'steps': 17153, 'loss/train': 1.987796664237976} 02/24/2022 23:47:53 - INFO - codeparrot_training - Step 17154: {'lr': 0.00038679038559964626, 'samples': 8783360, 'steps': 17154, 'loss/train': 1.763417363166809} 02/24/2022 23:47:57 - INFO - codeparrot_training - Step 17155: {'lr': 0.0003867766894839449, 'samples': 8783872, 'steps': 17155, 'loss/train': 1.6264234781265259} 02/24/2022 23:48:03 - INFO - codeparrot_training - Step 17156: {'lr': 0.0003867629927823357, 'samples': 8784384, 'steps': 17156, 'loss/train': 2.31601619720459} 02/24/2022 23:48:06 - INFO - codeparrot_training - Step 17157: {'lr': 0.00038674929549487714, 'samples': 8784896, 'steps': 17157, 'loss/train': 2.436521053314209} 02/24/2022 23:48:12 - INFO - codeparrot_training - Step 17158: {'lr': 0.00038673559762162816, 'samples': 8785408, 'steps': 17158, 'loss/train': 2.391249418258667} 02/24/2022 23:48:15 - INFO - codeparrot_training - Step 17159: {'lr': 0.0003867218991626472, 'samples': 8785920, 'steps': 17159, 'loss/train': 1.0393623113632202} 02/24/2022 23:48:21 - INFO - codeparrot_training - Step 17160: {'lr': 0.0003867082001179931, 'samples': 8786432, 'steps': 17160, 'loss/train': 2.1306588649749756} 02/24/2022 23:48:24 - INFO - codeparrot_training - Step 17161: {'lr': 0.0003866945004877245, 'samples': 8786944, 'steps': 17161, 'loss/train': 3.0967583656311035} 02/24/2022 23:48:30 - INFO - codeparrot_training - Step 17162: {'lr': 0.0003866808002719, 'samples': 8787456, 'steps': 17162, 'loss/train': 1.4777824878692627} 02/24/2022 23:48:33 - INFO - codeparrot_training - Step 17163: {'lr': 0.00038666709947057836, 'samples': 8787968, 'steps': 17163, 'loss/train': 2.7453489303588867} 02/24/2022 23:48:41 - INFO - codeparrot_training - Step 17164: {'lr': 0.0003866533980838183, 'samples': 8788480, 'steps': 17164, 'loss/train': 2.041107654571533} 02/24/2022 23:48:44 - INFO - codeparrot_training - Step 17165: {'lr': 0.0003866396961116785, 'samples': 8788992, 'steps': 17165, 'loss/train': 2.743147850036621} 02/24/2022 23:48:50 - INFO - codeparrot_training - Step 17166: {'lr': 0.00038662599355421756, 'samples': 8789504, 'steps': 17166, 'loss/train': 2.681450843811035} 02/24/2022 23:48:53 - INFO - codeparrot_training - Step 17167: {'lr': 0.00038661229041149427, 'samples': 8790016, 'steps': 17167, 'loss/train': 2.6153838634490967} 02/24/2022 23:48:59 - INFO - codeparrot_training - Step 17168: {'lr': 0.0003865985866835673, 'samples': 8790528, 'steps': 17168, 'loss/train': 2.2887461185455322} 02/24/2022 23:49:02 - INFO - codeparrot_training - Step 17169: {'lr': 0.0003865848823704954, 'samples': 8791040, 'steps': 17169, 'loss/train': 2.4888033866882324} 02/24/2022 23:49:08 - INFO - codeparrot_training - Step 17170: {'lr': 0.00038657117747233717, 'samples': 8791552, 'steps': 17170, 'loss/train': 1.12760329246521} 02/24/2022 23:49:11 - INFO - codeparrot_training - Step 17171: {'lr': 0.00038655747198915137, 'samples': 8792064, 'steps': 17171, 'loss/train': 3.0158510208129883} 02/24/2022 23:49:17 - INFO - codeparrot_training - Step 17172: {'lr': 0.0003865437659209968, 'samples': 8792576, 'steps': 17172, 'loss/train': 1.4976574182510376} 02/24/2022 23:49:20 - INFO - codeparrot_training - Step 17173: {'lr': 0.00038653005926793203, 'samples': 8793088, 'steps': 17173, 'loss/train': 1.0372263193130493} 02/24/2022 23:49:26 - INFO - codeparrot_training - Step 17174: {'lr': 0.0003865163520300159, 'samples': 8793600, 'steps': 17174, 'loss/train': 2.25982666015625} 02/24/2022 23:49:29 - INFO - codeparrot_training - Step 17175: {'lr': 0.00038650264420730707, 'samples': 8794112, 'steps': 17175, 'loss/train': 2.554713010787964} 02/24/2022 23:49:37 - INFO - codeparrot_training - Step 17176: {'lr': 0.00038648893579986424, 'samples': 8794624, 'steps': 17176, 'loss/train': 1.747277855873108} 02/24/2022 23:49:40 - INFO - codeparrot_training - Step 17177: {'lr': 0.00038647522680774603, 'samples': 8795136, 'steps': 17177, 'loss/train': 0.9550808072090149} 02/24/2022 23:49:46 - INFO - codeparrot_training - Step 17178: {'lr': 0.0003864615172310115, 'samples': 8795648, 'steps': 17178, 'loss/train': 2.0351920127868652} 02/24/2022 23:49:49 - INFO - codeparrot_training - Step 17179: {'lr': 0.000386447807069719, 'samples': 8796160, 'steps': 17179, 'loss/train': 3.6719768047332764} 02/24/2022 23:49:55 - INFO - codeparrot_training - Step 17180: {'lr': 0.0003864340963239275, 'samples': 8796672, 'steps': 17180, 'loss/train': 1.9507302045822144} 02/24/2022 23:49:58 - INFO - codeparrot_training - Step 17181: {'lr': 0.00038642038499369556, 'samples': 8797184, 'steps': 17181, 'loss/train': 1.3756699562072754} 02/24/2022 23:50:04 - INFO - codeparrot_training - Step 17182: {'lr': 0.0003864066730790821, 'samples': 8797696, 'steps': 17182, 'loss/train': 1.7046701908111572} 02/24/2022 23:50:07 - INFO - codeparrot_training - Step 17183: {'lr': 0.00038639296058014575, 'samples': 8798208, 'steps': 17183, 'loss/train': 1.9142571687698364} 02/24/2022 23:50:12 - INFO - codeparrot_training - Step 17184: {'lr': 0.0003863792474969453, 'samples': 8798720, 'steps': 17184, 'loss/train': 1.423899531364441} 02/24/2022 23:50:16 - INFO - codeparrot_training - Step 17185: {'lr': 0.00038636553382953944, 'samples': 8799232, 'steps': 17185, 'loss/train': 1.5073208808898926} 02/24/2022 23:50:21 - INFO - codeparrot_training - Step 17186: {'lr': 0.00038635181957798686, 'samples': 8799744, 'steps': 17186, 'loss/train': 1.9099687337875366} 02/24/2022 23:50:25 - INFO - codeparrot_training - Step 17187: {'lr': 0.00038633810474234643, 'samples': 8800256, 'steps': 17187, 'loss/train': 2.5165624618530273} 02/24/2022 23:50:30 - INFO - codeparrot_training - Step 17188: {'lr': 0.00038632438932267686, 'samples': 8800768, 'steps': 17188, 'loss/train': 3.1032791137695312} 02/24/2022 23:50:34 - INFO - codeparrot_training - Step 17189: {'lr': 0.0003863106733190369, 'samples': 8801280, 'steps': 17189, 'loss/train': 2.5249416828155518} 02/24/2022 23:50:41 - INFO - codeparrot_training - Step 17190: {'lr': 0.0003862969567314852, 'samples': 8801792, 'steps': 17190, 'loss/train': 0.4238697290420532} 02/24/2022 23:50:45 - INFO - codeparrot_training - Step 17191: {'lr': 0.0003862832395600808, 'samples': 8802304, 'steps': 17191, 'loss/train': 0.6999640464782715} 02/24/2022 23:50:50 - INFO - codeparrot_training - Step 17192: {'lr': 0.0003862695218048822, 'samples': 8802816, 'steps': 17192, 'loss/train': 1.8220654726028442} 02/24/2022 23:50:54 - INFO - codeparrot_training - Step 17193: {'lr': 0.00038625580346594824, 'samples': 8803328, 'steps': 17193, 'loss/train': 6.060995578765869} 02/24/2022 23:50:59 - INFO - codeparrot_training - Step 17194: {'lr': 0.00038624208454333763, 'samples': 8803840, 'steps': 17194, 'loss/train': 2.0100319385528564} 02/24/2022 23:51:03 - INFO - codeparrot_training - Step 17195: {'lr': 0.00038622836503710917, 'samples': 8804352, 'steps': 17195, 'loss/train': 1.8906749486923218} 02/24/2022 23:51:09 - INFO - codeparrot_training - Step 17196: {'lr': 0.00038621464494732174, 'samples': 8804864, 'steps': 17196, 'loss/train': 3.2008378505706787} 02/24/2022 23:51:12 - INFO - codeparrot_training - Step 17197: {'lr': 0.00038620092427403395, 'samples': 8805376, 'steps': 17197, 'loss/train': 1.7259007692337036} 02/24/2022 23:51:18 - INFO - codeparrot_training - Step 17198: {'lr': 0.0003861872030173047, 'samples': 8805888, 'steps': 17198, 'loss/train': 2.322706699371338} 02/24/2022 23:51:21 - INFO - codeparrot_training - Step 17199: {'lr': 0.0003861734811771928, 'samples': 8806400, 'steps': 17199, 'loss/train': 1.5288407802581787} 02/24/2022 23:51:27 - INFO - codeparrot_training - Step 17200: {'lr': 0.00038615975875375683, 'samples': 8806912, 'steps': 17200, 'loss/train': 1.030848503112793} 02/24/2022 23:51:30 - INFO - codeparrot_training - Step 17201: {'lr': 0.0003861460357470556, 'samples': 8807424, 'steps': 17201, 'loss/train': 1.8166879415512085} 02/24/2022 23:51:38 - INFO - codeparrot_training - Step 17202: {'lr': 0.0003861323121571482, 'samples': 8807936, 'steps': 17202, 'loss/train': 1.5446628332138062} 02/24/2022 23:51:41 - INFO - codeparrot_training - Step 17203: {'lr': 0.0003861185879840931, 'samples': 8808448, 'steps': 17203, 'loss/train': 1.6572048664093018} 02/24/2022 23:51:47 - INFO - codeparrot_training - Step 17204: {'lr': 0.00038610486322794915, 'samples': 8808960, 'steps': 17204, 'loss/train': 0.9908636212348938} 02/24/2022 23:51:51 - INFO - codeparrot_training - Step 17205: {'lr': 0.0003860911378887752, 'samples': 8809472, 'steps': 17205, 'loss/train': 2.543142080307007} 02/24/2022 23:51:56 - INFO - codeparrot_training - Step 17206: {'lr': 0.00038607741196663005, 'samples': 8809984, 'steps': 17206, 'loss/train': 1.8947288990020752} 02/24/2022 23:52:00 - INFO - codeparrot_training - Step 17207: {'lr': 0.0003860636854615725, 'samples': 8810496, 'steps': 17207, 'loss/train': 1.871793508529663} 02/24/2022 23:52:05 - INFO - codeparrot_training - Step 17208: {'lr': 0.0003860499583736613, 'samples': 8811008, 'steps': 17208, 'loss/train': 2.360804796218872} 02/24/2022 23:52:09 - INFO - codeparrot_training - Step 17209: {'lr': 0.00038603623070295536, 'samples': 8811520, 'steps': 17209, 'loss/train': 1.8559503555297852} 02/24/2022 23:52:14 - INFO - codeparrot_training - Step 17210: {'lr': 0.0003860225024495133, 'samples': 8812032, 'steps': 17210, 'loss/train': 2.4367282390594482} 02/24/2022 23:52:18 - INFO - codeparrot_training - Step 17211: {'lr': 0.000386008773613394, 'samples': 8812544, 'steps': 17211, 'loss/train': 2.1229898929595947} 02/24/2022 23:52:24 - INFO - codeparrot_training - Step 17212: {'lr': 0.0003859950441946564, 'samples': 8813056, 'steps': 17212, 'loss/train': 2.0907034873962402} 02/24/2022 23:52:27 - INFO - codeparrot_training - Step 17213: {'lr': 0.0003859813141933592, 'samples': 8813568, 'steps': 17213, 'loss/train': 2.2270755767822266} 02/24/2022 23:52:33 - INFO - codeparrot_training - Step 17214: {'lr': 0.0003859675836095612, 'samples': 8814080, 'steps': 17214, 'loss/train': 0.8479047417640686} 02/24/2022 23:52:36 - INFO - codeparrot_training - Step 17215: {'lr': 0.00038595385244332125, 'samples': 8814592, 'steps': 17215, 'loss/train': 2.7664647102355957} 02/24/2022 23:52:42 - INFO - codeparrot_training - Step 17216: {'lr': 0.00038594012069469814, 'samples': 8815104, 'steps': 17216, 'loss/train': 1.1028177738189697} 02/24/2022 23:52:45 - INFO - codeparrot_training - Step 17217: {'lr': 0.00038592638836375075, 'samples': 8815616, 'steps': 17217, 'loss/train': 3.339601993560791} 02/24/2022 23:52:51 - INFO - codeparrot_training - Step 17218: {'lr': 0.0003859126554505379, 'samples': 8816128, 'steps': 17218, 'loss/train': 2.159816265106201} 02/24/2022 23:52:54 - INFO - codeparrot_training - Step 17219: {'lr': 0.00038589892195511834, 'samples': 8816640, 'steps': 17219, 'loss/train': 4.061238765716553} 02/24/2022 23:53:00 - INFO - codeparrot_training - Step 17220: {'lr': 0.00038588518787755096, 'samples': 8817152, 'steps': 17220, 'loss/train': 1.8575063943862915} 02/24/2022 23:53:03 - INFO - codeparrot_training - Step 17221: {'lr': 0.00038587145321789456, 'samples': 8817664, 'steps': 17221, 'loss/train': 1.722158670425415} 02/24/2022 23:53:09 - INFO - codeparrot_training - Step 17222: {'lr': 0.00038585771797620803, 'samples': 8818176, 'steps': 17222, 'loss/train': 1.7295022010803223} 02/24/2022 23:53:13 - INFO - codeparrot_training - Step 17223: {'lr': 0.00038584398215255023, 'samples': 8818688, 'steps': 17223, 'loss/train': 1.8057955503463745} 02/24/2022 23:53:18 - INFO - codeparrot_training - Step 17224: {'lr': 0.0003858302457469799, 'samples': 8819200, 'steps': 17224, 'loss/train': 1.9161033630371094} 02/24/2022 23:53:22 - INFO - codeparrot_training - Step 17225: {'lr': 0.0003858165087595559, 'samples': 8819712, 'steps': 17225, 'loss/train': 1.9712824821472168} 02/24/2022 23:53:27 - INFO - codeparrot_training - Step 17226: {'lr': 0.00038580277119033715, 'samples': 8820224, 'steps': 17226, 'loss/train': 2.2184431552886963} 02/24/2022 23:53:31 - INFO - codeparrot_training - Step 17227: {'lr': 0.0003857890330393824, 'samples': 8820736, 'steps': 17227, 'loss/train': 1.8530975580215454} 02/24/2022 23:53:36 - INFO - codeparrot_training - Step 17228: {'lr': 0.0003857752943067506, 'samples': 8821248, 'steps': 17228, 'loss/train': 1.5442469120025635} 02/24/2022 23:53:40 - INFO - codeparrot_training - Step 17229: {'lr': 0.00038576155499250056, 'samples': 8821760, 'steps': 17229, 'loss/train': 2.025756359100342} 02/24/2022 23:53:45 - INFO - codeparrot_training - Step 17230: {'lr': 0.000385747815096691, 'samples': 8822272, 'steps': 17230, 'loss/train': 2.134267568588257} 02/24/2022 23:53:49 - INFO - codeparrot_training - Step 17231: {'lr': 0.00038573407461938103, 'samples': 8822784, 'steps': 17231, 'loss/train': 1.656584620475769} 02/24/2022 23:53:54 - INFO - codeparrot_training - Step 17232: {'lr': 0.0003857203335606294, 'samples': 8823296, 'steps': 17232, 'loss/train': 1.0241409540176392} 02/24/2022 23:53:58 - INFO - codeparrot_training - Step 17233: {'lr': 0.00038570659192049497, 'samples': 8823808, 'steps': 17233, 'loss/train': 1.7235249280929565} 02/24/2022 23:54:03 - INFO - codeparrot_training - Step 17234: {'lr': 0.0003856928496990364, 'samples': 8824320, 'steps': 17234, 'loss/train': 1.845015048980713} 02/24/2022 23:54:07 - INFO - codeparrot_training - Step 17235: {'lr': 0.000385679106896313, 'samples': 8824832, 'steps': 17235, 'loss/train': 2.526581048965454} 02/24/2022 23:54:13 - INFO - codeparrot_training - Step 17236: {'lr': 0.0003856653635123832, 'samples': 8825344, 'steps': 17236, 'loss/train': 2.620913028717041} 02/24/2022 23:54:17 - INFO - codeparrot_training - Step 17237: {'lr': 0.0003856516195473062, 'samples': 8825856, 'steps': 17237, 'loss/train': 1.7650483846664429} 02/24/2022 23:54:22 - INFO - codeparrot_training - Step 17238: {'lr': 0.0003856378750011407, 'samples': 8826368, 'steps': 17238, 'loss/train': 1.261871099472046} 02/24/2022 23:54:26 - INFO - codeparrot_training - Step 17239: {'lr': 0.0003856241298739456, 'samples': 8826880, 'steps': 17239, 'loss/train': 1.746508240699768} 02/24/2022 23:54:31 - INFO - codeparrot_training - Step 17240: {'lr': 0.0003856103841657797, 'samples': 8827392, 'steps': 17240, 'loss/train': 2.340733766555786} 02/24/2022 23:54:35 - INFO - codeparrot_training - Step 17241: {'lr': 0.0003855966378767021, 'samples': 8827904, 'steps': 17241, 'loss/train': 2.799269676208496} 02/24/2022 23:54:40 - INFO - codeparrot_training - Step 17242: {'lr': 0.00038558289100677144, 'samples': 8828416, 'steps': 17242, 'loss/train': 2.2604362964630127} 02/24/2022 23:54:44 - INFO - codeparrot_training - Step 17243: {'lr': 0.00038556914355604676, 'samples': 8828928, 'steps': 17243, 'loss/train': 2.04171085357666} 02/24/2022 23:54:49 - INFO - codeparrot_training - Step 17244: {'lr': 0.0003855553955245871, 'samples': 8829440, 'steps': 17244, 'loss/train': 2.540919542312622} 02/24/2022 23:54:53 - INFO - codeparrot_training - Step 17245: {'lr': 0.00038554164691245095, 'samples': 8829952, 'steps': 17245, 'loss/train': 1.5962390899658203} 02/24/2022 23:55:00 - INFO - codeparrot_training - Step 17246: {'lr': 0.00038552789771969755, 'samples': 8830464, 'steps': 17246, 'loss/train': 1.262508511543274} 02/24/2022 23:55:03 - INFO - codeparrot_training - Step 17247: {'lr': 0.00038551414794638555, 'samples': 8830976, 'steps': 17247, 'loss/train': 1.760473608970642} 02/24/2022 23:55:08 - INFO - codeparrot_training - Step 17248: {'lr': 0.00038550039759257404, 'samples': 8831488, 'steps': 17248, 'loss/train': 2.559582233428955} 02/24/2022 23:55:12 - INFO - codeparrot_training - Step 17249: {'lr': 0.0003854866466583219, 'samples': 8832000, 'steps': 17249, 'loss/train': 1.8721861839294434} 02/24/2022 23:55:17 - INFO - codeparrot_training - Step 17250: {'lr': 0.00038547289514368795, 'samples': 8832512, 'steps': 17250, 'loss/train': 2.282911539077759} 02/24/2022 23:55:21 - INFO - codeparrot_training - Step 17251: {'lr': 0.00038545914304873117, 'samples': 8833024, 'steps': 17251, 'loss/train': 2.229508638381958} 02/24/2022 23:55:26 - INFO - codeparrot_training - Step 17252: {'lr': 0.00038544539037351037, 'samples': 8833536, 'steps': 17252, 'loss/train': 2.107180118560791} 02/24/2022 23:55:30 - INFO - codeparrot_training - Step 17253: {'lr': 0.00038543163711808457, 'samples': 8834048, 'steps': 17253, 'loss/train': 2.538442850112915} 02/24/2022 23:55:35 - INFO - codeparrot_training - Step 17254: {'lr': 0.0003854178832825126, 'samples': 8834560, 'steps': 17254, 'loss/train': 2.3268327713012695} 02/24/2022 23:55:39 - INFO - codeparrot_training - Step 17255: {'lr': 0.0003854041288668534, 'samples': 8835072, 'steps': 17255, 'loss/train': 1.1370352506637573} 02/24/2022 23:55:45 - INFO - codeparrot_training - Step 17256: {'lr': 0.00038539037387116595, 'samples': 8835584, 'steps': 17256, 'loss/train': 1.4813798666000366} 02/24/2022 23:55:48 - INFO - codeparrot_training - Step 17257: {'lr': 0.0003853766182955092, 'samples': 8836096, 'steps': 17257, 'loss/train': 0.5485183000564575} 02/24/2022 23:55:55 - INFO - codeparrot_training - Step 17258: {'lr': 0.0003853628621399419, 'samples': 8836608, 'steps': 17258, 'loss/train': 1.8450720310211182} 02/24/2022 23:55:59 - INFO - codeparrot_training - Step 17259: {'lr': 0.00038534910540452305, 'samples': 8837120, 'steps': 17259, 'loss/train': 0.7206200361251831} 02/24/2022 23:56:04 - INFO - codeparrot_training - Step 17260: {'lr': 0.0003853353480893117, 'samples': 8837632, 'steps': 17260, 'loss/train': 2.4634640216827393} 02/24/2022 23:56:08 - INFO - codeparrot_training - Step 17261: {'lr': 0.0003853215901943667, 'samples': 8838144, 'steps': 17261, 'loss/train': 2.0845487117767334} 02/24/2022 23:56:13 - INFO - codeparrot_training - Step 17262: {'lr': 0.00038530783171974694, 'samples': 8838656, 'steps': 17262, 'loss/train': 2.1178362369537354} 02/24/2022 23:56:17 - INFO - codeparrot_training - Step 17263: {'lr': 0.0003852940726655114, 'samples': 8839168, 'steps': 17263, 'loss/train': 2.995131492614746} 02/24/2022 23:56:22 - INFO - codeparrot_training - Step 17264: {'lr': 0.000385280313031719, 'samples': 8839680, 'steps': 17264, 'loss/train': 2.0326082706451416} 02/24/2022 23:56:26 - INFO - codeparrot_training - Step 17265: {'lr': 0.0003852665528184287, 'samples': 8840192, 'steps': 17265, 'loss/train': 1.9007039070129395} 02/24/2022 23:56:31 - INFO - codeparrot_training - Step 17266: {'lr': 0.0003852527920256994, 'samples': 8840704, 'steps': 17266, 'loss/train': 1.9658913612365723} 02/24/2022 23:56:35 - INFO - codeparrot_training - Step 17267: {'lr': 0.00038523903065359013, 'samples': 8841216, 'steps': 17267, 'loss/train': 2.657074213027954} 02/24/2022 23:56:40 - INFO - codeparrot_training - Step 17268: {'lr': 0.0003852252687021598, 'samples': 8841728, 'steps': 17268, 'loss/train': 2.217015027999878} 02/24/2022 23:56:44 - INFO - codeparrot_training - Step 17269: {'lr': 0.00038521150617146737, 'samples': 8842240, 'steps': 17269, 'loss/train': 0.7021298408508301} 02/24/2022 23:56:50 - INFO - codeparrot_training - Step 17270: {'lr': 0.00038519774306157174, 'samples': 8842752, 'steps': 17270, 'loss/train': 2.22515869140625} 02/24/2022 23:56:53 - INFO - codeparrot_training - Step 17271: {'lr': 0.00038518397937253195, 'samples': 8843264, 'steps': 17271, 'loss/train': 0.89049232006073} 02/24/2022 23:56:59 - INFO - codeparrot_training - Step 17272: {'lr': 0.00038517021510440694, 'samples': 8843776, 'steps': 17272, 'loss/train': 2.688854217529297} 02/24/2022 23:57:02 - INFO - codeparrot_training - Step 17273: {'lr': 0.0003851564502572556, 'samples': 8844288, 'steps': 17273, 'loss/train': 1.1446787118911743} 02/24/2022 23:57:08 - INFO - codeparrot_training - Step 17274: {'lr': 0.00038514268483113694, 'samples': 8844800, 'steps': 17274, 'loss/train': 1.6221375465393066} 02/24/2022 23:57:12 - INFO - codeparrot_training - Step 17275: {'lr': 0.00038512891882610997, 'samples': 8845312, 'steps': 17275, 'loss/train': 2.0164897441864014} 02/24/2022 23:57:18 - INFO - codeparrot_training - Step 17276: {'lr': 0.0003851151522422336, 'samples': 8845824, 'steps': 17276, 'loss/train': 2.2144687175750732} 02/24/2022 23:57:21 - INFO - codeparrot_training - Step 17277: {'lr': 0.0003851013850795668, 'samples': 8846336, 'steps': 17277, 'loss/train': 2.7005553245544434} 02/24/2022 23:57:27 - INFO - codeparrot_training - Step 17278: {'lr': 0.00038508761733816864, 'samples': 8846848, 'steps': 17278, 'loss/train': 1.0483784675598145} 02/24/2022 23:57:30 - INFO - codeparrot_training - Step 17279: {'lr': 0.00038507384901809795, 'samples': 8847360, 'steps': 17279, 'loss/train': 2.0545334815979004} 02/24/2022 23:57:36 - INFO - codeparrot_training - Step 17280: {'lr': 0.00038506008011941376, 'samples': 8847872, 'steps': 17280, 'loss/train': 1.1230887174606323} 02/24/2022 23:57:39 - INFO - codeparrot_training - Step 17281: {'lr': 0.0003850463106421751, 'samples': 8848384, 'steps': 17281, 'loss/train': 1.8626708984375} 02/24/2022 23:57:45 - INFO - codeparrot_training - Step 17282: {'lr': 0.000385032540586441, 'samples': 8848896, 'steps': 17282, 'loss/train': 1.9536232948303223} 02/24/2022 23:57:48 - INFO - codeparrot_training - Step 17283: {'lr': 0.00038501876995227023, 'samples': 8849408, 'steps': 17283, 'loss/train': 2.213552236557007} 02/24/2022 23:57:54 - INFO - codeparrot_training - Step 17284: {'lr': 0.00038500499873972204, 'samples': 8849920, 'steps': 17284, 'loss/train': 1.6882789134979248} 02/24/2022 23:57:58 - INFO - codeparrot_training - Step 17285: {'lr': 0.0003849912269488552, 'samples': 8850432, 'steps': 17285, 'loss/train': 2.7493436336517334} 02/24/2022 23:58:03 - INFO - codeparrot_training - Step 17286: {'lr': 0.000384977454579729, 'samples': 8850944, 'steps': 17286, 'loss/train': 0.14187151193618774} 02/24/2022 23:58:07 - INFO - codeparrot_training - Step 17287: {'lr': 0.00038496368163240215, 'samples': 8851456, 'steps': 17287, 'loss/train': 1.7262598276138306} 02/24/2022 23:58:12 - INFO - codeparrot_training - Step 17288: {'lr': 0.00038494990810693366, 'samples': 8851968, 'steps': 17288, 'loss/train': 1.8332463502883911} 02/24/2022 23:58:15 - INFO - codeparrot_training - Step 17289: {'lr': 0.00038493613400338267, 'samples': 8852480, 'steps': 17289, 'loss/train': 2.5230519771575928} 02/24/2022 23:58:21 - INFO - codeparrot_training - Step 17290: {'lr': 0.0003849223593218082, 'samples': 8852992, 'steps': 17290, 'loss/train': 1.3553800582885742} 02/24/2022 23:58:24 - INFO - codeparrot_training - Step 17291: {'lr': 0.00038490858406226903, 'samples': 8853504, 'steps': 17291, 'loss/train': 2.525913953781128} 02/24/2022 23:58:30 - INFO - codeparrot_training - Step 17292: {'lr': 0.00038489480822482446, 'samples': 8854016, 'steps': 17292, 'loss/train': 1.7875498533248901} 02/24/2022 23:58:33 - INFO - codeparrot_training - Step 17293: {'lr': 0.00038488103180953326, 'samples': 8854528, 'steps': 17293, 'loss/train': 2.5612146854400635} 02/24/2022 23:58:40 - INFO - codeparrot_training - Step 17294: {'lr': 0.00038486725481645467, 'samples': 8855040, 'steps': 17294, 'loss/train': 1.9052436351776123} 02/24/2022 23:58:43 - INFO - codeparrot_training - Step 17295: {'lr': 0.00038485347724564746, 'samples': 8855552, 'steps': 17295, 'loss/train': 1.891305685043335} 02/24/2022 23:58:49 - INFO - codeparrot_training - Step 17296: {'lr': 0.0003848396990971709, 'samples': 8856064, 'steps': 17296, 'loss/train': 1.0746748447418213} 02/24/2022 23:58:52 - INFO - codeparrot_training - Step 17297: {'lr': 0.00038482592037108375, 'samples': 8856576, 'steps': 17297, 'loss/train': 2.259819269180298} 02/24/2022 23:58:58 - INFO - codeparrot_training - Step 17298: {'lr': 0.0003848121410674453, 'samples': 8857088, 'steps': 17298, 'loss/train': 1.4291638135910034} 02/24/2022 23:59:01 - INFO - codeparrot_training - Step 17299: {'lr': 0.0003847983611863144, 'samples': 8857600, 'steps': 17299, 'loss/train': 1.225612998008728} 02/24/2022 23:59:07 - INFO - codeparrot_training - Step 17300: {'lr': 0.0003847845807277501, 'samples': 8858112, 'steps': 17300, 'loss/train': 0.6676669120788574} 02/24/2022 23:59:10 - INFO - codeparrot_training - Step 17301: {'lr': 0.00038477079969181146, 'samples': 8858624, 'steps': 17301, 'loss/train': 2.7663023471832275} 02/24/2022 23:59:16 - INFO - codeparrot_training - Step 17302: {'lr': 0.00038475701807855753, 'samples': 8859136, 'steps': 17302, 'loss/train': 2.323258638381958} 02/24/2022 23:59:19 - INFO - codeparrot_training - Step 17303: {'lr': 0.00038474323588804727, 'samples': 8859648, 'steps': 17303, 'loss/train': 0.9643365144729614} 02/24/2022 23:59:26 - INFO - codeparrot_training - Step 17304: {'lr': 0.0003847294531203398, 'samples': 8860160, 'steps': 17304, 'loss/train': 2.4043641090393066} 02/24/2022 23:59:29 - INFO - codeparrot_training - Step 17305: {'lr': 0.0003847156697754942, 'samples': 8860672, 'steps': 17305, 'loss/train': 3.8596770763397217} 02/24/2022 23:59:35 - INFO - codeparrot_training - Step 17306: {'lr': 0.00038470188585356936, 'samples': 8861184, 'steps': 17306, 'loss/train': 1.4697163105010986} 02/24/2022 23:59:38 - INFO - codeparrot_training - Step 17307: {'lr': 0.00038468810135462445, 'samples': 8861696, 'steps': 17307, 'loss/train': 1.7332180738449097} 02/24/2022 23:59:44 - INFO - codeparrot_training - Step 17308: {'lr': 0.00038467431627871844, 'samples': 8862208, 'steps': 17308, 'loss/train': 1.3113980293273926} 02/24/2022 23:59:47 - INFO - codeparrot_training - Step 17309: {'lr': 0.0003846605306259105, 'samples': 8862720, 'steps': 17309, 'loss/train': 1.7961474657058716} 02/24/2022 23:59:53 - INFO - codeparrot_training - Step 17310: {'lr': 0.0003846467443962596, 'samples': 8863232, 'steps': 17310, 'loss/train': 2.2187306880950928} 02/24/2022 23:59:56 - INFO - codeparrot_training - Step 17311: {'lr': 0.00038463295758982475, 'samples': 8863744, 'steps': 17311, 'loss/train': 2.2249069213867188} 02/25/2022 00:00:02 - INFO - codeparrot_training - Step 17312: {'lr': 0.00038461917020666506, 'samples': 8864256, 'steps': 17312, 'loss/train': 1.267568588256836} 02/25/2022 00:00:05 - INFO - codeparrot_training - Step 17313: {'lr': 0.0003846053822468396, 'samples': 8864768, 'steps': 17313, 'loss/train': 1.2431365251541138} 02/25/2022 00:00:12 - INFO - codeparrot_training - Step 17314: {'lr': 0.00038459159371040743, 'samples': 8865280, 'steps': 17314, 'loss/train': 2.431715488433838} 02/25/2022 00:00:15 - INFO - codeparrot_training - Step 17315: {'lr': 0.0003845778045974276, 'samples': 8865792, 'steps': 17315, 'loss/train': 1.4574394226074219} 02/25/2022 00:00:21 - INFO - codeparrot_training - Step 17316: {'lr': 0.0003845640149079592, 'samples': 8866304, 'steps': 17316, 'loss/train': 1.674900770187378} 02/25/2022 00:00:24 - INFO - codeparrot_training - Step 17317: {'lr': 0.0003845502246420613, 'samples': 8866816, 'steps': 17317, 'loss/train': 1.8590947389602661} 02/25/2022 00:00:29 - INFO - codeparrot_training - Step 17318: {'lr': 0.00038453643379979295, 'samples': 8867328, 'steps': 17318, 'loss/train': 2.256387710571289} 02/25/2022 00:00:33 - INFO - codeparrot_training - Step 17319: {'lr': 0.00038452264238121326, 'samples': 8867840, 'steps': 17319, 'loss/train': 2.5044546127319336} 02/25/2022 00:00:39 - INFO - codeparrot_training - Step 17320: {'lr': 0.0003845088503863813, 'samples': 8868352, 'steps': 17320, 'loss/train': 1.5753520727157593} 02/25/2022 00:00:42 - INFO - codeparrot_training - Step 17321: {'lr': 0.0003844950578153561, 'samples': 8868864, 'steps': 17321, 'loss/train': 2.375957489013672} 02/25/2022 00:00:48 - INFO - codeparrot_training - Step 17322: {'lr': 0.00038448126466819675, 'samples': 8869376, 'steps': 17322, 'loss/train': 2.416022300720215} 02/25/2022 00:00:51 - INFO - codeparrot_training - Step 17323: {'lr': 0.00038446747094496243, 'samples': 8869888, 'steps': 17323, 'loss/train': 2.419193744659424} 02/25/2022 00:00:57 - INFO - codeparrot_training - Step 17324: {'lr': 0.00038445367664571216, 'samples': 8870400, 'steps': 17324, 'loss/train': 1.1529008150100708} 02/25/2022 00:01:00 - INFO - codeparrot_training - Step 17325: {'lr': 0.000384439881770505, 'samples': 8870912, 'steps': 17325, 'loss/train': 2.1947786808013916} 02/25/2022 00:01:06 - INFO - codeparrot_training - Step 17326: {'lr': 0.0003844260863194001, 'samples': 8871424, 'steps': 17326, 'loss/train': 0.752173662185669} 02/25/2022 00:01:09 - INFO - codeparrot_training - Step 17327: {'lr': 0.0003844122902924565, 'samples': 8871936, 'steps': 17327, 'loss/train': 2.003032684326172} 02/25/2022 00:01:15 - INFO - codeparrot_training - Step 17328: {'lr': 0.00038439849368973334, 'samples': 8872448, 'steps': 17328, 'loss/train': 1.9385327100753784} 02/25/2022 00:01:21 - INFO - codeparrot_training - Step 17329: {'lr': 0.0003843846965112897, 'samples': 8872960, 'steps': 17329, 'loss/train': 2.169112205505371} 02/25/2022 00:01:25 - INFO - codeparrot_training - Step 17330: {'lr': 0.0003843708987571847, 'samples': 8873472, 'steps': 17330, 'loss/train': 1.393505573272705} 02/25/2022 00:01:30 - INFO - codeparrot_training - Step 17331: {'lr': 0.0003843571004274775, 'samples': 8873984, 'steps': 17331, 'loss/train': 2.15645432472229} 02/25/2022 00:01:33 - INFO - codeparrot_training - Step 17332: {'lr': 0.0003843433015222271, 'samples': 8874496, 'steps': 17332, 'loss/train': 1.316759467124939} 02/25/2022 00:01:39 - INFO - codeparrot_training - Step 17333: {'lr': 0.0003843295020414926, 'samples': 8875008, 'steps': 17333, 'loss/train': 1.4930756092071533} 02/25/2022 00:01:43 - INFO - codeparrot_training - Step 17334: {'lr': 0.0003843157019853332, 'samples': 8875520, 'steps': 17334, 'loss/train': 1.4754931926727295} 02/25/2022 00:01:48 - INFO - codeparrot_training - Step 17335: {'lr': 0.00038430190135380803, 'samples': 8876032, 'steps': 17335, 'loss/train': 1.6922804117202759} 02/25/2022 00:01:52 - INFO - codeparrot_training - Step 17336: {'lr': 0.00038428810014697615, 'samples': 8876544, 'steps': 17336, 'loss/train': 1.763075351715088} 02/25/2022 00:01:57 - INFO - codeparrot_training - Step 17337: {'lr': 0.00038427429836489663, 'samples': 8877056, 'steps': 17337, 'loss/train': 2.5658202171325684} 02/25/2022 00:02:01 - INFO - codeparrot_training - Step 17338: {'lr': 0.00038426049600762867, 'samples': 8877568, 'steps': 17338, 'loss/train': 1.1538020372390747} 02/25/2022 00:02:07 - INFO - codeparrot_training - Step 17339: {'lr': 0.00038424669307523135, 'samples': 8878080, 'steps': 17339, 'loss/train': 1.071851134300232} 02/25/2022 00:02:11 - INFO - codeparrot_training - Step 17340: {'lr': 0.00038423288956776394, 'samples': 8878592, 'steps': 17340, 'loss/train': 1.8985893726348877} 02/25/2022 00:02:16 - INFO - codeparrot_training - Step 17341: {'lr': 0.00038421908548528534, 'samples': 8879104, 'steps': 17341, 'loss/train': 2.143930673599243} 02/25/2022 00:02:20 - INFO - codeparrot_training - Step 17342: {'lr': 0.0003842052808278549, 'samples': 8879616, 'steps': 17342, 'loss/train': 1.532324194908142} 02/25/2022 00:02:25 - INFO - codeparrot_training - Step 17343: {'lr': 0.0003841914755955315, 'samples': 8880128, 'steps': 17343, 'loss/train': 1.4769550561904907} 02/25/2022 00:02:29 - INFO - codeparrot_training - Step 17344: {'lr': 0.00038417766978837453, 'samples': 8880640, 'steps': 17344, 'loss/train': 1.7774338722229004} 02/25/2022 00:02:34 - INFO - codeparrot_training - Step 17345: {'lr': 0.00038416386340644305, 'samples': 8881152, 'steps': 17345, 'loss/train': 2.018763303756714} 02/25/2022 00:02:38 - INFO - codeparrot_training - Step 17346: {'lr': 0.00038415005644979616, 'samples': 8881664, 'steps': 17346, 'loss/train': 1.8801765441894531} 02/25/2022 00:02:43 - INFO - codeparrot_training - Step 17347: {'lr': 0.00038413624891849295, 'samples': 8882176, 'steps': 17347, 'loss/train': 1.8801686763763428} 02/25/2022 00:02:47 - INFO - codeparrot_training - Step 17348: {'lr': 0.00038412244081259273, 'samples': 8882688, 'steps': 17348, 'loss/train': 2.8764495849609375} 02/25/2022 00:02:53 - INFO - codeparrot_training - Step 17349: {'lr': 0.00038410863213215454, 'samples': 8883200, 'steps': 17349, 'loss/train': 2.4866943359375} 02/25/2022 00:02:56 - INFO - codeparrot_training - Step 17350: {'lr': 0.0003840948228772376, 'samples': 8883712, 'steps': 17350, 'loss/train': 1.2575478553771973} 02/25/2022 00:03:02 - INFO - codeparrot_training - Step 17351: {'lr': 0.00038408101304790096, 'samples': 8884224, 'steps': 17351, 'loss/train': 2.2058801651000977} 02/25/2022 00:03:06 - INFO - codeparrot_training - Step 17352: {'lr': 0.0003840672026442038, 'samples': 8884736, 'steps': 17352, 'loss/train': 2.506976842880249} 02/25/2022 00:03:09 - INFO - codeparrot_training - Step 17353: {'lr': 0.0003840533916662054, 'samples': 8885248, 'steps': 17353, 'loss/train': 2.8243627548217773} 02/25/2022 00:03:15 - INFO - codeparrot_training - Step 17354: {'lr': 0.00038403958011396476, 'samples': 8885760, 'steps': 17354, 'loss/train': 1.2112480401992798} 02/25/2022 00:03:18 - INFO - codeparrot_training - Step 17355: {'lr': 0.0003840257679875412, 'samples': 8886272, 'steps': 17355, 'loss/train': 2.1621241569519043} 02/25/2022 00:03:24 - INFO - codeparrot_training - Step 17356: {'lr': 0.00038401195528699374, 'samples': 8886784, 'steps': 17356, 'loss/train': 1.3151633739471436} 02/25/2022 00:03:27 - INFO - codeparrot_training - Step 17357: {'lr': 0.0003839981420123817, 'samples': 8887296, 'steps': 17357, 'loss/train': 2.6348929405212402} 02/25/2022 00:03:33 - INFO - codeparrot_training - Step 17358: {'lr': 0.00038398432816376404, 'samples': 8887808, 'steps': 17358, 'loss/train': 2.2323763370513916} 02/25/2022 00:03:36 - INFO - codeparrot_training - Step 17359: {'lr': 0.00038397051374120016, 'samples': 8888320, 'steps': 17359, 'loss/train': 2.9336204528808594} 02/25/2022 00:03:43 - INFO - codeparrot_training - Step 17360: {'lr': 0.00038395669874474915, 'samples': 8888832, 'steps': 17360, 'loss/train': 2.6371617317199707} 02/25/2022 00:03:46 - INFO - codeparrot_training - Step 17361: {'lr': 0.0003839428831744702, 'samples': 8889344, 'steps': 17361, 'loss/train': 2.2495625019073486} 02/25/2022 00:03:52 - INFO - codeparrot_training - Step 17362: {'lr': 0.0003839290670304224, 'samples': 8889856, 'steps': 17362, 'loss/train': 1.7084813117980957} 02/25/2022 00:03:55 - INFO - codeparrot_training - Step 17363: {'lr': 0.00038391525031266494, 'samples': 8890368, 'steps': 17363, 'loss/train': 1.162110447883606} 02/25/2022 00:04:01 - INFO - codeparrot_training - Step 17364: {'lr': 0.0003839014330212572, 'samples': 8890880, 'steps': 17364, 'loss/train': 1.8352888822555542} 02/25/2022 00:04:04 - INFO - codeparrot_training - Step 17365: {'lr': 0.00038388761515625815, 'samples': 8891392, 'steps': 17365, 'loss/train': 2.446359634399414} 02/25/2022 00:04:10 - INFO - codeparrot_training - Step 17366: {'lr': 0.0003838737967177271, 'samples': 8891904, 'steps': 17366, 'loss/train': 2.0594189167022705} 02/25/2022 00:04:13 - INFO - codeparrot_training - Step 17367: {'lr': 0.00038385997770572336, 'samples': 8892416, 'steps': 17367, 'loss/train': 2.2992804050445557} 02/25/2022 00:04:19 - INFO - codeparrot_training - Step 17368: {'lr': 0.0003838461581203058, 'samples': 8892928, 'steps': 17368, 'loss/train': 1.9460521936416626} 02/25/2022 00:04:22 - INFO - codeparrot_training - Step 17369: {'lr': 0.00038383233796153383, 'samples': 8893440, 'steps': 17369, 'loss/train': 1.9894386529922485} 02/25/2022 00:04:28 - INFO - codeparrot_training - Step 17370: {'lr': 0.00038381851722946663, 'samples': 8893952, 'steps': 17370, 'loss/train': 1.7741245031356812} 02/25/2022 00:04:31 - INFO - codeparrot_training - Step 17371: {'lr': 0.00038380469592416347, 'samples': 8894464, 'steps': 17371, 'loss/train': 0.7290230393409729} 02/25/2022 00:04:37 - INFO - codeparrot_training - Step 17372: {'lr': 0.00038379087404568333, 'samples': 8894976, 'steps': 17372, 'loss/train': 1.2639176845550537} 02/25/2022 00:04:40 - INFO - codeparrot_training - Step 17373: {'lr': 0.0003837770515940857, 'samples': 8895488, 'steps': 17373, 'loss/train': 2.028148889541626} 02/25/2022 00:04:46 - INFO - codeparrot_training - Step 17374: {'lr': 0.0003837632285694296, 'samples': 8896000, 'steps': 17374, 'loss/train': 2.300626516342163} 02/25/2022 00:04:49 - INFO - codeparrot_training - Step 17375: {'lr': 0.00038374940497177434, 'samples': 8896512, 'steps': 17375, 'loss/train': 2.2144341468811035} 02/25/2022 00:04:56 - INFO - codeparrot_training - Step 17376: {'lr': 0.000383735580801179, 'samples': 8897024, 'steps': 17376, 'loss/train': 1.6073862314224243} 02/25/2022 00:04:59 - INFO - codeparrot_training - Step 17377: {'lr': 0.00038372175605770305, 'samples': 8897536, 'steps': 17377, 'loss/train': 2.0909602642059326} 02/25/2022 00:05:05 - INFO - codeparrot_training - Step 17378: {'lr': 0.00038370793074140545, 'samples': 8898048, 'steps': 17378, 'loss/train': 1.5658361911773682} 02/25/2022 00:05:08 - INFO - codeparrot_training - Step 17379: {'lr': 0.00038369410485234557, 'samples': 8898560, 'steps': 17379, 'loss/train': 2.0146656036376953} 02/25/2022 00:05:14 - INFO - codeparrot_training - Step 17380: {'lr': 0.0003836802783905826, 'samples': 8899072, 'steps': 17380, 'loss/train': 2.606670618057251} 02/25/2022 00:05:17 - INFO - codeparrot_training - Step 17381: {'lr': 0.0003836664513561758, 'samples': 8899584, 'steps': 17381, 'loss/train': 1.9620325565338135} 02/25/2022 00:05:23 - INFO - codeparrot_training - Step 17382: {'lr': 0.0003836526237491843, 'samples': 8900096, 'steps': 17382, 'loss/train': 2.1051924228668213} 02/25/2022 00:05:26 - INFO - codeparrot_training - Step 17383: {'lr': 0.0003836387955696674, 'samples': 8900608, 'steps': 17383, 'loss/train': 2.5000431537628174} 02/25/2022 00:05:32 - INFO - codeparrot_training - Step 17384: {'lr': 0.00038362496681768434, 'samples': 8901120, 'steps': 17384, 'loss/train': 1.8268605470657349} 02/25/2022 00:05:35 - INFO - codeparrot_training - Step 17385: {'lr': 0.00038361113749329443, 'samples': 8901632, 'steps': 17385, 'loss/train': 1.5343314409255981} 02/25/2022 00:05:41 - INFO - codeparrot_training - Step 17386: {'lr': 0.00038359730759655674, 'samples': 8902144, 'steps': 17386, 'loss/train': 2.431920289993286} 02/25/2022 00:05:44 - INFO - codeparrot_training - Step 17387: {'lr': 0.00038358347712753063, 'samples': 8902656, 'steps': 17387, 'loss/train': 1.6329838037490845} 02/25/2022 00:05:51 - INFO - codeparrot_training - Step 17388: {'lr': 0.0003835696460862753, 'samples': 8903168, 'steps': 17388, 'loss/train': 1.812713861465454} 02/25/2022 00:05:55 - INFO - codeparrot_training - Step 17389: {'lr': 0.00038355581447285005, 'samples': 8903680, 'steps': 17389, 'loss/train': 1.0038838386535645} 02/25/2022 00:06:00 - INFO - codeparrot_training - Step 17390: {'lr': 0.00038354198228731414, 'samples': 8904192, 'steps': 17390, 'loss/train': 2.625652313232422} 02/25/2022 00:06:04 - INFO - codeparrot_training - Step 17391: {'lr': 0.0003835281495297267, 'samples': 8904704, 'steps': 17391, 'loss/train': 1.3525059223175049} 02/25/2022 00:06:09 - INFO - codeparrot_training - Step 17392: {'lr': 0.0003835143162001472, 'samples': 8905216, 'steps': 17392, 'loss/train': 1.672836184501648} 02/25/2022 00:06:13 - INFO - codeparrot_training - Step 17393: {'lr': 0.0003835004822986346, 'samples': 8905728, 'steps': 17393, 'loss/train': 1.8405547142028809} 02/25/2022 00:06:18 - INFO - codeparrot_training - Step 17394: {'lr': 0.00038348664782524846, 'samples': 8906240, 'steps': 17394, 'loss/train': 1.8644921779632568} 02/25/2022 00:06:22 - INFO - codeparrot_training - Step 17395: {'lr': 0.00038347281278004774, 'samples': 8906752, 'steps': 17395, 'loss/train': 2.2251172065734863} 02/25/2022 00:06:27 - INFO - codeparrot_training - Step 17396: {'lr': 0.0003834589771630921, 'samples': 8907264, 'steps': 17396, 'loss/train': 2.490211248397827} 02/25/2022 00:06:30 - INFO - codeparrot_training - Step 17397: {'lr': 0.0003834451409744404, 'samples': 8907776, 'steps': 17397, 'loss/train': 1.9458705186843872} 02/25/2022 00:06:37 - INFO - codeparrot_training - Step 17398: {'lr': 0.0003834313042141522, 'samples': 8908288, 'steps': 17398, 'loss/train': 2.2974541187286377} 02/25/2022 00:06:41 - INFO - codeparrot_training - Step 17399: {'lr': 0.0003834174668822865, 'samples': 8908800, 'steps': 17399, 'loss/train': 1.4726758003234863} 02/25/2022 00:06:46 - INFO - codeparrot_training - Step 17400: {'lr': 0.0003834036289789029, 'samples': 8909312, 'steps': 17400, 'loss/train': 2.2040772438049316} 02/25/2022 00:06:50 - INFO - codeparrot_training - Step 17401: {'lr': 0.0003833897905040604, 'samples': 8909824, 'steps': 17401, 'loss/train': 1.7959519624710083} 02/25/2022 00:06:55 - INFO - codeparrot_training - Step 17402: {'lr': 0.00038337595145781844, 'samples': 8910336, 'steps': 17402, 'loss/train': 1.9410135746002197} 02/25/2022 00:06:59 - INFO - codeparrot_training - Step 17403: {'lr': 0.00038336211184023634, 'samples': 8910848, 'steps': 17403, 'loss/train': 2.2259018421173096} 02/25/2022 00:07:04 - INFO - codeparrot_training - Step 17404: {'lr': 0.0003833482716513732, 'samples': 8911360, 'steps': 17404, 'loss/train': 2.1858718395233154} 02/25/2022 00:07:08 - INFO - codeparrot_training - Step 17405: {'lr': 0.0003833344308912885, 'samples': 8911872, 'steps': 17405, 'loss/train': 1.7019981145858765} 02/25/2022 00:07:13 - INFO - codeparrot_training - Step 17406: {'lr': 0.00038332058956004134, 'samples': 8912384, 'steps': 17406, 'loss/train': 1.642392873764038} 02/25/2022 00:07:17 - INFO - codeparrot_training - Step 17407: {'lr': 0.0003833067476576911, 'samples': 8912896, 'steps': 17407, 'loss/train': 0.7289206981658936} 02/25/2022 00:07:24 - INFO - codeparrot_training - Step 17408: {'lr': 0.0003832929051842972, 'samples': 8913408, 'steps': 17408, 'loss/train': 2.099308729171753} 02/25/2022 00:07:27 - INFO - codeparrot_training - Step 17409: {'lr': 0.0003832790621399187, 'samples': 8913920, 'steps': 17409, 'loss/train': 2.12369966506958} 02/25/2022 00:07:33 - INFO - codeparrot_training - Step 17410: {'lr': 0.00038326521852461505, 'samples': 8914432, 'steps': 17410, 'loss/train': 1.924633264541626} 02/25/2022 00:07:36 - INFO - codeparrot_training - Step 17411: {'lr': 0.0003832513743384456, 'samples': 8914944, 'steps': 17411, 'loss/train': 1.7784247398376465} 02/25/2022 00:07:42 - INFO - codeparrot_training - Step 17412: {'lr': 0.0003832375295814695, 'samples': 8915456, 'steps': 17412, 'loss/train': 1.3785922527313232} 02/25/2022 00:07:45 - INFO - codeparrot_training - Step 17413: {'lr': 0.0003832236842537461, 'samples': 8915968, 'steps': 17413, 'loss/train': 2.2821872234344482} 02/25/2022 00:07:51 - INFO - codeparrot_training - Step 17414: {'lr': 0.0003832098383553347, 'samples': 8916480, 'steps': 17414, 'loss/train': 2.455449342727661} 02/25/2022 00:07:54 - INFO - codeparrot_training - Step 17415: {'lr': 0.00038319599188629485, 'samples': 8916992, 'steps': 17415, 'loss/train': 1.7601726055145264} 02/25/2022 00:08:00 - INFO - codeparrot_training - Step 17416: {'lr': 0.00038318214484668557, 'samples': 8917504, 'steps': 17416, 'loss/train': 1.5574798583984375} 02/25/2022 00:08:05 - INFO - codeparrot_training - Step 17417: {'lr': 0.0003831682972365662, 'samples': 8918016, 'steps': 17417, 'loss/train': 2.0122427940368652} 02/25/2022 00:08:09 - INFO - codeparrot_training - Step 17418: {'lr': 0.0003831544490559962, 'samples': 8918528, 'steps': 17418, 'loss/train': 1.056150197982788} 02/25/2022 00:08:14 - INFO - codeparrot_training - Step 17419: {'lr': 0.00038314060030503476, 'samples': 8919040, 'steps': 17419, 'loss/train': 2.1905977725982666} 02/25/2022 00:08:18 - INFO - codeparrot_training - Step 17420: {'lr': 0.00038312675098374136, 'samples': 8919552, 'steps': 17420, 'loss/train': 1.6523983478546143} 02/25/2022 00:08:24 - INFO - codeparrot_training - Step 17421: {'lr': 0.0003831129010921751, 'samples': 8920064, 'steps': 17421, 'loss/train': 2.0863749980926514} 02/25/2022 00:08:28 - INFO - codeparrot_training - Step 17422: {'lr': 0.0003830990506303956, 'samples': 8920576, 'steps': 17422, 'loss/train': 1.2508740425109863} 02/25/2022 00:08:31 - INFO - codeparrot_training - Step 17423: {'lr': 0.0003830851995984619, 'samples': 8921088, 'steps': 17423, 'loss/train': 0.18153704702854156} 02/25/2022 00:08:36 - INFO - codeparrot_training - Step 17424: {'lr': 0.0003830713479964335, 'samples': 8921600, 'steps': 17424, 'loss/train': 2.6745219230651855} 02/25/2022 00:08:42 - INFO - codeparrot_training - Step 17425: {'lr': 0.0003830574958243697, 'samples': 8922112, 'steps': 17425, 'loss/train': 1.4275271892547607} 02/25/2022 00:08:46 - INFO - codeparrot_training - Step 17426: {'lr': 0.00038304364308232986, 'samples': 8922624, 'steps': 17426, 'loss/train': 1.6835452318191528} 02/25/2022 00:08:49 - INFO - codeparrot_training - Step 17427: {'lr': 0.0003830297897703733, 'samples': 8923136, 'steps': 17427, 'loss/train': 2.2621264457702637} 02/25/2022 00:08:55 - INFO - codeparrot_training - Step 17428: {'lr': 0.0003830159358885593, 'samples': 8923648, 'steps': 17428, 'loss/train': 2.3754281997680664} 02/25/2022 00:09:00 - INFO - codeparrot_training - Step 17429: {'lr': 0.00038300208143694737, 'samples': 8924160, 'steps': 17429, 'loss/train': 4.128789901733398} 02/25/2022 00:09:04 - INFO - codeparrot_training - Step 17430: {'lr': 0.00038298822641559673, 'samples': 8924672, 'steps': 17430, 'loss/train': 2.1815717220306396} 02/25/2022 00:09:07 - INFO - codeparrot_training - Step 17431: {'lr': 0.0003829743708245667, 'samples': 8925184, 'steps': 17431, 'loss/train': 2.9297618865966797} 02/25/2022 00:09:13 - INFO - codeparrot_training - Step 17432: {'lr': 0.0003829605146639167, 'samples': 8925696, 'steps': 17432, 'loss/train': 2.4314677715301514} 02/25/2022 00:09:19 - INFO - codeparrot_training - Step 17433: {'lr': 0.0003829466579337061, 'samples': 8926208, 'steps': 17433, 'loss/train': 0.6870249509811401} 02/25/2022 00:09:22 - INFO - codeparrot_training - Step 17434: {'lr': 0.00038293280063399427, 'samples': 8926720, 'steps': 17434, 'loss/train': 0.9593690037727356} 02/25/2022 00:09:28 - INFO - codeparrot_training - Step 17435: {'lr': 0.00038291894276484053, 'samples': 8927232, 'steps': 17435, 'loss/train': 1.496435284614563} 02/25/2022 00:09:31 - INFO - codeparrot_training - Step 17436: {'lr': 0.0003829050843263041, 'samples': 8927744, 'steps': 17436, 'loss/train': 1.831271767616272} 02/25/2022 00:09:37 - INFO - codeparrot_training - Step 17437: {'lr': 0.0003828912253184446, 'samples': 8928256, 'steps': 17437, 'loss/train': 1.7486910820007324} 02/25/2022 00:09:40 - INFO - codeparrot_training - Step 17438: {'lr': 0.0003828773657413213, 'samples': 8928768, 'steps': 17438, 'loss/train': 2.1680681705474854} 02/25/2022 00:09:46 - INFO - codeparrot_training - Step 17439: {'lr': 0.0003828635055949935, 'samples': 8929280, 'steps': 17439, 'loss/train': 1.663074016571045} 02/25/2022 00:09:49 - INFO - codeparrot_training - Step 17440: {'lr': 0.0003828496448795207, 'samples': 8929792, 'steps': 17440, 'loss/train': 1.912155270576477} 02/25/2022 00:09:55 - INFO - codeparrot_training - Step 17441: {'lr': 0.0003828357835949622, 'samples': 8930304, 'steps': 17441, 'loss/train': 1.4286575317382812} 02/25/2022 00:09:58 - INFO - codeparrot_training - Step 17442: {'lr': 0.00038282192174137744, 'samples': 8930816, 'steps': 17442, 'loss/train': 1.6696593761444092} 02/25/2022 00:10:05 - INFO - codeparrot_training - Step 17443: {'lr': 0.00038280805931882557, 'samples': 8931328, 'steps': 17443, 'loss/train': 0.6226693391799927} 02/25/2022 00:10:08 - INFO - codeparrot_training - Step 17444: {'lr': 0.0003827941963273663, 'samples': 8931840, 'steps': 17444, 'loss/train': 1.0359865427017212} 02/25/2022 00:10:14 - INFO - codeparrot_training - Step 17445: {'lr': 0.00038278033276705875, 'samples': 8932352, 'steps': 17445, 'loss/train': 2.0010087490081787} 02/25/2022 00:10:17 - INFO - codeparrot_training - Step 17446: {'lr': 0.0003827664686379625, 'samples': 8932864, 'steps': 17446, 'loss/train': 2.163616180419922} 02/25/2022 00:10:23 - INFO - codeparrot_training - Step 17447: {'lr': 0.00038275260394013676, 'samples': 8933376, 'steps': 17447, 'loss/train': 1.1750496625900269} 02/25/2022 00:10:26 - INFO - codeparrot_training - Step 17448: {'lr': 0.0003827387386736411, 'samples': 8933888, 'steps': 17448, 'loss/train': 0.9757341146469116} 02/25/2022 00:10:32 - INFO - codeparrot_training - Step 17449: {'lr': 0.0003827248728385349, 'samples': 8934400, 'steps': 17449, 'loss/train': 1.9917367696762085} 02/25/2022 00:10:35 - INFO - codeparrot_training - Step 17450: {'lr': 0.0003827110064348773, 'samples': 8934912, 'steps': 17450, 'loss/train': 2.0245180130004883} 02/25/2022 00:10:41 - INFO - codeparrot_training - Step 17451: {'lr': 0.000382697139462728, 'samples': 8935424, 'steps': 17451, 'loss/train': 1.805024266242981} 02/25/2022 00:10:44 - INFO - codeparrot_training - Step 17452: {'lr': 0.00038268327192214635, 'samples': 8935936, 'steps': 17452, 'loss/train': 2.43996524810791} 02/25/2022 00:10:51 - INFO - codeparrot_training - Step 17453: {'lr': 0.0003826694038131916, 'samples': 8936448, 'steps': 17453, 'loss/train': 1.3299962282180786} 02/25/2022 00:10:54 - INFO - codeparrot_training - Step 17454: {'lr': 0.00038265553513592334, 'samples': 8936960, 'steps': 17454, 'loss/train': 1.845405101776123} 02/25/2022 00:10:59 - INFO - codeparrot_training - Step 17455: {'lr': 0.00038264166589040084, 'samples': 8937472, 'steps': 17455, 'loss/train': 1.8641937971115112} 02/25/2022 00:11:03 - INFO - codeparrot_training - Step 17456: {'lr': 0.00038262779607668354, 'samples': 8937984, 'steps': 17456, 'loss/train': 3.083961009979248} 02/25/2022 00:11:09 - INFO - codeparrot_training - Step 17457: {'lr': 0.00038261392569483087, 'samples': 8938496, 'steps': 17457, 'loss/train': 3.5749051570892334} 02/25/2022 00:11:12 - INFO - codeparrot_training - Step 17458: {'lr': 0.0003826000547449023, 'samples': 8939008, 'steps': 17458, 'loss/train': 3.5318398475646973} 02/25/2022 00:11:18 - INFO - codeparrot_training - Step 17459: {'lr': 0.0003825861832269571, 'samples': 8939520, 'steps': 17459, 'loss/train': 1.3561949729919434} 02/25/2022 00:11:21 - INFO - codeparrot_training - Step 17460: {'lr': 0.00038257231114105495, 'samples': 8940032, 'steps': 17460, 'loss/train': 2.32259202003479} 02/25/2022 00:11:27 - INFO - codeparrot_training - Step 17461: {'lr': 0.00038255843848725504, 'samples': 8940544, 'steps': 17461, 'loss/train': 2.3007588386535645} 02/25/2022 00:11:30 - INFO - codeparrot_training - Step 17462: {'lr': 0.0003825445652656169, 'samples': 8941056, 'steps': 17462, 'loss/train': 6.096596717834473} 02/25/2022 00:11:36 - INFO - codeparrot_training - Step 17463: {'lr': 0.00038253069147619977, 'samples': 8941568, 'steps': 17463, 'loss/train': 0.38452979922294617} 02/25/2022 00:11:39 - INFO - codeparrot_training - Step 17464: {'lr': 0.00038251681711906345, 'samples': 8942080, 'steps': 17464, 'loss/train': 1.534771203994751} 02/25/2022 00:11:45 - INFO - codeparrot_training - Step 17465: {'lr': 0.00038250294219426706, 'samples': 8942592, 'steps': 17465, 'loss/train': 1.817731261253357} 02/25/2022 00:11:48 - INFO - codeparrot_training - Step 17466: {'lr': 0.00038248906670187017, 'samples': 8943104, 'steps': 17466, 'loss/train': 2.0167102813720703} 02/25/2022 00:11:53 - INFO - codeparrot_training - Step 17467: {'lr': 0.00038247519064193216, 'samples': 8943616, 'steps': 17467, 'loss/train': 2.3912782669067383} 02/25/2022 00:11:57 - INFO - codeparrot_training - Step 17468: {'lr': 0.0003824613140145125, 'samples': 8944128, 'steps': 17468, 'loss/train': 1.724741816520691} 02/25/2022 00:12:03 - INFO - codeparrot_training - Step 17469: {'lr': 0.00038244743681967066, 'samples': 8944640, 'steps': 17469, 'loss/train': 2.1507270336151123} 02/25/2022 00:12:07 - INFO - codeparrot_training - Step 17470: {'lr': 0.000382433559057466, 'samples': 8945152, 'steps': 17470, 'loss/train': 2.123882532119751} 02/25/2022 00:12:12 - INFO - codeparrot_training - Step 17471: {'lr': 0.00038241968072795805, 'samples': 8945664, 'steps': 17471, 'loss/train': 1.9180289506912231} 02/25/2022 00:12:16 - INFO - codeparrot_training - Step 17472: {'lr': 0.00038240580183120624, 'samples': 8946176, 'steps': 17472, 'loss/train': 2.4886069297790527} 02/25/2022 00:12:21 - INFO - codeparrot_training - Step 17473: {'lr': 0.0003823919223672701, 'samples': 8946688, 'steps': 17473, 'loss/train': 1.4735954999923706} 02/25/2022 00:12:25 - INFO - codeparrot_training - Step 17474: {'lr': 0.00038237804233620887, 'samples': 8947200, 'steps': 17474, 'loss/train': 0.1073988825082779} 02/25/2022 00:12:30 - INFO - codeparrot_training - Step 17475: {'lr': 0.0003823641617380823, 'samples': 8947712, 'steps': 17475, 'loss/train': 2.577849864959717} 02/25/2022 00:12:34 - INFO - codeparrot_training - Step 17476: {'lr': 0.00038235028057294953, 'samples': 8948224, 'steps': 17476, 'loss/train': 2.058762788772583} 02/25/2022 00:12:39 - INFO - codeparrot_training - Step 17477: {'lr': 0.0003823363988408703, 'samples': 8948736, 'steps': 17477, 'loss/train': 2.0384023189544678} 02/25/2022 00:12:43 - INFO - codeparrot_training - Step 17478: {'lr': 0.00038232251654190386, 'samples': 8949248, 'steps': 17478, 'loss/train': 0.9737243056297302} 02/25/2022 00:12:49 - INFO - codeparrot_training - Step 17479: {'lr': 0.0003823086336761099, 'samples': 8949760, 'steps': 17479, 'loss/train': 0.8889243602752686} 02/25/2022 00:12:52 - INFO - codeparrot_training - Step 17480: {'lr': 0.00038229475024354766, 'samples': 8950272, 'steps': 17480, 'loss/train': 2.84980845451355} 02/25/2022 00:12:58 - INFO - codeparrot_training - Step 17481: {'lr': 0.00038228086624427675, 'samples': 8950784, 'steps': 17481, 'loss/train': 2.54441499710083} 02/25/2022 00:13:01 - INFO - codeparrot_training - Step 17482: {'lr': 0.0003822669816783566, 'samples': 8951296, 'steps': 17482, 'loss/train': 2.1469645500183105} 02/25/2022 00:13:07 - INFO - codeparrot_training - Step 17483: {'lr': 0.0003822530965458467, 'samples': 8951808, 'steps': 17483, 'loss/train': 2.187218427658081} 02/25/2022 00:13:10 - INFO - codeparrot_training - Step 17484: {'lr': 0.0003822392108468066, 'samples': 8952320, 'steps': 17484, 'loss/train': 1.8644400835037231} 02/25/2022 00:13:16 - INFO - codeparrot_training - Step 17485: {'lr': 0.00038222532458129563, 'samples': 8952832, 'steps': 17485, 'loss/train': 1.0419584512710571} 02/25/2022 00:13:19 - INFO - codeparrot_training - Step 17486: {'lr': 0.0003822114377493734, 'samples': 8953344, 'steps': 17486, 'loss/train': 1.8946095705032349} 02/25/2022 00:13:25 - INFO - codeparrot_training - Step 17487: {'lr': 0.0003821975503510993, 'samples': 8953856, 'steps': 17487, 'loss/train': 2.175102710723877} 02/25/2022 00:13:28 - INFO - codeparrot_training - Step 17488: {'lr': 0.0003821836623865329, 'samples': 8954368, 'steps': 17488, 'loss/train': 2.0778934955596924} 02/25/2022 00:13:35 - INFO - codeparrot_training - Step 17489: {'lr': 0.0003821697738557337, 'samples': 8954880, 'steps': 17489, 'loss/train': 1.0143626928329468} 02/25/2022 00:13:38 - INFO - codeparrot_training - Step 17490: {'lr': 0.00038215588475876117, 'samples': 8955392, 'steps': 17490, 'loss/train': 0.8945921063423157} 02/25/2022 00:13:44 - INFO - codeparrot_training - Step 17491: {'lr': 0.0003821419950956747, 'samples': 8955904, 'steps': 17491, 'loss/train': 2.351945638656616} 02/25/2022 00:13:48 - INFO - codeparrot_training - Step 17492: {'lr': 0.00038212810486653394, 'samples': 8956416, 'steps': 17492, 'loss/train': 1.6855456829071045} 02/25/2022 00:13:53 - INFO - codeparrot_training - Step 17493: {'lr': 0.0003821142140713983, 'samples': 8956928, 'steps': 17493, 'loss/train': 2.2223989963531494} 02/25/2022 00:13:57 - INFO - codeparrot_training - Step 17494: {'lr': 0.0003821003227103274, 'samples': 8957440, 'steps': 17494, 'loss/train': 2.366349697113037} 02/25/2022 00:14:02 - INFO - codeparrot_training - Step 17495: {'lr': 0.00038208643078338055, 'samples': 8957952, 'steps': 17495, 'loss/train': 1.8043476343154907} 02/25/2022 00:14:06 - INFO - codeparrot_training - Step 17496: {'lr': 0.0003820725382906175, 'samples': 8958464, 'steps': 17496, 'loss/train': 0.7203257083892822} 02/25/2022 00:14:11 - INFO - codeparrot_training - Step 17497: {'lr': 0.0003820586452320975, 'samples': 8958976, 'steps': 17497, 'loss/train': 1.6372029781341553} 02/25/2022 00:14:15 - INFO - codeparrot_training - Step 17498: {'lr': 0.0003820447516078803, 'samples': 8959488, 'steps': 17498, 'loss/train': 1.2810819149017334} 02/25/2022 00:14:21 - INFO - codeparrot_training - Step 17499: {'lr': 0.0003820308574180253, 'samples': 8960000, 'steps': 17499, 'loss/train': 1.4078881740570068} 02/25/2022 00:14:25 - INFO - codeparrot_training - Step 17500: {'lr': 0.000382016962662592, 'samples': 8960512, 'steps': 17500, 'loss/train': 1.9532833099365234} 02/25/2022 00:14:30 - INFO - codeparrot_training - Step 17501: {'lr': 0.0003820030673416399, 'samples': 8961024, 'steps': 17501, 'loss/train': 2.0017523765563965} 02/25/2022 00:14:34 - INFO - codeparrot_training - Step 17502: {'lr': 0.0003819891714552287, 'samples': 8961536, 'steps': 17502, 'loss/train': 2.565995693206787} 02/25/2022 00:14:39 - INFO - codeparrot_training - Step 17503: {'lr': 0.00038197527500341777, 'samples': 8962048, 'steps': 17503, 'loss/train': 1.6449755430221558} 02/25/2022 00:14:43 - INFO - codeparrot_training - Step 17504: {'lr': 0.00038196137798626663, 'samples': 8962560, 'steps': 17504, 'loss/train': 1.8359822034835815} 02/25/2022 00:14:48 - INFO - codeparrot_training - Step 17505: {'lr': 0.00038194748040383487, 'samples': 8963072, 'steps': 17505, 'loss/train': 2.4941248893737793} 02/25/2022 00:14:52 - INFO - codeparrot_training - Step 17506: {'lr': 0.00038193358225618195, 'samples': 8963584, 'steps': 17506, 'loss/train': 3.04447865486145} 02/25/2022 00:14:57 - INFO - codeparrot_training - Step 17507: {'lr': 0.0003819196835433675, 'samples': 8964096, 'steps': 17507, 'loss/train': 1.7993943691253662} 02/25/2022 00:15:01 - INFO - codeparrot_training - Step 17508: {'lr': 0.000381905784265451, 'samples': 8964608, 'steps': 17508, 'loss/train': 1.528943657875061} 02/25/2022 00:15:08 - INFO - codeparrot_training - Step 17509: {'lr': 0.000381891884422492, 'samples': 8965120, 'steps': 17509, 'loss/train': 1.7222402095794678} 02/25/2022 00:15:11 - INFO - codeparrot_training - Step 17510: {'lr': 0.0003818779840145501, 'samples': 8965632, 'steps': 17510, 'loss/train': 2.2193057537078857} 02/25/2022 00:15:16 - INFO - codeparrot_training - Step 17511: {'lr': 0.00038186408304168474, 'samples': 8966144, 'steps': 17511, 'loss/train': 1.5934960842132568} 02/25/2022 00:15:20 - INFO - codeparrot_training - Step 17512: {'lr': 0.00038185018150395557, 'samples': 8966656, 'steps': 17512, 'loss/train': 1.2100094556808472} 02/25/2022 00:15:25 - INFO - codeparrot_training - Step 17513: {'lr': 0.000381836279401422, 'samples': 8967168, 'steps': 17513, 'loss/train': 2.0431087017059326} 02/25/2022 00:15:29 - INFO - codeparrot_training - Step 17514: {'lr': 0.00038182237673414375, 'samples': 8967680, 'steps': 17514, 'loss/train': 1.8947263956069946} 02/25/2022 00:15:35 - INFO - codeparrot_training - Step 17515: {'lr': 0.0003818084735021803, 'samples': 8968192, 'steps': 17515, 'loss/train': 1.9881848096847534} 02/25/2022 00:15:38 - INFO - codeparrot_training - Step 17516: {'lr': 0.00038179456970559116, 'samples': 8968704, 'steps': 17516, 'loss/train': 1.73069167137146} 02/25/2022 00:15:44 - INFO - codeparrot_training - Step 17517: {'lr': 0.00038178066534443587, 'samples': 8969216, 'steps': 17517, 'loss/train': 1.9151536226272583} 02/25/2022 00:15:47 - INFO - codeparrot_training - Step 17518: {'lr': 0.00038176676041877424, 'samples': 8969728, 'steps': 17518, 'loss/train': 2.415066957473755} 02/25/2022 00:15:53 - INFO - codeparrot_training - Step 17519: {'lr': 0.0003817528549286655, 'samples': 8970240, 'steps': 17519, 'loss/train': 2.235966920852661} 02/25/2022 00:15:56 - INFO - codeparrot_training - Step 17520: {'lr': 0.00038173894887416946, 'samples': 8970752, 'steps': 17520, 'loss/train': 1.8851059675216675} 02/25/2022 00:16:02 - INFO - codeparrot_training - Step 17521: {'lr': 0.0003817250422553455, 'samples': 8971264, 'steps': 17521, 'loss/train': 2.310234546661377} 02/25/2022 00:16:05 - INFO - codeparrot_training - Step 17522: {'lr': 0.0003817111350722533, 'samples': 8971776, 'steps': 17522, 'loss/train': 2.544466495513916} 02/25/2022 00:16:11 - INFO - codeparrot_training - Step 17523: {'lr': 0.0003816972273249525, 'samples': 8972288, 'steps': 17523, 'loss/train': 3.050013303756714} 02/25/2022 00:16:14 - INFO - codeparrot_training - Step 17524: {'lr': 0.00038168331901350253, 'samples': 8972800, 'steps': 17524, 'loss/train': 1.6691980361938477} 02/25/2022 00:16:20 - INFO - codeparrot_training - Step 17525: {'lr': 0.0003816694101379631, 'samples': 8973312, 'steps': 17525, 'loss/train': 1.4608827829360962} 02/25/2022 00:16:24 - INFO - codeparrot_training - Step 17526: {'lr': 0.0003816555006983936, 'samples': 8973824, 'steps': 17526, 'loss/train': 1.6421570777893066} 02/25/2022 00:16:29 - INFO - codeparrot_training - Step 17527: {'lr': 0.0003816415906948538, 'samples': 8974336, 'steps': 17527, 'loss/train': 2.1995041370391846} 02/25/2022 00:16:33 - INFO - codeparrot_training - Step 17528: {'lr': 0.00038162768012740323, 'samples': 8974848, 'steps': 17528, 'loss/train': 1.6848971843719482} 02/25/2022 00:16:38 - INFO - codeparrot_training - Step 17529: {'lr': 0.00038161376899610154, 'samples': 8975360, 'steps': 17529, 'loss/train': 3.63096284866333} 02/25/2022 00:16:42 - INFO - codeparrot_training - Step 17530: {'lr': 0.0003815998573010082, 'samples': 8975872, 'steps': 17530, 'loss/train': 2.397806167602539} 02/25/2022 00:16:47 - INFO - codeparrot_training - Step 17531: {'lr': 0.0003815859450421829, 'samples': 8976384, 'steps': 17531, 'loss/train': 3.6741068363189697} 02/25/2022 00:16:51 - INFO - codeparrot_training - Step 17532: {'lr': 0.00038157203221968514, 'samples': 8976896, 'steps': 17532, 'loss/train': 2.158233165740967} 02/25/2022 00:16:57 - INFO - codeparrot_training - Step 17533: {'lr': 0.00038155811883357454, 'samples': 8977408, 'steps': 17533, 'loss/train': 1.203019380569458} 02/25/2022 00:17:00 - INFO - codeparrot_training - Step 17534: {'lr': 0.0003815442048839108, 'samples': 8977920, 'steps': 17534, 'loss/train': 1.6776306629180908} 02/25/2022 00:17:06 - INFO - codeparrot_training - Step 17535: {'lr': 0.0003815302903707534, 'samples': 8978432, 'steps': 17535, 'loss/train': 2.1549835205078125} 02/25/2022 00:17:10 - INFO - codeparrot_training - Step 17536: {'lr': 0.0003815163752941621, 'samples': 8978944, 'steps': 17536, 'loss/train': 1.0324739217758179} 02/25/2022 00:17:15 - INFO - codeparrot_training - Step 17537: {'lr': 0.00038150245965419636, 'samples': 8979456, 'steps': 17537, 'loss/train': 1.4928556680679321} 02/25/2022 00:17:19 - INFO - codeparrot_training - Step 17538: {'lr': 0.0003814885434509158, 'samples': 8979968, 'steps': 17538, 'loss/train': 1.694128394126892} 02/25/2022 00:17:24 - INFO - codeparrot_training - Step 17539: {'lr': 0.0003814746266843801, 'samples': 8980480, 'steps': 17539, 'loss/train': 2.3358724117279053} 02/25/2022 00:17:28 - INFO - codeparrot_training - Step 17540: {'lr': 0.0003814607093546489, 'samples': 8980992, 'steps': 17540, 'loss/train': 0.5446189641952515} 02/25/2022 00:17:33 - INFO - codeparrot_training - Step 17541: {'lr': 0.00038144679146178166, 'samples': 8981504, 'steps': 17541, 'loss/train': 2.15185546875} 02/25/2022 00:17:37 - INFO - codeparrot_training - Step 17542: {'lr': 0.00038143287300583816, 'samples': 8982016, 'steps': 17542, 'loss/train': 1.3598071336746216} 02/25/2022 00:17:42 - INFO - codeparrot_training - Step 17543: {'lr': 0.00038141895398687806, 'samples': 8982528, 'steps': 17543, 'loss/train': 2.9764785766601562} 02/25/2022 00:17:46 - INFO - codeparrot_training - Step 17544: {'lr': 0.0003814050344049608, 'samples': 8983040, 'steps': 17544, 'loss/train': 1.0933226346969604} 02/25/2022 00:17:52 - INFO - codeparrot_training - Step 17545: {'lr': 0.00038139111426014607, 'samples': 8983552, 'steps': 17545, 'loss/train': 1.916902780532837} 02/25/2022 00:17:55 - INFO - codeparrot_training - Step 17546: {'lr': 0.00038137719355249355, 'samples': 8984064, 'steps': 17546, 'loss/train': 1.3877387046813965} 02/25/2022 00:18:01 - INFO - codeparrot_training - Step 17547: {'lr': 0.00038136327228206285, 'samples': 8984576, 'steps': 17547, 'loss/train': 1.461612343788147} 02/25/2022 00:18:04 - INFO - codeparrot_training - Step 17548: {'lr': 0.0003813493504489136, 'samples': 8985088, 'steps': 17548, 'loss/train': 2.3623855113983154} 02/25/2022 00:18:10 - INFO - codeparrot_training - Step 17549: {'lr': 0.0003813354280531055, 'samples': 8985600, 'steps': 17549, 'loss/train': 1.20237135887146} 02/25/2022 00:18:13 - INFO - codeparrot_training - Step 17550: {'lr': 0.00038132150509469806, 'samples': 8986112, 'steps': 17550, 'loss/train': 2.5042812824249268} 02/25/2022 00:18:19 - INFO - codeparrot_training - Step 17551: {'lr': 0.000381307581573751, 'samples': 8986624, 'steps': 17551, 'loss/train': 1.842660665512085} 02/25/2022 00:18:22 - INFO - codeparrot_training - Step 17552: {'lr': 0.00038129365749032395, 'samples': 8987136, 'steps': 17552, 'loss/train': 3.0439553260803223} 02/25/2022 00:18:28 - INFO - codeparrot_training - Step 17553: {'lr': 0.0003812797328444766, 'samples': 8987648, 'steps': 17553, 'loss/train': 2.318154811859131} 02/25/2022 00:18:31 - INFO - codeparrot_training - Step 17554: {'lr': 0.0003812658076362685, 'samples': 8988160, 'steps': 17554, 'loss/train': 2.430816411972046} 02/25/2022 00:18:37 - INFO - codeparrot_training - Step 17555: {'lr': 0.00038125188186575944, 'samples': 8988672, 'steps': 17555, 'loss/train': 2.5937442779541016} 02/25/2022 00:18:40 - INFO - codeparrot_training - Step 17556: {'lr': 0.00038123795553300893, 'samples': 8989184, 'steps': 17556, 'loss/train': 1.7834750413894653} 02/25/2022 00:18:46 - INFO - codeparrot_training - Step 17557: {'lr': 0.0003812240286380767, 'samples': 8989696, 'steps': 17557, 'loss/train': 1.567370891571045} 02/25/2022 00:18:50 - INFO - codeparrot_training - Step 17558: {'lr': 0.0003812101011810224, 'samples': 8990208, 'steps': 17558, 'loss/train': 2.460343360900879} 02/25/2022 00:18:58 - INFO - codeparrot_training - Step 17559: {'lr': 0.0003811961731619057, 'samples': 8990720, 'steps': 17559, 'loss/train': 2.5448851585388184} 02/25/2022 00:19:01 - INFO - codeparrot_training - Step 17560: {'lr': 0.0003811822445807863, 'samples': 8991232, 'steps': 17560, 'loss/train': 1.2904130220413208} 02/25/2022 00:19:07 - INFO - codeparrot_training - Step 17561: {'lr': 0.00038116831543772377, 'samples': 8991744, 'steps': 17561, 'loss/train': 2.130690097808838} 02/25/2022 00:19:10 - INFO - codeparrot_training - Step 17562: {'lr': 0.00038115438573277784, 'samples': 8992256, 'steps': 17562, 'loss/train': 2.2204508781433105} 02/25/2022 00:19:16 - INFO - codeparrot_training - Step 17563: {'lr': 0.0003811404554660082, 'samples': 8992768, 'steps': 17563, 'loss/train': 2.2739343643188477} 02/25/2022 00:19:19 - INFO - codeparrot_training - Step 17564: {'lr': 0.00038112652463747444, 'samples': 8993280, 'steps': 17564, 'loss/train': 2.316910982131958} 02/25/2022 00:19:25 - INFO - codeparrot_training - Step 17565: {'lr': 0.00038111259324723624, 'samples': 8993792, 'steps': 17565, 'loss/train': 1.7901859283447266} 02/25/2022 00:19:28 - INFO - codeparrot_training - Step 17566: {'lr': 0.0003810986612953534, 'samples': 8994304, 'steps': 17566, 'loss/train': 1.536866307258606} 02/25/2022 00:19:34 - INFO - codeparrot_training - Step 17567: {'lr': 0.0003810847287818855, 'samples': 8994816, 'steps': 17567, 'loss/train': 2.2681477069854736} 02/25/2022 00:19:37 - INFO - codeparrot_training - Step 17568: {'lr': 0.0003810707957068923, 'samples': 8995328, 'steps': 17568, 'loss/train': 2.4871232509613037} 02/25/2022 00:19:41 - INFO - codeparrot_training - Step 17569: {'lr': 0.0003810568620704334, 'samples': 8995840, 'steps': 17569, 'loss/train': 2.3494536876678467} 02/25/2022 00:19:48 - INFO - codeparrot_training - Step 17570: {'lr': 0.00038104292787256844, 'samples': 8996352, 'steps': 17570, 'loss/train': 2.346104860305786} 02/25/2022 00:19:52 - INFO - codeparrot_training - Step 17571: {'lr': 0.0003810289931133573, 'samples': 8996864, 'steps': 17571, 'loss/train': 2.085747241973877} 02/25/2022 00:19:57 - INFO - codeparrot_training - Step 17572: {'lr': 0.0003810150577928595, 'samples': 8997376, 'steps': 17572, 'loss/train': 1.8231298923492432} 02/25/2022 00:20:03 - INFO - codeparrot_training - Step 17573: {'lr': 0.0003810011219111348, 'samples': 8997888, 'steps': 17573, 'loss/train': 2.469087600708008} 02/25/2022 00:20:06 - INFO - codeparrot_training - Step 17574: {'lr': 0.00038098718546824287, 'samples': 8998400, 'steps': 17574, 'loss/train': 1.6692852973937988} 02/25/2022 00:20:10 - INFO - codeparrot_training - Step 17575: {'lr': 0.00038097324846424354, 'samples': 8998912, 'steps': 17575, 'loss/train': 2.06733775138855} 02/25/2022 00:20:15 - INFO - codeparrot_training - Step 17576: {'lr': 0.0003809593108991962, 'samples': 8999424, 'steps': 17576, 'loss/train': 1.8917266130447388} 02/25/2022 00:20:19 - INFO - codeparrot_training - Step 17577: {'lr': 0.0003809453727731609, 'samples': 8999936, 'steps': 17577, 'loss/train': 1.5943111181259155} 02/25/2022 00:20:24 - INFO - codeparrot_training - Step 17578: {'lr': 0.00038093143408619726, 'samples': 9000448, 'steps': 17578, 'loss/train': 1.589393973350525} 02/25/2022 00:20:28 - INFO - codeparrot_training - Step 17579: {'lr': 0.0003809174948383648, 'samples': 9000960, 'steps': 17579, 'loss/train': 2.8796637058258057} 02/25/2022 00:20:36 - INFO - codeparrot_training - Step 17580: {'lr': 0.0003809035550297234, 'samples': 9001472, 'steps': 17580, 'loss/train': 1.3443183898925781} 02/25/2022 00:20:39 - INFO - codeparrot_training - Step 17581: {'lr': 0.00038088961466033276, 'samples': 9001984, 'steps': 17581, 'loss/train': 2.3200275897979736} 02/25/2022 00:20:45 - INFO - codeparrot_training - Step 17582: {'lr': 0.00038087567373025255, 'samples': 9002496, 'steps': 17582, 'loss/train': 2.038069486618042} 02/25/2022 00:20:48 - INFO - codeparrot_training - Step 17583: {'lr': 0.0003808617322395425, 'samples': 9003008, 'steps': 17583, 'loss/train': 1.5535337924957275} 02/25/2022 00:20:54 - INFO - codeparrot_training - Step 17584: {'lr': 0.00038084779018826245, 'samples': 9003520, 'steps': 17584, 'loss/train': 1.5001914501190186} 02/25/2022 00:20:57 - INFO - codeparrot_training - Step 17585: {'lr': 0.00038083384757647186, 'samples': 9004032, 'steps': 17585, 'loss/train': 2.75010085105896} 02/25/2022 00:21:03 - INFO - codeparrot_training - Step 17586: {'lr': 0.0003808199044042308, 'samples': 9004544, 'steps': 17586, 'loss/train': 2.0771737098693848} 02/25/2022 00:21:06 - INFO - codeparrot_training - Step 17587: {'lr': 0.00038080596067159865, 'samples': 9005056, 'steps': 17587, 'loss/train': 2.6117703914642334} 02/25/2022 00:21:12 - INFO - codeparrot_training - Step 17588: {'lr': 0.0003807920163786353, 'samples': 9005568, 'steps': 17588, 'loss/train': 1.8987400531768799} 02/25/2022 00:21:15 - INFO - codeparrot_training - Step 17589: {'lr': 0.0003807780715254006, 'samples': 9006080, 'steps': 17589, 'loss/train': 1.979446291923523} 02/25/2022 00:21:21 - INFO - codeparrot_training - Step 17590: {'lr': 0.000380764126111954, 'samples': 9006592, 'steps': 17590, 'loss/train': 2.4507508277893066} 02/25/2022 00:21:24 - INFO - codeparrot_training - Step 17591: {'lr': 0.0003807501801383555, 'samples': 9007104, 'steps': 17591, 'loss/train': 2.1036577224731445} 02/25/2022 00:21:32 - INFO - codeparrot_training - Step 17592: {'lr': 0.0003807362336046648, 'samples': 9007616, 'steps': 17592, 'loss/train': 2.216003656387329} 02/25/2022 00:21:35 - INFO - codeparrot_training - Step 17593: {'lr': 0.00038072228651094155, 'samples': 9008128, 'steps': 17593, 'loss/train': 1.699267029762268} 02/25/2022 00:21:41 - INFO - codeparrot_training - Step 17594: {'lr': 0.0003807083388572455, 'samples': 9008640, 'steps': 17594, 'loss/train': 2.1328063011169434} 02/25/2022 00:21:44 - INFO - codeparrot_training - Step 17595: {'lr': 0.0003806943906436364, 'samples': 9009152, 'steps': 17595, 'loss/train': 2.195122003555298} 02/25/2022 00:21:50 - INFO - codeparrot_training - Step 17596: {'lr': 0.0003806804418701741, 'samples': 9009664, 'steps': 17596, 'loss/train': 1.8058260679244995} 02/25/2022 00:21:53 - INFO - codeparrot_training - Step 17597: {'lr': 0.0003806664925369183, 'samples': 9010176, 'steps': 17597, 'loss/train': 1.5706242322921753} 02/25/2022 00:21:59 - INFO - codeparrot_training - Step 17598: {'lr': 0.0003806525426439287, 'samples': 9010688, 'steps': 17598, 'loss/train': 1.4891386032104492} 02/25/2022 00:22:02 - INFO - codeparrot_training - Step 17599: {'lr': 0.00038063859219126514, 'samples': 9011200, 'steps': 17599, 'loss/train': 1.4190083742141724} 02/25/2022 00:22:09 - INFO - codeparrot_training - Step 17600: {'lr': 0.0003806246411789872, 'samples': 9011712, 'steps': 17600, 'loss/train': 1.6553376913070679} 02/25/2022 00:22:12 - INFO - codeparrot_training - Step 17601: {'lr': 0.00038061068960715494, 'samples': 9012224, 'steps': 17601, 'loss/train': 1.555208683013916} 02/25/2022 00:22:16 - INFO - codeparrot_training - Step 17602: {'lr': 0.00038059673747582783, 'samples': 9012736, 'steps': 17602, 'loss/train': 0.843410849571228} 02/25/2022 00:22:21 - INFO - codeparrot_training - Step 17603: {'lr': 0.00038058278478506584, 'samples': 9013248, 'steps': 17603, 'loss/train': 0.4604855477809906} 02/25/2022 00:22:25 - INFO - codeparrot_training - Step 17604: {'lr': 0.0003805688315349286, 'samples': 9013760, 'steps': 17604, 'loss/train': 1.3503074645996094} 02/25/2022 00:22:30 - INFO - codeparrot_training - Step 17605: {'lr': 0.00038055487772547603, 'samples': 9014272, 'steps': 17605, 'loss/train': 1.3851650953292847} 02/25/2022 00:22:34 - INFO - codeparrot_training - Step 17606: {'lr': 0.00038054092335676774, 'samples': 9014784, 'steps': 17606, 'loss/train': 2.7930288314819336} 02/25/2022 00:22:41 - INFO - codeparrot_training - Step 17607: {'lr': 0.00038052696842886364, 'samples': 9015296, 'steps': 17607, 'loss/train': 1.7360361814498901} 02/25/2022 00:22:45 - INFO - codeparrot_training - Step 17608: {'lr': 0.0003805130129418235, 'samples': 9015808, 'steps': 17608, 'loss/train': 1.829670786857605} 02/25/2022 00:22:50 - INFO - codeparrot_training - Step 17609: {'lr': 0.00038049905689570697, 'samples': 9016320, 'steps': 17609, 'loss/train': 2.2178890705108643} 02/25/2022 00:22:54 - INFO - codeparrot_training - Step 17610: {'lr': 0.00038048510029057393, 'samples': 9016832, 'steps': 17610, 'loss/train': 1.4339778423309326} 02/25/2022 00:22:59 - INFO - codeparrot_training - Step 17611: {'lr': 0.00038047114312648414, 'samples': 9017344, 'steps': 17611, 'loss/train': 2.061182975769043} 02/25/2022 00:23:05 - INFO - codeparrot_training - Step 17612: {'lr': 0.0003804571854034975, 'samples': 9017856, 'steps': 17612, 'loss/train': 1.550879716873169} 02/25/2022 00:23:08 - INFO - codeparrot_training - Step 17613: {'lr': 0.0003804432271216736, 'samples': 9018368, 'steps': 17613, 'loss/train': 1.9707456827163696} 02/25/2022 00:23:12 - INFO - codeparrot_training - Step 17614: {'lr': 0.0003804292682810724, 'samples': 9018880, 'steps': 17614, 'loss/train': 2.291849136352539} 02/25/2022 00:23:17 - INFO - codeparrot_training - Step 17615: {'lr': 0.00038041530888175356, 'samples': 9019392, 'steps': 17615, 'loss/train': 2.577650785446167} 02/25/2022 00:23:21 - INFO - codeparrot_training - Step 17616: {'lr': 0.00038040134892377696, 'samples': 9019904, 'steps': 17616, 'loss/train': 2.0950605869293213} 02/25/2022 00:23:28 - INFO - codeparrot_training - Step 17617: {'lr': 0.00038038738840720244, 'samples': 9020416, 'steps': 17617, 'loss/train': 2.4005303382873535} 02/25/2022 00:23:32 - INFO - codeparrot_training - Step 17618: {'lr': 0.0003803734273320897, 'samples': 9020928, 'steps': 17618, 'loss/train': 1.9828367233276367} 02/25/2022 00:23:37 - INFO - codeparrot_training - Step 17619: {'lr': 0.0003803594656984986, 'samples': 9021440, 'steps': 17619, 'loss/train': 2.0146870613098145} 02/25/2022 00:23:41 - INFO - codeparrot_training - Step 17620: {'lr': 0.000380345503506489, 'samples': 9021952, 'steps': 17620, 'loss/train': 2.2994463443756104} 02/25/2022 00:23:46 - INFO - codeparrot_training - Step 17621: {'lr': 0.00038033154075612063, 'samples': 9022464, 'steps': 17621, 'loss/train': 2.2996325492858887} 02/25/2022 00:23:50 - INFO - codeparrot_training - Step 17622: {'lr': 0.00038031757744745327, 'samples': 9022976, 'steps': 17622, 'loss/train': 2.190804958343506} 02/25/2022 00:23:55 - INFO - codeparrot_training - Step 17623: {'lr': 0.0003803036135805469, 'samples': 9023488, 'steps': 17623, 'loss/train': 2.294494152069092} 02/25/2022 00:23:59 - INFO - codeparrot_training - Step 17624: {'lr': 0.00038028964915546107, 'samples': 9024000, 'steps': 17624, 'loss/train': 1.5395901203155518} 02/25/2022 00:24:04 - INFO - codeparrot_training - Step 17625: {'lr': 0.00038027568417225586, 'samples': 9024512, 'steps': 17625, 'loss/train': 1.272788643836975} 02/25/2022 00:24:08 - INFO - codeparrot_training - Step 17626: {'lr': 0.00038026171863099093, 'samples': 9025024, 'steps': 17626, 'loss/train': 1.6162992715835571} 02/25/2022 00:24:15 - INFO - codeparrot_training - Step 17627: {'lr': 0.0003802477525317263, 'samples': 9025536, 'steps': 17627, 'loss/train': 2.5885162353515625} 02/25/2022 00:24:19 - INFO - codeparrot_training - Step 17628: {'lr': 0.00038023378587452144, 'samples': 9026048, 'steps': 17628, 'loss/train': 0.10160930454730988} 02/25/2022 00:24:24 - INFO - codeparrot_training - Step 17629: {'lr': 0.0003802198186594366, 'samples': 9026560, 'steps': 17629, 'loss/train': 2.385128974914551} 02/25/2022 00:24:27 - INFO - codeparrot_training - Step 17630: {'lr': 0.00038020585088653126, 'samples': 9027072, 'steps': 17630, 'loss/train': 1.407119631767273} 02/25/2022 00:24:33 - INFO - codeparrot_training - Step 17631: {'lr': 0.00038019188255586546, 'samples': 9027584, 'steps': 17631, 'loss/train': 1.7080103158950806} 02/25/2022 00:24:39 - INFO - codeparrot_training - Step 17632: {'lr': 0.00038017791366749896, 'samples': 9028096, 'steps': 17632, 'loss/train': 1.6353788375854492} 02/25/2022 00:24:42 - INFO - codeparrot_training - Step 17633: {'lr': 0.0003801639442214916, 'samples': 9028608, 'steps': 17633, 'loss/train': 1.9614137411117554} 02/25/2022 00:24:46 - INFO - codeparrot_training - Step 17634: {'lr': 0.0003801499742179033, 'samples': 9029120, 'steps': 17634, 'loss/train': 2.6898295879364014} 02/25/2022 00:24:51 - INFO - codeparrot_training - Step 17635: {'lr': 0.0003801360036567938, 'samples': 9029632, 'steps': 17635, 'loss/train': 1.8056670427322388} 02/25/2022 00:24:55 - INFO - codeparrot_training - Step 17636: {'lr': 0.000380122032538223, 'samples': 9030144, 'steps': 17636, 'loss/train': 2.6003236770629883} 02/25/2022 00:25:00 - INFO - codeparrot_training - Step 17637: {'lr': 0.0003801080608622507, 'samples': 9030656, 'steps': 17637, 'loss/train': 0.978050172328949} 02/25/2022 00:25:04 - INFO - codeparrot_training - Step 17638: {'lr': 0.0003800940886289368, 'samples': 9031168, 'steps': 17638, 'loss/train': 1.6452912092208862} 02/25/2022 00:25:11 - INFO - codeparrot_training - Step 17639: {'lr': 0.0003800801158383411, 'samples': 9031680, 'steps': 17639, 'loss/train': 1.1491564512252808} 02/25/2022 00:25:15 - INFO - codeparrot_training - Step 17640: {'lr': 0.00038006614249052353, 'samples': 9032192, 'steps': 17640, 'loss/train': 2.4591915607452393} 02/25/2022 00:25:20 - INFO - codeparrot_training - Step 17641: {'lr': 0.0003800521685855439, 'samples': 9032704, 'steps': 17641, 'loss/train': 0.5061256289482117} 02/25/2022 00:25:24 - INFO - codeparrot_training - Step 17642: {'lr': 0.000380038194123462, 'samples': 9033216, 'steps': 17642, 'loss/train': 2.544386148452759} 02/25/2022 00:25:29 - INFO - codeparrot_training - Step 17643: {'lr': 0.0003800242191043379, 'samples': 9033728, 'steps': 17643, 'loss/train': 3.2017595767974854} 02/25/2022 00:25:33 - INFO - codeparrot_training - Step 17644: {'lr': 0.00038001024352823123, 'samples': 9034240, 'steps': 17644, 'loss/train': 2.2780935764312744} 02/25/2022 00:25:38 - INFO - codeparrot_training - Step 17645: {'lr': 0.00037999626739520197, 'samples': 9034752, 'steps': 17645, 'loss/train': 2.3560986518859863} 02/25/2022 00:25:42 - INFO - codeparrot_training - Step 17646: {'lr': 0.00037998229070531, 'samples': 9035264, 'steps': 17646, 'loss/train': 1.8259999752044678} 02/25/2022 00:25:47 - INFO - codeparrot_training - Step 17647: {'lr': 0.0003799683134586152, 'samples': 9035776, 'steps': 17647, 'loss/train': 2.271980047225952} 02/25/2022 00:25:51 - INFO - codeparrot_training - Step 17648: {'lr': 0.0003799543356551773, 'samples': 9036288, 'steps': 17648, 'loss/train': 1.0147898197174072} 02/25/2022 00:25:56 - INFO - codeparrot_training - Step 17649: {'lr': 0.0003799403572950565, 'samples': 9036800, 'steps': 17649, 'loss/train': 1.4735788106918335} 02/25/2022 00:26:00 - INFO - codeparrot_training - Step 17650: {'lr': 0.00037992637837831235, 'samples': 9037312, 'steps': 17650, 'loss/train': 1.616865873336792} 02/25/2022 00:26:05 - INFO - codeparrot_training - Step 17651: {'lr': 0.00037991239890500483, 'samples': 9037824, 'steps': 17651, 'loss/train': 0.1368354707956314} 02/25/2022 00:26:09 - INFO - codeparrot_training - Step 17652: {'lr': 0.00037989841887519385, 'samples': 9038336, 'steps': 17652, 'loss/train': 1.9762451648712158} 02/25/2022 00:26:14 - INFO - codeparrot_training - Step 17653: {'lr': 0.00037988443828893936, 'samples': 9038848, 'steps': 17653, 'loss/train': 2.2848517894744873} 02/25/2022 00:26:18 - INFO - codeparrot_training - Step 17654: {'lr': 0.0003798704571463011, 'samples': 9039360, 'steps': 17654, 'loss/train': 2.704603433609009} 02/25/2022 00:26:25 - INFO - codeparrot_training - Step 17655: {'lr': 0.00037985647544733903, 'samples': 9039872, 'steps': 17655, 'loss/train': 2.0447399616241455} 02/25/2022 00:26:29 - INFO - codeparrot_training - Step 17656: {'lr': 0.0003798424931921131, 'samples': 9040384, 'steps': 17656, 'loss/train': 1.9364594221115112} 02/25/2022 00:26:34 - INFO - codeparrot_training - Step 17657: {'lr': 0.0003798285103806831, 'samples': 9040896, 'steps': 17657, 'loss/train': 1.139986515045166} 02/25/2022 00:26:38 - INFO - codeparrot_training - Step 17658: {'lr': 0.0003798145270131091, 'samples': 9041408, 'steps': 17658, 'loss/train': 0.7799100279808044} 02/25/2022 00:26:43 - INFO - codeparrot_training - Step 17659: {'lr': 0.00037980054308945076, 'samples': 9041920, 'steps': 17659, 'loss/train': 2.321681261062622} 02/25/2022 00:26:47 - INFO - codeparrot_training - Step 17660: {'lr': 0.00037978655860976826, 'samples': 9042432, 'steps': 17660, 'loss/train': 1.8950034379959106} 02/25/2022 00:26:52 - INFO - codeparrot_training - Step 17661: {'lr': 0.0003797725735741212, 'samples': 9042944, 'steps': 17661, 'loss/train': 1.894861102104187} 02/25/2022 00:26:59 - INFO - codeparrot_training - Step 17662: {'lr': 0.0003797585879825698, 'samples': 9043456, 'steps': 17662, 'loss/train': 1.2853599786758423} 02/25/2022 00:27:03 - INFO - codeparrot_training - Step 17663: {'lr': 0.00037974460183517366, 'samples': 9043968, 'steps': 17663, 'loss/train': 1.0120782852172852} 02/25/2022 00:27:08 - INFO - codeparrot_training - Step 17664: {'lr': 0.0003797306151319929, 'samples': 9044480, 'steps': 17664, 'loss/train': 1.5664876699447632} 02/25/2022 00:27:12 - INFO - codeparrot_training - Step 17665: {'lr': 0.00037971662787308734, 'samples': 9044992, 'steps': 17665, 'loss/train': 0.9982131719589233} 02/25/2022 00:27:18 - INFO - codeparrot_training - Step 17666: {'lr': 0.00037970264005851703, 'samples': 9045504, 'steps': 17666, 'loss/train': 3.467337131500244} 02/25/2022 00:27:22 - INFO - codeparrot_training - Step 17667: {'lr': 0.0003796886516883418, 'samples': 9046016, 'steps': 17667, 'loss/train': 1.6415878534317017} 02/25/2022 00:27:25 - INFO - codeparrot_training - Step 17668: {'lr': 0.0003796746627626214, 'samples': 9046528, 'steps': 17668, 'loss/train': 3.6356019973754883} 02/25/2022 00:27:31 - INFO - codeparrot_training - Step 17669: {'lr': 0.00037966067328141606, 'samples': 9047040, 'steps': 17669, 'loss/train': 2.9348771572113037} 02/25/2022 00:27:34 - INFO - codeparrot_training - Step 17670: {'lr': 0.0003796466832447856, 'samples': 9047552, 'steps': 17670, 'loss/train': 2.1387054920196533} 02/25/2022 00:27:40 - INFO - codeparrot_training - Step 17671: {'lr': 0.00037963269265278986, 'samples': 9048064, 'steps': 17671, 'loss/train': 2.7863388061523438} 02/25/2022 00:27:43 - INFO - codeparrot_training - Step 17672: {'lr': 0.0003796187015054888, 'samples': 9048576, 'steps': 17672, 'loss/train': 1.904708981513977} 02/25/2022 00:27:51 - INFO - codeparrot_training - Step 17673: {'lr': 0.0003796047098029424, 'samples': 9049088, 'steps': 17673, 'loss/train': 1.4503318071365356} 02/25/2022 00:27:54 - INFO - codeparrot_training - Step 17674: {'lr': 0.0003795907175452106, 'samples': 9049600, 'steps': 17674, 'loss/train': 0.6155263185501099} 02/25/2022 00:28:00 - INFO - codeparrot_training - Step 17675: {'lr': 0.0003795767247323533, 'samples': 9050112, 'steps': 17675, 'loss/train': 2.0112037658691406} 02/25/2022 00:28:03 - INFO - codeparrot_training - Step 17676: {'lr': 0.00037956273136443056, 'samples': 9050624, 'steps': 17676, 'loss/train': 1.6704438924789429} 02/25/2022 00:28:09 - INFO - codeparrot_training - Step 17677: {'lr': 0.000379548737441502, 'samples': 9051136, 'steps': 17677, 'loss/train': 3.0951290130615234} 02/25/2022 00:28:12 - INFO - codeparrot_training - Step 17678: {'lr': 0.00037953474296362796, 'samples': 9051648, 'steps': 17678, 'loss/train': 2.1795482635498047} 02/25/2022 00:28:18 - INFO - codeparrot_training - Step 17679: {'lr': 0.0003795207479308681, 'samples': 9052160, 'steps': 17679, 'loss/train': 2.0755298137664795} 02/25/2022 00:28:21 - INFO - codeparrot_training - Step 17680: {'lr': 0.00037950675234328256, 'samples': 9052672, 'steps': 17680, 'loss/train': 2.460949182510376} 02/25/2022 00:28:27 - INFO - codeparrot_training - Step 17681: {'lr': 0.00037949275620093124, 'samples': 9053184, 'steps': 17681, 'loss/train': 0.4283615052700043} 02/25/2022 00:28:30 - INFO - codeparrot_training - Step 17682: {'lr': 0.000379478759503874, 'samples': 9053696, 'steps': 17682, 'loss/train': 1.1838616132736206} 02/25/2022 00:28:36 - INFO - codeparrot_training - Step 17683: {'lr': 0.00037946476225217087, 'samples': 9054208, 'steps': 17683, 'loss/train': 0.9548600912094116} 02/25/2022 00:28:39 - INFO - codeparrot_training - Step 17684: {'lr': 0.0003794507644458819, 'samples': 9054720, 'steps': 17684, 'loss/train': 1.411368727684021} 02/25/2022 00:28:46 - INFO - codeparrot_training - Step 17685: {'lr': 0.00037943676608506683, 'samples': 9055232, 'steps': 17685, 'loss/train': 1.7294789552688599} 02/25/2022 00:28:50 - INFO - codeparrot_training - Step 17686: {'lr': 0.00037942276716978584, 'samples': 9055744, 'steps': 17686, 'loss/train': 1.638486385345459} 02/25/2022 00:28:55 - INFO - codeparrot_training - Step 17687: {'lr': 0.0003794087677000988, 'samples': 9056256, 'steps': 17687, 'loss/train': 2.057386636734009} 02/25/2022 00:28:59 - INFO - codeparrot_training - Step 17688: {'lr': 0.0003793947676760657, 'samples': 9056768, 'steps': 17688, 'loss/train': 1.859421730041504} 02/25/2022 00:29:04 - INFO - codeparrot_training - Step 17689: {'lr': 0.00037938076709774645, 'samples': 9057280, 'steps': 17689, 'loss/train': 2.4088358879089355} 02/25/2022 00:29:08 - INFO - codeparrot_training - Step 17690: {'lr': 0.0003793667659652011, 'samples': 9057792, 'steps': 17690, 'loss/train': 2.4362845420837402} 02/25/2022 00:29:13 - INFO - codeparrot_training - Step 17691: {'lr': 0.0003793527642784896, 'samples': 9058304, 'steps': 17691, 'loss/train': 2.44637131690979} 02/25/2022 00:29:17 - INFO - codeparrot_training - Step 17692: {'lr': 0.0003793387620376719, 'samples': 9058816, 'steps': 17692, 'loss/train': 1.7292999029159546} 02/25/2022 00:29:22 - INFO - codeparrot_training - Step 17693: {'lr': 0.0003793247592428081, 'samples': 9059328, 'steps': 17693, 'loss/train': 1.4550867080688477} 02/25/2022 00:29:26 - INFO - codeparrot_training - Step 17694: {'lr': 0.00037931075589395805, 'samples': 9059840, 'steps': 17694, 'loss/train': 1.7441051006317139} 02/25/2022 00:29:31 - INFO - codeparrot_training - Step 17695: {'lr': 0.00037929675199118183, 'samples': 9060352, 'steps': 17695, 'loss/train': 2.029517889022827} 02/25/2022 00:29:35 - INFO - codeparrot_training - Step 17696: {'lr': 0.0003792827475345393, 'samples': 9060864, 'steps': 17696, 'loss/train': 1.2710902690887451} 02/25/2022 00:29:40 - INFO - codeparrot_training - Step 17697: {'lr': 0.0003792687425240906, 'samples': 9061376, 'steps': 17697, 'loss/train': 1.5868641138076782} 02/25/2022 00:29:44 - INFO - codeparrot_training - Step 17698: {'lr': 0.0003792547369598956, 'samples': 9061888, 'steps': 17698, 'loss/train': 1.3753433227539062} 02/25/2022 00:29:51 - INFO - codeparrot_training - Step 17699: {'lr': 0.0003792407308420144, 'samples': 9062400, 'steps': 17699, 'loss/train': 1.7627010345458984} 02/25/2022 00:29:55 - INFO - codeparrot_training - Step 17700: {'lr': 0.00037922672417050685, 'samples': 9062912, 'steps': 17700, 'loss/train': 1.804733395576477} 02/25/2022 00:30:00 - INFO - codeparrot_training - Step 17701: {'lr': 0.00037921271694543317, 'samples': 9063424, 'steps': 17701, 'loss/train': 1.8018553256988525} 02/25/2022 00:30:04 - INFO - codeparrot_training - Step 17702: {'lr': 0.0003791987091668532, 'samples': 9063936, 'steps': 17702, 'loss/train': 1.9875469207763672} 02/25/2022 00:30:09 - INFO - codeparrot_training - Step 17703: {'lr': 0.00037918470083482693, 'samples': 9064448, 'steps': 17703, 'loss/train': 1.7927709817886353} 02/25/2022 00:30:13 - INFO - codeparrot_training - Step 17704: {'lr': 0.0003791706919494145, 'samples': 9064960, 'steps': 17704, 'loss/train': 2.337580680847168} 02/25/2022 00:30:18 - INFO - codeparrot_training - Step 17705: {'lr': 0.0003791566825106758, 'samples': 9065472, 'steps': 17705, 'loss/train': 2.726628065109253} 02/25/2022 00:30:22 - INFO - codeparrot_training - Step 17706: {'lr': 0.0003791426725186709, 'samples': 9065984, 'steps': 17706, 'loss/train': 2.2806732654571533} 02/25/2022 00:30:27 - INFO - codeparrot_training - Step 17707: {'lr': 0.0003791286619734597, 'samples': 9066496, 'steps': 17707, 'loss/train': 1.6257905960083008} 02/25/2022 00:30:31 - INFO - codeparrot_training - Step 17708: {'lr': 0.0003791146508751025, 'samples': 9067008, 'steps': 17708, 'loss/train': 0.14708131551742554} 02/25/2022 00:30:37 - INFO - codeparrot_training - Step 17709: {'lr': 0.00037910063922365903, 'samples': 9067520, 'steps': 17709, 'loss/train': 1.544282078742981} 02/25/2022 00:30:40 - INFO - codeparrot_training - Step 17710: {'lr': 0.00037908662701918944, 'samples': 9068032, 'steps': 17710, 'loss/train': 2.2766592502593994} 02/25/2022 00:30:43 - INFO - codeparrot_training - Step 17711: {'lr': 0.00037907261426175365, 'samples': 9068544, 'steps': 17711, 'loss/train': 2.3601431846618652} 02/25/2022 00:30:51 - INFO - codeparrot_training - Step 17712: {'lr': 0.0003790586009514119, 'samples': 9069056, 'steps': 17712, 'loss/train': 1.7010282278060913} 02/25/2022 00:30:54 - INFO - codeparrot_training - Step 17713: {'lr': 0.000379044587088224, 'samples': 9069568, 'steps': 17713, 'loss/train': 2.3169925212860107} 02/25/2022 00:31:00 - INFO - codeparrot_training - Step 17714: {'lr': 0.0003790305726722501, 'samples': 9070080, 'steps': 17714, 'loss/train': 1.6872769594192505} 02/25/2022 00:31:06 - INFO - codeparrot_training - Step 17715: {'lr': 0.00037901655770355015, 'samples': 9070592, 'steps': 17715, 'loss/train': 1.6401950120925903} 02/25/2022 00:31:09 - INFO - codeparrot_training - Step 17716: {'lr': 0.0003790025421821843, 'samples': 9071104, 'steps': 17716, 'loss/train': 2.0675594806671143} 02/25/2022 00:31:12 - INFO - codeparrot_training - Step 17717: {'lr': 0.0003789885261082124, 'samples': 9071616, 'steps': 17717, 'loss/train': 1.3518098592758179} 02/25/2022 00:31:18 - INFO - codeparrot_training - Step 17718: {'lr': 0.00037897450948169476, 'samples': 9072128, 'steps': 17718, 'loss/train': 2.6437594890594482} 02/25/2022 00:31:21 - INFO - codeparrot_training - Step 17719: {'lr': 0.0003789604923026912, 'samples': 9072640, 'steps': 17719, 'loss/train': 1.2582480907440186} 02/25/2022 00:31:27 - INFO - codeparrot_training - Step 17720: {'lr': 0.00037894647457126186, 'samples': 9073152, 'steps': 17720, 'loss/train': 1.8962912559509277} 02/25/2022 00:31:30 - INFO - codeparrot_training - Step 17721: {'lr': 0.0003789324562874668, 'samples': 9073664, 'steps': 17721, 'loss/train': 2.2809383869171143} 02/25/2022 00:31:38 - INFO - codeparrot_training - Step 17722: {'lr': 0.000378918437451366, 'samples': 9074176, 'steps': 17722, 'loss/train': 2.389406681060791} 02/25/2022 00:31:41 - INFO - codeparrot_training - Step 17723: {'lr': 0.00037890441806301954, 'samples': 9074688, 'steps': 17723, 'loss/train': 2.9667856693267822} 02/25/2022 00:31:47 - INFO - codeparrot_training - Step 17724: {'lr': 0.0003788903981224875, 'samples': 9075200, 'steps': 17724, 'loss/train': 1.3824249505996704} 02/25/2022 00:31:50 - INFO - codeparrot_training - Step 17725: {'lr': 0.00037887637762982996, 'samples': 9075712, 'steps': 17725, 'loss/train': 1.8328057527542114} 02/25/2022 00:31:56 - INFO - codeparrot_training - Step 17726: {'lr': 0.0003788623565851068, 'samples': 9076224, 'steps': 17726, 'loss/train': 1.9919962882995605} 02/25/2022 00:31:59 - INFO - codeparrot_training - Step 17727: {'lr': 0.00037884833498837833, 'samples': 9076736, 'steps': 17727, 'loss/train': 2.372586965560913} 02/25/2022 00:32:05 - INFO - codeparrot_training - Step 17728: {'lr': 0.00037883431283970454, 'samples': 9077248, 'steps': 17728, 'loss/train': 2.5369796752929688} 02/25/2022 00:32:08 - INFO - codeparrot_training - Step 17729: {'lr': 0.00037882029013914544, 'samples': 9077760, 'steps': 17729, 'loss/train': 2.93064022064209} 02/25/2022 00:32:14 - INFO - codeparrot_training - Step 17730: {'lr': 0.0003788062668867611, 'samples': 9078272, 'steps': 17730, 'loss/train': 2.5228629112243652} 02/25/2022 00:32:17 - INFO - codeparrot_training - Step 17731: {'lr': 0.00037879224308261163, 'samples': 9078784, 'steps': 17731, 'loss/train': 2.4482505321502686} 02/25/2022 00:32:23 - INFO - codeparrot_training - Step 17732: {'lr': 0.00037877821872675705, 'samples': 9079296, 'steps': 17732, 'loss/train': 2.4996213912963867} 02/25/2022 00:32:26 - INFO - codeparrot_training - Step 17733: {'lr': 0.0003787641938192575, 'samples': 9079808, 'steps': 17733, 'loss/train': 3.121630907058716} 02/25/2022 00:32:32 - INFO - codeparrot_training - Step 17734: {'lr': 0.00037875016836017304, 'samples': 9080320, 'steps': 17734, 'loss/train': 1.9725698232650757} 02/25/2022 00:32:36 - INFO - codeparrot_training - Step 17735: {'lr': 0.0003787361423495637, 'samples': 9080832, 'steps': 17735, 'loss/train': 1.4404319524765015} 02/25/2022 00:32:42 - INFO - codeparrot_training - Step 17736: {'lr': 0.0003787221157874897, 'samples': 9081344, 'steps': 17736, 'loss/train': 2.0056955814361572} 02/25/2022 00:32:45 - INFO - codeparrot_training - Step 17737: {'lr': 0.00037870808867401085, 'samples': 9081856, 'steps': 17737, 'loss/train': 1.2769489288330078} 02/25/2022 00:32:51 - INFO - codeparrot_training - Step 17738: {'lr': 0.00037869406100918756, 'samples': 9082368, 'steps': 17738, 'loss/train': 2.929872512817383} 02/25/2022 00:32:54 - INFO - codeparrot_training - Step 17739: {'lr': 0.0003786800327930797, 'samples': 9082880, 'steps': 17739, 'loss/train': 1.3697381019592285} 02/25/2022 00:33:00 - INFO - codeparrot_training - Step 17740: {'lr': 0.0003786660040257475, 'samples': 9083392, 'steps': 17740, 'loss/train': 3.458914279937744} 02/25/2022 00:33:03 - INFO - codeparrot_training - Step 17741: {'lr': 0.00037865197470725103, 'samples': 9083904, 'steps': 17741, 'loss/train': 2.1405208110809326} 02/25/2022 00:33:09 - INFO - codeparrot_training - Step 17742: {'lr': 0.0003786379448376503, 'samples': 9084416, 'steps': 17742, 'loss/train': 1.8302637338638306} 02/25/2022 00:33:12 - INFO - codeparrot_training - Step 17743: {'lr': 0.0003786239144170055, 'samples': 9084928, 'steps': 17743, 'loss/train': 2.5693111419677734} 02/25/2022 00:33:18 - INFO - codeparrot_training - Step 17744: {'lr': 0.0003786098834453766, 'samples': 9085440, 'steps': 17744, 'loss/train': 2.4527623653411865} 02/25/2022 00:33:21 - INFO - codeparrot_training - Step 17745: {'lr': 0.00037859585192282386, 'samples': 9085952, 'steps': 17745, 'loss/train': 2.138601779937744} 02/25/2022 00:33:27 - INFO - codeparrot_training - Step 17746: {'lr': 0.00037858181984940734, 'samples': 9086464, 'steps': 17746, 'loss/train': 2.146455764770508} 02/25/2022 00:33:30 - INFO - codeparrot_training - Step 17747: {'lr': 0.0003785677872251871, 'samples': 9086976, 'steps': 17747, 'loss/train': 2.2206130027770996} 02/25/2022 00:33:36 - INFO - codeparrot_training - Step 17748: {'lr': 0.0003785537540502233, 'samples': 9087488, 'steps': 17748, 'loss/train': 1.8951431512832642} 02/25/2022 00:33:40 - INFO - codeparrot_training - Step 17749: {'lr': 0.0003785397203245761, 'samples': 9088000, 'steps': 17749, 'loss/train': 2.0423741340637207} 02/25/2022 00:33:45 - INFO - codeparrot_training - Step 17750: {'lr': 0.0003785256860483054, 'samples': 9088512, 'steps': 17750, 'loss/train': 2.477142572402954} 02/25/2022 00:33:49 - INFO - codeparrot_training - Step 17751: {'lr': 0.0003785116512214716, 'samples': 9089024, 'steps': 17751, 'loss/train': 2.615281105041504} 02/25/2022 00:33:54 - INFO - codeparrot_training - Step 17752: {'lr': 0.0003784976158441347, 'samples': 9089536, 'steps': 17752, 'loss/train': 1.9404678344726562} 02/25/2022 00:33:58 - INFO - codeparrot_training - Step 17753: {'lr': 0.0003784835799163547, 'samples': 9090048, 'steps': 17753, 'loss/train': 1.0144660472869873} 02/25/2022 00:34:04 - INFO - codeparrot_training - Step 17754: {'lr': 0.00037846954343819195, 'samples': 9090560, 'steps': 17754, 'loss/train': 3.125117778778076} 02/25/2022 00:34:09 - INFO - codeparrot_training - Step 17755: {'lr': 0.00037845550640970636, 'samples': 9091072, 'steps': 17755, 'loss/train': 3.338714599609375} 02/25/2022 00:34:12 - INFO - codeparrot_training - Step 17756: {'lr': 0.0003784414688309583, 'samples': 9091584, 'steps': 17756, 'loss/train': 1.8071317672729492} 02/25/2022 00:34:19 - INFO - codeparrot_training - Step 17757: {'lr': 0.00037842743070200767, 'samples': 9092096, 'steps': 17757, 'loss/train': 4.624802112579346} 02/25/2022 00:34:22 - INFO - codeparrot_training - Step 17758: {'lr': 0.0003784133920229148, 'samples': 9092608, 'steps': 17758, 'loss/train': 1.204598307609558} 02/25/2022 00:34:28 - INFO - codeparrot_training - Step 17759: {'lr': 0.0003783993527937397, 'samples': 9093120, 'steps': 17759, 'loss/train': 2.5813825130462646} 02/25/2022 00:34:31 - INFO - codeparrot_training - Step 17760: {'lr': 0.0003783853130145425, 'samples': 9093632, 'steps': 17760, 'loss/train': 2.3898720741271973} 02/25/2022 00:34:35 - INFO - codeparrot_training - Step 17761: {'lr': 0.0003783712726853835, 'samples': 9094144, 'steps': 17761, 'loss/train': 1.5693180561065674} 02/25/2022 00:34:40 - INFO - codeparrot_training - Step 17762: {'lr': 0.00037835723180632263, 'samples': 9094656, 'steps': 17762, 'loss/train': 1.8705394268035889} 02/25/2022 00:34:44 - INFO - codeparrot_training - Step 17763: {'lr': 0.00037834319037742016, 'samples': 9095168, 'steps': 17763, 'loss/train': 4.796335697174072} 02/25/2022 00:34:49 - INFO - codeparrot_training - Step 17764: {'lr': 0.00037832914839873623, 'samples': 9095680, 'steps': 17764, 'loss/train': 1.8139954805374146} 02/25/2022 00:34:53 - INFO - codeparrot_training - Step 17765: {'lr': 0.0003783151058703309, 'samples': 9096192, 'steps': 17765, 'loss/train': 1.8321171998977661} 02/25/2022 00:34:58 - INFO - codeparrot_training - Step 17766: {'lr': 0.0003783010627922645, 'samples': 9096704, 'steps': 17766, 'loss/train': 1.9763429164886475} 02/25/2022 00:35:02 - INFO - codeparrot_training - Step 17767: {'lr': 0.0003782870191645971, 'samples': 9097216, 'steps': 17767, 'loss/train': 0.9249021410942078} 02/25/2022 00:35:08 - INFO - codeparrot_training - Step 17768: {'lr': 0.0003782729749873887, 'samples': 9097728, 'steps': 17768, 'loss/train': 1.7698930501937866} 02/25/2022 00:35:11 - INFO - codeparrot_training - Step 17769: {'lr': 0.00037825893026069977, 'samples': 9098240, 'steps': 17769, 'loss/train': 0.3880889117717743} 02/25/2022 00:35:17 - INFO - codeparrot_training - Step 17770: {'lr': 0.0003782448849845902, 'samples': 9098752, 'steps': 17770, 'loss/train': 1.4427440166473389} 02/25/2022 00:35:20 - INFO - codeparrot_training - Step 17771: {'lr': 0.0003782308391591203, 'samples': 9099264, 'steps': 17771, 'loss/train': 2.7742958068847656} 02/25/2022 00:35:26 - INFO - codeparrot_training - Step 17772: {'lr': 0.00037821679278435017, 'samples': 9099776, 'steps': 17772, 'loss/train': 2.4256370067596436} 02/25/2022 00:35:30 - INFO - codeparrot_training - Step 17773: {'lr': 0.0003782027458603401, 'samples': 9100288, 'steps': 17773, 'loss/train': 3.8415708541870117} 02/25/2022 00:35:35 - INFO - codeparrot_training - Step 17774: {'lr': 0.0003781886983871501, 'samples': 9100800, 'steps': 17774, 'loss/train': 2.436509132385254} 02/25/2022 00:35:39 - INFO - codeparrot_training - Step 17775: {'lr': 0.00037817465036484043, 'samples': 9101312, 'steps': 17775, 'loss/train': 1.5396196842193604} 02/25/2022 00:35:44 - INFO - codeparrot_training - Step 17776: {'lr': 0.0003781606017934713, 'samples': 9101824, 'steps': 17776, 'loss/train': 1.927518606185913} 02/25/2022 00:35:48 - INFO - codeparrot_training - Step 17777: {'lr': 0.0003781465526731028, 'samples': 9102336, 'steps': 17777, 'loss/train': 2.4948067665100098} 02/25/2022 00:35:53 - INFO - codeparrot_training - Step 17778: {'lr': 0.0003781325030037952, 'samples': 9102848, 'steps': 17778, 'loss/train': 1.948328971862793} 02/25/2022 00:35:57 - INFO - codeparrot_training - Step 17779: {'lr': 0.00037811845278560864, 'samples': 9103360, 'steps': 17779, 'loss/train': 0.8280380964279175} 02/25/2022 00:36:02 - INFO - codeparrot_training - Step 17780: {'lr': 0.0003781044020186033, 'samples': 9103872, 'steps': 17780, 'loss/train': 1.3666096925735474} 02/25/2022 00:36:06 - INFO - codeparrot_training - Step 17781: {'lr': 0.0003780903507028393, 'samples': 9104384, 'steps': 17781, 'loss/train': 0.5460395812988281} 02/25/2022 00:36:12 - INFO - codeparrot_training - Step 17782: {'lr': 0.00037807629883837703, 'samples': 9104896, 'steps': 17782, 'loss/train': 2.9699387550354004} 02/25/2022 00:36:16 - INFO - codeparrot_training - Step 17783: {'lr': 0.00037806224642527653, 'samples': 9105408, 'steps': 17783, 'loss/train': 1.9725273847579956} 02/25/2022 00:36:21 - INFO - codeparrot_training - Step 17784: {'lr': 0.000378048193463598, 'samples': 9105920, 'steps': 17784, 'loss/train': 2.8924503326416016} 02/25/2022 00:36:25 - INFO - codeparrot_training - Step 17785: {'lr': 0.0003780341399534017, 'samples': 9106432, 'steps': 17785, 'loss/train': 2.25705623626709} 02/25/2022 00:36:30 - INFO - codeparrot_training - Step 17786: {'lr': 0.00037802008589474777, 'samples': 9106944, 'steps': 17786, 'loss/train': 2.7668399810791016} 02/25/2022 00:36:33 - INFO - codeparrot_training - Step 17787: {'lr': 0.0003780060312876965, 'samples': 9107456, 'steps': 17787, 'loss/train': 3.03151273727417} 02/25/2022 00:36:39 - INFO - codeparrot_training - Step 17788: {'lr': 0.00037799197613230795, 'samples': 9107968, 'steps': 17788, 'loss/train': 1.6267716884613037} 02/25/2022 00:36:43 - INFO - codeparrot_training - Step 17789: {'lr': 0.00037797792042864247, 'samples': 9108480, 'steps': 17789, 'loss/train': 2.3754568099975586} 02/25/2022 00:36:48 - INFO - codeparrot_training - Step 17790: {'lr': 0.0003779638641767602, 'samples': 9108992, 'steps': 17790, 'loss/train': 1.3895710706710815} 02/25/2022 00:36:52 - INFO - codeparrot_training - Step 17791: {'lr': 0.0003779498073767214, 'samples': 9109504, 'steps': 17791, 'loss/train': 2.1284360885620117} 02/25/2022 00:36:57 - INFO - codeparrot_training - Step 17792: {'lr': 0.00037793575002858625, 'samples': 9110016, 'steps': 17792, 'loss/train': 1.511710524559021} 02/25/2022 00:37:01 - INFO - codeparrot_training - Step 17793: {'lr': 0.00037792169213241494, 'samples': 9110528, 'steps': 17793, 'loss/train': 1.6062425374984741} 02/25/2022 00:37:07 - INFO - codeparrot_training - Step 17794: {'lr': 0.00037790763368826774, 'samples': 9111040, 'steps': 17794, 'loss/train': 0.8982266783714294} 02/25/2022 00:37:11 - INFO - codeparrot_training - Step 17795: {'lr': 0.00037789357469620487, 'samples': 9111552, 'steps': 17795, 'loss/train': 2.3349978923797607} 02/25/2022 00:37:16 - INFO - codeparrot_training - Step 17796: {'lr': 0.0003778795151562865, 'samples': 9112064, 'steps': 17796, 'loss/train': 2.528998851776123} 02/25/2022 00:37:20 - INFO - codeparrot_training - Step 17797: {'lr': 0.00037786545506857295, 'samples': 9112576, 'steps': 17797, 'loss/train': 2.137895107269287} 02/25/2022 00:37:25 - INFO - codeparrot_training - Step 17798: {'lr': 0.0003778513944331243, 'samples': 9113088, 'steps': 17798, 'loss/train': 2.3629095554351807} 02/25/2022 00:37:29 - INFO - codeparrot_training - Step 17799: {'lr': 0.0003778373332500009, 'samples': 9113600, 'steps': 17799, 'loss/train': 2.8258047103881836} 02/25/2022 00:37:34 - INFO - codeparrot_training - Step 17800: {'lr': 0.00037782327151926297, 'samples': 9114112, 'steps': 17800, 'loss/train': 2.040865898132324} 02/25/2022 00:37:38 - INFO - codeparrot_training - Step 17801: {'lr': 0.00037780920924097085, 'samples': 9114624, 'steps': 17801, 'loss/train': 2.445117235183716} 02/25/2022 00:37:43 - INFO - codeparrot_training - Step 17802: {'lr': 0.00037779514641518455, 'samples': 9115136, 'steps': 17802, 'loss/train': 1.2249690294265747} 02/25/2022 00:37:47 - INFO - codeparrot_training - Step 17803: {'lr': 0.0003777810830419644, 'samples': 9115648, 'steps': 17803, 'loss/train': 2.1157913208007812} 02/25/2022 00:37:53 - INFO - codeparrot_training - Step 17804: {'lr': 0.00037776701912137066, 'samples': 9116160, 'steps': 17804, 'loss/train': 2.5650148391723633} 02/25/2022 00:37:56 - INFO - codeparrot_training - Step 17805: {'lr': 0.00037775295465346373, 'samples': 9116672, 'steps': 17805, 'loss/train': 0.37460461258888245} 02/25/2022 00:38:02 - INFO - codeparrot_training - Step 17806: {'lr': 0.0003777388896383035, 'samples': 9117184, 'steps': 17806, 'loss/train': 2.438077688217163} 02/25/2022 00:38:05 - INFO - codeparrot_training - Step 17807: {'lr': 0.00037772482407595056, 'samples': 9117696, 'steps': 17807, 'loss/train': 2.6476752758026123} 02/25/2022 00:38:11 - INFO - codeparrot_training - Step 17808: {'lr': 0.000377710757966465, 'samples': 9118208, 'steps': 17808, 'loss/train': 2.1442501544952393} 02/25/2022 00:38:15 - INFO - codeparrot_training - Step 17809: {'lr': 0.0003776966913099071, 'samples': 9118720, 'steps': 17809, 'loss/train': 2.276533603668213} 02/25/2022 00:38:20 - INFO - codeparrot_training - Step 17810: {'lr': 0.00037768262410633715, 'samples': 9119232, 'steps': 17810, 'loss/train': 1.7017161846160889} 02/25/2022 00:38:24 - INFO - codeparrot_training - Step 17811: {'lr': 0.0003776685563558153, 'samples': 9119744, 'steps': 17811, 'loss/train': 1.5580724477767944} 02/25/2022 00:38:29 - INFO - codeparrot_training - Step 17812: {'lr': 0.00037765448805840196, 'samples': 9120256, 'steps': 17812, 'loss/train': 1.3341552019119263} 02/25/2022 00:38:33 - INFO - codeparrot_training - Step 17813: {'lr': 0.00037764041921415736, 'samples': 9120768, 'steps': 17813, 'loss/train': 2.3488428592681885} 02/25/2022 00:38:39 - INFO - codeparrot_training - Step 17814: {'lr': 0.00037762634982314164, 'samples': 9121280, 'steps': 17814, 'loss/train': 1.8621599674224854} 02/25/2022 00:38:42 - INFO - codeparrot_training - Step 17815: {'lr': 0.00037761227988541523, 'samples': 9121792, 'steps': 17815, 'loss/train': 1.6176320314407349} 02/25/2022 00:38:48 - INFO - codeparrot_training - Step 17816: {'lr': 0.00037759820940103827, 'samples': 9122304, 'steps': 17816, 'loss/train': 2.340167760848999} 02/25/2022 00:38:51 - INFO - codeparrot_training - Step 17817: {'lr': 0.00037758413837007124, 'samples': 9122816, 'steps': 17817, 'loss/train': 0.27927979826927185} 02/25/2022 00:38:57 - INFO - codeparrot_training - Step 17818: {'lr': 0.0003775700667925741, 'samples': 9123328, 'steps': 17818, 'loss/train': 2.227670907974243} 02/25/2022 00:39:00 - INFO - codeparrot_training - Step 17819: {'lr': 0.0003775559946686075, 'samples': 9123840, 'steps': 17819, 'loss/train': 1.4832324981689453} 02/25/2022 00:39:06 - INFO - codeparrot_training - Step 17820: {'lr': 0.00037754192199823135, 'samples': 9124352, 'steps': 17820, 'loss/train': 1.5890132188796997} 02/25/2022 00:39:09 - INFO - codeparrot_training - Step 17821: {'lr': 0.00037752784878150613, 'samples': 9124864, 'steps': 17821, 'loss/train': 1.9960347414016724} 02/25/2022 00:39:15 - INFO - codeparrot_training - Step 17822: {'lr': 0.00037751377501849215, 'samples': 9125376, 'steps': 17822, 'loss/train': 1.1255578994750977} 02/25/2022 00:39:18 - INFO - codeparrot_training - Step 17823: {'lr': 0.0003774997007092496, 'samples': 9125888, 'steps': 17823, 'loss/train': 1.7893080711364746} 02/25/2022 00:39:24 - INFO - codeparrot_training - Step 17824: {'lr': 0.00037748562585383886, 'samples': 9126400, 'steps': 17824, 'loss/train': 2.536302089691162} 02/25/2022 00:39:27 - INFO - codeparrot_training - Step 17825: {'lr': 0.00037747155045232016, 'samples': 9126912, 'steps': 17825, 'loss/train': 2.4392430782318115} 02/25/2022 00:39:33 - INFO - codeparrot_training - Step 17826: {'lr': 0.0003774574745047539, 'samples': 9127424, 'steps': 17826, 'loss/train': 2.0956664085388184} 02/25/2022 00:39:36 - INFO - codeparrot_training - Step 17827: {'lr': 0.0003774433980112001, 'samples': 9127936, 'steps': 17827, 'loss/train': 1.8346408605575562} 02/25/2022 00:39:42 - INFO - codeparrot_training - Step 17828: {'lr': 0.00037742932097171945, 'samples': 9128448, 'steps': 17828, 'loss/train': 2.8361427783966064} 02/25/2022 00:39:45 - INFO - codeparrot_training - Step 17829: {'lr': 0.0003774152433863719, 'samples': 9128960, 'steps': 17829, 'loss/train': 3.5036838054656982} 02/25/2022 00:39:52 - INFO - codeparrot_training - Step 17830: {'lr': 0.000377401165255218, 'samples': 9129472, 'steps': 17830, 'loss/train': 2.1435322761535645} 02/25/2022 00:39:55 - INFO - codeparrot_training - Step 17831: {'lr': 0.0003773870865783179, 'samples': 9129984, 'steps': 17831, 'loss/train': 3.1499199867248535} 02/25/2022 00:40:01 - INFO - codeparrot_training - Step 17832: {'lr': 0.00037737300735573204, 'samples': 9130496, 'steps': 17832, 'loss/train': 2.5886032581329346} 02/25/2022 00:40:06 - INFO - codeparrot_training - Step 17833: {'lr': 0.00037735892758752063, 'samples': 9131008, 'steps': 17833, 'loss/train': 1.7489620447158813} 02/25/2022 00:40:10 - INFO - codeparrot_training - Step 17834: {'lr': 0.000377344847273744, 'samples': 9131520, 'steps': 17834, 'loss/train': 3.011117935180664} 02/25/2022 00:40:13 - INFO - codeparrot_training - Step 17835: {'lr': 0.0003773307664144625, 'samples': 9132032, 'steps': 17835, 'loss/train': 1.3230488300323486} 02/25/2022 00:40:19 - INFO - codeparrot_training - Step 17836: {'lr': 0.00037731668500973637, 'samples': 9132544, 'steps': 17836, 'loss/train': 1.1957640647888184} 02/25/2022 00:40:24 - INFO - codeparrot_training - Step 17837: {'lr': 0.00037730260305962604, 'samples': 9133056, 'steps': 17837, 'loss/train': 2.230207920074463} 02/25/2022 00:40:28 - INFO - codeparrot_training - Step 17838: {'lr': 0.00037728852056419183, 'samples': 9133568, 'steps': 17838, 'loss/train': 2.751715898513794} 02/25/2022 00:40:31 - INFO - codeparrot_training - Step 17839: {'lr': 0.000377274437523494, 'samples': 9134080, 'steps': 17839, 'loss/train': 2.4595115184783936} 02/25/2022 00:40:37 - INFO - codeparrot_training - Step 17840: {'lr': 0.00037726035393759286, 'samples': 9134592, 'steps': 17840, 'loss/train': 0.865892767906189} 02/25/2022 00:40:43 - INFO - codeparrot_training - Step 17841: {'lr': 0.00037724626980654877, 'samples': 9135104, 'steps': 17841, 'loss/train': 2.0343310832977295} 02/25/2022 00:40:46 - INFO - codeparrot_training - Step 17842: {'lr': 0.00037723218513042203, 'samples': 9135616, 'steps': 17842, 'loss/train': 1.599124550819397} 02/25/2022 00:40:52 - INFO - codeparrot_training - Step 17843: {'lr': 0.0003772180999092731, 'samples': 9136128, 'steps': 17843, 'loss/train': 1.4413633346557617} 02/25/2022 00:40:55 - INFO - codeparrot_training - Step 17844: {'lr': 0.00037720401414316213, 'samples': 9136640, 'steps': 17844, 'loss/train': 1.9628946781158447} 02/25/2022 00:41:01 - INFO - codeparrot_training - Step 17845: {'lr': 0.00037718992783214965, 'samples': 9137152, 'steps': 17845, 'loss/train': 2.9506173133850098} 02/25/2022 00:41:04 - INFO - codeparrot_training - Step 17846: {'lr': 0.0003771758409762958, 'samples': 9137664, 'steps': 17846, 'loss/train': 1.4582948684692383} 02/25/2022 00:41:10 - INFO - codeparrot_training - Step 17847: {'lr': 0.0003771617535756611, 'samples': 9138176, 'steps': 17847, 'loss/train': 2.0441431999206543} 02/25/2022 00:41:13 - INFO - codeparrot_training - Step 17848: {'lr': 0.00037714766563030585, 'samples': 9138688, 'steps': 17848, 'loss/train': 1.0176316499710083} 02/25/2022 00:41:19 - INFO - codeparrot_training - Step 17849: {'lr': 0.00037713357714029035, 'samples': 9139200, 'steps': 17849, 'loss/train': 1.4250121116638184} 02/25/2022 00:41:23 - INFO - codeparrot_training - Step 17850: {'lr': 0.000377119488105675, 'samples': 9139712, 'steps': 17850, 'loss/train': 2.64493727684021} 02/25/2022 00:41:28 - INFO - codeparrot_training - Step 17851: {'lr': 0.00037710539852652003, 'samples': 9140224, 'steps': 17851, 'loss/train': 1.5466663837432861} 02/25/2022 00:41:32 - INFO - codeparrot_training - Step 17852: {'lr': 0.00037709130840288605, 'samples': 9140736, 'steps': 17852, 'loss/train': 1.3424402475357056} 02/25/2022 00:41:38 - INFO - codeparrot_training - Step 17853: {'lr': 0.0003770772177348331, 'samples': 9141248, 'steps': 17853, 'loss/train': 2.6895925998687744} 02/25/2022 00:41:41 - INFO - codeparrot_training - Step 17854: {'lr': 0.0003770631265224218, 'samples': 9141760, 'steps': 17854, 'loss/train': 2.4510087966918945} 02/25/2022 00:41:45 - INFO - codeparrot_training - Step 17855: {'lr': 0.0003770490347657124, 'samples': 9142272, 'steps': 17855, 'loss/train': 1.26418936252594} 02/25/2022 00:41:50 - INFO - codeparrot_training - Step 17856: {'lr': 0.00037703494246476524, 'samples': 9142784, 'steps': 17856, 'loss/train': 1.832175374031067} 02/25/2022 00:41:54 - INFO - codeparrot_training - Step 17857: {'lr': 0.00037702084961964075, 'samples': 9143296, 'steps': 17857, 'loss/train': 1.06133234500885} 02/25/2022 00:41:59 - INFO - codeparrot_training - Step 17858: {'lr': 0.00037700675623039925, 'samples': 9143808, 'steps': 17858, 'loss/train': 1.9219799041748047} 02/25/2022 00:42:03 - INFO - codeparrot_training - Step 17859: {'lr': 0.00037699266229710115, 'samples': 9144320, 'steps': 17859, 'loss/train': 2.2191593647003174} 02/25/2022 00:42:09 - INFO - codeparrot_training - Step 17860: {'lr': 0.0003769785678198068, 'samples': 9144832, 'steps': 17860, 'loss/train': 2.746103525161743} 02/25/2022 00:42:13 - INFO - codeparrot_training - Step 17861: {'lr': 0.0003769644727985766, 'samples': 9145344, 'steps': 17861, 'loss/train': 1.5594803094863892} 02/25/2022 00:42:18 - INFO - codeparrot_training - Step 17862: {'lr': 0.00037695037723347094, 'samples': 9145856, 'steps': 17862, 'loss/train': 2.2601146697998047} 02/25/2022 00:42:21 - INFO - codeparrot_training - Step 17863: {'lr': 0.00037693628112455015, 'samples': 9146368, 'steps': 17863, 'loss/train': 2.0484261512756348} 02/25/2022 00:42:27 - INFO - codeparrot_training - Step 17864: {'lr': 0.0003769221844718746, 'samples': 9146880, 'steps': 17864, 'loss/train': 1.8653650283813477} 02/25/2022 00:42:30 - INFO - codeparrot_training - Step 17865: {'lr': 0.00037690808727550477, 'samples': 9147392, 'steps': 17865, 'loss/train': 1.382372498512268} 02/25/2022 00:42:36 - INFO - codeparrot_training - Step 17866: {'lr': 0.0003768939895355009, 'samples': 9147904, 'steps': 17866, 'loss/train': 1.6069279909133911} 02/25/2022 00:42:41 - INFO - codeparrot_training - Step 17867: {'lr': 0.0003768798912519236, 'samples': 9148416, 'steps': 17867, 'loss/train': 2.339858055114746} 02/25/2022 00:42:45 - INFO - codeparrot_training - Step 17868: {'lr': 0.0003768657924248331, 'samples': 9148928, 'steps': 17868, 'loss/train': 1.9556550979614258} 02/25/2022 00:42:50 - INFO - codeparrot_training - Step 17869: {'lr': 0.0003768516930542898, 'samples': 9149440, 'steps': 17869, 'loss/train': 2.3381431102752686} 02/25/2022 00:42:54 - INFO - codeparrot_training - Step 17870: {'lr': 0.00037683759314035414, 'samples': 9149952, 'steps': 17870, 'loss/train': 1.4832879304885864} 02/25/2022 00:42:59 - INFO - codeparrot_training - Step 17871: {'lr': 0.0003768234926830865, 'samples': 9150464, 'steps': 17871, 'loss/train': 1.8632103204727173} 02/25/2022 00:43:03 - INFO - codeparrot_training - Step 17872: {'lr': 0.0003768093916825473, 'samples': 9150976, 'steps': 17872, 'loss/train': 1.5347721576690674} 02/25/2022 00:43:08 - INFO - codeparrot_training - Step 17873: {'lr': 0.00037679529013879686, 'samples': 9151488, 'steps': 17873, 'loss/train': 1.3869832754135132} 02/25/2022 00:43:12 - INFO - codeparrot_training - Step 17874: {'lr': 0.00037678118805189575, 'samples': 9152000, 'steps': 17874, 'loss/train': 4.124237060546875} 02/25/2022 00:43:18 - INFO - codeparrot_training - Step 17875: {'lr': 0.0003767670854219043, 'samples': 9152512, 'steps': 17875, 'loss/train': 2.5834977626800537} 02/25/2022 00:43:22 - INFO - codeparrot_training - Step 17876: {'lr': 0.00037675298224888287, 'samples': 9153024, 'steps': 17876, 'loss/train': 1.7203011512756348} 02/25/2022 00:43:25 - INFO - codeparrot_training - Step 17877: {'lr': 0.0003767388785328919, 'samples': 9153536, 'steps': 17877, 'loss/train': 1.0873969793319702} 02/25/2022 00:43:31 - INFO - codeparrot_training - Step 17878: {'lr': 0.0003767247742739918, 'samples': 9154048, 'steps': 17878, 'loss/train': 2.1951661109924316} 02/25/2022 00:43:34 - INFO - codeparrot_training - Step 17879: {'lr': 0.0003767106694722431, 'samples': 9154560, 'steps': 17879, 'loss/train': 2.0824179649353027} 02/25/2022 00:43:40 - INFO - codeparrot_training - Step 17880: {'lr': 0.000376696564127706, 'samples': 9155072, 'steps': 17880, 'loss/train': 2.2230143547058105} 02/25/2022 00:43:43 - INFO - codeparrot_training - Step 17881: {'lr': 0.0003766824582404411, 'samples': 9155584, 'steps': 17881, 'loss/train': 1.6026670932769775} 02/25/2022 00:43:49 - INFO - codeparrot_training - Step 17882: {'lr': 0.00037666835181050887, 'samples': 9156096, 'steps': 17882, 'loss/train': 1.7113717794418335} 02/25/2022 00:43:52 - INFO - codeparrot_training - Step 17883: {'lr': 0.0003766542448379695, 'samples': 9156608, 'steps': 17883, 'loss/train': 3.091701030731201} 02/25/2022 00:43:58 - INFO - codeparrot_training - Step 17884: {'lr': 0.0003766401373228836, 'samples': 9157120, 'steps': 17884, 'loss/train': 2.518641233444214} 02/25/2022 00:44:01 - INFO - codeparrot_training - Step 17885: {'lr': 0.00037662602926531166, 'samples': 9157632, 'steps': 17885, 'loss/train': 2.174736738204956} 02/25/2022 00:44:08 - INFO - codeparrot_training - Step 17886: {'lr': 0.0003766119206653139, 'samples': 9158144, 'steps': 17886, 'loss/train': 2.477818012237549} 02/25/2022 00:44:11 - INFO - codeparrot_training - Step 17887: {'lr': 0.00037659781152295094, 'samples': 9158656, 'steps': 17887, 'loss/train': 2.4373779296875} 02/25/2022 00:44:17 - INFO - codeparrot_training - Step 17888: {'lr': 0.0003765837018382831, 'samples': 9159168, 'steps': 17888, 'loss/train': 1.989872932434082} 02/25/2022 00:44:20 - INFO - codeparrot_training - Step 17889: {'lr': 0.00037656959161137094, 'samples': 9159680, 'steps': 17889, 'loss/train': 2.429205894470215} 02/25/2022 00:44:26 - INFO - codeparrot_training - Step 17890: {'lr': 0.00037655548084227484, 'samples': 9160192, 'steps': 17890, 'loss/train': 2.52811861038208} 02/25/2022 00:44:29 - INFO - codeparrot_training - Step 17891: {'lr': 0.0003765413695310552, 'samples': 9160704, 'steps': 17891, 'loss/train': 3.137202501296997} 02/25/2022 00:44:35 - INFO - codeparrot_training - Step 17892: {'lr': 0.00037652725767777255, 'samples': 9161216, 'steps': 17892, 'loss/train': 2.6416451930999756} 02/25/2022 00:44:38 - INFO - codeparrot_training - Step 17893: {'lr': 0.00037651314528248724, 'samples': 9161728, 'steps': 17893, 'loss/train': 2.622027635574341} 02/25/2022 00:44:44 - INFO - codeparrot_training - Step 17894: {'lr': 0.00037649903234525996, 'samples': 9162240, 'steps': 17894, 'loss/train': 1.530945897102356} 02/25/2022 00:44:47 - INFO - codeparrot_training - Step 17895: {'lr': 0.00037648491886615077, 'samples': 9162752, 'steps': 17895, 'loss/train': 2.3728630542755127} 02/25/2022 00:44:54 - INFO - codeparrot_training - Step 17896: {'lr': 0.0003764708048452205, 'samples': 9163264, 'steps': 17896, 'loss/train': 1.969873070716858} 02/25/2022 00:44:57 - INFO - codeparrot_training - Step 17897: {'lr': 0.0003764566902825294, 'samples': 9163776, 'steps': 17897, 'loss/train': 1.7541239261627197} 02/25/2022 00:45:03 - INFO - codeparrot_training - Step 17898: {'lr': 0.0003764425751781381, 'samples': 9164288, 'steps': 17898, 'loss/train': 2.7798380851745605} 02/25/2022 00:45:08 - INFO - codeparrot_training - Step 17899: {'lr': 0.0003764284595321068, 'samples': 9164800, 'steps': 17899, 'loss/train': 1.6835497617721558} 02/25/2022 00:45:12 - INFO - codeparrot_training - Step 17900: {'lr': 0.0003764143433444962, 'samples': 9165312, 'steps': 17900, 'loss/train': 0.22605247795581818} 02/25/2022 00:45:15 - INFO - codeparrot_training - Step 17901: {'lr': 0.00037640022661536665, 'samples': 9165824, 'steps': 17901, 'loss/train': 2.232663631439209} 02/25/2022 00:45:22 - INFO - codeparrot_training - Step 17902: {'lr': 0.0003763861093447787, 'samples': 9166336, 'steps': 17902, 'loss/train': 1.1055772304534912} 02/25/2022 00:45:25 - INFO - codeparrot_training - Step 17903: {'lr': 0.0003763719915327928, 'samples': 9166848, 'steps': 17903, 'loss/train': 2.2600111961364746} 02/25/2022 00:45:31 - INFO - codeparrot_training - Step 17904: {'lr': 0.00037635787317946945, 'samples': 9167360, 'steps': 17904, 'loss/train': 1.1961216926574707} 02/25/2022 00:45:34 - INFO - codeparrot_training - Step 17905: {'lr': 0.000376343754284869, 'samples': 9167872, 'steps': 17905, 'loss/train': 3.1459691524505615} 02/25/2022 00:45:38 - INFO - codeparrot_training - Step 17906: {'lr': 0.00037632963484905213, 'samples': 9168384, 'steps': 17906, 'loss/train': 1.6319698095321655} 02/25/2022 00:45:44 - INFO - codeparrot_training - Step 17907: {'lr': 0.0003763155148720791, 'samples': 9168896, 'steps': 17907, 'loss/train': 1.7362861633300781} 02/25/2022 00:45:48 - INFO - codeparrot_training - Step 17908: {'lr': 0.00037630139435401055, 'samples': 9169408, 'steps': 17908, 'loss/train': 2.1172192096710205} 02/25/2022 00:45:53 - INFO - codeparrot_training - Step 17909: {'lr': 0.000376287273294907, 'samples': 9169920, 'steps': 17909, 'loss/train': 0.4598305821418762} 02/25/2022 00:45:57 - INFO - codeparrot_training - Step 17910: {'lr': 0.0003762731516948288, 'samples': 9170432, 'steps': 17910, 'loss/train': 2.358130693435669} 02/25/2022 00:46:02 - INFO - codeparrot_training - Step 17911: {'lr': 0.00037625902955383664, 'samples': 9170944, 'steps': 17911, 'loss/train': 1.4387531280517578} 02/25/2022 00:46:06 - INFO - codeparrot_training - Step 17912: {'lr': 0.0003762449068719907, 'samples': 9171456, 'steps': 17912, 'loss/train': 2.5760319232940674} 02/25/2022 00:46:12 - INFO - codeparrot_training - Step 17913: {'lr': 0.0003762307836493518, 'samples': 9171968, 'steps': 17913, 'loss/train': 1.6828365325927734} 02/25/2022 00:46:15 - INFO - codeparrot_training - Step 17914: {'lr': 0.00037621665988598024, 'samples': 9172480, 'steps': 17914, 'loss/train': 1.8636233806610107} 02/25/2022 00:46:21 - INFO - codeparrot_training - Step 17915: {'lr': 0.0003762025355819366, 'samples': 9172992, 'steps': 17915, 'loss/train': 2.0677895545959473} 02/25/2022 00:46:24 - INFO - codeparrot_training - Step 17916: {'lr': 0.0003761884107372814, 'samples': 9173504, 'steps': 17916, 'loss/train': 2.408090591430664} 02/25/2022 00:46:30 - INFO - codeparrot_training - Step 17917: {'lr': 0.0003761742853520751, 'samples': 9174016, 'steps': 17917, 'loss/train': 1.8922818899154663} 02/25/2022 00:46:34 - INFO - codeparrot_training - Step 17918: {'lr': 0.00037616015942637824, 'samples': 9174528, 'steps': 17918, 'loss/train': 1.2953591346740723} 02/25/2022 00:46:39 - INFO - codeparrot_training - Step 17919: {'lr': 0.0003761460329602513, 'samples': 9175040, 'steps': 17919, 'loss/train': 1.213797688484192} 02/25/2022 00:46:43 - INFO - codeparrot_training - Step 17920: {'lr': 0.0003761319059537548, 'samples': 9175552, 'steps': 17920, 'loss/train': 2.525622606277466} 02/25/2022 00:46:49 - INFO - codeparrot_training - Step 17921: {'lr': 0.0003761177784069493, 'samples': 9176064, 'steps': 17921, 'loss/train': 2.7263717651367188} 02/25/2022 00:46:52 - INFO - codeparrot_training - Step 17922: {'lr': 0.00037610365031989524, 'samples': 9176576, 'steps': 17922, 'loss/train': 1.6838401556015015} 02/25/2022 00:46:58 - INFO - codeparrot_training - Step 17923: {'lr': 0.0003760895216926532, 'samples': 9177088, 'steps': 17923, 'loss/train': 2.4757680892944336} 02/25/2022 00:47:01 - INFO - codeparrot_training - Step 17924: {'lr': 0.0003760753925252838, 'samples': 9177600, 'steps': 17924, 'loss/train': 1.7989904880523682} 02/25/2022 00:47:07 - INFO - codeparrot_training - Step 17925: {'lr': 0.00037606126281784725, 'samples': 9178112, 'steps': 17925, 'loss/train': 1.7875397205352783} 02/25/2022 00:47:10 - INFO - codeparrot_training - Step 17926: {'lr': 0.0003760471325704045, 'samples': 9178624, 'steps': 17926, 'loss/train': 2.589329719543457} 02/25/2022 00:47:16 - INFO - codeparrot_training - Step 17927: {'lr': 0.0003760330017830157, 'samples': 9179136, 'steps': 17927, 'loss/train': 1.5632318258285522} 02/25/2022 00:47:19 - INFO - codeparrot_training - Step 17928: {'lr': 0.00037601887045574155, 'samples': 9179648, 'steps': 17928, 'loss/train': 2.014723300933838} 02/25/2022 00:47:25 - INFO - codeparrot_training - Step 17929: {'lr': 0.0003760047385886426, 'samples': 9180160, 'steps': 17929, 'loss/train': 2.1561315059661865} 02/25/2022 00:47:28 - INFO - codeparrot_training - Step 17930: {'lr': 0.0003759906061817794, 'samples': 9180672, 'steps': 17930, 'loss/train': 2.223832845687866} 02/25/2022 00:47:34 - INFO - codeparrot_training - Step 17931: {'lr': 0.00037597647323521234, 'samples': 9181184, 'steps': 17931, 'loss/train': 2.039451837539673} 02/25/2022 00:47:38 - INFO - codeparrot_training - Step 17932: {'lr': 0.0003759623397490022, 'samples': 9181696, 'steps': 17932, 'loss/train': 1.986750602722168} 02/25/2022 00:47:44 - INFO - codeparrot_training - Step 17933: {'lr': 0.00037594820572320933, 'samples': 9182208, 'steps': 17933, 'loss/train': 2.7664713859558105} 02/25/2022 00:47:48 - INFO - codeparrot_training - Step 17934: {'lr': 0.0003759340711578944, 'samples': 9182720, 'steps': 17934, 'loss/train': 1.9041202068328857} 02/25/2022 00:47:53 - INFO - codeparrot_training - Step 17935: {'lr': 0.0003759199360531178, 'samples': 9183232, 'steps': 17935, 'loss/train': 1.7459298372268677} 02/25/2022 00:47:57 - INFO - codeparrot_training - Step 17936: {'lr': 0.00037590580040894024, 'samples': 9183744, 'steps': 17936, 'loss/train': 2.398015260696411} 02/25/2022 00:48:02 - INFO - codeparrot_training - Step 17937: {'lr': 0.0003758916642254222, 'samples': 9184256, 'steps': 17937, 'loss/train': 2.96384334564209} 02/25/2022 00:48:05 - INFO - codeparrot_training - Step 17938: {'lr': 0.00037587752750262426, 'samples': 9184768, 'steps': 17938, 'loss/train': 2.4360480308532715} 02/25/2022 00:48:11 - INFO - codeparrot_training - Step 17939: {'lr': 0.00037586339024060696, 'samples': 9185280, 'steps': 17939, 'loss/train': 2.36801815032959} 02/25/2022 00:48:15 - INFO - codeparrot_training - Step 17940: {'lr': 0.0003758492524394308, 'samples': 9185792, 'steps': 17940, 'loss/train': 0.4478470981121063} 02/25/2022 00:48:20 - INFO - codeparrot_training - Step 17941: {'lr': 0.0003758351140991565, 'samples': 9186304, 'steps': 17941, 'loss/train': 1.4435036182403564} 02/25/2022 00:48:24 - INFO - codeparrot_training - Step 17942: {'lr': 0.0003758209752198444, 'samples': 9186816, 'steps': 17942, 'loss/train': 1.9797041416168213} 02/25/2022 00:48:29 - INFO - codeparrot_training - Step 17943: {'lr': 0.0003758068358015553, 'samples': 9187328, 'steps': 17943, 'loss/train': 1.366119146347046} 02/25/2022 00:48:33 - INFO - codeparrot_training - Step 17944: {'lr': 0.0003757926958443496, 'samples': 9187840, 'steps': 17944, 'loss/train': 2.4365100860595703} 02/25/2022 00:48:39 - INFO - codeparrot_training - Step 17945: {'lr': 0.000375778555348288, 'samples': 9188352, 'steps': 17945, 'loss/train': 1.6386650800704956} 02/25/2022 00:48:42 - INFO - codeparrot_training - Step 17946: {'lr': 0.000375764414313431, 'samples': 9188864, 'steps': 17946, 'loss/train': 1.8946306705474854} 02/25/2022 00:48:48 - INFO - codeparrot_training - Step 17947: {'lr': 0.0003757502727398391, 'samples': 9189376, 'steps': 17947, 'loss/train': 2.3807075023651123} 02/25/2022 00:48:51 - INFO - codeparrot_training - Step 17948: {'lr': 0.00037573613062757304, 'samples': 9189888, 'steps': 17948, 'loss/train': 1.8239479064941406} 02/25/2022 00:48:57 - INFO - codeparrot_training - Step 17949: {'lr': 0.0003757219879766933, 'samples': 9190400, 'steps': 17949, 'loss/train': 1.410290002822876} 02/25/2022 00:49:02 - INFO - codeparrot_training - Step 17950: {'lr': 0.00037570784478726057, 'samples': 9190912, 'steps': 17950, 'loss/train': 1.792197346687317} 02/25/2022 00:49:06 - INFO - codeparrot_training - Step 17951: {'lr': 0.00037569370105933523, 'samples': 9191424, 'steps': 17951, 'loss/train': 2.7010374069213867} 02/25/2022 00:49:11 - INFO - codeparrot_training - Step 17952: {'lr': 0.00037567955679297806, 'samples': 9191936, 'steps': 17952, 'loss/train': 1.2868130207061768} 02/25/2022 00:49:15 - INFO - codeparrot_training - Step 17953: {'lr': 0.0003756654119882496, 'samples': 9192448, 'steps': 17953, 'loss/train': 2.4192488193511963} 02/25/2022 00:49:18 - INFO - codeparrot_training - Step 17954: {'lr': 0.0003756512666452103, 'samples': 9192960, 'steps': 17954, 'loss/train': 1.679328203201294} 02/25/2022 00:49:25 - INFO - codeparrot_training - Step 17955: {'lr': 0.0003756371207639209, 'samples': 9193472, 'steps': 17955, 'loss/train': 1.953917145729065} 02/25/2022 00:49:28 - INFO - codeparrot_training - Step 17956: {'lr': 0.00037562297434444203, 'samples': 9193984, 'steps': 17956, 'loss/train': 2.0183377265930176} 02/25/2022 00:49:34 - INFO - codeparrot_training - Step 17957: {'lr': 0.0003756088273868342, 'samples': 9194496, 'steps': 17957, 'loss/train': 2.3318753242492676} 02/25/2022 00:49:37 - INFO - codeparrot_training - Step 17958: {'lr': 0.00037559467989115806, 'samples': 9195008, 'steps': 17958, 'loss/train': 1.780102252960205} 02/25/2022 00:49:43 - INFO - codeparrot_training - Step 17959: {'lr': 0.00037558053185747416, 'samples': 9195520, 'steps': 17959, 'loss/train': 2.6560707092285156} 02/25/2022 00:49:48 - INFO - codeparrot_training - Step 17960: {'lr': 0.00037556638328584314, 'samples': 9196032, 'steps': 17960, 'loss/train': 1.8643437623977661} 02/25/2022 00:49:52 - INFO - codeparrot_training - Step 17961: {'lr': 0.00037555223417632565, 'samples': 9196544, 'steps': 17961, 'loss/train': 1.6963603496551514} 02/25/2022 00:49:57 - INFO - codeparrot_training - Step 17962: {'lr': 0.0003755380845289822, 'samples': 9197056, 'steps': 17962, 'loss/train': 1.8083080053329468} 02/25/2022 00:50:01 - INFO - codeparrot_training - Step 17963: {'lr': 0.0003755239343438735, 'samples': 9197568, 'steps': 17963, 'loss/train': 2.3275163173675537} 02/25/2022 00:50:07 - INFO - codeparrot_training - Step 17964: {'lr': 0.00037550978362106, 'samples': 9198080, 'steps': 17964, 'loss/train': 1.2991315126419067} 02/25/2022 00:50:10 - INFO - codeparrot_training - Step 17965: {'lr': 0.0003754956323606026, 'samples': 9198592, 'steps': 17965, 'loss/train': 1.908624291419983} 02/25/2022 00:50:16 - INFO - codeparrot_training - Step 17966: {'lr': 0.0003754814805625617, 'samples': 9199104, 'steps': 17966, 'loss/train': 2.5921530723571777} 02/25/2022 00:50:20 - INFO - codeparrot_training - Step 17967: {'lr': 0.00037546732822699803, 'samples': 9199616, 'steps': 17967, 'loss/train': 2.1310462951660156} 02/25/2022 00:50:25 - INFO - codeparrot_training - Step 17968: {'lr': 0.0003754531753539721, 'samples': 9200128, 'steps': 17968, 'loss/train': 1.7575037479400635} 02/25/2022 00:50:29 - INFO - codeparrot_training - Step 17969: {'lr': 0.0003754390219435446, 'samples': 9200640, 'steps': 17969, 'loss/train': 2.1864471435546875} 02/25/2022 00:50:34 - INFO - codeparrot_training - Step 17970: {'lr': 0.00037542486799577624, 'samples': 9201152, 'steps': 17970, 'loss/train': 0.6265259981155396} 02/25/2022 00:50:38 - INFO - codeparrot_training - Step 17971: {'lr': 0.00037541071351072746, 'samples': 9201664, 'steps': 17971, 'loss/train': 1.3466262817382812} 02/25/2022 00:50:43 - INFO - codeparrot_training - Step 17972: {'lr': 0.0003753965584884591, 'samples': 9202176, 'steps': 17972, 'loss/train': 1.4059536457061768} 02/25/2022 00:50:47 - INFO - codeparrot_training - Step 17973: {'lr': 0.00037538240292903167, 'samples': 9202688, 'steps': 17973, 'loss/train': 1.6436214447021484} 02/25/2022 00:50:52 - INFO - codeparrot_training - Step 17974: {'lr': 0.0003753682468325059, 'samples': 9203200, 'steps': 17974, 'loss/train': 3.1578054428100586} 02/25/2022 00:50:56 - INFO - codeparrot_training - Step 17975: {'lr': 0.0003753540901989422, 'samples': 9203712, 'steps': 17975, 'loss/train': 1.153119683265686} 02/25/2022 00:51:01 - INFO - codeparrot_training - Step 17976: {'lr': 0.00037533993302840153, 'samples': 9204224, 'steps': 17976, 'loss/train': 1.9482473134994507} 02/25/2022 00:51:05 - INFO - codeparrot_training - Step 17977: {'lr': 0.00037532577532094436, 'samples': 9204736, 'steps': 17977, 'loss/train': 2.373196840286255} 02/25/2022 00:51:11 - INFO - codeparrot_training - Step 17978: {'lr': 0.00037531161707663136, 'samples': 9205248, 'steps': 17978, 'loss/train': 2.9937961101531982} 02/25/2022 00:51:14 - INFO - codeparrot_training - Step 17979: {'lr': 0.0003752974582955232, 'samples': 9205760, 'steps': 17979, 'loss/train': 1.9035017490386963} 02/25/2022 00:51:20 - INFO - codeparrot_training - Step 17980: {'lr': 0.0003752832989776804, 'samples': 9206272, 'steps': 17980, 'loss/train': 2.5906925201416016} 02/25/2022 00:51:23 - INFO - codeparrot_training - Step 17981: {'lr': 0.0003752691391231639, 'samples': 9206784, 'steps': 17981, 'loss/train': 2.4567105770111084} 02/25/2022 00:51:29 - INFO - codeparrot_training - Step 17982: {'lr': 0.00037525497873203405, 'samples': 9207296, 'steps': 17982, 'loss/train': 2.204061985015869} 02/25/2022 00:51:32 - INFO - codeparrot_training - Step 17983: {'lr': 0.0003752408178043518, 'samples': 9207808, 'steps': 17983, 'loss/train': 1.612808108329773} 02/25/2022 00:51:38 - INFO - codeparrot_training - Step 17984: {'lr': 0.0003752266563401775, 'samples': 9208320, 'steps': 17984, 'loss/train': 1.8292522430419922} 02/25/2022 00:51:42 - INFO - codeparrot_training - Step 17985: {'lr': 0.00037521249433957203, 'samples': 9208832, 'steps': 17985, 'loss/train': 1.753325343132019} 02/25/2022 00:51:47 - INFO - codeparrot_training - Step 17986: {'lr': 0.000375198331802596, 'samples': 9209344, 'steps': 17986, 'loss/train': 0.6351202130317688} 02/25/2022 00:51:50 - INFO - codeparrot_training - Step 17987: {'lr': 0.00037518416872931007, 'samples': 9209856, 'steps': 17987, 'loss/train': 2.168978691101074} 02/25/2022 00:51:56 - INFO - codeparrot_training - Step 17988: {'lr': 0.00037517000511977486, 'samples': 9210368, 'steps': 17988, 'loss/train': 1.2864890098571777} 02/25/2022 00:51:59 - INFO - codeparrot_training - Step 17989: {'lr': 0.00037515584097405115, 'samples': 9210880, 'steps': 17989, 'loss/train': 3.4598395824432373} 02/25/2022 00:52:06 - INFO - codeparrot_training - Step 17990: {'lr': 0.00037514167629219955, 'samples': 9211392, 'steps': 17990, 'loss/train': 1.6512317657470703} 02/25/2022 00:52:09 - INFO - codeparrot_training - Step 17991: {'lr': 0.0003751275110742807, 'samples': 9211904, 'steps': 17991, 'loss/train': 2.2843728065490723} 02/25/2022 00:52:15 - INFO - codeparrot_training - Step 17992: {'lr': 0.00037511334532035537, 'samples': 9212416, 'steps': 17992, 'loss/train': 0.32034048438072205} 02/25/2022 00:52:18 - INFO - codeparrot_training - Step 17993: {'lr': 0.00037509917903048417, 'samples': 9212928, 'steps': 17993, 'loss/train': 2.830625295639038} 02/25/2022 00:52:24 - INFO - codeparrot_training - Step 17994: {'lr': 0.00037508501220472783, 'samples': 9213440, 'steps': 17994, 'loss/train': 3.263338565826416} 02/25/2022 00:52:27 - INFO - codeparrot_training - Step 17995: {'lr': 0.000375070844843147, 'samples': 9213952, 'steps': 17995, 'loss/train': 0.6421509981155396} 02/25/2022 00:52:33 - INFO - codeparrot_training - Step 17996: {'lr': 0.00037505667694580244, 'samples': 9214464, 'steps': 17996, 'loss/train': 2.5737671852111816} 02/25/2022 00:52:36 - INFO - codeparrot_training - Step 17997: {'lr': 0.00037504250851275466, 'samples': 9214976, 'steps': 17997, 'loss/train': 1.7033635377883911} 02/25/2022 00:52:42 - INFO - codeparrot_training - Step 17998: {'lr': 0.0003750283395440647, 'samples': 9215488, 'steps': 17998, 'loss/train': 1.7467429637908936} 02/25/2022 00:52:45 - INFO - codeparrot_training - Step 17999: {'lr': 0.0003750141700397928, 'samples': 9216000, 'steps': 17999, 'loss/train': 1.3792665004730225} 02/25/2022 00:52:45 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 00:53:02 - WARNING - huggingface_hub.repository - Several commits (18) will be pushed upstream. 02/25/2022 00:53:02 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 00:53:36 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 4fad2a5..4d2b52a floral-grass-11 -> floral-grass-11 02/25/2022 00:53:44 - INFO - codeparrot_training - Step 18000: {'lr': 0.000375, 'samples': 9216512, 'steps': 18000, 'loss/train': 2.864016532897949} 02/25/2022 00:53:47 - INFO - codeparrot_training - Step 18001: {'lr': 0.0003749858294247469, 'samples': 9217024, 'steps': 18001, 'loss/train': 0.5618553757667542} 02/25/2022 00:53:53 - INFO - codeparrot_training - Step 18002: {'lr': 0.0003749716583140942, 'samples': 9217536, 'steps': 18002, 'loss/train': 1.8189059495925903} 02/25/2022 00:53:56 - INFO - codeparrot_training - Step 18003: {'lr': 0.00037495748666810255, 'samples': 9218048, 'steps': 18003, 'loss/train': 2.4676153659820557} 02/25/2022 00:54:02 - INFO - codeparrot_training - Step 18004: {'lr': 0.0003749433144868327, 'samples': 9218560, 'steps': 18004, 'loss/train': 2.285482406616211} 02/25/2022 00:54:05 - INFO - codeparrot_training - Step 18005: {'lr': 0.00037492914177034544, 'samples': 9219072, 'steps': 18005, 'loss/train': 0.5833332538604736} 02/25/2022 00:54:11 - INFO - codeparrot_training - Step 18006: {'lr': 0.00037491496851870134, 'samples': 9219584, 'steps': 18006, 'loss/train': 2.044358491897583} 02/25/2022 00:54:14 - INFO - codeparrot_training - Step 18007: {'lr': 0.0003749007947319612, 'samples': 9220096, 'steps': 18007, 'loss/train': 1.815633773803711} 02/25/2022 00:54:20 - INFO - codeparrot_training - Step 18008: {'lr': 0.00037488662041018574, 'samples': 9220608, 'steps': 18008, 'loss/train': 1.6437115669250488} 02/25/2022 00:54:23 - INFO - codeparrot_training - Step 18009: {'lr': 0.0003748724455534356, 'samples': 9221120, 'steps': 18009, 'loss/train': 1.8296072483062744} 02/25/2022 00:54:30 - INFO - codeparrot_training - Step 18010: {'lr': 0.0003748582701617716, 'samples': 9221632, 'steps': 18010, 'loss/train': 0.8397614359855652} 02/25/2022 00:54:33 - INFO - codeparrot_training - Step 18011: {'lr': 0.00037484409423525446, 'samples': 9222144, 'steps': 18011, 'loss/train': 2.213873863220215} 02/25/2022 00:54:39 - INFO - codeparrot_training - Step 18012: {'lr': 0.00037482991777394476, 'samples': 9222656, 'steps': 18012, 'loss/train': 1.686563491821289} 02/25/2022 00:54:42 - INFO - codeparrot_training - Step 18013: {'lr': 0.00037481574077790334, 'samples': 9223168, 'steps': 18013, 'loss/train': 2.7907283306121826} 02/25/2022 00:54:48 - INFO - codeparrot_training - Step 18014: {'lr': 0.00037480156324719093, 'samples': 9223680, 'steps': 18014, 'loss/train': 1.6545108556747437} 02/25/2022 00:54:51 - INFO - codeparrot_training - Step 18015: {'lr': 0.00037478738518186835, 'samples': 9224192, 'steps': 18015, 'loss/train': 2.19451904296875} 02/25/2022 00:54:56 - INFO - codeparrot_training - Step 18016: {'lr': 0.00037477320658199615, 'samples': 9224704, 'steps': 18016, 'loss/train': 1.6612696647644043} 02/25/2022 00:55:00 - INFO - codeparrot_training - Step 18017: {'lr': 0.0003747590274476351, 'samples': 9225216, 'steps': 18017, 'loss/train': 1.947067141532898} 02/25/2022 00:55:05 - INFO - codeparrot_training - Step 18018: {'lr': 0.0003747448477788461, 'samples': 9225728, 'steps': 18018, 'loss/train': 1.3791682720184326} 02/25/2022 00:55:09 - INFO - codeparrot_training - Step 18019: {'lr': 0.0003747306675756898, 'samples': 9226240, 'steps': 18019, 'loss/train': 2.568746328353882} 02/25/2022 00:55:15 - INFO - codeparrot_training - Step 18020: {'lr': 0.00037471648683822683, 'samples': 9226752, 'steps': 18020, 'loss/train': 2.169994831085205} 02/25/2022 00:55:18 - INFO - codeparrot_training - Step 18021: {'lr': 0.00037470230556651814, 'samples': 9227264, 'steps': 18021, 'loss/train': 2.9739251136779785} 02/25/2022 00:55:23 - INFO - codeparrot_training - Step 18022: {'lr': 0.00037468812376062423, 'samples': 9227776, 'steps': 18022, 'loss/train': 0.4943702220916748} 02/25/2022 00:55:27 - INFO - codeparrot_training - Step 18023: {'lr': 0.00037467394142060614, 'samples': 9228288, 'steps': 18023, 'loss/train': 0.43165501952171326} 02/25/2022 00:55:32 - INFO - codeparrot_training - Step 18024: {'lr': 0.0003746597585465243, 'samples': 9228800, 'steps': 18024, 'loss/train': 2.010676145553589} 02/25/2022 00:55:36 - INFO - codeparrot_training - Step 18025: {'lr': 0.00037464557513843975, 'samples': 9229312, 'steps': 18025, 'loss/train': 1.948112964630127} 02/25/2022 00:55:42 - INFO - codeparrot_training - Step 18026: {'lr': 0.0003746313911964132, 'samples': 9229824, 'steps': 18026, 'loss/train': 1.3395642042160034} 02/25/2022 00:55:46 - INFO - codeparrot_training - Step 18027: {'lr': 0.00037461720672050524, 'samples': 9230336, 'steps': 18027, 'loss/train': 0.8749566078186035} 02/25/2022 00:55:51 - INFO - codeparrot_training - Step 18028: {'lr': 0.0003746030217107768, 'samples': 9230848, 'steps': 18028, 'loss/train': 2.4887499809265137} 02/25/2022 00:55:55 - INFO - codeparrot_training - Step 18029: {'lr': 0.0003745888361672885, 'samples': 9231360, 'steps': 18029, 'loss/train': 0.6735855340957642} 02/25/2022 00:56:00 - INFO - codeparrot_training - Step 18030: {'lr': 0.00037457465009010124, 'samples': 9231872, 'steps': 18030, 'loss/train': 2.5045032501220703} 02/25/2022 00:56:04 - INFO - codeparrot_training - Step 18031: {'lr': 0.00037456046347927576, 'samples': 9232384, 'steps': 18031, 'loss/train': 2.466980457305908} 02/25/2022 00:56:09 - INFO - codeparrot_training - Step 18032: {'lr': 0.0003745462763348727, 'samples': 9232896, 'steps': 18032, 'loss/train': 1.8540087938308716} 02/25/2022 00:56:13 - INFO - codeparrot_training - Step 18033: {'lr': 0.00037453208865695305, 'samples': 9233408, 'steps': 18033, 'loss/train': 1.4267619848251343} 02/25/2022 00:56:18 - INFO - codeparrot_training - Step 18034: {'lr': 0.0003745179004455774, 'samples': 9233920, 'steps': 18034, 'loss/train': 1.6516650915145874} 02/25/2022 00:56:22 - INFO - codeparrot_training - Step 18035: {'lr': 0.00037450371170080673, 'samples': 9234432, 'steps': 18035, 'loss/train': 2.1614670753479004} 02/25/2022 00:56:28 - INFO - codeparrot_training - Step 18036: {'lr': 0.00037448952242270155, 'samples': 9234944, 'steps': 18036, 'loss/train': 2.139948844909668} 02/25/2022 00:56:31 - INFO - codeparrot_training - Step 18037: {'lr': 0.0003744753326113229, 'samples': 9235456, 'steps': 18037, 'loss/train': 2.0563716888427734} 02/25/2022 00:56:37 - INFO - codeparrot_training - Step 18038: {'lr': 0.00037446114226673136, 'samples': 9235968, 'steps': 18038, 'loss/train': 1.928979754447937} 02/25/2022 00:56:42 - INFO - codeparrot_training - Step 18039: {'lr': 0.00037444695138898784, 'samples': 9236480, 'steps': 18039, 'loss/train': 3.129032850265503} 02/25/2022 00:56:46 - INFO - codeparrot_training - Step 18040: {'lr': 0.00037443275997815306, 'samples': 9236992, 'steps': 18040, 'loss/train': 1.6235578060150146} 02/25/2022 00:56:51 - INFO - codeparrot_training - Step 18041: {'lr': 0.00037441856803428785, 'samples': 9237504, 'steps': 18041, 'loss/train': 1.7870166301727295} 02/25/2022 00:56:55 - INFO - codeparrot_training - Step 18042: {'lr': 0.0003744043755574531, 'samples': 9238016, 'steps': 18042, 'loss/train': 2.188650608062744} 02/25/2022 00:57:00 - INFO - codeparrot_training - Step 18043: {'lr': 0.0003743901825477094, 'samples': 9238528, 'steps': 18043, 'loss/train': 1.6643199920654297} 02/25/2022 00:57:04 - INFO - codeparrot_training - Step 18044: {'lr': 0.0003743759890051177, 'samples': 9239040, 'steps': 18044, 'loss/train': 0.06709722429513931} 02/25/2022 00:57:10 - INFO - codeparrot_training - Step 18045: {'lr': 0.00037436179492973876, 'samples': 9239552, 'steps': 18045, 'loss/train': 1.8383413553237915} 02/25/2022 00:57:14 - INFO - codeparrot_training - Step 18046: {'lr': 0.0003743476003216334, 'samples': 9240064, 'steps': 18046, 'loss/train': 1.3675178289413452} 02/25/2022 00:57:19 - INFO - codeparrot_training - Step 18047: {'lr': 0.00037433340518086235, 'samples': 9240576, 'steps': 18047, 'loss/train': 2.4452977180480957} 02/25/2022 00:57:22 - INFO - codeparrot_training - Step 18048: {'lr': 0.0003743192095074866, 'samples': 9241088, 'steps': 18048, 'loss/train': 0.7475156188011169} 02/25/2022 00:57:28 - INFO - codeparrot_training - Step 18049: {'lr': 0.0003743050133015666, 'samples': 9241600, 'steps': 18049, 'loss/train': 1.7063541412353516} 02/25/2022 00:57:31 - INFO - codeparrot_training - Step 18050: {'lr': 0.0003742908165631636, 'samples': 9242112, 'steps': 18050, 'loss/train': 1.7238842248916626} 02/25/2022 00:57:37 - INFO - codeparrot_training - Step 18051: {'lr': 0.00037427661929233816, 'samples': 9242624, 'steps': 18051, 'loss/train': 1.30217707157135} 02/25/2022 00:57:40 - INFO - codeparrot_training - Step 18052: {'lr': 0.00037426242148915113, 'samples': 9243136, 'steps': 18052, 'loss/train': 2.37275767326355} 02/25/2022 00:57:46 - INFO - codeparrot_training - Step 18053: {'lr': 0.0003742482231536633, 'samples': 9243648, 'steps': 18053, 'loss/train': 2.9769768714904785} 02/25/2022 00:57:50 - INFO - codeparrot_training - Step 18054: {'lr': 0.00037423402428593555, 'samples': 9244160, 'steps': 18054, 'loss/train': 2.0993216037750244} 02/25/2022 00:57:56 - INFO - codeparrot_training - Step 18055: {'lr': 0.00037421982488602875, 'samples': 9244672, 'steps': 18055, 'loss/train': 0.5246959328651428} 02/25/2022 00:57:59 - INFO - codeparrot_training - Step 18056: {'lr': 0.0003742056249540036, 'samples': 9245184, 'steps': 18056, 'loss/train': 0.6965020895004272} 02/25/2022 00:58:05 - INFO - codeparrot_training - Step 18057: {'lr': 0.00037419142448992103, 'samples': 9245696, 'steps': 18057, 'loss/train': 1.2578572034835815} 02/25/2022 00:58:08 - INFO - codeparrot_training - Step 18058: {'lr': 0.0003741772234938418, 'samples': 9246208, 'steps': 18058, 'loss/train': 1.4348853826522827} 02/25/2022 00:58:14 - INFO - codeparrot_training - Step 18059: {'lr': 0.00037416302196582684, 'samples': 9246720, 'steps': 18059, 'loss/train': 2.3590095043182373} 02/25/2022 00:58:17 - INFO - codeparrot_training - Step 18060: {'lr': 0.00037414881990593686, 'samples': 9247232, 'steps': 18060, 'loss/train': 2.1980643272399902} 02/25/2022 00:58:23 - INFO - codeparrot_training - Step 18061: {'lr': 0.0003741346173142328, 'samples': 9247744, 'steps': 18061, 'loss/train': 1.9618237018585205} 02/25/2022 00:58:26 - INFO - codeparrot_training - Step 18062: {'lr': 0.0003741204141907754, 'samples': 9248256, 'steps': 18062, 'loss/train': 1.7716025114059448} 02/25/2022 00:58:32 - INFO - codeparrot_training - Step 18063: {'lr': 0.00037410621053562563, 'samples': 9248768, 'steps': 18063, 'loss/train': 1.532081127166748} 02/25/2022 00:58:35 - INFO - codeparrot_training - Step 18064: {'lr': 0.0003740920063488442, 'samples': 9249280, 'steps': 18064, 'loss/train': 0.5840127468109131} 02/25/2022 00:58:41 - INFO - codeparrot_training - Step 18065: {'lr': 0.0003740778016304921, 'samples': 9249792, 'steps': 18065, 'loss/train': 2.6977975368499756} 02/25/2022 00:58:44 - INFO - codeparrot_training - Step 18066: {'lr': 0.00037406359638063005, 'samples': 9250304, 'steps': 18066, 'loss/train': 1.6146961450576782} 02/25/2022 00:58:50 - INFO - codeparrot_training - Step 18067: {'lr': 0.00037404939059931884, 'samples': 9250816, 'steps': 18067, 'loss/train': 2.1225221157073975} 02/25/2022 00:58:53 - INFO - codeparrot_training - Step 18068: {'lr': 0.0003740351842866196, 'samples': 9251328, 'steps': 18068, 'loss/train': 1.1155716180801392} 02/25/2022 00:58:59 - INFO - codeparrot_training - Step 18069: {'lr': 0.0003740209774425929, 'samples': 9251840, 'steps': 18069, 'loss/train': 1.7514163255691528} 02/25/2022 00:59:02 - INFO - codeparrot_training - Step 18070: {'lr': 0.00037400677006729977, 'samples': 9252352, 'steps': 18070, 'loss/train': 2.209989547729492} 02/25/2022 00:59:09 - INFO - codeparrot_training - Step 18071: {'lr': 0.000373992562160801, 'samples': 9252864, 'steps': 18071, 'loss/train': 2.072781562805176} 02/25/2022 00:59:12 - INFO - codeparrot_training - Step 18072: {'lr': 0.0003739783537231575, 'samples': 9253376, 'steps': 18072, 'loss/train': 3.9846749305725098} 02/25/2022 00:59:18 - INFO - codeparrot_training - Step 18073: {'lr': 0.0003739641447544301, 'samples': 9253888, 'steps': 18073, 'loss/train': 0.84217768907547} 02/25/2022 00:59:21 - INFO - codeparrot_training - Step 18074: {'lr': 0.0003739499352546796, 'samples': 9254400, 'steps': 18074, 'loss/train': 2.2811882495880127} 02/25/2022 00:59:27 - INFO - codeparrot_training - Step 18075: {'lr': 0.00037393572522396697, 'samples': 9254912, 'steps': 18075, 'loss/train': 9.352462768554688} 02/25/2022 00:59:30 - INFO - codeparrot_training - Step 18076: {'lr': 0.00037392151466235303, 'samples': 9255424, 'steps': 18076, 'loss/train': 2.6928279399871826} 02/25/2022 00:59:36 - INFO - codeparrot_training - Step 18077: {'lr': 0.00037390730356989864, 'samples': 9255936, 'steps': 18077, 'loss/train': 1.6681896448135376} 02/25/2022 00:59:39 - INFO - codeparrot_training - Step 18078: {'lr': 0.0003738930919466648, 'samples': 9256448, 'steps': 18078, 'loss/train': 2.294260025024414} 02/25/2022 00:59:45 - INFO - codeparrot_training - Step 18079: {'lr': 0.0003738788797927122, 'samples': 9256960, 'steps': 18079, 'loss/train': 1.343259334564209} 02/25/2022 00:59:48 - INFO - codeparrot_training - Step 18080: {'lr': 0.0003738646671081019, 'samples': 9257472, 'steps': 18080, 'loss/train': 1.635291576385498} 02/25/2022 00:59:54 - INFO - codeparrot_training - Step 18081: {'lr': 0.0003738504538928947, 'samples': 9257984, 'steps': 18081, 'loss/train': 1.811547040939331} 02/25/2022 00:59:57 - INFO - codeparrot_training - Step 18082: {'lr': 0.0003738362401471514, 'samples': 9258496, 'steps': 18082, 'loss/train': 1.9558956623077393} 02/25/2022 01:00:03 - INFO - codeparrot_training - Step 18083: {'lr': 0.00037382202587093305, 'samples': 9259008, 'steps': 18083, 'loss/train': 1.6220651865005493} 02/25/2022 01:00:06 - INFO - codeparrot_training - Step 18084: {'lr': 0.0003738078110643004, 'samples': 9259520, 'steps': 18084, 'loss/train': 1.7357509136199951} 02/25/2022 01:00:19 - INFO - codeparrot_training - Step 18085: {'lr': 0.0003737935957273144, 'samples': 9260032, 'steps': 18085, 'loss/train': 1.1062289476394653} 02/25/2022 01:00:22 - INFO - codeparrot_training - Step 18086: {'lr': 0.000373779379860036, 'samples': 9260544, 'steps': 18086, 'loss/train': 2.552191734313965} 02/25/2022 01:00:28 - INFO - codeparrot_training - Step 18087: {'lr': 0.000373765163462526, 'samples': 9261056, 'steps': 18087, 'loss/train': 1.418000340461731} 02/25/2022 01:00:31 - INFO - codeparrot_training - Step 18088: {'lr': 0.0003737509465348453, 'samples': 9261568, 'steps': 18088, 'loss/train': 1.6852855682373047} 02/25/2022 01:00:36 - INFO - codeparrot_training - Step 18089: {'lr': 0.0003737367290770549, 'samples': 9262080, 'steps': 18089, 'loss/train': 1.9217792749404907} 02/25/2022 01:00:40 - INFO - codeparrot_training - Step 18090: {'lr': 0.0003737225110892156, 'samples': 9262592, 'steps': 18090, 'loss/train': 0.3693554401397705} 02/25/2022 01:00:46 - INFO - codeparrot_training - Step 18091: {'lr': 0.0003737082925713884, 'samples': 9263104, 'steps': 18091, 'loss/train': 3.2785496711730957} 02/25/2022 01:00:49 - INFO - codeparrot_training - Step 18092: {'lr': 0.00037369407352363417, 'samples': 9263616, 'steps': 18092, 'loss/train': 1.668378472328186} 02/25/2022 01:00:55 - INFO - codeparrot_training - Step 18093: {'lr': 0.0003736798539460138, 'samples': 9264128, 'steps': 18093, 'loss/train': 2.0754849910736084} 02/25/2022 01:00:58 - INFO - codeparrot_training - Step 18094: {'lr': 0.00037366563383858814, 'samples': 9264640, 'steps': 18094, 'loss/train': 1.7872614860534668} 02/25/2022 01:01:06 - INFO - codeparrot_training - Step 18095: {'lr': 0.0003736514132014182, 'samples': 9265152, 'steps': 18095, 'loss/train': 1.4480342864990234} 02/25/2022 01:01:09 - INFO - codeparrot_training - Step 18096: {'lr': 0.0003736371920345649, 'samples': 9265664, 'steps': 18096, 'loss/train': 2.033862829208374} 02/25/2022 01:01:15 - INFO - codeparrot_training - Step 18097: {'lr': 0.0003736229703380891, 'samples': 9266176, 'steps': 18097, 'loss/train': 1.2324012517929077} 02/25/2022 01:01:18 - INFO - codeparrot_training - Step 18098: {'lr': 0.0003736087481120518, 'samples': 9266688, 'steps': 18098, 'loss/train': 2.1741888523101807} 02/25/2022 01:01:24 - INFO - codeparrot_training - Step 18099: {'lr': 0.0003735945253565138, 'samples': 9267200, 'steps': 18099, 'loss/train': 1.3801395893096924} 02/25/2022 01:01:27 - INFO - codeparrot_training - Step 18100: {'lr': 0.00037358030207153617, 'samples': 9267712, 'steps': 18100, 'loss/train': 2.210556983947754} 02/25/2022 01:01:33 - INFO - codeparrot_training - Step 18101: {'lr': 0.0003735660782571797, 'samples': 9268224, 'steps': 18101, 'loss/train': 2.1449480056762695} 02/25/2022 01:01:36 - INFO - codeparrot_training - Step 18102: {'lr': 0.00037355185391350546, 'samples': 9268736, 'steps': 18102, 'loss/train': 1.519236445426941} 02/25/2022 01:01:42 - INFO - codeparrot_training - Step 18103: {'lr': 0.00037353762904057425, 'samples': 9269248, 'steps': 18103, 'loss/train': 2.09230899810791} 02/25/2022 01:01:45 - INFO - codeparrot_training - Step 18104: {'lr': 0.00037352340363844706, 'samples': 9269760, 'steps': 18104, 'loss/train': 1.587156057357788} 02/25/2022 01:01:51 - INFO - codeparrot_training - Step 18105: {'lr': 0.0003735091777071849, 'samples': 9270272, 'steps': 18105, 'loss/train': 1.9107987880706787} 02/25/2022 01:01:54 - INFO - codeparrot_training - Step 18106: {'lr': 0.0003734949512468486, 'samples': 9270784, 'steps': 18106, 'loss/train': 1.8355761766433716} 02/25/2022 01:02:02 - INFO - codeparrot_training - Step 18107: {'lr': 0.0003734807242574991, 'samples': 9271296, 'steps': 18107, 'loss/train': 2.575145721435547} 02/25/2022 01:02:05 - INFO - codeparrot_training - Step 18108: {'lr': 0.0003734664967391975, 'samples': 9271808, 'steps': 18108, 'loss/train': 2.6048264503479004} 02/25/2022 01:02:10 - INFO - codeparrot_training - Step 18109: {'lr': 0.0003734522686920045, 'samples': 9272320, 'steps': 18109, 'loss/train': 1.0448098182678223} 02/25/2022 01:02:14 - INFO - codeparrot_training - Step 18110: {'lr': 0.00037343804011598125, 'samples': 9272832, 'steps': 18110, 'loss/train': 2.380887508392334} 02/25/2022 01:02:20 - INFO - codeparrot_training - Step 18111: {'lr': 0.0003734238110111886, 'samples': 9273344, 'steps': 18111, 'loss/train': 2.5484042167663574} 02/25/2022 01:02:23 - INFO - codeparrot_training - Step 18112: {'lr': 0.0003734095813776876, 'samples': 9273856, 'steps': 18112, 'loss/train': 1.769220232963562} 02/25/2022 01:02:28 - INFO - codeparrot_training - Step 18113: {'lr': 0.00037339535121553907, 'samples': 9274368, 'steps': 18113, 'loss/train': 1.68701171875} 02/25/2022 01:02:32 - INFO - codeparrot_training - Step 18114: {'lr': 0.0003733811205248041, 'samples': 9274880, 'steps': 18114, 'loss/train': 1.1715240478515625} 02/25/2022 01:02:37 - INFO - codeparrot_training - Step 18115: {'lr': 0.0003733668893055435, 'samples': 9275392, 'steps': 18115, 'loss/train': 1.523885726928711} 02/25/2022 01:02:41 - INFO - codeparrot_training - Step 18116: {'lr': 0.00037335265755781844, 'samples': 9275904, 'steps': 18116, 'loss/train': 2.677492618560791} 02/25/2022 01:02:49 - INFO - codeparrot_training - Step 18117: {'lr': 0.0003733384252816897, 'samples': 9276416, 'steps': 18117, 'loss/train': 2.5146453380584717} 02/25/2022 01:02:52 - INFO - codeparrot_training - Step 18118: {'lr': 0.0003733241924772183, 'samples': 9276928, 'steps': 18118, 'loss/train': 2.061568260192871} 02/25/2022 01:02:58 - INFO - codeparrot_training - Step 18119: {'lr': 0.0003733099591444652, 'samples': 9277440, 'steps': 18119, 'loss/train': 2.324390172958374} 02/25/2022 01:03:01 - INFO - codeparrot_training - Step 18120: {'lr': 0.0003732957252834914, 'samples': 9277952, 'steps': 18120, 'loss/train': 2.0387895107269287} 02/25/2022 01:03:07 - INFO - codeparrot_training - Step 18121: {'lr': 0.0003732814908943579, 'samples': 9278464, 'steps': 18121, 'loss/train': 2.0450620651245117} 02/25/2022 01:03:10 - INFO - codeparrot_training - Step 18122: {'lr': 0.0003732672559771256, 'samples': 9278976, 'steps': 18122, 'loss/train': 2.033639907836914} 02/25/2022 01:03:16 - INFO - codeparrot_training - Step 18123: {'lr': 0.00037325302053185555, 'samples': 9279488, 'steps': 18123, 'loss/train': 1.8048869371414185} 02/25/2022 01:03:19 - INFO - codeparrot_training - Step 18124: {'lr': 0.0003732387845586087, 'samples': 9280000, 'steps': 18124, 'loss/train': 2.6541860103607178} 02/25/2022 01:03:25 - INFO - codeparrot_training - Step 18125: {'lr': 0.00037322454805744607, 'samples': 9280512, 'steps': 18125, 'loss/train': 1.932218313217163} 02/25/2022 01:03:28 - INFO - codeparrot_training - Step 18126: {'lr': 0.0003732103110284285, 'samples': 9281024, 'steps': 18126, 'loss/train': 2.185729503631592} 02/25/2022 01:03:34 - INFO - codeparrot_training - Step 18127: {'lr': 0.00037319607347161715, 'samples': 9281536, 'steps': 18127, 'loss/train': 2.0502755641937256} 02/25/2022 01:03:37 - INFO - codeparrot_training - Step 18128: {'lr': 0.0003731818353870729, 'samples': 9282048, 'steps': 18128, 'loss/train': 1.6255725622177124} 02/25/2022 01:03:43 - INFO - codeparrot_training - Step 18129: {'lr': 0.0003731675967748568, 'samples': 9282560, 'steps': 18129, 'loss/train': 1.9743597507476807} 02/25/2022 01:03:46 - INFO - codeparrot_training - Step 18130: {'lr': 0.0003731533576350298, 'samples': 9283072, 'steps': 18130, 'loss/train': 2.2525885105133057} 02/25/2022 01:03:53 - INFO - codeparrot_training - Step 18131: {'lr': 0.000373139117967653, 'samples': 9283584, 'steps': 18131, 'loss/train': 1.9093345403671265} 02/25/2022 01:03:57 - INFO - codeparrot_training - Step 18132: {'lr': 0.00037312487777278725, 'samples': 9284096, 'steps': 18132, 'loss/train': 1.726583480834961} 02/25/2022 01:04:03 - INFO - codeparrot_training - Step 18133: {'lr': 0.00037311063705049364, 'samples': 9284608, 'steps': 18133, 'loss/train': 1.5134916305541992} 02/25/2022 01:04:06 - INFO - codeparrot_training - Step 18134: {'lr': 0.00037309639580083314, 'samples': 9285120, 'steps': 18134, 'loss/train': 0.827488899230957} 02/25/2022 01:04:12 - INFO - codeparrot_training - Step 18135: {'lr': 0.0003730821540238668, 'samples': 9285632, 'steps': 18135, 'loss/train': 2.3526413440704346} 02/25/2022 01:04:15 - INFO - codeparrot_training - Step 18136: {'lr': 0.0003730679117196556, 'samples': 9286144, 'steps': 18136, 'loss/train': 1.6611604690551758} 02/25/2022 01:04:21 - INFO - codeparrot_training - Step 18137: {'lr': 0.0003730536688882605, 'samples': 9286656, 'steps': 18137, 'loss/train': 2.226381778717041} 02/25/2022 01:04:25 - INFO - codeparrot_training - Step 18138: {'lr': 0.00037303942552974255, 'samples': 9287168, 'steps': 18138, 'loss/train': 2.1728363037109375} 02/25/2022 01:04:30 - INFO - codeparrot_training - Step 18139: {'lr': 0.0003730251816441628, 'samples': 9287680, 'steps': 18139, 'loss/train': 1.7336833477020264} 02/25/2022 01:04:34 - INFO - codeparrot_training - Step 18140: {'lr': 0.00037301093723158223, 'samples': 9288192, 'steps': 18140, 'loss/train': 1.9051004648208618} 02/25/2022 01:04:39 - INFO - codeparrot_training - Step 18141: {'lr': 0.0003729966922920619, 'samples': 9288704, 'steps': 18141, 'loss/train': 2.0433247089385986} 02/25/2022 01:04:43 - INFO - codeparrot_training - Step 18142: {'lr': 0.00037298244682566273, 'samples': 9289216, 'steps': 18142, 'loss/train': 3.299739360809326} 02/25/2022 01:04:50 - INFO - codeparrot_training - Step 18143: {'lr': 0.0003729682008324459, 'samples': 9289728, 'steps': 18143, 'loss/train': 2.7595059871673584} 02/25/2022 01:04:53 - INFO - codeparrot_training - Step 18144: {'lr': 0.00037295395431247223, 'samples': 9290240, 'steps': 18144, 'loss/train': 1.1415661573410034} 02/25/2022 01:04:59 - INFO - codeparrot_training - Step 18145: {'lr': 0.000372939707265803, 'samples': 9290752, 'steps': 18145, 'loss/train': 1.5083061456680298} 02/25/2022 01:05:02 - INFO - codeparrot_training - Step 18146: {'lr': 0.000372925459692499, 'samples': 9291264, 'steps': 18146, 'loss/train': 1.9867349863052368} 02/25/2022 01:05:08 - INFO - codeparrot_training - Step 18147: {'lr': 0.0003729112115926214, 'samples': 9291776, 'steps': 18147, 'loss/train': 2.9592525959014893} 02/25/2022 01:05:12 - INFO - codeparrot_training - Step 18148: {'lr': 0.00037289696296623117, 'samples': 9292288, 'steps': 18148, 'loss/train': 2.8230485916137695} 02/25/2022 01:05:17 - INFO - codeparrot_training - Step 18149: {'lr': 0.00037288271381338936, 'samples': 9292800, 'steps': 18149, 'loss/train': 2.882969856262207} 02/25/2022 01:05:20 - INFO - codeparrot_training - Step 18150: {'lr': 0.0003728684641341571, 'samples': 9293312, 'steps': 18150, 'loss/train': 2.1075592041015625} 02/25/2022 01:05:26 - INFO - codeparrot_training - Step 18151: {'lr': 0.00037285421392859526, 'samples': 9293824, 'steps': 18151, 'loss/train': 0.2469644844532013} 02/25/2022 01:05:29 - INFO - codeparrot_training - Step 18152: {'lr': 0.00037283996319676505, 'samples': 9294336, 'steps': 18152, 'loss/train': 2.240318536758423} 02/25/2022 01:05:37 - INFO - codeparrot_training - Step 18153: {'lr': 0.0003728257119387274, 'samples': 9294848, 'steps': 18153, 'loss/train': 1.770466685295105} 02/25/2022 01:05:42 - INFO - codeparrot_training - Step 18154: {'lr': 0.0003728114601545435, 'samples': 9295360, 'steps': 18154, 'loss/train': 2.044111967086792} 02/25/2022 01:05:46 - INFO - codeparrot_training - Step 18155: {'lr': 0.0003727972078442742, 'samples': 9295872, 'steps': 18155, 'loss/train': 2.1879515647888184} 02/25/2022 01:05:49 - INFO - codeparrot_training - Step 18156: {'lr': 0.0003727829550079806, 'samples': 9296384, 'steps': 18156, 'loss/train': 1.160942792892456} 02/25/2022 01:05:55 - INFO - codeparrot_training - Step 18157: {'lr': 0.00037276870164572394, 'samples': 9296896, 'steps': 18157, 'loss/train': 2.1044564247131348} 02/25/2022 01:06:00 - INFO - codeparrot_training - Step 18158: {'lr': 0.00037275444775756506, 'samples': 9297408, 'steps': 18158, 'loss/train': 1.6638935804367065} 02/25/2022 01:06:04 - INFO - codeparrot_training - Step 18159: {'lr': 0.00037274019334356516, 'samples': 9297920, 'steps': 18159, 'loss/train': 2.2887117862701416} 02/25/2022 01:06:07 - INFO - codeparrot_training - Step 18160: {'lr': 0.0003727259384037852, 'samples': 9298432, 'steps': 18160, 'loss/train': 2.5881030559539795} 02/25/2022 01:06:13 - INFO - codeparrot_training - Step 18161: {'lr': 0.0003727116829382864, 'samples': 9298944, 'steps': 18161, 'loss/train': 2.3937418460845947} 02/25/2022 01:06:16 - INFO - codeparrot_training - Step 18162: {'lr': 0.00037269742694712963, 'samples': 9299456, 'steps': 18162, 'loss/train': 2.2580037117004395} 02/25/2022 01:06:24 - INFO - codeparrot_training - Step 18163: {'lr': 0.000372683170430376, 'samples': 9299968, 'steps': 18163, 'loss/train': 2.270491361618042} 02/25/2022 01:06:27 - INFO - codeparrot_training - Step 18164: {'lr': 0.00037266891338808667, 'samples': 9300480, 'steps': 18164, 'loss/train': 1.6043137311935425} 02/25/2022 01:06:32 - INFO - codeparrot_training - Step 18165: {'lr': 0.0003726546558203227, 'samples': 9300992, 'steps': 18165, 'loss/train': 3.3150887489318848} 02/25/2022 01:06:36 - INFO - codeparrot_training - Step 18166: {'lr': 0.0003726403977271451, 'samples': 9301504, 'steps': 18166, 'loss/train': 1.5193564891815186} 02/25/2022 01:06:41 - INFO - codeparrot_training - Step 18167: {'lr': 0.00037262613910861485, 'samples': 9302016, 'steps': 18167, 'loss/train': 0.18542303144931793} 02/25/2022 01:06:45 - INFO - codeparrot_training - Step 18168: {'lr': 0.0003726118799647933, 'samples': 9302528, 'steps': 18168, 'loss/train': 2.282599687576294} 02/25/2022 01:06:50 - INFO - codeparrot_training - Step 18169: {'lr': 0.0003725976202957414, 'samples': 9303040, 'steps': 18169, 'loss/train': 1.3491321802139282} 02/25/2022 01:06:54 - INFO - codeparrot_training - Step 18170: {'lr': 0.0003725833601015202, 'samples': 9303552, 'steps': 18170, 'loss/train': 1.3623782396316528} 02/25/2022 01:06:59 - INFO - codeparrot_training - Step 18171: {'lr': 0.0003725690993821908, 'samples': 9304064, 'steps': 18171, 'loss/train': 2.0347962379455566} 02/25/2022 01:07:03 - INFO - codeparrot_training - Step 18172: {'lr': 0.00037255483813781434, 'samples': 9304576, 'steps': 18172, 'loss/train': 1.257857084274292} 02/25/2022 01:07:08 - INFO - codeparrot_training - Step 18173: {'lr': 0.00037254057636845186, 'samples': 9305088, 'steps': 18173, 'loss/train': 2.6665821075439453} 02/25/2022 01:07:12 - INFO - codeparrot_training - Step 18174: {'lr': 0.0003725263140741644, 'samples': 9305600, 'steps': 18174, 'loss/train': 1.9697328805923462} 02/25/2022 01:07:17 - INFO - codeparrot_training - Step 18175: {'lr': 0.0003725120512550131, 'samples': 9306112, 'steps': 18175, 'loss/train': 1.4675523042678833} 02/25/2022 01:07:23 - INFO - codeparrot_training - Step 18176: {'lr': 0.00037249778791105916, 'samples': 9306624, 'steps': 18176, 'loss/train': 1.2928705215454102} 02/25/2022 01:07:26 - INFO - codeparrot_training - Step 18177: {'lr': 0.0003724835240423636, 'samples': 9307136, 'steps': 18177, 'loss/train': 2.392090320587158} 02/25/2022 01:07:34 - INFO - codeparrot_training - Step 18178: {'lr': 0.0003724692596489874, 'samples': 9307648, 'steps': 18178, 'loss/train': 2.3078391551971436} 02/25/2022 01:07:37 - INFO - codeparrot_training - Step 18179: {'lr': 0.0003724549947309919, 'samples': 9308160, 'steps': 18179, 'loss/train': 0.6930814385414124} 02/25/2022 01:07:43 - INFO - codeparrot_training - Step 18180: {'lr': 0.00037244072928843805, 'samples': 9308672, 'steps': 18180, 'loss/train': 1.8631397485733032} 02/25/2022 01:07:46 - INFO - codeparrot_training - Step 18181: {'lr': 0.000372426463321387, 'samples': 9309184, 'steps': 18181, 'loss/train': 2.373523712158203} 02/25/2022 01:07:51 - INFO - codeparrot_training - Step 18182: {'lr': 0.0003724121968298998, 'samples': 9309696, 'steps': 18182, 'loss/train': 1.6055680513381958} 02/25/2022 01:07:55 - INFO - codeparrot_training - Step 18183: {'lr': 0.0003723979298140377, 'samples': 9310208, 'steps': 18183, 'loss/train': 2.60477352142334} 02/25/2022 01:08:00 - INFO - codeparrot_training - Step 18184: {'lr': 0.0003723836622738617, 'samples': 9310720, 'steps': 18184, 'loss/train': 0.5717832446098328} 02/25/2022 01:08:04 - INFO - codeparrot_training - Step 18185: {'lr': 0.00037236939420943294, 'samples': 9311232, 'steps': 18185, 'loss/train': 1.2912105321884155} 02/25/2022 01:08:09 - INFO - codeparrot_training - Step 18186: {'lr': 0.0003723551256208125, 'samples': 9311744, 'steps': 18186, 'loss/train': 1.1674050092697144} 02/25/2022 01:08:13 - INFO - codeparrot_training - Step 18187: {'lr': 0.0003723408565080616, 'samples': 9312256, 'steps': 18187, 'loss/train': 1.5088348388671875} 02/25/2022 01:08:20 - INFO - codeparrot_training - Step 18188: {'lr': 0.00037232658687124135, 'samples': 9312768, 'steps': 18188, 'loss/train': 1.5856822729110718} 02/25/2022 01:08:24 - INFO - codeparrot_training - Step 18189: {'lr': 0.0003723123167104128, 'samples': 9313280, 'steps': 18189, 'loss/train': 2.3326053619384766} 02/25/2022 01:08:30 - INFO - codeparrot_training - Step 18190: {'lr': 0.0003722980460256372, 'samples': 9313792, 'steps': 18190, 'loss/train': 0.7577613592147827} 02/25/2022 01:08:33 - INFO - codeparrot_training - Step 18191: {'lr': 0.00037228377481697555, 'samples': 9314304, 'steps': 18191, 'loss/train': 1.4120705127716064} 02/25/2022 01:08:39 - INFO - codeparrot_training - Step 18192: {'lr': 0.0003722695030844891, 'samples': 9314816, 'steps': 18192, 'loss/train': 3.4470698833465576} 02/25/2022 01:08:42 - INFO - codeparrot_training - Step 18193: {'lr': 0.0003722552308282388, 'samples': 9315328, 'steps': 18193, 'loss/train': 2.4086358547210693} 02/25/2022 01:08:48 - INFO - codeparrot_training - Step 18194: {'lr': 0.000372240958048286, 'samples': 9315840, 'steps': 18194, 'loss/train': 1.9647432565689087} 02/25/2022 01:08:51 - INFO - codeparrot_training - Step 18195: {'lr': 0.0003722266847446918, 'samples': 9316352, 'steps': 18195, 'loss/train': 2.35617733001709} 02/25/2022 01:08:57 - INFO - codeparrot_training - Step 18196: {'lr': 0.00037221241091751716, 'samples': 9316864, 'steps': 18196, 'loss/train': 2.1925320625305176} 02/25/2022 01:09:00 - INFO - codeparrot_training - Step 18197: {'lr': 0.00037219813656682346, 'samples': 9317376, 'steps': 18197, 'loss/train': 1.495530605316162} 02/25/2022 01:09:07 - INFO - codeparrot_training - Step 18198: {'lr': 0.0003721838616926717, 'samples': 9317888, 'steps': 18198, 'loss/train': 1.1121848821640015} 02/25/2022 01:09:11 - INFO - codeparrot_training - Step 18199: {'lr': 0.0003721695862951231, 'samples': 9318400, 'steps': 18199, 'loss/train': 2.4066126346588135} 02/25/2022 01:09:16 - INFO - codeparrot_training - Step 18200: {'lr': 0.0003721553103742388, 'samples': 9318912, 'steps': 18200, 'loss/train': 2.2910492420196533} 02/25/2022 01:09:20 - INFO - codeparrot_training - Step 18201: {'lr': 0.0003721410339300799, 'samples': 9319424, 'steps': 18201, 'loss/train': 1.93354332447052} 02/25/2022 01:09:26 - INFO - codeparrot_training - Step 18202: {'lr': 0.0003721267569627076, 'samples': 9319936, 'steps': 18202, 'loss/train': 2.36238169670105} 02/25/2022 01:09:29 - INFO - codeparrot_training - Step 18203: {'lr': 0.00037211247947218306, 'samples': 9320448, 'steps': 18203, 'loss/train': 3.8021886348724365} 02/25/2022 01:09:35 - INFO - codeparrot_training - Step 18204: {'lr': 0.0003720982014585674, 'samples': 9320960, 'steps': 18204, 'loss/train': 2.211372137069702} 02/25/2022 01:09:38 - INFO - codeparrot_training - Step 18205: {'lr': 0.0003720839229219218, 'samples': 9321472, 'steps': 18205, 'loss/train': 2.2768352031707764} 02/25/2022 01:09:44 - INFO - codeparrot_training - Step 18206: {'lr': 0.00037206964386230754, 'samples': 9321984, 'steps': 18206, 'loss/train': 1.9894641637802124} 02/25/2022 01:09:47 - INFO - codeparrot_training - Step 18207: {'lr': 0.00037205536427978563, 'samples': 9322496, 'steps': 18207, 'loss/train': 1.787400722503662} 02/25/2022 01:09:54 - INFO - codeparrot_training - Step 18208: {'lr': 0.0003720410841744173, 'samples': 9323008, 'steps': 18208, 'loss/train': 1.9226490259170532} 02/25/2022 01:09:58 - INFO - codeparrot_training - Step 18209: {'lr': 0.0003720268035462637, 'samples': 9323520, 'steps': 18209, 'loss/train': 2.5580832958221436} 02/25/2022 01:10:04 - INFO - codeparrot_training - Step 18210: {'lr': 0.000372012522395386, 'samples': 9324032, 'steps': 18210, 'loss/train': 1.062124252319336} 02/25/2022 01:10:07 - INFO - codeparrot_training - Step 18211: {'lr': 0.00037199824072184546, 'samples': 9324544, 'steps': 18211, 'loss/train': 3.8143019676208496} 02/25/2022 01:10:13 - INFO - codeparrot_training - Step 18212: {'lr': 0.0003719839585257032, 'samples': 9325056, 'steps': 18212, 'loss/train': 2.7102842330932617} 02/25/2022 01:10:16 - INFO - codeparrot_training - Step 18213: {'lr': 0.00037196967580702036, 'samples': 9325568, 'steps': 18213, 'loss/train': 1.4733316898345947} 02/25/2022 01:10:22 - INFO - codeparrot_training - Step 18214: {'lr': 0.0003719553925658581, 'samples': 9326080, 'steps': 18214, 'loss/train': 1.295449137687683} 02/25/2022 01:10:25 - INFO - codeparrot_training - Step 18215: {'lr': 0.00037194110880227777, 'samples': 9326592, 'steps': 18215, 'loss/train': 1.6882466077804565} 02/25/2022 01:10:31 - INFO - codeparrot_training - Step 18216: {'lr': 0.0003719268245163404, 'samples': 9327104, 'steps': 18216, 'loss/train': 1.218934178352356} 02/25/2022 01:10:34 - INFO - codeparrot_training - Step 18217: {'lr': 0.0003719125397081072, 'samples': 9327616, 'steps': 18217, 'loss/train': 0.9610633254051208} 02/25/2022 01:10:40 - INFO - codeparrot_training - Step 18218: {'lr': 0.00037189825437763946, 'samples': 9328128, 'steps': 18218, 'loss/train': 2.1672823429107666} 02/25/2022 01:10:43 - INFO - codeparrot_training - Step 18219: {'lr': 0.0003718839685249983, 'samples': 9328640, 'steps': 18219, 'loss/train': 1.6228522062301636} 02/25/2022 01:10:49 - INFO - codeparrot_training - Step 18220: {'lr': 0.0003718696821502449, 'samples': 9329152, 'steps': 18220, 'loss/train': 2.0922420024871826} 02/25/2022 01:10:52 - INFO - codeparrot_training - Step 18221: {'lr': 0.0003718553952534405, 'samples': 9329664, 'steps': 18221, 'loss/train': 2.2527365684509277} 02/25/2022 01:10:58 - INFO - codeparrot_training - Step 18222: {'lr': 0.0003718411078346462, 'samples': 9330176, 'steps': 18222, 'loss/train': 1.614151120185852} 02/25/2022 01:11:01 - INFO - codeparrot_training - Step 18223: {'lr': 0.0003718268198939234, 'samples': 9330688, 'steps': 18223, 'loss/train': 1.6068544387817383} 02/25/2022 01:11:09 - INFO - codeparrot_training - Step 18224: {'lr': 0.0003718125314313331, 'samples': 9331200, 'steps': 18224, 'loss/train': 1.7785314321517944} 02/25/2022 01:11:12 - INFO - codeparrot_training - Step 18225: {'lr': 0.0003717982424469366, 'samples': 9331712, 'steps': 18225, 'loss/train': 3.118506908416748} 02/25/2022 01:11:18 - INFO - codeparrot_training - Step 18226: {'lr': 0.00037178395294079516, 'samples': 9332224, 'steps': 18226, 'loss/train': 1.8392834663391113} 02/25/2022 01:11:21 - INFO - codeparrot_training - Step 18227: {'lr': 0.0003717696629129699, 'samples': 9332736, 'steps': 18227, 'loss/train': 3.2869040966033936} 02/25/2022 01:11:27 - INFO - codeparrot_training - Step 18228: {'lr': 0.00037175537236352205, 'samples': 9333248, 'steps': 18228, 'loss/train': 1.5508891344070435} 02/25/2022 01:11:30 - INFO - codeparrot_training - Step 18229: {'lr': 0.0003717410812925129, 'samples': 9333760, 'steps': 18229, 'loss/train': 2.510763168334961} 02/25/2022 01:11:36 - INFO - codeparrot_training - Step 18230: {'lr': 0.0003717267897000036, 'samples': 9334272, 'steps': 18230, 'loss/train': 0.9270935654640198} 02/25/2022 01:11:39 - INFO - codeparrot_training - Step 18231: {'lr': 0.0003717124975860554, 'samples': 9334784, 'steps': 18231, 'loss/train': 1.5932364463806152} 02/25/2022 01:11:45 - INFO - codeparrot_training - Step 18232: {'lr': 0.00037169820495072935, 'samples': 9335296, 'steps': 18232, 'loss/train': 1.8895373344421387} 02/25/2022 01:11:48 - INFO - codeparrot_training - Step 18233: {'lr': 0.000371683911794087, 'samples': 9335808, 'steps': 18233, 'loss/train': 2.0789709091186523} 02/25/2022 01:11:56 - INFO - codeparrot_training - Step 18234: {'lr': 0.0003716696181161894, 'samples': 9336320, 'steps': 18234, 'loss/train': 2.076809883117676} 02/25/2022 01:11:59 - INFO - codeparrot_training - Step 18235: {'lr': 0.00037165532391709777, 'samples': 9336832, 'steps': 18235, 'loss/train': 1.8170396089553833} 02/25/2022 01:12:04 - INFO - codeparrot_training - Step 18236: {'lr': 0.00037164102919687335, 'samples': 9337344, 'steps': 18236, 'loss/train': 1.6767241954803467} 02/25/2022 01:12:08 - INFO - codeparrot_training - Step 18237: {'lr': 0.00037162673395557737, 'samples': 9337856, 'steps': 18237, 'loss/train': 2.6790950298309326} 02/25/2022 01:12:13 - INFO - codeparrot_training - Step 18238: {'lr': 0.0003716124381932711, 'samples': 9338368, 'steps': 18238, 'loss/train': 2.849553108215332} 02/25/2022 01:12:17 - INFO - codeparrot_training - Step 18239: {'lr': 0.00037159814191001586, 'samples': 9338880, 'steps': 18239, 'loss/train': 2.4096310138702393} 02/25/2022 01:12:23 - INFO - codeparrot_training - Step 18240: {'lr': 0.00037158384510587263, 'samples': 9339392, 'steps': 18240, 'loss/train': 1.4861860275268555} 02/25/2022 01:12:28 - INFO - codeparrot_training - Step 18241: {'lr': 0.0003715695477809029, 'samples': 9339904, 'steps': 18241, 'loss/train': 1.0371453762054443} 02/25/2022 01:12:31 - INFO - codeparrot_training - Step 18242: {'lr': 0.0003715552499351678, 'samples': 9340416, 'steps': 18242, 'loss/train': 2.462574005126953} 02/25/2022 01:12:38 - INFO - codeparrot_training - Step 18243: {'lr': 0.0003715409515687286, 'samples': 9340928, 'steps': 18243, 'loss/train': 1.9747743606567383} 02/25/2022 01:12:41 - INFO - codeparrot_training - Step 18244: {'lr': 0.00037152665268164664, 'samples': 9341440, 'steps': 18244, 'loss/train': 1.7964118719100952} 02/25/2022 01:12:45 - INFO - codeparrot_training - Step 18245: {'lr': 0.00037151235327398304, 'samples': 9341952, 'steps': 18245, 'loss/train': 1.9131324291229248} 02/25/2022 01:12:50 - INFO - codeparrot_training - Step 18246: {'lr': 0.000371498053345799, 'samples': 9342464, 'steps': 18246, 'loss/train': 3.0158257484436035} 02/25/2022 01:12:54 - INFO - codeparrot_training - Step 18247: {'lr': 0.000371483752897156, 'samples': 9342976, 'steps': 18247, 'loss/train': 2.997561454772949} 02/25/2022 01:12:59 - INFO - codeparrot_training - Step 18248: {'lr': 0.00037146945192811513, 'samples': 9343488, 'steps': 18248, 'loss/train': 1.6555982828140259} 02/25/2022 01:13:06 - INFO - codeparrot_training - Step 18249: {'lr': 0.0003714551504387378, 'samples': 9344000, 'steps': 18249, 'loss/train': 2.153865337371826} 02/25/2022 01:13:09 - INFO - codeparrot_training - Step 18250: {'lr': 0.000371440848429085, 'samples': 9344512, 'steps': 18250, 'loss/train': 1.8725228309631348} 02/25/2022 01:13:12 - INFO - codeparrot_training - Step 18251: {'lr': 0.0003714265458992183, 'samples': 9345024, 'steps': 18251, 'loss/train': 1.5715473890304565} 02/25/2022 01:13:18 - INFO - codeparrot_training - Step 18252: {'lr': 0.00037141224284919876, 'samples': 9345536, 'steps': 18252, 'loss/train': 2.6121826171875} 02/25/2022 01:13:21 - INFO - codeparrot_training - Step 18253: {'lr': 0.0003713979392790878, 'samples': 9346048, 'steps': 18253, 'loss/train': 1.0052279233932495} 02/25/2022 01:13:28 - INFO - codeparrot_training - Step 18254: {'lr': 0.0003713836351889465, 'samples': 9346560, 'steps': 18254, 'loss/train': 0.39789798855781555} 02/25/2022 01:13:31 - INFO - codeparrot_training - Step 18255: {'lr': 0.00037136933057883636, 'samples': 9347072, 'steps': 18255, 'loss/train': 3.3356668949127197} 02/25/2022 01:13:37 - INFO - codeparrot_training - Step 18256: {'lr': 0.0003713550254488185, 'samples': 9347584, 'steps': 18256, 'loss/train': 1.932038426399231} 02/25/2022 01:13:40 - INFO - codeparrot_training - Step 18257: {'lr': 0.0003713407197989543, 'samples': 9348096, 'steps': 18257, 'loss/train': 1.9442864656448364} 02/25/2022 01:13:46 - INFO - codeparrot_training - Step 18258: {'lr': 0.0003713264136293049, 'samples': 9348608, 'steps': 18258, 'loss/train': 1.2026679515838623} 02/25/2022 01:13:49 - INFO - codeparrot_training - Step 18259: {'lr': 0.0003713121069399317, 'samples': 9349120, 'steps': 18259, 'loss/train': 1.7753469944000244} 02/25/2022 01:13:55 - INFO - codeparrot_training - Step 18260: {'lr': 0.00037129779973089596, 'samples': 9349632, 'steps': 18260, 'loss/train': 2.330617904663086} 02/25/2022 01:13:58 - INFO - codeparrot_training - Step 18261: {'lr': 0.00037128349200225895, 'samples': 9350144, 'steps': 18261, 'loss/train': 1.6819695234298706} 02/25/2022 01:14:04 - INFO - codeparrot_training - Step 18262: {'lr': 0.000371269183754082, 'samples': 9350656, 'steps': 18262, 'loss/train': 2.2887349128723145} 02/25/2022 01:14:07 - INFO - codeparrot_training - Step 18263: {'lr': 0.00037125487498642636, 'samples': 9351168, 'steps': 18263, 'loss/train': 2.0672101974487305} 02/25/2022 01:14:13 - INFO - codeparrot_training - Step 18264: {'lr': 0.00037124056569935336, 'samples': 9351680, 'steps': 18264, 'loss/train': 2.353395938873291} 02/25/2022 01:14:16 - INFO - codeparrot_training - Step 18265: {'lr': 0.00037122625589292425, 'samples': 9352192, 'steps': 18265, 'loss/train': 2.8991599082946777} 02/25/2022 01:14:22 - INFO - codeparrot_training - Step 18266: {'lr': 0.0003712119455672004, 'samples': 9352704, 'steps': 18266, 'loss/train': 1.459247350692749} 02/25/2022 01:14:25 - INFO - codeparrot_training - Step 18267: {'lr': 0.000371197634722243, 'samples': 9353216, 'steps': 18267, 'loss/train': 1.7257615327835083} 02/25/2022 01:14:31 - INFO - codeparrot_training - Step 18268: {'lr': 0.0003711833233581134, 'samples': 9353728, 'steps': 18268, 'loss/train': 1.954116940498352} 02/25/2022 01:14:34 - INFO - codeparrot_training - Step 18269: {'lr': 0.000371169011474873, 'samples': 9354240, 'steps': 18269, 'loss/train': 1.8782039880752563} 02/25/2022 01:14:41 - INFO - codeparrot_training - Step 18270: {'lr': 0.00037115469907258303, 'samples': 9354752, 'steps': 18270, 'loss/train': 2.1509058475494385} 02/25/2022 01:14:44 - INFO - codeparrot_training - Step 18271: {'lr': 0.0003711403861513047, 'samples': 9355264, 'steps': 18271, 'loss/train': 1.794556975364685} 02/25/2022 01:14:50 - INFO - codeparrot_training - Step 18272: {'lr': 0.0003711260727110995, 'samples': 9355776, 'steps': 18272, 'loss/train': 1.8774168491363525} 02/25/2022 01:14:53 - INFO - codeparrot_training - Step 18273: {'lr': 0.00037111175875202863, 'samples': 9356288, 'steps': 18273, 'loss/train': 2.1876180171966553} 02/25/2022 01:14:59 - INFO - codeparrot_training - Step 18274: {'lr': 0.00037109744427415346, 'samples': 9356800, 'steps': 18274, 'loss/train': 1.2828644514083862} 02/25/2022 01:15:02 - INFO - codeparrot_training - Step 18275: {'lr': 0.0003710831292775353, 'samples': 9357312, 'steps': 18275, 'loss/train': 2.9989209175109863} 02/25/2022 01:15:08 - INFO - codeparrot_training - Step 18276: {'lr': 0.00037106881376223544, 'samples': 9357824, 'steps': 18276, 'loss/train': 1.2367514371871948} 02/25/2022 01:15:11 - INFO - codeparrot_training - Step 18277: {'lr': 0.00037105449772831527, 'samples': 9358336, 'steps': 18277, 'loss/train': 1.881469964981079} 02/25/2022 01:15:17 - INFO - codeparrot_training - Step 18278: {'lr': 0.00037104018117583605, 'samples': 9358848, 'steps': 18278, 'loss/train': 1.8572330474853516} 02/25/2022 01:15:20 - INFO - codeparrot_training - Step 18279: {'lr': 0.00037102586410485915, 'samples': 9359360, 'steps': 18279, 'loss/train': 2.6141300201416016} 02/25/2022 01:15:27 - INFO - codeparrot_training - Step 18280: {'lr': 0.00037101154651544583, 'samples': 9359872, 'steps': 18280, 'loss/train': 2.3476433753967285} 02/25/2022 01:15:30 - INFO - codeparrot_training - Step 18281: {'lr': 0.0003709972284076575, 'samples': 9360384, 'steps': 18281, 'loss/train': 2.8311078548431396} 02/25/2022 01:15:36 - INFO - codeparrot_training - Step 18282: {'lr': 0.0003709829097815555, 'samples': 9360896, 'steps': 18282, 'loss/train': 2.3941707611083984} 02/25/2022 01:15:39 - INFO - codeparrot_training - Step 18283: {'lr': 0.00037096859063720104, 'samples': 9361408, 'steps': 18283, 'loss/train': 2.813333749771118} 02/25/2022 01:15:45 - INFO - codeparrot_training - Step 18284: {'lr': 0.00037095427097465564, 'samples': 9361920, 'steps': 18284, 'loss/train': 2.8890011310577393} 02/25/2022 01:15:48 - INFO - codeparrot_training - Step 18285: {'lr': 0.0003709399507939805, 'samples': 9362432, 'steps': 18285, 'loss/train': 2.1595945358276367} 02/25/2022 01:15:54 - INFO - codeparrot_training - Step 18286: {'lr': 0.00037092563009523703, 'samples': 9362944, 'steps': 18286, 'loss/train': 0.4117693603038788} 02/25/2022 01:15:57 - INFO - codeparrot_training - Step 18287: {'lr': 0.0003709113088784865, 'samples': 9363456, 'steps': 18287, 'loss/train': 1.4524484872817993} 02/25/2022 01:16:03 - INFO - codeparrot_training - Step 18288: {'lr': 0.0003708969871437904, 'samples': 9363968, 'steps': 18288, 'loss/train': 1.7061798572540283} 02/25/2022 01:16:06 - INFO - codeparrot_training - Step 18289: {'lr': 0.00037088266489120996, 'samples': 9364480, 'steps': 18289, 'loss/train': 1.9754153490066528} 02/25/2022 01:16:12 - INFO - codeparrot_training - Step 18290: {'lr': 0.0003708683421208066, 'samples': 9364992, 'steps': 18290, 'loss/train': 2.352123975753784} 02/25/2022 01:16:15 - INFO - codeparrot_training - Step 18291: {'lr': 0.0003708540188326416, 'samples': 9365504, 'steps': 18291, 'loss/train': 2.7301297187805176} 02/25/2022 01:16:22 - INFO - codeparrot_training - Step 18292: {'lr': 0.0003708396950267764, 'samples': 9366016, 'steps': 18292, 'loss/train': 3.0772812366485596} 02/25/2022 01:16:26 - INFO - codeparrot_training - Step 18293: {'lr': 0.00037082537070327225, 'samples': 9366528, 'steps': 18293, 'loss/train': 1.3117905855178833} 02/25/2022 01:16:31 - INFO - codeparrot_training - Step 18294: {'lr': 0.0003708110458621906, 'samples': 9367040, 'steps': 18294, 'loss/train': 1.9054096937179565} 02/25/2022 01:16:35 - INFO - codeparrot_training - Step 18295: {'lr': 0.00037079672050359283, 'samples': 9367552, 'steps': 18295, 'loss/train': 1.9621068239212036} 02/25/2022 01:16:40 - INFO - codeparrot_training - Step 18296: {'lr': 0.00037078239462754023, 'samples': 9368064, 'steps': 18296, 'loss/train': 2.882730007171631} 02/25/2022 01:16:44 - INFO - codeparrot_training - Step 18297: {'lr': 0.00037076806823409426, 'samples': 9368576, 'steps': 18297, 'loss/train': 1.7667450904846191} 02/25/2022 01:16:49 - INFO - codeparrot_training - Step 18298: {'lr': 0.00037075374132331613, 'samples': 9369088, 'steps': 18298, 'loss/train': 1.369690179824829} 02/25/2022 01:16:53 - INFO - codeparrot_training - Step 18299: {'lr': 0.0003707394138952674, 'samples': 9369600, 'steps': 18299, 'loss/train': 1.5599039793014526} 02/25/2022 01:16:58 - INFO - codeparrot_training - Step 18300: {'lr': 0.0003707250859500093, 'samples': 9370112, 'steps': 18300, 'loss/train': 2.365187644958496} 02/25/2022 01:17:02 - INFO - codeparrot_training - Step 18301: {'lr': 0.00037071075748760336, 'samples': 9370624, 'steps': 18301, 'loss/train': 1.4861847162246704} 02/25/2022 01:17:08 - INFO - codeparrot_training - Step 18302: {'lr': 0.0003706964285081108, 'samples': 9371136, 'steps': 18302, 'loss/train': 2.0958995819091797} 02/25/2022 01:17:12 - INFO - codeparrot_training - Step 18303: {'lr': 0.0003706820990115931, 'samples': 9371648, 'steps': 18303, 'loss/train': 1.6913419961929321} 02/25/2022 01:17:17 - INFO - codeparrot_training - Step 18304: {'lr': 0.00037066776899811153, 'samples': 9372160, 'steps': 18304, 'loss/train': 3.7470314502716064} 02/25/2022 01:17:21 - INFO - codeparrot_training - Step 18305: {'lr': 0.00037065343846772765, 'samples': 9372672, 'steps': 18305, 'loss/train': 2.057295083999634} 02/25/2022 01:17:26 - INFO - codeparrot_training - Step 18306: {'lr': 0.0003706391074205027, 'samples': 9373184, 'steps': 18306, 'loss/train': 1.244897723197937} 02/25/2022 01:17:29 - INFO - codeparrot_training - Step 18307: {'lr': 0.00037062477585649814, 'samples': 9373696, 'steps': 18307, 'loss/train': 1.9053668975830078} 02/25/2022 01:17:35 - INFO - codeparrot_training - Step 18308: {'lr': 0.00037061044377577535, 'samples': 9374208, 'steps': 18308, 'loss/train': 1.5779110193252563} 02/25/2022 01:17:39 - INFO - codeparrot_training - Step 18309: {'lr': 0.00037059611117839565, 'samples': 9374720, 'steps': 18309, 'loss/train': 1.312774896621704} 02/25/2022 01:17:44 - INFO - codeparrot_training - Step 18310: {'lr': 0.0003705817780644206, 'samples': 9375232, 'steps': 18310, 'loss/train': 1.9697589874267578} 02/25/2022 01:17:47 - INFO - codeparrot_training - Step 18311: {'lr': 0.0003705674444339114, 'samples': 9375744, 'steps': 18311, 'loss/train': 0.5874767899513245} 02/25/2022 01:17:53 - INFO - codeparrot_training - Step 18312: {'lr': 0.0003705531102869297, 'samples': 9376256, 'steps': 18312, 'loss/train': 1.3228329420089722} 02/25/2022 01:17:57 - INFO - codeparrot_training - Step 18313: {'lr': 0.0003705387756235366, 'samples': 9376768, 'steps': 18313, 'loss/train': 2.041604518890381} 02/25/2022 01:18:02 - INFO - codeparrot_training - Step 18314: {'lr': 0.00037052444044379375, 'samples': 9377280, 'steps': 18314, 'loss/train': 1.872889518737793} 02/25/2022 01:18:06 - INFO - codeparrot_training - Step 18315: {'lr': 0.00037051010474776244, 'samples': 9377792, 'steps': 18315, 'loss/train': 1.7747011184692383} 02/25/2022 01:18:11 - INFO - codeparrot_training - Step 18316: {'lr': 0.0003704957685355041, 'samples': 9378304, 'steps': 18316, 'loss/train': 1.9528049230575562} 02/25/2022 01:18:15 - INFO - codeparrot_training - Step 18317: {'lr': 0.00037048143180708014, 'samples': 9378816, 'steps': 18317, 'loss/train': 2.45226788520813} 02/25/2022 01:18:21 - INFO - codeparrot_training - Step 18318: {'lr': 0.000370467094562552, 'samples': 9379328, 'steps': 18318, 'loss/train': 0.18193577229976654} 02/25/2022 01:18:24 - INFO - codeparrot_training - Step 18319: {'lr': 0.0003704527568019811, 'samples': 9379840, 'steps': 18319, 'loss/train': 3.0419232845306396} 02/25/2022 01:18:30 - INFO - codeparrot_training - Step 18320: {'lr': 0.0003704384185254288, 'samples': 9380352, 'steps': 18320, 'loss/train': 2.0626776218414307} 02/25/2022 01:18:33 - INFO - codeparrot_training - Step 18321: {'lr': 0.0003704240797329566, 'samples': 9380864, 'steps': 18321, 'loss/train': 2.1452231407165527} 02/25/2022 01:18:39 - INFO - codeparrot_training - Step 18322: {'lr': 0.00037040974042462584, 'samples': 9381376, 'steps': 18322, 'loss/train': 2.517463207244873} 02/25/2022 01:18:43 - INFO - codeparrot_training - Step 18323: {'lr': 0.000370395400600498, 'samples': 9381888, 'steps': 18323, 'loss/train': 1.62734854221344} 02/25/2022 01:18:48 - INFO - codeparrot_training - Step 18324: {'lr': 0.00037038106026063457, 'samples': 9382400, 'steps': 18324, 'loss/train': 2.8580076694488525} 02/25/2022 01:18:52 - INFO - codeparrot_training - Step 18325: {'lr': 0.0003703667194050968, 'samples': 9382912, 'steps': 18325, 'loss/train': 2.0689048767089844} 02/25/2022 01:18:57 - INFO - codeparrot_training - Step 18326: {'lr': 0.0003703523780339463, 'samples': 9383424, 'steps': 18326, 'loss/train': 1.462218165397644} 02/25/2022 01:19:00 - INFO - codeparrot_training - Step 18327: {'lr': 0.0003703380361472444, 'samples': 9383936, 'steps': 18327, 'loss/train': 1.5697070360183716} 02/25/2022 01:19:07 - INFO - codeparrot_training - Step 18328: {'lr': 0.00037032369374505255, 'samples': 9384448, 'steps': 18328, 'loss/train': 1.4214327335357666} 02/25/2022 01:19:11 - INFO - codeparrot_training - Step 18329: {'lr': 0.0003703093508274322, 'samples': 9384960, 'steps': 18329, 'loss/train': 1.856069803237915} 02/25/2022 01:19:16 - INFO - codeparrot_training - Step 18330: {'lr': 0.0003702950073944448, 'samples': 9385472, 'steps': 18330, 'loss/train': 1.7260596752166748} 02/25/2022 01:19:20 - INFO - codeparrot_training - Step 18331: {'lr': 0.00037028066344615176, 'samples': 9385984, 'steps': 18331, 'loss/train': 1.7860007286071777} 02/25/2022 01:19:25 - INFO - codeparrot_training - Step 18332: {'lr': 0.0003702663189826146, 'samples': 9386496, 'steps': 18332, 'loss/train': 2.374366521835327} 02/25/2022 01:19:29 - INFO - codeparrot_training - Step 18333: {'lr': 0.00037025197400389467, 'samples': 9387008, 'steps': 18333, 'loss/train': 2.837503671646118} 02/25/2022 01:19:34 - INFO - codeparrot_training - Step 18334: {'lr': 0.0003702376285100535, 'samples': 9387520, 'steps': 18334, 'loss/train': 1.8662089109420776} 02/25/2022 01:19:38 - INFO - codeparrot_training - Step 18335: {'lr': 0.00037022328250115244, 'samples': 9388032, 'steps': 18335, 'loss/train': 1.944802165031433} 02/25/2022 01:19:43 - INFO - codeparrot_training - Step 18336: {'lr': 0.00037020893597725313, 'samples': 9388544, 'steps': 18336, 'loss/train': 1.6886168718338013} 02/25/2022 01:19:47 - INFO - codeparrot_training - Step 18337: {'lr': 0.0003701945889384168, 'samples': 9389056, 'steps': 18337, 'loss/train': 1.4969483613967896} 02/25/2022 01:19:54 - INFO - codeparrot_training - Step 18338: {'lr': 0.00037018024138470515, 'samples': 9389568, 'steps': 18338, 'loss/train': 2.8870134353637695} 02/25/2022 01:19:57 - INFO - codeparrot_training - Step 18339: {'lr': 0.0003701658933161794, 'samples': 9390080, 'steps': 18339, 'loss/train': 2.0723161697387695} 02/25/2022 01:20:02 - INFO - codeparrot_training - Step 18340: {'lr': 0.00037015154473290113, 'samples': 9390592, 'steps': 18340, 'loss/train': 0.9998683333396912} 02/25/2022 01:20:06 - INFO - codeparrot_training - Step 18341: {'lr': 0.0003701371956349318, 'samples': 9391104, 'steps': 18341, 'loss/train': 1.3940317630767822} 02/25/2022 01:20:11 - INFO - codeparrot_training - Step 18342: {'lr': 0.00037012284602233294, 'samples': 9391616, 'steps': 18342, 'loss/train': 1.5923874378204346} 02/25/2022 01:20:15 - INFO - codeparrot_training - Step 18343: {'lr': 0.0003701084958951659, 'samples': 9392128, 'steps': 18343, 'loss/train': 1.4979335069656372} 02/25/2022 01:20:21 - INFO - codeparrot_training - Step 18344: {'lr': 0.0003700941452534922, 'samples': 9392640, 'steps': 18344, 'loss/train': 1.7824156284332275} 02/25/2022 01:20:24 - INFO - codeparrot_training - Step 18345: {'lr': 0.00037007979409737324, 'samples': 9393152, 'steps': 18345, 'loss/train': 1.1498560905456543} 02/25/2022 01:20:30 - INFO - codeparrot_training - Step 18346: {'lr': 0.0003700654424268707, 'samples': 9393664, 'steps': 18346, 'loss/train': 2.7380874156951904} 02/25/2022 01:20:36 - INFO - codeparrot_training - Step 18347: {'lr': 0.00037005109024204586, 'samples': 9394176, 'steps': 18347, 'loss/train': 2.328641176223755} 02/25/2022 01:20:39 - INFO - codeparrot_training - Step 18348: {'lr': 0.00037003673754296026, 'samples': 9394688, 'steps': 18348, 'loss/train': 1.8855458498001099} 02/25/2022 01:20:45 - INFO - codeparrot_training - Step 18349: {'lr': 0.00037002238432967547, 'samples': 9395200, 'steps': 18349, 'loss/train': 3.376471757888794} 02/25/2022 01:20:48 - INFO - codeparrot_training - Step 18350: {'lr': 0.0003700080306022528, 'samples': 9395712, 'steps': 18350, 'loss/train': 2.581509828567505} 02/25/2022 01:20:54 - INFO - codeparrot_training - Step 18351: {'lr': 0.00036999367636075386, 'samples': 9396224, 'steps': 18351, 'loss/train': 2.2691421508789062} 02/25/2022 01:20:58 - INFO - codeparrot_training - Step 18352: {'lr': 0.00036997932160524015, 'samples': 9396736, 'steps': 18352, 'loss/train': 1.8684520721435547} 02/25/2022 01:21:03 - INFO - codeparrot_training - Step 18353: {'lr': 0.00036996496633577314, 'samples': 9397248, 'steps': 18353, 'loss/train': 2.008300304412842} 02/25/2022 01:21:06 - INFO - codeparrot_training - Step 18354: {'lr': 0.00036995061055241426, 'samples': 9397760, 'steps': 18354, 'loss/train': 1.93635094165802} 02/25/2022 01:21:12 - INFO - codeparrot_training - Step 18355: {'lr': 0.000369936254255225, 'samples': 9398272, 'steps': 18355, 'loss/train': 1.8770588636398315} 02/25/2022 01:21:15 - INFO - codeparrot_training - Step 18356: {'lr': 0.000369921897444267, 'samples': 9398784, 'steps': 18356, 'loss/train': 2.5583128929138184} 02/25/2022 01:21:21 - INFO - codeparrot_training - Step 18357: {'lr': 0.00036990754011960165, 'samples': 9399296, 'steps': 18357, 'loss/train': 1.3393158912658691} 02/25/2022 01:21:25 - INFO - codeparrot_training - Step 18358: {'lr': 0.0003698931822812905, 'samples': 9399808, 'steps': 18358, 'loss/train': 0.09090162068605423} 02/25/2022 01:21:30 - INFO - codeparrot_training - Step 18359: {'lr': 0.000369878823929395, 'samples': 9400320, 'steps': 18359, 'loss/train': 1.9868463277816772} 02/25/2022 01:21:34 - INFO - codeparrot_training - Step 18360: {'lr': 0.00036986446506397666, 'samples': 9400832, 'steps': 18360, 'loss/train': 1.0877060890197754} 02/25/2022 01:21:39 - INFO - codeparrot_training - Step 18361: {'lr': 0.00036985010568509703, 'samples': 9401344, 'steps': 18361, 'loss/train': 1.9403364658355713} 02/25/2022 01:21:43 - INFO - codeparrot_training - Step 18362: {'lr': 0.00036983574579281764, 'samples': 9401856, 'steps': 18362, 'loss/train': 3.4670326709747314} 02/25/2022 01:21:49 - INFO - codeparrot_training - Step 18363: {'lr': 0.0003698213853871999, 'samples': 9402368, 'steps': 18363, 'loss/train': 2.404047727584839} 02/25/2022 01:21:52 - INFO - codeparrot_training - Step 18364: {'lr': 0.00036980702446830547, 'samples': 9402880, 'steps': 18364, 'loss/train': 1.217385172843933} 02/25/2022 01:21:58 - INFO - codeparrot_training - Step 18365: {'lr': 0.0003697926630361957, 'samples': 9403392, 'steps': 18365, 'loss/train': 2.240957498550415} 02/25/2022 01:22:01 - INFO - codeparrot_training - Step 18366: {'lr': 0.00036977830109093227, 'samples': 9403904, 'steps': 18366, 'loss/train': 2.3599369525909424} 02/25/2022 01:22:07 - INFO - codeparrot_training - Step 18367: {'lr': 0.0003697639386325766, 'samples': 9404416, 'steps': 18367, 'loss/train': 2.1313180923461914} 02/25/2022 01:22:10 - INFO - codeparrot_training - Step 18368: {'lr': 0.00036974957566119027, 'samples': 9404928, 'steps': 18368, 'loss/train': 2.2007174491882324} 02/25/2022 01:22:16 - INFO - codeparrot_training - Step 18369: {'lr': 0.00036973521217683475, 'samples': 9405440, 'steps': 18369, 'loss/train': 1.2404251098632812} 02/25/2022 01:22:19 - INFO - codeparrot_training - Step 18370: {'lr': 0.00036972084817957164, 'samples': 9405952, 'steps': 18370, 'loss/train': 1.6137281656265259} 02/25/2022 01:22:25 - INFO - codeparrot_training - Step 18371: {'lr': 0.0003697064836694624, 'samples': 9406464, 'steps': 18371, 'loss/train': 1.1886402368545532} 02/25/2022 01:22:28 - INFO - codeparrot_training - Step 18372: {'lr': 0.0003696921186465686, 'samples': 9406976, 'steps': 18372, 'loss/train': 1.5630158185958862} 02/25/2022 01:22:34 - INFO - codeparrot_training - Step 18373: {'lr': 0.00036967775311095186, 'samples': 9407488, 'steps': 18373, 'loss/train': 1.7254512310028076} 02/25/2022 01:22:38 - INFO - codeparrot_training - Step 18374: {'lr': 0.00036966338706267347, 'samples': 9408000, 'steps': 18374, 'loss/train': 1.644241213798523} 02/25/2022 01:22:43 - INFO - codeparrot_training - Step 18375: {'lr': 0.0003696490205017953, 'samples': 9408512, 'steps': 18375, 'loss/train': 1.9731807708740234} 02/25/2022 01:22:47 - INFO - codeparrot_training - Step 18376: {'lr': 0.00036963465342837855, 'samples': 9409024, 'steps': 18376, 'loss/train': 1.5577452182769775} 02/25/2022 01:22:52 - INFO - codeparrot_training - Step 18377: {'lr': 0.000369620285842485, 'samples': 9409536, 'steps': 18377, 'loss/train': 2.6976330280303955} 02/25/2022 01:22:56 - INFO - codeparrot_training - Step 18378: {'lr': 0.00036960591774417613, 'samples': 9410048, 'steps': 18378, 'loss/train': 2.6651177406311035} 02/25/2022 01:23:01 - INFO - codeparrot_training - Step 18379: {'lr': 0.00036959154913351357, 'samples': 9410560, 'steps': 18379, 'loss/train': 2.071716547012329} 02/25/2022 01:23:05 - INFO - codeparrot_training - Step 18380: {'lr': 0.0003695771800105586, 'samples': 9411072, 'steps': 18380, 'loss/train': 1.6388473510742188} 02/25/2022 01:23:11 - INFO - codeparrot_training - Step 18381: {'lr': 0.00036956281037537307, 'samples': 9411584, 'steps': 18381, 'loss/train': 2.0668752193450928} 02/25/2022 01:23:15 - INFO - codeparrot_training - Step 18382: {'lr': 0.00036954844022801846, 'samples': 9412096, 'steps': 18382, 'loss/train': 1.8403310775756836} 02/25/2022 01:23:20 - INFO - codeparrot_training - Step 18383: {'lr': 0.00036953406956855624, 'samples': 9412608, 'steps': 18383, 'loss/train': 2.3804545402526855} 02/25/2022 01:23:24 - INFO - codeparrot_training - Step 18384: {'lr': 0.0003695196983970481, 'samples': 9413120, 'steps': 18384, 'loss/train': 1.3367730379104614} 02/25/2022 01:23:29 - INFO - codeparrot_training - Step 18385: {'lr': 0.0003695053267135554, 'samples': 9413632, 'steps': 18385, 'loss/train': 2.2116525173187256} 02/25/2022 01:23:33 - INFO - codeparrot_training - Step 18386: {'lr': 0.00036949095451813997, 'samples': 9414144, 'steps': 18386, 'loss/train': 2.4977450370788574} 02/25/2022 01:23:39 - INFO - codeparrot_training - Step 18387: {'lr': 0.0003694765818108631, 'samples': 9414656, 'steps': 18387, 'loss/train': 2.2690670490264893} 02/25/2022 01:23:42 - INFO - codeparrot_training - Step 18388: {'lr': 0.00036946220859178656, 'samples': 9415168, 'steps': 18388, 'loss/train': 1.9999747276306152} 02/25/2022 01:23:48 - INFO - codeparrot_training - Step 18389: {'lr': 0.0003694478348609718, 'samples': 9415680, 'steps': 18389, 'loss/train': 1.8512916564941406} 02/25/2022 01:23:51 - INFO - codeparrot_training - Step 18390: {'lr': 0.00036943346061848054, 'samples': 9416192, 'steps': 18390, 'loss/train': 2.158099889755249} 02/25/2022 01:23:57 - INFO - codeparrot_training - Step 18391: {'lr': 0.00036941908586437416, 'samples': 9416704, 'steps': 18391, 'loss/train': 2.0486855506896973} 02/25/2022 01:24:00 - INFO - codeparrot_training - Step 18392: {'lr': 0.0003694047105987144, 'samples': 9417216, 'steps': 18392, 'loss/train': 1.4348971843719482} 02/25/2022 01:24:07 - INFO - codeparrot_training - Step 18393: {'lr': 0.00036939033482156277, 'samples': 9417728, 'steps': 18393, 'loss/train': 2.4456098079681396} 02/25/2022 01:24:10 - INFO - codeparrot_training - Step 18394: {'lr': 0.00036937595853298076, 'samples': 9418240, 'steps': 18394, 'loss/train': 4.3987884521484375} 02/25/2022 01:24:16 - INFO - codeparrot_training - Step 18395: {'lr': 0.0003693615817330302, 'samples': 9418752, 'steps': 18395, 'loss/train': 1.0911381244659424} 02/25/2022 01:24:19 - INFO - codeparrot_training - Step 18396: {'lr': 0.00036934720442177244, 'samples': 9419264, 'steps': 18396, 'loss/train': 2.5088391304016113} 02/25/2022 01:24:25 - INFO - codeparrot_training - Step 18397: {'lr': 0.0003693328265992692, 'samples': 9419776, 'steps': 18397, 'loss/train': 0.30967798829078674} 02/25/2022 01:24:28 - INFO - codeparrot_training - Step 18398: {'lr': 0.000369318448265582, 'samples': 9420288, 'steps': 18398, 'loss/train': 1.6470991373062134} 02/25/2022 01:24:34 - INFO - codeparrot_training - Step 18399: {'lr': 0.00036930406942077245, 'samples': 9420800, 'steps': 18399, 'loss/train': 0.20155486464500427} 02/25/2022 01:24:37 - INFO - codeparrot_training - Step 18400: {'lr': 0.0003692896900649021, 'samples': 9421312, 'steps': 18400, 'loss/train': 2.0221967697143555} 02/25/2022 01:24:43 - INFO - codeparrot_training - Step 18401: {'lr': 0.0003692753101980327, 'samples': 9421824, 'steps': 18401, 'loss/train': 2.021521806716919} 02/25/2022 01:24:46 - INFO - codeparrot_training - Step 18402: {'lr': 0.00036926092982022564, 'samples': 9422336, 'steps': 18402, 'loss/train': 2.6775269508361816} 02/25/2022 01:24:52 - INFO - codeparrot_training - Step 18403: {'lr': 0.0003692465489315427, 'samples': 9422848, 'steps': 18403, 'loss/train': 2.4501376152038574} 02/25/2022 01:24:55 - INFO - codeparrot_training - Step 18404: {'lr': 0.00036923216753204536, 'samples': 9423360, 'steps': 18404, 'loss/train': 0.5973928570747375} 02/25/2022 01:25:01 - INFO - codeparrot_training - Step 18405: {'lr': 0.0003692177856217953, 'samples': 9423872, 'steps': 18405, 'loss/train': 1.571923017501831} 02/25/2022 01:25:05 - INFO - codeparrot_training - Step 18406: {'lr': 0.00036920340320085413, 'samples': 9424384, 'steps': 18406, 'loss/train': 1.652537226676941} 02/25/2022 01:25:10 - INFO - codeparrot_training - Step 18407: {'lr': 0.00036918902026928334, 'samples': 9424896, 'steps': 18407, 'loss/train': 1.7127047777175903} 02/25/2022 01:25:14 - INFO - codeparrot_training - Step 18408: {'lr': 0.00036917463682714473, 'samples': 9425408, 'steps': 18408, 'loss/train': 2.864731788635254} 02/25/2022 01:25:20 - INFO - codeparrot_training - Step 18409: {'lr': 0.00036916025287449976, 'samples': 9425920, 'steps': 18409, 'loss/train': 2.3754286766052246} 02/25/2022 01:25:23 - INFO - codeparrot_training - Step 18410: {'lr': 0.0003691458684114102, 'samples': 9426432, 'steps': 18410, 'loss/train': 1.3476039171218872} 02/25/2022 01:25:29 - INFO - codeparrot_training - Step 18411: {'lr': 0.00036913148343793744, 'samples': 9426944, 'steps': 18411, 'loss/train': 1.0442968606948853} 02/25/2022 01:25:32 - INFO - codeparrot_training - Step 18412: {'lr': 0.00036911709795414336, 'samples': 9427456, 'steps': 18412, 'loss/train': 0.8767587542533875} 02/25/2022 01:25:38 - INFO - codeparrot_training - Step 18413: {'lr': 0.00036910271196008936, 'samples': 9427968, 'steps': 18413, 'loss/train': 7.2651214599609375} 02/25/2022 01:25:41 - INFO - codeparrot_training - Step 18414: {'lr': 0.0003690883254558372, 'samples': 9428480, 'steps': 18414, 'loss/train': 2.196528434753418} 02/25/2022 01:25:47 - INFO - codeparrot_training - Step 18415: {'lr': 0.0003690739384414485, 'samples': 9428992, 'steps': 18415, 'loss/train': 3.6021289825439453} 02/25/2022 01:25:51 - INFO - codeparrot_training - Step 18416: {'lr': 0.0003690595509169848, 'samples': 9429504, 'steps': 18416, 'loss/train': 1.0051133632659912} 02/25/2022 01:25:56 - INFO - codeparrot_training - Step 18417: {'lr': 0.00036904516288250786, 'samples': 9430016, 'steps': 18417, 'loss/train': 2.189079761505127} 02/25/2022 01:26:00 - INFO - codeparrot_training - Step 18418: {'lr': 0.0003690307743380791, 'samples': 9430528, 'steps': 18418, 'loss/train': 2.0821285247802734} 02/25/2022 01:26:06 - INFO - codeparrot_training - Step 18419: {'lr': 0.00036901638528376047, 'samples': 9431040, 'steps': 18419, 'loss/train': 1.8279738426208496} 02/25/2022 01:26:09 - INFO - codeparrot_training - Step 18420: {'lr': 0.00036900199571961336, 'samples': 9431552, 'steps': 18420, 'loss/train': 2.6134769916534424} 02/25/2022 01:26:15 - INFO - codeparrot_training - Step 18421: {'lr': 0.0003689876056456995, 'samples': 9432064, 'steps': 18421, 'loss/train': 1.9416747093200684} 02/25/2022 01:26:18 - INFO - codeparrot_training - Step 18422: {'lr': 0.0003689732150620805, 'samples': 9432576, 'steps': 18422, 'loss/train': 2.1659059524536133} 02/25/2022 01:26:24 - INFO - codeparrot_training - Step 18423: {'lr': 0.00036895882396881805, 'samples': 9433088, 'steps': 18423, 'loss/train': 2.4592130184173584} 02/25/2022 01:26:27 - INFO - codeparrot_training - Step 18424: {'lr': 0.0003689444323659737, 'samples': 9433600, 'steps': 18424, 'loss/train': 1.4516977071762085} 02/25/2022 01:26:33 - INFO - codeparrot_training - Step 18425: {'lr': 0.00036893004025360926, 'samples': 9434112, 'steps': 18425, 'loss/train': 1.8877735137939453} 02/25/2022 01:26:36 - INFO - codeparrot_training - Step 18426: {'lr': 0.0003689156476317862, 'samples': 9434624, 'steps': 18426, 'loss/train': 3.7378692626953125} 02/25/2022 01:26:42 - INFO - codeparrot_training - Step 18427: {'lr': 0.0003689012545005664, 'samples': 9435136, 'steps': 18427, 'loss/train': 1.6135536432266235} 02/25/2022 01:26:45 - INFO - codeparrot_training - Step 18428: {'lr': 0.0003688868608600113, 'samples': 9435648, 'steps': 18428, 'loss/train': 2.551117181777954} 02/25/2022 01:26:52 - INFO - codeparrot_training - Step 18429: {'lr': 0.0003688724667101826, 'samples': 9436160, 'steps': 18429, 'loss/train': 2.1527068614959717} 02/25/2022 01:26:55 - INFO - codeparrot_training - Step 18430: {'lr': 0.0003688580720511421, 'samples': 9436672, 'steps': 18430, 'loss/train': 1.1708322763442993} 02/25/2022 01:27:01 - INFO - codeparrot_training - Step 18431: {'lr': 0.0003688436768829512, 'samples': 9437184, 'steps': 18431, 'loss/train': 1.2201554775238037} 02/25/2022 01:27:04 - INFO - codeparrot_training - Step 18432: {'lr': 0.0003688292812056719, 'samples': 9437696, 'steps': 18432, 'loss/train': 2.417375326156616} 02/25/2022 01:27:10 - INFO - codeparrot_training - Step 18433: {'lr': 0.00036881488501936554, 'samples': 9438208, 'steps': 18433, 'loss/train': 0.9687504172325134} 02/25/2022 01:27:13 - INFO - codeparrot_training - Step 18434: {'lr': 0.00036880048832409407, 'samples': 9438720, 'steps': 18434, 'loss/train': 2.5395543575286865} 02/25/2022 01:27:19 - INFO - codeparrot_training - Step 18435: {'lr': 0.000368786091119919, 'samples': 9439232, 'steps': 18435, 'loss/train': 2.1375091075897217} 02/25/2022 01:27:22 - INFO - codeparrot_training - Step 18436: {'lr': 0.00036877169340690204, 'samples': 9439744, 'steps': 18436, 'loss/train': 2.1693179607391357} 02/25/2022 01:27:28 - INFO - codeparrot_training - Step 18437: {'lr': 0.0003687572951851048, 'samples': 9440256, 'steps': 18437, 'loss/train': 2.2370991706848145} 02/25/2022 01:27:31 - INFO - codeparrot_training - Step 18438: {'lr': 0.0003687428964545891, 'samples': 9440768, 'steps': 18438, 'loss/train': 1.4605731964111328} 02/25/2022 01:27:37 - INFO - codeparrot_training - Step 18439: {'lr': 0.00036872849721541643, 'samples': 9441280, 'steps': 18439, 'loss/train': 3.647575855255127} 02/25/2022 01:27:41 - INFO - codeparrot_training - Step 18440: {'lr': 0.0003687140974676486, 'samples': 9441792, 'steps': 18440, 'loss/train': 1.9040979146957397} 02/25/2022 01:27:46 - INFO - codeparrot_training - Step 18441: {'lr': 0.00036869969721134736, 'samples': 9442304, 'steps': 18441, 'loss/train': 1.2172081470489502} 02/25/2022 01:27:50 - INFO - codeparrot_training - Step 18442: {'lr': 0.0003686852964465742, 'samples': 9442816, 'steps': 18442, 'loss/train': 1.8815768957138062} 02/25/2022 01:27:55 - INFO - codeparrot_training - Step 18443: {'lr': 0.000368670895173391, 'samples': 9443328, 'steps': 18443, 'loss/train': 2.245439291000366} 02/25/2022 01:27:59 - INFO - codeparrot_training - Step 18444: {'lr': 0.00036865649339185935, 'samples': 9443840, 'steps': 18444, 'loss/train': 2.360513925552368} 02/25/2022 01:28:04 - INFO - codeparrot_training - Step 18445: {'lr': 0.000368642091102041, 'samples': 9444352, 'steps': 18445, 'loss/train': 1.7992020845413208} 02/25/2022 01:28:08 - INFO - codeparrot_training - Step 18446: {'lr': 0.0003686276883039975, 'samples': 9444864, 'steps': 18446, 'loss/train': 0.591636598110199} 02/25/2022 01:28:13 - INFO - codeparrot_training - Step 18447: {'lr': 0.0003686132849977908, 'samples': 9445376, 'steps': 18447, 'loss/train': 2.6358649730682373} 02/25/2022 01:28:17 - INFO - codeparrot_training - Step 18448: {'lr': 0.0003685988811834823, 'samples': 9445888, 'steps': 18448, 'loss/train': 1.8494524955749512} 02/25/2022 01:28:22 - INFO - codeparrot_training - Step 18449: {'lr': 0.00036858447686113395, 'samples': 9446400, 'steps': 18449, 'loss/train': 1.716161847114563} 02/25/2022 01:28:26 - INFO - codeparrot_training - Step 18450: {'lr': 0.0003685700720308073, 'samples': 9446912, 'steps': 18450, 'loss/train': 2.587435483932495} 02/25/2022 01:28:32 - INFO - codeparrot_training - Step 18451: {'lr': 0.0003685556666925641, 'samples': 9447424, 'steps': 18451, 'loss/train': 1.9122740030288696} 02/25/2022 01:28:37 - INFO - codeparrot_training - Step 18452: {'lr': 0.0003685412608464661, 'samples': 9447936, 'steps': 18452, 'loss/train': 1.8353451490402222} 02/25/2022 01:28:41 - INFO - codeparrot_training - Step 18453: {'lr': 0.00036852685449257505, 'samples': 9448448, 'steps': 18453, 'loss/train': 2.0929436683654785} 02/25/2022 01:28:46 - INFO - codeparrot_training - Step 18454: {'lr': 0.00036851244763095247, 'samples': 9448960, 'steps': 18454, 'loss/train': 2.102295398712158} 02/25/2022 01:28:50 - INFO - codeparrot_training - Step 18455: {'lr': 0.0003684980402616603, 'samples': 9449472, 'steps': 18455, 'loss/train': 2.165572166442871} 02/25/2022 01:28:55 - INFO - codeparrot_training - Step 18456: {'lr': 0.0003684836323847601, 'samples': 9449984, 'steps': 18456, 'loss/train': 1.3640385866165161} 02/25/2022 01:28:59 - INFO - codeparrot_training - Step 18457: {'lr': 0.0003684692240003137, 'samples': 9450496, 'steps': 18457, 'loss/train': 1.8594518899917603} 02/25/2022 01:29:04 - INFO - codeparrot_training - Step 18458: {'lr': 0.00036845481510838264, 'samples': 9451008, 'steps': 18458, 'loss/train': 2.5686378479003906} 02/25/2022 01:29:08 - INFO - codeparrot_training - Step 18459: {'lr': 0.00036844040570902886, 'samples': 9451520, 'steps': 18459, 'loss/train': 1.8522651195526123} 02/25/2022 01:29:14 - INFO - codeparrot_training - Step 18460: {'lr': 0.00036842599580231395, 'samples': 9452032, 'steps': 18460, 'loss/train': 2.4291305541992188} 02/25/2022 01:29:18 - INFO - codeparrot_training - Step 18461: {'lr': 0.0003684115853882997, 'samples': 9452544, 'steps': 18461, 'loss/train': 1.8021272420883179} 02/25/2022 01:29:23 - INFO - codeparrot_training - Step 18462: {'lr': 0.00036839717446704787, 'samples': 9453056, 'steps': 18462, 'loss/train': 2.070101737976074} 02/25/2022 01:29:27 - INFO - codeparrot_training - Step 18463: {'lr': 0.00036838276303862, 'samples': 9453568, 'steps': 18463, 'loss/train': 1.5154380798339844} 02/25/2022 01:29:32 - INFO - codeparrot_training - Step 18464: {'lr': 0.00036836835110307803, 'samples': 9454080, 'steps': 18464, 'loss/train': 2.5555546283721924} 02/25/2022 01:29:36 - INFO - codeparrot_training - Step 18465: {'lr': 0.0003683539386604837, 'samples': 9454592, 'steps': 18465, 'loss/train': 2.0288686752319336} 02/25/2022 01:29:41 - INFO - codeparrot_training - Step 18466: {'lr': 0.00036833952571089856, 'samples': 9455104, 'steps': 18466, 'loss/train': 2.535541296005249} 02/25/2022 01:29:45 - INFO - codeparrot_training - Step 18467: {'lr': 0.0003683251122543846, 'samples': 9455616, 'steps': 18467, 'loss/train': 2.0063090324401855} 02/25/2022 01:29:50 - INFO - codeparrot_training - Step 18468: {'lr': 0.0003683106982910033, 'samples': 9456128, 'steps': 18468, 'loss/train': 1.337662696838379} 02/25/2022 01:29:54 - INFO - codeparrot_training - Step 18469: {'lr': 0.0003682962838208166, 'samples': 9456640, 'steps': 18469, 'loss/train': 2.7879881858825684} 02/25/2022 01:29:59 - INFO - codeparrot_training - Step 18470: {'lr': 0.0003682818688438862, 'samples': 9457152, 'steps': 18470, 'loss/train': 2.137911319732666} 02/25/2022 01:30:03 - INFO - codeparrot_training - Step 18471: {'lr': 0.00036826745336027383, 'samples': 9457664, 'steps': 18471, 'loss/train': 1.447295904159546} 02/25/2022 01:30:08 - INFO - codeparrot_training - Step 18472: {'lr': 0.0003682530373700412, 'samples': 9458176, 'steps': 18472, 'loss/train': 0.09690675139427185} 02/25/2022 01:30:12 - INFO - codeparrot_training - Step 18473: {'lr': 0.00036823862087325017, 'samples': 9458688, 'steps': 18473, 'loss/train': 2.516244411468506} 02/25/2022 01:30:17 - INFO - codeparrot_training - Step 18474: {'lr': 0.00036822420386996237, 'samples': 9459200, 'steps': 18474, 'loss/train': 3.8018407821655273} 02/25/2022 01:30:21 - INFO - codeparrot_training - Step 18475: {'lr': 0.0003682097863602397, 'samples': 9459712, 'steps': 18475, 'loss/train': 1.629761815071106} 02/25/2022 01:30:27 - INFO - codeparrot_training - Step 18476: {'lr': 0.00036819536834414374, 'samples': 9460224, 'steps': 18476, 'loss/train': 1.3174259662628174} 02/25/2022 01:30:30 - INFO - codeparrot_training - Step 18477: {'lr': 0.0003681809498217364, 'samples': 9460736, 'steps': 18477, 'loss/train': 1.6340336799621582} 02/25/2022 01:30:36 - INFO - codeparrot_training - Step 18478: {'lr': 0.0003681665307930794, 'samples': 9461248, 'steps': 18478, 'loss/train': 2.07969069480896} 02/25/2022 01:30:39 - INFO - codeparrot_training - Step 18479: {'lr': 0.0003681521112582345, 'samples': 9461760, 'steps': 18479, 'loss/train': 2.468186140060425} 02/25/2022 01:30:45 - INFO - codeparrot_training - Step 18480: {'lr': 0.00036813769121726354, 'samples': 9462272, 'steps': 18480, 'loss/train': 1.7095630168914795} 02/25/2022 01:30:48 - INFO - codeparrot_training - Step 18481: {'lr': 0.00036812327067022813, 'samples': 9462784, 'steps': 18481, 'loss/train': 1.8921030759811401} 02/25/2022 01:30:54 - INFO - codeparrot_training - Step 18482: {'lr': 0.00036810884961719015, 'samples': 9463296, 'steps': 18482, 'loss/train': 3.1167805194854736} 02/25/2022 01:30:57 - INFO - codeparrot_training - Step 18483: {'lr': 0.0003680944280582114, 'samples': 9463808, 'steps': 18483, 'loss/train': 8.680535316467285} 02/25/2022 01:31:03 - INFO - codeparrot_training - Step 18484: {'lr': 0.0003680800059933536, 'samples': 9464320, 'steps': 18484, 'loss/train': 1.7377374172210693} 02/25/2022 01:31:06 - INFO - codeparrot_training - Step 18485: {'lr': 0.00036806558342267854, 'samples': 9464832, 'steps': 18485, 'loss/train': 1.5186407566070557} 02/25/2022 01:31:13 - INFO - codeparrot_training - Step 18486: {'lr': 0.0003680511603462481, 'samples': 9465344, 'steps': 18486, 'loss/train': 2.3167595863342285} 02/25/2022 01:31:16 - INFO - codeparrot_training - Step 18487: {'lr': 0.00036803673676412386, 'samples': 9465856, 'steps': 18487, 'loss/train': 1.307106375694275} 02/25/2022 01:31:22 - INFO - codeparrot_training - Step 18488: {'lr': 0.00036802231267636773, 'samples': 9466368, 'steps': 18488, 'loss/train': 1.9781852960586548} 02/25/2022 01:31:25 - INFO - codeparrot_training - Step 18489: {'lr': 0.0003680078880830415, 'samples': 9466880, 'steps': 18489, 'loss/train': 1.6012712717056274} 02/25/2022 01:31:31 - INFO - codeparrot_training - Step 18490: {'lr': 0.000367993462984207, 'samples': 9467392, 'steps': 18490, 'loss/train': 2.742845058441162} 02/25/2022 01:31:34 - INFO - codeparrot_training - Step 18491: {'lr': 0.0003679790373799259, 'samples': 9467904, 'steps': 18491, 'loss/train': 1.0837022066116333} 02/25/2022 01:31:40 - INFO - codeparrot_training - Step 18492: {'lr': 0.0003679646112702601, 'samples': 9468416, 'steps': 18492, 'loss/train': 2.023545026779175} 02/25/2022 01:31:43 - INFO - codeparrot_training - Step 18493: {'lr': 0.0003679501846552714, 'samples': 9468928, 'steps': 18493, 'loss/train': 1.9390811920166016} 02/25/2022 01:31:49 - INFO - codeparrot_training - Step 18494: {'lr': 0.00036793575753502153, 'samples': 9469440, 'steps': 18494, 'loss/train': 1.4034349918365479} 02/25/2022 01:31:52 - INFO - codeparrot_training - Step 18495: {'lr': 0.0003679213299095723, 'samples': 9469952, 'steps': 18495, 'loss/train': 0.9205752015113831} 02/25/2022 01:31:58 - INFO - codeparrot_training - Step 18496: {'lr': 0.00036790690177898556, 'samples': 9470464, 'steps': 18496, 'loss/train': 1.2389694452285767} 02/25/2022 01:32:02 - INFO - codeparrot_training - Step 18497: {'lr': 0.00036789247314332306, 'samples': 9470976, 'steps': 18497, 'loss/train': 0.17057999968528748} 02/25/2022 01:32:07 - INFO - codeparrot_training - Step 18498: {'lr': 0.00036787804400264666, 'samples': 9471488, 'steps': 18498, 'loss/train': 2.3652663230895996} 02/25/2022 01:32:11 - INFO - codeparrot_training - Step 18499: {'lr': 0.00036786361435701823, 'samples': 9472000, 'steps': 18499, 'loss/train': 0.6023238301277161} 02/25/2022 01:32:16 - INFO - codeparrot_training - Step 18500: {'lr': 0.0003678491842064995, 'samples': 9472512, 'steps': 18500, 'loss/train': 2.1793487071990967} 02/25/2022 01:32:20 - INFO - codeparrot_training - Step 18501: {'lr': 0.00036783475355115213, 'samples': 9473024, 'steps': 18501, 'loss/train': 1.4218807220458984} 02/25/2022 01:32:25 - INFO - codeparrot_training - Step 18502: {'lr': 0.0003678203223910382, 'samples': 9473536, 'steps': 18502, 'loss/train': 2.8462727069854736} 02/25/2022 01:32:29 - INFO - codeparrot_training - Step 18503: {'lr': 0.0003678058907262194, 'samples': 9474048, 'steps': 18503, 'loss/train': 1.7098971605300903} 02/25/2022 01:32:34 - INFO - codeparrot_training - Step 18504: {'lr': 0.00036779145855675763, 'samples': 9474560, 'steps': 18504, 'loss/train': 1.4468417167663574} 02/25/2022 01:32:38 - INFO - codeparrot_training - Step 18505: {'lr': 0.00036777702588271455, 'samples': 9475072, 'steps': 18505, 'loss/train': 2.733323574066162} 02/25/2022 01:32:44 - INFO - codeparrot_training - Step 18506: {'lr': 0.0003677625927041522, 'samples': 9475584, 'steps': 18506, 'loss/train': 1.9343047142028809} 02/25/2022 01:32:47 - INFO - codeparrot_training - Step 18507: {'lr': 0.0003677481590211322, 'samples': 9476096, 'steps': 18507, 'loss/train': 2.8910818099975586} 02/25/2022 01:32:53 - INFO - codeparrot_training - Step 18508: {'lr': 0.0003677337248337165, 'samples': 9476608, 'steps': 18508, 'loss/train': 2.4169318675994873} 02/25/2022 01:32:56 - INFO - codeparrot_training - Step 18509: {'lr': 0.0003677192901419669, 'samples': 9477120, 'steps': 18509, 'loss/train': 2.179732322692871} 02/25/2022 01:33:02 - INFO - codeparrot_training - Step 18510: {'lr': 0.0003677048549459453, 'samples': 9477632, 'steps': 18510, 'loss/train': 0.9500709176063538} 02/25/2022 01:33:07 - INFO - codeparrot_training - Step 18511: {'lr': 0.00036769041924571345, 'samples': 9478144, 'steps': 18511, 'loss/train': 2.284907817840576} 02/25/2022 01:33:11 - INFO - codeparrot_training - Step 18512: {'lr': 0.0003676759830413332, 'samples': 9478656, 'steps': 18512, 'loss/train': 2.174438238143921} 02/25/2022 01:33:16 - INFO - codeparrot_training - Step 18513: {'lr': 0.00036766154633286635, 'samples': 9479168, 'steps': 18513, 'loss/train': 1.991797685623169} 02/25/2022 01:33:20 - INFO - codeparrot_training - Step 18514: {'lr': 0.00036764710912037487, 'samples': 9479680, 'steps': 18514, 'loss/train': 2.8928022384643555} 02/25/2022 01:33:26 - INFO - codeparrot_training - Step 18515: {'lr': 0.00036763267140392053, 'samples': 9480192, 'steps': 18515, 'loss/train': 2.113471746444702} 02/25/2022 01:33:29 - INFO - codeparrot_training - Step 18516: {'lr': 0.0003676182331835651, 'samples': 9480704, 'steps': 18516, 'loss/train': 1.1642417907714844} 02/25/2022 01:33:35 - INFO - codeparrot_training - Step 18517: {'lr': 0.00036760379445937067, 'samples': 9481216, 'steps': 18517, 'loss/train': 1.633929967880249} 02/25/2022 01:33:38 - INFO - codeparrot_training - Step 18518: {'lr': 0.0003675893552313988, 'samples': 9481728, 'steps': 18518, 'loss/train': 1.928324580192566} 02/25/2022 01:33:44 - INFO - codeparrot_training - Step 18519: {'lr': 0.0003675749154997115, 'samples': 9482240, 'steps': 18519, 'loss/train': 1.4640511274337769} 02/25/2022 01:33:47 - INFO - codeparrot_training - Step 18520: {'lr': 0.00036756047526437057, 'samples': 9482752, 'steps': 18520, 'loss/train': 1.706226110458374} 02/25/2022 01:33:53 - INFO - codeparrot_training - Step 18521: {'lr': 0.00036754603452543796, 'samples': 9483264, 'steps': 18521, 'loss/train': 1.2723559141159058} 02/25/2022 01:33:57 - INFO - codeparrot_training - Step 18522: {'lr': 0.00036753159328297536, 'samples': 9483776, 'steps': 18522, 'loss/train': 2.3758411407470703} 02/25/2022 01:34:02 - INFO - codeparrot_training - Step 18523: {'lr': 0.00036751715153704483, 'samples': 9484288, 'steps': 18523, 'loss/train': 1.993325114250183} 02/25/2022 01:34:06 - INFO - codeparrot_training - Step 18524: {'lr': 0.0003675027092877081, 'samples': 9484800, 'steps': 18524, 'loss/train': 1.9763778448104858} 02/25/2022 01:34:11 - INFO - codeparrot_training - Step 18525: {'lr': 0.0003674882665350271, 'samples': 9485312, 'steps': 18525, 'loss/train': 1.1987676620483398} 02/25/2022 01:34:15 - INFO - codeparrot_training - Step 18526: {'lr': 0.0003674738232790636, 'samples': 9485824, 'steps': 18526, 'loss/train': 0.5050768852233887} 02/25/2022 01:34:20 - INFO - codeparrot_training - Step 18527: {'lr': 0.0003674593795198796, 'samples': 9486336, 'steps': 18527, 'loss/train': 2.358412265777588} 02/25/2022 01:34:24 - INFO - codeparrot_training - Step 18528: {'lr': 0.00036744493525753697, 'samples': 9486848, 'steps': 18528, 'loss/train': 0.9244344234466553} 02/25/2022 01:34:29 - INFO - codeparrot_training - Step 18529: {'lr': 0.00036743049049209743, 'samples': 9487360, 'steps': 18529, 'loss/train': 1.5023391246795654} 02/25/2022 01:34:33 - INFO - codeparrot_training - Step 18530: {'lr': 0.00036741604522362304, 'samples': 9487872, 'steps': 18530, 'loss/train': 2.2764010429382324} 02/25/2022 01:34:39 - INFO - codeparrot_training - Step 18531: {'lr': 0.00036740159945217556, 'samples': 9488384, 'steps': 18531, 'loss/train': 1.776261568069458} 02/25/2022 01:34:42 - INFO - codeparrot_training - Step 18532: {'lr': 0.0003673871531778169, 'samples': 9488896, 'steps': 18532, 'loss/train': 1.4440553188323975} 02/25/2022 01:34:48 - INFO - codeparrot_training - Step 18533: {'lr': 0.00036737270640060894, 'samples': 9489408, 'steps': 18533, 'loss/train': 2.702498197555542} 02/25/2022 01:34:51 - INFO - codeparrot_training - Step 18534: {'lr': 0.0003673582591206136, 'samples': 9489920, 'steps': 18534, 'loss/train': 2.5179083347320557} 02/25/2022 01:34:57 - INFO - codeparrot_training - Step 18535: {'lr': 0.00036734381133789277, 'samples': 9490432, 'steps': 18535, 'loss/train': 2.2687923908233643} 02/25/2022 01:35:00 - INFO - codeparrot_training - Step 18536: {'lr': 0.00036732936305250826, 'samples': 9490944, 'steps': 18536, 'loss/train': 2.0512826442718506} 02/25/2022 01:35:06 - INFO - codeparrot_training - Step 18537: {'lr': 0.00036731491426452204, 'samples': 9491456, 'steps': 18537, 'loss/train': 1.6194995641708374} 02/25/2022 01:35:09 - INFO - codeparrot_training - Step 18538: {'lr': 0.00036730046497399587, 'samples': 9491968, 'steps': 18538, 'loss/train': 1.5482797622680664} 02/25/2022 01:35:15 - INFO - codeparrot_training - Step 18539: {'lr': 0.0003672860151809919, 'samples': 9492480, 'steps': 18539, 'loss/train': 0.4164883494377136} 02/25/2022 01:35:19 - INFO - codeparrot_training - Step 18540: {'lr': 0.0003672715648855718, 'samples': 9492992, 'steps': 18540, 'loss/train': 2.125943660736084} 02/25/2022 01:35:25 - INFO - codeparrot_training - Step 18541: {'lr': 0.00036725711408779765, 'samples': 9493504, 'steps': 18541, 'loss/train': 2.3170721530914307} 02/25/2022 01:35:28 - INFO - codeparrot_training - Step 18542: {'lr': 0.0003672426627877312, 'samples': 9494016, 'steps': 18542, 'loss/train': 1.6657222509384155} 02/25/2022 01:35:34 - INFO - codeparrot_training - Step 18543: {'lr': 0.0003672282109854344, 'samples': 9494528, 'steps': 18543, 'loss/train': 2.4740726947784424} 02/25/2022 01:35:37 - INFO - codeparrot_training - Step 18544: {'lr': 0.00036721375868096925, 'samples': 9495040, 'steps': 18544, 'loss/train': 1.5595194101333618} 02/25/2022 01:35:43 - INFO - codeparrot_training - Step 18545: {'lr': 0.00036719930587439744, 'samples': 9495552, 'steps': 18545, 'loss/train': 2.0873467922210693} 02/25/2022 01:35:46 - INFO - codeparrot_training - Step 18546: {'lr': 0.00036718485256578116, 'samples': 9496064, 'steps': 18546, 'loss/train': 2.391516923904419} 02/25/2022 01:35:52 - INFO - codeparrot_training - Step 18547: {'lr': 0.00036717039875518203, 'samples': 9496576, 'steps': 18547, 'loss/train': 2.4634275436401367} 02/25/2022 01:35:56 - INFO - codeparrot_training - Step 18548: {'lr': 0.00036715594444266224, 'samples': 9497088, 'steps': 18548, 'loss/train': 2.5728814601898193} 02/25/2022 01:36:01 - INFO - codeparrot_training - Step 18549: {'lr': 0.00036714148962828353, 'samples': 9497600, 'steps': 18549, 'loss/train': 2.1516098976135254} 02/25/2022 01:36:04 - INFO - codeparrot_training - Step 18550: {'lr': 0.0003671270343121079, 'samples': 9498112, 'steps': 18550, 'loss/train': 0.10505091398954391} 02/25/2022 01:36:10 - INFO - codeparrot_training - Step 18551: {'lr': 0.0003671125784941972, 'samples': 9498624, 'steps': 18551, 'loss/train': 2.427187204360962} 02/25/2022 01:36:14 - INFO - codeparrot_training - Step 18552: {'lr': 0.00036709812217461347, 'samples': 9499136, 'steps': 18552, 'loss/train': 1.4499229192733765} 02/25/2022 01:36:17 - INFO - codeparrot_training - Step 18553: {'lr': 0.0003670836653534185, 'samples': 9499648, 'steps': 18553, 'loss/train': 2.1627843379974365} 02/25/2022 01:36:23 - INFO - codeparrot_training - Step 18554: {'lr': 0.0003670692080306743, 'samples': 9500160, 'steps': 18554, 'loss/train': 1.0516340732574463} 02/25/2022 01:36:27 - INFO - codeparrot_training - Step 18555: {'lr': 0.0003670547502064429, 'samples': 9500672, 'steps': 18555, 'loss/train': 2.418111562728882} 02/25/2022 01:36:32 - INFO - codeparrot_training - Step 18556: {'lr': 0.000367040291880786, 'samples': 9501184, 'steps': 18556, 'loss/train': 1.2454153299331665} 02/25/2022 01:36:36 - INFO - codeparrot_training - Step 18557: {'lr': 0.0003670258330537656, 'samples': 9501696, 'steps': 18557, 'loss/train': 1.701690673828125} 02/25/2022 01:36:41 - INFO - codeparrot_training - Step 18558: {'lr': 0.0003670113737254438, 'samples': 9502208, 'steps': 18558, 'loss/train': 1.7004296779632568} 02/25/2022 01:36:45 - INFO - codeparrot_training - Step 18559: {'lr': 0.0003669969138958824, 'samples': 9502720, 'steps': 18559, 'loss/train': 2.023686170578003} 02/25/2022 01:36:50 - INFO - codeparrot_training - Step 18560: {'lr': 0.00036698245356514336, 'samples': 9503232, 'steps': 18560, 'loss/train': 2.2197930812835693} 02/25/2022 01:36:56 - INFO - codeparrot_training - Step 18561: {'lr': 0.00036696799273328864, 'samples': 9503744, 'steps': 18561, 'loss/train': 0.8763425946235657} 02/25/2022 01:36:59 - INFO - codeparrot_training - Step 18562: {'lr': 0.0003669535314003802, 'samples': 9504256, 'steps': 18562, 'loss/train': 1.7204899787902832} 02/25/2022 01:37:05 - INFO - codeparrot_training - Step 18563: {'lr': 0.00036693906956647996, 'samples': 9504768, 'steps': 18563, 'loss/train': 2.2132744789123535} 02/25/2022 01:37:08 - INFO - codeparrot_training - Step 18564: {'lr': 0.0003669246072316498, 'samples': 9505280, 'steps': 18564, 'loss/train': 1.6493988037109375} 02/25/2022 01:37:14 - INFO - codeparrot_training - Step 18565: {'lr': 0.00036691014439595187, 'samples': 9505792, 'steps': 18565, 'loss/train': 2.3527889251708984} 02/25/2022 01:37:17 - INFO - codeparrot_training - Step 18566: {'lr': 0.00036689568105944794, 'samples': 9506304, 'steps': 18566, 'loss/train': 1.448954463005066} 02/25/2022 01:37:24 - INFO - codeparrot_training - Step 18567: {'lr': 0.0003668812172222001, 'samples': 9506816, 'steps': 18567, 'loss/train': 3.2489798069000244} 02/25/2022 01:37:28 - INFO - codeparrot_training - Step 18568: {'lr': 0.0003668667528842702, 'samples': 9507328, 'steps': 18568, 'loss/train': 2.65628981590271} 02/25/2022 01:37:33 - INFO - codeparrot_training - Step 18569: {'lr': 0.0003668522880457202, 'samples': 9507840, 'steps': 18569, 'loss/train': 1.5596976280212402} 02/25/2022 01:37:37 - INFO - codeparrot_training - Step 18570: {'lr': 0.0003668378227066121, 'samples': 9508352, 'steps': 18570, 'loss/train': 2.1314170360565186} 02/25/2022 01:37:40 - INFO - codeparrot_training - Step 18571: {'lr': 0.00036682335686700796, 'samples': 9508864, 'steps': 18571, 'loss/train': 0.7917835116386414} 02/25/2022 01:37:46 - INFO - codeparrot_training - Step 18572: {'lr': 0.00036680889052696954, 'samples': 9509376, 'steps': 18572, 'loss/train': 2.388180732727051} 02/25/2022 01:37:52 - INFO - codeparrot_training - Step 18573: {'lr': 0.00036679442368655897, 'samples': 9509888, 'steps': 18573, 'loss/train': 1.961667776107788} 02/25/2022 01:37:55 - INFO - codeparrot_training - Step 18574: {'lr': 0.00036677995634583815, 'samples': 9510400, 'steps': 18574, 'loss/train': 1.1142741441726685} 02/25/2022 01:38:01 - INFO - codeparrot_training - Step 18575: {'lr': 0.0003667654885048691, 'samples': 9510912, 'steps': 18575, 'loss/train': 1.5314449071884155} 02/25/2022 01:38:04 - INFO - codeparrot_training - Step 18576: {'lr': 0.00036675102016371386, 'samples': 9511424, 'steps': 18576, 'loss/train': 1.826163411140442} 02/25/2022 01:38:11 - INFO - codeparrot_training - Step 18577: {'lr': 0.0003667365513224342, 'samples': 9511936, 'steps': 18577, 'loss/train': 1.1742926836013794} 02/25/2022 01:38:14 - INFO - codeparrot_training - Step 18578: {'lr': 0.0003667220819810923, 'samples': 9512448, 'steps': 18578, 'loss/train': 1.7007622718811035} 02/25/2022 01:38:20 - INFO - codeparrot_training - Step 18579: {'lr': 0.00036670761213975, 'samples': 9512960, 'steps': 18579, 'loss/train': 2.1866090297698975} 02/25/2022 01:38:23 - INFO - codeparrot_training - Step 18580: {'lr': 0.0003666931417984694, 'samples': 9513472, 'steps': 18580, 'loss/train': 1.435978651046753} 02/25/2022 01:38:29 - INFO - codeparrot_training - Step 18581: {'lr': 0.00036667867095731244, 'samples': 9513984, 'steps': 18581, 'loss/train': 2.2813663482666016} 02/25/2022 01:38:32 - INFO - codeparrot_training - Step 18582: {'lr': 0.0003666641996163411, 'samples': 9514496, 'steps': 18582, 'loss/train': 0.24506095051765442} 02/25/2022 01:38:38 - INFO - codeparrot_training - Step 18583: {'lr': 0.0003666497277756173, 'samples': 9515008, 'steps': 18583, 'loss/train': 2.7871553897857666} 02/25/2022 01:38:42 - INFO - codeparrot_training - Step 18584: {'lr': 0.0003666352554352032, 'samples': 9515520, 'steps': 18584, 'loss/train': 1.028181552886963} 02/25/2022 01:38:45 - INFO - codeparrot_training - Step 18585: {'lr': 0.0003666207825951606, 'samples': 9516032, 'steps': 18585, 'loss/train': 2.455007553100586} 02/25/2022 01:38:51 - INFO - codeparrot_training - Step 18586: {'lr': 0.00036660630925555173, 'samples': 9516544, 'steps': 18586, 'loss/train': 2.088752508163452} 02/25/2022 01:38:54 - INFO - codeparrot_training - Step 18587: {'lr': 0.0003665918354164384, 'samples': 9517056, 'steps': 18587, 'loss/train': 2.1828858852386475} 02/25/2022 01:39:00 - INFO - codeparrot_training - Step 18588: {'lr': 0.00036657736107788264, 'samples': 9517568, 'steps': 18588, 'loss/train': 1.4636602401733398} 02/25/2022 01:39:03 - INFO - codeparrot_training - Step 18589: {'lr': 0.00036656288623994647, 'samples': 9518080, 'steps': 18589, 'loss/train': 1.9326659440994263} 02/25/2022 01:39:09 - INFO - codeparrot_training - Step 18590: {'lr': 0.000366548410902692, 'samples': 9518592, 'steps': 18590, 'loss/train': 1.6341596841812134} 02/25/2022 01:39:13 - INFO - codeparrot_training - Step 18591: {'lr': 0.00036653393506618106, 'samples': 9519104, 'steps': 18591, 'loss/train': 1.962467074394226} 02/25/2022 01:39:18 - INFO - codeparrot_training - Step 18592: {'lr': 0.00036651945873047574, 'samples': 9519616, 'steps': 18592, 'loss/train': 2.0787315368652344} 02/25/2022 01:39:22 - INFO - codeparrot_training - Step 18593: {'lr': 0.0003665049818956381, 'samples': 9520128, 'steps': 18593, 'loss/train': 2.3415849208831787} 02/25/2022 01:39:27 - INFO - codeparrot_training - Step 18594: {'lr': 0.0003664905045617301, 'samples': 9520640, 'steps': 18594, 'loss/train': 2.2822210788726807} 02/25/2022 01:39:31 - INFO - codeparrot_training - Step 18595: {'lr': 0.0003664760267288138, 'samples': 9521152, 'steps': 18595, 'loss/train': 3.195042371749878} 02/25/2022 01:39:36 - INFO - codeparrot_training - Step 18596: {'lr': 0.0003664615483969511, 'samples': 9521664, 'steps': 18596, 'loss/train': 1.5851309299468994} 02/25/2022 01:39:40 - INFO - codeparrot_training - Step 18597: {'lr': 0.0003664470695662042, 'samples': 9522176, 'steps': 18597, 'loss/train': 0.9775142669677734} 02/25/2022 01:39:45 - INFO - codeparrot_training - Step 18598: {'lr': 0.000366432590236635, 'samples': 9522688, 'steps': 18598, 'loss/train': 1.7166187763214111} 02/25/2022 01:39:49 - INFO - codeparrot_training - Step 18599: {'lr': 0.0003664181104083055, 'samples': 9523200, 'steps': 18599, 'loss/train': 1.5581121444702148} 02/25/2022 01:39:54 - INFO - codeparrot_training - Step 18600: {'lr': 0.00036640363008127785, 'samples': 9523712, 'steps': 18600, 'loss/train': 2.07365345954895} 02/25/2022 01:39:58 - INFO - codeparrot_training - Step 18601: {'lr': 0.000366389149255614, 'samples': 9524224, 'steps': 18601, 'loss/train': 1.609239935874939} 02/25/2022 01:40:06 - INFO - codeparrot_training - Step 18602: {'lr': 0.00036637466793137605, 'samples': 9524736, 'steps': 18602, 'loss/train': 1.6889331340789795} 02/25/2022 01:40:09 - INFO - codeparrot_training - Step 18603: {'lr': 0.0003663601861086259, 'samples': 9525248, 'steps': 18603, 'loss/train': 1.559266209602356} 02/25/2022 01:40:15 - INFO - codeparrot_training - Step 18604: {'lr': 0.00036634570378742565, 'samples': 9525760, 'steps': 18604, 'loss/train': 2.1429872512817383} 02/25/2022 01:40:18 - INFO - codeparrot_training - Step 18605: {'lr': 0.00036633122096783736, 'samples': 9526272, 'steps': 18605, 'loss/train': 1.1589276790618896} 02/25/2022 01:40:24 - INFO - codeparrot_training - Step 18606: {'lr': 0.00036631673764992307, 'samples': 9526784, 'steps': 18606, 'loss/train': 2.046861410140991} 02/25/2022 01:40:27 - INFO - codeparrot_training - Step 18607: {'lr': 0.00036630225383374476, 'samples': 9527296, 'steps': 18607, 'loss/train': 1.6538593769073486} 02/25/2022 01:40:33 - INFO - codeparrot_training - Step 18608: {'lr': 0.0003662877695193646, 'samples': 9527808, 'steps': 18608, 'loss/train': 1.6244672536849976} 02/25/2022 01:40:36 - INFO - codeparrot_training - Step 18609: {'lr': 0.0003662732847068445, 'samples': 9528320, 'steps': 18609, 'loss/train': 1.192000389099121} 02/25/2022 01:40:42 - INFO - codeparrot_training - Step 18610: {'lr': 0.00036625879939624663, 'samples': 9528832, 'steps': 18610, 'loss/train': 2.6161937713623047} 02/25/2022 01:40:45 - INFO - codeparrot_training - Step 18611: {'lr': 0.000366244313587633, 'samples': 9529344, 'steps': 18611, 'loss/train': 2.635939359664917} 02/25/2022 01:40:53 - INFO - codeparrot_training - Step 18612: {'lr': 0.0003662298272810655, 'samples': 9529856, 'steps': 18612, 'loss/train': 1.6780717372894287} 02/25/2022 01:40:56 - INFO - codeparrot_training - Step 18613: {'lr': 0.00036621534047660647, 'samples': 9530368, 'steps': 18613, 'loss/train': 2.310574769973755} 02/25/2022 01:41:02 - INFO - codeparrot_training - Step 18614: {'lr': 0.00036620085317431777, 'samples': 9530880, 'steps': 18614, 'loss/train': 1.6441854238510132} 02/25/2022 01:41:05 - INFO - codeparrot_training - Step 18615: {'lr': 0.0003661863653742615, 'samples': 9531392, 'steps': 18615, 'loss/train': 1.5445668697357178} 02/25/2022 01:41:11 - INFO - codeparrot_training - Step 18616: {'lr': 0.0003661718770764998, 'samples': 9531904, 'steps': 18616, 'loss/train': 3.113262414932251} 02/25/2022 01:41:14 - INFO - codeparrot_training - Step 18617: {'lr': 0.00036615738828109465, 'samples': 9532416, 'steps': 18617, 'loss/train': 2.387874126434326} 02/25/2022 01:41:20 - INFO - codeparrot_training - Step 18618: {'lr': 0.00036614289898810804, 'samples': 9532928, 'steps': 18618, 'loss/train': 1.5805346965789795} 02/25/2022 01:41:23 - INFO - codeparrot_training - Step 18619: {'lr': 0.00036612840919760225, 'samples': 9533440, 'steps': 18619, 'loss/train': 1.964694857597351} 02/25/2022 01:41:29 - INFO - codeparrot_training - Step 18620: {'lr': 0.00036611391890963913, 'samples': 9533952, 'steps': 18620, 'loss/train': 1.001628041267395} 02/25/2022 01:41:32 - INFO - codeparrot_training - Step 18621: {'lr': 0.00036609942812428087, 'samples': 9534464, 'steps': 18621, 'loss/train': 0.9977004528045654} 02/25/2022 01:41:40 - INFO - codeparrot_training - Step 18622: {'lr': 0.00036608493684158963, 'samples': 9534976, 'steps': 18622, 'loss/train': 1.417203426361084} 02/25/2022 01:41:43 - INFO - codeparrot_training - Step 18623: {'lr': 0.0003660704450616272, 'samples': 9535488, 'steps': 18623, 'loss/train': 2.5538036823272705} 02/25/2022 01:41:49 - INFO - codeparrot_training - Step 18624: {'lr': 0.00036605595278445605, 'samples': 9536000, 'steps': 18624, 'loss/train': 1.2743014097213745} 02/25/2022 01:41:52 - INFO - codeparrot_training - Step 18625: {'lr': 0.0003660414600101379, 'samples': 9536512, 'steps': 18625, 'loss/train': 0.45932769775390625} 02/25/2022 01:41:58 - INFO - codeparrot_training - Step 18626: {'lr': 0.00036602696673873505, 'samples': 9537024, 'steps': 18626, 'loss/train': 1.9580533504486084} 02/25/2022 01:42:01 - INFO - codeparrot_training - Step 18627: {'lr': 0.00036601247297030943, 'samples': 9537536, 'steps': 18627, 'loss/train': 2.022444248199463} 02/25/2022 01:42:07 - INFO - codeparrot_training - Step 18628: {'lr': 0.00036599797870492327, 'samples': 9538048, 'steps': 18628, 'loss/train': 2.1271891593933105} 02/25/2022 01:42:10 - INFO - codeparrot_training - Step 18629: {'lr': 0.0003659834839426387, 'samples': 9538560, 'steps': 18629, 'loss/train': 2.4939234256744385} 02/25/2022 01:42:16 - INFO - codeparrot_training - Step 18630: {'lr': 0.0003659689886835176, 'samples': 9539072, 'steps': 18630, 'loss/train': 1.5407270193099976} 02/25/2022 01:42:23 - INFO - codeparrot_training - Step 18631: {'lr': 0.00036595449292762215, 'samples': 9539584, 'steps': 18631, 'loss/train': 1.8431187868118286} 02/25/2022 01:42:26 - INFO - codeparrot_training - Step 18632: {'lr': 0.00036593999667501457, 'samples': 9540096, 'steps': 18632, 'loss/train': 1.1904419660568237} 02/25/2022 01:42:32 - INFO - codeparrot_training - Step 18633: {'lr': 0.0003659254999257568, 'samples': 9540608, 'steps': 18633, 'loss/train': 1.2419469356536865} 02/25/2022 01:42:35 - INFO - codeparrot_training - Step 18634: {'lr': 0.000365911002679911, 'samples': 9541120, 'steps': 18634, 'loss/train': 1.5306427478790283} 02/25/2022 01:42:41 - INFO - codeparrot_training - Step 18635: {'lr': 0.00036589650493753937, 'samples': 9541632, 'steps': 18635, 'loss/train': 2.0469346046447754} 02/25/2022 01:42:44 - INFO - codeparrot_training - Step 18636: {'lr': 0.00036588200669870376, 'samples': 9542144, 'steps': 18636, 'loss/train': 1.7330089807510376} 02/25/2022 01:42:50 - INFO - codeparrot_training - Step 18637: {'lr': 0.0003658675079634665, 'samples': 9542656, 'steps': 18637, 'loss/train': 2.1543710231781006} 02/25/2022 01:42:53 - INFO - codeparrot_training - Step 18638: {'lr': 0.0003658530087318896, 'samples': 9543168, 'steps': 18638, 'loss/train': 1.717761754989624} 02/25/2022 01:42:59 - INFO - codeparrot_training - Step 18639: {'lr': 0.00036583850900403527, 'samples': 9543680, 'steps': 18639, 'loss/train': 1.5691272020339966} 02/25/2022 01:43:02 - INFO - codeparrot_training - Step 18640: {'lr': 0.00036582400877996547, 'samples': 9544192, 'steps': 18640, 'loss/train': 2.1716856956481934} 02/25/2022 01:43:08 - INFO - codeparrot_training - Step 18641: {'lr': 0.0003658095080597424, 'samples': 9544704, 'steps': 18641, 'loss/train': 1.2038847208023071} 02/25/2022 01:43:11 - INFO - codeparrot_training - Step 18642: {'lr': 0.0003657950068434282, 'samples': 9545216, 'steps': 18642, 'loss/train': 2.194905996322632} 02/25/2022 01:43:17 - INFO - codeparrot_training - Step 18643: {'lr': 0.000365780505131085, 'samples': 9545728, 'steps': 18643, 'loss/train': 2.553783893585205} 02/25/2022 01:43:20 - INFO - codeparrot_training - Step 18644: {'lr': 0.00036576600292277477, 'samples': 9546240, 'steps': 18644, 'loss/train': 1.8347636461257935} 02/25/2022 01:43:26 - INFO - codeparrot_training - Step 18645: {'lr': 0.00036575150021855987, 'samples': 9546752, 'steps': 18645, 'loss/train': 1.0521267652511597} 02/25/2022 01:43:29 - INFO - codeparrot_training - Step 18646: {'lr': 0.00036573699701850223, 'samples': 9547264, 'steps': 18646, 'loss/train': 1.9159833192825317} 02/25/2022 01:43:36 - INFO - codeparrot_training - Step 18647: {'lr': 0.000365722493322664, 'samples': 9547776, 'steps': 18647, 'loss/train': 1.199291706085205} 02/25/2022 01:43:39 - INFO - codeparrot_training - Step 18648: {'lr': 0.0003657079891311075, 'samples': 9548288, 'steps': 18648, 'loss/train': 1.877380609512329} 02/25/2022 01:43:45 - INFO - codeparrot_training - Step 18649: {'lr': 0.00036569348444389456, 'samples': 9548800, 'steps': 18649, 'loss/train': 2.6515560150146484} 02/25/2022 01:43:48 - INFO - codeparrot_training - Step 18650: {'lr': 0.00036567897926108756, 'samples': 9549312, 'steps': 18650, 'loss/train': 3.0750679969787598} 02/25/2022 01:43:54 - INFO - codeparrot_training - Step 18651: {'lr': 0.00036566447358274846, 'samples': 9549824, 'steps': 18651, 'loss/train': 1.3747057914733887} 02/25/2022 01:43:58 - INFO - codeparrot_training - Step 18652: {'lr': 0.0003656499674089396, 'samples': 9550336, 'steps': 18652, 'loss/train': 2.389620304107666} 02/25/2022 01:44:03 - INFO - codeparrot_training - Step 18653: {'lr': 0.0003656354607397229, 'samples': 9550848, 'steps': 18653, 'loss/train': 0.7566519975662231} 02/25/2022 01:44:07 - INFO - codeparrot_training - Step 18654: {'lr': 0.00036562095357516066, 'samples': 9551360, 'steps': 18654, 'loss/train': 0.7429487705230713} 02/25/2022 01:44:12 - INFO - codeparrot_training - Step 18655: {'lr': 0.00036560644591531496, 'samples': 9551872, 'steps': 18655, 'loss/train': 1.5984597206115723} 02/25/2022 01:44:16 - INFO - codeparrot_training - Step 18656: {'lr': 0.00036559193776024794, 'samples': 9552384, 'steps': 18656, 'loss/train': 2.789930582046509} 02/25/2022 01:44:22 - INFO - codeparrot_training - Step 18657: {'lr': 0.0003655774291100218, 'samples': 9552896, 'steps': 18657, 'loss/train': 0.9962130784988403} 02/25/2022 01:44:25 - INFO - codeparrot_training - Step 18658: {'lr': 0.0003655629199646986, 'samples': 9553408, 'steps': 18658, 'loss/train': 1.7426908016204834} 02/25/2022 01:44:31 - INFO - codeparrot_training - Step 18659: {'lr': 0.00036554841032434063, 'samples': 9553920, 'steps': 18659, 'loss/train': 3.1881673336029053} 02/25/2022 01:44:35 - INFO - codeparrot_training - Step 18660: {'lr': 0.00036553390018900984, 'samples': 9554432, 'steps': 18660, 'loss/train': 1.6364705562591553} 02/25/2022 01:44:40 - INFO - codeparrot_training - Step 18661: {'lr': 0.0003655193895587686, 'samples': 9554944, 'steps': 18661, 'loss/train': 1.835903525352478} 02/25/2022 01:44:44 - INFO - codeparrot_training - Step 18662: {'lr': 0.000365504878433679, 'samples': 9555456, 'steps': 18662, 'loss/train': 2.0200867652893066} 02/25/2022 01:44:49 - INFO - codeparrot_training - Step 18663: {'lr': 0.00036549036681380307, 'samples': 9555968, 'steps': 18663, 'loss/train': 0.8442659378051758} 02/25/2022 01:44:53 - INFO - codeparrot_training - Step 18664: {'lr': 0.00036547585469920316, 'samples': 9556480, 'steps': 18664, 'loss/train': 1.6322911977767944} 02/25/2022 01:44:59 - INFO - codeparrot_training - Step 18665: {'lr': 0.00036546134208994137, 'samples': 9556992, 'steps': 18665, 'loss/train': 0.6029117703437805} 02/25/2022 01:45:02 - INFO - codeparrot_training - Step 18666: {'lr': 0.00036544682898607977, 'samples': 9557504, 'steps': 18666, 'loss/train': 0.2979816198348999} 02/25/2022 01:45:06 - INFO - codeparrot_training - Step 18667: {'lr': 0.00036543231538768066, 'samples': 9558016, 'steps': 18667, 'loss/train': 1.5411999225616455} 02/25/2022 01:45:12 - INFO - codeparrot_training - Step 18668: {'lr': 0.00036541780129480616, 'samples': 9558528, 'steps': 18668, 'loss/train': 2.4821763038635254} 02/25/2022 01:45:16 - INFO - codeparrot_training - Step 18669: {'lr': 0.0003654032867075185, 'samples': 9559040, 'steps': 18669, 'loss/train': 2.9620797634124756} 02/25/2022 01:45:21 - INFO - codeparrot_training - Step 18670: {'lr': 0.00036538877162587975, 'samples': 9559552, 'steps': 18670, 'loss/train': 1.7412275075912476} 02/25/2022 01:45:25 - INFO - codeparrot_training - Step 18671: {'lr': 0.00036537425604995214, 'samples': 9560064, 'steps': 18671, 'loss/train': 2.2659592628479004} 02/25/2022 01:45:30 - INFO - codeparrot_training - Step 18672: {'lr': 0.00036535973997979787, 'samples': 9560576, 'steps': 18672, 'loss/train': 1.7775391340255737} 02/25/2022 01:45:33 - INFO - codeparrot_training - Step 18673: {'lr': 0.0003653452234154791, 'samples': 9561088, 'steps': 18673, 'loss/train': 1.4945180416107178} 02/25/2022 01:45:39 - INFO - codeparrot_training - Step 18674: {'lr': 0.000365330706357058, 'samples': 9561600, 'steps': 18674, 'loss/train': 1.8599547147750854} 02/25/2022 01:45:43 - INFO - codeparrot_training - Step 18675: {'lr': 0.0003653161888045968, 'samples': 9562112, 'steps': 18675, 'loss/train': 1.500205636024475} 02/25/2022 01:45:48 - INFO - codeparrot_training - Step 18676: {'lr': 0.0003653016707581577, 'samples': 9562624, 'steps': 18676, 'loss/train': 2.0400757789611816} 02/25/2022 01:45:52 - INFO - codeparrot_training - Step 18677: {'lr': 0.00036528715221780276, 'samples': 9563136, 'steps': 18677, 'loss/train': 1.7568780183792114} 02/25/2022 01:45:57 - INFO - codeparrot_training - Step 18678: {'lr': 0.0003652726331835944, 'samples': 9563648, 'steps': 18678, 'loss/train': 2.6606838703155518} 02/25/2022 01:46:01 - INFO - codeparrot_training - Step 18679: {'lr': 0.00036525811365559457, 'samples': 9564160, 'steps': 18679, 'loss/train': 2.54681134223938} 02/25/2022 01:46:07 - INFO - codeparrot_training - Step 18680: {'lr': 0.0003652435936338656, 'samples': 9564672, 'steps': 18680, 'loss/train': 1.964200735092163} 02/25/2022 01:46:10 - INFO - codeparrot_training - Step 18681: {'lr': 0.0003652290731184697, 'samples': 9565184, 'steps': 18681, 'loss/train': 1.8075731992721558} 02/25/2022 01:46:16 - INFO - codeparrot_training - Step 18682: {'lr': 0.000365214552109469, 'samples': 9565696, 'steps': 18682, 'loss/train': 1.778092861175537} 02/25/2022 01:46:19 - INFO - codeparrot_training - Step 18683: {'lr': 0.0003652000306069258, 'samples': 9566208, 'steps': 18683, 'loss/train': 2.2482798099517822} 02/25/2022 01:46:25 - INFO - codeparrot_training - Step 18684: {'lr': 0.00036518550861090217, 'samples': 9566720, 'steps': 18684, 'loss/train': 2.2633185386657715} 02/25/2022 01:46:29 - INFO - codeparrot_training - Step 18685: {'lr': 0.0003651709861214605, 'samples': 9567232, 'steps': 18685, 'loss/train': 1.3000870943069458} 02/25/2022 01:46:34 - INFO - codeparrot_training - Step 18686: {'lr': 0.0003651564631386628, 'samples': 9567744, 'steps': 18686, 'loss/train': 2.1072256565093994} 02/25/2022 01:46:38 - INFO - codeparrot_training - Step 18687: {'lr': 0.0003651419396625714, 'samples': 9568256, 'steps': 18687, 'loss/train': 2.890693426132202} 02/25/2022 01:46:43 - INFO - codeparrot_training - Step 18688: {'lr': 0.0003651274156932485, 'samples': 9568768, 'steps': 18688, 'loss/train': 1.5545525550842285} 02/25/2022 01:46:47 - INFO - codeparrot_training - Step 18689: {'lr': 0.00036511289123075636, 'samples': 9569280, 'steps': 18689, 'loss/train': 1.9959642887115479} 02/25/2022 01:46:52 - INFO - codeparrot_training - Step 18690: {'lr': 0.0003650983662751571, 'samples': 9569792, 'steps': 18690, 'loss/train': 2.338176727294922} 02/25/2022 01:46:56 - INFO - codeparrot_training - Step 18691: {'lr': 0.000365083840826513, 'samples': 9570304, 'steps': 18691, 'loss/train': 0.6471648812294006} 02/25/2022 01:47:01 - INFO - codeparrot_training - Step 18692: {'lr': 0.00036506931488488627, 'samples': 9570816, 'steps': 18692, 'loss/train': 2.605759620666504} 02/25/2022 01:47:05 - INFO - codeparrot_training - Step 18693: {'lr': 0.0003650547884503391, 'samples': 9571328, 'steps': 18693, 'loss/train': 1.1564210653305054} 02/25/2022 01:47:11 - INFO - codeparrot_training - Step 18694: {'lr': 0.0003650402615229338, 'samples': 9571840, 'steps': 18694, 'loss/train': 1.6388384103775024} 02/25/2022 01:47:15 - INFO - codeparrot_training - Step 18695: {'lr': 0.00036502573410273243, 'samples': 9572352, 'steps': 18695, 'loss/train': 2.0362765789031982} 02/25/2022 01:47:20 - INFO - codeparrot_training - Step 18696: {'lr': 0.0003650112061897975, 'samples': 9572864, 'steps': 18696, 'loss/train': 2.113410234451294} 02/25/2022 01:47:24 - INFO - codeparrot_training - Step 18697: {'lr': 0.000364996677784191, 'samples': 9573376, 'steps': 18697, 'loss/train': 2.0579826831817627} 02/25/2022 01:47:29 - INFO - codeparrot_training - Step 18698: {'lr': 0.00036498214888597524, 'samples': 9573888, 'steps': 18698, 'loss/train': 2.890291452407837} 02/25/2022 01:47:33 - INFO - codeparrot_training - Step 18699: {'lr': 0.0003649676194952125, 'samples': 9574400, 'steps': 18699, 'loss/train': 2.3442039489746094} 02/25/2022 01:47:38 - INFO - codeparrot_training - Step 18700: {'lr': 0.000364953089611965, 'samples': 9574912, 'steps': 18700, 'loss/train': 2.574342727661133} 02/25/2022 01:47:42 - INFO - codeparrot_training - Step 18701: {'lr': 0.00036493855923629495, 'samples': 9575424, 'steps': 18701, 'loss/train': 3.9268972873687744} 02/25/2022 01:47:47 - INFO - codeparrot_training - Step 18702: {'lr': 0.00036492402836826453, 'samples': 9575936, 'steps': 18702, 'loss/train': 2.2675607204437256} 02/25/2022 01:47:54 - INFO - codeparrot_training - Step 18703: {'lr': 0.0003649094970079362, 'samples': 9576448, 'steps': 18703, 'loss/train': 2.112107992172241} 02/25/2022 01:47:57 - INFO - codeparrot_training - Step 18704: {'lr': 0.000364894965155372, 'samples': 9576960, 'steps': 18704, 'loss/train': 0.6963909864425659} 02/25/2022 01:48:03 - INFO - codeparrot_training - Step 18705: {'lr': 0.0003648804328106342, 'samples': 9577472, 'steps': 18705, 'loss/train': 2.042991876602173} 02/25/2022 01:48:07 - INFO - codeparrot_training - Step 18706: {'lr': 0.0003648658999737852, 'samples': 9577984, 'steps': 18706, 'loss/train': 1.4525494575500488} 02/25/2022 01:48:12 - INFO - codeparrot_training - Step 18707: {'lr': 0.0003648513666448871, 'samples': 9578496, 'steps': 18707, 'loss/train': 2.257282018661499} 02/25/2022 01:48:16 - INFO - codeparrot_training - Step 18708: {'lr': 0.0003648368328240022, 'samples': 9579008, 'steps': 18708, 'loss/train': 2.449894428253174} 02/25/2022 01:48:19 - INFO - codeparrot_training - Step 18709: {'lr': 0.00036482229851119287, 'samples': 9579520, 'steps': 18709, 'loss/train': 2.3471951484680176} 02/25/2022 01:48:25 - INFO - codeparrot_training - Step 18710: {'lr': 0.0003648077637065212, 'samples': 9580032, 'steps': 18710, 'loss/train': 2.30021071434021} 02/25/2022 01:48:28 - INFO - codeparrot_training - Step 18711: {'lr': 0.00036479322841004953, 'samples': 9580544, 'steps': 18711, 'loss/train': 1.9611283540725708} 02/25/2022 01:48:34 - INFO - codeparrot_training - Step 18712: {'lr': 0.0003647786926218401, 'samples': 9581056, 'steps': 18712, 'loss/train': 2.7539775371551514} 02/25/2022 01:48:37 - INFO - codeparrot_training - Step 18713: {'lr': 0.00036476415634195523, 'samples': 9581568, 'steps': 18713, 'loss/train': 0.16106773912906647} 02/25/2022 01:48:43 - INFO - codeparrot_training - Step 18714: {'lr': 0.00036474961957045715, 'samples': 9582080, 'steps': 18714, 'loss/train': 2.0612494945526123} 02/25/2022 01:48:47 - INFO - codeparrot_training - Step 18715: {'lr': 0.00036473508230740816, 'samples': 9582592, 'steps': 18715, 'loss/train': 2.2212347984313965} 02/25/2022 01:48:52 - INFO - codeparrot_training - Step 18716: {'lr': 0.00036472054455287053, 'samples': 9583104, 'steps': 18716, 'loss/train': 1.5926774740219116} 02/25/2022 01:48:56 - INFO - codeparrot_training - Step 18717: {'lr': 0.0003647060063069064, 'samples': 9583616, 'steps': 18717, 'loss/train': 1.8369320631027222} 02/25/2022 01:49:01 - INFO - codeparrot_training - Step 18718: {'lr': 0.0003646914675695783, 'samples': 9584128, 'steps': 18718, 'loss/train': 1.7059699296951294} 02/25/2022 01:49:05 - INFO - codeparrot_training - Step 18719: {'lr': 0.0003646769283409483, 'samples': 9584640, 'steps': 18719, 'loss/train': 2.582383155822754} 02/25/2022 01:49:11 - INFO - codeparrot_training - Step 18720: {'lr': 0.0003646623886210788, 'samples': 9585152, 'steps': 18720, 'loss/train': 2.0679750442504883} 02/25/2022 01:49:14 - INFO - codeparrot_training - Step 18721: {'lr': 0.00036464784841003196, 'samples': 9585664, 'steps': 18721, 'loss/train': 1.9153096675872803} 02/25/2022 01:49:20 - INFO - codeparrot_training - Step 18722: {'lr': 0.0003646333077078702, 'samples': 9586176, 'steps': 18722, 'loss/train': 1.320304274559021} 02/25/2022 01:49:23 - INFO - codeparrot_training - Step 18723: {'lr': 0.0003646187665146557, 'samples': 9586688, 'steps': 18723, 'loss/train': 2.1479573249816895} 02/25/2022 01:49:29 - INFO - codeparrot_training - Step 18724: {'lr': 0.00036460422483045084, 'samples': 9587200, 'steps': 18724, 'loss/train': 2.007450819015503} 02/25/2022 01:49:32 - INFO - codeparrot_training - Step 18725: {'lr': 0.0003645896826553178, 'samples': 9587712, 'steps': 18725, 'loss/train': 0.10037413239479065} 02/25/2022 01:49:38 - INFO - codeparrot_training - Step 18726: {'lr': 0.0003645751399893191, 'samples': 9588224, 'steps': 18726, 'loss/train': 2.3800787925720215} 02/25/2022 01:49:42 - INFO - codeparrot_training - Step 18727: {'lr': 0.0003645605968325167, 'samples': 9588736, 'steps': 18727, 'loss/train': 2.0359084606170654} 02/25/2022 01:49:47 - INFO - codeparrot_training - Step 18728: {'lr': 0.00036454605318497323, 'samples': 9589248, 'steps': 18728, 'loss/train': 2.33123779296875} 02/25/2022 01:49:51 - INFO - codeparrot_training - Step 18729: {'lr': 0.00036453150904675074, 'samples': 9589760, 'steps': 18729, 'loss/train': 1.9773380756378174} 02/25/2022 01:49:56 - INFO - codeparrot_training - Step 18730: {'lr': 0.00036451696441791164, 'samples': 9590272, 'steps': 18730, 'loss/train': 1.75650954246521} 02/25/2022 01:50:00 - INFO - codeparrot_training - Step 18731: {'lr': 0.0003645024192985183, 'samples': 9590784, 'steps': 18731, 'loss/train': 1.3827403783798218} 02/25/2022 01:50:05 - INFO - codeparrot_training - Step 18732: {'lr': 0.0003644878736886329, 'samples': 9591296, 'steps': 18732, 'loss/train': 0.9074762463569641} 02/25/2022 01:50:09 - INFO - codeparrot_training - Step 18733: {'lr': 0.0003644733275883179, 'samples': 9591808, 'steps': 18733, 'loss/train': 2.2755424976348877} 02/25/2022 01:50:14 - INFO - codeparrot_training - Step 18734: {'lr': 0.00036445878099763534, 'samples': 9592320, 'steps': 18734, 'loss/train': 1.992601990699768} 02/25/2022 01:50:18 - INFO - codeparrot_training - Step 18735: {'lr': 0.00036444423391664783, 'samples': 9592832, 'steps': 18735, 'loss/train': 1.757514476776123} 02/25/2022 01:50:23 - INFO - codeparrot_training - Step 18736: {'lr': 0.0003644296863454175, 'samples': 9593344, 'steps': 18736, 'loss/train': 2.7851083278656006} 02/25/2022 01:50:27 - INFO - codeparrot_training - Step 18737: {'lr': 0.0003644151382840068, 'samples': 9593856, 'steps': 18737, 'loss/train': 1.7062655687332153} 02/25/2022 01:50:32 - INFO - codeparrot_training - Step 18738: {'lr': 0.00036440058973247793, 'samples': 9594368, 'steps': 18738, 'loss/train': 1.3448009490966797} 02/25/2022 01:50:36 - INFO - codeparrot_training - Step 18739: {'lr': 0.0003643860406908933, 'samples': 9594880, 'steps': 18739, 'loss/train': 1.472324252128601} 02/25/2022 01:50:41 - INFO - codeparrot_training - Step 18740: {'lr': 0.0003643714911593151, 'samples': 9595392, 'steps': 18740, 'loss/train': 2.843435764312744} 02/25/2022 01:50:45 - INFO - codeparrot_training - Step 18741: {'lr': 0.00036435694113780585, 'samples': 9595904, 'steps': 18741, 'loss/train': 1.2746641635894775} 02/25/2022 01:50:51 - INFO - codeparrot_training - Step 18742: {'lr': 0.0003643423906264277, 'samples': 9596416, 'steps': 18742, 'loss/train': 1.2105600833892822} 02/25/2022 01:50:54 - INFO - codeparrot_training - Step 18743: {'lr': 0.0003643278396252431, 'samples': 9596928, 'steps': 18743, 'loss/train': 3.263894557952881} 02/25/2022 01:51:00 - INFO - codeparrot_training - Step 18744: {'lr': 0.0003643132881343144, 'samples': 9597440, 'steps': 18744, 'loss/train': 2.2403008937835693} 02/25/2022 01:51:03 - INFO - codeparrot_training - Step 18745: {'lr': 0.00036429873615370374, 'samples': 9597952, 'steps': 18745, 'loss/train': 1.7017784118652344} 02/25/2022 01:51:09 - INFO - codeparrot_training - Step 18746: {'lr': 0.00036428418368347363, 'samples': 9598464, 'steps': 18746, 'loss/train': 1.3000973463058472} 02/25/2022 01:51:12 - INFO - codeparrot_training - Step 18747: {'lr': 0.0003642696307236864, 'samples': 9598976, 'steps': 18747, 'loss/train': 1.8979524374008179} 02/25/2022 01:51:18 - INFO - codeparrot_training - Step 18748: {'lr': 0.0003642550772744044, 'samples': 9599488, 'steps': 18748, 'loss/train': 1.9144237041473389} 02/25/2022 01:51:21 - INFO - codeparrot_training - Step 18749: {'lr': 0.0003642405233356898, 'samples': 9600000, 'steps': 18749, 'loss/train': 1.6354658603668213} 02/25/2022 01:51:27 - INFO - codeparrot_training - Step 18750: {'lr': 0.00036422596890760517, 'samples': 9600512, 'steps': 18750, 'loss/train': 1.7698363065719604} 02/25/2022 01:51:30 - INFO - codeparrot_training - Step 18751: {'lr': 0.00036421141399021274, 'samples': 9601024, 'steps': 18751, 'loss/train': 0.7682331800460815} 02/25/2022 01:51:36 - INFO - codeparrot_training - Step 18752: {'lr': 0.00036419685858357485, 'samples': 9601536, 'steps': 18752, 'loss/train': 2.680948495864868} 02/25/2022 01:51:40 - INFO - codeparrot_training - Step 18753: {'lr': 0.00036418230268775393, 'samples': 9602048, 'steps': 18753, 'loss/train': 1.591759204864502} 02/25/2022 01:51:45 - INFO - codeparrot_training - Step 18754: {'lr': 0.0003641677463028123, 'samples': 9602560, 'steps': 18754, 'loss/train': 2.06766939163208} 02/25/2022 01:51:51 - INFO - codeparrot_training - Step 18755: {'lr': 0.0003641531894288122, 'samples': 9603072, 'steps': 18755, 'loss/train': 1.9107542037963867} 02/25/2022 01:51:54 - INFO - codeparrot_training - Step 18756: {'lr': 0.0003641386320658161, 'samples': 9603584, 'steps': 18756, 'loss/train': 0.9932055473327637} 02/25/2022 01:51:58 - INFO - codeparrot_training - Step 18757: {'lr': 0.00036412407421388646, 'samples': 9604096, 'steps': 18757, 'loss/train': 1.7812520265579224} 02/25/2022 01:52:03 - INFO - codeparrot_training - Step 18758: {'lr': 0.00036410951587308545, 'samples': 9604608, 'steps': 18758, 'loss/train': 2.319784641265869} 02/25/2022 01:52:09 - INFO - codeparrot_training - Step 18759: {'lr': 0.00036409495704347553, 'samples': 9605120, 'steps': 18759, 'loss/train': 3.9322290420532227} 02/25/2022 01:52:12 - INFO - codeparrot_training - Step 18760: {'lr': 0.000364080397725119, 'samples': 9605632, 'steps': 18760, 'loss/train': 1.9735074043273926} 02/25/2022 01:52:18 - INFO - codeparrot_training - Step 18761: {'lr': 0.00036406583791807824, 'samples': 9606144, 'steps': 18761, 'loss/train': 1.642637848854065} 02/25/2022 01:52:22 - INFO - codeparrot_training - Step 18762: {'lr': 0.0003640512776224157, 'samples': 9606656, 'steps': 18762, 'loss/train': 2.916508197784424} 02/25/2022 01:52:28 - INFO - codeparrot_training - Step 18763: {'lr': 0.0003640367168381937, 'samples': 9607168, 'steps': 18763, 'loss/train': 2.194520950317383} 02/25/2022 01:52:31 - INFO - codeparrot_training - Step 18764: {'lr': 0.0003640221555654747, 'samples': 9607680, 'steps': 18764, 'loss/train': 2.134033441543579} 02/25/2022 01:52:34 - INFO - codeparrot_training - Step 18765: {'lr': 0.00036400759380432083, 'samples': 9608192, 'steps': 18765, 'loss/train': 2.271791934967041} 02/25/2022 01:52:40 - INFO - codeparrot_training - Step 18766: {'lr': 0.00036399303155479476, 'samples': 9608704, 'steps': 18766, 'loss/train': 1.747043490409851} 02/25/2022 01:52:44 - INFO - codeparrot_training - Step 18767: {'lr': 0.00036397846881695866, 'samples': 9609216, 'steps': 18767, 'loss/train': 2.0768213272094727} 02/25/2022 01:52:49 - INFO - codeparrot_training - Step 18768: {'lr': 0.00036396390559087505, 'samples': 9609728, 'steps': 18768, 'loss/train': 2.3907644748687744} 02/25/2022 01:52:53 - INFO - codeparrot_training - Step 18769: {'lr': 0.0003639493418766062, 'samples': 9610240, 'steps': 18769, 'loss/train': 1.804693341255188} 02/25/2022 01:52:58 - INFO - codeparrot_training - Step 18770: {'lr': 0.0003639347776742146, 'samples': 9610752, 'steps': 18770, 'loss/train': 2.2344589233398438} 02/25/2022 01:53:02 - INFO - codeparrot_training - Step 18771: {'lr': 0.00036392021298376257, 'samples': 9611264, 'steps': 18771, 'loss/train': 1.1538872718811035} 02/25/2022 01:53:08 - INFO - codeparrot_training - Step 18772: {'lr': 0.0003639056478053125, 'samples': 9611776, 'steps': 18772, 'loss/train': 1.552165150642395} 02/25/2022 01:53:11 - INFO - codeparrot_training - Step 18773: {'lr': 0.0003638910821389268, 'samples': 9612288, 'steps': 18773, 'loss/train': 1.2958184480667114} 02/25/2022 01:53:17 - INFO - codeparrot_training - Step 18774: {'lr': 0.0003638765159846679, 'samples': 9612800, 'steps': 18774, 'loss/train': 1.857056975364685} 02/25/2022 01:53:20 - INFO - codeparrot_training - Step 18775: {'lr': 0.0003638619493425982, 'samples': 9613312, 'steps': 18775, 'loss/train': 0.22676512598991394} 02/25/2022 01:53:26 - INFO - codeparrot_training - Step 18776: {'lr': 0.00036384738221278, 'samples': 9613824, 'steps': 18776, 'loss/train': 1.1256705522537231} 02/25/2022 01:53:30 - INFO - codeparrot_training - Step 18777: {'lr': 0.0003638328145952758, 'samples': 9614336, 'steps': 18777, 'loss/train': 1.938619613647461} 02/25/2022 01:53:35 - INFO - codeparrot_training - Step 18778: {'lr': 0.0003638182464901479, 'samples': 9614848, 'steps': 18778, 'loss/train': 1.5208390951156616} 02/25/2022 01:53:39 - INFO - codeparrot_training - Step 18779: {'lr': 0.00036380367789745887, 'samples': 9615360, 'steps': 18779, 'loss/train': 2.1775331497192383} 02/25/2022 01:53:44 - INFO - codeparrot_training - Step 18780: {'lr': 0.00036378910881727095, 'samples': 9615872, 'steps': 18780, 'loss/train': 1.8853918313980103} 02/25/2022 01:53:48 - INFO - codeparrot_training - Step 18781: {'lr': 0.00036377453924964665, 'samples': 9616384, 'steps': 18781, 'loss/train': 2.315471649169922} 02/25/2022 01:53:53 - INFO - codeparrot_training - Step 18782: {'lr': 0.0003637599691946484, 'samples': 9616896, 'steps': 18782, 'loss/train': 1.3137383460998535} 02/25/2022 01:53:57 - INFO - codeparrot_training - Step 18783: {'lr': 0.00036374539865233847, 'samples': 9617408, 'steps': 18783, 'loss/train': 1.7385649681091309} 02/25/2022 01:54:02 - INFO - codeparrot_training - Step 18784: {'lr': 0.00036373082762277943, 'samples': 9617920, 'steps': 18784, 'loss/train': 2.9578235149383545} 02/25/2022 01:54:06 - INFO - codeparrot_training - Step 18785: {'lr': 0.00036371625610603366, 'samples': 9618432, 'steps': 18785, 'loss/train': 1.29164719581604} 02/25/2022 01:54:11 - INFO - codeparrot_training - Step 18786: {'lr': 0.0003637016841021635, 'samples': 9618944, 'steps': 18786, 'loss/train': 2.966223955154419} 02/25/2022 01:54:15 - INFO - codeparrot_training - Step 18787: {'lr': 0.00036368711161123145, 'samples': 9619456, 'steps': 18787, 'loss/train': 1.7146046161651611} 02/25/2022 01:54:21 - INFO - codeparrot_training - Step 18788: {'lr': 0.0003636725386332999, 'samples': 9619968, 'steps': 18788, 'loss/train': 1.6252731084823608} 02/25/2022 01:54:24 - INFO - codeparrot_training - Step 18789: {'lr': 0.0003636579651684313, 'samples': 9620480, 'steps': 18789, 'loss/train': 2.730056047439575} 02/25/2022 01:54:30 - INFO - codeparrot_training - Step 18790: {'lr': 0.0003636433912166881, 'samples': 9620992, 'steps': 18790, 'loss/train': 3.086652994155884} 02/25/2022 01:54:33 - INFO - codeparrot_training - Step 18791: {'lr': 0.00036362881677813266, 'samples': 9621504, 'steps': 18791, 'loss/train': 0.9975144267082214} 02/25/2022 01:54:39 - INFO - codeparrot_training - Step 18792: {'lr': 0.00036361424185282743, 'samples': 9622016, 'steps': 18792, 'loss/train': 1.7287423610687256} 02/25/2022 01:54:42 - INFO - codeparrot_training - Step 18793: {'lr': 0.00036359966644083485, 'samples': 9622528, 'steps': 18793, 'loss/train': 0.9263642430305481} 02/25/2022 01:54:48 - INFO - codeparrot_training - Step 18794: {'lr': 0.0003635850905422175, 'samples': 9623040, 'steps': 18794, 'loss/train': 1.5517878532409668} 02/25/2022 01:54:51 - INFO - codeparrot_training - Step 18795: {'lr': 0.00036357051415703755, 'samples': 9623552, 'steps': 18795, 'loss/train': 1.856420874595642} 02/25/2022 01:54:57 - INFO - codeparrot_training - Step 18796: {'lr': 0.00036355593728535767, 'samples': 9624064, 'steps': 18796, 'loss/train': 1.5313332080841064} 02/25/2022 01:55:00 - INFO - codeparrot_training - Step 18797: {'lr': 0.00036354135992724013, 'samples': 9624576, 'steps': 18797, 'loss/train': 1.5411643981933594} 02/25/2022 01:55:07 - INFO - codeparrot_training - Step 18798: {'lr': 0.0003635267820827476, 'samples': 9625088, 'steps': 18798, 'loss/train': 0.7101810574531555} 02/25/2022 01:55:10 - INFO - codeparrot_training - Step 18799: {'lr': 0.0003635122037519422, 'samples': 9625600, 'steps': 18799, 'loss/train': 2.0135602951049805} 02/25/2022 01:55:16 - INFO - codeparrot_training - Step 18800: {'lr': 0.00036349762493488667, 'samples': 9626112, 'steps': 18800, 'loss/train': 1.9933468103408813} 02/25/2022 01:55:19 - INFO - codeparrot_training - Step 18801: {'lr': 0.0003634830456316434, 'samples': 9626624, 'steps': 18801, 'loss/train': 1.921012043952942} 02/25/2022 01:55:25 - INFO - codeparrot_training - Step 18802: {'lr': 0.00036346846584227473, 'samples': 9627136, 'steps': 18802, 'loss/train': 2.269662380218506} 02/25/2022 01:55:28 - INFO - codeparrot_training - Step 18803: {'lr': 0.0003634538855668432, 'samples': 9627648, 'steps': 18803, 'loss/train': 1.8927403688430786} 02/25/2022 01:55:34 - INFO - codeparrot_training - Step 18804: {'lr': 0.00036343930480541123, 'samples': 9628160, 'steps': 18804, 'loss/train': 2.0032622814178467} 02/25/2022 01:55:37 - INFO - codeparrot_training - Step 18805: {'lr': 0.0003634247235580413, 'samples': 9628672, 'steps': 18805, 'loss/train': 2.2543907165527344} 02/25/2022 01:55:43 - INFO - codeparrot_training - Step 18806: {'lr': 0.0003634101418247959, 'samples': 9629184, 'steps': 18806, 'loss/train': 2.0594019889831543} 02/25/2022 01:55:46 - INFO - codeparrot_training - Step 18807: {'lr': 0.0003633955596057374, 'samples': 9629696, 'steps': 18807, 'loss/train': 3.0865230560302734} 02/25/2022 01:55:52 - INFO - codeparrot_training - Step 18808: {'lr': 0.00036338097690092843, 'samples': 9630208, 'steps': 18808, 'loss/train': 2.2409377098083496} 02/25/2022 01:55:56 - INFO - codeparrot_training - Step 18809: {'lr': 0.0003633663937104313, 'samples': 9630720, 'steps': 18809, 'loss/train': 1.6155309677124023} 02/25/2022 01:56:01 - INFO - codeparrot_training - Step 18810: {'lr': 0.0003633518100343085, 'samples': 9631232, 'steps': 18810, 'loss/train': 4.343753337860107} 02/25/2022 01:56:05 - INFO - codeparrot_training - Step 18811: {'lr': 0.0003633372258726226, 'samples': 9631744, 'steps': 18811, 'loss/train': 2.3139777183532715} 02/25/2022 01:56:11 - INFO - codeparrot_training - Step 18812: {'lr': 0.00036332264122543594, 'samples': 9632256, 'steps': 18812, 'loss/train': 1.5373393297195435} 02/25/2022 01:56:14 - INFO - codeparrot_training - Step 18813: {'lr': 0.0003633080560928111, 'samples': 9632768, 'steps': 18813, 'loss/train': 0.41178491711616516} 02/25/2022 01:56:20 - INFO - codeparrot_training - Step 18814: {'lr': 0.0003632934704748106, 'samples': 9633280, 'steps': 18814, 'loss/train': 1.730994701385498} 02/25/2022 01:56:23 - INFO - codeparrot_training - Step 18815: {'lr': 0.00036327888437149674, 'samples': 9633792, 'steps': 18815, 'loss/train': 2.9415700435638428} 02/25/2022 01:56:28 - INFO - codeparrot_training - Step 18816: {'lr': 0.00036326429778293223, 'samples': 9634304, 'steps': 18816, 'loss/train': 0.736484706401825} 02/25/2022 01:56:32 - INFO - codeparrot_training - Step 18817: {'lr': 0.00036324971070917934, 'samples': 9634816, 'steps': 18817, 'loss/train': 1.6742634773254395} 02/25/2022 01:56:38 - INFO - codeparrot_training - Step 18818: {'lr': 0.00036323512315030067, 'samples': 9635328, 'steps': 18818, 'loss/train': 1.4184633493423462} 02/25/2022 01:56:42 - INFO - codeparrot_training - Step 18819: {'lr': 0.0003632205351063587, 'samples': 9635840, 'steps': 18819, 'loss/train': 2.167532205581665} 02/25/2022 01:56:47 - INFO - codeparrot_training - Step 18820: {'lr': 0.0003632059465774159, 'samples': 9636352, 'steps': 18820, 'loss/train': 2.1325812339782715} 02/25/2022 01:56:50 - INFO - codeparrot_training - Step 18821: {'lr': 0.0003631913575635348, 'samples': 9636864, 'steps': 18821, 'loss/train': 1.9987629652023315} 02/25/2022 01:56:56 - INFO - codeparrot_training - Step 18822: {'lr': 0.00036317676806477784, 'samples': 9637376, 'steps': 18822, 'loss/train': 2.838286876678467} 02/25/2022 01:57:00 - INFO - codeparrot_training - Step 18823: {'lr': 0.0003631621780812075, 'samples': 9637888, 'steps': 18823, 'loss/train': 0.4584333300590515} 02/25/2022 01:57:05 - INFO - codeparrot_training - Step 18824: {'lr': 0.00036314758761288643, 'samples': 9638400, 'steps': 18824, 'loss/train': 2.645183801651001} 02/25/2022 01:57:11 - INFO - codeparrot_training - Step 18825: {'lr': 0.0003631329966598769, 'samples': 9638912, 'steps': 18825, 'loss/train': 2.121828079223633} 02/25/2022 01:57:14 - INFO - codeparrot_training - Step 18826: {'lr': 0.0003631184052222416, 'samples': 9639424, 'steps': 18826, 'loss/train': 0.2151978760957718} 02/25/2022 01:57:20 - INFO - codeparrot_training - Step 18827: {'lr': 0.00036310381330004296, 'samples': 9639936, 'steps': 18827, 'loss/train': 1.7303944826126099} 02/25/2022 01:57:23 - INFO - codeparrot_training - Step 18828: {'lr': 0.0003630892208933435, 'samples': 9640448, 'steps': 18828, 'loss/train': 0.845332145690918} 02/25/2022 01:57:29 - INFO - codeparrot_training - Step 18829: {'lr': 0.00036307462800220575, 'samples': 9640960, 'steps': 18829, 'loss/train': 2.6933765411376953} 02/25/2022 01:57:32 - INFO - codeparrot_training - Step 18830: {'lr': 0.0003630600346266922, 'samples': 9641472, 'steps': 18830, 'loss/train': 2.6756937503814697} 02/25/2022 01:57:38 - INFO - codeparrot_training - Step 18831: {'lr': 0.00036304544076686527, 'samples': 9641984, 'steps': 18831, 'loss/train': 1.8550463914871216} 02/25/2022 01:57:41 - INFO - codeparrot_training - Step 18832: {'lr': 0.0003630308464227877, 'samples': 9642496, 'steps': 18832, 'loss/train': 1.7160649299621582} 02/25/2022 01:57:48 - INFO - codeparrot_training - Step 18833: {'lr': 0.0003630162515945218, 'samples': 9643008, 'steps': 18833, 'loss/train': 0.8779038190841675} 02/25/2022 01:57:51 - INFO - codeparrot_training - Step 18834: {'lr': 0.00036300165628213015, 'samples': 9643520, 'steps': 18834, 'loss/train': 0.9629799723625183} 02/25/2022 01:57:57 - INFO - codeparrot_training - Step 18835: {'lr': 0.0003629870604856754, 'samples': 9644032, 'steps': 18835, 'loss/train': 1.703185796737671} 02/25/2022 01:58:00 - INFO - codeparrot_training - Step 18836: {'lr': 0.0003629724642052198, 'samples': 9644544, 'steps': 18836, 'loss/train': 1.658734679222107} 02/25/2022 01:58:06 - INFO - codeparrot_training - Step 18837: {'lr': 0.00036295786744082616, 'samples': 9645056, 'steps': 18837, 'loss/train': 3.145209312438965} 02/25/2022 01:58:09 - INFO - codeparrot_training - Step 18838: {'lr': 0.0003629432701925568, 'samples': 9645568, 'steps': 18838, 'loss/train': 1.8590718507766724} 02/25/2022 01:58:15 - INFO - codeparrot_training - Step 18839: {'lr': 0.0003629286724604744, 'samples': 9646080, 'steps': 18839, 'loss/train': 1.4280321598052979} 02/25/2022 01:58:18 - INFO - codeparrot_training - Step 18840: {'lr': 0.0003629140742446414, 'samples': 9646592, 'steps': 18840, 'loss/train': 1.3024110794067383} 02/25/2022 01:58:23 - INFO - codeparrot_training - Step 18841: {'lr': 0.00036289947554512034, 'samples': 9647104, 'steps': 18841, 'loss/train': 2.747363328933716} 02/25/2022 01:58:27 - INFO - codeparrot_training - Step 18842: {'lr': 0.0003628848763619738, 'samples': 9647616, 'steps': 18842, 'loss/train': 1.8926026821136475} 02/25/2022 01:58:33 - INFO - codeparrot_training - Step 18843: {'lr': 0.0003628702766952643, 'samples': 9648128, 'steps': 18843, 'loss/train': 2.5643370151519775} 02/25/2022 01:58:37 - INFO - codeparrot_training - Step 18844: {'lr': 0.00036285567654505433, 'samples': 9648640, 'steps': 18844, 'loss/train': 2.6005680561065674} 02/25/2022 01:58:42 - INFO - codeparrot_training - Step 18845: {'lr': 0.00036284107591140653, 'samples': 9649152, 'steps': 18845, 'loss/train': 2.3802216053009033} 02/25/2022 01:58:46 - INFO - codeparrot_training - Step 18846: {'lr': 0.0003628264747943834, 'samples': 9649664, 'steps': 18846, 'loss/train': 1.682621955871582} 02/25/2022 01:58:51 - INFO - codeparrot_training - Step 18847: {'lr': 0.0003628118731940475, 'samples': 9650176, 'steps': 18847, 'loss/train': 2.3060736656188965} 02/25/2022 01:58:55 - INFO - codeparrot_training - Step 18848: {'lr': 0.00036279727111046127, 'samples': 9650688, 'steps': 18848, 'loss/train': 2.185255765914917} 02/25/2022 01:59:00 - INFO - codeparrot_training - Step 18849: {'lr': 0.0003627826685436874, 'samples': 9651200, 'steps': 18849, 'loss/train': 2.854978084564209} 02/25/2022 01:59:04 - INFO - codeparrot_training - Step 18850: {'lr': 0.00036276806549378836, 'samples': 9651712, 'steps': 18850, 'loss/train': 1.5781532526016235} 02/25/2022 01:59:09 - INFO - codeparrot_training - Step 18851: {'lr': 0.0003627534619608268, 'samples': 9652224, 'steps': 18851, 'loss/train': 2.168938636779785} 02/25/2022 01:59:13 - INFO - codeparrot_training - Step 18852: {'lr': 0.00036273885794486514, 'samples': 9652736, 'steps': 18852, 'loss/train': 1.8750176429748535} 02/25/2022 01:59:19 - INFO - codeparrot_training - Step 18853: {'lr': 0.00036272425344596607, 'samples': 9653248, 'steps': 18853, 'loss/train': 2.0590384006500244} 02/25/2022 01:59:23 - INFO - codeparrot_training - Step 18854: {'lr': 0.000362709648464192, 'samples': 9653760, 'steps': 18854, 'loss/train': 1.1000386476516724} 02/25/2022 01:59:28 - INFO - codeparrot_training - Step 18855: {'lr': 0.00036269504299960573, 'samples': 9654272, 'steps': 18855, 'loss/train': 2.4076638221740723} 02/25/2022 01:59:32 - INFO - codeparrot_training - Step 18856: {'lr': 0.00036268043705226953, 'samples': 9654784, 'steps': 18856, 'loss/train': 2.177995204925537} 02/25/2022 01:59:37 - INFO - codeparrot_training - Step 18857: {'lr': 0.00036266583062224625, 'samples': 9655296, 'steps': 18857, 'loss/train': 2.24255108833313} 02/25/2022 01:59:41 - INFO - codeparrot_training - Step 18858: {'lr': 0.0003626512237095982, 'samples': 9655808, 'steps': 18858, 'loss/train': 2.230933427810669} 02/25/2022 01:59:46 - INFO - codeparrot_training - Step 18859: {'lr': 0.00036263661631438814, 'samples': 9656320, 'steps': 18859, 'loss/train': 2.467299699783325} 02/25/2022 01:59:50 - INFO - codeparrot_training - Step 18860: {'lr': 0.0003626220084366786, 'samples': 9656832, 'steps': 18860, 'loss/train': 1.9779781103134155} 02/25/2022 01:59:55 - INFO - codeparrot_training - Step 18861: {'lr': 0.00036260740007653216, 'samples': 9657344, 'steps': 18861, 'loss/train': 1.9783228635787964} 02/25/2022 01:59:59 - INFO - codeparrot_training - Step 18862: {'lr': 0.00036259279123401127, 'samples': 9657856, 'steps': 18862, 'loss/train': 3.0171501636505127} 02/25/2022 02:00:05 - INFO - codeparrot_training - Step 18863: {'lr': 0.0003625781819091787, 'samples': 9658368, 'steps': 18863, 'loss/train': 1.9603748321533203} 02/25/2022 02:00:08 - INFO - codeparrot_training - Step 18864: {'lr': 0.0003625635721020969, 'samples': 9658880, 'steps': 18864, 'loss/train': 2.323228120803833} 02/25/2022 02:00:14 - INFO - codeparrot_training - Step 18865: {'lr': 0.00036254896181282846, 'samples': 9659392, 'steps': 18865, 'loss/train': 0.31502848863601685} 02/25/2022 02:00:17 - INFO - codeparrot_training - Step 18866: {'lr': 0.0003625343510414362, 'samples': 9659904, 'steps': 18866, 'loss/train': 1.4574376344680786} 02/25/2022 02:00:23 - INFO - codeparrot_training - Step 18867: {'lr': 0.0003625197397879823, 'samples': 9660416, 'steps': 18867, 'loss/train': 2.1850383281707764} 02/25/2022 02:00:26 - INFO - codeparrot_training - Step 18868: {'lr': 0.0003625051280525297, 'samples': 9660928, 'steps': 18868, 'loss/train': 2.6404757499694824} 02/25/2022 02:00:30 - INFO - codeparrot_training - Step 18869: {'lr': 0.00036249051583514075, 'samples': 9661440, 'steps': 18869, 'loss/train': 2.3283097743988037} 02/25/2022 02:00:36 - INFO - codeparrot_training - Step 18870: {'lr': 0.00036247590313587827, 'samples': 9661952, 'steps': 18870, 'loss/train': 2.741659641265869} 02/25/2022 02:00:39 - INFO - codeparrot_training - Step 18871: {'lr': 0.0003624612899548046, 'samples': 9662464, 'steps': 18871, 'loss/train': 2.1134397983551025} 02/25/2022 02:00:45 - INFO - codeparrot_training - Step 18872: {'lr': 0.0003624466762919826, 'samples': 9662976, 'steps': 18872, 'loss/train': 1.6623711585998535} 02/25/2022 02:00:48 - INFO - codeparrot_training - Step 18873: {'lr': 0.00036243206214747466, 'samples': 9663488, 'steps': 18873, 'loss/train': 3.236616373062134} 02/25/2022 02:00:54 - INFO - codeparrot_training - Step 18874: {'lr': 0.00036241744752134347, 'samples': 9664000, 'steps': 18874, 'loss/train': 0.1615999937057495} 02/25/2022 02:00:57 - INFO - codeparrot_training - Step 18875: {'lr': 0.0003624028324136517, 'samples': 9664512, 'steps': 18875, 'loss/train': 2.288236379623413} 02/25/2022 02:01:03 - INFO - codeparrot_training - Step 18876: {'lr': 0.00036238821682446176, 'samples': 9665024, 'steps': 18876, 'loss/train': 1.4840154647827148} 02/25/2022 02:01:06 - INFO - codeparrot_training - Step 18877: {'lr': 0.0003623736007538365, 'samples': 9665536, 'steps': 18877, 'loss/train': 2.1744518280029297} 02/25/2022 02:01:11 - INFO - codeparrot_training - Step 18878: {'lr': 0.00036235898420183837, 'samples': 9666048, 'steps': 18878, 'loss/train': 2.1867895126342773} 02/25/2022 02:01:18 - INFO - codeparrot_training - Step 18879: {'lr': 0.0003623443671685301, 'samples': 9666560, 'steps': 18879, 'loss/train': 1.3891382217407227} 02/25/2022 02:01:21 - INFO - codeparrot_training - Step 18880: {'lr': 0.00036232974965397414, 'samples': 9667072, 'steps': 18880, 'loss/train': 1.3203290700912476} 02/25/2022 02:01:25 - INFO - codeparrot_training - Step 18881: {'lr': 0.00036231513165823324, 'samples': 9667584, 'steps': 18881, 'loss/train': 2.8153724670410156} 02/25/2022 02:01:30 - INFO - codeparrot_training - Step 18882: {'lr': 0.00036230051318137, 'samples': 9668096, 'steps': 18882, 'loss/train': 1.3842017650604248} 02/25/2022 02:01:34 - INFO - codeparrot_training - Step 18883: {'lr': 0.00036228589422344703, 'samples': 9668608, 'steps': 18883, 'loss/train': 0.4513271749019623} 02/25/2022 02:01:39 - INFO - codeparrot_training - Step 18884: {'lr': 0.0003622712747845269, 'samples': 9669120, 'steps': 18884, 'loss/train': 1.695831298828125} 02/25/2022 02:01:43 - INFO - codeparrot_training - Step 18885: {'lr': 0.0003622566548646723, 'samples': 9669632, 'steps': 18885, 'loss/train': 1.5047125816345215} 02/25/2022 02:01:48 - INFO - codeparrot_training - Step 18886: {'lr': 0.00036224203446394584, 'samples': 9670144, 'steps': 18886, 'loss/train': 3.071146011352539} 02/25/2022 02:01:54 - INFO - codeparrot_training - Step 18887: {'lr': 0.00036222741358241014, 'samples': 9670656, 'steps': 18887, 'loss/train': 2.0389113426208496} 02/25/2022 02:01:57 - INFO - codeparrot_training - Step 18888: {'lr': 0.0003622127922201278, 'samples': 9671168, 'steps': 18888, 'loss/train': 1.9491676092147827} 02/25/2022 02:02:01 - INFO - codeparrot_training - Step 18889: {'lr': 0.0003621981703771616, 'samples': 9671680, 'steps': 18889, 'loss/train': 2.0626060962677} 02/25/2022 02:02:07 - INFO - codeparrot_training - Step 18890: {'lr': 0.0003621835480535739, 'samples': 9672192, 'steps': 18890, 'loss/train': 1.7543541193008423} 02/25/2022 02:02:13 - INFO - codeparrot_training - Step 18891: {'lr': 0.00036216892524942764, 'samples': 9672704, 'steps': 18891, 'loss/train': 1.1323871612548828} 02/25/2022 02:02:16 - INFO - codeparrot_training - Step 18892: {'lr': 0.00036215430196478526, 'samples': 9673216, 'steps': 18892, 'loss/train': 1.2245393991470337} 02/25/2022 02:02:20 - INFO - codeparrot_training - Step 18893: {'lr': 0.0003621396781997095, 'samples': 9673728, 'steps': 18893, 'loss/train': 2.807162046432495} 02/25/2022 02:02:25 - INFO - codeparrot_training - Step 18894: {'lr': 0.00036212505395426297, 'samples': 9674240, 'steps': 18894, 'loss/train': 2.5276408195495605} 02/25/2022 02:02:29 - INFO - codeparrot_training - Step 18895: {'lr': 0.0003621104292285082, 'samples': 9674752, 'steps': 18895, 'loss/train': 2.152820587158203} 02/25/2022 02:02:34 - INFO - codeparrot_training - Step 18896: {'lr': 0.0003620958040225081, 'samples': 9675264, 'steps': 18896, 'loss/train': 2.154420852661133} 02/25/2022 02:02:38 - INFO - codeparrot_training - Step 18897: {'lr': 0.0003620811783363251, 'samples': 9675776, 'steps': 18897, 'loss/train': 0.9838770627975464} 02/25/2022 02:02:43 - INFO - codeparrot_training - Step 18898: {'lr': 0.0003620665521700219, 'samples': 9676288, 'steps': 18898, 'loss/train': 1.6690455675125122} 02/25/2022 02:02:47 - INFO - codeparrot_training - Step 18899: {'lr': 0.00036205192552366124, 'samples': 9676800, 'steps': 18899, 'loss/train': 2.0534627437591553} 02/25/2022 02:02:53 - INFO - codeparrot_training - Step 18900: {'lr': 0.0003620372983973057, 'samples': 9677312, 'steps': 18900, 'loss/train': 1.8415592908859253} 02/25/2022 02:02:56 - INFO - codeparrot_training - Step 18901: {'lr': 0.00036202267079101793, 'samples': 9677824, 'steps': 18901, 'loss/train': 2.4564619064331055} 02/25/2022 02:03:02 - INFO - codeparrot_training - Step 18902: {'lr': 0.0003620080427048605, 'samples': 9678336, 'steps': 18902, 'loss/train': 2.430135726928711} 02/25/2022 02:03:05 - INFO - codeparrot_training - Step 18903: {'lr': 0.00036199341413889637, 'samples': 9678848, 'steps': 18903, 'loss/train': 0.8723113536834717} 02/25/2022 02:03:11 - INFO - codeparrot_training - Step 18904: {'lr': 0.00036197878509318794, 'samples': 9679360, 'steps': 18904, 'loss/train': 2.141354560852051} 02/25/2022 02:03:14 - INFO - codeparrot_training - Step 18905: {'lr': 0.00036196415556779795, 'samples': 9679872, 'steps': 18905, 'loss/train': 1.7737194299697876} 02/25/2022 02:03:20 - INFO - codeparrot_training - Step 18906: {'lr': 0.00036194952556278906, 'samples': 9680384, 'steps': 18906, 'loss/train': 2.0249686241149902} 02/25/2022 02:03:23 - INFO - codeparrot_training - Step 18907: {'lr': 0.00036193489507822395, 'samples': 9680896, 'steps': 18907, 'loss/train': 2.179361343383789} 02/25/2022 02:03:29 - INFO - codeparrot_training - Step 18908: {'lr': 0.0003619202641141652, 'samples': 9681408, 'steps': 18908, 'loss/train': 1.6962487697601318} 02/25/2022 02:03:32 - INFO - codeparrot_training - Step 18909: {'lr': 0.00036190563267067576, 'samples': 9681920, 'steps': 18909, 'loss/train': 3.1392674446105957} 02/25/2022 02:03:39 - INFO - codeparrot_training - Step 18910: {'lr': 0.000361891000747818, 'samples': 9682432, 'steps': 18910, 'loss/train': 2.121316432952881} 02/25/2022 02:03:42 - INFO - codeparrot_training - Step 18911: {'lr': 0.0003618763683456548, 'samples': 9682944, 'steps': 18911, 'loss/train': 1.4225636720657349} 02/25/2022 02:03:48 - INFO - codeparrot_training - Step 18912: {'lr': 0.00036186173546424874, 'samples': 9683456, 'steps': 18912, 'loss/train': 0.2675212323665619} 02/25/2022 02:03:51 - INFO - codeparrot_training - Step 18913: {'lr': 0.00036184710210366243, 'samples': 9683968, 'steps': 18913, 'loss/train': 2.4865174293518066} 02/25/2022 02:03:56 - INFO - codeparrot_training - Step 18914: {'lr': 0.0003618324682639588, 'samples': 9684480, 'steps': 18914, 'loss/train': 1.8758796453475952} 02/25/2022 02:04:00 - INFO - codeparrot_training - Step 18915: {'lr': 0.0003618178339452003, 'samples': 9684992, 'steps': 18915, 'loss/train': 1.406438946723938} 02/25/2022 02:04:05 - INFO - codeparrot_training - Step 18916: {'lr': 0.00036180319914744973, 'samples': 9685504, 'steps': 18916, 'loss/train': 2.1090404987335205} 02/25/2022 02:04:09 - INFO - codeparrot_training - Step 18917: {'lr': 0.00036178856387076967, 'samples': 9686016, 'steps': 18917, 'loss/train': 2.6613447666168213} 02/25/2022 02:04:15 - INFO - codeparrot_training - Step 18918: {'lr': 0.00036177392811522304, 'samples': 9686528, 'steps': 18918, 'loss/train': 3.106773614883423} 02/25/2022 02:04:18 - INFO - codeparrot_training - Step 18919: {'lr': 0.0003617592918808723, 'samples': 9687040, 'steps': 18919, 'loss/train': 1.9209779500961304} 02/25/2022 02:04:24 - INFO - codeparrot_training - Step 18920: {'lr': 0.0003617446551677803, 'samples': 9687552, 'steps': 18920, 'loss/train': 1.841469645500183} 02/25/2022 02:04:27 - INFO - codeparrot_training - Step 18921: {'lr': 0.0003617300179760096, 'samples': 9688064, 'steps': 18921, 'loss/train': 1.5022215843200684} 02/25/2022 02:04:33 - INFO - codeparrot_training - Step 18922: {'lr': 0.0003617153803056231, 'samples': 9688576, 'steps': 18922, 'loss/train': 2.4341156482696533} 02/25/2022 02:04:36 - INFO - codeparrot_training - Step 18923: {'lr': 0.0003617007421566833, 'samples': 9689088, 'steps': 18923, 'loss/train': 2.744959831237793} 02/25/2022 02:04:42 - INFO - codeparrot_training - Step 18924: {'lr': 0.00036168610352925294, 'samples': 9689600, 'steps': 18924, 'loss/train': 1.898758053779602} 02/25/2022 02:04:45 - INFO - codeparrot_training - Step 18925: {'lr': 0.00036167146442339486, 'samples': 9690112, 'steps': 18925, 'loss/train': 0.7830764055252075} 02/25/2022 02:04:51 - INFO - codeparrot_training - Step 18926: {'lr': 0.00036165682483917165, 'samples': 9690624, 'steps': 18926, 'loss/train': 2.000889778137207} 02/25/2022 02:04:55 - INFO - codeparrot_training - Step 18927: {'lr': 0.0003616421847766461, 'samples': 9691136, 'steps': 18927, 'loss/train': 1.683374285697937} 02/25/2022 02:05:00 - INFO - codeparrot_training - Step 18928: {'lr': 0.00036162754423588086, 'samples': 9691648, 'steps': 18928, 'loss/train': 1.7564724683761597} 02/25/2022 02:05:04 - INFO - codeparrot_training - Step 18929: {'lr': 0.0003616129032169387, 'samples': 9692160, 'steps': 18929, 'loss/train': 1.1115750074386597} 02/25/2022 02:05:09 - INFO - codeparrot_training - Step 18930: {'lr': 0.00036159826171988223, 'samples': 9692672, 'steps': 18930, 'loss/train': 2.705695152282715} 02/25/2022 02:05:13 - INFO - codeparrot_training - Step 18931: {'lr': 0.00036158361974477434, 'samples': 9693184, 'steps': 18931, 'loss/train': 1.909696102142334} 02/25/2022 02:05:18 - INFO - codeparrot_training - Step 18932: {'lr': 0.0003615689772916776, 'samples': 9693696, 'steps': 18932, 'loss/train': 1.8387887477874756} 02/25/2022 02:05:22 - INFO - codeparrot_training - Step 18933: {'lr': 0.00036155433436065477, 'samples': 9694208, 'steps': 18933, 'loss/train': 2.6957552433013916} 02/25/2022 02:05:27 - INFO - codeparrot_training - Step 18934: {'lr': 0.0003615396909517686, 'samples': 9694720, 'steps': 18934, 'loss/train': 2.0331976413726807} 02/25/2022 02:05:31 - INFO - codeparrot_training - Step 18935: {'lr': 0.00036152504706508183, 'samples': 9695232, 'steps': 18935, 'loss/train': 1.6742535829544067} 02/25/2022 02:05:36 - INFO - codeparrot_training - Step 18936: {'lr': 0.00036151040270065716, 'samples': 9695744, 'steps': 18936, 'loss/train': 1.4416239261627197} 02/25/2022 02:05:40 - INFO - codeparrot_training - Step 18937: {'lr': 0.00036149575785855736, 'samples': 9696256, 'steps': 18937, 'loss/train': 1.6289310455322266} 02/25/2022 02:05:46 - INFO - codeparrot_training - Step 18938: {'lr': 0.00036148111253884514, 'samples': 9696768, 'steps': 18938, 'loss/train': 2.012981653213501} 02/25/2022 02:05:49 - INFO - codeparrot_training - Step 18939: {'lr': 0.00036146646674158317, 'samples': 9697280, 'steps': 18939, 'loss/train': 1.5505435466766357} 02/25/2022 02:05:55 - INFO - codeparrot_training - Step 18940: {'lr': 0.00036145182046683433, 'samples': 9697792, 'steps': 18940, 'loss/train': 3.1853725910186768} 02/25/2022 02:05:58 - INFO - codeparrot_training - Step 18941: {'lr': 0.0003614371737146612, 'samples': 9698304, 'steps': 18941, 'loss/train': 2.1582465171813965} 02/25/2022 02:06:04 - INFO - codeparrot_training - Step 18942: {'lr': 0.0003614225264851266, 'samples': 9698816, 'steps': 18942, 'loss/train': 1.6931086778640747} 02/25/2022 02:06:08 - INFO - codeparrot_training - Step 18943: {'lr': 0.00036140787877829336, 'samples': 9699328, 'steps': 18943, 'loss/train': 2.1213903427124023} 02/25/2022 02:06:13 - INFO - codeparrot_training - Step 18944: {'lr': 0.0003613932305942241, 'samples': 9699840, 'steps': 18944, 'loss/train': 2.4540419578552246} 02/25/2022 02:06:17 - INFO - codeparrot_training - Step 18945: {'lr': 0.0003613785819329816, 'samples': 9700352, 'steps': 18945, 'loss/train': 1.9778475761413574} 02/25/2022 02:06:22 - INFO - codeparrot_training - Step 18946: {'lr': 0.00036136393279462866, 'samples': 9700864, 'steps': 18946, 'loss/train': 1.6620349884033203} 02/25/2022 02:06:26 - INFO - codeparrot_training - Step 18947: {'lr': 0.00036134928317922796, 'samples': 9701376, 'steps': 18947, 'loss/train': 3.143666982650757} 02/25/2022 02:06:32 - INFO - codeparrot_training - Step 18948: {'lr': 0.0003613346330868423, 'samples': 9701888, 'steps': 18948, 'loss/train': 1.8337793350219727} 02/25/2022 02:06:35 - INFO - codeparrot_training - Step 18949: {'lr': 0.0003613199825175344, 'samples': 9702400, 'steps': 18949, 'loss/train': 2.0417895317077637} 02/25/2022 02:06:41 - INFO - codeparrot_training - Step 18950: {'lr': 0.0003613053314713671, 'samples': 9702912, 'steps': 18950, 'loss/train': 2.463352918624878} 02/25/2022 02:06:44 - INFO - codeparrot_training - Step 18951: {'lr': 0.00036129067994840303, 'samples': 9703424, 'steps': 18951, 'loss/train': 1.6299381256103516} 02/25/2022 02:06:50 - INFO - codeparrot_training - Step 18952: {'lr': 0.0003612760279487051, 'samples': 9703936, 'steps': 18952, 'loss/train': 1.932807445526123} 02/25/2022 02:06:54 - INFO - codeparrot_training - Step 18953: {'lr': 0.00036126137547233593, 'samples': 9704448, 'steps': 18953, 'loss/train': 3.009141683578491} 02/25/2022 02:06:59 - INFO - codeparrot_training - Step 18954: {'lr': 0.00036124672251935843, 'samples': 9704960, 'steps': 18954, 'loss/train': 1.8990607261657715} 02/25/2022 02:07:03 - INFO - codeparrot_training - Step 18955: {'lr': 0.00036123206908983527, 'samples': 9705472, 'steps': 18955, 'loss/train': 0.758811354637146} 02/25/2022 02:07:09 - INFO - codeparrot_training - Step 18956: {'lr': 0.00036121741518382915, 'samples': 9705984, 'steps': 18956, 'loss/train': 2.3784446716308594} 02/25/2022 02:07:12 - INFO - codeparrot_training - Step 18957: {'lr': 0.0003612027608014031, 'samples': 9706496, 'steps': 18957, 'loss/train': 0.8154066205024719} 02/25/2022 02:07:18 - INFO - codeparrot_training - Step 18958: {'lr': 0.0003611881059426196, 'samples': 9707008, 'steps': 18958, 'loss/train': 1.6953154802322388} 02/25/2022 02:07:22 - INFO - codeparrot_training - Step 18959: {'lr': 0.00036117345060754163, 'samples': 9707520, 'steps': 18959, 'loss/train': 0.1523435413837433} 02/25/2022 02:07:28 - INFO - codeparrot_training - Step 18960: {'lr': 0.00036115879479623185, 'samples': 9708032, 'steps': 18960, 'loss/train': 0.30938780307769775} 02/25/2022 02:07:31 - INFO - codeparrot_training - Step 18961: {'lr': 0.00036114413850875317, 'samples': 9708544, 'steps': 18961, 'loss/train': 0.3409813344478607} 02/25/2022 02:07:37 - INFO - codeparrot_training - Step 18962: {'lr': 0.00036112948174516824, 'samples': 9709056, 'steps': 18962, 'loss/train': 2.7156331539154053} 02/25/2022 02:07:40 - INFO - codeparrot_training - Step 18963: {'lr': 0.0003611148245055399, 'samples': 9709568, 'steps': 18963, 'loss/train': 1.6761538982391357} 02/25/2022 02:07:45 - INFO - codeparrot_training - Step 18964: {'lr': 0.000361100166789931, 'samples': 9710080, 'steps': 18964, 'loss/train': 2.658154010772705} 02/25/2022 02:07:49 - INFO - codeparrot_training - Step 18965: {'lr': 0.0003610855085984041, 'samples': 9710592, 'steps': 18965, 'loss/train': 2.2412421703338623} 02/25/2022 02:07:54 - INFO - codeparrot_training - Step 18966: {'lr': 0.0003610708499310223, 'samples': 9711104, 'steps': 18966, 'loss/train': 0.1282891035079956} 02/25/2022 02:07:58 - INFO - codeparrot_training - Step 18967: {'lr': 0.0003610561907878482, 'samples': 9711616, 'steps': 18967, 'loss/train': 1.6763033866882324} 02/25/2022 02:08:03 - INFO - codeparrot_training - Step 18968: {'lr': 0.00036104153116894465, 'samples': 9712128, 'steps': 18968, 'loss/train': 3.7002174854278564} 02/25/2022 02:08:07 - INFO - codeparrot_training - Step 18969: {'lr': 0.00036102687107437444, 'samples': 9712640, 'steps': 18969, 'loss/train': 1.2943414449691772} 02/25/2022 02:08:12 - INFO - codeparrot_training - Step 18970: {'lr': 0.0003610122105042004, 'samples': 9713152, 'steps': 18970, 'loss/train': 2.2200047969818115} 02/25/2022 02:08:16 - INFO - codeparrot_training - Step 18971: {'lr': 0.0003609975494584852, 'samples': 9713664, 'steps': 18971, 'loss/train': 1.8420921564102173} 02/25/2022 02:08:23 - INFO - codeparrot_training - Step 18972: {'lr': 0.0003609828879372918, 'samples': 9714176, 'steps': 18972, 'loss/train': 2.4288551807403564} 02/25/2022 02:08:26 - INFO - codeparrot_training - Step 18973: {'lr': 0.000360968225940683, 'samples': 9714688, 'steps': 18973, 'loss/train': 3.2310967445373535} 02/25/2022 02:08:32 - INFO - codeparrot_training - Step 18974: {'lr': 0.00036095356346872156, 'samples': 9715200, 'steps': 18974, 'loss/train': 1.9115612506866455} 02/25/2022 02:08:35 - INFO - codeparrot_training - Step 18975: {'lr': 0.0003609389005214702, 'samples': 9715712, 'steps': 18975, 'loss/train': 3.7304508686065674} 02/25/2022 02:08:41 - INFO - codeparrot_training - Step 18976: {'lr': 0.0003609242370989919, 'samples': 9716224, 'steps': 18976, 'loss/train': 2.1039819717407227} 02/25/2022 02:08:44 - INFO - codeparrot_training - Step 18977: {'lr': 0.0003609095732013493, 'samples': 9716736, 'steps': 18977, 'loss/train': 2.0949344635009766} 02/25/2022 02:08:50 - INFO - codeparrot_training - Step 18978: {'lr': 0.0003608949088286054, 'samples': 9717248, 'steps': 18978, 'loss/train': 1.1355477571487427} 02/25/2022 02:08:53 - INFO - codeparrot_training - Step 18979: {'lr': 0.00036088024398082285, 'samples': 9717760, 'steps': 18979, 'loss/train': 2.464592218399048} 02/25/2022 02:08:59 - INFO - codeparrot_training - Step 18980: {'lr': 0.00036086557865806464, 'samples': 9718272, 'steps': 18980, 'loss/train': 2.1163060665130615} 02/25/2022 02:09:02 - INFO - codeparrot_training - Step 18981: {'lr': 0.0003608509128603934, 'samples': 9718784, 'steps': 18981, 'loss/train': 2.080054521560669} 02/25/2022 02:09:09 - INFO - codeparrot_training - Step 18982: {'lr': 0.0003608362465878721, 'samples': 9719296, 'steps': 18982, 'loss/train': 2.635371446609497} 02/25/2022 02:09:12 - INFO - codeparrot_training - Step 18983: {'lr': 0.0003608215798405635, 'samples': 9719808, 'steps': 18983, 'loss/train': 1.5709997415542603} 02/25/2022 02:09:18 - INFO - codeparrot_training - Step 18984: {'lr': 0.0003608069126185304, 'samples': 9720320, 'steps': 18984, 'loss/train': 2.0796685218811035} 02/25/2022 02:09:21 - INFO - codeparrot_training - Step 18985: {'lr': 0.0003607922449218357, 'samples': 9720832, 'steps': 18985, 'loss/train': 1.9137529134750366} 02/25/2022 02:09:27 - INFO - codeparrot_training - Step 18986: {'lr': 0.00036077757675054224, 'samples': 9721344, 'steps': 18986, 'loss/train': 1.7957603931427002} 02/25/2022 02:09:30 - INFO - codeparrot_training - Step 18987: {'lr': 0.0003607629081047128, 'samples': 9721856, 'steps': 18987, 'loss/train': 1.5062298774719238} 02/25/2022 02:09:36 - INFO - codeparrot_training - Step 18988: {'lr': 0.0003607482389844102, 'samples': 9722368, 'steps': 18988, 'loss/train': 1.6406534910202026} 02/25/2022 02:09:39 - INFO - codeparrot_training - Step 18989: {'lr': 0.0003607335693896973, 'samples': 9722880, 'steps': 18989, 'loss/train': 1.2689883708953857} 02/25/2022 02:09:45 - INFO - codeparrot_training - Step 18990: {'lr': 0.00036071889932063706, 'samples': 9723392, 'steps': 18990, 'loss/train': 2.2161219120025635} 02/25/2022 02:09:49 - INFO - codeparrot_training - Step 18991: {'lr': 0.0003607042287772921, 'samples': 9723904, 'steps': 18991, 'loss/train': 1.3956328630447388} 02/25/2022 02:09:54 - INFO - codeparrot_training - Step 18992: {'lr': 0.0003606895577597254, 'samples': 9724416, 'steps': 18992, 'loss/train': 2.0705862045288086} 02/25/2022 02:09:58 - INFO - codeparrot_training - Step 18993: {'lr': 0.0003606748862679998, 'samples': 9724928, 'steps': 18993, 'loss/train': 1.8988370895385742} 02/25/2022 02:10:04 - INFO - codeparrot_training - Step 18994: {'lr': 0.00036066021430217817, 'samples': 9725440, 'steps': 18994, 'loss/train': 2.5313243865966797} 02/25/2022 02:10:07 - INFO - codeparrot_training - Step 18995: {'lr': 0.00036064554186232323, 'samples': 9725952, 'steps': 18995, 'loss/train': 2.888190269470215} 02/25/2022 02:10:13 - INFO - codeparrot_training - Step 18996: {'lr': 0.0003606308689484979, 'samples': 9726464, 'steps': 18996, 'loss/train': 2.065453290939331} 02/25/2022 02:10:16 - INFO - codeparrot_training - Step 18997: {'lr': 0.0003606161955607652, 'samples': 9726976, 'steps': 18997, 'loss/train': 1.7177544832229614} 02/25/2022 02:10:22 - INFO - codeparrot_training - Step 18998: {'lr': 0.0003606015216991877, 'samples': 9727488, 'steps': 18998, 'loss/train': 2.8631699085235596} 02/25/2022 02:10:25 - INFO - codeparrot_training - Step 18999: {'lr': 0.0003605868473638285, 'samples': 9728000, 'steps': 18999, 'loss/train': 1.1513526439666748} 02/25/2022 02:10:25 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 02:10:43 - WARNING - huggingface_hub.repository - Several commits (19) will be pushed upstream. 02/25/2022 02:10:43 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 02:11:16 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 4d2b52a..e9773d2 floral-grass-11 -> floral-grass-11 02/25/2022 02:11:22 - INFO - codeparrot_training - Step 19000: {'lr': 0.00036057217255475036, 'samples': 9728512, 'steps': 19000, 'loss/train': 1.9880913496017456} 02/25/2022 02:11:26 - INFO - codeparrot_training - Step 19001: {'lr': 0.0003605574972720161, 'samples': 9729024, 'steps': 19001, 'loss/train': 2.0283875465393066} 02/25/2022 02:11:31 - INFO - codeparrot_training - Step 19002: {'lr': 0.00036054282151568864, 'samples': 9729536, 'steps': 19002, 'loss/train': 1.7518035173416138} 02/25/2022 02:11:35 - INFO - codeparrot_training - Step 19003: {'lr': 0.0003605281452858308, 'samples': 9730048, 'steps': 19003, 'loss/train': 2.9040071964263916} 02/25/2022 02:11:42 - INFO - codeparrot_training - Step 19004: {'lr': 0.00036051346858250556, 'samples': 9730560, 'steps': 19004, 'loss/train': 1.6045945882797241} 02/25/2022 02:11:45 - INFO - codeparrot_training - Step 19005: {'lr': 0.00036049879140577566, 'samples': 9731072, 'steps': 19005, 'loss/train': 1.8744827508926392} 02/25/2022 02:11:51 - INFO - codeparrot_training - Step 19006: {'lr': 0.000360484113755704, 'samples': 9731584, 'steps': 19006, 'loss/train': 0.5272473692893982} 02/25/2022 02:11:54 - INFO - codeparrot_training - Step 19007: {'lr': 0.0003604694356323536, 'samples': 9732096, 'steps': 19007, 'loss/train': 2.0560686588287354} 02/25/2022 02:12:00 - INFO - codeparrot_training - Step 19008: {'lr': 0.00036045475703578705, 'samples': 9732608, 'steps': 19008, 'loss/train': 1.4259134531021118} 02/25/2022 02:12:03 - INFO - codeparrot_training - Step 19009: {'lr': 0.0003604400779660674, 'samples': 9733120, 'steps': 19009, 'loss/train': 1.5217783451080322} 02/25/2022 02:12:09 - INFO - codeparrot_training - Step 19010: {'lr': 0.00036042539842325765, 'samples': 9733632, 'steps': 19010, 'loss/train': 1.9425674676895142} 02/25/2022 02:12:12 - INFO - codeparrot_training - Step 19011: {'lr': 0.0003604107184074205, 'samples': 9734144, 'steps': 19011, 'loss/train': 2.312227249145508} 02/25/2022 02:12:18 - INFO - codeparrot_training - Step 19012: {'lr': 0.0003603960379186189, 'samples': 9734656, 'steps': 19012, 'loss/train': 2.4640073776245117} 02/25/2022 02:12:21 - INFO - codeparrot_training - Step 19013: {'lr': 0.0003603813569569157, 'samples': 9735168, 'steps': 19013, 'loss/train': 1.3639686107635498} 02/25/2022 02:12:27 - INFO - codeparrot_training - Step 19014: {'lr': 0.0003603666755223739, 'samples': 9735680, 'steps': 19014, 'loss/train': 3.058096408843994} 02/25/2022 02:12:30 - INFO - codeparrot_training - Step 19015: {'lr': 0.0003603519936150562, 'samples': 9736192, 'steps': 19015, 'loss/train': 1.3021838665008545} 02/25/2022 02:12:36 - INFO - codeparrot_training - Step 19016: {'lr': 0.00036033731123502567, 'samples': 9736704, 'steps': 19016, 'loss/train': 2.157147169113159} 02/25/2022 02:12:39 - INFO - codeparrot_training - Step 19017: {'lr': 0.00036032262838234507, 'samples': 9737216, 'steps': 19017, 'loss/train': 2.8765664100646973} 02/25/2022 02:12:46 - INFO - codeparrot_training - Step 19018: {'lr': 0.0003603079450570775, 'samples': 9737728, 'steps': 19018, 'loss/train': 2.0201380252838135} 02/25/2022 02:12:50 - INFO - codeparrot_training - Step 19019: {'lr': 0.00036029326125928556, 'samples': 9738240, 'steps': 19019, 'loss/train': 2.51269268989563} 02/25/2022 02:12:55 - INFO - codeparrot_training - Step 19020: {'lr': 0.00036027857698903235, 'samples': 9738752, 'steps': 19020, 'loss/train': 1.681691288948059} 02/25/2022 02:12:59 - INFO - codeparrot_training - Step 19021: {'lr': 0.00036026389224638077, 'samples': 9739264, 'steps': 19021, 'loss/train': 3.8848133087158203} 02/25/2022 02:13:04 - INFO - codeparrot_training - Step 19022: {'lr': 0.00036024920703139375, 'samples': 9739776, 'steps': 19022, 'loss/train': 3.7390666007995605} 02/25/2022 02:13:08 - INFO - codeparrot_training - Step 19023: {'lr': 0.00036023452134413406, 'samples': 9740288, 'steps': 19023, 'loss/train': 3.0671582221984863} 02/25/2022 02:13:13 - INFO - codeparrot_training - Step 19024: {'lr': 0.00036021983518466465, 'samples': 9740800, 'steps': 19024, 'loss/train': 1.5447514057159424} 02/25/2022 02:13:17 - INFO - codeparrot_training - Step 19025: {'lr': 0.00036020514855304855, 'samples': 9741312, 'steps': 19025, 'loss/train': 1.8689544200897217} 02/25/2022 02:13:22 - INFO - codeparrot_training - Step 19026: {'lr': 0.00036019046144934855, 'samples': 9741824, 'steps': 19026, 'loss/train': 1.758233666419983} 02/25/2022 02:13:25 - INFO - codeparrot_training - Step 19027: {'lr': 0.0003601757738736275, 'samples': 9742336, 'steps': 19027, 'loss/train': 2.0444414615631104} 02/25/2022 02:13:32 - INFO - codeparrot_training - Step 19028: {'lr': 0.0003601610858259485, 'samples': 9742848, 'steps': 19028, 'loss/train': 0.7698665261268616} 02/25/2022 02:13:35 - INFO - codeparrot_training - Step 19029: {'lr': 0.0003601463973063745, 'samples': 9743360, 'steps': 19029, 'loss/train': 1.7188829183578491} 02/25/2022 02:13:41 - INFO - codeparrot_training - Step 19030: {'lr': 0.0003601317083149682, 'samples': 9743872, 'steps': 19030, 'loss/train': 1.8872079849243164} 02/25/2022 02:13:44 - INFO - codeparrot_training - Step 19031: {'lr': 0.0003601170188517926, 'samples': 9744384, 'steps': 19031, 'loss/train': 2.456594944000244} 02/25/2022 02:13:50 - INFO - codeparrot_training - Step 19032: {'lr': 0.0003601023289169107, 'samples': 9744896, 'steps': 19032, 'loss/train': 2.471778631210327} 02/25/2022 02:13:53 - INFO - codeparrot_training - Step 19033: {'lr': 0.00036008763851038534, 'samples': 9745408, 'steps': 19033, 'loss/train': 1.528847575187683} 02/25/2022 02:13:59 - INFO - codeparrot_training - Step 19034: {'lr': 0.0003600729476322796, 'samples': 9745920, 'steps': 19034, 'loss/train': 0.5691848397254944} 02/25/2022 02:14:02 - INFO - codeparrot_training - Step 19035: {'lr': 0.0003600582562826562, 'samples': 9746432, 'steps': 19035, 'loss/train': 1.7924342155456543} 02/25/2022 02:14:08 - INFO - codeparrot_training - Step 19036: {'lr': 0.0003600435644615783, 'samples': 9746944, 'steps': 19036, 'loss/train': 1.8466625213623047} 02/25/2022 02:14:11 - INFO - codeparrot_training - Step 19037: {'lr': 0.0003600288721691085, 'samples': 9747456, 'steps': 19037, 'loss/train': 2.1249301433563232} 02/25/2022 02:14:17 - INFO - codeparrot_training - Step 19038: {'lr': 0.0003600141794053102, 'samples': 9747968, 'steps': 19038, 'loss/train': 0.725248396396637} 02/25/2022 02:14:20 - INFO - codeparrot_training - Step 19039: {'lr': 0.00035999948617024594, 'samples': 9748480, 'steps': 19039, 'loss/train': 0.7762969136238098} 02/25/2022 02:14:27 - INFO - codeparrot_training - Step 19040: {'lr': 0.0003599847924639788, 'samples': 9748992, 'steps': 19040, 'loss/train': 1.429334044456482} 02/25/2022 02:14:30 - INFO - codeparrot_training - Step 19041: {'lr': 0.0003599700982865718, 'samples': 9749504, 'steps': 19041, 'loss/train': 1.6418672800064087} 02/25/2022 02:14:36 - INFO - codeparrot_training - Step 19042: {'lr': 0.00035995540363808776, 'samples': 9750016, 'steps': 19042, 'loss/train': 2.384359359741211} 02/25/2022 02:14:39 - INFO - codeparrot_training - Step 19043: {'lr': 0.0003599407085185897, 'samples': 9750528, 'steps': 19043, 'loss/train': 1.5225509405136108} 02/25/2022 02:14:45 - INFO - codeparrot_training - Step 19044: {'lr': 0.00035992601292814065, 'samples': 9751040, 'steps': 19044, 'loss/train': 1.8368422985076904} 02/25/2022 02:14:48 - INFO - codeparrot_training - Step 19045: {'lr': 0.00035991131686680333, 'samples': 9751552, 'steps': 19045, 'loss/train': 1.9882451295852661} 02/25/2022 02:14:54 - INFO - codeparrot_training - Step 19046: {'lr': 0.00035989662033464087, 'samples': 9752064, 'steps': 19046, 'loss/train': 1.8748501539230347} 02/25/2022 02:14:57 - INFO - codeparrot_training - Step 19047: {'lr': 0.0003598819233317162, 'samples': 9752576, 'steps': 19047, 'loss/train': 0.7991703748703003} 02/25/2022 02:15:03 - INFO - codeparrot_training - Step 19048: {'lr': 0.0003598672258580923, 'samples': 9753088, 'steps': 19048, 'loss/train': 2.144665479660034} 02/25/2022 02:15:06 - INFO - codeparrot_training - Step 19049: {'lr': 0.00035985252791383203, 'samples': 9753600, 'steps': 19049, 'loss/train': 1.8054879903793335} 02/25/2022 02:15:12 - INFO - codeparrot_training - Step 19050: {'lr': 0.0003598378294989984, 'samples': 9754112, 'steps': 19050, 'loss/train': 1.294025182723999} 02/25/2022 02:15:16 - INFO - codeparrot_training - Step 19051: {'lr': 0.00035982313061365444, 'samples': 9754624, 'steps': 19051, 'loss/train': 2.0888357162475586} 02/25/2022 02:15:21 - INFO - codeparrot_training - Step 19052: {'lr': 0.00035980843125786306, 'samples': 9755136, 'steps': 19052, 'loss/train': 1.6288892030715942} 02/25/2022 02:15:25 - INFO - codeparrot_training - Step 19053: {'lr': 0.00035979373143168726, 'samples': 9755648, 'steps': 19053, 'loss/train': 2.3704967498779297} 02/25/2022 02:15:30 - INFO - codeparrot_training - Step 19054: {'lr': 0.0003597790311351898, 'samples': 9756160, 'steps': 19054, 'loss/train': 1.7540143728256226} 02/25/2022 02:15:34 - INFO - codeparrot_training - Step 19055: {'lr': 0.00035976433036843405, 'samples': 9756672, 'steps': 19055, 'loss/train': 1.3987394571304321} 02/25/2022 02:15:39 - INFO - codeparrot_training - Step 19056: {'lr': 0.00035974962913148267, 'samples': 9757184, 'steps': 19056, 'loss/train': 1.8322112560272217} 02/25/2022 02:15:45 - INFO - codeparrot_training - Step 19057: {'lr': 0.0003597349274243987, 'samples': 9757696, 'steps': 19057, 'loss/train': 1.9115732908248901} 02/25/2022 02:15:48 - INFO - codeparrot_training - Step 19058: {'lr': 0.0003597202252472452, 'samples': 9758208, 'steps': 19058, 'loss/train': 2.3142635822296143} 02/25/2022 02:15:54 - INFO - codeparrot_training - Step 19059: {'lr': 0.00035970552260008505, 'samples': 9758720, 'steps': 19059, 'loss/train': 1.919982671737671} 02/25/2022 02:15:57 - INFO - codeparrot_training - Step 19060: {'lr': 0.00035969081948298133, 'samples': 9759232, 'steps': 19060, 'loss/train': 2.7326436042785645} 02/25/2022 02:16:03 - INFO - codeparrot_training - Step 19061: {'lr': 0.0003596761158959969, 'samples': 9759744, 'steps': 19061, 'loss/train': 1.749724268913269} 02/25/2022 02:16:06 - INFO - codeparrot_training - Step 19062: {'lr': 0.00035966141183919485, 'samples': 9760256, 'steps': 19062, 'loss/train': 2.1644325256347656} 02/25/2022 02:16:12 - INFO - codeparrot_training - Step 19063: {'lr': 0.0003596467073126382, 'samples': 9760768, 'steps': 19063, 'loss/train': 2.2269372940063477} 02/25/2022 02:16:15 - INFO - codeparrot_training - Step 19064: {'lr': 0.0003596320023163898, 'samples': 9761280, 'steps': 19064, 'loss/train': 2.4966979026794434} 02/25/2022 02:16:22 - INFO - codeparrot_training - Step 19065: {'lr': 0.0003596172968505127, 'samples': 9761792, 'steps': 19065, 'loss/train': 1.328292965888977} 02/25/2022 02:16:25 - INFO - codeparrot_training - Step 19066: {'lr': 0.0003596025909150699, 'samples': 9762304, 'steps': 19066, 'loss/train': 1.5774143934249878} 02/25/2022 02:16:31 - INFO - codeparrot_training - Step 19067: {'lr': 0.00035958788451012446, 'samples': 9762816, 'steps': 19067, 'loss/train': 1.2293187379837036} 02/25/2022 02:16:34 - INFO - codeparrot_training - Step 19068: {'lr': 0.0003595731776357393, 'samples': 9763328, 'steps': 19068, 'loss/train': 0.6156061291694641} 02/25/2022 02:16:40 - INFO - codeparrot_training - Step 19069: {'lr': 0.00035955847029197746, 'samples': 9763840, 'steps': 19069, 'loss/train': 0.13419534265995026} 02/25/2022 02:16:43 - INFO - codeparrot_training - Step 19070: {'lr': 0.0003595437624789019, 'samples': 9764352, 'steps': 19070, 'loss/train': 1.035567283630371} 02/25/2022 02:16:49 - INFO - codeparrot_training - Step 19071: {'lr': 0.00035952905419657565, 'samples': 9764864, 'steps': 19071, 'loss/train': 1.6488547325134277} 02/25/2022 02:16:52 - INFO - codeparrot_training - Step 19072: {'lr': 0.0003595143454450617, 'samples': 9765376, 'steps': 19072, 'loss/train': 2.506180763244629} 02/25/2022 02:16:58 - INFO - codeparrot_training - Step 19073: {'lr': 0.0003594996362244231, 'samples': 9765888, 'steps': 19073, 'loss/train': 2.0034542083740234} 02/25/2022 02:17:01 - INFO - codeparrot_training - Step 19074: {'lr': 0.0003594849265347228, 'samples': 9766400, 'steps': 19074, 'loss/train': 1.6224825382232666} 02/25/2022 02:17:08 - INFO - codeparrot_training - Step 19075: {'lr': 0.00035947021637602384, 'samples': 9766912, 'steps': 19075, 'loss/train': 1.6428803205490112} 02/25/2022 02:17:11 - INFO - codeparrot_training - Step 19076: {'lr': 0.0003594555057483892, 'samples': 9767424, 'steps': 19076, 'loss/train': 1.0455342531204224} 02/25/2022 02:17:17 - INFO - codeparrot_training - Step 19077: {'lr': 0.00035944079465188194, 'samples': 9767936, 'steps': 19077, 'loss/train': 2.4600830078125} 02/25/2022 02:17:20 - INFO - codeparrot_training - Step 19078: {'lr': 0.0003594260830865651, 'samples': 9768448, 'steps': 19078, 'loss/train': 2.7476115226745605} 02/25/2022 02:17:26 - INFO - codeparrot_training - Step 19079: {'lr': 0.00035941137105250173, 'samples': 9768960, 'steps': 19079, 'loss/train': 2.278362989425659} 02/25/2022 02:17:29 - INFO - codeparrot_training - Step 19080: {'lr': 0.00035939665854975466, 'samples': 9769472, 'steps': 19080, 'loss/train': 1.9511642456054688} 02/25/2022 02:17:35 - INFO - codeparrot_training - Step 19081: {'lr': 0.0003593819455783871, 'samples': 9769984, 'steps': 19081, 'loss/train': 1.589357614517212} 02/25/2022 02:17:38 - INFO - codeparrot_training - Step 19082: {'lr': 0.000359367232138462, 'samples': 9770496, 'steps': 19082, 'loss/train': 1.426775336265564} 02/25/2022 02:17:44 - INFO - codeparrot_training - Step 19083: {'lr': 0.00035935251823004244, 'samples': 9771008, 'steps': 19083, 'loss/train': 1.2051615715026855} 02/25/2022 02:17:47 - INFO - codeparrot_training - Step 19084: {'lr': 0.00035933780385319136, 'samples': 9771520, 'steps': 19084, 'loss/train': 0.9950919151306152} 02/25/2022 02:17:54 - INFO - codeparrot_training - Step 19085: {'lr': 0.0003593230890079719, 'samples': 9772032, 'steps': 19085, 'loss/train': 1.2789618968963623} 02/25/2022 02:17:58 - INFO - codeparrot_training - Step 19086: {'lr': 0.0003593083736944471, 'samples': 9772544, 'steps': 19086, 'loss/train': 1.669321060180664} 02/25/2022 02:18:03 - INFO - codeparrot_training - Step 19087: {'lr': 0.00035929365791267974, 'samples': 9773056, 'steps': 19087, 'loss/train': 1.6883490085601807} 02/25/2022 02:18:07 - INFO - codeparrot_training - Step 19088: {'lr': 0.00035927894166273323, 'samples': 9773568, 'steps': 19088, 'loss/train': 0.8025223016738892} 02/25/2022 02:18:12 - INFO - codeparrot_training - Step 19089: {'lr': 0.00035926422494467035, 'samples': 9774080, 'steps': 19089, 'loss/train': 8.725171089172363} 02/25/2022 02:18:16 - INFO - codeparrot_training - Step 19090: {'lr': 0.0003592495077585543, 'samples': 9774592, 'steps': 19090, 'loss/train': 2.58119535446167} 02/25/2022 02:18:21 - INFO - codeparrot_training - Step 19091: {'lr': 0.000359234790104448, 'samples': 9775104, 'steps': 19091, 'loss/train': 2.629284381866455} 02/25/2022 02:18:25 - INFO - codeparrot_training - Step 19092: {'lr': 0.0003592200719824146, 'samples': 9775616, 'steps': 19092, 'loss/train': 2.485776424407959} 02/25/2022 02:18:30 - INFO - codeparrot_training - Step 19093: {'lr': 0.000359205353392517, 'samples': 9776128, 'steps': 19093, 'loss/train': 2.118177652359009} 02/25/2022 02:18:34 - INFO - codeparrot_training - Step 19094: {'lr': 0.00035919063433481835, 'samples': 9776640, 'steps': 19094, 'loss/train': 1.720985770225525} 02/25/2022 02:18:41 - INFO - codeparrot_training - Step 19095: {'lr': 0.0003591759148093818, 'samples': 9777152, 'steps': 19095, 'loss/train': 3.0862491130828857} 02/25/2022 02:18:44 - INFO - codeparrot_training - Step 19096: {'lr': 0.0003591611948162702, 'samples': 9777664, 'steps': 19096, 'loss/train': 1.9200283288955688} 02/25/2022 02:18:50 - INFO - codeparrot_training - Step 19097: {'lr': 0.0003591464743555467, 'samples': 9778176, 'steps': 19097, 'loss/train': 2.2101798057556152} 02/25/2022 02:18:53 - INFO - codeparrot_training - Step 19098: {'lr': 0.0003591317534272744, 'samples': 9778688, 'steps': 19098, 'loss/train': 1.1743124723434448} 02/25/2022 02:18:59 - INFO - codeparrot_training - Step 19099: {'lr': 0.0003591170320315164, 'samples': 9779200, 'steps': 19099, 'loss/train': 1.870015025138855} 02/25/2022 02:19:02 - INFO - codeparrot_training - Step 19100: {'lr': 0.0003591023101683355, 'samples': 9779712, 'steps': 19100, 'loss/train': 1.0984140634536743} 02/25/2022 02:19:08 - INFO - codeparrot_training - Step 19101: {'lr': 0.00035908758783779506, 'samples': 9780224, 'steps': 19101, 'loss/train': 5.222268104553223} 02/25/2022 02:19:11 - INFO - codeparrot_training - Step 19102: {'lr': 0.000359072865039958, 'samples': 9780736, 'steps': 19102, 'loss/train': 1.0986119508743286} 02/25/2022 02:19:17 - INFO - codeparrot_training - Step 19103: {'lr': 0.0003590581417748875, 'samples': 9781248, 'steps': 19103, 'loss/train': 1.209088921546936} 02/25/2022 02:19:20 - INFO - codeparrot_training - Step 19104: {'lr': 0.0003590434180426465, 'samples': 9781760, 'steps': 19104, 'loss/train': 0.7341287732124329} 02/25/2022 02:19:26 - INFO - codeparrot_training - Step 19105: {'lr': 0.00035902869384329803, 'samples': 9782272, 'steps': 19105, 'loss/train': 1.9990285634994507} 02/25/2022 02:19:29 - INFO - codeparrot_training - Step 19106: {'lr': 0.0003590139691769054, 'samples': 9782784, 'steps': 19106, 'loss/train': 2.1960625648498535} 02/25/2022 02:19:35 - INFO - codeparrot_training - Step 19107: {'lr': 0.00035899924404353146, 'samples': 9783296, 'steps': 19107, 'loss/train': 1.4845483303070068} 02/25/2022 02:19:38 - INFO - codeparrot_training - Step 19108: {'lr': 0.00035898451844323937, 'samples': 9783808, 'steps': 19108, 'loss/train': 1.3414267301559448} 02/25/2022 02:19:44 - INFO - codeparrot_training - Step 19109: {'lr': 0.0003589697923760923, 'samples': 9784320, 'steps': 19109, 'loss/train': 1.7662800550460815} 02/25/2022 02:19:47 - INFO - codeparrot_training - Step 19110: {'lr': 0.00035895506584215307, 'samples': 9784832, 'steps': 19110, 'loss/train': 1.7849555015563965} 02/25/2022 02:19:53 - INFO - codeparrot_training - Step 19111: {'lr': 0.0003589403388414851, 'samples': 9785344, 'steps': 19111, 'loss/train': 0.9868466258049011} 02/25/2022 02:19:57 - INFO - codeparrot_training - Step 19112: {'lr': 0.0003589256113741513, 'samples': 9785856, 'steps': 19112, 'loss/train': 1.3435007333755493} 02/25/2022 02:20:02 - INFO - codeparrot_training - Step 19113: {'lr': 0.00035891088344021464, 'samples': 9786368, 'steps': 19113, 'loss/train': 0.20858079195022583} 02/25/2022 02:20:06 - INFO - codeparrot_training - Step 19114: {'lr': 0.00035889615503973847, 'samples': 9786880, 'steps': 19114, 'loss/train': 4.169969081878662} 02/25/2022 02:20:11 - INFO - codeparrot_training - Step 19115: {'lr': 0.00035888142617278567, 'samples': 9787392, 'steps': 19115, 'loss/train': 1.5138440132141113} 02/25/2022 02:20:15 - INFO - codeparrot_training - Step 19116: {'lr': 0.00035886669683941945, 'samples': 9787904, 'steps': 19116, 'loss/train': 1.705866813659668} 02/25/2022 02:20:20 - INFO - codeparrot_training - Step 19117: {'lr': 0.00035885196703970286, 'samples': 9788416, 'steps': 19117, 'loss/train': 2.3572585582733154} 02/25/2022 02:20:24 - INFO - codeparrot_training - Step 19118: {'lr': 0.00035883723677369893, 'samples': 9788928, 'steps': 19118, 'loss/train': 1.2012253999710083} 02/25/2022 02:20:29 - INFO - codeparrot_training - Step 19119: {'lr': 0.000358822506041471, 'samples': 9789440, 'steps': 19119, 'loss/train': 0.9769202470779419} 02/25/2022 02:20:33 - INFO - codeparrot_training - Step 19120: {'lr': 0.0003588077748430819, 'samples': 9789952, 'steps': 19120, 'loss/train': 1.7477604150772095} 02/25/2022 02:20:39 - INFO - codeparrot_training - Step 19121: {'lr': 0.00035879304317859486, 'samples': 9790464, 'steps': 19121, 'loss/train': 2.4582486152648926} 02/25/2022 02:20:42 - INFO - codeparrot_training - Step 19122: {'lr': 0.00035877831104807296, 'samples': 9790976, 'steps': 19122, 'loss/train': 2.092266321182251} 02/25/2022 02:20:50 - INFO - codeparrot_training - Step 19123: {'lr': 0.00035876357845157934, 'samples': 9791488, 'steps': 19123, 'loss/train': 2.037261486053467} 02/25/2022 02:20:53 - INFO - codeparrot_training - Step 19124: {'lr': 0.00035874884538917705, 'samples': 9792000, 'steps': 19124, 'loss/train': 1.3121665716171265} 02/25/2022 02:20:59 - INFO - codeparrot_training - Step 19125: {'lr': 0.0003587341118609293, 'samples': 9792512, 'steps': 19125, 'loss/train': 1.3642024993896484} 02/25/2022 02:21:02 - INFO - codeparrot_training - Step 19126: {'lr': 0.00035871937786689914, 'samples': 9793024, 'steps': 19126, 'loss/train': 2.366123676300049} 02/25/2022 02:21:06 - INFO - codeparrot_training - Step 19127: {'lr': 0.00035870464340714966, 'samples': 9793536, 'steps': 19127, 'loss/train': 1.6785504817962646} 02/25/2022 02:21:11 - INFO - codeparrot_training - Step 19128: {'lr': 0.00035868990848174397, 'samples': 9794048, 'steps': 19128, 'loss/train': 2.1995596885681152} 02/25/2022 02:21:17 - INFO - codeparrot_training - Step 19129: {'lr': 0.00035867517309074527, 'samples': 9794560, 'steps': 19129, 'loss/train': 1.1732362508773804} 02/25/2022 02:21:21 - INFO - codeparrot_training - Step 19130: {'lr': 0.0003586604372342166, 'samples': 9795072, 'steps': 19130, 'loss/train': 2.041181802749634} 02/25/2022 02:21:24 - INFO - codeparrot_training - Step 19131: {'lr': 0.0003586457009122211, 'samples': 9795584, 'steps': 19131, 'loss/train': 4.172245979309082} 02/25/2022 02:21:30 - INFO - codeparrot_training - Step 19132: {'lr': 0.000358630964124822, 'samples': 9796096, 'steps': 19132, 'loss/train': 2.8790414333343506} 02/25/2022 02:21:37 - INFO - codeparrot_training - Step 19133: {'lr': 0.0003586162268720823, 'samples': 9796608, 'steps': 19133, 'loss/train': 2.4208638668060303} 02/25/2022 02:21:41 - INFO - codeparrot_training - Step 19134: {'lr': 0.00035860148915406513, 'samples': 9797120, 'steps': 19134, 'loss/train': 1.800406575202942} 02/25/2022 02:21:46 - INFO - codeparrot_training - Step 19135: {'lr': 0.00035858675097083373, 'samples': 9797632, 'steps': 19135, 'loss/train': 1.9086666107177734} 02/25/2022 02:21:50 - INFO - codeparrot_training - Step 19136: {'lr': 0.0003585720123224512, 'samples': 9798144, 'steps': 19136, 'loss/train': 0.8670181632041931} 02/25/2022 02:21:53 - INFO - codeparrot_training - Step 19137: {'lr': 0.0003585572732089806, 'samples': 9798656, 'steps': 19137, 'loss/train': 2.753687858581543} 02/25/2022 02:21:59 - INFO - codeparrot_training - Step 19138: {'lr': 0.00035854253363048507, 'samples': 9799168, 'steps': 19138, 'loss/train': 3.754897356033325} 02/25/2022 02:22:02 - INFO - codeparrot_training - Step 19139: {'lr': 0.00035852779358702783, 'samples': 9799680, 'steps': 19139, 'loss/train': 1.8329936265945435} 02/25/2022 02:22:08 - INFO - codeparrot_training - Step 19140: {'lr': 0.00035851305307867197, 'samples': 9800192, 'steps': 19140, 'loss/train': 1.8142180442810059} 02/25/2022 02:22:11 - INFO - codeparrot_training - Step 19141: {'lr': 0.0003584983121054807, 'samples': 9800704, 'steps': 19141, 'loss/train': 0.7222133874893188} 02/25/2022 02:22:17 - INFO - codeparrot_training - Step 19142: {'lr': 0.00035848357066751703, 'samples': 9801216, 'steps': 19142, 'loss/train': 1.2115339040756226} 02/25/2022 02:22:20 - INFO - codeparrot_training - Step 19143: {'lr': 0.0003584688287648442, 'samples': 9801728, 'steps': 19143, 'loss/train': 1.5355087518692017} 02/25/2022 02:22:28 - INFO - codeparrot_training - Step 19144: {'lr': 0.00035845408639752544, 'samples': 9802240, 'steps': 19144, 'loss/train': 1.1106281280517578} 02/25/2022 02:22:31 - INFO - codeparrot_training - Step 19145: {'lr': 0.00035843934356562376, 'samples': 9802752, 'steps': 19145, 'loss/train': 1.4328804016113281} 02/25/2022 02:22:37 - INFO - codeparrot_training - Step 19146: {'lr': 0.00035842460026920233, 'samples': 9803264, 'steps': 19146, 'loss/train': 1.9341048002243042} 02/25/2022 02:22:40 - INFO - codeparrot_training - Step 19147: {'lr': 0.00035840985650832435, 'samples': 9803776, 'steps': 19147, 'loss/train': 0.9028723239898682} 02/25/2022 02:22:46 - INFO - codeparrot_training - Step 19148: {'lr': 0.000358395112283053, 'samples': 9804288, 'steps': 19148, 'loss/train': 2.2050766944885254} 02/25/2022 02:22:49 - INFO - codeparrot_training - Step 19149: {'lr': 0.00035838036759345144, 'samples': 9804800, 'steps': 19149, 'loss/train': 1.6372251510620117} 02/25/2022 02:22:55 - INFO - codeparrot_training - Step 19150: {'lr': 0.0003583656224395827, 'samples': 9805312, 'steps': 19150, 'loss/train': 2.295809268951416} 02/25/2022 02:22:58 - INFO - codeparrot_training - Step 19151: {'lr': 0.00035835087682151016, 'samples': 9805824, 'steps': 19151, 'loss/train': 1.2212183475494385} 02/25/2022 02:23:04 - INFO - codeparrot_training - Step 19152: {'lr': 0.00035833613073929684, 'samples': 9806336, 'steps': 19152, 'loss/train': 1.8894333839416504} 02/25/2022 02:23:07 - INFO - codeparrot_training - Step 19153: {'lr': 0.00035832138419300585, 'samples': 9806848, 'steps': 19153, 'loss/train': 2.1929712295532227} 02/25/2022 02:23:15 - INFO - codeparrot_training - Step 19154: {'lr': 0.00035830663718270056, 'samples': 9807360, 'steps': 19154, 'loss/train': 1.6250994205474854} 02/25/2022 02:23:18 - INFO - codeparrot_training - Step 19155: {'lr': 0.00035829188970844397, 'samples': 9807872, 'steps': 19155, 'loss/train': 1.4156405925750732} 02/25/2022 02:23:24 - INFO - codeparrot_training - Step 19156: {'lr': 0.0003582771417702993, 'samples': 9808384, 'steps': 19156, 'loss/train': 1.7554818391799927} 02/25/2022 02:23:27 - INFO - codeparrot_training - Step 19157: {'lr': 0.0003582623933683297, 'samples': 9808896, 'steps': 19157, 'loss/train': 1.4958889484405518} 02/25/2022 02:23:33 - INFO - codeparrot_training - Step 19158: {'lr': 0.0003582476445025985, 'samples': 9809408, 'steps': 19158, 'loss/train': 2.397771120071411} 02/25/2022 02:23:36 - INFO - codeparrot_training - Step 19159: {'lr': 0.00035823289517316866, 'samples': 9809920, 'steps': 19159, 'loss/train': 2.149447441101074} 02/25/2022 02:23:42 - INFO - codeparrot_training - Step 19160: {'lr': 0.00035821814538010356, 'samples': 9810432, 'steps': 19160, 'loss/train': 0.8857444524765015} 02/25/2022 02:23:47 - INFO - codeparrot_training - Step 19161: {'lr': 0.00035820339512346614, 'samples': 9810944, 'steps': 19161, 'loss/train': 1.5802701711654663} 02/25/2022 02:23:51 - INFO - codeparrot_training - Step 19162: {'lr': 0.0003581886444033199, 'samples': 9811456, 'steps': 19162, 'loss/train': 1.6313536167144775} 02/25/2022 02:23:54 - INFO - codeparrot_training - Step 19163: {'lr': 0.00035817389321972777, 'samples': 9811968, 'steps': 19163, 'loss/train': 1.4544962644577026} 02/25/2022 02:24:00 - INFO - codeparrot_training - Step 19164: {'lr': 0.000358159141572753, 'samples': 9812480, 'steps': 19164, 'loss/train': 2.8580377101898193} 02/25/2022 02:24:05 - INFO - codeparrot_training - Step 19165: {'lr': 0.0003581443894624589, 'samples': 9812992, 'steps': 19165, 'loss/train': 1.9356400966644287} 02/25/2022 02:24:09 - INFO - codeparrot_training - Step 19166: {'lr': 0.0003581296368889085, 'samples': 9813504, 'steps': 19166, 'loss/train': 0.6926233768463135} 02/25/2022 02:24:14 - INFO - codeparrot_training - Step 19167: {'lr': 0.0003581148838521651, 'samples': 9814016, 'steps': 19167, 'loss/train': 2.0719878673553467} 02/25/2022 02:24:18 - INFO - codeparrot_training - Step 19168: {'lr': 0.00035810013035229187, 'samples': 9814528, 'steps': 19168, 'loss/train': 1.8100639581680298} 02/25/2022 02:24:25 - INFO - codeparrot_training - Step 19169: {'lr': 0.000358085376389352, 'samples': 9815040, 'steps': 19169, 'loss/train': 2.408417224884033} 02/25/2022 02:24:28 - INFO - codeparrot_training - Step 19170: {'lr': 0.0003580706219634087, 'samples': 9815552, 'steps': 19170, 'loss/train': 1.9094805717468262} 02/25/2022 02:24:34 - INFO - codeparrot_training - Step 19171: {'lr': 0.0003580558670745252, 'samples': 9816064, 'steps': 19171, 'loss/train': 2.222322702407837} 02/25/2022 02:24:38 - INFO - codeparrot_training - Step 19172: {'lr': 0.00035804111172276464, 'samples': 9816576, 'steps': 19172, 'loss/train': 1.8012343645095825} 02/25/2022 02:24:43 - INFO - codeparrot_training - Step 19173: {'lr': 0.00035802635590819035, 'samples': 9817088, 'steps': 19173, 'loss/train': 2.3293824195861816} 02/25/2022 02:24:47 - INFO - codeparrot_training - Step 19174: {'lr': 0.00035801159963086535, 'samples': 9817600, 'steps': 19174, 'loss/train': 1.8405214548110962} 02/25/2022 02:24:52 - INFO - codeparrot_training - Step 19175: {'lr': 0.000357996842890853, 'samples': 9818112, 'steps': 19175, 'loss/train': 0.4635801911354065} 02/25/2022 02:24:56 - INFO - codeparrot_training - Step 19176: {'lr': 0.00035798208568821647, 'samples': 9818624, 'steps': 19176, 'loss/train': 0.1690002977848053} 02/25/2022 02:25:01 - INFO - codeparrot_training - Step 19177: {'lr': 0.00035796732802301895, 'samples': 9819136, 'steps': 19177, 'loss/train': 2.4281907081604004} 02/25/2022 02:25:05 - INFO - codeparrot_training - Step 19178: {'lr': 0.00035795256989532367, 'samples': 9819648, 'steps': 19178, 'loss/train': 0.8867130279541016} 02/25/2022 02:25:12 - INFO - codeparrot_training - Step 19179: {'lr': 0.0003579378113051939, 'samples': 9820160, 'steps': 19179, 'loss/train': 1.3388879299163818} 02/25/2022 02:25:16 - INFO - codeparrot_training - Step 19180: {'lr': 0.0003579230522526928, 'samples': 9820672, 'steps': 19180, 'loss/train': 1.3108952045440674} 02/25/2022 02:25:21 - INFO - codeparrot_training - Step 19181: {'lr': 0.00035790829273788356, 'samples': 9821184, 'steps': 19181, 'loss/train': 1.4172788858413696} 02/25/2022 02:25:25 - INFO - codeparrot_training - Step 19182: {'lr': 0.0003578935327608295, 'samples': 9821696, 'steps': 19182, 'loss/train': 1.2417763471603394} 02/25/2022 02:25:31 - INFO - codeparrot_training - Step 19183: {'lr': 0.00035787877232159384, 'samples': 9822208, 'steps': 19183, 'loss/train': 2.0434372425079346} 02/25/2022 02:25:34 - INFO - codeparrot_training - Step 19184: {'lr': 0.00035786401142023975, 'samples': 9822720, 'steps': 19184, 'loss/train': 1.7178515195846558} 02/25/2022 02:25:38 - INFO - codeparrot_training - Step 19185: {'lr': 0.0003578492500568304, 'samples': 9823232, 'steps': 19185, 'loss/train': 1.2154598236083984} 02/25/2022 02:25:44 - INFO - codeparrot_training - Step 19186: {'lr': 0.00035783448823142926, 'samples': 9823744, 'steps': 19186, 'loss/train': 1.6298420429229736} 02/25/2022 02:25:47 - INFO - codeparrot_training - Step 19187: {'lr': 0.00035781972594409937, 'samples': 9824256, 'steps': 19187, 'loss/train': 2.3240180015563965} 02/25/2022 02:25:53 - INFO - codeparrot_training - Step 19188: {'lr': 0.000357804963194904, 'samples': 9824768, 'steps': 19188, 'loss/train': 0.30438581109046936} 02/25/2022 02:25:56 - INFO - codeparrot_training - Step 19189: {'lr': 0.00035779019998390636, 'samples': 9825280, 'steps': 19189, 'loss/train': 2.120345115661621} 02/25/2022 02:26:03 - INFO - codeparrot_training - Step 19190: {'lr': 0.00035777543631116977, 'samples': 9825792, 'steps': 19190, 'loss/train': 0.20522382855415344} 02/25/2022 02:26:07 - INFO - codeparrot_training - Step 19191: {'lr': 0.00035776067217675744, 'samples': 9826304, 'steps': 19191, 'loss/train': 1.9031397104263306} 02/25/2022 02:26:13 - INFO - codeparrot_training - Step 19192: {'lr': 0.0003577459075807326, 'samples': 9826816, 'steps': 19192, 'loss/train': 2.523228406906128} 02/25/2022 02:26:16 - INFO - codeparrot_training - Step 19193: {'lr': 0.00035773114252315844, 'samples': 9827328, 'steps': 19193, 'loss/train': 1.9494825601577759} 02/25/2022 02:26:22 - INFO - codeparrot_training - Step 19194: {'lr': 0.0003577163770040984, 'samples': 9827840, 'steps': 19194, 'loss/train': 1.5968466997146606} 02/25/2022 02:26:25 - INFO - codeparrot_training - Step 19195: {'lr': 0.00035770161102361553, 'samples': 9828352, 'steps': 19195, 'loss/train': 8.619075775146484} 02/25/2022 02:26:31 - INFO - codeparrot_training - Step 19196: {'lr': 0.0003576868445817732, 'samples': 9828864, 'steps': 19196, 'loss/train': 1.5660347938537598} 02/25/2022 02:26:34 - INFO - codeparrot_training - Step 19197: {'lr': 0.00035767207767863453, 'samples': 9829376, 'steps': 19197, 'loss/train': 0.23878076672554016} 02/25/2022 02:26:40 - INFO - codeparrot_training - Step 19198: {'lr': 0.00035765731031426294, 'samples': 9829888, 'steps': 19198, 'loss/train': 2.0039756298065186} 02/25/2022 02:26:43 - INFO - codeparrot_training - Step 19199: {'lr': 0.0003576425424887216, 'samples': 9830400, 'steps': 19199, 'loss/train': 1.539983868598938} 02/25/2022 02:26:50 - INFO - codeparrot_training - Step 19200: {'lr': 0.0003576277742020738, 'samples': 9830912, 'steps': 19200, 'loss/train': 2.3642733097076416} 02/25/2022 02:26:54 - INFO - codeparrot_training - Step 19201: {'lr': 0.0003576130054543828, 'samples': 9831424, 'steps': 19201, 'loss/train': 1.3938478231430054} 02/25/2022 02:26:59 - INFO - codeparrot_training - Step 19202: {'lr': 0.00035759823624571184, 'samples': 9831936, 'steps': 19202, 'loss/train': 3.6907947063446045} 02/25/2022 02:27:03 - INFO - codeparrot_training - Step 19203: {'lr': 0.00035758346657612417, 'samples': 9832448, 'steps': 19203, 'loss/train': 1.7668558359146118} 02/25/2022 02:27:09 - INFO - codeparrot_training - Step 19204: {'lr': 0.00035756869644568314, 'samples': 9832960, 'steps': 19204, 'loss/train': 2.613354206085205} 02/25/2022 02:27:12 - INFO - codeparrot_training - Step 19205: {'lr': 0.00035755392585445195, 'samples': 9833472, 'steps': 19205, 'loss/train': 2.5741567611694336} 02/25/2022 02:27:18 - INFO - codeparrot_training - Step 19206: {'lr': 0.0003575391548024939, 'samples': 9833984, 'steps': 19206, 'loss/train': 1.9574772119522095} 02/25/2022 02:27:21 - INFO - codeparrot_training - Step 19207: {'lr': 0.00035752438328987224, 'samples': 9834496, 'steps': 19207, 'loss/train': 2.3425042629241943} 02/25/2022 02:27:27 - INFO - codeparrot_training - Step 19208: {'lr': 0.00035750961131665034, 'samples': 9835008, 'steps': 19208, 'loss/train': 3.004188060760498} 02/25/2022 02:27:30 - INFO - codeparrot_training - Step 19209: {'lr': 0.0003574948388828913, 'samples': 9835520, 'steps': 19209, 'loss/train': 2.12380051612854} 02/25/2022 02:27:36 - INFO - codeparrot_training - Step 19210: {'lr': 0.0003574800659886586, 'samples': 9836032, 'steps': 19210, 'loss/train': 1.2154111862182617} 02/25/2022 02:27:39 - INFO - codeparrot_training - Step 19211: {'lr': 0.0003574652926340153, 'samples': 9836544, 'steps': 19211, 'loss/train': 1.6454066038131714} 02/25/2022 02:27:45 - INFO - codeparrot_training - Step 19212: {'lr': 0.0003574505188190249, 'samples': 9837056, 'steps': 19212, 'loss/train': 1.9090946912765503} 02/25/2022 02:27:48 - INFO - codeparrot_training - Step 19213: {'lr': 0.0003574357445437506, 'samples': 9837568, 'steps': 19213, 'loss/train': 0.6663073301315308} 02/25/2022 02:27:54 - INFO - codeparrot_training - Step 19214: {'lr': 0.0003574209698082556, 'samples': 9838080, 'steps': 19214, 'loss/train': 0.4567076861858368} 02/25/2022 02:27:57 - INFO - codeparrot_training - Step 19215: {'lr': 0.0003574061946126034, 'samples': 9838592, 'steps': 19215, 'loss/train': 1.997956395149231} 02/25/2022 02:28:04 - INFO - codeparrot_training - Step 19216: {'lr': 0.0003573914189568571, 'samples': 9839104, 'steps': 19216, 'loss/train': 2.9051742553710938} 02/25/2022 02:28:08 - INFO - codeparrot_training - Step 19217: {'lr': 0.0003573766428410801, 'samples': 9839616, 'steps': 19217, 'loss/train': 2.022897958755493} 02/25/2022 02:28:13 - INFO - codeparrot_training - Step 19218: {'lr': 0.00035736186626533556, 'samples': 9840128, 'steps': 19218, 'loss/train': 1.0769011974334717} 02/25/2022 02:28:17 - INFO - codeparrot_training - Step 19219: {'lr': 0.00035734708922968695, 'samples': 9840640, 'steps': 19219, 'loss/train': 2.1259169578552246} 02/25/2022 02:28:22 - INFO - codeparrot_training - Step 19220: {'lr': 0.0003573323117341975, 'samples': 9841152, 'steps': 19220, 'loss/train': 5.245190620422363} 02/25/2022 02:28:26 - INFO - codeparrot_training - Step 19221: {'lr': 0.0003573175337789305, 'samples': 9841664, 'steps': 19221, 'loss/train': 1.7913321256637573} 02/25/2022 02:28:31 - INFO - codeparrot_training - Step 19222: {'lr': 0.0003573027553639492, 'samples': 9842176, 'steps': 19222, 'loss/train': 2.099760055541992} 02/25/2022 02:28:35 - INFO - codeparrot_training - Step 19223: {'lr': 0.0003572879764893171, 'samples': 9842688, 'steps': 19223, 'loss/train': 1.3859553337097168} 02/25/2022 02:28:40 - INFO - codeparrot_training - Step 19224: {'lr': 0.0003572731971550973, 'samples': 9843200, 'steps': 19224, 'loss/train': 2.9749035835266113} 02/25/2022 02:28:44 - INFO - codeparrot_training - Step 19225: {'lr': 0.0003572584173613532, 'samples': 9843712, 'steps': 19225, 'loss/train': 1.63325035572052} 02/25/2022 02:28:51 - INFO - codeparrot_training - Step 19226: {'lr': 0.00035724363710814807, 'samples': 9844224, 'steps': 19226, 'loss/train': 2.308938980102539} 02/25/2022 02:28:54 - INFO - codeparrot_training - Step 19227: {'lr': 0.00035722885639554526, 'samples': 9844736, 'steps': 19227, 'loss/train': 1.8411939144134521} 02/25/2022 02:29:00 - INFO - codeparrot_training - Step 19228: {'lr': 0.0003572140752236081, 'samples': 9845248, 'steps': 19228, 'loss/train': 1.3928197622299194} 02/25/2022 02:29:04 - INFO - codeparrot_training - Step 19229: {'lr': 0.0003571992935923999, 'samples': 9845760, 'steps': 19229, 'loss/train': 2.289527416229248} 02/25/2022 02:29:09 - INFO - codeparrot_training - Step 19230: {'lr': 0.0003571845115019839, 'samples': 9846272, 'steps': 19230, 'loss/train': 1.6074715852737427} 02/25/2022 02:29:13 - INFO - codeparrot_training - Step 19231: {'lr': 0.0003571697289524235, 'samples': 9846784, 'steps': 19231, 'loss/train': 0.21067163348197937} 02/25/2022 02:29:18 - INFO - codeparrot_training - Step 19232: {'lr': 0.0003571549459437821, 'samples': 9847296, 'steps': 19232, 'loss/train': 1.397789716720581} 02/25/2022 02:29:22 - INFO - codeparrot_training - Step 19233: {'lr': 0.0003571401624761229, 'samples': 9847808, 'steps': 19233, 'loss/train': 1.697490930557251} 02/25/2022 02:29:28 - INFO - codeparrot_training - Step 19234: {'lr': 0.0003571253785495092, 'samples': 9848320, 'steps': 19234, 'loss/train': 1.3955912590026855} 02/25/2022 02:29:31 - INFO - codeparrot_training - Step 19235: {'lr': 0.00035711059416400446, 'samples': 9848832, 'steps': 19235, 'loss/train': 1.4052941799163818} 02/25/2022 02:29:35 - INFO - codeparrot_training - Step 19236: {'lr': 0.0003570958093196719, 'samples': 9849344, 'steps': 19236, 'loss/train': 0.811290979385376} 02/25/2022 02:29:42 - INFO - codeparrot_training - Step 19237: {'lr': 0.00035708102401657495, 'samples': 9849856, 'steps': 19237, 'loss/train': 2.0424890518188477} 02/25/2022 02:29:45 - INFO - codeparrot_training - Step 19238: {'lr': 0.00035706623825477687, 'samples': 9850368, 'steps': 19238, 'loss/train': 0.5274319648742676} 02/25/2022 02:29:51 - INFO - codeparrot_training - Step 19239: {'lr': 0.000357051452034341, 'samples': 9850880, 'steps': 19239, 'loss/train': 1.2534005641937256} 02/25/2022 02:29:54 - INFO - codeparrot_training - Step 19240: {'lr': 0.00035703666535533076, 'samples': 9851392, 'steps': 19240, 'loss/train': 0.9870202541351318} 02/25/2022 02:30:00 - INFO - codeparrot_training - Step 19241: {'lr': 0.0003570218782178094, 'samples': 9851904, 'steps': 19241, 'loss/train': 1.8944878578186035} 02/25/2022 02:30:03 - INFO - codeparrot_training - Step 19242: {'lr': 0.0003570070906218403, 'samples': 9852416, 'steps': 19242, 'loss/train': 2.7563107013702393} 02/25/2022 02:30:09 - INFO - codeparrot_training - Step 19243: {'lr': 0.00035699230256748684, 'samples': 9852928, 'steps': 19243, 'loss/train': 1.8566226959228516} 02/25/2022 02:30:12 - INFO - codeparrot_training - Step 19244: {'lr': 0.0003569775140548122, 'samples': 9853440, 'steps': 19244, 'loss/train': 1.938361406326294} 02/25/2022 02:30:18 - INFO - codeparrot_training - Step 19245: {'lr': 0.00035696272508388, 'samples': 9853952, 'steps': 19245, 'loss/train': 2.3620991706848145} 02/25/2022 02:30:21 - INFO - codeparrot_training - Step 19246: {'lr': 0.00035694793565475337, 'samples': 9854464, 'steps': 19246, 'loss/train': 2.646620988845825} 02/25/2022 02:30:29 - INFO - codeparrot_training - Step 19247: {'lr': 0.0003569331457674958, 'samples': 9854976, 'steps': 19247, 'loss/train': 2.3118441104888916} 02/25/2022 02:30:32 - INFO - codeparrot_training - Step 19248: {'lr': 0.00035691835542217055, 'samples': 9855488, 'steps': 19248, 'loss/train': 1.075470209121704} 02/25/2022 02:30:38 - INFO - codeparrot_training - Step 19249: {'lr': 0.00035690356461884104, 'samples': 9856000, 'steps': 19249, 'loss/train': 1.482853651046753} 02/25/2022 02:30:43 - INFO - codeparrot_training - Step 19250: {'lr': 0.0003568887733575705, 'samples': 9856512, 'steps': 19250, 'loss/train': 1.8895548582077026} 02/25/2022 02:30:46 - INFO - codeparrot_training - Step 19251: {'lr': 0.0003568739816384225, 'samples': 9857024, 'steps': 19251, 'loss/train': 1.1049559116363525} 02/25/2022 02:30:52 - INFO - codeparrot_training - Step 19252: {'lr': 0.00035685918946146036, 'samples': 9857536, 'steps': 19252, 'loss/train': 2.8000295162200928} 02/25/2022 02:30:56 - INFO - codeparrot_training - Step 19253: {'lr': 0.00035684439682674723, 'samples': 9858048, 'steps': 19253, 'loss/train': 1.9263113737106323} 02/25/2022 02:31:01 - INFO - codeparrot_training - Step 19254: {'lr': 0.00035682960373434677, 'samples': 9858560, 'steps': 19254, 'loss/train': 0.33952149748802185} 02/25/2022 02:31:05 - INFO - codeparrot_training - Step 19255: {'lr': 0.0003568148101843221, 'samples': 9859072, 'steps': 19255, 'loss/train': 2.2760066986083984} 02/25/2022 02:31:08 - INFO - codeparrot_training - Step 19256: {'lr': 0.0003568000161767368, 'samples': 9859584, 'steps': 19256, 'loss/train': 2.2122201919555664} 02/25/2022 02:31:14 - INFO - codeparrot_training - Step 19257: {'lr': 0.00035678522171165406, 'samples': 9860096, 'steps': 19257, 'loss/train': 2.1850576400756836} 02/25/2022 02:31:19 - INFO - codeparrot_training - Step 19258: {'lr': 0.0003567704267891374, 'samples': 9860608, 'steps': 19258, 'loss/train': 1.573604941368103} 02/25/2022 02:31:23 - INFO - codeparrot_training - Step 19259: {'lr': 0.00035675563140925, 'samples': 9861120, 'steps': 19259, 'loss/train': 2.432962417602539} 02/25/2022 02:31:28 - INFO - codeparrot_training - Step 19260: {'lr': 0.00035674083557205553, 'samples': 9861632, 'steps': 19260, 'loss/train': 1.551035761833191} 02/25/2022 02:31:32 - INFO - codeparrot_training - Step 19261: {'lr': 0.00035672603927761716, 'samples': 9862144, 'steps': 19261, 'loss/train': 1.7010586261749268} 02/25/2022 02:31:39 - INFO - codeparrot_training - Step 19262: {'lr': 0.0003567112425259984, 'samples': 9862656, 'steps': 19262, 'loss/train': 1.618138074874878} 02/25/2022 02:31:42 - INFO - codeparrot_training - Step 19263: {'lr': 0.00035669644531726244, 'samples': 9863168, 'steps': 19263, 'loss/train': 2.0759124755859375} 02/25/2022 02:31:48 - INFO - codeparrot_training - Step 19264: {'lr': 0.00035668164765147284, 'samples': 9863680, 'steps': 19264, 'loss/train': 1.9786914587020874} 02/25/2022 02:31:51 - INFO - codeparrot_training - Step 19265: {'lr': 0.00035666684952869295, 'samples': 9864192, 'steps': 19265, 'loss/train': 2.208364963531494} 02/25/2022 02:31:57 - INFO - codeparrot_training - Step 19266: {'lr': 0.00035665205094898613, 'samples': 9864704, 'steps': 19266, 'loss/train': 1.7359892129898071} 02/25/2022 02:32:00 - INFO - codeparrot_training - Step 19267: {'lr': 0.0003566372519124158, 'samples': 9865216, 'steps': 19267, 'loss/train': 2.4588685035705566} 02/25/2022 02:32:06 - INFO - codeparrot_training - Step 19268: {'lr': 0.00035662245241904533, 'samples': 9865728, 'steps': 19268, 'loss/train': 1.9101704359054565} 02/25/2022 02:32:09 - INFO - codeparrot_training - Step 19269: {'lr': 0.00035660765246893814, 'samples': 9866240, 'steps': 19269, 'loss/train': 1.653025507926941} 02/25/2022 02:32:15 - INFO - codeparrot_training - Step 19270: {'lr': 0.00035659285206215757, 'samples': 9866752, 'steps': 19270, 'loss/train': 2.1290361881256104} 02/25/2022 02:32:18 - INFO - codeparrot_training - Step 19271: {'lr': 0.0003565780511987672, 'samples': 9867264, 'steps': 19271, 'loss/train': 2.0186846256256104} 02/25/2022 02:32:25 - INFO - codeparrot_training - Step 19272: {'lr': 0.00035656324987883015, 'samples': 9867776, 'steps': 19272, 'loss/train': 1.522246241569519} 02/25/2022 02:32:29 - INFO - codeparrot_training - Step 19273: {'lr': 0.00035654844810241004, 'samples': 9868288, 'steps': 19273, 'loss/train': 2.132499933242798} 02/25/2022 02:32:34 - INFO - codeparrot_training - Step 19274: {'lr': 0.0003565336458695702, 'samples': 9868800, 'steps': 19274, 'loss/train': 2.6792471408843994} 02/25/2022 02:32:38 - INFO - codeparrot_training - Step 19275: {'lr': 0.0003565188431803741, 'samples': 9869312, 'steps': 19275, 'loss/train': 2.4057610034942627} 02/25/2022 02:32:44 - INFO - codeparrot_training - Step 19276: {'lr': 0.0003565040400348851, 'samples': 9869824, 'steps': 19276, 'loss/train': 1.4288160800933838} 02/25/2022 02:32:47 - INFO - codeparrot_training - Step 19277: {'lr': 0.0003564892364331665, 'samples': 9870336, 'steps': 19277, 'loss/train': 1.3318730592727661} 02/25/2022 02:32:53 - INFO - codeparrot_training - Step 19278: {'lr': 0.0003564744323752819, 'samples': 9870848, 'steps': 19278, 'loss/train': 0.21651515364646912} 02/25/2022 02:32:56 - INFO - codeparrot_training - Step 19279: {'lr': 0.00035645962786129464, 'samples': 9871360, 'steps': 19279, 'loss/train': 1.3973026275634766} 02/25/2022 02:33:02 - INFO - codeparrot_training - Step 19280: {'lr': 0.0003564448228912682, 'samples': 9871872, 'steps': 19280, 'loss/train': 2.7846038341522217} 02/25/2022 02:33:05 - INFO - codeparrot_training - Step 19281: {'lr': 0.00035643001746526586, 'samples': 9872384, 'steps': 19281, 'loss/train': 1.4548993110656738} 02/25/2022 02:33:13 - INFO - codeparrot_training - Step 19282: {'lr': 0.0003564152115833511, 'samples': 9872896, 'steps': 19282, 'loss/train': 1.5476588010787964} 02/25/2022 02:33:16 - INFO - codeparrot_training - Step 19283: {'lr': 0.00035640040524558746, 'samples': 9873408, 'steps': 19283, 'loss/train': 2.1892807483673096} 02/25/2022 02:33:22 - INFO - codeparrot_training - Step 19284: {'lr': 0.00035638559845203817, 'samples': 9873920, 'steps': 19284, 'loss/train': 1.2143034934997559} 02/25/2022 02:33:25 - INFO - codeparrot_training - Step 19285: {'lr': 0.00035637079120276683, 'samples': 9874432, 'steps': 19285, 'loss/train': 1.911623239517212} 02/25/2022 02:33:31 - INFO - codeparrot_training - Step 19286: {'lr': 0.00035635598349783676, 'samples': 9874944, 'steps': 19286, 'loss/train': 2.026838541030884} 02/25/2022 02:33:34 - INFO - codeparrot_training - Step 19287: {'lr': 0.0003563411753373115, 'samples': 9875456, 'steps': 19287, 'loss/train': 0.9892092347145081} 02/25/2022 02:33:40 - INFO - codeparrot_training - Step 19288: {'lr': 0.0003563263667212543, 'samples': 9875968, 'steps': 19288, 'loss/train': 1.7014323472976685} 02/25/2022 02:33:43 - INFO - codeparrot_training - Step 19289: {'lr': 0.0003563115576497288, 'samples': 9876480, 'steps': 19289, 'loss/train': 2.5118744373321533} 02/25/2022 02:33:49 - INFO - codeparrot_training - Step 19290: {'lr': 0.0003562967481227982, 'samples': 9876992, 'steps': 19290, 'loss/train': 2.366943359375} 02/25/2022 02:33:52 - INFO - codeparrot_training - Step 19291: {'lr': 0.0003562819381405262, 'samples': 9877504, 'steps': 19291, 'loss/train': 1.967867374420166} 02/25/2022 02:33:58 - INFO - codeparrot_training - Step 19292: {'lr': 0.0003562671277029761, 'samples': 9878016, 'steps': 19292, 'loss/train': 2.4265329837799072} 02/25/2022 02:34:01 - INFO - codeparrot_training - Step 19293: {'lr': 0.0003562523168102114, 'samples': 9878528, 'steps': 19293, 'loss/train': 0.9066846966743469} 02/25/2022 02:34:07 - INFO - codeparrot_training - Step 19294: {'lr': 0.00035623750546229547, 'samples': 9879040, 'steps': 19294, 'loss/train': 2.2698421478271484} 02/25/2022 02:34:11 - INFO - codeparrot_training - Step 19295: {'lr': 0.0003562226936592919, 'samples': 9879552, 'steps': 19295, 'loss/train': 2.1982572078704834} 02/25/2022 02:34:16 - INFO - codeparrot_training - Step 19296: {'lr': 0.0003562078814012639, 'samples': 9880064, 'steps': 19296, 'loss/train': 1.8500181436538696} 02/25/2022 02:34:20 - INFO - codeparrot_training - Step 19297: {'lr': 0.00035619306868827516, 'samples': 9880576, 'steps': 19297, 'loss/train': 1.2079533338546753} 02/25/2022 02:34:25 - INFO - codeparrot_training - Step 19298: {'lr': 0.00035617825552038894, 'samples': 9881088, 'steps': 19298, 'loss/train': 1.5644276142120361} 02/25/2022 02:34:29 - INFO - codeparrot_training - Step 19299: {'lr': 0.00035616344189766885, 'samples': 9881600, 'steps': 19299, 'loss/train': 1.6737005710601807} 02/25/2022 02:34:34 - INFO - codeparrot_training - Step 19300: {'lr': 0.0003561486278201783, 'samples': 9882112, 'steps': 19300, 'loss/train': 1.916688084602356} 02/25/2022 02:34:38 - INFO - codeparrot_training - Step 19301: {'lr': 0.00035613381328798065, 'samples': 9882624, 'steps': 19301, 'loss/train': 3.0540406703948975} 02/25/2022 02:34:43 - INFO - codeparrot_training - Step 19302: {'lr': 0.0003561189983011396, 'samples': 9883136, 'steps': 19302, 'loss/train': 1.4626095294952393} 02/25/2022 02:34:47 - INFO - codeparrot_training - Step 19303: {'lr': 0.00035610418285971835, 'samples': 9883648, 'steps': 19303, 'loss/train': 2.0717930793762207} 02/25/2022 02:34:52 - INFO - codeparrot_training - Step 19304: {'lr': 0.00035608936696378046, 'samples': 9884160, 'steps': 19304, 'loss/train': 0.9057942032814026} 02/25/2022 02:34:56 - INFO - codeparrot_training - Step 19305: {'lr': 0.00035607455061338947, 'samples': 9884672, 'steps': 19305, 'loss/train': 3.1500675678253174} 02/25/2022 02:35:01 - INFO - codeparrot_training - Step 19306: {'lr': 0.0003560597338086088, 'samples': 9885184, 'steps': 19306, 'loss/train': 1.528065800666809} 02/25/2022 02:35:05 - INFO - codeparrot_training - Step 19307: {'lr': 0.0003560449165495018, 'samples': 9885696, 'steps': 19307, 'loss/train': 1.8838484287261963} 02/25/2022 02:35:12 - INFO - codeparrot_training - Step 19308: {'lr': 0.0003560300988361321, 'samples': 9886208, 'steps': 19308, 'loss/train': 1.3809232711791992} 02/25/2022 02:35:15 - INFO - codeparrot_training - Step 19309: {'lr': 0.00035601528066856315, 'samples': 9886720, 'steps': 19309, 'loss/train': 2.148930788040161} 02/25/2022 02:35:21 - INFO - codeparrot_training - Step 19310: {'lr': 0.00035600046204685844, 'samples': 9887232, 'steps': 19310, 'loss/train': 1.2641681432724} 02/25/2022 02:35:24 - INFO - codeparrot_training - Step 19311: {'lr': 0.00035598564297108134, 'samples': 9887744, 'steps': 19311, 'loss/train': 2.1563804149627686} 02/25/2022 02:35:29 - INFO - codeparrot_training - Step 19312: {'lr': 0.0003559708234412954, 'samples': 9888256, 'steps': 19312, 'loss/train': 2.0812201499938965} 02/25/2022 02:35:33 - INFO - codeparrot_training - Step 19313: {'lr': 0.00035595600345756414, 'samples': 9888768, 'steps': 19313, 'loss/train': 1.456243872642517} 02/25/2022 02:35:39 - INFO - codeparrot_training - Step 19314: {'lr': 0.00035594118301995095, 'samples': 9889280, 'steps': 19314, 'loss/train': 0.8808965086936951} 02/25/2022 02:35:42 - INFO - codeparrot_training - Step 19315: {'lr': 0.00035592636212851945, 'samples': 9889792, 'steps': 19315, 'loss/train': 1.7614250183105469} 02/25/2022 02:35:48 - INFO - codeparrot_training - Step 19316: {'lr': 0.000355911540783333, 'samples': 9890304, 'steps': 19316, 'loss/train': 0.846782922744751} 02/25/2022 02:35:51 - INFO - codeparrot_training - Step 19317: {'lr': 0.00035589671898445517, 'samples': 9890816, 'steps': 19317, 'loss/train': 2.443593740463257} 02/25/2022 02:35:57 - INFO - codeparrot_training - Step 19318: {'lr': 0.0003558818967319494, 'samples': 9891328, 'steps': 19318, 'loss/train': 1.5117106437683105} 02/25/2022 02:36:00 - INFO - codeparrot_training - Step 19319: {'lr': 0.0003558670740258792, 'samples': 9891840, 'steps': 19319, 'loss/train': 1.2961843013763428} 02/25/2022 02:36:07 - INFO - codeparrot_training - Step 19320: {'lr': 0.00035585225086630807, 'samples': 9892352, 'steps': 19320, 'loss/train': 1.956886887550354} 02/25/2022 02:36:11 - INFO - codeparrot_training - Step 19321: {'lr': 0.00035583742725329954, 'samples': 9892864, 'steps': 19321, 'loss/train': 1.6915360689163208} 02/25/2022 02:36:16 - INFO - codeparrot_training - Step 19322: {'lr': 0.0003558226031869171, 'samples': 9893376, 'steps': 19322, 'loss/train': 0.9970734119415283} 02/25/2022 02:36:20 - INFO - codeparrot_training - Step 19323: {'lr': 0.00035580777866722415, 'samples': 9893888, 'steps': 19323, 'loss/train': 0.5012323260307312} 02/25/2022 02:36:25 - INFO - codeparrot_training - Step 19324: {'lr': 0.00035579295369428425, 'samples': 9894400, 'steps': 19324, 'loss/train': 1.9734588861465454} 02/25/2022 02:36:29 - INFO - codeparrot_training - Step 19325: {'lr': 0.000355778128268161, 'samples': 9894912, 'steps': 19325, 'loss/train': 1.9326817989349365} 02/25/2022 02:36:34 - INFO - codeparrot_training - Step 19326: {'lr': 0.0003557633023889179, 'samples': 9895424, 'steps': 19326, 'loss/train': 1.4279378652572632} 02/25/2022 02:36:37 - INFO - codeparrot_training - Step 19327: {'lr': 0.0003557484760566183, 'samples': 9895936, 'steps': 19327, 'loss/train': 1.6264982223510742} 02/25/2022 02:36:43 - INFO - codeparrot_training - Step 19328: {'lr': 0.0003557336492713258, 'samples': 9896448, 'steps': 19328, 'loss/train': 1.450286626815796} 02/25/2022 02:36:46 - INFO - codeparrot_training - Step 19329: {'lr': 0.000355718822033104, 'samples': 9896960, 'steps': 19329, 'loss/train': 2.410184383392334} 02/25/2022 02:36:53 - INFO - codeparrot_training - Step 19330: {'lr': 0.0003557039943420163, 'samples': 9897472, 'steps': 19330, 'loss/train': 1.1350754499435425} 02/25/2022 02:36:56 - INFO - codeparrot_training - Step 19331: {'lr': 0.00035568916619812624, 'samples': 9897984, 'steps': 19331, 'loss/train': 2.0210282802581787} 02/25/2022 02:37:02 - INFO - codeparrot_training - Step 19332: {'lr': 0.00035567433760149737, 'samples': 9898496, 'steps': 19332, 'loss/train': 0.6852190494537354} 02/25/2022 02:37:05 - INFO - codeparrot_training - Step 19333: {'lr': 0.0003556595085521931, 'samples': 9899008, 'steps': 19333, 'loss/train': 0.6045637130737305} 02/25/2022 02:37:11 - INFO - codeparrot_training - Step 19334: {'lr': 0.0003556446790502772, 'samples': 9899520, 'steps': 19334, 'loss/train': 2.5709099769592285} 02/25/2022 02:37:14 - INFO - codeparrot_training - Step 19335: {'lr': 0.00035562984909581297, 'samples': 9900032, 'steps': 19335, 'loss/train': 2.6259469985961914} 02/25/2022 02:37:20 - INFO - codeparrot_training - Step 19336: {'lr': 0.0003556150186888639, 'samples': 9900544, 'steps': 19336, 'loss/train': 1.0845199823379517} 02/25/2022 02:37:23 - INFO - codeparrot_training - Step 19337: {'lr': 0.00035560018782949384, 'samples': 9901056, 'steps': 19337, 'loss/train': 1.792197823524475} 02/25/2022 02:37:29 - INFO - codeparrot_training - Step 19338: {'lr': 0.0003555853565177659, 'samples': 9901568, 'steps': 19338, 'loss/train': 1.0452955961227417} 02/25/2022 02:37:32 - INFO - codeparrot_training - Step 19339: {'lr': 0.00035557052475374397, 'samples': 9902080, 'steps': 19339, 'loss/train': 2.3191845417022705} 02/25/2022 02:37:38 - INFO - codeparrot_training - Step 19340: {'lr': 0.00035555569253749135, 'samples': 9902592, 'steps': 19340, 'loss/train': 1.4723395109176636} 02/25/2022 02:37:41 - INFO - codeparrot_training - Step 19341: {'lr': 0.0003555408598690718, 'samples': 9903104, 'steps': 19341, 'loss/train': 2.416355609893799} 02/25/2022 02:37:47 - INFO - codeparrot_training - Step 19342: {'lr': 0.0003555260267485485, 'samples': 9903616, 'steps': 19342, 'loss/train': 5.271160125732422} 02/25/2022 02:37:50 - INFO - codeparrot_training - Step 19343: {'lr': 0.00035551119317598533, 'samples': 9904128, 'steps': 19343, 'loss/train': 1.906678318977356} 02/25/2022 02:37:56 - INFO - codeparrot_training - Step 19344: {'lr': 0.00035549635915144574, 'samples': 9904640, 'steps': 19344, 'loss/train': 1.712480902671814} 02/25/2022 02:38:00 - INFO - codeparrot_training - Step 19345: {'lr': 0.0003554815246749932, 'samples': 9905152, 'steps': 19345, 'loss/train': 0.8412127494812012} 02/25/2022 02:38:05 - INFO - codeparrot_training - Step 19346: {'lr': 0.00035546668974669127, 'samples': 9905664, 'steps': 19346, 'loss/train': 1.625795841217041} 02/25/2022 02:38:09 - INFO - codeparrot_training - Step 19347: {'lr': 0.00035545185436660357, 'samples': 9906176, 'steps': 19347, 'loss/train': 2.363006353378296} 02/25/2022 02:38:14 - INFO - codeparrot_training - Step 19348: {'lr': 0.00035543701853479366, 'samples': 9906688, 'steps': 19348, 'loss/train': 1.4702732563018799} 02/25/2022 02:38:18 - INFO - codeparrot_training - Step 19349: {'lr': 0.00035542218225132497, 'samples': 9907200, 'steps': 19349, 'loss/train': 1.820178747177124} 02/25/2022 02:38:23 - INFO - codeparrot_training - Step 19350: {'lr': 0.00035540734551626113, 'samples': 9907712, 'steps': 19350, 'loss/train': 1.2415663003921509} 02/25/2022 02:38:27 - INFO - codeparrot_training - Step 19351: {'lr': 0.00035539250832966574, 'samples': 9908224, 'steps': 19351, 'loss/train': 2.0476346015930176} 02/25/2022 02:38:32 - INFO - codeparrot_training - Step 19352: {'lr': 0.00035537767069160234, 'samples': 9908736, 'steps': 19352, 'loss/train': 2.33284592628479} 02/25/2022 02:38:36 - INFO - codeparrot_training - Step 19353: {'lr': 0.00035536283260213434, 'samples': 9909248, 'steps': 19353, 'loss/train': 2.6625912189483643} 02/25/2022 02:38:41 - INFO - codeparrot_training - Step 19354: {'lr': 0.0003553479940613255, 'samples': 9909760, 'steps': 19354, 'loss/train': 2.5236716270446777} 02/25/2022 02:38:45 - INFO - codeparrot_training - Step 19355: {'lr': 0.00035533315506923924, 'samples': 9910272, 'steps': 19355, 'loss/train': 2.708268165588379} 02/25/2022 02:38:51 - INFO - codeparrot_training - Step 19356: {'lr': 0.0003553183156259393, 'samples': 9910784, 'steps': 19356, 'loss/train': 1.909191370010376} 02/25/2022 02:38:54 - INFO - codeparrot_training - Step 19357: {'lr': 0.00035530347573148904, 'samples': 9911296, 'steps': 19357, 'loss/train': 2.39231276512146} 02/25/2022 02:39:00 - INFO - codeparrot_training - Step 19358: {'lr': 0.0003552886353859522, 'samples': 9911808, 'steps': 19358, 'loss/train': 2.1169612407684326} 02/25/2022 02:39:03 - INFO - codeparrot_training - Step 19359: {'lr': 0.00035527379458939225, 'samples': 9912320, 'steps': 19359, 'loss/train': 2.254922389984131} 02/25/2022 02:39:09 - INFO - codeparrot_training - Step 19360: {'lr': 0.00035525895334187274, 'samples': 9912832, 'steps': 19360, 'loss/train': 1.441593050956726} 02/25/2022 02:39:12 - INFO - codeparrot_training - Step 19361: {'lr': 0.0003552441116434574, 'samples': 9913344, 'steps': 19361, 'loss/train': 2.53437876701355} 02/25/2022 02:39:18 - INFO - codeparrot_training - Step 19362: {'lr': 0.0003552292694942096, 'samples': 9913856, 'steps': 19362, 'loss/train': 3.4571027755737305} 02/25/2022 02:39:21 - INFO - codeparrot_training - Step 19363: {'lr': 0.0003552144268941931, 'samples': 9914368, 'steps': 19363, 'loss/train': 2.1294069290161133} 02/25/2022 02:39:27 - INFO - codeparrot_training - Step 19364: {'lr': 0.00035519958384347134, 'samples': 9914880, 'steps': 19364, 'loss/train': 3.0064258575439453} 02/25/2022 02:39:31 - INFO - codeparrot_training - Step 19365: {'lr': 0.000355184740342108, 'samples': 9915392, 'steps': 19365, 'loss/train': 2.2653021812438965} 02/25/2022 02:39:37 - INFO - codeparrot_training - Step 19366: {'lr': 0.00035516989639016664, 'samples': 9915904, 'steps': 19366, 'loss/train': 0.742005467414856} 02/25/2022 02:39:40 - INFO - codeparrot_training - Step 19367: {'lr': 0.00035515505198771086, 'samples': 9916416, 'steps': 19367, 'loss/train': 2.0157525539398193} 02/25/2022 02:39:46 - INFO - codeparrot_training - Step 19368: {'lr': 0.0003551402071348042, 'samples': 9916928, 'steps': 19368, 'loss/train': 1.8428932428359985} 02/25/2022 02:39:49 - INFO - codeparrot_training - Step 19369: {'lr': 0.0003551253618315103, 'samples': 9917440, 'steps': 19369, 'loss/train': 0.5470841526985168} 02/25/2022 02:39:55 - INFO - codeparrot_training - Step 19370: {'lr': 0.0003551105160778927, 'samples': 9917952, 'steps': 19370, 'loss/train': 2.891383171081543} 02/25/2022 02:39:58 - INFO - codeparrot_training - Step 19371: {'lr': 0.000355095669874015, 'samples': 9918464, 'steps': 19371, 'loss/train': 2.265749931335449} 02/25/2022 02:40:04 - INFO - codeparrot_training - Step 19372: {'lr': 0.00035508082321994097, 'samples': 9918976, 'steps': 19372, 'loss/train': 0.6018871068954468} 02/25/2022 02:40:08 - INFO - codeparrot_training - Step 19373: {'lr': 0.00035506597611573387, 'samples': 9919488, 'steps': 19373, 'loss/train': 1.6407136917114258} 02/25/2022 02:40:13 - INFO - codeparrot_training - Step 19374: {'lr': 0.0003550511285614576, 'samples': 9920000, 'steps': 19374, 'loss/train': 1.9741944074630737} 02/25/2022 02:40:17 - INFO - codeparrot_training - Step 19375: {'lr': 0.0003550362805571756, 'samples': 9920512, 'steps': 19375, 'loss/train': 0.6745988130569458} 02/25/2022 02:40:23 - INFO - codeparrot_training - Step 19376: {'lr': 0.00035502143210295163, 'samples': 9921024, 'steps': 19376, 'loss/train': 1.1684695482254028} 02/25/2022 02:40:26 - INFO - codeparrot_training - Step 19377: {'lr': 0.000355006583198849, 'samples': 9921536, 'steps': 19377, 'loss/train': 1.9040051698684692} 02/25/2022 02:40:32 - INFO - codeparrot_training - Step 19378: {'lr': 0.00035499173384493174, 'samples': 9922048, 'steps': 19378, 'loss/train': 2.65724515914917} 02/25/2022 02:40:35 - INFO - codeparrot_training - Step 19379: {'lr': 0.00035497688404126306, 'samples': 9922560, 'steps': 19379, 'loss/train': 2.56378436088562} 02/25/2022 02:40:41 - INFO - codeparrot_training - Step 19380: {'lr': 0.00035496203378790683, 'samples': 9923072, 'steps': 19380, 'loss/train': 1.8325612545013428} 02/25/2022 02:40:44 - INFO - codeparrot_training - Step 19381: {'lr': 0.0003549471830849265, 'samples': 9923584, 'steps': 19381, 'loss/train': 2.72446870803833} 02/25/2022 02:40:50 - INFO - codeparrot_training - Step 19382: {'lr': 0.00035493233193238584, 'samples': 9924096, 'steps': 19382, 'loss/train': 2.2228333950042725} 02/25/2022 02:40:53 - INFO - codeparrot_training - Step 19383: {'lr': 0.00035491748033034836, 'samples': 9924608, 'steps': 19383, 'loss/train': 2.292468547821045} 02/25/2022 02:40:59 - INFO - codeparrot_training - Step 19384: {'lr': 0.00035490262827887764, 'samples': 9925120, 'steps': 19384, 'loss/train': 1.4534056186676025} 02/25/2022 02:41:02 - INFO - codeparrot_training - Step 19385: {'lr': 0.0003548877757780375, 'samples': 9925632, 'steps': 19385, 'loss/train': 2.4866769313812256} 02/25/2022 02:41:09 - INFO - codeparrot_training - Step 19386: {'lr': 0.00035487292282789136, 'samples': 9926144, 'steps': 19386, 'loss/train': 2.3067026138305664} 02/25/2022 02:41:12 - INFO - codeparrot_training - Step 19387: {'lr': 0.000354858069428503, 'samples': 9926656, 'steps': 19387, 'loss/train': 1.1646324396133423} 02/25/2022 02:41:18 - INFO - codeparrot_training - Step 19388: {'lr': 0.0003548432155799358, 'samples': 9927168, 'steps': 19388, 'loss/train': 2.183593511581421} 02/25/2022 02:41:21 - INFO - codeparrot_training - Step 19389: {'lr': 0.0003548283612822537, 'samples': 9927680, 'steps': 19389, 'loss/train': 0.9730269312858582} 02/25/2022 02:41:27 - INFO - codeparrot_training - Step 19390: {'lr': 0.0003548135065355201, 'samples': 9928192, 'steps': 19390, 'loss/train': 2.060657262802124} 02/25/2022 02:41:30 - INFO - codeparrot_training - Step 19391: {'lr': 0.0003547986513397988, 'samples': 9928704, 'steps': 19391, 'loss/train': 2.292754888534546} 02/25/2022 02:41:36 - INFO - codeparrot_training - Step 19392: {'lr': 0.0003547837956951533, 'samples': 9929216, 'steps': 19392, 'loss/train': 2.0760703086853027} 02/25/2022 02:41:41 - INFO - codeparrot_training - Step 19393: {'lr': 0.00035476893960164734, 'samples': 9929728, 'steps': 19393, 'loss/train': 2.142906665802002} 02/25/2022 02:41:45 - INFO - codeparrot_training - Step 19394: {'lr': 0.00035475408305934444, 'samples': 9930240, 'steps': 19394, 'loss/train': 2.280719518661499} 02/25/2022 02:41:50 - INFO - codeparrot_training - Step 19395: {'lr': 0.0003547392260683084, 'samples': 9930752, 'steps': 19395, 'loss/train': 0.554097056388855} 02/25/2022 02:41:54 - INFO - codeparrot_training - Step 19396: {'lr': 0.0003547243686286027, 'samples': 9931264, 'steps': 19396, 'loss/train': 2.9527838230133057} 02/25/2022 02:41:57 - INFO - codeparrot_training - Step 19397: {'lr': 0.000354709510740291, 'samples': 9931776, 'steps': 19397, 'loss/train': 0.9633437991142273} 02/25/2022 02:42:03 - INFO - codeparrot_training - Step 19398: {'lr': 0.0003546946524034371, 'samples': 9932288, 'steps': 19398, 'loss/train': 3.556969404220581} 02/25/2022 02:42:06 - INFO - codeparrot_training - Step 19399: {'lr': 0.00035467979361810455, 'samples': 9932800, 'steps': 19399, 'loss/train': 1.7744941711425781} 02/25/2022 02:42:12 - INFO - codeparrot_training - Step 19400: {'lr': 0.00035466493438435703, 'samples': 9933312, 'steps': 19400, 'loss/train': 1.5338068008422852} 02/25/2022 02:42:18 - INFO - codeparrot_training - Step 19401: {'lr': 0.00035465007470225813, 'samples': 9933824, 'steps': 19401, 'loss/train': 1.9150464534759521} 02/25/2022 02:42:21 - INFO - codeparrot_training - Step 19402: {'lr': 0.0003546352145718715, 'samples': 9934336, 'steps': 19402, 'loss/train': 1.7839462757110596} 02/25/2022 02:42:27 - INFO - codeparrot_training - Step 19403: {'lr': 0.0003546203539932609, 'samples': 9934848, 'steps': 19403, 'loss/train': 2.4478936195373535} 02/25/2022 02:42:30 - INFO - codeparrot_training - Step 19404: {'lr': 0.0003546054929664899, 'samples': 9935360, 'steps': 19404, 'loss/train': 2.5634331703186035} 02/25/2022 02:42:34 - INFO - codeparrot_training - Step 19405: {'lr': 0.0003545906314916222, 'samples': 9935872, 'steps': 19405, 'loss/train': 1.7341463565826416} 02/25/2022 02:42:39 - INFO - codeparrot_training - Step 19406: {'lr': 0.00035457576956872145, 'samples': 9936384, 'steps': 19406, 'loss/train': 2.2536680698394775} 02/25/2022 02:42:43 - INFO - codeparrot_training - Step 19407: {'lr': 0.00035456090719785126, 'samples': 9936896, 'steps': 19407, 'loss/train': 1.537829041481018} 02/25/2022 02:42:48 - INFO - codeparrot_training - Step 19408: {'lr': 0.00035454604437907536, 'samples': 9937408, 'steps': 19408, 'loss/train': 1.2349534034729004} 02/25/2022 02:42:54 - INFO - codeparrot_training - Step 19409: {'lr': 0.0003545311811124574, 'samples': 9937920, 'steps': 19409, 'loss/train': 2.5081257820129395} 02/25/2022 02:42:57 - INFO - codeparrot_training - Step 19410: {'lr': 0.0003545163173980611, 'samples': 9938432, 'steps': 19410, 'loss/train': 2.3050005435943604} 02/25/2022 02:43:03 - INFO - codeparrot_training - Step 19411: {'lr': 0.0003545014532359501, 'samples': 9938944, 'steps': 19411, 'loss/train': 0.22542831301689148} 02/25/2022 02:43:07 - INFO - codeparrot_training - Step 19412: {'lr': 0.000354486588626188, 'samples': 9939456, 'steps': 19412, 'loss/train': 1.9168795347213745} 02/25/2022 02:43:13 - INFO - codeparrot_training - Step 19413: {'lr': 0.0003544717235688385, 'samples': 9939968, 'steps': 19413, 'loss/train': 2.002089738845825} 02/25/2022 02:43:16 - INFO - codeparrot_training - Step 19414: {'lr': 0.00035445685806396543, 'samples': 9940480, 'steps': 19414, 'loss/train': 2.94095778465271} 02/25/2022 02:43:21 - INFO - codeparrot_training - Step 19415: {'lr': 0.00035444199211163226, 'samples': 9940992, 'steps': 19415, 'loss/train': 2.3083767890930176} 02/25/2022 02:43:25 - INFO - codeparrot_training - Step 19416: {'lr': 0.0003544271257119028, 'samples': 9941504, 'steps': 19416, 'loss/train': 2.683513879776001} 02/25/2022 02:43:30 - INFO - codeparrot_training - Step 19417: {'lr': 0.00035441225886484066, 'samples': 9942016, 'steps': 19417, 'loss/train': 1.9332672357559204} 02/25/2022 02:43:34 - INFO - codeparrot_training - Step 19418: {'lr': 0.0003543973915705095, 'samples': 9942528, 'steps': 19418, 'loss/train': 1.6198923587799072} 02/25/2022 02:43:39 - INFO - codeparrot_training - Step 19419: {'lr': 0.0003543825238289732, 'samples': 9943040, 'steps': 19419, 'loss/train': 0.8241666555404663} 02/25/2022 02:43:43 - INFO - codeparrot_training - Step 19420: {'lr': 0.0003543676556402952, 'samples': 9943552, 'steps': 19420, 'loss/train': 1.546484112739563} 02/25/2022 02:43:48 - INFO - codeparrot_training - Step 19421: {'lr': 0.00035435278700453934, 'samples': 9944064, 'steps': 19421, 'loss/train': 2.1739501953125} 02/25/2022 02:43:52 - INFO - codeparrot_training - Step 19422: {'lr': 0.0003543379179217693, 'samples': 9944576, 'steps': 19422, 'loss/train': 3.12174129486084} 02/25/2022 02:43:58 - INFO - codeparrot_training - Step 19423: {'lr': 0.0003543230483920487, 'samples': 9945088, 'steps': 19423, 'loss/train': 0.8596010208129883} 02/25/2022 02:44:01 - INFO - codeparrot_training - Step 19424: {'lr': 0.0003543081784154414, 'samples': 9945600, 'steps': 19424, 'loss/train': 2.4726855754852295} 02/25/2022 02:44:07 - INFO - codeparrot_training - Step 19425: {'lr': 0.00035429330799201085, 'samples': 9946112, 'steps': 19425, 'loss/train': 1.9892871379852295} 02/25/2022 02:44:10 - INFO - codeparrot_training - Step 19426: {'lr': 0.00035427843712182097, 'samples': 9946624, 'steps': 19426, 'loss/train': 2.166886806488037} 02/25/2022 02:44:16 - INFO - codeparrot_training - Step 19427: {'lr': 0.0003542635658049353, 'samples': 9947136, 'steps': 19427, 'loss/train': 2.1598165035247803} 02/25/2022 02:44:19 - INFO - codeparrot_training - Step 19428: {'lr': 0.0003542486940414177, 'samples': 9947648, 'steps': 19428, 'loss/train': 0.9542031288146973} 02/25/2022 02:44:25 - INFO - codeparrot_training - Step 19429: {'lr': 0.00035423382183133174, 'samples': 9948160, 'steps': 19429, 'loss/train': 2.6849005222320557} 02/25/2022 02:44:28 - INFO - codeparrot_training - Step 19430: {'lr': 0.00035421894917474125, 'samples': 9948672, 'steps': 19430, 'loss/train': 2.466521739959717} 02/25/2022 02:44:34 - INFO - codeparrot_training - Step 19431: {'lr': 0.0003542040760717098, 'samples': 9949184, 'steps': 19431, 'loss/train': 1.6099371910095215} 02/25/2022 02:44:38 - INFO - codeparrot_training - Step 19432: {'lr': 0.0003541892025223012, 'samples': 9949696, 'steps': 19432, 'loss/train': 0.8487576246261597} 02/25/2022 02:44:43 - INFO - codeparrot_training - Step 19433: {'lr': 0.00035417432852657916, 'samples': 9950208, 'steps': 19433, 'loss/train': 0.7949591279029846} 02/25/2022 02:44:46 - INFO - codeparrot_training - Step 19434: {'lr': 0.00035415945408460737, 'samples': 9950720, 'steps': 19434, 'loss/train': 1.8293412923812866} 02/25/2022 02:44:52 - INFO - codeparrot_training - Step 19435: {'lr': 0.0003541445791964496, 'samples': 9951232, 'steps': 19435, 'loss/train': 1.6136757135391235} 02/25/2022 02:44:55 - INFO - codeparrot_training - Step 19436: {'lr': 0.0003541297038621694, 'samples': 9951744, 'steps': 19436, 'loss/train': 2.080000638961792} 02/25/2022 02:45:02 - INFO - codeparrot_training - Step 19437: {'lr': 0.0003541148280818307, 'samples': 9952256, 'steps': 19437, 'loss/train': 1.9385898113250732} 02/25/2022 02:45:05 - INFO - codeparrot_training - Step 19438: {'lr': 0.00035409995185549717, 'samples': 9952768, 'steps': 19438, 'loss/train': 2.651024103164673} 02/25/2022 02:45:11 - INFO - codeparrot_training - Step 19439: {'lr': 0.00035408507518323244, 'samples': 9953280, 'steps': 19439, 'loss/train': 1.0612624883651733} 02/25/2022 02:45:14 - INFO - codeparrot_training - Step 19440: {'lr': 0.00035407019806510035, 'samples': 9953792, 'steps': 19440, 'loss/train': 2.1233155727386475} 02/25/2022 02:45:20 - INFO - codeparrot_training - Step 19441: {'lr': 0.0003540553205011645, 'samples': 9954304, 'steps': 19441, 'loss/train': 2.258026123046875} 02/25/2022 02:45:23 - INFO - codeparrot_training - Step 19442: {'lr': 0.00035404044249148873, 'samples': 9954816, 'steps': 19442, 'loss/train': 2.034329652786255} 02/25/2022 02:45:29 - INFO - codeparrot_training - Step 19443: {'lr': 0.0003540255640361368, 'samples': 9955328, 'steps': 19443, 'loss/train': 0.7765913009643555} 02/25/2022 02:45:32 - INFO - codeparrot_training - Step 19444: {'lr': 0.0003540106851351723, 'samples': 9955840, 'steps': 19444, 'loss/train': 1.3520792722702026} 02/25/2022 02:45:38 - INFO - codeparrot_training - Step 19445: {'lr': 0.00035399580578865907, 'samples': 9956352, 'steps': 19445, 'loss/train': 0.47456982731819153} 02/25/2022 02:45:41 - INFO - codeparrot_training - Step 19446: {'lr': 0.00035398092599666086, 'samples': 9956864, 'steps': 19446, 'loss/train': 2.323485851287842} 02/25/2022 02:45:48 - INFO - codeparrot_training - Step 19447: {'lr': 0.00035396604575924133, 'samples': 9957376, 'steps': 19447, 'loss/train': 2.045677661895752} 02/25/2022 02:45:51 - INFO - codeparrot_training - Step 19448: {'lr': 0.00035395116507646435, 'samples': 9957888, 'steps': 19448, 'loss/train': 1.7655506134033203} 02/25/2022 02:45:57 - INFO - codeparrot_training - Step 19449: {'lr': 0.00035393628394839356, 'samples': 9958400, 'steps': 19449, 'loss/train': 3.113555431365967} 02/25/2022 02:46:00 - INFO - codeparrot_training - Step 19450: {'lr': 0.00035392140237509276, 'samples': 9958912, 'steps': 19450, 'loss/train': 1.6115671396255493} 02/25/2022 02:46:06 - INFO - codeparrot_training - Step 19451: {'lr': 0.0003539065203566256, 'samples': 9959424, 'steps': 19451, 'loss/train': 2.853212356567383} 02/25/2022 02:46:09 - INFO - codeparrot_training - Step 19452: {'lr': 0.0003538916378930559, 'samples': 9959936, 'steps': 19452, 'loss/train': 2.6921048164367676} 02/25/2022 02:46:15 - INFO - codeparrot_training - Step 19453: {'lr': 0.0003538767549844475, 'samples': 9960448, 'steps': 19453, 'loss/train': 2.059866428375244} 02/25/2022 02:46:18 - INFO - codeparrot_training - Step 19454: {'lr': 0.000353861871630864, 'samples': 9960960, 'steps': 19454, 'loss/train': 3.673832893371582} 02/25/2022 02:46:24 - INFO - codeparrot_training - Step 19455: {'lr': 0.00035384698783236923, 'samples': 9961472, 'steps': 19455, 'loss/train': 1.805040955543518} 02/25/2022 02:46:27 - INFO - codeparrot_training - Step 19456: {'lr': 0.00035383210358902695, 'samples': 9961984, 'steps': 19456, 'loss/train': 1.6846179962158203} 02/25/2022 02:46:34 - INFO - codeparrot_training - Step 19457: {'lr': 0.0003538172189009009, 'samples': 9962496, 'steps': 19457, 'loss/train': 2.2305312156677246} 02/25/2022 02:46:37 - INFO - codeparrot_training - Step 19458: {'lr': 0.00035380233376805487, 'samples': 9963008, 'steps': 19458, 'loss/train': 1.8792197704315186} 02/25/2022 02:46:43 - INFO - codeparrot_training - Step 19459: {'lr': 0.00035378744819055264, 'samples': 9963520, 'steps': 19459, 'loss/train': 1.164211630821228} 02/25/2022 02:46:46 - INFO - codeparrot_training - Step 19460: {'lr': 0.0003537725621684578, 'samples': 9964032, 'steps': 19460, 'loss/train': 2.3472208976745605} 02/25/2022 02:46:52 - INFO - codeparrot_training - Step 19461: {'lr': 0.0003537576757018344, 'samples': 9964544, 'steps': 19461, 'loss/train': 2.610691785812378} 02/25/2022 02:46:55 - INFO - codeparrot_training - Step 19462: {'lr': 0.0003537427887907459, 'samples': 9965056, 'steps': 19462, 'loss/train': 2.193692445755005} 02/25/2022 02:47:01 - INFO - codeparrot_training - Step 19463: {'lr': 0.0003537279014352565, 'samples': 9965568, 'steps': 19463, 'loss/train': 1.923625111579895} 02/25/2022 02:47:04 - INFO - codeparrot_training - Step 19464: {'lr': 0.00035371301363542945, 'samples': 9966080, 'steps': 19464, 'loss/train': 2.0617213249206543} 02/25/2022 02:47:10 - INFO - codeparrot_training - Step 19465: {'lr': 0.00035369812539132894, 'samples': 9966592, 'steps': 19465, 'loss/train': 1.2559000253677368} 02/25/2022 02:47:13 - INFO - codeparrot_training - Step 19466: {'lr': 0.0003536832367030185, 'samples': 9967104, 'steps': 19466, 'loss/train': 1.83268141746521} 02/25/2022 02:47:20 - INFO - codeparrot_training - Step 19467: {'lr': 0.00035366834757056203, 'samples': 9967616, 'steps': 19467, 'loss/train': 1.2091857194900513} 02/25/2022 02:47:24 - INFO - codeparrot_training - Step 19468: {'lr': 0.0003536534579940233, 'samples': 9968128, 'steps': 19468, 'loss/train': 2.146350383758545} 02/25/2022 02:47:29 - INFO - codeparrot_training - Step 19469: {'lr': 0.0003536385679734659, 'samples': 9968640, 'steps': 19469, 'loss/train': 2.4267187118530273} 02/25/2022 02:47:33 - INFO - codeparrot_training - Step 19470: {'lr': 0.000353623677508954, 'samples': 9969152, 'steps': 19470, 'loss/train': 2.47650146484375} 02/25/2022 02:47:38 - INFO - codeparrot_training - Step 19471: {'lr': 0.00035360878660055107, 'samples': 9969664, 'steps': 19471, 'loss/train': 2.0601797103881836} 02/25/2022 02:47:42 - INFO - codeparrot_training - Step 19472: {'lr': 0.0003535938952483211, 'samples': 9970176, 'steps': 19472, 'loss/train': 2.7199342250823975} 02/25/2022 02:47:47 - INFO - codeparrot_training - Step 19473: {'lr': 0.00035357900345232764, 'samples': 9970688, 'steps': 19473, 'loss/train': 1.1251051425933838} 02/25/2022 02:47:51 - INFO - codeparrot_training - Step 19474: {'lr': 0.0003535641112126347, 'samples': 9971200, 'steps': 19474, 'loss/train': 1.7359498739242554} 02/25/2022 02:47:56 - INFO - codeparrot_training - Step 19475: {'lr': 0.00035354921852930596, 'samples': 9971712, 'steps': 19475, 'loss/train': 1.335972785949707} 02/25/2022 02:48:00 - INFO - codeparrot_training - Step 19476: {'lr': 0.0003535343254024053, 'samples': 9972224, 'steps': 19476, 'loss/train': 0.9973539113998413} 02/25/2022 02:48:06 - INFO - codeparrot_training - Step 19477: {'lr': 0.00035351943183199643, 'samples': 9972736, 'steps': 19477, 'loss/train': 2.1096160411834717} 02/25/2022 02:48:09 - INFO - codeparrot_training - Step 19478: {'lr': 0.0003535045378181432, 'samples': 9973248, 'steps': 19478, 'loss/train': 2.2010080814361572} 02/25/2022 02:48:15 - INFO - codeparrot_training - Step 19479: {'lr': 0.0003534896433609093, 'samples': 9973760, 'steps': 19479, 'loss/train': 2.0903377532958984} 02/25/2022 02:48:18 - INFO - codeparrot_training - Step 19480: {'lr': 0.0003534747484603587, 'samples': 9974272, 'steps': 19480, 'loss/train': 0.5457518696784973} 02/25/2022 02:48:24 - INFO - codeparrot_training - Step 19481: {'lr': 0.00035345985311655513, 'samples': 9974784, 'steps': 19481, 'loss/train': 2.3291385173797607} 02/25/2022 02:48:27 - INFO - codeparrot_training - Step 19482: {'lr': 0.0003534449573295624, 'samples': 9975296, 'steps': 19482, 'loss/train': 1.7085214853286743} 02/25/2022 02:48:34 - INFO - codeparrot_training - Step 19483: {'lr': 0.0003534300610994444, 'samples': 9975808, 'steps': 19483, 'loss/train': 0.7055985331535339} 02/25/2022 02:48:38 - INFO - codeparrot_training - Step 19484: {'lr': 0.00035341516442626475, 'samples': 9976320, 'steps': 19484, 'loss/train': 2.070014238357544} 02/25/2022 02:48:44 - INFO - codeparrot_training - Step 19485: {'lr': 0.00035340026731008745, 'samples': 9976832, 'steps': 19485, 'loss/train': 1.5858851671218872} 02/25/2022 02:48:47 - INFO - codeparrot_training - Step 19486: {'lr': 0.0003533853697509762, 'samples': 9977344, 'steps': 19486, 'loss/train': 1.0990734100341797} 02/25/2022 02:48:52 - INFO - codeparrot_training - Step 19487: {'lr': 0.0003533704717489949, 'samples': 9977856, 'steps': 19487, 'loss/train': 1.49764084815979} 02/25/2022 02:48:56 - INFO - codeparrot_training - Step 19488: {'lr': 0.00035335557330420724, 'samples': 9978368, 'steps': 19488, 'loss/train': 1.5058784484863281} 02/25/2022 02:49:01 - INFO - codeparrot_training - Step 19489: {'lr': 0.0003533406744166772, 'samples': 9978880, 'steps': 19489, 'loss/train': 1.4632439613342285} 02/25/2022 02:49:05 - INFO - codeparrot_training - Step 19490: {'lr': 0.00035332577508646846, 'samples': 9979392, 'steps': 19490, 'loss/train': 2.2159199714660645} 02/25/2022 02:49:10 - INFO - codeparrot_training - Step 19491: {'lr': 0.000353310875313645, 'samples': 9979904, 'steps': 19491, 'loss/train': 1.598966121673584} 02/25/2022 02:49:14 - INFO - codeparrot_training - Step 19492: {'lr': 0.00035329597509827046, 'samples': 9980416, 'steps': 19492, 'loss/train': 1.0308691263198853} 02/25/2022 02:49:21 - INFO - codeparrot_training - Step 19493: {'lr': 0.00035328107444040875, 'samples': 9980928, 'steps': 19493, 'loss/train': 2.353309392929077} 02/25/2022 02:49:27 - INFO - codeparrot_training - Step 19494: {'lr': 0.0003532661733401238, 'samples': 9981440, 'steps': 19494, 'loss/train': 0.5372552275657654} 02/25/2022 02:49:30 - INFO - codeparrot_training - Step 19495: {'lr': 0.0003532512717974793, 'samples': 9981952, 'steps': 19495, 'loss/train': 2.1641781330108643} 02/25/2022 02:49:36 - INFO - codeparrot_training - Step 19496: {'lr': 0.00035323636981253914, 'samples': 9982464, 'steps': 19496, 'loss/train': 0.7208479046821594} 02/25/2022 02:49:39 - INFO - codeparrot_training - Step 19497: {'lr': 0.00035322146738536714, 'samples': 9982976, 'steps': 19497, 'loss/train': 1.4075491428375244} 02/25/2022 02:49:45 - INFO - codeparrot_training - Step 19498: {'lr': 0.00035320656451602725, 'samples': 9983488, 'steps': 19498, 'loss/train': 2.5539238452911377} 02/25/2022 02:49:48 - INFO - codeparrot_training - Step 19499: {'lr': 0.00035319166120458315, 'samples': 9984000, 'steps': 19499, 'loss/train': 1.254536509513855} 02/25/2022 02:49:54 - INFO - codeparrot_training - Step 19500: {'lr': 0.00035317675745109866, 'samples': 9984512, 'steps': 19500, 'loss/train': 3.8736679553985596} 02/25/2022 02:49:57 - INFO - codeparrot_training - Step 19501: {'lr': 0.0003531618532556378, 'samples': 9985024, 'steps': 19501, 'loss/train': 2.3909385204315186} 02/25/2022 02:50:03 - INFO - codeparrot_training - Step 19502: {'lr': 0.00035314694861826427, 'samples': 9985536, 'steps': 19502, 'loss/train': 2.1281533241271973} 02/25/2022 02:50:07 - INFO - codeparrot_training - Step 19503: {'lr': 0.00035313204353904203, 'samples': 9986048, 'steps': 19503, 'loss/train': 1.9034149646759033} 02/25/2022 02:50:10 - INFO - codeparrot_training - Step 19504: {'lr': 0.0003531171380180348, 'samples': 9986560, 'steps': 19504, 'loss/train': 3.7592263221740723} 02/25/2022 02:50:18 - INFO - codeparrot_training - Step 19505: {'lr': 0.00035310223205530655, 'samples': 9987072, 'steps': 19505, 'loss/train': 2.2944185733795166} 02/25/2022 02:50:21 - INFO - codeparrot_training - Step 19506: {'lr': 0.00035308732565092097, 'samples': 9987584, 'steps': 19506, 'loss/train': 1.685953974723816} 02/25/2022 02:50:27 - INFO - codeparrot_training - Step 19507: {'lr': 0.0003530724188049421, 'samples': 9988096, 'steps': 19507, 'loss/train': 2.1296699047088623} 02/25/2022 02:50:30 - INFO - codeparrot_training - Step 19508: {'lr': 0.0003530575115174337, 'samples': 9988608, 'steps': 19508, 'loss/train': 1.947312593460083} 02/25/2022 02:50:36 - INFO - codeparrot_training - Step 19509: {'lr': 0.0003530426037884597, 'samples': 9989120, 'steps': 19509, 'loss/train': 2.0738906860351562} 02/25/2022 02:50:39 - INFO - codeparrot_training - Step 19510: {'lr': 0.00035302769561808387, 'samples': 9989632, 'steps': 19510, 'loss/train': 1.4905050992965698} 02/25/2022 02:50:45 - INFO - codeparrot_training - Step 19511: {'lr': 0.00035301278700637005, 'samples': 9990144, 'steps': 19511, 'loss/train': 1.7383122444152832} 02/25/2022 02:50:48 - INFO - codeparrot_training - Step 19512: {'lr': 0.0003529978779533822, 'samples': 9990656, 'steps': 19512, 'loss/train': 2.7670764923095703} 02/25/2022 02:50:54 - INFO - codeparrot_training - Step 19513: {'lr': 0.0003529829684591842, 'samples': 9991168, 'steps': 19513, 'loss/train': 1.805279016494751} 02/25/2022 02:50:57 - INFO - codeparrot_training - Step 19514: {'lr': 0.0003529680585238398, 'samples': 9991680, 'steps': 19514, 'loss/train': 2.214223861694336} 02/25/2022 02:51:04 - INFO - codeparrot_training - Step 19515: {'lr': 0.000352953148147413, 'samples': 9992192, 'steps': 19515, 'loss/train': 2.2694575786590576} 02/25/2022 02:51:08 - INFO - codeparrot_training - Step 19516: {'lr': 0.0003529382373299675, 'samples': 9992704, 'steps': 19516, 'loss/train': 2.026510238647461} 02/25/2022 02:51:13 - INFO - codeparrot_training - Step 19517: {'lr': 0.0003529233260715673, 'samples': 9993216, 'steps': 19517, 'loss/train': 2.6090493202209473} 02/25/2022 02:51:17 - INFO - codeparrot_training - Step 19518: {'lr': 0.00035290841437227627, 'samples': 9993728, 'steps': 19518, 'loss/train': 1.5069223642349243} 02/25/2022 02:51:23 - INFO - codeparrot_training - Step 19519: {'lr': 0.00035289350223215827, 'samples': 9994240, 'steps': 19519, 'loss/train': 3.166726589202881} 02/25/2022 02:51:26 - INFO - codeparrot_training - Step 19520: {'lr': 0.00035287858965127723, 'samples': 9994752, 'steps': 19520, 'loss/train': 0.9958294034004211} 02/25/2022 02:51:32 - INFO - codeparrot_training - Step 19521: {'lr': 0.0003528636766296969, 'samples': 9995264, 'steps': 19521, 'loss/train': 1.962761402130127} 02/25/2022 02:51:35 - INFO - codeparrot_training - Step 19522: {'lr': 0.0003528487631674813, 'samples': 9995776, 'steps': 19522, 'loss/train': 1.7990299463272095} 02/25/2022 02:51:41 - INFO - codeparrot_training - Step 19523: {'lr': 0.00035283384926469426, 'samples': 9996288, 'steps': 19523, 'loss/train': 1.3673853874206543} 02/25/2022 02:51:44 - INFO - codeparrot_training - Step 19524: {'lr': 0.00035281893492139965, 'samples': 9996800, 'steps': 19524, 'loss/train': 2.6090004444122314} 02/25/2022 02:51:52 - INFO - codeparrot_training - Step 19525: {'lr': 0.0003528040201376613, 'samples': 9997312, 'steps': 19525, 'loss/train': 2.215991973876953} 02/25/2022 02:51:55 - INFO - codeparrot_training - Step 19526: {'lr': 0.00035278910491354335, 'samples': 9997824, 'steps': 19526, 'loss/train': 2.9071054458618164} 02/25/2022 02:52:01 - INFO - codeparrot_training - Step 19527: {'lr': 0.0003527741892491093, 'samples': 9998336, 'steps': 19527, 'loss/train': 0.8466321229934692} 02/25/2022 02:52:04 - INFO - codeparrot_training - Step 19528: {'lr': 0.00035275927314442344, 'samples': 9998848, 'steps': 19528, 'loss/train': 1.6073174476623535} 02/25/2022 02:52:10 - INFO - codeparrot_training - Step 19529: {'lr': 0.00035274435659954936, 'samples': 9999360, 'steps': 19529, 'loss/train': 2.08135986328125} 02/25/2022 02:52:13 - INFO - codeparrot_training - Step 19530: {'lr': 0.00035272943961455106, 'samples': 9999872, 'steps': 19530, 'loss/train': 1.9042640924453735} 02/25/2022 02:52:19 - INFO - codeparrot_training - Step 19531: {'lr': 0.00035271452218949256, 'samples': 10000384, 'steps': 19531, 'loss/train': 0.2562394440174103} 02/25/2022 02:52:22 - INFO - codeparrot_training - Step 19532: {'lr': 0.0003526996043244376, 'samples': 10000896, 'steps': 19532, 'loss/train': 2.480272054672241} 02/25/2022 02:52:28 - INFO - codeparrot_training - Step 19533: {'lr': 0.0003526846860194502, 'samples': 10001408, 'steps': 19533, 'loss/train': 3.416160821914673} 02/25/2022 02:52:31 - INFO - codeparrot_training - Step 19534: {'lr': 0.0003526697672745942, 'samples': 10001920, 'steps': 19534, 'loss/train': 1.7121726274490356} 02/25/2022 02:52:37 - INFO - codeparrot_training - Step 19535: {'lr': 0.0003526548480899335, 'samples': 10002432, 'steps': 19535, 'loss/train': 0.3300989866256714} 02/25/2022 02:52:40 - INFO - codeparrot_training - Step 19536: {'lr': 0.000352639928465532, 'samples': 10002944, 'steps': 19536, 'loss/train': 2.5870585441589355} 02/25/2022 02:52:46 - INFO - codeparrot_training - Step 19537: {'lr': 0.00035262500840145365, 'samples': 10003456, 'steps': 19537, 'loss/train': 1.162061095237732} 02/25/2022 02:52:49 - INFO - codeparrot_training - Step 19538: {'lr': 0.00035261008789776237, 'samples': 10003968, 'steps': 19538, 'loss/train': 2.3007102012634277} 02/25/2022 02:52:55 - INFO - codeparrot_training - Step 19539: {'lr': 0.0003525951669545221, 'samples': 10004480, 'steps': 19539, 'loss/train': 1.0688713788986206} 02/25/2022 02:52:58 - INFO - codeparrot_training - Step 19540: {'lr': 0.00035258024557179664, 'samples': 10004992, 'steps': 19540, 'loss/train': 5.207286834716797} 02/25/2022 02:53:06 - INFO - codeparrot_training - Step 19541: {'lr': 0.00035256532374965, 'samples': 10005504, 'steps': 19541, 'loss/train': 2.109147310256958} 02/25/2022 02:53:09 - INFO - codeparrot_training - Step 19542: {'lr': 0.0003525504014881461, 'samples': 10006016, 'steps': 19542, 'loss/train': 1.3966560363769531} 02/25/2022 02:53:15 - INFO - codeparrot_training - Step 19543: {'lr': 0.0003525354787873488, 'samples': 10006528, 'steps': 19543, 'loss/train': 2.1022613048553467} 02/25/2022 02:53:18 - INFO - codeparrot_training - Step 19544: {'lr': 0.0003525205556473221, 'samples': 10007040, 'steps': 19544, 'loss/train': 3.664903402328491} 02/25/2022 02:53:24 - INFO - codeparrot_training - Step 19545: {'lr': 0.0003525056320681299, 'samples': 10007552, 'steps': 19545, 'loss/train': 2.0755460262298584} 02/25/2022 02:53:27 - INFO - codeparrot_training - Step 19546: {'lr': 0.00035249070804983616, 'samples': 10008064, 'steps': 19546, 'loss/train': 0.9586907625198364} 02/25/2022 02:53:33 - INFO - codeparrot_training - Step 19547: {'lr': 0.00035247578359250473, 'samples': 10008576, 'steps': 19547, 'loss/train': 1.0740753412246704} 02/25/2022 02:53:36 - INFO - codeparrot_training - Step 19548: {'lr': 0.0003524608586961996, 'samples': 10009088, 'steps': 19548, 'loss/train': 1.3718539476394653} 02/25/2022 02:53:42 - INFO - codeparrot_training - Step 19549: {'lr': 0.00035244593336098464, 'samples': 10009600, 'steps': 19549, 'loss/train': 1.3043261766433716} 02/25/2022 02:53:45 - INFO - codeparrot_training - Step 19550: {'lr': 0.0003524310075869239, 'samples': 10010112, 'steps': 19550, 'loss/train': 1.4652314186096191} 02/25/2022 02:53:53 - INFO - codeparrot_training - Step 19551: {'lr': 0.0003524160813740812, 'samples': 10010624, 'steps': 19551, 'loss/train': 2.622642993927002} 02/25/2022 02:53:56 - INFO - codeparrot_training - Step 19552: {'lr': 0.00035240115472252056, 'samples': 10011136, 'steps': 19552, 'loss/train': 1.0312288999557495} 02/25/2022 02:54:02 - INFO - codeparrot_training - Step 19553: {'lr': 0.00035238622763230574, 'samples': 10011648, 'steps': 19553, 'loss/train': 0.9093035459518433} 02/25/2022 02:54:05 - INFO - codeparrot_training - Step 19554: {'lr': 0.000352371300103501, 'samples': 10012160, 'steps': 19554, 'loss/train': 2.064634084701538} 02/25/2022 02:54:11 - INFO - codeparrot_training - Step 19555: {'lr': 0.0003523563721361701, 'samples': 10012672, 'steps': 19555, 'loss/train': 1.5709894895553589} 02/25/2022 02:54:14 - INFO - codeparrot_training - Step 19556: {'lr': 0.0003523414437303769, 'samples': 10013184, 'steps': 19556, 'loss/train': 2.32289719581604} 02/25/2022 02:54:20 - INFO - codeparrot_training - Step 19557: {'lr': 0.00035232651488618564, 'samples': 10013696, 'steps': 19557, 'loss/train': 0.8737673163414001} 02/25/2022 02:54:23 - INFO - codeparrot_training - Step 19558: {'lr': 0.00035231158560365987, 'samples': 10014208, 'steps': 19558, 'loss/train': 2.232379913330078} 02/25/2022 02:54:29 - INFO - codeparrot_training - Step 19559: {'lr': 0.0003522966558828639, 'samples': 10014720, 'steps': 19559, 'loss/train': 2.5879552364349365} 02/25/2022 02:54:32 - INFO - codeparrot_training - Step 19560: {'lr': 0.00035228172572386146, 'samples': 10015232, 'steps': 19560, 'loss/train': 0.694169282913208} 02/25/2022 02:54:40 - INFO - codeparrot_training - Step 19561: {'lr': 0.00035226679512671664, 'samples': 10015744, 'steps': 19561, 'loss/train': 1.174360990524292} 02/25/2022 02:54:43 - INFO - codeparrot_training - Step 19562: {'lr': 0.00035225186409149333, 'samples': 10016256, 'steps': 19562, 'loss/train': 3.9466090202331543} 02/25/2022 02:54:49 - INFO - codeparrot_training - Step 19563: {'lr': 0.00035223693261825554, 'samples': 10016768, 'steps': 19563, 'loss/train': 1.5544453859329224} 02/25/2022 02:54:52 - INFO - codeparrot_training - Step 19564: {'lr': 0.0003522220007070671, 'samples': 10017280, 'steps': 19564, 'loss/train': 2.1561708450317383} 02/25/2022 02:54:58 - INFO - codeparrot_training - Step 19565: {'lr': 0.0003522070683579921, 'samples': 10017792, 'steps': 19565, 'loss/train': 1.282261848449707} 02/25/2022 02:55:01 - INFO - codeparrot_training - Step 19566: {'lr': 0.00035219213557109453, 'samples': 10018304, 'steps': 19566, 'loss/train': 2.3898255825042725} 02/25/2022 02:55:07 - INFO - codeparrot_training - Step 19567: {'lr': 0.00035217720234643823, 'samples': 10018816, 'steps': 19567, 'loss/train': 2.159210205078125} 02/25/2022 02:55:10 - INFO - codeparrot_training - Step 19568: {'lr': 0.0003521622686840873, 'samples': 10019328, 'steps': 19568, 'loss/train': 1.8321435451507568} 02/25/2022 02:55:16 - INFO - codeparrot_training - Step 19569: {'lr': 0.00035214733458410557, 'samples': 10019840, 'steps': 19569, 'loss/train': 0.6593620181083679} 02/25/2022 02:55:19 - INFO - codeparrot_training - Step 19570: {'lr': 0.00035213240004655714, 'samples': 10020352, 'steps': 19570, 'loss/train': 1.808738350868225} 02/25/2022 02:55:25 - INFO - codeparrot_training - Step 19571: {'lr': 0.000352117465071506, 'samples': 10020864, 'steps': 19571, 'loss/train': 0.38315537571907043} 02/25/2022 02:55:28 - INFO - codeparrot_training - Step 19572: {'lr': 0.000352102529659016, 'samples': 10021376, 'steps': 19572, 'loss/train': 1.3931598663330078} 02/25/2022 02:55:34 - INFO - codeparrot_training - Step 19573: {'lr': 0.00035208759380915116, 'samples': 10021888, 'steps': 19573, 'loss/train': 2.9910523891448975} 02/25/2022 02:55:37 - INFO - codeparrot_training - Step 19574: {'lr': 0.0003520726575219756, 'samples': 10022400, 'steps': 19574, 'loss/train': 1.307809591293335} 02/25/2022 02:55:43 - INFO - codeparrot_training - Step 19575: {'lr': 0.00035205772079755307, 'samples': 10022912, 'steps': 19575, 'loss/train': 2.1897032260894775} 02/25/2022 02:55:50 - INFO - codeparrot_training - Step 19576: {'lr': 0.0003520427836359477, 'samples': 10023424, 'steps': 19576, 'loss/train': 1.7167686223983765} 02/25/2022 02:55:54 - INFO - codeparrot_training - Step 19577: {'lr': 0.00035202784603722344, 'samples': 10023936, 'steps': 19577, 'loss/train': 2.5758914947509766} 02/25/2022 02:55:59 - INFO - codeparrot_training - Step 19578: {'lr': 0.00035201290800144423, 'samples': 10024448, 'steps': 19578, 'loss/train': 2.058234691619873} 02/25/2022 02:56:03 - INFO - codeparrot_training - Step 19579: {'lr': 0.00035199796952867425, 'samples': 10024960, 'steps': 19579, 'loss/train': 2.095686674118042} 02/25/2022 02:56:08 - INFO - codeparrot_training - Step 19580: {'lr': 0.0003519830306189773, 'samples': 10025472, 'steps': 19580, 'loss/train': 2.260281562805176} 02/25/2022 02:56:12 - INFO - codeparrot_training - Step 19581: {'lr': 0.0003519680912724174, 'samples': 10025984, 'steps': 19581, 'loss/train': 1.3114290237426758} 02/25/2022 02:56:17 - INFO - codeparrot_training - Step 19582: {'lr': 0.0003519531514890586, 'samples': 10026496, 'steps': 19582, 'loss/train': 1.6117550134658813} 02/25/2022 02:56:21 - INFO - codeparrot_training - Step 19583: {'lr': 0.00035193821126896493, 'samples': 10027008, 'steps': 19583, 'loss/train': 2.0356404781341553} 02/25/2022 02:56:26 - INFO - codeparrot_training - Step 19584: {'lr': 0.0003519232706122002, 'samples': 10027520, 'steps': 19584, 'loss/train': 1.8880693912506104} 02/25/2022 02:56:30 - INFO - codeparrot_training - Step 19585: {'lr': 0.0003519083295188287, 'samples': 10028032, 'steps': 19585, 'loss/train': 0.7321484088897705} 02/25/2022 02:56:38 - INFO - codeparrot_training - Step 19586: {'lr': 0.0003518933879889142, 'samples': 10028544, 'steps': 19586, 'loss/train': 1.8996468782424927} 02/25/2022 02:56:41 - INFO - codeparrot_training - Step 19587: {'lr': 0.00035187844602252075, 'samples': 10029056, 'steps': 19587, 'loss/train': 1.8454076051712036} 02/25/2022 02:56:47 - INFO - codeparrot_training - Step 19588: {'lr': 0.0003518635036197124, 'samples': 10029568, 'steps': 19588, 'loss/train': 1.3699305057525635} 02/25/2022 02:56:50 - INFO - codeparrot_training - Step 19589: {'lr': 0.00035184856078055316, 'samples': 10030080, 'steps': 19589, 'loss/train': 3.159069299697876} 02/25/2022 02:56:56 - INFO - codeparrot_training - Step 19590: {'lr': 0.000351833617505107, 'samples': 10030592, 'steps': 19590, 'loss/train': 1.8408854007720947} 02/25/2022 02:56:59 - INFO - codeparrot_training - Step 19591: {'lr': 0.000351818673793438, 'samples': 10031104, 'steps': 19591, 'loss/train': 1.092026948928833} 02/25/2022 02:57:05 - INFO - codeparrot_training - Step 19592: {'lr': 0.00035180372964561013, 'samples': 10031616, 'steps': 19592, 'loss/train': 2.6419565677642822} 02/25/2022 02:57:08 - INFO - codeparrot_training - Step 19593: {'lr': 0.00035178878506168733, 'samples': 10032128, 'steps': 19593, 'loss/train': 1.9302358627319336} 02/25/2022 02:57:13 - INFO - codeparrot_training - Step 19594: {'lr': 0.0003517738400417338, 'samples': 10032640, 'steps': 19594, 'loss/train': 4.34604024887085} 02/25/2022 02:57:17 - INFO - codeparrot_training - Step 19595: {'lr': 0.0003517588945858134, 'samples': 10033152, 'steps': 19595, 'loss/train': 2.429274559020996} 02/25/2022 02:57:25 - INFO - codeparrot_training - Step 19596: {'lr': 0.00035174394869399024, 'samples': 10033664, 'steps': 19596, 'loss/train': 2.749152421951294} 02/25/2022 02:57:28 - INFO - codeparrot_training - Step 19597: {'lr': 0.0003517290023663283, 'samples': 10034176, 'steps': 19597, 'loss/train': 2.5418922901153564} 02/25/2022 02:57:34 - INFO - codeparrot_training - Step 19598: {'lr': 0.0003517140556028916, 'samples': 10034688, 'steps': 19598, 'loss/train': 2.3519039154052734} 02/25/2022 02:57:37 - INFO - codeparrot_training - Step 19599: {'lr': 0.0003516991084037442, 'samples': 10035200, 'steps': 19599, 'loss/train': 1.9998564720153809} 02/25/2022 02:57:43 - INFO - codeparrot_training - Step 19600: {'lr': 0.0003516841607689501, 'samples': 10035712, 'steps': 19600, 'loss/train': 1.041944980621338} 02/25/2022 02:57:46 - INFO - codeparrot_training - Step 19601: {'lr': 0.0003516692126985733, 'samples': 10036224, 'steps': 19601, 'loss/train': 2.174870729446411} 02/25/2022 02:57:52 - INFO - codeparrot_training - Step 19602: {'lr': 0.00035165426419267795, 'samples': 10036736, 'steps': 19602, 'loss/train': 2.2092790603637695} 02/25/2022 02:57:55 - INFO - codeparrot_training - Step 19603: {'lr': 0.00035163931525132797, 'samples': 10037248, 'steps': 19603, 'loss/train': 1.6803193092346191} 02/25/2022 02:58:01 - INFO - codeparrot_training - Step 19604: {'lr': 0.00035162436587458744, 'samples': 10037760, 'steps': 19604, 'loss/train': 1.847691297531128} 02/25/2022 02:58:04 - INFO - codeparrot_training - Step 19605: {'lr': 0.00035160941606252044, 'samples': 10038272, 'steps': 19605, 'loss/train': 2.574420690536499} 02/25/2022 02:58:12 - INFO - codeparrot_training - Step 19606: {'lr': 0.00035159446581519094, 'samples': 10038784, 'steps': 19606, 'loss/train': 1.9931769371032715} 02/25/2022 02:58:15 - INFO - codeparrot_training - Step 19607: {'lr': 0.000351579515132663, 'samples': 10039296, 'steps': 19607, 'loss/train': 2.1745822429656982} 02/25/2022 02:58:21 - INFO - codeparrot_training - Step 19608: {'lr': 0.0003515645640150006, 'samples': 10039808, 'steps': 19608, 'loss/train': 2.42423939704895} 02/25/2022 02:58:24 - INFO - codeparrot_training - Step 19609: {'lr': 0.000351549612462268, 'samples': 10040320, 'steps': 19609, 'loss/train': 1.997499704360962} 02/25/2022 02:58:30 - INFO - codeparrot_training - Step 19610: {'lr': 0.00035153466047452904, 'samples': 10040832, 'steps': 19610, 'loss/train': 2.232370376586914} 02/25/2022 02:58:33 - INFO - codeparrot_training - Step 19611: {'lr': 0.00035151970805184785, 'samples': 10041344, 'steps': 19611, 'loss/train': 1.0818361043930054} 02/25/2022 02:58:39 - INFO - codeparrot_training - Step 19612: {'lr': 0.00035150475519428844, 'samples': 10041856, 'steps': 19612, 'loss/train': 2.621471405029297} 02/25/2022 02:58:43 - INFO - codeparrot_training - Step 19613: {'lr': 0.00035148980190191485, 'samples': 10042368, 'steps': 19613, 'loss/train': 3.741616725921631} 02/25/2022 02:58:46 - INFO - codeparrot_training - Step 19614: {'lr': 0.00035147484817479126, 'samples': 10042880, 'steps': 19614, 'loss/train': 2.7486398220062256} 02/25/2022 02:58:52 - INFO - codeparrot_training - Step 19615: {'lr': 0.00035145989401298163, 'samples': 10043392, 'steps': 19615, 'loss/train': 2.101335287094116} 02/25/2022 02:58:55 - INFO - codeparrot_training - Step 19616: {'lr': 0.00035144493941655, 'samples': 10043904, 'steps': 19616, 'loss/train': 2.3324625492095947} 02/25/2022 02:59:01 - INFO - codeparrot_training - Step 19617: {'lr': 0.0003514299843855605, 'samples': 10044416, 'steps': 19617, 'loss/train': 0.20641575753688812} 02/25/2022 02:59:04 - INFO - codeparrot_training - Step 19618: {'lr': 0.0003514150289200771, 'samples': 10044928, 'steps': 19618, 'loss/train': 2.2910702228546143} 02/25/2022 02:59:11 - INFO - codeparrot_training - Step 19619: {'lr': 0.00035140007302016395, 'samples': 10045440, 'steps': 19619, 'loss/train': 1.4463056325912476} 02/25/2022 02:59:15 - INFO - codeparrot_training - Step 19620: {'lr': 0.0003513851166858851, 'samples': 10045952, 'steps': 19620, 'loss/train': 2.1278014183044434} 02/25/2022 02:59:21 - INFO - codeparrot_training - Step 19621: {'lr': 0.0003513701599173046, 'samples': 10046464, 'steps': 19621, 'loss/train': 1.4461511373519897} 02/25/2022 02:59:24 - INFO - codeparrot_training - Step 19622: {'lr': 0.0003513552027144865, 'samples': 10046976, 'steps': 19622, 'loss/train': 1.8281314373016357} 02/25/2022 02:59:30 - INFO - codeparrot_training - Step 19623: {'lr': 0.00035134024507749487, 'samples': 10047488, 'steps': 19623, 'loss/train': 1.8796703815460205} 02/25/2022 02:59:34 - INFO - codeparrot_training - Step 19624: {'lr': 0.0003513252870063939, 'samples': 10048000, 'steps': 19624, 'loss/train': 2.665205955505371} 02/25/2022 02:59:39 - INFO - codeparrot_training - Step 19625: {'lr': 0.00035131032850124744, 'samples': 10048512, 'steps': 19625, 'loss/train': 2.2983086109161377} 02/25/2022 02:59:43 - INFO - codeparrot_training - Step 19626: {'lr': 0.0003512953695621198, 'samples': 10049024, 'steps': 19626, 'loss/train': 2.309155225753784} 02/25/2022 02:59:48 - INFO - codeparrot_training - Step 19627: {'lr': 0.00035128041018907496, 'samples': 10049536, 'steps': 19627, 'loss/train': 2.0799875259399414} 02/25/2022 02:59:52 - INFO - codeparrot_training - Step 19628: {'lr': 0.000351265450382177, 'samples': 10050048, 'steps': 19628, 'loss/train': 3.1819663047790527} 02/25/2022 02:59:57 - INFO - codeparrot_training - Step 19629: {'lr': 0.00035125049014148995, 'samples': 10050560, 'steps': 19629, 'loss/train': 2.152209758758545} 02/25/2022 03:00:01 - INFO - codeparrot_training - Step 19630: {'lr': 0.000351235529467078, 'samples': 10051072, 'steps': 19630, 'loss/train': 0.7848385572433472} 02/25/2022 03:00:06 - INFO - codeparrot_training - Step 19631: {'lr': 0.00035122056835900517, 'samples': 10051584, 'steps': 19631, 'loss/train': 2.8216657638549805} 02/25/2022 03:00:10 - INFO - codeparrot_training - Step 19632: {'lr': 0.0003512056068173356, 'samples': 10052096, 'steps': 19632, 'loss/train': 1.5875760316848755} 02/25/2022 03:00:17 - INFO - codeparrot_training - Step 19633: {'lr': 0.00035119064484213333, 'samples': 10052608, 'steps': 19633, 'loss/train': 2.5184178352355957} 02/25/2022 03:00:20 - INFO - codeparrot_training - Step 19634: {'lr': 0.0003511756824334625, 'samples': 10053120, 'steps': 19634, 'loss/train': 1.962084412574768} 02/25/2022 03:00:26 - INFO - codeparrot_training - Step 19635: {'lr': 0.0003511607195913872, 'samples': 10053632, 'steps': 19635, 'loss/train': 2.3503575325012207} 02/25/2022 03:00:29 - INFO - codeparrot_training - Step 19636: {'lr': 0.00035114575631597136, 'samples': 10054144, 'steps': 19636, 'loss/train': 1.3409106731414795} 02/25/2022 03:00:35 - INFO - codeparrot_training - Step 19637: {'lr': 0.0003511307926072793, 'samples': 10054656, 'steps': 19637, 'loss/train': 1.7063028812408447} 02/25/2022 03:00:38 - INFO - codeparrot_training - Step 19638: {'lr': 0.00035111582846537507, 'samples': 10055168, 'steps': 19638, 'loss/train': 2.3895628452301025} 02/25/2022 03:00:45 - INFO - codeparrot_training - Step 19639: {'lr': 0.00035110086389032264, 'samples': 10055680, 'steps': 19639, 'loss/train': 2.257286548614502} 02/25/2022 03:00:49 - INFO - codeparrot_training - Step 19640: {'lr': 0.0003510858988821863, 'samples': 10056192, 'steps': 19640, 'loss/train': 1.6644878387451172} 02/25/2022 03:00:54 - INFO - codeparrot_training - Step 19641: {'lr': 0.00035107093344103, 'samples': 10056704, 'steps': 19641, 'loss/train': 1.656775951385498} 02/25/2022 03:00:58 - INFO - codeparrot_training - Step 19642: {'lr': 0.000351055967566918, 'samples': 10057216, 'steps': 19642, 'loss/train': 0.5723735094070435} 02/25/2022 03:01:03 - INFO - codeparrot_training - Step 19643: {'lr': 0.0003510410012599142, 'samples': 10057728, 'steps': 19643, 'loss/train': 0.9587002396583557} 02/25/2022 03:01:06 - INFO - codeparrot_training - Step 19644: {'lr': 0.00035102603452008297, 'samples': 10058240, 'steps': 19644, 'loss/train': 1.5231226682662964} 02/25/2022 03:01:12 - INFO - codeparrot_training - Step 19645: {'lr': 0.00035101106734748824, 'samples': 10058752, 'steps': 19645, 'loss/train': 1.4624035358428955} 02/25/2022 03:01:16 - INFO - codeparrot_training - Step 19646: {'lr': 0.00035099609974219417, 'samples': 10059264, 'steps': 19646, 'loss/train': 1.7937335968017578} 02/25/2022 03:01:21 - INFO - codeparrot_training - Step 19647: {'lr': 0.00035098113170426484, 'samples': 10059776, 'steps': 19647, 'loss/train': 2.059809446334839} 02/25/2022 03:01:25 - INFO - codeparrot_training - Step 19648: {'lr': 0.0003509661632337645, 'samples': 10060288, 'steps': 19648, 'loss/train': 2.2621843814849854} 02/25/2022 03:01:31 - INFO - codeparrot_training - Step 19649: {'lr': 0.00035095119433075706, 'samples': 10060800, 'steps': 19649, 'loss/train': 4.312873363494873} 02/25/2022 03:01:35 - INFO - codeparrot_training - Step 19650: {'lr': 0.00035093622499530677, 'samples': 10061312, 'steps': 19650, 'loss/train': 4.907009601593018} 02/25/2022 03:01:40 - INFO - codeparrot_training - Step 19651: {'lr': 0.0003509212552274778, 'samples': 10061824, 'steps': 19651, 'loss/train': 2.3816497325897217} 02/25/2022 03:01:44 - INFO - codeparrot_training - Step 19652: {'lr': 0.0003509062850273342, 'samples': 10062336, 'steps': 19652, 'loss/train': 1.1412864923477173} 02/25/2022 03:01:49 - INFO - codeparrot_training - Step 19653: {'lr': 0.00035089131439494013, 'samples': 10062848, 'steps': 19653, 'loss/train': 1.5915791988372803} 02/25/2022 03:01:53 - INFO - codeparrot_training - Step 19654: {'lr': 0.00035087634333035966, 'samples': 10063360, 'steps': 19654, 'loss/train': 1.409009575843811} 02/25/2022 03:01:58 - INFO - codeparrot_training - Step 19655: {'lr': 0.00035086137183365707, 'samples': 10063872, 'steps': 19655, 'loss/train': 2.1974523067474365} 02/25/2022 03:02:02 - INFO - codeparrot_training - Step 19656: {'lr': 0.0003508463999048963, 'samples': 10064384, 'steps': 19656, 'loss/train': 1.8320226669311523} 02/25/2022 03:02:07 - INFO - codeparrot_training - Step 19657: {'lr': 0.0003508314275441416, 'samples': 10064896, 'steps': 19657, 'loss/train': 1.9070558547973633} 02/25/2022 03:02:11 - INFO - codeparrot_training - Step 19658: {'lr': 0.0003508164547514571, 'samples': 10065408, 'steps': 19658, 'loss/train': 1.9889055490493774} 02/25/2022 03:02:17 - INFO - codeparrot_training - Step 19659: {'lr': 0.00035080148152690687, 'samples': 10065920, 'steps': 19659, 'loss/train': 2.0737147331237793} 02/25/2022 03:02:21 - INFO - codeparrot_training - Step 19660: {'lr': 0.0003507865078705551, 'samples': 10066432, 'steps': 19660, 'loss/train': 2.672455310821533} 02/25/2022 03:02:26 - INFO - codeparrot_training - Step 19661: {'lr': 0.00035077153378246603, 'samples': 10066944, 'steps': 19661, 'loss/train': 0.9656722545623779} 02/25/2022 03:02:30 - INFO - codeparrot_training - Step 19662: {'lr': 0.0003507565592627036, 'samples': 10067456, 'steps': 19662, 'loss/train': 2.1578664779663086} 02/25/2022 03:02:35 - INFO - codeparrot_training - Step 19663: {'lr': 0.0003507415843113321, 'samples': 10067968, 'steps': 19663, 'loss/train': 3.3387155532836914} 02/25/2022 03:02:39 - INFO - codeparrot_training - Step 19664: {'lr': 0.00035072660892841566, 'samples': 10068480, 'steps': 19664, 'loss/train': 2.3395957946777344} 02/25/2022 03:02:44 - INFO - codeparrot_training - Step 19665: {'lr': 0.0003507116331140184, 'samples': 10068992, 'steps': 19665, 'loss/train': 0.7336011528968811} 02/25/2022 03:02:48 - INFO - codeparrot_training - Step 19666: {'lr': 0.00035069665686820453, 'samples': 10069504, 'steps': 19666, 'loss/train': 2.6456384658813477} 02/25/2022 03:02:53 - INFO - codeparrot_training - Step 19667: {'lr': 0.0003506816801910381, 'samples': 10070016, 'steps': 19667, 'loss/train': 2.104161024093628} 02/25/2022 03:02:57 - INFO - codeparrot_training - Step 19668: {'lr': 0.00035066670308258333, 'samples': 10070528, 'steps': 19668, 'loss/train': 1.2089543342590332} 02/25/2022 03:03:02 - INFO - codeparrot_training - Step 19669: {'lr': 0.00035065172554290435, 'samples': 10071040, 'steps': 19669, 'loss/train': 1.1023633480072021} 02/25/2022 03:03:06 - INFO - codeparrot_training - Step 19670: {'lr': 0.0003506367475720654, 'samples': 10071552, 'steps': 19670, 'loss/train': 2.4401018619537354} 02/25/2022 03:03:12 - INFO - codeparrot_training - Step 19671: {'lr': 0.0003506217691701305, 'samples': 10072064, 'steps': 19671, 'loss/train': 1.69490385055542} 02/25/2022 03:03:16 - INFO - codeparrot_training - Step 19672: {'lr': 0.000350606790337164, 'samples': 10072576, 'steps': 19672, 'loss/train': 1.5114063024520874} 02/25/2022 03:03:21 - INFO - codeparrot_training - Step 19673: {'lr': 0.00035059181107322977, 'samples': 10073088, 'steps': 19673, 'loss/train': 2.0572352409362793} 02/25/2022 03:03:25 - INFO - codeparrot_training - Step 19674: {'lr': 0.00035057683137839236, 'samples': 10073600, 'steps': 19674, 'loss/train': 1.143776535987854} 02/25/2022 03:03:30 - INFO - codeparrot_training - Step 19675: {'lr': 0.00035056185125271566, 'samples': 10074112, 'steps': 19675, 'loss/train': 2.620931386947632} 02/25/2022 03:03:34 - INFO - codeparrot_training - Step 19676: {'lr': 0.0003505468706962639, 'samples': 10074624, 'steps': 19676, 'loss/train': 3.6897389888763428} 02/25/2022 03:03:39 - INFO - codeparrot_training - Step 19677: {'lr': 0.0003505318897091013, 'samples': 10075136, 'steps': 19677, 'loss/train': 2.638714551925659} 02/25/2022 03:03:45 - INFO - codeparrot_training - Step 19678: {'lr': 0.000350516908291292, 'samples': 10075648, 'steps': 19678, 'loss/train': 1.2624706029891968} 02/25/2022 03:03:48 - INFO - codeparrot_training - Step 19679: {'lr': 0.00035050192644290023, 'samples': 10076160, 'steps': 19679, 'loss/train': 1.9817955493927002} 02/25/2022 03:03:54 - INFO - codeparrot_training - Step 19680: {'lr': 0.00035048694416399005, 'samples': 10076672, 'steps': 19680, 'loss/train': 2.1430068016052246} 02/25/2022 03:03:58 - INFO - codeparrot_training - Step 19681: {'lr': 0.0003504719614546258, 'samples': 10077184, 'steps': 19681, 'loss/train': 2.407341241836548} 02/25/2022 03:04:03 - INFO - codeparrot_training - Step 19682: {'lr': 0.00035045697831487146, 'samples': 10077696, 'steps': 19682, 'loss/train': 1.3414525985717773} 02/25/2022 03:04:07 - INFO - codeparrot_training - Step 19683: {'lr': 0.00035044199474479137, 'samples': 10078208, 'steps': 19683, 'loss/train': 1.7528821229934692} 02/25/2022 03:04:13 - INFO - codeparrot_training - Step 19684: {'lr': 0.00035042701074444965, 'samples': 10078720, 'steps': 19684, 'loss/train': 1.60835599899292} 02/25/2022 03:04:16 - INFO - codeparrot_training - Step 19685: {'lr': 0.00035041202631391056, 'samples': 10079232, 'steps': 19685, 'loss/train': 1.390791893005371} 02/25/2022 03:04:22 - INFO - codeparrot_training - Step 19686: {'lr': 0.0003503970414532382, 'samples': 10079744, 'steps': 19686, 'loss/train': 2.100835084915161} 02/25/2022 03:04:25 - INFO - codeparrot_training - Step 19687: {'lr': 0.00035038205616249674, 'samples': 10080256, 'steps': 19687, 'loss/train': 2.3965134620666504} 02/25/2022 03:04:31 - INFO - codeparrot_training - Step 19688: {'lr': 0.00035036707044175055, 'samples': 10080768, 'steps': 19688, 'loss/train': 1.8352265357971191} 02/25/2022 03:04:34 - INFO - codeparrot_training - Step 19689: {'lr': 0.00035035208429106356, 'samples': 10081280, 'steps': 19689, 'loss/train': 0.6369045376777649} 02/25/2022 03:04:40 - INFO - codeparrot_training - Step 19690: {'lr': 0.0003503370977105002, 'samples': 10081792, 'steps': 19690, 'loss/train': 1.8793842792510986} 02/25/2022 03:04:43 - INFO - codeparrot_training - Step 19691: {'lr': 0.00035032211070012455, 'samples': 10082304, 'steps': 19691, 'loss/train': 0.7447012662887573} 02/25/2022 03:04:49 - INFO - codeparrot_training - Step 19692: {'lr': 0.00035030712326000084, 'samples': 10082816, 'steps': 19692, 'loss/train': 2.0413427352905273} 02/25/2022 03:04:52 - INFO - codeparrot_training - Step 19693: {'lr': 0.00035029213539019324, 'samples': 10083328, 'steps': 19693, 'loss/train': 0.8735166192054749} 02/25/2022 03:04:58 - INFO - codeparrot_training - Step 19694: {'lr': 0.000350277147090766, 'samples': 10083840, 'steps': 19694, 'loss/train': 2.240682363510132} 02/25/2022 03:05:02 - INFO - codeparrot_training - Step 19695: {'lr': 0.0003502621583617833, 'samples': 10084352, 'steps': 19695, 'loss/train': 1.1770226955413818} 02/25/2022 03:05:07 - INFO - codeparrot_training - Step 19696: {'lr': 0.00035024716920330933, 'samples': 10084864, 'steps': 19696, 'loss/train': 0.5625348091125488} 02/25/2022 03:05:11 - INFO - codeparrot_training - Step 19697: {'lr': 0.0003502321796154084, 'samples': 10085376, 'steps': 19697, 'loss/train': 1.4729421138763428} 02/25/2022 03:05:16 - INFO - codeparrot_training - Step 19698: {'lr': 0.00035021718959814453, 'samples': 10085888, 'steps': 19698, 'loss/train': 0.38844922184944153} 02/25/2022 03:05:20 - INFO - codeparrot_training - Step 19699: {'lr': 0.0003502021991515821, 'samples': 10086400, 'steps': 19699, 'loss/train': 2.6359524726867676} 02/25/2022 03:05:25 - INFO - codeparrot_training - Step 19700: {'lr': 0.0003501872082757852, 'samples': 10086912, 'steps': 19700, 'loss/train': 2.1211774349212646} 02/25/2022 03:05:29 - INFO - codeparrot_training - Step 19701: {'lr': 0.00035017221697081826, 'samples': 10087424, 'steps': 19701, 'loss/train': 1.5974019765853882} 02/25/2022 03:05:34 - INFO - codeparrot_training - Step 19702: {'lr': 0.0003501572252367452, 'samples': 10087936, 'steps': 19702, 'loss/train': 2.3705458641052246} 02/25/2022 03:05:38 - INFO - codeparrot_training - Step 19703: {'lr': 0.00035014223307363045, 'samples': 10088448, 'steps': 19703, 'loss/train': 0.5771686434745789} 02/25/2022 03:05:43 - INFO - codeparrot_training - Step 19704: {'lr': 0.0003501272404815382, 'samples': 10088960, 'steps': 19704, 'loss/train': 0.4329005777835846} 02/25/2022 03:05:47 - INFO - codeparrot_training - Step 19705: {'lr': 0.0003501122474605326, 'samples': 10089472, 'steps': 19705, 'loss/train': 0.6630956530570984} 02/25/2022 03:05:53 - INFO - codeparrot_training - Step 19706: {'lr': 0.00035009725401067795, 'samples': 10089984, 'steps': 19706, 'loss/train': 2.5255801677703857} 02/25/2022 03:05:57 - INFO - codeparrot_training - Step 19707: {'lr': 0.00035008226013203845, 'samples': 10090496, 'steps': 19707, 'loss/train': 1.8529889583587646} 02/25/2022 03:06:02 - INFO - codeparrot_training - Step 19708: {'lr': 0.0003500672658246783, 'samples': 10091008, 'steps': 19708, 'loss/train': 2.2475297451019287} 02/25/2022 03:06:06 - INFO - codeparrot_training - Step 19709: {'lr': 0.0003500522710886618, 'samples': 10091520, 'steps': 19709, 'loss/train': 2.3407280445098877} 02/25/2022 03:06:12 - INFO - codeparrot_training - Step 19710: {'lr': 0.0003500372759240531, 'samples': 10092032, 'steps': 19710, 'loss/train': 1.3891665935516357} 02/25/2022 03:06:15 - INFO - codeparrot_training - Step 19711: {'lr': 0.00035002228033091643, 'samples': 10092544, 'steps': 19711, 'loss/train': 1.2385997772216797} 02/25/2022 03:06:21 - INFO - codeparrot_training - Step 19712: {'lr': 0.00035000728430931616, 'samples': 10093056, 'steps': 19712, 'loss/train': 1.5949411392211914} 02/25/2022 03:06:24 - INFO - codeparrot_training - Step 19713: {'lr': 0.00034999228785931644, 'samples': 10093568, 'steps': 19713, 'loss/train': 1.7539089918136597} 02/25/2022 03:06:30 - INFO - codeparrot_training - Step 19714: {'lr': 0.0003499772909809815, 'samples': 10094080, 'steps': 19714, 'loss/train': 0.2511043846607208} 02/25/2022 03:06:33 - INFO - codeparrot_training - Step 19715: {'lr': 0.0003499622936743756, 'samples': 10094592, 'steps': 19715, 'loss/train': 1.0134738683700562} 02/25/2022 03:06:39 - INFO - codeparrot_training - Step 19716: {'lr': 0.0003499472959395629, 'samples': 10095104, 'steps': 19716, 'loss/train': 1.9131131172180176} 02/25/2022 03:06:43 - INFO - codeparrot_training - Step 19717: {'lr': 0.00034993229777660785, 'samples': 10095616, 'steps': 19717, 'loss/train': 2.655975341796875} 02/25/2022 03:06:48 - INFO - codeparrot_training - Step 19718: {'lr': 0.0003499172991855744, 'samples': 10096128, 'steps': 19718, 'loss/train': 2.1876728534698486} 02/25/2022 03:06:52 - INFO - codeparrot_training - Step 19719: {'lr': 0.00034990230016652713, 'samples': 10096640, 'steps': 19719, 'loss/train': 1.8051618337631226} 02/25/2022 03:06:57 - INFO - codeparrot_training - Step 19720: {'lr': 0.00034988730071953, 'samples': 10097152, 'steps': 19720, 'loss/train': 1.1078022718429565} 02/25/2022 03:07:01 - INFO - codeparrot_training - Step 19721: {'lr': 0.0003498723008446475, 'samples': 10097664, 'steps': 19721, 'loss/train': 1.6421740055084229} 02/25/2022 03:07:06 - INFO - codeparrot_training - Step 19722: {'lr': 0.0003498573005419437, 'samples': 10098176, 'steps': 19722, 'loss/train': 1.2230674028396606} 02/25/2022 03:07:09 - INFO - codeparrot_training - Step 19723: {'lr': 0.000349842299811483, 'samples': 10098688, 'steps': 19723, 'loss/train': 1.839769721031189} 02/25/2022 03:07:15 - INFO - codeparrot_training - Step 19724: {'lr': 0.00034982729865332953, 'samples': 10099200, 'steps': 19724, 'loss/train': 2.547917366027832} 02/25/2022 03:07:19 - INFO - codeparrot_training - Step 19725: {'lr': 0.00034981229706754755, 'samples': 10099712, 'steps': 19725, 'loss/train': 3.229694366455078} 02/25/2022 03:07:25 - INFO - codeparrot_training - Step 19726: {'lr': 0.0003497972950542015, 'samples': 10100224, 'steps': 19726, 'loss/train': 0.8456030488014221} 02/25/2022 03:07:28 - INFO - codeparrot_training - Step 19727: {'lr': 0.0003497822926133555, 'samples': 10100736, 'steps': 19727, 'loss/train': 1.3858362436294556} 02/25/2022 03:07:34 - INFO - codeparrot_training - Step 19728: {'lr': 0.0003497672897450738, 'samples': 10101248, 'steps': 19728, 'loss/train': 1.8247514963150024} 02/25/2022 03:07:37 - INFO - codeparrot_training - Step 19729: {'lr': 0.0003497522864494208, 'samples': 10101760, 'steps': 19729, 'loss/train': 1.6876375675201416} 02/25/2022 03:07:43 - INFO - codeparrot_training - Step 19730: {'lr': 0.0003497372827264606, 'samples': 10102272, 'steps': 19730, 'loss/train': 1.3257802724838257} 02/25/2022 03:07:46 - INFO - codeparrot_training - Step 19731: {'lr': 0.0003497222785762576, 'samples': 10102784, 'steps': 19731, 'loss/train': 2.0304720401763916} 02/25/2022 03:07:52 - INFO - codeparrot_training - Step 19732: {'lr': 0.000349707273998876, 'samples': 10103296, 'steps': 19732, 'loss/train': 2.7126681804656982} 02/25/2022 03:07:55 - INFO - codeparrot_training - Step 19733: {'lr': 0.00034969226899438013, 'samples': 10103808, 'steps': 19733, 'loss/train': 1.9379675388336182} 02/25/2022 03:08:01 - INFO - codeparrot_training - Step 19734: {'lr': 0.00034967726356283416, 'samples': 10104320, 'steps': 19734, 'loss/train': 2.1391079425811768} 02/25/2022 03:08:04 - INFO - codeparrot_training - Step 19735: {'lr': 0.00034966225770430244, 'samples': 10104832, 'steps': 19735, 'loss/train': 1.4799776077270508} 02/25/2022 03:08:10 - INFO - codeparrot_training - Step 19736: {'lr': 0.00034964725141884936, 'samples': 10105344, 'steps': 19736, 'loss/train': 2.34786319732666} 02/25/2022 03:08:13 - INFO - codeparrot_training - Step 19737: {'lr': 0.000349632244706539, 'samples': 10105856, 'steps': 19737, 'loss/train': 1.2513939142227173} 02/25/2022 03:08:19 - INFO - codeparrot_training - Step 19738: {'lr': 0.0003496172375674358, 'samples': 10106368, 'steps': 19738, 'loss/train': 2.008321762084961} 02/25/2022 03:08:22 - INFO - codeparrot_training - Step 19739: {'lr': 0.0003496022300016039, 'samples': 10106880, 'steps': 19739, 'loss/train': 2.2683513164520264} 02/25/2022 03:08:28 - INFO - codeparrot_training - Step 19740: {'lr': 0.00034958722200910777, 'samples': 10107392, 'steps': 19740, 'loss/train': 1.4257746934890747} 02/25/2022 03:08:32 - INFO - codeparrot_training - Step 19741: {'lr': 0.00034957221359001154, 'samples': 10107904, 'steps': 19741, 'loss/train': 2.0964877605438232} 02/25/2022 03:08:38 - INFO - codeparrot_training - Step 19742: {'lr': 0.0003495572047443796, 'samples': 10108416, 'steps': 19742, 'loss/train': 2.4307374954223633} 02/25/2022 03:08:41 - INFO - codeparrot_training - Step 19743: {'lr': 0.0003495421954722762, 'samples': 10108928, 'steps': 19743, 'loss/train': 1.7479761838912964} 02/25/2022 03:08:45 - INFO - codeparrot_training - Step 19744: {'lr': 0.0003495271857737657, 'samples': 10109440, 'steps': 19744, 'loss/train': 0.9561011791229248} 02/25/2022 03:08:50 - INFO - codeparrot_training - Step 19745: {'lr': 0.00034951217564891226, 'samples': 10109952, 'steps': 19745, 'loss/train': 2.7571966648101807} 02/25/2022 03:08:54 - INFO - codeparrot_training - Step 19746: {'lr': 0.0003494971650977802, 'samples': 10110464, 'steps': 19746, 'loss/train': 1.5141698122024536} 02/25/2022 03:08:59 - INFO - codeparrot_training - Step 19747: {'lr': 0.00034948215412043405, 'samples': 10110976, 'steps': 19747, 'loss/train': 1.3123365640640259} 02/25/2022 03:09:03 - INFO - codeparrot_training - Step 19748: {'lr': 0.00034946714271693783, 'samples': 10111488, 'steps': 19748, 'loss/train': 1.062041997909546} 02/25/2022 03:09:08 - INFO - codeparrot_training - Step 19749: {'lr': 0.00034945213088735595, 'samples': 10112000, 'steps': 19749, 'loss/train': 1.904049277305603} 02/25/2022 03:09:12 - INFO - codeparrot_training - Step 19750: {'lr': 0.00034943711863175277, 'samples': 10112512, 'steps': 19750, 'loss/train': 1.5080755949020386} 02/25/2022 03:09:17 - INFO - codeparrot_training - Step 19751: {'lr': 0.0003494221059501925, 'samples': 10113024, 'steps': 19751, 'loss/train': 1.5641701221466064} 02/25/2022 03:09:21 - INFO - codeparrot_training - Step 19752: {'lr': 0.0003494070928427395, 'samples': 10113536, 'steps': 19752, 'loss/train': 2.83099365234375} 02/25/2022 03:09:27 - INFO - codeparrot_training - Step 19753: {'lr': 0.00034939207930945816, 'samples': 10114048, 'steps': 19753, 'loss/train': 2.6274471282958984} 02/25/2022 03:09:30 - INFO - codeparrot_training - Step 19754: {'lr': 0.00034937706535041263, 'samples': 10114560, 'steps': 19754, 'loss/train': 1.526594877243042} 02/25/2022 03:09:36 - INFO - codeparrot_training - Step 19755: {'lr': 0.0003493620509656674, 'samples': 10115072, 'steps': 19755, 'loss/train': 1.7514238357543945} 02/25/2022 03:09:41 - INFO - codeparrot_training - Step 19756: {'lr': 0.00034934703615528657, 'samples': 10115584, 'steps': 19756, 'loss/train': 2.7907276153564453} 02/25/2022 03:09:45 - INFO - codeparrot_training - Step 19757: {'lr': 0.0003493320209193347, 'samples': 10116096, 'steps': 19757, 'loss/train': 2.051156759262085} 02/25/2022 03:09:51 - INFO - codeparrot_training - Step 19758: {'lr': 0.0003493170052578759, 'samples': 10116608, 'steps': 19758, 'loss/train': 1.8145142793655396} 02/25/2022 03:09:54 - INFO - codeparrot_training - Step 19759: {'lr': 0.00034930198917097467, 'samples': 10117120, 'steps': 19759, 'loss/train': 2.934638023376465} 02/25/2022 03:09:57 - INFO - codeparrot_training - Step 19760: {'lr': 0.00034928697265869515, 'samples': 10117632, 'steps': 19760, 'loss/train': 1.642092227935791} 02/25/2022 03:10:03 - INFO - codeparrot_training - Step 19761: {'lr': 0.0003492719557211018, 'samples': 10118144, 'steps': 19761, 'loss/train': 2.2578444480895996} 02/25/2022 03:10:06 - INFO - codeparrot_training - Step 19762: {'lr': 0.0003492569383582589, 'samples': 10118656, 'steps': 19762, 'loss/train': 2.276261329650879} 02/25/2022 03:10:13 - INFO - codeparrot_training - Step 19763: {'lr': 0.0003492419205702309, 'samples': 10119168, 'steps': 19763, 'loss/train': 1.4965124130249023} 02/25/2022 03:10:16 - INFO - codeparrot_training - Step 19764: {'lr': 0.00034922690235708197, 'samples': 10119680, 'steps': 19764, 'loss/train': 2.066959857940674} 02/25/2022 03:10:22 - INFO - codeparrot_training - Step 19765: {'lr': 0.0003492118837188765, 'samples': 10120192, 'steps': 19765, 'loss/train': 2.168503522872925} 02/25/2022 03:10:25 - INFO - codeparrot_training - Step 19766: {'lr': 0.0003491968646556788, 'samples': 10120704, 'steps': 19766, 'loss/train': 2.2297017574310303} 02/25/2022 03:10:31 - INFO - codeparrot_training - Step 19767: {'lr': 0.00034918184516755324, 'samples': 10121216, 'steps': 19767, 'loss/train': 1.77622389793396} 02/25/2022 03:10:34 - INFO - codeparrot_training - Step 19768: {'lr': 0.00034916682525456416, 'samples': 10121728, 'steps': 19768, 'loss/train': 2.5508880615234375} 02/25/2022 03:10:40 - INFO - codeparrot_training - Step 19769: {'lr': 0.00034915180491677583, 'samples': 10122240, 'steps': 19769, 'loss/train': 0.801866352558136} 02/25/2022 03:10:43 - INFO - codeparrot_training - Step 19770: {'lr': 0.00034913678415425276, 'samples': 10122752, 'steps': 19770, 'loss/train': 2.4061455726623535} 02/25/2022 03:10:49 - INFO - codeparrot_training - Step 19771: {'lr': 0.00034912176296705903, 'samples': 10123264, 'steps': 19771, 'loss/train': 2.761408805847168} 02/25/2022 03:10:52 - INFO - codeparrot_training - Step 19772: {'lr': 0.00034910674135525926, 'samples': 10123776, 'steps': 19772, 'loss/train': 1.70564603805542} 02/25/2022 03:10:58 - INFO - codeparrot_training - Step 19773: {'lr': 0.0003490917193189177, 'samples': 10124288, 'steps': 19773, 'loss/train': 1.8040701150894165} 02/25/2022 03:11:04 - INFO - codeparrot_training - Step 19774: {'lr': 0.00034907669685809855, 'samples': 10124800, 'steps': 19774, 'loss/train': 1.7402268648147583} 02/25/2022 03:11:07 - INFO - codeparrot_training - Step 19775: {'lr': 0.0003490616739728664, 'samples': 10125312, 'steps': 19775, 'loss/train': 1.9939600229263306} 02/25/2022 03:11:11 - INFO - codeparrot_training - Step 19776: {'lr': 0.00034904665066328545, 'samples': 10125824, 'steps': 19776, 'loss/train': 2.2742834091186523} 02/25/2022 03:11:17 - INFO - codeparrot_training - Step 19777: {'lr': 0.00034903162692942013, 'samples': 10126336, 'steps': 19777, 'loss/train': 0.256330281496048} 02/25/2022 03:11:22 - INFO - codeparrot_training - Step 19778: {'lr': 0.0003490166027713348, 'samples': 10126848, 'steps': 19778, 'loss/train': 2.058600425720215} 02/25/2022 03:11:25 - INFO - codeparrot_training - Step 19779: {'lr': 0.0003490015781890937, 'samples': 10127360, 'steps': 19779, 'loss/train': 2.046503782272339} 02/25/2022 03:11:31 - INFO - codeparrot_training - Step 19780: {'lr': 0.00034898655318276134, 'samples': 10127872, 'steps': 19780, 'loss/train': 1.8526111841201782} 02/25/2022 03:11:34 - INFO - codeparrot_training - Step 19781: {'lr': 0.00034897152775240206, 'samples': 10128384, 'steps': 19781, 'loss/train': 4.541769027709961} 02/25/2022 03:11:40 - INFO - codeparrot_training - Step 19782: {'lr': 0.00034895650189808005, 'samples': 10128896, 'steps': 19782, 'loss/train': 2.008277177810669} 02/25/2022 03:11:43 - INFO - codeparrot_training - Step 19783: {'lr': 0.0003489414756198598, 'samples': 10129408, 'steps': 19783, 'loss/train': 2.068462371826172} 02/25/2022 03:11:49 - INFO - codeparrot_training - Step 19784: {'lr': 0.0003489264489178058, 'samples': 10129920, 'steps': 19784, 'loss/train': 1.4204087257385254} 02/25/2022 03:11:52 - INFO - codeparrot_training - Step 19785: {'lr': 0.0003489114217919823, 'samples': 10130432, 'steps': 19785, 'loss/train': 1.5703151226043701} 02/25/2022 03:11:58 - INFO - codeparrot_training - Step 19786: {'lr': 0.00034889639424245364, 'samples': 10130944, 'steps': 19786, 'loss/train': 1.9224170446395874} 02/25/2022 03:12:01 - INFO - codeparrot_training - Step 19787: {'lr': 0.00034888136626928427, 'samples': 10131456, 'steps': 19787, 'loss/train': 2.3439974784851074} 02/25/2022 03:12:08 - INFO - codeparrot_training - Step 19788: {'lr': 0.00034886633787253846, 'samples': 10131968, 'steps': 19788, 'loss/train': 1.9139883518218994} 02/25/2022 03:12:11 - INFO - codeparrot_training - Step 19789: {'lr': 0.00034885130905228063, 'samples': 10132480, 'steps': 19789, 'loss/train': 1.4127751588821411} 02/25/2022 03:12:15 - INFO - codeparrot_training - Step 19790: {'lr': 0.00034883627980857523, 'samples': 10132992, 'steps': 19790, 'loss/train': 1.917236328125} 02/25/2022 03:12:20 - INFO - codeparrot_training - Step 19791: {'lr': 0.0003488212501414866, 'samples': 10133504, 'steps': 19791, 'loss/train': 1.3780182600021362} 02/25/2022 03:12:24 - INFO - codeparrot_training - Step 19792: {'lr': 0.00034880622005107915, 'samples': 10134016, 'steps': 19792, 'loss/train': 0.6669219136238098} 02/25/2022 03:12:29 - INFO - codeparrot_training - Step 19793: {'lr': 0.00034879118953741716, 'samples': 10134528, 'steps': 19793, 'loss/train': 1.807779312133789} 02/25/2022 03:12:33 - INFO - codeparrot_training - Step 19794: {'lr': 0.0003487761586005651, 'samples': 10135040, 'steps': 19794, 'loss/train': 2.2414004802703857} 02/25/2022 03:12:38 - INFO - codeparrot_training - Step 19795: {'lr': 0.0003487611272405873, 'samples': 10135552, 'steps': 19795, 'loss/train': 3.0148677825927734} 02/25/2022 03:12:44 - INFO - codeparrot_training - Step 19796: {'lr': 0.00034874609545754826, 'samples': 10136064, 'steps': 19796, 'loss/train': 2.7842581272125244} 02/25/2022 03:12:47 - INFO - codeparrot_training - Step 19797: {'lr': 0.00034873106325151234, 'samples': 10136576, 'steps': 19797, 'loss/train': 0.2121119499206543} 02/25/2022 03:12:54 - INFO - codeparrot_training - Step 19798: {'lr': 0.0003487160306225438, 'samples': 10137088, 'steps': 19798, 'loss/train': 1.2804843187332153} 02/25/2022 03:12:58 - INFO - codeparrot_training - Step 19799: {'lr': 0.00034870099757070716, 'samples': 10137600, 'steps': 19799, 'loss/train': 2.1828575134277344} 02/25/2022 03:13:01 - INFO - codeparrot_training - Step 19800: {'lr': 0.0003486859640960668, 'samples': 10138112, 'steps': 19800, 'loss/train': 0.934968888759613} 02/25/2022 03:13:07 - INFO - codeparrot_training - Step 19801: {'lr': 0.0003486709301986871, 'samples': 10138624, 'steps': 19801, 'loss/train': 3.027498483657837} 02/25/2022 03:13:10 - INFO - codeparrot_training - Step 19802: {'lr': 0.00034865589587863247, 'samples': 10139136, 'steps': 19802, 'loss/train': 2.7640810012817383} 02/25/2022 03:13:16 - INFO - codeparrot_training - Step 19803: {'lr': 0.0003486408611359673, 'samples': 10139648, 'steps': 19803, 'loss/train': 1.5279866456985474} 02/25/2022 03:13:19 - INFO - codeparrot_training - Step 19804: {'lr': 0.00034862582597075607, 'samples': 10140160, 'steps': 19804, 'loss/train': 2.3333680629730225} 02/25/2022 03:13:25 - INFO - codeparrot_training - Step 19805: {'lr': 0.00034861079038306305, 'samples': 10140672, 'steps': 19805, 'loss/train': 1.9807062149047852} 02/25/2022 03:13:28 - INFO - codeparrot_training - Step 19806: {'lr': 0.00034859575437295277, 'samples': 10141184, 'steps': 19806, 'loss/train': 2.265599250793457} 02/25/2022 03:13:34 - INFO - codeparrot_training - Step 19807: {'lr': 0.00034858071794048953, 'samples': 10141696, 'steps': 19807, 'loss/train': 2.709702253341675} 02/25/2022 03:13:37 - INFO - codeparrot_training - Step 19808: {'lr': 0.0003485656810857378, 'samples': 10142208, 'steps': 19808, 'loss/train': 1.8666470050811768} 02/25/2022 03:13:43 - INFO - codeparrot_training - Step 19809: {'lr': 0.00034855064380876193, 'samples': 10142720, 'steps': 19809, 'loss/train': 1.3959065675735474} 02/25/2022 03:13:47 - INFO - codeparrot_training - Step 19810: {'lr': 0.00034853560610962654, 'samples': 10143232, 'steps': 19810, 'loss/train': 1.6216082572937012} 02/25/2022 03:13:52 - INFO - codeparrot_training - Step 19811: {'lr': 0.0003485205679883958, 'samples': 10143744, 'steps': 19811, 'loss/train': 2.1622729301452637} 02/25/2022 03:13:56 - INFO - codeparrot_training - Step 19812: {'lr': 0.00034850552944513426, 'samples': 10144256, 'steps': 19812, 'loss/train': 2.237422227859497} 02/25/2022 03:14:01 - INFO - codeparrot_training - Step 19813: {'lr': 0.00034849049047990633, 'samples': 10144768, 'steps': 19813, 'loss/train': 2.4103262424468994} 02/25/2022 03:14:05 - INFO - codeparrot_training - Step 19814: {'lr': 0.0003484754510927764, 'samples': 10145280, 'steps': 19814, 'loss/train': 2.142704486846924} 02/25/2022 03:14:11 - INFO - codeparrot_training - Step 19815: {'lr': 0.00034846041128380886, 'samples': 10145792, 'steps': 19815, 'loss/train': 1.2305799722671509} 02/25/2022 03:14:14 - INFO - codeparrot_training - Step 19816: {'lr': 0.0003484453710530682, 'samples': 10146304, 'steps': 19816, 'loss/train': 0.07420670986175537} 02/25/2022 03:14:20 - INFO - codeparrot_training - Step 19817: {'lr': 0.0003484303304006189, 'samples': 10146816, 'steps': 19817, 'loss/train': 1.7484591007232666} 02/25/2022 03:14:23 - INFO - codeparrot_training - Step 19818: {'lr': 0.0003484152893265253, 'samples': 10147328, 'steps': 19818, 'loss/train': 1.968850016593933} 02/25/2022 03:14:29 - INFO - codeparrot_training - Step 19819: {'lr': 0.00034840024783085177, 'samples': 10147840, 'steps': 19819, 'loss/train': 2.4700424671173096} 02/25/2022 03:14:33 - INFO - codeparrot_training - Step 19820: {'lr': 0.00034838520591366285, 'samples': 10148352, 'steps': 19820, 'loss/train': 3.0214126110076904} 02/25/2022 03:14:38 - INFO - codeparrot_training - Step 19821: {'lr': 0.00034837016357502297, 'samples': 10148864, 'steps': 19821, 'loss/train': 1.083014965057373} 02/25/2022 03:14:42 - INFO - codeparrot_training - Step 19822: {'lr': 0.0003483551208149965, 'samples': 10149376, 'steps': 19822, 'loss/train': 0.3195439577102661} 02/25/2022 03:14:48 - INFO - codeparrot_training - Step 19823: {'lr': 0.00034834007763364803, 'samples': 10149888, 'steps': 19823, 'loss/train': 1.2361674308776855} 02/25/2022 03:14:51 - INFO - codeparrot_training - Step 19824: {'lr': 0.00034832503403104184, 'samples': 10150400, 'steps': 19824, 'loss/train': 1.9308634996414185} 02/25/2022 03:14:57 - INFO - codeparrot_training - Step 19825: {'lr': 0.00034830999000724246, 'samples': 10150912, 'steps': 19825, 'loss/train': 1.9489511251449585} 02/25/2022 03:15:00 - INFO - codeparrot_training - Step 19826: {'lr': 0.00034829494556231423, 'samples': 10151424, 'steps': 19826, 'loss/train': 2.585437297821045} 02/25/2022 03:15:06 - INFO - codeparrot_training - Step 19827: {'lr': 0.00034827990069632173, 'samples': 10151936, 'steps': 19827, 'loss/train': 3.128664493560791} 02/25/2022 03:15:09 - INFO - codeparrot_training - Step 19828: {'lr': 0.0003482648554093293, 'samples': 10152448, 'steps': 19828, 'loss/train': 2.040684223175049} 02/25/2022 03:15:15 - INFO - codeparrot_training - Step 19829: {'lr': 0.0003482498097014015, 'samples': 10152960, 'steps': 19829, 'loss/train': 2.1903162002563477} 02/25/2022 03:15:18 - INFO - codeparrot_training - Step 19830: {'lr': 0.0003482347635726026, 'samples': 10153472, 'steps': 19830, 'loss/train': 1.3641180992126465} 02/25/2022 03:15:24 - INFO - codeparrot_training - Step 19831: {'lr': 0.00034821971702299716, 'samples': 10153984, 'steps': 19831, 'loss/train': 2.8737971782684326} 02/25/2022 03:15:27 - INFO - codeparrot_training - Step 19832: {'lr': 0.0003482046700526498, 'samples': 10154496, 'steps': 19832, 'loss/train': 2.024074077606201} 02/25/2022 03:15:33 - INFO - codeparrot_training - Step 19833: {'lr': 0.0003481896226616246, 'samples': 10155008, 'steps': 19833, 'loss/train': 1.6594007015228271} 02/25/2022 03:15:36 - INFO - codeparrot_training - Step 19834: {'lr': 0.00034817457484998644, 'samples': 10155520, 'steps': 19834, 'loss/train': 2.130175828933716} 02/25/2022 03:15:42 - INFO - codeparrot_training - Step 19835: {'lr': 0.00034815952661779946, 'samples': 10156032, 'steps': 19835, 'loss/train': 2.5905861854553223} 02/25/2022 03:15:46 - INFO - codeparrot_training - Step 19836: {'lr': 0.00034814447796512824, 'samples': 10156544, 'steps': 19836, 'loss/train': 1.9122241735458374} 02/25/2022 03:15:51 - INFO - codeparrot_training - Step 19837: {'lr': 0.0003481294288920373, 'samples': 10157056, 'steps': 19837, 'loss/train': 2.485546112060547} 02/25/2022 03:15:55 - INFO - codeparrot_training - Step 19838: {'lr': 0.00034811437939859106, 'samples': 10157568, 'steps': 19838, 'loss/train': 1.7731329202651978} 02/25/2022 03:16:00 - INFO - codeparrot_training - Step 19839: {'lr': 0.0003480993294848539, 'samples': 10158080, 'steps': 19839, 'loss/train': 1.0564831495285034} 02/25/2022 03:16:04 - INFO - codeparrot_training - Step 19840: {'lr': 0.00034808427915089036, 'samples': 10158592, 'steps': 19840, 'loss/train': 1.7406283617019653} 02/25/2022 03:16:09 - INFO - codeparrot_training - Step 19841: {'lr': 0.000348069228396765, 'samples': 10159104, 'steps': 19841, 'loss/train': 2.0025475025177} 02/25/2022 03:16:13 - INFO - codeparrot_training - Step 19842: {'lr': 0.00034805417722254213, 'samples': 10159616, 'steps': 19842, 'loss/train': 2.2896740436553955} 02/25/2022 03:16:18 - INFO - codeparrot_training - Step 19843: {'lr': 0.00034803912562828633, 'samples': 10160128, 'steps': 19843, 'loss/train': 2.4485151767730713} 02/25/2022 03:16:25 - INFO - codeparrot_training - Step 19844: {'lr': 0.000348024073614062, 'samples': 10160640, 'steps': 19844, 'loss/train': 2.3837153911590576} 02/25/2022 03:16:28 - INFO - codeparrot_training - Step 19845: {'lr': 0.0003480090211799337, 'samples': 10161152, 'steps': 19845, 'loss/train': 1.630618929862976} 02/25/2022 03:16:34 - INFO - codeparrot_training - Step 19846: {'lr': 0.0003479939683259659, 'samples': 10161664, 'steps': 19846, 'loss/train': 1.957349181175232} 02/25/2022 03:16:37 - INFO - codeparrot_training - Step 19847: {'lr': 0.000347978915052223, 'samples': 10162176, 'steps': 19847, 'loss/train': 1.1609623432159424} 02/25/2022 03:16:41 - INFO - codeparrot_training - Step 19848: {'lr': 0.0003479638613587696, 'samples': 10162688, 'steps': 19848, 'loss/train': 2.6124157905578613} 02/25/2022 03:16:46 - INFO - codeparrot_training - Step 19849: {'lr': 0.0003479488072456701, 'samples': 10163200, 'steps': 19849, 'loss/train': 2.1950464248657227} 02/25/2022 03:16:52 - INFO - codeparrot_training - Step 19850: {'lr': 0.000347933752712989, 'samples': 10163712, 'steps': 19850, 'loss/train': 2.1847610473632812} 02/25/2022 03:16:55 - INFO - codeparrot_training - Step 19851: {'lr': 0.00034791869776079084, 'samples': 10164224, 'steps': 19851, 'loss/train': 2.757704257965088} 02/25/2022 03:17:01 - INFO - codeparrot_training - Step 19852: {'lr': 0.00034790364238914003, 'samples': 10164736, 'steps': 19852, 'loss/train': 2.469663143157959} 02/25/2022 03:17:04 - INFO - codeparrot_training - Step 19853: {'lr': 0.0003478885865981011, 'samples': 10165248, 'steps': 19853, 'loss/train': 1.9874293804168701} 02/25/2022 03:17:10 - INFO - codeparrot_training - Step 19854: {'lr': 0.0003478735303877386, 'samples': 10165760, 'steps': 19854, 'loss/train': 1.5458056926727295} 02/25/2022 03:17:14 - INFO - codeparrot_training - Step 19855: {'lr': 0.0003478584737581169, 'samples': 10166272, 'steps': 19855, 'loss/train': 1.4001595973968506} 02/25/2022 03:17:19 - INFO - codeparrot_training - Step 19856: {'lr': 0.00034784341670930066, 'samples': 10166784, 'steps': 19856, 'loss/train': 1.5884674787521362} 02/25/2022 03:17:23 - INFO - codeparrot_training - Step 19857: {'lr': 0.00034782835924135417, 'samples': 10167296, 'steps': 19857, 'loss/train': 3.9674744606018066} 02/25/2022 03:17:28 - INFO - codeparrot_training - Step 19858: {'lr': 0.0003478133013543422, 'samples': 10167808, 'steps': 19858, 'loss/train': 1.4108281135559082} 02/25/2022 03:17:32 - INFO - codeparrot_training - Step 19859: {'lr': 0.000347798243048329, 'samples': 10168320, 'steps': 19859, 'loss/train': 1.3814793825149536} 02/25/2022 03:17:37 - INFO - codeparrot_training - Step 19860: {'lr': 0.00034778318432337926, 'samples': 10168832, 'steps': 19860, 'loss/train': 3.2302088737487793} 02/25/2022 03:17:41 - INFO - codeparrot_training - Step 19861: {'lr': 0.0003477681251795573, 'samples': 10169344, 'steps': 19861, 'loss/train': 1.1204324960708618} 02/25/2022 03:17:46 - INFO - codeparrot_training - Step 19862: {'lr': 0.0003477530656169278, 'samples': 10169856, 'steps': 19862, 'loss/train': 1.9495435953140259} 02/25/2022 03:17:50 - INFO - codeparrot_training - Step 19863: {'lr': 0.00034773800563555517, 'samples': 10170368, 'steps': 19863, 'loss/train': 1.4796690940856934} 02/25/2022 03:17:55 - INFO - codeparrot_training - Step 19864: {'lr': 0.0003477229452355041, 'samples': 10170880, 'steps': 19864, 'loss/train': 1.983762502670288} 02/25/2022 03:17:59 - INFO - codeparrot_training - Step 19865: {'lr': 0.00034770788441683875, 'samples': 10171392, 'steps': 19865, 'loss/train': 3.638152837753296} 02/25/2022 03:18:05 - INFO - codeparrot_training - Step 19866: {'lr': 0.00034769282317962405, 'samples': 10171904, 'steps': 19866, 'loss/train': 1.7491425275802612} 02/25/2022 03:18:09 - INFO - codeparrot_training - Step 19867: {'lr': 0.00034767776152392417, 'samples': 10172416, 'steps': 19867, 'loss/train': 2.5320382118225098} 02/25/2022 03:18:14 - INFO - codeparrot_training - Step 19868: {'lr': 0.0003476626994498038, 'samples': 10172928, 'steps': 19868, 'loss/train': 1.660361409187317} 02/25/2022 03:18:18 - INFO - codeparrot_training - Step 19869: {'lr': 0.00034764763695732746, 'samples': 10173440, 'steps': 19869, 'loss/train': 1.7201393842697144} 02/25/2022 03:18:23 - INFO - codeparrot_training - Step 19870: {'lr': 0.0003476325740465597, 'samples': 10173952, 'steps': 19870, 'loss/train': 1.8269997835159302} 02/25/2022 03:18:27 - INFO - codeparrot_training - Step 19871: {'lr': 0.0003476175107175649, 'samples': 10174464, 'steps': 19871, 'loss/train': 2.901841402053833} 02/25/2022 03:18:33 - INFO - codeparrot_training - Step 19872: {'lr': 0.00034760244697040776, 'samples': 10174976, 'steps': 19872, 'loss/train': 2.4298744201660156} 02/25/2022 03:18:36 - INFO - codeparrot_training - Step 19873: {'lr': 0.00034758738280515265, 'samples': 10175488, 'steps': 19873, 'loss/train': 2.4408974647521973} 02/25/2022 03:18:42 - INFO - codeparrot_training - Step 19874: {'lr': 0.00034757231822186426, 'samples': 10176000, 'steps': 19874, 'loss/train': 0.612687349319458} 02/25/2022 03:18:45 - INFO - codeparrot_training - Step 19875: {'lr': 0.00034755725322060705, 'samples': 10176512, 'steps': 19875, 'loss/train': 1.9196265935897827} 02/25/2022 03:18:51 - INFO - codeparrot_training - Step 19876: {'lr': 0.00034754218780144546, 'samples': 10177024, 'steps': 19876, 'loss/train': 2.2596445083618164} 02/25/2022 03:18:54 - INFO - codeparrot_training - Step 19877: {'lr': 0.00034752712196444417, 'samples': 10177536, 'steps': 19877, 'loss/train': 5.748746395111084} 02/25/2022 03:19:00 - INFO - codeparrot_training - Step 19878: {'lr': 0.00034751205570966764, 'samples': 10178048, 'steps': 19878, 'loss/train': 1.9081982374191284} 02/25/2022 03:19:03 - INFO - codeparrot_training - Step 19879: {'lr': 0.0003474969890371805, 'samples': 10178560, 'steps': 19879, 'loss/train': 3.318934440612793} 02/25/2022 03:19:11 - INFO - codeparrot_training - Step 19880: {'lr': 0.0003474819219470471, 'samples': 10179072, 'steps': 19880, 'loss/train': 1.8040556907653809} 02/25/2022 03:19:14 - INFO - codeparrot_training - Step 19881: {'lr': 0.0003474668544393321, 'samples': 10179584, 'steps': 19881, 'loss/train': 1.8916703462600708} 02/25/2022 03:19:20 - INFO - codeparrot_training - Step 19882: {'lr': 0.00034745178651410014, 'samples': 10180096, 'steps': 19882, 'loss/train': 2.4303483963012695} 02/25/2022 03:19:23 - INFO - codeparrot_training - Step 19883: {'lr': 0.0003474367181714156, 'samples': 10180608, 'steps': 19883, 'loss/train': 1.925785779953003} 02/25/2022 03:19:29 - INFO - codeparrot_training - Step 19884: {'lr': 0.0003474216494113431, 'samples': 10181120, 'steps': 19884, 'loss/train': 2.178867816925049} 02/25/2022 03:19:32 - INFO - codeparrot_training - Step 19885: {'lr': 0.00034740658023394723, 'samples': 10181632, 'steps': 19885, 'loss/train': 2.8438267707824707} 02/25/2022 03:19:38 - INFO - codeparrot_training - Step 19886: {'lr': 0.0003473915106392925, 'samples': 10182144, 'steps': 19886, 'loss/train': 1.2732062339782715} 02/25/2022 03:19:41 - INFO - codeparrot_training - Step 19887: {'lr': 0.00034737644062744343, 'samples': 10182656, 'steps': 19887, 'loss/train': 1.9972949028015137} 02/25/2022 03:19:47 - INFO - codeparrot_training - Step 19888: {'lr': 0.0003473613701984646, 'samples': 10183168, 'steps': 19888, 'loss/train': 1.4803389310836792} 02/25/2022 03:19:50 - INFO - codeparrot_training - Step 19889: {'lr': 0.0003473462993524206, 'samples': 10183680, 'steps': 19889, 'loss/train': 1.1882394552230835} 02/25/2022 03:19:56 - INFO - codeparrot_training - Step 19890: {'lr': 0.000347331228089376, 'samples': 10184192, 'steps': 19890, 'loss/train': 1.7638345956802368} 02/25/2022 03:20:00 - INFO - codeparrot_training - Step 19891: {'lr': 0.0003473161564093953, 'samples': 10184704, 'steps': 19891, 'loss/train': 2.086263656616211} 02/25/2022 03:20:05 - INFO - codeparrot_training - Step 19892: {'lr': 0.0003473010843125431, 'samples': 10185216, 'steps': 19892, 'loss/train': 2.5646424293518066} 02/25/2022 03:20:09 - INFO - codeparrot_training - Step 19893: {'lr': 0.00034728601179888395, 'samples': 10185728, 'steps': 19893, 'loss/train': 2.2138781547546387} 02/25/2022 03:20:14 - INFO - codeparrot_training - Step 19894: {'lr': 0.00034727093886848236, 'samples': 10186240, 'steps': 19894, 'loss/train': 2.007449150085449} 02/25/2022 03:20:18 - INFO - codeparrot_training - Step 19895: {'lr': 0.000347255865521403, 'samples': 10186752, 'steps': 19895, 'loss/train': 1.0181776285171509} 02/25/2022 03:20:23 - INFO - codeparrot_training - Step 19896: {'lr': 0.0003472407917577104, 'samples': 10187264, 'steps': 19896, 'loss/train': 1.5233153104782104} 02/25/2022 03:20:27 - INFO - codeparrot_training - Step 19897: {'lr': 0.0003472257175774691, 'samples': 10187776, 'steps': 19897, 'loss/train': 0.6901962161064148} 02/25/2022 03:20:33 - INFO - codeparrot_training - Step 19898: {'lr': 0.00034721064298074366, 'samples': 10188288, 'steps': 19898, 'loss/train': 2.7319552898406982} 02/25/2022 03:20:36 - INFO - codeparrot_training - Step 19899: {'lr': 0.0003471955679675988, 'samples': 10188800, 'steps': 19899, 'loss/train': 0.7970505952835083} 02/25/2022 03:20:42 - INFO - codeparrot_training - Step 19900: {'lr': 0.0003471804925380989, 'samples': 10189312, 'steps': 19900, 'loss/train': 1.6189533472061157} 02/25/2022 03:20:45 - INFO - codeparrot_training - Step 19901: {'lr': 0.0003471654166923087, 'samples': 10189824, 'steps': 19901, 'loss/train': 2.3713557720184326} 02/25/2022 03:20:51 - INFO - codeparrot_training - Step 19902: {'lr': 0.00034715034043029263, 'samples': 10190336, 'steps': 19902, 'loss/train': 2.5521323680877686} 02/25/2022 03:20:55 - INFO - codeparrot_training - Step 19903: {'lr': 0.00034713526375211546, 'samples': 10190848, 'steps': 19903, 'loss/train': 1.1911510229110718} 02/25/2022 03:21:00 - INFO - codeparrot_training - Step 19904: {'lr': 0.00034712018665784155, 'samples': 10191360, 'steps': 19904, 'loss/train': 1.551103949546814} 02/25/2022 03:21:04 - INFO - codeparrot_training - Step 19905: {'lr': 0.00034710510914753563, 'samples': 10191872, 'steps': 19905, 'loss/train': 1.9401558637619019} 02/25/2022 03:21:09 - INFO - codeparrot_training - Step 19906: {'lr': 0.00034709003122126227, 'samples': 10192384, 'steps': 19906, 'loss/train': 2.7029035091400146} 02/25/2022 03:21:13 - INFO - codeparrot_training - Step 19907: {'lr': 0.000347074952879086, 'samples': 10192896, 'steps': 19907, 'loss/train': 1.5081888437271118} 02/25/2022 03:21:18 - INFO - codeparrot_training - Step 19908: {'lr': 0.0003470598741210715, 'samples': 10193408, 'steps': 19908, 'loss/train': 1.5659664869308472} 02/25/2022 03:21:22 - INFO - codeparrot_training - Step 19909: {'lr': 0.00034704479494728337, 'samples': 10193920, 'steps': 19909, 'loss/train': 0.2435697615146637} 02/25/2022 03:21:27 - INFO - codeparrot_training - Step 19910: {'lr': 0.00034702971535778614, 'samples': 10194432, 'steps': 19910, 'loss/train': 1.5257058143615723} 02/25/2022 03:21:31 - INFO - codeparrot_training - Step 19911: {'lr': 0.00034701463535264434, 'samples': 10194944, 'steps': 19911, 'loss/train': 1.5345323085784912} 02/25/2022 03:21:37 - INFO - codeparrot_training - Step 19912: {'lr': 0.0003469995549319227, 'samples': 10195456, 'steps': 19912, 'loss/train': 2.294989824295044} 02/25/2022 03:21:40 - INFO - codeparrot_training - Step 19913: {'lr': 0.0003469844740956858, 'samples': 10195968, 'steps': 19913, 'loss/train': 2.1156210899353027} 02/25/2022 03:21:46 - INFO - codeparrot_training - Step 19914: {'lr': 0.0003469693928439982, 'samples': 10196480, 'steps': 19914, 'loss/train': 1.6220123767852783} 02/25/2022 03:21:49 - INFO - codeparrot_training - Step 19915: {'lr': 0.00034695431117692446, 'samples': 10196992, 'steps': 19915, 'loss/train': 0.9607713222503662} 02/25/2022 03:21:55 - INFO - codeparrot_training - Step 19916: {'lr': 0.0003469392290945292, 'samples': 10197504, 'steps': 19916, 'loss/train': 1.727417230606079} 02/25/2022 03:21:58 - INFO - codeparrot_training - Step 19917: {'lr': 0.00034692414659687714, 'samples': 10198016, 'steps': 19917, 'loss/train': 1.8917516469955444} 02/25/2022 03:22:04 - INFO - codeparrot_training - Step 19918: {'lr': 0.0003469090636840328, 'samples': 10198528, 'steps': 19918, 'loss/train': 1.4184610843658447} 02/25/2022 03:22:07 - INFO - codeparrot_training - Step 19919: {'lr': 0.0003468939803560608, 'samples': 10199040, 'steps': 19919, 'loss/train': 1.2965744733810425} 02/25/2022 03:22:13 - INFO - codeparrot_training - Step 19920: {'lr': 0.00034687889661302575, 'samples': 10199552, 'steps': 19920, 'loss/train': 0.7373710870742798} 02/25/2022 03:22:16 - INFO - codeparrot_training - Step 19921: {'lr': 0.0003468638124549923, 'samples': 10200064, 'steps': 19921, 'loss/train': 3.1162328720092773} 02/25/2022 03:22:22 - INFO - codeparrot_training - Step 19922: {'lr': 0.00034684872788202497, 'samples': 10200576, 'steps': 19922, 'loss/train': 2.807286024093628} 02/25/2022 03:22:25 - INFO - codeparrot_training - Step 19923: {'lr': 0.0003468336428941885, 'samples': 10201088, 'steps': 19923, 'loss/train': 0.7555232048034668} 02/25/2022 03:22:31 - INFO - codeparrot_training - Step 19924: {'lr': 0.00034681855749154743, 'samples': 10201600, 'steps': 19924, 'loss/train': 2.127145767211914} 02/25/2022 03:22:35 - INFO - codeparrot_training - Step 19925: {'lr': 0.00034680347167416643, 'samples': 10202112, 'steps': 19925, 'loss/train': 2.4664297103881836} 02/25/2022 03:22:41 - INFO - codeparrot_training - Step 19926: {'lr': 0.00034678838544211003, 'samples': 10202624, 'steps': 19926, 'loss/train': 1.6883305311203003} 02/25/2022 03:22:44 - INFO - codeparrot_training - Step 19927: {'lr': 0.000346773298795443, 'samples': 10203136, 'steps': 19927, 'loss/train': 2.5180041790008545} 02/25/2022 03:22:50 - INFO - codeparrot_training - Step 19928: {'lr': 0.00034675821173422983, 'samples': 10203648, 'steps': 19928, 'loss/train': 1.777349829673767} 02/25/2022 03:22:55 - INFO - codeparrot_training - Step 19929: {'lr': 0.0003467431242585352, 'samples': 10204160, 'steps': 19929, 'loss/train': 2.1454718112945557} 02/25/2022 03:22:59 - INFO - codeparrot_training - Step 19930: {'lr': 0.0003467280363684238, 'samples': 10204672, 'steps': 19930, 'loss/train': 1.5410213470458984} 02/25/2022 03:23:02 - INFO - codeparrot_training - Step 19931: {'lr': 0.0003467129480639601, 'samples': 10205184, 'steps': 19931, 'loss/train': 2.920882225036621} 02/25/2022 03:23:08 - INFO - codeparrot_training - Step 19932: {'lr': 0.000346697859345209, 'samples': 10205696, 'steps': 19932, 'loss/train': 3.7892770767211914} 02/25/2022 03:23:11 - INFO - codeparrot_training - Step 19933: {'lr': 0.00034668277021223493, 'samples': 10206208, 'steps': 19933, 'loss/train': 1.9957493543624878} 02/25/2022 03:23:17 - INFO - codeparrot_training - Step 19934: {'lr': 0.0003466676806651025, 'samples': 10206720, 'steps': 19934, 'loss/train': 0.9894843101501465} 02/25/2022 03:23:20 - INFO - codeparrot_training - Step 19935: {'lr': 0.0003466525907038765, 'samples': 10207232, 'steps': 19935, 'loss/train': 2.076892614364624} 02/25/2022 03:23:26 - INFO - codeparrot_training - Step 19936: {'lr': 0.0003466375003286214, 'samples': 10207744, 'steps': 19936, 'loss/train': 0.9773076176643372} 02/25/2022 03:23:29 - INFO - codeparrot_training - Step 19937: {'lr': 0.00034662240953940205, 'samples': 10208256, 'steps': 19937, 'loss/train': 2.639242172241211} 02/25/2022 03:23:35 - INFO - codeparrot_training - Step 19938: {'lr': 0.0003466073183362829, 'samples': 10208768, 'steps': 19938, 'loss/train': 1.8597412109375} 02/25/2022 03:23:39 - INFO - codeparrot_training - Step 19939: {'lr': 0.00034659222671932865, 'samples': 10209280, 'steps': 19939, 'loss/train': 2.3973875045776367} 02/25/2022 03:23:44 - INFO - codeparrot_training - Step 19940: {'lr': 0.000346577134688604, 'samples': 10209792, 'steps': 19940, 'loss/train': 1.4727375507354736} 02/25/2022 03:23:48 - INFO - codeparrot_training - Step 19941: {'lr': 0.0003465620422441737, 'samples': 10210304, 'steps': 19941, 'loss/train': 1.929829478263855} 02/25/2022 03:23:54 - INFO - codeparrot_training - Step 19942: {'lr': 0.00034654694938610205, 'samples': 10210816, 'steps': 19942, 'loss/train': 2.0960168838500977} 02/25/2022 03:23:57 - INFO - codeparrot_training - Step 19943: {'lr': 0.00034653185611445403, 'samples': 10211328, 'steps': 19943, 'loss/train': 1.7377725839614868} 02/25/2022 03:24:03 - INFO - codeparrot_training - Step 19944: {'lr': 0.0003465167624292942, 'samples': 10211840, 'steps': 19944, 'loss/train': 3.166250705718994} 02/25/2022 03:24:08 - INFO - codeparrot_training - Step 19945: {'lr': 0.0003465016683306872, 'samples': 10212352, 'steps': 19945, 'loss/train': 2.699152708053589} 02/25/2022 03:24:12 - INFO - codeparrot_training - Step 19946: {'lr': 0.0003464865738186977, 'samples': 10212864, 'steps': 19946, 'loss/train': 0.7595689296722412} 02/25/2022 03:24:15 - INFO - codeparrot_training - Step 19947: {'lr': 0.0003464714788933904, 'samples': 10213376, 'steps': 19947, 'loss/train': 2.344856023788452} 02/25/2022 03:24:21 - INFO - codeparrot_training - Step 19948: {'lr': 0.0003464563835548298, 'samples': 10213888, 'steps': 19948, 'loss/train': 1.5992101430892944} 02/25/2022 03:24:25 - INFO - codeparrot_training - Step 19949: {'lr': 0.0003464412878030808, 'samples': 10214400, 'steps': 19949, 'loss/train': 2.587639808654785} 02/25/2022 03:24:30 - INFO - codeparrot_training - Step 19950: {'lr': 0.0003464261916382079, 'samples': 10214912, 'steps': 19950, 'loss/train': 2.1051950454711914} 02/25/2022 03:24:34 - INFO - codeparrot_training - Step 19951: {'lr': 0.0003464110950602758, 'samples': 10215424, 'steps': 19951, 'loss/train': 2.285545825958252} 02/25/2022 03:24:39 - INFO - codeparrot_training - Step 19952: {'lr': 0.00034639599806934917, 'samples': 10215936, 'steps': 19952, 'loss/train': 2.064997911453247} 02/25/2022 03:24:43 - INFO - codeparrot_training - Step 19953: {'lr': 0.0003463809006654927, 'samples': 10216448, 'steps': 19953, 'loss/train': 2.6954777240753174} 02/25/2022 03:24:48 - INFO - codeparrot_training - Step 19954: {'lr': 0.0003463658028487711, 'samples': 10216960, 'steps': 19954, 'loss/train': 2.060100555419922} 02/25/2022 03:24:52 - INFO - codeparrot_training - Step 19955: {'lr': 0.0003463507046192489, 'samples': 10217472, 'steps': 19955, 'loss/train': 3.039870500564575} 02/25/2022 03:24:57 - INFO - codeparrot_training - Step 19956: {'lr': 0.000346335605976991, 'samples': 10217984, 'steps': 19956, 'loss/train': 1.7583907842636108} 02/25/2022 03:25:01 - INFO - codeparrot_training - Step 19957: {'lr': 0.00034632050692206184, 'samples': 10218496, 'steps': 19957, 'loss/train': 1.969556450843811} 02/25/2022 03:25:07 - INFO - codeparrot_training - Step 19958: {'lr': 0.0003463054074545263, 'samples': 10219008, 'steps': 19958, 'loss/train': 1.9443734884262085} 02/25/2022 03:25:10 - INFO - codeparrot_training - Step 19959: {'lr': 0.00034629030757444895, 'samples': 10219520, 'steps': 19959, 'loss/train': 2.1260299682617188} 02/25/2022 03:25:16 - INFO - codeparrot_training - Step 19960: {'lr': 0.00034627520728189456, 'samples': 10220032, 'steps': 19960, 'loss/train': 1.9024779796600342} 02/25/2022 03:25:22 - INFO - codeparrot_training - Step 19961: {'lr': 0.00034626010657692766, 'samples': 10220544, 'steps': 19961, 'loss/train': 1.2565720081329346} 02/25/2022 03:25:25 - INFO - codeparrot_training - Step 19962: {'lr': 0.0003462450054596131, 'samples': 10221056, 'steps': 19962, 'loss/train': 1.2504113912582397} 02/25/2022 03:25:31 - INFO - codeparrot_training - Step 19963: {'lr': 0.0003462299039300154, 'samples': 10221568, 'steps': 19963, 'loss/train': 1.6858859062194824} 02/25/2022 03:25:34 - INFO - codeparrot_training - Step 19964: {'lr': 0.0003462148019881994, 'samples': 10222080, 'steps': 19964, 'loss/train': 1.0568069219589233} 02/25/2022 03:25:40 - INFO - codeparrot_training - Step 19965: {'lr': 0.0003461996996342298, 'samples': 10222592, 'steps': 19965, 'loss/train': 0.959199070930481} 02/25/2022 03:25:43 - INFO - codeparrot_training - Step 19966: {'lr': 0.00034618459686817113, 'samples': 10223104, 'steps': 19966, 'loss/train': 2.508249044418335} 02/25/2022 03:25:49 - INFO - codeparrot_training - Step 19967: {'lr': 0.0003461694936900883, 'samples': 10223616, 'steps': 19967, 'loss/train': 2.3394620418548584} 02/25/2022 03:25:52 - INFO - codeparrot_training - Step 19968: {'lr': 0.0003461543901000458, 'samples': 10224128, 'steps': 19968, 'loss/train': 2.048583507537842} 02/25/2022 03:25:58 - INFO - codeparrot_training - Step 19969: {'lr': 0.00034613928609810845, 'samples': 10224640, 'steps': 19969, 'loss/train': 0.318865031003952} 02/25/2022 03:26:01 - INFO - codeparrot_training - Step 19970: {'lr': 0.0003461241816843409, 'samples': 10225152, 'steps': 19970, 'loss/train': 2.4739160537719727} 02/25/2022 03:26:07 - INFO - codeparrot_training - Step 19971: {'lr': 0.00034610907685880794, 'samples': 10225664, 'steps': 19971, 'loss/train': 1.484741449356079} 02/25/2022 03:26:10 - INFO - codeparrot_training - Step 19972: {'lr': 0.00034609397162157417, 'samples': 10226176, 'steps': 19972, 'loss/train': 1.639478325843811} 02/25/2022 03:26:16 - INFO - codeparrot_training - Step 19973: {'lr': 0.0003460788659727044, 'samples': 10226688, 'steps': 19973, 'loss/train': 2.4710071086883545} 02/25/2022 03:26:20 - INFO - codeparrot_training - Step 19974: {'lr': 0.0003460637599122632, 'samples': 10227200, 'steps': 19974, 'loss/train': 0.15729981660842896} 02/25/2022 03:26:25 - INFO - codeparrot_training - Step 19975: {'lr': 0.0003460486534403154, 'samples': 10227712, 'steps': 19975, 'loss/train': 1.7175712585449219} 02/25/2022 03:26:29 - INFO - codeparrot_training - Step 19976: {'lr': 0.0003460335465569256, 'samples': 10228224, 'steps': 19976, 'loss/train': 2.6938610076904297} 02/25/2022 03:26:34 - INFO - codeparrot_training - Step 19977: {'lr': 0.0003460184392621587, 'samples': 10228736, 'steps': 19977, 'loss/train': 2.1446146965026855} 02/25/2022 03:26:38 - INFO - codeparrot_training - Step 19978: {'lr': 0.0003460033315560792, 'samples': 10229248, 'steps': 19978, 'loss/train': 1.577078938484192} 02/25/2022 03:26:43 - INFO - codeparrot_training - Step 19979: {'lr': 0.00034598822343875197, 'samples': 10229760, 'steps': 19979, 'loss/train': 2.1239805221557617} 02/25/2022 03:26:47 - INFO - codeparrot_training - Step 19980: {'lr': 0.0003459731149102417, 'samples': 10230272, 'steps': 19980, 'loss/train': 3.171340227127075} 02/25/2022 03:26:52 - INFO - codeparrot_training - Step 19981: {'lr': 0.000345958005970613, 'samples': 10230784, 'steps': 19981, 'loss/train': 1.18368661403656} 02/25/2022 03:26:56 - INFO - codeparrot_training - Step 19982: {'lr': 0.0003459428966199307, 'samples': 10231296, 'steps': 19982, 'loss/train': 2.0071804523468018} 02/25/2022 03:27:02 - INFO - codeparrot_training - Step 19983: {'lr': 0.0003459277868582595, 'samples': 10231808, 'steps': 19983, 'loss/train': 2.0091047286987305} 02/25/2022 03:27:05 - INFO - codeparrot_training - Step 19984: {'lr': 0.00034591267668566413, 'samples': 10232320, 'steps': 19984, 'loss/train': 2.1538147926330566} 02/25/2022 03:27:11 - INFO - codeparrot_training - Step 19985: {'lr': 0.00034589756610220923, 'samples': 10232832, 'steps': 19985, 'loss/train': 1.5419145822525024} 02/25/2022 03:27:14 - INFO - codeparrot_training - Step 19986: {'lr': 0.0003458824551079597, 'samples': 10233344, 'steps': 19986, 'loss/train': 2.545056104660034} 02/25/2022 03:27:20 - INFO - codeparrot_training - Step 19987: {'lr': 0.00034586734370298017, 'samples': 10233856, 'steps': 19987, 'loss/train': 1.6556711196899414} 02/25/2022 03:27:23 - INFO - codeparrot_training - Step 19988: {'lr': 0.00034585223188733535, 'samples': 10234368, 'steps': 19988, 'loss/train': 2.288792848587036} 02/25/2022 03:27:29 - INFO - codeparrot_training - Step 19989: {'lr': 0.00034583711966109005, 'samples': 10234880, 'steps': 19989, 'loss/train': 1.9010210037231445} 02/25/2022 03:27:32 - INFO - codeparrot_training - Step 19990: {'lr': 0.0003458220070243089, 'samples': 10235392, 'steps': 19990, 'loss/train': 0.43585482239723206} 02/25/2022 03:27:38 - INFO - codeparrot_training - Step 19991: {'lr': 0.0003458068939770567, 'samples': 10235904, 'steps': 19991, 'loss/train': 2.1353538036346436} 02/25/2022 03:27:41 - INFO - codeparrot_training - Step 19992: {'lr': 0.00034579178051939827, 'samples': 10236416, 'steps': 19992, 'loss/train': 2.5798909664154053} 02/25/2022 03:27:48 - INFO - codeparrot_training - Step 19993: {'lr': 0.00034577666665139815, 'samples': 10236928, 'steps': 19993, 'loss/train': 1.2594544887542725} 02/25/2022 03:27:51 - INFO - codeparrot_training - Step 19994: {'lr': 0.0003457615523731213, 'samples': 10237440, 'steps': 19994, 'loss/train': 2.1348679065704346} 02/25/2022 03:27:57 - INFO - codeparrot_training - Step 19995: {'lr': 0.00034574643768463237, 'samples': 10237952, 'steps': 19995, 'loss/train': 2.14424467086792} 02/25/2022 03:28:00 - INFO - codeparrot_training - Step 19996: {'lr': 0.00034573132258599606, 'samples': 10238464, 'steps': 19996, 'loss/train': 1.6995242834091187} 02/25/2022 03:28:06 - INFO - codeparrot_training - Step 19997: {'lr': 0.00034571620707727713, 'samples': 10238976, 'steps': 19997, 'loss/train': 1.9517722129821777} 02/25/2022 03:28:09 - INFO - codeparrot_training - Step 19998: {'lr': 0.0003457010911585404, 'samples': 10239488, 'steps': 19998, 'loss/train': 1.497208595275879} 02/25/2022 03:28:15 - INFO - codeparrot_training - Step 19999: {'lr': 0.00034568597482985067, 'samples': 10240000, 'steps': 19999, 'loss/train': 1.5917458534240723} 02/25/2022 03:28:15 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 03:28:32 - WARNING - huggingface_hub.repository - Several commits (20) will be pushed upstream. 02/25/2022 03:28:32 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 03:29:06 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy e9773d2..ad14ca4 floral-grass-11 -> floral-grass-11 02/25/2022 03:29:10 - INFO - codeparrot_training - Step 20000: {'lr': 0.0003456708580912725, 'samples': 10240512, 'steps': 20000, 'loss/train': 2.0182840824127197} 02/25/2022 03:29:16 - INFO - codeparrot_training - Step 20001: {'lr': 0.00034565574094287075, 'samples': 10241024, 'steps': 20001, 'loss/train': 2.2768688201904297} 02/25/2022 03:29:19 - INFO - codeparrot_training - Step 20002: {'lr': 0.0003456406233847102, 'samples': 10241536, 'steps': 20002, 'loss/train': 2.4322569370269775} 02/25/2022 03:29:26 - INFO - codeparrot_training - Step 20003: {'lr': 0.00034562550541685557, 'samples': 10242048, 'steps': 20003, 'loss/train': 1.8150594234466553} 02/25/2022 03:29:29 - INFO - codeparrot_training - Step 20004: {'lr': 0.0003456103870393717, 'samples': 10242560, 'steps': 20004, 'loss/train': 2.9891395568847656} 02/25/2022 03:29:35 - INFO - codeparrot_training - Step 20005: {'lr': 0.0003455952682523232, 'samples': 10243072, 'steps': 20005, 'loss/train': 1.8301373720169067} 02/25/2022 03:29:38 - INFO - codeparrot_training - Step 20006: {'lr': 0.00034558014905577506, 'samples': 10243584, 'steps': 20006, 'loss/train': 2.319890022277832} 02/25/2022 03:29:44 - INFO - codeparrot_training - Step 20007: {'lr': 0.00034556502944979177, 'samples': 10244096, 'steps': 20007, 'loss/train': 1.6531423330307007} 02/25/2022 03:29:47 - INFO - codeparrot_training - Step 20008: {'lr': 0.0003455499094344383, 'samples': 10244608, 'steps': 20008, 'loss/train': 1.5679006576538086} 02/25/2022 03:29:53 - INFO - codeparrot_training - Step 20009: {'lr': 0.00034553478900977943, 'samples': 10245120, 'steps': 20009, 'loss/train': 2.504625082015991} 02/25/2022 03:29:56 - INFO - codeparrot_training - Step 20010: {'lr': 0.0003455196681758798, 'samples': 10245632, 'steps': 20010, 'loss/train': 1.9351191520690918} 02/25/2022 03:30:02 - INFO - codeparrot_training - Step 20011: {'lr': 0.00034550454693280417, 'samples': 10246144, 'steps': 20011, 'loss/train': 2.481072425842285} 02/25/2022 03:30:05 - INFO - codeparrot_training - Step 20012: {'lr': 0.0003454894252806175, 'samples': 10246656, 'steps': 20012, 'loss/train': 1.2526456117630005} 02/25/2022 03:30:11 - INFO - codeparrot_training - Step 20013: {'lr': 0.0003454743032193844, 'samples': 10247168, 'steps': 20013, 'loss/train': 1.3796939849853516} 02/25/2022 03:30:15 - INFO - codeparrot_training - Step 20014: {'lr': 0.00034545918074916965, 'samples': 10247680, 'steps': 20014, 'loss/train': 3.00093674659729} 02/25/2022 03:30:20 - INFO - codeparrot_training - Step 20015: {'lr': 0.00034544405787003817, 'samples': 10248192, 'steps': 20015, 'loss/train': 1.7029036283493042} 02/25/2022 03:30:24 - INFO - codeparrot_training - Step 20016: {'lr': 0.0003454289345820546, 'samples': 10248704, 'steps': 20016, 'loss/train': 1.8101145029067993} 02/25/2022 03:30:29 - INFO - codeparrot_training - Step 20017: {'lr': 0.00034541381088528376, 'samples': 10249216, 'steps': 20017, 'loss/train': 1.9798136949539185} 02/25/2022 03:30:33 - INFO - codeparrot_training - Step 20018: {'lr': 0.00034539868677979055, 'samples': 10249728, 'steps': 20018, 'loss/train': 1.6781688928604126} 02/25/2022 03:30:39 - INFO - codeparrot_training - Step 20019: {'lr': 0.0003453835622656396, 'samples': 10250240, 'steps': 20019, 'loss/train': 1.1882104873657227} 02/25/2022 03:30:42 - INFO - codeparrot_training - Step 20020: {'lr': 0.0003453684373428957, 'samples': 10250752, 'steps': 20020, 'loss/train': 1.8127082586288452} 02/25/2022 03:30:48 - INFO - codeparrot_training - Step 20021: {'lr': 0.0003453533120116238, 'samples': 10251264, 'steps': 20021, 'loss/train': 1.358256459236145} 02/25/2022 03:30:51 - INFO - codeparrot_training - Step 20022: {'lr': 0.0003453381862718886, 'samples': 10251776, 'steps': 20022, 'loss/train': 1.9331316947937012} 02/25/2022 03:30:57 - INFO - codeparrot_training - Step 20023: {'lr': 0.00034532306012375474, 'samples': 10252288, 'steps': 20023, 'loss/train': 1.6333743333816528} 02/25/2022 03:31:00 - INFO - codeparrot_training - Step 20024: {'lr': 0.00034530793356728727, 'samples': 10252800, 'steps': 20024, 'loss/train': 2.4872119426727295} 02/25/2022 03:31:06 - INFO - codeparrot_training - Step 20025: {'lr': 0.00034529280660255084, 'samples': 10253312, 'steps': 20025, 'loss/train': 2.5141241550445557} 02/25/2022 03:31:09 - INFO - codeparrot_training - Step 20026: {'lr': 0.00034527767922961034, 'samples': 10253824, 'steps': 20026, 'loss/train': 1.9492285251617432} 02/25/2022 03:31:15 - INFO - codeparrot_training - Step 20027: {'lr': 0.0003452625514485305, 'samples': 10254336, 'steps': 20027, 'loss/train': 1.420654535293579} 02/25/2022 03:31:18 - INFO - codeparrot_training - Step 20028: {'lr': 0.0003452474232593761, 'samples': 10254848, 'steps': 20028, 'loss/train': 2.066033363342285} 02/25/2022 03:31:24 - INFO - codeparrot_training - Step 20029: {'lr': 0.00034523229466221195, 'samples': 10255360, 'steps': 20029, 'loss/train': 1.3243342638015747} 02/25/2022 03:31:28 - INFO - codeparrot_training - Step 20030: {'lr': 0.00034521716565710293, 'samples': 10255872, 'steps': 20030, 'loss/train': 0.8642081022262573} 02/25/2022 03:31:34 - INFO - codeparrot_training - Step 20031: {'lr': 0.00034520203624411385, 'samples': 10256384, 'steps': 20031, 'loss/train': 1.956555962562561} 02/25/2022 03:31:37 - INFO - codeparrot_training - Step 20032: {'lr': 0.0003451869064233094, 'samples': 10256896, 'steps': 20032, 'loss/train': 2.77828049659729} 02/25/2022 03:31:43 - INFO - codeparrot_training - Step 20033: {'lr': 0.0003451717761947545, 'samples': 10257408, 'steps': 20033, 'loss/train': 2.0623152256011963} 02/25/2022 03:31:46 - INFO - codeparrot_training - Step 20034: {'lr': 0.0003451566455585139, 'samples': 10257920, 'steps': 20034, 'loss/train': 1.875267505645752} 02/25/2022 03:31:52 - INFO - codeparrot_training - Step 20035: {'lr': 0.00034514151451465254, 'samples': 10258432, 'steps': 20035, 'loss/train': 2.443436622619629} 02/25/2022 03:31:55 - INFO - codeparrot_training - Step 20036: {'lr': 0.00034512638306323506, 'samples': 10258944, 'steps': 20036, 'loss/train': 1.7062523365020752} 02/25/2022 03:32:01 - INFO - codeparrot_training - Step 20037: {'lr': 0.0003451112512043264, 'samples': 10259456, 'steps': 20037, 'loss/train': 1.2771670818328857} 02/25/2022 03:32:04 - INFO - codeparrot_training - Step 20038: {'lr': 0.0003450961189379913, 'samples': 10259968, 'steps': 20038, 'loss/train': 1.9417378902435303} 02/25/2022 03:32:11 - INFO - codeparrot_training - Step 20039: {'lr': 0.0003450809862642947, 'samples': 10260480, 'steps': 20039, 'loss/train': 2.025756359100342} 02/25/2022 03:32:14 - INFO - codeparrot_training - Step 20040: {'lr': 0.0003450658531833013, 'samples': 10260992, 'steps': 20040, 'loss/train': 1.3042536973953247} 02/25/2022 03:32:20 - INFO - codeparrot_training - Step 20041: {'lr': 0.00034505071969507595, 'samples': 10261504, 'steps': 20041, 'loss/train': 2.375941514968872} 02/25/2022 03:32:25 - INFO - codeparrot_training - Step 20042: {'lr': 0.0003450355857996835, 'samples': 10262016, 'steps': 20042, 'loss/train': 2.3680078983306885} 02/25/2022 03:32:29 - INFO - codeparrot_training - Step 20043: {'lr': 0.0003450204514971888, 'samples': 10262528, 'steps': 20043, 'loss/train': 0.9760146737098694} 02/25/2022 03:32:34 - INFO - codeparrot_training - Step 20044: {'lr': 0.0003450053167876566, 'samples': 10263040, 'steps': 20044, 'loss/train': 1.9572149515151978} 02/25/2022 03:32:38 - INFO - codeparrot_training - Step 20045: {'lr': 0.0003449901816711519, 'samples': 10263552, 'steps': 20045, 'loss/train': 0.6744064092636108} 02/25/2022 03:32:43 - INFO - codeparrot_training - Step 20046: {'lr': 0.00034497504614773935, 'samples': 10264064, 'steps': 20046, 'loss/train': 2.525388240814209} 02/25/2022 03:32:47 - INFO - codeparrot_training - Step 20047: {'lr': 0.0003449599102174839, 'samples': 10264576, 'steps': 20047, 'loss/train': 2.0306506156921387} 02/25/2022 03:32:53 - INFO - codeparrot_training - Step 20048: {'lr': 0.0003449447738804503, 'samples': 10265088, 'steps': 20048, 'loss/train': 1.8904715776443481} 02/25/2022 03:32:57 - INFO - codeparrot_training - Step 20049: {'lr': 0.0003449296371367034, 'samples': 10265600, 'steps': 20049, 'loss/train': 1.5907604694366455} 02/25/2022 03:33:02 - INFO - codeparrot_training - Step 20050: {'lr': 0.0003449144999863082, 'samples': 10266112, 'steps': 20050, 'loss/train': 2.379424571990967} 02/25/2022 03:33:06 - INFO - codeparrot_training - Step 20051: {'lr': 0.00034489936242932935, 'samples': 10266624, 'steps': 20051, 'loss/train': 0.9318121075630188} 02/25/2022 03:33:11 - INFO - codeparrot_training - Step 20052: {'lr': 0.00034488422446583177, 'samples': 10267136, 'steps': 20052, 'loss/train': 2.723862886428833} 02/25/2022 03:33:15 - INFO - codeparrot_training - Step 20053: {'lr': 0.0003448690860958803, 'samples': 10267648, 'steps': 20053, 'loss/train': 1.6078450679779053} 02/25/2022 03:33:20 - INFO - codeparrot_training - Step 20054: {'lr': 0.00034485394731953976, 'samples': 10268160, 'steps': 20054, 'loss/train': 2.144737720489502} 02/25/2022 03:33:24 - INFO - codeparrot_training - Step 20055: {'lr': 0.00034483880813687505, 'samples': 10268672, 'steps': 20055, 'loss/train': 1.9698745012283325} 02/25/2022 03:33:29 - INFO - codeparrot_training - Step 20056: {'lr': 0.0003448236685479511, 'samples': 10269184, 'steps': 20056, 'loss/train': 1.7840856313705444} 02/25/2022 03:33:33 - INFO - codeparrot_training - Step 20057: {'lr': 0.0003448085285528326, 'samples': 10269696, 'steps': 20057, 'loss/train': 1.2451926469802856} 02/25/2022 03:33:38 - INFO - codeparrot_training - Step 20058: {'lr': 0.00034479338815158447, 'samples': 10270208, 'steps': 20058, 'loss/train': 1.1563801765441895} 02/25/2022 03:33:42 - INFO - codeparrot_training - Step 20059: {'lr': 0.0003447782473442715, 'samples': 10270720, 'steps': 20059, 'loss/train': 1.4472498893737793} 02/25/2022 03:33:47 - INFO - codeparrot_training - Step 20060: {'lr': 0.00034476310613095867, 'samples': 10271232, 'steps': 20060, 'loss/train': 1.7619688510894775} 02/25/2022 03:33:51 - INFO - codeparrot_training - Step 20061: {'lr': 0.00034474796451171075, 'samples': 10271744, 'steps': 20061, 'loss/train': 2.4976987838745117} 02/25/2022 03:33:57 - INFO - codeparrot_training - Step 20062: {'lr': 0.00034473282248659266, 'samples': 10272256, 'steps': 20062, 'loss/train': 1.5934425592422485} 02/25/2022 03:34:00 - INFO - codeparrot_training - Step 20063: {'lr': 0.00034471768005566925, 'samples': 10272768, 'steps': 20063, 'loss/train': 0.8754019141197205} 02/25/2022 03:34:06 - INFO - codeparrot_training - Step 20064: {'lr': 0.00034470253721900535, 'samples': 10273280, 'steps': 20064, 'loss/train': 2.1164333820343018} 02/25/2022 03:34:10 - INFO - codeparrot_training - Step 20065: {'lr': 0.0003446873939766659, 'samples': 10273792, 'steps': 20065, 'loss/train': 2.306887149810791} 02/25/2022 03:34:15 - INFO - codeparrot_training - Step 20066: {'lr': 0.0003446722503287157, 'samples': 10274304, 'steps': 20066, 'loss/train': 1.4776066541671753} 02/25/2022 03:34:19 - INFO - codeparrot_training - Step 20067: {'lr': 0.0003446571062752196, 'samples': 10274816, 'steps': 20067, 'loss/train': 1.2129466533660889} 02/25/2022 03:34:24 - INFO - codeparrot_training - Step 20068: {'lr': 0.0003446419618162425, 'samples': 10275328, 'steps': 20068, 'loss/train': 3.2345340251922607} 02/25/2022 03:34:28 - INFO - codeparrot_training - Step 20069: {'lr': 0.0003446268169518494, 'samples': 10275840, 'steps': 20069, 'loss/train': 1.928001046180725} 02/25/2022 03:34:33 - INFO - codeparrot_training - Step 20070: {'lr': 0.00034461167168210494, 'samples': 10276352, 'steps': 20070, 'loss/train': 1.7225770950317383} 02/25/2022 03:34:37 - INFO - codeparrot_training - Step 20071: {'lr': 0.00034459652600707423, 'samples': 10276864, 'steps': 20071, 'loss/train': 2.394929885864258} 02/25/2022 03:34:42 - INFO - codeparrot_training - Step 20072: {'lr': 0.000344581379926822, 'samples': 10277376, 'steps': 20072, 'loss/train': 2.9181714057922363} 02/25/2022 03:34:46 - INFO - codeparrot_training - Step 20073: {'lr': 0.0003445662334414131, 'samples': 10277888, 'steps': 20073, 'loss/train': 1.212462306022644} 02/25/2022 03:34:52 - INFO - codeparrot_training - Step 20074: {'lr': 0.00034455108655091256, 'samples': 10278400, 'steps': 20074, 'loss/train': 1.367550253868103} 02/25/2022 03:34:55 - INFO - codeparrot_training - Step 20075: {'lr': 0.00034453593925538515, 'samples': 10278912, 'steps': 20075, 'loss/train': 2.1880338191986084} 02/25/2022 03:35:01 - INFO - codeparrot_training - Step 20076: {'lr': 0.00034452079155489586, 'samples': 10279424, 'steps': 20076, 'loss/train': 1.9069173336029053} 02/25/2022 03:35:04 - INFO - codeparrot_training - Step 20077: {'lr': 0.00034450564344950944, 'samples': 10279936, 'steps': 20077, 'loss/train': 2.4891440868377686} 02/25/2022 03:35:10 - INFO - codeparrot_training - Step 20078: {'lr': 0.00034449049493929086, 'samples': 10280448, 'steps': 20078, 'loss/train': 1.4974418878555298} 02/25/2022 03:35:13 - INFO - codeparrot_training - Step 20079: {'lr': 0.00034447534602430503, 'samples': 10280960, 'steps': 20079, 'loss/train': 1.76010000705719} 02/25/2022 03:35:19 - INFO - codeparrot_training - Step 20080: {'lr': 0.00034446019670461683, 'samples': 10281472, 'steps': 20080, 'loss/train': 1.767886757850647} 02/25/2022 03:35:22 - INFO - codeparrot_training - Step 20081: {'lr': 0.0003444450469802911, 'samples': 10281984, 'steps': 20081, 'loss/train': 2.418989419937134} 02/25/2022 03:35:28 - INFO - codeparrot_training - Step 20082: {'lr': 0.0003444298968513928, 'samples': 10282496, 'steps': 20082, 'loss/train': 1.963512659072876} 02/25/2022 03:35:31 - INFO - codeparrot_training - Step 20083: {'lr': 0.0003444147463179868, 'samples': 10283008, 'steps': 20083, 'loss/train': 2.308830499649048} 02/25/2022 03:35:38 - INFO - codeparrot_training - Step 20084: {'lr': 0.00034439959538013805, 'samples': 10283520, 'steps': 20084, 'loss/train': 1.3150640726089478} 02/25/2022 03:35:41 - INFO - codeparrot_training - Step 20085: {'lr': 0.00034438444403791135, 'samples': 10284032, 'steps': 20085, 'loss/train': 2.6701104640960693} 02/25/2022 03:35:47 - INFO - codeparrot_training - Step 20086: {'lr': 0.00034436929229137163, 'samples': 10284544, 'steps': 20086, 'loss/train': 0.08370508253574371} 02/25/2022 03:35:50 - INFO - codeparrot_training - Step 20087: {'lr': 0.00034435414014058393, 'samples': 10285056, 'steps': 20087, 'loss/train': 2.3114542961120605} 02/25/2022 03:35:56 - INFO - codeparrot_training - Step 20088: {'lr': 0.000344338987585613, 'samples': 10285568, 'steps': 20088, 'loss/train': 2.2398841381073} 02/25/2022 03:35:59 - INFO - codeparrot_training - Step 20089: {'lr': 0.0003443238346265238, 'samples': 10286080, 'steps': 20089, 'loss/train': 1.71213698387146} 02/25/2022 03:36:05 - INFO - codeparrot_training - Step 20090: {'lr': 0.0003443086812633812, 'samples': 10286592, 'steps': 20090, 'loss/train': 2.1920006275177} 02/25/2022 03:36:08 - INFO - codeparrot_training - Step 20091: {'lr': 0.00034429352749625026, 'samples': 10287104, 'steps': 20091, 'loss/train': 2.539146661758423} 02/25/2022 03:36:14 - INFO - codeparrot_training - Step 20092: {'lr': 0.00034427837332519573, 'samples': 10287616, 'steps': 20092, 'loss/train': 2.2923905849456787} 02/25/2022 03:36:17 - INFO - codeparrot_training - Step 20093: {'lr': 0.0003442632187502826, 'samples': 10288128, 'steps': 20093, 'loss/train': 1.10002863407135} 02/25/2022 03:36:23 - INFO - codeparrot_training - Step 20094: {'lr': 0.00034424806377157576, 'samples': 10288640, 'steps': 20094, 'loss/train': 1.6613848209381104} 02/25/2022 03:36:26 - INFO - codeparrot_training - Step 20095: {'lr': 0.0003442329083891402, 'samples': 10289152, 'steps': 20095, 'loss/train': 0.9307950139045715} 02/25/2022 03:36:32 - INFO - codeparrot_training - Step 20096: {'lr': 0.00034421775260304067, 'samples': 10289664, 'steps': 20096, 'loss/train': 1.4342178106307983} 02/25/2022 03:36:36 - INFO - codeparrot_training - Step 20097: {'lr': 0.0003442025964133422, 'samples': 10290176, 'steps': 20097, 'loss/train': 2.288285732269287} 02/25/2022 03:36:41 - INFO - codeparrot_training - Step 20098: {'lr': 0.0003441874398201099, 'samples': 10290688, 'steps': 20098, 'loss/train': 2.137441396713257} 02/25/2022 03:36:45 - INFO - codeparrot_training - Step 20099: {'lr': 0.00034417228282340837, 'samples': 10291200, 'steps': 20099, 'loss/train': 1.7621381282806396} 02/25/2022 03:36:52 - INFO - codeparrot_training - Step 20100: {'lr': 0.0003441571254233027, 'samples': 10291712, 'steps': 20100, 'loss/train': 1.829404592514038} 02/25/2022 03:36:55 - INFO - codeparrot_training - Step 20101: {'lr': 0.00034414196761985784, 'samples': 10292224, 'steps': 20101, 'loss/train': 2.217130422592163} 02/25/2022 03:37:01 - INFO - codeparrot_training - Step 20102: {'lr': 0.00034412680941313866, 'samples': 10292736, 'steps': 20102, 'loss/train': 1.9657161235809326} 02/25/2022 03:37:04 - INFO - codeparrot_training - Step 20103: {'lr': 0.00034411165080321007, 'samples': 10293248, 'steps': 20103, 'loss/train': 2.4528114795684814} 02/25/2022 03:37:10 - INFO - codeparrot_training - Step 20104: {'lr': 0.00034409649179013716, 'samples': 10293760, 'steps': 20104, 'loss/train': 2.3077969551086426} 02/25/2022 03:37:13 - INFO - codeparrot_training - Step 20105: {'lr': 0.00034408133237398466, 'samples': 10294272, 'steps': 20105, 'loss/train': 1.5373939275741577} 02/25/2022 03:37:19 - INFO - codeparrot_training - Step 20106: {'lr': 0.0003440661725548176, 'samples': 10294784, 'steps': 20106, 'loss/train': 0.37341976165771484} 02/25/2022 03:37:22 - INFO - codeparrot_training - Step 20107: {'lr': 0.00034405101233270105, 'samples': 10295296, 'steps': 20107, 'loss/train': 1.6152325868606567} 02/25/2022 03:37:28 - INFO - codeparrot_training - Step 20108: {'lr': 0.0003440358517076997, 'samples': 10295808, 'steps': 20108, 'loss/train': 2.2198026180267334} 02/25/2022 03:37:31 - INFO - codeparrot_training - Step 20109: {'lr': 0.00034402069067987874, 'samples': 10296320, 'steps': 20109, 'loss/train': 2.169184684753418} 02/25/2022 03:37:37 - INFO - codeparrot_training - Step 20110: {'lr': 0.0003440055292493029, 'samples': 10296832, 'steps': 20110, 'loss/train': 1.4872303009033203} 02/25/2022 03:37:40 - INFO - codeparrot_training - Step 20111: {'lr': 0.0003439903674160373, 'samples': 10297344, 'steps': 20111, 'loss/train': 2.643341064453125} 02/25/2022 03:37:47 - INFO - codeparrot_training - Step 20112: {'lr': 0.0003439752051801467, 'samples': 10297856, 'steps': 20112, 'loss/train': 2.191432237625122} 02/25/2022 03:37:50 - INFO - codeparrot_training - Step 20113: {'lr': 0.0003439600425416963, 'samples': 10298368, 'steps': 20113, 'loss/train': 3.3072595596313477} 02/25/2022 03:37:56 - INFO - codeparrot_training - Step 20114: {'lr': 0.00034394487950075076, 'samples': 10298880, 'steps': 20114, 'loss/train': 2.21701717376709} 02/25/2022 03:37:59 - INFO - codeparrot_training - Step 20115: {'lr': 0.0003439297160573753, 'samples': 10299392, 'steps': 20115, 'loss/train': 2.4579732418060303} 02/25/2022 03:38:05 - INFO - codeparrot_training - Step 20116: {'lr': 0.0003439145522116347, 'samples': 10299904, 'steps': 20116, 'loss/train': 1.6692771911621094} 02/25/2022 03:38:08 - INFO - codeparrot_training - Step 20117: {'lr': 0.000343899387963594, 'samples': 10300416, 'steps': 20117, 'loss/train': 2.643324613571167} 02/25/2022 03:38:14 - INFO - codeparrot_training - Step 20118: {'lr': 0.00034388422331331817, 'samples': 10300928, 'steps': 20118, 'loss/train': 3.2394938468933105} 02/25/2022 03:38:18 - INFO - codeparrot_training - Step 20119: {'lr': 0.0003438690582608721, 'samples': 10301440, 'steps': 20119, 'loss/train': 1.951217770576477} 02/25/2022 03:38:23 - INFO - codeparrot_training - Step 20120: {'lr': 0.00034385389280632077, 'samples': 10301952, 'steps': 20120, 'loss/train': 2.131113052368164} 02/25/2022 03:38:27 - INFO - codeparrot_training - Step 20121: {'lr': 0.00034383872694972916, 'samples': 10302464, 'steps': 20121, 'loss/train': 1.0880255699157715} 02/25/2022 03:38:33 - INFO - codeparrot_training - Step 20122: {'lr': 0.0003438235606911623, 'samples': 10302976, 'steps': 20122, 'loss/train': 0.5588898658752441} 02/25/2022 03:38:37 - INFO - codeparrot_training - Step 20123: {'lr': 0.0003438083940306851, 'samples': 10303488, 'steps': 20123, 'loss/train': 1.6169017553329468} 02/25/2022 03:38:42 - INFO - codeparrot_training - Step 20124: {'lr': 0.0003437932269683625, 'samples': 10304000, 'steps': 20124, 'loss/train': 0.9220380783081055} 02/25/2022 03:38:46 - INFO - codeparrot_training - Step 20125: {'lr': 0.0003437780595042595, 'samples': 10304512, 'steps': 20125, 'loss/train': 3.7603652477264404} 02/25/2022 03:38:52 - INFO - codeparrot_training - Step 20126: {'lr': 0.0003437628916384411, 'samples': 10305024, 'steps': 20126, 'loss/train': 2.475961685180664} 02/25/2022 03:38:55 - INFO - codeparrot_training - Step 20127: {'lr': 0.0003437477233709722, 'samples': 10305536, 'steps': 20127, 'loss/train': 2.18354868888855} 02/25/2022 03:39:01 - INFO - codeparrot_training - Step 20128: {'lr': 0.0003437325547019179, 'samples': 10306048, 'steps': 20128, 'loss/train': 1.8523273468017578} 02/25/2022 03:39:04 - INFO - codeparrot_training - Step 20129: {'lr': 0.000343717385631343, 'samples': 10306560, 'steps': 20129, 'loss/train': 0.9113268256187439} 02/25/2022 03:39:10 - INFO - codeparrot_training - Step 20130: {'lr': 0.00034370221615931265, 'samples': 10307072, 'steps': 20130, 'loss/train': 2.594956874847412} 02/25/2022 03:39:13 - INFO - codeparrot_training - Step 20131: {'lr': 0.0003436870462858917, 'samples': 10307584, 'steps': 20131, 'loss/train': 2.4714298248291016} 02/25/2022 03:39:19 - INFO - codeparrot_training - Step 20132: {'lr': 0.0003436718760111452, 'samples': 10308096, 'steps': 20132, 'loss/train': 2.5058271884918213} 02/25/2022 03:39:23 - INFO - codeparrot_training - Step 20133: {'lr': 0.00034365670533513813, 'samples': 10308608, 'steps': 20133, 'loss/train': 1.8360416889190674} 02/25/2022 03:39:28 - INFO - codeparrot_training - Step 20134: {'lr': 0.00034364153425793547, 'samples': 10309120, 'steps': 20134, 'loss/train': 1.4394134283065796} 02/25/2022 03:39:32 - INFO - codeparrot_training - Step 20135: {'lr': 0.0003436263627796023, 'samples': 10309632, 'steps': 20135, 'loss/train': 2.3129072189331055} 02/25/2022 03:39:38 - INFO - codeparrot_training - Step 20136: {'lr': 0.00034361119090020343, 'samples': 10310144, 'steps': 20136, 'loss/train': 1.7954285144805908} 02/25/2022 03:39:41 - INFO - codeparrot_training - Step 20137: {'lr': 0.000343596018619804, 'samples': 10310656, 'steps': 20137, 'loss/train': 2.490025281906128} 02/25/2022 03:39:47 - INFO - codeparrot_training - Step 20138: {'lr': 0.00034358084593846886, 'samples': 10311168, 'steps': 20138, 'loss/train': 2.0469114780426025} 02/25/2022 03:39:50 - INFO - codeparrot_training - Step 20139: {'lr': 0.00034356567285626316, 'samples': 10311680, 'steps': 20139, 'loss/train': 2.0434672832489014} 02/25/2022 03:39:56 - INFO - codeparrot_training - Step 20140: {'lr': 0.0003435504993732518, 'samples': 10312192, 'steps': 20140, 'loss/train': 1.5098412036895752} 02/25/2022 03:39:59 - INFO - codeparrot_training - Step 20141: {'lr': 0.00034353532548949984, 'samples': 10312704, 'steps': 20141, 'loss/train': 2.270737648010254} 02/25/2022 03:40:05 - INFO - codeparrot_training - Step 20142: {'lr': 0.0003435201512050722, 'samples': 10313216, 'steps': 20142, 'loss/train': 1.4375869035720825} 02/25/2022 03:40:08 - INFO - codeparrot_training - Step 20143: {'lr': 0.000343504976520034, 'samples': 10313728, 'steps': 20143, 'loss/train': 1.4480029344558716} 02/25/2022 03:40:14 - INFO - codeparrot_training - Step 20144: {'lr': 0.0003434898014344501, 'samples': 10314240, 'steps': 20144, 'loss/train': 1.4094799757003784} 02/25/2022 03:40:17 - INFO - codeparrot_training - Step 20145: {'lr': 0.00034347462594838565, 'samples': 10314752, 'steps': 20145, 'loss/train': 1.9889744520187378} 02/25/2022 03:40:23 - INFO - codeparrot_training - Step 20146: {'lr': 0.0003434594500619055, 'samples': 10315264, 'steps': 20146, 'loss/train': 0.23489277064800262} 02/25/2022 03:40:26 - INFO - codeparrot_training - Step 20147: {'lr': 0.00034344427377507475, 'samples': 10315776, 'steps': 20147, 'loss/train': 1.9060548543930054} 02/25/2022 03:40:33 - INFO - codeparrot_training - Step 20148: {'lr': 0.00034342909708795846, 'samples': 10316288, 'steps': 20148, 'loss/train': 2.104743003845215} 02/25/2022 03:40:36 - INFO - codeparrot_training - Step 20149: {'lr': 0.0003434139200006216, 'samples': 10316800, 'steps': 20149, 'loss/train': 2.5730905532836914} 02/25/2022 03:40:40 - INFO - codeparrot_training - Step 20150: {'lr': 0.0003433987425131291, 'samples': 10317312, 'steps': 20150, 'loss/train': 0.40831297636032104} 02/25/2022 03:40:45 - INFO - codeparrot_training - Step 20151: {'lr': 0.0003433835646255461, 'samples': 10317824, 'steps': 20151, 'loss/train': 2.559032440185547} 02/25/2022 03:40:49 - INFO - codeparrot_training - Step 20152: {'lr': 0.0003433683863379375, 'samples': 10318336, 'steps': 20152, 'loss/train': 2.9120593070983887} 02/25/2022 03:40:54 - INFO - codeparrot_training - Step 20153: {'lr': 0.00034335320765036843, 'samples': 10318848, 'steps': 20153, 'loss/train': 1.8782655000686646} 02/25/2022 03:40:58 - INFO - codeparrot_training - Step 20154: {'lr': 0.0003433380285629039, 'samples': 10319360, 'steps': 20154, 'loss/train': 2.706453323364258} 02/25/2022 03:41:03 - INFO - codeparrot_training - Step 20155: {'lr': 0.0003433228490756088, 'samples': 10319872, 'steps': 20155, 'loss/train': 1.754799246788025} 02/25/2022 03:41:06 - INFO - codeparrot_training - Step 20156: {'lr': 0.00034330766918854827, 'samples': 10320384, 'steps': 20156, 'loss/train': 3.3620426654815674} 02/25/2022 03:41:14 - INFO - codeparrot_training - Step 20157: {'lr': 0.0003432924889017873, 'samples': 10320896, 'steps': 20157, 'loss/train': 1.7797236442565918} 02/25/2022 03:41:18 - INFO - codeparrot_training - Step 20158: {'lr': 0.000343277308215391, 'samples': 10321408, 'steps': 20158, 'loss/train': 1.5693120956420898} 02/25/2022 03:41:23 - INFO - codeparrot_training - Step 20159: {'lr': 0.0003432621271294243, 'samples': 10321920, 'steps': 20159, 'loss/train': 2.9955568313598633} 02/25/2022 03:41:27 - INFO - codeparrot_training - Step 20160: {'lr': 0.00034324694564395226, 'samples': 10322432, 'steps': 20160, 'loss/train': 2.8591437339782715} 02/25/2022 03:41:32 - INFO - codeparrot_training - Step 20161: {'lr': 0.0003432317637590399, 'samples': 10322944, 'steps': 20161, 'loss/train': 2.9914727210998535} 02/25/2022 03:41:36 - INFO - codeparrot_training - Step 20162: {'lr': 0.0003432165814747523, 'samples': 10323456, 'steps': 20162, 'loss/train': 0.8629205226898193} 02/25/2022 03:41:41 - INFO - codeparrot_training - Step 20163: {'lr': 0.0003432013987911544, 'samples': 10323968, 'steps': 20163, 'loss/train': 2.3496475219726562} 02/25/2022 03:41:45 - INFO - codeparrot_training - Step 20164: {'lr': 0.0003431862157083114, 'samples': 10324480, 'steps': 20164, 'loss/train': 3.192779541015625} 02/25/2022 03:41:50 - INFO - codeparrot_training - Step 20165: {'lr': 0.0003431710322262882, 'samples': 10324992, 'steps': 20165, 'loss/train': 1.9700579643249512} 02/25/2022 03:41:58 - INFO - codeparrot_training - Step 20166: {'lr': 0.0003431558483451498, 'samples': 10325504, 'steps': 20166, 'loss/train': 1.953985571861267} 02/25/2022 03:42:01 - INFO - codeparrot_training - Step 20167: {'lr': 0.00034314066406496146, 'samples': 10326016, 'steps': 20167, 'loss/train': 2.1392829418182373} 02/25/2022 03:42:07 - INFO - codeparrot_training - Step 20168: {'lr': 0.00034312547938578796, 'samples': 10326528, 'steps': 20168, 'loss/train': 1.6107432842254639} 02/25/2022 03:42:10 - INFO - codeparrot_training - Step 20169: {'lr': 0.0003431102943076946, 'samples': 10327040, 'steps': 20169, 'loss/train': 2.2947282791137695} 02/25/2022 03:42:16 - INFO - codeparrot_training - Step 20170: {'lr': 0.00034309510883074625, 'samples': 10327552, 'steps': 20170, 'loss/train': 2.008180856704712} 02/25/2022 03:42:19 - INFO - codeparrot_training - Step 20171: {'lr': 0.0003430799229550079, 'samples': 10328064, 'steps': 20171, 'loss/train': 1.8287910223007202} 02/25/2022 03:42:25 - INFO - codeparrot_training - Step 20172: {'lr': 0.0003430647366805449, 'samples': 10328576, 'steps': 20172, 'loss/train': 0.12379389256238937} 02/25/2022 03:42:28 - INFO - codeparrot_training - Step 20173: {'lr': 0.000343049550007422, 'samples': 10329088, 'steps': 20173, 'loss/train': 2.1541004180908203} 02/25/2022 03:42:34 - INFO - codeparrot_training - Step 20174: {'lr': 0.0003430343629357044, 'samples': 10329600, 'steps': 20174, 'loss/train': 0.17658719420433044} 02/25/2022 03:42:37 - INFO - codeparrot_training - Step 20175: {'lr': 0.0003430191754654572, 'samples': 10330112, 'steps': 20175, 'loss/train': 2.2133290767669678} 02/25/2022 03:42:43 - INFO - codeparrot_training - Step 20176: {'lr': 0.0003430039875967454, 'samples': 10330624, 'steps': 20176, 'loss/train': 2.2580955028533936} 02/25/2022 03:42:46 - INFO - codeparrot_training - Step 20177: {'lr': 0.00034298879932963397, 'samples': 10331136, 'steps': 20177, 'loss/train': 5.419597148895264} 02/25/2022 03:42:53 - INFO - codeparrot_training - Step 20178: {'lr': 0.0003429736106641881, 'samples': 10331648, 'steps': 20178, 'loss/train': 1.6949840784072876} 02/25/2022 03:42:57 - INFO - codeparrot_training - Step 20179: {'lr': 0.0003429584216004728, 'samples': 10332160, 'steps': 20179, 'loss/train': 2.1994357109069824} 02/25/2022 03:43:02 - INFO - codeparrot_training - Step 20180: {'lr': 0.0003429432321385531, 'samples': 10332672, 'steps': 20180, 'loss/train': 1.9700783491134644} 02/25/2022 03:43:06 - INFO - codeparrot_training - Step 20181: {'lr': 0.00034292804227849407, 'samples': 10333184, 'steps': 20181, 'loss/train': 2.1304335594177246} 02/25/2022 03:43:11 - INFO - codeparrot_training - Step 20182: {'lr': 0.0003429128520203608, 'samples': 10333696, 'steps': 20182, 'loss/train': 1.1833218336105347} 02/25/2022 03:43:15 - INFO - codeparrot_training - Step 20183: {'lr': 0.00034289766136421854, 'samples': 10334208, 'steps': 20183, 'loss/train': 2.419492244720459} 02/25/2022 03:43:20 - INFO - codeparrot_training - Step 20184: {'lr': 0.000342882470310132, 'samples': 10334720, 'steps': 20184, 'loss/train': 2.2310307025909424} 02/25/2022 03:43:24 - INFO - codeparrot_training - Step 20185: {'lr': 0.0003428672788581666, 'samples': 10335232, 'steps': 20185, 'loss/train': 2.2726104259490967} 02/25/2022 03:43:29 - INFO - codeparrot_training - Step 20186: {'lr': 0.0003428520870083872, 'samples': 10335744, 'steps': 20186, 'loss/train': 1.7123125791549683} 02/25/2022 03:43:33 - INFO - codeparrot_training - Step 20187: {'lr': 0.000342836894760859, 'samples': 10336256, 'steps': 20187, 'loss/train': 1.6534618139266968} 02/25/2022 03:43:40 - INFO - codeparrot_training - Step 20188: {'lr': 0.00034282170211564697, 'samples': 10336768, 'steps': 20188, 'loss/train': 1.998224139213562} 02/25/2022 03:43:44 - INFO - codeparrot_training - Step 20189: {'lr': 0.0003428065090728163, 'samples': 10337280, 'steps': 20189, 'loss/train': 2.199709415435791} 02/25/2022 03:43:49 - INFO - codeparrot_training - Step 20190: {'lr': 0.0003427913156324319, 'samples': 10337792, 'steps': 20190, 'loss/train': 1.9417345523834229} 02/25/2022 03:43:53 - INFO - codeparrot_training - Step 20191: {'lr': 0.00034277612179455907, 'samples': 10338304, 'steps': 20191, 'loss/train': 1.9399032592773438} 02/25/2022 03:43:58 - INFO - codeparrot_training - Step 20192: {'lr': 0.00034276092755926275, 'samples': 10338816, 'steps': 20192, 'loss/train': 0.8594242930412292} 02/25/2022 03:44:02 - INFO - codeparrot_training - Step 20193: {'lr': 0.0003427457329266081, 'samples': 10339328, 'steps': 20193, 'loss/train': 0.14682406187057495} 02/25/2022 03:44:07 - INFO - codeparrot_training - Step 20194: {'lr': 0.0003427305378966601, 'samples': 10339840, 'steps': 20194, 'loss/train': 1.7434056997299194} 02/25/2022 03:44:11 - INFO - codeparrot_training - Step 20195: {'lr': 0.00034271534246948403, 'samples': 10340352, 'steps': 20195, 'loss/train': 2.130207061767578} 02/25/2022 03:44:16 - INFO - codeparrot_training - Step 20196: {'lr': 0.0003427001466451448, 'samples': 10340864, 'steps': 20196, 'loss/train': 1.4441578388214111} 02/25/2022 03:44:20 - INFO - codeparrot_training - Step 20197: {'lr': 0.00034268495042370767, 'samples': 10341376, 'steps': 20197, 'loss/train': 2.3099284172058105} 02/25/2022 03:44:25 - INFO - codeparrot_training - Step 20198: {'lr': 0.00034266975380523756, 'samples': 10341888, 'steps': 20198, 'loss/train': 2.0511014461517334} 02/25/2022 03:44:29 - INFO - codeparrot_training - Step 20199: {'lr': 0.00034265455678979967, 'samples': 10342400, 'steps': 20199, 'loss/train': 1.2763123512268066} 02/25/2022 03:44:34 - INFO - codeparrot_training - Step 20200: {'lr': 0.0003426393593774591, 'samples': 10342912, 'steps': 20200, 'loss/train': 2.256762742996216} 02/25/2022 03:44:38 - INFO - codeparrot_training - Step 20201: {'lr': 0.0003426241615682809, 'samples': 10343424, 'steps': 20201, 'loss/train': 1.6343567371368408} 02/25/2022 03:44:43 - INFO - codeparrot_training - Step 20202: {'lr': 0.0003426089633623302, 'samples': 10343936, 'steps': 20202, 'loss/train': 1.9044687747955322} 02/25/2022 03:44:47 - INFO - codeparrot_training - Step 20203: {'lr': 0.0003425937647596721, 'samples': 10344448, 'steps': 20203, 'loss/train': 2.21189546585083} 02/25/2022 03:44:54 - INFO - codeparrot_training - Step 20204: {'lr': 0.0003425785657603718, 'samples': 10344960, 'steps': 20204, 'loss/train': 2.301870107650757} 02/25/2022 03:44:57 - INFO - codeparrot_training - Step 20205: {'lr': 0.0003425633663644942, 'samples': 10345472, 'steps': 20205, 'loss/train': 1.391209602355957} 02/25/2022 03:45:03 - INFO - codeparrot_training - Step 20206: {'lr': 0.00034254816657210455, 'samples': 10345984, 'steps': 20206, 'loss/train': 3.39180850982666} 02/25/2022 03:45:06 - INFO - codeparrot_training - Step 20207: {'lr': 0.00034253296638326805, 'samples': 10346496, 'steps': 20207, 'loss/train': 2.0950658321380615} 02/25/2022 03:45:12 - INFO - codeparrot_training - Step 20208: {'lr': 0.0003425177657980496, 'samples': 10347008, 'steps': 20208, 'loss/train': 0.19738437235355377} 02/25/2022 03:45:16 - INFO - codeparrot_training - Step 20209: {'lr': 0.0003425025648165145, 'samples': 10347520, 'steps': 20209, 'loss/train': 2.274714708328247} 02/25/2022 03:45:21 - INFO - codeparrot_training - Step 20210: {'lr': 0.00034248736343872767, 'samples': 10348032, 'steps': 20210, 'loss/train': 1.759999394416809} 02/25/2022 03:45:25 - INFO - codeparrot_training - Step 20211: {'lr': 0.0003424721616647544, 'samples': 10348544, 'steps': 20211, 'loss/train': 2.694755792617798} 02/25/2022 03:45:30 - INFO - codeparrot_training - Step 20212: {'lr': 0.00034245695949465977, 'samples': 10349056, 'steps': 20212, 'loss/train': 2.084501028060913} 02/25/2022 03:45:34 - INFO - codeparrot_training - Step 20213: {'lr': 0.00034244175692850894, 'samples': 10349568, 'steps': 20213, 'loss/train': 1.7472513914108276} 02/25/2022 03:45:41 - INFO - codeparrot_training - Step 20214: {'lr': 0.00034242655396636687, 'samples': 10350080, 'steps': 20214, 'loss/train': 1.5567471981048584} 02/25/2022 03:45:44 - INFO - codeparrot_training - Step 20215: {'lr': 0.0003424113506082989, 'samples': 10350592, 'steps': 20215, 'loss/train': 1.5938544273376465} 02/25/2022 03:45:50 - INFO - codeparrot_training - Step 20216: {'lr': 0.00034239614685436994, 'samples': 10351104, 'steps': 20216, 'loss/train': 1.9916096925735474} 02/25/2022 03:45:53 - INFO - codeparrot_training - Step 20217: {'lr': 0.00034238094270464523, 'samples': 10351616, 'steps': 20217, 'loss/train': 1.6987683773040771} 02/25/2022 03:45:59 - INFO - codeparrot_training - Step 20218: {'lr': 0.00034236573815918993, 'samples': 10352128, 'steps': 20218, 'loss/train': 1.4841560125350952} 02/25/2022 03:46:02 - INFO - codeparrot_training - Step 20219: {'lr': 0.00034235053321806915, 'samples': 10352640, 'steps': 20219, 'loss/train': 1.3522506952285767} 02/25/2022 03:46:08 - INFO - codeparrot_training - Step 20220: {'lr': 0.00034233532788134803, 'samples': 10353152, 'steps': 20220, 'loss/train': 4.422814846038818} 02/25/2022 03:46:11 - INFO - codeparrot_training - Step 20221: {'lr': 0.0003423201221490916, 'samples': 10353664, 'steps': 20221, 'loss/train': 1.8960996866226196} 02/25/2022 03:46:17 - INFO - codeparrot_training - Step 20222: {'lr': 0.00034230491602136513, 'samples': 10354176, 'steps': 20222, 'loss/train': 2.271332263946533} 02/25/2022 03:46:20 - INFO - codeparrot_training - Step 20223: {'lr': 0.0003422897094982337, 'samples': 10354688, 'steps': 20223, 'loss/train': 2.104485511779785} 02/25/2022 03:46:28 - INFO - codeparrot_training - Step 20224: {'lr': 0.0003422745025797626, 'samples': 10355200, 'steps': 20224, 'loss/train': 2.424830436706543} 02/25/2022 03:46:31 - INFO - codeparrot_training - Step 20225: {'lr': 0.00034225929526601664, 'samples': 10355712, 'steps': 20225, 'loss/train': 1.6067159175872803} 02/25/2022 03:46:37 - INFO - codeparrot_training - Step 20226: {'lr': 0.0003422440875570612, 'samples': 10356224, 'steps': 20226, 'loss/train': 1.3527050018310547} 02/25/2022 03:46:40 - INFO - codeparrot_training - Step 20227: {'lr': 0.00034222887945296144, 'samples': 10356736, 'steps': 20227, 'loss/train': 2.335644245147705} 02/25/2022 03:46:45 - INFO - codeparrot_training - Step 20228: {'lr': 0.0003422136709537824, 'samples': 10357248, 'steps': 20228, 'loss/train': 1.4278925657272339} 02/25/2022 03:46:49 - INFO - codeparrot_training - Step 20229: {'lr': 0.00034219846205958926, 'samples': 10357760, 'steps': 20229, 'loss/train': 1.9828650951385498} 02/25/2022 03:46:54 - INFO - codeparrot_training - Step 20230: {'lr': 0.0003421832527704471, 'samples': 10358272, 'steps': 20230, 'loss/train': 0.9953842163085938} 02/25/2022 03:46:58 - INFO - codeparrot_training - Step 20231: {'lr': 0.0003421680430864214, 'samples': 10358784, 'steps': 20231, 'loss/train': 1.9220088720321655} 02/25/2022 03:47:03 - INFO - codeparrot_training - Step 20232: {'lr': 0.0003421528330075769, 'samples': 10359296, 'steps': 20232, 'loss/train': 2.1465003490448} 02/25/2022 03:47:07 - INFO - codeparrot_training - Step 20233: {'lr': 0.00034213762253397896, 'samples': 10359808, 'steps': 20233, 'loss/train': 2.023115873336792} 02/25/2022 03:47:14 - INFO - codeparrot_training - Step 20234: {'lr': 0.0003421224116656927, 'samples': 10360320, 'steps': 20234, 'loss/train': 1.7697272300720215} 02/25/2022 03:47:18 - INFO - codeparrot_training - Step 20235: {'lr': 0.0003421072004027833, 'samples': 10360832, 'steps': 20235, 'loss/train': 0.3598443865776062} 02/25/2022 03:47:23 - INFO - codeparrot_training - Step 20236: {'lr': 0.00034209198874531586, 'samples': 10361344, 'steps': 20236, 'loss/train': 1.957742691040039} 02/25/2022 03:47:27 - INFO - codeparrot_training - Step 20237: {'lr': 0.00034207677669335565, 'samples': 10361856, 'steps': 20237, 'loss/train': 2.067368268966675} 02/25/2022 03:47:32 - INFO - codeparrot_training - Step 20238: {'lr': 0.0003420615642469678, 'samples': 10362368, 'steps': 20238, 'loss/train': 2.1140098571777344} 02/25/2022 03:47:36 - INFO - codeparrot_training - Step 20239: {'lr': 0.00034204635140621726, 'samples': 10362880, 'steps': 20239, 'loss/train': 1.9461902379989624} 02/25/2022 03:47:41 - INFO - codeparrot_training - Step 20240: {'lr': 0.00034203113817116957, 'samples': 10363392, 'steps': 20240, 'loss/train': 1.085320234298706} 02/25/2022 03:47:47 - INFO - codeparrot_training - Step 20241: {'lr': 0.0003420159245418896, 'samples': 10363904, 'steps': 20241, 'loss/train': 0.9491880536079407} 02/25/2022 03:47:50 - INFO - codeparrot_training - Step 20242: {'lr': 0.0003420007105184426, 'samples': 10364416, 'steps': 20242, 'loss/train': 1.3015691041946411} 02/25/2022 03:47:57 - INFO - codeparrot_training - Step 20243: {'lr': 0.0003419854961008938, 'samples': 10364928, 'steps': 20243, 'loss/train': 1.5790596008300781} 02/25/2022 03:48:01 - INFO - codeparrot_training - Step 20244: {'lr': 0.0003419702812893084, 'samples': 10365440, 'steps': 20244, 'loss/train': 1.3433915376663208} 02/25/2022 03:48:04 - INFO - codeparrot_training - Step 20245: {'lr': 0.0003419550660837515, 'samples': 10365952, 'steps': 20245, 'loss/train': 2.1597776412963867} 02/25/2022 03:48:10 - INFO - codeparrot_training - Step 20246: {'lr': 0.0003419398504842883, 'samples': 10366464, 'steps': 20246, 'loss/train': 3.4824254512786865} 02/25/2022 03:48:13 - INFO - codeparrot_training - Step 20247: {'lr': 0.00034192463449098386, 'samples': 10366976, 'steps': 20247, 'loss/train': 3.133544445037842} 02/25/2022 03:48:19 - INFO - codeparrot_training - Step 20248: {'lr': 0.00034190941810390365, 'samples': 10367488, 'steps': 20248, 'loss/train': 1.955594539642334} 02/25/2022 03:48:22 - INFO - codeparrot_training - Step 20249: {'lr': 0.00034189420132311256, 'samples': 10368000, 'steps': 20249, 'loss/train': 2.202923536300659} 02/25/2022 03:48:28 - INFO - codeparrot_training - Step 20250: {'lr': 0.0003418789841486759, 'samples': 10368512, 'steps': 20250, 'loss/train': 2.6444976329803467} 02/25/2022 03:48:31 - INFO - codeparrot_training - Step 20251: {'lr': 0.0003418637665806589, 'samples': 10369024, 'steps': 20251, 'loss/train': 2.010192632675171} 02/25/2022 03:48:38 - INFO - codeparrot_training - Step 20252: {'lr': 0.0003418485486191267, 'samples': 10369536, 'steps': 20252, 'loss/train': 2.4391818046569824} 02/25/2022 03:48:41 - INFO - codeparrot_training - Step 20253: {'lr': 0.0003418333302641444, 'samples': 10370048, 'steps': 20253, 'loss/train': 2.3971216678619385} 02/25/2022 03:48:47 - INFO - codeparrot_training - Step 20254: {'lr': 0.0003418181115157774, 'samples': 10370560, 'steps': 20254, 'loss/train': 0.6539112329483032} 02/25/2022 03:48:50 - INFO - codeparrot_training - Step 20255: {'lr': 0.00034180289237409063, 'samples': 10371072, 'steps': 20255, 'loss/train': 1.6873406171798706} 02/25/2022 03:48:56 - INFO - codeparrot_training - Step 20256: {'lr': 0.00034178767283914944, 'samples': 10371584, 'steps': 20256, 'loss/train': 1.764180064201355} 02/25/2022 03:48:59 - INFO - codeparrot_training - Step 20257: {'lr': 0.000341772452911019, 'samples': 10372096, 'steps': 20257, 'loss/train': 1.1096210479736328} 02/25/2022 03:49:05 - INFO - codeparrot_training - Step 20258: {'lr': 0.0003417572325897646, 'samples': 10372608, 'steps': 20258, 'loss/train': 1.8763506412506104} 02/25/2022 03:49:08 - INFO - codeparrot_training - Step 20259: {'lr': 0.00034174201187545133, 'samples': 10373120, 'steps': 20259, 'loss/train': 1.6311677694320679} 02/25/2022 03:49:16 - INFO - codeparrot_training - Step 20260: {'lr': 0.00034172679076814437, 'samples': 10373632, 'steps': 20260, 'loss/train': 2.2753138542175293} 02/25/2022 03:49:19 - INFO - codeparrot_training - Step 20261: {'lr': 0.00034171156926790904, 'samples': 10374144, 'steps': 20261, 'loss/train': 2.1644623279571533} 02/25/2022 03:49:25 - INFO - codeparrot_training - Step 20262: {'lr': 0.00034169634737481034, 'samples': 10374656, 'steps': 20262, 'loss/train': 2.2250335216522217} 02/25/2022 03:49:28 - INFO - codeparrot_training - Step 20263: {'lr': 0.0003416811250889137, 'samples': 10375168, 'steps': 20263, 'loss/train': 1.5773553848266602} 02/25/2022 03:49:34 - INFO - codeparrot_training - Step 20264: {'lr': 0.00034166590241028425, 'samples': 10375680, 'steps': 20264, 'loss/train': 1.0202758312225342} 02/25/2022 03:49:37 - INFO - codeparrot_training - Step 20265: {'lr': 0.0003416506793389871, 'samples': 10376192, 'steps': 20265, 'loss/train': 2.670820474624634} 02/25/2022 03:49:43 - INFO - codeparrot_training - Step 20266: {'lr': 0.0003416354558750876, 'samples': 10376704, 'steps': 20266, 'loss/train': 2.014662981033325} 02/25/2022 03:49:46 - INFO - codeparrot_training - Step 20267: {'lr': 0.0003416202320186508, 'samples': 10377216, 'steps': 20267, 'loss/train': 2.6723196506500244} 02/25/2022 03:49:52 - INFO - codeparrot_training - Step 20268: {'lr': 0.0003416050077697422, 'samples': 10377728, 'steps': 20268, 'loss/train': 1.0188957452774048} 02/25/2022 03:49:56 - INFO - codeparrot_training - Step 20269: {'lr': 0.0003415897831284267, 'samples': 10378240, 'steps': 20269, 'loss/train': 2.786364793777466} 02/25/2022 03:50:03 - INFO - codeparrot_training - Step 20270: {'lr': 0.0003415745580947697, 'samples': 10378752, 'steps': 20270, 'loss/train': 2.278608798980713} 02/25/2022 03:50:06 - INFO - codeparrot_training - Step 20271: {'lr': 0.0003415593326688364, 'samples': 10379264, 'steps': 20271, 'loss/train': 1.9659326076507568} 02/25/2022 03:50:12 - INFO - codeparrot_training - Step 20272: {'lr': 0.0003415441068506919, 'samples': 10379776, 'steps': 20272, 'loss/train': 0.8071721196174622} 02/25/2022 03:50:15 - INFO - codeparrot_training - Step 20273: {'lr': 0.0003415288806404016, 'samples': 10380288, 'steps': 20273, 'loss/train': 4.123531341552734} 02/25/2022 03:50:21 - INFO - codeparrot_training - Step 20274: {'lr': 0.00034151365403803065, 'samples': 10380800, 'steps': 20274, 'loss/train': 1.7842228412628174} 02/25/2022 03:50:24 - INFO - codeparrot_training - Step 20275: {'lr': 0.0003414984270436442, 'samples': 10381312, 'steps': 20275, 'loss/train': 2.567279577255249} 02/25/2022 03:50:30 - INFO - codeparrot_training - Step 20276: {'lr': 0.00034148319965730757, 'samples': 10381824, 'steps': 20276, 'loss/train': 2.0263853073120117} 02/25/2022 03:50:33 - INFO - codeparrot_training - Step 20277: {'lr': 0.000341467971879086, 'samples': 10382336, 'steps': 20277, 'loss/train': 1.390875220298767} 02/25/2022 03:50:39 - INFO - codeparrot_training - Step 20278: {'lr': 0.0003414527437090446, 'samples': 10382848, 'steps': 20278, 'loss/train': 1.2159847021102905} 02/25/2022 03:50:42 - INFO - codeparrot_training - Step 20279: {'lr': 0.00034143751514724874, 'samples': 10383360, 'steps': 20279, 'loss/train': 2.333939552307129} 02/25/2022 03:50:49 - INFO - codeparrot_training - Step 20280: {'lr': 0.0003414222861937636, 'samples': 10383872, 'steps': 20280, 'loss/train': 1.8507686853408813} 02/25/2022 03:50:53 - INFO - codeparrot_training - Step 20281: {'lr': 0.00034140705684865437, 'samples': 10384384, 'steps': 20281, 'loss/train': 1.3636988401412964} 02/25/2022 03:50:58 - INFO - codeparrot_training - Step 20282: {'lr': 0.0003413918271119864, 'samples': 10384896, 'steps': 20282, 'loss/train': 2.3370273113250732} 02/25/2022 03:51:02 - INFO - codeparrot_training - Step 20283: {'lr': 0.00034137659698382485, 'samples': 10385408, 'steps': 20283, 'loss/train': 2.8440072536468506} 02/25/2022 03:51:07 - INFO - codeparrot_training - Step 20284: {'lr': 0.0003413613664642349, 'samples': 10385920, 'steps': 20284, 'loss/train': 1.3650819063186646} 02/25/2022 03:51:11 - INFO - codeparrot_training - Step 20285: {'lr': 0.00034134613555328195, 'samples': 10386432, 'steps': 20285, 'loss/train': 1.5565146207809448} 02/25/2022 03:51:16 - INFO - codeparrot_training - Step 20286: {'lr': 0.00034133090425103114, 'samples': 10386944, 'steps': 20286, 'loss/train': 1.8609282970428467} 02/25/2022 03:51:20 - INFO - codeparrot_training - Step 20287: {'lr': 0.00034131567255754776, 'samples': 10387456, 'steps': 20287, 'loss/train': 1.7913402318954468} 02/25/2022 03:51:25 - INFO - codeparrot_training - Step 20288: {'lr': 0.000341300440472897, 'samples': 10387968, 'steps': 20288, 'loss/train': 2.0067458152770996} 02/25/2022 03:51:29 - INFO - codeparrot_training - Step 20289: {'lr': 0.0003412852079971441, 'samples': 10388480, 'steps': 20289, 'loss/train': 2.3061718940734863} 02/25/2022 03:51:34 - INFO - codeparrot_training - Step 20290: {'lr': 0.0003412699751303544, 'samples': 10388992, 'steps': 20290, 'loss/train': 2.2157087326049805} 02/25/2022 03:51:38 - INFO - codeparrot_training - Step 20291: {'lr': 0.00034125474187259307, 'samples': 10389504, 'steps': 20291, 'loss/train': 2.6220147609710693} 02/25/2022 03:51:44 - INFO - codeparrot_training - Step 20292: {'lr': 0.0003412395082239255, 'samples': 10390016, 'steps': 20292, 'loss/train': 2.666356325149536} 02/25/2022 03:51:47 - INFO - codeparrot_training - Step 20293: {'lr': 0.0003412242741844168, 'samples': 10390528, 'steps': 20293, 'loss/train': 2.0440683364868164} 02/25/2022 03:51:53 - INFO - codeparrot_training - Step 20294: {'lr': 0.0003412090397541323, 'samples': 10391040, 'steps': 20294, 'loss/train': 1.1251710653305054} 02/25/2022 03:51:56 - INFO - codeparrot_training - Step 20295: {'lr': 0.0003411938049331372, 'samples': 10391552, 'steps': 20295, 'loss/train': 1.8887560367584229} 02/25/2022 03:52:04 - INFO - codeparrot_training - Step 20296: {'lr': 0.0003411785697214968, 'samples': 10392064, 'steps': 20296, 'loss/train': 3.546196222305298} 02/25/2022 03:52:07 - INFO - codeparrot_training - Step 20297: {'lr': 0.00034116333411927637, 'samples': 10392576, 'steps': 20297, 'loss/train': 1.8653291463851929} 02/25/2022 03:52:13 - INFO - codeparrot_training - Step 20298: {'lr': 0.0003411480981265411, 'samples': 10393088, 'steps': 20298, 'loss/train': 0.7127207517623901} 02/25/2022 03:52:16 - INFO - codeparrot_training - Step 20299: {'lr': 0.0003411328617433564, 'samples': 10393600, 'steps': 20299, 'loss/train': 2.2414886951446533} 02/25/2022 03:52:22 - INFO - codeparrot_training - Step 20300: {'lr': 0.0003411176249697875, 'samples': 10394112, 'steps': 20300, 'loss/train': 2.163728952407837} 02/25/2022 03:52:25 - INFO - codeparrot_training - Step 20301: {'lr': 0.0003411023878058995, 'samples': 10394624, 'steps': 20301, 'loss/train': 2.2823126316070557} 02/25/2022 03:52:31 - INFO - codeparrot_training - Step 20302: {'lr': 0.0003410871502517579, 'samples': 10395136, 'steps': 20302, 'loss/train': 3.952664613723755} 02/25/2022 03:52:34 - INFO - codeparrot_training - Step 20303: {'lr': 0.00034107191230742776, 'samples': 10395648, 'steps': 20303, 'loss/train': 1.834477186203003} 02/25/2022 03:52:40 - INFO - codeparrot_training - Step 20304: {'lr': 0.00034105667397297455, 'samples': 10396160, 'steps': 20304, 'loss/train': 1.780929684638977} 02/25/2022 03:52:43 - INFO - codeparrot_training - Step 20305: {'lr': 0.0003410414352484635, 'samples': 10396672, 'steps': 20305, 'loss/train': 1.5590553283691406} 02/25/2022 03:52:51 - INFO - codeparrot_training - Step 20306: {'lr': 0.00034102619613395974, 'samples': 10397184, 'steps': 20306, 'loss/train': 2.6603586673736572} 02/25/2022 03:52:54 - INFO - codeparrot_training - Step 20307: {'lr': 0.00034101095662952873, 'samples': 10397696, 'steps': 20307, 'loss/train': 1.601564884185791} 02/25/2022 03:53:00 - INFO - codeparrot_training - Step 20308: {'lr': 0.00034099571673523564, 'samples': 10398208, 'steps': 20308, 'loss/train': 2.099993944168091} 02/25/2022 03:53:03 - INFO - codeparrot_training - Step 20309: {'lr': 0.0003409804764511459, 'samples': 10398720, 'steps': 20309, 'loss/train': 0.9904488325119019} 02/25/2022 03:53:09 - INFO - codeparrot_training - Step 20310: {'lr': 0.00034096523577732457, 'samples': 10399232, 'steps': 20310, 'loss/train': 1.1795721054077148} 02/25/2022 03:53:12 - INFO - codeparrot_training - Step 20311: {'lr': 0.00034094999471383713, 'samples': 10399744, 'steps': 20311, 'loss/train': 2.3907694816589355} 02/25/2022 03:53:18 - INFO - codeparrot_training - Step 20312: {'lr': 0.00034093475326074874, 'samples': 10400256, 'steps': 20312, 'loss/train': 2.467639207839966} 02/25/2022 03:53:21 - INFO - codeparrot_training - Step 20313: {'lr': 0.00034091951141812483, 'samples': 10400768, 'steps': 20313, 'loss/train': 1.0897190570831299} 02/25/2022 03:53:27 - INFO - codeparrot_training - Step 20314: {'lr': 0.00034090426918603045, 'samples': 10401280, 'steps': 20314, 'loss/train': 0.9841932058334351} 02/25/2022 03:53:30 - INFO - codeparrot_training - Step 20315: {'lr': 0.00034088902656453116, 'samples': 10401792, 'steps': 20315, 'loss/train': 2.4677000045776367} 02/25/2022 03:53:37 - INFO - codeparrot_training - Step 20316: {'lr': 0.00034087378355369215, 'samples': 10402304, 'steps': 20316, 'loss/train': 2.2717084884643555} 02/25/2022 03:53:41 - INFO - codeparrot_training - Step 20317: {'lr': 0.00034085854015357864, 'samples': 10402816, 'steps': 20317, 'loss/train': 0.8332167863845825} 02/25/2022 03:53:46 - INFO - codeparrot_training - Step 20318: {'lr': 0.000340843296364256, 'samples': 10403328, 'steps': 20318, 'loss/train': 1.9185445308685303} 02/25/2022 03:53:50 - INFO - codeparrot_training - Step 20319: {'lr': 0.00034082805218578954, 'samples': 10403840, 'steps': 20319, 'loss/train': 2.3251028060913086} 02/25/2022 03:53:55 - INFO - codeparrot_training - Step 20320: {'lr': 0.00034081280761824465, 'samples': 10404352, 'steps': 20320, 'loss/train': 2.1127960681915283} 02/25/2022 03:53:59 - INFO - codeparrot_training - Step 20321: {'lr': 0.0003407975626616864, 'samples': 10404864, 'steps': 20321, 'loss/train': 2.142221689224243} 02/25/2022 03:54:04 - INFO - codeparrot_training - Step 20322: {'lr': 0.00034078231731618025, 'samples': 10405376, 'steps': 20322, 'loss/train': 2.7477948665618896} 02/25/2022 03:54:08 - INFO - codeparrot_training - Step 20323: {'lr': 0.00034076707158179145, 'samples': 10405888, 'steps': 20323, 'loss/train': 1.6565980911254883} 02/25/2022 03:54:13 - INFO - codeparrot_training - Step 20324: {'lr': 0.0003407518254585854, 'samples': 10406400, 'steps': 20324, 'loss/train': 2.412675380706787} 02/25/2022 03:54:17 - INFO - codeparrot_training - Step 20325: {'lr': 0.0003407365789466273, 'samples': 10406912, 'steps': 20325, 'loss/train': 2.3295187950134277} 02/25/2022 03:54:23 - INFO - codeparrot_training - Step 20326: {'lr': 0.00034072133204598255, 'samples': 10407424, 'steps': 20326, 'loss/train': 1.271855115890503} 02/25/2022 03:54:26 - INFO - codeparrot_training - Step 20327: {'lr': 0.00034070608475671643, 'samples': 10407936, 'steps': 20327, 'loss/train': 1.4647499322891235} 02/25/2022 03:54:32 - INFO - codeparrot_training - Step 20328: {'lr': 0.0003406908370788942, 'samples': 10408448, 'steps': 20328, 'loss/train': 0.8187311887741089} 02/25/2022 03:54:35 - INFO - codeparrot_training - Step 20329: {'lr': 0.0003406755890125813, 'samples': 10408960, 'steps': 20329, 'loss/train': 2.774731397628784} 02/25/2022 03:54:41 - INFO - codeparrot_training - Step 20330: {'lr': 0.00034066034055784284, 'samples': 10409472, 'steps': 20330, 'loss/train': 1.4913020133972168} 02/25/2022 03:54:44 - INFO - codeparrot_training - Step 20331: {'lr': 0.00034064509171474435, 'samples': 10409984, 'steps': 20331, 'loss/train': 3.659651756286621} 02/25/2022 03:54:50 - INFO - codeparrot_training - Step 20332: {'lr': 0.00034062984248335105, 'samples': 10410496, 'steps': 20332, 'loss/train': 1.6191192865371704} 02/25/2022 03:54:53 - INFO - codeparrot_training - Step 20333: {'lr': 0.0003406145928637283, 'samples': 10411008, 'steps': 20333, 'loss/train': 1.7046446800231934} 02/25/2022 03:54:59 - INFO - codeparrot_training - Step 20334: {'lr': 0.0003405993428559414, 'samples': 10411520, 'steps': 20334, 'loss/train': 0.5736855268478394} 02/25/2022 03:55:02 - INFO - codeparrot_training - Step 20335: {'lr': 0.0003405840924600557, 'samples': 10412032, 'steps': 20335, 'loss/train': 3.1948325634002686} 02/25/2022 03:55:08 - INFO - codeparrot_training - Step 20336: {'lr': 0.0003405688416761364, 'samples': 10412544, 'steps': 20336, 'loss/train': 1.9165589809417725} 02/25/2022 03:55:11 - INFO - codeparrot_training - Step 20337: {'lr': 0.0003405535905042491, 'samples': 10413056, 'steps': 20337, 'loss/train': 2.527679204940796} 02/25/2022 03:55:17 - INFO - codeparrot_training - Step 20338: {'lr': 0.0003405383389444589, 'samples': 10413568, 'steps': 20338, 'loss/train': 1.2761796712875366} 02/25/2022 03:55:20 - INFO - codeparrot_training - Step 20339: {'lr': 0.00034052308699683114, 'samples': 10414080, 'steps': 20339, 'loss/train': 1.6958613395690918} 02/25/2022 03:55:27 - INFO - codeparrot_training - Step 20340: {'lr': 0.0003405078346614313, 'samples': 10414592, 'steps': 20340, 'loss/train': 2.393470287322998} 02/25/2022 03:55:31 - INFO - codeparrot_training - Step 20341: {'lr': 0.00034049258193832464, 'samples': 10415104, 'steps': 20341, 'loss/train': 2.105696678161621} 02/25/2022 03:55:36 - INFO - codeparrot_training - Step 20342: {'lr': 0.00034047732882757655, 'samples': 10415616, 'steps': 20342, 'loss/train': 2.420139789581299} 02/25/2022 03:55:40 - INFO - codeparrot_training - Step 20343: {'lr': 0.00034046207532925215, 'samples': 10416128, 'steps': 20343, 'loss/train': 1.5979433059692383} 02/25/2022 03:55:45 - INFO - codeparrot_training - Step 20344: {'lr': 0.0003404468214434171, 'samples': 10416640, 'steps': 20344, 'loss/train': 1.823500633239746} 02/25/2022 03:55:49 - INFO - codeparrot_training - Step 20345: {'lr': 0.00034043156717013647, 'samples': 10417152, 'steps': 20345, 'loss/train': 2.3961737155914307} 02/25/2022 03:55:55 - INFO - codeparrot_training - Step 20346: {'lr': 0.0003404163125094758, 'samples': 10417664, 'steps': 20346, 'loss/train': 1.849566102027893} 02/25/2022 03:56:00 - INFO - codeparrot_training - Step 20347: {'lr': 0.0003404010574615003, 'samples': 10418176, 'steps': 20347, 'loss/train': 2.186707019805908} 02/25/2022 03:56:03 - INFO - codeparrot_training - Step 20348: {'lr': 0.00034038580202627543, 'samples': 10418688, 'steps': 20348, 'loss/train': 3.459792137145996} 02/25/2022 03:56:09 - INFO - codeparrot_training - Step 20349: {'lr': 0.0003403705462038665, 'samples': 10419200, 'steps': 20349, 'loss/train': 0.9187582731246948} 02/25/2022 03:56:13 - INFO - codeparrot_training - Step 20350: {'lr': 0.0003403552899943388, 'samples': 10419712, 'steps': 20350, 'loss/train': 2.584151029586792} 02/25/2022 03:56:19 - INFO - codeparrot_training - Step 20351: {'lr': 0.0003403400333977577, 'samples': 10420224, 'steps': 20351, 'loss/train': 1.807735562324524} 02/25/2022 03:56:22 - INFO - codeparrot_training - Step 20352: {'lr': 0.00034032477641418856, 'samples': 10420736, 'steps': 20352, 'loss/train': 1.9062429666519165} 02/25/2022 03:56:28 - INFO - codeparrot_training - Step 20353: {'lr': 0.00034030951904369687, 'samples': 10421248, 'steps': 20353, 'loss/train': 1.157932996749878} 02/25/2022 03:56:31 - INFO - codeparrot_training - Step 20354: {'lr': 0.0003402942612863478, 'samples': 10421760, 'steps': 20354, 'loss/train': 3.330751657485962} 02/25/2022 03:56:37 - INFO - codeparrot_training - Step 20355: {'lr': 0.00034027900314220684, 'samples': 10422272, 'steps': 20355, 'loss/train': 0.6594069004058838} 02/25/2022 03:56:40 - INFO - codeparrot_training - Step 20356: {'lr': 0.00034026374461133927, 'samples': 10422784, 'steps': 20356, 'loss/train': 0.8760718107223511} 02/25/2022 03:56:46 - INFO - codeparrot_training - Step 20357: {'lr': 0.0003402484856938105, 'samples': 10423296, 'steps': 20357, 'loss/train': 2.344481945037842} 02/25/2022 03:56:50 - INFO - codeparrot_training - Step 20358: {'lr': 0.00034023322638968587, 'samples': 10423808, 'steps': 20358, 'loss/train': 1.5700024366378784} 02/25/2022 03:56:55 - INFO - codeparrot_training - Step 20359: {'lr': 0.0003402179666990307, 'samples': 10424320, 'steps': 20359, 'loss/train': 1.1055718660354614} 02/25/2022 03:56:59 - INFO - codeparrot_training - Step 20360: {'lr': 0.00034020270662191046, 'samples': 10424832, 'steps': 20360, 'loss/train': 2.3525965213775635} 02/25/2022 03:57:04 - INFO - codeparrot_training - Step 20361: {'lr': 0.0003401874461583905, 'samples': 10425344, 'steps': 20361, 'loss/train': 1.6585325002670288} 02/25/2022 03:57:08 - INFO - codeparrot_training - Step 20362: {'lr': 0.00034017218530853606, 'samples': 10425856, 'steps': 20362, 'loss/train': 1.6461007595062256} 02/25/2022 03:57:14 - INFO - codeparrot_training - Step 20363: {'lr': 0.0003401569240724126, 'samples': 10426368, 'steps': 20363, 'loss/train': 2.7128148078918457} 02/25/2022 03:57:17 - INFO - codeparrot_training - Step 20364: {'lr': 0.0003401416624500856, 'samples': 10426880, 'steps': 20364, 'loss/train': 1.3656822443008423} 02/25/2022 03:57:23 - INFO - codeparrot_training - Step 20365: {'lr': 0.0003401264004416203, 'samples': 10427392, 'steps': 20365, 'loss/train': 2.0054006576538086} 02/25/2022 03:57:26 - INFO - codeparrot_training - Step 20366: {'lr': 0.00034011113804708216, 'samples': 10427904, 'steps': 20366, 'loss/train': 1.8891280889511108} 02/25/2022 03:57:32 - INFO - codeparrot_training - Step 20367: {'lr': 0.0003400958752665365, 'samples': 10428416, 'steps': 20367, 'loss/train': 1.4771785736083984} 02/25/2022 03:57:36 - INFO - codeparrot_training - Step 20368: {'lr': 0.0003400806121000487, 'samples': 10428928, 'steps': 20368, 'loss/train': 2.1553425788879395} 02/25/2022 03:57:41 - INFO - codeparrot_training - Step 20369: {'lr': 0.00034006534854768417, 'samples': 10429440, 'steps': 20369, 'loss/train': 2.0339150428771973} 02/25/2022 03:57:45 - INFO - codeparrot_training - Step 20370: {'lr': 0.00034005008460950825, 'samples': 10429952, 'steps': 20370, 'loss/train': 1.69804048538208} 02/25/2022 03:57:52 - INFO - codeparrot_training - Step 20371: {'lr': 0.00034003482028558644, 'samples': 10430464, 'steps': 20371, 'loss/train': 1.744829773902893} 02/25/2022 03:57:55 - INFO - codeparrot_training - Step 20372: {'lr': 0.000340019555575984, 'samples': 10430976, 'steps': 20372, 'loss/train': 1.6603659391403198} 02/25/2022 03:58:00 - INFO - codeparrot_training - Step 20373: {'lr': 0.00034000429048076637, 'samples': 10431488, 'steps': 20373, 'loss/train': 1.9485360383987427} 02/25/2022 03:58:04 - INFO - codeparrot_training - Step 20374: {'lr': 0.00033998902499999894, 'samples': 10432000, 'steps': 20374, 'loss/train': 0.7774336934089661} 02/25/2022 03:58:09 - INFO - codeparrot_training - Step 20375: {'lr': 0.00033997375913374705, 'samples': 10432512, 'steps': 20375, 'loss/train': 0.4261077344417572} 02/25/2022 03:58:13 - INFO - codeparrot_training - Step 20376: {'lr': 0.0003399584928820762, 'samples': 10433024, 'steps': 20376, 'loss/train': 2.7587804794311523} 02/25/2022 03:58:18 - INFO - codeparrot_training - Step 20377: {'lr': 0.00033994322624505177, 'samples': 10433536, 'steps': 20377, 'loss/train': 1.6619898080825806} 02/25/2022 03:58:22 - INFO - codeparrot_training - Step 20378: {'lr': 0.000339927959222739, 'samples': 10434048, 'steps': 20378, 'loss/train': 1.8428360223770142} 02/25/2022 03:58:27 - INFO - codeparrot_training - Step 20379: {'lr': 0.0003399126918152036, 'samples': 10434560, 'steps': 20379, 'loss/train': 1.369884967803955} 02/25/2022 03:58:31 - INFO - codeparrot_training - Step 20380: {'lr': 0.00033989742402251055, 'samples': 10435072, 'steps': 20380, 'loss/train': 0.5787926912307739} 02/25/2022 03:58:36 - INFO - codeparrot_training - Step 20381: {'lr': 0.00033988215584472564, 'samples': 10435584, 'steps': 20381, 'loss/train': 1.7602312564849854} 02/25/2022 03:58:40 - INFO - codeparrot_training - Step 20382: {'lr': 0.00033986688728191404, 'samples': 10436096, 'steps': 20382, 'loss/train': 2.696516275405884} 02/25/2022 03:58:45 - INFO - codeparrot_training - Step 20383: {'lr': 0.0003398516183341413, 'samples': 10436608, 'steps': 20383, 'loss/train': 1.0951000452041626} 02/25/2022 03:58:49 - INFO - codeparrot_training - Step 20384: {'lr': 0.0003398363490014727, 'samples': 10437120, 'steps': 20384, 'loss/train': 0.23739130795001984} 02/25/2022 03:58:54 - INFO - codeparrot_training - Step 20385: {'lr': 0.0003398210792839738, 'samples': 10437632, 'steps': 20385, 'loss/train': 2.0905277729034424} 02/25/2022 03:58:58 - INFO - codeparrot_training - Step 20386: {'lr': 0.0003398058091817098, 'samples': 10438144, 'steps': 20386, 'loss/train': 1.4892735481262207} 02/25/2022 03:59:04 - INFO - codeparrot_training - Step 20387: {'lr': 0.0003397905386947462, 'samples': 10438656, 'steps': 20387, 'loss/train': 3.576022148132324} 02/25/2022 03:59:08 - INFO - codeparrot_training - Step 20388: {'lr': 0.00033977526782314854, 'samples': 10439168, 'steps': 20388, 'loss/train': 2.563326120376587} 02/25/2022 03:59:13 - INFO - codeparrot_training - Step 20389: {'lr': 0.00033975999656698206, 'samples': 10439680, 'steps': 20389, 'loss/train': 2.2927839756011963} 02/25/2022 03:59:16 - INFO - codeparrot_training - Step 20390: {'lr': 0.00033974472492631234, 'samples': 10440192, 'steps': 20390, 'loss/train': 2.5966854095458984} 02/25/2022 03:59:22 - INFO - codeparrot_training - Step 20391: {'lr': 0.0003397294529012047, 'samples': 10440704, 'steps': 20391, 'loss/train': 0.21449001133441925} 02/25/2022 03:59:25 - INFO - codeparrot_training - Step 20392: {'lr': 0.0003397141804917246, 'samples': 10441216, 'steps': 20392, 'loss/train': 1.3566102981567383} 02/25/2022 03:59:31 - INFO - codeparrot_training - Step 20393: {'lr': 0.00033969890769793736, 'samples': 10441728, 'steps': 20393, 'loss/train': 1.2585475444793701} 02/25/2022 03:59:34 - INFO - codeparrot_training - Step 20394: {'lr': 0.0003396836345199086, 'samples': 10442240, 'steps': 20394, 'loss/train': 2.7500460147857666} 02/25/2022 03:59:40 - INFO - codeparrot_training - Step 20395: {'lr': 0.0003396683609577035, 'samples': 10442752, 'steps': 20395, 'loss/train': 0.9440882205963135} 02/25/2022 03:59:43 - INFO - codeparrot_training - Step 20396: {'lr': 0.0003396530870113877, 'samples': 10443264, 'steps': 20396, 'loss/train': 3.391152858734131} 02/25/2022 03:59:49 - INFO - codeparrot_training - Step 20397: {'lr': 0.0003396378126810264, 'samples': 10443776, 'steps': 20397, 'loss/train': 1.4591679573059082} 02/25/2022 03:59:53 - INFO - codeparrot_training - Step 20398: {'lr': 0.0003396225379666854, 'samples': 10444288, 'steps': 20398, 'loss/train': 1.8293768167495728} 02/25/2022 03:59:58 - INFO - codeparrot_training - Step 20399: {'lr': 0.00033960726286842973, 'samples': 10444800, 'steps': 20399, 'loss/train': 3.2655420303344727} 02/25/2022 04:00:02 - INFO - codeparrot_training - Step 20400: {'lr': 0.00033959198738632503, 'samples': 10445312, 'steps': 20400, 'loss/train': 1.8054386377334595} 02/25/2022 04:00:07 - INFO - codeparrot_training - Step 20401: {'lr': 0.00033957671152043677, 'samples': 10445824, 'steps': 20401, 'loss/train': 1.1139875650405884} 02/25/2022 04:00:11 - INFO - codeparrot_training - Step 20402: {'lr': 0.0003395614352708303, 'samples': 10446336, 'steps': 20402, 'loss/train': 2.0459933280944824} 02/25/2022 04:00:16 - INFO - codeparrot_training - Step 20403: {'lr': 0.00033954615863757105, 'samples': 10446848, 'steps': 20403, 'loss/train': 0.432099848985672} 02/25/2022 04:00:20 - INFO - codeparrot_training - Step 20404: {'lr': 0.0003395308816207245, 'samples': 10447360, 'steps': 20404, 'loss/train': 1.4184174537658691} 02/25/2022 04:00:25 - INFO - codeparrot_training - Step 20405: {'lr': 0.0003395156042203561, 'samples': 10447872, 'steps': 20405, 'loss/train': 2.241032361984253} 02/25/2022 04:00:29 - INFO - codeparrot_training - Step 20406: {'lr': 0.0003395003264365313, 'samples': 10448384, 'steps': 20406, 'loss/train': 2.641526699066162} 02/25/2022 04:00:34 - INFO - codeparrot_training - Step 20407: {'lr': 0.00033948504826931546, 'samples': 10448896, 'steps': 20407, 'loss/train': 0.43966493010520935} 02/25/2022 04:00:38 - INFO - codeparrot_training - Step 20408: {'lr': 0.0003394697697187741, 'samples': 10449408, 'steps': 20408, 'loss/train': 2.6849939823150635} 02/25/2022 04:00:44 - INFO - codeparrot_training - Step 20409: {'lr': 0.0003394544907849727, 'samples': 10449920, 'steps': 20409, 'loss/train': 1.4712785482406616} 02/25/2022 04:00:48 - INFO - codeparrot_training - Step 20410: {'lr': 0.0003394392114679766, 'samples': 10450432, 'steps': 20410, 'loss/train': 3.2722394466400146} 02/25/2022 04:00:53 - INFO - codeparrot_training - Step 20411: {'lr': 0.00033942393176785134, 'samples': 10450944, 'steps': 20411, 'loss/train': 1.9418766498565674} 02/25/2022 04:00:56 - INFO - codeparrot_training - Step 20412: {'lr': 0.00033940865168466237, 'samples': 10451456, 'steps': 20412, 'loss/train': 1.7702662944793701} 02/25/2022 04:01:02 - INFO - codeparrot_training - Step 20413: {'lr': 0.00033939337121847513, 'samples': 10451968, 'steps': 20413, 'loss/train': 2.305882215499878} 02/25/2022 04:01:05 - INFO - codeparrot_training - Step 20414: {'lr': 0.00033937809036935505, 'samples': 10452480, 'steps': 20414, 'loss/train': 2.1831846237182617} 02/25/2022 04:01:11 - INFO - codeparrot_training - Step 20415: {'lr': 0.0003393628091373677, 'samples': 10452992, 'steps': 20415, 'loss/train': 2.453648567199707} 02/25/2022 04:01:14 - INFO - codeparrot_training - Step 20416: {'lr': 0.00033934752752257834, 'samples': 10453504, 'steps': 20416, 'loss/train': 1.6907323598861694} 02/25/2022 04:01:20 - INFO - codeparrot_training - Step 20417: {'lr': 0.00033933224552505257, 'samples': 10454016, 'steps': 20417, 'loss/train': 2.2379722595214844} 02/25/2022 04:01:23 - INFO - codeparrot_training - Step 20418: {'lr': 0.0003393169631448559, 'samples': 10454528, 'steps': 20418, 'loss/train': 1.8829885721206665} 02/25/2022 04:01:32 - INFO - codeparrot_training - Step 20419: {'lr': 0.0003393016803820536, 'samples': 10455040, 'steps': 20419, 'loss/train': 1.9877028465270996} 02/25/2022 04:01:37 - INFO - codeparrot_training - Step 20420: {'lr': 0.0003392863972367114, 'samples': 10455552, 'steps': 20420, 'loss/train': 1.5622564554214478} 02/25/2022 04:01:41 - INFO - codeparrot_training - Step 20421: {'lr': 0.0003392711137088945, 'samples': 10456064, 'steps': 20421, 'loss/train': 1.0281120538711548} 02/25/2022 04:01:46 - INFO - codeparrot_training - Step 20422: {'lr': 0.00033925582979866853, 'samples': 10456576, 'steps': 20422, 'loss/train': 1.0365203619003296} 02/25/2022 04:01:50 - INFO - codeparrot_training - Step 20423: {'lr': 0.00033924054550609894, 'samples': 10457088, 'steps': 20423, 'loss/train': 1.7628027200698853} 02/25/2022 04:01:55 - INFO - codeparrot_training - Step 20424: {'lr': 0.00033922526083125115, 'samples': 10457600, 'steps': 20424, 'loss/train': 0.5534242391586304} 02/25/2022 04:01:59 - INFO - codeparrot_training - Step 20425: {'lr': 0.00033920997577419076, 'samples': 10458112, 'steps': 20425, 'loss/train': 1.0568207502365112} 02/25/2022 04:02:04 - INFO - codeparrot_training - Step 20426: {'lr': 0.00033919469033498304, 'samples': 10458624, 'steps': 20426, 'loss/train': 2.7125191688537598} 02/25/2022 04:02:08 - INFO - codeparrot_training - Step 20427: {'lr': 0.0003391794045136937, 'samples': 10459136, 'steps': 20427, 'loss/train': 2.322615623474121} 02/25/2022 04:02:13 - INFO - codeparrot_training - Step 20428: {'lr': 0.00033916411831038805, 'samples': 10459648, 'steps': 20428, 'loss/train': 2.0178771018981934} 02/25/2022 04:02:17 - INFO - codeparrot_training - Step 20429: {'lr': 0.0003391488317251317, 'samples': 10460160, 'steps': 20429, 'loss/train': 2.501131772994995} 02/25/2022 04:02:22 - INFO - codeparrot_training - Step 20430: {'lr': 0.00033913354475798995, 'samples': 10460672, 'steps': 20430, 'loss/train': 2.1958131790161133} 02/25/2022 04:02:26 - INFO - codeparrot_training - Step 20431: {'lr': 0.0003391182574090285, 'samples': 10461184, 'steps': 20431, 'loss/train': 2.0659499168395996} 02/25/2022 04:02:31 - INFO - codeparrot_training - Step 20432: {'lr': 0.00033910296967831267, 'samples': 10461696, 'steps': 20432, 'loss/train': 2.09885573387146} 02/25/2022 04:02:35 - INFO - codeparrot_training - Step 20433: {'lr': 0.00033908768156590806, 'samples': 10462208, 'steps': 20433, 'loss/train': 1.8103362321853638} 02/25/2022 04:02:42 - INFO - codeparrot_training - Step 20434: {'lr': 0.0003390723930718801, 'samples': 10462720, 'steps': 20434, 'loss/train': 0.6538064479827881} 02/25/2022 04:02:45 - INFO - codeparrot_training - Step 20435: {'lr': 0.0003390571041962943, 'samples': 10463232, 'steps': 20435, 'loss/train': 1.8231159448623657} 02/25/2022 04:02:51 - INFO - codeparrot_training - Step 20436: {'lr': 0.0003390418149392161, 'samples': 10463744, 'steps': 20436, 'loss/train': 1.4817583560943604} 02/25/2022 04:02:54 - INFO - codeparrot_training - Step 20437: {'lr': 0.000339026525300711, 'samples': 10464256, 'steps': 20437, 'loss/train': 0.34716546535491943} 02/25/2022 04:03:00 - INFO - codeparrot_training - Step 20438: {'lr': 0.0003390112352808447, 'samples': 10464768, 'steps': 20438, 'loss/train': 1.7235552072525024} 02/25/2022 04:03:03 - INFO - codeparrot_training - Step 20439: {'lr': 0.00033899594487968234, 'samples': 10465280, 'steps': 20439, 'loss/train': 1.886596441268921} 02/25/2022 04:03:09 - INFO - codeparrot_training - Step 20440: {'lr': 0.00033898065409728974, 'samples': 10465792, 'steps': 20440, 'loss/train': 2.01859712600708} 02/25/2022 04:03:12 - INFO - codeparrot_training - Step 20441: {'lr': 0.0003389653629337322, 'samples': 10466304, 'steps': 20441, 'loss/train': 1.9983035326004028} 02/25/2022 04:03:20 - INFO - codeparrot_training - Step 20442: {'lr': 0.00033895007138907534, 'samples': 10466816, 'steps': 20442, 'loss/train': 1.7958946228027344} 02/25/2022 04:03:23 - INFO - codeparrot_training - Step 20443: {'lr': 0.00033893477946338456, 'samples': 10467328, 'steps': 20443, 'loss/train': 2.8162474632263184} 02/25/2022 04:03:29 - INFO - codeparrot_training - Step 20444: {'lr': 0.0003389194871567255, 'samples': 10467840, 'steps': 20444, 'loss/train': 0.8238244652748108} 02/25/2022 04:03:32 - INFO - codeparrot_training - Step 20445: {'lr': 0.00033890419446916354, 'samples': 10468352, 'steps': 20445, 'loss/train': 2.4720194339752197} 02/25/2022 04:03:36 - INFO - codeparrot_training - Step 20446: {'lr': 0.00033888890140076433, 'samples': 10468864, 'steps': 20446, 'loss/train': 8.994832992553711} 02/25/2022 04:03:41 - INFO - codeparrot_training - Step 20447: {'lr': 0.00033887360795159315, 'samples': 10469376, 'steps': 20447, 'loss/train': 1.3861021995544434} 02/25/2022 04:03:45 - INFO - codeparrot_training - Step 20448: {'lr': 0.00033885831412171577, 'samples': 10469888, 'steps': 20448, 'loss/train': 2.0955758094787598} 02/25/2022 04:03:50 - INFO - codeparrot_training - Step 20449: {'lr': 0.00033884301991119755, 'samples': 10470400, 'steps': 20449, 'loss/train': 2.7549216747283936} 02/25/2022 04:03:54 - INFO - codeparrot_training - Step 20450: {'lr': 0.00033882772532010404, 'samples': 10470912, 'steps': 20450, 'loss/train': 1.8993308544158936} 02/25/2022 04:03:59 - INFO - codeparrot_training - Step 20451: {'lr': 0.0003388124303485007, 'samples': 10471424, 'steps': 20451, 'loss/train': 2.9791548252105713} 02/25/2022 04:04:03 - INFO - codeparrot_training - Step 20452: {'lr': 0.0003387971349964532, 'samples': 10471936, 'steps': 20452, 'loss/train': 0.8059250116348267} 02/25/2022 04:04:09 - INFO - codeparrot_training - Step 20453: {'lr': 0.00033878183926402695, 'samples': 10472448, 'steps': 20453, 'loss/train': 2.2308528423309326} 02/25/2022 04:04:12 - INFO - codeparrot_training - Step 20454: {'lr': 0.00033876654315128746, 'samples': 10472960, 'steps': 20454, 'loss/train': 2.264249801635742} 02/25/2022 04:04:19 - INFO - codeparrot_training - Step 20455: {'lr': 0.0003387512466583004, 'samples': 10473472, 'steps': 20455, 'loss/train': 1.5208839178085327} 02/25/2022 04:04:23 - INFO - codeparrot_training - Step 20456: {'lr': 0.0003387359497851311, 'samples': 10473984, 'steps': 20456, 'loss/train': 0.6904934048652649} 02/25/2022 04:04:28 - INFO - codeparrot_training - Step 20457: {'lr': 0.0003387206525318451, 'samples': 10474496, 'steps': 20457, 'loss/train': 0.7754421234130859} 02/25/2022 04:04:32 - INFO - codeparrot_training - Step 20458: {'lr': 0.00033870535489850806, 'samples': 10475008, 'steps': 20458, 'loss/train': 2.2456939220428467} 02/25/2022 04:04:37 - INFO - codeparrot_training - Step 20459: {'lr': 0.00033869005688518547, 'samples': 10475520, 'steps': 20459, 'loss/train': 1.1401348114013672} 02/25/2022 04:04:41 - INFO - codeparrot_training - Step 20460: {'lr': 0.0003386747584919428, 'samples': 10476032, 'steps': 20460, 'loss/train': 1.9550672769546509} 02/25/2022 04:04:46 - INFO - codeparrot_training - Step 20461: {'lr': 0.0003386594597188456, 'samples': 10476544, 'steps': 20461, 'loss/train': 2.4731264114379883} 02/25/2022 04:04:50 - INFO - codeparrot_training - Step 20462: {'lr': 0.0003386441605659595, 'samples': 10477056, 'steps': 20462, 'loss/train': 2.4738430976867676} 02/25/2022 04:04:55 - INFO - codeparrot_training - Step 20463: {'lr': 0.00033862886103334987, 'samples': 10477568, 'steps': 20463, 'loss/train': 1.706005334854126} 02/25/2022 04:04:59 - INFO - codeparrot_training - Step 20464: {'lr': 0.0003386135611210824, 'samples': 10478080, 'steps': 20464, 'loss/train': 1.3999069929122925} 02/25/2022 04:05:05 - INFO - codeparrot_training - Step 20465: {'lr': 0.0003385982608292226, 'samples': 10478592, 'steps': 20465, 'loss/train': 1.4723130464553833} 02/25/2022 04:05:09 - INFO - codeparrot_training - Step 20466: {'lr': 0.000338582960157836, 'samples': 10479104, 'steps': 20466, 'loss/train': 2.181034803390503} 02/25/2022 04:05:14 - INFO - codeparrot_training - Step 20467: {'lr': 0.000338567659106988, 'samples': 10479616, 'steps': 20467, 'loss/train': 1.3326058387756348} 02/25/2022 04:05:18 - INFO - codeparrot_training - Step 20468: {'lr': 0.0003385523576767444, 'samples': 10480128, 'steps': 20468, 'loss/train': 1.0458500385284424} 02/25/2022 04:05:23 - INFO - codeparrot_training - Step 20469: {'lr': 0.00033853705586717056, 'samples': 10480640, 'steps': 20469, 'loss/train': 1.040419340133667} 02/25/2022 04:05:27 - INFO - codeparrot_training - Step 20470: {'lr': 0.0003385217536783321, 'samples': 10481152, 'steps': 20470, 'loss/train': 2.120584011077881} 02/25/2022 04:05:32 - INFO - codeparrot_training - Step 20471: {'lr': 0.00033850645111029455, 'samples': 10481664, 'steps': 20471, 'loss/train': 1.3446996212005615} 02/25/2022 04:05:36 - INFO - codeparrot_training - Step 20472: {'lr': 0.00033849114816312343, 'samples': 10482176, 'steps': 20472, 'loss/train': 1.733984351158142} 02/25/2022 04:05:41 - INFO - codeparrot_training - Step 20473: {'lr': 0.0003384758448368844, 'samples': 10482688, 'steps': 20473, 'loss/train': 2.818788528442383} 02/25/2022 04:05:45 - INFO - codeparrot_training - Step 20474: {'lr': 0.00033846054113164287, 'samples': 10483200, 'steps': 20474, 'loss/train': 1.5297060012817383} 02/25/2022 04:05:50 - INFO - codeparrot_training - Step 20475: {'lr': 0.00033844523704746455, 'samples': 10483712, 'steps': 20475, 'loss/train': 2.1493990421295166} 02/25/2022 04:05:54 - INFO - codeparrot_training - Step 20476: {'lr': 0.00033842993258441487, 'samples': 10484224, 'steps': 20476, 'loss/train': 2.7441694736480713} 02/25/2022 04:05:59 - INFO - codeparrot_training - Step 20477: {'lr': 0.0003384146277425595, 'samples': 10484736, 'steps': 20477, 'loss/train': 1.7860933542251587} 02/25/2022 04:06:03 - INFO - codeparrot_training - Step 20478: {'lr': 0.0003383993225219638, 'samples': 10485248, 'steps': 20478, 'loss/train': 1.4116884469985962} 02/25/2022 04:06:08 - INFO - codeparrot_training - Step 20479: {'lr': 0.0003383840169226936, 'samples': 10485760, 'steps': 20479, 'loss/train': 2.0376620292663574} 02/25/2022 04:06:12 - INFO - codeparrot_training - Step 20480: {'lr': 0.00033836871094481433, 'samples': 10486272, 'steps': 20480, 'loss/train': 2.797713279724121} 02/25/2022 04:06:19 - INFO - codeparrot_training - Step 20481: {'lr': 0.00033835340458839155, 'samples': 10486784, 'steps': 20481, 'loss/train': 2.2535343170166016} 02/25/2022 04:06:22 - INFO - codeparrot_training - Step 20482: {'lr': 0.0003383380978534908, 'samples': 10487296, 'steps': 20482, 'loss/train': 2.2196648120880127} 02/25/2022 04:06:28 - INFO - codeparrot_training - Step 20483: {'lr': 0.0003383227907401777, 'samples': 10487808, 'steps': 20483, 'loss/train': 2.666508913040161} 02/25/2022 04:06:31 - INFO - codeparrot_training - Step 20484: {'lr': 0.0003383074832485178, 'samples': 10488320, 'steps': 20484, 'loss/train': 1.3228881359100342} 02/25/2022 04:06:37 - INFO - codeparrot_training - Step 20485: {'lr': 0.0003382921753785767, 'samples': 10488832, 'steps': 20485, 'loss/train': 1.1112384796142578} 02/25/2022 04:06:40 - INFO - codeparrot_training - Step 20486: {'lr': 0.0003382768671304199, 'samples': 10489344, 'steps': 20486, 'loss/train': 2.3861844539642334} 02/25/2022 04:06:46 - INFO - codeparrot_training - Step 20487: {'lr': 0.0003382615585041131, 'samples': 10489856, 'steps': 20487, 'loss/train': 2.106295347213745} 02/25/2022 04:06:50 - INFO - codeparrot_training - Step 20488: {'lr': 0.0003382462494997219, 'samples': 10490368, 'steps': 20488, 'loss/train': 2.3959200382232666} 02/25/2022 04:06:55 - INFO - codeparrot_training - Step 20489: {'lr': 0.00033823094011731165, 'samples': 10490880, 'steps': 20489, 'loss/train': 1.983829379081726} 02/25/2022 04:06:59 - INFO - codeparrot_training - Step 20490: {'lr': 0.0003382156303569481, 'samples': 10491392, 'steps': 20490, 'loss/train': 2.0471601486206055} 02/25/2022 04:07:06 - INFO - codeparrot_training - Step 20491: {'lr': 0.0003382003202186968, 'samples': 10491904, 'steps': 20491, 'loss/train': 1.9037071466445923} 02/25/2022 04:07:09 - INFO - codeparrot_training - Step 20492: {'lr': 0.0003381850097026234, 'samples': 10492416, 'steps': 20492, 'loss/train': 2.2065999507904053} 02/25/2022 04:07:15 - INFO - codeparrot_training - Step 20493: {'lr': 0.0003381696988087933, 'samples': 10492928, 'steps': 20493, 'loss/train': 2.1461875438690186} 02/25/2022 04:07:18 - INFO - codeparrot_training - Step 20494: {'lr': 0.00033815438753727236, 'samples': 10493440, 'steps': 20494, 'loss/train': 1.7670491933822632} 02/25/2022 04:07:24 - INFO - codeparrot_training - Step 20495: {'lr': 0.00033813907588812586, 'samples': 10493952, 'steps': 20495, 'loss/train': 2.1959335803985596} 02/25/2022 04:07:27 - INFO - codeparrot_training - Step 20496: {'lr': 0.0003381237638614196, 'samples': 10494464, 'steps': 20496, 'loss/train': 0.7912722826004028} 02/25/2022 04:07:33 - INFO - codeparrot_training - Step 20497: {'lr': 0.00033810845145721915, 'samples': 10494976, 'steps': 20497, 'loss/train': 2.4614434242248535} 02/25/2022 04:07:36 - INFO - codeparrot_training - Step 20498: {'lr': 0.0003380931386755901, 'samples': 10495488, 'steps': 20498, 'loss/train': 1.9480470418930054} 02/25/2022 04:07:42 - INFO - codeparrot_training - Step 20499: {'lr': 0.00033807782551659795, 'samples': 10496000, 'steps': 20499, 'loss/train': 1.1926898956298828} 02/25/2022 04:07:45 - INFO - codeparrot_training - Step 20500: {'lr': 0.0003380625119803084, 'samples': 10496512, 'steps': 20500, 'loss/train': 2.2023305892944336} 02/25/2022 04:07:51 - INFO - codeparrot_training - Step 20501: {'lr': 0.00033804719806678694, 'samples': 10497024, 'steps': 20501, 'loss/train': 1.3211919069290161} 02/25/2022 04:07:55 - INFO - codeparrot_training - Step 20502: {'lr': 0.00033803188377609937, 'samples': 10497536, 'steps': 20502, 'loss/train': 2.2933125495910645} 02/25/2022 04:08:00 - INFO - codeparrot_training - Step 20503: {'lr': 0.00033801656910831116, 'samples': 10498048, 'steps': 20503, 'loss/train': 0.5051615238189697} 02/25/2022 04:08:04 - INFO - codeparrot_training - Step 20504: {'lr': 0.0003380012540634878, 'samples': 10498560, 'steps': 20504, 'loss/train': 0.7014504671096802} 02/25/2022 04:08:09 - INFO - codeparrot_training - Step 20505: {'lr': 0.0003379859386416951, 'samples': 10499072, 'steps': 20505, 'loss/train': 1.2564024925231934} 02/25/2022 04:08:13 - INFO - codeparrot_training - Step 20506: {'lr': 0.00033797062284299847, 'samples': 10499584, 'steps': 20506, 'loss/train': 1.6168928146362305} 02/25/2022 04:08:18 - INFO - codeparrot_training - Step 20507: {'lr': 0.0003379553066674637, 'samples': 10500096, 'steps': 20507, 'loss/train': 0.8878385424613953} 02/25/2022 04:08:22 - INFO - codeparrot_training - Step 20508: {'lr': 0.0003379399901151563, 'samples': 10500608, 'steps': 20508, 'loss/train': 1.9881776571273804} 02/25/2022 04:08:28 - INFO - codeparrot_training - Step 20509: {'lr': 0.00033792467318614195, 'samples': 10501120, 'steps': 20509, 'loss/train': 2.4221129417419434} 02/25/2022 04:08:31 - INFO - codeparrot_training - Step 20510: {'lr': 0.0003379093558804862, 'samples': 10501632, 'steps': 20510, 'loss/train': 1.0279136896133423} 02/25/2022 04:08:37 - INFO - codeparrot_training - Step 20511: {'lr': 0.00033789403819825466, 'samples': 10502144, 'steps': 20511, 'loss/train': 1.7528560161590576} 02/25/2022 04:08:41 - INFO - codeparrot_training - Step 20512: {'lr': 0.00033787872013951297, 'samples': 10502656, 'steps': 20512, 'loss/train': 1.3501797914505005} 02/25/2022 04:08:46 - INFO - codeparrot_training - Step 20513: {'lr': 0.0003378634017043268, 'samples': 10503168, 'steps': 20513, 'loss/train': 1.4831734895706177} 02/25/2022 04:08:50 - INFO - codeparrot_training - Step 20514: {'lr': 0.0003378480828927617, 'samples': 10503680, 'steps': 20514, 'loss/train': 3.4787306785583496} 02/25/2022 04:08:55 - INFO - codeparrot_training - Step 20515: {'lr': 0.00033783276370488323, 'samples': 10504192, 'steps': 20515, 'loss/train': 1.2933127880096436} 02/25/2022 04:08:59 - INFO - codeparrot_training - Step 20516: {'lr': 0.00033781744414075723, 'samples': 10504704, 'steps': 20516, 'loss/train': 1.977235198020935} 02/25/2022 04:09:04 - INFO - codeparrot_training - Step 20517: {'lr': 0.00033780212420044903, 'samples': 10505216, 'steps': 20517, 'loss/train': 2.3855020999908447} 02/25/2022 04:09:08 - INFO - codeparrot_training - Step 20518: {'lr': 0.0003377868038840245, 'samples': 10505728, 'steps': 20518, 'loss/train': 1.3925323486328125} 02/25/2022 04:09:13 - INFO - codeparrot_training - Step 20519: {'lr': 0.00033777148319154917, 'samples': 10506240, 'steps': 20519, 'loss/train': 1.3762468099594116} 02/25/2022 04:09:17 - INFO - codeparrot_training - Step 20520: {'lr': 0.0003377561621230887, 'samples': 10506752, 'steps': 20520, 'loss/train': 1.8767329454421997} 02/25/2022 04:09:22 - INFO - codeparrot_training - Step 20521: {'lr': 0.0003377408406787086, 'samples': 10507264, 'steps': 20521, 'loss/train': 1.657964825630188} 02/25/2022 04:09:26 - INFO - codeparrot_training - Step 20522: {'lr': 0.00033772551885847467, 'samples': 10507776, 'steps': 20522, 'loss/train': 2.1231024265289307} 02/25/2022 04:09:31 - INFO - codeparrot_training - Step 20523: {'lr': 0.0003377101966624525, 'samples': 10508288, 'steps': 20523, 'loss/train': 1.6197705268859863} 02/25/2022 04:09:35 - INFO - codeparrot_training - Step 20524: {'lr': 0.0003376948740907077, 'samples': 10508800, 'steps': 20524, 'loss/train': 0.15179258584976196} 02/25/2022 04:09:40 - INFO - codeparrot_training - Step 20525: {'lr': 0.0003376795511433058, 'samples': 10509312, 'steps': 20525, 'loss/train': 2.460319995880127} 02/25/2022 04:09:44 - INFO - codeparrot_training - Step 20526: {'lr': 0.00033766422782031263, 'samples': 10509824, 'steps': 20526, 'loss/train': 1.6952224969863892} 02/25/2022 04:09:50 - INFO - codeparrot_training - Step 20527: {'lr': 0.0003376489041217937, 'samples': 10510336, 'steps': 20527, 'loss/train': 1.53373122215271} 02/25/2022 04:09:54 - INFO - codeparrot_training - Step 20528: {'lr': 0.00033763358004781474, 'samples': 10510848, 'steps': 20528, 'loss/train': 2.0647308826446533} 02/25/2022 04:09:59 - INFO - codeparrot_training - Step 20529: {'lr': 0.00033761825559844137, 'samples': 10511360, 'steps': 20529, 'loss/train': 1.2797430753707886} 02/25/2022 04:10:03 - INFO - codeparrot_training - Step 20530: {'lr': 0.00033760293077373917, 'samples': 10511872, 'steps': 20530, 'loss/train': 2.235442638397217} 02/25/2022 04:10:08 - INFO - codeparrot_training - Step 20531: {'lr': 0.0003375876055737739, 'samples': 10512384, 'steps': 20531, 'loss/train': 1.8197238445281982} 02/25/2022 04:10:11 - INFO - codeparrot_training - Step 20532: {'lr': 0.000337572279998611, 'samples': 10512896, 'steps': 20532, 'loss/train': 0.5820606350898743} 02/25/2022 04:10:17 - INFO - codeparrot_training - Step 20533: {'lr': 0.00033755695404831635, 'samples': 10513408, 'steps': 20533, 'loss/train': 2.4444525241851807} 02/25/2022 04:10:21 - INFO - codeparrot_training - Step 20534: {'lr': 0.00033754162772295555, 'samples': 10513920, 'steps': 20534, 'loss/train': 2.326087713241577} 02/25/2022 04:10:26 - INFO - codeparrot_training - Step 20535: {'lr': 0.0003375263010225941, 'samples': 10514432, 'steps': 20535, 'loss/train': 2.1294429302215576} 02/25/2022 04:10:30 - INFO - codeparrot_training - Step 20536: {'lr': 0.00033751097394729793, 'samples': 10514944, 'steps': 20536, 'loss/train': 1.9030009508132935} 02/25/2022 04:10:36 - INFO - codeparrot_training - Step 20537: {'lr': 0.0003374956464971324, 'samples': 10515456, 'steps': 20537, 'loss/train': 2.6954758167266846} 02/25/2022 04:10:39 - INFO - codeparrot_training - Step 20538: {'lr': 0.00033748031867216334, 'samples': 10515968, 'steps': 20538, 'loss/train': 0.29608097672462463} 02/25/2022 04:10:45 - INFO - codeparrot_training - Step 20539: {'lr': 0.0003374649904724564, 'samples': 10516480, 'steps': 20539, 'loss/train': 1.9044148921966553} 02/25/2022 04:10:48 - INFO - codeparrot_training - Step 20540: {'lr': 0.0003374496618980772, 'samples': 10516992, 'steps': 20540, 'loss/train': 2.0163896083831787} 02/25/2022 04:10:54 - INFO - codeparrot_training - Step 20541: {'lr': 0.0003374343329490914, 'samples': 10517504, 'steps': 20541, 'loss/train': 1.5897040367126465} 02/25/2022 04:10:58 - INFO - codeparrot_training - Step 20542: {'lr': 0.00033741900362556474, 'samples': 10518016, 'steps': 20542, 'loss/train': 2.1951777935028076} 02/25/2022 04:11:03 - INFO - codeparrot_training - Step 20543: {'lr': 0.00033740367392756274, 'samples': 10518528, 'steps': 20543, 'loss/train': 2.0441653728485107} 02/25/2022 04:11:07 - INFO - codeparrot_training - Step 20544: {'lr': 0.0003373883438551512, 'samples': 10519040, 'steps': 20544, 'loss/train': 1.952451467514038} 02/25/2022 04:11:12 - INFO - codeparrot_training - Step 20545: {'lr': 0.0003373730134083958, 'samples': 10519552, 'steps': 20545, 'loss/train': 1.6012649536132812} 02/25/2022 04:11:16 - INFO - codeparrot_training - Step 20546: {'lr': 0.00033735768258736206, 'samples': 10520064, 'steps': 20546, 'loss/train': 3.1981472969055176} 02/25/2022 04:11:23 - INFO - codeparrot_training - Step 20547: {'lr': 0.0003373423513921158, 'samples': 10520576, 'steps': 20547, 'loss/train': 2.0102579593658447} 02/25/2022 04:11:26 - INFO - codeparrot_training - Step 20548: {'lr': 0.00033732701982272257, 'samples': 10521088, 'steps': 20548, 'loss/train': 2.3970634937286377} 02/25/2022 04:11:32 - INFO - codeparrot_training - Step 20549: {'lr': 0.0003373116878792482, 'samples': 10521600, 'steps': 20549, 'loss/train': 2.4481210708618164} 02/25/2022 04:11:35 - INFO - codeparrot_training - Step 20550: {'lr': 0.00033729635556175826, 'samples': 10522112, 'steps': 20550, 'loss/train': 1.7564014196395874} 02/25/2022 04:11:41 - INFO - codeparrot_training - Step 20551: {'lr': 0.0003372810228703184, 'samples': 10522624, 'steps': 20551, 'loss/train': 1.8961580991744995} 02/25/2022 04:11:44 - INFO - codeparrot_training - Step 20552: {'lr': 0.0003372656898049944, 'samples': 10523136, 'steps': 20552, 'loss/train': 2.156228542327881} 02/25/2022 04:11:50 - INFO - codeparrot_training - Step 20553: {'lr': 0.00033725035636585196, 'samples': 10523648, 'steps': 20553, 'loss/train': 2.398817777633667} 02/25/2022 04:11:54 - INFO - codeparrot_training - Step 20554: {'lr': 0.0003372350225529566, 'samples': 10524160, 'steps': 20554, 'loss/train': 1.324874997138977} 02/25/2022 04:11:59 - INFO - codeparrot_training - Step 20555: {'lr': 0.00033721968836637414, 'samples': 10524672, 'steps': 20555, 'loss/train': 0.8835898041725159} 02/25/2022 04:12:03 - INFO - codeparrot_training - Step 20556: {'lr': 0.0003372043538061702, 'samples': 10525184, 'steps': 20556, 'loss/train': 2.0057356357574463} 02/25/2022 04:12:08 - INFO - codeparrot_training - Step 20557: {'lr': 0.00033718901887241047, 'samples': 10525696, 'steps': 20557, 'loss/train': 2.072836399078369} 02/25/2022 04:12:12 - INFO - codeparrot_training - Step 20558: {'lr': 0.00033717368356516077, 'samples': 10526208, 'steps': 20558, 'loss/train': 1.0213122367858887} 02/25/2022 04:12:19 - INFO - codeparrot_training - Step 20559: {'lr': 0.0003371583478844866, 'samples': 10526720, 'steps': 20559, 'loss/train': 1.5481805801391602} 02/25/2022 04:12:22 - INFO - codeparrot_training - Step 20560: {'lr': 0.0003371430118304538, 'samples': 10527232, 'steps': 20560, 'loss/train': 2.0219507217407227} 02/25/2022 04:12:28 - INFO - codeparrot_training - Step 20561: {'lr': 0.000337127675403128, 'samples': 10527744, 'steps': 20561, 'loss/train': 2.5098512172698975} 02/25/2022 04:12:31 - INFO - codeparrot_training - Step 20562: {'lr': 0.00033711233860257494, 'samples': 10528256, 'steps': 20562, 'loss/train': 1.6744624376296997} 02/25/2022 04:12:37 - INFO - codeparrot_training - Step 20563: {'lr': 0.00033709700142886025, 'samples': 10528768, 'steps': 20563, 'loss/train': 2.8262696266174316} 02/25/2022 04:12:40 - INFO - codeparrot_training - Step 20564: {'lr': 0.00033708166388204963, 'samples': 10529280, 'steps': 20564, 'loss/train': 2.3715226650238037} 02/25/2022 04:12:46 - INFO - codeparrot_training - Step 20565: {'lr': 0.00033706632596220885, 'samples': 10529792, 'steps': 20565, 'loss/train': 2.365448474884033} 02/25/2022 04:12:49 - INFO - codeparrot_training - Step 20566: {'lr': 0.00033705098766940354, 'samples': 10530304, 'steps': 20566, 'loss/train': 1.470035433769226} 02/25/2022 04:12:55 - INFO - codeparrot_training - Step 20567: {'lr': 0.00033703564900369943, 'samples': 10530816, 'steps': 20567, 'loss/train': 2.040496826171875} 02/25/2022 04:12:58 - INFO - codeparrot_training - Step 20568: {'lr': 0.0003370203099651623, 'samples': 10531328, 'steps': 20568, 'loss/train': 1.470913052558899} 02/25/2022 04:13:04 - INFO - codeparrot_training - Step 20569: {'lr': 0.0003370049705538578, 'samples': 10531840, 'steps': 20569, 'loss/train': 2.430475950241089} 02/25/2022 04:13:07 - INFO - codeparrot_training - Step 20570: {'lr': 0.00033698963076985155, 'samples': 10532352, 'steps': 20570, 'loss/train': 1.7727771997451782} 02/25/2022 04:13:13 - INFO - codeparrot_training - Step 20571: {'lr': 0.00033697429061320936, 'samples': 10532864, 'steps': 20571, 'loss/train': 1.7067110538482666} 02/25/2022 04:13:16 - INFO - codeparrot_training - Step 20572: {'lr': 0.00033695895008399704, 'samples': 10533376, 'steps': 20572, 'loss/train': 2.1659817695617676} 02/25/2022 04:13:23 - INFO - codeparrot_training - Step 20573: {'lr': 0.00033694360918228006, 'samples': 10533888, 'steps': 20573, 'loss/train': 1.9703102111816406} 02/25/2022 04:13:27 - INFO - codeparrot_training - Step 20574: {'lr': 0.0003369282679081243, 'samples': 10534400, 'steps': 20574, 'loss/train': 1.6992661952972412} 02/25/2022 04:13:32 - INFO - codeparrot_training - Step 20575: {'lr': 0.00033691292626159556, 'samples': 10534912, 'steps': 20575, 'loss/train': 1.243703007698059} 02/25/2022 04:13:36 - INFO - codeparrot_training - Step 20576: {'lr': 0.0003368975842427592, 'samples': 10535424, 'steps': 20576, 'loss/train': 1.986406683921814} 02/25/2022 04:13:41 - INFO - codeparrot_training - Step 20577: {'lr': 0.00033688224185168136, 'samples': 10535936, 'steps': 20577, 'loss/train': 1.7609435319900513} 02/25/2022 04:13:45 - INFO - codeparrot_training - Step 20578: {'lr': 0.0003368668990884275, 'samples': 10536448, 'steps': 20578, 'loss/train': 3.246948719024658} 02/25/2022 04:13:50 - INFO - codeparrot_training - Step 20579: {'lr': 0.0003368515559530635, 'samples': 10536960, 'steps': 20579, 'loss/train': 1.6103166341781616} 02/25/2022 04:13:54 - INFO - codeparrot_training - Step 20580: {'lr': 0.00033683621244565506, 'samples': 10537472, 'steps': 20580, 'loss/train': 2.2472264766693115} 02/25/2022 04:13:59 - INFO - codeparrot_training - Step 20581: {'lr': 0.00033682086856626773, 'samples': 10537984, 'steps': 20581, 'loss/train': 1.9969643354415894} 02/25/2022 04:14:03 - INFO - codeparrot_training - Step 20582: {'lr': 0.00033680552431496744, 'samples': 10538496, 'steps': 20582, 'loss/train': 1.7149379253387451} 02/25/2022 04:14:10 - INFO - codeparrot_training - Step 20583: {'lr': 0.0003367901796918198, 'samples': 10539008, 'steps': 20583, 'loss/train': 2.339693784713745} 02/25/2022 04:14:13 - INFO - codeparrot_training - Step 20584: {'lr': 0.0003367748346968907, 'samples': 10539520, 'steps': 20584, 'loss/train': 2.123481512069702} 02/25/2022 04:14:19 - INFO - codeparrot_training - Step 20585: {'lr': 0.0003367594893302457, 'samples': 10540032, 'steps': 20585, 'loss/train': 1.932574987411499} 02/25/2022 04:14:22 - INFO - codeparrot_training - Step 20586: {'lr': 0.00033674414359195055, 'samples': 10540544, 'steps': 20586, 'loss/train': 1.6969020366668701} 02/25/2022 04:14:28 - INFO - codeparrot_training - Step 20587: {'lr': 0.00033672879748207105, 'samples': 10541056, 'steps': 20587, 'loss/train': 3.1708507537841797} 02/25/2022 04:14:31 - INFO - codeparrot_training - Step 20588: {'lr': 0.000336713451000673, 'samples': 10541568, 'steps': 20588, 'loss/train': 1.717882513999939} 02/25/2022 04:14:37 - INFO - codeparrot_training - Step 20589: {'lr': 0.000336698104147822, 'samples': 10542080, 'steps': 20589, 'loss/train': 1.6928091049194336} 02/25/2022 04:14:40 - INFO - codeparrot_training - Step 20590: {'lr': 0.0003366827569235838, 'samples': 10542592, 'steps': 20590, 'loss/train': 2.0823044776916504} 02/25/2022 04:14:46 - INFO - codeparrot_training - Step 20591: {'lr': 0.00033666740932802424, 'samples': 10543104, 'steps': 20591, 'loss/train': 1.821470022201538} 02/25/2022 04:14:49 - INFO - codeparrot_training - Step 20592: {'lr': 0.0003366520613612091, 'samples': 10543616, 'steps': 20592, 'loss/train': 0.9725879430770874} 02/25/2022 04:14:56 - INFO - codeparrot_training - Step 20593: {'lr': 0.00033663671302320387, 'samples': 10544128, 'steps': 20593, 'loss/train': 2.4706289768218994} 02/25/2022 04:14:59 - INFO - codeparrot_training - Step 20594: {'lr': 0.0003366213643140745, 'samples': 10544640, 'steps': 20594, 'loss/train': 0.7303314805030823} 02/25/2022 04:15:05 - INFO - codeparrot_training - Step 20595: {'lr': 0.0003366060152338869, 'samples': 10545152, 'steps': 20595, 'loss/train': 1.5464123487472534} 02/25/2022 04:15:08 - INFO - codeparrot_training - Step 20596: {'lr': 0.0003365906657827065, 'samples': 10545664, 'steps': 20596, 'loss/train': 1.7940410375595093} 02/25/2022 04:15:14 - INFO - codeparrot_training - Step 20597: {'lr': 0.00033657531596059914, 'samples': 10546176, 'steps': 20597, 'loss/train': 0.9959035515785217} 02/25/2022 04:15:17 - INFO - codeparrot_training - Step 20598: {'lr': 0.00033655996576763064, 'samples': 10546688, 'steps': 20598, 'loss/train': 1.9002448320388794} 02/25/2022 04:15:23 - INFO - codeparrot_training - Step 20599: {'lr': 0.00033654461520386683, 'samples': 10547200, 'steps': 20599, 'loss/train': 1.2800064086914062} 02/25/2022 04:15:26 - INFO - codeparrot_training - Step 20600: {'lr': 0.0003365292642693733, 'samples': 10547712, 'steps': 20600, 'loss/train': 1.8706591129302979} 02/25/2022 04:15:32 - INFO - codeparrot_training - Step 20601: {'lr': 0.00033651391296421587, 'samples': 10548224, 'steps': 20601, 'loss/train': 2.04245924949646} 02/25/2022 04:15:35 - INFO - codeparrot_training - Step 20602: {'lr': 0.0003364985612884603, 'samples': 10548736, 'steps': 20602, 'loss/train': 1.962165355682373} 02/25/2022 04:15:41 - INFO - codeparrot_training - Step 20603: {'lr': 0.0003364832092421724, 'samples': 10549248, 'steps': 20603, 'loss/train': 2.599982261657715} 02/25/2022 04:15:44 - INFO - codeparrot_training - Step 20604: {'lr': 0.0003364678568254178, 'samples': 10549760, 'steps': 20604, 'loss/train': 2.2657630443573} 02/25/2022 04:15:50 - INFO - codeparrot_training - Step 20605: {'lr': 0.00033645250403826246, 'samples': 10550272, 'steps': 20605, 'loss/train': 2.3945748805999756} 02/25/2022 04:15:54 - INFO - codeparrot_training - Step 20606: {'lr': 0.00033643715088077205, 'samples': 10550784, 'steps': 20606, 'loss/train': 1.7973207235336304} 02/25/2022 04:15:59 - INFO - codeparrot_training - Step 20607: {'lr': 0.0003364217973530123, 'samples': 10551296, 'steps': 20607, 'loss/train': 1.2333757877349854} 02/25/2022 04:16:03 - INFO - codeparrot_training - Step 20608: {'lr': 0.00033640644345504897, 'samples': 10551808, 'steps': 20608, 'loss/train': 2.023709535598755} 02/25/2022 04:16:08 - INFO - codeparrot_training - Step 20609: {'lr': 0.0003363910891869479, 'samples': 10552320, 'steps': 20609, 'loss/train': 1.9283435344696045} 02/25/2022 04:16:12 - INFO - codeparrot_training - Step 20610: {'lr': 0.00033637573454877486, 'samples': 10552832, 'steps': 20610, 'loss/train': 2.3888638019561768} 02/25/2022 04:16:17 - INFO - codeparrot_training - Step 20611: {'lr': 0.0003363603795405956, 'samples': 10553344, 'steps': 20611, 'loss/train': 1.7015565633773804} 02/25/2022 04:16:24 - INFO - codeparrot_training - Step 20612: {'lr': 0.0003363450241624759, 'samples': 10553856, 'steps': 20612, 'loss/train': 1.7890573740005493} 02/25/2022 04:16:27 - INFO - codeparrot_training - Step 20613: {'lr': 0.00033632966841448144, 'samples': 10554368, 'steps': 20613, 'loss/train': 1.6532474756240845} 02/25/2022 04:16:30 - INFO - codeparrot_training - Step 20614: {'lr': 0.0003363143122966782, 'samples': 10554880, 'steps': 20614, 'loss/train': 0.4228179454803467} 02/25/2022 04:16:38 - INFO - codeparrot_training - Step 20615: {'lr': 0.00033629895580913183, 'samples': 10555392, 'steps': 20615, 'loss/train': 1.9388201236724854} 02/25/2022 04:16:41 - INFO - codeparrot_training - Step 20616: {'lr': 0.00033628359895190814, 'samples': 10555904, 'steps': 20616, 'loss/train': 5.145500183105469} 02/25/2022 04:16:47 - INFO - codeparrot_training - Step 20617: {'lr': 0.0003362682417250729, 'samples': 10556416, 'steps': 20617, 'loss/train': 1.5483806133270264} 02/25/2022 04:16:50 - INFO - codeparrot_training - Step 20618: {'lr': 0.00033625288412869185, 'samples': 10556928, 'steps': 20618, 'loss/train': 1.572220802307129} 02/25/2022 04:16:56 - INFO - codeparrot_training - Step 20619: {'lr': 0.00033623752616283094, 'samples': 10557440, 'steps': 20619, 'loss/train': 3.715501546859741} 02/25/2022 04:16:59 - INFO - codeparrot_training - Step 20620: {'lr': 0.0003362221678275558, 'samples': 10557952, 'steps': 20620, 'loss/train': 2.956888198852539} 02/25/2022 04:17:05 - INFO - codeparrot_training - Step 20621: {'lr': 0.00033620680912293226, 'samples': 10558464, 'steps': 20621, 'loss/train': 1.7934989929199219} 02/25/2022 04:17:08 - INFO - codeparrot_training - Step 20622: {'lr': 0.0003361914500490261, 'samples': 10558976, 'steps': 20622, 'loss/train': 1.4772794246673584} 02/25/2022 04:17:14 - INFO - codeparrot_training - Step 20623: {'lr': 0.0003361760906059032, 'samples': 10559488, 'steps': 20623, 'loss/train': 1.341872215270996} 02/25/2022 04:17:17 - INFO - codeparrot_training - Step 20624: {'lr': 0.00033616073079362923, 'samples': 10560000, 'steps': 20624, 'loss/train': 0.7913236021995544} 02/25/2022 04:17:23 - INFO - codeparrot_training - Step 20625: {'lr': 0.00033614537061227014, 'samples': 10560512, 'steps': 20625, 'loss/train': 1.3911631107330322} 02/25/2022 04:17:26 - INFO - codeparrot_training - Step 20626: {'lr': 0.00033613001006189156, 'samples': 10561024, 'steps': 20626, 'loss/train': 2.240656852722168} 02/25/2022 04:17:32 - INFO - codeparrot_training - Step 20627: {'lr': 0.0003361146491425594, 'samples': 10561536, 'steps': 20627, 'loss/train': 2.7742202281951904} 02/25/2022 04:17:35 - INFO - codeparrot_training - Step 20628: {'lr': 0.0003360992878543394, 'samples': 10562048, 'steps': 20628, 'loss/train': 2.2700459957122803} 02/25/2022 04:17:42 - INFO - codeparrot_training - Step 20629: {'lr': 0.00033608392619729745, 'samples': 10562560, 'steps': 20629, 'loss/train': 2.4273712635040283} 02/25/2022 04:17:46 - INFO - codeparrot_training - Step 20630: {'lr': 0.0003360685641714992, 'samples': 10563072, 'steps': 20630, 'loss/train': 2.5046215057373047} 02/25/2022 04:17:51 - INFO - codeparrot_training - Step 20631: {'lr': 0.0003360532017770106, 'samples': 10563584, 'steps': 20631, 'loss/train': 1.9252820014953613} 02/25/2022 04:17:55 - INFO - codeparrot_training - Step 20632: {'lr': 0.00033603783901389744, 'samples': 10564096, 'steps': 20632, 'loss/train': 2.424588918685913} 02/25/2022 04:18:00 - INFO - codeparrot_training - Step 20633: {'lr': 0.00033602247588222545, 'samples': 10564608, 'steps': 20633, 'loss/train': 1.0199012756347656} 02/25/2022 04:18:04 - INFO - codeparrot_training - Step 20634: {'lr': 0.00033600711238206056, 'samples': 10565120, 'steps': 20634, 'loss/train': 2.2843220233917236} 02/25/2022 04:18:09 - INFO - codeparrot_training - Step 20635: {'lr': 0.00033599174851346844, 'samples': 10565632, 'steps': 20635, 'loss/train': 2.708117723464966} 02/25/2022 04:18:13 - INFO - codeparrot_training - Step 20636: {'lr': 0.000335976384276515, 'samples': 10566144, 'steps': 20636, 'loss/train': 0.8021120429039001} 02/25/2022 04:18:18 - INFO - codeparrot_training - Step 20637: {'lr': 0.000335961019671266, 'samples': 10566656, 'steps': 20637, 'loss/train': 2.1264638900756836} 02/25/2022 04:18:22 - INFO - codeparrot_training - Step 20638: {'lr': 0.00033594565469778734, 'samples': 10567168, 'steps': 20638, 'loss/train': 1.562037706375122} 02/25/2022 04:18:29 - INFO - codeparrot_training - Step 20639: {'lr': 0.00033593028935614475, 'samples': 10567680, 'steps': 20639, 'loss/train': 2.5130465030670166} 02/25/2022 04:18:32 - INFO - codeparrot_training - Step 20640: {'lr': 0.0003359149236464041, 'samples': 10568192, 'steps': 20640, 'loss/train': 2.2507386207580566} 02/25/2022 04:18:38 - INFO - codeparrot_training - Step 20641: {'lr': 0.0003358995575686311, 'samples': 10568704, 'steps': 20641, 'loss/train': 3.4444360733032227} 02/25/2022 04:18:41 - INFO - codeparrot_training - Step 20642: {'lr': 0.00033588419112289176, 'samples': 10569216, 'steps': 20642, 'loss/train': 1.8604896068572998} 02/25/2022 04:18:47 - INFO - codeparrot_training - Step 20643: {'lr': 0.00033586882430925184, 'samples': 10569728, 'steps': 20643, 'loss/train': 1.009259819984436} 02/25/2022 04:18:50 - INFO - codeparrot_training - Step 20644: {'lr': 0.0003358534571277771, 'samples': 10570240, 'steps': 20644, 'loss/train': 1.8278828859329224} 02/25/2022 04:18:56 - INFO - codeparrot_training - Step 20645: {'lr': 0.0003358380895785335, 'samples': 10570752, 'steps': 20645, 'loss/train': 0.7926627397537231} 02/25/2022 04:19:00 - INFO - codeparrot_training - Step 20646: {'lr': 0.00033582272166158666, 'samples': 10571264, 'steps': 20646, 'loss/train': 2.104950428009033} 02/25/2022 04:19:05 - INFO - codeparrot_training - Step 20647: {'lr': 0.00033580735337700266, 'samples': 10571776, 'steps': 20647, 'loss/train': 1.9371141195297241} 02/25/2022 04:19:09 - INFO - codeparrot_training - Step 20648: {'lr': 0.00033579198472484707, 'samples': 10572288, 'steps': 20648, 'loss/train': 2.2995235919952393} 02/25/2022 04:19:14 - INFO - codeparrot_training - Step 20649: {'lr': 0.0003357766157051859, 'samples': 10572800, 'steps': 20649, 'loss/train': 1.6594722270965576} 02/25/2022 04:19:18 - INFO - codeparrot_training - Step 20650: {'lr': 0.000335761246318085, 'samples': 10573312, 'steps': 20650, 'loss/train': 2.056654930114746} 02/25/2022 04:19:24 - INFO - codeparrot_training - Step 20651: {'lr': 0.0003357458765636101, 'samples': 10573824, 'steps': 20651, 'loss/train': 1.4294756650924683} 02/25/2022 04:19:28 - INFO - codeparrot_training - Step 20652: {'lr': 0.00033573050644182713, 'samples': 10574336, 'steps': 20652, 'loss/train': 2.060487985610962} 02/25/2022 04:19:33 - INFO - codeparrot_training - Step 20653: {'lr': 0.00033571513595280185, 'samples': 10574848, 'steps': 20653, 'loss/train': 2.9714717864990234} 02/25/2022 04:19:37 - INFO - codeparrot_training - Step 20654: {'lr': 0.0003356997650966002, 'samples': 10575360, 'steps': 20654, 'loss/train': 2.1979050636291504} 02/25/2022 04:19:40 - INFO - codeparrot_training - Step 20655: {'lr': 0.0003356843938732879, 'samples': 10575872, 'steps': 20655, 'loss/train': 1.7492620944976807} 02/25/2022 04:19:46 - INFO - codeparrot_training - Step 20656: {'lr': 0.0003356690222829309, 'samples': 10576384, 'steps': 20656, 'loss/train': 1.6251932382583618} 02/25/2022 04:19:50 - INFO - codeparrot_training - Step 20657: {'lr': 0.00033565365032559496, 'samples': 10576896, 'steps': 20657, 'loss/train': 1.6758636236190796} 02/25/2022 04:19:55 - INFO - codeparrot_training - Step 20658: {'lr': 0.00033563827800134604, 'samples': 10577408, 'steps': 20658, 'loss/train': 1.3154969215393066} 02/25/2022 04:19:58 - INFO - codeparrot_training - Step 20659: {'lr': 0.00033562290531025, 'samples': 10577920, 'steps': 20659, 'loss/train': 0.9799546003341675} 02/25/2022 04:20:06 - INFO - codeparrot_training - Step 20660: {'lr': 0.0003356075322523725, 'samples': 10578432, 'steps': 20660, 'loss/train': 2.0282552242279053} 02/25/2022 04:20:09 - INFO - codeparrot_training - Step 20661: {'lr': 0.00033559215882777955, 'samples': 10578944, 'steps': 20661, 'loss/train': 8.620622634887695} 02/25/2022 04:20:14 - INFO - codeparrot_training - Step 20662: {'lr': 0.000335576785036537, 'samples': 10579456, 'steps': 20662, 'loss/train': 2.532790184020996} 02/25/2022 04:20:18 - INFO - codeparrot_training - Step 20663: {'lr': 0.0003355614108787106, 'samples': 10579968, 'steps': 20663, 'loss/train': 2.788933753967285} 02/25/2022 04:20:23 - INFO - codeparrot_training - Step 20664: {'lr': 0.00033554603635436645, 'samples': 10580480, 'steps': 20664, 'loss/train': 2.4055988788604736} 02/25/2022 04:20:29 - INFO - codeparrot_training - Step 20665: {'lr': 0.00033553066146357, 'samples': 10580992, 'steps': 20665, 'loss/train': 2.019536256790161} 02/25/2022 04:20:33 - INFO - codeparrot_training - Step 20666: {'lr': 0.0003355152862063875, 'samples': 10581504, 'steps': 20666, 'loss/train': 2.2609477043151855} 02/25/2022 04:20:38 - INFO - codeparrot_training - Step 20667: {'lr': 0.0003354999105828847, 'samples': 10582016, 'steps': 20667, 'loss/train': 2.2286131381988525} 02/25/2022 04:20:42 - INFO - codeparrot_training - Step 20668: {'lr': 0.0003354845345931274, 'samples': 10582528, 'steps': 20668, 'loss/train': 1.742092490196228} 02/25/2022 04:20:49 - INFO - codeparrot_training - Step 20669: {'lr': 0.0003354691582371815, 'samples': 10583040, 'steps': 20669, 'loss/train': 2.0620768070220947} 02/25/2022 04:20:53 - INFO - codeparrot_training - Step 20670: {'lr': 0.00033545378151511287, 'samples': 10583552, 'steps': 20670, 'loss/train': 1.5710407495498657} 02/25/2022 04:20:58 - INFO - codeparrot_training - Step 20671: {'lr': 0.00033543840442698737, 'samples': 10584064, 'steps': 20671, 'loss/train': 1.565824270248413} 02/25/2022 04:21:02 - INFO - codeparrot_training - Step 20672: {'lr': 0.0003354230269728709, 'samples': 10584576, 'steps': 20672, 'loss/train': 1.8649389743804932} 02/25/2022 04:21:07 - INFO - codeparrot_training - Step 20673: {'lr': 0.0003354076491528292, 'samples': 10585088, 'steps': 20673, 'loss/train': 1.282605528831482} 02/25/2022 04:21:11 - INFO - codeparrot_training - Step 20674: {'lr': 0.00033539227096692837, 'samples': 10585600, 'steps': 20674, 'loss/train': 2.198899269104004} 02/25/2022 04:21:16 - INFO - codeparrot_training - Step 20675: {'lr': 0.00033537689241523407, 'samples': 10586112, 'steps': 20675, 'loss/train': 1.7681479454040527} 02/25/2022 04:21:20 - INFO - codeparrot_training - Step 20676: {'lr': 0.00033536151349781236, 'samples': 10586624, 'steps': 20676, 'loss/train': 2.2473134994506836} 02/25/2022 04:21:25 - INFO - codeparrot_training - Step 20677: {'lr': 0.000335346134214729, 'samples': 10587136, 'steps': 20677, 'loss/train': 2.4820892810821533} 02/25/2022 04:21:29 - INFO - codeparrot_training - Step 20678: {'lr': 0.0003353307545660499, 'samples': 10587648, 'steps': 20678, 'loss/train': 1.0811653137207031} 02/25/2022 04:21:36 - INFO - codeparrot_training - Step 20679: {'lr': 0.0003353153745518409, 'samples': 10588160, 'steps': 20679, 'loss/train': 2.558427095413208} 02/25/2022 04:21:40 - INFO - codeparrot_training - Step 20680: {'lr': 0.000335299994172168, 'samples': 10588672, 'steps': 20680, 'loss/train': 1.893054723739624} 02/25/2022 04:21:45 - INFO - codeparrot_training - Step 20681: {'lr': 0.0003352846134270969, 'samples': 10589184, 'steps': 20681, 'loss/train': 2.0280046463012695} 02/25/2022 04:21:49 - INFO - codeparrot_training - Step 20682: {'lr': 0.0003352692323166938, 'samples': 10589696, 'steps': 20682, 'loss/train': 1.0919030904769897} 02/25/2022 04:21:52 - INFO - codeparrot_training - Step 20683: {'lr': 0.0003352538508410242, 'samples': 10590208, 'steps': 20683, 'loss/train': 0.13893944025039673} 02/25/2022 04:21:58 - INFO - codeparrot_training - Step 20684: {'lr': 0.00033523846900015427, 'samples': 10590720, 'steps': 20684, 'loss/train': 1.4504388570785522} 02/25/2022 04:22:03 - INFO - codeparrot_training - Step 20685: {'lr': 0.0003352230867941497, 'samples': 10591232, 'steps': 20685, 'loss/train': 1.847858190536499} 02/25/2022 04:22:07 - INFO - codeparrot_training - Step 20686: {'lr': 0.0003352077042230766, 'samples': 10591744, 'steps': 20686, 'loss/train': 2.4221556186676025} 02/25/2022 04:22:12 - INFO - codeparrot_training - Step 20687: {'lr': 0.0003351923212870007, 'samples': 10592256, 'steps': 20687, 'loss/train': 1.736351728439331} 02/25/2022 04:22:16 - INFO - codeparrot_training - Step 20688: {'lr': 0.000335176937985988, 'samples': 10592768, 'steps': 20688, 'loss/train': 2.496405839920044} 02/25/2022 04:22:21 - INFO - codeparrot_training - Step 20689: {'lr': 0.0003351615543201042, 'samples': 10593280, 'steps': 20689, 'loss/train': 1.488968014717102} 02/25/2022 04:22:25 - INFO - codeparrot_training - Step 20690: {'lr': 0.00033514617028941547, 'samples': 10593792, 'steps': 20690, 'loss/train': 1.6677732467651367} 02/25/2022 04:22:32 - INFO - codeparrot_training - Step 20691: {'lr': 0.0003351307858939875, 'samples': 10594304, 'steps': 20691, 'loss/train': 1.5482953786849976} 02/25/2022 04:22:35 - INFO - codeparrot_training - Step 20692: {'lr': 0.0003351154011338864, 'samples': 10594816, 'steps': 20692, 'loss/train': 1.4778696298599243} 02/25/2022 04:22:41 - INFO - codeparrot_training - Step 20693: {'lr': 0.00033510001600917783, 'samples': 10595328, 'steps': 20693, 'loss/train': 2.131817579269409} 02/25/2022 04:22:44 - INFO - codeparrot_training - Step 20694: {'lr': 0.00033508463051992786, 'samples': 10595840, 'steps': 20694, 'loss/train': 0.7110695242881775} 02/25/2022 04:22:50 - INFO - codeparrot_training - Step 20695: {'lr': 0.00033506924466620235, 'samples': 10596352, 'steps': 20695, 'loss/train': 0.950127363204956} 02/25/2022 04:22:53 - INFO - codeparrot_training - Step 20696: {'lr': 0.0003350538584480672, 'samples': 10596864, 'steps': 20696, 'loss/train': 0.5178962349891663} 02/25/2022 04:22:59 - INFO - codeparrot_training - Step 20697: {'lr': 0.0003350384718655884, 'samples': 10597376, 'steps': 20697, 'loss/train': 2.0694234371185303} 02/25/2022 04:23:03 - INFO - codeparrot_training - Step 20698: {'lr': 0.0003350230849188317, 'samples': 10597888, 'steps': 20698, 'loss/train': 2.4739716053009033} 02/25/2022 04:23:08 - INFO - codeparrot_training - Step 20699: {'lr': 0.00033500769760786314, 'samples': 10598400, 'steps': 20699, 'loss/train': 3.235154867172241} 02/25/2022 04:23:12 - INFO - codeparrot_training - Step 20700: {'lr': 0.0003349923099327485, 'samples': 10598912, 'steps': 20700, 'loss/train': 1.994461178779602} 02/25/2022 04:23:19 - INFO - codeparrot_training - Step 20701: {'lr': 0.000334976921893554, 'samples': 10599424, 'steps': 20701, 'loss/train': 1.927176594734192} 02/25/2022 04:23:22 - INFO - codeparrot_training - Step 20702: {'lr': 0.0003349615334903452, 'samples': 10599936, 'steps': 20702, 'loss/train': 2.4826650619506836} 02/25/2022 04:23:28 - INFO - codeparrot_training - Step 20703: {'lr': 0.00033494614472318816, 'samples': 10600448, 'steps': 20703, 'loss/train': 1.3483468294143677} 02/25/2022 04:23:31 - INFO - codeparrot_training - Step 20704: {'lr': 0.00033493075559214885, 'samples': 10600960, 'steps': 20704, 'loss/train': 1.7731810808181763} 02/25/2022 04:23:37 - INFO - codeparrot_training - Step 20705: {'lr': 0.00033491536609729313, 'samples': 10601472, 'steps': 20705, 'loss/train': 1.9997228384017944} 02/25/2022 04:23:40 - INFO - codeparrot_training - Step 20706: {'lr': 0.000334899976238687, 'samples': 10601984, 'steps': 20706, 'loss/train': 1.9523364305496216} 02/25/2022 04:23:46 - INFO - codeparrot_training - Step 20707: {'lr': 0.00033488458601639624, 'samples': 10602496, 'steps': 20707, 'loss/train': 2.0029444694519043} 02/25/2022 04:23:49 - INFO - codeparrot_training - Step 20708: {'lr': 0.000334869195430487, 'samples': 10603008, 'steps': 20708, 'loss/train': 1.8645951747894287} 02/25/2022 04:23:55 - INFO - codeparrot_training - Step 20709: {'lr': 0.00033485380448102496, 'samples': 10603520, 'steps': 20709, 'loss/train': 2.003495931625366} 02/25/2022 04:23:58 - INFO - codeparrot_training - Step 20710: {'lr': 0.0003348384131680762, 'samples': 10604032, 'steps': 20710, 'loss/train': 1.5402758121490479} 02/25/2022 04:24:04 - INFO - codeparrot_training - Step 20711: {'lr': 0.0003348230214917066, 'samples': 10604544, 'steps': 20711, 'loss/train': 2.003594160079956} 02/25/2022 04:24:07 - INFO - codeparrot_training - Step 20712: {'lr': 0.0003348076294519822, 'samples': 10605056, 'steps': 20712, 'loss/train': 2.126431465148926} 02/25/2022 04:24:13 - INFO - codeparrot_training - Step 20713: {'lr': 0.0003347922370489687, 'samples': 10605568, 'steps': 20713, 'loss/train': 2.2130186557769775} 02/25/2022 04:24:16 - INFO - codeparrot_training - Step 20714: {'lr': 0.00033477684428273233, 'samples': 10606080, 'steps': 20714, 'loss/train': 0.691471517086029} 02/25/2022 04:24:22 - INFO - codeparrot_training - Step 20715: {'lr': 0.0003347614511533388, 'samples': 10606592, 'steps': 20715, 'loss/train': 3.0492777824401855} 02/25/2022 04:24:25 - INFO - codeparrot_training - Step 20716: {'lr': 0.0003347460576608541, 'samples': 10607104, 'steps': 20716, 'loss/train': 2.4379959106445312} 02/25/2022 04:24:32 - INFO - codeparrot_training - Step 20717: {'lr': 0.00033473066380534423, 'samples': 10607616, 'steps': 20717, 'loss/train': 2.1472010612487793} 02/25/2022 04:24:36 - INFO - codeparrot_training - Step 20718: {'lr': 0.00033471526958687514, 'samples': 10608128, 'steps': 20718, 'loss/train': 2.1485254764556885} 02/25/2022 04:24:41 - INFO - codeparrot_training - Step 20719: {'lr': 0.0003346998750055127, 'samples': 10608640, 'steps': 20719, 'loss/train': 0.8854549527168274} 02/25/2022 04:24:45 - INFO - codeparrot_training - Step 20720: {'lr': 0.0003346844800613229, 'samples': 10609152, 'steps': 20720, 'loss/train': 2.275102138519287} 02/25/2022 04:24:50 - INFO - codeparrot_training - Step 20721: {'lr': 0.0003346690847543717, 'samples': 10609664, 'steps': 20721, 'loss/train': 1.079146385192871} 02/25/2022 04:24:54 - INFO - codeparrot_training - Step 20722: {'lr': 0.00033465368908472496, 'samples': 10610176, 'steps': 20722, 'loss/train': 1.4751988649368286} 02/25/2022 04:24:59 - INFO - codeparrot_training - Step 20723: {'lr': 0.00033463829305244874, 'samples': 10610688, 'steps': 20723, 'loss/train': 3.0920238494873047} 02/25/2022 04:25:03 - INFO - codeparrot_training - Step 20724: {'lr': 0.0003346228966576089, 'samples': 10611200, 'steps': 20724, 'loss/train': 0.6148373484611511} 02/25/2022 04:25:08 - INFO - codeparrot_training - Step 20725: {'lr': 0.0003346074999002715, 'samples': 10611712, 'steps': 20725, 'loss/train': 2.074270248413086} 02/25/2022 04:25:12 - INFO - codeparrot_training - Step 20726: {'lr': 0.0003345921027805024, 'samples': 10612224, 'steps': 20726, 'loss/train': 1.8014246225357056} 02/25/2022 04:25:19 - INFO - codeparrot_training - Step 20727: {'lr': 0.00033457670529836756, 'samples': 10612736, 'steps': 20727, 'loss/train': 1.2105252742767334} 02/25/2022 04:25:22 - INFO - codeparrot_training - Step 20728: {'lr': 0.0003345613074539331, 'samples': 10613248, 'steps': 20728, 'loss/train': 2.1932897567749023} 02/25/2022 04:25:28 - INFO - codeparrot_training - Step 20729: {'lr': 0.00033454590924726467, 'samples': 10613760, 'steps': 20729, 'loss/train': 1.5145426988601685} 02/25/2022 04:25:31 - INFO - codeparrot_training - Step 20730: {'lr': 0.0003345305106784286, 'samples': 10614272, 'steps': 20730, 'loss/train': 2.0614726543426514} 02/25/2022 04:25:37 - INFO - codeparrot_training - Step 20731: {'lr': 0.00033451511174749057, 'samples': 10614784, 'steps': 20731, 'loss/train': 1.3937602043151855} 02/25/2022 04:25:40 - INFO - codeparrot_training - Step 20732: {'lr': 0.0003344997124545166, 'samples': 10615296, 'steps': 20732, 'loss/train': 1.9721789360046387} 02/25/2022 04:25:46 - INFO - codeparrot_training - Step 20733: {'lr': 0.0003344843127995728, 'samples': 10615808, 'steps': 20733, 'loss/train': 2.206768035888672} 02/25/2022 04:25:50 - INFO - codeparrot_training - Step 20734: {'lr': 0.00033446891278272493, 'samples': 10616320, 'steps': 20734, 'loss/train': 3.2114272117614746} 02/25/2022 04:25:55 - INFO - codeparrot_training - Step 20735: {'lr': 0.0003344535124040391, 'samples': 10616832, 'steps': 20735, 'loss/train': 1.652105689048767} 02/25/2022 04:25:59 - INFO - codeparrot_training - Step 20736: {'lr': 0.0003344381116635812, 'samples': 10617344, 'steps': 20736, 'loss/train': 2.0425808429718018} 02/25/2022 04:26:06 - INFO - codeparrot_training - Step 20737: {'lr': 0.0003344227105614173, 'samples': 10617856, 'steps': 20737, 'loss/train': 1.4232568740844727} 02/25/2022 04:26:09 - INFO - codeparrot_training - Step 20738: {'lr': 0.0003344073090976132, 'samples': 10618368, 'steps': 20738, 'loss/train': 1.394872784614563} 02/25/2022 04:26:15 - INFO - codeparrot_training - Step 20739: {'lr': 0.00033439190727223517, 'samples': 10618880, 'steps': 20739, 'loss/train': 2.1726560592651367} 02/25/2022 04:26:18 - INFO - codeparrot_training - Step 20740: {'lr': 0.00033437650508534887, 'samples': 10619392, 'steps': 20740, 'loss/train': 1.9124764204025269} 02/25/2022 04:26:25 - INFO - codeparrot_training - Step 20741: {'lr': 0.0003343611025370205, 'samples': 10619904, 'steps': 20741, 'loss/train': 0.954076886177063} 02/25/2022 04:26:28 - INFO - codeparrot_training - Step 20742: {'lr': 0.00033434569962731593, 'samples': 10620416, 'steps': 20742, 'loss/train': 2.2607502937316895} 02/25/2022 04:26:31 - INFO - codeparrot_training - Step 20743: {'lr': 0.0003343302963563012, 'samples': 10620928, 'steps': 20743, 'loss/train': 2.2857413291931152} 02/25/2022 04:26:37 - INFO - codeparrot_training - Step 20744: {'lr': 0.00033431489272404215, 'samples': 10621440, 'steps': 20744, 'loss/train': 1.923913836479187} 02/25/2022 04:26:41 - INFO - codeparrot_training - Step 20745: {'lr': 0.00033429948873060496, 'samples': 10621952, 'steps': 20745, 'loss/train': 1.737326741218567} 02/25/2022 04:26:46 - INFO - codeparrot_training - Step 20746: {'lr': 0.0003342840843760555, 'samples': 10622464, 'steps': 20746, 'loss/train': 2.188119411468506} 02/25/2022 04:26:50 - INFO - codeparrot_training - Step 20747: {'lr': 0.00033426867966045984, 'samples': 10622976, 'steps': 20747, 'loss/train': 1.7467223405838013} 02/25/2022 04:26:57 - INFO - codeparrot_training - Step 20748: {'lr': 0.00033425327458388375, 'samples': 10623488, 'steps': 20748, 'loss/train': 2.5961050987243652} 02/25/2022 04:27:00 - INFO - codeparrot_training - Step 20749: {'lr': 0.0003342378691463936, 'samples': 10624000, 'steps': 20749, 'loss/train': 2.309401512145996} 02/25/2022 04:27:06 - INFO - codeparrot_training - Step 20750: {'lr': 0.00033422246334805503, 'samples': 10624512, 'steps': 20750, 'loss/train': 2.1629879474639893} 02/25/2022 04:27:09 - INFO - codeparrot_training - Step 20751: {'lr': 0.0003342070571889342, 'samples': 10625024, 'steps': 20751, 'loss/train': 1.0394816398620605} 02/25/2022 04:27:15 - INFO - codeparrot_training - Step 20752: {'lr': 0.00033419165066909707, 'samples': 10625536, 'steps': 20752, 'loss/train': 2.7774131298065186} 02/25/2022 04:27:18 - INFO - codeparrot_training - Step 20753: {'lr': 0.0003341762437886097, 'samples': 10626048, 'steps': 20753, 'loss/train': 3.2035017013549805} 02/25/2022 04:27:24 - INFO - codeparrot_training - Step 20754: {'lr': 0.0003341608365475379, 'samples': 10626560, 'steps': 20754, 'loss/train': 0.35955071449279785} 02/25/2022 04:27:27 - INFO - codeparrot_training - Step 20755: {'lr': 0.00033414542894594793, 'samples': 10627072, 'steps': 20755, 'loss/train': 1.6717816591262817} 02/25/2022 04:27:33 - INFO - codeparrot_training - Step 20756: {'lr': 0.00033413002098390567, 'samples': 10627584, 'steps': 20756, 'loss/train': 2.539670467376709} 02/25/2022 04:27:36 - INFO - codeparrot_training - Step 20757: {'lr': 0.00033411461266147705, 'samples': 10628096, 'steps': 20757, 'loss/train': 1.463529348373413} 02/25/2022 04:27:42 - INFO - codeparrot_training - Step 20758: {'lr': 0.00033409920397872814, 'samples': 10628608, 'steps': 20758, 'loss/train': 1.6585992574691772} 02/25/2022 04:27:45 - INFO - codeparrot_training - Step 20759: {'lr': 0.00033408379493572493, 'samples': 10629120, 'steps': 20759, 'loss/train': 2.5601906776428223} 02/25/2022 04:27:51 - INFO - codeparrot_training - Step 20760: {'lr': 0.0003340683855325335, 'samples': 10629632, 'steps': 20760, 'loss/train': 1.451882243156433} 02/25/2022 04:27:54 - INFO - codeparrot_training - Step 20761: {'lr': 0.00033405297576921976, 'samples': 10630144, 'steps': 20761, 'loss/train': 2.0742735862731934} 02/25/2022 04:28:00 - INFO - codeparrot_training - Step 20762: {'lr': 0.00033403756564584974, 'samples': 10630656, 'steps': 20762, 'loss/train': 1.939662218093872} 02/25/2022 04:28:03 - INFO - codeparrot_training - Step 20763: {'lr': 0.0003340221551624896, 'samples': 10631168, 'steps': 20763, 'loss/train': 1.9721509218215942} 02/25/2022 04:28:11 - INFO - codeparrot_training - Step 20764: {'lr': 0.0003340067443192051, 'samples': 10631680, 'steps': 20764, 'loss/train': 1.8509892225265503} 02/25/2022 04:28:14 - INFO - codeparrot_training - Step 20765: {'lr': 0.0003339913331160624, 'samples': 10632192, 'steps': 20765, 'loss/train': 1.9136699438095093} 02/25/2022 04:28:20 - INFO - codeparrot_training - Step 20766: {'lr': 0.0003339759215531275, 'samples': 10632704, 'steps': 20766, 'loss/train': 2.413472890853882} 02/25/2022 04:28:23 - INFO - codeparrot_training - Step 20767: {'lr': 0.0003339605096304664, 'samples': 10633216, 'steps': 20767, 'loss/train': 2.148566246032715} 02/25/2022 04:28:29 - INFO - codeparrot_training - Step 20768: {'lr': 0.00033394509734814516, 'samples': 10633728, 'steps': 20768, 'loss/train': 2.2628605365753174} 02/25/2022 04:28:33 - INFO - codeparrot_training - Step 20769: {'lr': 0.00033392968470622987, 'samples': 10634240, 'steps': 20769, 'loss/train': 1.1457905769348145} 02/25/2022 04:28:38 - INFO - codeparrot_training - Step 20770: {'lr': 0.0003339142717047863, 'samples': 10634752, 'steps': 20770, 'loss/train': 2.2070224285125732} 02/25/2022 04:28:42 - INFO - codeparrot_training - Step 20771: {'lr': 0.0003338988583438808, 'samples': 10635264, 'steps': 20771, 'loss/train': 2.3740336894989014} 02/25/2022 04:28:47 - INFO - codeparrot_training - Step 20772: {'lr': 0.0003338834446235791, 'samples': 10635776, 'steps': 20772, 'loss/train': 1.4844857454299927} 02/25/2022 04:28:51 - INFO - codeparrot_training - Step 20773: {'lr': 0.00033386803054394744, 'samples': 10636288, 'steps': 20773, 'loss/train': 1.525451898574829} 02/25/2022 04:28:58 - INFO - codeparrot_training - Step 20774: {'lr': 0.0003338526161050517, 'samples': 10636800, 'steps': 20774, 'loss/train': 2.371129274368286} 02/25/2022 04:29:01 - INFO - codeparrot_training - Step 20775: {'lr': 0.00033383720130695794, 'samples': 10637312, 'steps': 20775, 'loss/train': 1.7599045038223267} 02/25/2022 04:29:07 - INFO - codeparrot_training - Step 20776: {'lr': 0.0003338217861497324, 'samples': 10637824, 'steps': 20776, 'loss/train': 2.037294387817383} 02/25/2022 04:29:10 - INFO - codeparrot_training - Step 20777: {'lr': 0.0003338063706334408, 'samples': 10638336, 'steps': 20777, 'loss/train': 1.5171146392822266} 02/25/2022 04:29:16 - INFO - codeparrot_training - Step 20778: {'lr': 0.00033379095475814937, 'samples': 10638848, 'steps': 20778, 'loss/train': 2.194605827331543} 02/25/2022 04:29:19 - INFO - codeparrot_training - Step 20779: {'lr': 0.00033377553852392404, 'samples': 10639360, 'steps': 20779, 'loss/train': 1.1043100357055664} 02/25/2022 04:29:25 - INFO - codeparrot_training - Step 20780: {'lr': 0.000333760121930831, 'samples': 10639872, 'steps': 20780, 'loss/train': 2.4773428440093994} 02/25/2022 04:29:28 - INFO - codeparrot_training - Step 20781: {'lr': 0.00033374470497893614, 'samples': 10640384, 'steps': 20781, 'loss/train': 2.133458137512207} 02/25/2022 04:29:34 - INFO - codeparrot_training - Step 20782: {'lr': 0.0003337292876683056, 'samples': 10640896, 'steps': 20782, 'loss/train': 3.3295552730560303} 02/25/2022 04:29:37 - INFO - codeparrot_training - Step 20783: {'lr': 0.0003337138699990053, 'samples': 10641408, 'steps': 20783, 'loss/train': 2.7440850734710693} 02/25/2022 04:29:45 - INFO - codeparrot_training - Step 20784: {'lr': 0.00033369845197110144, 'samples': 10641920, 'steps': 20784, 'loss/train': 1.9222973585128784} 02/25/2022 04:29:48 - INFO - codeparrot_training - Step 20785: {'lr': 0.00033368303358465994, 'samples': 10642432, 'steps': 20785, 'loss/train': 1.8609380722045898} 02/25/2022 04:29:54 - INFO - codeparrot_training - Step 20786: {'lr': 0.00033366761483974693, 'samples': 10642944, 'steps': 20786, 'loss/train': 2.008422613143921} 02/25/2022 04:29:57 - INFO - codeparrot_training - Step 20787: {'lr': 0.0003336521957364284, 'samples': 10643456, 'steps': 20787, 'loss/train': 2.1725199222564697} 02/25/2022 04:30:03 - INFO - codeparrot_training - Step 20788: {'lr': 0.0003336367762747704, 'samples': 10643968, 'steps': 20788, 'loss/train': 1.5638988018035889} 02/25/2022 04:30:06 - INFO - codeparrot_training - Step 20789: {'lr': 0.0003336213564548391, 'samples': 10644480, 'steps': 20789, 'loss/train': 1.1934763193130493} 02/25/2022 04:30:12 - INFO - codeparrot_training - Step 20790: {'lr': 0.0003336059362767004, 'samples': 10644992, 'steps': 20790, 'loss/train': 2.389613389968872} 02/25/2022 04:30:15 - INFO - codeparrot_training - Step 20791: {'lr': 0.0003335905157404204, 'samples': 10645504, 'steps': 20791, 'loss/train': 1.5832748413085938} 02/25/2022 04:30:21 - INFO - codeparrot_training - Step 20792: {'lr': 0.0003335750948460652, 'samples': 10646016, 'steps': 20792, 'loss/train': 2.5373685359954834} 02/25/2022 04:30:28 - INFO - codeparrot_training - Step 20793: {'lr': 0.0003335596735937009, 'samples': 10646528, 'steps': 20793, 'loss/train': 2.4140779972076416} 02/25/2022 04:30:31 - INFO - codeparrot_training - Step 20794: {'lr': 0.0003335442519833933, 'samples': 10647040, 'steps': 20794, 'loss/train': 2.4741437435150146} 02/25/2022 04:30:37 - INFO - codeparrot_training - Step 20795: {'lr': 0.00033352883001520884, 'samples': 10647552, 'steps': 20795, 'loss/train': 1.6465550661087036} 02/25/2022 04:30:40 - INFO - codeparrot_training - Step 20796: {'lr': 0.0003335134076892133, 'samples': 10648064, 'steps': 20796, 'loss/train': 2.7344467639923096} 02/25/2022 04:30:46 - INFO - codeparrot_training - Step 20797: {'lr': 0.0003334979850054729, 'samples': 10648576, 'steps': 20797, 'loss/train': 1.6852879524230957} 02/25/2022 04:30:50 - INFO - codeparrot_training - Step 20798: {'lr': 0.0003334825619640536, 'samples': 10649088, 'steps': 20798, 'loss/train': 1.8291069269180298} 02/25/2022 04:30:53 - INFO - codeparrot_training - Step 20799: {'lr': 0.0003334671385650215, 'samples': 10649600, 'steps': 20799, 'loss/train': 1.322557806968689} 02/25/2022 04:31:00 - INFO - codeparrot_training - Step 20800: {'lr': 0.0003334517148084427, 'samples': 10650112, 'steps': 20800, 'loss/train': 2.409087657928467} 02/25/2022 04:31:04 - INFO - codeparrot_training - Step 20801: {'lr': 0.00033343629069438333, 'samples': 10650624, 'steps': 20801, 'loss/train': 3.682776927947998} 02/25/2022 04:31:09 - INFO - codeparrot_training - Step 20802: {'lr': 0.0003334208662229093, 'samples': 10651136, 'steps': 20802, 'loss/train': 1.8637281656265259} 02/25/2022 04:31:13 - INFO - codeparrot_training - Step 20803: {'lr': 0.0003334054413940868, 'samples': 10651648, 'steps': 20803, 'loss/train': 1.6451548337936401} 02/25/2022 04:31:18 - INFO - codeparrot_training - Step 20804: {'lr': 0.0003333900162079818, 'samples': 10652160, 'steps': 20804, 'loss/train': 2.9036614894866943} 02/25/2022 04:31:22 - INFO - codeparrot_training - Step 20805: {'lr': 0.00033337459066466057, 'samples': 10652672, 'steps': 20805, 'loss/train': 2.6945085525512695} 02/25/2022 04:31:27 - INFO - codeparrot_training - Step 20806: {'lr': 0.000333359164764189, 'samples': 10653184, 'steps': 20806, 'loss/train': 2.3912503719329834} 02/25/2022 04:31:31 - INFO - codeparrot_training - Step 20807: {'lr': 0.00033334373850663323, 'samples': 10653696, 'steps': 20807, 'loss/train': 1.7559146881103516} 02/25/2022 04:31:36 - INFO - codeparrot_training - Step 20808: {'lr': 0.00033332831189205936, 'samples': 10654208, 'steps': 20808, 'loss/train': 2.333390951156616} 02/25/2022 04:31:40 - INFO - codeparrot_training - Step 20809: {'lr': 0.00033331288492053344, 'samples': 10654720, 'steps': 20809, 'loss/train': 1.834153175354004} 02/25/2022 04:31:47 - INFO - codeparrot_training - Step 20810: {'lr': 0.0003332974575921217, 'samples': 10655232, 'steps': 20810, 'loss/train': 1.6094986200332642} 02/25/2022 04:31:50 - INFO - codeparrot_training - Step 20811: {'lr': 0.00033328202990688996, 'samples': 10655744, 'steps': 20811, 'loss/train': 2.034541130065918} 02/25/2022 04:31:56 - INFO - codeparrot_training - Step 20812: {'lr': 0.0003332666018649044, 'samples': 10656256, 'steps': 20812, 'loss/train': 0.3137263357639313} 02/25/2022 04:32:00 - INFO - codeparrot_training - Step 20813: {'lr': 0.00033325117346623135, 'samples': 10656768, 'steps': 20813, 'loss/train': 1.6814106702804565} 02/25/2022 04:32:05 - INFO - codeparrot_training - Step 20814: {'lr': 0.00033323574471093656, 'samples': 10657280, 'steps': 20814, 'loss/train': 2.4355385303497314} 02/25/2022 04:32:09 - INFO - codeparrot_training - Step 20815: {'lr': 0.0003332203155990863, 'samples': 10657792, 'steps': 20815, 'loss/train': 1.4117026329040527} 02/25/2022 04:32:14 - INFO - codeparrot_training - Step 20816: {'lr': 0.00033320488613074666, 'samples': 10658304, 'steps': 20816, 'loss/train': 2.3932127952575684} 02/25/2022 04:32:18 - INFO - codeparrot_training - Step 20817: {'lr': 0.00033318945630598373, 'samples': 10658816, 'steps': 20817, 'loss/train': 1.157488226890564} 02/25/2022 04:32:23 - INFO - codeparrot_training - Step 20818: {'lr': 0.00033317402612486355, 'samples': 10659328, 'steps': 20818, 'loss/train': 2.4078705310821533} 02/25/2022 04:32:27 - INFO - codeparrot_training - Step 20819: {'lr': 0.00033315859558745225, 'samples': 10659840, 'steps': 20819, 'loss/train': 1.5141855478286743} 02/25/2022 04:32:34 - INFO - codeparrot_training - Step 20820: {'lr': 0.0003331431646938159, 'samples': 10660352, 'steps': 20820, 'loss/train': 1.9648964405059814} 02/25/2022 04:32:37 - INFO - codeparrot_training - Step 20821: {'lr': 0.00033312773344402075, 'samples': 10660864, 'steps': 20821, 'loss/train': 2.2263917922973633} 02/25/2022 04:32:43 - INFO - codeparrot_training - Step 20822: {'lr': 0.00033311230183813266, 'samples': 10661376, 'steps': 20822, 'loss/train': 2.011302947998047} 02/25/2022 04:32:46 - INFO - codeparrot_training - Step 20823: {'lr': 0.0003330968698762179, 'samples': 10661888, 'steps': 20823, 'loss/train': 1.5979396104812622} 02/25/2022 04:32:52 - INFO - codeparrot_training - Step 20824: {'lr': 0.0003330814375583426, 'samples': 10662400, 'steps': 20824, 'loss/train': 2.817964792251587} 02/25/2022 04:32:55 - INFO - codeparrot_training - Step 20825: {'lr': 0.00033306600488457264, 'samples': 10662912, 'steps': 20825, 'loss/train': 1.537620186805725} 02/25/2022 04:33:01 - INFO - codeparrot_training - Step 20826: {'lr': 0.00033305057185497444, 'samples': 10663424, 'steps': 20826, 'loss/train': 1.2470121383666992} 02/25/2022 04:33:05 - INFO - codeparrot_training - Step 20827: {'lr': 0.0003330351384696139, 'samples': 10663936, 'steps': 20827, 'loss/train': 2.086422920227051} 02/25/2022 04:33:10 - INFO - codeparrot_training - Step 20828: {'lr': 0.00033301970472855724, 'samples': 10664448, 'steps': 20828, 'loss/train': 1.9239081144332886} 02/25/2022 04:33:13 - INFO - codeparrot_training - Step 20829: {'lr': 0.0003330042706318705, 'samples': 10664960, 'steps': 20829, 'loss/train': 2.0555431842803955} 02/25/2022 04:33:21 - INFO - codeparrot_training - Step 20830: {'lr': 0.00033298883617961984, 'samples': 10665472, 'steps': 20830, 'loss/train': 1.4617338180541992} 02/25/2022 04:33:24 - INFO - codeparrot_training - Step 20831: {'lr': 0.0003329734013718713, 'samples': 10665984, 'steps': 20831, 'loss/train': 0.9457871913909912} 02/25/2022 04:33:30 - INFO - codeparrot_training - Step 20832: {'lr': 0.0003329579662086911, 'samples': 10666496, 'steps': 20832, 'loss/train': 1.6988276243209839} 02/25/2022 04:33:33 - INFO - codeparrot_training - Step 20833: {'lr': 0.00033294253069014534, 'samples': 10667008, 'steps': 20833, 'loss/train': 2.575636625289917} 02/25/2022 04:33:39 - INFO - codeparrot_training - Step 20834: {'lr': 0.0003329270948163001, 'samples': 10667520, 'steps': 20834, 'loss/train': 2.3679113388061523} 02/25/2022 04:33:42 - INFO - codeparrot_training - Step 20835: {'lr': 0.0003329116585872215, 'samples': 10668032, 'steps': 20835, 'loss/train': 0.6654316186904907} 02/25/2022 04:33:48 - INFO - codeparrot_training - Step 20836: {'lr': 0.00033289622200297563, 'samples': 10668544, 'steps': 20836, 'loss/train': 3.990220069885254} 02/25/2022 04:33:51 - INFO - codeparrot_training - Step 20837: {'lr': 0.0003328807850636287, 'samples': 10669056, 'steps': 20837, 'loss/train': 1.8589555025100708} 02/25/2022 04:33:58 - INFO - codeparrot_training - Step 20838: {'lr': 0.0003328653477692469, 'samples': 10669568, 'steps': 20838, 'loss/train': 1.8131022453308105} 02/25/2022 04:34:01 - INFO - codeparrot_training - Step 20839: {'lr': 0.0003328499101198962, 'samples': 10670080, 'steps': 20839, 'loss/train': 2.7114906311035156} 02/25/2022 04:34:05 - INFO - codeparrot_training - Step 20840: {'lr': 0.0003328344721156427, 'samples': 10670592, 'steps': 20840, 'loss/train': 2.376781702041626} 02/25/2022 04:34:10 - INFO - codeparrot_training - Step 20841: {'lr': 0.00033281903375655277, 'samples': 10671104, 'steps': 20841, 'loss/train': 1.2486603260040283} 02/25/2022 04:34:13 - INFO - codeparrot_training - Step 20842: {'lr': 0.0003328035950426923, 'samples': 10671616, 'steps': 20842, 'loss/train': 2.7645821571350098} 02/25/2022 04:34:19 - INFO - codeparrot_training - Step 20843: {'lr': 0.0003327881559741276, 'samples': 10672128, 'steps': 20843, 'loss/train': 1.2805066108703613} 02/25/2022 04:34:23 - INFO - codeparrot_training - Step 20844: {'lr': 0.00033277271655092467, 'samples': 10672640, 'steps': 20844, 'loss/train': 2.83009672164917} 02/25/2022 04:34:28 - INFO - codeparrot_training - Step 20845: {'lr': 0.0003327572767731497, 'samples': 10673152, 'steps': 20845, 'loss/train': 2.5204989910125732} 02/25/2022 04:34:31 - INFO - codeparrot_training - Step 20846: {'lr': 0.0003327418366408689, 'samples': 10673664, 'steps': 20846, 'loss/train': 2.482987642288208} 02/25/2022 04:34:39 - INFO - codeparrot_training - Step 20847: {'lr': 0.0003327263961541483, 'samples': 10674176, 'steps': 20847, 'loss/train': 1.3386704921722412} 02/25/2022 04:34:42 - INFO - codeparrot_training - Step 20848: {'lr': 0.0003327109553130541, 'samples': 10674688, 'steps': 20848, 'loss/train': 2.4186148643493652} 02/25/2022 04:34:48 - INFO - codeparrot_training - Step 20849: {'lr': 0.0003326955141176524, 'samples': 10675200, 'steps': 20849, 'loss/train': 2.323169469833374} 02/25/2022 04:34:51 - INFO - codeparrot_training - Step 20850: {'lr': 0.0003326800725680094, 'samples': 10675712, 'steps': 20850, 'loss/train': 2.3529465198516846} 02/25/2022 04:34:57 - INFO - codeparrot_training - Step 20851: {'lr': 0.0003326646306641912, 'samples': 10676224, 'steps': 20851, 'loss/train': 1.5705105066299438} 02/25/2022 04:35:00 - INFO - codeparrot_training - Step 20852: {'lr': 0.000332649188406264, 'samples': 10676736, 'steps': 20852, 'loss/train': 0.962972104549408} 02/25/2022 04:35:06 - INFO - codeparrot_training - Step 20853: {'lr': 0.0003326337457942939, 'samples': 10677248, 'steps': 20853, 'loss/train': 2.0908255577087402} 02/25/2022 04:35:09 - INFO - codeparrot_training - Step 20854: {'lr': 0.00033261830282834716, 'samples': 10677760, 'steps': 20854, 'loss/train': 1.5958484411239624} 02/25/2022 04:35:15 - INFO - codeparrot_training - Step 20855: {'lr': 0.00033260285950848965, 'samples': 10678272, 'steps': 20855, 'loss/train': 1.704235553741455} 02/25/2022 04:35:18 - INFO - codeparrot_training - Step 20856: {'lr': 0.0003325874158347879, 'samples': 10678784, 'steps': 20856, 'loss/train': 1.8372846841812134} 02/25/2022 04:35:25 - INFO - codeparrot_training - Step 20857: {'lr': 0.0003325719718073078, 'samples': 10679296, 'steps': 20857, 'loss/train': 2.416452169418335} 02/25/2022 04:35:29 - INFO - codeparrot_training - Step 20858: {'lr': 0.00033255652742611566, 'samples': 10679808, 'steps': 20858, 'loss/train': 1.8830294609069824} 02/25/2022 04:35:34 - INFO - codeparrot_training - Step 20859: {'lr': 0.0003325410826912775, 'samples': 10680320, 'steps': 20859, 'loss/train': 3.635756015777588} 02/25/2022 04:35:38 - INFO - codeparrot_training - Step 20860: {'lr': 0.0003325256376028595, 'samples': 10680832, 'steps': 20860, 'loss/train': 1.945672869682312} 02/25/2022 04:35:43 - INFO - codeparrot_training - Step 20861: {'lr': 0.000332510192160928, 'samples': 10681344, 'steps': 20861, 'loss/train': 1.0041775703430176} 02/25/2022 04:35:47 - INFO - codeparrot_training - Step 20862: {'lr': 0.000332494746365549, 'samples': 10681856, 'steps': 20862, 'loss/train': 2.1228537559509277} 02/25/2022 04:35:52 - INFO - codeparrot_training - Step 20863: {'lr': 0.00033247930021678866, 'samples': 10682368, 'steps': 20863, 'loss/train': 2.3177340030670166} 02/25/2022 04:35:56 - INFO - codeparrot_training - Step 20864: {'lr': 0.0003324638537147132, 'samples': 10682880, 'steps': 20864, 'loss/train': 1.7031713724136353} 02/25/2022 04:36:01 - INFO - codeparrot_training - Step 20865: {'lr': 0.00033244840685938884, 'samples': 10683392, 'steps': 20865, 'loss/train': 2.3604841232299805} 02/25/2022 04:36:05 - INFO - codeparrot_training - Step 20866: {'lr': 0.0003324329596508816, 'samples': 10683904, 'steps': 20866, 'loss/train': 1.0252118110656738} 02/25/2022 04:36:12 - INFO - codeparrot_training - Step 20867: {'lr': 0.0003324175120892579, 'samples': 10684416, 'steps': 20867, 'loss/train': 1.966038465499878} 02/25/2022 04:36:15 - INFO - codeparrot_training - Step 20868: {'lr': 0.00033240206417458354, 'samples': 10684928, 'steps': 20868, 'loss/train': 1.7910404205322266} 02/25/2022 04:36:21 - INFO - codeparrot_training - Step 20869: {'lr': 0.00033238661590692496, 'samples': 10685440, 'steps': 20869, 'loss/train': 2.265265703201294} 02/25/2022 04:36:24 - INFO - codeparrot_training - Step 20870: {'lr': 0.00033237116728634833, 'samples': 10685952, 'steps': 20870, 'loss/train': 1.8578381538391113} 02/25/2022 04:36:30 - INFO - codeparrot_training - Step 20871: {'lr': 0.0003323557183129197, 'samples': 10686464, 'steps': 20871, 'loss/train': 2.419618606567383} 02/25/2022 04:36:33 - INFO - codeparrot_training - Step 20872: {'lr': 0.0003323402689867054, 'samples': 10686976, 'steps': 20872, 'loss/train': 1.4632823467254639} 02/25/2022 04:36:39 - INFO - codeparrot_training - Step 20873: {'lr': 0.0003323248193077715, 'samples': 10687488, 'steps': 20873, 'loss/train': 1.7312337160110474} 02/25/2022 04:36:42 - INFO - codeparrot_training - Step 20874: {'lr': 0.0003323093692761842, 'samples': 10688000, 'steps': 20874, 'loss/train': 2.6617040634155273} 02/25/2022 04:36:48 - INFO - codeparrot_training - Step 20875: {'lr': 0.00033229391889200974, 'samples': 10688512, 'steps': 20875, 'loss/train': 2.7804622650146484} 02/25/2022 04:36:51 - INFO - codeparrot_training - Step 20876: {'lr': 0.00033227846815531424, 'samples': 10689024, 'steps': 20876, 'loss/train': 2.038386583328247} 02/25/2022 04:36:57 - INFO - codeparrot_training - Step 20877: {'lr': 0.0003322630170661639, 'samples': 10689536, 'steps': 20877, 'loss/train': 1.1443202495574951} 02/25/2022 04:37:00 - INFO - codeparrot_training - Step 20878: {'lr': 0.0003322475656246249, 'samples': 10690048, 'steps': 20878, 'loss/train': 2.632556438446045} 02/25/2022 04:37:08 - INFO - codeparrot_training - Step 20879: {'lr': 0.0003322321138307635, 'samples': 10690560, 'steps': 20879, 'loss/train': 1.0574843883514404} 02/25/2022 04:37:11 - INFO - codeparrot_training - Step 20880: {'lr': 0.0003322166616846458, 'samples': 10691072, 'steps': 20880, 'loss/train': 1.4141308069229126} 02/25/2022 04:37:17 - INFO - codeparrot_training - Step 20881: {'lr': 0.0003322012091863381, 'samples': 10691584, 'steps': 20881, 'loss/train': 1.7375237941741943} 02/25/2022 04:37:20 - INFO - codeparrot_training - Step 20882: {'lr': 0.0003321857563359064, 'samples': 10692096, 'steps': 20882, 'loss/train': 2.5988423824310303} 02/25/2022 04:37:26 - INFO - codeparrot_training - Step 20883: {'lr': 0.00033217030313341704, 'samples': 10692608, 'steps': 20883, 'loss/train': 2.4692912101745605} 02/25/2022 04:37:29 - INFO - codeparrot_training - Step 20884: {'lr': 0.00033215484957893626, 'samples': 10693120, 'steps': 20884, 'loss/train': 1.7564690113067627} 02/25/2022 04:37:35 - INFO - codeparrot_training - Step 20885: {'lr': 0.0003321393956725302, 'samples': 10693632, 'steps': 20885, 'loss/train': 1.5622729063034058} 02/25/2022 04:37:38 - INFO - codeparrot_training - Step 20886: {'lr': 0.00033212394141426493, 'samples': 10694144, 'steps': 20886, 'loss/train': 1.2269775867462158} 02/25/2022 04:37:44 - INFO - codeparrot_training - Step 20887: {'lr': 0.00033210848680420693, 'samples': 10694656, 'steps': 20887, 'loss/train': 2.2369186878204346} 02/25/2022 04:37:47 - INFO - codeparrot_training - Step 20888: {'lr': 0.00033209303184242214, 'samples': 10695168, 'steps': 20888, 'loss/train': 2.591826915740967} 02/25/2022 04:37:55 - INFO - codeparrot_training - Step 20889: {'lr': 0.0003320775765289769, 'samples': 10695680, 'steps': 20889, 'loss/train': 2.1632304191589355} 02/25/2022 04:37:58 - INFO - codeparrot_training - Step 20890: {'lr': 0.0003320621208639374, 'samples': 10696192, 'steps': 20890, 'loss/train': 0.8858595490455627} 02/25/2022 04:38:04 - INFO - codeparrot_training - Step 20891: {'lr': 0.00033204666484736977, 'samples': 10696704, 'steps': 20891, 'loss/train': 1.6609606742858887} 02/25/2022 04:38:07 - INFO - codeparrot_training - Step 20892: {'lr': 0.0003320312084793404, 'samples': 10697216, 'steps': 20892, 'loss/train': 2.259467840194702} 02/25/2022 04:38:13 - INFO - codeparrot_training - Step 20893: {'lr': 0.0003320157517599153, 'samples': 10697728, 'steps': 20893, 'loss/train': 2.0813517570495605} 02/25/2022 04:38:16 - INFO - codeparrot_training - Step 20894: {'lr': 0.00033200029468916076, 'samples': 10698240, 'steps': 20894, 'loss/train': 1.8910647630691528} 02/25/2022 04:38:22 - INFO - codeparrot_training - Step 20895: {'lr': 0.00033198483726714294, 'samples': 10698752, 'steps': 20895, 'loss/train': 1.3603250980377197} 02/25/2022 04:38:25 - INFO - codeparrot_training - Step 20896: {'lr': 0.00033196937949392824, 'samples': 10699264, 'steps': 20896, 'loss/train': 2.342127561569214} 02/25/2022 04:38:31 - INFO - codeparrot_training - Step 20897: {'lr': 0.00033195392136958264, 'samples': 10699776, 'steps': 20897, 'loss/train': 2.5612542629241943} 02/25/2022 04:38:34 - INFO - codeparrot_training - Step 20898: {'lr': 0.00033193846289417253, 'samples': 10700288, 'steps': 20898, 'loss/train': 1.9002939462661743} 02/25/2022 04:38:40 - INFO - codeparrot_training - Step 20899: {'lr': 0.00033192300406776406, 'samples': 10700800, 'steps': 20899, 'loss/train': 2.211519718170166} 02/25/2022 04:38:43 - INFO - codeparrot_training - Step 20900: {'lr': 0.0003319075448904234, 'samples': 10701312, 'steps': 20900, 'loss/train': 2.207306146621704} 02/25/2022 04:38:49 - INFO - codeparrot_training - Step 20901: {'lr': 0.00033189208536221683, 'samples': 10701824, 'steps': 20901, 'loss/train': 2.002225160598755} 02/25/2022 04:38:52 - INFO - codeparrot_training - Step 20902: {'lr': 0.00033187662548321063, 'samples': 10702336, 'steps': 20902, 'loss/train': 1.6608514785766602} 02/25/2022 04:38:59 - INFO - codeparrot_training - Step 20903: {'lr': 0.00033186116525347093, 'samples': 10702848, 'steps': 20903, 'loss/train': 2.467921495437622} 02/25/2022 04:39:03 - INFO - codeparrot_training - Step 20904: {'lr': 0.00033184570467306403, 'samples': 10703360, 'steps': 20904, 'loss/train': 1.6327680349349976} 02/25/2022 04:39:08 - INFO - codeparrot_training - Step 20905: {'lr': 0.0003318302437420561, 'samples': 10703872, 'steps': 20905, 'loss/train': 1.6022164821624756} 02/25/2022 04:39:12 - INFO - codeparrot_training - Step 20906: {'lr': 0.0003318147824605133, 'samples': 10704384, 'steps': 20906, 'loss/train': 1.7568657398223877} 02/25/2022 04:39:18 - INFO - codeparrot_training - Step 20907: {'lr': 0.0003317993208285021, 'samples': 10704896, 'steps': 20907, 'loss/train': 2.583235502243042} 02/25/2022 04:39:21 - INFO - codeparrot_training - Step 20908: {'lr': 0.0003317838588460884, 'samples': 10705408, 'steps': 20908, 'loss/train': 2.3424243927001953} 02/25/2022 04:39:27 - INFO - codeparrot_training - Step 20909: {'lr': 0.0003317683965133388, 'samples': 10705920, 'steps': 20909, 'loss/train': 2.498840093612671} 02/25/2022 04:39:30 - INFO - codeparrot_training - Step 20910: {'lr': 0.0003317529338303192, 'samples': 10706432, 'steps': 20910, 'loss/train': 2.320816993713379} 02/25/2022 04:39:36 - INFO - codeparrot_training - Step 20911: {'lr': 0.00033173747079709616, 'samples': 10706944, 'steps': 20911, 'loss/train': 2.204019069671631} 02/25/2022 04:39:39 - INFO - codeparrot_training - Step 20912: {'lr': 0.0003317220074137356, 'samples': 10707456, 'steps': 20912, 'loss/train': 2.279435634613037} 02/25/2022 04:39:46 - INFO - codeparrot_training - Step 20913: {'lr': 0.000331706543680304, 'samples': 10707968, 'steps': 20913, 'loss/train': 1.2946723699569702} 02/25/2022 04:39:50 - INFO - codeparrot_training - Step 20914: {'lr': 0.0003316910795968675, 'samples': 10708480, 'steps': 20914, 'loss/train': 1.8910942077636719} 02/25/2022 04:39:55 - INFO - codeparrot_training - Step 20915: {'lr': 0.00033167561516349233, 'samples': 10708992, 'steps': 20915, 'loss/train': 1.9200794696807861} 02/25/2022 04:39:59 - INFO - codeparrot_training - Step 20916: {'lr': 0.0003316601503802448, 'samples': 10709504, 'steps': 20916, 'loss/train': 1.2175666093826294} 02/25/2022 04:40:04 - INFO - codeparrot_training - Step 20917: {'lr': 0.00033164468524719105, 'samples': 10710016, 'steps': 20917, 'loss/train': 2.0007216930389404} 02/25/2022 04:40:08 - INFO - codeparrot_training - Step 20918: {'lr': 0.00033162921976439744, 'samples': 10710528, 'steps': 20918, 'loss/train': 2.225008726119995} 02/25/2022 04:40:13 - INFO - codeparrot_training - Step 20919: {'lr': 0.00033161375393193015, 'samples': 10711040, 'steps': 20919, 'loss/train': 1.6899045705795288} 02/25/2022 04:40:17 - INFO - codeparrot_training - Step 20920: {'lr': 0.00033159828774985547, 'samples': 10711552, 'steps': 20920, 'loss/train': 1.5959523916244507} 02/25/2022 04:40:23 - INFO - codeparrot_training - Step 20921: {'lr': 0.0003315828212182396, 'samples': 10712064, 'steps': 20921, 'loss/train': 1.3249160051345825} 02/25/2022 04:40:26 - INFO - codeparrot_training - Step 20922: {'lr': 0.00033156735433714893, 'samples': 10712576, 'steps': 20922, 'loss/train': 1.5032868385314941} 02/25/2022 04:40:32 - INFO - codeparrot_training - Step 20923: {'lr': 0.00033155188710664945, 'samples': 10713088, 'steps': 20923, 'loss/train': 1.9754023551940918} 02/25/2022 04:40:35 - INFO - codeparrot_training - Step 20924: {'lr': 0.00033153641952680767, 'samples': 10713600, 'steps': 20924, 'loss/train': 1.177696943283081} 02/25/2022 04:40:43 - INFO - codeparrot_training - Step 20925: {'lr': 0.0003315209515976898, 'samples': 10714112, 'steps': 20925, 'loss/train': 2.143282413482666} 02/25/2022 04:40:46 - INFO - codeparrot_training - Step 20926: {'lr': 0.000331505483319362, 'samples': 10714624, 'steps': 20926, 'loss/train': 3.3589823246002197} 02/25/2022 04:40:52 - INFO - codeparrot_training - Step 20927: {'lr': 0.0003314900146918906, 'samples': 10715136, 'steps': 20927, 'loss/train': 2.3602800369262695} 02/25/2022 04:40:55 - INFO - codeparrot_training - Step 20928: {'lr': 0.0003314745457153419, 'samples': 10715648, 'steps': 20928, 'loss/train': 0.9475632309913635} 02/25/2022 04:41:01 - INFO - codeparrot_training - Step 20929: {'lr': 0.00033145907638978207, 'samples': 10716160, 'steps': 20929, 'loss/train': 0.13922227919101715} 02/25/2022 04:41:04 - INFO - codeparrot_training - Step 20930: {'lr': 0.00033144360671527747, 'samples': 10716672, 'steps': 20930, 'loss/train': 2.3799941539764404} 02/25/2022 04:41:10 - INFO - codeparrot_training - Step 20931: {'lr': 0.0003314281366918943, 'samples': 10717184, 'steps': 20931, 'loss/train': 1.9544856548309326} 02/25/2022 04:41:13 - INFO - codeparrot_training - Step 20932: {'lr': 0.0003314126663196988, 'samples': 10717696, 'steps': 20932, 'loss/train': 1.8646478652954102} 02/25/2022 04:41:19 - INFO - codeparrot_training - Step 20933: {'lr': 0.0003313971955987573, 'samples': 10718208, 'steps': 20933, 'loss/train': 3.3239803314208984} 02/25/2022 04:41:22 - INFO - codeparrot_training - Step 20934: {'lr': 0.0003313817245291361, 'samples': 10718720, 'steps': 20934, 'loss/train': 0.28450432419776917} 02/25/2022 04:41:30 - INFO - codeparrot_training - Step 20935: {'lr': 0.0003313662531109014, 'samples': 10719232, 'steps': 20935, 'loss/train': 1.54340660572052} 02/25/2022 04:41:33 - INFO - codeparrot_training - Step 20936: {'lr': 0.00033135078134411956, 'samples': 10719744, 'steps': 20936, 'loss/train': 1.963658332824707} 02/25/2022 04:41:39 - INFO - codeparrot_training - Step 20937: {'lr': 0.0003313353092288568, 'samples': 10720256, 'steps': 20937, 'loss/train': 1.5553035736083984} 02/25/2022 04:41:42 - INFO - codeparrot_training - Step 20938: {'lr': 0.00033131983676517934, 'samples': 10720768, 'steps': 20938, 'loss/train': 1.9589520692825317} 02/25/2022 04:41:48 - INFO - codeparrot_training - Step 20939: {'lr': 0.0003313043639531536, 'samples': 10721280, 'steps': 20939, 'loss/train': 1.6424849033355713} 02/25/2022 04:41:51 - INFO - codeparrot_training - Step 20940: {'lr': 0.00033128889079284574, 'samples': 10721792, 'steps': 20940, 'loss/train': 1.3562875986099243} 02/25/2022 04:41:57 - INFO - codeparrot_training - Step 20941: {'lr': 0.0003312734172843221, 'samples': 10722304, 'steps': 20941, 'loss/train': 0.8118757009506226} 02/25/2022 04:42:00 - INFO - codeparrot_training - Step 20942: {'lr': 0.0003312579434276489, 'samples': 10722816, 'steps': 20942, 'loss/train': 1.8427571058273315} 02/25/2022 04:42:06 - INFO - codeparrot_training - Step 20943: {'lr': 0.0003312424692228925, 'samples': 10723328, 'steps': 20943, 'loss/train': 1.7319566011428833} 02/25/2022 04:42:10 - INFO - codeparrot_training - Step 20944: {'lr': 0.0003312269946701191, 'samples': 10723840, 'steps': 20944, 'loss/train': 2.4748294353485107} 02/25/2022 04:42:17 - INFO - codeparrot_training - Step 20945: {'lr': 0.0003312115197693951, 'samples': 10724352, 'steps': 20945, 'loss/train': 2.1030755043029785} 02/25/2022 04:42:20 - INFO - codeparrot_training - Step 20946: {'lr': 0.00033119604452078676, 'samples': 10724864, 'steps': 20946, 'loss/train': 2.4600119590759277} 02/25/2022 04:42:26 - INFO - codeparrot_training - Step 20947: {'lr': 0.00033118056892436035, 'samples': 10725376, 'steps': 20947, 'loss/train': 1.407612919807434} 02/25/2022 04:42:29 - INFO - codeparrot_training - Step 20948: {'lr': 0.00033116509298018217, 'samples': 10725888, 'steps': 20948, 'loss/train': 2.043806552886963} 02/25/2022 04:42:35 - INFO - codeparrot_training - Step 20949: {'lr': 0.00033114961668831845, 'samples': 10726400, 'steps': 20949, 'loss/train': 0.8096098899841309} 02/25/2022 04:42:38 - INFO - codeparrot_training - Step 20950: {'lr': 0.00033113414004883556, 'samples': 10726912, 'steps': 20950, 'loss/train': 1.237052321434021} 02/25/2022 04:42:44 - INFO - codeparrot_training - Step 20951: {'lr': 0.0003311186630617998, 'samples': 10727424, 'steps': 20951, 'loss/train': 2.1905226707458496} 02/25/2022 04:42:47 - INFO - codeparrot_training - Step 20952: {'lr': 0.00033110318572727743, 'samples': 10727936, 'steps': 20952, 'loss/train': 1.2035577297210693} 02/25/2022 04:42:53 - INFO - codeparrot_training - Step 20953: {'lr': 0.0003310877080453348, 'samples': 10728448, 'steps': 20953, 'loss/train': 2.5778281688690186} 02/25/2022 04:42:56 - INFO - codeparrot_training - Step 20954: {'lr': 0.00033107223001603814, 'samples': 10728960, 'steps': 20954, 'loss/train': 1.4431873559951782} 02/25/2022 04:43:02 - INFO - codeparrot_training - Step 20955: {'lr': 0.00033105675163945373, 'samples': 10729472, 'steps': 20955, 'loss/train': 2.041386127471924} 02/25/2022 04:43:05 - INFO - codeparrot_training - Step 20956: {'lr': 0.000331041272915648, 'samples': 10729984, 'steps': 20956, 'loss/train': 1.9621171951293945} 02/25/2022 04:43:11 - INFO - codeparrot_training - Step 20957: {'lr': 0.00033102579384468723, 'samples': 10730496, 'steps': 20957, 'loss/train': 1.3029671907424927} 02/25/2022 04:43:14 - INFO - codeparrot_training - Step 20958: {'lr': 0.0003310103144266376, 'samples': 10731008, 'steps': 20958, 'loss/train': 1.6323598623275757} 02/25/2022 04:43:20 - INFO - codeparrot_training - Step 20959: {'lr': 0.00033099483466156554, 'samples': 10731520, 'steps': 20959, 'loss/train': 2.477983236312866} 02/25/2022 04:43:23 - INFO - codeparrot_training - Step 20960: {'lr': 0.00033097935454953737, 'samples': 10732032, 'steps': 20960, 'loss/train': 0.12386111170053482} 02/25/2022 04:43:30 - INFO - codeparrot_training - Step 20961: {'lr': 0.00033096387409061937, 'samples': 10732544, 'steps': 20961, 'loss/train': 1.303353190422058} 02/25/2022 04:43:34 - INFO - codeparrot_training - Step 20962: {'lr': 0.00033094839328487777, 'samples': 10733056, 'steps': 20962, 'loss/train': 1.988746166229248} 02/25/2022 04:43:39 - INFO - codeparrot_training - Step 20963: {'lr': 0.000330932912132379, 'samples': 10733568, 'steps': 20963, 'loss/train': 2.444192409515381} 02/25/2022 04:43:43 - INFO - codeparrot_training - Step 20964: {'lr': 0.0003309174306331893, 'samples': 10734080, 'steps': 20964, 'loss/train': 2.0982134342193604} 02/25/2022 04:43:48 - INFO - codeparrot_training - Step 20965: {'lr': 0.00033090194878737504, 'samples': 10734592, 'steps': 20965, 'loss/train': 1.7582303285598755} 02/25/2022 04:43:52 - INFO - codeparrot_training - Step 20966: {'lr': 0.0003308864665950025, 'samples': 10735104, 'steps': 20966, 'loss/train': 1.6458404064178467} 02/25/2022 04:43:58 - INFO - codeparrot_training - Step 20967: {'lr': 0.0003308709840561381, 'samples': 10735616, 'steps': 20967, 'loss/train': 0.9111060500144958} 02/25/2022 04:44:01 - INFO - codeparrot_training - Step 20968: {'lr': 0.00033085550117084795, 'samples': 10736128, 'steps': 20968, 'loss/train': 2.4431984424591064} 02/25/2022 04:44:07 - INFO - codeparrot_training - Step 20969: {'lr': 0.0003308400179391986, 'samples': 10736640, 'steps': 20969, 'loss/train': 1.752737045288086} 02/25/2022 04:44:10 - INFO - codeparrot_training - Step 20970: {'lr': 0.00033082453436125627, 'samples': 10737152, 'steps': 20970, 'loss/train': 3.0392465591430664} 02/25/2022 04:44:17 - INFO - codeparrot_training - Step 20971: {'lr': 0.00033080905043708734, 'samples': 10737664, 'steps': 20971, 'loss/train': 1.7490004301071167} 02/25/2022 04:44:21 - INFO - codeparrot_training - Step 20972: {'lr': 0.000330793566166758, 'samples': 10738176, 'steps': 20972, 'loss/train': 2.33172345161438} 02/25/2022 04:44:26 - INFO - codeparrot_training - Step 20973: {'lr': 0.00033077808155033473, 'samples': 10738688, 'steps': 20973, 'loss/train': 3.273465633392334} 02/25/2022 04:44:30 - INFO - codeparrot_training - Step 20974: {'lr': 0.0003307625965878838, 'samples': 10739200, 'steps': 20974, 'loss/train': 1.5874143838882446} 02/25/2022 04:44:35 - INFO - codeparrot_training - Step 20975: {'lr': 0.00033074711127947153, 'samples': 10739712, 'steps': 20975, 'loss/train': 1.7917135953903198} 02/25/2022 04:44:39 - INFO - codeparrot_training - Step 20976: {'lr': 0.0003307316256251644, 'samples': 10740224, 'steps': 20976, 'loss/train': 2.751756191253662} 02/25/2022 04:44:44 - INFO - codeparrot_training - Step 20977: {'lr': 0.0003307161396250285, 'samples': 10740736, 'steps': 20977, 'loss/train': 1.7633999586105347} 02/25/2022 04:44:48 - INFO - codeparrot_training - Step 20978: {'lr': 0.00033070065327913035, 'samples': 10741248, 'steps': 20978, 'loss/train': 2.336181163787842} 02/25/2022 04:44:53 - INFO - codeparrot_training - Step 20979: {'lr': 0.00033068516658753624, 'samples': 10741760, 'steps': 20979, 'loss/train': 1.8295520544052124} 02/25/2022 04:44:57 - INFO - codeparrot_training - Step 20980: {'lr': 0.00033066967955031236, 'samples': 10742272, 'steps': 20980, 'loss/train': 1.3669748306274414} 02/25/2022 04:45:04 - INFO - codeparrot_training - Step 20981: {'lr': 0.0003306541921675253, 'samples': 10742784, 'steps': 20981, 'loss/train': 2.227687120437622} 02/25/2022 04:45:07 - INFO - codeparrot_training - Step 20982: {'lr': 0.0003306387044392413, 'samples': 10743296, 'steps': 20982, 'loss/train': 1.1107168197631836} 02/25/2022 04:45:13 - INFO - codeparrot_training - Step 20983: {'lr': 0.0003306232163655267, 'samples': 10743808, 'steps': 20983, 'loss/train': 1.6783326864242554} 02/25/2022 04:45:17 - INFO - codeparrot_training - Step 20984: {'lr': 0.00033060772794644776, 'samples': 10744320, 'steps': 20984, 'loss/train': 1.659469485282898} 02/25/2022 04:45:22 - INFO - codeparrot_training - Step 20985: {'lr': 0.000330592239182071, 'samples': 10744832, 'steps': 20985, 'loss/train': 2.3079824447631836} 02/25/2022 04:45:25 - INFO - codeparrot_training - Step 20986: {'lr': 0.0003305767500724626, 'samples': 10745344, 'steps': 20986, 'loss/train': 2.3354249000549316} 02/25/2022 04:45:31 - INFO - codeparrot_training - Step 20987: {'lr': 0.00033056126061768905, 'samples': 10745856, 'steps': 20987, 'loss/train': 1.2813560962677002} 02/25/2022 04:45:35 - INFO - codeparrot_training - Step 20988: {'lr': 0.00033054577081781654, 'samples': 10746368, 'steps': 20988, 'loss/train': 2.155677318572998} 02/25/2022 04:45:40 - INFO - codeparrot_training - Step 20989: {'lr': 0.00033053028067291166, 'samples': 10746880, 'steps': 20989, 'loss/train': 2.8354949951171875} 02/25/2022 04:45:44 - INFO - codeparrot_training - Step 20990: {'lr': 0.00033051479018304054, 'samples': 10747392, 'steps': 20990, 'loss/train': 1.4095090627670288} 02/25/2022 04:45:51 - INFO - codeparrot_training - Step 20991: {'lr': 0.0003304992993482697, 'samples': 10747904, 'steps': 20991, 'loss/train': 1.4379342794418335} 02/25/2022 04:45:54 - INFO - codeparrot_training - Step 20992: {'lr': 0.0003304838081686653, 'samples': 10748416, 'steps': 20992, 'loss/train': 2.53092622756958} 02/25/2022 04:46:00 - INFO - codeparrot_training - Step 20993: {'lr': 0.0003304683166442939, 'samples': 10748928, 'steps': 20993, 'loss/train': 1.5771119594573975} 02/25/2022 04:46:03 - INFO - codeparrot_training - Step 20994: {'lr': 0.0003304528247752218, 'samples': 10749440, 'steps': 20994, 'loss/train': 2.071444511413574} 02/25/2022 04:46:09 - INFO - codeparrot_training - Step 20995: {'lr': 0.0003304373325615153, 'samples': 10749952, 'steps': 20995, 'loss/train': 1.9969342947006226} 02/25/2022 04:46:12 - INFO - codeparrot_training - Step 20996: {'lr': 0.00033042184000324086, 'samples': 10750464, 'steps': 20996, 'loss/train': 2.3822269439697266} 02/25/2022 04:46:18 - INFO - codeparrot_training - Step 20997: {'lr': 0.00033040634710046474, 'samples': 10750976, 'steps': 20997, 'loss/train': 2.328132152557373} 02/25/2022 04:46:23 - INFO - codeparrot_training - Step 20998: {'lr': 0.0003303908538532534, 'samples': 10751488, 'steps': 20998, 'loss/train': 2.365709066390991} 02/25/2022 04:46:27 - INFO - codeparrot_training - Step 20999: {'lr': 0.00033037536026167313, 'samples': 10752000, 'steps': 20999, 'loss/train': 2.366279363632202} 02/25/2022 04:46:27 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 04:46:44 - WARNING - huggingface_hub.repository - Several commits (21) will be pushed upstream. 02/25/2022 04:46:44 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 04:47:19 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy ad14ca4..1098822 floral-grass-11 -> floral-grass-11 02/25/2022 04:47:26 - INFO - codeparrot_training - Step 21000: {'lr': 0.0003303598663257904, 'samples': 10752512, 'steps': 21000, 'loss/train': 2.4860072135925293} 02/25/2022 04:47:29 - INFO - codeparrot_training - Step 21001: {'lr': 0.00033034437204567145, 'samples': 10753024, 'steps': 21001, 'loss/train': 1.9799772500991821} 02/25/2022 04:47:35 - INFO - codeparrot_training - Step 21002: {'lr': 0.00033032887742138285, 'samples': 10753536, 'steps': 21002, 'loss/train': 1.8810853958129883} 02/25/2022 04:47:38 - INFO - codeparrot_training - Step 21003: {'lr': 0.0003303133824529907, 'samples': 10754048, 'steps': 21003, 'loss/train': 2.583627939224243} 02/25/2022 04:47:44 - INFO - codeparrot_training - Step 21004: {'lr': 0.00033029788714056165, 'samples': 10754560, 'steps': 21004, 'loss/train': 2.34503436088562} 02/25/2022 04:47:47 - INFO - codeparrot_training - Step 21005: {'lr': 0.0003302823914841618, 'samples': 10755072, 'steps': 21005, 'loss/train': 0.7265483736991882} 02/25/2022 04:47:54 - INFO - codeparrot_training - Step 21006: {'lr': 0.00033026689548385776, 'samples': 10755584, 'steps': 21006, 'loss/train': 0.3984239399433136} 02/25/2022 04:47:58 - INFO - codeparrot_training - Step 21007: {'lr': 0.00033025139913971585, 'samples': 10756096, 'steps': 21007, 'loss/train': 1.0154517889022827} 02/25/2022 04:48:03 - INFO - codeparrot_training - Step 21008: {'lr': 0.00033023590245180237, 'samples': 10756608, 'steps': 21008, 'loss/train': 2.5615317821502686} 02/25/2022 04:48:07 - INFO - codeparrot_training - Step 21009: {'lr': 0.00033022040542018385, 'samples': 10757120, 'steps': 21009, 'loss/train': 1.0869808197021484} 02/25/2022 04:48:12 - INFO - codeparrot_training - Step 21010: {'lr': 0.0003302049080449265, 'samples': 10757632, 'steps': 21010, 'loss/train': 3.8830811977386475} 02/25/2022 04:48:16 - INFO - codeparrot_training - Step 21011: {'lr': 0.0003301894103260968, 'samples': 10758144, 'steps': 21011, 'loss/train': 1.8186558485031128} 02/25/2022 04:48:21 - INFO - codeparrot_training - Step 21012: {'lr': 0.0003301739122637611, 'samples': 10758656, 'steps': 21012, 'loss/train': 1.922775387763977} 02/25/2022 04:48:25 - INFO - codeparrot_training - Step 21013: {'lr': 0.00033015841385798596, 'samples': 10759168, 'steps': 21013, 'loss/train': 1.9774383306503296} 02/25/2022 04:48:30 - INFO - codeparrot_training - Step 21014: {'lr': 0.00033014291510883746, 'samples': 10759680, 'steps': 21014, 'loss/train': 1.828243613243103} 02/25/2022 04:48:34 - INFO - codeparrot_training - Step 21015: {'lr': 0.0003301274160163823, 'samples': 10760192, 'steps': 21015, 'loss/train': 2.2705297470092773} 02/25/2022 04:48:41 - INFO - codeparrot_training - Step 21016: {'lr': 0.00033011191658068663, 'samples': 10760704, 'steps': 21016, 'loss/train': 2.7779653072357178} 02/25/2022 04:48:45 - INFO - codeparrot_training - Step 21017: {'lr': 0.00033009641680181696, 'samples': 10761216, 'steps': 21017, 'loss/train': 0.4270228445529938} 02/25/2022 04:48:50 - INFO - codeparrot_training - Step 21018: {'lr': 0.00033008091667983974, 'samples': 10761728, 'steps': 21018, 'loss/train': 2.462270975112915} 02/25/2022 04:48:54 - INFO - codeparrot_training - Step 21019: {'lr': 0.0003300654162148213, 'samples': 10762240, 'steps': 21019, 'loss/train': 1.0128071308135986} 02/25/2022 04:48:59 - INFO - codeparrot_training - Step 21020: {'lr': 0.00033004991540682793, 'samples': 10762752, 'steps': 21020, 'loss/train': 0.6379898190498352} 02/25/2022 04:49:03 - INFO - codeparrot_training - Step 21021: {'lr': 0.00033003441425592627, 'samples': 10763264, 'steps': 21021, 'loss/train': 1.2960740327835083} 02/25/2022 04:49:08 - INFO - codeparrot_training - Step 21022: {'lr': 0.00033001891276218247, 'samples': 10763776, 'steps': 21022, 'loss/train': 1.744156002998352} 02/25/2022 04:49:12 - INFO - codeparrot_training - Step 21023: {'lr': 0.0003300034109256632, 'samples': 10764288, 'steps': 21023, 'loss/train': 1.010110855102539} 02/25/2022 04:49:17 - INFO - codeparrot_training - Step 21024: {'lr': 0.00032998790874643456, 'samples': 10764800, 'steps': 21024, 'loss/train': 0.2765105366706848} 02/25/2022 04:49:21 - INFO - codeparrot_training - Step 21025: {'lr': 0.00032997240622456326, 'samples': 10765312, 'steps': 21025, 'loss/train': 1.4589354991912842} 02/25/2022 04:49:28 - INFO - codeparrot_training - Step 21026: {'lr': 0.0003299569033601155, 'samples': 10765824, 'steps': 21026, 'loss/train': 1.5648595094680786} 02/25/2022 04:49:32 - INFO - codeparrot_training - Step 21027: {'lr': 0.0003299414001531578, 'samples': 10766336, 'steps': 21027, 'loss/train': 2.5697457790374756} 02/25/2022 04:49:37 - INFO - codeparrot_training - Step 21028: {'lr': 0.0003299258966037565, 'samples': 10766848, 'steps': 21028, 'loss/train': 2.44779896736145} 02/25/2022 04:49:41 - INFO - codeparrot_training - Step 21029: {'lr': 0.000329910392711978, 'samples': 10767360, 'steps': 21029, 'loss/train': 1.4116836786270142} 02/25/2022 04:49:46 - INFO - codeparrot_training - Step 21030: {'lr': 0.0003298948884778887, 'samples': 10767872, 'steps': 21030, 'loss/train': 2.0573225021362305} 02/25/2022 04:49:50 - INFO - codeparrot_training - Step 21031: {'lr': 0.00032987938390155523, 'samples': 10768384, 'steps': 21031, 'loss/train': 1.0688015222549438} 02/25/2022 04:49:55 - INFO - codeparrot_training - Step 21032: {'lr': 0.00032986387898304375, 'samples': 10768896, 'steps': 21032, 'loss/train': 1.879749059677124} 02/25/2022 04:49:59 - INFO - codeparrot_training - Step 21033: {'lr': 0.00032984837372242084, 'samples': 10769408, 'steps': 21033, 'loss/train': 2.6197383403778076} 02/25/2022 04:50:04 - INFO - codeparrot_training - Step 21034: {'lr': 0.0003298328681197528, 'samples': 10769920, 'steps': 21034, 'loss/train': 1.3640557527542114} 02/25/2022 04:50:08 - INFO - codeparrot_training - Step 21035: {'lr': 0.00032981736217510603, 'samples': 10770432, 'steps': 21035, 'loss/train': 0.510646402835846} 02/25/2022 04:50:13 - INFO - codeparrot_training - Step 21036: {'lr': 0.0003298018558885471, 'samples': 10770944, 'steps': 21036, 'loss/train': 2.5600621700286865} 02/25/2022 04:50:17 - INFO - codeparrot_training - Step 21037: {'lr': 0.0003297863492601424, 'samples': 10771456, 'steps': 21037, 'loss/train': 1.5862576961517334} 02/25/2022 04:50:24 - INFO - codeparrot_training - Step 21038: {'lr': 0.0003297708422899582, 'samples': 10771968, 'steps': 21038, 'loss/train': 2.7320468425750732} 02/25/2022 04:50:27 - INFO - codeparrot_training - Step 21039: {'lr': 0.0003297553349780612, 'samples': 10772480, 'steps': 21039, 'loss/train': 1.9368572235107422} 02/25/2022 04:50:33 - INFO - codeparrot_training - Step 21040: {'lr': 0.0003297398273245175, 'samples': 10772992, 'steps': 21040, 'loss/train': 1.3581600189208984} 02/25/2022 04:50:37 - INFO - codeparrot_training - Step 21041: {'lr': 0.0003297243193293938, 'samples': 10773504, 'steps': 21041, 'loss/train': 1.6091434955596924} 02/25/2022 04:50:42 - INFO - codeparrot_training - Step 21042: {'lr': 0.0003297088109927564, 'samples': 10774016, 'steps': 21042, 'loss/train': 1.7612583637237549} 02/25/2022 04:50:46 - INFO - codeparrot_training - Step 21043: {'lr': 0.00032969330231467177, 'samples': 10774528, 'steps': 21043, 'loss/train': 2.51531720161438} 02/25/2022 04:50:51 - INFO - codeparrot_training - Step 21044: {'lr': 0.0003296777932952064, 'samples': 10775040, 'steps': 21044, 'loss/train': 1.769242525100708} 02/25/2022 04:50:55 - INFO - codeparrot_training - Step 21045: {'lr': 0.0003296622839344265, 'samples': 10775552, 'steps': 21045, 'loss/train': 1.7496014833450317} 02/25/2022 04:51:01 - INFO - codeparrot_training - Step 21046: {'lr': 0.00032964677423239885, 'samples': 10776064, 'steps': 21046, 'loss/train': 0.8651546239852905} 02/25/2022 04:51:04 - INFO - codeparrot_training - Step 21047: {'lr': 0.0003296312641891896, 'samples': 10776576, 'steps': 21047, 'loss/train': 2.37147855758667} 02/25/2022 04:51:10 - INFO - codeparrot_training - Step 21048: {'lr': 0.0003296157538048654, 'samples': 10777088, 'steps': 21048, 'loss/train': 2.2832863330841064} 02/25/2022 04:51:13 - INFO - codeparrot_training - Step 21049: {'lr': 0.0003296002430794925, 'samples': 10777600, 'steps': 21049, 'loss/train': 1.270258903503418} 02/25/2022 04:51:18 - INFO - codeparrot_training - Step 21050: {'lr': 0.00032958473201313745, 'samples': 10778112, 'steps': 21050, 'loss/train': 2.1549997329711914} 02/25/2022 04:51:22 - INFO - codeparrot_training - Step 21051: {'lr': 0.0003295692206058667, 'samples': 10778624, 'steps': 21051, 'loss/train': 2.7490174770355225} 02/25/2022 04:51:29 - INFO - codeparrot_training - Step 21052: {'lr': 0.00032955370885774665, 'samples': 10779136, 'steps': 21052, 'loss/train': 1.6762350797653198} 02/25/2022 04:51:33 - INFO - codeparrot_training - Step 21053: {'lr': 0.0003295381967688438, 'samples': 10779648, 'steps': 21053, 'loss/train': 1.6603776216506958} 02/25/2022 04:51:38 - INFO - codeparrot_training - Step 21054: {'lr': 0.0003295226843392245, 'samples': 10780160, 'steps': 21054, 'loss/train': 2.773618459701538} 02/25/2022 04:51:42 - INFO - codeparrot_training - Step 21055: {'lr': 0.0003295071715689554, 'samples': 10780672, 'steps': 21055, 'loss/train': 1.3425158262252808} 02/25/2022 04:51:47 - INFO - codeparrot_training - Step 21056: {'lr': 0.0003294916584581027, 'samples': 10781184, 'steps': 21056, 'loss/train': 1.8481996059417725} 02/25/2022 04:51:51 - INFO - codeparrot_training - Step 21057: {'lr': 0.00032947614500673306, 'samples': 10781696, 'steps': 21057, 'loss/train': 1.491181492805481} 02/25/2022 04:51:56 - INFO - codeparrot_training - Step 21058: {'lr': 0.0003294606312149128, 'samples': 10782208, 'steps': 21058, 'loss/train': 1.9579530954360962} 02/25/2022 04:52:00 - INFO - codeparrot_training - Step 21059: {'lr': 0.00032944511708270853, 'samples': 10782720, 'steps': 21059, 'loss/train': 2.503943681716919} 02/25/2022 04:52:05 - INFO - codeparrot_training - Step 21060: {'lr': 0.00032942960261018653, 'samples': 10783232, 'steps': 21060, 'loss/train': 3.124640464782715} 02/25/2022 04:52:09 - INFO - codeparrot_training - Step 21061: {'lr': 0.0003294140877974133, 'samples': 10783744, 'steps': 21061, 'loss/train': 1.4339001178741455} 02/25/2022 04:52:14 - INFO - codeparrot_training - Step 21062: {'lr': 0.00032939857264445535, 'samples': 10784256, 'steps': 21062, 'loss/train': 2.0752813816070557} 02/25/2022 04:52:18 - INFO - codeparrot_training - Step 21063: {'lr': 0.00032938305715137916, 'samples': 10784768, 'steps': 21063, 'loss/train': 1.7039581537246704} 02/25/2022 04:52:25 - INFO - codeparrot_training - Step 21064: {'lr': 0.00032936754131825113, 'samples': 10785280, 'steps': 21064, 'loss/train': 1.726387619972229} 02/25/2022 04:52:29 - INFO - codeparrot_training - Step 21065: {'lr': 0.0003293520251451378, 'samples': 10785792, 'steps': 21065, 'loss/train': 2.150524377822876} 02/25/2022 04:52:34 - INFO - codeparrot_training - Step 21066: {'lr': 0.00032933650863210557, 'samples': 10786304, 'steps': 21066, 'loss/train': 3.0293538570404053} 02/25/2022 04:52:38 - INFO - codeparrot_training - Step 21067: {'lr': 0.00032932099177922095, 'samples': 10786816, 'steps': 21067, 'loss/train': 2.0431742668151855} 02/25/2022 04:52:43 - INFO - codeparrot_training - Step 21068: {'lr': 0.00032930547458655035, 'samples': 10787328, 'steps': 21068, 'loss/train': 2.278754949569702} 02/25/2022 04:52:47 - INFO - codeparrot_training - Step 21069: {'lr': 0.0003292899570541603, 'samples': 10787840, 'steps': 21069, 'loss/train': 1.818558692932129} 02/25/2022 04:52:52 - INFO - codeparrot_training - Step 21070: {'lr': 0.00032927443918211717, 'samples': 10788352, 'steps': 21070, 'loss/train': 1.162871241569519} 02/25/2022 04:52:56 - INFO - codeparrot_training - Step 21071: {'lr': 0.0003292589209704876, 'samples': 10788864, 'steps': 21071, 'loss/train': 2.547092914581299} 02/25/2022 04:53:01 - INFO - codeparrot_training - Step 21072: {'lr': 0.000329243402419338, 'samples': 10789376, 'steps': 21072, 'loss/train': 1.8579257726669312} 02/25/2022 04:53:05 - INFO - codeparrot_training - Step 21073: {'lr': 0.00032922788352873474, 'samples': 10789888, 'steps': 21073, 'loss/train': 2.371066093444824} 02/25/2022 04:53:12 - INFO - codeparrot_training - Step 21074: {'lr': 0.00032921236429874444, 'samples': 10790400, 'steps': 21074, 'loss/train': 2.1397194862365723} 02/25/2022 04:53:15 - INFO - codeparrot_training - Step 21075: {'lr': 0.0003291968447294335, 'samples': 10790912, 'steps': 21075, 'loss/train': 1.8102643489837646} 02/25/2022 04:53:21 - INFO - codeparrot_training - Step 21076: {'lr': 0.0003291813248208685, 'samples': 10791424, 'steps': 21076, 'loss/train': 1.8460663557052612} 02/25/2022 04:53:24 - INFO - codeparrot_training - Step 21077: {'lr': 0.00032916580457311573, 'samples': 10791936, 'steps': 21077, 'loss/train': 3.097266674041748} 02/25/2022 04:53:30 - INFO - codeparrot_training - Step 21078: {'lr': 0.00032915028398624186, 'samples': 10792448, 'steps': 21078, 'loss/train': 1.4640593528747559} 02/25/2022 04:53:33 - INFO - codeparrot_training - Step 21079: {'lr': 0.0003291347630603133, 'samples': 10792960, 'steps': 21079, 'loss/train': 2.1472299098968506} 02/25/2022 04:53:39 - INFO - codeparrot_training - Step 21080: {'lr': 0.00032911924179539653, 'samples': 10793472, 'steps': 21080, 'loss/train': 2.6353187561035156} 02/25/2022 04:53:42 - INFO - codeparrot_training - Step 21081: {'lr': 0.0003291037201915581, 'samples': 10793984, 'steps': 21081, 'loss/train': 0.8202171921730042} 02/25/2022 04:53:48 - INFO - codeparrot_training - Step 21082: {'lr': 0.0003290881982488644, 'samples': 10794496, 'steps': 21082, 'loss/train': 3.2697031497955322} 02/25/2022 04:53:51 - INFO - codeparrot_training - Step 21083: {'lr': 0.000329072675967382, 'samples': 10795008, 'steps': 21083, 'loss/train': 1.1767582893371582} 02/25/2022 04:53:59 - INFO - codeparrot_training - Step 21084: {'lr': 0.00032905715334717735, 'samples': 10795520, 'steps': 21084, 'loss/train': 2.0095412731170654} 02/25/2022 04:54:02 - INFO - codeparrot_training - Step 21085: {'lr': 0.000329041630388317, 'samples': 10796032, 'steps': 21085, 'loss/train': 2.0435166358947754} 02/25/2022 04:54:08 - INFO - codeparrot_training - Step 21086: {'lr': 0.00032902610709086727, 'samples': 10796544, 'steps': 21086, 'loss/train': 8.830096244812012} 02/25/2022 04:54:11 - INFO - codeparrot_training - Step 21087: {'lr': 0.00032901058345489494, 'samples': 10797056, 'steps': 21087, 'loss/train': 1.9922130107879639} 02/25/2022 04:54:17 - INFO - codeparrot_training - Step 21088: {'lr': 0.00032899505948046624, 'samples': 10797568, 'steps': 21088, 'loss/train': 2.054474353790283} 02/25/2022 04:54:20 - INFO - codeparrot_training - Step 21089: {'lr': 0.0003289795351676479, 'samples': 10798080, 'steps': 21089, 'loss/train': 2.3793280124664307} 02/25/2022 04:54:26 - INFO - codeparrot_training - Step 21090: {'lr': 0.0003289640105165063, 'samples': 10798592, 'steps': 21090, 'loss/train': 2.679232358932495} 02/25/2022 04:54:29 - INFO - codeparrot_training - Step 21091: {'lr': 0.0003289484855271078, 'samples': 10799104, 'steps': 21091, 'loss/train': 1.937221884727478} 02/25/2022 04:54:35 - INFO - codeparrot_training - Step 21092: {'lr': 0.0003289329601995192, 'samples': 10799616, 'steps': 21092, 'loss/train': 1.3778191804885864} 02/25/2022 04:54:38 - INFO - codeparrot_training - Step 21093: {'lr': 0.00032891743453380685, 'samples': 10800128, 'steps': 21093, 'loss/train': 1.8153339624404907} 02/25/2022 04:54:46 - INFO - codeparrot_training - Step 21094: {'lr': 0.00032890190853003727, 'samples': 10800640, 'steps': 21094, 'loss/train': 2.862828254699707} 02/25/2022 04:54:49 - INFO - codeparrot_training - Step 21095: {'lr': 0.0003288863821882769, 'samples': 10801152, 'steps': 21095, 'loss/train': 2.6834182739257812} 02/25/2022 04:54:55 - INFO - codeparrot_training - Step 21096: {'lr': 0.0003288708555085924, 'samples': 10801664, 'steps': 21096, 'loss/train': 1.8154313564300537} 02/25/2022 04:54:58 - INFO - codeparrot_training - Step 21097: {'lr': 0.00032885532849105014, 'samples': 10802176, 'steps': 21097, 'loss/train': 1.644252896308899} 02/25/2022 04:55:04 - INFO - codeparrot_training - Step 21098: {'lr': 0.0003288398011357168, 'samples': 10802688, 'steps': 21098, 'loss/train': 2.0015244483947754} 02/25/2022 04:55:07 - INFO - codeparrot_training - Step 21099: {'lr': 0.0003288242734426586, 'samples': 10803200, 'steps': 21099, 'loss/train': 2.689619779586792} 02/25/2022 04:55:13 - INFO - codeparrot_training - Step 21100: {'lr': 0.00032880874541194244, 'samples': 10803712, 'steps': 21100, 'loss/train': 1.8975166082382202} 02/25/2022 04:55:16 - INFO - codeparrot_training - Step 21101: {'lr': 0.0003287932170436345, 'samples': 10804224, 'steps': 21101, 'loss/train': 2.0911428928375244} 02/25/2022 04:55:22 - INFO - codeparrot_training - Step 21102: {'lr': 0.00032877768833780146, 'samples': 10804736, 'steps': 21102, 'loss/train': 2.481397867202759} 02/25/2022 04:55:25 - INFO - codeparrot_training - Step 21103: {'lr': 0.0003287621592945099, 'samples': 10805248, 'steps': 21103, 'loss/train': 1.8134071826934814} 02/25/2022 04:55:31 - INFO - codeparrot_training - Step 21104: {'lr': 0.0003287466299138262, 'samples': 10805760, 'steps': 21104, 'loss/train': 2.027007818222046} 02/25/2022 04:55:34 - INFO - codeparrot_training - Step 21105: {'lr': 0.000328731100195817, 'samples': 10806272, 'steps': 21105, 'loss/train': 1.9264589548110962} 02/25/2022 04:55:40 - INFO - codeparrot_training - Step 21106: {'lr': 0.00032871557014054864, 'samples': 10806784, 'steps': 21106, 'loss/train': 1.8219751119613647} 02/25/2022 04:55:43 - INFO - codeparrot_training - Step 21107: {'lr': 0.00032870003974808787, 'samples': 10807296, 'steps': 21107, 'loss/train': 1.6427385807037354} 02/25/2022 04:55:50 - INFO - codeparrot_training - Step 21108: {'lr': 0.0003286845090185011, 'samples': 10807808, 'steps': 21108, 'loss/train': 1.62216055393219} 02/25/2022 04:55:54 - INFO - codeparrot_training - Step 21109: {'lr': 0.0003286689779518549, 'samples': 10808320, 'steps': 21109, 'loss/train': 2.5330448150634766} 02/25/2022 04:55:59 - INFO - codeparrot_training - Step 21110: {'lr': 0.00032865344654821576, 'samples': 10808832, 'steps': 21110, 'loss/train': 1.924042820930481} 02/25/2022 04:56:03 - INFO - codeparrot_training - Step 21111: {'lr': 0.0003286379148076502, 'samples': 10809344, 'steps': 21111, 'loss/train': 2.2303214073181152} 02/25/2022 04:56:08 - INFO - codeparrot_training - Step 21112: {'lr': 0.00032862238273022483, 'samples': 10809856, 'steps': 21112, 'loss/train': 2.805931568145752} 02/25/2022 04:56:12 - INFO - codeparrot_training - Step 21113: {'lr': 0.0003286068503160061, 'samples': 10810368, 'steps': 21113, 'loss/train': 1.8496837615966797} 02/25/2022 04:56:17 - INFO - codeparrot_training - Step 21114: {'lr': 0.0003285913175650606, 'samples': 10810880, 'steps': 21114, 'loss/train': 1.8455113172531128} 02/25/2022 04:56:21 - INFO - codeparrot_training - Step 21115: {'lr': 0.00032857578447745484, 'samples': 10811392, 'steps': 21115, 'loss/train': 0.6516101956367493} 02/25/2022 04:56:26 - INFO - codeparrot_training - Step 21116: {'lr': 0.00032856025105325537, 'samples': 10811904, 'steps': 21116, 'loss/train': 1.9704476594924927} 02/25/2022 04:56:30 - INFO - codeparrot_training - Step 21117: {'lr': 0.00032854471729252876, 'samples': 10812416, 'steps': 21117, 'loss/train': 0.9231125116348267} 02/25/2022 04:56:35 - INFO - codeparrot_training - Step 21118: {'lr': 0.00032852918319534153, 'samples': 10812928, 'steps': 21118, 'loss/train': 2.7022061347961426} 02/25/2022 04:56:39 - INFO - codeparrot_training - Step 21119: {'lr': 0.00032851364876176014, 'samples': 10813440, 'steps': 21119, 'loss/train': 1.6050589084625244} 02/25/2022 04:56:46 - INFO - codeparrot_training - Step 21120: {'lr': 0.00032849811399185126, 'samples': 10813952, 'steps': 21120, 'loss/train': 1.2908967733383179} 02/25/2022 04:56:50 - INFO - codeparrot_training - Step 21121: {'lr': 0.0003284825788856814, 'samples': 10814464, 'steps': 21121, 'loss/train': 2.0518288612365723} 02/25/2022 04:56:55 - INFO - codeparrot_training - Step 21122: {'lr': 0.00032846704344331707, 'samples': 10814976, 'steps': 21122, 'loss/train': 1.7153793573379517} 02/25/2022 04:56:59 - INFO - codeparrot_training - Step 21123: {'lr': 0.00032845150766482484, 'samples': 10815488, 'steps': 21123, 'loss/train': 2.383584499359131} 02/25/2022 04:57:04 - INFO - codeparrot_training - Step 21124: {'lr': 0.0003284359715502713, 'samples': 10816000, 'steps': 21124, 'loss/train': 2.0794625282287598} 02/25/2022 04:57:08 - INFO - codeparrot_training - Step 21125: {'lr': 0.0003284204350997229, 'samples': 10816512, 'steps': 21125, 'loss/train': 2.3251895904541016} 02/25/2022 04:57:13 - INFO - codeparrot_training - Step 21126: {'lr': 0.0003284048983132463, 'samples': 10817024, 'steps': 21126, 'loss/train': 1.4006412029266357} 02/25/2022 04:57:17 - INFO - codeparrot_training - Step 21127: {'lr': 0.000328389361190908, 'samples': 10817536, 'steps': 21127, 'loss/train': 3.051767587661743} 02/25/2022 04:57:22 - INFO - codeparrot_training - Step 21128: {'lr': 0.0003283738237327745, 'samples': 10818048, 'steps': 21128, 'loss/train': 2.5093495845794678} 02/25/2022 04:57:26 - INFO - codeparrot_training - Step 21129: {'lr': 0.0003283582859389125, 'samples': 10818560, 'steps': 21129, 'loss/train': 1.7888586521148682} 02/25/2022 04:57:33 - INFO - codeparrot_training - Step 21130: {'lr': 0.0003283427478093885, 'samples': 10819072, 'steps': 21130, 'loss/train': 2.1881496906280518} 02/25/2022 04:57:36 - INFO - codeparrot_training - Step 21131: {'lr': 0.0003283272093442691, 'samples': 10819584, 'steps': 21131, 'loss/train': 2.4616143703460693} 02/25/2022 04:57:42 - INFO - codeparrot_training - Step 21132: {'lr': 0.00032831167054362065, 'samples': 10820096, 'steps': 21132, 'loss/train': 1.0961674451828003} 02/25/2022 04:57:45 - INFO - codeparrot_training - Step 21133: {'lr': 0.00032829613140751004, 'samples': 10820608, 'steps': 21133, 'loss/train': 1.1025972366333008} 02/25/2022 04:57:51 - INFO - codeparrot_training - Step 21134: {'lr': 0.0003282805919360035, 'samples': 10821120, 'steps': 21134, 'loss/train': 2.0721688270568848} 02/25/2022 04:57:54 - INFO - codeparrot_training - Step 21135: {'lr': 0.0003282650521291679, 'samples': 10821632, 'steps': 21135, 'loss/train': 1.9505776166915894} 02/25/2022 04:58:00 - INFO - codeparrot_training - Step 21136: {'lr': 0.0003282495119870695, 'samples': 10822144, 'steps': 21136, 'loss/train': 1.6411024332046509} 02/25/2022 04:58:04 - INFO - codeparrot_training - Step 21137: {'lr': 0.00032823397150977523, 'samples': 10822656, 'steps': 21137, 'loss/train': 1.9987449645996094} 02/25/2022 04:58:09 - INFO - codeparrot_training - Step 21138: {'lr': 0.00032821843069735134, 'samples': 10823168, 'steps': 21138, 'loss/train': 1.6433777809143066} 02/25/2022 04:58:13 - INFO - codeparrot_training - Step 21139: {'lr': 0.00032820288954986453, 'samples': 10823680, 'steps': 21139, 'loss/train': 1.414487600326538} 02/25/2022 04:58:20 - INFO - codeparrot_training - Step 21140: {'lr': 0.00032818734806738147, 'samples': 10824192, 'steps': 21140, 'loss/train': 1.5155943632125854} 02/25/2022 04:58:23 - INFO - codeparrot_training - Step 21141: {'lr': 0.0003281718062499686, 'samples': 10824704, 'steps': 21141, 'loss/train': 3.074453353881836} 02/25/2022 04:58:29 - INFO - codeparrot_training - Step 21142: {'lr': 0.0003281562640976925, 'samples': 10825216, 'steps': 21142, 'loss/train': 2.4185543060302734} 02/25/2022 04:58:32 - INFO - codeparrot_training - Step 21143: {'lr': 0.0003281407216106198, 'samples': 10825728, 'steps': 21143, 'loss/train': 1.5754749774932861} 02/25/2022 04:58:38 - INFO - codeparrot_training - Step 21144: {'lr': 0.0003281251787888171, 'samples': 10826240, 'steps': 21144, 'loss/train': 2.1936450004577637} 02/25/2022 04:58:41 - INFO - codeparrot_training - Step 21145: {'lr': 0.00032810963563235083, 'samples': 10826752, 'steps': 21145, 'loss/train': 0.7120290398597717} 02/25/2022 04:58:47 - INFO - codeparrot_training - Step 21146: {'lr': 0.00032809409214128784, 'samples': 10827264, 'steps': 21146, 'loss/train': 1.374009132385254} 02/25/2022 04:58:50 - INFO - codeparrot_training - Step 21147: {'lr': 0.0003280785483156944, 'samples': 10827776, 'steps': 21147, 'loss/train': 2.015887975692749} 02/25/2022 04:58:56 - INFO - codeparrot_training - Step 21148: {'lr': 0.0003280630041556374, 'samples': 10828288, 'steps': 21148, 'loss/train': 2.152256488800049} 02/25/2022 04:58:59 - INFO - codeparrot_training - Step 21149: {'lr': 0.0003280474596611832, 'samples': 10828800, 'steps': 21149, 'loss/train': 1.2193974256515503} 02/25/2022 04:59:05 - INFO - codeparrot_training - Step 21150: {'lr': 0.0003280319148323985, 'samples': 10829312, 'steps': 21150, 'loss/train': 1.5585376024246216} 02/25/2022 04:59:08 - INFO - codeparrot_training - Step 21151: {'lr': 0.00032801636966934975, 'samples': 10829824, 'steps': 21151, 'loss/train': 2.268585681915283} 02/25/2022 04:59:14 - INFO - codeparrot_training - Step 21152: {'lr': 0.0003280008241721038, 'samples': 10830336, 'steps': 21152, 'loss/train': 2.3331964015960693} 02/25/2022 04:59:17 - INFO - codeparrot_training - Step 21153: {'lr': 0.0003279852783407269, 'samples': 10830848, 'steps': 21153, 'loss/train': 0.49396517872810364} 02/25/2022 04:59:23 - INFO - codeparrot_training - Step 21154: {'lr': 0.00032796973217528595, 'samples': 10831360, 'steps': 21154, 'loss/train': 1.787941575050354} 02/25/2022 04:59:26 - INFO - codeparrot_training - Step 21155: {'lr': 0.00032795418567584746, 'samples': 10831872, 'steps': 21155, 'loss/train': 1.124595046043396} 02/25/2022 04:59:34 - INFO - codeparrot_training - Step 21156: {'lr': 0.00032793863884247794, 'samples': 10832384, 'steps': 21156, 'loss/train': 1.5419896841049194} 02/25/2022 04:59:37 - INFO - codeparrot_training - Step 21157: {'lr': 0.00032792309167524404, 'samples': 10832896, 'steps': 21157, 'loss/train': 8.86019229888916} 02/25/2022 04:59:43 - INFO - codeparrot_training - Step 21158: {'lr': 0.00032790754417421237, 'samples': 10833408, 'steps': 21158, 'loss/train': 2.938297748565674} 02/25/2022 04:59:46 - INFO - codeparrot_training - Step 21159: {'lr': 0.00032789199633944955, 'samples': 10833920, 'steps': 21159, 'loss/train': 2.118612289428711} 02/25/2022 04:59:52 - INFO - codeparrot_training - Step 21160: {'lr': 0.00032787644817102205, 'samples': 10834432, 'steps': 21160, 'loss/train': 1.8754611015319824} 02/25/2022 04:59:55 - INFO - codeparrot_training - Step 21161: {'lr': 0.00032786089966899666, 'samples': 10834944, 'steps': 21161, 'loss/train': 1.1040573120117188} 02/25/2022 05:00:01 - INFO - codeparrot_training - Step 21162: {'lr': 0.0003278453508334399, 'samples': 10835456, 'steps': 21162, 'loss/train': 1.3849550485610962} 02/25/2022 05:00:04 - INFO - codeparrot_training - Step 21163: {'lr': 0.00032782980166441836, 'samples': 10835968, 'steps': 21163, 'loss/train': 1.6671080589294434} 02/25/2022 05:00:10 - INFO - codeparrot_training - Step 21164: {'lr': 0.00032781425216199864, 'samples': 10836480, 'steps': 21164, 'loss/train': 0.5320336222648621} 02/25/2022 05:00:17 - INFO - codeparrot_training - Step 21165: {'lr': 0.00032779870232624737, 'samples': 10836992, 'steps': 21165, 'loss/train': 2.697252035140991} 02/25/2022 05:00:20 - INFO - codeparrot_training - Step 21166: {'lr': 0.0003277831521572312, 'samples': 10837504, 'steps': 21166, 'loss/train': 2.136936664581299} 02/25/2022 05:00:26 - INFO - codeparrot_training - Step 21167: {'lr': 0.00032776760165501663, 'samples': 10838016, 'steps': 21167, 'loss/train': 2.4128663539886475} 02/25/2022 05:00:29 - INFO - codeparrot_training - Step 21168: {'lr': 0.00032775205081967047, 'samples': 10838528, 'steps': 21168, 'loss/train': 2.4195454120635986} 02/25/2022 05:00:35 - INFO - codeparrot_training - Step 21169: {'lr': 0.00032773649965125914, 'samples': 10839040, 'steps': 21169, 'loss/train': 2.199857234954834} 02/25/2022 05:00:38 - INFO - codeparrot_training - Step 21170: {'lr': 0.0003277209481498493, 'samples': 10839552, 'steps': 21170, 'loss/train': 2.2646963596343994} 02/25/2022 05:00:44 - INFO - codeparrot_training - Step 21171: {'lr': 0.00032770539631550767, 'samples': 10840064, 'steps': 21171, 'loss/train': 2.768000364303589} 02/25/2022 05:00:47 - INFO - codeparrot_training - Step 21172: {'lr': 0.0003276898441483008, 'samples': 10840576, 'steps': 21172, 'loss/train': 1.887044906616211} 02/25/2022 05:00:53 - INFO - codeparrot_training - Step 21173: {'lr': 0.0003276742916482952, 'samples': 10841088, 'steps': 21173, 'loss/train': 2.0410633087158203} 02/25/2022 05:00:56 - INFO - codeparrot_training - Step 21174: {'lr': 0.00032765873881555765, 'samples': 10841600, 'steps': 21174, 'loss/train': 1.6916298866271973} 02/25/2022 05:01:03 - INFO - codeparrot_training - Step 21175: {'lr': 0.0003276431856501548, 'samples': 10842112, 'steps': 21175, 'loss/train': 1.9030319452285767} 02/25/2022 05:01:07 - INFO - codeparrot_training - Step 21176: {'lr': 0.0003276276321521531, 'samples': 10842624, 'steps': 21176, 'loss/train': 2.1897199153900146} 02/25/2022 05:01:12 - INFO - codeparrot_training - Step 21177: {'lr': 0.00032761207832161934, 'samples': 10843136, 'steps': 21177, 'loss/train': 1.475611925125122} 02/25/2022 05:01:16 - INFO - codeparrot_training - Step 21178: {'lr': 0.00032759652415862003, 'samples': 10843648, 'steps': 21178, 'loss/train': 2.7832348346710205} 02/25/2022 05:01:21 - INFO - codeparrot_training - Step 21179: {'lr': 0.00032758096966322185, 'samples': 10844160, 'steps': 21179, 'loss/train': 2.833794355392456} 02/25/2022 05:01:25 - INFO - codeparrot_training - Step 21180: {'lr': 0.00032756541483549146, 'samples': 10844672, 'steps': 21180, 'loss/train': 1.7616722583770752} 02/25/2022 05:01:30 - INFO - codeparrot_training - Step 21181: {'lr': 0.00032754985967549545, 'samples': 10845184, 'steps': 21181, 'loss/train': 2.2143542766571045} 02/25/2022 05:01:34 - INFO - codeparrot_training - Step 21182: {'lr': 0.0003275343041833005, 'samples': 10845696, 'steps': 21182, 'loss/train': 2.070802688598633} 02/25/2022 05:01:39 - INFO - codeparrot_training - Step 21183: {'lr': 0.00032751874835897316, 'samples': 10846208, 'steps': 21183, 'loss/train': 1.9777973890304565} 02/25/2022 05:01:43 - INFO - codeparrot_training - Step 21184: {'lr': 0.0003275031922025801, 'samples': 10846720, 'steps': 21184, 'loss/train': 1.5701086521148682} 02/25/2022 05:01:48 - INFO - codeparrot_training - Step 21185: {'lr': 0.00032748763571418805, 'samples': 10847232, 'steps': 21185, 'loss/train': 1.8671663999557495} 02/25/2022 05:01:52 - INFO - codeparrot_training - Step 21186: {'lr': 0.00032747207889386354, 'samples': 10847744, 'steps': 21186, 'loss/train': 1.347330093383789} 02/25/2022 05:01:59 - INFO - codeparrot_training - Step 21187: {'lr': 0.00032745652174167325, 'samples': 10848256, 'steps': 21187, 'loss/train': 1.8753383159637451} 02/25/2022 05:02:02 - INFO - codeparrot_training - Step 21188: {'lr': 0.00032744096425768376, 'samples': 10848768, 'steps': 21188, 'loss/train': 0.994610071182251} 02/25/2022 05:02:08 - INFO - codeparrot_training - Step 21189: {'lr': 0.00032742540644196185, 'samples': 10849280, 'steps': 21189, 'loss/train': 1.4592911005020142} 02/25/2022 05:02:11 - INFO - codeparrot_training - Step 21190: {'lr': 0.00032740984829457404, 'samples': 10849792, 'steps': 21190, 'loss/train': 1.7477768659591675} 02/25/2022 05:02:17 - INFO - codeparrot_training - Step 21191: {'lr': 0.00032739428981558706, 'samples': 10850304, 'steps': 21191, 'loss/train': 1.7833974361419678} 02/25/2022 05:02:20 - INFO - codeparrot_training - Step 21192: {'lr': 0.0003273787310050675, 'samples': 10850816, 'steps': 21192, 'loss/train': 2.370821237564087} 02/25/2022 05:02:26 - INFO - codeparrot_training - Step 21193: {'lr': 0.00032736317186308193, 'samples': 10851328, 'steps': 21193, 'loss/train': 2.1287100315093994} 02/25/2022 05:02:29 - INFO - codeparrot_training - Step 21194: {'lr': 0.00032734761238969724, 'samples': 10851840, 'steps': 21194, 'loss/train': 1.648510456085205} 02/25/2022 05:02:35 - INFO - codeparrot_training - Step 21195: {'lr': 0.00032733205258497994, 'samples': 10852352, 'steps': 21195, 'loss/train': 2.151329755783081} 02/25/2022 05:02:38 - INFO - codeparrot_training - Step 21196: {'lr': 0.0003273164924489966, 'samples': 10852864, 'steps': 21196, 'loss/train': 2.278211832046509} 02/25/2022 05:02:44 - INFO - codeparrot_training - Step 21197: {'lr': 0.000327300931981814, 'samples': 10853376, 'steps': 21197, 'loss/train': 1.2565044164657593} 02/25/2022 05:02:47 - INFO - codeparrot_training - Step 21198: {'lr': 0.0003272853711834987, 'samples': 10853888, 'steps': 21198, 'loss/train': 1.0751640796661377} 02/25/2022 05:02:53 - INFO - codeparrot_training - Step 21199: {'lr': 0.00032726981005411747, 'samples': 10854400, 'steps': 21199, 'loss/train': 2.5215718746185303} 02/25/2022 05:02:57 - INFO - codeparrot_training - Step 21200: {'lr': 0.00032725424859373687, 'samples': 10854912, 'steps': 21200, 'loss/train': 2.211833953857422} 02/25/2022 05:03:03 - INFO - codeparrot_training - Step 21201: {'lr': 0.0003272386868024236, 'samples': 10855424, 'steps': 21201, 'loss/train': 2.257880210876465} 02/25/2022 05:03:06 - INFO - codeparrot_training - Step 21202: {'lr': 0.00032722312468024434, 'samples': 10855936, 'steps': 21202, 'loss/train': 3.337616205215454} 02/25/2022 05:03:12 - INFO - codeparrot_training - Step 21203: {'lr': 0.00032720756222726576, 'samples': 10856448, 'steps': 21203, 'loss/train': 2.967454195022583} 02/25/2022 05:03:15 - INFO - codeparrot_training - Step 21204: {'lr': 0.0003271919994435545, 'samples': 10856960, 'steps': 21204, 'loss/train': 1.9554553031921387} 02/25/2022 05:03:21 - INFO - codeparrot_training - Step 21205: {'lr': 0.0003271764363291772, 'samples': 10857472, 'steps': 21205, 'loss/train': 1.900269627571106} 02/25/2022 05:03:24 - INFO - codeparrot_training - Step 21206: {'lr': 0.00032716087288420047, 'samples': 10857984, 'steps': 21206, 'loss/train': 1.0107386112213135} 02/25/2022 05:03:30 - INFO - codeparrot_training - Step 21207: {'lr': 0.0003271453091086912, 'samples': 10858496, 'steps': 21207, 'loss/train': 1.9128812551498413} 02/25/2022 05:03:33 - INFO - codeparrot_training - Step 21208: {'lr': 0.0003271297450027159, 'samples': 10859008, 'steps': 21208, 'loss/train': 2.0418624877929688} 02/25/2022 05:03:39 - INFO - codeparrot_training - Step 21209: {'lr': 0.0003271141805663412, 'samples': 10859520, 'steps': 21209, 'loss/train': 2.390519618988037} 02/25/2022 05:03:43 - INFO - codeparrot_training - Step 21210: {'lr': 0.00032709861579963384, 'samples': 10860032, 'steps': 21210, 'loss/train': 2.144798517227173} 02/25/2022 05:03:48 - INFO - codeparrot_training - Step 21211: {'lr': 0.00032708305070266053, 'samples': 10860544, 'steps': 21211, 'loss/train': 1.5633108615875244} 02/25/2022 05:03:52 - INFO - codeparrot_training - Step 21212: {'lr': 0.00032706748527548793, 'samples': 10861056, 'steps': 21212, 'loss/train': 2.3127710819244385} 02/25/2022 05:03:57 - INFO - codeparrot_training - Step 21213: {'lr': 0.0003270519195181826, 'samples': 10861568, 'steps': 21213, 'loss/train': 1.4027512073516846} 02/25/2022 05:04:01 - INFO - codeparrot_training - Step 21214: {'lr': 0.00032703635343081137, 'samples': 10862080, 'steps': 21214, 'loss/train': 2.843304395675659} 02/25/2022 05:04:06 - INFO - codeparrot_training - Step 21215: {'lr': 0.0003270207870134408, 'samples': 10862592, 'steps': 21215, 'loss/train': 1.4319182634353638} 02/25/2022 05:04:10 - INFO - codeparrot_training - Step 21216: {'lr': 0.00032700522026613785, 'samples': 10863104, 'steps': 21216, 'loss/train': 1.7666003704071045} 02/25/2022 05:04:15 - INFO - codeparrot_training - Step 21217: {'lr': 0.00032698965318896876, 'samples': 10863616, 'steps': 21217, 'loss/train': 2.059915542602539} 02/25/2022 05:04:19 - INFO - codeparrot_training - Step 21218: {'lr': 0.00032697408578200054, 'samples': 10864128, 'steps': 21218, 'loss/train': 2.2525713443756104} 02/25/2022 05:04:25 - INFO - codeparrot_training - Step 21219: {'lr': 0.00032695851804529977, 'samples': 10864640, 'steps': 21219, 'loss/train': 1.1876782178878784} 02/25/2022 05:04:28 - INFO - codeparrot_training - Step 21220: {'lr': 0.0003269429499789331, 'samples': 10865152, 'steps': 21220, 'loss/train': 0.6411092877388} 02/25/2022 05:04:34 - INFO - codeparrot_training - Step 21221: {'lr': 0.0003269273815829673, 'samples': 10865664, 'steps': 21221, 'loss/train': 2.443061113357544} 02/25/2022 05:04:37 - INFO - codeparrot_training - Step 21222: {'lr': 0.00032691181285746904, 'samples': 10866176, 'steps': 21222, 'loss/train': 2.3866283893585205} 02/25/2022 05:04:43 - INFO - codeparrot_training - Step 21223: {'lr': 0.000326896243802505, 'samples': 10866688, 'steps': 21223, 'loss/train': 0.8457977175712585} 02/25/2022 05:04:46 - INFO - codeparrot_training - Step 21224: {'lr': 0.0003268806744181418, 'samples': 10867200, 'steps': 21224, 'loss/train': 1.7025806903839111} 02/25/2022 05:04:52 - INFO - codeparrot_training - Step 21225: {'lr': 0.0003268651047044463, 'samples': 10867712, 'steps': 21225, 'loss/train': 1.9069551229476929} 02/25/2022 05:04:55 - INFO - codeparrot_training - Step 21226: {'lr': 0.00032684953466148505, 'samples': 10868224, 'steps': 21226, 'loss/train': 1.2650535106658936} 02/25/2022 05:05:01 - INFO - codeparrot_training - Step 21227: {'lr': 0.00032683396428932483, 'samples': 10868736, 'steps': 21227, 'loss/train': 2.2936995029449463} 02/25/2022 05:05:04 - INFO - codeparrot_training - Step 21228: {'lr': 0.0003268183935880322, 'samples': 10869248, 'steps': 21228, 'loss/train': 2.6150200366973877} 02/25/2022 05:05:11 - INFO - codeparrot_training - Step 21229: {'lr': 0.0003268028225576741, 'samples': 10869760, 'steps': 21229, 'loss/train': 1.4549059867858887} 02/25/2022 05:05:14 - INFO - codeparrot_training - Step 21230: {'lr': 0.00032678725119831696, 'samples': 10870272, 'steps': 21230, 'loss/train': 2.556772470474243} 02/25/2022 05:05:20 - INFO - codeparrot_training - Step 21231: {'lr': 0.0003267716795100278, 'samples': 10870784, 'steps': 21231, 'loss/train': 1.857347846031189} 02/25/2022 05:05:23 - INFO - codeparrot_training - Step 21232: {'lr': 0.000326756107492873, 'samples': 10871296, 'steps': 21232, 'loss/train': 1.5590025186538696} 02/25/2022 05:05:29 - INFO - codeparrot_training - Step 21233: {'lr': 0.00032674053514691946, 'samples': 10871808, 'steps': 21233, 'loss/train': 2.4627904891967773} 02/25/2022 05:05:32 - INFO - codeparrot_training - Step 21234: {'lr': 0.00032672496247223377, 'samples': 10872320, 'steps': 21234, 'loss/train': 0.885090708732605} 02/25/2022 05:05:38 - INFO - codeparrot_training - Step 21235: {'lr': 0.00032670938946888277, 'samples': 10872832, 'steps': 21235, 'loss/train': 3.0869076251983643} 02/25/2022 05:05:43 - INFO - codeparrot_training - Step 21236: {'lr': 0.00032669381613693307, 'samples': 10873344, 'steps': 21236, 'loss/train': 1.5500586032867432} 02/25/2022 05:05:47 - INFO - codeparrot_training - Step 21237: {'lr': 0.0003266782424764515, 'samples': 10873856, 'steps': 21237, 'loss/train': 1.2396222352981567} 02/25/2022 05:05:52 - INFO - codeparrot_training - Step 21238: {'lr': 0.0003266626684875046, 'samples': 10874368, 'steps': 21238, 'loss/train': 3.0672528743743896} 02/25/2022 05:05:56 - INFO - codeparrot_training - Step 21239: {'lr': 0.0003266470941701592, 'samples': 10874880, 'steps': 21239, 'loss/train': 1.9160960912704468} 02/25/2022 05:06:02 - INFO - codeparrot_training - Step 21240: {'lr': 0.00032663151952448194, 'samples': 10875392, 'steps': 21240, 'loss/train': 3.1761932373046875} 02/25/2022 05:06:05 - INFO - codeparrot_training - Step 21241: {'lr': 0.00032661594455053966, 'samples': 10875904, 'steps': 21241, 'loss/train': 2.4112744331359863} 02/25/2022 05:06:08 - INFO - codeparrot_training - Step 21242: {'lr': 0.0003266003692483989, 'samples': 10876416, 'steps': 21242, 'loss/train': 1.534583330154419} 02/25/2022 05:06:15 - INFO - codeparrot_training - Step 21243: {'lr': 0.0003265847936181266, 'samples': 10876928, 'steps': 21243, 'loss/train': 1.4044123888015747} 02/25/2022 05:06:19 - INFO - codeparrot_training - Step 21244: {'lr': 0.0003265692176597893, 'samples': 10877440, 'steps': 21244, 'loss/train': 1.7013826370239258} 02/25/2022 05:06:25 - INFO - codeparrot_training - Step 21245: {'lr': 0.0003265536413734538, 'samples': 10877952, 'steps': 21245, 'loss/train': 1.1996538639068604} 02/25/2022 05:06:28 - INFO - codeparrot_training - Step 21246: {'lr': 0.0003265380647591868, 'samples': 10878464, 'steps': 21246, 'loss/train': 2.779387950897217} 02/25/2022 05:06:34 - INFO - codeparrot_training - Step 21247: {'lr': 0.000326522487817055, 'samples': 10878976, 'steps': 21247, 'loss/train': 1.4230057001113892} 02/25/2022 05:06:37 - INFO - codeparrot_training - Step 21248: {'lr': 0.00032650691054712523, 'samples': 10879488, 'steps': 21248, 'loss/train': 3.7371480464935303} 02/25/2022 05:06:43 - INFO - codeparrot_training - Step 21249: {'lr': 0.0003264913329494641, 'samples': 10880000, 'steps': 21249, 'loss/train': 0.5596840977668762} 02/25/2022 05:06:46 - INFO - codeparrot_training - Step 21250: {'lr': 0.0003264757550241384, 'samples': 10880512, 'steps': 21250, 'loss/train': 0.5821067690849304} 02/25/2022 05:06:52 - INFO - codeparrot_training - Step 21251: {'lr': 0.00032646017677121484, 'samples': 10881024, 'steps': 21251, 'loss/train': 1.5946199893951416} 02/25/2022 05:06:55 - INFO - codeparrot_training - Step 21252: {'lr': 0.00032644459819076017, 'samples': 10881536, 'steps': 21252, 'loss/train': 2.3842592239379883} 02/25/2022 05:07:01 - INFO - codeparrot_training - Step 21253: {'lr': 0.00032642901928284115, 'samples': 10882048, 'steps': 21253, 'loss/train': 0.673209547996521} 02/25/2022 05:07:04 - INFO - codeparrot_training - Step 21254: {'lr': 0.0003264134400475244, 'samples': 10882560, 'steps': 21254, 'loss/train': 2.3186614513397217} 02/25/2022 05:07:11 - INFO - codeparrot_training - Step 21255: {'lr': 0.00032639786048487685, 'samples': 10883072, 'steps': 21255, 'loss/train': 2.7216572761535645} 02/25/2022 05:07:14 - INFO - codeparrot_training - Step 21256: {'lr': 0.000326382280594965, 'samples': 10883584, 'steps': 21256, 'loss/train': 2.0145411491394043} 02/25/2022 05:07:20 - INFO - codeparrot_training - Step 21257: {'lr': 0.00032636670037785583, 'samples': 10884096, 'steps': 21257, 'loss/train': 1.768389344215393} 02/25/2022 05:07:24 - INFO - codeparrot_training - Step 21258: {'lr': 0.00032635111983361586, 'samples': 10884608, 'steps': 21258, 'loss/train': 1.6506032943725586} 02/25/2022 05:07:29 - INFO - codeparrot_training - Step 21259: {'lr': 0.0003263355389623119, 'samples': 10885120, 'steps': 21259, 'loss/train': 1.3988884687423706} 02/25/2022 05:07:33 - INFO - codeparrot_training - Step 21260: {'lr': 0.0003263199577640109, 'samples': 10885632, 'steps': 21260, 'loss/train': 2.0579917430877686} 02/25/2022 05:07:38 - INFO - codeparrot_training - Step 21261: {'lr': 0.00032630437623877936, 'samples': 10886144, 'steps': 21261, 'loss/train': 2.7820398807525635} 02/25/2022 05:07:41 - INFO - codeparrot_training - Step 21262: {'lr': 0.00032628879438668414, 'samples': 10886656, 'steps': 21262, 'loss/train': 2.115137815475464} 02/25/2022 05:07:47 - INFO - codeparrot_training - Step 21263: {'lr': 0.00032627321220779184, 'samples': 10887168, 'steps': 21263, 'loss/train': 1.4926291704177856} 02/25/2022 05:07:50 - INFO - codeparrot_training - Step 21264: {'lr': 0.0003262576297021694, 'samples': 10887680, 'steps': 21264, 'loss/train': 2.3401951789855957} 02/25/2022 05:07:57 - INFO - codeparrot_training - Step 21265: {'lr': 0.00032624204686988343, 'samples': 10888192, 'steps': 21265, 'loss/train': 1.9107630252838135} 02/25/2022 05:08:00 - INFO - codeparrot_training - Step 21266: {'lr': 0.00032622646371100085, 'samples': 10888704, 'steps': 21266, 'loss/train': 1.1766200065612793} 02/25/2022 05:08:06 - INFO - codeparrot_training - Step 21267: {'lr': 0.00032621088022558823, 'samples': 10889216, 'steps': 21267, 'loss/train': 1.8639490604400635} 02/25/2022 05:08:09 - INFO - codeparrot_training - Step 21268: {'lr': 0.00032619529641371236, 'samples': 10889728, 'steps': 21268, 'loss/train': 2.769702434539795} 02/25/2022 05:08:15 - INFO - codeparrot_training - Step 21269: {'lr': 0.0003261797122754401, 'samples': 10890240, 'steps': 21269, 'loss/train': 2.4219839572906494} 02/25/2022 05:08:18 - INFO - codeparrot_training - Step 21270: {'lr': 0.00032616412781083813, 'samples': 10890752, 'steps': 21270, 'loss/train': 1.6763619184494019} 02/25/2022 05:08:24 - INFO - codeparrot_training - Step 21271: {'lr': 0.00032614854301997323, 'samples': 10891264, 'steps': 21271, 'loss/train': 2.125734806060791} 02/25/2022 05:08:27 - INFO - codeparrot_training - Step 21272: {'lr': 0.0003261329579029121, 'samples': 10891776, 'steps': 21272, 'loss/train': 2.938096523284912} 02/25/2022 05:08:33 - INFO - codeparrot_training - Step 21273: {'lr': 0.00032611737245972155, 'samples': 10892288, 'steps': 21273, 'loss/train': 2.756913661956787} 02/25/2022 05:08:36 - INFO - codeparrot_training - Step 21274: {'lr': 0.0003261017866904684, 'samples': 10892800, 'steps': 21274, 'loss/train': 0.14569512009620667} 02/25/2022 05:08:42 - INFO - codeparrot_training - Step 21275: {'lr': 0.0003260862005952193, 'samples': 10893312, 'steps': 21275, 'loss/train': 1.828144907951355} 02/25/2022 05:08:45 - INFO - codeparrot_training - Step 21276: {'lr': 0.00032607061417404113, 'samples': 10893824, 'steps': 21276, 'loss/train': 1.659318208694458} 02/25/2022 05:08:51 - INFO - codeparrot_training - Step 21277: {'lr': 0.0003260550274270007, 'samples': 10894336, 'steps': 21277, 'loss/train': 2.3319449424743652} 02/25/2022 05:08:54 - INFO - codeparrot_training - Step 21278: {'lr': 0.0003260394403541645, 'samples': 10894848, 'steps': 21278, 'loss/train': 1.7328227758407593} 02/25/2022 05:09:01 - INFO - codeparrot_training - Step 21279: {'lr': 0.00032602385295559953, 'samples': 10895360, 'steps': 21279, 'loss/train': 1.6713923215866089} 02/25/2022 05:09:04 - INFO - codeparrot_training - Step 21280: {'lr': 0.00032600826523137254, 'samples': 10895872, 'steps': 21280, 'loss/train': 2.395864963531494} 02/25/2022 05:09:10 - INFO - codeparrot_training - Step 21281: {'lr': 0.00032599267718155036, 'samples': 10896384, 'steps': 21281, 'loss/train': 1.2828642129898071} 02/25/2022 05:09:13 - INFO - codeparrot_training - Step 21282: {'lr': 0.0003259770888061995, 'samples': 10896896, 'steps': 21282, 'loss/train': 1.498377799987793} 02/25/2022 05:09:18 - INFO - codeparrot_training - Step 21283: {'lr': 0.0003259615001053871, 'samples': 10897408, 'steps': 21283, 'loss/train': 2.6201424598693848} 02/25/2022 05:09:22 - INFO - codeparrot_training - Step 21284: {'lr': 0.0003259459110791797, 'samples': 10897920, 'steps': 21284, 'loss/train': 2.3451859951019287} 02/25/2022 05:09:28 - INFO - codeparrot_training - Step 21285: {'lr': 0.0003259303217276441, 'samples': 10898432, 'steps': 21285, 'loss/train': 2.6213128566741943} 02/25/2022 05:09:31 - INFO - codeparrot_training - Step 21286: {'lr': 0.0003259147320508471, 'samples': 10898944, 'steps': 21286, 'loss/train': 2.1762027740478516} 02/25/2022 05:09:37 - INFO - codeparrot_training - Step 21287: {'lr': 0.0003258991420488555, 'samples': 10899456, 'steps': 21287, 'loss/train': 2.7658660411834717} 02/25/2022 05:09:40 - INFO - codeparrot_training - Step 21288: {'lr': 0.0003258835517217361, 'samples': 10899968, 'steps': 21288, 'loss/train': 1.5106842517852783} 02/25/2022 05:09:46 - INFO - codeparrot_training - Step 21289: {'lr': 0.0003258679610695556, 'samples': 10900480, 'steps': 21289, 'loss/train': 2.2157793045043945} 02/25/2022 05:09:49 - INFO - codeparrot_training - Step 21290: {'lr': 0.000325852370092381, 'samples': 10900992, 'steps': 21290, 'loss/train': 1.886971116065979} 02/25/2022 05:09:55 - INFO - codeparrot_training - Step 21291: {'lr': 0.00032583677879027877, 'samples': 10901504, 'steps': 21291, 'loss/train': 0.7840352654457092} 02/25/2022 05:09:59 - INFO - codeparrot_training - Step 21292: {'lr': 0.00032582118716331594, 'samples': 10902016, 'steps': 21292, 'loss/train': 1.838408350944519} 02/25/2022 05:10:04 - INFO - codeparrot_training - Step 21293: {'lr': 0.00032580559521155916, 'samples': 10902528, 'steps': 21293, 'loss/train': 2.238321304321289} 02/25/2022 05:10:08 - INFO - codeparrot_training - Step 21294: {'lr': 0.00032579000293507537, 'samples': 10903040, 'steps': 21294, 'loss/train': 1.330949068069458} 02/25/2022 05:10:13 - INFO - codeparrot_training - Step 21295: {'lr': 0.0003257744103339312, 'samples': 10903552, 'steps': 21295, 'loss/train': 1.9749799966812134} 02/25/2022 05:10:17 - INFO - codeparrot_training - Step 21296: {'lr': 0.00032575881740819353, 'samples': 10904064, 'steps': 21296, 'loss/train': 2.121551275253296} 02/25/2022 05:10:22 - INFO - codeparrot_training - Step 21297: {'lr': 0.0003257432241579291, 'samples': 10904576, 'steps': 21297, 'loss/train': 1.8827018737792969} 02/25/2022 05:10:26 - INFO - codeparrot_training - Step 21298: {'lr': 0.00032572763058320484, 'samples': 10905088, 'steps': 21298, 'loss/train': 2.8160476684570312} 02/25/2022 05:10:31 - INFO - codeparrot_training - Step 21299: {'lr': 0.00032571203668408744, 'samples': 10905600, 'steps': 21299, 'loss/train': 2.4017887115478516} 02/25/2022 05:10:37 - INFO - codeparrot_training - Step 21300: {'lr': 0.0003256964424606437, 'samples': 10906112, 'steps': 21300, 'loss/train': 1.7928662300109863} 02/25/2022 05:10:41 - INFO - codeparrot_training - Step 21301: {'lr': 0.0003256808479129404, 'samples': 10906624, 'steps': 21301, 'loss/train': 2.3589839935302734} 02/25/2022 05:10:46 - INFO - codeparrot_training - Step 21302: {'lr': 0.0003256652530410444, 'samples': 10907136, 'steps': 21302, 'loss/train': 2.0396180152893066} 02/25/2022 05:10:50 - INFO - codeparrot_training - Step 21303: {'lr': 0.00032564965784502255, 'samples': 10907648, 'steps': 21303, 'loss/train': 3.246683120727539} 02/25/2022 05:10:56 - INFO - codeparrot_training - Step 21304: {'lr': 0.0003256340623249415, 'samples': 10908160, 'steps': 21304, 'loss/train': 2.8570749759674072} 02/25/2022 05:10:59 - INFO - codeparrot_training - Step 21305: {'lr': 0.00032561846648086816, 'samples': 10908672, 'steps': 21305, 'loss/train': 3.0938119888305664} 02/25/2022 05:11:02 - INFO - codeparrot_training - Step 21306: {'lr': 0.0003256028703128693, 'samples': 10909184, 'steps': 21306, 'loss/train': 2.3098132610321045} 02/25/2022 05:11:08 - INFO - codeparrot_training - Step 21307: {'lr': 0.0003255872738210118, 'samples': 10909696, 'steps': 21307, 'loss/train': 1.508814811706543} 02/25/2022 05:11:14 - INFO - codeparrot_training - Step 21308: {'lr': 0.0003255716770053624, 'samples': 10910208, 'steps': 21308, 'loss/train': 2.027381420135498} 02/25/2022 05:11:18 - INFO - codeparrot_training - Step 21309: {'lr': 0.0003255560798659879, 'samples': 10910720, 'steps': 21309, 'loss/train': 2.1558380126953125} 02/25/2022 05:11:21 - INFO - codeparrot_training - Step 21310: {'lr': 0.0003255404824029552, 'samples': 10911232, 'steps': 21310, 'loss/train': 2.34118914604187} 02/25/2022 05:11:28 - INFO - codeparrot_training - Step 21311: {'lr': 0.00032552488461633103, 'samples': 10911744, 'steps': 21311, 'loss/train': 2.317178726196289} 02/25/2022 05:11:32 - INFO - codeparrot_training - Step 21312: {'lr': 0.00032550928650618225, 'samples': 10912256, 'steps': 21312, 'loss/train': 2.280752182006836} 02/25/2022 05:11:37 - INFO - codeparrot_training - Step 21313: {'lr': 0.0003254936880725757, 'samples': 10912768, 'steps': 21313, 'loss/train': 1.3541532754898071} 02/25/2022 05:11:40 - INFO - codeparrot_training - Step 21314: {'lr': 0.0003254780893155781, 'samples': 10913280, 'steps': 21314, 'loss/train': 1.4611402750015259} 02/25/2022 05:11:46 - INFO - codeparrot_training - Step 21315: {'lr': 0.00032546249023525636, 'samples': 10913792, 'steps': 21315, 'loss/train': 1.0965029001235962} 02/25/2022 05:11:50 - INFO - codeparrot_training - Step 21316: {'lr': 0.00032544689083167727, 'samples': 10914304, 'steps': 21316, 'loss/train': 2.3084423542022705} 02/25/2022 05:11:55 - INFO - codeparrot_training - Step 21317: {'lr': 0.00032543129110490764, 'samples': 10914816, 'steps': 21317, 'loss/train': 2.16798734664917} 02/25/2022 05:11:58 - INFO - codeparrot_training - Step 21318: {'lr': 0.00032541569105501433, 'samples': 10915328, 'steps': 21318, 'loss/train': 0.34300774335861206} 02/25/2022 05:12:04 - INFO - codeparrot_training - Step 21319: {'lr': 0.0003254000906820642, 'samples': 10915840, 'steps': 21319, 'loss/train': 1.1201236248016357} 02/25/2022 05:12:07 - INFO - codeparrot_training - Step 21320: {'lr': 0.0003253844899861239, 'samples': 10916352, 'steps': 21320, 'loss/train': 0.29911482334136963} 02/25/2022 05:12:13 - INFO - codeparrot_training - Step 21321: {'lr': 0.00032536888896726046, 'samples': 10916864, 'steps': 21321, 'loss/train': 2.0089926719665527} 02/25/2022 05:12:16 - INFO - codeparrot_training - Step 21322: {'lr': 0.00032535328762554064, 'samples': 10917376, 'steps': 21322, 'loss/train': 1.7739598751068115} 02/25/2022 05:12:22 - INFO - codeparrot_training - Step 21323: {'lr': 0.00032533768596103123, 'samples': 10917888, 'steps': 21323, 'loss/train': 1.973223090171814} 02/25/2022 05:12:25 - INFO - codeparrot_training - Step 21324: {'lr': 0.0003253220839737991, 'samples': 10918400, 'steps': 21324, 'loss/train': 2.3111445903778076} 02/25/2022 05:12:32 - INFO - codeparrot_training - Step 21325: {'lr': 0.00032530648166391115, 'samples': 10918912, 'steps': 21325, 'loss/train': 2.7867369651794434} 02/25/2022 05:12:35 - INFO - codeparrot_training - Step 21326: {'lr': 0.00032529087903143407, 'samples': 10919424, 'steps': 21326, 'loss/train': 1.4398127794265747} 02/25/2022 05:12:41 - INFO - codeparrot_training - Step 21327: {'lr': 0.00032527527607643475, 'samples': 10919936, 'steps': 21327, 'loss/train': 1.3045326471328735} 02/25/2022 05:12:44 - INFO - codeparrot_training - Step 21328: {'lr': 0.00032525967279898015, 'samples': 10920448, 'steps': 21328, 'loss/train': 1.9129797220230103} 02/25/2022 05:12:50 - INFO - codeparrot_training - Step 21329: {'lr': 0.0003252440691991369, 'samples': 10920960, 'steps': 21329, 'loss/train': 1.9628323316574097} 02/25/2022 05:12:53 - INFO - codeparrot_training - Step 21330: {'lr': 0.000325228465276972, 'samples': 10921472, 'steps': 21330, 'loss/train': 0.9939577579498291} 02/25/2022 05:12:59 - INFO - codeparrot_training - Step 21331: {'lr': 0.00032521286103255224, 'samples': 10921984, 'steps': 21331, 'loss/train': 2.592625617980957} 02/25/2022 05:13:02 - INFO - codeparrot_training - Step 21332: {'lr': 0.0003251972564659445, 'samples': 10922496, 'steps': 21332, 'loss/train': 0.34064191579818726} 02/25/2022 05:13:08 - INFO - codeparrot_training - Step 21333: {'lr': 0.00032518165157721554, 'samples': 10923008, 'steps': 21333, 'loss/train': 2.357318639755249} 02/25/2022 05:13:12 - INFO - codeparrot_training - Step 21334: {'lr': 0.00032516604636643234, 'samples': 10923520, 'steps': 21334, 'loss/train': 1.158659815788269} 02/25/2022 05:13:15 - INFO - codeparrot_training - Step 21335: {'lr': 0.00032515044083366153, 'samples': 10924032, 'steps': 21335, 'loss/train': 1.301227331161499} 02/25/2022 05:13:20 - INFO - codeparrot_training - Step 21336: {'lr': 0.0003251348349789702, 'samples': 10924544, 'steps': 21336, 'loss/train': 1.601779580116272} 02/25/2022 05:13:24 - INFO - codeparrot_training - Step 21337: {'lr': 0.00032511922880242505, 'samples': 10925056, 'steps': 21337, 'loss/train': 1.9722219705581665} 02/25/2022 05:13:30 - INFO - codeparrot_training - Step 21338: {'lr': 0.00032510362230409295, 'samples': 10925568, 'steps': 21338, 'loss/train': 1.6293399333953857} 02/25/2022 05:13:34 - INFO - codeparrot_training - Step 21339: {'lr': 0.0003250880154840408, 'samples': 10926080, 'steps': 21339, 'loss/train': 2.160531520843506} 02/25/2022 05:13:39 - INFO - codeparrot_training - Step 21340: {'lr': 0.0003250724083423355, 'samples': 10926592, 'steps': 21340, 'loss/train': 1.19654381275177} 02/25/2022 05:13:43 - INFO - codeparrot_training - Step 21341: {'lr': 0.00032505680087904375, 'samples': 10927104, 'steps': 21341, 'loss/train': 0.9483667016029358} 02/25/2022 05:13:49 - INFO - codeparrot_training - Step 21342: {'lr': 0.0003250411930942326, 'samples': 10927616, 'steps': 21342, 'loss/train': 2.331989049911499} 02/25/2022 05:13:52 - INFO - codeparrot_training - Step 21343: {'lr': 0.00032502558498796876, 'samples': 10928128, 'steps': 21343, 'loss/train': 2.6511056423187256} 02/25/2022 05:13:58 - INFO - codeparrot_training - Step 21344: {'lr': 0.00032500997656031907, 'samples': 10928640, 'steps': 21344, 'loss/train': 2.0560503005981445} 02/25/2022 05:14:01 - INFO - codeparrot_training - Step 21345: {'lr': 0.0003249943678113505, 'samples': 10929152, 'steps': 21345, 'loss/train': 1.9853085279464722} 02/25/2022 05:14:07 - INFO - codeparrot_training - Step 21346: {'lr': 0.00032497875874112995, 'samples': 10929664, 'steps': 21346, 'loss/train': 1.886953592300415} 02/25/2022 05:14:10 - INFO - codeparrot_training - Step 21347: {'lr': 0.0003249631493497241, 'samples': 10930176, 'steps': 21347, 'loss/train': 2.412015676498413} 02/25/2022 05:14:16 - INFO - codeparrot_training - Step 21348: {'lr': 0.0003249475396371999, 'samples': 10930688, 'steps': 21348, 'loss/train': 1.7576769590377808} 02/25/2022 05:14:20 - INFO - codeparrot_training - Step 21349: {'lr': 0.00032493192960362437, 'samples': 10931200, 'steps': 21349, 'loss/train': 1.962602972984314} 02/25/2022 05:14:25 - INFO - codeparrot_training - Step 21350: {'lr': 0.00032491631924906416, 'samples': 10931712, 'steps': 21350, 'loss/train': 1.3196920156478882} 02/25/2022 05:14:29 - INFO - codeparrot_training - Step 21351: {'lr': 0.0003249007085735863, 'samples': 10932224, 'steps': 21351, 'loss/train': 2.489952325820923} 02/25/2022 05:14:34 - INFO - codeparrot_training - Step 21352: {'lr': 0.0003248850975772575, 'samples': 10932736, 'steps': 21352, 'loss/train': 1.9956902265548706} 02/25/2022 05:14:38 - INFO - codeparrot_training - Step 21353: {'lr': 0.00032486948626014476, 'samples': 10933248, 'steps': 21353, 'loss/train': 0.5638812780380249} 02/25/2022 05:14:44 - INFO - codeparrot_training - Step 21354: {'lr': 0.00032485387462231484, 'samples': 10933760, 'steps': 21354, 'loss/train': 1.377614140510559} 02/25/2022 05:14:47 - INFO - codeparrot_training - Step 21355: {'lr': 0.0003248382626638348, 'samples': 10934272, 'steps': 21355, 'loss/train': 1.6708141565322876} 02/25/2022 05:14:52 - INFO - codeparrot_training - Step 21356: {'lr': 0.0003248226503847714, 'samples': 10934784, 'steps': 21356, 'loss/train': 1.6795576810836792} 02/25/2022 05:14:56 - INFO - codeparrot_training - Step 21357: {'lr': 0.00032480703778519146, 'samples': 10935296, 'steps': 21357, 'loss/train': 2.426008939743042} 02/25/2022 05:15:02 - INFO - codeparrot_training - Step 21358: {'lr': 0.00032479142486516193, 'samples': 10935808, 'steps': 21358, 'loss/train': 1.647079348564148} 02/25/2022 05:15:06 - INFO - codeparrot_training - Step 21359: {'lr': 0.00032477581162474974, 'samples': 10936320, 'steps': 21359, 'loss/train': 1.9271972179412842} 02/25/2022 05:15:11 - INFO - codeparrot_training - Step 21360: {'lr': 0.0003247601980640217, 'samples': 10936832, 'steps': 21360, 'loss/train': 0.4454154670238495} 02/25/2022 05:15:15 - INFO - codeparrot_training - Step 21361: {'lr': 0.0003247445841830446, 'samples': 10937344, 'steps': 21361, 'loss/train': 2.4724366664886475} 02/25/2022 05:15:20 - INFO - codeparrot_training - Step 21362: {'lr': 0.0003247289699818856, 'samples': 10937856, 'steps': 21362, 'loss/train': 2.876556634902954} 02/25/2022 05:15:24 - INFO - codeparrot_training - Step 21363: {'lr': 0.0003247133554606113, 'samples': 10938368, 'steps': 21363, 'loss/train': 1.9389662742614746} 02/25/2022 05:15:29 - INFO - codeparrot_training - Step 21364: {'lr': 0.0003246977406192888, 'samples': 10938880, 'steps': 21364, 'loss/train': 1.802810549736023} 02/25/2022 05:15:33 - INFO - codeparrot_training - Step 21365: {'lr': 0.00032468212545798484, 'samples': 10939392, 'steps': 21365, 'loss/train': 1.1155353784561157} 02/25/2022 05:15:38 - INFO - codeparrot_training - Step 21366: {'lr': 0.0003246665099767664, 'samples': 10939904, 'steps': 21366, 'loss/train': 2.016072988510132} 02/25/2022 05:15:42 - INFO - codeparrot_training - Step 21367: {'lr': 0.0003246508941757004, 'samples': 10940416, 'steps': 21367, 'loss/train': 1.5841889381408691} 02/25/2022 05:15:47 - INFO - codeparrot_training - Step 21368: {'lr': 0.0003246352780548536, 'samples': 10940928, 'steps': 21368, 'loss/train': 1.494071125984192} 02/25/2022 05:15:51 - INFO - codeparrot_training - Step 21369: {'lr': 0.0003246196616142929, 'samples': 10941440, 'steps': 21369, 'loss/train': 1.6460036039352417} 02/25/2022 05:15:56 - INFO - codeparrot_training - Step 21370: {'lr': 0.0003246040448540854, 'samples': 10941952, 'steps': 21370, 'loss/train': 0.8125658631324768} 02/25/2022 05:16:00 - INFO - codeparrot_training - Step 21371: {'lr': 0.00032458842777429776, 'samples': 10942464, 'steps': 21371, 'loss/train': 1.5591201782226562} 02/25/2022 05:16:05 - INFO - codeparrot_training - Step 21372: {'lr': 0.00032457281037499706, 'samples': 10942976, 'steps': 21372, 'loss/train': 1.009738564491272} 02/25/2022 05:16:09 - INFO - codeparrot_training - Step 21373: {'lr': 0.0003245571926562501, 'samples': 10943488, 'steps': 21373, 'loss/train': 2.5773777961730957} 02/25/2022 05:16:15 - INFO - codeparrot_training - Step 21374: {'lr': 0.0003245415746181237, 'samples': 10944000, 'steps': 21374, 'loss/train': 1.5520761013031006} 02/25/2022 05:16:18 - INFO - codeparrot_training - Step 21375: {'lr': 0.00032452595626068504, 'samples': 10944512, 'steps': 21375, 'loss/train': 2.8511366844177246} 02/25/2022 05:16:24 - INFO - codeparrot_training - Step 21376: {'lr': 0.0003245103375840007, 'samples': 10945024, 'steps': 21376, 'loss/train': 2.3014512062072754} 02/25/2022 05:16:27 - INFO - codeparrot_training - Step 21377: {'lr': 0.0003244947185881378, 'samples': 10945536, 'steps': 21377, 'loss/train': 1.6308921575546265} 02/25/2022 05:16:33 - INFO - codeparrot_training - Step 21378: {'lr': 0.00032447909927316317, 'samples': 10946048, 'steps': 21378, 'loss/train': 1.7439149618148804} 02/25/2022 05:16:36 - INFO - codeparrot_training - Step 21379: {'lr': 0.00032446347963914376, 'samples': 10946560, 'steps': 21379, 'loss/train': 0.8212968707084656} 02/25/2022 05:16:42 - INFO - codeparrot_training - Step 21380: {'lr': 0.0003244478596861464, 'samples': 10947072, 'steps': 21380, 'loss/train': 2.1549930572509766} 02/25/2022 05:16:46 - INFO - codeparrot_training - Step 21381: {'lr': 0.000324432239414238, 'samples': 10947584, 'steps': 21381, 'loss/train': 2.2635746002197266} 02/25/2022 05:16:51 - INFO - codeparrot_training - Step 21382: {'lr': 0.0003244166188234856, 'samples': 10948096, 'steps': 21382, 'loss/train': 0.44461098313331604} 02/25/2022 05:16:55 - INFO - codeparrot_training - Step 21383: {'lr': 0.000324400997913956, 'samples': 10948608, 'steps': 21383, 'loss/train': 1.6025444269180298} 02/25/2022 05:17:01 - INFO - codeparrot_training - Step 21384: {'lr': 0.0003243853766857162, 'samples': 10949120, 'steps': 21384, 'loss/train': 2.54162859916687} 02/25/2022 05:17:05 - INFO - codeparrot_training - Step 21385: {'lr': 0.000324369755138833, 'samples': 10949632, 'steps': 21385, 'loss/train': 1.9391682147979736} 02/25/2022 05:17:10 - INFO - codeparrot_training - Step 21386: {'lr': 0.0003243541332733734, 'samples': 10950144, 'steps': 21386, 'loss/train': 2.18705677986145} 02/25/2022 05:17:14 - INFO - codeparrot_training - Step 21387: {'lr': 0.00032433851108940433, 'samples': 10950656, 'steps': 21387, 'loss/train': 3.22505259513855} 02/25/2022 05:17:19 - INFO - codeparrot_training - Step 21388: {'lr': 0.0003243228885869927, 'samples': 10951168, 'steps': 21388, 'loss/train': 3.6134586334228516} 02/25/2022 05:17:23 - INFO - codeparrot_training - Step 21389: {'lr': 0.0003243072657662054, 'samples': 10951680, 'steps': 21389, 'loss/train': 1.7234821319580078} 02/25/2022 05:17:28 - INFO - codeparrot_training - Step 21390: {'lr': 0.00032429164262710934, 'samples': 10952192, 'steps': 21390, 'loss/train': 2.198101758956909} 02/25/2022 05:17:32 - INFO - codeparrot_training - Step 21391: {'lr': 0.0003242760191697714, 'samples': 10952704, 'steps': 21391, 'loss/train': 0.9394028186798096} 02/25/2022 05:17:37 - INFO - codeparrot_training - Step 21392: {'lr': 0.0003242603953942587, 'samples': 10953216, 'steps': 21392, 'loss/train': 1.871528148651123} 02/25/2022 05:17:41 - INFO - codeparrot_training - Step 21393: {'lr': 0.00032424477130063806, 'samples': 10953728, 'steps': 21393, 'loss/train': 1.598361849784851} 02/25/2022 05:17:48 - INFO - codeparrot_training - Step 21394: {'lr': 0.0003242291468889763, 'samples': 10954240, 'steps': 21394, 'loss/train': 2.190145492553711} 02/25/2022 05:17:51 - INFO - codeparrot_training - Step 21395: {'lr': 0.0003242135221593405, 'samples': 10954752, 'steps': 21395, 'loss/train': 1.4833991527557373} 02/25/2022 05:17:57 - INFO - codeparrot_training - Step 21396: {'lr': 0.0003241978971117976, 'samples': 10955264, 'steps': 21396, 'loss/train': 1.6480028629302979} 02/25/2022 05:18:00 - INFO - codeparrot_training - Step 21397: {'lr': 0.0003241822717464144, 'samples': 10955776, 'steps': 21397, 'loss/train': 1.575247883796692} 02/25/2022 05:18:05 - INFO - codeparrot_training - Step 21398: {'lr': 0.0003241666460632579, 'samples': 10956288, 'steps': 21398, 'loss/train': 1.7289868593215942} 02/25/2022 05:18:09 - INFO - codeparrot_training - Step 21399: {'lr': 0.00032415102006239506, 'samples': 10956800, 'steps': 21399, 'loss/train': 1.049124836921692} 02/25/2022 05:18:15 - INFO - codeparrot_training - Step 21400: {'lr': 0.0003241353937438927, 'samples': 10957312, 'steps': 21400, 'loss/train': 2.3949666023254395} 02/25/2022 05:18:18 - INFO - codeparrot_training - Step 21401: {'lr': 0.000324119767107818, 'samples': 10957824, 'steps': 21401, 'loss/train': 1.219030737876892} 02/25/2022 05:18:23 - INFO - codeparrot_training - Step 21402: {'lr': 0.0003241041401542377, 'samples': 10958336, 'steps': 21402, 'loss/train': 1.8786659240722656} 02/25/2022 05:18:27 - INFO - codeparrot_training - Step 21403: {'lr': 0.0003240885128832188, 'samples': 10958848, 'steps': 21403, 'loss/train': 2.7543208599090576} 02/25/2022 05:18:33 - INFO - codeparrot_training - Step 21404: {'lr': 0.0003240728852948281, 'samples': 10959360, 'steps': 21404, 'loss/train': 1.6549605131149292} 02/25/2022 05:18:37 - INFO - codeparrot_training - Step 21405: {'lr': 0.00032405725738913284, 'samples': 10959872, 'steps': 21405, 'loss/train': 2.694563865661621} 02/25/2022 05:18:42 - INFO - codeparrot_training - Step 21406: {'lr': 0.0003240416291661998, 'samples': 10960384, 'steps': 21406, 'loss/train': 1.2894922494888306} 02/25/2022 05:18:46 - INFO - codeparrot_training - Step 21407: {'lr': 0.0003240260006260959, 'samples': 10960896, 'steps': 21407, 'loss/train': 2.1315720081329346} 02/25/2022 05:18:51 - INFO - codeparrot_training - Step 21408: {'lr': 0.0003240103717688881, 'samples': 10961408, 'steps': 21408, 'loss/train': 0.43893662095069885} 02/25/2022 05:18:55 - INFO - codeparrot_training - Step 21409: {'lr': 0.00032399474259464336, 'samples': 10961920, 'steps': 21409, 'loss/train': 1.3445855379104614} 02/25/2022 05:19:00 - INFO - codeparrot_training - Step 21410: {'lr': 0.0003239791131034287, 'samples': 10962432, 'steps': 21410, 'loss/train': 1.8981529474258423} 02/25/2022 05:19:04 - INFO - codeparrot_training - Step 21411: {'lr': 0.00032396348329531097, 'samples': 10962944, 'steps': 21411, 'loss/train': 2.38852596282959} 02/25/2022 05:19:09 - INFO - codeparrot_training - Step 21412: {'lr': 0.0003239478531703571, 'samples': 10963456, 'steps': 21412, 'loss/train': 2.345404863357544} 02/25/2022 05:19:13 - INFO - codeparrot_training - Step 21413: {'lr': 0.0003239322227286343, 'samples': 10963968, 'steps': 21413, 'loss/train': 2.1669015884399414} 02/25/2022 05:19:18 - INFO - codeparrot_training - Step 21414: {'lr': 0.0003239165919702092, 'samples': 10964480, 'steps': 21414, 'loss/train': 2.0201916694641113} 02/25/2022 05:19:22 - INFO - codeparrot_training - Step 21415: {'lr': 0.0003239009608951489, 'samples': 10964992, 'steps': 21415, 'loss/train': 1.5376311540603638} 02/25/2022 05:19:27 - INFO - codeparrot_training - Step 21416: {'lr': 0.0003238853295035203, 'samples': 10965504, 'steps': 21416, 'loss/train': 2.01336932182312} 02/25/2022 05:19:31 - INFO - codeparrot_training - Step 21417: {'lr': 0.0003238696977953905, 'samples': 10966016, 'steps': 21417, 'loss/train': 2.6084911823272705} 02/25/2022 05:19:36 - INFO - codeparrot_training - Step 21418: {'lr': 0.0003238540657708263, 'samples': 10966528, 'steps': 21418, 'loss/train': 1.6179522275924683} 02/25/2022 05:19:40 - INFO - codeparrot_training - Step 21419: {'lr': 0.0003238384334298948, 'samples': 10967040, 'steps': 21419, 'loss/train': 3.299560308456421} 02/25/2022 05:19:46 - INFO - codeparrot_training - Step 21420: {'lr': 0.0003238228007726628, 'samples': 10967552, 'steps': 21420, 'loss/train': 1.4794039726257324} 02/25/2022 05:19:50 - INFO - codeparrot_training - Step 21421: {'lr': 0.00032380716779919745, 'samples': 10968064, 'steps': 21421, 'loss/train': 1.9403265714645386} 02/25/2022 05:19:55 - INFO - codeparrot_training - Step 21422: {'lr': 0.00032379153450956555, 'samples': 10968576, 'steps': 21422, 'loss/train': 1.8774206638336182} 02/25/2022 05:19:59 - INFO - codeparrot_training - Step 21423: {'lr': 0.0003237759009038342, 'samples': 10969088, 'steps': 21423, 'loss/train': 1.940247654914856} 02/25/2022 05:20:04 - INFO - codeparrot_training - Step 21424: {'lr': 0.0003237602669820703, 'samples': 10969600, 'steps': 21424, 'loss/train': 1.4444103240966797} 02/25/2022 05:20:08 - INFO - codeparrot_training - Step 21425: {'lr': 0.00032374463274434097, 'samples': 10970112, 'steps': 21425, 'loss/train': 2.090341091156006} 02/25/2022 05:20:13 - INFO - codeparrot_training - Step 21426: {'lr': 0.0003237289981907129, 'samples': 10970624, 'steps': 21426, 'loss/train': 1.3708947896957397} 02/25/2022 05:20:17 - INFO - codeparrot_training - Step 21427: {'lr': 0.00032371336332125323, 'samples': 10971136, 'steps': 21427, 'loss/train': 1.4762861728668213} 02/25/2022 05:20:22 - INFO - codeparrot_training - Step 21428: {'lr': 0.0003236977281360289, 'samples': 10971648, 'steps': 21428, 'loss/train': 2.433065891265869} 02/25/2022 05:20:26 - INFO - codeparrot_training - Step 21429: {'lr': 0.00032368209263510694, 'samples': 10972160, 'steps': 21429, 'loss/train': 0.7332336902618408} 02/25/2022 05:20:33 - INFO - codeparrot_training - Step 21430: {'lr': 0.00032366645681855435, 'samples': 10972672, 'steps': 21430, 'loss/train': 2.2203409671783447} 02/25/2022 05:20:36 - INFO - codeparrot_training - Step 21431: {'lr': 0.0003236508206864379, 'samples': 10973184, 'steps': 21431, 'loss/train': 1.782835841178894} 02/25/2022 05:20:42 - INFO - codeparrot_training - Step 21432: {'lr': 0.0003236351842388249, 'samples': 10973696, 'steps': 21432, 'loss/train': 2.293973684310913} 02/25/2022 05:20:45 - INFO - codeparrot_training - Step 21433: {'lr': 0.00032361954747578203, 'samples': 10974208, 'steps': 21433, 'loss/train': 1.5398344993591309} 02/25/2022 05:20:51 - INFO - codeparrot_training - Step 21434: {'lr': 0.00032360391039737646, 'samples': 10974720, 'steps': 21434, 'loss/train': 2.6571176052093506} 02/25/2022 05:20:54 - INFO - codeparrot_training - Step 21435: {'lr': 0.00032358827300367504, 'samples': 10975232, 'steps': 21435, 'loss/train': 1.572434902191162} 02/25/2022 05:21:00 - INFO - codeparrot_training - Step 21436: {'lr': 0.0003235726352947449, 'samples': 10975744, 'steps': 21436, 'loss/train': 1.5795124769210815} 02/25/2022 05:21:03 - INFO - codeparrot_training - Step 21437: {'lr': 0.0003235569972706529, 'samples': 10976256, 'steps': 21437, 'loss/train': 1.3008886575698853} 02/25/2022 05:21:09 - INFO - codeparrot_training - Step 21438: {'lr': 0.00032354135893146614, 'samples': 10976768, 'steps': 21438, 'loss/train': 1.5897101163864136} 02/25/2022 05:21:12 - INFO - codeparrot_training - Step 21439: {'lr': 0.0003235257202772515, 'samples': 10977280, 'steps': 21439, 'loss/train': 1.3608970642089844} 02/25/2022 05:21:19 - INFO - codeparrot_training - Step 21440: {'lr': 0.000323510081308076, 'samples': 10977792, 'steps': 21440, 'loss/train': 2.2303550243377686} 02/25/2022 05:21:22 - INFO - codeparrot_training - Step 21441: {'lr': 0.00032349444202400666, 'samples': 10978304, 'steps': 21441, 'loss/train': 2.004150390625} 02/25/2022 05:21:28 - INFO - codeparrot_training - Step 21442: {'lr': 0.0003234788024251105, 'samples': 10978816, 'steps': 21442, 'loss/train': 1.6814168691635132} 02/25/2022 05:21:33 - INFO - codeparrot_training - Step 21443: {'lr': 0.00032346316251145445, 'samples': 10979328, 'steps': 21443, 'loss/train': 1.9356404542922974} 02/25/2022 05:21:37 - INFO - codeparrot_training - Step 21444: {'lr': 0.0003234475222831056, 'samples': 10979840, 'steps': 21444, 'loss/train': 1.8436837196350098} 02/25/2022 05:21:42 - INFO - codeparrot_training - Step 21445: {'lr': 0.0003234318817401309, 'samples': 10980352, 'steps': 21445, 'loss/train': 1.9931660890579224} 02/25/2022 05:21:46 - INFO - codeparrot_training - Step 21446: {'lr': 0.00032341624088259727, 'samples': 10980864, 'steps': 21446, 'loss/train': 1.9036604166030884} 02/25/2022 05:21:49 - INFO - codeparrot_training - Step 21447: {'lr': 0.0003234005997105718, 'samples': 10981376, 'steps': 21447, 'loss/train': 1.9068318605422974} 02/25/2022 05:21:55 - INFO - codeparrot_training - Step 21448: {'lr': 0.0003233849582241214, 'samples': 10981888, 'steps': 21448, 'loss/train': 2.5308852195739746} 02/25/2022 05:21:58 - INFO - codeparrot_training - Step 21449: {'lr': 0.0003233693164233132, 'samples': 10982400, 'steps': 21449, 'loss/train': 0.7553144097328186} 02/25/2022 05:22:04 - INFO - codeparrot_training - Step 21450: {'lr': 0.00032335367430821416, 'samples': 10982912, 'steps': 21450, 'loss/train': 2.3498120307922363} 02/25/2022 05:22:07 - INFO - codeparrot_training - Step 21451: {'lr': 0.00032333803187889135, 'samples': 10983424, 'steps': 21451, 'loss/train': 2.21490478515625} 02/25/2022 05:22:14 - INFO - codeparrot_training - Step 21452: {'lr': 0.0003233223891354116, 'samples': 10983936, 'steps': 21452, 'loss/train': 1.1874514818191528} 02/25/2022 05:22:18 - INFO - codeparrot_training - Step 21453: {'lr': 0.000323306746077842, 'samples': 10984448, 'steps': 21453, 'loss/train': 1.4692349433898926} 02/25/2022 05:22:23 - INFO - codeparrot_training - Step 21454: {'lr': 0.00032329110270624956, 'samples': 10984960, 'steps': 21454, 'loss/train': 2.027933120727539} 02/25/2022 05:22:27 - INFO - codeparrot_training - Step 21455: {'lr': 0.00032327545902070137, 'samples': 10985472, 'steps': 21455, 'loss/train': 2.1449429988861084} 02/25/2022 05:22:32 - INFO - codeparrot_training - Step 21456: {'lr': 0.00032325981502126435, 'samples': 10985984, 'steps': 21456, 'loss/train': 2.320558786392212} 02/25/2022 05:22:36 - INFO - codeparrot_training - Step 21457: {'lr': 0.0003232441707080056, 'samples': 10986496, 'steps': 21457, 'loss/train': 3.3822484016418457} 02/25/2022 05:22:41 - INFO - codeparrot_training - Step 21458: {'lr': 0.00032322852608099203, 'samples': 10987008, 'steps': 21458, 'loss/train': 0.48733872175216675} 02/25/2022 05:22:45 - INFO - codeparrot_training - Step 21459: {'lr': 0.00032321288114029074, 'samples': 10987520, 'steps': 21459, 'loss/train': 2.1243081092834473} 02/25/2022 05:22:50 - INFO - codeparrot_training - Step 21460: {'lr': 0.00032319723588596875, 'samples': 10988032, 'steps': 21460, 'loss/train': 1.4248861074447632} 02/25/2022 05:22:54 - INFO - codeparrot_training - Step 21461: {'lr': 0.00032318159031809293, 'samples': 10988544, 'steps': 21461, 'loss/train': 1.7653262615203857} 02/25/2022 05:22:59 - INFO - codeparrot_training - Step 21462: {'lr': 0.00032316594443673047, 'samples': 10989056, 'steps': 21462, 'loss/train': 1.8202306032180786} 02/25/2022 05:23:03 - INFO - codeparrot_training - Step 21463: {'lr': 0.0003231502982419483, 'samples': 10989568, 'steps': 21463, 'loss/train': 1.9894782304763794} 02/25/2022 05:23:08 - INFO - codeparrot_training - Step 21464: {'lr': 0.0003231346517338135, 'samples': 10990080, 'steps': 21464, 'loss/train': 1.9325813055038452} 02/25/2022 05:23:12 - INFO - codeparrot_training - Step 21465: {'lr': 0.0003231190049123931, 'samples': 10990592, 'steps': 21465, 'loss/train': 1.9395051002502441} 02/25/2022 05:23:18 - INFO - codeparrot_training - Step 21466: {'lr': 0.00032310335777775413, 'samples': 10991104, 'steps': 21466, 'loss/train': 2.093909502029419} 02/25/2022 05:23:21 - INFO - codeparrot_training - Step 21467: {'lr': 0.00032308771032996353, 'samples': 10991616, 'steps': 21467, 'loss/train': 2.245347499847412} 02/25/2022 05:23:27 - INFO - codeparrot_training - Step 21468: {'lr': 0.0003230720625690884, 'samples': 10992128, 'steps': 21468, 'loss/train': 1.9406821727752686} 02/25/2022 05:23:30 - INFO - codeparrot_training - Step 21469: {'lr': 0.0003230564144951958, 'samples': 10992640, 'steps': 21469, 'loss/train': 2.02470064163208} 02/25/2022 05:23:36 - INFO - codeparrot_training - Step 21470: {'lr': 0.00032304076610835267, 'samples': 10993152, 'steps': 21470, 'loss/train': 0.7104726433753967} 02/25/2022 05:23:42 - INFO - codeparrot_training - Step 21471: {'lr': 0.0003230251174086261, 'samples': 10993664, 'steps': 21471, 'loss/train': 2.2231509685516357} 02/25/2022 05:23:45 - INFO - codeparrot_training - Step 21472: {'lr': 0.0003230094683960831, 'samples': 10994176, 'steps': 21472, 'loss/train': 2.0974111557006836} 02/25/2022 05:23:49 - INFO - codeparrot_training - Step 21473: {'lr': 0.0003229938190707908, 'samples': 10994688, 'steps': 21473, 'loss/train': 2.2401421070098877} 02/25/2022 05:23:54 - INFO - codeparrot_training - Step 21474: {'lr': 0.00032297816943281605, 'samples': 10995200, 'steps': 21474, 'loss/train': 1.2822587490081787} 02/25/2022 05:23:58 - INFO - codeparrot_training - Step 21475: {'lr': 0.00032296251948222605, 'samples': 10995712, 'steps': 21475, 'loss/train': 1.9859706163406372} 02/25/2022 05:24:04 - INFO - codeparrot_training - Step 21476: {'lr': 0.0003229468692190878, 'samples': 10996224, 'steps': 21476, 'loss/train': 1.5835591554641724} 02/25/2022 05:24:07 - INFO - codeparrot_training - Step 21477: {'lr': 0.00032293121864346823, 'samples': 10996736, 'steps': 21477, 'loss/train': 2.6338560581207275} 02/25/2022 05:24:13 - INFO - codeparrot_training - Step 21478: {'lr': 0.00032291556775543463, 'samples': 10997248, 'steps': 21478, 'loss/train': 0.9179028868675232} 02/25/2022 05:24:16 - INFO - codeparrot_training - Step 21479: {'lr': 0.0003228999165550537, 'samples': 10997760, 'steps': 21479, 'loss/train': 9.65449333190918} 02/25/2022 05:24:22 - INFO - codeparrot_training - Step 21480: {'lr': 0.0003228842650423929, 'samples': 10998272, 'steps': 21480, 'loss/train': 2.430394411087036} 02/25/2022 05:24:25 - INFO - codeparrot_training - Step 21481: {'lr': 0.0003228686132175189, 'samples': 10998784, 'steps': 21481, 'loss/train': 2.339261531829834} 02/25/2022 05:24:31 - INFO - codeparrot_training - Step 21482: {'lr': 0.0003228529610804989, 'samples': 10999296, 'steps': 21482, 'loss/train': 1.9135621786117554} 02/25/2022 05:24:34 - INFO - codeparrot_training - Step 21483: {'lr': 0.00032283730863140003, 'samples': 10999808, 'steps': 21483, 'loss/train': 2.3155276775360107} 02/25/2022 05:24:40 - INFO - codeparrot_training - Step 21484: {'lr': 0.0003228216558702892, 'samples': 11000320, 'steps': 21484, 'loss/train': 0.3337833285331726} 02/25/2022 05:24:44 - INFO - codeparrot_training - Step 21485: {'lr': 0.00032280600279723355, 'samples': 11000832, 'steps': 21485, 'loss/train': 2.135376214981079} 02/25/2022 05:24:51 - INFO - codeparrot_training - Step 21486: {'lr': 0.00032279034941230014, 'samples': 11001344, 'steps': 21486, 'loss/train': 2.3484268188476562} 02/25/2022 05:24:54 - INFO - codeparrot_training - Step 21487: {'lr': 0.00032277469571555587, 'samples': 11001856, 'steps': 21487, 'loss/train': 1.3024400472640991} 02/25/2022 05:24:58 - INFO - codeparrot_training - Step 21488: {'lr': 0.0003227590417070679, 'samples': 11002368, 'steps': 21488, 'loss/train': 2.356600522994995} 02/25/2022 05:25:03 - INFO - codeparrot_training - Step 21489: {'lr': 0.00032274338738690344, 'samples': 11002880, 'steps': 21489, 'loss/train': 2.263683319091797} 02/25/2022 05:25:07 - INFO - codeparrot_training - Step 21490: {'lr': 0.00032272773275512933, 'samples': 11003392, 'steps': 21490, 'loss/train': 1.0298773050308228} 02/25/2022 05:25:12 - INFO - codeparrot_training - Step 21491: {'lr': 0.0003227120778118127, 'samples': 11003904, 'steps': 21491, 'loss/train': 2.260049343109131} 02/25/2022 05:25:16 - INFO - codeparrot_training - Step 21492: {'lr': 0.00032269642255702065, 'samples': 11004416, 'steps': 21492, 'loss/train': 0.27687039971351624} 02/25/2022 05:25:21 - INFO - codeparrot_training - Step 21493: {'lr': 0.00032268076699082024, 'samples': 11004928, 'steps': 21493, 'loss/train': 2.0938594341278076} 02/25/2022 05:25:25 - INFO - codeparrot_training - Step 21494: {'lr': 0.0003226651111132784, 'samples': 11005440, 'steps': 21494, 'loss/train': 1.9103293418884277} 02/25/2022 05:25:30 - INFO - codeparrot_training - Step 21495: {'lr': 0.0003226494549244624, 'samples': 11005952, 'steps': 21495, 'loss/train': 2.0065977573394775} 02/25/2022 05:25:34 - INFO - codeparrot_training - Step 21496: {'lr': 0.00032263379842443915, 'samples': 11006464, 'steps': 21496, 'loss/train': 2.1660172939300537} 02/25/2022 05:25:39 - INFO - codeparrot_training - Step 21497: {'lr': 0.0003226181416132758, 'samples': 11006976, 'steps': 21497, 'loss/train': 1.0341888666152954} 02/25/2022 05:25:43 - INFO - codeparrot_training - Step 21498: {'lr': 0.00032260248449103937, 'samples': 11007488, 'steps': 21498, 'loss/train': 1.4313536882400513} 02/25/2022 05:25:50 - INFO - codeparrot_training - Step 21499: {'lr': 0.00032258682705779695, 'samples': 11008000, 'steps': 21499, 'loss/train': 1.1632885932922363} 02/25/2022 05:25:53 - INFO - codeparrot_training - Step 21500: {'lr': 0.00032257116931361555, 'samples': 11008512, 'steps': 21500, 'loss/train': 1.889289379119873} 02/25/2022 05:25:59 - INFO - codeparrot_training - Step 21501: {'lr': 0.0003225555112585624, 'samples': 11009024, 'steps': 21501, 'loss/train': 2.3959970474243164} 02/25/2022 05:26:04 - INFO - codeparrot_training - Step 21502: {'lr': 0.0003225398528927045, 'samples': 11009536, 'steps': 21502, 'loss/train': 1.7308250665664673} 02/25/2022 05:26:08 - INFO - codeparrot_training - Step 21503: {'lr': 0.00032252419421610883, 'samples': 11010048, 'steps': 21503, 'loss/train': 1.6846668720245361} 02/25/2022 05:26:13 - INFO - codeparrot_training - Step 21504: {'lr': 0.0003225085352288426, 'samples': 11010560, 'steps': 21504, 'loss/train': 1.2956887483596802} 02/25/2022 05:26:17 - INFO - codeparrot_training - Step 21505: {'lr': 0.00032249287593097274, 'samples': 11011072, 'steps': 21505, 'loss/train': 0.7393494844436646} 02/25/2022 05:26:22 - INFO - codeparrot_training - Step 21506: {'lr': 0.00032247721632256657, 'samples': 11011584, 'steps': 21506, 'loss/train': 2.6542041301727295} 02/25/2022 05:26:25 - INFO - codeparrot_training - Step 21507: {'lr': 0.0003224615564036908, 'samples': 11012096, 'steps': 21507, 'loss/train': 1.7451930046081543} 02/25/2022 05:26:32 - INFO - codeparrot_training - Step 21508: {'lr': 0.00032244589617441287, 'samples': 11012608, 'steps': 21508, 'loss/train': 0.37942302227020264} 02/25/2022 05:26:35 - INFO - codeparrot_training - Step 21509: {'lr': 0.0003224302356347997, 'samples': 11013120, 'steps': 21509, 'loss/train': 1.7924957275390625} 02/25/2022 05:26:41 - INFO - codeparrot_training - Step 21510: {'lr': 0.0003224145747849185, 'samples': 11013632, 'steps': 21510, 'loss/train': 1.631548285484314} 02/25/2022 05:26:44 - INFO - codeparrot_training - Step 21511: {'lr': 0.0003223989136248361, 'samples': 11014144, 'steps': 21511, 'loss/train': 2.365044355392456} 02/25/2022 05:26:50 - INFO - codeparrot_training - Step 21512: {'lr': 0.0003223832521546198, 'samples': 11014656, 'steps': 21512, 'loss/train': 2.034379005432129} 02/25/2022 05:26:53 - INFO - codeparrot_training - Step 21513: {'lr': 0.0003223675903743366, 'samples': 11015168, 'steps': 21513, 'loss/train': 1.1576682329177856} 02/25/2022 05:26:59 - INFO - codeparrot_training - Step 21514: {'lr': 0.0003223519282840537, 'samples': 11015680, 'steps': 21514, 'loss/train': 2.333731174468994} 02/25/2022 05:27:02 - INFO - codeparrot_training - Step 21515: {'lr': 0.00032233626588383806, 'samples': 11016192, 'steps': 21515, 'loss/train': 1.3007678985595703} 02/25/2022 05:27:08 - INFO - codeparrot_training - Step 21516: {'lr': 0.00032232060317375684, 'samples': 11016704, 'steps': 21516, 'loss/train': 2.1323704719543457} 02/25/2022 05:27:11 - INFO - codeparrot_training - Step 21517: {'lr': 0.00032230494015387715, 'samples': 11017216, 'steps': 21517, 'loss/train': 3.6352202892303467} 02/25/2022 05:27:17 - INFO - codeparrot_training - Step 21518: {'lr': 0.000322289276824266, 'samples': 11017728, 'steps': 21518, 'loss/train': 2.469447612762451} 02/25/2022 05:27:20 - INFO - codeparrot_training - Step 21519: {'lr': 0.0003222736131849906, 'samples': 11018240, 'steps': 21519, 'loss/train': 2.392763614654541} 02/25/2022 05:27:26 - INFO - codeparrot_training - Step 21520: {'lr': 0.0003222579492361179, 'samples': 11018752, 'steps': 21520, 'loss/train': 2.1672446727752686} 02/25/2022 05:27:29 - INFO - codeparrot_training - Step 21521: {'lr': 0.0003222422849777152, 'samples': 11019264, 'steps': 21521, 'loss/train': 1.3211790323257446} 02/25/2022 05:27:36 - INFO - codeparrot_training - Step 21522: {'lr': 0.0003222266204098494, 'samples': 11019776, 'steps': 21522, 'loss/train': 1.3548214435577393} 02/25/2022 05:27:39 - INFO - codeparrot_training - Step 21523: {'lr': 0.0003222109555325877, 'samples': 11020288, 'steps': 21523, 'loss/train': 2.0841355323791504} 02/25/2022 05:27:45 - INFO - codeparrot_training - Step 21524: {'lr': 0.00032219529034599725, 'samples': 11020800, 'steps': 21524, 'loss/train': 1.3415675163269043} 02/25/2022 05:27:48 - INFO - codeparrot_training - Step 21525: {'lr': 0.00032217962485014506, 'samples': 11021312, 'steps': 21525, 'loss/train': 1.6736340522766113} 02/25/2022 05:27:54 - INFO - codeparrot_training - Step 21526: {'lr': 0.0003221639590450983, 'samples': 11021824, 'steps': 21526, 'loss/train': 2.2644736766815186} 02/25/2022 05:27:57 - INFO - codeparrot_training - Step 21527: {'lr': 0.00032214829293092406, 'samples': 11022336, 'steps': 21527, 'loss/train': 1.2068449258804321} 02/25/2022 05:28:03 - INFO - codeparrot_training - Step 21528: {'lr': 0.0003221326265076894, 'samples': 11022848, 'steps': 21528, 'loss/train': 1.7141304016113281} 02/25/2022 05:28:06 - INFO - codeparrot_training - Step 21529: {'lr': 0.00032211695977546153, 'samples': 11023360, 'steps': 21529, 'loss/train': 1.680727243423462} 02/25/2022 05:28:12 - INFO - codeparrot_training - Step 21530: {'lr': 0.0003221012927343075, 'samples': 11023872, 'steps': 21530, 'loss/train': 1.3877259492874146} 02/25/2022 05:28:15 - INFO - codeparrot_training - Step 21531: {'lr': 0.0003220856253842944, 'samples': 11024384, 'steps': 21531, 'loss/train': 2.5571298599243164} 02/25/2022 05:28:21 - INFO - codeparrot_training - Step 21532: {'lr': 0.00032206995772548943, 'samples': 11024896, 'steps': 21532, 'loss/train': 2.1010231971740723} 02/25/2022 05:28:25 - INFO - codeparrot_training - Step 21533: {'lr': 0.00032205428975795955, 'samples': 11025408, 'steps': 21533, 'loss/train': 1.8804799318313599} 02/25/2022 05:28:28 - INFO - codeparrot_training - Step 21534: {'lr': 0.000322038621481772, 'samples': 11025920, 'steps': 21534, 'loss/train': 2.0093066692352295} 02/25/2022 05:28:35 - INFO - codeparrot_training - Step 21535: {'lr': 0.0003220229528969939, 'samples': 11026432, 'steps': 21535, 'loss/train': 1.9713863134384155} 02/25/2022 05:28:40 - INFO - codeparrot_training - Step 21536: {'lr': 0.00032200728400369233, 'samples': 11026944, 'steps': 21536, 'loss/train': 1.868619680404663} 02/25/2022 05:28:44 - INFO - codeparrot_training - Step 21537: {'lr': 0.0003219916148019344, 'samples': 11027456, 'steps': 21537, 'loss/train': 2.5183682441711426} 02/25/2022 05:28:50 - INFO - codeparrot_training - Step 21538: {'lr': 0.0003219759452917872, 'samples': 11027968, 'steps': 21538, 'loss/train': 2.6670279502868652} 02/25/2022 05:28:53 - INFO - codeparrot_training - Step 21539: {'lr': 0.000321960275473318, 'samples': 11028480, 'steps': 21539, 'loss/train': 1.315918207168579} 02/25/2022 05:28:57 - INFO - codeparrot_training - Step 21540: {'lr': 0.0003219446053465938, 'samples': 11028992, 'steps': 21540, 'loss/train': 1.5764955282211304} 02/25/2022 05:29:02 - INFO - codeparrot_training - Step 21541: {'lr': 0.0003219289349116818, 'samples': 11029504, 'steps': 21541, 'loss/train': 2.478363513946533} 02/25/2022 05:29:06 - INFO - codeparrot_training - Step 21542: {'lr': 0.0003219132641686491, 'samples': 11030016, 'steps': 21542, 'loss/train': 3.195913791656494} 02/25/2022 05:29:11 - INFO - codeparrot_training - Step 21543: {'lr': 0.0003218975931175627, 'samples': 11030528, 'steps': 21543, 'loss/train': 2.328716993331909} 02/25/2022 05:29:15 - INFO - codeparrot_training - Step 21544: {'lr': 0.0003218819217584899, 'samples': 11031040, 'steps': 21544, 'loss/train': 1.7310285568237305} 02/25/2022 05:29:21 - INFO - codeparrot_training - Step 21545: {'lr': 0.0003218662500914977, 'samples': 11031552, 'steps': 21545, 'loss/train': 2.249018907546997} 02/25/2022 05:29:25 - INFO - codeparrot_training - Step 21546: {'lr': 0.0003218505781166534, 'samples': 11032064, 'steps': 21546, 'loss/train': 1.839613676071167} 02/25/2022 05:29:30 - INFO - codeparrot_training - Step 21547: {'lr': 0.000321834905834024, 'samples': 11032576, 'steps': 21547, 'loss/train': 0.6879871487617493} 02/25/2022 05:29:34 - INFO - codeparrot_training - Step 21548: {'lr': 0.00032181923324367675, 'samples': 11033088, 'steps': 21548, 'loss/train': 2.6076583862304688} 02/25/2022 05:29:40 - INFO - codeparrot_training - Step 21549: {'lr': 0.0003218035603456786, 'samples': 11033600, 'steps': 21549, 'loss/train': 0.9620224833488464} 02/25/2022 05:29:43 - INFO - codeparrot_training - Step 21550: {'lr': 0.00032178788714009687, 'samples': 11034112, 'steps': 21550, 'loss/train': 2.3495805263519287} 02/25/2022 05:29:48 - INFO - codeparrot_training - Step 21551: {'lr': 0.00032177221362699853, 'samples': 11034624, 'steps': 21551, 'loss/train': 2.0556299686431885} 02/25/2022 05:29:52 - INFO - codeparrot_training - Step 21552: {'lr': 0.0003217565398064509, 'samples': 11035136, 'steps': 21552, 'loss/train': 1.6750190258026123} 02/25/2022 05:29:57 - INFO - codeparrot_training - Step 21553: {'lr': 0.000321740865678521, 'samples': 11035648, 'steps': 21553, 'loss/train': 2.536710500717163} 02/25/2022 05:30:01 - INFO - codeparrot_training - Step 21554: {'lr': 0.00032172519124327607, 'samples': 11036160, 'steps': 21554, 'loss/train': 2.2535555362701416} 02/25/2022 05:30:08 - INFO - codeparrot_training - Step 21555: {'lr': 0.00032170951650078316, 'samples': 11036672, 'steps': 21555, 'loss/train': 1.9997724294662476} 02/25/2022 05:30:11 - INFO - codeparrot_training - Step 21556: {'lr': 0.0003216938414511095, 'samples': 11037184, 'steps': 21556, 'loss/train': 0.8170745372772217} 02/25/2022 05:30:17 - INFO - codeparrot_training - Step 21557: {'lr': 0.0003216781660943221, 'samples': 11037696, 'steps': 21557, 'loss/train': 3.0403549671173096} 02/25/2022 05:30:20 - INFO - codeparrot_training - Step 21558: {'lr': 0.0003216624904304882, 'samples': 11038208, 'steps': 21558, 'loss/train': 0.5783214569091797} 02/25/2022 05:30:26 - INFO - codeparrot_training - Step 21559: {'lr': 0.000321646814459675, 'samples': 11038720, 'steps': 21559, 'loss/train': 1.0909634828567505} 02/25/2022 05:30:30 - INFO - codeparrot_training - Step 21560: {'lr': 0.0003216311381819496, 'samples': 11039232, 'steps': 21560, 'loss/train': 1.8134702444076538} 02/25/2022 05:30:35 - INFO - codeparrot_training - Step 21561: {'lr': 0.00032161546159737917, 'samples': 11039744, 'steps': 21561, 'loss/train': 1.2008908987045288} 02/25/2022 05:30:39 - INFO - codeparrot_training - Step 21562: {'lr': 0.0003215997847060307, 'samples': 11040256, 'steps': 21562, 'loss/train': 1.9453626871109009} 02/25/2022 05:30:44 - INFO - codeparrot_training - Step 21563: {'lr': 0.00032158410750797163, 'samples': 11040768, 'steps': 21563, 'loss/train': 2.7384119033813477} 02/25/2022 05:30:48 - INFO - codeparrot_training - Step 21564: {'lr': 0.000321568430003269, 'samples': 11041280, 'steps': 21564, 'loss/train': 2.250610589981079} 02/25/2022 05:30:53 - INFO - codeparrot_training - Step 21565: {'lr': 0.00032155275219198986, 'samples': 11041792, 'steps': 21565, 'loss/train': 2.0307776927948} 02/25/2022 05:30:57 - INFO - codeparrot_training - Step 21566: {'lr': 0.0003215370740742014, 'samples': 11042304, 'steps': 21566, 'loss/train': 2.0014281272888184} 02/25/2022 05:31:02 - INFO - codeparrot_training - Step 21567: {'lr': 0.00032152139564997097, 'samples': 11042816, 'steps': 21567, 'loss/train': 2.3043441772460938} 02/25/2022 05:31:06 - INFO - codeparrot_training - Step 21568: {'lr': 0.0003215057169193655, 'samples': 11043328, 'steps': 21568, 'loss/train': 1.6961584091186523} 02/25/2022 05:31:12 - INFO - codeparrot_training - Step 21569: {'lr': 0.00032149003788245223, 'samples': 11043840, 'steps': 21569, 'loss/train': 1.4521780014038086} 02/25/2022 05:31:15 - INFO - codeparrot_training - Step 21570: {'lr': 0.0003214743585392984, 'samples': 11044352, 'steps': 21570, 'loss/train': 2.091576337814331} 02/25/2022 05:31:21 - INFO - codeparrot_training - Step 21571: {'lr': 0.0003214586788899711, 'samples': 11044864, 'steps': 21571, 'loss/train': 0.19392921030521393} 02/25/2022 05:31:24 - INFO - codeparrot_training - Step 21572: {'lr': 0.00032144299893453743, 'samples': 11045376, 'steps': 21572, 'loss/train': 2.0229432582855225} 02/25/2022 05:31:30 - INFO - codeparrot_training - Step 21573: {'lr': 0.00032142731867306466, 'samples': 11045888, 'steps': 21573, 'loss/train': 1.1942694187164307} 02/25/2022 05:31:33 - INFO - codeparrot_training - Step 21574: {'lr': 0.00032141163810562, 'samples': 11046400, 'steps': 21574, 'loss/train': 1.114532470703125} 02/25/2022 05:31:39 - INFO - codeparrot_training - Step 21575: {'lr': 0.00032139595723227054, 'samples': 11046912, 'steps': 21575, 'loss/train': 1.5500388145446777} 02/25/2022 05:31:43 - INFO - codeparrot_training - Step 21576: {'lr': 0.0003213802760530835, 'samples': 11047424, 'steps': 21576, 'loss/train': 1.2036854028701782} 02/25/2022 05:31:48 - INFO - codeparrot_training - Step 21577: {'lr': 0.000321364594568126, 'samples': 11047936, 'steps': 21577, 'loss/train': 1.8385207653045654} 02/25/2022 05:31:51 - INFO - codeparrot_training - Step 21578: {'lr': 0.00032134891277746527, 'samples': 11048448, 'steps': 21578, 'loss/train': 1.7082643508911133} 02/25/2022 05:31:58 - INFO - codeparrot_training - Step 21579: {'lr': 0.0003213332306811684, 'samples': 11048960, 'steps': 21579, 'loss/train': 1.6035127639770508} 02/25/2022 05:32:01 - INFO - codeparrot_training - Step 21580: {'lr': 0.0003213175482793026, 'samples': 11049472, 'steps': 21580, 'loss/train': 2.3959131240844727} 02/25/2022 05:32:07 - INFO - codeparrot_training - Step 21581: {'lr': 0.00032130186557193506, 'samples': 11049984, 'steps': 21581, 'loss/train': 2.015110969543457} 02/25/2022 05:32:10 - INFO - codeparrot_training - Step 21582: {'lr': 0.0003212861825591331, 'samples': 11050496, 'steps': 21582, 'loss/train': 1.8352669477462769} 02/25/2022 05:32:16 - INFO - codeparrot_training - Step 21583: {'lr': 0.00032127049924096364, 'samples': 11051008, 'steps': 21583, 'loss/train': 1.3586941957473755} 02/25/2022 05:32:19 - INFO - codeparrot_training - Step 21584: {'lr': 0.00032125481561749405, 'samples': 11051520, 'steps': 21584, 'loss/train': 2.635873556137085} 02/25/2022 05:32:25 - INFO - codeparrot_training - Step 21585: {'lr': 0.00032123913168879146, 'samples': 11052032, 'steps': 21585, 'loss/train': 1.7228937149047852} 02/25/2022 05:32:29 - INFO - codeparrot_training - Step 21586: {'lr': 0.00032122344745492303, 'samples': 11052544, 'steps': 21586, 'loss/train': 2.5237181186676025} 02/25/2022 05:32:34 - INFO - codeparrot_training - Step 21587: {'lr': 0.00032120776291595594, 'samples': 11053056, 'steps': 21587, 'loss/train': 1.6375033855438232} 02/25/2022 05:32:38 - INFO - codeparrot_training - Step 21588: {'lr': 0.00032119207807195747, 'samples': 11053568, 'steps': 21588, 'loss/train': 2.575406551361084} 02/25/2022 05:32:41 - INFO - codeparrot_training - Step 21589: {'lr': 0.0003211763929229947, 'samples': 11054080, 'steps': 21589, 'loss/train': 1.9104794263839722} 02/25/2022 05:32:47 - INFO - codeparrot_training - Step 21590: {'lr': 0.00032116070746913484, 'samples': 11054592, 'steps': 21590, 'loss/train': 1.5984069108963013} 02/25/2022 05:32:50 - INFO - codeparrot_training - Step 21591: {'lr': 0.0003211450217104452, 'samples': 11055104, 'steps': 21591, 'loss/train': 1.8124760389328003} 02/25/2022 05:32:56 - INFO - codeparrot_training - Step 21592: {'lr': 0.00032112933564699275, 'samples': 11055616, 'steps': 21592, 'loss/train': 1.4645565748214722} 02/25/2022 05:33:00 - INFO - codeparrot_training - Step 21593: {'lr': 0.0003211136492788449, 'samples': 11056128, 'steps': 21593, 'loss/train': 2.189774990081787} 02/25/2022 05:33:05 - INFO - codeparrot_training - Step 21594: {'lr': 0.0003210979626060687, 'samples': 11056640, 'steps': 21594, 'loss/train': 2.0911288261413574} 02/25/2022 05:33:09 - INFO - codeparrot_training - Step 21595: {'lr': 0.00032108227562873147, 'samples': 11057152, 'steps': 21595, 'loss/train': 1.86595618724823} 02/25/2022 05:33:14 - INFO - codeparrot_training - Step 21596: {'lr': 0.0003210665883469003, 'samples': 11057664, 'steps': 21596, 'loss/train': 1.6551841497421265} 02/25/2022 05:33:18 - INFO - codeparrot_training - Step 21597: {'lr': 0.0003210509007606424, 'samples': 11058176, 'steps': 21597, 'loss/train': 1.5344195365905762} 02/25/2022 05:33:23 - INFO - codeparrot_training - Step 21598: {'lr': 0.00032103521287002505, 'samples': 11058688, 'steps': 21598, 'loss/train': 1.24522066116333} 02/25/2022 05:33:26 - INFO - codeparrot_training - Step 21599: {'lr': 0.0003210195246751154, 'samples': 11059200, 'steps': 21599, 'loss/train': 1.8943443298339844} 02/25/2022 05:33:32 - INFO - codeparrot_training - Step 21600: {'lr': 0.0003210038361759807, 'samples': 11059712, 'steps': 21600, 'loss/train': 1.5027269124984741} 02/25/2022 05:33:35 - INFO - codeparrot_training - Step 21601: {'lr': 0.000320988147372688, 'samples': 11060224, 'steps': 21601, 'loss/train': 1.9278641939163208} 02/25/2022 05:33:42 - INFO - codeparrot_training - Step 21602: {'lr': 0.00032097245826530476, 'samples': 11060736, 'steps': 21602, 'loss/train': 1.8387091159820557} 02/25/2022 05:33:45 - INFO - codeparrot_training - Step 21603: {'lr': 0.00032095676885389793, 'samples': 11061248, 'steps': 21603, 'loss/train': 2.8980751037597656} 02/25/2022 05:33:51 - INFO - codeparrot_training - Step 21604: {'lr': 0.00032094107913853485, 'samples': 11061760, 'steps': 21604, 'loss/train': 1.0872629880905151} 02/25/2022 05:33:54 - INFO - codeparrot_training - Step 21605: {'lr': 0.00032092538911928276, 'samples': 11062272, 'steps': 21605, 'loss/train': 1.9053674936294556} 02/25/2022 05:34:00 - INFO - codeparrot_training - Step 21606: {'lr': 0.00032090969879620886, 'samples': 11062784, 'steps': 21606, 'loss/train': 1.8309128284454346} 02/25/2022 05:34:03 - INFO - codeparrot_training - Step 21607: {'lr': 0.00032089400816938016, 'samples': 11063296, 'steps': 21607, 'loss/train': 2.072174072265625} 02/25/2022 05:34:09 - INFO - codeparrot_training - Step 21608: {'lr': 0.0003208783172388642, 'samples': 11063808, 'steps': 21608, 'loss/train': 1.8215605020523071} 02/25/2022 05:34:12 - INFO - codeparrot_training - Step 21609: {'lr': 0.000320862626004728, 'samples': 11064320, 'steps': 21609, 'loss/train': 1.2355942726135254} 02/25/2022 05:34:18 - INFO - codeparrot_training - Step 21610: {'lr': 0.00032084693446703875, 'samples': 11064832, 'steps': 21610, 'loss/train': 1.642086386680603} 02/25/2022 05:34:21 - INFO - codeparrot_training - Step 21611: {'lr': 0.00032083124262586384, 'samples': 11065344, 'steps': 21611, 'loss/train': 2.1018474102020264} 02/25/2022 05:34:27 - INFO - codeparrot_training - Step 21612: {'lr': 0.0003208155504812703, 'samples': 11065856, 'steps': 21612, 'loss/train': 2.3018088340759277} 02/25/2022 05:34:30 - INFO - codeparrot_training - Step 21613: {'lr': 0.00032079985803332546, 'samples': 11066368, 'steps': 21613, 'loss/train': 2.333974838256836} 02/25/2022 05:34:36 - INFO - codeparrot_training - Step 21614: {'lr': 0.0003207841652820964, 'samples': 11066880, 'steps': 21614, 'loss/train': 2.717080593109131} 02/25/2022 05:34:39 - INFO - codeparrot_training - Step 21615: {'lr': 0.0003207684722276506, 'samples': 11067392, 'steps': 21615, 'loss/train': 1.194378137588501} 02/25/2022 05:34:46 - INFO - codeparrot_training - Step 21616: {'lr': 0.00032075277887005503, 'samples': 11067904, 'steps': 21616, 'loss/train': 1.5053054094314575} 02/25/2022 05:34:49 - INFO - codeparrot_training - Step 21617: {'lr': 0.0003207370852093771, 'samples': 11068416, 'steps': 21617, 'loss/train': 2.3491132259368896} 02/25/2022 05:34:55 - INFO - codeparrot_training - Step 21618: {'lr': 0.00032072139124568396, 'samples': 11068928, 'steps': 21618, 'loss/train': 1.3829879760742188} 02/25/2022 05:34:58 - INFO - codeparrot_training - Step 21619: {'lr': 0.0003207056969790428, 'samples': 11069440, 'steps': 21619, 'loss/train': 1.9224377870559692} 02/25/2022 05:35:04 - INFO - codeparrot_training - Step 21620: {'lr': 0.0003206900024095208, 'samples': 11069952, 'steps': 21620, 'loss/train': 0.983349621295929} 02/25/2022 05:35:07 - INFO - codeparrot_training - Step 21621: {'lr': 0.0003206743075371854, 'samples': 11070464, 'steps': 21621, 'loss/train': 1.8232043981552124} 02/25/2022 05:35:13 - INFO - codeparrot_training - Step 21622: {'lr': 0.0003206586123621037, 'samples': 11070976, 'steps': 21622, 'loss/train': 1.5382137298583984} 02/25/2022 05:35:17 - INFO - codeparrot_training - Step 21623: {'lr': 0.00032064291688434286, 'samples': 11071488, 'steps': 21623, 'loss/train': 2.208237409591675} 02/25/2022 05:35:22 - INFO - codeparrot_training - Step 21624: {'lr': 0.00032062722110397034, 'samples': 11072000, 'steps': 21624, 'loss/train': 2.0548887252807617} 02/25/2022 05:35:26 - INFO - codeparrot_training - Step 21625: {'lr': 0.0003206115250210531, 'samples': 11072512, 'steps': 21625, 'loss/train': 2.4550108909606934} 02/25/2022 05:35:32 - INFO - codeparrot_training - Step 21626: {'lr': 0.00032059582863565864, 'samples': 11073024, 'steps': 21626, 'loss/train': 1.967165470123291} 02/25/2022 05:35:35 - INFO - codeparrot_training - Step 21627: {'lr': 0.0003205801319478539, 'samples': 11073536, 'steps': 21627, 'loss/train': 2.440612316131592} 02/25/2022 05:35:41 - INFO - codeparrot_training - Step 21628: {'lr': 0.00032056443495770637, 'samples': 11074048, 'steps': 21628, 'loss/train': 2.2766144275665283} 02/25/2022 05:35:44 - INFO - codeparrot_training - Step 21629: {'lr': 0.0003205487376652833, 'samples': 11074560, 'steps': 21629, 'loss/train': 1.8009629249572754} 02/25/2022 05:35:50 - INFO - codeparrot_training - Step 21630: {'lr': 0.0003205330400706517, 'samples': 11075072, 'steps': 21630, 'loss/train': 2.550490140914917} 02/25/2022 05:35:53 - INFO - codeparrot_training - Step 21631: {'lr': 0.000320517342173879, 'samples': 11075584, 'steps': 21631, 'loss/train': 2.653726100921631} 02/25/2022 05:35:59 - INFO - codeparrot_training - Step 21632: {'lr': 0.0003205016439750323, 'samples': 11076096, 'steps': 21632, 'loss/train': 1.3426618576049805} 02/25/2022 05:36:02 - INFO - codeparrot_training - Step 21633: {'lr': 0.00032048594547417916, 'samples': 11076608, 'steps': 21633, 'loss/train': 2.6711585521698} 02/25/2022 05:36:08 - INFO - codeparrot_training - Step 21634: {'lr': 0.00032047024667138644, 'samples': 11077120, 'steps': 21634, 'loss/train': 1.8430368900299072} 02/25/2022 05:36:11 - INFO - codeparrot_training - Step 21635: {'lr': 0.00032045454756672164, 'samples': 11077632, 'steps': 21635, 'loss/train': 1.7331066131591797} 02/25/2022 05:36:17 - INFO - codeparrot_training - Step 21636: {'lr': 0.00032043884816025187, 'samples': 11078144, 'steps': 21636, 'loss/train': 1.765415072441101} 02/25/2022 05:36:20 - INFO - codeparrot_training - Step 21637: {'lr': 0.0003204231484520445, 'samples': 11078656, 'steps': 21637, 'loss/train': 2.5659048557281494} 02/25/2022 05:36:26 - INFO - codeparrot_training - Step 21638: {'lr': 0.0003204074484421667, 'samples': 11079168, 'steps': 21638, 'loss/train': 3.0074050426483154} 02/25/2022 05:36:30 - INFO - codeparrot_training - Step 21639: {'lr': 0.0003203917481306857, 'samples': 11079680, 'steps': 21639, 'loss/train': 2.3378326892852783} 02/25/2022 05:36:35 - INFO - codeparrot_training - Step 21640: {'lr': 0.0003203760475176689, 'samples': 11080192, 'steps': 21640, 'loss/train': 1.4756957292556763} 02/25/2022 05:36:39 - INFO - codeparrot_training - Step 21641: {'lr': 0.00032036034660318344, 'samples': 11080704, 'steps': 21641, 'loss/train': 1.7736973762512207} 02/25/2022 05:36:44 - INFO - codeparrot_training - Step 21642: {'lr': 0.00032034464538729647, 'samples': 11081216, 'steps': 21642, 'loss/train': 2.2646687030792236} 02/25/2022 05:36:48 - INFO - codeparrot_training - Step 21643: {'lr': 0.0003203289438700755, 'samples': 11081728, 'steps': 21643, 'loss/train': 2.2318098545074463} 02/25/2022 05:36:53 - INFO - codeparrot_training - Step 21644: {'lr': 0.0003203132420515876, 'samples': 11082240, 'steps': 21644, 'loss/train': 2.0150458812713623} 02/25/2022 05:36:57 - INFO - codeparrot_training - Step 21645: {'lr': 0.0003202975399319002, 'samples': 11082752, 'steps': 21645, 'loss/train': 2.0222859382629395} 02/25/2022 05:37:02 - INFO - codeparrot_training - Step 21646: {'lr': 0.00032028183751108035, 'samples': 11083264, 'steps': 21646, 'loss/train': 1.8015633821487427} 02/25/2022 05:37:06 - INFO - codeparrot_training - Step 21647: {'lr': 0.00032026613478919547, 'samples': 11083776, 'steps': 21647, 'loss/train': 1.9266560077667236} 02/25/2022 05:37:12 - INFO - codeparrot_training - Step 21648: {'lr': 0.0003202504317663128, 'samples': 11084288, 'steps': 21648, 'loss/train': 0.4452267289161682} 02/25/2022 05:37:16 - INFO - codeparrot_training - Step 21649: {'lr': 0.0003202347284424995, 'samples': 11084800, 'steps': 21649, 'loss/train': 1.8500224351882935} 02/25/2022 05:37:21 - INFO - codeparrot_training - Step 21650: {'lr': 0.00032021902481782304, 'samples': 11085312, 'steps': 21650, 'loss/train': 1.7770010232925415} 02/25/2022 05:37:25 - INFO - codeparrot_training - Step 21651: {'lr': 0.0003202033208923505, 'samples': 11085824, 'steps': 21651, 'loss/train': 1.1370052099227905} 02/25/2022 05:37:30 - INFO - codeparrot_training - Step 21652: {'lr': 0.0003201876166661493, 'samples': 11086336, 'steps': 21652, 'loss/train': 1.2044696807861328} 02/25/2022 05:37:34 - INFO - codeparrot_training - Step 21653: {'lr': 0.00032017191213928653, 'samples': 11086848, 'steps': 21653, 'loss/train': 1.6115232706069946} 02/25/2022 05:37:39 - INFO - codeparrot_training - Step 21654: {'lr': 0.0003201562073118297, 'samples': 11087360, 'steps': 21654, 'loss/train': 1.9590775966644287} 02/25/2022 05:37:43 - INFO - codeparrot_training - Step 21655: {'lr': 0.00032014050218384584, 'samples': 11087872, 'steps': 21655, 'loss/train': 1.4783846139907837} 02/25/2022 05:37:48 - INFO - codeparrot_training - Step 21656: {'lr': 0.0003201247967554024, 'samples': 11088384, 'steps': 21656, 'loss/train': 2.3855648040771484} 02/25/2022 05:37:52 - INFO - codeparrot_training - Step 21657: {'lr': 0.0003201090910265666, 'samples': 11088896, 'steps': 21657, 'loss/train': 0.8637088537216187} 02/25/2022 05:37:57 - INFO - codeparrot_training - Step 21658: {'lr': 0.0003200933849974056, 'samples': 11089408, 'steps': 21658, 'loss/train': 2.0209999084472656} 02/25/2022 05:38:01 - INFO - codeparrot_training - Step 21659: {'lr': 0.000320077678667987, 'samples': 11089920, 'steps': 21659, 'loss/train': 2.6658084392547607} 02/25/2022 05:38:06 - INFO - codeparrot_training - Step 21660: {'lr': 0.00032006197203837775, 'samples': 11090432, 'steps': 21660, 'loss/train': 2.1893503665924072} 02/25/2022 05:38:10 - INFO - codeparrot_training - Step 21661: {'lr': 0.00032004626510864526, 'samples': 11090944, 'steps': 21661, 'loss/train': 1.649872899055481} 02/25/2022 05:38:16 - INFO - codeparrot_training - Step 21662: {'lr': 0.00032003055787885684, 'samples': 11091456, 'steps': 21662, 'loss/train': 0.434133380651474} 02/25/2022 05:38:19 - INFO - codeparrot_training - Step 21663: {'lr': 0.00032001485034907975, 'samples': 11091968, 'steps': 21663, 'loss/train': 1.8698738813400269} 02/25/2022 05:38:25 - INFO - codeparrot_training - Step 21664: {'lr': 0.00031999914251938124, 'samples': 11092480, 'steps': 21664, 'loss/train': 2.088303327560425} 02/25/2022 05:38:29 - INFO - codeparrot_training - Step 21665: {'lr': 0.00031998343438982866, 'samples': 11092992, 'steps': 21665, 'loss/train': 2.7469441890716553} 02/25/2022 05:38:34 - INFO - codeparrot_training - Step 21666: {'lr': 0.0003199677259604893, 'samples': 11093504, 'steps': 21666, 'loss/train': 0.9768959879875183} 02/25/2022 05:38:38 - INFO - codeparrot_training - Step 21667: {'lr': 0.0003199520172314304, 'samples': 11094016, 'steps': 21667, 'loss/train': 2.5853993892669678} 02/25/2022 05:38:44 - INFO - codeparrot_training - Step 21668: {'lr': 0.00031993630820271925, 'samples': 11094528, 'steps': 21668, 'loss/train': 2.009589195251465} 02/25/2022 05:38:48 - INFO - codeparrot_training - Step 21669: {'lr': 0.00031992059887442316, 'samples': 11095040, 'steps': 21669, 'loss/train': 1.9536654949188232} 02/25/2022 05:38:51 - INFO - codeparrot_training - Step 21670: {'lr': 0.00031990488924660943, 'samples': 11095552, 'steps': 21670, 'loss/train': 2.4060659408569336} 02/25/2022 05:38:57 - INFO - codeparrot_training - Step 21671: {'lr': 0.00031988917931934536, 'samples': 11096064, 'steps': 21671, 'loss/train': 2.093985080718994} 02/25/2022 05:39:00 - INFO - codeparrot_training - Step 21672: {'lr': 0.0003198734690926982, 'samples': 11096576, 'steps': 21672, 'loss/train': 2.3026843070983887} 02/25/2022 05:39:06 - INFO - codeparrot_training - Step 21673: {'lr': 0.00031985775856673536, 'samples': 11097088, 'steps': 21673, 'loss/train': 0.8858436346054077} 02/25/2022 05:39:09 - INFO - codeparrot_training - Step 21674: {'lr': 0.000319842047741524, 'samples': 11097600, 'steps': 21674, 'loss/train': 2.1434948444366455} 02/25/2022 05:39:16 - INFO - codeparrot_training - Step 21675: {'lr': 0.0003198263366171315, 'samples': 11098112, 'steps': 21675, 'loss/train': 2.1019387245178223} 02/25/2022 05:39:19 - INFO - codeparrot_training - Step 21676: {'lr': 0.00031981062519362513, 'samples': 11098624, 'steps': 21676, 'loss/train': 2.29005765914917} 02/25/2022 05:39:24 - INFO - codeparrot_training - Step 21677: {'lr': 0.00031979491347107226, 'samples': 11099136, 'steps': 21677, 'loss/train': 2.097097873687744} 02/25/2022 05:39:28 - INFO - codeparrot_training - Step 21678: {'lr': 0.0003197792014495402, 'samples': 11099648, 'steps': 21678, 'loss/train': 2.0289864540100098} 02/25/2022 05:39:34 - INFO - codeparrot_training - Step 21679: {'lr': 0.0003197634891290961, 'samples': 11100160, 'steps': 21679, 'loss/train': 1.9478580951690674} 02/25/2022 05:39:37 - INFO - codeparrot_training - Step 21680: {'lr': 0.00031974777650980735, 'samples': 11100672, 'steps': 21680, 'loss/train': 1.9184279441833496} 02/25/2022 05:39:42 - INFO - codeparrot_training - Step 21681: {'lr': 0.0003197320635917413, 'samples': 11101184, 'steps': 21681, 'loss/train': 1.909900426864624} 02/25/2022 05:39:46 - INFO - codeparrot_training - Step 21682: {'lr': 0.0003197163503749652, 'samples': 11101696, 'steps': 21682, 'loss/train': 1.452235460281372} 02/25/2022 05:39:52 - INFO - codeparrot_training - Step 21683: {'lr': 0.00031970063685954645, 'samples': 11102208, 'steps': 21683, 'loss/train': 1.1591506004333496} 02/25/2022 05:39:55 - INFO - codeparrot_training - Step 21684: {'lr': 0.0003196849230455523, 'samples': 11102720, 'steps': 21684, 'loss/train': 2.5111916065216064} 02/25/2022 05:40:01 - INFO - codeparrot_training - Step 21685: {'lr': 0.0003196692089330501, 'samples': 11103232, 'steps': 21685, 'loss/train': 4.647315502166748} 02/25/2022 05:40:05 - INFO - codeparrot_training - Step 21686: {'lr': 0.000319653494522107, 'samples': 11103744, 'steps': 21686, 'loss/train': 1.4676721096038818} 02/25/2022 05:40:10 - INFO - codeparrot_training - Step 21687: {'lr': 0.00031963777981279057, 'samples': 11104256, 'steps': 21687, 'loss/train': 2.463695764541626} 02/25/2022 05:40:14 - INFO - codeparrot_training - Step 21688: {'lr': 0.00031962206480516794, 'samples': 11104768, 'steps': 21688, 'loss/train': 2.2198917865753174} 02/25/2022 05:40:19 - INFO - codeparrot_training - Step 21689: {'lr': 0.00031960634949930656, 'samples': 11105280, 'steps': 21689, 'loss/train': 3.032831907272339} 02/25/2022 05:40:23 - INFO - codeparrot_training - Step 21690: {'lr': 0.0003195906338952736, 'samples': 11105792, 'steps': 21690, 'loss/train': 1.944473385810852} 02/25/2022 05:40:29 - INFO - codeparrot_training - Step 21691: {'lr': 0.00031957491799313646, 'samples': 11106304, 'steps': 21691, 'loss/train': 1.865497350692749} 02/25/2022 05:40:32 - INFO - codeparrot_training - Step 21692: {'lr': 0.0003195592017929625, 'samples': 11106816, 'steps': 21692, 'loss/train': 1.2514517307281494} 02/25/2022 05:40:38 - INFO - codeparrot_training - Step 21693: {'lr': 0.000319543485294819, 'samples': 11107328, 'steps': 21693, 'loss/train': 2.192732810974121} 02/25/2022 05:40:41 - INFO - codeparrot_training - Step 21694: {'lr': 0.0003195277684987733, 'samples': 11107840, 'steps': 21694, 'loss/train': 1.3806431293487549} 02/25/2022 05:40:47 - INFO - codeparrot_training - Step 21695: {'lr': 0.0003195120514048927, 'samples': 11108352, 'steps': 21695, 'loss/train': 1.8311817646026611} 02/25/2022 05:40:51 - INFO - codeparrot_training - Step 21696: {'lr': 0.00031949633401324464, 'samples': 11108864, 'steps': 21696, 'loss/train': 3.015684127807617} 02/25/2022 05:40:56 - INFO - codeparrot_training - Step 21697: {'lr': 0.00031948061632389624, 'samples': 11109376, 'steps': 21697, 'loss/train': 2.6459567546844482} 02/25/2022 05:41:00 - INFO - codeparrot_training - Step 21698: {'lr': 0.00031946489833691494, 'samples': 11109888, 'steps': 21698, 'loss/train': 2.950221061706543} 02/25/2022 05:41:06 - INFO - codeparrot_training - Step 21699: {'lr': 0.0003194491800523681, 'samples': 11110400, 'steps': 21699, 'loss/train': 8.44400691986084} 02/25/2022 05:41:09 - INFO - codeparrot_training - Step 21700: {'lr': 0.0003194334614703231, 'samples': 11110912, 'steps': 21700, 'loss/train': 1.637054204940796} 02/25/2022 05:41:14 - INFO - codeparrot_training - Step 21701: {'lr': 0.0003194177425908471, 'samples': 11111424, 'steps': 21701, 'loss/train': 2.4237711429595947} 02/25/2022 05:41:18 - INFO - codeparrot_training - Step 21702: {'lr': 0.0003194020234140076, 'samples': 11111936, 'steps': 21702, 'loss/train': 0.9742262363433838} 02/25/2022 05:41:24 - INFO - codeparrot_training - Step 21703: {'lr': 0.00031938630393987176, 'samples': 11112448, 'steps': 21703, 'loss/train': 3.2086498737335205} 02/25/2022 05:41:27 - INFO - codeparrot_training - Step 21704: {'lr': 0.00031937058416850716, 'samples': 11112960, 'steps': 21704, 'loss/train': 2.247377395629883} 02/25/2022 05:41:31 - INFO - codeparrot_training - Step 21705: {'lr': 0.00031935486409998096, 'samples': 11113472, 'steps': 21705, 'loss/train': 2.8686368465423584} 02/25/2022 05:41:37 - INFO - codeparrot_training - Step 21706: {'lr': 0.0003193391437343605, 'samples': 11113984, 'steps': 21706, 'loss/train': 1.051864743232727} 02/25/2022 05:41:40 - INFO - codeparrot_training - Step 21707: {'lr': 0.0003193234230717132, 'samples': 11114496, 'steps': 21707, 'loss/train': 2.027564287185669} 02/25/2022 05:41:46 - INFO - codeparrot_training - Step 21708: {'lr': 0.00031930770211210637, 'samples': 11115008, 'steps': 21708, 'loss/train': 1.927280306816101} 02/25/2022 05:41:49 - INFO - codeparrot_training - Step 21709: {'lr': 0.0003192919808556073, 'samples': 11115520, 'steps': 21709, 'loss/train': 2.390582323074341} 02/25/2022 05:41:57 - INFO - codeparrot_training - Step 21710: {'lr': 0.00031927625930228343, 'samples': 11116032, 'steps': 21710, 'loss/train': 0.09722858667373657} 02/25/2022 05:42:00 - INFO - codeparrot_training - Step 21711: {'lr': 0.00031926053745220213, 'samples': 11116544, 'steps': 21711, 'loss/train': 1.2596288919448853} 02/25/2022 05:42:06 - INFO - codeparrot_training - Step 21712: {'lr': 0.0003192448153054306, 'samples': 11117056, 'steps': 21712, 'loss/train': 1.5834779739379883} 02/25/2022 05:42:09 - INFO - codeparrot_training - Step 21713: {'lr': 0.0003192290928620363, 'samples': 11117568, 'steps': 21713, 'loss/train': 0.8703785538673401} 02/25/2022 05:42:15 - INFO - codeparrot_training - Step 21714: {'lr': 0.0003192133701220865, 'samples': 11118080, 'steps': 21714, 'loss/train': 1.3689699172973633} 02/25/2022 05:42:18 - INFO - codeparrot_training - Step 21715: {'lr': 0.0003191976470856487, 'samples': 11118592, 'steps': 21715, 'loss/train': 0.7610230445861816} 02/25/2022 05:42:24 - INFO - codeparrot_training - Step 21716: {'lr': 0.00031918192375279006, 'samples': 11119104, 'steps': 21716, 'loss/train': 2.0348331928253174} 02/25/2022 05:42:27 - INFO - codeparrot_training - Step 21717: {'lr': 0.00031916620012357804, 'samples': 11119616, 'steps': 21717, 'loss/train': 1.950169324874878} 02/25/2022 05:42:33 - INFO - codeparrot_training - Step 21718: {'lr': 0.00031915047619808, 'samples': 11120128, 'steps': 21718, 'loss/train': 1.8343373537063599} 02/25/2022 05:42:36 - INFO - codeparrot_training - Step 21719: {'lr': 0.0003191347519763633, 'samples': 11120640, 'steps': 21719, 'loss/train': 4.004819869995117} 02/25/2022 05:42:44 - INFO - codeparrot_training - Step 21720: {'lr': 0.00031911902745849525, 'samples': 11121152, 'steps': 21720, 'loss/train': 3.9783334732055664} 02/25/2022 05:42:47 - INFO - codeparrot_training - Step 21721: {'lr': 0.0003191033026445432, 'samples': 11121664, 'steps': 21721, 'loss/train': 0.32352808117866516} 02/25/2022 05:42:53 - INFO - codeparrot_training - Step 21722: {'lr': 0.00031908757753457465, 'samples': 11122176, 'steps': 21722, 'loss/train': 1.5259758234024048} 02/25/2022 05:42:56 - INFO - codeparrot_training - Step 21723: {'lr': 0.00031907185212865673, 'samples': 11122688, 'steps': 21723, 'loss/train': 2.676894426345825} 02/25/2022 05:43:02 - INFO - codeparrot_training - Step 21724: {'lr': 0.000319056126426857, 'samples': 11123200, 'steps': 21724, 'loss/train': 2.128580093383789} 02/25/2022 05:43:05 - INFO - codeparrot_training - Step 21725: {'lr': 0.0003190404004292427, 'samples': 11123712, 'steps': 21725, 'loss/train': 2.2001075744628906} 02/25/2022 05:43:11 - INFO - codeparrot_training - Step 21726: {'lr': 0.00031902467413588134, 'samples': 11124224, 'steps': 21726, 'loss/train': 1.6967453956604004} 02/25/2022 05:43:15 - INFO - codeparrot_training - Step 21727: {'lr': 0.00031900894754684006, 'samples': 11124736, 'steps': 21727, 'loss/train': 1.6169955730438232} 02/25/2022 05:43:20 - INFO - codeparrot_training - Step 21728: {'lr': 0.0003189932206621865, 'samples': 11125248, 'steps': 21728, 'loss/train': 2.5277345180511475} 02/25/2022 05:43:23 - INFO - codeparrot_training - Step 21729: {'lr': 0.00031897749348198777, 'samples': 11125760, 'steps': 21729, 'loss/train': 1.373136281967163} 02/25/2022 05:43:31 - INFO - codeparrot_training - Step 21730: {'lr': 0.0003189617660063114, 'samples': 11126272, 'steps': 21730, 'loss/train': 2.139944553375244} 02/25/2022 05:43:34 - INFO - codeparrot_training - Step 21731: {'lr': 0.0003189460382352248, 'samples': 11126784, 'steps': 21731, 'loss/train': 2.2382009029388428} 02/25/2022 05:43:40 - INFO - codeparrot_training - Step 21732: {'lr': 0.00031893031016879515, 'samples': 11127296, 'steps': 21732, 'loss/train': 2.19052791595459} 02/25/2022 05:43:43 - INFO - codeparrot_training - Step 21733: {'lr': 0.00031891458180709003, 'samples': 11127808, 'steps': 21733, 'loss/train': 1.669534683227539} 02/25/2022 05:43:49 - INFO - codeparrot_training - Step 21734: {'lr': 0.0003188988531501766, 'samples': 11128320, 'steps': 21734, 'loss/train': 2.308239698410034} 02/25/2022 05:43:52 - INFO - codeparrot_training - Step 21735: {'lr': 0.00031888312419812255, 'samples': 11128832, 'steps': 21735, 'loss/train': 1.716054081916809} 02/25/2022 05:43:58 - INFO - codeparrot_training - Step 21736: {'lr': 0.00031886739495099494, 'samples': 11129344, 'steps': 21736, 'loss/train': 0.9718392491340637} 02/25/2022 05:44:01 - INFO - codeparrot_training - Step 21737: {'lr': 0.00031885166540886135, 'samples': 11129856, 'steps': 21737, 'loss/train': 1.8353155851364136} 02/25/2022 05:44:08 - INFO - codeparrot_training - Step 21738: {'lr': 0.000318835935571789, 'samples': 11130368, 'steps': 21738, 'loss/train': 1.8165950775146484} 02/25/2022 05:44:12 - INFO - codeparrot_training - Step 21739: {'lr': 0.0003188202054398454, 'samples': 11130880, 'steps': 21739, 'loss/train': 2.2578961849212646} 02/25/2022 05:44:17 - INFO - codeparrot_training - Step 21740: {'lr': 0.00031880447501309787, 'samples': 11131392, 'steps': 21740, 'loss/train': 2.956209182739258} 02/25/2022 05:44:21 - INFO - codeparrot_training - Step 21741: {'lr': 0.0003187887442916139, 'samples': 11131904, 'steps': 21741, 'loss/train': 0.956165075302124} 02/25/2022 05:44:26 - INFO - codeparrot_training - Step 21742: {'lr': 0.0003187730132754607, 'samples': 11132416, 'steps': 21742, 'loss/train': 1.874299168586731} 02/25/2022 05:44:30 - INFO - codeparrot_training - Step 21743: {'lr': 0.0003187572819647058, 'samples': 11132928, 'steps': 21743, 'loss/train': 3.6773388385772705} 02/25/2022 05:44:35 - INFO - codeparrot_training - Step 21744: {'lr': 0.00031874155035941656, 'samples': 11133440, 'steps': 21744, 'loss/train': 2.2933857440948486} 02/25/2022 05:44:39 - INFO - codeparrot_training - Step 21745: {'lr': 0.00031872581845966024, 'samples': 11133952, 'steps': 21745, 'loss/train': 0.9773945808410645} 02/25/2022 05:44:44 - INFO - codeparrot_training - Step 21746: {'lr': 0.0003187100862655044, 'samples': 11134464, 'steps': 21746, 'loss/train': 2.891629457473755} 02/25/2022 05:44:48 - INFO - codeparrot_training - Step 21747: {'lr': 0.00031869435377701637, 'samples': 11134976, 'steps': 21747, 'loss/train': 1.495321273803711} 02/25/2022 05:44:53 - INFO - codeparrot_training - Step 21748: {'lr': 0.0003186786209942636, 'samples': 11135488, 'steps': 21748, 'loss/train': 1.6084660291671753} 02/25/2022 05:44:57 - INFO - codeparrot_training - Step 21749: {'lr': 0.00031866288791731334, 'samples': 11136000, 'steps': 21749, 'loss/train': 1.2702771425247192} 02/25/2022 05:45:02 - INFO - codeparrot_training - Step 21750: {'lr': 0.0003186471545462331, 'samples': 11136512, 'steps': 21750, 'loss/train': 1.8765714168548584} 02/25/2022 05:45:06 - INFO - codeparrot_training - Step 21751: {'lr': 0.0003186314208810902, 'samples': 11137024, 'steps': 21751, 'loss/train': 1.386092185974121} 02/25/2022 05:45:12 - INFO - codeparrot_training - Step 21752: {'lr': 0.0003186156869219522, 'samples': 11137536, 'steps': 21752, 'loss/train': 1.234541416168213} 02/25/2022 05:45:15 - INFO - codeparrot_training - Step 21753: {'lr': 0.0003185999526688863, 'samples': 11138048, 'steps': 21753, 'loss/train': 2.1140732765197754} 02/25/2022 05:45:21 - INFO - codeparrot_training - Step 21754: {'lr': 0.00031858421812196, 'samples': 11138560, 'steps': 21754, 'loss/train': 1.5866442918777466} 02/25/2022 05:45:24 - INFO - codeparrot_training - Step 21755: {'lr': 0.0003185684832812407, 'samples': 11139072, 'steps': 21755, 'loss/train': 1.4904589653015137} 02/25/2022 05:45:32 - INFO - codeparrot_training - Step 21756: {'lr': 0.00031855274814679576, 'samples': 11139584, 'steps': 21756, 'loss/train': 2.090688467025757} 02/25/2022 05:45:35 - INFO - codeparrot_training - Step 21757: {'lr': 0.0003185370127186926, 'samples': 11140096, 'steps': 21757, 'loss/train': 2.320434331893921} 02/25/2022 05:45:41 - INFO - codeparrot_training - Step 21758: {'lr': 0.00031852127699699874, 'samples': 11140608, 'steps': 21758, 'loss/train': 1.4662401676177979} 02/25/2022 05:45:44 - INFO - codeparrot_training - Step 21759: {'lr': 0.0003185055409817814, 'samples': 11141120, 'steps': 21759, 'loss/train': 3.2484476566314697} 02/25/2022 05:45:50 - INFO - codeparrot_training - Step 21760: {'lr': 0.0003184898046731082, 'samples': 11141632, 'steps': 21760, 'loss/train': 2.333574056625366} 02/25/2022 05:45:53 - INFO - codeparrot_training - Step 21761: {'lr': 0.0003184740680710462, 'samples': 11142144, 'steps': 21761, 'loss/train': 2.1570136547088623} 02/25/2022 05:45:58 - INFO - codeparrot_training - Step 21762: {'lr': 0.00031845833117566326, 'samples': 11142656, 'steps': 21762, 'loss/train': 1.5060921907424927} 02/25/2022 05:46:02 - INFO - codeparrot_training - Step 21763: {'lr': 0.0003184425939870264, 'samples': 11143168, 'steps': 21763, 'loss/train': 2.389431953430176} 02/25/2022 05:46:07 - INFO - codeparrot_training - Step 21764: {'lr': 0.0003184268565052033, 'samples': 11143680, 'steps': 21764, 'loss/train': 1.9260852336883545} 02/25/2022 05:46:11 - INFO - codeparrot_training - Step 21765: {'lr': 0.00031841111873026124, 'samples': 11144192, 'steps': 21765, 'loss/train': 1.991213083267212} 02/25/2022 05:46:18 - INFO - codeparrot_training - Step 21766: {'lr': 0.0003183953806622677, 'samples': 11144704, 'steps': 21766, 'loss/train': 2.0757369995117188} 02/25/2022 05:46:22 - INFO - codeparrot_training - Step 21767: {'lr': 0.00031837964230129004, 'samples': 11145216, 'steps': 21767, 'loss/train': 1.8996134996414185} 02/25/2022 05:46:27 - INFO - codeparrot_training - Step 21768: {'lr': 0.0003183639036473957, 'samples': 11145728, 'steps': 21768, 'loss/train': 2.2683517932891846} 02/25/2022 05:46:30 - INFO - codeparrot_training - Step 21769: {'lr': 0.0003183481647006521, 'samples': 11146240, 'steps': 21769, 'loss/train': 1.8951406478881836} 02/25/2022 05:46:36 - INFO - codeparrot_training - Step 21770: {'lr': 0.0003183324254611267, 'samples': 11146752, 'steps': 21770, 'loss/train': 1.2873823642730713} 02/25/2022 05:46:39 - INFO - codeparrot_training - Step 21771: {'lr': 0.00031831668592888684, 'samples': 11147264, 'steps': 21771, 'loss/train': 1.0109927654266357} 02/25/2022 05:46:45 - INFO - codeparrot_training - Step 21772: {'lr': 0.0003183009461040001, 'samples': 11147776, 'steps': 21772, 'loss/train': 1.9222068786621094} 02/25/2022 05:46:48 - INFO - codeparrot_training - Step 21773: {'lr': 0.0003182852059865337, 'samples': 11148288, 'steps': 21773, 'loss/train': 2.1747031211853027} 02/25/2022 05:46:54 - INFO - codeparrot_training - Step 21774: {'lr': 0.0003182694655765551, 'samples': 11148800, 'steps': 21774, 'loss/train': 1.476946473121643} 02/25/2022 05:46:57 - INFO - codeparrot_training - Step 21775: {'lr': 0.00031825372487413186, 'samples': 11149312, 'steps': 21775, 'loss/train': 2.472576141357422} 02/25/2022 05:47:05 - INFO - codeparrot_training - Step 21776: {'lr': 0.00031823798387933133, 'samples': 11149824, 'steps': 21776, 'loss/train': 2.259316921234131} 02/25/2022 05:47:08 - INFO - codeparrot_training - Step 21777: {'lr': 0.00031822224259222095, 'samples': 11150336, 'steps': 21777, 'loss/train': 2.6426916122436523} 02/25/2022 05:47:14 - INFO - codeparrot_training - Step 21778: {'lr': 0.0003182065010128682, 'samples': 11150848, 'steps': 21778, 'loss/train': 1.7708532810211182} 02/25/2022 05:47:17 - INFO - codeparrot_training - Step 21779: {'lr': 0.0003181907591413403, 'samples': 11151360, 'steps': 21779, 'loss/train': 1.8966010808944702} 02/25/2022 05:47:23 - INFO - codeparrot_training - Step 21780: {'lr': 0.00031817501697770496, 'samples': 11151872, 'steps': 21780, 'loss/train': 1.9514433145523071} 02/25/2022 05:47:26 - INFO - codeparrot_training - Step 21781: {'lr': 0.00031815927452202955, 'samples': 11152384, 'steps': 21781, 'loss/train': 1.747356653213501} 02/25/2022 05:47:32 - INFO - codeparrot_training - Step 21782: {'lr': 0.0003181435317743813, 'samples': 11152896, 'steps': 21782, 'loss/train': 1.8831177949905396} 02/25/2022 05:47:37 - INFO - codeparrot_training - Step 21783: {'lr': 0.00031812778873482796, 'samples': 11153408, 'steps': 21783, 'loss/train': 1.084555983543396} 02/25/2022 05:47:41 - INFO - codeparrot_training - Step 21784: {'lr': 0.00031811204540343666, 'samples': 11153920, 'steps': 21784, 'loss/train': 1.1655285358428955} 02/25/2022 05:47:48 - INFO - codeparrot_training - Step 21785: {'lr': 0.00031809630178027506, 'samples': 11154432, 'steps': 21785, 'loss/train': 2.3911197185516357} 02/25/2022 05:47:51 - INFO - codeparrot_training - Step 21786: {'lr': 0.0003180805578654105, 'samples': 11154944, 'steps': 21786, 'loss/train': 1.0611038208007812} 02/25/2022 05:47:57 - INFO - codeparrot_training - Step 21787: {'lr': 0.0003180648136589105, 'samples': 11155456, 'steps': 21787, 'loss/train': 2.358529567718506} 02/25/2022 05:48:00 - INFO - codeparrot_training - Step 21788: {'lr': 0.00031804906916084235, 'samples': 11155968, 'steps': 21788, 'loss/train': 1.7644219398498535} 02/25/2022 05:48:04 - INFO - codeparrot_training - Step 21789: {'lr': 0.0003180333243712737, 'samples': 11156480, 'steps': 21789, 'loss/train': 1.7161561250686646} 02/25/2022 05:48:09 - INFO - codeparrot_training - Step 21790: {'lr': 0.00031801757929027187, 'samples': 11156992, 'steps': 21790, 'loss/train': 2.701889991760254} 02/25/2022 05:48:13 - INFO - codeparrot_training - Step 21791: {'lr': 0.0003180018339179043, 'samples': 11157504, 'steps': 21791, 'loss/train': 1.7461918592453003} 02/25/2022 05:48:19 - INFO - codeparrot_training - Step 21792: {'lr': 0.00031798608825423847, 'samples': 11158016, 'steps': 21792, 'loss/train': 2.433109998703003} 02/25/2022 05:48:23 - INFO - codeparrot_training - Step 21793: {'lr': 0.0003179703422993418, 'samples': 11158528, 'steps': 21793, 'loss/train': 1.8091360330581665} 02/25/2022 05:48:26 - INFO - codeparrot_training - Step 21794: {'lr': 0.00031795459605328183, 'samples': 11159040, 'steps': 21794, 'loss/train': 1.9812487363815308} 02/25/2022 05:48:32 - INFO - codeparrot_training - Step 21795: {'lr': 0.0003179388495161259, 'samples': 11159552, 'steps': 21795, 'loss/train': 1.5676612854003906} 02/25/2022 05:48:35 - INFO - codeparrot_training - Step 21796: {'lr': 0.00031792310268794155, 'samples': 11160064, 'steps': 21796, 'loss/train': 1.1732467412948608} 02/25/2022 05:48:43 - INFO - codeparrot_training - Step 21797: {'lr': 0.0003179073555687961, 'samples': 11160576, 'steps': 21797, 'loss/train': 1.2560960054397583} 02/25/2022 05:48:46 - INFO - codeparrot_training - Step 21798: {'lr': 0.00031789160815875724, 'samples': 11161088, 'steps': 21798, 'loss/train': 1.4300634860992432} 02/25/2022 05:48:52 - INFO - codeparrot_training - Step 21799: {'lr': 0.0003178758604578922, 'samples': 11161600, 'steps': 21799, 'loss/train': 1.5342192649841309} 02/25/2022 05:48:55 - INFO - codeparrot_training - Step 21800: {'lr': 0.00031786011246626855, 'samples': 11162112, 'steps': 21800, 'loss/train': 2.1601996421813965} 02/25/2022 05:49:01 - INFO - codeparrot_training - Step 21801: {'lr': 0.00031784436418395373, 'samples': 11162624, 'steps': 21801, 'loss/train': 0.8945356607437134} 02/25/2022 05:49:04 - INFO - codeparrot_training - Step 21802: {'lr': 0.0003178286156110152, 'samples': 11163136, 'steps': 21802, 'loss/train': 1.3482789993286133} 02/25/2022 05:49:09 - INFO - codeparrot_training - Step 21803: {'lr': 0.00031781286674752043, 'samples': 11163648, 'steps': 21803, 'loss/train': 2.4322028160095215} 02/25/2022 05:49:13 - INFO - codeparrot_training - Step 21804: {'lr': 0.00031779711759353683, 'samples': 11164160, 'steps': 21804, 'loss/train': 2.3651041984558105} 02/25/2022 05:49:18 - INFO - codeparrot_training - Step 21805: {'lr': 0.00031778136814913195, 'samples': 11164672, 'steps': 21805, 'loss/train': 2.624145030975342} 02/25/2022 05:49:22 - INFO - codeparrot_training - Step 21806: {'lr': 0.0003177656184143732, 'samples': 11165184, 'steps': 21806, 'loss/train': 1.7672096490859985} 02/25/2022 05:49:27 - INFO - codeparrot_training - Step 21807: {'lr': 0.0003177498683893281, 'samples': 11165696, 'steps': 21807, 'loss/train': 1.8487001657485962} 02/25/2022 05:49:31 - INFO - codeparrot_training - Step 21808: {'lr': 0.000317734118074064, 'samples': 11166208, 'steps': 21808, 'loss/train': 1.944106936454773} 02/25/2022 05:49:36 - INFO - codeparrot_training - Step 21809: {'lr': 0.00031771836746864854, 'samples': 11166720, 'steps': 21809, 'loss/train': 2.9717795848846436} 02/25/2022 05:49:40 - INFO - codeparrot_training - Step 21810: {'lr': 0.000317702616573149, 'samples': 11167232, 'steps': 21810, 'loss/train': 2.2071993350982666} 02/25/2022 05:49:45 - INFO - codeparrot_training - Step 21811: {'lr': 0.000317686865387633, 'samples': 11167744, 'steps': 21811, 'loss/train': 2.3968911170959473} 02/25/2022 05:49:53 - INFO - codeparrot_training - Step 21812: {'lr': 0.0003176711139121679, 'samples': 11168256, 'steps': 21812, 'loss/train': 0.9619163870811462} 02/25/2022 05:49:56 - INFO - codeparrot_training - Step 21813: {'lr': 0.00031765536214682134, 'samples': 11168768, 'steps': 21813, 'loss/train': 2.2722325325012207} 02/25/2022 05:50:02 - INFO - codeparrot_training - Step 21814: {'lr': 0.00031763961009166055, 'samples': 11169280, 'steps': 21814, 'loss/train': 1.8562426567077637} 02/25/2022 05:50:05 - INFO - codeparrot_training - Step 21815: {'lr': 0.00031762385774675324, 'samples': 11169792, 'steps': 21815, 'loss/train': 2.3930931091308594} 02/25/2022 05:50:11 - INFO - codeparrot_training - Step 21816: {'lr': 0.0003176081051121668, 'samples': 11170304, 'steps': 21816, 'loss/train': 1.8064790964126587} 02/25/2022 05:50:14 - INFO - codeparrot_training - Step 21817: {'lr': 0.0003175923521879687, 'samples': 11170816, 'steps': 21817, 'loss/train': 1.9096622467041016} 02/25/2022 05:50:20 - INFO - codeparrot_training - Step 21818: {'lr': 0.0003175765989742264, 'samples': 11171328, 'steps': 21818, 'loss/train': 3.021843910217285} 02/25/2022 05:50:23 - INFO - codeparrot_training - Step 21819: {'lr': 0.0003175608454710074, 'samples': 11171840, 'steps': 21819, 'loss/train': 1.4630122184753418} 02/25/2022 05:50:29 - INFO - codeparrot_training - Step 21820: {'lr': 0.00031754509167837927, 'samples': 11172352, 'steps': 21820, 'loss/train': 2.1704537868499756} 02/25/2022 05:50:32 - INFO - codeparrot_training - Step 21821: {'lr': 0.00031752933759640937, 'samples': 11172864, 'steps': 21821, 'loss/train': 1.9772887229919434} 02/25/2022 05:50:39 - INFO - codeparrot_training - Step 21822: {'lr': 0.0003175135832251652, 'samples': 11173376, 'steps': 21822, 'loss/train': 2.621196985244751} 02/25/2022 05:50:43 - INFO - codeparrot_training - Step 21823: {'lr': 0.00031749782856471426, 'samples': 11173888, 'steps': 21823, 'loss/train': 1.4463396072387695} 02/25/2022 05:50:48 - INFO - codeparrot_training - Step 21824: {'lr': 0.00031748207361512415, 'samples': 11174400, 'steps': 21824, 'loss/train': 1.6398717164993286} 02/25/2022 05:50:52 - INFO - codeparrot_training - Step 21825: {'lr': 0.00031746631837646216, 'samples': 11174912, 'steps': 21825, 'loss/train': 1.392532229423523} 02/25/2022 05:50:57 - INFO - codeparrot_training - Step 21826: {'lr': 0.000317450562848796, 'samples': 11175424, 'steps': 21826, 'loss/train': 1.6127662658691406} 02/25/2022 05:51:01 - INFO - codeparrot_training - Step 21827: {'lr': 0.00031743480703219293, 'samples': 11175936, 'steps': 21827, 'loss/train': 1.640582799911499} 02/25/2022 05:51:07 - INFO - codeparrot_training - Step 21828: {'lr': 0.00031741905092672057, 'samples': 11176448, 'steps': 21828, 'loss/train': 2.6265556812286377} 02/25/2022 05:51:10 - INFO - codeparrot_training - Step 21829: {'lr': 0.0003174032945324465, 'samples': 11176960, 'steps': 21829, 'loss/train': 2.2672958374023438} 02/25/2022 05:51:16 - INFO - codeparrot_training - Step 21830: {'lr': 0.00031738753784943803, 'samples': 11177472, 'steps': 21830, 'loss/train': 1.5970215797424316} 02/25/2022 05:51:19 - INFO - codeparrot_training - Step 21831: {'lr': 0.0003173717808777628, 'samples': 11177984, 'steps': 21831, 'loss/train': 1.926913857460022} 02/25/2022 05:51:27 - INFO - codeparrot_training - Step 21832: {'lr': 0.00031735602361748815, 'samples': 11178496, 'steps': 21832, 'loss/train': 3.0766592025756836} 02/25/2022 05:51:30 - INFO - codeparrot_training - Step 21833: {'lr': 0.00031734026606868184, 'samples': 11179008, 'steps': 21833, 'loss/train': 1.5882474184036255} 02/25/2022 05:51:36 - INFO - codeparrot_training - Step 21834: {'lr': 0.0003173245082314111, 'samples': 11179520, 'steps': 21834, 'loss/train': 0.5971845388412476} 02/25/2022 05:51:40 - INFO - codeparrot_training - Step 21835: {'lr': 0.0003173087501057436, 'samples': 11180032, 'steps': 21835, 'loss/train': 2.0837650299072266} 02/25/2022 05:51:43 - INFO - codeparrot_training - Step 21836: {'lr': 0.00031729299169174673, 'samples': 11180544, 'steps': 21836, 'loss/train': 1.2923085689544678} 02/25/2022 05:51:49 - INFO - codeparrot_training - Step 21837: {'lr': 0.0003172772329894882, 'samples': 11181056, 'steps': 21837, 'loss/train': 2.2904369831085205} 02/25/2022 05:51:52 - INFO - codeparrot_training - Step 21838: {'lr': 0.0003172614739990352, 'samples': 11181568, 'steps': 21838, 'loss/train': 1.8894999027252197} 02/25/2022 05:51:58 - INFO - codeparrot_training - Step 21839: {'lr': 0.0003172457147204554, 'samples': 11182080, 'steps': 21839, 'loss/train': 1.8371168375015259} 02/25/2022 05:52:01 - INFO - codeparrot_training - Step 21840: {'lr': 0.0003172299551538164, 'samples': 11182592, 'steps': 21840, 'loss/train': 1.8612170219421387} 02/25/2022 05:52:07 - INFO - codeparrot_training - Step 21841: {'lr': 0.0003172141952991856, 'samples': 11183104, 'steps': 21841, 'loss/train': 1.421238660812378} 02/25/2022 05:52:10 - INFO - codeparrot_training - Step 21842: {'lr': 0.00031719843515663055, 'samples': 11183616, 'steps': 21842, 'loss/train': 1.301273226737976} 02/25/2022 05:52:16 - INFO - codeparrot_training - Step 21843: {'lr': 0.0003171826747262187, 'samples': 11184128, 'steps': 21843, 'loss/train': 1.9404352903366089} 02/25/2022 05:52:19 - INFO - codeparrot_training - Step 21844: {'lr': 0.0003171669140080177, 'samples': 11184640, 'steps': 21844, 'loss/train': 1.9364114999771118} 02/25/2022 05:52:27 - INFO - codeparrot_training - Step 21845: {'lr': 0.00031715115300209477, 'samples': 11185152, 'steps': 21845, 'loss/train': 1.4868065118789673} 02/25/2022 05:52:30 - INFO - codeparrot_training - Step 21846: {'lr': 0.0003171353917085178, 'samples': 11185664, 'steps': 21846, 'loss/train': 1.7475389242172241} 02/25/2022 05:52:36 - INFO - codeparrot_training - Step 21847: {'lr': 0.00031711963012735414, 'samples': 11186176, 'steps': 21847, 'loss/train': 2.692734718322754} 02/25/2022 05:52:39 - INFO - codeparrot_training - Step 21848: {'lr': 0.0003171038682586712, 'samples': 11186688, 'steps': 21848, 'loss/train': 2.5753443241119385} 02/25/2022 05:52:45 - INFO - codeparrot_training - Step 21849: {'lr': 0.0003170881061025366, 'samples': 11187200, 'steps': 21849, 'loss/train': 2.5939724445343018} 02/25/2022 05:52:48 - INFO - codeparrot_training - Step 21850: {'lr': 0.00031707234365901786, 'samples': 11187712, 'steps': 21850, 'loss/train': 1.4615073204040527} 02/25/2022 05:52:54 - INFO - codeparrot_training - Step 21851: {'lr': 0.0003170565809281826, 'samples': 11188224, 'steps': 21851, 'loss/train': 1.6854385137557983} 02/25/2022 05:52:57 - INFO - codeparrot_training - Step 21852: {'lr': 0.0003170408179100981, 'samples': 11188736, 'steps': 21852, 'loss/train': 1.514777421951294} 02/25/2022 05:53:03 - INFO - codeparrot_training - Step 21853: {'lr': 0.0003170250546048321, 'samples': 11189248, 'steps': 21853, 'loss/train': 1.7585700750350952} 02/25/2022 05:53:06 - INFO - codeparrot_training - Step 21854: {'lr': 0.000317009291012452, 'samples': 11189760, 'steps': 21854, 'loss/train': 2.0186338424682617} 02/25/2022 05:53:12 - INFO - codeparrot_training - Step 21855: {'lr': 0.00031699352713302544, 'samples': 11190272, 'steps': 21855, 'loss/train': 2.488243341445923} 02/25/2022 05:53:15 - INFO - codeparrot_training - Step 21856: {'lr': 0.00031697776296661987, 'samples': 11190784, 'steps': 21856, 'loss/train': 1.6587148904800415} 02/25/2022 05:53:21 - INFO - codeparrot_training - Step 21857: {'lr': 0.0003169619985133028, 'samples': 11191296, 'steps': 21857, 'loss/train': 1.4487026929855347} 02/25/2022 05:53:24 - INFO - codeparrot_training - Step 21858: {'lr': 0.0003169462337731418, 'samples': 11191808, 'steps': 21858, 'loss/train': 1.4821003675460815} 02/25/2022 05:53:32 - INFO - codeparrot_training - Step 21859: {'lr': 0.0003169304687462044, 'samples': 11192320, 'steps': 21859, 'loss/train': 1.9078369140625} 02/25/2022 05:53:35 - INFO - codeparrot_training - Step 21860: {'lr': 0.00031691470343255814, 'samples': 11192832, 'steps': 21860, 'loss/train': 2.3228437900543213} 02/25/2022 05:53:41 - INFO - codeparrot_training - Step 21861: {'lr': 0.00031689893783227053, 'samples': 11193344, 'steps': 21861, 'loss/train': 1.3735699653625488} 02/25/2022 05:53:44 - INFO - codeparrot_training - Step 21862: {'lr': 0.00031688317194540904, 'samples': 11193856, 'steps': 21862, 'loss/train': 3.6704025268554688} 02/25/2022 05:53:50 - INFO - codeparrot_training - Step 21863: {'lr': 0.0003168674057720413, 'samples': 11194368, 'steps': 21863, 'loss/train': 1.6292989253997803} 02/25/2022 05:53:53 - INFO - codeparrot_training - Step 21864: {'lr': 0.0003168516393122349, 'samples': 11194880, 'steps': 21864, 'loss/train': 0.9050735831260681} 02/25/2022 05:53:59 - INFO - codeparrot_training - Step 21865: {'lr': 0.0003168358725660573, 'samples': 11195392, 'steps': 21865, 'loss/train': 2.204618215560913} 02/25/2022 05:54:02 - INFO - codeparrot_training - Step 21866: {'lr': 0.000316820105533576, 'samples': 11195904, 'steps': 21866, 'loss/train': 0.9024485349655151} 02/25/2022 05:54:08 - INFO - codeparrot_training - Step 21867: {'lr': 0.0003168043382148586, 'samples': 11196416, 'steps': 21867, 'loss/train': 2.4066126346588135} 02/25/2022 05:54:11 - INFO - codeparrot_training - Step 21868: {'lr': 0.0003167885706099726, 'samples': 11196928, 'steps': 21868, 'loss/train': 1.3682433366775513} 02/25/2022 05:54:17 - INFO - codeparrot_training - Step 21869: {'lr': 0.0003167728027189856, 'samples': 11197440, 'steps': 21869, 'loss/train': 2.008517026901245} 02/25/2022 05:54:20 - INFO - codeparrot_training - Step 21870: {'lr': 0.00031675703454196513, 'samples': 11197952, 'steps': 21870, 'loss/train': 1.7025039196014404} 02/25/2022 05:54:27 - INFO - codeparrot_training - Step 21871: {'lr': 0.00031674126607897867, 'samples': 11198464, 'steps': 21871, 'loss/train': 2.643584728240967} 02/25/2022 05:54:31 - INFO - codeparrot_training - Step 21872: {'lr': 0.00031672549733009395, 'samples': 11198976, 'steps': 21872, 'loss/train': 1.5366795063018799} 02/25/2022 05:54:36 - INFO - codeparrot_training - Step 21873: {'lr': 0.00031670972829537825, 'samples': 11199488, 'steps': 21873, 'loss/train': 2.6930928230285645} 02/25/2022 05:54:40 - INFO - codeparrot_training - Step 21874: {'lr': 0.0003166939589748993, 'samples': 11200000, 'steps': 21874, 'loss/train': 2.4976396560668945} 02/25/2022 05:54:45 - INFO - codeparrot_training - Step 21875: {'lr': 0.0003166781893687246, 'samples': 11200512, 'steps': 21875, 'loss/train': 1.1043554544448853} 02/25/2022 05:54:49 - INFO - codeparrot_training - Step 21876: {'lr': 0.00031666241947692173, 'samples': 11201024, 'steps': 21876, 'loss/train': 1.8815280199050903} 02/25/2022 05:54:54 - INFO - codeparrot_training - Step 21877: {'lr': 0.0003166466492995582, 'samples': 11201536, 'steps': 21877, 'loss/train': 2.5244462490081787} 02/25/2022 05:54:58 - INFO - codeparrot_training - Step 21878: {'lr': 0.0003166308788367016, 'samples': 11202048, 'steps': 21878, 'loss/train': 1.9727551937103271} 02/25/2022 05:55:03 - INFO - codeparrot_training - Step 21879: {'lr': 0.00031661510808841947, 'samples': 11202560, 'steps': 21879, 'loss/train': 1.600012183189392} 02/25/2022 05:55:07 - INFO - codeparrot_training - Step 21880: {'lr': 0.0003165993370547794, 'samples': 11203072, 'steps': 21880, 'loss/train': 1.368313193321228} 02/25/2022 05:55:13 - INFO - codeparrot_training - Step 21881: {'lr': 0.0003165835657358489, 'samples': 11203584, 'steps': 21881, 'loss/train': 1.874405860900879} 02/25/2022 05:55:17 - INFO - codeparrot_training - Step 21882: {'lr': 0.00031656779413169543, 'samples': 11204096, 'steps': 21882, 'loss/train': 2.786487102508545} 02/25/2022 05:55:22 - INFO - codeparrot_training - Step 21883: {'lr': 0.00031655202224238686, 'samples': 11204608, 'steps': 21883, 'loss/train': 2.423970937728882} 02/25/2022 05:55:26 - INFO - codeparrot_training - Step 21884: {'lr': 0.0003165362500679905, 'samples': 11205120, 'steps': 21884, 'loss/train': 0.9125109910964966} 02/25/2022 05:55:31 - INFO - codeparrot_training - Step 21885: {'lr': 0.00031652047760857393, 'samples': 11205632, 'steps': 21885, 'loss/train': 1.6199257373809814} 02/25/2022 05:55:35 - INFO - codeparrot_training - Step 21886: {'lr': 0.0003165047048642047, 'samples': 11206144, 'steps': 21886, 'loss/train': 2.3350679874420166} 02/25/2022 05:55:40 - INFO - codeparrot_training - Step 21887: {'lr': 0.00031648893183495053, 'samples': 11206656, 'steps': 21887, 'loss/train': 1.8431631326675415} 02/25/2022 05:55:44 - INFO - codeparrot_training - Step 21888: {'lr': 0.0003164731585208789, 'samples': 11207168, 'steps': 21888, 'loss/train': 1.8499499559402466} 02/25/2022 05:55:49 - INFO - codeparrot_training - Step 21889: {'lr': 0.00031645738492205736, 'samples': 11207680, 'steps': 21889, 'loss/train': 1.6582543849945068} 02/25/2022 05:55:53 - INFO - codeparrot_training - Step 21890: {'lr': 0.0003164416110385534, 'samples': 11208192, 'steps': 21890, 'loss/train': 2.1935536861419678} 02/25/2022 05:55:59 - INFO - codeparrot_training - Step 21891: {'lr': 0.0003164258368704347, 'samples': 11208704, 'steps': 21891, 'loss/train': 5.17179012298584} 02/25/2022 05:56:03 - INFO - codeparrot_training - Step 21892: {'lr': 0.00031641006241776886, 'samples': 11209216, 'steps': 21892, 'loss/train': 2.0094187259674072} 02/25/2022 05:56:08 - INFO - codeparrot_training - Step 21893: {'lr': 0.0003163942876806234, 'samples': 11209728, 'steps': 21893, 'loss/train': 1.6515058279037476} 02/25/2022 05:56:12 - INFO - codeparrot_training - Step 21894: {'lr': 0.00031637851265906594, 'samples': 11210240, 'steps': 21894, 'loss/train': 1.353257417678833} 02/25/2022 05:56:17 - INFO - codeparrot_training - Step 21895: {'lr': 0.0003163627373531639, 'samples': 11210752, 'steps': 21895, 'loss/train': 2.082242965698242} 02/25/2022 05:56:21 - INFO - codeparrot_training - Step 21896: {'lr': 0.000316346961762985, 'samples': 11211264, 'steps': 21896, 'loss/train': 1.565011978149414} 02/25/2022 05:56:26 - INFO - codeparrot_training - Step 21897: {'lr': 0.00031633118588859677, 'samples': 11211776, 'steps': 21897, 'loss/train': 2.1991031169891357} 02/25/2022 05:56:30 - INFO - codeparrot_training - Step 21898: {'lr': 0.00031631540973006683, 'samples': 11212288, 'steps': 21898, 'loss/train': 1.5421130657196045} 02/25/2022 05:56:35 - INFO - codeparrot_training - Step 21899: {'lr': 0.0003162996332874627, 'samples': 11212800, 'steps': 21899, 'loss/train': 1.5429047346115112} 02/25/2022 05:56:39 - INFO - codeparrot_training - Step 21900: {'lr': 0.000316283856560852, 'samples': 11213312, 'steps': 21900, 'loss/train': 1.492174744606018} 02/25/2022 05:56:45 - INFO - codeparrot_training - Step 21901: {'lr': 0.00031626807955030236, 'samples': 11213824, 'steps': 21901, 'loss/train': 0.4489256739616394} 02/25/2022 05:56:48 - INFO - codeparrot_training - Step 21902: {'lr': 0.00031625230225588123, 'samples': 11214336, 'steps': 21902, 'loss/train': 1.5887861251831055} 02/25/2022 05:56:54 - INFO - codeparrot_training - Step 21903: {'lr': 0.0003162365246776564, 'samples': 11214848, 'steps': 21903, 'loss/train': 2.179230213165283} 02/25/2022 05:56:57 - INFO - codeparrot_training - Step 21904: {'lr': 0.0003162207468156952, 'samples': 11215360, 'steps': 21904, 'loss/train': 2.1936845779418945} 02/25/2022 05:57:03 - INFO - codeparrot_training - Step 21905: {'lr': 0.00031620496867006543, 'samples': 11215872, 'steps': 21905, 'loss/train': 1.948491096496582} 02/25/2022 05:57:06 - INFO - codeparrot_training - Step 21906: {'lr': 0.0003161891902408345, 'samples': 11216384, 'steps': 21906, 'loss/train': 1.604262351989746} 02/25/2022 05:57:13 - INFO - codeparrot_training - Step 21907: {'lr': 0.00031617341152807024, 'samples': 11216896, 'steps': 21907, 'loss/train': 1.1284022331237793} 02/25/2022 05:57:16 - INFO - codeparrot_training - Step 21908: {'lr': 0.00031615763253183996, 'samples': 11217408, 'steps': 21908, 'loss/train': 1.889078140258789} 02/25/2022 05:57:22 - INFO - codeparrot_training - Step 21909: {'lr': 0.00031614185325221143, 'samples': 11217920, 'steps': 21909, 'loss/train': 2.71724271774292} 02/25/2022 05:57:25 - INFO - codeparrot_training - Step 21910: {'lr': 0.0003161260736892523, 'samples': 11218432, 'steps': 21910, 'loss/train': 2.512601613998413} 02/25/2022 05:57:31 - INFO - codeparrot_training - Step 21911: {'lr': 0.00031611029384302997, 'samples': 11218944, 'steps': 21911, 'loss/train': 1.430791974067688} 02/25/2022 05:57:34 - INFO - codeparrot_training - Step 21912: {'lr': 0.0003160945137136121, 'samples': 11219456, 'steps': 21912, 'loss/train': 1.1917078495025635} 02/25/2022 05:57:40 - INFO - codeparrot_training - Step 21913: {'lr': 0.0003160787333010664, 'samples': 11219968, 'steps': 21913, 'loss/train': 1.4198331832885742} 02/25/2022 05:57:43 - INFO - codeparrot_training - Step 21914: {'lr': 0.00031606295260546037, 'samples': 11220480, 'steps': 21914, 'loss/train': 1.0272213220596313} 02/25/2022 05:57:49 - INFO - codeparrot_training - Step 21915: {'lr': 0.00031604717162686156, 'samples': 11220992, 'steps': 21915, 'loss/train': 1.8996394872665405} 02/25/2022 05:57:52 - INFO - codeparrot_training - Step 21916: {'lr': 0.00031603139036533775, 'samples': 11221504, 'steps': 21916, 'loss/train': 1.981220006942749} 02/25/2022 05:57:58 - INFO - codeparrot_training - Step 21917: {'lr': 0.0003160156088209564, 'samples': 11222016, 'steps': 21917, 'loss/train': 1.6076395511627197} 02/25/2022 05:58:02 - INFO - codeparrot_training - Step 21918: {'lr': 0.0003159998269937851, 'samples': 11222528, 'steps': 21918, 'loss/train': 1.584609031677246} 02/25/2022 05:58:07 - INFO - codeparrot_training - Step 21919: {'lr': 0.0003159840448838915, 'samples': 11223040, 'steps': 21919, 'loss/train': 2.040466547012329} 02/25/2022 05:58:11 - INFO - codeparrot_training - Step 21920: {'lr': 0.00031596826249134324, 'samples': 11223552, 'steps': 21920, 'loss/train': 2.3362655639648438} 02/25/2022 05:58:17 - INFO - codeparrot_training - Step 21921: {'lr': 0.0003159524798162079, 'samples': 11224064, 'steps': 21921, 'loss/train': 1.411509394645691} 02/25/2022 05:58:20 - INFO - codeparrot_training - Step 21922: {'lr': 0.000315936696858553, 'samples': 11224576, 'steps': 21922, 'loss/train': 1.3273935317993164} 02/25/2022 05:58:26 - INFO - codeparrot_training - Step 21923: {'lr': 0.00031592091361844633, 'samples': 11225088, 'steps': 21923, 'loss/train': 1.3492838144302368} 02/25/2022 05:58:29 - INFO - codeparrot_training - Step 21924: {'lr': 0.0003159051300959553, 'samples': 11225600, 'steps': 21924, 'loss/train': 1.7265515327453613} 02/25/2022 05:58:35 - INFO - codeparrot_training - Step 21925: {'lr': 0.0003158893462911477, 'samples': 11226112, 'steps': 21925, 'loss/train': 1.8900066614151} 02/25/2022 05:58:38 - INFO - codeparrot_training - Step 21926: {'lr': 0.00031587356220409105, 'samples': 11226624, 'steps': 21926, 'loss/train': 1.0663362741470337} 02/25/2022 05:58:44 - INFO - codeparrot_training - Step 21927: {'lr': 0.0003158577778348529, 'samples': 11227136, 'steps': 21927, 'loss/train': 1.4324028491973877} 02/25/2022 05:58:48 - INFO - codeparrot_training - Step 21928: {'lr': 0.000315841993183501, 'samples': 11227648, 'steps': 21928, 'loss/train': 2.416149616241455} 02/25/2022 05:58:53 - INFO - codeparrot_training - Step 21929: {'lr': 0.000315826208250103, 'samples': 11228160, 'steps': 21929, 'loss/train': 1.052208662033081} 02/25/2022 05:58:57 - INFO - codeparrot_training - Step 21930: {'lr': 0.0003158104230347263, 'samples': 11228672, 'steps': 21930, 'loss/train': 2.507903575897217} 02/25/2022 05:59:02 - INFO - codeparrot_training - Step 21931: {'lr': 0.0003157946375374387, 'samples': 11229184, 'steps': 21931, 'loss/train': 1.5963064432144165} 02/25/2022 05:59:06 - INFO - codeparrot_training - Step 21932: {'lr': 0.0003157788517583077, 'samples': 11229696, 'steps': 21932, 'loss/train': 1.5115591287612915} 02/25/2022 05:59:11 - INFO - codeparrot_training - Step 21933: {'lr': 0.00031576306569740107, 'samples': 11230208, 'steps': 21933, 'loss/train': 1.7534328699111938} 02/25/2022 05:59:15 - INFO - codeparrot_training - Step 21934: {'lr': 0.00031574727935478633, 'samples': 11230720, 'steps': 21934, 'loss/train': 2.5700552463531494} 02/25/2022 05:59:20 - INFO - codeparrot_training - Step 21935: {'lr': 0.0003157314927305311, 'samples': 11231232, 'steps': 21935, 'loss/train': 2.3193955421447754} 02/25/2022 05:59:24 - INFO - codeparrot_training - Step 21936: {'lr': 0.00031571570582470304, 'samples': 11231744, 'steps': 21936, 'loss/train': 2.091813564300537} 02/25/2022 05:59:30 - INFO - codeparrot_training - Step 21937: {'lr': 0.00031569991863736975, 'samples': 11232256, 'steps': 21937, 'loss/train': 0.7968567609786987} 02/25/2022 05:59:34 - INFO - codeparrot_training - Step 21938: {'lr': 0.00031568413116859887, 'samples': 11232768, 'steps': 21938, 'loss/train': 2.351186990737915} 02/25/2022 05:59:39 - INFO - codeparrot_training - Step 21939: {'lr': 0.000315668343418458, 'samples': 11233280, 'steps': 21939, 'loss/train': 1.532464861869812} 02/25/2022 05:59:43 - INFO - codeparrot_training - Step 21940: {'lr': 0.0003156525553870149, 'samples': 11233792, 'steps': 21940, 'loss/train': 1.3910305500030518} 02/25/2022 05:59:46 - INFO - codeparrot_training - Step 21941: {'lr': 0.000315636767074337, 'samples': 11234304, 'steps': 21941, 'loss/train': 2.2487399578094482} 02/25/2022 05:59:52 - INFO - codeparrot_training - Step 21942: {'lr': 0.00031562097848049207, 'samples': 11234816, 'steps': 21942, 'loss/train': 1.386334776878357} 02/25/2022 05:59:55 - INFO - codeparrot_training - Step 21943: {'lr': 0.0003156051896055477, 'samples': 11235328, 'steps': 21943, 'loss/train': 1.9307063817977905} 02/25/2022 06:00:01 - INFO - codeparrot_training - Step 21944: {'lr': 0.0003155894004495716, 'samples': 11235840, 'steps': 21944, 'loss/train': 2.247981071472168} 02/25/2022 06:00:04 - INFO - codeparrot_training - Step 21945: {'lr': 0.0003155736110126312, 'samples': 11236352, 'steps': 21945, 'loss/train': 1.0795257091522217} 02/25/2022 06:00:10 - INFO - codeparrot_training - Step 21946: {'lr': 0.0003155578212947944, 'samples': 11236864, 'steps': 21946, 'loss/train': 0.4491962492465973} 02/25/2022 06:00:13 - INFO - codeparrot_training - Step 21947: {'lr': 0.0003155420312961286, 'samples': 11237376, 'steps': 21947, 'loss/train': 2.270949602127075} 02/25/2022 06:00:19 - INFO - codeparrot_training - Step 21948: {'lr': 0.00031552624101670155, 'samples': 11237888, 'steps': 21948, 'loss/train': 1.7403013706207275} 02/25/2022 06:00:23 - INFO - codeparrot_training - Step 21949: {'lr': 0.000315510450456581, 'samples': 11238400, 'steps': 21949, 'loss/train': 1.7849270105361938} 02/25/2022 06:00:28 - INFO - codeparrot_training - Step 21950: {'lr': 0.0003154946596158343, 'samples': 11238912, 'steps': 21950, 'loss/train': 2.225351572036743} 02/25/2022 06:00:32 - INFO - codeparrot_training - Step 21951: {'lr': 0.0003154788684945295, 'samples': 11239424, 'steps': 21951, 'loss/train': 1.7404541969299316} 02/25/2022 06:00:38 - INFO - codeparrot_training - Step 21952: {'lr': 0.00031546307709273393, 'samples': 11239936, 'steps': 21952, 'loss/train': 1.300930142402649} 02/25/2022 06:00:41 - INFO - codeparrot_training - Step 21953: {'lr': 0.0003154472854105153, 'samples': 11240448, 'steps': 21953, 'loss/train': 0.9668514132499695} 02/25/2022 06:00:47 - INFO - codeparrot_training - Step 21954: {'lr': 0.0003154314934479413, 'samples': 11240960, 'steps': 21954, 'loss/train': 3.051269292831421} 02/25/2022 06:00:50 - INFO - codeparrot_training - Step 21955: {'lr': 0.00031541570120507956, 'samples': 11241472, 'steps': 21955, 'loss/train': 1.9547733068466187} 02/25/2022 06:00:56 - INFO - codeparrot_training - Step 21956: {'lr': 0.0003153999086819977, 'samples': 11241984, 'steps': 21956, 'loss/train': 1.877734899520874} 02/25/2022 06:00:59 - INFO - codeparrot_training - Step 21957: {'lr': 0.00031538411587876344, 'samples': 11242496, 'steps': 21957, 'loss/train': 1.4035985469818115} 02/25/2022 06:01:05 - INFO - codeparrot_training - Step 21958: {'lr': 0.0003153683227954443, 'samples': 11243008, 'steps': 21958, 'loss/train': 1.8795199394226074} 02/25/2022 06:01:08 - INFO - codeparrot_training - Step 21959: {'lr': 0.00031535252943210815, 'samples': 11243520, 'steps': 21959, 'loss/train': 1.7047089338302612} 02/25/2022 06:01:14 - INFO - codeparrot_training - Step 21960: {'lr': 0.0003153367357888224, 'samples': 11244032, 'steps': 21960, 'loss/train': 2.137213945388794} 02/25/2022 06:01:17 - INFO - codeparrot_training - Step 21961: {'lr': 0.00031532094186565487, 'samples': 11244544, 'steps': 21961, 'loss/train': 1.8743616342544556} 02/25/2022 06:01:23 - INFO - codeparrot_training - Step 21962: {'lr': 0.0003153051476626732, 'samples': 11245056, 'steps': 21962, 'loss/train': 1.7010562419891357} 02/25/2022 06:01:26 - INFO - codeparrot_training - Step 21963: {'lr': 0.00031528935317994493, 'samples': 11245568, 'steps': 21963, 'loss/train': 1.934483289718628} 02/25/2022 06:01:33 - INFO - codeparrot_training - Step 21964: {'lr': 0.00031527355841753793, 'samples': 11246080, 'steps': 21964, 'loss/train': 2.77974271774292} 02/25/2022 06:01:36 - INFO - codeparrot_training - Step 21965: {'lr': 0.0003152577633755196, 'samples': 11246592, 'steps': 21965, 'loss/train': 2.4961276054382324} 02/25/2022 06:01:42 - INFO - codeparrot_training - Step 21966: {'lr': 0.0003152419680539578, 'samples': 11247104, 'steps': 21966, 'loss/train': 2.3780453205108643} 02/25/2022 06:01:45 - INFO - codeparrot_training - Step 21967: {'lr': 0.0003152261724529201, 'samples': 11247616, 'steps': 21967, 'loss/train': 2.140078544616699} 02/25/2022 06:01:51 - INFO - codeparrot_training - Step 21968: {'lr': 0.0003152103765724743, 'samples': 11248128, 'steps': 21968, 'loss/train': 1.5463610887527466} 02/25/2022 06:01:54 - INFO - codeparrot_training - Step 21969: {'lr': 0.0003151945804126878, 'samples': 11248640, 'steps': 21969, 'loss/train': 1.4795680046081543} 02/25/2022 06:02:00 - INFO - codeparrot_training - Step 21970: {'lr': 0.00031517878397362847, 'samples': 11249152, 'steps': 21970, 'loss/train': 1.890347957611084} 02/25/2022 06:02:03 - INFO - codeparrot_training - Step 21971: {'lr': 0.00031516298725536396, 'samples': 11249664, 'steps': 21971, 'loss/train': 1.8169739246368408} 02/25/2022 06:02:09 - INFO - codeparrot_training - Step 21972: {'lr': 0.00031514719025796183, 'samples': 11250176, 'steps': 21972, 'loss/train': 2.1855034828186035} 02/25/2022 06:02:15 - INFO - codeparrot_training - Step 21973: {'lr': 0.0003151313929814899, 'samples': 11250688, 'steps': 21973, 'loss/train': 1.1833609342575073} 02/25/2022 06:02:18 - INFO - codeparrot_training - Step 21974: {'lr': 0.0003151155954260157, 'samples': 11251200, 'steps': 21974, 'loss/train': 2.4833056926727295} 02/25/2022 06:02:24 - INFO - codeparrot_training - Step 21975: {'lr': 0.00031509979759160707, 'samples': 11251712, 'steps': 21975, 'loss/train': 0.6434096097946167} 02/25/2022 06:02:27 - INFO - codeparrot_training - Step 21976: {'lr': 0.00031508399947833155, 'samples': 11252224, 'steps': 21976, 'loss/train': 2.0098721981048584} 02/25/2022 06:02:33 - INFO - codeparrot_training - Step 21977: {'lr': 0.0003150682010862568, 'samples': 11252736, 'steps': 21977, 'loss/train': 1.5231654644012451} 02/25/2022 06:02:37 - INFO - codeparrot_training - Step 21978: {'lr': 0.00031505240241545054, 'samples': 11253248, 'steps': 21978, 'loss/train': 2.677313804626465} 02/25/2022 06:02:40 - INFO - codeparrot_training - Step 21979: {'lr': 0.0003150366034659804, 'samples': 11253760, 'steps': 21979, 'loss/train': 1.5920900106430054} 02/25/2022 06:02:46 - INFO - codeparrot_training - Step 21980: {'lr': 0.00031502080423791417, 'samples': 11254272, 'steps': 21980, 'loss/train': 2.176215887069702} 02/25/2022 06:02:49 - INFO - codeparrot_training - Step 21981: {'lr': 0.00031500500473131943, 'samples': 11254784, 'steps': 21981, 'loss/train': 2.0931975841522217} 02/25/2022 06:02:55 - INFO - codeparrot_training - Step 21982: {'lr': 0.00031498920494626396, 'samples': 11255296, 'steps': 21982, 'loss/train': 1.997752070426941} 02/25/2022 06:02:58 - INFO - codeparrot_training - Step 21983: {'lr': 0.00031497340488281534, 'samples': 11255808, 'steps': 21983, 'loss/train': 1.963989496231079} 02/25/2022 06:03:04 - INFO - codeparrot_training - Step 21984: {'lr': 0.0003149576045410412, 'samples': 11256320, 'steps': 21984, 'loss/train': 0.0632946640253067} 02/25/2022 06:03:08 - INFO - codeparrot_training - Step 21985: {'lr': 0.00031494180392100936, 'samples': 11256832, 'steps': 21985, 'loss/train': 1.512750267982483} 02/25/2022 06:03:13 - INFO - codeparrot_training - Step 21986: {'lr': 0.00031492600302278746, 'samples': 11257344, 'steps': 21986, 'loss/train': 1.5110458135604858} 02/25/2022 06:03:17 - INFO - codeparrot_training - Step 21987: {'lr': 0.00031491020184644316, 'samples': 11257856, 'steps': 21987, 'loss/train': 2.3577587604522705} 02/25/2022 06:03:22 - INFO - codeparrot_training - Step 21988: {'lr': 0.0003148944003920442, 'samples': 11258368, 'steps': 21988, 'loss/train': 1.7468795776367188} 02/25/2022 06:03:26 - INFO - codeparrot_training - Step 21989: {'lr': 0.00031487859865965824, 'samples': 11258880, 'steps': 21989, 'loss/train': 0.12696422636508942} 02/25/2022 06:03:32 - INFO - codeparrot_training - Step 21990: {'lr': 0.000314862796649353, 'samples': 11259392, 'steps': 21990, 'loss/train': 1.486364483833313} 02/25/2022 06:03:35 - INFO - codeparrot_training - Step 21991: {'lr': 0.000314846994361196, 'samples': 11259904, 'steps': 21991, 'loss/train': 2.426990032196045} 02/25/2022 06:03:41 - INFO - codeparrot_training - Step 21992: {'lr': 0.0003148311917952552, 'samples': 11260416, 'steps': 21992, 'loss/train': 1.7946034669876099} 02/25/2022 06:03:44 - INFO - codeparrot_training - Step 21993: {'lr': 0.00031481538895159806, 'samples': 11260928, 'steps': 21993, 'loss/train': 0.2279038429260254} 02/25/2022 06:03:50 - INFO - codeparrot_training - Step 21994: {'lr': 0.00031479958583029247, 'samples': 11261440, 'steps': 21994, 'loss/train': 2.484208345413208} 02/25/2022 06:03:53 - INFO - codeparrot_training - Step 21995: {'lr': 0.0003147837824314059, 'samples': 11261952, 'steps': 21995, 'loss/train': 2.4712982177734375} 02/25/2022 06:03:59 - INFO - codeparrot_training - Step 21996: {'lr': 0.00031476797875500627, 'samples': 11262464, 'steps': 21996, 'loss/train': 6.382114887237549} 02/25/2022 06:04:03 - INFO - codeparrot_training - Step 21997: {'lr': 0.00031475217480116124, 'samples': 11262976, 'steps': 21997, 'loss/train': 3.1666152477264404} 02/25/2022 06:04:08 - INFO - codeparrot_training - Step 21998: {'lr': 0.00031473637056993837, 'samples': 11263488, 'steps': 21998, 'loss/train': 2.339315176010132} 02/25/2022 06:04:12 - INFO - codeparrot_training - Step 21999: {'lr': 0.0003147205660614055, 'samples': 11264000, 'steps': 21999, 'loss/train': 2.7713630199432373} 02/25/2022 06:04:12 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 06:04:30 - WARNING - huggingface_hub.repository - Several commits (22) will be pushed upstream. 02/25/2022 06:04:30 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 06:05:08 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 1098822..de93bae floral-grass-11 -> floral-grass-11 02/25/2022 06:05:16 - INFO - codeparrot_training - Step 22000: {'lr': 0.00031470476127563017, 'samples': 11264512, 'steps': 22000, 'loss/train': 2.5917716026306152} 02/25/2022 06:05:19 - INFO - codeparrot_training - Step 22001: {'lr': 0.00031468895621268036, 'samples': 11265024, 'steps': 22001, 'loss/train': 2.364344358444214} 02/25/2022 06:05:25 - INFO - codeparrot_training - Step 22002: {'lr': 0.00031467315087262344, 'samples': 11265536, 'steps': 22002, 'loss/train': 2.4098546504974365} 02/25/2022 06:05:28 - INFO - codeparrot_training - Step 22003: {'lr': 0.0003146573452555274, 'samples': 11266048, 'steps': 22003, 'loss/train': 1.268041968345642} 02/25/2022 06:05:34 - INFO - codeparrot_training - Step 22004: {'lr': 0.0003146415393614597, 'samples': 11266560, 'steps': 22004, 'loss/train': 1.6502779722213745} 02/25/2022 06:05:37 - INFO - codeparrot_training - Step 22005: {'lr': 0.00031462573319048827, 'samples': 11267072, 'steps': 22005, 'loss/train': 2.3571839332580566} 02/25/2022 06:05:43 - INFO - codeparrot_training - Step 22006: {'lr': 0.0003146099267426806, 'samples': 11267584, 'steps': 22006, 'loss/train': 1.2613369226455688} 02/25/2022 06:05:46 - INFO - codeparrot_training - Step 22007: {'lr': 0.00031459412001810474, 'samples': 11268096, 'steps': 22007, 'loss/train': 1.7785935401916504} 02/25/2022 06:05:52 - INFO - codeparrot_training - Step 22008: {'lr': 0.000314578313016828, 'samples': 11268608, 'steps': 22008, 'loss/train': 2.1543264389038086} 02/25/2022 06:05:55 - INFO - codeparrot_training - Step 22009: {'lr': 0.00031456250573891825, 'samples': 11269120, 'steps': 22009, 'loss/train': 1.5843356847763062} 02/25/2022 06:06:02 - INFO - codeparrot_training - Step 22010: {'lr': 0.0003145466981844434, 'samples': 11269632, 'steps': 22010, 'loss/train': 2.041003465652466} 02/25/2022 06:06:05 - INFO - codeparrot_training - Step 22011: {'lr': 0.00031453089035347084, 'samples': 11270144, 'steps': 22011, 'loss/train': 1.2788344621658325} 02/25/2022 06:06:11 - INFO - codeparrot_training - Step 22012: {'lr': 0.0003145150822460685, 'samples': 11270656, 'steps': 22012, 'loss/train': 1.6295868158340454} 02/25/2022 06:06:15 - INFO - codeparrot_training - Step 22013: {'lr': 0.00031449927386230397, 'samples': 11271168, 'steps': 22013, 'loss/train': 2.3790078163146973} 02/25/2022 06:06:18 - INFO - codeparrot_training - Step 22014: {'lr': 0.0003144834652022451, 'samples': 11271680, 'steps': 22014, 'loss/train': 2.0947465896606445} 02/25/2022 06:06:24 - INFO - codeparrot_training - Step 22015: {'lr': 0.0003144676562659595, 'samples': 11272192, 'steps': 22015, 'loss/train': 0.5828628540039062} 02/25/2022 06:06:27 - INFO - codeparrot_training - Step 22016: {'lr': 0.000314451847053515, 'samples': 11272704, 'steps': 22016, 'loss/train': 0.27841484546661377} 02/25/2022 06:06:33 - INFO - codeparrot_training - Step 22017: {'lr': 0.0003144360375649792, 'samples': 11273216, 'steps': 22017, 'loss/train': 2.2132320404052734} 02/25/2022 06:06:36 - INFO - codeparrot_training - Step 22018: {'lr': 0.0003144202278004199, 'samples': 11273728, 'steps': 22018, 'loss/train': 1.1934279203414917} 02/25/2022 06:06:42 - INFO - codeparrot_training - Step 22019: {'lr': 0.0003144044177599047, 'samples': 11274240, 'steps': 22019, 'loss/train': 2.107335090637207} 02/25/2022 06:06:45 - INFO - codeparrot_training - Step 22020: {'lr': 0.00031438860744350156, 'samples': 11274752, 'steps': 22020, 'loss/train': 1.3147578239440918} 02/25/2022 06:06:51 - INFO - codeparrot_training - Step 22021: {'lr': 0.000314372796851278, 'samples': 11275264, 'steps': 22021, 'loss/train': 2.705869436264038} 02/25/2022 06:06:55 - INFO - codeparrot_training - Step 22022: {'lr': 0.00031435698598330185, 'samples': 11275776, 'steps': 22022, 'loss/train': 2.0425286293029785} 02/25/2022 06:07:00 - INFO - codeparrot_training - Step 22023: {'lr': 0.0003143411748396408, 'samples': 11276288, 'steps': 22023, 'loss/train': 2.396756649017334} 02/25/2022 06:07:04 - INFO - codeparrot_training - Step 22024: {'lr': 0.00031432536342036255, 'samples': 11276800, 'steps': 22024, 'loss/train': 1.321191430091858} 02/25/2022 06:07:09 - INFO - codeparrot_training - Step 22025: {'lr': 0.00031430955172553497, 'samples': 11277312, 'steps': 22025, 'loss/train': 2.703444480895996} 02/25/2022 06:07:13 - INFO - codeparrot_training - Step 22026: {'lr': 0.00031429373975522555, 'samples': 11277824, 'steps': 22026, 'loss/train': 2.1531193256378174} 02/25/2022 06:07:18 - INFO - codeparrot_training - Step 22027: {'lr': 0.00031427792750950227, 'samples': 11278336, 'steps': 22027, 'loss/train': 1.317335605621338} 02/25/2022 06:07:22 - INFO - codeparrot_training - Step 22028: {'lr': 0.0003142621149884327, 'samples': 11278848, 'steps': 22028, 'loss/train': 0.8760442137718201} 02/25/2022 06:07:27 - INFO - codeparrot_training - Step 22029: {'lr': 0.00031424630219208474, 'samples': 11279360, 'steps': 22029, 'loss/train': 1.9641828536987305} 02/25/2022 06:07:31 - INFO - codeparrot_training - Step 22030: {'lr': 0.00031423048912052585, 'samples': 11279872, 'steps': 22030, 'loss/train': 2.4220521450042725} 02/25/2022 06:07:37 - INFO - codeparrot_training - Step 22031: {'lr': 0.0003142146757738241, 'samples': 11280384, 'steps': 22031, 'loss/train': 1.4144482612609863} 02/25/2022 06:07:40 - INFO - codeparrot_training - Step 22032: {'lr': 0.000314198862152047, 'samples': 11280896, 'steps': 22032, 'loss/train': 1.148971676826477} 02/25/2022 06:07:46 - INFO - codeparrot_training - Step 22033: {'lr': 0.00031418304825526236, 'samples': 11281408, 'steps': 22033, 'loss/train': 2.127129077911377} 02/25/2022 06:07:50 - INFO - codeparrot_training - Step 22034: {'lr': 0.0003141672340835379, 'samples': 11281920, 'steps': 22034, 'loss/train': 1.512710452079773} 02/25/2022 06:07:53 - INFO - codeparrot_training - Step 22035: {'lr': 0.0003141514196369414, 'samples': 11282432, 'steps': 22035, 'loss/train': 2.4291930198669434} 02/25/2022 06:07:59 - INFO - codeparrot_training - Step 22036: {'lr': 0.0003141356049155406, 'samples': 11282944, 'steps': 22036, 'loss/train': 1.2321540117263794} 02/25/2022 06:08:02 - INFO - codeparrot_training - Step 22037: {'lr': 0.00031411978991940324, 'samples': 11283456, 'steps': 22037, 'loss/train': 1.7294766902923584} 02/25/2022 06:08:08 - INFO - codeparrot_training - Step 22038: {'lr': 0.000314103974648597, 'samples': 11283968, 'steps': 22038, 'loss/train': 1.537401556968689} 02/25/2022 06:08:12 - INFO - codeparrot_training - Step 22039: {'lr': 0.0003140881591031898, 'samples': 11284480, 'steps': 22039, 'loss/train': 1.3926559686660767} 02/25/2022 06:08:18 - INFO - codeparrot_training - Step 22040: {'lr': 0.0003140723432832492, 'samples': 11284992, 'steps': 22040, 'loss/train': 1.8262780904769897} 02/25/2022 06:08:21 - INFO - codeparrot_training - Step 22041: {'lr': 0.00031405652718884304, 'samples': 11285504, 'steps': 22041, 'loss/train': 3.38923716545105} 02/25/2022 06:08:24 - INFO - codeparrot_training - Step 22042: {'lr': 0.00031404071082003903, 'samples': 11286016, 'steps': 22042, 'loss/train': 1.4119133949279785} 02/25/2022 06:08:30 - INFO - codeparrot_training - Step 22043: {'lr': 0.000314024894176905, 'samples': 11286528, 'steps': 22043, 'loss/train': 2.4981746673583984} 02/25/2022 06:08:34 - INFO - codeparrot_training - Step 22044: {'lr': 0.00031400907725950865, 'samples': 11287040, 'steps': 22044, 'loss/train': 1.7532103061676025} 02/25/2022 06:08:40 - INFO - codeparrot_training - Step 22045: {'lr': 0.00031399326006791765, 'samples': 11287552, 'steps': 22045, 'loss/train': 2.3768534660339355} 02/25/2022 06:08:43 - INFO - codeparrot_training - Step 22046: {'lr': 0.00031397744260219996, 'samples': 11288064, 'steps': 22046, 'loss/train': 2.332976818084717} 02/25/2022 06:08:49 - INFO - codeparrot_training - Step 22047: {'lr': 0.00031396162486242317, 'samples': 11288576, 'steps': 22047, 'loss/train': 2.721773862838745} 02/25/2022 06:08:52 - INFO - codeparrot_training - Step 22048: {'lr': 0.0003139458068486551, 'samples': 11289088, 'steps': 22048, 'loss/train': 2.5646514892578125} 02/25/2022 06:08:58 - INFO - codeparrot_training - Step 22049: {'lr': 0.0003139299885609635, 'samples': 11289600, 'steps': 22049, 'loss/train': 2.6611459255218506} 02/25/2022 06:09:02 - INFO - codeparrot_training - Step 22050: {'lr': 0.00031391416999941606, 'samples': 11290112, 'steps': 22050, 'loss/train': 1.656847357749939} 02/25/2022 06:09:07 - INFO - codeparrot_training - Step 22051: {'lr': 0.00031389835116408073, 'samples': 11290624, 'steps': 22051, 'loss/train': 0.536063015460968} 02/25/2022 06:09:11 - INFO - codeparrot_training - Step 22052: {'lr': 0.000313882532055025, 'samples': 11291136, 'steps': 22052, 'loss/train': 2.067643404006958} 02/25/2022 06:09:16 - INFO - codeparrot_training - Step 22053: {'lr': 0.000313866712672317, 'samples': 11291648, 'steps': 22053, 'loss/train': 0.26975131034851074} 02/25/2022 06:09:20 - INFO - codeparrot_training - Step 22054: {'lr': 0.0003138508930160241, 'samples': 11292160, 'steps': 22054, 'loss/train': 2.1808390617370605} 02/25/2022 06:09:25 - INFO - codeparrot_training - Step 22055: {'lr': 0.0003138350730862144, 'samples': 11292672, 'steps': 22055, 'loss/train': 1.4270166158676147} 02/25/2022 06:09:29 - INFO - codeparrot_training - Step 22056: {'lr': 0.00031381925288295536, 'samples': 11293184, 'steps': 22056, 'loss/train': 2.1215500831604004} 02/25/2022 06:09:34 - INFO - codeparrot_training - Step 22057: {'lr': 0.000313803432406315, 'samples': 11293696, 'steps': 22057, 'loss/train': 1.5915385484695435} 02/25/2022 06:09:37 - INFO - codeparrot_training - Step 22058: {'lr': 0.000313787611656361, 'samples': 11294208, 'steps': 22058, 'loss/train': 2.786083698272705} 02/25/2022 06:09:43 - INFO - codeparrot_training - Step 22059: {'lr': 0.00031377179063316106, 'samples': 11294720, 'steps': 22059, 'loss/train': 1.9742531776428223} 02/25/2022 06:09:46 - INFO - codeparrot_training - Step 22060: {'lr': 0.0003137559693367831, 'samples': 11295232, 'steps': 22060, 'loss/train': 1.7338865995407104} 02/25/2022 06:09:53 - INFO - codeparrot_training - Step 22061: {'lr': 0.0003137401477672947, 'samples': 11295744, 'steps': 22061, 'loss/train': 2.4261152744293213} 02/25/2022 06:09:56 - INFO - codeparrot_training - Step 22062: {'lr': 0.00031372432592476383, 'samples': 11296256, 'steps': 22062, 'loss/train': 1.083591103553772} 02/25/2022 06:10:02 - INFO - codeparrot_training - Step 22063: {'lr': 0.0003137085038092582, 'samples': 11296768, 'steps': 22063, 'loss/train': 2.613905668258667} 02/25/2022 06:10:07 - INFO - codeparrot_training - Step 22064: {'lr': 0.00031369268142084555, 'samples': 11297280, 'steps': 22064, 'loss/train': 1.549139380455017} 02/25/2022 06:10:11 - INFO - codeparrot_training - Step 22065: {'lr': 0.00031367685875959375, 'samples': 11297792, 'steps': 22065, 'loss/train': 2.2094218730926514} 02/25/2022 06:10:14 - INFO - codeparrot_training - Step 22066: {'lr': 0.0003136610358255704, 'samples': 11298304, 'steps': 22066, 'loss/train': 2.8319671154022217} 02/25/2022 06:10:20 - INFO - codeparrot_training - Step 22067: {'lr': 0.00031364521261884336, 'samples': 11298816, 'steps': 22067, 'loss/train': 1.9258968830108643} 02/25/2022 06:10:23 - INFO - codeparrot_training - Step 22068: {'lr': 0.00031362938913948046, 'samples': 11299328, 'steps': 22068, 'loss/train': 1.7411415576934814} 02/25/2022 06:10:30 - INFO - codeparrot_training - Step 22069: {'lr': 0.0003136135653875495, 'samples': 11299840, 'steps': 22069, 'loss/train': 1.7595075368881226} 02/25/2022 06:10:33 - INFO - codeparrot_training - Step 22070: {'lr': 0.00031359774136311823, 'samples': 11300352, 'steps': 22070, 'loss/train': 1.4619450569152832} 02/25/2022 06:10:39 - INFO - codeparrot_training - Step 22071: {'lr': 0.0003135819170662545, 'samples': 11300864, 'steps': 22071, 'loss/train': 1.2482984066009521} 02/25/2022 06:10:42 - INFO - codeparrot_training - Step 22072: {'lr': 0.00031356609249702587, 'samples': 11301376, 'steps': 22072, 'loss/train': 1.7502570152282715} 02/25/2022 06:10:48 - INFO - codeparrot_training - Step 22073: {'lr': 0.0003135502676555004, 'samples': 11301888, 'steps': 22073, 'loss/train': 1.2703635692596436} 02/25/2022 06:10:51 - INFO - codeparrot_training - Step 22074: {'lr': 0.0003135344425417457, 'samples': 11302400, 'steps': 22074, 'loss/train': 1.7558841705322266} 02/25/2022 06:10:57 - INFO - codeparrot_training - Step 22075: {'lr': 0.0003135186171558297, 'samples': 11302912, 'steps': 22075, 'loss/train': 1.6188462972640991} 02/25/2022 06:11:00 - INFO - codeparrot_training - Step 22076: {'lr': 0.00031350279149782004, 'samples': 11303424, 'steps': 22076, 'loss/train': 1.442162275314331} 02/25/2022 06:11:06 - INFO - codeparrot_training - Step 22077: {'lr': 0.0003134869655677846, 'samples': 11303936, 'steps': 22077, 'loss/train': 2.6089706420898438} 02/25/2022 06:11:09 - INFO - codeparrot_training - Step 22078: {'lr': 0.00031347113936579116, 'samples': 11304448, 'steps': 22078, 'loss/train': 2.9049012660980225} 02/25/2022 06:11:16 - INFO - codeparrot_training - Step 22079: {'lr': 0.00031345531289190756, 'samples': 11304960, 'steps': 22079, 'loss/train': 0.896712601184845} 02/25/2022 06:11:19 - INFO - codeparrot_training - Step 22080: {'lr': 0.0003134394861462014, 'samples': 11305472, 'steps': 22080, 'loss/train': 0.7773104906082153} 02/25/2022 06:11:25 - INFO - codeparrot_training - Step 22081: {'lr': 0.0003134236591287407, 'samples': 11305984, 'steps': 22081, 'loss/train': 1.9527983665466309} 02/25/2022 06:11:28 - INFO - codeparrot_training - Step 22082: {'lr': 0.0003134078318395933, 'samples': 11306496, 'steps': 22082, 'loss/train': 2.2992241382598877} 02/25/2022 06:11:34 - INFO - codeparrot_training - Step 22083: {'lr': 0.00031339200427882676, 'samples': 11307008, 'steps': 22083, 'loss/train': 8.847299575805664} 02/25/2022 06:11:37 - INFO - codeparrot_training - Step 22084: {'lr': 0.00031337617644650907, 'samples': 11307520, 'steps': 22084, 'loss/train': 2.2997782230377197} 02/25/2022 06:11:43 - INFO - codeparrot_training - Step 22085: {'lr': 0.00031336034834270786, 'samples': 11308032, 'steps': 22085, 'loss/train': 1.500760793685913} 02/25/2022 06:11:47 - INFO - codeparrot_training - Step 22086: {'lr': 0.00031334451996749117, 'samples': 11308544, 'steps': 22086, 'loss/train': 2.2298805713653564} 02/25/2022 06:11:52 - INFO - codeparrot_training - Step 22087: {'lr': 0.00031332869132092654, 'samples': 11309056, 'steps': 22087, 'loss/train': 2.9004907608032227} 02/25/2022 06:11:56 - INFO - codeparrot_training - Step 22088: {'lr': 0.00031331286240308205, 'samples': 11309568, 'steps': 22088, 'loss/train': 2.2678866386413574} 02/25/2022 06:12:01 - INFO - codeparrot_training - Step 22089: {'lr': 0.00031329703321402526, 'samples': 11310080, 'steps': 22089, 'loss/train': 2.3201212882995605} 02/25/2022 06:12:05 - INFO - codeparrot_training - Step 22090: {'lr': 0.00031328120375382414, 'samples': 11310592, 'steps': 22090, 'loss/train': 2.5450994968414307} 02/25/2022 06:12:11 - INFO - codeparrot_training - Step 22091: {'lr': 0.0003132653740225464, 'samples': 11311104, 'steps': 22091, 'loss/train': 2.115913152694702} 02/25/2022 06:12:15 - INFO - codeparrot_training - Step 22092: {'lr': 0.0003132495440202599, 'samples': 11311616, 'steps': 22092, 'loss/train': 1.537142276763916} 02/25/2022 06:12:20 - INFO - codeparrot_training - Step 22093: {'lr': 0.0003132337137470324, 'samples': 11312128, 'steps': 22093, 'loss/train': 2.2702367305755615} 02/25/2022 06:12:24 - INFO - codeparrot_training - Step 22094: {'lr': 0.00031321788320293176, 'samples': 11312640, 'steps': 22094, 'loss/train': 2.2105822563171387} 02/25/2022 06:12:29 - INFO - codeparrot_training - Step 22095: {'lr': 0.00031320205238802583, 'samples': 11313152, 'steps': 22095, 'loss/train': 0.5547376275062561} 02/25/2022 06:12:33 - INFO - codeparrot_training - Step 22096: {'lr': 0.0003131862213023823, 'samples': 11313664, 'steps': 22096, 'loss/train': 3.415804862976074} 02/25/2022 06:12:38 - INFO - codeparrot_training - Step 22097: {'lr': 0.0003131703899460692, 'samples': 11314176, 'steps': 22097, 'loss/train': 2.715240001678467} 02/25/2022 06:12:42 - INFO - codeparrot_training - Step 22098: {'lr': 0.0003131545583191541, 'samples': 11314688, 'steps': 22098, 'loss/train': 0.7137669324874878} 02/25/2022 06:12:47 - INFO - codeparrot_training - Step 22099: {'lr': 0.00031313872642170493, 'samples': 11315200, 'steps': 22099, 'loss/train': 1.9726505279541016} 02/25/2022 06:12:51 - INFO - codeparrot_training - Step 22100: {'lr': 0.0003131228942537895, 'samples': 11315712, 'steps': 22100, 'loss/train': 2.0028703212738037} 02/25/2022 06:12:56 - INFO - codeparrot_training - Step 22101: {'lr': 0.00031310706181547567, 'samples': 11316224, 'steps': 22101, 'loss/train': 0.7319560647010803} 02/25/2022 06:13:00 - INFO - codeparrot_training - Step 22102: {'lr': 0.0003130912291068312, 'samples': 11316736, 'steps': 22102, 'loss/train': 1.1150904893875122} 02/25/2022 06:13:05 - INFO - codeparrot_training - Step 22103: {'lr': 0.000313075396127924, 'samples': 11317248, 'steps': 22103, 'loss/train': 1.8547435998916626} 02/25/2022 06:13:09 - INFO - codeparrot_training - Step 22104: {'lr': 0.0003130595628788217, 'samples': 11317760, 'steps': 22104, 'loss/train': 2.63070011138916} 02/25/2022 06:13:14 - INFO - codeparrot_training - Step 22105: {'lr': 0.00031304372935959235, 'samples': 11318272, 'steps': 22105, 'loss/train': 2.5418074131011963} 02/25/2022 06:13:18 - INFO - codeparrot_training - Step 22106: {'lr': 0.00031302789557030364, 'samples': 11318784, 'steps': 22106, 'loss/train': 1.7129013538360596} 02/25/2022 06:13:24 - INFO - codeparrot_training - Step 22107: {'lr': 0.00031301206151102353, 'samples': 11319296, 'steps': 22107, 'loss/train': 1.9842053651809692} 02/25/2022 06:13:27 - INFO - codeparrot_training - Step 22108: {'lr': 0.00031299622718181964, 'samples': 11319808, 'steps': 22108, 'loss/train': 1.3814620971679688} 02/25/2022 06:13:33 - INFO - codeparrot_training - Step 22109: {'lr': 0.00031298039258276, 'samples': 11320320, 'steps': 22109, 'loss/train': 1.87350332736969} 02/25/2022 06:13:36 - INFO - codeparrot_training - Step 22110: {'lr': 0.0003129645577139123, 'samples': 11320832, 'steps': 22110, 'loss/train': 0.9727237224578857} 02/25/2022 06:13:42 - INFO - codeparrot_training - Step 22111: {'lr': 0.00031294872257534446, 'samples': 11321344, 'steps': 22111, 'loss/train': 1.3958938121795654} 02/25/2022 06:13:48 - INFO - codeparrot_training - Step 22112: {'lr': 0.0003129328871671243, 'samples': 11321856, 'steps': 22112, 'loss/train': 1.9277312755584717} 02/25/2022 06:13:51 - INFO - codeparrot_training - Step 22113: {'lr': 0.0003129170514893196, 'samples': 11322368, 'steps': 22113, 'loss/train': 2.7265465259552} 02/25/2022 06:13:55 - INFO - codeparrot_training - Step 22114: {'lr': 0.0003129012155419983, 'samples': 11322880, 'steps': 22114, 'loss/train': 1.2648028135299683} 02/25/2022 06:14:02 - INFO - codeparrot_training - Step 22115: {'lr': 0.00031288537932522807, 'samples': 11323392, 'steps': 22115, 'loss/train': 1.6724538803100586} 02/25/2022 06:14:05 - INFO - codeparrot_training - Step 22116: {'lr': 0.0003128695428390769, 'samples': 11323904, 'steps': 22116, 'loss/train': 1.838234543800354} 02/25/2022 06:14:11 - INFO - codeparrot_training - Step 22117: {'lr': 0.0003128537060836125, 'samples': 11324416, 'steps': 22117, 'loss/train': 1.3269188404083252} 02/25/2022 06:14:14 - INFO - codeparrot_training - Step 22118: {'lr': 0.0003128378690589028, 'samples': 11324928, 'steps': 22118, 'loss/train': 0.9925171732902527} 02/25/2022 06:14:20 - INFO - codeparrot_training - Step 22119: {'lr': 0.00031282203176501573, 'samples': 11325440, 'steps': 22119, 'loss/train': 0.9470146298408508} 02/25/2022 06:14:23 - INFO - codeparrot_training - Step 22120: {'lr': 0.0003128061942020189, 'samples': 11325952, 'steps': 22120, 'loss/train': 2.619854211807251} 02/25/2022 06:14:29 - INFO - codeparrot_training - Step 22121: {'lr': 0.00031279035636998037, 'samples': 11326464, 'steps': 22121, 'loss/train': 1.4416135549545288} 02/25/2022 06:14:32 - INFO - codeparrot_training - Step 22122: {'lr': 0.0003127745182689678, 'samples': 11326976, 'steps': 22122, 'loss/train': 1.8916935920715332} 02/25/2022 06:14:38 - INFO - codeparrot_training - Step 22123: {'lr': 0.00031275867989904923, 'samples': 11327488, 'steps': 22123, 'loss/train': 2.0850741863250732} 02/25/2022 06:14:41 - INFO - codeparrot_training - Step 22124: {'lr': 0.0003127428412602923, 'samples': 11328000, 'steps': 22124, 'loss/train': 1.721575140953064} 02/25/2022 06:14:47 - INFO - codeparrot_training - Step 22125: {'lr': 0.00031272700235276507, 'samples': 11328512, 'steps': 22125, 'loss/train': 2.039088726043701} 02/25/2022 06:14:50 - INFO - codeparrot_training - Step 22126: {'lr': 0.00031271116317653513, 'samples': 11329024, 'steps': 22126, 'loss/train': 2.3093457221984863} 02/25/2022 06:14:57 - INFO - codeparrot_training - Step 22127: {'lr': 0.00031269532373167063, 'samples': 11329536, 'steps': 22127, 'loss/train': 1.8865641355514526} 02/25/2022 06:15:01 - INFO - codeparrot_training - Step 22128: {'lr': 0.0003126794840182392, 'samples': 11330048, 'steps': 22128, 'loss/train': 1.6816167831420898} 02/25/2022 06:15:06 - INFO - codeparrot_training - Step 22129: {'lr': 0.00031266364403630874, 'samples': 11330560, 'steps': 22129, 'loss/train': 1.5853948593139648} 02/25/2022 06:15:10 - INFO - codeparrot_training - Step 22130: {'lr': 0.0003126478037859472, 'samples': 11331072, 'steps': 22130, 'loss/train': 1.2221128940582275} 02/25/2022 06:15:15 - INFO - codeparrot_training - Step 22131: {'lr': 0.0003126319632672223, 'samples': 11331584, 'steps': 22131, 'loss/train': 1.6561076641082764} 02/25/2022 06:15:19 - INFO - codeparrot_training - Step 22132: {'lr': 0.00031261612248020197, 'samples': 11332096, 'steps': 22132, 'loss/train': 2.2861557006835938} 02/25/2022 06:15:24 - INFO - codeparrot_training - Step 22133: {'lr': 0.00031260028142495404, 'samples': 11332608, 'steps': 22133, 'loss/train': 2.2173848152160645} 02/25/2022 06:15:28 - INFO - codeparrot_training - Step 22134: {'lr': 0.0003125844401015465, 'samples': 11333120, 'steps': 22134, 'loss/train': 0.941705584526062} 02/25/2022 06:15:33 - INFO - codeparrot_training - Step 22135: {'lr': 0.0003125685985100469, 'samples': 11333632, 'steps': 22135, 'loss/train': 1.8156362771987915} 02/25/2022 06:15:37 - INFO - codeparrot_training - Step 22136: {'lr': 0.0003125527566505234, 'samples': 11334144, 'steps': 22136, 'loss/train': 1.6366169452667236} 02/25/2022 06:15:43 - INFO - codeparrot_training - Step 22137: {'lr': 0.0003125369145230438, 'samples': 11334656, 'steps': 22137, 'loss/train': 2.258678674697876} 02/25/2022 06:15:46 - INFO - codeparrot_training - Step 22138: {'lr': 0.0003125210721276758, 'samples': 11335168, 'steps': 22138, 'loss/train': 1.7923755645751953} 02/25/2022 06:15:52 - INFO - codeparrot_training - Step 22139: {'lr': 0.0003125052294644874, 'samples': 11335680, 'steps': 22139, 'loss/train': 0.8070864677429199} 02/25/2022 06:15:55 - INFO - codeparrot_training - Step 22140: {'lr': 0.0003124893865335466, 'samples': 11336192, 'steps': 22140, 'loss/train': 2.3999979496002197} 02/25/2022 06:16:01 - INFO - codeparrot_training - Step 22141: {'lr': 0.00031247354333492096, 'samples': 11336704, 'steps': 22141, 'loss/train': 0.48616787791252136} 02/25/2022 06:16:04 - INFO - codeparrot_training - Step 22142: {'lr': 0.00031245769986867845, 'samples': 11337216, 'steps': 22142, 'loss/train': 0.303365021944046} 02/25/2022 06:16:10 - INFO - codeparrot_training - Step 22143: {'lr': 0.0003124418561348871, 'samples': 11337728, 'steps': 22143, 'loss/train': 2.018758535385132} 02/25/2022 06:16:13 - INFO - codeparrot_training - Step 22144: {'lr': 0.0003124260121336146, 'samples': 11338240, 'steps': 22144, 'loss/train': 2.5217652320861816} 02/25/2022 06:16:19 - INFO - codeparrot_training - Step 22145: {'lr': 0.000312410167864929, 'samples': 11338752, 'steps': 22145, 'loss/train': 0.6321930885314941} 02/25/2022 06:16:22 - INFO - codeparrot_training - Step 22146: {'lr': 0.00031239432332889796, 'samples': 11339264, 'steps': 22146, 'loss/train': 1.5276532173156738} 02/25/2022 06:16:28 - INFO - codeparrot_training - Step 22147: {'lr': 0.00031237847852558947, 'samples': 11339776, 'steps': 22147, 'loss/train': 3.1110036373138428} 02/25/2022 06:16:31 - INFO - codeparrot_training - Step 22148: {'lr': 0.00031236263345507133, 'samples': 11340288, 'steps': 22148, 'loss/train': 1.7653257846832275} 02/25/2022 06:16:36 - INFO - codeparrot_training - Step 22149: {'lr': 0.0003123467881174116, 'samples': 11340800, 'steps': 22149, 'loss/train': 1.5785020589828491} 02/25/2022 06:16:42 - INFO - codeparrot_training - Step 22150: {'lr': 0.0003123309425126779, 'samples': 11341312, 'steps': 22150, 'loss/train': 1.8085739612579346} 02/25/2022 06:16:45 - INFO - codeparrot_training - Step 22151: {'lr': 0.00031231509664093833, 'samples': 11341824, 'steps': 22151, 'loss/train': 2.071397066116333} 02/25/2022 06:16:52 - INFO - codeparrot_training - Step 22152: {'lr': 0.00031229925050226063, 'samples': 11342336, 'steps': 22152, 'loss/train': 2.453071117401123} 02/25/2022 06:16:55 - INFO - codeparrot_training - Step 22153: {'lr': 0.0003122834040967127, 'samples': 11342848, 'steps': 22153, 'loss/train': 1.8134874105453491} 02/25/2022 06:17:01 - INFO - codeparrot_training - Step 22154: {'lr': 0.00031226755742436255, 'samples': 11343360, 'steps': 22154, 'loss/train': 2.8095500469207764} 02/25/2022 06:17:04 - INFO - codeparrot_training - Step 22155: {'lr': 0.0003122517104852778, 'samples': 11343872, 'steps': 22155, 'loss/train': 1.751172423362732} 02/25/2022 06:17:10 - INFO - codeparrot_training - Step 22156: {'lr': 0.0003122358632795266, 'samples': 11344384, 'steps': 22156, 'loss/train': 1.9667421579360962} 02/25/2022 06:17:13 - INFO - codeparrot_training - Step 22157: {'lr': 0.00031222001580717663, 'samples': 11344896, 'steps': 22157, 'loss/train': 0.8395716547966003} 02/25/2022 06:17:19 - INFO - codeparrot_training - Step 22158: {'lr': 0.000312204168068296, 'samples': 11345408, 'steps': 22158, 'loss/train': 2.086977005004883} 02/25/2022 06:17:22 - INFO - codeparrot_training - Step 22159: {'lr': 0.00031218832006295235, 'samples': 11345920, 'steps': 22159, 'loss/train': 1.1217615604400635} 02/25/2022 06:17:27 - INFO - codeparrot_training - Step 22160: {'lr': 0.0003121724717912138, 'samples': 11346432, 'steps': 22160, 'loss/train': 1.074906349182129} 02/25/2022 06:17:31 - INFO - codeparrot_training - Step 22161: {'lr': 0.000312156623253148, 'samples': 11346944, 'steps': 22161, 'loss/train': 1.9170221090316772} 02/25/2022 06:17:37 - INFO - codeparrot_training - Step 22162: {'lr': 0.00031214077444882297, 'samples': 11347456, 'steps': 22162, 'loss/train': 1.7256731986999512} 02/25/2022 06:17:40 - INFO - codeparrot_training - Step 22163: {'lr': 0.0003121249253783067, 'samples': 11347968, 'steps': 22163, 'loss/train': 1.3001811504364014} 02/25/2022 06:17:46 - INFO - codeparrot_training - Step 22164: {'lr': 0.00031210907604166686, 'samples': 11348480, 'steps': 22164, 'loss/train': 1.930328607559204} 02/25/2022 06:17:49 - INFO - codeparrot_training - Step 22165: {'lr': 0.0003120932264389715, 'samples': 11348992, 'steps': 22165, 'loss/train': 1.9401648044586182} 02/25/2022 06:17:55 - INFO - codeparrot_training - Step 22166: {'lr': 0.0003120773765702885, 'samples': 11349504, 'steps': 22166, 'loss/train': 1.5868417024612427} 02/25/2022 06:17:59 - INFO - codeparrot_training - Step 22167: {'lr': 0.00031206152643568577, 'samples': 11350016, 'steps': 22167, 'loss/train': 2.1357803344726562} 02/25/2022 06:18:04 - INFO - codeparrot_training - Step 22168: {'lr': 0.00031204567603523105, 'samples': 11350528, 'steps': 22168, 'loss/train': 2.0386104583740234} 02/25/2022 06:18:08 - INFO - codeparrot_training - Step 22169: {'lr': 0.00031202982536899246, 'samples': 11351040, 'steps': 22169, 'loss/train': 1.4887747764587402} 02/25/2022 06:18:11 - INFO - codeparrot_training - Step 22170: {'lr': 0.0003120139744370377, 'samples': 11351552, 'steps': 22170, 'loss/train': 2.29206919670105} 02/25/2022 06:18:17 - INFO - codeparrot_training - Step 22171: {'lr': 0.0003119981232394349, 'samples': 11352064, 'steps': 22171, 'loss/train': 1.4098385572433472} 02/25/2022 06:18:20 - INFO - codeparrot_training - Step 22172: {'lr': 0.0003119822717762517, 'samples': 11352576, 'steps': 22172, 'loss/train': 1.6300408840179443} 02/25/2022 06:18:26 - INFO - codeparrot_training - Step 22173: {'lr': 0.0003119664200475562, 'samples': 11353088, 'steps': 22173, 'loss/train': 1.2397927045822144} 02/25/2022 06:18:30 - INFO - codeparrot_training - Step 22174: {'lr': 0.0003119505680534162, 'samples': 11353600, 'steps': 22174, 'loss/train': 1.7889060974121094} 02/25/2022 06:18:35 - INFO - codeparrot_training - Step 22175: {'lr': 0.00031193471579389967, 'samples': 11354112, 'steps': 22175, 'loss/train': 1.9722788333892822} 02/25/2022 06:18:39 - INFO - codeparrot_training - Step 22176: {'lr': 0.00031191886326907445, 'samples': 11354624, 'steps': 22176, 'loss/train': 2.2971391677856445} 02/25/2022 06:18:44 - INFO - codeparrot_training - Step 22177: {'lr': 0.0003119030104790085, 'samples': 11355136, 'steps': 22177, 'loss/train': 1.3095362186431885} 02/25/2022 06:18:48 - INFO - codeparrot_training - Step 22178: {'lr': 0.00031188715742376966, 'samples': 11355648, 'steps': 22178, 'loss/train': 1.6352877616882324} 02/25/2022 06:18:53 - INFO - codeparrot_training - Step 22179: {'lr': 0.0003118713041034259, 'samples': 11356160, 'steps': 22179, 'loss/train': 2.841221570968628} 02/25/2022 06:18:57 - INFO - codeparrot_training - Step 22180: {'lr': 0.0003118554505180452, 'samples': 11356672, 'steps': 22180, 'loss/train': 3.101526975631714} 02/25/2022 06:19:02 - INFO - codeparrot_training - Step 22181: {'lr': 0.0003118395966676953, 'samples': 11357184, 'steps': 22181, 'loss/train': 2.118783950805664} 02/25/2022 06:19:06 - INFO - codeparrot_training - Step 22182: {'lr': 0.00031182374255244426, 'samples': 11357696, 'steps': 22182, 'loss/train': 2.5226573944091797} 02/25/2022 06:19:12 - INFO - codeparrot_training - Step 22183: {'lr': 0.0003118078881723599, 'samples': 11358208, 'steps': 22183, 'loss/train': 1.4040274620056152} 02/25/2022 06:19:16 - INFO - codeparrot_training - Step 22184: {'lr': 0.0003117920335275102, 'samples': 11358720, 'steps': 22184, 'loss/train': 2.7390952110290527} 02/25/2022 06:19:21 - INFO - codeparrot_training - Step 22185: {'lr': 0.000311776178617963, 'samples': 11359232, 'steps': 22185, 'loss/train': 1.879380226135254} 02/25/2022 06:19:25 - INFO - codeparrot_training - Step 22186: {'lr': 0.0003117603234437864, 'samples': 11359744, 'steps': 22186, 'loss/train': 1.598555564880371} 02/25/2022 06:19:30 - INFO - codeparrot_training - Step 22187: {'lr': 0.0003117444680050481, 'samples': 11360256, 'steps': 22187, 'loss/train': 1.8586094379425049} 02/25/2022 06:19:34 - INFO - codeparrot_training - Step 22188: {'lr': 0.00031172861230181607, 'samples': 11360768, 'steps': 22188, 'loss/train': 1.9133294820785522} 02/25/2022 06:19:39 - INFO - codeparrot_training - Step 22189: {'lr': 0.0003117127563341583, 'samples': 11361280, 'steps': 22189, 'loss/train': 3.0111265182495117} 02/25/2022 06:19:43 - INFO - codeparrot_training - Step 22190: {'lr': 0.0003116969001021427, 'samples': 11361792, 'steps': 22190, 'loss/train': 2.1985645294189453} 02/25/2022 06:19:48 - INFO - codeparrot_training - Step 22191: {'lr': 0.00031168104360583716, 'samples': 11362304, 'steps': 22191, 'loss/train': 2.684141159057617} 02/25/2022 06:19:52 - INFO - codeparrot_training - Step 22192: {'lr': 0.00031166518684530966, 'samples': 11362816, 'steps': 22192, 'loss/train': 1.6079938411712646} 02/25/2022 06:19:57 - INFO - codeparrot_training - Step 22193: {'lr': 0.000311649329820628, 'samples': 11363328, 'steps': 22193, 'loss/train': 2.01727294921875} 02/25/2022 06:20:01 - INFO - codeparrot_training - Step 22194: {'lr': 0.0003116334725318602, 'samples': 11363840, 'steps': 22194, 'loss/train': 2.1195273399353027} 02/25/2022 06:20:07 - INFO - codeparrot_training - Step 22195: {'lr': 0.00031161761497907416, 'samples': 11364352, 'steps': 22195, 'loss/train': 1.7520968914031982} 02/25/2022 06:20:10 - INFO - codeparrot_training - Step 22196: {'lr': 0.00031160175716233793, 'samples': 11364864, 'steps': 22196, 'loss/train': 1.125702977180481} 02/25/2022 06:20:16 - INFO - codeparrot_training - Step 22197: {'lr': 0.0003115858990817192, 'samples': 11365376, 'steps': 22197, 'loss/train': 1.9171411991119385} 02/25/2022 06:20:21 - INFO - codeparrot_training - Step 22198: {'lr': 0.0003115700407372861, 'samples': 11365888, 'steps': 22198, 'loss/train': 2.5504097938537598} 02/25/2022 06:20:25 - INFO - codeparrot_training - Step 22199: {'lr': 0.00031155418212910647, 'samples': 11366400, 'steps': 22199, 'loss/train': 1.4988442659378052} 02/25/2022 06:20:31 - INFO - codeparrot_training - Step 22200: {'lr': 0.00031153832325724826, 'samples': 11366912, 'steps': 22200, 'loss/train': 2.427340507507324} 02/25/2022 06:20:34 - INFO - codeparrot_training - Step 22201: {'lr': 0.0003115224641217795, 'samples': 11367424, 'steps': 22201, 'loss/train': 1.7603408098220825} 02/25/2022 06:20:37 - INFO - codeparrot_training - Step 22202: {'lr': 0.0003115066047227679, 'samples': 11367936, 'steps': 22202, 'loss/train': 1.614784836769104} 02/25/2022 06:20:43 - INFO - codeparrot_training - Step 22203: {'lr': 0.00031149074506028163, 'samples': 11368448, 'steps': 22203, 'loss/train': 2.251432180404663} 02/25/2022 06:20:47 - INFO - codeparrot_training - Step 22204: {'lr': 0.00031147488513438853, 'samples': 11368960, 'steps': 22204, 'loss/train': 1.5265086889266968} 02/25/2022 06:20:52 - INFO - codeparrot_training - Step 22205: {'lr': 0.00031145902494515655, 'samples': 11369472, 'steps': 22205, 'loss/train': 1.5114670991897583} 02/25/2022 06:20:56 - INFO - codeparrot_training - Step 22206: {'lr': 0.0003114431644926536, 'samples': 11369984, 'steps': 22206, 'loss/train': 2.287856340408325} 02/25/2022 06:21:01 - INFO - codeparrot_training - Step 22207: {'lr': 0.00031142730377694763, 'samples': 11370496, 'steps': 22207, 'loss/train': 1.7188853025436401} 02/25/2022 06:21:04 - INFO - codeparrot_training - Step 22208: {'lr': 0.0003114114427981066, 'samples': 11371008, 'steps': 22208, 'loss/train': 2.1920838356018066} 02/25/2022 06:21:12 - INFO - codeparrot_training - Step 22209: {'lr': 0.00031139558155619844, 'samples': 11371520, 'steps': 22209, 'loss/train': 2.505232334136963} 02/25/2022 06:21:15 - INFO - codeparrot_training - Step 22210: {'lr': 0.0003113797200512912, 'samples': 11372032, 'steps': 22210, 'loss/train': 2.104592800140381} 02/25/2022 06:21:20 - INFO - codeparrot_training - Step 22211: {'lr': 0.0003113638582834526, 'samples': 11372544, 'steps': 22211, 'loss/train': 1.9102720022201538} 02/25/2022 06:21:24 - INFO - codeparrot_training - Step 22212: {'lr': 0.00031134799625275077, 'samples': 11373056, 'steps': 22212, 'loss/train': 2.0225603580474854} 02/25/2022 06:21:29 - INFO - codeparrot_training - Step 22213: {'lr': 0.0003113321339592536, 'samples': 11373568, 'steps': 22213, 'loss/train': 2.8301875591278076} 02/25/2022 06:21:33 - INFO - codeparrot_training - Step 22214: {'lr': 0.000311316271403029, 'samples': 11374080, 'steps': 22214, 'loss/train': 0.8736248016357422} 02/25/2022 06:21:38 - INFO - codeparrot_training - Step 22215: {'lr': 0.00031130040858414506, 'samples': 11374592, 'steps': 22215, 'loss/train': 1.6126081943511963} 02/25/2022 06:21:42 - INFO - codeparrot_training - Step 22216: {'lr': 0.00031128454550266956, 'samples': 11375104, 'steps': 22216, 'loss/train': 1.715529203414917} 02/25/2022 06:21:47 - INFO - codeparrot_training - Step 22217: {'lr': 0.0003112686821586706, 'samples': 11375616, 'steps': 22217, 'loss/train': 1.7984647750854492} 02/25/2022 06:21:51 - INFO - codeparrot_training - Step 22218: {'lr': 0.000311252818552216, 'samples': 11376128, 'steps': 22218, 'loss/train': 2.9339468479156494} 02/25/2022 06:21:57 - INFO - codeparrot_training - Step 22219: {'lr': 0.00031123695468337375, 'samples': 11376640, 'steps': 22219, 'loss/train': 2.3963327407836914} 02/25/2022 06:22:01 - INFO - codeparrot_training - Step 22220: {'lr': 0.00031122109055221187, 'samples': 11377152, 'steps': 22220, 'loss/train': 2.3818411827087402} 02/25/2022 06:22:06 - INFO - codeparrot_training - Step 22221: {'lr': 0.00031120522615879834, 'samples': 11377664, 'steps': 22221, 'loss/train': 8.73620891571045} 02/25/2022 06:22:10 - INFO - codeparrot_training - Step 22222: {'lr': 0.00031118936150320093, 'samples': 11378176, 'steps': 22222, 'loss/train': 1.4981305599212646} 02/25/2022 06:22:16 - INFO - codeparrot_training - Step 22223: {'lr': 0.00031117349658548783, 'samples': 11378688, 'steps': 22223, 'loss/train': 1.4015003442764282} 02/25/2022 06:22:19 - INFO - codeparrot_training - Step 22224: {'lr': 0.00031115763140572686, 'samples': 11379200, 'steps': 22224, 'loss/train': 1.9452404975891113} 02/25/2022 06:22:25 - INFO - codeparrot_training - Step 22225: {'lr': 0.000311141765963986, 'samples': 11379712, 'steps': 22225, 'loss/train': 1.7791557312011719} 02/25/2022 06:22:28 - INFO - codeparrot_training - Step 22226: {'lr': 0.00031112590026033323, 'samples': 11380224, 'steps': 22226, 'loss/train': 2.105590343475342} 02/25/2022 06:22:34 - INFO - codeparrot_training - Step 22227: {'lr': 0.00031111003429483647, 'samples': 11380736, 'steps': 22227, 'loss/train': 2.030669927597046} 02/25/2022 06:22:37 - INFO - codeparrot_training - Step 22228: {'lr': 0.00031109416806756387, 'samples': 11381248, 'steps': 22228, 'loss/train': 1.9703563451766968} 02/25/2022 06:22:43 - INFO - codeparrot_training - Step 22229: {'lr': 0.0003110783015785831, 'samples': 11381760, 'steps': 22229, 'loss/train': 1.6472063064575195} 02/25/2022 06:22:46 - INFO - codeparrot_training - Step 22230: {'lr': 0.00031106243482796234, 'samples': 11382272, 'steps': 22230, 'loss/train': 1.0945477485656738} 02/25/2022 06:22:52 - INFO - codeparrot_training - Step 22231: {'lr': 0.0003110465678157695, 'samples': 11382784, 'steps': 22231, 'loss/train': 2.600090980529785} 02/25/2022 06:22:55 - INFO - codeparrot_training - Step 22232: {'lr': 0.0003110307005420726, 'samples': 11383296, 'steps': 22232, 'loss/train': 1.1109521389007568} 02/25/2022 06:23:03 - INFO - codeparrot_training - Step 22233: {'lr': 0.00031101483300693944, 'samples': 11383808, 'steps': 22233, 'loss/train': 1.9960497617721558} 02/25/2022 06:23:06 - INFO - codeparrot_training - Step 22234: {'lr': 0.00031099896521043826, 'samples': 11384320, 'steps': 22234, 'loss/train': 1.584606409072876} 02/25/2022 06:23:12 - INFO - codeparrot_training - Step 22235: {'lr': 0.0003109830971526367, 'samples': 11384832, 'steps': 22235, 'loss/train': 3.2077078819274902} 02/25/2022 06:23:16 - INFO - codeparrot_training - Step 22236: {'lr': 0.0003109672288336031, 'samples': 11385344, 'steps': 22236, 'loss/train': 0.8509630560874939} 02/25/2022 06:23:21 - INFO - codeparrot_training - Step 22237: {'lr': 0.00031095136025340514, 'samples': 11385856, 'steps': 22237, 'loss/train': 2.147021770477295} 02/25/2022 06:23:25 - INFO - codeparrot_training - Step 22238: {'lr': 0.00031093549141211096, 'samples': 11386368, 'steps': 22238, 'loss/train': 2.7605669498443604} 02/25/2022 06:23:30 - INFO - codeparrot_training - Step 22239: {'lr': 0.00031091962230978844, 'samples': 11386880, 'steps': 22239, 'loss/train': 2.381690263748169} 02/25/2022 06:23:34 - INFO - codeparrot_training - Step 22240: {'lr': 0.0003109037529465056, 'samples': 11387392, 'steps': 22240, 'loss/train': 2.512786626815796} 02/25/2022 06:23:39 - INFO - codeparrot_training - Step 22241: {'lr': 0.0003108878833223305, 'samples': 11387904, 'steps': 22241, 'loss/train': 1.117661714553833} 02/25/2022 06:23:43 - INFO - codeparrot_training - Step 22242: {'lr': 0.00031087201343733096, 'samples': 11388416, 'steps': 22242, 'loss/train': 1.766898274421692} 02/25/2022 06:23:50 - INFO - codeparrot_training - Step 22243: {'lr': 0.00031085614329157515, 'samples': 11388928, 'steps': 22243, 'loss/train': 1.4396743774414062} 02/25/2022 06:23:53 - INFO - codeparrot_training - Step 22244: {'lr': 0.00031084027288513083, 'samples': 11389440, 'steps': 22244, 'loss/train': 1.9462149143218994} 02/25/2022 06:23:59 - INFO - codeparrot_training - Step 22245: {'lr': 0.0003108244022180661, 'samples': 11389952, 'steps': 22245, 'loss/train': 1.7819745540618896} 02/25/2022 06:24:02 - INFO - codeparrot_training - Step 22246: {'lr': 0.000310808531290449, 'samples': 11390464, 'steps': 22246, 'loss/train': 2.3998100757598877} 02/25/2022 06:24:08 - INFO - codeparrot_training - Step 22247: {'lr': 0.00031079266010234746, 'samples': 11390976, 'steps': 22247, 'loss/train': 2.242130994796753} 02/25/2022 06:24:11 - INFO - codeparrot_training - Step 22248: {'lr': 0.00031077678865382944, 'samples': 11391488, 'steps': 22248, 'loss/train': 0.9377748370170593} 02/25/2022 06:24:17 - INFO - codeparrot_training - Step 22249: {'lr': 0.000310760916944963, 'samples': 11392000, 'steps': 22249, 'loss/train': 2.5221896171569824} 02/25/2022 06:24:20 - INFO - codeparrot_training - Step 22250: {'lr': 0.000310745044975816, 'samples': 11392512, 'steps': 22250, 'loss/train': 1.9791074991226196} 02/25/2022 06:24:26 - INFO - codeparrot_training - Step 22251: {'lr': 0.00031072917274645656, 'samples': 11393024, 'steps': 22251, 'loss/train': 1.8299438953399658} 02/25/2022 06:24:29 - INFO - codeparrot_training - Step 22252: {'lr': 0.00031071330025695266, 'samples': 11393536, 'steps': 22252, 'loss/train': 2.5882506370544434} 02/25/2022 06:24:35 - INFO - codeparrot_training - Step 22253: {'lr': 0.0003106974275073722, 'samples': 11394048, 'steps': 22253, 'loss/train': 1.676809549331665} 02/25/2022 06:24:39 - INFO - codeparrot_training - Step 22254: {'lr': 0.0003106815544977833, 'samples': 11394560, 'steps': 22254, 'loss/train': 1.1664416790008545} 02/25/2022 06:24:46 - INFO - codeparrot_training - Step 22255: {'lr': 0.00031066568122825383, 'samples': 11395072, 'steps': 22255, 'loss/train': 2.075968027114868} 02/25/2022 06:24:49 - INFO - codeparrot_training - Step 22256: {'lr': 0.0003106498076988519, 'samples': 11395584, 'steps': 22256, 'loss/train': 2.494464159011841} 02/25/2022 06:24:55 - INFO - codeparrot_training - Step 22257: {'lr': 0.0003106339339096454, 'samples': 11396096, 'steps': 22257, 'loss/train': 1.869131326675415} 02/25/2022 06:24:58 - INFO - codeparrot_training - Step 22258: {'lr': 0.0003106180598607024, 'samples': 11396608, 'steps': 22258, 'loss/train': 2.144507646560669} 02/25/2022 06:25:04 - INFO - codeparrot_training - Step 22259: {'lr': 0.00031060218555209094, 'samples': 11397120, 'steps': 22259, 'loss/train': 1.9071341753005981} 02/25/2022 06:25:07 - INFO - codeparrot_training - Step 22260: {'lr': 0.0003105863109838789, 'samples': 11397632, 'steps': 22260, 'loss/train': 1.9524785280227661} 02/25/2022 06:25:13 - INFO - codeparrot_training - Step 22261: {'lr': 0.0003105704361561343, 'samples': 11398144, 'steps': 22261, 'loss/train': 1.5737897157669067} 02/25/2022 06:25:16 - INFO - codeparrot_training - Step 22262: {'lr': 0.00031055456106892526, 'samples': 11398656, 'steps': 22262, 'loss/train': 2.6269309520721436} 02/25/2022 06:25:22 - INFO - codeparrot_training - Step 22263: {'lr': 0.0003105386857223197, 'samples': 11399168, 'steps': 22263, 'loss/train': 1.1933808326721191} 02/25/2022 06:25:25 - INFO - codeparrot_training - Step 22264: {'lr': 0.0003105228101163856, 'samples': 11399680, 'steps': 22264, 'loss/train': 1.6305445432662964} 02/25/2022 06:25:33 - INFO - codeparrot_training - Step 22265: {'lr': 0.000310506934251191, 'samples': 11400192, 'steps': 22265, 'loss/train': 2.9968018531799316} 02/25/2022 06:25:36 - INFO - codeparrot_training - Step 22266: {'lr': 0.0003104910581268039, 'samples': 11400704, 'steps': 22266, 'loss/train': 1.2612520456314087} 02/25/2022 06:25:42 - INFO - codeparrot_training - Step 22267: {'lr': 0.00031047518174329234, 'samples': 11401216, 'steps': 22267, 'loss/train': 1.0289467573165894} 02/25/2022 06:25:45 - INFO - codeparrot_training - Step 22268: {'lr': 0.00031045930510072427, 'samples': 11401728, 'steps': 22268, 'loss/train': 1.7810465097427368} 02/25/2022 06:25:51 - INFO - codeparrot_training - Step 22269: {'lr': 0.00031044342819916784, 'samples': 11402240, 'steps': 22269, 'loss/train': 2.5308055877685547} 02/25/2022 06:25:54 - INFO - codeparrot_training - Step 22270: {'lr': 0.0003104275510386908, 'samples': 11402752, 'steps': 22270, 'loss/train': 3.20619797706604} 02/25/2022 06:26:00 - INFO - codeparrot_training - Step 22271: {'lr': 0.0003104116736193615, 'samples': 11403264, 'steps': 22271, 'loss/train': 0.8564321398735046} 02/25/2022 06:26:03 - INFO - codeparrot_training - Step 22272: {'lr': 0.00031039579594124763, 'samples': 11403776, 'steps': 22272, 'loss/train': 2.959088087081909} 02/25/2022 06:26:09 - INFO - codeparrot_training - Step 22273: {'lr': 0.0003103799180044174, 'samples': 11404288, 'steps': 22273, 'loss/train': 1.844323754310608} 02/25/2022 06:26:12 - INFO - codeparrot_training - Step 22274: {'lr': 0.00031036403980893874, 'samples': 11404800, 'steps': 22274, 'loss/train': 1.3666863441467285} 02/25/2022 06:26:18 - INFO - codeparrot_training - Step 22275: {'lr': 0.0003103481613548797, 'samples': 11405312, 'steps': 22275, 'loss/train': 1.5148844718933105} 02/25/2022 06:26:21 - INFO - codeparrot_training - Step 22276: {'lr': 0.00031033228264230834, 'samples': 11405824, 'steps': 22276, 'loss/train': 1.8171871900558472} 02/25/2022 06:26:27 - INFO - codeparrot_training - Step 22277: {'lr': 0.0003103164036712926, 'samples': 11406336, 'steps': 22277, 'loss/train': 1.1824021339416504} 02/25/2022 06:26:30 - INFO - codeparrot_training - Step 22278: {'lr': 0.0003103005244419006, 'samples': 11406848, 'steps': 22278, 'loss/train': 1.8386484384536743} 02/25/2022 06:26:37 - INFO - codeparrot_training - Step 22279: {'lr': 0.00031028464495420026, 'samples': 11407360, 'steps': 22279, 'loss/train': 0.8456698656082153} 02/25/2022 06:26:41 - INFO - codeparrot_training - Step 22280: {'lr': 0.0003102687652082597, 'samples': 11407872, 'steps': 22280, 'loss/train': 2.0187506675720215} 02/25/2022 06:26:47 - INFO - codeparrot_training - Step 22281: {'lr': 0.00031025288520414686, 'samples': 11408384, 'steps': 22281, 'loss/train': 1.7374745607376099} 02/25/2022 06:26:50 - INFO - codeparrot_training - Step 22282: {'lr': 0.0003102370049419297, 'samples': 11408896, 'steps': 22282, 'loss/train': 0.9777387380599976} 02/25/2022 06:26:56 - INFO - codeparrot_training - Step 22283: {'lr': 0.0003102211244216764, 'samples': 11409408, 'steps': 22283, 'loss/train': 1.8480573892593384} 02/25/2022 06:26:59 - INFO - codeparrot_training - Step 22284: {'lr': 0.000310205243643455, 'samples': 11409920, 'steps': 22284, 'loss/train': 1.9416848421096802} 02/25/2022 06:27:05 - INFO - codeparrot_training - Step 22285: {'lr': 0.00031018936260733337, 'samples': 11410432, 'steps': 22285, 'loss/train': 1.6788299083709717} 02/25/2022 06:27:08 - INFO - codeparrot_training - Step 22286: {'lr': 0.00031017348131337963, 'samples': 11410944, 'steps': 22286, 'loss/train': 1.5511828660964966} 02/25/2022 06:27:14 - INFO - codeparrot_training - Step 22287: {'lr': 0.00031015759976166186, 'samples': 11411456, 'steps': 22287, 'loss/train': 0.8448664546012878} 02/25/2022 06:27:17 - INFO - codeparrot_training - Step 22288: {'lr': 0.00031014171795224794, 'samples': 11411968, 'steps': 22288, 'loss/train': 1.511526107788086} 02/25/2022 06:27:25 - INFO - codeparrot_training - Step 22289: {'lr': 0.00031012583588520607, 'samples': 11412480, 'steps': 22289, 'loss/train': 0.6437299847602844} 02/25/2022 06:27:28 - INFO - codeparrot_training - Step 22290: {'lr': 0.00031010995356060416, 'samples': 11412992, 'steps': 22290, 'loss/train': 2.339708089828491} 02/25/2022 06:27:34 - INFO - codeparrot_training - Step 22291: {'lr': 0.00031009407097851036, 'samples': 11413504, 'steps': 22291, 'loss/train': 1.4100689888000488} 02/25/2022 06:27:37 - INFO - codeparrot_training - Step 22292: {'lr': 0.0003100781881389926, 'samples': 11414016, 'steps': 22292, 'loss/train': 1.462053894996643} 02/25/2022 06:27:42 - INFO - codeparrot_training - Step 22293: {'lr': 0.00031006230504211895, 'samples': 11414528, 'steps': 22293, 'loss/train': 1.3496007919311523} 02/25/2022 06:27:46 - INFO - codeparrot_training - Step 22294: {'lr': 0.0003100464216879574, 'samples': 11415040, 'steps': 22294, 'loss/train': 0.5380419492721558} 02/25/2022 06:27:51 - INFO - codeparrot_training - Step 22295: {'lr': 0.0003100305380765762, 'samples': 11415552, 'steps': 22295, 'loss/train': 1.5153319835662842} 02/25/2022 06:27:55 - INFO - codeparrot_training - Step 22296: {'lr': 0.00031001465420804316, 'samples': 11416064, 'steps': 22296, 'loss/train': 2.651585102081299} 02/25/2022 06:28:00 - INFO - codeparrot_training - Step 22297: {'lr': 0.0003099987700824264, 'samples': 11416576, 'steps': 22297, 'loss/train': 1.4443249702453613} 02/25/2022 06:28:04 - INFO - codeparrot_training - Step 22298: {'lr': 0.00030998288569979393, 'samples': 11417088, 'steps': 22298, 'loss/train': 1.9768311977386475} 02/25/2022 06:28:09 - INFO - codeparrot_training - Step 22299: {'lr': 0.0003099670010602138, 'samples': 11417600, 'steps': 22299, 'loss/train': 1.9314266443252563} 02/25/2022 06:28:13 - INFO - codeparrot_training - Step 22300: {'lr': 0.00030995111616375417, 'samples': 11418112, 'steps': 22300, 'loss/train': 2.394134998321533} 02/25/2022 06:28:20 - INFO - codeparrot_training - Step 22301: {'lr': 0.00030993523101048294, 'samples': 11418624, 'steps': 22301, 'loss/train': 1.2647587060928345} 02/25/2022 06:28:24 - INFO - codeparrot_training - Step 22302: {'lr': 0.0003099193456004682, 'samples': 11419136, 'steps': 22302, 'loss/train': 0.9063059687614441} 02/25/2022 06:28:29 - INFO - codeparrot_training - Step 22303: {'lr': 0.00030990345993377807, 'samples': 11419648, 'steps': 22303, 'loss/train': 1.9191433191299438} 02/25/2022 06:28:33 - INFO - codeparrot_training - Step 22304: {'lr': 0.0003098875740104805, 'samples': 11420160, 'steps': 22304, 'loss/train': 2.153932571411133} 02/25/2022 06:28:38 - INFO - codeparrot_training - Step 22305: {'lr': 0.00030987168783064355, 'samples': 11420672, 'steps': 22305, 'loss/train': 0.8334032893180847} 02/25/2022 06:28:42 - INFO - codeparrot_training - Step 22306: {'lr': 0.0003098558013943353, 'samples': 11421184, 'steps': 22306, 'loss/train': 2.282377243041992} 02/25/2022 06:28:47 - INFO - codeparrot_training - Step 22307: {'lr': 0.00030983991470162386, 'samples': 11421696, 'steps': 22307, 'loss/train': 2.2862932682037354} 02/25/2022 06:28:51 - INFO - codeparrot_training - Step 22308: {'lr': 0.00030982402775257725, 'samples': 11422208, 'steps': 22308, 'loss/train': 0.5140820145606995} 02/25/2022 06:28:56 - INFO - codeparrot_training - Step 22309: {'lr': 0.0003098081405472634, 'samples': 11422720, 'steps': 22309, 'loss/train': 2.301920175552368} 02/25/2022 06:29:00 - INFO - codeparrot_training - Step 22310: {'lr': 0.0003097922530857505, 'samples': 11423232, 'steps': 22310, 'loss/train': 1.5764747858047485} 02/25/2022 06:29:07 - INFO - codeparrot_training - Step 22311: {'lr': 0.0003097763653681066, 'samples': 11423744, 'steps': 22311, 'loss/train': 1.5203237533569336} 02/25/2022 06:29:11 - INFO - codeparrot_training - Step 22312: {'lr': 0.00030976047739439974, 'samples': 11424256, 'steps': 22312, 'loss/train': 0.7984387278556824} 02/25/2022 06:29:16 - INFO - codeparrot_training - Step 22313: {'lr': 0.000309744589164698, 'samples': 11424768, 'steps': 22313, 'loss/train': 1.663845181465149} 02/25/2022 06:29:20 - INFO - codeparrot_training - Step 22314: {'lr': 0.00030972870067906934, 'samples': 11425280, 'steps': 22314, 'loss/train': 1.4970442056655884} 02/25/2022 06:29:25 - INFO - codeparrot_training - Step 22315: {'lr': 0.00030971281193758197, 'samples': 11425792, 'steps': 22315, 'loss/train': 2.329415798187256} 02/25/2022 06:29:29 - INFO - codeparrot_training - Step 22316: {'lr': 0.00030969692294030376, 'samples': 11426304, 'steps': 22316, 'loss/train': 1.0922491550445557} 02/25/2022 06:29:34 - INFO - codeparrot_training - Step 22317: {'lr': 0.000309681033687303, 'samples': 11426816, 'steps': 22317, 'loss/train': 1.4815539121627808} 02/25/2022 06:29:38 - INFO - codeparrot_training - Step 22318: {'lr': 0.0003096651441786476, 'samples': 11427328, 'steps': 22318, 'loss/train': 7.239895820617676} 02/25/2022 06:29:43 - INFO - codeparrot_training - Step 22319: {'lr': 0.0003096492544144056, 'samples': 11427840, 'steps': 22319, 'loss/train': 1.5799311399459839} 02/25/2022 06:29:47 - INFO - codeparrot_training - Step 22320: {'lr': 0.00030963336439464523, 'samples': 11428352, 'steps': 22320, 'loss/train': 2.619276523590088} 02/25/2022 06:29:54 - INFO - codeparrot_training - Step 22321: {'lr': 0.0003096174741194344, 'samples': 11428864, 'steps': 22321, 'loss/train': 1.6923171281814575} 02/25/2022 06:29:57 - INFO - codeparrot_training - Step 22322: {'lr': 0.00030960158358884127, 'samples': 11429376, 'steps': 22322, 'loss/train': 2.112273931503296} 02/25/2022 06:30:03 - INFO - codeparrot_training - Step 22323: {'lr': 0.0003095856928029339, 'samples': 11429888, 'steps': 22323, 'loss/train': 0.857179582118988} 02/25/2022 06:30:06 - INFO - codeparrot_training - Step 22324: {'lr': 0.00030956980176178033, 'samples': 11430400, 'steps': 22324, 'loss/train': 1.9736242294311523} 02/25/2022 06:30:12 - INFO - codeparrot_training - Step 22325: {'lr': 0.00030955391046544865, 'samples': 11430912, 'steps': 22325, 'loss/train': 2.1934549808502197} 02/25/2022 06:30:16 - INFO - codeparrot_training - Step 22326: {'lr': 0.0003095380189140069, 'samples': 11431424, 'steps': 22326, 'loss/train': 3.762875556945801} 02/25/2022 06:30:21 - INFO - codeparrot_training - Step 22327: {'lr': 0.00030952212710752325, 'samples': 11431936, 'steps': 22327, 'loss/train': 2.293222188949585} 02/25/2022 06:30:25 - INFO - codeparrot_training - Step 22328: {'lr': 0.00030950623504606565, 'samples': 11432448, 'steps': 22328, 'loss/train': 1.8966045379638672} 02/25/2022 06:30:31 - INFO - codeparrot_training - Step 22329: {'lr': 0.0003094903427297023, 'samples': 11432960, 'steps': 22329, 'loss/train': 1.7513986825942993} 02/25/2022 06:30:35 - INFO - codeparrot_training - Step 22330: {'lr': 0.00030947445015850114, 'samples': 11433472, 'steps': 22330, 'loss/train': 2.750603199005127} 02/25/2022 06:30:38 - INFO - codeparrot_training - Step 22331: {'lr': 0.0003094585573325303, 'samples': 11433984, 'steps': 22331, 'loss/train': 2.1039719581604004} 02/25/2022 06:30:44 - INFO - codeparrot_training - Step 22332: {'lr': 0.00030944266425185794, 'samples': 11434496, 'steps': 22332, 'loss/train': 3.583482265472412} 02/25/2022 06:30:47 - INFO - codeparrot_training - Step 22333: {'lr': 0.000309426770916552, 'samples': 11435008, 'steps': 22333, 'loss/train': 0.25457778573036194} 02/25/2022 06:30:53 - INFO - codeparrot_training - Step 22334: {'lr': 0.0003094108773266808, 'samples': 11435520, 'steps': 22334, 'loss/train': 1.7558698654174805} 02/25/2022 06:30:56 - INFO - codeparrot_training - Step 22335: {'lr': 0.00030939498348231214, 'samples': 11436032, 'steps': 22335, 'loss/train': 1.6925777196884155} 02/25/2022 06:31:02 - INFO - codeparrot_training - Step 22336: {'lr': 0.00030937908938351424, 'samples': 11436544, 'steps': 22336, 'loss/train': 1.1856589317321777} 02/25/2022 06:31:05 - INFO - codeparrot_training - Step 22337: {'lr': 0.0003093631950303552, 'samples': 11437056, 'steps': 22337, 'loss/train': 1.2924731969833374} 02/25/2022 06:31:13 - INFO - codeparrot_training - Step 22338: {'lr': 0.00030934730042290304, 'samples': 11437568, 'steps': 22338, 'loss/train': 1.7724864482879639} 02/25/2022 06:31:16 - INFO - codeparrot_training - Step 22339: {'lr': 0.00030933140556122597, 'samples': 11438080, 'steps': 22339, 'loss/train': 1.5146642923355103} 02/25/2022 06:31:22 - INFO - codeparrot_training - Step 22340: {'lr': 0.00030931551044539196, 'samples': 11438592, 'steps': 22340, 'loss/train': 2.4913101196289062} 02/25/2022 06:31:25 - INFO - codeparrot_training - Step 22341: {'lr': 0.00030929961507546915, 'samples': 11439104, 'steps': 22341, 'loss/train': 2.6212687492370605} 02/25/2022 06:31:31 - INFO - codeparrot_training - Step 22342: {'lr': 0.0003092837194515256, 'samples': 11439616, 'steps': 22342, 'loss/train': 1.9838272333145142} 02/25/2022 06:31:34 - INFO - codeparrot_training - Step 22343: {'lr': 0.00030926782357362944, 'samples': 11440128, 'steps': 22343, 'loss/train': 2.157978057861328} 02/25/2022 06:31:40 - INFO - codeparrot_training - Step 22344: {'lr': 0.0003092519274418487, 'samples': 11440640, 'steps': 22344, 'loss/train': 3.0668160915374756} 02/25/2022 06:31:43 - INFO - codeparrot_training - Step 22345: {'lr': 0.0003092360310562516, 'samples': 11441152, 'steps': 22345, 'loss/train': 1.5791271924972534} 02/25/2022 06:31:49 - INFO - codeparrot_training - Step 22346: {'lr': 0.00030922013441690607, 'samples': 11441664, 'steps': 22346, 'loss/train': 2.672255277633667} 02/25/2022 06:31:53 - INFO - codeparrot_training - Step 22347: {'lr': 0.0003092042375238803, 'samples': 11442176, 'steps': 22347, 'loss/train': 3.058821439743042} 02/25/2022 06:31:56 - INFO - codeparrot_training - Step 22348: {'lr': 0.00030918834037724235, 'samples': 11442688, 'steps': 22348, 'loss/train': 2.3949058055877686} 02/25/2022 06:32:03 - INFO - codeparrot_training - Step 22349: {'lr': 0.0003091724429770604, 'samples': 11443200, 'steps': 22349, 'loss/train': 1.860541582107544} 02/25/2022 06:32:07 - INFO - codeparrot_training - Step 22350: {'lr': 0.0003091565453234025, 'samples': 11443712, 'steps': 22350, 'loss/train': 2.587707757949829} 02/25/2022 06:32:12 - INFO - codeparrot_training - Step 22351: {'lr': 0.0003091406474163367, 'samples': 11444224, 'steps': 22351, 'loss/train': 2.50187611579895} 02/25/2022 06:32:16 - INFO - codeparrot_training - Step 22352: {'lr': 0.0003091247492559312, 'samples': 11444736, 'steps': 22352, 'loss/train': 1.755619764328003} 02/25/2022 06:32:21 - INFO - codeparrot_training - Step 22353: {'lr': 0.00030910885084225396, 'samples': 11445248, 'steps': 22353, 'loss/train': 2.078305959701538} 02/25/2022 06:32:25 - INFO - codeparrot_training - Step 22354: {'lr': 0.0003090929521753733, 'samples': 11445760, 'steps': 22354, 'loss/train': 1.433426022529602} 02/25/2022 06:32:30 - INFO - codeparrot_training - Step 22355: {'lr': 0.00030907705325535704, 'samples': 11446272, 'steps': 22355, 'loss/train': 1.9524354934692383} 02/25/2022 06:32:34 - INFO - codeparrot_training - Step 22356: {'lr': 0.0003090611540822736, 'samples': 11446784, 'steps': 22356, 'loss/train': 1.4155231714248657} 02/25/2022 06:32:39 - INFO - codeparrot_training - Step 22357: {'lr': 0.0003090452546561908, 'samples': 11447296, 'steps': 22357, 'loss/train': 1.8659199476242065} 02/25/2022 06:32:42 - INFO - codeparrot_training - Step 22358: {'lr': 0.000309029354977177, 'samples': 11447808, 'steps': 22358, 'loss/train': 1.7597140073776245} 02/25/2022 06:32:50 - INFO - codeparrot_training - Step 22359: {'lr': 0.00030901345504530007, 'samples': 11448320, 'steps': 22359, 'loss/train': 1.6660492420196533} 02/25/2022 06:32:53 - INFO - codeparrot_training - Step 22360: {'lr': 0.0003089975548606283, 'samples': 11448832, 'steps': 22360, 'loss/train': 1.9426867961883545} 02/25/2022 06:32:59 - INFO - codeparrot_training - Step 22361: {'lr': 0.0003089816544232298, 'samples': 11449344, 'steps': 22361, 'loss/train': 2.3940839767456055} 02/25/2022 06:33:02 - INFO - codeparrot_training - Step 22362: {'lr': 0.00030896575373317247, 'samples': 11449856, 'steps': 22362, 'loss/train': 1.415296196937561} 02/25/2022 06:33:08 - INFO - codeparrot_training - Step 22363: {'lr': 0.0003089498527905247, 'samples': 11450368, 'steps': 22363, 'loss/train': 0.4083438515663147} 02/25/2022 06:33:11 - INFO - codeparrot_training - Step 22364: {'lr': 0.00030893395159535444, 'samples': 11450880, 'steps': 22364, 'loss/train': 1.8211491107940674} 02/25/2022 06:33:17 - INFO - codeparrot_training - Step 22365: {'lr': 0.00030891805014772987, 'samples': 11451392, 'steps': 22365, 'loss/train': 1.9234308004379272} 02/25/2022 06:33:20 - INFO - codeparrot_training - Step 22366: {'lr': 0.000308902148447719, 'samples': 11451904, 'steps': 22366, 'loss/train': 2.0892865657806396} 02/25/2022 06:33:26 - INFO - codeparrot_training - Step 22367: {'lr': 0.00030888624649539015, 'samples': 11452416, 'steps': 22367, 'loss/train': 1.4807833433151245} 02/25/2022 06:33:29 - INFO - codeparrot_training - Step 22368: {'lr': 0.0003088703442908112, 'samples': 11452928, 'steps': 22368, 'loss/train': 2.2732961177825928} 02/25/2022 06:33:37 - INFO - codeparrot_training - Step 22369: {'lr': 0.0003088544418340505, 'samples': 11453440, 'steps': 22369, 'loss/train': 0.744574248790741} 02/25/2022 06:33:40 - INFO - codeparrot_training - Step 22370: {'lr': 0.000308838539125176, 'samples': 11453952, 'steps': 22370, 'loss/train': 2.6158666610717773} 02/25/2022 06:33:46 - INFO - codeparrot_training - Step 22371: {'lr': 0.00030882263616425587, 'samples': 11454464, 'steps': 22371, 'loss/train': 2.2508530616760254} 02/25/2022 06:33:49 - INFO - codeparrot_training - Step 22372: {'lr': 0.0003088067329513583, 'samples': 11454976, 'steps': 22372, 'loss/train': 2.4914891719818115} 02/25/2022 06:33:55 - INFO - codeparrot_training - Step 22373: {'lr': 0.0003087908294865513, 'samples': 11455488, 'steps': 22373, 'loss/train': 1.806899070739746} 02/25/2022 06:33:58 - INFO - codeparrot_training - Step 22374: {'lr': 0.00030877492576990306, 'samples': 11456000, 'steps': 22374, 'loss/train': 1.1799976825714111} 02/25/2022 06:34:04 - INFO - codeparrot_training - Step 22375: {'lr': 0.0003087590218014817, 'samples': 11456512, 'steps': 22375, 'loss/train': 2.5723395347595215} 02/25/2022 06:34:07 - INFO - codeparrot_training - Step 22376: {'lr': 0.00030874311758135535, 'samples': 11457024, 'steps': 22376, 'loss/train': 1.8258635997772217} 02/25/2022 06:34:13 - INFO - codeparrot_training - Step 22377: {'lr': 0.00030872721310959216, 'samples': 11457536, 'steps': 22377, 'loss/train': 1.6009079217910767} 02/25/2022 06:34:19 - INFO - codeparrot_training - Step 22378: {'lr': 0.00030871130838626025, 'samples': 11458048, 'steps': 22378, 'loss/train': 1.4505228996276855} 02/25/2022 06:34:22 - INFO - codeparrot_training - Step 22379: {'lr': 0.0003086954034114277, 'samples': 11458560, 'steps': 22379, 'loss/train': 2.5557217597961426} 02/25/2022 06:34:26 - INFO - codeparrot_training - Step 22380: {'lr': 0.00030867949818516274, 'samples': 11459072, 'steps': 22380, 'loss/train': 2.4533960819244385} 02/25/2022 06:34:31 - INFO - codeparrot_training - Step 22381: {'lr': 0.00030866359270753337, 'samples': 11459584, 'steps': 22381, 'loss/train': 0.8030892014503479} 02/25/2022 06:34:35 - INFO - codeparrot_training - Step 22382: {'lr': 0.00030864768697860784, 'samples': 11460096, 'steps': 22382, 'loss/train': 1.9053959846496582} 02/25/2022 06:34:40 - INFO - codeparrot_training - Step 22383: {'lr': 0.0003086317809984542, 'samples': 11460608, 'steps': 22383, 'loss/train': 3.842362642288208} 02/25/2022 06:34:44 - INFO - codeparrot_training - Step 22384: {'lr': 0.0003086158747671406, 'samples': 11461120, 'steps': 22384, 'loss/train': 0.8807100057601929} 02/25/2022 06:34:51 - INFO - codeparrot_training - Step 22385: {'lr': 0.0003085999682847353, 'samples': 11461632, 'steps': 22385, 'loss/train': 1.4275264739990234} 02/25/2022 06:34:54 - INFO - codeparrot_training - Step 22386: {'lr': 0.00030858406155130625, 'samples': 11462144, 'steps': 22386, 'loss/train': 1.7387988567352295} 02/25/2022 06:35:00 - INFO - codeparrot_training - Step 22387: {'lr': 0.00030856815456692177, 'samples': 11462656, 'steps': 22387, 'loss/train': 2.484557628631592} 02/25/2022 06:35:03 - INFO - codeparrot_training - Step 22388: {'lr': 0.00030855224733164987, 'samples': 11463168, 'steps': 22388, 'loss/train': 1.8729536533355713} 02/25/2022 06:35:09 - INFO - codeparrot_training - Step 22389: {'lr': 0.00030853633984555875, 'samples': 11463680, 'steps': 22389, 'loss/train': 2.02634859085083} 02/25/2022 06:35:12 - INFO - codeparrot_training - Step 22390: {'lr': 0.0003085204321087165, 'samples': 11464192, 'steps': 22390, 'loss/train': 8.760651588439941} 02/25/2022 06:35:18 - INFO - codeparrot_training - Step 22391: {'lr': 0.00030850452412119135, 'samples': 11464704, 'steps': 22391, 'loss/train': 1.6826472282409668} 02/25/2022 06:35:21 - INFO - codeparrot_training - Step 22392: {'lr': 0.00030848861588305136, 'samples': 11465216, 'steps': 22392, 'loss/train': 2.1556930541992188} 02/25/2022 06:35:27 - INFO - codeparrot_training - Step 22393: {'lr': 0.0003084727073943648, 'samples': 11465728, 'steps': 22393, 'loss/train': 2.5755672454833984} 02/25/2022 06:35:31 - INFO - codeparrot_training - Step 22394: {'lr': 0.0003084567986551996, 'samples': 11466240, 'steps': 22394, 'loss/train': 2.094748020172119} 02/25/2022 06:35:38 - INFO - codeparrot_training - Step 22395: {'lr': 0.0003084408896656241, 'samples': 11466752, 'steps': 22395, 'loss/train': 1.9773863554000854} 02/25/2022 06:35:41 - INFO - codeparrot_training - Step 22396: {'lr': 0.0003084249804257064, 'samples': 11467264, 'steps': 22396, 'loss/train': 1.619940996170044} 02/25/2022 06:35:47 - INFO - codeparrot_training - Step 22397: {'lr': 0.00030840907093551456, 'samples': 11467776, 'steps': 22397, 'loss/train': 1.7091479301452637} 02/25/2022 06:35:50 - INFO - codeparrot_training - Step 22398: {'lr': 0.0003083931611951169, 'samples': 11468288, 'steps': 22398, 'loss/train': 0.6624564528465271} 02/25/2022 06:35:56 - INFO - codeparrot_training - Step 22399: {'lr': 0.0003083772512045814, 'samples': 11468800, 'steps': 22399, 'loss/train': 1.4868395328521729} 02/25/2022 06:35:59 - INFO - codeparrot_training - Step 22400: {'lr': 0.0003083613409639764, 'samples': 11469312, 'steps': 22400, 'loss/train': 2.3430211544036865} 02/25/2022 06:36:05 - INFO - codeparrot_training - Step 22401: {'lr': 0.0003083454304733698, 'samples': 11469824, 'steps': 22401, 'loss/train': 2.3103830814361572} 02/25/2022 06:36:08 - INFO - codeparrot_training - Step 22402: {'lr': 0.00030832951973283, 'samples': 11470336, 'steps': 22402, 'loss/train': 1.3244562149047852} 02/25/2022 06:36:14 - INFO - codeparrot_training - Step 22403: {'lr': 0.000308313608742425, 'samples': 11470848, 'steps': 22403, 'loss/train': 1.2192047834396362} 02/25/2022 06:36:17 - INFO - codeparrot_training - Step 22404: {'lr': 0.00030829769750222315, 'samples': 11471360, 'steps': 22404, 'loss/train': 1.416987657546997} 02/25/2022 06:36:23 - INFO - codeparrot_training - Step 22405: {'lr': 0.00030828178601229235, 'samples': 11471872, 'steps': 22405, 'loss/train': 1.8768166303634644} 02/25/2022 06:36:27 - INFO - codeparrot_training - Step 22406: {'lr': 0.00030826587427270095, 'samples': 11472384, 'steps': 22406, 'loss/train': 1.5243351459503174} 02/25/2022 06:36:32 - INFO - codeparrot_training - Step 22407: {'lr': 0.000308249962283517, 'samples': 11472896, 'steps': 22407, 'loss/train': 1.476779580116272} 02/25/2022 06:36:36 - INFO - codeparrot_training - Step 22408: {'lr': 0.0003082340500448087, 'samples': 11473408, 'steps': 22408, 'loss/train': 2.0404796600341797} 02/25/2022 06:36:42 - INFO - codeparrot_training - Step 22409: {'lr': 0.0003082181375566443, 'samples': 11473920, 'steps': 22409, 'loss/train': 2.1903507709503174} 02/25/2022 06:36:45 - INFO - codeparrot_training - Step 22410: {'lr': 0.0003082022248190918, 'samples': 11474432, 'steps': 22410, 'loss/train': 1.9735642671585083} 02/25/2022 06:36:51 - INFO - codeparrot_training - Step 22411: {'lr': 0.00030818631183221945, 'samples': 11474944, 'steps': 22411, 'loss/train': 1.6710984706878662} 02/25/2022 06:36:54 - INFO - codeparrot_training - Step 22412: {'lr': 0.0003081703985960955, 'samples': 11475456, 'steps': 22412, 'loss/train': 2.2112793922424316} 02/25/2022 06:37:00 - INFO - codeparrot_training - Step 22413: {'lr': 0.00030815448511078796, 'samples': 11475968, 'steps': 22413, 'loss/train': 1.548136830329895} 02/25/2022 06:37:03 - INFO - codeparrot_training - Step 22414: {'lr': 0.0003081385713763651, 'samples': 11476480, 'steps': 22414, 'loss/train': 1.352513074874878} 02/25/2022 06:37:09 - INFO - codeparrot_training - Step 22415: {'lr': 0.00030812265739289503, 'samples': 11476992, 'steps': 22415, 'loss/train': 2.00291109085083} 02/25/2022 06:37:13 - INFO - codeparrot_training - Step 22416: {'lr': 0.000308106743160446, 'samples': 11477504, 'steps': 22416, 'loss/train': 1.8740191459655762} 02/25/2022 06:37:18 - INFO - codeparrot_training - Step 22417: {'lr': 0.00030809082867908614, 'samples': 11478016, 'steps': 22417, 'loss/train': 1.6866564750671387} 02/25/2022 06:37:22 - INFO - codeparrot_training - Step 22418: {'lr': 0.0003080749139488836, 'samples': 11478528, 'steps': 22418, 'loss/train': 1.7677775621414185} 02/25/2022 06:37:27 - INFO - codeparrot_training - Step 22419: {'lr': 0.0003080589989699066, 'samples': 11479040, 'steps': 22419, 'loss/train': 5.318525791168213} 02/25/2022 06:37:31 - INFO - codeparrot_training - Step 22420: {'lr': 0.00030804308374222315, 'samples': 11479552, 'steps': 22420, 'loss/train': 1.7014484405517578} 02/25/2022 06:37:36 - INFO - codeparrot_training - Step 22421: {'lr': 0.00030802716826590164, 'samples': 11480064, 'steps': 22421, 'loss/train': 1.4869434833526611} 02/25/2022 06:37:39 - INFO - codeparrot_training - Step 22422: {'lr': 0.0003080112525410102, 'samples': 11480576, 'steps': 22422, 'loss/train': 2.2668511867523193} 02/25/2022 06:37:45 - INFO - codeparrot_training - Step 22423: {'lr': 0.0003079953365676169, 'samples': 11481088, 'steps': 22423, 'loss/train': 0.7269768714904785} 02/25/2022 06:37:49 - INFO - codeparrot_training - Step 22424: {'lr': 0.00030797942034579013, 'samples': 11481600, 'steps': 22424, 'loss/train': 1.968507170677185} 02/25/2022 06:37:55 - INFO - codeparrot_training - Step 22425: {'lr': 0.0003079635038755978, 'samples': 11482112, 'steps': 22425, 'loss/train': 0.9637570381164551} 02/25/2022 06:37:58 - INFO - codeparrot_training - Step 22426: {'lr': 0.0003079475871571083, 'samples': 11482624, 'steps': 22426, 'loss/train': 1.7919520139694214} 02/25/2022 06:38:04 - INFO - codeparrot_training - Step 22427: {'lr': 0.0003079316701903897, 'samples': 11483136, 'steps': 22427, 'loss/train': 1.6670854091644287} 02/25/2022 06:38:07 - INFO - codeparrot_training - Step 22428: {'lr': 0.0003079157529755102, 'samples': 11483648, 'steps': 22428, 'loss/train': 0.763849139213562} 02/25/2022 06:38:13 - INFO - codeparrot_training - Step 22429: {'lr': 0.0003078998355125381, 'samples': 11484160, 'steps': 22429, 'loss/train': 1.8103898763656616} 02/25/2022 06:38:16 - INFO - codeparrot_training - Step 22430: {'lr': 0.0003078839178015414, 'samples': 11484672, 'steps': 22430, 'loss/train': 1.5743341445922852} 02/25/2022 06:38:22 - INFO - codeparrot_training - Step 22431: {'lr': 0.0003078679998425884, 'samples': 11485184, 'steps': 22431, 'loss/train': 2.0739946365356445} 02/25/2022 06:38:25 - INFO - codeparrot_training - Step 22432: {'lr': 0.0003078520816357472, 'samples': 11485696, 'steps': 22432, 'loss/train': 2.7872555255889893} 02/25/2022 06:38:31 - INFO - codeparrot_training - Step 22433: {'lr': 0.00030783616318108613, 'samples': 11486208, 'steps': 22433, 'loss/train': 1.7650415897369385} 02/25/2022 06:38:35 - INFO - codeparrot_training - Step 22434: {'lr': 0.0003078202444786733, 'samples': 11486720, 'steps': 22434, 'loss/train': 1.876613736152649} 02/25/2022 06:38:40 - INFO - codeparrot_training - Step 22435: {'lr': 0.0003078043255285769, 'samples': 11487232, 'steps': 22435, 'loss/train': 0.8877896666526794} 02/25/2022 06:38:44 - INFO - codeparrot_training - Step 22436: {'lr': 0.00030778840633086514, 'samples': 11487744, 'steps': 22436, 'loss/train': 3.0490896701812744} 02/25/2022 06:38:49 - INFO - codeparrot_training - Step 22437: {'lr': 0.00030777248688560615, 'samples': 11488256, 'steps': 22437, 'loss/train': 2.067582368850708} 02/25/2022 06:38:53 - INFO - codeparrot_training - Step 22438: {'lr': 0.0003077565671928682, 'samples': 11488768, 'steps': 22438, 'loss/train': 2.9970216751098633} 02/25/2022 06:38:58 - INFO - codeparrot_training - Step 22439: {'lr': 0.00030774064725271944, 'samples': 11489280, 'steps': 22439, 'loss/train': 0.872682511806488} 02/25/2022 06:39:02 - INFO - codeparrot_training - Step 22440: {'lr': 0.00030772472706522806, 'samples': 11489792, 'steps': 22440, 'loss/train': 9.422595024108887} 02/25/2022 06:39:05 - INFO - codeparrot_training - Step 22441: {'lr': 0.00030770880663046236, 'samples': 11490304, 'steps': 22441, 'loss/train': 2.216752052307129} 02/25/2022 06:39:11 - INFO - codeparrot_training - Step 22442: {'lr': 0.00030769288594849044, 'samples': 11490816, 'steps': 22442, 'loss/train': 1.1536403894424438} 02/25/2022 06:39:17 - INFO - codeparrot_training - Step 22443: {'lr': 0.0003076769650193805, 'samples': 11491328, 'steps': 22443, 'loss/train': 2.079317331314087} 02/25/2022 06:39:20 - INFO - codeparrot_training - Step 22444: {'lr': 0.0003076610438432007, 'samples': 11491840, 'steps': 22444, 'loss/train': 1.0838345289230347} 02/25/2022 06:39:26 - INFO - codeparrot_training - Step 22445: {'lr': 0.00030764512242001927, 'samples': 11492352, 'steps': 22445, 'loss/train': 1.7356657981872559} 02/25/2022 06:39:29 - INFO - codeparrot_training - Step 22446: {'lr': 0.00030762920074990457, 'samples': 11492864, 'steps': 22446, 'loss/train': 1.089828372001648} 02/25/2022 06:39:33 - INFO - codeparrot_training - Step 22447: {'lr': 0.00030761327883292456, 'samples': 11493376, 'steps': 22447, 'loss/train': 2.4420320987701416} 02/25/2022 06:39:39 - INFO - codeparrot_training - Step 22448: {'lr': 0.00030759735666914767, 'samples': 11493888, 'steps': 22448, 'loss/train': 1.6346549987792969} 02/25/2022 06:39:42 - INFO - codeparrot_training - Step 22449: {'lr': 0.00030758143425864187, 'samples': 11494400, 'steps': 22449, 'loss/train': 0.16934257745742798} 02/25/2022 06:39:47 - INFO - codeparrot_training - Step 22450: {'lr': 0.00030756551160147563, 'samples': 11494912, 'steps': 22450, 'loss/train': 2.335850715637207} 02/25/2022 06:39:51 - INFO - codeparrot_training - Step 22451: {'lr': 0.0003075495886977169, 'samples': 11495424, 'steps': 22451, 'loss/train': 2.307386636734009} 02/25/2022 06:39:57 - INFO - codeparrot_training - Step 22452: {'lr': 0.0003075336655474341, 'samples': 11495936, 'steps': 22452, 'loss/train': 1.7017930746078491} 02/25/2022 06:40:01 - INFO - codeparrot_training - Step 22453: {'lr': 0.0003075177421506952, 'samples': 11496448, 'steps': 22453, 'loss/train': 1.9930847883224487} 02/25/2022 06:40:06 - INFO - codeparrot_training - Step 22454: {'lr': 0.0003075018185075687, 'samples': 11496960, 'steps': 22454, 'loss/train': 2.1117916107177734} 02/25/2022 06:40:12 - INFO - codeparrot_training - Step 22455: {'lr': 0.0003074858946181226, 'samples': 11497472, 'steps': 22455, 'loss/train': 1.4634958505630493} 02/25/2022 06:40:15 - INFO - codeparrot_training - Step 22456: {'lr': 0.0003074699704824252, 'samples': 11497984, 'steps': 22456, 'loss/train': 2.485975980758667} 02/25/2022 06:40:21 - INFO - codeparrot_training - Step 22457: {'lr': 0.0003074540461005447, 'samples': 11498496, 'steps': 22457, 'loss/train': 2.007704496383667} 02/25/2022 06:40:24 - INFO - codeparrot_training - Step 22458: {'lr': 0.00030743812147254935, 'samples': 11499008, 'steps': 22458, 'loss/train': 1.2556360960006714} 02/25/2022 06:40:30 - INFO - codeparrot_training - Step 22459: {'lr': 0.0003074221965985073, 'samples': 11499520, 'steps': 22459, 'loss/train': 2.854020595550537} 02/25/2022 06:40:33 - INFO - codeparrot_training - Step 22460: {'lr': 0.0003074062714784867, 'samples': 11500032, 'steps': 22460, 'loss/train': 1.9867686033248901} 02/25/2022 06:40:39 - INFO - codeparrot_training - Step 22461: {'lr': 0.000307390346112556, 'samples': 11500544, 'steps': 22461, 'loss/train': 2.059323310852051} 02/25/2022 06:40:43 - INFO - codeparrot_training - Step 22462: {'lr': 0.0003073744205007832, 'samples': 11501056, 'steps': 22462, 'loss/train': 1.5115104913711548} 02/25/2022 06:40:48 - INFO - codeparrot_training - Step 22463: {'lr': 0.0003073584946432366, 'samples': 11501568, 'steps': 22463, 'loss/train': 1.714331865310669} 02/25/2022 06:40:52 - INFO - codeparrot_training - Step 22464: {'lr': 0.00030734256853998446, 'samples': 11502080, 'steps': 22464, 'loss/train': 2.7452924251556396} 02/25/2022 06:40:57 - INFO - codeparrot_training - Step 22465: {'lr': 0.00030732664219109497, 'samples': 11502592, 'steps': 22465, 'loss/train': 1.7802088260650635} 02/25/2022 06:41:01 - INFO - codeparrot_training - Step 22466: {'lr': 0.00030731071559663624, 'samples': 11503104, 'steps': 22466, 'loss/train': 2.6749212741851807} 02/25/2022 06:41:06 - INFO - codeparrot_training - Step 22467: {'lr': 0.0003072947887566768, 'samples': 11503616, 'steps': 22467, 'loss/train': 2.516126871109009} 02/25/2022 06:41:10 - INFO - codeparrot_training - Step 22468: {'lr': 0.0003072788616712845, 'samples': 11504128, 'steps': 22468, 'loss/train': 2.287215232849121} 02/25/2022 06:41:15 - INFO - codeparrot_training - Step 22469: {'lr': 0.0003072629343405278, 'samples': 11504640, 'steps': 22469, 'loss/train': 2.2294540405273438} 02/25/2022 06:41:19 - INFO - codeparrot_training - Step 22470: {'lr': 0.00030724700676447485, 'samples': 11505152, 'steps': 22470, 'loss/train': 2.8658690452575684} 02/25/2022 06:41:24 - INFO - codeparrot_training - Step 22471: {'lr': 0.00030723107894319393, 'samples': 11505664, 'steps': 22471, 'loss/train': 1.7094885110855103} 02/25/2022 06:41:28 - INFO - codeparrot_training - Step 22472: {'lr': 0.00030721515087675326, 'samples': 11506176, 'steps': 22472, 'loss/train': 2.535473585128784} 02/25/2022 06:41:33 - INFO - codeparrot_training - Step 22473: {'lr': 0.00030719922256522105, 'samples': 11506688, 'steps': 22473, 'loss/train': 2.112297534942627} 02/25/2022 06:41:37 - INFO - codeparrot_training - Step 22474: {'lr': 0.0003071832940086655, 'samples': 11507200, 'steps': 22474, 'loss/train': 1.5509376525878906} 02/25/2022 06:41:42 - INFO - codeparrot_training - Step 22475: {'lr': 0.0003071673652071549, 'samples': 11507712, 'steps': 22475, 'loss/train': 1.7644466161727905} 02/25/2022 06:41:46 - INFO - codeparrot_training - Step 22476: {'lr': 0.0003071514361607575, 'samples': 11508224, 'steps': 22476, 'loss/train': 1.521186351776123} 02/25/2022 06:41:52 - INFO - codeparrot_training - Step 22477: {'lr': 0.0003071355068695414, 'samples': 11508736, 'steps': 22477, 'loss/train': 1.7518539428710938} 02/25/2022 06:41:56 - INFO - codeparrot_training - Step 22478: {'lr': 0.000307119577333575, 'samples': 11509248, 'steps': 22478, 'loss/train': 1.415287971496582} 02/25/2022 06:42:01 - INFO - codeparrot_training - Step 22479: {'lr': 0.0003071036475529264, 'samples': 11509760, 'steps': 22479, 'loss/train': 2.7647347450256348} 02/25/2022 06:42:05 - INFO - codeparrot_training - Step 22480: {'lr': 0.00030708771752766396, 'samples': 11510272, 'steps': 22480, 'loss/train': 0.5685231685638428} 02/25/2022 06:42:10 - INFO - codeparrot_training - Step 22481: {'lr': 0.0003070717872578558, 'samples': 11510784, 'steps': 22481, 'loss/train': 2.328578472137451} 02/25/2022 06:42:14 - INFO - codeparrot_training - Step 22482: {'lr': 0.0003070558567435703, 'samples': 11511296, 'steps': 22482, 'loss/train': 1.4170125722885132} 02/25/2022 06:42:19 - INFO - codeparrot_training - Step 22483: {'lr': 0.00030703992598487564, 'samples': 11511808, 'steps': 22483, 'loss/train': 2.0182225704193115} 02/25/2022 06:42:23 - INFO - codeparrot_training - Step 22484: {'lr': 0.00030702399498184005, 'samples': 11512320, 'steps': 22484, 'loss/train': 2.1395626068115234} 02/25/2022 06:42:28 - INFO - codeparrot_training - Step 22485: {'lr': 0.0003070080637345317, 'samples': 11512832, 'steps': 22485, 'loss/train': 2.0802996158599854} 02/25/2022 06:42:32 - INFO - codeparrot_training - Step 22486: {'lr': 0.00030699213224301896, 'samples': 11513344, 'steps': 22486, 'loss/train': 1.6864551305770874} 02/25/2022 06:42:38 - INFO - codeparrot_training - Step 22487: {'lr': 0.00030697620050737, 'samples': 11513856, 'steps': 22487, 'loss/train': 1.5342485904693604} 02/25/2022 06:42:42 - INFO - codeparrot_training - Step 22488: {'lr': 0.00030696026852765313, 'samples': 11514368, 'steps': 22488, 'loss/train': 2.027613401412964} 02/25/2022 06:42:47 - INFO - codeparrot_training - Step 22489: {'lr': 0.00030694433630393654, 'samples': 11514880, 'steps': 22489, 'loss/train': 2.235450267791748} 02/25/2022 06:42:51 - INFO - codeparrot_training - Step 22490: {'lr': 0.00030692840383628845, 'samples': 11515392, 'steps': 22490, 'loss/train': 2.5430638790130615} 02/25/2022 06:42:56 - INFO - codeparrot_training - Step 22491: {'lr': 0.0003069124711247772, 'samples': 11515904, 'steps': 22491, 'loss/train': 1.5179001092910767} 02/25/2022 06:42:59 - INFO - codeparrot_training - Step 22492: {'lr': 0.000306896538169471, 'samples': 11516416, 'steps': 22492, 'loss/train': 1.5861786603927612} 02/25/2022 06:43:05 - INFO - codeparrot_training - Step 22493: {'lr': 0.000306880604970438, 'samples': 11516928, 'steps': 22493, 'loss/train': 2.1795713901519775} 02/25/2022 06:43:09 - INFO - codeparrot_training - Step 22494: {'lr': 0.00030686467152774667, 'samples': 11517440, 'steps': 22494, 'loss/train': 1.5836273431777954} 02/25/2022 06:43:14 - INFO - codeparrot_training - Step 22495: {'lr': 0.0003068487378414651, 'samples': 11517952, 'steps': 22495, 'loss/train': 1.8670686483383179} 02/25/2022 06:43:18 - INFO - codeparrot_training - Step 22496: {'lr': 0.0003068328039116616, 'samples': 11518464, 'steps': 22496, 'loss/train': 2.420645236968994} 02/25/2022 06:43:23 - INFO - codeparrot_training - Step 22497: {'lr': 0.0003068168697384044, 'samples': 11518976, 'steps': 22497, 'loss/train': 1.6798491477966309} 02/25/2022 06:43:27 - INFO - codeparrot_training - Step 22498: {'lr': 0.0003068009353217618, 'samples': 11519488, 'steps': 22498, 'loss/train': 1.7972851991653442} 02/25/2022 06:43:33 - INFO - codeparrot_training - Step 22499: {'lr': 0.00030678500066180206, 'samples': 11520000, 'steps': 22499, 'loss/train': 1.7461363077163696} 02/25/2022 06:43:36 - INFO - codeparrot_training - Step 22500: {'lr': 0.0003067690657585933, 'samples': 11520512, 'steps': 22500, 'loss/train': 2.513232707977295} 02/25/2022 06:43:42 - INFO - codeparrot_training - Step 22501: {'lr': 0.000306753130612204, 'samples': 11521024, 'steps': 22501, 'loss/train': 1.3351699113845825} 02/25/2022 06:43:45 - INFO - codeparrot_training - Step 22502: {'lr': 0.0003067371952227022, 'samples': 11521536, 'steps': 22502, 'loss/train': 0.22736585140228271} 02/25/2022 06:43:51 - INFO - codeparrot_training - Step 22503: {'lr': 0.0003067212595901564, 'samples': 11522048, 'steps': 22503, 'loss/train': 2.130584716796875} 02/25/2022 06:43:54 - INFO - codeparrot_training - Step 22504: {'lr': 0.00030670532371463463, 'samples': 11522560, 'steps': 22504, 'loss/train': 1.7992980480194092} 02/25/2022 06:44:00 - INFO - codeparrot_training - Step 22505: {'lr': 0.0003066893875962053, 'samples': 11523072, 'steps': 22505, 'loss/train': 1.839167833328247} 02/25/2022 06:44:03 - INFO - codeparrot_training - Step 22506: {'lr': 0.0003066734512349366, 'samples': 11523584, 'steps': 22506, 'loss/train': 0.9062573909759521} 02/25/2022 06:44:09 - INFO - codeparrot_training - Step 22507: {'lr': 0.00030665751463089687, 'samples': 11524096, 'steps': 22507, 'loss/train': 2.3805291652679443} 02/25/2022 06:44:12 - INFO - codeparrot_training - Step 22508: {'lr': 0.0003066415777841543, 'samples': 11524608, 'steps': 22508, 'loss/train': 2.0209567546844482} 02/25/2022 06:44:18 - INFO - codeparrot_training - Step 22509: {'lr': 0.0003066256406947773, 'samples': 11525120, 'steps': 22509, 'loss/train': 2.0216166973114014} 02/25/2022 06:44:22 - INFO - codeparrot_training - Step 22510: {'lr': 0.0003066097033628339, 'samples': 11525632, 'steps': 22510, 'loss/train': 3.337493658065796} 02/25/2022 06:44:27 - INFO - codeparrot_training - Step 22511: {'lr': 0.0003065937657883926, 'samples': 11526144, 'steps': 22511, 'loss/train': 2.563512086868286} 02/25/2022 06:44:31 - INFO - codeparrot_training - Step 22512: {'lr': 0.0003065778279715215, 'samples': 11526656, 'steps': 22512, 'loss/train': 1.7350542545318604} 02/25/2022 06:44:36 - INFO - codeparrot_training - Step 22513: {'lr': 0.000306561889912289, 'samples': 11527168, 'steps': 22513, 'loss/train': 1.6011396646499634} 02/25/2022 06:44:40 - INFO - codeparrot_training - Step 22514: {'lr': 0.0003065459516107633, 'samples': 11527680, 'steps': 22514, 'loss/train': 2.065385580062866} 02/25/2022 06:44:45 - INFO - codeparrot_training - Step 22515: {'lr': 0.0003065300130670127, 'samples': 11528192, 'steps': 22515, 'loss/train': 1.7015151977539062} 02/25/2022 06:44:49 - INFO - codeparrot_training - Step 22516: {'lr': 0.0003065140742811055, 'samples': 11528704, 'steps': 22516, 'loss/train': 1.1297411918640137} 02/25/2022 06:44:54 - INFO - codeparrot_training - Step 22517: {'lr': 0.0003064981352531099, 'samples': 11529216, 'steps': 22517, 'loss/train': 1.5894063711166382} 02/25/2022 06:44:58 - INFO - codeparrot_training - Step 22518: {'lr': 0.0003064821959830943, 'samples': 11529728, 'steps': 22518, 'loss/train': 2.866004228591919} 02/25/2022 06:45:04 - INFO - codeparrot_training - Step 22519: {'lr': 0.00030646625647112686, 'samples': 11530240, 'steps': 22519, 'loss/train': 1.7332532405853271} 02/25/2022 06:45:08 - INFO - codeparrot_training - Step 22520: {'lr': 0.000306450316717276, 'samples': 11530752, 'steps': 22520, 'loss/train': 1.2880399227142334} 02/25/2022 06:45:13 - INFO - codeparrot_training - Step 22521: {'lr': 0.0003064343767216098, 'samples': 11531264, 'steps': 22521, 'loss/train': 3.3929717540740967} 02/25/2022 06:45:17 - INFO - codeparrot_training - Step 22522: {'lr': 0.00030641843648419664, 'samples': 11531776, 'steps': 22522, 'loss/train': 1.7112188339233398} 02/25/2022 06:45:22 - INFO - codeparrot_training - Step 22523: {'lr': 0.0003064024960051048, 'samples': 11532288, 'steps': 22523, 'loss/train': 2.9232795238494873} 02/25/2022 06:45:26 - INFO - codeparrot_training - Step 22524: {'lr': 0.00030638655528440273, 'samples': 11532800, 'steps': 22524, 'loss/train': 1.5316267013549805} 02/25/2022 06:45:31 - INFO - codeparrot_training - Step 22525: {'lr': 0.0003063706143221584, 'samples': 11533312, 'steps': 22525, 'loss/train': 1.7459677457809448} 02/25/2022 06:45:35 - INFO - codeparrot_training - Step 22526: {'lr': 0.00030635467311844033, 'samples': 11533824, 'steps': 22526, 'loss/train': 1.469390869140625} 02/25/2022 06:45:40 - INFO - codeparrot_training - Step 22527: {'lr': 0.00030633873167331674, 'samples': 11534336, 'steps': 22527, 'loss/train': 2.3531734943389893} 02/25/2022 06:45:44 - INFO - codeparrot_training - Step 22528: {'lr': 0.0003063227899868559, 'samples': 11534848, 'steps': 22528, 'loss/train': 1.3716802597045898} 02/25/2022 06:45:49 - INFO - codeparrot_training - Step 22529: {'lr': 0.00030630684805912613, 'samples': 11535360, 'steps': 22529, 'loss/train': 2.041776657104492} 02/25/2022 06:45:53 - INFO - codeparrot_training - Step 22530: {'lr': 0.00030629090589019567, 'samples': 11535872, 'steps': 22530, 'loss/train': 2.340679168701172} 02/25/2022 06:45:58 - INFO - codeparrot_training - Step 22531: {'lr': 0.00030627496348013285, 'samples': 11536384, 'steps': 22531, 'loss/train': 2.0246407985687256} 02/25/2022 06:46:02 - INFO - codeparrot_training - Step 22532: {'lr': 0.0003062590208290059, 'samples': 11536896, 'steps': 22532, 'loss/train': 2.2053232192993164} 02/25/2022 06:46:07 - INFO - codeparrot_training - Step 22533: {'lr': 0.00030624307793688334, 'samples': 11537408, 'steps': 22533, 'loss/train': 2.2110037803649902} 02/25/2022 06:46:11 - INFO - codeparrot_training - Step 22534: {'lr': 0.00030622713480383314, 'samples': 11537920, 'steps': 22534, 'loss/train': 1.9606841802597046} 02/25/2022 06:46:17 - INFO - codeparrot_training - Step 22535: {'lr': 0.0003062111914299238, 'samples': 11538432, 'steps': 22535, 'loss/train': 2.604990005493164} 02/25/2022 06:46:20 - INFO - codeparrot_training - Step 22536: {'lr': 0.0003061952478152236, 'samples': 11538944, 'steps': 22536, 'loss/train': 1.3748756647109985} 02/25/2022 06:46:26 - INFO - codeparrot_training - Step 22537: {'lr': 0.0003061793039598007, 'samples': 11539456, 'steps': 22537, 'loss/train': 1.0223584175109863} 02/25/2022 06:46:29 - INFO - codeparrot_training - Step 22538: {'lr': 0.0003061633598637236, 'samples': 11539968, 'steps': 22538, 'loss/train': 1.9234931468963623} 02/25/2022 06:46:35 - INFO - codeparrot_training - Step 22539: {'lr': 0.00030614741552706045, 'samples': 11540480, 'steps': 22539, 'loss/train': 2.226428508758545} 02/25/2022 06:46:38 - INFO - codeparrot_training - Step 22540: {'lr': 0.0003061314709498796, 'samples': 11540992, 'steps': 22540, 'loss/train': 0.689652144908905} 02/25/2022 06:46:44 - INFO - codeparrot_training - Step 22541: {'lr': 0.00030611552613224935, 'samples': 11541504, 'steps': 22541, 'loss/train': 2.6056675910949707} 02/25/2022 06:46:47 - INFO - codeparrot_training - Step 22542: {'lr': 0.00030609958107423804, 'samples': 11542016, 'steps': 22542, 'loss/train': 2.230349540710449} 02/25/2022 06:46:54 - INFO - codeparrot_training - Step 22543: {'lr': 0.0003060836357759139, 'samples': 11542528, 'steps': 22543, 'loss/train': 1.6245497465133667} 02/25/2022 06:46:57 - INFO - codeparrot_training - Step 22544: {'lr': 0.00030606769023734534, 'samples': 11543040, 'steps': 22544, 'loss/train': 1.721929907798767} 02/25/2022 06:47:03 - INFO - codeparrot_training - Step 22545: {'lr': 0.0003060517444586005, 'samples': 11543552, 'steps': 22545, 'loss/train': 1.7743027210235596} 02/25/2022 06:47:06 - INFO - codeparrot_training - Step 22546: {'lr': 0.0003060357984397479, 'samples': 11544064, 'steps': 22546, 'loss/train': 2.000258684158325} 02/25/2022 06:47:12 - INFO - codeparrot_training - Step 22547: {'lr': 0.00030601985218085565, 'samples': 11544576, 'steps': 22547, 'loss/train': 8.809915542602539} 02/25/2022 06:47:15 - INFO - codeparrot_training - Step 22548: {'lr': 0.00030600390568199213, 'samples': 11545088, 'steps': 22548, 'loss/train': 3.437715768814087} 02/25/2022 06:47:21 - INFO - codeparrot_training - Step 22549: {'lr': 0.00030598795894322574, 'samples': 11545600, 'steps': 22549, 'loss/train': 2.2694318294525146} 02/25/2022 06:47:24 - INFO - codeparrot_training - Step 22550: {'lr': 0.00030597201196462466, 'samples': 11546112, 'steps': 22550, 'loss/train': 0.8924354314804077} 02/25/2022 06:47:30 - INFO - codeparrot_training - Step 22551: {'lr': 0.00030595606474625726, 'samples': 11546624, 'steps': 22551, 'loss/train': 1.3325576782226562} 02/25/2022 06:47:33 - INFO - codeparrot_training - Step 22552: {'lr': 0.00030594011728819184, 'samples': 11547136, 'steps': 22552, 'loss/train': 1.0910444259643555} 02/25/2022 06:47:39 - INFO - codeparrot_training - Step 22553: {'lr': 0.00030592416959049666, 'samples': 11547648, 'steps': 22553, 'loss/train': 0.49670591950416565} 02/25/2022 06:47:42 - INFO - codeparrot_training - Step 22554: {'lr': 0.00030590822165324017, 'samples': 11548160, 'steps': 22554, 'loss/train': 2.146921157836914} 02/25/2022 06:47:48 - INFO - codeparrot_training - Step 22555: {'lr': 0.00030589227347649063, 'samples': 11548672, 'steps': 22555, 'loss/train': 2.43127179145813} 02/25/2022 06:47:52 - INFO - codeparrot_training - Step 22556: {'lr': 0.00030587632506031624, 'samples': 11549184, 'steps': 22556, 'loss/train': 2.116577625274658} 02/25/2022 06:47:58 - INFO - codeparrot_training - Step 22557: {'lr': 0.0003058603764047855, 'samples': 11549696, 'steps': 22557, 'loss/train': 1.4954054355621338} 02/25/2022 06:48:01 - INFO - codeparrot_training - Step 22558: {'lr': 0.00030584442750996666, 'samples': 11550208, 'steps': 22558, 'loss/train': 2.4073333740234375} 02/25/2022 06:48:06 - INFO - codeparrot_training - Step 22559: {'lr': 0.000305828478375928, 'samples': 11550720, 'steps': 22559, 'loss/train': 1.8182624578475952} 02/25/2022 06:48:10 - INFO - codeparrot_training - Step 22560: {'lr': 0.00030581252900273786, 'samples': 11551232, 'steps': 22560, 'loss/train': 1.1319001913070679} 02/25/2022 06:48:15 - INFO - codeparrot_training - Step 22561: {'lr': 0.0003057965793904646, 'samples': 11551744, 'steps': 22561, 'loss/train': 2.161604642868042} 02/25/2022 06:48:19 - INFO - codeparrot_training - Step 22562: {'lr': 0.00030578062953917645, 'samples': 11552256, 'steps': 22562, 'loss/train': 1.9175020456314087} 02/25/2022 06:48:24 - INFO - codeparrot_training - Step 22563: {'lr': 0.00030576467944894186, 'samples': 11552768, 'steps': 22563, 'loss/train': 2.0127949714660645} 02/25/2022 06:48:28 - INFO - codeparrot_training - Step 22564: {'lr': 0.000305748729119829, 'samples': 11553280, 'steps': 22564, 'loss/train': 1.4060845375061035} 02/25/2022 06:48:33 - INFO - codeparrot_training - Step 22565: {'lr': 0.00030573277855190634, 'samples': 11553792, 'steps': 22565, 'loss/train': 2.1600468158721924} 02/25/2022 06:48:37 - INFO - codeparrot_training - Step 22566: {'lr': 0.0003057168277452422, 'samples': 11554304, 'steps': 22566, 'loss/train': 1.054341435432434} 02/25/2022 06:48:42 - INFO - codeparrot_training - Step 22567: {'lr': 0.0003057008766999048, 'samples': 11554816, 'steps': 22567, 'loss/train': 1.209962010383606} 02/25/2022 06:48:46 - INFO - codeparrot_training - Step 22568: {'lr': 0.0003056849254159625, 'samples': 11555328, 'steps': 22568, 'loss/train': 1.5850061178207397} 02/25/2022 06:48:52 - INFO - codeparrot_training - Step 22569: {'lr': 0.00030566897389348375, 'samples': 11555840, 'steps': 22569, 'loss/train': 3.0002191066741943} 02/25/2022 06:48:56 - INFO - codeparrot_training - Step 22570: {'lr': 0.00030565302213253677, 'samples': 11556352, 'steps': 22570, 'loss/train': 2.1758604049682617} 02/25/2022 06:49:01 - INFO - codeparrot_training - Step 22571: {'lr': 0.0003056370701331899, 'samples': 11556864, 'steps': 22571, 'loss/train': 2.4038541316986084} 02/25/2022 06:49:05 - INFO - codeparrot_training - Step 22572: {'lr': 0.0003056211178955115, 'samples': 11557376, 'steps': 22572, 'loss/train': 1.7791229486465454} 02/25/2022 06:49:10 - INFO - codeparrot_training - Step 22573: {'lr': 0.00030560516541956983, 'samples': 11557888, 'steps': 22573, 'loss/train': 1.9129211902618408} 02/25/2022 06:49:14 - INFO - codeparrot_training - Step 22574: {'lr': 0.0003055892127054334, 'samples': 11558400, 'steps': 22574, 'loss/train': 1.517530918121338} 02/25/2022 06:49:19 - INFO - codeparrot_training - Step 22575: {'lr': 0.00030557325975317037, 'samples': 11558912, 'steps': 22575, 'loss/train': 1.8406928777694702} 02/25/2022 06:49:23 - INFO - codeparrot_training - Step 22576: {'lr': 0.00030555730656284914, 'samples': 11559424, 'steps': 22576, 'loss/train': 1.956127643585205} 02/25/2022 06:49:28 - INFO - codeparrot_training - Step 22577: {'lr': 0.000305541353134538, 'samples': 11559936, 'steps': 22577, 'loss/train': 1.8516039848327637} 02/25/2022 06:49:32 - INFO - codeparrot_training - Step 22578: {'lr': 0.0003055253994683054, 'samples': 11560448, 'steps': 22578, 'loss/train': 1.3101212978363037} 02/25/2022 06:49:37 - INFO - codeparrot_training - Step 22579: {'lr': 0.0003055094455642196, 'samples': 11560960, 'steps': 22579, 'loss/train': 1.2690118551254272} 02/25/2022 06:49:41 - INFO - codeparrot_training - Step 22580: {'lr': 0.000305493491422349, 'samples': 11561472, 'steps': 22580, 'loss/train': 1.714158058166504} 02/25/2022 06:49:47 - INFO - codeparrot_training - Step 22581: {'lr': 0.0003054775370427619, 'samples': 11561984, 'steps': 22581, 'loss/train': 1.7413963079452515} 02/25/2022 06:49:50 - INFO - codeparrot_training - Step 22582: {'lr': 0.00030546158242552657, 'samples': 11562496, 'steps': 22582, 'loss/train': 1.9933078289031982} 02/25/2022 06:49:56 - INFO - codeparrot_training - Step 22583: {'lr': 0.00030544562757071154, 'samples': 11563008, 'steps': 22583, 'loss/train': 1.5428571701049805} 02/25/2022 06:49:59 - INFO - codeparrot_training - Step 22584: {'lr': 0.000305429672478385, 'samples': 11563520, 'steps': 22584, 'loss/train': 1.779737949371338} 02/25/2022 06:50:05 - INFO - codeparrot_training - Step 22585: {'lr': 0.0003054137171486153, 'samples': 11564032, 'steps': 22585, 'loss/train': 1.7626997232437134} 02/25/2022 06:50:08 - INFO - codeparrot_training - Step 22586: {'lr': 0.0003053977615814709, 'samples': 11564544, 'steps': 22586, 'loss/train': 2.107056140899658} 02/25/2022 06:50:14 - INFO - codeparrot_training - Step 22587: {'lr': 0.00030538180577702005, 'samples': 11565056, 'steps': 22587, 'loss/train': 1.5872962474822998} 02/25/2022 06:50:17 - INFO - codeparrot_training - Step 22588: {'lr': 0.00030536584973533113, 'samples': 11565568, 'steps': 22588, 'loss/train': 1.234094262123108} 02/25/2022 06:50:23 - INFO - codeparrot_training - Step 22589: {'lr': 0.0003053498934564725, 'samples': 11566080, 'steps': 22589, 'loss/train': 1.825188159942627} 02/25/2022 06:50:27 - INFO - codeparrot_training - Step 22590: {'lr': 0.00030533393694051256, 'samples': 11566592, 'steps': 22590, 'loss/train': 1.8414422273635864} 02/25/2022 06:50:33 - INFO - codeparrot_training - Step 22591: {'lr': 0.0003053179801875195, 'samples': 11567104, 'steps': 22591, 'loss/train': 0.3257031738758087} 02/25/2022 06:50:36 - INFO - codeparrot_training - Step 22592: {'lr': 0.00030530202319756184, 'samples': 11567616, 'steps': 22592, 'loss/train': 1.338649034500122} 02/25/2022 06:50:41 - INFO - codeparrot_training - Step 22593: {'lr': 0.0003052860659707078, 'samples': 11568128, 'steps': 22593, 'loss/train': 2.513063669204712} 02/25/2022 06:50:47 - INFO - codeparrot_training - Step 22594: {'lr': 0.0003052701085070259, 'samples': 11568640, 'steps': 22594, 'loss/train': 2.3820440769195557} 02/25/2022 06:50:50 - INFO - codeparrot_training - Step 22595: {'lr': 0.00030525415080658437, 'samples': 11569152, 'steps': 22595, 'loss/train': 2.8618366718292236} 02/25/2022 06:50:56 - INFO - codeparrot_training - Step 22596: {'lr': 0.0003052381928694516, 'samples': 11569664, 'steps': 22596, 'loss/train': 2.079991340637207} 02/25/2022 06:50:59 - INFO - codeparrot_training - Step 22597: {'lr': 0.00030522223469569594, 'samples': 11570176, 'steps': 22597, 'loss/train': 1.8576678037643433} 02/25/2022 06:51:05 - INFO - codeparrot_training - Step 22598: {'lr': 0.00030520627628538577, 'samples': 11570688, 'steps': 22598, 'loss/train': 2.1208832263946533} 02/25/2022 06:51:08 - INFO - codeparrot_training - Step 22599: {'lr': 0.0003051903176385894, 'samples': 11571200, 'steps': 22599, 'loss/train': 1.9460619688034058} 02/25/2022 06:51:15 - INFO - codeparrot_training - Step 22600: {'lr': 0.00030517435875537536, 'samples': 11571712, 'steps': 22600, 'loss/train': 0.8828034400939941} 02/25/2022 06:51:18 - INFO - codeparrot_training - Step 22601: {'lr': 0.0003051583996358118, 'samples': 11572224, 'steps': 22601, 'loss/train': 2.2202556133270264} 02/25/2022 06:51:24 - INFO - codeparrot_training - Step 22602: {'lr': 0.00030514244027996705, 'samples': 11572736, 'steps': 22602, 'loss/train': 1.241764783859253} 02/25/2022 06:51:27 - INFO - codeparrot_training - Step 22603: {'lr': 0.00030512648068790985, 'samples': 11573248, 'steps': 22603, 'loss/train': 1.4387774467468262} 02/25/2022 06:51:33 - INFO - codeparrot_training - Step 22604: {'lr': 0.0003051105208597081, 'samples': 11573760, 'steps': 22604, 'loss/train': 1.8418468236923218} 02/25/2022 06:51:36 - INFO - codeparrot_training - Step 22605: {'lr': 0.00030509456079543044, 'samples': 11574272, 'steps': 22605, 'loss/train': 1.8847973346710205} 02/25/2022 06:51:42 - INFO - codeparrot_training - Step 22606: {'lr': 0.0003050786004951452, 'samples': 11574784, 'steps': 22606, 'loss/train': 1.0247842073440552} 02/25/2022 06:51:45 - INFO - codeparrot_training - Step 22607: {'lr': 0.00030506263995892075, 'samples': 11575296, 'steps': 22607, 'loss/train': 2.8790903091430664} 02/25/2022 06:51:51 - INFO - codeparrot_training - Step 22608: {'lr': 0.0003050466791868254, 'samples': 11575808, 'steps': 22608, 'loss/train': 2.366002321243286} 02/25/2022 06:51:54 - INFO - codeparrot_training - Step 22609: {'lr': 0.0003050307181789276, 'samples': 11576320, 'steps': 22609, 'loss/train': 0.2187565416097641} 02/25/2022 06:52:00 - INFO - codeparrot_training - Step 22610: {'lr': 0.0003050147569352956, 'samples': 11576832, 'steps': 22610, 'loss/train': 1.1164970397949219} 02/25/2022 06:52:03 - INFO - codeparrot_training - Step 22611: {'lr': 0.0003049987954559979, 'samples': 11577344, 'steps': 22611, 'loss/train': 2.433436870574951} 02/25/2022 06:52:09 - INFO - codeparrot_training - Step 22612: {'lr': 0.0003049828337411028, 'samples': 11577856, 'steps': 22612, 'loss/train': 1.5017539262771606} 02/25/2022 06:52:12 - INFO - codeparrot_training - Step 22613: {'lr': 0.00030496687179067865, 'samples': 11578368, 'steps': 22613, 'loss/train': 2.2317848205566406} 02/25/2022 06:52:18 - INFO - codeparrot_training - Step 22614: {'lr': 0.000304950909604794, 'samples': 11578880, 'steps': 22614, 'loss/train': 2.0177857875823975} 02/25/2022 06:52:22 - INFO - codeparrot_training - Step 22615: {'lr': 0.000304934947183517, 'samples': 11579392, 'steps': 22615, 'loss/train': 0.43425843119621277} 02/25/2022 06:52:27 - INFO - codeparrot_training - Step 22616: {'lr': 0.00030491898452691626, 'samples': 11579904, 'steps': 22616, 'loss/train': 1.6392916440963745} 02/25/2022 06:52:31 - INFO - codeparrot_training - Step 22617: {'lr': 0.0003049030216350599, 'samples': 11580416, 'steps': 22617, 'loss/train': 2.577558755874634} 02/25/2022 06:52:36 - INFO - codeparrot_training - Step 22618: {'lr': 0.00030488705850801646, 'samples': 11580928, 'steps': 22618, 'loss/train': 1.6446527242660522} 02/25/2022 06:52:40 - INFO - codeparrot_training - Step 22619: {'lr': 0.00030487109514585426, 'samples': 11581440, 'steps': 22619, 'loss/train': 2.1492116451263428} 02/25/2022 06:52:45 - INFO - codeparrot_training - Step 22620: {'lr': 0.0003048551315486418, 'samples': 11581952, 'steps': 22620, 'loss/train': 2.0937840938568115} 02/25/2022 06:52:49 - INFO - codeparrot_training - Step 22621: {'lr': 0.00030483916771644734, 'samples': 11582464, 'steps': 22621, 'loss/train': 1.7621278762817383} 02/25/2022 06:52:54 - INFO - codeparrot_training - Step 22622: {'lr': 0.0003048232036493392, 'samples': 11582976, 'steps': 22622, 'loss/train': 4.088159084320068} 02/25/2022 06:52:58 - INFO - codeparrot_training - Step 22623: {'lr': 0.00030480723934738597, 'samples': 11583488, 'steps': 22623, 'loss/train': 0.8258746266365051} 02/25/2022 06:53:03 - INFO - codeparrot_training - Step 22624: {'lr': 0.00030479127481065595, 'samples': 11584000, 'steps': 22624, 'loss/train': 1.5218144655227661} 02/25/2022 06:53:07 - INFO - codeparrot_training - Step 22625: {'lr': 0.0003047753100392174, 'samples': 11584512, 'steps': 22625, 'loss/train': 2.1196799278259277} 02/25/2022 06:53:13 - INFO - codeparrot_training - Step 22626: {'lr': 0.00030475934503313893, 'samples': 11585024, 'steps': 22626, 'loss/train': 1.9896446466445923} 02/25/2022 06:53:17 - INFO - codeparrot_training - Step 22627: {'lr': 0.0003047433797924888, 'samples': 11585536, 'steps': 22627, 'loss/train': 1.3188180923461914} 02/25/2022 06:53:22 - INFO - codeparrot_training - Step 22628: {'lr': 0.00030472741431733535, 'samples': 11586048, 'steps': 22628, 'loss/train': 1.964693307876587} 02/25/2022 06:53:26 - INFO - codeparrot_training - Step 22629: {'lr': 0.0003047114486077471, 'samples': 11586560, 'steps': 22629, 'loss/train': 2.071418046951294} 02/25/2022 06:53:31 - INFO - codeparrot_training - Step 22630: {'lr': 0.0003046954826637923, 'samples': 11587072, 'steps': 22630, 'loss/train': 1.6267136335372925} 02/25/2022 06:53:34 - INFO - codeparrot_training - Step 22631: {'lr': 0.0003046795164855395, 'samples': 11587584, 'steps': 22631, 'loss/train': 1.8542166948318481} 02/25/2022 06:53:40 - INFO - codeparrot_training - Step 22632: {'lr': 0.000304663550073057, 'samples': 11588096, 'steps': 22632, 'loss/train': 1.6350902318954468} 02/25/2022 06:53:43 - INFO - codeparrot_training - Step 22633: {'lr': 0.00030464758342641315, 'samples': 11588608, 'steps': 22633, 'loss/train': 2.111021041870117} 02/25/2022 06:53:49 - INFO - codeparrot_training - Step 22634: {'lr': 0.00030463161654567647, 'samples': 11589120, 'steps': 22634, 'loss/train': 1.3240230083465576} 02/25/2022 06:53:53 - INFO - codeparrot_training - Step 22635: {'lr': 0.00030461564943091524, 'samples': 11589632, 'steps': 22635, 'loss/train': 1.9727163314819336} 02/25/2022 06:53:59 - INFO - codeparrot_training - Step 22636: {'lr': 0.0003045996820821979, 'samples': 11590144, 'steps': 22636, 'loss/train': 2.2866928577423096} 02/25/2022 06:54:02 - INFO - codeparrot_training - Step 22637: {'lr': 0.00030458371449959293, 'samples': 11590656, 'steps': 22637, 'loss/train': 1.6979786157608032} 02/25/2022 06:54:08 - INFO - codeparrot_training - Step 22638: {'lr': 0.00030456774668316864, 'samples': 11591168, 'steps': 22638, 'loss/train': 1.0800909996032715} 02/25/2022 06:54:11 - INFO - codeparrot_training - Step 22639: {'lr': 0.0003045517786329934, 'samples': 11591680, 'steps': 22639, 'loss/train': 1.8978074789047241} 02/25/2022 06:54:17 - INFO - codeparrot_training - Step 22640: {'lr': 0.0003045358103491357, 'samples': 11592192, 'steps': 22640, 'loss/train': 1.728190541267395} 02/25/2022 06:54:20 - INFO - codeparrot_training - Step 22641: {'lr': 0.00030451984183166384, 'samples': 11592704, 'steps': 22641, 'loss/train': 2.319401741027832} 02/25/2022 06:54:26 - INFO - codeparrot_training - Step 22642: {'lr': 0.0003045038730806464, 'samples': 11593216, 'steps': 22642, 'loss/train': 2.149230718612671} 02/25/2022 06:54:29 - INFO - codeparrot_training - Step 22643: {'lr': 0.00030448790409615155, 'samples': 11593728, 'steps': 22643, 'loss/train': 1.6373363733291626} 02/25/2022 06:54:35 - INFO - codeparrot_training - Step 22644: {'lr': 0.00030447193487824796, 'samples': 11594240, 'steps': 22644, 'loss/train': 2.55241322517395} 02/25/2022 06:54:38 - INFO - codeparrot_training - Step 22645: {'lr': 0.00030445596542700383, 'samples': 11594752, 'steps': 22645, 'loss/train': 1.3099442720413208} 02/25/2022 06:54:44 - INFO - codeparrot_training - Step 22646: {'lr': 0.00030443999574248764, 'samples': 11595264, 'steps': 22646, 'loss/train': 1.0037115812301636} 02/25/2022 06:54:48 - INFO - codeparrot_training - Step 22647: {'lr': 0.00030442402582476775, 'samples': 11595776, 'steps': 22647, 'loss/train': 2.4970531463623047} 02/25/2022 06:54:53 - INFO - codeparrot_training - Step 22648: {'lr': 0.00030440805567391274, 'samples': 11596288, 'steps': 22648, 'loss/train': 1.5511350631713867} 02/25/2022 06:54:57 - INFO - codeparrot_training - Step 22649: {'lr': 0.00030439208528999074, 'samples': 11596800, 'steps': 22649, 'loss/train': 1.1514108180999756} 02/25/2022 06:55:02 - INFO - codeparrot_training - Step 22650: {'lr': 0.0003043761146730704, 'samples': 11597312, 'steps': 22650, 'loss/train': 1.9396324157714844} 02/25/2022 06:55:06 - INFO - codeparrot_training - Step 22651: {'lr': 0.00030436014382322004, 'samples': 11597824, 'steps': 22651, 'loss/train': 0.1052892878651619} 02/25/2022 06:55:12 - INFO - codeparrot_training - Step 22652: {'lr': 0.00030434417274050805, 'samples': 11598336, 'steps': 22652, 'loss/train': 2.3675334453582764} 02/25/2022 06:55:15 - INFO - codeparrot_training - Step 22653: {'lr': 0.00030432820142500296, 'samples': 11598848, 'steps': 22653, 'loss/train': 2.3279995918273926} 02/25/2022 06:55:20 - INFO - codeparrot_training - Step 22654: {'lr': 0.00030431222987677305, 'samples': 11599360, 'steps': 22654, 'loss/train': 2.154399871826172} 02/25/2022 06:55:24 - INFO - codeparrot_training - Step 22655: {'lr': 0.0003042962580958868, 'samples': 11599872, 'steps': 22655, 'loss/train': 1.4444462060928345} 02/25/2022 06:55:29 - INFO - codeparrot_training - Step 22656: {'lr': 0.00030428028608241257, 'samples': 11600384, 'steps': 22656, 'loss/train': 2.0426950454711914} 02/25/2022 06:55:33 - INFO - codeparrot_training - Step 22657: {'lr': 0.00030426431383641893, 'samples': 11600896, 'steps': 22657, 'loss/train': 3.1355977058410645} 02/25/2022 06:55:39 - INFO - codeparrot_training - Step 22658: {'lr': 0.00030424834135797413, 'samples': 11601408, 'steps': 22658, 'loss/train': 1.8279017210006714} 02/25/2022 06:55:42 - INFO - codeparrot_training - Step 22659: {'lr': 0.00030423236864714676, 'samples': 11601920, 'steps': 22659, 'loss/train': 1.784300684928894} 02/25/2022 06:55:47 - INFO - codeparrot_training - Step 22660: {'lr': 0.00030421639570400505, 'samples': 11602432, 'steps': 22660, 'loss/train': 1.6259421110153198} 02/25/2022 06:55:51 - INFO - codeparrot_training - Step 22661: {'lr': 0.0003042004225286176, 'samples': 11602944, 'steps': 22661, 'loss/train': 1.819698452949524} 02/25/2022 06:55:57 - INFO - codeparrot_training - Step 22662: {'lr': 0.00030418444912105256, 'samples': 11603456, 'steps': 22662, 'loss/train': 2.13950777053833} 02/25/2022 06:56:01 - INFO - codeparrot_training - Step 22663: {'lr': 0.0003041684754813787, 'samples': 11603968, 'steps': 22663, 'loss/train': 1.4907573461532593} 02/25/2022 06:56:06 - INFO - codeparrot_training - Step 22664: {'lr': 0.0003041525016096643, 'samples': 11604480, 'steps': 22664, 'loss/train': 1.2841544151306152} 02/25/2022 06:56:10 - INFO - codeparrot_training - Step 22665: {'lr': 0.0003041365275059777, 'samples': 11604992, 'steps': 22665, 'loss/train': 1.6701160669326782} 02/25/2022 06:56:15 - INFO - codeparrot_training - Step 22666: {'lr': 0.0003041205531703875, 'samples': 11605504, 'steps': 22666, 'loss/train': 2.068481683731079} 02/25/2022 06:56:19 - INFO - codeparrot_training - Step 22667: {'lr': 0.0003041045786029619, 'samples': 11606016, 'steps': 22667, 'loss/train': 2.0479958057403564} 02/25/2022 06:56:24 - INFO - codeparrot_training - Step 22668: {'lr': 0.0003040886038037696, 'samples': 11606528, 'steps': 22668, 'loss/train': 2.3006644248962402} 02/25/2022 06:56:28 - INFO - codeparrot_training - Step 22669: {'lr': 0.00030407262877287883, 'samples': 11607040, 'steps': 22669, 'loss/train': 0.9008424878120422} 02/25/2022 06:56:33 - INFO - codeparrot_training - Step 22670: {'lr': 0.00030405665351035816, 'samples': 11607552, 'steps': 22670, 'loss/train': 1.136783480644226} 02/25/2022 06:56:37 - INFO - codeparrot_training - Step 22671: {'lr': 0.0003040406780162759, 'samples': 11608064, 'steps': 22671, 'loss/train': 2.953453302383423} 02/25/2022 06:56:43 - INFO - codeparrot_training - Step 22672: {'lr': 0.00030402470229070054, 'samples': 11608576, 'steps': 22672, 'loss/train': 2.416731357574463} 02/25/2022 06:56:47 - INFO - codeparrot_training - Step 22673: {'lr': 0.00030400872633370047, 'samples': 11609088, 'steps': 22673, 'loss/train': 2.1524155139923096} 02/25/2022 06:56:52 - INFO - codeparrot_training - Step 22674: {'lr': 0.0003039927501453442, 'samples': 11609600, 'steps': 22674, 'loss/train': 1.6534168720245361} 02/25/2022 06:56:56 - INFO - codeparrot_training - Step 22675: {'lr': 0.0003039767737257002, 'samples': 11610112, 'steps': 22675, 'loss/train': 2.829561471939087} 02/25/2022 06:57:01 - INFO - codeparrot_training - Step 22676: {'lr': 0.0003039607970748368, 'samples': 11610624, 'steps': 22676, 'loss/train': 1.9732459783554077} 02/25/2022 06:57:05 - INFO - codeparrot_training - Step 22677: {'lr': 0.00030394482019282246, 'samples': 11611136, 'steps': 22677, 'loss/train': 2.2965991497039795} 02/25/2022 06:57:10 - INFO - codeparrot_training - Step 22678: {'lr': 0.0003039288430797256, 'samples': 11611648, 'steps': 22678, 'loss/train': 2.116220235824585} 02/25/2022 06:57:14 - INFO - codeparrot_training - Step 22679: {'lr': 0.0003039128657356147, 'samples': 11612160, 'steps': 22679, 'loss/train': 1.1863832473754883} 02/25/2022 06:57:19 - INFO - codeparrot_training - Step 22680: {'lr': 0.0003038968881605583, 'samples': 11612672, 'steps': 22680, 'loss/train': 1.2606350183486938} 02/25/2022 06:57:23 - INFO - codeparrot_training - Step 22681: {'lr': 0.00030388091035462466, 'samples': 11613184, 'steps': 22681, 'loss/train': 1.0160527229309082} 02/25/2022 06:57:29 - INFO - codeparrot_training - Step 22682: {'lr': 0.00030386493231788234, 'samples': 11613696, 'steps': 22682, 'loss/train': 1.9279738664627075} 02/25/2022 06:57:33 - INFO - codeparrot_training - Step 22683: {'lr': 0.0003038489540503997, 'samples': 11614208, 'steps': 22683, 'loss/train': 2.256998300552368} 02/25/2022 06:57:38 - INFO - codeparrot_training - Step 22684: {'lr': 0.0003038329755522453, 'samples': 11614720, 'steps': 22684, 'loss/train': 0.8462879061698914} 02/25/2022 06:57:42 - INFO - codeparrot_training - Step 22685: {'lr': 0.0003038169968234875, 'samples': 11615232, 'steps': 22685, 'loss/train': 2.325924873352051} 02/25/2022 06:57:47 - INFO - codeparrot_training - Step 22686: {'lr': 0.0003038010178641948, 'samples': 11615744, 'steps': 22686, 'loss/train': 1.0872704982757568} 02/25/2022 06:57:51 - INFO - codeparrot_training - Step 22687: {'lr': 0.00030378503867443555, 'samples': 11616256, 'steps': 22687, 'loss/train': 1.750091314315796} 02/25/2022 06:57:56 - INFO - codeparrot_training - Step 22688: {'lr': 0.0003037690592542784, 'samples': 11616768, 'steps': 22688, 'loss/train': 2.5829105377197266} 02/25/2022 06:58:00 - INFO - codeparrot_training - Step 22689: {'lr': 0.0003037530796037916, 'samples': 11617280, 'steps': 22689, 'loss/train': 1.859145998954773} 02/25/2022 06:58:05 - INFO - codeparrot_training - Step 22690: {'lr': 0.0003037370997230436, 'samples': 11617792, 'steps': 22690, 'loss/train': 1.6390535831451416} 02/25/2022 06:58:09 - INFO - codeparrot_training - Step 22691: {'lr': 0.000303721119612103, 'samples': 11618304, 'steps': 22691, 'loss/train': 1.4562523365020752} 02/25/2022 06:58:15 - INFO - codeparrot_training - Step 22692: {'lr': 0.00030370513927103826, 'samples': 11618816, 'steps': 22692, 'loss/train': 2.485083818435669} 02/25/2022 06:58:18 - INFO - codeparrot_training - Step 22693: {'lr': 0.0003036891586999176, 'samples': 11619328, 'steps': 22693, 'loss/train': 2.0450894832611084} 02/25/2022 06:58:24 - INFO - codeparrot_training - Step 22694: {'lr': 0.00030367317789880985, 'samples': 11619840, 'steps': 22694, 'loss/train': 2.5536746978759766} 02/25/2022 06:58:27 - INFO - codeparrot_training - Step 22695: {'lr': 0.000303657196867783, 'samples': 11620352, 'steps': 22695, 'loss/train': 1.9993013143539429} 02/25/2022 06:58:33 - INFO - codeparrot_training - Step 22696: {'lr': 0.0003036412156069059, 'samples': 11620864, 'steps': 22696, 'loss/train': 1.2640838623046875} 02/25/2022 06:58:36 - INFO - codeparrot_training - Step 22697: {'lr': 0.00030362523411624686, 'samples': 11621376, 'steps': 22697, 'loss/train': 1.169634461402893} 02/25/2022 06:58:42 - INFO - codeparrot_training - Step 22698: {'lr': 0.0003036092523958743, 'samples': 11621888, 'steps': 22698, 'loss/train': 1.7449144124984741} 02/25/2022 06:58:45 - INFO - codeparrot_training - Step 22699: {'lr': 0.00030359327044585685, 'samples': 11622400, 'steps': 22699, 'loss/train': 1.1600278615951538} 02/25/2022 06:58:51 - INFO - codeparrot_training - Step 22700: {'lr': 0.00030357728826626266, 'samples': 11622912, 'steps': 22700, 'loss/train': 1.6789931058883667} 02/25/2022 06:58:55 - INFO - codeparrot_training - Step 22701: {'lr': 0.0003035613058571605, 'samples': 11623424, 'steps': 22701, 'loss/train': 1.6943378448486328} 02/25/2022 06:59:00 - INFO - codeparrot_training - Step 22702: {'lr': 0.00030354532321861865, 'samples': 11623936, 'steps': 22702, 'loss/train': 1.4822139739990234} 02/25/2022 06:59:03 - INFO - codeparrot_training - Step 22703: {'lr': 0.0003035293403507057, 'samples': 11624448, 'steps': 22703, 'loss/train': 2.8570351600646973} 02/25/2022 06:59:09 - INFO - codeparrot_training - Step 22704: {'lr': 0.00030351335725349, 'samples': 11624960, 'steps': 22704, 'loss/train': 2.175259828567505} 02/25/2022 06:59:13 - INFO - codeparrot_training - Step 22705: {'lr': 0.0003034973739270401, 'samples': 11625472, 'steps': 22705, 'loss/train': 1.9089807271957397} 02/25/2022 06:59:18 - INFO - codeparrot_training - Step 22706: {'lr': 0.0003034813903714244, 'samples': 11625984, 'steps': 22706, 'loss/train': 1.247238278388977} 02/25/2022 06:59:21 - INFO - codeparrot_training - Step 22707: {'lr': 0.0003034654065867115, 'samples': 11626496, 'steps': 22707, 'loss/train': 2.4615378379821777} 02/25/2022 06:59:28 - INFO - codeparrot_training - Step 22708: {'lr': 0.0003034494225729697, 'samples': 11627008, 'steps': 22708, 'loss/train': 1.6899162530899048} 02/25/2022 06:59:33 - INFO - codeparrot_training - Step 22709: {'lr': 0.0003034334383302676, 'samples': 11627520, 'steps': 22709, 'loss/train': 1.9336423873901367} 02/25/2022 06:59:37 - INFO - codeparrot_training - Step 22710: {'lr': 0.0003034174538586735, 'samples': 11628032, 'steps': 22710, 'loss/train': 1.7221230268478394} 02/25/2022 06:59:42 - INFO - codeparrot_training - Step 22711: {'lr': 0.00030340146915825605, 'samples': 11628544, 'steps': 22711, 'loss/train': 1.6769905090332031} 02/25/2022 06:59:46 - INFO - codeparrot_training - Step 22712: {'lr': 0.00030338548422908373, 'samples': 11629056, 'steps': 22712, 'loss/train': 2.2398524284362793} 02/25/2022 06:59:52 - INFO - codeparrot_training - Step 22713: {'lr': 0.00030336949907122483, 'samples': 11629568, 'steps': 22713, 'loss/train': 1.357518196105957} 02/25/2022 06:59:55 - INFO - codeparrot_training - Step 22714: {'lr': 0.0003033535136847481, 'samples': 11630080, 'steps': 22714, 'loss/train': 0.7835584282875061} 02/25/2022 06:59:59 - INFO - codeparrot_training - Step 22715: {'lr': 0.0003033375280697218, 'samples': 11630592, 'steps': 22715, 'loss/train': 1.5173217058181763} 02/25/2022 07:00:04 - INFO - codeparrot_training - Step 22716: {'lr': 0.0003033215422262144, 'samples': 11631104, 'steps': 22716, 'loss/train': 1.8177915811538696} 02/25/2022 07:00:08 - INFO - codeparrot_training - Step 22717: {'lr': 0.0003033055561542945, 'samples': 11631616, 'steps': 22717, 'loss/train': 0.5128215551376343} 02/25/2022 07:00:14 - INFO - codeparrot_training - Step 22718: {'lr': 0.00030328956985403043, 'samples': 11632128, 'steps': 22718, 'loss/train': 2.6350560188293457} 02/25/2022 07:00:18 - INFO - codeparrot_training - Step 22719: {'lr': 0.0003032735833254909, 'samples': 11632640, 'steps': 22719, 'loss/train': 2.1741247177124023} 02/25/2022 07:00:23 - INFO - codeparrot_training - Step 22720: {'lr': 0.0003032575965687442, 'samples': 11633152, 'steps': 22720, 'loss/train': 1.713106393814087} 02/25/2022 07:00:27 - INFO - codeparrot_training - Step 22721: {'lr': 0.0003032416095838588, 'samples': 11633664, 'steps': 22721, 'loss/train': 8.00403118133545} 02/25/2022 07:00:32 - INFO - codeparrot_training - Step 22722: {'lr': 0.00030322562237090336, 'samples': 11634176, 'steps': 22722, 'loss/train': 1.9899433851242065} 02/25/2022 07:00:36 - INFO - codeparrot_training - Step 22723: {'lr': 0.00030320963492994616, 'samples': 11634688, 'steps': 22723, 'loss/train': 2.51070237159729} 02/25/2022 07:00:41 - INFO - codeparrot_training - Step 22724: {'lr': 0.00030319364726105584, 'samples': 11635200, 'steps': 22724, 'loss/train': 2.5846736431121826} 02/25/2022 07:00:45 - INFO - codeparrot_training - Step 22725: {'lr': 0.00030317765936430086, 'samples': 11635712, 'steps': 22725, 'loss/train': 1.0974221229553223} 02/25/2022 07:00:50 - INFO - codeparrot_training - Step 22726: {'lr': 0.0003031616712397496, 'samples': 11636224, 'steps': 22726, 'loss/train': 1.6972495317459106} 02/25/2022 07:00:54 - INFO - codeparrot_training - Step 22727: {'lr': 0.00030314568288747064, 'samples': 11636736, 'steps': 22727, 'loss/train': 1.8170210123062134} 02/25/2022 07:01:00 - INFO - codeparrot_training - Step 22728: {'lr': 0.00030312969430753244, 'samples': 11637248, 'steps': 22728, 'loss/train': 1.1301255226135254} 02/25/2022 07:01:03 - INFO - codeparrot_training - Step 22729: {'lr': 0.00030311370550000356, 'samples': 11637760, 'steps': 22729, 'loss/train': 1.0847762823104858} 02/25/2022 07:01:09 - INFO - codeparrot_training - Step 22730: {'lr': 0.0003030977164649523, 'samples': 11638272, 'steps': 22730, 'loss/train': 2.3306732177734375} 02/25/2022 07:01:12 - INFO - codeparrot_training - Step 22731: {'lr': 0.0003030817272024474, 'samples': 11638784, 'steps': 22731, 'loss/train': 1.3591725826263428} 02/25/2022 07:01:18 - INFO - codeparrot_training - Step 22732: {'lr': 0.0003030657377125572, 'samples': 11639296, 'steps': 22732, 'loss/train': 2.274630308151245} 02/25/2022 07:01:21 - INFO - codeparrot_training - Step 22733: {'lr': 0.0003030497479953503, 'samples': 11639808, 'steps': 22733, 'loss/train': 1.75478994846344} 02/25/2022 07:01:27 - INFO - codeparrot_training - Step 22734: {'lr': 0.00030303375805089503, 'samples': 11640320, 'steps': 22734, 'loss/train': 2.390540599822998} 02/25/2022 07:01:31 - INFO - codeparrot_training - Step 22735: {'lr': 0.00030301776787925995, 'samples': 11640832, 'steps': 22735, 'loss/train': 1.2266637086868286} 02/25/2022 07:01:34 - INFO - codeparrot_training - Step 22736: {'lr': 0.00030300177748051373, 'samples': 11641344, 'steps': 22736, 'loss/train': 0.8526584506034851} 02/25/2022 07:01:40 - INFO - codeparrot_training - Step 22737: {'lr': 0.0003029857868547246, 'samples': 11641856, 'steps': 22737, 'loss/train': 1.6764588356018066} 02/25/2022 07:01:43 - INFO - codeparrot_training - Step 22738: {'lr': 0.0003029697960019612, 'samples': 11642368, 'steps': 22738, 'loss/train': 1.7571521997451782} 02/25/2022 07:01:50 - INFO - codeparrot_training - Step 22739: {'lr': 0.0003029538049222921, 'samples': 11642880, 'steps': 22739, 'loss/train': 1.9561799764633179} 02/25/2022 07:01:53 - INFO - codeparrot_training - Step 22740: {'lr': 0.00030293781361578567, 'samples': 11643392, 'steps': 22740, 'loss/train': 0.8752596974372864} 02/25/2022 07:01:59 - INFO - codeparrot_training - Step 22741: {'lr': 0.0003029218220825104, 'samples': 11643904, 'steps': 22741, 'loss/train': 1.9061131477355957} 02/25/2022 07:02:02 - INFO - codeparrot_training - Step 22742: {'lr': 0.00030290583032253486, 'samples': 11644416, 'steps': 22742, 'loss/train': 1.623331904411316} 02/25/2022 07:02:08 - INFO - codeparrot_training - Step 22743: {'lr': 0.00030288983833592757, 'samples': 11644928, 'steps': 22743, 'loss/train': 1.826436996459961} 02/25/2022 07:02:11 - INFO - codeparrot_training - Step 22744: {'lr': 0.00030287384612275704, 'samples': 11645440, 'steps': 22744, 'loss/train': 1.921907663345337} 02/25/2022 07:02:17 - INFO - codeparrot_training - Step 22745: {'lr': 0.00030285785368309174, 'samples': 11645952, 'steps': 22745, 'loss/train': 2.0599167346954346} 02/25/2022 07:02:20 - INFO - codeparrot_training - Step 22746: {'lr': 0.0003028418610170001, 'samples': 11646464, 'steps': 22746, 'loss/train': 2.42594051361084} 02/25/2022 07:02:26 - INFO - codeparrot_training - Step 22747: {'lr': 0.00030282586812455076, 'samples': 11646976, 'steps': 22747, 'loss/train': 1.217502474784851} 02/25/2022 07:02:29 - INFO - codeparrot_training - Step 22748: {'lr': 0.00030280987500581213, 'samples': 11647488, 'steps': 22748, 'loss/train': 2.151460886001587} 02/25/2022 07:02:35 - INFO - codeparrot_training - Step 22749: {'lr': 0.00030279388166085287, 'samples': 11648000, 'steps': 22749, 'loss/train': 8.756054878234863} 02/25/2022 07:02:38 - INFO - codeparrot_training - Step 22750: {'lr': 0.0003027778880897413, 'samples': 11648512, 'steps': 22750, 'loss/train': 2.2929177284240723} 02/25/2022 07:02:44 - INFO - codeparrot_training - Step 22751: {'lr': 0.000302761894292546, 'samples': 11649024, 'steps': 22751, 'loss/train': 1.9948689937591553} 02/25/2022 07:02:47 - INFO - codeparrot_training - Step 22752: {'lr': 0.0003027459002693356, 'samples': 11649536, 'steps': 22752, 'loss/train': 1.965183973312378} 02/25/2022 07:02:53 - INFO - codeparrot_training - Step 22753: {'lr': 0.00030272990602017843, 'samples': 11650048, 'steps': 22753, 'loss/train': 4.539176940917969} 02/25/2022 07:02:56 - INFO - codeparrot_training - Step 22754: {'lr': 0.0003027139115451431, 'samples': 11650560, 'steps': 22754, 'loss/train': 2.828314781188965} 02/25/2022 07:03:04 - INFO - codeparrot_training - Step 22755: {'lr': 0.0003026979168442982, 'samples': 11651072, 'steps': 22755, 'loss/train': 0.9160630702972412} 02/25/2022 07:03:08 - INFO - codeparrot_training - Step 22756: {'lr': 0.000302681921917712, 'samples': 11651584, 'steps': 22756, 'loss/train': 2.709104299545288} 02/25/2022 07:03:14 - INFO - codeparrot_training - Step 22757: {'lr': 0.00030266592676545326, 'samples': 11652096, 'steps': 22757, 'loss/train': 2.413318157196045} 02/25/2022 07:03:17 - INFO - codeparrot_training - Step 22758: {'lr': 0.0003026499313875903, 'samples': 11652608, 'steps': 22758, 'loss/train': 1.2083826065063477} 02/25/2022 07:03:23 - INFO - codeparrot_training - Step 22759: {'lr': 0.00030263393578419196, 'samples': 11653120, 'steps': 22759, 'loss/train': 1.7300212383270264} 02/25/2022 07:03:26 - INFO - codeparrot_training - Step 22760: {'lr': 0.0003026179399553264, 'samples': 11653632, 'steps': 22760, 'loss/train': 1.4456440210342407} 02/25/2022 07:03:32 - INFO - codeparrot_training - Step 22761: {'lr': 0.0003026019439010624, 'samples': 11654144, 'steps': 22761, 'loss/train': 1.550153136253357} 02/25/2022 07:03:36 - INFO - codeparrot_training - Step 22762: {'lr': 0.0003025859476214683, 'samples': 11654656, 'steps': 22762, 'loss/train': 0.6863503456115723} 02/25/2022 07:03:41 - INFO - codeparrot_training - Step 22763: {'lr': 0.00030256995111661275, 'samples': 11655168, 'steps': 22763, 'loss/train': 1.0480544567108154} 02/25/2022 07:03:45 - INFO - codeparrot_training - Step 22764: {'lr': 0.0003025539543865642, 'samples': 11655680, 'steps': 22764, 'loss/train': 2.2686541080474854} 02/25/2022 07:03:50 - INFO - codeparrot_training - Step 22765: {'lr': 0.00030253795743139113, 'samples': 11656192, 'steps': 22765, 'loss/train': 2.4163551330566406} 02/25/2022 07:03:54 - INFO - codeparrot_training - Step 22766: {'lr': 0.0003025219602511622, 'samples': 11656704, 'steps': 22766, 'loss/train': 2.0099222660064697} 02/25/2022 07:04:01 - INFO - codeparrot_training - Step 22767: {'lr': 0.00030250596284594583, 'samples': 11657216, 'steps': 22767, 'loss/train': 1.368360996246338} 02/25/2022 07:04:05 - INFO - codeparrot_training - Step 22768: {'lr': 0.0003024899652158107, 'samples': 11657728, 'steps': 22768, 'loss/train': 2.145542621612549} 02/25/2022 07:04:10 - INFO - codeparrot_training - Step 22769: {'lr': 0.00030247396736082506, 'samples': 11658240, 'steps': 22769, 'loss/train': 1.8363025188446045} 02/25/2022 07:04:14 - INFO - codeparrot_training - Step 22770: {'lr': 0.0003024579692810577, 'samples': 11658752, 'steps': 22770, 'loss/train': 2.3849079608917236} 02/25/2022 07:04:19 - INFO - codeparrot_training - Step 22771: {'lr': 0.00030244197097657705, 'samples': 11659264, 'steps': 22771, 'loss/train': 1.414366602897644} 02/25/2022 07:04:23 - INFO - codeparrot_training - Step 22772: {'lr': 0.0003024259724474516, 'samples': 11659776, 'steps': 22772, 'loss/train': 3.4421193599700928} 02/25/2022 07:04:28 - INFO - codeparrot_training - Step 22773: {'lr': 0.00030240997369375, 'samples': 11660288, 'steps': 22773, 'loss/train': 1.5428099632263184} 02/25/2022 07:04:32 - INFO - codeparrot_training - Step 22774: {'lr': 0.0003023939747155406, 'samples': 11660800, 'steps': 22774, 'loss/train': 0.8700119853019714} 02/25/2022 07:04:37 - INFO - codeparrot_training - Step 22775: {'lr': 0.00030237797551289225, 'samples': 11661312, 'steps': 22775, 'loss/train': 2.581195831298828} 02/25/2022 07:04:41 - INFO - codeparrot_training - Step 22776: {'lr': 0.0003023619760858731, 'samples': 11661824, 'steps': 22776, 'loss/train': 2.4814250469207764} 02/25/2022 07:04:48 - INFO - codeparrot_training - Step 22777: {'lr': 0.000302345976434552, 'samples': 11662336, 'steps': 22777, 'loss/train': 1.3026471138000488} 02/25/2022 07:04:51 - INFO - codeparrot_training - Step 22778: {'lr': 0.0003023299765589973, 'samples': 11662848, 'steps': 22778, 'loss/train': 1.9169707298278809} 02/25/2022 07:04:57 - INFO - codeparrot_training - Step 22779: {'lr': 0.0003023139764592776, 'samples': 11663360, 'steps': 22779, 'loss/train': 3.4005424976348877} 02/25/2022 07:05:00 - INFO - codeparrot_training - Step 22780: {'lr': 0.0003022979761354614, 'samples': 11663872, 'steps': 22780, 'loss/train': 1.7546783685684204} 02/25/2022 07:05:06 - INFO - codeparrot_training - Step 22781: {'lr': 0.00030228197558761737, 'samples': 11664384, 'steps': 22781, 'loss/train': 2.767155885696411} 02/25/2022 07:05:09 - INFO - codeparrot_training - Step 22782: {'lr': 0.00030226597481581387, 'samples': 11664896, 'steps': 22782, 'loss/train': 1.376948356628418} 02/25/2022 07:05:15 - INFO - codeparrot_training - Step 22783: {'lr': 0.0003022499738201195, 'samples': 11665408, 'steps': 22783, 'loss/train': 8.588897705078125} 02/25/2022 07:05:18 - INFO - codeparrot_training - Step 22784: {'lr': 0.0003022339726006029, 'samples': 11665920, 'steps': 22784, 'loss/train': 2.3826470375061035} 02/25/2022 07:05:24 - INFO - codeparrot_training - Step 22785: {'lr': 0.0003022179711573326, 'samples': 11666432, 'steps': 22785, 'loss/train': 2.0467541217803955} 02/25/2022 07:05:28 - INFO - codeparrot_training - Step 22786: {'lr': 0.000302201969490377, 'samples': 11666944, 'steps': 22786, 'loss/train': 2.019573211669922} 02/25/2022 07:05:31 - INFO - codeparrot_training - Step 22787: {'lr': 0.00030218596759980476, 'samples': 11667456, 'steps': 22787, 'loss/train': 2.2680153846740723} 02/25/2022 07:05:39 - INFO - codeparrot_training - Step 22788: {'lr': 0.00030216996548568443, 'samples': 11667968, 'steps': 22788, 'loss/train': 1.7450085878372192} 02/25/2022 07:05:43 - INFO - codeparrot_training - Step 22789: {'lr': 0.0003021539631480845, 'samples': 11668480, 'steps': 22789, 'loss/train': 2.415402889251709} 02/25/2022 07:05:46 - INFO - codeparrot_training - Step 22790: {'lr': 0.0003021379605870736, 'samples': 11668992, 'steps': 22790, 'loss/train': 2.3866066932678223} 02/25/2022 07:05:51 - INFO - codeparrot_training - Step 22791: {'lr': 0.00030212195780272025, 'samples': 11669504, 'steps': 22791, 'loss/train': 2.0148651599884033} 02/25/2022 07:05:55 - INFO - codeparrot_training - Step 22792: {'lr': 0.000302105954795093, 'samples': 11670016, 'steps': 22792, 'loss/train': 1.82499361038208} 02/25/2022 07:06:00 - INFO - codeparrot_training - Step 22793: {'lr': 0.00030208995156426024, 'samples': 11670528, 'steps': 22793, 'loss/train': 2.2363693714141846} 02/25/2022 07:06:04 - INFO - codeparrot_training - Step 22794: {'lr': 0.0003020739481102908, 'samples': 11671040, 'steps': 22794, 'loss/train': 2.268592119216919} 02/25/2022 07:06:09 - INFO - codeparrot_training - Step 22795: {'lr': 0.00030205794443325296, 'samples': 11671552, 'steps': 22795, 'loss/train': 1.5524967908859253} 02/25/2022 07:06:13 - INFO - codeparrot_training - Step 22796: {'lr': 0.00030204194053321556, 'samples': 11672064, 'steps': 22796, 'loss/train': 1.9256868362426758} 02/25/2022 07:06:18 - INFO - codeparrot_training - Step 22797: {'lr': 0.00030202593641024696, 'samples': 11672576, 'steps': 22797, 'loss/train': 1.6537578105926514} 02/25/2022 07:06:22 - INFO - codeparrot_training - Step 22798: {'lr': 0.0003020099320644157, 'samples': 11673088, 'steps': 22798, 'loss/train': 0.9273232817649841} 02/25/2022 07:06:29 - INFO - codeparrot_training - Step 22799: {'lr': 0.00030199392749579053, 'samples': 11673600, 'steps': 22799, 'loss/train': 2.043025255203247} 02/25/2022 07:06:33 - INFO - codeparrot_training - Step 22800: {'lr': 0.0003019779227044398, 'samples': 11674112, 'steps': 22800, 'loss/train': 2.1142122745513916} 02/25/2022 07:06:38 - INFO - codeparrot_training - Step 22801: {'lr': 0.0003019619176904322, 'samples': 11674624, 'steps': 22801, 'loss/train': 1.5707728862762451} 02/25/2022 07:06:42 - INFO - codeparrot_training - Step 22802: {'lr': 0.00030194591245383625, 'samples': 11675136, 'steps': 22802, 'loss/train': 1.275285005569458} 02/25/2022 07:06:47 - INFO - codeparrot_training - Step 22803: {'lr': 0.00030192990699472053, 'samples': 11675648, 'steps': 22803, 'loss/train': 1.4428529739379883} 02/25/2022 07:06:51 - INFO - codeparrot_training - Step 22804: {'lr': 0.00030191390131315357, 'samples': 11676160, 'steps': 22804, 'loss/train': 1.102246880531311} 02/25/2022 07:06:56 - INFO - codeparrot_training - Step 22805: {'lr': 0.0003018978954092039, 'samples': 11676672, 'steps': 22805, 'loss/train': 1.888580560684204} 02/25/2022 07:07:02 - INFO - codeparrot_training - Step 22806: {'lr': 0.0003018818892829401, 'samples': 11677184, 'steps': 22806, 'loss/train': 1.2770750522613525} 02/25/2022 07:07:05 - INFO - codeparrot_training - Step 22807: {'lr': 0.00030186588293443077, 'samples': 11677696, 'steps': 22807, 'loss/train': 1.3058511018753052} 02/25/2022 07:07:11 - INFO - codeparrot_training - Step 22808: {'lr': 0.0003018498763637445, 'samples': 11678208, 'steps': 22808, 'loss/train': 1.4961966276168823} 02/25/2022 07:07:14 - INFO - codeparrot_training - Step 22809: {'lr': 0.0003018338695709498, 'samples': 11678720, 'steps': 22809, 'loss/train': 1.1272844076156616} 02/25/2022 07:07:18 - INFO - codeparrot_training - Step 22810: {'lr': 0.0003018178625561153, 'samples': 11679232, 'steps': 22810, 'loss/train': 0.9606180787086487} 02/25/2022 07:07:24 - INFO - codeparrot_training - Step 22811: {'lr': 0.0003018018553193095, 'samples': 11679744, 'steps': 22811, 'loss/train': 1.904482126235962} 02/25/2022 07:07:27 - INFO - codeparrot_training - Step 22812: {'lr': 0.00030178584786060106, 'samples': 11680256, 'steps': 22812, 'loss/train': 0.4109419584274292} 02/25/2022 07:07:33 - INFO - codeparrot_training - Step 22813: {'lr': 0.00030176984018005836, 'samples': 11680768, 'steps': 22813, 'loss/train': 0.9216345548629761} 02/25/2022 07:07:36 - INFO - codeparrot_training - Step 22814: {'lr': 0.0003017538322777502, 'samples': 11681280, 'steps': 22814, 'loss/train': 1.3572323322296143} 02/25/2022 07:07:43 - INFO - codeparrot_training - Step 22815: {'lr': 0.00030173782415374503, 'samples': 11681792, 'steps': 22815, 'loss/train': 1.9806339740753174} 02/25/2022 07:07:47 - INFO - codeparrot_training - Step 22816: {'lr': 0.00030172181580811146, 'samples': 11682304, 'steps': 22816, 'loss/train': 3.126826763153076} 02/25/2022 07:07:52 - INFO - codeparrot_training - Step 22817: {'lr': 0.0003017058072409181, 'samples': 11682816, 'steps': 22817, 'loss/train': 1.5081586837768555} 02/25/2022 07:07:56 - INFO - codeparrot_training - Step 22818: {'lr': 0.0003016897984522334, 'samples': 11683328, 'steps': 22818, 'loss/train': 1.6155692338943481} 02/25/2022 07:08:01 - INFO - codeparrot_training - Step 22819: {'lr': 0.00030167378944212606, 'samples': 11683840, 'steps': 22819, 'loss/train': 3.465468406677246} 02/25/2022 07:08:05 - INFO - codeparrot_training - Step 22820: {'lr': 0.0003016577802106645, 'samples': 11684352, 'steps': 22820, 'loss/train': 2.298074722290039} 02/25/2022 07:08:10 - INFO - codeparrot_training - Step 22821: {'lr': 0.0003016417707579176, 'samples': 11684864, 'steps': 22821, 'loss/train': 1.7258412837982178} 02/25/2022 07:08:14 - INFO - codeparrot_training - Step 22822: {'lr': 0.00030162576108395364, 'samples': 11685376, 'steps': 22822, 'loss/train': 2.8233084678649902} 02/25/2022 07:08:21 - INFO - codeparrot_training - Step 22823: {'lr': 0.0003016097511888414, 'samples': 11685888, 'steps': 22823, 'loss/train': 2.577441930770874} 02/25/2022 07:08:25 - INFO - codeparrot_training - Step 22824: {'lr': 0.0003015937410726493, 'samples': 11686400, 'steps': 22824, 'loss/train': 2.5049593448638916} 02/25/2022 07:08:30 - INFO - codeparrot_training - Step 22825: {'lr': 0.000301577730735446, 'samples': 11686912, 'steps': 22825, 'loss/train': 2.476512908935547} 02/25/2022 07:08:34 - INFO - codeparrot_training - Step 22826: {'lr': 0.00030156172017730006, 'samples': 11687424, 'steps': 22826, 'loss/train': 0.84358811378479} 02/25/2022 07:08:39 - INFO - codeparrot_training - Step 22827: {'lr': 0.00030154570939828015, 'samples': 11687936, 'steps': 22827, 'loss/train': 2.8828892707824707} 02/25/2022 07:08:43 - INFO - codeparrot_training - Step 22828: {'lr': 0.0003015296983984547, 'samples': 11688448, 'steps': 22828, 'loss/train': 5.055017471313477} 02/25/2022 07:08:48 - INFO - codeparrot_training - Step 22829: {'lr': 0.00030151368717789244, 'samples': 11688960, 'steps': 22829, 'loss/train': 1.475337028503418} 02/25/2022 07:08:52 - INFO - codeparrot_training - Step 22830: {'lr': 0.0003014976757366619, 'samples': 11689472, 'steps': 22830, 'loss/train': 2.0757195949554443} 02/25/2022 07:08:57 - INFO - codeparrot_training - Step 22831: {'lr': 0.0003014816640748316, 'samples': 11689984, 'steps': 22831, 'loss/train': 2.0199716091156006} 02/25/2022 07:09:01 - INFO - codeparrot_training - Step 22832: {'lr': 0.00030146565219247033, 'samples': 11690496, 'steps': 22832, 'loss/train': 1.2075138092041016} 02/25/2022 07:09:06 - INFO - codeparrot_training - Step 22833: {'lr': 0.0003014496400896465, 'samples': 11691008, 'steps': 22833, 'loss/train': 3.0877788066864014} 02/25/2022 07:09:10 - INFO - codeparrot_training - Step 22834: {'lr': 0.0003014336277664287, 'samples': 11691520, 'steps': 22834, 'loss/train': 2.5128939151763916} 02/25/2022 07:09:17 - INFO - codeparrot_training - Step 22835: {'lr': 0.0003014176152228856, 'samples': 11692032, 'steps': 22835, 'loss/train': 1.7883402109146118} 02/25/2022 07:09:21 - INFO - codeparrot_training - Step 22836: {'lr': 0.00030140160245908584, 'samples': 11692544, 'steps': 22836, 'loss/train': 1.8974905014038086} 02/25/2022 07:09:26 - INFO - codeparrot_training - Step 22837: {'lr': 0.0003013855894750978, 'samples': 11693056, 'steps': 22837, 'loss/train': 2.3971943855285645} 02/25/2022 07:09:30 - INFO - codeparrot_training - Step 22838: {'lr': 0.0003013695762709903, 'samples': 11693568, 'steps': 22838, 'loss/train': 2.310260057449341} 02/25/2022 07:09:35 - INFO - codeparrot_training - Step 22839: {'lr': 0.0003013535628468318, 'samples': 11694080, 'steps': 22839, 'loss/train': 3.591212034225464} 02/25/2022 07:09:39 - INFO - codeparrot_training - Step 22840: {'lr': 0.000301337549202691, 'samples': 11694592, 'steps': 22840, 'loss/train': 2.19614315032959} 02/25/2022 07:09:44 - INFO - codeparrot_training - Step 22841: {'lr': 0.0003013215353386364, 'samples': 11695104, 'steps': 22841, 'loss/train': 1.233171820640564} 02/25/2022 07:09:48 - INFO - codeparrot_training - Step 22842: {'lr': 0.00030130552125473667, 'samples': 11695616, 'steps': 22842, 'loss/train': 0.5252600908279419} 02/25/2022 07:09:53 - INFO - codeparrot_training - Step 22843: {'lr': 0.00030128950695106034, 'samples': 11696128, 'steps': 22843, 'loss/train': 2.377997875213623} 02/25/2022 07:09:57 - INFO - codeparrot_training - Step 22844: {'lr': 0.00030127349242767607, 'samples': 11696640, 'steps': 22844, 'loss/train': 1.7373830080032349} 02/25/2022 07:10:04 - INFO - codeparrot_training - Step 22845: {'lr': 0.0003012574776846524, 'samples': 11697152, 'steps': 22845, 'loss/train': 0.16632595658302307} 02/25/2022 07:10:07 - INFO - codeparrot_training - Step 22846: {'lr': 0.00030124146272205804, 'samples': 11697664, 'steps': 22846, 'loss/train': 1.4264466762542725} 02/25/2022 07:10:13 - INFO - codeparrot_training - Step 22847: {'lr': 0.00030122544753996143, 'samples': 11698176, 'steps': 22847, 'loss/train': 2.196805238723755} 02/25/2022 07:10:16 - INFO - codeparrot_training - Step 22848: {'lr': 0.00030120943213843136, 'samples': 11698688, 'steps': 22848, 'loss/train': 1.886110782623291} 02/25/2022 07:10:22 - INFO - codeparrot_training - Step 22849: {'lr': 0.0003011934165175363, 'samples': 11699200, 'steps': 22849, 'loss/train': 2.2160489559173584} 02/25/2022 07:10:26 - INFO - codeparrot_training - Step 22850: {'lr': 0.00030117740067734495, 'samples': 11699712, 'steps': 22850, 'loss/train': 1.6635175943374634} 02/25/2022 07:10:31 - INFO - codeparrot_training - Step 22851: {'lr': 0.0003011613846179258, 'samples': 11700224, 'steps': 22851, 'loss/train': 1.7605781555175781} 02/25/2022 07:10:35 - INFO - codeparrot_training - Step 22852: {'lr': 0.0003011453683393476, 'samples': 11700736, 'steps': 22852, 'loss/train': 0.5744600296020508} 02/25/2022 07:10:40 - INFO - codeparrot_training - Step 22853: {'lr': 0.0003011293518416788, 'samples': 11701248, 'steps': 22853, 'loss/train': 0.6563553810119629} 02/25/2022 07:10:44 - INFO - codeparrot_training - Step 22854: {'lr': 0.00030111333512498813, 'samples': 11701760, 'steps': 22854, 'loss/train': 2.2709381580352783} 02/25/2022 07:10:51 - INFO - codeparrot_training - Step 22855: {'lr': 0.00030109731818934413, 'samples': 11702272, 'steps': 22855, 'loss/train': 1.6041088104248047} 02/25/2022 07:10:54 - INFO - codeparrot_training - Step 22856: {'lr': 0.00030108130103481554, 'samples': 11702784, 'steps': 22856, 'loss/train': 1.9734938144683838} 02/25/2022 07:11:00 - INFO - codeparrot_training - Step 22857: {'lr': 0.0003010652836614707, 'samples': 11703296, 'steps': 22857, 'loss/train': 2.181708812713623} 02/25/2022 07:11:03 - INFO - codeparrot_training - Step 22858: {'lr': 0.00030104926606937856, 'samples': 11703808, 'steps': 22858, 'loss/train': 1.8034135103225708} 02/25/2022 07:11:09 - INFO - codeparrot_training - Step 22859: {'lr': 0.0003010332482586075, 'samples': 11704320, 'steps': 22859, 'loss/train': 1.9789999723434448} 02/25/2022 07:11:13 - INFO - codeparrot_training - Step 22860: {'lr': 0.0003010172302292263, 'samples': 11704832, 'steps': 22860, 'loss/train': 2.2439792156219482} 02/25/2022 07:11:18 - INFO - codeparrot_training - Step 22861: {'lr': 0.00030100121198130335, 'samples': 11705344, 'steps': 22861, 'loss/train': 0.6301648616790771} 02/25/2022 07:11:22 - INFO - codeparrot_training - Step 22862: {'lr': 0.0003009851935149075, 'samples': 11705856, 'steps': 22862, 'loss/train': 1.5742487907409668} 02/25/2022 07:11:27 - INFO - codeparrot_training - Step 22863: {'lr': 0.0003009691748301072, 'samples': 11706368, 'steps': 22863, 'loss/train': 0.7274398803710938} 02/25/2022 07:11:31 - INFO - codeparrot_training - Step 22864: {'lr': 0.0003009531559269713, 'samples': 11706880, 'steps': 22864, 'loss/train': 3.308727502822876} 02/25/2022 07:11:36 - INFO - codeparrot_training - Step 22865: {'lr': 0.00030093713680556805, 'samples': 11707392, 'steps': 22865, 'loss/train': 0.9269284605979919} 02/25/2022 07:11:40 - INFO - codeparrot_training - Step 22866: {'lr': 0.0003009211174659664, 'samples': 11707904, 'steps': 22866, 'loss/train': 1.522827386856079} 02/25/2022 07:11:45 - INFO - codeparrot_training - Step 22867: {'lr': 0.00030090509790823476, 'samples': 11708416, 'steps': 22867, 'loss/train': 1.4396413564682007} 02/25/2022 07:11:49 - INFO - codeparrot_training - Step 22868: {'lr': 0.0003008890781324419, 'samples': 11708928, 'steps': 22868, 'loss/train': 2.7254207134246826} 02/25/2022 07:11:54 - INFO - codeparrot_training - Step 22869: {'lr': 0.0003008730581386564, 'samples': 11709440, 'steps': 22869, 'loss/train': 1.9003928899765015} 02/25/2022 07:11:58 - INFO - codeparrot_training - Step 22870: {'lr': 0.00030085703792694687, 'samples': 11709952, 'steps': 22870, 'loss/train': 1.429430365562439} 02/25/2022 07:12:05 - INFO - codeparrot_training - Step 22871: {'lr': 0.00030084101749738195, 'samples': 11710464, 'steps': 22871, 'loss/train': 1.5849227905273438} 02/25/2022 07:12:09 - INFO - codeparrot_training - Step 22872: {'lr': 0.00030082499685003025, 'samples': 11710976, 'steps': 22872, 'loss/train': 1.8664072751998901} 02/25/2022 07:12:14 - INFO - codeparrot_training - Step 22873: {'lr': 0.0003008089759849604, 'samples': 11711488, 'steps': 22873, 'loss/train': 2.0416150093078613} 02/25/2022 07:12:18 - INFO - codeparrot_training - Step 22874: {'lr': 0.000300792954902241, 'samples': 11712000, 'steps': 22874, 'loss/train': 1.8513036966323853} 02/25/2022 07:12:23 - INFO - codeparrot_training - Step 22875: {'lr': 0.00030077693360194076, 'samples': 11712512, 'steps': 22875, 'loss/train': 3.3205151557922363} 02/25/2022 07:12:27 - INFO - codeparrot_training - Step 22876: {'lr': 0.0003007609120841282, 'samples': 11713024, 'steps': 22876, 'loss/train': 2.2465617656707764} 02/25/2022 07:12:32 - INFO - codeparrot_training - Step 22877: {'lr': 0.0003007448903488721, 'samples': 11713536, 'steps': 22877, 'loss/train': 1.8106093406677246} 02/25/2022 07:12:36 - INFO - codeparrot_training - Step 22878: {'lr': 0.00030072886839624093, 'samples': 11714048, 'steps': 22878, 'loss/train': 1.3157438039779663} 02/25/2022 07:12:41 - INFO - codeparrot_training - Step 22879: {'lr': 0.0003007128462263034, 'samples': 11714560, 'steps': 22879, 'loss/train': 0.8539958000183105} 02/25/2022 07:12:45 - INFO - codeparrot_training - Step 22880: {'lr': 0.0003006968238391281, 'samples': 11715072, 'steps': 22880, 'loss/train': 3.133535623550415} 02/25/2022 07:12:52 - INFO - codeparrot_training - Step 22881: {'lr': 0.00030068080123478376, 'samples': 11715584, 'steps': 22881, 'loss/train': 0.6200644969940186} 02/25/2022 07:12:55 - INFO - codeparrot_training - Step 22882: {'lr': 0.000300664778413339, 'samples': 11716096, 'steps': 22882, 'loss/train': 2.0040032863616943} 02/25/2022 07:13:01 - INFO - codeparrot_training - Step 22883: {'lr': 0.00030064875537486236, 'samples': 11716608, 'steps': 22883, 'loss/train': 0.9974362254142761} 02/25/2022 07:13:04 - INFO - codeparrot_training - Step 22884: {'lr': 0.00030063273211942254, 'samples': 11717120, 'steps': 22884, 'loss/train': 1.7254163026809692} 02/25/2022 07:13:10 - INFO - codeparrot_training - Step 22885: {'lr': 0.0003006167086470882, 'samples': 11717632, 'steps': 22885, 'loss/train': 2.538034677505493} 02/25/2022 07:13:13 - INFO - codeparrot_training - Step 22886: {'lr': 0.00030060068495792793, 'samples': 11718144, 'steps': 22886, 'loss/train': 2.075181245803833} 02/25/2022 07:13:19 - INFO - codeparrot_training - Step 22887: {'lr': 0.0003005846610520104, 'samples': 11718656, 'steps': 22887, 'loss/train': 1.473665475845337} 02/25/2022 07:13:22 - INFO - codeparrot_training - Step 22888: {'lr': 0.00030056863692940426, 'samples': 11719168, 'steps': 22888, 'loss/train': 1.9633461236953735} 02/25/2022 07:13:28 - INFO - codeparrot_training - Step 22889: {'lr': 0.00030055261259017807, 'samples': 11719680, 'steps': 22889, 'loss/train': 0.8838751316070557} 02/25/2022 07:13:31 - INFO - codeparrot_training - Step 22890: {'lr': 0.00030053658803440064, 'samples': 11720192, 'steps': 22890, 'loss/train': 2.0789523124694824} 02/25/2022 07:13:39 - INFO - codeparrot_training - Step 22891: {'lr': 0.00030052056326214046, 'samples': 11720704, 'steps': 22891, 'loss/train': 1.9491651058197021} 02/25/2022 07:13:42 - INFO - codeparrot_training - Step 22892: {'lr': 0.00030050453827346627, 'samples': 11721216, 'steps': 22892, 'loss/train': 1.710429310798645} 02/25/2022 07:13:48 - INFO - codeparrot_training - Step 22893: {'lr': 0.0003004885130684467, 'samples': 11721728, 'steps': 22893, 'loss/train': 2.135662794113159} 02/25/2022 07:13:51 - INFO - codeparrot_training - Step 22894: {'lr': 0.00030047248764715023, 'samples': 11722240, 'steps': 22894, 'loss/train': 0.6546010971069336} 02/25/2022 07:13:57 - INFO - codeparrot_training - Step 22895: {'lr': 0.0003004564620096457, 'samples': 11722752, 'steps': 22895, 'loss/train': 2.801112174987793} 02/25/2022 07:14:00 - INFO - codeparrot_training - Step 22896: {'lr': 0.00030044043615600174, 'samples': 11723264, 'steps': 22896, 'loss/train': 1.4357532262802124} 02/25/2022 07:14:06 - INFO - codeparrot_training - Step 22897: {'lr': 0.000300424410086287, 'samples': 11723776, 'steps': 22897, 'loss/train': 2.0789871215820312} 02/25/2022 07:14:09 - INFO - codeparrot_training - Step 22898: {'lr': 0.00030040838380057005, 'samples': 11724288, 'steps': 22898, 'loss/train': 1.5147205591201782} 02/25/2022 07:14:15 - INFO - codeparrot_training - Step 22899: {'lr': 0.00030039235729891964, 'samples': 11724800, 'steps': 22899, 'loss/train': 2.2267744541168213} 02/25/2022 07:14:18 - INFO - codeparrot_training - Step 22900: {'lr': 0.0003003763305814043, 'samples': 11725312, 'steps': 22900, 'loss/train': 1.8106625080108643} 02/25/2022 07:14:24 - INFO - codeparrot_training - Step 22901: {'lr': 0.00030036030364809284, 'samples': 11725824, 'steps': 22901, 'loss/train': 1.4188302755355835} 02/25/2022 07:14:27 - INFO - codeparrot_training - Step 22902: {'lr': 0.00030034427649905377, 'samples': 11726336, 'steps': 22902, 'loss/train': 1.9424360990524292} 02/25/2022 07:14:33 - INFO - codeparrot_training - Step 22903: {'lr': 0.0003003282491343559, 'samples': 11726848, 'steps': 22903, 'loss/train': 2.2287771701812744} 02/25/2022 07:14:36 - INFO - codeparrot_training - Step 22904: {'lr': 0.00030031222155406763, 'samples': 11727360, 'steps': 22904, 'loss/train': 1.9816529750823975} 02/25/2022 07:14:42 - INFO - codeparrot_training - Step 22905: {'lr': 0.00030029619375825784, 'samples': 11727872, 'steps': 22905, 'loss/train': 1.6138992309570312} 02/25/2022 07:14:45 - INFO - codeparrot_training - Step 22906: {'lr': 0.00030028016574699517, 'samples': 11728384, 'steps': 22906, 'loss/train': 2.4598488807678223} 02/25/2022 07:14:53 - INFO - codeparrot_training - Step 22907: {'lr': 0.0003002641375203482, 'samples': 11728896, 'steps': 22907, 'loss/train': 2.3908636569976807} 02/25/2022 07:14:56 - INFO - codeparrot_training - Step 22908: {'lr': 0.0003002481090783856, 'samples': 11729408, 'steps': 22908, 'loss/train': 1.5112683773040771} 02/25/2022 07:15:02 - INFO - codeparrot_training - Step 22909: {'lr': 0.0003002320804211761, 'samples': 11729920, 'steps': 22909, 'loss/train': 1.861320972442627} 02/25/2022 07:15:05 - INFO - codeparrot_training - Step 22910: {'lr': 0.00030021605154878836, 'samples': 11730432, 'steps': 22910, 'loss/train': 1.053736686706543} 02/25/2022 07:15:10 - INFO - codeparrot_training - Step 22911: {'lr': 0.0003002000224612909, 'samples': 11730944, 'steps': 22911, 'loss/train': 1.0247892141342163} 02/25/2022 07:15:14 - INFO - codeparrot_training - Step 22912: {'lr': 0.0003001839931587526, 'samples': 11731456, 'steps': 22912, 'loss/train': 1.086834192276001} 02/25/2022 07:15:20 - INFO - codeparrot_training - Step 22913: {'lr': 0.0003001679636412419, 'samples': 11731968, 'steps': 22913, 'loss/train': 0.8913392424583435} 02/25/2022 07:15:23 - INFO - codeparrot_training - Step 22914: {'lr': 0.0003001519339088277, 'samples': 11732480, 'steps': 22914, 'loss/train': 1.4678913354873657} 02/25/2022 07:15:29 - INFO - codeparrot_training - Step 22915: {'lr': 0.00030013590396157843, 'samples': 11732992, 'steps': 22915, 'loss/train': 2.1191389560699463} 02/25/2022 07:15:33 - INFO - codeparrot_training - Step 22916: {'lr': 0.0003001198737995628, 'samples': 11733504, 'steps': 22916, 'loss/train': 2.638381004333496} 02/25/2022 07:15:36 - INFO - codeparrot_training - Step 22917: {'lr': 0.0003001038434228497, 'samples': 11734016, 'steps': 22917, 'loss/train': 2.4855711460113525} 02/25/2022 07:15:43 - INFO - codeparrot_training - Step 22918: {'lr': 0.00030008781283150755, 'samples': 11734528, 'steps': 22918, 'loss/train': 2.275350570678711} 02/25/2022 07:15:47 - INFO - codeparrot_training - Step 22919: {'lr': 0.0003000717820256052, 'samples': 11735040, 'steps': 22919, 'loss/train': 1.6692391633987427} 02/25/2022 07:15:52 - INFO - codeparrot_training - Step 22920: {'lr': 0.00030005575100521117, 'samples': 11735552, 'steps': 22920, 'loss/train': 2.2804603576660156} 02/25/2022 07:15:56 - INFO - codeparrot_training - Step 22921: {'lr': 0.0003000397197703942, 'samples': 11736064, 'steps': 22921, 'loss/train': 2.3221590518951416} 02/25/2022 07:16:02 - INFO - codeparrot_training - Step 22922: {'lr': 0.00030002368832122295, 'samples': 11736576, 'steps': 22922, 'loss/train': 2.876267433166504} 02/25/2022 07:16:05 - INFO - codeparrot_training - Step 22923: {'lr': 0.00030000765665776617, 'samples': 11737088, 'steps': 22923, 'loss/train': 8.902372360229492} 02/25/2022 07:16:11 - INFO - codeparrot_training - Step 22924: {'lr': 0.0002999916247800924, 'samples': 11737600, 'steps': 22924, 'loss/train': 2.0848114490509033} 02/25/2022 07:16:16 - INFO - codeparrot_training - Step 22925: {'lr': 0.00029997559268827044, 'samples': 11738112, 'steps': 22925, 'loss/train': 2.377110719680786} 02/25/2022 07:16:20 - INFO - codeparrot_training - Step 22926: {'lr': 0.0002999595603823689, 'samples': 11738624, 'steps': 22926, 'loss/train': 0.3089900016784668} 02/25/2022 07:16:26 - INFO - codeparrot_training - Step 22927: {'lr': 0.00029994352786245643, 'samples': 11739136, 'steps': 22927, 'loss/train': 1.363265872001648} 02/25/2022 07:16:29 - INFO - codeparrot_training - Step 22928: {'lr': 0.0002999274951286017, 'samples': 11739648, 'steps': 22928, 'loss/train': 1.5553009510040283} 02/25/2022 07:16:35 - INFO - codeparrot_training - Step 22929: {'lr': 0.0002999114621808735, 'samples': 11740160, 'steps': 22929, 'loss/train': 1.5626431703567505} 02/25/2022 07:16:38 - INFO - codeparrot_training - Step 22930: {'lr': 0.0002998954290193405, 'samples': 11740672, 'steps': 22930, 'loss/train': 1.7593611478805542} 02/25/2022 07:16:43 - INFO - codeparrot_training - Step 22931: {'lr': 0.00029987939564407124, 'samples': 11741184, 'steps': 22931, 'loss/train': 1.8623770475387573} 02/25/2022 07:16:47 - INFO - codeparrot_training - Step 22932: {'lr': 0.00029986336205513456, 'samples': 11741696, 'steps': 22932, 'loss/train': 1.893572211265564} 02/25/2022 07:16:53 - INFO - codeparrot_training - Step 22933: {'lr': 0.00029984732825259904, 'samples': 11742208, 'steps': 22933, 'loss/train': 0.6987547874450684} 02/25/2022 07:16:56 - INFO - codeparrot_training - Step 22934: {'lr': 0.00029983129423653333, 'samples': 11742720, 'steps': 22934, 'loss/train': 1.8675495386123657} 02/25/2022 07:17:02 - INFO - codeparrot_training - Step 22935: {'lr': 0.00029981526000700626, 'samples': 11743232, 'steps': 22935, 'loss/train': 2.0508322715759277} 02/25/2022 07:17:05 - INFO - codeparrot_training - Step 22936: {'lr': 0.0002997992255640864, 'samples': 11743744, 'steps': 22936, 'loss/train': 3.23275089263916} 02/25/2022 07:17:10 - INFO - codeparrot_training - Step 22937: {'lr': 0.0002997831909078425, 'samples': 11744256, 'steps': 22937, 'loss/train': 1.193077564239502} 02/25/2022 07:17:14 - INFO - codeparrot_training - Step 22938: {'lr': 0.00029976715603834315, 'samples': 11744768, 'steps': 22938, 'loss/train': 2.0761213302612305} 02/25/2022 07:17:20 - INFO - codeparrot_training - Step 22939: {'lr': 0.00029975112095565723, 'samples': 11745280, 'steps': 22939, 'loss/train': 0.33442795276641846} 02/25/2022 07:17:24 - INFO - codeparrot_training - Step 22940: {'lr': 0.00029973508565985316, 'samples': 11745792, 'steps': 22940, 'loss/train': 1.3022972345352173} 02/25/2022 07:17:29 - INFO - codeparrot_training - Step 22941: {'lr': 0.0002997190501509999, 'samples': 11746304, 'steps': 22941, 'loss/train': 2.723212480545044} 02/25/2022 07:17:33 - INFO - codeparrot_training - Step 22942: {'lr': 0.00029970301442916594, 'samples': 11746816, 'steps': 22942, 'loss/train': 0.7927182912826538} 02/25/2022 07:17:38 - INFO - codeparrot_training - Step 22943: {'lr': 0.00029968697849442006, 'samples': 11747328, 'steps': 22943, 'loss/train': 1.8818817138671875} 02/25/2022 07:17:42 - INFO - codeparrot_training - Step 22944: {'lr': 0.0002996709423468309, 'samples': 11747840, 'steps': 22944, 'loss/train': 2.369805335998535} 02/25/2022 07:17:47 - INFO - codeparrot_training - Step 22945: {'lr': 0.00029965490598646727, 'samples': 11748352, 'steps': 22945, 'loss/train': 1.4444445371627808} 02/25/2022 07:17:51 - INFO - codeparrot_training - Step 22946: {'lr': 0.00029963886941339774, 'samples': 11748864, 'steps': 22946, 'loss/train': 2.2752997875213623} 02/25/2022 07:17:56 - INFO - codeparrot_training - Step 22947: {'lr': 0.0002996228326276911, 'samples': 11749376, 'steps': 22947, 'loss/train': 1.4901851415634155} 02/25/2022 07:18:00 - INFO - codeparrot_training - Step 22948: {'lr': 0.0002996067956294159, 'samples': 11749888, 'steps': 22948, 'loss/train': 1.6299219131469727} 02/25/2022 07:18:05 - INFO - codeparrot_training - Step 22949: {'lr': 0.0002995907584186411, 'samples': 11750400, 'steps': 22949, 'loss/train': 0.42957594990730286} 02/25/2022 07:18:09 - INFO - codeparrot_training - Step 22950: {'lr': 0.00029957472099543516, 'samples': 11750912, 'steps': 22950, 'loss/train': 3.1684324741363525} 02/25/2022 07:18:14 - INFO - codeparrot_training - Step 22951: {'lr': 0.00029955868335986686, 'samples': 11751424, 'steps': 22951, 'loss/train': 1.6765109300613403} 02/25/2022 07:18:18 - INFO - codeparrot_training - Step 22952: {'lr': 0.0002995426455120049, 'samples': 11751936, 'steps': 22952, 'loss/train': 2.0211105346679688} 02/25/2022 07:18:25 - INFO - codeparrot_training - Step 22953: {'lr': 0.0002995266074519179, 'samples': 11752448, 'steps': 22953, 'loss/train': 1.6748194694519043} 02/25/2022 07:18:28 - INFO - codeparrot_training - Step 22954: {'lr': 0.00029951056917967476, 'samples': 11752960, 'steps': 22954, 'loss/train': 3.1632049083709717} 02/25/2022 07:18:34 - INFO - codeparrot_training - Step 22955: {'lr': 0.000299494530695344, 'samples': 11753472, 'steps': 22955, 'loss/train': 1.021226406097412} 02/25/2022 07:18:37 - INFO - codeparrot_training - Step 22956: {'lr': 0.0002994784919989944, 'samples': 11753984, 'steps': 22956, 'loss/train': 1.8655604124069214} 02/25/2022 07:18:41 - INFO - codeparrot_training - Step 22957: {'lr': 0.00029946245309069464, 'samples': 11754496, 'steps': 22957, 'loss/train': 1.7605639696121216} 02/25/2022 07:18:46 - INFO - codeparrot_training - Step 22958: {'lr': 0.0002994464139705135, 'samples': 11755008, 'steps': 22958, 'loss/train': 1.7276771068572998} 02/25/2022 07:18:50 - INFO - codeparrot_training - Step 22959: {'lr': 0.00029943037463851953, 'samples': 11755520, 'steps': 22959, 'loss/train': 1.2175084352493286} 02/25/2022 07:18:55 - INFO - codeparrot_training - Step 22960: {'lr': 0.00029941433509478153, 'samples': 11756032, 'steps': 22960, 'loss/train': 1.2298845052719116} 02/25/2022 07:18:59 - INFO - codeparrot_training - Step 22961: {'lr': 0.00029939829533936823, 'samples': 11756544, 'steps': 22961, 'loss/train': 1.3954131603240967} 02/25/2022 07:19:04 - INFO - codeparrot_training - Step 22962: {'lr': 0.00029938225537234843, 'samples': 11757056, 'steps': 22962, 'loss/train': 0.8541293740272522} 02/25/2022 07:19:08 - INFO - codeparrot_training - Step 22963: {'lr': 0.0002993662151937906, 'samples': 11757568, 'steps': 22963, 'loss/train': 1.5400323867797852} 02/25/2022 07:19:14 - INFO - codeparrot_training - Step 22964: {'lr': 0.00029935017480376357, 'samples': 11758080, 'steps': 22964, 'loss/train': 1.5345503091812134} 02/25/2022 07:19:18 - INFO - codeparrot_training - Step 22965: {'lr': 0.00029933413420233615, 'samples': 11758592, 'steps': 22965, 'loss/train': 2.783712863922119} 02/25/2022 07:19:23 - INFO - codeparrot_training - Step 22966: {'lr': 0.0002993180933895769, 'samples': 11759104, 'steps': 22966, 'loss/train': 1.6588160991668701} 02/25/2022 07:19:29 - INFO - codeparrot_training - Step 22967: {'lr': 0.00029930205236555464, 'samples': 11759616, 'steps': 22967, 'loss/train': 2.9335856437683105} 02/25/2022 07:19:33 - INFO - codeparrot_training - Step 22968: {'lr': 0.000299286011130338, 'samples': 11760128, 'steps': 22968, 'loss/train': 2.218116283416748} 02/25/2022 07:19:36 - INFO - codeparrot_training - Step 22969: {'lr': 0.00029926996968399576, 'samples': 11760640, 'steps': 22969, 'loss/train': 2.2955267429351807} 02/25/2022 07:19:42 - INFO - codeparrot_training - Step 22970: {'lr': 0.0002992539280265966, 'samples': 11761152, 'steps': 22970, 'loss/train': 2.4329981803894043} 02/25/2022 07:19:46 - INFO - codeparrot_training - Step 22971: {'lr': 0.00029923788615820936, 'samples': 11761664, 'steps': 22971, 'loss/train': 2.0191218852996826} 02/25/2022 07:19:51 - INFO - codeparrot_training - Step 22972: {'lr': 0.0002992218440789025, 'samples': 11762176, 'steps': 22972, 'loss/train': 2.4375100135803223} 02/25/2022 07:19:54 - INFO - codeparrot_training - Step 22973: {'lr': 0.00029920580178874497, 'samples': 11762688, 'steps': 22973, 'loss/train': 1.9143273830413818} 02/25/2022 07:20:01 - INFO - codeparrot_training - Step 22974: {'lr': 0.00029918975928780537, 'samples': 11763200, 'steps': 22974, 'loss/train': 1.580680251121521} 02/25/2022 07:20:04 - INFO - codeparrot_training - Step 22975: {'lr': 0.0002991737165761525, 'samples': 11763712, 'steps': 22975, 'loss/train': 2.1008057594299316} 02/25/2022 07:20:10 - INFO - codeparrot_training - Step 22976: {'lr': 0.000299157673653855, 'samples': 11764224, 'steps': 22976, 'loss/train': 2.170525550842285} 02/25/2022 07:20:13 - INFO - codeparrot_training - Step 22977: {'lr': 0.0002991416305209817, 'samples': 11764736, 'steps': 22977, 'loss/train': 2.5588033199310303} 02/25/2022 07:20:19 - INFO - codeparrot_training - Step 22978: {'lr': 0.0002991255871776012, 'samples': 11765248, 'steps': 22978, 'loss/train': 1.8432810306549072} 02/25/2022 07:20:22 - INFO - codeparrot_training - Step 22979: {'lr': 0.0002991095436237823, 'samples': 11765760, 'steps': 22979, 'loss/train': 2.135591745376587} 02/25/2022 07:20:28 - INFO - codeparrot_training - Step 22980: {'lr': 0.00029909349985959377, 'samples': 11766272, 'steps': 22980, 'loss/train': 2.1418495178222656} 02/25/2022 07:20:32 - INFO - codeparrot_training - Step 22981: {'lr': 0.00029907745588510416, 'samples': 11766784, 'steps': 22981, 'loss/train': 2.6582584381103516} 02/25/2022 07:20:37 - INFO - codeparrot_training - Step 22982: {'lr': 0.00029906141170038243, 'samples': 11767296, 'steps': 22982, 'loss/train': 0.7367690801620483} 02/25/2022 07:20:41 - INFO - codeparrot_training - Step 22983: {'lr': 0.00029904536730549706, 'samples': 11767808, 'steps': 22983, 'loss/train': 2.2438948154449463} 02/25/2022 07:20:47 - INFO - codeparrot_training - Step 22984: {'lr': 0.00029902932270051705, 'samples': 11768320, 'steps': 22984, 'loss/train': 1.4486254453659058} 02/25/2022 07:20:51 - INFO - codeparrot_training - Step 22985: {'lr': 0.00029901327788551087, 'samples': 11768832, 'steps': 22985, 'loss/train': 1.742058277130127} 02/25/2022 07:20:56 - INFO - codeparrot_training - Step 22986: {'lr': 0.0002989972328605475, 'samples': 11769344, 'steps': 22986, 'loss/train': 1.4603650569915771} 02/25/2022 07:21:00 - INFO - codeparrot_training - Step 22987: {'lr': 0.0002989811876256954, 'samples': 11769856, 'steps': 22987, 'loss/train': 2.547775983810425} 02/25/2022 07:21:05 - INFO - codeparrot_training - Step 22988: {'lr': 0.0002989651421810235, 'samples': 11770368, 'steps': 22988, 'loss/train': 2.5148606300354004} 02/25/2022 07:21:09 - INFO - codeparrot_training - Step 22989: {'lr': 0.00029894909652660053, 'samples': 11770880, 'steps': 22989, 'loss/train': 2.832615613937378} 02/25/2022 07:21:14 - INFO - codeparrot_training - Step 22990: {'lr': 0.0002989330506624951, 'samples': 11771392, 'steps': 22990, 'loss/train': 1.343993067741394} 02/25/2022 07:21:18 - INFO - codeparrot_training - Step 22991: {'lr': 0.0002989170045887761, 'samples': 11771904, 'steps': 22991, 'loss/train': 1.6383126974105835} 02/25/2022 07:21:24 - INFO - codeparrot_training - Step 22992: {'lr': 0.00029890095830551204, 'samples': 11772416, 'steps': 22992, 'loss/train': 1.9948525428771973} 02/25/2022 07:21:27 - INFO - codeparrot_training - Step 22993: {'lr': 0.00029888491181277195, 'samples': 11772928, 'steps': 22993, 'loss/train': 0.6399610638618469} 02/25/2022 07:21:33 - INFO - codeparrot_training - Step 22994: {'lr': 0.00029886886511062434, 'samples': 11773440, 'steps': 22994, 'loss/train': 2.6558916568756104} 02/25/2022 07:21:36 - INFO - codeparrot_training - Step 22995: {'lr': 0.0002988528181991381, 'samples': 11773952, 'steps': 22995, 'loss/train': 0.29033002257347107} 02/25/2022 07:21:42 - INFO - codeparrot_training - Step 22996: {'lr': 0.00029883677107838183, 'samples': 11774464, 'steps': 22996, 'loss/train': 1.6300299167633057} 02/25/2022 07:21:46 - INFO - codeparrot_training - Step 22997: {'lr': 0.0002988207237484244, 'samples': 11774976, 'steps': 22997, 'loss/train': 2.483008623123169} 02/25/2022 07:21:51 - INFO - codeparrot_training - Step 22998: {'lr': 0.0002988046762093344, 'samples': 11775488, 'steps': 22998, 'loss/train': 1.0141639709472656} 02/25/2022 07:21:55 - INFO - codeparrot_training - Step 22999: {'lr': 0.00029878862846118075, 'samples': 11776000, 'steps': 22999, 'loss/train': 0.8634439706802368} 02/25/2022 07:21:55 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 07:22:12 - WARNING - huggingface_hub.repository - Several commits (23) will be pushed upstream. 02/25/2022 07:22:12 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 07:22:46 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy de93bae..04e6e2d floral-grass-11 -> floral-grass-11 02/25/2022 07:22:53 - INFO - codeparrot_training - Step 23000: {'lr': 0.0002987725805040321, 'samples': 11776512, 'steps': 23000, 'loss/train': 2.382962703704834} 02/25/2022 07:22:56 - INFO - codeparrot_training - Step 23001: {'lr': 0.00029875653233795715, 'samples': 11777024, 'steps': 23001, 'loss/train': 1.9941284656524658} 02/25/2022 07:23:02 - INFO - codeparrot_training - Step 23002: {'lr': 0.0002987404839630248, 'samples': 11777536, 'steps': 23002, 'loss/train': 2.2633652687072754} 02/25/2022 07:23:05 - INFO - codeparrot_training - Step 23003: {'lr': 0.00029872443537930357, 'samples': 11778048, 'steps': 23003, 'loss/train': 2.39821457862854} 02/25/2022 07:23:09 - INFO - codeparrot_training - Step 23004: {'lr': 0.0002987083865868624, 'samples': 11778560, 'steps': 23004, 'loss/train': 2.1304678916931152} 02/25/2022 07:23:14 - INFO - codeparrot_training - Step 23005: {'lr': 0.0002986923375857699, 'samples': 11779072, 'steps': 23005, 'loss/train': 0.8211341500282288} 02/25/2022 07:23:18 - INFO - codeparrot_training - Step 23006: {'lr': 0.00029867628837609503, 'samples': 11779584, 'steps': 23006, 'loss/train': 2.1896445751190186} 02/25/2022 07:23:23 - INFO - codeparrot_training - Step 23007: {'lr': 0.0002986602389579062, 'samples': 11780096, 'steps': 23007, 'loss/train': 1.629866123199463} 02/25/2022 07:23:29 - INFO - codeparrot_training - Step 23008: {'lr': 0.0002986441893312726, 'samples': 11780608, 'steps': 23008, 'loss/train': 1.587494969367981} 02/25/2022 07:23:32 - INFO - codeparrot_training - Step 23009: {'lr': 0.0002986281394962626, 'samples': 11781120, 'steps': 23009, 'loss/train': 1.5454816818237305} 02/25/2022 07:23:39 - INFO - codeparrot_training - Step 23010: {'lr': 0.00029861208945294507, 'samples': 11781632, 'steps': 23010, 'loss/train': 2.3994152545928955} 02/25/2022 07:23:42 - INFO - codeparrot_training - Step 23011: {'lr': 0.00029859603920138876, 'samples': 11782144, 'steps': 23011, 'loss/train': 1.8347333669662476} 02/25/2022 07:23:48 - INFO - codeparrot_training - Step 23012: {'lr': 0.00029857998874166253, 'samples': 11782656, 'steps': 23012, 'loss/train': 1.088434100151062} 02/25/2022 07:23:51 - INFO - codeparrot_training - Step 23013: {'lr': 0.00029856393807383504, 'samples': 11783168, 'steps': 23013, 'loss/train': 1.871364712715149} 02/25/2022 07:23:57 - INFO - codeparrot_training - Step 23014: {'lr': 0.000298547887197975, 'samples': 11783680, 'steps': 23014, 'loss/train': 1.372658133506775} 02/25/2022 07:24:00 - INFO - codeparrot_training - Step 23015: {'lr': 0.0002985318361141513, 'samples': 11784192, 'steps': 23015, 'loss/train': 1.7402955293655396} 02/25/2022 07:24:06 - INFO - codeparrot_training - Step 23016: {'lr': 0.0002985157848224326, 'samples': 11784704, 'steps': 23016, 'loss/train': 1.1968350410461426} 02/25/2022 07:24:09 - INFO - codeparrot_training - Step 23017: {'lr': 0.00029849973332288763, 'samples': 11785216, 'steps': 23017, 'loss/train': 2.6905784606933594} 02/25/2022 07:24:15 - INFO - codeparrot_training - Step 23018: {'lr': 0.00029848368161558526, 'samples': 11785728, 'steps': 23018, 'loss/train': 1.9995077848434448} 02/25/2022 07:24:18 - INFO - codeparrot_training - Step 23019: {'lr': 0.0002984676297005942, 'samples': 11786240, 'steps': 23019, 'loss/train': 1.9210776090621948} 02/25/2022 07:24:24 - INFO - codeparrot_training - Step 23020: {'lr': 0.0002984515775779832, 'samples': 11786752, 'steps': 23020, 'loss/train': 2.007884979248047} 02/25/2022 07:24:28 - INFO - codeparrot_training - Step 23021: {'lr': 0.00029843552524782104, 'samples': 11787264, 'steps': 23021, 'loss/train': 3.0704941749572754} 02/25/2022 07:24:33 - INFO - codeparrot_training - Step 23022: {'lr': 0.00029841947271017647, 'samples': 11787776, 'steps': 23022, 'loss/train': 1.3323493003845215} 02/25/2022 07:24:37 - INFO - codeparrot_training - Step 23023: {'lr': 0.0002984034199651182, 'samples': 11788288, 'steps': 23023, 'loss/train': 1.605269193649292} 02/25/2022 07:24:42 - INFO - codeparrot_training - Step 23024: {'lr': 0.00029838736701271514, 'samples': 11788800, 'steps': 23024, 'loss/train': 2.2570393085479736} 02/25/2022 07:24:46 - INFO - codeparrot_training - Step 23025: {'lr': 0.00029837131385303587, 'samples': 11789312, 'steps': 23025, 'loss/train': 1.5309020280838013} 02/25/2022 07:24:52 - INFO - codeparrot_training - Step 23026: {'lr': 0.0002983552604861493, 'samples': 11789824, 'steps': 23026, 'loss/train': 2.29774808883667} 02/25/2022 07:24:55 - INFO - codeparrot_training - Step 23027: {'lr': 0.0002983392069121241, 'samples': 11790336, 'steps': 23027, 'loss/train': 1.4504623413085938} 02/25/2022 07:25:00 - INFO - codeparrot_training - Step 23028: {'lr': 0.00029832315313102915, 'samples': 11790848, 'steps': 23028, 'loss/train': 1.5331591367721558} 02/25/2022 07:25:04 - INFO - codeparrot_training - Step 23029: {'lr': 0.00029830709914293306, 'samples': 11791360, 'steps': 23029, 'loss/train': 2.652798652648926} 02/25/2022 07:25:09 - INFO - codeparrot_training - Step 23030: {'lr': 0.00029829104494790483, 'samples': 11791872, 'steps': 23030, 'loss/train': 2.51054310798645} 02/25/2022 07:25:13 - INFO - codeparrot_training - Step 23031: {'lr': 0.00029827499054601306, 'samples': 11792384, 'steps': 23031, 'loss/train': 1.9142118692398071} 02/25/2022 07:25:19 - INFO - codeparrot_training - Step 23032: {'lr': 0.0002982589359373265, 'samples': 11792896, 'steps': 23032, 'loss/train': 2.2765953540802} 02/25/2022 07:25:23 - INFO - codeparrot_training - Step 23033: {'lr': 0.000298242881121914, 'samples': 11793408, 'steps': 23033, 'loss/train': 2.8762152194976807} 02/25/2022 07:25:28 - INFO - codeparrot_training - Step 23034: {'lr': 0.00029822682609984436, 'samples': 11793920, 'steps': 23034, 'loss/train': 1.7421038150787354} 02/25/2022 07:25:31 - INFO - codeparrot_training - Step 23035: {'lr': 0.00029821077087118625, 'samples': 11794432, 'steps': 23035, 'loss/train': 1.9507066011428833} 02/25/2022 07:25:37 - INFO - codeparrot_training - Step 23036: {'lr': 0.00029819471543600856, 'samples': 11794944, 'steps': 23036, 'loss/train': 2.204113245010376} 02/25/2022 07:25:41 - INFO - codeparrot_training - Step 23037: {'lr': 0.00029817865979437996, 'samples': 11795456, 'steps': 23037, 'loss/train': 2.5255625247955322} 02/25/2022 07:25:46 - INFO - codeparrot_training - Step 23038: {'lr': 0.0002981626039463693, 'samples': 11795968, 'steps': 23038, 'loss/train': 1.2746597528457642} 02/25/2022 07:25:50 - INFO - codeparrot_training - Step 23039: {'lr': 0.0002981465478920454, 'samples': 11796480, 'steps': 23039, 'loss/train': 1.636218547821045} 02/25/2022 07:25:55 - INFO - codeparrot_training - Step 23040: {'lr': 0.00029813049163147687, 'samples': 11796992, 'steps': 23040, 'loss/train': 2.1896657943725586} 02/25/2022 07:25:59 - INFO - codeparrot_training - Step 23041: {'lr': 0.0002981144351647327, 'samples': 11797504, 'steps': 23041, 'loss/train': 1.9750356674194336} 02/25/2022 07:26:04 - INFO - codeparrot_training - Step 23042: {'lr': 0.0002980983784918815, 'samples': 11798016, 'steps': 23042, 'loss/train': 3.3246102333068848} 02/25/2022 07:26:08 - INFO - codeparrot_training - Step 23043: {'lr': 0.0002980823216129921, 'samples': 11798528, 'steps': 23043, 'loss/train': 1.5700490474700928} 02/25/2022 07:26:13 - INFO - codeparrot_training - Step 23044: {'lr': 0.00029806626452813333, 'samples': 11799040, 'steps': 23044, 'loss/train': 1.8791347742080688} 02/25/2022 07:26:17 - INFO - codeparrot_training - Step 23045: {'lr': 0.000298050207237374, 'samples': 11799552, 'steps': 23045, 'loss/train': 1.9777239561080933} 02/25/2022 07:26:23 - INFO - codeparrot_training - Step 23046: {'lr': 0.0002980341497407828, 'samples': 11800064, 'steps': 23046, 'loss/train': 1.7885113954544067} 02/25/2022 07:26:27 - INFO - codeparrot_training - Step 23047: {'lr': 0.0002980180920384286, 'samples': 11800576, 'steps': 23047, 'loss/train': 1.5340200662612915} 02/25/2022 07:26:32 - INFO - codeparrot_training - Step 23048: {'lr': 0.00029800203413038, 'samples': 11801088, 'steps': 23048, 'loss/train': 1.889414668083191} 02/25/2022 07:26:36 - INFO - codeparrot_training - Step 23049: {'lr': 0.000297985976016706, 'samples': 11801600, 'steps': 23049, 'loss/train': 1.3475145101547241} 02/25/2022 07:26:41 - INFO - codeparrot_training - Step 23050: {'lr': 0.0002979699176974754, 'samples': 11802112, 'steps': 23050, 'loss/train': 1.8376847505569458} 02/25/2022 07:26:45 - INFO - codeparrot_training - Step 23051: {'lr': 0.0002979538591727568, 'samples': 11802624, 'steps': 23051, 'loss/train': 1.7431894540786743} 02/25/2022 07:26:50 - INFO - codeparrot_training - Step 23052: {'lr': 0.00029793780044261916, 'samples': 11803136, 'steps': 23052, 'loss/train': 2.2214648723602295} 02/25/2022 07:26:54 - INFO - codeparrot_training - Step 23053: {'lr': 0.0002979217415071311, 'samples': 11803648, 'steps': 23053, 'loss/train': 2.104339122772217} 02/25/2022 07:26:59 - INFO - codeparrot_training - Step 23054: {'lr': 0.00029790568236636166, 'samples': 11804160, 'steps': 23054, 'loss/train': 1.9334293603897095} 02/25/2022 07:27:03 - INFO - codeparrot_training - Step 23055: {'lr': 0.0002978896230203794, 'samples': 11804672, 'steps': 23055, 'loss/train': 1.538697600364685} 02/25/2022 07:27:09 - INFO - codeparrot_training - Step 23056: {'lr': 0.0002978735634692533, 'samples': 11805184, 'steps': 23056, 'loss/train': 2.4216134548187256} 02/25/2022 07:27:12 - INFO - codeparrot_training - Step 23057: {'lr': 0.0002978575037130519, 'samples': 11805696, 'steps': 23057, 'loss/train': 1.93329656124115} 02/25/2022 07:27:18 - INFO - codeparrot_training - Step 23058: {'lr': 0.0002978414437518443, 'samples': 11806208, 'steps': 23058, 'loss/train': 0.9649325013160706} 02/25/2022 07:27:21 - INFO - codeparrot_training - Step 23059: {'lr': 0.00029782538358569905, 'samples': 11806720, 'steps': 23059, 'loss/train': 0.5590830445289612} 02/25/2022 07:27:27 - INFO - codeparrot_training - Step 23060: {'lr': 0.0002978093232146851, 'samples': 11807232, 'steps': 23060, 'loss/train': 2.119703769683838} 02/25/2022 07:27:30 - INFO - codeparrot_training - Step 23061: {'lr': 0.00029779326263887113, 'samples': 11807744, 'steps': 23061, 'loss/train': 1.141127586364746} 02/25/2022 07:27:36 - INFO - codeparrot_training - Step 23062: {'lr': 0.00029777720185832605, 'samples': 11808256, 'steps': 23062, 'loss/train': 2.6110951900482178} 02/25/2022 07:27:40 - INFO - codeparrot_training - Step 23063: {'lr': 0.0002977611408731186, 'samples': 11808768, 'steps': 23063, 'loss/train': 2.351412534713745} 02/25/2022 07:27:43 - INFO - codeparrot_training - Step 23064: {'lr': 0.0002977450796833176, 'samples': 11809280, 'steps': 23064, 'loss/train': 0.686683177947998} 02/25/2022 07:27:49 - INFO - codeparrot_training - Step 23065: {'lr': 0.0002977290182889918, 'samples': 11809792, 'steps': 23065, 'loss/train': 1.3830899000167847} 02/25/2022 07:27:52 - INFO - codeparrot_training - Step 23066: {'lr': 0.00029771295669021, 'samples': 11810304, 'steps': 23066, 'loss/train': 1.66257643699646} 02/25/2022 07:27:59 - INFO - codeparrot_training - Step 23067: {'lr': 0.00029769689488704117, 'samples': 11810816, 'steps': 23067, 'loss/train': 1.7433782815933228} 02/25/2022 07:28:02 - INFO - codeparrot_training - Step 23068: {'lr': 0.00029768083287955394, 'samples': 11811328, 'steps': 23068, 'loss/train': 0.8610262274742126} 02/25/2022 07:28:07 - INFO - codeparrot_training - Step 23069: {'lr': 0.00029766477066781716, 'samples': 11811840, 'steps': 23069, 'loss/train': 2.1658577919006348} 02/25/2022 07:28:11 - INFO - codeparrot_training - Step 23070: {'lr': 0.0002976487082518996, 'samples': 11812352, 'steps': 23070, 'loss/train': 1.1951266527175903} 02/25/2022 07:28:17 - INFO - codeparrot_training - Step 23071: {'lr': 0.00029763264563187016, 'samples': 11812864, 'steps': 23071, 'loss/train': 1.8942818641662598} 02/25/2022 07:28:20 - INFO - codeparrot_training - Step 23072: {'lr': 0.0002976165828077975, 'samples': 11813376, 'steps': 23072, 'loss/train': 1.3101755380630493} 02/25/2022 07:28:26 - INFO - codeparrot_training - Step 23073: {'lr': 0.00029760051977975057, 'samples': 11813888, 'steps': 23073, 'loss/train': 1.9660675525665283} 02/25/2022 07:28:29 - INFO - codeparrot_training - Step 23074: {'lr': 0.00029758445654779814, 'samples': 11814400, 'steps': 23074, 'loss/train': 1.7070472240447998} 02/25/2022 07:28:35 - INFO - codeparrot_training - Step 23075: {'lr': 0.000297568393112009, 'samples': 11814912, 'steps': 23075, 'loss/train': 1.3628267049789429} 02/25/2022 07:28:38 - INFO - codeparrot_training - Step 23076: {'lr': 0.00029755232947245195, 'samples': 11815424, 'steps': 23076, 'loss/train': 1.8507574796676636} 02/25/2022 07:28:44 - INFO - codeparrot_training - Step 23077: {'lr': 0.0002975362656291958, 'samples': 11815936, 'steps': 23077, 'loss/train': 2.1007096767425537} 02/25/2022 07:28:47 - INFO - codeparrot_training - Step 23078: {'lr': 0.0002975202015823095, 'samples': 11816448, 'steps': 23078, 'loss/train': 0.45156094431877136} 02/25/2022 07:28:53 - INFO - codeparrot_training - Step 23079: {'lr': 0.0002975041373318617, 'samples': 11816960, 'steps': 23079, 'loss/train': 1.8716130256652832} 02/25/2022 07:28:57 - INFO - codeparrot_training - Step 23080: {'lr': 0.0002974880728779212, 'samples': 11817472, 'steps': 23080, 'loss/train': 1.1723854541778564} 02/25/2022 07:29:02 - INFO - codeparrot_training - Step 23081: {'lr': 0.00029747200822055684, 'samples': 11817984, 'steps': 23081, 'loss/train': 0.7784999012947083} 02/25/2022 07:29:05 - INFO - codeparrot_training - Step 23082: {'lr': 0.0002974559433598376, 'samples': 11818496, 'steps': 23082, 'loss/train': 1.902015209197998} 02/25/2022 07:29:11 - INFO - codeparrot_training - Step 23083: {'lr': 0.0002974398782958321, 'samples': 11819008, 'steps': 23083, 'loss/train': 1.523607611656189} 02/25/2022 07:29:15 - INFO - codeparrot_training - Step 23084: {'lr': 0.00029742381302860923, 'samples': 11819520, 'steps': 23084, 'loss/train': 2.626331329345703} 02/25/2022 07:29:20 - INFO - codeparrot_training - Step 23085: {'lr': 0.00029740774755823777, 'samples': 11820032, 'steps': 23085, 'loss/train': 2.2582669258117676} 02/25/2022 07:29:24 - INFO - codeparrot_training - Step 23086: {'lr': 0.0002973916818847866, 'samples': 11820544, 'steps': 23086, 'loss/train': 3.8212239742279053} 02/25/2022 07:29:29 - INFO - codeparrot_training - Step 23087: {'lr': 0.00029737561600832454, 'samples': 11821056, 'steps': 23087, 'loss/train': 2.4119222164154053} 02/25/2022 07:29:33 - INFO - codeparrot_training - Step 23088: {'lr': 0.00029735954992892035, 'samples': 11821568, 'steps': 23088, 'loss/train': 2.844414234161377} 02/25/2022 07:29:38 - INFO - codeparrot_training - Step 23089: {'lr': 0.00029734348364664285, 'samples': 11822080, 'steps': 23089, 'loss/train': 1.7163355350494385} 02/25/2022 07:29:42 - INFO - codeparrot_training - Step 23090: {'lr': 0.000297327417161561, 'samples': 11822592, 'steps': 23090, 'loss/train': 1.9642480611801147} 02/25/2022 07:29:47 - INFO - codeparrot_training - Step 23091: {'lr': 0.0002973113504737435, 'samples': 11823104, 'steps': 23091, 'loss/train': 2.682631492614746} 02/25/2022 07:29:51 - INFO - codeparrot_training - Step 23092: {'lr': 0.00029729528358325914, 'samples': 11823616, 'steps': 23092, 'loss/train': 0.8743570446968079} 02/25/2022 07:29:57 - INFO - codeparrot_training - Step 23093: {'lr': 0.00029727921649017687, 'samples': 11824128, 'steps': 23093, 'loss/train': 0.8451046943664551} 02/25/2022 07:30:00 - INFO - codeparrot_training - Step 23094: {'lr': 0.0002972631491945653, 'samples': 11824640, 'steps': 23094, 'loss/train': 1.8207341432571411} 02/25/2022 07:30:06 - INFO - codeparrot_training - Step 23095: {'lr': 0.00029724708169649364, 'samples': 11825152, 'steps': 23095, 'loss/train': 2.172386646270752} 02/25/2022 07:30:10 - INFO - codeparrot_training - Step 23096: {'lr': 0.0002972310139960303, 'samples': 11825664, 'steps': 23096, 'loss/train': 1.44020676612854} 02/25/2022 07:30:15 - INFO - codeparrot_training - Step 23097: {'lr': 0.00029721494609324435, 'samples': 11826176, 'steps': 23097, 'loss/train': 1.4003266096115112} 02/25/2022 07:30:19 - INFO - codeparrot_training - Step 23098: {'lr': 0.00029719887798820453, 'samples': 11826688, 'steps': 23098, 'loss/train': 1.6491516828536987} 02/25/2022 07:30:24 - INFO - codeparrot_training - Step 23099: {'lr': 0.00029718280968097976, 'samples': 11827200, 'steps': 23099, 'loss/train': 1.5048080682754517} 02/25/2022 07:30:28 - INFO - codeparrot_training - Step 23100: {'lr': 0.00029716674117163884, 'samples': 11827712, 'steps': 23100, 'loss/train': 1.9512022733688354} 02/25/2022 07:30:33 - INFO - codeparrot_training - Step 23101: {'lr': 0.0002971506724602505, 'samples': 11828224, 'steps': 23101, 'loss/train': 1.9504514932632446} 02/25/2022 07:30:37 - INFO - codeparrot_training - Step 23102: {'lr': 0.0002971346035468837, 'samples': 11828736, 'steps': 23102, 'loss/train': 1.9367523193359375} 02/25/2022 07:30:43 - INFO - codeparrot_training - Step 23103: {'lr': 0.0002971185344316072, 'samples': 11829248, 'steps': 23103, 'loss/train': 1.373284101486206} 02/25/2022 07:30:46 - INFO - codeparrot_training - Step 23104: {'lr': 0.0002971024651144899, 'samples': 11829760, 'steps': 23104, 'loss/train': 1.8134773969650269} 02/25/2022 07:30:52 - INFO - codeparrot_training - Step 23105: {'lr': 0.0002970863955956005, 'samples': 11830272, 'steps': 23105, 'loss/train': 2.2017300128936768} 02/25/2022 07:30:55 - INFO - codeparrot_training - Step 23106: {'lr': 0.00029707032587500805, 'samples': 11830784, 'steps': 23106, 'loss/train': 1.5983376502990723} 02/25/2022 07:31:01 - INFO - codeparrot_training - Step 23107: {'lr': 0.00029705425595278126, 'samples': 11831296, 'steps': 23107, 'loss/train': 1.9444851875305176} 02/25/2022 07:31:04 - INFO - codeparrot_training - Step 23108: {'lr': 0.0002970381858289889, 'samples': 11831808, 'steps': 23108, 'loss/train': 0.9139845967292786} 02/25/2022 07:31:10 - INFO - codeparrot_training - Step 23109: {'lr': 0.0002970221155037, 'samples': 11832320, 'steps': 23109, 'loss/train': 1.834768295288086} 02/25/2022 07:31:13 - INFO - codeparrot_training - Step 23110: {'lr': 0.0002970060449769832, 'samples': 11832832, 'steps': 23110, 'loss/train': 1.7454090118408203} 02/25/2022 07:31:19 - INFO - codeparrot_training - Step 23111: {'lr': 0.00029698997424890746, 'samples': 11833344, 'steps': 23111, 'loss/train': 1.0150907039642334} 02/25/2022 07:31:22 - INFO - codeparrot_training - Step 23112: {'lr': 0.0002969739033195415, 'samples': 11833856, 'steps': 23112, 'loss/train': 2.812715768814087} 02/25/2022 07:31:29 - INFO - codeparrot_training - Step 23113: {'lr': 0.0002969578321889544, 'samples': 11834368, 'steps': 23113, 'loss/train': 1.3748753070831299} 02/25/2022 07:31:32 - INFO - codeparrot_training - Step 23114: {'lr': 0.00029694176085721474, 'samples': 11834880, 'steps': 23114, 'loss/train': 1.3746641874313354} 02/25/2022 07:31:38 - INFO - codeparrot_training - Step 23115: {'lr': 0.0002969256893243916, 'samples': 11835392, 'steps': 23115, 'loss/train': 2.402064800262451} 02/25/2022 07:31:41 - INFO - codeparrot_training - Step 23116: {'lr': 0.0002969096175905536, 'samples': 11835904, 'steps': 23116, 'loss/train': 0.9432001113891602} 02/25/2022 07:31:47 - INFO - codeparrot_training - Step 23117: {'lr': 0.00029689354565576976, 'samples': 11836416, 'steps': 23117, 'loss/train': 1.957139492034912} 02/25/2022 07:31:50 - INFO - codeparrot_training - Step 23118: {'lr': 0.0002968774735201088, 'samples': 11836928, 'steps': 23118, 'loss/train': 1.8836572170257568} 02/25/2022 07:31:56 - INFO - codeparrot_training - Step 23119: {'lr': 0.0002968614011836397, 'samples': 11837440, 'steps': 23119, 'loss/train': 1.0959608554840088} 02/25/2022 07:31:59 - INFO - codeparrot_training - Step 23120: {'lr': 0.0002968453286464312, 'samples': 11837952, 'steps': 23120, 'loss/train': 2.70177960395813} 02/25/2022 07:32:05 - INFO - codeparrot_training - Step 23121: {'lr': 0.00029682925590855213, 'samples': 11838464, 'steps': 23121, 'loss/train': 2.2461211681365967} 02/25/2022 07:32:08 - INFO - codeparrot_training - Step 23122: {'lr': 0.0002968131829700715, 'samples': 11838976, 'steps': 23122, 'loss/train': 1.3319445848464966} 02/25/2022 07:32:14 - INFO - codeparrot_training - Step 23123: {'lr': 0.000296797109831058, 'samples': 11839488, 'steps': 23123, 'loss/train': 1.2728564739227295} 02/25/2022 07:32:17 - INFO - codeparrot_training - Step 23124: {'lr': 0.00029678103649158057, 'samples': 11840000, 'steps': 23124, 'loss/train': 1.7950807809829712} 02/25/2022 07:32:24 - INFO - codeparrot_training - Step 23125: {'lr': 0.00029676496295170804, 'samples': 11840512, 'steps': 23125, 'loss/train': 1.8831136226654053} 02/25/2022 07:32:27 - INFO - codeparrot_training - Step 23126: {'lr': 0.0002967488892115092, 'samples': 11841024, 'steps': 23126, 'loss/train': 1.8715931177139282} 02/25/2022 07:32:33 - INFO - codeparrot_training - Step 23127: {'lr': 0.000296732815271053, 'samples': 11841536, 'steps': 23127, 'loss/train': 2.4496424198150635} 02/25/2022 07:32:36 - INFO - codeparrot_training - Step 23128: {'lr': 0.00029671674113040833, 'samples': 11842048, 'steps': 23128, 'loss/train': 1.496080994606018} 02/25/2022 07:32:42 - INFO - codeparrot_training - Step 23129: {'lr': 0.00029670066678964385, 'samples': 11842560, 'steps': 23129, 'loss/train': 2.152468204498291} 02/25/2022 07:32:45 - INFO - codeparrot_training - Step 23130: {'lr': 0.0002966845922488286, 'samples': 11843072, 'steps': 23130, 'loss/train': 2.4569578170776367} 02/25/2022 07:32:51 - INFO - codeparrot_training - Step 23131: {'lr': 0.00029666851750803137, 'samples': 11843584, 'steps': 23131, 'loss/train': 2.5395867824554443} 02/25/2022 07:32:55 - INFO - codeparrot_training - Step 23132: {'lr': 0.00029665244256732107, 'samples': 11844096, 'steps': 23132, 'loss/train': 1.6447722911834717} 02/25/2022 07:33:00 - INFO - codeparrot_training - Step 23133: {'lr': 0.0002966363674267665, 'samples': 11844608, 'steps': 23133, 'loss/train': 1.27051842212677} 02/25/2022 07:33:04 - INFO - codeparrot_training - Step 23134: {'lr': 0.00029662029208643646, 'samples': 11845120, 'steps': 23134, 'loss/train': 1.4961507320404053} 02/25/2022 07:33:09 - INFO - codeparrot_training - Step 23135: {'lr': 0.0002966042165464, 'samples': 11845632, 'steps': 23135, 'loss/train': 1.6537866592407227} 02/25/2022 07:33:13 - INFO - codeparrot_training - Step 23136: {'lr': 0.0002965881408067258, 'samples': 11846144, 'steps': 23136, 'loss/train': 1.022843599319458} 02/25/2022 07:33:18 - INFO - codeparrot_training - Step 23137: {'lr': 0.0002965720648674829, 'samples': 11846656, 'steps': 23137, 'loss/train': 1.7009072303771973} 02/25/2022 07:33:22 - INFO - codeparrot_training - Step 23138: {'lr': 0.00029655598872873994, 'samples': 11847168, 'steps': 23138, 'loss/train': 0.5573568344116211} 02/25/2022 07:33:27 - INFO - codeparrot_training - Step 23139: {'lr': 0.000296539912390566, 'samples': 11847680, 'steps': 23139, 'loss/train': 2.214735746383667} 02/25/2022 07:33:31 - INFO - codeparrot_training - Step 23140: {'lr': 0.0002965238358530298, 'samples': 11848192, 'steps': 23140, 'loss/train': 1.9246439933776855} 02/25/2022 07:33:37 - INFO - codeparrot_training - Step 23141: {'lr': 0.0002965077591162003, 'samples': 11848704, 'steps': 23141, 'loss/train': 1.6096656322479248} 02/25/2022 07:33:40 - INFO - codeparrot_training - Step 23142: {'lr': 0.00029649168218014627, 'samples': 11849216, 'steps': 23142, 'loss/train': 2.5338127613067627} 02/25/2022 07:33:46 - INFO - codeparrot_training - Step 23143: {'lr': 0.0002964756050449367, 'samples': 11849728, 'steps': 23143, 'loss/train': 1.336146593093872} 02/25/2022 07:33:49 - INFO - codeparrot_training - Step 23144: {'lr': 0.0002964595277106403, 'samples': 11850240, 'steps': 23144, 'loss/train': 0.9666948318481445} 02/25/2022 07:33:55 - INFO - codeparrot_training - Step 23145: {'lr': 0.0002964434501773262, 'samples': 11850752, 'steps': 23145, 'loss/train': 1.3791836500167847} 02/25/2022 07:33:58 - INFO - codeparrot_training - Step 23146: {'lr': 0.00029642737244506295, 'samples': 11851264, 'steps': 23146, 'loss/train': 1.6816203594207764} 02/25/2022 07:34:04 - INFO - codeparrot_training - Step 23147: {'lr': 0.0002964112945139196, 'samples': 11851776, 'steps': 23147, 'loss/train': 2.0669057369232178} 02/25/2022 07:34:07 - INFO - codeparrot_training - Step 23148: {'lr': 0.0002963952163839651, 'samples': 11852288, 'steps': 23148, 'loss/train': 2.180706262588501} 02/25/2022 07:34:13 - INFO - codeparrot_training - Step 23149: {'lr': 0.00029637913805526816, 'samples': 11852800, 'steps': 23149, 'loss/train': 2.2974071502685547} 02/25/2022 07:34:16 - INFO - codeparrot_training - Step 23150: {'lr': 0.0002963630595278977, 'samples': 11853312, 'steps': 23150, 'loss/train': 1.8673449754714966} 02/25/2022 07:34:23 - INFO - codeparrot_training - Step 23151: {'lr': 0.0002963469808019227, 'samples': 11853824, 'steps': 23151, 'loss/train': 1.8603227138519287} 02/25/2022 07:34:26 - INFO - codeparrot_training - Step 23152: {'lr': 0.00029633090187741187, 'samples': 11854336, 'steps': 23152, 'loss/train': 2.259190320968628} 02/25/2022 07:34:32 - INFO - codeparrot_training - Step 23153: {'lr': 0.00029631482275443424, 'samples': 11854848, 'steps': 23153, 'loss/train': 3.105224370956421} 02/25/2022 07:34:35 - INFO - codeparrot_training - Step 23154: {'lr': 0.0002962987434330585, 'samples': 11855360, 'steps': 23154, 'loss/train': 2.0582025051116943} 02/25/2022 07:34:41 - INFO - codeparrot_training - Step 23155: {'lr': 0.0002962826639133538, 'samples': 11855872, 'steps': 23155, 'loss/train': 0.9695888161659241} 02/25/2022 07:34:44 - INFO - codeparrot_training - Step 23156: {'lr': 0.00029626658419538873, 'samples': 11856384, 'steps': 23156, 'loss/train': 2.500014543533325} 02/25/2022 07:34:50 - INFO - codeparrot_training - Step 23157: {'lr': 0.0002962505042792324, 'samples': 11856896, 'steps': 23157, 'loss/train': 1.32759428024292} 02/25/2022 07:34:53 - INFO - codeparrot_training - Step 23158: {'lr': 0.0002962344241649535, 'samples': 11857408, 'steps': 23158, 'loss/train': 1.75245201587677} 02/25/2022 07:34:59 - INFO - codeparrot_training - Step 23159: {'lr': 0.000296218343852621, 'samples': 11857920, 'steps': 23159, 'loss/train': 2.4808712005615234} 02/25/2022 07:35:02 - INFO - codeparrot_training - Step 23160: {'lr': 0.0002962022633423039, 'samples': 11858432, 'steps': 23160, 'loss/train': 2.1747500896453857} 02/25/2022 07:35:09 - INFO - codeparrot_training - Step 23161: {'lr': 0.00029618618263407094, 'samples': 11858944, 'steps': 23161, 'loss/train': 2.921461820602417} 02/25/2022 07:35:12 - INFO - codeparrot_training - Step 23162: {'lr': 0.00029617010172799095, 'samples': 11859456, 'steps': 23162, 'loss/train': 1.1186691522598267} 02/25/2022 07:35:18 - INFO - codeparrot_training - Step 23163: {'lr': 0.00029615402062413307, 'samples': 11859968, 'steps': 23163, 'loss/train': 1.9791722297668457} 02/25/2022 07:35:21 - INFO - codeparrot_training - Step 23164: {'lr': 0.00029613793932256583, 'samples': 11860480, 'steps': 23164, 'loss/train': 1.3354239463806152} 02/25/2022 07:35:27 - INFO - codeparrot_training - Step 23165: {'lr': 0.00029612185782335845, 'samples': 11860992, 'steps': 23165, 'loss/train': 2.0327491760253906} 02/25/2022 07:35:30 - INFO - codeparrot_training - Step 23166: {'lr': 0.00029610577612657963, 'samples': 11861504, 'steps': 23166, 'loss/train': 1.3488519191741943} 02/25/2022 07:35:36 - INFO - codeparrot_training - Step 23167: {'lr': 0.0002960896942322983, 'samples': 11862016, 'steps': 23167, 'loss/train': 1.3818823099136353} 02/25/2022 07:35:39 - INFO - codeparrot_training - Step 23168: {'lr': 0.0002960736121405834, 'samples': 11862528, 'steps': 23168, 'loss/train': 2.542564868927002} 02/25/2022 07:35:44 - INFO - codeparrot_training - Step 23169: {'lr': 0.00029605752985150367, 'samples': 11863040, 'steps': 23169, 'loss/train': 1.3945231437683105} 02/25/2022 07:35:48 - INFO - codeparrot_training - Step 23170: {'lr': 0.00029604144736512816, 'samples': 11863552, 'steps': 23170, 'loss/train': 1.9836618900299072} 02/25/2022 07:35:55 - INFO - codeparrot_training - Step 23171: {'lr': 0.00029602536468152575, 'samples': 11864064, 'steps': 23171, 'loss/train': 1.4883031845092773} 02/25/2022 07:36:00 - INFO - codeparrot_training - Step 23172: {'lr': 0.00029600928180076525, 'samples': 11864576, 'steps': 23172, 'loss/train': 0.9137559533119202} 02/25/2022 07:36:04 - INFO - codeparrot_training - Step 23173: {'lr': 0.0002959931987229156, 'samples': 11865088, 'steps': 23173, 'loss/train': 1.8890084028244019} 02/25/2022 07:36:07 - INFO - codeparrot_training - Step 23174: {'lr': 0.0002959771154480457, 'samples': 11865600, 'steps': 23174, 'loss/train': 1.723131775856018} 02/25/2022 07:36:13 - INFO - codeparrot_training - Step 23175: {'lr': 0.0002959610319762244, 'samples': 11866112, 'steps': 23175, 'loss/train': 0.6872315406799316} 02/25/2022 07:36:18 - INFO - codeparrot_training - Step 23176: {'lr': 0.0002959449483075207, 'samples': 11866624, 'steps': 23176, 'loss/train': 1.8419396877288818} 02/25/2022 07:36:22 - INFO - codeparrot_training - Step 23177: {'lr': 0.0002959288644420034, 'samples': 11867136, 'steps': 23177, 'loss/train': 0.9334335327148438} 02/25/2022 07:36:27 - INFO - codeparrot_training - Step 23178: {'lr': 0.00029591278037974147, 'samples': 11867648, 'steps': 23178, 'loss/train': 2.110797882080078} 02/25/2022 07:36:31 - INFO - codeparrot_training - Step 23179: {'lr': 0.0002958966961208037, 'samples': 11868160, 'steps': 23179, 'loss/train': 1.0112143754959106} 02/25/2022 07:36:36 - INFO - codeparrot_training - Step 23180: {'lr': 0.0002958806116652591, 'samples': 11868672, 'steps': 23180, 'loss/train': 1.5760059356689453} 02/25/2022 07:36:40 - INFO - codeparrot_training - Step 23181: {'lr': 0.0002958645270131765, 'samples': 11869184, 'steps': 23181, 'loss/train': 1.6010059118270874} 02/25/2022 07:36:45 - INFO - codeparrot_training - Step 23182: {'lr': 0.0002958484421646248, 'samples': 11869696, 'steps': 23182, 'loss/train': 2.161012649536133} 02/25/2022 07:36:49 - INFO - codeparrot_training - Step 23183: {'lr': 0.000295832357119673, 'samples': 11870208, 'steps': 23183, 'loss/train': 3.2022106647491455} 02/25/2022 07:36:55 - INFO - codeparrot_training - Step 23184: {'lr': 0.00029581627187838993, 'samples': 11870720, 'steps': 23184, 'loss/train': 2.231543779373169} 02/25/2022 07:36:58 - INFO - codeparrot_training - Step 23185: {'lr': 0.0002958001864408445, 'samples': 11871232, 'steps': 23185, 'loss/train': 1.2037922143936157} 02/25/2022 07:37:04 - INFO - codeparrot_training - Step 23186: {'lr': 0.0002957841008071056, 'samples': 11871744, 'steps': 23186, 'loss/train': 2.15334153175354} 02/25/2022 07:37:08 - INFO - codeparrot_training - Step 23187: {'lr': 0.0002957680149772422, 'samples': 11872256, 'steps': 23187, 'loss/train': 2.383380651473999} 02/25/2022 07:37:13 - INFO - codeparrot_training - Step 23188: {'lr': 0.000295751928951323, 'samples': 11872768, 'steps': 23188, 'loss/train': 2.431396961212158} 02/25/2022 07:37:17 - INFO - codeparrot_training - Step 23189: {'lr': 0.0002957358427294172, 'samples': 11873280, 'steps': 23189, 'loss/train': 2.157102584838867} 02/25/2022 07:37:22 - INFO - codeparrot_training - Step 23190: {'lr': 0.00029571975631159355, 'samples': 11873792, 'steps': 23190, 'loss/train': 1.9784764051437378} 02/25/2022 07:37:26 - INFO - codeparrot_training - Step 23191: {'lr': 0.000295703669697921, 'samples': 11874304, 'steps': 23191, 'loss/train': 1.3329405784606934} 02/25/2022 07:37:31 - INFO - codeparrot_training - Step 23192: {'lr': 0.0002956875828884684, 'samples': 11874816, 'steps': 23192, 'loss/train': 1.9540085792541504} 02/25/2022 07:37:35 - INFO - codeparrot_training - Step 23193: {'lr': 0.0002956714958833047, 'samples': 11875328, 'steps': 23193, 'loss/train': 1.9511375427246094} 02/25/2022 07:37:40 - INFO - codeparrot_training - Step 23194: {'lr': 0.00029565540868249884, 'samples': 11875840, 'steps': 23194, 'loss/train': 2.8634321689605713} 02/25/2022 07:37:44 - INFO - codeparrot_training - Step 23195: {'lr': 0.0002956393212861197, 'samples': 11876352, 'steps': 23195, 'loss/train': 1.680410385131836} 02/25/2022 07:37:50 - INFO - codeparrot_training - Step 23196: {'lr': 0.0002956232336942362, 'samples': 11876864, 'steps': 23196, 'loss/train': 2.1952054500579834} 02/25/2022 07:37:53 - INFO - codeparrot_training - Step 23197: {'lr': 0.0002956071459069173, 'samples': 11877376, 'steps': 23197, 'loss/train': 2.2575438022613525} 02/25/2022 07:37:59 - INFO - codeparrot_training - Step 23198: {'lr': 0.0002955910579242318, 'samples': 11877888, 'steps': 23198, 'loss/train': 1.839053988456726} 02/25/2022 07:38:03 - INFO - codeparrot_training - Step 23199: {'lr': 0.0002955749697462487, 'samples': 11878400, 'steps': 23199, 'loss/train': 2.897972822189331} 02/25/2022 07:38:08 - INFO - codeparrot_training - Step 23200: {'lr': 0.0002955588813730369, 'samples': 11878912, 'steps': 23200, 'loss/train': 2.0402991771698} 02/25/2022 07:38:12 - INFO - codeparrot_training - Step 23201: {'lr': 0.0002955427928046653, 'samples': 11879424, 'steps': 23201, 'loss/train': 2.769561767578125} 02/25/2022 07:38:17 - INFO - codeparrot_training - Step 23202: {'lr': 0.0002955267040412029, 'samples': 11879936, 'steps': 23202, 'loss/train': 1.3466453552246094} 02/25/2022 07:38:21 - INFO - codeparrot_training - Step 23203: {'lr': 0.0002955106150827185, 'samples': 11880448, 'steps': 23203, 'loss/train': 2.1680798530578613} 02/25/2022 07:38:26 - INFO - codeparrot_training - Step 23204: {'lr': 0.0002954945259292811, 'samples': 11880960, 'steps': 23204, 'loss/train': 2.1050631999969482} 02/25/2022 07:38:30 - INFO - codeparrot_training - Step 23205: {'lr': 0.0002954784365809596, 'samples': 11881472, 'steps': 23205, 'loss/train': 2.2999377250671387} 02/25/2022 07:38:36 - INFO - codeparrot_training - Step 23206: {'lr': 0.0002954623470378229, 'samples': 11881984, 'steps': 23206, 'loss/train': 1.374948501586914} 02/25/2022 07:38:39 - INFO - codeparrot_training - Step 23207: {'lr': 0.00029544625729993997, 'samples': 11882496, 'steps': 23207, 'loss/train': 1.9059501886367798} 02/25/2022 07:38:45 - INFO - codeparrot_training - Step 23208: {'lr': 0.0002954301673673797, 'samples': 11883008, 'steps': 23208, 'loss/train': 1.356425404548645} 02/25/2022 07:38:48 - INFO - codeparrot_training - Step 23209: {'lr': 0.00029541407724021095, 'samples': 11883520, 'steps': 23209, 'loss/train': 1.6585997343063354} 02/25/2022 07:38:54 - INFO - codeparrot_training - Step 23210: {'lr': 0.00029539798691850274, 'samples': 11884032, 'steps': 23210, 'loss/train': 1.3325096368789673} 02/25/2022 07:38:58 - INFO - codeparrot_training - Step 23211: {'lr': 0.00029538189640232406, 'samples': 11884544, 'steps': 23211, 'loss/train': 2.0629372596740723} 02/25/2022 07:39:03 - INFO - codeparrot_training - Step 23212: {'lr': 0.00029536580569174364, 'samples': 11885056, 'steps': 23212, 'loss/train': 2.1043946743011475} 02/25/2022 07:39:07 - INFO - codeparrot_training - Step 23213: {'lr': 0.0002953497147868306, 'samples': 11885568, 'steps': 23213, 'loss/train': 1.8474342823028564} 02/25/2022 07:39:12 - INFO - codeparrot_training - Step 23214: {'lr': 0.0002953336236876537, 'samples': 11886080, 'steps': 23214, 'loss/train': 2.839002847671509} 02/25/2022 07:39:16 - INFO - codeparrot_training - Step 23215: {'lr': 0.000295317532394282, 'samples': 11886592, 'steps': 23215, 'loss/train': 0.8753924369812012} 02/25/2022 07:39:21 - INFO - codeparrot_training - Step 23216: {'lr': 0.0002953014409067844, 'samples': 11887104, 'steps': 23216, 'loss/train': 1.5727424621582031} 02/25/2022 07:39:25 - INFO - codeparrot_training - Step 23217: {'lr': 0.00029528534922522974, 'samples': 11887616, 'steps': 23217, 'loss/train': 1.8975706100463867} 02/25/2022 07:39:30 - INFO - codeparrot_training - Step 23218: {'lr': 0.00029526925734968705, 'samples': 11888128, 'steps': 23218, 'loss/train': 1.2779752016067505} 02/25/2022 07:39:34 - INFO - codeparrot_training - Step 23219: {'lr': 0.0002952531652802252, 'samples': 11888640, 'steps': 23219, 'loss/train': 2.38446044921875} 02/25/2022 07:39:39 - INFO - codeparrot_training - Step 23220: {'lr': 0.00029523707301691327, 'samples': 11889152, 'steps': 23220, 'loss/train': 2.2417595386505127} 02/25/2022 07:39:43 - INFO - codeparrot_training - Step 23221: {'lr': 0.00029522098055982, 'samples': 11889664, 'steps': 23221, 'loss/train': 1.8170905113220215} 02/25/2022 07:39:49 - INFO - codeparrot_training - Step 23222: {'lr': 0.00029520488790901446, 'samples': 11890176, 'steps': 23222, 'loss/train': 0.8516761064529419} 02/25/2022 07:39:53 - INFO - codeparrot_training - Step 23223: {'lr': 0.00029518879506456556, 'samples': 11890688, 'steps': 23223, 'loss/train': 0.8684076070785522} 02/25/2022 07:39:58 - INFO - codeparrot_training - Step 23224: {'lr': 0.0002951727020265421, 'samples': 11891200, 'steps': 23224, 'loss/train': 2.3690929412841797} 02/25/2022 07:40:02 - INFO - codeparrot_training - Step 23225: {'lr': 0.0002951566087950132, 'samples': 11891712, 'steps': 23225, 'loss/train': 1.27492356300354} 02/25/2022 07:40:07 - INFO - codeparrot_training - Step 23226: {'lr': 0.0002951405153700477, 'samples': 11892224, 'steps': 23226, 'loss/train': 2.8572022914886475} 02/25/2022 07:40:11 - INFO - codeparrot_training - Step 23227: {'lr': 0.0002951244217517146, 'samples': 11892736, 'steps': 23227, 'loss/train': 2.1676816940307617} 02/25/2022 07:40:16 - INFO - codeparrot_training - Step 23228: {'lr': 0.0002951083279400828, 'samples': 11893248, 'steps': 23228, 'loss/train': 1.5893467664718628} 02/25/2022 07:40:20 - INFO - codeparrot_training - Step 23229: {'lr': 0.00029509223393522124, 'samples': 11893760, 'steps': 23229, 'loss/train': 1.9131907224655151} 02/25/2022 07:40:25 - INFO - codeparrot_training - Step 23230: {'lr': 0.00029507613973719883, 'samples': 11894272, 'steps': 23230, 'loss/train': 1.2191270589828491} 02/25/2022 07:40:29 - INFO - codeparrot_training - Step 23231: {'lr': 0.0002950600453460846, 'samples': 11894784, 'steps': 23231, 'loss/train': 1.5991896390914917} 02/25/2022 07:40:34 - INFO - codeparrot_training - Step 23232: {'lr': 0.0002950439507619474, 'samples': 11895296, 'steps': 23232, 'loss/train': 2.732994318008423} 02/25/2022 07:40:38 - INFO - codeparrot_training - Step 23233: {'lr': 0.00029502785598485624, 'samples': 11895808, 'steps': 23233, 'loss/train': 3.0105879306793213} 02/25/2022 07:40:44 - INFO - codeparrot_training - Step 23234: {'lr': 0.00029501176101488004, 'samples': 11896320, 'steps': 23234, 'loss/train': 0.6744531393051147} 02/25/2022 07:40:47 - INFO - codeparrot_training - Step 23235: {'lr': 0.00029499566585208776, 'samples': 11896832, 'steps': 23235, 'loss/train': 2.4094879627227783} 02/25/2022 07:40:53 - INFO - codeparrot_training - Step 23236: {'lr': 0.00029497957049654834, 'samples': 11897344, 'steps': 23236, 'loss/train': 0.689185619354248} 02/25/2022 07:40:56 - INFO - codeparrot_training - Step 23237: {'lr': 0.0002949634749483307, 'samples': 11897856, 'steps': 23237, 'loss/train': 1.8878679275512695} 02/25/2022 07:41:02 - INFO - codeparrot_training - Step 23238: {'lr': 0.0002949473792075039, 'samples': 11898368, 'steps': 23238, 'loss/train': 1.487168312072754} 02/25/2022 07:41:05 - INFO - codeparrot_training - Step 23239: {'lr': 0.00029493128327413664, 'samples': 11898880, 'steps': 23239, 'loss/train': 0.883429229259491} 02/25/2022 07:41:11 - INFO - codeparrot_training - Step 23240: {'lr': 0.0002949151871482982, 'samples': 11899392, 'steps': 23240, 'loss/train': 2.5060760974884033} 02/25/2022 07:41:14 - INFO - codeparrot_training - Step 23241: {'lr': 0.00029489909083005723, 'samples': 11899904, 'steps': 23241, 'loss/train': 0.9486697316169739} 02/25/2022 07:41:20 - INFO - codeparrot_training - Step 23242: {'lr': 0.0002948829943194829, 'samples': 11900416, 'steps': 23242, 'loss/train': 1.5521860122680664} 02/25/2022 07:41:23 - INFO - codeparrot_training - Step 23243: {'lr': 0.00029486689761664396, 'samples': 11900928, 'steps': 23243, 'loss/train': 1.8223611116409302} 02/25/2022 07:41:30 - INFO - codeparrot_training - Step 23244: {'lr': 0.0002948508007216096, 'samples': 11901440, 'steps': 23244, 'loss/train': 0.6826063394546509} 02/25/2022 07:41:34 - INFO - codeparrot_training - Step 23245: {'lr': 0.00029483470363444856, 'samples': 11901952, 'steps': 23245, 'loss/train': 2.939300298690796} 02/25/2022 07:41:39 - INFO - codeparrot_training - Step 23246: {'lr': 0.00029481860635522994, 'samples': 11902464, 'steps': 23246, 'loss/train': 1.946687936782837} 02/25/2022 07:41:43 - INFO - codeparrot_training - Step 23247: {'lr': 0.0002948025088840226, 'samples': 11902976, 'steps': 23247, 'loss/train': 1.5869731903076172} 02/25/2022 07:41:48 - INFO - codeparrot_training - Step 23248: {'lr': 0.0002947864112208956, 'samples': 11903488, 'steps': 23248, 'loss/train': 2.951584577560425} 02/25/2022 07:41:52 - INFO - codeparrot_training - Step 23249: {'lr': 0.0002947703133659178, 'samples': 11904000, 'steps': 23249, 'loss/train': 1.869722843170166} 02/25/2022 07:41:57 - INFO - codeparrot_training - Step 23250: {'lr': 0.00029475421531915827, 'samples': 11904512, 'steps': 23250, 'loss/train': 3.3536550998687744} 02/25/2022 07:42:01 - INFO - codeparrot_training - Step 23251: {'lr': 0.00029473811708068576, 'samples': 11905024, 'steps': 23251, 'loss/train': 1.612632393836975} 02/25/2022 07:42:06 - INFO - codeparrot_training - Step 23252: {'lr': 0.0002947220186505694, 'samples': 11905536, 'steps': 23252, 'loss/train': 1.134292721748352} 02/25/2022 07:42:10 - INFO - codeparrot_training - Step 23253: {'lr': 0.00029470592002887815, 'samples': 11906048, 'steps': 23253, 'loss/train': 1.8171945810317993} 02/25/2022 07:42:16 - INFO - codeparrot_training - Step 23254: {'lr': 0.00029468982121568096, 'samples': 11906560, 'steps': 23254, 'loss/train': 2.15289044380188} 02/25/2022 07:42:20 - INFO - codeparrot_training - Step 23255: {'lr': 0.0002946737222110467, 'samples': 11907072, 'steps': 23255, 'loss/train': 1.3837889432907104} 02/25/2022 07:42:25 - INFO - codeparrot_training - Step 23256: {'lr': 0.0002946576230150444, 'samples': 11907584, 'steps': 23256, 'loss/train': 1.6993913650512695} 02/25/2022 07:42:29 - INFO - codeparrot_training - Step 23257: {'lr': 0.00029464152362774305, 'samples': 11908096, 'steps': 23257, 'loss/train': 1.5446391105651855} 02/25/2022 07:42:34 - INFO - codeparrot_training - Step 23258: {'lr': 0.00029462542404921156, 'samples': 11908608, 'steps': 23258, 'loss/train': 2.648787498474121} 02/25/2022 07:42:38 - INFO - codeparrot_training - Step 23259: {'lr': 0.00029460932427951897, 'samples': 11909120, 'steps': 23259, 'loss/train': 0.8352705836296082} 02/25/2022 07:42:44 - INFO - codeparrot_training - Step 23260: {'lr': 0.00029459322431873416, 'samples': 11909632, 'steps': 23260, 'loss/train': 1.9974157810211182} 02/25/2022 07:42:47 - INFO - codeparrot_training - Step 23261: {'lr': 0.00029457712416692617, 'samples': 11910144, 'steps': 23261, 'loss/train': 1.5597301721572876} 02/25/2022 07:42:53 - INFO - codeparrot_training - Step 23262: {'lr': 0.0002945610238241639, 'samples': 11910656, 'steps': 23262, 'loss/train': 4.257477760314941} 02/25/2022 07:42:56 - INFO - codeparrot_training - Step 23263: {'lr': 0.0002945449232905164, 'samples': 11911168, 'steps': 23263, 'loss/train': 1.993708848953247} 02/25/2022 07:43:02 - INFO - codeparrot_training - Step 23264: {'lr': 0.0002945288225660525, 'samples': 11911680, 'steps': 23264, 'loss/train': 2.6475870609283447} 02/25/2022 07:43:05 - INFO - codeparrot_training - Step 23265: {'lr': 0.00029451272165084144, 'samples': 11912192, 'steps': 23265, 'loss/train': 2.3297805786132812} 02/25/2022 07:43:11 - INFO - codeparrot_training - Step 23266: {'lr': 0.00029449662054495184, 'samples': 11912704, 'steps': 23266, 'loss/train': 2.599456787109375} 02/25/2022 07:43:14 - INFO - codeparrot_training - Step 23267: {'lr': 0.0002944805192484529, 'samples': 11913216, 'steps': 23267, 'loss/train': 1.3924566507339478} 02/25/2022 07:43:20 - INFO - codeparrot_training - Step 23268: {'lr': 0.0002944644177614136, 'samples': 11913728, 'steps': 23268, 'loss/train': 1.5712859630584717} 02/25/2022 07:43:24 - INFO - codeparrot_training - Step 23269: {'lr': 0.00029444831608390276, 'samples': 11914240, 'steps': 23269, 'loss/train': 2.215167284011841} 02/25/2022 07:43:29 - INFO - codeparrot_training - Step 23270: {'lr': 0.0002944322142159895, 'samples': 11914752, 'steps': 23270, 'loss/train': 2.082902431488037} 02/25/2022 07:43:33 - INFO - codeparrot_training - Step 23271: {'lr': 0.0002944161121577427, 'samples': 11915264, 'steps': 23271, 'loss/train': 1.5175353288650513} 02/25/2022 07:43:38 - INFO - codeparrot_training - Step 23272: {'lr': 0.0002944000099092314, 'samples': 11915776, 'steps': 23272, 'loss/train': 1.6339221000671387} 02/25/2022 07:43:42 - INFO - codeparrot_training - Step 23273: {'lr': 0.0002943839074705246, 'samples': 11916288, 'steps': 23273, 'loss/train': 1.7660932540893555} 02/25/2022 07:43:47 - INFO - codeparrot_training - Step 23274: {'lr': 0.00029436780484169125, 'samples': 11916800, 'steps': 23274, 'loss/train': 1.6506413221359253} 02/25/2022 07:43:51 - INFO - codeparrot_training - Step 23275: {'lr': 0.0002943517020228003, 'samples': 11917312, 'steps': 23275, 'loss/train': 1.7423412799835205} 02/25/2022 07:43:58 - INFO - codeparrot_training - Step 23276: {'lr': 0.00029433559901392067, 'samples': 11917824, 'steps': 23276, 'loss/train': 1.5406180620193481} 02/25/2022 07:44:02 - INFO - codeparrot_training - Step 23277: {'lr': 0.0002943194958151214, 'samples': 11918336, 'steps': 23277, 'loss/train': 1.117648720741272} 02/25/2022 07:44:07 - INFO - codeparrot_training - Step 23278: {'lr': 0.00029430339242647157, 'samples': 11918848, 'steps': 23278, 'loss/train': 2.628079414367676} 02/25/2022 07:44:11 - INFO - codeparrot_training - Step 23279: {'lr': 0.00029428728884804, 'samples': 11919360, 'steps': 23279, 'loss/train': 2.8892574310302734} 02/25/2022 07:44:16 - INFO - codeparrot_training - Step 23280: {'lr': 0.0002942711850798959, 'samples': 11919872, 'steps': 23280, 'loss/train': 1.594959020614624} 02/25/2022 07:44:20 - INFO - codeparrot_training - Step 23281: {'lr': 0.00029425508112210794, 'samples': 11920384, 'steps': 23281, 'loss/train': 1.6724756956100464} 02/25/2022 07:44:25 - INFO - codeparrot_training - Step 23282: {'lr': 0.00029423897697474535, 'samples': 11920896, 'steps': 23282, 'loss/train': 1.522385597229004} 02/25/2022 07:44:29 - INFO - codeparrot_training - Step 23283: {'lr': 0.00029422287263787696, 'samples': 11921408, 'steps': 23283, 'loss/train': 3.1397314071655273} 02/25/2022 07:44:34 - INFO - codeparrot_training - Step 23284: {'lr': 0.0002942067681115719, 'samples': 11921920, 'steps': 23284, 'loss/train': 1.8952866792678833} 02/25/2022 07:44:38 - INFO - codeparrot_training - Step 23285: {'lr': 0.0002941906633958989, 'samples': 11922432, 'steps': 23285, 'loss/train': 2.027068853378296} 02/25/2022 07:44:45 - INFO - codeparrot_training - Step 23286: {'lr': 0.0002941745584909274, 'samples': 11922944, 'steps': 23286, 'loss/train': 2.7364048957824707} 02/25/2022 07:44:49 - INFO - codeparrot_training - Step 23287: {'lr': 0.00029415845339672596, 'samples': 11923456, 'steps': 23287, 'loss/train': 1.1098403930664062} 02/25/2022 07:44:54 - INFO - codeparrot_training - Step 23288: {'lr': 0.00029414234811336377, 'samples': 11923968, 'steps': 23288, 'loss/train': 1.479647159576416} 02/25/2022 07:44:57 - INFO - codeparrot_training - Step 23289: {'lr': 0.00029412624264090983, 'samples': 11924480, 'steps': 23289, 'loss/train': 0.11885195225477219} 02/25/2022 07:45:03 - INFO - codeparrot_training - Step 23290: {'lr': 0.00029411013697943294, 'samples': 11924992, 'steps': 23290, 'loss/train': 1.6844847202301025} 02/25/2022 07:45:07 - INFO - codeparrot_training - Step 23291: {'lr': 0.00029409403112900227, 'samples': 11925504, 'steps': 23291, 'loss/train': 0.6062667369842529} 02/25/2022 07:45:12 - INFO - codeparrot_training - Step 23292: {'lr': 0.00029407792508968683, 'samples': 11926016, 'steps': 23292, 'loss/train': 1.4477728605270386} 02/25/2022 07:45:16 - INFO - codeparrot_training - Step 23293: {'lr': 0.0002940618188615555, 'samples': 11926528, 'steps': 23293, 'loss/train': 0.6699808239936829} 02/25/2022 07:45:21 - INFO - codeparrot_training - Step 23294: {'lr': 0.0002940457124446774, 'samples': 11927040, 'steps': 23294, 'loss/train': 2.055760145187378} 02/25/2022 07:45:25 - INFO - codeparrot_training - Step 23295: {'lr': 0.0002940296058391214, 'samples': 11927552, 'steps': 23295, 'loss/train': 1.6895854473114014} 02/25/2022 07:45:32 - INFO - codeparrot_training - Step 23296: {'lr': 0.00029401349904495654, 'samples': 11928064, 'steps': 23296, 'loss/train': 0.45486074686050415} 02/25/2022 07:45:35 - INFO - codeparrot_training - Step 23297: {'lr': 0.00029399739206225186, 'samples': 11928576, 'steps': 23297, 'loss/train': 1.7504163980484009} 02/25/2022 07:45:41 - INFO - codeparrot_training - Step 23298: {'lr': 0.00029398128489107635, 'samples': 11929088, 'steps': 23298, 'loss/train': 1.576931118965149} 02/25/2022 07:45:44 - INFO - codeparrot_training - Step 23299: {'lr': 0.0002939651775314989, 'samples': 11929600, 'steps': 23299, 'loss/train': 2.00386905670166} 02/25/2022 07:45:50 - INFO - codeparrot_training - Step 23300: {'lr': 0.0002939490699835887, 'samples': 11930112, 'steps': 23300, 'loss/train': 1.555681586265564} 02/25/2022 07:45:53 - INFO - codeparrot_training - Step 23301: {'lr': 0.0002939329622474145, 'samples': 11930624, 'steps': 23301, 'loss/train': 2.8473918437957764} 02/25/2022 07:46:00 - INFO - codeparrot_training - Step 23302: {'lr': 0.00029391685432304554, 'samples': 11931136, 'steps': 23302, 'loss/train': 1.720230221748352} 02/25/2022 07:46:03 - INFO - codeparrot_training - Step 23303: {'lr': 0.0002939007462105507, 'samples': 11931648, 'steps': 23303, 'loss/train': 2.24920654296875} 02/25/2022 07:46:09 - INFO - codeparrot_training - Step 23304: {'lr': 0.000293884637909999, 'samples': 11932160, 'steps': 23304, 'loss/train': 1.97749924659729} 02/25/2022 07:46:12 - INFO - codeparrot_training - Step 23305: {'lr': 0.00029386852942145944, 'samples': 11932672, 'steps': 23305, 'loss/train': 1.7793421745300293} 02/25/2022 07:46:19 - INFO - codeparrot_training - Step 23306: {'lr': 0.000293852420745001, 'samples': 11933184, 'steps': 23306, 'loss/train': 2.5778260231018066} 02/25/2022 07:46:23 - INFO - codeparrot_training - Step 23307: {'lr': 0.0002938363118806928, 'samples': 11933696, 'steps': 23307, 'loss/train': 1.3569884300231934} 02/25/2022 07:46:28 - INFO - codeparrot_training - Step 23308: {'lr': 0.00029382020282860366, 'samples': 11934208, 'steps': 23308, 'loss/train': 2.33257794380188} 02/25/2022 07:46:32 - INFO - codeparrot_training - Step 23309: {'lr': 0.00029380409358880277, 'samples': 11934720, 'steps': 23309, 'loss/train': 2.3145346641540527} 02/25/2022 07:46:37 - INFO - codeparrot_training - Step 23310: {'lr': 0.000293787984161359, 'samples': 11935232, 'steps': 23310, 'loss/train': 1.8709235191345215} 02/25/2022 07:46:41 - INFO - codeparrot_training - Step 23311: {'lr': 0.0002937718745463414, 'samples': 11935744, 'steps': 23311, 'loss/train': 1.5701366662979126} 02/25/2022 07:46:46 - INFO - codeparrot_training - Step 23312: {'lr': 0.00029375576474381903, 'samples': 11936256, 'steps': 23312, 'loss/train': 2.1795248985290527} 02/25/2022 07:46:50 - INFO - codeparrot_training - Step 23313: {'lr': 0.0002937396547538609, 'samples': 11936768, 'steps': 23313, 'loss/train': 2.784252643585205} 02/25/2022 07:46:55 - INFO - codeparrot_training - Step 23314: {'lr': 0.00029372354457653585, 'samples': 11937280, 'steps': 23314, 'loss/train': 2.8848493099212646} 02/25/2022 07:46:59 - INFO - codeparrot_training - Step 23315: {'lr': 0.00029370743421191305, 'samples': 11937792, 'steps': 23315, 'loss/train': 1.4853767156600952} 02/25/2022 07:47:04 - INFO - codeparrot_training - Step 23316: {'lr': 0.0002936913236600616, 'samples': 11938304, 'steps': 23316, 'loss/train': 1.9631294012069702} 02/25/2022 07:47:08 - INFO - codeparrot_training - Step 23317: {'lr': 0.0002936752129210503, 'samples': 11938816, 'steps': 23317, 'loss/train': 2.023798704147339} 02/25/2022 07:47:13 - INFO - codeparrot_training - Step 23318: {'lr': 0.00029365910199494823, 'samples': 11939328, 'steps': 23318, 'loss/train': 2.9043664932250977} 02/25/2022 07:47:17 - INFO - codeparrot_training - Step 23319: {'lr': 0.00029364299088182445, 'samples': 11939840, 'steps': 23319, 'loss/train': 0.6101219058036804} 02/25/2022 07:47:22 - INFO - codeparrot_training - Step 23320: {'lr': 0.000293626879581748, 'samples': 11940352, 'steps': 23320, 'loss/train': 1.6024255752563477} 02/25/2022 07:47:26 - INFO - codeparrot_training - Step 23321: {'lr': 0.00029361076809478784, 'samples': 11940864, 'steps': 23321, 'loss/train': 1.73313307762146} 02/25/2022 07:47:33 - INFO - codeparrot_training - Step 23322: {'lr': 0.00029359465642101297, 'samples': 11941376, 'steps': 23322, 'loss/train': 1.649749517440796} 02/25/2022 07:47:37 - INFO - codeparrot_training - Step 23323: {'lr': 0.0002935785445604924, 'samples': 11941888, 'steps': 23323, 'loss/train': 2.128927230834961} 02/25/2022 07:47:42 - INFO - codeparrot_training - Step 23324: {'lr': 0.0002935624325132953, 'samples': 11942400, 'steps': 23324, 'loss/train': 1.4314079284667969} 02/25/2022 07:47:46 - INFO - codeparrot_training - Step 23325: {'lr': 0.0002935463202794905, 'samples': 11942912, 'steps': 23325, 'loss/train': 2.885136365890503} 02/25/2022 07:47:52 - INFO - codeparrot_training - Step 23326: {'lr': 0.00029353020785914707, 'samples': 11943424, 'steps': 23326, 'loss/train': 1.5936180353164673} 02/25/2022 07:47:55 - INFO - codeparrot_training - Step 23327: {'lr': 0.00029351409525233416, 'samples': 11943936, 'steps': 23327, 'loss/train': 1.7138837575912476} 02/25/2022 07:48:00 - INFO - codeparrot_training - Step 23328: {'lr': 0.0002934979824591205, 'samples': 11944448, 'steps': 23328, 'loss/train': 0.6495726704597473} 02/25/2022 07:48:04 - INFO - codeparrot_training - Step 23329: {'lr': 0.0002934818694795755, 'samples': 11944960, 'steps': 23329, 'loss/train': 1.9211796522140503} 02/25/2022 07:48:10 - INFO - codeparrot_training - Step 23330: {'lr': 0.0002934657563137678, 'samples': 11945472, 'steps': 23330, 'loss/train': 1.0002098083496094} 02/25/2022 07:48:13 - INFO - codeparrot_training - Step 23331: {'lr': 0.00029344964296176666, 'samples': 11945984, 'steps': 23331, 'loss/train': 0.7236199378967285} 02/25/2022 07:48:20 - INFO - codeparrot_training - Step 23332: {'lr': 0.00029343352942364106, 'samples': 11946496, 'steps': 23332, 'loss/train': 2.111624240875244} 02/25/2022 07:48:24 - INFO - codeparrot_training - Step 23333: {'lr': 0.00029341741569946007, 'samples': 11947008, 'steps': 23333, 'loss/train': 1.6773176193237305} 02/25/2022 07:48:29 - INFO - codeparrot_training - Step 23334: {'lr': 0.00029340130178929255, 'samples': 11947520, 'steps': 23334, 'loss/train': 2.680154323577881} 02/25/2022 07:48:33 - INFO - codeparrot_training - Step 23335: {'lr': 0.0002933851876932077, 'samples': 11948032, 'steps': 23335, 'loss/train': 2.562551975250244} 02/25/2022 07:48:38 - INFO - codeparrot_training - Step 23336: {'lr': 0.00029336907341127443, 'samples': 11948544, 'steps': 23336, 'loss/train': 1.8527193069458008} 02/25/2022 07:48:42 - INFO - codeparrot_training - Step 23337: {'lr': 0.0002933529589435619, 'samples': 11949056, 'steps': 23337, 'loss/train': 2.284768581390381} 02/25/2022 07:48:45 - INFO - codeparrot_training - Step 23338: {'lr': 0.000293336844290139, 'samples': 11949568, 'steps': 23338, 'loss/train': 1.9783297777175903} 02/25/2022 07:48:51 - INFO - codeparrot_training - Step 23339: {'lr': 0.0002933207294510748, 'samples': 11950080, 'steps': 23339, 'loss/train': 2.1956281661987305} 02/25/2022 07:48:56 - INFO - codeparrot_training - Step 23340: {'lr': 0.0002933046144264384, 'samples': 11950592, 'steps': 23340, 'loss/train': 1.364457368850708} 02/25/2022 07:49:00 - INFO - codeparrot_training - Step 23341: {'lr': 0.0002932884992162987, 'samples': 11951104, 'steps': 23341, 'loss/train': 1.0895816087722778} 02/25/2022 07:49:03 - INFO - codeparrot_training - Step 23342: {'lr': 0.00029327238382072495, 'samples': 11951616, 'steps': 23342, 'loss/train': 1.349489450454712} 02/25/2022 07:49:11 - INFO - codeparrot_training - Step 23343: {'lr': 0.000293256268239786, 'samples': 11952128, 'steps': 23343, 'loss/train': 2.8808159828186035} 02/25/2022 07:49:14 - INFO - codeparrot_training - Step 23344: {'lr': 0.00029324015247355093, 'samples': 11952640, 'steps': 23344, 'loss/train': 2.297421932220459} 02/25/2022 07:49:20 - INFO - codeparrot_training - Step 23345: {'lr': 0.0002932240365220887, 'samples': 11953152, 'steps': 23345, 'loss/train': 1.592259407043457} 02/25/2022 07:49:23 - INFO - codeparrot_training - Step 23346: {'lr': 0.00029320792038546855, 'samples': 11953664, 'steps': 23346, 'loss/train': 2.0885281562805176} 02/25/2022 07:49:29 - INFO - codeparrot_training - Step 23347: {'lr': 0.0002931918040637593, 'samples': 11954176, 'steps': 23347, 'loss/train': 2.349640369415283} 02/25/2022 07:49:34 - INFO - codeparrot_training - Step 23348: {'lr': 0.0002931756875570301, 'samples': 11954688, 'steps': 23348, 'loss/train': 1.9957149028778076} 02/25/2022 07:49:38 - INFO - codeparrot_training - Step 23349: {'lr': 0.0002931595708653499, 'samples': 11955200, 'steps': 23349, 'loss/train': 1.07041597366333} 02/25/2022 07:49:43 - INFO - codeparrot_training - Step 23350: {'lr': 0.00029314345398878796, 'samples': 11955712, 'steps': 23350, 'loss/train': 2.185713291168213} 02/25/2022 07:49:47 - INFO - codeparrot_training - Step 23351: {'lr': 0.00029312733692741307, 'samples': 11956224, 'steps': 23351, 'loss/train': 1.176406979560852} 02/25/2022 07:49:54 - INFO - codeparrot_training - Step 23352: {'lr': 0.00029311121968129435, 'samples': 11956736, 'steps': 23352, 'loss/train': 1.394729495048523} 02/25/2022 07:49:57 - INFO - codeparrot_training - Step 23353: {'lr': 0.00029309510225050087, 'samples': 11957248, 'steps': 23353, 'loss/train': 1.989072322845459} 02/25/2022 07:50:03 - INFO - codeparrot_training - Step 23354: {'lr': 0.00029307898463510164, 'samples': 11957760, 'steps': 23354, 'loss/train': 1.8369868993759155} 02/25/2022 07:50:06 - INFO - codeparrot_training - Step 23355: {'lr': 0.00029306286683516573, 'samples': 11958272, 'steps': 23355, 'loss/train': 1.9826176166534424} 02/25/2022 07:50:12 - INFO - codeparrot_training - Step 23356: {'lr': 0.00029304674885076215, 'samples': 11958784, 'steps': 23356, 'loss/train': 3.8771159648895264} 02/25/2022 07:50:15 - INFO - codeparrot_training - Step 23357: {'lr': 0.00029303063068196006, 'samples': 11959296, 'steps': 23357, 'loss/train': 1.3828481435775757} 02/25/2022 07:50:21 - INFO - codeparrot_training - Step 23358: {'lr': 0.0002930145123288283, 'samples': 11959808, 'steps': 23358, 'loss/train': 1.342498779296875} 02/25/2022 07:50:24 - INFO - codeparrot_training - Step 23359: {'lr': 0.00029299839379143613, 'samples': 11960320, 'steps': 23359, 'loss/train': 2.260441541671753} 02/25/2022 07:50:30 - INFO - codeparrot_training - Step 23360: {'lr': 0.0002929822750698524, 'samples': 11960832, 'steps': 23360, 'loss/train': 2.9931976795196533} 02/25/2022 07:50:33 - INFO - codeparrot_training - Step 23361: {'lr': 0.0002929661561641463, 'samples': 11961344, 'steps': 23361, 'loss/train': 1.3829420804977417} 02/25/2022 07:50:39 - INFO - codeparrot_training - Step 23362: {'lr': 0.0002929500370743868, 'samples': 11961856, 'steps': 23362, 'loss/train': 1.6675212383270264} 02/25/2022 07:50:43 - INFO - codeparrot_training - Step 23363: {'lr': 0.000292933917800643, 'samples': 11962368, 'steps': 23363, 'loss/train': 1.8965108394622803} 02/25/2022 07:50:48 - INFO - codeparrot_training - Step 23364: {'lr': 0.0002929177983429839, 'samples': 11962880, 'steps': 23364, 'loss/train': 2.617595911026001} 02/25/2022 07:50:52 - INFO - codeparrot_training - Step 23365: {'lr': 0.0002929016787014786, 'samples': 11963392, 'steps': 23365, 'loss/train': 1.4088730812072754} 02/25/2022 07:50:57 - INFO - codeparrot_training - Step 23366: {'lr': 0.0002928855588761962, 'samples': 11963904, 'steps': 23366, 'loss/train': 2.2544336318969727} 02/25/2022 07:51:01 - INFO - codeparrot_training - Step 23367: {'lr': 0.0002928694388672056, 'samples': 11964416, 'steps': 23367, 'loss/train': 1.7133792638778687} 02/25/2022 07:51:08 - INFO - codeparrot_training - Step 23368: {'lr': 0.00029285331867457597, 'samples': 11964928, 'steps': 23368, 'loss/train': 1.6071662902832031} 02/25/2022 07:51:11 - INFO - codeparrot_training - Step 23369: {'lr': 0.00029283719829837636, 'samples': 11965440, 'steps': 23369, 'loss/train': 1.1860008239746094} 02/25/2022 07:51:17 - INFO - codeparrot_training - Step 23370: {'lr': 0.0002928210777386757, 'samples': 11965952, 'steps': 23370, 'loss/train': 1.4075809717178345} 02/25/2022 07:51:21 - INFO - codeparrot_training - Step 23371: {'lr': 0.0002928049569955432, 'samples': 11966464, 'steps': 23371, 'loss/train': 2.2653560638427734} 02/25/2022 07:51:24 - INFO - codeparrot_training - Step 23372: {'lr': 0.0002927888360690478, 'samples': 11966976, 'steps': 23372, 'loss/train': 1.7175580263137817} 02/25/2022 07:51:30 - INFO - codeparrot_training - Step 23373: {'lr': 0.00029277271495925874, 'samples': 11967488, 'steps': 23373, 'loss/train': 1.8998998403549194} 02/25/2022 07:51:33 - INFO - codeparrot_training - Step 23374: {'lr': 0.00029275659366624493, 'samples': 11968000, 'steps': 23374, 'loss/train': 2.1208808422088623} 02/25/2022 07:51:39 - INFO - codeparrot_training - Step 23375: {'lr': 0.00029274047219007534, 'samples': 11968512, 'steps': 23375, 'loss/train': 2.0657408237457275} 02/25/2022 07:51:42 - INFO - codeparrot_training - Step 23376: {'lr': 0.00029272435053081917, 'samples': 11969024, 'steps': 23376, 'loss/train': 1.7744271755218506} 02/25/2022 07:51:48 - INFO - codeparrot_training - Step 23377: {'lr': 0.0002927082286885455, 'samples': 11969536, 'steps': 23377, 'loss/train': 2.682598352432251} 02/25/2022 07:51:51 - INFO - codeparrot_training - Step 23378: {'lr': 0.0002926921066633233, 'samples': 11970048, 'steps': 23378, 'loss/train': 2.908313751220703} 02/25/2022 07:52:00 - INFO - codeparrot_training - Step 23379: {'lr': 0.0002926759844552217, 'samples': 11970560, 'steps': 23379, 'loss/train': 2.297487735748291} 02/25/2022 07:52:04 - INFO - codeparrot_training - Step 23380: {'lr': 0.0002926598620643097, 'samples': 11971072, 'steps': 23380, 'loss/train': 2.234861373901367} 02/25/2022 07:52:09 - INFO - codeparrot_training - Step 23381: {'lr': 0.0002926437394906564, 'samples': 11971584, 'steps': 23381, 'loss/train': 2.625631809234619} 02/25/2022 07:52:12 - INFO - codeparrot_training - Step 23382: {'lr': 0.00029262761673433087, 'samples': 11972096, 'steps': 23382, 'loss/train': 2.7523982524871826} 02/25/2022 07:52:18 - INFO - codeparrot_training - Step 23383: {'lr': 0.0002926114937954022, 'samples': 11972608, 'steps': 23383, 'loss/train': 1.088431477546692} 02/25/2022 07:52:21 - INFO - codeparrot_training - Step 23384: {'lr': 0.0002925953706739394, 'samples': 11973120, 'steps': 23384, 'loss/train': 1.984533667564392} 02/25/2022 07:52:27 - INFO - codeparrot_training - Step 23385: {'lr': 0.0002925792473700116, 'samples': 11973632, 'steps': 23385, 'loss/train': 1.199676275253296} 02/25/2022 07:52:30 - INFO - codeparrot_training - Step 23386: {'lr': 0.00029256312388368773, 'samples': 11974144, 'steps': 23386, 'loss/train': 2.3697826862335205} 02/25/2022 07:52:36 - INFO - codeparrot_training - Step 23387: {'lr': 0.000292547000215037, 'samples': 11974656, 'steps': 23387, 'loss/train': 3.095698118209839} 02/25/2022 07:52:39 - INFO - codeparrot_training - Step 23388: {'lr': 0.0002925308763641284, 'samples': 11975168, 'steps': 23388, 'loss/train': 2.0039196014404297} 02/25/2022 07:52:45 - INFO - codeparrot_training - Step 23389: {'lr': 0.00029251475233103104, 'samples': 11975680, 'steps': 23389, 'loss/train': 2.377756357192993} 02/25/2022 07:52:48 - INFO - codeparrot_training - Step 23390: {'lr': 0.000292498628115814, 'samples': 11976192, 'steps': 23390, 'loss/train': 2.5976409912109375} 02/25/2022 07:52:56 - INFO - codeparrot_training - Step 23391: {'lr': 0.00029248250371854636, 'samples': 11976704, 'steps': 23391, 'loss/train': 2.7413089275360107} 02/25/2022 07:53:01 - INFO - codeparrot_training - Step 23392: {'lr': 0.0002924663791392971, 'samples': 11977216, 'steps': 23392, 'loss/train': 1.9561271667480469} 02/25/2022 07:53:05 - INFO - codeparrot_training - Step 23393: {'lr': 0.0002924502543781354, 'samples': 11977728, 'steps': 23393, 'loss/train': 1.8099284172058105} 02/25/2022 07:53:10 - INFO - codeparrot_training - Step 23394: {'lr': 0.0002924341294351303, 'samples': 11978240, 'steps': 23394, 'loss/train': 1.7902275323867798} 02/25/2022 07:53:14 - INFO - codeparrot_training - Step 23395: {'lr': 0.0002924180043103508, 'samples': 11978752, 'steps': 23395, 'loss/train': 2.017207622528076} 02/25/2022 07:53:19 - INFO - codeparrot_training - Step 23396: {'lr': 0.0002924018790038662, 'samples': 11979264, 'steps': 23396, 'loss/train': 2.2842447757720947} 02/25/2022 07:53:23 - INFO - codeparrot_training - Step 23397: {'lr': 0.0002923857535157452, 'samples': 11979776, 'steps': 23397, 'loss/train': 1.391904592514038} 02/25/2022 07:53:26 - INFO - codeparrot_training - Step 23398: {'lr': 0.00029236962784605727, 'samples': 11980288, 'steps': 23398, 'loss/train': 1.9955264329910278} 02/25/2022 07:53:32 - INFO - codeparrot_training - Step 23399: {'lr': 0.00029235350199487115, 'samples': 11980800, 'steps': 23399, 'loss/train': 2.3995463848114014} 02/25/2022 07:53:39 - INFO - codeparrot_training - Step 23400: {'lr': 0.0002923373759622561, 'samples': 11981312, 'steps': 23400, 'loss/train': 1.9457592964172363} 02/25/2022 07:53:43 - INFO - codeparrot_training - Step 23401: {'lr': 0.00029232124974828126, 'samples': 11981824, 'steps': 23401, 'loss/train': 2.4070663452148438} 02/25/2022 07:53:46 - INFO - codeparrot_training - Step 23402: {'lr': 0.0002923051233530156, 'samples': 11982336, 'steps': 23402, 'loss/train': 1.2458875179290771} 02/25/2022 07:53:52 - INFO - codeparrot_training - Step 23403: {'lr': 0.00029228899677652826, 'samples': 11982848, 'steps': 23403, 'loss/train': 2.039367914199829} 02/25/2022 07:53:55 - INFO - codeparrot_training - Step 23404: {'lr': 0.0002922728700188882, 'samples': 11983360, 'steps': 23404, 'loss/train': 2.762005090713501} 02/25/2022 07:54:01 - INFO - codeparrot_training - Step 23405: {'lr': 0.00029225674308016464, 'samples': 11983872, 'steps': 23405, 'loss/train': 1.8602744340896606} 02/25/2022 07:54:06 - INFO - codeparrot_training - Step 23406: {'lr': 0.00029224061596042663, 'samples': 11984384, 'steps': 23406, 'loss/train': 1.7666420936584473} 02/25/2022 07:54:10 - INFO - codeparrot_training - Step 23407: {'lr': 0.0002922244886597432, 'samples': 11984896, 'steps': 23407, 'loss/train': 2.2174315452575684} 02/25/2022 07:54:15 - INFO - codeparrot_training - Step 23408: {'lr': 0.00029220836117818346, 'samples': 11985408, 'steps': 23408, 'loss/train': 2.378347635269165} 02/25/2022 07:54:19 - INFO - codeparrot_training - Step 23409: {'lr': 0.00029219223351581653, 'samples': 11985920, 'steps': 23409, 'loss/train': 2.434917449951172} 02/25/2022 07:54:24 - INFO - codeparrot_training - Step 23410: {'lr': 0.00029217610567271147, 'samples': 11986432, 'steps': 23410, 'loss/train': 5.452525615692139} 02/25/2022 07:54:28 - INFO - codeparrot_training - Step 23411: {'lr': 0.00029215997764893734, 'samples': 11986944, 'steps': 23411, 'loss/train': 0.46564772725105286} 02/25/2022 07:54:33 - INFO - codeparrot_training - Step 23412: {'lr': 0.00029214384944456327, 'samples': 11987456, 'steps': 23412, 'loss/train': 1.8892968893051147} 02/25/2022 07:54:37 - INFO - codeparrot_training - Step 23413: {'lr': 0.00029212772105965837, 'samples': 11987968, 'steps': 23413, 'loss/train': 2.130406379699707} 02/25/2022 07:54:44 - INFO - codeparrot_training - Step 23414: {'lr': 0.0002921115924942916, 'samples': 11988480, 'steps': 23414, 'loss/train': 0.36277735233306885} 02/25/2022 07:54:48 - INFO - codeparrot_training - Step 23415: {'lr': 0.0002920954637485322, 'samples': 11988992, 'steps': 23415, 'loss/train': 1.677271842956543} 02/25/2022 07:54:53 - INFO - codeparrot_training - Step 23416: {'lr': 0.0002920793348224493, 'samples': 11989504, 'steps': 23416, 'loss/train': 0.4123888611793518} 02/25/2022 07:54:56 - INFO - codeparrot_training - Step 23417: {'lr': 0.0002920632057161117, 'samples': 11990016, 'steps': 23417, 'loss/train': 1.8849787712097168} 02/25/2022 07:55:02 - INFO - codeparrot_training - Step 23418: {'lr': 0.00029204707642958876, 'samples': 11990528, 'steps': 23418, 'loss/train': 1.4175535440444946} 02/25/2022 07:55:06 - INFO - codeparrot_training - Step 23419: {'lr': 0.0002920309469629495, 'samples': 11991040, 'steps': 23419, 'loss/train': 0.1740378588438034} 02/25/2022 07:55:11 - INFO - codeparrot_training - Step 23420: {'lr': 0.000292014817316263, 'samples': 11991552, 'steps': 23420, 'loss/train': 1.7785595655441284} 02/25/2022 07:55:15 - INFO - codeparrot_training - Step 23421: {'lr': 0.00029199868748959836, 'samples': 11992064, 'steps': 23421, 'loss/train': 1.7453330755233765} 02/25/2022 07:55:20 - INFO - codeparrot_training - Step 23422: {'lr': 0.00029198255748302473, 'samples': 11992576, 'steps': 23422, 'loss/train': 1.1835572719573975} 02/25/2022 07:55:24 - INFO - codeparrot_training - Step 23423: {'lr': 0.00029196642729661113, 'samples': 11993088, 'steps': 23423, 'loss/train': 2.376190662384033} 02/25/2022 07:55:29 - INFO - codeparrot_training - Step 23424: {'lr': 0.0002919502969304266, 'samples': 11993600, 'steps': 23424, 'loss/train': 1.0783319473266602} 02/25/2022 07:55:33 - INFO - codeparrot_training - Step 23425: {'lr': 0.0002919341663845404, 'samples': 11994112, 'steps': 23425, 'loss/train': 2.0160794258117676} 02/25/2022 07:55:40 - INFO - codeparrot_training - Step 23426: {'lr': 0.00029191803565902153, 'samples': 11994624, 'steps': 23426, 'loss/train': 2.896087646484375} 02/25/2022 07:55:43 - INFO - codeparrot_training - Step 23427: {'lr': 0.00029190190475393915, 'samples': 11995136, 'steps': 23427, 'loss/train': 1.8752520084381104} 02/25/2022 07:55:49 - INFO - codeparrot_training - Step 23428: {'lr': 0.0002918857736693622, 'samples': 11995648, 'steps': 23428, 'loss/train': 1.5570167303085327} 02/25/2022 07:55:52 - INFO - codeparrot_training - Step 23429: {'lr': 0.00029186964240536, 'samples': 11996160, 'steps': 23429, 'loss/train': 1.6154569387435913} 02/25/2022 07:55:58 - INFO - codeparrot_training - Step 23430: {'lr': 0.0002918535109620015, 'samples': 11996672, 'steps': 23430, 'loss/train': 2.3734962940216064} 02/25/2022 07:56:01 - INFO - codeparrot_training - Step 23431: {'lr': 0.0002918373793393559, 'samples': 11997184, 'steps': 23431, 'loss/train': 2.393853187561035} 02/25/2022 07:56:07 - INFO - codeparrot_training - Step 23432: {'lr': 0.0002918212475374922, 'samples': 11997696, 'steps': 23432, 'loss/train': 1.748063087463379} 02/25/2022 07:56:10 - INFO - codeparrot_training - Step 23433: {'lr': 0.0002918051155564796, 'samples': 11998208, 'steps': 23433, 'loss/train': 1.5458824634552002} 02/25/2022 07:56:16 - INFO - codeparrot_training - Step 23434: {'lr': 0.0002917889833963871, 'samples': 11998720, 'steps': 23434, 'loss/train': 2.2787227630615234} 02/25/2022 07:56:19 - INFO - codeparrot_training - Step 23435: {'lr': 0.00029177285105728393, 'samples': 11999232, 'steps': 23435, 'loss/train': 2.189514636993408} 02/25/2022 07:56:26 - INFO - codeparrot_training - Step 23436: {'lr': 0.00029175671853923907, 'samples': 11999744, 'steps': 23436, 'loss/train': 1.7956748008728027} 02/25/2022 07:56:30 - INFO - codeparrot_training - Step 23437: {'lr': 0.00029174058584232167, 'samples': 12000256, 'steps': 23437, 'loss/train': 1.8041551113128662} 02/25/2022 07:56:35 - INFO - codeparrot_training - Step 23438: {'lr': 0.00029172445296660096, 'samples': 12000768, 'steps': 23438, 'loss/train': 2.3593952655792236} 02/25/2022 07:56:39 - INFO - codeparrot_training - Step 23439: {'lr': 0.0002917083199121459, 'samples': 12001280, 'steps': 23439, 'loss/train': 0.9599772095680237} 02/25/2022 07:56:44 - INFO - codeparrot_training - Step 23440: {'lr': 0.00029169218667902556, 'samples': 12001792, 'steps': 23440, 'loss/train': 0.8404916524887085} 02/25/2022 07:56:48 - INFO - codeparrot_training - Step 23441: {'lr': 0.0002916760532673092, 'samples': 12002304, 'steps': 23441, 'loss/train': 2.7631075382232666} 02/25/2022 07:56:53 - INFO - codeparrot_training - Step 23442: {'lr': 0.0002916599196770659, 'samples': 12002816, 'steps': 23442, 'loss/train': 2.1840200424194336} 02/25/2022 07:56:57 - INFO - codeparrot_training - Step 23443: {'lr': 0.00029164378590836466, 'samples': 12003328, 'steps': 23443, 'loss/train': 1.6124993562698364} 02/25/2022 07:57:02 - INFO - codeparrot_training - Step 23444: {'lr': 0.0002916276519612747, 'samples': 12003840, 'steps': 23444, 'loss/train': 0.8980755805969238} 02/25/2022 07:57:06 - INFO - codeparrot_training - Step 23445: {'lr': 0.00029161151783586506, 'samples': 12004352, 'steps': 23445, 'loss/train': 1.6769890785217285} 02/25/2022 07:57:12 - INFO - codeparrot_training - Step 23446: {'lr': 0.0002915953835322049, 'samples': 12004864, 'steps': 23446, 'loss/train': 1.6724718809127808} 02/25/2022 07:57:15 - INFO - codeparrot_training - Step 23447: {'lr': 0.0002915792490503634, 'samples': 12005376, 'steps': 23447, 'loss/train': 2.1814796924591064} 02/25/2022 07:57:21 - INFO - codeparrot_training - Step 23448: {'lr': 0.0002915631143904095, 'samples': 12005888, 'steps': 23448, 'loss/train': 0.5531330704689026} 02/25/2022 07:57:24 - INFO - codeparrot_training - Step 23449: {'lr': 0.0002915469795524125, 'samples': 12006400, 'steps': 23449, 'loss/train': 2.236222505569458} 02/25/2022 07:57:30 - INFO - codeparrot_training - Step 23450: {'lr': 0.00029153084453644135, 'samples': 12006912, 'steps': 23450, 'loss/train': 1.8865327835083008} 02/25/2022 07:57:33 - INFO - codeparrot_training - Step 23451: {'lr': 0.0002915147093425653, 'samples': 12007424, 'steps': 23451, 'loss/train': 1.2760186195373535} 02/25/2022 07:57:39 - INFO - codeparrot_training - Step 23452: {'lr': 0.0002914985739708534, 'samples': 12007936, 'steps': 23452, 'loss/train': 1.7267462015151978} 02/25/2022 07:57:42 - INFO - codeparrot_training - Step 23453: {'lr': 0.00029148243842137486, 'samples': 12008448, 'steps': 23453, 'loss/train': 1.568948745727539} 02/25/2022 07:57:48 - INFO - codeparrot_training - Step 23454: {'lr': 0.0002914663026941986, 'samples': 12008960, 'steps': 23454, 'loss/train': 2.4859275817871094} 02/25/2022 07:57:51 - INFO - codeparrot_training - Step 23455: {'lr': 0.00029145016678939394, 'samples': 12009472, 'steps': 23455, 'loss/train': 2.7000420093536377} 02/25/2022 07:57:57 - INFO - codeparrot_training - Step 23456: {'lr': 0.00029143403070702994, 'samples': 12009984, 'steps': 23456, 'loss/train': 1.298871636390686} 02/25/2022 07:58:01 - INFO - codeparrot_training - Step 23457: {'lr': 0.0002914178944471757, 'samples': 12010496, 'steps': 23457, 'loss/train': 1.8028844594955444} 02/25/2022 07:58:06 - INFO - codeparrot_training - Step 23458: {'lr': 0.00029140175800990034, 'samples': 12011008, 'steps': 23458, 'loss/train': 2.457233428955078} 02/25/2022 07:58:10 - INFO - codeparrot_training - Step 23459: {'lr': 0.000291385621395273, 'samples': 12011520, 'steps': 23459, 'loss/train': 1.549008846282959} 02/25/2022 07:58:15 - INFO - codeparrot_training - Step 23460: {'lr': 0.0002913694846033628, 'samples': 12012032, 'steps': 23460, 'loss/train': 0.4102102518081665} 02/25/2022 07:58:19 - INFO - codeparrot_training - Step 23461: {'lr': 0.00029135334763423883, 'samples': 12012544, 'steps': 23461, 'loss/train': 1.9378471374511719} 02/25/2022 07:58:25 - INFO - codeparrot_training - Step 23462: {'lr': 0.0002913372104879703, 'samples': 12013056, 'steps': 23462, 'loss/train': 2.656374454498291} 02/25/2022 07:58:28 - INFO - codeparrot_training - Step 23463: {'lr': 0.00029132107316462625, 'samples': 12013568, 'steps': 23463, 'loss/train': 1.1116570234298706} 02/25/2022 07:58:34 - INFO - codeparrot_training - Step 23464: {'lr': 0.00029130493566427586, 'samples': 12014080, 'steps': 23464, 'loss/train': 2.106980562210083} 02/25/2022 07:58:37 - INFO - codeparrot_training - Step 23465: {'lr': 0.0002912887979869882, 'samples': 12014592, 'steps': 23465, 'loss/train': 1.8519765138626099} 02/25/2022 07:58:43 - INFO - codeparrot_training - Step 23466: {'lr': 0.00029127266013283254, 'samples': 12015104, 'steps': 23466, 'loss/train': 2.661858320236206} 02/25/2022 07:58:47 - INFO - codeparrot_training - Step 23467: {'lr': 0.0002912565221018778, 'samples': 12015616, 'steps': 23467, 'loss/train': 1.2358543872833252} 02/25/2022 07:58:52 - INFO - codeparrot_training - Step 23468: {'lr': 0.00029124038389419325, 'samples': 12016128, 'steps': 23468, 'loss/train': 1.6122462749481201} 02/25/2022 07:58:56 - INFO - codeparrot_training - Step 23469: {'lr': 0.00029122424550984805, 'samples': 12016640, 'steps': 23469, 'loss/train': 1.77985680103302} 02/25/2022 07:59:01 - INFO - codeparrot_training - Step 23470: {'lr': 0.00029120810694891126, 'samples': 12017152, 'steps': 23470, 'loss/train': 1.2456860542297363} 02/25/2022 07:59:05 - INFO - codeparrot_training - Step 23471: {'lr': 0.00029119196821145193, 'samples': 12017664, 'steps': 23471, 'loss/train': 1.7298152446746826} 02/25/2022 07:59:11 - INFO - codeparrot_training - Step 23472: {'lr': 0.00029117582929753935, 'samples': 12018176, 'steps': 23472, 'loss/train': 2.3893356323242188} 02/25/2022 07:59:14 - INFO - codeparrot_training - Step 23473: {'lr': 0.00029115969020724263, 'samples': 12018688, 'steps': 23473, 'loss/train': 1.2058300971984863} 02/25/2022 07:59:20 - INFO - codeparrot_training - Step 23474: {'lr': 0.0002911435509406308, 'samples': 12019200, 'steps': 23474, 'loss/train': 2.3196771144866943} 02/25/2022 07:59:23 - INFO - codeparrot_training - Step 23475: {'lr': 0.0002911274114977731, 'samples': 12019712, 'steps': 23475, 'loss/train': 2.146773338317871} 02/25/2022 07:59:29 - INFO - codeparrot_training - Step 23476: {'lr': 0.0002911112718787386, 'samples': 12020224, 'steps': 23476, 'loss/train': 1.6535544395446777} 02/25/2022 07:59:32 - INFO - codeparrot_training - Step 23477: {'lr': 0.0002910951320835965, 'samples': 12020736, 'steps': 23477, 'loss/train': 2.479684829711914} 02/25/2022 07:59:38 - INFO - codeparrot_training - Step 23478: {'lr': 0.0002910789921124159, 'samples': 12021248, 'steps': 23478, 'loss/train': 1.8687198162078857} 02/25/2022 07:59:43 - INFO - codeparrot_training - Step 23479: {'lr': 0.00029106285196526594, 'samples': 12021760, 'steps': 23479, 'loss/train': 2.0104100704193115} 02/25/2022 07:59:47 - INFO - codeparrot_training - Step 23480: {'lr': 0.00029104671164221574, 'samples': 12022272, 'steps': 23480, 'loss/train': 2.8713185787200928} 02/25/2022 07:59:53 - INFO - codeparrot_training - Step 23481: {'lr': 0.0002910305711433345, 'samples': 12022784, 'steps': 23481, 'loss/train': 2.610586166381836} 02/25/2022 07:59:57 - INFO - codeparrot_training - Step 23482: {'lr': 0.00029101443046869127, 'samples': 12023296, 'steps': 23482, 'loss/train': 2.129425287246704} 02/25/2022 08:00:02 - INFO - codeparrot_training - Step 23483: {'lr': 0.0002909982896183553, 'samples': 12023808, 'steps': 23483, 'loss/train': 0.7625100612640381} 02/25/2022 08:00:06 - INFO - codeparrot_training - Step 23484: {'lr': 0.0002909821485923956, 'samples': 12024320, 'steps': 23484, 'loss/train': 2.345798969268799} 02/25/2022 08:00:11 - INFO - codeparrot_training - Step 23485: {'lr': 0.0002909660073908814, 'samples': 12024832, 'steps': 23485, 'loss/train': 2.2357516288757324} 02/25/2022 08:00:15 - INFO - codeparrot_training - Step 23486: {'lr': 0.00029094986601388195, 'samples': 12025344, 'steps': 23486, 'loss/train': 2.2574121952056885} 02/25/2022 08:00:21 - INFO - codeparrot_training - Step 23487: {'lr': 0.00029093372446146613, 'samples': 12025856, 'steps': 23487, 'loss/train': 1.0914568901062012} 02/25/2022 08:00:24 - INFO - codeparrot_training - Step 23488: {'lr': 0.0002909175827337033, 'samples': 12026368, 'steps': 23488, 'loss/train': 2.441570997238159} 02/25/2022 08:00:30 - INFO - codeparrot_training - Step 23489: {'lr': 0.00029090144083066254, 'samples': 12026880, 'steps': 23489, 'loss/train': 1.7172644138336182} 02/25/2022 08:00:33 - INFO - codeparrot_training - Step 23490: {'lr': 0.00029088529875241306, 'samples': 12027392, 'steps': 23490, 'loss/train': 1.4030110836029053} 02/25/2022 08:00:39 - INFO - codeparrot_training - Step 23491: {'lr': 0.00029086915649902383, 'samples': 12027904, 'steps': 23491, 'loss/train': 2.096893548965454} 02/25/2022 08:00:43 - INFO - codeparrot_training - Step 23492: {'lr': 0.0002908530140705642, 'samples': 12028416, 'steps': 23492, 'loss/train': 2.14057993888855} 02/25/2022 08:00:48 - INFO - codeparrot_training - Step 23493: {'lr': 0.0002908368714671031, 'samples': 12028928, 'steps': 23493, 'loss/train': 1.6170334815979004} 02/25/2022 08:00:52 - INFO - codeparrot_training - Step 23494: {'lr': 0.00029082072868870997, 'samples': 12029440, 'steps': 23494, 'loss/train': 1.5164546966552734} 02/25/2022 08:00:57 - INFO - codeparrot_training - Step 23495: {'lr': 0.00029080458573545366, 'samples': 12029952, 'steps': 23495, 'loss/train': 1.3293206691741943} 02/25/2022 08:01:01 - INFO - codeparrot_training - Step 23496: {'lr': 0.0002907884426074036, 'samples': 12030464, 'steps': 23496, 'loss/train': 1.4230456352233887} 02/25/2022 08:01:07 - INFO - codeparrot_training - Step 23497: {'lr': 0.0002907722993046287, 'samples': 12030976, 'steps': 23497, 'loss/train': 1.4491885900497437} 02/25/2022 08:01:10 - INFO - codeparrot_training - Step 23498: {'lr': 0.0002907561558271983, 'samples': 12031488, 'steps': 23498, 'loss/train': 1.9440513849258423} 02/25/2022 08:01:15 - INFO - codeparrot_training - Step 23499: {'lr': 0.0002907400121751814, 'samples': 12032000, 'steps': 23499, 'loss/train': 1.219433307647705} 02/25/2022 08:01:19 - INFO - codeparrot_training - Step 23500: {'lr': 0.0002907238683486472, 'samples': 12032512, 'steps': 23500, 'loss/train': 1.6441736221313477} 02/25/2022 08:01:24 - INFO - codeparrot_training - Step 23501: {'lr': 0.0002907077243476649, 'samples': 12033024, 'steps': 23501, 'loss/train': 1.0706543922424316} 02/25/2022 08:01:28 - INFO - codeparrot_training - Step 23502: {'lr': 0.0002906915801723037, 'samples': 12033536, 'steps': 23502, 'loss/train': 1.668981909751892} 02/25/2022 08:01:33 - INFO - codeparrot_training - Step 23503: {'lr': 0.00029067543582263265, 'samples': 12034048, 'steps': 23503, 'loss/train': 1.8567200899124146} 02/25/2022 08:01:37 - INFO - codeparrot_training - Step 23504: {'lr': 0.00029065929129872095, 'samples': 12034560, 'steps': 23504, 'loss/train': 2.8706367015838623} 02/25/2022 08:01:42 - INFO - codeparrot_training - Step 23505: {'lr': 0.0002906431466006378, 'samples': 12035072, 'steps': 23505, 'loss/train': 2.2313928604125977} 02/25/2022 08:01:46 - INFO - codeparrot_training - Step 23506: {'lr': 0.0002906270017284522, 'samples': 12035584, 'steps': 23506, 'loss/train': 1.8907411098480225} 02/25/2022 08:01:53 - INFO - codeparrot_training - Step 23507: {'lr': 0.0002906108566822335, 'samples': 12036096, 'steps': 23507, 'loss/train': 0.9114428758621216} 02/25/2022 08:01:56 - INFO - codeparrot_training - Step 23508: {'lr': 0.0002905947114620508, 'samples': 12036608, 'steps': 23508, 'loss/train': 2.088773727416992} 02/25/2022 08:02:02 - INFO - codeparrot_training - Step 23509: {'lr': 0.0002905785660679732, 'samples': 12037120, 'steps': 23509, 'loss/train': 1.829337477684021} 02/25/2022 08:02:05 - INFO - codeparrot_training - Step 23510: {'lr': 0.00029056242050006995, 'samples': 12037632, 'steps': 23510, 'loss/train': 1.8250982761383057} 02/25/2022 08:02:11 - INFO - codeparrot_training - Step 23511: {'lr': 0.0002905462747584101, 'samples': 12038144, 'steps': 23511, 'loss/train': 2.116312026977539} 02/25/2022 08:02:14 - INFO - codeparrot_training - Step 23512: {'lr': 0.00029053012884306295, 'samples': 12038656, 'steps': 23512, 'loss/train': 2.298887252807617} 02/25/2022 08:02:20 - INFO - codeparrot_training - Step 23513: {'lr': 0.00029051398275409756, 'samples': 12039168, 'steps': 23513, 'loss/train': 2.1299140453338623} 02/25/2022 08:02:23 - INFO - codeparrot_training - Step 23514: {'lr': 0.0002904978364915832, 'samples': 12039680, 'steps': 23514, 'loss/train': 1.426316499710083} 02/25/2022 08:02:29 - INFO - codeparrot_training - Step 23515: {'lr': 0.0002904816900555889, 'samples': 12040192, 'steps': 23515, 'loss/train': 2.437614917755127} 02/25/2022 08:02:32 - INFO - codeparrot_training - Step 23516: {'lr': 0.000290465543446184, 'samples': 12040704, 'steps': 23516, 'loss/train': 2.7329211235046387} 02/25/2022 08:02:39 - INFO - codeparrot_training - Step 23517: {'lr': 0.0002904493966634374, 'samples': 12041216, 'steps': 23517, 'loss/train': 1.8629857301712036} 02/25/2022 08:02:42 - INFO - codeparrot_training - Step 23518: {'lr': 0.0002904332497074186, 'samples': 12041728, 'steps': 23518, 'loss/train': 2.3979287147521973} 02/25/2022 08:02:48 - INFO - codeparrot_training - Step 23519: {'lr': 0.00029041710257819643, 'samples': 12042240, 'steps': 23519, 'loss/train': 1.4235047101974487} 02/25/2022 08:02:51 - INFO - codeparrot_training - Step 23520: {'lr': 0.0002904009552758403, 'samples': 12042752, 'steps': 23520, 'loss/train': 2.374789237976074} 02/25/2022 08:02:57 - INFO - codeparrot_training - Step 23521: {'lr': 0.0002903848078004193, 'samples': 12043264, 'steps': 23521, 'loss/train': 2.884300708770752} 02/25/2022 08:03:00 - INFO - codeparrot_training - Step 23522: {'lr': 0.00029036866015200256, 'samples': 12043776, 'steps': 23522, 'loss/train': 1.8297744989395142} 02/25/2022 08:03:06 - INFO - codeparrot_training - Step 23523: {'lr': 0.0002903525123306594, 'samples': 12044288, 'steps': 23523, 'loss/train': 2.233595609664917} 02/25/2022 08:03:09 - INFO - codeparrot_training - Step 23524: {'lr': 0.0002903363643364588, 'samples': 12044800, 'steps': 23524, 'loss/train': 1.9957176446914673} 02/25/2022 08:03:15 - INFO - codeparrot_training - Step 23525: {'lr': 0.0002903202161694701, 'samples': 12045312, 'steps': 23525, 'loss/train': 2.4707109928131104} 02/25/2022 08:03:18 - INFO - codeparrot_training - Step 23526: {'lr': 0.0002903040678297624, 'samples': 12045824, 'steps': 23526, 'loss/train': 2.564110279083252} 02/25/2022 08:03:25 - INFO - codeparrot_training - Step 23527: {'lr': 0.00029028791931740483, 'samples': 12046336, 'steps': 23527, 'loss/train': 2.197964668273926} 02/25/2022 08:03:28 - INFO - codeparrot_training - Step 23528: {'lr': 0.0002902717706324666, 'samples': 12046848, 'steps': 23528, 'loss/train': 2.2101938724517822} 02/25/2022 08:03:33 - INFO - codeparrot_training - Step 23529: {'lr': 0.00029025562177501695, 'samples': 12047360, 'steps': 23529, 'loss/train': 2.7704222202301025} 02/25/2022 08:03:37 - INFO - codeparrot_training - Step 23530: {'lr': 0.0002902394727451249, 'samples': 12047872, 'steps': 23530, 'loss/train': 2.554912567138672} 02/25/2022 08:03:42 - INFO - codeparrot_training - Step 23531: {'lr': 0.00029022332354285986, 'samples': 12048384, 'steps': 23531, 'loss/train': 1.7678285837173462} 02/25/2022 08:03:46 - INFO - codeparrot_training - Step 23532: {'lr': 0.00029020717416829075, 'samples': 12048896, 'steps': 23532, 'loss/train': 2.1888411045074463} 02/25/2022 08:03:51 - INFO - codeparrot_training - Step 23533: {'lr': 0.0002901910246214869, 'samples': 12049408, 'steps': 23533, 'loss/train': 1.2653511762619019} 02/25/2022 08:03:55 - INFO - codeparrot_training - Step 23534: {'lr': 0.0002901748749025175, 'samples': 12049920, 'steps': 23534, 'loss/train': 2.6030540466308594} 02/25/2022 08:04:00 - INFO - codeparrot_training - Step 23535: {'lr': 0.00029015872501145164, 'samples': 12050432, 'steps': 23535, 'loss/train': 1.6123956441879272} 02/25/2022 08:04:04 - INFO - codeparrot_training - Step 23536: {'lr': 0.00029014257494835863, 'samples': 12050944, 'steps': 23536, 'loss/train': 1.9868042469024658} 02/25/2022 08:04:10 - INFO - codeparrot_training - Step 23537: {'lr': 0.0002901264247133075, 'samples': 12051456, 'steps': 23537, 'loss/train': 2.39188814163208} 02/25/2022 08:04:14 - INFO - codeparrot_training - Step 23538: {'lr': 0.00029011027430636755, 'samples': 12051968, 'steps': 23538, 'loss/train': 1.4962033033370972} 02/25/2022 08:04:19 - INFO - codeparrot_training - Step 23539: {'lr': 0.00029009412372760793, 'samples': 12052480, 'steps': 23539, 'loss/train': 0.9243882894515991} 02/25/2022 08:04:23 - INFO - codeparrot_training - Step 23540: {'lr': 0.00029007797297709784, 'samples': 12052992, 'steps': 23540, 'loss/train': 2.113966226577759} 02/25/2022 08:04:28 - INFO - codeparrot_training - Step 23541: {'lr': 0.00029006182205490634, 'samples': 12053504, 'steps': 23541, 'loss/train': 2.3709867000579834} 02/25/2022 08:04:31 - INFO - codeparrot_training - Step 23542: {'lr': 0.0002900456709611028, 'samples': 12054016, 'steps': 23542, 'loss/train': 1.5875611305236816} 02/25/2022 08:04:37 - INFO - codeparrot_training - Step 23543: {'lr': 0.0002900295196957563, 'samples': 12054528, 'steps': 23543, 'loss/train': 1.8569351434707642} 02/25/2022 08:04:41 - INFO - codeparrot_training - Step 23544: {'lr': 0.00029001336825893603, 'samples': 12055040, 'steps': 23544, 'loss/train': 2.1762990951538086} 02/25/2022 08:04:46 - INFO - codeparrot_training - Step 23545: {'lr': 0.0002899972166507112, 'samples': 12055552, 'steps': 23545, 'loss/train': 2.8551700115203857} 02/25/2022 08:04:50 - INFO - codeparrot_training - Step 23546: {'lr': 0.000289981064871151, 'samples': 12056064, 'steps': 23546, 'loss/train': 0.300271213054657} 02/25/2022 08:04:55 - INFO - codeparrot_training - Step 23547: {'lr': 0.00028996491292032465, 'samples': 12056576, 'steps': 23547, 'loss/train': 2.8087878227233887} 02/25/2022 08:04:58 - INFO - codeparrot_training - Step 23548: {'lr': 0.00028994876079830125, 'samples': 12057088, 'steps': 23548, 'loss/train': 2.1102256774902344} 02/25/2022 08:05:04 - INFO - codeparrot_training - Step 23549: {'lr': 0.00028993260850515015, 'samples': 12057600, 'steps': 23549, 'loss/train': 1.7570375204086304} 02/25/2022 08:05:08 - INFO - codeparrot_training - Step 23550: {'lr': 0.0002899164560409403, 'samples': 12058112, 'steps': 23550, 'loss/train': 2.206887722015381} 02/25/2022 08:05:13 - INFO - codeparrot_training - Step 23551: {'lr': 0.00028990030340574107, 'samples': 12058624, 'steps': 23551, 'loss/train': 2.4764509201049805} 02/25/2022 08:05:17 - INFO - codeparrot_training - Step 23552: {'lr': 0.0002898841505996216, 'samples': 12059136, 'steps': 23552, 'loss/train': 1.5376133918762207} 02/25/2022 08:05:23 - INFO - codeparrot_training - Step 23553: {'lr': 0.0002898679976226511, 'samples': 12059648, 'steps': 23553, 'loss/train': 1.9802324771881104} 02/25/2022 08:05:26 - INFO - codeparrot_training - Step 23554: {'lr': 0.00028985184447489874, 'samples': 12060160, 'steps': 23554, 'loss/train': 2.137173891067505} 02/25/2022 08:05:32 - INFO - codeparrot_training - Step 23555: {'lr': 0.00028983569115643385, 'samples': 12060672, 'steps': 23555, 'loss/train': 1.8138989210128784} 02/25/2022 08:05:35 - INFO - codeparrot_training - Step 23556: {'lr': 0.0002898195376673254, 'samples': 12061184, 'steps': 23556, 'loss/train': 2.3573219776153564} 02/25/2022 08:05:41 - INFO - codeparrot_training - Step 23557: {'lr': 0.0002898033840076427, 'samples': 12061696, 'steps': 23557, 'loss/train': 2.5795705318450928} 02/25/2022 08:05:44 - INFO - codeparrot_training - Step 23558: {'lr': 0.00028978723017745496, 'samples': 12062208, 'steps': 23558, 'loss/train': 1.5039584636688232} 02/25/2022 08:05:50 - INFO - codeparrot_training - Step 23559: {'lr': 0.0002897710761768313, 'samples': 12062720, 'steps': 23559, 'loss/train': 1.0971256494522095} 02/25/2022 08:05:53 - INFO - codeparrot_training - Step 23560: {'lr': 0.00028975492200584106, 'samples': 12063232, 'steps': 23560, 'loss/train': 2.4428679943084717} 02/25/2022 08:05:59 - INFO - codeparrot_training - Step 23561: {'lr': 0.00028973876766455334, 'samples': 12063744, 'steps': 23561, 'loss/train': 1.3993030786514282} 02/25/2022 08:06:02 - INFO - codeparrot_training - Step 23562: {'lr': 0.00028972261315303736, 'samples': 12064256, 'steps': 23562, 'loss/train': 0.6624284982681274} 02/25/2022 08:06:08 - INFO - codeparrot_training - Step 23563: {'lr': 0.0002897064584713623, 'samples': 12064768, 'steps': 23563, 'loss/train': 2.4249911308288574} 02/25/2022 08:06:11 - INFO - codeparrot_training - Step 23564: {'lr': 0.0002896903036195974, 'samples': 12065280, 'steps': 23564, 'loss/train': 0.613732099533081} 02/25/2022 08:06:18 - INFO - codeparrot_training - Step 23565: {'lr': 0.0002896741485978118, 'samples': 12065792, 'steps': 23565, 'loss/train': 1.7938508987426758} 02/25/2022 08:06:21 - INFO - codeparrot_training - Step 23566: {'lr': 0.0002896579934060748, 'samples': 12066304, 'steps': 23566, 'loss/train': 0.8712823987007141} 02/25/2022 08:06:27 - INFO - codeparrot_training - Step 23567: {'lr': 0.00028964183804445554, 'samples': 12066816, 'steps': 23567, 'loss/train': 2.06577730178833} 02/25/2022 08:06:30 - INFO - codeparrot_training - Step 23568: {'lr': 0.00028962568251302326, 'samples': 12067328, 'steps': 23568, 'loss/train': 1.487418293952942} 02/25/2022 08:06:36 - INFO - codeparrot_training - Step 23569: {'lr': 0.00028960952681184705, 'samples': 12067840, 'steps': 23569, 'loss/train': 1.7848950624465942} 02/25/2022 08:06:39 - INFO - codeparrot_training - Step 23570: {'lr': 0.00028959337094099617, 'samples': 12068352, 'steps': 23570, 'loss/train': 2.7480626106262207} 02/25/2022 08:06:44 - INFO - codeparrot_training - Step 23571: {'lr': 0.00028957721490054, 'samples': 12068864, 'steps': 23571, 'loss/train': 1.9302492141723633} 02/25/2022 08:06:51 - INFO - codeparrot_training - Step 23572: {'lr': 0.00028956105869054757, 'samples': 12069376, 'steps': 23572, 'loss/train': 1.0581475496292114} 02/25/2022 08:06:54 - INFO - codeparrot_training - Step 23573: {'lr': 0.0002895449023110881, 'samples': 12069888, 'steps': 23573, 'loss/train': 2.0345520973205566} 02/25/2022 08:07:00 - INFO - codeparrot_training - Step 23574: {'lr': 0.00028952874576223083, 'samples': 12070400, 'steps': 23574, 'loss/train': 1.625321865081787} 02/25/2022 08:07:03 - INFO - codeparrot_training - Step 23575: {'lr': 0.000289512589044045, 'samples': 12070912, 'steps': 23575, 'loss/train': 2.2414658069610596} 02/25/2022 08:07:09 - INFO - codeparrot_training - Step 23576: {'lr': 0.0002894964321565997, 'samples': 12071424, 'steps': 23576, 'loss/train': 0.6521539092063904} 02/25/2022 08:07:12 - INFO - codeparrot_training - Step 23577: {'lr': 0.00028948027509996435, 'samples': 12071936, 'steps': 23577, 'loss/train': 0.5160545706748962} 02/25/2022 08:07:18 - INFO - codeparrot_training - Step 23578: {'lr': 0.0002894641178742079, 'samples': 12072448, 'steps': 23578, 'loss/train': 2.873347520828247} 02/25/2022 08:07:21 - INFO - codeparrot_training - Step 23579: {'lr': 0.0002894479604793998, 'samples': 12072960, 'steps': 23579, 'loss/train': 2.0298352241516113} 02/25/2022 08:07:27 - INFO - codeparrot_training - Step 23580: {'lr': 0.0002894318029156091, 'samples': 12073472, 'steps': 23580, 'loss/train': 2.342482328414917} 02/25/2022 08:07:30 - INFO - codeparrot_training - Step 23581: {'lr': 0.0002894156451829051, 'samples': 12073984, 'steps': 23581, 'loss/train': 2.3614048957824707} 02/25/2022 08:07:36 - INFO - codeparrot_training - Step 23582: {'lr': 0.000289399487281357, 'samples': 12074496, 'steps': 23582, 'loss/train': 1.5407274961471558} 02/25/2022 08:07:39 - INFO - codeparrot_training - Step 23583: {'lr': 0.000289383329211034, 'samples': 12075008, 'steps': 23583, 'loss/train': 1.610329270362854} 02/25/2022 08:07:45 - INFO - codeparrot_training - Step 23584: {'lr': 0.0002893671709720054, 'samples': 12075520, 'steps': 23584, 'loss/train': 1.6335129737854004} 02/25/2022 08:07:49 - INFO - codeparrot_training - Step 23585: {'lr': 0.0002893510125643403, 'samples': 12076032, 'steps': 23585, 'loss/train': 1.1203486919403076} 02/25/2022 08:07:54 - INFO - codeparrot_training - Step 23586: {'lr': 0.0002893348539881079, 'samples': 12076544, 'steps': 23586, 'loss/train': 1.49395751953125} 02/25/2022 08:07:58 - INFO - codeparrot_training - Step 23587: {'lr': 0.0002893186952433775, 'samples': 12077056, 'steps': 23587, 'loss/train': 1.9389312267303467} 02/25/2022 08:08:03 - INFO - codeparrot_training - Step 23588: {'lr': 0.00028930253633021826, 'samples': 12077568, 'steps': 23588, 'loss/train': 2.972951650619507} 02/25/2022 08:08:07 - INFO - codeparrot_training - Step 23589: {'lr': 0.0002892863772486995, 'samples': 12078080, 'steps': 23589, 'loss/train': 1.420477271080017} 02/25/2022 08:08:12 - INFO - codeparrot_training - Step 23590: {'lr': 0.0002892702179988903, 'samples': 12078592, 'steps': 23590, 'loss/train': 0.7500631809234619} 02/25/2022 08:08:16 - INFO - codeparrot_training - Step 23591: {'lr': 0.00028925405858086007, 'samples': 12079104, 'steps': 23591, 'loss/train': 2.364516496658325} 02/25/2022 08:08:21 - INFO - codeparrot_training - Step 23592: {'lr': 0.0002892378989946779, 'samples': 12079616, 'steps': 23592, 'loss/train': 1.313019871711731} 02/25/2022 08:08:25 - INFO - codeparrot_training - Step 23593: {'lr': 0.00028922173924041296, 'samples': 12080128, 'steps': 23593, 'loss/train': 1.4040350914001465} 02/25/2022 08:08:31 - INFO - codeparrot_training - Step 23594: {'lr': 0.00028920557931813454, 'samples': 12080640, 'steps': 23594, 'loss/train': 1.1124212741851807} 02/25/2022 08:08:34 - INFO - codeparrot_training - Step 23595: {'lr': 0.00028918941922791185, 'samples': 12081152, 'steps': 23595, 'loss/train': 1.933236002922058} 02/25/2022 08:08:39 - INFO - codeparrot_training - Step 23596: {'lr': 0.00028917325896981417, 'samples': 12081664, 'steps': 23596, 'loss/train': 1.9197604656219482} 02/25/2022 08:08:43 - INFO - codeparrot_training - Step 23597: {'lr': 0.0002891570985439108, 'samples': 12082176, 'steps': 23597, 'loss/train': 2.299419403076172} 02/25/2022 08:08:49 - INFO - codeparrot_training - Step 23598: {'lr': 0.00028914093795027063, 'samples': 12082688, 'steps': 23598, 'loss/train': 1.5218538045883179} 02/25/2022 08:08:53 - INFO - codeparrot_training - Step 23599: {'lr': 0.0002891247771889633, 'samples': 12083200, 'steps': 23599, 'loss/train': 2.170194625854492} 02/25/2022 08:08:58 - INFO - codeparrot_training - Step 23600: {'lr': 0.00028910861626005774, 'samples': 12083712, 'steps': 23600, 'loss/train': 1.9556257724761963} 02/25/2022 08:09:02 - INFO - codeparrot_training - Step 23601: {'lr': 0.0002890924551636234, 'samples': 12084224, 'steps': 23601, 'loss/train': 1.3168898820877075} 02/25/2022 08:09:07 - INFO - codeparrot_training - Step 23602: {'lr': 0.00028907629389972924, 'samples': 12084736, 'steps': 23602, 'loss/train': 1.6897366046905518} 02/25/2022 08:09:11 - INFO - codeparrot_training - Step 23603: {'lr': 0.00028906013246844474, 'samples': 12085248, 'steps': 23603, 'loss/train': 2.4212253093719482} 02/25/2022 08:09:16 - INFO - codeparrot_training - Step 23604: {'lr': 0.000289043970869839, 'samples': 12085760, 'steps': 23604, 'loss/train': 1.745396614074707} 02/25/2022 08:09:20 - INFO - codeparrot_training - Step 23605: {'lr': 0.0002890278091039813, 'samples': 12086272, 'steps': 23605, 'loss/train': 1.6462994813919067} 02/25/2022 08:09:25 - INFO - codeparrot_training - Step 23606: {'lr': 0.00028901164717094085, 'samples': 12086784, 'steps': 23606, 'loss/train': 0.9969839453697205} 02/25/2022 08:09:29 - INFO - codeparrot_training - Step 23607: {'lr': 0.0002889954850707869, 'samples': 12087296, 'steps': 23607, 'loss/train': 2.06868052482605} 02/25/2022 08:09:34 - INFO - codeparrot_training - Step 23608: {'lr': 0.0002889793228035887, 'samples': 12087808, 'steps': 23608, 'loss/train': 0.4268155097961426} 02/25/2022 08:09:37 - INFO - codeparrot_training - Step 23609: {'lr': 0.0002889631603694154, 'samples': 12088320, 'steps': 23609, 'loss/train': 2.3262276649475098} 02/25/2022 08:09:44 - INFO - codeparrot_training - Step 23610: {'lr': 0.00028894699776833637, 'samples': 12088832, 'steps': 23610, 'loss/train': 2.541623115539551} 02/25/2022 08:09:47 - INFO - codeparrot_training - Step 23611: {'lr': 0.0002889308350004207, 'samples': 12089344, 'steps': 23611, 'loss/train': 2.428924083709717} 02/25/2022 08:09:53 - INFO - codeparrot_training - Step 23612: {'lr': 0.00028891467206573773, 'samples': 12089856, 'steps': 23612, 'loss/train': 2.2161452770233154} 02/25/2022 08:09:56 - INFO - codeparrot_training - Step 23613: {'lr': 0.00028889850896435666, 'samples': 12090368, 'steps': 23613, 'loss/train': 1.9017736911773682} 02/25/2022 08:10:02 - INFO - codeparrot_training - Step 23614: {'lr': 0.00028888234569634673, 'samples': 12090880, 'steps': 23614, 'loss/train': 0.44609636068344116} 02/25/2022 08:10:05 - INFO - codeparrot_training - Step 23615: {'lr': 0.00028886618226177716, 'samples': 12091392, 'steps': 23615, 'loss/train': 0.5661726593971252} 02/25/2022 08:10:11 - INFO - codeparrot_training - Step 23616: {'lr': 0.00028885001866071723, 'samples': 12091904, 'steps': 23616, 'loss/train': 1.4779706001281738} 02/25/2022 08:10:14 - INFO - codeparrot_training - Step 23617: {'lr': 0.0002888338548932361, 'samples': 12092416, 'steps': 23617, 'loss/train': 1.9706391096115112} 02/25/2022 08:10:20 - INFO - codeparrot_training - Step 23618: {'lr': 0.00028881769095940316, 'samples': 12092928, 'steps': 23618, 'loss/train': 1.127479910850525} 02/25/2022 08:10:23 - INFO - codeparrot_training - Step 23619: {'lr': 0.0002888015268592875, 'samples': 12093440, 'steps': 23619, 'loss/train': 0.5383728742599487} 02/25/2022 08:10:29 - INFO - codeparrot_training - Step 23620: {'lr': 0.0002887853625929584, 'samples': 12093952, 'steps': 23620, 'loss/train': 1.414881944656372} 02/25/2022 08:10:33 - INFO - codeparrot_training - Step 23621: {'lr': 0.00028876919816048517, 'samples': 12094464, 'steps': 23621, 'loss/train': 2.3857603073120117} 02/25/2022 08:10:38 - INFO - codeparrot_training - Step 23622: {'lr': 0.00028875303356193697, 'samples': 12094976, 'steps': 23622, 'loss/train': 2.302809476852417} 02/25/2022 08:10:42 - INFO - codeparrot_training - Step 23623: {'lr': 0.0002887368687973831, 'samples': 12095488, 'steps': 23623, 'loss/train': 2.480724811553955} 02/25/2022 08:10:47 - INFO - codeparrot_training - Step 23624: {'lr': 0.00028872070386689274, 'samples': 12096000, 'steps': 23624, 'loss/train': 2.1395230293273926} 02/25/2022 08:10:51 - INFO - codeparrot_training - Step 23625: {'lr': 0.00028870453877053527, 'samples': 12096512, 'steps': 23625, 'loss/train': 1.15674889087677} 02/25/2022 08:10:56 - INFO - codeparrot_training - Step 23626: {'lr': 0.00028868837350837977, 'samples': 12097024, 'steps': 23626, 'loss/train': 1.0673731565475464} 02/25/2022 08:11:00 - INFO - codeparrot_training - Step 23627: {'lr': 0.0002886722080804956, 'samples': 12097536, 'steps': 23627, 'loss/train': 2.1376001834869385} 02/25/2022 08:11:05 - INFO - codeparrot_training - Step 23628: {'lr': 0.000288656042486952, 'samples': 12098048, 'steps': 23628, 'loss/train': 2.4457335472106934} 02/25/2022 08:11:09 - INFO - codeparrot_training - Step 23629: {'lr': 0.0002886398767278181, 'samples': 12098560, 'steps': 23629, 'loss/train': 2.3115906715393066} 02/25/2022 08:11:15 - INFO - codeparrot_training - Step 23630: {'lr': 0.0002886237108031633, 'samples': 12099072, 'steps': 23630, 'loss/train': 1.891711950302124} 02/25/2022 08:11:19 - INFO - codeparrot_training - Step 23631: {'lr': 0.0002886075447130568, 'samples': 12099584, 'steps': 23631, 'loss/train': 2.793041706085205} 02/25/2022 08:11:24 - INFO - codeparrot_training - Step 23632: {'lr': 0.00028859137845756784, 'samples': 12100096, 'steps': 23632, 'loss/train': 2.1435625553131104} 02/25/2022 08:11:28 - INFO - codeparrot_training - Step 23633: {'lr': 0.00028857521203676565, 'samples': 12100608, 'steps': 23633, 'loss/train': 1.3745698928833008} 02/25/2022 08:11:33 - INFO - codeparrot_training - Step 23634: {'lr': 0.0002885590454507195, 'samples': 12101120, 'steps': 23634, 'loss/train': 1.522180199623108} 02/25/2022 08:11:37 - INFO - codeparrot_training - Step 23635: {'lr': 0.00028854287869949856, 'samples': 12101632, 'steps': 23635, 'loss/train': 2.0995702743530273} 02/25/2022 08:11:42 - INFO - codeparrot_training - Step 23636: {'lr': 0.00028852671178317233, 'samples': 12102144, 'steps': 23636, 'loss/train': 1.5845879316329956} 02/25/2022 08:11:46 - INFO - codeparrot_training - Step 23637: {'lr': 0.00028851054470180977, 'samples': 12102656, 'steps': 23637, 'loss/train': 2.264266014099121} 02/25/2022 08:11:51 - INFO - codeparrot_training - Step 23638: {'lr': 0.0002884943774554803, 'samples': 12103168, 'steps': 23638, 'loss/train': 6.310962677001953} 02/25/2022 08:11:55 - INFO - codeparrot_training - Step 23639: {'lr': 0.0002884782100442531, 'samples': 12103680, 'steps': 23639, 'loss/train': 2.1470947265625} 02/25/2022 08:12:00 - INFO - codeparrot_training - Step 23640: {'lr': 0.0002884620424681976, 'samples': 12104192, 'steps': 23640, 'loss/train': 1.6811007261276245} 02/25/2022 08:12:04 - INFO - codeparrot_training - Step 23641: {'lr': 0.0002884458747273828, 'samples': 12104704, 'steps': 23641, 'loss/train': 1.70032799243927} 02/25/2022 08:12:09 - INFO - codeparrot_training - Step 23642: {'lr': 0.0002884297068218781, 'samples': 12105216, 'steps': 23642, 'loss/train': 0.9124624729156494} 02/25/2022 08:12:13 - INFO - codeparrot_training - Step 23643: {'lr': 0.00028841353875175274, 'samples': 12105728, 'steps': 23643, 'loss/train': 1.7605984210968018} 02/25/2022 08:12:19 - INFO - codeparrot_training - Step 23644: {'lr': 0.000288397370517076, 'samples': 12106240, 'steps': 23644, 'loss/train': 2.816835403442383} 02/25/2022 08:12:22 - INFO - codeparrot_training - Step 23645: {'lr': 0.0002883812021179171, 'samples': 12106752, 'steps': 23645, 'loss/train': 2.8557376861572266} 02/25/2022 08:12:28 - INFO - codeparrot_training - Step 23646: {'lr': 0.0002883650335543453, 'samples': 12107264, 'steps': 23646, 'loss/train': 0.38947081565856934} 02/25/2022 08:12:32 - INFO - codeparrot_training - Step 23647: {'lr': 0.00028834886482643, 'samples': 12107776, 'steps': 23647, 'loss/train': 2.050731897354126} 02/25/2022 08:12:37 - INFO - codeparrot_training - Step 23648: {'lr': 0.00028833269593424017, 'samples': 12108288, 'steps': 23648, 'loss/train': 1.3458151817321777} 02/25/2022 08:12:41 - INFO - codeparrot_training - Step 23649: {'lr': 0.0002883165268778454, 'samples': 12108800, 'steps': 23649, 'loss/train': 1.7565962076187134} 02/25/2022 08:12:46 - INFO - codeparrot_training - Step 23650: {'lr': 0.00028830035765731464, 'samples': 12109312, 'steps': 23650, 'loss/train': 1.6757336854934692} 02/25/2022 08:12:50 - INFO - codeparrot_training - Step 23651: {'lr': 0.0002882841882727174, 'samples': 12109824, 'steps': 23651, 'loss/train': 1.373449683189392} 02/25/2022 08:12:56 - INFO - codeparrot_training - Step 23652: {'lr': 0.00028826801872412284, 'samples': 12110336, 'steps': 23652, 'loss/train': 1.4238611459732056} 02/25/2022 08:12:59 - INFO - codeparrot_training - Step 23653: {'lr': 0.00028825184901160023, 'samples': 12110848, 'steps': 23653, 'loss/train': 1.948917269706726} 02/25/2022 08:13:05 - INFO - codeparrot_training - Step 23654: {'lr': 0.00028823567913521884, 'samples': 12111360, 'steps': 23654, 'loss/train': 1.3194708824157715} 02/25/2022 08:13:08 - INFO - codeparrot_training - Step 23655: {'lr': 0.0002882195090950479, 'samples': 12111872, 'steps': 23655, 'loss/train': 0.4681048095226288} 02/25/2022 08:13:14 - INFO - codeparrot_training - Step 23656: {'lr': 0.00028820333889115684, 'samples': 12112384, 'steps': 23656, 'loss/train': 2.575822591781616} 02/25/2022 08:13:18 - INFO - codeparrot_training - Step 23657: {'lr': 0.0002881871685236147, 'samples': 12112896, 'steps': 23657, 'loss/train': 2.1207542419433594} 02/25/2022 08:13:23 - INFO - codeparrot_training - Step 23658: {'lr': 0.0002881709979924909, 'samples': 12113408, 'steps': 23658, 'loss/train': 1.9079558849334717} 02/25/2022 08:13:27 - INFO - codeparrot_training - Step 23659: {'lr': 0.00028815482729785467, 'samples': 12113920, 'steps': 23659, 'loss/train': 2.139220952987671} 02/25/2022 08:13:32 - INFO - codeparrot_training - Step 23660: {'lr': 0.00028813865643977527, 'samples': 12114432, 'steps': 23660, 'loss/train': 1.6365087032318115} 02/25/2022 08:13:36 - INFO - codeparrot_training - Step 23661: {'lr': 0.000288122485418322, 'samples': 12114944, 'steps': 23661, 'loss/train': 0.6145601868629456} 02/25/2022 08:13:41 - INFO - codeparrot_training - Step 23662: {'lr': 0.0002881063142335641, 'samples': 12115456, 'steps': 23662, 'loss/train': 1.2969300746917725} 02/25/2022 08:13:45 - INFO - codeparrot_training - Step 23663: {'lr': 0.0002880901428855708, 'samples': 12115968, 'steps': 23663, 'loss/train': 1.2844479084014893} 02/25/2022 08:13:50 - INFO - codeparrot_training - Step 23664: {'lr': 0.0002880739713744114, 'samples': 12116480, 'steps': 23664, 'loss/train': 1.845025658607483} 02/25/2022 08:13:54 - INFO - codeparrot_training - Step 23665: {'lr': 0.00028805779970015525, 'samples': 12116992, 'steps': 23665, 'loss/train': 0.7780728340148926} 02/25/2022 08:14:00 - INFO - codeparrot_training - Step 23666: {'lr': 0.00028804162786287156, 'samples': 12117504, 'steps': 23666, 'loss/train': 0.5093708634376526} 02/25/2022 08:14:04 - INFO - codeparrot_training - Step 23667: {'lr': 0.0002880254558626297, 'samples': 12118016, 'steps': 23667, 'loss/train': 1.828165888786316} 02/25/2022 08:14:09 - INFO - codeparrot_training - Step 23668: {'lr': 0.0002880092836994987, 'samples': 12118528, 'steps': 23668, 'loss/train': 2.9405670166015625} 02/25/2022 08:14:13 - INFO - codeparrot_training - Step 23669: {'lr': 0.0002879931113735482, 'samples': 12119040, 'steps': 23669, 'loss/train': 1.4543344974517822} 02/25/2022 08:14:18 - INFO - codeparrot_training - Step 23670: {'lr': 0.0002879769388848471, 'samples': 12119552, 'steps': 23670, 'loss/train': 2.3173696994781494} 02/25/2022 08:14:22 - INFO - codeparrot_training - Step 23671: {'lr': 0.0002879607662334649, 'samples': 12120064, 'steps': 23671, 'loss/train': 0.4551342725753784} 02/25/2022 08:14:28 - INFO - codeparrot_training - Step 23672: {'lr': 0.0002879445934194709, 'samples': 12120576, 'steps': 23672, 'loss/train': 0.9114976525306702} 02/25/2022 08:14:31 - INFO - codeparrot_training - Step 23673: {'lr': 0.0002879284204429342, 'samples': 12121088, 'steps': 23673, 'loss/train': 2.378455400466919} 02/25/2022 08:14:36 - INFO - codeparrot_training - Step 23674: {'lr': 0.0002879122473039243, 'samples': 12121600, 'steps': 23674, 'loss/train': 3.390928030014038} 02/25/2022 08:14:40 - INFO - codeparrot_training - Step 23675: {'lr': 0.0002878960740025103, 'samples': 12122112, 'steps': 23675, 'loss/train': 1.4086164236068726} 02/25/2022 08:14:46 - INFO - codeparrot_training - Step 23676: {'lr': 0.00028787990053876156, 'samples': 12122624, 'steps': 23676, 'loss/train': 1.6998674869537354} 02/25/2022 08:14:49 - INFO - codeparrot_training - Step 23677: {'lr': 0.00028786372691274735, 'samples': 12123136, 'steps': 23677, 'loss/train': 1.3767874240875244} 02/25/2022 08:14:55 - INFO - codeparrot_training - Step 23678: {'lr': 0.00028784755312453704, 'samples': 12123648, 'steps': 23678, 'loss/train': 2.796020269393921} 02/25/2022 08:14:58 - INFO - codeparrot_training - Step 23679: {'lr': 0.0002878313791741997, 'samples': 12124160, 'steps': 23679, 'loss/train': 1.5031064748764038} 02/25/2022 08:15:05 - INFO - codeparrot_training - Step 23680: {'lr': 0.00028781520506180485, 'samples': 12124672, 'steps': 23680, 'loss/train': 1.7649303674697876} 02/25/2022 08:15:08 - INFO - codeparrot_training - Step 23681: {'lr': 0.0002877990307874216, 'samples': 12125184, 'steps': 23681, 'loss/train': 1.710687279701233} 02/25/2022 08:15:11 - INFO - codeparrot_training - Step 23682: {'lr': 0.00028778285635111935, 'samples': 12125696, 'steps': 23682, 'loss/train': 2.431793689727783} 02/25/2022 08:15:18 - INFO - codeparrot_training - Step 23683: {'lr': 0.0002877666817529673, 'samples': 12126208, 'steps': 23683, 'loss/train': 0.13264045119285583} 02/25/2022 08:15:21 - INFO - codeparrot_training - Step 23684: {'lr': 0.0002877505069930348, 'samples': 12126720, 'steps': 23684, 'loss/train': 0.6525019407272339} 02/25/2022 08:15:27 - INFO - codeparrot_training - Step 23685: {'lr': 0.0002877343320713911, 'samples': 12127232, 'steps': 23685, 'loss/train': 1.9331852197647095} 02/25/2022 08:15:30 - INFO - codeparrot_training - Step 23686: {'lr': 0.0002877181569881055, 'samples': 12127744, 'steps': 23686, 'loss/train': 1.9713950157165527} 02/25/2022 08:15:35 - INFO - codeparrot_training - Step 23687: {'lr': 0.00028770198174324737, 'samples': 12128256, 'steps': 23687, 'loss/train': 1.7827495336532593} 02/25/2022 08:15:39 - INFO - codeparrot_training - Step 23688: {'lr': 0.00028768580633688586, 'samples': 12128768, 'steps': 23688, 'loss/train': 1.46895170211792} 02/25/2022 08:15:45 - INFO - codeparrot_training - Step 23689: {'lr': 0.00028766963076909033, 'samples': 12129280, 'steps': 23689, 'loss/train': 0.99400794506073} 02/25/2022 08:15:48 - INFO - codeparrot_training - Step 23690: {'lr': 0.00028765345503993, 'samples': 12129792, 'steps': 23690, 'loss/train': 1.38034188747406} 02/25/2022 08:15:54 - INFO - codeparrot_training - Step 23691: {'lr': 0.0002876372791494743, 'samples': 12130304, 'steps': 23691, 'loss/train': 1.8652857542037964} 02/25/2022 08:15:57 - INFO - codeparrot_training - Step 23692: {'lr': 0.00028762110309779246, 'samples': 12130816, 'steps': 23692, 'loss/train': 2.032402515411377} 02/25/2022 08:16:03 - INFO - codeparrot_training - Step 23693: {'lr': 0.0002876049268849537, 'samples': 12131328, 'steps': 23693, 'loss/train': 1.7879059314727783} 02/25/2022 08:16:07 - INFO - codeparrot_training - Step 23694: {'lr': 0.0002875887505110274, 'samples': 12131840, 'steps': 23694, 'loss/train': 2.240527868270874} 02/25/2022 08:16:13 - INFO - codeparrot_training - Step 23695: {'lr': 0.00028757257397608285, 'samples': 12132352, 'steps': 23695, 'loss/train': 1.4836267232894897} 02/25/2022 08:16:16 - INFO - codeparrot_training - Step 23696: {'lr': 0.0002875563972801893, 'samples': 12132864, 'steps': 23696, 'loss/train': 2.52135968208313} 02/25/2022 08:16:19 - INFO - codeparrot_training - Step 23697: {'lr': 0.0002875402204234161, 'samples': 12133376, 'steps': 23697, 'loss/train': 2.2787137031555176} 02/25/2022 08:16:25 - INFO - codeparrot_training - Step 23698: {'lr': 0.0002875240434058324, 'samples': 12133888, 'steps': 23698, 'loss/train': 2.455629587173462} 02/25/2022 08:16:31 - INFO - codeparrot_training - Step 23699: {'lr': 0.0002875078662275078, 'samples': 12134400, 'steps': 23699, 'loss/train': 2.860032796859741} 02/25/2022 08:16:34 - INFO - codeparrot_training - Step 23700: {'lr': 0.00028749168888851125, 'samples': 12134912, 'steps': 23700, 'loss/train': 1.3428969383239746} 02/25/2022 08:16:39 - INFO - codeparrot_training - Step 23701: {'lr': 0.0002874755113889123, 'samples': 12135424, 'steps': 23701, 'loss/train': 1.2920112609863281} 02/25/2022 08:16:43 - INFO - codeparrot_training - Step 23702: {'lr': 0.0002874593337287801, 'samples': 12135936, 'steps': 23702, 'loss/train': 1.5097090005874634} 02/25/2022 08:16:49 - INFO - codeparrot_training - Step 23703: {'lr': 0.00028744315590818406, 'samples': 12136448, 'steps': 23703, 'loss/train': 1.324911117553711} 02/25/2022 08:16:53 - INFO - codeparrot_training - Step 23704: {'lr': 0.0002874269779271934, 'samples': 12136960, 'steps': 23704, 'loss/train': 2.5079987049102783} 02/25/2022 08:16:56 - INFO - codeparrot_training - Step 23705: {'lr': 0.0002874107997858775, 'samples': 12137472, 'steps': 23705, 'loss/train': 2.4839026927948} 02/25/2022 08:17:02 - INFO - codeparrot_training - Step 23706: {'lr': 0.00028739462148430554, 'samples': 12137984, 'steps': 23706, 'loss/train': 1.077143669128418} 02/25/2022 08:17:05 - INFO - codeparrot_training - Step 23707: {'lr': 0.0002873784430225469, 'samples': 12138496, 'steps': 23707, 'loss/train': 2.0203192234039307} 02/25/2022 08:17:11 - INFO - codeparrot_training - Step 23708: {'lr': 0.000287362264400671, 'samples': 12139008, 'steps': 23708, 'loss/train': 1.7209244966506958} 02/25/2022 08:17:14 - INFO - codeparrot_training - Step 23709: {'lr': 0.00028734608561874686, 'samples': 12139520, 'steps': 23709, 'loss/train': 1.2826449871063232} 02/25/2022 08:17:20 - INFO - codeparrot_training - Step 23710: {'lr': 0.0002873299066768441, 'samples': 12140032, 'steps': 23710, 'loss/train': 1.8688377141952515} 02/25/2022 08:17:24 - INFO - codeparrot_training - Step 23711: {'lr': 0.0002873137275750317, 'samples': 12140544, 'steps': 23711, 'loss/train': 1.4672523736953735} 02/25/2022 08:17:29 - INFO - codeparrot_training - Step 23712: {'lr': 0.0002872975483133793, 'samples': 12141056, 'steps': 23712, 'loss/train': 2.293625593185425} 02/25/2022 08:17:33 - INFO - codeparrot_training - Step 23713: {'lr': 0.00028728136889195595, 'samples': 12141568, 'steps': 23713, 'loss/train': 1.154145359992981} 02/25/2022 08:17:39 - INFO - codeparrot_training - Step 23714: {'lr': 0.0002872651893108311, 'samples': 12142080, 'steps': 23714, 'loss/train': 1.65375554561615} 02/25/2022 08:17:42 - INFO - codeparrot_training - Step 23715: {'lr': 0.000287249009570074, 'samples': 12142592, 'steps': 23715, 'loss/train': 2.7515904903411865} 02/25/2022 08:17:48 - INFO - codeparrot_training - Step 23716: {'lr': 0.000287232829669754, 'samples': 12143104, 'steps': 23716, 'loss/train': 1.6771210432052612} 02/25/2022 08:17:51 - INFO - codeparrot_training - Step 23717: {'lr': 0.0002872166496099403, 'samples': 12143616, 'steps': 23717, 'loss/train': 2.27978777885437} 02/25/2022 08:17:57 - INFO - codeparrot_training - Step 23718: {'lr': 0.0002872004693907024, 'samples': 12144128, 'steps': 23718, 'loss/train': 1.7869230508804321} 02/25/2022 08:18:00 - INFO - codeparrot_training - Step 23719: {'lr': 0.0002871842890121094, 'samples': 12144640, 'steps': 23719, 'loss/train': 1.5900501012802124} 02/25/2022 08:18:06 - INFO - codeparrot_training - Step 23720: {'lr': 0.0002871681084742308, 'samples': 12145152, 'steps': 23720, 'loss/train': 1.6827583312988281} 02/25/2022 08:18:09 - INFO - codeparrot_training - Step 23721: {'lr': 0.0002871519277771358, 'samples': 12145664, 'steps': 23721, 'loss/train': 2.6957075595855713} 02/25/2022 08:18:15 - INFO - codeparrot_training - Step 23722: {'lr': 0.0002871357469208937, 'samples': 12146176, 'steps': 23722, 'loss/train': 1.01897394657135} 02/25/2022 08:18:18 - INFO - codeparrot_training - Step 23723: {'lr': 0.0002871195659055739, 'samples': 12146688, 'steps': 23723, 'loss/train': 3.624803066253662} 02/25/2022 08:18:25 - INFO - codeparrot_training - Step 23724: {'lr': 0.0002871033847312456, 'samples': 12147200, 'steps': 23724, 'loss/train': 2.4313771724700928} 02/25/2022 08:18:28 - INFO - codeparrot_training - Step 23725: {'lr': 0.0002870872033979782, 'samples': 12147712, 'steps': 23725, 'loss/train': 2.0268619060516357} 02/25/2022 08:18:34 - INFO - codeparrot_training - Step 23726: {'lr': 0.000287071021905841, 'samples': 12148224, 'steps': 23726, 'loss/train': 1.992823839187622} 02/25/2022 08:18:37 - INFO - codeparrot_training - Step 23727: {'lr': 0.00028705484025490333, 'samples': 12148736, 'steps': 23727, 'loss/train': 1.1792811155319214} 02/25/2022 08:18:43 - INFO - codeparrot_training - Step 23728: {'lr': 0.0002870386584452345, 'samples': 12149248, 'steps': 23728, 'loss/train': 1.0752718448638916} 02/25/2022 08:18:46 - INFO - codeparrot_training - Step 23729: {'lr': 0.00028702247647690383, 'samples': 12149760, 'steps': 23729, 'loss/train': 1.4516328573226929} 02/25/2022 08:18:52 - INFO - codeparrot_training - Step 23730: {'lr': 0.0002870062943499806, 'samples': 12150272, 'steps': 23730, 'loss/train': 2.6758906841278076} 02/25/2022 08:18:55 - INFO - codeparrot_training - Step 23731: {'lr': 0.0002869901120645341, 'samples': 12150784, 'steps': 23731, 'loss/train': 1.6887352466583252} 02/25/2022 08:19:01 - INFO - codeparrot_training - Step 23732: {'lr': 0.0002869739296206338, 'samples': 12151296, 'steps': 23732, 'loss/train': 2.6629281044006348} 02/25/2022 08:19:04 - INFO - codeparrot_training - Step 23733: {'lr': 0.0002869577470183489, 'samples': 12151808, 'steps': 23733, 'loss/train': 1.8930538892745972} 02/25/2022 08:19:10 - INFO - codeparrot_training - Step 23734: {'lr': 0.00028694156425774874, 'samples': 12152320, 'steps': 23734, 'loss/train': 1.591787338256836} 02/25/2022 08:19:13 - INFO - codeparrot_training - Step 23735: {'lr': 0.00028692538133890267, 'samples': 12152832, 'steps': 23735, 'loss/train': 0.7416547536849976} 02/25/2022 08:19:19 - INFO - codeparrot_training - Step 23736: {'lr': 0.00028690919826188, 'samples': 12153344, 'steps': 23736, 'loss/train': 2.022280693054199} 02/25/2022 08:19:22 - INFO - codeparrot_training - Step 23737: {'lr': 0.00028689301502674995, 'samples': 12153856, 'steps': 23737, 'loss/train': 1.56686532497406} 02/25/2022 08:19:28 - INFO - codeparrot_training - Step 23738: {'lr': 0.00028687683163358197, 'samples': 12154368, 'steps': 23738, 'loss/train': 2.0775949954986572} 02/25/2022 08:19:31 - INFO - codeparrot_training - Step 23739: {'lr': 0.00028686064808244546, 'samples': 12154880, 'steps': 23739, 'loss/train': 2.261827230453491} 02/25/2022 08:19:38 - INFO - codeparrot_training - Step 23740: {'lr': 0.0002868444643734095, 'samples': 12155392, 'steps': 23740, 'loss/train': 1.521373987197876} 02/25/2022 08:19:41 - INFO - codeparrot_training - Step 23741: {'lr': 0.00028682828050654365, 'samples': 12155904, 'steps': 23741, 'loss/train': 1.4633139371871948} 02/25/2022 08:19:46 - INFO - codeparrot_training - Step 23742: {'lr': 0.0002868120964819171, 'samples': 12156416, 'steps': 23742, 'loss/train': 1.9177744388580322} 02/25/2022 08:19:50 - INFO - codeparrot_training - Step 23743: {'lr': 0.0002867959122995992, 'samples': 12156928, 'steps': 23743, 'loss/train': 2.566213846206665} 02/25/2022 08:19:55 - INFO - codeparrot_training - Step 23744: {'lr': 0.0002867797279596593, 'samples': 12157440, 'steps': 23744, 'loss/train': 1.6049220561981201} 02/25/2022 08:19:59 - INFO - codeparrot_training - Step 23745: {'lr': 0.0002867635434621668, 'samples': 12157952, 'steps': 23745, 'loss/train': 1.4053068161010742} 02/25/2022 08:20:04 - INFO - codeparrot_training - Step 23746: {'lr': 0.0002867473588071909, 'samples': 12158464, 'steps': 23746, 'loss/train': 1.5826157331466675} 02/25/2022 08:20:08 - INFO - codeparrot_training - Step 23747: {'lr': 0.00028673117399480096, 'samples': 12158976, 'steps': 23747, 'loss/train': 1.0591213703155518} 02/25/2022 08:20:13 - INFO - codeparrot_training - Step 23748: {'lr': 0.00028671498902506636, 'samples': 12159488, 'steps': 23748, 'loss/train': 1.4980249404907227} 02/25/2022 08:20:17 - INFO - codeparrot_training - Step 23749: {'lr': 0.00028669880389805647, 'samples': 12160000, 'steps': 23749, 'loss/train': 1.593460202217102} 02/25/2022 08:20:23 - INFO - codeparrot_training - Step 23750: {'lr': 0.00028668261861384045, 'samples': 12160512, 'steps': 23750, 'loss/train': 1.9531420469284058} 02/25/2022 08:20:27 - INFO - codeparrot_training - Step 23751: {'lr': 0.00028666643317248777, 'samples': 12161024, 'steps': 23751, 'loss/train': 1.8310507535934448} 02/25/2022 08:20:32 - INFO - codeparrot_training - Step 23752: {'lr': 0.00028665024757406775, 'samples': 12161536, 'steps': 23752, 'loss/train': 2.0640852451324463} 02/25/2022 08:20:36 - INFO - codeparrot_training - Step 23753: {'lr': 0.0002866340618186497, 'samples': 12162048, 'steps': 23753, 'loss/train': 0.6460937857627869} 02/25/2022 08:20:41 - INFO - codeparrot_training - Step 23754: {'lr': 0.00028661787590630297, 'samples': 12162560, 'steps': 23754, 'loss/train': 3.5005781650543213} 02/25/2022 08:20:45 - INFO - codeparrot_training - Step 23755: {'lr': 0.00028660168983709683, 'samples': 12163072, 'steps': 23755, 'loss/train': 1.9756114482879639} 02/25/2022 08:20:50 - INFO - codeparrot_training - Step 23756: {'lr': 0.00028658550361110075, 'samples': 12163584, 'steps': 23756, 'loss/train': 1.8267971277236938} 02/25/2022 08:20:54 - INFO - codeparrot_training - Step 23757: {'lr': 0.000286569317228384, 'samples': 12164096, 'steps': 23757, 'loss/train': 1.715086817741394} 02/25/2022 08:20:59 - INFO - codeparrot_training - Step 23758: {'lr': 0.00028655313068901586, 'samples': 12164608, 'steps': 23758, 'loss/train': 1.671566367149353} 02/25/2022 08:21:03 - INFO - codeparrot_training - Step 23759: {'lr': 0.0002865369439930657, 'samples': 12165120, 'steps': 23759, 'loss/train': 2.253812313079834} 02/25/2022 08:21:06 - INFO - codeparrot_training - Step 23760: {'lr': 0.00028652075714060294, 'samples': 12165632, 'steps': 23760, 'loss/train': 2.096128463745117} 02/25/2022 08:21:13 - INFO - codeparrot_training - Step 23761: {'lr': 0.0002865045701316968, 'samples': 12166144, 'steps': 23761, 'loss/train': 2.3430652618408203} 02/25/2022 08:21:17 - INFO - codeparrot_training - Step 23762: {'lr': 0.00028648838296641666, 'samples': 12166656, 'steps': 23762, 'loss/train': 0.8878803253173828} 02/25/2022 08:21:22 - INFO - codeparrot_training - Step 23763: {'lr': 0.00028647219564483195, 'samples': 12167168, 'steps': 23763, 'loss/train': 3.3712682723999023} 02/25/2022 08:21:26 - INFO - codeparrot_training - Step 23764: {'lr': 0.00028645600816701186, 'samples': 12167680, 'steps': 23764, 'loss/train': 1.7934439182281494} 02/25/2022 08:21:31 - INFO - codeparrot_training - Step 23765: {'lr': 0.00028643982053302584, 'samples': 12168192, 'steps': 23765, 'loss/train': 1.5670610666275024} 02/25/2022 08:21:35 - INFO - codeparrot_training - Step 23766: {'lr': 0.00028642363274294317, 'samples': 12168704, 'steps': 23766, 'loss/train': 2.365324020385742} 02/25/2022 08:21:40 - INFO - codeparrot_training - Step 23767: {'lr': 0.0002864074447968333, 'samples': 12169216, 'steps': 23767, 'loss/train': 2.1852331161499023} 02/25/2022 08:21:44 - INFO - codeparrot_training - Step 23768: {'lr': 0.0002863912566947654, 'samples': 12169728, 'steps': 23768, 'loss/train': 1.7493542432785034} 02/25/2022 08:21:49 - INFO - codeparrot_training - Step 23769: {'lr': 0.000286375068436809, 'samples': 12170240, 'steps': 23769, 'loss/train': 2.3759031295776367} 02/25/2022 08:21:53 - INFO - codeparrot_training - Step 23770: {'lr': 0.00028635888002303324, 'samples': 12170752, 'steps': 23770, 'loss/train': 1.806571364402771} 02/25/2022 08:21:59 - INFO - codeparrot_training - Step 23771: {'lr': 0.00028634269145350765, 'samples': 12171264, 'steps': 23771, 'loss/train': 1.4930880069732666} 02/25/2022 08:22:04 - INFO - codeparrot_training - Step 23772: {'lr': 0.00028632650272830153, 'samples': 12171776, 'steps': 23772, 'loss/train': 2.107416868209839} 02/25/2022 08:22:08 - INFO - codeparrot_training - Step 23773: {'lr': 0.00028631031384748426, 'samples': 12172288, 'steps': 23773, 'loss/train': 1.9796546697616577} 02/25/2022 08:22:13 - INFO - codeparrot_training - Step 23774: {'lr': 0.000286294124811125, 'samples': 12172800, 'steps': 23774, 'loss/train': 0.703578531742096} 02/25/2022 08:22:17 - INFO - codeparrot_training - Step 23775: {'lr': 0.0002862779356192933, 'samples': 12173312, 'steps': 23775, 'loss/train': 1.4263368844985962} 02/25/2022 08:22:22 - INFO - codeparrot_training - Step 23776: {'lr': 0.0002862617462720584, 'samples': 12173824, 'steps': 23776, 'loss/train': 1.785409927368164} 02/25/2022 08:22:26 - INFO - codeparrot_training - Step 23777: {'lr': 0.00028624555676948975, 'samples': 12174336, 'steps': 23777, 'loss/train': 2.2191672325134277} 02/25/2022 08:22:31 - INFO - codeparrot_training - Step 23778: {'lr': 0.00028622936711165665, 'samples': 12174848, 'steps': 23778, 'loss/train': 2.310695171356201} 02/25/2022 08:22:35 - INFO - codeparrot_training - Step 23779: {'lr': 0.00028621317729862837, 'samples': 12175360, 'steps': 23779, 'loss/train': 3.5981569290161133} 02/25/2022 08:22:38 - INFO - codeparrot_training - Step 23780: {'lr': 0.00028619698733047444, 'samples': 12175872, 'steps': 23780, 'loss/train': 2.387065887451172} 02/25/2022 08:22:44 - INFO - codeparrot_training - Step 23781: {'lr': 0.0002861807972072641, 'samples': 12176384, 'steps': 23781, 'loss/train': 3.240861654281616} 02/25/2022 08:22:49 - INFO - codeparrot_training - Step 23782: {'lr': 0.0002861646069290667, 'samples': 12176896, 'steps': 23782, 'loss/train': 2.517465114593506} 02/25/2022 08:22:53 - INFO - codeparrot_training - Step 23783: {'lr': 0.0002861484164959515, 'samples': 12177408, 'steps': 23783, 'loss/train': 2.3863937854766846} 02/25/2022 08:22:58 - INFO - codeparrot_training - Step 23784: {'lr': 0.0002861322259079881, 'samples': 12177920, 'steps': 23784, 'loss/train': 2.1893911361694336} 02/25/2022 08:23:02 - INFO - codeparrot_training - Step 23785: {'lr': 0.00028611603516524566, 'samples': 12178432, 'steps': 23785, 'loss/train': 1.6996592283248901} 02/25/2022 08:23:08 - INFO - codeparrot_training - Step 23786: {'lr': 0.00028609984426779364, 'samples': 12178944, 'steps': 23786, 'loss/train': 1.5738521814346313} 02/25/2022 08:23:11 - INFO - codeparrot_training - Step 23787: {'lr': 0.0002860836532157012, 'samples': 12179456, 'steps': 23787, 'loss/train': 2.2220871448516846} 02/25/2022 08:23:17 - INFO - codeparrot_training - Step 23788: {'lr': 0.000286067462009038, 'samples': 12179968, 'steps': 23788, 'loss/train': 2.3630354404449463} 02/25/2022 08:23:20 - INFO - codeparrot_training - Step 23789: {'lr': 0.00028605127064787315, 'samples': 12180480, 'steps': 23789, 'loss/train': 2.5233869552612305} 02/25/2022 08:23:26 - INFO - codeparrot_training - Step 23790: {'lr': 0.00028603507913227617, 'samples': 12180992, 'steps': 23790, 'loss/train': 0.14146028459072113} 02/25/2022 08:23:29 - INFO - codeparrot_training - Step 23791: {'lr': 0.00028601888746231633, 'samples': 12181504, 'steps': 23791, 'loss/train': 1.0562998056411743} 02/25/2022 08:23:35 - INFO - codeparrot_training - Step 23792: {'lr': 0.00028600269563806304, 'samples': 12182016, 'steps': 23792, 'loss/train': 2.630549907684326} 02/25/2022 08:23:39 - INFO - codeparrot_training - Step 23793: {'lr': 0.0002859865036595856, 'samples': 12182528, 'steps': 23793, 'loss/train': 1.8632588386535645} 02/25/2022 08:23:44 - INFO - codeparrot_training - Step 23794: {'lr': 0.0002859703115269534, 'samples': 12183040, 'steps': 23794, 'loss/train': 1.6509627103805542} 02/25/2022 08:23:48 - INFO - codeparrot_training - Step 23795: {'lr': 0.0002859541192402359, 'samples': 12183552, 'steps': 23795, 'loss/train': 0.23247288167476654} 02/25/2022 08:23:54 - INFO - codeparrot_training - Step 23796: {'lr': 0.00028593792679950227, 'samples': 12184064, 'steps': 23796, 'loss/train': 2.736579418182373} 02/25/2022 08:23:57 - INFO - codeparrot_training - Step 23797: {'lr': 0.00028592173420482206, 'samples': 12184576, 'steps': 23797, 'loss/train': 2.8215363025665283} 02/25/2022 08:24:03 - INFO - codeparrot_training - Step 23798: {'lr': 0.0002859055414562644, 'samples': 12185088, 'steps': 23798, 'loss/train': 1.9680349826812744} 02/25/2022 08:24:06 - INFO - codeparrot_training - Step 23799: {'lr': 0.00028588934855389885, 'samples': 12185600, 'steps': 23799, 'loss/train': 1.837709665298462} 02/25/2022 08:24:14 - INFO - codeparrot_training - Step 23800: {'lr': 0.0002858731554977948, 'samples': 12186112, 'steps': 23800, 'loss/train': 1.3905572891235352} 02/25/2022 08:24:17 - INFO - codeparrot_training - Step 23801: {'lr': 0.00028585696228802153, 'samples': 12186624, 'steps': 23801, 'loss/train': 1.984190821647644} 02/25/2022 08:24:23 - INFO - codeparrot_training - Step 23802: {'lr': 0.0002858407689246484, 'samples': 12187136, 'steps': 23802, 'loss/train': 2.0714988708496094} 02/25/2022 08:24:26 - INFO - codeparrot_training - Step 23803: {'lr': 0.0002858245754077448, 'samples': 12187648, 'steps': 23803, 'loss/train': 0.27793917059898376} 02/25/2022 08:24:32 - INFO - codeparrot_training - Step 23804: {'lr': 0.0002858083817373801, 'samples': 12188160, 'steps': 23804, 'loss/train': 1.0633853673934937} 02/25/2022 08:24:35 - INFO - codeparrot_training - Step 23805: {'lr': 0.00028579218791362367, 'samples': 12188672, 'steps': 23805, 'loss/train': 2.295172691345215} 02/25/2022 08:24:41 - INFO - codeparrot_training - Step 23806: {'lr': 0.0002857759939365449, 'samples': 12189184, 'steps': 23806, 'loss/train': 7.6252875328063965} 02/25/2022 08:24:44 - INFO - codeparrot_training - Step 23807: {'lr': 0.000285759799806213, 'samples': 12189696, 'steps': 23807, 'loss/train': 1.5053744316101074} 02/25/2022 08:24:51 - INFO - codeparrot_training - Step 23808: {'lr': 0.00028574360552269755, 'samples': 12190208, 'steps': 23808, 'loss/train': 1.590543270111084} 02/25/2022 08:24:54 - INFO - codeparrot_training - Step 23809: {'lr': 0.0002857274110860679, 'samples': 12190720, 'steps': 23809, 'loss/train': 2.0183393955230713} 02/25/2022 08:24:57 - INFO - codeparrot_training - Step 23810: {'lr': 0.00028571121649639337, 'samples': 12191232, 'steps': 23810, 'loss/train': 2.259247064590454} 02/25/2022 08:25:05 - INFO - codeparrot_training - Step 23811: {'lr': 0.0002856950217537432, 'samples': 12191744, 'steps': 23811, 'loss/train': 2.4867947101593018} 02/25/2022 08:25:08 - INFO - codeparrot_training - Step 23812: {'lr': 0.000285678826858187, 'samples': 12192256, 'steps': 23812, 'loss/train': 2.487729072570801} 02/25/2022 08:25:14 - INFO - codeparrot_training - Step 23813: {'lr': 0.000285662631809794, 'samples': 12192768, 'steps': 23813, 'loss/train': 3.44779372215271} 02/25/2022 08:25:18 - INFO - codeparrot_training - Step 23814: {'lr': 0.0002856464366086336, 'samples': 12193280, 'steps': 23814, 'loss/train': 0.6749687194824219} 02/25/2022 08:25:23 - INFO - codeparrot_training - Step 23815: {'lr': 0.0002856302412547752, 'samples': 12193792, 'steps': 23815, 'loss/train': 1.9551236629486084} 02/25/2022 08:25:27 - INFO - codeparrot_training - Step 23816: {'lr': 0.0002856140457482882, 'samples': 12194304, 'steps': 23816, 'loss/train': 1.7070772647857666} 02/25/2022 08:25:32 - INFO - codeparrot_training - Step 23817: {'lr': 0.0002855978500892419, 'samples': 12194816, 'steps': 23817, 'loss/train': 0.27305731177330017} 02/25/2022 08:25:36 - INFO - codeparrot_training - Step 23818: {'lr': 0.00028558165427770567, 'samples': 12195328, 'steps': 23818, 'loss/train': 1.9458577632904053} 02/25/2022 08:25:41 - INFO - codeparrot_training - Step 23819: {'lr': 0.00028556545831374903, 'samples': 12195840, 'steps': 23819, 'loss/train': 1.327791452407837} 02/25/2022 08:25:45 - INFO - codeparrot_training - Step 23820: {'lr': 0.0002855492621974411, 'samples': 12196352, 'steps': 23820, 'loss/train': 1.8198851346969604} 02/25/2022 08:25:52 - INFO - codeparrot_training - Step 23821: {'lr': 0.0002855330659288516, 'samples': 12196864, 'steps': 23821, 'loss/train': 0.4827019274234772} 02/25/2022 08:25:56 - INFO - codeparrot_training - Step 23822: {'lr': 0.00028551686950804964, 'samples': 12197376, 'steps': 23822, 'loss/train': 1.7372627258300781} 02/25/2022 08:26:01 - INFO - codeparrot_training - Step 23823: {'lr': 0.0002855006729351046, 'samples': 12197888, 'steps': 23823, 'loss/train': 1.6372865438461304} 02/25/2022 08:26:05 - INFO - codeparrot_training - Step 23824: {'lr': 0.0002854844762100861, 'samples': 12198400, 'steps': 23824, 'loss/train': 2.4423463344573975} 02/25/2022 08:26:10 - INFO - codeparrot_training - Step 23825: {'lr': 0.0002854682793330633, 'samples': 12198912, 'steps': 23825, 'loss/train': 2.299920082092285} 02/25/2022 08:26:14 - INFO - codeparrot_training - Step 23826: {'lr': 0.0002854520823041057, 'samples': 12199424, 'steps': 23826, 'loss/train': 1.8253332376480103} 02/25/2022 08:26:19 - INFO - codeparrot_training - Step 23827: {'lr': 0.0002854358851232826, 'samples': 12199936, 'steps': 23827, 'loss/train': 2.4432296752929688} 02/25/2022 08:26:23 - INFO - codeparrot_training - Step 23828: {'lr': 0.0002854196877906635, 'samples': 12200448, 'steps': 23828, 'loss/train': 1.1840457916259766} 02/25/2022 08:26:28 - INFO - codeparrot_training - Step 23829: {'lr': 0.00028540349030631756, 'samples': 12200960, 'steps': 23829, 'loss/train': 2.380634069442749} 02/25/2022 08:26:32 - INFO - codeparrot_training - Step 23830: {'lr': 0.0002853872926703144, 'samples': 12201472, 'steps': 23830, 'loss/train': 2.720336675643921} 02/25/2022 08:26:39 - INFO - codeparrot_training - Step 23831: {'lr': 0.0002853710948827233, 'samples': 12201984, 'steps': 23831, 'loss/train': 2.087082624435425} 02/25/2022 08:26:43 - INFO - codeparrot_training - Step 23832: {'lr': 0.00028535489694361365, 'samples': 12202496, 'steps': 23832, 'loss/train': 0.8275420069694519} 02/25/2022 08:26:48 - INFO - codeparrot_training - Step 23833: {'lr': 0.00028533869885305485, 'samples': 12203008, 'steps': 23833, 'loss/train': 1.8307034969329834} 02/25/2022 08:26:52 - INFO - codeparrot_training - Step 23834: {'lr': 0.0002853225006111163, 'samples': 12203520, 'steps': 23834, 'loss/train': 2.4690442085266113} 02/25/2022 08:26:57 - INFO - codeparrot_training - Step 23835: {'lr': 0.00028530630221786736, 'samples': 12204032, 'steps': 23835, 'loss/train': 1.9603184461593628} 02/25/2022 08:27:01 - INFO - codeparrot_training - Step 23836: {'lr': 0.00028529010367337745, 'samples': 12204544, 'steps': 23836, 'loss/train': 2.428802490234375} 02/25/2022 08:27:06 - INFO - codeparrot_training - Step 23837: {'lr': 0.0002852739049777159, 'samples': 12205056, 'steps': 23837, 'loss/train': 2.112835645675659} 02/25/2022 08:27:10 - INFO - codeparrot_training - Step 23838: {'lr': 0.00028525770613095213, 'samples': 12205568, 'steps': 23838, 'loss/train': 1.6155688762664795} 02/25/2022 08:27:15 - INFO - codeparrot_training - Step 23839: {'lr': 0.00028524150713315566, 'samples': 12206080, 'steps': 23839, 'loss/train': 8.739619255065918} 02/25/2022 08:27:19 - INFO - codeparrot_training - Step 23840: {'lr': 0.00028522530798439564, 'samples': 12206592, 'steps': 23840, 'loss/train': 1.8128621578216553} 02/25/2022 08:27:24 - INFO - codeparrot_training - Step 23841: {'lr': 0.0002852091086847417, 'samples': 12207104, 'steps': 23841, 'loss/train': 2.49391770362854} 02/25/2022 08:27:28 - INFO - codeparrot_training - Step 23842: {'lr': 0.000285192909234263, 'samples': 12207616, 'steps': 23842, 'loss/train': 1.474942684173584} 02/25/2022 08:27:35 - INFO - codeparrot_training - Step 23843: {'lr': 0.0002851767096330291, 'samples': 12208128, 'steps': 23843, 'loss/train': 2.480138063430786} 02/25/2022 08:27:39 - INFO - codeparrot_training - Step 23844: {'lr': 0.00028516050988110935, 'samples': 12208640, 'steps': 23844, 'loss/train': 2.6836564540863037} 02/25/2022 08:27:44 - INFO - codeparrot_training - Step 23845: {'lr': 0.00028514430997857317, 'samples': 12209152, 'steps': 23845, 'loss/train': 1.6603347063064575} 02/25/2022 08:27:48 - INFO - codeparrot_training - Step 23846: {'lr': 0.0002851281099254899, 'samples': 12209664, 'steps': 23846, 'loss/train': 0.36443275213241577} 02/25/2022 08:27:53 - INFO - codeparrot_training - Step 23847: {'lr': 0.000285111909721929, 'samples': 12210176, 'steps': 23847, 'loss/train': 2.6149418354034424} 02/25/2022 08:27:57 - INFO - codeparrot_training - Step 23848: {'lr': 0.0002850957093679597, 'samples': 12210688, 'steps': 23848, 'loss/train': 0.4315800368785858} 02/25/2022 08:28:02 - INFO - codeparrot_training - Step 23849: {'lr': 0.0002850795088636516, 'samples': 12211200, 'steps': 23849, 'loss/train': 2.4588608741760254} 02/25/2022 08:28:06 - INFO - codeparrot_training - Step 23850: {'lr': 0.0002850633082090741, 'samples': 12211712, 'steps': 23850, 'loss/train': 1.8169078826904297} 02/25/2022 08:28:12 - INFO - codeparrot_training - Step 23851: {'lr': 0.00028504710740429647, 'samples': 12212224, 'steps': 23851, 'loss/train': 0.9353863596916199} 02/25/2022 08:28:15 - INFO - codeparrot_training - Step 23852: {'lr': 0.0002850309064493882, 'samples': 12212736, 'steps': 23852, 'loss/train': 8.571444511413574} 02/25/2022 08:28:21 - INFO - codeparrot_training - Step 23853: {'lr': 0.00028501470534441855, 'samples': 12213248, 'steps': 23853, 'loss/train': 2.3966128826141357} 02/25/2022 08:28:24 - INFO - codeparrot_training - Step 23854: {'lr': 0.00028499850408945704, 'samples': 12213760, 'steps': 23854, 'loss/train': 1.5944530963897705} 02/25/2022 08:28:30 - INFO - codeparrot_training - Step 23855: {'lr': 0.0002849823026845731, 'samples': 12214272, 'steps': 23855, 'loss/train': 1.1452698707580566} 02/25/2022 08:28:33 - INFO - codeparrot_training - Step 23856: {'lr': 0.00028496610112983605, 'samples': 12214784, 'steps': 23856, 'loss/train': 1.9814021587371826} 02/25/2022 08:28:40 - INFO - codeparrot_training - Step 23857: {'lr': 0.00028494989942531537, 'samples': 12215296, 'steps': 23857, 'loss/train': 1.8425670862197876} 02/25/2022 08:28:44 - INFO - codeparrot_training - Step 23858: {'lr': 0.0002849336975710804, 'samples': 12215808, 'steps': 23858, 'loss/train': 2.0652003288269043} 02/25/2022 08:28:50 - INFO - codeparrot_training - Step 23859: {'lr': 0.0002849174955672005, 'samples': 12216320, 'steps': 23859, 'loss/train': 2.181178092956543} 02/25/2022 08:28:53 - INFO - codeparrot_training - Step 23860: {'lr': 0.0002849012934137452, 'samples': 12216832, 'steps': 23860, 'loss/train': 2.166696786880493} 02/25/2022 08:28:59 - INFO - codeparrot_training - Step 23861: {'lr': 0.0002848850911107838, 'samples': 12217344, 'steps': 23861, 'loss/train': 2.069958209991455} 02/25/2022 08:29:02 - INFO - codeparrot_training - Step 23862: {'lr': 0.0002848688886583858, 'samples': 12217856, 'steps': 23862, 'loss/train': 2.1374595165252686} 02/25/2022 08:29:08 - INFO - codeparrot_training - Step 23863: {'lr': 0.0002848526860566205, 'samples': 12218368, 'steps': 23863, 'loss/train': 2.445357084274292} 02/25/2022 08:29:11 - INFO - codeparrot_training - Step 23864: {'lr': 0.00028483648330555737, 'samples': 12218880, 'steps': 23864, 'loss/train': 0.9217566251754761} 02/25/2022 08:29:17 - INFO - codeparrot_training - Step 23865: {'lr': 0.0002848202804052659, 'samples': 12219392, 'steps': 23865, 'loss/train': 1.1582810878753662} 02/25/2022 08:29:20 - INFO - codeparrot_training - Step 23866: {'lr': 0.00028480407735581527, 'samples': 12219904, 'steps': 23866, 'loss/train': 1.6610405445098877} 02/25/2022 08:29:27 - INFO - codeparrot_training - Step 23867: {'lr': 0.00028478787415727515, 'samples': 12220416, 'steps': 23867, 'loss/train': 2.0354859828948975} 02/25/2022 08:29:31 - INFO - codeparrot_training - Step 23868: {'lr': 0.00028477167080971465, 'samples': 12220928, 'steps': 23868, 'loss/train': 2.06433367729187} 02/25/2022 08:29:36 - INFO - codeparrot_training - Step 23869: {'lr': 0.0002847554673132035, 'samples': 12221440, 'steps': 23869, 'loss/train': 2.5286731719970703} 02/25/2022 08:29:40 - INFO - codeparrot_training - Step 23870: {'lr': 0.00028473926366781095, 'samples': 12221952, 'steps': 23870, 'loss/train': 2.0785348415374756} 02/25/2022 08:29:46 - INFO - codeparrot_training - Step 23871: {'lr': 0.00028472305987360636, 'samples': 12222464, 'steps': 23871, 'loss/train': 3.007127523422241} 02/25/2022 08:29:50 - INFO - codeparrot_training - Step 23872: {'lr': 0.00028470685593065927, 'samples': 12222976, 'steps': 23872, 'loss/train': 1.1006990671157837} 02/25/2022 08:29:53 - INFO - codeparrot_training - Step 23873: {'lr': 0.00028469065183903895, 'samples': 12223488, 'steps': 23873, 'loss/train': 2.1557517051696777} 02/25/2022 08:29:59 - INFO - codeparrot_training - Step 23874: {'lr': 0.000284674447598815, 'samples': 12224000, 'steps': 23874, 'loss/train': 1.7379764318466187} 02/25/2022 08:30:02 - INFO - codeparrot_training - Step 23875: {'lr': 0.00028465824321005667, 'samples': 12224512, 'steps': 23875, 'loss/train': 1.7558883428573608} 02/25/2022 08:30:08 - INFO - codeparrot_training - Step 23876: {'lr': 0.0002846420386728334, 'samples': 12225024, 'steps': 23876, 'loss/train': 2.2542593479156494} 02/25/2022 08:30:11 - INFO - codeparrot_training - Step 23877: {'lr': 0.0002846258339872147, 'samples': 12225536, 'steps': 23877, 'loss/train': 1.8808306455612183} 02/25/2022 08:30:18 - INFO - codeparrot_training - Step 23878: {'lr': 0.0002846096291532699, 'samples': 12226048, 'steps': 23878, 'loss/train': 1.2494032382965088} 02/25/2022 08:30:22 - INFO - codeparrot_training - Step 23879: {'lr': 0.0002845934241710684, 'samples': 12226560, 'steps': 23879, 'loss/train': 0.8155087828636169} 02/25/2022 08:30:27 - INFO - codeparrot_training - Step 23880: {'lr': 0.0002845772190406798, 'samples': 12227072, 'steps': 23880, 'loss/train': 0.6408429145812988} 02/25/2022 08:30:31 - INFO - codeparrot_training - Step 23881: {'lr': 0.0002845610137621732, 'samples': 12227584, 'steps': 23881, 'loss/train': 3.0057241916656494} 02/25/2022 08:30:36 - INFO - codeparrot_training - Step 23882: {'lr': 0.0002845448083356183, 'samples': 12228096, 'steps': 23882, 'loss/train': 0.3577134311199188} 02/25/2022 08:30:40 - INFO - codeparrot_training - Step 23883: {'lr': 0.00028452860276108436, 'samples': 12228608, 'steps': 23883, 'loss/train': 1.316670536994934} 02/25/2022 08:30:46 - INFO - codeparrot_training - Step 23884: {'lr': 0.0002845123970386408, 'samples': 12229120, 'steps': 23884, 'loss/train': 1.7536288499832153} 02/25/2022 08:30:49 - INFO - codeparrot_training - Step 23885: {'lr': 0.00028449619116835715, 'samples': 12229632, 'steps': 23885, 'loss/train': 2.04780650138855} 02/25/2022 08:30:55 - INFO - codeparrot_training - Step 23886: {'lr': 0.0002844799851503028, 'samples': 12230144, 'steps': 23886, 'loss/train': 1.4486092329025269} 02/25/2022 08:30:58 - INFO - codeparrot_training - Step 23887: {'lr': 0.0002844637789845471, 'samples': 12230656, 'steps': 23887, 'loss/train': 1.6248226165771484} 02/25/2022 08:31:04 - INFO - codeparrot_training - Step 23888: {'lr': 0.0002844475726711595, 'samples': 12231168, 'steps': 23888, 'loss/train': 1.3015556335449219} 02/25/2022 08:31:07 - INFO - codeparrot_training - Step 23889: {'lr': 0.0002844313662102095, 'samples': 12231680, 'steps': 23889, 'loss/train': 1.826593041419983} 02/25/2022 08:31:13 - INFO - codeparrot_training - Step 23890: {'lr': 0.0002844151596017665, 'samples': 12232192, 'steps': 23890, 'loss/train': 1.816049575805664} 02/25/2022 08:31:16 - INFO - codeparrot_training - Step 23891: {'lr': 0.0002843989528458997, 'samples': 12232704, 'steps': 23891, 'loss/train': 1.9060184955596924} 02/25/2022 08:31:22 - INFO - codeparrot_training - Step 23892: {'lr': 0.0002843827459426789, 'samples': 12233216, 'steps': 23892, 'loss/train': 2.3060052394866943} 02/25/2022 08:31:25 - INFO - codeparrot_training - Step 23893: {'lr': 0.00028436653889217316, 'samples': 12233728, 'steps': 23893, 'loss/train': 1.5798581838607788} 02/25/2022 08:31:33 - INFO - codeparrot_training - Step 23894: {'lr': 0.00028435033169445223, 'samples': 12234240, 'steps': 23894, 'loss/train': 0.9531805515289307} 02/25/2022 08:31:36 - INFO - codeparrot_training - Step 23895: {'lr': 0.0002843341243495853, 'samples': 12234752, 'steps': 23895, 'loss/train': 1.2489651441574097} 02/25/2022 08:31:42 - INFO - codeparrot_training - Step 23896: {'lr': 0.0002843179168576419, 'samples': 12235264, 'steps': 23896, 'loss/train': 2.063264846801758} 02/25/2022 08:31:45 - INFO - codeparrot_training - Step 23897: {'lr': 0.00028430170921869147, 'samples': 12235776, 'steps': 23897, 'loss/train': 2.0053510665893555} 02/25/2022 08:31:51 - INFO - codeparrot_training - Step 23898: {'lr': 0.0002842855014328034, 'samples': 12236288, 'steps': 23898, 'loss/train': 1.8234277963638306} 02/25/2022 08:31:54 - INFO - codeparrot_training - Step 23899: {'lr': 0.0002842692935000471, 'samples': 12236800, 'steps': 23899, 'loss/train': 2.18677020072937} 02/25/2022 08:31:59 - INFO - codeparrot_training - Step 23900: {'lr': 0.00028425308542049207, 'samples': 12237312, 'steps': 23900, 'loss/train': 1.6970452070236206} 02/25/2022 08:32:03 - INFO - codeparrot_training - Step 23901: {'lr': 0.0002842368771942077, 'samples': 12237824, 'steps': 23901, 'loss/train': 1.960959553718567} 02/25/2022 08:32:08 - INFO - codeparrot_training - Step 23902: {'lr': 0.00028422066882126336, 'samples': 12238336, 'steps': 23902, 'loss/train': 1.6810353994369507} 02/25/2022 08:32:12 - INFO - codeparrot_training - Step 23903: {'lr': 0.0002842044603017285, 'samples': 12238848, 'steps': 23903, 'loss/train': 1.7343754768371582} 02/25/2022 08:32:19 - INFO - codeparrot_training - Step 23904: {'lr': 0.00028418825163567275, 'samples': 12239360, 'steps': 23904, 'loss/train': 2.1843326091766357} 02/25/2022 08:32:23 - INFO - codeparrot_training - Step 23905: {'lr': 0.0002841720428231653, 'samples': 12239872, 'steps': 23905, 'loss/train': 0.16531874239444733} 02/25/2022 08:32:28 - INFO - codeparrot_training - Step 23906: {'lr': 0.00028415583386427566, 'samples': 12240384, 'steps': 23906, 'loss/train': 1.6627346277236938} 02/25/2022 08:32:32 - INFO - codeparrot_training - Step 23907: {'lr': 0.0002841396247590733, 'samples': 12240896, 'steps': 23907, 'loss/train': 1.393295407295227} 02/25/2022 08:32:38 - INFO - codeparrot_training - Step 23908: {'lr': 0.00028412341550762755, 'samples': 12241408, 'steps': 23908, 'loss/train': 2.3162336349487305} 02/25/2022 08:32:41 - INFO - codeparrot_training - Step 23909: {'lr': 0.00028410720611000804, 'samples': 12241920, 'steps': 23909, 'loss/train': 1.3320385217666626} 02/25/2022 08:32:47 - INFO - codeparrot_training - Step 23910: {'lr': 0.000284090996566284, 'samples': 12242432, 'steps': 23910, 'loss/train': 2.113345146179199} 02/25/2022 08:32:50 - INFO - codeparrot_training - Step 23911: {'lr': 0.00028407478687652503, 'samples': 12242944, 'steps': 23911, 'loss/train': 3.181025505065918} 02/25/2022 08:32:56 - INFO - codeparrot_training - Step 23912: {'lr': 0.0002840585770408004, 'samples': 12243456, 'steps': 23912, 'loss/train': 2.1168301105499268} 02/25/2022 08:32:59 - INFO - codeparrot_training - Step 23913: {'lr': 0.00028404236705917974, 'samples': 12243968, 'steps': 23913, 'loss/train': 1.8107494115829468} 02/25/2022 08:33:05 - INFO - codeparrot_training - Step 23914: {'lr': 0.00028402615693173236, 'samples': 12244480, 'steps': 23914, 'loss/train': 2.187854290008545} 02/25/2022 08:33:08 - INFO - codeparrot_training - Step 23915: {'lr': 0.00028400994665852777, 'samples': 12244992, 'steps': 23915, 'loss/train': 4.253845691680908} 02/25/2022 08:33:15 - INFO - codeparrot_training - Step 23916: {'lr': 0.00028399373623963525, 'samples': 12245504, 'steps': 23916, 'loss/train': 2.3905253410339355} 02/25/2022 08:33:19 - INFO - codeparrot_training - Step 23917: {'lr': 0.0002839775256751244, 'samples': 12246016, 'steps': 23917, 'loss/train': 1.4984098672866821} 02/25/2022 08:33:24 - INFO - codeparrot_training - Step 23918: {'lr': 0.00028396131496506466, 'samples': 12246528, 'steps': 23918, 'loss/train': 0.7051137089729309} 02/25/2022 08:33:28 - INFO - codeparrot_training - Step 23919: {'lr': 0.00028394510410952544, 'samples': 12247040, 'steps': 23919, 'loss/train': 1.7828171253204346} 02/25/2022 08:33:33 - INFO - codeparrot_training - Step 23920: {'lr': 0.0002839288931085761, 'samples': 12247552, 'steps': 23920, 'loss/train': 1.4712936878204346} 02/25/2022 08:33:37 - INFO - codeparrot_training - Step 23921: {'lr': 0.0002839126819622862, 'samples': 12248064, 'steps': 23921, 'loss/train': 2.1698801517486572} 02/25/2022 08:33:42 - INFO - codeparrot_training - Step 23922: {'lr': 0.00028389647067072517, 'samples': 12248576, 'steps': 23922, 'loss/train': 1.0560258626937866} 02/25/2022 08:33:46 - INFO - codeparrot_training - Step 23923: {'lr': 0.00028388025923396234, 'samples': 12249088, 'steps': 23923, 'loss/train': 1.5148906707763672} 02/25/2022 08:33:53 - INFO - codeparrot_training - Step 23924: {'lr': 0.0002838640476520673, 'samples': 12249600, 'steps': 23924, 'loss/train': 1.8574568033218384} 02/25/2022 08:33:57 - INFO - codeparrot_training - Step 23925: {'lr': 0.00028384783592510945, 'samples': 12250112, 'steps': 23925, 'loss/train': 1.7618740797042847} 02/25/2022 08:34:02 - INFO - codeparrot_training - Step 23926: {'lr': 0.00028383162405315823, 'samples': 12250624, 'steps': 23926, 'loss/train': 0.8848698139190674} 02/25/2022 08:34:06 - INFO - codeparrot_training - Step 23927: {'lr': 0.00028381541203628295, 'samples': 12251136, 'steps': 23927, 'loss/train': 1.5072880983352661} 02/25/2022 08:34:11 - INFO - codeparrot_training - Step 23928: {'lr': 0.0002837991998745533, 'samples': 12251648, 'steps': 23928, 'loss/train': 1.8419615030288696} 02/25/2022 08:34:15 - INFO - codeparrot_training - Step 23929: {'lr': 0.0002837829875680386, 'samples': 12252160, 'steps': 23929, 'loss/train': 2.556342124938965} 02/25/2022 08:34:20 - INFO - codeparrot_training - Step 23930: {'lr': 0.00028376677511680827, 'samples': 12252672, 'steps': 23930, 'loss/train': 2.056863784790039} 02/25/2022 08:34:24 - INFO - codeparrot_training - Step 23931: {'lr': 0.0002837505625209318, 'samples': 12253184, 'steps': 23931, 'loss/train': 2.4866621494293213} 02/25/2022 08:34:29 - INFO - codeparrot_training - Step 23932: {'lr': 0.0002837343497804787, 'samples': 12253696, 'steps': 23932, 'loss/train': 1.559464693069458} 02/25/2022 08:34:33 - INFO - codeparrot_training - Step 23933: {'lr': 0.0002837181368955183, 'samples': 12254208, 'steps': 23933, 'loss/train': 1.319364070892334} 02/25/2022 08:34:38 - INFO - codeparrot_training - Step 23934: {'lr': 0.0002837019238661201, 'samples': 12254720, 'steps': 23934, 'loss/train': 2.5055911540985107} 02/25/2022 08:34:42 - INFO - codeparrot_training - Step 23935: {'lr': 0.00028368571069235354, 'samples': 12255232, 'steps': 23935, 'loss/train': 0.5492614507675171} 02/25/2022 08:34:47 - INFO - codeparrot_training - Step 23936: {'lr': 0.00028366949737428814, 'samples': 12255744, 'steps': 23936, 'loss/train': 2.809084892272949} 02/25/2022 08:34:51 - INFO - codeparrot_training - Step 23937: {'lr': 0.00028365328391199334, 'samples': 12256256, 'steps': 23937, 'loss/train': 1.8783550262451172} 02/25/2022 08:34:56 - INFO - codeparrot_training - Step 23938: {'lr': 0.0002836370703055385, 'samples': 12256768, 'steps': 23938, 'loss/train': 2.457873821258545} 02/25/2022 08:35:00 - INFO - codeparrot_training - Step 23939: {'lr': 0.0002836208565549932, 'samples': 12257280, 'steps': 23939, 'loss/train': 1.9005979299545288} 02/25/2022 08:35:07 - INFO - codeparrot_training - Step 23940: {'lr': 0.00028360464266042674, 'samples': 12257792, 'steps': 23940, 'loss/train': 1.6991955041885376} 02/25/2022 08:35:10 - INFO - codeparrot_training - Step 23941: {'lr': 0.00028358842862190873, 'samples': 12258304, 'steps': 23941, 'loss/train': 2.1745004653930664} 02/25/2022 08:35:16 - INFO - codeparrot_training - Step 23942: {'lr': 0.00028357221443950847, 'samples': 12258816, 'steps': 23942, 'loss/train': 2.757969617843628} 02/25/2022 08:35:19 - INFO - codeparrot_training - Step 23943: {'lr': 0.00028355600011329557, 'samples': 12259328, 'steps': 23943, 'loss/train': 1.3630497455596924} 02/25/2022 08:35:25 - INFO - codeparrot_training - Step 23944: {'lr': 0.00028353978564333936, 'samples': 12259840, 'steps': 23944, 'loss/train': 1.3697361946105957} 02/25/2022 08:35:28 - INFO - codeparrot_training - Step 23945: {'lr': 0.0002835235710297094, 'samples': 12260352, 'steps': 23945, 'loss/train': 1.5021661520004272} 02/25/2022 08:35:34 - INFO - codeparrot_training - Step 23946: {'lr': 0.0002835073562724751, 'samples': 12260864, 'steps': 23946, 'loss/train': 1.128860354423523} 02/25/2022 08:35:38 - INFO - codeparrot_training - Step 23947: {'lr': 0.00028349114137170593, 'samples': 12261376, 'steps': 23947, 'loss/train': 0.5817089676856995} 02/25/2022 08:35:43 - INFO - codeparrot_training - Step 23948: {'lr': 0.0002834749263274714, 'samples': 12261888, 'steps': 23948, 'loss/train': 1.8628727197647095} 02/25/2022 08:35:47 - INFO - codeparrot_training - Step 23949: {'lr': 0.00028345871113984086, 'samples': 12262400, 'steps': 23949, 'loss/train': 1.4499738216400146} 02/25/2022 08:35:54 - INFO - codeparrot_training - Step 23950: {'lr': 0.0002834424958088838, 'samples': 12262912, 'steps': 23950, 'loss/train': 1.1090614795684814} 02/25/2022 08:35:57 - INFO - codeparrot_training - Step 23951: {'lr': 0.00028342628033466974, 'samples': 12263424, 'steps': 23951, 'loss/train': 1.9155235290527344} 02/25/2022 08:36:03 - INFO - codeparrot_training - Step 23952: {'lr': 0.00028341006471726816, 'samples': 12263936, 'steps': 23952, 'loss/train': 1.125829815864563} 02/25/2022 08:36:06 - INFO - codeparrot_training - Step 23953: {'lr': 0.0002833938489567484, 'samples': 12264448, 'steps': 23953, 'loss/train': 1.8911327123641968} 02/25/2022 08:36:12 - INFO - codeparrot_training - Step 23954: {'lr': 0.00028337763305318, 'samples': 12264960, 'steps': 23954, 'loss/train': 2.3532049655914307} 02/25/2022 08:36:15 - INFO - codeparrot_training - Step 23955: {'lr': 0.00028336141700663244, 'samples': 12265472, 'steps': 23955, 'loss/train': 2.2518019676208496} 02/25/2022 08:36:21 - INFO - codeparrot_training - Step 23956: {'lr': 0.00028334520081717507, 'samples': 12265984, 'steps': 23956, 'loss/train': 1.2189196348190308} 02/25/2022 08:36:24 - INFO - codeparrot_training - Step 23957: {'lr': 0.0002833289844848776, 'samples': 12266496, 'steps': 23957, 'loss/train': 1.7994390726089478} 02/25/2022 08:36:30 - INFO - codeparrot_training - Step 23958: {'lr': 0.0002833127680098092, 'samples': 12267008, 'steps': 23958, 'loss/train': 0.6896646618843079} 02/25/2022 08:36:33 - INFO - codeparrot_training - Step 23959: {'lr': 0.0002832965513920396, 'samples': 12267520, 'steps': 23959, 'loss/train': 1.7736015319824219} 02/25/2022 08:36:39 - INFO - codeparrot_training - Step 23960: {'lr': 0.0002832803346316381, 'samples': 12268032, 'steps': 23960, 'loss/train': 2.5266411304473877} 02/25/2022 08:36:42 - INFO - codeparrot_training - Step 23961: {'lr': 0.0002832641177286742, 'samples': 12268544, 'steps': 23961, 'loss/train': 1.6412752866744995} 02/25/2022 08:36:50 - INFO - codeparrot_training - Step 23962: {'lr': 0.0002832479006832174, 'samples': 12269056, 'steps': 23962, 'loss/train': 1.9535921812057495} 02/25/2022 08:36:53 - INFO - codeparrot_training - Step 23963: {'lr': 0.0002832316834953372, 'samples': 12269568, 'steps': 23963, 'loss/train': 1.6700303554534912} 02/25/2022 08:36:59 - INFO - codeparrot_training - Step 23964: {'lr': 0.0002832154661651029, 'samples': 12270080, 'steps': 23964, 'loss/train': 1.6835803985595703} 02/25/2022 08:37:02 - INFO - codeparrot_training - Step 23965: {'lr': 0.00028319924869258425, 'samples': 12270592, 'steps': 23965, 'loss/train': 1.9687668085098267} 02/25/2022 08:37:08 - INFO - codeparrot_training - Step 23966: {'lr': 0.0002831830310778504, 'samples': 12271104, 'steps': 23966, 'loss/train': 1.5070534944534302} 02/25/2022 08:37:13 - INFO - codeparrot_training - Step 23967: {'lr': 0.0002831668133209711, 'samples': 12271616, 'steps': 23967, 'loss/train': 1.837750792503357} 02/25/2022 08:37:17 - INFO - codeparrot_training - Step 23968: {'lr': 0.0002831505954220156, 'samples': 12272128, 'steps': 23968, 'loss/train': 1.968056559562683} 02/25/2022 08:37:22 - INFO - codeparrot_training - Step 23969: {'lr': 0.00028313437738105353, 'samples': 12272640, 'steps': 23969, 'loss/train': 1.7962291240692139} 02/25/2022 08:37:26 - INFO - codeparrot_training - Step 23970: {'lr': 0.0002831181591981543, 'samples': 12273152, 'steps': 23970, 'loss/train': 2.4770469665527344} 02/25/2022 08:37:33 - INFO - codeparrot_training - Step 23971: {'lr': 0.0002831019408733874, 'samples': 12273664, 'steps': 23971, 'loss/train': 2.0747499465942383} 02/25/2022 08:37:36 - INFO - codeparrot_training - Step 23972: {'lr': 0.00028308572240682233, 'samples': 12274176, 'steps': 23972, 'loss/train': 0.2593514919281006} 02/25/2022 08:37:42 - INFO - codeparrot_training - Step 23973: {'lr': 0.00028306950379852844, 'samples': 12274688, 'steps': 23973, 'loss/train': 2.389361619949341} 02/25/2022 08:37:45 - INFO - codeparrot_training - Step 23974: {'lr': 0.0002830532850485754, 'samples': 12275200, 'steps': 23974, 'loss/train': 2.4268078804016113} 02/25/2022 08:37:51 - INFO - codeparrot_training - Step 23975: {'lr': 0.0002830370661570325, 'samples': 12275712, 'steps': 23975, 'loss/train': 1.82704496383667} 02/25/2022 08:37:54 - INFO - codeparrot_training - Step 23976: {'lr': 0.00028302084712396937, 'samples': 12276224, 'steps': 23976, 'loss/train': 1.6886565685272217} 02/25/2022 08:38:00 - INFO - codeparrot_training - Step 23977: {'lr': 0.00028300462794945535, 'samples': 12276736, 'steps': 23977, 'loss/train': 2.3774254322052} 02/25/2022 08:38:03 - INFO - codeparrot_training - Step 23978: {'lr': 0.00028298840863356006, 'samples': 12277248, 'steps': 23978, 'loss/train': 2.3185224533081055} 02/25/2022 08:38:09 - INFO - codeparrot_training - Step 23979: {'lr': 0.0002829721891763529, 'samples': 12277760, 'steps': 23979, 'loss/train': 1.9011954069137573} 02/25/2022 08:38:12 - INFO - codeparrot_training - Step 23980: {'lr': 0.00028295596957790325, 'samples': 12278272, 'steps': 23980, 'loss/train': 1.0226720571517944} 02/25/2022 08:38:18 - INFO - codeparrot_training - Step 23981: {'lr': 0.0002829397498382808, 'samples': 12278784, 'steps': 23981, 'loss/train': 1.6950478553771973} 02/25/2022 08:38:22 - INFO - codeparrot_training - Step 23982: {'lr': 0.00028292352995755487, 'samples': 12279296, 'steps': 23982, 'loss/train': 1.0232689380645752} 02/25/2022 08:38:27 - INFO - codeparrot_training - Step 23983: {'lr': 0.000282907309935795, 'samples': 12279808, 'steps': 23983, 'loss/train': 0.05856531485915184} 02/25/2022 08:38:30 - INFO - codeparrot_training - Step 23984: {'lr': 0.00028289108977307066, 'samples': 12280320, 'steps': 23984, 'loss/train': 2.1872799396514893} 02/25/2022 08:38:36 - INFO - codeparrot_training - Step 23985: {'lr': 0.00028287486946945137, 'samples': 12280832, 'steps': 23985, 'loss/train': 1.0601203441619873} 02/25/2022 08:38:39 - INFO - codeparrot_training - Step 23986: {'lr': 0.0002828586490250065, 'samples': 12281344, 'steps': 23986, 'loss/train': 3.0550875663757324} 02/25/2022 08:38:47 - INFO - codeparrot_training - Step 23987: {'lr': 0.00028284242843980566, 'samples': 12281856, 'steps': 23987, 'loss/train': 3.133599042892456} 02/25/2022 08:38:50 - INFO - codeparrot_training - Step 23988: {'lr': 0.00028282620771391824, 'samples': 12282368, 'steps': 23988, 'loss/train': 1.8133797645568848} 02/25/2022 08:38:56 - INFO - codeparrot_training - Step 23989: {'lr': 0.00028280998684741387, 'samples': 12282880, 'steps': 23989, 'loss/train': 0.9786564111709595} 02/25/2022 08:38:59 - INFO - codeparrot_training - Step 23990: {'lr': 0.00028279376584036187, 'samples': 12283392, 'steps': 23990, 'loss/train': 2.0829660892486572} 02/25/2022 08:39:05 - INFO - codeparrot_training - Step 23991: {'lr': 0.0002827775446928318, 'samples': 12283904, 'steps': 23991, 'loss/train': 0.30979397892951965} 02/25/2022 08:39:08 - INFO - codeparrot_training - Step 23992: {'lr': 0.00028276132340489306, 'samples': 12284416, 'steps': 23992, 'loss/train': 2.194631814956665} 02/25/2022 08:39:14 - INFO - codeparrot_training - Step 23993: {'lr': 0.0002827451019766153, 'samples': 12284928, 'steps': 23993, 'loss/train': 1.5144309997558594} 02/25/2022 08:39:18 - INFO - codeparrot_training - Step 23994: {'lr': 0.00028272888040806795, 'samples': 12285440, 'steps': 23994, 'loss/train': 0.11840825527906418} 02/25/2022 08:39:23 - INFO - codeparrot_training - Step 23995: {'lr': 0.0002827126586993204, 'samples': 12285952, 'steps': 23995, 'loss/train': 1.9413007497787476} 02/25/2022 08:39:27 - INFO - codeparrot_training - Step 23996: {'lr': 0.0002826964368504422, 'samples': 12286464, 'steps': 23996, 'loss/train': 1.617460012435913} 02/25/2022 08:39:34 - INFO - codeparrot_training - Step 23997: {'lr': 0.0002826802148615029, 'samples': 12286976, 'steps': 23997, 'loss/train': 3.1694045066833496} 02/25/2022 08:39:37 - INFO - codeparrot_training - Step 23998: {'lr': 0.00028266399273257193, 'samples': 12287488, 'steps': 23998, 'loss/train': 2.2219560146331787} 02/25/2022 08:39:43 - INFO - codeparrot_training - Step 23999: {'lr': 0.0002826477704637188, 'samples': 12288000, 'steps': 23999, 'loss/train': 1.6088899374008179} 02/25/2022 08:39:43 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 08:40:01 - WARNING - huggingface_hub.repository - Several commits (24) will be pushed upstream. 02/25/2022 08:40:01 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 08:40:34 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 04e6e2d..7a862ae floral-grass-11 -> floral-grass-11 02/25/2022 08:40:39 - INFO - codeparrot_training - Step 24000: {'lr': 0.000282631548055013, 'samples': 12288512, 'steps': 24000, 'loss/train': 1.9354079961776733} 02/25/2022 08:40:45 - INFO - codeparrot_training - Step 24001: {'lr': 0.00028261532550652387, 'samples': 12289024, 'steps': 24001, 'loss/train': 0.9326981902122498} 02/25/2022 08:40:48 - INFO - codeparrot_training - Step 24002: {'lr': 0.0002825991028183212, 'samples': 12289536, 'steps': 24002, 'loss/train': 5.711711883544922} 02/25/2022 08:40:54 - INFO - codeparrot_training - Step 24003: {'lr': 0.00028258287999047423, 'samples': 12290048, 'steps': 24003, 'loss/train': 2.30151104927063} 02/25/2022 08:40:57 - INFO - codeparrot_training - Step 24004: {'lr': 0.0002825666570230526, 'samples': 12290560, 'steps': 24004, 'loss/train': 1.609424114227295} 02/25/2022 08:41:03 - INFO - codeparrot_training - Step 24005: {'lr': 0.00028255043391612575, 'samples': 12291072, 'steps': 24005, 'loss/train': 1.54843270778656} 02/25/2022 08:41:06 - INFO - codeparrot_training - Step 24006: {'lr': 0.0002825342106697631, 'samples': 12291584, 'steps': 24006, 'loss/train': 1.9663410186767578} 02/25/2022 08:41:14 - INFO - codeparrot_training - Step 24007: {'lr': 0.00028251798728403434, 'samples': 12292096, 'steps': 24007, 'loss/train': 1.3868088722229004} 02/25/2022 08:41:17 - INFO - codeparrot_training - Step 24008: {'lr': 0.0002825017637590088, 'samples': 12292608, 'steps': 24008, 'loss/train': 2.250570774078369} 02/25/2022 08:41:23 - INFO - codeparrot_training - Step 24009: {'lr': 0.00028248554009475604, 'samples': 12293120, 'steps': 24009, 'loss/train': 1.775123119354248} 02/25/2022 08:41:26 - INFO - codeparrot_training - Step 24010: {'lr': 0.0002824693162913454, 'samples': 12293632, 'steps': 24010, 'loss/train': 1.6687016487121582} 02/25/2022 08:41:32 - INFO - codeparrot_training - Step 24011: {'lr': 0.00028245309234884667, 'samples': 12294144, 'steps': 24011, 'loss/train': 1.0860693454742432} 02/25/2022 08:41:35 - INFO - codeparrot_training - Step 24012: {'lr': 0.0002824368682673292, 'samples': 12294656, 'steps': 24012, 'loss/train': 2.729572057723999} 02/25/2022 08:41:41 - INFO - codeparrot_training - Step 24013: {'lr': 0.00028242064404686247, 'samples': 12295168, 'steps': 24013, 'loss/train': 2.1110994815826416} 02/25/2022 08:41:44 - INFO - codeparrot_training - Step 24014: {'lr': 0.00028240441968751596, 'samples': 12295680, 'steps': 24014, 'loss/train': 1.3538614511489868} 02/25/2022 08:41:50 - INFO - codeparrot_training - Step 24015: {'lr': 0.00028238819518935926, 'samples': 12296192, 'steps': 24015, 'loss/train': 1.8347688913345337} 02/25/2022 08:41:53 - INFO - codeparrot_training - Step 24016: {'lr': 0.0002823719705524617, 'samples': 12296704, 'steps': 24016, 'loss/train': 0.36350807547569275} 02/25/2022 08:42:01 - INFO - codeparrot_training - Step 24017: {'lr': 0.000282355745776893, 'samples': 12297216, 'steps': 24017, 'loss/train': 2.2671284675598145} 02/25/2022 08:42:04 - INFO - codeparrot_training - Step 24018: {'lr': 0.0002823395208627226, 'samples': 12297728, 'steps': 24018, 'loss/train': 1.6415660381317139} 02/25/2022 08:42:10 - INFO - codeparrot_training - Step 24019: {'lr': 0.0002823232958100199, 'samples': 12298240, 'steps': 24019, 'loss/train': 2.3679404258728027} 02/25/2022 08:42:13 - INFO - codeparrot_training - Step 24020: {'lr': 0.0002823070706188544, 'samples': 12298752, 'steps': 24020, 'loss/train': 2.10951828956604} 02/25/2022 08:42:19 - INFO - codeparrot_training - Step 24021: {'lr': 0.0002822908452892958, 'samples': 12299264, 'steps': 24021, 'loss/train': 1.0121402740478516} 02/25/2022 08:42:22 - INFO - codeparrot_training - Step 24022: {'lr': 0.0002822746198214133, 'samples': 12299776, 'steps': 24022, 'loss/train': 0.5913536548614502} 02/25/2022 08:42:28 - INFO - codeparrot_training - Step 24023: {'lr': 0.00028225839421527676, 'samples': 12300288, 'steps': 24023, 'loss/train': 2.159191846847534} 02/25/2022 08:42:31 - INFO - codeparrot_training - Step 24024: {'lr': 0.00028224216847095543, 'samples': 12300800, 'steps': 24024, 'loss/train': 0.8222342133522034} 02/25/2022 08:42:37 - INFO - codeparrot_training - Step 24025: {'lr': 0.00028222594258851884, 'samples': 12301312, 'steps': 24025, 'loss/train': 2.244511365890503} 02/25/2022 08:42:40 - INFO - codeparrot_training - Step 24026: {'lr': 0.0002822097165680367, 'samples': 12301824, 'steps': 24026, 'loss/train': 0.4128205478191376} 02/25/2022 08:42:46 - INFO - codeparrot_training - Step 24027: {'lr': 0.0002821934904095782, 'samples': 12302336, 'steps': 24027, 'loss/train': 1.5412288904190063} 02/25/2022 08:42:49 - INFO - codeparrot_training - Step 24028: {'lr': 0.0002821772641132131, 'samples': 12302848, 'steps': 24028, 'loss/train': 0.3498699963092804} 02/25/2022 08:42:55 - INFO - codeparrot_training - Step 24029: {'lr': 0.00028216103767901075, 'samples': 12303360, 'steps': 24029, 'loss/train': 2.6806092262268066} 02/25/2022 08:42:58 - INFO - codeparrot_training - Step 24030: {'lr': 0.0002821448111070407, 'samples': 12303872, 'steps': 24030, 'loss/train': 2.172480344772339} 02/25/2022 08:43:06 - INFO - codeparrot_training - Step 24031: {'lr': 0.00028212858439737254, 'samples': 12304384, 'steps': 24031, 'loss/train': 1.7248648405075073} 02/25/2022 08:43:09 - INFO - codeparrot_training - Step 24032: {'lr': 0.00028211235755007575, 'samples': 12304896, 'steps': 24032, 'loss/train': 2.882354974746704} 02/25/2022 08:43:15 - INFO - codeparrot_training - Step 24033: {'lr': 0.00028209613056521984, 'samples': 12305408, 'steps': 24033, 'loss/train': 2.4785149097442627} 02/25/2022 08:43:18 - INFO - codeparrot_training - Step 24034: {'lr': 0.00028207990344287414, 'samples': 12305920, 'steps': 24034, 'loss/train': 1.556368112564087} 02/25/2022 08:43:23 - INFO - codeparrot_training - Step 24035: {'lr': 0.0002820636761831085, 'samples': 12306432, 'steps': 24035, 'loss/train': 1.5226795673370361} 02/25/2022 08:43:27 - INFO - codeparrot_training - Step 24036: {'lr': 0.00028204744878599215, 'samples': 12306944, 'steps': 24036, 'loss/train': 0.5659157633781433} 02/25/2022 08:43:33 - INFO - codeparrot_training - Step 24037: {'lr': 0.00028203122125159476, 'samples': 12307456, 'steps': 24037, 'loss/train': 0.859870970249176} 02/25/2022 08:43:36 - INFO - codeparrot_training - Step 24038: {'lr': 0.0002820149935799858, 'samples': 12307968, 'steps': 24038, 'loss/train': 2.4015917778015137} 02/25/2022 08:43:40 - INFO - codeparrot_training - Step 24039: {'lr': 0.0002819987657712347, 'samples': 12308480, 'steps': 24039, 'loss/train': 2.179159641265869} 02/25/2022 08:43:46 - INFO - codeparrot_training - Step 24040: {'lr': 0.0002819825378254111, 'samples': 12308992, 'steps': 24040, 'loss/train': 2.623436450958252} 02/25/2022 08:43:49 - INFO - codeparrot_training - Step 24041: {'lr': 0.0002819663097425844, 'samples': 12309504, 'steps': 24041, 'loss/train': 2.718327045440674} 02/25/2022 08:43:55 - INFO - codeparrot_training - Step 24042: {'lr': 0.0002819500815228243, 'samples': 12310016, 'steps': 24042, 'loss/train': 1.3318506479263306} 02/25/2022 08:43:58 - INFO - codeparrot_training - Step 24043: {'lr': 0.00028193385316620007, 'samples': 12310528, 'steps': 24043, 'loss/train': 2.337672472000122} 02/25/2022 08:44:05 - INFO - codeparrot_training - Step 24044: {'lr': 0.00028191762467278146, 'samples': 12311040, 'steps': 24044, 'loss/train': 1.1867364645004272} 02/25/2022 08:44:09 - INFO - codeparrot_training - Step 24045: {'lr': 0.0002819013960426378, 'samples': 12311552, 'steps': 24045, 'loss/train': 1.3193825483322144} 02/25/2022 08:44:14 - INFO - codeparrot_training - Step 24046: {'lr': 0.00028188516727583874, 'samples': 12312064, 'steps': 24046, 'loss/train': 1.7453577518463135} 02/25/2022 08:44:18 - INFO - codeparrot_training - Step 24047: {'lr': 0.00028186893837245375, 'samples': 12312576, 'steps': 24047, 'loss/train': 2.1129283905029297} 02/25/2022 08:44:23 - INFO - codeparrot_training - Step 24048: {'lr': 0.00028185270933255236, 'samples': 12313088, 'steps': 24048, 'loss/train': 1.895471215248108} 02/25/2022 08:44:27 - INFO - codeparrot_training - Step 24049: {'lr': 0.000281836480156204, 'samples': 12313600, 'steps': 24049, 'loss/train': 0.7452694177627563} 02/25/2022 08:44:32 - INFO - codeparrot_training - Step 24050: {'lr': 0.0002818202508434783, 'samples': 12314112, 'steps': 24050, 'loss/train': 2.0025129318237305} 02/25/2022 08:44:36 - INFO - codeparrot_training - Step 24051: {'lr': 0.0002818040213944448, 'samples': 12314624, 'steps': 24051, 'loss/train': 0.6810865998268127} 02/25/2022 08:44:41 - INFO - codeparrot_training - Step 24052: {'lr': 0.0002817877918091729, 'samples': 12315136, 'steps': 24052, 'loss/train': 2.5121686458587646} 02/25/2022 08:44:45 - INFO - codeparrot_training - Step 24053: {'lr': 0.00028177156208773226, 'samples': 12315648, 'steps': 24053, 'loss/train': 0.9702437520027161} 02/25/2022 08:44:52 - INFO - codeparrot_training - Step 24054: {'lr': 0.0002817553322301922, 'samples': 12316160, 'steps': 24054, 'loss/train': 2.570195198059082} 02/25/2022 08:44:56 - INFO - codeparrot_training - Step 24055: {'lr': 0.0002817391022366226, 'samples': 12316672, 'steps': 24055, 'loss/train': 2.3012843132019043} 02/25/2022 08:45:01 - INFO - codeparrot_training - Step 24056: {'lr': 0.0002817228721070926, 'samples': 12317184, 'steps': 24056, 'loss/train': 0.297405481338501} 02/25/2022 08:45:07 - INFO - codeparrot_training - Step 24057: {'lr': 0.00028170664184167194, 'samples': 12317696, 'steps': 24057, 'loss/train': 1.7904223203659058} 02/25/2022 08:45:10 - INFO - codeparrot_training - Step 24058: {'lr': 0.0002816904114404301, 'samples': 12318208, 'steps': 24058, 'loss/train': 1.4671951532363892} 02/25/2022 08:45:14 - INFO - codeparrot_training - Step 24059: {'lr': 0.0002816741809034366, 'samples': 12318720, 'steps': 24059, 'loss/train': 0.8111671209335327} 02/25/2022 08:45:19 - INFO - codeparrot_training - Step 24060: {'lr': 0.000281657950230761, 'samples': 12319232, 'steps': 24060, 'loss/train': 1.8155766725540161} 02/25/2022 08:45:25 - INFO - codeparrot_training - Step 24061: {'lr': 0.0002816417194224727, 'samples': 12319744, 'steps': 24061, 'loss/train': 1.7982791662216187} 02/25/2022 08:45:28 - INFO - codeparrot_training - Step 24062: {'lr': 0.0002816254884786414, 'samples': 12320256, 'steps': 24062, 'loss/train': 1.531999945640564} 02/25/2022 08:45:32 - INFO - codeparrot_training - Step 24063: {'lr': 0.00028160925739933654, 'samples': 12320768, 'steps': 24063, 'loss/train': 2.4597690105438232} 02/25/2022 08:45:39 - INFO - codeparrot_training - Step 24064: {'lr': 0.0002815930261846276, 'samples': 12321280, 'steps': 24064, 'loss/train': 1.4569636583328247} 02/25/2022 08:45:42 - INFO - codeparrot_training - Step 24065: {'lr': 0.00028157679483458423, 'samples': 12321792, 'steps': 24065, 'loss/train': 2.049741506576538} 02/25/2022 08:45:48 - INFO - codeparrot_training - Step 24066: {'lr': 0.00028156056334927595, 'samples': 12322304, 'steps': 24066, 'loss/train': 1.324506402015686} 02/25/2022 08:45:51 - INFO - codeparrot_training - Step 24067: {'lr': 0.00028154433172877216, 'samples': 12322816, 'steps': 24067, 'loss/train': 2.5777499675750732} 02/25/2022 08:45:57 - INFO - codeparrot_training - Step 24068: {'lr': 0.0002815280999731424, 'samples': 12323328, 'steps': 24068, 'loss/train': 1.1260238885879517} 02/25/2022 08:46:00 - INFO - codeparrot_training - Step 24069: {'lr': 0.00028151186808245636, 'samples': 12323840, 'steps': 24069, 'loss/train': 2.5874171257019043} 02/25/2022 08:46:06 - INFO - codeparrot_training - Step 24070: {'lr': 0.0002814956360567835, 'samples': 12324352, 'steps': 24070, 'loss/train': 1.1112257242202759} 02/25/2022 08:46:10 - INFO - codeparrot_training - Step 24071: {'lr': 0.0002814794038961933, 'samples': 12324864, 'steps': 24071, 'loss/train': 0.7731790542602539} 02/25/2022 08:46:15 - INFO - codeparrot_training - Step 24072: {'lr': 0.00028146317160075533, 'samples': 12325376, 'steps': 24072, 'loss/train': 1.5927790403366089} 02/25/2022 08:46:19 - INFO - codeparrot_training - Step 24073: {'lr': 0.00028144693917053916, 'samples': 12325888, 'steps': 24073, 'loss/train': 2.5427446365356445} 02/25/2022 08:46:24 - INFO - codeparrot_training - Step 24074: {'lr': 0.00028143070660561424, 'samples': 12326400, 'steps': 24074, 'loss/train': 1.7923914194107056} 02/25/2022 08:46:28 - INFO - codeparrot_training - Step 24075: {'lr': 0.00028141447390605016, 'samples': 12326912, 'steps': 24075, 'loss/train': 1.227791428565979} 02/25/2022 08:46:33 - INFO - codeparrot_training - Step 24076: {'lr': 0.0002813982410719165, 'samples': 12327424, 'steps': 24076, 'loss/train': 2.6001620292663574} 02/25/2022 08:46:37 - INFO - codeparrot_training - Step 24077: {'lr': 0.0002813820081032827, 'samples': 12327936, 'steps': 24077, 'loss/train': 4.775791168212891} 02/25/2022 08:46:42 - INFO - codeparrot_training - Step 24078: {'lr': 0.0002813657750002183, 'samples': 12328448, 'steps': 24078, 'loss/train': 1.6923785209655762} 02/25/2022 08:46:46 - INFO - codeparrot_training - Step 24079: {'lr': 0.0002813495417627929, 'samples': 12328960, 'steps': 24079, 'loss/train': 2.515749216079712} 02/25/2022 08:46:53 - INFO - codeparrot_training - Step 24080: {'lr': 0.00028133330839107606, 'samples': 12329472, 'steps': 24080, 'loss/train': 2.0844130516052246} 02/25/2022 08:46:57 - INFO - codeparrot_training - Step 24081: {'lr': 0.0002813170748851373, 'samples': 12329984, 'steps': 24081, 'loss/train': 2.1875863075256348} 02/25/2022 08:47:02 - INFO - codeparrot_training - Step 24082: {'lr': 0.00028130084124504607, 'samples': 12330496, 'steps': 24082, 'loss/train': 1.7935336828231812} 02/25/2022 08:47:06 - INFO - codeparrot_training - Step 24083: {'lr': 0.000281284607470872, 'samples': 12331008, 'steps': 24083, 'loss/train': 2.8879871368408203} 02/25/2022 08:47:11 - INFO - codeparrot_training - Step 24084: {'lr': 0.00028126837356268463, 'samples': 12331520, 'steps': 24084, 'loss/train': 2.0636024475097656} 02/25/2022 08:47:15 - INFO - codeparrot_training - Step 24085: {'lr': 0.00028125213952055343, 'samples': 12332032, 'steps': 24085, 'loss/train': 2.0726478099823} 02/25/2022 08:47:20 - INFO - codeparrot_training - Step 24086: {'lr': 0.00028123590534454804, 'samples': 12332544, 'steps': 24086, 'loss/train': 2.4974093437194824} 02/25/2022 08:47:24 - INFO - codeparrot_training - Step 24087: {'lr': 0.00028121967103473795, 'samples': 12333056, 'steps': 24087, 'loss/train': 2.1765761375427246} 02/25/2022 08:47:29 - INFO - codeparrot_training - Step 24088: {'lr': 0.0002812034365911926, 'samples': 12333568, 'steps': 24088, 'loss/train': 1.0200955867767334} 02/25/2022 08:47:33 - INFO - codeparrot_training - Step 24089: {'lr': 0.00028118720201398173, 'samples': 12334080, 'steps': 24089, 'loss/train': 2.440505266189575} 02/25/2022 08:47:40 - INFO - codeparrot_training - Step 24090: {'lr': 0.00028117096730317475, 'samples': 12334592, 'steps': 24090, 'loss/train': 0.2194325178861618} 02/25/2022 08:47:43 - INFO - codeparrot_training - Step 24091: {'lr': 0.0002811547324588413, 'samples': 12335104, 'steps': 24091, 'loss/train': 1.6860175132751465} 02/25/2022 08:47:49 - INFO - codeparrot_training - Step 24092: {'lr': 0.0002811384974810508, 'samples': 12335616, 'steps': 24092, 'loss/train': 1.37722909450531} 02/25/2022 08:47:52 - INFO - codeparrot_training - Step 24093: {'lr': 0.0002811222623698729, 'samples': 12336128, 'steps': 24093, 'loss/train': 2.72214937210083} 02/25/2022 08:47:58 - INFO - codeparrot_training - Step 24094: {'lr': 0.00028110602712537713, 'samples': 12336640, 'steps': 24094, 'loss/train': 2.743868827819824} 02/25/2022 08:48:01 - INFO - codeparrot_training - Step 24095: {'lr': 0.000281089791747633, 'samples': 12337152, 'steps': 24095, 'loss/train': 2.0026705265045166} 02/25/2022 08:48:07 - INFO - codeparrot_training - Step 24096: {'lr': 0.00028107355623671, 'samples': 12337664, 'steps': 24096, 'loss/train': 2.309119701385498} 02/25/2022 08:48:11 - INFO - codeparrot_training - Step 24097: {'lr': 0.00028105732059267787, 'samples': 12338176, 'steps': 24097, 'loss/train': 1.4898655414581299} 02/25/2022 08:48:16 - INFO - codeparrot_training - Step 24098: {'lr': 0.00028104108481560603, 'samples': 12338688, 'steps': 24098, 'loss/train': 2.7868711948394775} 02/25/2022 08:48:20 - INFO - codeparrot_training - Step 24099: {'lr': 0.00028102484890556395, 'samples': 12339200, 'steps': 24099, 'loss/train': 1.7177258729934692} 02/25/2022 08:48:27 - INFO - codeparrot_training - Step 24100: {'lr': 0.00028100861286262135, 'samples': 12339712, 'steps': 24100, 'loss/train': 1.4562634229660034} 02/25/2022 08:48:30 - INFO - codeparrot_training - Step 24101: {'lr': 0.0002809923766868476, 'samples': 12340224, 'steps': 24101, 'loss/train': 1.6887115240097046} 02/25/2022 08:48:36 - INFO - codeparrot_training - Step 24102: {'lr': 0.00028097614037831244, 'samples': 12340736, 'steps': 24102, 'loss/train': 1.382042646408081} 02/25/2022 08:48:39 - INFO - codeparrot_training - Step 24103: {'lr': 0.00028095990393708534, 'samples': 12341248, 'steps': 24103, 'loss/train': 2.1060309410095215} 02/25/2022 08:48:45 - INFO - codeparrot_training - Step 24104: {'lr': 0.00028094366736323577, 'samples': 12341760, 'steps': 24104, 'loss/train': 1.7109125852584839} 02/25/2022 08:48:49 - INFO - codeparrot_training - Step 24105: {'lr': 0.0002809274306568335, 'samples': 12342272, 'steps': 24105, 'loss/train': 0.7774963974952698} 02/25/2022 08:48:54 - INFO - codeparrot_training - Step 24106: {'lr': 0.00028091119381794776, 'samples': 12342784, 'steps': 24106, 'loss/train': 1.1881859302520752} 02/25/2022 08:48:58 - INFO - codeparrot_training - Step 24107: {'lr': 0.0002808949568466484, 'samples': 12343296, 'steps': 24107, 'loss/train': 1.824115514755249} 02/25/2022 08:49:03 - INFO - codeparrot_training - Step 24108: {'lr': 0.0002808787197430048, 'samples': 12343808, 'steps': 24108, 'loss/train': 1.1005526781082153} 02/25/2022 08:49:07 - INFO - codeparrot_training - Step 24109: {'lr': 0.0002808624825070866, 'samples': 12344320, 'steps': 24109, 'loss/train': 1.4673726558685303} 02/25/2022 08:49:14 - INFO - codeparrot_training - Step 24110: {'lr': 0.0002808462451389633, 'samples': 12344832, 'steps': 24110, 'loss/train': 2.0260980129241943} 02/25/2022 08:49:17 - INFO - codeparrot_training - Step 24111: {'lr': 0.0002808300076387045, 'samples': 12345344, 'steps': 24111, 'loss/train': 2.2493178844451904} 02/25/2022 08:49:23 - INFO - codeparrot_training - Step 24112: {'lr': 0.0002808137700063797, 'samples': 12345856, 'steps': 24112, 'loss/train': 2.1967365741729736} 02/25/2022 08:49:26 - INFO - codeparrot_training - Step 24113: {'lr': 0.0002807975322420585, 'samples': 12346368, 'steps': 24113, 'loss/train': 0.9307685494422913} 02/25/2022 08:49:32 - INFO - codeparrot_training - Step 24114: {'lr': 0.0002807812943458105, 'samples': 12346880, 'steps': 24114, 'loss/train': 1.0833745002746582} 02/25/2022 08:49:35 - INFO - codeparrot_training - Step 24115: {'lr': 0.00028076505631770515, 'samples': 12347392, 'steps': 24115, 'loss/train': 1.9819477796554565} 02/25/2022 08:49:41 - INFO - codeparrot_training - Step 24116: {'lr': 0.0002807488181578121, 'samples': 12347904, 'steps': 24116, 'loss/train': 1.9693245887756348} 02/25/2022 08:49:44 - INFO - codeparrot_training - Step 24117: {'lr': 0.00028073257986620085, 'samples': 12348416, 'steps': 24117, 'loss/train': 1.4338395595550537} 02/25/2022 08:49:50 - INFO - codeparrot_training - Step 24118: {'lr': 0.00028071634144294106, 'samples': 12348928, 'steps': 24118, 'loss/train': 0.9974554181098938} 02/25/2022 08:49:53 - INFO - codeparrot_training - Step 24119: {'lr': 0.0002807001028881021, 'samples': 12349440, 'steps': 24119, 'loss/train': 1.834402322769165} 02/25/2022 08:49:59 - INFO - codeparrot_training - Step 24120: {'lr': 0.0002806838642017537, 'samples': 12349952, 'steps': 24120, 'loss/train': 1.0758941173553467} 02/25/2022 08:50:03 - INFO - codeparrot_training - Step 24121: {'lr': 0.0002806676253839654, 'samples': 12350464, 'steps': 24121, 'loss/train': 1.8781483173370361} 02/25/2022 08:50:08 - INFO - codeparrot_training - Step 24122: {'lr': 0.00028065138643480677, 'samples': 12350976, 'steps': 24122, 'loss/train': 2.4904329776763916} 02/25/2022 08:50:12 - INFO - codeparrot_training - Step 24123: {'lr': 0.00028063514735434726, 'samples': 12351488, 'steps': 24123, 'loss/train': 2.470557928085327} 02/25/2022 08:50:17 - INFO - codeparrot_training - Step 24124: {'lr': 0.0002806189081426565, 'samples': 12352000, 'steps': 24124, 'loss/train': 8.82812213897705} 02/25/2022 08:50:21 - INFO - codeparrot_training - Step 24125: {'lr': 0.0002806026687998041, 'samples': 12352512, 'steps': 24125, 'loss/train': 1.196666955947876} 02/25/2022 08:50:28 - INFO - codeparrot_training - Step 24126: {'lr': 0.0002805864293258595, 'samples': 12353024, 'steps': 24126, 'loss/train': 2.797055721282959} 02/25/2022 08:50:32 - INFO - codeparrot_training - Step 24127: {'lr': 0.00028057018972089243, 'samples': 12353536, 'steps': 24127, 'loss/train': 1.391169786453247} 02/25/2022 08:50:37 - INFO - codeparrot_training - Step 24128: {'lr': 0.00028055394998497237, 'samples': 12354048, 'steps': 24128, 'loss/train': 2.859102725982666} 02/25/2022 08:50:41 - INFO - codeparrot_training - Step 24129: {'lr': 0.00028053771011816894, 'samples': 12354560, 'steps': 24129, 'loss/train': 1.148810863494873} 02/25/2022 08:50:47 - INFO - codeparrot_training - Step 24130: {'lr': 0.00028052147012055153, 'samples': 12355072, 'steps': 24130, 'loss/train': 1.575769305229187} 02/25/2022 08:50:50 - INFO - codeparrot_training - Step 24131: {'lr': 0.0002805052299921899, 'samples': 12355584, 'steps': 24131, 'loss/train': 0.7827603816986084} 02/25/2022 08:50:54 - INFO - codeparrot_training - Step 24132: {'lr': 0.0002804889897331536, 'samples': 12356096, 'steps': 24132, 'loss/train': 2.9821131229400635} 02/25/2022 08:50:59 - INFO - codeparrot_training - Step 24133: {'lr': 0.0002804727493435121, 'samples': 12356608, 'steps': 24133, 'loss/train': 0.4320419430732727} 02/25/2022 08:51:03 - INFO - codeparrot_training - Step 24134: {'lr': 0.000280456508823335, 'samples': 12357120, 'steps': 24134, 'loss/train': 0.1283150017261505} 02/25/2022 08:51:08 - INFO - codeparrot_training - Step 24135: {'lr': 0.0002804402681726919, 'samples': 12357632, 'steps': 24135, 'loss/train': 2.6496336460113525} 02/25/2022 08:51:12 - INFO - codeparrot_training - Step 24136: {'lr': 0.00028042402739165235, 'samples': 12358144, 'steps': 24136, 'loss/train': 1.8936935663223267} 02/25/2022 08:51:20 - INFO - codeparrot_training - Step 24137: {'lr': 0.000280407786480286, 'samples': 12358656, 'steps': 24137, 'loss/train': 1.015517234802246} 02/25/2022 08:51:23 - INFO - codeparrot_training - Step 24138: {'lr': 0.00028039154543866223, 'samples': 12359168, 'steps': 24138, 'loss/train': 1.958033800125122} 02/25/2022 08:51:29 - INFO - codeparrot_training - Step 24139: {'lr': 0.0002803753042668508, 'samples': 12359680, 'steps': 24139, 'loss/train': 0.6777181625366211} 02/25/2022 08:51:32 - INFO - codeparrot_training - Step 24140: {'lr': 0.0002803590629649212, 'samples': 12360192, 'steps': 24140, 'loss/train': 0.3564806878566742} 02/25/2022 08:51:38 - INFO - codeparrot_training - Step 24141: {'lr': 0.000280342821532943, 'samples': 12360704, 'steps': 24141, 'loss/train': 2.056265115737915} 02/25/2022 08:51:41 - INFO - codeparrot_training - Step 24142: {'lr': 0.0002803265799709858, 'samples': 12361216, 'steps': 24142, 'loss/train': 0.9112293720245361} 02/25/2022 08:51:47 - INFO - codeparrot_training - Step 24143: {'lr': 0.0002803103382791191, 'samples': 12361728, 'steps': 24143, 'loss/train': 1.7687865495681763} 02/25/2022 08:51:50 - INFO - codeparrot_training - Step 24144: {'lr': 0.00028029409645741264, 'samples': 12362240, 'steps': 24144, 'loss/train': 1.9292594194412231} 02/25/2022 08:51:56 - INFO - codeparrot_training - Step 24145: {'lr': 0.00028027785450593585, 'samples': 12362752, 'steps': 24145, 'loss/train': 2.1720035076141357} 02/25/2022 08:51:59 - INFO - codeparrot_training - Step 24146: {'lr': 0.00028026161242475833, 'samples': 12363264, 'steps': 24146, 'loss/train': 0.971410870552063} 02/25/2022 08:52:07 - INFO - codeparrot_training - Step 24147: {'lr': 0.0002802453702139497, 'samples': 12363776, 'steps': 24147, 'loss/train': 1.8997070789337158} 02/25/2022 08:52:10 - INFO - codeparrot_training - Step 24148: {'lr': 0.0002802291278735795, 'samples': 12364288, 'steps': 24148, 'loss/train': 1.9926778078079224} 02/25/2022 08:52:16 - INFO - codeparrot_training - Step 24149: {'lr': 0.0002802128854037173, 'samples': 12364800, 'steps': 24149, 'loss/train': 2.148219108581543} 02/25/2022 08:52:19 - INFO - codeparrot_training - Step 24150: {'lr': 0.00028019664280443275, 'samples': 12365312, 'steps': 24150, 'loss/train': 1.7260531187057495} 02/25/2022 08:52:24 - INFO - codeparrot_training - Step 24151: {'lr': 0.00028018040007579524, 'samples': 12365824, 'steps': 24151, 'loss/train': 2.203099250793457} 02/25/2022 08:52:28 - INFO - codeparrot_training - Step 24152: {'lr': 0.00028016415721787463, 'samples': 12366336, 'steps': 24152, 'loss/train': 1.587680697441101} 02/25/2022 08:52:34 - INFO - codeparrot_training - Step 24153: {'lr': 0.0002801479142307403, 'samples': 12366848, 'steps': 24153, 'loss/train': 1.4639865159988403} 02/25/2022 08:52:37 - INFO - codeparrot_training - Step 24154: {'lr': 0.0002801316711144619, 'samples': 12367360, 'steps': 24154, 'loss/train': 0.7426556348800659} 02/25/2022 08:52:43 - INFO - codeparrot_training - Step 24155: {'lr': 0.00028011542786910896, 'samples': 12367872, 'steps': 24155, 'loss/train': 1.4798780679702759} 02/25/2022 08:52:46 - INFO - codeparrot_training - Step 24156: {'lr': 0.00028009918449475104, 'samples': 12368384, 'steps': 24156, 'loss/train': 3.189852476119995} 02/25/2022 08:52:53 - INFO - codeparrot_training - Step 24157: {'lr': 0.0002800829409914578, 'samples': 12368896, 'steps': 24157, 'loss/train': 2.7021100521087646} 02/25/2022 08:52:57 - INFO - codeparrot_training - Step 24158: {'lr': 0.0002800666973592988, 'samples': 12369408, 'steps': 24158, 'loss/train': 1.9429315328598022} 02/25/2022 08:53:02 - INFO - codeparrot_training - Step 24159: {'lr': 0.0002800504535983436, 'samples': 12369920, 'steps': 24159, 'loss/train': 2.165562152862549} 02/25/2022 08:53:06 - INFO - codeparrot_training - Step 24160: {'lr': 0.00028003420970866175, 'samples': 12370432, 'steps': 24160, 'loss/train': 0.9998244047164917} 02/25/2022 08:53:11 - INFO - codeparrot_training - Step 24161: {'lr': 0.00028001796569032304, 'samples': 12370944, 'steps': 24161, 'loss/train': 1.5083441734313965} 02/25/2022 08:53:15 - INFO - codeparrot_training - Step 24162: {'lr': 0.00028000172154339675, 'samples': 12371456, 'steps': 24162, 'loss/train': 1.707032322883606} 02/25/2022 08:53:20 - INFO - codeparrot_training - Step 24163: {'lr': 0.00027998547726795265, 'samples': 12371968, 'steps': 24163, 'loss/train': 1.2037488222122192} 02/25/2022 08:53:24 - INFO - codeparrot_training - Step 24164: {'lr': 0.00027996923286406037, 'samples': 12372480, 'steps': 24164, 'loss/train': 2.5205039978027344} 02/25/2022 08:53:29 - INFO - codeparrot_training - Step 24165: {'lr': 0.00027995298833178927, 'samples': 12372992, 'steps': 24165, 'loss/train': 1.5232199430465698} 02/25/2022 08:53:33 - INFO - codeparrot_training - Step 24166: {'lr': 0.0002799367436712092, 'samples': 12373504, 'steps': 24166, 'loss/train': 1.2361977100372314} 02/25/2022 08:53:38 - INFO - codeparrot_training - Step 24167: {'lr': 0.00027992049888238957, 'samples': 12374016, 'steps': 24167, 'loss/train': 1.5076152086257935} 02/25/2022 08:53:42 - INFO - codeparrot_training - Step 24168: {'lr': 0.00027990425396540007, 'samples': 12374528, 'steps': 24168, 'loss/train': 2.016110897064209} 02/25/2022 08:53:47 - INFO - codeparrot_training - Step 24169: {'lr': 0.0002798880089203102, 'samples': 12375040, 'steps': 24169, 'loss/train': 1.945264458656311} 02/25/2022 08:53:51 - INFO - codeparrot_training - Step 24170: {'lr': 0.0002798717637471896, 'samples': 12375552, 'steps': 24170, 'loss/train': 1.2609217166900635} 02/25/2022 08:53:56 - INFO - codeparrot_training - Step 24171: {'lr': 0.0002798555184461078, 'samples': 12376064, 'steps': 24171, 'loss/train': 1.3531477451324463} 02/25/2022 08:53:59 - INFO - codeparrot_training - Step 24172: {'lr': 0.0002798392730171345, 'samples': 12376576, 'steps': 24172, 'loss/train': 1.294921875} 02/25/2022 08:54:07 - INFO - codeparrot_training - Step 24173: {'lr': 0.00027982302746033924, 'samples': 12377088, 'steps': 24173, 'loss/train': 0.43970680236816406} 02/25/2022 08:54:10 - INFO - codeparrot_training - Step 24174: {'lr': 0.0002798067817757916, 'samples': 12377600, 'steps': 24174, 'loss/train': 1.9545875787734985} 02/25/2022 08:54:16 - INFO - codeparrot_training - Step 24175: {'lr': 0.00027979053596356105, 'samples': 12378112, 'steps': 24175, 'loss/train': 1.5943564176559448} 02/25/2022 08:54:19 - INFO - codeparrot_training - Step 24176: {'lr': 0.00027977429002371744, 'samples': 12378624, 'steps': 24176, 'loss/train': 2.2447307109832764} 02/25/2022 08:54:25 - INFO - codeparrot_training - Step 24177: {'lr': 0.0002797580439563302, 'samples': 12379136, 'steps': 24177, 'loss/train': 2.1768507957458496} 02/25/2022 08:54:28 - INFO - codeparrot_training - Step 24178: {'lr': 0.0002797417977614689, 'samples': 12379648, 'steps': 24178, 'loss/train': 1.9363279342651367} 02/25/2022 08:54:34 - INFO - codeparrot_training - Step 24179: {'lr': 0.00027972555143920326, 'samples': 12380160, 'steps': 24179, 'loss/train': 2.437059164047241} 02/25/2022 08:54:37 - INFO - codeparrot_training - Step 24180: {'lr': 0.00027970930498960275, 'samples': 12380672, 'steps': 24180, 'loss/train': 1.9060323238372803} 02/25/2022 08:54:43 - INFO - codeparrot_training - Step 24181: {'lr': 0.000279693058412737, 'samples': 12381184, 'steps': 24181, 'loss/train': 1.8979460000991821} 02/25/2022 08:54:46 - INFO - codeparrot_training - Step 24182: {'lr': 0.00027967681170867567, 'samples': 12381696, 'steps': 24182, 'loss/train': 2.1599743366241455} 02/25/2022 08:54:54 - INFO - codeparrot_training - Step 24183: {'lr': 0.00027966056487748825, 'samples': 12382208, 'steps': 24183, 'loss/train': 2.8617851734161377} 02/25/2022 08:54:57 - INFO - codeparrot_training - Step 24184: {'lr': 0.0002796443179192444, 'samples': 12382720, 'steps': 24184, 'loss/train': 2.365355968475342} 02/25/2022 08:55:03 - INFO - codeparrot_training - Step 24185: {'lr': 0.0002796280708340137, 'samples': 12383232, 'steps': 24185, 'loss/train': 2.2310194969177246} 02/25/2022 08:55:06 - INFO - codeparrot_training - Step 24186: {'lr': 0.0002796118236218658, 'samples': 12383744, 'steps': 24186, 'loss/train': 1.3441805839538574} 02/25/2022 08:55:12 - INFO - codeparrot_training - Step 24187: {'lr': 0.00027959557628287016, 'samples': 12384256, 'steps': 24187, 'loss/train': 0.7696624398231506} 02/25/2022 08:55:15 - INFO - codeparrot_training - Step 24188: {'lr': 0.0002795793288170965, 'samples': 12384768, 'steps': 24188, 'loss/train': 2.259233236312866} 02/25/2022 08:55:21 - INFO - codeparrot_training - Step 24189: {'lr': 0.00027956308122461436, 'samples': 12385280, 'steps': 24189, 'loss/train': 1.4260845184326172} 02/25/2022 08:55:24 - INFO - codeparrot_training - Step 24190: {'lr': 0.0002795468335054935, 'samples': 12385792, 'steps': 24190, 'loss/train': 1.8014477491378784} 02/25/2022 08:55:30 - INFO - codeparrot_training - Step 24191: {'lr': 0.00027953058565980326, 'samples': 12386304, 'steps': 24191, 'loss/train': 1.4506462812423706} 02/25/2022 08:55:33 - INFO - codeparrot_training - Step 24192: {'lr': 0.0002795143376876134, 'samples': 12386816, 'steps': 24192, 'loss/train': 2.0312018394470215} 02/25/2022 08:55:39 - INFO - codeparrot_training - Step 24193: {'lr': 0.00027949808958899354, 'samples': 12387328, 'steps': 24193, 'loss/train': 2.3549644947052} 02/25/2022 08:55:42 - INFO - codeparrot_training - Step 24194: {'lr': 0.00027948184136401317, 'samples': 12387840, 'steps': 24194, 'loss/train': 1.9346637725830078} 02/25/2022 08:55:50 - INFO - codeparrot_training - Step 24195: {'lr': 0.00027946559301274206, 'samples': 12388352, 'steps': 24195, 'loss/train': 3.3463943004608154} 02/25/2022 08:55:53 - INFO - codeparrot_training - Step 24196: {'lr': 0.0002794493445352496, 'samples': 12388864, 'steps': 24196, 'loss/train': 1.8509702682495117} 02/25/2022 08:55:59 - INFO - codeparrot_training - Step 24197: {'lr': 0.0002794330959316055, 'samples': 12389376, 'steps': 24197, 'loss/train': 1.0540375709533691} 02/25/2022 08:56:02 - INFO - codeparrot_training - Step 24198: {'lr': 0.0002794168472018794, 'samples': 12389888, 'steps': 24198, 'loss/train': 1.5441783666610718} 02/25/2022 08:56:08 - INFO - codeparrot_training - Step 24199: {'lr': 0.0002794005983461408, 'samples': 12390400, 'steps': 24199, 'loss/train': 2.2779369354248047} 02/25/2022 08:56:11 - INFO - codeparrot_training - Step 24200: {'lr': 0.00027938434936445943, 'samples': 12390912, 'steps': 24200, 'loss/train': 0.6469489932060242} 02/25/2022 08:56:17 - INFO - codeparrot_training - Step 24201: {'lr': 0.00027936810025690483, 'samples': 12391424, 'steps': 24201, 'loss/train': 1.9743132591247559} 02/25/2022 08:56:20 - INFO - codeparrot_training - Step 24202: {'lr': 0.0002793518510235466, 'samples': 12391936, 'steps': 24202, 'loss/train': 2.8360402584075928} 02/25/2022 08:56:26 - INFO - codeparrot_training - Step 24203: {'lr': 0.00027933560166445445, 'samples': 12392448, 'steps': 24203, 'loss/train': 2.3457517623901367} 02/25/2022 08:56:29 - INFO - codeparrot_training - Step 24204: {'lr': 0.00027931935217969777, 'samples': 12392960, 'steps': 24204, 'loss/train': 2.4369144439697266} 02/25/2022 08:56:36 - INFO - codeparrot_training - Step 24205: {'lr': 0.00027930310256934636, 'samples': 12393472, 'steps': 24205, 'loss/train': 1.5425649881362915} 02/25/2022 08:56:40 - INFO - codeparrot_training - Step 24206: {'lr': 0.0002792868528334697, 'samples': 12393984, 'steps': 24206, 'loss/train': 2.003851890563965} 02/25/2022 08:56:45 - INFO - codeparrot_training - Step 24207: {'lr': 0.0002792706029721376, 'samples': 12394496, 'steps': 24207, 'loss/train': 1.8645727634429932} 02/25/2022 08:56:49 - INFO - codeparrot_training - Step 24208: {'lr': 0.0002792543529854194, 'samples': 12395008, 'steps': 24208, 'loss/train': 1.888032078742981} 02/25/2022 08:56:55 - INFO - codeparrot_training - Step 24209: {'lr': 0.0002792381028733849, 'samples': 12395520, 'steps': 24209, 'loss/train': 1.743841290473938} 02/25/2022 08:56:58 - INFO - codeparrot_training - Step 24210: {'lr': 0.0002792218526361036, 'samples': 12396032, 'steps': 24210, 'loss/train': 1.3721891641616821} 02/25/2022 08:57:04 - INFO - codeparrot_training - Step 24211: {'lr': 0.00027920560227364516, 'samples': 12396544, 'steps': 24211, 'loss/train': 2.106753349304199} 02/25/2022 08:57:07 - INFO - codeparrot_training - Step 24212: {'lr': 0.00027918935178607927, 'samples': 12397056, 'steps': 24212, 'loss/train': 1.7442359924316406} 02/25/2022 08:57:12 - INFO - codeparrot_training - Step 24213: {'lr': 0.00027917310117347543, 'samples': 12397568, 'steps': 24213, 'loss/train': 2.296968460083008} 02/25/2022 08:57:16 - INFO - codeparrot_training - Step 24214: {'lr': 0.00027915685043590325, 'samples': 12398080, 'steps': 24214, 'loss/train': 2.3627583980560303} 02/25/2022 08:57:22 - INFO - codeparrot_training - Step 24215: {'lr': 0.00027914059957343245, 'samples': 12398592, 'steps': 24215, 'loss/train': 1.9211373329162598} 02/25/2022 08:57:27 - INFO - codeparrot_training - Step 24216: {'lr': 0.00027912434858613257, 'samples': 12399104, 'steps': 24216, 'loss/train': 1.789819598197937} 02/25/2022 08:57:31 - INFO - codeparrot_training - Step 24217: {'lr': 0.00027910809747407316, 'samples': 12399616, 'steps': 24217, 'loss/train': 1.4690529108047485} 02/25/2022 08:57:38 - INFO - codeparrot_training - Step 24218: {'lr': 0.00027909184623732395, 'samples': 12400128, 'steps': 24218, 'loss/train': 3.1276214122772217} 02/25/2022 08:57:42 - INFO - codeparrot_training - Step 24219: {'lr': 0.00027907559487595453, 'samples': 12400640, 'steps': 24219, 'loss/train': 2.853700637817383} 02/25/2022 08:57:47 - INFO - codeparrot_training - Step 24220: {'lr': 0.00027905934339003446, 'samples': 12401152, 'steps': 24220, 'loss/train': 1.4881477355957031} 02/25/2022 08:57:51 - INFO - codeparrot_training - Step 24221: {'lr': 0.0002790430917796334, 'samples': 12401664, 'steps': 24221, 'loss/train': 2.0138003826141357} 02/25/2022 08:57:56 - INFO - codeparrot_training - Step 24222: {'lr': 0.0002790268400448209, 'samples': 12402176, 'steps': 24222, 'loss/train': 2.5834553241729736} 02/25/2022 08:58:00 - INFO - codeparrot_training - Step 24223: {'lr': 0.00027901058818566673, 'samples': 12402688, 'steps': 24223, 'loss/train': 1.3534088134765625} 02/25/2022 08:58:05 - INFO - codeparrot_training - Step 24224: {'lr': 0.00027899433620224033, 'samples': 12403200, 'steps': 24224, 'loss/train': 1.4188522100448608} 02/25/2022 08:58:09 - INFO - codeparrot_training - Step 24225: {'lr': 0.00027897808409461147, 'samples': 12403712, 'steps': 24225, 'loss/train': 0.9269300103187561} 02/25/2022 08:58:14 - INFO - codeparrot_training - Step 24226: {'lr': 0.00027896183186284964, 'samples': 12404224, 'steps': 24226, 'loss/train': 2.1332437992095947} 02/25/2022 08:58:18 - INFO - codeparrot_training - Step 24227: {'lr': 0.00027894557950702457, 'samples': 12404736, 'steps': 24227, 'loss/train': 3.8392045497894287} 02/25/2022 08:58:23 - INFO - codeparrot_training - Step 24228: {'lr': 0.0002789293270272058, 'samples': 12405248, 'steps': 24228, 'loss/train': 1.892857551574707} 02/25/2022 08:58:27 - INFO - codeparrot_training - Step 24229: {'lr': 0.00027891307442346304, 'samples': 12405760, 'steps': 24229, 'loss/train': 1.1620746850967407} 02/25/2022 08:58:34 - INFO - codeparrot_training - Step 24230: {'lr': 0.0002788968216958657, 'samples': 12406272, 'steps': 24230, 'loss/train': 1.9652231931686401} 02/25/2022 08:58:38 - INFO - codeparrot_training - Step 24231: {'lr': 0.0002788805688444837, 'samples': 12406784, 'steps': 24231, 'loss/train': 2.0279951095581055} 02/25/2022 08:58:43 - INFO - codeparrot_training - Step 24232: {'lr': 0.00027886431586938645, 'samples': 12407296, 'steps': 24232, 'loss/train': 1.5853351354599} 02/25/2022 08:58:47 - INFO - codeparrot_training - Step 24233: {'lr': 0.00027884806277064366, 'samples': 12407808, 'steps': 24233, 'loss/train': 1.0278211832046509} 02/25/2022 08:58:50 - INFO - codeparrot_training - Step 24234: {'lr': 0.00027883180954832486, 'samples': 12408320, 'steps': 24234, 'loss/train': 3.804316282272339} 02/25/2022 08:58:56 - INFO - codeparrot_training - Step 24235: {'lr': 0.0002788155562024999, 'samples': 12408832, 'steps': 24235, 'loss/train': 5.114134311676025} 02/25/2022 08:58:59 - INFO - codeparrot_training - Step 24236: {'lr': 0.0002787993027332381, 'samples': 12409344, 'steps': 24236, 'loss/train': 1.403770923614502} 02/25/2022 08:59:05 - INFO - codeparrot_training - Step 24237: {'lr': 0.00027878304914060934, 'samples': 12409856, 'steps': 24237, 'loss/train': 2.5707404613494873} 02/25/2022 08:59:10 - INFO - codeparrot_training - Step 24238: {'lr': 0.00027876679542468305, 'samples': 12410368, 'steps': 24238, 'loss/train': 0.8141780495643616} 02/25/2022 08:59:14 - INFO - codeparrot_training - Step 24239: {'lr': 0.00027875054158552895, 'samples': 12410880, 'steps': 24239, 'loss/train': 2.051060914993286} 02/25/2022 08:59:22 - INFO - codeparrot_training - Step 24240: {'lr': 0.0002787342876232167, 'samples': 12411392, 'steps': 24240, 'loss/train': 1.2634204626083374} 02/25/2022 08:59:25 - INFO - codeparrot_training - Step 24241: {'lr': 0.0002787180335378158, 'samples': 12411904, 'steps': 24241, 'loss/train': 2.1626992225646973} 02/25/2022 08:59:29 - INFO - codeparrot_training - Step 24242: {'lr': 0.00027870177932939606, 'samples': 12412416, 'steps': 24242, 'loss/train': 2.298090934753418} 02/25/2022 08:59:34 - INFO - codeparrot_training - Step 24243: {'lr': 0.000278685524998027, 'samples': 12412928, 'steps': 24243, 'loss/train': 2.382725715637207} 02/25/2022 08:59:38 - INFO - codeparrot_training - Step 24244: {'lr': 0.0002786692705437783, 'samples': 12413440, 'steps': 24244, 'loss/train': 3.057950258255005} 02/25/2022 08:59:44 - INFO - codeparrot_training - Step 24245: {'lr': 0.00027865301596671945, 'samples': 12413952, 'steps': 24245, 'loss/train': 2.18320369720459} 02/25/2022 08:59:47 - INFO - codeparrot_training - Step 24246: {'lr': 0.0002786367612669202, 'samples': 12414464, 'steps': 24246, 'loss/train': 1.99215829372406} 02/25/2022 08:59:53 - INFO - codeparrot_training - Step 24247: {'lr': 0.00027862050644445016, 'samples': 12414976, 'steps': 24247, 'loss/train': 1.9238471984863281} 02/25/2022 08:59:56 - INFO - codeparrot_training - Step 24248: {'lr': 0.00027860425149937894, 'samples': 12415488, 'steps': 24248, 'loss/train': 1.4706083536148071} 02/25/2022 09:00:02 - INFO - codeparrot_training - Step 24249: {'lr': 0.00027858799643177624, 'samples': 12416000, 'steps': 24249, 'loss/train': 1.3288410902023315} 02/25/2022 09:00:05 - INFO - codeparrot_training - Step 24250: {'lr': 0.00027857174124171165, 'samples': 12416512, 'steps': 24250, 'loss/train': 1.8691834211349487} 02/25/2022 09:00:12 - INFO - codeparrot_training - Step 24251: {'lr': 0.0002785554859292548, 'samples': 12417024, 'steps': 24251, 'loss/train': 0.7318936586380005} 02/25/2022 09:00:16 - INFO - codeparrot_training - Step 24252: {'lr': 0.0002785392304944752, 'samples': 12417536, 'steps': 24252, 'loss/train': 1.5632015466690063} 02/25/2022 09:00:21 - INFO - codeparrot_training - Step 24253: {'lr': 0.0002785229749374427, 'samples': 12418048, 'steps': 24253, 'loss/train': 2.637326717376709} 02/25/2022 09:00:25 - INFO - codeparrot_training - Step 24254: {'lr': 0.0002785067192582268, 'samples': 12418560, 'steps': 24254, 'loss/train': 2.509840250015259} 02/25/2022 09:00:30 - INFO - codeparrot_training - Step 24255: {'lr': 0.0002784904634568972, 'samples': 12419072, 'steps': 24255, 'loss/train': 1.9652732610702515} 02/25/2022 09:00:34 - INFO - codeparrot_training - Step 24256: {'lr': 0.0002784742075335235, 'samples': 12419584, 'steps': 24256, 'loss/train': 1.8331021070480347} 02/25/2022 09:00:39 - INFO - codeparrot_training - Step 24257: {'lr': 0.0002784579514881753, 'samples': 12420096, 'steps': 24257, 'loss/train': 1.8726516962051392} 02/25/2022 09:00:43 - INFO - codeparrot_training - Step 24258: {'lr': 0.0002784416953209223, 'samples': 12420608, 'steps': 24258, 'loss/train': 2.4869985580444336} 02/25/2022 09:00:48 - INFO - codeparrot_training - Step 24259: {'lr': 0.00027842543903183406, 'samples': 12421120, 'steps': 24259, 'loss/train': 2.487274646759033} 02/25/2022 09:00:52 - INFO - codeparrot_training - Step 24260: {'lr': 0.0002784091826209803, 'samples': 12421632, 'steps': 24260, 'loss/train': 2.5869758129119873} 02/25/2022 09:00:57 - INFO - codeparrot_training - Step 24261: {'lr': 0.0002783929260884306, 'samples': 12422144, 'steps': 24261, 'loss/train': 1.8004097938537598} 02/25/2022 09:01:01 - INFO - codeparrot_training - Step 24262: {'lr': 0.00027837666943425466, 'samples': 12422656, 'steps': 24262, 'loss/train': 1.555433750152588} 02/25/2022 09:01:06 - INFO - codeparrot_training - Step 24263: {'lr': 0.00027836041265852206, 'samples': 12423168, 'steps': 24263, 'loss/train': 3.006253957748413} 02/25/2022 09:01:10 - INFO - codeparrot_training - Step 24264: {'lr': 0.0002783441557613025, 'samples': 12423680, 'steps': 24264, 'loss/train': 2.2995831966400146} 02/25/2022 09:01:15 - INFO - codeparrot_training - Step 24265: {'lr': 0.0002783278987426655, 'samples': 12424192, 'steps': 24265, 'loss/train': 1.8213459253311157} 02/25/2022 09:01:19 - INFO - codeparrot_training - Step 24266: {'lr': 0.00027831164160268087, 'samples': 12424704, 'steps': 24266, 'loss/train': 1.57319974899292} 02/25/2022 09:01:26 - INFO - codeparrot_training - Step 24267: {'lr': 0.00027829538434141803, 'samples': 12425216, 'steps': 24267, 'loss/train': 4.088140487670898} 02/25/2022 09:01:30 - INFO - codeparrot_training - Step 24268: {'lr': 0.00027827912695894686, 'samples': 12425728, 'steps': 24268, 'loss/train': 0.7101935148239136} 02/25/2022 09:01:35 - INFO - codeparrot_training - Step 24269: {'lr': 0.00027826286945533687, 'samples': 12426240, 'steps': 24269, 'loss/train': 0.826137125492096} 02/25/2022 09:01:39 - INFO - codeparrot_training - Step 24270: {'lr': 0.0002782466118306577, 'samples': 12426752, 'steps': 24270, 'loss/train': 2.115061044692993} 02/25/2022 09:01:44 - INFO - codeparrot_training - Step 24271: {'lr': 0.00027823035408497897, 'samples': 12427264, 'steps': 24271, 'loss/train': 2.5070087909698486} 02/25/2022 09:01:48 - INFO - codeparrot_training - Step 24272: {'lr': 0.0002782140962183704, 'samples': 12427776, 'steps': 24272, 'loss/train': 1.9987293481826782} 02/25/2022 09:01:53 - INFO - codeparrot_training - Step 24273: {'lr': 0.0002781978382309017, 'samples': 12428288, 'steps': 24273, 'loss/train': 1.8195164203643799} 02/25/2022 09:01:57 - INFO - codeparrot_training - Step 24274: {'lr': 0.00027818158012264226, 'samples': 12428800, 'steps': 24274, 'loss/train': 1.359501600265503} 02/25/2022 09:02:02 - INFO - codeparrot_training - Step 24275: {'lr': 0.00027816532189366193, 'samples': 12429312, 'steps': 24275, 'loss/train': 1.328649878501892} 02/25/2022 09:02:06 - INFO - codeparrot_training - Step 24276: {'lr': 0.00027814906354403033, 'samples': 12429824, 'steps': 24276, 'loss/train': 1.9147900342941284} 02/25/2022 09:02:13 - INFO - codeparrot_training - Step 24277: {'lr': 0.00027813280507381713, 'samples': 12430336, 'steps': 24277, 'loss/train': 2.292353630065918} 02/25/2022 09:02:16 - INFO - codeparrot_training - Step 24278: {'lr': 0.0002781165464830918, 'samples': 12430848, 'steps': 24278, 'loss/train': 1.9917237758636475} 02/25/2022 09:02:22 - INFO - codeparrot_training - Step 24279: {'lr': 0.0002781002877719243, 'samples': 12431360, 'steps': 24279, 'loss/train': 0.934857964515686} 02/25/2022 09:02:25 - INFO - codeparrot_training - Step 24280: {'lr': 0.0002780840289403839, 'samples': 12431872, 'steps': 24280, 'loss/train': 1.9658485651016235} 02/25/2022 09:02:31 - INFO - codeparrot_training - Step 24281: {'lr': 0.0002780677699885405, 'samples': 12432384, 'steps': 24281, 'loss/train': 1.8287357091903687} 02/25/2022 09:02:34 - INFO - codeparrot_training - Step 24282: {'lr': 0.0002780515109164637, 'samples': 12432896, 'steps': 24282, 'loss/train': 2.986912250518799} 02/25/2022 09:02:40 - INFO - codeparrot_training - Step 24283: {'lr': 0.00027803525172422316, 'samples': 12433408, 'steps': 24283, 'loss/train': 0.8380170464515686} 02/25/2022 09:02:43 - INFO - codeparrot_training - Step 24284: {'lr': 0.0002780189924118885, 'samples': 12433920, 'steps': 24284, 'loss/train': 1.5723103284835815} 02/25/2022 09:02:51 - INFO - codeparrot_training - Step 24285: {'lr': 0.00027800273297952935, 'samples': 12434432, 'steps': 24285, 'loss/train': 1.9735093116760254} 02/25/2022 09:02:54 - INFO - codeparrot_training - Step 24286: {'lr': 0.0002779864734272154, 'samples': 12434944, 'steps': 24286, 'loss/train': 1.3630189895629883} 02/25/2022 09:03:00 - INFO - codeparrot_training - Step 24287: {'lr': 0.0002779702137550162, 'samples': 12435456, 'steps': 24287, 'loss/train': 1.713210940361023} 02/25/2022 09:03:03 - INFO - codeparrot_training - Step 24288: {'lr': 0.0002779539539630016, 'samples': 12435968, 'steps': 24288, 'loss/train': 1.8743717670440674} 02/25/2022 09:03:09 - INFO - codeparrot_training - Step 24289: {'lr': 0.00027793769405124103, 'samples': 12436480, 'steps': 24289, 'loss/train': 2.632967472076416} 02/25/2022 09:03:12 - INFO - codeparrot_training - Step 24290: {'lr': 0.00027792143401980435, 'samples': 12436992, 'steps': 24290, 'loss/train': 2.1291074752807617} 02/25/2022 09:03:18 - INFO - codeparrot_training - Step 24291: {'lr': 0.0002779051738687611, 'samples': 12437504, 'steps': 24291, 'loss/train': 2.324388027191162} 02/25/2022 09:03:22 - INFO - codeparrot_training - Step 24292: {'lr': 0.0002778889135981809, 'samples': 12438016, 'steps': 24292, 'loss/train': 1.976311206817627} 02/25/2022 09:03:27 - INFO - codeparrot_training - Step 24293: {'lr': 0.00027787265320813344, 'samples': 12438528, 'steps': 24293, 'loss/train': 2.4051411151885986} 02/25/2022 09:03:31 - INFO - codeparrot_training - Step 24294: {'lr': 0.00027785639269868844, 'samples': 12439040, 'steps': 24294, 'loss/train': 2.4954209327697754} 02/25/2022 09:03:38 - INFO - codeparrot_training - Step 24295: {'lr': 0.00027784013206991545, 'samples': 12439552, 'steps': 24295, 'loss/train': 1.896340250968933} 02/25/2022 09:03:41 - INFO - codeparrot_training - Step 24296: {'lr': 0.0002778238713218842, 'samples': 12440064, 'steps': 24296, 'loss/train': 1.7325503826141357} 02/25/2022 09:03:47 - INFO - codeparrot_training - Step 24297: {'lr': 0.0002778076104546643, 'samples': 12440576, 'steps': 24297, 'loss/train': 2.351745843887329} 02/25/2022 09:03:50 - INFO - codeparrot_training - Step 24298: {'lr': 0.0002777913494683255, 'samples': 12441088, 'steps': 24298, 'loss/train': 2.510953664779663} 02/25/2022 09:03:56 - INFO - codeparrot_training - Step 24299: {'lr': 0.0002777750883629373, 'samples': 12441600, 'steps': 24299, 'loss/train': 2.0129010677337646} 02/25/2022 09:03:59 - INFO - codeparrot_training - Step 24300: {'lr': 0.0002777588271385694, 'samples': 12442112, 'steps': 24300, 'loss/train': 2.266029119491577} 02/25/2022 09:04:05 - INFO - codeparrot_training - Step 24301: {'lr': 0.0002777425657952916, 'samples': 12442624, 'steps': 24301, 'loss/train': 2.1747663021087646} 02/25/2022 09:04:08 - INFO - codeparrot_training - Step 24302: {'lr': 0.0002777263043331734, 'samples': 12443136, 'steps': 24302, 'loss/train': 0.8633140921592712} 02/25/2022 09:04:14 - INFO - codeparrot_training - Step 24303: {'lr': 0.00027771004275228465, 'samples': 12443648, 'steps': 24303, 'loss/train': 0.796806275844574} 02/25/2022 09:04:18 - INFO - codeparrot_training - Step 24304: {'lr': 0.00027769378105269466, 'samples': 12444160, 'steps': 24304, 'loss/train': 2.556018829345703} 02/25/2022 09:04:23 - INFO - codeparrot_training - Step 24305: {'lr': 0.0002776775192344735, 'samples': 12444672, 'steps': 24305, 'loss/train': 2.546994209289551} 02/25/2022 09:04:27 - INFO - codeparrot_training - Step 24306: {'lr': 0.0002776612572976905, 'samples': 12445184, 'steps': 24306, 'loss/train': 1.7073020935058594} 02/25/2022 09:04:32 - INFO - codeparrot_training - Step 24307: {'lr': 0.0002776449952424155, 'samples': 12445696, 'steps': 24307, 'loss/train': 1.3263360261917114} 02/25/2022 09:04:36 - INFO - codeparrot_training - Step 24308: {'lr': 0.0002776287330687181, 'samples': 12446208, 'steps': 24308, 'loss/train': 2.3595807552337646} 02/25/2022 09:04:41 - INFO - codeparrot_training - Step 24309: {'lr': 0.000277612470776668, 'samples': 12446720, 'steps': 24309, 'loss/train': 1.9263052940368652} 02/25/2022 09:04:45 - INFO - codeparrot_training - Step 24310: {'lr': 0.0002775962083663349, 'samples': 12447232, 'steps': 24310, 'loss/train': 1.4907540082931519} 02/25/2022 09:04:50 - INFO - codeparrot_training - Step 24311: {'lr': 0.00027757994583778827, 'samples': 12447744, 'steps': 24311, 'loss/train': 2.1533026695251465} 02/25/2022 09:04:54 - INFO - codeparrot_training - Step 24312: {'lr': 0.000277563683191098, 'samples': 12448256, 'steps': 24312, 'loss/train': 2.024930477142334} 02/25/2022 09:05:01 - INFO - codeparrot_training - Step 24313: {'lr': 0.00027754742042633367, 'samples': 12448768, 'steps': 24313, 'loss/train': 2.051766872406006} 02/25/2022 09:05:04 - INFO - codeparrot_training - Step 24314: {'lr': 0.00027753115754356497, 'samples': 12449280, 'steps': 24314, 'loss/train': 2.1391053199768066} 02/25/2022 09:05:10 - INFO - codeparrot_training - Step 24315: {'lr': 0.0002775148945428614, 'samples': 12449792, 'steps': 24315, 'loss/train': 2.4742958545684814} 02/25/2022 09:05:13 - INFO - codeparrot_training - Step 24316: {'lr': 0.00027749863142429294, 'samples': 12450304, 'steps': 24316, 'loss/train': 1.7500590085983276} 02/25/2022 09:05:19 - INFO - codeparrot_training - Step 24317: {'lr': 0.00027748236818792894, 'samples': 12450816, 'steps': 24317, 'loss/train': 1.2185124158859253} 02/25/2022 09:05:22 - INFO - codeparrot_training - Step 24318: {'lr': 0.0002774661048338393, 'samples': 12451328, 'steps': 24318, 'loss/train': 2.03092098236084} 02/25/2022 09:05:30 - INFO - codeparrot_training - Step 24319: {'lr': 0.0002774498413620935, 'samples': 12451840, 'steps': 24319, 'loss/train': 2.5360498428344727} 02/25/2022 09:05:33 - INFO - codeparrot_training - Step 24320: {'lr': 0.0002774335777727613, 'samples': 12452352, 'steps': 24320, 'loss/train': 2.1589839458465576} 02/25/2022 09:05:39 - INFO - codeparrot_training - Step 24321: {'lr': 0.0002774173140659124, 'samples': 12452864, 'steps': 24321, 'loss/train': 1.8409605026245117} 02/25/2022 09:05:42 - INFO - codeparrot_training - Step 24322: {'lr': 0.00027740105024161646, 'samples': 12453376, 'steps': 24322, 'loss/train': 2.1168031692504883} 02/25/2022 09:05:48 - INFO - codeparrot_training - Step 24323: {'lr': 0.00027738478629994306, 'samples': 12453888, 'steps': 24323, 'loss/train': 1.936193585395813} 02/25/2022 09:05:51 - INFO - codeparrot_training - Step 24324: {'lr': 0.00027736852224096196, 'samples': 12454400, 'steps': 24324, 'loss/train': 2.5072245597839355} 02/25/2022 09:05:57 - INFO - codeparrot_training - Step 24325: {'lr': 0.0002773522580647428, 'samples': 12454912, 'steps': 24325, 'loss/train': 3.1440672874450684} 02/25/2022 09:06:00 - INFO - codeparrot_training - Step 24326: {'lr': 0.00027733599377135527, 'samples': 12455424, 'steps': 24326, 'loss/train': 1.4173699617385864} 02/25/2022 09:06:06 - INFO - codeparrot_training - Step 24327: {'lr': 0.00027731972936086895, 'samples': 12455936, 'steps': 24327, 'loss/train': 1.0513856410980225} 02/25/2022 09:06:09 - INFO - codeparrot_training - Step 24328: {'lr': 0.00027730346483335373, 'samples': 12456448, 'steps': 24328, 'loss/train': 1.9477031230926514} 02/25/2022 09:06:17 - INFO - codeparrot_training - Step 24329: {'lr': 0.00027728720018887896, 'samples': 12456960, 'steps': 24329, 'loss/train': 2.603245496749878} 02/25/2022 09:06:20 - INFO - codeparrot_training - Step 24330: {'lr': 0.0002772709354275146, 'samples': 12457472, 'steps': 24330, 'loss/train': 1.859300136566162} 02/25/2022 09:06:26 - INFO - codeparrot_training - Step 24331: {'lr': 0.0002772546705493302, 'samples': 12457984, 'steps': 24331, 'loss/train': 2.2015562057495117} 02/25/2022 09:06:29 - INFO - codeparrot_training - Step 24332: {'lr': 0.0002772384055543954, 'samples': 12458496, 'steps': 24332, 'loss/train': 2.3628835678100586} 02/25/2022 09:06:35 - INFO - codeparrot_training - Step 24333: {'lr': 0.00027722214044278, 'samples': 12459008, 'steps': 24333, 'loss/train': 1.6578948497772217} 02/25/2022 09:06:38 - INFO - codeparrot_training - Step 24334: {'lr': 0.00027720587521455354, 'samples': 12459520, 'steps': 24334, 'loss/train': 1.805680513381958} 02/25/2022 09:06:44 - INFO - codeparrot_training - Step 24335: {'lr': 0.00027718960986978575, 'samples': 12460032, 'steps': 24335, 'loss/train': 2.547279119491577} 02/25/2022 09:06:47 - INFO - codeparrot_training - Step 24336: {'lr': 0.0002771733444085463, 'samples': 12460544, 'steps': 24336, 'loss/train': 1.863051414489746} 02/25/2022 09:06:53 - INFO - codeparrot_training - Step 24337: {'lr': 0.00027715707883090485, 'samples': 12461056, 'steps': 24337, 'loss/train': 1.6383551359176636} 02/25/2022 09:06:57 - INFO - codeparrot_training - Step 24338: {'lr': 0.00027714081313693115, 'samples': 12461568, 'steps': 24338, 'loss/train': 1.4323002099990845} 02/25/2022 09:07:04 - INFO - codeparrot_training - Step 24339: {'lr': 0.0002771245473266948, 'samples': 12462080, 'steps': 24339, 'loss/train': 2.596574068069458} 02/25/2022 09:07:08 - INFO - codeparrot_training - Step 24340: {'lr': 0.00027710828140026553, 'samples': 12462592, 'steps': 24340, 'loss/train': 1.6517757177352905} 02/25/2022 09:07:13 - INFO - codeparrot_training - Step 24341: {'lr': 0.0002770920153577129, 'samples': 12463104, 'steps': 24341, 'loss/train': 1.9912539720535278} 02/25/2022 09:07:17 - INFO - codeparrot_training - Step 24342: {'lr': 0.00027707574919910683, 'samples': 12463616, 'steps': 24342, 'loss/train': 2.359637498855591} 02/25/2022 09:07:22 - INFO - codeparrot_training - Step 24343: {'lr': 0.0002770594829245167, 'samples': 12464128, 'steps': 24343, 'loss/train': 1.962052583694458} 02/25/2022 09:07:26 - INFO - codeparrot_training - Step 24344: {'lr': 0.00027704321653401244, 'samples': 12464640, 'steps': 24344, 'loss/train': 2.239534854888916} 02/25/2022 09:07:31 - INFO - codeparrot_training - Step 24345: {'lr': 0.00027702695002766357, 'samples': 12465152, 'steps': 24345, 'loss/train': 0.7575122714042664} 02/25/2022 09:07:35 - INFO - codeparrot_training - Step 24346: {'lr': 0.0002770106834055398, 'samples': 12465664, 'steps': 24346, 'loss/train': 1.8983911275863647} 02/25/2022 09:07:40 - INFO - codeparrot_training - Step 24347: {'lr': 0.000276994416667711, 'samples': 12466176, 'steps': 24347, 'loss/train': 2.28899884223938} 02/25/2022 09:07:44 - INFO - codeparrot_training - Step 24348: {'lr': 0.0002769781498142465, 'samples': 12466688, 'steps': 24348, 'loss/train': 2.139613151550293} 02/25/2022 09:07:49 - INFO - codeparrot_training - Step 24349: {'lr': 0.0002769618828452163, 'samples': 12467200, 'steps': 24349, 'loss/train': 2.573157787322998} 02/25/2022 09:07:53 - INFO - codeparrot_training - Step 24350: {'lr': 0.00027694561576068985, 'samples': 12467712, 'steps': 24350, 'loss/train': 2.0471861362457275} 02/25/2022 09:08:00 - INFO - codeparrot_training - Step 24351: {'lr': 0.00027692934856073705, 'samples': 12468224, 'steps': 24351, 'loss/train': 2.3244287967681885} 02/25/2022 09:08:04 - INFO - codeparrot_training - Step 24352: {'lr': 0.0002769130812454274, 'samples': 12468736, 'steps': 24352, 'loss/train': 2.1188852787017822} 02/25/2022 09:08:09 - INFO - codeparrot_training - Step 24353: {'lr': 0.0002768968138148307, 'samples': 12469248, 'steps': 24353, 'loss/train': 2.082770586013794} 02/25/2022 09:08:12 - INFO - codeparrot_training - Step 24354: {'lr': 0.0002768805462690165, 'samples': 12469760, 'steps': 24354, 'loss/train': 2.0079801082611084} 02/25/2022 09:08:18 - INFO - codeparrot_training - Step 24355: {'lr': 0.0002768642786080546, 'samples': 12470272, 'steps': 24355, 'loss/train': 1.2906962633132935} 02/25/2022 09:08:21 - INFO - codeparrot_training - Step 24356: {'lr': 0.0002768480108320147, 'samples': 12470784, 'steps': 24356, 'loss/train': 2.0639445781707764} 02/25/2022 09:08:27 - INFO - codeparrot_training - Step 24357: {'lr': 0.0002768317429409664, 'samples': 12471296, 'steps': 24357, 'loss/train': 1.953052043914795} 02/25/2022 09:08:30 - INFO - codeparrot_training - Step 24358: {'lr': 0.00027681547493497955, 'samples': 12471808, 'steps': 24358, 'loss/train': 1.8935571908950806} 02/25/2022 09:08:36 - INFO - codeparrot_training - Step 24359: {'lr': 0.00027679920681412365, 'samples': 12472320, 'steps': 24359, 'loss/train': 1.1786301136016846} 02/25/2022 09:08:40 - INFO - codeparrot_training - Step 24360: {'lr': 0.0002767829385784684, 'samples': 12472832, 'steps': 24360, 'loss/train': 2.4036977291107178} 02/25/2022 09:08:47 - INFO - codeparrot_training - Step 24361: {'lr': 0.0002767666702280836, 'samples': 12473344, 'steps': 24361, 'loss/train': 2.4852397441864014} 02/25/2022 09:08:50 - INFO - codeparrot_training - Step 24362: {'lr': 0.0002767504017630389, 'samples': 12473856, 'steps': 24362, 'loss/train': 1.4464774131774902} 02/25/2022 09:08:56 - INFO - codeparrot_training - Step 24363: {'lr': 0.00027673413318340397, 'samples': 12474368, 'steps': 24363, 'loss/train': 0.43508830666542053} 02/25/2022 09:08:59 - INFO - codeparrot_training - Step 24364: {'lr': 0.0002767178644892485, 'samples': 12474880, 'steps': 24364, 'loss/train': 1.3368829488754272} 02/25/2022 09:09:05 - INFO - codeparrot_training - Step 24365: {'lr': 0.00027670159568064215, 'samples': 12475392, 'steps': 24365, 'loss/train': 1.8249205350875854} 02/25/2022 09:09:08 - INFO - codeparrot_training - Step 24366: {'lr': 0.00027668532675765466, 'samples': 12475904, 'steps': 24366, 'loss/train': 2.646982431411743} 02/25/2022 09:09:14 - INFO - codeparrot_training - Step 24367: {'lr': 0.00027666905772035573, 'samples': 12476416, 'steps': 24367, 'loss/train': 1.5984379053115845} 02/25/2022 09:09:17 - INFO - codeparrot_training - Step 24368: {'lr': 0.00027665278856881496, 'samples': 12476928, 'steps': 24368, 'loss/train': 2.1124837398529053} 02/25/2022 09:09:23 - INFO - codeparrot_training - Step 24369: {'lr': 0.0002766365193031022, 'samples': 12477440, 'steps': 24369, 'loss/train': 1.7307562828063965} 02/25/2022 09:09:26 - INFO - codeparrot_training - Step 24370: {'lr': 0.00027662024992328697, 'samples': 12477952, 'steps': 24370, 'loss/train': 1.9396568536758423} 02/25/2022 09:09:32 - INFO - codeparrot_training - Step 24371: {'lr': 0.00027660398042943907, 'samples': 12478464, 'steps': 24371, 'loss/train': 2.0503547191619873} 02/25/2022 09:09:35 - INFO - codeparrot_training - Step 24372: {'lr': 0.0002765877108216282, 'samples': 12478976, 'steps': 24372, 'loss/train': 0.8696369528770447} 02/25/2022 09:09:41 - INFO - codeparrot_training - Step 24373: {'lr': 0.000276571441099924, 'samples': 12479488, 'steps': 24373, 'loss/train': 1.862486481666565} 02/25/2022 09:09:45 - INFO - codeparrot_training - Step 24374: {'lr': 0.00027655517126439616, 'samples': 12480000, 'steps': 24374, 'loss/train': 3.0517146587371826} 02/25/2022 09:09:48 - INFO - codeparrot_training - Step 24375: {'lr': 0.00027653890131511445, 'samples': 12480512, 'steps': 24375, 'loss/train': 1.9236218929290771} 02/25/2022 09:09:55 - INFO - codeparrot_training - Step 24376: {'lr': 0.00027652263125214845, 'samples': 12481024, 'steps': 24376, 'loss/train': 1.5654569864273071} 02/25/2022 09:09:59 - INFO - codeparrot_training - Step 24377: {'lr': 0.0002765063610755679, 'samples': 12481536, 'steps': 24377, 'loss/train': 1.2738064527511597} 02/25/2022 09:10:04 - INFO - codeparrot_training - Step 24378: {'lr': 0.0002764900907854426, 'samples': 12482048, 'steps': 24378, 'loss/train': 2.406679630279541} 02/25/2022 09:10:08 - INFO - codeparrot_training - Step 24379: {'lr': 0.0002764738203818421, 'samples': 12482560, 'steps': 24379, 'loss/train': 1.2675119638442993} 02/25/2022 09:10:13 - INFO - codeparrot_training - Step 24380: {'lr': 0.0002764575498648362, 'samples': 12483072, 'steps': 24380, 'loss/train': 1.303342342376709} 02/25/2022 09:10:19 - INFO - codeparrot_training - Step 24381: {'lr': 0.00027644127923449446, 'samples': 12483584, 'steps': 24381, 'loss/train': 1.9849861860275269} 02/25/2022 09:10:22 - INFO - codeparrot_training - Step 24382: {'lr': 0.0002764250084908868, 'samples': 12484096, 'steps': 24382, 'loss/train': 1.4672366380691528} 02/25/2022 09:10:28 - INFO - codeparrot_training - Step 24383: {'lr': 0.00027640873763408273, 'samples': 12484608, 'steps': 24383, 'loss/train': 1.2897140979766846} 02/25/2022 09:10:31 - INFO - codeparrot_training - Step 24384: {'lr': 0.00027639246666415207, 'samples': 12485120, 'steps': 24384, 'loss/train': 1.5851019620895386} 02/25/2022 09:10:39 - INFO - codeparrot_training - Step 24385: {'lr': 0.0002763761955811644, 'samples': 12485632, 'steps': 24385, 'loss/train': 2.6295554637908936} 02/25/2022 09:10:42 - INFO - codeparrot_training - Step 24386: {'lr': 0.00027635992438518954, 'samples': 12486144, 'steps': 24386, 'loss/train': 1.5678340196609497} 02/25/2022 09:10:48 - INFO - codeparrot_training - Step 24387: {'lr': 0.00027634365307629705, 'samples': 12486656, 'steps': 24387, 'loss/train': 2.0508193969726562} 02/25/2022 09:10:51 - INFO - codeparrot_training - Step 24388: {'lr': 0.00027632738165455685, 'samples': 12487168, 'steps': 24388, 'loss/train': 1.8973125219345093} 02/25/2022 09:10:57 - INFO - codeparrot_training - Step 24389: {'lr': 0.00027631111012003836, 'samples': 12487680, 'steps': 24389, 'loss/train': 2.5513358116149902} 02/25/2022 09:11:00 - INFO - codeparrot_training - Step 24390: {'lr': 0.0002762948384728115, 'samples': 12488192, 'steps': 24390, 'loss/train': 0.49132177233695984} 02/25/2022 09:11:06 - INFO - codeparrot_training - Step 24391: {'lr': 0.00027627856671294586, 'samples': 12488704, 'steps': 24391, 'loss/train': 1.7337344884872437} 02/25/2022 09:11:09 - INFO - codeparrot_training - Step 24392: {'lr': 0.00027626229484051126, 'samples': 12489216, 'steps': 24392, 'loss/train': 1.959826946258545} 02/25/2022 09:11:15 - INFO - codeparrot_training - Step 24393: {'lr': 0.00027624602285557725, 'samples': 12489728, 'steps': 24393, 'loss/train': 1.6941568851470947} 02/25/2022 09:11:18 - INFO - codeparrot_training - Step 24394: {'lr': 0.00027622975075821364, 'samples': 12490240, 'steps': 24394, 'loss/train': 1.2322165966033936} 02/25/2022 09:11:24 - INFO - codeparrot_training - Step 24395: {'lr': 0.00027621347854849015, 'samples': 12490752, 'steps': 24395, 'loss/train': 2.565326690673828} 02/25/2022 09:11:27 - INFO - codeparrot_training - Step 24396: {'lr': 0.0002761972062264764, 'samples': 12491264, 'steps': 24396, 'loss/train': 0.1975572109222412} 02/25/2022 09:11:35 - INFO - codeparrot_training - Step 24397: {'lr': 0.0002761809337922422, 'samples': 12491776, 'steps': 24397, 'loss/train': 2.5141966342926025} 02/25/2022 09:11:38 - INFO - codeparrot_training - Step 24398: {'lr': 0.0002761646612458571, 'samples': 12492288, 'steps': 24398, 'loss/train': 1.4061397314071655} 02/25/2022 09:11:44 - INFO - codeparrot_training - Step 24399: {'lr': 0.00027614838858739093, 'samples': 12492800, 'steps': 24399, 'loss/train': 1.66769540309906} 02/25/2022 09:11:47 - INFO - codeparrot_training - Step 24400: {'lr': 0.0002761321158169134, 'samples': 12493312, 'steps': 24400, 'loss/train': 2.850393772125244} 02/25/2022 09:11:53 - INFO - codeparrot_training - Step 24401: {'lr': 0.00027611584293449413, 'samples': 12493824, 'steps': 24401, 'loss/train': 1.828554630279541} 02/25/2022 09:11:56 - INFO - codeparrot_training - Step 24402: {'lr': 0.0002760995699402029, 'samples': 12494336, 'steps': 24402, 'loss/train': 2.0508873462677} 02/25/2022 09:12:02 - INFO - codeparrot_training - Step 24403: {'lr': 0.0002760832968341094, 'samples': 12494848, 'steps': 24403, 'loss/train': 1.4734331369400024} 02/25/2022 09:12:05 - INFO - codeparrot_training - Step 24404: {'lr': 0.00027606702361628337, 'samples': 12495360, 'steps': 24404, 'loss/train': 2.2672884464263916} 02/25/2022 09:12:11 - INFO - codeparrot_training - Step 24405: {'lr': 0.00027605075028679446, 'samples': 12495872, 'steps': 24405, 'loss/train': 2.1354756355285645} 02/25/2022 09:12:14 - INFO - codeparrot_training - Step 24406: {'lr': 0.0002760344768457124, 'samples': 12496384, 'steps': 24406, 'loss/train': 2.779613971710205} 02/25/2022 09:12:22 - INFO - codeparrot_training - Step 24407: {'lr': 0.0002760182032931069, 'samples': 12496896, 'steps': 24407, 'loss/train': 1.361061692237854} 02/25/2022 09:12:25 - INFO - codeparrot_training - Step 24408: {'lr': 0.00027600192962904773, 'samples': 12497408, 'steps': 24408, 'loss/train': 2.293539524078369} 02/25/2022 09:12:31 - INFO - codeparrot_training - Step 24409: {'lr': 0.0002759856558536045, 'samples': 12497920, 'steps': 24409, 'loss/train': 2.024662971496582} 02/25/2022 09:12:34 - INFO - codeparrot_training - Step 24410: {'lr': 0.000275969381966847, 'samples': 12498432, 'steps': 24410, 'loss/train': 1.0239968299865723} 02/25/2022 09:12:40 - INFO - codeparrot_training - Step 24411: {'lr': 0.0002759531079688449, 'samples': 12498944, 'steps': 24411, 'loss/train': 1.3101378679275513} 02/25/2022 09:12:43 - INFO - codeparrot_training - Step 24412: {'lr': 0.000275936833859668, 'samples': 12499456, 'steps': 24412, 'loss/train': 1.9156187772750854} 02/25/2022 09:12:49 - INFO - codeparrot_training - Step 24413: {'lr': 0.0002759205596393859, 'samples': 12499968, 'steps': 24413, 'loss/train': 1.6916019916534424} 02/25/2022 09:12:52 - INFO - codeparrot_training - Step 24414: {'lr': 0.0002759042853080683, 'samples': 12500480, 'steps': 24414, 'loss/train': 0.9117424488067627} 02/25/2022 09:12:58 - INFO - codeparrot_training - Step 24415: {'lr': 0.000275888010865785, 'samples': 12500992, 'steps': 24415, 'loss/train': 1.7600107192993164} 02/25/2022 09:13:01 - INFO - codeparrot_training - Step 24416: {'lr': 0.00027587173631260563, 'samples': 12501504, 'steps': 24416, 'loss/train': 2.988499164581299} 02/25/2022 09:13:07 - INFO - codeparrot_training - Step 24417: {'lr': 0.0002758554616486, 'samples': 12502016, 'steps': 24417, 'loss/train': 1.1107361316680908} 02/25/2022 09:13:10 - INFO - codeparrot_training - Step 24418: {'lr': 0.0002758391868738378, 'samples': 12502528, 'steps': 24418, 'loss/train': 1.1253814697265625} 02/25/2022 09:13:16 - INFO - codeparrot_training - Step 24419: {'lr': 0.0002758229119883888, 'samples': 12503040, 'steps': 24419, 'loss/train': 1.8688368797302246} 02/25/2022 09:13:19 - INFO - codeparrot_training - Step 24420: {'lr': 0.0002758066369923225, 'samples': 12503552, 'steps': 24420, 'loss/train': 3.5818772315979004} 02/25/2022 09:13:25 - INFO - codeparrot_training - Step 24421: {'lr': 0.00027579036188570883, 'samples': 12504064, 'steps': 24421, 'loss/train': 1.5507477521896362} 02/25/2022 09:13:28 - INFO - codeparrot_training - Step 24422: {'lr': 0.00027577408666861744, 'samples': 12504576, 'steps': 24422, 'loss/train': 0.26538732647895813} 02/25/2022 09:13:36 - INFO - codeparrot_training - Step 24423: {'lr': 0.00027575781134111805, 'samples': 12505088, 'steps': 24423, 'loss/train': 2.144705295562744} 02/25/2022 09:13:39 - INFO - codeparrot_training - Step 24424: {'lr': 0.00027574153590328033, 'samples': 12505600, 'steps': 24424, 'loss/train': 0.6362213492393494} 02/25/2022 09:13:44 - INFO - codeparrot_training - Step 24425: {'lr': 0.0002757252603551741, 'samples': 12506112, 'steps': 24425, 'loss/train': 1.3682286739349365} 02/25/2022 09:13:48 - INFO - codeparrot_training - Step 24426: {'lr': 0.00027570898469686896, 'samples': 12506624, 'steps': 24426, 'loss/train': 2.466688871383667} 02/25/2022 09:13:54 - INFO - codeparrot_training - Step 24427: {'lr': 0.00027569270892843474, 'samples': 12507136, 'steps': 24427, 'loss/train': 1.8645234107971191} 02/25/2022 09:13:57 - INFO - codeparrot_training - Step 24428: {'lr': 0.0002756764330499411, 'samples': 12507648, 'steps': 24428, 'loss/train': 2.1611015796661377} 02/25/2022 09:14:03 - INFO - codeparrot_training - Step 24429: {'lr': 0.00027566015706145775, 'samples': 12508160, 'steps': 24429, 'loss/train': 1.5633702278137207} 02/25/2022 09:14:06 - INFO - codeparrot_training - Step 24430: {'lr': 0.00027564388096305446, 'samples': 12508672, 'steps': 24430, 'loss/train': 2.6833713054656982} 02/25/2022 09:14:12 - INFO - codeparrot_training - Step 24431: {'lr': 0.00027562760475480095, 'samples': 12509184, 'steps': 24431, 'loss/train': 2.556901693344116} 02/25/2022 09:14:15 - INFO - codeparrot_training - Step 24432: {'lr': 0.0002756113284367669, 'samples': 12509696, 'steps': 24432, 'loss/train': 2.1286306381225586} 02/25/2022 09:14:23 - INFO - codeparrot_training - Step 24433: {'lr': 0.00027559505200902204, 'samples': 12510208, 'steps': 24433, 'loss/train': 2.053403854370117} 02/25/2022 09:14:26 - INFO - codeparrot_training - Step 24434: {'lr': 0.00027557877547163613, 'samples': 12510720, 'steps': 24434, 'loss/train': 2.115370750427246} 02/25/2022 09:14:32 - INFO - codeparrot_training - Step 24435: {'lr': 0.0002755624988246788, 'samples': 12511232, 'steps': 24435, 'loss/train': 2.078890800476074} 02/25/2022 09:14:35 - INFO - codeparrot_training - Step 24436: {'lr': 0.0002755462220682199, 'samples': 12511744, 'steps': 24436, 'loss/train': 1.6449601650238037} 02/25/2022 09:14:41 - INFO - codeparrot_training - Step 24437: {'lr': 0.0002755299452023291, 'samples': 12512256, 'steps': 24437, 'loss/train': 2.1148951053619385} 02/25/2022 09:14:44 - INFO - codeparrot_training - Step 24438: {'lr': 0.0002755136682270761, 'samples': 12512768, 'steps': 24438, 'loss/train': 2.4138970375061035} 02/25/2022 09:14:50 - INFO - codeparrot_training - Step 24439: {'lr': 0.00027549739114253067, 'samples': 12513280, 'steps': 24439, 'loss/train': 2.2177722454071045} 02/25/2022 09:14:53 - INFO - codeparrot_training - Step 24440: {'lr': 0.00027548111394876254, 'samples': 12513792, 'steps': 24440, 'loss/train': 1.8243474960327148} 02/25/2022 09:14:59 - INFO - codeparrot_training - Step 24441: {'lr': 0.00027546483664584137, 'samples': 12514304, 'steps': 24441, 'loss/train': 2.0535154342651367} 02/25/2022 09:15:03 - INFO - codeparrot_training - Step 24442: {'lr': 0.0002754485592338369, 'samples': 12514816, 'steps': 24442, 'loss/train': 1.728975534439087} 02/25/2022 09:15:10 - INFO - codeparrot_training - Step 24443: {'lr': 0.0002754322817128189, 'samples': 12515328, 'steps': 24443, 'loss/train': 0.6173520088195801} 02/25/2022 09:15:13 - INFO - codeparrot_training - Step 24444: {'lr': 0.0002754160040828571, 'samples': 12515840, 'steps': 24444, 'loss/train': 1.6655817031860352} 02/25/2022 09:15:19 - INFO - codeparrot_training - Step 24445: {'lr': 0.00027539972634402124, 'samples': 12516352, 'steps': 24445, 'loss/train': 1.7090250253677368} 02/25/2022 09:15:22 - INFO - codeparrot_training - Step 24446: {'lr': 0.0002753834484963809, 'samples': 12516864, 'steps': 24446, 'loss/train': 0.6559537053108215} 02/25/2022 09:15:28 - INFO - codeparrot_training - Step 24447: {'lr': 0.00027536717054000605, 'samples': 12517376, 'steps': 24447, 'loss/train': 2.229844093322754} 02/25/2022 09:15:31 - INFO - codeparrot_training - Step 24448: {'lr': 0.00027535089247496627, 'samples': 12517888, 'steps': 24448, 'loss/train': 1.687962293624878} 02/25/2022 09:15:37 - INFO - codeparrot_training - Step 24449: {'lr': 0.0002753346143013313, 'samples': 12518400, 'steps': 24449, 'loss/train': 1.6801073551177979} 02/25/2022 09:15:40 - INFO - codeparrot_training - Step 24450: {'lr': 0.00027531833601917086, 'samples': 12518912, 'steps': 24450, 'loss/train': 1.8685766458511353} 02/25/2022 09:15:48 - INFO - codeparrot_training - Step 24451: {'lr': 0.00027530205762855476, 'samples': 12519424, 'steps': 24451, 'loss/train': 1.3363102674484253} 02/25/2022 09:15:51 - INFO - codeparrot_training - Step 24452: {'lr': 0.0002752857791295526, 'samples': 12519936, 'steps': 24452, 'loss/train': 2.0327751636505127} 02/25/2022 09:15:57 - INFO - codeparrot_training - Step 24453: {'lr': 0.0002752695005222343, 'samples': 12520448, 'steps': 24453, 'loss/train': 2.1356115341186523} 02/25/2022 09:16:00 - INFO - codeparrot_training - Step 24454: {'lr': 0.0002752532218066694, 'samples': 12520960, 'steps': 24454, 'loss/train': 0.9559314846992493} 02/25/2022 09:16:06 - INFO - codeparrot_training - Step 24455: {'lr': 0.0002752369429829278, 'samples': 12521472, 'steps': 24455, 'loss/train': 1.6489819288253784} 02/25/2022 09:16:09 - INFO - codeparrot_training - Step 24456: {'lr': 0.00027522066405107906, 'samples': 12521984, 'steps': 24456, 'loss/train': 0.7534356117248535} 02/25/2022 09:16:15 - INFO - codeparrot_training - Step 24457: {'lr': 0.00027520438501119304, 'samples': 12522496, 'steps': 24457, 'loss/train': 2.2548422813415527} 02/25/2022 09:16:18 - INFO - codeparrot_training - Step 24458: {'lr': 0.0002751881058633394, 'samples': 12523008, 'steps': 24458, 'loss/train': 2.4279603958129883} 02/25/2022 09:16:24 - INFO - codeparrot_training - Step 24459: {'lr': 0.000275171826607588, 'samples': 12523520, 'steps': 24459, 'loss/train': 2.2388763427734375} 02/25/2022 09:16:27 - INFO - codeparrot_training - Step 24460: {'lr': 0.0002751555472440084, 'samples': 12524032, 'steps': 24460, 'loss/train': 2.459434986114502} 02/25/2022 09:16:33 - INFO - codeparrot_training - Step 24461: {'lr': 0.00027513926777267045, 'samples': 12524544, 'steps': 24461, 'loss/train': 2.8915839195251465} 02/25/2022 09:16:36 - INFO - codeparrot_training - Step 24462: {'lr': 0.00027512298819364387, 'samples': 12525056, 'steps': 24462, 'loss/train': 2.3681576251983643} 02/25/2022 09:16:42 - INFO - codeparrot_training - Step 24463: {'lr': 0.0002751067085069984, 'samples': 12525568, 'steps': 24463, 'loss/train': 1.0342020988464355} 02/25/2022 09:16:45 - INFO - codeparrot_training - Step 24464: {'lr': 0.0002750904287128037, 'samples': 12526080, 'steps': 24464, 'loss/train': 0.5249140858650208} 02/25/2022 09:16:51 - INFO - codeparrot_training - Step 24465: {'lr': 0.0002750741488111297, 'samples': 12526592, 'steps': 24465, 'loss/train': 1.663329005241394} 02/25/2022 09:16:54 - INFO - codeparrot_training - Step 24466: {'lr': 0.00027505786880204587, 'samples': 12527104, 'steps': 24466, 'loss/train': 1.4527299404144287} 02/25/2022 09:17:02 - INFO - codeparrot_training - Step 24467: {'lr': 0.0002750415886856222, 'samples': 12527616, 'steps': 24467, 'loss/train': 1.3786159753799438} 02/25/2022 09:17:05 - INFO - codeparrot_training - Step 24468: {'lr': 0.0002750253084619282, 'samples': 12528128, 'steps': 24468, 'loss/train': 1.8913116455078125} 02/25/2022 09:17:11 - INFO - codeparrot_training - Step 24469: {'lr': 0.00027500902813103385, 'samples': 12528640, 'steps': 24469, 'loss/train': 1.9512361288070679} 02/25/2022 09:17:14 - INFO - codeparrot_training - Step 24470: {'lr': 0.00027499274769300863, 'samples': 12529152, 'steps': 24470, 'loss/train': 2.614431142807007} 02/25/2022 09:17:19 - INFO - codeparrot_training - Step 24471: {'lr': 0.0002749764671479225, 'samples': 12529664, 'steps': 24471, 'loss/train': 1.3377066850662231} 02/25/2022 09:17:23 - INFO - codeparrot_training - Step 24472: {'lr': 0.00027496018649584506, 'samples': 12530176, 'steps': 24472, 'loss/train': 2.480290412902832} 02/25/2022 09:17:29 - INFO - codeparrot_training - Step 24473: {'lr': 0.00027494390573684624, 'samples': 12530688, 'steps': 24473, 'loss/train': 1.292910099029541} 02/25/2022 09:17:32 - INFO - codeparrot_training - Step 24474: {'lr': 0.0002749276248709955, 'samples': 12531200, 'steps': 24474, 'loss/train': 0.7167650461196899} 02/25/2022 09:17:37 - INFO - codeparrot_training - Step 24475: {'lr': 0.00027491134389836283, 'samples': 12531712, 'steps': 24475, 'loss/train': 1.7323970794677734} 02/25/2022 09:17:41 - INFO - codeparrot_training - Step 24476: {'lr': 0.00027489506281901777, 'samples': 12532224, 'steps': 24476, 'loss/train': 2.0276267528533936} 02/25/2022 09:17:46 - INFO - codeparrot_training - Step 24477: {'lr': 0.0002748787816330302, 'samples': 12532736, 'steps': 24477, 'loss/train': 1.820704698562622} 02/25/2022 09:17:50 - INFO - codeparrot_training - Step 24478: {'lr': 0.0002748625003404699, 'samples': 12533248, 'steps': 24478, 'loss/train': 1.592031717300415} 02/25/2022 09:17:57 - INFO - codeparrot_training - Step 24479: {'lr': 0.0002748462189414065, 'samples': 12533760, 'steps': 24479, 'loss/train': 2.2885358333587646} 02/25/2022 09:18:01 - INFO - codeparrot_training - Step 24480: {'lr': 0.0002748299374359098, 'samples': 12534272, 'steps': 24480, 'loss/train': 1.6717679500579834} 02/25/2022 09:18:06 - INFO - codeparrot_training - Step 24481: {'lr': 0.00027481365582404947, 'samples': 12534784, 'steps': 24481, 'loss/train': 1.6308966875076294} 02/25/2022 09:18:10 - INFO - codeparrot_training - Step 24482: {'lr': 0.00027479737410589537, 'samples': 12535296, 'steps': 24482, 'loss/train': 2.250183582305908} 02/25/2022 09:18:15 - INFO - codeparrot_training - Step 24483: {'lr': 0.0002747810922815172, 'samples': 12535808, 'steps': 24483, 'loss/train': 1.4832526445388794} 02/25/2022 09:18:18 - INFO - codeparrot_training - Step 24484: {'lr': 0.0002747648103509847, 'samples': 12536320, 'steps': 24484, 'loss/train': 1.9258743524551392} 02/25/2022 09:18:24 - INFO - codeparrot_training - Step 24485: {'lr': 0.0002747485283143676, 'samples': 12536832, 'steps': 24485, 'loss/train': 1.9167473316192627} 02/25/2022 09:18:27 - INFO - codeparrot_training - Step 24486: {'lr': 0.00027473224617173564, 'samples': 12537344, 'steps': 24486, 'loss/train': 1.9225873947143555} 02/25/2022 09:18:33 - INFO - codeparrot_training - Step 24487: {'lr': 0.0002747159639231586, 'samples': 12537856, 'steps': 24487, 'loss/train': 1.394760012626648} 02/25/2022 09:18:36 - INFO - codeparrot_training - Step 24488: {'lr': 0.00027469968156870625, 'samples': 12538368, 'steps': 24488, 'loss/train': 1.8857632875442505} 02/25/2022 09:18:43 - INFO - codeparrot_training - Step 24489: {'lr': 0.0002746833991084483, 'samples': 12538880, 'steps': 24489, 'loss/train': 0.7643253803253174} 02/25/2022 09:18:46 - INFO - codeparrot_training - Step 24490: {'lr': 0.0002746671165424545, 'samples': 12539392, 'steps': 24490, 'loss/train': 2.0565500259399414} 02/25/2022 09:18:52 - INFO - codeparrot_training - Step 24491: {'lr': 0.0002746508338707946, 'samples': 12539904, 'steps': 24491, 'loss/train': 2.6000890731811523} 02/25/2022 09:18:55 - INFO - codeparrot_training - Step 24492: {'lr': 0.00027463455109353837, 'samples': 12540416, 'steps': 24492, 'loss/train': 1.6585088968276978} 02/25/2022 09:19:01 - INFO - codeparrot_training - Step 24493: {'lr': 0.00027461826821075554, 'samples': 12540928, 'steps': 24493, 'loss/train': 1.6845992803573608} 02/25/2022 09:19:04 - INFO - codeparrot_training - Step 24494: {'lr': 0.00027460198522251584, 'samples': 12541440, 'steps': 24494, 'loss/train': 1.7877154350280762} 02/25/2022 09:19:10 - INFO - codeparrot_training - Step 24495: {'lr': 0.0002745857021288891, 'samples': 12541952, 'steps': 24495, 'loss/train': 2.11248517036438} 02/25/2022 09:19:13 - INFO - codeparrot_training - Step 24496: {'lr': 0.000274569418929945, 'samples': 12542464, 'steps': 24496, 'loss/train': 1.868991732597351} 02/25/2022 09:19:19 - INFO - codeparrot_training - Step 24497: {'lr': 0.0002745531356257533, 'samples': 12542976, 'steps': 24497, 'loss/train': 1.831846833229065} 02/25/2022 09:19:22 - INFO - codeparrot_training - Step 24498: {'lr': 0.0002745368522163837, 'samples': 12543488, 'steps': 24498, 'loss/train': 2.2027790546417236} 02/25/2022 09:19:28 - INFO - codeparrot_training - Step 24499: {'lr': 0.0002745205687019061, 'samples': 12544000, 'steps': 24499, 'loss/train': 3.2684645652770996} 02/25/2022 09:19:32 - INFO - codeparrot_training - Step 24500: {'lr': 0.0002745042850823902, 'samples': 12544512, 'steps': 24500, 'loss/train': 1.1635981798171997} 02/25/2022 09:19:37 - INFO - codeparrot_training - Step 24501: {'lr': 0.00027448800135790563, 'samples': 12545024, 'steps': 24501, 'loss/train': 1.4591107368469238} 02/25/2022 09:19:41 - INFO - codeparrot_training - Step 24502: {'lr': 0.00027447171752852237, 'samples': 12545536, 'steps': 24502, 'loss/train': 1.7812879085540771} 02/25/2022 09:19:46 - INFO - codeparrot_training - Step 24503: {'lr': 0.00027445543359430987, 'samples': 12546048, 'steps': 24503, 'loss/train': 1.4782919883728027} 02/25/2022 09:19:50 - INFO - codeparrot_training - Step 24504: {'lr': 0.0002744391495553382, 'samples': 12546560, 'steps': 24504, 'loss/train': 1.7535756826400757} 02/25/2022 09:19:55 - INFO - codeparrot_training - Step 24505: {'lr': 0.00027442286541167686, 'samples': 12547072, 'steps': 24505, 'loss/train': 2.996457576751709} 02/25/2022 09:19:59 - INFO - codeparrot_training - Step 24506: {'lr': 0.00027440658116339577, 'samples': 12547584, 'steps': 24506, 'loss/train': 2.0868303775787354} 02/25/2022 09:20:04 - INFO - codeparrot_training - Step 24507: {'lr': 0.00027439029681056463, 'samples': 12548096, 'steps': 24507, 'loss/train': 1.03550386428833} 02/25/2022 09:20:08 - INFO - codeparrot_training - Step 24508: {'lr': 0.00027437401235325327, 'samples': 12548608, 'steps': 24508, 'loss/train': 2.2904536724090576} 02/25/2022 09:20:14 - INFO - codeparrot_training - Step 24509: {'lr': 0.0002743577277915313, 'samples': 12549120, 'steps': 24509, 'loss/train': 2.1216845512390137} 02/25/2022 09:20:17 - INFO - codeparrot_training - Step 24510: {'lr': 0.0002743414431254686, 'samples': 12549632, 'steps': 24510, 'loss/train': 2.9121487140655518} 02/25/2022 09:20:23 - INFO - codeparrot_training - Step 24511: {'lr': 0.00027432515835513485, 'samples': 12550144, 'steps': 24511, 'loss/train': 2.58669114112854} 02/25/2022 09:20:26 - INFO - codeparrot_training - Step 24512: {'lr': 0.00027430887348059993, 'samples': 12550656, 'steps': 24512, 'loss/train': 2.023613691329956} 02/25/2022 09:20:32 - INFO - codeparrot_training - Step 24513: {'lr': 0.0002742925885019334, 'samples': 12551168, 'steps': 24513, 'loss/train': 1.340361475944519} 02/25/2022 09:20:36 - INFO - codeparrot_training - Step 24514: {'lr': 0.0002742763034192051, 'samples': 12551680, 'steps': 24514, 'loss/train': 0.6868749856948853} 02/25/2022 09:20:42 - INFO - codeparrot_training - Step 24515: {'lr': 0.000274260018232485, 'samples': 12552192, 'steps': 24515, 'loss/train': 2.7125587463378906} 02/25/2022 09:20:45 - INFO - codeparrot_training - Step 24516: {'lr': 0.00027424373294184255, 'samples': 12552704, 'steps': 24516, 'loss/train': 2.369844436645508} 02/25/2022 09:20:51 - INFO - codeparrot_training - Step 24517: {'lr': 0.00027422744754734775, 'samples': 12553216, 'steps': 24517, 'loss/train': 2.318622350692749} 02/25/2022 09:20:54 - INFO - codeparrot_training - Step 24518: {'lr': 0.00027421116204907013, 'samples': 12553728, 'steps': 24518, 'loss/train': 1.1846846342086792} 02/25/2022 09:21:00 - INFO - codeparrot_training - Step 24519: {'lr': 0.0002741948764470797, 'samples': 12554240, 'steps': 24519, 'loss/train': 1.3930130004882812} 02/25/2022 09:21:03 - INFO - codeparrot_training - Step 24520: {'lr': 0.000274178590741446, 'samples': 12554752, 'steps': 24520, 'loss/train': 2.1973109245300293} 02/25/2022 09:21:09 - INFO - codeparrot_training - Step 24521: {'lr': 0.00027416230493223896, 'samples': 12555264, 'steps': 24521, 'loss/train': 2.070000171661377} 02/25/2022 09:21:13 - INFO - codeparrot_training - Step 24522: {'lr': 0.0002741460190195283, 'samples': 12555776, 'steps': 24522, 'loss/train': 2.4560000896453857} 02/25/2022 09:21:18 - INFO - codeparrot_training - Step 24523: {'lr': 0.00027412973300338376, 'samples': 12556288, 'steps': 24523, 'loss/train': 1.6704992055892944} 02/25/2022 09:21:22 - INFO - codeparrot_training - Step 24524: {'lr': 0.000274113446883875, 'samples': 12556800, 'steps': 24524, 'loss/train': 1.6079808473587036} 02/25/2022 09:21:28 - INFO - codeparrot_training - Step 24525: {'lr': 0.0002740971606610719, 'samples': 12557312, 'steps': 24525, 'loss/train': 1.5220167636871338} 02/25/2022 09:21:31 - INFO - codeparrot_training - Step 24526: {'lr': 0.00027408087433504437, 'samples': 12557824, 'steps': 24526, 'loss/train': 2.134115219116211} 02/25/2022 09:21:35 - INFO - codeparrot_training - Step 24527: {'lr': 0.0002740645879058619, 'samples': 12558336, 'steps': 24527, 'loss/train': 2.4043474197387695} 02/25/2022 09:21:40 - INFO - codeparrot_training - Step 24528: {'lr': 0.0002740483013735944, 'samples': 12558848, 'steps': 24528, 'loss/train': 1.037956714630127} 02/25/2022 09:21:44 - INFO - codeparrot_training - Step 24529: {'lr': 0.00027403201473831165, 'samples': 12559360, 'steps': 24529, 'loss/train': 2.3583619594573975} 02/25/2022 09:21:49 - INFO - codeparrot_training - Step 24530: {'lr': 0.00027401572800008335, 'samples': 12559872, 'steps': 24530, 'loss/train': 1.2241411209106445} 02/25/2022 09:21:53 - INFO - codeparrot_training - Step 24531: {'lr': 0.0002739994411589792, 'samples': 12560384, 'steps': 24531, 'loss/train': 2.054564952850342} 02/25/2022 09:21:59 - INFO - codeparrot_training - Step 24532: {'lr': 0.00027398315421506926, 'samples': 12560896, 'steps': 24532, 'loss/train': 2.1500895023345947} 02/25/2022 09:22:02 - INFO - codeparrot_training - Step 24533: {'lr': 0.00027396686716842295, 'samples': 12561408, 'steps': 24533, 'loss/train': 1.61677086353302} 02/25/2022 09:22:07 - INFO - codeparrot_training - Step 24534: {'lr': 0.0002739505800191103, 'samples': 12561920, 'steps': 24534, 'loss/train': 1.5186175107955933} 02/25/2022 09:22:11 - INFO - codeparrot_training - Step 24535: {'lr': 0.0002739342927672009, 'samples': 12562432, 'steps': 24535, 'loss/train': 1.5359241962432861} 02/25/2022 09:22:17 - INFO - codeparrot_training - Step 24536: {'lr': 0.00027391800541276464, 'samples': 12562944, 'steps': 24536, 'loss/train': 1.7430449724197388} 02/25/2022 09:22:21 - INFO - codeparrot_training - Step 24537: {'lr': 0.00027390171795587114, 'samples': 12563456, 'steps': 24537, 'loss/train': 1.097501516342163} 02/25/2022 09:22:26 - INFO - codeparrot_training - Step 24538: {'lr': 0.0002738854303965903, 'samples': 12563968, 'steps': 24538, 'loss/train': 1.8211660385131836} 02/25/2022 09:22:32 - INFO - codeparrot_training - Step 24539: {'lr': 0.0002738691427349919, 'samples': 12564480, 'steps': 24539, 'loss/train': 1.9708224534988403} 02/25/2022 09:22:36 - INFO - codeparrot_training - Step 24540: {'lr': 0.00027385285497114563, 'samples': 12564992, 'steps': 24540, 'loss/train': 1.9764999151229858} 02/25/2022 09:22:41 - INFO - codeparrot_training - Step 24541: {'lr': 0.0002738365671051214, 'samples': 12565504, 'steps': 24541, 'loss/train': 1.355342984199524} 02/25/2022 09:22:45 - INFO - codeparrot_training - Step 24542: {'lr': 0.0002738202791369889, 'samples': 12566016, 'steps': 24542, 'loss/train': 2.0183141231536865} 02/25/2022 09:22:50 - INFO - codeparrot_training - Step 24543: {'lr': 0.00027380399106681773, 'samples': 12566528, 'steps': 24543, 'loss/train': 1.5322482585906982} 02/25/2022 09:22:54 - INFO - codeparrot_training - Step 24544: {'lr': 0.00027378770289467795, 'samples': 12567040, 'steps': 24544, 'loss/train': 2.3565030097961426} 02/25/2022 09:23:00 - INFO - codeparrot_training - Step 24545: {'lr': 0.00027377141462063916, 'samples': 12567552, 'steps': 24545, 'loss/train': 1.427379846572876} 02/25/2022 09:23:03 - INFO - codeparrot_training - Step 24546: {'lr': 0.0002737551262447712, 'samples': 12568064, 'steps': 24546, 'loss/train': 1.826188325881958} 02/25/2022 09:23:09 - INFO - codeparrot_training - Step 24547: {'lr': 0.00027373883776714373, 'samples': 12568576, 'steps': 24547, 'loss/train': 1.5599071979522705} 02/25/2022 09:23:12 - INFO - codeparrot_training - Step 24548: {'lr': 0.00027372254918782673, 'samples': 12569088, 'steps': 24548, 'loss/train': 1.6204804182052612} 02/25/2022 09:23:18 - INFO - codeparrot_training - Step 24549: {'lr': 0.0002737062605068898, 'samples': 12569600, 'steps': 24549, 'loss/train': 1.1979007720947266} 02/25/2022 09:23:21 - INFO - codeparrot_training - Step 24550: {'lr': 0.0002736899717244028, 'samples': 12570112, 'steps': 24550, 'loss/train': 0.6492655873298645} 02/25/2022 09:23:27 - INFO - codeparrot_training - Step 24551: {'lr': 0.0002736736828404355, 'samples': 12570624, 'steps': 24551, 'loss/train': 2.2705295085906982} 02/25/2022 09:23:31 - INFO - codeparrot_training - Step 24552: {'lr': 0.0002736573938550577, 'samples': 12571136, 'steps': 24552, 'loss/train': 3.0706064701080322} 02/25/2022 09:23:36 - INFO - codeparrot_training - Step 24553: {'lr': 0.000273641104768339, 'samples': 12571648, 'steps': 24553, 'loss/train': 1.422694444656372} 02/25/2022 09:23:40 - INFO - codeparrot_training - Step 24554: {'lr': 0.0002736248155803494, 'samples': 12572160, 'steps': 24554, 'loss/train': 3.0465056896209717} 02/25/2022 09:23:45 - INFO - codeparrot_training - Step 24555: {'lr': 0.00027360852629115855, 'samples': 12572672, 'steps': 24555, 'loss/train': 5.385129928588867} 02/25/2022 09:23:49 - INFO - codeparrot_training - Step 24556: {'lr': 0.00027359223690083637, 'samples': 12573184, 'steps': 24556, 'loss/train': 1.9099751710891724} 02/25/2022 09:23:54 - INFO - codeparrot_training - Step 24557: {'lr': 0.0002735759474094524, 'samples': 12573696, 'steps': 24557, 'loss/train': 1.010520577430725} 02/25/2022 09:23:58 - INFO - codeparrot_training - Step 24558: {'lr': 0.0002735596578170767, 'samples': 12574208, 'steps': 24558, 'loss/train': 2.2503061294555664} 02/25/2022 09:24:03 - INFO - codeparrot_training - Step 24559: {'lr': 0.00027354336812377875, 'samples': 12574720, 'steps': 24559, 'loss/train': 1.7974720001220703} 02/25/2022 09:24:07 - INFO - codeparrot_training - Step 24560: {'lr': 0.0002735270783296286, 'samples': 12575232, 'steps': 24560, 'loss/train': 1.950779914855957} 02/25/2022 09:24:13 - INFO - codeparrot_training - Step 24561: {'lr': 0.0002735107884346959, 'samples': 12575744, 'steps': 24561, 'loss/train': 1.9560264348983765} 02/25/2022 09:24:17 - INFO - codeparrot_training - Step 24562: {'lr': 0.0002734944984390504, 'samples': 12576256, 'steps': 24562, 'loss/train': 2.540931224822998} 02/25/2022 09:24:22 - INFO - codeparrot_training - Step 24563: {'lr': 0.000273478208342762, 'samples': 12576768, 'steps': 24563, 'loss/train': 0.7010085582733154} 02/25/2022 09:24:26 - INFO - codeparrot_training - Step 24564: {'lr': 0.0002734619181459003, 'samples': 12577280, 'steps': 24564, 'loss/train': 1.374031901359558} 02/25/2022 09:24:32 - INFO - codeparrot_training - Step 24565: {'lr': 0.00027344562784853535, 'samples': 12577792, 'steps': 24565, 'loss/train': 0.8947920203208923} 02/25/2022 09:24:35 - INFO - codeparrot_training - Step 24566: {'lr': 0.0002734293374507367, 'samples': 12578304, 'steps': 24566, 'loss/train': 2.3546180725097656} 02/25/2022 09:24:41 - INFO - codeparrot_training - Step 24567: {'lr': 0.00027341304695257417, 'samples': 12578816, 'steps': 24567, 'loss/train': 1.5401376485824585} 02/25/2022 09:24:44 - INFO - codeparrot_training - Step 24568: {'lr': 0.0002733967563541176, 'samples': 12579328, 'steps': 24568, 'loss/train': 1.9574304819107056} 02/25/2022 09:24:48 - INFO - codeparrot_training - Step 24569: {'lr': 0.0002733804656554368, 'samples': 12579840, 'steps': 24569, 'loss/train': 1.685356855392456} 02/25/2022 09:24:53 - INFO - codeparrot_training - Step 24570: {'lr': 0.0002733641748566015, 'samples': 12580352, 'steps': 24570, 'loss/train': 0.5905944108963013} 02/25/2022 09:25:00 - INFO - codeparrot_training - Step 24571: {'lr': 0.0002733478839576815, 'samples': 12580864, 'steps': 24571, 'loss/train': 2.1740591526031494} 02/25/2022 09:25:03 - INFO - codeparrot_training - Step 24572: {'lr': 0.0002733315929587465, 'samples': 12581376, 'steps': 24572, 'loss/train': 1.9099316596984863} 02/25/2022 09:25:09 - INFO - codeparrot_training - Step 24573: {'lr': 0.00027331530185986636, 'samples': 12581888, 'steps': 24573, 'loss/train': 2.815541982650757} 02/25/2022 09:25:12 - INFO - codeparrot_training - Step 24574: {'lr': 0.00027329901066111095, 'samples': 12582400, 'steps': 24574, 'loss/train': 2.305406332015991} 02/25/2022 09:25:16 - INFO - codeparrot_training - Step 24575: {'lr': 0.00027328271936254996, 'samples': 12582912, 'steps': 24575, 'loss/train': 1.4654350280761719} 02/25/2022 09:25:21 - INFO - codeparrot_training - Step 24576: {'lr': 0.00027326642796425316, 'samples': 12583424, 'steps': 24576, 'loss/train': 3.1276206970214844} 02/25/2022 09:25:25 - INFO - codeparrot_training - Step 24577: {'lr': 0.00027325013646629036, 'samples': 12583936, 'steps': 24577, 'loss/train': 2.556285858154297} 02/25/2022 09:25:30 - INFO - codeparrot_training - Step 24578: {'lr': 0.00027323384486873146, 'samples': 12584448, 'steps': 24578, 'loss/train': 1.7402993440628052} 02/25/2022 09:25:34 - INFO - codeparrot_training - Step 24579: {'lr': 0.00027321755317164605, 'samples': 12584960, 'steps': 24579, 'loss/train': 2.208979606628418} 02/25/2022 09:25:40 - INFO - codeparrot_training - Step 24580: {'lr': 0.0002732012613751041, 'samples': 12585472, 'steps': 24580, 'loss/train': 1.5136798620224} 02/25/2022 09:25:45 - INFO - codeparrot_training - Step 24581: {'lr': 0.0002731849694791752, 'samples': 12585984, 'steps': 24581, 'loss/train': 0.6692734956741333} 02/25/2022 09:25:48 - INFO - codeparrot_training - Step 24582: {'lr': 0.0002731686774839294, 'samples': 12586496, 'steps': 24582, 'loss/train': 2.0074896812438965} 02/25/2022 09:25:55 - INFO - codeparrot_training - Step 24583: {'lr': 0.00027315238538943616, 'samples': 12587008, 'steps': 24583, 'loss/train': 1.5673069953918457} 02/25/2022 09:25:58 - INFO - codeparrot_training - Step 24584: {'lr': 0.0002731360931957656, 'samples': 12587520, 'steps': 24584, 'loss/train': 0.8167309165000916} 02/25/2022 09:26:04 - INFO - codeparrot_training - Step 24585: {'lr': 0.00027311980090298727, 'samples': 12588032, 'steps': 24585, 'loss/train': 1.6703777313232422} 02/25/2022 09:26:07 - INFO - codeparrot_training - Step 24586: {'lr': 0.0002731035085111711, 'samples': 12588544, 'steps': 24586, 'loss/train': 1.4617022275924683} 02/25/2022 09:26:13 - INFO - codeparrot_training - Step 24587: {'lr': 0.00027308721602038684, 'samples': 12589056, 'steps': 24587, 'loss/train': 2.047581434249878} 02/25/2022 09:26:16 - INFO - codeparrot_training - Step 24588: {'lr': 0.0002730709234307043, 'samples': 12589568, 'steps': 24588, 'loss/train': 1.6226900815963745} 02/25/2022 09:26:22 - INFO - codeparrot_training - Step 24589: {'lr': 0.00027305463074219323, 'samples': 12590080, 'steps': 24589, 'loss/train': 2.5973119735717773} 02/25/2022 09:26:25 - INFO - codeparrot_training - Step 24590: {'lr': 0.0002730383379549234, 'samples': 12590592, 'steps': 24590, 'loss/train': 10.280702590942383} 02/25/2022 09:26:31 - INFO - codeparrot_training - Step 24591: {'lr': 0.0002730220450689647, 'samples': 12591104, 'steps': 24591, 'loss/train': 1.4964972734451294} 02/25/2022 09:26:34 - INFO - codeparrot_training - Step 24592: {'lr': 0.00027300575208438684, 'samples': 12591616, 'steps': 24592, 'loss/train': 2.003296375274658} 02/25/2022 09:26:40 - INFO - codeparrot_training - Step 24593: {'lr': 0.00027298945900125965, 'samples': 12592128, 'steps': 24593, 'loss/train': 1.6805553436279297} 02/25/2022 09:26:43 - INFO - codeparrot_training - Step 24594: {'lr': 0.00027297316581965285, 'samples': 12592640, 'steps': 24594, 'loss/train': 2.3026275634765625} 02/25/2022 09:26:49 - INFO - codeparrot_training - Step 24595: {'lr': 0.0002729568725396364, 'samples': 12593152, 'steps': 24595, 'loss/train': 1.7576102018356323} 02/25/2022 09:26:52 - INFO - codeparrot_training - Step 24596: {'lr': 0.00027294057916127997, 'samples': 12593664, 'steps': 24596, 'loss/train': 2.323209762573242} 02/25/2022 09:26:58 - INFO - codeparrot_training - Step 24597: {'lr': 0.0002729242856846533, 'samples': 12594176, 'steps': 24597, 'loss/train': 2.3953471183776855} 02/25/2022 09:27:02 - INFO - codeparrot_training - Step 24598: {'lr': 0.00027290799210982644, 'samples': 12594688, 'steps': 24598, 'loss/train': 1.495805025100708} 02/25/2022 09:27:07 - INFO - codeparrot_training - Step 24599: {'lr': 0.0002728916984368689, 'samples': 12595200, 'steps': 24599, 'loss/train': 0.1257612407207489} 02/25/2022 09:27:11 - INFO - codeparrot_training - Step 24600: {'lr': 0.00027287540466585064, 'samples': 12595712, 'steps': 24600, 'loss/train': 2.241560220718384} 02/25/2022 09:27:16 - INFO - codeparrot_training - Step 24601: {'lr': 0.00027285911079684134, 'samples': 12596224, 'steps': 24601, 'loss/train': 2.553394079208374} 02/25/2022 09:27:20 - INFO - codeparrot_training - Step 24602: {'lr': 0.000272842816829911, 'samples': 12596736, 'steps': 24602, 'loss/train': 1.7803586721420288} 02/25/2022 09:27:25 - INFO - codeparrot_training - Step 24603: {'lr': 0.00027282652276512914, 'samples': 12597248, 'steps': 24603, 'loss/train': 1.9855388402938843} 02/25/2022 09:27:29 - INFO - codeparrot_training - Step 24604: {'lr': 0.00027281022860256576, 'samples': 12597760, 'steps': 24604, 'loss/train': 2.154768228530884} 02/25/2022 09:27:34 - INFO - codeparrot_training - Step 24605: {'lr': 0.0002727939343422906, 'samples': 12598272, 'steps': 24605, 'loss/train': 2.0404345989227295} 02/25/2022 09:27:38 - INFO - codeparrot_training - Step 24606: {'lr': 0.0002727776399843735, 'samples': 12598784, 'steps': 24606, 'loss/train': 2.2131264209747314} 02/25/2022 09:27:44 - INFO - codeparrot_training - Step 24607: {'lr': 0.00027276134552888415, 'samples': 12599296, 'steps': 24607, 'loss/train': 2.552511692047119} 02/25/2022 09:27:47 - INFO - codeparrot_training - Step 24608: {'lr': 0.0002727450509758925, 'samples': 12599808, 'steps': 24608, 'loss/train': 2.0152928829193115} 02/25/2022 09:27:53 - INFO - codeparrot_training - Step 24609: {'lr': 0.0002727287563254682, 'samples': 12600320, 'steps': 24609, 'loss/train': 1.5814917087554932} 02/25/2022 09:27:56 - INFO - codeparrot_training - Step 24610: {'lr': 0.00027271246157768123, 'samples': 12600832, 'steps': 24610, 'loss/train': 2.7921433448791504} 02/25/2022 09:28:02 - INFO - codeparrot_training - Step 24611: {'lr': 0.0002726961667326012, 'samples': 12601344, 'steps': 24611, 'loss/train': 1.7712713479995728} 02/25/2022 09:28:05 - INFO - codeparrot_training - Step 24612: {'lr': 0.000272679871790298, 'samples': 12601856, 'steps': 24612, 'loss/train': 1.2757468223571777} 02/25/2022 09:28:11 - INFO - codeparrot_training - Step 24613: {'lr': 0.00027266357675084145, 'samples': 12602368, 'steps': 24613, 'loss/train': 2.2437453269958496} 02/25/2022 09:28:14 - INFO - codeparrot_training - Step 24614: {'lr': 0.00027264728161430137, 'samples': 12602880, 'steps': 24614, 'loss/train': 0.7529964447021484} 02/25/2022 09:28:20 - INFO - codeparrot_training - Step 24615: {'lr': 0.00027263098638074753, 'samples': 12603392, 'steps': 24615, 'loss/train': 1.377153754234314} 02/25/2022 09:28:23 - INFO - codeparrot_training - Step 24616: {'lr': 0.0002726146910502496, 'samples': 12603904, 'steps': 24616, 'loss/train': 2.3080670833587646} 02/25/2022 09:28:30 - INFO - codeparrot_training - Step 24617: {'lr': 0.0002725983956228777, 'samples': 12604416, 'steps': 24617, 'loss/train': 0.46623751521110535} 02/25/2022 09:28:33 - INFO - codeparrot_training - Step 24618: {'lr': 0.0002725821000987013, 'samples': 12604928, 'steps': 24618, 'loss/train': 1.3109983205795288} 02/25/2022 09:28:39 - INFO - codeparrot_training - Step 24619: {'lr': 0.00027256580447779043, 'samples': 12605440, 'steps': 24619, 'loss/train': 2.421919107437134} 02/25/2022 09:28:42 - INFO - codeparrot_training - Step 24620: {'lr': 0.0002725495087602148, 'samples': 12605952, 'steps': 24620, 'loss/train': 1.560776948928833} 02/25/2022 09:28:48 - INFO - codeparrot_training - Step 24621: {'lr': 0.0002725332129460442, 'samples': 12606464, 'steps': 24621, 'loss/train': 1.8630948066711426} 02/25/2022 09:28:51 - INFO - codeparrot_training - Step 24622: {'lr': 0.00027251691703534853, 'samples': 12606976, 'steps': 24622, 'loss/train': 2.2826967239379883} 02/25/2022 09:28:57 - INFO - codeparrot_training - Step 24623: {'lr': 0.0002725006210281975, 'samples': 12607488, 'steps': 24623, 'loss/train': 1.3786349296569824} 02/25/2022 09:29:00 - INFO - codeparrot_training - Step 24624: {'lr': 0.000272484324924661, 'samples': 12608000, 'steps': 24624, 'loss/train': 2.12591552734375} 02/25/2022 09:29:06 - INFO - codeparrot_training - Step 24625: {'lr': 0.00027246802872480877, 'samples': 12608512, 'steps': 24625, 'loss/train': 1.7557646036148071} 02/25/2022 09:29:09 - INFO - codeparrot_training - Step 24626: {'lr': 0.0002724517324287106, 'samples': 12609024, 'steps': 24626, 'loss/train': 1.381507396697998} 02/25/2022 09:29:15 - INFO - codeparrot_training - Step 24627: {'lr': 0.00027243543603643636, 'samples': 12609536, 'steps': 24627, 'loss/train': 1.5252448320388794} 02/25/2022 09:29:19 - INFO - codeparrot_training - Step 24628: {'lr': 0.00027241913954805587, 'samples': 12610048, 'steps': 24628, 'loss/train': 2.1617088317871094} 02/25/2022 09:29:24 - INFO - codeparrot_training - Step 24629: {'lr': 0.00027240284296363887, 'samples': 12610560, 'steps': 24629, 'loss/train': 2.0953404903411865} 02/25/2022 09:29:28 - INFO - codeparrot_training - Step 24630: {'lr': 0.00027238654628325524, 'samples': 12611072, 'steps': 24630, 'loss/train': 1.753469467163086} 02/25/2022 09:29:33 - INFO - codeparrot_training - Step 24631: {'lr': 0.00027237024950697473, 'samples': 12611584, 'steps': 24631, 'loss/train': 2.017042398452759} 02/25/2022 09:29:37 - INFO - codeparrot_training - Step 24632: {'lr': 0.0002723539526348671, 'samples': 12612096, 'steps': 24632, 'loss/train': 2.876164674758911} 02/25/2022 09:29:42 - INFO - codeparrot_training - Step 24633: {'lr': 0.00027233765566700235, 'samples': 12612608, 'steps': 24633, 'loss/train': 2.098737955093384} 02/25/2022 09:29:46 - INFO - codeparrot_training - Step 24634: {'lr': 0.00027232135860345017, 'samples': 12613120, 'steps': 24634, 'loss/train': 2.8700051307678223} 02/25/2022 09:29:51 - INFO - codeparrot_training - Step 24635: {'lr': 0.00027230506144428036, 'samples': 12613632, 'steps': 24635, 'loss/train': 1.7338645458221436} 02/25/2022 09:29:55 - INFO - codeparrot_training - Step 24636: {'lr': 0.0002722887641895627, 'samples': 12614144, 'steps': 24636, 'loss/train': 0.9063979983329773} 02/25/2022 09:30:00 - INFO - codeparrot_training - Step 24637: {'lr': 0.0002722724668393671, 'samples': 12614656, 'steps': 24637, 'loss/train': 2.547739267349243} 02/25/2022 09:30:04 - INFO - codeparrot_training - Step 24638: {'lr': 0.00027225616939376325, 'samples': 12615168, 'steps': 24638, 'loss/train': 2.4339704513549805} 02/25/2022 09:30:09 - INFO - codeparrot_training - Step 24639: {'lr': 0.00027223987185282113, 'samples': 12615680, 'steps': 24639, 'loss/train': 2.8214831352233887} 02/25/2022 09:30:13 - INFO - codeparrot_training - Step 24640: {'lr': 0.00027222357421661044, 'samples': 12616192, 'steps': 24640, 'loss/train': 2.649697780609131} 02/25/2022 09:30:18 - INFO - codeparrot_training - Step 24641: {'lr': 0.000272207276485201, 'samples': 12616704, 'steps': 24641, 'loss/train': 1.6627308130264282} 02/25/2022 09:30:22 - INFO - codeparrot_training - Step 24642: {'lr': 0.0002721909786586626, 'samples': 12617216, 'steps': 24642, 'loss/train': 1.9770973920822144} 02/25/2022 09:30:28 - INFO - codeparrot_training - Step 24643: {'lr': 0.00027217468073706516, 'samples': 12617728, 'steps': 24643, 'loss/train': 3.204784393310547} 02/25/2022 09:30:32 - INFO - codeparrot_training - Step 24644: {'lr': 0.0002721583827204784, 'samples': 12618240, 'steps': 24644, 'loss/train': 1.5256271362304688} 02/25/2022 09:30:37 - INFO - codeparrot_training - Step 24645: {'lr': 0.0002721420846089722, 'samples': 12618752, 'steps': 24645, 'loss/train': 0.4611816704273224} 02/25/2022 09:30:41 - INFO - codeparrot_training - Step 24646: {'lr': 0.00027212578640261627, 'samples': 12619264, 'steps': 24646, 'loss/train': 2.130063772201538} 02/25/2022 09:30:46 - INFO - codeparrot_training - Step 24647: {'lr': 0.0002721094881014805, 'samples': 12619776, 'steps': 24647, 'loss/train': 1.3082990646362305} 02/25/2022 09:30:50 - INFO - codeparrot_training - Step 24648: {'lr': 0.0002720931897056347, 'samples': 12620288, 'steps': 24648, 'loss/train': 1.4419804811477661} 02/25/2022 09:30:55 - INFO - codeparrot_training - Step 24649: {'lr': 0.0002720768912151487, 'samples': 12620800, 'steps': 24649, 'loss/train': 2.1319429874420166} 02/25/2022 09:30:59 - INFO - codeparrot_training - Step 24650: {'lr': 0.00027206059263009243, 'samples': 12621312, 'steps': 24650, 'loss/train': 2.5869035720825195} 02/25/2022 09:31:04 - INFO - codeparrot_training - Step 24651: {'lr': 0.00027204429395053545, 'samples': 12621824, 'steps': 24651, 'loss/train': 1.911998987197876} 02/25/2022 09:31:08 - INFO - codeparrot_training - Step 24652: {'lr': 0.0002720279951765478, 'samples': 12622336, 'steps': 24652, 'loss/train': 1.693935751914978} 02/25/2022 09:31:13 - INFO - codeparrot_training - Step 24653: {'lr': 0.0002720116963081991, 'samples': 12622848, 'steps': 24653, 'loss/train': 1.788081407546997} 02/25/2022 09:31:16 - INFO - codeparrot_training - Step 24654: {'lr': 0.00027199539734555937, 'samples': 12623360, 'steps': 24654, 'loss/train': 0.9297451376914978} 02/25/2022 09:31:23 - INFO - codeparrot_training - Step 24655: {'lr': 0.0002719790982886983, 'samples': 12623872, 'steps': 24655, 'loss/train': 1.7606264352798462} 02/25/2022 09:31:26 - INFO - codeparrot_training - Step 24656: {'lr': 0.00027196279913768587, 'samples': 12624384, 'steps': 24656, 'loss/train': 1.6372466087341309} 02/25/2022 09:31:32 - INFO - codeparrot_training - Step 24657: {'lr': 0.00027194649989259164, 'samples': 12624896, 'steps': 24657, 'loss/train': 1.7308844327926636} 02/25/2022 09:31:35 - INFO - codeparrot_training - Step 24658: {'lr': 0.0002719302005534856, 'samples': 12625408, 'steps': 24658, 'loss/train': 1.886879324913025} 02/25/2022 09:31:41 - INFO - codeparrot_training - Step 24659: {'lr': 0.0002719139011204376, 'samples': 12625920, 'steps': 24659, 'loss/train': 2.9859209060668945} 02/25/2022 09:31:45 - INFO - codeparrot_training - Step 24660: {'lr': 0.00027189760159351735, 'samples': 12626432, 'steps': 24660, 'loss/train': 1.5303181409835815} 02/25/2022 09:31:50 - INFO - codeparrot_training - Step 24661: {'lr': 0.00027188130197279477, 'samples': 12626944, 'steps': 24661, 'loss/train': 1.9257639646530151} 02/25/2022 09:31:54 - INFO - codeparrot_training - Step 24662: {'lr': 0.00027186500225833955, 'samples': 12627456, 'steps': 24662, 'loss/train': 2.1705539226531982} 02/25/2022 09:31:59 - INFO - codeparrot_training - Step 24663: {'lr': 0.00027184870245022173, 'samples': 12627968, 'steps': 24663, 'loss/train': 1.9565694332122803} 02/25/2022 09:32:03 - INFO - codeparrot_training - Step 24664: {'lr': 0.0002718324025485109, 'samples': 12628480, 'steps': 24664, 'loss/train': 1.956503987312317} 02/25/2022 09:32:09 - INFO - codeparrot_training - Step 24665: {'lr': 0.0002718161025532771, 'samples': 12628992, 'steps': 24665, 'loss/train': 1.2346445322036743} 02/25/2022 09:32:13 - INFO - codeparrot_training - Step 24666: {'lr': 0.00027179980246459, 'samples': 12629504, 'steps': 24666, 'loss/train': 1.7928624153137207} 02/25/2022 09:32:18 - INFO - codeparrot_training - Step 24667: {'lr': 0.0002717835022825194, 'samples': 12630016, 'steps': 24667, 'loss/train': 1.4517059326171875} 02/25/2022 09:32:22 - INFO - codeparrot_training - Step 24668: {'lr': 0.0002717672020071352, 'samples': 12630528, 'steps': 24668, 'loss/train': 1.3011809587478638} 02/25/2022 09:32:27 - INFO - codeparrot_training - Step 24669: {'lr': 0.00027175090163850736, 'samples': 12631040, 'steps': 24669, 'loss/train': 1.4450370073318481} 02/25/2022 09:32:31 - INFO - codeparrot_training - Step 24670: {'lr': 0.0002717346011767054, 'samples': 12631552, 'steps': 24670, 'loss/train': 1.423353672027588} 02/25/2022 09:32:36 - INFO - codeparrot_training - Step 24671: {'lr': 0.00027171830062179943, 'samples': 12632064, 'steps': 24671, 'loss/train': 1.6652315855026245} 02/25/2022 09:32:40 - INFO - codeparrot_training - Step 24672: {'lr': 0.0002717019999738591, 'samples': 12632576, 'steps': 24672, 'loss/train': 0.451820969581604} 02/25/2022 09:32:45 - INFO - codeparrot_training - Step 24673: {'lr': 0.0002716856992329543, 'samples': 12633088, 'steps': 24673, 'loss/train': 1.8710403442382812} 02/25/2022 09:32:49 - INFO - codeparrot_training - Step 24674: {'lr': 0.00027166939839915486, 'samples': 12633600, 'steps': 24674, 'loss/train': 1.1517184972763062} 02/25/2022 09:32:55 - INFO - codeparrot_training - Step 24675: {'lr': 0.0002716530974725306, 'samples': 12634112, 'steps': 24675, 'loss/train': 1.9069888591766357} 02/25/2022 09:32:59 - INFO - codeparrot_training - Step 24676: {'lr': 0.0002716367964531513, 'samples': 12634624, 'steps': 24676, 'loss/train': 0.27837708592414856} 02/25/2022 09:33:04 - INFO - codeparrot_training - Step 24677: {'lr': 0.0002716204953410869, 'samples': 12635136, 'steps': 24677, 'loss/train': 1.7817667722702026} 02/25/2022 09:33:08 - INFO - codeparrot_training - Step 24678: {'lr': 0.00027160419413640714, 'samples': 12635648, 'steps': 24678, 'loss/train': 2.4504878520965576} 02/25/2022 09:33:13 - INFO - codeparrot_training - Step 24679: {'lr': 0.0002715878928391818, 'samples': 12636160, 'steps': 24679, 'loss/train': 1.368591547012329} 02/25/2022 09:33:17 - INFO - codeparrot_training - Step 24680: {'lr': 0.0002715715914494809, 'samples': 12636672, 'steps': 24680, 'loss/train': 2.594334602355957} 02/25/2022 09:33:22 - INFO - codeparrot_training - Step 24681: {'lr': 0.00027155528996737404, 'samples': 12637184, 'steps': 24681, 'loss/train': 1.9636738300323486} 02/25/2022 09:33:25 - INFO - codeparrot_training - Step 24682: {'lr': 0.00027153898839293124, 'samples': 12637696, 'steps': 24682, 'loss/train': 1.0482842922210693} 02/25/2022 09:33:31 - INFO - codeparrot_training - Step 24683: {'lr': 0.0002715226867262223, 'samples': 12638208, 'steps': 24683, 'loss/train': 1.1728090047836304} 02/25/2022 09:33:34 - INFO - codeparrot_training - Step 24684: {'lr': 0.00027150638496731684, 'samples': 12638720, 'steps': 24684, 'loss/train': 2.0531039237976074} 02/25/2022 09:33:41 - INFO - codeparrot_training - Step 24685: {'lr': 0.000271490083116285, 'samples': 12639232, 'steps': 24685, 'loss/train': 2.5731382369995117} 02/25/2022 09:33:44 - INFO - codeparrot_training - Step 24686: {'lr': 0.0002714737811731964, 'samples': 12639744, 'steps': 24686, 'loss/train': 2.7383646965026855} 02/25/2022 09:33:50 - INFO - codeparrot_training - Step 24687: {'lr': 0.00027145747913812096, 'samples': 12640256, 'steps': 24687, 'loss/train': 2.3733439445495605} 02/25/2022 09:33:53 - INFO - codeparrot_training - Step 24688: {'lr': 0.00027144117701112844, 'samples': 12640768, 'steps': 24688, 'loss/train': 1.9890474081039429} 02/25/2022 09:33:59 - INFO - codeparrot_training - Step 24689: {'lr': 0.00027142487479228883, 'samples': 12641280, 'steps': 24689, 'loss/train': 1.2939120531082153} 02/25/2022 09:34:02 - INFO - codeparrot_training - Step 24690: {'lr': 0.0002714085724816718, 'samples': 12641792, 'steps': 24690, 'loss/train': 2.0192711353302} 02/25/2022 09:34:08 - INFO - codeparrot_training - Step 24691: {'lr': 0.0002713922700793473, 'samples': 12642304, 'steps': 24691, 'loss/train': 2.2185118198394775} 02/25/2022 09:34:11 - INFO - codeparrot_training - Step 24692: {'lr': 0.000271375967585385, 'samples': 12642816, 'steps': 24692, 'loss/train': 2.1645827293395996} 02/25/2022 09:34:17 - INFO - codeparrot_training - Step 24693: {'lr': 0.000271359664999855, 'samples': 12643328, 'steps': 24693, 'loss/train': 2.1087687015533447} 02/25/2022 09:34:20 - INFO - codeparrot_training - Step 24694: {'lr': 0.0002713433623228268, 'samples': 12643840, 'steps': 24694, 'loss/train': 3.318579912185669} 02/25/2022 09:34:26 - INFO - codeparrot_training - Step 24695: {'lr': 0.00027132705955437047, 'samples': 12644352, 'steps': 24695, 'loss/train': 1.543757677078247} 02/25/2022 09:34:29 - INFO - codeparrot_training - Step 24696: {'lr': 0.00027131075669455584, 'samples': 12644864, 'steps': 24696, 'loss/train': 2.4322471618652344} 02/25/2022 09:34:35 - INFO - codeparrot_training - Step 24697: {'lr': 0.00027129445374345264, 'samples': 12645376, 'steps': 24697, 'loss/train': 1.7436842918395996} 02/25/2022 09:34:38 - INFO - codeparrot_training - Step 24698: {'lr': 0.00027127815070113084, 'samples': 12645888, 'steps': 24698, 'loss/train': 1.5629328489303589} 02/25/2022 09:34:44 - INFO - codeparrot_training - Step 24699: {'lr': 0.0002712618475676601, 'samples': 12646400, 'steps': 24699, 'loss/train': 1.822376012802124} 02/25/2022 09:34:47 - INFO - codeparrot_training - Step 24700: {'lr': 0.00027124554434311045, 'samples': 12646912, 'steps': 24700, 'loss/train': 2.0030741691589355} 02/25/2022 09:34:54 - INFO - codeparrot_training - Step 24701: {'lr': 0.00027122924102755154, 'samples': 12647424, 'steps': 24701, 'loss/train': 2.359635829925537} 02/25/2022 09:34:57 - INFO - codeparrot_training - Step 24702: {'lr': 0.0002712129376210534, 'samples': 12647936, 'steps': 24702, 'loss/train': 1.3259538412094116} 02/25/2022 09:35:03 - INFO - codeparrot_training - Step 24703: {'lr': 0.00027119663412368566, 'samples': 12648448, 'steps': 24703, 'loss/train': 2.0318992137908936} 02/25/2022 09:35:06 - INFO - codeparrot_training - Step 24704: {'lr': 0.0002711803305355184, 'samples': 12648960, 'steps': 24704, 'loss/train': 1.5608978271484375} 02/25/2022 09:35:12 - INFO - codeparrot_training - Step 24705: {'lr': 0.0002711640268566212, 'samples': 12649472, 'steps': 24705, 'loss/train': 2.191253185272217} 02/25/2022 09:35:15 - INFO - codeparrot_training - Step 24706: {'lr': 0.0002711477230870641, 'samples': 12649984, 'steps': 24706, 'loss/train': 1.4878144264221191} 02/25/2022 09:35:21 - INFO - codeparrot_training - Step 24707: {'lr': 0.0002711314192269169, 'samples': 12650496, 'steps': 24707, 'loss/train': 1.6701326370239258} 02/25/2022 09:35:24 - INFO - codeparrot_training - Step 24708: {'lr': 0.0002711151152762493, 'samples': 12651008, 'steps': 24708, 'loss/train': 1.852627158164978} 02/25/2022 09:35:30 - INFO - codeparrot_training - Step 24709: {'lr': 0.00027109881123513137, 'samples': 12651520, 'steps': 24709, 'loss/train': 1.2789192199707031} 02/25/2022 09:35:33 - INFO - codeparrot_training - Step 24710: {'lr': 0.00027108250710363276, 'samples': 12652032, 'steps': 24710, 'loss/train': 2.080981969833374} 02/25/2022 09:35:39 - INFO - codeparrot_training - Step 24711: {'lr': 0.0002710662028818234, 'samples': 12652544, 'steps': 24711, 'loss/train': 2.297793388366699} 02/25/2022 09:35:43 - INFO - codeparrot_training - Step 24712: {'lr': 0.00027104989856977315, 'samples': 12653056, 'steps': 24712, 'loss/train': 2.144005537033081} 02/25/2022 09:35:48 - INFO - codeparrot_training - Step 24713: {'lr': 0.0002710335941675518, 'samples': 12653568, 'steps': 24713, 'loss/train': 2.7341160774230957} 02/25/2022 09:35:52 - INFO - codeparrot_training - Step 24714: {'lr': 0.0002710172896752292, 'samples': 12654080, 'steps': 24714, 'loss/train': 1.7736080884933472} 02/25/2022 09:35:57 - INFO - codeparrot_training - Step 24715: {'lr': 0.00027100098509287525, 'samples': 12654592, 'steps': 24715, 'loss/train': 2.310661792755127} 02/25/2022 09:36:01 - INFO - codeparrot_training - Step 24716: {'lr': 0.0002709846804205597, 'samples': 12655104, 'steps': 24716, 'loss/train': 2.078089475631714} 02/25/2022 09:36:06 - INFO - codeparrot_training - Step 24717: {'lr': 0.0002709683756583524, 'samples': 12655616, 'steps': 24717, 'loss/train': 1.4521042108535767} 02/25/2022 09:36:10 - INFO - codeparrot_training - Step 24718: {'lr': 0.00027095207080632335, 'samples': 12656128, 'steps': 24718, 'loss/train': 1.8167824745178223} 02/25/2022 09:36:15 - INFO - codeparrot_training - Step 24719: {'lr': 0.00027093576586454223, 'samples': 12656640, 'steps': 24719, 'loss/train': 1.5202302932739258} 02/25/2022 09:36:19 - INFO - codeparrot_training - Step 24720: {'lr': 0.00027091946083307894, 'samples': 12657152, 'steps': 24720, 'loss/train': 1.656925082206726} 02/25/2022 09:36:25 - INFO - codeparrot_training - Step 24721: {'lr': 0.00027090315571200326, 'samples': 12657664, 'steps': 24721, 'loss/train': 1.1279889345169067} 02/25/2022 09:36:28 - INFO - codeparrot_training - Step 24722: {'lr': 0.00027088685050138516, 'samples': 12658176, 'steps': 24722, 'loss/train': 1.7200313806533813} 02/25/2022 09:36:34 - INFO - codeparrot_training - Step 24723: {'lr': 0.0002708705452012944, 'samples': 12658688, 'steps': 24723, 'loss/train': 1.6260935068130493} 02/25/2022 09:36:37 - INFO - codeparrot_training - Step 24724: {'lr': 0.0002708542398118009, 'samples': 12659200, 'steps': 24724, 'loss/train': 1.7535221576690674} 02/25/2022 09:36:43 - INFO - codeparrot_training - Step 24725: {'lr': 0.0002708379343329744, 'samples': 12659712, 'steps': 24725, 'loss/train': 1.6087149381637573} 02/25/2022 09:36:47 - INFO - codeparrot_training - Step 24726: {'lr': 0.00027082162876488486, 'samples': 12660224, 'steps': 24726, 'loss/train': 1.9861663579940796} 02/25/2022 09:36:52 - INFO - codeparrot_training - Step 24727: {'lr': 0.000270805323107602, 'samples': 12660736, 'steps': 24727, 'loss/train': 2.083125114440918} 02/25/2022 09:36:56 - INFO - codeparrot_training - Step 24728: {'lr': 0.0002707890173611958, 'samples': 12661248, 'steps': 24728, 'loss/train': 1.7620930671691895} 02/25/2022 09:37:01 - INFO - codeparrot_training - Step 24729: {'lr': 0.000270772711525736, 'samples': 12661760, 'steps': 24729, 'loss/train': 1.5488576889038086} 02/25/2022 09:37:05 - INFO - codeparrot_training - Step 24730: {'lr': 0.00027075640560129255, 'samples': 12662272, 'steps': 24730, 'loss/train': 1.8740463256835938} 02/25/2022 09:37:11 - INFO - codeparrot_training - Step 24731: {'lr': 0.00027074009958793523, 'samples': 12662784, 'steps': 24731, 'loss/train': 2.655433416366577} 02/25/2022 09:37:14 - INFO - codeparrot_training - Step 24732: {'lr': 0.0002707237934857339, 'samples': 12663296, 'steps': 24732, 'loss/train': 2.241673469543457} 02/25/2022 09:37:20 - INFO - codeparrot_training - Step 24733: {'lr': 0.0002707074872947585, 'samples': 12663808, 'steps': 24733, 'loss/train': 2.175262928009033} 02/25/2022 09:37:23 - INFO - codeparrot_training - Step 24734: {'lr': 0.0002706911810150787, 'samples': 12664320, 'steps': 24734, 'loss/train': 2.5209100246429443} 02/25/2022 09:37:29 - INFO - codeparrot_training - Step 24735: {'lr': 0.0002706748746467645, 'samples': 12664832, 'steps': 24735, 'loss/train': 1.610283374786377} 02/25/2022 09:37:32 - INFO - codeparrot_training - Step 24736: {'lr': 0.00027065856818988567, 'samples': 12665344, 'steps': 24736, 'loss/train': 2.2202906608581543} 02/25/2022 09:37:38 - INFO - codeparrot_training - Step 24737: {'lr': 0.00027064226164451213, 'samples': 12665856, 'steps': 24737, 'loss/train': 1.3246862888336182} 02/25/2022 09:37:41 - INFO - codeparrot_training - Step 24738: {'lr': 0.0002706259550107136, 'samples': 12666368, 'steps': 24738, 'loss/train': 2.9630515575408936} 02/25/2022 09:37:47 - INFO - codeparrot_training - Step 24739: {'lr': 0.00027060964828856015, 'samples': 12666880, 'steps': 24739, 'loss/train': 0.3877182602882385} 02/25/2022 09:37:50 - INFO - codeparrot_training - Step 24740: {'lr': 0.0002705933414781214, 'samples': 12667392, 'steps': 24740, 'loss/train': 1.803843379020691} 02/25/2022 09:37:56 - INFO - codeparrot_training - Step 24741: {'lr': 0.00027057703457946747, 'samples': 12667904, 'steps': 24741, 'loss/train': 1.636519193649292} 02/25/2022 09:37:59 - INFO - codeparrot_training - Step 24742: {'lr': 0.0002705607275926679, 'samples': 12668416, 'steps': 24742, 'loss/train': 1.6374881267547607} 02/25/2022 09:38:05 - INFO - codeparrot_training - Step 24743: {'lr': 0.0002705444205177928, 'samples': 12668928, 'steps': 24743, 'loss/train': 1.6519156694412231} 02/25/2022 09:38:09 - INFO - codeparrot_training - Step 24744: {'lr': 0.0002705281133549119, 'samples': 12669440, 'steps': 24744, 'loss/train': 1.5621296167373657} 02/25/2022 09:38:14 - INFO - codeparrot_training - Step 24745: {'lr': 0.000270511806104095, 'samples': 12669952, 'steps': 24745, 'loss/train': 0.5297259092330933} 02/25/2022 09:38:17 - INFO - codeparrot_training - Step 24746: {'lr': 0.0002704954987654122, 'samples': 12670464, 'steps': 24746, 'loss/train': 1.9659370183944702} 02/25/2022 09:38:24 - INFO - codeparrot_training - Step 24747: {'lr': 0.00027047919133893304, 'samples': 12670976, 'steps': 24747, 'loss/train': 2.1376090049743652} 02/25/2022 09:38:27 - INFO - codeparrot_training - Step 24748: {'lr': 0.0002704628838247276, 'samples': 12671488, 'steps': 24748, 'loss/train': 1.825350046157837} 02/25/2022 09:38:33 - INFO - codeparrot_training - Step 24749: {'lr': 0.0002704465762228656, 'samples': 12672000, 'steps': 24749, 'loss/train': 1.8036125898361206} 02/25/2022 09:38:36 - INFO - codeparrot_training - Step 24750: {'lr': 0.00027043026853341707, 'samples': 12672512, 'steps': 24750, 'loss/train': 0.12093210220336914} 02/25/2022 09:38:42 - INFO - codeparrot_training - Step 24751: {'lr': 0.0002704139607564517, 'samples': 12673024, 'steps': 24751, 'loss/train': 1.0983920097351074} 02/25/2022 09:38:47 - INFO - codeparrot_training - Step 24752: {'lr': 0.00027039765289203944, 'samples': 12673536, 'steps': 24752, 'loss/train': 1.9977049827575684} 02/25/2022 09:38:51 - INFO - codeparrot_training - Step 24753: {'lr': 0.00027038134494025, 'samples': 12674048, 'steps': 24753, 'loss/train': 2.3789122104644775} 02/25/2022 09:38:56 - INFO - codeparrot_training - Step 24754: {'lr': 0.00027036503690115347, 'samples': 12674560, 'steps': 24754, 'loss/train': 3.5160694122314453} 02/25/2022 09:39:00 - INFO - codeparrot_training - Step 24755: {'lr': 0.0002703487287748195, 'samples': 12675072, 'steps': 24755, 'loss/train': 2.3558614253997803} 02/25/2022 09:39:06 - INFO - codeparrot_training - Step 24756: {'lr': 0.00027033242056131806, 'samples': 12675584, 'steps': 24756, 'loss/train': 2.561534881591797} 02/25/2022 09:39:10 - INFO - codeparrot_training - Step 24757: {'lr': 0.00027031611226071905, 'samples': 12676096, 'steps': 24757, 'loss/train': 1.4274990558624268} 02/25/2022 09:39:15 - INFO - codeparrot_training - Step 24758: {'lr': 0.0002702998038730923, 'samples': 12676608, 'steps': 24758, 'loss/train': 1.7446138858795166} 02/25/2022 09:39:19 - INFO - codeparrot_training - Step 24759: {'lr': 0.0002702834953985075, 'samples': 12677120, 'steps': 24759, 'loss/train': 1.9842373132705688} 02/25/2022 09:39:24 - INFO - codeparrot_training - Step 24760: {'lr': 0.0002702671868370347, 'samples': 12677632, 'steps': 24760, 'loss/train': 2.2683663368225098} 02/25/2022 09:39:28 - INFO - codeparrot_training - Step 24761: {'lr': 0.0002702508781887437, 'samples': 12678144, 'steps': 24761, 'loss/train': 2.706632614135742} 02/25/2022 09:39:33 - INFO - codeparrot_training - Step 24762: {'lr': 0.00027023456945370446, 'samples': 12678656, 'steps': 24762, 'loss/train': 1.7192577123641968} 02/25/2022 09:39:37 - INFO - codeparrot_training - Step 24763: {'lr': 0.0002702182606319866, 'samples': 12679168, 'steps': 24763, 'loss/train': 2.067255735397339} 02/25/2022 09:39:42 - INFO - codeparrot_training - Step 24764: {'lr': 0.00027020195172366025, 'samples': 12679680, 'steps': 24764, 'loss/train': 2.072300910949707} 02/25/2022 09:39:46 - INFO - codeparrot_training - Step 24765: {'lr': 0.0002701856427287951, 'samples': 12680192, 'steps': 24765, 'loss/train': 1.1736483573913574} 02/25/2022 09:39:52 - INFO - codeparrot_training - Step 24766: {'lr': 0.0002701693336474611, 'samples': 12680704, 'steps': 24766, 'loss/train': 1.6502591371536255} 02/25/2022 09:39:55 - INFO - codeparrot_training - Step 24767: {'lr': 0.000270153024479728, 'samples': 12681216, 'steps': 24767, 'loss/train': 2.540256977081299} 02/25/2022 09:40:01 - INFO - codeparrot_training - Step 24768: {'lr': 0.0002701367152256658, 'samples': 12681728, 'steps': 24768, 'loss/train': 1.3141244649887085} 02/25/2022 09:40:04 - INFO - codeparrot_training - Step 24769: {'lr': 0.0002701204058853443, 'samples': 12682240, 'steps': 24769, 'loss/train': 2.6552889347076416} 02/25/2022 09:40:10 - INFO - codeparrot_training - Step 24770: {'lr': 0.0002701040964588334, 'samples': 12682752, 'steps': 24770, 'loss/train': 0.9351255297660828} 02/25/2022 09:40:14 - INFO - codeparrot_training - Step 24771: {'lr': 0.0002700877869462029, 'samples': 12683264, 'steps': 24771, 'loss/train': 3.129750967025757} 02/25/2022 09:40:17 - INFO - codeparrot_training - Step 24772: {'lr': 0.0002700714773475227, 'samples': 12683776, 'steps': 24772, 'loss/train': 2.1949305534362793} 02/25/2022 09:40:23 - INFO - codeparrot_training - Step 24773: {'lr': 0.00027005516766286265, 'samples': 12684288, 'steps': 24773, 'loss/train': 1.2875661849975586} 02/25/2022 09:40:26 - INFO - codeparrot_training - Step 24774: {'lr': 0.00027003885789229264, 'samples': 12684800, 'steps': 24774, 'loss/train': 2.469822406768799} 02/25/2022 09:40:32 - INFO - codeparrot_training - Step 24775: {'lr': 0.00027002254803588254, 'samples': 12685312, 'steps': 24775, 'loss/train': 1.4646978378295898} 02/25/2022 09:40:35 - INFO - codeparrot_training - Step 24776: {'lr': 0.00027000623809370224, 'samples': 12685824, 'steps': 24776, 'loss/train': 0.8952632546424866} 02/25/2022 09:40:41 - INFO - codeparrot_training - Step 24777: {'lr': 0.0002699899280658215, 'samples': 12686336, 'steps': 24777, 'loss/train': 1.0153728723526} 02/25/2022 09:40:45 - INFO - codeparrot_training - Step 24778: {'lr': 0.00026997361795231027, 'samples': 12686848, 'steps': 24778, 'loss/train': 2.2385873794555664} 02/25/2022 09:40:50 - INFO - codeparrot_training - Step 24779: {'lr': 0.0002699573077532384, 'samples': 12687360, 'steps': 24779, 'loss/train': 1.4095805883407593} 02/25/2022 09:40:54 - INFO - codeparrot_training - Step 24780: {'lr': 0.0002699409974686758, 'samples': 12687872, 'steps': 24780, 'loss/train': 0.6144838929176331} 02/25/2022 09:40:59 - INFO - codeparrot_training - Step 24781: {'lr': 0.0002699246870986923, 'samples': 12688384, 'steps': 24781, 'loss/train': 1.347374677658081} 02/25/2022 09:41:03 - INFO - codeparrot_training - Step 24782: {'lr': 0.00026990837664335763, 'samples': 12688896, 'steps': 24782, 'loss/train': 1.4676218032836914} 02/25/2022 09:41:08 - INFO - codeparrot_training - Step 24783: {'lr': 0.00026989206610274197, 'samples': 12689408, 'steps': 24783, 'loss/train': 2.816321611404419} 02/25/2022 09:41:12 - INFO - codeparrot_training - Step 24784: {'lr': 0.00026987575547691495, 'samples': 12689920, 'steps': 24784, 'loss/train': 1.5207104682922363} 02/25/2022 09:41:17 - INFO - codeparrot_training - Step 24785: {'lr': 0.0002698594447659465, 'samples': 12690432, 'steps': 24785, 'loss/train': 1.9138569831848145} 02/25/2022 09:41:23 - INFO - codeparrot_training - Step 24786: {'lr': 0.0002698431339699065, 'samples': 12690944, 'steps': 24786, 'loss/train': 2.3683454990386963} 02/25/2022 09:41:26 - INFO - codeparrot_training - Step 24787: {'lr': 0.00026982682308886483, 'samples': 12691456, 'steps': 24787, 'loss/train': 1.6553800106048584} 02/25/2022 09:41:32 - INFO - codeparrot_training - Step 24788: {'lr': 0.00026981051212289134, 'samples': 12691968, 'steps': 24788, 'loss/train': 1.0435470342636108} 02/25/2022 09:41:35 - INFO - codeparrot_training - Step 24789: {'lr': 0.00026979420107205594, 'samples': 12692480, 'steps': 24789, 'loss/train': 1.8520112037658691} 02/25/2022 09:41:41 - INFO - codeparrot_training - Step 24790: {'lr': 0.0002697778899364284, 'samples': 12692992, 'steps': 24790, 'loss/train': 1.8776912689208984} 02/25/2022 09:41:44 - INFO - codeparrot_training - Step 24791: {'lr': 0.0002697615787160787, 'samples': 12693504, 'steps': 24791, 'loss/train': 1.3093762397766113} 02/25/2022 09:41:51 - INFO - codeparrot_training - Step 24792: {'lr': 0.0002697452674110766, 'samples': 12694016, 'steps': 24792, 'loss/train': 1.1372944116592407} 02/25/2022 09:41:54 - INFO - codeparrot_training - Step 24793: {'lr': 0.0002697289560214921, 'samples': 12694528, 'steps': 24793, 'loss/train': 2.164684534072876} 02/25/2022 09:42:00 - INFO - codeparrot_training - Step 24794: {'lr': 0.000269712644547395, 'samples': 12695040, 'steps': 24794, 'loss/train': 2.130544424057007} 02/25/2022 09:42:03 - INFO - codeparrot_training - Step 24795: {'lr': 0.0002696963329888552, 'samples': 12695552, 'steps': 24795, 'loss/train': 1.784342646598816} 02/25/2022 09:42:09 - INFO - codeparrot_training - Step 24796: {'lr': 0.00026968002134594265, 'samples': 12696064, 'steps': 24796, 'loss/train': 2.583003044128418} 02/25/2022 09:42:12 - INFO - codeparrot_training - Step 24797: {'lr': 0.000269663709618727, 'samples': 12696576, 'steps': 24797, 'loss/train': 1.6890408992767334} 02/25/2022 09:42:18 - INFO - codeparrot_training - Step 24798: {'lr': 0.00026964739780727836, 'samples': 12697088, 'steps': 24798, 'loss/train': 2.2527852058410645} 02/25/2022 09:42:21 - INFO - codeparrot_training - Step 24799: {'lr': 0.00026963108591166645, 'samples': 12697600, 'steps': 24799, 'loss/train': 1.5724927186965942} 02/25/2022 09:42:27 - INFO - codeparrot_training - Step 24800: {'lr': 0.00026961477393196127, 'samples': 12698112, 'steps': 24800, 'loss/train': 2.0026538372039795} 02/25/2022 09:42:30 - INFO - codeparrot_training - Step 24801: {'lr': 0.00026959846186823253, 'samples': 12698624, 'steps': 24801, 'loss/train': 0.0752396211028099} 02/25/2022 09:42:36 - INFO - codeparrot_training - Step 24802: {'lr': 0.0002695821497205503, 'samples': 12699136, 'steps': 24802, 'loss/train': 2.7291741371154785} 02/25/2022 09:42:40 - INFO - codeparrot_training - Step 24803: {'lr': 0.0002695658374889843, 'samples': 12699648, 'steps': 24803, 'loss/train': 2.9530017375946045} 02/25/2022 09:42:45 - INFO - codeparrot_training - Step 24804: {'lr': 0.0002695495251736045, 'samples': 12700160, 'steps': 24804, 'loss/train': 1.4212478399276733} 02/25/2022 09:42:49 - INFO - codeparrot_training - Step 24805: {'lr': 0.0002695332127744807, 'samples': 12700672, 'steps': 24805, 'loss/train': 1.149299144744873} 02/25/2022 09:42:54 - INFO - codeparrot_training - Step 24806: {'lr': 0.00026951690029168286, 'samples': 12701184, 'steps': 24806, 'loss/train': 1.0933799743652344} 02/25/2022 09:42:58 - INFO - codeparrot_training - Step 24807: {'lr': 0.0002695005877252808, 'samples': 12701696, 'steps': 24807, 'loss/train': 1.3683966398239136} 02/25/2022 09:43:03 - INFO - codeparrot_training - Step 24808: {'lr': 0.00026948427507534435, 'samples': 12702208, 'steps': 24808, 'loss/train': 2.027744770050049} 02/25/2022 09:43:07 - INFO - codeparrot_training - Step 24809: {'lr': 0.00026946796234194356, 'samples': 12702720, 'steps': 24809, 'loss/train': 2.7946319580078125} 02/25/2022 09:43:12 - INFO - codeparrot_training - Step 24810: {'lr': 0.0002694516495251481, 'samples': 12703232, 'steps': 24810, 'loss/train': 2.4352428913116455} 02/25/2022 09:43:16 - INFO - codeparrot_training - Step 24811: {'lr': 0.00026943533662502803, 'samples': 12703744, 'steps': 24811, 'loss/train': 1.74042809009552} 02/25/2022 09:43:22 - INFO - codeparrot_training - Step 24812: {'lr': 0.0002694190236416531, 'samples': 12704256, 'steps': 24812, 'loss/train': 1.9915233850479126} 02/25/2022 09:43:25 - INFO - codeparrot_training - Step 24813: {'lr': 0.0002694027105750933, 'samples': 12704768, 'steps': 24813, 'loss/train': 2.4247989654541016} 02/25/2022 09:43:31 - INFO - codeparrot_training - Step 24814: {'lr': 0.00026938639742541835, 'samples': 12705280, 'steps': 24814, 'loss/train': 1.9830337762832642} 02/25/2022 09:43:34 - INFO - codeparrot_training - Step 24815: {'lr': 0.0002693700841926983, 'samples': 12705792, 'steps': 24815, 'loss/train': 1.444778561592102} 02/25/2022 09:43:40 - INFO - codeparrot_training - Step 24816: {'lr': 0.00026935377087700297, 'samples': 12706304, 'steps': 24816, 'loss/train': 2.545281171798706} 02/25/2022 09:43:43 - INFO - codeparrot_training - Step 24817: {'lr': 0.00026933745747840214, 'samples': 12706816, 'steps': 24817, 'loss/train': 2.891022205352783} 02/25/2022 09:43:49 - INFO - codeparrot_training - Step 24818: {'lr': 0.0002693211439969659, 'samples': 12707328, 'steps': 24818, 'loss/train': 1.9779574871063232} 02/25/2022 09:43:52 - INFO - codeparrot_training - Step 24819: {'lr': 0.00026930483043276394, 'samples': 12707840, 'steps': 24819, 'loss/train': 1.5759530067443848} 02/25/2022 09:43:58 - INFO - codeparrot_training - Step 24820: {'lr': 0.0002692885167858663, 'samples': 12708352, 'steps': 24820, 'loss/train': 2.7322146892547607} 02/25/2022 09:44:01 - INFO - codeparrot_training - Step 24821: {'lr': 0.0002692722030563427, 'samples': 12708864, 'steps': 24821, 'loss/train': 2.908230781555176} 02/25/2022 09:44:07 - INFO - codeparrot_training - Step 24822: {'lr': 0.00026925588924426317, 'samples': 12709376, 'steps': 24822, 'loss/train': 1.0032862424850464} 02/25/2022 09:44:10 - INFO - codeparrot_training - Step 24823: {'lr': 0.0002692395753496974, 'samples': 12709888, 'steps': 24823, 'loss/train': 1.842634916305542} 02/25/2022 09:44:16 - INFO - codeparrot_training - Step 24824: {'lr': 0.00026922326137271554, 'samples': 12710400, 'steps': 24824, 'loss/train': 2.001751184463501} 02/25/2022 09:44:20 - INFO - codeparrot_training - Step 24825: {'lr': 0.0002692069473133872, 'samples': 12710912, 'steps': 24825, 'loss/train': 1.853408932685852} 02/25/2022 09:44:26 - INFO - codeparrot_training - Step 24826: {'lr': 0.0002691906331717825, 'samples': 12711424, 'steps': 24826, 'loss/train': 0.14718309044837952} 02/25/2022 09:44:29 - INFO - codeparrot_training - Step 24827: {'lr': 0.0002691743189479712, 'samples': 12711936, 'steps': 24827, 'loss/train': 2.2776758670806885} 02/25/2022 09:44:35 - INFO - codeparrot_training - Step 24828: {'lr': 0.0002691580046420231, 'samples': 12712448, 'steps': 24828, 'loss/train': 1.8902251720428467} 02/25/2022 09:44:38 - INFO - codeparrot_training - Step 24829: {'lr': 0.00026914169025400833, 'samples': 12712960, 'steps': 24829, 'loss/train': 1.1861977577209473} 02/25/2022 09:44:46 - INFO - codeparrot_training - Step 24830: {'lr': 0.0002691253757839965, 'samples': 12713472, 'steps': 24830, 'loss/train': 1.8247123956680298} 02/25/2022 09:44:49 - INFO - codeparrot_training - Step 24831: {'lr': 0.0002691090612320578, 'samples': 12713984, 'steps': 24831, 'loss/train': 1.8515534400939941} 02/25/2022 09:44:55 - INFO - codeparrot_training - Step 24832: {'lr': 0.0002690927465982619, 'samples': 12714496, 'steps': 24832, 'loss/train': 2.1409778594970703} 02/25/2022 09:44:58 - INFO - codeparrot_training - Step 24833: {'lr': 0.0002690764318826787, 'samples': 12715008, 'steps': 24833, 'loss/train': 0.8388583064079285} 02/25/2022 09:45:04 - INFO - codeparrot_training - Step 24834: {'lr': 0.00026906011708537807, 'samples': 12715520, 'steps': 24834, 'loss/train': 1.1354522705078125} 02/25/2022 09:45:08 - INFO - codeparrot_training - Step 24835: {'lr': 0.0002690438022064301, 'samples': 12716032, 'steps': 24835, 'loss/train': 1.9713672399520874} 02/25/2022 09:45:13 - INFO - codeparrot_training - Step 24836: {'lr': 0.00026902748724590435, 'samples': 12716544, 'steps': 24836, 'loss/train': 2.2519001960754395} 02/25/2022 09:45:17 - INFO - codeparrot_training - Step 24837: {'lr': 0.00026901117220387105, 'samples': 12717056, 'steps': 24837, 'loss/train': 1.6257095336914062} 02/25/2022 09:45:22 - INFO - codeparrot_training - Step 24838: {'lr': 0.0002689948570803998, 'samples': 12717568, 'steps': 24838, 'loss/train': 1.3859540224075317} 02/25/2022 09:45:26 - INFO - codeparrot_training - Step 24839: {'lr': 0.00026897854187556066, 'samples': 12718080, 'steps': 24839, 'loss/train': 1.5837857723236084} 02/25/2022 09:45:31 - INFO - codeparrot_training - Step 24840: {'lr': 0.00026896222658942347, 'samples': 12718592, 'steps': 24840, 'loss/train': 2.2093896865844727} 02/25/2022 09:45:35 - INFO - codeparrot_training - Step 24841: {'lr': 0.00026894591122205813, 'samples': 12719104, 'steps': 24841, 'loss/train': 1.0886955261230469} 02/25/2022 09:45:42 - INFO - codeparrot_training - Step 24842: {'lr': 0.0002689295957735346, 'samples': 12719616, 'steps': 24842, 'loss/train': 1.7096246480941772} 02/25/2022 09:45:46 - INFO - codeparrot_training - Step 24843: {'lr': 0.0002689132802439226, 'samples': 12720128, 'steps': 24843, 'loss/train': 1.952592134475708} 02/25/2022 09:45:51 - INFO - codeparrot_training - Step 24844: {'lr': 0.0002688969646332921, 'samples': 12720640, 'steps': 24844, 'loss/train': 0.9055593609809875} 02/25/2022 09:45:55 - INFO - codeparrot_training - Step 24845: {'lr': 0.00026888064894171307, 'samples': 12721152, 'steps': 24845, 'loss/train': 1.931763768196106} 02/25/2022 09:46:00 - INFO - codeparrot_training - Step 24846: {'lr': 0.0002688643331692553, 'samples': 12721664, 'steps': 24846, 'loss/train': 2.015608549118042} 02/25/2022 09:46:04 - INFO - codeparrot_training - Step 24847: {'lr': 0.00026884801731598873, 'samples': 12722176, 'steps': 24847, 'loss/train': 2.2528727054595947} 02/25/2022 09:46:09 - INFO - codeparrot_training - Step 24848: {'lr': 0.00026883170138198323, 'samples': 12722688, 'steps': 24848, 'loss/train': 2.5241613388061523} 02/25/2022 09:46:13 - INFO - codeparrot_training - Step 24849: {'lr': 0.0002688153853673087, 'samples': 12723200, 'steps': 24849, 'loss/train': 1.9175479412078857} 02/25/2022 09:46:18 - INFO - codeparrot_training - Step 24850: {'lr': 0.000268799069272035, 'samples': 12723712, 'steps': 24850, 'loss/train': 1.4548085927963257} 02/25/2022 09:46:22 - INFO - codeparrot_training - Step 24851: {'lr': 0.00026878275309623215, 'samples': 12724224, 'steps': 24851, 'loss/train': 2.0615615844726562} 02/25/2022 09:46:29 - INFO - codeparrot_training - Step 24852: {'lr': 0.00026876643683996983, 'samples': 12724736, 'steps': 24852, 'loss/train': 1.844257116317749} 02/25/2022 09:46:32 - INFO - codeparrot_training - Step 24853: {'lr': 0.0002687501205033181, 'samples': 12725248, 'steps': 24853, 'loss/train': 1.035587191581726} 02/25/2022 09:46:38 - INFO - codeparrot_training - Step 24854: {'lr': 0.00026873380408634677, 'samples': 12725760, 'steps': 24854, 'loss/train': 1.9487367868423462} 02/25/2022 09:46:41 - INFO - codeparrot_training - Step 24855: {'lr': 0.0002687174875891259, 'samples': 12726272, 'steps': 24855, 'loss/train': 1.5183360576629639} 02/25/2022 09:46:47 - INFO - codeparrot_training - Step 24856: {'lr': 0.00026870117101172517, 'samples': 12726784, 'steps': 24856, 'loss/train': 2.2734506130218506} 02/25/2022 09:46:51 - INFO - codeparrot_training - Step 24857: {'lr': 0.0002686848543542146, 'samples': 12727296, 'steps': 24857, 'loss/train': 1.6893467903137207} 02/25/2022 09:46:56 - INFO - codeparrot_training - Step 24858: {'lr': 0.0002686685376166639, 'samples': 12727808, 'steps': 24858, 'loss/train': 1.6245406866073608} 02/25/2022 09:47:00 - INFO - codeparrot_training - Step 24859: {'lr': 0.0002686522207991433, 'samples': 12728320, 'steps': 24859, 'loss/train': 1.5985651016235352} 02/25/2022 09:47:05 - INFO - codeparrot_training - Step 24860: {'lr': 0.00026863590390172244, 'samples': 12728832, 'steps': 24860, 'loss/train': 1.3258453607559204} 02/25/2022 09:47:08 - INFO - codeparrot_training - Step 24861: {'lr': 0.0002686195869244713, 'samples': 12729344, 'steps': 24861, 'loss/train': 1.8844633102416992} 02/25/2022 09:47:16 - INFO - codeparrot_training - Step 24862: {'lr': 0.0002686032698674597, 'samples': 12729856, 'steps': 24862, 'loss/train': 2.275313138961792} 02/25/2022 09:47:19 - INFO - codeparrot_training - Step 24863: {'lr': 0.00026858695273075764, 'samples': 12730368, 'steps': 24863, 'loss/train': 2.365968942642212} 02/25/2022 09:47:25 - INFO - codeparrot_training - Step 24864: {'lr': 0.000268570635514435, 'samples': 12730880, 'steps': 24864, 'loss/train': 2.141315221786499} 02/25/2022 09:47:28 - INFO - codeparrot_training - Step 24865: {'lr': 0.0002685543182185616, 'samples': 12731392, 'steps': 24865, 'loss/train': 1.274315595626831} 02/25/2022 09:47:34 - INFO - codeparrot_training - Step 24866: {'lr': 0.00026853800084320747, 'samples': 12731904, 'steps': 24866, 'loss/train': 1.9560257196426392} 02/25/2022 09:47:37 - INFO - codeparrot_training - Step 24867: {'lr': 0.0002685216833884423, 'samples': 12732416, 'steps': 24867, 'loss/train': 1.280389666557312} 02/25/2022 09:47:43 - INFO - codeparrot_training - Step 24868: {'lr': 0.0002685053658543363, 'samples': 12732928, 'steps': 24868, 'loss/train': 1.3529318571090698} 02/25/2022 09:47:47 - INFO - codeparrot_training - Step 24869: {'lr': 0.00026848904824095904, 'samples': 12733440, 'steps': 24869, 'loss/train': 2.7981669902801514} 02/25/2022 09:47:52 - INFO - codeparrot_training - Step 24870: {'lr': 0.00026847273054838065, 'samples': 12733952, 'steps': 24870, 'loss/train': 2.1037886142730713} 02/25/2022 09:47:56 - INFO - codeparrot_training - Step 24871: {'lr': 0.0002684564127766709, 'samples': 12734464, 'steps': 24871, 'loss/train': 1.8249258995056152} 02/25/2022 09:48:01 - INFO - codeparrot_training - Step 24872: {'lr': 0.00026844009492589977, 'samples': 12734976, 'steps': 24872, 'loss/train': 1.8887410163879395} 02/25/2022 09:48:05 - INFO - codeparrot_training - Step 24873: {'lr': 0.00026842377699613714, 'samples': 12735488, 'steps': 24873, 'loss/train': 2.372438430786133} 02/25/2022 09:48:08 - INFO - codeparrot_training - Step 24874: {'lr': 0.0002684074589874529, 'samples': 12736000, 'steps': 24874, 'loss/train': 1.2539584636688232} 02/25/2022 09:48:16 - INFO - codeparrot_training - Step 24875: {'lr': 0.0002683911408999169, 'samples': 12736512, 'steps': 24875, 'loss/train': 2.336827278137207} 02/25/2022 09:48:19 - INFO - codeparrot_training - Step 24876: {'lr': 0.00026837482273359907, 'samples': 12737024, 'steps': 24876, 'loss/train': 2.133115530014038} 02/25/2022 09:48:25 - INFO - codeparrot_training - Step 24877: {'lr': 0.0002683585044885694, 'samples': 12737536, 'steps': 24877, 'loss/train': 2.400137186050415} 02/25/2022 09:48:28 - INFO - codeparrot_training - Step 24878: {'lr': 0.0002683421861648977, 'samples': 12738048, 'steps': 24878, 'loss/train': 1.3767703771591187} 02/25/2022 09:48:34 - INFO - codeparrot_training - Step 24879: {'lr': 0.0002683258677626539, 'samples': 12738560, 'steps': 24879, 'loss/train': 2.032876491546631} 02/25/2022 09:48:37 - INFO - codeparrot_training - Step 24880: {'lr': 0.00026830954928190793, 'samples': 12739072, 'steps': 24880, 'loss/train': 1.6309301853179932} 02/25/2022 09:48:43 - INFO - codeparrot_training - Step 24881: {'lr': 0.0002682932307227297, 'samples': 12739584, 'steps': 24881, 'loss/train': 0.4427894651889801} 02/25/2022 09:48:48 - INFO - codeparrot_training - Step 24882: {'lr': 0.00026827691208518897, 'samples': 12740096, 'steps': 24882, 'loss/train': 1.8768963813781738} 02/25/2022 09:48:52 - INFO - codeparrot_training - Step 24883: {'lr': 0.0002682605933693558, 'samples': 12740608, 'steps': 24883, 'loss/train': 1.7223317623138428} 02/25/2022 09:48:55 - INFO - codeparrot_training - Step 24884: {'lr': 0.00026824427457530005, 'samples': 12741120, 'steps': 24884, 'loss/train': 1.645879864692688} 02/25/2022 09:49:01 - INFO - codeparrot_training - Step 24885: {'lr': 0.00026822795570309165, 'samples': 12741632, 'steps': 24885, 'loss/train': 1.984671711921692} 02/25/2022 09:49:06 - INFO - codeparrot_training - Step 24886: {'lr': 0.0002682116367528004, 'samples': 12742144, 'steps': 24886, 'loss/train': 1.413815975189209} 02/25/2022 09:49:10 - INFO - codeparrot_training - Step 24887: {'lr': 0.0002681953177244964, 'samples': 12742656, 'steps': 24887, 'loss/train': 0.42039358615875244} 02/25/2022 09:49:13 - INFO - codeparrot_training - Step 24888: {'lr': 0.00026817899861824934, 'samples': 12743168, 'steps': 24888, 'loss/train': 1.6924446821212769} 02/25/2022 09:49:20 - INFO - codeparrot_training - Step 24889: {'lr': 0.00026816267943412925, 'samples': 12743680, 'steps': 24889, 'loss/train': 1.9290411472320557} 02/25/2022 09:49:26 - INFO - codeparrot_training - Step 24890: {'lr': 0.000268146360172206, 'samples': 12744192, 'steps': 24890, 'loss/train': 2.25826096534729} 02/25/2022 09:49:29 - INFO - codeparrot_training - Step 24891: {'lr': 0.0002681300408325495, 'samples': 12744704, 'steps': 24891, 'loss/train': 1.389148473739624} 02/25/2022 09:49:35 - INFO - codeparrot_training - Step 24892: {'lr': 0.00026811372141522964, 'samples': 12745216, 'steps': 24892, 'loss/train': 1.5085041522979736} 02/25/2022 09:49:38 - INFO - codeparrot_training - Step 24893: {'lr': 0.00026809740192031644, 'samples': 12745728, 'steps': 24893, 'loss/train': 0.5152866840362549} 02/25/2022 09:49:42 - INFO - codeparrot_training - Step 24894: {'lr': 0.0002680810823478797, 'samples': 12746240, 'steps': 24894, 'loss/train': 1.534449815750122} 02/25/2022 09:49:47 - INFO - codeparrot_training - Step 24895: {'lr': 0.0002680647626979893, 'samples': 12746752, 'steps': 24895, 'loss/train': 2.183363199234009} 02/25/2022 09:49:53 - INFO - codeparrot_training - Step 24896: {'lr': 0.00026804844297071524, 'samples': 12747264, 'steps': 24896, 'loss/train': 2.2756173610687256} 02/25/2022 09:49:56 - INFO - codeparrot_training - Step 24897: {'lr': 0.0002680321231661273, 'samples': 12747776, 'steps': 24897, 'loss/train': 1.3891857862472534} 02/25/2022 09:50:04 - INFO - codeparrot_training - Step 24898: {'lr': 0.00026801580328429555, 'samples': 12748288, 'steps': 24898, 'loss/train': 0.12074144184589386} 02/25/2022 09:50:07 - INFO - codeparrot_training - Step 24899: {'lr': 0.0002679994833252897, 'samples': 12748800, 'steps': 24899, 'loss/train': 1.473228096961975} 02/25/2022 09:50:13 - INFO - codeparrot_training - Step 24900: {'lr': 0.0002679831632891799, 'samples': 12749312, 'steps': 24900, 'loss/train': 2.710495948791504} 02/25/2022 09:50:16 - INFO - codeparrot_training - Step 24901: {'lr': 0.00026796684317603584, 'samples': 12749824, 'steps': 24901, 'loss/train': 1.7328565120697021} 02/25/2022 09:50:22 - INFO - codeparrot_training - Step 24902: {'lr': 0.0002679505229859276, 'samples': 12750336, 'steps': 24902, 'loss/train': 1.5025874376296997} 02/25/2022 09:50:25 - INFO - codeparrot_training - Step 24903: {'lr': 0.00026793420271892503, 'samples': 12750848, 'steps': 24903, 'loss/train': 1.9491593837738037} 02/25/2022 09:50:31 - INFO - codeparrot_training - Step 24904: {'lr': 0.000267917882375098, 'samples': 12751360, 'steps': 24904, 'loss/train': 3.3263776302337646} 02/25/2022 09:50:34 - INFO - codeparrot_training - Step 24905: {'lr': 0.00026790156195451647, 'samples': 12751872, 'steps': 24905, 'loss/train': 2.0799412727355957} 02/25/2022 09:50:40 - INFO - codeparrot_training - Step 24906: {'lr': 0.0002678852414572503, 'samples': 12752384, 'steps': 24906, 'loss/train': 0.7658665776252747} 02/25/2022 09:50:43 - INFO - codeparrot_training - Step 24907: {'lr': 0.0002678689208833695, 'samples': 12752896, 'steps': 24907, 'loss/train': 3.0476977825164795} 02/25/2022 09:50:50 - INFO - codeparrot_training - Step 24908: {'lr': 0.0002678526002329438, 'samples': 12753408, 'steps': 24908, 'loss/train': 1.5017503499984741} 02/25/2022 09:50:54 - INFO - codeparrot_training - Step 24909: {'lr': 0.00026783627950604334, 'samples': 12753920, 'steps': 24909, 'loss/train': 2.349515199661255} 02/25/2022 09:50:59 - INFO - codeparrot_training - Step 24910: {'lr': 0.0002678199587027379, 'samples': 12754432, 'steps': 24910, 'loss/train': 1.9906333684921265} 02/25/2022 09:51:03 - INFO - codeparrot_training - Step 24911: {'lr': 0.0002678036378230974, 'samples': 12754944, 'steps': 24911, 'loss/train': 0.7913855910301208} 02/25/2022 09:51:08 - INFO - codeparrot_training - Step 24912: {'lr': 0.0002677873168671918, 'samples': 12755456, 'steps': 24912, 'loss/train': 1.7299515008926392} 02/25/2022 09:51:12 - INFO - codeparrot_training - Step 24913: {'lr': 0.00026777099583509084, 'samples': 12755968, 'steps': 24913, 'loss/train': 3.56026554107666} 02/25/2022 09:51:17 - INFO - codeparrot_training - Step 24914: {'lr': 0.00026775467472686475, 'samples': 12756480, 'steps': 24914, 'loss/train': 1.643774151802063} 02/25/2022 09:51:20 - INFO - codeparrot_training - Step 24915: {'lr': 0.0002677383535425832, 'samples': 12756992, 'steps': 24915, 'loss/train': 1.1217113733291626} 02/25/2022 09:51:26 - INFO - codeparrot_training - Step 24916: {'lr': 0.00026772203228231617, 'samples': 12757504, 'steps': 24916, 'loss/train': 1.094772458076477} 02/25/2022 09:51:29 - INFO - codeparrot_training - Step 24917: {'lr': 0.0002677057109461336, 'samples': 12758016, 'steps': 24917, 'loss/train': 2.081655740737915} 02/25/2022 09:51:37 - INFO - codeparrot_training - Step 24918: {'lr': 0.0002676893895341054, 'samples': 12758528, 'steps': 24918, 'loss/train': 2.005033254623413} 02/25/2022 09:51:40 - INFO - codeparrot_training - Step 24919: {'lr': 0.0002676730680463014, 'samples': 12759040, 'steps': 24919, 'loss/train': 1.8220857381820679} 02/25/2022 09:51:46 - INFO - codeparrot_training - Step 24920: {'lr': 0.0002676567464827917, 'samples': 12759552, 'steps': 24920, 'loss/train': 1.720070719718933} 02/25/2022 09:51:49 - INFO - codeparrot_training - Step 24921: {'lr': 0.00026764042484364603, 'samples': 12760064, 'steps': 24921, 'loss/train': 2.696366310119629} 02/25/2022 09:51:55 - INFO - codeparrot_training - Step 24922: {'lr': 0.0002676241031289344, 'samples': 12760576, 'steps': 24922, 'loss/train': 1.7974830865859985} 02/25/2022 09:51:58 - INFO - codeparrot_training - Step 24923: {'lr': 0.0002676077813387267, 'samples': 12761088, 'steps': 24923, 'loss/train': 0.584373950958252} 02/25/2022 09:52:04 - INFO - codeparrot_training - Step 24924: {'lr': 0.00026759145947309284, 'samples': 12761600, 'steps': 24924, 'loss/train': 1.8534595966339111} 02/25/2022 09:52:07 - INFO - codeparrot_training - Step 24925: {'lr': 0.0002675751375321028, 'samples': 12762112, 'steps': 24925, 'loss/train': 2.036557674407959} 02/25/2022 09:52:13 - INFO - codeparrot_training - Step 24926: {'lr': 0.0002675588155158264, 'samples': 12762624, 'steps': 24926, 'loss/train': 2.2146878242492676} 02/25/2022 09:52:16 - INFO - codeparrot_training - Step 24927: {'lr': 0.0002675424934243337, 'samples': 12763136, 'steps': 24927, 'loss/train': 2.439091444015503} 02/25/2022 09:52:22 - INFO - codeparrot_training - Step 24928: {'lr': 0.0002675261712576944, 'samples': 12763648, 'steps': 24928, 'loss/train': 1.9425225257873535} 02/25/2022 09:52:25 - INFO - codeparrot_training - Step 24929: {'lr': 0.00026750984901597865, 'samples': 12764160, 'steps': 24929, 'loss/train': 1.7834248542785645} 02/25/2022 09:52:31 - INFO - codeparrot_training - Step 24930: {'lr': 0.0002674935266992562, 'samples': 12764672, 'steps': 24930, 'loss/train': 1.4715917110443115} 02/25/2022 09:52:34 - INFO - codeparrot_training - Step 24931: {'lr': 0.0002674772043075971, 'samples': 12765184, 'steps': 24931, 'loss/train': 2.0394773483276367} 02/25/2022 09:52:40 - INFO - codeparrot_training - Step 24932: {'lr': 0.00026746088184107116, 'samples': 12765696, 'steps': 24932, 'loss/train': 1.8661472797393799} 02/25/2022 09:52:43 - INFO - codeparrot_training - Step 24933: {'lr': 0.00026744455929974837, 'samples': 12766208, 'steps': 24933, 'loss/train': 1.2441176176071167} 02/25/2022 09:52:51 - INFO - codeparrot_training - Step 24934: {'lr': 0.0002674282366836986, 'samples': 12766720, 'steps': 24934, 'loss/train': 1.8973724842071533} 02/25/2022 09:52:54 - INFO - codeparrot_training - Step 24935: {'lr': 0.00026741191399299186, 'samples': 12767232, 'steps': 24935, 'loss/train': 2.035752773284912} 02/25/2022 09:53:00 - INFO - codeparrot_training - Step 24936: {'lr': 0.00026739559122769795, 'samples': 12767744, 'steps': 24936, 'loss/train': 2.289137601852417} 02/25/2022 09:53:03 - INFO - codeparrot_training - Step 24937: {'lr': 0.0002673792683878869, 'samples': 12768256, 'steps': 24937, 'loss/train': 1.8673814535140991} 02/25/2022 09:53:09 - INFO - codeparrot_training - Step 24938: {'lr': 0.0002673629454736285, 'samples': 12768768, 'steps': 24938, 'loss/train': 1.6944831609725952} 02/25/2022 09:53:12 - INFO - codeparrot_training - Step 24939: {'lr': 0.0002673466224849928, 'samples': 12769280, 'steps': 24939, 'loss/train': 1.2597445249557495} 02/25/2022 09:53:18 - INFO - codeparrot_training - Step 24940: {'lr': 0.00026733029942204974, 'samples': 12769792, 'steps': 24940, 'loss/train': 1.0115681886672974} 02/25/2022 09:53:21 - INFO - codeparrot_training - Step 24941: {'lr': 0.00026731397628486906, 'samples': 12770304, 'steps': 24941, 'loss/train': 1.9106547832489014} 02/25/2022 09:53:27 - INFO - codeparrot_training - Step 24942: {'lr': 0.00026729765307352093, 'samples': 12770816, 'steps': 24942, 'loss/train': 1.6540048122406006} 02/25/2022 09:53:30 - INFO - codeparrot_training - Step 24943: {'lr': 0.00026728132978807507, 'samples': 12771328, 'steps': 24943, 'loss/train': 2.1333460807800293} 02/25/2022 09:53:38 - INFO - codeparrot_training - Step 24944: {'lr': 0.0002672650064286015, 'samples': 12771840, 'steps': 24944, 'loss/train': 1.900303602218628} 02/25/2022 09:53:41 - INFO - codeparrot_training - Step 24945: {'lr': 0.00026724868299517006, 'samples': 12772352, 'steps': 24945, 'loss/train': 0.08900638669729233} 02/25/2022 09:53:47 - INFO - codeparrot_training - Step 24946: {'lr': 0.00026723235948785084, 'samples': 12772864, 'steps': 24946, 'loss/train': 1.2578349113464355} 02/25/2022 09:53:50 - INFO - codeparrot_training - Step 24947: {'lr': 0.0002672160359067136, 'samples': 12773376, 'steps': 24947, 'loss/train': 1.4240481853485107} 02/25/2022 09:53:56 - INFO - codeparrot_training - Step 24948: {'lr': 0.00026719971225182835, 'samples': 12773888, 'steps': 24948, 'loss/train': 1.9180265665054321} 02/25/2022 09:53:59 - INFO - codeparrot_training - Step 24949: {'lr': 0.00026718338852326504, 'samples': 12774400, 'steps': 24949, 'loss/train': 1.72811758518219} 02/25/2022 09:54:05 - INFO - codeparrot_training - Step 24950: {'lr': 0.0002671670647210934, 'samples': 12774912, 'steps': 24950, 'loss/train': 2.247767210006714} 02/25/2022 09:54:08 - INFO - codeparrot_training - Step 24951: {'lr': 0.0002671507408453837, 'samples': 12775424, 'steps': 24951, 'loss/train': 2.412048578262329} 02/25/2022 09:54:14 - INFO - codeparrot_training - Step 24952: {'lr': 0.0002671344168962055, 'samples': 12775936, 'steps': 24952, 'loss/train': 1.3504868745803833} 02/25/2022 09:54:17 - INFO - codeparrot_training - Step 24953: {'lr': 0.00026711809287362903, 'samples': 12776448, 'steps': 24953, 'loss/train': 1.8935706615447998} 02/25/2022 09:54:25 - INFO - codeparrot_training - Step 24954: {'lr': 0.000267101768777724, 'samples': 12776960, 'steps': 24954, 'loss/train': 1.974847674369812} 02/25/2022 09:54:28 - INFO - codeparrot_training - Step 24955: {'lr': 0.0002670854446085605, 'samples': 12777472, 'steps': 24955, 'loss/train': 3.0134706497192383} 02/25/2022 09:54:34 - INFO - codeparrot_training - Step 24956: {'lr': 0.00026706912036620836, 'samples': 12777984, 'steps': 24956, 'loss/train': 1.155659556388855} 02/25/2022 09:54:37 - INFO - codeparrot_training - Step 24957: {'lr': 0.0002670527960507375, 'samples': 12778496, 'steps': 24957, 'loss/train': 1.9154446125030518} 02/25/2022 09:54:43 - INFO - codeparrot_training - Step 24958: {'lr': 0.00026703647166221786, 'samples': 12779008, 'steps': 24958, 'loss/train': 1.2526443004608154} 02/25/2022 09:54:46 - INFO - codeparrot_training - Step 24959: {'lr': 0.0002670201472007194, 'samples': 12779520, 'steps': 24959, 'loss/train': 2.276221513748169} 02/25/2022 09:54:52 - INFO - codeparrot_training - Step 24960: {'lr': 0.00026700382266631206, 'samples': 12780032, 'steps': 24960, 'loss/train': 2.6500132083892822} 02/25/2022 09:54:55 - INFO - codeparrot_training - Step 24961: {'lr': 0.00026698749805906567, 'samples': 12780544, 'steps': 24961, 'loss/train': 2.2446229457855225} 02/25/2022 09:55:01 - INFO - codeparrot_training - Step 24962: {'lr': 0.00026697117337905034, 'samples': 12781056, 'steps': 24962, 'loss/train': 2.034681558609009} 02/25/2022 09:55:04 - INFO - codeparrot_training - Step 24963: {'lr': 0.00026695484862633583, 'samples': 12781568, 'steps': 24963, 'loss/train': 1.2532463073730469} 02/25/2022 09:55:11 - INFO - codeparrot_training - Step 24964: {'lr': 0.00026693852380099215, 'samples': 12782080, 'steps': 24964, 'loss/train': 1.6912041902542114} 02/25/2022 09:55:15 - INFO - codeparrot_training - Step 24965: {'lr': 0.0002669221989030892, 'samples': 12782592, 'steps': 24965, 'loss/train': 1.4864130020141602} 02/25/2022 09:55:20 - INFO - codeparrot_training - Step 24966: {'lr': 0.00026690587393269694, 'samples': 12783104, 'steps': 24966, 'loss/train': 2.3117659091949463} 02/25/2022 09:55:24 - INFO - codeparrot_training - Step 24967: {'lr': 0.0002668895488898853, 'samples': 12783616, 'steps': 24967, 'loss/train': 1.5027096271514893} 02/25/2022 09:55:29 - INFO - codeparrot_training - Step 24968: {'lr': 0.00026687322377472416, 'samples': 12784128, 'steps': 24968, 'loss/train': 1.3746724128723145} 02/25/2022 09:55:33 - INFO - codeparrot_training - Step 24969: {'lr': 0.00026685689858728346, 'samples': 12784640, 'steps': 24969, 'loss/train': 2.1076157093048096} 02/25/2022 09:55:38 - INFO - codeparrot_training - Step 24970: {'lr': 0.0002668405733276332, 'samples': 12785152, 'steps': 24970, 'loss/train': 2.6847290992736816} 02/25/2022 09:55:42 - INFO - codeparrot_training - Step 24971: {'lr': 0.00026682424799584324, 'samples': 12785664, 'steps': 24971, 'loss/train': 1.9541301727294922} 02/25/2022 09:55:48 - INFO - codeparrot_training - Step 24972: {'lr': 0.00026680792259198353, 'samples': 12786176, 'steps': 24972, 'loss/train': 8.260965347290039} 02/25/2022 09:55:51 - INFO - codeparrot_training - Step 24973: {'lr': 0.0002667915971161241, 'samples': 12786688, 'steps': 24973, 'loss/train': 0.8277385830879211} 02/25/2022 09:55:57 - INFO - codeparrot_training - Step 24974: {'lr': 0.00026677527156833473, 'samples': 12787200, 'steps': 24974, 'loss/train': 2.2886950969696045} 02/25/2022 09:56:00 - INFO - codeparrot_training - Step 24975: {'lr': 0.0002667589459486855, 'samples': 12787712, 'steps': 24975, 'loss/train': 2.0280275344848633} 02/25/2022 09:56:06 - INFO - codeparrot_training - Step 24976: {'lr': 0.00026674262025724627, 'samples': 12788224, 'steps': 24976, 'loss/train': 2.3432979583740234} 02/25/2022 09:56:09 - INFO - codeparrot_training - Step 24977: {'lr': 0.00026672629449408684, 'samples': 12788736, 'steps': 24977, 'loss/train': 2.0472216606140137} 02/25/2022 09:56:15 - INFO - codeparrot_training - Step 24978: {'lr': 0.0002667099686592774, 'samples': 12789248, 'steps': 24978, 'loss/train': 2.7924671173095703} 02/25/2022 09:56:18 - INFO - codeparrot_training - Step 24979: {'lr': 0.00026669364275288773, 'samples': 12789760, 'steps': 24979, 'loss/train': 0.6600639224052429} 02/25/2022 09:56:25 - INFO - codeparrot_training - Step 24980: {'lr': 0.0002666773167749878, 'samples': 12790272, 'steps': 24980, 'loss/train': 1.716858148574829} 02/25/2022 09:56:29 - INFO - codeparrot_training - Step 24981: {'lr': 0.00026666099072564746, 'samples': 12790784, 'steps': 24981, 'loss/train': 2.035151958465576} 02/25/2022 09:56:34 - INFO - codeparrot_training - Step 24982: {'lr': 0.00026664466460493686, 'samples': 12791296, 'steps': 24982, 'loss/train': 1.325531005859375} 02/25/2022 09:56:38 - INFO - codeparrot_training - Step 24983: {'lr': 0.0002666283384129257, 'samples': 12791808, 'steps': 24983, 'loss/train': 1.8949499130249023} 02/25/2022 09:56:43 - INFO - codeparrot_training - Step 24984: {'lr': 0.00026661201214968404, 'samples': 12792320, 'steps': 24984, 'loss/train': 1.9902961254119873} 02/25/2022 09:56:47 - INFO - codeparrot_training - Step 24985: {'lr': 0.0002665956858152818, 'samples': 12792832, 'steps': 24985, 'loss/train': 0.4217319190502167} 02/25/2022 09:56:52 - INFO - codeparrot_training - Step 24986: {'lr': 0.00026657935940978896, 'samples': 12793344, 'steps': 24986, 'loss/train': 3.0633857250213623} 02/25/2022 09:56:56 - INFO - codeparrot_training - Step 24987: {'lr': 0.00026656303293327534, 'samples': 12793856, 'steps': 24987, 'loss/train': 3.4580883979797363} 02/25/2022 09:57:01 - INFO - codeparrot_training - Step 24988: {'lr': 0.00026654670638581095, 'samples': 12794368, 'steps': 24988, 'loss/train': 1.8375736474990845} 02/25/2022 09:57:05 - INFO - codeparrot_training - Step 24989: {'lr': 0.00026653037976746575, 'samples': 12794880, 'steps': 24989, 'loss/train': 2.4298338890075684} 02/25/2022 09:57:12 - INFO - codeparrot_training - Step 24990: {'lr': 0.0002665140530783097, 'samples': 12795392, 'steps': 24990, 'loss/train': 1.2055655717849731} 02/25/2022 09:57:15 - INFO - codeparrot_training - Step 24991: {'lr': 0.00026649772631841257, 'samples': 12795904, 'steps': 24991, 'loss/train': 2.2322161197662354} 02/25/2022 09:57:21 - INFO - codeparrot_training - Step 24992: {'lr': 0.0002664813994878445, 'samples': 12796416, 'steps': 24992, 'loss/train': 1.461451768875122} 02/25/2022 09:57:25 - INFO - codeparrot_training - Step 24993: {'lr': 0.0002664650725866753, 'samples': 12796928, 'steps': 24993, 'loss/train': 1.0411195755004883} 02/25/2022 09:57:30 - INFO - codeparrot_training - Step 24994: {'lr': 0.00026644874561497506, 'samples': 12797440, 'steps': 24994, 'loss/train': 2.5259485244750977} 02/25/2022 09:57:34 - INFO - codeparrot_training - Step 24995: {'lr': 0.0002664324185728135, 'samples': 12797952, 'steps': 24995, 'loss/train': 2.2317144870758057} 02/25/2022 09:57:39 - INFO - codeparrot_training - Step 24996: {'lr': 0.0002664160914602607, 'samples': 12798464, 'steps': 24996, 'loss/train': 1.6283648014068604} 02/25/2022 09:57:43 - INFO - codeparrot_training - Step 24997: {'lr': 0.0002663997642773866, 'samples': 12798976, 'steps': 24997, 'loss/train': 1.6553277969360352} 02/25/2022 09:57:48 - INFO - codeparrot_training - Step 24998: {'lr': 0.0002663834370242611, 'samples': 12799488, 'steps': 24998, 'loss/train': 2.7944555282592773} 02/25/2022 09:57:52 - INFO - codeparrot_training - Step 24999: {'lr': 0.00026636710970095426, 'samples': 12800000, 'steps': 24999, 'loss/train': 0.8392263650894165} 02/25/2022 09:57:52 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 09:58:09 - WARNING - huggingface_hub.repository - Several commits (25) will be pushed upstream. 02/25/2022 09:58:09 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 09:58:43 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 7a862ae..5549929 floral-grass-11 -> floral-grass-11 02/25/2022 09:58:50 - INFO - codeparrot_training - Step 25000: {'lr': 0.0002663507823075358, 'samples': 12800512, 'steps': 25000, 'loss/train': 1.160456895828247} 02/25/2022 09:58:53 - INFO - codeparrot_training - Step 25001: {'lr': 0.00026633445484407583, 'samples': 12801024, 'steps': 25001, 'loss/train': 2.628060817718506} 02/25/2022 09:59:01 - INFO - codeparrot_training - Step 25002: {'lr': 0.0002663181273106442, 'samples': 12801536, 'steps': 25002, 'loss/train': 1.6362329721450806} 02/25/2022 09:59:04 - INFO - codeparrot_training - Step 25003: {'lr': 0.00026630179970731094, 'samples': 12802048, 'steps': 25003, 'loss/train': 1.5520554780960083} 02/25/2022 09:59:10 - INFO - codeparrot_training - Step 25004: {'lr': 0.0002662854720341459, 'samples': 12802560, 'steps': 25004, 'loss/train': 0.42571038007736206} 02/25/2022 09:59:13 - INFO - codeparrot_training - Step 25005: {'lr': 0.00026626914429121913, 'samples': 12803072, 'steps': 25005, 'loss/train': 1.7985339164733887} 02/25/2022 09:59:19 - INFO - codeparrot_training - Step 25006: {'lr': 0.00026625281647860047, 'samples': 12803584, 'steps': 25006, 'loss/train': 2.1499977111816406} 02/25/2022 09:59:22 - INFO - codeparrot_training - Step 25007: {'lr': 0.0002662364885963599, 'samples': 12804096, 'steps': 25007, 'loss/train': 1.6239352226257324} 02/25/2022 09:59:28 - INFO - codeparrot_training - Step 25008: {'lr': 0.0002662201606445674, 'samples': 12804608, 'steps': 25008, 'loss/train': 1.7747437953948975} 02/25/2022 09:59:31 - INFO - codeparrot_training - Step 25009: {'lr': 0.00026620383262329286, 'samples': 12805120, 'steps': 25009, 'loss/train': 1.72773277759552} 02/25/2022 09:59:37 - INFO - codeparrot_training - Step 25010: {'lr': 0.0002661875045326062, 'samples': 12805632, 'steps': 25010, 'loss/train': 2.1759560108184814} 02/25/2022 09:59:40 - INFO - codeparrot_training - Step 25011: {'lr': 0.0002661711763725775, 'samples': 12806144, 'steps': 25011, 'loss/train': 1.2365885972976685} 02/25/2022 09:59:48 - INFO - codeparrot_training - Step 25012: {'lr': 0.0002661548481432766, 'samples': 12806656, 'steps': 25012, 'loss/train': 2.0251924991607666} 02/25/2022 09:59:51 - INFO - codeparrot_training - Step 25013: {'lr': 0.00026613851984477337, 'samples': 12807168, 'steps': 25013, 'loss/train': 1.984740972518921} 02/25/2022 09:59:57 - INFO - codeparrot_training - Step 25014: {'lr': 0.00026612219147713795, 'samples': 12807680, 'steps': 25014, 'loss/train': 2.4250524044036865} 02/25/2022 10:00:01 - INFO - codeparrot_training - Step 25015: {'lr': 0.00026610586304044015, 'samples': 12808192, 'steps': 25015, 'loss/train': 0.9071022272109985} 02/25/2022 10:00:06 - INFO - codeparrot_training - Step 25016: {'lr': 0.0002660895345347499, 'samples': 12808704, 'steps': 25016, 'loss/train': 1.991665005683899} 02/25/2022 10:00:10 - INFO - codeparrot_training - Step 25017: {'lr': 0.0002660732059601372, 'samples': 12809216, 'steps': 25017, 'loss/train': 2.069371461868286} 02/25/2022 10:00:15 - INFO - codeparrot_training - Step 25018: {'lr': 0.0002660568773166721, 'samples': 12809728, 'steps': 25018, 'loss/train': 2.193485975265503} 02/25/2022 10:00:19 - INFO - codeparrot_training - Step 25019: {'lr': 0.00026604054860442436, 'samples': 12810240, 'steps': 25019, 'loss/train': 2.2914202213287354} 02/25/2022 10:00:24 - INFO - codeparrot_training - Step 25020: {'lr': 0.000266024219823464, 'samples': 12810752, 'steps': 25020, 'loss/train': 1.6088972091674805} 02/25/2022 10:00:28 - INFO - codeparrot_training - Step 25021: {'lr': 0.00026600789097386095, 'samples': 12811264, 'steps': 25021, 'loss/train': 0.6779914498329163} 02/25/2022 10:00:35 - INFO - codeparrot_training - Step 25022: {'lr': 0.00026599156205568523, 'samples': 12811776, 'steps': 25022, 'loss/train': 1.6544021368026733} 02/25/2022 10:00:38 - INFO - codeparrot_training - Step 25023: {'lr': 0.00026597523306900675, 'samples': 12812288, 'steps': 25023, 'loss/train': 2.0362980365753174} 02/25/2022 10:00:44 - INFO - codeparrot_training - Step 25024: {'lr': 0.0002659589040138954, 'samples': 12812800, 'steps': 25024, 'loss/train': 1.6817017793655396} 02/25/2022 10:00:47 - INFO - codeparrot_training - Step 25025: {'lr': 0.00026594257489042115, 'samples': 12813312, 'steps': 25025, 'loss/train': 1.5721395015716553} 02/25/2022 10:00:53 - INFO - codeparrot_training - Step 25026: {'lr': 0.00026592624569865396, 'samples': 12813824, 'steps': 25026, 'loss/train': 2.136380672454834} 02/25/2022 10:00:56 - INFO - codeparrot_training - Step 25027: {'lr': 0.00026590991643866393, 'samples': 12814336, 'steps': 25027, 'loss/train': 1.866845965385437} 02/25/2022 10:01:02 - INFO - codeparrot_training - Step 25028: {'lr': 0.0002658935871105207, 'samples': 12814848, 'steps': 25028, 'loss/train': 2.767653226852417} 02/25/2022 10:01:05 - INFO - codeparrot_training - Step 25029: {'lr': 0.0002658772577142945, 'samples': 12815360, 'steps': 25029, 'loss/train': 0.9264286756515503} 02/25/2022 10:01:11 - INFO - codeparrot_training - Step 25030: {'lr': 0.0002658609282500551, 'samples': 12815872, 'steps': 25030, 'loss/train': 1.4567736387252808} 02/25/2022 10:01:14 - INFO - codeparrot_training - Step 25031: {'lr': 0.0002658445987178726, 'samples': 12816384, 'steps': 25031, 'loss/train': 2.3768463134765625} 02/25/2022 10:01:20 - INFO - codeparrot_training - Step 25032: {'lr': 0.00026582826911781675, 'samples': 12816896, 'steps': 25032, 'loss/train': 2.361173152923584} 02/25/2022 10:01:23 - INFO - codeparrot_training - Step 25033: {'lr': 0.0002658119394499577, 'samples': 12817408, 'steps': 25033, 'loss/train': 2.2039384841918945} 02/25/2022 10:01:29 - INFO - codeparrot_training - Step 25034: {'lr': 0.0002657956097143653, 'samples': 12817920, 'steps': 25034, 'loss/train': 2.3813908100128174} 02/25/2022 10:01:32 - INFO - codeparrot_training - Step 25035: {'lr': 0.0002657792799111095, 'samples': 12818432, 'steps': 25035, 'loss/train': 1.2311429977416992} 02/25/2022 10:01:40 - INFO - codeparrot_training - Step 25036: {'lr': 0.00026576295004026034, 'samples': 12818944, 'steps': 25036, 'loss/train': 0.7064680457115173} 02/25/2022 10:01:43 - INFO - codeparrot_training - Step 25037: {'lr': 0.00026574662010188767, 'samples': 12819456, 'steps': 25037, 'loss/train': 2.1231205463409424} 02/25/2022 10:01:49 - INFO - codeparrot_training - Step 25038: {'lr': 0.0002657302900960615, 'samples': 12819968, 'steps': 25038, 'loss/train': 1.8891308307647705} 02/25/2022 10:01:52 - INFO - codeparrot_training - Step 25039: {'lr': 0.00026571396002285174, 'samples': 12820480, 'steps': 25039, 'loss/train': 1.2767367362976074} 02/25/2022 10:01:58 - INFO - codeparrot_training - Step 25040: {'lr': 0.0002656976298823284, 'samples': 12820992, 'steps': 25040, 'loss/train': 0.625769853591919} 02/25/2022 10:02:02 - INFO - codeparrot_training - Step 25041: {'lr': 0.00026568129967456135, 'samples': 12821504, 'steps': 25041, 'loss/train': 2.32356595993042} 02/25/2022 10:02:07 - INFO - codeparrot_training - Step 25042: {'lr': 0.0002656649693996206, 'samples': 12822016, 'steps': 25042, 'loss/train': 1.1074837446212769} 02/25/2022 10:02:11 - INFO - codeparrot_training - Step 25043: {'lr': 0.00026564863905757606, 'samples': 12822528, 'steps': 25043, 'loss/train': 2.6152665615081787} 02/25/2022 10:02:16 - INFO - codeparrot_training - Step 25044: {'lr': 0.00026563230864849784, 'samples': 12823040, 'steps': 25044, 'loss/train': 2.2992584705352783} 02/25/2022 10:02:20 - INFO - codeparrot_training - Step 25045: {'lr': 0.0002656159781724557, 'samples': 12823552, 'steps': 25045, 'loss/train': 1.691097617149353} 02/25/2022 10:02:25 - INFO - codeparrot_training - Step 25046: {'lr': 0.0002655996476295196, 'samples': 12824064, 'steps': 25046, 'loss/train': 0.944744884967804} 02/25/2022 10:02:29 - INFO - codeparrot_training - Step 25047: {'lr': 0.0002655833170197597, 'samples': 12824576, 'steps': 25047, 'loss/train': 1.5726711750030518} 02/25/2022 10:02:36 - INFO - codeparrot_training - Step 25048: {'lr': 0.0002655669863432457, 'samples': 12825088, 'steps': 25048, 'loss/train': 2.2993552684783936} 02/25/2022 10:02:40 - INFO - codeparrot_training - Step 25049: {'lr': 0.00026555065560004775, 'samples': 12825600, 'steps': 25049, 'loss/train': 1.6567405462265015} 02/25/2022 10:02:46 - INFO - codeparrot_training - Step 25050: {'lr': 0.0002655343247902357, 'samples': 12826112, 'steps': 25050, 'loss/train': 2.1947453022003174} 02/25/2022 10:02:49 - INFO - codeparrot_training - Step 25051: {'lr': 0.00026551799391387953, 'samples': 12826624, 'steps': 25051, 'loss/train': 1.6891319751739502} 02/25/2022 10:02:52 - INFO - codeparrot_training - Step 25052: {'lr': 0.0002655016629710492, 'samples': 12827136, 'steps': 25052, 'loss/train': 2.944282293319702} 02/25/2022 10:02:58 - INFO - codeparrot_training - Step 25053: {'lr': 0.0002654853319618147, 'samples': 12827648, 'steps': 25053, 'loss/train': 0.4813360571861267} 02/25/2022 10:03:02 - INFO - codeparrot_training - Step 25054: {'lr': 0.00026546900088624594, 'samples': 12828160, 'steps': 25054, 'loss/train': 3.6540863513946533} 02/25/2022 10:03:07 - INFO - codeparrot_training - Step 25055: {'lr': 0.0002654526697444129, 'samples': 12828672, 'steps': 25055, 'loss/train': 1.298710584640503} 02/25/2022 10:03:11 - INFO - codeparrot_training - Step 25056: {'lr': 0.0002654363385363855, 'samples': 12829184, 'steps': 25056, 'loss/train': 1.1219193935394287} 02/25/2022 10:03:17 - INFO - codeparrot_training - Step 25057: {'lr': 0.0002654200072622337, 'samples': 12829696, 'steps': 25057, 'loss/train': 1.8067675828933716} 02/25/2022 10:03:20 - INFO - codeparrot_training - Step 25058: {'lr': 0.00026540367592202754, 'samples': 12830208, 'steps': 25058, 'loss/train': 1.7024645805358887} 02/25/2022 10:03:24 - INFO - codeparrot_training - Step 25059: {'lr': 0.0002653873445158369, 'samples': 12830720, 'steps': 25059, 'loss/train': 2.1636903285980225} 02/25/2022 10:03:31 - INFO - codeparrot_training - Step 25060: {'lr': 0.0002653710130437318, 'samples': 12831232, 'steps': 25060, 'loss/train': 2.718723773956299} 02/25/2022 10:03:35 - INFO - codeparrot_training - Step 25061: {'lr': 0.0002653546815057821, 'samples': 12831744, 'steps': 25061, 'loss/train': 2.3811492919921875} 02/25/2022 10:03:40 - INFO - codeparrot_training - Step 25062: {'lr': 0.0002653383499020579, 'samples': 12832256, 'steps': 25062, 'loss/train': 8.567477226257324} 02/25/2022 10:03:44 - INFO - codeparrot_training - Step 25063: {'lr': 0.00026532201823262897, 'samples': 12832768, 'steps': 25063, 'loss/train': 1.0298140048980713} 02/25/2022 10:03:49 - INFO - codeparrot_training - Step 25064: {'lr': 0.00026530568649756547, 'samples': 12833280, 'steps': 25064, 'loss/train': 1.4718226194381714} 02/25/2022 10:03:53 - INFO - codeparrot_training - Step 25065: {'lr': 0.0002652893546969373, 'samples': 12833792, 'steps': 25065, 'loss/train': 2.306047201156616} 02/25/2022 10:03:58 - INFO - codeparrot_training - Step 25066: {'lr': 0.0002652730228308143, 'samples': 12834304, 'steps': 25066, 'loss/train': 2.2164902687072754} 02/25/2022 10:04:02 - INFO - codeparrot_training - Step 25067: {'lr': 0.00026525669089926657, 'samples': 12834816, 'steps': 25067, 'loss/train': 2.468912363052368} 02/25/2022 10:04:07 - INFO - codeparrot_training - Step 25068: {'lr': 0.00026524035890236397, 'samples': 12835328, 'steps': 25068, 'loss/train': 1.9229764938354492} 02/25/2022 10:04:11 - INFO - codeparrot_training - Step 25069: {'lr': 0.0002652240268401766, 'samples': 12835840, 'steps': 25069, 'loss/train': 2.1440024375915527} 02/25/2022 10:04:18 - INFO - codeparrot_training - Step 25070: {'lr': 0.00026520769471277423, 'samples': 12836352, 'steps': 25070, 'loss/train': 2.160217523574829} 02/25/2022 10:04:21 - INFO - codeparrot_training - Step 25071: {'lr': 0.000265191362520227, 'samples': 12836864, 'steps': 25071, 'loss/train': 1.2895935773849487} 02/25/2022 10:04:27 - INFO - codeparrot_training - Step 25072: {'lr': 0.00026517503026260474, 'samples': 12837376, 'steps': 25072, 'loss/train': 1.391907811164856} 02/25/2022 10:04:30 - INFO - codeparrot_training - Step 25073: {'lr': 0.00026515869793997755, 'samples': 12837888, 'steps': 25073, 'loss/train': 2.3654892444610596} 02/25/2022 10:04:36 - INFO - codeparrot_training - Step 25074: {'lr': 0.0002651423655524152, 'samples': 12838400, 'steps': 25074, 'loss/train': 1.4736377000808716} 02/25/2022 10:04:40 - INFO - codeparrot_training - Step 25075: {'lr': 0.00026512603309998783, 'samples': 12838912, 'steps': 25075, 'loss/train': 2.848262071609497} 02/25/2022 10:04:45 - INFO - codeparrot_training - Step 25076: {'lr': 0.00026510970058276533, 'samples': 12839424, 'steps': 25076, 'loss/train': 1.776846170425415} 02/25/2022 10:04:49 - INFO - codeparrot_training - Step 25077: {'lr': 0.00026509336800081774, 'samples': 12839936, 'steps': 25077, 'loss/train': 4.0198974609375} 02/25/2022 10:04:54 - INFO - codeparrot_training - Step 25078: {'lr': 0.00026507703535421485, 'samples': 12840448, 'steps': 25078, 'loss/train': 0.8756927847862244} 02/25/2022 10:04:58 - INFO - codeparrot_training - Step 25079: {'lr': 0.0002650607026430268, 'samples': 12840960, 'steps': 25079, 'loss/train': 1.9865827560424805} 02/25/2022 10:05:03 - INFO - codeparrot_training - Step 25080: {'lr': 0.0002650443698673234, 'samples': 12841472, 'steps': 25080, 'loss/train': 1.4674131870269775} 02/25/2022 10:05:07 - INFO - codeparrot_training - Step 25081: {'lr': 0.0002650280370271747, 'samples': 12841984, 'steps': 25081, 'loss/train': 1.5579302310943604} 02/25/2022 10:05:13 - INFO - codeparrot_training - Step 25082: {'lr': 0.00026501170412265076, 'samples': 12842496, 'steps': 25082, 'loss/train': 1.2404475212097168} 02/25/2022 10:05:16 - INFO - codeparrot_training - Step 25083: {'lr': 0.00026499537115382135, 'samples': 12843008, 'steps': 25083, 'loss/train': 1.9711076021194458} 02/25/2022 10:05:19 - INFO - codeparrot_training - Step 25084: {'lr': 0.0002649790381207566, 'samples': 12843520, 'steps': 25084, 'loss/train': 0.4490146338939667} 02/25/2022 10:05:27 - INFO - codeparrot_training - Step 25085: {'lr': 0.0002649627050235263, 'samples': 12844032, 'steps': 25085, 'loss/train': 2.262723922729492} 02/25/2022 10:05:30 - INFO - codeparrot_training - Step 25086: {'lr': 0.00026494637186220065, 'samples': 12844544, 'steps': 25086, 'loss/train': 1.8037751913070679} 02/25/2022 10:05:36 - INFO - codeparrot_training - Step 25087: {'lr': 0.00026493003863684944, 'samples': 12845056, 'steps': 25087, 'loss/train': 2.374884605407715} 02/25/2022 10:05:39 - INFO - codeparrot_training - Step 25088: {'lr': 0.0002649137053475427, 'samples': 12845568, 'steps': 25088, 'loss/train': 4.739823818206787} 02/25/2022 10:05:45 - INFO - codeparrot_training - Step 25089: {'lr': 0.00026489737199435036, 'samples': 12846080, 'steps': 25089, 'loss/train': 1.5920213460922241} 02/25/2022 10:05:50 - INFO - codeparrot_training - Step 25090: {'lr': 0.0002648810385773424, 'samples': 12846592, 'steps': 25090, 'loss/train': 2.736701726913452} 02/25/2022 10:05:54 - INFO - codeparrot_training - Step 25091: {'lr': 0.0002648647050965888, 'samples': 12847104, 'steps': 25091, 'loss/train': 1.961225986480713} 02/25/2022 10:05:59 - INFO - codeparrot_training - Step 25092: {'lr': 0.00026484837155215957, 'samples': 12847616, 'steps': 25092, 'loss/train': 1.0222641229629517} 02/25/2022 10:06:03 - INFO - codeparrot_training - Step 25093: {'lr': 0.00026483203794412457, 'samples': 12848128, 'steps': 25093, 'loss/train': 1.9658944606781006} 02/25/2022 10:06:08 - INFO - codeparrot_training - Step 25094: {'lr': 0.00026481570427255385, 'samples': 12848640, 'steps': 25094, 'loss/train': 1.1277885437011719} 02/25/2022 10:06:12 - INFO - codeparrot_training - Step 25095: {'lr': 0.0002647993705375174, 'samples': 12849152, 'steps': 25095, 'loss/train': 2.2916805744171143} 02/25/2022 10:06:19 - INFO - codeparrot_training - Step 25096: {'lr': 0.00026478303673908507, 'samples': 12849664, 'steps': 25096, 'loss/train': 1.7349023818969727} 02/25/2022 10:06:23 - INFO - codeparrot_training - Step 25097: {'lr': 0.00026476670287732697, 'samples': 12850176, 'steps': 25097, 'loss/train': 2.313267707824707} 02/25/2022 10:06:28 - INFO - codeparrot_training - Step 25098: {'lr': 0.000264750368952313, 'samples': 12850688, 'steps': 25098, 'loss/train': 2.211817741394043} 02/25/2022 10:06:32 - INFO - codeparrot_training - Step 25099: {'lr': 0.00026473403496411307, 'samples': 12851200, 'steps': 25099, 'loss/train': 1.7130725383758545} 02/25/2022 10:06:37 - INFO - codeparrot_training - Step 25100: {'lr': 0.00026471770091279724, 'samples': 12851712, 'steps': 25100, 'loss/train': 2.3953895568847656} 02/25/2022 10:06:41 - INFO - codeparrot_training - Step 25101: {'lr': 0.0002647013667984355, 'samples': 12852224, 'steps': 25101, 'loss/train': 2.012718915939331} 02/25/2022 10:06:46 - INFO - codeparrot_training - Step 25102: {'lr': 0.00026468503262109774, 'samples': 12852736, 'steps': 25102, 'loss/train': 1.41632080078125} 02/25/2022 10:06:50 - INFO - codeparrot_training - Step 25103: {'lr': 0.00026466869838085393, 'samples': 12853248, 'steps': 25103, 'loss/train': 2.2961106300354004} 02/25/2022 10:06:55 - INFO - codeparrot_training - Step 25104: {'lr': 0.0002646523640777741, 'samples': 12853760, 'steps': 25104, 'loss/train': 1.8013173341751099} 02/25/2022 10:06:58 - INFO - codeparrot_training - Step 25105: {'lr': 0.00026463602971192824, 'samples': 12854272, 'steps': 25105, 'loss/train': 1.9846426248550415} 02/25/2022 10:07:06 - INFO - codeparrot_training - Step 25106: {'lr': 0.0002646196952833862, 'samples': 12854784, 'steps': 25106, 'loss/train': 2.1611316204071045} 02/25/2022 10:07:09 - INFO - codeparrot_training - Step 25107: {'lr': 0.0002646033607922181, 'samples': 12855296, 'steps': 25107, 'loss/train': 1.252687931060791} 02/25/2022 10:07:15 - INFO - codeparrot_training - Step 25108: {'lr': 0.0002645870262384938, 'samples': 12855808, 'steps': 25108, 'loss/train': 1.8214967250823975} 02/25/2022 10:07:18 - INFO - codeparrot_training - Step 25109: {'lr': 0.00026457069162228337, 'samples': 12856320, 'steps': 25109, 'loss/train': 1.959747076034546} 02/25/2022 10:07:24 - INFO - codeparrot_training - Step 25110: {'lr': 0.0002645543569436566, 'samples': 12856832, 'steps': 25110, 'loss/train': 2.5783262252807617} 02/25/2022 10:07:27 - INFO - codeparrot_training - Step 25111: {'lr': 0.0002645380222026837, 'samples': 12857344, 'steps': 25111, 'loss/train': 2.056340217590332} 02/25/2022 10:07:33 - INFO - codeparrot_training - Step 25112: {'lr': 0.0002645216873994345, 'samples': 12857856, 'steps': 25112, 'loss/train': 0.7776860594749451} 02/25/2022 10:07:36 - INFO - codeparrot_training - Step 25113: {'lr': 0.00026450535253397895, 'samples': 12858368, 'steps': 25113, 'loss/train': 1.2803047895431519} 02/25/2022 10:07:42 - INFO - codeparrot_training - Step 25114: {'lr': 0.0002644890176063872, 'samples': 12858880, 'steps': 25114, 'loss/train': 0.702643096446991} 02/25/2022 10:07:45 - INFO - codeparrot_training - Step 25115: {'lr': 0.00026447268261672895, 'samples': 12859392, 'steps': 25115, 'loss/train': 2.390354633331299} 02/25/2022 10:07:51 - INFO - codeparrot_training - Step 25116: {'lr': 0.0002644563475650744, 'samples': 12859904, 'steps': 25116, 'loss/train': 0.7148985266685486} 02/25/2022 10:07:54 - INFO - codeparrot_training - Step 25117: {'lr': 0.0002644400124514934, 'samples': 12860416, 'steps': 25117, 'loss/train': 1.1707137823104858} 02/25/2022 10:08:00 - INFO - codeparrot_training - Step 25118: {'lr': 0.000264423677276056, 'samples': 12860928, 'steps': 25118, 'loss/train': 1.7243993282318115} 02/25/2022 10:08:03 - INFO - codeparrot_training - Step 25119: {'lr': 0.0002644073420388322, 'samples': 12861440, 'steps': 25119, 'loss/train': 1.88140070438385} 02/25/2022 10:08:09 - INFO - codeparrot_training - Step 25120: {'lr': 0.00026439100673989184, 'samples': 12861952, 'steps': 25120, 'loss/train': 0.8988780975341797} 02/25/2022 10:08:12 - INFO - codeparrot_training - Step 25121: {'lr': 0.00026437467137930507, 'samples': 12862464, 'steps': 25121, 'loss/train': 1.7940598726272583} 02/25/2022 10:08:20 - INFO - codeparrot_training - Step 25122: {'lr': 0.00026435833595714163, 'samples': 12862976, 'steps': 25122, 'loss/train': 1.2678987979888916} 02/25/2022 10:08:23 - INFO - codeparrot_training - Step 25123: {'lr': 0.0002643420004734718, 'samples': 12863488, 'steps': 25123, 'loss/train': 2.009763479232788} 02/25/2022 10:08:29 - INFO - codeparrot_training - Step 25124: {'lr': 0.00026432566492836523, 'samples': 12864000, 'steps': 25124, 'loss/train': 1.8347705602645874} 02/25/2022 10:08:32 - INFO - codeparrot_training - Step 25125: {'lr': 0.00026430932932189225, 'samples': 12864512, 'steps': 25125, 'loss/train': 1.8948403596878052} 02/25/2022 10:08:38 - INFO - codeparrot_training - Step 25126: {'lr': 0.0002642929936541225, 'samples': 12865024, 'steps': 25126, 'loss/train': 2.1912829875946045} 02/25/2022 10:08:41 - INFO - codeparrot_training - Step 25127: {'lr': 0.00026427665792512614, 'samples': 12865536, 'steps': 25127, 'loss/train': 2.6858644485473633} 02/25/2022 10:08:47 - INFO - codeparrot_training - Step 25128: {'lr': 0.0002642603221349731, 'samples': 12866048, 'steps': 25128, 'loss/train': 1.4050952196121216} 02/25/2022 10:08:50 - INFO - codeparrot_training - Step 25129: {'lr': 0.0002642439862837334, 'samples': 12866560, 'steps': 25129, 'loss/train': 1.9154839515686035} 02/25/2022 10:08:56 - INFO - codeparrot_training - Step 25130: {'lr': 0.00026422765037147696, 'samples': 12867072, 'steps': 25130, 'loss/train': 2.706439971923828} 02/25/2022 10:09:00 - INFO - codeparrot_training - Step 25131: {'lr': 0.0002642113143982738, 'samples': 12867584, 'steps': 25131, 'loss/train': 1.623214840888977} 02/25/2022 10:09:07 - INFO - codeparrot_training - Step 25132: {'lr': 0.0002641949783641939, 'samples': 12868096, 'steps': 25132, 'loss/train': 2.114255428314209} 02/25/2022 10:09:10 - INFO - codeparrot_training - Step 25133: {'lr': 0.00026417864226930713, 'samples': 12868608, 'steps': 25133, 'loss/train': 2.5305700302124023} 02/25/2022 10:09:16 - INFO - codeparrot_training - Step 25134: {'lr': 0.00026416230611368363, 'samples': 12869120, 'steps': 25134, 'loss/train': 1.8685394525527954} 02/25/2022 10:09:19 - INFO - codeparrot_training - Step 25135: {'lr': 0.00026414596989739333, 'samples': 12869632, 'steps': 25135, 'loss/train': 2.091590166091919} 02/25/2022 10:09:25 - INFO - codeparrot_training - Step 25136: {'lr': 0.0002641296336205062, 'samples': 12870144, 'steps': 25136, 'loss/train': 1.804237723350525} 02/25/2022 10:09:28 - INFO - codeparrot_training - Step 25137: {'lr': 0.0002641132972830921, 'samples': 12870656, 'steps': 25137, 'loss/train': 1.5900814533233643} 02/25/2022 10:09:34 - INFO - codeparrot_training - Step 25138: {'lr': 0.00026409696088522123, 'samples': 12871168, 'steps': 25138, 'loss/train': 1.7084108591079712} 02/25/2022 10:09:37 - INFO - codeparrot_training - Step 25139: {'lr': 0.00026408062442696335, 'samples': 12871680, 'steps': 25139, 'loss/train': 1.98611319065094} 02/25/2022 10:09:43 - INFO - codeparrot_training - Step 25140: {'lr': 0.00026406428790838864, 'samples': 12872192, 'steps': 25140, 'loss/train': 2.0422236919403076} 02/25/2022 10:09:46 - INFO - codeparrot_training - Step 25141: {'lr': 0.00026404795132956694, 'samples': 12872704, 'steps': 25141, 'loss/train': 1.176757574081421} 02/25/2022 10:09:53 - INFO - codeparrot_training - Step 25142: {'lr': 0.00026403161469056827, 'samples': 12873216, 'steps': 25142, 'loss/train': 0.999722957611084} 02/25/2022 10:09:57 - INFO - codeparrot_training - Step 25143: {'lr': 0.0002640152779914626, 'samples': 12873728, 'steps': 25143, 'loss/train': 2.1472854614257812} 02/25/2022 10:10:02 - INFO - codeparrot_training - Step 25144: {'lr': 0.00026399894123232, 'samples': 12874240, 'steps': 25144, 'loss/train': 1.7768681049346924} 02/25/2022 10:10:06 - INFO - codeparrot_training - Step 25145: {'lr': 0.00026398260441321035, 'samples': 12874752, 'steps': 25145, 'loss/train': 2.6299691200256348} 02/25/2022 10:10:11 - INFO - codeparrot_training - Step 25146: {'lr': 0.0002639662675342036, 'samples': 12875264, 'steps': 25146, 'loss/train': 1.055228590965271} 02/25/2022 10:10:15 - INFO - codeparrot_training - Step 25147: {'lr': 0.00026394993059536995, 'samples': 12875776, 'steps': 25147, 'loss/train': 6.421290397644043} 02/25/2022 10:10:20 - INFO - codeparrot_training - Step 25148: {'lr': 0.00026393359359677904, 'samples': 12876288, 'steps': 25148, 'loss/train': 1.5929527282714844} 02/25/2022 10:10:24 - INFO - codeparrot_training - Step 25149: {'lr': 0.0002639172565385012, 'samples': 12876800, 'steps': 25149, 'loss/train': 1.9797353744506836} 02/25/2022 10:10:29 - INFO - codeparrot_training - Step 25150: {'lr': 0.00026390091942060613, 'samples': 12877312, 'steps': 25150, 'loss/train': 0.9531596302986145} 02/25/2022 10:10:33 - INFO - codeparrot_training - Step 25151: {'lr': 0.00026388458224316404, 'samples': 12877824, 'steps': 25151, 'loss/train': 1.5998244285583496} 02/25/2022 10:10:40 - INFO - codeparrot_training - Step 25152: {'lr': 0.0002638682450062448, 'samples': 12878336, 'steps': 25152, 'loss/train': 2.1040232181549072} 02/25/2022 10:10:44 - INFO - codeparrot_training - Step 25153: {'lr': 0.0002638519077099183, 'samples': 12878848, 'steps': 25153, 'loss/train': 2.62042498588562} 02/25/2022 10:10:49 - INFO - codeparrot_training - Step 25154: {'lr': 0.0002638355703542547, 'samples': 12879360, 'steps': 25154, 'loss/train': 3.3423335552215576} 02/25/2022 10:10:53 - INFO - codeparrot_training - Step 25155: {'lr': 0.00026381923293932385, 'samples': 12879872, 'steps': 25155, 'loss/train': 1.4967269897460938} 02/25/2022 10:10:58 - INFO - codeparrot_training - Step 25156: {'lr': 0.0002638028954651959, 'samples': 12880384, 'steps': 25156, 'loss/train': 0.7448533177375793} 02/25/2022 10:11:02 - INFO - codeparrot_training - Step 25157: {'lr': 0.0002637865579319406, 'samples': 12880896, 'steps': 25157, 'loss/train': 2.8340001106262207} 02/25/2022 10:11:07 - INFO - codeparrot_training - Step 25158: {'lr': 0.0002637702203396282, 'samples': 12881408, 'steps': 25158, 'loss/train': 1.175994873046875} 02/25/2022 10:11:11 - INFO - codeparrot_training - Step 25159: {'lr': 0.00026375388268832844, 'samples': 12881920, 'steps': 25159, 'loss/train': 1.030189871788025} 02/25/2022 10:11:16 - INFO - codeparrot_training - Step 25160: {'lr': 0.00026373754497811147, 'samples': 12882432, 'steps': 25160, 'loss/train': 2.870720148086548} 02/25/2022 10:11:19 - INFO - codeparrot_training - Step 25161: {'lr': 0.0002637212072090472, 'samples': 12882944, 'steps': 25161, 'loss/train': 2.220973014831543} 02/25/2022 10:11:25 - INFO - codeparrot_training - Step 25162: {'lr': 0.0002637048693812056, 'samples': 12883456, 'steps': 25162, 'loss/train': 1.8070130348205566} 02/25/2022 10:11:29 - INFO - codeparrot_training - Step 25163: {'lr': 0.00026368853149465676, 'samples': 12883968, 'steps': 25163, 'loss/train': 0.9473209977149963} 02/25/2022 10:11:35 - INFO - codeparrot_training - Step 25164: {'lr': 0.00026367219354947056, 'samples': 12884480, 'steps': 25164, 'loss/train': 2.1047441959381104} 02/25/2022 10:11:38 - INFO - codeparrot_training - Step 25165: {'lr': 0.000263655855545717, 'samples': 12884992, 'steps': 25165, 'loss/train': 0.9830390810966492} 02/25/2022 10:11:44 - INFO - codeparrot_training - Step 25166: {'lr': 0.00026363951748346615, 'samples': 12885504, 'steps': 25166, 'loss/train': 0.845406711101532} 02/25/2022 10:11:47 - INFO - codeparrot_training - Step 25167: {'lr': 0.0002636231793627879, 'samples': 12886016, 'steps': 25167, 'loss/train': 1.1542160511016846} 02/25/2022 10:11:55 - INFO - codeparrot_training - Step 25168: {'lr': 0.00026360684118375226, 'samples': 12886528, 'steps': 25168, 'loss/train': 1.5078092813491821} 02/25/2022 10:11:58 - INFO - codeparrot_training - Step 25169: {'lr': 0.0002635905029464293, 'samples': 12887040, 'steps': 25169, 'loss/train': 2.3469903469085693} 02/25/2022 10:12:04 - INFO - codeparrot_training - Step 25170: {'lr': 0.0002635741646508889, 'samples': 12887552, 'steps': 25170, 'loss/train': 1.9248242378234863} 02/25/2022 10:12:07 - INFO - codeparrot_training - Step 25171: {'lr': 0.0002635578262972011, 'samples': 12888064, 'steps': 25171, 'loss/train': 2.318403959274292} 02/25/2022 10:12:13 - INFO - codeparrot_training - Step 25172: {'lr': 0.0002635414878854359, 'samples': 12888576, 'steps': 25172, 'loss/train': 2.0437896251678467} 02/25/2022 10:12:16 - INFO - codeparrot_training - Step 25173: {'lr': 0.0002635251494156632, 'samples': 12889088, 'steps': 25173, 'loss/train': 1.870969533920288} 02/25/2022 10:12:22 - INFO - codeparrot_training - Step 25174: {'lr': 0.00026350881088795306, 'samples': 12889600, 'steps': 25174, 'loss/train': 1.3610689640045166} 02/25/2022 10:12:25 - INFO - codeparrot_training - Step 25175: {'lr': 0.0002634924723023755, 'samples': 12890112, 'steps': 25175, 'loss/train': 0.3429269790649414} 02/25/2022 10:12:29 - INFO - codeparrot_training - Step 25176: {'lr': 0.0002634761336590005, 'samples': 12890624, 'steps': 25176, 'loss/train': 0.12544433772563934} 02/25/2022 10:12:34 - INFO - codeparrot_training - Step 25177: {'lr': 0.00026345979495789797, 'samples': 12891136, 'steps': 25177, 'loss/train': 1.1430953741073608} 02/25/2022 10:12:41 - INFO - codeparrot_training - Step 25178: {'lr': 0.00026344345619913796, 'samples': 12891648, 'steps': 25178, 'loss/train': 1.7195504903793335} 02/25/2022 10:12:45 - INFO - codeparrot_training - Step 25179: {'lr': 0.0002634271173827904, 'samples': 12892160, 'steps': 25179, 'loss/train': 2.5008950233459473} 02/25/2022 10:12:50 - INFO - codeparrot_training - Step 25180: {'lr': 0.0002634107785089254, 'samples': 12892672, 'steps': 25180, 'loss/train': 1.8234107494354248} 02/25/2022 10:12:54 - INFO - codeparrot_training - Step 25181: {'lr': 0.0002633944395776129, 'samples': 12893184, 'steps': 25181, 'loss/train': 1.1687787771224976} 02/25/2022 10:12:59 - INFO - codeparrot_training - Step 25182: {'lr': 0.00026337810058892286, 'samples': 12893696, 'steps': 25182, 'loss/train': 1.5779426097869873} 02/25/2022 10:13:03 - INFO - codeparrot_training - Step 25183: {'lr': 0.0002633617615429252, 'samples': 12894208, 'steps': 25183, 'loss/train': 0.3335775136947632} 02/25/2022 10:13:08 - INFO - codeparrot_training - Step 25184: {'lr': 0.0002633454224396901, 'samples': 12894720, 'steps': 25184, 'loss/train': 1.4372093677520752} 02/25/2022 10:13:12 - INFO - codeparrot_training - Step 25185: {'lr': 0.00026332908327928746, 'samples': 12895232, 'steps': 25185, 'loss/train': 1.827237606048584} 02/25/2022 10:13:17 - INFO - codeparrot_training - Step 25186: {'lr': 0.00026331274406178726, 'samples': 12895744, 'steps': 25186, 'loss/train': 0.3919581174850464} 02/25/2022 10:13:21 - INFO - codeparrot_training - Step 25187: {'lr': 0.00026329640478725937, 'samples': 12896256, 'steps': 25187, 'loss/train': 2.2953128814697266} 02/25/2022 10:13:28 - INFO - codeparrot_training - Step 25188: {'lr': 0.00026328006545577406, 'samples': 12896768, 'steps': 25188, 'loss/train': 2.5481157302856445} 02/25/2022 10:13:32 - INFO - codeparrot_training - Step 25189: {'lr': 0.000263263726067401, 'samples': 12897280, 'steps': 25189, 'loss/train': 1.0703096389770508} 02/25/2022 10:13:37 - INFO - codeparrot_training - Step 25190: {'lr': 0.0002632473866222104, 'samples': 12897792, 'steps': 25190, 'loss/train': 2.317592144012451} 02/25/2022 10:13:41 - INFO - codeparrot_training - Step 25191: {'lr': 0.0002632310471202723, 'samples': 12898304, 'steps': 25191, 'loss/train': 2.086141586303711} 02/25/2022 10:13:44 - INFO - codeparrot_training - Step 25192: {'lr': 0.0002632147075616566, 'samples': 12898816, 'steps': 25192, 'loss/train': 1.1146897077560425} 02/25/2022 10:13:50 - INFO - codeparrot_training - Step 25193: {'lr': 0.0002631983679464332, 'samples': 12899328, 'steps': 25193, 'loss/train': 1.962989330291748} 02/25/2022 10:13:53 - INFO - codeparrot_training - Step 25194: {'lr': 0.0002631820282746722, 'samples': 12899840, 'steps': 25194, 'loss/train': 2.2453997135162354} 02/25/2022 10:13:59 - INFO - codeparrot_training - Step 25195: {'lr': 0.0002631656885464436, 'samples': 12900352, 'steps': 25195, 'loss/train': 0.8386679291725159} 02/25/2022 10:14:03 - INFO - codeparrot_training - Step 25196: {'lr': 0.00026314934876181734, 'samples': 12900864, 'steps': 25196, 'loss/train': 2.2704102993011475} 02/25/2022 10:14:08 - INFO - codeparrot_training - Step 25197: {'lr': 0.0002631330089208634, 'samples': 12901376, 'steps': 25197, 'loss/train': 2.329030990600586} 02/25/2022 10:14:12 - INFO - codeparrot_training - Step 25198: {'lr': 0.00026311666902365194, 'samples': 12901888, 'steps': 25198, 'loss/train': 1.9652130603790283} 02/25/2022 10:14:17 - INFO - codeparrot_training - Step 25199: {'lr': 0.00026310032907025276, 'samples': 12902400, 'steps': 25199, 'loss/train': 2.265171766281128} 02/25/2022 10:14:21 - INFO - codeparrot_training - Step 25200: {'lr': 0.000263083989060736, 'samples': 12902912, 'steps': 25200, 'loss/train': 2.438291072845459} 02/25/2022 10:14:28 - INFO - codeparrot_training - Step 25201: {'lr': 0.00026306764899517154, 'samples': 12903424, 'steps': 25201, 'loss/train': 1.3084348440170288} 02/25/2022 10:14:31 - INFO - codeparrot_training - Step 25202: {'lr': 0.0002630513088736294, 'samples': 12903936, 'steps': 25202, 'loss/train': 2.0516445636749268} 02/25/2022 10:14:37 - INFO - codeparrot_training - Step 25203: {'lr': 0.0002630349686961796, 'samples': 12904448, 'steps': 25203, 'loss/train': 1.963873267173767} 02/25/2022 10:14:40 - INFO - codeparrot_training - Step 25204: {'lr': 0.00026301862846289214, 'samples': 12904960, 'steps': 25204, 'loss/train': 2.644075393676758} 02/25/2022 10:14:46 - INFO - codeparrot_training - Step 25205: {'lr': 0.00026300228817383703, 'samples': 12905472, 'steps': 25205, 'loss/train': 2.1028757095336914} 02/25/2022 10:14:50 - INFO - codeparrot_training - Step 25206: {'lr': 0.00026298594782908424, 'samples': 12905984, 'steps': 25206, 'loss/train': 3.740410327911377} 02/25/2022 10:14:55 - INFO - codeparrot_training - Step 25207: {'lr': 0.00026296960742870374, 'samples': 12906496, 'steps': 25207, 'loss/train': 1.3810604810714722} 02/25/2022 10:14:59 - INFO - codeparrot_training - Step 25208: {'lr': 0.00026295326697276563, 'samples': 12907008, 'steps': 25208, 'loss/train': 1.3692246675491333} 02/25/2022 10:15:04 - INFO - codeparrot_training - Step 25209: {'lr': 0.00026293692646133974, 'samples': 12907520, 'steps': 25209, 'loss/train': 1.5162711143493652} 02/25/2022 10:15:08 - INFO - codeparrot_training - Step 25210: {'lr': 0.00026292058589449626, 'samples': 12908032, 'steps': 25210, 'loss/train': 2.312471866607666} 02/25/2022 10:15:15 - INFO - codeparrot_training - Step 25211: {'lr': 0.000262904245272305, 'samples': 12908544, 'steps': 25211, 'loss/train': 1.8811147212982178} 02/25/2022 10:15:19 - INFO - codeparrot_training - Step 25212: {'lr': 0.0002628879045948361, 'samples': 12909056, 'steps': 25212, 'loss/train': 1.6245062351226807} 02/25/2022 10:15:24 - INFO - codeparrot_training - Step 25213: {'lr': 0.0002628715638621595, 'samples': 12909568, 'steps': 25213, 'loss/train': 1.7548400163650513} 02/25/2022 10:15:28 - INFO - codeparrot_training - Step 25214: {'lr': 0.00026285522307434517, 'samples': 12910080, 'steps': 25214, 'loss/train': 2.358091354370117} 02/25/2022 10:15:33 - INFO - codeparrot_training - Step 25215: {'lr': 0.0002628388822314632, 'samples': 12910592, 'steps': 25215, 'loss/train': 2.141122817993164} 02/25/2022 10:15:37 - INFO - codeparrot_training - Step 25216: {'lr': 0.0002628225413335836, 'samples': 12911104, 'steps': 25216, 'loss/train': 1.2398630380630493} 02/25/2022 10:15:42 - INFO - codeparrot_training - Step 25217: {'lr': 0.0002628062003807762, 'samples': 12911616, 'steps': 25217, 'loss/train': 0.9681863784790039} 02/25/2022 10:15:46 - INFO - codeparrot_training - Step 25218: {'lr': 0.0002627898593731111, 'samples': 12912128, 'steps': 25218, 'loss/train': 1.0605509281158447} 02/25/2022 10:15:51 - INFO - codeparrot_training - Step 25219: {'lr': 0.0002627735183106584, 'samples': 12912640, 'steps': 25219, 'loss/train': 1.580093502998352} 02/25/2022 10:15:55 - INFO - codeparrot_training - Step 25220: {'lr': 0.00026275717719348793, 'samples': 12913152, 'steps': 25220, 'loss/train': 0.6143463850021362} 02/25/2022 10:16:00 - INFO - codeparrot_training - Step 25221: {'lr': 0.0002627408360216698, 'samples': 12913664, 'steps': 25221, 'loss/train': 0.4043956696987152} 02/25/2022 10:16:04 - INFO - codeparrot_training - Step 25222: {'lr': 0.00026272449479527386, 'samples': 12914176, 'steps': 25222, 'loss/train': 0.45767590403556824} 02/25/2022 10:16:09 - INFO - codeparrot_training - Step 25223: {'lr': 0.0002627081535143704, 'samples': 12914688, 'steps': 25223, 'loss/train': 1.9036891460418701} 02/25/2022 10:16:13 - INFO - codeparrot_training - Step 25224: {'lr': 0.0002626918121790291, 'samples': 12915200, 'steps': 25224, 'loss/train': 2.329961061477661} 02/25/2022 10:16:20 - INFO - codeparrot_training - Step 25225: {'lr': 0.0002626754707893202, 'samples': 12915712, 'steps': 25225, 'loss/train': 0.49185848236083984} 02/25/2022 10:16:24 - INFO - codeparrot_training - Step 25226: {'lr': 0.00026265912934531353, 'samples': 12916224, 'steps': 25226, 'loss/train': 1.6422491073608398} 02/25/2022 10:16:29 - INFO - codeparrot_training - Step 25227: {'lr': 0.00026264278784707924, 'samples': 12916736, 'steps': 25227, 'loss/train': 2.4964475631713867} 02/25/2022 10:16:33 - INFO - codeparrot_training - Step 25228: {'lr': 0.00026262644629468726, 'samples': 12917248, 'steps': 25228, 'loss/train': 1.4067500829696655} 02/25/2022 10:16:38 - INFO - codeparrot_training - Step 25229: {'lr': 0.0002626101046882076, 'samples': 12917760, 'steps': 25229, 'loss/train': 2.486905336380005} 02/25/2022 10:16:42 - INFO - codeparrot_training - Step 25230: {'lr': 0.00026259376302771023, 'samples': 12918272, 'steps': 25230, 'loss/train': 1.2278902530670166} 02/25/2022 10:16:47 - INFO - codeparrot_training - Step 25231: {'lr': 0.0002625774213132651, 'samples': 12918784, 'steps': 25231, 'loss/train': 1.9769572019577026} 02/25/2022 10:16:51 - INFO - codeparrot_training - Step 25232: {'lr': 0.0002625610795449424, 'samples': 12919296, 'steps': 25232, 'loss/train': 1.467758059501648} 02/25/2022 10:16:56 - INFO - codeparrot_training - Step 25233: {'lr': 0.00026254473772281196, 'samples': 12919808, 'steps': 25233, 'loss/train': 1.4294638633728027} 02/25/2022 10:17:00 - INFO - codeparrot_training - Step 25234: {'lr': 0.0002625283958469439, 'samples': 12920320, 'steps': 25234, 'loss/train': 0.9792097806930542} 02/25/2022 10:17:07 - INFO - codeparrot_training - Step 25235: {'lr': 0.00026251205391740815, 'samples': 12920832, 'steps': 25235, 'loss/train': 2.814891815185547} 02/25/2022 10:17:10 - INFO - codeparrot_training - Step 25236: {'lr': 0.00026249571193427473, 'samples': 12921344, 'steps': 25236, 'loss/train': 1.8181837797164917} 02/25/2022 10:17:16 - INFO - codeparrot_training - Step 25237: {'lr': 0.0002624793698976136, 'samples': 12921856, 'steps': 25237, 'loss/train': 1.7550987005233765} 02/25/2022 10:17:19 - INFO - codeparrot_training - Step 25238: {'lr': 0.0002624630278074949, 'samples': 12922368, 'steps': 25238, 'loss/train': 2.3792221546173096} 02/25/2022 10:17:25 - INFO - codeparrot_training - Step 25239: {'lr': 0.0002624466856639885, 'samples': 12922880, 'steps': 25239, 'loss/train': 1.0399821996688843} 02/25/2022 10:17:28 - INFO - codeparrot_training - Step 25240: {'lr': 0.0002624303434671645, 'samples': 12923392, 'steps': 25240, 'loss/train': 1.0097362995147705} 02/25/2022 10:17:34 - INFO - codeparrot_training - Step 25241: {'lr': 0.00026241400121709287, 'samples': 12923904, 'steps': 25241, 'loss/train': 2.1287190914154053} 02/25/2022 10:17:37 - INFO - codeparrot_training - Step 25242: {'lr': 0.0002623976589138435, 'samples': 12924416, 'steps': 25242, 'loss/train': 1.6061726808547974} 02/25/2022 10:17:43 - INFO - codeparrot_training - Step 25243: {'lr': 0.00026238131655748656, 'samples': 12924928, 'steps': 25243, 'loss/train': 1.814828872680664} 02/25/2022 10:17:46 - INFO - codeparrot_training - Step 25244: {'lr': 0.0002623649741480919, 'samples': 12925440, 'steps': 25244, 'loss/train': 2.7730724811553955} 02/25/2022 10:17:54 - INFO - codeparrot_training - Step 25245: {'lr': 0.00026234863168572974, 'samples': 12925952, 'steps': 25245, 'loss/train': 2.665889263153076} 02/25/2022 10:17:57 - INFO - codeparrot_training - Step 25246: {'lr': 0.0002623322891704699, 'samples': 12926464, 'steps': 25246, 'loss/train': 2.974480390548706} 02/25/2022 10:18:03 - INFO - codeparrot_training - Step 25247: {'lr': 0.0002623159466023825, 'samples': 12926976, 'steps': 25247, 'loss/train': 0.7214266061782837} 02/25/2022 10:18:06 - INFO - codeparrot_training - Step 25248: {'lr': 0.0002622996039815374, 'samples': 12927488, 'steps': 25248, 'loss/train': 1.5576725006103516} 02/25/2022 10:18:12 - INFO - codeparrot_training - Step 25249: {'lr': 0.0002622832613080048, 'samples': 12928000, 'steps': 25249, 'loss/train': 2.1740376949310303} 02/25/2022 10:18:15 - INFO - codeparrot_training - Step 25250: {'lr': 0.00026226691858185456, 'samples': 12928512, 'steps': 25250, 'loss/train': 2.48518967628479} 02/25/2022 10:18:21 - INFO - codeparrot_training - Step 25251: {'lr': 0.0002622505758031567, 'samples': 12929024, 'steps': 25251, 'loss/train': 2.667180299758911} 02/25/2022 10:18:24 - INFO - codeparrot_training - Step 25252: {'lr': 0.00026223423297198136, 'samples': 12929536, 'steps': 25252, 'loss/train': 2.1488027572631836} 02/25/2022 10:18:30 - INFO - codeparrot_training - Step 25253: {'lr': 0.00026221789008839836, 'samples': 12930048, 'steps': 25253, 'loss/train': 1.2392905950546265} 02/25/2022 10:18:33 - INFO - codeparrot_training - Step 25254: {'lr': 0.00026220154715247783, 'samples': 12930560, 'steps': 25254, 'loss/train': 1.8072820901870728} 02/25/2022 10:18:39 - INFO - codeparrot_training - Step 25255: {'lr': 0.00026218520416428976, 'samples': 12931072, 'steps': 25255, 'loss/train': 2.21661376953125} 02/25/2022 10:18:42 - INFO - codeparrot_training - Step 25256: {'lr': 0.00026216886112390413, 'samples': 12931584, 'steps': 25256, 'loss/train': 2.3472323417663574} 02/25/2022 10:18:49 - INFO - codeparrot_training - Step 25257: {'lr': 0.00026215251803139093, 'samples': 12932096, 'steps': 25257, 'loss/train': 1.6080751419067383} 02/25/2022 10:18:53 - INFO - codeparrot_training - Step 25258: {'lr': 0.0002621361748868203, 'samples': 12932608, 'steps': 25258, 'loss/train': 1.6127243041992188} 02/25/2022 10:18:58 - INFO - codeparrot_training - Step 25259: {'lr': 0.00026211983169026207, 'samples': 12933120, 'steps': 25259, 'loss/train': 1.5720276832580566} 02/25/2022 10:19:02 - INFO - codeparrot_training - Step 25260: {'lr': 0.0002621034884417864, 'samples': 12933632, 'steps': 25260, 'loss/train': 1.5244604349136353} 02/25/2022 10:19:07 - INFO - codeparrot_training - Step 25261: {'lr': 0.0002620871451414631, 'samples': 12934144, 'steps': 25261, 'loss/train': 1.9730055332183838} 02/25/2022 10:19:13 - INFO - codeparrot_training - Step 25262: {'lr': 0.00026207080178936243, 'samples': 12934656, 'steps': 25262, 'loss/train': 1.9861608743667603} 02/25/2022 10:19:16 - INFO - codeparrot_training - Step 25263: {'lr': 0.0002620544583855542, 'samples': 12935168, 'steps': 25263, 'loss/train': 0.6329479813575745} 02/25/2022 10:19:22 - INFO - codeparrot_training - Step 25264: {'lr': 0.00026203811493010854, 'samples': 12935680, 'steps': 25264, 'loss/train': 2.566152572631836} 02/25/2022 10:19:25 - INFO - codeparrot_training - Step 25265: {'lr': 0.0002620217714230954, 'samples': 12936192, 'steps': 25265, 'loss/train': 1.6782448291778564} 02/25/2022 10:19:31 - INFO - codeparrot_training - Step 25266: {'lr': 0.0002620054278645848, 'samples': 12936704, 'steps': 25266, 'loss/train': 1.4395540952682495} 02/25/2022 10:19:34 - INFO - codeparrot_training - Step 25267: {'lr': 0.00026198908425464684, 'samples': 12937216, 'steps': 25267, 'loss/train': 2.275939464569092} 02/25/2022 10:19:40 - INFO - codeparrot_training - Step 25268: {'lr': 0.00026197274059335137, 'samples': 12937728, 'steps': 25268, 'loss/train': 1.917593002319336} 02/25/2022 10:19:43 - INFO - codeparrot_training - Step 25269: {'lr': 0.0002619563968807685, 'samples': 12938240, 'steps': 25269, 'loss/train': 1.9233434200286865} 02/25/2022 10:19:49 - INFO - codeparrot_training - Step 25270: {'lr': 0.00026194005311696824, 'samples': 12938752, 'steps': 25270, 'loss/train': 1.6113722324371338} 02/25/2022 10:19:52 - INFO - codeparrot_training - Step 25271: {'lr': 0.0002619237093020206, 'samples': 12939264, 'steps': 25271, 'loss/train': 0.8846839666366577} 02/25/2022 10:20:00 - INFO - codeparrot_training - Step 25272: {'lr': 0.00026190736543599545, 'samples': 12939776, 'steps': 25272, 'loss/train': 1.7859965562820435} 02/25/2022 10:20:03 - INFO - codeparrot_training - Step 25273: {'lr': 0.0002618910215189631, 'samples': 12940288, 'steps': 25273, 'loss/train': 1.5689723491668701} 02/25/2022 10:20:08 - INFO - codeparrot_training - Step 25274: {'lr': 0.0002618746775509933, 'samples': 12940800, 'steps': 25274, 'loss/train': 2.146951913833618} 02/25/2022 10:20:12 - INFO - codeparrot_training - Step 25275: {'lr': 0.0002618583335321561, 'samples': 12941312, 'steps': 25275, 'loss/train': 2.030069351196289} 02/25/2022 10:20:18 - INFO - codeparrot_training - Step 25276: {'lr': 0.00026184198946252164, 'samples': 12941824, 'steps': 25276, 'loss/train': 1.7635247707366943} 02/25/2022 10:20:21 - INFO - codeparrot_training - Step 25277: {'lr': 0.00026182564534215985, 'samples': 12942336, 'steps': 25277, 'loss/train': 0.9114202857017517} 02/25/2022 10:20:27 - INFO - codeparrot_training - Step 25278: {'lr': 0.00026180930117114076, 'samples': 12942848, 'steps': 25278, 'loss/train': 2.3035519123077393} 02/25/2022 10:20:30 - INFO - codeparrot_training - Step 25279: {'lr': 0.0002617929569495343, 'samples': 12943360, 'steps': 25279, 'loss/train': 1.680801510810852} 02/25/2022 10:20:36 - INFO - codeparrot_training - Step 25280: {'lr': 0.00026177661267741067, 'samples': 12943872, 'steps': 25280, 'loss/train': 1.2134040594100952} 02/25/2022 10:20:39 - INFO - codeparrot_training - Step 25281: {'lr': 0.00026176026835483975, 'samples': 12944384, 'steps': 25281, 'loss/train': 0.7380163669586182} 02/25/2022 10:20:46 - INFO - codeparrot_training - Step 25282: {'lr': 0.00026174392398189153, 'samples': 12944896, 'steps': 25282, 'loss/train': 2.2398416996002197} 02/25/2022 10:20:50 - INFO - codeparrot_training - Step 25283: {'lr': 0.00026172757955863606, 'samples': 12945408, 'steps': 25283, 'loss/train': 1.9276384115219116} 02/25/2022 10:20:55 - INFO - codeparrot_training - Step 25284: {'lr': 0.0002617112350851434, 'samples': 12945920, 'steps': 25284, 'loss/train': 1.5832734107971191} 02/25/2022 10:20:59 - INFO - codeparrot_training - Step 25285: {'lr': 0.00026169489056148353, 'samples': 12946432, 'steps': 25285, 'loss/train': 2.6555488109588623} 02/25/2022 10:21:04 - INFO - codeparrot_training - Step 25286: {'lr': 0.0002616785459877264, 'samples': 12946944, 'steps': 25286, 'loss/train': 2.1728549003601074} 02/25/2022 10:21:08 - INFO - codeparrot_training - Step 25287: {'lr': 0.00026166220136394214, 'samples': 12947456, 'steps': 25287, 'loss/train': 2.062012195587158} 02/25/2022 10:21:13 - INFO - codeparrot_training - Step 25288: {'lr': 0.0002616458566902007, 'samples': 12947968, 'steps': 25288, 'loss/train': 2.2282443046569824} 02/25/2022 10:21:17 - INFO - codeparrot_training - Step 25289: {'lr': 0.00026162951196657215, 'samples': 12948480, 'steps': 25289, 'loss/train': 2.558828830718994} 02/25/2022 10:21:22 - INFO - codeparrot_training - Step 25290: {'lr': 0.00026161316719312637, 'samples': 12948992, 'steps': 25290, 'loss/train': 2.0352771282196045} 02/25/2022 10:21:26 - INFO - codeparrot_training - Step 25291: {'lr': 0.0002615968223699336, 'samples': 12949504, 'steps': 25291, 'loss/train': 2.081367015838623} 02/25/2022 10:21:33 - INFO - codeparrot_training - Step 25292: {'lr': 0.0002615804774970636, 'samples': 12950016, 'steps': 25292, 'loss/train': 1.9572584629058838} 02/25/2022 10:21:36 - INFO - codeparrot_training - Step 25293: {'lr': 0.0002615641325745866, 'samples': 12950528, 'steps': 25293, 'loss/train': 1.195730209350586} 02/25/2022 10:21:42 - INFO - codeparrot_training - Step 25294: {'lr': 0.00026154778760257246, 'samples': 12951040, 'steps': 25294, 'loss/train': 0.3673703670501709} 02/25/2022 10:21:45 - INFO - codeparrot_training - Step 25295: {'lr': 0.0002615314425810913, 'samples': 12951552, 'steps': 25295, 'loss/train': 1.7312372922897339} 02/25/2022 10:21:51 - INFO - codeparrot_training - Step 25296: {'lr': 0.0002615150975102131, 'samples': 12952064, 'steps': 25296, 'loss/train': 2.1584668159484863} 02/25/2022 10:21:54 - INFO - codeparrot_training - Step 25297: {'lr': 0.00026149875239000786, 'samples': 12952576, 'steps': 25297, 'loss/train': 2.6947379112243652} 02/25/2022 10:22:00 - INFO - codeparrot_training - Step 25298: {'lr': 0.0002614824072205456, 'samples': 12953088, 'steps': 25298, 'loss/train': 1.9793378114700317} 02/25/2022 10:22:03 - INFO - codeparrot_training - Step 25299: {'lr': 0.00026146606200189637, 'samples': 12953600, 'steps': 25299, 'loss/train': 2.1179208755493164} 02/25/2022 10:22:09 - INFO - codeparrot_training - Step 25300: {'lr': 0.0002614497167341302, 'samples': 12954112, 'steps': 25300, 'loss/train': 2.183959484100342} 02/25/2022 10:22:12 - INFO - codeparrot_training - Step 25301: {'lr': 0.00026143337141731703, 'samples': 12954624, 'steps': 25301, 'loss/train': 1.6719461679458618} 02/25/2022 10:22:20 - INFO - codeparrot_training - Step 25302: {'lr': 0.000261417026051527, 'samples': 12955136, 'steps': 25302, 'loss/train': 1.7023910284042358} 02/25/2022 10:22:23 - INFO - codeparrot_training - Step 25303: {'lr': 0.00026140068063683, 'samples': 12955648, 'steps': 25303, 'loss/train': 0.7320264577865601} 02/25/2022 10:22:29 - INFO - codeparrot_training - Step 25304: {'lr': 0.00026138433517329616, 'samples': 12956160, 'steps': 25304, 'loss/train': 2.1170542240142822} 02/25/2022 10:22:32 - INFO - codeparrot_training - Step 25305: {'lr': 0.0002613679896609954, 'samples': 12956672, 'steps': 25305, 'loss/train': 1.7571150064468384} 02/25/2022 10:22:38 - INFO - codeparrot_training - Step 25306: {'lr': 0.0002613516440999978, 'samples': 12957184, 'steps': 25306, 'loss/train': 1.5068050622940063} 02/25/2022 10:22:41 - INFO - codeparrot_training - Step 25307: {'lr': 0.0002613352984903733, 'samples': 12957696, 'steps': 25307, 'loss/train': 2.0442020893096924} 02/25/2022 10:22:47 - INFO - codeparrot_training - Step 25308: {'lr': 0.0002613189528321921, 'samples': 12958208, 'steps': 25308, 'loss/train': 2.2437171936035156} 02/25/2022 10:22:51 - INFO - codeparrot_training - Step 25309: {'lr': 0.00026130260712552394, 'samples': 12958720, 'steps': 25309, 'loss/train': 1.404790997505188} 02/25/2022 10:22:56 - INFO - codeparrot_training - Step 25310: {'lr': 0.00026128626137043913, 'samples': 12959232, 'steps': 25310, 'loss/train': 1.9766829013824463} 02/25/2022 10:22:59 - INFO - codeparrot_training - Step 25311: {'lr': 0.0002612699155670075, 'samples': 12959744, 'steps': 25311, 'loss/train': 1.4593075513839722} 02/25/2022 10:23:05 - INFO - codeparrot_training - Step 25312: {'lr': 0.0002612535697152991, 'samples': 12960256, 'steps': 25312, 'loss/train': 1.157163381576538} 02/25/2022 10:23:08 - INFO - codeparrot_training - Step 25313: {'lr': 0.0002612372238153841, 'samples': 12960768, 'steps': 25313, 'loss/train': 1.974704384803772} 02/25/2022 10:23:14 - INFO - codeparrot_training - Step 25314: {'lr': 0.00026122087786733226, 'samples': 12961280, 'steps': 25314, 'loss/train': 2.5316545963287354} 02/25/2022 10:23:17 - INFO - codeparrot_training - Step 25315: {'lr': 0.00026120453187121386, 'samples': 12961792, 'steps': 25315, 'loss/train': 1.7827507257461548} 02/25/2022 10:23:23 - INFO - codeparrot_training - Step 25316: {'lr': 0.00026118818582709875, 'samples': 12962304, 'steps': 25316, 'loss/train': 1.0745261907577515} 02/25/2022 10:23:27 - INFO - codeparrot_training - Step 25317: {'lr': 0.000261171839735057, 'samples': 12962816, 'steps': 25317, 'loss/train': 0.8575783371925354} 02/25/2022 10:23:34 - INFO - codeparrot_training - Step 25318: {'lr': 0.00026115549359515867, 'samples': 12963328, 'steps': 25318, 'loss/train': 2.160360813140869} 02/25/2022 10:23:37 - INFO - codeparrot_training - Step 25319: {'lr': 0.00026113914740747373, 'samples': 12963840, 'steps': 25319, 'loss/train': 1.8212852478027344} 02/25/2022 10:23:43 - INFO - codeparrot_training - Step 25320: {'lr': 0.00026112280117207223, 'samples': 12964352, 'steps': 25320, 'loss/train': 2.2596120834350586} 02/25/2022 10:23:46 - INFO - codeparrot_training - Step 25321: {'lr': 0.00026110645488902417, 'samples': 12964864, 'steps': 25321, 'loss/train': 1.876841425895691} 02/25/2022 10:23:52 - INFO - codeparrot_training - Step 25322: {'lr': 0.00026109010855839953, 'samples': 12965376, 'steps': 25322, 'loss/train': 2.182778835296631} 02/25/2022 10:23:55 - INFO - codeparrot_training - Step 25323: {'lr': 0.00026107376218026846, 'samples': 12965888, 'steps': 25323, 'loss/train': 1.8846172094345093} 02/25/2022 10:24:01 - INFO - codeparrot_training - Step 25324: {'lr': 0.0002610574157547009, 'samples': 12966400, 'steps': 25324, 'loss/train': 1.6929960250854492} 02/25/2022 10:24:04 - INFO - codeparrot_training - Step 25325: {'lr': 0.00026104106928176684, 'samples': 12966912, 'steps': 25325, 'loss/train': 1.2628182172775269} 02/25/2022 10:24:10 - INFO - codeparrot_training - Step 25326: {'lr': 0.0002610247227615364, 'samples': 12967424, 'steps': 25326, 'loss/train': 3.2269182205200195} 02/25/2022 10:24:13 - INFO - codeparrot_training - Step 25327: {'lr': 0.00026100837619407954, 'samples': 12967936, 'steps': 25327, 'loss/train': 2.1335484981536865} 02/25/2022 10:24:21 - INFO - codeparrot_training - Step 25328: {'lr': 0.0002609920295794662, 'samples': 12968448, 'steps': 25328, 'loss/train': 1.5020781755447388} 02/25/2022 10:24:25 - INFO - codeparrot_training - Step 25329: {'lr': 0.00026097568291776665, 'samples': 12968960, 'steps': 25329, 'loss/train': 1.9993840456008911} 02/25/2022 10:24:30 - INFO - codeparrot_training - Step 25330: {'lr': 0.0002609593362090507, 'samples': 12969472, 'steps': 25330, 'loss/train': 2.5293819904327393} 02/25/2022 10:24:34 - INFO - codeparrot_training - Step 25331: {'lr': 0.0002609429894533884, 'samples': 12969984, 'steps': 25331, 'loss/train': 1.8078157901763916} 02/25/2022 10:24:39 - INFO - codeparrot_training - Step 25332: {'lr': 0.00026092664265084983, 'samples': 12970496, 'steps': 25332, 'loss/train': 2.028881788253784} 02/25/2022 10:24:43 - INFO - codeparrot_training - Step 25333: {'lr': 0.0002609102958015049, 'samples': 12971008, 'steps': 25333, 'loss/train': 2.8448386192321777} 02/25/2022 10:24:48 - INFO - codeparrot_training - Step 25334: {'lr': 0.0002608939489054239, 'samples': 12971520, 'steps': 25334, 'loss/train': 1.7702529430389404} 02/25/2022 10:24:52 - INFO - codeparrot_training - Step 25335: {'lr': 0.00026087760196267655, 'samples': 12972032, 'steps': 25335, 'loss/train': 1.3746795654296875} 02/25/2022 10:24:58 - INFO - codeparrot_training - Step 25336: {'lr': 0.000260861254973333, 'samples': 12972544, 'steps': 25336, 'loss/train': 0.6420924067497253} 02/25/2022 10:25:01 - INFO - codeparrot_training - Step 25337: {'lr': 0.00026084490793746337, 'samples': 12973056, 'steps': 25337, 'loss/train': 1.5441663265228271} 02/25/2022 10:25:07 - INFO - codeparrot_training - Step 25338: {'lr': 0.00026082856085513756, 'samples': 12973568, 'steps': 25338, 'loss/train': 2.481992721557617} 02/25/2022 10:25:11 - INFO - codeparrot_training - Step 25339: {'lr': 0.0002608122137264257, 'samples': 12974080, 'steps': 25339, 'loss/train': 2.3662033081054688} 02/25/2022 10:25:17 - INFO - codeparrot_training - Step 25340: {'lr': 0.0002607958665513976, 'samples': 12974592, 'steps': 25340, 'loss/train': 0.8369866013526917} 02/25/2022 10:25:20 - INFO - codeparrot_training - Step 25341: {'lr': 0.00026077951933012355, 'samples': 12975104, 'steps': 25341, 'loss/train': 1.852104902267456} 02/25/2022 10:25:26 - INFO - codeparrot_training - Step 25342: {'lr': 0.0002607631720626734, 'samples': 12975616, 'steps': 25342, 'loss/train': 1.4531915187835693} 02/25/2022 10:25:29 - INFO - codeparrot_training - Step 25343: {'lr': 0.0002607468247491173, 'samples': 12976128, 'steps': 25343, 'loss/train': 1.8628367185592651} 02/25/2022 10:25:36 - INFO - codeparrot_training - Step 25344: {'lr': 0.00026073047738952513, 'samples': 12976640, 'steps': 25344, 'loss/train': 0.22304286062717438} 02/25/2022 10:25:40 - INFO - codeparrot_training - Step 25345: {'lr': 0.0002607141299839671, 'samples': 12977152, 'steps': 25345, 'loss/train': 1.4542042016983032} 02/25/2022 10:25:45 - INFO - codeparrot_training - Step 25346: {'lr': 0.0002606977825325131, 'samples': 12977664, 'steps': 25346, 'loss/train': 3.1794064044952393} 02/25/2022 10:25:49 - INFO - codeparrot_training - Step 25347: {'lr': 0.0002606814350352331, 'samples': 12978176, 'steps': 25347, 'loss/train': 1.249501347541809} 02/25/2022 10:25:54 - INFO - codeparrot_training - Step 25348: {'lr': 0.00026066508749219734, 'samples': 12978688, 'steps': 25348, 'loss/train': 2.259679079055786} 02/25/2022 10:25:58 - INFO - codeparrot_training - Step 25349: {'lr': 0.0002606487399034757, 'samples': 12979200, 'steps': 25349, 'loss/train': 2.393623113632202} 02/25/2022 10:26:03 - INFO - codeparrot_training - Step 25350: {'lr': 0.0002606323922691383, 'samples': 12979712, 'steps': 25350, 'loss/train': 2.1388514041900635} 02/25/2022 10:26:07 - INFO - codeparrot_training - Step 25351: {'lr': 0.000260616044589255, 'samples': 12980224, 'steps': 25351, 'loss/train': 2.245131254196167} 02/25/2022 10:26:12 - INFO - codeparrot_training - Step 25352: {'lr': 0.00026059969686389605, 'samples': 12980736, 'steps': 25352, 'loss/train': 1.0605303049087524} 02/25/2022 10:26:16 - INFO - codeparrot_training - Step 25353: {'lr': 0.0002605833490931313, 'samples': 12981248, 'steps': 25353, 'loss/train': 1.6841325759887695} 02/25/2022 10:26:22 - INFO - codeparrot_training - Step 25354: {'lr': 0.00026056700127703085, 'samples': 12981760, 'steps': 25354, 'loss/train': 1.1205074787139893} 02/25/2022 10:26:26 - INFO - codeparrot_training - Step 25355: {'lr': 0.0002605506534156647, 'samples': 12982272, 'steps': 25355, 'loss/train': 1.686244010925293} 02/25/2022 10:26:31 - INFO - codeparrot_training - Step 25356: {'lr': 0.00026053430550910297, 'samples': 12982784, 'steps': 25356, 'loss/train': 2.243374824523926} 02/25/2022 10:26:35 - INFO - codeparrot_training - Step 25357: {'lr': 0.0002605179575574155, 'samples': 12983296, 'steps': 25357, 'loss/train': 1.817054033279419} 02/25/2022 10:26:40 - INFO - codeparrot_training - Step 25358: {'lr': 0.0002605016095606726, 'samples': 12983808, 'steps': 25358, 'loss/train': 1.4486634731292725} 02/25/2022 10:26:44 - INFO - codeparrot_training - Step 25359: {'lr': 0.00026048526151894395, 'samples': 12984320, 'steps': 25359, 'loss/train': 0.8246952295303345} 02/25/2022 10:26:49 - INFO - codeparrot_training - Step 25360: {'lr': 0.0002604689134322999, 'samples': 12984832, 'steps': 25360, 'loss/train': 1.6542701721191406} 02/25/2022 10:26:53 - INFO - codeparrot_training - Step 25361: {'lr': 0.00026045256530081033, 'samples': 12985344, 'steps': 25361, 'loss/train': 2.278477907180786} 02/25/2022 10:26:58 - INFO - codeparrot_training - Step 25362: {'lr': 0.00026043621712454524, 'samples': 12985856, 'steps': 25362, 'loss/train': 1.476204514503479} 02/25/2022 10:27:02 - INFO - codeparrot_training - Step 25363: {'lr': 0.00026041986890357476, 'samples': 12986368, 'steps': 25363, 'loss/train': 2.968668222427368} 02/25/2022 10:27:08 - INFO - codeparrot_training - Step 25364: {'lr': 0.00026040352063796886, 'samples': 12986880, 'steps': 25364, 'loss/train': 1.6224629878997803} 02/25/2022 10:27:12 - INFO - codeparrot_training - Step 25365: {'lr': 0.00026038717232779765, 'samples': 12987392, 'steps': 25365, 'loss/train': 0.9365993738174438} 02/25/2022 10:27:17 - INFO - codeparrot_training - Step 25366: {'lr': 0.00026037082397313094, 'samples': 12987904, 'steps': 25366, 'loss/train': 2.612879991531372} 02/25/2022 10:27:21 - INFO - codeparrot_training - Step 25367: {'lr': 0.000260354475574039, 'samples': 12988416, 'steps': 25367, 'loss/train': 1.475862979888916} 02/25/2022 10:27:26 - INFO - codeparrot_training - Step 25368: {'lr': 0.0002603381271305918, 'samples': 12988928, 'steps': 25368, 'loss/train': 2.3788938522338867} 02/25/2022 10:27:30 - INFO - codeparrot_training - Step 25369: {'lr': 0.0002603217786428593, 'samples': 12989440, 'steps': 25369, 'loss/train': 1.380410075187683} 02/25/2022 10:27:35 - INFO - codeparrot_training - Step 25370: {'lr': 0.0002603054301109117, 'samples': 12989952, 'steps': 25370, 'loss/train': 2.6743950843811035} 02/25/2022 10:27:39 - INFO - codeparrot_training - Step 25371: {'lr': 0.00026028908153481875, 'samples': 12990464, 'steps': 25371, 'loss/train': 2.0259957313537598} 02/25/2022 10:27:44 - INFO - codeparrot_training - Step 25372: {'lr': 0.0002602727329146507, 'samples': 12990976, 'steps': 25372, 'loss/train': 1.5719716548919678} 02/25/2022 10:27:48 - INFO - codeparrot_training - Step 25373: {'lr': 0.00026025638425047746, 'samples': 12991488, 'steps': 25373, 'loss/train': 1.8571712970733643} 02/25/2022 10:27:55 - INFO - codeparrot_training - Step 25374: {'lr': 0.00026024003554236925, 'samples': 12992000, 'steps': 25374, 'loss/train': 2.4841511249542236} 02/25/2022 10:27:58 - INFO - codeparrot_training - Step 25375: {'lr': 0.0002602236867903959, 'samples': 12992512, 'steps': 25375, 'loss/train': 3.207655191421509} 02/25/2022 10:28:04 - INFO - codeparrot_training - Step 25376: {'lr': 0.00026020733799462755, 'samples': 12993024, 'steps': 25376, 'loss/train': 2.3679535388946533} 02/25/2022 10:28:07 - INFO - codeparrot_training - Step 25377: {'lr': 0.0002601909891551342, 'samples': 12993536, 'steps': 25377, 'loss/train': 2.1880011558532715} 02/25/2022 10:28:13 - INFO - codeparrot_training - Step 25378: {'lr': 0.0002601746402719859, 'samples': 12994048, 'steps': 25378, 'loss/train': 2.410015106201172} 02/25/2022 10:28:16 - INFO - codeparrot_training - Step 25379: {'lr': 0.0002601582913452526, 'samples': 12994560, 'steps': 25379, 'loss/train': 1.9797877073287964} 02/25/2022 10:28:22 - INFO - codeparrot_training - Step 25380: {'lr': 0.0002601419423750045, 'samples': 12995072, 'steps': 25380, 'loss/train': 1.0747326612472534} 02/25/2022 10:28:25 - INFO - codeparrot_training - Step 25381: {'lr': 0.00026012559336131147, 'samples': 12995584, 'steps': 25381, 'loss/train': 2.887636423110962} 02/25/2022 10:28:31 - INFO - codeparrot_training - Step 25382: {'lr': 0.0002601092443042437, 'samples': 12996096, 'steps': 25382, 'loss/train': 1.7689247131347656} 02/25/2022 10:28:34 - INFO - codeparrot_training - Step 25383: {'lr': 0.000260092895203871, 'samples': 12996608, 'steps': 25383, 'loss/train': 2.6680214405059814} 02/25/2022 10:28:40 - INFO - codeparrot_training - Step 25384: {'lr': 0.0002600765460602636, 'samples': 12997120, 'steps': 25384, 'loss/train': 1.4685745239257812} 02/25/2022 10:28:44 - INFO - codeparrot_training - Step 25385: {'lr': 0.0002600601968734915, 'samples': 12997632, 'steps': 25385, 'loss/train': 2.89216685295105} 02/25/2022 10:28:49 - INFO - codeparrot_training - Step 25386: {'lr': 0.00026004384764362473, 'samples': 12998144, 'steps': 25386, 'loss/train': 2.2213428020477295} 02/25/2022 10:28:53 - INFO - codeparrot_training - Step 25387: {'lr': 0.0002600274983707333, 'samples': 12998656, 'steps': 25387, 'loss/train': 1.8447260856628418} 02/25/2022 10:28:58 - INFO - codeparrot_training - Step 25388: {'lr': 0.0002600111490548872, 'samples': 12999168, 'steps': 25388, 'loss/train': 1.543138861656189} 02/25/2022 10:29:02 - INFO - codeparrot_training - Step 25389: {'lr': 0.00025999479969615656, 'samples': 12999680, 'steps': 25389, 'loss/train': 2.214895725250244} 02/25/2022 10:29:07 - INFO - codeparrot_training - Step 25390: {'lr': 0.00025997845029461134, 'samples': 13000192, 'steps': 25390, 'loss/train': 1.2128826379776} 02/25/2022 10:29:11 - INFO - codeparrot_training - Step 25391: {'lr': 0.00025996210085032167, 'samples': 13000704, 'steps': 25391, 'loss/train': 2.0411376953125} 02/25/2022 10:29:16 - INFO - codeparrot_training - Step 25392: {'lr': 0.00025994575136335747, 'samples': 13001216, 'steps': 25392, 'loss/train': 0.09121988713741302} 02/25/2022 10:29:20 - INFO - codeparrot_training - Step 25393: {'lr': 0.0002599294018337889, 'samples': 13001728, 'steps': 25393, 'loss/train': 1.353220820426941} 02/25/2022 10:29:25 - INFO - codeparrot_training - Step 25394: {'lr': 0.0002599130522616858, 'samples': 13002240, 'steps': 25394, 'loss/train': 1.7172094583511353} 02/25/2022 10:29:29 - INFO - codeparrot_training - Step 25395: {'lr': 0.00025989670264711846, 'samples': 13002752, 'steps': 25395, 'loss/train': 2.8422091007232666} 02/25/2022 10:29:34 - INFO - codeparrot_training - Step 25396: {'lr': 0.0002598803529901567, 'samples': 13003264, 'steps': 25396, 'loss/train': 1.205527901649475} 02/25/2022 10:29:38 - INFO - codeparrot_training - Step 25397: {'lr': 0.0002598640032908706, 'samples': 13003776, 'steps': 25397, 'loss/train': 1.7367963790893555} 02/25/2022 10:29:43 - INFO - codeparrot_training - Step 25398: {'lr': 0.0002598476535493304, 'samples': 13004288, 'steps': 25398, 'loss/train': 1.991734504699707} 02/25/2022 10:29:47 - INFO - codeparrot_training - Step 25399: {'lr': 0.0002598313037656058, 'samples': 13004800, 'steps': 25399, 'loss/train': 2.408982992172241} 02/25/2022 10:29:53 - INFO - codeparrot_training - Step 25400: {'lr': 0.00025981495393976716, 'samples': 13005312, 'steps': 25400, 'loss/train': 2.187908887863159} 02/25/2022 10:29:57 - INFO - codeparrot_training - Step 25401: {'lr': 0.0002597986040718843, 'samples': 13005824, 'steps': 25401, 'loss/train': 1.5372228622436523} 02/25/2022 10:30:02 - INFO - codeparrot_training - Step 25402: {'lr': 0.0002597822541620274, 'samples': 13006336, 'steps': 25402, 'loss/train': 1.9885507822036743} 02/25/2022 10:30:06 - INFO - codeparrot_training - Step 25403: {'lr': 0.00025976590421026636, 'samples': 13006848, 'steps': 25403, 'loss/train': 2.3181846141815186} 02/25/2022 10:30:11 - INFO - codeparrot_training - Step 25404: {'lr': 0.00025974955421667134, 'samples': 13007360, 'steps': 25404, 'loss/train': 2.372411012649536} 02/25/2022 10:30:15 - INFO - codeparrot_training - Step 25405: {'lr': 0.00025973320418131227, 'samples': 13007872, 'steps': 25405, 'loss/train': 1.2195004224777222} 02/25/2022 10:30:20 - INFO - codeparrot_training - Step 25406: {'lr': 0.00025971685410425933, 'samples': 13008384, 'steps': 25406, 'loss/train': 8.741043090820312} 02/25/2022 10:30:24 - INFO - codeparrot_training - Step 25407: {'lr': 0.00025970050398558235, 'samples': 13008896, 'steps': 25407, 'loss/train': 0.9202667474746704} 02/25/2022 10:30:29 - INFO - codeparrot_training - Step 25408: {'lr': 0.00025968415382535153, 'samples': 13009408, 'steps': 25408, 'loss/train': 2.025219440460205} 02/25/2022 10:30:33 - INFO - codeparrot_training - Step 25409: {'lr': 0.000259667803623637, 'samples': 13009920, 'steps': 25409, 'loss/train': 2.1659293174743652} 02/25/2022 10:30:39 - INFO - codeparrot_training - Step 25410: {'lr': 0.00025965145338050855, 'samples': 13010432, 'steps': 25410, 'loss/train': 2.5328400135040283} 02/25/2022 10:30:43 - INFO - codeparrot_training - Step 25411: {'lr': 0.00025963510309603635, 'samples': 13010944, 'steps': 25411, 'loss/train': 2.10746431350708} 02/25/2022 10:30:48 - INFO - codeparrot_training - Step 25412: {'lr': 0.0002596187527702904, 'samples': 13011456, 'steps': 25412, 'loss/train': 1.6793832778930664} 02/25/2022 10:30:51 - INFO - codeparrot_training - Step 25413: {'lr': 0.0002596024024033408, 'samples': 13011968, 'steps': 25413, 'loss/train': 1.3334057331085205} 02/25/2022 10:30:57 - INFO - codeparrot_training - Step 25414: {'lr': 0.00025958605199525756, 'samples': 13012480, 'steps': 25414, 'loss/train': 1.6516681909561157} 02/25/2022 10:31:01 - INFO - codeparrot_training - Step 25415: {'lr': 0.0002595697015461107, 'samples': 13012992, 'steps': 25415, 'loss/train': 1.3304407596588135} 02/25/2022 10:31:07 - INFO - codeparrot_training - Step 25416: {'lr': 0.00025955335105597036, 'samples': 13013504, 'steps': 25416, 'loss/train': 1.3077303171157837} 02/25/2022 10:31:10 - INFO - codeparrot_training - Step 25417: {'lr': 0.0002595370005249064, 'samples': 13014016, 'steps': 25417, 'loss/train': 2.26249361038208} 02/25/2022 10:31:14 - INFO - codeparrot_training - Step 25418: {'lr': 0.00025952064995298895, 'samples': 13014528, 'steps': 25418, 'loss/train': 0.6702812910079956} 02/25/2022 10:31:19 - INFO - codeparrot_training - Step 25419: {'lr': 0.0002595042993402882, 'samples': 13015040, 'steps': 25419, 'loss/train': 0.7806688547134399} 02/25/2022 10:31:23 - INFO - codeparrot_training - Step 25420: {'lr': 0.0002594879486868739, 'samples': 13015552, 'steps': 25420, 'loss/train': 1.027009129524231} 02/25/2022 10:31:29 - INFO - codeparrot_training - Step 25421: {'lr': 0.00025947159799281623, 'samples': 13016064, 'steps': 25421, 'loss/train': 2.146104335784912} 02/25/2022 10:31:33 - INFO - codeparrot_training - Step 25422: {'lr': 0.0002594552472581853, 'samples': 13016576, 'steps': 25422, 'loss/train': 2.3594746589660645} 02/25/2022 10:31:38 - INFO - codeparrot_training - Step 25423: {'lr': 0.00025943889648305106, 'samples': 13017088, 'steps': 25423, 'loss/train': 2.610564947128296} 02/25/2022 10:31:42 - INFO - codeparrot_training - Step 25424: {'lr': 0.0002594225456674837, 'samples': 13017600, 'steps': 25424, 'loss/train': 0.26570653915405273} 02/25/2022 10:31:47 - INFO - codeparrot_training - Step 25425: {'lr': 0.000259406194811553, 'samples': 13018112, 'steps': 25425, 'loss/train': 1.6636470556259155} 02/25/2022 10:31:51 - INFO - codeparrot_training - Step 25426: {'lr': 0.0002593898439153293, 'samples': 13018624, 'steps': 25426, 'loss/train': 1.9744212627410889} 02/25/2022 10:31:56 - INFO - codeparrot_training - Step 25427: {'lr': 0.00025937349297888235, 'samples': 13019136, 'steps': 25427, 'loss/train': 1.8284608125686646} 02/25/2022 10:32:00 - INFO - codeparrot_training - Step 25428: {'lr': 0.0002593571420022824, 'samples': 13019648, 'steps': 25428, 'loss/train': 1.9696407318115234} 02/25/2022 10:32:05 - INFO - codeparrot_training - Step 25429: {'lr': 0.0002593407909855994, 'samples': 13020160, 'steps': 25429, 'loss/train': 1.9532917737960815} 02/25/2022 10:32:09 - INFO - codeparrot_training - Step 25430: {'lr': 0.00025932443992890343, 'samples': 13020672, 'steps': 25430, 'loss/train': 0.7508223652839661} 02/25/2022 10:32:16 - INFO - codeparrot_training - Step 25431: {'lr': 0.0002593080888322645, 'samples': 13021184, 'steps': 25431, 'loss/train': 1.1213818788528442} 02/25/2022 10:32:19 - INFO - codeparrot_training - Step 25432: {'lr': 0.00025929173769575266, 'samples': 13021696, 'steps': 25432, 'loss/train': 0.7965890169143677} 02/25/2022 10:32:25 - INFO - codeparrot_training - Step 25433: {'lr': 0.000259275386519438, 'samples': 13022208, 'steps': 25433, 'loss/train': 1.337225317955017} 02/25/2022 10:32:28 - INFO - codeparrot_training - Step 25434: {'lr': 0.0002592590353033905, 'samples': 13022720, 'steps': 25434, 'loss/train': 1.908464789390564} 02/25/2022 10:32:34 - INFO - codeparrot_training - Step 25435: {'lr': 0.0002592426840476803, 'samples': 13023232, 'steps': 25435, 'loss/train': 0.9522150754928589} 02/25/2022 10:32:37 - INFO - codeparrot_training - Step 25436: {'lr': 0.0002592263327523773, 'samples': 13023744, 'steps': 25436, 'loss/train': 2.873217821121216} 02/25/2022 10:32:43 - INFO - codeparrot_training - Step 25437: {'lr': 0.0002592099814175517, 'samples': 13024256, 'steps': 25437, 'loss/train': 1.4420586824417114} 02/25/2022 10:32:46 - INFO - codeparrot_training - Step 25438: {'lr': 0.00025919363004327337, 'samples': 13024768, 'steps': 25438, 'loss/train': 2.89656138420105} 02/25/2022 10:32:52 - INFO - codeparrot_training - Step 25439: {'lr': 0.0002591772786296125, 'samples': 13025280, 'steps': 25439, 'loss/train': 1.6785086393356323} 02/25/2022 10:32:55 - INFO - codeparrot_training - Step 25440: {'lr': 0.0002591609271766391, 'samples': 13025792, 'steps': 25440, 'loss/train': 1.2279216051101685} 02/25/2022 10:33:01 - INFO - codeparrot_training - Step 25441: {'lr': 0.0002591445756844232, 'samples': 13026304, 'steps': 25441, 'loss/train': 2.342471122741699} 02/25/2022 10:33:04 - INFO - codeparrot_training - Step 25442: {'lr': 0.0002591282241530348, 'samples': 13026816, 'steps': 25442, 'loss/train': 1.7308307886123657} 02/25/2022 10:33:10 - INFO - codeparrot_training - Step 25443: {'lr': 0.000259111872582544, 'samples': 13027328, 'steps': 25443, 'loss/train': 2.144907236099243} 02/25/2022 10:33:13 - INFO - codeparrot_training - Step 25444: {'lr': 0.0002590955209730208, 'samples': 13027840, 'steps': 25444, 'loss/train': 1.9072288274765015} 02/25/2022 10:33:19 - INFO - codeparrot_training - Step 25445: {'lr': 0.0002590791693245353, 'samples': 13028352, 'steps': 25445, 'loss/train': 1.843860149383545} 02/25/2022 10:33:22 - INFO - codeparrot_training - Step 25446: {'lr': 0.00025906281763715753, 'samples': 13028864, 'steps': 25446, 'loss/train': 2.8474838733673096} 02/25/2022 10:33:29 - INFO - codeparrot_training - Step 25447: {'lr': 0.0002590464659109575, 'samples': 13029376, 'steps': 25447, 'loss/train': 2.140130043029785} 02/25/2022 10:33:32 - INFO - codeparrot_training - Step 25448: {'lr': 0.00025903011414600536, 'samples': 13029888, 'steps': 25448, 'loss/train': 1.7169541120529175} 02/25/2022 10:33:38 - INFO - codeparrot_training - Step 25449: {'lr': 0.00025901376234237103, 'samples': 13030400, 'steps': 25449, 'loss/train': 2.234224796295166} 02/25/2022 10:33:41 - INFO - codeparrot_training - Step 25450: {'lr': 0.0002589974105001246, 'samples': 13030912, 'steps': 25450, 'loss/train': 2.063070297241211} 02/25/2022 10:33:47 - INFO - codeparrot_training - Step 25451: {'lr': 0.00025898105861933614, 'samples': 13031424, 'steps': 25451, 'loss/train': 1.272411823272705} 02/25/2022 10:33:50 - INFO - codeparrot_training - Step 25452: {'lr': 0.00025896470670007567, 'samples': 13031936, 'steps': 25452, 'loss/train': 2.5773468017578125} 02/25/2022 10:33:56 - INFO - codeparrot_training - Step 25453: {'lr': 0.0002589483547424132, 'samples': 13032448, 'steps': 25453, 'loss/train': 1.8566616773605347} 02/25/2022 10:33:59 - INFO - codeparrot_training - Step 25454: {'lr': 0.0002589320027464189, 'samples': 13032960, 'steps': 25454, 'loss/train': 0.7591987252235413} 02/25/2022 10:34:05 - INFO - codeparrot_training - Step 25455: {'lr': 0.00025891565071216267, 'samples': 13033472, 'steps': 25455, 'loss/train': 2.1075520515441895} 02/25/2022 10:34:08 - INFO - codeparrot_training - Step 25456: {'lr': 0.0002588992986397146, 'samples': 13033984, 'steps': 25456, 'loss/train': 1.6278977394104004} 02/25/2022 10:34:14 - INFO - codeparrot_training - Step 25457: {'lr': 0.0002588829465291448, 'samples': 13034496, 'steps': 25457, 'loss/train': 2.203155279159546} 02/25/2022 10:34:18 - INFO - codeparrot_training - Step 25458: {'lr': 0.0002588665943805233, 'samples': 13035008, 'steps': 25458, 'loss/train': 3.2736563682556152} 02/25/2022 10:34:23 - INFO - codeparrot_training - Step 25459: {'lr': 0.00025885024219392017, 'samples': 13035520, 'steps': 25459, 'loss/train': 1.5789657831192017} 02/25/2022 10:34:27 - INFO - codeparrot_training - Step 25460: {'lr': 0.00025883388996940533, 'samples': 13036032, 'steps': 25460, 'loss/train': 1.5168770551681519} 02/25/2022 10:34:32 - INFO - codeparrot_training - Step 25461: {'lr': 0.00025881753770704897, 'samples': 13036544, 'steps': 25461, 'loss/train': 1.689923882484436} 02/25/2022 10:34:36 - INFO - codeparrot_training - Step 25462: {'lr': 0.000258801185406921, 'samples': 13037056, 'steps': 25462, 'loss/train': 1.134519100189209} 02/25/2022 10:34:41 - INFO - codeparrot_training - Step 25463: {'lr': 0.0002587848330690916, 'samples': 13037568, 'steps': 25463, 'loss/train': 1.8639558553695679} 02/25/2022 10:34:45 - INFO - codeparrot_training - Step 25464: {'lr': 0.0002587684806936307, 'samples': 13038080, 'steps': 25464, 'loss/train': 1.2560575008392334} 02/25/2022 10:34:50 - INFO - codeparrot_training - Step 25465: {'lr': 0.0002587521282806085, 'samples': 13038592, 'steps': 25465, 'loss/train': 1.5340209007263184} 02/25/2022 10:34:54 - INFO - codeparrot_training - Step 25466: {'lr': 0.00025873577583009495, 'samples': 13039104, 'steps': 25466, 'loss/train': 1.7172353267669678} 02/25/2022 10:35:00 - INFO - codeparrot_training - Step 25467: {'lr': 0.0002587194233421601, 'samples': 13039616, 'steps': 25467, 'loss/train': 0.3781517446041107} 02/25/2022 10:35:04 - INFO - codeparrot_training - Step 25468: {'lr': 0.00025870307081687395, 'samples': 13040128, 'steps': 25468, 'loss/train': 1.8215484619140625} 02/25/2022 10:35:09 - INFO - codeparrot_training - Step 25469: {'lr': 0.0002586867182543066, 'samples': 13040640, 'steps': 25469, 'loss/train': 2.0719873905181885} 02/25/2022 10:35:13 - INFO - codeparrot_training - Step 25470: {'lr': 0.0002586703656545282, 'samples': 13041152, 'steps': 25470, 'loss/train': 2.7393150329589844} 02/25/2022 10:35:18 - INFO - codeparrot_training - Step 25471: {'lr': 0.00025865401301760867, 'samples': 13041664, 'steps': 25471, 'loss/train': 1.7830309867858887} 02/25/2022 10:35:22 - INFO - codeparrot_training - Step 25472: {'lr': 0.00025863766034361815, 'samples': 13042176, 'steps': 25472, 'loss/train': 1.3046330213546753} 02/25/2022 10:35:27 - INFO - codeparrot_training - Step 25473: {'lr': 0.00025862130763262646, 'samples': 13042688, 'steps': 25473, 'loss/train': 1.2730748653411865} 02/25/2022 10:35:31 - INFO - codeparrot_training - Step 25474: {'lr': 0.00025860495488470403, 'samples': 13043200, 'steps': 25474, 'loss/train': 3.5933492183685303} 02/25/2022 10:35:36 - INFO - codeparrot_training - Step 25475: {'lr': 0.00025858860209992057, 'samples': 13043712, 'steps': 25475, 'loss/train': 1.6586380004882812} 02/25/2022 10:35:40 - INFO - codeparrot_training - Step 25476: {'lr': 0.00025857224927834633, 'samples': 13044224, 'steps': 25476, 'loss/train': 1.2413740158081055} 02/25/2022 10:35:45 - INFO - codeparrot_training - Step 25477: {'lr': 0.0002585558964200513, 'samples': 13044736, 'steps': 25477, 'loss/train': 1.6203744411468506} 02/25/2022 10:35:49 - INFO - codeparrot_training - Step 25478: {'lr': 0.0002585395435251055, 'samples': 13045248, 'steps': 25478, 'loss/train': 2.581801652908325} 02/25/2022 10:35:55 - INFO - codeparrot_training - Step 25479: {'lr': 0.000258523190593579, 'samples': 13045760, 'steps': 25479, 'loss/train': 2.717923641204834} 02/25/2022 10:36:00 - INFO - codeparrot_training - Step 25480: {'lr': 0.0002585068376255418, 'samples': 13046272, 'steps': 25480, 'loss/train': 8.688414573669434} 02/25/2022 10:36:04 - INFO - codeparrot_training - Step 25481: {'lr': 0.00025849048462106414, 'samples': 13046784, 'steps': 25481, 'loss/train': 2.695458173751831} 02/25/2022 10:36:09 - INFO - codeparrot_training - Step 25482: {'lr': 0.00025847413158021587, 'samples': 13047296, 'steps': 25482, 'loss/train': 2.860206127166748} 02/25/2022 10:36:13 - INFO - codeparrot_training - Step 25483: {'lr': 0.00025845777850306716, 'samples': 13047808, 'steps': 25483, 'loss/train': 2.300733804702759} 02/25/2022 10:36:18 - INFO - codeparrot_training - Step 25484: {'lr': 0.000258441425389688, 'samples': 13048320, 'steps': 25484, 'loss/train': 2.0404443740844727} 02/25/2022 10:36:22 - INFO - codeparrot_training - Step 25485: {'lr': 0.0002584250722401484, 'samples': 13048832, 'steps': 25485, 'loss/train': 1.8768339157104492} 02/25/2022 10:36:27 - INFO - codeparrot_training - Step 25486: {'lr': 0.00025840871905451847, 'samples': 13049344, 'steps': 25486, 'loss/train': 2.9098217487335205} 02/25/2022 10:36:31 - INFO - codeparrot_training - Step 25487: {'lr': 0.0002583923658328683, 'samples': 13049856, 'steps': 25487, 'loss/train': 1.714423418045044} 02/25/2022 10:36:36 - INFO - codeparrot_training - Step 25488: {'lr': 0.00025837601257526786, 'samples': 13050368, 'steps': 25488, 'loss/train': 1.2151185274124146} 02/25/2022 10:36:40 - INFO - codeparrot_training - Step 25489: {'lr': 0.0002583596592817873, 'samples': 13050880, 'steps': 25489, 'loss/train': 2.2535359859466553} 02/25/2022 10:36:45 - INFO - codeparrot_training - Step 25490: {'lr': 0.00025834330595249654, 'samples': 13051392, 'steps': 25490, 'loss/train': 0.9353277087211609} 02/25/2022 10:36:49 - INFO - codeparrot_training - Step 25491: {'lr': 0.0002583269525874658, 'samples': 13051904, 'steps': 25491, 'loss/train': 1.742711067199707} 02/25/2022 10:36:55 - INFO - codeparrot_training - Step 25492: {'lr': 0.00025831059918676497, 'samples': 13052416, 'steps': 25492, 'loss/train': 1.9470714330673218} 02/25/2022 10:36:59 - INFO - codeparrot_training - Step 25493: {'lr': 0.00025829424575046414, 'samples': 13052928, 'steps': 25493, 'loss/train': 1.2629696130752563} 02/25/2022 10:37:04 - INFO - codeparrot_training - Step 25494: {'lr': 0.0002582778922786335, 'samples': 13053440, 'steps': 25494, 'loss/train': 0.8298726677894592} 02/25/2022 10:37:07 - INFO - codeparrot_training - Step 25495: {'lr': 0.0002582615387713429, 'samples': 13053952, 'steps': 25495, 'loss/train': 2.4662070274353027} 02/25/2022 10:37:13 - INFO - codeparrot_training - Step 25496: {'lr': 0.00025824518522866253, 'samples': 13054464, 'steps': 25496, 'loss/train': 2.401947498321533} 02/25/2022 10:37:16 - INFO - codeparrot_training - Step 25497: {'lr': 0.0002582288316506624, 'samples': 13054976, 'steps': 25497, 'loss/train': 1.985283374786377} 02/25/2022 10:37:22 - INFO - codeparrot_training - Step 25498: {'lr': 0.0002582124780374126, 'samples': 13055488, 'steps': 25498, 'loss/train': 1.9059538841247559} 02/25/2022 10:37:26 - INFO - codeparrot_training - Step 25499: {'lr': 0.00025819612438898314, 'samples': 13056000, 'steps': 25499, 'loss/train': 2.1930618286132812} 02/25/2022 10:37:31 - INFO - codeparrot_training - Step 25500: {'lr': 0.00025817977070544405, 'samples': 13056512, 'steps': 25500, 'loss/train': 2.1014113426208496} 02/25/2022 10:37:34 - INFO - codeparrot_training - Step 25501: {'lr': 0.00025816341698686545, 'samples': 13057024, 'steps': 25501, 'loss/train': 2.0710501670837402} 02/25/2022 10:37:41 - INFO - codeparrot_training - Step 25502: {'lr': 0.0002581470632333173, 'samples': 13057536, 'steps': 25502, 'loss/train': 1.6427345275878906} 02/25/2022 10:37:44 - INFO - codeparrot_training - Step 25503: {'lr': 0.00025813070944486985, 'samples': 13058048, 'steps': 25503, 'loss/train': 2.5381133556365967} 02/25/2022 10:37:50 - INFO - codeparrot_training - Step 25504: {'lr': 0.0002581143556215929, 'samples': 13058560, 'steps': 25504, 'loss/train': 1.2489573955535889} 02/25/2022 10:37:53 - INFO - codeparrot_training - Step 25505: {'lr': 0.0002580980017635567, 'samples': 13059072, 'steps': 25505, 'loss/train': 2.8317482471466064} 02/25/2022 10:37:59 - INFO - codeparrot_training - Step 25506: {'lr': 0.0002580816478708312, 'samples': 13059584, 'steps': 25506, 'loss/train': 3.079897403717041} 02/25/2022 10:38:02 - INFO - codeparrot_training - Step 25507: {'lr': 0.0002580652939434865, 'samples': 13060096, 'steps': 25507, 'loss/train': 2.1846425533294678} 02/25/2022 10:38:08 - INFO - codeparrot_training - Step 25508: {'lr': 0.0002580489399815926, 'samples': 13060608, 'steps': 25508, 'loss/train': 1.4465380907058716} 02/25/2022 10:38:11 - INFO - codeparrot_training - Step 25509: {'lr': 0.00025803258598521966, 'samples': 13061120, 'steps': 25509, 'loss/train': 2.062087059020996} 02/25/2022 10:38:17 - INFO - codeparrot_training - Step 25510: {'lr': 0.0002580162319544376, 'samples': 13061632, 'steps': 25510, 'loss/train': 0.6560931205749512} 02/25/2022 10:38:20 - INFO - codeparrot_training - Step 25511: {'lr': 0.00025799987788931666, 'samples': 13062144, 'steps': 25511, 'loss/train': 2.174177885055542} 02/25/2022 10:38:27 - INFO - codeparrot_training - Step 25512: {'lr': 0.0002579835237899267, 'samples': 13062656, 'steps': 25512, 'loss/train': 2.0699105262756348} 02/25/2022 10:38:30 - INFO - codeparrot_training - Step 25513: {'lr': 0.00025796716965633787, 'samples': 13063168, 'steps': 25513, 'loss/train': 1.134191870689392} 02/25/2022 10:38:36 - INFO - codeparrot_training - Step 25514: {'lr': 0.0002579508154886202, 'samples': 13063680, 'steps': 25514, 'loss/train': 2.9658265113830566} 02/25/2022 10:38:39 - INFO - codeparrot_training - Step 25515: {'lr': 0.00025793446128684377, 'samples': 13064192, 'steps': 25515, 'loss/train': 0.21962374448776245} 02/25/2022 10:38:45 - INFO - codeparrot_training - Step 25516: {'lr': 0.00025791810705107866, 'samples': 13064704, 'steps': 25516, 'loss/train': 0.7573524117469788} 02/25/2022 10:38:48 - INFO - codeparrot_training - Step 25517: {'lr': 0.00025790175278139483, 'samples': 13065216, 'steps': 25517, 'loss/train': 2.088284730911255} 02/25/2022 10:38:54 - INFO - codeparrot_training - Step 25518: {'lr': 0.0002578853984778625, 'samples': 13065728, 'steps': 25518, 'loss/train': 2.1003429889678955} 02/25/2022 10:38:58 - INFO - codeparrot_training - Step 25519: {'lr': 0.0002578690441405515, 'samples': 13066240, 'steps': 25519, 'loss/train': 2.253528356552124} 02/25/2022 10:39:01 - INFO - codeparrot_training - Step 25520: {'lr': 0.00025785268976953206, 'samples': 13066752, 'steps': 25520, 'loss/train': 1.3939143419265747} 02/25/2022 10:39:07 - INFO - codeparrot_training - Step 25521: {'lr': 0.0002578363353648742, 'samples': 13067264, 'steps': 25521, 'loss/train': 0.7397488355636597} 02/25/2022 10:39:11 - INFO - codeparrot_training - Step 25522: {'lr': 0.000257819980926648, 'samples': 13067776, 'steps': 25522, 'loss/train': 2.4867022037506104} 02/25/2022 10:39:16 - INFO - codeparrot_training - Step 25523: {'lr': 0.00025780362645492344, 'samples': 13068288, 'steps': 25523, 'loss/train': 2.077606678009033} 02/25/2022 10:39:20 - INFO - codeparrot_training - Step 25524: {'lr': 0.0002577872719497707, 'samples': 13068800, 'steps': 25524, 'loss/train': 2.550997257232666} 02/25/2022 10:39:26 - INFO - codeparrot_training - Step 25525: {'lr': 0.0002577709174112597, 'samples': 13069312, 'steps': 25525, 'loss/train': 1.8374128341674805} 02/25/2022 10:39:29 - INFO - codeparrot_training - Step 25526: {'lr': 0.0002577545628394606, 'samples': 13069824, 'steps': 25526, 'loss/train': 1.7259000539779663} 02/25/2022 10:39:35 - INFO - codeparrot_training - Step 25527: {'lr': 0.00025773820823444334, 'samples': 13070336, 'steps': 25527, 'loss/train': 2.8765740394592285} 02/25/2022 10:39:38 - INFO - codeparrot_training - Step 25528: {'lr': 0.0002577218535962781, 'samples': 13070848, 'steps': 25528, 'loss/train': 1.0596556663513184} 02/25/2022 10:39:44 - INFO - codeparrot_training - Step 25529: {'lr': 0.00025770549892503486, 'samples': 13071360, 'steps': 25529, 'loss/train': 1.5723588466644287} 02/25/2022 10:39:47 - INFO - codeparrot_training - Step 25530: {'lr': 0.0002576891442207837, 'samples': 13071872, 'steps': 25530, 'loss/train': 3.7963387966156006} 02/25/2022 10:39:53 - INFO - codeparrot_training - Step 25531: {'lr': 0.00025767278948359473, 'samples': 13072384, 'steps': 25531, 'loss/train': 1.7266310453414917} 02/25/2022 10:39:56 - INFO - codeparrot_training - Step 25532: {'lr': 0.00025765643471353794, 'samples': 13072896, 'steps': 25532, 'loss/train': 1.6487746238708496} 02/25/2022 10:40:02 - INFO - codeparrot_training - Step 25533: {'lr': 0.00025764007991068344, 'samples': 13073408, 'steps': 25533, 'loss/train': 1.7094881534576416} 02/25/2022 10:40:05 - INFO - codeparrot_training - Step 25534: {'lr': 0.0002576237250751012, 'samples': 13073920, 'steps': 25534, 'loss/train': 0.9303370714187622} 02/25/2022 10:40:11 - INFO - codeparrot_training - Step 25535: {'lr': 0.0002576073702068615, 'samples': 13074432, 'steps': 25535, 'loss/train': 1.7391544580459595} 02/25/2022 10:40:15 - INFO - codeparrot_training - Step 25536: {'lr': 0.00025759101530603405, 'samples': 13074944, 'steps': 25536, 'loss/train': 2.6711604595184326} 02/25/2022 10:40:20 - INFO - codeparrot_training - Step 25537: {'lr': 0.00025757466037268925, 'samples': 13075456, 'steps': 25537, 'loss/train': 1.4803251028060913} 02/25/2022 10:40:24 - INFO - codeparrot_training - Step 25538: {'lr': 0.0002575583054068969, 'samples': 13075968, 'steps': 25538, 'loss/train': 1.5431030988693237} 02/25/2022 10:40:29 - INFO - codeparrot_training - Step 25539: {'lr': 0.00025754195040872727, 'samples': 13076480, 'steps': 25539, 'loss/train': 2.096067190170288} 02/25/2022 10:40:33 - INFO - codeparrot_training - Step 25540: {'lr': 0.00025752559537825027, 'samples': 13076992, 'steps': 25540, 'loss/train': 2.3013758659362793} 02/25/2022 10:40:38 - INFO - codeparrot_training - Step 25541: {'lr': 0.00025750924031553603, 'samples': 13077504, 'steps': 25541, 'loss/train': 2.5413990020751953} 02/25/2022 10:40:42 - INFO - codeparrot_training - Step 25542: {'lr': 0.0002574928852206545, 'samples': 13078016, 'steps': 25542, 'loss/train': 3.4210944175720215} 02/25/2022 10:40:48 - INFO - codeparrot_training - Step 25543: {'lr': 0.00025747653009367596, 'samples': 13078528, 'steps': 25543, 'loss/train': 2.311305284500122} 02/25/2022 10:40:51 - INFO - codeparrot_training - Step 25544: {'lr': 0.0002574601749346702, 'samples': 13079040, 'steps': 25544, 'loss/train': 2.3376216888427734} 02/25/2022 10:40:57 - INFO - codeparrot_training - Step 25545: {'lr': 0.0002574438197437076, 'samples': 13079552, 'steps': 25545, 'loss/train': 2.332993268966675} 02/25/2022 10:41:00 - INFO - codeparrot_training - Step 25546: {'lr': 0.00025742746452085794, 'samples': 13080064, 'steps': 25546, 'loss/train': 0.3118111491203308} 02/25/2022 10:41:06 - INFO - codeparrot_training - Step 25547: {'lr': 0.00025741110926619134, 'samples': 13080576, 'steps': 25547, 'loss/train': 2.3225464820861816} 02/25/2022 10:41:09 - INFO - codeparrot_training - Step 25548: {'lr': 0.00025739475397977795, 'samples': 13081088, 'steps': 25548, 'loss/train': 2.4193713665008545} 02/25/2022 10:41:15 - INFO - codeparrot_training - Step 25549: {'lr': 0.0002573783986616877, 'samples': 13081600, 'steps': 25549, 'loss/train': 1.776007056236267} 02/25/2022 10:41:18 - INFO - codeparrot_training - Step 25550: {'lr': 0.00025736204331199084, 'samples': 13082112, 'steps': 25550, 'loss/train': 2.2091004848480225} 02/25/2022 10:41:24 - INFO - codeparrot_training - Step 25551: {'lr': 0.00025734568793075725, 'samples': 13082624, 'steps': 25551, 'loss/train': 1.9936878681182861} 02/25/2022 10:41:28 - INFO - codeparrot_training - Step 25552: {'lr': 0.00025732933251805713, 'samples': 13083136, 'steps': 25552, 'loss/train': 2.126516342163086} 02/25/2022 10:41:33 - INFO - codeparrot_training - Step 25553: {'lr': 0.00025731297707396047, 'samples': 13083648, 'steps': 25553, 'loss/train': 1.8014073371887207} 02/25/2022 10:41:37 - INFO - codeparrot_training - Step 25554: {'lr': 0.00025729662159853725, 'samples': 13084160, 'steps': 25554, 'loss/train': 2.103489398956299} 02/25/2022 10:41:42 - INFO - codeparrot_training - Step 25555: {'lr': 0.0002572802660918577, 'samples': 13084672, 'steps': 25555, 'loss/train': 1.1406726837158203} 02/25/2022 10:41:46 - INFO - codeparrot_training - Step 25556: {'lr': 0.0002572639105539918, 'samples': 13085184, 'steps': 25556, 'loss/train': 1.2123444080352783} 02/25/2022 10:41:51 - INFO - codeparrot_training - Step 25557: {'lr': 0.0002572475549850096, 'samples': 13085696, 'steps': 25557, 'loss/train': 1.4706628322601318} 02/25/2022 10:41:55 - INFO - codeparrot_training - Step 25558: {'lr': 0.00025723119938498115, 'samples': 13086208, 'steps': 25558, 'loss/train': 1.7352365255355835} 02/25/2022 10:42:00 - INFO - codeparrot_training - Step 25559: {'lr': 0.0002572148437539766, 'samples': 13086720, 'steps': 25559, 'loss/train': 0.4805527329444885} 02/25/2022 10:42:04 - INFO - codeparrot_training - Step 25560: {'lr': 0.0002571984880920659, 'samples': 13087232, 'steps': 25560, 'loss/train': 2.853137493133545} 02/25/2022 10:42:10 - INFO - codeparrot_training - Step 25561: {'lr': 0.0002571821323993192, 'samples': 13087744, 'steps': 25561, 'loss/train': 2.0273165702819824} 02/25/2022 10:42:13 - INFO - codeparrot_training - Step 25562: {'lr': 0.00025716577667580647, 'samples': 13088256, 'steps': 25562, 'loss/train': 1.693200945854187} 02/25/2022 10:42:19 - INFO - codeparrot_training - Step 25563: {'lr': 0.0002571494209215979, 'samples': 13088768, 'steps': 25563, 'loss/train': 1.1780664920806885} 02/25/2022 10:42:22 - INFO - codeparrot_training - Step 25564: {'lr': 0.0002571330651367634, 'samples': 13089280, 'steps': 25564, 'loss/train': 2.799821376800537} 02/25/2022 10:42:28 - INFO - codeparrot_training - Step 25565: {'lr': 0.0002571167093213731, 'samples': 13089792, 'steps': 25565, 'loss/train': 1.9448604583740234} 02/25/2022 10:42:31 - INFO - codeparrot_training - Step 25566: {'lr': 0.0002571003534754972, 'samples': 13090304, 'steps': 25566, 'loss/train': 0.9562610387802124} 02/25/2022 10:42:37 - INFO - codeparrot_training - Step 25567: {'lr': 0.0002570839975992055, 'samples': 13090816, 'steps': 25567, 'loss/train': 2.371833562850952} 02/25/2022 10:42:40 - INFO - codeparrot_training - Step 25568: {'lr': 0.00025706764169256837, 'samples': 13091328, 'steps': 25568, 'loss/train': 2.011049270629883} 02/25/2022 10:42:46 - INFO - codeparrot_training - Step 25569: {'lr': 0.0002570512857556556, 'samples': 13091840, 'steps': 25569, 'loss/train': 0.7666170597076416} 02/25/2022 10:42:49 - INFO - codeparrot_training - Step 25570: {'lr': 0.0002570349297885374, 'samples': 13092352, 'steps': 25570, 'loss/train': 1.8144721984863281} 02/25/2022 10:42:56 - INFO - codeparrot_training - Step 25571: {'lr': 0.00025701857379128366, 'samples': 13092864, 'steps': 25571, 'loss/train': 1.5858922004699707} 02/25/2022 10:42:59 - INFO - codeparrot_training - Step 25572: {'lr': 0.0002570022177639648, 'samples': 13093376, 'steps': 25572, 'loss/train': 2.163553476333618} 02/25/2022 10:43:05 - INFO - codeparrot_training - Step 25573: {'lr': 0.0002569858617066505, 'samples': 13093888, 'steps': 25573, 'loss/train': 2.0017151832580566} 02/25/2022 10:43:08 - INFO - codeparrot_training - Step 25574: {'lr': 0.00025696950561941104, 'samples': 13094400, 'steps': 25574, 'loss/train': 2.418836832046509} 02/25/2022 10:43:14 - INFO - codeparrot_training - Step 25575: {'lr': 0.00025695314950231643, 'samples': 13094912, 'steps': 25575, 'loss/train': 2.1170639991760254} 02/25/2022 10:43:17 - INFO - codeparrot_training - Step 25576: {'lr': 0.0002569367933554367, 'samples': 13095424, 'steps': 25576, 'loss/train': 1.3791933059692383} 02/25/2022 10:43:23 - INFO - codeparrot_training - Step 25577: {'lr': 0.000256920437178842, 'samples': 13095936, 'steps': 25577, 'loss/train': 2.83950138092041} 02/25/2022 10:43:26 - INFO - codeparrot_training - Step 25578: {'lr': 0.00025690408097260234, 'samples': 13096448, 'steps': 25578, 'loss/train': 1.680790901184082} 02/25/2022 10:43:32 - INFO - codeparrot_training - Step 25579: {'lr': 0.0002568877247367878, 'samples': 13096960, 'steps': 25579, 'loss/train': 2.302349090576172} 02/25/2022 10:43:35 - INFO - codeparrot_training - Step 25580: {'lr': 0.0002568713684714684, 'samples': 13097472, 'steps': 25580, 'loss/train': 2.396230459213257} 02/25/2022 10:43:41 - INFO - codeparrot_training - Step 25581: {'lr': 0.00025685501217671423, 'samples': 13097984, 'steps': 25581, 'loss/train': 2.152153253555298} 02/25/2022 10:43:44 - INFO - codeparrot_training - Step 25582: {'lr': 0.00025683865585259533, 'samples': 13098496, 'steps': 25582, 'loss/train': 2.87414288520813} 02/25/2022 10:43:50 - INFO - codeparrot_training - Step 25583: {'lr': 0.0002568222994991819, 'samples': 13099008, 'steps': 25583, 'loss/train': 0.6613590121269226} 02/25/2022 10:43:53 - INFO - codeparrot_training - Step 25584: {'lr': 0.0002568059431165438, 'samples': 13099520, 'steps': 25584, 'loss/train': 0.7476187348365784} 02/25/2022 10:43:59 - INFO - codeparrot_training - Step 25585: {'lr': 0.00025678958670475135, 'samples': 13100032, 'steps': 25585, 'loss/train': 1.3253560066223145} 02/25/2022 10:44:02 - INFO - codeparrot_training - Step 25586: {'lr': 0.0002567732302638744, 'samples': 13100544, 'steps': 25586, 'loss/train': 1.8754878044128418} 02/25/2022 10:44:09 - INFO - codeparrot_training - Step 25587: {'lr': 0.00025675687379398304, 'samples': 13101056, 'steps': 25587, 'loss/train': 0.08983694016933441} 02/25/2022 10:44:12 - INFO - codeparrot_training - Step 25588: {'lr': 0.0002567405172951474, 'samples': 13101568, 'steps': 25588, 'loss/train': 1.6174893379211426} 02/25/2022 10:44:18 - INFO - codeparrot_training - Step 25589: {'lr': 0.0002567241607674375, 'samples': 13102080, 'steps': 25589, 'loss/train': 0.8111506104469299} 02/25/2022 10:44:21 - INFO - codeparrot_training - Step 25590: {'lr': 0.0002567078042109235, 'samples': 13102592, 'steps': 25590, 'loss/train': 1.9698349237442017} 02/25/2022 10:44:27 - INFO - codeparrot_training - Step 25591: {'lr': 0.0002566914476256753, 'samples': 13103104, 'steps': 25591, 'loss/train': 0.5227755308151245} 02/25/2022 10:44:30 - INFO - codeparrot_training - Step 25592: {'lr': 0.00025667509101176317, 'samples': 13103616, 'steps': 25592, 'loss/train': 1.7507622241973877} 02/25/2022 10:44:36 - INFO - codeparrot_training - Step 25593: {'lr': 0.00025665873436925697, 'samples': 13104128, 'steps': 25593, 'loss/train': 2.006484270095825} 02/25/2022 10:44:39 - INFO - codeparrot_training - Step 25594: {'lr': 0.000256642377698227, 'samples': 13104640, 'steps': 25594, 'loss/train': 1.7868516445159912} 02/25/2022 10:44:45 - INFO - codeparrot_training - Step 25595: {'lr': 0.000256626020998743, 'samples': 13105152, 'steps': 25595, 'loss/train': 2.7415452003479004} 02/25/2022 10:44:48 - INFO - codeparrot_training - Step 25596: {'lr': 0.0002566096642708754, 'samples': 13105664, 'steps': 25596, 'loss/train': 2.0796501636505127} 02/25/2022 10:44:55 - INFO - codeparrot_training - Step 25597: {'lr': 0.00025659330751469394, 'samples': 13106176, 'steps': 25597, 'loss/train': 2.1918537616729736} 02/25/2022 10:44:58 - INFO - codeparrot_training - Step 25598: {'lr': 0.000256576950730269, 'samples': 13106688, 'steps': 25598, 'loss/train': 1.1648391485214233} 02/25/2022 10:45:04 - INFO - codeparrot_training - Step 25599: {'lr': 0.0002565605939176704, 'samples': 13107200, 'steps': 25599, 'loss/train': 1.7362256050109863} 02/25/2022 10:45:07 - INFO - codeparrot_training - Step 25600: {'lr': 0.00025654423707696834, 'samples': 13107712, 'steps': 25600, 'loss/train': 2.187987804412842} 02/25/2022 10:45:13 - INFO - codeparrot_training - Step 25601: {'lr': 0.0002565278802082328, 'samples': 13108224, 'steps': 25601, 'loss/train': 1.9063791036605835} 02/25/2022 10:45:16 - INFO - codeparrot_training - Step 25602: {'lr': 0.00025651152331153393, 'samples': 13108736, 'steps': 25602, 'loss/train': 1.6950427293777466} 02/25/2022 10:45:22 - INFO - codeparrot_training - Step 25603: {'lr': 0.00025649516638694174, 'samples': 13109248, 'steps': 25603, 'loss/train': 2.131453037261963} 02/25/2022 10:45:25 - INFO - codeparrot_training - Step 25604: {'lr': 0.00025647880943452633, 'samples': 13109760, 'steps': 25604, 'loss/train': 2.075115919113159} 02/25/2022 10:45:32 - INFO - codeparrot_training - Step 25605: {'lr': 0.0002564624524543578, 'samples': 13110272, 'steps': 25605, 'loss/train': 1.14596426486969} 02/25/2022 10:45:35 - INFO - codeparrot_training - Step 25606: {'lr': 0.0002564460954465061, 'samples': 13110784, 'steps': 25606, 'loss/train': 3.5504369735717773} 02/25/2022 10:45:41 - INFO - codeparrot_training - Step 25607: {'lr': 0.0002564297384110414, 'samples': 13111296, 'steps': 25607, 'loss/train': 2.123778820037842} 02/25/2022 10:45:45 - INFO - codeparrot_training - Step 25608: {'lr': 0.00025641338134803376, 'samples': 13111808, 'steps': 25608, 'loss/train': 1.2274852991104126} 02/25/2022 10:45:50 - INFO - codeparrot_training - Step 25609: {'lr': 0.0002563970242575533, 'samples': 13112320, 'steps': 25609, 'loss/train': 1.6785922050476074} 02/25/2022 10:45:54 - INFO - codeparrot_training - Step 25610: {'lr': 0.0002563806671396699, 'samples': 13112832, 'steps': 25610, 'loss/train': 1.8916770219802856} 02/25/2022 10:45:59 - INFO - codeparrot_training - Step 25611: {'lr': 0.0002563643099944538, 'samples': 13113344, 'steps': 25611, 'loss/train': 1.98416006565094} 02/25/2022 10:46:03 - INFO - codeparrot_training - Step 25612: {'lr': 0.000256347952821975, 'samples': 13113856, 'steps': 25612, 'loss/train': 0.3088620603084564} 02/25/2022 10:46:08 - INFO - codeparrot_training - Step 25613: {'lr': 0.0002563315956223036, 'samples': 13114368, 'steps': 25613, 'loss/train': 1.8856010437011719} 02/25/2022 10:46:12 - INFO - codeparrot_training - Step 25614: {'lr': 0.0002563152383955096, 'samples': 13114880, 'steps': 25614, 'loss/train': 2.6615824699401855} 02/25/2022 10:46:17 - INFO - codeparrot_training - Step 25615: {'lr': 0.0002562988811416632, 'samples': 13115392, 'steps': 25615, 'loss/train': 1.9493383169174194} 02/25/2022 10:46:21 - INFO - codeparrot_training - Step 25616: {'lr': 0.0002562825238608344, 'samples': 13115904, 'steps': 25616, 'loss/train': 1.2353770732879639} 02/25/2022 10:46:27 - INFO - codeparrot_training - Step 25617: {'lr': 0.0002562661665530932, 'samples': 13116416, 'steps': 25617, 'loss/train': 1.5191400051116943} 02/25/2022 10:46:30 - INFO - codeparrot_training - Step 25618: {'lr': 0.00025624980921850975, 'samples': 13116928, 'steps': 25618, 'loss/train': 1.785825490951538} 02/25/2022 10:46:36 - INFO - codeparrot_training - Step 25619: {'lr': 0.00025623345185715415, 'samples': 13117440, 'steps': 25619, 'loss/train': 1.1005932092666626} 02/25/2022 10:46:39 - INFO - codeparrot_training - Step 25620: {'lr': 0.00025621709446909644, 'samples': 13117952, 'steps': 25620, 'loss/train': 2.3417887687683105} 02/25/2022 10:46:45 - INFO - codeparrot_training - Step 25621: {'lr': 0.00025620073705440654, 'samples': 13118464, 'steps': 25621, 'loss/train': 2.985455274581909} 02/25/2022 10:46:48 - INFO - codeparrot_training - Step 25622: {'lr': 0.0002561843796131548, 'samples': 13118976, 'steps': 25622, 'loss/train': 2.6752543449401855} 02/25/2022 10:46:54 - INFO - codeparrot_training - Step 25623: {'lr': 0.000256168022145411, 'samples': 13119488, 'steps': 25623, 'loss/train': 1.1425602436065674} 02/25/2022 10:46:58 - INFO - codeparrot_training - Step 25624: {'lr': 0.0002561516646512454, 'samples': 13120000, 'steps': 25624, 'loss/train': 1.9599618911743164} 02/25/2022 10:47:03 - INFO - codeparrot_training - Step 25625: {'lr': 0.0002561353071307281, 'samples': 13120512, 'steps': 25625, 'loss/train': 1.5613296031951904} 02/25/2022 10:47:07 - INFO - codeparrot_training - Step 25626: {'lr': 0.000256118949583929, 'samples': 13121024, 'steps': 25626, 'loss/train': 1.5623118877410889} 02/25/2022 10:47:12 - INFO - codeparrot_training - Step 25627: {'lr': 0.0002561025920109183, 'samples': 13121536, 'steps': 25627, 'loss/train': 2.1120214462280273} 02/25/2022 10:47:16 - INFO - codeparrot_training - Step 25628: {'lr': 0.0002560862344117661, 'samples': 13122048, 'steps': 25628, 'loss/train': 2.0736422538757324} 02/25/2022 10:47:21 - INFO - codeparrot_training - Step 25629: {'lr': 0.0002560698767865423, 'samples': 13122560, 'steps': 25629, 'loss/train': 2.025491237640381} 02/25/2022 10:47:25 - INFO - codeparrot_training - Step 25630: {'lr': 0.0002560535191353171, 'samples': 13123072, 'steps': 25630, 'loss/train': 3.3473308086395264} 02/25/2022 10:47:30 - INFO - codeparrot_training - Step 25631: {'lr': 0.00025603716145816056, 'samples': 13123584, 'steps': 25631, 'loss/train': 1.3574202060699463} 02/25/2022 10:47:34 - INFO - codeparrot_training - Step 25632: {'lr': 0.00025602080375514277, 'samples': 13124096, 'steps': 25632, 'loss/train': 0.2443811148405075} 02/25/2022 10:47:40 - INFO - codeparrot_training - Step 25633: {'lr': 0.0002560044460263337, 'samples': 13124608, 'steps': 25633, 'loss/train': 1.4887213706970215} 02/25/2022 10:47:44 - INFO - codeparrot_training - Step 25634: {'lr': 0.0002559880882718035, 'samples': 13125120, 'steps': 25634, 'loss/train': 1.1063514947891235} 02/25/2022 10:47:49 - INFO - codeparrot_training - Step 25635: {'lr': 0.00025597173049162234, 'samples': 13125632, 'steps': 25635, 'loss/train': 1.3391457796096802} 02/25/2022 10:47:52 - INFO - codeparrot_training - Step 25636: {'lr': 0.0002559553726858601, 'samples': 13126144, 'steps': 25636, 'loss/train': 2.2346572875976562} 02/25/2022 10:47:58 - INFO - codeparrot_training - Step 25637: {'lr': 0.0002559390148545869, 'samples': 13126656, 'steps': 25637, 'loss/train': 1.7385978698730469} 02/25/2022 10:48:01 - INFO - codeparrot_training - Step 25638: {'lr': 0.0002559226569978729, 'samples': 13127168, 'steps': 25638, 'loss/train': 1.7197811603546143} 02/25/2022 10:48:07 - INFO - codeparrot_training - Step 25639: {'lr': 0.00025590629911578805, 'samples': 13127680, 'steps': 25639, 'loss/train': 0.9838075041770935} 02/25/2022 10:48:10 - INFO - codeparrot_training - Step 25640: {'lr': 0.0002558899412084026, 'samples': 13128192, 'steps': 25640, 'loss/train': 0.9118172526359558} 02/25/2022 10:48:16 - INFO - codeparrot_training - Step 25641: {'lr': 0.0002558735832757864, 'samples': 13128704, 'steps': 25641, 'loss/train': 0.9940716028213501} 02/25/2022 10:48:19 - INFO - codeparrot_training - Step 25642: {'lr': 0.0002558572253180097, 'samples': 13129216, 'steps': 25642, 'loss/train': 1.4485918283462524} 02/25/2022 10:48:25 - INFO - codeparrot_training - Step 25643: {'lr': 0.0002558408673351425, 'samples': 13129728, 'steps': 25643, 'loss/train': 2.0387818813323975} 02/25/2022 10:48:29 - INFO - codeparrot_training - Step 25644: {'lr': 0.0002558245093272549, 'samples': 13130240, 'steps': 25644, 'loss/train': 1.5015616416931152} 02/25/2022 10:48:35 - INFO - codeparrot_training - Step 25645: {'lr': 0.0002558081512944169, 'samples': 13130752, 'steps': 25645, 'loss/train': 1.6553453207015991} 02/25/2022 10:48:38 - INFO - codeparrot_training - Step 25646: {'lr': 0.00025579179323669866, 'samples': 13131264, 'steps': 25646, 'loss/train': 1.626403570175171} 02/25/2022 10:48:42 - INFO - codeparrot_training - Step 25647: {'lr': 0.00025577543515417015, 'samples': 13131776, 'steps': 25647, 'loss/train': 1.8631216287612915} 02/25/2022 10:48:47 - INFO - codeparrot_training - Step 25648: {'lr': 0.0002557590770469016, 'samples': 13132288, 'steps': 25648, 'loss/train': 1.6985031366348267} 02/25/2022 10:48:51 - INFO - codeparrot_training - Step 25649: {'lr': 0.0002557427189149629, 'samples': 13132800, 'steps': 25649, 'loss/train': 2.2670018672943115} 02/25/2022 10:48:56 - INFO - codeparrot_training - Step 25650: {'lr': 0.0002557263607584243, 'samples': 13133312, 'steps': 25650, 'loss/train': 1.4436798095703125} 02/25/2022 10:49:00 - INFO - codeparrot_training - Step 25651: {'lr': 0.0002557100025773558, 'samples': 13133824, 'steps': 25651, 'loss/train': 2.4478187561035156} 02/25/2022 10:49:05 - INFO - codeparrot_training - Step 25652: {'lr': 0.00025569364437182736, 'samples': 13134336, 'steps': 25652, 'loss/train': 1.841557502746582} 02/25/2022 10:49:09 - INFO - codeparrot_training - Step 25653: {'lr': 0.00025567728614190926, 'samples': 13134848, 'steps': 25653, 'loss/train': 1.109722375869751} 02/25/2022 10:49:15 - INFO - codeparrot_training - Step 25654: {'lr': 0.0002556609278876714, 'samples': 13135360, 'steps': 25654, 'loss/train': 1.647121548652649} 02/25/2022 10:49:19 - INFO - codeparrot_training - Step 25655: {'lr': 0.00025564456960918395, 'samples': 13135872, 'steps': 25655, 'loss/train': 1.0943329334259033} 02/25/2022 10:49:24 - INFO - codeparrot_training - Step 25656: {'lr': 0.0002556282113065169, 'samples': 13136384, 'steps': 25656, 'loss/train': 1.7512942552566528} 02/25/2022 10:49:30 - INFO - codeparrot_training - Step 25657: {'lr': 0.0002556118529797405, 'samples': 13136896, 'steps': 25657, 'loss/train': 0.0709896832704544} 02/25/2022 10:49:33 - INFO - codeparrot_training - Step 25658: {'lr': 0.00025559549462892463, 'samples': 13137408, 'steps': 25658, 'loss/train': 1.7430973052978516} 02/25/2022 10:49:39 - INFO - codeparrot_training - Step 25659: {'lr': 0.00025557913625413945, 'samples': 13137920, 'steps': 25659, 'loss/train': 1.395140528678894} 02/25/2022 10:49:42 - INFO - codeparrot_training - Step 25660: {'lr': 0.000255562777855455, 'samples': 13138432, 'steps': 25660, 'loss/train': 3.4155142307281494} 02/25/2022 10:49:48 - INFO - codeparrot_training - Step 25661: {'lr': 0.00025554641943294145, 'samples': 13138944, 'steps': 25661, 'loss/train': 1.91545832157135} 02/25/2022 10:49:51 - INFO - codeparrot_training - Step 25662: {'lr': 0.0002555300609866687, 'samples': 13139456, 'steps': 25662, 'loss/train': 1.1756900548934937} 02/25/2022 10:49:57 - INFO - codeparrot_training - Step 25663: {'lr': 0.00025551370251670694, 'samples': 13139968, 'steps': 25663, 'loss/train': 1.3493777513504028} 02/25/2022 10:50:01 - INFO - codeparrot_training - Step 25664: {'lr': 0.0002554973440231263, 'samples': 13140480, 'steps': 25664, 'loss/train': 2.425305128097534} 02/25/2022 10:50:07 - INFO - codeparrot_training - Step 25665: {'lr': 0.0002554809855059967, 'samples': 13140992, 'steps': 25665, 'loss/train': 1.7686724662780762} 02/25/2022 10:50:10 - INFO - codeparrot_training - Step 25666: {'lr': 0.00025546462696538836, 'samples': 13141504, 'steps': 25666, 'loss/train': 3.03287672996521} 02/25/2022 10:50:15 - INFO - codeparrot_training - Step 25667: {'lr': 0.00025544826840137125, 'samples': 13142016, 'steps': 25667, 'loss/train': 2.772830009460449} 02/25/2022 10:50:19 - INFO - codeparrot_training - Step 25668: {'lr': 0.00025543190981401555, 'samples': 13142528, 'steps': 25668, 'loss/train': 0.9455603361129761} 02/25/2022 10:50:24 - INFO - codeparrot_training - Step 25669: {'lr': 0.0002554155512033912, 'samples': 13143040, 'steps': 25669, 'loss/train': 1.3380768299102783} 02/25/2022 10:50:28 - INFO - codeparrot_training - Step 25670: {'lr': 0.00025539919256956843, 'samples': 13143552, 'steps': 25670, 'loss/train': 1.609128713607788} 02/25/2022 10:50:34 - INFO - codeparrot_training - Step 25671: {'lr': 0.00025538283391261714, 'samples': 13144064, 'steps': 25671, 'loss/train': 1.4452869892120361} 02/25/2022 10:50:37 - INFO - codeparrot_training - Step 25672: {'lr': 0.0002553664752326076, 'samples': 13144576, 'steps': 25672, 'loss/train': 2.1352651119232178} 02/25/2022 10:50:42 - INFO - codeparrot_training - Step 25673: {'lr': 0.0002553501165296097, 'samples': 13145088, 'steps': 25673, 'loss/train': 2.2420411109924316} 02/25/2022 10:50:46 - INFO - codeparrot_training - Step 25674: {'lr': 0.00025533375780369366, 'samples': 13145600, 'steps': 25674, 'loss/train': 1.3607232570648193} 02/25/2022 10:50:52 - INFO - codeparrot_training - Step 25675: {'lr': 0.00025531739905492953, 'samples': 13146112, 'steps': 25675, 'loss/train': 1.0559757947921753} 02/25/2022 10:50:55 - INFO - codeparrot_training - Step 25676: {'lr': 0.0002553010402833872, 'samples': 13146624, 'steps': 25676, 'loss/train': 0.0966961681842804} 02/25/2022 10:51:01 - INFO - codeparrot_training - Step 25677: {'lr': 0.00025528468148913703, 'samples': 13147136, 'steps': 25677, 'loss/train': 2.1042630672454834} 02/25/2022 10:51:04 - INFO - codeparrot_training - Step 25678: {'lr': 0.00025526832267224885, 'samples': 13147648, 'steps': 25678, 'loss/train': 1.3894081115722656} 02/25/2022 10:51:10 - INFO - codeparrot_training - Step 25679: {'lr': 0.000255251963832793, 'samples': 13148160, 'steps': 25679, 'loss/train': 2.479046106338501} 02/25/2022 10:51:13 - INFO - codeparrot_training - Step 25680: {'lr': 0.00025523560497083924, 'samples': 13148672, 'steps': 25680, 'loss/train': 1.7690578699111938} 02/25/2022 10:51:20 - INFO - codeparrot_training - Step 25681: {'lr': 0.00025521924608645796, 'samples': 13149184, 'steps': 25681, 'loss/train': 2.472615957260132} 02/25/2022 10:51:23 - INFO - codeparrot_training - Step 25682: {'lr': 0.00025520288717971896, 'samples': 13149696, 'steps': 25682, 'loss/train': 1.6594858169555664} 02/25/2022 10:51:29 - INFO - codeparrot_training - Step 25683: {'lr': 0.00025518652825069256, 'samples': 13150208, 'steps': 25683, 'loss/train': 2.097919464111328} 02/25/2022 10:51:32 - INFO - codeparrot_training - Step 25684: {'lr': 0.00025517016929944863, 'samples': 13150720, 'steps': 25684, 'loss/train': 1.5367844104766846} 02/25/2022 10:51:39 - INFO - codeparrot_training - Step 25685: {'lr': 0.0002551538103260574, 'samples': 13151232, 'steps': 25685, 'loss/train': 1.544798493385315} 02/25/2022 10:51:42 - INFO - codeparrot_training - Step 25686: {'lr': 0.0002551374513305888, 'samples': 13151744, 'steps': 25686, 'loss/train': 1.5396851301193237} 02/25/2022 10:51:45 - INFO - codeparrot_training - Step 25687: {'lr': 0.000255121092313113, 'samples': 13152256, 'steps': 25687, 'loss/train': 0.9987378716468811} 02/25/2022 10:51:51 - INFO - codeparrot_training - Step 25688: {'lr': 0.00025510473327370014, 'samples': 13152768, 'steps': 25688, 'loss/train': 1.808637022972107} 02/25/2022 10:51:55 - INFO - codeparrot_training - Step 25689: {'lr': 0.00025508837421242016, 'samples': 13153280, 'steps': 25689, 'loss/train': 0.5900411605834961} 02/25/2022 10:52:01 - INFO - codeparrot_training - Step 25690: {'lr': 0.00025507201512934327, 'samples': 13153792, 'steps': 25690, 'loss/train': 1.537652611732483} 02/25/2022 10:52:05 - INFO - codeparrot_training - Step 25691: {'lr': 0.0002550556560245394, 'samples': 13154304, 'steps': 25691, 'loss/train': 1.1949650049209595} 02/25/2022 10:52:10 - INFO - codeparrot_training - Step 25692: {'lr': 0.0002550392968980788, 'samples': 13154816, 'steps': 25692, 'loss/train': 1.3613845109939575} 02/25/2022 10:52:14 - INFO - codeparrot_training - Step 25693: {'lr': 0.0002550229377500313, 'samples': 13155328, 'steps': 25693, 'loss/train': 2.4665143489837646} 02/25/2022 10:52:20 - INFO - codeparrot_training - Step 25694: {'lr': 0.00025500657858046734, 'samples': 13155840, 'steps': 25694, 'loss/train': 2.102137804031372} 02/25/2022 10:52:23 - INFO - codeparrot_training - Step 25695: {'lr': 0.0002549902193894566, 'samples': 13156352, 'steps': 25695, 'loss/train': 1.3893893957138062} 02/25/2022 10:52:29 - INFO - codeparrot_training - Step 25696: {'lr': 0.00025497386017706947, 'samples': 13156864, 'steps': 25696, 'loss/train': 1.0654795169830322} 02/25/2022 10:52:32 - INFO - codeparrot_training - Step 25697: {'lr': 0.00025495750094337586, 'samples': 13157376, 'steps': 25697, 'loss/train': 2.170686721801758} 02/25/2022 10:52:38 - INFO - codeparrot_training - Step 25698: {'lr': 0.00025494114168844576, 'samples': 13157888, 'steps': 25698, 'loss/train': 2.075395345687866} 02/25/2022 10:52:41 - INFO - codeparrot_training - Step 25699: {'lr': 0.00025492478241234955, 'samples': 13158400, 'steps': 25699, 'loss/train': 2.4617505073547363} 02/25/2022 10:52:47 - INFO - codeparrot_training - Step 25700: {'lr': 0.00025490842311515704, 'samples': 13158912, 'steps': 25700, 'loss/train': 2.340740203857422} 02/25/2022 10:52:50 - INFO - codeparrot_training - Step 25701: {'lr': 0.00025489206379693854, 'samples': 13159424, 'steps': 25701, 'loss/train': 1.5465506315231323} 02/25/2022 10:52:57 - INFO - codeparrot_training - Step 25702: {'lr': 0.0002548757044577638, 'samples': 13159936, 'steps': 25702, 'loss/train': 1.6311246156692505} 02/25/2022 10:53:00 - INFO - codeparrot_training - Step 25703: {'lr': 0.00025485934509770326, 'samples': 13160448, 'steps': 25703, 'loss/train': 2.338343620300293} 02/25/2022 10:53:05 - INFO - codeparrot_training - Step 25704: {'lr': 0.00025484298571682676, 'samples': 13160960, 'steps': 25704, 'loss/train': 1.9064362049102783} 02/25/2022 10:53:09 - INFO - codeparrot_training - Step 25705: {'lr': 0.00025482662631520444, 'samples': 13161472, 'steps': 25705, 'loss/train': 1.0886472463607788} 02/25/2022 10:53:15 - INFO - codeparrot_training - Step 25706: {'lr': 0.0002548102668929064, 'samples': 13161984, 'steps': 25706, 'loss/train': 1.0186115503311157} 02/25/2022 10:53:18 - INFO - codeparrot_training - Step 25707: {'lr': 0.0002547939074500027, 'samples': 13162496, 'steps': 25707, 'loss/train': 2.081252098083496} 02/25/2022 10:53:24 - INFO - codeparrot_training - Step 25708: {'lr': 0.0002547775479865634, 'samples': 13163008, 'steps': 25708, 'loss/train': 2.5716304779052734} 02/25/2022 10:53:27 - INFO - codeparrot_training - Step 25709: {'lr': 0.00025476118850265863, 'samples': 13163520, 'steps': 25709, 'loss/train': 2.8330962657928467} 02/25/2022 10:53:33 - INFO - codeparrot_training - Step 25710: {'lr': 0.0002547448289983584, 'samples': 13164032, 'steps': 25710, 'loss/train': 2.475098133087158} 02/25/2022 10:53:36 - INFO - codeparrot_training - Step 25711: {'lr': 0.00025472846947373286, 'samples': 13164544, 'steps': 25711, 'loss/train': 1.94191575050354} 02/25/2022 10:53:42 - INFO - codeparrot_training - Step 25712: {'lr': 0.0002547121099288521, 'samples': 13165056, 'steps': 25712, 'loss/train': 2.3849785327911377} 02/25/2022 10:53:46 - INFO - codeparrot_training - Step 25713: {'lr': 0.00025469575036378607, 'samples': 13165568, 'steps': 25713, 'loss/train': 0.8719168305397034} 02/25/2022 10:53:51 - INFO - codeparrot_training - Step 25714: {'lr': 0.00025467939077860496, 'samples': 13166080, 'steps': 25714, 'loss/train': 2.9024498462677} 02/25/2022 10:53:55 - INFO - codeparrot_training - Step 25715: {'lr': 0.0002546630311733789, 'samples': 13166592, 'steps': 25715, 'loss/train': 1.0577749013900757} 02/25/2022 10:54:00 - INFO - codeparrot_training - Step 25716: {'lr': 0.00025464667154817777, 'samples': 13167104, 'steps': 25716, 'loss/train': 2.32977032661438} 02/25/2022 10:54:04 - INFO - codeparrot_training - Step 25717: {'lr': 0.0002546303119030719, 'samples': 13167616, 'steps': 25717, 'loss/train': 2.5572292804718018} 02/25/2022 10:54:10 - INFO - codeparrot_training - Step 25718: {'lr': 0.00025461395223813117, 'samples': 13168128, 'steps': 25718, 'loss/train': 0.13165880739688873} 02/25/2022 10:54:13 - INFO - codeparrot_training - Step 25719: {'lr': 0.0002545975925534258, 'samples': 13168640, 'steps': 25719, 'loss/train': 2.654447078704834} 02/25/2022 10:54:19 - INFO - codeparrot_training - Step 25720: {'lr': 0.00025458123284902574, 'samples': 13169152, 'steps': 25720, 'loss/train': 2.6887803077697754} 02/25/2022 10:54:22 - INFO - codeparrot_training - Step 25721: {'lr': 0.00025456487312500116, 'samples': 13169664, 'steps': 25721, 'loss/train': 2.5403947830200195} 02/25/2022 10:54:28 - INFO - codeparrot_training - Step 25722: {'lr': 0.00025454851338142207, 'samples': 13170176, 'steps': 25722, 'loss/train': 2.2072603702545166} 02/25/2022 10:54:31 - INFO - codeparrot_training - Step 25723: {'lr': 0.0002545321536183586, 'samples': 13170688, 'steps': 25723, 'loss/train': 2.1591055393218994} 02/25/2022 10:54:37 - INFO - codeparrot_training - Step 25724: {'lr': 0.00025451579383588084, 'samples': 13171200, 'steps': 25724, 'loss/train': 1.5913097858428955} 02/25/2022 10:54:40 - INFO - codeparrot_training - Step 25725: {'lr': 0.00025449943403405896, 'samples': 13171712, 'steps': 25725, 'loss/train': 0.959050714969635} 02/25/2022 10:54:46 - INFO - codeparrot_training - Step 25726: {'lr': 0.00025448307421296277, 'samples': 13172224, 'steps': 25726, 'loss/train': 2.009406328201294} 02/25/2022 10:54:50 - INFO - codeparrot_training - Step 25727: {'lr': 0.0002544667143726626, 'samples': 13172736, 'steps': 25727, 'loss/train': 2.427276134490967} 02/25/2022 10:54:55 - INFO - codeparrot_training - Step 25728: {'lr': 0.0002544503545132284, 'samples': 13173248, 'steps': 25728, 'loss/train': 1.9406293630599976} 02/25/2022 10:54:59 - INFO - codeparrot_training - Step 25729: {'lr': 0.0002544339946347303, 'samples': 13173760, 'steps': 25729, 'loss/train': 2.9963250160217285} 02/25/2022 10:55:04 - INFO - codeparrot_training - Step 25730: {'lr': 0.00025441763473723846, 'samples': 13174272, 'steps': 25730, 'loss/train': 2.4323890209198} 02/25/2022 10:55:08 - INFO - codeparrot_training - Step 25731: {'lr': 0.0002544012748208228, 'samples': 13174784, 'steps': 25731, 'loss/train': 2.374079704284668} 02/25/2022 10:55:13 - INFO - codeparrot_training - Step 25732: {'lr': 0.0002543849148855534, 'samples': 13175296, 'steps': 25732, 'loss/train': 2.516752243041992} 02/25/2022 10:55:17 - INFO - codeparrot_training - Step 25733: {'lr': 0.0002543685549315006, 'samples': 13175808, 'steps': 25733, 'loss/train': 2.384999990463257} 02/25/2022 10:55:22 - INFO - codeparrot_training - Step 25734: {'lr': 0.0002543521949587341, 'samples': 13176320, 'steps': 25734, 'loss/train': 0.9762688279151917} 02/25/2022 10:55:26 - INFO - codeparrot_training - Step 25735: {'lr': 0.0002543358349673243, 'samples': 13176832, 'steps': 25735, 'loss/train': 1.6819815635681152} 02/25/2022 10:55:31 - INFO - codeparrot_training - Step 25736: {'lr': 0.00025431947495734117, 'samples': 13177344, 'steps': 25736, 'loss/train': 1.8829540014266968} 02/25/2022 10:55:35 - INFO - codeparrot_training - Step 25737: {'lr': 0.00025430311492885473, 'samples': 13177856, 'steps': 25737, 'loss/train': 2.006132125854492} 02/25/2022 10:55:41 - INFO - codeparrot_training - Step 25738: {'lr': 0.00025428675488193517, 'samples': 13178368, 'steps': 25738, 'loss/train': 1.673203706741333} 02/25/2022 10:55:44 - INFO - codeparrot_training - Step 25739: {'lr': 0.0002542703948166524, 'samples': 13178880, 'steps': 25739, 'loss/train': 1.9815729856491089} 02/25/2022 10:55:50 - INFO - codeparrot_training - Step 25740: {'lr': 0.00025425403473307675, 'samples': 13179392, 'steps': 25740, 'loss/train': 1.899364948272705} 02/25/2022 10:55:53 - INFO - codeparrot_training - Step 25741: {'lr': 0.00025423767463127807, 'samples': 13179904, 'steps': 25741, 'loss/train': 1.8814778327941895} 02/25/2022 10:55:59 - INFO - codeparrot_training - Step 25742: {'lr': 0.00025422131451132665, 'samples': 13180416, 'steps': 25742, 'loss/train': 2.028073310852051} 02/25/2022 10:56:02 - INFO - codeparrot_training - Step 25743: {'lr': 0.0002542049543732923, 'samples': 13180928, 'steps': 25743, 'loss/train': 2.6303043365478516} 02/25/2022 10:56:08 - INFO - codeparrot_training - Step 25744: {'lr': 0.00025418859421724537, 'samples': 13181440, 'steps': 25744, 'loss/train': 0.8574888110160828} 02/25/2022 10:56:11 - INFO - codeparrot_training - Step 25745: {'lr': 0.0002541722340432558, 'samples': 13181952, 'steps': 25745, 'loss/train': 2.7316718101501465} 02/25/2022 10:56:17 - INFO - codeparrot_training - Step 25746: {'lr': 0.00025415587385139364, 'samples': 13182464, 'steps': 25746, 'loss/train': 2.671398162841797} 02/25/2022 10:56:20 - INFO - codeparrot_training - Step 25747: {'lr': 0.00025413951364172915, 'samples': 13182976, 'steps': 25747, 'loss/train': 0.6429460048675537} 02/25/2022 10:56:26 - INFO - codeparrot_training - Step 25748: {'lr': 0.0002541231534143322, 'samples': 13183488, 'steps': 25748, 'loss/train': 2.8606956005096436} 02/25/2022 10:56:30 - INFO - codeparrot_training - Step 25749: {'lr': 0.0002541067931692731, 'samples': 13184000, 'steps': 25749, 'loss/train': 1.670526146888733} 02/25/2022 10:56:35 - INFO - codeparrot_training - Step 25750: {'lr': 0.0002540904329066217, 'samples': 13184512, 'steps': 25750, 'loss/train': 1.62129807472229} 02/25/2022 10:56:39 - INFO - codeparrot_training - Step 25751: {'lr': 0.0002540740726264482, 'samples': 13185024, 'steps': 25751, 'loss/train': 1.9914284944534302} 02/25/2022 10:56:44 - INFO - codeparrot_training - Step 25752: {'lr': 0.0002540577123288227, 'samples': 13185536, 'steps': 25752, 'loss/train': 1.3021748065948486} 02/25/2022 10:56:48 - INFO - codeparrot_training - Step 25753: {'lr': 0.0002540413520138152, 'samples': 13186048, 'steps': 25753, 'loss/train': 1.5482739210128784} 02/25/2022 10:56:54 - INFO - codeparrot_training - Step 25754: {'lr': 0.00025402499168149587, 'samples': 13186560, 'steps': 25754, 'loss/train': 0.9066489934921265} 02/25/2022 10:56:57 - INFO - codeparrot_training - Step 25755: {'lr': 0.00025400863133193477, 'samples': 13187072, 'steps': 25755, 'loss/train': 1.965298056602478} 02/25/2022 10:57:02 - INFO - codeparrot_training - Step 25756: {'lr': 0.0002539922709652019, 'samples': 13187584, 'steps': 25756, 'loss/train': 2.1001808643341064} 02/25/2022 10:57:06 - INFO - codeparrot_training - Step 25757: {'lr': 0.0002539759105813675, 'samples': 13188096, 'steps': 25757, 'loss/train': 2.9081852436065674} 02/25/2022 10:57:12 - INFO - codeparrot_training - Step 25758: {'lr': 0.0002539595501805014, 'samples': 13188608, 'steps': 25758, 'loss/train': 2.226444721221924} 02/25/2022 10:57:16 - INFO - codeparrot_training - Step 25759: {'lr': 0.000253943189762674, 'samples': 13189120, 'steps': 25759, 'loss/train': 2.1923131942749023} 02/25/2022 10:57:21 - INFO - codeparrot_training - Step 25760: {'lr': 0.0002539268293279552, 'samples': 13189632, 'steps': 25760, 'loss/train': 1.6481832265853882} 02/25/2022 10:57:25 - INFO - codeparrot_training - Step 25761: {'lr': 0.00025391046887641506, 'samples': 13190144, 'steps': 25761, 'loss/train': 2.0524251461029053} 02/25/2022 10:57:30 - INFO - codeparrot_training - Step 25762: {'lr': 0.00025389410840812374, 'samples': 13190656, 'steps': 25762, 'loss/train': 1.869503140449524} 02/25/2022 10:57:33 - INFO - codeparrot_training - Step 25763: {'lr': 0.0002538777479231513, 'samples': 13191168, 'steps': 25763, 'loss/train': 0.918232262134552} 02/25/2022 10:57:39 - INFO - codeparrot_training - Step 25764: {'lr': 0.0002538613874215679, 'samples': 13191680, 'steps': 25764, 'loss/train': 2.697035551071167} 02/25/2022 10:57:43 - INFO - codeparrot_training - Step 25765: {'lr': 0.0002538450269034434, 'samples': 13192192, 'steps': 25765, 'loss/train': 2.201368808746338} 02/25/2022 10:57:48 - INFO - codeparrot_training - Step 25766: {'lr': 0.0002538286663688481, 'samples': 13192704, 'steps': 25766, 'loss/train': 3.3679611682891846} 02/25/2022 10:57:52 - INFO - codeparrot_training - Step 25767: {'lr': 0.00025381230581785196, 'samples': 13193216, 'steps': 25767, 'loss/train': 2.0616753101348877} 02/25/2022 10:57:57 - INFO - codeparrot_training - Step 25768: {'lr': 0.00025379594525052517, 'samples': 13193728, 'steps': 25768, 'loss/train': 1.1646887063980103} 02/25/2022 10:58:01 - INFO - codeparrot_training - Step 25769: {'lr': 0.00025377958466693773, 'samples': 13194240, 'steps': 25769, 'loss/train': 0.4731895625591278} 02/25/2022 10:58:06 - INFO - codeparrot_training - Step 25770: {'lr': 0.0002537632240671598, 'samples': 13194752, 'steps': 25770, 'loss/train': 1.464052438735962} 02/25/2022 10:58:10 - INFO - codeparrot_training - Step 25771: {'lr': 0.0002537468634512613, 'samples': 13195264, 'steps': 25771, 'loss/train': 2.3432819843292236} 02/25/2022 10:58:16 - INFO - codeparrot_training - Step 25772: {'lr': 0.00025373050281931247, 'samples': 13195776, 'steps': 25772, 'loss/train': 1.3197146654129028} 02/25/2022 10:58:19 - INFO - codeparrot_training - Step 25773: {'lr': 0.0002537141421713834, 'samples': 13196288, 'steps': 25773, 'loss/train': 1.645582675933838} 02/25/2022 10:58:25 - INFO - codeparrot_training - Step 25774: {'lr': 0.00025369778150754406, 'samples': 13196800, 'steps': 25774, 'loss/train': 1.7776597738265991} 02/25/2022 10:58:28 - INFO - codeparrot_training - Step 25775: {'lr': 0.00025368142082786465, 'samples': 13197312, 'steps': 25775, 'loss/train': 1.6517839431762695} 02/25/2022 10:58:34 - INFO - codeparrot_training - Step 25776: {'lr': 0.0002536650601324152, 'samples': 13197824, 'steps': 25776, 'loss/train': 0.1663322001695633} 02/25/2022 10:58:37 - INFO - codeparrot_training - Step 25777: {'lr': 0.0002536486994212658, 'samples': 13198336, 'steps': 25777, 'loss/train': 1.4063405990600586} 02/25/2022 10:58:43 - INFO - codeparrot_training - Step 25778: {'lr': 0.00025363233869448647, 'samples': 13198848, 'steps': 25778, 'loss/train': 2.351147413253784} 02/25/2022 10:58:46 - INFO - codeparrot_training - Step 25779: {'lr': 0.00025361597795214743, 'samples': 13199360, 'steps': 25779, 'loss/train': 1.40060293674469} 02/25/2022 10:58:52 - INFO - codeparrot_training - Step 25780: {'lr': 0.0002535996171943186, 'samples': 13199872, 'steps': 25780, 'loss/train': 1.617052674293518} 02/25/2022 10:58:55 - INFO - codeparrot_training - Step 25781: {'lr': 0.0002535832564210703, 'samples': 13200384, 'steps': 25781, 'loss/train': 1.1847692728042603} 02/25/2022 10:59:01 - INFO - codeparrot_training - Step 25782: {'lr': 0.0002535668956324723, 'samples': 13200896, 'steps': 25782, 'loss/train': 0.6098423600196838} 02/25/2022 10:59:04 - INFO - codeparrot_training - Step 25783: {'lr': 0.0002535505348285949, 'samples': 13201408, 'steps': 25783, 'loss/train': 1.836410641670227} 02/25/2022 10:59:11 - INFO - codeparrot_training - Step 25784: {'lr': 0.0002535341740095082, 'samples': 13201920, 'steps': 25784, 'loss/train': 2.2008070945739746} 02/25/2022 10:59:14 - INFO - codeparrot_training - Step 25785: {'lr': 0.0002535178131752822, 'samples': 13202432, 'steps': 25785, 'loss/train': 1.6727392673492432} 02/25/2022 10:59:20 - INFO - codeparrot_training - Step 25786: {'lr': 0.000253501452325987, 'samples': 13202944, 'steps': 25786, 'loss/train': 0.9830268025398254} 02/25/2022 10:59:23 - INFO - codeparrot_training - Step 25787: {'lr': 0.0002534850914616926, 'samples': 13203456, 'steps': 25787, 'loss/train': 1.4031126499176025} 02/25/2022 10:59:29 - INFO - codeparrot_training - Step 25788: {'lr': 0.0002534687305824693, 'samples': 13203968, 'steps': 25788, 'loss/train': 1.7922031879425049} 02/25/2022 10:59:32 - INFO - codeparrot_training - Step 25789: {'lr': 0.00025345236968838704, 'samples': 13204480, 'steps': 25789, 'loss/train': 1.715986728668213} 02/25/2022 10:59:38 - INFO - codeparrot_training - Step 25790: {'lr': 0.0002534360087795159, 'samples': 13204992, 'steps': 25790, 'loss/train': 2.3838462829589844} 02/25/2022 10:59:41 - INFO - codeparrot_training - Step 25791: {'lr': 0.000253419647855926, 'samples': 13205504, 'steps': 25791, 'loss/train': 2.225346088409424} 02/25/2022 10:59:47 - INFO - codeparrot_training - Step 25792: {'lr': 0.0002534032869176874, 'samples': 13206016, 'steps': 25792, 'loss/train': 0.12485877424478531} 02/25/2022 10:59:50 - INFO - codeparrot_training - Step 25793: {'lr': 0.0002533869259648702, 'samples': 13206528, 'steps': 25793, 'loss/train': 1.1403968334197998} 02/25/2022 10:59:56 - INFO - codeparrot_training - Step 25794: {'lr': 0.00025337056499754446, 'samples': 13207040, 'steps': 25794, 'loss/train': 2.668804883956909} 02/25/2022 11:00:00 - INFO - codeparrot_training - Step 25795: {'lr': 0.0002533542040157803, 'samples': 13207552, 'steps': 25795, 'loss/train': 1.8052493333816528} 02/25/2022 11:00:06 - INFO - codeparrot_training - Step 25796: {'lr': 0.0002533378430196478, 'samples': 13208064, 'steps': 25796, 'loss/train': 1.5465325117111206} 02/25/2022 11:00:09 - INFO - codeparrot_training - Step 25797: {'lr': 0.0002533214820092171, 'samples': 13208576, 'steps': 25797, 'loss/train': 1.316029667854309} 02/25/2022 11:00:15 - INFO - codeparrot_training - Step 25798: {'lr': 0.0002533051209845581, 'samples': 13209088, 'steps': 25798, 'loss/train': 2.12951922416687} 02/25/2022 11:00:18 - INFO - codeparrot_training - Step 25799: {'lr': 0.0002532887599457411, 'samples': 13209600, 'steps': 25799, 'loss/train': 1.361382246017456} 02/25/2022 11:00:24 - INFO - codeparrot_training - Step 25800: {'lr': 0.0002532723988928361, 'samples': 13210112, 'steps': 25800, 'loss/train': 1.951144814491272} 02/25/2022 11:00:27 - INFO - codeparrot_training - Step 25801: {'lr': 0.0002532560378259132, 'samples': 13210624, 'steps': 25801, 'loss/train': 2.426379680633545} 02/25/2022 11:00:33 - INFO - codeparrot_training - Step 25802: {'lr': 0.0002532396767450424, 'samples': 13211136, 'steps': 25802, 'loss/train': 1.969315767288208} 02/25/2022 11:00:36 - INFO - codeparrot_training - Step 25803: {'lr': 0.00025322331565029395, 'samples': 13211648, 'steps': 25803, 'loss/train': 1.8718268871307373} 02/25/2022 11:00:43 - INFO - codeparrot_training - Step 25804: {'lr': 0.00025320695454173777, 'samples': 13212160, 'steps': 25804, 'loss/train': 2.588383197784424} 02/25/2022 11:00:46 - INFO - codeparrot_training - Step 25805: {'lr': 0.000253190593419444, 'samples': 13212672, 'steps': 25805, 'loss/train': 2.333868980407715} 02/25/2022 11:00:51 - INFO - codeparrot_training - Step 25806: {'lr': 0.0002531742322834828, 'samples': 13213184, 'steps': 25806, 'loss/train': 0.33307725191116333} 02/25/2022 11:00:55 - INFO - codeparrot_training - Step 25807: {'lr': 0.0002531578711339241, 'samples': 13213696, 'steps': 25807, 'loss/train': 2.445645332336426} 02/25/2022 11:01:00 - INFO - codeparrot_training - Step 25808: {'lr': 0.0002531415099708382, 'samples': 13214208, 'steps': 25808, 'loss/train': 1.2026604413986206} 02/25/2022 11:01:04 - INFO - codeparrot_training - Step 25809: {'lr': 0.000253125148794295, 'samples': 13214720, 'steps': 25809, 'loss/train': 1.6700819730758667} 02/25/2022 11:01:09 - INFO - codeparrot_training - Step 25810: {'lr': 0.0002531087876043647, 'samples': 13215232, 'steps': 25810, 'loss/train': 1.4728336334228516} 02/25/2022 11:01:13 - INFO - codeparrot_training - Step 25811: {'lr': 0.0002530924264011172, 'samples': 13215744, 'steps': 25811, 'loss/train': 0.7189762592315674} 02/25/2022 11:01:18 - INFO - codeparrot_training - Step 25812: {'lr': 0.00025307606518462297, 'samples': 13216256, 'steps': 25812, 'loss/train': 2.2786972522735596} 02/25/2022 11:01:22 - INFO - codeparrot_training - Step 25813: {'lr': 0.00025305970395495165, 'samples': 13216768, 'steps': 25813, 'loss/train': 1.4087247848510742} 02/25/2022 11:01:27 - INFO - codeparrot_training - Step 25814: {'lr': 0.00025304334271217363, 'samples': 13217280, 'steps': 25814, 'loss/train': 1.1288540363311768} 02/25/2022 11:01:31 - INFO - codeparrot_training - Step 25815: {'lr': 0.00025302698145635884, 'samples': 13217792, 'steps': 25815, 'loss/train': 1.194296956062317} 02/25/2022 11:01:36 - INFO - codeparrot_training - Step 25816: {'lr': 0.0002530106201875775, 'samples': 13218304, 'steps': 25816, 'loss/train': 2.638526678085327} 02/25/2022 11:01:40 - INFO - codeparrot_training - Step 25817: {'lr': 0.0002529942589058995, 'samples': 13218816, 'steps': 25817, 'loss/train': 1.6297551393508911} 02/25/2022 11:01:45 - INFO - codeparrot_training - Step 25818: {'lr': 0.0002529778976113952, 'samples': 13219328, 'steps': 25818, 'loss/train': 1.6843180656433105} 02/25/2022 11:01:49 - INFO - codeparrot_training - Step 25819: {'lr': 0.00025296153630413435, 'samples': 13219840, 'steps': 25819, 'loss/train': 1.806922197341919} 02/25/2022 11:01:55 - INFO - codeparrot_training - Step 25820: {'lr': 0.00025294517498418727, 'samples': 13220352, 'steps': 25820, 'loss/train': 1.7788069248199463} 02/25/2022 11:01:58 - INFO - codeparrot_training - Step 25821: {'lr': 0.00025292881365162413, 'samples': 13220864, 'steps': 25821, 'loss/train': 1.8311032056808472} 02/25/2022 11:02:04 - INFO - codeparrot_training - Step 25822: {'lr': 0.0002529124523065147, 'samples': 13221376, 'steps': 25822, 'loss/train': 1.8440948724746704} 02/25/2022 11:02:07 - INFO - codeparrot_training - Step 25823: {'lr': 0.0002528960909489294, 'samples': 13221888, 'steps': 25823, 'loss/train': 1.6727509498596191} 02/25/2022 11:02:13 - INFO - codeparrot_training - Step 25824: {'lr': 0.0002528797295789381, 'samples': 13222400, 'steps': 25824, 'loss/train': 1.6291556358337402} 02/25/2022 11:02:16 - INFO - codeparrot_training - Step 25825: {'lr': 0.000252863368196611, 'samples': 13222912, 'steps': 25825, 'loss/train': 1.479705810546875} 02/25/2022 11:02:22 - INFO - codeparrot_training - Step 25826: {'lr': 0.00025284700680201804, 'samples': 13223424, 'steps': 25826, 'loss/train': 1.7157994508743286} 02/25/2022 11:02:25 - INFO - codeparrot_training - Step 25827: {'lr': 0.0002528306453952295, 'samples': 13223936, 'steps': 25827, 'loss/train': 3.061528205871582} 02/25/2022 11:02:31 - INFO - codeparrot_training - Step 25828: {'lr': 0.00025281428397631536, 'samples': 13224448, 'steps': 25828, 'loss/train': 2.138091802597046} 02/25/2022 11:02:34 - INFO - codeparrot_training - Step 25829: {'lr': 0.0002527979225453457, 'samples': 13224960, 'steps': 25829, 'loss/train': 1.6424891948699951} 02/25/2022 11:02:41 - INFO - codeparrot_training - Step 25830: {'lr': 0.0002527815611023906, 'samples': 13225472, 'steps': 25830, 'loss/train': 2.46722412109375} 02/25/2022 11:02:44 - INFO - codeparrot_training - Step 25831: {'lr': 0.00025276519964752015, 'samples': 13225984, 'steps': 25831, 'loss/train': 1.6439441442489624} 02/25/2022 11:02:50 - INFO - codeparrot_training - Step 25832: {'lr': 0.00025274883818080456, 'samples': 13226496, 'steps': 25832, 'loss/train': 1.950255274772644} 02/25/2022 11:02:53 - INFO - codeparrot_training - Step 25833: {'lr': 0.0002527324767023138, 'samples': 13227008, 'steps': 25833, 'loss/train': 1.8935997486114502} 02/25/2022 11:02:59 - INFO - codeparrot_training - Step 25834: {'lr': 0.00025271611521211795, 'samples': 13227520, 'steps': 25834, 'loss/train': 2.3201212882995605} 02/25/2022 11:03:02 - INFO - codeparrot_training - Step 25835: {'lr': 0.0002526997537102871, 'samples': 13228032, 'steps': 25835, 'loss/train': 3.44230318069458} 02/25/2022 11:03:08 - INFO - codeparrot_training - Step 25836: {'lr': 0.0002526833921968914, 'samples': 13228544, 'steps': 25836, 'loss/train': 0.8415111899375916} 02/25/2022 11:03:11 - INFO - codeparrot_training - Step 25837: {'lr': 0.00025266703067200084, 'samples': 13229056, 'steps': 25837, 'loss/train': 1.803125262260437} 02/25/2022 11:03:17 - INFO - codeparrot_training - Step 25838: {'lr': 0.0002526506691356857, 'samples': 13229568, 'steps': 25838, 'loss/train': 1.0776190757751465} 02/25/2022 11:03:20 - INFO - codeparrot_training - Step 25839: {'lr': 0.00025263430758801583, 'samples': 13230080, 'steps': 25839, 'loss/train': 2.6610002517700195} 02/25/2022 11:03:26 - INFO - codeparrot_training - Step 25840: {'lr': 0.00025261794602906147, 'samples': 13230592, 'steps': 25840, 'loss/train': 2.604844331741333} 02/25/2022 11:03:30 - INFO - codeparrot_training - Step 25841: {'lr': 0.00025260158445889266, 'samples': 13231104, 'steps': 25841, 'loss/train': 2.0738120079040527} 02/25/2022 11:03:35 - INFO - codeparrot_training - Step 25842: {'lr': 0.00025258522287757945, 'samples': 13231616, 'steps': 25842, 'loss/train': 2.1464474201202393} 02/25/2022 11:03:39 - INFO - codeparrot_training - Step 25843: {'lr': 0.000252568861285192, 'samples': 13232128, 'steps': 25843, 'loss/train': 1.261229157447815} 02/25/2022 11:03:44 - INFO - codeparrot_training - Step 25844: {'lr': 0.00025255249968180035, 'samples': 13232640, 'steps': 25844, 'loss/train': 1.1351053714752197} 02/25/2022 11:03:48 - INFO - codeparrot_training - Step 25845: {'lr': 0.00025253613806747454, 'samples': 13233152, 'steps': 25845, 'loss/train': 1.0389478206634521} 02/25/2022 11:03:54 - INFO - codeparrot_training - Step 25846: {'lr': 0.0002525197764422848, 'samples': 13233664, 'steps': 25846, 'loss/train': 1.6159627437591553} 02/25/2022 11:03:57 - INFO - codeparrot_training - Step 25847: {'lr': 0.00025250341480630113, 'samples': 13234176, 'steps': 25847, 'loss/train': 1.8487048149108887} 02/25/2022 11:04:03 - INFO - codeparrot_training - Step 25848: {'lr': 0.0002524870531595936, 'samples': 13234688, 'steps': 25848, 'loss/train': 1.6222976446151733} 02/25/2022 11:04:06 - INFO - codeparrot_training - Step 25849: {'lr': 0.00025247069150223235, 'samples': 13235200, 'steps': 25849, 'loss/train': 1.6385929584503174} 02/25/2022 11:04:12 - INFO - codeparrot_training - Step 25850: {'lr': 0.0002524543298342875, 'samples': 13235712, 'steps': 25850, 'loss/train': 0.9132198691368103} 02/25/2022 11:04:16 - INFO - codeparrot_training - Step 25851: {'lr': 0.00025243796815582894, 'samples': 13236224, 'steps': 25851, 'loss/train': 1.9199039936065674} 02/25/2022 11:04:21 - INFO - codeparrot_training - Step 25852: {'lr': 0.000252421606466927, 'samples': 13236736, 'steps': 25852, 'loss/train': 2.31599497795105} 02/25/2022 11:04:25 - INFO - codeparrot_training - Step 25853: {'lr': 0.00025240524476765166, 'samples': 13237248, 'steps': 25853, 'loss/train': 2.0127487182617188} 02/25/2022 11:04:30 - INFO - codeparrot_training - Step 25854: {'lr': 0.00025238888305807294, 'samples': 13237760, 'steps': 25854, 'loss/train': 1.9955291748046875} 02/25/2022 11:04:34 - INFO - codeparrot_training - Step 25855: {'lr': 0.00025237252133826103, 'samples': 13238272, 'steps': 25855, 'loss/train': 3.0744364261627197} 02/25/2022 11:04:39 - INFO - codeparrot_training - Step 25856: {'lr': 0.0002523561596082861, 'samples': 13238784, 'steps': 25856, 'loss/train': 1.3415467739105225} 02/25/2022 11:04:43 - INFO - codeparrot_training - Step 25857: {'lr': 0.00025233979786821797, 'samples': 13239296, 'steps': 25857, 'loss/train': 1.6282848119735718} 02/25/2022 11:04:48 - INFO - codeparrot_training - Step 25858: {'lr': 0.00025232343611812695, 'samples': 13239808, 'steps': 25858, 'loss/train': 1.8890135288238525} 02/25/2022 11:04:52 - INFO - codeparrot_training - Step 25859: {'lr': 0.00025230707435808307, 'samples': 13240320, 'steps': 25859, 'loss/train': 1.5632286071777344} 02/25/2022 11:05:00 - INFO - codeparrot_training - Step 25860: {'lr': 0.0002522907125881565, 'samples': 13240832, 'steps': 25860, 'loss/train': 2.3817331790924072} 02/25/2022 11:05:03 - INFO - codeparrot_training - Step 25861: {'lr': 0.0002522743508084171, 'samples': 13241344, 'steps': 25861, 'loss/train': 1.6820285320281982} 02/25/2022 11:05:09 - INFO - codeparrot_training - Step 25862: {'lr': 0.0002522579890189352, 'samples': 13241856, 'steps': 25862, 'loss/train': 2.2996323108673096} 02/25/2022 11:05:12 - INFO - codeparrot_training - Step 25863: {'lr': 0.0002522416272197808, 'samples': 13242368, 'steps': 25863, 'loss/train': 1.5716899633407593} 02/25/2022 11:05:18 - INFO - codeparrot_training - Step 25864: {'lr': 0.0002522252654110239, 'samples': 13242880, 'steps': 25864, 'loss/train': 1.6403475999832153} 02/25/2022 11:05:22 - INFO - codeparrot_training - Step 25865: {'lr': 0.00025220890359273473, 'samples': 13243392, 'steps': 25865, 'loss/train': 2.276691198348999} 02/25/2022 11:05:27 - INFO - codeparrot_training - Step 25866: {'lr': 0.00025219254176498334, 'samples': 13243904, 'steps': 25866, 'loss/train': 3.263749122619629} 02/25/2022 11:05:31 - INFO - codeparrot_training - Step 25867: {'lr': 0.0002521761799278398, 'samples': 13244416, 'steps': 25867, 'loss/train': 0.7852783799171448} 02/25/2022 11:05:38 - INFO - codeparrot_training - Step 25868: {'lr': 0.0002521598180813741, 'samples': 13244928, 'steps': 25868, 'loss/train': 2.3589694499969482} 02/25/2022 11:05:41 - INFO - codeparrot_training - Step 25869: {'lr': 0.0002521434562256565, 'samples': 13245440, 'steps': 25869, 'loss/train': 0.21988695859909058} 02/25/2022 11:05:47 - INFO - codeparrot_training - Step 25870: {'lr': 0.000252127094360757, 'samples': 13245952, 'steps': 25870, 'loss/train': 1.8397248983383179} 02/25/2022 11:05:51 - INFO - codeparrot_training - Step 25871: {'lr': 0.00025211073248674573, 'samples': 13246464, 'steps': 25871, 'loss/train': 1.4425493478775024} 02/25/2022 11:05:57 - INFO - codeparrot_training - Step 25872: {'lr': 0.00025209437060369266, 'samples': 13246976, 'steps': 25872, 'loss/train': 2.2448723316192627} 02/25/2022 11:06:00 - INFO - codeparrot_training - Step 25873: {'lr': 0.0002520780087116681, 'samples': 13247488, 'steps': 25873, 'loss/train': 1.781081199645996} 02/25/2022 11:06:04 - INFO - codeparrot_training - Step 25874: {'lr': 0.0002520616468107419, 'samples': 13248000, 'steps': 25874, 'loss/train': 0.39445847272872925} 02/25/2022 11:06:09 - INFO - codeparrot_training - Step 25875: {'lr': 0.0002520452849009843, 'samples': 13248512, 'steps': 25875, 'loss/train': 1.5589098930358887} 02/25/2022 11:06:13 - INFO - codeparrot_training - Step 25876: {'lr': 0.0002520289229824653, 'samples': 13249024, 'steps': 25876, 'loss/train': 1.2916417121887207} 02/25/2022 11:06:19 - INFO - codeparrot_training - Step 25877: {'lr': 0.0002520125610552552, 'samples': 13249536, 'steps': 25877, 'loss/train': 2.270116090774536} 02/25/2022 11:06:22 - INFO - codeparrot_training - Step 25878: {'lr': 0.0002519961991194238, 'samples': 13250048, 'steps': 25878, 'loss/train': 1.5044080018997192} 02/25/2022 11:06:28 - INFO - codeparrot_training - Step 25879: {'lr': 0.0002519798371750413, 'samples': 13250560, 'steps': 25879, 'loss/train': 2.329695224761963} 02/25/2022 11:06:31 - INFO - codeparrot_training - Step 25880: {'lr': 0.0002519634752221778, 'samples': 13251072, 'steps': 25880, 'loss/train': 2.5209453105926514} 02/25/2022 11:06:38 - INFO - codeparrot_training - Step 25881: {'lr': 0.00025194711326090346, 'samples': 13251584, 'steps': 25881, 'loss/train': 2.630035638809204} 02/25/2022 11:06:42 - INFO - codeparrot_training - Step 25882: {'lr': 0.0002519307512912883, 'samples': 13252096, 'steps': 25882, 'loss/train': 1.6330238580703735} 02/25/2022 11:06:47 - INFO - codeparrot_training - Step 25883: {'lr': 0.00025191438931340237, 'samples': 13252608, 'steps': 25883, 'loss/train': 1.349502444267273} 02/25/2022 11:06:51 - INFO - codeparrot_training - Step 25884: {'lr': 0.0002518980273273159, 'samples': 13253120, 'steps': 25884, 'loss/train': 2.3931283950805664} 02/25/2022 11:06:56 - INFO - codeparrot_training - Step 25885: {'lr': 0.00025188166533309875, 'samples': 13253632, 'steps': 25885, 'loss/train': 2.9504313468933105} 02/25/2022 11:07:00 - INFO - codeparrot_training - Step 25886: {'lr': 0.0002518653033308212, 'samples': 13254144, 'steps': 25886, 'loss/train': 1.3233942985534668} 02/25/2022 11:07:06 - INFO - codeparrot_training - Step 25887: {'lr': 0.0002518489413205533, 'samples': 13254656, 'steps': 25887, 'loss/train': 2.4223709106445312} 02/25/2022 11:07:09 - INFO - codeparrot_training - Step 25888: {'lr': 0.00025183257930236515, 'samples': 13255168, 'steps': 25888, 'loss/train': 2.405268430709839} 02/25/2022 11:07:15 - INFO - codeparrot_training - Step 25889: {'lr': 0.00025181621727632673, 'samples': 13255680, 'steps': 25889, 'loss/train': 0.924758791923523} 02/25/2022 11:07:18 - INFO - codeparrot_training - Step 25890: {'lr': 0.0002517998552425083, 'samples': 13256192, 'steps': 25890, 'loss/train': 1.5506850481033325} 02/25/2022 11:07:25 - INFO - codeparrot_training - Step 25891: {'lr': 0.00025178349320097984, 'samples': 13256704, 'steps': 25891, 'loss/train': 0.4020494818687439} 02/25/2022 11:07:29 - INFO - codeparrot_training - Step 25892: {'lr': 0.00025176713115181143, 'samples': 13257216, 'steps': 25892, 'loss/train': 2.080244779586792} 02/25/2022 11:07:35 - INFO - codeparrot_training - Step 25893: {'lr': 0.0002517507690950733, 'samples': 13257728, 'steps': 25893, 'loss/train': 2.1497750282287598} 02/25/2022 11:07:38 - INFO - codeparrot_training - Step 25894: {'lr': 0.0002517344070308353, 'samples': 13258240, 'steps': 25894, 'loss/train': 4.769745349884033} 02/25/2022 11:07:44 - INFO - codeparrot_training - Step 25895: {'lr': 0.0002517180449591677, 'samples': 13258752, 'steps': 25895, 'loss/train': 1.1782888174057007} 02/25/2022 11:07:47 - INFO - codeparrot_training - Step 25896: {'lr': 0.00025170168288014046, 'samples': 13259264, 'steps': 25896, 'loss/train': 1.297042727470398} 02/25/2022 11:07:53 - INFO - codeparrot_training - Step 25897: {'lr': 0.0002516853207938238, 'samples': 13259776, 'steps': 25897, 'loss/train': 1.8245781660079956} 02/25/2022 11:07:56 - INFO - codeparrot_training - Step 25898: {'lr': 0.0002516689587002878, 'samples': 13260288, 'steps': 25898, 'loss/train': 2.8517932891845703} 02/25/2022 11:08:02 - INFO - codeparrot_training - Step 25899: {'lr': 0.00025165259659960256, 'samples': 13260800, 'steps': 25899, 'loss/train': 1.672888994216919} 02/25/2022 11:08:05 - INFO - codeparrot_training - Step 25900: {'lr': 0.00025163623449183796, 'samples': 13261312, 'steps': 25900, 'loss/train': 2.4763741493225098} 02/25/2022 11:08:11 - INFO - codeparrot_training - Step 25901: {'lr': 0.00025161987237706443, 'samples': 13261824, 'steps': 25901, 'loss/train': 0.5543332099914551} 02/25/2022 11:08:14 - INFO - codeparrot_training - Step 25902: {'lr': 0.00025160351025535173, 'samples': 13262336, 'steps': 25902, 'loss/train': 0.34986764192581177} 02/25/2022 11:08:20 - INFO - codeparrot_training - Step 25903: {'lr': 0.0002515871481267702, 'samples': 13262848, 'steps': 25903, 'loss/train': 2.4872703552246094} 02/25/2022 11:08:23 - INFO - codeparrot_training - Step 25904: {'lr': 0.00025157078599138976, 'samples': 13263360, 'steps': 25904, 'loss/train': 1.520838737487793} 02/25/2022 11:08:29 - INFO - codeparrot_training - Step 25905: {'lr': 0.00025155442384928057, 'samples': 13263872, 'steps': 25905, 'loss/train': 2.6658027172088623} 02/25/2022 11:08:32 - INFO - codeparrot_training - Step 25906: {'lr': 0.0002515380617005128, 'samples': 13264384, 'steps': 25906, 'loss/train': 2.430997848510742} 02/25/2022 11:08:40 - INFO - codeparrot_training - Step 25907: {'lr': 0.0002515216995451564, 'samples': 13264896, 'steps': 25907, 'loss/train': 0.6682925820350647} 02/25/2022 11:08:43 - INFO - codeparrot_training - Step 25908: {'lr': 0.0002515053373832816, 'samples': 13265408, 'steps': 25908, 'loss/train': 1.8046756982803345} 02/25/2022 11:08:49 - INFO - codeparrot_training - Step 25909: {'lr': 0.0002514889752149583, 'samples': 13265920, 'steps': 25909, 'loss/train': 0.4461532533168793} 02/25/2022 11:08:52 - INFO - codeparrot_training - Step 25910: {'lr': 0.00025147261304025683, 'samples': 13266432, 'steps': 25910, 'loss/train': 2.2284903526306152} 02/25/2022 11:08:58 - INFO - codeparrot_training - Step 25911: {'lr': 0.000251456250859247, 'samples': 13266944, 'steps': 25911, 'loss/train': 2.7277920246124268} 02/25/2022 11:09:01 - INFO - codeparrot_training - Step 25912: {'lr': 0.00025143988867199916, 'samples': 13267456, 'steps': 25912, 'loss/train': 4.894824981689453} 02/25/2022 11:09:07 - INFO - codeparrot_training - Step 25913: {'lr': 0.00025142352647858324, 'samples': 13267968, 'steps': 25913, 'loss/train': 1.703282356262207} 02/25/2022 11:09:10 - INFO - codeparrot_training - Step 25914: {'lr': 0.00025140716427906945, 'samples': 13268480, 'steps': 25914, 'loss/train': 2.0290658473968506} 02/25/2022 11:09:16 - INFO - codeparrot_training - Step 25915: {'lr': 0.00025139080207352776, 'samples': 13268992, 'steps': 25915, 'loss/train': 1.3233006000518799} 02/25/2022 11:09:19 - INFO - codeparrot_training - Step 25916: {'lr': 0.00025137443986202827, 'samples': 13269504, 'steps': 25916, 'loss/train': 1.7200279235839844} 02/25/2022 11:09:27 - INFO - codeparrot_training - Step 25917: {'lr': 0.00025135807764464123, 'samples': 13270016, 'steps': 25917, 'loss/train': 2.1949801445007324} 02/25/2022 11:09:30 - INFO - codeparrot_training - Step 25918: {'lr': 0.0002513417154214366, 'samples': 13270528, 'steps': 25918, 'loss/train': 0.7022582292556763} 02/25/2022 11:09:36 - INFO - codeparrot_training - Step 25919: {'lr': 0.0002513253531924844, 'samples': 13271040, 'steps': 25919, 'loss/train': 3.2486822605133057} 02/25/2022 11:09:39 - INFO - codeparrot_training - Step 25920: {'lr': 0.0002513089909578549, 'samples': 13271552, 'steps': 25920, 'loss/train': 1.7495293617248535} 02/25/2022 11:09:43 - INFO - codeparrot_training - Step 25921: {'lr': 0.0002512926287176181, 'samples': 13272064, 'steps': 25921, 'loss/train': 1.521016001701355} 02/25/2022 11:09:48 - INFO - codeparrot_training - Step 25922: {'lr': 0.000251276266471844, 'samples': 13272576, 'steps': 25922, 'loss/train': 1.9212186336517334} 02/25/2022 11:09:52 - INFO - codeparrot_training - Step 25923: {'lr': 0.00025125990422060284, 'samples': 13273088, 'steps': 25923, 'loss/train': 2.0351526737213135} 02/25/2022 11:09:57 - INFO - codeparrot_training - Step 25924: {'lr': 0.00025124354196396465, 'samples': 13273600, 'steps': 25924, 'loss/train': 1.6816900968551636} 02/25/2022 11:10:03 - INFO - codeparrot_training - Step 25925: {'lr': 0.00025122717970199956, 'samples': 13274112, 'steps': 25925, 'loss/train': 2.1449434757232666} 02/25/2022 11:10:07 - INFO - codeparrot_training - Step 25926: {'lr': 0.0002512108174347775, 'samples': 13274624, 'steps': 25926, 'loss/train': 2.3340272903442383} 02/25/2022 11:10:10 - INFO - codeparrot_training - Step 25927: {'lr': 0.0002511944551623688, 'samples': 13275136, 'steps': 25927, 'loss/train': 1.4073207378387451} 02/25/2022 11:10:17 - INFO - codeparrot_training - Step 25928: {'lr': 0.00025117809288484333, 'samples': 13275648, 'steps': 25928, 'loss/train': 1.188170075416565} 02/25/2022 11:10:21 - INFO - codeparrot_training - Step 25929: {'lr': 0.0002511617306022713, 'samples': 13276160, 'steps': 25929, 'loss/train': 2.3164262771606445} 02/25/2022 11:10:26 - INFO - codeparrot_training - Step 25930: {'lr': 0.0002511453683147229, 'samples': 13276672, 'steps': 25930, 'loss/train': 1.4057706594467163} 02/25/2022 11:10:30 - INFO - codeparrot_training - Step 25931: {'lr': 0.000251129006022268, 'samples': 13277184, 'steps': 25931, 'loss/train': 1.5364357233047485} 02/25/2022 11:10:35 - INFO - codeparrot_training - Step 25932: {'lr': 0.00025111264372497686, 'samples': 13277696, 'steps': 25932, 'loss/train': 2.583714008331299} 02/25/2022 11:10:39 - INFO - codeparrot_training - Step 25933: {'lr': 0.0002510962814229195, 'samples': 13278208, 'steps': 25933, 'loss/train': 2.0228497982025146} 02/25/2022 11:10:44 - INFO - codeparrot_training - Step 25934: {'lr': 0.000251079919116166, 'samples': 13278720, 'steps': 25934, 'loss/train': 0.3589111864566803} 02/25/2022 11:10:48 - INFO - codeparrot_training - Step 25935: {'lr': 0.00025106355680478646, 'samples': 13279232, 'steps': 25935, 'loss/train': 1.6998690366744995} 02/25/2022 11:10:54 - INFO - codeparrot_training - Step 25936: {'lr': 0.000251047194488851, 'samples': 13279744, 'steps': 25936, 'loss/train': 2.191589117050171} 02/25/2022 11:10:57 - INFO - codeparrot_training - Step 25937: {'lr': 0.0002510308321684297, 'samples': 13280256, 'steps': 25937, 'loss/train': 2.3912060260772705} 02/25/2022 11:11:03 - INFO - codeparrot_training - Step 25938: {'lr': 0.0002510144698435926, 'samples': 13280768, 'steps': 25938, 'loss/train': 1.7736722230911255} 02/25/2022 11:11:06 - INFO - codeparrot_training - Step 25939: {'lr': 0.00025099810751440986, 'samples': 13281280, 'steps': 25939, 'loss/train': 1.5722779035568237} 02/25/2022 11:11:14 - INFO - codeparrot_training - Step 25940: {'lr': 0.0002509817451809515, 'samples': 13281792, 'steps': 25940, 'loss/train': 1.761673092842102} 02/25/2022 11:11:17 - INFO - codeparrot_training - Step 25941: {'lr': 0.00025096538284328775, 'samples': 13282304, 'steps': 25941, 'loss/train': 2.2593841552734375} 02/25/2022 11:11:23 - INFO - codeparrot_training - Step 25942: {'lr': 0.0002509490205014885, 'samples': 13282816, 'steps': 25942, 'loss/train': 1.6592916250228882} 02/25/2022 11:11:26 - INFO - codeparrot_training - Step 25943: {'lr': 0.00025093265815562406, 'samples': 13283328, 'steps': 25943, 'loss/train': 0.317325621843338} 02/25/2022 11:11:32 - INFO - codeparrot_training - Step 25944: {'lr': 0.00025091629580576433, 'samples': 13283840, 'steps': 25944, 'loss/train': 1.5023730993270874} 02/25/2022 11:11:35 - INFO - codeparrot_training - Step 25945: {'lr': 0.00025089993345197946, 'samples': 13284352, 'steps': 25945, 'loss/train': 1.326545238494873} 02/25/2022 11:11:41 - INFO - codeparrot_training - Step 25946: {'lr': 0.0002508835710943396, 'samples': 13284864, 'steps': 25946, 'loss/train': 2.2905256748199463} 02/25/2022 11:11:44 - INFO - codeparrot_training - Step 25947: {'lr': 0.00025086720873291485, 'samples': 13285376, 'steps': 25947, 'loss/train': 0.7675917148590088} 02/25/2022 11:11:49 - INFO - codeparrot_training - Step 25948: {'lr': 0.0002508508463677752, 'samples': 13285888, 'steps': 25948, 'loss/train': 2.64031720161438} 02/25/2022 11:11:53 - INFO - codeparrot_training - Step 25949: {'lr': 0.0002508344839989908, 'samples': 13286400, 'steps': 25949, 'loss/train': 1.8297836780548096} 02/25/2022 11:12:00 - INFO - codeparrot_training - Step 25950: {'lr': 0.0002508181216266317, 'samples': 13286912, 'steps': 25950, 'loss/train': 1.6454377174377441} 02/25/2022 11:12:04 - INFO - codeparrot_training - Step 25951: {'lr': 0.0002508017592507681, 'samples': 13287424, 'steps': 25951, 'loss/train': 1.8232076168060303} 02/25/2022 11:12:09 - INFO - codeparrot_training - Step 25952: {'lr': 0.0002507853968714699, 'samples': 13287936, 'steps': 25952, 'loss/train': 3.2949485778808594} 02/25/2022 11:12:13 - INFO - codeparrot_training - Step 25953: {'lr': 0.0002507690344888074, 'samples': 13288448, 'steps': 25953, 'loss/train': 0.8985562324523926} 02/25/2022 11:12:19 - INFO - codeparrot_training - Step 25954: {'lr': 0.00025075267210285056, 'samples': 13288960, 'steps': 25954, 'loss/train': 1.9168519973754883} 02/25/2022 11:12:22 - INFO - codeparrot_training - Step 25955: {'lr': 0.0002507363097136696, 'samples': 13289472, 'steps': 25955, 'loss/train': 2.2396509647369385} 02/25/2022 11:12:28 - INFO - codeparrot_training - Step 25956: {'lr': 0.0002507199473213344, 'samples': 13289984, 'steps': 25956, 'loss/train': 1.6588709354400635} 02/25/2022 11:12:31 - INFO - codeparrot_training - Step 25957: {'lr': 0.00025070358492591523, 'samples': 13290496, 'steps': 25957, 'loss/train': 1.9297422170639038} 02/25/2022 11:12:37 - INFO - codeparrot_training - Step 25958: {'lr': 0.00025068722252748215, 'samples': 13291008, 'steps': 25958, 'loss/train': 1.6260007619857788} 02/25/2022 11:12:40 - INFO - codeparrot_training - Step 25959: {'lr': 0.0002506708601261052, 'samples': 13291520, 'steps': 25959, 'loss/train': 3.115778923034668} 02/25/2022 11:12:46 - INFO - codeparrot_training - Step 25960: {'lr': 0.00025065449772185456, 'samples': 13292032, 'steps': 25960, 'loss/train': 2.6047306060791016} 02/25/2022 11:12:49 - INFO - codeparrot_training - Step 25961: {'lr': 0.0002506381353148002, 'samples': 13292544, 'steps': 25961, 'loss/train': 2.872133493423462} 02/25/2022 11:12:55 - INFO - codeparrot_training - Step 25962: {'lr': 0.00025062177290501226, 'samples': 13293056, 'steps': 25962, 'loss/train': 1.9665627479553223} 02/25/2022 11:12:58 - INFO - codeparrot_training - Step 25963: {'lr': 0.0002506054104925608, 'samples': 13293568, 'steps': 25963, 'loss/train': 2.273334503173828} 02/25/2022 11:13:06 - INFO - codeparrot_training - Step 25964: {'lr': 0.00025058904807751604, 'samples': 13294080, 'steps': 25964, 'loss/train': 2.1211843490600586} 02/25/2022 11:13:09 - INFO - codeparrot_training - Step 25965: {'lr': 0.00025057268565994794, 'samples': 13294592, 'steps': 25965, 'loss/train': 1.1610602140426636} 02/25/2022 11:13:15 - INFO - codeparrot_training - Step 25966: {'lr': 0.00025055632323992666, 'samples': 13295104, 'steps': 25966, 'loss/train': 1.9810941219329834} 02/25/2022 11:13:18 - INFO - codeparrot_training - Step 25967: {'lr': 0.00025053996081752226, 'samples': 13295616, 'steps': 25967, 'loss/train': 2.5076656341552734} 02/25/2022 11:13:24 - INFO - codeparrot_training - Step 25968: {'lr': 0.0002505235983928048, 'samples': 13296128, 'steps': 25968, 'loss/train': 1.9034626483917236} 02/25/2022 11:13:27 - INFO - codeparrot_training - Step 25969: {'lr': 0.00025050723596584454, 'samples': 13296640, 'steps': 25969, 'loss/train': 1.7650463581085205} 02/25/2022 11:13:33 - INFO - codeparrot_training - Step 25970: {'lr': 0.0002504908735367113, 'samples': 13297152, 'steps': 25970, 'loss/train': 3.1255433559417725} 02/25/2022 11:13:36 - INFO - codeparrot_training - Step 25971: {'lr': 0.0002504745111054754, 'samples': 13297664, 'steps': 25971, 'loss/train': 1.3024663925170898} 02/25/2022 11:13:42 - INFO - codeparrot_training - Step 25972: {'lr': 0.00025045814867220677, 'samples': 13298176, 'steps': 25972, 'loss/train': 1.6699652671813965} 02/25/2022 11:13:46 - INFO - codeparrot_training - Step 25973: {'lr': 0.00025044178623697565, 'samples': 13298688, 'steps': 25973, 'loss/train': 2.245182991027832} 02/25/2022 11:13:53 - INFO - codeparrot_training - Step 25974: {'lr': 0.000250425423799852, 'samples': 13299200, 'steps': 25974, 'loss/train': 1.8966944217681885} 02/25/2022 11:13:56 - INFO - codeparrot_training - Step 25975: {'lr': 0.00025040906136090604, 'samples': 13299712, 'steps': 25975, 'loss/train': 1.5828367471694946} 02/25/2022 11:14:02 - INFO - codeparrot_training - Step 25976: {'lr': 0.00025039269892020773, 'samples': 13300224, 'steps': 25976, 'loss/train': 1.1790169477462769} 02/25/2022 11:14:05 - INFO - codeparrot_training - Step 25977: {'lr': 0.0002503763364778273, 'samples': 13300736, 'steps': 25977, 'loss/train': 1.0677714347839355} 02/25/2022 11:14:11 - INFO - codeparrot_training - Step 25978: {'lr': 0.0002503599740338347, 'samples': 13301248, 'steps': 25978, 'loss/train': 2.010186195373535} 02/25/2022 11:14:14 - INFO - codeparrot_training - Step 25979: {'lr': 0.00025034361158830006, 'samples': 13301760, 'steps': 25979, 'loss/train': 1.9555803537368774} 02/25/2022 11:14:20 - INFO - codeparrot_training - Step 25980: {'lr': 0.0002503272491412936, 'samples': 13302272, 'steps': 25980, 'loss/train': 3.3068318367004395} 02/25/2022 11:14:24 - INFO - codeparrot_training - Step 25981: {'lr': 0.00025031088669288514, 'samples': 13302784, 'steps': 25981, 'loss/train': 1.796302318572998} 02/25/2022 11:14:29 - INFO - codeparrot_training - Step 25982: {'lr': 0.00025029452424314507, 'samples': 13303296, 'steps': 25982, 'loss/train': 1.937954068183899} 02/25/2022 11:14:33 - INFO - codeparrot_training - Step 25983: {'lr': 0.0002502781617921433, 'samples': 13303808, 'steps': 25983, 'loss/train': 1.954081416130066} 02/25/2022 11:14:38 - INFO - codeparrot_training - Step 25984: {'lr': 0.00025026179933995, 'samples': 13304320, 'steps': 25984, 'loss/train': 1.7263826131820679} 02/25/2022 11:14:42 - INFO - codeparrot_training - Step 25985: {'lr': 0.0002502454368866352, 'samples': 13304832, 'steps': 25985, 'loss/train': 0.9804202318191528} 02/25/2022 11:14:49 - INFO - codeparrot_training - Step 25986: {'lr': 0.0002502290744322691, 'samples': 13305344, 'steps': 25986, 'loss/train': 1.898422360420227} 02/25/2022 11:14:52 - INFO - codeparrot_training - Step 25987: {'lr': 0.00025021271197692163, 'samples': 13305856, 'steps': 25987, 'loss/train': 1.9202721118927002} 02/25/2022 11:14:58 - INFO - codeparrot_training - Step 25988: {'lr': 0.000250196349520663, 'samples': 13306368, 'steps': 25988, 'loss/train': 2.741056442260742} 02/25/2022 11:15:01 - INFO - codeparrot_training - Step 25989: {'lr': 0.0002501799870635633, 'samples': 13306880, 'steps': 25989, 'loss/train': 1.5168230533599854} 02/25/2022 11:15:07 - INFO - codeparrot_training - Step 25990: {'lr': 0.00025016362460569255, 'samples': 13307392, 'steps': 25990, 'loss/train': 0.5687598586082458} 02/25/2022 11:15:11 - INFO - codeparrot_training - Step 25991: {'lr': 0.0002501472621471209, 'samples': 13307904, 'steps': 25991, 'loss/train': 0.9630624651908875} 02/25/2022 11:15:16 - INFO - codeparrot_training - Step 25992: {'lr': 0.0002501308996879184, 'samples': 13308416, 'steps': 25992, 'loss/train': 1.777406930923462} 02/25/2022 11:15:20 - INFO - codeparrot_training - Step 25993: {'lr': 0.00025011453722815526, 'samples': 13308928, 'steps': 25993, 'loss/train': 1.7278110980987549} 02/25/2022 11:15:25 - INFO - codeparrot_training - Step 25994: {'lr': 0.0002500981747679014, 'samples': 13309440, 'steps': 25994, 'loss/train': 2.7779786586761475} 02/25/2022 11:15:28 - INFO - codeparrot_training - Step 25995: {'lr': 0.00025008181230722705, 'samples': 13309952, 'steps': 25995, 'loss/train': 3.2408502101898193} 02/25/2022 11:15:36 - INFO - codeparrot_training - Step 25996: {'lr': 0.00025006544984620217, 'samples': 13310464, 'steps': 25996, 'loss/train': 2.555023193359375} 02/25/2022 11:15:39 - INFO - codeparrot_training - Step 25997: {'lr': 0.00025004908738489697, 'samples': 13310976, 'steps': 25997, 'loss/train': 0.3586549162864685} 02/25/2022 11:15:45 - INFO - codeparrot_training - Step 25998: {'lr': 0.0002500327249233815, 'samples': 13311488, 'steps': 25998, 'loss/train': 1.2565412521362305} 02/25/2022 11:15:48 - INFO - codeparrot_training - Step 25999: {'lr': 0.0002500163624617258, 'samples': 13312000, 'steps': 25999, 'loss/train': 1.7097930908203125} 02/25/2022 11:15:48 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 11:16:06 - WARNING - huggingface_hub.repository - Several commits (26) will be pushed upstream. 02/25/2022 11:16:06 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 11:16:40 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 5549929..86c4c54 floral-grass-11 -> floral-grass-11 02/25/2022 11:16:47 - INFO - codeparrot_training - Step 26000: {'lr': 0.00025, 'samples': 13312512, 'steps': 26000, 'loss/train': 0.37809640169143677} 02/25/2022 11:16:51 - INFO - codeparrot_training - Step 26001: {'lr': 0.00024998363753827424, 'samples': 13313024, 'steps': 26001, 'loss/train': 1.5100133419036865} 02/25/2022 11:16:56 - INFO - codeparrot_training - Step 26002: {'lr': 0.0002499672750766185, 'samples': 13313536, 'steps': 26002, 'loss/train': 2.3396451473236084} 02/25/2022 11:17:00 - INFO - codeparrot_training - Step 26003: {'lr': 0.0002499509126151031, 'samples': 13314048, 'steps': 26003, 'loss/train': 1.9778226613998413} 02/25/2022 11:17:05 - INFO - codeparrot_training - Step 26004: {'lr': 0.0002499345501537979, 'samples': 13314560, 'steps': 26004, 'loss/train': 0.13023696839809418} 02/25/2022 11:17:09 - INFO - codeparrot_training - Step 26005: {'lr': 0.000249918187692773, 'samples': 13315072, 'steps': 26005, 'loss/train': 2.3946428298950195} 02/25/2022 11:17:14 - INFO - codeparrot_training - Step 26006: {'lr': 0.0002499018252320986, 'samples': 13315584, 'steps': 26006, 'loss/train': 1.8516818284988403} 02/25/2022 11:17:18 - INFO - codeparrot_training - Step 26007: {'lr': 0.0002498854627718448, 'samples': 13316096, 'steps': 26007, 'loss/train': 2.227653980255127} 02/25/2022 11:17:24 - INFO - codeparrot_training - Step 26008: {'lr': 0.0002498691003120816, 'samples': 13316608, 'steps': 26008, 'loss/train': 0.6200409531593323} 02/25/2022 11:17:27 - INFO - codeparrot_training - Step 26009: {'lr': 0.0002498527378528791, 'samples': 13317120, 'steps': 26009, 'loss/train': 1.5642789602279663} 02/25/2022 11:17:33 - INFO - codeparrot_training - Step 26010: {'lr': 0.0002498363753943074, 'samples': 13317632, 'steps': 26010, 'loss/train': 1.069700002670288} 02/25/2022 11:17:36 - INFO - codeparrot_training - Step 26011: {'lr': 0.0002498200129364368, 'samples': 13318144, 'steps': 26011, 'loss/train': 1.6406219005584717} 02/25/2022 11:17:43 - INFO - codeparrot_training - Step 26012: {'lr': 0.00024980365047933705, 'samples': 13318656, 'steps': 26012, 'loss/train': 1.312057614326477} 02/25/2022 11:17:47 - INFO - codeparrot_training - Step 26013: {'lr': 0.0002497872880230784, 'samples': 13319168, 'steps': 26013, 'loss/train': 2.089871406555176} 02/25/2022 11:17:52 - INFO - codeparrot_training - Step 26014: {'lr': 0.000249770925567731, 'samples': 13319680, 'steps': 26014, 'loss/train': 1.568368911743164} 02/25/2022 11:17:56 - INFO - codeparrot_training - Step 26015: {'lr': 0.00024975456311336484, 'samples': 13320192, 'steps': 26015, 'loss/train': 2.7935009002685547} 02/25/2022 11:18:01 - INFO - codeparrot_training - Step 26016: {'lr': 0.00024973820066005005, 'samples': 13320704, 'steps': 26016, 'loss/train': 2.1883440017700195} 02/25/2022 11:18:05 - INFO - codeparrot_training - Step 26017: {'lr': 0.00024972183820785675, 'samples': 13321216, 'steps': 26017, 'loss/train': 2.4874684810638428} 02/25/2022 11:18:10 - INFO - codeparrot_training - Step 26018: {'lr': 0.00024970547575685494, 'samples': 13321728, 'steps': 26018, 'loss/train': 1.6686476469039917} 02/25/2022 11:18:14 - INFO - codeparrot_training - Step 26019: {'lr': 0.00024968911330711487, 'samples': 13322240, 'steps': 26019, 'loss/train': 1.7602189779281616} 02/25/2022 11:18:19 - INFO - codeparrot_training - Step 26020: {'lr': 0.00024967275085870653, 'samples': 13322752, 'steps': 26020, 'loss/train': 1.928260087966919} 02/25/2022 11:18:23 - INFO - codeparrot_training - Step 26021: {'lr': 0.0002496563884117, 'samples': 13323264, 'steps': 26021, 'loss/train': 2.0191547870635986} 02/25/2022 11:18:30 - INFO - codeparrot_training - Step 26022: {'lr': 0.00024964002596616544, 'samples': 13323776, 'steps': 26022, 'loss/train': 1.2096259593963623} 02/25/2022 11:18:34 - INFO - codeparrot_training - Step 26023: {'lr': 0.0002496236635221728, 'samples': 13324288, 'steps': 26023, 'loss/train': 1.2350517511367798} 02/25/2022 11:18:39 - INFO - codeparrot_training - Step 26024: {'lr': 0.00024960730107979233, 'samples': 13324800, 'steps': 26024, 'loss/train': 3.2793495655059814} 02/25/2022 11:18:43 - INFO - codeparrot_training - Step 26025: {'lr': 0.000249590938639094, 'samples': 13325312, 'steps': 26025, 'loss/train': 2.6286778450012207} 02/25/2022 11:18:48 - INFO - codeparrot_training - Step 26026: {'lr': 0.00024957457620014805, 'samples': 13325824, 'steps': 26026, 'loss/train': 2.493715524673462} 02/25/2022 11:18:52 - INFO - codeparrot_training - Step 26027: {'lr': 0.00024955821376302436, 'samples': 13326336, 'steps': 26027, 'loss/train': 1.7127689123153687} 02/25/2022 11:18:57 - INFO - codeparrot_training - Step 26028: {'lr': 0.0002495418513277932, 'samples': 13326848, 'steps': 26028, 'loss/train': 2.8942880630493164} 02/25/2022 11:19:01 - INFO - codeparrot_training - Step 26029: {'lr': 0.0002495254888945247, 'samples': 13327360, 'steps': 26029, 'loss/train': 1.8060437440872192} 02/25/2022 11:19:06 - INFO - codeparrot_training - Step 26030: {'lr': 0.00024950912646328876, 'samples': 13327872, 'steps': 26030, 'loss/train': 0.9384188055992126} 02/25/2022 11:19:10 - INFO - codeparrot_training - Step 26031: {'lr': 0.0002494927640341555, 'samples': 13328384, 'steps': 26031, 'loss/train': 1.7117457389831543} 02/25/2022 11:19:17 - INFO - codeparrot_training - Step 26032: {'lr': 0.00024947640160719514, 'samples': 13328896, 'steps': 26032, 'loss/train': 1.9165449142456055} 02/25/2022 11:19:20 - INFO - codeparrot_training - Step 26033: {'lr': 0.0002494600391824778, 'samples': 13329408, 'steps': 26033, 'loss/train': 1.6713306903839111} 02/25/2022 11:19:26 - INFO - codeparrot_training - Step 26034: {'lr': 0.0002494436767600734, 'samples': 13329920, 'steps': 26034, 'loss/train': 3.214975118637085} 02/25/2022 11:19:30 - INFO - codeparrot_training - Step 26035: {'lr': 0.00024942731434005207, 'samples': 13330432, 'steps': 26035, 'loss/train': 0.9205017685890198} 02/25/2022 11:19:35 - INFO - codeparrot_training - Step 26036: {'lr': 0.00024941095192248397, 'samples': 13330944, 'steps': 26036, 'loss/train': 1.7402057647705078} 02/25/2022 11:19:39 - INFO - codeparrot_training - Step 26037: {'lr': 0.0002493945895074391, 'samples': 13331456, 'steps': 26037, 'loss/train': 2.598801851272583} 02/25/2022 11:19:44 - INFO - codeparrot_training - Step 26038: {'lr': 0.00024937822709498786, 'samples': 13331968, 'steps': 26038, 'loss/train': 2.4083430767059326} 02/25/2022 11:19:48 - INFO - codeparrot_training - Step 26039: {'lr': 0.0002493618646851999, 'samples': 13332480, 'steps': 26039, 'loss/train': 1.244691014289856} 02/25/2022 11:19:53 - INFO - codeparrot_training - Step 26040: {'lr': 0.0002493455022781455, 'samples': 13332992, 'steps': 26040, 'loss/train': 3.324190139770508} 02/25/2022 11:19:57 - INFO - codeparrot_training - Step 26041: {'lr': 0.00024932913987389476, 'samples': 13333504, 'steps': 26041, 'loss/train': 1.1800880432128906} 02/25/2022 11:20:04 - INFO - codeparrot_training - Step 26042: {'lr': 0.0002493127774725179, 'samples': 13334016, 'steps': 26042, 'loss/train': 0.3366134762763977} 02/25/2022 11:20:07 - INFO - codeparrot_training - Step 26043: {'lr': 0.0002492964150740848, 'samples': 13334528, 'steps': 26043, 'loss/train': 1.7161277532577515} 02/25/2022 11:20:13 - INFO - codeparrot_training - Step 26044: {'lr': 0.00024928005267866563, 'samples': 13335040, 'steps': 26044, 'loss/train': 1.6405366659164429} 02/25/2022 11:20:16 - INFO - codeparrot_training - Step 26045: {'lr': 0.00024926369028633043, 'samples': 13335552, 'steps': 26045, 'loss/train': 2.173854351043701} 02/25/2022 11:20:22 - INFO - codeparrot_training - Step 26046: {'lr': 0.0002492473278971495, 'samples': 13336064, 'steps': 26046, 'loss/train': 2.255094051361084} 02/25/2022 11:20:26 - INFO - codeparrot_training - Step 26047: {'lr': 0.00024923096551119267, 'samples': 13336576, 'steps': 26047, 'loss/train': 2.2586750984191895} 02/25/2022 11:20:31 - INFO - codeparrot_training - Step 26048: {'lr': 0.0002492146031285301, 'samples': 13337088, 'steps': 26048, 'loss/train': 2.7552316188812256} 02/25/2022 11:20:35 - INFO - codeparrot_training - Step 26049: {'lr': 0.000249198240749232, 'samples': 13337600, 'steps': 26049, 'loss/train': 2.1905179023742676} 02/25/2022 11:20:40 - INFO - codeparrot_training - Step 26050: {'lr': 0.0002491818783733683, 'samples': 13338112, 'steps': 26050, 'loss/train': 2.317699909210205} 02/25/2022 11:20:44 - INFO - codeparrot_training - Step 26051: {'lr': 0.0002491655160010093, 'samples': 13338624, 'steps': 26051, 'loss/train': 1.9859191179275513} 02/25/2022 11:20:49 - INFO - codeparrot_training - Step 26052: {'lr': 0.0002491491536322249, 'samples': 13339136, 'steps': 26052, 'loss/train': 1.8399955034255981} 02/25/2022 11:20:53 - INFO - codeparrot_training - Step 26053: {'lr': 0.0002491327912670852, 'samples': 13339648, 'steps': 26053, 'loss/train': 1.77420175075531} 02/25/2022 11:20:59 - INFO - codeparrot_training - Step 26054: {'lr': 0.0002491164289056604, 'samples': 13340160, 'steps': 26054, 'loss/train': 2.473858118057251} 02/25/2022 11:21:02 - INFO - codeparrot_training - Step 26055: {'lr': 0.0002491000665480206, 'samples': 13340672, 'steps': 26055, 'loss/train': 2.050698757171631} 02/25/2022 11:21:07 - INFO - codeparrot_training - Step 26056: {'lr': 0.00024908370419423573, 'samples': 13341184, 'steps': 26056, 'loss/train': 1.9810444116592407} 02/25/2022 11:21:11 - INFO - codeparrot_training - Step 26057: {'lr': 0.000249067341844376, 'samples': 13341696, 'steps': 26057, 'loss/train': 1.2977389097213745} 02/25/2022 11:21:18 - INFO - codeparrot_training - Step 26058: {'lr': 0.00024905097949851144, 'samples': 13342208, 'steps': 26058, 'loss/train': 2.452275037765503} 02/25/2022 11:21:22 - INFO - codeparrot_training - Step 26059: {'lr': 0.0002490346171567124, 'samples': 13342720, 'steps': 26059, 'loss/train': 2.565781831741333} 02/25/2022 11:21:28 - INFO - codeparrot_training - Step 26060: {'lr': 0.0002490182548190485, 'samples': 13343232, 'steps': 26060, 'loss/train': 2.2402310371398926} 02/25/2022 11:21:31 - INFO - codeparrot_training - Step 26061: {'lr': 0.0002490018924855902, 'samples': 13343744, 'steps': 26061, 'loss/train': 1.701701283454895} 02/25/2022 11:21:37 - INFO - codeparrot_training - Step 26062: {'lr': 0.00024898553015640745, 'samples': 13344256, 'steps': 26062, 'loss/train': 1.6444568634033203} 02/25/2022 11:21:40 - INFO - codeparrot_training - Step 26063: {'lr': 0.00024896916783157035, 'samples': 13344768, 'steps': 26063, 'loss/train': 1.571385145187378} 02/25/2022 11:21:46 - INFO - codeparrot_training - Step 26064: {'lr': 0.0002489528055111491, 'samples': 13345280, 'steps': 26064, 'loss/train': 1.5612801313400269} 02/25/2022 11:21:49 - INFO - codeparrot_training - Step 26065: {'lr': 0.00024893644319521355, 'samples': 13345792, 'steps': 26065, 'loss/train': 1.6536020040512085} 02/25/2022 11:21:55 - INFO - codeparrot_training - Step 26066: {'lr': 0.00024892008088383405, 'samples': 13346304, 'steps': 26066, 'loss/train': 0.6262559294700623} 02/25/2022 11:21:58 - INFO - codeparrot_training - Step 26067: {'lr': 0.0002489037185770805, 'samples': 13346816, 'steps': 26067, 'loss/train': 1.5273154973983765} 02/25/2022 11:22:05 - INFO - codeparrot_training - Step 26068: {'lr': 0.0002488873562750232, 'samples': 13347328, 'steps': 26068, 'loss/train': 1.6312168836593628} 02/25/2022 11:22:09 - INFO - codeparrot_training - Step 26069: {'lr': 0.00024887099397773204, 'samples': 13347840, 'steps': 26069, 'loss/train': 1.3209587335586548} 02/25/2022 11:22:14 - INFO - codeparrot_training - Step 26070: {'lr': 0.0002488546316852771, 'samples': 13348352, 'steps': 26070, 'loss/train': 2.204545497894287} 02/25/2022 11:22:18 - INFO - codeparrot_training - Step 26071: {'lr': 0.00024883826939772866, 'samples': 13348864, 'steps': 26071, 'loss/train': 1.6728625297546387} 02/25/2022 11:22:23 - INFO - codeparrot_training - Step 26072: {'lr': 0.0002488219071151567, 'samples': 13349376, 'steps': 26072, 'loss/train': 1.9980379343032837} 02/25/2022 11:22:27 - INFO - codeparrot_training - Step 26073: {'lr': 0.0002488055448376313, 'samples': 13349888, 'steps': 26073, 'loss/train': 2.60790753364563} 02/25/2022 11:22:32 - INFO - codeparrot_training - Step 26074: {'lr': 0.00024878918256522256, 'samples': 13350400, 'steps': 26074, 'loss/train': 1.8955438137054443} 02/25/2022 11:22:36 - INFO - codeparrot_training - Step 26075: {'lr': 0.0002487728202980005, 'samples': 13350912, 'steps': 26075, 'loss/train': 1.9609507322311401} 02/25/2022 11:22:41 - INFO - codeparrot_training - Step 26076: {'lr': 0.00024875645803603536, 'samples': 13351424, 'steps': 26076, 'loss/train': 2.889387845993042} 02/25/2022 11:22:45 - INFO - codeparrot_training - Step 26077: {'lr': 0.0002487400957793972, 'samples': 13351936, 'steps': 26077, 'loss/train': 1.2577725648880005} 02/25/2022 11:22:52 - INFO - codeparrot_training - Step 26078: {'lr': 0.00024872373352815603, 'samples': 13352448, 'steps': 26078, 'loss/train': 1.705277919769287} 02/25/2022 11:22:56 - INFO - codeparrot_training - Step 26079: {'lr': 0.000248707371282382, 'samples': 13352960, 'steps': 26079, 'loss/train': 1.1163451671600342} 02/25/2022 11:23:01 - INFO - codeparrot_training - Step 26080: {'lr': 0.00024869100904214507, 'samples': 13353472, 'steps': 26080, 'loss/train': 1.8644014596939087} 02/25/2022 11:23:04 - INFO - codeparrot_training - Step 26081: {'lr': 0.00024867464680751564, 'samples': 13353984, 'steps': 26081, 'loss/train': 1.5341618061065674} 02/25/2022 11:23:10 - INFO - codeparrot_training - Step 26082: {'lr': 0.0002486582845785635, 'samples': 13354496, 'steps': 26082, 'loss/train': 2.293531656265259} 02/25/2022 11:23:13 - INFO - codeparrot_training - Step 26083: {'lr': 0.0002486419223553588, 'samples': 13355008, 'steps': 26083, 'loss/train': 1.3669458627700806} 02/25/2022 11:23:19 - INFO - codeparrot_training - Step 26084: {'lr': 0.00024862556013797164, 'samples': 13355520, 'steps': 26084, 'loss/train': 0.8123820424079895} 02/25/2022 11:23:22 - INFO - codeparrot_training - Step 26085: {'lr': 0.0002486091979264722, 'samples': 13356032, 'steps': 26085, 'loss/train': 0.5838306546211243} 02/25/2022 11:23:28 - INFO - codeparrot_training - Step 26086: {'lr': 0.0002485928357209306, 'samples': 13356544, 'steps': 26086, 'loss/train': 1.9058812856674194} 02/25/2022 11:23:31 - INFO - codeparrot_training - Step 26087: {'lr': 0.00024857647352141677, 'samples': 13357056, 'steps': 26087, 'loss/train': 1.8455145359039307} 02/25/2022 11:23:39 - INFO - codeparrot_training - Step 26088: {'lr': 0.00024856011132800085, 'samples': 13357568, 'steps': 26088, 'loss/train': 0.8172327280044556} 02/25/2022 11:23:42 - INFO - codeparrot_training - Step 26089: {'lr': 0.00024854374914075295, 'samples': 13358080, 'steps': 26089, 'loss/train': 0.4047224223613739} 02/25/2022 11:23:48 - INFO - codeparrot_training - Step 26090: {'lr': 0.0002485273869597433, 'samples': 13358592, 'steps': 26090, 'loss/train': 0.8914444446563721} 02/25/2022 11:23:51 - INFO - codeparrot_training - Step 26091: {'lr': 0.00024851102478504173, 'samples': 13359104, 'steps': 26091, 'loss/train': 2.2594966888427734} 02/25/2022 11:23:57 - INFO - codeparrot_training - Step 26092: {'lr': 0.0002484946626167185, 'samples': 13359616, 'steps': 26092, 'loss/train': 1.4258530139923096} 02/25/2022 11:24:00 - INFO - codeparrot_training - Step 26093: {'lr': 0.00024847830045484357, 'samples': 13360128, 'steps': 26093, 'loss/train': 2.382063150405884} 02/25/2022 11:24:06 - INFO - codeparrot_training - Step 26094: {'lr': 0.0002484619382994873, 'samples': 13360640, 'steps': 26094, 'loss/train': 2.10003924369812} 02/25/2022 11:24:09 - INFO - codeparrot_training - Step 26095: {'lr': 0.00024844557615071944, 'samples': 13361152, 'steps': 26095, 'loss/train': 1.9866585731506348} 02/25/2022 11:24:15 - INFO - codeparrot_training - Step 26096: {'lr': 0.00024842921400861025, 'samples': 13361664, 'steps': 26096, 'loss/train': 2.330418109893799} 02/25/2022 11:24:18 - INFO - codeparrot_training - Step 26097: {'lr': 0.0002484128518732298, 'samples': 13362176, 'steps': 26097, 'loss/train': 2.1247408390045166} 02/25/2022 11:24:26 - INFO - codeparrot_training - Step 26098: {'lr': 0.0002483964897446482, 'samples': 13362688, 'steps': 26098, 'loss/train': 0.8035333156585693} 02/25/2022 11:24:29 - INFO - codeparrot_training - Step 26099: {'lr': 0.0002483801276229357, 'samples': 13363200, 'steps': 26099, 'loss/train': 0.8666520118713379} 02/25/2022 11:24:35 - INFO - codeparrot_training - Step 26100: {'lr': 0.00024836376550816205, 'samples': 13363712, 'steps': 26100, 'loss/train': 0.7444463968276978} 02/25/2022 11:24:38 - INFO - codeparrot_training - Step 26101: {'lr': 0.0002483474034003975, 'samples': 13364224, 'steps': 26101, 'loss/train': 0.4754897654056549} 02/25/2022 11:24:44 - INFO - codeparrot_training - Step 26102: {'lr': 0.00024833104129971226, 'samples': 13364736, 'steps': 26102, 'loss/train': 1.634103775024414} 02/25/2022 11:24:47 - INFO - codeparrot_training - Step 26103: {'lr': 0.00024831467920617624, 'samples': 13365248, 'steps': 26103, 'loss/train': 2.524019241333008} 02/25/2022 11:24:53 - INFO - codeparrot_training - Step 26104: {'lr': 0.00024829831711985955, 'samples': 13365760, 'steps': 26104, 'loss/train': 1.8770508766174316} 02/25/2022 11:24:56 - INFO - codeparrot_training - Step 26105: {'lr': 0.0002482819550408324, 'samples': 13366272, 'steps': 26105, 'loss/train': 2.1426119804382324} 02/25/2022 11:25:02 - INFO - codeparrot_training - Step 26106: {'lr': 0.0002482655929691648, 'samples': 13366784, 'steps': 26106, 'loss/train': 2.0250260829925537} 02/25/2022 11:25:05 - INFO - codeparrot_training - Step 26107: {'lr': 0.0002482492309049268, 'samples': 13367296, 'steps': 26107, 'loss/train': 0.9565497636795044} 02/25/2022 11:25:11 - INFO - codeparrot_training - Step 26108: {'lr': 0.0002482328688481886, 'samples': 13367808, 'steps': 26108, 'loss/train': 1.8798900842666626} 02/25/2022 11:25:14 - INFO - codeparrot_training - Step 26109: {'lr': 0.0002482165067990202, 'samples': 13368320, 'steps': 26109, 'loss/train': 2.214495897293091} 02/25/2022 11:25:20 - INFO - codeparrot_training - Step 26110: {'lr': 0.0002482001447574917, 'samples': 13368832, 'steps': 26110, 'loss/train': 1.5836482048034668} 02/25/2022 11:25:23 - INFO - codeparrot_training - Step 26111: {'lr': 0.0002481837827236732, 'samples': 13369344, 'steps': 26111, 'loss/train': 1.7975934743881226} 02/25/2022 11:25:29 - INFO - codeparrot_training - Step 26112: {'lr': 0.00024816742069763486, 'samples': 13369856, 'steps': 26112, 'loss/train': 1.5957379341125488} 02/25/2022 11:25:32 - INFO - codeparrot_training - Step 26113: {'lr': 0.0002481510586794467, 'samples': 13370368, 'steps': 26113, 'loss/train': 2.000986099243164} 02/25/2022 11:25:39 - INFO - codeparrot_training - Step 26114: {'lr': 0.0002481346966691788, 'samples': 13370880, 'steps': 26114, 'loss/train': 2.064115047454834} 02/25/2022 11:25:43 - INFO - codeparrot_training - Step 26115: {'lr': 0.0002481183346669012, 'samples': 13371392, 'steps': 26115, 'loss/train': 3.854135513305664} 02/25/2022 11:25:48 - INFO - codeparrot_training - Step 26116: {'lr': 0.0002481019726726842, 'samples': 13371904, 'steps': 26116, 'loss/train': 2.0776100158691406} 02/25/2022 11:25:52 - INFO - codeparrot_training - Step 26117: {'lr': 0.00024808561068659764, 'samples': 13372416, 'steps': 26117, 'loss/train': 0.347175657749176} 02/25/2022 11:25:57 - INFO - codeparrot_training - Step 26118: {'lr': 0.00024806924870871173, 'samples': 13372928, 'steps': 26118, 'loss/train': 2.3343749046325684} 02/25/2022 11:26:01 - INFO - codeparrot_training - Step 26119: {'lr': 0.00024805288673909656, 'samples': 13373440, 'steps': 26119, 'loss/train': 1.73350989818573} 02/25/2022 11:26:06 - INFO - codeparrot_training - Step 26120: {'lr': 0.00024803652477782225, 'samples': 13373952, 'steps': 26120, 'loss/train': 1.7164617776870728} 02/25/2022 11:26:10 - INFO - codeparrot_training - Step 26121: {'lr': 0.00024802016282495876, 'samples': 13374464, 'steps': 26121, 'loss/train': 1.5382933616638184} 02/25/2022 11:26:15 - INFO - codeparrot_training - Step 26122: {'lr': 0.00024800380088057627, 'samples': 13374976, 'steps': 26122, 'loss/train': 2.393277645111084} 02/25/2022 11:26:19 - INFO - codeparrot_training - Step 26123: {'lr': 0.0002479874389447449, 'samples': 13375488, 'steps': 26123, 'loss/train': 2.6691489219665527} 02/25/2022 11:26:26 - INFO - codeparrot_training - Step 26124: {'lr': 0.00024797107701753464, 'samples': 13376000, 'steps': 26124, 'loss/train': 1.599758267402649} 02/25/2022 11:26:30 - INFO - codeparrot_training - Step 26125: {'lr': 0.0002479547150990158, 'samples': 13376512, 'steps': 26125, 'loss/train': 2.0483505725860596} 02/25/2022 11:26:35 - INFO - codeparrot_training - Step 26126: {'lr': 0.0002479383531892582, 'samples': 13377024, 'steps': 26126, 'loss/train': 1.5155377388000488} 02/25/2022 11:26:38 - INFO - codeparrot_training - Step 26127: {'lr': 0.000247921991288332, 'samples': 13377536, 'steps': 26127, 'loss/train': 2.8289942741394043} 02/25/2022 11:26:44 - INFO - codeparrot_training - Step 26128: {'lr': 0.00024790562939630735, 'samples': 13378048, 'steps': 26128, 'loss/train': 1.9080466032028198} 02/25/2022 11:26:48 - INFO - codeparrot_training - Step 26129: {'lr': 0.00024788926751325444, 'samples': 13378560, 'steps': 26129, 'loss/train': 2.094698667526245} 02/25/2022 11:26:53 - INFO - codeparrot_training - Step 26130: {'lr': 0.00024787290563924307, 'samples': 13379072, 'steps': 26130, 'loss/train': 1.41063392162323} 02/25/2022 11:26:57 - INFO - codeparrot_training - Step 26131: {'lr': 0.00024785654377434355, 'samples': 13379584, 'steps': 26131, 'loss/train': 2.8320565223693848} 02/25/2022 11:27:02 - INFO - codeparrot_training - Step 26132: {'lr': 0.00024784018191862593, 'samples': 13380096, 'steps': 26132, 'loss/train': 1.170541524887085} 02/25/2022 11:27:06 - INFO - codeparrot_training - Step 26133: {'lr': 0.00024782382007216034, 'samples': 13380608, 'steps': 26133, 'loss/train': 1.952217936515808} 02/25/2022 11:27:11 - INFO - codeparrot_training - Step 26134: {'lr': 0.0002478074582350168, 'samples': 13381120, 'steps': 26134, 'loss/train': 1.4973058700561523} 02/25/2022 11:27:15 - INFO - codeparrot_training - Step 26135: {'lr': 0.0002477910964072653, 'samples': 13381632, 'steps': 26135, 'loss/train': 1.983148455619812} 02/25/2022 11:27:22 - INFO - codeparrot_training - Step 26136: {'lr': 0.0002477747345889761, 'samples': 13382144, 'steps': 26136, 'loss/train': 1.711746096611023} 02/25/2022 11:27:25 - INFO - codeparrot_training - Step 26137: {'lr': 0.00024775837278021923, 'samples': 13382656, 'steps': 26137, 'loss/train': 1.007331132888794} 02/25/2022 11:27:31 - INFO - codeparrot_training - Step 26138: {'lr': 0.00024774201098106487, 'samples': 13383168, 'steps': 26138, 'loss/train': 2.1550087928771973} 02/25/2022 11:27:34 - INFO - codeparrot_training - Step 26139: {'lr': 0.0002477256491915829, 'samples': 13383680, 'steps': 26139, 'loss/train': 2.03678035736084} 02/25/2022 11:27:40 - INFO - codeparrot_training - Step 26140: {'lr': 0.00024770928741184355, 'samples': 13384192, 'steps': 26140, 'loss/train': 1.0919735431671143} 02/25/2022 11:27:44 - INFO - codeparrot_training - Step 26141: {'lr': 0.0002476929256419169, 'samples': 13384704, 'steps': 26141, 'loss/train': 1.9188615083694458} 02/25/2022 11:27:49 - INFO - codeparrot_training - Step 26142: {'lr': 0.0002476765638818731, 'samples': 13385216, 'steps': 26142, 'loss/train': 2.8201305866241455} 02/25/2022 11:27:53 - INFO - codeparrot_training - Step 26143: {'lr': 0.0002476602021317821, 'samples': 13385728, 'steps': 26143, 'loss/train': 2.4097628593444824} 02/25/2022 11:27:58 - INFO - codeparrot_training - Step 26144: {'lr': 0.000247643840391714, 'samples': 13386240, 'steps': 26144, 'loss/train': 1.867837905883789} 02/25/2022 11:28:01 - INFO - codeparrot_training - Step 26145: {'lr': 0.000247627478661739, 'samples': 13386752, 'steps': 26145, 'loss/train': 1.6900721788406372} 02/25/2022 11:28:09 - INFO - codeparrot_training - Step 26146: {'lr': 0.0002476111169419271, 'samples': 13387264, 'steps': 26146, 'loss/train': 1.1938014030456543} 02/25/2022 11:28:12 - INFO - codeparrot_training - Step 26147: {'lr': 0.00024759475523234846, 'samples': 13387776, 'steps': 26147, 'loss/train': 2.093402862548828} 02/25/2022 11:28:18 - INFO - codeparrot_training - Step 26148: {'lr': 0.0002475783935330731, 'samples': 13388288, 'steps': 26148, 'loss/train': 1.6534773111343384} 02/25/2022 11:28:21 - INFO - codeparrot_training - Step 26149: {'lr': 0.0002475620318441711, 'samples': 13388800, 'steps': 26149, 'loss/train': 2.011162281036377} 02/25/2022 11:28:27 - INFO - codeparrot_training - Step 26150: {'lr': 0.0002475456701657126, 'samples': 13389312, 'steps': 26150, 'loss/train': 1.3796783685684204} 02/25/2022 11:28:30 - INFO - codeparrot_training - Step 26151: {'lr': 0.0002475293084977677, 'samples': 13389824, 'steps': 26151, 'loss/train': 2.144446849822998} 02/25/2022 11:28:36 - INFO - codeparrot_training - Step 26152: {'lr': 0.00024751294684040645, 'samples': 13390336, 'steps': 26152, 'loss/train': 2.7847137451171875} 02/25/2022 11:28:39 - INFO - codeparrot_training - Step 26153: {'lr': 0.00024749658519369894, 'samples': 13390848, 'steps': 26153, 'loss/train': 2.33437180519104} 02/25/2022 11:28:43 - INFO - codeparrot_training - Step 26154: {'lr': 0.0002474802235577152, 'samples': 13391360, 'steps': 26154, 'loss/train': 1.6049247980117798} 02/25/2022 11:28:48 - INFO - codeparrot_training - Step 26155: {'lr': 0.0002474638619325255, 'samples': 13391872, 'steps': 26155, 'loss/train': 1.2048485279083252} 02/25/2022 11:28:52 - INFO - codeparrot_training - Step 26156: {'lr': 0.0002474475003181997, 'samples': 13392384, 'steps': 26156, 'loss/train': 0.9827079176902771} 02/25/2022 11:28:58 - INFO - codeparrot_training - Step 26157: {'lr': 0.0002474311387148081, 'samples': 13392896, 'steps': 26157, 'loss/train': 2.077143430709839} 02/25/2022 11:29:01 - INFO - codeparrot_training - Step 26158: {'lr': 0.00024741477712242056, 'samples': 13393408, 'steps': 26158, 'loss/train': 1.3191636800765991} 02/25/2022 11:29:06 - INFO - codeparrot_training - Step 26159: {'lr': 0.00024739841554110735, 'samples': 13393920, 'steps': 26159, 'loss/train': 1.5726323127746582} 02/25/2022 11:29:10 - INFO - codeparrot_training - Step 26160: {'lr': 0.00024738205397093865, 'samples': 13394432, 'steps': 26160, 'loss/train': 1.984808087348938} 02/25/2022 11:29:17 - INFO - codeparrot_training - Step 26161: {'lr': 0.00024736569241198424, 'samples': 13394944, 'steps': 26161, 'loss/train': 1.4291894435882568} 02/25/2022 11:29:23 - INFO - codeparrot_training - Step 26162: {'lr': 0.00024734933086431436, 'samples': 13395456, 'steps': 26162, 'loss/train': 1.5518579483032227} 02/25/2022 11:29:26 - INFO - codeparrot_training - Step 26163: {'lr': 0.0002473329693279991, 'samples': 13395968, 'steps': 26163, 'loss/train': 2.2656655311584473} 02/25/2022 11:29:32 - INFO - codeparrot_training - Step 26164: {'lr': 0.00024731660780310865, 'samples': 13396480, 'steps': 26164, 'loss/train': 2.4494128227233887} 02/25/2022 11:29:35 - INFO - codeparrot_training - Step 26165: {'lr': 0.0002473002462897129, 'samples': 13396992, 'steps': 26165, 'loss/train': 2.3789594173431396} 02/25/2022 11:29:41 - INFO - codeparrot_training - Step 26166: {'lr': 0.0002472838847878821, 'samples': 13397504, 'steps': 26166, 'loss/train': 2.240251302719116} 02/25/2022 11:29:44 - INFO - codeparrot_training - Step 26167: {'lr': 0.00024726752329768623, 'samples': 13398016, 'steps': 26167, 'loss/train': 3.5440189838409424} 02/25/2022 11:29:50 - INFO - codeparrot_training - Step 26168: {'lr': 0.0002472511618191955, 'samples': 13398528, 'steps': 26168, 'loss/train': 0.8982509970664978} 02/25/2022 11:29:53 - INFO - codeparrot_training - Step 26169: {'lr': 0.00024723480035247986, 'samples': 13399040, 'steps': 26169, 'loss/train': 1.172585129737854} 02/25/2022 11:29:59 - INFO - codeparrot_training - Step 26170: {'lr': 0.00024721843889760945, 'samples': 13399552, 'steps': 26170, 'loss/train': 2.0878639221191406} 02/25/2022 11:30:02 - INFO - codeparrot_training - Step 26171: {'lr': 0.0002472020774546543, 'samples': 13400064, 'steps': 26171, 'loss/train': 3.1603128910064697} 02/25/2022 11:30:10 - INFO - codeparrot_training - Step 26172: {'lr': 0.00024718571602368465, 'samples': 13400576, 'steps': 26172, 'loss/train': 0.9676345586776733} 02/25/2022 11:30:13 - INFO - codeparrot_training - Step 26173: {'lr': 0.00024716935460477056, 'samples': 13401088, 'steps': 26173, 'loss/train': 2.520599126815796} 02/25/2022 11:30:19 - INFO - codeparrot_training - Step 26174: {'lr': 0.00024715299319798197, 'samples': 13401600, 'steps': 26174, 'loss/train': 3.573441505432129} 02/25/2022 11:30:22 - INFO - codeparrot_training - Step 26175: {'lr': 0.000247136631803389, 'samples': 13402112, 'steps': 26175, 'loss/train': 1.077135443687439} 02/25/2022 11:30:28 - INFO - codeparrot_training - Step 26176: {'lr': 0.0002471202704210619, 'samples': 13402624, 'steps': 26176, 'loss/train': 1.184415578842163} 02/25/2022 11:30:31 - INFO - codeparrot_training - Step 26177: {'lr': 0.0002471039090510707, 'samples': 13403136, 'steps': 26177, 'loss/train': 1.4721198081970215} 02/25/2022 11:30:37 - INFO - codeparrot_training - Step 26178: {'lr': 0.0002470875476934853, 'samples': 13403648, 'steps': 26178, 'loss/train': 1.9805235862731934} 02/25/2022 11:30:40 - INFO - codeparrot_training - Step 26179: {'lr': 0.00024707118634837593, 'samples': 13404160, 'steps': 26179, 'loss/train': 2.18442440032959} 02/25/2022 11:30:46 - INFO - codeparrot_training - Step 26180: {'lr': 0.0002470548250158127, 'samples': 13404672, 'steps': 26180, 'loss/train': 0.30659595131874084} 02/25/2022 11:30:49 - INFO - codeparrot_training - Step 26181: {'lr': 0.0002470384636958657, 'samples': 13405184, 'steps': 26181, 'loss/train': 1.3039799928665161} 02/25/2022 11:30:57 - INFO - codeparrot_training - Step 26182: {'lr': 0.0002470221023886049, 'samples': 13405696, 'steps': 26182, 'loss/train': 1.2231236696243286} 02/25/2022 11:31:00 - INFO - codeparrot_training - Step 26183: {'lr': 0.00024700574109410054, 'samples': 13406208, 'steps': 26183, 'loss/train': 1.3229402303695679} 02/25/2022 11:31:06 - INFO - codeparrot_training - Step 26184: {'lr': 0.00024698937981242254, 'samples': 13406720, 'steps': 26184, 'loss/train': 2.06085467338562} 02/25/2022 11:31:10 - INFO - codeparrot_training - Step 26185: {'lr': 0.00024697301854364117, 'samples': 13407232, 'steps': 26185, 'loss/train': 1.0619232654571533} 02/25/2022 11:31:15 - INFO - codeparrot_training - Step 26186: {'lr': 0.00024695665728782643, 'samples': 13407744, 'steps': 26186, 'loss/train': 1.3313326835632324} 02/25/2022 11:31:18 - INFO - codeparrot_training - Step 26187: {'lr': 0.0002469402960450484, 'samples': 13408256, 'steps': 26187, 'loss/train': 1.696230411529541} 02/25/2022 11:31:24 - INFO - codeparrot_training - Step 26188: {'lr': 0.0002469239348153771, 'samples': 13408768, 'steps': 26188, 'loss/train': 1.9529732465744019} 02/25/2022 11:31:27 - INFO - codeparrot_training - Step 26189: {'lr': 0.00024690757359888273, 'samples': 13409280, 'steps': 26189, 'loss/train': 1.9038735628128052} 02/25/2022 11:31:33 - INFO - codeparrot_training - Step 26190: {'lr': 0.0002468912123956354, 'samples': 13409792, 'steps': 26190, 'loss/train': 1.583982229232788} 02/25/2022 11:31:36 - INFO - codeparrot_training - Step 26191: {'lr': 0.00024687485120570505, 'samples': 13410304, 'steps': 26191, 'loss/train': 1.7699180841445923} 02/25/2022 11:31:42 - INFO - codeparrot_training - Step 26192: {'lr': 0.0002468584900291618, 'samples': 13410816, 'steps': 26192, 'loss/train': 0.6914160251617432} 02/25/2022 11:31:45 - INFO - codeparrot_training - Step 26193: {'lr': 0.0002468421288660759, 'samples': 13411328, 'steps': 26193, 'loss/train': 0.5209804773330688} 02/25/2022 11:31:51 - INFO - codeparrot_training - Step 26194: {'lr': 0.00024682576771651725, 'samples': 13411840, 'steps': 26194, 'loss/train': 1.2797808647155762} 02/25/2022 11:31:54 - INFO - codeparrot_training - Step 26195: {'lr': 0.000246809406580556, 'samples': 13412352, 'steps': 26195, 'loss/train': 1.3434840440750122} 02/25/2022 11:32:02 - INFO - codeparrot_training - Step 26196: {'lr': 0.00024679304545826224, 'samples': 13412864, 'steps': 26196, 'loss/train': 2.2019340991973877} 02/25/2022 11:32:06 - INFO - codeparrot_training - Step 26197: {'lr': 0.0002467766843497061, 'samples': 13413376, 'steps': 26197, 'loss/train': 1.9646128416061401} 02/25/2022 11:32:11 - INFO - codeparrot_training - Step 26198: {'lr': 0.0002467603232549576, 'samples': 13413888, 'steps': 26198, 'loss/train': 1.7637488842010498} 02/25/2022 11:32:15 - INFO - codeparrot_training - Step 26199: {'lr': 0.00024674396217408683, 'samples': 13414400, 'steps': 26199, 'loss/train': 0.3719959855079651} 02/25/2022 11:32:20 - INFO - codeparrot_training - Step 26200: {'lr': 0.00024672760110716396, 'samples': 13414912, 'steps': 26200, 'loss/train': 2.594597101211548} 02/25/2022 11:32:24 - INFO - codeparrot_training - Step 26201: {'lr': 0.0002467112400542589, 'samples': 13415424, 'steps': 26201, 'loss/train': 0.5499547123908997} 02/25/2022 11:32:29 - INFO - codeparrot_training - Step 26202: {'lr': 0.00024669487901544186, 'samples': 13415936, 'steps': 26202, 'loss/train': 2.2180166244506836} 02/25/2022 11:32:33 - INFO - codeparrot_training - Step 26203: {'lr': 0.000246678517990783, 'samples': 13416448, 'steps': 26203, 'loss/train': 1.4721049070358276} 02/25/2022 11:32:38 - INFO - codeparrot_training - Step 26204: {'lr': 0.00024666215698035225, 'samples': 13416960, 'steps': 26204, 'loss/train': 0.854203462600708} 02/25/2022 11:32:42 - INFO - codeparrot_training - Step 26205: {'lr': 0.00024664579598421976, 'samples': 13417472, 'steps': 26205, 'loss/train': 2.214334487915039} 02/25/2022 11:32:47 - INFO - codeparrot_training - Step 26206: {'lr': 0.00024662943500245555, 'samples': 13417984, 'steps': 26206, 'loss/train': 2.644735813140869} 02/25/2022 11:32:51 - INFO - codeparrot_training - Step 26207: {'lr': 0.0002466130740351298, 'samples': 13418496, 'steps': 26207, 'loss/train': 1.8945960998535156} 02/25/2022 11:32:58 - INFO - codeparrot_training - Step 26208: {'lr': 0.00024659671308231273, 'samples': 13419008, 'steps': 26208, 'loss/train': 0.8003060221672058} 02/25/2022 11:33:02 - INFO - codeparrot_training - Step 26209: {'lr': 0.0002465803521440741, 'samples': 13419520, 'steps': 26209, 'loss/train': 2.819276809692383} 02/25/2022 11:33:07 - INFO - codeparrot_training - Step 26210: {'lr': 0.00024656399122048415, 'samples': 13420032, 'steps': 26210, 'loss/train': 1.4604949951171875} 02/25/2022 11:33:11 - INFO - codeparrot_training - Step 26211: {'lr': 0.00024654763031161297, 'samples': 13420544, 'steps': 26211, 'loss/train': 1.8633092641830444} 02/25/2022 11:33:16 - INFO - codeparrot_training - Step 26212: {'lr': 0.0002465312694175308, 'samples': 13421056, 'steps': 26212, 'loss/train': 2.3561911582946777} 02/25/2022 11:33:20 - INFO - codeparrot_training - Step 26213: {'lr': 0.0002465149085383074, 'samples': 13421568, 'steps': 26213, 'loss/train': 1.8537622690200806} 02/25/2022 11:33:26 - INFO - codeparrot_training - Step 26214: {'lr': 0.00024649854767401307, 'samples': 13422080, 'steps': 26214, 'loss/train': 1.671692967414856} 02/25/2022 11:33:29 - INFO - codeparrot_training - Step 26215: {'lr': 0.0002464821868247178, 'samples': 13422592, 'steps': 26215, 'loss/train': 3.491947889328003} 02/25/2022 11:33:35 - INFO - codeparrot_training - Step 26216: {'lr': 0.0002464658259904919, 'samples': 13423104, 'steps': 26216, 'loss/train': 2.350614547729492} 02/25/2022 11:33:38 - INFO - codeparrot_training - Step 26217: {'lr': 0.0002464494651714051, 'samples': 13423616, 'steps': 26217, 'loss/train': 1.8202143907546997} 02/25/2022 11:33:46 - INFO - codeparrot_training - Step 26218: {'lr': 0.0002464331043675277, 'samples': 13424128, 'steps': 26218, 'loss/train': 1.8527570962905884} 02/25/2022 11:33:49 - INFO - codeparrot_training - Step 26219: {'lr': 0.0002464167435789298, 'samples': 13424640, 'steps': 26219, 'loss/train': 1.8715065717697144} 02/25/2022 11:33:55 - INFO - codeparrot_training - Step 26220: {'lr': 0.0002464003828056814, 'samples': 13425152, 'steps': 26220, 'loss/train': 2.158860445022583} 02/25/2022 11:33:58 - INFO - codeparrot_training - Step 26221: {'lr': 0.00024638402204785263, 'samples': 13425664, 'steps': 26221, 'loss/train': 1.5496877431869507} 02/25/2022 11:34:03 - INFO - codeparrot_training - Step 26222: {'lr': 0.00024636766130551354, 'samples': 13426176, 'steps': 26222, 'loss/train': 1.6209758520126343} 02/25/2022 11:34:07 - INFO - codeparrot_training - Step 26223: {'lr': 0.0002463513005787343, 'samples': 13426688, 'steps': 26223, 'loss/train': 1.8798103332519531} 02/25/2022 11:34:12 - INFO - codeparrot_training - Step 26224: {'lr': 0.0002463349398675848, 'samples': 13427200, 'steps': 26224, 'loss/train': 2.4048264026641846} 02/25/2022 11:34:16 - INFO - codeparrot_training - Step 26225: {'lr': 0.0002463185791721354, 'samples': 13427712, 'steps': 26225, 'loss/train': 1.4891506433486938} 02/25/2022 11:34:22 - INFO - codeparrot_training - Step 26226: {'lr': 0.00024630221849245595, 'samples': 13428224, 'steps': 26226, 'loss/train': 2.109239339828491} 02/25/2022 11:34:25 - INFO - codeparrot_training - Step 26227: {'lr': 0.00024628585782861663, 'samples': 13428736, 'steps': 26227, 'loss/train': 1.5284104347229004} 02/25/2022 11:34:32 - INFO - codeparrot_training - Step 26228: {'lr': 0.0002462694971806875, 'samples': 13429248, 'steps': 26228, 'loss/train': 0.9382879734039307} 02/25/2022 11:34:36 - INFO - codeparrot_training - Step 26229: {'lr': 0.0002462531365487388, 'samples': 13429760, 'steps': 26229, 'loss/train': 2.3886308670043945} 02/25/2022 11:34:41 - INFO - codeparrot_training - Step 26230: {'lr': 0.0002462367759328403, 'samples': 13430272, 'steps': 26230, 'loss/train': 1.0029067993164062} 02/25/2022 11:34:45 - INFO - codeparrot_training - Step 26231: {'lr': 0.00024622041533306233, 'samples': 13430784, 'steps': 26231, 'loss/train': 2.5881235599517822} 02/25/2022 11:34:50 - INFO - codeparrot_training - Step 26232: {'lr': 0.00024620405474947484, 'samples': 13431296, 'steps': 26232, 'loss/train': 1.576241135597229} 02/25/2022 11:34:54 - INFO - codeparrot_training - Step 26233: {'lr': 0.00024618769418214805, 'samples': 13431808, 'steps': 26233, 'loss/train': 2.874534845352173} 02/25/2022 11:34:59 - INFO - codeparrot_training - Step 26234: {'lr': 0.000246171333631152, 'samples': 13432320, 'steps': 26234, 'loss/train': 1.5983017683029175} 02/25/2022 11:35:03 - INFO - codeparrot_training - Step 26235: {'lr': 0.0002461549730965567, 'samples': 13432832, 'steps': 26235, 'loss/train': 1.9189375638961792} 02/25/2022 11:35:08 - INFO - codeparrot_training - Step 26236: {'lr': 0.0002461386125784322, 'samples': 13433344, 'steps': 26236, 'loss/train': 1.0371625423431396} 02/25/2022 11:35:12 - INFO - codeparrot_training - Step 26237: {'lr': 0.0002461222520768487, 'samples': 13433856, 'steps': 26237, 'loss/train': 1.8823490142822266} 02/25/2022 11:35:19 - INFO - codeparrot_training - Step 26238: {'lr': 0.00024610589159187627, 'samples': 13434368, 'steps': 26238, 'loss/train': 1.2497329711914062} 02/25/2022 11:35:23 - INFO - codeparrot_training - Step 26239: {'lr': 0.00024608953112358495, 'samples': 13434880, 'steps': 26239, 'loss/train': 0.6016479730606079} 02/25/2022 11:35:28 - INFO - codeparrot_training - Step 26240: {'lr': 0.0002460731706720449, 'samples': 13435392, 'steps': 26240, 'loss/train': 1.3934191465377808} 02/25/2022 11:35:32 - INFO - codeparrot_training - Step 26241: {'lr': 0.000246056810237326, 'samples': 13435904, 'steps': 26241, 'loss/train': 2.158846855163574} 02/25/2022 11:35:37 - INFO - codeparrot_training - Step 26242: {'lr': 0.00024604044981949854, 'samples': 13436416, 'steps': 26242, 'loss/train': 2.3282394409179688} 02/25/2022 11:35:41 - INFO - codeparrot_training - Step 26243: {'lr': 0.00024602408941863256, 'samples': 13436928, 'steps': 26243, 'loss/train': 2.113534688949585} 02/25/2022 11:35:46 - INFO - codeparrot_training - Step 26244: {'lr': 0.00024600772903479815, 'samples': 13437440, 'steps': 26244, 'loss/train': 2.697965621948242} 02/25/2022 11:35:50 - INFO - codeparrot_training - Step 26245: {'lr': 0.0002459913686680653, 'samples': 13437952, 'steps': 26245, 'loss/train': 1.2481744289398193} 02/25/2022 11:35:55 - INFO - codeparrot_training - Step 26246: {'lr': 0.00024597500831850415, 'samples': 13438464, 'steps': 26246, 'loss/train': 1.79423987865448} 02/25/2022 11:35:59 - INFO - codeparrot_training - Step 26247: {'lr': 0.00024595864798618484, 'samples': 13438976, 'steps': 26247, 'loss/train': 1.2637900114059448} 02/25/2022 11:36:04 - INFO - codeparrot_training - Step 26248: {'lr': 0.0002459422876711774, 'samples': 13439488, 'steps': 26248, 'loss/train': 2.9381160736083984} 02/25/2022 11:36:08 - INFO - codeparrot_training - Step 26249: {'lr': 0.00024592592737355184, 'samples': 13440000, 'steps': 26249, 'loss/train': 1.3678526878356934} 02/25/2022 11:36:13 - INFO - codeparrot_training - Step 26250: {'lr': 0.0002459095670933783, 'samples': 13440512, 'steps': 26250, 'loss/train': 2.3651723861694336} 02/25/2022 11:36:17 - INFO - codeparrot_training - Step 26251: {'lr': 0.00024589320683072704, 'samples': 13441024, 'steps': 26251, 'loss/train': 2.900285482406616} 02/25/2022 11:36:22 - INFO - codeparrot_training - Step 26252: {'lr': 0.0002458768465856678, 'samples': 13441536, 'steps': 26252, 'loss/train': 0.2583363354206085} 02/25/2022 11:36:26 - INFO - codeparrot_training - Step 26253: {'lr': 0.00024586048635827086, 'samples': 13442048, 'steps': 26253, 'loss/train': 2.3059117794036865} 02/25/2022 11:36:33 - INFO - codeparrot_training - Step 26254: {'lr': 0.0002458441261486063, 'samples': 13442560, 'steps': 26254, 'loss/train': 1.6535823345184326} 02/25/2022 11:36:36 - INFO - codeparrot_training - Step 26255: {'lr': 0.0002458277659567442, 'samples': 13443072, 'steps': 26255, 'loss/train': 1.945255160331726} 02/25/2022 11:36:42 - INFO - codeparrot_training - Step 26256: {'lr': 0.0002458114057827547, 'samples': 13443584, 'steps': 26256, 'loss/train': 2.845904588699341} 02/25/2022 11:36:45 - INFO - codeparrot_training - Step 26257: {'lr': 0.00024579504562670775, 'samples': 13444096, 'steps': 26257, 'loss/train': 2.2146105766296387} 02/25/2022 11:36:51 - INFO - codeparrot_training - Step 26258: {'lr': 0.0002457786854886734, 'samples': 13444608, 'steps': 26258, 'loss/train': 2.3608477115631104} 02/25/2022 11:36:54 - INFO - codeparrot_training - Step 26259: {'lr': 0.00024576232536872194, 'samples': 13445120, 'steps': 26259, 'loss/train': 2.943692684173584} 02/25/2022 11:37:00 - INFO - codeparrot_training - Step 26260: {'lr': 0.00024574596526692337, 'samples': 13445632, 'steps': 26260, 'loss/train': 3.278921365737915} 02/25/2022 11:37:03 - INFO - codeparrot_training - Step 26261: {'lr': 0.0002457296051833476, 'samples': 13446144, 'steps': 26261, 'loss/train': 1.5675383806228638} 02/25/2022 11:37:09 - INFO - codeparrot_training - Step 26262: {'lr': 0.0002457132451180649, 'samples': 13446656, 'steps': 26262, 'loss/train': 1.985438346862793} 02/25/2022 11:37:12 - INFO - codeparrot_training - Step 26263: {'lr': 0.0002456968850711453, 'samples': 13447168, 'steps': 26263, 'loss/train': 1.7740424871444702} 02/25/2022 11:37:19 - INFO - codeparrot_training - Step 26264: {'lr': 0.00024568052504265895, 'samples': 13447680, 'steps': 26264, 'loss/train': 0.7943164110183716} 02/25/2022 11:37:23 - INFO - codeparrot_training - Step 26265: {'lr': 0.00024566416503267577, 'samples': 13448192, 'steps': 26265, 'loss/train': 1.2856733798980713} 02/25/2022 11:37:28 - INFO - codeparrot_training - Step 26266: {'lr': 0.0002456478050412659, 'samples': 13448704, 'steps': 26266, 'loss/train': 0.706490695476532} 02/25/2022 11:37:32 - INFO - codeparrot_training - Step 26267: {'lr': 0.0002456314450684995, 'samples': 13449216, 'steps': 26267, 'loss/train': 1.5952187776565552} 02/25/2022 11:37:37 - INFO - codeparrot_training - Step 26268: {'lr': 0.00024561508511444655, 'samples': 13449728, 'steps': 26268, 'loss/train': 1.5375702381134033} 02/25/2022 11:37:41 - INFO - codeparrot_training - Step 26269: {'lr': 0.0002455987251791773, 'samples': 13450240, 'steps': 26269, 'loss/train': 1.4487030506134033} 02/25/2022 11:37:47 - INFO - codeparrot_training - Step 26270: {'lr': 0.0002455823652627616, 'samples': 13450752, 'steps': 26270, 'loss/train': 2.676496982574463} 02/25/2022 11:37:50 - INFO - codeparrot_training - Step 26271: {'lr': 0.0002455660053652697, 'samples': 13451264, 'steps': 26271, 'loss/train': 2.1709210872650146} 02/25/2022 11:37:56 - INFO - codeparrot_training - Step 26272: {'lr': 0.0002455496454867716, 'samples': 13451776, 'steps': 26272, 'loss/train': 1.0727542638778687} 02/25/2022 11:37:59 - INFO - codeparrot_training - Step 26273: {'lr': 0.00024553328562733746, 'samples': 13452288, 'steps': 26273, 'loss/train': 1.678377389907837} 02/25/2022 11:38:06 - INFO - codeparrot_training - Step 26274: {'lr': 0.0002455169257870373, 'samples': 13452800, 'steps': 26274, 'loss/train': 1.5258363485336304} 02/25/2022 11:38:10 - INFO - codeparrot_training - Step 26275: {'lr': 0.0002455005659659411, 'samples': 13453312, 'steps': 26275, 'loss/train': 0.6886319518089294} 02/25/2022 11:38:16 - INFO - codeparrot_training - Step 26276: {'lr': 0.0002454842061641191, 'samples': 13453824, 'steps': 26276, 'loss/train': 0.9576643109321594} 02/25/2022 11:38:19 - INFO - codeparrot_training - Step 26277: {'lr': 0.00024546784638164145, 'samples': 13454336, 'steps': 26277, 'loss/train': 2.579251289367676} 02/25/2022 11:38:25 - INFO - codeparrot_training - Step 26278: {'lr': 0.00024545148661857794, 'samples': 13454848, 'steps': 26278, 'loss/train': 2.191575288772583} 02/25/2022 11:38:28 - INFO - codeparrot_training - Step 26279: {'lr': 0.0002454351268749989, 'samples': 13455360, 'steps': 26279, 'loss/train': 1.9742505550384521} 02/25/2022 11:38:34 - INFO - codeparrot_training - Step 26280: {'lr': 0.0002454187671509743, 'samples': 13455872, 'steps': 26280, 'loss/train': 2.0906577110290527} 02/25/2022 11:38:37 - INFO - codeparrot_training - Step 26281: {'lr': 0.0002454024074465743, 'samples': 13456384, 'steps': 26281, 'loss/train': 2.5209059715270996} 02/25/2022 11:38:41 - INFO - codeparrot_training - Step 26282: {'lr': 0.0002453860477618689, 'samples': 13456896, 'steps': 26282, 'loss/train': 1.6772773265838623} 02/25/2022 11:38:46 - INFO - codeparrot_training - Step 26283: {'lr': 0.0002453696880969281, 'samples': 13457408, 'steps': 26283, 'loss/train': 3.2000346183776855} 02/25/2022 11:38:50 - INFO - codeparrot_training - Step 26284: {'lr': 0.00024535332845182224, 'samples': 13457920, 'steps': 26284, 'loss/train': 1.639349102973938} 02/25/2022 11:38:57 - INFO - codeparrot_training - Step 26285: {'lr': 0.00024533696882662117, 'samples': 13458432, 'steps': 26285, 'loss/train': 0.9806617498397827} 02/25/2022 11:39:00 - INFO - codeparrot_training - Step 26286: {'lr': 0.00024532060922139505, 'samples': 13458944, 'steps': 26286, 'loss/train': 2.0610392093658447} 02/25/2022 11:39:06 - INFO - codeparrot_training - Step 26287: {'lr': 0.000245304249636214, 'samples': 13459456, 'steps': 26287, 'loss/train': 1.7213953733444214} 02/25/2022 11:39:09 - INFO - codeparrot_training - Step 26288: {'lr': 0.000245287890071148, 'samples': 13459968, 'steps': 26288, 'loss/train': 1.8167425394058228} 02/25/2022 11:39:15 - INFO - codeparrot_training - Step 26289: {'lr': 0.00024527153052626715, 'samples': 13460480, 'steps': 26289, 'loss/train': 1.2684626579284668} 02/25/2022 11:39:18 - INFO - codeparrot_training - Step 26290: {'lr': 0.00024525517100164166, 'samples': 13460992, 'steps': 26290, 'loss/train': 0.6238476037979126} 02/25/2022 11:39:24 - INFO - codeparrot_training - Step 26291: {'lr': 0.0002452388114973415, 'samples': 13461504, 'steps': 26291, 'loss/train': 0.7544644474983215} 02/25/2022 11:39:27 - INFO - codeparrot_training - Step 26292: {'lr': 0.0002452224520134367, 'samples': 13462016, 'steps': 26292, 'loss/train': 2.487161159515381} 02/25/2022 11:39:33 - INFO - codeparrot_training - Step 26293: {'lr': 0.0002452060925499973, 'samples': 13462528, 'steps': 26293, 'loss/train': 0.49551865458488464} 02/25/2022 11:39:36 - INFO - codeparrot_training - Step 26294: {'lr': 0.0002451897331070936, 'samples': 13463040, 'steps': 26294, 'loss/train': 1.293086290359497} 02/25/2022 11:39:42 - INFO - codeparrot_training - Step 26295: {'lr': 0.0002451733736847957, 'samples': 13463552, 'steps': 26295, 'loss/train': 2.158167839050293} 02/25/2022 11:39:45 - INFO - codeparrot_training - Step 26296: {'lr': 0.00024515701428317336, 'samples': 13464064, 'steps': 26296, 'loss/train': 1.394509196281433} 02/25/2022 11:39:51 - INFO - codeparrot_training - Step 26297: {'lr': 0.0002451406549022968, 'samples': 13464576, 'steps': 26297, 'loss/train': 3.0340778827667236} 02/25/2022 11:39:54 - INFO - codeparrot_training - Step 26298: {'lr': 0.00024512429554223613, 'samples': 13465088, 'steps': 26298, 'loss/train': 2.00824236869812} 02/25/2022 11:40:00 - INFO - codeparrot_training - Step 26299: {'lr': 0.0002451079362030616, 'samples': 13465600, 'steps': 26299, 'loss/train': 1.4786911010742188} 02/25/2022 11:40:03 - INFO - codeparrot_training - Step 26300: {'lr': 0.00024509157688484297, 'samples': 13466112, 'steps': 26300, 'loss/train': 1.0193793773651123} 02/25/2022 11:40:11 - INFO - codeparrot_training - Step 26301: {'lr': 0.00024507521758765046, 'samples': 13466624, 'steps': 26301, 'loss/train': 1.8714008331298828} 02/25/2022 11:40:14 - INFO - codeparrot_training - Step 26302: {'lr': 0.0002450588583115542, 'samples': 13467136, 'steps': 26302, 'loss/train': 2.2302093505859375} 02/25/2022 11:40:20 - INFO - codeparrot_training - Step 26303: {'lr': 0.00024504249905662415, 'samples': 13467648, 'steps': 26303, 'loss/train': 2.8510708808898926} 02/25/2022 11:40:24 - INFO - codeparrot_training - Step 26304: {'lr': 0.00024502613982293065, 'samples': 13468160, 'steps': 26304, 'loss/train': 2.325540781021118} 02/25/2022 11:40:29 - INFO - codeparrot_training - Step 26305: {'lr': 0.00024500978061054346, 'samples': 13468672, 'steps': 26305, 'loss/train': 1.4643923044204712} 02/25/2022 11:40:33 - INFO - codeparrot_training - Step 26306: {'lr': 0.0002449934214195327, 'samples': 13469184, 'steps': 26306, 'loss/train': 1.4209997653961182} 02/25/2022 11:40:38 - INFO - codeparrot_training - Step 26307: {'lr': 0.00024497706224996864, 'samples': 13469696, 'steps': 26307, 'loss/train': 2.1823487281799316} 02/25/2022 11:40:42 - INFO - codeparrot_training - Step 26308: {'lr': 0.0002449607031019213, 'samples': 13470208, 'steps': 26308, 'loss/train': 1.8023405075073242} 02/25/2022 11:40:47 - INFO - codeparrot_training - Step 26309: {'lr': 0.00024494434397546067, 'samples': 13470720, 'steps': 26309, 'loss/train': 0.4026634991168976} 02/25/2022 11:40:51 - INFO - codeparrot_training - Step 26310: {'lr': 0.00024492798487065674, 'samples': 13471232, 'steps': 26310, 'loss/train': 1.7011061906814575} 02/25/2022 11:40:58 - INFO - codeparrot_training - Step 26311: {'lr': 0.0002449116257875798, 'samples': 13471744, 'steps': 26311, 'loss/train': 2.4335412979125977} 02/25/2022 11:41:01 - INFO - codeparrot_training - Step 26312: {'lr': 0.0002448952667262999, 'samples': 13472256, 'steps': 26312, 'loss/train': 2.1463046073913574} 02/25/2022 11:41:07 - INFO - codeparrot_training - Step 26313: {'lr': 0.00024487890768688705, 'samples': 13472768, 'steps': 26313, 'loss/train': 1.31080961227417} 02/25/2022 11:41:10 - INFO - codeparrot_training - Step 26314: {'lr': 0.0002448625486694112, 'samples': 13473280, 'steps': 26314, 'loss/train': 1.2081104516983032} 02/25/2022 11:41:16 - INFO - codeparrot_training - Step 26315: {'lr': 0.00024484618967394263, 'samples': 13473792, 'steps': 26315, 'loss/train': 0.958557665348053} 02/25/2022 11:41:20 - INFO - codeparrot_training - Step 26316: {'lr': 0.0002448298307005514, 'samples': 13474304, 'steps': 26316, 'loss/train': 0.6188236474990845} 02/25/2022 11:41:25 - INFO - codeparrot_training - Step 26317: {'lr': 0.00024481347174930756, 'samples': 13474816, 'steps': 26317, 'loss/train': 0.2930773198604584} 02/25/2022 11:41:28 - INFO - codeparrot_training - Step 26318: {'lr': 0.00024479711282028105, 'samples': 13475328, 'steps': 26318, 'loss/train': 1.8447990417480469} 02/25/2022 11:41:34 - INFO - codeparrot_training - Step 26319: {'lr': 0.0002447807539135421, 'samples': 13475840, 'steps': 26319, 'loss/train': 2.0253796577453613} 02/25/2022 11:41:37 - INFO - codeparrot_training - Step 26320: {'lr': 0.00024476439502916077, 'samples': 13476352, 'steps': 26320, 'loss/train': 2.498218297958374} 02/25/2022 11:41:45 - INFO - codeparrot_training - Step 26321: {'lr': 0.0002447480361672071, 'samples': 13476864, 'steps': 26321, 'loss/train': 1.2911486625671387} 02/25/2022 11:41:48 - INFO - codeparrot_training - Step 26322: {'lr': 0.00024473167732775116, 'samples': 13477376, 'steps': 26322, 'loss/train': 1.895642638206482} 02/25/2022 11:41:54 - INFO - codeparrot_training - Step 26323: {'lr': 0.00024471531851086303, 'samples': 13477888, 'steps': 26323, 'loss/train': 1.9397436380386353} 02/25/2022 11:41:57 - INFO - codeparrot_training - Step 26324: {'lr': 0.00024469895971661283, 'samples': 13478400, 'steps': 26324, 'loss/train': 0.09864596277475357} 02/25/2022 11:42:03 - INFO - codeparrot_training - Step 26325: {'lr': 0.0002446826009450706, 'samples': 13478912, 'steps': 26325, 'loss/train': 2.038851022720337} 02/25/2022 11:42:06 - INFO - codeparrot_training - Step 26326: {'lr': 0.0002446662421963064, 'samples': 13479424, 'steps': 26326, 'loss/train': 2.157456874847412} 02/25/2022 11:42:12 - INFO - codeparrot_training - Step 26327: {'lr': 0.00024464988347039037, 'samples': 13479936, 'steps': 26327, 'loss/train': 1.946167230606079} 02/25/2022 11:42:15 - INFO - codeparrot_training - Step 26328: {'lr': 0.0002446335247673925, 'samples': 13480448, 'steps': 26328, 'loss/train': 1.6931893825531006} 02/25/2022 11:42:21 - INFO - codeparrot_training - Step 26329: {'lr': 0.0002446171660873828, 'samples': 13480960, 'steps': 26329, 'loss/train': 1.933598518371582} 02/25/2022 11:42:24 - INFO - codeparrot_training - Step 26330: {'lr': 0.00024460080743043163, 'samples': 13481472, 'steps': 26330, 'loss/train': 0.47926944494247437} 02/25/2022 11:42:31 - INFO - codeparrot_training - Step 26331: {'lr': 0.0002445844487966088, 'samples': 13481984, 'steps': 26331, 'loss/train': 2.051633834838867} 02/25/2022 11:42:35 - INFO - codeparrot_training - Step 26332: {'lr': 0.0002445680901859845, 'samples': 13482496, 'steps': 26332, 'loss/train': 0.803810179233551} 02/25/2022 11:42:40 - INFO - codeparrot_training - Step 26333: {'lr': 0.0002445517315986287, 'samples': 13483008, 'steps': 26333, 'loss/train': 1.4540663957595825} 02/25/2022 11:42:44 - INFO - codeparrot_training - Step 26334: {'lr': 0.00024453537303461176, 'samples': 13483520, 'steps': 26334, 'loss/train': 1.1829906702041626} 02/25/2022 11:42:49 - INFO - codeparrot_training - Step 26335: {'lr': 0.00024451901449400334, 'samples': 13484032, 'steps': 26335, 'loss/train': 0.08518210053443909} 02/25/2022 11:42:53 - INFO - codeparrot_training - Step 26336: {'lr': 0.00024450265597687374, 'samples': 13484544, 'steps': 26336, 'loss/train': 0.9057542085647583} 02/25/2022 11:42:58 - INFO - codeparrot_training - Step 26337: {'lr': 0.000244486297483293, 'samples': 13485056, 'steps': 26337, 'loss/train': 2.377490997314453} 02/25/2022 11:43:02 - INFO - codeparrot_training - Step 26338: {'lr': 0.00024446993901333137, 'samples': 13485568, 'steps': 26338, 'loss/train': 1.8935253620147705} 02/25/2022 11:43:08 - INFO - codeparrot_training - Step 26339: {'lr': 0.0002444535805670587, 'samples': 13486080, 'steps': 26339, 'loss/train': 1.525955080986023} 02/25/2022 11:43:11 - INFO - codeparrot_training - Step 26340: {'lr': 0.000244437222144545, 'samples': 13486592, 'steps': 26340, 'loss/train': 3.2384912967681885} 02/25/2022 11:43:16 - INFO - codeparrot_training - Step 26341: {'lr': 0.00024442086374586056, 'samples': 13487104, 'steps': 26341, 'loss/train': 2.4522252082824707} 02/25/2022 11:43:20 - INFO - codeparrot_training - Step 26342: {'lr': 0.0002444045053710754, 'samples': 13487616, 'steps': 26342, 'loss/train': 1.6107395887374878} 02/25/2022 11:43:27 - INFO - codeparrot_training - Step 26343: {'lr': 0.0002443881470202596, 'samples': 13488128, 'steps': 26343, 'loss/train': 1.1788272857666016} 02/25/2022 11:43:31 - INFO - codeparrot_training - Step 26344: {'lr': 0.0002443717886934831, 'samples': 13488640, 'steps': 26344, 'loss/train': 3.0571553707122803} 02/25/2022 11:43:36 - INFO - codeparrot_training - Step 26345: {'lr': 0.00024435543039081606, 'samples': 13489152, 'steps': 26345, 'loss/train': 2.716193199157715} 02/25/2022 11:43:40 - INFO - codeparrot_training - Step 26346: {'lr': 0.0002443390721123286, 'samples': 13489664, 'steps': 26346, 'loss/train': 2.6670377254486084} 02/25/2022 11:43:45 - INFO - codeparrot_training - Step 26347: {'lr': 0.00024432271385809085, 'samples': 13490176, 'steps': 26347, 'loss/train': 1.6349900960922241} 02/25/2022 11:43:49 - INFO - codeparrot_training - Step 26348: {'lr': 0.0002443063556281727, 'samples': 13490688, 'steps': 26348, 'loss/train': 0.5903484225273132} 02/25/2022 11:43:55 - INFO - codeparrot_training - Step 26349: {'lr': 0.0002442899974226443, 'samples': 13491200, 'steps': 26349, 'loss/train': 2.429945945739746} 02/25/2022 11:43:58 - INFO - codeparrot_training - Step 26350: {'lr': 0.00024427363924157567, 'samples': 13491712, 'steps': 26350, 'loss/train': 1.7602263689041138} 02/25/2022 11:44:02 - INFO - codeparrot_training - Step 26351: {'lr': 0.0002442572810850371, 'samples': 13492224, 'steps': 26351, 'loss/train': 7.8003716468811035} 02/25/2022 11:44:09 - INFO - codeparrot_training - Step 26352: {'lr': 0.0002442409229530985, 'samples': 13492736, 'steps': 26352, 'loss/train': 2.0946173667907715} 02/25/2022 11:44:13 - INFO - codeparrot_training - Step 26353: {'lr': 0.00024422456484582986, 'samples': 13493248, 'steps': 26353, 'loss/train': 1.4148906469345093} 02/25/2022 11:44:18 - INFO - codeparrot_training - Step 26354: {'lr': 0.0002442082067633014, 'samples': 13493760, 'steps': 26354, 'loss/train': 1.2416362762451172} 02/25/2022 11:44:22 - INFO - codeparrot_training - Step 26355: {'lr': 0.00024419184870558313, 'samples': 13494272, 'steps': 26355, 'loss/train': 2.064422607421875} 02/25/2022 11:44:27 - INFO - codeparrot_training - Step 26356: {'lr': 0.00024417549067274523, 'samples': 13494784, 'steps': 26356, 'loss/train': 1.3636423349380493} 02/25/2022 11:44:31 - INFO - codeparrot_training - Step 26357: {'lr': 0.00024415913266485754, 'samples': 13495296, 'steps': 26357, 'loss/train': 1.8285813331604004} 02/25/2022 11:44:36 - INFO - codeparrot_training - Step 26358: {'lr': 0.0002441427746819903, 'samples': 13495808, 'steps': 26358, 'loss/train': 0.4868795573711395} 02/25/2022 11:44:40 - INFO - codeparrot_training - Step 26359: {'lr': 0.00024412641672421357, 'samples': 13496320, 'steps': 26359, 'loss/train': 1.3161351680755615} 02/25/2022 11:44:45 - INFO - codeparrot_training - Step 26360: {'lr': 0.0002441100587915975, 'samples': 13496832, 'steps': 26360, 'loss/train': 1.8202041387557983} 02/25/2022 11:44:49 - INFO - codeparrot_training - Step 26361: {'lr': 0.000244093700884212, 'samples': 13497344, 'steps': 26361, 'loss/train': 2.2062909603118896} 02/25/2022 11:44:54 - INFO - codeparrot_training - Step 26362: {'lr': 0.00024407734300212715, 'samples': 13497856, 'steps': 26362, 'loss/train': 1.8599963188171387} 02/25/2022 11:44:58 - INFO - codeparrot_training - Step 26363: {'lr': 0.0002440609851454131, 'samples': 13498368, 'steps': 26363, 'loss/train': 1.5320758819580078} 02/25/2022 11:45:04 - INFO - codeparrot_training - Step 26364: {'lr': 0.00024404462731413996, 'samples': 13498880, 'steps': 26364, 'loss/train': 2.612672805786133} 02/25/2022 11:45:07 - INFO - codeparrot_training - Step 26365: {'lr': 0.00024402826950837775, 'samples': 13499392, 'steps': 26365, 'loss/train': 0.6196591854095459} 02/25/2022 11:45:10 - INFO - codeparrot_training - Step 26366: {'lr': 0.00024401191172819647, 'samples': 13499904, 'steps': 26366, 'loss/train': 1.7340437173843384} 02/25/2022 11:45:16 - INFO - codeparrot_training - Step 26367: {'lr': 0.00024399555397366633, 'samples': 13500416, 'steps': 26367, 'loss/train': 1.1996020078659058} 02/25/2022 11:45:20 - INFO - codeparrot_training - Step 26368: {'lr': 0.0002439791962448573, 'samples': 13500928, 'steps': 26368, 'loss/train': 2.166757583618164} 02/25/2022 11:45:27 - INFO - codeparrot_training - Step 26369: {'lr': 0.00024396283854183947, 'samples': 13501440, 'steps': 26369, 'loss/train': 1.2331719398498535} 02/25/2022 11:45:30 - INFO - codeparrot_training - Step 26370: {'lr': 0.0002439464808646829, 'samples': 13501952, 'steps': 26370, 'loss/train': 1.3486862182617188} 02/25/2022 11:45:36 - INFO - codeparrot_training - Step 26371: {'lr': 0.00024393012321345775, 'samples': 13502464, 'steps': 26371, 'loss/train': 2.0946381092071533} 02/25/2022 11:45:39 - INFO - codeparrot_training - Step 26372: {'lr': 0.00024391376558823398, 'samples': 13502976, 'steps': 26372, 'loss/train': 0.9696707725524902} 02/25/2022 11:45:45 - INFO - codeparrot_training - Step 26373: {'lr': 0.00024389740798908173, 'samples': 13503488, 'steps': 26373, 'loss/train': 0.8916708827018738} 02/25/2022 11:45:48 - INFO - codeparrot_training - Step 26374: {'lr': 0.00024388105041607105, 'samples': 13504000, 'steps': 26374, 'loss/train': 0.22761280834674835} 02/25/2022 11:45:54 - INFO - codeparrot_training - Step 26375: {'lr': 0.00024386469286927196, 'samples': 13504512, 'steps': 26375, 'loss/train': 2.7063379287719727} 02/25/2022 11:45:57 - INFO - codeparrot_training - Step 26376: {'lr': 0.00024384833534875458, 'samples': 13505024, 'steps': 26376, 'loss/train': 2.295356035232544} 02/25/2022 11:46:05 - INFO - codeparrot_training - Step 26377: {'lr': 0.00024383197785458899, 'samples': 13505536, 'steps': 26377, 'loss/train': 1.3243085145950317} 02/25/2022 11:46:08 - INFO - codeparrot_training - Step 26378: {'lr': 0.00024381562038684534, 'samples': 13506048, 'steps': 26378, 'loss/train': 1.9024004936218262} 02/25/2022 11:46:14 - INFO - codeparrot_training - Step 26379: {'lr': 0.00024379926294559352, 'samples': 13506560, 'steps': 26379, 'loss/train': 2.1351585388183594} 02/25/2022 11:46:17 - INFO - codeparrot_training - Step 26380: {'lr': 0.00024378290553090366, 'samples': 13507072, 'steps': 26380, 'loss/train': 2.3966636657714844} 02/25/2022 11:46:23 - INFO - codeparrot_training - Step 26381: {'lr': 0.00024376654814284586, 'samples': 13507584, 'steps': 26381, 'loss/train': 2.458189010620117} 02/25/2022 11:46:26 - INFO - codeparrot_training - Step 26382: {'lr': 0.0002437501907814903, 'samples': 13508096, 'steps': 26382, 'loss/train': 2.2841715812683105} 02/25/2022 11:46:32 - INFO - codeparrot_training - Step 26383: {'lr': 0.00024373383344690686, 'samples': 13508608, 'steps': 26383, 'loss/train': 1.6360108852386475} 02/25/2022 11:46:35 - INFO - codeparrot_training - Step 26384: {'lr': 0.00024371747613916565, 'samples': 13509120, 'steps': 26384, 'loss/train': 1.940095067024231} 02/25/2022 11:46:41 - INFO - codeparrot_training - Step 26385: {'lr': 0.00024370111885833678, 'samples': 13509632, 'steps': 26385, 'loss/train': 2.4187495708465576} 02/25/2022 11:46:44 - INFO - codeparrot_training - Step 26386: {'lr': 0.00024368476160449047, 'samples': 13510144, 'steps': 26386, 'loss/train': 1.6374770402908325} 02/25/2022 11:46:50 - INFO - codeparrot_training - Step 26387: {'lr': 0.00024366840437769647, 'samples': 13510656, 'steps': 26387, 'loss/train': 3.7390332221984863} 02/25/2022 11:46:54 - INFO - codeparrot_training - Step 26388: {'lr': 0.00024365204717802507, 'samples': 13511168, 'steps': 26388, 'loss/train': 1.6569944620132446} 02/25/2022 11:46:59 - INFO - codeparrot_training - Step 26389: {'lr': 0.0002436356900055462, 'samples': 13511680, 'steps': 26389, 'loss/train': 1.7931761741638184} 02/25/2022 11:47:03 - INFO - codeparrot_training - Step 26390: {'lr': 0.0002436193328603301, 'samples': 13512192, 'steps': 26390, 'loss/train': 1.3246029615402222} 02/25/2022 11:47:08 - INFO - codeparrot_training - Step 26391: {'lr': 0.00024360297574244682, 'samples': 13512704, 'steps': 26391, 'loss/train': 2.185159683227539} 02/25/2022 11:47:14 - INFO - codeparrot_training - Step 26392: {'lr': 0.00024358661865196628, 'samples': 13513216, 'steps': 26392, 'loss/train': 1.2739105224609375} 02/25/2022 11:47:17 - INFO - codeparrot_training - Step 26393: {'lr': 0.00024357026158895858, 'samples': 13513728, 'steps': 26393, 'loss/train': 3.099457263946533} 02/25/2022 11:47:23 - INFO - codeparrot_training - Step 26394: {'lr': 0.00024355390455349388, 'samples': 13514240, 'steps': 26394, 'loss/train': 0.4554407000541687} 02/25/2022 11:47:26 - INFO - codeparrot_training - Step 26395: {'lr': 0.0002435375475456423, 'samples': 13514752, 'steps': 26395, 'loss/train': 0.24246011674404144} 02/25/2022 11:47:32 - INFO - codeparrot_training - Step 26396: {'lr': 0.0002435211905654737, 'samples': 13515264, 'steps': 26396, 'loss/train': 2.035609006881714} 02/25/2022 11:47:35 - INFO - codeparrot_training - Step 26397: {'lr': 0.00024350483361305827, 'samples': 13515776, 'steps': 26397, 'loss/train': 0.31787604093551636} 02/25/2022 11:47:42 - INFO - codeparrot_training - Step 26398: {'lr': 0.00024348847668846608, 'samples': 13516288, 'steps': 26398, 'loss/train': 1.9995020627975464} 02/25/2022 11:47:45 - INFO - codeparrot_training - Step 26399: {'lr': 0.00024347211979176727, 'samples': 13516800, 'steps': 26399, 'loss/train': 1.246475100517273} 02/25/2022 11:47:49 - INFO - codeparrot_training - Step 26400: {'lr': 0.00024345576292303175, 'samples': 13517312, 'steps': 26400, 'loss/train': 1.504394292831421} 02/25/2022 11:47:54 - INFO - codeparrot_training - Step 26401: {'lr': 0.00024343940608232965, 'samples': 13517824, 'steps': 26401, 'loss/train': 0.5194740295410156} 02/25/2022 11:47:58 - INFO - codeparrot_training - Step 26402: {'lr': 0.00024342304926973105, 'samples': 13518336, 'steps': 26402, 'loss/train': 2.1852293014526367} 02/25/2022 11:48:03 - INFO - codeparrot_training - Step 26403: {'lr': 0.00024340669248530602, 'samples': 13518848, 'steps': 26403, 'loss/train': 1.5815045833587646} 02/25/2022 11:48:07 - INFO - codeparrot_training - Step 26404: {'lr': 0.00024339033572912472, 'samples': 13519360, 'steps': 26404, 'loss/train': 1.9526680707931519} 02/25/2022 11:48:12 - INFO - codeparrot_training - Step 26405: {'lr': 0.000243373979001257, 'samples': 13519872, 'steps': 26405, 'loss/train': 2.006246328353882} 02/25/2022 11:48:19 - INFO - codeparrot_training - Step 26406: {'lr': 0.0002433576223017731, 'samples': 13520384, 'steps': 26406, 'loss/train': 1.5348105430603027} 02/25/2022 11:48:23 - INFO - codeparrot_training - Step 26407: {'lr': 0.00024334126563074307, 'samples': 13520896, 'steps': 26407, 'loss/train': 1.5503588914871216} 02/25/2022 11:48:28 - INFO - codeparrot_training - Step 26408: {'lr': 0.00024332490898823695, 'samples': 13521408, 'steps': 26408, 'loss/train': 1.5304791927337646} 02/25/2022 11:48:32 - INFO - codeparrot_training - Step 26409: {'lr': 0.00024330855237432472, 'samples': 13521920, 'steps': 26409, 'loss/train': 2.740084409713745} 02/25/2022 11:48:37 - INFO - codeparrot_training - Step 26410: {'lr': 0.00024329219578907653, 'samples': 13522432, 'steps': 26410, 'loss/train': 2.324345827102661} 02/25/2022 11:48:41 - INFO - codeparrot_training - Step 26411: {'lr': 0.00024327583923256253, 'samples': 13522944, 'steps': 26411, 'loss/train': 2.3731019496917725} 02/25/2022 11:48:46 - INFO - codeparrot_training - Step 26412: {'lr': 0.00024325948270485263, 'samples': 13523456, 'steps': 26412, 'loss/train': 1.6274141073226929} 02/25/2022 11:48:50 - INFO - codeparrot_training - Step 26413: {'lr': 0.00024324312620601702, 'samples': 13523968, 'steps': 26413, 'loss/train': 2.2079548835754395} 02/25/2022 11:48:56 - INFO - codeparrot_training - Step 26414: {'lr': 0.00024322676973612565, 'samples': 13524480, 'steps': 26414, 'loss/train': 1.9055914878845215} 02/25/2022 11:48:59 - INFO - codeparrot_training - Step 26415: {'lr': 0.00024321041329524872, 'samples': 13524992, 'steps': 26415, 'loss/train': 0.7796139121055603} 02/25/2022 11:49:03 - INFO - codeparrot_training - Step 26416: {'lr': 0.00024319405688345613, 'samples': 13525504, 'steps': 26416, 'loss/train': 2.0578885078430176} 02/25/2022 11:49:08 - INFO - codeparrot_training - Step 26417: {'lr': 0.00024317770050081815, 'samples': 13526016, 'steps': 26417, 'loss/train': 2.581758737564087} 02/25/2022 11:49:11 - INFO - codeparrot_training - Step 26418: {'lr': 0.00024316134414740468, 'samples': 13526528, 'steps': 26418, 'loss/train': 2.315044641494751} 02/25/2022 11:49:17 - INFO - codeparrot_training - Step 26419: {'lr': 0.0002431449878232858, 'samples': 13527040, 'steps': 26419, 'loss/train': 3.263165235519409} 02/25/2022 11:49:20 - INFO - codeparrot_training - Step 26420: {'lr': 0.00024312863152853165, 'samples': 13527552, 'steps': 26420, 'loss/train': 1.6306297779083252} 02/25/2022 11:49:26 - INFO - codeparrot_training - Step 26421: {'lr': 0.0002431122752632123, 'samples': 13528064, 'steps': 26421, 'loss/train': 1.2548456192016602} 02/25/2022 11:49:33 - INFO - codeparrot_training - Step 26422: {'lr': 0.00024309591902739775, 'samples': 13528576, 'steps': 26422, 'loss/train': 1.6222890615463257} 02/25/2022 11:49:36 - INFO - codeparrot_training - Step 26423: {'lr': 0.00024307956282115803, 'samples': 13529088, 'steps': 26423, 'loss/train': 2.542529344558716} 02/25/2022 11:49:42 - INFO - codeparrot_training - Step 26424: {'lr': 0.0002430632066445633, 'samples': 13529600, 'steps': 26424, 'loss/train': 2.220149040222168} 02/25/2022 11:49:45 - INFO - codeparrot_training - Step 26425: {'lr': 0.00024304685049768358, 'samples': 13530112, 'steps': 26425, 'loss/train': 2.2881112098693848} 02/25/2022 11:49:52 - INFO - codeparrot_training - Step 26426: {'lr': 0.00024303049438058905, 'samples': 13530624, 'steps': 26426, 'loss/train': 1.0089293718338013} 02/25/2022 11:49:55 - INFO - codeparrot_training - Step 26427: {'lr': 0.00024301413829334957, 'samples': 13531136, 'steps': 26427, 'loss/train': 2.1103029251098633} 02/25/2022 11:49:59 - INFO - codeparrot_training - Step 26428: {'lr': 0.00024299778223603528, 'samples': 13531648, 'steps': 26428, 'loss/train': 2.3049960136413574} 02/25/2022 11:50:02 - INFO - codeparrot_training - Step 26429: {'lr': 0.00024298142620871627, 'samples': 13532160, 'steps': 26429, 'loss/train': 1.0803604125976562} 02/25/2022 11:50:08 - INFO - codeparrot_training - Step 26430: {'lr': 0.00024296507021146274, 'samples': 13532672, 'steps': 26430, 'loss/train': 2.7948157787323} 02/25/2022 11:50:11 - INFO - codeparrot_training - Step 26431: {'lr': 0.0002429487142443445, 'samples': 13533184, 'steps': 26431, 'loss/train': 1.4330811500549316} 02/25/2022 11:50:17 - INFO - codeparrot_training - Step 26432: {'lr': 0.00024293235830743172, 'samples': 13533696, 'steps': 26432, 'loss/train': 1.6584124565124512} 02/25/2022 11:50:20 - INFO - codeparrot_training - Step 26433: {'lr': 0.00024291600240079444, 'samples': 13534208, 'steps': 26433, 'loss/train': 2.803561210632324} 02/25/2022 11:50:27 - INFO - codeparrot_training - Step 26434: {'lr': 0.0002428996465245029, 'samples': 13534720, 'steps': 26434, 'loss/train': 1.9357272386550903} 02/25/2022 11:50:30 - INFO - codeparrot_training - Step 26435: {'lr': 0.00024288329067862692, 'samples': 13535232, 'steps': 26435, 'loss/train': 1.6152267456054688} 02/25/2022 11:50:36 - INFO - codeparrot_training - Step 26436: {'lr': 0.00024286693486323663, 'samples': 13535744, 'steps': 26436, 'loss/train': 1.9259852170944214} 02/25/2022 11:50:39 - INFO - codeparrot_training - Step 26437: {'lr': 0.00024285057907840217, 'samples': 13536256, 'steps': 26437, 'loss/train': 1.3499150276184082} 02/25/2022 11:50:45 - INFO - codeparrot_training - Step 26438: {'lr': 0.00024283422332419352, 'samples': 13536768, 'steps': 26438, 'loss/train': 2.0157687664031982} 02/25/2022 11:50:48 - INFO - codeparrot_training - Step 26439: {'lr': 0.00024281786760068093, 'samples': 13537280, 'steps': 26439, 'loss/train': 1.5157722234725952} 02/25/2022 11:50:54 - INFO - codeparrot_training - Step 26440: {'lr': 0.00024280151190793415, 'samples': 13537792, 'steps': 26440, 'loss/train': 2.3307814598083496} 02/25/2022 11:50:59 - INFO - codeparrot_training - Step 26441: {'lr': 0.00024278515624602344, 'samples': 13538304, 'steps': 26441, 'loss/train': 2.3740296363830566} 02/25/2022 11:51:02 - INFO - codeparrot_training - Step 26442: {'lr': 0.00024276880061501884, 'samples': 13538816, 'steps': 26442, 'loss/train': 0.3079553246498108} 02/25/2022 11:51:09 - INFO - codeparrot_training - Step 26443: {'lr': 0.00024275244501499048, 'samples': 13539328, 'steps': 26443, 'loss/train': 0.6742960810661316} 02/25/2022 11:51:12 - INFO - codeparrot_training - Step 26444: {'lr': 0.00024273608944600826, 'samples': 13539840, 'steps': 26444, 'loss/train': 1.4264261722564697} 02/25/2022 11:51:18 - INFO - codeparrot_training - Step 26445: {'lr': 0.00024271973390814234, 'samples': 13540352, 'steps': 26445, 'loss/train': 1.2122957706451416} 02/25/2022 11:51:21 - INFO - codeparrot_training - Step 26446: {'lr': 0.00024270337840146274, 'samples': 13540864, 'steps': 26446, 'loss/train': 2.329277992248535} 02/25/2022 11:51:25 - INFO - codeparrot_training - Step 26447: {'lr': 0.00024268702292603968, 'samples': 13541376, 'steps': 26447, 'loss/train': 2.3386948108673096} 02/25/2022 11:51:30 - INFO - codeparrot_training - Step 26448: {'lr': 0.00024267066748194293, 'samples': 13541888, 'steps': 26448, 'loss/train': 2.032050132751465} 02/25/2022 11:51:34 - INFO - codeparrot_training - Step 26449: {'lr': 0.00024265431206924276, 'samples': 13542400, 'steps': 26449, 'loss/train': 1.2951445579528809} 02/25/2022 11:51:39 - INFO - codeparrot_training - Step 26450: {'lr': 0.0002426379566880092, 'samples': 13542912, 'steps': 26450, 'loss/train': 7.37010383605957} 02/25/2022 11:51:43 - INFO - codeparrot_training - Step 26451: {'lr': 0.0002426216013383123, 'samples': 13543424, 'steps': 26451, 'loss/train': 1.44121515750885} 02/25/2022 11:51:48 - INFO - codeparrot_training - Step 26452: {'lr': 0.00024260524602022216, 'samples': 13543936, 'steps': 26452, 'loss/train': 1.873138666152954} 02/25/2022 11:51:52 - INFO - codeparrot_training - Step 26453: {'lr': 0.00024258889073380875, 'samples': 13544448, 'steps': 26453, 'loss/train': 2.3655571937561035} 02/25/2022 11:51:58 - INFO - codeparrot_training - Step 26454: {'lr': 0.00024257253547914213, 'samples': 13544960, 'steps': 26454, 'loss/train': 1.5407308340072632} 02/25/2022 11:52:04 - INFO - codeparrot_training - Step 26455: {'lr': 0.0002425561802562925, 'samples': 13545472, 'steps': 26455, 'loss/train': 3.0276730060577393} 02/25/2022 11:52:08 - INFO - codeparrot_training - Step 26456: {'lr': 0.0002425398250653298, 'samples': 13545984, 'steps': 26456, 'loss/train': 1.8760740756988525} 02/25/2022 11:52:11 - INFO - codeparrot_training - Step 26457: {'lr': 0.0002425234699063241, 'samples': 13546496, 'steps': 26457, 'loss/train': 0.7985270619392395} 02/25/2022 11:52:16 - INFO - codeparrot_training - Step 26458: {'lr': 0.00024250711477934552, 'samples': 13547008, 'steps': 26458, 'loss/train': 2.397965908050537} 02/25/2022 11:52:20 - INFO - codeparrot_training - Step 26459: {'lr': 0.00024249075968446404, 'samples': 13547520, 'steps': 26459, 'loss/train': 0.9945634007453918} 02/25/2022 11:52:26 - INFO - codeparrot_training - Step 26460: {'lr': 0.00024247440462174974, 'samples': 13548032, 'steps': 26460, 'loss/train': 1.7033792734146118} 02/25/2022 11:52:29 - INFO - codeparrot_training - Step 26461: {'lr': 0.00024245804959127277, 'samples': 13548544, 'steps': 26461, 'loss/train': 1.7728068828582764} 02/25/2022 11:52:35 - INFO - codeparrot_training - Step 26462: {'lr': 0.00024244169459310312, 'samples': 13549056, 'steps': 26462, 'loss/train': 1.7886940240859985} 02/25/2022 11:52:38 - INFO - codeparrot_training - Step 26463: {'lr': 0.00024242533962731078, 'samples': 13549568, 'steps': 26463, 'loss/train': 1.7180354595184326} 02/25/2022 11:52:44 - INFO - codeparrot_training - Step 26464: {'lr': 0.0002424089846939659, 'samples': 13550080, 'steps': 26464, 'loss/train': 1.7602565288543701} 02/25/2022 11:52:48 - INFO - codeparrot_training - Step 26465: {'lr': 0.0002423926297931386, 'samples': 13550592, 'steps': 26465, 'loss/train': 0.8669648170471191} 02/25/2022 11:52:53 - INFO - codeparrot_training - Step 26466: {'lr': 0.0002423762749248988, 'samples': 13551104, 'steps': 26466, 'loss/train': 3.1445229053497314} 02/25/2022 11:52:57 - INFO - codeparrot_training - Step 26467: {'lr': 0.00024235992008931657, 'samples': 13551616, 'steps': 26467, 'loss/train': 1.8835163116455078} 02/25/2022 11:53:02 - INFO - codeparrot_training - Step 26468: {'lr': 0.00024234356528646204, 'samples': 13552128, 'steps': 26468, 'loss/train': 2.2562899589538574} 02/25/2022 11:53:06 - INFO - codeparrot_training - Step 26469: {'lr': 0.00024232721051640536, 'samples': 13552640, 'steps': 26469, 'loss/train': 1.772828221321106} 02/25/2022 11:53:11 - INFO - codeparrot_training - Step 26470: {'lr': 0.00024231085577921635, 'samples': 13553152, 'steps': 26470, 'loss/train': 2.3923168182373047} 02/25/2022 11:53:15 - INFO - codeparrot_training - Step 26471: {'lr': 0.00024229450107496518, 'samples': 13553664, 'steps': 26471, 'loss/train': 1.8775622844696045} 02/25/2022 11:53:20 - INFO - codeparrot_training - Step 26472: {'lr': 0.00024227814640372195, 'samples': 13554176, 'steps': 26472, 'loss/train': 1.8366479873657227} 02/25/2022 11:53:24 - INFO - codeparrot_training - Step 26473: {'lr': 0.00024226179176555665, 'samples': 13554688, 'steps': 26473, 'loss/train': 0.8986123204231262} 02/25/2022 11:53:30 - INFO - codeparrot_training - Step 26474: {'lr': 0.00024224543716053952, 'samples': 13555200, 'steps': 26474, 'loss/train': 3.697416067123413} 02/25/2022 11:53:34 - INFO - codeparrot_training - Step 26475: {'lr': 0.00024222908258874035, 'samples': 13555712, 'steps': 26475, 'loss/train': 1.9967089891433716} 02/25/2022 11:53:39 - INFO - codeparrot_training - Step 26476: {'lr': 0.00024221272805022935, 'samples': 13556224, 'steps': 26476, 'loss/train': 2.262087345123291} 02/25/2022 11:53:43 - INFO - codeparrot_training - Step 26477: {'lr': 0.0002421963735450765, 'samples': 13556736, 'steps': 26477, 'loss/train': 2.1178677082061768} 02/25/2022 11:53:48 - INFO - codeparrot_training - Step 26478: {'lr': 0.00024218001907335207, 'samples': 13557248, 'steps': 26478, 'loss/train': 0.641125500202179} 02/25/2022 11:53:52 - INFO - codeparrot_training - Step 26479: {'lr': 0.00024216366463512582, 'samples': 13557760, 'steps': 26479, 'loss/train': 1.3325393199920654} 02/25/2022 11:53:58 - INFO - codeparrot_training - Step 26480: {'lr': 0.00024214731023046793, 'samples': 13558272, 'steps': 26480, 'loss/train': 2.29300594329834} 02/25/2022 11:54:01 - INFO - codeparrot_training - Step 26481: {'lr': 0.0002421309558594485, 'samples': 13558784, 'steps': 26481, 'loss/train': 1.9313616752624512} 02/25/2022 11:54:07 - INFO - codeparrot_training - Step 26482: {'lr': 0.00024211460152213763, 'samples': 13559296, 'steps': 26482, 'loss/train': 1.8698166608810425} 02/25/2022 11:54:10 - INFO - codeparrot_training - Step 26483: {'lr': 0.0002420982472186052, 'samples': 13559808, 'steps': 26483, 'loss/train': 2.829249620437622} 02/25/2022 11:54:16 - INFO - codeparrot_training - Step 26484: {'lr': 0.0002420818929489214, 'samples': 13560320, 'steps': 26484, 'loss/train': 1.379737138748169} 02/25/2022 11:54:19 - INFO - codeparrot_training - Step 26485: {'lr': 0.00024206553871315622, 'samples': 13560832, 'steps': 26485, 'loss/train': 2.0421934127807617} 02/25/2022 11:54:25 - INFO - codeparrot_training - Step 26486: {'lr': 0.0002420491845113798, 'samples': 13561344, 'steps': 26486, 'loss/train': 2.2395713329315186} 02/25/2022 11:54:29 - INFO - codeparrot_training - Step 26487: {'lr': 0.00024203283034366223, 'samples': 13561856, 'steps': 26487, 'loss/train': 2.547132730484009} 02/25/2022 11:54:34 - INFO - codeparrot_training - Step 26488: {'lr': 0.00024201647621007336, 'samples': 13562368, 'steps': 26488, 'loss/train': 1.558421015739441} 02/25/2022 11:54:38 - INFO - codeparrot_training - Step 26489: {'lr': 0.0002420001221106834, 'samples': 13562880, 'steps': 26489, 'loss/train': 0.7302256226539612} 02/25/2022 11:54:44 - INFO - codeparrot_training - Step 26490: {'lr': 0.00024198376804556235, 'samples': 13563392, 'steps': 26490, 'loss/train': 3.0497522354125977} 02/25/2022 11:54:48 - INFO - codeparrot_training - Step 26491: {'lr': 0.00024196741401478044, 'samples': 13563904, 'steps': 26491, 'loss/train': 1.8120146989822388} 02/25/2022 11:54:53 - INFO - codeparrot_training - Step 26492: {'lr': 0.00024195106001840741, 'samples': 13564416, 'steps': 26492, 'loss/train': 1.8885564804077148} 02/25/2022 11:54:57 - INFO - codeparrot_training - Step 26493: {'lr': 0.0002419347060565135, 'samples': 13564928, 'steps': 26493, 'loss/train': 1.7711570262908936} 02/25/2022 11:55:02 - INFO - codeparrot_training - Step 26494: {'lr': 0.0002419183521291688, 'samples': 13565440, 'steps': 26494, 'loss/train': 1.6625829935073853} 02/25/2022 11:55:06 - INFO - codeparrot_training - Step 26495: {'lr': 0.0002419019982364434, 'samples': 13565952, 'steps': 26495, 'loss/train': 2.5402770042419434} 02/25/2022 11:55:11 - INFO - codeparrot_training - Step 26496: {'lr': 0.00024188564437840714, 'samples': 13566464, 'steps': 26496, 'loss/train': 2.300276041030884} 02/25/2022 11:55:15 - INFO - codeparrot_training - Step 26497: {'lr': 0.0002418692905551302, 'samples': 13566976, 'steps': 26497, 'loss/train': 1.101785659790039} 02/25/2022 11:55:20 - INFO - codeparrot_training - Step 26498: {'lr': 0.00024185293676668267, 'samples': 13567488, 'steps': 26498, 'loss/train': 2.2778050899505615} 02/25/2022 11:55:24 - INFO - codeparrot_training - Step 26499: {'lr': 0.0002418365830131346, 'samples': 13568000, 'steps': 26499, 'loss/train': 1.111402988433838} 02/25/2022 11:55:30 - INFO - codeparrot_training - Step 26500: {'lr': 0.00024182022929455598, 'samples': 13568512, 'steps': 26500, 'loss/train': 3.105008363723755} 02/25/2022 11:55:34 - INFO - codeparrot_training - Step 26501: {'lr': 0.00024180387561101692, 'samples': 13569024, 'steps': 26501, 'loss/train': 2.5942699909210205} 02/25/2022 11:55:39 - INFO - codeparrot_training - Step 26502: {'lr': 0.00024178752196258747, 'samples': 13569536, 'steps': 26502, 'loss/train': 1.6470201015472412} 02/25/2022 11:55:43 - INFO - codeparrot_training - Step 26503: {'lr': 0.0002417711683493376, 'samples': 13570048, 'steps': 26503, 'loss/train': 2.0359835624694824} 02/25/2022 11:55:48 - INFO - codeparrot_training - Step 26504: {'lr': 0.0002417548147713375, 'samples': 13570560, 'steps': 26504, 'loss/train': 2.046633243560791} 02/25/2022 11:55:52 - INFO - codeparrot_training - Step 26505: {'lr': 0.00024173846122865718, 'samples': 13571072, 'steps': 26505, 'loss/train': 1.1537288427352905} 02/25/2022 11:55:57 - INFO - codeparrot_training - Step 26506: {'lr': 0.00024172210772136656, 'samples': 13571584, 'steps': 26506, 'loss/train': 0.40817639231681824} 02/25/2022 11:56:01 - INFO - codeparrot_training - Step 26507: {'lr': 0.00024170575424953584, 'samples': 13572096, 'steps': 26507, 'loss/train': 1.8084417581558228} 02/25/2022 11:56:06 - INFO - codeparrot_training - Step 26508: {'lr': 0.0002416894008132351, 'samples': 13572608, 'steps': 26508, 'loss/train': 1.5863398313522339} 02/25/2022 11:56:10 - INFO - codeparrot_training - Step 26509: {'lr': 0.00024167304741253432, 'samples': 13573120, 'steps': 26509, 'loss/train': 1.686205506324768} 02/25/2022 11:56:15 - INFO - codeparrot_training - Step 26510: {'lr': 0.00024165669404750347, 'samples': 13573632, 'steps': 26510, 'loss/train': 1.3457974195480347} 02/25/2022 11:56:19 - INFO - codeparrot_training - Step 26511: {'lr': 0.00024164034071821273, 'samples': 13574144, 'steps': 26511, 'loss/train': 1.401214838027954} 02/25/2022 11:56:25 - INFO - codeparrot_training - Step 26512: {'lr': 0.00024162398742473212, 'samples': 13574656, 'steps': 26512, 'loss/train': 1.4683468341827393} 02/25/2022 11:56:28 - INFO - codeparrot_training - Step 26513: {'lr': 0.00024160763416713178, 'samples': 13575168, 'steps': 26513, 'loss/train': 2.3101515769958496} 02/25/2022 11:56:34 - INFO - codeparrot_training - Step 26514: {'lr': 0.00024159128094548157, 'samples': 13575680, 'steps': 26514, 'loss/train': 1.2724518775939941} 02/25/2022 11:56:38 - INFO - codeparrot_training - Step 26515: {'lr': 0.00024157492775985162, 'samples': 13576192, 'steps': 26515, 'loss/train': 2.1959407329559326} 02/25/2022 11:56:43 - INFO - codeparrot_training - Step 26516: {'lr': 0.00024155857461031203, 'samples': 13576704, 'steps': 26516, 'loss/train': 2.2702581882476807} 02/25/2022 11:56:47 - INFO - codeparrot_training - Step 26517: {'lr': 0.00024154222149693294, 'samples': 13577216, 'steps': 26517, 'loss/train': 2.8432254791259766} 02/25/2022 11:56:52 - INFO - codeparrot_training - Step 26518: {'lr': 0.00024152586841978417, 'samples': 13577728, 'steps': 26518, 'loss/train': 1.9278030395507812} 02/25/2022 11:56:56 - INFO - codeparrot_training - Step 26519: {'lr': 0.00024150951537893587, 'samples': 13578240, 'steps': 26519, 'loss/train': 2.2340476512908936} 02/25/2022 11:57:01 - INFO - codeparrot_training - Step 26520: {'lr': 0.00024149316237445813, 'samples': 13578752, 'steps': 26520, 'loss/train': 1.3008180856704712} 02/25/2022 11:57:04 - INFO - codeparrot_training - Step 26521: {'lr': 0.00024147680940642097, 'samples': 13579264, 'steps': 26521, 'loss/train': 0.985638439655304} 02/25/2022 11:57:10 - INFO - codeparrot_training - Step 26522: {'lr': 0.0002414604564748946, 'samples': 13579776, 'steps': 26522, 'loss/train': 1.7963171005249023} 02/25/2022 11:57:13 - INFO - codeparrot_training - Step 26523: {'lr': 0.00024144410357994876, 'samples': 13580288, 'steps': 26523, 'loss/train': 2.6810050010681152} 02/25/2022 11:57:19 - INFO - codeparrot_training - Step 26524: {'lr': 0.00024142775072165368, 'samples': 13580800, 'steps': 26524, 'loss/train': 1.9601081609725952} 02/25/2022 11:57:22 - INFO - codeparrot_training - Step 26525: {'lr': 0.00024141139790007942, 'samples': 13581312, 'steps': 26525, 'loss/train': 1.6294368505477905} 02/25/2022 11:57:29 - INFO - codeparrot_training - Step 26526: {'lr': 0.0002413950451152961, 'samples': 13581824, 'steps': 26526, 'loss/train': 1.6196589469909668} 02/25/2022 11:57:32 - INFO - codeparrot_training - Step 26527: {'lr': 0.00024137869236737352, 'samples': 13582336, 'steps': 26527, 'loss/train': 2.0140879154205322} 02/25/2022 11:57:38 - INFO - codeparrot_training - Step 26528: {'lr': 0.00024136233965638194, 'samples': 13582848, 'steps': 26528, 'loss/train': 2.0122475624084473} 02/25/2022 11:57:41 - INFO - codeparrot_training - Step 26529: {'lr': 0.00024134598698239134, 'samples': 13583360, 'steps': 26529, 'loss/train': 2.5290005207061768} 02/25/2022 11:57:47 - INFO - codeparrot_training - Step 26530: {'lr': 0.00024132963434547188, 'samples': 13583872, 'steps': 26530, 'loss/train': 1.0112128257751465} 02/25/2022 11:57:50 - INFO - codeparrot_training - Step 26531: {'lr': 0.00024131328174569342, 'samples': 13584384, 'steps': 26531, 'loss/train': 2.014789581298828} 02/25/2022 11:57:56 - INFO - codeparrot_training - Step 26532: {'lr': 0.0002412969291831261, 'samples': 13584896, 'steps': 26532, 'loss/train': 1.5335077047348022} 02/25/2022 11:57:59 - INFO - codeparrot_training - Step 26533: {'lr': 0.00024128057665783996, 'samples': 13585408, 'steps': 26533, 'loss/train': 1.3351154327392578} 02/25/2022 11:58:05 - INFO - codeparrot_training - Step 26534: {'lr': 0.00024126422416990506, 'samples': 13585920, 'steps': 26534, 'loss/train': 2.1637916564941406} 02/25/2022 11:58:08 - INFO - codeparrot_training - Step 26535: {'lr': 0.0002412478717193916, 'samples': 13586432, 'steps': 26535, 'loss/train': 2.0140459537506104} 02/25/2022 11:58:15 - INFO - codeparrot_training - Step 26536: {'lr': 0.0002412315193063693, 'samples': 13586944, 'steps': 26536, 'loss/train': 1.7236000299453735} 02/25/2022 11:58:20 - INFO - codeparrot_training - Step 26537: {'lr': 0.00024121516693090841, 'samples': 13587456, 'steps': 26537, 'loss/train': 1.813808560371399} 02/25/2022 11:58:24 - INFO - codeparrot_training - Step 26538: {'lr': 0.00024119881459307906, 'samples': 13587968, 'steps': 26538, 'loss/train': 0.23139937222003937} 02/25/2022 11:58:27 - INFO - codeparrot_training - Step 26539: {'lr': 0.00024118246229295115, 'samples': 13588480, 'steps': 26539, 'loss/train': 1.9699782133102417} 02/25/2022 11:58:33 - INFO - codeparrot_training - Step 26540: {'lr': 0.0002411661100305947, 'samples': 13588992, 'steps': 26540, 'loss/train': 2.3278486728668213} 02/25/2022 11:58:36 - INFO - codeparrot_training - Step 26541: {'lr': 0.00024114975780607987, 'samples': 13589504, 'steps': 26541, 'loss/train': 2.3825831413269043} 02/25/2022 11:58:42 - INFO - codeparrot_training - Step 26542: {'lr': 0.0002411334056194767, 'samples': 13590016, 'steps': 26542, 'loss/train': 1.9745625257492065} 02/25/2022 11:58:45 - INFO - codeparrot_training - Step 26543: {'lr': 0.00024111705347085521, 'samples': 13590528, 'steps': 26543, 'loss/train': 2.179685592651367} 02/25/2022 11:58:51 - INFO - codeparrot_training - Step 26544: {'lr': 0.0002411007013602854, 'samples': 13591040, 'steps': 26544, 'loss/train': 1.745808482170105} 02/25/2022 11:58:56 - INFO - codeparrot_training - Step 26545: {'lr': 0.0002410843492878374, 'samples': 13591552, 'steps': 26545, 'loss/train': 0.9869518280029297} 02/25/2022 11:59:00 - INFO - codeparrot_training - Step 26546: {'lr': 0.00024106799725358117, 'samples': 13592064, 'steps': 26546, 'loss/train': 1.8209220170974731} 02/25/2022 11:59:06 - INFO - codeparrot_training - Step 26547: {'lr': 0.0002410516452575868, 'samples': 13592576, 'steps': 26547, 'loss/train': 1.198609471321106} 02/25/2022 11:59:09 - INFO - codeparrot_training - Step 26548: {'lr': 0.00024103529329992437, 'samples': 13593088, 'steps': 26548, 'loss/train': 2.8850834369659424} 02/25/2022 11:59:15 - INFO - codeparrot_training - Step 26549: {'lr': 0.00024101894138066395, 'samples': 13593600, 'steps': 26549, 'loss/train': 0.6228379607200623} 02/25/2022 11:59:18 - INFO - codeparrot_training - Step 26550: {'lr': 0.00024100258949987544, 'samples': 13594112, 'steps': 26550, 'loss/train': 1.8354586362838745} 02/25/2022 11:59:24 - INFO - codeparrot_training - Step 26551: {'lr': 0.00024098623765762898, 'samples': 13594624, 'steps': 26551, 'loss/train': 2.662301778793335} 02/25/2022 11:59:27 - INFO - codeparrot_training - Step 26552: {'lr': 0.00024096988585399474, 'samples': 13595136, 'steps': 26552, 'loss/train': 1.998597264289856} 02/25/2022 11:59:33 - INFO - codeparrot_training - Step 26553: {'lr': 0.00024095353408904252, 'samples': 13595648, 'steps': 26553, 'loss/train': 2.0784735679626465} 02/25/2022 11:59:36 - INFO - codeparrot_training - Step 26554: {'lr': 0.00024093718236284248, 'samples': 13596160, 'steps': 26554, 'loss/train': 1.9538110494613647} 02/25/2022 11:59:42 - INFO - codeparrot_training - Step 26555: {'lr': 0.00024092083067546468, 'samples': 13596672, 'steps': 26555, 'loss/train': 1.279066801071167} 02/25/2022 11:59:45 - INFO - codeparrot_training - Step 26556: {'lr': 0.00024090447902697928, 'samples': 13597184, 'steps': 26556, 'loss/train': 0.8897513151168823} 02/25/2022 11:59:52 - INFO - codeparrot_training - Step 26557: {'lr': 0.0002408881274174561, 'samples': 13597696, 'steps': 26557, 'loss/train': 2.133594036102295} 02/25/2022 11:59:55 - INFO - codeparrot_training - Step 26558: {'lr': 0.00024087177584696526, 'samples': 13598208, 'steps': 26558, 'loss/train': 3.3634629249572754} 02/25/2022 12:00:01 - INFO - codeparrot_training - Step 26559: {'lr': 0.00024085542431557687, 'samples': 13598720, 'steps': 26559, 'loss/train': 0.5618734955787659} 02/25/2022 12:00:04 - INFO - codeparrot_training - Step 26560: {'lr': 0.0002408390728233609, 'samples': 13599232, 'steps': 26560, 'loss/train': 1.6674375534057617} 02/25/2022 12:00:07 - INFO - codeparrot_training - Step 26561: {'lr': 0.00024082272137038757, 'samples': 13599744, 'steps': 26561, 'loss/train': 2.2926013469696045} 02/25/2022 12:00:13 - INFO - codeparrot_training - Step 26562: {'lr': 0.00024080636995672667, 'samples': 13600256, 'steps': 26562, 'loss/train': 1.0135136842727661} 02/25/2022 12:00:19 - INFO - codeparrot_training - Step 26563: {'lr': 0.00024079001858244835, 'samples': 13600768, 'steps': 26563, 'loss/train': 2.3266797065734863} 02/25/2022 12:00:22 - INFO - codeparrot_training - Step 26564: {'lr': 0.0002407736672476227, 'samples': 13601280, 'steps': 26564, 'loss/train': 1.8169890642166138} 02/25/2022 12:00:28 - INFO - codeparrot_training - Step 26565: {'lr': 0.0002407573159523198, 'samples': 13601792, 'steps': 26565, 'loss/train': 0.08260375261306763} 02/25/2022 12:00:32 - INFO - codeparrot_training - Step 26566: {'lr': 0.00024074096469660952, 'samples': 13602304, 'steps': 26566, 'loss/train': 1.8742878437042236} 02/25/2022 12:00:35 - INFO - codeparrot_training - Step 26567: {'lr': 0.00024072461348056205, 'samples': 13602816, 'steps': 26567, 'loss/train': 1.9476584196090698} 02/25/2022 12:00:41 - INFO - codeparrot_training - Step 26568: {'lr': 0.00024070826230424732, 'samples': 13603328, 'steps': 26568, 'loss/train': 2.019618272781372} 02/25/2022 12:00:44 - INFO - codeparrot_training - Step 26569: {'lr': 0.00024069191116773552, 'samples': 13603840, 'steps': 26569, 'loss/train': 2.0052618980407715} 02/25/2022 12:00:50 - INFO - codeparrot_training - Step 26570: {'lr': 0.00024067556007109666, 'samples': 13604352, 'steps': 26570, 'loss/train': 1.7393321990966797} 02/25/2022 12:00:53 - INFO - codeparrot_training - Step 26571: {'lr': 0.00024065920901440068, 'samples': 13604864, 'steps': 26571, 'loss/train': 2.554619789123535} 02/25/2022 12:00:59 - INFO - codeparrot_training - Step 26572: {'lr': 0.00024064285799771766, 'samples': 13605376, 'steps': 26572, 'loss/train': 1.22853684425354} 02/25/2022 12:01:03 - INFO - codeparrot_training - Step 26573: {'lr': 0.00024062650702111766, 'samples': 13605888, 'steps': 26573, 'loss/train': 1.5425652265548706} 02/25/2022 12:01:08 - INFO - codeparrot_training - Step 26574: {'lr': 0.00024061015608467084, 'samples': 13606400, 'steps': 26574, 'loss/train': 1.59976327419281} 02/25/2022 12:01:12 - INFO - codeparrot_training - Step 26575: {'lr': 0.00024059380518844702, 'samples': 13606912, 'steps': 26575, 'loss/train': 1.6789342164993286} 02/25/2022 12:01:17 - INFO - codeparrot_training - Step 26576: {'lr': 0.00024057745433251636, 'samples': 13607424, 'steps': 26576, 'loss/train': 1.1474515199661255} 02/25/2022 12:01:21 - INFO - codeparrot_training - Step 26577: {'lr': 0.00024056110351694887, 'samples': 13607936, 'steps': 26577, 'loss/train': 2.047783374786377} 02/25/2022 12:01:26 - INFO - codeparrot_training - Step 26578: {'lr': 0.00024054475274181474, 'samples': 13608448, 'steps': 26578, 'loss/train': 1.6341447830200195} 02/25/2022 12:01:30 - INFO - codeparrot_training - Step 26579: {'lr': 0.0002405284020071838, 'samples': 13608960, 'steps': 26579, 'loss/train': 2.2441558837890625} 02/25/2022 12:01:35 - INFO - codeparrot_training - Step 26580: {'lr': 0.00024051205131312618, 'samples': 13609472, 'steps': 26580, 'loss/train': 0.5366752743721008} 02/25/2022 12:01:39 - INFO - codeparrot_training - Step 26581: {'lr': 0.00024049570065971188, 'samples': 13609984, 'steps': 26581, 'loss/train': 1.8090510368347168} 02/25/2022 12:01:44 - INFO - codeparrot_training - Step 26582: {'lr': 0.00024047935004701106, 'samples': 13610496, 'steps': 26582, 'loss/train': 1.4878827333450317} 02/25/2022 12:01:48 - INFO - codeparrot_training - Step 26583: {'lr': 0.0002404629994750937, 'samples': 13611008, 'steps': 26583, 'loss/train': 1.8696717023849487} 02/25/2022 12:01:54 - INFO - codeparrot_training - Step 26584: {'lr': 0.0002404466489440297, 'samples': 13611520, 'steps': 26584, 'loss/train': 1.2563058137893677} 02/25/2022 12:01:57 - INFO - codeparrot_training - Step 26585: {'lr': 0.00024043029845388934, 'samples': 13612032, 'steps': 26585, 'loss/train': 2.8918843269348145} 02/25/2022 12:02:03 - INFO - codeparrot_training - Step 26586: {'lr': 0.00024041394800474247, 'samples': 13612544, 'steps': 26586, 'loss/train': 1.777243971824646} 02/25/2022 12:02:06 - INFO - codeparrot_training - Step 26587: {'lr': 0.00024039759759665925, 'samples': 13613056, 'steps': 26587, 'loss/train': 1.2043194770812988} 02/25/2022 12:02:12 - INFO - codeparrot_training - Step 26588: {'lr': 0.00024038124722970962, 'samples': 13613568, 'steps': 26588, 'loss/train': 1.223273515701294} 02/25/2022 12:02:15 - INFO - codeparrot_training - Step 26589: {'lr': 0.00024036489690396374, 'samples': 13614080, 'steps': 26589, 'loss/train': 2.6900808811187744} 02/25/2022 12:02:21 - INFO - codeparrot_training - Step 26590: {'lr': 0.00024034854661949152, 'samples': 13614592, 'steps': 26590, 'loss/train': 1.5867611169815063} 02/25/2022 12:02:24 - INFO - codeparrot_training - Step 26591: {'lr': 0.00024033219637636312, 'samples': 13615104, 'steps': 26591, 'loss/train': 1.285199522972107} 02/25/2022 12:02:30 - INFO - codeparrot_training - Step 26592: {'lr': 0.0002403158461746485, 'samples': 13615616, 'steps': 26592, 'loss/train': 0.11368382722139359} 02/25/2022 12:02:33 - INFO - codeparrot_training - Step 26593: {'lr': 0.00024029949601441766, 'samples': 13616128, 'steps': 26593, 'loss/train': 2.3916549682617188} 02/25/2022 12:02:40 - INFO - codeparrot_training - Step 26594: {'lr': 0.00024028314589574074, 'samples': 13616640, 'steps': 26594, 'loss/train': 0.6202870011329651} 02/25/2022 12:02:43 - INFO - codeparrot_training - Step 26595: {'lr': 0.00024026679581868772, 'samples': 13617152, 'steps': 26595, 'loss/train': 2.531552791595459} 02/25/2022 12:02:49 - INFO - codeparrot_training - Step 26596: {'lr': 0.00024025044578332875, 'samples': 13617664, 'steps': 26596, 'loss/train': 2.273005247116089} 02/25/2022 12:02:52 - INFO - codeparrot_training - Step 26597: {'lr': 0.00024023409578973368, 'samples': 13618176, 'steps': 26597, 'loss/train': 2.321377992630005} 02/25/2022 12:02:58 - INFO - codeparrot_training - Step 26598: {'lr': 0.00024021774583797266, 'samples': 13618688, 'steps': 26598, 'loss/train': 0.043440669775009155} 02/25/2022 12:03:01 - INFO - codeparrot_training - Step 26599: {'lr': 0.00024020139592811568, 'samples': 13619200, 'steps': 26599, 'loss/train': 2.193847179412842} 02/25/2022 12:03:06 - INFO - codeparrot_training - Step 26600: {'lr': 0.00024018504606023293, 'samples': 13619712, 'steps': 26600, 'loss/train': 1.485547423362732} 02/25/2022 12:03:10 - INFO - codeparrot_training - Step 26601: {'lr': 0.00024016869623439422, 'samples': 13620224, 'steps': 26601, 'loss/train': 2.1026217937469482} 02/25/2022 12:03:16 - INFO - codeparrot_training - Step 26602: {'lr': 0.00024015234645066968, 'samples': 13620736, 'steps': 26602, 'loss/train': 1.9695422649383545} 02/25/2022 12:03:19 - INFO - codeparrot_training - Step 26603: {'lr': 0.00024013599670912936, 'samples': 13621248, 'steps': 26603, 'loss/train': 1.4434484243392944} 02/25/2022 12:03:25 - INFO - codeparrot_training - Step 26604: {'lr': 0.00024011964700984342, 'samples': 13621760, 'steps': 26604, 'loss/train': 1.5674806833267212} 02/25/2022 12:03:29 - INFO - codeparrot_training - Step 26605: {'lr': 0.00024010329735288166, 'samples': 13622272, 'steps': 26605, 'loss/train': 2.466535806655884} 02/25/2022 12:03:34 - INFO - codeparrot_training - Step 26606: {'lr': 0.00024008694773831424, 'samples': 13622784, 'steps': 26606, 'loss/train': 2.4010603427886963} 02/25/2022 12:03:38 - INFO - codeparrot_training - Step 26607: {'lr': 0.0002400705981662112, 'samples': 13623296, 'steps': 26607, 'loss/train': 2.2689950466156006} 02/25/2022 12:03:43 - INFO - codeparrot_training - Step 26608: {'lr': 0.00024005424863664251, 'samples': 13623808, 'steps': 26608, 'loss/train': 1.5158312320709229} 02/25/2022 12:03:47 - INFO - codeparrot_training - Step 26609: {'lr': 0.00024003789914967842, 'samples': 13624320, 'steps': 26609, 'loss/train': 2.184361219406128} 02/25/2022 12:03:52 - INFO - codeparrot_training - Step 26610: {'lr': 0.0002400215497053887, 'samples': 13624832, 'steps': 26610, 'loss/train': 1.9128063917160034} 02/25/2022 12:03:56 - INFO - codeparrot_training - Step 26611: {'lr': 0.00024000520030384345, 'samples': 13625344, 'steps': 26611, 'loss/train': 3.688497304916382} 02/25/2022 12:04:01 - INFO - codeparrot_training - Step 26612: {'lr': 0.00023998885094511277, 'samples': 13625856, 'steps': 26612, 'loss/train': 0.3894404172897339} 02/25/2022 12:04:05 - INFO - codeparrot_training - Step 26613: {'lr': 0.00023997250162926682, 'samples': 13626368, 'steps': 26613, 'loss/train': 2.4020166397094727} 02/25/2022 12:04:10 - INFO - codeparrot_training - Step 26614: {'lr': 0.00023995615235637534, 'samples': 13626880, 'steps': 26614, 'loss/train': 0.3250059187412262} 02/25/2022 12:04:14 - INFO - codeparrot_training - Step 26615: {'lr': 0.00023993980312650852, 'samples': 13627392, 'steps': 26615, 'loss/train': 1.8597544431686401} 02/25/2022 12:04:19 - INFO - codeparrot_training - Step 26616: {'lr': 0.00023992345393973637, 'samples': 13627904, 'steps': 26616, 'loss/train': 3.1517093181610107} 02/25/2022 12:04:23 - INFO - codeparrot_training - Step 26617: {'lr': 0.00023990710479612897, 'samples': 13628416, 'steps': 26617, 'loss/train': 2.3661489486694336} 02/25/2022 12:04:28 - INFO - codeparrot_training - Step 26618: {'lr': 0.00023989075569575644, 'samples': 13628928, 'steps': 26618, 'loss/train': 2.699223518371582} 02/25/2022 12:04:31 - INFO - codeparrot_training - Step 26619: {'lr': 0.0002398744066386886, 'samples': 13629440, 'steps': 26619, 'loss/train': 1.8562602996826172} 02/25/2022 12:04:38 - INFO - codeparrot_training - Step 26620: {'lr': 0.00023985805762499553, 'samples': 13629952, 'steps': 26620, 'loss/train': 0.43890470266342163} 02/25/2022 12:04:41 - INFO - codeparrot_training - Step 26621: {'lr': 0.00023984170865474737, 'samples': 13630464, 'steps': 26621, 'loss/train': 1.8763319253921509} 02/25/2022 12:04:47 - INFO - codeparrot_training - Step 26622: {'lr': 0.00023982535972801424, 'samples': 13630976, 'steps': 26622, 'loss/train': 1.9783025979995728} 02/25/2022 12:04:50 - INFO - codeparrot_training - Step 26623: {'lr': 0.00023980901084486587, 'samples': 13631488, 'steps': 26623, 'loss/train': 1.3606804609298706} 02/25/2022 12:04:56 - INFO - codeparrot_training - Step 26624: {'lr': 0.00023979266200537251, 'samples': 13632000, 'steps': 26624, 'loss/train': 1.8004802465438843} 02/25/2022 12:04:59 - INFO - codeparrot_training - Step 26625: {'lr': 0.00023977631320960417, 'samples': 13632512, 'steps': 26625, 'loss/train': 1.6889280080795288} 02/25/2022 12:05:05 - INFO - codeparrot_training - Step 26626: {'lr': 0.00023975996445763084, 'samples': 13633024, 'steps': 26626, 'loss/train': 1.457277536392212} 02/25/2022 12:05:08 - INFO - codeparrot_training - Step 26627: {'lr': 0.00023974361574952255, 'samples': 13633536, 'steps': 26627, 'loss/train': 1.3851258754730225} 02/25/2022 12:05:14 - INFO - codeparrot_training - Step 26628: {'lr': 0.00023972726708534932, 'samples': 13634048, 'steps': 26628, 'loss/train': 1.564491868019104} 02/25/2022 12:05:17 - INFO - codeparrot_training - Step 26629: {'lr': 0.00023971091846518132, 'samples': 13634560, 'steps': 26629, 'loss/train': 2.001295804977417} 02/25/2022 12:05:23 - INFO - codeparrot_training - Step 26630: {'lr': 0.00023969456988908838, 'samples': 13635072, 'steps': 26630, 'loss/train': 2.423020362854004} 02/25/2022 12:05:27 - INFO - codeparrot_training - Step 26631: {'lr': 0.0002396782213571407, 'samples': 13635584, 'steps': 26631, 'loss/train': 1.8267040252685547} 02/25/2022 12:05:32 - INFO - codeparrot_training - Step 26632: {'lr': 0.00023966187286940827, 'samples': 13636096, 'steps': 26632, 'loss/train': 1.217692255973816} 02/25/2022 12:05:36 - INFO - codeparrot_training - Step 26633: {'lr': 0.000239645524425961, 'samples': 13636608, 'steps': 26633, 'loss/train': 0.6895974278450012} 02/25/2022 12:05:41 - INFO - codeparrot_training - Step 26634: {'lr': 0.00023962917602686904, 'samples': 13637120, 'steps': 26634, 'loss/train': 1.9834662675857544} 02/25/2022 12:05:45 - INFO - codeparrot_training - Step 26635: {'lr': 0.00023961282767220244, 'samples': 13637632, 'steps': 26635, 'loss/train': 1.8184022903442383} 02/25/2022 12:05:51 - INFO - codeparrot_training - Step 26636: {'lr': 0.00023959647936203118, 'samples': 13638144, 'steps': 26636, 'loss/train': 2.4523892402648926} 02/25/2022 12:05:54 - INFO - codeparrot_training - Step 26637: {'lr': 0.00023958013109642523, 'samples': 13638656, 'steps': 26637, 'loss/train': 1.2354689836502075} 02/25/2022 12:06:00 - INFO - codeparrot_training - Step 26638: {'lr': 0.0002395637828754547, 'samples': 13639168, 'steps': 26638, 'loss/train': 2.018993616104126} 02/25/2022 12:06:04 - INFO - codeparrot_training - Step 26639: {'lr': 0.00023954743469918976, 'samples': 13639680, 'steps': 26639, 'loss/train': 1.320568323135376} 02/25/2022 12:06:10 - INFO - codeparrot_training - Step 26640: {'lr': 0.00023953108656770016, 'samples': 13640192, 'steps': 26640, 'loss/train': 2.031628131866455} 02/25/2022 12:06:13 - INFO - codeparrot_training - Step 26641: {'lr': 0.00023951473848105603, 'samples': 13640704, 'steps': 26641, 'loss/train': 1.6969976425170898} 02/25/2022 12:06:17 - INFO - codeparrot_training - Step 26642: {'lr': 0.00023949839043932748, 'samples': 13641216, 'steps': 26642, 'loss/train': 0.2128089964389801} 02/25/2022 12:06:22 - INFO - codeparrot_training - Step 26643: {'lr': 0.00023948204244258445, 'samples': 13641728, 'steps': 26643, 'loss/train': 1.0688501596450806} 02/25/2022 12:06:26 - INFO - codeparrot_training - Step 26644: {'lr': 0.00023946569449089715, 'samples': 13642240, 'steps': 26644, 'loss/train': 2.347038507461548} 02/25/2022 12:06:32 - INFO - codeparrot_training - Step 26645: {'lr': 0.00023944934658433535, 'samples': 13642752, 'steps': 26645, 'loss/train': 1.2107418775558472} 02/25/2022 12:06:35 - INFO - codeparrot_training - Step 26646: {'lr': 0.00023943299872296916, 'samples': 13643264, 'steps': 26646, 'loss/train': 2.8251256942749023} 02/25/2022 12:06:40 - INFO - codeparrot_training - Step 26647: {'lr': 0.00023941665090686871, 'samples': 13643776, 'steps': 26647, 'loss/train': 2.206286668777466} 02/25/2022 12:06:44 - INFO - codeparrot_training - Step 26648: {'lr': 0.00023940030313610402, 'samples': 13644288, 'steps': 26648, 'loss/train': 1.6813606023788452} 02/25/2022 12:06:49 - INFO - codeparrot_training - Step 26649: {'lr': 0.000239383955410745, 'samples': 13644800, 'steps': 26649, 'loss/train': 1.980342149734497} 02/25/2022 12:06:53 - INFO - codeparrot_training - Step 26650: {'lr': 0.00023936760773086174, 'samples': 13645312, 'steps': 26650, 'loss/train': 2.2128231525421143} 02/25/2022 12:06:58 - INFO - codeparrot_training - Step 26651: {'lr': 0.00023935126009652424, 'samples': 13645824, 'steps': 26651, 'loss/train': 1.8103755712509155} 02/25/2022 12:07:02 - INFO - codeparrot_training - Step 26652: {'lr': 0.0002393349125078027, 'samples': 13646336, 'steps': 26652, 'loss/train': 2.147063970565796} 02/25/2022 12:07:07 - INFO - codeparrot_training - Step 26653: {'lr': 0.0002393185649647669, 'samples': 13646848, 'steps': 26653, 'loss/train': 1.9703105688095093} 02/25/2022 12:07:11 - INFO - codeparrot_training - Step 26654: {'lr': 0.00023930221746748696, 'samples': 13647360, 'steps': 26654, 'loss/train': 2.709378719329834} 02/25/2022 12:07:17 - INFO - codeparrot_training - Step 26655: {'lr': 0.00023928587001603295, 'samples': 13647872, 'steps': 26655, 'loss/train': 1.3376750946044922} 02/25/2022 12:07:21 - INFO - codeparrot_training - Step 26656: {'lr': 0.00023926952261047482, 'samples': 13648384, 'steps': 26656, 'loss/train': 1.840340495109558} 02/25/2022 12:07:26 - INFO - codeparrot_training - Step 26657: {'lr': 0.0002392531752508828, 'samples': 13648896, 'steps': 26657, 'loss/train': 1.5656640529632568} 02/25/2022 12:07:30 - INFO - codeparrot_training - Step 26658: {'lr': 0.00023923682793732665, 'samples': 13649408, 'steps': 26658, 'loss/train': 2.85665225982666} 02/25/2022 12:07:35 - INFO - codeparrot_training - Step 26659: {'lr': 0.00023922048066987652, 'samples': 13649920, 'steps': 26659, 'loss/train': 1.539355993270874} 02/25/2022 12:07:39 - INFO - codeparrot_training - Step 26660: {'lr': 0.0002392041334486024, 'samples': 13650432, 'steps': 26660, 'loss/train': 2.2422823905944824} 02/25/2022 12:07:44 - INFO - codeparrot_training - Step 26661: {'lr': 0.00023918778627357446, 'samples': 13650944, 'steps': 26661, 'loss/train': 1.8062001466751099} 02/25/2022 12:07:48 - INFO - codeparrot_training - Step 26662: {'lr': 0.0002391714391448625, 'samples': 13651456, 'steps': 26662, 'loss/train': 1.9824061393737793} 02/25/2022 12:07:53 - INFO - codeparrot_training - Step 26663: {'lr': 0.00023915509206253667, 'samples': 13651968, 'steps': 26663, 'loss/train': 1.5481623411178589} 02/25/2022 12:07:57 - INFO - codeparrot_training - Step 26664: {'lr': 0.00023913874502666697, 'samples': 13652480, 'steps': 26664, 'loss/train': 0.2746760845184326} 02/25/2022 12:08:02 - INFO - codeparrot_training - Step 26665: {'lr': 0.00023912239803732354, 'samples': 13652992, 'steps': 26665, 'loss/train': 8.848331451416016} 02/25/2022 12:08:06 - INFO - codeparrot_training - Step 26666: {'lr': 0.00023910605109457624, 'samples': 13653504, 'steps': 26666, 'loss/train': 2.227452516555786} 02/25/2022 12:08:12 - INFO - codeparrot_training - Step 26667: {'lr': 0.0002390897041984951, 'samples': 13654016, 'steps': 26667, 'loss/train': 2.1106221675872803} 02/25/2022 12:08:15 - INFO - codeparrot_training - Step 26668: {'lr': 0.00023907335734915024, 'samples': 13654528, 'steps': 26668, 'loss/train': 0.7590692043304443} 02/25/2022 12:08:21 - INFO - codeparrot_training - Step 26669: {'lr': 0.00023905701054661166, 'samples': 13655040, 'steps': 26669, 'loss/train': 1.7530666589736938} 02/25/2022 12:08:24 - INFO - codeparrot_training - Step 26670: {'lr': 0.00023904066379094941, 'samples': 13655552, 'steps': 26670, 'loss/train': 1.820556879043579} 02/25/2022 12:08:30 - INFO - codeparrot_training - Step 26671: {'lr': 0.00023902431708223342, 'samples': 13656064, 'steps': 26671, 'loss/train': 2.796388626098633} 02/25/2022 12:08:33 - INFO - codeparrot_training - Step 26672: {'lr': 0.00023900797042053382, 'samples': 13656576, 'steps': 26672, 'loss/train': 1.490715503692627} 02/25/2022 12:08:39 - INFO - codeparrot_training - Step 26673: {'lr': 0.00023899162380592052, 'samples': 13657088, 'steps': 26673, 'loss/train': 0.7696727514266968} 02/25/2022 12:08:42 - INFO - codeparrot_training - Step 26674: {'lr': 0.00023897527723846365, 'samples': 13657600, 'steps': 26674, 'loss/train': 1.3364040851593018} 02/25/2022 12:08:48 - INFO - codeparrot_training - Step 26675: {'lr': 0.0002389589307182332, 'samples': 13658112, 'steps': 26675, 'loss/train': 1.5182671546936035} 02/25/2022 12:08:51 - INFO - codeparrot_training - Step 26676: {'lr': 0.00023894258424529916, 'samples': 13658624, 'steps': 26676, 'loss/train': 1.4722493886947632} 02/25/2022 12:08:58 - INFO - codeparrot_training - Step 26677: {'lr': 0.00023892623781973158, 'samples': 13659136, 'steps': 26677, 'loss/train': 0.5243614912033081} 02/25/2022 12:09:02 - INFO - codeparrot_training - Step 26678: {'lr': 0.00023890989144160045, 'samples': 13659648, 'steps': 26678, 'loss/train': 2.0051987171173096} 02/25/2022 12:09:07 - INFO - codeparrot_training - Step 26679: {'lr': 0.0002388935451109759, 'samples': 13660160, 'steps': 26679, 'loss/train': 0.909983217716217} 02/25/2022 12:09:13 - INFO - codeparrot_training - Step 26680: {'lr': 0.00023887719882792786, 'samples': 13660672, 'steps': 26680, 'loss/train': 1.8194165229797363} 02/25/2022 12:09:16 - INFO - codeparrot_training - Step 26681: {'lr': 0.00023886085259252628, 'samples': 13661184, 'steps': 26681, 'loss/train': 1.961308240890503} 02/25/2022 12:09:22 - INFO - codeparrot_training - Step 26682: {'lr': 0.00023884450640484134, 'samples': 13661696, 'steps': 26682, 'loss/train': 2.4702696800231934} 02/25/2022 12:09:25 - INFO - codeparrot_training - Step 26683: {'lr': 0.00023882816026494306, 'samples': 13662208, 'steps': 26683, 'loss/train': 1.9763925075531006} 02/25/2022 12:09:31 - INFO - codeparrot_training - Step 26684: {'lr': 0.00023881181417290129, 'samples': 13662720, 'steps': 26684, 'loss/train': 1.8367406129837036} 02/25/2022 12:09:34 - INFO - codeparrot_training - Step 26685: {'lr': 0.00023879546812878618, 'samples': 13663232, 'steps': 26685, 'loss/train': 1.3323067426681519} 02/25/2022 12:09:40 - INFO - codeparrot_training - Step 26686: {'lr': 0.0002387791221326677, 'samples': 13663744, 'steps': 26686, 'loss/train': 2.693667411804199} 02/25/2022 12:09:44 - INFO - codeparrot_training - Step 26687: {'lr': 0.000238762776184616, 'samples': 13664256, 'steps': 26687, 'loss/train': 1.696393370628357} 02/25/2022 12:09:49 - INFO - codeparrot_training - Step 26688: {'lr': 0.0002387464302847009, 'samples': 13664768, 'steps': 26688, 'loss/train': 1.588341474533081} 02/25/2022 12:09:53 - INFO - codeparrot_training - Step 26689: {'lr': 0.00023873008443299254, 'samples': 13665280, 'steps': 26689, 'loss/train': 1.1500643491744995} 02/25/2022 12:09:58 - INFO - codeparrot_training - Step 26690: {'lr': 0.00023871373862956088, 'samples': 13665792, 'steps': 26690, 'loss/train': 2.7455103397369385} 02/25/2022 12:10:02 - INFO - codeparrot_training - Step 26691: {'lr': 0.00023869739287447602, 'samples': 13666304, 'steps': 26691, 'loss/train': 2.0966193675994873} 02/25/2022 12:10:07 - INFO - codeparrot_training - Step 26692: {'lr': 0.00023868104716780802, 'samples': 13666816, 'steps': 26692, 'loss/train': 1.7433397769927979} 02/25/2022 12:10:11 - INFO - codeparrot_training - Step 26693: {'lr': 0.00023866470150962673, 'samples': 13667328, 'steps': 26693, 'loss/train': 1.6447679996490479} 02/25/2022 12:10:16 - INFO - codeparrot_training - Step 26694: {'lr': 0.00023864835590000225, 'samples': 13667840, 'steps': 26694, 'loss/train': 2.9193737506866455} 02/25/2022 12:10:20 - INFO - codeparrot_training - Step 26695: {'lr': 0.00023863201033900462, 'samples': 13668352, 'steps': 26695, 'loss/train': 2.1003835201263428} 02/25/2022 12:10:25 - INFO - codeparrot_training - Step 26696: {'lr': 0.00023861566482670393, 'samples': 13668864, 'steps': 26696, 'loss/train': 0.4295022189617157} 02/25/2022 12:10:29 - INFO - codeparrot_training - Step 26697: {'lr': 0.00023859931936317003, 'samples': 13669376, 'steps': 26697, 'loss/train': 0.7564507722854614} 02/25/2022 12:10:34 - INFO - codeparrot_training - Step 26698: {'lr': 0.00023858297394847302, 'samples': 13669888, 'steps': 26698, 'loss/train': 2.08242130279541} 02/25/2022 12:10:38 - INFO - codeparrot_training - Step 26699: {'lr': 0.00023856662858268293, 'samples': 13670400, 'steps': 26699, 'loss/train': 1.5856281518936157} 02/25/2022 12:10:43 - INFO - codeparrot_training - Step 26700: {'lr': 0.00023855028326586985, 'samples': 13670912, 'steps': 26700, 'loss/train': 0.8329271674156189} 02/25/2022 12:10:47 - INFO - codeparrot_training - Step 26701: {'lr': 0.00023853393799810367, 'samples': 13671424, 'steps': 26701, 'loss/train': 0.9876444339752197} 02/25/2022 12:10:53 - INFO - codeparrot_training - Step 26702: {'lr': 0.00023851759277945442, 'samples': 13671936, 'steps': 26702, 'loss/train': 2.265570640563965} 02/25/2022 12:10:57 - INFO - codeparrot_training - Step 26703: {'lr': 0.00023850124760999217, 'samples': 13672448, 'steps': 26703, 'loss/train': 1.5717158317565918} 02/25/2022 12:11:02 - INFO - codeparrot_training - Step 26704: {'lr': 0.0002384849024897869, 'samples': 13672960, 'steps': 26704, 'loss/train': 1.2222161293029785} 02/25/2022 12:11:06 - INFO - codeparrot_training - Step 26705: {'lr': 0.0002384685574189088, 'samples': 13673472, 'steps': 26705, 'loss/train': 1.661441683769226} 02/25/2022 12:11:11 - INFO - codeparrot_training - Step 26706: {'lr': 0.00023845221239742758, 'samples': 13673984, 'steps': 26706, 'loss/train': 2.177128791809082} 02/25/2022 12:11:15 - INFO - codeparrot_training - Step 26707: {'lr': 0.00023843586742541346, 'samples': 13674496, 'steps': 26707, 'loss/train': 1.0078840255737305} 02/25/2022 12:11:20 - INFO - codeparrot_training - Step 26708: {'lr': 0.0002384195225029364, 'samples': 13675008, 'steps': 26708, 'loss/train': 1.187821388244629} 02/25/2022 12:11:24 - INFO - codeparrot_training - Step 26709: {'lr': 0.0002384031776300665, 'samples': 13675520, 'steps': 26709, 'loss/train': 1.7327485084533691} 02/25/2022 12:11:29 - INFO - codeparrot_training - Step 26710: {'lr': 0.00023838683280687364, 'samples': 13676032, 'steps': 26710, 'loss/train': 2.5005273818969727} 02/25/2022 12:11:33 - INFO - codeparrot_training - Step 26711: {'lr': 0.0002383704880334279, 'samples': 13676544, 'steps': 26711, 'loss/train': 2.4751930236816406} 02/25/2022 12:11:36 - INFO - codeparrot_training - Step 26712: {'lr': 0.00023835414330979928, 'samples': 13677056, 'steps': 26712, 'loss/train': 2.3055968284606934} 02/25/2022 12:11:42 - INFO - codeparrot_training - Step 26713: {'lr': 0.00023833779863605793, 'samples': 13677568, 'steps': 26713, 'loss/train': 1.997613549232483} 02/25/2022 12:11:48 - INFO - codeparrot_training - Step 26714: {'lr': 0.00023832145401227363, 'samples': 13678080, 'steps': 26714, 'loss/train': 2.6117136478424072} 02/25/2022 12:11:51 - INFO - codeparrot_training - Step 26715: {'lr': 0.00023830510943851653, 'samples': 13678592, 'steps': 26715, 'loss/train': 1.553622841835022} 02/25/2022 12:11:57 - INFO - codeparrot_training - Step 26716: {'lr': 0.00023828876491485665, 'samples': 13679104, 'steps': 26716, 'loss/train': 2.3320722579956055} 02/25/2022 12:12:00 - INFO - codeparrot_training - Step 26717: {'lr': 0.00023827242044136395, 'samples': 13679616, 'steps': 26717, 'loss/train': 2.0440430641174316} 02/25/2022 12:12:06 - INFO - codeparrot_training - Step 26718: {'lr': 0.00023825607601810853, 'samples': 13680128, 'steps': 26718, 'loss/train': 1.2034085988998413} 02/25/2022 12:12:09 - INFO - codeparrot_training - Step 26719: {'lr': 0.0002382397316451603, 'samples': 13680640, 'steps': 26719, 'loss/train': 2.004425048828125} 02/25/2022 12:12:15 - INFO - codeparrot_training - Step 26720: {'lr': 0.00023822338732258937, 'samples': 13681152, 'steps': 26720, 'loss/train': 0.8579483032226562} 02/25/2022 12:12:18 - INFO - codeparrot_training - Step 26721: {'lr': 0.00023820704305046564, 'samples': 13681664, 'steps': 26721, 'loss/train': 1.320696234703064} 02/25/2022 12:12:25 - INFO - codeparrot_training - Step 26722: {'lr': 0.00023819069882885928, 'samples': 13682176, 'steps': 26722, 'loss/train': 1.1861369609832764} 02/25/2022 12:12:28 - INFO - codeparrot_training - Step 26723: {'lr': 0.00023817435465784022, 'samples': 13682688, 'steps': 26723, 'loss/train': 2.611652135848999} 02/25/2022 12:12:34 - INFO - codeparrot_training - Step 26724: {'lr': 0.00023815801053747837, 'samples': 13683200, 'steps': 26724, 'loss/train': 1.8311712741851807} 02/25/2022 12:12:37 - INFO - codeparrot_training - Step 26725: {'lr': 0.0002381416664678439, 'samples': 13683712, 'steps': 26725, 'loss/train': 2.5182249546051025} 02/25/2022 12:12:43 - INFO - codeparrot_training - Step 26726: {'lr': 0.00023812532244900673, 'samples': 13684224, 'steps': 26726, 'loss/train': 1.261854887008667} 02/25/2022 12:12:46 - INFO - codeparrot_training - Step 26727: {'lr': 0.00023810897848103703, 'samples': 13684736, 'steps': 26727, 'loss/train': 2.2077341079711914} 02/25/2022 12:12:52 - INFO - codeparrot_training - Step 26728: {'lr': 0.00023809263456400456, 'samples': 13685248, 'steps': 26728, 'loss/train': 2.747763156890869} 02/25/2022 12:12:55 - INFO - codeparrot_training - Step 26729: {'lr': 0.00023807629069797948, 'samples': 13685760, 'steps': 26729, 'loss/train': 1.0291587114334106} 02/25/2022 12:13:01 - INFO - codeparrot_training - Step 26730: {'lr': 0.00023805994688303177, 'samples': 13686272, 'steps': 26730, 'loss/train': 2.311230182647705} 02/25/2022 12:13:05 - INFO - codeparrot_training - Step 26731: {'lr': 0.00023804360311923157, 'samples': 13686784, 'steps': 26731, 'loss/train': 2.2481613159179688} 02/25/2022 12:13:08 - INFO - codeparrot_training - Step 26732: {'lr': 0.00023802725940664867, 'samples': 13687296, 'steps': 26732, 'loss/train': 2.0449531078338623} 02/25/2022 12:13:14 - INFO - codeparrot_training - Step 26733: {'lr': 0.00023801091574535322, 'samples': 13687808, 'steps': 26733, 'loss/train': 1.5576388835906982} 02/25/2022 12:13:18 - INFO - codeparrot_training - Step 26734: {'lr': 0.00023799457213541516, 'samples': 13688320, 'steps': 26734, 'loss/train': 1.0398763418197632} 02/25/2022 12:13:23 - INFO - codeparrot_training - Step 26735: {'lr': 0.00023797822857690465, 'samples': 13688832, 'steps': 26735, 'loss/train': 2.205842971801758} 02/25/2022 12:13:27 - INFO - codeparrot_training - Step 26736: {'lr': 0.00023796188506989153, 'samples': 13689344, 'steps': 26736, 'loss/train': 2.0860581398010254} 02/25/2022 12:13:32 - INFO - codeparrot_training - Step 26737: {'lr': 0.00023794554161444582, 'samples': 13689856, 'steps': 26737, 'loss/train': 0.382570743560791} 02/25/2022 12:13:36 - INFO - codeparrot_training - Step 26738: {'lr': 0.0002379291982106376, 'samples': 13690368, 'steps': 26738, 'loss/train': 0.4131981134414673} 02/25/2022 12:13:41 - INFO - codeparrot_training - Step 26739: {'lr': 0.00023791285485853686, 'samples': 13690880, 'steps': 26739, 'loss/train': 1.860669493675232} 02/25/2022 12:13:45 - INFO - codeparrot_training - Step 26740: {'lr': 0.00023789651155821373, 'samples': 13691392, 'steps': 26740, 'loss/train': 1.9211090803146362} 02/25/2022 12:13:50 - INFO - codeparrot_training - Step 26741: {'lr': 0.00023788016830973797, 'samples': 13691904, 'steps': 26741, 'loss/train': 2.520002841949463} 02/25/2022 12:13:54 - INFO - codeparrot_training - Step 26742: {'lr': 0.00023786382511317976, 'samples': 13692416, 'steps': 26742, 'loss/train': 1.8196336030960083} 02/25/2022 12:14:00 - INFO - codeparrot_training - Step 26743: {'lr': 0.00023784748196860903, 'samples': 13692928, 'steps': 26743, 'loss/train': 2.385216474533081} 02/25/2022 12:14:05 - INFO - codeparrot_training - Step 26744: {'lr': 0.00023783113887609596, 'samples': 13693440, 'steps': 26744, 'loss/train': 1.892880916595459} 02/25/2022 12:14:09 - INFO - codeparrot_training - Step 26745: {'lr': 0.0002378147958357103, 'samples': 13693952, 'steps': 26745, 'loss/train': 1.8789676427841187} 02/25/2022 12:14:14 - INFO - codeparrot_training - Step 26746: {'lr': 0.00023779845284752218, 'samples': 13694464, 'steps': 26746, 'loss/train': 1.783031940460205} 02/25/2022 12:14:18 - INFO - codeparrot_training - Step 26747: {'lr': 0.00023778210991160165, 'samples': 13694976, 'steps': 26747, 'loss/train': 1.2729867696762085} 02/25/2022 12:14:24 - INFO - codeparrot_training - Step 26748: {'lr': 0.0002377657670280188, 'samples': 13695488, 'steps': 26748, 'loss/train': 1.7813818454742432} 02/25/2022 12:14:27 - INFO - codeparrot_training - Step 26749: {'lr': 0.00023774942419684333, 'samples': 13696000, 'steps': 26749, 'loss/train': 1.202014446258545} 02/25/2022 12:14:31 - INFO - codeparrot_training - Step 26750: {'lr': 0.0002377330814181455, 'samples': 13696512, 'steps': 26750, 'loss/train': 1.443450927734375} 02/25/2022 12:14:36 - INFO - codeparrot_training - Step 26751: {'lr': 0.00023771673869199525, 'samples': 13697024, 'steps': 26751, 'loss/train': 0.7878649830818176} 02/25/2022 12:14:40 - INFO - codeparrot_training - Step 26752: {'lr': 0.00023770039601846257, 'samples': 13697536, 'steps': 26752, 'loss/train': 1.4083608388900757} 02/25/2022 12:14:45 - INFO - codeparrot_training - Step 26753: {'lr': 0.00023768405339761762, 'samples': 13698048, 'steps': 26753, 'loss/train': 2.113192081451416} 02/25/2022 12:14:49 - INFO - codeparrot_training - Step 26754: {'lr': 0.00023766771082953014, 'samples': 13698560, 'steps': 26754, 'loss/train': 1.9707279205322266} 02/25/2022 12:14:54 - INFO - codeparrot_training - Step 26755: {'lr': 0.0002376513683142703, 'samples': 13699072, 'steps': 26755, 'loss/train': 0.7499202489852905} 02/25/2022 12:15:00 - INFO - codeparrot_training - Step 26756: {'lr': 0.0002376350258519081, 'samples': 13699584, 'steps': 26756, 'loss/train': 2.2111752033233643} 02/25/2022 12:15:03 - INFO - codeparrot_training - Step 26757: {'lr': 0.00023761868344251356, 'samples': 13700096, 'steps': 26757, 'loss/train': 1.2791670560836792} 02/25/2022 12:15:07 - INFO - codeparrot_training - Step 26758: {'lr': 0.00023760234108615656, 'samples': 13700608, 'steps': 26758, 'loss/train': 0.6070950031280518} 02/25/2022 12:15:13 - INFO - codeparrot_training - Step 26759: {'lr': 0.00023758599878290722, 'samples': 13701120, 'steps': 26759, 'loss/train': 1.8183608055114746} 02/25/2022 12:15:17 - INFO - codeparrot_training - Step 26760: {'lr': 0.00023756965653283556, 'samples': 13701632, 'steps': 26760, 'loss/train': 1.6736971139907837} 02/25/2022 12:15:22 - INFO - codeparrot_training - Step 26761: {'lr': 0.00023755331433601153, 'samples': 13702144, 'steps': 26761, 'loss/train': 1.8272558450698853} 02/25/2022 12:15:26 - INFO - codeparrot_training - Step 26762: {'lr': 0.00023753697219250513, 'samples': 13702656, 'steps': 26762, 'loss/train': 2.012540102005005} 02/25/2022 12:15:31 - INFO - codeparrot_training - Step 26763: {'lr': 0.0002375206301023864, 'samples': 13703168, 'steps': 26763, 'loss/train': 1.3723292350769043} 02/25/2022 12:15:35 - INFO - codeparrot_training - Step 26764: {'lr': 0.0002375042880657253, 'samples': 13703680, 'steps': 26764, 'loss/train': 2.2657365798950195} 02/25/2022 12:15:40 - INFO - codeparrot_training - Step 26765: {'lr': 0.00023748794608259186, 'samples': 13704192, 'steps': 26765, 'loss/train': 1.8205993175506592} 02/25/2022 12:15:44 - INFO - codeparrot_training - Step 26766: {'lr': 0.00023747160415305612, 'samples': 13704704, 'steps': 26766, 'loss/train': 1.750273585319519} 02/25/2022 12:15:49 - INFO - codeparrot_training - Step 26767: {'lr': 0.00023745526227718808, 'samples': 13705216, 'steps': 26767, 'loss/train': 1.5082825422286987} 02/25/2022 12:15:53 - INFO - codeparrot_training - Step 26768: {'lr': 0.00023743892045505763, 'samples': 13705728, 'steps': 26768, 'loss/train': 1.7371412515640259} 02/25/2022 12:15:59 - INFO - codeparrot_training - Step 26769: {'lr': 0.00023742257868673486, 'samples': 13706240, 'steps': 26769, 'loss/train': 1.9277106523513794} 02/25/2022 12:16:03 - INFO - codeparrot_training - Step 26770: {'lr': 0.00023740623697228992, 'samples': 13706752, 'steps': 26770, 'loss/train': 1.839339256286621} 02/25/2022 12:16:08 - INFO - codeparrot_training - Step 26771: {'lr': 0.0002373898953117925, 'samples': 13707264, 'steps': 26771, 'loss/train': 2.537174940109253} 02/25/2022 12:16:12 - INFO - codeparrot_training - Step 26772: {'lr': 0.00023737355370531278, 'samples': 13707776, 'steps': 26772, 'loss/train': 1.7278813123703003} 02/25/2022 12:16:17 - INFO - codeparrot_training - Step 26773: {'lr': 0.00023735721215292074, 'samples': 13708288, 'steps': 26773, 'loss/train': 1.8701725006103516} 02/25/2022 12:16:21 - INFO - codeparrot_training - Step 26774: {'lr': 0.0002373408706546865, 'samples': 13708800, 'steps': 26774, 'loss/train': 1.7911018133163452} 02/25/2022 12:16:26 - INFO - codeparrot_training - Step 26775: {'lr': 0.00023732452921067988, 'samples': 13709312, 'steps': 26775, 'loss/train': 2.0169239044189453} 02/25/2022 12:16:30 - INFO - codeparrot_training - Step 26776: {'lr': 0.00023730818782097092, 'samples': 13709824, 'steps': 26776, 'loss/train': 1.7721707820892334} 02/25/2022 12:16:35 - INFO - codeparrot_training - Step 26777: {'lr': 0.00023729184648562965, 'samples': 13710336, 'steps': 26777, 'loss/train': 1.4347071647644043} 02/25/2022 12:16:39 - INFO - codeparrot_training - Step 26778: {'lr': 0.00023727550520472607, 'samples': 13710848, 'steps': 26778, 'loss/train': 2.0113892555236816} 02/25/2022 12:16:45 - INFO - codeparrot_training - Step 26779: {'lr': 0.00023725916397833032, 'samples': 13711360, 'steps': 26779, 'loss/train': 2.8242876529693604} 02/25/2022 12:16:49 - INFO - codeparrot_training - Step 26780: {'lr': 0.00023724282280651214, 'samples': 13711872, 'steps': 26780, 'loss/train': 1.09941828250885} 02/25/2022 12:16:54 - INFO - codeparrot_training - Step 26781: {'lr': 0.00023722648168934165, 'samples': 13712384, 'steps': 26781, 'loss/train': 4.165886402130127} 02/25/2022 12:16:58 - INFO - codeparrot_training - Step 26782: {'lr': 0.00023721014062688886, 'samples': 13712896, 'steps': 26782, 'loss/train': 1.1417133808135986} 02/25/2022 12:17:03 - INFO - codeparrot_training - Step 26783: {'lr': 0.00023719379961922388, 'samples': 13713408, 'steps': 26783, 'loss/train': 1.5448094606399536} 02/25/2022 12:17:07 - INFO - codeparrot_training - Step 26784: {'lr': 0.0002371774586664165, 'samples': 13713920, 'steps': 26784, 'loss/train': 2.5177090167999268} 02/25/2022 12:17:12 - INFO - codeparrot_training - Step 26785: {'lr': 0.00023716111776853679, 'samples': 13714432, 'steps': 26785, 'loss/train': 1.6660338640213013} 02/25/2022 12:17:16 - INFO - codeparrot_training - Step 26786: {'lr': 0.00023714477692565478, 'samples': 13714944, 'steps': 26786, 'loss/train': 2.424240827560425} 02/25/2022 12:17:21 - INFO - codeparrot_training - Step 26787: {'lr': 0.00023712843613784047, 'samples': 13715456, 'steps': 26787, 'loss/train': 1.3898310661315918} 02/25/2022 12:17:25 - INFO - codeparrot_training - Step 26788: {'lr': 0.00023711209540516396, 'samples': 13715968, 'steps': 26788, 'loss/train': 2.6359150409698486} 02/25/2022 12:17:31 - INFO - codeparrot_training - Step 26789: {'lr': 0.00023709575472769505, 'samples': 13716480, 'steps': 26789, 'loss/train': 2.998382091522217} 02/25/2022 12:17:35 - INFO - codeparrot_training - Step 26790: {'lr': 0.0002370794141055038, 'samples': 13716992, 'steps': 26790, 'loss/train': 2.6061904430389404} 02/25/2022 12:17:40 - INFO - codeparrot_training - Step 26791: {'lr': 0.00023706307353866025, 'samples': 13717504, 'steps': 26791, 'loss/train': 0.48793825507164} 02/25/2022 12:17:44 - INFO - codeparrot_training - Step 26792: {'lr': 0.00023704673302723449, 'samples': 13718016, 'steps': 26792, 'loss/train': 1.3789302110671997} 02/25/2022 12:17:49 - INFO - codeparrot_training - Step 26793: {'lr': 0.0002370303925712963, 'samples': 13718528, 'steps': 26793, 'loss/train': 1.8587357997894287} 02/25/2022 12:17:53 - INFO - codeparrot_training - Step 26794: {'lr': 0.00023701405217091583, 'samples': 13719040, 'steps': 26794, 'loss/train': 1.4766002893447876} 02/25/2022 12:17:58 - INFO - codeparrot_training - Step 26795: {'lr': 0.00023699771182616298, 'samples': 13719552, 'steps': 26795, 'loss/train': 0.9420468807220459} 02/25/2022 12:18:02 - INFO - codeparrot_training - Step 26796: {'lr': 0.00023698137153710795, 'samples': 13720064, 'steps': 26796, 'loss/train': 1.0802677869796753} 02/25/2022 12:18:07 - INFO - codeparrot_training - Step 26797: {'lr': 0.00023696503130382044, 'samples': 13720576, 'steps': 26797, 'loss/train': 1.8587385416030884} 02/25/2022 12:18:11 - INFO - codeparrot_training - Step 26798: {'lr': 0.00023694869112637063, 'samples': 13721088, 'steps': 26798, 'loss/train': 0.8755258321762085} 02/25/2022 12:18:16 - INFO - codeparrot_training - Step 26799: {'lr': 0.0002369323510048285, 'samples': 13721600, 'steps': 26799, 'loss/train': 1.7706990242004395} 02/25/2022 12:18:19 - INFO - codeparrot_training - Step 26800: {'lr': 0.00023691601093926405, 'samples': 13722112, 'steps': 26800, 'loss/train': 1.8594273328781128} 02/25/2022 12:18:25 - INFO - codeparrot_training - Step 26801: {'lr': 0.00023689967092974728, 'samples': 13722624, 'steps': 26801, 'loss/train': 2.4129745960235596} 02/25/2022 12:18:28 - INFO - codeparrot_training - Step 26802: {'lr': 0.00023688333097634807, 'samples': 13723136, 'steps': 26802, 'loss/train': 1.707031488418579} 02/25/2022 12:18:34 - INFO - codeparrot_training - Step 26803: {'lr': 0.00023686699107913658, 'samples': 13723648, 'steps': 26803, 'loss/train': 1.1110788583755493} 02/25/2022 12:18:37 - INFO - codeparrot_training - Step 26804: {'lr': 0.00023685065123818267, 'samples': 13724160, 'steps': 26804, 'loss/train': 2.5814056396484375} 02/25/2022 12:18:44 - INFO - codeparrot_training - Step 26805: {'lr': 0.00023683431145355647, 'samples': 13724672, 'steps': 26805, 'loss/train': 2.2575063705444336} 02/25/2022 12:18:47 - INFO - codeparrot_training - Step 26806: {'lr': 0.0002368179717253278, 'samples': 13725184, 'steps': 26806, 'loss/train': 0.48791399598121643} 02/25/2022 12:18:53 - INFO - codeparrot_training - Step 26807: {'lr': 0.00023680163205356687, 'samples': 13725696, 'steps': 26807, 'loss/train': 1.4099923372268677} 02/25/2022 12:18:56 - INFO - codeparrot_training - Step 26808: {'lr': 0.00023678529243834347, 'samples': 13726208, 'steps': 26808, 'loss/train': 1.836795449256897} 02/25/2022 12:19:02 - INFO - codeparrot_training - Step 26809: {'lr': 0.00023676895287972773, 'samples': 13726720, 'steps': 26809, 'loss/train': 1.2395943403244019} 02/25/2022 12:19:05 - INFO - codeparrot_training - Step 26810: {'lr': 0.00023675261337778957, 'samples': 13727232, 'steps': 26810, 'loss/train': 2.0408270359039307} 02/25/2022 12:19:10 - INFO - codeparrot_training - Step 26811: {'lr': 0.000236736273932599, 'samples': 13727744, 'steps': 26811, 'loss/train': 2.1260390281677246} 02/25/2022 12:19:14 - INFO - codeparrot_training - Step 26812: {'lr': 0.00023671993454422603, 'samples': 13728256, 'steps': 26812, 'loss/train': 1.8632508516311646} 02/25/2022 12:19:20 - INFO - codeparrot_training - Step 26813: {'lr': 0.00023670359521274059, 'samples': 13728768, 'steps': 26813, 'loss/train': 1.5697566270828247} 02/25/2022 12:19:23 - INFO - codeparrot_training - Step 26814: {'lr': 0.0002366872559382129, 'samples': 13729280, 'steps': 26814, 'loss/train': 2.268803834915161} 02/25/2022 12:19:29 - INFO - codeparrot_training - Step 26815: {'lr': 0.00023667091672071263, 'samples': 13729792, 'steps': 26815, 'loss/train': 2.0402042865753174} 02/25/2022 12:19:33 - INFO - codeparrot_training - Step 26816: {'lr': 0.0002366545775603099, 'samples': 13730304, 'steps': 26816, 'loss/train': 1.8463525772094727} 02/25/2022 12:19:38 - INFO - codeparrot_training - Step 26817: {'lr': 0.00023663823845707476, 'samples': 13730816, 'steps': 26817, 'loss/train': 1.9438560009002686} 02/25/2022 12:19:42 - INFO - codeparrot_training - Step 26818: {'lr': 0.00023662189941107726, 'samples': 13731328, 'steps': 26818, 'loss/train': 1.1925404071807861} 02/25/2022 12:19:47 - INFO - codeparrot_training - Step 26819: {'lr': 0.00023660556042238718, 'samples': 13731840, 'steps': 26819, 'loss/train': 1.205910563468933} 02/25/2022 12:19:51 - INFO - codeparrot_training - Step 26820: {'lr': 0.0002365892214910746, 'samples': 13732352, 'steps': 26820, 'loss/train': 2.38867449760437} 02/25/2022 12:19:57 - INFO - codeparrot_training - Step 26821: {'lr': 0.00023657288261720957, 'samples': 13732864, 'steps': 26821, 'loss/train': 1.930776596069336} 02/25/2022 12:20:00 - INFO - codeparrot_training - Step 26822: {'lr': 0.00023655654380086216, 'samples': 13733376, 'steps': 26822, 'loss/train': 2.257798194885254} 02/25/2022 12:20:05 - INFO - codeparrot_training - Step 26823: {'lr': 0.00023654020504210215, 'samples': 13733888, 'steps': 26823, 'loss/train': 1.8421812057495117} 02/25/2022 12:20:09 - INFO - codeparrot_training - Step 26824: {'lr': 0.0002365238663409996, 'samples': 13734400, 'steps': 26824, 'loss/train': 1.8362871408462524} 02/25/2022 12:20:16 - INFO - codeparrot_training - Step 26825: {'lr': 0.0002365075276976245, 'samples': 13734912, 'steps': 26825, 'loss/train': 1.683214783668518} 02/25/2022 12:20:19 - INFO - codeparrot_training - Step 26826: {'lr': 0.00023649118911204693, 'samples': 13735424, 'steps': 26826, 'loss/train': 2.1748716831207275} 02/25/2022 12:20:25 - INFO - codeparrot_training - Step 26827: {'lr': 0.0002364748505843369, 'samples': 13735936, 'steps': 26827, 'loss/train': 1.2752037048339844} 02/25/2022 12:20:28 - INFO - codeparrot_training - Step 26828: {'lr': 0.0002364585121145642, 'samples': 13736448, 'steps': 26828, 'loss/train': 1.4916974306106567} 02/25/2022 12:20:34 - INFO - codeparrot_training - Step 26829: {'lr': 0.00023644217370279893, 'samples': 13736960, 'steps': 26829, 'loss/train': 4.189927101135254} 02/25/2022 12:20:37 - INFO - codeparrot_training - Step 26830: {'lr': 0.0002364258353491111, 'samples': 13737472, 'steps': 26830, 'loss/train': 2.129122018814087} 02/25/2022 12:20:43 - INFO - codeparrot_training - Step 26831: {'lr': 0.00023640949705357078, 'samples': 13737984, 'steps': 26831, 'loss/train': 1.9369949102401733} 02/25/2022 12:20:46 - INFO - codeparrot_training - Step 26832: {'lr': 0.00023639315881624775, 'samples': 13738496, 'steps': 26832, 'loss/train': 0.6894099116325378} 02/25/2022 12:20:52 - INFO - codeparrot_training - Step 26833: {'lr': 0.0002363768206372121, 'samples': 13739008, 'steps': 26833, 'loss/train': 0.18040867149829865} 02/25/2022 12:20:55 - INFO - codeparrot_training - Step 26834: {'lr': 0.00023636048251653386, 'samples': 13739520, 'steps': 26834, 'loss/train': 0.7886021137237549} 02/25/2022 12:21:01 - INFO - codeparrot_training - Step 26835: {'lr': 0.00023634414445428297, 'samples': 13740032, 'steps': 26835, 'loss/train': 2.280454158782959} 02/25/2022 12:21:05 - INFO - codeparrot_training - Step 26836: {'lr': 0.0002363278064505295, 'samples': 13740544, 'steps': 26836, 'loss/train': 2.167886257171631} 02/25/2022 12:21:10 - INFO - codeparrot_training - Step 26837: {'lr': 0.00023631146850534328, 'samples': 13741056, 'steps': 26837, 'loss/train': 2.3823812007904053} 02/25/2022 12:21:14 - INFO - codeparrot_training - Step 26838: {'lr': 0.00023629513061879437, 'samples': 13741568, 'steps': 26838, 'loss/train': 1.8117777109146118} 02/25/2022 12:21:19 - INFO - codeparrot_training - Step 26839: {'lr': 0.0002362787927909528, 'samples': 13742080, 'steps': 26839, 'loss/train': 1.6279933452606201} 02/25/2022 12:21:23 - INFO - codeparrot_training - Step 26840: {'lr': 0.00023626245502188863, 'samples': 13742592, 'steps': 26840, 'loss/train': 1.2928667068481445} 02/25/2022 12:21:29 - INFO - codeparrot_training - Step 26841: {'lr': 0.0002362461173116716, 'samples': 13743104, 'steps': 26841, 'loss/train': 0.9796396493911743} 02/25/2022 12:21:32 - INFO - codeparrot_training - Step 26842: {'lr': 0.00023622977966037184, 'samples': 13743616, 'steps': 26842, 'loss/train': 1.7577086687088013} 02/25/2022 12:21:38 - INFO - codeparrot_training - Step 26843: {'lr': 0.0002362134420680594, 'samples': 13744128, 'steps': 26843, 'loss/train': 1.9885544776916504} 02/25/2022 12:21:41 - INFO - codeparrot_training - Step 26844: {'lr': 0.0002361971045348042, 'samples': 13744640, 'steps': 26844, 'loss/train': 1.3850375413894653} 02/25/2022 12:21:47 - INFO - codeparrot_training - Step 26845: {'lr': 0.00023618076706067616, 'samples': 13745152, 'steps': 26845, 'loss/train': 1.7462857961654663} 02/25/2022 12:21:50 - INFO - codeparrot_training - Step 26846: {'lr': 0.00023616442964574533, 'samples': 13745664, 'steps': 26846, 'loss/train': 1.104557991027832} 02/25/2022 12:21:56 - INFO - codeparrot_training - Step 26847: {'lr': 0.00023614809229008173, 'samples': 13746176, 'steps': 26847, 'loss/train': 1.7320835590362549} 02/25/2022 12:21:59 - INFO - codeparrot_training - Step 26848: {'lr': 0.00023613175499375528, 'samples': 13746688, 'steps': 26848, 'loss/train': 1.4686414003372192} 02/25/2022 12:22:05 - INFO - codeparrot_training - Step 26849: {'lr': 0.000236115417756836, 'samples': 13747200, 'steps': 26849, 'loss/train': 1.4412899017333984} 02/25/2022 12:22:08 - INFO - codeparrot_training - Step 26850: {'lr': 0.00023609908057939388, 'samples': 13747712, 'steps': 26850, 'loss/train': 1.3898965120315552} 02/25/2022 12:22:14 - INFO - codeparrot_training - Step 26851: {'lr': 0.00023608274346149886, 'samples': 13748224, 'steps': 26851, 'loss/train': 2.3681750297546387} 02/25/2022 12:22:18 - INFO - codeparrot_training - Step 26852: {'lr': 0.00023606640640322092, 'samples': 13748736, 'steps': 26852, 'loss/train': 2.6245272159576416} 02/25/2022 12:22:23 - INFO - codeparrot_training - Step 26853: {'lr': 0.00023605006940463012, 'samples': 13749248, 'steps': 26853, 'loss/train': 1.9355206489562988} 02/25/2022 12:22:27 - INFO - codeparrot_training - Step 26854: {'lr': 0.0002360337324657964, 'samples': 13749760, 'steps': 26854, 'loss/train': 1.4596617221832275} 02/25/2022 12:22:32 - INFO - codeparrot_training - Step 26855: {'lr': 0.00023601739558678968, 'samples': 13750272, 'steps': 26855, 'loss/train': 1.1350302696228027} 02/25/2022 12:22:36 - INFO - codeparrot_training - Step 26856: {'lr': 0.00023600105876768, 'samples': 13750784, 'steps': 26856, 'loss/train': 1.9619797468185425} 02/25/2022 12:22:41 - INFO - codeparrot_training - Step 26857: {'lr': 0.00023598472200853747, 'samples': 13751296, 'steps': 26857, 'loss/train': 0.891645610332489} 02/25/2022 12:22:45 - INFO - codeparrot_training - Step 26858: {'lr': 0.0002359683853094318, 'samples': 13751808, 'steps': 26858, 'loss/train': 2.0408740043640137} 02/25/2022 12:22:50 - INFO - codeparrot_training - Step 26859: {'lr': 0.00023595204867043313, 'samples': 13752320, 'steps': 26859, 'loss/train': 1.3322694301605225} 02/25/2022 12:22:57 - INFO - codeparrot_training - Step 26860: {'lr': 0.00023593571209161142, 'samples': 13752832, 'steps': 26860, 'loss/train': 0.8712801337242126} 02/25/2022 12:23:00 - INFO - codeparrot_training - Step 26861: {'lr': 0.00023591937557303664, 'samples': 13753344, 'steps': 26861, 'loss/train': 1.4019553661346436} 02/25/2022 12:23:04 - INFO - codeparrot_training - Step 26862: {'lr': 0.0002359030391147789, 'samples': 13753856, 'steps': 26862, 'loss/train': 0.35387876629829407} 02/25/2022 12:23:09 - INFO - codeparrot_training - Step 26863: {'lr': 0.00023588670271690792, 'samples': 13754368, 'steps': 26863, 'loss/train': 0.9114536046981812} 02/25/2022 12:23:15 - INFO - codeparrot_training - Step 26864: {'lr': 0.00023587036637949389, 'samples': 13754880, 'steps': 26864, 'loss/train': 0.6249039173126221} 02/25/2022 12:23:19 - INFO - codeparrot_training - Step 26865: {'lr': 0.00023585403010260668, 'samples': 13755392, 'steps': 26865, 'loss/train': 1.0310841798782349} 02/25/2022 12:23:22 - INFO - codeparrot_training - Step 26866: {'lr': 0.00023583769388631643, 'samples': 13755904, 'steps': 26866, 'loss/train': 0.3120317757129669} 02/25/2022 12:23:28 - INFO - codeparrot_training - Step 26867: {'lr': 0.0002358213577306929, 'samples': 13756416, 'steps': 26867, 'loss/train': 1.968385100364685} 02/25/2022 12:23:31 - INFO - codeparrot_training - Step 26868: {'lr': 0.00023580502163580617, 'samples': 13756928, 'steps': 26868, 'loss/train': 3.264969825744629} 02/25/2022 12:23:37 - INFO - codeparrot_training - Step 26869: {'lr': 0.00023578868560172623, 'samples': 13757440, 'steps': 26869, 'loss/train': 1.6674898862838745} 02/25/2022 12:23:40 - INFO - codeparrot_training - Step 26870: {'lr': 0.00023577234962852316, 'samples': 13757952, 'steps': 26870, 'loss/train': 1.49990713596344} 02/25/2022 12:23:46 - INFO - codeparrot_training - Step 26871: {'lr': 0.0002357560137162667, 'samples': 13758464, 'steps': 26871, 'loss/train': 2.062405824661255} 02/25/2022 12:23:49 - INFO - codeparrot_training - Step 26872: {'lr': 0.00023573967786502694, 'samples': 13758976, 'steps': 26872, 'loss/train': 0.9157112836837769} 02/25/2022 12:23:55 - INFO - codeparrot_training - Step 26873: {'lr': 0.0002357233420748739, 'samples': 13759488, 'steps': 26873, 'loss/train': 2.401214838027954} 02/25/2022 12:24:01 - INFO - codeparrot_training - Step 26874: {'lr': 0.00023570700634587753, 'samples': 13760000, 'steps': 26874, 'loss/train': 1.9598432779312134} 02/25/2022 12:24:04 - INFO - codeparrot_training - Step 26875: {'lr': 0.0002356906706781079, 'samples': 13760512, 'steps': 26875, 'loss/train': 1.7263938188552856} 02/25/2022 12:24:10 - INFO - codeparrot_training - Step 26876: {'lr': 0.00023567433507163478, 'samples': 13761024, 'steps': 26876, 'loss/train': 1.5991458892822266} 02/25/2022 12:24:13 - INFO - codeparrot_training - Step 26877: {'lr': 0.0002356579995265283, 'samples': 13761536, 'steps': 26877, 'loss/train': 2.3382513523101807} 02/25/2022 12:24:19 - INFO - codeparrot_training - Step 26878: {'lr': 0.00023564166404285833, 'samples': 13762048, 'steps': 26878, 'loss/train': 1.8738468885421753} 02/25/2022 12:24:22 - INFO - codeparrot_training - Step 26879: {'lr': 0.00023562532862069505, 'samples': 13762560, 'steps': 26879, 'loss/train': 0.6636435985565186} 02/25/2022 12:24:29 - INFO - codeparrot_training - Step 26880: {'lr': 0.0002356089932601082, 'samples': 13763072, 'steps': 26880, 'loss/train': 1.8751325607299805} 02/25/2022 12:24:32 - INFO - codeparrot_training - Step 26881: {'lr': 0.00023559265796116785, 'samples': 13763584, 'steps': 26881, 'loss/train': 1.8324499130249023} 02/25/2022 12:24:36 - INFO - codeparrot_training - Step 26882: {'lr': 0.00023557632272394396, 'samples': 13764096, 'steps': 26882, 'loss/train': 1.8953220844268799} 02/25/2022 12:24:41 - INFO - codeparrot_training - Step 26883: {'lr': 0.00023555998754850666, 'samples': 13764608, 'steps': 26883, 'loss/train': 1.5912364721298218} 02/25/2022 12:24:47 - INFO - codeparrot_training - Step 26884: {'lr': 0.00023554365243492567, 'samples': 13765120, 'steps': 26884, 'loss/train': 1.2592664957046509} 02/25/2022 12:24:50 - INFO - codeparrot_training - Step 26885: {'lr': 0.0002355273173832711, 'samples': 13765632, 'steps': 26885, 'loss/train': 2.073204517364502} 02/25/2022 12:24:56 - INFO - codeparrot_training - Step 26886: {'lr': 0.00023551098239361288, 'samples': 13766144, 'steps': 26886, 'loss/train': 1.4266589879989624} 02/25/2022 12:24:59 - INFO - codeparrot_training - Step 26887: {'lr': 0.00023549464746602106, 'samples': 13766656, 'steps': 26887, 'loss/train': 1.1164153814315796} 02/25/2022 12:25:04 - INFO - codeparrot_training - Step 26888: {'lr': 0.00023547831260056556, 'samples': 13767168, 'steps': 26888, 'loss/train': 1.8492226600646973} 02/25/2022 12:25:08 - INFO - codeparrot_training - Step 26889: {'lr': 0.00023546197779731636, 'samples': 13767680, 'steps': 26889, 'loss/train': 2.067951202392578} 02/25/2022 12:25:16 - INFO - codeparrot_training - Step 26890: {'lr': 0.00023544564305634342, 'samples': 13768192, 'steps': 26890, 'loss/train': 2.8574769496917725} 02/25/2022 12:25:19 - INFO - codeparrot_training - Step 26891: {'lr': 0.0002354293083777167, 'samples': 13768704, 'steps': 26891, 'loss/train': 1.9289116859436035} 02/25/2022 12:25:25 - INFO - codeparrot_training - Step 26892: {'lr': 0.00023541297376150625, 'samples': 13769216, 'steps': 26892, 'loss/train': 2.142730951309204} 02/25/2022 12:25:28 - INFO - codeparrot_training - Step 26893: {'lr': 0.00023539663920778191, 'samples': 13769728, 'steps': 26893, 'loss/train': 2.585052013397217} 02/25/2022 12:25:34 - INFO - codeparrot_training - Step 26894: {'lr': 0.00023538030471661383, 'samples': 13770240, 'steps': 26894, 'loss/train': 3.1235265731811523} 02/25/2022 12:25:37 - INFO - codeparrot_training - Step 26895: {'lr': 0.0002353639702880718, 'samples': 13770752, 'steps': 26895, 'loss/train': 1.5745577812194824} 02/25/2022 12:25:43 - INFO - codeparrot_training - Step 26896: {'lr': 0.00023534763592222586, 'samples': 13771264, 'steps': 26896, 'loss/train': 2.261033535003662} 02/25/2022 12:25:46 - INFO - codeparrot_training - Step 26897: {'lr': 0.0002353313016191461, 'samples': 13771776, 'steps': 26897, 'loss/train': 1.7269883155822754} 02/25/2022 12:25:52 - INFO - codeparrot_training - Step 26898: {'lr': 0.00023531496737890233, 'samples': 13772288, 'steps': 26898, 'loss/train': 2.141082763671875} 02/25/2022 12:25:55 - INFO - codeparrot_training - Step 26899: {'lr': 0.00023529863320156453, 'samples': 13772800, 'steps': 26899, 'loss/train': 1.8374658823013306} 02/25/2022 12:26:02 - INFO - codeparrot_training - Step 26900: {'lr': 0.00023528229908720272, 'samples': 13773312, 'steps': 26900, 'loss/train': 3.225510835647583} 02/25/2022 12:26:06 - INFO - codeparrot_training - Step 26901: {'lr': 0.00023526596503588702, 'samples': 13773824, 'steps': 26901, 'loss/train': 1.1997783184051514} 02/25/2022 12:26:11 - INFO - codeparrot_training - Step 26902: {'lr': 0.0002352496310476871, 'samples': 13774336, 'steps': 26902, 'loss/train': 1.8035966157913208} 02/25/2022 12:26:15 - INFO - codeparrot_training - Step 26903: {'lr': 0.00023523329712267307, 'samples': 13774848, 'steps': 26903, 'loss/train': 1.8447291851043701} 02/25/2022 12:26:20 - INFO - codeparrot_training - Step 26904: {'lr': 0.0002352169632609149, 'samples': 13775360, 'steps': 26904, 'loss/train': 2.5429704189300537} 02/25/2022 12:26:24 - INFO - codeparrot_training - Step 26905: {'lr': 0.0002352006294624827, 'samples': 13775872, 'steps': 26905, 'loss/train': 2.946312189102173} 02/25/2022 12:26:29 - INFO - codeparrot_training - Step 26906: {'lr': 0.0002351842957274462, 'samples': 13776384, 'steps': 26906, 'loss/train': 2.117699384689331} 02/25/2022 12:26:33 - INFO - codeparrot_training - Step 26907: {'lr': 0.00023516796205587547, 'samples': 13776896, 'steps': 26907, 'loss/train': 2.334052324295044} 02/25/2022 12:26:38 - INFO - codeparrot_training - Step 26908: {'lr': 0.00023515162844784046, 'samples': 13777408, 'steps': 26908, 'loss/train': 1.7366881370544434} 02/25/2022 12:26:42 - INFO - codeparrot_training - Step 26909: {'lr': 0.00023513529490341116, 'samples': 13777920, 'steps': 26909, 'loss/train': 1.0266656875610352} 02/25/2022 12:26:47 - INFO - codeparrot_training - Step 26910: {'lr': 0.00023511896142265764, 'samples': 13778432, 'steps': 26910, 'loss/train': 2.2760567665100098} 02/25/2022 12:26:51 - INFO - codeparrot_training - Step 26911: {'lr': 0.00023510262800564968, 'samples': 13778944, 'steps': 26911, 'loss/train': 1.6876857280731201} 02/25/2022 12:26:58 - INFO - codeparrot_training - Step 26912: {'lr': 0.00023508629465245735, 'samples': 13779456, 'steps': 26912, 'loss/train': 1.7996442317962646} 02/25/2022 12:27:01 - INFO - codeparrot_training - Step 26913: {'lr': 0.00023506996136315057, 'samples': 13779968, 'steps': 26913, 'loss/train': 1.8048832416534424} 02/25/2022 12:27:07 - INFO - codeparrot_training - Step 26914: {'lr': 0.00023505362813779944, 'samples': 13780480, 'steps': 26914, 'loss/train': 0.961259663105011} 02/25/2022 12:27:11 - INFO - codeparrot_training - Step 26915: {'lr': 0.0002350372949764737, 'samples': 13780992, 'steps': 26915, 'loss/train': 2.397836208343506} 02/25/2022 12:27:16 - INFO - codeparrot_training - Step 26916: {'lr': 0.00023502096187924347, 'samples': 13781504, 'steps': 26916, 'loss/train': 1.7659590244293213} 02/25/2022 12:27:20 - INFO - codeparrot_training - Step 26917: {'lr': 0.00023500462884617866, 'samples': 13782016, 'steps': 26917, 'loss/train': 2.1485817432403564} 02/25/2022 12:27:25 - INFO - codeparrot_training - Step 26918: {'lr': 0.00023498829587734936, 'samples': 13782528, 'steps': 26918, 'loss/train': 2.0062687397003174} 02/25/2022 12:27:29 - INFO - codeparrot_training - Step 26919: {'lr': 0.00023497196297282534, 'samples': 13783040, 'steps': 26919, 'loss/train': 1.2176316976547241} 02/25/2022 12:27:34 - INFO - codeparrot_training - Step 26920: {'lr': 0.00023495563013267666, 'samples': 13783552, 'steps': 26920, 'loss/train': 0.33599233627319336} 02/25/2022 12:27:38 - INFO - codeparrot_training - Step 26921: {'lr': 0.00023493929735697328, 'samples': 13784064, 'steps': 26921, 'loss/train': 1.470510721206665} 02/25/2022 12:27:45 - INFO - codeparrot_training - Step 26922: {'lr': 0.00023492296464578516, 'samples': 13784576, 'steps': 26922, 'loss/train': 2.0926873683929443} 02/25/2022 12:27:48 - INFO - codeparrot_training - Step 26923: {'lr': 0.00023490663199918238, 'samples': 13785088, 'steps': 26923, 'loss/train': 1.7308963537216187} 02/25/2022 12:27:54 - INFO - codeparrot_training - Step 26924: {'lr': 0.00023489029941723468, 'samples': 13785600, 'steps': 26924, 'loss/train': 2.3819479942321777} 02/25/2022 12:27:57 - INFO - codeparrot_training - Step 26925: {'lr': 0.00023487396690001218, 'samples': 13786112, 'steps': 26925, 'loss/train': 1.6707313060760498} 02/25/2022 12:28:03 - INFO - codeparrot_training - Step 26926: {'lr': 0.00023485763444758476, 'samples': 13786624, 'steps': 26926, 'loss/train': 1.0905370712280273} 02/25/2022 12:28:06 - INFO - codeparrot_training - Step 26927: {'lr': 0.00023484130206002257, 'samples': 13787136, 'steps': 26927, 'loss/train': 2.322423219680786} 02/25/2022 12:28:12 - INFO - codeparrot_training - Step 26928: {'lr': 0.00023482496973739527, 'samples': 13787648, 'steps': 26928, 'loss/train': 1.0728545188903809} 02/25/2022 12:28:15 - INFO - codeparrot_training - Step 26929: {'lr': 0.00023480863747977305, 'samples': 13788160, 'steps': 26929, 'loss/train': 1.530009150505066} 02/25/2022 12:28:21 - INFO - codeparrot_training - Step 26930: {'lr': 0.0002347923052872258, 'samples': 13788672, 'steps': 26930, 'loss/train': 1.2424969673156738} 02/25/2022 12:28:24 - INFO - codeparrot_training - Step 26931: {'lr': 0.00023477597315982353, 'samples': 13789184, 'steps': 26931, 'loss/train': 1.1929811239242554} 02/25/2022 12:28:32 - INFO - codeparrot_training - Step 26932: {'lr': 0.00023475964109763607, 'samples': 13789696, 'steps': 26932, 'loss/train': 2.6617064476013184} 02/25/2022 12:28:35 - INFO - codeparrot_training - Step 26933: {'lr': 0.00023474330910073347, 'samples': 13790208, 'steps': 26933, 'loss/train': 2.2720558643341064} 02/25/2022 12:28:41 - INFO - codeparrot_training - Step 26934: {'lr': 0.00023472697716918575, 'samples': 13790720, 'steps': 26934, 'loss/train': 2.345998764038086} 02/25/2022 12:28:44 - INFO - codeparrot_training - Step 26935: {'lr': 0.00023471064530306273, 'samples': 13791232, 'steps': 26935, 'loss/train': 1.6002155542373657} 02/25/2022 12:28:50 - INFO - codeparrot_training - Step 26936: {'lr': 0.00023469431350243457, 'samples': 13791744, 'steps': 26936, 'loss/train': 1.923892617225647} 02/25/2022 12:28:53 - INFO - codeparrot_training - Step 26937: {'lr': 0.00023467798176737104, 'samples': 13792256, 'steps': 26937, 'loss/train': 2.699946880340576} 02/25/2022 12:28:59 - INFO - codeparrot_training - Step 26938: {'lr': 0.00023466165009794216, 'samples': 13792768, 'steps': 26938, 'loss/train': 2.1960859298706055} 02/25/2022 12:29:02 - INFO - codeparrot_training - Step 26939: {'lr': 0.00023464531849421788, 'samples': 13793280, 'steps': 26939, 'loss/train': 2.641878604888916} 02/25/2022 12:29:08 - INFO - codeparrot_training - Step 26940: {'lr': 0.00023462898695626825, 'samples': 13793792, 'steps': 26940, 'loss/train': 2.054353713989258} 02/25/2022 12:29:11 - INFO - codeparrot_training - Step 26941: {'lr': 0.00023461265548416316, 'samples': 13794304, 'steps': 26941, 'loss/train': 2.514080286026001} 02/25/2022 12:29:17 - INFO - codeparrot_training - Step 26942: {'lr': 0.0002345963240779725, 'samples': 13794816, 'steps': 26942, 'loss/train': 0.7865099906921387} 02/25/2022 12:29:20 - INFO - codeparrot_training - Step 26943: {'lr': 0.0002345799927377663, 'samples': 13795328, 'steps': 26943, 'loss/train': 1.6671823263168335} 02/25/2022 12:29:26 - INFO - codeparrot_training - Step 26944: {'lr': 0.0002345636614636145, 'samples': 13795840, 'steps': 26944, 'loss/train': 0.3675834536552429} 02/25/2022 12:29:29 - INFO - codeparrot_training - Step 26945: {'lr': 0.00023454733025558722, 'samples': 13796352, 'steps': 26945, 'loss/train': 2.8286986351013184} 02/25/2022 12:29:37 - INFO - codeparrot_training - Step 26946: {'lr': 0.00023453099911375415, 'samples': 13796864, 'steps': 26946, 'loss/train': 1.433905005455017} 02/25/2022 12:29:40 - INFO - codeparrot_training - Step 26947: {'lr': 0.00023451466803818533, 'samples': 13797376, 'steps': 26947, 'loss/train': 1.6915913820266724} 02/25/2022 12:29:46 - INFO - codeparrot_training - Step 26948: {'lr': 0.00023449833702895079, 'samples': 13797888, 'steps': 26948, 'loss/train': 1.1234729290008545} 02/25/2022 12:29:49 - INFO - codeparrot_training - Step 26949: {'lr': 0.00023448200608612056, 'samples': 13798400, 'steps': 26949, 'loss/train': 1.8688406944274902} 02/25/2022 12:29:55 - INFO - codeparrot_training - Step 26950: {'lr': 0.00023446567520976437, 'samples': 13798912, 'steps': 26950, 'loss/train': 1.9066269397735596} 02/25/2022 12:29:58 - INFO - codeparrot_training - Step 26951: {'lr': 0.00023444934439995231, 'samples': 13799424, 'steps': 26951, 'loss/train': 2.5175890922546387} 02/25/2022 12:30:04 - INFO - codeparrot_training - Step 26952: {'lr': 0.00023443301365675429, 'samples': 13799936, 'steps': 26952, 'loss/train': 1.9502217769622803} 02/25/2022 12:30:07 - INFO - codeparrot_training - Step 26953: {'lr': 0.0002344166829802404, 'samples': 13800448, 'steps': 26953, 'loss/train': 2.8203048706054688} 02/25/2022 12:30:13 - INFO - codeparrot_training - Step 26954: {'lr': 0.00023440035237048043, 'samples': 13800960, 'steps': 26954, 'loss/train': 2.954355001449585} 02/25/2022 12:30:16 - INFO - codeparrot_training - Step 26955: {'lr': 0.00023438402182754438, 'samples': 13801472, 'steps': 26955, 'loss/train': 1.731742262840271} 02/25/2022 12:30:23 - INFO - codeparrot_training - Step 26956: {'lr': 0.0002343676913515022, 'samples': 13801984, 'steps': 26956, 'loss/train': 1.950888752937317} 02/25/2022 12:30:27 - INFO - codeparrot_training - Step 26957: {'lr': 0.00023435136094242386, 'samples': 13802496, 'steps': 26957, 'loss/train': 2.819667339324951} 02/25/2022 12:30:32 - INFO - codeparrot_training - Step 26958: {'lr': 0.00023433503060037947, 'samples': 13803008, 'steps': 26958, 'loss/train': 1.6244242191314697} 02/25/2022 12:30:36 - INFO - codeparrot_training - Step 26959: {'lr': 0.0002343187003254387, 'samples': 13803520, 'steps': 26959, 'loss/train': 1.6507328748703003} 02/25/2022 12:30:41 - INFO - codeparrot_training - Step 26960: {'lr': 0.00023430237011767165, 'samples': 13804032, 'steps': 26960, 'loss/train': 2.173668384552002} 02/25/2022 12:30:45 - INFO - codeparrot_training - Step 26961: {'lr': 0.00023428603997714825, 'samples': 13804544, 'steps': 26961, 'loss/train': 2.7793047428131104} 02/25/2022 12:30:50 - INFO - codeparrot_training - Step 26962: {'lr': 0.0002342697099039386, 'samples': 13805056, 'steps': 26962, 'loss/train': 1.5307743549346924} 02/25/2022 12:30:54 - INFO - codeparrot_training - Step 26963: {'lr': 0.00023425337989811237, 'samples': 13805568, 'steps': 26963, 'loss/train': 2.0633704662323} 02/25/2022 12:30:59 - INFO - codeparrot_training - Step 26964: {'lr': 0.00023423704995973967, 'samples': 13806080, 'steps': 26964, 'loss/train': 1.6551611423492432} 02/25/2022 12:31:03 - INFO - codeparrot_training - Step 26965: {'lr': 0.00023422072008889047, 'samples': 13806592, 'steps': 26965, 'loss/train': 2.9720993041992188} 02/25/2022 12:31:06 - INFO - codeparrot_training - Step 26966: {'lr': 0.0002342043902856348, 'samples': 13807104, 'steps': 26966, 'loss/train': 2.187894105911255} 02/25/2022 12:31:12 - INFO - codeparrot_training - Step 26967: {'lr': 0.00023418806055004238, 'samples': 13807616, 'steps': 26967, 'loss/train': 1.8310469388961792} 02/25/2022 12:31:19 - INFO - codeparrot_training - Step 26968: {'lr': 0.00023417173088218328, 'samples': 13808128, 'steps': 26968, 'loss/train': 1.8833154439926147} 02/25/2022 12:31:23 - INFO - codeparrot_training - Step 26969: {'lr': 0.00023415540128212748, 'samples': 13808640, 'steps': 26969, 'loss/train': 2.8030288219451904} 02/25/2022 12:31:28 - INFO - codeparrot_training - Step 26970: {'lr': 0.00023413907174994495, 'samples': 13809152, 'steps': 26970, 'loss/train': 1.8186744451522827} 02/25/2022 12:31:32 - INFO - codeparrot_training - Step 26971: {'lr': 0.0002341227422857056, 'samples': 13809664, 'steps': 26971, 'loss/train': 2.4994451999664307} 02/25/2022 12:31:35 - INFO - codeparrot_training - Step 26972: {'lr': 0.00023410641288947935, 'samples': 13810176, 'steps': 26972, 'loss/train': 1.3450109958648682} 02/25/2022 12:31:41 - INFO - codeparrot_training - Step 26973: {'lr': 0.00023409008356133616, 'samples': 13810688, 'steps': 26973, 'loss/train': 1.844714641571045} 02/25/2022 12:31:46 - INFO - codeparrot_training - Step 26974: {'lr': 0.00023407375430134603, 'samples': 13811200, 'steps': 26974, 'loss/train': 1.5350931882858276} 02/25/2022 12:31:50 - INFO - codeparrot_training - Step 26975: {'lr': 0.0002340574251095789, 'samples': 13811712, 'steps': 26975, 'loss/train': 2.808093309402466} 02/25/2022 12:31:56 - INFO - codeparrot_training - Step 26976: {'lr': 0.0002340410959861047, 'samples': 13812224, 'steps': 26976, 'loss/train': 2.0313754081726074} 02/25/2022 12:31:59 - INFO - codeparrot_training - Step 26977: {'lr': 0.0002340247669309933, 'samples': 13812736, 'steps': 26977, 'loss/train': 2.2129156589508057} 02/25/2022 12:32:02 - INFO - codeparrot_training - Step 26978: {'lr': 0.0002340084379443148, 'samples': 13813248, 'steps': 26978, 'loss/train': 1.1918896436691284} 02/25/2022 12:32:08 - INFO - codeparrot_training - Step 26979: {'lr': 0.00023399210902613906, 'samples': 13813760, 'steps': 26979, 'loss/train': 1.6910289525985718} 02/25/2022 12:32:11 - INFO - codeparrot_training - Step 26980: {'lr': 0.00023397578017653604, 'samples': 13814272, 'steps': 26980, 'loss/train': 0.5626059174537659} 02/25/2022 12:32:17 - INFO - codeparrot_training - Step 26981: {'lr': 0.0002339594513955757, 'samples': 13814784, 'steps': 26981, 'loss/train': 1.6927043199539185} 02/25/2022 12:32:20 - INFO - codeparrot_training - Step 26982: {'lr': 0.00023394312268332793, 'samples': 13815296, 'steps': 26982, 'loss/train': 0.8931909203529358} 02/25/2022 12:32:28 - INFO - codeparrot_training - Step 26983: {'lr': 0.00023392679403986272, 'samples': 13815808, 'steps': 26983, 'loss/train': 1.8689600229263306} 02/25/2022 12:32:31 - INFO - codeparrot_training - Step 26984: {'lr': 0.0002339104654652501, 'samples': 13816320, 'steps': 26984, 'loss/train': 1.7950832843780518} 02/25/2022 12:32:37 - INFO - codeparrot_training - Step 26985: {'lr': 0.0002338941369595599, 'samples': 13816832, 'steps': 26985, 'loss/train': 1.7799584865570068} 02/25/2022 12:32:40 - INFO - codeparrot_training - Step 26986: {'lr': 0.0002338778085228621, 'samples': 13817344, 'steps': 26986, 'loss/train': 2.5141048431396484} 02/25/2022 12:32:46 - INFO - codeparrot_training - Step 26987: {'lr': 0.0002338614801552266, 'samples': 13817856, 'steps': 26987, 'loss/train': 1.3375355005264282} 02/25/2022 12:32:49 - INFO - codeparrot_training - Step 26988: {'lr': 0.00023384515185672353, 'samples': 13818368, 'steps': 26988, 'loss/train': 2.8453927040100098} 02/25/2022 12:32:55 - INFO - codeparrot_training - Step 26989: {'lr': 0.00023382882362742258, 'samples': 13818880, 'steps': 26989, 'loss/train': 2.148735284805298} 02/25/2022 12:32:58 - INFO - codeparrot_training - Step 26990: {'lr': 0.0002338124954673938, 'samples': 13819392, 'steps': 26990, 'loss/train': 1.4600436687469482} 02/25/2022 12:33:04 - INFO - codeparrot_training - Step 26991: {'lr': 0.00023379616737670718, 'samples': 13819904, 'steps': 26991, 'loss/train': 1.4878910779953003} 02/25/2022 12:33:07 - INFO - codeparrot_training - Step 26992: {'lr': 0.0002337798393554326, 'samples': 13820416, 'steps': 26992, 'loss/train': 2.499819040298462} 02/25/2022 12:33:13 - INFO - codeparrot_training - Step 26993: {'lr': 0.0002337635114036402, 'samples': 13820928, 'steps': 26993, 'loss/train': 2.2658467292785645} 02/25/2022 12:33:16 - INFO - codeparrot_training - Step 26994: {'lr': 0.0002337471835213996, 'samples': 13821440, 'steps': 26994, 'loss/train': 1.5784450769424438} 02/25/2022 12:33:23 - INFO - codeparrot_training - Step 26995: {'lr': 0.0002337308557087809, 'samples': 13821952, 'steps': 26995, 'loss/train': 1.4939954280853271} 02/25/2022 12:33:27 - INFO - codeparrot_training - Step 26996: {'lr': 0.00023371452796585408, 'samples': 13822464, 'steps': 26996, 'loss/train': 2.2780017852783203} 02/25/2022 12:33:32 - INFO - codeparrot_training - Step 26997: {'lr': 0.00023369820029268916, 'samples': 13822976, 'steps': 26997, 'loss/train': 1.8589674234390259} 02/25/2022 12:33:36 - INFO - codeparrot_training - Step 26998: {'lr': 0.00023368187268935588, 'samples': 13823488, 'steps': 26998, 'loss/train': 2.1118319034576416} 02/25/2022 12:33:41 - INFO - codeparrot_training - Step 26999: {'lr': 0.00023366554515592423, 'samples': 13824000, 'steps': 26999, 'loss/train': 1.4386999607086182} 02/25/2022 12:33:42 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 12:33:59 - WARNING - huggingface_hub.repository - Several commits (27) will be pushed upstream. 02/25/2022 12:33:59 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 12:34:41 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 86c4c54..9d342aa floral-grass-11 -> floral-grass-11 02/25/2022 12:34:45 - INFO - codeparrot_training - Step 27000: {'lr': 0.00023364921769246423, 'samples': 13824512, 'steps': 27000, 'loss/train': 1.4343920946121216} 02/25/2022 12:34:51 - INFO - codeparrot_training - Step 27001: {'lr': 0.0002336328902990459, 'samples': 13825024, 'steps': 27001, 'loss/train': 0.8831506371498108} 02/25/2022 12:34:54 - INFO - codeparrot_training - Step 27002: {'lr': 0.0002336165629757389, 'samples': 13825536, 'steps': 27002, 'loss/train': 1.8967748880386353} 02/25/2022 12:35:00 - INFO - codeparrot_training - Step 27003: {'lr': 0.0002336002357226134, 'samples': 13826048, 'steps': 27003, 'loss/train': 1.9600099325180054} 02/25/2022 12:35:03 - INFO - codeparrot_training - Step 27004: {'lr': 0.00023358390853973928, 'samples': 13826560, 'steps': 27004, 'loss/train': 1.4554760456085205} 02/25/2022 12:35:11 - INFO - codeparrot_training - Step 27005: {'lr': 0.0002335675814271865, 'samples': 13827072, 'steps': 27005, 'loss/train': 0.9962917566299438} 02/25/2022 12:35:14 - INFO - codeparrot_training - Step 27006: {'lr': 0.00023355125438502506, 'samples': 13827584, 'steps': 27006, 'loss/train': 2.461592197418213} 02/25/2022 12:35:20 - INFO - codeparrot_training - Step 27007: {'lr': 0.00023353492741332472, 'samples': 13828096, 'steps': 27007, 'loss/train': 1.1618081331253052} 02/25/2022 12:35:23 - INFO - codeparrot_training - Step 27008: {'lr': 0.00023351860051215554, 'samples': 13828608, 'steps': 27008, 'loss/train': 2.4684417247772217} 02/25/2022 12:35:29 - INFO - codeparrot_training - Step 27009: {'lr': 0.00023350227368158742, 'samples': 13829120, 'steps': 27009, 'loss/train': 2.1745755672454834} 02/25/2022 12:35:32 - INFO - codeparrot_training - Step 27010: {'lr': 0.00023348594692169044, 'samples': 13829632, 'steps': 27010, 'loss/train': 1.3420857191085815} 02/25/2022 12:35:38 - INFO - codeparrot_training - Step 27011: {'lr': 0.00023346962023253431, 'samples': 13830144, 'steps': 27011, 'loss/train': 1.9390747547149658} 02/25/2022 12:35:41 - INFO - codeparrot_training - Step 27012: {'lr': 0.00023345329361418906, 'samples': 13830656, 'steps': 27012, 'loss/train': 2.150869131088257} 02/25/2022 12:35:47 - INFO - codeparrot_training - Step 27013: {'lr': 0.0002334369670667247, 'samples': 13831168, 'steps': 27013, 'loss/train': 1.68816339969635} 02/25/2022 12:35:50 - INFO - codeparrot_training - Step 27014: {'lr': 0.00023342064059021116, 'samples': 13831680, 'steps': 27014, 'loss/train': 1.979455590248108} 02/25/2022 12:35:58 - INFO - codeparrot_training - Step 27015: {'lr': 0.00023340431418471824, 'samples': 13832192, 'steps': 27015, 'loss/train': 0.5500654578208923} 02/25/2022 12:36:01 - INFO - codeparrot_training - Step 27016: {'lr': 0.00023338798785031597, 'samples': 13832704, 'steps': 27016, 'loss/train': 1.2035255432128906} 02/25/2022 12:36:07 - INFO - codeparrot_training - Step 27017: {'lr': 0.00023337166158707428, 'samples': 13833216, 'steps': 27017, 'loss/train': 1.976059079170227} 02/25/2022 12:36:10 - INFO - codeparrot_training - Step 27018: {'lr': 0.0002333553353950632, 'samples': 13833728, 'steps': 27018, 'loss/train': 2.331285238265991} 02/25/2022 12:36:16 - INFO - codeparrot_training - Step 27019: {'lr': 0.00023333900927435255, 'samples': 13834240, 'steps': 27019, 'loss/train': 1.1922298669815063} 02/25/2022 12:36:19 - INFO - codeparrot_training - Step 27020: {'lr': 0.00023332268322501226, 'samples': 13834752, 'steps': 27020, 'loss/train': 1.0364292860031128} 02/25/2022 12:36:25 - INFO - codeparrot_training - Step 27021: {'lr': 0.00023330635724711234, 'samples': 13835264, 'steps': 27021, 'loss/train': 2.257805585861206} 02/25/2022 12:36:28 - INFO - codeparrot_training - Step 27022: {'lr': 0.00023329003134072262, 'samples': 13835776, 'steps': 27022, 'loss/train': 2.1434290409088135} 02/25/2022 12:36:34 - INFO - codeparrot_training - Step 27023: {'lr': 0.00023327370550591314, 'samples': 13836288, 'steps': 27023, 'loss/train': 1.5563085079193115} 02/25/2022 12:36:37 - INFO - codeparrot_training - Step 27024: {'lr': 0.0002332573797427538, 'samples': 13836800, 'steps': 27024, 'loss/train': 1.5208728313446045} 02/25/2022 12:36:43 - INFO - codeparrot_training - Step 27025: {'lr': 0.00023324105405131453, 'samples': 13837312, 'steps': 27025, 'loss/train': 2.117149591445923} 02/25/2022 12:36:46 - INFO - codeparrot_training - Step 27026: {'lr': 0.00023322472843166522, 'samples': 13837824, 'steps': 27026, 'loss/train': 1.9643398523330688} 02/25/2022 12:36:52 - INFO - codeparrot_training - Step 27027: {'lr': 0.0002332084028838759, 'samples': 13838336, 'steps': 27027, 'loss/train': 1.7869101762771606} 02/25/2022 12:36:55 - INFO - codeparrot_training - Step 27028: {'lr': 0.00023319207740801645, 'samples': 13838848, 'steps': 27028, 'loss/train': 2.148238182067871} 02/25/2022 12:37:01 - INFO - codeparrot_training - Step 27029: {'lr': 0.00023317575200415677, 'samples': 13839360, 'steps': 27029, 'loss/train': 1.5981947183609009} 02/25/2022 12:37:04 - INFO - codeparrot_training - Step 27030: {'lr': 0.0002331594266723668, 'samples': 13839872, 'steps': 27030, 'loss/train': 1.2656347751617432} 02/25/2022 12:37:12 - INFO - codeparrot_training - Step 27031: {'lr': 0.00023314310141271652, 'samples': 13840384, 'steps': 27031, 'loss/train': 0.17614024877548218} 02/25/2022 12:37:15 - INFO - codeparrot_training - Step 27032: {'lr': 0.00023312677622527595, 'samples': 13840896, 'steps': 27032, 'loss/train': 2.560861349105835} 02/25/2022 12:37:21 - INFO - codeparrot_training - Step 27033: {'lr': 0.00023311045111011478, 'samples': 13841408, 'steps': 27033, 'loss/train': 2.1723649501800537} 02/25/2022 12:37:24 - INFO - codeparrot_training - Step 27034: {'lr': 0.0002330941260673031, 'samples': 13841920, 'steps': 27034, 'loss/train': 0.8607158660888672} 02/25/2022 12:37:30 - INFO - codeparrot_training - Step 27035: {'lr': 0.0002330778010969108, 'samples': 13842432, 'steps': 27035, 'loss/train': 0.4541480839252472} 02/25/2022 12:37:33 - INFO - codeparrot_training - Step 27036: {'lr': 0.00023306147619900794, 'samples': 13842944, 'steps': 27036, 'loss/train': 1.3741477727890015} 02/25/2022 12:37:39 - INFO - codeparrot_training - Step 27037: {'lr': 0.00023304515137366423, 'samples': 13843456, 'steps': 27037, 'loss/train': 1.7156343460083008} 02/25/2022 12:37:42 - INFO - codeparrot_training - Step 27038: {'lr': 0.0002330288266209497, 'samples': 13843968, 'steps': 27038, 'loss/train': 1.9308557510375977} 02/25/2022 12:37:48 - INFO - codeparrot_training - Step 27039: {'lr': 0.0002330125019409343, 'samples': 13844480, 'steps': 27039, 'loss/train': 2.123274564743042} 02/25/2022 12:37:51 - INFO - codeparrot_training - Step 27040: {'lr': 0.00023299617733368806, 'samples': 13844992, 'steps': 27040, 'loss/train': 1.9808441400527954} 02/25/2022 12:37:58 - INFO - codeparrot_training - Step 27041: {'lr': 0.00023297985279928065, 'samples': 13845504, 'steps': 27041, 'loss/train': 2.7426066398620605} 02/25/2022 12:38:02 - INFO - codeparrot_training - Step 27042: {'lr': 0.00023296352833778217, 'samples': 13846016, 'steps': 27042, 'loss/train': 2.7557485103607178} 02/25/2022 12:38:07 - INFO - codeparrot_training - Step 27043: {'lr': 0.00023294720394926254, 'samples': 13846528, 'steps': 27043, 'loss/train': 2.155454635620117} 02/25/2022 12:38:11 - INFO - codeparrot_training - Step 27044: {'lr': 0.00023293087963379168, 'samples': 13847040, 'steps': 27044, 'loss/train': 2.1809327602386475} 02/25/2022 12:38:16 - INFO - codeparrot_training - Step 27045: {'lr': 0.00023291455539143957, 'samples': 13847552, 'steps': 27045, 'loss/train': 1.895490288734436} 02/25/2022 12:38:20 - INFO - codeparrot_training - Step 27046: {'lr': 0.00023289823122227602, 'samples': 13848064, 'steps': 27046, 'loss/train': 1.6413599252700806} 02/25/2022 12:38:26 - INFO - codeparrot_training - Step 27047: {'lr': 0.000232881907126371, 'samples': 13848576, 'steps': 27047, 'loss/train': 3.226123809814453} 02/25/2022 12:38:29 - INFO - codeparrot_training - Step 27048: {'lr': 0.00023286558310379443, 'samples': 13849088, 'steps': 27048, 'loss/train': 2.4897942543029785} 02/25/2022 12:38:35 - INFO - codeparrot_training - Step 27049: {'lr': 0.00023284925915461642, 'samples': 13849600, 'steps': 27049, 'loss/train': 1.9303382635116577} 02/25/2022 12:38:38 - INFO - codeparrot_training - Step 27050: {'lr': 0.00023283293527890658, 'samples': 13850112, 'steps': 27050, 'loss/train': 2.461439847946167} 02/25/2022 12:38:45 - INFO - codeparrot_training - Step 27051: {'lr': 0.00023281661147673502, 'samples': 13850624, 'steps': 27051, 'loss/train': 1.6616508960723877} 02/25/2022 12:38:49 - INFO - codeparrot_training - Step 27052: {'lr': 0.00023280028774817164, 'samples': 13851136, 'steps': 27052, 'loss/train': 2.972146987915039} 02/25/2022 12:38:54 - INFO - codeparrot_training - Step 27053: {'lr': 0.00023278396409328638, 'samples': 13851648, 'steps': 27053, 'loss/train': 2.261521100997925} 02/25/2022 12:38:58 - INFO - codeparrot_training - Step 27054: {'lr': 0.00023276764051214925, 'samples': 13852160, 'steps': 27054, 'loss/train': 2.037123918533325} 02/25/2022 12:39:03 - INFO - codeparrot_training - Step 27055: {'lr': 0.00023275131700482992, 'samples': 13852672, 'steps': 27055, 'loss/train': 1.8184711933135986} 02/25/2022 12:39:09 - INFO - codeparrot_training - Step 27056: {'lr': 0.00023273499357139853, 'samples': 13853184, 'steps': 27056, 'loss/train': 0.9392097592353821} 02/25/2022 12:39:12 - INFO - codeparrot_training - Step 27057: {'lr': 0.00023271867021192494, 'samples': 13853696, 'steps': 27057, 'loss/train': 2.3508214950561523} 02/25/2022 12:39:18 - INFO - codeparrot_training - Step 27058: {'lr': 0.00023270234692647919, 'samples': 13854208, 'steps': 27058, 'loss/train': 2.173530101776123} 02/25/2022 12:39:21 - INFO - codeparrot_training - Step 27059: {'lr': 0.00023268602371513095, 'samples': 13854720, 'steps': 27059, 'loss/train': 1.7950519323349} 02/25/2022 12:39:29 - INFO - codeparrot_training - Step 27060: {'lr': 0.0002326697005779503, 'samples': 13855232, 'steps': 27060, 'loss/train': 1.7231035232543945} 02/25/2022 12:39:32 - INFO - codeparrot_training - Step 27061: {'lr': 0.0002326533775150072, 'samples': 13855744, 'steps': 27061, 'loss/train': 1.129690408706665} 02/25/2022 12:39:37 - INFO - codeparrot_training - Step 27062: {'lr': 0.00023263705452637154, 'samples': 13856256, 'steps': 27062, 'loss/train': 1.405643105506897} 02/25/2022 12:39:41 - INFO - codeparrot_training - Step 27063: {'lr': 0.00023262073161211316, 'samples': 13856768, 'steps': 27063, 'loss/train': 2.770814895629883} 02/25/2022 12:39:46 - INFO - codeparrot_training - Step 27064: {'lr': 0.00023260440877230206, 'samples': 13857280, 'steps': 27064, 'loss/train': 1.6392837762832642} 02/25/2022 12:39:50 - INFO - codeparrot_training - Step 27065: {'lr': 0.0002325880860070082, 'samples': 13857792, 'steps': 27065, 'loss/train': 2.566894292831421} 02/25/2022 12:39:55 - INFO - codeparrot_training - Step 27066: {'lr': 0.00023257176331630135, 'samples': 13858304, 'steps': 27066, 'loss/train': 2.1903414726257324} 02/25/2022 12:39:59 - INFO - codeparrot_training - Step 27067: {'lr': 0.00023255544070025164, 'samples': 13858816, 'steps': 27067, 'loss/train': 2.4081480503082275} 02/25/2022 12:40:04 - INFO - codeparrot_training - Step 27068: {'lr': 0.00023253911815892888, 'samples': 13859328, 'steps': 27068, 'loss/train': 2.0360560417175293} 02/25/2022 12:40:08 - INFO - codeparrot_training - Step 27069: {'lr': 0.00023252279569240292, 'samples': 13859840, 'steps': 27069, 'loss/train': 2.400786876678467} 02/25/2022 12:40:13 - INFO - codeparrot_training - Step 27070: {'lr': 0.00023250647330074376, 'samples': 13860352, 'steps': 27070, 'loss/train': 2.2619104385375977} 02/25/2022 12:40:17 - INFO - codeparrot_training - Step 27071: {'lr': 0.00023249015098402137, 'samples': 13860864, 'steps': 27071, 'loss/train': 1.9302499294281006} 02/25/2022 12:40:24 - INFO - codeparrot_training - Step 27072: {'lr': 0.00023247382874230562, 'samples': 13861376, 'steps': 27072, 'loss/train': 2.0327818393707275} 02/25/2022 12:40:28 - INFO - codeparrot_training - Step 27073: {'lr': 0.00023245750657566636, 'samples': 13861888, 'steps': 27073, 'loss/train': 1.0842987298965454} 02/25/2022 12:40:33 - INFO - codeparrot_training - Step 27074: {'lr': 0.0002324411844841736, 'samples': 13862400, 'steps': 27074, 'loss/train': 0.9487189650535583} 02/25/2022 12:40:37 - INFO - codeparrot_training - Step 27075: {'lr': 0.0002324248624678973, 'samples': 13862912, 'steps': 27075, 'loss/train': 0.672633171081543} 02/25/2022 12:40:42 - INFO - codeparrot_training - Step 27076: {'lr': 0.00023240854052690722, 'samples': 13863424, 'steps': 27076, 'loss/train': 1.6048749685287476} 02/25/2022 12:40:46 - INFO - codeparrot_training - Step 27077: {'lr': 0.00023239221866127333, 'samples': 13863936, 'steps': 27077, 'loss/train': 1.9225966930389404} 02/25/2022 12:40:51 - INFO - codeparrot_training - Step 27078: {'lr': 0.0002323758968710656, 'samples': 13864448, 'steps': 27078, 'loss/train': 2.140515089035034} 02/25/2022 12:40:55 - INFO - codeparrot_training - Step 27079: {'lr': 0.00023235957515635398, 'samples': 13864960, 'steps': 27079, 'loss/train': 2.1723084449768066} 02/25/2022 12:41:00 - INFO - codeparrot_training - Step 27080: {'lr': 0.0002323432535172084, 'samples': 13865472, 'steps': 27080, 'loss/train': 2.795393466949463} 02/25/2022 12:41:04 - INFO - codeparrot_training - Step 27081: {'lr': 0.0002323269319536986, 'samples': 13865984, 'steps': 27081, 'loss/train': 2.176541566848755} 02/25/2022 12:41:09 - INFO - codeparrot_training - Step 27082: {'lr': 0.00023231061046589464, 'samples': 13866496, 'steps': 27082, 'loss/train': 2.1633191108703613} 02/25/2022 12:41:13 - INFO - codeparrot_training - Step 27083: {'lr': 0.0002322942890538664, 'samples': 13867008, 'steps': 27083, 'loss/train': 1.303206205368042} 02/25/2022 12:41:18 - INFO - codeparrot_training - Step 27084: {'lr': 0.0002322779677176839, 'samples': 13867520, 'steps': 27084, 'loss/train': 2.6134519577026367} 02/25/2022 12:41:22 - INFO - codeparrot_training - Step 27085: {'lr': 0.00023226164645741689, 'samples': 13868032, 'steps': 27085, 'loss/train': 2.4531805515289307} 02/25/2022 12:41:27 - INFO - codeparrot_training - Step 27086: {'lr': 0.0002322453252731353, 'samples': 13868544, 'steps': 27086, 'loss/train': 2.132291793823242} 02/25/2022 12:41:31 - INFO - codeparrot_training - Step 27087: {'lr': 0.00023222900416490912, 'samples': 13869056, 'steps': 27087, 'loss/train': 1.0688408613204956} 02/25/2022 12:41:38 - INFO - codeparrot_training - Step 27088: {'lr': 0.00023221268313280836, 'samples': 13869568, 'steps': 27088, 'loss/train': 1.203363299369812} 02/25/2022 12:41:41 - INFO - codeparrot_training - Step 27089: {'lr': 0.00023219636217690267, 'samples': 13870080, 'steps': 27089, 'loss/train': 2.549347400665283} 02/25/2022 12:41:48 - INFO - codeparrot_training - Step 27090: {'lr': 0.00023218004129726218, 'samples': 13870592, 'steps': 27090, 'loss/train': 2.318603754043579} 02/25/2022 12:41:51 - INFO - codeparrot_training - Step 27091: {'lr': 0.00023216372049395667, 'samples': 13871104, 'steps': 27091, 'loss/train': 2.312936305999756} 02/25/2022 12:41:55 - INFO - codeparrot_training - Step 27092: {'lr': 0.00023214739976705614, 'samples': 13871616, 'steps': 27092, 'loss/train': 2.578765869140625} 02/25/2022 12:42:00 - INFO - codeparrot_training - Step 27093: {'lr': 0.00023213107911663062, 'samples': 13872128, 'steps': 27093, 'loss/train': 1.841038465499878} 02/25/2022 12:42:04 - INFO - codeparrot_training - Step 27094: {'lr': 0.00023211475854274975, 'samples': 13872640, 'steps': 27094, 'loss/train': 2.7046167850494385} 02/25/2022 12:42:09 - INFO - codeparrot_training - Step 27095: {'lr': 0.00023209843804548357, 'samples': 13873152, 'steps': 27095, 'loss/train': 2.2283904552459717} 02/25/2022 12:42:13 - INFO - codeparrot_training - Step 27096: {'lr': 0.000232082117624902, 'samples': 13873664, 'steps': 27096, 'loss/train': 2.249000072479248} 02/25/2022 12:42:18 - INFO - codeparrot_training - Step 27097: {'lr': 0.00023206579728107506, 'samples': 13874176, 'steps': 27097, 'loss/train': 1.733628511428833} 02/25/2022 12:42:22 - INFO - codeparrot_training - Step 27098: {'lr': 0.00023204947701407243, 'samples': 13874688, 'steps': 27098, 'loss/train': 1.3768359422683716} 02/25/2022 12:42:29 - INFO - codeparrot_training - Step 27099: {'lr': 0.00023203315682396414, 'samples': 13875200, 'steps': 27099, 'loss/train': 0.6457328200340271} 02/25/2022 12:42:32 - INFO - codeparrot_training - Step 27100: {'lr': 0.00023201683671082013, 'samples': 13875712, 'steps': 27100, 'loss/train': 2.564887762069702} 02/25/2022 12:42:38 - INFO - codeparrot_training - Step 27101: {'lr': 0.00023200051667471031, 'samples': 13876224, 'steps': 27101, 'loss/train': 1.8683359622955322} 02/25/2022 12:42:41 - INFO - codeparrot_training - Step 27102: {'lr': 0.00023198419671570456, 'samples': 13876736, 'steps': 27102, 'loss/train': 1.8324161767959595} 02/25/2022 12:42:47 - INFO - codeparrot_training - Step 27103: {'lr': 0.00023196787683387276, 'samples': 13877248, 'steps': 27103, 'loss/train': 0.6329008936882019} 02/25/2022 12:42:50 - INFO - codeparrot_training - Step 27104: {'lr': 0.00023195155702928483, 'samples': 13877760, 'steps': 27104, 'loss/train': 2.7817440032958984} 02/25/2022 12:42:56 - INFO - codeparrot_training - Step 27105: {'lr': 0.00023193523730201076, 'samples': 13878272, 'steps': 27105, 'loss/train': 2.1233136653900146} 02/25/2022 12:42:59 - INFO - codeparrot_training - Step 27106: {'lr': 0.0002319189176521204, 'samples': 13878784, 'steps': 27106, 'loss/train': 2.404804229736328} 02/25/2022 12:43:05 - INFO - codeparrot_training - Step 27107: {'lr': 0.00023190259807968357, 'samples': 13879296, 'steps': 27107, 'loss/train': 1.9561907052993774} 02/25/2022 12:43:08 - INFO - codeparrot_training - Step 27108: {'lr': 0.00023188627858477035, 'samples': 13879808, 'steps': 27108, 'loss/train': 2.178588628768921} 02/25/2022 12:43:15 - INFO - codeparrot_training - Step 27109: {'lr': 0.00023186995916745051, 'samples': 13880320, 'steps': 27109, 'loss/train': 1.5745863914489746} 02/25/2022 12:43:19 - INFO - codeparrot_training - Step 27110: {'lr': 0.00023185363982779406, 'samples': 13880832, 'steps': 27110, 'loss/train': 1.8510769605636597} 02/25/2022 12:43:24 - INFO - codeparrot_training - Step 27111: {'lr': 0.0002318373205658708, 'samples': 13881344, 'steps': 27111, 'loss/train': 1.5771312713623047} 02/25/2022 12:43:28 - INFO - codeparrot_training - Step 27112: {'lr': 0.00023182100138175073, 'samples': 13881856, 'steps': 27112, 'loss/train': 2.4008595943450928} 02/25/2022 12:43:34 - INFO - codeparrot_training - Step 27113: {'lr': 0.00023180468227550368, 'samples': 13882368, 'steps': 27113, 'loss/train': 2.3287513256073} 02/25/2022 12:43:37 - INFO - codeparrot_training - Step 27114: {'lr': 0.00023178836324719957, 'samples': 13882880, 'steps': 27114, 'loss/train': 2.1267924308776855} 02/25/2022 12:43:43 - INFO - codeparrot_training - Step 27115: {'lr': 0.00023177204429690847, 'samples': 13883392, 'steps': 27115, 'loss/train': 2.1555447578430176} 02/25/2022 12:43:46 - INFO - codeparrot_training - Step 27116: {'lr': 0.00023175572542469998, 'samples': 13883904, 'steps': 27116, 'loss/train': 2.275451183319092} 02/25/2022 12:43:53 - INFO - codeparrot_training - Step 27117: {'lr': 0.0002317394066306442, 'samples': 13884416, 'steps': 27117, 'loss/train': 1.5330990552902222} 02/25/2022 12:43:57 - INFO - codeparrot_training - Step 27118: {'lr': 0.00023172308791481104, 'samples': 13884928, 'steps': 27118, 'loss/train': 2.3393394947052} 02/25/2022 12:44:02 - INFO - codeparrot_training - Step 27119: {'lr': 0.00023170676927727045, 'samples': 13885440, 'steps': 27119, 'loss/train': 8.617653846740723} 02/25/2022 12:44:06 - INFO - codeparrot_training - Step 27120: {'lr': 0.00023169045071809213, 'samples': 13885952, 'steps': 27120, 'loss/train': 2.4430043697357178} 02/25/2022 12:44:11 - INFO - codeparrot_training - Step 27121: {'lr': 0.00023167413223734612, 'samples': 13886464, 'steps': 27121, 'loss/train': 2.00612473487854} 02/25/2022 12:44:15 - INFO - codeparrot_training - Step 27122: {'lr': 0.00023165781383510228, 'samples': 13886976, 'steps': 27122, 'loss/train': 0.4821772575378418} 02/25/2022 12:44:21 - INFO - codeparrot_training - Step 27123: {'lr': 0.00023164149551143067, 'samples': 13887488, 'steps': 27123, 'loss/train': 2.098259687423706} 02/25/2022 12:44:24 - INFO - codeparrot_training - Step 27124: {'lr': 0.00023162517726640097, 'samples': 13888000, 'steps': 27124, 'loss/train': 0.997495174407959} 02/25/2022 12:44:30 - INFO - codeparrot_training - Step 27125: {'lr': 0.00023160885910008318, 'samples': 13888512, 'steps': 27125, 'loss/train': 1.4886780977249146} 02/25/2022 12:44:33 - INFO - codeparrot_training - Step 27126: {'lr': 0.00023159254101254718, 'samples': 13889024, 'steps': 27126, 'loss/train': 2.1669259071350098} 02/25/2022 12:44:39 - INFO - codeparrot_training - Step 27127: {'lr': 0.00023157622300386287, 'samples': 13889536, 'steps': 27127, 'loss/train': 1.8937004804611206} 02/25/2022 12:44:42 - INFO - codeparrot_training - Step 27128: {'lr': 0.00023155990507410032, 'samples': 13890048, 'steps': 27128, 'loss/train': 2.1748459339141846} 02/25/2022 12:44:48 - INFO - codeparrot_training - Step 27129: {'lr': 0.00023154358722332917, 'samples': 13890560, 'steps': 27129, 'loss/train': 2.040041446685791} 02/25/2022 12:44:51 - INFO - codeparrot_training - Step 27130: {'lr': 0.0002315272694516194, 'samples': 13891072, 'steps': 27130, 'loss/train': 2.510655403137207} 02/25/2022 12:44:57 - INFO - codeparrot_training - Step 27131: {'lr': 0.00023151095175904095, 'samples': 13891584, 'steps': 27131, 'loss/train': 2.11201548576355} 02/25/2022 12:45:00 - INFO - codeparrot_training - Step 27132: {'lr': 0.00023149463414566383, 'samples': 13892096, 'steps': 27132, 'loss/train': 1.9070948362350464} 02/25/2022 12:45:06 - INFO - codeparrot_training - Step 27133: {'lr': 0.0002314783166115577, 'samples': 13892608, 'steps': 27133, 'loss/train': 1.5832808017730713} 02/25/2022 12:45:09 - INFO - codeparrot_training - Step 27134: {'lr': 0.0002314619991567926, 'samples': 13893120, 'steps': 27134, 'loss/train': 1.957318663597107} 02/25/2022 12:45:16 - INFO - codeparrot_training - Step 27135: {'lr': 0.0002314456817814384, 'samples': 13893632, 'steps': 27135, 'loss/train': 1.5475295782089233} 02/25/2022 12:45:20 - INFO - codeparrot_training - Step 27136: {'lr': 0.0002314293644855651, 'samples': 13894144, 'steps': 27136, 'loss/train': 1.2550175189971924} 02/25/2022 12:45:25 - INFO - codeparrot_training - Step 27137: {'lr': 0.00023141304726924242, 'samples': 13894656, 'steps': 27137, 'loss/train': 0.10717252641916275} 02/25/2022 12:45:29 - INFO - codeparrot_training - Step 27138: {'lr': 0.00023139673013254034, 'samples': 13895168, 'steps': 27138, 'loss/train': 1.3263659477233887} 02/25/2022 12:45:35 - INFO - codeparrot_training - Step 27139: {'lr': 0.00023138041307552875, 'samples': 13895680, 'steps': 27139, 'loss/train': 1.5434885025024414} 02/25/2022 12:45:38 - INFO - codeparrot_training - Step 27140: {'lr': 0.00023136409609827757, 'samples': 13896192, 'steps': 27140, 'loss/train': 1.9732574224472046} 02/25/2022 12:45:43 - INFO - codeparrot_training - Step 27141: {'lr': 0.00023134777920085677, 'samples': 13896704, 'steps': 27141, 'loss/train': 1.5463266372680664} 02/25/2022 12:45:47 - INFO - codeparrot_training - Step 27142: {'lr': 0.00023133146238333609, 'samples': 13897216, 'steps': 27142, 'loss/train': 2.0716617107391357} 02/25/2022 12:45:52 - INFO - codeparrot_training - Step 27143: {'lr': 0.00023131514564578547, 'samples': 13897728, 'steps': 27143, 'loss/train': 2.009819269180298} 02/25/2022 12:45:56 - INFO - codeparrot_training - Step 27144: {'lr': 0.00023129882898827484, 'samples': 13898240, 'steps': 27144, 'loss/train': 1.7620108127593994} 02/25/2022 12:46:03 - INFO - codeparrot_training - Step 27145: {'lr': 0.0002312825124108742, 'samples': 13898752, 'steps': 27145, 'loss/train': 1.8677525520324707} 02/25/2022 12:46:07 - INFO - codeparrot_training - Step 27146: {'lr': 0.0002312661959136532, 'samples': 13899264, 'steps': 27146, 'loss/train': 1.3752330541610718} 02/25/2022 12:46:13 - INFO - codeparrot_training - Step 27147: {'lr': 0.0002312498794966819, 'samples': 13899776, 'steps': 27147, 'loss/train': 2.3319666385650635} 02/25/2022 12:46:16 - INFO - codeparrot_training - Step 27148: {'lr': 0.0002312335631600302, 'samples': 13900288, 'steps': 27148, 'loss/train': 2.804506301879883} 02/25/2022 12:46:20 - INFO - codeparrot_training - Step 27149: {'lr': 0.000231217246903768, 'samples': 13900800, 'steps': 27149, 'loss/train': 1.6366115808486938} 02/25/2022 12:46:25 - INFO - codeparrot_training - Step 27150: {'lr': 0.00023120093072796506, 'samples': 13901312, 'steps': 27150, 'loss/train': 1.6016454696655273} 02/25/2022 12:46:31 - INFO - codeparrot_training - Step 27151: {'lr': 0.00023118461463269137, 'samples': 13901824, 'steps': 27151, 'loss/train': 1.7640063762664795} 02/25/2022 12:46:34 - INFO - codeparrot_training - Step 27152: {'lr': 0.00023116829861801686, 'samples': 13902336, 'steps': 27152, 'loss/train': 2.5405426025390625} 02/25/2022 12:46:38 - INFO - codeparrot_training - Step 27153: {'lr': 0.0002311519826840113, 'samples': 13902848, 'steps': 27153, 'loss/train': 2.9520487785339355} 02/25/2022 12:46:43 - INFO - codeparrot_training - Step 27154: {'lr': 0.00023113566683074477, 'samples': 13903360, 'steps': 27154, 'loss/train': 1.6504452228546143} 02/25/2022 12:46:47 - INFO - codeparrot_training - Step 27155: {'lr': 0.00023111935105828702, 'samples': 13903872, 'steps': 27155, 'loss/train': 2.4168541431427} 02/25/2022 12:46:55 - INFO - codeparrot_training - Step 27156: {'lr': 0.00023110303536670793, 'samples': 13904384, 'steps': 27156, 'loss/train': 2.660118579864502} 02/25/2022 12:46:58 - INFO - codeparrot_training - Step 27157: {'lr': 0.00023108671975607741, 'samples': 13904896, 'steps': 27157, 'loss/train': 2.314059019088745} 02/25/2022 12:47:04 - INFO - codeparrot_training - Step 27158: {'lr': 0.00023107040422646547, 'samples': 13905408, 'steps': 27158, 'loss/train': 1.920183777809143} 02/25/2022 12:47:07 - INFO - codeparrot_training - Step 27159: {'lr': 0.0002310540887779419, 'samples': 13905920, 'steps': 27159, 'loss/train': 1.5839715003967285} 02/25/2022 12:47:13 - INFO - codeparrot_training - Step 27160: {'lr': 0.0002310377734105765, 'samples': 13906432, 'steps': 27160, 'loss/train': 1.7582656145095825} 02/25/2022 12:47:16 - INFO - codeparrot_training - Step 27161: {'lr': 0.0002310214581244393, 'samples': 13906944, 'steps': 27161, 'loss/train': 1.8979930877685547} 02/25/2022 12:47:22 - INFO - codeparrot_training - Step 27162: {'lr': 0.00023100514291960017, 'samples': 13907456, 'steps': 27162, 'loss/train': 2.1478185653686523} 02/25/2022 12:47:25 - INFO - codeparrot_training - Step 27163: {'lr': 0.00023098882779612904, 'samples': 13907968, 'steps': 27163, 'loss/train': 2.982717514038086} 02/25/2022 12:47:31 - INFO - codeparrot_training - Step 27164: {'lr': 0.00023097251275409564, 'samples': 13908480, 'steps': 27164, 'loss/train': 1.919095754623413} 02/25/2022 12:47:34 - INFO - codeparrot_training - Step 27165: {'lr': 0.00023095619779356998, 'samples': 13908992, 'steps': 27165, 'loss/train': 1.1683413982391357} 02/25/2022 12:47:42 - INFO - codeparrot_training - Step 27166: {'lr': 0.00023093988291462186, 'samples': 13909504, 'steps': 27166, 'loss/train': 2.4578590393066406} 02/25/2022 12:47:45 - INFO - codeparrot_training - Step 27167: {'lr': 0.0002309235681173214, 'samples': 13910016, 'steps': 27167, 'loss/train': 1.0059542655944824} 02/25/2022 12:47:51 - INFO - codeparrot_training - Step 27168: {'lr': 0.00023090725340173818, 'samples': 13910528, 'steps': 27168, 'loss/train': 1.317365050315857} 02/25/2022 12:47:54 - INFO - codeparrot_training - Step 27169: {'lr': 0.00023089093876794224, 'samples': 13911040, 'steps': 27169, 'loss/train': 2.1020314693450928} 02/25/2022 12:48:00 - INFO - codeparrot_training - Step 27170: {'lr': 0.0002308746242160034, 'samples': 13911552, 'steps': 27170, 'loss/train': 1.8312489986419678} 02/25/2022 12:48:03 - INFO - codeparrot_training - Step 27171: {'lr': 0.00023085830974599176, 'samples': 13912064, 'steps': 27171, 'loss/train': 2.5547494888305664} 02/25/2022 12:48:09 - INFO - codeparrot_training - Step 27172: {'lr': 0.00023084199535797692, 'samples': 13912576, 'steps': 27172, 'loss/train': 2.441981792449951} 02/25/2022 12:48:12 - INFO - codeparrot_training - Step 27173: {'lr': 0.00023082568105202889, 'samples': 13913088, 'steps': 27173, 'loss/train': 1.776422142982483} 02/25/2022 12:48:18 - INFO - codeparrot_training - Step 27174: {'lr': 0.00023080936682821755, 'samples': 13913600, 'steps': 27174, 'loss/train': 1.1487456560134888} 02/25/2022 12:48:21 - INFO - codeparrot_training - Step 27175: {'lr': 0.00023079305268661277, 'samples': 13914112, 'steps': 27175, 'loss/train': 2.4722201824188232} 02/25/2022 12:48:27 - INFO - codeparrot_training - Step 27176: {'lr': 0.0002307767386272846, 'samples': 13914624, 'steps': 27176, 'loss/train': 0.8710139393806458} 02/25/2022 12:48:30 - INFO - codeparrot_training - Step 27177: {'lr': 0.00023076042465030265, 'samples': 13915136, 'steps': 27177, 'loss/train': 1.998792052268982} 02/25/2022 12:48:36 - INFO - codeparrot_training - Step 27178: {'lr': 0.00023074411075573692, 'samples': 13915648, 'steps': 27178, 'loss/train': 0.7048488855361938} 02/25/2022 12:48:39 - INFO - codeparrot_training - Step 27179: {'lr': 0.00023072779694365732, 'samples': 13916160, 'steps': 27179, 'loss/train': 1.5979152917861938} 02/25/2022 12:48:45 - INFO - codeparrot_training - Step 27180: {'lr': 0.00023071148321413383, 'samples': 13916672, 'steps': 27180, 'loss/train': 2.3554227352142334} 02/25/2022 12:48:48 - INFO - codeparrot_training - Step 27181: {'lr': 0.00023069516956723607, 'samples': 13917184, 'steps': 27181, 'loss/train': 1.9561262130737305} 02/25/2022 12:48:55 - INFO - codeparrot_training - Step 27182: {'lr': 0.0002306788560030341, 'samples': 13917696, 'steps': 27182, 'loss/train': 1.8612322807312012} 02/25/2022 12:48:59 - INFO - codeparrot_training - Step 27183: {'lr': 0.0002306625425215978, 'samples': 13918208, 'steps': 27183, 'loss/train': 2.59479022026062} 02/25/2022 12:49:04 - INFO - codeparrot_training - Step 27184: {'lr': 0.00023064622912299712, 'samples': 13918720, 'steps': 27184, 'loss/train': 1.6724401712417603} 02/25/2022 12:49:08 - INFO - codeparrot_training - Step 27185: {'lr': 0.00023062991580730176, 'samples': 13919232, 'steps': 27185, 'loss/train': 2.0189878940582275} 02/25/2022 12:49:13 - INFO - codeparrot_training - Step 27186: {'lr': 0.00023061360257458166, 'samples': 13919744, 'steps': 27186, 'loss/train': 1.8643494844436646} 02/25/2022 12:49:17 - INFO - codeparrot_training - Step 27187: {'lr': 0.00023059728942490673, 'samples': 13920256, 'steps': 27187, 'loss/train': 2.181403160095215} 02/25/2022 12:49:22 - INFO - codeparrot_training - Step 27188: {'lr': 0.00023058097635834693, 'samples': 13920768, 'steps': 27188, 'loss/train': 2.475731372833252} 02/25/2022 12:49:26 - INFO - codeparrot_training - Step 27189: {'lr': 0.00023056466337497206, 'samples': 13921280, 'steps': 27189, 'loss/train': 2.5900321006774902} 02/25/2022 12:49:31 - INFO - codeparrot_training - Step 27190: {'lr': 0.00023054835047485193, 'samples': 13921792, 'steps': 27190, 'loss/train': 2.261118173599243} 02/25/2022 12:49:35 - INFO - codeparrot_training - Step 27191: {'lr': 0.00023053203765805648, 'samples': 13922304, 'steps': 27191, 'loss/train': 1.4558427333831787} 02/25/2022 12:49:42 - INFO - codeparrot_training - Step 27192: {'lr': 0.00023051572492465566, 'samples': 13922816, 'steps': 27192, 'loss/train': 1.7706620693206787} 02/25/2022 12:49:45 - INFO - codeparrot_training - Step 27193: {'lr': 0.00023049941227471933, 'samples': 13923328, 'steps': 27193, 'loss/train': 1.2718539237976074} 02/25/2022 12:49:51 - INFO - codeparrot_training - Step 27194: {'lr': 0.0002304830997083172, 'samples': 13923840, 'steps': 27194, 'loss/train': 1.5267246961593628} 02/25/2022 12:49:54 - INFO - codeparrot_training - Step 27195: {'lr': 0.0002304667872255194, 'samples': 13924352, 'steps': 27195, 'loss/train': 1.4698207378387451} 02/25/2022 12:50:00 - INFO - codeparrot_training - Step 27196: {'lr': 0.00023045047482639556, 'samples': 13924864, 'steps': 27196, 'loss/train': 1.576555848121643} 02/25/2022 12:50:03 - INFO - codeparrot_training - Step 27197: {'lr': 0.00023043416251101575, 'samples': 13925376, 'steps': 27197, 'loss/train': 0.3115084171295166} 02/25/2022 12:50:09 - INFO - codeparrot_training - Step 27198: {'lr': 0.00023041785027944975, 'samples': 13925888, 'steps': 27198, 'loss/train': 2.4912776947021484} 02/25/2022 12:50:12 - INFO - codeparrot_training - Step 27199: {'lr': 0.0002304015381317675, 'samples': 13926400, 'steps': 27199, 'loss/train': 0.9662851691246033} 02/25/2022 12:50:18 - INFO - codeparrot_training - Step 27200: {'lr': 0.0002303852260680388, 'samples': 13926912, 'steps': 27200, 'loss/train': 1.828028678894043} 02/25/2022 12:50:21 - INFO - codeparrot_training - Step 27201: {'lr': 0.0002303689140883335, 'samples': 13927424, 'steps': 27201, 'loss/train': 1.5456184148788452} 02/25/2022 12:50:29 - INFO - codeparrot_training - Step 27202: {'lr': 0.00023035260219272168, 'samples': 13927936, 'steps': 27202, 'loss/train': 2.411005735397339} 02/25/2022 12:50:32 - INFO - codeparrot_training - Step 27203: {'lr': 0.00023033629038127302, 'samples': 13928448, 'steps': 27203, 'loss/train': 2.0583863258361816} 02/25/2022 12:50:38 - INFO - codeparrot_training - Step 27204: {'lr': 0.0002303199786540574, 'samples': 13928960, 'steps': 27204, 'loss/train': 2.002237319946289} 02/25/2022 12:50:41 - INFO - codeparrot_training - Step 27205: {'lr': 0.00023030366701114476, 'samples': 13929472, 'steps': 27205, 'loss/train': 2.032560110092163} 02/25/2022 12:50:47 - INFO - codeparrot_training - Step 27206: {'lr': 0.00023028735545260504, 'samples': 13929984, 'steps': 27206, 'loss/train': 1.6462011337280273} 02/25/2022 12:50:50 - INFO - codeparrot_training - Step 27207: {'lr': 0.00023027104397850795, 'samples': 13930496, 'steps': 27207, 'loss/train': 2.0393178462982178} 02/25/2022 12:50:56 - INFO - codeparrot_training - Step 27208: {'lr': 0.00023025473258892344, 'samples': 13931008, 'steps': 27208, 'loss/train': 1.538419246673584} 02/25/2022 12:50:59 - INFO - codeparrot_training - Step 27209: {'lr': 0.00023023842128392136, 'samples': 13931520, 'steps': 27209, 'loss/train': 1.1460812091827393} 02/25/2022 12:51:05 - INFO - codeparrot_training - Step 27210: {'lr': 0.0002302221100635716, 'samples': 13932032, 'steps': 27210, 'loss/train': 1.9465043544769287} 02/25/2022 12:51:08 - INFO - codeparrot_training - Step 27211: {'lr': 0.00023020579892794418, 'samples': 13932544, 'steps': 27211, 'loss/train': 2.175814151763916} 02/25/2022 12:51:16 - INFO - codeparrot_training - Step 27212: {'lr': 0.00023018948787710872, 'samples': 13933056, 'steps': 27212, 'loss/train': 2.371870279312134} 02/25/2022 12:51:19 - INFO - codeparrot_training - Step 27213: {'lr': 0.00023017317691113518, 'samples': 13933568, 'steps': 27213, 'loss/train': 2.3585050106048584} 02/25/2022 12:51:24 - INFO - codeparrot_training - Step 27214: {'lr': 0.00023015686603009347, 'samples': 13934080, 'steps': 27214, 'loss/train': 1.5567394495010376} 02/25/2022 12:51:28 - INFO - codeparrot_training - Step 27215: {'lr': 0.00023014055523405355, 'samples': 13934592, 'steps': 27215, 'loss/train': 1.6095361709594727} 02/25/2022 12:51:33 - INFO - codeparrot_training - Step 27216: {'lr': 0.00023012424452308508, 'samples': 13935104, 'steps': 27216, 'loss/train': 1.4049408435821533} 02/25/2022 12:51:37 - INFO - codeparrot_training - Step 27217: {'lr': 0.00023010793389725804, 'samples': 13935616, 'steps': 27217, 'loss/train': 2.5144336223602295} 02/25/2022 12:51:43 - INFO - codeparrot_training - Step 27218: {'lr': 0.0002300916233566423, 'samples': 13936128, 'steps': 27218, 'loss/train': 1.5060300827026367} 02/25/2022 12:51:46 - INFO - codeparrot_training - Step 27219: {'lr': 0.00023007531290130783, 'samples': 13936640, 'steps': 27219, 'loss/train': 2.409358501434326} 02/25/2022 12:51:51 - INFO - codeparrot_training - Step 27220: {'lr': 0.00023005900253132429, 'samples': 13937152, 'steps': 27220, 'loss/train': 2.524146795272827} 02/25/2022 12:51:55 - INFO - codeparrot_training - Step 27221: {'lr': 0.0002300426922467616, 'samples': 13937664, 'steps': 27221, 'loss/train': 1.2848683595657349} 02/25/2022 12:52:01 - INFO - codeparrot_training - Step 27222: {'lr': 0.00023002638204768975, 'samples': 13938176, 'steps': 27222, 'loss/train': 1.17233145236969} 02/25/2022 12:52:04 - INFO - codeparrot_training - Step 27223: {'lr': 0.0002300100719341785, 'samples': 13938688, 'steps': 27223, 'loss/train': 1.143632411956787} 02/25/2022 12:52:10 - INFO - codeparrot_training - Step 27224: {'lr': 0.00022999376190629788, 'samples': 13939200, 'steps': 27224, 'loss/train': 2.301959753036499} 02/25/2022 12:52:13 - INFO - codeparrot_training - Step 27225: {'lr': 0.0002299774519641175, 'samples': 13939712, 'steps': 27225, 'loss/train': 1.6440844535827637} 02/25/2022 12:52:18 - INFO - codeparrot_training - Step 27226: {'lr': 0.00022996114210770735, 'samples': 13940224, 'steps': 27226, 'loss/train': 1.3331406116485596} 02/25/2022 12:52:22 - INFO - codeparrot_training - Step 27227: {'lr': 0.00022994483233713733, 'samples': 13940736, 'steps': 27227, 'loss/train': 2.461721658706665} 02/25/2022 12:52:29 - INFO - codeparrot_training - Step 27228: {'lr': 0.00022992852265247738, 'samples': 13941248, 'steps': 27228, 'loss/train': 2.0594546794891357} 02/25/2022 12:52:33 - INFO - codeparrot_training - Step 27229: {'lr': 0.00022991221305379718, 'samples': 13941760, 'steps': 27229, 'loss/train': 2.064624071121216} 02/25/2022 12:52:38 - INFO - codeparrot_training - Step 27230: {'lr': 0.00022989590354116664, 'samples': 13942272, 'steps': 27230, 'loss/train': 1.986768364906311} 02/25/2022 12:52:42 - INFO - codeparrot_training - Step 27231: {'lr': 0.0002298795941146557, 'samples': 13942784, 'steps': 27231, 'loss/train': 1.9297752380371094} 02/25/2022 12:52:47 - INFO - codeparrot_training - Step 27232: {'lr': 0.00022986328477433426, 'samples': 13943296, 'steps': 27232, 'loss/train': 2.690293073654175} 02/25/2022 12:52:53 - INFO - codeparrot_training - Step 27233: {'lr': 0.00022984697552027203, 'samples': 13943808, 'steps': 27233, 'loss/train': 2.1552116870880127} 02/25/2022 12:52:56 - INFO - codeparrot_training - Step 27234: {'lr': 0.00022983066635253897, 'samples': 13944320, 'steps': 27234, 'loss/train': 2.2502670288085938} 02/25/2022 12:53:02 - INFO - codeparrot_training - Step 27235: {'lr': 0.00022981435727120498, 'samples': 13944832, 'steps': 27235, 'loss/train': 1.5809038877487183} 02/25/2022 12:53:05 - INFO - codeparrot_training - Step 27236: {'lr': 0.0002297980482763398, 'samples': 13945344, 'steps': 27236, 'loss/train': 1.8932820558547974} 02/25/2022 12:53:13 - INFO - codeparrot_training - Step 27237: {'lr': 0.00022978173936801343, 'samples': 13945856, 'steps': 27237, 'loss/train': 1.3279972076416016} 02/25/2022 12:53:16 - INFO - codeparrot_training - Step 27238: {'lr': 0.00022976543054629563, 'samples': 13946368, 'steps': 27238, 'loss/train': 4.221709728240967} 02/25/2022 12:53:22 - INFO - codeparrot_training - Step 27239: {'lr': 0.00022974912181125633, 'samples': 13946880, 'steps': 27239, 'loss/train': 2.5972936153411865} 02/25/2022 12:53:25 - INFO - codeparrot_training - Step 27240: {'lr': 0.0002297328131629653, 'samples': 13947392, 'steps': 27240, 'loss/train': 1.7605942487716675} 02/25/2022 12:53:31 - INFO - codeparrot_training - Step 27241: {'lr': 0.00022971650460149254, 'samples': 13947904, 'steps': 27241, 'loss/train': 1.8697713613510132} 02/25/2022 12:53:34 - INFO - codeparrot_training - Step 27242: {'lr': 0.0002297001961269078, 'samples': 13948416, 'steps': 27242, 'loss/train': 1.192578911781311} 02/25/2022 12:53:40 - INFO - codeparrot_training - Step 27243: {'lr': 0.000229683887739281, 'samples': 13948928, 'steps': 27243, 'loss/train': 0.8933893442153931} 02/25/2022 12:53:43 - INFO - codeparrot_training - Step 27244: {'lr': 0.0002296675794386819, 'samples': 13949440, 'steps': 27244, 'loss/train': 2.0607001781463623} 02/25/2022 12:53:49 - INFO - codeparrot_training - Step 27245: {'lr': 0.0002296512712251805, 'samples': 13949952, 'steps': 27245, 'loss/train': 1.4638926982879639} 02/25/2022 12:53:52 - INFO - codeparrot_training - Step 27246: {'lr': 0.00022963496309884662, 'samples': 13950464, 'steps': 27246, 'loss/train': 2.27056622505188} 02/25/2022 12:53:58 - INFO - codeparrot_training - Step 27247: {'lr': 0.00022961865505975002, 'samples': 13950976, 'steps': 27247, 'loss/train': 1.632817029953003} 02/25/2022 12:54:01 - INFO - codeparrot_training - Step 27248: {'lr': 0.00022960234710796062, 'samples': 13951488, 'steps': 27248, 'loss/train': 1.8588624000549316} 02/25/2022 12:54:09 - INFO - codeparrot_training - Step 27249: {'lr': 0.00022958603924354832, 'samples': 13952000, 'steps': 27249, 'loss/train': 1.4994919300079346} 02/25/2022 12:54:12 - INFO - codeparrot_training - Step 27250: {'lr': 0.00022956973146658302, 'samples': 13952512, 'steps': 27250, 'loss/train': 1.7028942108154297} 02/25/2022 12:54:17 - INFO - codeparrot_training - Step 27251: {'lr': 0.00022955342377713439, 'samples': 13953024, 'steps': 27251, 'loss/train': 1.4811912775039673} 02/25/2022 12:54:21 - INFO - codeparrot_training - Step 27252: {'lr': 0.00022953711617527243, 'samples': 13953536, 'steps': 27252, 'loss/train': 1.858709692955017} 02/25/2022 12:54:26 - INFO - codeparrot_training - Step 27253: {'lr': 0.00022952080866106694, 'samples': 13954048, 'steps': 27253, 'loss/train': 1.3391867876052856} 02/25/2022 12:54:30 - INFO - codeparrot_training - Step 27254: {'lr': 0.00022950450123458794, 'samples': 13954560, 'steps': 27254, 'loss/train': 2.081342935562134} 02/25/2022 12:54:35 - INFO - codeparrot_training - Step 27255: {'lr': 0.00022948819389590502, 'samples': 13955072, 'steps': 27255, 'loss/train': 1.4298616647720337} 02/25/2022 12:54:39 - INFO - codeparrot_training - Step 27256: {'lr': 0.00022947188664508816, 'samples': 13955584, 'steps': 27256, 'loss/train': 1.8034216165542603} 02/25/2022 12:54:45 - INFO - codeparrot_training - Step 27257: {'lr': 0.00022945557948220725, 'samples': 13956096, 'steps': 27257, 'loss/train': 2.344057559967041} 02/25/2022 12:54:48 - INFO - codeparrot_training - Step 27258: {'lr': 0.00022943927240733216, 'samples': 13956608, 'steps': 27258, 'loss/train': 1.4996782541275024} 02/25/2022 12:54:55 - INFO - codeparrot_training - Step 27259: {'lr': 0.00022942296542053265, 'samples': 13957120, 'steps': 27259, 'loss/train': 1.955847144126892} 02/25/2022 12:54:59 - INFO - codeparrot_training - Step 27260: {'lr': 0.0002294066585218786, 'samples': 13957632, 'steps': 27260, 'loss/train': 3.158378839492798} 02/25/2022 12:55:04 - INFO - codeparrot_training - Step 27261: {'lr': 0.00022939035171143992, 'samples': 13958144, 'steps': 27261, 'loss/train': 1.7249529361724854} 02/25/2022 12:55:08 - INFO - codeparrot_training - Step 27262: {'lr': 0.00022937404498928637, 'samples': 13958656, 'steps': 27262, 'loss/train': 2.572111129760742} 02/25/2022 12:55:13 - INFO - codeparrot_training - Step 27263: {'lr': 0.000229357738355488, 'samples': 13959168, 'steps': 27263, 'loss/train': 1.3766676187515259} 02/25/2022 12:55:17 - INFO - codeparrot_training - Step 27264: {'lr': 0.0002293414318101144, 'samples': 13959680, 'steps': 27264, 'loss/train': 2.379448652267456} 02/25/2022 12:55:22 - INFO - codeparrot_training - Step 27265: {'lr': 0.00022932512535323558, 'samples': 13960192, 'steps': 27265, 'loss/train': 1.549857258796692} 02/25/2022 12:55:26 - INFO - codeparrot_training - Step 27266: {'lr': 0.0002293088189849213, 'samples': 13960704, 'steps': 27266, 'loss/train': 1.948725938796997} 02/25/2022 12:55:31 - INFO - codeparrot_training - Step 27267: {'lr': 0.00022929251270524165, 'samples': 13961216, 'steps': 27267, 'loss/train': 1.6696780920028687} 02/25/2022 12:55:35 - INFO - codeparrot_training - Step 27268: {'lr': 0.00022927620651426616, 'samples': 13961728, 'steps': 27268, 'loss/train': 2.392348051071167} 02/25/2022 12:55:41 - INFO - codeparrot_training - Step 27269: {'lr': 0.00022925990041206478, 'samples': 13962240, 'steps': 27269, 'loss/train': 1.8572660684585571} 02/25/2022 12:55:44 - INFO - codeparrot_training - Step 27270: {'lr': 0.00022924359439870746, 'samples': 13962752, 'steps': 27270, 'loss/train': 2.9471664428710938} 02/25/2022 12:55:50 - INFO - codeparrot_training - Step 27271: {'lr': 0.00022922728847426396, 'samples': 13963264, 'steps': 27271, 'loss/train': 1.9245359897613525} 02/25/2022 12:55:53 - INFO - codeparrot_training - Step 27272: {'lr': 0.00022921098263880427, 'samples': 13963776, 'steps': 27272, 'loss/train': 1.6871438026428223} 02/25/2022 12:55:58 - INFO - codeparrot_training - Step 27273: {'lr': 0.00022919467689239804, 'samples': 13964288, 'steps': 27273, 'loss/train': 1.7842377424240112} 02/25/2022 12:56:02 - INFO - codeparrot_training - Step 27274: {'lr': 0.0002291783712351152, 'samples': 13964800, 'steps': 27274, 'loss/train': 2.3293793201446533} 02/25/2022 12:56:10 - INFO - codeparrot_training - Step 27275: {'lr': 0.0002291620656670256, 'samples': 13965312, 'steps': 27275, 'loss/train': 2.4978673458099365} 02/25/2022 12:56:13 - INFO - codeparrot_training - Step 27276: {'lr': 0.00022914576018819922, 'samples': 13965824, 'steps': 27276, 'loss/train': 1.9577503204345703} 02/25/2022 12:56:19 - INFO - codeparrot_training - Step 27277: {'lr': 0.00022912945479870565, 'samples': 13966336, 'steps': 27277, 'loss/train': 1.4334455728530884} 02/25/2022 12:56:22 - INFO - codeparrot_training - Step 27278: {'lr': 0.00022911314949861488, 'samples': 13966848, 'steps': 27278, 'loss/train': 2.327921152114868} 02/25/2022 12:56:28 - INFO - codeparrot_training - Step 27279: {'lr': 0.00022909684428799675, 'samples': 13967360, 'steps': 27279, 'loss/train': 2.6758856773376465} 02/25/2022 12:56:31 - INFO - codeparrot_training - Step 27280: {'lr': 0.00022908053916692117, 'samples': 13967872, 'steps': 27280, 'loss/train': 2.4031214714050293} 02/25/2022 12:56:37 - INFO - codeparrot_training - Step 27281: {'lr': 0.00022906423413545786, 'samples': 13968384, 'steps': 27281, 'loss/train': 1.354489803314209} 02/25/2022 12:56:40 - INFO - codeparrot_training - Step 27282: {'lr': 0.00022904792919367666, 'samples': 13968896, 'steps': 27282, 'loss/train': 1.9781063795089722} 02/25/2022 12:56:44 - INFO - codeparrot_training - Step 27283: {'lr': 0.0002290316243416476, 'samples': 13969408, 'steps': 27283, 'loss/train': 1.4685373306274414} 02/25/2022 12:56:49 - INFO - codeparrot_training - Step 27284: {'lr': 0.00022901531957944033, 'samples': 13969920, 'steps': 27284, 'loss/train': 1.7582650184631348} 02/25/2022 12:56:53 - INFO - codeparrot_training - Step 27285: {'lr': 0.00022899901490712482, 'samples': 13970432, 'steps': 27285, 'loss/train': 2.623544692993164} 02/25/2022 12:57:00 - INFO - codeparrot_training - Step 27286: {'lr': 0.00022898271032477087, 'samples': 13970944, 'steps': 27286, 'loss/train': 1.838005781173706} 02/25/2022 12:57:04 - INFO - codeparrot_training - Step 27287: {'lr': 0.00022896640583244823, 'samples': 13971456, 'steps': 27287, 'loss/train': 1.701197624206543} 02/25/2022 12:57:09 - INFO - codeparrot_training - Step 27288: {'lr': 0.00022895010143022686, 'samples': 13971968, 'steps': 27288, 'loss/train': 2.727379083633423} 02/25/2022 12:57:13 - INFO - codeparrot_training - Step 27289: {'lr': 0.0002289337971181766, 'samples': 13972480, 'steps': 27289, 'loss/train': 0.43912503123283386} 02/25/2022 12:57:18 - INFO - codeparrot_training - Step 27290: {'lr': 0.00022891749289636725, 'samples': 13972992, 'steps': 27290, 'loss/train': 2.4275786876678467} 02/25/2022 12:57:22 - INFO - codeparrot_training - Step 27291: {'lr': 0.00022890118876486864, 'samples': 13973504, 'steps': 27291, 'loss/train': 1.52302086353302} 02/25/2022 12:57:27 - INFO - codeparrot_training - Step 27292: {'lr': 0.00022888488472375067, 'samples': 13974016, 'steps': 27292, 'loss/train': 1.3313792943954468} 02/25/2022 12:57:31 - INFO - codeparrot_training - Step 27293: {'lr': 0.0002288685807730832, 'samples': 13974528, 'steps': 27293, 'loss/train': 1.9958349466323853} 02/25/2022 12:57:36 - INFO - codeparrot_training - Step 27294: {'lr': 0.00022885227691293595, 'samples': 13975040, 'steps': 27294, 'loss/train': 1.7047677040100098} 02/25/2022 12:57:40 - INFO - codeparrot_training - Step 27295: {'lr': 0.0002288359731433788, 'samples': 13975552, 'steps': 27295, 'loss/train': 2.0008506774902344} 02/25/2022 12:57:47 - INFO - codeparrot_training - Step 27296: {'lr': 0.00022881966946448166, 'samples': 13976064, 'steps': 27296, 'loss/train': 1.6596100330352783} 02/25/2022 12:57:50 - INFO - codeparrot_training - Step 27297: {'lr': 0.0002288033658763143, 'samples': 13976576, 'steps': 27297, 'loss/train': 1.7757279872894287} 02/25/2022 12:57:56 - INFO - codeparrot_training - Step 27298: {'lr': 0.0002287870623789467, 'samples': 13977088, 'steps': 27298, 'loss/train': 1.891709566116333} 02/25/2022 12:57:59 - INFO - codeparrot_training - Step 27299: {'lr': 0.00022877075897244847, 'samples': 13977600, 'steps': 27299, 'loss/train': 1.0397292375564575} 02/25/2022 12:58:05 - INFO - codeparrot_training - Step 27300: {'lr': 0.0002287544556568896, 'samples': 13978112, 'steps': 27300, 'loss/train': 3.0717060565948486} 02/25/2022 12:58:09 - INFO - codeparrot_training - Step 27301: {'lr': 0.00022873815243233987, 'samples': 13978624, 'steps': 27301, 'loss/train': 1.8659656047821045} 02/25/2022 12:58:14 - INFO - codeparrot_training - Step 27302: {'lr': 0.00022872184929886928, 'samples': 13979136, 'steps': 27302, 'loss/train': 2.4029877185821533} 02/25/2022 12:58:18 - INFO - codeparrot_training - Step 27303: {'lr': 0.00022870554625654737, 'samples': 13979648, 'steps': 27303, 'loss/train': 1.7859495878219604} 02/25/2022 12:58:23 - INFO - codeparrot_training - Step 27304: {'lr': 0.0002286892433054442, 'samples': 13980160, 'steps': 27304, 'loss/train': 2.1265759468078613} 02/25/2022 12:58:27 - INFO - codeparrot_training - Step 27305: {'lr': 0.0002286729404456295, 'samples': 13980672, 'steps': 27305, 'loss/train': 2.41939640045166} 02/25/2022 12:58:34 - INFO - codeparrot_training - Step 27306: {'lr': 0.0002286566376771733, 'samples': 13981184, 'steps': 27306, 'loss/train': 1.5045340061187744} 02/25/2022 12:58:37 - INFO - codeparrot_training - Step 27307: {'lr': 0.00022864033500014514, 'samples': 13981696, 'steps': 27307, 'loss/train': 2.2248873710632324} 02/25/2022 12:58:43 - INFO - codeparrot_training - Step 27308: {'lr': 0.00022862403241461502, 'samples': 13982208, 'steps': 27308, 'loss/train': 1.2735166549682617} 02/25/2022 12:58:46 - INFO - codeparrot_training - Step 27309: {'lr': 0.00022860772992065273, 'samples': 13982720, 'steps': 27309, 'loss/train': 1.8813750743865967} 02/25/2022 12:58:52 - INFO - codeparrot_training - Step 27310: {'lr': 0.0002285914275183282, 'samples': 13983232, 'steps': 27310, 'loss/train': 1.3298530578613281} 02/25/2022 12:58:55 - INFO - codeparrot_training - Step 27311: {'lr': 0.00022857512520771124, 'samples': 13983744, 'steps': 27311, 'loss/train': 1.666682481765747} 02/25/2022 12:59:01 - INFO - codeparrot_training - Step 27312: {'lr': 0.00022855882298887154, 'samples': 13984256, 'steps': 27312, 'loss/train': 1.1620745658874512} 02/25/2022 12:59:04 - INFO - codeparrot_training - Step 27313: {'lr': 0.00022854252086187905, 'samples': 13984768, 'steps': 27313, 'loss/train': 1.8478984832763672} 02/25/2022 12:59:10 - INFO - codeparrot_training - Step 27314: {'lr': 0.00022852621882680357, 'samples': 13985280, 'steps': 27314, 'loss/train': 0.6565882563591003} 02/25/2022 12:59:13 - INFO - codeparrot_training - Step 27315: {'lr': 0.0002285099168837151, 'samples': 13985792, 'steps': 27315, 'loss/train': 2.1913156509399414} 02/25/2022 12:59:20 - INFO - codeparrot_training - Step 27316: {'lr': 0.00022849361503268317, 'samples': 13986304, 'steps': 27316, 'loss/train': 2.696385145187378} 02/25/2022 12:59:24 - INFO - codeparrot_training - Step 27317: {'lr': 0.0002284773132737778, 'samples': 13986816, 'steps': 27317, 'loss/train': 1.8935333490371704} 02/25/2022 12:59:30 - INFO - codeparrot_training - Step 27318: {'lr': 0.00022846101160706875, 'samples': 13987328, 'steps': 27318, 'loss/train': 1.5423376560211182} 02/25/2022 12:59:33 - INFO - codeparrot_training - Step 27319: {'lr': 0.00022844471003262597, 'samples': 13987840, 'steps': 27319, 'loss/train': 0.28522610664367676} 02/25/2022 12:59:39 - INFO - codeparrot_training - Step 27320: {'lr': 0.00022842840855051918, 'samples': 13988352, 'steps': 27320, 'loss/train': 1.079040765762329} 02/25/2022 12:59:42 - INFO - codeparrot_training - Step 27321: {'lr': 0.0002284121071608182, 'samples': 13988864, 'steps': 27321, 'loss/train': 0.2602623701095581} 02/25/2022 12:59:48 - INFO - codeparrot_training - Step 27322: {'lr': 0.0002283958058635929, 'samples': 13989376, 'steps': 27322, 'loss/train': 1.2343724966049194} 02/25/2022 12:59:51 - INFO - codeparrot_training - Step 27323: {'lr': 0.00022837950465891317, 'samples': 13989888, 'steps': 27323, 'loss/train': 1.2429561614990234} 02/25/2022 12:59:57 - INFO - codeparrot_training - Step 27324: {'lr': 0.00022836320354684876, 'samples': 13990400, 'steps': 27324, 'loss/train': 0.31624525785446167} 02/25/2022 13:00:00 - INFO - codeparrot_training - Step 27325: {'lr': 0.00022834690252746947, 'samples': 13990912, 'steps': 27325, 'loss/train': 2.132636547088623} 02/25/2022 13:00:06 - INFO - codeparrot_training - Step 27326: {'lr': 0.00022833060160084524, 'samples': 13991424, 'steps': 27326, 'loss/train': 0.9071618318557739} 02/25/2022 13:00:09 - INFO - codeparrot_training - Step 27327: {'lr': 0.00022831430076704573, 'samples': 13991936, 'steps': 27327, 'loss/train': 2.5429437160491943} 02/25/2022 13:00:15 - INFO - codeparrot_training - Step 27328: {'lr': 0.00022829800002614094, 'samples': 13992448, 'steps': 27328, 'loss/train': 1.949150800704956} 02/25/2022 13:00:18 - INFO - codeparrot_training - Step 27329: {'lr': 0.0002282816993782006, 'samples': 13992960, 'steps': 27329, 'loss/train': 1.4601025581359863} 02/25/2022 13:00:24 - INFO - codeparrot_training - Step 27330: {'lr': 0.0002282653988232946, 'samples': 13993472, 'steps': 27330, 'loss/train': 0.41124340891838074} 02/25/2022 13:00:27 - INFO - codeparrot_training - Step 27331: {'lr': 0.0002282490983614927, 'samples': 13993984, 'steps': 27331, 'loss/train': 1.2325631380081177} 02/25/2022 13:00:37 - INFO - codeparrot_training - Step 27332: {'lr': 0.00022823279799286472, 'samples': 13994496, 'steps': 27332, 'loss/train': 1.7864919900894165} 02/25/2022 13:00:40 - INFO - codeparrot_training - Step 27333: {'lr': 0.00022821649771748067, 'samples': 13995008, 'steps': 27333, 'loss/train': 2.319870710372925} 02/25/2022 13:00:44 - INFO - codeparrot_training - Step 27334: {'lr': 0.0002282001975354101, 'samples': 13995520, 'steps': 27334, 'loss/train': 1.877912998199463} 02/25/2022 13:00:49 - INFO - codeparrot_training - Step 27335: {'lr': 0.00022818389744672297, 'samples': 13996032, 'steps': 27335, 'loss/train': 2.3093326091766357} 02/25/2022 13:00:53 - INFO - codeparrot_training - Step 27336: {'lr': 0.00022816759745148906, 'samples': 13996544, 'steps': 27336, 'loss/train': 1.4386746883392334} 02/25/2022 13:00:58 - INFO - codeparrot_training - Step 27337: {'lr': 0.00022815129754977836, 'samples': 13997056, 'steps': 27337, 'loss/train': 2.5465614795684814} 02/25/2022 13:01:02 - INFO - codeparrot_training - Step 27338: {'lr': 0.00022813499774166046, 'samples': 13997568, 'steps': 27338, 'loss/train': 2.212357997894287} 02/25/2022 13:01:07 - INFO - codeparrot_training - Step 27339: {'lr': 0.00022811869802720532, 'samples': 13998080, 'steps': 27339, 'loss/train': 1.942797064781189} 02/25/2022 13:01:11 - INFO - codeparrot_training - Step 27340: {'lr': 0.00022810239840648268, 'samples': 13998592, 'steps': 27340, 'loss/train': 2.6351046562194824} 02/25/2022 13:01:16 - INFO - codeparrot_training - Step 27341: {'lr': 0.00022808609887956254, 'samples': 13999104, 'steps': 27341, 'loss/train': 1.2209769487380981} 02/25/2022 13:01:20 - INFO - codeparrot_training - Step 27342: {'lr': 0.00022806979944651446, 'samples': 13999616, 'steps': 27342, 'loss/train': 1.6552488803863525} 02/25/2022 13:01:27 - INFO - codeparrot_training - Step 27343: {'lr': 0.0002280535001074084, 'samples': 14000128, 'steps': 27343, 'loss/train': 1.8452142477035522} 02/25/2022 13:01:31 - INFO - codeparrot_training - Step 27344: {'lr': 0.00022803720086231422, 'samples': 14000640, 'steps': 27344, 'loss/train': 1.879871129989624} 02/25/2022 13:01:36 - INFO - codeparrot_training - Step 27345: {'lr': 0.00022802090171130166, 'samples': 14001152, 'steps': 27345, 'loss/train': 1.9572960138320923} 02/25/2022 13:01:40 - INFO - codeparrot_training - Step 27346: {'lr': 0.0002280046026544407, 'samples': 14001664, 'steps': 27346, 'loss/train': 1.7477837800979614} 02/25/2022 13:01:45 - INFO - codeparrot_training - Step 27347: {'lr': 0.00022798830369180094, 'samples': 14002176, 'steps': 27347, 'loss/train': 2.547405242919922} 02/25/2022 13:01:49 - INFO - codeparrot_training - Step 27348: {'lr': 0.00022797200482345227, 'samples': 14002688, 'steps': 27348, 'loss/train': 1.622405767440796} 02/25/2022 13:01:54 - INFO - codeparrot_training - Step 27349: {'lr': 0.00022795570604946454, 'samples': 14003200, 'steps': 27349, 'loss/train': 2.198040723800659} 02/25/2022 13:01:58 - INFO - codeparrot_training - Step 27350: {'lr': 0.00022793940736990766, 'samples': 14003712, 'steps': 27350, 'loss/train': 1.9251022338867188} 02/25/2022 13:02:03 - INFO - codeparrot_training - Step 27351: {'lr': 0.0002279231087848513, 'samples': 14004224, 'steps': 27351, 'loss/train': 1.9350048303604126} 02/25/2022 13:02:07 - INFO - codeparrot_training - Step 27352: {'lr': 0.0002279068102943653, 'samples': 14004736, 'steps': 27352, 'loss/train': 2.227863073348999} 02/25/2022 13:02:14 - INFO - codeparrot_training - Step 27353: {'lr': 0.0002278905118985195, 'samples': 14005248, 'steps': 27353, 'loss/train': 1.5180689096450806} 02/25/2022 13:02:18 - INFO - codeparrot_training - Step 27354: {'lr': 0.00022787421359738387, 'samples': 14005760, 'steps': 27354, 'loss/train': 1.5994786024093628} 02/25/2022 13:02:23 - INFO - codeparrot_training - Step 27355: {'lr': 0.00022785791539102794, 'samples': 14006272, 'steps': 27355, 'loss/train': 3.0293421745300293} 02/25/2022 13:02:27 - INFO - codeparrot_training - Step 27356: {'lr': 0.00022784161727952166, 'samples': 14006784, 'steps': 27356, 'loss/train': 2.085681438446045} 02/25/2022 13:02:32 - INFO - codeparrot_training - Step 27357: {'lr': 0.00022782531926293488, 'samples': 14007296, 'steps': 27357, 'loss/train': 2.179945707321167} 02/25/2022 13:02:36 - INFO - codeparrot_training - Step 27358: {'lr': 0.00022780902134133738, 'samples': 14007808, 'steps': 27358, 'loss/train': 1.0897220373153687} 02/25/2022 13:02:41 - INFO - codeparrot_training - Step 27359: {'lr': 0.0002277927235147991, 'samples': 14008320, 'steps': 27359, 'loss/train': 1.6038397550582886} 02/25/2022 13:02:45 - INFO - codeparrot_training - Step 27360: {'lr': 0.00022777642578338965, 'samples': 14008832, 'steps': 27360, 'loss/train': 2.2412548065185547} 02/25/2022 13:02:50 - INFO - codeparrot_training - Step 27361: {'lr': 0.00022776012814717888, 'samples': 14009344, 'steps': 27361, 'loss/train': 1.0821096897125244} 02/25/2022 13:02:54 - INFO - codeparrot_training - Step 27362: {'lr': 0.0002277438306062367, 'samples': 14009856, 'steps': 27362, 'loss/train': 1.806030511856079} 02/25/2022 13:03:01 - INFO - codeparrot_training - Step 27363: {'lr': 0.00022772753316063302, 'samples': 14010368, 'steps': 27363, 'loss/train': 2.419410467147827} 02/25/2022 13:03:04 - INFO - codeparrot_training - Step 27364: {'lr': 0.00022771123581043738, 'samples': 14010880, 'steps': 27364, 'loss/train': 2.57742977142334} 02/25/2022 13:03:10 - INFO - codeparrot_training - Step 27365: {'lr': 0.0002276949385557197, 'samples': 14011392, 'steps': 27365, 'loss/train': 2.073284149169922} 02/25/2022 13:03:13 - INFO - codeparrot_training - Step 27366: {'lr': 0.0002276786413965499, 'samples': 14011904, 'steps': 27366, 'loss/train': 2.0743749141693115} 02/25/2022 13:03:19 - INFO - codeparrot_training - Step 27367: {'lr': 0.00022766234433299764, 'samples': 14012416, 'steps': 27367, 'loss/train': 1.8015978336334229} 02/25/2022 13:03:22 - INFO - codeparrot_training - Step 27368: {'lr': 0.0002276460473651329, 'samples': 14012928, 'steps': 27368, 'loss/train': 1.7960762977600098} 02/25/2022 13:03:28 - INFO - codeparrot_training - Step 27369: {'lr': 0.0002276297504930253, 'samples': 14013440, 'steps': 27369, 'loss/train': 8.626423835754395} 02/25/2022 13:03:31 - INFO - codeparrot_training - Step 27370: {'lr': 0.00022761345371674483, 'samples': 14013952, 'steps': 27370, 'loss/train': 1.6085503101348877} 02/25/2022 13:03:37 - INFO - codeparrot_training - Step 27371: {'lr': 0.00022759715703636114, 'samples': 14014464, 'steps': 27371, 'loss/train': 2.1694164276123047} 02/25/2022 13:03:40 - INFO - codeparrot_training - Step 27372: {'lr': 0.00022758086045194417, 'samples': 14014976, 'steps': 27372, 'loss/train': 2.093355178833008} 02/25/2022 13:03:46 - INFO - codeparrot_training - Step 27373: {'lr': 0.00022756456396356368, 'samples': 14015488, 'steps': 27373, 'loss/train': 2.26033616065979} 02/25/2022 13:03:50 - INFO - codeparrot_training - Step 27374: {'lr': 0.00022754826757128942, 'samples': 14016000, 'steps': 27374, 'loss/train': 2.317751407623291} 02/25/2022 13:03:55 - INFO - codeparrot_training - Step 27375: {'lr': 0.00022753197127519127, 'samples': 14016512, 'steps': 27375, 'loss/train': 0.4168996810913086} 02/25/2022 13:03:59 - INFO - codeparrot_training - Step 27376: {'lr': 0.00022751567507533905, 'samples': 14017024, 'steps': 27376, 'loss/train': 1.4742178916931152} 02/25/2022 13:04:04 - INFO - codeparrot_training - Step 27377: {'lr': 0.00022749937897180257, 'samples': 14017536, 'steps': 27377, 'loss/train': 2.1209206581115723} 02/25/2022 13:04:08 - INFO - codeparrot_training - Step 27378: {'lr': 0.0002274830829646515, 'samples': 14018048, 'steps': 27378, 'loss/train': 0.5565173625946045} 02/25/2022 13:04:15 - INFO - codeparrot_training - Step 27379: {'lr': 0.0002274667870539558, 'samples': 14018560, 'steps': 27379, 'loss/train': 2.070378065109253} 02/25/2022 13:04:18 - INFO - codeparrot_training - Step 27380: {'lr': 0.0002274504912397852, 'samples': 14019072, 'steps': 27380, 'loss/train': 2.1325156688690186} 02/25/2022 13:04:24 - INFO - codeparrot_training - Step 27381: {'lr': 0.00022743419552220963, 'samples': 14019584, 'steps': 27381, 'loss/train': 1.6749143600463867} 02/25/2022 13:04:27 - INFO - codeparrot_training - Step 27382: {'lr': 0.00022741789990129874, 'samples': 14020096, 'steps': 27382, 'loss/train': 1.7678847312927246} 02/25/2022 13:04:33 - INFO - codeparrot_training - Step 27383: {'lr': 0.00022740160437712236, 'samples': 14020608, 'steps': 27383, 'loss/train': 1.9126551151275635} 02/25/2022 13:04:36 - INFO - codeparrot_training - Step 27384: {'lr': 0.00022738530894975034, 'samples': 14021120, 'steps': 27384, 'loss/train': 0.7265369296073914} 02/25/2022 13:04:42 - INFO - codeparrot_training - Step 27385: {'lr': 0.00022736901361925261, 'samples': 14021632, 'steps': 27385, 'loss/train': 1.8522812128067017} 02/25/2022 13:04:45 - INFO - codeparrot_training - Step 27386: {'lr': 0.00022735271838569872, 'samples': 14022144, 'steps': 27386, 'loss/train': 1.4959352016448975} 02/25/2022 13:04:51 - INFO - codeparrot_training - Step 27387: {'lr': 0.00022733642324915856, 'samples': 14022656, 'steps': 27387, 'loss/train': 2.1490225791931152} 02/25/2022 13:04:54 - INFO - codeparrot_training - Step 27388: {'lr': 0.00022732012820970198, 'samples': 14023168, 'steps': 27388, 'loss/train': 2.401810646057129} 02/25/2022 13:05:02 - INFO - codeparrot_training - Step 27389: {'lr': 0.0002273038332673989, 'samples': 14023680, 'steps': 27389, 'loss/train': 2.265002965927124} 02/25/2022 13:05:05 - INFO - codeparrot_training - Step 27390: {'lr': 0.00022728753842231886, 'samples': 14024192, 'steps': 27390, 'loss/train': 1.6038258075714111} 02/25/2022 13:05:10 - INFO - codeparrot_training - Step 27391: {'lr': 0.0002272712436745318, 'samples': 14024704, 'steps': 27391, 'loss/train': 1.2745909690856934} 02/25/2022 13:05:14 - INFO - codeparrot_training - Step 27392: {'lr': 0.0002272549490241075, 'samples': 14025216, 'steps': 27392, 'loss/train': 1.4223980903625488} 02/25/2022 13:05:19 - INFO - codeparrot_training - Step 27393: {'lr': 0.0002272386544711158, 'samples': 14025728, 'steps': 27393, 'loss/train': 2.1912713050842285} 02/25/2022 13:05:23 - INFO - codeparrot_training - Step 27394: {'lr': 0.00022722236001562656, 'samples': 14026240, 'steps': 27394, 'loss/train': 1.4348034858703613} 02/25/2022 13:05:28 - INFO - codeparrot_training - Step 27395: {'lr': 0.00022720606565770942, 'samples': 14026752, 'steps': 27395, 'loss/train': 1.9353013038635254} 02/25/2022 13:05:32 - INFO - codeparrot_training - Step 27396: {'lr': 0.00022718977139743425, 'samples': 14027264, 'steps': 27396, 'loss/train': 1.378192663192749} 02/25/2022 13:05:37 - INFO - codeparrot_training - Step 27397: {'lr': 0.00022717347723487087, 'samples': 14027776, 'steps': 27397, 'loss/train': 1.6723642349243164} 02/25/2022 13:05:41 - INFO - codeparrot_training - Step 27398: {'lr': 0.00022715718317008912, 'samples': 14028288, 'steps': 27398, 'loss/train': 2.1466236114501953} 02/25/2022 13:05:48 - INFO - codeparrot_training - Step 27399: {'lr': 0.00022714088920315867, 'samples': 14028800, 'steps': 27399, 'loss/train': 2.059225559234619} 02/25/2022 13:05:53 - INFO - codeparrot_training - Step 27400: {'lr': 0.00022712459533414943, 'samples': 14029312, 'steps': 27400, 'loss/train': 1.8407567739486694} 02/25/2022 13:05:57 - INFO - codeparrot_training - Step 27401: {'lr': 0.00022710830156313108, 'samples': 14029824, 'steps': 27401, 'loss/train': 1.5227664709091187} 02/25/2022 13:06:02 - INFO - codeparrot_training - Step 27402: {'lr': 0.00022709200789017368, 'samples': 14030336, 'steps': 27402, 'loss/train': 1.9594672918319702} 02/25/2022 13:06:06 - INFO - codeparrot_training - Step 27403: {'lr': 0.00022707571431534668, 'samples': 14030848, 'steps': 27403, 'loss/train': 1.805530309677124} 02/25/2022 13:06:12 - INFO - codeparrot_training - Step 27404: {'lr': 0.00022705942083872004, 'samples': 14031360, 'steps': 27404, 'loss/train': 2.0772674083709717} 02/25/2022 13:06:15 - INFO - codeparrot_training - Step 27405: {'lr': 0.0002270431274603636, 'samples': 14031872, 'steps': 27405, 'loss/train': 2.0689711570739746} 02/25/2022 13:06:19 - INFO - codeparrot_training - Step 27406: {'lr': 0.00022702683418034713, 'samples': 14032384, 'steps': 27406, 'loss/train': 1.2383630275726318} 02/25/2022 13:06:24 - INFO - codeparrot_training - Step 27407: {'lr': 0.00022701054099874044, 'samples': 14032896, 'steps': 27407, 'loss/train': 1.6797540187835693} 02/25/2022 13:06:30 - INFO - codeparrot_training - Step 27408: {'lr': 0.00022699424791561325, 'samples': 14033408, 'steps': 27408, 'loss/train': 3.1989011764526367} 02/25/2022 13:06:33 - INFO - codeparrot_training - Step 27409: {'lr': 0.00022697795493103535, 'samples': 14033920, 'steps': 27409, 'loss/train': 2.340059995651245} 02/25/2022 13:06:40 - INFO - codeparrot_training - Step 27410: {'lr': 0.00022696166204507666, 'samples': 14034432, 'steps': 27410, 'loss/train': 1.4605093002319336} 02/25/2022 13:06:44 - INFO - codeparrot_training - Step 27411: {'lr': 0.00022694536925780688, 'samples': 14034944, 'steps': 27411, 'loss/train': 1.4475537538528442} 02/25/2022 13:06:47 - INFO - codeparrot_training - Step 27412: {'lr': 0.00022692907656929575, 'samples': 14035456, 'steps': 27412, 'loss/train': 0.7247291207313538} 02/25/2022 13:06:53 - INFO - codeparrot_training - Step 27413: {'lr': 0.0002269127839796132, 'samples': 14035968, 'steps': 27413, 'loss/train': 1.9053713083267212} 02/25/2022 13:06:57 - INFO - codeparrot_training - Step 27414: {'lr': 0.00022689649148882894, 'samples': 14036480, 'steps': 27414, 'loss/train': 1.8588893413543701} 02/25/2022 13:07:02 - INFO - codeparrot_training - Step 27415: {'lr': 0.00022688019909701277, 'samples': 14036992, 'steps': 27415, 'loss/train': 1.2183656692504883} 02/25/2022 13:07:06 - INFO - codeparrot_training - Step 27416: {'lr': 0.00022686390680423446, 'samples': 14037504, 'steps': 27416, 'loss/train': 2.2951855659484863} 02/25/2022 13:07:11 - INFO - codeparrot_training - Step 27417: {'lr': 0.00022684761461056385, 'samples': 14038016, 'steps': 27417, 'loss/train': 1.1336044073104858} 02/25/2022 13:07:15 - INFO - codeparrot_training - Step 27418: {'lr': 0.0002268313225160707, 'samples': 14038528, 'steps': 27418, 'loss/train': 1.936003565788269} 02/25/2022 13:07:20 - INFO - codeparrot_training - Step 27419: {'lr': 0.00022681503052082478, 'samples': 14039040, 'steps': 27419, 'loss/train': 1.238242745399475} 02/25/2022 13:07:24 - INFO - codeparrot_training - Step 27420: {'lr': 0.00022679873862489603, 'samples': 14039552, 'steps': 27420, 'loss/train': 2.4736294746398926} 02/25/2022 13:07:31 - INFO - codeparrot_training - Step 27421: {'lr': 0.000226782446828354, 'samples': 14040064, 'steps': 27421, 'loss/train': 1.6786586046218872} 02/25/2022 13:07:34 - INFO - codeparrot_training - Step 27422: {'lr': 0.00022676615513126858, 'samples': 14040576, 'steps': 27422, 'loss/train': 1.5163426399230957} 02/25/2022 13:07:40 - INFO - codeparrot_training - Step 27423: {'lr': 0.00022674986353370957, 'samples': 14041088, 'steps': 27423, 'loss/train': 1.3789907693862915} 02/25/2022 13:07:43 - INFO - codeparrot_training - Step 27424: {'lr': 0.0002267335720357469, 'samples': 14041600, 'steps': 27424, 'loss/train': 2.1570146083831787} 02/25/2022 13:07:49 - INFO - codeparrot_training - Step 27425: {'lr': 0.0002267172806374501, 'samples': 14042112, 'steps': 27425, 'loss/train': 2.1576268672943115} 02/25/2022 13:07:54 - INFO - codeparrot_training - Step 27426: {'lr': 0.0002267009893388891, 'samples': 14042624, 'steps': 27426, 'loss/train': 1.5178781747817993} 02/25/2022 13:07:57 - INFO - codeparrot_training - Step 27427: {'lr': 0.00022668469814013362, 'samples': 14043136, 'steps': 27427, 'loss/train': 2.3554844856262207} 02/25/2022 13:08:03 - INFO - codeparrot_training - Step 27428: {'lr': 0.00022666840704125353, 'samples': 14043648, 'steps': 27428, 'loss/train': 1.4940383434295654} 02/25/2022 13:08:06 - INFO - codeparrot_training - Step 27429: {'lr': 0.00022665211604231864, 'samples': 14044160, 'steps': 27429, 'loss/train': 2.6759610176086426} 02/25/2022 13:08:13 - INFO - codeparrot_training - Step 27430: {'lr': 0.00022663582514339858, 'samples': 14044672, 'steps': 27430, 'loss/train': 1.6067527532577515} 02/25/2022 13:08:17 - INFO - codeparrot_training - Step 27431: {'lr': 0.00022661953434456323, 'samples': 14045184, 'steps': 27431, 'loss/train': 2.3536789417266846} 02/25/2022 13:08:22 - INFO - codeparrot_training - Step 27432: {'lr': 0.00022660324364588236, 'samples': 14045696, 'steps': 27432, 'loss/train': 2.7798709869384766} 02/25/2022 13:08:26 - INFO - codeparrot_training - Step 27433: {'lr': 0.00022658695304742592, 'samples': 14046208, 'steps': 27433, 'loss/train': 2.226771354675293} 02/25/2022 13:08:31 - INFO - codeparrot_training - Step 27434: {'lr': 0.00022657066254926336, 'samples': 14046720, 'steps': 27434, 'loss/train': 2.5460848808288574} 02/25/2022 13:08:35 - INFO - codeparrot_training - Step 27435: {'lr': 0.0002265543721514647, 'samples': 14047232, 'steps': 27435, 'loss/train': 2.075054883956909} 02/25/2022 13:08:40 - INFO - codeparrot_training - Step 27436: {'lr': 0.00022653808185409962, 'samples': 14047744, 'steps': 27436, 'loss/train': 2.302901268005371} 02/25/2022 13:08:44 - INFO - codeparrot_training - Step 27437: {'lr': 0.0002265217916572381, 'samples': 14048256, 'steps': 27437, 'loss/train': 0.9972638487815857} 02/25/2022 13:08:49 - INFO - codeparrot_training - Step 27438: {'lr': 0.00022650550156094962, 'samples': 14048768, 'steps': 27438, 'loss/train': 1.3234316110610962} 02/25/2022 13:08:53 - INFO - codeparrot_training - Step 27439: {'lr': 0.00022648921156530414, 'samples': 14049280, 'steps': 27439, 'loss/train': 0.09567834436893463} 02/25/2022 13:09:00 - INFO - codeparrot_training - Step 27440: {'lr': 0.00022647292167037142, 'samples': 14049792, 'steps': 27440, 'loss/train': 1.3499737977981567} 02/25/2022 13:09:03 - INFO - codeparrot_training - Step 27441: {'lr': 0.0002264566318762212, 'samples': 14050304, 'steps': 27441, 'loss/train': 2.0918891429901123} 02/25/2022 13:09:09 - INFO - codeparrot_training - Step 27442: {'lr': 0.00022644034218292342, 'samples': 14050816, 'steps': 27442, 'loss/train': 1.148611307144165} 02/25/2022 13:09:12 - INFO - codeparrot_training - Step 27443: {'lr': 0.00022642405259054764, 'samples': 14051328, 'steps': 27443, 'loss/train': 1.0863064527511597} 02/25/2022 13:09:18 - INFO - codeparrot_training - Step 27444: {'lr': 0.0002264077630991637, 'samples': 14051840, 'steps': 27444, 'loss/train': 1.713740587234497} 02/25/2022 13:09:21 - INFO - codeparrot_training - Step 27445: {'lr': 0.0002263914737088414, 'samples': 14052352, 'steps': 27445, 'loss/train': 2.2682785987854004} 02/25/2022 13:09:27 - INFO - codeparrot_training - Step 27446: {'lr': 0.00022637518441965068, 'samples': 14052864, 'steps': 27446, 'loss/train': 2.4317922592163086} 02/25/2022 13:09:30 - INFO - codeparrot_training - Step 27447: {'lr': 0.00022635889523166106, 'samples': 14053376, 'steps': 27447, 'loss/train': 1.267711877822876} 02/25/2022 13:09:36 - INFO - codeparrot_training - Step 27448: {'lr': 0.0002263426061449424, 'samples': 14053888, 'steps': 27448, 'loss/train': 3.0414998531341553} 02/25/2022 13:09:39 - INFO - codeparrot_training - Step 27449: {'lr': 0.0002263263171595645, 'samples': 14054400, 'steps': 27449, 'loss/train': 1.6580809354782104} 02/25/2022 13:09:45 - INFO - codeparrot_training - Step 27450: {'lr': 0.00022631002827559727, 'samples': 14054912, 'steps': 27450, 'loss/train': 2.4516777992248535} 02/25/2022 13:09:48 - INFO - codeparrot_training - Step 27451: {'lr': 0.00022629373949311024, 'samples': 14055424, 'steps': 27451, 'loss/train': 2.7914505004882812} 02/25/2022 13:09:54 - INFO - codeparrot_training - Step 27452: {'lr': 0.0002262774508121733, 'samples': 14055936, 'steps': 27452, 'loss/train': 3.0902316570281982} 02/25/2022 13:09:57 - INFO - codeparrot_training - Step 27453: {'lr': 0.00022626116223285628, 'samples': 14056448, 'steps': 27453, 'loss/train': 0.8027417063713074} 02/25/2022 13:10:03 - INFO - codeparrot_training - Step 27454: {'lr': 0.00022624487375522888, 'samples': 14056960, 'steps': 27454, 'loss/train': 0.4505730867385864} 02/25/2022 13:10:06 - INFO - codeparrot_training - Step 27455: {'lr': 0.00022622858537936093, 'samples': 14057472, 'steps': 27455, 'loss/train': 2.1103124618530273} 02/25/2022 13:10:12 - INFO - codeparrot_training - Step 27456: {'lr': 0.00022621229710532208, 'samples': 14057984, 'steps': 27456, 'loss/train': 1.5171276330947876} 02/25/2022 13:10:16 - INFO - codeparrot_training - Step 27457: {'lr': 0.00022619600893318228, 'samples': 14058496, 'steps': 27457, 'loss/train': 2.027768850326538} 02/25/2022 13:10:21 - INFO - codeparrot_training - Step 27458: {'lr': 0.00022617972086301117, 'samples': 14059008, 'steps': 27458, 'loss/train': 2.2001190185546875} 02/25/2022 13:10:25 - INFO - codeparrot_training - Step 27459: {'lr': 0.00022616343289487862, 'samples': 14059520, 'steps': 27459, 'loss/train': 1.8636726140975952} 02/25/2022 13:10:30 - INFO - codeparrot_training - Step 27460: {'lr': 0.00022614714502885435, 'samples': 14060032, 'steps': 27460, 'loss/train': 1.8095732927322388} 02/25/2022 13:10:34 - INFO - codeparrot_training - Step 27461: {'lr': 0.0002261308572650081, 'samples': 14060544, 'steps': 27461, 'loss/train': 2.16621994972229} 02/25/2022 13:10:39 - INFO - codeparrot_training - Step 27462: {'lr': 0.00022611456960340968, 'samples': 14061056, 'steps': 27462, 'loss/train': 1.7925037145614624} 02/25/2022 13:10:43 - INFO - codeparrot_training - Step 27463: {'lr': 0.0002260982820441289, 'samples': 14061568, 'steps': 27463, 'loss/train': 1.3859803676605225} 02/25/2022 13:10:48 - INFO - codeparrot_training - Step 27464: {'lr': 0.0002260819945872355, 'samples': 14062080, 'steps': 27464, 'loss/train': 1.592763900756836} 02/25/2022 13:10:52 - INFO - codeparrot_training - Step 27465: {'lr': 0.0002260657072327992, 'samples': 14062592, 'steps': 27465, 'loss/train': 1.6806966066360474} 02/25/2022 13:10:58 - INFO - codeparrot_training - Step 27466: {'lr': 0.00022604941998088978, 'samples': 14063104, 'steps': 27466, 'loss/train': 2.4284791946411133} 02/25/2022 13:11:01 - INFO - codeparrot_training - Step 27467: {'lr': 0.00022603313283157703, 'samples': 14063616, 'steps': 27467, 'loss/train': 0.4122775197029114} 02/25/2022 13:11:07 - INFO - codeparrot_training - Step 27468: {'lr': 0.00022601684578493083, 'samples': 14064128, 'steps': 27468, 'loss/train': 2.1889266967773438} 02/25/2022 13:11:10 - INFO - codeparrot_training - Step 27469: {'lr': 0.00022600055884102079, 'samples': 14064640, 'steps': 27469, 'loss/train': 1.9168899059295654} 02/25/2022 13:11:16 - INFO - codeparrot_training - Step 27470: {'lr': 0.0002259842719999167, 'samples': 14065152, 'steps': 27470, 'loss/train': 2.676759719848633} 02/25/2022 13:11:19 - INFO - codeparrot_training - Step 27471: {'lr': 0.00022596798526168838, 'samples': 14065664, 'steps': 27471, 'loss/train': 2.6157772541046143} 02/25/2022 13:11:25 - INFO - codeparrot_training - Step 27472: {'lr': 0.0002259516986264057, 'samples': 14066176, 'steps': 27472, 'loss/train': 2.164128065109253} 02/25/2022 13:11:28 - INFO - codeparrot_training - Step 27473: {'lr': 0.00022593541209413814, 'samples': 14066688, 'steps': 27473, 'loss/train': 3.1501500606536865} 02/25/2022 13:11:34 - INFO - codeparrot_training - Step 27474: {'lr': 0.0002259191256649557, 'samples': 14067200, 'steps': 27474, 'loss/train': 0.9105191230773926} 02/25/2022 13:11:37 - INFO - codeparrot_training - Step 27475: {'lr': 0.00022590283933892805, 'samples': 14067712, 'steps': 27475, 'loss/train': 1.7122735977172852} 02/25/2022 13:11:43 - INFO - codeparrot_training - Step 27476: {'lr': 0.00022588655311612496, 'samples': 14068224, 'steps': 27476, 'loss/train': 1.426735520362854} 02/25/2022 13:11:46 - INFO - codeparrot_training - Step 27477: {'lr': 0.00022587026699661636, 'samples': 14068736, 'steps': 27477, 'loss/train': 1.4208389520645142} 02/25/2022 13:11:53 - INFO - codeparrot_training - Step 27478: {'lr': 0.00022585398098047177, 'samples': 14069248, 'steps': 27478, 'loss/train': 1.742519497871399} 02/25/2022 13:11:56 - INFO - codeparrot_training - Step 27479: {'lr': 0.00022583769506776105, 'samples': 14069760, 'steps': 27479, 'loss/train': 1.02162766456604} 02/25/2022 13:12:02 - INFO - codeparrot_training - Step 27480: {'lr': 0.00022582140925855396, 'samples': 14070272, 'steps': 27480, 'loss/train': 1.2883697748184204} 02/25/2022 13:12:05 - INFO - codeparrot_training - Step 27481: {'lr': 0.0002258051235529204, 'samples': 14070784, 'steps': 27481, 'loss/train': 1.4234154224395752} 02/25/2022 13:12:11 - INFO - codeparrot_training - Step 27482: {'lr': 0.00022578883795092988, 'samples': 14071296, 'steps': 27482, 'loss/train': 2.000788450241089} 02/25/2022 13:12:14 - INFO - codeparrot_training - Step 27483: {'lr': 0.00022577255245265232, 'samples': 14071808, 'steps': 27483, 'loss/train': 2.4077391624450684} 02/25/2022 13:12:20 - INFO - codeparrot_training - Step 27484: {'lr': 0.00022575626705815743, 'samples': 14072320, 'steps': 27484, 'loss/train': 2.5601675510406494} 02/25/2022 13:12:23 - INFO - codeparrot_training - Step 27485: {'lr': 0.0002257399817675151, 'samples': 14072832, 'steps': 27485, 'loss/train': 2.8967151641845703} 02/25/2022 13:12:29 - INFO - codeparrot_training - Step 27486: {'lr': 0.00022572369658079488, 'samples': 14073344, 'steps': 27486, 'loss/train': 1.1514675617218018} 02/25/2022 13:12:32 - INFO - codeparrot_training - Step 27487: {'lr': 0.00022570741149806665, 'samples': 14073856, 'steps': 27487, 'loss/train': 0.5491908192634583} 02/25/2022 13:12:38 - INFO - codeparrot_training - Step 27488: {'lr': 0.00022569112651940016, 'samples': 14074368, 'steps': 27488, 'loss/train': 2.202266216278076} 02/25/2022 13:12:42 - INFO - codeparrot_training - Step 27489: {'lr': 0.00022567484164486514, 'samples': 14074880, 'steps': 27489, 'loss/train': 1.4990229606628418} 02/25/2022 13:12:47 - INFO - codeparrot_training - Step 27490: {'lr': 0.0002256585568745315, 'samples': 14075392, 'steps': 27490, 'loss/train': 1.4296541213989258} 02/25/2022 13:12:51 - INFO - codeparrot_training - Step 27491: {'lr': 0.00022564227220846876, 'samples': 14075904, 'steps': 27491, 'loss/train': 1.1576011180877686} 02/25/2022 13:12:56 - INFO - codeparrot_training - Step 27492: {'lr': 0.00022562598764674677, 'samples': 14076416, 'steps': 27492, 'loss/train': 1.212106466293335} 02/25/2022 13:13:00 - INFO - codeparrot_training - Step 27493: {'lr': 0.00022560970318943538, 'samples': 14076928, 'steps': 27493, 'loss/train': 1.497833013534546} 02/25/2022 13:13:05 - INFO - codeparrot_training - Step 27494: {'lr': 0.00022559341883660427, 'samples': 14077440, 'steps': 27494, 'loss/train': 0.039732251316308975} 02/25/2022 13:13:09 - INFO - codeparrot_training - Step 27495: {'lr': 0.00022557713458832318, 'samples': 14077952, 'steps': 27495, 'loss/train': 1.7172825336456299} 02/25/2022 13:13:14 - INFO - codeparrot_training - Step 27496: {'lr': 0.00022556085044466185, 'samples': 14078464, 'steps': 27496, 'loss/train': 2.050816297531128} 02/25/2022 13:13:18 - INFO - codeparrot_training - Step 27497: {'lr': 0.00022554456640569017, 'samples': 14078976, 'steps': 27497, 'loss/train': 1.374283790588379} 02/25/2022 13:13:24 - INFO - codeparrot_training - Step 27498: {'lr': 0.00022552828247147778, 'samples': 14079488, 'steps': 27498, 'loss/train': 0.40927544236183167} 02/25/2022 13:13:27 - INFO - codeparrot_training - Step 27499: {'lr': 0.0002255119986420944, 'samples': 14080000, 'steps': 27499, 'loss/train': 1.6877408027648926} 02/25/2022 13:13:33 - INFO - codeparrot_training - Step 27500: {'lr': 0.00022549571491760985, 'samples': 14080512, 'steps': 27500, 'loss/train': 2.191965341567993} 02/25/2022 13:13:36 - INFO - codeparrot_training - Step 27501: {'lr': 0.00022547943129809392, 'samples': 14081024, 'steps': 27501, 'loss/train': 1.6670523881912231} 02/25/2022 13:13:42 - INFO - codeparrot_training - Step 27502: {'lr': 0.00022546314778361626, 'samples': 14081536, 'steps': 27502, 'loss/train': 2.7569758892059326} 02/25/2022 13:13:45 - INFO - codeparrot_training - Step 27503: {'lr': 0.00022544686437424676, 'samples': 14082048, 'steps': 27503, 'loss/train': 2.689985752105713} 02/25/2022 13:13:51 - INFO - codeparrot_training - Step 27504: {'lr': 0.0002254305810700551, 'samples': 14082560, 'steps': 27504, 'loss/train': 2.3572473526000977} 02/25/2022 13:13:54 - INFO - codeparrot_training - Step 27505: {'lr': 0.00022541429787111095, 'samples': 14083072, 'steps': 27505, 'loss/train': 1.786815881729126} 02/25/2022 13:14:00 - INFO - codeparrot_training - Step 27506: {'lr': 0.00022539801477748414, 'samples': 14083584, 'steps': 27506, 'loss/train': 2.169814348220825} 02/25/2022 13:14:03 - INFO - codeparrot_training - Step 27507: {'lr': 0.00022538173178924452, 'samples': 14084096, 'steps': 27507, 'loss/train': 2.227525234222412} 02/25/2022 13:14:09 - INFO - codeparrot_training - Step 27508: {'lr': 0.00022536544890646172, 'samples': 14084608, 'steps': 27508, 'loss/train': 1.5764561891555786} 02/25/2022 13:14:12 - INFO - codeparrot_training - Step 27509: {'lr': 0.00022534916612920543, 'samples': 14085120, 'steps': 27509, 'loss/train': 2.4514665603637695} 02/25/2022 13:14:18 - INFO - codeparrot_training - Step 27510: {'lr': 0.00022533288345754553, 'samples': 14085632, 'steps': 27510, 'loss/train': 0.28191226720809937} 02/25/2022 13:14:22 - INFO - codeparrot_training - Step 27511: {'lr': 0.00022531660089155178, 'samples': 14086144, 'steps': 27511, 'loss/train': 1.7199852466583252} 02/25/2022 13:14:27 - INFO - codeparrot_training - Step 27512: {'lr': 0.0002253003184312938, 'samples': 14086656, 'steps': 27512, 'loss/train': 2.033536195755005} 02/25/2022 13:14:31 - INFO - codeparrot_training - Step 27513: {'lr': 0.00022528403607684143, 'samples': 14087168, 'steps': 27513, 'loss/train': 2.2799434661865234} 02/25/2022 13:14:37 - INFO - codeparrot_training - Step 27514: {'lr': 0.00022526775382826437, 'samples': 14087680, 'steps': 27514, 'loss/train': 1.563528299331665} 02/25/2022 13:14:41 - INFO - codeparrot_training - Step 27515: {'lr': 0.0002252514716856324, 'samples': 14088192, 'steps': 27515, 'loss/train': 2.22799015045166} 02/25/2022 13:14:46 - INFO - codeparrot_training - Step 27516: {'lr': 0.0002252351896490154, 'samples': 14088704, 'steps': 27516, 'loss/train': 1.7148692607879639} 02/25/2022 13:14:50 - INFO - codeparrot_training - Step 27517: {'lr': 0.00022521890771848286, 'samples': 14089216, 'steps': 27517, 'loss/train': 1.7799091339111328} 02/25/2022 13:14:55 - INFO - codeparrot_training - Step 27518: {'lr': 0.00022520262589410464, 'samples': 14089728, 'steps': 27518, 'loss/train': 1.9172215461730957} 02/25/2022 13:14:59 - INFO - codeparrot_training - Step 27519: {'lr': 0.0002251863441759505, 'samples': 14090240, 'steps': 27519, 'loss/train': 1.4067130088806152} 02/25/2022 13:15:04 - INFO - codeparrot_training - Step 27520: {'lr': 0.00022517006256409032, 'samples': 14090752, 'steps': 27520, 'loss/train': 2.225858449935913} 02/25/2022 13:15:08 - INFO - codeparrot_training - Step 27521: {'lr': 0.00022515378105859358, 'samples': 14091264, 'steps': 27521, 'loss/train': 1.9427075386047363} 02/25/2022 13:15:14 - INFO - codeparrot_training - Step 27522: {'lr': 0.00022513749965953015, 'samples': 14091776, 'steps': 27522, 'loss/train': 1.470167636871338} 02/25/2022 13:15:17 - INFO - codeparrot_training - Step 27523: {'lr': 0.00022512121836696977, 'samples': 14092288, 'steps': 27523, 'loss/train': 2.064854621887207} 02/25/2022 13:15:23 - INFO - codeparrot_training - Step 27524: {'lr': 0.0002251049371809823, 'samples': 14092800, 'steps': 27524, 'loss/train': 0.22656604647636414} 02/25/2022 13:15:27 - INFO - codeparrot_training - Step 27525: {'lr': 0.0002250886561016373, 'samples': 14093312, 'steps': 27525, 'loss/train': 1.4379996061325073} 02/25/2022 13:15:32 - INFO - codeparrot_training - Step 27526: {'lr': 0.00022507237512900454, 'samples': 14093824, 'steps': 27526, 'loss/train': 1.5008889436721802} 02/25/2022 13:15:36 - INFO - codeparrot_training - Step 27527: {'lr': 0.00022505609426315382, 'samples': 14094336, 'steps': 27527, 'loss/train': 1.8783501386642456} 02/25/2022 13:15:41 - INFO - codeparrot_training - Step 27528: {'lr': 0.0002250398135041549, 'samples': 14094848, 'steps': 27528, 'loss/train': 1.4466270208358765} 02/25/2022 13:15:45 - INFO - codeparrot_training - Step 27529: {'lr': 0.00022502353285207757, 'samples': 14095360, 'steps': 27529, 'loss/train': 2.2592251300811768} 02/25/2022 13:15:50 - INFO - codeparrot_training - Step 27530: {'lr': 0.0002250072523069914, 'samples': 14095872, 'steps': 27530, 'loss/train': 1.8453835248947144} 02/25/2022 13:15:54 - INFO - codeparrot_training - Step 27531: {'lr': 0.00022499097186896622, 'samples': 14096384, 'steps': 27531, 'loss/train': 1.4378182888031006} 02/25/2022 13:16:00 - INFO - codeparrot_training - Step 27532: {'lr': 0.00022497469153807175, 'samples': 14096896, 'steps': 27532, 'loss/train': 2.338637590408325} 02/25/2022 13:16:04 - INFO - codeparrot_training - Step 27533: {'lr': 0.0002249584113143779, 'samples': 14097408, 'steps': 27533, 'loss/train': 3.065415143966675} 02/25/2022 13:16:09 - INFO - codeparrot_training - Step 27534: {'lr': 0.00022494213119795414, 'samples': 14097920, 'steps': 27534, 'loss/train': 1.4547630548477173} 02/25/2022 13:16:13 - INFO - codeparrot_training - Step 27535: {'lr': 0.00022492585118887034, 'samples': 14098432, 'steps': 27535, 'loss/train': 1.37712562084198} 02/25/2022 13:16:18 - INFO - codeparrot_training - Step 27536: {'lr': 0.00022490957128719626, 'samples': 14098944, 'steps': 27536, 'loss/train': 2.382920980453491} 02/25/2022 13:16:22 - INFO - codeparrot_training - Step 27537: {'lr': 0.00022489329149300163, 'samples': 14099456, 'steps': 27537, 'loss/train': 2.580874443054199} 02/25/2022 13:16:27 - INFO - codeparrot_training - Step 27538: {'lr': 0.00022487701180635617, 'samples': 14099968, 'steps': 27538, 'loss/train': 1.6556426286697388} 02/25/2022 13:16:31 - INFO - codeparrot_training - Step 27539: {'lr': 0.00022486073222732956, 'samples': 14100480, 'steps': 27539, 'loss/train': 1.8759794235229492} 02/25/2022 13:16:36 - INFO - codeparrot_training - Step 27540: {'lr': 0.00022484445275599158, 'samples': 14100992, 'steps': 27540, 'loss/train': 3.4298949241638184} 02/25/2022 13:16:40 - INFO - codeparrot_training - Step 27541: {'lr': 0.00022482817339241208, 'samples': 14101504, 'steps': 27541, 'loss/train': 2.441953659057617} 02/25/2022 13:16:45 - INFO - codeparrot_training - Step 27542: {'lr': 0.00022481189413666065, 'samples': 14102016, 'steps': 27542, 'loss/train': 2.5302250385284424} 02/25/2022 13:16:49 - INFO - codeparrot_training - Step 27543: {'lr': 0.00022479561498880702, 'samples': 14102528, 'steps': 27543, 'loss/train': 2.505784034729004} 02/25/2022 13:16:54 - INFO - codeparrot_training - Step 27544: {'lr': 0.000224779335948921, 'samples': 14103040, 'steps': 27544, 'loss/train': 2.3544015884399414} 02/25/2022 13:16:58 - INFO - codeparrot_training - Step 27545: {'lr': 0.00022476305701707227, 'samples': 14103552, 'steps': 27545, 'loss/train': 1.5262361764907837} 02/25/2022 13:17:03 - INFO - codeparrot_training - Step 27546: {'lr': 0.00022474677819333064, 'samples': 14104064, 'steps': 27546, 'loss/train': 2.424870729446411} 02/25/2022 13:17:07 - INFO - codeparrot_training - Step 27547: {'lr': 0.00022473049947776576, 'samples': 14104576, 'steps': 27547, 'loss/train': 1.2375322580337524} 02/25/2022 13:17:13 - INFO - codeparrot_training - Step 27548: {'lr': 0.0002247142208704474, 'samples': 14105088, 'steps': 27548, 'loss/train': 2.844965934753418} 02/25/2022 13:17:16 - INFO - codeparrot_training - Step 27549: {'lr': 0.00022469794237144528, 'samples': 14105600, 'steps': 27549, 'loss/train': 2.602905750274658} 02/25/2022 13:17:22 - INFO - codeparrot_training - Step 27550: {'lr': 0.00022468166398082913, 'samples': 14106112, 'steps': 27550, 'loss/train': 0.5203708410263062} 02/25/2022 13:17:25 - INFO - codeparrot_training - Step 27551: {'lr': 0.00022466538569866878, 'samples': 14106624, 'steps': 27551, 'loss/train': 0.8628764748573303} 02/25/2022 13:17:31 - INFO - codeparrot_training - Step 27552: {'lr': 0.00022464910752503382, 'samples': 14107136, 'steps': 27552, 'loss/train': 2.5456087589263916} 02/25/2022 13:17:34 - INFO - codeparrot_training - Step 27553: {'lr': 0.00022463282945999396, 'samples': 14107648, 'steps': 27553, 'loss/train': 2.410552501678467} 02/25/2022 13:17:40 - INFO - codeparrot_training - Step 27554: {'lr': 0.00022461655150361908, 'samples': 14108160, 'steps': 27554, 'loss/train': 0.7931026220321655} 02/25/2022 13:17:44 - INFO - codeparrot_training - Step 27555: {'lr': 0.00022460027365597888, 'samples': 14108672, 'steps': 27555, 'loss/train': 0.6397621035575867} 02/25/2022 13:17:49 - INFO - codeparrot_training - Step 27556: {'lr': 0.00022458399591714296, 'samples': 14109184, 'steps': 27556, 'loss/train': 1.2442971467971802} 02/25/2022 13:17:52 - INFO - codeparrot_training - Step 27557: {'lr': 0.00022456771828718112, 'samples': 14109696, 'steps': 27557, 'loss/train': 1.470343828201294} 02/25/2022 13:17:58 - INFO - codeparrot_training - Step 27558: {'lr': 0.0002245514407661631, 'samples': 14110208, 'steps': 27558, 'loss/train': 2.870821475982666} 02/25/2022 13:18:01 - INFO - codeparrot_training - Step 27559: {'lr': 0.00022453516335415875, 'samples': 14110720, 'steps': 27559, 'loss/train': 2.0164613723754883} 02/25/2022 13:18:08 - INFO - codeparrot_training - Step 27560: {'lr': 0.00022451888605123756, 'samples': 14111232, 'steps': 27560, 'loss/train': 1.6098765134811401} 02/25/2022 13:18:11 - INFO - codeparrot_training - Step 27561: {'lr': 0.00022450260885746934, 'samples': 14111744, 'steps': 27561, 'loss/train': 2.1574995517730713} 02/25/2022 13:18:17 - INFO - codeparrot_training - Step 27562: {'lr': 0.0002244863317729239, 'samples': 14112256, 'steps': 27562, 'loss/train': 0.12546542286872864} 02/25/2022 13:18:20 - INFO - codeparrot_training - Step 27563: {'lr': 0.00022447005479767087, 'samples': 14112768, 'steps': 27563, 'loss/train': 2.8106637001037598} 02/25/2022 13:18:26 - INFO - codeparrot_training - Step 27564: {'lr': 0.00022445377793178014, 'samples': 14113280, 'steps': 27564, 'loss/train': 1.8404120206832886} 02/25/2022 13:18:29 - INFO - codeparrot_training - Step 27565: {'lr': 0.0002244375011753212, 'samples': 14113792, 'steps': 27565, 'loss/train': 2.5010669231414795} 02/25/2022 13:18:35 - INFO - codeparrot_training - Step 27566: {'lr': 0.0002244212245283639, 'samples': 14114304, 'steps': 27566, 'loss/train': 2.2299082279205322} 02/25/2022 13:18:38 - INFO - codeparrot_training - Step 27567: {'lr': 0.00022440494799097797, 'samples': 14114816, 'steps': 27567, 'loss/train': 1.8206995725631714} 02/25/2022 13:18:45 - INFO - codeparrot_training - Step 27568: {'lr': 0.0002243886715632332, 'samples': 14115328, 'steps': 27568, 'loss/train': 2.437713146209717} 02/25/2022 13:18:48 - INFO - codeparrot_training - Step 27569: {'lr': 0.00022437239524519912, 'samples': 14115840, 'steps': 27569, 'loss/train': 2.008140802383423} 02/25/2022 13:18:54 - INFO - codeparrot_training - Step 27570: {'lr': 0.00022435611903694555, 'samples': 14116352, 'steps': 27570, 'loss/train': 1.4823436737060547} 02/25/2022 13:18:57 - INFO - codeparrot_training - Step 27571: {'lr': 0.00022433984293854226, 'samples': 14116864, 'steps': 27571, 'loss/train': 1.2989506721496582} 02/25/2022 13:19:03 - INFO - codeparrot_training - Step 27572: {'lr': 0.00022432356695005902, 'samples': 14117376, 'steps': 27572, 'loss/train': 1.6583824157714844} 02/25/2022 13:19:08 - INFO - codeparrot_training - Step 27573: {'lr': 0.00022430729107156532, 'samples': 14117888, 'steps': 27573, 'loss/train': 2.2382700443267822} 02/25/2022 13:19:12 - INFO - codeparrot_training - Step 27574: {'lr': 0.0002242910153031311, 'samples': 14118400, 'steps': 27574, 'loss/train': 1.2156875133514404} 02/25/2022 13:19:17 - INFO - codeparrot_training - Step 27575: {'lr': 0.00022427473964482597, 'samples': 14118912, 'steps': 27575, 'loss/train': 2.770169496536255} 02/25/2022 13:19:21 - INFO - codeparrot_training - Step 27576: {'lr': 0.00022425846409671968, 'samples': 14119424, 'steps': 27576, 'loss/train': 2.0766983032226562} 02/25/2022 13:19:28 - INFO - codeparrot_training - Step 27577: {'lr': 0.00022424218865888207, 'samples': 14119936, 'steps': 27577, 'loss/train': 2.6259078979492188} 02/25/2022 13:19:31 - INFO - codeparrot_training - Step 27578: {'lr': 0.00022422591333138265, 'samples': 14120448, 'steps': 27578, 'loss/train': 0.47769051790237427} 02/25/2022 13:19:35 - INFO - codeparrot_training - Step 27579: {'lr': 0.0002242096381142912, 'samples': 14120960, 'steps': 27579, 'loss/train': 2.3289449214935303} 02/25/2022 13:19:40 - INFO - codeparrot_training - Step 27580: {'lr': 0.00022419336300767752, 'samples': 14121472, 'steps': 27580, 'loss/train': 1.2013404369354248} 02/25/2022 13:19:44 - INFO - codeparrot_training - Step 27581: {'lr': 0.00022417708801161136, 'samples': 14121984, 'steps': 27581, 'loss/train': 1.5715900659561157} 02/25/2022 13:19:49 - INFO - codeparrot_training - Step 27582: {'lr': 0.00022416081312616224, 'samples': 14122496, 'steps': 27582, 'loss/train': 2.195190668106079} 02/25/2022 13:19:53 - INFO - codeparrot_training - Step 27583: {'lr': 0.0002241445383514, 'samples': 14123008, 'steps': 27583, 'loss/train': 0.9663413763046265} 02/25/2022 13:19:58 - INFO - codeparrot_training - Step 27584: {'lr': 0.00022412826368739438, 'samples': 14123520, 'steps': 27584, 'loss/train': 1.4118077754974365} 02/25/2022 13:20:02 - INFO - codeparrot_training - Step 27585: {'lr': 0.00022411198913421506, 'samples': 14124032, 'steps': 27585, 'loss/train': 1.2280981540679932} 02/25/2022 13:20:07 - INFO - codeparrot_training - Step 27586: {'lr': 0.00022409571469193178, 'samples': 14124544, 'steps': 27586, 'loss/train': 1.1711260080337524} 02/25/2022 13:20:11 - INFO - codeparrot_training - Step 27587: {'lr': 0.00022407944036061418, 'samples': 14125056, 'steps': 27587, 'loss/train': 1.842634677886963} 02/25/2022 13:20:16 - INFO - codeparrot_training - Step 27588: {'lr': 0.00022406316614033205, 'samples': 14125568, 'steps': 27588, 'loss/train': 2.152214288711548} 02/25/2022 13:20:20 - INFO - codeparrot_training - Step 27589: {'lr': 0.0002240468920311551, 'samples': 14126080, 'steps': 27589, 'loss/train': 1.5584194660186768} 02/25/2022 13:20:25 - INFO - codeparrot_training - Step 27590: {'lr': 0.000224030618033153, 'samples': 14126592, 'steps': 27590, 'loss/train': 2.6940431594848633} 02/25/2022 13:20:29 - INFO - codeparrot_training - Step 27591: {'lr': 0.00022401434414639552, 'samples': 14127104, 'steps': 27591, 'loss/train': 2.9364380836486816} 02/25/2022 13:20:34 - INFO - codeparrot_training - Step 27592: {'lr': 0.0002239980703709523, 'samples': 14127616, 'steps': 27592, 'loss/train': 1.4578418731689453} 02/25/2022 13:20:38 - INFO - codeparrot_training - Step 27593: {'lr': 0.00022398179670689305, 'samples': 14128128, 'steps': 27593, 'loss/train': 1.9327623844146729} 02/25/2022 13:20:44 - INFO - codeparrot_training - Step 27594: {'lr': 0.00022396552315428762, 'samples': 14128640, 'steps': 27594, 'loss/train': 0.12504425644874573} 02/25/2022 13:20:47 - INFO - codeparrot_training - Step 27595: {'lr': 0.0002239492497132056, 'samples': 14129152, 'steps': 27595, 'loss/train': 2.0389366149902344} 02/25/2022 13:20:53 - INFO - codeparrot_training - Step 27596: {'lr': 0.00022393297638371667, 'samples': 14129664, 'steps': 27596, 'loss/train': 1.5920017957687378} 02/25/2022 13:20:56 - INFO - codeparrot_training - Step 27597: {'lr': 0.0002239167031658906, 'samples': 14130176, 'steps': 27597, 'loss/train': 2.036208391189575} 02/25/2022 13:21:02 - INFO - codeparrot_training - Step 27598: {'lr': 0.00022390043005979707, 'samples': 14130688, 'steps': 27598, 'loss/train': 1.9205272197723389} 02/25/2022 13:21:07 - INFO - codeparrot_training - Step 27599: {'lr': 0.00022388415706550593, 'samples': 14131200, 'steps': 27599, 'loss/train': 2.547344446182251} 02/25/2022 13:21:11 - INFO - codeparrot_training - Step 27600: {'lr': 0.00022386788418308668, 'samples': 14131712, 'steps': 27600, 'loss/train': 1.4409884214401245} 02/25/2022 13:21:17 - INFO - codeparrot_training - Step 27601: {'lr': 0.0002238516114126091, 'samples': 14132224, 'steps': 27601, 'loss/train': 2.344510793685913} 02/25/2022 13:21:21 - INFO - codeparrot_training - Step 27602: {'lr': 0.0002238353387541429, 'samples': 14132736, 'steps': 27602, 'loss/train': 1.9597134590148926} 02/25/2022 13:21:24 - INFO - codeparrot_training - Step 27603: {'lr': 0.00022381906620775794, 'samples': 14133248, 'steps': 27603, 'loss/train': 2.3054986000061035} 02/25/2022 13:21:30 - INFO - codeparrot_training - Step 27604: {'lr': 0.00022380279377352363, 'samples': 14133760, 'steps': 27604, 'loss/train': 1.4429665803909302} 02/25/2022 13:21:34 - INFO - codeparrot_training - Step 27605: {'lr': 0.0002237865214515099, 'samples': 14134272, 'steps': 27605, 'loss/train': 2.250612258911133} 02/25/2022 13:21:39 - INFO - codeparrot_training - Step 27606: {'lr': 0.00022377024924178632, 'samples': 14134784, 'steps': 27606, 'loss/train': 1.073330044746399} 02/25/2022 13:21:43 - INFO - codeparrot_training - Step 27607: {'lr': 0.00022375397714442281, 'samples': 14135296, 'steps': 27607, 'loss/train': 1.6177748441696167} 02/25/2022 13:21:48 - INFO - codeparrot_training - Step 27608: {'lr': 0.00022373770515948883, 'samples': 14135808, 'steps': 27608, 'loss/train': 2.7134149074554443} 02/25/2022 13:21:52 - INFO - codeparrot_training - Step 27609: {'lr': 0.00022372143328705413, 'samples': 14136320, 'steps': 27609, 'loss/train': 1.0845777988433838} 02/25/2022 13:21:57 - INFO - codeparrot_training - Step 27610: {'lr': 0.0002237051615271885, 'samples': 14136832, 'steps': 27610, 'loss/train': 1.7530558109283447} 02/25/2022 13:22:01 - INFO - codeparrot_training - Step 27611: {'lr': 0.00022368888987996162, 'samples': 14137344, 'steps': 27611, 'loss/train': 2.2539851665496826} 02/25/2022 13:22:06 - INFO - codeparrot_training - Step 27612: {'lr': 0.00022367261834544327, 'samples': 14137856, 'steps': 27612, 'loss/train': 1.0081212520599365} 02/25/2022 13:22:10 - INFO - codeparrot_training - Step 27613: {'lr': 0.00022365634692370296, 'samples': 14138368, 'steps': 27613, 'loss/train': 2.083918809890747} 02/25/2022 13:22:15 - INFO - codeparrot_training - Step 27614: {'lr': 0.0002236400756148105, 'samples': 14138880, 'steps': 27614, 'loss/train': 1.8575726747512817} 02/25/2022 13:22:19 - INFO - codeparrot_training - Step 27615: {'lr': 0.0002236238044188356, 'samples': 14139392, 'steps': 27615, 'loss/train': 2.190290927886963} 02/25/2022 13:22:26 - INFO - codeparrot_training - Step 27616: {'lr': 0.00022360753333584805, 'samples': 14139904, 'steps': 27616, 'loss/train': 1.7230314016342163} 02/25/2022 13:22:29 - INFO - codeparrot_training - Step 27617: {'lr': 0.0002235912623659173, 'samples': 14140416, 'steps': 27617, 'loss/train': 2.0915002822875977} 02/25/2022 13:22:35 - INFO - codeparrot_training - Step 27618: {'lr': 0.00022357499150911324, 'samples': 14140928, 'steps': 27618, 'loss/train': 2.457686185836792} 02/25/2022 13:22:38 - INFO - codeparrot_training - Step 27619: {'lr': 0.0002235587207655055, 'samples': 14141440, 'steps': 27619, 'loss/train': 1.8092637062072754} 02/25/2022 13:22:44 - INFO - codeparrot_training - Step 27620: {'lr': 0.00022354245013516392, 'samples': 14141952, 'steps': 27620, 'loss/train': 2.0804600715637207} 02/25/2022 13:22:47 - INFO - codeparrot_training - Step 27621: {'lr': 0.00022352617961815795, 'samples': 14142464, 'steps': 27621, 'loss/train': 2.0949928760528564} 02/25/2022 13:22:52 - INFO - codeparrot_training - Step 27622: {'lr': 0.00022350990921455747, 'samples': 14142976, 'steps': 27622, 'loss/train': 1.9814836978912354} 02/25/2022 13:22:56 - INFO - codeparrot_training - Step 27623: {'lr': 0.0002234936389244321, 'samples': 14143488, 'steps': 27623, 'loss/train': 1.7035192251205444} 02/25/2022 13:23:01 - INFO - codeparrot_training - Step 27624: {'lr': 0.00022347736874785162, 'samples': 14144000, 'steps': 27624, 'loss/train': 0.5222650766372681} 02/25/2022 13:23:05 - INFO - codeparrot_training - Step 27625: {'lr': 0.00022346109868488567, 'samples': 14144512, 'steps': 27625, 'loss/train': 2.2711825370788574} 02/25/2022 13:23:11 - INFO - codeparrot_training - Step 27626: {'lr': 0.0002234448287356039, 'samples': 14145024, 'steps': 27626, 'loss/train': 2.2342886924743652} 02/25/2022 13:23:15 - INFO - codeparrot_training - Step 27627: {'lr': 0.00022342855890007603, 'samples': 14145536, 'steps': 27627, 'loss/train': 1.1420929431915283} 02/25/2022 13:23:20 - INFO - codeparrot_training - Step 27628: {'lr': 0.00022341228917837185, 'samples': 14146048, 'steps': 27628, 'loss/train': 2.265047073364258} 02/25/2022 13:23:24 - INFO - codeparrot_training - Step 27629: {'lr': 0.00022339601957056097, 'samples': 14146560, 'steps': 27629, 'loss/train': 2.317077398300171} 02/25/2022 13:23:29 - INFO - codeparrot_training - Step 27630: {'lr': 0.00022337975007671304, 'samples': 14147072, 'steps': 27630, 'loss/train': 1.0609493255615234} 02/25/2022 13:23:33 - INFO - codeparrot_training - Step 27631: {'lr': 0.0002233634806968979, 'samples': 14147584, 'steps': 27631, 'loss/train': 1.6585279703140259} 02/25/2022 13:23:38 - INFO - codeparrot_training - Step 27632: {'lr': 0.00022334721143118502, 'samples': 14148096, 'steps': 27632, 'loss/train': 1.9878816604614258} 02/25/2022 13:23:42 - INFO - codeparrot_training - Step 27633: {'lr': 0.00022333094227964436, 'samples': 14148608, 'steps': 27633, 'loss/train': 1.5142663717269897} 02/25/2022 13:23:47 - INFO - codeparrot_training - Step 27634: {'lr': 0.00022331467324234537, 'samples': 14149120, 'steps': 27634, 'loss/train': 0.7418177127838135} 02/25/2022 13:23:51 - INFO - codeparrot_training - Step 27635: {'lr': 0.00022329840431935792, 'samples': 14149632, 'steps': 27635, 'loss/train': 1.6578922271728516} 02/25/2022 13:23:56 - INFO - codeparrot_training - Step 27636: {'lr': 0.00022328213551075154, 'samples': 14150144, 'steps': 27636, 'loss/train': 0.968309760093689} 02/25/2022 13:24:00 - INFO - codeparrot_training - Step 27637: {'lr': 0.00022326586681659607, 'samples': 14150656, 'steps': 27637, 'loss/train': 2.5778255462646484} 02/25/2022 13:24:05 - INFO - codeparrot_training - Step 27638: {'lr': 0.00022324959823696118, 'samples': 14151168, 'steps': 27638, 'loss/train': 0.8112436532974243} 02/25/2022 13:24:09 - INFO - codeparrot_training - Step 27639: {'lr': 0.00022323332977191643, 'samples': 14151680, 'steps': 27639, 'loss/train': 2.3479654788970947} 02/25/2022 13:24:15 - INFO - codeparrot_training - Step 27640: {'lr': 0.00022321706142153163, 'samples': 14152192, 'steps': 27640, 'loss/train': 2.084033250808716} 02/25/2022 13:24:18 - INFO - codeparrot_training - Step 27641: {'lr': 0.00022320079318587639, 'samples': 14152704, 'steps': 27641, 'loss/train': 2.2912707328796387} 02/25/2022 13:24:25 - INFO - codeparrot_training - Step 27642: {'lr': 0.00022318452506502057, 'samples': 14153216, 'steps': 27642, 'loss/train': 1.9373067617416382} 02/25/2022 13:24:28 - INFO - codeparrot_training - Step 27643: {'lr': 0.00022316825705903363, 'samples': 14153728, 'steps': 27643, 'loss/train': 1.187343716621399} 02/25/2022 13:24:34 - INFO - codeparrot_training - Step 27644: {'lr': 0.00022315198916798533, 'samples': 14154240, 'steps': 27644, 'loss/train': 1.7621794939041138} 02/25/2022 13:24:37 - INFO - codeparrot_training - Step 27645: {'lr': 0.0002231357213919454, 'samples': 14154752, 'steps': 27645, 'loss/train': 2.896050214767456} 02/25/2022 13:24:43 - INFO - codeparrot_training - Step 27646: {'lr': 0.0002231194537309835, 'samples': 14155264, 'steps': 27646, 'loss/train': 1.8517706394195557} 02/25/2022 13:24:46 - INFO - codeparrot_training - Step 27647: {'lr': 0.00022310318618516944, 'samples': 14155776, 'steps': 27647, 'loss/train': 1.801209807395935} 02/25/2022 13:24:52 - INFO - codeparrot_training - Step 27648: {'lr': 0.0002230869187545727, 'samples': 14156288, 'steps': 27648, 'loss/train': 1.8718500137329102} 02/25/2022 13:24:55 - INFO - codeparrot_training - Step 27649: {'lr': 0.00022307065143926304, 'samples': 14156800, 'steps': 27649, 'loss/train': 1.6194086074829102} 02/25/2022 13:25:01 - INFO - codeparrot_training - Step 27650: {'lr': 0.00022305438423931017, 'samples': 14157312, 'steps': 27650, 'loss/train': 1.940527319908142} 02/25/2022 13:25:04 - INFO - codeparrot_training - Step 27651: {'lr': 0.00022303811715478384, 'samples': 14157824, 'steps': 27651, 'loss/train': 1.5621339082717896} 02/25/2022 13:25:11 - INFO - codeparrot_training - Step 27652: {'lr': 0.00022302185018575356, 'samples': 14158336, 'steps': 27652, 'loss/train': 1.719569206237793} 02/25/2022 13:25:14 - INFO - codeparrot_training - Step 27653: {'lr': 0.0002230055833322891, 'samples': 14158848, 'steps': 27653, 'loss/train': 1.676282286643982} 02/25/2022 13:25:20 - INFO - codeparrot_training - Step 27654: {'lr': 0.00022298931659446014, 'samples': 14159360, 'steps': 27654, 'loss/train': 1.22589910030365} 02/25/2022 13:25:23 - INFO - codeparrot_training - Step 27655: {'lr': 0.00022297304997233653, 'samples': 14159872, 'steps': 27655, 'loss/train': 2.559494733810425} 02/25/2022 13:25:29 - INFO - codeparrot_training - Step 27656: {'lr': 0.00022295678346598763, 'samples': 14160384, 'steps': 27656, 'loss/train': 2.6130824089050293} 02/25/2022 13:25:32 - INFO - codeparrot_training - Step 27657: {'lr': 0.0002229405170754833, 'samples': 14160896, 'steps': 27657, 'loss/train': 2.796952486038208} 02/25/2022 13:25:38 - INFO - codeparrot_training - Step 27658: {'lr': 0.0002229242508008932, 'samples': 14161408, 'steps': 27658, 'loss/train': 1.19169020652771} 02/25/2022 13:25:41 - INFO - codeparrot_training - Step 27659: {'lr': 0.00022290798464228703, 'samples': 14161920, 'steps': 27659, 'loss/train': 1.6232154369354248} 02/25/2022 13:25:47 - INFO - codeparrot_training - Step 27660: {'lr': 0.00022289171859973456, 'samples': 14162432, 'steps': 27660, 'loss/train': 2.304831027984619} 02/25/2022 13:25:50 - INFO - codeparrot_training - Step 27661: {'lr': 0.00022287545267330524, 'samples': 14162944, 'steps': 27661, 'loss/train': 0.9531778693199158} 02/25/2022 13:25:57 - INFO - codeparrot_training - Step 27662: {'lr': 0.00022285918686306886, 'samples': 14163456, 'steps': 27662, 'loss/train': 2.490593910217285} 02/25/2022 13:26:00 - INFO - codeparrot_training - Step 27663: {'lr': 0.0002228429211690951, 'samples': 14163968, 'steps': 27663, 'loss/train': 1.9293419122695923} 02/25/2022 13:26:05 - INFO - codeparrot_training - Step 27664: {'lr': 0.00022282665559145376, 'samples': 14164480, 'steps': 27664, 'loss/train': 1.6510881185531616} 02/25/2022 13:26:09 - INFO - codeparrot_training - Step 27665: {'lr': 0.0002228103901302143, 'samples': 14164992, 'steps': 27665, 'loss/train': 1.121003270149231} 02/25/2022 13:26:14 - INFO - codeparrot_training - Step 27666: {'lr': 0.0002227941247854465, 'samples': 14165504, 'steps': 27666, 'loss/train': 1.3819140195846558} 02/25/2022 13:26:18 - INFO - codeparrot_training - Step 27667: {'lr': 0.00022277785955722, 'samples': 14166016, 'steps': 27667, 'loss/train': 0.46884042024612427} 02/25/2022 13:26:23 - INFO - codeparrot_training - Step 27668: {'lr': 0.00022276159444560464, 'samples': 14166528, 'steps': 27668, 'loss/train': 2.10331130027771} 02/25/2022 13:26:27 - INFO - codeparrot_training - Step 27669: {'lr': 0.00022274532945066987, 'samples': 14167040, 'steps': 27669, 'loss/train': 1.8764166831970215} 02/25/2022 13:26:32 - INFO - codeparrot_training - Step 27670: {'lr': 0.0002227290645724854, 'samples': 14167552, 'steps': 27670, 'loss/train': 1.1372950077056885} 02/25/2022 13:26:36 - INFO - codeparrot_training - Step 27671: {'lr': 0.00022271279981112105, 'samples': 14168064, 'steps': 27671, 'loss/train': 1.5634915828704834} 02/25/2022 13:26:41 - INFO - codeparrot_training - Step 27672: {'lr': 0.00022269653516664633, 'samples': 14168576, 'steps': 27672, 'loss/train': 1.8037562370300293} 02/25/2022 13:26:45 - INFO - codeparrot_training - Step 27673: {'lr': 0.00022268027063913104, 'samples': 14169088, 'steps': 27673, 'loss/train': 2.0806803703308105} 02/25/2022 13:26:50 - INFO - codeparrot_training - Step 27674: {'lr': 0.00022266400622864474, 'samples': 14169600, 'steps': 27674, 'loss/train': 1.8873063325881958} 02/25/2022 13:26:54 - INFO - codeparrot_training - Step 27675: {'lr': 0.00022264774193525723, 'samples': 14170112, 'steps': 27675, 'loss/train': 0.3434310853481293} 02/25/2022 13:26:59 - INFO - codeparrot_training - Step 27676: {'lr': 0.00022263147775903805, 'samples': 14170624, 'steps': 27676, 'loss/train': 1.2676708698272705} 02/25/2022 13:27:03 - INFO - codeparrot_training - Step 27677: {'lr': 0.00022261521370005698, 'samples': 14171136, 'steps': 27677, 'loss/train': 1.3135557174682617} 02/25/2022 13:27:09 - INFO - codeparrot_training - Step 27678: {'lr': 0.00022259894975838363, 'samples': 14171648, 'steps': 27678, 'loss/train': 2.106416940689087} 02/25/2022 13:27:13 - INFO - codeparrot_training - Step 27679: {'lr': 0.0002225826859340876, 'samples': 14172160, 'steps': 27679, 'loss/train': 2.1538820266723633} 02/25/2022 13:27:18 - INFO - codeparrot_training - Step 27680: {'lr': 0.00022256642222723868, 'samples': 14172672, 'steps': 27680, 'loss/train': 1.9056397676467896} 02/25/2022 13:27:22 - INFO - codeparrot_training - Step 27681: {'lr': 0.00022255015863790656, 'samples': 14173184, 'steps': 27681, 'loss/train': 1.9012720584869385} 02/25/2022 13:27:27 - INFO - codeparrot_training - Step 27682: {'lr': 0.00022253389516616083, 'samples': 14173696, 'steps': 27682, 'loss/train': 1.3327257633209229} 02/25/2022 13:27:31 - INFO - codeparrot_training - Step 27683: {'lr': 0.00022251763181207107, 'samples': 14174208, 'steps': 27683, 'loss/train': 2.3490893840789795} 02/25/2022 13:27:36 - INFO - codeparrot_training - Step 27684: {'lr': 0.0002225013685757071, 'samples': 14174720, 'steps': 27684, 'loss/train': 1.998063325881958} 02/25/2022 13:27:40 - INFO - codeparrot_training - Step 27685: {'lr': 0.00022248510545713851, 'samples': 14175232, 'steps': 27685, 'loss/train': 1.465667724609375} 02/25/2022 13:27:45 - INFO - codeparrot_training - Step 27686: {'lr': 0.00022246884245643512, 'samples': 14175744, 'steps': 27686, 'loss/train': 1.754370927810669} 02/25/2022 13:27:49 - INFO - codeparrot_training - Step 27687: {'lr': 0.00022245257957366634, 'samples': 14176256, 'steps': 27687, 'loss/train': 1.1433238983154297} 02/25/2022 13:27:55 - INFO - codeparrot_training - Step 27688: {'lr': 0.00022243631680890198, 'samples': 14176768, 'steps': 27688, 'loss/train': 1.6080366373062134} 02/25/2022 13:27:59 - INFO - codeparrot_training - Step 27689: {'lr': 0.00022242005416221166, 'samples': 14177280, 'steps': 27689, 'loss/train': 1.9073879718780518} 02/25/2022 13:28:04 - INFO - codeparrot_training - Step 27690: {'lr': 0.00022240379163366523, 'samples': 14177792, 'steps': 27690, 'loss/train': 1.2257037162780762} 02/25/2022 13:28:08 - INFO - codeparrot_training - Step 27691: {'lr': 0.00022238752922333207, 'samples': 14178304, 'steps': 27691, 'loss/train': 1.4254138469696045} 02/25/2022 13:28:13 - INFO - codeparrot_training - Step 27692: {'lr': 0.00022237126693128192, 'samples': 14178816, 'steps': 27692, 'loss/train': 1.561631679534912} 02/25/2022 13:28:17 - INFO - codeparrot_training - Step 27693: {'lr': 0.00022235500475758453, 'samples': 14179328, 'steps': 27693, 'loss/train': 1.6028903722763062} 02/25/2022 13:28:22 - INFO - codeparrot_training - Step 27694: {'lr': 0.0002223387427023095, 'samples': 14179840, 'steps': 27694, 'loss/train': 0.7694383263587952} 02/25/2022 13:28:26 - INFO - codeparrot_training - Step 27695: {'lr': 0.00022232248076552662, 'samples': 14180352, 'steps': 27695, 'loss/train': 1.859541893005371} 02/25/2022 13:28:31 - INFO - codeparrot_training - Step 27696: {'lr': 0.00022230621894730535, 'samples': 14180864, 'steps': 27696, 'loss/train': 2.5287795066833496} 02/25/2022 13:28:35 - INFO - codeparrot_training - Step 27697: {'lr': 0.00022228995724771545, 'samples': 14181376, 'steps': 27697, 'loss/train': 1.9428517818450928} 02/25/2022 13:28:41 - INFO - codeparrot_training - Step 27698: {'lr': 0.00022227369566682657, 'samples': 14181888, 'steps': 27698, 'loss/train': 2.6602895259857178} 02/25/2022 13:28:44 - INFO - codeparrot_training - Step 27699: {'lr': 0.00022225743420470844, 'samples': 14182400, 'steps': 27699, 'loss/train': 4.13417911529541} 02/25/2022 13:28:50 - INFO - codeparrot_training - Step 27700: {'lr': 0.0002222411728614306, 'samples': 14182912, 'steps': 27700, 'loss/train': 2.261253833770752} 02/25/2022 13:28:53 - INFO - codeparrot_training - Step 27701: {'lr': 0.00022222491163706275, 'samples': 14183424, 'steps': 27701, 'loss/train': 1.140095829963684} 02/25/2022 13:28:59 - INFO - codeparrot_training - Step 27702: {'lr': 0.00022220865053167456, 'samples': 14183936, 'steps': 27702, 'loss/train': 2.3190813064575195} 02/25/2022 13:29:02 - INFO - codeparrot_training - Step 27703: {'lr': 0.00022219238954533578, 'samples': 14184448, 'steps': 27703, 'loss/train': 0.9984862208366394} 02/25/2022 13:29:08 - INFO - codeparrot_training - Step 27704: {'lr': 0.0002221761286781159, 'samples': 14184960, 'steps': 27704, 'loss/train': 1.4892514944076538} 02/25/2022 13:29:11 - INFO - codeparrot_training - Step 27705: {'lr': 0.00022215986793008459, 'samples': 14185472, 'steps': 27705, 'loss/train': 2.2885193824768066} 02/25/2022 13:29:17 - INFO - codeparrot_training - Step 27706: {'lr': 0.0002221436073013116, 'samples': 14185984, 'steps': 27706, 'loss/train': 1.7201766967773438} 02/25/2022 13:29:20 - INFO - codeparrot_training - Step 27707: {'lr': 0.00022212734679186651, 'samples': 14186496, 'steps': 27707, 'loss/train': 1.9517145156860352} 02/25/2022 13:29:27 - INFO - codeparrot_training - Step 27708: {'lr': 0.00022211108640181917, 'samples': 14187008, 'steps': 27708, 'loss/train': 2.1709697246551514} 02/25/2022 13:29:30 - INFO - codeparrot_training - Step 27709: {'lr': 0.00022209482613123898, 'samples': 14187520, 'steps': 27709, 'loss/train': 2.9296810626983643} 02/25/2022 13:29:36 - INFO - codeparrot_training - Step 27710: {'lr': 0.0002220785659801957, 'samples': 14188032, 'steps': 27710, 'loss/train': 1.9803681373596191} 02/25/2022 13:29:39 - INFO - codeparrot_training - Step 27711: {'lr': 0.00022206230594875898, 'samples': 14188544, 'steps': 27711, 'loss/train': 1.7493771314620972} 02/25/2022 13:29:45 - INFO - codeparrot_training - Step 27712: {'lr': 0.0002220460460369985, 'samples': 14189056, 'steps': 27712, 'loss/train': 1.181638240814209} 02/25/2022 13:29:48 - INFO - codeparrot_training - Step 27713: {'lr': 0.00022202978624498383, 'samples': 14189568, 'steps': 27713, 'loss/train': 1.7782877683639526} 02/25/2022 13:29:54 - INFO - codeparrot_training - Step 27714: {'lr': 0.00022201352657278466, 'samples': 14190080, 'steps': 27714, 'loss/train': 1.7174371480941772} 02/25/2022 13:29:57 - INFO - codeparrot_training - Step 27715: {'lr': 0.00022199726702047074, 'samples': 14190592, 'steps': 27715, 'loss/train': 1.833799958229065} 02/25/2022 13:30:03 - INFO - codeparrot_training - Step 27716: {'lr': 0.0002219810075881116, 'samples': 14191104, 'steps': 27716, 'loss/train': 0.9955262541770935} 02/25/2022 13:30:06 - INFO - codeparrot_training - Step 27717: {'lr': 0.0002219647482757769, 'samples': 14191616, 'steps': 27717, 'loss/train': 1.6329665184020996} 02/25/2022 13:30:12 - INFO - codeparrot_training - Step 27718: {'lr': 0.00022194848908353634, 'samples': 14192128, 'steps': 27718, 'loss/train': 0.23492471873760223} 02/25/2022 13:30:15 - INFO - codeparrot_training - Step 27719: {'lr': 0.00022193223001145952, 'samples': 14192640, 'steps': 27719, 'loss/train': 1.9844173192977905} 02/25/2022 13:30:21 - INFO - codeparrot_training - Step 27720: {'lr': 0.00022191597105961612, 'samples': 14193152, 'steps': 27720, 'loss/train': 1.7243579626083374} 02/25/2022 13:30:24 - INFO - codeparrot_training - Step 27721: {'lr': 0.00022189971222807582, 'samples': 14193664, 'steps': 27721, 'loss/train': 2.4163992404937744} 02/25/2022 13:30:30 - INFO - codeparrot_training - Step 27722: {'lr': 0.00022188345351690822, 'samples': 14194176, 'steps': 27722, 'loss/train': 1.9439784288406372} 02/25/2022 13:30:33 - INFO - codeparrot_training - Step 27723: {'lr': 0.00022186719492618294, 'samples': 14194688, 'steps': 27723, 'loss/train': 2.3920094966888428} 02/25/2022 13:30:40 - INFO - codeparrot_training - Step 27724: {'lr': 0.00022185093645596965, 'samples': 14195200, 'steps': 27724, 'loss/train': 2.495496988296509} 02/25/2022 13:30:44 - INFO - codeparrot_training - Step 27725: {'lr': 0.0002218346781063381, 'samples': 14195712, 'steps': 27725, 'loss/train': 2.224153995513916} 02/25/2022 13:30:49 - INFO - codeparrot_training - Step 27726: {'lr': 0.0002218184198773578, 'samples': 14196224, 'steps': 27726, 'loss/train': 2.4931423664093018} 02/25/2022 13:30:53 - INFO - codeparrot_training - Step 27727: {'lr': 0.0002218021617690984, 'samples': 14196736, 'steps': 27727, 'loss/train': 2.061262369155884} 02/25/2022 13:30:58 - INFO - codeparrot_training - Step 27728: {'lr': 0.00022178590378162956, 'samples': 14197248, 'steps': 27728, 'loss/train': 1.544169545173645} 02/25/2022 13:31:01 - INFO - codeparrot_training - Step 27729: {'lr': 0.00022176964591502112, 'samples': 14197760, 'steps': 27729, 'loss/train': 2.0824270248413086} 02/25/2022 13:31:07 - INFO - codeparrot_training - Step 27730: {'lr': 0.0002217533881693424, 'samples': 14198272, 'steps': 27730, 'loss/train': 1.380448818206787} 02/25/2022 13:31:11 - INFO - codeparrot_training - Step 27731: {'lr': 0.00022173713054466322, 'samples': 14198784, 'steps': 27731, 'loss/train': 2.2570340633392334} 02/25/2022 13:31:16 - INFO - codeparrot_training - Step 27732: {'lr': 0.00022172087304105317, 'samples': 14199296, 'steps': 27732, 'loss/train': 1.6462903022766113} 02/25/2022 13:31:20 - INFO - codeparrot_training - Step 27733: {'lr': 0.00022170461565858193, 'samples': 14199808, 'steps': 27733, 'loss/train': 1.5064811706542969} 02/25/2022 13:31:25 - INFO - codeparrot_training - Step 27734: {'lr': 0.00022168835839731925, 'samples': 14200320, 'steps': 27734, 'loss/train': 2.2610645294189453} 02/25/2022 13:31:29 - INFO - codeparrot_training - Step 27735: {'lr': 0.00022167210125733454, 'samples': 14200832, 'steps': 27735, 'loss/train': 0.16780051589012146} 02/25/2022 13:31:35 - INFO - codeparrot_training - Step 27736: {'lr': 0.00022165584423869755, 'samples': 14201344, 'steps': 27736, 'loss/train': 2.669257402420044} 02/25/2022 13:31:38 - INFO - codeparrot_training - Step 27737: {'lr': 0.00022163958734147793, 'samples': 14201856, 'steps': 27737, 'loss/train': 2.5278546810150146} 02/25/2022 13:31:44 - INFO - codeparrot_training - Step 27738: {'lr': 0.0002216233305657454, 'samples': 14202368, 'steps': 27738, 'loss/train': 1.9732301235198975} 02/25/2022 13:31:47 - INFO - codeparrot_training - Step 27739: {'lr': 0.00022160707391156943, 'samples': 14202880, 'steps': 27739, 'loss/train': 1.6554172039031982} 02/25/2022 13:31:53 - INFO - codeparrot_training - Step 27740: {'lr': 0.00022159081737901975, 'samples': 14203392, 'steps': 27740, 'loss/train': 0.4357997179031372} 02/25/2022 13:31:56 - INFO - codeparrot_training - Step 27741: {'lr': 0.00022157456096816595, 'samples': 14203904, 'steps': 27741, 'loss/train': 1.1607322692871094} 02/25/2022 13:32:01 - INFO - codeparrot_training - Step 27742: {'lr': 0.00022155830467907774, 'samples': 14204416, 'steps': 27742, 'loss/train': 1.7959980964660645} 02/25/2022 13:32:05 - INFO - codeparrot_training - Step 27743: {'lr': 0.0002215420485118248, 'samples': 14204928, 'steps': 27743, 'loss/train': 1.7010483741760254} 02/25/2022 13:32:11 - INFO - codeparrot_training - Step 27744: {'lr': 0.00022152579246647659, 'samples': 14205440, 'steps': 27744, 'loss/train': 1.758897304534912} 02/25/2022 13:32:14 - INFO - codeparrot_training - Step 27745: {'lr': 0.00022150953654310283, 'samples': 14205952, 'steps': 27745, 'loss/train': 1.5634516477584839} 02/25/2022 13:32:20 - INFO - codeparrot_training - Step 27746: {'lr': 0.0002214932807417732, 'samples': 14206464, 'steps': 27746, 'loss/train': 2.1271111965179443} 02/25/2022 13:32:24 - INFO - codeparrot_training - Step 27747: {'lr': 0.00022147702506255737, 'samples': 14206976, 'steps': 27747, 'loss/train': 1.813204288482666} 02/25/2022 13:32:29 - INFO - codeparrot_training - Step 27748: {'lr': 0.00022146076950552482, 'samples': 14207488, 'steps': 27748, 'loss/train': 2.859630584716797} 02/25/2022 13:32:33 - INFO - codeparrot_training - Step 27749: {'lr': 0.00022144451407074528, 'samples': 14208000, 'steps': 27749, 'loss/train': 2.5048768520355225} 02/25/2022 13:32:38 - INFO - codeparrot_training - Step 27750: {'lr': 0.00022142825875828836, 'samples': 14208512, 'steps': 27750, 'loss/train': 0.8993305563926697} 02/25/2022 13:32:42 - INFO - codeparrot_training - Step 27751: {'lr': 0.00022141200356822383, 'samples': 14209024, 'steps': 27751, 'loss/train': 1.6798063516616821} 02/25/2022 13:32:47 - INFO - codeparrot_training - Step 27752: {'lr': 0.0002213957485006211, 'samples': 14209536, 'steps': 27752, 'loss/train': 0.7667948007583618} 02/25/2022 13:32:51 - INFO - codeparrot_training - Step 27753: {'lr': 0.00022137949355554987, 'samples': 14210048, 'steps': 27753, 'loss/train': 1.3859878778457642} 02/25/2022 13:32:56 - INFO - codeparrot_training - Step 27754: {'lr': 0.0002213632387330798, 'samples': 14210560, 'steps': 27754, 'loss/train': 1.7311384677886963} 02/25/2022 13:33:00 - INFO - codeparrot_training - Step 27755: {'lr': 0.00022134698403328062, 'samples': 14211072, 'steps': 27755, 'loss/train': 2.2139503955841064} 02/25/2022 13:33:06 - INFO - codeparrot_training - Step 27756: {'lr': 0.00022133072945622182, 'samples': 14211584, 'steps': 27756, 'loss/train': 2.4205338954925537} 02/25/2022 13:33:09 - INFO - codeparrot_training - Step 27757: {'lr': 0.00022131447500197305, 'samples': 14212096, 'steps': 27757, 'loss/train': 1.3299930095672607} 02/25/2022 13:33:15 - INFO - codeparrot_training - Step 27758: {'lr': 0.00022129822067060398, 'samples': 14212608, 'steps': 27758, 'loss/train': 1.6065117120742798} 02/25/2022 13:33:18 - INFO - codeparrot_training - Step 27759: {'lr': 0.0002212819664621842, 'samples': 14213120, 'steps': 27759, 'loss/train': 1.617329716682434} 02/25/2022 13:33:24 - INFO - codeparrot_training - Step 27760: {'lr': 0.00022126571237678338, 'samples': 14213632, 'steps': 27760, 'loss/train': 2.0306341648101807} 02/25/2022 13:33:27 - INFO - codeparrot_training - Step 27761: {'lr': 0.0002212494584144711, 'samples': 14214144, 'steps': 27761, 'loss/train': 2.2177698612213135} 02/25/2022 13:33:33 - INFO - codeparrot_training - Step 27762: {'lr': 0.00022123320457531704, 'samples': 14214656, 'steps': 27762, 'loss/train': 2.270902395248413} 02/25/2022 13:33:36 - INFO - codeparrot_training - Step 27763: {'lr': 0.00022121695085939075, 'samples': 14215168, 'steps': 27763, 'loss/train': 1.4836783409118652} 02/25/2022 13:33:42 - INFO - codeparrot_training - Step 27764: {'lr': 0.00022120069726676194, 'samples': 14215680, 'steps': 27764, 'loss/train': 2.291496753692627} 02/25/2022 13:33:45 - INFO - codeparrot_training - Step 27765: {'lr': 0.00022118444379750016, 'samples': 14216192, 'steps': 27765, 'loss/train': 2.231790781021118} 02/25/2022 13:33:51 - INFO - codeparrot_training - Step 27766: {'lr': 0.00022116819045167512, 'samples': 14216704, 'steps': 27766, 'loss/train': 1.5718729496002197} 02/25/2022 13:33:54 - INFO - codeparrot_training - Step 27767: {'lr': 0.00022115193722935638, 'samples': 14217216, 'steps': 27767, 'loss/train': 1.8208876848220825} 02/25/2022 13:34:00 - INFO - codeparrot_training - Step 27768: {'lr': 0.00022113568413061356, 'samples': 14217728, 'steps': 27768, 'loss/train': 1.6542704105377197} 02/25/2022 13:34:03 - INFO - codeparrot_training - Step 27769: {'lr': 0.00022111943115551638, 'samples': 14218240, 'steps': 27769, 'loss/train': 2.653064012527466} 02/25/2022 13:34:09 - INFO - codeparrot_training - Step 27770: {'lr': 0.00022110317830413432, 'samples': 14218752, 'steps': 27770, 'loss/train': 1.756883144378662} 02/25/2022 13:34:12 - INFO - codeparrot_training - Step 27771: {'lr': 0.00022108692557653702, 'samples': 14219264, 'steps': 27771, 'loss/train': 2.680830240249634} 02/25/2022 13:34:19 - INFO - codeparrot_training - Step 27772: {'lr': 0.0002210706729727942, 'samples': 14219776, 'steps': 27772, 'loss/train': 2.6151442527770996} 02/25/2022 13:34:22 - INFO - codeparrot_training - Step 27773: {'lr': 0.00022105442049297552, 'samples': 14220288, 'steps': 27773, 'loss/train': 2.0104522705078125} 02/25/2022 13:34:27 - INFO - codeparrot_training - Step 27774: {'lr': 0.0002210381681371504, 'samples': 14220800, 'steps': 27774, 'loss/train': 1.2885103225708008} 02/25/2022 13:34:31 - INFO - codeparrot_training - Step 27775: {'lr': 0.00022102191590538857, 'samples': 14221312, 'steps': 27775, 'loss/train': 0.9257943630218506} 02/25/2022 13:34:36 - INFO - codeparrot_training - Step 27776: {'lr': 0.00022100566379775965, 'samples': 14221824, 'steps': 27776, 'loss/train': 1.4682072401046753} 02/25/2022 13:34:40 - INFO - codeparrot_training - Step 27777: {'lr': 0.0002209894118143334, 'samples': 14222336, 'steps': 27777, 'loss/train': 1.0441948175430298} 02/25/2022 13:34:45 - INFO - codeparrot_training - Step 27778: {'lr': 0.00022097315995517913, 'samples': 14222848, 'steps': 27778, 'loss/train': 1.6112158298492432} 02/25/2022 13:34:49 - INFO - codeparrot_training - Step 27779: {'lr': 0.00022095690822036666, 'samples': 14223360, 'steps': 27779, 'loss/train': 2.0746285915374756} 02/25/2022 13:34:54 - INFO - codeparrot_training - Step 27780: {'lr': 0.00022094065660996557, 'samples': 14223872, 'steps': 27780, 'loss/train': 1.873259425163269} 02/25/2022 13:34:58 - INFO - codeparrot_training - Step 27781: {'lr': 0.0002209244051240455, 'samples': 14224384, 'steps': 27781, 'loss/train': 1.9643899202346802} 02/25/2022 13:35:04 - INFO - codeparrot_training - Step 27782: {'lr': 0.00022090815376267611, 'samples': 14224896, 'steps': 27782, 'loss/train': 1.788137674331665} 02/25/2022 13:35:09 - INFO - codeparrot_training - Step 27783: {'lr': 0.0002208919025259269, 'samples': 14225408, 'steps': 27783, 'loss/train': 2.2936220169067383} 02/25/2022 13:35:13 - INFO - codeparrot_training - Step 27784: {'lr': 0.00022087565141386747, 'samples': 14225920, 'steps': 27784, 'loss/train': 1.3192137479782104} 02/25/2022 13:35:19 - INFO - codeparrot_training - Step 27785: {'lr': 0.00022085940042656754, 'samples': 14226432, 'steps': 27785, 'loss/train': 1.7573145627975464} 02/25/2022 13:35:22 - INFO - codeparrot_training - Step 27786: {'lr': 0.00022084314956409678, 'samples': 14226944, 'steps': 27786, 'loss/train': 1.472084403038025} 02/25/2022 13:35:25 - INFO - codeparrot_training - Step 27787: {'lr': 0.00022082689882652463, 'samples': 14227456, 'steps': 27787, 'loss/train': 4.177149772644043} 02/25/2022 13:35:31 - INFO - codeparrot_training - Step 27788: {'lr': 0.00022081064821392074, 'samples': 14227968, 'steps': 27788, 'loss/train': 2.5439677238464355} 02/25/2022 13:35:35 - INFO - codeparrot_training - Step 27789: {'lr': 0.0002207943977263548, 'samples': 14228480, 'steps': 27789, 'loss/train': 1.398103952407837} 02/25/2022 13:35:40 - INFO - codeparrot_training - Step 27790: {'lr': 0.00022077814736389648, 'samples': 14228992, 'steps': 27790, 'loss/train': 2.349363327026367} 02/25/2022 13:35:43 - INFO - codeparrot_training - Step 27791: {'lr': 0.00022076189712661522, 'samples': 14229504, 'steps': 27791, 'loss/train': 1.1061385869979858} 02/25/2022 13:35:50 - INFO - codeparrot_training - Step 27792: {'lr': 0.00022074564701458067, 'samples': 14230016, 'steps': 27792, 'loss/train': 1.1201121807098389} 02/25/2022 13:35:53 - INFO - codeparrot_training - Step 27793: {'lr': 0.00022072939702786246, 'samples': 14230528, 'steps': 27793, 'loss/train': 2.7665772438049316} 02/25/2022 13:35:59 - INFO - codeparrot_training - Step 27794: {'lr': 0.00022071314716653028, 'samples': 14231040, 'steps': 27794, 'loss/train': 2.7143783569335938} 02/25/2022 13:36:02 - INFO - codeparrot_training - Step 27795: {'lr': 0.00022069689743065373, 'samples': 14231552, 'steps': 27795, 'loss/train': 1.6863172054290771} 02/25/2022 13:36:08 - INFO - codeparrot_training - Step 27796: {'lr': 0.0002206806478203023, 'samples': 14232064, 'steps': 27796, 'loss/train': 0.559899091720581} 02/25/2022 13:36:12 - INFO - codeparrot_training - Step 27797: {'lr': 0.00022066439833554565, 'samples': 14232576, 'steps': 27797, 'loss/train': 2.693077564239502} 02/25/2022 13:36:17 - INFO - codeparrot_training - Step 27798: {'lr': 0.0002206481489764534, 'samples': 14233088, 'steps': 27798, 'loss/train': 1.8381447792053223} 02/25/2022 13:36:21 - INFO - codeparrot_training - Step 27799: {'lr': 0.00022063189974309524, 'samples': 14233600, 'steps': 27799, 'loss/train': 1.5035371780395508} 02/25/2022 13:36:26 - INFO - codeparrot_training - Step 27800: {'lr': 0.00022061565063554063, 'samples': 14234112, 'steps': 27800, 'loss/train': 0.786113440990448} 02/25/2022 13:36:30 - INFO - codeparrot_training - Step 27801: {'lr': 0.00022059940165385919, 'samples': 14234624, 'steps': 27801, 'loss/train': 2.77750825881958} 02/25/2022 13:36:36 - INFO - codeparrot_training - Step 27802: {'lr': 0.0002205831527981207, 'samples': 14235136, 'steps': 27802, 'loss/train': 1.6873021125793457} 02/25/2022 13:36:39 - INFO - codeparrot_training - Step 27803: {'lr': 0.00022056690406839453, 'samples': 14235648, 'steps': 27803, 'loss/train': 1.3729861974716187} 02/25/2022 13:36:45 - INFO - codeparrot_training - Step 27804: {'lr': 0.00022055065546475048, 'samples': 14236160, 'steps': 27804, 'loss/train': 2.353654146194458} 02/25/2022 13:36:48 - INFO - codeparrot_training - Step 27805: {'lr': 0.000220534406987258, 'samples': 14236672, 'steps': 27805, 'loss/train': 1.5033620595932007} 02/25/2022 13:36:54 - INFO - codeparrot_training - Step 27806: {'lr': 0.00022051815863598684, 'samples': 14237184, 'steps': 27806, 'loss/train': 2.432969331741333} 02/25/2022 13:36:57 - INFO - codeparrot_training - Step 27807: {'lr': 0.00022050191041100647, 'samples': 14237696, 'steps': 27807, 'loss/train': 1.6492234468460083} 02/25/2022 13:37:03 - INFO - codeparrot_training - Step 27808: {'lr': 0.0002204856623123866, 'samples': 14238208, 'steps': 27808, 'loss/train': 1.7908642292022705} 02/25/2022 13:37:06 - INFO - codeparrot_training - Step 27809: {'lr': 0.0002204694143401968, 'samples': 14238720, 'steps': 27809, 'loss/train': 4.232202529907227} 02/25/2022 13:37:12 - INFO - codeparrot_training - Step 27810: {'lr': 0.00022045316649450656, 'samples': 14239232, 'steps': 27810, 'loss/train': 1.6632262468338013} 02/25/2022 13:37:15 - INFO - codeparrot_training - Step 27811: {'lr': 0.00022043691877538557, 'samples': 14239744, 'steps': 27811, 'loss/train': 1.4221631288528442} 02/25/2022 13:37:21 - INFO - codeparrot_training - Step 27812: {'lr': 0.0002204206711829035, 'samples': 14240256, 'steps': 27812, 'loss/train': 1.8015223741531372} 02/25/2022 13:37:24 - INFO - codeparrot_training - Step 27813: {'lr': 0.0002204044237171299, 'samples': 14240768, 'steps': 27813, 'loss/train': 1.09882652759552} 02/25/2022 13:37:30 - INFO - codeparrot_training - Step 27814: {'lr': 0.00022038817637813428, 'samples': 14241280, 'steps': 27814, 'loss/train': 2.1022181510925293} 02/25/2022 13:37:33 - INFO - codeparrot_training - Step 27815: {'lr': 0.00022037192916598633, 'samples': 14241792, 'steps': 27815, 'loss/train': 1.5143153667449951} 02/25/2022 13:37:39 - INFO - codeparrot_training - Step 27816: {'lr': 0.0002203556820807556, 'samples': 14242304, 'steps': 27816, 'loss/train': 1.922299861907959} 02/25/2022 13:37:42 - INFO - codeparrot_training - Step 27817: {'lr': 0.00022033943512251184, 'samples': 14242816, 'steps': 27817, 'loss/train': 2.7210869789123535} 02/25/2022 13:37:48 - INFO - codeparrot_training - Step 27818: {'lr': 0.00022032318829132442, 'samples': 14243328, 'steps': 27818, 'loss/train': 1.6526563167572021} 02/25/2022 13:37:52 - INFO - codeparrot_training - Step 27819: {'lr': 0.00022030694158726302, 'samples': 14243840, 'steps': 27819, 'loss/train': 1.861775279045105} 02/25/2022 13:37:57 - INFO - codeparrot_training - Step 27820: {'lr': 0.00022029069501039726, 'samples': 14244352, 'steps': 27820, 'loss/train': 2.442487955093384} 02/25/2022 13:38:01 - INFO - codeparrot_training - Step 27821: {'lr': 0.00022027444856079684, 'samples': 14244864, 'steps': 27821, 'loss/train': 0.662951648235321} 02/25/2022 13:38:06 - INFO - codeparrot_training - Step 27822: {'lr': 0.00022025820223853113, 'samples': 14245376, 'steps': 27822, 'loss/train': 2.484048366546631} 02/25/2022 13:38:10 - INFO - codeparrot_training - Step 27823: {'lr': 0.00022024195604366983, 'samples': 14245888, 'steps': 27823, 'loss/train': 2.381814956665039} 02/25/2022 13:38:15 - INFO - codeparrot_training - Step 27824: {'lr': 0.00022022570997628254, 'samples': 14246400, 'steps': 27824, 'loss/train': 2.6971611976623535} 02/25/2022 13:38:19 - INFO - codeparrot_training - Step 27825: {'lr': 0.00022020946403643899, 'samples': 14246912, 'steps': 27825, 'loss/train': 2.3923892974853516} 02/25/2022 13:38:24 - INFO - codeparrot_training - Step 27826: {'lr': 0.00022019321822420852, 'samples': 14247424, 'steps': 27826, 'loss/train': 2.2321488857269287} 02/25/2022 13:38:28 - INFO - codeparrot_training - Step 27827: {'lr': 0.00022017697253966082, 'samples': 14247936, 'steps': 27827, 'loss/train': 2.395418882369995} 02/25/2022 13:38:34 - INFO - codeparrot_training - Step 27828: {'lr': 0.00022016072698286551, 'samples': 14248448, 'steps': 27828, 'loss/train': 1.973207712173462} 02/25/2022 13:38:37 - INFO - codeparrot_training - Step 27829: {'lr': 0.00022014448155389217, 'samples': 14248960, 'steps': 27829, 'loss/train': 0.8779698610305786} 02/25/2022 13:38:43 - INFO - codeparrot_training - Step 27830: {'lr': 0.0002201282362528105, 'samples': 14249472, 'steps': 27830, 'loss/train': 1.4527925252914429} 02/25/2022 13:38:46 - INFO - codeparrot_training - Step 27831: {'lr': 0.00022011199107968988, 'samples': 14249984, 'steps': 27831, 'loss/train': 2.2932536602020264} 02/25/2022 13:38:52 - INFO - codeparrot_training - Step 27832: {'lr': 0.00022009574603459997, 'samples': 14250496, 'steps': 27832, 'loss/train': 1.9249000549316406} 02/25/2022 13:38:56 - INFO - codeparrot_training - Step 27833: {'lr': 0.0002200795011176104, 'samples': 14251008, 'steps': 27833, 'loss/train': 1.3450337648391724} 02/25/2022 13:39:01 - INFO - codeparrot_training - Step 27834: {'lr': 0.00022006325632879087, 'samples': 14251520, 'steps': 27834, 'loss/train': 2.4711389541625977} 02/25/2022 13:39:04 - INFO - codeparrot_training - Step 27835: {'lr': 0.00022004701166821074, 'samples': 14252032, 'steps': 27835, 'loss/train': 2.395136833190918} 02/25/2022 13:39:10 - INFO - codeparrot_training - Step 27836: {'lr': 0.0002200307671359397, 'samples': 14252544, 'steps': 27836, 'loss/train': 1.8185944557189941} 02/25/2022 13:39:13 - INFO - codeparrot_training - Step 27837: {'lr': 0.0002200145227320473, 'samples': 14253056, 'steps': 27837, 'loss/train': 1.2070010900497437} 02/25/2022 13:39:20 - INFO - codeparrot_training - Step 27838: {'lr': 0.00021999827845660332, 'samples': 14253568, 'steps': 27838, 'loss/train': 1.2528291940689087} 02/25/2022 13:39:23 - INFO - codeparrot_training - Step 27839: {'lr': 0.00021998203430967706, 'samples': 14254080, 'steps': 27839, 'loss/train': 1.6341338157653809} 02/25/2022 13:39:29 - INFO - codeparrot_training - Step 27840: {'lr': 0.00021996579029133823, 'samples': 14254592, 'steps': 27840, 'loss/train': 2.2483770847320557} 02/25/2022 13:39:32 - INFO - codeparrot_training - Step 27841: {'lr': 0.00021994954640165644, 'samples': 14255104, 'steps': 27841, 'loss/train': 1.8836464881896973} 02/25/2022 13:39:38 - INFO - codeparrot_training - Step 27842: {'lr': 0.00021993330264070127, 'samples': 14255616, 'steps': 27842, 'loss/train': 1.5909324884414673} 02/25/2022 13:39:42 - INFO - codeparrot_training - Step 27843: {'lr': 0.0002199170590085423, 'samples': 14256128, 'steps': 27843, 'loss/train': 2.447415828704834} 02/25/2022 13:39:47 - INFO - codeparrot_training - Step 27844: {'lr': 0.00021990081550524903, 'samples': 14256640, 'steps': 27844, 'loss/train': 0.5877178311347961} 02/25/2022 13:39:51 - INFO - codeparrot_training - Step 27845: {'lr': 0.00021988457213089114, 'samples': 14257152, 'steps': 27845, 'loss/train': 0.4085596799850464} 02/25/2022 13:39:56 - INFO - codeparrot_training - Step 27846: {'lr': 0.0002198683288855382, 'samples': 14257664, 'steps': 27846, 'loss/train': 2.3782947063446045} 02/25/2022 13:40:00 - INFO - codeparrot_training - Step 27847: {'lr': 0.0002198520857692598, 'samples': 14258176, 'steps': 27847, 'loss/train': 1.4745988845825195} 02/25/2022 13:40:06 - INFO - codeparrot_training - Step 27848: {'lr': 0.00021983584278212543, 'samples': 14258688, 'steps': 27848, 'loss/train': 2.190983295440674} 02/25/2022 13:40:10 - INFO - codeparrot_training - Step 27849: {'lr': 0.00021981959992420475, 'samples': 14259200, 'steps': 27849, 'loss/train': 1.711606502532959} 02/25/2022 13:40:15 - INFO - codeparrot_training - Step 27850: {'lr': 0.0002198033571955673, 'samples': 14259712, 'steps': 27850, 'loss/train': 1.4930320978164673} 02/25/2022 13:40:18 - INFO - codeparrot_training - Step 27851: {'lr': 0.0002197871145962827, 'samples': 14260224, 'steps': 27851, 'loss/train': 2.3477847576141357} 02/25/2022 13:40:24 - INFO - codeparrot_training - Step 27852: {'lr': 0.00021977087212642052, 'samples': 14260736, 'steps': 27852, 'loss/train': 1.8510905504226685} 02/25/2022 13:40:27 - INFO - codeparrot_training - Step 27853: {'lr': 0.00021975462978605036, 'samples': 14261248, 'steps': 27853, 'loss/train': 1.3851715326309204} 02/25/2022 13:40:33 - INFO - codeparrot_training - Step 27854: {'lr': 0.00021973838757524168, 'samples': 14261760, 'steps': 27854, 'loss/train': 0.8404471278190613} 02/25/2022 13:40:37 - INFO - codeparrot_training - Step 27855: {'lr': 0.00021972214549406414, 'samples': 14262272, 'steps': 27855, 'loss/train': 1.822780966758728} 02/25/2022 13:40:42 - INFO - codeparrot_training - Step 27856: {'lr': 0.00021970590354258743, 'samples': 14262784, 'steps': 27856, 'loss/train': 1.6116869449615479} 02/25/2022 13:40:46 - INFO - codeparrot_training - Step 27857: {'lr': 0.0002196896617208809, 'samples': 14263296, 'steps': 27857, 'loss/train': 0.8649618029594421} 02/25/2022 13:40:51 - INFO - codeparrot_training - Step 27858: {'lr': 0.00021967342002901424, 'samples': 14263808, 'steps': 27858, 'loss/train': 2.0634584426879883} 02/25/2022 13:40:55 - INFO - codeparrot_training - Step 27859: {'lr': 0.00021965717846705702, 'samples': 14264320, 'steps': 27859, 'loss/train': 1.420082688331604} 02/25/2022 13:41:00 - INFO - codeparrot_training - Step 27860: {'lr': 0.00021964093703507893, 'samples': 14264832, 'steps': 27860, 'loss/train': 1.009238362312317} 02/25/2022 13:41:04 - INFO - codeparrot_training - Step 27861: {'lr': 0.00021962469573314928, 'samples': 14265344, 'steps': 27861, 'loss/train': 2.0793399810791016} 02/25/2022 13:41:09 - INFO - codeparrot_training - Step 27862: {'lr': 0.00021960845456133783, 'samples': 14265856, 'steps': 27862, 'loss/train': 1.3136025667190552} 02/25/2022 13:41:13 - INFO - codeparrot_training - Step 27863: {'lr': 0.0002195922135197141, 'samples': 14266368, 'steps': 27863, 'loss/train': 1.86370050907135} 02/25/2022 13:41:19 - INFO - codeparrot_training - Step 27864: {'lr': 0.00021957597260834763, 'samples': 14266880, 'steps': 27864, 'loss/train': 2.7393109798431396} 02/25/2022 13:41:22 - INFO - codeparrot_training - Step 27865: {'lr': 0.00021955973182730818, 'samples': 14267392, 'steps': 27865, 'loss/train': 1.668273687362671} 02/25/2022 13:41:28 - INFO - codeparrot_training - Step 27866: {'lr': 0.00021954349117666506, 'samples': 14267904, 'steps': 27866, 'loss/train': 1.5442965030670166} 02/25/2022 13:41:31 - INFO - codeparrot_training - Step 27867: {'lr': 0.00021952725065648796, 'samples': 14268416, 'steps': 27867, 'loss/train': 2.2530710697174072} 02/25/2022 13:41:37 - INFO - codeparrot_training - Step 27868: {'lr': 0.00021951101026684643, 'samples': 14268928, 'steps': 27868, 'loss/train': 2.1887032985687256} 02/25/2022 13:41:40 - INFO - codeparrot_training - Step 27869: {'lr': 0.00021949477000781018, 'samples': 14269440, 'steps': 27869, 'loss/train': 2.6383321285247803} 02/25/2022 13:41:46 - INFO - codeparrot_training - Step 27870: {'lr': 0.0002194785298794485, 'samples': 14269952, 'steps': 27870, 'loss/train': 1.1898847818374634} 02/25/2022 13:41:50 - INFO - codeparrot_training - Step 27871: {'lr': 0.00021946228988183115, 'samples': 14270464, 'steps': 27871, 'loss/train': 1.9612748622894287} 02/25/2022 13:41:55 - INFO - codeparrot_training - Step 27872: {'lr': 0.00021944605001502761, 'samples': 14270976, 'steps': 27872, 'loss/train': 0.6058495044708252} 02/25/2022 13:41:59 - INFO - codeparrot_training - Step 27873: {'lr': 0.00021942981027910763, 'samples': 14271488, 'steps': 27873, 'loss/train': 1.4544835090637207} 02/25/2022 13:42:05 - INFO - codeparrot_training - Step 27874: {'lr': 0.00021941357067414052, 'samples': 14272000, 'steps': 27874, 'loss/train': 1.8225550651550293} 02/25/2022 13:42:08 - INFO - codeparrot_training - Step 27875: {'lr': 0.00021939733120019599, 'samples': 14272512, 'steps': 27875, 'loss/train': 1.949411153793335} 02/25/2022 13:42:14 - INFO - codeparrot_training - Step 27876: {'lr': 0.0002193810918573435, 'samples': 14273024, 'steps': 27876, 'loss/train': 1.260650873184204} 02/25/2022 13:42:17 - INFO - codeparrot_training - Step 27877: {'lr': 0.00021936485264565275, 'samples': 14273536, 'steps': 27877, 'loss/train': 2.270949125289917} 02/25/2022 13:42:23 - INFO - codeparrot_training - Step 27878: {'lr': 0.00021934861356519335, 'samples': 14274048, 'steps': 27878, 'loss/train': 2.3504598140716553} 02/25/2022 13:42:26 - INFO - codeparrot_training - Step 27879: {'lr': 0.00021933237461603462, 'samples': 14274560, 'steps': 27879, 'loss/train': 0.7530683875083923} 02/25/2022 13:42:32 - INFO - codeparrot_training - Step 27880: {'lr': 0.00021931613579824626, 'samples': 14275072, 'steps': 27880, 'loss/train': 1.8609378337860107} 02/25/2022 13:42:35 - INFO - codeparrot_training - Step 27881: {'lr': 0.00021929989711189787, 'samples': 14275584, 'steps': 27881, 'loss/train': 1.4496419429779053} 02/25/2022 13:42:41 - INFO - codeparrot_training - Step 27882: {'lr': 0.00021928365855705906, 'samples': 14276096, 'steps': 27882, 'loss/train': 2.198673725128174} 02/25/2022 13:42:44 - INFO - codeparrot_training - Step 27883: {'lr': 0.00021926742013379918, 'samples': 14276608, 'steps': 27883, 'loss/train': 1.1382017135620117} 02/25/2022 13:42:50 - INFO - codeparrot_training - Step 27884: {'lr': 0.00021925118184218793, 'samples': 14277120, 'steps': 27884, 'loss/train': 1.882972240447998} 02/25/2022 13:42:54 - INFO - codeparrot_training - Step 27885: {'lr': 0.00021923494368229486, 'samples': 14277632, 'steps': 27885, 'loss/train': 1.4330956935882568} 02/25/2022 13:42:59 - INFO - codeparrot_training - Step 27886: {'lr': 0.00021921870565418962, 'samples': 14278144, 'steps': 27886, 'loss/train': 1.6983836889266968} 02/25/2022 13:43:03 - INFO - codeparrot_training - Step 27887: {'lr': 0.00021920246775794156, 'samples': 14278656, 'steps': 27887, 'loss/train': 1.799717903137207} 02/25/2022 13:43:08 - INFO - codeparrot_training - Step 27888: {'lr': 0.00021918622999362035, 'samples': 14279168, 'steps': 27888, 'loss/train': 1.973542332649231} 02/25/2022 13:43:12 - INFO - codeparrot_training - Step 27889: {'lr': 0.00021916999236129558, 'samples': 14279680, 'steps': 27889, 'loss/train': 2.15346622467041} 02/25/2022 13:43:17 - INFO - codeparrot_training - Step 27890: {'lr': 0.00021915375486103675, 'samples': 14280192, 'steps': 27890, 'loss/train': 1.8841772079467773} 02/25/2022 13:43:21 - INFO - codeparrot_training - Step 27891: {'lr': 0.00021913751749291346, 'samples': 14280704, 'steps': 27891, 'loss/train': 1.1203455924987793} 02/25/2022 13:43:26 - INFO - codeparrot_training - Step 27892: {'lr': 0.00021912128025699523, 'samples': 14281216, 'steps': 27892, 'loss/train': 0.9009745121002197} 02/25/2022 13:43:30 - INFO - codeparrot_training - Step 27893: {'lr': 0.00021910504315335167, 'samples': 14281728, 'steps': 27893, 'loss/train': 1.8594086170196533} 02/25/2022 13:43:36 - INFO - codeparrot_training - Step 27894: {'lr': 0.00021908880618205223, 'samples': 14282240, 'steps': 27894, 'loss/train': 1.881860613822937} 02/25/2022 13:43:39 - INFO - codeparrot_training - Step 27895: {'lr': 0.0002190725693431666, 'samples': 14282752, 'steps': 27895, 'loss/train': 1.39854896068573} 02/25/2022 13:43:45 - INFO - codeparrot_training - Step 27896: {'lr': 0.00021905633263676424, 'samples': 14283264, 'steps': 27896, 'loss/train': 1.7973921298980713} 02/25/2022 13:43:48 - INFO - codeparrot_training - Step 27897: {'lr': 0.0002190400960629147, 'samples': 14283776, 'steps': 27897, 'loss/train': 2.187462091445923} 02/25/2022 13:43:54 - INFO - codeparrot_training - Step 27898: {'lr': 0.00021902385962168752, 'samples': 14284288, 'steps': 27898, 'loss/train': 2.111696481704712} 02/25/2022 13:43:58 - INFO - codeparrot_training - Step 27899: {'lr': 0.00021900762331315238, 'samples': 14284800, 'steps': 27899, 'loss/train': 1.4585720300674438} 02/25/2022 13:44:03 - INFO - codeparrot_training - Step 27900: {'lr': 0.00021899138713737875, 'samples': 14285312, 'steps': 27900, 'loss/train': 2.317547082901001} 02/25/2022 13:44:07 - INFO - codeparrot_training - Step 27901: {'lr': 0.00021897515109443609, 'samples': 14285824, 'steps': 27901, 'loss/train': 0.6042247414588928} 02/25/2022 13:44:12 - INFO - codeparrot_training - Step 27902: {'lr': 0.00021895891518439403, 'samples': 14286336, 'steps': 27902, 'loss/train': 1.846108078956604} 02/25/2022 13:44:16 - INFO - codeparrot_training - Step 27903: {'lr': 0.00021894267940732211, 'samples': 14286848, 'steps': 27903, 'loss/train': 2.2347066402435303} 02/25/2022 13:44:22 - INFO - codeparrot_training - Step 27904: {'lr': 0.00021892644376329001, 'samples': 14287360, 'steps': 27904, 'loss/train': 1.7606449127197266} 02/25/2022 13:44:25 - INFO - codeparrot_training - Step 27905: {'lr': 0.00021891020825236707, 'samples': 14287872, 'steps': 27905, 'loss/train': 2.2213141918182373} 02/25/2022 13:44:31 - INFO - codeparrot_training - Step 27906: {'lr': 0.0002188939728746229, 'samples': 14288384, 'steps': 27906, 'loss/train': 2.346630811691284} 02/25/2022 13:44:34 - INFO - codeparrot_training - Step 27907: {'lr': 0.0002188777376301271, 'samples': 14288896, 'steps': 27907, 'loss/train': 1.9391924142837524} 02/25/2022 13:44:40 - INFO - codeparrot_training - Step 27908: {'lr': 0.00021886150251894927, 'samples': 14289408, 'steps': 27908, 'loss/train': 2.502089500427246} 02/25/2022 13:44:43 - INFO - codeparrot_training - Step 27909: {'lr': 0.00021884526754115878, 'samples': 14289920, 'steps': 27909, 'loss/train': 0.837447464466095} 02/25/2022 13:44:49 - INFO - codeparrot_training - Step 27910: {'lr': 0.00021882903269682526, 'samples': 14290432, 'steps': 27910, 'loss/train': 1.8735260963439941} 02/25/2022 13:44:52 - INFO - codeparrot_training - Step 27911: {'lr': 0.00021881279798601828, 'samples': 14290944, 'steps': 27911, 'loss/train': 1.3187624216079712} 02/25/2022 13:44:58 - INFO - codeparrot_training - Step 27912: {'lr': 0.00021879656340880734, 'samples': 14291456, 'steps': 27912, 'loss/train': 2.862642526626587} 02/25/2022 13:45:02 - INFO - codeparrot_training - Step 27913: {'lr': 0.00021878032896526216, 'samples': 14291968, 'steps': 27913, 'loss/train': 3.502828598022461} 02/25/2022 13:45:05 - INFO - codeparrot_training - Step 27914: {'lr': 0.00021876409465545202, 'samples': 14292480, 'steps': 27914, 'loss/train': 2.9667201042175293} 02/25/2022 13:45:10 - INFO - codeparrot_training - Step 27915: {'lr': 0.00021874786047944658, 'samples': 14292992, 'steps': 27915, 'loss/train': 1.2065900564193726} 02/25/2022 13:45:14 - INFO - codeparrot_training - Step 27916: {'lr': 0.00021873162643731535, 'samples': 14293504, 'steps': 27916, 'loss/train': 1.3866487741470337} 02/25/2022 13:45:19 - INFO - codeparrot_training - Step 27917: {'lr': 0.00021871539252912807, 'samples': 14294016, 'steps': 27917, 'loss/train': 2.808729410171509} 02/25/2022 13:45:23 - INFO - codeparrot_training - Step 27918: {'lr': 0.00021869915875495397, 'samples': 14294528, 'steps': 27918, 'loss/train': 1.9199239015579224} 02/25/2022 13:45:28 - INFO - codeparrot_training - Step 27919: {'lr': 0.00021868292511486274, 'samples': 14295040, 'steps': 27919, 'loss/train': 1.946401834487915} 02/25/2022 13:45:32 - INFO - codeparrot_training - Step 27920: {'lr': 0.00021866669160892392, 'samples': 14295552, 'steps': 27920, 'loss/train': 1.187985897064209} 02/25/2022 13:45:37 - INFO - codeparrot_training - Step 27921: {'lr': 0.00021865045823720713, 'samples': 14296064, 'steps': 27921, 'loss/train': 2.733861207962036} 02/25/2022 13:45:41 - INFO - codeparrot_training - Step 27922: {'lr': 0.00021863422499978174, 'samples': 14296576, 'steps': 27922, 'loss/train': 2.041826009750366} 02/25/2022 13:45:46 - INFO - codeparrot_training - Step 27923: {'lr': 0.00021861799189671737, 'samples': 14297088, 'steps': 27923, 'loss/train': 1.9650715589523315} 02/25/2022 13:45:49 - INFO - codeparrot_training - Dataset epoch: 1 02/25/2022 13:45:54 - INFO - codeparrot_training - Step 27924: {'lr': 0.00021860175892808353, 'samples': 14297600, 'steps': 27924, 'loss/train': 1.7330572605133057} 02/25/2022 13:45:57 - INFO - codeparrot_training - Step 27925: {'lr': 0.00021858552609394983, 'samples': 14298112, 'steps': 27925, 'loss/train': 2.0157382488250732} 02/25/2022 13:46:01 - INFO - codeparrot_training - Step 27926: {'lr': 0.00021856929339438583, 'samples': 14298624, 'steps': 27926, 'loss/train': 2.117175579071045} 02/25/2022 13:46:06 - INFO - codeparrot_training - Step 27927: {'lr': 0.00021855306082946093, 'samples': 14299136, 'steps': 27927, 'loss/train': 1.6364167928695679} 02/25/2022 13:46:10 - INFO - codeparrot_training - Step 27928: {'lr': 0.00021853682839924468, 'samples': 14299648, 'steps': 27928, 'loss/train': 1.7808254957199097} 02/25/2022 13:46:16 - INFO - codeparrot_training - Step 27929: {'lr': 0.00021852059610380677, 'samples': 14300160, 'steps': 27929, 'loss/train': 1.8157947063446045} 02/25/2022 13:46:20 - INFO - codeparrot_training - Step 27930: {'lr': 0.00021850436394321658, 'samples': 14300672, 'steps': 27930, 'loss/train': 0.9372547268867493} 02/25/2022 13:46:25 - INFO - codeparrot_training - Step 27931: {'lr': 0.00021848813191754365, 'samples': 14301184, 'steps': 27931, 'loss/train': 1.892242670059204} 02/25/2022 13:46:29 - INFO - codeparrot_training - Step 27932: {'lr': 0.00021847190002685757, 'samples': 14301696, 'steps': 27932, 'loss/train': 2.2397842407226562} 02/25/2022 13:46:34 - INFO - codeparrot_training - Step 27933: {'lr': 0.00021845566827122793, 'samples': 14302208, 'steps': 27933, 'loss/train': 0.16233785450458527} 02/25/2022 13:46:38 - INFO - codeparrot_training - Step 27934: {'lr': 0.0002184394366507242, 'samples': 14302720, 'steps': 27934, 'loss/train': 2.0924746990203857} 02/25/2022 13:46:43 - INFO - codeparrot_training - Step 27935: {'lr': 0.0002184232051654158, 'samples': 14303232, 'steps': 27935, 'loss/train': 1.54048752784729} 02/25/2022 13:46:47 - INFO - codeparrot_training - Step 27936: {'lr': 0.00021840697381537245, 'samples': 14303744, 'steps': 27936, 'loss/train': 2.669825792312622} 02/25/2022 13:46:52 - INFO - codeparrot_training - Step 27937: {'lr': 0.0002183907426006635, 'samples': 14304256, 'steps': 27937, 'loss/train': 1.4830563068389893} 02/25/2022 13:46:56 - INFO - codeparrot_training - Step 27938: {'lr': 0.00021837451152135863, 'samples': 14304768, 'steps': 27938, 'loss/train': 1.3057787418365479} 02/25/2022 13:47:01 - INFO - codeparrot_training - Step 27939: {'lr': 0.00021835828057752732, 'samples': 14305280, 'steps': 27939, 'loss/train': 1.0629407167434692} 02/25/2022 13:47:05 - INFO - codeparrot_training - Step 27940: {'lr': 0.0002183420497692391, 'samples': 14305792, 'steps': 27940, 'loss/train': 2.2861998081207275} 02/25/2022 13:47:11 - INFO - codeparrot_training - Step 27941: {'lr': 0.00021832581909656342, 'samples': 14306304, 'steps': 27941, 'loss/train': 0.4250176250934601} 02/25/2022 13:47:14 - INFO - codeparrot_training - Step 27942: {'lr': 0.0002183095885595699, 'samples': 14306816, 'steps': 27942, 'loss/train': 1.5778084993362427} 02/25/2022 13:47:20 - INFO - codeparrot_training - Step 27943: {'lr': 0.00021829335815832813, 'samples': 14307328, 'steps': 27943, 'loss/train': 2.0959837436676025} 02/25/2022 13:47:23 - INFO - codeparrot_training - Step 27944: {'lr': 0.00021827712789290743, 'samples': 14307840, 'steps': 27944, 'loss/train': 1.300779938697815} 02/25/2022 13:47:29 - INFO - codeparrot_training - Step 27945: {'lr': 0.0002182608977633775, 'samples': 14308352, 'steps': 27945, 'loss/train': 1.7630836963653564} 02/25/2022 13:47:34 - INFO - codeparrot_training - Step 27946: {'lr': 0.00021824466776980772, 'samples': 14308864, 'steps': 27946, 'loss/train': 1.1993800401687622} 02/25/2022 13:47:38 - INFO - codeparrot_training - Step 27947: {'lr': 0.00021822843791226786, 'samples': 14309376, 'steps': 27947, 'loss/train': 0.8562624454498291} 02/25/2022 13:47:43 - INFO - codeparrot_training - Step 27948: {'lr': 0.00021821220819082714, 'samples': 14309888, 'steps': 27948, 'loss/train': 1.0514682531356812} 02/25/2022 13:47:47 - INFO - codeparrot_training - Step 27949: {'lr': 0.00021819597860555525, 'samples': 14310400, 'steps': 27949, 'loss/train': 1.1465142965316772} 02/25/2022 13:47:53 - INFO - codeparrot_training - Step 27950: {'lr': 0.00021817974915652172, 'samples': 14310912, 'steps': 27950, 'loss/train': 2.1919548511505127} 02/25/2022 13:47:56 - INFO - codeparrot_training - Step 27951: {'lr': 0.000218163519843796, 'samples': 14311424, 'steps': 27951, 'loss/train': 2.164722204208374} 02/25/2022 13:48:02 - INFO - codeparrot_training - Step 27952: {'lr': 0.00021814729066744776, 'samples': 14311936, 'steps': 27952, 'loss/train': 1.4317976236343384} 02/25/2022 13:48:05 - INFO - codeparrot_training - Step 27953: {'lr': 0.00021813106162754632, 'samples': 14312448, 'steps': 27953, 'loss/train': 2.4428582191467285} 02/25/2022 13:48:11 - INFO - codeparrot_training - Step 27954: {'lr': 0.00021811483272416127, 'samples': 14312960, 'steps': 27954, 'loss/train': 2.5472686290740967} 02/25/2022 13:48:15 - INFO - codeparrot_training - Step 27955: {'lr': 0.00021809860395736216, 'samples': 14313472, 'steps': 27955, 'loss/train': 2.027923107147217} 02/25/2022 13:48:20 - INFO - codeparrot_training - Step 27956: {'lr': 0.00021808237532721864, 'samples': 14313984, 'steps': 27956, 'loss/train': 1.3858634233474731} 02/25/2022 13:48:24 - INFO - codeparrot_training - Step 27957: {'lr': 0.00021806614683379994, 'samples': 14314496, 'steps': 27957, 'loss/train': 2.2144172191619873} 02/25/2022 13:48:29 - INFO - codeparrot_training - Step 27958: {'lr': 0.00021804991847717577, 'samples': 14315008, 'steps': 27958, 'loss/train': 1.7784770727157593} 02/25/2022 13:48:33 - INFO - codeparrot_training - Step 27959: {'lr': 0.00021803369025741556, 'samples': 14315520, 'steps': 27959, 'loss/train': 1.1121454238891602} 02/25/2022 13:48:39 - INFO - codeparrot_training - Step 27960: {'lr': 0.0002180174621745889, 'samples': 14316032, 'steps': 27960, 'loss/train': 2.161804437637329} 02/25/2022 13:48:42 - INFO - codeparrot_training - Step 27961: {'lr': 0.00021800123422876537, 'samples': 14316544, 'steps': 27961, 'loss/train': 2.4507758617401123} 02/25/2022 13:48:48 - INFO - codeparrot_training - Step 27962: {'lr': 0.00021798500642001428, 'samples': 14317056, 'steps': 27962, 'loss/train': 1.5650053024291992} 02/25/2022 13:48:51 - INFO - codeparrot_training - Step 27963: {'lr': 0.00021796877874840525, 'samples': 14317568, 'steps': 27963, 'loss/train': 1.986311912536621} 02/25/2022 13:48:57 - INFO - codeparrot_training - Step 27964: {'lr': 0.0002179525512140078, 'samples': 14318080, 'steps': 27964, 'loss/train': 0.8944240808486938} 02/25/2022 13:49:00 - INFO - codeparrot_training - Step 27965: {'lr': 0.0002179363238168916, 'samples': 14318592, 'steps': 27965, 'loss/train': 1.7376790046691895} 02/25/2022 13:49:06 - INFO - codeparrot_training - Step 27966: {'lr': 0.00021792009655712585, 'samples': 14319104, 'steps': 27966, 'loss/train': 1.2342053651809692} 02/25/2022 13:49:09 - INFO - codeparrot_training - Step 27967: {'lr': 0.00021790386943478025, 'samples': 14319616, 'steps': 27967, 'loss/train': 2.6067380905151367} 02/25/2022 13:49:15 - INFO - codeparrot_training - Step 27968: {'lr': 0.00021788764244992426, 'samples': 14320128, 'steps': 27968, 'loss/train': 2.0745351314544678} 02/25/2022 13:49:18 - INFO - codeparrot_training - Step 27969: {'lr': 0.00021787141560262752, 'samples': 14320640, 'steps': 27969, 'loss/train': 1.684960126876831} 02/25/2022 13:49:24 - INFO - codeparrot_training - Step 27970: {'lr': 0.00021785518889295936, 'samples': 14321152, 'steps': 27970, 'loss/train': 0.7392134070396423} 02/25/2022 13:49:27 - INFO - codeparrot_training - Step 27971: {'lr': 0.00021783896232098932, 'samples': 14321664, 'steps': 27971, 'loss/train': 0.48542502522468567} 02/25/2022 13:49:33 - INFO - codeparrot_training - Step 27972: {'lr': 0.00021782273588678697, 'samples': 14322176, 'steps': 27972, 'loss/train': 1.3457531929016113} 02/25/2022 13:49:37 - INFO - codeparrot_training - Step 27973: {'lr': 0.00021780650959042186, 'samples': 14322688, 'steps': 27973, 'loss/train': 1.9171829223632812} 02/25/2022 13:49:42 - INFO - codeparrot_training - Step 27974: {'lr': 0.00021779028343196343, 'samples': 14323200, 'steps': 27974, 'loss/train': 1.6579058170318604} 02/25/2022 13:49:46 - INFO - codeparrot_training - Step 27975: {'lr': 0.00021777405741148115, 'samples': 14323712, 'steps': 27975, 'loss/train': 0.8871021270751953} 02/25/2022 13:49:51 - INFO - codeparrot_training - Step 27976: {'lr': 0.00021775783152904463, 'samples': 14324224, 'steps': 27976, 'loss/train': 2.3760933876037598} 02/25/2022 13:49:55 - INFO - codeparrot_training - Step 27977: {'lr': 0.00021774160578472328, 'samples': 14324736, 'steps': 27977, 'loss/train': 1.3642617464065552} 02/25/2022 13:50:00 - INFO - codeparrot_training - Step 27978: {'lr': 0.00021772538017858668, 'samples': 14325248, 'steps': 27978, 'loss/train': 1.0976110696792603} 02/25/2022 13:50:04 - INFO - codeparrot_training - Step 27979: {'lr': 0.00021770915471070428, 'samples': 14325760, 'steps': 27979, 'loss/train': 1.7964541912078857} 02/25/2022 13:50:09 - INFO - codeparrot_training - Step 27980: {'lr': 0.00021769292938114563, 'samples': 14326272, 'steps': 27980, 'loss/train': 2.5346992015838623} 02/25/2022 13:50:13 - INFO - codeparrot_training - Step 27981: {'lr': 0.00021767670418998015, 'samples': 14326784, 'steps': 27981, 'loss/train': 2.2278366088867188} 02/25/2022 13:50:18 - INFO - codeparrot_training - Step 27982: {'lr': 0.0002176604791372775, 'samples': 14327296, 'steps': 27982, 'loss/train': 1.601348876953125} 02/25/2022 13:50:22 - INFO - codeparrot_training - Step 27983: {'lr': 0.00021764425422310705, 'samples': 14327808, 'steps': 27983, 'loss/train': 1.6496527194976807} 02/25/2022 13:50:27 - INFO - codeparrot_training - Step 27984: {'lr': 0.00021762802944753828, 'samples': 14328320, 'steps': 27984, 'loss/train': 1.5080457925796509} 02/25/2022 13:50:31 - INFO - codeparrot_training - Step 27985: {'lr': 0.0002176118048106408, 'samples': 14328832, 'steps': 27985, 'loss/train': 1.2395840883255005} 02/25/2022 13:50:37 - INFO - codeparrot_training - Step 27986: {'lr': 0.00021759558031248403, 'samples': 14329344, 'steps': 27986, 'loss/train': 0.9466220140457153} 02/25/2022 13:50:43 - INFO - codeparrot_training - Step 27987: {'lr': 0.00021757935595313762, 'samples': 14329856, 'steps': 27987, 'loss/train': 1.5399867296218872} 02/25/2022 13:50:46 - INFO - codeparrot_training - Step 27988: {'lr': 0.00021756313173267085, 'samples': 14330368, 'steps': 27988, 'loss/train': 2.2046821117401123} 02/25/2022 13:50:52 - INFO - codeparrot_training - Step 27989: {'lr': 0.00021754690765115331, 'samples': 14330880, 'steps': 27989, 'loss/train': 2.1169803142547607} 02/25/2022 13:50:55 - INFO - codeparrot_training - Step 27990: {'lr': 0.00021753068370865454, 'samples': 14331392, 'steps': 27990, 'loss/train': 1.42330002784729} 02/25/2022 13:51:01 - INFO - codeparrot_training - Step 27991: {'lr': 0.0002175144599052441, 'samples': 14331904, 'steps': 27991, 'loss/train': 1.930044174194336} 02/25/2022 13:51:04 - INFO - codeparrot_training - Step 27992: {'lr': 0.0002174982362409913, 'samples': 14332416, 'steps': 27992, 'loss/train': 1.9812712669372559} 02/25/2022 13:51:10 - INFO - codeparrot_training - Step 27993: {'lr': 0.00021748201271596575, 'samples': 14332928, 'steps': 27993, 'loss/train': 1.6960402727127075} 02/25/2022 13:51:13 - INFO - codeparrot_training - Step 27994: {'lr': 0.00021746578933023688, 'samples': 14333440, 'steps': 27994, 'loss/train': 1.3897360563278198} 02/25/2022 13:51:19 - INFO - codeparrot_training - Step 27995: {'lr': 0.0002174495660838744, 'samples': 14333952, 'steps': 27995, 'loss/train': 1.360162615776062} 02/25/2022 13:51:23 - INFO - codeparrot_training - Step 27996: {'lr': 0.0002174333429769475, 'samples': 14334464, 'steps': 27996, 'loss/train': 2.202347993850708} 02/25/2022 13:51:28 - INFO - codeparrot_training - Step 27997: {'lr': 0.00021741712000952583, 'samples': 14334976, 'steps': 27997, 'loss/train': 1.8725557327270508} 02/25/2022 13:51:32 - INFO - codeparrot_training - Step 27998: {'lr': 0.00021740089718167886, 'samples': 14335488, 'steps': 27998, 'loss/train': 1.9601777791976929} 02/25/2022 13:51:37 - INFO - codeparrot_training - Step 27999: {'lr': 0.0002173846744934761, 'samples': 14336000, 'steps': 27999, 'loss/train': 1.8428465127944946} 02/25/2022 13:51:37 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 13:51:55 - WARNING - huggingface_hub.repository - Several commits (28) will be pushed upstream. 02/25/2022 13:51:55 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 13:52:28 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 9d342aa..739c14e floral-grass-11 -> floral-grass-11 02/25/2022 13:52:33 - INFO - codeparrot_training - Step 28000: {'lr': 0.00021736845194498717, 'samples': 14336512, 'steps': 28000, 'loss/train': 1.673930048942566} 02/25/2022 13:52:38 - INFO - codeparrot_training - Step 28001: {'lr': 0.00021735222953628131, 'samples': 14337024, 'steps': 28001, 'loss/train': 1.4876612424850464} 02/25/2022 13:52:42 - INFO - codeparrot_training - Step 28002: {'lr': 0.0002173360072674281, 'samples': 14337536, 'steps': 28002, 'loss/train': 0.10827385634183884} 02/25/2022 13:52:47 - INFO - codeparrot_training - Step 28003: {'lr': 0.0002173197851384971, 'samples': 14338048, 'steps': 28003, 'loss/train': 2.374160051345825} 02/25/2022 13:52:51 - INFO - codeparrot_training - Step 28004: {'lr': 0.00021730356314955785, 'samples': 14338560, 'steps': 28004, 'loss/train': 1.1584421396255493} 02/25/2022 13:52:57 - INFO - codeparrot_training - Step 28005: {'lr': 0.00021728734130067968, 'samples': 14339072, 'steps': 28005, 'loss/train': 2.1272032260894775} 02/25/2022 13:53:00 - INFO - codeparrot_training - Step 28006: {'lr': 0.00021727111959193212, 'samples': 14339584, 'steps': 28006, 'loss/train': 1.7936766147613525} 02/25/2022 13:53:06 - INFO - codeparrot_training - Step 28007: {'lr': 0.0002172548980233847, 'samples': 14340096, 'steps': 28007, 'loss/train': 1.1587542295455933} 02/25/2022 13:53:09 - INFO - codeparrot_training - Step 28008: {'lr': 0.00021723867659510698, 'samples': 14340608, 'steps': 28008, 'loss/train': 1.640569806098938} 02/25/2022 13:53:15 - INFO - codeparrot_training - Step 28009: {'lr': 0.0002172224553071683, 'samples': 14341120, 'steps': 28009, 'loss/train': 1.0196444988250732} 02/25/2022 13:53:18 - INFO - codeparrot_training - Step 28010: {'lr': 0.0002172062341596382, 'samples': 14341632, 'steps': 28010, 'loss/train': 2.137467622756958} 02/25/2022 13:53:24 - INFO - codeparrot_training - Step 28011: {'lr': 0.0002171900131525862, 'samples': 14342144, 'steps': 28011, 'loss/train': 2.705371856689453} 02/25/2022 13:53:27 - INFO - codeparrot_training - Step 28012: {'lr': 0.00021717379228608171, 'samples': 14342656, 'steps': 28012, 'loss/train': 1.1543686389923096} 02/25/2022 13:53:33 - INFO - codeparrot_training - Step 28013: {'lr': 0.0002171575715601944, 'samples': 14343168, 'steps': 28013, 'loss/train': 1.592047929763794} 02/25/2022 13:53:36 - INFO - codeparrot_training - Step 28014: {'lr': 0.00021714135097499355, 'samples': 14343680, 'steps': 28014, 'loss/train': 2.2105391025543213} 02/25/2022 13:53:42 - INFO - codeparrot_training - Step 28015: {'lr': 0.0002171251305305487, 'samples': 14344192, 'steps': 28015, 'loss/train': 1.9797700643539429} 02/25/2022 13:53:45 - INFO - codeparrot_training - Step 28016: {'lr': 0.00021710891022692937, 'samples': 14344704, 'steps': 28016, 'loss/train': 1.2600021362304688} 02/25/2022 13:53:52 - INFO - codeparrot_training - Step 28017: {'lr': 0.00021709269006420508, 'samples': 14345216, 'steps': 28017, 'loss/train': 2.207559108734131} 02/25/2022 13:53:55 - INFO - codeparrot_training - Step 28018: {'lr': 0.00021707647004244517, 'samples': 14345728, 'steps': 28018, 'loss/train': 0.6188715100288391} 02/25/2022 13:54:00 - INFO - codeparrot_training - Step 28019: {'lr': 0.00021706025016171923, 'samples': 14346240, 'steps': 28019, 'loss/train': 1.5557671785354614} 02/25/2022 13:54:04 - INFO - codeparrot_training - Step 28020: {'lr': 0.00021704403042209676, 'samples': 14346752, 'steps': 28020, 'loss/train': 2.0231847763061523} 02/25/2022 13:54:09 - INFO - codeparrot_training - Step 28021: {'lr': 0.00021702781082364714, 'samples': 14347264, 'steps': 28021, 'loss/train': 1.9406682252883911} 02/25/2022 13:54:13 - INFO - codeparrot_training - Step 28022: {'lr': 0.00021701159136643997, 'samples': 14347776, 'steps': 28022, 'loss/train': 0.7393404841423035} 02/25/2022 13:54:18 - INFO - codeparrot_training - Step 28023: {'lr': 0.00021699537205054466, 'samples': 14348288, 'steps': 28023, 'loss/train': 1.646751046180725} 02/25/2022 13:54:22 - INFO - codeparrot_training - Step 28024: {'lr': 0.00021697915287603064, 'samples': 14348800, 'steps': 28024, 'loss/train': 1.502605676651001} 02/25/2022 13:54:27 - INFO - codeparrot_training - Step 28025: {'lr': 0.00021696293384296746, 'samples': 14349312, 'steps': 28025, 'loss/train': 1.6017626523971558} 02/25/2022 13:54:31 - INFO - codeparrot_training - Step 28026: {'lr': 0.00021694671495142462, 'samples': 14349824, 'steps': 28026, 'loss/train': 1.9990664720535278} 02/25/2022 13:54:37 - INFO - codeparrot_training - Step 28027: {'lr': 0.00021693049620147157, 'samples': 14350336, 'steps': 28027, 'loss/train': 1.449980616569519} 02/25/2022 13:54:40 - INFO - codeparrot_training - Step 28028: {'lr': 0.0002169142775931777, 'samples': 14350848, 'steps': 28028, 'loss/train': 2.0600173473358154} 02/25/2022 13:54:46 - INFO - codeparrot_training - Step 28029: {'lr': 0.00021689805912661258, 'samples': 14351360, 'steps': 28029, 'loss/train': 1.0579495429992676} 02/25/2022 13:54:49 - INFO - codeparrot_training - Step 28030: {'lr': 0.00021688184080184572, 'samples': 14351872, 'steps': 28030, 'loss/train': 1.7171986103057861} 02/25/2022 13:54:55 - INFO - codeparrot_training - Step 28031: {'lr': 0.00021686562261894653, 'samples': 14352384, 'steps': 28031, 'loss/train': 1.11579167842865} 02/25/2022 13:54:58 - INFO - codeparrot_training - Step 28032: {'lr': 0.00021684940457798442, 'samples': 14352896, 'steps': 28032, 'loss/train': 1.6769918203353882} 02/25/2022 13:55:04 - INFO - codeparrot_training - Step 28033: {'lr': 0.00021683318667902897, 'samples': 14353408, 'steps': 28033, 'loss/train': 2.0899550914764404} 02/25/2022 13:55:07 - INFO - codeparrot_training - Step 28034: {'lr': 0.0002168169689221496, 'samples': 14353920, 'steps': 28034, 'loss/train': 0.5463775396347046} 02/25/2022 13:55:13 - INFO - codeparrot_training - Step 28035: {'lr': 0.0002168007513074159, 'samples': 14354432, 'steps': 28035, 'loss/train': 1.1548123359680176} 02/25/2022 13:55:19 - INFO - codeparrot_training - Step 28036: {'lr': 0.0002167845338348971, 'samples': 14354944, 'steps': 28036, 'loss/train': 1.4615651369094849} 02/25/2022 13:55:22 - INFO - codeparrot_training - Step 28037: {'lr': 0.00021676831650466286, 'samples': 14355456, 'steps': 28037, 'loss/train': 0.10096308588981628} 02/25/2022 13:55:28 - INFO - codeparrot_training - Step 28038: {'lr': 0.00021675209931678257, 'samples': 14355968, 'steps': 28038, 'loss/train': 1.5023850202560425} 02/25/2022 13:55:31 - INFO - codeparrot_training - Step 28039: {'lr': 0.00021673588227132587, 'samples': 14356480, 'steps': 28039, 'loss/train': 1.06998610496521} 02/25/2022 13:55:37 - INFO - codeparrot_training - Step 28040: {'lr': 0.00021671966536836195, 'samples': 14356992, 'steps': 28040, 'loss/train': 1.199450969696045} 02/25/2022 13:55:41 - INFO - codeparrot_training - Step 28041: {'lr': 0.00021670344860796043, 'samples': 14357504, 'steps': 28041, 'loss/train': 1.4964430332183838} 02/25/2022 13:55:46 - INFO - codeparrot_training - Step 28042: {'lr': 0.00021668723199019075, 'samples': 14358016, 'steps': 28042, 'loss/train': 1.9237947463989258} 02/25/2022 13:55:50 - INFO - codeparrot_training - Step 28043: {'lr': 0.00021667101551512254, 'samples': 14358528, 'steps': 28043, 'loss/train': 1.155117392539978} 02/25/2022 13:55:55 - INFO - codeparrot_training - Step 28044: {'lr': 0.00021665479918282494, 'samples': 14359040, 'steps': 28044, 'loss/train': 1.9901700019836426} 02/25/2022 13:55:59 - INFO - codeparrot_training - Step 28045: {'lr': 0.00021663858299336765, 'samples': 14359552, 'steps': 28045, 'loss/train': 1.1686725616455078} 02/25/2022 13:56:04 - INFO - codeparrot_training - Step 28046: {'lr': 0.00021662236694682002, 'samples': 14360064, 'steps': 28046, 'loss/train': 1.9542475938796997} 02/25/2022 13:56:08 - INFO - codeparrot_training - Step 28047: {'lr': 0.00021660615104325163, 'samples': 14360576, 'steps': 28047, 'loss/train': 1.1862233877182007} 02/25/2022 13:56:13 - INFO - codeparrot_training - Step 28048: {'lr': 0.00021658993528273195, 'samples': 14361088, 'steps': 28048, 'loss/train': 1.810849905014038} 02/25/2022 13:56:17 - INFO - codeparrot_training - Step 28049: {'lr': 0.0002165737196653303, 'samples': 14361600, 'steps': 28049, 'loss/train': 2.0611181259155273} 02/25/2022 13:56:23 - INFO - codeparrot_training - Step 28050: {'lr': 0.00021655750419111622, 'samples': 14362112, 'steps': 28050, 'loss/train': 2.0399246215820312} 02/25/2022 13:56:26 - INFO - codeparrot_training - Step 28051: {'lr': 0.00021654128886015918, 'samples': 14362624, 'steps': 28051, 'loss/train': 2.108231782913208} 02/25/2022 13:56:32 - INFO - codeparrot_training - Step 28052: {'lr': 0.0002165250736725287, 'samples': 14363136, 'steps': 28052, 'loss/train': 2.020242214202881} 02/25/2022 13:56:36 - INFO - codeparrot_training - Step 28053: {'lr': 0.0002165088586282941, 'samples': 14363648, 'steps': 28053, 'loss/train': 0.4527716040611267} 02/25/2022 13:56:41 - INFO - codeparrot_training - Step 28054: {'lr': 0.0002164926437275249, 'samples': 14364160, 'steps': 28054, 'loss/train': 2.0973050594329834} 02/25/2022 13:56:45 - INFO - codeparrot_training - Step 28055: {'lr': 0.00021647642897029058, 'samples': 14364672, 'steps': 28055, 'loss/train': 1.2976244688034058} 02/25/2022 13:56:50 - INFO - codeparrot_training - Step 28056: {'lr': 0.00021646021435666073, 'samples': 14365184, 'steps': 28056, 'loss/train': 2.2881805896759033} 02/25/2022 13:56:54 - INFO - codeparrot_training - Step 28057: {'lr': 0.00021644399988670452, 'samples': 14365696, 'steps': 28057, 'loss/train': 1.6598111391067505} 02/25/2022 13:56:59 - INFO - codeparrot_training - Step 28058: {'lr': 0.0002164277855604916, 'samples': 14366208, 'steps': 28058, 'loss/train': 1.6773346662521362} 02/25/2022 13:57:03 - INFO - codeparrot_training - Step 28059: {'lr': 0.00021641157137809134, 'samples': 14366720, 'steps': 28059, 'loss/train': 1.7139997482299805} 02/25/2022 13:57:08 - INFO - codeparrot_training - Step 28060: {'lr': 0.00021639535733957333, 'samples': 14367232, 'steps': 28060, 'loss/train': 1.2584538459777832} 02/25/2022 13:57:12 - INFO - codeparrot_training - Step 28061: {'lr': 0.00021637914344500693, 'samples': 14367744, 'steps': 28061, 'loss/train': 1.7919621467590332} 02/25/2022 13:57:18 - INFO - codeparrot_training - Step 28062: {'lr': 0.00021636292969446152, 'samples': 14368256, 'steps': 28062, 'loss/train': 2.3028078079223633} 02/25/2022 13:57:21 - INFO - codeparrot_training - Step 28063: {'lr': 0.00021634671608800667, 'samples': 14368768, 'steps': 28063, 'loss/train': 1.750710368156433} 02/25/2022 13:57:27 - INFO - codeparrot_training - Step 28064: {'lr': 0.00021633050262571187, 'samples': 14369280, 'steps': 28064, 'loss/train': 2.099971294403076} 02/25/2022 13:57:30 - INFO - codeparrot_training - Step 28065: {'lr': 0.0002163142893076465, 'samples': 14369792, 'steps': 28065, 'loss/train': 1.5793737173080444} 02/25/2022 13:57:36 - INFO - codeparrot_training - Step 28066: {'lr': 0.00021629807613387993, 'samples': 14370304, 'steps': 28066, 'loss/train': 1.7580320835113525} 02/25/2022 13:57:39 - INFO - codeparrot_training - Step 28067: {'lr': 0.0002162818631044818, 'samples': 14370816, 'steps': 28067, 'loss/train': 2.207139015197754} 02/25/2022 13:57:45 - INFO - codeparrot_training - Step 28068: {'lr': 0.00021626565021952136, 'samples': 14371328, 'steps': 28068, 'loss/train': 2.1355602741241455} 02/25/2022 13:57:48 - INFO - codeparrot_training - Step 28069: {'lr': 0.0002162494374790682, 'samples': 14371840, 'steps': 28069, 'loss/train': 1.7528828382492065} 02/25/2022 13:57:54 - INFO - codeparrot_training - Step 28070: {'lr': 0.00021623322488319174, 'samples': 14372352, 'steps': 28070, 'loss/train': 0.9418584704399109} 02/25/2022 13:57:57 - INFO - codeparrot_training - Step 28071: {'lr': 0.0002162170124319615, 'samples': 14372864, 'steps': 28071, 'loss/train': 1.112986445426941} 02/25/2022 13:58:03 - INFO - codeparrot_training - Step 28072: {'lr': 0.00021620080012544674, 'samples': 14373376, 'steps': 28072, 'loss/train': 2.258427619934082} 02/25/2022 13:58:07 - INFO - codeparrot_training - Step 28073: {'lr': 0.000216184587963717, 'samples': 14373888, 'steps': 28073, 'loss/train': 1.416995882987976} 02/25/2022 13:58:12 - INFO - codeparrot_training - Step 28074: {'lr': 0.00021616837594684192, 'samples': 14374400, 'steps': 28074, 'loss/train': 1.9147346019744873} 02/25/2022 13:58:16 - INFO - codeparrot_training - Step 28075: {'lr': 0.00021615216407489064, 'samples': 14374912, 'steps': 28075, 'loss/train': 0.5222363471984863} 02/25/2022 13:58:21 - INFO - codeparrot_training - Step 28076: {'lr': 0.0002161359523479327, 'samples': 14375424, 'steps': 28076, 'loss/train': 1.954379916191101} 02/25/2022 13:58:25 - INFO - codeparrot_training - Step 28077: {'lr': 0.00021611974076603764, 'samples': 14375936, 'steps': 28077, 'loss/train': 1.7154027223587036} 02/25/2022 13:58:30 - INFO - codeparrot_training - Step 28078: {'lr': 0.00021610352932927495, 'samples': 14376448, 'steps': 28078, 'loss/train': 1.7833881378173828} 02/25/2022 13:58:34 - INFO - codeparrot_training - Step 28079: {'lr': 0.00021608731803771387, 'samples': 14376960, 'steps': 28079, 'loss/train': 0.8669403791427612} 02/25/2022 13:58:39 - INFO - codeparrot_training - Step 28080: {'lr': 0.00021607110689142393, 'samples': 14377472, 'steps': 28080, 'loss/train': 1.6418979167938232} 02/25/2022 13:58:43 - INFO - codeparrot_training - Step 28081: {'lr': 0.0002160548958904746, 'samples': 14377984, 'steps': 28081, 'loss/train': 0.0733647346496582} 02/25/2022 13:58:50 - INFO - codeparrot_training - Step 28082: {'lr': 0.00021603868503493535, 'samples': 14378496, 'steps': 28082, 'loss/train': 1.7847167253494263} 02/25/2022 13:58:53 - INFO - codeparrot_training - Step 28083: {'lr': 0.00021602247432487564, 'samples': 14379008, 'steps': 28083, 'loss/train': 1.5454208850860596} 02/25/2022 13:58:59 - INFO - codeparrot_training - Step 28084: {'lr': 0.00021600626376036482, 'samples': 14379520, 'steps': 28084, 'loss/train': 2.453158378601074} 02/25/2022 13:59:02 - INFO - codeparrot_training - Step 28085: {'lr': 0.00021599005334147232, 'samples': 14380032, 'steps': 28085, 'loss/train': 1.43528413772583} 02/25/2022 13:59:08 - INFO - codeparrot_training - Step 28086: {'lr': 0.00021597384306826765, 'samples': 14380544, 'steps': 28086, 'loss/train': 1.4690775871276855} 02/25/2022 13:59:11 - INFO - codeparrot_training - Step 28087: {'lr': 0.00021595763294082033, 'samples': 14381056, 'steps': 28087, 'loss/train': 1.5960325002670288} 02/25/2022 13:59:17 - INFO - codeparrot_training - Step 28088: {'lr': 0.0002159414229591996, 'samples': 14381568, 'steps': 28088, 'loss/train': 2.019350528717041} 02/25/2022 13:59:20 - INFO - codeparrot_training - Step 28089: {'lr': 0.000215925213123475, 'samples': 14382080, 'steps': 28089, 'loss/train': 1.195811152458191} 02/25/2022 13:59:26 - INFO - codeparrot_training - Step 28090: {'lr': 0.00021590900343371598, 'samples': 14382592, 'steps': 28090, 'loss/train': 1.4054591655731201} 02/25/2022 13:59:29 - INFO - codeparrot_training - Step 28091: {'lr': 0.00021589279388999205, 'samples': 14383104, 'steps': 28091, 'loss/train': 1.5456644296646118} 02/25/2022 13:59:35 - INFO - codeparrot_training - Step 28092: {'lr': 0.00021587658449237246, 'samples': 14383616, 'steps': 28092, 'loss/train': 1.7446430921554565} 02/25/2022 13:59:38 - INFO - codeparrot_training - Step 28093: {'lr': 0.00021586037524092677, 'samples': 14384128, 'steps': 28093, 'loss/train': 2.2373547554016113} 02/25/2022 13:59:44 - INFO - codeparrot_training - Step 28094: {'lr': 0.00021584416613572435, 'samples': 14384640, 'steps': 28094, 'loss/train': 1.3391170501708984} 02/25/2022 13:59:47 - INFO - codeparrot_training - Step 28095: {'lr': 0.0002158279571768347, 'samples': 14385152, 'steps': 28095, 'loss/train': 1.7505850791931152} 02/25/2022 13:59:53 - INFO - codeparrot_training - Step 28096: {'lr': 0.00021581174836432735, 'samples': 14385664, 'steps': 28096, 'loss/train': 1.226311445236206} 02/25/2022 13:59:56 - INFO - codeparrot_training - Step 28097: {'lr': 0.00021579553969827147, 'samples': 14386176, 'steps': 28097, 'loss/train': 1.9549120664596558} 02/25/2022 14:00:02 - INFO - codeparrot_training - Step 28098: {'lr': 0.00021577933117873668, 'samples': 14386688, 'steps': 28098, 'loss/train': 2.2918050289154053} 02/25/2022 14:00:08 - INFO - codeparrot_training - Step 28099: {'lr': 0.00021576312280579236, 'samples': 14387200, 'steps': 28099, 'loss/train': 2.2413296699523926} 02/25/2022 14:00:11 - INFO - codeparrot_training - Step 28100: {'lr': 0.00021574691457950805, 'samples': 14387712, 'steps': 28100, 'loss/train': 1.1642554998397827} 02/25/2022 14:00:17 - INFO - codeparrot_training - Step 28101: {'lr': 0.00021573070649995296, 'samples': 14388224, 'steps': 28101, 'loss/train': 1.2302623987197876} 02/25/2022 14:00:20 - INFO - codeparrot_training - Step 28102: {'lr': 0.00021571449856719666, 'samples': 14388736, 'steps': 28102, 'loss/train': 1.9054666757583618} 02/25/2022 14:00:26 - INFO - codeparrot_training - Step 28103: {'lr': 0.00021569829078130854, 'samples': 14389248, 'steps': 28103, 'loss/train': 1.915951132774353} 02/25/2022 14:00:29 - INFO - codeparrot_training - Step 28104: {'lr': 0.00021568208314235818, 'samples': 14389760, 'steps': 28104, 'loss/train': 1.8752994537353516} 02/25/2022 14:00:35 - INFO - codeparrot_training - Step 28105: {'lr': 0.00021566587565041476, 'samples': 14390272, 'steps': 28105, 'loss/train': 1.3552035093307495} 02/25/2022 14:00:38 - INFO - codeparrot_training - Step 28106: {'lr': 0.00021564966830554783, 'samples': 14390784, 'steps': 28106, 'loss/train': 1.3413447141647339} 02/25/2022 14:00:45 - INFO - codeparrot_training - Step 28107: {'lr': 0.00021563346110782685, 'samples': 14391296, 'steps': 28107, 'loss/train': 2.0431149005889893} 02/25/2022 14:00:48 - INFO - codeparrot_training - Step 28108: {'lr': 0.0002156172540573212, 'samples': 14391808, 'steps': 28108, 'loss/train': 0.9471141695976257} 02/25/2022 14:00:54 - INFO - codeparrot_training - Step 28109: {'lr': 0.00021560104715410034, 'samples': 14392320, 'steps': 28109, 'loss/train': 1.4455173015594482} 02/25/2022 14:00:57 - INFO - codeparrot_training - Step 28110: {'lr': 0.00021558484039823362, 'samples': 14392832, 'steps': 28110, 'loss/train': 1.6967476606369019} 02/25/2022 14:01:02 - INFO - codeparrot_training - Step 28111: {'lr': 0.00021556863378979055, 'samples': 14393344, 'steps': 28111, 'loss/train': 2.3741543292999268} 02/25/2022 14:01:06 - INFO - codeparrot_training - Step 28112: {'lr': 0.0002155524273288405, 'samples': 14393856, 'steps': 28112, 'loss/train': 2.576460599899292} 02/25/2022 14:01:11 - INFO - codeparrot_training - Step 28113: {'lr': 0.00021553622101545296, 'samples': 14394368, 'steps': 28113, 'loss/train': 1.945274829864502} 02/25/2022 14:01:15 - INFO - codeparrot_training - Step 28114: {'lr': 0.00021552001484969726, 'samples': 14394880, 'steps': 28114, 'loss/train': 1.6612763404846191} 02/25/2022 14:01:20 - INFO - codeparrot_training - Step 28115: {'lr': 0.00021550380883164286, 'samples': 14395392, 'steps': 28115, 'loss/train': 1.6042780876159668} 02/25/2022 14:01:24 - INFO - codeparrot_training - Step 28116: {'lr': 0.0002154876029613592, 'samples': 14395904, 'steps': 28116, 'loss/train': 0.7500272989273071} 02/25/2022 14:01:30 - INFO - codeparrot_training - Step 28117: {'lr': 0.00021547139723891565, 'samples': 14396416, 'steps': 28117, 'loss/train': 1.5988813638687134} 02/25/2022 14:01:33 - INFO - codeparrot_training - Step 28118: {'lr': 0.00021545519166438182, 'samples': 14396928, 'steps': 28118, 'loss/train': 1.2883379459381104} 02/25/2022 14:01:39 - INFO - codeparrot_training - Step 28119: {'lr': 0.00021543898623782684, 'samples': 14397440, 'steps': 28119, 'loss/train': 1.5287202596664429} 02/25/2022 14:01:42 - INFO - codeparrot_training - Step 28120: {'lr': 0.00021542278095932027, 'samples': 14397952, 'steps': 28120, 'loss/train': 1.446513056755066} 02/25/2022 14:01:48 - INFO - codeparrot_training - Step 28121: {'lr': 0.00021540657582893154, 'samples': 14398464, 'steps': 28121, 'loss/train': 2.0345067977905273} 02/25/2022 14:01:51 - INFO - codeparrot_training - Step 28122: {'lr': 0.00021539037084673015, 'samples': 14398976, 'steps': 28122, 'loss/train': 2.1813249588012695} 02/25/2022 14:01:57 - INFO - codeparrot_training - Step 28123: {'lr': 0.00021537416601278533, 'samples': 14399488, 'steps': 28123, 'loss/train': 0.6046134233474731} 02/25/2022 14:02:00 - INFO - codeparrot_training - Step 28124: {'lr': 0.00021535796132716658, 'samples': 14400000, 'steps': 28124, 'loss/train': 1.187265157699585} 02/25/2022 14:02:06 - INFO - codeparrot_training - Step 28125: {'lr': 0.00021534175678994335, 'samples': 14400512, 'steps': 28125, 'loss/train': 1.6002784967422485} 02/25/2022 14:02:09 - INFO - codeparrot_training - Step 28126: {'lr': 0.00021532555240118509, 'samples': 14401024, 'steps': 28126, 'loss/train': 1.6754015684127808} 02/25/2022 14:02:16 - INFO - codeparrot_training - Step 28127: {'lr': 0.00021530934816096106, 'samples': 14401536, 'steps': 28127, 'loss/train': 2.282870054244995} 02/25/2022 14:02:19 - INFO - codeparrot_training - Step 28128: {'lr': 0.00021529314406934077, 'samples': 14402048, 'steps': 28128, 'loss/train': 2.4172914028167725} 02/25/2022 14:02:25 - INFO - codeparrot_training - Step 28129: {'lr': 0.00021527694012639362, 'samples': 14402560, 'steps': 28129, 'loss/train': 1.4151328802108765} 02/25/2022 14:02:28 - INFO - codeparrot_training - Step 28130: {'lr': 0.00021526073633218907, 'samples': 14403072, 'steps': 28130, 'loss/train': 1.3612678050994873} 02/25/2022 14:02:34 - INFO - codeparrot_training - Step 28131: {'lr': 0.00021524453268679657, 'samples': 14403584, 'steps': 28131, 'loss/train': 2.0022971630096436} 02/25/2022 14:02:37 - INFO - codeparrot_training - Step 28132: {'lr': 0.00021522832919028533, 'samples': 14404096, 'steps': 28132, 'loss/train': 1.8268616199493408} 02/25/2022 14:02:43 - INFO - codeparrot_training - Step 28133: {'lr': 0.00021521212584272494, 'samples': 14404608, 'steps': 28133, 'loss/train': 1.985948085784912} 02/25/2022 14:02:46 - INFO - codeparrot_training - Step 28134: {'lr': 0.00021519592264418472, 'samples': 14405120, 'steps': 28134, 'loss/train': 0.5921363234519958} 02/25/2022 14:02:52 - INFO - codeparrot_training - Step 28135: {'lr': 0.00021517971959473422, 'samples': 14405632, 'steps': 28135, 'loss/train': 1.3431898355484009} 02/25/2022 14:02:55 - INFO - codeparrot_training - Step 28136: {'lr': 0.00021516351669444267, 'samples': 14406144, 'steps': 28136, 'loss/train': 1.3506590127944946} 02/25/2022 14:03:01 - INFO - codeparrot_training - Step 28137: {'lr': 0.00021514731394337952, 'samples': 14406656, 'steps': 28137, 'loss/train': 2.289846658706665} 02/25/2022 14:03:04 - INFO - codeparrot_training - Step 28138: {'lr': 0.00021513111134161423, 'samples': 14407168, 'steps': 28138, 'loss/train': 1.7058439254760742} 02/25/2022 14:03:10 - INFO - codeparrot_training - Step 28139: {'lr': 0.00021511490888921628, 'samples': 14407680, 'steps': 28139, 'loss/train': 1.9068092107772827} 02/25/2022 14:03:14 - INFO - codeparrot_training - Step 28140: {'lr': 0.00021509870658625489, 'samples': 14408192, 'steps': 28140, 'loss/train': 1.0219967365264893} 02/25/2022 14:03:19 - INFO - codeparrot_training - Step 28141: {'lr': 0.00021508250443279952, 'samples': 14408704, 'steps': 28141, 'loss/train': 2.337024450302124} 02/25/2022 14:03:23 - INFO - codeparrot_training - Step 28142: {'lr': 0.00021506630242891967, 'samples': 14409216, 'steps': 28142, 'loss/train': 2.6537768840789795} 02/25/2022 14:03:28 - INFO - codeparrot_training - Step 28143: {'lr': 0.00021505010057468467, 'samples': 14409728, 'steps': 28143, 'loss/train': 2.428022861480713} 02/25/2022 14:03:31 - INFO - codeparrot_training - Step 28144: {'lr': 0.00021503389887016404, 'samples': 14410240, 'steps': 28144, 'loss/train': 8.52154541015625} 02/25/2022 14:03:37 - INFO - codeparrot_training - Step 28145: {'lr': 0.00021501769731542694, 'samples': 14410752, 'steps': 28145, 'loss/train': 1.2380784749984741} 02/25/2022 14:03:41 - INFO - codeparrot_training - Step 28146: {'lr': 0.00021500149591054297, 'samples': 14411264, 'steps': 28146, 'loss/train': 1.3599052429199219} 02/25/2022 14:03:46 - INFO - codeparrot_training - Step 28147: {'lr': 0.0002149852946555815, 'samples': 14411776, 'steps': 28147, 'loss/train': 1.460645318031311} 02/25/2022 14:03:53 - INFO - codeparrot_training - Step 28148: {'lr': 0.00021496909355061194, 'samples': 14412288, 'steps': 28148, 'loss/train': 2.548835277557373} 02/25/2022 14:03:56 - INFO - codeparrot_training - Step 28149: {'lr': 0.0002149528925957036, 'samples': 14412800, 'steps': 28149, 'loss/train': 1.4063996076583862} 02/25/2022 14:04:02 - INFO - codeparrot_training - Step 28150: {'lr': 0.00021493669179092594, 'samples': 14413312, 'steps': 28150, 'loss/train': 1.708360195159912} 02/25/2022 14:04:05 - INFO - codeparrot_training - Step 28151: {'lr': 0.00021492049113634837, 'samples': 14413824, 'steps': 28151, 'loss/train': 2.007326126098633} 02/25/2022 14:04:11 - INFO - codeparrot_training - Step 28152: {'lr': 0.0002149042906320403, 'samples': 14414336, 'steps': 28152, 'loss/train': 2.248398780822754} 02/25/2022 14:04:14 - INFO - codeparrot_training - Step 28153: {'lr': 0.00021488809027807106, 'samples': 14414848, 'steps': 28153, 'loss/train': 0.7525196671485901} 02/25/2022 14:04:19 - INFO - codeparrot_training - Step 28154: {'lr': 0.00021487189007451016, 'samples': 14415360, 'steps': 28154, 'loss/train': 1.4810278415679932} 02/25/2022 14:04:23 - INFO - codeparrot_training - Step 28155: {'lr': 0.00021485569002142684, 'samples': 14415872, 'steps': 28155, 'loss/train': 1.97084379196167} 02/25/2022 14:04:28 - INFO - codeparrot_training - Step 28156: {'lr': 0.00021483949011889066, 'samples': 14416384, 'steps': 28156, 'loss/train': 1.7304683923721313} 02/25/2022 14:04:32 - INFO - codeparrot_training - Step 28157: {'lr': 0.0002148232903669709, 'samples': 14416896, 'steps': 28157, 'loss/train': 1.599805235862732} 02/25/2022 14:04:37 - INFO - codeparrot_training - Step 28158: {'lr': 0.00021480709076573704, 'samples': 14417408, 'steps': 28158, 'loss/train': 0.06547779589891434} 02/25/2022 14:04:41 - INFO - codeparrot_training - Step 28159: {'lr': 0.00021479089131525836, 'samples': 14417920, 'steps': 28159, 'loss/train': 2.0388407707214355} 02/25/2022 14:04:46 - INFO - codeparrot_training - Step 28160: {'lr': 0.00021477469201560434, 'samples': 14418432, 'steps': 28160, 'loss/train': 1.4564509391784668} 02/25/2022 14:04:50 - INFO - codeparrot_training - Step 28161: {'lr': 0.00021475849286684446, 'samples': 14418944, 'steps': 28161, 'loss/train': 1.5271575450897217} 02/25/2022 14:04:56 - INFO - codeparrot_training - Step 28162: {'lr': 0.00021474229386904786, 'samples': 14419456, 'steps': 28162, 'loss/train': 1.5319429636001587} 02/25/2022 14:05:00 - INFO - codeparrot_training - Step 28163: {'lr': 0.00021472609502228414, 'samples': 14419968, 'steps': 28163, 'loss/train': 1.6581001281738281} 02/25/2022 14:05:05 - INFO - codeparrot_training - Step 28164: {'lr': 0.0002147098963266226, 'samples': 14420480, 'steps': 28164, 'loss/train': 1.5576667785644531} 02/25/2022 14:05:08 - INFO - codeparrot_training - Step 28165: {'lr': 0.00021469369778213273, 'samples': 14420992, 'steps': 28165, 'loss/train': 1.6378532648086548} 02/25/2022 14:05:14 - INFO - codeparrot_training - Step 28166: {'lr': 0.00021467749938888377, 'samples': 14421504, 'steps': 28166, 'loss/train': 1.3086446523666382} 02/25/2022 14:05:17 - INFO - codeparrot_training - Step 28167: {'lr': 0.0002146613011469452, 'samples': 14422016, 'steps': 28167, 'loss/train': 1.9520518779754639} 02/25/2022 14:05:23 - INFO - codeparrot_training - Step 28168: {'lr': 0.00021464510305638638, 'samples': 14422528, 'steps': 28168, 'loss/train': 2.518522024154663} 02/25/2022 14:05:26 - INFO - codeparrot_training - Step 28169: {'lr': 0.0002146289051172767, 'samples': 14423040, 'steps': 28169, 'loss/train': 1.546021580696106} 02/25/2022 14:05:32 - INFO - codeparrot_training - Step 28170: {'lr': 0.0002146127073296857, 'samples': 14423552, 'steps': 28170, 'loss/train': 2.0718612670898438} 02/25/2022 14:05:35 - INFO - codeparrot_training - Step 28171: {'lr': 0.0002145965096936825, 'samples': 14424064, 'steps': 28171, 'loss/train': 1.227023959159851} 02/25/2022 14:05:41 - INFO - codeparrot_training - Step 28172: {'lr': 0.0002145803122093366, 'samples': 14424576, 'steps': 28172, 'loss/train': 1.546911597251892} 02/25/2022 14:05:44 - INFO - codeparrot_training - Step 28173: {'lr': 0.0002145641148767174, 'samples': 14425088, 'steps': 28173, 'loss/train': 1.7818413972854614} 02/25/2022 14:05:51 - INFO - codeparrot_training - Step 28174: {'lr': 0.0002145479176958944, 'samples': 14425600, 'steps': 28174, 'loss/train': 2.3829448223114014} 02/25/2022 14:05:54 - INFO - codeparrot_training - Step 28175: {'lr': 0.00021453172066693673, 'samples': 14426112, 'steps': 28175, 'loss/train': 1.5756821632385254} 02/25/2022 14:06:00 - INFO - codeparrot_training - Step 28176: {'lr': 0.0002145155237899139, 'samples': 14426624, 'steps': 28176, 'loss/train': 2.3233695030212402} 02/25/2022 14:06:03 - INFO - codeparrot_training - Step 28177: {'lr': 0.00021449932706489534, 'samples': 14427136, 'steps': 28177, 'loss/train': 1.4959889650344849} 02/25/2022 14:06:09 - INFO - codeparrot_training - Step 28178: {'lr': 0.00021448313049195037, 'samples': 14427648, 'steps': 28178, 'loss/train': 1.8766247034072876} 02/25/2022 14:06:12 - INFO - codeparrot_training - Step 28179: {'lr': 0.00021446693407114852, 'samples': 14428160, 'steps': 28179, 'loss/train': 1.7371975183486938} 02/25/2022 14:06:17 - INFO - codeparrot_training - Step 28180: {'lr': 0.00021445073780255888, 'samples': 14428672, 'steps': 28180, 'loss/train': 1.3295888900756836} 02/25/2022 14:06:21 - INFO - codeparrot_training - Step 28181: {'lr': 0.00021443454168625103, 'samples': 14429184, 'steps': 28181, 'loss/train': 1.6605846881866455} 02/25/2022 14:06:26 - INFO - codeparrot_training - Step 28182: {'lr': 0.0002144183457222943, 'samples': 14429696, 'steps': 28182, 'loss/train': 1.1189130544662476} 02/25/2022 14:06:30 - INFO - codeparrot_training - Step 28183: {'lr': 0.00021440214991075822, 'samples': 14430208, 'steps': 28183, 'loss/train': 1.8553526401519775} 02/25/2022 14:06:36 - INFO - codeparrot_training - Step 28184: {'lr': 0.00021438595425171187, 'samples': 14430720, 'steps': 28184, 'loss/train': 1.5230867862701416} 02/25/2022 14:06:39 - INFO - codeparrot_training - Step 28185: {'lr': 0.0002143697587452248, 'samples': 14431232, 'steps': 28185, 'loss/train': 2.4427521228790283} 02/25/2022 14:06:45 - INFO - codeparrot_training - Step 28186: {'lr': 0.00021435356339136638, 'samples': 14431744, 'steps': 28186, 'loss/train': 2.2740325927734375} 02/25/2022 14:06:48 - INFO - codeparrot_training - Step 28187: {'lr': 0.00021433736819020607, 'samples': 14432256, 'steps': 28187, 'loss/train': 2.3454790115356445} 02/25/2022 14:06:54 - INFO - codeparrot_training - Step 28188: {'lr': 0.00021432117314181304, 'samples': 14432768, 'steps': 28188, 'loss/train': 2.020791530609131} 02/25/2022 14:06:57 - INFO - codeparrot_training - Step 28189: {'lr': 0.00021430497824625677, 'samples': 14433280, 'steps': 28189, 'loss/train': 0.047374699264764786} 02/25/2022 14:07:03 - INFO - codeparrot_training - Step 28190: {'lr': 0.00021428878350360667, 'samples': 14433792, 'steps': 28190, 'loss/train': 1.9009246826171875} 02/25/2022 14:07:06 - INFO - codeparrot_training - Step 28191: {'lr': 0.0002142725889139321, 'samples': 14434304, 'steps': 28191, 'loss/train': 2.777763605117798} 02/25/2022 14:07:12 - INFO - codeparrot_training - Step 28192: {'lr': 0.00021425639447730243, 'samples': 14434816, 'steps': 28192, 'loss/train': 2.066321611404419} 02/25/2022 14:07:15 - INFO - codeparrot_training - Step 28193: {'lr': 0.00021424020019378699, 'samples': 14435328, 'steps': 28193, 'loss/train': 1.9836395978927612} 02/25/2022 14:07:22 - INFO - codeparrot_training - Step 28194: {'lr': 0.0002142240060634552, 'samples': 14435840, 'steps': 28194, 'loss/train': 1.4912992715835571} 02/25/2022 14:07:26 - INFO - codeparrot_training - Step 28195: {'lr': 0.00021420781208637634, 'samples': 14436352, 'steps': 28195, 'loss/train': 1.8374536037445068} 02/25/2022 14:07:31 - INFO - codeparrot_training - Step 28196: {'lr': 0.00021419161826261997, 'samples': 14436864, 'steps': 28196, 'loss/train': 1.4489614963531494} 02/25/2022 14:07:35 - INFO - codeparrot_training - Step 28197: {'lr': 0.0002141754245922552, 'samples': 14437376, 'steps': 28197, 'loss/train': 0.8596668839454651} 02/25/2022 14:07:40 - INFO - codeparrot_training - Step 28198: {'lr': 0.00021415923107535162, 'samples': 14437888, 'steps': 28198, 'loss/train': 1.210690975189209} 02/25/2022 14:07:44 - INFO - codeparrot_training - Step 28199: {'lr': 0.00021414303771197845, 'samples': 14438400, 'steps': 28199, 'loss/train': 1.900983214378357} 02/25/2022 14:07:49 - INFO - codeparrot_training - Step 28200: {'lr': 0.0002141268445022052, 'samples': 14438912, 'steps': 28200, 'loss/train': 1.0254895687103271} 02/25/2022 14:07:53 - INFO - codeparrot_training - Step 28201: {'lr': 0.00021411065144610113, 'samples': 14439424, 'steps': 28201, 'loss/train': 1.9845155477523804} 02/25/2022 14:07:59 - INFO - codeparrot_training - Step 28202: {'lr': 0.0002140944585437356, 'samples': 14439936, 'steps': 28202, 'loss/train': 1.625124454498291} 02/25/2022 14:08:02 - INFO - codeparrot_training - Step 28203: {'lr': 0.00021407826579517803, 'samples': 14440448, 'steps': 28203, 'loss/train': 1.8188352584838867} 02/25/2022 14:08:08 - INFO - codeparrot_training - Step 28204: {'lr': 0.00021406207320049772, 'samples': 14440960, 'steps': 28204, 'loss/train': 0.9709446430206299} 02/25/2022 14:08:11 - INFO - codeparrot_training - Step 28205: {'lr': 0.00021404588075976422, 'samples': 14441472, 'steps': 28205, 'loss/train': 1.3550825119018555} 02/25/2022 14:08:17 - INFO - codeparrot_training - Step 28206: {'lr': 0.0002140296884730466, 'samples': 14441984, 'steps': 28206, 'loss/train': 1.4155069589614868} 02/25/2022 14:08:20 - INFO - codeparrot_training - Step 28207: {'lr': 0.0002140134963404144, 'samples': 14442496, 'steps': 28207, 'loss/train': 1.8521008491516113} 02/25/2022 14:08:25 - INFO - codeparrot_training - Step 28208: {'lr': 0.00021399730436193694, 'samples': 14443008, 'steps': 28208, 'loss/train': 2.0159757137298584} 02/25/2022 14:08:29 - INFO - codeparrot_training - Step 28209: {'lr': 0.00021398111253768373, 'samples': 14443520, 'steps': 28209, 'loss/train': 0.8082032203674316} 02/25/2022 14:08:35 - INFO - codeparrot_training - Step 28210: {'lr': 0.00021396492086772387, 'samples': 14444032, 'steps': 28210, 'loss/train': 1.9070026874542236} 02/25/2022 14:08:39 - INFO - codeparrot_training - Step 28211: {'lr': 0.00021394872935212684, 'samples': 14444544, 'steps': 28211, 'loss/train': 1.7314732074737549} 02/25/2022 14:08:44 - INFO - codeparrot_training - Step 28212: {'lr': 0.000213932537990962, 'samples': 14445056, 'steps': 28212, 'loss/train': 1.6581107378005981} 02/25/2022 14:08:47 - INFO - codeparrot_training - Step 28213: {'lr': 0.00021391634678429887, 'samples': 14445568, 'steps': 28213, 'loss/train': 0.4519192576408386} 02/25/2022 14:08:53 - INFO - codeparrot_training - Step 28214: {'lr': 0.0002139001557322065, 'samples': 14446080, 'steps': 28214, 'loss/train': 1.3191540241241455} 02/25/2022 14:08:56 - INFO - codeparrot_training - Step 28215: {'lr': 0.0002138839648347544, 'samples': 14446592, 'steps': 28215, 'loss/train': 2.546160936355591} 02/25/2022 14:09:02 - INFO - codeparrot_training - Step 28216: {'lr': 0.00021386777409201194, 'samples': 14447104, 'steps': 28216, 'loss/train': 0.9705095887184143} 02/25/2022 14:09:05 - INFO - codeparrot_training - Step 28217: {'lr': 0.00021385158350404845, 'samples': 14447616, 'steps': 28217, 'loss/train': 1.3036950826644897} 02/25/2022 14:09:11 - INFO - codeparrot_training - Step 28218: {'lr': 0.00021383539307093344, 'samples': 14448128, 'steps': 28218, 'loss/train': 2.0621285438537598} 02/25/2022 14:09:17 - INFO - codeparrot_training - Step 28219: {'lr': 0.00021381920279273597, 'samples': 14448640, 'steps': 28219, 'loss/train': 1.7898664474487305} 02/25/2022 14:09:21 - INFO - codeparrot_training - Step 28220: {'lr': 0.00021380301266952557, 'samples': 14449152, 'steps': 28220, 'loss/train': 2.425165891647339} 02/25/2022 14:09:24 - INFO - codeparrot_training - Step 28221: {'lr': 0.00021378682270137156, 'samples': 14449664, 'steps': 28221, 'loss/train': 1.5211080312728882} 02/25/2022 14:09:30 - INFO - codeparrot_training - Step 28222: {'lr': 0.0002137706328883434, 'samples': 14450176, 'steps': 28222, 'loss/train': 1.5055103302001953} 02/25/2022 14:09:33 - INFO - codeparrot_training - Step 28223: {'lr': 0.0002137544432305103, 'samples': 14450688, 'steps': 28223, 'loss/train': 1.73633873462677} 02/25/2022 14:09:39 - INFO - codeparrot_training - Step 28224: {'lr': 0.0002137382537279416, 'samples': 14451200, 'steps': 28224, 'loss/train': 1.9062696695327759} 02/25/2022 14:09:42 - INFO - codeparrot_training - Step 28225: {'lr': 0.00021372206438070672, 'samples': 14451712, 'steps': 28225, 'loss/train': 1.1663005352020264} 02/25/2022 14:09:48 - INFO - codeparrot_training - Step 28226: {'lr': 0.000213705875188875, 'samples': 14452224, 'steps': 28226, 'loss/train': 2.040790319442749} 02/25/2022 14:09:51 - INFO - codeparrot_training - Step 28227: {'lr': 0.0002136896861525159, 'samples': 14452736, 'steps': 28227, 'loss/train': 1.808836579322815} 02/25/2022 14:09:57 - INFO - codeparrot_training - Step 28228: {'lr': 0.0002136734972716985, 'samples': 14453248, 'steps': 28228, 'loss/train': 1.0167431831359863} 02/25/2022 14:10:00 - INFO - codeparrot_training - Step 28229: {'lr': 0.00021365730854649236, 'samples': 14453760, 'steps': 28229, 'loss/train': 1.3448762893676758} 02/25/2022 14:10:07 - INFO - codeparrot_training - Step 28230: {'lr': 0.0002136411199769667, 'samples': 14454272, 'steps': 28230, 'loss/train': 1.4268722534179688} 02/25/2022 14:10:10 - INFO - codeparrot_training - Step 28231: {'lr': 0.00021362493156319113, 'samples': 14454784, 'steps': 28231, 'loss/train': 1.303395390510559} 02/25/2022 14:10:16 - INFO - codeparrot_training - Step 28232: {'lr': 0.00021360874330523467, 'samples': 14455296, 'steps': 28232, 'loss/train': 1.3476988077163696} 02/25/2022 14:10:19 - INFO - codeparrot_training - Step 28233: {'lr': 0.00021359255520316675, 'samples': 14455808, 'steps': 28233, 'loss/train': 1.2929127216339111} 02/25/2022 14:10:24 - INFO - codeparrot_training - Step 28234: {'lr': 0.00021357636725705687, 'samples': 14456320, 'steps': 28234, 'loss/train': 1.5819498300552368} 02/25/2022 14:10:30 - INFO - codeparrot_training - Step 28235: {'lr': 0.00021356017946697425, 'samples': 14456832, 'steps': 28235, 'loss/train': 1.2295715808868408} 02/25/2022 14:10:33 - INFO - codeparrot_training - Step 28236: {'lr': 0.0002135439918329882, 'samples': 14457344, 'steps': 28236, 'loss/train': 2.0276858806610107} 02/25/2022 14:10:39 - INFO - codeparrot_training - Step 28237: {'lr': 0.00021352780435516811, 'samples': 14457856, 'steps': 28237, 'loss/train': 0.05608693137764931} 02/25/2022 14:10:42 - INFO - codeparrot_training - Step 28238: {'lr': 0.00021351161703358335, 'samples': 14458368, 'steps': 28238, 'loss/train': 2.012699842453003} 02/25/2022 14:10:49 - INFO - codeparrot_training - Step 28239: {'lr': 0.0002134954298683032, 'samples': 14458880, 'steps': 28239, 'loss/train': 1.3082733154296875} 02/25/2022 14:10:53 - INFO - codeparrot_training - Step 28240: {'lr': 0.00021347924285939712, 'samples': 14459392, 'steps': 28240, 'loss/train': 1.3898792266845703} 02/25/2022 14:10:56 - INFO - codeparrot_training - Step 28241: {'lr': 0.00021346305600693433, 'samples': 14459904, 'steps': 28241, 'loss/train': 2.0165929794311523} 02/25/2022 14:11:02 - INFO - codeparrot_training - Step 28242: {'lr': 0.00021344686931098418, 'samples': 14460416, 'steps': 28242, 'loss/train': 1.545802354812622} 02/25/2022 14:11:07 - INFO - codeparrot_training - Step 28243: {'lr': 0.00021343068277161604, 'samples': 14460928, 'steps': 28243, 'loss/train': 2.3445956707000732} 02/25/2022 14:11:11 - INFO - codeparrot_training - Step 28244: {'lr': 0.00021341449638889926, 'samples': 14461440, 'steps': 28244, 'loss/train': 0.15547709167003632} 02/25/2022 14:11:16 - INFO - codeparrot_training - Step 28245: {'lr': 0.00021339831016290318, 'samples': 14461952, 'steps': 28245, 'loss/train': 3.4157631397247314} 02/25/2022 14:11:20 - INFO - codeparrot_training - Step 28246: {'lr': 0.00021338212409369707, 'samples': 14462464, 'steps': 28246, 'loss/train': 3.1715946197509766} 02/25/2022 14:11:25 - INFO - codeparrot_training - Step 28247: {'lr': 0.00021336593818135032, 'samples': 14462976, 'steps': 28247, 'loss/train': 2.5878636837005615} 02/25/2022 14:11:29 - INFO - codeparrot_training - Step 28248: {'lr': 0.00021334975242593234, 'samples': 14463488, 'steps': 28248, 'loss/train': 1.7279274463653564} 02/25/2022 14:11:32 - INFO - codeparrot_training - Step 28249: {'lr': 0.0002133335668275123, 'samples': 14464000, 'steps': 28249, 'loss/train': 1.9622348546981812} 02/25/2022 14:11:38 - INFO - codeparrot_training - Step 28250: {'lr': 0.00021331738138615958, 'samples': 14464512, 'steps': 28250, 'loss/train': 0.05712484195828438} 02/25/2022 14:11:41 - INFO - codeparrot_training - Step 28251: {'lr': 0.0002133011961019436, 'samples': 14465024, 'steps': 28251, 'loss/train': 1.6140528917312622} 02/25/2022 14:11:48 - INFO - codeparrot_training - Step 28252: {'lr': 0.0002132850109749336, 'samples': 14465536, 'steps': 28252, 'loss/train': 1.1634860038757324} 02/25/2022 14:11:53 - INFO - codeparrot_training - Step 28253: {'lr': 0.0002132688260051991, 'samples': 14466048, 'steps': 28253, 'loss/train': 2.5930566787719727} 02/25/2022 14:11:56 - INFO - codeparrot_training - Step 28254: {'lr': 0.00021325264119280914, 'samples': 14466560, 'steps': 28254, 'loss/train': 1.196531891822815} 02/25/2022 14:12:02 - INFO - codeparrot_training - Step 28255: {'lr': 0.00021323645653783323, 'samples': 14467072, 'steps': 28255, 'loss/train': 1.8630634546279907} 02/25/2022 14:12:05 - INFO - codeparrot_training - Step 28256: {'lr': 0.00021322027204034063, 'samples': 14467584, 'steps': 28256, 'loss/train': 2.4226880073547363} 02/25/2022 14:12:11 - INFO - codeparrot_training - Step 28257: {'lr': 0.00021320408770040083, 'samples': 14468096, 'steps': 28257, 'loss/train': 1.9604405164718628} 02/25/2022 14:12:14 - INFO - codeparrot_training - Step 28258: {'lr': 0.00021318790351808296, 'samples': 14468608, 'steps': 28258, 'loss/train': 1.3375805616378784} 02/25/2022 14:12:20 - INFO - codeparrot_training - Step 28259: {'lr': 0.00021317171949345636, 'samples': 14469120, 'steps': 28259, 'loss/train': 0.15957719087600708} 02/25/2022 14:12:23 - INFO - codeparrot_training - Step 28260: {'lr': 0.00021315553562659045, 'samples': 14469632, 'steps': 28260, 'loss/train': 1.3576756715774536} 02/25/2022 14:12:29 - INFO - codeparrot_training - Step 28261: {'lr': 0.00021313935191755466, 'samples': 14470144, 'steps': 28261, 'loss/train': 1.406958818435669} 02/25/2022 14:12:32 - INFO - codeparrot_training - Step 28262: {'lr': 0.00021312316836641804, 'samples': 14470656, 'steps': 28262, 'loss/train': 2.1285762786865234} 02/25/2022 14:12:38 - INFO - codeparrot_training - Step 28263: {'lr': 0.00021310698497325006, 'samples': 14471168, 'steps': 28263, 'loss/train': 1.821584701538086} 02/25/2022 14:12:41 - INFO - codeparrot_training - Step 28264: {'lr': 0.00021309080173812007, 'samples': 14471680, 'steps': 28264, 'loss/train': 1.6576480865478516} 02/25/2022 14:12:47 - INFO - codeparrot_training - Step 28265: {'lr': 0.00021307461866109734, 'samples': 14472192, 'steps': 28265, 'loss/train': 1.8209234476089478} 02/25/2022 14:12:51 - INFO - codeparrot_training - Step 28266: {'lr': 0.00021305843574225133, 'samples': 14472704, 'steps': 28266, 'loss/train': 0.9990397691726685} 02/25/2022 14:12:56 - INFO - codeparrot_training - Step 28267: {'lr': 0.00021304225298165113, 'samples': 14473216, 'steps': 28267, 'loss/train': 2.634413719177246} 02/25/2022 14:13:00 - INFO - codeparrot_training - Step 28268: {'lr': 0.0002130260703793662, 'samples': 14473728, 'steps': 28268, 'loss/train': 2.0726587772369385} 02/25/2022 14:13:05 - INFO - codeparrot_training - Step 28269: {'lr': 0.00021300988793546586, 'samples': 14474240, 'steps': 28269, 'loss/train': 0.06204100698232651} 02/25/2022 14:13:09 - INFO - codeparrot_training - Step 28270: {'lr': 0.00021299370565001948, 'samples': 14474752, 'steps': 28270, 'loss/train': 1.9901347160339355} 02/25/2022 14:13:14 - INFO - codeparrot_training - Step 28271: {'lr': 0.00021297752352309623, 'samples': 14475264, 'steps': 28271, 'loss/train': 2.7706313133239746} 02/25/2022 14:13:18 - INFO - codeparrot_training - Step 28272: {'lr': 0.00021296134155476552, 'samples': 14475776, 'steps': 28272, 'loss/train': 0.9279763698577881} 02/25/2022 14:13:23 - INFO - codeparrot_training - Step 28273: {'lr': 0.00021294515974509666, 'samples': 14476288, 'steps': 28273, 'loss/train': 2.64377760887146} 02/25/2022 14:13:26 - INFO - codeparrot_training - Step 28274: {'lr': 0.00021292897809415906, 'samples': 14476800, 'steps': 28274, 'loss/train': 2.3142173290252686} 02/25/2022 14:13:33 - INFO - codeparrot_training - Step 28275: {'lr': 0.00021291279660202187, 'samples': 14477312, 'steps': 28275, 'loss/train': 1.709380865097046} 02/25/2022 14:13:36 - INFO - codeparrot_training - Step 28276: {'lr': 0.00021289661526875445, 'samples': 14477824, 'steps': 28276, 'loss/train': 2.758554458618164} 02/25/2022 14:13:42 - INFO - codeparrot_training - Step 28277: {'lr': 0.00021288043409442616, 'samples': 14478336, 'steps': 28277, 'loss/train': 0.8211163878440857} 02/25/2022 14:13:45 - INFO - codeparrot_training - Step 28278: {'lr': 0.00021286425307910636, 'samples': 14478848, 'steps': 28278, 'loss/train': 1.3632938861846924} 02/25/2022 14:13:50 - INFO - codeparrot_training - Step 28279: {'lr': 0.00021284807222286432, 'samples': 14479360, 'steps': 28279, 'loss/train': 0.996660590171814} 02/25/2022 14:13:54 - INFO - codeparrot_training - Step 28280: {'lr': 0.00021283189152576927, 'samples': 14479872, 'steps': 28280, 'loss/train': 1.7499308586120605} 02/25/2022 14:13:59 - INFO - codeparrot_training - Step 28281: {'lr': 0.00021281571098789065, 'samples': 14480384, 'steps': 28281, 'loss/train': 0.7764137387275696} 02/25/2022 14:14:03 - INFO - codeparrot_training - Step 28282: {'lr': 0.00021279953060929766, 'samples': 14480896, 'steps': 28282, 'loss/train': 2.001917600631714} 02/25/2022 14:14:08 - INFO - codeparrot_training - Step 28283: {'lr': 0.0002127833503900597, 'samples': 14481408, 'steps': 28283, 'loss/train': 2.2516567707061768} 02/25/2022 14:14:12 - INFO - codeparrot_training - Step 28284: {'lr': 0.00021276717033024607, 'samples': 14481920, 'steps': 28284, 'loss/train': 2.323160409927368} 02/25/2022 14:14:19 - INFO - codeparrot_training - Step 28285: {'lr': 0.00021275099042992606, 'samples': 14482432, 'steps': 28285, 'loss/train': 1.4083870649337769} 02/25/2022 14:14:22 - INFO - codeparrot_training - Step 28286: {'lr': 0.0002127348106891689, 'samples': 14482944, 'steps': 28286, 'loss/train': 1.5999479293823242} 02/25/2022 14:14:27 - INFO - codeparrot_training - Step 28287: {'lr': 0.000212718631108044, 'samples': 14483456, 'steps': 28287, 'loss/train': 1.4723689556121826} 02/25/2022 14:14:33 - INFO - codeparrot_training - Step 28288: {'lr': 0.00021270245168662077, 'samples': 14483968, 'steps': 28288, 'loss/train': 2.1287407875061035} 02/25/2022 14:14:36 - INFO - codeparrot_training - Step 28289: {'lr': 0.0002126862724249683, 'samples': 14484480, 'steps': 28289, 'loss/train': 1.5229589939117432} 02/25/2022 14:14:42 - INFO - codeparrot_training - Step 28290: {'lr': 0.000212670093323156, 'samples': 14484992, 'steps': 28290, 'loss/train': 1.637123465538025} 02/25/2022 14:14:46 - INFO - codeparrot_training - Step 28291: {'lr': 0.00021265391438125313, 'samples': 14485504, 'steps': 28291, 'loss/train': 1.69315767288208} 02/25/2022 14:14:49 - INFO - codeparrot_training - Step 28292: {'lr': 0.00021263773559932915, 'samples': 14486016, 'steps': 28292, 'loss/train': 1.6702494621276855} 02/25/2022 14:14:54 - INFO - codeparrot_training - Step 28293: {'lr': 0.00021262155697745311, 'samples': 14486528, 'steps': 28293, 'loss/train': 1.4113856554031372} 02/25/2022 14:14:58 - INFO - codeparrot_training - Step 28294: {'lr': 0.00021260537851569447, 'samples': 14487040, 'steps': 28294, 'loss/train': 1.4689005613327026} 02/25/2022 14:15:03 - INFO - codeparrot_training - Step 28295: {'lr': 0.00021258920021412252, 'samples': 14487552, 'steps': 28295, 'loss/train': 0.07687009871006012} 02/25/2022 14:15:10 - INFO - codeparrot_training - Step 28296: {'lr': 0.0002125730220728067, 'samples': 14488064, 'steps': 28296, 'loss/train': 1.0091983079910278} 02/25/2022 14:15:13 - INFO - codeparrot_training - Step 28297: {'lr': 0.000212556844091816, 'samples': 14488576, 'steps': 28297, 'loss/train': 1.4685746431350708} 02/25/2022 14:15:19 - INFO - codeparrot_training - Step 28298: {'lr': 0.00021254066627121994, 'samples': 14489088, 'steps': 28298, 'loss/train': 1.0950254201889038} 02/25/2022 14:15:22 - INFO - codeparrot_training - Step 28299: {'lr': 0.00021252448861108772, 'samples': 14489600, 'steps': 28299, 'loss/train': 2.378859043121338} 02/25/2022 14:15:28 - INFO - codeparrot_training - Step 28300: {'lr': 0.0002125083111114887, 'samples': 14490112, 'steps': 28300, 'loss/train': 0.11248381435871124} 02/25/2022 14:15:31 - INFO - codeparrot_training - Step 28301: {'lr': 0.00021249213377249232, 'samples': 14490624, 'steps': 28301, 'loss/train': 0.8231486678123474} 02/25/2022 14:15:37 - INFO - codeparrot_training - Step 28302: {'lr': 0.0002124759565941676, 'samples': 14491136, 'steps': 28302, 'loss/train': 2.3650870323181152} 02/25/2022 14:15:40 - INFO - codeparrot_training - Step 28303: {'lr': 0.00021245977957658397, 'samples': 14491648, 'steps': 28303, 'loss/train': 1.709525227546692} 02/25/2022 14:15:46 - INFO - codeparrot_training - Step 28304: {'lr': 0.00021244360271981073, 'samples': 14492160, 'steps': 28304, 'loss/train': 1.2625675201416016} 02/25/2022 14:15:49 - INFO - codeparrot_training - Step 28305: {'lr': 0.00021242742602391727, 'samples': 14492672, 'steps': 28305, 'loss/train': 2.158215045928955} 02/25/2022 14:15:56 - INFO - codeparrot_training - Step 28306: {'lr': 0.00021241124948897266, 'samples': 14493184, 'steps': 28306, 'loss/train': 1.7731049060821533} 02/25/2022 14:15:59 - INFO - codeparrot_training - Step 28307: {'lr': 0.00021239507311504634, 'samples': 14493696, 'steps': 28307, 'loss/train': 2.3433847427368164} 02/25/2022 14:16:05 - INFO - codeparrot_training - Step 28308: {'lr': 0.00021237889690220758, 'samples': 14494208, 'steps': 28308, 'loss/train': 0.08356847614049911} 02/25/2022 14:16:08 - INFO - codeparrot_training - Step 28309: {'lr': 0.0002123627208505258, 'samples': 14494720, 'steps': 28309, 'loss/train': 2.075960159301758} 02/25/2022 14:16:14 - INFO - codeparrot_training - Step 28310: {'lr': 0.00021234654496007005, 'samples': 14495232, 'steps': 28310, 'loss/train': 1.9082318544387817} 02/25/2022 14:16:17 - INFO - codeparrot_training - Step 28311: {'lr': 0.00021233036923090976, 'samples': 14495744, 'steps': 28311, 'loss/train': 2.581702709197998} 02/25/2022 14:16:23 - INFO - codeparrot_training - Step 28312: {'lr': 0.00021231419366311418, 'samples': 14496256, 'steps': 28312, 'loss/train': 1.8835728168487549} 02/25/2022 14:16:26 - INFO - codeparrot_training - Step 28313: {'lr': 0.00021229801825675267, 'samples': 14496768, 'steps': 28313, 'loss/train': 1.6758482456207275} 02/25/2022 14:16:32 - INFO - codeparrot_training - Step 28314: {'lr': 0.00021228184301189454, 'samples': 14497280, 'steps': 28314, 'loss/train': 1.2920202016830444} 02/25/2022 14:16:35 - INFO - codeparrot_training - Step 28315: {'lr': 0.00021226566792860893, 'samples': 14497792, 'steps': 28315, 'loss/train': 0.6129186749458313} 02/25/2022 14:16:41 - INFO - codeparrot_training - Step 28316: {'lr': 0.00021224949300696522, 'samples': 14498304, 'steps': 28316, 'loss/train': 1.807091474533081} 02/25/2022 14:16:44 - INFO - codeparrot_training - Step 28317: {'lr': 0.0002122333182470327, 'samples': 14498816, 'steps': 28317, 'loss/train': 1.946730613708496} 02/25/2022 14:16:50 - INFO - codeparrot_training - Step 28318: {'lr': 0.00021221714364888072, 'samples': 14499328, 'steps': 28318, 'loss/train': 1.6418988704681396} 02/25/2022 14:16:53 - INFO - codeparrot_training - Step 28319: {'lr': 0.00021220096921257845, 'samples': 14499840, 'steps': 28319, 'loss/train': 2.0740702152252197} 02/25/2022 14:16:59 - INFO - codeparrot_training - Step 28320: {'lr': 0.0002121847949381952, 'samples': 14500352, 'steps': 28320, 'loss/train': 1.0231614112854004} 02/25/2022 14:17:02 - INFO - codeparrot_training - Step 28321: {'lr': 0.00021216862082580032, 'samples': 14500864, 'steps': 28321, 'loss/train': 1.1979848146438599} 02/25/2022 14:17:08 - INFO - codeparrot_training - Step 28322: {'lr': 0.00021215244687546308, 'samples': 14501376, 'steps': 28322, 'loss/train': 1.2794530391693115} 02/25/2022 14:17:12 - INFO - codeparrot_training - Step 28323: {'lr': 0.00021213627308725266, 'samples': 14501888, 'steps': 28323, 'loss/train': 1.1398054361343384} 02/25/2022 14:17:17 - INFO - codeparrot_training - Step 28324: {'lr': 0.00021212009946123845, 'samples': 14502400, 'steps': 28324, 'loss/train': 1.1762428283691406} 02/25/2022 14:17:21 - INFO - codeparrot_training - Step 28325: {'lr': 0.00021210392599748973, 'samples': 14502912, 'steps': 28325, 'loss/train': 1.7300468683242798} 02/25/2022 14:17:26 - INFO - codeparrot_training - Step 28326: {'lr': 0.00021208775269607576, 'samples': 14503424, 'steps': 28326, 'loss/train': 1.592652440071106} 02/25/2022 14:17:30 - INFO - codeparrot_training - Step 28327: {'lr': 0.0002120715795570658, 'samples': 14503936, 'steps': 28327, 'loss/train': 2.032846212387085} 02/25/2022 14:17:35 - INFO - codeparrot_training - Step 28328: {'lr': 0.00021205540658052912, 'samples': 14504448, 'steps': 28328, 'loss/train': 1.2168664932250977} 02/25/2022 14:17:39 - INFO - codeparrot_training - Step 28329: {'lr': 0.00021203923376653512, 'samples': 14504960, 'steps': 28329, 'loss/train': 1.053036093711853} 02/25/2022 14:17:44 - INFO - codeparrot_training - Step 28330: {'lr': 0.0002120230611151529, 'samples': 14505472, 'steps': 28330, 'loss/train': 2.599642276763916} 02/25/2022 14:17:48 - INFO - codeparrot_training - Step 28331: {'lr': 0.0002120068886264519, 'samples': 14505984, 'steps': 28331, 'loss/train': 2.286259889602661} 02/25/2022 14:17:54 - INFO - codeparrot_training - Step 28332: {'lr': 0.00021199071630050132, 'samples': 14506496, 'steps': 28332, 'loss/train': 1.5866584777832031} 02/25/2022 14:17:59 - INFO - codeparrot_training - Step 28333: {'lr': 0.0002119745441373704, 'samples': 14507008, 'steps': 28333, 'loss/train': 2.208592653274536} 02/25/2022 14:18:03 - INFO - codeparrot_training - Step 28334: {'lr': 0.0002119583721371284, 'samples': 14507520, 'steps': 28334, 'loss/train': 2.1119301319122314} 02/25/2022 14:18:06 - INFO - codeparrot_training - Step 28335: {'lr': 0.0002119422002998447, 'samples': 14508032, 'steps': 28335, 'loss/train': 2.3421809673309326} 02/25/2022 14:18:12 - INFO - codeparrot_training - Step 28336: {'lr': 0.00021192602862558864, 'samples': 14508544, 'steps': 28336, 'loss/train': 0.6300588250160217} 02/25/2022 14:18:17 - INFO - codeparrot_training - Step 28337: {'lr': 0.00021190985711442924, 'samples': 14509056, 'steps': 28337, 'loss/train': 1.2629461288452148} 02/25/2022 14:18:21 - INFO - codeparrot_training - Step 28338: {'lr': 0.00021189368576643599, 'samples': 14509568, 'steps': 28338, 'loss/train': 1.7022731304168701} 02/25/2022 14:18:26 - INFO - codeparrot_training - Step 28339: {'lr': 0.00021187751458167804, 'samples': 14510080, 'steps': 28339, 'loss/train': 1.6350210905075073} 02/25/2022 14:18:30 - INFO - codeparrot_training - Step 28340: {'lr': 0.0002118613435602248, 'samples': 14510592, 'steps': 28340, 'loss/train': 1.8421428203582764} 02/25/2022 14:18:36 - INFO - codeparrot_training - Step 28341: {'lr': 0.00021184517270214537, 'samples': 14511104, 'steps': 28341, 'loss/train': 2.856041431427002} 02/25/2022 14:18:39 - INFO - codeparrot_training - Step 28342: {'lr': 0.0002118290020075091, 'samples': 14511616, 'steps': 28342, 'loss/train': 1.5928322076797485} 02/25/2022 14:18:43 - INFO - codeparrot_training - Step 28343: {'lr': 0.00021181283147638527, 'samples': 14512128, 'steps': 28343, 'loss/train': 0.04897914454340935} 02/25/2022 14:18:48 - INFO - codeparrot_training - Step 28344: {'lr': 0.00021179666110884328, 'samples': 14512640, 'steps': 28344, 'loss/train': 2.2004902362823486} 02/25/2022 14:18:54 - INFO - codeparrot_training - Step 28345: {'lr': 0.00021178049090495211, 'samples': 14513152, 'steps': 28345, 'loss/train': 1.6638797521591187} 02/25/2022 14:18:57 - INFO - codeparrot_training - Step 28346: {'lr': 0.0002117643208647812, 'samples': 14513664, 'steps': 28346, 'loss/train': 0.49903514981269836} 02/25/2022 14:19:03 - INFO - codeparrot_training - Step 28347: {'lr': 0.00021174815098839978, 'samples': 14514176, 'steps': 28347, 'loss/train': 1.2643673419952393} 02/25/2022 14:19:06 - INFO - codeparrot_training - Step 28348: {'lr': 0.00021173198127587717, 'samples': 14514688, 'steps': 28348, 'loss/train': 1.9727510213851929} 02/25/2022 14:19:12 - INFO - codeparrot_training - Step 28349: {'lr': 0.00021171581172728268, 'samples': 14515200, 'steps': 28349, 'loss/train': 1.769645094871521} 02/25/2022 14:19:15 - INFO - codeparrot_training - Step 28350: {'lr': 0.0002116996423426854, 'samples': 14515712, 'steps': 28350, 'loss/train': 0.595661461353302} 02/25/2022 14:19:21 - INFO - codeparrot_training - Step 28351: {'lr': 0.00021168347312215468, 'samples': 14516224, 'steps': 28351, 'loss/train': 0.16420549154281616} 02/25/2022 14:19:25 - INFO - codeparrot_training - Step 28352: {'lr': 0.0002116673040657598, 'samples': 14516736, 'steps': 28352, 'loss/train': 1.470096230506897} 02/25/2022 14:19:30 - INFO - codeparrot_training - Step 28353: {'lr': 0.00021165113517357016, 'samples': 14517248, 'steps': 28353, 'loss/train': 1.6985714435577393} 02/25/2022 14:19:34 - INFO - codeparrot_training - Step 28354: {'lr': 0.00021163496644565472, 'samples': 14517760, 'steps': 28354, 'loss/train': 1.5188720226287842} 02/25/2022 14:19:39 - INFO - codeparrot_training - Step 28355: {'lr': 0.00021161879788208295, 'samples': 14518272, 'steps': 28355, 'loss/train': 1.8870947360992432} 02/25/2022 14:19:43 - INFO - codeparrot_training - Step 28356: {'lr': 0.000211602629482924, 'samples': 14518784, 'steps': 28356, 'loss/train': 0.9333053231239319} 02/25/2022 14:19:48 - INFO - codeparrot_training - Step 28357: {'lr': 0.00021158646124824735, 'samples': 14519296, 'steps': 28357, 'loss/train': 1.6203453540802002} 02/25/2022 14:19:52 - INFO - codeparrot_training - Step 28358: {'lr': 0.00021157029317812198, 'samples': 14519808, 'steps': 28358, 'loss/train': 2.8714053630828857} 02/25/2022 14:19:57 - INFO - codeparrot_training - Step 28359: {'lr': 0.00021155412527261726, 'samples': 14520320, 'steps': 28359, 'loss/train': 1.8954418897628784} 02/25/2022 14:20:01 - INFO - codeparrot_training - Step 28360: {'lr': 0.00021153795753180247, 'samples': 14520832, 'steps': 28360, 'loss/train': 0.6464634537696838} 02/25/2022 14:20:06 - INFO - codeparrot_training - Step 28361: {'lr': 0.00021152178995574686, 'samples': 14521344, 'steps': 28361, 'loss/train': 1.4953765869140625} 02/25/2022 14:20:10 - INFO - codeparrot_training - Step 28362: {'lr': 0.00021150562254451978, 'samples': 14521856, 'steps': 28362, 'loss/train': 1.9135023355484009} 02/25/2022 14:20:15 - INFO - codeparrot_training - Step 28363: {'lr': 0.0002114894552981903, 'samples': 14522368, 'steps': 28363, 'loss/train': 2.067519187927246} 02/25/2022 14:20:19 - INFO - codeparrot_training - Step 28364: {'lr': 0.00021147328821682776, 'samples': 14522880, 'steps': 28364, 'loss/train': 1.184251070022583} 02/25/2022 14:20:24 - INFO - codeparrot_training - Step 28365: {'lr': 0.00021145712130050145, 'samples': 14523392, 'steps': 28365, 'loss/train': 1.4790740013122559} 02/25/2022 14:20:28 - INFO - codeparrot_training - Step 28366: {'lr': 0.00021144095454928062, 'samples': 14523904, 'steps': 28366, 'loss/train': 1.4340696334838867} 02/25/2022 14:20:34 - INFO - codeparrot_training - Step 28367: {'lr': 0.00021142478796323444, 'samples': 14524416, 'steps': 28367, 'loss/train': 1.3393166065216064} 02/25/2022 14:20:37 - INFO - codeparrot_training - Step 28368: {'lr': 0.0002114086215424322, 'samples': 14524928, 'steps': 28368, 'loss/train': 1.8404030799865723} 02/25/2022 14:20:43 - INFO - codeparrot_training - Step 28369: {'lr': 0.00021139245528694324, 'samples': 14525440, 'steps': 28369, 'loss/train': 1.5250625610351562} 02/25/2022 14:20:46 - INFO - codeparrot_training - Step 28370: {'lr': 0.00021137628919683674, 'samples': 14525952, 'steps': 28370, 'loss/train': 1.276877760887146} 02/25/2022 14:20:52 - INFO - codeparrot_training - Step 28371: {'lr': 0.00021136012327218191, 'samples': 14526464, 'steps': 28371, 'loss/train': 1.39301335811615} 02/25/2022 14:20:55 - INFO - codeparrot_training - Step 28372: {'lr': 0.00021134395751304808, 'samples': 14526976, 'steps': 28372, 'loss/train': 1.5065416097640991} 02/25/2022 14:21:01 - INFO - codeparrot_training - Step 28373: {'lr': 0.0002113277919195044, 'samples': 14527488, 'steps': 28373, 'loss/train': 1.7488592863082886} 02/25/2022 14:21:04 - INFO - codeparrot_training - Step 28374: {'lr': 0.00021131162649162022, 'samples': 14528000, 'steps': 28374, 'loss/train': 0.4177989959716797} 02/25/2022 14:21:10 - INFO - codeparrot_training - Step 28375: {'lr': 0.00021129546122946477, 'samples': 14528512, 'steps': 28375, 'loss/train': 1.301292061805725} 02/25/2022 14:21:13 - INFO - codeparrot_training - Step 28376: {'lr': 0.00021127929613310725, 'samples': 14529024, 'steps': 28376, 'loss/train': 1.3771742582321167} 02/25/2022 14:21:19 - INFO - codeparrot_training - Step 28377: {'lr': 0.0002112631312026169, 'samples': 14529536, 'steps': 28377, 'loss/train': 1.7349754571914673} 02/25/2022 14:21:23 - INFO - codeparrot_training - Step 28378: {'lr': 0.00021124696643806302, 'samples': 14530048, 'steps': 28378, 'loss/train': 2.1916744709014893} 02/25/2022 14:21:28 - INFO - codeparrot_training - Step 28379: {'lr': 0.00021123080183951492, 'samples': 14530560, 'steps': 28379, 'loss/train': 1.386434555053711} 02/25/2022 14:21:32 - INFO - codeparrot_training - Step 28380: {'lr': 0.00021121463740704166, 'samples': 14531072, 'steps': 28380, 'loss/train': 1.5551437139511108} 02/25/2022 14:21:37 - INFO - codeparrot_training - Step 28381: {'lr': 0.00021119847314071254, 'samples': 14531584, 'steps': 28381, 'loss/train': 0.06534381210803986} 02/25/2022 14:21:41 - INFO - codeparrot_training - Step 28382: {'lr': 0.00021118230904059688, 'samples': 14532096, 'steps': 28382, 'loss/train': 1.5706872940063477} 02/25/2022 14:21:46 - INFO - codeparrot_training - Step 28383: {'lr': 0.00021116614510676397, 'samples': 14532608, 'steps': 28383, 'loss/train': 1.1185634136199951} 02/25/2022 14:21:50 - INFO - codeparrot_training - Step 28384: {'lr': 0.00021114998133928286, 'samples': 14533120, 'steps': 28384, 'loss/train': 1.580344796180725} 02/25/2022 14:21:55 - INFO - codeparrot_training - Step 28385: {'lr': 0.00021113381773822288, 'samples': 14533632, 'steps': 28385, 'loss/train': 1.5345349311828613} 02/25/2022 14:22:02 - INFO - codeparrot_training - Step 28386: {'lr': 0.0002111176543036533, 'samples': 14534144, 'steps': 28386, 'loss/train': 1.2951834201812744} 02/25/2022 14:22:05 - INFO - codeparrot_training - Step 28387: {'lr': 0.00021110149103564335, 'samples': 14534656, 'steps': 28387, 'loss/train': 2.4283063411712646} 02/25/2022 14:22:11 - INFO - codeparrot_training - Step 28388: {'lr': 0.00021108532793426236, 'samples': 14535168, 'steps': 28388, 'loss/train': 1.3708207607269287} 02/25/2022 14:22:14 - INFO - codeparrot_training - Step 28389: {'lr': 0.00021106916499957936, 'samples': 14535680, 'steps': 28389, 'loss/train': 2.2369604110717773} 02/25/2022 14:22:19 - INFO - codeparrot_training - Step 28390: {'lr': 0.0002110530022316637, 'samples': 14536192, 'steps': 28390, 'loss/train': 1.4980615377426147} 02/25/2022 14:22:23 - INFO - codeparrot_training - Step 28391: {'lr': 0.00021103683963058457, 'samples': 14536704, 'steps': 28391, 'loss/train': 1.6785829067230225} 02/25/2022 14:22:28 - INFO - codeparrot_training - Step 28392: {'lr': 0.0002110206771964114, 'samples': 14537216, 'steps': 28392, 'loss/train': 2.1099720001220703} 02/25/2022 14:22:32 - INFO - codeparrot_training - Step 28393: {'lr': 0.00021100451492921316, 'samples': 14537728, 'steps': 28393, 'loss/train': 1.256303310394287} 02/25/2022 14:22:38 - INFO - codeparrot_training - Step 28394: {'lr': 0.0002109883528290592, 'samples': 14538240, 'steps': 28394, 'loss/train': 8.790576934814453} 02/25/2022 14:22:41 - INFO - codeparrot_training - Step 28395: {'lr': 0.00021097219089601872, 'samples': 14538752, 'steps': 28395, 'loss/train': 1.1408854722976685} 02/25/2022 14:22:47 - INFO - codeparrot_training - Step 28396: {'lr': 0.00021095602913016097, 'samples': 14539264, 'steps': 28396, 'loss/train': 0.04697303846478462} 02/25/2022 14:22:50 - INFO - codeparrot_training - Step 28397: {'lr': 0.00021093986753155538, 'samples': 14539776, 'steps': 28397, 'loss/train': 1.5680654048919678} 02/25/2022 14:22:56 - INFO - codeparrot_training - Step 28398: {'lr': 0.00021092370610027082, 'samples': 14540288, 'steps': 28398, 'loss/train': 1.721678376197815} 02/25/2022 14:23:00 - INFO - codeparrot_training - Step 28399: {'lr': 0.0002109075448363767, 'samples': 14540800, 'steps': 28399, 'loss/train': 2.1614668369293213} 02/25/2022 14:23:05 - INFO - codeparrot_training - Step 28400: {'lr': 0.00021089138373994224, 'samples': 14541312, 'steps': 28400, 'loss/train': 1.2734172344207764} 02/25/2022 14:23:09 - INFO - codeparrot_training - Step 28401: {'lr': 0.0002108752228110368, 'samples': 14541824, 'steps': 28401, 'loss/train': 2.3260772228240967} 02/25/2022 14:23:14 - INFO - codeparrot_training - Step 28402: {'lr': 0.00021085906204972938, 'samples': 14542336, 'steps': 28402, 'loss/train': 2.5203962326049805} 02/25/2022 14:23:18 - INFO - codeparrot_training - Step 28403: {'lr': 0.0002108429014560893, 'samples': 14542848, 'steps': 28403, 'loss/train': 1.7350388765335083} 02/25/2022 14:23:23 - INFO - codeparrot_training - Step 28404: {'lr': 0.00021082674103018579, 'samples': 14543360, 'steps': 28404, 'loss/train': 1.0933187007904053} 02/25/2022 14:23:27 - INFO - codeparrot_training - Step 28405: {'lr': 0.0002108105807720882, 'samples': 14543872, 'steps': 28405, 'loss/train': 1.4100477695465088} 02/25/2022 14:23:32 - INFO - codeparrot_training - Step 28406: {'lr': 0.00021079442068186552, 'samples': 14544384, 'steps': 28406, 'loss/train': 2.218083620071411} 02/25/2022 14:23:36 - INFO - codeparrot_training - Step 28407: {'lr': 0.0002107782607595871, 'samples': 14544896, 'steps': 28407, 'loss/train': 2.2843759059906006} 02/25/2022 14:23:41 - INFO - codeparrot_training - Step 28408: {'lr': 0.00021076210100532216, 'samples': 14545408, 'steps': 28408, 'loss/train': 1.5812772512435913} 02/25/2022 14:23:45 - INFO - codeparrot_training - Step 28409: {'lr': 0.00021074594141913997, 'samples': 14545920, 'steps': 28409, 'loss/train': 2.3902242183685303} 02/25/2022 14:23:50 - INFO - codeparrot_training - Step 28410: {'lr': 0.0002107297820011097, 'samples': 14546432, 'steps': 28410, 'loss/train': 1.7851753234863281} 02/25/2022 14:23:54 - INFO - codeparrot_training - Step 28411: {'lr': 0.00021071362275130052, 'samples': 14546944, 'steps': 28411, 'loss/train': 1.677545428276062} 02/25/2022 14:24:00 - INFO - codeparrot_training - Step 28412: {'lr': 0.00021069746366978177, 'samples': 14547456, 'steps': 28412, 'loss/train': 0.8821040391921997} 02/25/2022 14:24:03 - INFO - codeparrot_training - Step 28413: {'lr': 0.00021068130475662255, 'samples': 14547968, 'steps': 28413, 'loss/train': 1.6094051599502563} 02/25/2022 14:24:07 - INFO - codeparrot_training - Step 28414: {'lr': 0.00021066514601189218, 'samples': 14548480, 'steps': 28414, 'loss/train': 2.2515792846679688} 02/25/2022 14:24:13 - INFO - codeparrot_training - Step 28415: {'lr': 0.00021064898743565976, 'samples': 14548992, 'steps': 28415, 'loss/train': 0.9202149510383606} 02/25/2022 14:24:16 - INFO - codeparrot_training - Step 28416: {'lr': 0.00021063282902799468, 'samples': 14549504, 'steps': 28416, 'loss/train': 3.0163931846618652} 02/25/2022 14:24:22 - INFO - codeparrot_training - Step 28417: {'lr': 0.000210616670788966, 'samples': 14550016, 'steps': 28417, 'loss/train': 0.9904517531394958} 02/25/2022 14:24:25 - INFO - codeparrot_training - Step 28418: {'lr': 0.00021060051271864304, 'samples': 14550528, 'steps': 28418, 'loss/train': 1.882892370223999} 02/25/2022 14:24:31 - INFO - codeparrot_training - Step 28419: {'lr': 0.00021058435481709496, 'samples': 14551040, 'steps': 28419, 'loss/train': 1.5398706197738647} 02/25/2022 14:24:34 - INFO - codeparrot_training - Step 28420: {'lr': 0.00021056819708439092, 'samples': 14551552, 'steps': 28420, 'loss/train': 1.811499834060669} 02/25/2022 14:24:40 - INFO - codeparrot_training - Step 28421: {'lr': 0.00021055203952060022, 'samples': 14552064, 'steps': 28421, 'loss/train': 1.8703547716140747} 02/25/2022 14:24:45 - INFO - codeparrot_training - Step 28422: {'lr': 0.0002105358821257921, 'samples': 14552576, 'steps': 28422, 'loss/train': 1.6546560525894165} 02/25/2022 14:24:49 - INFO - codeparrot_training - Step 28423: {'lr': 0.0002105197249000358, 'samples': 14553088, 'steps': 28423, 'loss/train': 2.8906209468841553} 02/25/2022 14:24:55 - INFO - codeparrot_training - Step 28424: {'lr': 0.00021050356784340033, 'samples': 14553600, 'steps': 28424, 'loss/train': 1.5693621635437012} 02/25/2022 14:24:58 - INFO - codeparrot_training - Step 28425: {'lr': 0.00021048741095595506, 'samples': 14554112, 'steps': 28425, 'loss/train': 1.0829867124557495} 02/25/2022 14:25:04 - INFO - codeparrot_training - Step 28426: {'lr': 0.00021047125423776918, 'samples': 14554624, 'steps': 28426, 'loss/train': 0.5926124453544617} 02/25/2022 14:25:07 - INFO - codeparrot_training - Step 28427: {'lr': 0.000210455097688912, 'samples': 14555136, 'steps': 28427, 'loss/train': 1.9430432319641113} 02/25/2022 14:25:13 - INFO - codeparrot_training - Step 28428: {'lr': 0.00021043894130945252, 'samples': 14555648, 'steps': 28428, 'loss/train': 2.2305712699890137} 02/25/2022 14:25:16 - INFO - codeparrot_training - Step 28429: {'lr': 0.00021042278509946004, 'samples': 14556160, 'steps': 28429, 'loss/train': 1.6274597644805908} 02/25/2022 14:25:22 - INFO - codeparrot_training - Step 28430: {'lr': 0.00021040662905900376, 'samples': 14556672, 'steps': 28430, 'loss/train': 2.3341946601867676} 02/25/2022 14:25:25 - INFO - codeparrot_training - Step 28431: {'lr': 0.00021039047318815307, 'samples': 14557184, 'steps': 28431, 'loss/train': 2.4299092292785645} 02/25/2022 14:25:31 - INFO - codeparrot_training - Step 28432: {'lr': 0.00021037431748697688, 'samples': 14557696, 'steps': 28432, 'loss/train': 2.862508773803711} 02/25/2022 14:25:34 - INFO - codeparrot_training - Step 28433: {'lr': 0.00021035816195554452, 'samples': 14558208, 'steps': 28433, 'loss/train': 1.7706307172775269} 02/25/2022 14:25:40 - INFO - codeparrot_training - Step 28434: {'lr': 0.00021034200659392522, 'samples': 14558720, 'steps': 28434, 'loss/train': 1.8307759761810303} 02/25/2022 14:25:44 - INFO - codeparrot_training - Step 28435: {'lr': 0.00021032585140218817, 'samples': 14559232, 'steps': 28435, 'loss/train': 0.9823546409606934} 02/25/2022 14:25:49 - INFO - codeparrot_training - Step 28436: {'lr': 0.0002103096963804027, 'samples': 14559744, 'steps': 28436, 'loss/train': 1.3510633707046509} 02/25/2022 14:25:53 - INFO - codeparrot_training - Step 28437: {'lr': 0.00021029354152863776, 'samples': 14560256, 'steps': 28437, 'loss/train': 2.3421902656555176} 02/25/2022 14:25:58 - INFO - codeparrot_training - Step 28438: {'lr': 0.00021027738684696267, 'samples': 14560768, 'steps': 28438, 'loss/train': 1.8667206764221191} 02/25/2022 14:26:02 - INFO - codeparrot_training - Step 28439: {'lr': 0.00021026123233544667, 'samples': 14561280, 'steps': 28439, 'loss/train': 0.923180103302002} 02/25/2022 14:26:07 - INFO - codeparrot_training - Step 28440: {'lr': 0.000210245077994159, 'samples': 14561792, 'steps': 28440, 'loss/train': 3.013241767883301} 02/25/2022 14:26:11 - INFO - codeparrot_training - Step 28441: {'lr': 0.00021022892382316873, 'samples': 14562304, 'steps': 28441, 'loss/train': 1.757417917251587} 02/25/2022 14:26:16 - INFO - codeparrot_training - Step 28442: {'lr': 0.00021021276982254508, 'samples': 14562816, 'steps': 28442, 'loss/train': 1.7152827978134155} 02/25/2022 14:26:20 - INFO - codeparrot_training - Step 28443: {'lr': 0.0002101966159923573, 'samples': 14563328, 'steps': 28443, 'loss/train': 2.326441526412964} 02/25/2022 14:26:25 - INFO - codeparrot_training - Step 28444: {'lr': 0.0002101804623326746, 'samples': 14563840, 'steps': 28444, 'loss/train': 1.1961010694503784} 02/25/2022 14:26:29 - INFO - codeparrot_training - Step 28445: {'lr': 0.00021016430884356627, 'samples': 14564352, 'steps': 28445, 'loss/train': 1.5915242433547974} 02/25/2022 14:26:36 - INFO - codeparrot_training - Step 28446: {'lr': 0.00021014815552510124, 'samples': 14564864, 'steps': 28446, 'loss/train': 1.7820900678634644} 02/25/2022 14:26:39 - INFO - codeparrot_training - Step 28447: {'lr': 0.0002101320023773489, 'samples': 14565376, 'steps': 28447, 'loss/train': 2.502531051635742} 02/25/2022 14:26:45 - INFO - codeparrot_training - Step 28448: {'lr': 0.00021011584940037838, 'samples': 14565888, 'steps': 28448, 'loss/train': 2.571626663208008} 02/25/2022 14:26:48 - INFO - codeparrot_training - Step 28449: {'lr': 0.00021009969659425902, 'samples': 14566400, 'steps': 28449, 'loss/train': 1.7852882146835327} 02/25/2022 14:26:53 - INFO - codeparrot_training - Step 28450: {'lr': 0.00021008354395905978, 'samples': 14566912, 'steps': 28450, 'loss/train': 1.5485172271728516} 02/25/2022 14:26:57 - INFO - codeparrot_training - Step 28451: {'lr': 0.00021006739149484995, 'samples': 14567424, 'steps': 28451, 'loss/train': 2.1825942993164062} 02/25/2022 14:27:03 - INFO - codeparrot_training - Step 28452: {'lr': 0.00021005123920169878, 'samples': 14567936, 'steps': 28452, 'loss/train': 2.541562795639038} 02/25/2022 14:27:06 - INFO - codeparrot_training - Step 28453: {'lr': 0.00021003508707967544, 'samples': 14568448, 'steps': 28453, 'loss/train': 3.115570068359375} 02/25/2022 14:27:11 - INFO - codeparrot_training - Step 28454: {'lr': 0.000210018935128849, 'samples': 14568960, 'steps': 28454, 'loss/train': 1.3251688480377197} 02/25/2022 14:27:15 - INFO - codeparrot_training - Step 28455: {'lr': 0.0002100027833492888, 'samples': 14569472, 'steps': 28455, 'loss/train': 1.4058822393417358} 02/25/2022 14:27:21 - INFO - codeparrot_training - Step 28456: {'lr': 0.000209986631741064, 'samples': 14569984, 'steps': 28456, 'loss/train': 1.6258450746536255} 02/25/2022 14:27:25 - INFO - codeparrot_training - Step 28457: {'lr': 0.00020997048030424372, 'samples': 14570496, 'steps': 28457, 'loss/train': 1.88962721824646} 02/25/2022 14:27:30 - INFO - codeparrot_training - Step 28458: {'lr': 0.00020995432903889725, 'samples': 14571008, 'steps': 28458, 'loss/train': 1.9786540269851685} 02/25/2022 14:27:34 - INFO - codeparrot_training - Step 28459: {'lr': 0.0002099381779450937, 'samples': 14571520, 'steps': 28459, 'loss/train': 1.2691636085510254} 02/25/2022 14:27:39 - INFO - codeparrot_training - Step 28460: {'lr': 0.00020992202702290225, 'samples': 14572032, 'steps': 28460, 'loss/train': 1.8930144309997559} 02/25/2022 14:27:45 - INFO - codeparrot_training - Step 28461: {'lr': 0.00020990587627239208, 'samples': 14572544, 'steps': 28461, 'loss/train': 1.8002294301986694} 02/25/2022 14:27:48 - INFO - codeparrot_training - Step 28462: {'lr': 0.00020988972569363246, 'samples': 14573056, 'steps': 28462, 'loss/train': 2.0443029403686523} 02/25/2022 14:27:52 - INFO - codeparrot_training - Step 28463: {'lr': 0.00020987357528669254, 'samples': 14573568, 'steps': 28463, 'loss/train': 0.8098900318145752} 02/25/2022 14:27:57 - INFO - codeparrot_training - Step 28464: {'lr': 0.00020985742505164143, 'samples': 14574080, 'steps': 28464, 'loss/train': 4.687154769897461} 02/25/2022 14:28:01 - INFO - codeparrot_training - Step 28465: {'lr': 0.00020984127498854834, 'samples': 14574592, 'steps': 28465, 'loss/train': 1.154121994972229} 02/25/2022 14:28:07 - INFO - codeparrot_training - Step 28466: {'lr': 0.0002098251250974826, 'samples': 14575104, 'steps': 28466, 'loss/train': 1.4480441808700562} 02/25/2022 14:28:13 - INFO - codeparrot_training - Step 28467: {'lr': 0.00020980897537851314, 'samples': 14575616, 'steps': 28467, 'loss/train': 0.6143372654914856} 02/25/2022 14:28:16 - INFO - codeparrot_training - Step 28468: {'lr': 0.00020979282583170932, 'samples': 14576128, 'steps': 28468, 'loss/train': 1.7119415998458862} 02/25/2022 14:28:22 - INFO - codeparrot_training - Step 28469: {'lr': 0.00020977667645714023, 'samples': 14576640, 'steps': 28469, 'loss/train': 1.733964443206787} 02/25/2022 14:28:25 - INFO - codeparrot_training - Step 28470: {'lr': 0.00020976052725487507, 'samples': 14577152, 'steps': 28470, 'loss/train': 1.044413447380066} 02/25/2022 14:28:31 - INFO - codeparrot_training - Step 28471: {'lr': 0.00020974437822498317, 'samples': 14577664, 'steps': 28471, 'loss/train': 2.8491978645324707} 02/25/2022 14:28:34 - INFO - codeparrot_training - Step 28472: {'lr': 0.00020972822936753344, 'samples': 14578176, 'steps': 28472, 'loss/train': 1.7251328229904175} 02/25/2022 14:28:40 - INFO - codeparrot_training - Step 28473: {'lr': 0.00020971208068259518, 'samples': 14578688, 'steps': 28473, 'loss/train': 0.9348810315132141} 02/25/2022 14:28:43 - INFO - codeparrot_training - Step 28474: {'lr': 0.0002096959321702376, 'samples': 14579200, 'steps': 28474, 'loss/train': 1.3053526878356934} 02/25/2022 14:28:50 - INFO - codeparrot_training - Step 28475: {'lr': 0.00020967978383052994, 'samples': 14579712, 'steps': 28475, 'loss/train': 1.5097630023956299} 02/25/2022 14:28:53 - INFO - codeparrot_training - Step 28476: {'lr': 0.0002096636356635412, 'samples': 14580224, 'steps': 28476, 'loss/train': 2.7082297801971436} 02/25/2022 14:28:59 - INFO - codeparrot_training - Step 28477: {'lr': 0.0002096474876693406, 'samples': 14580736, 'steps': 28477, 'loss/train': 0.1345982849597931} 02/25/2022 14:29:03 - INFO - codeparrot_training - Step 28478: {'lr': 0.00020963133984799737, 'samples': 14581248, 'steps': 28478, 'loss/train': 1.7590373754501343} 02/25/2022 14:29:08 - INFO - codeparrot_training - Step 28479: {'lr': 0.00020961519219958079, 'samples': 14581760, 'steps': 28479, 'loss/train': 2.068002939224243} 02/25/2022 14:29:12 - INFO - codeparrot_training - Step 28480: {'lr': 0.00020959904472415974, 'samples': 14582272, 'steps': 28480, 'loss/train': 1.5228018760681152} 02/25/2022 14:29:17 - INFO - codeparrot_training - Step 28481: {'lr': 0.0002095828974218036, 'samples': 14582784, 'steps': 28481, 'loss/train': 0.8721629977226257} 02/25/2022 14:29:21 - INFO - codeparrot_training - Step 28482: {'lr': 0.0002095667502925815, 'samples': 14583296, 'steps': 28482, 'loss/train': 0.08726678043603897} 02/25/2022 14:29:26 - INFO - codeparrot_training - Step 28483: {'lr': 0.00020955060333656257, 'samples': 14583808, 'steps': 28483, 'loss/train': 2.0495686531066895} 02/25/2022 14:29:30 - INFO - codeparrot_training - Step 28484: {'lr': 0.00020953445655381615, 'samples': 14584320, 'steps': 28484, 'loss/train': 2.4478495121002197} 02/25/2022 14:29:35 - INFO - codeparrot_training - Step 28485: {'lr': 0.00020951830994441113, 'samples': 14584832, 'steps': 28485, 'loss/train': 1.9732154607772827} 02/25/2022 14:29:39 - INFO - codeparrot_training - Step 28486: {'lr': 0.00020950216350841682, 'samples': 14585344, 'steps': 28486, 'loss/train': 0.6213319301605225} 02/25/2022 14:29:44 - INFO - codeparrot_training - Step 28487: {'lr': 0.00020948601724590237, 'samples': 14585856, 'steps': 28487, 'loss/train': 1.034710168838501} 02/25/2022 14:29:48 - INFO - codeparrot_training - Step 28488: {'lr': 0.0002094698711569371, 'samples': 14586368, 'steps': 28488, 'loss/train': 1.8198992013931274} 02/25/2022 14:29:53 - INFO - codeparrot_training - Step 28489: {'lr': 0.00020945372524158992, 'samples': 14586880, 'steps': 28489, 'loss/train': 0.9788869619369507} 02/25/2022 14:29:57 - INFO - codeparrot_training - Step 28490: {'lr': 0.0002094375794999301, 'samples': 14587392, 'steps': 28490, 'loss/train': 1.9800159931182861} 02/25/2022 14:30:02 - INFO - codeparrot_training - Step 28491: {'lr': 0.0002094214339320268, 'samples': 14587904, 'steps': 28491, 'loss/train': 1.5009678602218628} 02/25/2022 14:30:06 - INFO - codeparrot_training - Step 28492: {'lr': 0.00020940528853794928, 'samples': 14588416, 'steps': 28492, 'loss/train': 0.9870285391807556} 02/25/2022 14:30:12 - INFO - codeparrot_training - Step 28493: {'lr': 0.00020938914331776657, 'samples': 14588928, 'steps': 28493, 'loss/train': 0.7175915837287903} 02/25/2022 14:30:15 - INFO - codeparrot_training - Step 28494: {'lr': 0.00020937299827154782, 'samples': 14589440, 'steps': 28494, 'loss/train': 1.7921266555786133} 02/25/2022 14:30:21 - INFO - codeparrot_training - Step 28495: {'lr': 0.00020935685339936228, 'samples': 14589952, 'steps': 28495, 'loss/train': 1.127036690711975} 02/25/2022 14:30:24 - INFO - codeparrot_training - Step 28496: {'lr': 0.0002093407087012791, 'samples': 14590464, 'steps': 28496, 'loss/train': 0.12431463599205017} 02/25/2022 14:30:30 - INFO - codeparrot_training - Step 28497: {'lr': 0.0002093245641773674, 'samples': 14590976, 'steps': 28497, 'loss/train': 1.3516435623168945} 02/25/2022 14:30:33 - INFO - codeparrot_training - Step 28498: {'lr': 0.00020930841982769635, 'samples': 14591488, 'steps': 28498, 'loss/train': 1.6913585662841797} 02/25/2022 14:30:38 - INFO - codeparrot_training - Step 28499: {'lr': 0.00020929227565233513, 'samples': 14592000, 'steps': 28499, 'loss/train': 1.9581748247146606} 02/25/2022 14:30:42 - INFO - codeparrot_training - Step 28500: {'lr': 0.00020927613165135284, 'samples': 14592512, 'steps': 28500, 'loss/train': 2.6589584350585938} 02/25/2022 14:30:48 - INFO - codeparrot_training - Step 28501: {'lr': 0.00020925998782481868, 'samples': 14593024, 'steps': 28501, 'loss/train': 1.4079232215881348} 02/25/2022 14:30:51 - INFO - codeparrot_training - Step 28502: {'lr': 0.0002092438441728018, 'samples': 14593536, 'steps': 28502, 'loss/train': 1.5161478519439697} 02/25/2022 14:30:57 - INFO - codeparrot_training - Step 28503: {'lr': 0.00020922770069537136, 'samples': 14594048, 'steps': 28503, 'loss/train': 1.4210277795791626} 02/25/2022 14:31:01 - INFO - codeparrot_training - Step 28504: {'lr': 0.00020921155739259646, 'samples': 14594560, 'steps': 28504, 'loss/train': 3.5918595790863037} 02/25/2022 14:31:06 - INFO - codeparrot_training - Step 28505: {'lr': 0.0002091954142645463, 'samples': 14595072, 'steps': 28505, 'loss/train': 2.795027732849121} 02/25/2022 14:31:10 - INFO - codeparrot_training - Step 28506: {'lr': 0.00020917927131129015, 'samples': 14595584, 'steps': 28506, 'loss/train': 1.9583570957183838} 02/25/2022 14:31:15 - INFO - codeparrot_training - Step 28507: {'lr': 0.0002091631285328969, 'samples': 14596096, 'steps': 28507, 'loss/train': 2.0473642349243164} 02/25/2022 14:31:19 - INFO - codeparrot_training - Step 28508: {'lr': 0.00020914698592943586, 'samples': 14596608, 'steps': 28508, 'loss/train': 1.3696526288986206} 02/25/2022 14:31:24 - INFO - codeparrot_training - Step 28509: {'lr': 0.00020913084350097618, 'samples': 14597120, 'steps': 28509, 'loss/train': 2.081183433532715} 02/25/2022 14:31:27 - INFO - codeparrot_training - Step 28510: {'lr': 0.00020911470124758706, 'samples': 14597632, 'steps': 28510, 'loss/train': 1.8736523389816284} 02/25/2022 14:31:33 - INFO - codeparrot_training - Step 28511: {'lr': 0.00020909855916933747, 'samples': 14598144, 'steps': 28511, 'loss/train': 1.8217201232910156} 02/25/2022 14:31:40 - INFO - codeparrot_training - Step 28512: {'lr': 0.0002090824172662967, 'samples': 14598656, 'steps': 28512, 'loss/train': 1.6766507625579834} 02/25/2022 14:31:43 - INFO - codeparrot_training - Step 28513: {'lr': 0.00020906627553853382, 'samples': 14599168, 'steps': 28513, 'loss/train': 1.1767115592956543} 02/25/2022 14:31:49 - INFO - codeparrot_training - Step 28514: {'lr': 0.00020905013398611817, 'samples': 14599680, 'steps': 28514, 'loss/train': 2.170016288757324} 02/25/2022 14:31:52 - INFO - codeparrot_training - Step 28515: {'lr': 0.0002090339926091186, 'samples': 14600192, 'steps': 28515, 'loss/train': 2.7046866416931152} 02/25/2022 14:31:58 - INFO - codeparrot_training - Step 28516: {'lr': 0.0002090178514076044, 'samples': 14600704, 'steps': 28516, 'loss/train': 2.140939235687256} 02/25/2022 14:32:01 - INFO - codeparrot_training - Step 28517: {'lr': 0.00020900171038164473, 'samples': 14601216, 'steps': 28517, 'loss/train': 1.3000941276550293} 02/25/2022 14:32:07 - INFO - codeparrot_training - Step 28518: {'lr': 0.00020898556953130872, 'samples': 14601728, 'steps': 28518, 'loss/train': 1.7874553203582764} 02/25/2022 14:32:10 - INFO - codeparrot_training - Step 28519: {'lr': 0.00020896942885666558, 'samples': 14602240, 'steps': 28519, 'loss/train': 0.09542856365442276} 02/25/2022 14:32:16 - INFO - codeparrot_training - Step 28520: {'lr': 0.0002089532883577843, 'samples': 14602752, 'steps': 28520, 'loss/train': 1.963853120803833} 02/25/2022 14:32:19 - INFO - codeparrot_training - Step 28521: {'lr': 0.00020893714803473407, 'samples': 14603264, 'steps': 28521, 'loss/train': 2.3917295932769775} 02/25/2022 14:32:26 - INFO - codeparrot_training - Step 28522: {'lr': 0.00020892100788758407, 'samples': 14603776, 'steps': 28522, 'loss/train': 1.9010334014892578} 02/25/2022 14:32:29 - INFO - codeparrot_training - Step 28523: {'lr': 0.00020890486791640356, 'samples': 14604288, 'steps': 28523, 'loss/train': 2.143414258956909} 02/25/2022 14:32:35 - INFO - codeparrot_training - Step 28524: {'lr': 0.00020888872812126143, 'samples': 14604800, 'steps': 28524, 'loss/train': 2.0403006076812744} 02/25/2022 14:32:38 - INFO - codeparrot_training - Step 28525: {'lr': 0.00020887258850222693, 'samples': 14605312, 'steps': 28525, 'loss/train': 3.372239112854004} 02/25/2022 14:32:44 - INFO - codeparrot_training - Step 28526: {'lr': 0.0002088564490593692, 'samples': 14605824, 'steps': 28526, 'loss/train': 1.5008960962295532} 02/25/2022 14:32:48 - INFO - codeparrot_training - Step 28527: {'lr': 0.00020884030979275749, 'samples': 14606336, 'steps': 28527, 'loss/train': 2.028052568435669} 02/25/2022 14:32:53 - INFO - codeparrot_training - Step 28528: {'lr': 0.0002088241707024607, 'samples': 14606848, 'steps': 28528, 'loss/train': 1.7060691118240356} 02/25/2022 14:32:57 - INFO - codeparrot_training - Step 28529: {'lr': 0.0002088080317885481, 'samples': 14607360, 'steps': 28529, 'loss/train': 0.13562779128551483} 02/25/2022 14:33:02 - INFO - codeparrot_training - Step 28530: {'lr': 0.0002087918930510888, 'samples': 14607872, 'steps': 28530, 'loss/train': 2.797316312789917} 02/25/2022 14:33:06 - INFO - codeparrot_training - Step 28531: {'lr': 0.00020877575449015196, 'samples': 14608384, 'steps': 28531, 'loss/train': 2.104167938232422} 02/25/2022 14:33:11 - INFO - codeparrot_training - Step 28532: {'lr': 0.0002087596161058068, 'samples': 14608896, 'steps': 28532, 'loss/train': 2.1439998149871826} 02/25/2022 14:33:15 - INFO - codeparrot_training - Step 28533: {'lr': 0.00020874347789812223, 'samples': 14609408, 'steps': 28533, 'loss/train': 1.8332749605178833} 02/25/2022 14:33:21 - INFO - codeparrot_training - Step 28534: {'lr': 0.00020872733986716753, 'samples': 14609920, 'steps': 28534, 'loss/train': 0.17677833139896393} 02/25/2022 14:33:24 - INFO - codeparrot_training - Step 28535: {'lr': 0.00020871120201301175, 'samples': 14610432, 'steps': 28535, 'loss/train': 2.381044626235962} 02/25/2022 14:33:30 - INFO - codeparrot_training - Step 28536: {'lr': 0.00020869506433572423, 'samples': 14610944, 'steps': 28536, 'loss/train': 1.4228692054748535} 02/25/2022 14:33:33 - INFO - codeparrot_training - Step 28537: {'lr': 0.0002086789268353738, 'samples': 14611456, 'steps': 28537, 'loss/train': 1.1511629819869995} 02/25/2022 14:33:40 - INFO - codeparrot_training - Step 28538: {'lr': 0.00020866278951202976, 'samples': 14611968, 'steps': 28538, 'loss/train': 1.707629680633545} 02/25/2022 14:33:43 - INFO - codeparrot_training - Step 28539: {'lr': 0.00020864665236576123, 'samples': 14612480, 'steps': 28539, 'loss/train': 1.5443429946899414} 02/25/2022 14:33:48 - INFO - codeparrot_training - Step 28540: {'lr': 0.0002086305153966373, 'samples': 14612992, 'steps': 28540, 'loss/train': 1.6012040376663208} 02/25/2022 14:33:52 - INFO - codeparrot_training - Step 28541: {'lr': 0.00020861437860472706, 'samples': 14613504, 'steps': 28541, 'loss/train': 1.6184651851654053} 02/25/2022 14:33:57 - INFO - codeparrot_training - Step 28542: {'lr': 0.0002085982419900997, 'samples': 14614016, 'steps': 28542, 'loss/train': 2.0229878425598145} 02/25/2022 14:34:01 - INFO - codeparrot_training - Step 28543: {'lr': 0.00020858210555282436, 'samples': 14614528, 'steps': 28543, 'loss/train': 1.0092744827270508} 02/25/2022 14:34:06 - INFO - codeparrot_training - Step 28544: {'lr': 0.00020856596929297007, 'samples': 14615040, 'steps': 28544, 'loss/train': 1.4197784662246704} 02/25/2022 14:34:10 - INFO - codeparrot_training - Step 28545: {'lr': 0.00020854983321060607, 'samples': 14615552, 'steps': 28545, 'loss/train': 2.2018327713012695} 02/25/2022 14:34:15 - INFO - codeparrot_training - Step 28546: {'lr': 0.00020853369730580143, 'samples': 14616064, 'steps': 28546, 'loss/train': 2.091029644012451} 02/25/2022 14:34:19 - INFO - codeparrot_training - Step 28547: {'lr': 0.00020851756157862523, 'samples': 14616576, 'steps': 28547, 'loss/train': 1.0809040069580078} 02/25/2022 14:34:25 - INFO - codeparrot_training - Step 28548: {'lr': 0.00020850142602914657, 'samples': 14617088, 'steps': 28548, 'loss/train': 1.5546313524246216} 02/25/2022 14:34:28 - INFO - codeparrot_training - Step 28549: {'lr': 0.0002084852906574347, 'samples': 14617600, 'steps': 28549, 'loss/train': 1.2265230417251587} 02/25/2022 14:34:34 - INFO - codeparrot_training - Step 28550: {'lr': 0.0002084691554635587, 'samples': 14618112, 'steps': 28550, 'loss/train': 2.349365472793579} 02/25/2022 14:34:39 - INFO - codeparrot_training - Step 28551: {'lr': 0.00020845302044758757, 'samples': 14618624, 'steps': 28551, 'loss/train': 1.6875478029251099} 02/25/2022 14:34:43 - INFO - codeparrot_training - Step 28552: {'lr': 0.0002084368856095905, 'samples': 14619136, 'steps': 28552, 'loss/train': 1.7619774341583252} 02/25/2022 14:34:48 - INFO - codeparrot_training - Step 28553: {'lr': 0.0002084207509496366, 'samples': 14619648, 'steps': 28553, 'loss/train': 1.9773417711257935} 02/25/2022 14:34:52 - INFO - codeparrot_training - Step 28554: {'lr': 0.00020840461646779514, 'samples': 14620160, 'steps': 28554, 'loss/train': 1.5791746377944946} 02/25/2022 14:34:57 - INFO - codeparrot_training - Step 28555: {'lr': 0.00020838848216413498, 'samples': 14620672, 'steps': 28555, 'loss/train': 1.9619215726852417} 02/25/2022 14:35:01 - INFO - codeparrot_training - Step 28556: {'lr': 0.00020837234803872535, 'samples': 14621184, 'steps': 28556, 'loss/train': 2.4131245613098145} 02/25/2022 14:35:08 - INFO - codeparrot_training - Step 28557: {'lr': 0.00020835621409163535, 'samples': 14621696, 'steps': 28557, 'loss/train': 1.222489356994629} 02/25/2022 14:35:11 - INFO - codeparrot_training - Step 28558: {'lr': 0.00020834008032293423, 'samples': 14622208, 'steps': 28558, 'loss/train': 2.3088722229003906} 02/25/2022 14:35:17 - INFO - codeparrot_training - Step 28559: {'lr': 0.00020832394673269084, 'samples': 14622720, 'steps': 28559, 'loss/train': 1.247127652168274} 02/25/2022 14:35:20 - INFO - codeparrot_training - Step 28560: {'lr': 0.00020830781332097445, 'samples': 14623232, 'steps': 28560, 'loss/train': 2.018090009689331} 02/25/2022 14:35:26 - INFO - codeparrot_training - Step 28561: {'lr': 0.00020829168008785416, 'samples': 14623744, 'steps': 28561, 'loss/train': 1.283644199371338} 02/25/2022 14:35:29 - INFO - codeparrot_training - Step 28562: {'lr': 0.00020827554703339916, 'samples': 14624256, 'steps': 28562, 'loss/train': 0.7227175235748291} 02/25/2022 14:35:35 - INFO - codeparrot_training - Step 28563: {'lr': 0.00020825941415767834, 'samples': 14624768, 'steps': 28563, 'loss/train': 1.8993874788284302} 02/25/2022 14:35:38 - INFO - codeparrot_training - Step 28564: {'lr': 0.00020824328146076097, 'samples': 14625280, 'steps': 28564, 'loss/train': 0.7130821943283081} 02/25/2022 14:35:43 - INFO - codeparrot_training - Step 28565: {'lr': 0.0002082271489427161, 'samples': 14625792, 'steps': 28565, 'loss/train': 1.4089535474777222} 02/25/2022 14:35:47 - INFO - codeparrot_training - Step 28566: {'lr': 0.00020821101660361288, 'samples': 14626304, 'steps': 28566, 'loss/train': 2.0551435947418213} 02/25/2022 14:35:52 - INFO - codeparrot_training - Step 28567: {'lr': 0.0002081948844435205, 'samples': 14626816, 'steps': 28567, 'loss/train': 0.628305971622467} 02/25/2022 14:35:56 - INFO - codeparrot_training - Step 28568: {'lr': 0.00020817875246250783, 'samples': 14627328, 'steps': 28568, 'loss/train': 1.8296648263931274} 02/25/2022 14:36:02 - INFO - codeparrot_training - Step 28569: {'lr': 0.00020816262066064413, 'samples': 14627840, 'steps': 28569, 'loss/train': 2.1563613414764404} 02/25/2022 14:36:06 - INFO - codeparrot_training - Step 28570: {'lr': 0.0002081464890379985, 'samples': 14628352, 'steps': 28570, 'loss/train': 0.7723124027252197} 02/25/2022 14:36:12 - INFO - codeparrot_training - Step 28571: {'lr': 0.00020813035759464007, 'samples': 14628864, 'steps': 28571, 'loss/train': 1.6140035390853882} 02/25/2022 14:36:15 - INFO - codeparrot_training - Step 28572: {'lr': 0.00020811422633063782, 'samples': 14629376, 'steps': 28572, 'loss/train': 1.7647329568862915} 02/25/2022 14:36:21 - INFO - codeparrot_training - Step 28573: {'lr': 0.00020809809524606091, 'samples': 14629888, 'steps': 28573, 'loss/train': 1.649605631828308} 02/25/2022 14:36:24 - INFO - codeparrot_training - Step 28574: {'lr': 0.00020808196434097848, 'samples': 14630400, 'steps': 28574, 'loss/train': 1.9374405145645142} 02/25/2022 14:36:30 - INFO - codeparrot_training - Step 28575: {'lr': 0.00020806583361545965, 'samples': 14630912, 'steps': 28575, 'loss/train': 0.14211036264896393} 02/25/2022 14:36:33 - INFO - codeparrot_training - Step 28576: {'lr': 0.00020804970306957343, 'samples': 14631424, 'steps': 28576, 'loss/train': 2.1012015342712402} 02/25/2022 14:36:39 - INFO - codeparrot_training - Step 28577: {'lr': 0.00020803357270338896, 'samples': 14631936, 'steps': 28577, 'loss/train': 1.3976686000823975} 02/25/2022 14:36:42 - INFO - codeparrot_training - Step 28578: {'lr': 0.0002080174425169753, 'samples': 14632448, 'steps': 28578, 'loss/train': 2.407975673675537} 02/25/2022 14:36:49 - INFO - codeparrot_training - Step 28579: {'lr': 0.00020800131251040167, 'samples': 14632960, 'steps': 28579, 'loss/train': 1.723849892616272} 02/25/2022 14:36:52 - INFO - codeparrot_training - Step 28580: {'lr': 0.00020798518268373706, 'samples': 14633472, 'steps': 28580, 'loss/train': 1.3217285871505737} 02/25/2022 14:36:58 - INFO - codeparrot_training - Step 28581: {'lr': 0.00020796905303705054, 'samples': 14633984, 'steps': 28581, 'loss/train': 1.6374187469482422} 02/25/2022 14:37:01 - INFO - codeparrot_training - Step 28582: {'lr': 0.00020795292357041127, 'samples': 14634496, 'steps': 28582, 'loss/train': 1.4681124687194824} 02/25/2022 14:37:07 - INFO - codeparrot_training - Step 28583: {'lr': 0.00020793679428388835, 'samples': 14635008, 'steps': 28583, 'loss/train': 1.5293604135513306} 02/25/2022 14:37:10 - INFO - codeparrot_training - Step 28584: {'lr': 0.00020792066517755087, 'samples': 14635520, 'steps': 28584, 'loss/train': 2.2930104732513428} 02/25/2022 14:37:16 - INFO - codeparrot_training - Step 28585: {'lr': 0.0002079045362514678, 'samples': 14636032, 'steps': 28585, 'loss/train': 2.5324337482452393} 02/25/2022 14:37:19 - INFO - codeparrot_training - Step 28586: {'lr': 0.00020788840750570844, 'samples': 14636544, 'steps': 28586, 'loss/train': 0.18094457685947418} 02/25/2022 14:37:25 - INFO - codeparrot_training - Step 28587: {'lr': 0.0002078722789403417, 'samples': 14637056, 'steps': 28587, 'loss/train': 1.896743655204773} 02/25/2022 14:37:28 - INFO - codeparrot_training - Step 28588: {'lr': 0.00020785615055543677, 'samples': 14637568, 'steps': 28588, 'loss/train': 1.4646371603012085} 02/25/2022 14:37:34 - INFO - codeparrot_training - Step 28589: {'lr': 0.0002078400223510627, 'samples': 14638080, 'steps': 28589, 'loss/train': 1.76212477684021} 02/25/2022 14:37:37 - INFO - codeparrot_training - Step 28590: {'lr': 0.0002078238943272886, 'samples': 14638592, 'steps': 28590, 'loss/train': 0.7759493589401245} 02/25/2022 14:37:43 - INFO - codeparrot_training - Step 28591: {'lr': 0.00020780776648418348, 'samples': 14639104, 'steps': 28591, 'loss/train': 1.2317774295806885} 02/25/2022 14:37:46 - INFO - codeparrot_training - Step 28592: {'lr': 0.00020779163882181655, 'samples': 14639616, 'steps': 28592, 'loss/train': 1.692404866218567} 02/25/2022 14:37:52 - INFO - codeparrot_training - Step 28593: {'lr': 0.00020777551134025683, 'samples': 14640128, 'steps': 28593, 'loss/train': 2.312106132507324} 02/25/2022 14:37:55 - INFO - codeparrot_training - Step 28594: {'lr': 0.00020775938403957346, 'samples': 14640640, 'steps': 28594, 'loss/train': 2.040748119354248} 02/25/2022 14:38:02 - INFO - codeparrot_training - Step 28595: {'lr': 0.00020774325691983537, 'samples': 14641152, 'steps': 28595, 'loss/train': 1.5753356218338013} 02/25/2022 14:38:06 - INFO - codeparrot_training - Step 28596: {'lr': 0.00020772712998111178, 'samples': 14641664, 'steps': 28596, 'loss/train': 1.6934571266174316} 02/25/2022 14:38:11 - INFO - codeparrot_training - Step 28597: {'lr': 0.00020771100322347185, 'samples': 14642176, 'steps': 28597, 'loss/train': 1.2630354166030884} 02/25/2022 14:38:15 - INFO - codeparrot_training - Step 28598: {'lr': 0.00020769487664698444, 'samples': 14642688, 'steps': 28598, 'loss/train': 0.935442328453064} 02/25/2022 14:38:20 - INFO - codeparrot_training - Step 28599: {'lr': 0.00020767875025171876, 'samples': 14643200, 'steps': 28599, 'loss/train': 1.4629456996917725} 02/25/2022 14:38:24 - INFO - codeparrot_training - Step 28600: {'lr': 0.00020766262403774385, 'samples': 14643712, 'steps': 28600, 'loss/train': 1.1431686878204346} 02/25/2022 14:38:29 - INFO - codeparrot_training - Step 28601: {'lr': 0.00020764649800512883, 'samples': 14644224, 'steps': 28601, 'loss/train': 1.6016335487365723} 02/25/2022 14:38:33 - INFO - codeparrot_training - Step 28602: {'lr': 0.00020763037215394288, 'samples': 14644736, 'steps': 28602, 'loss/train': 0.3140012323856354} 02/25/2022 14:38:39 - INFO - codeparrot_training - Step 28603: {'lr': 0.00020761424648425484, 'samples': 14645248, 'steps': 28603, 'loss/train': 0.8882197737693787} 02/25/2022 14:38:42 - INFO - codeparrot_training - Step 28604: {'lr': 0.0002075981209961339, 'samples': 14645760, 'steps': 28604, 'loss/train': 1.863071084022522} 02/25/2022 14:38:48 - INFO - codeparrot_training - Step 28605: {'lr': 0.00020758199568964917, 'samples': 14646272, 'steps': 28605, 'loss/train': 1.4822843074798584} 02/25/2022 14:38:51 - INFO - codeparrot_training - Step 28606: {'lr': 0.00020756587056486978, 'samples': 14646784, 'steps': 28606, 'loss/train': 2.31294846534729} 02/25/2022 14:38:57 - INFO - codeparrot_training - Step 28607: {'lr': 0.00020754974562186467, 'samples': 14647296, 'steps': 28607, 'loss/train': 1.0011610984802246} 02/25/2022 14:39:00 - INFO - codeparrot_training - Step 28608: {'lr': 0.00020753362086070292, 'samples': 14647808, 'steps': 28608, 'loss/train': 1.8208271265029907} 02/25/2022 14:39:06 - INFO - codeparrot_training - Step 28609: {'lr': 0.00020751749628145368, 'samples': 14648320, 'steps': 28609, 'loss/train': 1.6482386589050293} 02/25/2022 14:39:09 - INFO - codeparrot_training - Step 28610: {'lr': 0.0002075013718841861, 'samples': 14648832, 'steps': 28610, 'loss/train': 1.8164997100830078} 02/25/2022 14:39:15 - INFO - codeparrot_training - Step 28611: {'lr': 0.000207485247668969, 'samples': 14649344, 'steps': 28611, 'loss/train': 0.45227307081222534} 02/25/2022 14:39:18 - INFO - codeparrot_training - Step 28612: {'lr': 0.00020746912363587164, 'samples': 14649856, 'steps': 28612, 'loss/train': 1.3443998098373413} 02/25/2022 14:39:24 - INFO - codeparrot_training - Step 28613: {'lr': 0.00020745299978496305, 'samples': 14650368, 'steps': 28613, 'loss/train': 2.8215768337249756} 02/25/2022 14:39:27 - INFO - codeparrot_training - Step 28614: {'lr': 0.00020743687611631228, 'samples': 14650880, 'steps': 28614, 'loss/train': 2.068390369415283} 02/25/2022 14:39:33 - INFO - codeparrot_training - Step 28615: {'lr': 0.00020742075262998854, 'samples': 14651392, 'steps': 28615, 'loss/train': 2.1479415893554688} 02/25/2022 14:39:37 - INFO - codeparrot_training - Step 28616: {'lr': 0.00020740462932606067, 'samples': 14651904, 'steps': 28616, 'loss/train': 2.1982851028442383} 02/25/2022 14:39:42 - INFO - codeparrot_training - Step 28617: {'lr': 0.00020738850620459782, 'samples': 14652416, 'steps': 28617, 'loss/train': 1.437268853187561} 02/25/2022 14:39:46 - INFO - codeparrot_training - Step 28618: {'lr': 0.00020737238326566909, 'samples': 14652928, 'steps': 28618, 'loss/train': 1.3859179019927979} 02/25/2022 14:39:51 - INFO - codeparrot_training - Step 28619: {'lr': 0.00020735626050934368, 'samples': 14653440, 'steps': 28619, 'loss/train': 1.6420785188674927} 02/25/2022 14:39:55 - INFO - codeparrot_training - Step 28620: {'lr': 0.00020734013793569035, 'samples': 14653952, 'steps': 28620, 'loss/train': 2.492062568664551} 02/25/2022 14:40:00 - INFO - codeparrot_training - Step 28621: {'lr': 0.00020732401554477835, 'samples': 14654464, 'steps': 28621, 'loss/train': 0.6783117055892944} 02/25/2022 14:40:04 - INFO - codeparrot_training - Step 28622: {'lr': 0.0002073078933366767, 'samples': 14654976, 'steps': 28622, 'loss/train': 1.1454886198043823} 02/25/2022 14:40:09 - INFO - codeparrot_training - Step 28623: {'lr': 0.0002072917713114546, 'samples': 14655488, 'steps': 28623, 'loss/train': 0.911496639251709} 02/25/2022 14:40:15 - INFO - codeparrot_training - Step 28624: {'lr': 0.00020727564946918087, 'samples': 14656000, 'steps': 28624, 'loss/train': 2.397329092025757} 02/25/2022 14:40:19 - INFO - codeparrot_training - Step 28625: {'lr': 0.00020725952780992467, 'samples': 14656512, 'steps': 28625, 'loss/train': 1.5379050970077515} 02/25/2022 14:40:24 - INFO - codeparrot_training - Step 28626: {'lr': 0.00020724340633375513, 'samples': 14657024, 'steps': 28626, 'loss/train': 1.7232321500778198} 02/25/2022 14:40:28 - INFO - codeparrot_training - Step 28627: {'lr': 0.0002072272850407413, 'samples': 14657536, 'steps': 28627, 'loss/train': 2.2017805576324463} 02/25/2022 14:40:33 - INFO - codeparrot_training - Step 28628: {'lr': 0.00020721116393095218, 'samples': 14658048, 'steps': 28628, 'loss/train': 1.135353922843933} 02/25/2022 14:40:37 - INFO - codeparrot_training - Step 28629: {'lr': 0.0002071950430044568, 'samples': 14658560, 'steps': 28629, 'loss/train': 1.7855346202850342} 02/25/2022 14:40:43 - INFO - codeparrot_training - Step 28630: {'lr': 0.00020717892226132433, 'samples': 14659072, 'steps': 28630, 'loss/train': 1.7313529253005981} 02/25/2022 14:40:46 - INFO - codeparrot_training - Step 28631: {'lr': 0.00020716280170162373, 'samples': 14659584, 'steps': 28631, 'loss/train': 1.4980180263519287} 02/25/2022 14:40:51 - INFO - codeparrot_training - Step 28632: {'lr': 0.00020714668132542407, 'samples': 14660096, 'steps': 28632, 'loss/train': 0.990715503692627} 02/25/2022 14:40:55 - INFO - codeparrot_training - Step 28633: {'lr': 0.00020713056113279442, 'samples': 14660608, 'steps': 28633, 'loss/train': 1.4652318954467773} 02/25/2022 14:41:01 - INFO - codeparrot_training - Step 28634: {'lr': 0.00020711444112380388, 'samples': 14661120, 'steps': 28634, 'loss/train': 1.8936306238174438} 02/25/2022 14:41:04 - INFO - codeparrot_training - Step 28635: {'lr': 0.0002070983212985214, 'samples': 14661632, 'steps': 28635, 'loss/train': 2.2042338848114014} 02/25/2022 14:41:10 - INFO - codeparrot_training - Step 28636: {'lr': 0.0002070822016570161, 'samples': 14662144, 'steps': 28636, 'loss/train': 1.4645395278930664} 02/25/2022 14:41:13 - INFO - codeparrot_training - Step 28637: {'lr': 0.00020706608219935707, 'samples': 14662656, 'steps': 28637, 'loss/train': 2.1972157955169678} 02/25/2022 14:41:19 - INFO - codeparrot_training - Step 28638: {'lr': 0.00020704996292561324, 'samples': 14663168, 'steps': 28638, 'loss/train': 1.8572359085083008} 02/25/2022 14:41:22 - INFO - codeparrot_training - Step 28639: {'lr': 0.00020703384383585375, 'samples': 14663680, 'steps': 28639, 'loss/train': 1.9268752336502075} 02/25/2022 14:41:28 - INFO - codeparrot_training - Step 28640: {'lr': 0.00020701772493014758, 'samples': 14664192, 'steps': 28640, 'loss/train': 2.246455669403076} 02/25/2022 14:41:32 - INFO - codeparrot_training - Step 28641: {'lr': 0.000207001606208564, 'samples': 14664704, 'steps': 28641, 'loss/train': 2.3508694171905518} 02/25/2022 14:41:35 - INFO - codeparrot_training - Step 28642: {'lr': 0.00020698548767117173, 'samples': 14665216, 'steps': 28642, 'loss/train': 2.2325327396392822} 02/25/2022 14:41:41 - INFO - codeparrot_training - Step 28643: {'lr': 0.00020696936931804, 'samples': 14665728, 'steps': 28643, 'loss/train': 1.6687617301940918} 02/25/2022 14:41:44 - INFO - codeparrot_training - Step 28644: {'lr': 0.0002069532511492378, 'samples': 14666240, 'steps': 28644, 'loss/train': 1.7231296300888062} 02/25/2022 14:41:50 - INFO - codeparrot_training - Step 28645: {'lr': 0.00020693713316483433, 'samples': 14666752, 'steps': 28645, 'loss/train': 1.0525126457214355} 02/25/2022 14:41:53 - INFO - codeparrot_training - Step 28646: {'lr': 0.00020692101536489842, 'samples': 14667264, 'steps': 28646, 'loss/train': 1.8687127828598022} 02/25/2022 14:41:59 - INFO - codeparrot_training - Step 28647: {'lr': 0.00020690489774949917, 'samples': 14667776, 'steps': 28647, 'loss/train': 1.6920956373214722} 02/25/2022 14:42:02 - INFO - codeparrot_training - Step 28648: {'lr': 0.00020688878031870566, 'samples': 14668288, 'steps': 28648, 'loss/train': 2.2957491874694824} 02/25/2022 14:42:08 - INFO - codeparrot_training - Step 28649: {'lr': 0.00020687266307258702, 'samples': 14668800, 'steps': 28649, 'loss/train': 1.5056724548339844} 02/25/2022 14:42:11 - INFO - codeparrot_training - Step 28650: {'lr': 0.00020685654601121213, 'samples': 14669312, 'steps': 28650, 'loss/train': 3.3470239639282227} 02/25/2022 14:42:18 - INFO - codeparrot_training - Step 28651: {'lr': 0.00020684042913465006, 'samples': 14669824, 'steps': 28651, 'loss/train': 2.2611420154571533} 02/25/2022 14:42:23 - INFO - codeparrot_training - Step 28652: {'lr': 0.0002068243124429699, 'samples': 14670336, 'steps': 28652, 'loss/train': 1.4850214719772339} 02/25/2022 14:42:27 - INFO - codeparrot_training - Step 28653: {'lr': 0.00020680819593624068, 'samples': 14670848, 'steps': 28653, 'loss/train': 0.2812162935733795} 02/25/2022 14:42:30 - INFO - codeparrot_training - Step 28654: {'lr': 0.00020679207961453154, 'samples': 14671360, 'steps': 28654, 'loss/train': 2.114703416824341} 02/25/2022 14:42:36 - INFO - codeparrot_training - Step 28655: {'lr': 0.0002067759634779113, 'samples': 14671872, 'steps': 28655, 'loss/train': 1.8977705240249634} 02/25/2022 14:42:39 - INFO - codeparrot_training - Step 28656: {'lr': 0.0002067598475264491, 'samples': 14672384, 'steps': 28656, 'loss/train': 3.3362317085266113} 02/25/2022 14:42:45 - INFO - codeparrot_training - Step 28657: {'lr': 0.00020674373176021398, 'samples': 14672896, 'steps': 28657, 'loss/train': 2.0299575328826904} 02/25/2022 14:42:50 - INFO - codeparrot_training - Step 28658: {'lr': 0.0002067276161792751, 'samples': 14673408, 'steps': 28658, 'loss/train': 1.2983375787734985} 02/25/2022 14:42:54 - INFO - codeparrot_training - Step 28659: {'lr': 0.00020671150078370126, 'samples': 14673920, 'steps': 28659, 'loss/train': 1.2460395097732544} 02/25/2022 14:43:00 - INFO - codeparrot_training - Step 28660: {'lr': 0.0002066953855735616, 'samples': 14674432, 'steps': 28660, 'loss/train': 1.99422287940979} 02/25/2022 14:43:03 - INFO - codeparrot_training - Step 28661: {'lr': 0.00020667927054892517, 'samples': 14674944, 'steps': 28661, 'loss/train': 1.4847508668899536} 02/25/2022 14:43:09 - INFO - codeparrot_training - Step 28662: {'lr': 0.000206663155709861, 'samples': 14675456, 'steps': 28662, 'loss/train': 2.0936968326568604} 02/25/2022 14:43:12 - INFO - codeparrot_training - Step 28663: {'lr': 0.00020664704105643819, 'samples': 14675968, 'steps': 28663, 'loss/train': 1.7086166143417358} 02/25/2022 14:43:18 - INFO - codeparrot_training - Step 28664: {'lr': 0.00020663092658872558, 'samples': 14676480, 'steps': 28664, 'loss/train': 1.5557129383087158} 02/25/2022 14:43:21 - INFO - codeparrot_training - Step 28665: {'lr': 0.00020661481230679232, 'samples': 14676992, 'steps': 28665, 'loss/train': 1.6385211944580078} 02/25/2022 14:43:27 - INFO - codeparrot_training - Step 28666: {'lr': 0.00020659869821070743, 'samples': 14677504, 'steps': 28666, 'loss/train': 2.0725016593933105} 02/25/2022 14:43:30 - INFO - codeparrot_training - Step 28667: {'lr': 0.00020658258430054005, 'samples': 14678016, 'steps': 28667, 'loss/train': 0.871519148349762} 02/25/2022 14:43:36 - INFO - codeparrot_training - Step 28668: {'lr': 0.00020656647057635895, 'samples': 14678528, 'steps': 28668, 'loss/train': 1.3905963897705078} 02/25/2022 14:43:39 - INFO - codeparrot_training - Step 28669: {'lr': 0.00020655035703823335, 'samples': 14679040, 'steps': 28669, 'loss/train': 2.8349263668060303} 02/25/2022 14:43:45 - INFO - codeparrot_training - Step 28670: {'lr': 0.00020653424368623224, 'samples': 14679552, 'steps': 28670, 'loss/train': 2.7927067279815674} 02/25/2022 14:43:49 - INFO - codeparrot_training - Step 28671: {'lr': 0.00020651813052042465, 'samples': 14680064, 'steps': 28671, 'loss/train': 2.0033702850341797} 02/25/2022 14:43:54 - INFO - codeparrot_training - Step 28672: {'lr': 0.0002065020175408795, 'samples': 14680576, 'steps': 28672, 'loss/train': 1.4825894832611084} 02/25/2022 14:43:58 - INFO - codeparrot_training - Step 28673: {'lr': 0.00020648590474766593, 'samples': 14681088, 'steps': 28673, 'loss/train': 1.778352975845337} 02/25/2022 14:44:03 - INFO - codeparrot_training - Step 28674: {'lr': 0.00020646979214085297, 'samples': 14681600, 'steps': 28674, 'loss/train': 1.3953208923339844} 02/25/2022 14:44:07 - INFO - codeparrot_training - Step 28675: {'lr': 0.00020645367972050953, 'samples': 14682112, 'steps': 28675, 'loss/train': 1.6745994091033936} 02/25/2022 14:44:12 - INFO - codeparrot_training - Step 28676: {'lr': 0.00020643756748670475, 'samples': 14682624, 'steps': 28676, 'loss/train': 2.226574420928955} 02/25/2022 14:44:16 - INFO - codeparrot_training - Step 28677: {'lr': 0.00020642145543950762, 'samples': 14683136, 'steps': 28677, 'loss/train': 2.3025829792022705} 02/25/2022 14:44:21 - INFO - codeparrot_training - Step 28678: {'lr': 0.00020640534357898707, 'samples': 14683648, 'steps': 28678, 'loss/train': 1.069924235343933} 02/25/2022 14:44:25 - INFO - codeparrot_training - Step 28679: {'lr': 0.0002063892319052122, 'samples': 14684160, 'steps': 28679, 'loss/train': 2.2852087020874023} 02/25/2022 14:44:30 - INFO - codeparrot_training - Step 28680: {'lr': 0.00020637312041825204, 'samples': 14684672, 'steps': 28680, 'loss/train': 1.590694785118103} 02/25/2022 14:44:34 - INFO - codeparrot_training - Step 28681: {'lr': 0.0002063570091181756, 'samples': 14685184, 'steps': 28681, 'loss/train': 1.619643211364746} 02/25/2022 14:44:39 - INFO - codeparrot_training - Step 28682: {'lr': 0.00020634089800505178, 'samples': 14685696, 'steps': 28682, 'loss/train': 1.892874836921692} 02/25/2022 14:44:43 - INFO - codeparrot_training - Step 28683: {'lr': 0.00020632478707894976, 'samples': 14686208, 'steps': 28683, 'loss/train': 0.04007269814610481} 02/25/2022 14:44:48 - INFO - codeparrot_training - Step 28684: {'lr': 0.00020630867633993855, 'samples': 14686720, 'steps': 28684, 'loss/train': 0.06422439217567444} 02/25/2022 14:44:52 - INFO - codeparrot_training - Step 28685: {'lr': 0.00020629256578808696, 'samples': 14687232, 'steps': 28685, 'loss/train': 2.165538787841797} 02/25/2022 14:44:58 - INFO - codeparrot_training - Step 28686: {'lr': 0.0002062764554234642, 'samples': 14687744, 'steps': 28686, 'loss/train': 2.598635196685791} 02/25/2022 14:45:02 - INFO - codeparrot_training - Step 28687: {'lr': 0.00020626034524613918, 'samples': 14688256, 'steps': 28687, 'loss/train': 2.199892044067383} 02/25/2022 14:45:07 - INFO - codeparrot_training - Step 28688: {'lr': 0.00020624423525618098, 'samples': 14688768, 'steps': 28688, 'loss/train': 1.1122218370437622} 02/25/2022 14:45:11 - INFO - codeparrot_training - Step 28689: {'lr': 0.00020622812545365864, 'samples': 14689280, 'steps': 28689, 'loss/train': 1.9154456853866577} 02/25/2022 14:45:16 - INFO - codeparrot_training - Step 28690: {'lr': 0.00020621201583864108, 'samples': 14689792, 'steps': 28690, 'loss/train': 2.39687180519104} 02/25/2022 14:45:20 - INFO - codeparrot_training - Step 28691: {'lr': 0.0002061959064111973, 'samples': 14690304, 'steps': 28691, 'loss/train': 1.426198959350586} 02/25/2022 14:45:25 - INFO - codeparrot_training - Step 28692: {'lr': 0.00020617979717139635, 'samples': 14690816, 'steps': 28692, 'loss/train': 0.7376397252082825} 02/25/2022 14:45:29 - INFO - codeparrot_training - Step 28693: {'lr': 0.00020616368811930734, 'samples': 14691328, 'steps': 28693, 'loss/train': 1.5889759063720703} 02/25/2022 14:45:34 - INFO - codeparrot_training - Step 28694: {'lr': 0.00020614757925499906, 'samples': 14691840, 'steps': 28694, 'loss/train': 0.9628810286521912} 02/25/2022 14:45:37 - INFO - codeparrot_training - Step 28695: {'lr': 0.0002061314705785406, 'samples': 14692352, 'steps': 28695, 'loss/train': 2.215167284011841} 02/25/2022 14:45:44 - INFO - codeparrot_training - Step 28696: {'lr': 0.00020611536209000104, 'samples': 14692864, 'steps': 28696, 'loss/train': 1.7047783136367798} 02/25/2022 14:45:47 - INFO - codeparrot_training - Step 28697: {'lr': 0.0002060992537894494, 'samples': 14693376, 'steps': 28697, 'loss/train': 1.2807097434997559} 02/25/2022 14:45:53 - INFO - codeparrot_training - Step 28698: {'lr': 0.00020608314567695452, 'samples': 14693888, 'steps': 28698, 'loss/train': 0.7985394597053528} 02/25/2022 14:45:56 - INFO - codeparrot_training - Step 28699: {'lr': 0.0002060670377525855, 'samples': 14694400, 'steps': 28699, 'loss/train': 0.030608918517827988} 02/25/2022 14:46:02 - INFO - codeparrot_training - Step 28700: {'lr': 0.00020605093001641137, 'samples': 14694912, 'steps': 28700, 'loss/train': 3.155197858810425} 02/25/2022 14:46:05 - INFO - codeparrot_training - Step 28701: {'lr': 0.0002060348224685011, 'samples': 14695424, 'steps': 28701, 'loss/train': 1.1235108375549316} 02/25/2022 14:46:11 - INFO - codeparrot_training - Step 28702: {'lr': 0.00020601871510892374, 'samples': 14695936, 'steps': 28702, 'loss/train': 1.8598308563232422} 02/25/2022 14:46:14 - INFO - codeparrot_training - Step 28703: {'lr': 0.00020600260793774815, 'samples': 14696448, 'steps': 28703, 'loss/train': 1.8477109670639038} 02/25/2022 14:46:20 - INFO - codeparrot_training - Step 28704: {'lr': 0.00020598650095504344, 'samples': 14696960, 'steps': 28704, 'loss/train': 1.6313574314117432} 02/25/2022 14:46:23 - INFO - codeparrot_training - Step 28705: {'lr': 0.0002059703941608786, 'samples': 14697472, 'steps': 28705, 'loss/train': 2.326218366622925} 02/25/2022 14:46:31 - INFO - codeparrot_training - Step 28706: {'lr': 0.0002059542875553227, 'samples': 14697984, 'steps': 28706, 'loss/train': 2.12508487701416} 02/25/2022 14:46:34 - INFO - codeparrot_training - Step 28707: {'lr': 0.00020593818113844454, 'samples': 14698496, 'steps': 28707, 'loss/train': 2.466625452041626} 02/25/2022 14:46:39 - INFO - codeparrot_training - Step 28708: {'lr': 0.0002059220749103132, 'samples': 14699008, 'steps': 28708, 'loss/train': 2.1992530822753906} 02/25/2022 14:46:43 - INFO - codeparrot_training - Step 28709: {'lr': 0.00020590596887099771, 'samples': 14699520, 'steps': 28709, 'loss/train': 4.6187028884887695} 02/25/2022 14:46:48 - INFO - codeparrot_training - Step 28710: {'lr': 0.00020588986302056713, 'samples': 14700032, 'steps': 28710, 'loss/train': 2.0518155097961426} 02/25/2022 14:46:54 - INFO - codeparrot_training - Step 28711: {'lr': 0.00020587375735909032, 'samples': 14700544, 'steps': 28711, 'loss/train': 1.4279311895370483} 02/25/2022 14:46:57 - INFO - codeparrot_training - Step 28712: {'lr': 0.00020585765188663627, 'samples': 14701056, 'steps': 28712, 'loss/train': 2.015519857406616} 02/25/2022 14:47:03 - INFO - codeparrot_training - Step 28713: {'lr': 0.00020584154660327407, 'samples': 14701568, 'steps': 28713, 'loss/train': 1.6761972904205322} 02/25/2022 14:47:06 - INFO - codeparrot_training - Step 28714: {'lr': 0.00020582544150907268, 'samples': 14702080, 'steps': 28714, 'loss/train': 1.7252320051193237} 02/25/2022 14:47:13 - INFO - codeparrot_training - Step 28715: {'lr': 0.00020580933660410106, 'samples': 14702592, 'steps': 28715, 'loss/train': 2.269017219543457} 02/25/2022 14:47:16 - INFO - codeparrot_training - Step 28716: {'lr': 0.00020579323188842818, 'samples': 14703104, 'steps': 28716, 'loss/train': 1.3190195560455322} 02/25/2022 14:47:21 - INFO - codeparrot_training - Step 28717: {'lr': 0.0002057771273621231, 'samples': 14703616, 'steps': 28717, 'loss/train': 1.3510839939117432} 02/25/2022 14:47:25 - INFO - codeparrot_training - Step 28718: {'lr': 0.00020576102302525474, 'samples': 14704128, 'steps': 28718, 'loss/train': 0.7739048600196838} 02/25/2022 14:47:30 - INFO - codeparrot_training - Step 28719: {'lr': 0.00020574491887789212, 'samples': 14704640, 'steps': 28719, 'loss/train': 2.20670223236084} 02/25/2022 14:47:34 - INFO - codeparrot_training - Step 28720: {'lr': 0.0002057288149201042, 'samples': 14705152, 'steps': 28720, 'loss/train': 2.3893651962280273} 02/25/2022 14:47:39 - INFO - codeparrot_training - Step 28721: {'lr': 0.00020571271115196, 'samples': 14705664, 'steps': 28721, 'loss/train': 1.216496467590332} 02/25/2022 14:47:43 - INFO - codeparrot_training - Step 28722: {'lr': 0.00020569660757352847, 'samples': 14706176, 'steps': 28722, 'loss/train': 1.737797498703003} 02/25/2022 14:47:48 - INFO - codeparrot_training - Step 28723: {'lr': 0.00020568050418487855, 'samples': 14706688, 'steps': 28723, 'loss/train': 1.9807382822036743} 02/25/2022 14:47:52 - INFO - codeparrot_training - Step 28724: {'lr': 0.00020566440098607943, 'samples': 14707200, 'steps': 28724, 'loss/train': 1.1350288391113281} 02/25/2022 14:47:57 - INFO - codeparrot_training - Step 28725: {'lr': 0.0002056482979771998, 'samples': 14707712, 'steps': 28725, 'loss/train': 1.7946361303329468} 02/25/2022 14:48:01 - INFO - codeparrot_training - Step 28726: {'lr': 0.0002056321951583088, 'samples': 14708224, 'steps': 28726, 'loss/train': 1.7340128421783447} 02/25/2022 14:48:06 - INFO - codeparrot_training - Step 28727: {'lr': 0.0002056160925294754, 'samples': 14708736, 'steps': 28727, 'loss/train': 1.778532862663269} 02/25/2022 14:48:10 - INFO - codeparrot_training - Step 28728: {'lr': 0.00020559999009076864, 'samples': 14709248, 'steps': 28728, 'loss/train': 0.8869004249572754} 02/25/2022 14:48:15 - INFO - codeparrot_training - Step 28729: {'lr': 0.00020558388784225733, 'samples': 14709760, 'steps': 28729, 'loss/train': 1.7055113315582275} 02/25/2022 14:48:19 - INFO - codeparrot_training - Step 28730: {'lr': 0.00020556778578401054, 'samples': 14710272, 'steps': 28730, 'loss/train': 1.5329267978668213} 02/25/2022 14:48:26 - INFO - codeparrot_training - Step 28731: {'lr': 0.00020555168391609726, 'samples': 14710784, 'steps': 28731, 'loss/train': 1.6859725713729858} 02/25/2022 14:48:29 - INFO - codeparrot_training - Step 28732: {'lr': 0.00020553558223858654, 'samples': 14711296, 'steps': 28732, 'loss/train': 0.2816525101661682} 02/25/2022 14:48:34 - INFO - codeparrot_training - Step 28733: {'lr': 0.00020551948075154714, 'samples': 14711808, 'steps': 28733, 'loss/train': 1.9571174383163452} 02/25/2022 14:48:38 - INFO - codeparrot_training - Step 28734: {'lr': 0.00020550337945504822, 'samples': 14712320, 'steps': 28734, 'loss/train': 1.4706687927246094} 02/25/2022 14:48:43 - INFO - codeparrot_training - Step 28735: {'lr': 0.00020548727834915866, 'samples': 14712832, 'steps': 28735, 'loss/train': 0.6561219692230225} 02/25/2022 14:48:47 - INFO - codeparrot_training - Step 28736: {'lr': 0.00020547117743394743, 'samples': 14713344, 'steps': 28736, 'loss/train': 2.74123477935791} 02/25/2022 14:48:52 - INFO - codeparrot_training - Step 28737: {'lr': 0.00020545507670948369, 'samples': 14713856, 'steps': 28737, 'loss/train': 1.3911833763122559} 02/25/2022 14:48:56 - INFO - codeparrot_training - Step 28738: {'lr': 0.00020543897617583614, 'samples': 14714368, 'steps': 28738, 'loss/train': 0.653398334980011} 02/25/2022 14:49:01 - INFO - codeparrot_training - Step 28739: {'lr': 0.00020542287583307387, 'samples': 14714880, 'steps': 28739, 'loss/train': 1.9609392881393433} 02/25/2022 14:49:05 - INFO - codeparrot_training - Step 28740: {'lr': 0.00020540677568126585, 'samples': 14715392, 'steps': 28740, 'loss/train': 1.7053741216659546} 02/25/2022 14:49:10 - INFO - codeparrot_training - Step 28741: {'lr': 0.00020539067572048113, 'samples': 14715904, 'steps': 28741, 'loss/train': 1.6404954195022583} 02/25/2022 14:49:14 - INFO - codeparrot_training - Step 28742: {'lr': 0.00020537457595078847, 'samples': 14716416, 'steps': 28742, 'loss/train': 0.9432055950164795} 02/25/2022 14:49:20 - INFO - codeparrot_training - Step 28743: {'lr': 0.00020535847637225698, 'samples': 14716928, 'steps': 28743, 'loss/train': 1.2730201482772827} 02/25/2022 14:49:24 - INFO - codeparrot_training - Step 28744: {'lr': 0.0002053423769849556, 'samples': 14717440, 'steps': 28744, 'loss/train': 1.872104287147522} 02/25/2022 14:49:29 - INFO - codeparrot_training - Step 28745: {'lr': 0.0002053262777889534, 'samples': 14717952, 'steps': 28745, 'loss/train': 0.7880858182907104} 02/25/2022 14:49:32 - INFO - codeparrot_training - Step 28746: {'lr': 0.00020531017878431916, 'samples': 14718464, 'steps': 28746, 'loss/train': 1.3543180227279663} 02/25/2022 14:49:38 - INFO - codeparrot_training - Step 28747: {'lr': 0.0002052940799711219, 'samples': 14718976, 'steps': 28747, 'loss/train': 2.4413537979125977} 02/25/2022 14:49:41 - INFO - codeparrot_training - Step 28748: {'lr': 0.0002052779813494306, 'samples': 14719488, 'steps': 28748, 'loss/train': 1.8142073154449463} 02/25/2022 14:49:47 - INFO - codeparrot_training - Step 28749: {'lr': 0.00020526188291931425, 'samples': 14720000, 'steps': 28749, 'loss/train': 1.770838975906372} 02/25/2022 14:49:50 - INFO - codeparrot_training - Step 28750: {'lr': 0.00020524578468084187, 'samples': 14720512, 'steps': 28750, 'loss/train': 1.8709532022476196} 02/25/2022 14:49:56 - INFO - codeparrot_training - Step 28751: {'lr': 0.00020522968663408227, 'samples': 14721024, 'steps': 28751, 'loss/train': 1.670910120010376} 02/25/2022 14:49:59 - INFO - codeparrot_training - Step 28752: {'lr': 0.00020521358877910443, 'samples': 14721536, 'steps': 28752, 'loss/train': 1.01792573928833} 02/25/2022 14:50:06 - INFO - codeparrot_training - Step 28753: {'lr': 0.00020519749111597735, 'samples': 14722048, 'steps': 28753, 'loss/train': 1.2357524633407593} 02/25/2022 14:50:10 - INFO - codeparrot_training - Step 28754: {'lr': 0.00020518139364477013, 'samples': 14722560, 'steps': 28754, 'loss/train': 1.5755587816238403} 02/25/2022 14:50:15 - INFO - codeparrot_training - Step 28755: {'lr': 0.00020516529636555148, 'samples': 14723072, 'steps': 28755, 'loss/train': 1.5529495477676392} 02/25/2022 14:50:18 - INFO - codeparrot_training - Step 28756: {'lr': 0.00020514919927839043, 'samples': 14723584, 'steps': 28756, 'loss/train': 1.8930851221084595} 02/25/2022 14:50:24 - INFO - codeparrot_training - Step 28757: {'lr': 0.00020513310238335605, 'samples': 14724096, 'steps': 28757, 'loss/train': 1.8603923320770264} 02/25/2022 14:50:30 - INFO - codeparrot_training - Step 28758: {'lr': 0.00020511700568051722, 'samples': 14724608, 'steps': 28758, 'loss/train': 1.7283425331115723} 02/25/2022 14:50:33 - INFO - codeparrot_training - Step 28759: {'lr': 0.0002051009091699428, 'samples': 14725120, 'steps': 28759, 'loss/train': 1.3955087661743164} 02/25/2022 14:50:39 - INFO - codeparrot_training - Step 28760: {'lr': 0.00020508481285170185, 'samples': 14725632, 'steps': 28760, 'loss/train': 0.8470161557197571} 02/25/2022 14:50:42 - INFO - codeparrot_training - Step 28761: {'lr': 0.00020506871672586335, 'samples': 14726144, 'steps': 28761, 'loss/train': 1.517698884010315} 02/25/2022 14:50:49 - INFO - codeparrot_training - Step 28762: {'lr': 0.00020505262079249616, 'samples': 14726656, 'steps': 28762, 'loss/train': 2.106020450592041} 02/25/2022 14:50:52 - INFO - codeparrot_training - Step 28763: {'lr': 0.00020503652505166932, 'samples': 14727168, 'steps': 28763, 'loss/train': 1.446267008781433} 02/25/2022 14:50:58 - INFO - codeparrot_training - Step 28764: {'lr': 0.00020502042950345172, 'samples': 14727680, 'steps': 28764, 'loss/train': 2.2972755432128906} 02/25/2022 14:51:01 - INFO - codeparrot_training - Step 28765: {'lr': 0.00020500433414791225, 'samples': 14728192, 'steps': 28765, 'loss/train': 0.03696531429886818} 02/25/2022 14:51:05 - INFO - codeparrot_training - Step 28766: {'lr': 0.00020498823898511994, 'samples': 14728704, 'steps': 28766, 'loss/train': 2.1747655868530273} 02/25/2022 14:51:11 - INFO - codeparrot_training - Step 28767: {'lr': 0.0002049721440151438, 'samples': 14729216, 'steps': 28767, 'loss/train': 1.4541162252426147} 02/25/2022 14:51:14 - INFO - codeparrot_training - Step 28768: {'lr': 0.00020495604923805266, 'samples': 14729728, 'steps': 28768, 'loss/train': 1.9620437622070312} 02/25/2022 14:51:20 - INFO - codeparrot_training - Step 28769: {'lr': 0.00020493995465391547, 'samples': 14730240, 'steps': 28769, 'loss/train': 0.8052703142166138} 02/25/2022 14:51:23 - INFO - codeparrot_training - Step 28770: {'lr': 0.00020492386026280118, 'samples': 14730752, 'steps': 28770, 'loss/train': 1.440199613571167} 02/25/2022 14:51:28 - INFO - codeparrot_training - Step 28771: {'lr': 0.0002049077660647788, 'samples': 14731264, 'steps': 28771, 'loss/train': 2.5613863468170166} 02/25/2022 14:51:32 - INFO - codeparrot_training - Step 28772: {'lr': 0.0002048916720599173, 'samples': 14731776, 'steps': 28772, 'loss/train': 1.9654618501663208} 02/25/2022 14:51:39 - INFO - codeparrot_training - Step 28773: {'lr': 0.00020487557824828544, 'samples': 14732288, 'steps': 28773, 'loss/train': 0.8163354396820068} 02/25/2022 14:51:42 - INFO - codeparrot_training - Step 28774: {'lr': 0.0002048594846299523, 'samples': 14732800, 'steps': 28774, 'loss/train': 0.8364977836608887} 02/25/2022 14:51:48 - INFO - codeparrot_training - Step 28775: {'lr': 0.0002048433912049868, 'samples': 14733312, 'steps': 28775, 'loss/train': 1.1681852340698242} 02/25/2022 14:51:51 - INFO - codeparrot_training - Step 28776: {'lr': 0.00020482729797345798, 'samples': 14733824, 'steps': 28776, 'loss/train': 2.1070070266723633} 02/25/2022 14:51:57 - INFO - codeparrot_training - Step 28777: {'lr': 0.00020481120493543454, 'samples': 14734336, 'steps': 28777, 'loss/train': 1.93741774559021} 02/25/2022 14:52:00 - INFO - codeparrot_training - Step 28778: {'lr': 0.00020479511209098555, 'samples': 14734848, 'steps': 28778, 'loss/train': 0.30311065912246704} 02/25/2022 14:52:06 - INFO - codeparrot_training - Step 28779: {'lr': 0.00020477901944017995, 'samples': 14735360, 'steps': 28779, 'loss/train': 1.9404871463775635} 02/25/2022 14:52:09 - INFO - codeparrot_training - Step 28780: {'lr': 0.0002047629269830868, 'samples': 14735872, 'steps': 28780, 'loss/train': 1.0799428224563599} 02/25/2022 14:52:15 - INFO - codeparrot_training - Step 28781: {'lr': 0.00020474683471977478, 'samples': 14736384, 'steps': 28781, 'loss/train': 1.9959062337875366} 02/25/2022 14:52:18 - INFO - codeparrot_training - Step 28782: {'lr': 0.00020473074265031296, 'samples': 14736896, 'steps': 28782, 'loss/train': 1.7825422286987305} 02/25/2022 14:52:24 - INFO - codeparrot_training - Step 28783: {'lr': 0.00020471465077477027, 'samples': 14737408, 'steps': 28783, 'loss/train': 2.2013068199157715} 02/25/2022 14:52:27 - INFO - codeparrot_training - Step 28784: {'lr': 0.00020469855909321564, 'samples': 14737920, 'steps': 28784, 'loss/train': 3.418954849243164} 02/25/2022 14:52:33 - INFO - codeparrot_training - Step 28785: {'lr': 0.00020468246760571807, 'samples': 14738432, 'steps': 28785, 'loss/train': 0.06748632341623306} 02/25/2022 14:52:36 - INFO - codeparrot_training - Step 28786: {'lr': 0.00020466637631234635, 'samples': 14738944, 'steps': 28786, 'loss/train': 1.76181960105896} 02/25/2022 14:52:42 - INFO - codeparrot_training - Step 28787: {'lr': 0.00020465028521316948, 'samples': 14739456, 'steps': 28787, 'loss/train': 1.4854650497436523} 02/25/2022 14:52:45 - INFO - codeparrot_training - Step 28788: {'lr': 0.00020463419430825637, 'samples': 14739968, 'steps': 28788, 'loss/train': 2.1689870357513428} 02/25/2022 14:52:52 - INFO - codeparrot_training - Step 28789: {'lr': 0.00020461810359767604, 'samples': 14740480, 'steps': 28789, 'loss/train': 1.7968506813049316} 02/25/2022 14:52:55 - INFO - codeparrot_training - Step 28790: {'lr': 0.00020460201308149727, 'samples': 14740992, 'steps': 28790, 'loss/train': 1.7370885610580444} 02/25/2022 14:53:00 - INFO - codeparrot_training - Step 28791: {'lr': 0.00020458592275978906, 'samples': 14741504, 'steps': 28791, 'loss/train': 2.430738925933838} 02/25/2022 14:53:04 - INFO - codeparrot_training - Step 28792: {'lr': 0.00020456983263262036, 'samples': 14742016, 'steps': 28792, 'loss/train': 1.5756193399429321} 02/25/2022 14:53:09 - INFO - codeparrot_training - Step 28793: {'lr': 0.00020455374270006018, 'samples': 14742528, 'steps': 28793, 'loss/train': 2.4550929069519043} 02/25/2022 14:53:13 - INFO - codeparrot_training - Step 28794: {'lr': 0.0002045376529621772, 'samples': 14743040, 'steps': 28794, 'loss/train': 1.1958998441696167} 02/25/2022 14:53:18 - INFO - codeparrot_training - Step 28795: {'lr': 0.00020452156341904047, 'samples': 14743552, 'steps': 28795, 'loss/train': 2.3857009410858154} 02/25/2022 14:53:24 - INFO - codeparrot_training - Step 28796: {'lr': 0.00020450547407071894, 'samples': 14744064, 'steps': 28796, 'loss/train': 1.6521241664886475} 02/25/2022 14:53:27 - INFO - codeparrot_training - Step 28797: {'lr': 0.00020448938491728153, 'samples': 14744576, 'steps': 28797, 'loss/train': 1.5824264287948608} 02/25/2022 14:53:34 - INFO - codeparrot_training - Step 28798: {'lr': 0.00020447329595879718, 'samples': 14745088, 'steps': 28798, 'loss/train': 1.3231697082519531} 02/25/2022 14:53:37 - INFO - codeparrot_training - Step 28799: {'lr': 0.00020445720719533474, 'samples': 14745600, 'steps': 28799, 'loss/train': 2.106687545776367} 02/25/2022 14:53:43 - INFO - codeparrot_training - Step 28800: {'lr': 0.00020444111862696313, 'samples': 14746112, 'steps': 28800, 'loss/train': 1.9485704898834229} 02/25/2022 14:53:46 - INFO - codeparrot_training - Step 28801: {'lr': 0.00020442503025375138, 'samples': 14746624, 'steps': 28801, 'loss/train': 1.8105411529541016} 02/25/2022 14:53:52 - INFO - codeparrot_training - Step 28802: {'lr': 0.0002044089420757683, 'samples': 14747136, 'steps': 28802, 'loss/train': 2.4100394248962402} 02/25/2022 14:53:55 - INFO - codeparrot_training - Step 28803: {'lr': 0.0002043928540930828, 'samples': 14747648, 'steps': 28803, 'loss/train': 0.9466906189918518} 02/25/2022 14:54:01 - INFO - codeparrot_training - Step 28804: {'lr': 0.00020437676630576386, 'samples': 14748160, 'steps': 28804, 'loss/train': 0.7663961052894592} 02/25/2022 14:54:04 - INFO - codeparrot_training - Step 28805: {'lr': 0.00020436067871388032, 'samples': 14748672, 'steps': 28805, 'loss/train': 1.7867168188095093} 02/25/2022 14:54:09 - INFO - codeparrot_training - Step 28806: {'lr': 0.00020434459131750122, 'samples': 14749184, 'steps': 28806, 'loss/train': 1.5550402402877808} 02/25/2022 14:54:13 - INFO - codeparrot_training - Step 28807: {'lr': 0.00020432850411669531, 'samples': 14749696, 'steps': 28807, 'loss/train': 3.034815788269043} 02/25/2022 14:54:19 - INFO - codeparrot_training - Step 28808: {'lr': 0.00020431241711153165, 'samples': 14750208, 'steps': 28808, 'loss/train': 0.093503437936306} 02/25/2022 14:54:23 - INFO - codeparrot_training - Step 28809: {'lr': 0.00020429633030207906, 'samples': 14750720, 'steps': 28809, 'loss/train': 0.9339956641197205} 02/25/2022 14:54:28 - INFO - codeparrot_training - Step 28810: {'lr': 0.00020428024368840644, 'samples': 14751232, 'steps': 28810, 'loss/train': 1.3292827606201172} 02/25/2022 14:54:32 - INFO - codeparrot_training - Step 28811: {'lr': 0.00020426415727058288, 'samples': 14751744, 'steps': 28811, 'loss/train': 2.028827667236328} 02/25/2022 14:54:37 - INFO - codeparrot_training - Step 28812: {'lr': 0.000204248071048677, 'samples': 14752256, 'steps': 28812, 'loss/train': 1.6217399835586548} 02/25/2022 14:54:41 - INFO - codeparrot_training - Step 28813: {'lr': 0.0002042319850227579, 'samples': 14752768, 'steps': 28813, 'loss/train': 0.9365517497062683} 02/25/2022 14:54:46 - INFO - codeparrot_training - Step 28814: {'lr': 0.0002042158991928944, 'samples': 14753280, 'steps': 28814, 'loss/train': 2.1713790893554688} 02/25/2022 14:54:50 - INFO - codeparrot_training - Step 28815: {'lr': 0.0002041998135591556, 'samples': 14753792, 'steps': 28815, 'loss/train': 1.6756418943405151} 02/25/2022 14:54:55 - INFO - codeparrot_training - Step 28816: {'lr': 0.00020418372812161013, 'samples': 14754304, 'steps': 28816, 'loss/train': 0.9748415350914001} 02/25/2022 14:54:59 - INFO - codeparrot_training - Step 28817: {'lr': 0.000204167642880327, 'samples': 14754816, 'steps': 28817, 'loss/train': 1.5152543783187866} 02/25/2022 14:55:05 - INFO - codeparrot_training - Step 28818: {'lr': 0.00020415155783537513, 'samples': 14755328, 'steps': 28818, 'loss/train': 1.4748494625091553} 02/25/2022 14:55:08 - INFO - codeparrot_training - Step 28819: {'lr': 0.00020413547298682346, 'samples': 14755840, 'steps': 28819, 'loss/train': 2.737989902496338} 02/25/2022 14:55:14 - INFO - codeparrot_training - Step 28820: {'lr': 0.00020411938833474097, 'samples': 14756352, 'steps': 28820, 'loss/train': 1.906535267829895} 02/25/2022 14:55:17 - INFO - codeparrot_training - Step 28821: {'lr': 0.00020410330387919632, 'samples': 14756864, 'steps': 28821, 'loss/train': 1.3603588342666626} 02/25/2022 14:55:23 - INFO - codeparrot_training - Step 28822: {'lr': 0.00020408721962025857, 'samples': 14757376, 'steps': 28822, 'loss/train': 1.3775073289871216} 02/25/2022 14:55:26 - INFO - codeparrot_training - Step 28823: {'lr': 0.00020407113555799655, 'samples': 14757888, 'steps': 28823, 'loss/train': 1.9446040391921997} 02/25/2022 14:55:32 - INFO - codeparrot_training - Step 28824: {'lr': 0.00020405505169247934, 'samples': 14758400, 'steps': 28824, 'loss/train': 1.7067701816558838} 02/25/2022 14:55:35 - INFO - codeparrot_training - Step 28825: {'lr': 0.00020403896802377559, 'samples': 14758912, 'steps': 28825, 'loss/train': 1.4373440742492676} 02/25/2022 14:55:41 - INFO - codeparrot_training - Step 28826: {'lr': 0.0002040228845519543, 'samples': 14759424, 'steps': 28826, 'loss/train': 2.052150249481201} 02/25/2022 14:55:44 - INFO - codeparrot_training - Step 28827: {'lr': 0.00020400680127708438, 'samples': 14759936, 'steps': 28827, 'loss/train': 2.708298921585083} 02/25/2022 14:55:50 - INFO - codeparrot_training - Step 28828: {'lr': 0.00020399071819923481, 'samples': 14760448, 'steps': 28828, 'loss/train': 1.6759121417999268} 02/25/2022 14:55:53 - INFO - codeparrot_training - Step 28829: {'lr': 0.0002039746353184743, 'samples': 14760960, 'steps': 28829, 'loss/train': 1.826778769493103} 02/25/2022 14:55:59 - INFO - codeparrot_training - Step 28830: {'lr': 0.00020395855263487185, 'samples': 14761472, 'steps': 28830, 'loss/train': 1.1359045505523682} 02/25/2022 14:56:02 - INFO - codeparrot_training - Step 28831: {'lr': 0.0002039424701484963, 'samples': 14761984, 'steps': 28831, 'loss/train': 1.5218032598495483} 02/25/2022 14:56:07 - INFO - codeparrot_training - Step 28832: {'lr': 0.00020392638785941665, 'samples': 14762496, 'steps': 28832, 'loss/train': 2.3321585655212402} 02/25/2022 14:56:14 - INFO - codeparrot_training - Step 28833: {'lr': 0.00020391030576770178, 'samples': 14763008, 'steps': 28833, 'loss/train': 1.690425157546997} 02/25/2022 14:56:17 - INFO - codeparrot_training - Step 28834: {'lr': 0.00020389422387342044, 'samples': 14763520, 'steps': 28834, 'loss/train': 2.3165602684020996} 02/25/2022 14:56:23 - INFO - codeparrot_training - Step 28835: {'lr': 0.00020387814217664158, 'samples': 14764032, 'steps': 28835, 'loss/train': 1.0011096000671387} 02/25/2022 14:56:26 - INFO - codeparrot_training - Step 28836: {'lr': 0.00020386206067743413, 'samples': 14764544, 'steps': 28836, 'loss/train': 2.806399345397949} 02/25/2022 14:56:32 - INFO - codeparrot_training - Step 28837: {'lr': 0.00020384597937586708, 'samples': 14765056, 'steps': 28837, 'loss/train': 1.8127254247665405} 02/25/2022 14:56:35 - INFO - codeparrot_training - Step 28838: {'lr': 0.00020382989827200907, 'samples': 14765568, 'steps': 28838, 'loss/train': 1.8910161256790161} 02/25/2022 14:56:41 - INFO - codeparrot_training - Step 28839: {'lr': 0.0002038138173659291, 'samples': 14766080, 'steps': 28839, 'loss/train': 1.3368737697601318} 02/25/2022 14:56:44 - INFO - codeparrot_training - Step 28840: {'lr': 0.0002037977366576961, 'samples': 14766592, 'steps': 28840, 'loss/train': 0.7702274322509766} 02/25/2022 14:56:50 - INFO - codeparrot_training - Step 28841: {'lr': 0.00020378165614737904, 'samples': 14767104, 'steps': 28841, 'loss/train': 0.7103793025016785} 02/25/2022 14:56:53 - INFO - codeparrot_training - Step 28842: {'lr': 0.00020376557583504657, 'samples': 14767616, 'steps': 28842, 'loss/train': 1.5509947538375854} 02/25/2022 14:56:59 - INFO - codeparrot_training - Step 28843: {'lr': 0.0002037494957207677, 'samples': 14768128, 'steps': 28843, 'loss/train': 2.027299404144287} 02/25/2022 14:57:03 - INFO - codeparrot_training - Step 28844: {'lr': 0.00020373341580461133, 'samples': 14768640, 'steps': 28844, 'loss/train': 1.8981330394744873} 02/25/2022 14:57:08 - INFO - codeparrot_training - Step 28845: {'lr': 0.00020371733608664627, 'samples': 14769152, 'steps': 28845, 'loss/train': 1.3378561735153198} 02/25/2022 14:57:12 - INFO - codeparrot_training - Step 28846: {'lr': 0.00020370125656694153, 'samples': 14769664, 'steps': 28846, 'loss/train': 0.5291807055473328} 02/25/2022 14:57:17 - INFO - codeparrot_training - Step 28847: {'lr': 0.00020368517724556583, 'samples': 14770176, 'steps': 28847, 'loss/train': 2.3157451152801514} 02/25/2022 14:57:21 - INFO - codeparrot_training - Step 28848: {'lr': 0.00020366909812258817, 'samples': 14770688, 'steps': 28848, 'loss/train': 1.742571234703064} 02/25/2022 14:57:26 - INFO - codeparrot_training - Step 28849: {'lr': 0.00020365301919807733, 'samples': 14771200, 'steps': 28849, 'loss/train': 1.5665849447250366} 02/25/2022 14:57:30 - INFO - codeparrot_training - Step 28850: {'lr': 0.00020363694047210228, 'samples': 14771712, 'steps': 28850, 'loss/train': 2.2817270755767822} 02/25/2022 14:57:35 - INFO - codeparrot_training - Step 28851: {'lr': 0.00020362086194473185, 'samples': 14772224, 'steps': 28851, 'loss/train': 1.2029011249542236} 02/25/2022 14:57:39 - INFO - codeparrot_training - Step 28852: {'lr': 0.00020360478361603493, 'samples': 14772736, 'steps': 28852, 'loss/train': 1.2510825395584106} 02/25/2022 14:57:45 - INFO - codeparrot_training - Step 28853: {'lr': 0.00020358870548608035, 'samples': 14773248, 'steps': 28853, 'loss/train': 2.434044361114502} 02/25/2022 14:57:49 - INFO - codeparrot_training - Step 28854: {'lr': 0.0002035726275549371, 'samples': 14773760, 'steps': 28854, 'loss/train': 2.3309528827667236} 02/25/2022 14:57:54 - INFO - codeparrot_training - Step 28855: {'lr': 0.0002035565498226739, 'samples': 14774272, 'steps': 28855, 'loss/train': 1.2794013023376465} 02/25/2022 14:57:57 - INFO - codeparrot_training - Step 28856: {'lr': 0.00020354047228935969, 'samples': 14774784, 'steps': 28856, 'loss/train': 1.783035397529602} 02/25/2022 14:58:03 - INFO - codeparrot_training - Step 28857: {'lr': 0.00020352439495506335, 'samples': 14775296, 'steps': 28857, 'loss/train': 1.3094043731689453} 02/25/2022 14:58:06 - INFO - codeparrot_training - Step 28858: {'lr': 0.00020350831781985372, 'samples': 14775808, 'steps': 28858, 'loss/train': 1.8641000986099243} 02/25/2022 14:58:12 - INFO - codeparrot_training - Step 28859: {'lr': 0.0002034922408837998, 'samples': 14776320, 'steps': 28859, 'loss/train': 1.677760362625122} 02/25/2022 14:58:15 - INFO - codeparrot_training - Step 28860: {'lr': 0.00020347616414697023, 'samples': 14776832, 'steps': 28860, 'loss/train': 1.5541274547576904} 02/25/2022 14:58:21 - INFO - codeparrot_training - Step 28861: {'lr': 0.00020346008760943404, 'samples': 14777344, 'steps': 28861, 'loss/train': 2.8858442306518555} 02/25/2022 14:58:24 - INFO - codeparrot_training - Step 28862: {'lr': 0.00020344401127126005, 'samples': 14777856, 'steps': 28862, 'loss/train': 1.2962510585784912} 02/25/2022 14:58:30 - INFO - codeparrot_training - Step 28863: {'lr': 0.00020342793513251724, 'samples': 14778368, 'steps': 28863, 'loss/train': 1.4813169240951538} 02/25/2022 14:58:33 - INFO - codeparrot_training - Step 28864: {'lr': 0.00020341185919327425, 'samples': 14778880, 'steps': 28864, 'loss/train': 1.480462908744812} 02/25/2022 14:58:40 - INFO - codeparrot_training - Step 28865: {'lr': 0.00020339578345360005, 'samples': 14779392, 'steps': 28865, 'loss/train': 1.7878892421722412} 02/25/2022 14:58:43 - INFO - codeparrot_training - Step 28866: {'lr': 0.00020337970791356352, 'samples': 14779904, 'steps': 28866, 'loss/train': 2.87617564201355} 02/25/2022 14:58:49 - INFO - codeparrot_training - Step 28867: {'lr': 0.00020336363257323354, 'samples': 14780416, 'steps': 28867, 'loss/train': 2.882838726043701} 02/25/2022 14:58:52 - INFO - codeparrot_training - Step 28868: {'lr': 0.00020334755743267903, 'samples': 14780928, 'steps': 28868, 'loss/train': 1.6327271461486816} 02/25/2022 14:58:58 - INFO - codeparrot_training - Step 28869: {'lr': 0.00020333148249196867, 'samples': 14781440, 'steps': 28869, 'loss/train': 1.1009503602981567} 02/25/2022 14:59:01 - INFO - codeparrot_training - Step 28870: {'lr': 0.00020331540775117142, 'samples': 14781952, 'steps': 28870, 'loss/train': 0.37435033917427063} 02/25/2022 14:59:07 - INFO - codeparrot_training - Step 28871: {'lr': 0.00020329933321035616, 'samples': 14782464, 'steps': 28871, 'loss/train': 1.4789108037948608} 02/25/2022 14:59:10 - INFO - codeparrot_training - Step 28872: {'lr': 0.00020328325886959182, 'samples': 14782976, 'steps': 28872, 'loss/train': 1.1462639570236206} 02/25/2022 14:59:16 - INFO - codeparrot_training - Step 28873: {'lr': 0.00020326718472894704, 'samples': 14783488, 'steps': 28873, 'loss/train': 1.7303941249847412} 02/25/2022 14:59:19 - INFO - codeparrot_training - Step 28874: {'lr': 0.00020325111078849082, 'samples': 14784000, 'steps': 28874, 'loss/train': 1.615073323249817} 02/25/2022 14:59:25 - INFO - codeparrot_training - Step 28875: {'lr': 0.000203235037048292, 'samples': 14784512, 'steps': 28875, 'loss/train': 2.763493299484253} 02/25/2022 14:59:29 - INFO - codeparrot_training - Step 28876: {'lr': 0.0002032189635084195, 'samples': 14785024, 'steps': 28876, 'loss/train': 1.6377590894699097} 02/25/2022 14:59:34 - INFO - codeparrot_training - Step 28877: {'lr': 0.00020320289016894207, 'samples': 14785536, 'steps': 28877, 'loss/train': 2.0550026893615723} 02/25/2022 14:59:38 - INFO - codeparrot_training - Step 28878: {'lr': 0.00020318681702992852, 'samples': 14786048, 'steps': 28878, 'loss/train': 1.940559983253479} 02/25/2022 14:59:43 - INFO - codeparrot_training - Step 28879: {'lr': 0.00020317074409144785, 'samples': 14786560, 'steps': 28879, 'loss/train': 2.354809522628784} 02/25/2022 14:59:49 - INFO - codeparrot_training - Step 28880: {'lr': 0.0002031546713535688, 'samples': 14787072, 'steps': 28880, 'loss/train': 1.8954113721847534} 02/25/2022 14:59:52 - INFO - codeparrot_training - Step 28881: {'lr': 0.00020313859881636038, 'samples': 14787584, 'steps': 28881, 'loss/train': 2.0843141078948975} 02/25/2022 14:59:58 - INFO - codeparrot_training - Step 28882: {'lr': 0.0002031225264798912, 'samples': 14788096, 'steps': 28882, 'loss/train': 2.0360560417175293} 02/25/2022 15:00:01 - INFO - codeparrot_training - Step 28883: {'lr': 0.00020310645434423025, 'samples': 14788608, 'steps': 28883, 'loss/train': 2.3703370094299316} 02/25/2022 15:00:07 - INFO - codeparrot_training - Step 28884: {'lr': 0.00020309038240944643, 'samples': 14789120, 'steps': 28884, 'loss/train': 1.0343043804168701} 02/25/2022 15:00:10 - INFO - codeparrot_training - Step 28885: {'lr': 0.0002030743106756085, 'samples': 14789632, 'steps': 28885, 'loss/train': 1.93584144115448} 02/25/2022 15:00:16 - INFO - codeparrot_training - Step 28886: {'lr': 0.00020305823914278527, 'samples': 14790144, 'steps': 28886, 'loss/train': 0.6957326531410217} 02/25/2022 15:00:19 - INFO - codeparrot_training - Step 28887: {'lr': 0.00020304216781104562, 'samples': 14790656, 'steps': 28887, 'loss/train': 8.838553428649902} 02/25/2022 15:00:23 - INFO - codeparrot_training - Step 28888: {'lr': 0.00020302609668045848, 'samples': 14791168, 'steps': 28888, 'loss/train': 1.8295924663543701} 02/25/2022 15:00:28 - INFO - codeparrot_training - Step 28889: {'lr': 0.00020301002575109266, 'samples': 14791680, 'steps': 28889, 'loss/train': 1.9290168285369873} 02/25/2022 15:00:31 - INFO - codeparrot_training - Step 28890: {'lr': 0.00020299395502301689, 'samples': 14792192, 'steps': 28890, 'loss/train': 0.6554014086723328} 02/25/2022 15:00:38 - INFO - codeparrot_training - Step 28891: {'lr': 0.00020297788449630006, 'samples': 14792704, 'steps': 28891, 'loss/train': 1.4383426904678345} 02/25/2022 15:00:43 - INFO - codeparrot_training - Step 28892: {'lr': 0.0002029618141710111, 'samples': 14793216, 'steps': 28892, 'loss/train': 0.5175086259841919} 02/25/2022 15:00:47 - INFO - codeparrot_training - Step 28893: {'lr': 0.00020294574404721878, 'samples': 14793728, 'steps': 28893, 'loss/train': 0.8374757766723633} 02/25/2022 15:00:52 - INFO - codeparrot_training - Step 28894: {'lr': 0.00020292967412499196, 'samples': 14794240, 'steps': 28894, 'loss/train': 2.3668553829193115} 02/25/2022 15:00:56 - INFO - codeparrot_training - Step 28895: {'lr': 0.00020291360440439948, 'samples': 14794752, 'steps': 28895, 'loss/train': 1.3878892660140991} 02/25/2022 15:01:01 - INFO - codeparrot_training - Step 28896: {'lr': 0.00020289753488551016, 'samples': 14795264, 'steps': 28896, 'loss/train': 1.7678484916687012} 02/25/2022 15:01:05 - INFO - codeparrot_training - Step 28897: {'lr': 0.00020288146556839278, 'samples': 14795776, 'steps': 28897, 'loss/train': 1.3984142541885376} 02/25/2022 15:01:10 - INFO - codeparrot_training - Step 28898: {'lr': 0.00020286539645311634, 'samples': 14796288, 'steps': 28898, 'loss/train': 2.6917362213134766} 02/25/2022 15:01:13 - INFO - codeparrot_training - Step 28899: {'lr': 0.00020284932753974956, 'samples': 14796800, 'steps': 28899, 'loss/train': 1.7898812294006348} 02/25/2022 15:01:20 - INFO - codeparrot_training - Step 28900: {'lr': 0.00020283325882836122, 'samples': 14797312, 'steps': 28900, 'loss/train': 1.6224220991134644} 02/25/2022 15:01:23 - INFO - codeparrot_training - Step 28901: {'lr': 0.00020281719031902025, 'samples': 14797824, 'steps': 28901, 'loss/train': 1.9690816402435303} 02/25/2022 15:01:29 - INFO - codeparrot_training - Step 28902: {'lr': 0.0002028011220117955, 'samples': 14798336, 'steps': 28902, 'loss/train': 4.891348838806152} 02/25/2022 15:01:32 - INFO - codeparrot_training - Step 28903: {'lr': 0.00020278505390675572, 'samples': 14798848, 'steps': 28903, 'loss/train': 1.8070660829544067} 02/25/2022 15:01:38 - INFO - codeparrot_training - Step 28904: {'lr': 0.00020276898600396975, 'samples': 14799360, 'steps': 28904, 'loss/train': 2.5544748306274414} 02/25/2022 15:01:41 - INFO - codeparrot_training - Step 28905: {'lr': 0.00020275291830350645, 'samples': 14799872, 'steps': 28905, 'loss/train': 1.1793084144592285} 02/25/2022 15:01:47 - INFO - codeparrot_training - Step 28906: {'lr': 0.00020273685080543464, 'samples': 14800384, 'steps': 28906, 'loss/train': 1.445697546005249} 02/25/2022 15:01:50 - INFO - codeparrot_training - Step 28907: {'lr': 0.00020272078350982325, 'samples': 14800896, 'steps': 28907, 'loss/train': 1.6257308721542358} 02/25/2022 15:01:56 - INFO - codeparrot_training - Step 28908: {'lr': 0.00020270471641674093, 'samples': 14801408, 'steps': 28908, 'loss/train': 0.0301054734736681} 02/25/2022 15:01:59 - INFO - codeparrot_training - Step 28909: {'lr': 0.00020268864952625656, 'samples': 14801920, 'steps': 28909, 'loss/train': 1.8452787399291992} 02/25/2022 15:02:06 - INFO - codeparrot_training - Step 28910: {'lr': 0.000202672582838439, 'samples': 14802432, 'steps': 28910, 'loss/train': 2.0845794677734375} 02/25/2022 15:02:09 - INFO - codeparrot_training - Step 28911: {'lr': 0.00020265651635335719, 'samples': 14802944, 'steps': 28911, 'loss/train': 1.7245670557022095} 02/25/2022 15:02:15 - INFO - codeparrot_training - Step 28912: {'lr': 0.00020264045007107974, 'samples': 14803456, 'steps': 28912, 'loss/train': 1.579886794090271} 02/25/2022 15:02:18 - INFO - codeparrot_training - Step 28913: {'lr': 0.00020262438399167552, 'samples': 14803968, 'steps': 28913, 'loss/train': 1.6542655229568481} 02/25/2022 15:02:24 - INFO - codeparrot_training - Step 28914: {'lr': 0.0002026083181152134, 'samples': 14804480, 'steps': 28914, 'loss/train': 1.8935295343399048} 02/25/2022 15:02:28 - INFO - codeparrot_training - Step 28915: {'lr': 0.00020259225244176232, 'samples': 14804992, 'steps': 28915, 'loss/train': 1.7291104793548584} 02/25/2022 15:02:33 - INFO - codeparrot_training - Step 28916: {'lr': 0.00020257618697139086, 'samples': 14805504, 'steps': 28916, 'loss/train': 1.0121861696243286} 02/25/2022 15:02:37 - INFO - codeparrot_training - Step 28917: {'lr': 0.00020256012170416797, 'samples': 14806016, 'steps': 28917, 'loss/train': 1.8899732828140259} 02/25/2022 15:02:42 - INFO - codeparrot_training - Step 28918: {'lr': 0.00020254405664016245, 'samples': 14806528, 'steps': 28918, 'loss/train': 1.450128197669983} 02/25/2022 15:02:45 - INFO - codeparrot_training - Step 28919: {'lr': 0.0002025279917794431, 'samples': 14807040, 'steps': 28919, 'loss/train': 2.908566951751709} 02/25/2022 15:02:52 - INFO - codeparrot_training - Step 28920: {'lr': 0.0002025119271220789, 'samples': 14807552, 'steps': 28920, 'loss/train': 2.1035959720611572} 02/25/2022 15:02:55 - INFO - codeparrot_training - Step 28921: {'lr': 0.0002024958626681384, 'samples': 14808064, 'steps': 28921, 'loss/train': 1.8606675863265991} 02/25/2022 15:03:01 - INFO - codeparrot_training - Step 28922: {'lr': 0.00020247979841769053, 'samples': 14808576, 'steps': 28922, 'loss/train': 2.044541358947754} 02/25/2022 15:03:04 - INFO - codeparrot_training - Step 28923: {'lr': 0.00020246373437080413, 'samples': 14809088, 'steps': 28923, 'loss/train': 1.2998839616775513} 02/25/2022 15:03:10 - INFO - codeparrot_training - Step 28924: {'lr': 0.0002024476705275481, 'samples': 14809600, 'steps': 28924, 'loss/train': 0.5418252348899841} 02/25/2022 15:03:13 - INFO - codeparrot_training - Step 28925: {'lr': 0.00020243160688799105, 'samples': 14810112, 'steps': 28925, 'loss/train': 1.4740431308746338} 02/25/2022 15:03:19 - INFO - codeparrot_training - Step 28926: {'lr': 0.0002024155434522019, 'samples': 14810624, 'steps': 28926, 'loss/train': 1.7650465965270996} 02/25/2022 15:03:22 - INFO - codeparrot_training - Step 28927: {'lr': 0.00020239948022024941, 'samples': 14811136, 'steps': 28927, 'loss/train': 0.3023514449596405} 02/25/2022 15:03:28 - INFO - codeparrot_training - Step 28928: {'lr': 0.00020238341719220254, 'samples': 14811648, 'steps': 28928, 'loss/train': 2.5048794746398926} 02/25/2022 15:03:31 - INFO - codeparrot_training - Step 28929: {'lr': 0.00020236735436812996, 'samples': 14812160, 'steps': 28929, 'loss/train': 1.2107259035110474} 02/25/2022 15:03:37 - INFO - codeparrot_training - Step 28930: {'lr': 0.00020235129174810045, 'samples': 14812672, 'steps': 28930, 'loss/train': 1.3697885274887085} 02/25/2022 15:03:40 - INFO - codeparrot_training - Step 28931: {'lr': 0.0002023352293321829, 'samples': 14813184, 'steps': 28931, 'loss/train': 1.601084589958191} 02/25/2022 15:03:46 - INFO - codeparrot_training - Step 28932: {'lr': 0.00020231916712044613, 'samples': 14813696, 'steps': 28932, 'loss/train': 1.7738384008407593} 02/25/2022 15:03:50 - INFO - codeparrot_training - Step 28933: {'lr': 0.0002023031051129589, 'samples': 14814208, 'steps': 28933, 'loss/train': 2.0779154300689697} 02/25/2022 15:03:55 - INFO - codeparrot_training - Step 28934: {'lr': 0.00020228704330978997, 'samples': 14814720, 'steps': 28934, 'loss/train': 1.2004581689834595} 02/25/2022 15:03:59 - INFO - codeparrot_training - Step 28935: {'lr': 0.00020227098171100828, 'samples': 14815232, 'steps': 28935, 'loss/train': 0.09606925398111343} 02/25/2022 15:04:05 - INFO - codeparrot_training - Step 28936: {'lr': 0.00020225492031668247, 'samples': 14815744, 'steps': 28936, 'loss/train': 0.42571696639060974} 02/25/2022 15:04:08 - INFO - codeparrot_training - Step 28937: {'lr': 0.00020223885912688146, 'samples': 14816256, 'steps': 28937, 'loss/train': 2.1236870288848877} 02/25/2022 15:04:14 - INFO - codeparrot_training - Step 28938: {'lr': 0.00020222279814167398, 'samples': 14816768, 'steps': 28938, 'loss/train': 2.339834451675415} 02/25/2022 15:04:17 - INFO - codeparrot_training - Step 28939: {'lr': 0.0002022067373611289, 'samples': 14817280, 'steps': 28939, 'loss/train': 1.5197266340255737} 02/25/2022 15:04:23 - INFO - codeparrot_training - Step 28940: {'lr': 0.00020219067678531495, 'samples': 14817792, 'steps': 28940, 'loss/train': 2.158062696456909} 02/25/2022 15:04:26 - INFO - codeparrot_training - Step 28941: {'lr': 0.00020217461641430096, 'samples': 14818304, 'steps': 28941, 'loss/train': 1.1631979942321777} 02/25/2022 15:04:32 - INFO - codeparrot_training - Step 28942: {'lr': 0.0002021585562481558, 'samples': 14818816, 'steps': 28942, 'loss/train': 2.30741286277771} 02/25/2022 15:04:35 - INFO - codeparrot_training - Step 28943: {'lr': 0.0002021424962869481, 'samples': 14819328, 'steps': 28943, 'loss/train': 1.9290807247161865} 02/25/2022 15:04:41 - INFO - codeparrot_training - Step 28944: {'lr': 0.00020212643653074678, 'samples': 14819840, 'steps': 28944, 'loss/train': 1.7719894647598267} 02/25/2022 15:04:44 - INFO - codeparrot_training - Step 28945: {'lr': 0.00020211037697962056, 'samples': 14820352, 'steps': 28945, 'loss/train': 1.351855993270874} 02/25/2022 15:04:50 - INFO - codeparrot_training - Step 28946: {'lr': 0.0002020943176336384, 'samples': 14820864, 'steps': 28946, 'loss/train': 0.9579593539237976} 02/25/2022 15:04:54 - INFO - codeparrot_training - Step 28947: {'lr': 0.00020207825849286886, 'samples': 14821376, 'steps': 28947, 'loss/train': 2.261284828186035} 02/25/2022 15:04:59 - INFO - codeparrot_training - Step 28948: {'lr': 0.00020206219955738088, 'samples': 14821888, 'steps': 28948, 'loss/train': 0.973355233669281} 02/25/2022 15:05:03 - INFO - codeparrot_training - Step 28949: {'lr': 0.0002020461408272432, 'samples': 14822400, 'steps': 28949, 'loss/train': 1.0824620723724365} 02/25/2022 15:05:08 - INFO - codeparrot_training - Step 28950: {'lr': 0.00020203008230252473, 'samples': 14822912, 'steps': 28950, 'loss/train': 1.2817915678024292} 02/25/2022 15:05:12 - INFO - codeparrot_training - Step 28951: {'lr': 0.000202014023983294, 'samples': 14823424, 'steps': 28951, 'loss/train': 2.090498924255371} 02/25/2022 15:05:17 - INFO - codeparrot_training - Step 28952: {'lr': 0.00020199796586962003, 'samples': 14823936, 'steps': 28952, 'loss/train': 1.5789719820022583} 02/25/2022 15:05:21 - INFO - codeparrot_training - Step 28953: {'lr': 0.0002019819079615715, 'samples': 14824448, 'steps': 28953, 'loss/train': 2.967611789703369} 02/25/2022 15:05:26 - INFO - codeparrot_training - Step 28954: {'lr': 0.00020196585025921722, 'samples': 14824960, 'steps': 28954, 'loss/train': 1.1173735857009888} 02/25/2022 15:05:30 - INFO - codeparrot_training - Step 28955: {'lr': 0.0002019497927626261, 'samples': 14825472, 'steps': 28955, 'loss/train': 1.1480498313903809} 02/25/2022 15:05:36 - INFO - codeparrot_training - Step 28956: {'lr': 0.00020193373547186668, 'samples': 14825984, 'steps': 28956, 'loss/train': 1.3480191230773926} 02/25/2022 15:05:39 - INFO - codeparrot_training - Step 28957: {'lr': 0.00020191767838700792, 'samples': 14826496, 'steps': 28957, 'loss/train': 1.5765042304992676} 02/25/2022 15:05:45 - INFO - codeparrot_training - Step 28958: {'lr': 0.00020190162150811853, 'samples': 14827008, 'steps': 28958, 'loss/train': 1.3374444246292114} 02/25/2022 15:05:48 - INFO - codeparrot_training - Step 28959: {'lr': 0.00020188556483526743, 'samples': 14827520, 'steps': 28959, 'loss/train': 1.4985771179199219} 02/25/2022 15:05:54 - INFO - codeparrot_training - Step 28960: {'lr': 0.0002018695083685232, 'samples': 14828032, 'steps': 28960, 'loss/train': 2.812014579772949} 02/25/2022 15:05:57 - INFO - codeparrot_training - Step 28961: {'lr': 0.00020185345210795466, 'samples': 14828544, 'steps': 28961, 'loss/train': 1.8284573554992676} 02/25/2022 15:06:03 - INFO - codeparrot_training - Step 28962: {'lr': 0.00020183739605363069, 'samples': 14829056, 'steps': 28962, 'loss/train': 1.4798394441604614} 02/25/2022 15:06:06 - INFO - codeparrot_training - Step 28963: {'lr': 0.00020182134020562014, 'samples': 14829568, 'steps': 28963, 'loss/train': 1.6920136213302612} 02/25/2022 15:06:12 - INFO - codeparrot_training - Step 28964: {'lr': 0.00020180528456399153, 'samples': 14830080, 'steps': 28964, 'loss/train': 1.226874828338623} 02/25/2022 15:06:15 - INFO - codeparrot_training - Step 28965: {'lr': 0.00020178922912881378, 'samples': 14830592, 'steps': 28965, 'loss/train': 2.674668788909912} 02/25/2022 15:06:23 - INFO - codeparrot_training - Step 28966: {'lr': 0.00020177317390015568, 'samples': 14831104, 'steps': 28966, 'loss/train': 2.2185351848602295} 02/25/2022 15:06:26 - INFO - codeparrot_training - Step 28967: {'lr': 0.00020175711887808598, 'samples': 14831616, 'steps': 28967, 'loss/train': 1.9698344469070435} 02/25/2022 15:06:32 - INFO - codeparrot_training - Step 28968: {'lr': 0.0002017410640626736, 'samples': 14832128, 'steps': 28968, 'loss/train': 1.7015665769577026} 02/25/2022 15:06:37 - INFO - codeparrot_training - Step 28969: {'lr': 0.00020172500945398703, 'samples': 14832640, 'steps': 28969, 'loss/train': 2.2540230751037598} 02/25/2022 15:06:41 - INFO - codeparrot_training - Step 28970: {'lr': 0.00020170895505209523, 'samples': 14833152, 'steps': 28970, 'loss/train': 0.05459139123558998} 02/25/2022 15:06:46 - INFO - codeparrot_training - Step 28971: {'lr': 0.00020169290085706687, 'samples': 14833664, 'steps': 28971, 'loss/train': 1.7054938077926636} 02/25/2022 15:06:50 - INFO - codeparrot_training - Step 28972: {'lr': 0.00020167684686897094, 'samples': 14834176, 'steps': 28972, 'loss/train': 1.4739301204681396} 02/25/2022 15:06:55 - INFO - codeparrot_training - Step 28973: {'lr': 0.00020166079308787593, 'samples': 14834688, 'steps': 28973, 'loss/train': 1.7997164726257324} 02/25/2022 15:06:59 - INFO - codeparrot_training - Step 28974: {'lr': 0.00020164473951385077, 'samples': 14835200, 'steps': 28974, 'loss/train': 1.6903244256973267} 02/25/2022 15:07:04 - INFO - codeparrot_training - Step 28975: {'lr': 0.0002016286861469642, 'samples': 14835712, 'steps': 28975, 'loss/train': 2.206655263900757} 02/25/2022 15:07:08 - INFO - codeparrot_training - Step 28976: {'lr': 0.00020161263298728495, 'samples': 14836224, 'steps': 28976, 'loss/train': 3.481577157974243} 02/25/2022 15:07:14 - INFO - codeparrot_training - Step 28977: {'lr': 0.00020159658003488185, 'samples': 14836736, 'steps': 28977, 'loss/train': 1.877198576927185} 02/25/2022 15:07:17 - INFO - codeparrot_training - Step 28978: {'lr': 0.00020158052728982357, 'samples': 14837248, 'steps': 28978, 'loss/train': 1.6590062379837036} 02/25/2022 15:07:23 - INFO - codeparrot_training - Step 28979: {'lr': 0.000201564474752179, 'samples': 14837760, 'steps': 28979, 'loss/train': 1.7465327978134155} 02/25/2022 15:07:26 - INFO - codeparrot_training - Step 28980: {'lr': 0.00020154842242201682, 'samples': 14838272, 'steps': 28980, 'loss/train': 1.3785573244094849} 02/25/2022 15:07:32 - INFO - codeparrot_training - Step 28981: {'lr': 0.00020153237029940584, 'samples': 14838784, 'steps': 28981, 'loss/train': 1.5153241157531738} 02/25/2022 15:07:35 - INFO - codeparrot_training - Step 28982: {'lr': 0.00020151631838441478, 'samples': 14839296, 'steps': 28982, 'loss/train': 1.3197001218795776} 02/25/2022 15:07:42 - INFO - codeparrot_training - Step 28983: {'lr': 0.00020150026667711238, 'samples': 14839808, 'steps': 28983, 'loss/train': 0.2718430161476135} 02/25/2022 15:07:46 - INFO - codeparrot_training - Step 28984: {'lr': 0.00020148421517756743, 'samples': 14840320, 'steps': 28984, 'loss/train': 1.6264054775238037} 02/25/2022 15:07:51 - INFO - codeparrot_training - Step 28985: {'lr': 0.00020146816388584877, 'samples': 14840832, 'steps': 28985, 'loss/train': 2.099219799041748} 02/25/2022 15:07:55 - INFO - codeparrot_training - Step 28986: {'lr': 0.00020145211280202504, 'samples': 14841344, 'steps': 28986, 'loss/train': 1.719549298286438} 02/25/2022 15:08:00 - INFO - codeparrot_training - Step 28987: {'lr': 0.00020143606192616503, 'samples': 14841856, 'steps': 28987, 'loss/train': 0.9109772443771362} 02/25/2022 15:08:04 - INFO - codeparrot_training - Step 28988: {'lr': 0.0002014200112583375, 'samples': 14842368, 'steps': 28988, 'loss/train': 0.3388788402080536} 02/25/2022 15:08:09 - INFO - codeparrot_training - Step 28989: {'lr': 0.00020140396079861123, 'samples': 14842880, 'steps': 28989, 'loss/train': 1.8075129985809326} 02/25/2022 15:08:13 - INFO - codeparrot_training - Step 28990: {'lr': 0.00020138791054705505, 'samples': 14843392, 'steps': 28990, 'loss/train': 2.0343501567840576} 02/25/2022 15:08:18 - INFO - codeparrot_training - Step 28991: {'lr': 0.0002013718605037375, 'samples': 14843904, 'steps': 28991, 'loss/train': 1.957627773284912} 02/25/2022 15:08:22 - INFO - codeparrot_training - Step 28992: {'lr': 0.0002013558106687275, 'samples': 14844416, 'steps': 28992, 'loss/train': 1.6540087461471558} 02/25/2022 15:08:28 - INFO - codeparrot_training - Step 28993: {'lr': 0.00020133976104209372, 'samples': 14844928, 'steps': 28993, 'loss/train': 1.4951592683792114} 02/25/2022 15:08:32 - INFO - codeparrot_training - Step 28994: {'lr': 0.00020132371162390512, 'samples': 14845440, 'steps': 28994, 'loss/train': 1.3136905431747437} 02/25/2022 15:08:38 - INFO - codeparrot_training - Step 28995: {'lr': 0.0002013076624142301, 'samples': 14845952, 'steps': 28995, 'loss/train': 1.4378952980041504} 02/25/2022 15:08:41 - INFO - codeparrot_training - Step 28996: {'lr': 0.00020129161341313765, 'samples': 14846464, 'steps': 28996, 'loss/train': 0.17468242347240448} 02/25/2022 15:08:46 - INFO - codeparrot_training - Step 28997: {'lr': 0.00020127556462069644, 'samples': 14846976, 'steps': 28997, 'loss/train': 0.5771901607513428} 02/25/2022 15:08:50 - INFO - codeparrot_training - Step 28998: {'lr': 0.00020125951603697535, 'samples': 14847488, 'steps': 28998, 'loss/train': 1.5860698223114014} 02/25/2022 15:08:55 - INFO - codeparrot_training - Step 28999: {'lr': 0.0002012434676620429, 'samples': 14848000, 'steps': 28999, 'loss/train': 0.6781620979309082} 02/25/2022 15:08:55 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 15:09:13 - WARNING - huggingface_hub.repository - Several commits (29) will be pushed upstream. 02/25/2022 15:09:13 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 15:09:53 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 739c14e..95f22fa floral-grass-11 -> floral-grass-11 02/25/2022 15:09:58 - INFO - codeparrot_training - Step 29000: {'lr': 0.00020122741949596797, 'samples': 14848512, 'steps': 29000, 'loss/train': 0.9625966548919678} 02/25/2022 15:10:03 - INFO - codeparrot_training - Step 29001: {'lr': 0.00020121137153881924, 'samples': 14849024, 'steps': 29001, 'loss/train': 3.0134713649749756} 02/25/2022 15:10:07 - INFO - codeparrot_training - Step 29002: {'lr': 0.00020119532379066557, 'samples': 14849536, 'steps': 29002, 'loss/train': 1.2822824716567993} 02/25/2022 15:10:13 - INFO - codeparrot_training - Step 29003: {'lr': 0.0002011792762515757, 'samples': 14850048, 'steps': 29003, 'loss/train': 2.368891477584839} 02/25/2022 15:10:16 - INFO - codeparrot_training - Step 29004: {'lr': 0.0002011632289216182, 'samples': 14850560, 'steps': 29004, 'loss/train': 2.1708297729492188} 02/25/2022 15:10:22 - INFO - codeparrot_training - Step 29005: {'lr': 0.00020114718180086195, 'samples': 14851072, 'steps': 29005, 'loss/train': 2.9230146408081055} 02/25/2022 15:10:25 - INFO - codeparrot_training - Step 29006: {'lr': 0.0002011311348893756, 'samples': 14851584, 'steps': 29006, 'loss/train': 1.6136120557785034} 02/25/2022 15:10:31 - INFO - codeparrot_training - Step 29007: {'lr': 0.00020111508818722814, 'samples': 14852096, 'steps': 29007, 'loss/train': 4.313667297363281} 02/25/2022 15:10:35 - INFO - codeparrot_training - Step 29008: {'lr': 0.00020109904169448797, 'samples': 14852608, 'steps': 29008, 'loss/train': 0.40597519278526306} 02/25/2022 15:10:40 - INFO - codeparrot_training - Step 29009: {'lr': 0.00020108299541122398, 'samples': 14853120, 'steps': 29009, 'loss/train': 1.30682373046875} 02/25/2022 15:10:44 - INFO - codeparrot_training - Step 29010: {'lr': 0.00020106694933750492, 'samples': 14853632, 'steps': 29010, 'loss/train': 2.0005903244018555} 02/25/2022 15:10:49 - INFO - codeparrot_training - Step 29011: {'lr': 0.00020105090347339962, 'samples': 14854144, 'steps': 29011, 'loss/train': 1.2578452825546265} 02/25/2022 15:10:53 - INFO - codeparrot_training - Step 29012: {'lr': 0.00020103485781897658, 'samples': 14854656, 'steps': 29012, 'loss/train': 1.7885220050811768} 02/25/2022 15:10:58 - INFO - codeparrot_training - Step 29013: {'lr': 0.00020101881237430465, 'samples': 14855168, 'steps': 29013, 'loss/train': 1.321237564086914} 02/25/2022 15:11:02 - INFO - codeparrot_training - Step 29014: {'lr': 0.00020100276713945259, 'samples': 14855680, 'steps': 29014, 'loss/train': 0.7380515336990356} 02/25/2022 15:11:08 - INFO - codeparrot_training - Step 29015: {'lr': 0.00020098672211448914, 'samples': 14856192, 'steps': 29015, 'loss/train': 1.0846383571624756} 02/25/2022 15:11:11 - INFO - codeparrot_training - Step 29016: {'lr': 0.00020097067729948304, 'samples': 14856704, 'steps': 29016, 'loss/train': 1.660983681678772} 02/25/2022 15:11:17 - INFO - codeparrot_training - Step 29017: {'lr': 0.00020095463269450295, 'samples': 14857216, 'steps': 29017, 'loss/train': 1.8459781408309937} 02/25/2022 15:11:20 - INFO - codeparrot_training - Step 29018: {'lr': 0.00020093858829961763, 'samples': 14857728, 'steps': 29018, 'loss/train': 2.4474856853485107} 02/25/2022 15:11:26 - INFO - codeparrot_training - Step 29019: {'lr': 0.00020092254411489585, 'samples': 14858240, 'steps': 29019, 'loss/train': 2.3780765533447266} 02/25/2022 15:11:29 - INFO - codeparrot_training - Step 29020: {'lr': 0.00020090650014040635, 'samples': 14858752, 'steps': 29020, 'loss/train': 1.4674705266952515} 02/25/2022 15:11:35 - INFO - codeparrot_training - Step 29021: {'lr': 0.00020089045637621772, 'samples': 14859264, 'steps': 29021, 'loss/train': 1.916792869567871} 02/25/2022 15:11:38 - INFO - codeparrot_training - Step 29022: {'lr': 0.00020087441282239885, 'samples': 14859776, 'steps': 29022, 'loss/train': 2.1213696002960205} 02/25/2022 15:11:44 - INFO - codeparrot_training - Step 29023: {'lr': 0.00020085836947901834, 'samples': 14860288, 'steps': 29023, 'loss/train': 1.4662578105926514} 02/25/2022 15:11:47 - INFO - codeparrot_training - Step 29024: {'lr': 0.00020084232634614503, 'samples': 14860800, 'steps': 29024, 'loss/train': 1.986836552619934} 02/25/2022 15:11:53 - INFO - codeparrot_training - Step 29025: {'lr': 0.00020082628342384752, 'samples': 14861312, 'steps': 29025, 'loss/train': 2.443887948989868} 02/25/2022 15:11:56 - INFO - codeparrot_training - Step 29026: {'lr': 0.00020081024071219467, 'samples': 14861824, 'steps': 29026, 'loss/train': 1.7471908330917358} 02/25/2022 15:12:02 - INFO - codeparrot_training - Step 29027: {'lr': 0.00020079419821125504, 'samples': 14862336, 'steps': 29027, 'loss/train': 2.382305860519409} 02/25/2022 15:12:05 - INFO - codeparrot_training - Step 29028: {'lr': 0.0002007781559210975, 'samples': 14862848, 'steps': 29028, 'loss/train': 1.7041689157485962} 02/25/2022 15:12:12 - INFO - codeparrot_training - Step 29029: {'lr': 0.00020076211384179078, 'samples': 14863360, 'steps': 29029, 'loss/train': 1.5009945631027222} 02/25/2022 15:12:15 - INFO - codeparrot_training - Step 29030: {'lr': 0.0002007460719734034, 'samples': 14863872, 'steps': 29030, 'loss/train': 2.061202049255371} 02/25/2022 15:12:21 - INFO - codeparrot_training - Step 29031: {'lr': 0.00020073003031600422, 'samples': 14864384, 'steps': 29031, 'loss/train': 1.6635651588439941} 02/25/2022 15:12:26 - INFO - codeparrot_training - Step 29032: {'lr': 0.00020071398886966198, 'samples': 14864896, 'steps': 29032, 'loss/train': 2.2170233726501465} 02/25/2022 15:12:30 - INFO - codeparrot_training - Step 29033: {'lr': 0.00020069794763444545, 'samples': 14865408, 'steps': 29033, 'loss/train': 0.6922512650489807} 02/25/2022 15:12:35 - INFO - codeparrot_training - Step 29034: {'lr': 0.00020068190661042315, 'samples': 14865920, 'steps': 29034, 'loss/train': 1.869996190071106} 02/25/2022 15:12:39 - INFO - codeparrot_training - Step 29035: {'lr': 0.00020066586579766389, 'samples': 14866432, 'steps': 29035, 'loss/train': 1.3881815671920776} 02/25/2022 15:12:44 - INFO - codeparrot_training - Step 29036: {'lr': 0.0002006498251962364, 'samples': 14866944, 'steps': 29036, 'loss/train': 2.233100652694702} 02/25/2022 15:12:48 - INFO - codeparrot_training - Step 29037: {'lr': 0.0002006337848062094, 'samples': 14867456, 'steps': 29037, 'loss/train': 0.6549373269081116} 02/25/2022 15:12:53 - INFO - codeparrot_training - Step 29038: {'lr': 0.0002006177446276517, 'samples': 14867968, 'steps': 29038, 'loss/train': 2.5598695278167725} 02/25/2022 15:12:57 - INFO - codeparrot_training - Step 29039: {'lr': 0.00020060170466063178, 'samples': 14868480, 'steps': 29039, 'loss/train': 2.1152689456939697} 02/25/2022 15:13:03 - INFO - codeparrot_training - Step 29040: {'lr': 0.00020058566490521845, 'samples': 14868992, 'steps': 29040, 'loss/train': 1.7224782705307007} 02/25/2022 15:13:06 - INFO - codeparrot_training - Step 29041: {'lr': 0.00020056962536148048, 'samples': 14869504, 'steps': 29041, 'loss/train': 1.044476866722107} 02/25/2022 15:13:12 - INFO - codeparrot_training - Step 29042: {'lr': 0.00020055358602948665, 'samples': 14870016, 'steps': 29042, 'loss/train': 0.026592062786221504} 02/25/2022 15:13:15 - INFO - codeparrot_training - Step 29043: {'lr': 0.00020053754690930542, 'samples': 14870528, 'steps': 29043, 'loss/train': 1.5946857929229736} 02/25/2022 15:13:21 - INFO - codeparrot_training - Step 29044: {'lr': 0.00020052150800100565, 'samples': 14871040, 'steps': 29044, 'loss/train': 2.780329704284668} 02/25/2022 15:13:25 - INFO - codeparrot_training - Step 29045: {'lr': 0.00020050546930465603, 'samples': 14871552, 'steps': 29045, 'loss/train': 0.9196075797080994} 02/25/2022 15:13:30 - INFO - codeparrot_training - Step 29046: {'lr': 0.00020048943082032533, 'samples': 14872064, 'steps': 29046, 'loss/train': 0.9458131790161133} 02/25/2022 15:13:34 - INFO - codeparrot_training - Step 29047: {'lr': 0.00020047339254808215, 'samples': 14872576, 'steps': 29047, 'loss/train': 1.834579586982727} 02/25/2022 15:13:40 - INFO - codeparrot_training - Step 29048: {'lr': 0.0002004573544879952, 'samples': 14873088, 'steps': 29048, 'loss/train': 1.1861038208007812} 02/25/2022 15:13:44 - INFO - codeparrot_training - Step 29049: {'lr': 0.00020044131664013318, 'samples': 14873600, 'steps': 29049, 'loss/train': 1.1611157655715942} 02/25/2022 15:13:49 - INFO - codeparrot_training - Step 29050: {'lr': 0.00020042527900456485, 'samples': 14874112, 'steps': 29050, 'loss/train': 1.5956506729125977} 02/25/2022 15:13:53 - INFO - codeparrot_training - Step 29051: {'lr': 0.00020040924158135897, 'samples': 14874624, 'steps': 29051, 'loss/train': 1.0758423805236816} 02/25/2022 15:13:58 - INFO - codeparrot_training - Step 29052: {'lr': 0.00020039320437058409, 'samples': 14875136, 'steps': 29052, 'loss/train': 1.390692949295044} 02/25/2022 15:14:02 - INFO - codeparrot_training - Step 29053: {'lr': 0.00020037716737230892, 'samples': 14875648, 'steps': 29053, 'loss/train': 1.2980892658233643} 02/25/2022 15:14:08 - INFO - codeparrot_training - Step 29054: {'lr': 0.00020036113058660224, 'samples': 14876160, 'steps': 29054, 'loss/train': 1.3055460453033447} 02/25/2022 15:14:11 - INFO - codeparrot_training - Step 29055: {'lr': 0.00020034509401353282, 'samples': 14876672, 'steps': 29055, 'loss/train': 0.11481061577796936} 02/25/2022 15:14:17 - INFO - codeparrot_training - Step 29056: {'lr': 0.00020032905765316914, 'samples': 14877184, 'steps': 29056, 'loss/train': 1.6125483512878418} 02/25/2022 15:14:20 - INFO - codeparrot_training - Step 29057: {'lr': 0.00020031302150558, 'samples': 14877696, 'steps': 29057, 'loss/train': 2.3205788135528564} 02/25/2022 15:14:26 - INFO - codeparrot_training - Step 29058: {'lr': 0.0002002969855708341, 'samples': 14878208, 'steps': 29058, 'loss/train': 1.8551654815673828} 02/25/2022 15:14:29 - INFO - codeparrot_training - Step 29059: {'lr': 0.0002002809498490002, 'samples': 14878720, 'steps': 29059, 'loss/train': 1.5854817628860474} 02/25/2022 15:14:35 - INFO - codeparrot_training - Step 29060: {'lr': 0.00020026491434014688, 'samples': 14879232, 'steps': 29060, 'loss/train': 2.3203086853027344} 02/25/2022 15:14:38 - INFO - codeparrot_training - Step 29061: {'lr': 0.00020024887904434286, 'samples': 14879744, 'steps': 29061, 'loss/train': 1.2007865905761719} 02/25/2022 15:14:44 - INFO - codeparrot_training - Step 29062: {'lr': 0.00020023284396165686, 'samples': 14880256, 'steps': 29062, 'loss/train': 1.4887027740478516} 02/25/2022 15:14:47 - INFO - codeparrot_training - Step 29063: {'lr': 0.00020021680909215755, 'samples': 14880768, 'steps': 29063, 'loss/train': 2.570343017578125} 02/25/2022 15:14:53 - INFO - codeparrot_training - Step 29064: {'lr': 0.00020020077443591363, 'samples': 14881280, 'steps': 29064, 'loss/train': 1.6657326221466064} 02/25/2022 15:14:57 - INFO - codeparrot_training - Step 29065: {'lr': 0.00020018473999299375, 'samples': 14881792, 'steps': 29065, 'loss/train': 2.588829755783081} 02/25/2022 15:15:03 - INFO - codeparrot_training - Step 29066: {'lr': 0.00020016870576346668, 'samples': 14882304, 'steps': 29066, 'loss/train': 1.1490429639816284} 02/25/2022 15:15:06 - INFO - codeparrot_training - Step 29067: {'lr': 0.00020015267174740102, 'samples': 14882816, 'steps': 29067, 'loss/train': 1.9579099416732788} 02/25/2022 15:15:12 - INFO - codeparrot_training - Step 29068: {'lr': 0.0002001366379448655, 'samples': 14883328, 'steps': 29068, 'loss/train': 0.08411605656147003} 02/25/2022 15:15:15 - INFO - codeparrot_training - Step 29069: {'lr': 0.0002001206043559288, 'samples': 14883840, 'steps': 29069, 'loss/train': 2.3117170333862305} 02/25/2022 15:15:21 - INFO - codeparrot_training - Step 29070: {'lr': 0.00020010457098065953, 'samples': 14884352, 'steps': 29070, 'loss/train': 1.3876031637191772} 02/25/2022 15:15:24 - INFO - codeparrot_training - Step 29071: {'lr': 0.00020008853781912646, 'samples': 14884864, 'steps': 29071, 'loss/train': 2.474170207977295} 02/25/2022 15:15:30 - INFO - codeparrot_training - Step 29072: {'lr': 0.00020007250487139827, 'samples': 14885376, 'steps': 29072, 'loss/train': 0.9779680371284485} 02/25/2022 15:15:33 - INFO - codeparrot_training - Step 29073: {'lr': 0.00020005647213754363, 'samples': 14885888, 'steps': 29073, 'loss/train': 1.6422357559204102} 02/25/2022 15:15:39 - INFO - codeparrot_training - Step 29074: {'lr': 0.00020004043961763116, 'samples': 14886400, 'steps': 29074, 'loss/train': 2.2321534156799316} 02/25/2022 15:15:42 - INFO - codeparrot_training - Step 29075: {'lr': 0.00020002440731172957, 'samples': 14886912, 'steps': 29075, 'loss/train': 1.9346281290054321} 02/25/2022 15:15:49 - INFO - codeparrot_training - Step 29076: {'lr': 0.00020000837521990755, 'samples': 14887424, 'steps': 29076, 'loss/train': 1.2649942636489868} 02/25/2022 15:15:52 - INFO - codeparrot_training - Step 29077: {'lr': 0.0001999923433422339, 'samples': 14887936, 'steps': 29077, 'loss/train': 1.595371127128601} 02/25/2022 15:15:58 - INFO - codeparrot_training - Step 29078: {'lr': 0.00019997631167877706, 'samples': 14888448, 'steps': 29078, 'loss/train': 2.4133219718933105} 02/25/2022 15:16:01 - INFO - codeparrot_training - Step 29079: {'lr': 0.0001999602802296058, 'samples': 14888960, 'steps': 29079, 'loss/train': 2.6212804317474365} 02/25/2022 15:16:07 - INFO - codeparrot_training - Step 29080: {'lr': 0.00019994424899478884, 'samples': 14889472, 'steps': 29080, 'loss/train': 1.28313148021698} 02/25/2022 15:16:10 - INFO - codeparrot_training - Step 29081: {'lr': 0.0001999282179743949, 'samples': 14889984, 'steps': 29081, 'loss/train': 0.08394929021596909} 02/25/2022 15:16:16 - INFO - codeparrot_training - Step 29082: {'lr': 0.00019991218716849246, 'samples': 14890496, 'steps': 29082, 'loss/train': 2.2583987712860107} 02/25/2022 15:16:19 - INFO - codeparrot_training - Step 29083: {'lr': 0.00019989615657715033, 'samples': 14891008, 'steps': 29083, 'loss/train': 1.5651559829711914} 02/25/2022 15:16:25 - INFO - codeparrot_training - Step 29084: {'lr': 0.00019988012620043716, 'samples': 14891520, 'steps': 29084, 'loss/train': 2.0740163326263428} 02/25/2022 15:16:28 - INFO - codeparrot_training - Step 29085: {'lr': 0.00019986409603842158, 'samples': 14892032, 'steps': 29085, 'loss/train': 2.237011194229126} 02/25/2022 15:16:35 - INFO - codeparrot_training - Step 29086: {'lr': 0.00019984806609117247, 'samples': 14892544, 'steps': 29086, 'loss/train': 1.4755436182022095} 02/25/2022 15:16:38 - INFO - codeparrot_training - Step 29087: {'lr': 0.00019983203635875812, 'samples': 14893056, 'steps': 29087, 'loss/train': 1.7659515142440796} 02/25/2022 15:16:44 - INFO - codeparrot_training - Step 29088: {'lr': 0.00019981600684124744, 'samples': 14893568, 'steps': 29088, 'loss/train': 2.350338935852051} 02/25/2022 15:16:47 - INFO - codeparrot_training - Step 29089: {'lr': 0.00019979997753870907, 'samples': 14894080, 'steps': 29089, 'loss/train': 2.077547311782837} 02/25/2022 15:16:53 - INFO - codeparrot_training - Step 29090: {'lr': 0.00019978394845121173, 'samples': 14894592, 'steps': 29090, 'loss/train': 1.985950231552124} 02/25/2022 15:16:56 - INFO - codeparrot_training - Step 29091: {'lr': 0.00019976791957882392, 'samples': 14895104, 'steps': 29091, 'loss/train': 1.5815422534942627} 02/25/2022 15:17:02 - INFO - codeparrot_training - Step 29092: {'lr': 0.0001997518909216144, 'samples': 14895616, 'steps': 29092, 'loss/train': 2.0188963413238525} 02/25/2022 15:17:05 - INFO - codeparrot_training - Step 29093: {'lr': 0.0001997358624796518, 'samples': 14896128, 'steps': 29093, 'loss/train': 2.1548421382904053} 02/25/2022 15:17:11 - INFO - codeparrot_training - Step 29094: {'lr': 0.00019971983425300495, 'samples': 14896640, 'steps': 29094, 'loss/train': 2.8655264377593994} 02/25/2022 15:17:14 - INFO - codeparrot_training - Step 29095: {'lr': 0.00019970380624174223, 'samples': 14897152, 'steps': 29095, 'loss/train': 1.9505460262298584} 02/25/2022 15:17:21 - INFO - codeparrot_training - Step 29096: {'lr': 0.0001996877784459324, 'samples': 14897664, 'steps': 29096, 'loss/train': 1.06796395778656} 02/25/2022 15:17:27 - INFO - codeparrot_training - Step 29097: {'lr': 0.00019967175086564422, 'samples': 14898176, 'steps': 29097, 'loss/train': 2.072253704071045} 02/25/2022 15:17:30 - INFO - codeparrot_training - Step 29098: {'lr': 0.00019965572350094625, 'samples': 14898688, 'steps': 29098, 'loss/train': 0.9562056064605713} 02/25/2022 15:17:36 - INFO - codeparrot_training - Step 29099: {'lr': 0.00019963969635190725, 'samples': 14899200, 'steps': 29099, 'loss/train': 1.3275948762893677} 02/25/2022 15:17:39 - INFO - codeparrot_training - Step 29100: {'lr': 0.00019962366941859573, 'samples': 14899712, 'steps': 29100, 'loss/train': 1.411186695098877} 02/25/2022 15:17:44 - INFO - codeparrot_training - Step 29101: {'lr': 0.0001996076427010804, 'samples': 14900224, 'steps': 29101, 'loss/train': 0.07481779903173447} 02/25/2022 15:17:48 - INFO - codeparrot_training - Step 29102: {'lr': 0.00019959161619943002, 'samples': 14900736, 'steps': 29102, 'loss/train': 1.9677882194519043} 02/25/2022 15:17:53 - INFO - codeparrot_training - Step 29103: {'lr': 0.00019957558991371308, 'samples': 14901248, 'steps': 29103, 'loss/train': 1.8027527332305908} 02/25/2022 15:17:57 - INFO - codeparrot_training - Step 29104: {'lr': 0.0001995595638439983, 'samples': 14901760, 'steps': 29104, 'loss/train': 0.16070152819156647} 02/25/2022 15:18:02 - INFO - codeparrot_training - Step 29105: {'lr': 0.0001995435379903543, 'samples': 14902272, 'steps': 29105, 'loss/train': 1.4582862854003906} 02/25/2022 15:18:06 - INFO - codeparrot_training - Step 29106: {'lr': 0.00019952751235284983, 'samples': 14902784, 'steps': 29106, 'loss/train': 3.9073355197906494} 02/25/2022 15:18:11 - INFO - codeparrot_training - Step 29107: {'lr': 0.00019951148693155346, 'samples': 14903296, 'steps': 29107, 'loss/train': 0.9861534833908081} 02/25/2022 15:18:15 - INFO - codeparrot_training - Step 29108: {'lr': 0.0001994954617265338, 'samples': 14903808, 'steps': 29108, 'loss/train': 1.1505014896392822} 02/25/2022 15:18:20 - INFO - codeparrot_training - Step 29109: {'lr': 0.00019947943673785963, 'samples': 14904320, 'steps': 29109, 'loss/train': 1.6635161638259888} 02/25/2022 15:18:24 - INFO - codeparrot_training - Step 29110: {'lr': 0.00019946341196559942, 'samples': 14904832, 'steps': 29110, 'loss/train': 2.057328462600708} 02/25/2022 15:18:30 - INFO - codeparrot_training - Step 29111: {'lr': 0.00019944738740982192, 'samples': 14905344, 'steps': 29111, 'loss/train': 1.3910659551620483} 02/25/2022 15:18:34 - INFO - codeparrot_training - Step 29112: {'lr': 0.00019943136307059577, 'samples': 14905856, 'steps': 29112, 'loss/train': 2.2051427364349365} 02/25/2022 15:18:39 - INFO - codeparrot_training - Step 29113: {'lr': 0.00019941533894798967, 'samples': 14906368, 'steps': 29113, 'loss/train': 1.7288485765457153} 02/25/2022 15:18:43 - INFO - codeparrot_training - Step 29114: {'lr': 0.00019939931504207208, 'samples': 14906880, 'steps': 29114, 'loss/train': 2.2145586013793945} 02/25/2022 15:18:48 - INFO - codeparrot_training - Step 29115: {'lr': 0.0001993832913529118, 'samples': 14907392, 'steps': 29115, 'loss/train': 1.4452009201049805} 02/25/2022 15:18:52 - INFO - codeparrot_training - Step 29116: {'lr': 0.0001993672678805775, 'samples': 14907904, 'steps': 29116, 'loss/train': 2.066502809524536} 02/25/2022 15:18:57 - INFO - codeparrot_training - Step 29117: {'lr': 0.0001993512446251377, 'samples': 14908416, 'steps': 29117, 'loss/train': 0.6649622321128845} 02/25/2022 15:19:01 - INFO - codeparrot_training - Step 29118: {'lr': 0.00019933522158666108, 'samples': 14908928, 'steps': 29118, 'loss/train': 3.3403406143188477} 02/25/2022 15:19:06 - INFO - codeparrot_training - Step 29119: {'lr': 0.00019931919876521622, 'samples': 14909440, 'steps': 29119, 'loss/train': 1.5797144174575806} 02/25/2022 15:19:10 - INFO - codeparrot_training - Step 29120: {'lr': 0.00019930317616087195, 'samples': 14909952, 'steps': 29120, 'loss/train': 1.4475345611572266} 02/25/2022 15:19:16 - INFO - codeparrot_training - Step 29121: {'lr': 0.0001992871537736967, 'samples': 14910464, 'steps': 29121, 'loss/train': 1.509169101715088} 02/25/2022 15:19:19 - INFO - codeparrot_training - Step 29122: {'lr': 0.00019927113160375914, 'samples': 14910976, 'steps': 29122, 'loss/train': 1.3574632406234741} 02/25/2022 15:19:25 - INFO - codeparrot_training - Step 29123: {'lr': 0.00019925510965112798, 'samples': 14911488, 'steps': 29123, 'loss/train': 2.3240227699279785} 02/25/2022 15:19:28 - INFO - codeparrot_training - Step 29124: {'lr': 0.00019923908791587177, 'samples': 14912000, 'steps': 29124, 'loss/train': 2.033957004547119} 02/25/2022 15:19:34 - INFO - codeparrot_training - Step 29125: {'lr': 0.00019922306639805933, 'samples': 14912512, 'steps': 29125, 'loss/train': 1.8728193044662476} 02/25/2022 15:19:37 - INFO - codeparrot_training - Step 29126: {'lr': 0.00019920704509775904, 'samples': 14913024, 'steps': 29126, 'loss/train': 1.3663140535354614} 02/25/2022 15:19:43 - INFO - codeparrot_training - Step 29127: {'lr': 0.00019919102401503964, 'samples': 14913536, 'steps': 29127, 'loss/train': 2.560606002807617} 02/25/2022 15:19:46 - INFO - codeparrot_training - Step 29128: {'lr': 0.0001991750031499698, 'samples': 14914048, 'steps': 29128, 'loss/train': 0.6288747787475586} 02/25/2022 15:19:52 - INFO - codeparrot_training - Step 29129: {'lr': 0.00019915898250261814, 'samples': 14914560, 'steps': 29129, 'loss/train': 1.8293359279632568} 02/25/2022 15:19:55 - INFO - codeparrot_training - Step 29130: {'lr': 0.0001991429620730532, 'samples': 14915072, 'steps': 29130, 'loss/train': 1.4865890741348267} 02/25/2022 15:20:02 - INFO - codeparrot_training - Step 29131: {'lr': 0.0001991269418613436, 'samples': 14915584, 'steps': 29131, 'loss/train': 1.2068628072738647} 02/25/2022 15:20:05 - INFO - codeparrot_training - Step 29132: {'lr': 0.00019911092186755808, 'samples': 14916096, 'steps': 29132, 'loss/train': 3.7082173824310303} 02/25/2022 15:20:10 - INFO - codeparrot_training - Step 29133: {'lr': 0.0001990949020917653, 'samples': 14916608, 'steps': 29133, 'loss/train': 1.8108493089675903} 02/25/2022 15:20:14 - INFO - codeparrot_training - Step 29134: {'lr': 0.00019907888253403372, 'samples': 14917120, 'steps': 29134, 'loss/train': 1.4859955310821533} 02/25/2022 15:20:20 - INFO - codeparrot_training - Step 29135: {'lr': 0.00019906286319443202, 'samples': 14917632, 'steps': 29135, 'loss/train': 0.7468170523643494} 02/25/2022 15:20:23 - INFO - codeparrot_training - Step 29136: {'lr': 0.00019904684407302882, 'samples': 14918144, 'steps': 29136, 'loss/train': 2.2013497352600098} 02/25/2022 15:20:29 - INFO - codeparrot_training - Step 29137: {'lr': 0.00019903082516989275, 'samples': 14918656, 'steps': 29137, 'loss/train': 1.5608460903167725} 02/25/2022 15:20:34 - INFO - codeparrot_training - Step 29138: {'lr': 0.0001990148064850926, 'samples': 14919168, 'steps': 29138, 'loss/train': 1.5630799531936646} 02/25/2022 15:20:38 - INFO - codeparrot_training - Step 29139: {'lr': 0.0001989987880186967, 'samples': 14919680, 'steps': 29139, 'loss/train': 1.295097827911377} 02/25/2022 15:20:44 - INFO - codeparrot_training - Step 29140: {'lr': 0.0001989827697707738, 'samples': 14920192, 'steps': 29140, 'loss/train': 2.077880859375} 02/25/2022 15:20:47 - INFO - codeparrot_training - Step 29141: {'lr': 0.0001989667517413925, 'samples': 14920704, 'steps': 29141, 'loss/train': 0.9966211915016174} 02/25/2022 15:20:53 - INFO - codeparrot_training - Step 29142: {'lr': 0.00019895073393062153, 'samples': 14921216, 'steps': 29142, 'loss/train': 0.6767629384994507} 02/25/2022 15:20:56 - INFO - codeparrot_training - Step 29143: {'lr': 0.0001989347163385293, 'samples': 14921728, 'steps': 29143, 'loss/train': 1.9332820177078247} 02/25/2022 15:21:02 - INFO - codeparrot_training - Step 29144: {'lr': 0.00019891869896518455, 'samples': 14922240, 'steps': 29144, 'loss/train': 2.0521185398101807} 02/25/2022 15:21:05 - INFO - codeparrot_training - Step 29145: {'lr': 0.00019890268181065585, 'samples': 14922752, 'steps': 29145, 'loss/train': 0.08609345555305481} 02/25/2022 15:21:11 - INFO - codeparrot_training - Step 29146: {'lr': 0.00019888666487501193, 'samples': 14923264, 'steps': 29146, 'loss/train': 4.454577922821045} 02/25/2022 15:21:14 - INFO - codeparrot_training - Step 29147: {'lr': 0.00019887064815832126, 'samples': 14923776, 'steps': 29147, 'loss/train': 0.8347738981246948} 02/25/2022 15:21:20 - INFO - codeparrot_training - Step 29148: {'lr': 0.00019885463166065248, 'samples': 14924288, 'steps': 29148, 'loss/train': 1.6879264116287231} 02/25/2022 15:21:23 - INFO - codeparrot_training - Step 29149: {'lr': 0.00019883861538207427, 'samples': 14924800, 'steps': 29149, 'loss/train': 2.151691198348999} 02/25/2022 15:21:29 - INFO - codeparrot_training - Step 29150: {'lr': 0.00019882259932265512, 'samples': 14925312, 'steps': 29150, 'loss/train': 1.461527705192566} 02/25/2022 15:21:33 - INFO - codeparrot_training - Step 29151: {'lr': 0.00019880658348246375, 'samples': 14925824, 'steps': 29151, 'loss/train': 0.7858912944793701} 02/25/2022 15:21:36 - INFO - codeparrot_training - Step 29152: {'lr': 0.00019879056786156865, 'samples': 14926336, 'steps': 29152, 'loss/train': 1.5283597707748413} 02/25/2022 15:21:42 - INFO - codeparrot_training - Step 29153: {'lr': 0.00019877455246003858, 'samples': 14926848, 'steps': 29153, 'loss/train': 1.1754670143127441} 02/25/2022 15:21:46 - INFO - codeparrot_training - Step 29154: {'lr': 0.00019875853727794202, 'samples': 14927360, 'steps': 29154, 'loss/train': 1.646106481552124} 02/25/2022 15:21:51 - INFO - codeparrot_training - Step 29155: {'lr': 0.00019874252231534765, 'samples': 14927872, 'steps': 29155, 'loss/train': 1.1370307207107544} 02/25/2022 15:21:55 - INFO - codeparrot_training - Step 29156: {'lr': 0.00019872650757232397, 'samples': 14928384, 'steps': 29156, 'loss/train': 1.2444932460784912} 02/25/2022 15:22:00 - INFO - codeparrot_training - Step 29157: {'lr': 0.00019871049304893973, 'samples': 14928896, 'steps': 29157, 'loss/train': 1.2231873273849487} 02/25/2022 15:22:04 - INFO - codeparrot_training - Step 29158: {'lr': 0.0001986944787452634, 'samples': 14929408, 'steps': 29158, 'loss/train': 1.5079692602157593} 02/25/2022 15:22:09 - INFO - codeparrot_training - Step 29159: {'lr': 0.0001986784646613636, 'samples': 14929920, 'steps': 29159, 'loss/train': 2.841705799102783} 02/25/2022 15:22:13 - INFO - codeparrot_training - Step 29160: {'lr': 0.0001986624507973091, 'samples': 14930432, 'steps': 29160, 'loss/train': 1.5490890741348267} 02/25/2022 15:22:18 - INFO - codeparrot_training - Step 29161: {'lr': 0.00019864643715316823, 'samples': 14930944, 'steps': 29161, 'loss/train': 0.963879406452179} 02/25/2022 15:22:22 - INFO - codeparrot_training - Step 29162: {'lr': 0.00019863042372900974, 'samples': 14931456, 'steps': 29162, 'loss/train': 1.7841893434524536} 02/25/2022 15:22:27 - INFO - codeparrot_training - Step 29163: {'lr': 0.00019861441052490216, 'samples': 14931968, 'steps': 29163, 'loss/train': 1.379916787147522} 02/25/2022 15:22:31 - INFO - codeparrot_training - Step 29164: {'lr': 0.0001985983975409143, 'samples': 14932480, 'steps': 29164, 'loss/train': 1.6264894008636475} 02/25/2022 15:22:36 - INFO - codeparrot_training - Step 29165: {'lr': 0.00019858238477711443, 'samples': 14932992, 'steps': 29165, 'loss/train': 1.7188951969146729} 02/25/2022 15:22:40 - INFO - codeparrot_training - Step 29166: {'lr': 0.0001985663722335713, 'samples': 14933504, 'steps': 29166, 'loss/train': 0.8646218180656433} 02/25/2022 15:22:46 - INFO - codeparrot_training - Step 29167: {'lr': 0.00019855035991035353, 'samples': 14934016, 'steps': 29167, 'loss/train': 2.0107421875} 02/25/2022 15:22:51 - INFO - codeparrot_training - Step 29168: {'lr': 0.00019853434780752973, 'samples': 14934528, 'steps': 29168, 'loss/train': 1.2153934240341187} 02/25/2022 15:22:55 - INFO - codeparrot_training - Step 29169: {'lr': 0.0001985183359251684, 'samples': 14935040, 'steps': 29169, 'loss/train': 2.0026893615722656} 02/25/2022 15:23:00 - INFO - codeparrot_training - Step 29170: {'lr': 0.00019850232426333814, 'samples': 14935552, 'steps': 29170, 'loss/train': 2.3233540058135986} 02/25/2022 15:23:04 - INFO - codeparrot_training - Step 29171: {'lr': 0.00019848631282210754, 'samples': 14936064, 'steps': 29171, 'loss/train': 0.8999312520027161} 02/25/2022 15:23:09 - INFO - codeparrot_training - Step 29172: {'lr': 0.00019847030160154527, 'samples': 14936576, 'steps': 29172, 'loss/train': 1.1251603364944458} 02/25/2022 15:23:13 - INFO - codeparrot_training - Step 29173: {'lr': 0.00019845429060171997, 'samples': 14937088, 'steps': 29173, 'loss/train': 2.0283477306365967} 02/25/2022 15:23:18 - INFO - codeparrot_training - Step 29174: {'lr': 0.00019843827982269998, 'samples': 14937600, 'steps': 29174, 'loss/train': 0.05940178036689758} 02/25/2022 15:23:22 - INFO - codeparrot_training - Step 29175: {'lr': 0.00019842226926455403, 'samples': 14938112, 'steps': 29175, 'loss/train': 1.4810538291931152} 02/25/2022 15:23:28 - INFO - codeparrot_training - Step 29176: {'lr': 0.00019840625892735074, 'samples': 14938624, 'steps': 29176, 'loss/train': 0.6489487886428833} 02/25/2022 15:23:32 - INFO - codeparrot_training - Step 29177: {'lr': 0.00019839024881115872, 'samples': 14939136, 'steps': 29177, 'loss/train': 1.5926954746246338} 02/25/2022 15:23:35 - INFO - codeparrot_training - Step 29178: {'lr': 0.0001983742389160464, 'samples': 14939648, 'steps': 29178, 'loss/train': 2.5636343955993652} 02/25/2022 15:23:41 - INFO - codeparrot_training - Step 29179: {'lr': 0.00019835822924208244, 'samples': 14940160, 'steps': 29179, 'loss/train': 1.109387993812561} 02/25/2022 15:23:46 - INFO - codeparrot_training - Step 29180: {'lr': 0.00019834221978933542, 'samples': 14940672, 'steps': 29180, 'loss/train': 1.7216293811798096} 02/25/2022 15:23:50 - INFO - codeparrot_training - Step 29181: {'lr': 0.00019832621055787406, 'samples': 14941184, 'steps': 29181, 'loss/train': 0.7553069591522217} 02/25/2022 15:23:55 - INFO - codeparrot_training - Step 29182: {'lr': 0.0001983102015477667, 'samples': 14941696, 'steps': 29182, 'loss/train': 1.003391981124878} 02/25/2022 15:23:59 - INFO - codeparrot_training - Step 29183: {'lr': 0.000198294192759082, 'samples': 14942208, 'steps': 29183, 'loss/train': 1.844997525215149} 02/25/2022 15:24:04 - INFO - codeparrot_training - Step 29184: {'lr': 0.00019827818419188855, 'samples': 14942720, 'steps': 29184, 'loss/train': 1.0950952768325806} 02/25/2022 15:24:08 - INFO - codeparrot_training - Step 29185: {'lr': 0.00019826217584625495, 'samples': 14943232, 'steps': 29185, 'loss/train': 1.5986212491989136} 02/25/2022 15:24:13 - INFO - codeparrot_training - Step 29186: {'lr': 0.00019824616772224987, 'samples': 14943744, 'steps': 29186, 'loss/train': 1.4953640699386597} 02/25/2022 15:24:17 - INFO - codeparrot_training - Step 29187: {'lr': 0.00019823015981994165, 'samples': 14944256, 'steps': 29187, 'loss/train': 2.254401445388794} 02/25/2022 15:24:20 - INFO - codeparrot_training - Step 29188: {'lr': 0.00019821415213939903, 'samples': 14944768, 'steps': 29188, 'loss/train': 2.495269775390625} 02/25/2022 15:24:26 - INFO - codeparrot_training - Step 29189: {'lr': 0.0001981981446806905, 'samples': 14945280, 'steps': 29189, 'loss/train': 1.8421341180801392} 02/25/2022 15:24:30 - INFO - codeparrot_training - Step 29190: {'lr': 0.0001981821374438848, 'samples': 14945792, 'steps': 29190, 'loss/train': 2.6240923404693604} 02/25/2022 15:24:35 - INFO - codeparrot_training - Step 29191: {'lr': 0.00019816613042905024, 'samples': 14946304, 'steps': 29191, 'loss/train': 1.244981050491333} 02/25/2022 15:24:39 - INFO - codeparrot_training - Step 29192: {'lr': 0.0001981501236362555, 'samples': 14946816, 'steps': 29192, 'loss/train': 2.169217109680176} 02/25/2022 15:24:44 - INFO - codeparrot_training - Step 29193: {'lr': 0.00019813411706556927, 'samples': 14947328, 'steps': 29193, 'loss/train': 1.9341119527816772} 02/25/2022 15:24:48 - INFO - codeparrot_training - Step 29194: {'lr': 0.00019811811071705993, 'samples': 14947840, 'steps': 29194, 'loss/train': 1.773614525794983} 02/25/2022 15:24:53 - INFO - codeparrot_training - Step 29195: {'lr': 0.0001981021045907962, 'samples': 14948352, 'steps': 29195, 'loss/train': 1.3399029970169067} 02/25/2022 15:24:57 - INFO - codeparrot_training - Step 29196: {'lr': 0.00019808609868684652, 'samples': 14948864, 'steps': 29196, 'loss/train': 1.0027785301208496} 02/25/2022 15:25:02 - INFO - codeparrot_training - Step 29197: {'lr': 0.00019807009300527954, 'samples': 14949376, 'steps': 29197, 'loss/train': 0.34942182898521423} 02/25/2022 15:25:09 - INFO - codeparrot_training - Step 29198: {'lr': 0.00019805408754616376, 'samples': 14949888, 'steps': 29198, 'loss/train': 1.4731298685073853} 02/25/2022 15:25:12 - INFO - codeparrot_training - Step 29199: {'lr': 0.00019803808230956782, 'samples': 14950400, 'steps': 29199, 'loss/train': 0.944177508354187} 02/25/2022 15:25:18 - INFO - codeparrot_training - Step 29200: {'lr': 0.0001980220772955602, 'samples': 14950912, 'steps': 29200, 'loss/train': 0.8870585560798645} 02/25/2022 15:25:21 - INFO - codeparrot_training - Step 29201: {'lr': 0.00019800607250420948, 'samples': 14951424, 'steps': 29201, 'loss/train': 0.7162878513336182} 02/25/2022 15:25:27 - INFO - codeparrot_training - Step 29202: {'lr': 0.00019799006793558423, 'samples': 14951936, 'steps': 29202, 'loss/train': 1.5842257738113403} 02/25/2022 15:25:30 - INFO - codeparrot_training - Step 29203: {'lr': 0.00019797406358975308, 'samples': 14952448, 'steps': 29203, 'loss/train': 1.0783542394638062} 02/25/2022 15:25:36 - INFO - codeparrot_training - Step 29204: {'lr': 0.00019795805946678453, 'samples': 14952960, 'steps': 29204, 'loss/train': 1.986425518989563} 02/25/2022 15:25:39 - INFO - codeparrot_training - Step 29205: {'lr': 0.00019794205556674706, 'samples': 14953472, 'steps': 29205, 'loss/train': 1.8771569728851318} 02/25/2022 15:25:44 - INFO - codeparrot_training - Step 29206: {'lr': 0.0001979260518897093, 'samples': 14953984, 'steps': 29206, 'loss/train': 1.699537992477417} 02/25/2022 15:25:48 - INFO - codeparrot_training - Step 29207: {'lr': 0.00019791004843573975, 'samples': 14954496, 'steps': 29207, 'loss/train': 1.7367717027664185} 02/25/2022 15:25:54 - INFO - codeparrot_training - Step 29208: {'lr': 0.00019789404520490717, 'samples': 14955008, 'steps': 29208, 'loss/train': 2.343014717102051} 02/25/2022 15:25:57 - INFO - codeparrot_training - Step 29209: {'lr': 0.00019787804219727984, 'samples': 14955520, 'steps': 29209, 'loss/train': 1.4697836637496948} 02/25/2022 15:26:03 - INFO - codeparrot_training - Step 29210: {'lr': 0.0001978620394129264, 'samples': 14956032, 'steps': 29210, 'loss/train': 0.46326011419296265} 02/25/2022 15:26:06 - INFO - codeparrot_training - Step 29211: {'lr': 0.00019784603685191546, 'samples': 14956544, 'steps': 29211, 'loss/train': 2.3311257362365723} 02/25/2022 15:26:12 - INFO - codeparrot_training - Step 29212: {'lr': 0.00019783003451431564, 'samples': 14957056, 'steps': 29212, 'loss/train': 1.293544888496399} 02/25/2022 15:26:15 - INFO - codeparrot_training - Step 29213: {'lr': 0.00019781403240019528, 'samples': 14957568, 'steps': 29213, 'loss/train': 1.6733356714248657} 02/25/2022 15:26:21 - INFO - codeparrot_training - Step 29214: {'lr': 0.00019779803050962302, 'samples': 14958080, 'steps': 29214, 'loss/train': 1.7695726156234741} 02/25/2022 15:26:25 - INFO - codeparrot_training - Step 29215: {'lr': 0.00019778202884266743, 'samples': 14958592, 'steps': 29215, 'loss/train': 1.8135424852371216} 02/25/2022 15:26:30 - INFO - codeparrot_training - Step 29216: {'lr': 0.00019776602739939714, 'samples': 14959104, 'steps': 29216, 'loss/train': 0.8080177307128906} 02/25/2022 15:26:34 - INFO - codeparrot_training - Step 29217: {'lr': 0.00019775002617988054, 'samples': 14959616, 'steps': 29217, 'loss/train': 1.9845391511917114} 02/25/2022 15:26:39 - INFO - codeparrot_training - Step 29218: {'lr': 0.00019773402518418617, 'samples': 14960128, 'steps': 29218, 'loss/train': 1.1778076887130737} 02/25/2022 15:26:43 - INFO - codeparrot_training - Step 29219: {'lr': 0.00019771802441238267, 'samples': 14960640, 'steps': 29219, 'loss/train': 2.2341270446777344} 02/25/2022 15:26:48 - INFO - codeparrot_training - Step 29220: {'lr': 0.00019770202386453856, 'samples': 14961152, 'steps': 29220, 'loss/train': 2.2118546962738037} 02/25/2022 15:26:52 - INFO - codeparrot_training - Step 29221: {'lr': 0.0001976860235407225, 'samples': 14961664, 'steps': 29221, 'loss/train': 1.772839903831482} 02/25/2022 15:26:57 - INFO - codeparrot_training - Step 29222: {'lr': 0.00019767002344100277, 'samples': 14962176, 'steps': 29222, 'loss/train': 2.4914770126342773} 02/25/2022 15:27:01 - INFO - codeparrot_training - Step 29223: {'lr': 0.00019765402356544803, 'samples': 14962688, 'steps': 29223, 'loss/train': 0.49232980608940125} 02/25/2022 15:27:07 - INFO - codeparrot_training - Step 29224: {'lr': 0.00019763802391412687, 'samples': 14963200, 'steps': 29224, 'loss/train': 1.1145492792129517} 02/25/2022 15:27:10 - INFO - codeparrot_training - Step 29225: {'lr': 0.00019762202448710787, 'samples': 14963712, 'steps': 29225, 'loss/train': 0.11281704157590866} 02/25/2022 15:27:16 - INFO - codeparrot_training - Step 29226: {'lr': 0.00019760602528445936, 'samples': 14964224, 'steps': 29226, 'loss/train': 1.1031262874603271} 02/25/2022 15:27:19 - INFO - codeparrot_training - Step 29227: {'lr': 0.00019759002630625002, 'samples': 14964736, 'steps': 29227, 'loss/train': 3.045053482055664} 02/25/2022 15:27:25 - INFO - codeparrot_training - Step 29228: {'lr': 0.00019757402755254838, 'samples': 14965248, 'steps': 29228, 'loss/train': 2.555513620376587} 02/25/2022 15:27:30 - INFO - codeparrot_training - Step 29229: {'lr': 0.00019755802902342304, 'samples': 14965760, 'steps': 29229, 'loss/train': 1.16560697555542} 02/25/2022 15:27:34 - INFO - codeparrot_training - Step 29230: {'lr': 0.00019754203071894235, 'samples': 14966272, 'steps': 29230, 'loss/train': 1.9016810655593872} 02/25/2022 15:27:37 - INFO - codeparrot_training - Step 29231: {'lr': 0.00019752603263917495, 'samples': 14966784, 'steps': 29231, 'loss/train': 0.2916000485420227} 02/25/2022 15:27:43 - INFO - codeparrot_training - Step 29232: {'lr': 0.0001975100347841894, 'samples': 14967296, 'steps': 29232, 'loss/train': 1.4350587129592896} 02/25/2022 15:27:46 - INFO - codeparrot_training - Step 29233: {'lr': 0.0001974940371540542, 'samples': 14967808, 'steps': 29233, 'loss/train': 1.8296558856964111} 02/25/2022 15:27:53 - INFO - codeparrot_training - Step 29234: {'lr': 0.00019747803974883788, 'samples': 14968320, 'steps': 29234, 'loss/train': 1.3263262510299683} 02/25/2022 15:27:56 - INFO - codeparrot_training - Step 29235: {'lr': 0.0001974620425686089, 'samples': 14968832, 'steps': 29235, 'loss/train': 2.071936845779419} 02/25/2022 15:28:02 - INFO - codeparrot_training - Step 29236: {'lr': 0.00019744604561343587, 'samples': 14969344, 'steps': 29236, 'loss/train': 1.1329994201660156} 02/25/2022 15:28:05 - INFO - codeparrot_training - Step 29237: {'lr': 0.00019743004888338734, 'samples': 14969856, 'steps': 29237, 'loss/train': 2.0754735469818115} 02/25/2022 15:28:11 - INFO - codeparrot_training - Step 29238: {'lr': 0.00019741405237853177, 'samples': 14970368, 'steps': 29238, 'loss/train': 0.10996003448963165} 02/25/2022 15:28:14 - INFO - codeparrot_training - Step 29239: {'lr': 0.00019739805609893767, 'samples': 14970880, 'steps': 29239, 'loss/train': 2.6604561805725098} 02/25/2022 15:28:20 - INFO - codeparrot_training - Step 29240: {'lr': 0.00019738206004467362, 'samples': 14971392, 'steps': 29240, 'loss/train': 1.1484644412994385} 02/25/2022 15:28:23 - INFO - codeparrot_training - Step 29241: {'lr': 0.0001973660642158081, 'samples': 14971904, 'steps': 29241, 'loss/train': 2.341256856918335} 02/25/2022 15:28:29 - INFO - codeparrot_training - Step 29242: {'lr': 0.00019735006861240962, 'samples': 14972416, 'steps': 29242, 'loss/train': 1.289214015007019} 02/25/2022 15:28:35 - INFO - codeparrot_training - Step 29243: {'lr': 0.00019733407323454678, 'samples': 14972928, 'steps': 29243, 'loss/train': 1.597387433052063} 02/25/2022 15:28:38 - INFO - codeparrot_training - Step 29244: {'lr': 0.00019731807808228807, 'samples': 14973440, 'steps': 29244, 'loss/train': 1.108201503753662} 02/25/2022 15:28:44 - INFO - codeparrot_training - Step 29245: {'lr': 0.00019730208315570192, 'samples': 14973952, 'steps': 29245, 'loss/train': 1.306506872177124} 02/25/2022 15:28:48 - INFO - codeparrot_training - Step 29246: {'lr': 0.0001972860884548569, 'samples': 14974464, 'steps': 29246, 'loss/train': 1.9439525604248047} 02/25/2022 15:28:51 - INFO - codeparrot_training - Step 29247: {'lr': 0.00019727009397982166, 'samples': 14974976, 'steps': 29247, 'loss/train': 1.989899754524231} 02/25/2022 15:28:57 - INFO - codeparrot_training - Step 29248: {'lr': 0.0001972540997306645, 'samples': 14975488, 'steps': 29248, 'loss/train': 1.3352165222167969} 02/25/2022 15:29:00 - INFO - codeparrot_training - Step 29249: {'lr': 0.000197238105707454, 'samples': 14976000, 'steps': 29249, 'loss/train': 1.0574227571487427} 02/25/2022 15:29:06 - INFO - codeparrot_training - Step 29250: {'lr': 0.00019722211191025872, 'samples': 14976512, 'steps': 29250, 'loss/train': 1.839902639389038} 02/25/2022 15:29:09 - INFO - codeparrot_training - Step 29251: {'lr': 0.00019720611833914725, 'samples': 14977024, 'steps': 29251, 'loss/train': 0.0934521034359932} 02/25/2022 15:29:15 - INFO - codeparrot_training - Step 29252: {'lr': 0.0001971901249941879, 'samples': 14977536, 'steps': 29252, 'loss/train': 1.4760456085205078} 02/25/2022 15:29:18 - INFO - codeparrot_training - Step 29253: {'lr': 0.00019717413187544927, 'samples': 14978048, 'steps': 29253, 'loss/train': 2.273589611053467} 02/25/2022 15:29:24 - INFO - codeparrot_training - Step 29254: {'lr': 0.00019715813898299993, 'samples': 14978560, 'steps': 29254, 'loss/train': 1.7405003309249878} 02/25/2022 15:29:27 - INFO - codeparrot_training - Step 29255: {'lr': 0.0001971421463169083, 'samples': 14979072, 'steps': 29255, 'loss/train': 2.853459596633911} 02/25/2022 15:29:32 - INFO - codeparrot_training - Step 29256: {'lr': 0.00019712615387724306, 'samples': 14979584, 'steps': 29256, 'loss/train': 1.4149203300476074} 02/25/2022 15:29:36 - INFO - codeparrot_training - Step 29257: {'lr': 0.00019711016166407247, 'samples': 14980096, 'steps': 29257, 'loss/train': 1.1924718618392944} 02/25/2022 15:29:41 - INFO - codeparrot_training - Step 29258: {'lr': 0.00019709416967746515, 'samples': 14980608, 'steps': 29258, 'loss/train': 0.6543747186660767} 02/25/2022 15:29:48 - INFO - codeparrot_training - Step 29259: {'lr': 0.0001970781779174896, 'samples': 14981120, 'steps': 29259, 'loss/train': 1.014675498008728} 02/25/2022 15:29:52 - INFO - codeparrot_training - Step 29260: {'lr': 0.00019706218638421448, 'samples': 14981632, 'steps': 29260, 'loss/train': 2.5467488765716553} 02/25/2022 15:29:57 - INFO - codeparrot_training - Step 29261: {'lr': 0.00019704619507770799, 'samples': 14982144, 'steps': 29261, 'loss/train': 2.1105871200561523} 02/25/2022 15:30:01 - INFO - codeparrot_training - Step 29262: {'lr': 0.00019703020399803881, 'samples': 14982656, 'steps': 29262, 'loss/train': 3.2613344192504883} 02/25/2022 15:30:06 - INFO - codeparrot_training - Step 29263: {'lr': 0.0001970142131452754, 'samples': 14983168, 'steps': 29263, 'loss/train': 1.2367548942565918} 02/25/2022 15:30:10 - INFO - codeparrot_training - Step 29264: {'lr': 0.0001969982225194864, 'samples': 14983680, 'steps': 29264, 'loss/train': 1.9033775329589844} 02/25/2022 15:30:15 - INFO - codeparrot_training - Step 29265: {'lr': 0.00019698223212074006, 'samples': 14984192, 'steps': 29265, 'loss/train': 1.407662034034729} 02/25/2022 15:30:19 - INFO - codeparrot_training - Step 29266: {'lr': 0.00019696624194910504, 'samples': 14984704, 'steps': 29266, 'loss/train': 1.8565514087677002} 02/25/2022 15:30:24 - INFO - codeparrot_training - Step 29267: {'lr': 0.00019695025200464978, 'samples': 14985216, 'steps': 29267, 'loss/train': 1.8945730924606323} 02/25/2022 15:30:28 - INFO - codeparrot_training - Step 29268: {'lr': 0.00019693426228744278, 'samples': 14985728, 'steps': 29268, 'loss/train': 1.8950185775756836} 02/25/2022 15:30:34 - INFO - codeparrot_training - Step 29269: {'lr': 0.00019691827279755267, 'samples': 14986240, 'steps': 29269, 'loss/train': 0.8250755667686462} 02/25/2022 15:30:38 - INFO - codeparrot_training - Step 29270: {'lr': 0.0001969022835350477, 'samples': 14986752, 'steps': 29270, 'loss/train': 3.0484871864318848} 02/25/2022 15:30:43 - INFO - codeparrot_training - Step 29271: {'lr': 0.0001968862944999965, 'samples': 14987264, 'steps': 29271, 'loss/train': 0.8977957963943481} 02/25/2022 15:30:47 - INFO - codeparrot_training - Step 29272: {'lr': 0.00019687030569246755, 'samples': 14987776, 'steps': 29272, 'loss/train': 2.0800747871398926} 02/25/2022 15:30:52 - INFO - codeparrot_training - Step 29273: {'lr': 0.00019685431711252943, 'samples': 14988288, 'steps': 29273, 'loss/train': 2.110466957092285} 02/25/2022 15:30:56 - INFO - codeparrot_training - Step 29274: {'lr': 0.00019683832876025048, 'samples': 14988800, 'steps': 29274, 'loss/train': 1.6376875638961792} 02/25/2022 15:31:01 - INFO - codeparrot_training - Step 29275: {'lr': 0.0001968223406356992, 'samples': 14989312, 'steps': 29275, 'loss/train': 1.8917607069015503} 02/25/2022 15:31:05 - INFO - codeparrot_training - Step 29276: {'lr': 0.00019680635273894415, 'samples': 14989824, 'steps': 29276, 'loss/train': 1.366753101348877} 02/25/2022 15:31:10 - INFO - codeparrot_training - Step 29277: {'lr': 0.0001967903650700539, 'samples': 14990336, 'steps': 29277, 'loss/train': 1.572314739227295} 02/25/2022 15:31:13 - INFO - codeparrot_training - Step 29278: {'lr': 0.00019677437762909673, 'samples': 14990848, 'steps': 29278, 'loss/train': 1.0510770082473755} 02/25/2022 15:31:20 - INFO - codeparrot_training - Step 29279: {'lr': 0.00019675839041614118, 'samples': 14991360, 'steps': 29279, 'loss/train': 2.573420763015747} 02/25/2022 15:31:24 - INFO - codeparrot_training - Step 29280: {'lr': 0.00019674240343125587, 'samples': 14991872, 'steps': 29280, 'loss/train': 1.5276926755905151} 02/25/2022 15:31:29 - INFO - codeparrot_training - Step 29281: {'lr': 0.00019672641667450915, 'samples': 14992384, 'steps': 29281, 'loss/train': 0.049334824085235596} 02/25/2022 15:31:33 - INFO - codeparrot_training - Step 29282: {'lr': 0.0001967104301459696, 'samples': 14992896, 'steps': 29282, 'loss/train': 1.5207090377807617} 02/25/2022 15:31:38 - INFO - codeparrot_training - Step 29283: {'lr': 0.00019669444384570557, 'samples': 14993408, 'steps': 29283, 'loss/train': 1.9674229621887207} 02/25/2022 15:31:42 - INFO - codeparrot_training - Step 29284: {'lr': 0.0001966784577737857, 'samples': 14993920, 'steps': 29284, 'loss/train': 1.933316946029663} 02/25/2022 15:31:47 - INFO - codeparrot_training - Step 29285: {'lr': 0.00019666247193027829, 'samples': 14994432, 'steps': 29285, 'loss/train': 3.544548749923706} 02/25/2022 15:31:51 - INFO - codeparrot_training - Step 29286: {'lr': 0.00019664648631525199, 'samples': 14994944, 'steps': 29286, 'loss/train': 2.274874687194824} 02/25/2022 15:31:56 - INFO - codeparrot_training - Step 29287: {'lr': 0.00019663050092877515, 'samples': 14995456, 'steps': 29287, 'loss/train': 1.3459421396255493} 02/25/2022 15:32:00 - INFO - codeparrot_training - Step 29288: {'lr': 0.00019661451577091633, 'samples': 14995968, 'steps': 29288, 'loss/train': 1.2078479528427124} 02/25/2022 15:32:06 - INFO - codeparrot_training - Step 29289: {'lr': 0.0001965985308417439, 'samples': 14996480, 'steps': 29289, 'loss/train': 1.757480502128601} 02/25/2022 15:32:09 - INFO - codeparrot_training - Step 29290: {'lr': 0.0001965825461413265, 'samples': 14996992, 'steps': 29290, 'loss/train': 0.9850122332572937} 02/25/2022 15:32:15 - INFO - codeparrot_training - Step 29291: {'lr': 0.00019656656166973252, 'samples': 14997504, 'steps': 29291, 'loss/train': 6.274078369140625} 02/25/2022 15:32:18 - INFO - codeparrot_training - Step 29292: {'lr': 0.00019655057742703032, 'samples': 14998016, 'steps': 29292, 'loss/train': 2.1278178691864014} 02/25/2022 15:32:24 - INFO - codeparrot_training - Step 29293: {'lr': 0.00019653459341328853, 'samples': 14998528, 'steps': 29293, 'loss/train': 1.459757685661316} 02/25/2022 15:32:27 - INFO - codeparrot_training - Step 29294: {'lr': 0.00019651860962857553, 'samples': 14999040, 'steps': 29294, 'loss/train': 2.2623355388641357} 02/25/2022 15:32:33 - INFO - codeparrot_training - Step 29295: {'lr': 0.00019650262607295995, 'samples': 14999552, 'steps': 29295, 'loss/train': 2.201338529586792} 02/25/2022 15:32:37 - INFO - codeparrot_training - Step 29296: {'lr': 0.00019648664274651002, 'samples': 15000064, 'steps': 29296, 'loss/train': 2.0335028171539307} 02/25/2022 15:32:42 - INFO - codeparrot_training - Step 29297: {'lr': 0.00019647065964929433, 'samples': 15000576, 'steps': 29297, 'loss/train': 2.2171027660369873} 02/25/2022 15:32:46 - INFO - codeparrot_training - Step 29298: {'lr': 0.00019645467678138133, 'samples': 15001088, 'steps': 29298, 'loss/train': 1.4360616207122803} 02/25/2022 15:32:51 - INFO - codeparrot_training - Step 29299: {'lr': 0.00019643869414283957, 'samples': 15001600, 'steps': 29299, 'loss/train': 1.6715625524520874} 02/25/2022 15:32:55 - INFO - codeparrot_training - Step 29300: {'lr': 0.00019642271173373735, 'samples': 15002112, 'steps': 29300, 'loss/train': 1.9193049669265747} 02/25/2022 15:33:00 - INFO - codeparrot_training - Step 29301: {'lr': 0.00019640672955414324, 'samples': 15002624, 'steps': 29301, 'loss/train': 1.0728017091751099} 02/25/2022 15:33:04 - INFO - codeparrot_training - Step 29302: {'lr': 0.00019639074760412567, 'samples': 15003136, 'steps': 29302, 'loss/train': 0.8722072243690491} 02/25/2022 15:33:09 - INFO - codeparrot_training - Step 29303: {'lr': 0.00019637476588375312, 'samples': 15003648, 'steps': 29303, 'loss/train': 2.069180488586426} 02/25/2022 15:33:13 - INFO - codeparrot_training - Step 29304: {'lr': 0.00019635878439309412, 'samples': 15004160, 'steps': 29304, 'loss/train': 0.8410064578056335} 02/25/2022 15:33:19 - INFO - codeparrot_training - Step 29305: {'lr': 0.00019634280313221697, 'samples': 15004672, 'steps': 29305, 'loss/train': 1.8132877349853516} 02/25/2022 15:33:22 - INFO - codeparrot_training - Step 29306: {'lr': 0.00019632682210119021, 'samples': 15005184, 'steps': 29306, 'loss/train': 0.051910340785980225} 02/25/2022 15:33:28 - INFO - codeparrot_training - Step 29307: {'lr': 0.0001963108413000823, 'samples': 15005696, 'steps': 29307, 'loss/train': 1.4317172765731812} 02/25/2022 15:33:31 - INFO - codeparrot_training - Step 29308: {'lr': 0.00019629486072896183, 'samples': 15006208, 'steps': 29308, 'loss/train': 2.275067090988159} 02/25/2022 15:33:37 - INFO - codeparrot_training - Step 29309: {'lr': 0.00019627888038789698, 'samples': 15006720, 'steps': 29309, 'loss/train': 1.3946512937545776} 02/25/2022 15:33:40 - INFO - codeparrot_training - Step 29310: {'lr': 0.00019626290027695636, 'samples': 15007232, 'steps': 29310, 'loss/train': 0.10922368615865707} 02/25/2022 15:33:46 - INFO - codeparrot_training - Step 29311: {'lr': 0.00019624692039620843, 'samples': 15007744, 'steps': 29311, 'loss/train': 0.8043493628501892} 02/25/2022 15:33:50 - INFO - codeparrot_training - Step 29312: {'lr': 0.00019623094074572173, 'samples': 15008256, 'steps': 29312, 'loss/train': 1.902226448059082} 02/25/2022 15:33:55 - INFO - codeparrot_training - Step 29313: {'lr': 0.00019621496132556446, 'samples': 15008768, 'steps': 29313, 'loss/train': 1.8014180660247803} 02/25/2022 15:33:58 - INFO - codeparrot_training - Step 29314: {'lr': 0.00019619898213580527, 'samples': 15009280, 'steps': 29314, 'loss/train': 2.7981128692626953} 02/25/2022 15:34:04 - INFO - codeparrot_training - Step 29315: {'lr': 0.0001961830031765125, 'samples': 15009792, 'steps': 29315, 'loss/train': 1.6252634525299072} 02/25/2022 15:34:07 - INFO - codeparrot_training - Step 29316: {'lr': 0.0001961670244477547, 'samples': 15010304, 'steps': 29316, 'loss/train': 1.3770695924758911} 02/25/2022 15:34:14 - INFO - codeparrot_training - Step 29317: {'lr': 0.00019615104594960037, 'samples': 15010816, 'steps': 29317, 'loss/train': 1.3650835752487183} 02/25/2022 15:34:17 - INFO - codeparrot_training - Step 29318: {'lr': 0.00019613506768211776, 'samples': 15011328, 'steps': 29318, 'loss/train': 0.790747344493866} 02/25/2022 15:34:23 - INFO - codeparrot_training - Step 29319: {'lr': 0.0001961190896453754, 'samples': 15011840, 'steps': 29319, 'loss/train': 0.8436750769615173} 02/25/2022 15:34:26 - INFO - codeparrot_training - Step 29320: {'lr': 0.0001961031118394418, 'samples': 15012352, 'steps': 29320, 'loss/train': 1.0435761213302612} 02/25/2022 15:34:32 - INFO - codeparrot_training - Step 29321: {'lr': 0.00019608713426438534, 'samples': 15012864, 'steps': 29321, 'loss/train': 1.2243672609329224} 02/25/2022 15:34:35 - INFO - codeparrot_training - Step 29322: {'lr': 0.00019607115692027446, 'samples': 15013376, 'steps': 29322, 'loss/train': 1.5739738941192627} 02/25/2022 15:34:41 - INFO - codeparrot_training - Step 29323: {'lr': 0.00019605517980717763, 'samples': 15013888, 'steps': 29323, 'loss/train': 1.5121190547943115} 02/25/2022 15:34:44 - INFO - codeparrot_training - Step 29324: {'lr': 0.0001960392029251633, 'samples': 15014400, 'steps': 29324, 'loss/train': 1.2646921873092651} 02/25/2022 15:34:50 - INFO - codeparrot_training - Step 29325: {'lr': 0.00019602322627429992, 'samples': 15014912, 'steps': 29325, 'loss/train': 1.9140151739120483} 02/25/2022 15:34:56 - INFO - codeparrot_training - Step 29326: {'lr': 0.0001960072498546558, 'samples': 15015424, 'steps': 29326, 'loss/train': 2.251607894897461} 02/25/2022 15:34:59 - INFO - codeparrot_training - Step 29327: {'lr': 0.00019599127366629957, 'samples': 15015936, 'steps': 29327, 'loss/train': 1.3281522989273071} 02/25/2022 15:35:05 - INFO - codeparrot_training - Step 29328: {'lr': 0.0001959752977092995, 'samples': 15016448, 'steps': 29328, 'loss/train': 1.7271817922592163} 02/25/2022 15:35:08 - INFO - codeparrot_training - Step 29329: {'lr': 0.00019595932198372413, 'samples': 15016960, 'steps': 29329, 'loss/train': 1.743116021156311} 02/25/2022 15:35:14 - INFO - codeparrot_training - Step 29330: {'lr': 0.00019594334648964187, 'samples': 15017472, 'steps': 29330, 'loss/train': 1.4380152225494385} 02/25/2022 15:35:18 - INFO - codeparrot_training - Step 29331: {'lr': 0.00019592737122712118, 'samples': 15017984, 'steps': 29331, 'loss/train': 2.723742723464966} 02/25/2022 15:35:21 - INFO - codeparrot_training - Step 29332: {'lr': 0.00019591139619623042, 'samples': 15018496, 'steps': 29332, 'loss/train': 1.4156464338302612} 02/25/2022 15:35:27 - INFO - codeparrot_training - Step 29333: {'lr': 0.00019589542139703803, 'samples': 15019008, 'steps': 29333, 'loss/train': 0.8630192875862122} 02/25/2022 15:35:30 - INFO - codeparrot_training - Step 29334: {'lr': 0.00019587944682961263, 'samples': 15019520, 'steps': 29334, 'loss/train': 1.9504932165145874} 02/25/2022 15:35:36 - INFO - codeparrot_training - Step 29335: {'lr': 0.00019586347249402235, 'samples': 15020032, 'steps': 29335, 'loss/train': 1.0513288974761963} 02/25/2022 15:35:39 - INFO - codeparrot_training - Step 29336: {'lr': 0.00019584749839033575, 'samples': 15020544, 'steps': 29336, 'loss/train': 2.233471155166626} 02/25/2022 15:35:45 - INFO - codeparrot_training - Step 29337: {'lr': 0.0001958315245186213, 'samples': 15021056, 'steps': 29337, 'loss/train': 1.8266321420669556} 02/25/2022 15:35:49 - INFO - codeparrot_training - Step 29338: {'lr': 0.00019581555087894747, 'samples': 15021568, 'steps': 29338, 'loss/train': 1.6159883737564087} 02/25/2022 15:35:54 - INFO - codeparrot_training - Step 29339: {'lr': 0.00019579957747138256, 'samples': 15022080, 'steps': 29339, 'loss/train': 1.0751569271087646} 02/25/2022 15:35:58 - INFO - codeparrot_training - Step 29340: {'lr': 0.00019578360429599502, 'samples': 15022592, 'steps': 29340, 'loss/train': 1.4261159896850586} 02/25/2022 15:36:03 - INFO - codeparrot_training - Step 29341: {'lr': 0.0001957676313528533, 'samples': 15023104, 'steps': 29341, 'loss/train': 1.552079677581787} 02/25/2022 15:36:09 - INFO - codeparrot_training - Step 29342: {'lr': 0.00019575165864202582, 'samples': 15023616, 'steps': 29342, 'loss/train': 1.9177117347717285} 02/25/2022 15:36:12 - INFO - codeparrot_training - Step 29343: {'lr': 0.00019573568616358114, 'samples': 15024128, 'steps': 29343, 'loss/train': 1.2271636724472046} 02/25/2022 15:36:18 - INFO - codeparrot_training - Step 29344: {'lr': 0.00019571971391758744, 'samples': 15024640, 'steps': 29344, 'loss/train': 2.9983057975769043} 02/25/2022 15:36:21 - INFO - codeparrot_training - Step 29345: {'lr': 0.0001957037419041132, 'samples': 15025152, 'steps': 29345, 'loss/train': 0.7506029009819031} 02/25/2022 15:36:27 - INFO - codeparrot_training - Step 29346: {'lr': 0.00019568777012322696, 'samples': 15025664, 'steps': 29346, 'loss/train': 1.364060878753662} 02/25/2022 15:36:30 - INFO - codeparrot_training - Step 29347: {'lr': 0.00019567179857499714, 'samples': 15026176, 'steps': 29347, 'loss/train': 2.2714734077453613} 02/25/2022 15:36:36 - INFO - codeparrot_training - Step 29348: {'lr': 0.00019565582725949198, 'samples': 15026688, 'steps': 29348, 'loss/train': 1.7118453979492188} 02/25/2022 15:36:39 - INFO - codeparrot_training - Step 29349: {'lr': 0.00019563985617678, 'samples': 15027200, 'steps': 29349, 'loss/train': 1.5523933172225952} 02/25/2022 15:36:45 - INFO - codeparrot_training - Step 29350: {'lr': 0.00019562388532692963, 'samples': 15027712, 'steps': 29350, 'loss/train': 1.0967680215835571} 02/25/2022 15:36:48 - INFO - codeparrot_training - Step 29351: {'lr': 0.00019560791471000925, 'samples': 15028224, 'steps': 29351, 'loss/train': 1.3886933326721191} 02/25/2022 15:36:54 - INFO - codeparrot_training - Step 29352: {'lr': 0.0001955919443260874, 'samples': 15028736, 'steps': 29352, 'loss/train': 0.7895808219909668} 02/25/2022 15:36:58 - INFO - codeparrot_training - Step 29353: {'lr': 0.0001955759741752323, 'samples': 15029248, 'steps': 29353, 'loss/train': 1.2784093618392944} 02/25/2022 15:37:03 - INFO - codeparrot_training - Step 29354: {'lr': 0.00019556000425751242, 'samples': 15029760, 'steps': 29354, 'loss/train': 1.3381541967391968} 02/25/2022 15:37:07 - INFO - codeparrot_training - Step 29355: {'lr': 0.00019554403457299618, 'samples': 15030272, 'steps': 29355, 'loss/train': 2.130079507827759} 02/25/2022 15:37:12 - INFO - codeparrot_training - Step 29356: {'lr': 0.00019552806512175214, 'samples': 15030784, 'steps': 29356, 'loss/train': 1.572493553161621} 02/25/2022 15:37:16 - INFO - codeparrot_training - Step 29357: {'lr': 0.00019551209590384848, 'samples': 15031296, 'steps': 29357, 'loss/train': 1.359002947807312} 02/25/2022 15:37:21 - INFO - codeparrot_training - Step 29358: {'lr': 0.00019549612691935363, 'samples': 15031808, 'steps': 29358, 'loss/train': 2.0528948307037354} 02/25/2022 15:37:25 - INFO - codeparrot_training - Step 29359: {'lr': 0.00019548015816833614, 'samples': 15032320, 'steps': 29359, 'loss/train': 1.699252963066101} 02/25/2022 15:37:30 - INFO - codeparrot_training - Step 29360: {'lr': 0.00019546418965086444, 'samples': 15032832, 'steps': 29360, 'loss/train': 1.3062719106674194} 02/25/2022 15:37:34 - INFO - codeparrot_training - Step 29361: {'lr': 0.00019544822136700668, 'samples': 15033344, 'steps': 29361, 'loss/train': 1.5943058729171753} 02/25/2022 15:37:40 - INFO - codeparrot_training - Step 29362: {'lr': 0.00019543225331683145, 'samples': 15033856, 'steps': 29362, 'loss/train': 3.369662046432495} 02/25/2022 15:37:43 - INFO - codeparrot_training - Step 29363: {'lr': 0.0001954162855004071, 'samples': 15034368, 'steps': 29363, 'loss/train': 1.9614683389663696} 02/25/2022 15:37:49 - INFO - codeparrot_training - Step 29364: {'lr': 0.00019540031791780214, 'samples': 15034880, 'steps': 29364, 'loss/train': 2.183573007583618} 02/25/2022 15:37:52 - INFO - codeparrot_training - Step 29365: {'lr': 0.00019538435056908485, 'samples': 15035392, 'steps': 29365, 'loss/train': 1.6338070631027222} 02/25/2022 15:37:58 - INFO - codeparrot_training - Step 29366: {'lr': 0.00019536838345432362, 'samples': 15035904, 'steps': 29366, 'loss/train': 0.5155447721481323} 02/25/2022 15:38:01 - INFO - codeparrot_training - Step 29367: {'lr': 0.00019535241657358694, 'samples': 15036416, 'steps': 29367, 'loss/train': 1.9426361322402954} 02/25/2022 15:38:07 - INFO - codeparrot_training - Step 29368: {'lr': 0.0001953364499269431, 'samples': 15036928, 'steps': 29368, 'loss/train': 0.9669619798660278} 02/25/2022 15:38:10 - INFO - codeparrot_training - Step 29369: {'lr': 0.0001953204835144606, 'samples': 15037440, 'steps': 29369, 'loss/train': 2.10241436958313} 02/25/2022 15:38:16 - INFO - codeparrot_training - Step 29370: {'lr': 0.0001953045173362077, 'samples': 15037952, 'steps': 29370, 'loss/train': 1.0091904401779175} 02/25/2022 15:38:19 - INFO - codeparrot_training - Step 29371: {'lr': 0.000195288551392253, 'samples': 15038464, 'steps': 29371, 'loss/train': 2.2979555130004883} 02/25/2022 15:38:25 - INFO - codeparrot_training - Step 29372: {'lr': 0.0001952725856826647, 'samples': 15038976, 'steps': 29372, 'loss/train': 2.239908218383789} 02/25/2022 15:38:29 - INFO - codeparrot_training - Step 29373: {'lr': 0.00019525662020751129, 'samples': 15039488, 'steps': 29373, 'loss/train': 1.7329139709472656} 02/25/2022 15:38:34 - INFO - codeparrot_training - Step 29374: {'lr': 0.00019524065496686114, 'samples': 15040000, 'steps': 29374, 'loss/train': 1.4030593633651733} 02/25/2022 15:38:38 - INFO - codeparrot_training - Step 29375: {'lr': 0.00019522468996078257, 'samples': 15040512, 'steps': 29375, 'loss/train': 2.3346424102783203} 02/25/2022 15:38:43 - INFO - codeparrot_training - Step 29376: {'lr': 0.00019520872518934406, 'samples': 15041024, 'steps': 29376, 'loss/train': 2.5046608448028564} 02/25/2022 15:38:47 - INFO - codeparrot_training - Step 29377: {'lr': 0.00019519276065261399, 'samples': 15041536, 'steps': 29377, 'loss/train': 1.0978639125823975} 02/25/2022 15:38:52 - INFO - codeparrot_training - Step 29378: {'lr': 0.00019517679635066082, 'samples': 15042048, 'steps': 29378, 'loss/train': 1.665866494178772} 02/25/2022 15:38:56 - INFO - codeparrot_training - Step 29379: {'lr': 0.00019516083228355275, 'samples': 15042560, 'steps': 29379, 'loss/train': 1.7864251136779785} 02/25/2022 15:39:01 - INFO - codeparrot_training - Step 29380: {'lr': 0.00019514486845135823, 'samples': 15043072, 'steps': 29380, 'loss/train': 2.371065616607666} 02/25/2022 15:39:04 - INFO - codeparrot_training - Step 29381: {'lr': 0.0001951289048541457, 'samples': 15043584, 'steps': 29381, 'loss/train': 2.5646724700927734} 02/25/2022 15:39:11 - INFO - codeparrot_training - Step 29382: {'lr': 0.0001951129414919836, 'samples': 15044096, 'steps': 29382, 'loss/train': 1.8295261859893799} 02/25/2022 15:39:14 - INFO - codeparrot_training - Step 29383: {'lr': 0.00019509697836494016, 'samples': 15044608, 'steps': 29383, 'loss/train': 2.626011848449707} 02/25/2022 15:39:20 - INFO - codeparrot_training - Step 29384: {'lr': 0.00019508101547308383, 'samples': 15045120, 'steps': 29384, 'loss/train': 1.493629813194275} 02/25/2022 15:39:25 - INFO - codeparrot_training - Step 29385: {'lr': 0.00019506505281648295, 'samples': 15045632, 'steps': 29385, 'loss/train': 1.805092453956604} 02/25/2022 15:39:29 - INFO - codeparrot_training - Step 29386: {'lr': 0.00019504909039520609, 'samples': 15046144, 'steps': 29386, 'loss/train': 1.8686336278915405} 02/25/2022 15:39:34 - INFO - codeparrot_training - Step 29387: {'lr': 0.00019503312820932133, 'samples': 15046656, 'steps': 29387, 'loss/train': 3.2920823097229004} 02/25/2022 15:39:38 - INFO - codeparrot_training - Step 29388: {'lr': 0.00019501716625889724, 'samples': 15047168, 'steps': 29388, 'loss/train': 1.1561026573181152} 02/25/2022 15:39:43 - INFO - codeparrot_training - Step 29389: {'lr': 0.0001950012045440021, 'samples': 15047680, 'steps': 29389, 'loss/train': 2.1098732948303223} 02/25/2022 15:39:47 - INFO - codeparrot_training - Step 29390: {'lr': 0.00019498524306470438, 'samples': 15048192, 'steps': 29390, 'loss/train': 1.7553433179855347} 02/25/2022 15:39:53 - INFO - codeparrot_training - Step 29391: {'lr': 0.00019496928182107252, 'samples': 15048704, 'steps': 29391, 'loss/train': 0.07703979313373566} 02/25/2022 15:39:57 - INFO - codeparrot_training - Step 29392: {'lr': 0.00019495332081317463, 'samples': 15049216, 'steps': 29392, 'loss/train': 2.7321858406066895} 02/25/2022 15:40:02 - INFO - codeparrot_training - Step 29393: {'lr': 0.00019493736004107926, 'samples': 15049728, 'steps': 29393, 'loss/train': 1.851007342338562} 02/25/2022 15:40:06 - INFO - codeparrot_training - Step 29394: {'lr': 0.00019492139950485476, 'samples': 15050240, 'steps': 29394, 'loss/train': 1.576150894165039} 02/25/2022 15:40:11 - INFO - codeparrot_training - Step 29395: {'lr': 0.0001949054392045696, 'samples': 15050752, 'steps': 29395, 'loss/train': 1.3481794595718384} 02/25/2022 15:40:15 - INFO - codeparrot_training - Step 29396: {'lr': 0.00019488947914029193, 'samples': 15051264, 'steps': 29396, 'loss/train': 2.079875946044922} 02/25/2022 15:40:20 - INFO - codeparrot_training - Step 29397: {'lr': 0.00019487351931209024, 'samples': 15051776, 'steps': 29397, 'loss/train': 2.4200375080108643} 02/25/2022 15:40:24 - INFO - codeparrot_training - Step 29398: {'lr': 0.00019485755972003288, 'samples': 15052288, 'steps': 29398, 'loss/train': 1.5261116027832031} 02/25/2022 15:40:29 - INFO - codeparrot_training - Step 29399: {'lr': 0.00019484160036418836, 'samples': 15052800, 'steps': 29399, 'loss/train': 2.0629138946533203} 02/25/2022 15:40:33 - INFO - codeparrot_training - Step 29400: {'lr': 0.00019482564124462476, 'samples': 15053312, 'steps': 29400, 'loss/train': 2.50455641746521} 02/25/2022 15:40:38 - INFO - codeparrot_training - Step 29401: {'lr': 0.00019480968236141062, 'samples': 15053824, 'steps': 29401, 'loss/train': 1.1354399919509888} 02/25/2022 15:40:41 - INFO - codeparrot_training - Step 29402: {'lr': 0.00019479372371461427, 'samples': 15054336, 'steps': 29402, 'loss/train': 2.0294108390808105} 02/25/2022 15:40:47 - INFO - codeparrot_training - Step 29403: {'lr': 0.00019477776530430407, 'samples': 15054848, 'steps': 29403, 'loss/train': 0.8522345423698425} 02/25/2022 15:40:50 - INFO - codeparrot_training - Step 29404: {'lr': 0.00019476180713054851, 'samples': 15055360, 'steps': 29404, 'loss/train': 3.1087419986724854} 02/25/2022 15:40:56 - INFO - codeparrot_training - Step 29405: {'lr': 0.0001947458491934157, 'samples': 15055872, 'steps': 29405, 'loss/train': 2.4730608463287354} 02/25/2022 15:40:59 - INFO - codeparrot_training - Step 29406: {'lr': 0.00019472989149297415, 'samples': 15056384, 'steps': 29406, 'loss/train': 1.6713955402374268} 02/25/2022 15:41:05 - INFO - codeparrot_training - Step 29407: {'lr': 0.00019471393402929223, 'samples': 15056896, 'steps': 29407, 'loss/train': 1.9496151208877563} 02/25/2022 15:41:09 - INFO - codeparrot_training - Step 29408: {'lr': 0.00019469797680243827, 'samples': 15057408, 'steps': 29408, 'loss/train': 1.1949479579925537} 02/25/2022 15:41:14 - INFO - codeparrot_training - Step 29409: {'lr': 0.00019468201981248057, 'samples': 15057920, 'steps': 29409, 'loss/train': 1.6739237308502197} 02/25/2022 15:41:18 - INFO - codeparrot_training - Step 29410: {'lr': 0.0001946660630594875, 'samples': 15058432, 'steps': 29410, 'loss/train': 1.9693725109100342} 02/25/2022 15:41:24 - INFO - codeparrot_training - Step 29411: {'lr': 0.00019465010654352751, 'samples': 15058944, 'steps': 29411, 'loss/train': 2.389209508895874} 02/25/2022 15:41:27 - INFO - codeparrot_training - Step 29412: {'lr': 0.00019463415026466885, 'samples': 15059456, 'steps': 29412, 'loss/train': 0.26547902822494507} 02/25/2022 15:41:33 - INFO - codeparrot_training - Step 29413: {'lr': 0.00019461819422297998, 'samples': 15059968, 'steps': 29413, 'loss/train': 2.3412394523620605} 02/25/2022 15:41:36 - INFO - codeparrot_training - Step 29414: {'lr': 0.00019460223841852908, 'samples': 15060480, 'steps': 29414, 'loss/train': 0.8751606345176697} 02/25/2022 15:41:42 - INFO - codeparrot_training - Step 29415: {'lr': 0.00019458628285138469, 'samples': 15060992, 'steps': 29415, 'loss/train': 1.8667467832565308} 02/25/2022 15:41:45 - INFO - codeparrot_training - Step 29416: {'lr': 0.00019457032752161503, 'samples': 15061504, 'steps': 29416, 'loss/train': 1.6235064268112183} 02/25/2022 15:41:52 - INFO - codeparrot_training - Step 29417: {'lr': 0.0001945543724292885, 'samples': 15062016, 'steps': 29417, 'loss/train': 1.1945921182632446} 02/25/2022 15:41:55 - INFO - codeparrot_training - Step 29418: {'lr': 0.00019453841757447344, 'samples': 15062528, 'steps': 29418, 'loss/train': 1.0461317300796509} 02/25/2022 15:42:00 - INFO - codeparrot_training - Step 29419: {'lr': 0.00019452246295723814, 'samples': 15063040, 'steps': 29419, 'loss/train': 1.671424150466919} 02/25/2022 15:42:04 - INFO - codeparrot_training - Step 29420: {'lr': 0.00019450650857765102, 'samples': 15063552, 'steps': 29420, 'loss/train': 1.7188565731048584} 02/25/2022 15:42:09 - INFO - codeparrot_training - Step 29421: {'lr': 0.0001944905544357804, 'samples': 15064064, 'steps': 29421, 'loss/train': 0.5750018358230591} 02/25/2022 15:42:13 - INFO - codeparrot_training - Step 29422: {'lr': 0.00019447460053169464, 'samples': 15064576, 'steps': 29422, 'loss/train': 1.7493497133255005} 02/25/2022 15:42:19 - INFO - codeparrot_training - Step 29423: {'lr': 0.00019445864686546202, 'samples': 15065088, 'steps': 29423, 'loss/train': 2.12671160697937} 02/25/2022 15:42:24 - INFO - codeparrot_training - Step 29424: {'lr': 0.0001944426934371509, 'samples': 15065600, 'steps': 29424, 'loss/train': 1.4279041290283203} 02/25/2022 15:42:28 - INFO - codeparrot_training - Step 29425: {'lr': 0.00019442674024682967, 'samples': 15066112, 'steps': 29425, 'loss/train': 1.9587042331695557} 02/25/2022 15:42:33 - INFO - codeparrot_training - Step 29426: {'lr': 0.0001944107872945667, 'samples': 15066624, 'steps': 29426, 'loss/train': 1.8342225551605225} 02/25/2022 15:42:36 - INFO - codeparrot_training - Step 29427: {'lr': 0.0001943948345804302, 'samples': 15067136, 'steps': 29427, 'loss/train': 1.7993742227554321} 02/25/2022 15:42:43 - INFO - codeparrot_training - Step 29428: {'lr': 0.00019437888210448854, 'samples': 15067648, 'steps': 29428, 'loss/train': 2.1386380195617676} 02/25/2022 15:42:46 - INFO - codeparrot_training - Step 29429: {'lr': 0.00019436292986681008, 'samples': 15068160, 'steps': 29429, 'loss/train': 1.4390922784805298} 02/25/2022 15:42:52 - INFO - codeparrot_training - Step 29430: {'lr': 0.0001943469778674633, 'samples': 15068672, 'steps': 29430, 'loss/train': 1.1585594415664673} 02/25/2022 15:42:55 - INFO - codeparrot_training - Step 29431: {'lr': 0.0001943310261065163, 'samples': 15069184, 'steps': 29431, 'loss/train': 2.0845022201538086} 02/25/2022 15:43:01 - INFO - codeparrot_training - Step 29432: {'lr': 0.00019431507458403749, 'samples': 15069696, 'steps': 29432, 'loss/train': 3.0166690349578857} 02/25/2022 15:43:04 - INFO - codeparrot_training - Step 29433: {'lr': 0.0001942991233000952, 'samples': 15070208, 'steps': 29433, 'loss/train': 0.1814434826374054} 02/25/2022 15:43:10 - INFO - codeparrot_training - Step 29434: {'lr': 0.0001942831722547579, 'samples': 15070720, 'steps': 29434, 'loss/train': 1.883793592453003} 02/25/2022 15:43:13 - INFO - codeparrot_training - Step 29435: {'lr': 0.0001942672214480937, 'samples': 15071232, 'steps': 29435, 'loss/train': 1.7238143682479858} 02/25/2022 15:43:19 - INFO - codeparrot_training - Step 29436: {'lr': 0.00019425127088017102, 'samples': 15071744, 'steps': 29436, 'loss/train': 1.0666111707687378} 02/25/2022 15:43:22 - INFO - codeparrot_training - Step 29437: {'lr': 0.0001942353205510582, 'samples': 15072256, 'steps': 29437, 'loss/train': 1.8013737201690674} 02/25/2022 15:43:29 - INFO - codeparrot_training - Step 29438: {'lr': 0.00019421937046082356, 'samples': 15072768, 'steps': 29438, 'loss/train': 2.0546984672546387} 02/25/2022 15:43:32 - INFO - codeparrot_training - Step 29439: {'lr': 0.0001942034206095355, 'samples': 15073280, 'steps': 29439, 'loss/train': 1.9984092712402344} 02/25/2022 15:43:36 - INFO - codeparrot_training - Step 29440: {'lr': 0.0001941874709972622, 'samples': 15073792, 'steps': 29440, 'loss/train': 2.270327091217041} 02/25/2022 15:43:41 - INFO - codeparrot_training - Step 29441: {'lr': 0.00019417152162407204, 'samples': 15074304, 'steps': 29441, 'loss/train': 0.8403199315071106} 02/25/2022 15:43:45 - INFO - codeparrot_training - Step 29442: {'lr': 0.00019415557249003333, 'samples': 15074816, 'steps': 29442, 'loss/train': 1.4506248235702515} 02/25/2022 15:43:50 - INFO - codeparrot_training - Step 29443: {'lr': 0.00019413962359521452, 'samples': 15075328, 'steps': 29443, 'loss/train': 1.1142593622207642} 02/25/2022 15:43:54 - INFO - codeparrot_training - Step 29444: {'lr': 0.00019412367493968374, 'samples': 15075840, 'steps': 29444, 'loss/train': 1.184648871421814} 02/25/2022 15:43:59 - INFO - codeparrot_training - Step 29445: {'lr': 0.0001941077265235094, 'samples': 15076352, 'steps': 29445, 'loss/train': 0.9934272170066833} 02/25/2022 15:44:05 - INFO - codeparrot_training - Step 29446: {'lr': 0.00019409177834675982, 'samples': 15076864, 'steps': 29446, 'loss/train': 1.4621813297271729} 02/25/2022 15:44:08 - INFO - codeparrot_training - Step 29447: {'lr': 0.00019407583040950338, 'samples': 15077376, 'steps': 29447, 'loss/train': 1.6000694036483765} 02/25/2022 15:44:12 - INFO - codeparrot_training - Step 29448: {'lr': 0.00019405988271180825, 'samples': 15077888, 'steps': 29448, 'loss/train': 9.644929885864258} 02/25/2022 15:44:17 - INFO - codeparrot_training - Step 29449: {'lr': 0.0001940439352537428, 'samples': 15078400, 'steps': 29449, 'loss/train': 1.9958058595657349} 02/25/2022 15:44:21 - INFO - codeparrot_training - Step 29450: {'lr': 0.00019402798803537538, 'samples': 15078912, 'steps': 29450, 'loss/train': 1.8296359777450562} 02/25/2022 15:44:26 - INFO - codeparrot_training - Step 29451: {'lr': 0.00019401204105677433, 'samples': 15079424, 'steps': 29451, 'loss/train': 1.334689974784851} 02/25/2022 15:44:30 - INFO - codeparrot_training - Step 29452: {'lr': 0.0001939960943180079, 'samples': 15079936, 'steps': 29452, 'loss/train': 1.03721022605896} 02/25/2022 15:44:36 - INFO - codeparrot_training - Step 29453: {'lr': 0.0001939801478191444, 'samples': 15080448, 'steps': 29453, 'loss/train': 1.6068592071533203} 02/25/2022 15:44:41 - INFO - codeparrot_training - Step 29454: {'lr': 0.00019396420156025213, 'samples': 15080960, 'steps': 29454, 'loss/train': 1.4263707399368286} 02/25/2022 15:44:45 - INFO - codeparrot_training - Step 29455: {'lr': 0.0001939482555413995, 'samples': 15081472, 'steps': 29455, 'loss/train': 1.5901168584823608} 02/25/2022 15:44:50 - INFO - codeparrot_training - Step 29456: {'lr': 0.00019393230976265475, 'samples': 15081984, 'steps': 29456, 'loss/train': 1.0355278253555298} 02/25/2022 15:44:54 - INFO - codeparrot_training - Step 29457: {'lr': 0.0001939163642240861, 'samples': 15082496, 'steps': 29457, 'loss/train': 2.120537757873535} 02/25/2022 15:44:59 - INFO - codeparrot_training - Step 29458: {'lr': 0.00019390041892576202, 'samples': 15083008, 'steps': 29458, 'loss/train': 1.9373555183410645} 02/25/2022 15:45:03 - INFO - codeparrot_training - Step 29459: {'lr': 0.00019388447386775066, 'samples': 15083520, 'steps': 29459, 'loss/train': 1.7710665464401245} 02/25/2022 15:45:08 - INFO - codeparrot_training - Step 29460: {'lr': 0.0001938685290501204, 'samples': 15084032, 'steps': 29460, 'loss/train': 1.3460595607757568} 02/25/2022 15:45:12 - INFO - codeparrot_training - Step 29461: {'lr': 0.00019385258447293958, 'samples': 15084544, 'steps': 29461, 'loss/train': 0.07201723009347916} 02/25/2022 15:45:18 - INFO - codeparrot_training - Step 29462: {'lr': 0.00019383664013627647, 'samples': 15085056, 'steps': 29462, 'loss/train': 1.4551562070846558} 02/25/2022 15:45:21 - INFO - codeparrot_training - Step 29463: {'lr': 0.0001938206960401993, 'samples': 15085568, 'steps': 29463, 'loss/train': 0.5092214345932007} 02/25/2022 15:45:27 - INFO - codeparrot_training - Step 29464: {'lr': 0.00019380475218477644, 'samples': 15086080, 'steps': 29464, 'loss/train': 2.0804381370544434} 02/25/2022 15:45:30 - INFO - codeparrot_training - Step 29465: {'lr': 0.0001937888085700763, 'samples': 15086592, 'steps': 29465, 'loss/train': 2.38860821723938} 02/25/2022 15:45:36 - INFO - codeparrot_training - Step 29466: {'lr': 0.00019377286519616692, 'samples': 15087104, 'steps': 29466, 'loss/train': 1.8411568403244019} 02/25/2022 15:45:39 - INFO - codeparrot_training - Step 29467: {'lr': 0.00019375692206311673, 'samples': 15087616, 'steps': 29467, 'loss/train': 0.9012255072593689} 02/25/2022 15:45:45 - INFO - codeparrot_training - Step 29468: {'lr': 0.00019374097917099404, 'samples': 15088128, 'steps': 29468, 'loss/train': 1.1848349571228027} 02/25/2022 15:45:48 - INFO - codeparrot_training - Step 29469: {'lr': 0.00019372503651986721, 'samples': 15088640, 'steps': 29469, 'loss/train': 2.2058956623077393} 02/25/2022 15:45:54 - INFO - codeparrot_training - Step 29470: {'lr': 0.0001937090941098044, 'samples': 15089152, 'steps': 29470, 'loss/train': 1.4756314754486084} 02/25/2022 15:45:57 - INFO - codeparrot_training - Step 29471: {'lr': 0.0001936931519408739, 'samples': 15089664, 'steps': 29471, 'loss/train': 0.39692986011505127} 02/25/2022 15:46:04 - INFO - codeparrot_training - Step 29472: {'lr': 0.00019367721001314412, 'samples': 15090176, 'steps': 29472, 'loss/train': 1.7813066244125366} 02/25/2022 15:46:07 - INFO - codeparrot_training - Step 29473: {'lr': 0.00019366126832668325, 'samples': 15090688, 'steps': 29473, 'loss/train': 2.5521273612976074} 02/25/2022 15:46:13 - INFO - codeparrot_training - Step 29474: {'lr': 0.00019364532688155974, 'samples': 15091200, 'steps': 29474, 'loss/train': 1.4444141387939453} 02/25/2022 15:46:16 - INFO - codeparrot_training - Step 29475: {'lr': 0.00019362938567784162, 'samples': 15091712, 'steps': 29475, 'loss/train': 1.8973220586776733} 02/25/2022 15:46:22 - INFO - codeparrot_training - Step 29476: {'lr': 0.00019361344471559733, 'samples': 15092224, 'steps': 29476, 'loss/train': 2.45803165435791} 02/25/2022 15:46:25 - INFO - codeparrot_training - Step 29477: {'lr': 0.00019359750399489513, 'samples': 15092736, 'steps': 29477, 'loss/train': 0.8903579711914062} 02/25/2022 15:46:31 - INFO - codeparrot_training - Step 29478: {'lr': 0.00019358156351580343, 'samples': 15093248, 'steps': 29478, 'loss/train': 1.7931666374206543} 02/25/2022 15:46:34 - INFO - codeparrot_training - Step 29479: {'lr': 0.00019356562327839028, 'samples': 15093760, 'steps': 29479, 'loss/train': 2.4955785274505615} 02/25/2022 15:46:40 - INFO - codeparrot_training - Step 29480: {'lr': 0.0001935496832827241, 'samples': 15094272, 'steps': 29480, 'loss/train': 2.1500279903411865} 02/25/2022 15:46:43 - INFO - codeparrot_training - Step 29481: {'lr': 0.00019353374352887312, 'samples': 15094784, 'steps': 29481, 'loss/train': 1.5815048217773438} 02/25/2022 15:46:49 - INFO - codeparrot_training - Step 29482: {'lr': 0.00019351780401690577, 'samples': 15095296, 'steps': 29482, 'loss/train': 2.6938416957855225} 02/25/2022 15:46:52 - INFO - codeparrot_training - Step 29483: {'lr': 0.00019350186474689015, 'samples': 15095808, 'steps': 29483, 'loss/train': 0.2253606915473938} 02/25/2022 15:46:58 - INFO - codeparrot_training - Step 29484: {'lr': 0.00019348592571889455, 'samples': 15096320, 'steps': 29484, 'loss/train': 1.763107180595398} 02/25/2022 15:47:02 - INFO - codeparrot_training - Step 29485: {'lr': 0.00019346998693298728, 'samples': 15096832, 'steps': 29485, 'loss/train': 2.435241937637329} 02/25/2022 15:47:07 - INFO - codeparrot_training - Step 29486: {'lr': 0.0001934540483892367, 'samples': 15097344, 'steps': 29486, 'loss/train': 2.925766706466675} 02/25/2022 15:47:11 - INFO - codeparrot_training - Step 29487: {'lr': 0.0001934381100877111, 'samples': 15097856, 'steps': 29487, 'loss/train': 1.9237571954727173} 02/25/2022 15:47:17 - INFO - codeparrot_training - Step 29488: {'lr': 0.00019342217202847857, 'samples': 15098368, 'steps': 29488, 'loss/train': 1.8808329105377197} 02/25/2022 15:47:20 - INFO - codeparrot_training - Step 29489: {'lr': 0.00019340623421160746, 'samples': 15098880, 'steps': 29489, 'loss/train': 0.7876949310302734} 02/25/2022 15:47:26 - INFO - codeparrot_training - Step 29490: {'lr': 0.00019339029663716613, 'samples': 15099392, 'steps': 29490, 'loss/train': 1.7941373586654663} 02/25/2022 15:47:29 - INFO - codeparrot_training - Step 29491: {'lr': 0.00019337435930522284, 'samples': 15099904, 'steps': 29491, 'loss/train': 1.0998692512512207} 02/25/2022 15:47:35 - INFO - codeparrot_training - Step 29492: {'lr': 0.00019335842221584573, 'samples': 15100416, 'steps': 29492, 'loss/train': 1.1510341167449951} 02/25/2022 15:47:38 - INFO - codeparrot_training - Step 29493: {'lr': 0.00019334248536910314, 'samples': 15100928, 'steps': 29493, 'loss/train': 1.662171483039856} 02/25/2022 15:47:44 - INFO - codeparrot_training - Step 29494: {'lr': 0.0001933265487650634, 'samples': 15101440, 'steps': 29494, 'loss/train': 0.19030936062335968} 02/25/2022 15:47:47 - INFO - codeparrot_training - Step 29495: {'lr': 0.0001933106124037948, 'samples': 15101952, 'steps': 29495, 'loss/train': 2.00911808013916} 02/25/2022 15:47:53 - INFO - codeparrot_training - Step 29496: {'lr': 0.00019329467628536543, 'samples': 15102464, 'steps': 29496, 'loss/train': 1.2284132242202759} 02/25/2022 15:47:56 - INFO - codeparrot_training - Step 29497: {'lr': 0.00019327874040984367, 'samples': 15102976, 'steps': 29497, 'loss/train': 1.6208912134170532} 02/25/2022 15:48:03 - INFO - codeparrot_training - Step 29498: {'lr': 0.00019326280477729782, 'samples': 15103488, 'steps': 29498, 'loss/train': 2.196002960205078} 02/25/2022 15:48:06 - INFO - codeparrot_training - Step 29499: {'lr': 0.00019324686938779607, 'samples': 15104000, 'steps': 29499, 'loss/train': 1.610750675201416} 02/25/2022 15:48:11 - INFO - codeparrot_training - Step 29500: {'lr': 0.00019323093424140672, 'samples': 15104512, 'steps': 29500, 'loss/train': 2.109882354736328} 02/25/2022 15:48:15 - INFO - codeparrot_training - Step 29501: {'lr': 0.000193214999338198, 'samples': 15105024, 'steps': 29501, 'loss/train': 2.144434928894043} 02/25/2022 15:48:20 - INFO - codeparrot_training - Step 29502: {'lr': 0.00019319906467823823, 'samples': 15105536, 'steps': 29502, 'loss/train': 2.1958749294281006} 02/25/2022 15:48:24 - INFO - codeparrot_training - Step 29503: {'lr': 0.0001931831302615956, 'samples': 15106048, 'steps': 29503, 'loss/train': 1.651770830154419} 02/25/2022 15:48:29 - INFO - codeparrot_training - Step 29504: {'lr': 0.00019316719608833844, 'samples': 15106560, 'steps': 29504, 'loss/train': 1.9949365854263306} 02/25/2022 15:48:33 - INFO - codeparrot_training - Step 29505: {'lr': 0.00019315126215853495, 'samples': 15107072, 'steps': 29505, 'loss/train': 1.9250798225402832} 02/25/2022 15:48:39 - INFO - codeparrot_training - Step 29506: {'lr': 0.00019313532847225334, 'samples': 15107584, 'steps': 29506, 'loss/train': 1.9694836139678955} 02/25/2022 15:48:42 - INFO - codeparrot_training - Step 29507: {'lr': 0.00019311939502956196, 'samples': 15108096, 'steps': 29507, 'loss/train': 2.1735928058624268} 02/25/2022 15:48:48 - INFO - codeparrot_training - Step 29508: {'lr': 0.00019310346183052906, 'samples': 15108608, 'steps': 29508, 'loss/train': 0.9547359943389893} 02/25/2022 15:48:52 - INFO - codeparrot_training - Step 29509: {'lr': 0.00019308752887522288, 'samples': 15109120, 'steps': 29509, 'loss/train': 0.9513463377952576} 02/25/2022 15:48:58 - INFO - codeparrot_training - Step 29510: {'lr': 0.00019307159616371159, 'samples': 15109632, 'steps': 29510, 'loss/train': 2.1413819789886475} 02/25/2022 15:49:01 - INFO - codeparrot_training - Step 29511: {'lr': 0.0001930556636960635, 'samples': 15110144, 'steps': 29511, 'loss/train': 2.062797784805298} 02/25/2022 15:49:07 - INFO - codeparrot_training - Step 29512: {'lr': 0.00019303973147234688, 'samples': 15110656, 'steps': 29512, 'loss/train': 1.8565301895141602} 02/25/2022 15:49:10 - INFO - codeparrot_training - Step 29513: {'lr': 0.00019302379949263004, 'samples': 15111168, 'steps': 29513, 'loss/train': 2.2574539184570312} 02/25/2022 15:49:16 - INFO - codeparrot_training - Step 29514: {'lr': 0.0001930078677569811, 'samples': 15111680, 'steps': 29514, 'loss/train': 2.2472429275512695} 02/25/2022 15:49:19 - INFO - codeparrot_training - Step 29515: {'lr': 0.0001929919362654683, 'samples': 15112192, 'steps': 29515, 'loss/train': 2.3249053955078125} 02/25/2022 15:49:25 - INFO - codeparrot_training - Step 29516: {'lr': 0.00019297600501816, 'samples': 15112704, 'steps': 29516, 'loss/train': 1.8411891460418701} 02/25/2022 15:49:28 - INFO - codeparrot_training - Step 29517: {'lr': 0.00019296007401512446, 'samples': 15113216, 'steps': 29517, 'loss/train': 2.0202596187591553} 02/25/2022 15:49:34 - INFO - codeparrot_training - Step 29518: {'lr': 0.00019294414325642974, 'samples': 15113728, 'steps': 29518, 'loss/train': 2.335059881210327} 02/25/2022 15:49:37 - INFO - codeparrot_training - Step 29519: {'lr': 0.00019292821274214417, 'samples': 15114240, 'steps': 29519, 'loss/train': 2.269178867340088} 02/25/2022 15:49:43 - INFO - codeparrot_training - Step 29520: {'lr': 0.00019291228247233605, 'samples': 15114752, 'steps': 29520, 'loss/train': 1.6253571510314941} 02/25/2022 15:49:47 - INFO - codeparrot_training - Step 29521: {'lr': 0.0001928963524470736, 'samples': 15115264, 'steps': 29521, 'loss/train': 1.1992791891098022} 02/25/2022 15:49:54 - INFO - codeparrot_training - Step 29522: {'lr': 0.00019288042266642512, 'samples': 15115776, 'steps': 29522, 'loss/train': 0.9252859354019165} 02/25/2022 15:49:57 - INFO - codeparrot_training - Step 29523: {'lr': 0.00019286449313045867, 'samples': 15116288, 'steps': 29523, 'loss/train': 1.6279892921447754} 02/25/2022 15:50:01 - INFO - codeparrot_training - Step 29524: {'lr': 0.0001928485638392426, 'samples': 15116800, 'steps': 29524, 'loss/train': 1.7121692895889282} 02/25/2022 15:50:06 - INFO - codeparrot_training - Step 29525: {'lr': 0.00019283263479284512, 'samples': 15117312, 'steps': 29525, 'loss/train': 1.0754340887069702} 02/25/2022 15:50:10 - INFO - codeparrot_training - Step 29526: {'lr': 0.0001928167059913346, 'samples': 15117824, 'steps': 29526, 'loss/train': 1.8617494106292725} 02/25/2022 15:50:15 - INFO - codeparrot_training - Step 29527: {'lr': 0.00019280077743477904, 'samples': 15118336, 'steps': 29527, 'loss/train': 2.3506860733032227} 02/25/2022 15:50:19 - INFO - codeparrot_training - Step 29528: {'lr': 0.00019278484912324678, 'samples': 15118848, 'steps': 29528, 'loss/train': 2.4082860946655273} 02/25/2022 15:50:24 - INFO - codeparrot_training - Step 29529: {'lr': 0.00019276892105680606, 'samples': 15119360, 'steps': 29529, 'loss/train': 1.2569336891174316} 02/25/2022 15:50:30 - INFO - codeparrot_training - Step 29530: {'lr': 0.00019275299323552524, 'samples': 15119872, 'steps': 29530, 'loss/train': 2.0977346897125244} 02/25/2022 15:50:33 - INFO - codeparrot_training - Step 29531: {'lr': 0.00019273706565947228, 'samples': 15120384, 'steps': 29531, 'loss/train': 1.2070053815841675} 02/25/2022 15:50:40 - INFO - codeparrot_training - Step 29532: {'lr': 0.00019272113832871556, 'samples': 15120896, 'steps': 29532, 'loss/train': 2.329711675643921} 02/25/2022 15:50:43 - INFO - codeparrot_training - Step 29533: {'lr': 0.0001927052112433233, 'samples': 15121408, 'steps': 29533, 'loss/train': 1.1488196849822998} 02/25/2022 15:50:49 - INFO - codeparrot_training - Step 29534: {'lr': 0.0001926892844033637, 'samples': 15121920, 'steps': 29534, 'loss/train': 1.0951377153396606} 02/25/2022 15:50:52 - INFO - codeparrot_training - Step 29535: {'lr': 0.00019267335780890515, 'samples': 15122432, 'steps': 29535, 'loss/train': 5.3136820793151855} 02/25/2022 15:50:57 - INFO - codeparrot_training - Step 29536: {'lr': 0.0001926574314600156, 'samples': 15122944, 'steps': 29536, 'loss/train': 2.738659620285034} 02/25/2022 15:51:01 - INFO - codeparrot_training - Step 29537: {'lr': 0.00019264150535676342, 'samples': 15123456, 'steps': 29537, 'loss/train': 0.5020974278450012} 02/25/2022 15:51:06 - INFO - codeparrot_training - Step 29538: {'lr': 0.00019262557949921685, 'samples': 15123968, 'steps': 29538, 'loss/train': 0.9084689617156982} 02/25/2022 15:51:10 - INFO - codeparrot_training - Step 29539: {'lr': 0.0001926096538874441, 'samples': 15124480, 'steps': 29539, 'loss/train': 1.8927985429763794} 02/25/2022 15:51:15 - INFO - codeparrot_training - Step 29540: {'lr': 0.0001925937285215133, 'samples': 15124992, 'steps': 29540, 'loss/train': 1.073237657546997} 02/25/2022 15:51:19 - INFO - codeparrot_training - Step 29541: {'lr': 0.00019257780340149275, 'samples': 15125504, 'steps': 29541, 'loss/train': 2.1037042140960693} 02/25/2022 15:51:25 - INFO - codeparrot_training - Step 29542: {'lr': 0.0001925618785274507, 'samples': 15126016, 'steps': 29542, 'loss/train': 1.9683488607406616} 02/25/2022 15:51:29 - INFO - codeparrot_training - Step 29543: {'lr': 0.00019254595389945535, 'samples': 15126528, 'steps': 29543, 'loss/train': 0.8652065396308899} 02/25/2022 15:51:34 - INFO - codeparrot_training - Step 29544: {'lr': 0.0001925300295175748, 'samples': 15127040, 'steps': 29544, 'loss/train': 1.3263882398605347} 02/25/2022 15:51:37 - INFO - codeparrot_training - Step 29545: {'lr': 0.00019251410538187745, 'samples': 15127552, 'steps': 29545, 'loss/train': 2.102966070175171} 02/25/2022 15:51:43 - INFO - codeparrot_training - Step 29546: {'lr': 0.00019249818149243134, 'samples': 15128064, 'steps': 29546, 'loss/train': 1.1589629650115967} 02/25/2022 15:51:46 - INFO - codeparrot_training - Step 29547: {'lr': 0.00019248225784930477, 'samples': 15128576, 'steps': 29547, 'loss/train': 1.2822563648223877} 02/25/2022 15:51:52 - INFO - codeparrot_training - Step 29548: {'lr': 0.00019246633445256598, 'samples': 15129088, 'steps': 29548, 'loss/train': 1.6994547843933105} 02/25/2022 15:51:55 - INFO - codeparrot_training - Step 29549: {'lr': 0.00019245041130228313, 'samples': 15129600, 'steps': 29549, 'loss/train': 1.9016213417053223} 02/25/2022 15:52:01 - INFO - codeparrot_training - Step 29550: {'lr': 0.00019243448839852443, 'samples': 15130112, 'steps': 29550, 'loss/train': 2.334958076477051} 02/25/2022 15:52:04 - INFO - codeparrot_training - Step 29551: {'lr': 0.00019241856574135808, 'samples': 15130624, 'steps': 29551, 'loss/train': 1.5798158645629883} 02/25/2022 15:52:10 - INFO - codeparrot_training - Step 29552: {'lr': 0.00019240264333085245, 'samples': 15131136, 'steps': 29552, 'loss/train': 4.411936283111572} 02/25/2022 15:52:14 - INFO - codeparrot_training - Step 29553: {'lr': 0.00019238672116707545, 'samples': 15131648, 'steps': 29553, 'loss/train': 0.17052246630191803} 02/25/2022 15:52:20 - INFO - codeparrot_training - Step 29554: {'lr': 0.00019237079925009547, 'samples': 15132160, 'steps': 29554, 'loss/train': 2.920450210571289} 02/25/2022 15:52:23 - INFO - codeparrot_training - Step 29555: {'lr': 0.00019235487757998069, 'samples': 15132672, 'steps': 29555, 'loss/train': 2.9069278240203857} 02/25/2022 15:52:29 - INFO - codeparrot_training - Step 29556: {'lr': 0.00019233895615679942, 'samples': 15133184, 'steps': 29556, 'loss/train': 3.27300763130188} 02/25/2022 15:52:32 - INFO - codeparrot_training - Step 29557: {'lr': 0.00019232303498061963, 'samples': 15133696, 'steps': 29557, 'loss/train': 0.46360430121421814} 02/25/2022 15:52:38 - INFO - codeparrot_training - Step 29558: {'lr': 0.00019230711405150962, 'samples': 15134208, 'steps': 29558, 'loss/train': 2.1411924362182617} 02/25/2022 15:52:41 - INFO - codeparrot_training - Step 29559: {'lr': 0.00019229119336953765, 'samples': 15134720, 'steps': 29559, 'loss/train': 0.42131873965263367} 02/25/2022 15:52:47 - INFO - codeparrot_training - Step 29560: {'lr': 0.00019227527293477187, 'samples': 15135232, 'steps': 29560, 'loss/train': 2.43323016166687} 02/25/2022 15:52:50 - INFO - codeparrot_training - Step 29561: {'lr': 0.00019225935274728062, 'samples': 15135744, 'steps': 29561, 'loss/train': 0.8900261521339417} 02/25/2022 15:52:56 - INFO - codeparrot_training - Step 29562: {'lr': 0.00019224343280713186, 'samples': 15136256, 'steps': 29562, 'loss/train': 0.5050694346427917} 02/25/2022 15:52:59 - INFO - codeparrot_training - Step 29563: {'lr': 0.00019222751311439386, 'samples': 15136768, 'steps': 29563, 'loss/train': 1.0143170356750488} 02/25/2022 15:53:05 - INFO - codeparrot_training - Step 29564: {'lr': 0.00019221159366913487, 'samples': 15137280, 'steps': 29564, 'loss/train': 0.9791566133499146} 02/25/2022 15:53:08 - INFO - codeparrot_training - Step 29565: {'lr': 0.00019219567447142317, 'samples': 15137792, 'steps': 29565, 'loss/train': 1.795219898223877} 02/25/2022 15:53:14 - INFO - codeparrot_training - Step 29566: {'lr': 0.00019217975552132674, 'samples': 15138304, 'steps': 29566, 'loss/train': 1.645015835762024} 02/25/2022 15:53:17 - INFO - codeparrot_training - Step 29567: {'lr': 0.00019216383681891386, 'samples': 15138816, 'steps': 29567, 'loss/train': 2.1263914108276367} 02/25/2022 15:53:24 - INFO - codeparrot_training - Step 29568: {'lr': 0.00019214791836425277, 'samples': 15139328, 'steps': 29568, 'loss/train': 1.382897138595581} 02/25/2022 15:53:27 - INFO - codeparrot_training - Step 29569: {'lr': 0.00019213200015741162, 'samples': 15139840, 'steps': 29569, 'loss/train': 1.7313216924667358} 02/25/2022 15:53:33 - INFO - codeparrot_training - Step 29570: {'lr': 0.0001921160821984587, 'samples': 15140352, 'steps': 29570, 'loss/train': 2.05067777633667} 02/25/2022 15:53:36 - INFO - codeparrot_training - Step 29571: {'lr': 0.000192100164487462, 'samples': 15140864, 'steps': 29571, 'loss/train': 1.8695744276046753} 02/25/2022 15:53:42 - INFO - codeparrot_training - Step 29572: {'lr': 0.00019208424702448978, 'samples': 15141376, 'steps': 29572, 'loss/train': 1.1515432596206665} 02/25/2022 15:53:45 - INFO - codeparrot_training - Step 29573: {'lr': 0.0001920683298096103, 'samples': 15141888, 'steps': 29573, 'loss/train': 1.4647974967956543} 02/25/2022 15:53:51 - INFO - codeparrot_training - Step 29574: {'lr': 0.0001920524128428918, 'samples': 15142400, 'steps': 29574, 'loss/train': 1.6085587739944458} 02/25/2022 15:53:54 - INFO - codeparrot_training - Step 29575: {'lr': 0.00019203649612440225, 'samples': 15142912, 'steps': 29575, 'loss/train': 2.2466847896575928} 02/25/2022 15:54:00 - INFO - codeparrot_training - Step 29576: {'lr': 0.00019202057965420993, 'samples': 15143424, 'steps': 29576, 'loss/train': 2.081496000289917} 02/25/2022 15:54:03 - INFO - codeparrot_training - Step 29577: {'lr': 0.00019200466343238305, 'samples': 15143936, 'steps': 29577, 'loss/train': 2.046339750289917} 02/25/2022 15:54:09 - INFO - codeparrot_training - Step 29578: {'lr': 0.00019198874745898986, 'samples': 15144448, 'steps': 29578, 'loss/train': 1.7536154985427856} 02/25/2022 15:54:12 - INFO - codeparrot_training - Step 29579: {'lr': 0.00019197283173409837, 'samples': 15144960, 'steps': 29579, 'loss/train': 0.6974378228187561} 02/25/2022 15:54:19 - INFO - codeparrot_training - Step 29580: {'lr': 0.00019195691625777686, 'samples': 15145472, 'steps': 29580, 'loss/train': 2.0256919860839844} 02/25/2022 15:54:22 - INFO - codeparrot_training - Step 29581: {'lr': 0.0001919410010300935, 'samples': 15145984, 'steps': 29581, 'loss/train': 1.1672112941741943} 02/25/2022 15:54:28 - INFO - codeparrot_training - Step 29582: {'lr': 0.00019192508605111647, 'samples': 15146496, 'steps': 29582, 'loss/train': 1.1593413352966309} 02/25/2022 15:54:31 - INFO - codeparrot_training - Step 29583: {'lr': 0.00019190917132091393, 'samples': 15147008, 'steps': 29583, 'loss/train': 1.7980064153671265} 02/25/2022 15:54:37 - INFO - codeparrot_training - Step 29584: {'lr': 0.000191893256839554, 'samples': 15147520, 'steps': 29584, 'loss/train': 0.8032127022743225} 02/25/2022 15:54:40 - INFO - codeparrot_training - Step 29585: {'lr': 0.00019187734260710498, 'samples': 15148032, 'steps': 29585, 'loss/train': 1.225101351737976} 02/25/2022 15:54:45 - INFO - codeparrot_training - Step 29586: {'lr': 0.00019186142862363491, 'samples': 15148544, 'steps': 29586, 'loss/train': 1.4679492712020874} 02/25/2022 15:54:49 - INFO - codeparrot_training - Step 29587: {'lr': 0.00019184551488921208, 'samples': 15149056, 'steps': 29587, 'loss/train': 0.7630745768547058} 02/25/2022 15:54:54 - INFO - codeparrot_training - Step 29588: {'lr': 0.00019182960140390454, 'samples': 15149568, 'steps': 29588, 'loss/train': 1.9443163871765137} 02/25/2022 15:54:58 - INFO - codeparrot_training - Step 29589: {'lr': 0.00019181368816778056, 'samples': 15150080, 'steps': 29589, 'loss/train': 2.063915729522705} 02/25/2022 15:55:04 - INFO - codeparrot_training - Step 29590: {'lr': 0.0001917977751809082, 'samples': 15150592, 'steps': 29590, 'loss/train': 2.680968999862671} 02/25/2022 15:55:08 - INFO - codeparrot_training - Step 29591: {'lr': 0.00019178186244335577, 'samples': 15151104, 'steps': 29591, 'loss/train': 1.933132290840149} 02/25/2022 15:55:13 - INFO - codeparrot_training - Step 29592: {'lr': 0.00019176594995519135, 'samples': 15151616, 'steps': 29592, 'loss/train': 2.1474061012268066} 02/25/2022 15:55:17 - INFO - codeparrot_training - Step 29593: {'lr': 0.00019175003771648307, 'samples': 15152128, 'steps': 29593, 'loss/train': 1.4122283458709717} 02/25/2022 15:55:22 - INFO - codeparrot_training - Step 29594: {'lr': 0.0001917341257272991, 'samples': 15152640, 'steps': 29594, 'loss/train': 2.515367031097412} 02/25/2022 15:55:26 - INFO - codeparrot_training - Step 29595: {'lr': 0.00019171821398770764, 'samples': 15153152, 'steps': 29595, 'loss/train': 2.163191556930542} 02/25/2022 15:55:31 - INFO - codeparrot_training - Step 29596: {'lr': 0.00019170230249777696, 'samples': 15153664, 'steps': 29596, 'loss/train': 0.5583831071853638} 02/25/2022 15:55:35 - INFO - codeparrot_training - Step 29597: {'lr': 0.00019168639125757497, 'samples': 15154176, 'steps': 29597, 'loss/train': 1.909466028213501} 02/25/2022 15:55:40 - INFO - codeparrot_training - Step 29598: {'lr': 0.00019167048026716999, 'samples': 15154688, 'steps': 29598, 'loss/train': 1.9099794626235962} 02/25/2022 15:55:44 - INFO - codeparrot_training - Step 29599: {'lr': 0.00019165456952663015, 'samples': 15155200, 'steps': 29599, 'loss/train': 0.5206064581871033} 02/25/2022 15:55:50 - INFO - codeparrot_training - Step 29600: {'lr': 0.00019163865903602372, 'samples': 15155712, 'steps': 29600, 'loss/train': 1.6779024600982666} 02/25/2022 15:55:54 - INFO - codeparrot_training - Step 29601: {'lr': 0.00019162274879541864, 'samples': 15156224, 'steps': 29601, 'loss/train': 2.0416195392608643} 02/25/2022 15:55:59 - INFO - codeparrot_training - Step 29602: {'lr': 0.00019160683880488314, 'samples': 15156736, 'steps': 29602, 'loss/train': 1.99224853515625} 02/25/2022 15:56:03 - INFO - codeparrot_training - Step 29603: {'lr': 0.00019159092906448542, 'samples': 15157248, 'steps': 29603, 'loss/train': 1.9454398155212402} 02/25/2022 15:56:08 - INFO - codeparrot_training - Step 29604: {'lr': 0.00019157501957429372, 'samples': 15157760, 'steps': 29604, 'loss/train': 0.9746183156967163} 02/25/2022 15:56:12 - INFO - codeparrot_training - Step 29605: {'lr': 0.00019155911033437597, 'samples': 15158272, 'steps': 29605, 'loss/train': 2.7291858196258545} 02/25/2022 15:56:17 - INFO - codeparrot_training - Step 29606: {'lr': 0.00019154320134480044, 'samples': 15158784, 'steps': 29606, 'loss/train': 1.805185079574585} 02/25/2022 15:56:21 - INFO - codeparrot_training - Step 29607: {'lr': 0.0001915272926056353, 'samples': 15159296, 'steps': 29607, 'loss/train': 1.6780824661254883} 02/25/2022 15:56:26 - INFO - codeparrot_training - Step 29608: {'lr': 0.00019151138411694862, 'samples': 15159808, 'steps': 29608, 'loss/train': 1.4848872423171997} 02/25/2022 15:56:30 - INFO - codeparrot_training - Step 29609: {'lr': 0.0001914954758788087, 'samples': 15160320, 'steps': 29609, 'loss/train': 2.320767879486084} 02/25/2022 15:56:36 - INFO - codeparrot_training - Step 29610: {'lr': 0.00019147956789128352, 'samples': 15160832, 'steps': 29610, 'loss/train': 2.444650173187256} 02/25/2022 15:56:39 - INFO - codeparrot_training - Step 29611: {'lr': 0.00019146366015444126, 'samples': 15161344, 'steps': 29611, 'loss/train': 1.49068284034729} 02/25/2022 15:56:45 - INFO - codeparrot_training - Step 29612: {'lr': 0.00019144775266835012, 'samples': 15161856, 'steps': 29612, 'loss/train': 1.059166431427002} 02/25/2022 15:56:48 - INFO - codeparrot_training - Step 29613: {'lr': 0.00019143184543307833, 'samples': 15162368, 'steps': 29613, 'loss/train': 1.8150075674057007} 02/25/2022 15:56:54 - INFO - codeparrot_training - Step 29614: {'lr': 0.00019141593844869376, 'samples': 15162880, 'steps': 29614, 'loss/train': 1.6683028936386108} 02/25/2022 15:56:57 - INFO - codeparrot_training - Step 29615: {'lr': 0.00019140003171526474, 'samples': 15163392, 'steps': 29615, 'loss/train': 1.4695719480514526} 02/25/2022 15:57:03 - INFO - codeparrot_training - Step 29616: {'lr': 0.00019138412523285936, 'samples': 15163904, 'steps': 29616, 'loss/train': 1.930280327796936} 02/25/2022 15:57:06 - INFO - codeparrot_training - Step 29617: {'lr': 0.0001913682190015458, 'samples': 15164416, 'steps': 29617, 'loss/train': 2.031759738922119} 02/25/2022 15:57:12 - INFO - codeparrot_training - Step 29618: {'lr': 0.00019135231302139227, 'samples': 15164928, 'steps': 29618, 'loss/train': 0.242783322930336} 02/25/2022 15:57:15 - INFO - codeparrot_training - Step 29619: {'lr': 0.00019133640729246672, 'samples': 15165440, 'steps': 29619, 'loss/train': 1.2959227561950684} 02/25/2022 15:57:21 - INFO - codeparrot_training - Step 29620: {'lr': 0.00019132050181483736, 'samples': 15165952, 'steps': 29620, 'loss/train': 1.4619669914245605} 02/25/2022 15:57:26 - INFO - codeparrot_training - Step 29621: {'lr': 0.0001913045965885723, 'samples': 15166464, 'steps': 29621, 'loss/train': 2.0575144290924072} 02/25/2022 15:57:30 - INFO - codeparrot_training - Step 29622: {'lr': 0.00019128869161373985, 'samples': 15166976, 'steps': 29622, 'loss/train': 1.9949980974197388} 02/25/2022 15:57:35 - INFO - codeparrot_training - Step 29623: {'lr': 0.0001912727868904079, 'samples': 15167488, 'steps': 29623, 'loss/train': 1.4256365299224854} 02/25/2022 15:57:39 - INFO - codeparrot_training - Step 29624: {'lr': 0.00019125688241864464, 'samples': 15168000, 'steps': 29624, 'loss/train': 1.6261653900146484} 02/25/2022 15:57:45 - INFO - codeparrot_training - Step 29625: {'lr': 0.00019124097819851835, 'samples': 15168512, 'steps': 29625, 'loss/train': 2.0156874656677246} 02/25/2022 15:57:48 - INFO - codeparrot_training - Step 29626: {'lr': 0.00019122507423009703, 'samples': 15169024, 'steps': 29626, 'loss/train': 1.8535398244857788} 02/25/2022 15:57:54 - INFO - codeparrot_training - Step 29627: {'lr': 0.00019120917051344878, 'samples': 15169536, 'steps': 29627, 'loss/train': 1.51189124584198} 02/25/2022 15:57:57 - INFO - codeparrot_training - Step 29628: {'lr': 0.00019119326704864176, 'samples': 15170048, 'steps': 29628, 'loss/train': 1.3984068632125854} 02/25/2022 15:58:03 - INFO - codeparrot_training - Step 29629: {'lr': 0.0001911773638357442, 'samples': 15170560, 'steps': 29629, 'loss/train': 1.2846662998199463} 02/25/2022 15:58:06 - INFO - codeparrot_training - Step 29630: {'lr': 0.00019116146087482403, 'samples': 15171072, 'steps': 29630, 'loss/train': 1.957324743270874} 02/25/2022 15:58:12 - INFO - codeparrot_training - Step 29631: {'lr': 0.00019114555816594956, 'samples': 15171584, 'steps': 29631, 'loss/train': 2.1340928077697754} 02/25/2022 15:58:15 - INFO - codeparrot_training - Step 29632: {'lr': 0.00019112965570918883, 'samples': 15172096, 'steps': 29632, 'loss/train': 2.267246723175049} 02/25/2022 15:58:22 - INFO - codeparrot_training - Step 29633: {'lr': 0.00019111375350460991, 'samples': 15172608, 'steps': 29633, 'loss/train': 2.070429801940918} 02/25/2022 15:58:25 - INFO - codeparrot_training - Step 29634: {'lr': 0.00019109785155228096, 'samples': 15173120, 'steps': 29634, 'loss/train': 0.7860600352287292} 02/25/2022 15:58:31 - INFO - codeparrot_training - Step 29635: {'lr': 0.00019108194985227017, 'samples': 15173632, 'steps': 29635, 'loss/train': 1.2916388511657715} 02/25/2022 15:58:34 - INFO - codeparrot_training - Step 29636: {'lr': 0.00019106604840464562, 'samples': 15174144, 'steps': 29636, 'loss/train': 1.569584608078003} 02/25/2022 15:58:40 - INFO - codeparrot_training - Step 29637: {'lr': 0.00019105014720947533, 'samples': 15174656, 'steps': 29637, 'loss/train': 1.7774972915649414} 02/25/2022 15:58:43 - INFO - codeparrot_training - Step 29638: {'lr': 0.00019103424626682746, 'samples': 15175168, 'steps': 29638, 'loss/train': 1.153586745262146} 02/25/2022 15:58:49 - INFO - codeparrot_training - Step 29639: {'lr': 0.00019101834557677034, 'samples': 15175680, 'steps': 29639, 'loss/train': 2.7185306549072266} 02/25/2022 15:58:52 - INFO - codeparrot_training - Step 29640: {'lr': 0.00019100244513937175, 'samples': 15176192, 'steps': 29640, 'loss/train': 2.2741262912750244} 02/25/2022 15:58:58 - INFO - codeparrot_training - Step 29641: {'lr': 0.00019098654495469991, 'samples': 15176704, 'steps': 29641, 'loss/train': 1.2731679677963257} 02/25/2022 15:59:01 - INFO - codeparrot_training - Step 29642: {'lr': 0.00019097064502282302, 'samples': 15177216, 'steps': 29642, 'loss/train': 1.6074756383895874} 02/25/2022 15:59:07 - INFO - codeparrot_training - Step 29643: {'lr': 0.00019095474534380912, 'samples': 15177728, 'steps': 29643, 'loss/train': 1.6521459817886353} 02/25/2022 15:59:10 - INFO - codeparrot_training - Step 29644: {'lr': 0.00019093884591772647, 'samples': 15178240, 'steps': 29644, 'loss/train': 1.7591832876205444} 02/25/2022 15:59:17 - INFO - codeparrot_training - Step 29645: {'lr': 0.00019092294674464294, 'samples': 15178752, 'steps': 29645, 'loss/train': 1.0980162620544434} 02/25/2022 15:59:20 - INFO - codeparrot_training - Step 29646: {'lr': 0.00019090704782462675, 'samples': 15179264, 'steps': 29646, 'loss/train': 2.2975497245788574} 02/25/2022 15:59:24 - INFO - codeparrot_training - Step 29647: {'lr': 0.000190891149157746, 'samples': 15179776, 'steps': 29647, 'loss/train': 1.4089785814285278} 02/25/2022 15:59:29 - INFO - codeparrot_training - Step 29648: {'lr': 0.0001908752507440689, 'samples': 15180288, 'steps': 29648, 'loss/train': 1.8052784204483032} 02/25/2022 15:59:32 - INFO - codeparrot_training - Step 29649: {'lr': 0.00019085935258366334, 'samples': 15180800, 'steps': 29649, 'loss/train': 1.8595654964447021} 02/25/2022 15:59:38 - INFO - codeparrot_training - Step 29650: {'lr': 0.00019084345467659754, 'samples': 15181312, 'steps': 29650, 'loss/train': 1.6302999258041382} 02/25/2022 15:59:44 - INFO - codeparrot_training - Step 29651: {'lr': 0.00019082755702293957, 'samples': 15181824, 'steps': 29651, 'loss/train': 1.4600595235824585} 02/25/2022 15:59:47 - INFO - codeparrot_training - Step 29652: {'lr': 0.00019081165962275771, 'samples': 15182336, 'steps': 29652, 'loss/train': 1.6980838775634766} 02/25/2022 15:59:52 - INFO - codeparrot_training - Step 29653: {'lr': 0.0001907957624761198, 'samples': 15182848, 'steps': 29653, 'loss/train': 1.9229475259780884} 02/25/2022 15:59:56 - INFO - codeparrot_training - Step 29654: {'lr': 0.00019077986558309402, 'samples': 15183360, 'steps': 29654, 'loss/train': 1.6512038707733154} 02/25/2022 16:00:02 - INFO - codeparrot_training - Step 29655: {'lr': 0.00019076396894374847, 'samples': 15183872, 'steps': 29655, 'loss/train': 1.4573386907577515} 02/25/2022 16:00:05 - INFO - codeparrot_training - Step 29656: {'lr': 0.0001907480725581513, 'samples': 15184384, 'steps': 29656, 'loss/train': 1.6344777345657349} 02/25/2022 16:00:11 - INFO - codeparrot_training - Step 29657: {'lr': 0.00019073217642637068, 'samples': 15184896, 'steps': 29657, 'loss/train': 1.3940215110778809} 02/25/2022 16:00:14 - INFO - codeparrot_training - Step 29658: {'lr': 0.00019071628054847446, 'samples': 15185408, 'steps': 29658, 'loss/train': 1.614280343055725} 02/25/2022 16:00:20 - INFO - codeparrot_training - Step 29659: {'lr': 0.0001907003849245309, 'samples': 15185920, 'steps': 29659, 'loss/train': 1.55470871925354} 02/25/2022 16:00:23 - INFO - codeparrot_training - Step 29660: {'lr': 0.00019068448955460805, 'samples': 15186432, 'steps': 29660, 'loss/train': 1.646599531173706} 02/25/2022 16:00:29 - INFO - codeparrot_training - Step 29661: {'lr': 0.0001906685944387741, 'samples': 15186944, 'steps': 29661, 'loss/train': 2.3054816722869873} 02/25/2022 16:00:32 - INFO - codeparrot_training - Step 29662: {'lr': 0.00019065269957709697, 'samples': 15187456, 'steps': 29662, 'loss/train': 0.7035332918167114} 02/25/2022 16:00:38 - INFO - codeparrot_training - Step 29663: {'lr': 0.00019063680496964484, 'samples': 15187968, 'steps': 29663, 'loss/train': 2.4976718425750732} 02/25/2022 16:00:41 - INFO - codeparrot_training - Step 29664: {'lr': 0.00019062091061648577, 'samples': 15188480, 'steps': 29664, 'loss/train': 1.5096818208694458} 02/25/2022 16:00:47 - INFO - codeparrot_training - Step 29665: {'lr': 0.00019060501651768796, 'samples': 15188992, 'steps': 29665, 'loss/train': 2.1369731426239014} 02/25/2022 16:00:50 - INFO - codeparrot_training - Step 29666: {'lr': 0.0001905891226733193, 'samples': 15189504, 'steps': 29666, 'loss/train': 0.33467504382133484} 02/25/2022 16:00:56 - INFO - codeparrot_training - Step 29667: {'lr': 0.000190573229083448, 'samples': 15190016, 'steps': 29667, 'loss/train': 1.9126348495483398} 02/25/2022 16:01:00 - INFO - codeparrot_training - Step 29668: {'lr': 0.00019055733574814207, 'samples': 15190528, 'steps': 29668, 'loss/train': 0.6411345601081848} 02/25/2022 16:01:05 - INFO - codeparrot_training - Step 29669: {'lr': 0.00019054144266746973, 'samples': 15191040, 'steps': 29669, 'loss/train': 2.2083663940429688} 02/25/2022 16:01:09 - INFO - codeparrot_training - Step 29670: {'lr': 0.00019052554984149895, 'samples': 15191552, 'steps': 29670, 'loss/train': 2.313786268234253} 02/25/2022 16:01:12 - INFO - codeparrot_training - Step 29671: {'lr': 0.0001905096572702978, 'samples': 15192064, 'steps': 29671, 'loss/train': 1.0212492942810059} 02/25/2022 16:01:18 - INFO - codeparrot_training - Step 29672: {'lr': 0.0001904937649539344, 'samples': 15192576, 'steps': 29672, 'loss/train': 2.3255090713500977} 02/25/2022 16:01:22 - INFO - codeparrot_training - Step 29673: {'lr': 0.0001904778728924768, 'samples': 15193088, 'steps': 29673, 'loss/train': 1.4593840837478638} 02/25/2022 16:01:27 - INFO - codeparrot_training - Step 29674: {'lr': 0.0001904619810859931, 'samples': 15193600, 'steps': 29674, 'loss/train': 2.344268321990967} 02/25/2022 16:01:31 - INFO - codeparrot_training - Step 29675: {'lr': 0.00019044608953455136, 'samples': 15194112, 'steps': 29675, 'loss/train': 1.2264584302902222} 02/25/2022 16:01:36 - INFO - codeparrot_training - Step 29676: {'lr': 0.0001904301982382197, 'samples': 15194624, 'steps': 29676, 'loss/train': 1.537766695022583} 02/25/2022 16:01:40 - INFO - codeparrot_training - Step 29677: {'lr': 0.0001904143071970661, 'samples': 15195136, 'steps': 29677, 'loss/train': 2.050400972366333} 02/25/2022 16:01:45 - INFO - codeparrot_training - Step 29678: {'lr': 0.0001903984164111587, 'samples': 15195648, 'steps': 29678, 'loss/train': 0.5048926472663879} 02/25/2022 16:01:49 - INFO - codeparrot_training - Step 29679: {'lr': 0.00019038252588056558, 'samples': 15196160, 'steps': 29679, 'loss/train': 2.1396775245666504} 02/25/2022 16:01:54 - INFO - codeparrot_training - Step 29680: {'lr': 0.00019036663560535483, 'samples': 15196672, 'steps': 29680, 'loss/train': 1.6611448526382446} 02/25/2022 16:02:01 - INFO - codeparrot_training - Step 29681: {'lr': 0.0001903507455855944, 'samples': 15197184, 'steps': 29681, 'loss/train': 1.4303745031356812} 02/25/2022 16:02:04 - INFO - codeparrot_training - Step 29682: {'lr': 0.00019033485582135244, 'samples': 15197696, 'steps': 29682, 'loss/train': 1.0385608673095703} 02/25/2022 16:02:09 - INFO - codeparrot_training - Step 29683: {'lr': 0.00019031896631269713, 'samples': 15198208, 'steps': 29683, 'loss/train': 1.137509822845459} 02/25/2022 16:02:13 - INFO - codeparrot_training - Step 29684: {'lr': 0.00019030307705969628, 'samples': 15198720, 'steps': 29684, 'loss/train': 0.40029680728912354} 02/25/2022 16:02:18 - INFO - codeparrot_training - Step 29685: {'lr': 0.00019028718806241813, 'samples': 15199232, 'steps': 29685, 'loss/train': 1.3261994123458862} 02/25/2022 16:02:22 - INFO - codeparrot_training - Step 29686: {'lr': 0.00019027129932093067, 'samples': 15199744, 'steps': 29686, 'loss/train': 1.7638357877731323} 02/25/2022 16:02:28 - INFO - codeparrot_training - Step 29687: {'lr': 0.00019025541083530213, 'samples': 15200256, 'steps': 29687, 'loss/train': 2.1726949214935303} 02/25/2022 16:02:31 - INFO - codeparrot_training - Step 29688: {'lr': 0.00019023952260560032, 'samples': 15200768, 'steps': 29688, 'loss/train': 1.969434142112732} 02/25/2022 16:02:37 - INFO - codeparrot_training - Step 29689: {'lr': 0.0001902236346318934, 'samples': 15201280, 'steps': 29689, 'loss/train': 1.3506629467010498} 02/25/2022 16:02:40 - INFO - codeparrot_training - Step 29690: {'lr': 0.00019020774691424946, 'samples': 15201792, 'steps': 29690, 'loss/train': 1.8096134662628174} 02/25/2022 16:02:46 - INFO - codeparrot_training - Step 29691: {'lr': 0.00019019185945273655, 'samples': 15202304, 'steps': 29691, 'loss/train': 1.6162241697311401} 02/25/2022 16:02:50 - INFO - codeparrot_training - Step 29692: {'lr': 0.00019017597224742287, 'samples': 15202816, 'steps': 29692, 'loss/train': 1.4075875282287598} 02/25/2022 16:02:55 - INFO - codeparrot_training - Step 29693: {'lr': 0.00019016008529837617, 'samples': 15203328, 'steps': 29693, 'loss/train': 1.740389347076416} 02/25/2022 16:02:59 - INFO - codeparrot_training - Step 29694: {'lr': 0.00019014419860566468, 'samples': 15203840, 'steps': 29694, 'loss/train': 1.511661410331726} 02/25/2022 16:03:04 - INFO - codeparrot_training - Step 29695: {'lr': 0.00019012831216935644, 'samples': 15204352, 'steps': 29695, 'loss/train': 0.8117575645446777} 02/25/2022 16:03:08 - INFO - codeparrot_training - Step 29696: {'lr': 0.0001901124259895196, 'samples': 15204864, 'steps': 29696, 'loss/train': 2.414487600326538} 02/25/2022 16:03:13 - INFO - codeparrot_training - Step 29697: {'lr': 0.000190096540066222, 'samples': 15205376, 'steps': 29697, 'loss/train': 1.35346257686615} 02/25/2022 16:03:17 - INFO - codeparrot_training - Step 29698: {'lr': 0.00019008065439953182, 'samples': 15205888, 'steps': 29698, 'loss/train': 8.438970565795898} 02/25/2022 16:03:22 - INFO - codeparrot_training - Step 29699: {'lr': 0.00019006476898951707, 'samples': 15206400, 'steps': 29699, 'loss/train': 2.0487515926361084} 02/25/2022 16:03:26 - INFO - codeparrot_training - Step 29700: {'lr': 0.00019004888383624595, 'samples': 15206912, 'steps': 29700, 'loss/train': 1.7922396659851074} 02/25/2022 16:03:32 - INFO - codeparrot_training - Step 29701: {'lr': 0.00019003299893978622, 'samples': 15207424, 'steps': 29701, 'loss/train': 1.2144306898117065} 02/25/2022 16:03:35 - INFO - codeparrot_training - Step 29702: {'lr': 0.00019001711430020613, 'samples': 15207936, 'steps': 29702, 'loss/train': 1.787750244140625} 02/25/2022 16:03:41 - INFO - codeparrot_training - Step 29703: {'lr': 0.00019000122991757367, 'samples': 15208448, 'steps': 29703, 'loss/train': 1.7397289276123047} 02/25/2022 16:03:45 - INFO - codeparrot_training - Step 29704: {'lr': 0.00018998534579195685, 'samples': 15208960, 'steps': 29704, 'loss/train': 2.310115098953247} 02/25/2022 16:03:48 - INFO - codeparrot_training - Step 29705: {'lr': 0.0001899694619234239, 'samples': 15209472, 'steps': 29705, 'loss/train': 2.9601659774780273} 02/25/2022 16:03:54 - INFO - codeparrot_training - Step 29706: {'lr': 0.00018995357831204257, 'samples': 15209984, 'steps': 29706, 'loss/train': 1.3097292184829712} 02/25/2022 16:03:57 - INFO - codeparrot_training - Step 29707: {'lr': 0.00018993769495788109, 'samples': 15210496, 'steps': 29707, 'loss/train': 1.6220203638076782} 02/25/2022 16:04:02 - INFO - codeparrot_training - Step 29708: {'lr': 0.00018992181186100744, 'samples': 15211008, 'steps': 29708, 'loss/train': 1.0518784523010254} 02/25/2022 16:04:06 - INFO - codeparrot_training - Step 29709: {'lr': 0.00018990592902148973, 'samples': 15211520, 'steps': 29709, 'loss/train': 1.4611061811447144} 02/25/2022 16:04:11 - INFO - codeparrot_training - Step 29710: {'lr': 0.00018989004643939588, 'samples': 15212032, 'steps': 29710, 'loss/train': 1.481844425201416} 02/25/2022 16:04:15 - INFO - codeparrot_training - Step 29711: {'lr': 0.00018987416411479397, 'samples': 15212544, 'steps': 29711, 'loss/train': 2.341019630432129} 02/25/2022 16:04:20 - INFO - codeparrot_training - Step 29712: {'lr': 0.00018985828204775204, 'samples': 15213056, 'steps': 29712, 'loss/train': 2.197713613510132} 02/25/2022 16:04:24 - INFO - codeparrot_training - Step 29713: {'lr': 0.00018984240023833826, 'samples': 15213568, 'steps': 29713, 'loss/train': 1.190040946006775} 02/25/2022 16:04:30 - INFO - codeparrot_training - Step 29714: {'lr': 0.00018982651868662044, 'samples': 15214080, 'steps': 29714, 'loss/train': 1.5683180093765259} 02/25/2022 16:04:33 - INFO - codeparrot_training - Step 29715: {'lr': 0.00018981063739266667, 'samples': 15214592, 'steps': 29715, 'loss/train': 1.4655780792236328} 02/25/2022 16:04:39 - INFO - codeparrot_training - Step 29716: {'lr': 0.0001897947563565451, 'samples': 15215104, 'steps': 29716, 'loss/train': 2.7867512702941895} 02/25/2022 16:04:42 - INFO - codeparrot_training - Step 29717: {'lr': 0.00018977887557832356, 'samples': 15215616, 'steps': 29717, 'loss/train': 0.9705775380134583} 02/25/2022 16:04:48 - INFO - codeparrot_training - Step 29718: {'lr': 0.00018976299505807031, 'samples': 15216128, 'steps': 29718, 'loss/train': 2.2573275566101074} 02/25/2022 16:04:54 - INFO - codeparrot_training - Step 29719: {'lr': 0.00018974711479585323, 'samples': 15216640, 'steps': 29719, 'loss/train': 1.5685479640960693} 02/25/2022 16:04:57 - INFO - codeparrot_training - Step 29720: {'lr': 0.00018973123479174036, 'samples': 15217152, 'steps': 29720, 'loss/train': 1.150031328201294} 02/25/2022 16:05:03 - INFO - codeparrot_training - Step 29721: {'lr': 0.00018971535504579972, 'samples': 15217664, 'steps': 29721, 'loss/train': 2.5471959114074707} 02/25/2022 16:05:06 - INFO - codeparrot_training - Step 29722: {'lr': 0.0001896994755580994, 'samples': 15218176, 'steps': 29722, 'loss/train': 2.855038642883301} 02/25/2022 16:05:12 - INFO - codeparrot_training - Step 29723: {'lr': 0.0001896835963287074, 'samples': 15218688, 'steps': 29723, 'loss/train': 1.5021971464157104} 02/25/2022 16:05:15 - INFO - codeparrot_training - Step 29724: {'lr': 0.00018966771735769167, 'samples': 15219200, 'steps': 29724, 'loss/train': 2.3879992961883545} 02/25/2022 16:05:21 - INFO - codeparrot_training - Step 29725: {'lr': 0.00018965183864512027, 'samples': 15219712, 'steps': 29725, 'loss/train': 2.1341824531555176} 02/25/2022 16:05:24 - INFO - codeparrot_training - Step 29726: {'lr': 0.00018963596019106125, 'samples': 15220224, 'steps': 29726, 'loss/train': 1.0678937435150146} 02/25/2022 16:05:31 - INFO - codeparrot_training - Step 29727: {'lr': 0.0001896200819955827, 'samples': 15220736, 'steps': 29727, 'loss/train': 2.567450761795044} 02/25/2022 16:05:34 - INFO - codeparrot_training - Step 29728: {'lr': 0.00018960420405875244, 'samples': 15221248, 'steps': 29728, 'loss/train': 1.4132837057113647} 02/25/2022 16:05:40 - INFO - codeparrot_training - Step 29729: {'lr': 0.00018958832638063858, 'samples': 15221760, 'steps': 29729, 'loss/train': 0.6862665414810181} 02/25/2022 16:05:43 - INFO - codeparrot_training - Step 29730: {'lr': 0.00018957244896130915, 'samples': 15222272, 'steps': 29730, 'loss/train': 1.5507272481918335} 02/25/2022 16:05:49 - INFO - codeparrot_training - Step 29731: {'lr': 0.00018955657180083228, 'samples': 15222784, 'steps': 29731, 'loss/train': 1.1663380861282349} 02/25/2022 16:05:52 - INFO - codeparrot_training - Step 29732: {'lr': 0.00018954069489927574, 'samples': 15223296, 'steps': 29732, 'loss/train': 0.8850343823432922} 02/25/2022 16:05:58 - INFO - codeparrot_training - Step 29733: {'lr': 0.0001895248182567077, 'samples': 15223808, 'steps': 29733, 'loss/train': 1.8836568593978882} 02/25/2022 16:06:01 - INFO - codeparrot_training - Step 29734: {'lr': 0.0001895089418731961, 'samples': 15224320, 'steps': 29734, 'loss/train': 0.9981403946876526} 02/25/2022 16:06:07 - INFO - codeparrot_training - Step 29735: {'lr': 0.0001894930657488091, 'samples': 15224832, 'steps': 29735, 'loss/train': 2.9812047481536865} 02/25/2022 16:06:10 - INFO - codeparrot_training - Step 29736: {'lr': 0.0001894771898836145, 'samples': 15225344, 'steps': 29736, 'loss/train': 0.3940974175930023} 02/25/2022 16:06:17 - INFO - codeparrot_training - Step 29737: {'lr': 0.00018946131427768039, 'samples': 15225856, 'steps': 29737, 'loss/train': 1.7756891250610352} 02/25/2022 16:06:21 - INFO - codeparrot_training - Step 29738: {'lr': 0.00018944543893107478, 'samples': 15226368, 'steps': 29738, 'loss/train': 1.3673560619354248} 02/25/2022 16:06:26 - INFO - codeparrot_training - Step 29739: {'lr': 0.00018942956384386566, 'samples': 15226880, 'steps': 29739, 'loss/train': 2.3634588718414307} 02/25/2022 16:06:30 - INFO - codeparrot_training - Step 29740: {'lr': 0.00018941368901612121, 'samples': 15227392, 'steps': 29740, 'loss/train': 1.5838547945022583} 02/25/2022 16:06:35 - INFO - codeparrot_training - Step 29741: {'lr': 0.00018939781444790916, 'samples': 15227904, 'steps': 29741, 'loss/train': 2.242109775543213} 02/25/2022 16:06:39 - INFO - codeparrot_training - Step 29742: {'lr': 0.0001893819401392976, 'samples': 15228416, 'steps': 29742, 'loss/train': 0.8927356004714966} 02/25/2022 16:06:44 - INFO - codeparrot_training - Step 29743: {'lr': 0.0001893660660903546, 'samples': 15228928, 'steps': 29743, 'loss/train': 1.2056217193603516} 02/25/2022 16:06:48 - INFO - codeparrot_training - Step 29744: {'lr': 0.0001893501923011482, 'samples': 15229440, 'steps': 29744, 'loss/train': 1.2550374269485474} 02/25/2022 16:06:53 - INFO - codeparrot_training - Step 29745: {'lr': 0.00018933431877174624, 'samples': 15229952, 'steps': 29745, 'loss/train': 1.7424426078796387} 02/25/2022 16:06:57 - INFO - codeparrot_training - Step 29746: {'lr': 0.00018931844550221675, 'samples': 15230464, 'steps': 29746, 'loss/train': 0.5975298285484314} 02/25/2022 16:07:03 - INFO - codeparrot_training - Step 29747: {'lr': 0.00018930257249262776, 'samples': 15230976, 'steps': 29747, 'loss/train': 0.07923591136932373} 02/25/2022 16:07:06 - INFO - codeparrot_training - Step 29748: {'lr': 0.00018928669974304746, 'samples': 15231488, 'steps': 29748, 'loss/train': 1.3820666074752808} 02/25/2022 16:07:12 - INFO - codeparrot_training - Step 29749: {'lr': 0.0001892708272535435, 'samples': 15232000, 'steps': 29749, 'loss/train': 1.4223815202713013} 02/25/2022 16:07:15 - INFO - codeparrot_training - Step 29750: {'lr': 0.00018925495502418406, 'samples': 15232512, 'steps': 29750, 'loss/train': 1.9621717929840088} 02/25/2022 16:07:21 - INFO - codeparrot_training - Step 29751: {'lr': 0.00018923908305503706, 'samples': 15233024, 'steps': 29751, 'loss/train': 2.3909809589385986} 02/25/2022 16:07:24 - INFO - codeparrot_training - Step 29752: {'lr': 0.00018922321134617055, 'samples': 15233536, 'steps': 29752, 'loss/train': 1.416527271270752} 02/25/2022 16:07:30 - INFO - codeparrot_training - Step 29753: {'lr': 0.00018920733989765263, 'samples': 15234048, 'steps': 29753, 'loss/train': 1.3958063125610352} 02/25/2022 16:07:33 - INFO - codeparrot_training - Step 29754: {'lr': 0.00018919146870955103, 'samples': 15234560, 'steps': 29754, 'loss/train': 1.333345651626587} 02/25/2022 16:07:39 - INFO - codeparrot_training - Step 29755: {'lr': 0.00018917559778193387, 'samples': 15235072, 'steps': 29755, 'loss/train': 1.0761003494262695} 02/25/2022 16:07:42 - INFO - codeparrot_training - Step 29756: {'lr': 0.00018915972711486923, 'samples': 15235584, 'steps': 29756, 'loss/train': 1.9296795129776} 02/25/2022 16:07:48 - INFO - codeparrot_training - Step 29757: {'lr': 0.000189143856708425, 'samples': 15236096, 'steps': 29757, 'loss/train': 1.7058906555175781} 02/25/2022 16:07:51 - INFO - codeparrot_training - Step 29758: {'lr': 0.00018912798656266905, 'samples': 15236608, 'steps': 29758, 'loss/train': 1.8765857219696045} 02/25/2022 16:07:57 - INFO - codeparrot_training - Step 29759: {'lr': 0.00018911211667766952, 'samples': 15237120, 'steps': 29759, 'loss/train': 1.8540867567062378} 02/25/2022 16:08:02 - INFO - codeparrot_training - Step 29760: {'lr': 0.00018909624705349437, 'samples': 15237632, 'steps': 29760, 'loss/train': 1.778419017791748} 02/25/2022 16:08:06 - INFO - codeparrot_training - Step 29761: {'lr': 0.0001890803776902116, 'samples': 15238144, 'steps': 29761, 'loss/train': 1.5091160535812378} 02/25/2022 16:08:11 - INFO - codeparrot_training - Step 29762: {'lr': 0.00018906450858788908, 'samples': 15238656, 'steps': 29762, 'loss/train': 1.6516075134277344} 02/25/2022 16:08:15 - INFO - codeparrot_training - Step 29763: {'lr': 0.0001890486397465949, 'samples': 15239168, 'steps': 29763, 'loss/train': 1.1941330432891846} 02/25/2022 16:08:21 - INFO - codeparrot_training - Step 29764: {'lr': 0.00018903277116639694, 'samples': 15239680, 'steps': 29764, 'loss/train': 1.5156298875808716} 02/25/2022 16:08:24 - INFO - codeparrot_training - Step 29765: {'lr': 0.00018901690284736322, 'samples': 15240192, 'steps': 29765, 'loss/train': 1.579389214515686} 02/25/2022 16:08:30 - INFO - codeparrot_training - Step 29766: {'lr': 0.0001890010347895618, 'samples': 15240704, 'steps': 29766, 'loss/train': 2.8881235122680664} 02/25/2022 16:08:33 - INFO - codeparrot_training - Step 29767: {'lr': 0.00018898516699306057, 'samples': 15241216, 'steps': 29767, 'loss/train': 1.1252330541610718} 02/25/2022 16:08:39 - INFO - codeparrot_training - Step 29768: {'lr': 0.00018896929945792746, 'samples': 15241728, 'steps': 29768, 'loss/train': 0.9265627861022949} 02/25/2022 16:08:43 - INFO - codeparrot_training - Step 29769: {'lr': 0.00018895343218423047, 'samples': 15242240, 'steps': 29769, 'loss/train': 0.06501796841621399} 02/25/2022 16:08:48 - INFO - codeparrot_training - Step 29770: {'lr': 0.00018893756517203772, 'samples': 15242752, 'steps': 29770, 'loss/train': 1.299767017364502} 02/25/2022 16:08:51 - INFO - codeparrot_training - Step 29771: {'lr': 0.00018892169842141694, 'samples': 15243264, 'steps': 29771, 'loss/train': 3.187598943710327} 02/25/2022 16:08:57 - INFO - codeparrot_training - Step 29772: {'lr': 0.00018890583193243622, 'samples': 15243776, 'steps': 29772, 'loss/train': 0.026948869228363037} 02/25/2022 16:09:00 - INFO - codeparrot_training - Step 29773: {'lr': 0.0001888899657051635, 'samples': 15244288, 'steps': 29773, 'loss/train': 2.140406370162964} 02/25/2022 16:09:07 - INFO - codeparrot_training - Step 29774: {'lr': 0.00018887409973966686, 'samples': 15244800, 'steps': 29774, 'loss/train': 1.4617607593536377} 02/25/2022 16:09:10 - INFO - codeparrot_training - Step 29775: {'lr': 0.0001888582340360141, 'samples': 15245312, 'steps': 29775, 'loss/train': 1.877639889717102} 02/25/2022 16:09:16 - INFO - codeparrot_training - Step 29776: {'lr': 0.0001888423685942732, 'samples': 15245824, 'steps': 29776, 'loss/train': 1.2678022384643555} 02/25/2022 16:09:19 - INFO - codeparrot_training - Step 29777: {'lr': 0.0001888265034145122, 'samples': 15246336, 'steps': 29777, 'loss/train': 1.7724556922912598} 02/25/2022 16:09:25 - INFO - codeparrot_training - Step 29778: {'lr': 0.00018881063849679903, 'samples': 15246848, 'steps': 29778, 'loss/train': 2.483582019805908} 02/25/2022 16:09:28 - INFO - codeparrot_training - Step 29779: {'lr': 0.00018879477384120178, 'samples': 15247360, 'steps': 29779, 'loss/train': 1.9547593593597412} 02/25/2022 16:09:34 - INFO - codeparrot_training - Step 29780: {'lr': 0.00018877890944778814, 'samples': 15247872, 'steps': 29780, 'loss/train': 2.127516269683838} 02/25/2022 16:09:37 - INFO - codeparrot_training - Step 29781: {'lr': 0.00018876304531662626, 'samples': 15248384, 'steps': 29781, 'loss/train': 1.4735227823257446} 02/25/2022 16:09:43 - INFO - codeparrot_training - Step 29782: {'lr': 0.00018874718144778403, 'samples': 15248896, 'steps': 29782, 'loss/train': 8.687186241149902} 02/25/2022 16:09:46 - INFO - codeparrot_training - Step 29783: {'lr': 0.00018873131784132952, 'samples': 15249408, 'steps': 29783, 'loss/train': 2.3048532009124756} 02/25/2022 16:09:52 - INFO - codeparrot_training - Step 29784: {'lr': 0.00018871545449733048, 'samples': 15249920, 'steps': 29784, 'loss/train': 1.9706388711929321} 02/25/2022 16:09:56 - INFO - codeparrot_training - Step 29785: {'lr': 0.00018869959141585497, 'samples': 15250432, 'steps': 29785, 'loss/train': 2.3992395401000977} 02/25/2022 16:10:01 - INFO - codeparrot_training - Step 29786: {'lr': 0.00018868372859697096, 'samples': 15250944, 'steps': 29786, 'loss/train': 2.2633328437805176} 02/25/2022 16:10:05 - INFO - codeparrot_training - Step 29787: {'lr': 0.00018866786604074636, 'samples': 15251456, 'steps': 29787, 'loss/train': 2.0165457725524902} 02/25/2022 16:10:10 - INFO - codeparrot_training - Step 29788: {'lr': 0.0001886520037472493, 'samples': 15251968, 'steps': 29788, 'loss/train': 1.6836521625518799} 02/25/2022 16:10:14 - INFO - codeparrot_training - Step 29789: {'lr': 0.00018863614171654743, 'samples': 15252480, 'steps': 29789, 'loss/train': 3.434588670730591} 02/25/2022 16:10:19 - INFO - codeparrot_training - Step 29790: {'lr': 0.00018862027994870887, 'samples': 15252992, 'steps': 29790, 'loss/train': 1.123921275138855} 02/25/2022 16:10:23 - INFO - codeparrot_training - Step 29791: {'lr': 0.0001886044184438015, 'samples': 15253504, 'steps': 29791, 'loss/train': 1.8841543197631836} 02/25/2022 16:10:28 - INFO - codeparrot_training - Step 29792: {'lr': 0.00018858855720189346, 'samples': 15254016, 'steps': 29792, 'loss/train': 2.458669900894165} 02/25/2022 16:10:32 - INFO - codeparrot_training - Step 29793: {'lr': 0.00018857269622305243, 'samples': 15254528, 'steps': 29793, 'loss/train': 0.5346103310585022} 02/25/2022 16:10:38 - INFO - codeparrot_training - Step 29794: {'lr': 0.00018855683550734643, 'samples': 15255040, 'steps': 29794, 'loss/train': 1.3085296154022217} 02/25/2022 16:10:41 - INFO - codeparrot_training - Step 29795: {'lr': 0.00018854097505484346, 'samples': 15255552, 'steps': 29795, 'loss/train': 1.4604032039642334} 02/25/2022 16:10:47 - INFO - codeparrot_training - Step 29796: {'lr': 0.00018852511486561156, 'samples': 15256064, 'steps': 29796, 'loss/train': 1.089059591293335} 02/25/2022 16:10:50 - INFO - codeparrot_training - Step 29797: {'lr': 0.0001885092549397184, 'samples': 15256576, 'steps': 29797, 'loss/train': 1.461957573890686} 02/25/2022 16:10:56 - INFO - codeparrot_training - Step 29798: {'lr': 0.0001884933952772321, 'samples': 15257088, 'steps': 29798, 'loss/train': 1.9711018800735474} 02/25/2022 16:10:59 - INFO - codeparrot_training - Step 29799: {'lr': 0.00018847753587822053, 'samples': 15257600, 'steps': 29799, 'loss/train': 1.9362515211105347} 02/25/2022 16:11:05 - INFO - codeparrot_training - Step 29800: {'lr': 0.00018846167674275175, 'samples': 15258112, 'steps': 29800, 'loss/train': 1.3598463535308838} 02/25/2022 16:11:08 - INFO - codeparrot_training - Step 29801: {'lr': 0.0001884458178708936, 'samples': 15258624, 'steps': 29801, 'loss/train': 1.8957009315490723} 02/25/2022 16:11:14 - INFO - codeparrot_training - Step 29802: {'lr': 0.00018842995926271393, 'samples': 15259136, 'steps': 29802, 'loss/train': 2.2282369136810303} 02/25/2022 16:11:17 - INFO - codeparrot_training - Step 29803: {'lr': 0.00018841410091828087, 'samples': 15259648, 'steps': 29803, 'loss/train': 1.5105842351913452} 02/25/2022 16:11:23 - INFO - codeparrot_training - Step 29804: {'lr': 0.00018839824283766216, 'samples': 15260160, 'steps': 29804, 'loss/train': 1.987215280532837} 02/25/2022 16:11:26 - INFO - codeparrot_training - Step 29805: {'lr': 0.00018838238502092585, 'samples': 15260672, 'steps': 29805, 'loss/train': 1.5523611307144165} 02/25/2022 16:11:32 - INFO - codeparrot_training - Step 29806: {'lr': 0.00018836652746813984, 'samples': 15261184, 'steps': 29806, 'loss/train': 1.0876392126083374} 02/25/2022 16:11:35 - INFO - codeparrot_training - Step 29807: {'lr': 0.00018835067017937206, 'samples': 15261696, 'steps': 29807, 'loss/train': 2.51267409324646} 02/25/2022 16:11:41 - INFO - codeparrot_training - Step 29808: {'lr': 0.0001883348131546904, 'samples': 15262208, 'steps': 29808, 'loss/train': 1.921204686164856} 02/25/2022 16:11:44 - INFO - codeparrot_training - Step 29809: {'lr': 0.00018831895639416288, 'samples': 15262720, 'steps': 29809, 'loss/train': 2.3218698501586914} 02/25/2022 16:11:50 - INFO - codeparrot_training - Step 29810: {'lr': 0.00018830309989785736, 'samples': 15263232, 'steps': 29810, 'loss/train': 1.2432307004928589} 02/25/2022 16:11:54 - INFO - codeparrot_training - Step 29811: {'lr': 0.00018828724366584172, 'samples': 15263744, 'steps': 29811, 'loss/train': 1.2498656511306763} 02/25/2022 16:12:00 - INFO - codeparrot_training - Step 29812: {'lr': 0.00018827138769818391, 'samples': 15264256, 'steps': 29812, 'loss/train': 1.3791866302490234} 02/25/2022 16:12:03 - INFO - codeparrot_training - Step 29813: {'lr': 0.0001882555319949519, 'samples': 15264768, 'steps': 29813, 'loss/train': 1.654335379600525} 02/25/2022 16:12:09 - INFO - codeparrot_training - Step 29814: {'lr': 0.00018823967655621368, 'samples': 15265280, 'steps': 29814, 'loss/train': 0.4004649519920349} 02/25/2022 16:12:12 - INFO - codeparrot_training - Step 29815: {'lr': 0.000188223821382037, 'samples': 15265792, 'steps': 29815, 'loss/train': 1.8729444742202759} 02/25/2022 16:12:18 - INFO - codeparrot_training - Step 29816: {'lr': 0.00018820796647248982, 'samples': 15266304, 'steps': 29816, 'loss/train': 1.3981703519821167} 02/25/2022 16:12:21 - INFO - codeparrot_training - Step 29817: {'lr': 0.00018819211182764006, 'samples': 15266816, 'steps': 29817, 'loss/train': 3.234819173812866} 02/25/2022 16:12:27 - INFO - codeparrot_training - Step 29818: {'lr': 0.0001881762574475558, 'samples': 15267328, 'steps': 29818, 'loss/train': 1.0996278524398804} 02/25/2022 16:12:30 - INFO - codeparrot_training - Step 29819: {'lr': 0.00018816040333230472, 'samples': 15267840, 'steps': 29819, 'loss/train': 1.0675584077835083} 02/25/2022 16:12:37 - INFO - codeparrot_training - Step 29820: {'lr': 0.00018814454948195487, 'samples': 15268352, 'steps': 29820, 'loss/train': 2.1080851554870605} 02/25/2022 16:12:40 - INFO - codeparrot_training - Step 29821: {'lr': 0.0001881286958965741, 'samples': 15268864, 'steps': 29821, 'loss/train': 1.9768056869506836} 02/25/2022 16:12:45 - INFO - codeparrot_training - Step 29822: {'lr': 0.00018811284257623043, 'samples': 15269376, 'steps': 29822, 'loss/train': 1.400217056274414} 02/25/2022 16:12:49 - INFO - codeparrot_training - Step 29823: {'lr': 0.0001880969895209916, 'samples': 15269888, 'steps': 29823, 'loss/train': 0.8915282487869263} 02/25/2022 16:12:55 - INFO - codeparrot_training - Step 29824: {'lr': 0.00018808113673092564, 'samples': 15270400, 'steps': 29824, 'loss/train': 1.4705545902252197} 02/25/2022 16:12:58 - INFO - codeparrot_training - Step 29825: {'lr': 0.0001880652842061004, 'samples': 15270912, 'steps': 29825, 'loss/train': 1.9595308303833008} 02/25/2022 16:13:04 - INFO - codeparrot_training - Step 29826: {'lr': 0.0001880494319465838, 'samples': 15271424, 'steps': 29826, 'loss/train': 1.4071155786514282} 02/25/2022 16:13:07 - INFO - codeparrot_training - Step 29827: {'lr': 0.00018803357995244392, 'samples': 15271936, 'steps': 29827, 'loss/train': 1.5162628889083862} 02/25/2022 16:13:13 - INFO - codeparrot_training - Step 29828: {'lr': 0.00018801772822374835, 'samples': 15272448, 'steps': 29828, 'loss/train': 2.0283267498016357} 02/25/2022 16:13:16 - INFO - codeparrot_training - Step 29829: {'lr': 0.0001880018767605652, 'samples': 15272960, 'steps': 29829, 'loss/train': 1.5230984687805176} 02/25/2022 16:13:22 - INFO - codeparrot_training - Step 29830: {'lr': 0.00018798602556296227, 'samples': 15273472, 'steps': 29830, 'loss/train': 2.1076390743255615} 02/25/2022 16:13:26 - INFO - codeparrot_training - Step 29831: {'lr': 0.00018797017463100766, 'samples': 15273984, 'steps': 29831, 'loss/train': 2.198701858520508} 02/25/2022 16:13:31 - INFO - codeparrot_training - Step 29832: {'lr': 0.000187954323964769, 'samples': 15274496, 'steps': 29832, 'loss/train': 0.7351523637771606} 02/25/2022 16:13:35 - INFO - codeparrot_training - Step 29833: {'lr': 0.00018793847356431432, 'samples': 15275008, 'steps': 29833, 'loss/train': 2.197909355163574} 02/25/2022 16:13:40 - INFO - codeparrot_training - Step 29834: {'lr': 0.0001879226234297115, 'samples': 15275520, 'steps': 29834, 'loss/train': 1.4007171392440796} 02/25/2022 16:13:46 - INFO - codeparrot_training - Step 29835: {'lr': 0.00018790677356102847, 'samples': 15276032, 'steps': 29835, 'loss/train': 1.749891996383667} 02/25/2022 16:13:49 - INFO - codeparrot_training - Step 29836: {'lr': 0.0001878909239583332, 'samples': 15276544, 'steps': 29836, 'loss/train': 1.1606929302215576} 02/25/2022 16:13:55 - INFO - codeparrot_training - Step 29837: {'lr': 0.0001878750746216934, 'samples': 15277056, 'steps': 29837, 'loss/train': 2.2587380409240723} 02/25/2022 16:13:58 - INFO - codeparrot_training - Step 29838: {'lr': 0.000187859225551177, 'samples': 15277568, 'steps': 29838, 'loss/train': 2.2455410957336426} 02/25/2022 16:14:04 - INFO - codeparrot_training - Step 29839: {'lr': 0.00018784337674685201, 'samples': 15278080, 'steps': 29839, 'loss/train': 2.0633840560913086} 02/25/2022 16:14:08 - INFO - codeparrot_training - Step 29840: {'lr': 0.00018782752820878634, 'samples': 15278592, 'steps': 29840, 'loss/train': 1.5649887323379517} 02/25/2022 16:14:13 - INFO - codeparrot_training - Step 29841: {'lr': 0.0001878116799370477, 'samples': 15279104, 'steps': 29841, 'loss/train': 1.1138519048690796} 02/25/2022 16:14:17 - INFO - codeparrot_training - Step 29842: {'lr': 0.00018779583193170407, 'samples': 15279616, 'steps': 29842, 'loss/train': 1.6223907470703125} 02/25/2022 16:14:22 - INFO - codeparrot_training - Step 29843: {'lr': 0.00018777998419282338, 'samples': 15280128, 'steps': 29843, 'loss/train': 2.567366600036621} 02/25/2022 16:14:26 - INFO - codeparrot_training - Step 29844: {'lr': 0.0001877641367204735, 'samples': 15280640, 'steps': 29844, 'loss/train': 1.1580398082733154} 02/25/2022 16:14:31 - INFO - codeparrot_training - Step 29845: {'lr': 0.00018774828951472223, 'samples': 15281152, 'steps': 29845, 'loss/train': 2.1439363956451416} 02/25/2022 16:14:35 - INFO - codeparrot_training - Step 29846: {'lr': 0.00018773244257563754, 'samples': 15281664, 'steps': 29846, 'loss/train': 1.6922271251678467} 02/25/2022 16:14:40 - INFO - codeparrot_training - Step 29847: {'lr': 0.00018771659590328732, 'samples': 15282176, 'steps': 29847, 'loss/train': 1.844558835029602} 02/25/2022 16:14:43 - INFO - codeparrot_training - Step 29848: {'lr': 0.0001877007494977394, 'samples': 15282688, 'steps': 29848, 'loss/train': 1.942375659942627} 02/25/2022 16:14:49 - INFO - codeparrot_training - Step 29849: {'lr': 0.0001876849033590617, 'samples': 15283200, 'steps': 29849, 'loss/train': 1.2581619024276733} 02/25/2022 16:14:52 - INFO - codeparrot_training - Step 29850: {'lr': 0.00018766905748732212, 'samples': 15283712, 'steps': 29850, 'loss/train': 2.7367358207702637} 02/25/2022 16:14:58 - INFO - codeparrot_training - Step 29851: {'lr': 0.00018765321188258846, 'samples': 15284224, 'steps': 29851, 'loss/train': 1.8691809177398682} 02/25/2022 16:15:02 - INFO - codeparrot_training - Step 29852: {'lr': 0.00018763736654492863, 'samples': 15284736, 'steps': 29852, 'loss/train': 1.7424983978271484} 02/25/2022 16:15:07 - INFO - codeparrot_training - Step 29853: {'lr': 0.0001876215214744106, 'samples': 15285248, 'steps': 29853, 'loss/train': 2.612626314163208} 02/25/2022 16:15:11 - INFO - codeparrot_training - Step 29854: {'lr': 0.00018760567667110213, 'samples': 15285760, 'steps': 29854, 'loss/train': 1.677430510520935} 02/25/2022 16:15:17 - INFO - codeparrot_training - Step 29855: {'lr': 0.00018758983213507108, 'samples': 15286272, 'steps': 29855, 'loss/train': 1.66335928440094} 02/25/2022 16:15:20 - INFO - codeparrot_training - Step 29856: {'lr': 0.00018757398786638536, 'samples': 15286784, 'steps': 29856, 'loss/train': 1.9988110065460205} 02/25/2022 16:15:26 - INFO - codeparrot_training - Step 29857: {'lr': 0.00018755814386511296, 'samples': 15287296, 'steps': 29857, 'loss/train': 2.092353343963623} 02/25/2022 16:15:29 - INFO - codeparrot_training - Step 29858: {'lr': 0.00018754230013132156, 'samples': 15287808, 'steps': 29858, 'loss/train': 0.9833886623382568} 02/25/2022 16:15:35 - INFO - codeparrot_training - Step 29859: {'lr': 0.0001875264566650791, 'samples': 15288320, 'steps': 29859, 'loss/train': 1.6744505167007446} 02/25/2022 16:15:38 - INFO - codeparrot_training - Step 29860: {'lr': 0.00018751061346645347, 'samples': 15288832, 'steps': 29860, 'loss/train': 1.8313595056533813} 02/25/2022 16:15:44 - INFO - codeparrot_training - Step 29861: {'lr': 0.00018749477053551253, 'samples': 15289344, 'steps': 29861, 'loss/train': 0.6466661095619202} 02/25/2022 16:15:47 - INFO - codeparrot_training - Step 29862: {'lr': 0.00018747892787232423, 'samples': 15289856, 'steps': 29862, 'loss/train': 1.9630857706069946} 02/25/2022 16:15:53 - INFO - codeparrot_training - Step 29863: {'lr': 0.0001874630854769563, 'samples': 15290368, 'steps': 29863, 'loss/train': 2.6020519733428955} 02/25/2022 16:15:56 - INFO - codeparrot_training - Step 29864: {'lr': 0.0001874472433494766, 'samples': 15290880, 'steps': 29864, 'loss/train': 1.771794080734253} 02/25/2022 16:16:02 - INFO - codeparrot_training - Step 29865: {'lr': 0.00018743140148995307, 'samples': 15291392, 'steps': 29865, 'loss/train': 1.7185612916946411} 02/25/2022 16:16:06 - INFO - codeparrot_training - Step 29866: {'lr': 0.00018741555989845366, 'samples': 15291904, 'steps': 29866, 'loss/train': 1.4438657760620117} 02/25/2022 16:16:11 - INFO - codeparrot_training - Step 29867: {'lr': 0.00018739971857504597, 'samples': 15292416, 'steps': 29867, 'loss/train': 1.6309866905212402} 02/25/2022 16:16:15 - INFO - codeparrot_training - Step 29868: {'lr': 0.00018738387751979804, 'samples': 15292928, 'steps': 29868, 'loss/train': 0.8186879754066467} 02/25/2022 16:16:20 - INFO - codeparrot_training - Step 29869: {'lr': 0.0001873680367327777, 'samples': 15293440, 'steps': 29869, 'loss/train': 2.065586566925049} 02/25/2022 16:16:26 - INFO - codeparrot_training - Step 29870: {'lr': 0.0001873521962140529, 'samples': 15293952, 'steps': 29870, 'loss/train': 1.1738897562026978} 02/25/2022 16:16:29 - INFO - codeparrot_training - Step 29871: {'lr': 0.0001873363559636913, 'samples': 15294464, 'steps': 29871, 'loss/train': 1.2587348222732544} 02/25/2022 16:16:35 - INFO - codeparrot_training - Step 29872: {'lr': 0.00018732051598176086, 'samples': 15294976, 'steps': 29872, 'loss/train': 2.7985191345214844} 02/25/2022 16:16:38 - INFO - codeparrot_training - Step 29873: {'lr': 0.00018730467626832943, 'samples': 15295488, 'steps': 29873, 'loss/train': 1.3569531440734863} 02/25/2022 16:16:44 - INFO - codeparrot_training - Step 29874: {'lr': 0.00018728883682346482, 'samples': 15296000, 'steps': 29874, 'loss/train': 0.3122064471244812} 02/25/2022 16:16:47 - INFO - codeparrot_training - Step 29875: {'lr': 0.00018727299764723505, 'samples': 15296512, 'steps': 29875, 'loss/train': 1.2292375564575195} 02/25/2022 16:16:53 - INFO - codeparrot_training - Step 29876: {'lr': 0.0001872571587397077, 'samples': 15297024, 'steps': 29876, 'loss/train': 1.993888020515442} 02/25/2022 16:16:57 - INFO - codeparrot_training - Step 29877: {'lr': 0.00018724132010095084, 'samples': 15297536, 'steps': 29877, 'loss/train': 1.970518708229065} 02/25/2022 16:17:03 - INFO - codeparrot_training - Step 29878: {'lr': 0.00018722548173103216, 'samples': 15298048, 'steps': 29878, 'loss/train': 2.3291068077087402} 02/25/2022 16:17:06 - INFO - codeparrot_training - Step 29879: {'lr': 0.00018720964363001975, 'samples': 15298560, 'steps': 29879, 'loss/train': 0.7262600064277649} 02/25/2022 16:17:12 - INFO - codeparrot_training - Step 29880: {'lr': 0.00018719380579798112, 'samples': 15299072, 'steps': 29880, 'loss/train': 2.0507781505584717} 02/25/2022 16:17:15 - INFO - codeparrot_training - Step 29881: {'lr': 0.00018717796823498434, 'samples': 15299584, 'steps': 29881, 'loss/train': 1.4369564056396484} 02/25/2022 16:17:21 - INFO - codeparrot_training - Step 29882: {'lr': 0.00018716213094109715, 'samples': 15300096, 'steps': 29882, 'loss/train': 2.242964267730713} 02/25/2022 16:17:24 - INFO - codeparrot_training - Step 29883: {'lr': 0.00018714629391638758, 'samples': 15300608, 'steps': 29883, 'loss/train': 1.3677055835723877} 02/25/2022 16:17:30 - INFO - codeparrot_training - Step 29884: {'lr': 0.0001871304571609232, 'samples': 15301120, 'steps': 29884, 'loss/train': 1.5618127584457397} 02/25/2022 16:17:33 - INFO - codeparrot_training - Step 29885: {'lr': 0.00018711462067477197, 'samples': 15301632, 'steps': 29885, 'loss/train': 2.276064157485962} 02/25/2022 16:17:40 - INFO - codeparrot_training - Step 29886: {'lr': 0.00018709878445800177, 'samples': 15302144, 'steps': 29886, 'loss/train': 0.7911723852157593} 02/25/2022 16:17:43 - INFO - codeparrot_training - Step 29887: {'lr': 0.0001870829485106804, 'samples': 15302656, 'steps': 29887, 'loss/train': 0.036495327949523926} 02/25/2022 16:17:49 - INFO - codeparrot_training - Step 29888: {'lr': 0.00018706711283287576, 'samples': 15303168, 'steps': 29888, 'loss/train': 1.5549920797348022} 02/25/2022 16:17:52 - INFO - codeparrot_training - Step 29889: {'lr': 0.00018705127742465555, 'samples': 15303680, 'steps': 29889, 'loss/train': 1.387128233909607} 02/25/2022 16:17:57 - INFO - codeparrot_training - Step 29890: {'lr': 0.00018703544228608775, 'samples': 15304192, 'steps': 29890, 'loss/train': 1.522879958152771} 02/25/2022 16:18:01 - INFO - codeparrot_training - Step 29891: {'lr': 0.00018701960741724003, 'samples': 15304704, 'steps': 29891, 'loss/train': 2.4289002418518066} 02/25/2022 16:18:07 - INFO - codeparrot_training - Step 29892: {'lr': 0.00018700377281818037, 'samples': 15305216, 'steps': 29892, 'loss/train': 0.4750358462333679} 02/25/2022 16:18:10 - INFO - codeparrot_training - Step 29893: {'lr': 0.00018698793848897656, 'samples': 15305728, 'steps': 29893, 'loss/train': 1.9496701955795288} 02/25/2022 16:18:15 - INFO - codeparrot_training - Step 29894: {'lr': 0.00018697210442969637, 'samples': 15306240, 'steps': 29894, 'loss/train': 2.053495407104492} 02/25/2022 16:18:19 - INFO - codeparrot_training - Step 29895: {'lr': 0.00018695627064040766, 'samples': 15306752, 'steps': 29895, 'loss/train': 1.6171382665634155} 02/25/2022 16:18:25 - INFO - codeparrot_training - Step 29896: {'lr': 0.00018694043712117828, 'samples': 15307264, 'steps': 29896, 'loss/train': 1.5707803964614868} 02/25/2022 16:18:28 - INFO - codeparrot_training - Step 29897: {'lr': 0.00018692460387207616, 'samples': 15307776, 'steps': 29897, 'loss/train': 1.1517490148544312} 02/25/2022 16:18:34 - INFO - codeparrot_training - Step 29898: {'lr': 0.0001869087708931689, 'samples': 15308288, 'steps': 29898, 'loss/train': 3.257429838180542} 02/25/2022 16:18:37 - INFO - codeparrot_training - Step 29899: {'lr': 0.00018689293818452437, 'samples': 15308800, 'steps': 29899, 'loss/train': 1.906327486038208} 02/25/2022 16:18:43 - INFO - codeparrot_training - Step 29900: {'lr': 0.00018687710574621051, 'samples': 15309312, 'steps': 29900, 'loss/train': 2.666914701461792} 02/25/2022 16:18:47 - INFO - codeparrot_training - Step 29901: {'lr': 0.0001868612735782952, 'samples': 15309824, 'steps': 29901, 'loss/train': 1.3863444328308105} 02/25/2022 16:18:53 - INFO - codeparrot_training - Step 29902: {'lr': 0.000186845441680846, 'samples': 15310336, 'steps': 29902, 'loss/train': 1.2310504913330078} 02/25/2022 16:18:57 - INFO - codeparrot_training - Step 29903: {'lr': 0.0001868296100539309, 'samples': 15310848, 'steps': 29903, 'loss/train': 1.6549617052078247} 02/25/2022 16:19:02 - INFO - codeparrot_training - Step 29904: {'lr': 0.00018681377869761767, 'samples': 15311360, 'steps': 29904, 'loss/train': 1.6674079895019531} 02/25/2022 16:19:06 - INFO - codeparrot_training - Step 29905: {'lr': 0.00018679794761197426, 'samples': 15311872, 'steps': 29905, 'loss/train': 0.11368287354707718} 02/25/2022 16:19:11 - INFO - codeparrot_training - Step 29906: {'lr': 0.00018678211679706828, 'samples': 15312384, 'steps': 29906, 'loss/train': 1.024168610572815} 02/25/2022 16:19:15 - INFO - codeparrot_training - Step 29907: {'lr': 0.00018676628625296764, 'samples': 15312896, 'steps': 29907, 'loss/train': 2.765545606613159} 02/25/2022 16:19:20 - INFO - codeparrot_training - Step 29908: {'lr': 0.00018675045597974014, 'samples': 15313408, 'steps': 29908, 'loss/train': 1.0893880128860474} 02/25/2022 16:19:24 - INFO - codeparrot_training - Step 29909: {'lr': 0.0001867346259774536, 'samples': 15313920, 'steps': 29909, 'loss/train': 1.813191294670105} 02/25/2022 16:19:29 - INFO - codeparrot_training - Step 29910: {'lr': 0.00018671879624617592, 'samples': 15314432, 'steps': 29910, 'loss/train': 1.8024532794952393} 02/25/2022 16:19:33 - INFO - codeparrot_training - Step 29911: {'lr': 0.00018670296678597475, 'samples': 15314944, 'steps': 29911, 'loss/train': 1.6160545349121094} 02/25/2022 16:19:39 - INFO - codeparrot_training - Step 29912: {'lr': 0.00018668713759691796, 'samples': 15315456, 'steps': 29912, 'loss/train': 1.8470666408538818} 02/25/2022 16:19:43 - INFO - codeparrot_training - Step 29913: {'lr': 0.0001866713086790734, 'samples': 15315968, 'steps': 29913, 'loss/train': 1.7411481142044067} 02/25/2022 16:19:48 - INFO - codeparrot_training - Step 29914: {'lr': 0.00018665548003250892, 'samples': 15316480, 'steps': 29914, 'loss/train': 1.9195104837417603} 02/25/2022 16:19:52 - INFO - codeparrot_training - Step 29915: {'lr': 0.00018663965165729212, 'samples': 15316992, 'steps': 29915, 'loss/train': 2.3886818885803223} 02/25/2022 16:19:57 - INFO - codeparrot_training - Step 29916: {'lr': 0.00018662382355349097, 'samples': 15317504, 'steps': 29916, 'loss/train': 1.9789098501205444} 02/25/2022 16:20:01 - INFO - codeparrot_training - Step 29917: {'lr': 0.00018660799572117322, 'samples': 15318016, 'steps': 29917, 'loss/train': 2.1190907955169678} 02/25/2022 16:20:06 - INFO - codeparrot_training - Step 29918: {'lr': 0.00018659216816040681, 'samples': 15318528, 'steps': 29918, 'loss/train': 1.5890998840332031} 02/25/2022 16:20:10 - INFO - codeparrot_training - Step 29919: {'lr': 0.0001865763408712593, 'samples': 15319040, 'steps': 29919, 'loss/train': 0.4515683650970459} 02/25/2022 16:20:15 - INFO - codeparrot_training - Step 29920: {'lr': 0.0001865605138537986, 'samples': 15319552, 'steps': 29920, 'loss/train': 1.984484076499939} 02/25/2022 16:20:18 - INFO - codeparrot_training - Step 29921: {'lr': 0.00018654468710809253, 'samples': 15320064, 'steps': 29921, 'loss/train': 1.7808912992477417} 02/25/2022 16:20:25 - INFO - codeparrot_training - Step 29922: {'lr': 0.00018652886063420888, 'samples': 15320576, 'steps': 29922, 'loss/train': 1.6435508728027344} 02/25/2022 16:20:31 - INFO - codeparrot_training - Step 29923: {'lr': 0.0001865130344322155, 'samples': 15321088, 'steps': 29923, 'loss/train': 1.0891015529632568} 02/25/2022 16:20:34 - INFO - codeparrot_training - Step 29924: {'lr': 0.00018649720850218005, 'samples': 15321600, 'steps': 29924, 'loss/train': 1.3649837970733643} 02/25/2022 16:20:40 - INFO - codeparrot_training - Step 29925: {'lr': 0.00018648138284417038, 'samples': 15322112, 'steps': 29925, 'loss/train': 2.976989984512329} 02/25/2022 16:20:43 - INFO - codeparrot_training - Step 29926: {'lr': 0.00018646555745825428, 'samples': 15322624, 'steps': 29926, 'loss/train': 1.773169994354248} 02/25/2022 16:20:47 - INFO - codeparrot_training - Step 29927: {'lr': 0.0001864497323444997, 'samples': 15323136, 'steps': 29927, 'loss/train': 0.02213803492486477} 02/25/2022 16:20:52 - INFO - codeparrot_training - Step 29928: {'lr': 0.00018643390750297417, 'samples': 15323648, 'steps': 29928, 'loss/train': 1.4953334331512451} 02/25/2022 16:20:56 - INFO - codeparrot_training - Step 29929: {'lr': 0.00018641808293374558, 'samples': 15324160, 'steps': 29929, 'loss/train': 1.6758784055709839} 02/25/2022 16:21:02 - INFO - codeparrot_training - Step 29930: {'lr': 0.00018640225863688178, 'samples': 15324672, 'steps': 29930, 'loss/train': 1.9071877002716064} 02/25/2022 16:21:05 - INFO - codeparrot_training - Step 29931: {'lr': 0.00018638643461245052, 'samples': 15325184, 'steps': 29931, 'loss/train': 0.02334209717810154} 02/25/2022 16:21:12 - INFO - codeparrot_training - Step 29932: {'lr': 0.00018637061086051955, 'samples': 15325696, 'steps': 29932, 'loss/train': 1.3272088766098022} 02/25/2022 16:21:15 - INFO - codeparrot_training - Step 29933: {'lr': 0.00018635478738115662, 'samples': 15326208, 'steps': 29933, 'loss/train': 1.2069669961929321} 02/25/2022 16:21:21 - INFO - codeparrot_training - Step 29934: {'lr': 0.0001863389641744297, 'samples': 15326720, 'steps': 29934, 'loss/train': 2.0012550354003906} 02/25/2022 16:21:24 - INFO - codeparrot_training - Step 29935: {'lr': 0.00018632314124040635, 'samples': 15327232, 'steps': 29935, 'loss/train': 1.572792887687683} 02/25/2022 16:21:30 - INFO - codeparrot_training - Step 29936: {'lr': 0.00018630731857915452, 'samples': 15327744, 'steps': 29936, 'loss/train': 2.4771556854248047} 02/25/2022 16:21:33 - INFO - codeparrot_training - Step 29937: {'lr': 0.00018629149619074186, 'samples': 15328256, 'steps': 29937, 'loss/train': 1.747625708580017} 02/25/2022 16:21:39 - INFO - codeparrot_training - Step 29938: {'lr': 0.00018627567407523615, 'samples': 15328768, 'steps': 29938, 'loss/train': 1.184655785560608} 02/25/2022 16:21:44 - INFO - codeparrot_training - Step 29939: {'lr': 0.00018625985223270526, 'samples': 15329280, 'steps': 29939, 'loss/train': 2.02947998046875} 02/25/2022 16:21:48 - INFO - codeparrot_training - Step 29940: {'lr': 0.00018624403066321696, 'samples': 15329792, 'steps': 29940, 'loss/train': 0.8118106722831726} 02/25/2022 16:21:54 - INFO - codeparrot_training - Step 29941: {'lr': 0.00018622820936683898, 'samples': 15330304, 'steps': 29941, 'loss/train': 0.9568519592285156} 02/25/2022 16:21:58 - INFO - codeparrot_training - Step 29942: {'lr': 0.00018621238834363906, 'samples': 15330816, 'steps': 29942, 'loss/train': 1.9766855239868164} 02/25/2022 16:22:01 - INFO - codeparrot_training - Step 29943: {'lr': 0.00018619656759368504, 'samples': 15331328, 'steps': 29943, 'loss/train': 1.9341517686843872} 02/25/2022 16:22:07 - INFO - codeparrot_training - Step 29944: {'lr': 0.00018618074711704462, 'samples': 15331840, 'steps': 29944, 'loss/train': 1.0070363283157349} 02/25/2022 16:22:10 - INFO - codeparrot_training - Step 29945: {'lr': 0.00018616492691378572, 'samples': 15332352, 'steps': 29945, 'loss/train': 1.2417387962341309} 02/25/2022 16:22:16 - INFO - codeparrot_training - Step 29946: {'lr': 0.00018614910698397591, 'samples': 15332864, 'steps': 29946, 'loss/train': 2.1367080211639404} 02/25/2022 16:22:19 - INFO - codeparrot_training - Step 29947: {'lr': 0.00018613328732768305, 'samples': 15333376, 'steps': 29947, 'loss/train': 1.1206823587417603} 02/25/2022 16:22:25 - INFO - codeparrot_training - Step 29948: {'lr': 0.00018611746794497492, 'samples': 15333888, 'steps': 29948, 'loss/train': 1.4667154550552368} 02/25/2022 16:22:28 - INFO - codeparrot_training - Step 29949: {'lr': 0.00018610164883591936, 'samples': 15334400, 'steps': 29949, 'loss/train': 1.8145860433578491} 02/25/2022 16:22:34 - INFO - codeparrot_training - Step 29950: {'lr': 0.00018608583000058395, 'samples': 15334912, 'steps': 29950, 'loss/train': 0.9149900674819946} 02/25/2022 16:22:37 - INFO - codeparrot_training - Step 29951: {'lr': 0.00018607001143903658, 'samples': 15335424, 'steps': 29951, 'loss/train': 8.858648300170898} 02/25/2022 16:22:43 - INFO - codeparrot_training - Step 29952: {'lr': 0.0001860541931513449, 'samples': 15335936, 'steps': 29952, 'loss/train': 1.905663251876831} 02/25/2022 16:22:46 - INFO - codeparrot_training - Step 29953: {'lr': 0.00018603837513757692, 'samples': 15336448, 'steps': 29953, 'loss/train': 1.3951255083084106} 02/25/2022 16:22:52 - INFO - codeparrot_training - Step 29954: {'lr': 0.0001860225573978001, 'samples': 15336960, 'steps': 29954, 'loss/train': 1.651943564414978} 02/25/2022 16:22:56 - INFO - codeparrot_training - Step 29955: {'lr': 0.00018600673993208233, 'samples': 15337472, 'steps': 29955, 'loss/train': 1.7826783657073975} 02/25/2022 16:23:01 - INFO - codeparrot_training - Step 29956: {'lr': 0.0001859909227404914, 'samples': 15337984, 'steps': 29956, 'loss/train': 1.4395363330841064} 02/25/2022 16:23:05 - INFO - codeparrot_training - Step 29957: {'lr': 0.000185975105823095, 'samples': 15338496, 'steps': 29957, 'loss/train': 0.8780284523963928} 02/25/2022 16:23:11 - INFO - codeparrot_training - Step 29958: {'lr': 0.000185959289179961, 'samples': 15339008, 'steps': 29958, 'loss/train': 1.5655696392059326} 02/25/2022 16:23:14 - INFO - codeparrot_training - Step 29959: {'lr': 0.00018594347281115703, 'samples': 15339520, 'steps': 29959, 'loss/train': 1.1873611211776733} 02/25/2022 16:23:20 - INFO - codeparrot_training - Step 29960: {'lr': 0.00018592765671675081, 'samples': 15340032, 'steps': 29960, 'loss/train': 1.6124969720840454} 02/25/2022 16:23:23 - INFO - codeparrot_training - Step 29961: {'lr': 0.00018591184089681025, 'samples': 15340544, 'steps': 29961, 'loss/train': 1.7341086864471436} 02/25/2022 16:23:29 - INFO - codeparrot_training - Step 29962: {'lr': 0.00018589602535140303, 'samples': 15341056, 'steps': 29962, 'loss/train': 1.8591160774230957} 02/25/2022 16:23:34 - INFO - codeparrot_training - Step 29963: {'lr': 0.00018588021008059683, 'samples': 15341568, 'steps': 29963, 'loss/train': 2.252802848815918} 02/25/2022 16:23:38 - INFO - codeparrot_training - Step 29964: {'lr': 0.00018586439508445942, 'samples': 15342080, 'steps': 29964, 'loss/train': 1.975447416305542} 02/25/2022 16:23:43 - INFO - codeparrot_training - Step 29965: {'lr': 0.00018584858036305862, 'samples': 15342592, 'steps': 29965, 'loss/train': 0.9785515666007996} 02/25/2022 16:23:47 - INFO - codeparrot_training - Step 29966: {'lr': 0.0001858327659164622, 'samples': 15343104, 'steps': 29966, 'loss/train': 1.7467445135116577} 02/25/2022 16:23:53 - INFO - codeparrot_training - Step 29967: {'lr': 0.00018581695174473773, 'samples': 15343616, 'steps': 29967, 'loss/train': 2.2341299057006836} 02/25/2022 16:23:56 - INFO - codeparrot_training - Step 29968: {'lr': 0.00018580113784795306, 'samples': 15344128, 'steps': 29968, 'loss/train': 2.1433560848236084} 02/25/2022 16:24:02 - INFO - codeparrot_training - Step 29969: {'lr': 0.00018578532422617593, 'samples': 15344640, 'steps': 29969, 'loss/train': 1.321009635925293} 02/25/2022 16:24:05 - INFO - codeparrot_training - Step 29970: {'lr': 0.00018576951087947413, 'samples': 15345152, 'steps': 29970, 'loss/train': 1.6169174909591675} 02/25/2022 16:24:11 - INFO - codeparrot_training - Step 29971: {'lr': 0.0001857536978079154, 'samples': 15345664, 'steps': 29971, 'loss/train': 1.31284499168396} 02/25/2022 16:24:15 - INFO - codeparrot_training - Step 29972: {'lr': 0.0001857378850115673, 'samples': 15346176, 'steps': 29972, 'loss/train': 1.9284205436706543} 02/25/2022 16:24:20 - INFO - codeparrot_training - Step 29973: {'lr': 0.00018572207249049774, 'samples': 15346688, 'steps': 29973, 'loss/train': 1.9425864219665527} 02/25/2022 16:24:24 - INFO - codeparrot_training - Step 29974: {'lr': 0.00018570626024477443, 'samples': 15347200, 'steps': 29974, 'loss/train': 2.069139003753662} 02/25/2022 16:24:29 - INFO - codeparrot_training - Step 29975: {'lr': 0.00018569044827446512, 'samples': 15347712, 'steps': 29975, 'loss/train': 1.0136204957962036} 02/25/2022 16:24:33 - INFO - codeparrot_training - Step 29976: {'lr': 0.00018567463657963746, 'samples': 15348224, 'steps': 29976, 'loss/train': 0.7865616083145142} 02/25/2022 16:24:39 - INFO - codeparrot_training - Step 29977: {'lr': 0.0001856588251603592, 'samples': 15348736, 'steps': 29977, 'loss/train': 1.5731464624404907} 02/25/2022 16:24:42 - INFO - codeparrot_training - Step 29978: {'lr': 0.0001856430140166982, 'samples': 15349248, 'steps': 29978, 'loss/train': 1.0307871103286743} 02/25/2022 16:24:48 - INFO - codeparrot_training - Step 29979: {'lr': 0.00018562720314872205, 'samples': 15349760, 'steps': 29979, 'loss/train': 1.8848859071731567} 02/25/2022 16:24:51 - INFO - codeparrot_training - Step 29980: {'lr': 0.00018561139255649848, 'samples': 15350272, 'steps': 29980, 'loss/train': 1.8795145750045776} 02/25/2022 16:24:57 - INFO - codeparrot_training - Step 29981: {'lr': 0.00018559558224009532, 'samples': 15350784, 'steps': 29981, 'loss/train': 1.0919599533081055} 02/25/2022 16:25:00 - INFO - codeparrot_training - Step 29982: {'lr': 0.00018557977219958016, 'samples': 15351296, 'steps': 29982, 'loss/train': 1.9145148992538452} 02/25/2022 16:25:06 - INFO - codeparrot_training - Step 29983: {'lr': 0.00018556396243502084, 'samples': 15351808, 'steps': 29983, 'loss/train': 1.362343668937683} 02/25/2022 16:25:09 - INFO - codeparrot_training - Step 29984: {'lr': 0.00018554815294648505, 'samples': 15352320, 'steps': 29984, 'loss/train': 4.103315830230713} 02/25/2022 16:25:15 - INFO - codeparrot_training - Step 29985: {'lr': 0.00018553234373404055, 'samples': 15352832, 'steps': 29985, 'loss/train': 0.41832172870635986} 02/25/2022 16:25:18 - INFO - codeparrot_training - Step 29986: {'lr': 0.00018551653479775493, 'samples': 15353344, 'steps': 29986, 'loss/train': 0.9002289175987244} 02/25/2022 16:25:24 - INFO - codeparrot_training - Step 29987: {'lr': 0.00018550072613769604, 'samples': 15353856, 'steps': 29987, 'loss/train': 2.3278775215148926} 02/25/2022 16:25:27 - INFO - codeparrot_training - Step 29988: {'lr': 0.0001854849177539316, 'samples': 15354368, 'steps': 29988, 'loss/train': 1.7065527439117432} 02/25/2022 16:25:33 - INFO - codeparrot_training - Step 29989: {'lr': 0.00018546910964652925, 'samples': 15354880, 'steps': 29989, 'loss/train': 1.6275399923324585} 02/25/2022 16:25:37 - INFO - codeparrot_training - Step 29990: {'lr': 0.0001854533018155567, 'samples': 15355392, 'steps': 29990, 'loss/train': 2.2125675678253174} 02/25/2022 16:25:42 - INFO - codeparrot_training - Step 29991: {'lr': 0.00018543749426108174, 'samples': 15355904, 'steps': 29991, 'loss/train': 1.320961356163025} 02/25/2022 16:25:46 - INFO - codeparrot_training - Step 29992: {'lr': 0.00018542168698317202, 'samples': 15356416, 'steps': 29992, 'loss/train': 1.1490603685379028} 02/25/2022 16:25:51 - INFO - codeparrot_training - Step 29993: {'lr': 0.0001854058799818954, 'samples': 15356928, 'steps': 29993, 'loss/train': 1.6798889636993408} 02/25/2022 16:25:55 - INFO - codeparrot_training - Step 29994: {'lr': 0.00018539007325731939, 'samples': 15357440, 'steps': 29994, 'loss/train': 0.7727236747741699} 02/25/2022 16:26:00 - INFO - codeparrot_training - Step 29995: {'lr': 0.00018537426680951177, 'samples': 15357952, 'steps': 29995, 'loss/train': 2.1184613704681396} 02/25/2022 16:26:04 - INFO - codeparrot_training - Step 29996: {'lr': 0.00018535846063854027, 'samples': 15358464, 'steps': 29996, 'loss/train': 2.0593769550323486} 02/25/2022 16:26:09 - INFO - codeparrot_training - Step 29997: {'lr': 0.00018534265474447274, 'samples': 15358976, 'steps': 29997, 'loss/train': 1.5953224897384644} 02/25/2022 16:26:13 - INFO - codeparrot_training - Step 29998: {'lr': 0.00018532684912737663, 'samples': 15359488, 'steps': 29998, 'loss/train': 2.062704086303711} 02/25/2022 16:26:18 - INFO - codeparrot_training - Step 29999: {'lr': 0.00018531104378731973, 'samples': 15360000, 'steps': 29999, 'loss/train': 1.6037700176239014} 02/25/2022 16:26:18 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 16:26:36 - WARNING - huggingface_hub.repository - Several commits (30) will be pushed upstream. 02/25/2022 16:26:36 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 16:27:10 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 95f22fa..af58905 floral-grass-11 -> floral-grass-11 02/25/2022 16:27:15 - INFO - codeparrot_training - Step 30000: {'lr': 0.0001852952387243698, 'samples': 15360512, 'steps': 30000, 'loss/train': 2.1731338500976562} 02/25/2022 16:27:21 - INFO - codeparrot_training - Step 30001: {'lr': 0.0001852794339385946, 'samples': 15361024, 'steps': 30001, 'loss/train': 1.118362307548523} 02/25/2022 16:27:24 - INFO - codeparrot_training - Step 30002: {'lr': 0.0001852636294300617, 'samples': 15361536, 'steps': 30002, 'loss/train': 1.3688808679580688} 02/25/2022 16:27:30 - INFO - codeparrot_training - Step 30003: {'lr': 0.00018524782519883882, 'samples': 15362048, 'steps': 30003, 'loss/train': 2.811722993850708} 02/25/2022 16:27:34 - INFO - codeparrot_training - Step 30004: {'lr': 0.00018523202124499368, 'samples': 15362560, 'steps': 30004, 'loss/train': 1.7936128377914429} 02/25/2022 16:27:39 - INFO - codeparrot_training - Step 30005: {'lr': 0.00018521621756859404, 'samples': 15363072, 'steps': 30005, 'loss/train': 0.579941987991333} 02/25/2022 16:27:43 - INFO - codeparrot_training - Step 30006: {'lr': 0.00018520041416970762, 'samples': 15363584, 'steps': 30006, 'loss/train': 2.729137420654297} 02/25/2022 16:27:48 - INFO - codeparrot_training - Step 30007: {'lr': 0.00018518461104840195, 'samples': 15364096, 'steps': 30007, 'loss/train': 2.9498283863067627} 02/25/2022 16:27:52 - INFO - codeparrot_training - Step 30008: {'lr': 0.00018516880820474484, 'samples': 15364608, 'steps': 30008, 'loss/train': 0.9807052612304688} 02/25/2022 16:27:58 - INFO - codeparrot_training - Step 30009: {'lr': 0.000185153005638804, 'samples': 15365120, 'steps': 30009, 'loss/train': 2.155941963195801} 02/25/2022 16:28:01 - INFO - codeparrot_training - Step 30010: {'lr': 0.00018513720335064715, 'samples': 15365632, 'steps': 30010, 'loss/train': 1.6149916648864746} 02/25/2022 16:28:06 - INFO - codeparrot_training - Step 30011: {'lr': 0.00018512140134034182, 'samples': 15366144, 'steps': 30011, 'loss/train': 2.8154728412628174} 02/25/2022 16:28:10 - INFO - codeparrot_training - Step 30012: {'lr': 0.0001851055996079558, 'samples': 15366656, 'steps': 30012, 'loss/train': 1.0076522827148438} 02/25/2022 16:28:16 - INFO - codeparrot_training - Step 30013: {'lr': 0.0001850897981535568, 'samples': 15367168, 'steps': 30013, 'loss/train': 1.431535005569458} 02/25/2022 16:28:20 - INFO - codeparrot_training - Step 30014: {'lr': 0.0001850739969772126, 'samples': 15367680, 'steps': 30014, 'loss/train': 0.21688517928123474} 02/25/2022 16:28:25 - INFO - codeparrot_training - Step 30015: {'lr': 0.0001850581960789907, 'samples': 15368192, 'steps': 30015, 'loss/train': 2.0599780082702637} 02/25/2022 16:28:29 - INFO - codeparrot_training - Step 30016: {'lr': 0.00018504239545895885, 'samples': 15368704, 'steps': 30016, 'loss/train': 0.7041217088699341} 02/25/2022 16:28:34 - INFO - codeparrot_training - Step 30017: {'lr': 0.00018502659511718473, 'samples': 15369216, 'steps': 30017, 'loss/train': 1.148532748222351} 02/25/2022 16:28:38 - INFO - codeparrot_training - Step 30018: {'lr': 0.00018501079505373614, 'samples': 15369728, 'steps': 30018, 'loss/train': 1.7754034996032715} 02/25/2022 16:28:43 - INFO - codeparrot_training - Step 30019: {'lr': 0.00018499499526868063, 'samples': 15370240, 'steps': 30019, 'loss/train': 0.9794594645500183} 02/25/2022 16:28:47 - INFO - codeparrot_training - Step 30020: {'lr': 0.00018497919576208587, 'samples': 15370752, 'steps': 30020, 'loss/train': 2.097980499267578} 02/25/2022 16:28:53 - INFO - codeparrot_training - Step 30021: {'lr': 0.00018496339653401963, 'samples': 15371264, 'steps': 30021, 'loss/train': 1.9253982305526733} 02/25/2022 16:28:56 - INFO - codeparrot_training - Step 30022: {'lr': 0.00018494759758454953, 'samples': 15371776, 'steps': 30022, 'loss/train': 2.1501739025115967} 02/25/2022 16:29:02 - INFO - codeparrot_training - Step 30023: {'lr': 0.0001849317989137433, 'samples': 15372288, 'steps': 30023, 'loss/train': 0.7061108946800232} 02/25/2022 16:29:06 - INFO - codeparrot_training - Step 30024: {'lr': 0.0001849160005216685, 'samples': 15372800, 'steps': 30024, 'loss/train': 1.9436016082763672} 02/25/2022 16:29:11 - INFO - codeparrot_training - Step 30025: {'lr': 0.000184900202408393, 'samples': 15373312, 'steps': 30025, 'loss/train': 1.7332801818847656} 02/25/2022 16:29:15 - INFO - codeparrot_training - Step 30026: {'lr': 0.00018488440457398425, 'samples': 15373824, 'steps': 30026, 'loss/train': 1.6052758693695068} 02/25/2022 16:29:21 - INFO - codeparrot_training - Step 30027: {'lr': 0.00018486860701851014, 'samples': 15374336, 'steps': 30027, 'loss/train': 1.5934809446334839} 02/25/2022 16:29:24 - INFO - codeparrot_training - Step 30028: {'lr': 0.00018485280974203823, 'samples': 15374848, 'steps': 30028, 'loss/train': 1.5203590393066406} 02/25/2022 16:29:29 - INFO - codeparrot_training - Step 30029: {'lr': 0.0001848370127446361, 'samples': 15375360, 'steps': 30029, 'loss/train': 0.6706119775772095} 02/25/2022 16:29:33 - INFO - codeparrot_training - Step 30030: {'lr': 0.00018482121602637154, 'samples': 15375872, 'steps': 30030, 'loss/train': 1.453978419303894} 02/25/2022 16:29:38 - INFO - codeparrot_training - Step 30031: {'lr': 0.0001848054195873122, 'samples': 15376384, 'steps': 30031, 'loss/train': 1.6087207794189453} 02/25/2022 16:29:42 - INFO - codeparrot_training - Step 30032: {'lr': 0.00018478962342752584, 'samples': 15376896, 'steps': 30032, 'loss/train': 1.5389703512191772} 02/25/2022 16:29:47 - INFO - codeparrot_training - Step 30033: {'lr': 0.0001847738275470799, 'samples': 15377408, 'steps': 30033, 'loss/train': 1.95013427734375} 02/25/2022 16:29:53 - INFO - codeparrot_training - Step 30034: {'lr': 0.0001847580319460422, 'samples': 15377920, 'steps': 30034, 'loss/train': 2.7690927982330322} 02/25/2022 16:29:56 - INFO - codeparrot_training - Step 30035: {'lr': 0.00018474223662448037, 'samples': 15378432, 'steps': 30035, 'loss/train': 2.2924814224243164} 02/25/2022 16:30:02 - INFO - codeparrot_training - Step 30036: {'lr': 0.00018472644158246219, 'samples': 15378944, 'steps': 30036, 'loss/train': 1.685707926750183} 02/25/2022 16:30:05 - INFO - codeparrot_training - Step 30037: {'lr': 0.0001847106468200551, 'samples': 15379456, 'steps': 30037, 'loss/train': 0.7427926063537598} 02/25/2022 16:30:11 - INFO - codeparrot_training - Step 30038: {'lr': 0.00018469485233732686, 'samples': 15379968, 'steps': 30038, 'loss/train': 0.8560056686401367} 02/25/2022 16:30:15 - INFO - codeparrot_training - Step 30039: {'lr': 0.00018467905813434514, 'samples': 15380480, 'steps': 30039, 'loss/train': 2.108710765838623} 02/25/2022 16:30:20 - INFO - codeparrot_training - Step 30040: {'lr': 0.00018466326421117768, 'samples': 15380992, 'steps': 30040, 'loss/train': 2.2106690406799316} 02/25/2022 16:30:24 - INFO - codeparrot_training - Step 30041: {'lr': 0.00018464747056789194, 'samples': 15381504, 'steps': 30041, 'loss/train': 1.280234694480896} 02/25/2022 16:30:29 - INFO - codeparrot_training - Step 30042: {'lr': 0.0001846316772045557, 'samples': 15382016, 'steps': 30042, 'loss/train': 0.9699698090553284} 02/25/2022 16:30:33 - INFO - codeparrot_training - Step 30043: {'lr': 0.00018461588412123662, 'samples': 15382528, 'steps': 30043, 'loss/train': 1.8953702449798584} 02/25/2022 16:30:39 - INFO - codeparrot_training - Step 30044: {'lr': 0.00018460009131800233, 'samples': 15383040, 'steps': 30044, 'loss/train': 1.057846188545227} 02/25/2022 16:30:42 - INFO - codeparrot_training - Step 30045: {'lr': 0.00018458429879492056, 'samples': 15383552, 'steps': 30045, 'loss/train': 1.8462986946105957} 02/25/2022 16:30:48 - INFO - codeparrot_training - Step 30046: {'lr': 0.00018456850655205876, 'samples': 15384064, 'steps': 30046, 'loss/train': 2.8983097076416016} 02/25/2022 16:30:51 - INFO - codeparrot_training - Step 30047: {'lr': 0.00018455271458948474, 'samples': 15384576, 'steps': 30047, 'loss/train': 1.0029246807098389} 02/25/2022 16:30:57 - INFO - codeparrot_training - Step 30048: {'lr': 0.00018453692290726608, 'samples': 15385088, 'steps': 30048, 'loss/train': 2.83892560005188} 02/25/2022 16:31:01 - INFO - codeparrot_training - Step 30049: {'lr': 0.0001845211315054706, 'samples': 15385600, 'steps': 30049, 'loss/train': 1.5945707559585571} 02/25/2022 16:31:06 - INFO - codeparrot_training - Step 30050: {'lr': 0.00018450534038416566, 'samples': 15386112, 'steps': 30050, 'loss/train': 1.7259442806243896} 02/25/2022 16:31:10 - INFO - codeparrot_training - Step 30051: {'lr': 0.00018448954954341907, 'samples': 15386624, 'steps': 30051, 'loss/train': 1.0598602294921875} 02/25/2022 16:31:15 - INFO - codeparrot_training - Step 30052: {'lr': 0.00018447375898329843, 'samples': 15387136, 'steps': 30052, 'loss/train': 2.3258752822875977} 02/25/2022 16:31:19 - INFO - codeparrot_training - Step 30053: {'lr': 0.0001844579687038714, 'samples': 15387648, 'steps': 30053, 'loss/train': 1.9298633337020874} 02/25/2022 16:31:24 - INFO - codeparrot_training - Step 30054: {'lr': 0.00018444217870520574, 'samples': 15388160, 'steps': 30054, 'loss/train': 1.5542007684707642} 02/25/2022 16:31:28 - INFO - codeparrot_training - Step 30055: {'lr': 0.00018442638898736886, 'samples': 15388672, 'steps': 30055, 'loss/train': 0.4983118176460266} 02/25/2022 16:31:33 - INFO - codeparrot_training - Step 30056: {'lr': 0.0001844105995504285, 'samples': 15389184, 'steps': 30056, 'loss/train': 1.224015474319458} 02/25/2022 16:31:37 - INFO - codeparrot_training - Step 30057: {'lr': 0.00018439481039445228, 'samples': 15389696, 'steps': 30057, 'loss/train': 1.2572087049484253} 02/25/2022 16:31:43 - INFO - codeparrot_training - Step 30058: {'lr': 0.000184379021519508, 'samples': 15390208, 'steps': 30058, 'loss/train': 1.927124261856079} 02/25/2022 16:31:46 - INFO - codeparrot_training - Step 30059: {'lr': 0.00018436323292566305, 'samples': 15390720, 'steps': 30059, 'loss/train': 4.666149139404297} 02/25/2022 16:31:52 - INFO - codeparrot_training - Step 30060: {'lr': 0.00018434744461298514, 'samples': 15391232, 'steps': 30060, 'loss/train': 1.5695821046829224} 02/25/2022 16:31:56 - INFO - codeparrot_training - Step 30061: {'lr': 0.000184331656581542, 'samples': 15391744, 'steps': 30061, 'loss/train': 0.10898319631814957} 02/25/2022 16:32:01 - INFO - codeparrot_training - Step 30062: {'lr': 0.0001843158688314012, 'samples': 15392256, 'steps': 30062, 'loss/train': 1.9409795999526978} 02/25/2022 16:32:05 - INFO - codeparrot_training - Step 30063: {'lr': 0.0001843000813626303, 'samples': 15392768, 'steps': 30063, 'loss/train': 1.1655218601226807} 02/25/2022 16:32:12 - INFO - codeparrot_training - Step 30064: {'lr': 0.000184284294175297, 'samples': 15393280, 'steps': 30064, 'loss/train': 1.15315842628479} 02/25/2022 16:32:15 - INFO - codeparrot_training - Step 30065: {'lr': 0.00018426850726946896, 'samples': 15393792, 'steps': 30065, 'loss/train': 1.391595721244812} 02/25/2022 16:32:21 - INFO - codeparrot_training - Step 30066: {'lr': 0.0001842527206452137, 'samples': 15394304, 'steps': 30066, 'loss/train': 0.05137182027101517} 02/25/2022 16:32:24 - INFO - codeparrot_training - Step 30067: {'lr': 0.00018423693430259897, 'samples': 15394816, 'steps': 30067, 'loss/train': 1.3020931482315063} 02/25/2022 16:32:30 - INFO - codeparrot_training - Step 30068: {'lr': 0.00018422114824169234, 'samples': 15395328, 'steps': 30068, 'loss/train': 2.1964986324310303} 02/25/2022 16:32:33 - INFO - codeparrot_training - Step 30069: {'lr': 0.00018420536246256138, 'samples': 15395840, 'steps': 30069, 'loss/train': 1.3870564699172974} 02/25/2022 16:32:39 - INFO - codeparrot_training - Step 30070: {'lr': 0.00018418957696527373, 'samples': 15396352, 'steps': 30070, 'loss/train': 0.6149616837501526} 02/25/2022 16:32:42 - INFO - codeparrot_training - Step 30071: {'lr': 0.00018417379174989708, 'samples': 15396864, 'steps': 30071, 'loss/train': 2.8009674549102783} 02/25/2022 16:32:48 - INFO - codeparrot_training - Step 30072: {'lr': 0.00018415800681649902, 'samples': 15397376, 'steps': 30072, 'loss/train': 1.3102842569351196} 02/25/2022 16:32:51 - INFO - codeparrot_training - Step 30073: {'lr': 0.00018414222216514708, 'samples': 15397888, 'steps': 30073, 'loss/train': 2.7271907329559326} 02/25/2022 16:32:58 - INFO - codeparrot_training - Step 30074: {'lr': 0.000184126437795909, 'samples': 15398400, 'steps': 30074, 'loss/train': 1.596235990524292} 02/25/2022 16:33:01 - INFO - codeparrot_training - Step 30075: {'lr': 0.00018411065370885237, 'samples': 15398912, 'steps': 30075, 'loss/train': 1.2019269466400146} 02/25/2022 16:33:07 - INFO - codeparrot_training - Step 30076: {'lr': 0.00018409486990404473, 'samples': 15399424, 'steps': 30076, 'loss/train': 1.3568345308303833} 02/25/2022 16:33:10 - INFO - codeparrot_training - Step 30077: {'lr': 0.00018407908638155374, 'samples': 15399936, 'steps': 30077, 'loss/train': 2.513878583908081} 02/25/2022 16:33:16 - INFO - codeparrot_training - Step 30078: {'lr': 0.000184063303141447, 'samples': 15400448, 'steps': 30078, 'loss/train': 1.8428399562835693} 02/25/2022 16:33:19 - INFO - codeparrot_training - Step 30079: {'lr': 0.00018404752018379213, 'samples': 15400960, 'steps': 30079, 'loss/train': 1.758124828338623} 02/25/2022 16:33:25 - INFO - codeparrot_training - Step 30080: {'lr': 0.00018403173750865685, 'samples': 15401472, 'steps': 30080, 'loss/train': 1.529345989227295} 02/25/2022 16:33:28 - INFO - codeparrot_training - Step 30081: {'lr': 0.00018401595511610854, 'samples': 15401984, 'steps': 30081, 'loss/train': 1.8872168064117432} 02/25/2022 16:33:34 - INFO - codeparrot_training - Step 30082: {'lr': 0.00018400017300621493, 'samples': 15402496, 'steps': 30082, 'loss/train': 1.0820785760879517} 02/25/2022 16:33:37 - INFO - codeparrot_training - Step 30083: {'lr': 0.00018398439117904363, 'samples': 15403008, 'steps': 30083, 'loss/train': 2.262575626373291} 02/25/2022 16:33:44 - INFO - codeparrot_training - Step 30084: {'lr': 0.00018396860963466234, 'samples': 15403520, 'steps': 30084, 'loss/train': 1.8471931219100952} 02/25/2022 16:33:48 - INFO - codeparrot_training - Step 30085: {'lr': 0.00018395282837313845, 'samples': 15404032, 'steps': 30085, 'loss/train': 1.480950951576233} 02/25/2022 16:33:53 - INFO - codeparrot_training - Step 30086: {'lr': 0.0001839370473945397, 'samples': 15404544, 'steps': 30086, 'loss/train': 2.2740652561187744} 02/25/2022 16:33:57 - INFO - codeparrot_training - Step 30087: {'lr': 0.0001839212666989336, 'samples': 15405056, 'steps': 30087, 'loss/train': 1.5621719360351562} 02/25/2022 16:34:02 - INFO - codeparrot_training - Step 30088: {'lr': 0.00018390548628638797, 'samples': 15405568, 'steps': 30088, 'loss/train': 0.9227740168571472} 02/25/2022 16:34:06 - INFO - codeparrot_training - Step 30089: {'lr': 0.00018388970615697013, 'samples': 15406080, 'steps': 30089, 'loss/train': 1.8394383192062378} 02/25/2022 16:34:12 - INFO - codeparrot_training - Step 30090: {'lr': 0.00018387392631074778, 'samples': 15406592, 'steps': 30090, 'loss/train': 1.1881377696990967} 02/25/2022 16:34:15 - INFO - codeparrot_training - Step 30091: {'lr': 0.00018385814674778855, 'samples': 15407104, 'steps': 30091, 'loss/train': 1.5937190055847168} 02/25/2022 16:34:21 - INFO - codeparrot_training - Step 30092: {'lr': 0.00018384236746816002, 'samples': 15407616, 'steps': 30092, 'loss/train': 1.8211866617202759} 02/25/2022 16:34:24 - INFO - codeparrot_training - Step 30093: {'lr': 0.00018382658847192987, 'samples': 15408128, 'steps': 30093, 'loss/train': 0.9190847873687744} 02/25/2022 16:34:30 - INFO - codeparrot_training - Step 30094: {'lr': 0.0001838108097591655, 'samples': 15408640, 'steps': 30094, 'loss/train': 2.038853406906128} 02/25/2022 16:34:33 - INFO - codeparrot_training - Step 30095: {'lr': 0.00018379503132993463, 'samples': 15409152, 'steps': 30095, 'loss/train': 1.6370124816894531} 02/25/2022 16:34:39 - INFO - codeparrot_training - Step 30096: {'lr': 0.00018377925318430478, 'samples': 15409664, 'steps': 30096, 'loss/train': 1.0914748907089233} 02/25/2022 16:34:43 - INFO - codeparrot_training - Step 30097: {'lr': 0.00018376347532234372, 'samples': 15410176, 'steps': 30097, 'loss/train': 1.6376678943634033} 02/25/2022 16:34:48 - INFO - codeparrot_training - Step 30098: {'lr': 0.00018374769774411878, 'samples': 15410688, 'steps': 30098, 'loss/train': 1.219774842262268} 02/25/2022 16:34:52 - INFO - codeparrot_training - Step 30099: {'lr': 0.00018373192044969766, 'samples': 15411200, 'steps': 30099, 'loss/train': 0.8286318778991699} 02/25/2022 16:34:57 - INFO - codeparrot_training - Step 30100: {'lr': 0.00018371614343914796, 'samples': 15411712, 'steps': 30100, 'loss/train': 2.9472908973693848} 02/25/2022 16:35:01 - INFO - codeparrot_training - Step 30101: {'lr': 0.0001837003667125373, 'samples': 15412224, 'steps': 30101, 'loss/train': 1.432822823524475} 02/25/2022 16:35:06 - INFO - codeparrot_training - Step 30102: {'lr': 0.00018368459026993323, 'samples': 15412736, 'steps': 30102, 'loss/train': 2.500108242034912} 02/25/2022 16:35:10 - INFO - codeparrot_training - Step 30103: {'lr': 0.00018366881411140327, 'samples': 15413248, 'steps': 30103, 'loss/train': 2.1952555179595947} 02/25/2022 16:35:15 - INFO - codeparrot_training - Step 30104: {'lr': 0.00018365303823701502, 'samples': 15413760, 'steps': 30104, 'loss/train': 1.3362709283828735} 02/25/2022 16:35:19 - INFO - codeparrot_training - Step 30105: {'lr': 0.00018363726264683617, 'samples': 15414272, 'steps': 30105, 'loss/train': 2.6070902347564697} 02/25/2022 16:35:25 - INFO - codeparrot_training - Step 30106: {'lr': 0.0001836214873409342, 'samples': 15414784, 'steps': 30106, 'loss/train': 1.4674382209777832} 02/25/2022 16:35:28 - INFO - codeparrot_training - Step 30107: {'lr': 0.00018360571231937662, 'samples': 15415296, 'steps': 30107, 'loss/train': 1.2712163925170898} 02/25/2022 16:35:34 - INFO - codeparrot_training - Step 30108: {'lr': 0.00018358993758223118, 'samples': 15415808, 'steps': 30108, 'loss/train': 8.703932762145996} 02/25/2022 16:35:37 - INFO - codeparrot_training - Step 30109: {'lr': 0.0001835741631295653, 'samples': 15416320, 'steps': 30109, 'loss/train': 2.3264341354370117} 02/25/2022 16:35:43 - INFO - codeparrot_training - Step 30110: {'lr': 0.00018355838896144666, 'samples': 15416832, 'steps': 30110, 'loss/train': 1.7961509227752686} 02/25/2022 16:35:47 - INFO - codeparrot_training - Step 30111: {'lr': 0.00018354261507794273, 'samples': 15417344, 'steps': 30111, 'loss/train': 2.467033624649048} 02/25/2022 16:35:52 - INFO - codeparrot_training - Step 30112: {'lr': 0.00018352684147912118, 'samples': 15417856, 'steps': 30112, 'loss/train': 0.7049136161804199} 02/25/2022 16:35:55 - INFO - codeparrot_training - Step 30113: {'lr': 0.00018351106816504948, 'samples': 15418368, 'steps': 30113, 'loss/train': 0.9948156476020813} 02/25/2022 16:36:01 - INFO - codeparrot_training - Step 30114: {'lr': 0.00018349529513579527, 'samples': 15418880, 'steps': 30114, 'loss/train': 1.4332759380340576} 02/25/2022 16:36:04 - INFO - codeparrot_training - Step 30115: {'lr': 0.0001834795223914262, 'samples': 15419392, 'steps': 30115, 'loss/train': 1.420559287071228} 02/25/2022 16:36:10 - INFO - codeparrot_training - Step 30116: {'lr': 0.0001834637499320096, 'samples': 15419904, 'steps': 30116, 'loss/train': 1.6374807357788086} 02/25/2022 16:36:16 - INFO - codeparrot_training - Step 30117: {'lr': 0.00018344797775761318, 'samples': 15420416, 'steps': 30117, 'loss/train': 2.265371799468994} 02/25/2022 16:36:19 - INFO - codeparrot_training - Step 30118: {'lr': 0.0001834322058683045, 'samples': 15420928, 'steps': 30118, 'loss/train': 2.8234853744506836} 02/25/2022 16:36:25 - INFO - codeparrot_training - Step 30119: {'lr': 0.00018341643426415122, 'samples': 15421440, 'steps': 30119, 'loss/train': 1.768696904182434} 02/25/2022 16:36:28 - INFO - codeparrot_training - Step 30120: {'lr': 0.00018340066294522068, 'samples': 15421952, 'steps': 30120, 'loss/train': 1.665058970451355} 02/25/2022 16:36:35 - INFO - codeparrot_training - Step 30121: {'lr': 0.00018338489191158057, 'samples': 15422464, 'steps': 30121, 'loss/train': 0.6417982578277588} 02/25/2022 16:36:38 - INFO - codeparrot_training - Step 30122: {'lr': 0.0001833691211632984, 'samples': 15422976, 'steps': 30122, 'loss/train': 2.398552894592285} 02/25/2022 16:36:44 - INFO - codeparrot_training - Step 30123: {'lr': 0.0001833533507004419, 'samples': 15423488, 'steps': 30123, 'loss/train': 0.0898883044719696} 02/25/2022 16:36:47 - INFO - codeparrot_training - Step 30124: {'lr': 0.00018333758052307834, 'samples': 15424000, 'steps': 30124, 'loss/train': 1.6271077394485474} 02/25/2022 16:36:53 - INFO - codeparrot_training - Step 30125: {'lr': 0.00018332181063127542, 'samples': 15424512, 'steps': 30125, 'loss/train': 2.1353557109832764} 02/25/2022 16:36:57 - INFO - codeparrot_training - Step 30126: {'lr': 0.00018330604102510074, 'samples': 15425024, 'steps': 30126, 'loss/train': 1.6372706890106201} 02/25/2022 16:37:02 - INFO - codeparrot_training - Step 30127: {'lr': 0.00018329027170462176, 'samples': 15425536, 'steps': 30127, 'loss/train': 1.4246652126312256} 02/25/2022 16:37:06 - INFO - codeparrot_training - Step 30128: {'lr': 0.00018327450266990617, 'samples': 15426048, 'steps': 30128, 'loss/train': 2.0891523361206055} 02/25/2022 16:37:11 - INFO - codeparrot_training - Step 30129: {'lr': 0.00018325873392102131, 'samples': 15426560, 'steps': 30129, 'loss/train': 2.1926770210266113} 02/25/2022 16:37:15 - INFO - codeparrot_training - Step 30130: {'lr': 0.00018324296545803488, 'samples': 15427072, 'steps': 30130, 'loss/train': 0.9406639933586121} 02/25/2022 16:37:21 - INFO - codeparrot_training - Step 30131: {'lr': 0.00018322719728101439, 'samples': 15427584, 'steps': 30131, 'loss/train': 1.8344426155090332} 02/25/2022 16:37:25 - INFO - codeparrot_training - Step 30132: {'lr': 0.00018321142939002743, 'samples': 15428096, 'steps': 30132, 'loss/train': 1.2943284511566162} 02/25/2022 16:37:30 - INFO - codeparrot_training - Step 30133: {'lr': 0.00018319566178514145, 'samples': 15428608, 'steps': 30133, 'loss/train': 2.014268159866333} 02/25/2022 16:37:34 - INFO - codeparrot_training - Step 30134: {'lr': 0.00018317989446642403, 'samples': 15429120, 'steps': 30134, 'loss/train': 0.475665807723999} 02/25/2022 16:37:39 - INFO - codeparrot_training - Step 30135: {'lr': 0.00018316412743394272, 'samples': 15429632, 'steps': 30135, 'loss/train': 2.0615270137786865} 02/25/2022 16:37:43 - INFO - codeparrot_training - Step 30136: {'lr': 0.0001831483606877652, 'samples': 15430144, 'steps': 30136, 'loss/train': 3.243621826171875} 02/25/2022 16:37:48 - INFO - codeparrot_training - Step 30137: {'lr': 0.00018313259422795872, 'samples': 15430656, 'steps': 30137, 'loss/train': 1.5465216636657715} 02/25/2022 16:37:52 - INFO - codeparrot_training - Step 30138: {'lr': 0.000183116828054591, 'samples': 15431168, 'steps': 30138, 'loss/train': 2.214568614959717} 02/25/2022 16:37:57 - INFO - codeparrot_training - Step 30139: {'lr': 0.00018310106216772954, 'samples': 15431680, 'steps': 30139, 'loss/train': 0.936568558216095} 02/25/2022 16:38:01 - INFO - codeparrot_training - Step 30140: {'lr': 0.0001830852965674419, 'samples': 15432192, 'steps': 30140, 'loss/train': 1.6838629245758057} 02/25/2022 16:38:08 - INFO - codeparrot_training - Step 30141: {'lr': 0.00018306953125379568, 'samples': 15432704, 'steps': 30141, 'loss/train': 1.531569480895996} 02/25/2022 16:38:11 - INFO - codeparrot_training - Step 30142: {'lr': 0.00018305376622685827, 'samples': 15433216, 'steps': 30142, 'loss/train': 1.35321044921875} 02/25/2022 16:38:17 - INFO - codeparrot_training - Step 30143: {'lr': 0.00018303800148669724, 'samples': 15433728, 'steps': 30143, 'loss/train': 2.2404191493988037} 02/25/2022 16:38:20 - INFO - codeparrot_training - Step 30144: {'lr': 0.00018302223703338017, 'samples': 15434240, 'steps': 30144, 'loss/train': 1.8038151264190674} 02/25/2022 16:38:26 - INFO - codeparrot_training - Step 30145: {'lr': 0.00018300647286697465, 'samples': 15434752, 'steps': 30145, 'loss/train': 2.1525137424468994} 02/25/2022 16:38:29 - INFO - codeparrot_training - Step 30146: {'lr': 0.00018299070898754804, 'samples': 15435264, 'steps': 30146, 'loss/train': 0.23471519351005554} 02/25/2022 16:38:34 - INFO - codeparrot_training - Step 30147: {'lr': 0.00018297494539516794, 'samples': 15435776, 'steps': 30147, 'loss/train': 1.5956913232803345} 02/25/2022 16:38:38 - INFO - codeparrot_training - Step 30148: {'lr': 0.00018295918208990195, 'samples': 15436288, 'steps': 30148, 'loss/train': 1.5869932174682617} 02/25/2022 16:38:43 - INFO - codeparrot_training - Step 30149: {'lr': 0.00018294341907181756, 'samples': 15436800, 'steps': 30149, 'loss/train': 0.8938074111938477} 02/25/2022 16:38:47 - INFO - codeparrot_training - Step 30150: {'lr': 0.00018292765634098218, 'samples': 15437312, 'steps': 30150, 'loss/train': 1.9406918287277222} 02/25/2022 16:38:53 - INFO - codeparrot_training - Step 30151: {'lr': 0.00018291189389746344, 'samples': 15437824, 'steps': 30151, 'loss/train': 2.3271782398223877} 02/25/2022 16:38:57 - INFO - codeparrot_training - Step 30152: {'lr': 0.00018289613174132888, 'samples': 15438336, 'steps': 30152, 'loss/train': 1.7094000577926636} 02/25/2022 16:39:02 - INFO - codeparrot_training - Step 30153: {'lr': 0.00018288036987264595, 'samples': 15438848, 'steps': 30153, 'loss/train': 1.9177099466323853} 02/25/2022 16:39:06 - INFO - codeparrot_training - Step 30154: {'lr': 0.00018286460829148224, 'samples': 15439360, 'steps': 30154, 'loss/train': 1.5429911613464355} 02/25/2022 16:39:11 - INFO - codeparrot_training - Step 30155: {'lr': 0.00018284884699790524, 'samples': 15439872, 'steps': 30155, 'loss/train': 1.754862904548645} 02/25/2022 16:39:15 - INFO - codeparrot_training - Step 30156: {'lr': 0.0001828330859919824, 'samples': 15440384, 'steps': 30156, 'loss/train': 1.721798300743103} 02/25/2022 16:39:20 - INFO - codeparrot_training - Step 30157: {'lr': 0.00018281732527378132, 'samples': 15440896, 'steps': 30157, 'loss/train': 1.8194557428359985} 02/25/2022 16:39:24 - INFO - codeparrot_training - Step 30158: {'lr': 0.0001828015648433695, 'samples': 15441408, 'steps': 30158, 'loss/train': 0.8753766417503357} 02/25/2022 16:39:29 - INFO - codeparrot_training - Step 30159: {'lr': 0.00018278580470081443, 'samples': 15441920, 'steps': 30159, 'loss/train': 1.3507989645004272} 02/25/2022 16:39:33 - INFO - codeparrot_training - Step 30160: {'lr': 0.00018277004484618358, 'samples': 15442432, 'steps': 30160, 'loss/train': 1.6233487129211426} 02/25/2022 16:39:39 - INFO - codeparrot_training - Step 30161: {'lr': 0.00018275428527954457, 'samples': 15442944, 'steps': 30161, 'loss/train': 1.5625134706497192} 02/25/2022 16:39:45 - INFO - codeparrot_training - Step 30162: {'lr': 0.0001827385260009648, 'samples': 15443456, 'steps': 30162, 'loss/train': 0.8783797025680542} 02/25/2022 16:39:48 - INFO - codeparrot_training - Step 30163: {'lr': 0.00018272276701051192, 'samples': 15443968, 'steps': 30163, 'loss/train': 1.6139253377914429} 02/25/2022 16:39:52 - INFO - codeparrot_training - Step 30164: {'lr': 0.00018270700830825325, 'samples': 15444480, 'steps': 30164, 'loss/train': 1.8954344987869263} 02/25/2022 16:39:57 - INFO - codeparrot_training - Step 30165: {'lr': 0.00018269124989425643, 'samples': 15444992, 'steps': 30165, 'loss/train': 1.599933385848999} 02/25/2022 16:40:01 - INFO - codeparrot_training - Step 30166: {'lr': 0.00018267549176858889, 'samples': 15445504, 'steps': 30166, 'loss/train': 1.280036211013794} 02/25/2022 16:40:06 - INFO - codeparrot_training - Step 30167: {'lr': 0.00018265973393131823, 'samples': 15446016, 'steps': 30167, 'loss/train': 0.6672527194023132} 02/25/2022 16:40:10 - INFO - codeparrot_training - Step 30168: {'lr': 0.00018264397638251183, 'samples': 15446528, 'steps': 30168, 'loss/train': 1.7261629104614258} 02/25/2022 16:40:15 - INFO - codeparrot_training - Step 30169: {'lr': 0.00018262821912223726, 'samples': 15447040, 'steps': 30169, 'loss/train': 1.4885536432266235} 02/25/2022 16:40:19 - INFO - codeparrot_training - Step 30170: {'lr': 0.00018261246215056198, 'samples': 15447552, 'steps': 30170, 'loss/train': 1.1082655191421509} 02/25/2022 16:40:25 - INFO - codeparrot_training - Step 30171: {'lr': 0.00018259670546755364, 'samples': 15448064, 'steps': 30171, 'loss/train': 0.9089599251747131} 02/25/2022 16:40:28 - INFO - codeparrot_training - Step 30172: {'lr': 0.0001825809490732795, 'samples': 15448576, 'steps': 30172, 'loss/train': 1.4234005212783813} 02/25/2022 16:40:34 - INFO - codeparrot_training - Step 30173: {'lr': 0.00018256519296780714, 'samples': 15449088, 'steps': 30173, 'loss/train': 1.1667343378067017} 02/25/2022 16:40:37 - INFO - codeparrot_training - Step 30174: {'lr': 0.0001825494371512041, 'samples': 15449600, 'steps': 30174, 'loss/train': 1.395885705947876} 02/25/2022 16:40:43 - INFO - codeparrot_training - Step 30175: {'lr': 0.00018253368162353783, 'samples': 15450112, 'steps': 30175, 'loss/train': 0.7092799544334412} 02/25/2022 16:40:46 - INFO - codeparrot_training - Step 30176: {'lr': 0.00018251792638487597, 'samples': 15450624, 'steps': 30176, 'loss/train': 2.2143876552581787} 02/25/2022 16:40:53 - INFO - codeparrot_training - Step 30177: {'lr': 0.00018250217143528578, 'samples': 15451136, 'steps': 30177, 'loss/train': 2.141566514968872} 02/25/2022 16:40:56 - INFO - codeparrot_training - Step 30178: {'lr': 0.00018248641677483485, 'samples': 15451648, 'steps': 30178, 'loss/train': 0.09129561483860016} 02/25/2022 16:41:02 - INFO - codeparrot_training - Step 30179: {'lr': 0.00018247066240359067, 'samples': 15452160, 'steps': 30179, 'loss/train': 2.096642017364502} 02/25/2022 16:41:05 - INFO - codeparrot_training - Step 30180: {'lr': 0.00018245490832162085, 'samples': 15452672, 'steps': 30180, 'loss/train': 2.144042730331421} 02/25/2022 16:41:11 - INFO - codeparrot_training - Step 30181: {'lr': 0.00018243915452899264, 'samples': 15453184, 'steps': 30181, 'loss/train': 1.9913325309753418} 02/25/2022 16:41:14 - INFO - codeparrot_training - Step 30182: {'lr': 0.00018242340102577365, 'samples': 15453696, 'steps': 30182, 'loss/train': 1.9597300291061401} 02/25/2022 16:41:20 - INFO - codeparrot_training - Step 30183: {'lr': 0.00018240764781203133, 'samples': 15454208, 'steps': 30183, 'loss/train': 1.1077646017074585} 02/25/2022 16:41:23 - INFO - codeparrot_training - Step 30184: {'lr': 0.0001823918948878333, 'samples': 15454720, 'steps': 30184, 'loss/train': 2.382155656814575} 02/25/2022 16:41:29 - INFO - codeparrot_training - Step 30185: {'lr': 0.00018237614225324682, 'samples': 15455232, 'steps': 30185, 'loss/train': 1.6744897365570068} 02/25/2022 16:41:32 - INFO - codeparrot_training - Step 30186: {'lr': 0.00018236038990833948, 'samples': 15455744, 'steps': 30186, 'loss/train': 1.266266107559204} 02/25/2022 16:41:38 - INFO - codeparrot_training - Step 30187: {'lr': 0.00018234463785317872, 'samples': 15456256, 'steps': 30187, 'loss/train': 1.7273356914520264} 02/25/2022 16:41:41 - INFO - codeparrot_training - Step 30188: {'lr': 0.00018232888608783217, 'samples': 15456768, 'steps': 30188, 'loss/train': 1.1520874500274658} 02/25/2022 16:41:48 - INFO - codeparrot_training - Step 30189: {'lr': 0.00018231313461236713, 'samples': 15457280, 'steps': 30189, 'loss/train': 1.7684825658798218} 02/25/2022 16:41:53 - INFO - codeparrot_training - Step 30190: {'lr': 0.00018229738342685105, 'samples': 15457792, 'steps': 30190, 'loss/train': 0.6701033711433411} 02/25/2022 16:41:57 - INFO - codeparrot_training - Step 30191: {'lr': 0.00018228163253135155, 'samples': 15458304, 'steps': 30191, 'loss/train': 1.2233402729034424} 02/25/2022 16:42:02 - INFO - codeparrot_training - Step 30192: {'lr': 0.00018226588192593604, 'samples': 15458816, 'steps': 30192, 'loss/train': 1.1996300220489502} 02/25/2022 16:42:06 - INFO - codeparrot_training - Step 30193: {'lr': 0.000182250131610672, 'samples': 15459328, 'steps': 30193, 'loss/train': 1.3650349378585815} 02/25/2022 16:42:11 - INFO - codeparrot_training - Step 30194: {'lr': 0.00018223438158562684, 'samples': 15459840, 'steps': 30194, 'loss/train': 0.8478029370307922} 02/25/2022 16:42:15 - INFO - codeparrot_training - Step 30195: {'lr': 0.0001822186318508681, 'samples': 15460352, 'steps': 30195, 'loss/train': 1.6683921813964844} 02/25/2022 16:42:21 - INFO - codeparrot_training - Step 30196: {'lr': 0.00018220288240646315, 'samples': 15460864, 'steps': 30196, 'loss/train': 0.8926817774772644} 02/25/2022 16:42:25 - INFO - codeparrot_training - Step 30197: {'lr': 0.00018218713325247963, 'samples': 15461376, 'steps': 30197, 'loss/train': 2.3401453495025635} 02/25/2022 16:42:30 - INFO - codeparrot_training - Step 30198: {'lr': 0.00018217138438898482, 'samples': 15461888, 'steps': 30198, 'loss/train': 2.5961201190948486} 02/25/2022 16:42:34 - INFO - codeparrot_training - Step 30199: {'lr': 0.0001821556358160463, 'samples': 15462400, 'steps': 30199, 'loss/train': 0.856624960899353} 02/25/2022 16:42:39 - INFO - codeparrot_training - Step 30200: {'lr': 0.00018213988753373146, 'samples': 15462912, 'steps': 30200, 'loss/train': 2.5985987186431885} 02/25/2022 16:42:43 - INFO - codeparrot_training - Step 30201: {'lr': 0.00018212413954210777, 'samples': 15463424, 'steps': 30201, 'loss/train': 2.1162166595458984} 02/25/2022 16:42:48 - INFO - codeparrot_training - Step 30202: {'lr': 0.00018210839184124277, 'samples': 15463936, 'steps': 30202, 'loss/train': 0.44784021377563477} 02/25/2022 16:42:52 - INFO - codeparrot_training - Step 30203: {'lr': 0.00018209264443120392, 'samples': 15464448, 'steps': 30203, 'loss/train': 1.6257777214050293} 02/25/2022 16:42:57 - INFO - codeparrot_training - Step 30204: {'lr': 0.00018207689731205852, 'samples': 15464960, 'steps': 30204, 'loss/train': 1.9447648525238037} 02/25/2022 16:43:01 - INFO - codeparrot_training - Step 30205: {'lr': 0.0001820611504838741, 'samples': 15465472, 'steps': 30205, 'loss/train': 1.8288772106170654} 02/25/2022 16:43:06 - INFO - codeparrot_training - Step 30206: {'lr': 0.00018204540394671826, 'samples': 15465984, 'steps': 30206, 'loss/train': 1.1476339101791382} 02/25/2022 16:43:10 - INFO - codeparrot_training - Step 30207: {'lr': 0.00018202965770065823, 'samples': 15466496, 'steps': 30207, 'loss/train': 2.9437742233276367} 02/25/2022 16:43:15 - INFO - codeparrot_training - Step 30208: {'lr': 0.0001820139117457616, 'samples': 15467008, 'steps': 30208, 'loss/train': 1.3765581846237183} 02/25/2022 16:43:19 - INFO - codeparrot_training - Step 30209: {'lr': 0.00018199816608209575, 'samples': 15467520, 'steps': 30209, 'loss/train': 2.2918102741241455} 02/25/2022 16:43:24 - INFO - codeparrot_training - Step 30210: {'lr': 0.00018198242070972814, 'samples': 15468032, 'steps': 30210, 'loss/train': 0.5545539259910583} 02/25/2022 16:43:28 - INFO - codeparrot_training - Step 30211: {'lr': 0.00018196667562872637, 'samples': 15468544, 'steps': 30211, 'loss/train': 2.459815740585327} 02/25/2022 16:43:34 - INFO - codeparrot_training - Step 30212: {'lr': 0.00018195093083915766, 'samples': 15469056, 'steps': 30212, 'loss/train': 1.0434744358062744} 02/25/2022 16:43:38 - INFO - codeparrot_training - Step 30213: {'lr': 0.00018193518634108954, 'samples': 15469568, 'steps': 30213, 'loss/train': 0.13225218653678894} 02/25/2022 16:43:43 - INFO - codeparrot_training - Step 30214: {'lr': 0.0001819194421345895, 'samples': 15470080, 'steps': 30214, 'loss/train': 0.9162151217460632} 02/25/2022 16:43:46 - INFO - codeparrot_training - Step 30215: {'lr': 0.00018190369821972503, 'samples': 15470592, 'steps': 30215, 'loss/train': 2.204357147216797} 02/25/2022 16:43:52 - INFO - codeparrot_training - Step 30216: {'lr': 0.00018188795459656338, 'samples': 15471104, 'steps': 30216, 'loss/train': 2.0624074935913086} 02/25/2022 16:43:56 - INFO - codeparrot_training - Step 30217: {'lr': 0.00018187221126517213, 'samples': 15471616, 'steps': 30217, 'loss/train': 1.7415975332260132} 02/25/2022 16:44:01 - INFO - codeparrot_training - Step 30218: {'lr': 0.0001818564682256187, 'samples': 15472128, 'steps': 30218, 'loss/train': 1.7843259572982788} 02/25/2022 16:44:05 - INFO - codeparrot_training - Step 30219: {'lr': 0.0001818407254779706, 'samples': 15472640, 'steps': 30219, 'loss/train': 1.7959237098693848} 02/25/2022 16:44:10 - INFO - codeparrot_training - Step 30220: {'lr': 0.00018182498302229507, 'samples': 15473152, 'steps': 30220, 'loss/train': 1.3945038318634033} 02/25/2022 16:44:14 - INFO - codeparrot_training - Step 30221: {'lr': 0.00018180924085865968, 'samples': 15473664, 'steps': 30221, 'loss/train': 0.21227788925170898} 02/25/2022 16:44:20 - INFO - codeparrot_training - Step 30222: {'lr': 0.00018179349898713189, 'samples': 15474176, 'steps': 30222, 'loss/train': 1.6112300157546997} 02/25/2022 16:44:23 - INFO - codeparrot_training - Step 30223: {'lr': 0.000181777757407779, 'samples': 15474688, 'steps': 30223, 'loss/train': 1.6448408365249634} 02/25/2022 16:44:29 - INFO - codeparrot_training - Step 30224: {'lr': 0.00018176201612066874, 'samples': 15475200, 'steps': 30224, 'loss/train': 0.7363759279251099} 02/25/2022 16:44:32 - INFO - codeparrot_training - Step 30225: {'lr': 0.00018174627512586818, 'samples': 15475712, 'steps': 30225, 'loss/train': 1.2504916191101074} 02/25/2022 16:44:38 - INFO - codeparrot_training - Step 30226: {'lr': 0.00018173053442344493, 'samples': 15476224, 'steps': 30226, 'loss/train': 0.6949149966239929} 02/25/2022 16:44:41 - INFO - codeparrot_training - Step 30227: {'lr': 0.00018171479401346636, 'samples': 15476736, 'steps': 30227, 'loss/train': 2.5589053630828857} 02/25/2022 16:44:47 - INFO - codeparrot_training - Step 30228: {'lr': 0.00018169905389600006, 'samples': 15477248, 'steps': 30228, 'loss/train': 2.4627411365509033} 02/25/2022 16:44:51 - INFO - codeparrot_training - Step 30229: {'lr': 0.00018168331407111322, 'samples': 15477760, 'steps': 30229, 'loss/train': 2.2167105674743652} 02/25/2022 16:44:56 - INFO - codeparrot_training - Step 30230: {'lr': 0.00018166757453887334, 'samples': 15478272, 'steps': 30230, 'loss/train': 1.3010408878326416} 02/25/2022 16:44:59 - INFO - codeparrot_training - Step 30231: {'lr': 0.0001816518352993479, 'samples': 15478784, 'steps': 30231, 'loss/train': 1.871283769607544} 02/25/2022 16:45:06 - INFO - codeparrot_training - Step 30232: {'lr': 0.00018163609635260436, 'samples': 15479296, 'steps': 30232, 'loss/train': 2.2058842182159424} 02/25/2022 16:45:10 - INFO - codeparrot_training - Step 30233: {'lr': 0.00018162035769871002, 'samples': 15479808, 'steps': 30233, 'loss/train': 3.6509897708892822} 02/25/2022 16:45:15 - INFO - codeparrot_training - Step 30234: {'lr': 0.00018160461933773234, 'samples': 15480320, 'steps': 30234, 'loss/train': 1.5391018390655518} 02/25/2022 16:45:19 - INFO - codeparrot_training - Step 30235: {'lr': 0.0001815888812697388, 'samples': 15480832, 'steps': 30235, 'loss/train': 2.2007896900177} 02/25/2022 16:45:24 - INFO - codeparrot_training - Step 30236: {'lr': 0.00018157314349479672, 'samples': 15481344, 'steps': 30236, 'loss/train': 2.0087921619415283} 02/25/2022 16:45:28 - INFO - codeparrot_training - Step 30237: {'lr': 0.0001815574060129736, 'samples': 15481856, 'steps': 30237, 'loss/train': 1.4928714036941528} 02/25/2022 16:45:33 - INFO - codeparrot_training - Step 30238: {'lr': 0.0001815416688243368, 'samples': 15482368, 'steps': 30238, 'loss/train': 1.8697328567504883} 02/25/2022 16:45:37 - INFO - codeparrot_training - Step 30239: {'lr': 0.00018152593192895377, 'samples': 15482880, 'steps': 30239, 'loss/train': 2.0418519973754883} 02/25/2022 16:45:42 - INFO - codeparrot_training - Step 30240: {'lr': 0.00018151019532689188, 'samples': 15483392, 'steps': 30240, 'loss/train': 1.92549467086792} 02/25/2022 16:45:46 - INFO - codeparrot_training - Step 30241: {'lr': 0.00018149445901821862, 'samples': 15483904, 'steps': 30241, 'loss/train': 1.6473177671432495} 02/25/2022 16:45:51 - INFO - codeparrot_training - Step 30242: {'lr': 0.00018147872300300127, 'samples': 15484416, 'steps': 30242, 'loss/train': 1.3054625988006592} 02/25/2022 16:45:55 - INFO - codeparrot_training - Step 30243: {'lr': 0.0001814629872813074, 'samples': 15484928, 'steps': 30243, 'loss/train': 1.6677675247192383} 02/25/2022 16:46:02 - INFO - codeparrot_training - Step 30244: {'lr': 0.00018144725185320423, 'samples': 15485440, 'steps': 30244, 'loss/train': 1.9955741167068481} 02/25/2022 16:46:05 - INFO - codeparrot_training - Step 30245: {'lr': 0.00018143151671875935, 'samples': 15485952, 'steps': 30245, 'loss/train': 1.2038229703903198} 02/25/2022 16:46:11 - INFO - codeparrot_training - Step 30246: {'lr': 0.00018141578187804008, 'samples': 15486464, 'steps': 30246, 'loss/train': 2.1339006423950195} 02/25/2022 16:46:14 - INFO - codeparrot_training - Step 30247: {'lr': 0.00018140004733111376, 'samples': 15486976, 'steps': 30247, 'loss/train': 0.936119794845581} 02/25/2022 16:46:20 - INFO - codeparrot_training - Step 30248: {'lr': 0.00018138431307804784, 'samples': 15487488, 'steps': 30248, 'loss/train': 2.3088300228118896} 02/25/2022 16:46:23 - INFO - codeparrot_training - Step 30249: {'lr': 0.00018136857911890976, 'samples': 15488000, 'steps': 30249, 'loss/train': 2.0716497898101807} 02/25/2022 16:46:29 - INFO - codeparrot_training - Step 30250: {'lr': 0.000181352845453767, 'samples': 15488512, 'steps': 30250, 'loss/train': 1.789767861366272} 02/25/2022 16:46:32 - INFO - codeparrot_training - Step 30251: {'lr': 0.00018133711208268672, 'samples': 15489024, 'steps': 30251, 'loss/train': 2.851513624191284} 02/25/2022 16:46:38 - INFO - codeparrot_training - Step 30252: {'lr': 0.00018132137900573648, 'samples': 15489536, 'steps': 30252, 'loss/train': 1.2172852754592896} 02/25/2022 16:46:44 - INFO - codeparrot_training - Step 30253: {'lr': 0.0001813056462229836, 'samples': 15490048, 'steps': 30253, 'loss/train': 3.98465633392334} 02/25/2022 16:46:47 - INFO - codeparrot_training - Step 30254: {'lr': 0.00018128991373449565, 'samples': 15490560, 'steps': 30254, 'loss/train': 1.589063048362732} 02/25/2022 16:46:53 - INFO - codeparrot_training - Step 30255: {'lr': 0.00018127418154033982, 'samples': 15491072, 'steps': 30255, 'loss/train': 1.040554165840149} 02/25/2022 16:46:57 - INFO - codeparrot_training - Step 30256: {'lr': 0.00018125844964058353, 'samples': 15491584, 'steps': 30256, 'loss/train': 0.8568814992904663} 02/25/2022 16:47:00 - INFO - codeparrot_training - Step 30257: {'lr': 0.00018124271803529425, 'samples': 15492096, 'steps': 30257, 'loss/train': 2.070145606994629} 02/25/2022 16:47:06 - INFO - codeparrot_training - Step 30258: {'lr': 0.00018122698672453939, 'samples': 15492608, 'steps': 30258, 'loss/train': 2.6813528537750244} 02/25/2022 16:47:12 - INFO - codeparrot_training - Step 30259: {'lr': 0.0001812112557083862, 'samples': 15493120, 'steps': 30259, 'loss/train': 2.105872869491577} 02/25/2022 16:47:16 - INFO - codeparrot_training - Step 30260: {'lr': 0.00018119552498690214, 'samples': 15493632, 'steps': 30260, 'loss/train': 1.447824239730835} 02/25/2022 16:47:19 - INFO - codeparrot_training - Step 30261: {'lr': 0.0001811797945601546, 'samples': 15494144, 'steps': 30261, 'loss/train': 2.2766470909118652} 02/25/2022 16:47:23 - INFO - codeparrot_training - Step 30262: {'lr': 0.000181164064428211, 'samples': 15494656, 'steps': 30262, 'loss/train': 2.3196182250976562} 02/25/2022 16:47:28 - INFO - codeparrot_training - Step 30263: {'lr': 0.00018114833459113877, 'samples': 15495168, 'steps': 30263, 'loss/train': 1.186671257019043} 02/25/2022 16:47:32 - INFO - codeparrot_training - Step 30264: {'lr': 0.00018113260504900512, 'samples': 15495680, 'steps': 30264, 'loss/train': 1.2272653579711914} 02/25/2022 16:47:37 - INFO - codeparrot_training - Step 30265: {'lr': 0.00018111687580187752, 'samples': 15496192, 'steps': 30265, 'loss/train': 0.9229240417480469} 02/25/2022 16:47:41 - INFO - codeparrot_training - Step 30266: {'lr': 0.00018110114684982332, 'samples': 15496704, 'steps': 30266, 'loss/train': 2.104604721069336} 02/25/2022 16:47:46 - INFO - codeparrot_training - Step 30267: {'lr': 0.0001810854181929101, 'samples': 15497216, 'steps': 30267, 'loss/train': 1.3323432207107544} 02/25/2022 16:47:50 - INFO - codeparrot_training - Step 30268: {'lr': 0.00018106968983120486, 'samples': 15497728, 'steps': 30268, 'loss/train': 1.8889106512069702} 02/25/2022 16:47:56 - INFO - codeparrot_training - Step 30269: {'lr': 0.00018105396176477527, 'samples': 15498240, 'steps': 30269, 'loss/train': 1.221634030342102} 02/25/2022 16:47:59 - INFO - codeparrot_training - Step 30270: {'lr': 0.00018103823399368859, 'samples': 15498752, 'steps': 30270, 'loss/train': 1.6162779331207275} 02/25/2022 16:48:05 - INFO - codeparrot_training - Step 30271: {'lr': 0.0001810225065180122, 'samples': 15499264, 'steps': 30271, 'loss/train': 2.724073648452759} 02/25/2022 16:48:08 - INFO - codeparrot_training - Step 30272: {'lr': 0.00018100677933781362, 'samples': 15499776, 'steps': 30272, 'loss/train': 2.6476962566375732} 02/25/2022 16:48:14 - INFO - codeparrot_training - Step 30273: {'lr': 0.00018099105245315995, 'samples': 15500288, 'steps': 30273, 'loss/train': 1.5384633541107178} 02/25/2022 16:48:17 - INFO - codeparrot_training - Step 30274: {'lr': 0.00018097532586411872, 'samples': 15500800, 'steps': 30274, 'loss/train': 1.2873787879943848} 02/25/2022 16:48:23 - INFO - codeparrot_training - Step 30275: {'lr': 0.0001809595995707573, 'samples': 15501312, 'steps': 30275, 'loss/train': 1.1256585121154785} 02/25/2022 16:48:26 - INFO - codeparrot_training - Step 30276: {'lr': 0.0001809438735731431, 'samples': 15501824, 'steps': 30276, 'loss/train': 1.6302610635757446} 02/25/2022 16:48:32 - INFO - codeparrot_training - Step 30277: {'lr': 0.00018092814787134333, 'samples': 15502336, 'steps': 30277, 'loss/train': 1.5706431865692139} 02/25/2022 16:48:35 - INFO - codeparrot_training - Step 30278: {'lr': 0.00018091242246542544, 'samples': 15502848, 'steps': 30278, 'loss/train': 2.1778671741485596} 02/25/2022 16:48:41 - INFO - codeparrot_training - Step 30279: {'lr': 0.00018089669735545682, 'samples': 15503360, 'steps': 30279, 'loss/train': 1.2678357362747192} 02/25/2022 16:48:45 - INFO - codeparrot_training - Step 30280: {'lr': 0.00018088097254150484, 'samples': 15503872, 'steps': 30280, 'loss/train': 1.8698092699050903} 02/25/2022 16:48:50 - INFO - codeparrot_training - Step 30281: {'lr': 0.00018086524802363678, 'samples': 15504384, 'steps': 30281, 'loss/train': 1.727962613105774} 02/25/2022 16:48:54 - INFO - codeparrot_training - Step 30282: {'lr': 0.00018084952380192, 'samples': 15504896, 'steps': 30282, 'loss/train': 1.9903841018676758} 02/25/2022 16:48:59 - INFO - codeparrot_training - Step 30283: {'lr': 0.00018083379987642202, 'samples': 15505408, 'steps': 30283, 'loss/train': 2.727038860321045} 02/25/2022 16:49:03 - INFO - codeparrot_training - Step 30284: {'lr': 0.00018081807624720998, 'samples': 15505920, 'steps': 30284, 'loss/train': 1.9048691987991333} 02/25/2022 16:49:08 - INFO - codeparrot_training - Step 30285: {'lr': 0.00018080235291435137, 'samples': 15506432, 'steps': 30285, 'loss/train': 1.6210334300994873} 02/25/2022 16:49:12 - INFO - codeparrot_training - Step 30286: {'lr': 0.00018078662987791355, 'samples': 15506944, 'steps': 30286, 'loss/train': 1.8738731145858765} 02/25/2022 16:49:18 - INFO - codeparrot_training - Step 30287: {'lr': 0.00018077090713796372, 'samples': 15507456, 'steps': 30287, 'loss/train': 1.5903934240341187} 02/25/2022 16:49:21 - INFO - codeparrot_training - Step 30288: {'lr': 0.00018075518469456943, 'samples': 15507968, 'steps': 30288, 'loss/train': 0.026191094890236855} 02/25/2022 16:49:27 - INFO - codeparrot_training - Step 30289: {'lr': 0.00018073946254779793, 'samples': 15508480, 'steps': 30289, 'loss/train': 1.4007740020751953} 02/25/2022 16:49:33 - INFO - codeparrot_training - Step 30290: {'lr': 0.00018072374069771658, 'samples': 15508992, 'steps': 30290, 'loss/train': 1.5824601650238037} 02/25/2022 16:49:36 - INFO - codeparrot_training - Step 30291: {'lr': 0.00018070801914439271, 'samples': 15509504, 'steps': 30291, 'loss/train': 1.3173692226409912} 02/25/2022 16:49:42 - INFO - codeparrot_training - Step 30292: {'lr': 0.00018069229788789364, 'samples': 15510016, 'steps': 30292, 'loss/train': 1.6421961784362793} 02/25/2022 16:49:45 - INFO - codeparrot_training - Step 30293: {'lr': 0.0001806765769282869, 'samples': 15510528, 'steps': 30293, 'loss/train': 0.9650719165802002} 02/25/2022 16:49:51 - INFO - codeparrot_training - Step 30294: {'lr': 0.00018066085626563958, 'samples': 15511040, 'steps': 30294, 'loss/train': 1.5381430387496948} 02/25/2022 16:49:54 - INFO - codeparrot_training - Step 30295: {'lr': 0.00018064513590001913, 'samples': 15511552, 'steps': 30295, 'loss/train': 1.9772381782531738} 02/25/2022 16:50:00 - INFO - codeparrot_training - Step 30296: {'lr': 0.0001806294158314929, 'samples': 15512064, 'steps': 30296, 'loss/train': 1.8971456289291382} 02/25/2022 16:50:03 - INFO - codeparrot_training - Step 30297: {'lr': 0.0001806136960601282, 'samples': 15512576, 'steps': 30297, 'loss/train': 1.6689056158065796} 02/25/2022 16:50:09 - INFO - codeparrot_training - Step 30298: {'lr': 0.00018059797658599253, 'samples': 15513088, 'steps': 30298, 'loss/train': 2.2771003246307373} 02/25/2022 16:50:12 - INFO - codeparrot_training - Step 30299: {'lr': 0.00018058225740915295, 'samples': 15513600, 'steps': 30299, 'loss/train': 0.5486544966697693} 02/25/2022 16:50:18 - INFO - codeparrot_training - Step 30300: {'lr': 0.00018056653852967697, 'samples': 15514112, 'steps': 30300, 'loss/train': 0.883249819278717} 02/25/2022 16:50:22 - INFO - codeparrot_training - Step 30301: {'lr': 0.00018055081994763187, 'samples': 15514624, 'steps': 30301, 'loss/train': 2.0749897956848145} 02/25/2022 16:50:27 - INFO - codeparrot_training - Step 30302: {'lr': 0.00018053510166308513, 'samples': 15515136, 'steps': 30302, 'loss/train': 0.5682500004768372} 02/25/2022 16:50:31 - INFO - codeparrot_training - Step 30303: {'lr': 0.00018051938367610382, 'samples': 15515648, 'steps': 30303, 'loss/train': 0.42886900901794434} 02/25/2022 16:50:36 - INFO - codeparrot_training - Step 30304: {'lr': 0.00018050366598675545, 'samples': 15516160, 'steps': 30304, 'loss/train': 1.0510327816009521} 02/25/2022 16:50:40 - INFO - codeparrot_training - Step 30305: {'lr': 0.00018048794859510726, 'samples': 15516672, 'steps': 30305, 'loss/train': 1.8106976747512817} 02/25/2022 16:50:45 - INFO - codeparrot_training - Step 30306: {'lr': 0.00018047223150122676, 'samples': 15517184, 'steps': 30306, 'loss/train': 1.6108672618865967} 02/25/2022 16:50:49 - INFO - codeparrot_training - Step 30307: {'lr': 0.00018045651470518105, 'samples': 15517696, 'steps': 30307, 'loss/train': 1.9195351600646973} 02/25/2022 16:50:54 - INFO - codeparrot_training - Step 30308: {'lr': 0.00018044079820703752, 'samples': 15518208, 'steps': 30308, 'loss/train': 1.5020885467529297} 02/25/2022 16:50:58 - INFO - codeparrot_training - Step 30309: {'lr': 0.00018042508200686352, 'samples': 15518720, 'steps': 30309, 'loss/train': 0.6649724841117859} 02/25/2022 16:51:03 - INFO - codeparrot_training - Step 30310: {'lr': 0.00018040936610472642, 'samples': 15519232, 'steps': 30310, 'loss/train': 1.338865041732788} 02/25/2022 16:51:07 - INFO - codeparrot_training - Step 30311: {'lr': 0.00018039365050069356, 'samples': 15519744, 'steps': 30311, 'loss/train': 1.6641895771026611} 02/25/2022 16:51:12 - INFO - codeparrot_training - Step 30312: {'lr': 0.00018037793519483212, 'samples': 15520256, 'steps': 30312, 'loss/train': 1.285415530204773} 02/25/2022 16:51:16 - INFO - codeparrot_training - Step 30313: {'lr': 0.0001803622201872095, 'samples': 15520768, 'steps': 30313, 'loss/train': 1.5049984455108643} 02/25/2022 16:51:21 - INFO - codeparrot_training - Step 30314: {'lr': 0.000180346505477893, 'samples': 15521280, 'steps': 30314, 'loss/train': 1.6976279020309448} 02/25/2022 16:51:25 - INFO - codeparrot_training - Step 30315: {'lr': 0.00018033079106695004, 'samples': 15521792, 'steps': 30315, 'loss/train': 1.6904067993164062} 02/25/2022 16:51:31 - INFO - codeparrot_training - Step 30316: {'lr': 0.00018031507695444776, 'samples': 15522304, 'steps': 30316, 'loss/train': 1.3612467050552368} 02/25/2022 16:51:34 - INFO - codeparrot_training - Step 30317: {'lr': 0.00018029936314045358, 'samples': 15522816, 'steps': 30317, 'loss/train': 0.8877971768379211} 02/25/2022 16:51:40 - INFO - codeparrot_training - Step 30318: {'lr': 0.00018028364962503478, 'samples': 15523328, 'steps': 30318, 'loss/train': 2.2065677642822266} 02/25/2022 16:51:43 - INFO - codeparrot_training - Step 30319: {'lr': 0.00018026793640825873, 'samples': 15523840, 'steps': 30319, 'loss/train': 1.7737586498260498} 02/25/2022 16:51:49 - INFO - codeparrot_training - Step 30320: {'lr': 0.0001802522234901927, 'samples': 15524352, 'steps': 30320, 'loss/train': 1.4549020528793335} 02/25/2022 16:51:52 - INFO - codeparrot_training - Step 30321: {'lr': 0.00018023651087090397, 'samples': 15524864, 'steps': 30321, 'loss/train': 1.487540364265442} 02/25/2022 16:51:58 - INFO - codeparrot_training - Step 30322: {'lr': 0.00018022079855045987, 'samples': 15525376, 'steps': 30322, 'loss/train': 2.4746651649475098} 02/25/2022 16:52:01 - INFO - codeparrot_training - Step 30323: {'lr': 0.00018020508652892775, 'samples': 15525888, 'steps': 30323, 'loss/train': 2.464829921722412} 02/25/2022 16:52:07 - INFO - codeparrot_training - Step 30324: {'lr': 0.00018018937480637488, 'samples': 15526400, 'steps': 30324, 'loss/train': 2.0563764572143555} 02/25/2022 16:52:13 - INFO - codeparrot_training - Step 30325: {'lr': 0.0001801736633828685, 'samples': 15526912, 'steps': 30325, 'loss/train': 2.1133432388305664} 02/25/2022 16:52:17 - INFO - codeparrot_training - Step 30326: {'lr': 0.00018015795225847603, 'samples': 15527424, 'steps': 30326, 'loss/train': 1.5929611921310425} 02/25/2022 16:52:22 - INFO - codeparrot_training - Step 30327: {'lr': 0.00018014224143326468, 'samples': 15527936, 'steps': 30327, 'loss/train': 0.8876922726631165} 02/25/2022 16:52:25 - INFO - codeparrot_training - Step 30328: {'lr': 0.00018012653090730183, 'samples': 15528448, 'steps': 30328, 'loss/train': 0.8034305572509766} 02/25/2022 16:52:31 - INFO - codeparrot_training - Step 30329: {'lr': 0.00018011082068065468, 'samples': 15528960, 'steps': 30329, 'loss/train': 1.4819104671478271} 02/25/2022 16:52:34 - INFO - codeparrot_training - Step 30330: {'lr': 0.00018009511075339063, 'samples': 15529472, 'steps': 30330, 'loss/train': 1.6907936334609985} 02/25/2022 16:52:40 - INFO - codeparrot_training - Step 30331: {'lr': 0.00018007940112557688, 'samples': 15529984, 'steps': 30331, 'loss/train': 1.6368550062179565} 02/25/2022 16:52:43 - INFO - codeparrot_training - Step 30332: {'lr': 0.00018006369179728078, 'samples': 15530496, 'steps': 30332, 'loss/train': 2.366891622543335} 02/25/2022 16:52:49 - INFO - codeparrot_training - Step 30333: {'lr': 0.0001800479827685697, 'samples': 15531008, 'steps': 30333, 'loss/train': 0.8492269515991211} 02/25/2022 16:52:52 - INFO - codeparrot_training - Step 30334: {'lr': 0.00018003227403951078, 'samples': 15531520, 'steps': 30334, 'loss/train': 2.5445034503936768} 02/25/2022 16:52:59 - INFO - codeparrot_training - Step 30335: {'lr': 0.00018001656561017135, 'samples': 15532032, 'steps': 30335, 'loss/train': 0.2546338438987732} 02/25/2022 16:53:02 - INFO - codeparrot_training - Step 30336: {'lr': 0.00018000085748061872, 'samples': 15532544, 'steps': 30336, 'loss/train': 1.3956319093704224} 02/25/2022 16:53:08 - INFO - codeparrot_training - Step 30337: {'lr': 0.00017998514965092034, 'samples': 15533056, 'steps': 30337, 'loss/train': 1.7547041177749634} 02/25/2022 16:53:11 - INFO - codeparrot_training - Step 30338: {'lr': 0.0001799694421211432, 'samples': 15533568, 'steps': 30338, 'loss/train': 1.4463893175125122} 02/25/2022 16:53:17 - INFO - codeparrot_training - Step 30339: {'lr': 0.00017995373489135475, 'samples': 15534080, 'steps': 30339, 'loss/train': 2.6473276615142822} 02/25/2022 16:53:20 - INFO - codeparrot_training - Step 30340: {'lr': 0.00017993802796162226, 'samples': 15534592, 'steps': 30340, 'loss/train': 2.3492541313171387} 02/25/2022 16:53:26 - INFO - codeparrot_training - Step 30341: {'lr': 0.0001799223213320131, 'samples': 15535104, 'steps': 30341, 'loss/train': 0.3014780580997467} 02/25/2022 16:53:29 - INFO - codeparrot_training - Step 30342: {'lr': 0.0001799066150025944, 'samples': 15535616, 'steps': 30342, 'loss/train': 1.133432149887085} 02/25/2022 16:53:35 - INFO - codeparrot_training - Step 30343: {'lr': 0.00017989090897343346, 'samples': 15536128, 'steps': 30343, 'loss/train': 1.5439274311065674} 02/25/2022 16:53:38 - INFO - codeparrot_training - Step 30344: {'lr': 0.0001798752032445976, 'samples': 15536640, 'steps': 30344, 'loss/train': 2.585175037384033} 02/25/2022 16:53:45 - INFO - codeparrot_training - Step 30345: {'lr': 0.00017985949781615412, 'samples': 15537152, 'steps': 30345, 'loss/train': 3.327129364013672} 02/25/2022 16:53:48 - INFO - codeparrot_training - Step 30346: {'lr': 0.00017984379268817043, 'samples': 15537664, 'steps': 30346, 'loss/train': 2.091188669204712} 02/25/2022 16:53:54 - INFO - codeparrot_training - Step 30347: {'lr': 0.00017982808786071348, 'samples': 15538176, 'steps': 30347, 'loss/train': 1.7906984090805054} 02/25/2022 16:53:57 - INFO - codeparrot_training - Step 30348: {'lr': 0.00017981238333385075, 'samples': 15538688, 'steps': 30348, 'loss/train': 1.2289783954620361} 02/25/2022 16:54:03 - INFO - codeparrot_training - Step 30349: {'lr': 0.00017979667910764945, 'samples': 15539200, 'steps': 30349, 'loss/train': 0.6899938583374023} 02/25/2022 16:54:06 - INFO - codeparrot_training - Step 30350: {'lr': 0.00017978097518217702, 'samples': 15539712, 'steps': 30350, 'loss/train': 2.473595380783081} 02/25/2022 16:54:12 - INFO - codeparrot_training - Step 30351: {'lr': 0.0001797652715575005, 'samples': 15540224, 'steps': 30351, 'loss/train': 0.7746754288673401} 02/25/2022 16:54:15 - INFO - codeparrot_training - Step 30352: {'lr': 0.00017974956823368728, 'samples': 15540736, 'steps': 30352, 'loss/train': 1.9135740995407104} 02/25/2022 16:54:21 - INFO - codeparrot_training - Step 30353: {'lr': 0.00017973386521080454, 'samples': 15541248, 'steps': 30353, 'loss/train': 2.9119625091552734} 02/25/2022 16:54:24 - INFO - codeparrot_training - Step 30354: {'lr': 0.00017971816248891972, 'samples': 15541760, 'steps': 30354, 'loss/train': 1.8786157369613647} 02/25/2022 16:54:30 - INFO - codeparrot_training - Step 30355: {'lr': 0.0001797024600680999, 'samples': 15542272, 'steps': 30355, 'loss/train': 1.6714987754821777} 02/25/2022 16:54:33 - INFO - codeparrot_training - Step 30356: {'lr': 0.00017968675794841242, 'samples': 15542784, 'steps': 30356, 'loss/train': 0.6088492274284363} 02/25/2022 16:54:39 - INFO - codeparrot_training - Step 30357: {'lr': 0.00017967105612992453, 'samples': 15543296, 'steps': 30357, 'loss/train': 2.0997838973999023} 02/25/2022 16:54:42 - INFO - codeparrot_training - Step 30358: {'lr': 0.0001796553546127035, 'samples': 15543808, 'steps': 30358, 'loss/train': 2.0839684009552} 02/25/2022 16:54:48 - INFO - codeparrot_training - Step 30359: {'lr': 0.0001796396533968167, 'samples': 15544320, 'steps': 30359, 'loss/train': 1.047836422920227} 02/25/2022 16:54:51 - INFO - codeparrot_training - Step 30360: {'lr': 0.00017962395248233116, 'samples': 15544832, 'steps': 30360, 'loss/train': 0.5357745289802551} 02/25/2022 16:54:58 - INFO - codeparrot_training - Step 30361: {'lr': 0.0001796082518693143, 'samples': 15545344, 'steps': 30361, 'loss/train': 0.5331573486328125} 02/25/2022 16:55:02 - INFO - codeparrot_training - Step 30362: {'lr': 0.00017959255155783333, 'samples': 15545856, 'steps': 30362, 'loss/train': 2.125424861907959} 02/25/2022 16:55:07 - INFO - codeparrot_training - Step 30363: {'lr': 0.00017957685154795558, 'samples': 15546368, 'steps': 30363, 'loss/train': 2.351050853729248} 02/25/2022 16:55:11 - INFO - codeparrot_training - Step 30364: {'lr': 0.00017956115183974817, 'samples': 15546880, 'steps': 30364, 'loss/train': 0.18436282873153687} 02/25/2022 16:55:16 - INFO - codeparrot_training - Step 30365: {'lr': 0.0001795454524332784, 'samples': 15547392, 'steps': 30365, 'loss/train': 1.8483694791793823} 02/25/2022 16:55:20 - INFO - codeparrot_training - Step 30366: {'lr': 0.0001795297533286136, 'samples': 15547904, 'steps': 30366, 'loss/train': 1.4149712324142456} 02/25/2022 16:55:25 - INFO - codeparrot_training - Step 30367: {'lr': 0.0001795140545258209, 'samples': 15548416, 'steps': 30367, 'loss/train': 2.066742420196533} 02/25/2022 16:55:29 - INFO - codeparrot_training - Step 30368: {'lr': 0.00017949835602496767, 'samples': 15548928, 'steps': 30368, 'loss/train': 2.4230031967163086} 02/25/2022 16:55:34 - INFO - codeparrot_training - Step 30369: {'lr': 0.00017948265782612106, 'samples': 15549440, 'steps': 30369, 'loss/train': 1.5072993040084839} 02/25/2022 16:55:38 - INFO - codeparrot_training - Step 30370: {'lr': 0.00017946695992934836, 'samples': 15549952, 'steps': 30370, 'loss/train': 1.5139480829238892} 02/25/2022 16:55:45 - INFO - codeparrot_training - Step 30371: {'lr': 0.0001794512623347168, 'samples': 15550464, 'steps': 30371, 'loss/train': 1.3498259782791138} 02/25/2022 16:55:48 - INFO - codeparrot_training - Step 30372: {'lr': 0.00017943556504229364, 'samples': 15550976, 'steps': 30372, 'loss/train': 1.6495883464813232} 02/25/2022 16:55:54 - INFO - codeparrot_training - Step 30373: {'lr': 0.00017941986805214614, 'samples': 15551488, 'steps': 30373, 'loss/train': 0.6480066180229187} 02/25/2022 16:55:57 - INFO - codeparrot_training - Step 30374: {'lr': 0.00017940417136434145, 'samples': 15552000, 'steps': 30374, 'loss/train': 1.8776949644088745} 02/25/2022 16:56:03 - INFO - codeparrot_training - Step 30375: {'lr': 0.0001793884749789469, 'samples': 15552512, 'steps': 30375, 'loss/train': 1.4403725862503052} 02/25/2022 16:56:06 - INFO - codeparrot_training - Step 30376: {'lr': 0.00017937277889602972, 'samples': 15553024, 'steps': 30376, 'loss/train': 2.1115121841430664} 02/25/2022 16:56:13 - INFO - codeparrot_training - Step 30377: {'lr': 0.00017935708311565712, 'samples': 15553536, 'steps': 30377, 'loss/train': 2.657585859298706} 02/25/2022 16:56:16 - INFO - codeparrot_training - Step 30378: {'lr': 0.00017934138763789633, 'samples': 15554048, 'steps': 30378, 'loss/train': 2.2927820682525635} 02/25/2022 16:56:19 - INFO - codeparrot_training - Step 30379: {'lr': 0.00017932569246281458, 'samples': 15554560, 'steps': 30379, 'loss/train': 1.2131829261779785} 02/25/2022 16:56:25 - INFO - codeparrot_training - Step 30380: {'lr': 0.0001793099975904791, 'samples': 15555072, 'steps': 30380, 'loss/train': 2.9934585094451904} 02/25/2022 16:56:28 - INFO - codeparrot_training - Step 30381: {'lr': 0.0001792943030209573, 'samples': 15555584, 'steps': 30381, 'loss/train': 2.340916156768799} 02/25/2022 16:56:35 - INFO - codeparrot_training - Step 30382: {'lr': 0.0001792786087543161, 'samples': 15556096, 'steps': 30382, 'loss/train': 2.8370912075042725} 02/25/2022 16:56:38 - INFO - codeparrot_training - Step 30383: {'lr': 0.0001792629147906229, 'samples': 15556608, 'steps': 30383, 'loss/train': 1.3860429525375366} 02/25/2022 16:56:44 - INFO - codeparrot_training - Step 30384: {'lr': 0.00017924722112994492, 'samples': 15557120, 'steps': 30384, 'loss/train': 1.2881313562393188} 02/25/2022 16:56:49 - INFO - codeparrot_training - Step 30385: {'lr': 0.0001792315277723495, 'samples': 15557632, 'steps': 30385, 'loss/train': 1.1875712871551514} 02/25/2022 16:56:53 - INFO - codeparrot_training - Step 30386: {'lr': 0.0001792158347179036, 'samples': 15558144, 'steps': 30386, 'loss/train': 1.2800253629684448} 02/25/2022 16:56:58 - INFO - codeparrot_training - Step 30387: {'lr': 0.0001792001419666746, 'samples': 15558656, 'steps': 30387, 'loss/train': 3.454211950302124} 02/25/2022 16:57:02 - INFO - codeparrot_training - Step 30388: {'lr': 0.00017918444951872973, 'samples': 15559168, 'steps': 30388, 'loss/train': 2.270298719406128} 02/25/2022 16:57:07 - INFO - codeparrot_training - Step 30389: {'lr': 0.00017916875737413628, 'samples': 15559680, 'steps': 30389, 'loss/train': 0.14630478620529175} 02/25/2022 16:57:10 - INFO - codeparrot_training - Step 30390: {'lr': 0.00017915306553296128, 'samples': 15560192, 'steps': 30390, 'loss/train': 0.29109862446784973} 02/25/2022 16:57:16 - INFO - codeparrot_training - Step 30391: {'lr': 0.00017913737399527206, 'samples': 15560704, 'steps': 30391, 'loss/train': 1.2275121212005615} 02/25/2022 16:57:20 - INFO - codeparrot_training - Step 30392: {'lr': 0.00017912168276113582, 'samples': 15561216, 'steps': 30392, 'loss/train': 1.7555817365646362} 02/25/2022 16:57:25 - INFO - codeparrot_training - Step 30393: {'lr': 0.00017910599183061976, 'samples': 15561728, 'steps': 30393, 'loss/train': 0.1005677655339241} 02/25/2022 16:57:29 - INFO - codeparrot_training - Step 30394: {'lr': 0.00017909030120379126, 'samples': 15562240, 'steps': 30394, 'loss/train': 1.9121339321136475} 02/25/2022 16:57:34 - INFO - codeparrot_training - Step 30395: {'lr': 0.00017907461088071728, 'samples': 15562752, 'steps': 30395, 'loss/train': 1.8280134201049805} 02/25/2022 16:57:38 - INFO - codeparrot_training - Step 30396: {'lr': 0.00017905892086146513, 'samples': 15563264, 'steps': 30396, 'loss/train': 1.857879638671875} 02/25/2022 16:57:44 - INFO - codeparrot_training - Step 30397: {'lr': 0.00017904323114610206, 'samples': 15563776, 'steps': 30397, 'loss/train': 1.7603507041931152} 02/25/2022 16:57:47 - INFO - codeparrot_training - Step 30398: {'lr': 0.00017902754173469536, 'samples': 15564288, 'steps': 30398, 'loss/train': 1.5191551446914673} 02/25/2022 16:57:53 - INFO - codeparrot_training - Step 30399: {'lr': 0.000179011852627312, 'samples': 15564800, 'steps': 30399, 'loss/train': 1.4686052799224854} 02/25/2022 16:57:56 - INFO - codeparrot_training - Step 30400: {'lr': 0.00017899616382401935, 'samples': 15565312, 'steps': 30400, 'loss/train': 0.4663267433643341} 02/25/2022 16:58:02 - INFO - codeparrot_training - Step 30401: {'lr': 0.0001789804753248846, 'samples': 15565824, 'steps': 30401, 'loss/train': 2.4124910831451416} 02/25/2022 16:58:05 - INFO - codeparrot_training - Step 30402: {'lr': 0.00017896478712997499, 'samples': 15566336, 'steps': 30402, 'loss/train': 2.231232166290283} 02/25/2022 16:58:11 - INFO - codeparrot_training - Step 30403: {'lr': 0.00017894909923935761, 'samples': 15566848, 'steps': 30403, 'loss/train': 1.9327001571655273} 02/25/2022 16:58:14 - INFO - codeparrot_training - Step 30404: {'lr': 0.00017893341165309973, 'samples': 15567360, 'steps': 30404, 'loss/train': 1.5463478565216064} 02/25/2022 16:58:20 - INFO - codeparrot_training - Step 30405: {'lr': 0.00017891772437126854, 'samples': 15567872, 'steps': 30405, 'loss/train': 1.7060670852661133} 02/25/2022 16:58:23 - INFO - codeparrot_training - Step 30406: {'lr': 0.0001789020373939313, 'samples': 15568384, 'steps': 30406, 'loss/train': 1.8175512552261353} 02/25/2022 16:58:30 - INFO - codeparrot_training - Step 30407: {'lr': 0.00017888635072115517, 'samples': 15568896, 'steps': 30407, 'loss/train': 1.597469687461853} 02/25/2022 16:58:33 - INFO - codeparrot_training - Step 30408: {'lr': 0.00017887066435300729, 'samples': 15569408, 'steps': 30408, 'loss/train': 2.3395121097564697} 02/25/2022 16:58:39 - INFO - codeparrot_training - Step 30409: {'lr': 0.00017885497828955488, 'samples': 15569920, 'steps': 30409, 'loss/train': 2.5138654708862305} 02/25/2022 16:58:42 - INFO - codeparrot_training - Step 30410: {'lr': 0.00017883929253086517, 'samples': 15570432, 'steps': 30410, 'loss/train': 0.031385187059640884} 02/25/2022 16:58:48 - INFO - codeparrot_training - Step 30411: {'lr': 0.0001788236070770054, 'samples': 15570944, 'steps': 30411, 'loss/train': 1.58345365524292} 02/25/2022 16:58:51 - INFO - codeparrot_training - Step 30412: {'lr': 0.0001788079219280426, 'samples': 15571456, 'steps': 30412, 'loss/train': 1.1391631364822388} 02/25/2022 16:58:57 - INFO - codeparrot_training - Step 30413: {'lr': 0.00017879223708404412, 'samples': 15571968, 'steps': 30413, 'loss/train': 2.277214765548706} 02/25/2022 16:59:00 - INFO - codeparrot_training - Step 30414: {'lr': 0.00017877655254507703, 'samples': 15572480, 'steps': 30414, 'loss/train': 1.3200619220733643} 02/25/2022 16:59:06 - INFO - codeparrot_training - Step 30415: {'lr': 0.00017876086831120864, 'samples': 15572992, 'steps': 30415, 'loss/train': 2.1675829887390137} 02/25/2022 16:59:09 - INFO - codeparrot_training - Step 30416: {'lr': 0.00017874518438250596, 'samples': 15573504, 'steps': 30416, 'loss/train': 2.615511894226074} 02/25/2022 16:59:15 - INFO - codeparrot_training - Step 30417: {'lr': 0.0001787295007590364, 'samples': 15574016, 'steps': 30417, 'loss/train': 2.0820140838623047} 02/25/2022 16:59:18 - INFO - codeparrot_training - Step 30418: {'lr': 0.00017871381744086694, 'samples': 15574528, 'steps': 30418, 'loss/train': 3.1339359283447266} 02/25/2022 16:59:24 - INFO - codeparrot_training - Step 30419: {'lr': 0.00017869813442806487, 'samples': 15575040, 'steps': 30419, 'loss/train': 1.7506332397460938} 02/25/2022 16:59:28 - INFO - codeparrot_training - Step 30420: {'lr': 0.00017868245172069746, 'samples': 15575552, 'steps': 30420, 'loss/train': 2.532691478729248} 02/25/2022 16:59:33 - INFO - codeparrot_training - Step 30421: {'lr': 0.0001786667693188317, 'samples': 15576064, 'steps': 30421, 'loss/train': 0.6444061398506165} 02/25/2022 16:59:37 - INFO - codeparrot_training - Step 30422: {'lr': 0.00017865108722253483, 'samples': 15576576, 'steps': 30422, 'loss/train': 2.0239057540893555} 02/25/2022 16:59:42 - INFO - codeparrot_training - Step 30423: {'lr': 0.000178635405431874, 'samples': 15577088, 'steps': 30423, 'loss/train': 2.295332431793213} 02/25/2022 16:59:46 - INFO - codeparrot_training - Step 30424: {'lr': 0.00017861972394691655, 'samples': 15577600, 'steps': 30424, 'loss/train': 1.6566064357757568} 02/25/2022 16:59:52 - INFO - codeparrot_training - Step 30425: {'lr': 0.00017860404276772945, 'samples': 15578112, 'steps': 30425, 'loss/train': 0.6063607335090637} 02/25/2022 16:59:55 - INFO - codeparrot_training - Step 30426: {'lr': 0.00017858836189438, 'samples': 15578624, 'steps': 30426, 'loss/train': 2.3397676944732666} 02/25/2022 17:00:01 - INFO - codeparrot_training - Step 30427: {'lr': 0.00017857268132693527, 'samples': 15579136, 'steps': 30427, 'loss/train': 1.2900431156158447} 02/25/2022 17:00:04 - INFO - codeparrot_training - Step 30428: {'lr': 0.00017855700106546253, 'samples': 15579648, 'steps': 30428, 'loss/train': 1.8489912748336792} 02/25/2022 17:00:11 - INFO - codeparrot_training - Step 30429: {'lr': 0.00017854132111002903, 'samples': 15580160, 'steps': 30429, 'loss/train': 0.22888192534446716} 02/25/2022 17:00:16 - INFO - codeparrot_training - Step 30430: {'lr': 0.00017852564146070166, 'samples': 15580672, 'steps': 30430, 'loss/train': 1.965506672859192} 02/25/2022 17:00:20 - INFO - codeparrot_training - Step 30431: {'lr': 0.00017850996211754778, 'samples': 15581184, 'steps': 30431, 'loss/train': 1.6344127655029297} 02/25/2022 17:00:25 - INFO - codeparrot_training - Step 30432: {'lr': 0.00017849428308063452, 'samples': 15581696, 'steps': 30432, 'loss/train': 0.405988484621048} 02/25/2022 17:00:29 - INFO - codeparrot_training - Step 30433: {'lr': 0.00017847860435002915, 'samples': 15582208, 'steps': 30433, 'loss/train': 1.8703230619430542} 02/25/2022 17:00:34 - INFO - codeparrot_training - Step 30434: {'lr': 0.0001784629259257986, 'samples': 15582720, 'steps': 30434, 'loss/train': 2.751441478729248} 02/25/2022 17:00:38 - INFO - codeparrot_training - Step 30435: {'lr': 0.0001784472478080102, 'samples': 15583232, 'steps': 30435, 'loss/train': 0.8970136642456055} 02/25/2022 17:00:43 - INFO - codeparrot_training - Step 30436: {'lr': 0.00017843156999673105, 'samples': 15583744, 'steps': 30436, 'loss/train': 2.1291606426239014} 02/25/2022 17:00:47 - INFO - codeparrot_training - Step 30437: {'lr': 0.00017841589249202843, 'samples': 15584256, 'steps': 30437, 'loss/train': 1.4831438064575195} 02/25/2022 17:00:54 - INFO - codeparrot_training - Step 30438: {'lr': 0.0001784002152939693, 'samples': 15584768, 'steps': 30438, 'loss/train': 1.8092501163482666} 02/25/2022 17:00:57 - INFO - codeparrot_training - Step 30439: {'lr': 0.00017838453840262093, 'samples': 15585280, 'steps': 30439, 'loss/train': 1.9374003410339355} 02/25/2022 17:01:03 - INFO - codeparrot_training - Step 30440: {'lr': 0.0001783688618180504, 'samples': 15585792, 'steps': 30440, 'loss/train': 1.5201354026794434} 02/25/2022 17:01:06 - INFO - codeparrot_training - Step 30441: {'lr': 0.00017835318554032498, 'samples': 15586304, 'steps': 30441, 'loss/train': 0.5519653558731079} 02/25/2022 17:01:11 - INFO - codeparrot_training - Step 30442: {'lr': 0.00017833750956951184, 'samples': 15586816, 'steps': 30442, 'loss/train': 2.6692910194396973} 02/25/2022 17:01:15 - INFO - codeparrot_training - Step 30443: {'lr': 0.00017832183390567794, 'samples': 15587328, 'steps': 30443, 'loss/train': 1.865640640258789} 02/25/2022 17:01:20 - INFO - codeparrot_training - Step 30444: {'lr': 0.00017830615854889055, 'samples': 15587840, 'steps': 30444, 'loss/train': 2.4657881259918213} 02/25/2022 17:01:24 - INFO - codeparrot_training - Step 30445: {'lr': 0.00017829048349921685, 'samples': 15588352, 'steps': 30445, 'loss/train': 1.0936853885650635} 02/25/2022 17:01:29 - INFO - codeparrot_training - Step 30446: {'lr': 0.000178274808756724, 'samples': 15588864, 'steps': 30446, 'loss/train': 2.338529586791992} 02/25/2022 17:01:33 - INFO - codeparrot_training - Step 30447: {'lr': 0.000178259134321479, 'samples': 15589376, 'steps': 30447, 'loss/train': 1.8069361448287964} 02/25/2022 17:01:38 - INFO - codeparrot_training - Step 30448: {'lr': 0.0001782434601935491, 'samples': 15589888, 'steps': 30448, 'loss/train': 0.4443916380405426} 02/25/2022 17:01:42 - INFO - codeparrot_training - Step 30449: {'lr': 0.0001782277863730014, 'samples': 15590400, 'steps': 30449, 'loss/train': 1.14827299118042} 02/25/2022 17:01:47 - INFO - codeparrot_training - Step 30450: {'lr': 0.00017821211285990325, 'samples': 15590912, 'steps': 30450, 'loss/train': 2.1597254276275635} 02/25/2022 17:01:51 - INFO - codeparrot_training - Step 30451: {'lr': 0.00017819643965432145, 'samples': 15591424, 'steps': 30451, 'loss/train': 1.946171522140503} 02/25/2022 17:01:56 - INFO - codeparrot_training - Step 30452: {'lr': 0.00017818076675632334, 'samples': 15591936, 'steps': 30452, 'loss/train': 0.8167882561683655} 02/25/2022 17:02:00 - INFO - codeparrot_training - Step 30453: {'lr': 0.00017816509416597603, 'samples': 15592448, 'steps': 30453, 'loss/train': 1.5983017683029175} 02/25/2022 17:02:07 - INFO - codeparrot_training - Step 30454: {'lr': 0.0001781494218833466, 'samples': 15592960, 'steps': 30454, 'loss/train': 2.4102554321289062} 02/25/2022 17:02:10 - INFO - codeparrot_training - Step 30455: {'lr': 0.0001781337499085023, 'samples': 15593472, 'steps': 30455, 'loss/train': 0.025332242250442505} 02/25/2022 17:02:16 - INFO - codeparrot_training - Step 30456: {'lr': 0.00017811807824151014, 'samples': 15593984, 'steps': 30456, 'loss/train': 1.1768425703048706} 02/25/2022 17:02:19 - INFO - codeparrot_training - Step 30457: {'lr': 0.00017810240688243738, 'samples': 15594496, 'steps': 30457, 'loss/train': 2.0345492362976074} 02/25/2022 17:02:25 - INFO - codeparrot_training - Step 30458: {'lr': 0.000178086735831351, 'samples': 15595008, 'steps': 30458, 'loss/train': 0.7213736772537231} 02/25/2022 17:02:28 - INFO - codeparrot_training - Step 30459: {'lr': 0.00017807106508831828, 'samples': 15595520, 'steps': 30459, 'loss/train': 2.2966644763946533} 02/25/2022 17:02:34 - INFO - codeparrot_training - Step 30460: {'lr': 0.00017805539465340624, 'samples': 15596032, 'steps': 30460, 'loss/train': 2.0306010246276855} 02/25/2022 17:02:37 - INFO - codeparrot_training - Step 30461: {'lr': 0.000178039724526682, 'samples': 15596544, 'steps': 30461, 'loss/train': 0.999924898147583} 02/25/2022 17:02:43 - INFO - codeparrot_training - Step 30462: {'lr': 0.00017802405470821275, 'samples': 15597056, 'steps': 30462, 'loss/train': 1.6579713821411133} 02/25/2022 17:02:46 - INFO - codeparrot_training - Step 30463: {'lr': 0.00017800838519806567, 'samples': 15597568, 'steps': 30463, 'loss/train': 2.093268871307373} 02/25/2022 17:02:53 - INFO - codeparrot_training - Step 30464: {'lr': 0.0001779927159963078, 'samples': 15598080, 'steps': 30464, 'loss/train': 1.3233659267425537} 02/25/2022 17:02:56 - INFO - codeparrot_training - Step 30465: {'lr': 0.00017797704710300615, 'samples': 15598592, 'steps': 30465, 'loss/train': 2.512624740600586} 02/25/2022 17:03:02 - INFO - codeparrot_training - Step 30466: {'lr': 0.00017796137851822802, 'samples': 15599104, 'steps': 30466, 'loss/train': 1.8047192096710205} 02/25/2022 17:03:05 - INFO - codeparrot_training - Step 30467: {'lr': 0.0001779457102420405, 'samples': 15599616, 'steps': 30467, 'loss/train': 1.194332480430603} 02/25/2022 17:03:10 - INFO - codeparrot_training - Step 30468: {'lr': 0.00017793004227451071, 'samples': 15600128, 'steps': 30468, 'loss/train': 1.56063973903656} 02/25/2022 17:03:14 - INFO - codeparrot_training - Step 30469: {'lr': 0.00017791437461570563, 'samples': 15600640, 'steps': 30469, 'loss/train': 3.1466665267944336} 02/25/2022 17:03:19 - INFO - codeparrot_training - Step 30470: {'lr': 0.00017789870726569256, 'samples': 15601152, 'steps': 30470, 'loss/train': 1.4906448125839233} 02/25/2022 17:03:23 - INFO - codeparrot_training - Step 30471: {'lr': 0.00017788304022453848, 'samples': 15601664, 'steps': 30471, 'loss/train': 2.7139439582824707} 02/25/2022 17:03:28 - INFO - codeparrot_training - Step 30472: {'lr': 0.00017786737349231063, 'samples': 15602176, 'steps': 30472, 'loss/train': 1.6032030582427979} 02/25/2022 17:03:32 - INFO - codeparrot_training - Step 30473: {'lr': 0.000177851707069076, 'samples': 15602688, 'steps': 30473, 'loss/train': 2.177809953689575} 02/25/2022 17:03:38 - INFO - codeparrot_training - Step 30474: {'lr': 0.00017783604095490172, 'samples': 15603200, 'steps': 30474, 'loss/train': 1.9201726913452148} 02/25/2022 17:03:41 - INFO - codeparrot_training - Step 30475: {'lr': 0.00017782037514985495, 'samples': 15603712, 'steps': 30475, 'loss/train': 2.48826265335083} 02/25/2022 17:03:47 - INFO - codeparrot_training - Step 30476: {'lr': 0.0001778047096540027, 'samples': 15604224, 'steps': 30476, 'loss/train': 2.207005739212036} 02/25/2022 17:03:51 - INFO - codeparrot_training - Step 30477: {'lr': 0.00017778904446741234, 'samples': 15604736, 'steps': 30477, 'loss/train': 1.539368987083435} 02/25/2022 17:03:56 - INFO - codeparrot_training - Step 30478: {'lr': 0.00017777337959015065, 'samples': 15605248, 'steps': 30478, 'loss/train': 1.6143447160720825} 02/25/2022 17:03:59 - INFO - codeparrot_training - Step 30479: {'lr': 0.00017775771502228486, 'samples': 15605760, 'steps': 30479, 'loss/train': 1.0108176469802856} 02/25/2022 17:04:05 - INFO - codeparrot_training - Step 30480: {'lr': 0.00017774205076388205, 'samples': 15606272, 'steps': 30480, 'loss/train': 0.540143609046936} 02/25/2022 17:04:08 - INFO - codeparrot_training - Step 30481: {'lr': 0.00017772638681500953, 'samples': 15606784, 'steps': 30481, 'loss/train': 1.286149024963379} 02/25/2022 17:04:14 - INFO - codeparrot_training - Step 30482: {'lr': 0.00017771072317573406, 'samples': 15607296, 'steps': 30482, 'loss/train': 2.2609012126922607} 02/25/2022 17:04:19 - INFO - codeparrot_training - Step 30483: {'lr': 0.00017769505984612292, 'samples': 15607808, 'steps': 30483, 'loss/train': 1.8654690980911255} 02/25/2022 17:04:23 - INFO - codeparrot_training - Step 30484: {'lr': 0.00017767939682624317, 'samples': 15608320, 'steps': 30484, 'loss/train': 2.398495674133301} 02/25/2022 17:04:29 - INFO - codeparrot_training - Step 30485: {'lr': 0.000177663734116162, 'samples': 15608832, 'steps': 30485, 'loss/train': 2.141735315322876} 02/25/2022 17:04:32 - INFO - codeparrot_training - Step 30486: {'lr': 0.0001776480717159464, 'samples': 15609344, 'steps': 30486, 'loss/train': 2.3866524696350098} 02/25/2022 17:04:38 - INFO - codeparrot_training - Step 30487: {'lr': 0.0001776324096256634, 'samples': 15609856, 'steps': 30487, 'loss/train': 2.574432849884033} 02/25/2022 17:04:41 - INFO - codeparrot_training - Step 30488: {'lr': 0.0001776167478453802, 'samples': 15610368, 'steps': 30488, 'loss/train': 1.2526462078094482} 02/25/2022 17:04:47 - INFO - codeparrot_training - Step 30489: {'lr': 0.0001776010863751639, 'samples': 15610880, 'steps': 30489, 'loss/train': 2.8575878143310547} 02/25/2022 17:04:51 - INFO - codeparrot_training - Step 30490: {'lr': 0.00017758542521508164, 'samples': 15611392, 'steps': 30490, 'loss/train': 2.34999418258667} 02/25/2022 17:04:54 - INFO - codeparrot_training - Step 30491: {'lr': 0.00017756976436520032, 'samples': 15611904, 'steps': 30491, 'loss/train': 1.517310380935669} 02/25/2022 17:05:00 - INFO - codeparrot_training - Step 30492: {'lr': 0.0001775541038255871, 'samples': 15612416, 'steps': 30492, 'loss/train': 1.964463710784912} 02/25/2022 17:05:05 - INFO - codeparrot_training - Step 30493: {'lr': 0.00017753844359630921, 'samples': 15612928, 'steps': 30493, 'loss/train': 1.4008477926254272} 02/25/2022 17:05:09 - INFO - codeparrot_training - Step 30494: {'lr': 0.00017752278367743357, 'samples': 15613440, 'steps': 30494, 'loss/train': 1.7811930179595947} 02/25/2022 17:05:14 - INFO - codeparrot_training - Step 30495: {'lr': 0.0001775071240690273, 'samples': 15613952, 'steps': 30495, 'loss/train': 1.6813509464263916} 02/25/2022 17:05:18 - INFO - codeparrot_training - Step 30496: {'lr': 0.00017749146477115746, 'samples': 15614464, 'steps': 30496, 'loss/train': 1.497260570526123} 02/25/2022 17:05:23 - INFO - codeparrot_training - Step 30497: {'lr': 0.0001774758057838912, 'samples': 15614976, 'steps': 30497, 'loss/train': 1.5442038774490356} 02/25/2022 17:05:27 - INFO - codeparrot_training - Step 30498: {'lr': 0.00017746014710729558, 'samples': 15615488, 'steps': 30498, 'loss/train': 1.6781425476074219} 02/25/2022 17:05:33 - INFO - codeparrot_training - Step 30499: {'lr': 0.00017744448874143758, 'samples': 15616000, 'steps': 30499, 'loss/train': 1.3502012491226196} 02/25/2022 17:05:36 - INFO - codeparrot_training - Step 30500: {'lr': 0.00017742883068638446, 'samples': 15616512, 'steps': 30500, 'loss/train': 1.7134546041488647} 02/25/2022 17:05:42 - INFO - codeparrot_training - Step 30501: {'lr': 0.0001774131729422031, 'samples': 15617024, 'steps': 30501, 'loss/train': 1.6110897064208984} 02/25/2022 17:05:45 - INFO - codeparrot_training - Step 30502: {'lr': 0.00017739751550896067, 'samples': 15617536, 'steps': 30502, 'loss/train': 1.4763764142990112} 02/25/2022 17:05:51 - INFO - codeparrot_training - Step 30503: {'lr': 0.00017738185838672425, 'samples': 15618048, 'steps': 30503, 'loss/train': 0.5471266508102417} 02/25/2022 17:05:54 - INFO - codeparrot_training - Step 30504: {'lr': 0.0001773662015755609, 'samples': 15618560, 'steps': 30504, 'loss/train': 0.8086187839508057} 02/25/2022 17:06:00 - INFO - codeparrot_training - Step 30505: {'lr': 0.00017735054507553766, 'samples': 15619072, 'steps': 30505, 'loss/train': 1.5706974267959595} 02/25/2022 17:06:03 - INFO - codeparrot_training - Step 30506: {'lr': 0.00017733488888672155, 'samples': 15619584, 'steps': 30506, 'loss/train': 0.7623317837715149} 02/25/2022 17:06:09 - INFO - codeparrot_training - Step 30507: {'lr': 0.00017731923300917982, 'samples': 15620096, 'steps': 30507, 'loss/train': 0.09161017090082169} 02/25/2022 17:06:12 - INFO - codeparrot_training - Step 30508: {'lr': 0.00017730357744297938, 'samples': 15620608, 'steps': 30508, 'loss/train': 1.4521379470825195} 02/25/2022 17:06:18 - INFO - codeparrot_training - Step 30509: {'lr': 0.0001772879221881873, 'samples': 15621120, 'steps': 30509, 'loss/train': 1.8131828308105469} 02/25/2022 17:06:22 - INFO - codeparrot_training - Step 30510: {'lr': 0.00017727226724487068, 'samples': 15621632, 'steps': 30510, 'loss/train': 1.8775689601898193} 02/25/2022 17:06:28 - INFO - codeparrot_training - Step 30511: {'lr': 0.00017725661261309666, 'samples': 15622144, 'steps': 30511, 'loss/train': 1.881587266921997} 02/25/2022 17:06:31 - INFO - codeparrot_training - Step 30512: {'lr': 0.0001772409582929321, 'samples': 15622656, 'steps': 30512, 'loss/train': 1.0350226163864136} 02/25/2022 17:06:37 - INFO - codeparrot_training - Step 30513: {'lr': 0.00017722530428444417, 'samples': 15623168, 'steps': 30513, 'loss/train': 1.8111931085586548} 02/25/2022 17:06:40 - INFO - codeparrot_training - Step 30514: {'lr': 0.00017720965058769995, 'samples': 15623680, 'steps': 30514, 'loss/train': 2.302652597427368} 02/25/2022 17:06:46 - INFO - codeparrot_training - Step 30515: {'lr': 0.00017719399720276646, 'samples': 15624192, 'steps': 30515, 'loss/train': 1.6977014541625977} 02/25/2022 17:06:49 - INFO - codeparrot_training - Step 30516: {'lr': 0.00017717834412971087, 'samples': 15624704, 'steps': 30516, 'loss/train': 2.1854710578918457} 02/25/2022 17:06:55 - INFO - codeparrot_training - Step 30517: {'lr': 0.00017716269136860003, 'samples': 15625216, 'steps': 30517, 'loss/train': 1.097787618637085} 02/25/2022 17:06:58 - INFO - codeparrot_training - Step 30518: {'lr': 0.0001771470389195011, 'samples': 15625728, 'steps': 30518, 'loss/train': 1.9076862335205078} 02/25/2022 17:07:04 - INFO - codeparrot_training - Step 30519: {'lr': 0.00017713138678248112, 'samples': 15626240, 'steps': 30519, 'loss/train': 2.024108648300171} 02/25/2022 17:07:08 - INFO - codeparrot_training - Step 30520: {'lr': 0.00017711573495760724, 'samples': 15626752, 'steps': 30520, 'loss/train': 1.320483922958374} 02/25/2022 17:07:13 - INFO - codeparrot_training - Step 30521: {'lr': 0.0001771000834449463, 'samples': 15627264, 'steps': 30521, 'loss/train': 1.3672194480895996} 02/25/2022 17:07:17 - INFO - codeparrot_training - Step 30522: {'lr': 0.00017708443224456543, 'samples': 15627776, 'steps': 30522, 'loss/train': 1.7192773818969727} 02/25/2022 17:07:22 - INFO - codeparrot_training - Step 30523: {'lr': 0.00017706878135653176, 'samples': 15628288, 'steps': 30523, 'loss/train': 1.680015206336975} 02/25/2022 17:07:26 - INFO - codeparrot_training - Step 30524: {'lr': 0.00017705313078091235, 'samples': 15628800, 'steps': 30524, 'loss/train': 1.914446473121643} 02/25/2022 17:07:32 - INFO - codeparrot_training - Step 30525: {'lr': 0.00017703748051777401, 'samples': 15629312, 'steps': 30525, 'loss/train': 1.5716078281402588} 02/25/2022 17:07:35 - INFO - codeparrot_training - Step 30526: {'lr': 0.00017702183056718398, 'samples': 15629824, 'steps': 30526, 'loss/train': 1.9488903284072876} 02/25/2022 17:07:40 - INFO - codeparrot_training - Step 30527: {'lr': 0.00017700618092920928, 'samples': 15630336, 'steps': 30527, 'loss/train': 2.2982089519500732} 02/25/2022 17:07:44 - INFO - codeparrot_training - Step 30528: {'lr': 0.00017699053160391686, 'samples': 15630848, 'steps': 30528, 'loss/train': 2.0038654804229736} 02/25/2022 17:07:50 - INFO - codeparrot_training - Step 30529: {'lr': 0.00017697488259137397, 'samples': 15631360, 'steps': 30529, 'loss/train': 1.3951107263565063} 02/25/2022 17:07:53 - INFO - codeparrot_training - Step 30530: {'lr': 0.0001769592338916474, 'samples': 15631872, 'steps': 30530, 'loss/train': 1.6343458890914917} 02/25/2022 17:07:59 - INFO - codeparrot_training - Step 30531: {'lr': 0.00017694358550480424, 'samples': 15632384, 'steps': 30531, 'loss/train': 1.818979024887085} 02/25/2022 17:08:02 - INFO - codeparrot_training - Step 30532: {'lr': 0.00017692793743091158, 'samples': 15632896, 'steps': 30532, 'loss/train': 1.8359817266464233} 02/25/2022 17:08:08 - INFO - codeparrot_training - Step 30533: {'lr': 0.00017691228967003654, 'samples': 15633408, 'steps': 30533, 'loss/train': 1.6695060729980469} 02/25/2022 17:08:11 - INFO - codeparrot_training - Step 30534: {'lr': 0.00017689664222224593, 'samples': 15633920, 'steps': 30534, 'loss/train': 1.3436036109924316} 02/25/2022 17:08:17 - INFO - codeparrot_training - Step 30535: {'lr': 0.0001768809950876069, 'samples': 15634432, 'steps': 30535, 'loss/train': 1.8788748979568481} 02/25/2022 17:08:21 - INFO - codeparrot_training - Step 30536: {'lr': 0.00017686534826618646, 'samples': 15634944, 'steps': 30536, 'loss/train': 0.854434072971344} 02/25/2022 17:08:27 - INFO - codeparrot_training - Step 30537: {'lr': 0.0001768497017580517, 'samples': 15635456, 'steps': 30537, 'loss/train': 0.04006991535425186} 02/25/2022 17:08:30 - INFO - codeparrot_training - Step 30538: {'lr': 0.00017683405556326962, 'samples': 15635968, 'steps': 30538, 'loss/train': 1.76677405834198} 02/25/2022 17:08:36 - INFO - codeparrot_training - Step 30539: {'lr': 0.00017681840968190713, 'samples': 15636480, 'steps': 30539, 'loss/train': 0.7660912275314331} 02/25/2022 17:08:39 - INFO - codeparrot_training - Step 30540: {'lr': 0.00017680276411403135, 'samples': 15636992, 'steps': 30540, 'loss/train': 3.143925428390503} 02/25/2022 17:08:45 - INFO - codeparrot_training - Step 30541: {'lr': 0.00017678711885970932, 'samples': 15637504, 'steps': 30541, 'loss/train': 2.062979221343994} 02/25/2022 17:08:48 - INFO - codeparrot_training - Step 30542: {'lr': 0.00017677147391900806, 'samples': 15638016, 'steps': 30542, 'loss/train': 0.058564506471157074} 02/25/2022 17:08:54 - INFO - codeparrot_training - Step 30543: {'lr': 0.00017675582929199446, 'samples': 15638528, 'steps': 30543, 'loss/train': 0.9685401320457458} 02/25/2022 17:08:57 - INFO - codeparrot_training - Step 30544: {'lr': 0.00017674018497873566, 'samples': 15639040, 'steps': 30544, 'loss/train': 2.0894265174865723} 02/25/2022 17:09:04 - INFO - codeparrot_training - Step 30545: {'lr': 0.00017672454097929864, 'samples': 15639552, 'steps': 30545, 'loss/train': 1.9795327186584473} 02/25/2022 17:09:07 - INFO - codeparrot_training - Step 30546: {'lr': 0.00017670889729375048, 'samples': 15640064, 'steps': 30546, 'loss/train': 1.4028739929199219} 02/25/2022 17:09:13 - INFO - codeparrot_training - Step 30547: {'lr': 0.00017669325392215808, 'samples': 15640576, 'steps': 30547, 'loss/train': 0.38878780603408813} 02/25/2022 17:09:16 - INFO - codeparrot_training - Step 30548: {'lr': 0.0001766776108645885, 'samples': 15641088, 'steps': 30548, 'loss/train': 1.4374879598617554} 02/25/2022 17:09:21 - INFO - codeparrot_training - Step 30549: {'lr': 0.00017666196812110875, 'samples': 15641600, 'steps': 30549, 'loss/train': 2.805933952331543} 02/25/2022 17:09:25 - INFO - codeparrot_training - Step 30550: {'lr': 0.00017664632569178583, 'samples': 15642112, 'steps': 30550, 'loss/train': 1.6329450607299805} 02/25/2022 17:09:30 - INFO - codeparrot_training - Step 30551: {'lr': 0.0001766306835766868, 'samples': 15642624, 'steps': 30551, 'loss/train': 2.0024988651275635} 02/25/2022 17:09:34 - INFO - codeparrot_training - Step 30552: {'lr': 0.0001766150417758786, 'samples': 15643136, 'steps': 30552, 'loss/train': 1.7042843103408813} 02/25/2022 17:09:39 - INFO - codeparrot_training - Step 30553: {'lr': 0.00017659940028942825, 'samples': 15643648, 'steps': 30553, 'loss/train': 2.143972396850586} 02/25/2022 17:09:43 - INFO - codeparrot_training - Step 30554: {'lr': 0.00017658375911740277, 'samples': 15644160, 'steps': 30554, 'loss/train': 1.4576095342636108} 02/25/2022 17:09:48 - INFO - codeparrot_training - Step 30555: {'lr': 0.00017656811825986923, 'samples': 15644672, 'steps': 30555, 'loss/train': 1.351486086845398} 02/25/2022 17:09:52 - INFO - codeparrot_training - Step 30556: {'lr': 0.00017655247771689446, 'samples': 15645184, 'steps': 30556, 'loss/train': 2.148231029510498} 02/25/2022 17:09:58 - INFO - codeparrot_training - Step 30557: {'lr': 0.00017653683748854553, 'samples': 15645696, 'steps': 30557, 'loss/train': 1.3356378078460693} 02/25/2022 17:10:02 - INFO - codeparrot_training - Step 30558: {'lr': 0.00017652119757488949, 'samples': 15646208, 'steps': 30558, 'loss/train': 1.282679796218872} 02/25/2022 17:10:07 - INFO - codeparrot_training - Step 30559: {'lr': 0.0001765055579759934, 'samples': 15646720, 'steps': 30559, 'loss/train': 1.8678544759750366} 02/25/2022 17:10:10 - INFO - codeparrot_training - Step 30560: {'lr': 0.00017648991869192405, 'samples': 15647232, 'steps': 30560, 'loss/train': 1.091941237449646} 02/25/2022 17:10:16 - INFO - codeparrot_training - Step 30561: {'lr': 0.00017647427972274858, 'samples': 15647744, 'steps': 30561, 'loss/train': 2.1383249759674072} 02/25/2022 17:10:19 - INFO - codeparrot_training - Step 30562: {'lr': 0.00017645864106853392, 'samples': 15648256, 'steps': 30562, 'loss/train': 2.4352896213531494} 02/25/2022 17:10:25 - INFO - codeparrot_training - Step 30563: {'lr': 0.0001764430027293471, 'samples': 15648768, 'steps': 30563, 'loss/train': 1.6953250169754028} 02/25/2022 17:10:28 - INFO - codeparrot_training - Step 30564: {'lr': 0.0001764273647052552, 'samples': 15649280, 'steps': 30564, 'loss/train': 2.5382022857666016} 02/25/2022 17:10:34 - INFO - codeparrot_training - Step 30565: {'lr': 0.000176411726996325, 'samples': 15649792, 'steps': 30565, 'loss/train': 0.9479430913925171} 02/25/2022 17:10:37 - INFO - codeparrot_training - Step 30566: {'lr': 0.0001763960896026236, 'samples': 15650304, 'steps': 30566, 'loss/train': 1.8191521167755127} 02/25/2022 17:10:44 - INFO - codeparrot_training - Step 30567: {'lr': 0.00017638045252421798, 'samples': 15650816, 'steps': 30567, 'loss/train': 1.8058654069900513} 02/25/2022 17:10:47 - INFO - codeparrot_training - Step 30568: {'lr': 0.00017636481576117522, 'samples': 15651328, 'steps': 30568, 'loss/train': 1.1017128229141235} 02/25/2022 17:10:53 - INFO - codeparrot_training - Step 30569: {'lr': 0.0001763491793135621, 'samples': 15651840, 'steps': 30569, 'loss/train': 2.018596649169922} 02/25/2022 17:10:56 - INFO - codeparrot_training - Step 30570: {'lr': 0.00017633354318144574, 'samples': 15652352, 'steps': 30570, 'loss/train': 1.7992150783538818} 02/25/2022 17:11:02 - INFO - codeparrot_training - Step 30571: {'lr': 0.00017631790736489307, 'samples': 15652864, 'steps': 30571, 'loss/train': 0.16332975029945374} 02/25/2022 17:11:05 - INFO - codeparrot_training - Step 30572: {'lr': 0.00017630227186397118, 'samples': 15653376, 'steps': 30572, 'loss/train': 1.3902952671051025} 02/25/2022 17:11:11 - INFO - codeparrot_training - Step 30573: {'lr': 0.00017628663667874683, 'samples': 15653888, 'steps': 30573, 'loss/train': 1.0414601564407349} 02/25/2022 17:11:14 - INFO - codeparrot_training - Step 30574: {'lr': 0.00017627100180928718, 'samples': 15654400, 'steps': 30574, 'loss/train': 2.0970213413238525} 02/25/2022 17:11:20 - INFO - codeparrot_training - Step 30575: {'lr': 0.0001762553672556591, 'samples': 15654912, 'steps': 30575, 'loss/train': 2.0458486080169678} 02/25/2022 17:11:23 - INFO - codeparrot_training - Step 30576: {'lr': 0.00017623973301792963, 'samples': 15655424, 'steps': 30576, 'loss/train': 1.5893223285675049} 02/25/2022 17:11:29 - INFO - codeparrot_training - Step 30577: {'lr': 0.00017622409909616582, 'samples': 15655936, 'steps': 30577, 'loss/train': 1.4068567752838135} 02/25/2022 17:11:34 - INFO - codeparrot_training - Step 30578: {'lr': 0.00017620846549043446, 'samples': 15656448, 'steps': 30578, 'loss/train': 1.2612806558609009} 02/25/2022 17:11:38 - INFO - codeparrot_training - Step 30579: {'lr': 0.00017619283220080258, 'samples': 15656960, 'steps': 30579, 'loss/train': 1.2561819553375244} 02/25/2022 17:11:44 - INFO - codeparrot_training - Step 30580: {'lr': 0.0001761771992273372, 'samples': 15657472, 'steps': 30580, 'loss/train': 1.2824877500534058} 02/25/2022 17:11:48 - INFO - codeparrot_training - Step 30581: {'lr': 0.00017616156657010532, 'samples': 15657984, 'steps': 30581, 'loss/train': 1.796589970588684} 02/25/2022 17:11:53 - INFO - codeparrot_training - Step 30582: {'lr': 0.00017614593422917373, 'samples': 15658496, 'steps': 30582, 'loss/train': 1.1098741292953491} 02/25/2022 17:11:57 - INFO - codeparrot_training - Step 30583: {'lr': 0.00017613030220460957, 'samples': 15659008, 'steps': 30583, 'loss/train': 0.7071213126182556} 02/25/2022 17:12:02 - INFO - codeparrot_training - Step 30584: {'lr': 0.00017611467049647976, 'samples': 15659520, 'steps': 30584, 'loss/train': 2.6004810333251953} 02/25/2022 17:12:06 - INFO - codeparrot_training - Step 30585: {'lr': 0.00017609903910485118, 'samples': 15660032, 'steps': 30585, 'loss/train': 1.9710332155227661} 02/25/2022 17:12:11 - INFO - codeparrot_training - Step 30586: {'lr': 0.0001760834080297909, 'samples': 15660544, 'steps': 30586, 'loss/train': 0.32962745428085327} 02/25/2022 17:12:15 - INFO - codeparrot_training - Step 30587: {'lr': 0.0001760677772713658, 'samples': 15661056, 'steps': 30587, 'loss/train': 2.142575740814209} 02/25/2022 17:12:20 - INFO - codeparrot_training - Step 30588: {'lr': 0.00017605214682964287, 'samples': 15661568, 'steps': 30588, 'loss/train': 0.9723318219184875} 02/25/2022 17:12:24 - INFO - codeparrot_training - Step 30589: {'lr': 0.00017603651670468907, 'samples': 15662080, 'steps': 30589, 'loss/train': 1.0869251489639282} 02/25/2022 17:12:30 - INFO - codeparrot_training - Step 30590: {'lr': 0.00017602088689657136, 'samples': 15662592, 'steps': 30590, 'loss/train': 3.337860345840454} 02/25/2022 17:12:34 - INFO - codeparrot_training - Step 30591: {'lr': 0.00017600525740535668, 'samples': 15663104, 'steps': 30591, 'loss/train': 0.8833999633789062} 02/25/2022 17:12:39 - INFO - codeparrot_training - Step 30592: {'lr': 0.00017598962823111192, 'samples': 15663616, 'steps': 30592, 'loss/train': 1.1802172660827637} 02/25/2022 17:12:43 - INFO - codeparrot_training - Step 30593: {'lr': 0.00017597399937390414, 'samples': 15664128, 'steps': 30593, 'loss/train': 1.6584898233413696} 02/25/2022 17:12:48 - INFO - codeparrot_training - Step 30594: {'lr': 0.00017595837083380026, 'samples': 15664640, 'steps': 30594, 'loss/train': 1.86874520778656} 02/25/2022 17:12:52 - INFO - codeparrot_training - Step 30595: {'lr': 0.0001759427426108672, 'samples': 15665152, 'steps': 30595, 'loss/train': 3.5682506561279297} 02/25/2022 17:12:57 - INFO - codeparrot_training - Step 30596: {'lr': 0.00017592711470517186, 'samples': 15665664, 'steps': 30596, 'loss/train': 1.6682144403457642} 02/25/2022 17:13:01 - INFO - codeparrot_training - Step 30597: {'lr': 0.0001759114871167813, 'samples': 15666176, 'steps': 30597, 'loss/train': 0.6292073130607605} 02/25/2022 17:13:06 - INFO - codeparrot_training - Step 30598: {'lr': 0.00017589585984576235, 'samples': 15666688, 'steps': 30598, 'loss/train': 0.6108160614967346} 02/25/2022 17:13:10 - INFO - codeparrot_training - Step 30599: {'lr': 0.00017588023289218208, 'samples': 15667200, 'steps': 30599, 'loss/train': 1.347469687461853} 02/25/2022 17:13:15 - INFO - codeparrot_training - Step 30600: {'lr': 0.0001758646062561073, 'samples': 15667712, 'steps': 30600, 'loss/train': 0.13789142668247223} 02/25/2022 17:13:19 - INFO - codeparrot_training - Step 30601: {'lr': 0.000175848979937605, 'samples': 15668224, 'steps': 30601, 'loss/train': 1.4443327188491821} 02/25/2022 17:13:25 - INFO - codeparrot_training - Step 30602: {'lr': 0.0001758333539367421, 'samples': 15668736, 'steps': 30602, 'loss/train': 1.9768928289413452} 02/25/2022 17:13:28 - INFO - codeparrot_training - Step 30603: {'lr': 0.0001758177282535857, 'samples': 15669248, 'steps': 30603, 'loss/train': 1.5133295059204102} 02/25/2022 17:13:34 - INFO - codeparrot_training - Step 30604: {'lr': 0.00017580210288820248, 'samples': 15669760, 'steps': 30604, 'loss/train': 3.128160238265991} 02/25/2022 17:13:37 - INFO - codeparrot_training - Step 30605: {'lr': 0.0001757864778406595, 'samples': 15670272, 'steps': 30605, 'loss/train': 1.8558063507080078} 02/25/2022 17:13:43 - INFO - codeparrot_training - Step 30606: {'lr': 0.00017577085311102364, 'samples': 15670784, 'steps': 30606, 'loss/train': 0.8960566520690918} 02/25/2022 17:13:46 - INFO - codeparrot_training - Step 30607: {'lr': 0.00017575522869936203, 'samples': 15671296, 'steps': 30607, 'loss/train': 0.9792315363883972} 02/25/2022 17:13:53 - INFO - codeparrot_training - Step 30608: {'lr': 0.00017573960460574132, 'samples': 15671808, 'steps': 30608, 'loss/train': 1.8218752145767212} 02/25/2022 17:13:56 - INFO - codeparrot_training - Step 30609: {'lr': 0.00017572398083022856, 'samples': 15672320, 'steps': 30609, 'loss/train': 0.870544970035553} 02/25/2022 17:14:02 - INFO - codeparrot_training - Step 30610: {'lr': 0.0001757083573728907, 'samples': 15672832, 'steps': 30610, 'loss/train': 0.1313645988702774} 02/25/2022 17:14:06 - INFO - codeparrot_training - Step 30611: {'lr': 0.0001756927342337946, 'samples': 15673344, 'steps': 30611, 'loss/train': 1.1649972200393677} 02/25/2022 17:14:11 - INFO - codeparrot_training - Step 30612: {'lr': 0.0001756771114130074, 'samples': 15673856, 'steps': 30612, 'loss/train': 2.251112937927246} 02/25/2022 17:14:15 - INFO - codeparrot_training - Step 30613: {'lr': 0.00017566148891059568, 'samples': 15674368, 'steps': 30613, 'loss/train': 0.2405310571193695} 02/25/2022 17:14:20 - INFO - codeparrot_training - Step 30614: {'lr': 0.0001756458667266266, 'samples': 15674880, 'steps': 30614, 'loss/train': 1.2505325078964233} 02/25/2022 17:14:24 - INFO - codeparrot_training - Step 30615: {'lr': 0.00017563024486116698, 'samples': 15675392, 'steps': 30615, 'loss/train': 1.6216870546340942} 02/25/2022 17:14:29 - INFO - codeparrot_training - Step 30616: {'lr': 0.0001756146233142839, 'samples': 15675904, 'steps': 30616, 'loss/train': 1.4520866870880127} 02/25/2022 17:14:33 - INFO - codeparrot_training - Step 30617: {'lr': 0.00017559900208604404, 'samples': 15676416, 'steps': 30617, 'loss/train': 0.8889603614807129} 02/25/2022 17:14:39 - INFO - codeparrot_training - Step 30618: {'lr': 0.00017558338117651447, 'samples': 15676928, 'steps': 30618, 'loss/train': 1.3003312349319458} 02/25/2022 17:14:42 - INFO - codeparrot_training - Step 30619: {'lr': 0.000175567760585762, 'samples': 15677440, 'steps': 30619, 'loss/train': 1.8794946670532227} 02/25/2022 17:14:48 - INFO - codeparrot_training - Step 30620: {'lr': 0.00017555214031385376, 'samples': 15677952, 'steps': 30620, 'loss/train': 0.7604542970657349} 02/25/2022 17:14:51 - INFO - codeparrot_training - Step 30621: {'lr': 0.00017553652036085638, 'samples': 15678464, 'steps': 30621, 'loss/train': 1.744019627571106} 02/25/2022 17:14:57 - INFO - codeparrot_training - Step 30622: {'lr': 0.00017552090072683692, 'samples': 15678976, 'steps': 30622, 'loss/train': 1.0111384391784668} 02/25/2022 17:15:00 - INFO - codeparrot_training - Step 30623: {'lr': 0.00017550528141186222, 'samples': 15679488, 'steps': 30623, 'loss/train': 0.04356483370065689} 02/25/2022 17:15:06 - INFO - codeparrot_training - Step 30624: {'lr': 0.00017548966241599933, 'samples': 15680000, 'steps': 30624, 'loss/train': 1.5989981889724731} 02/25/2022 17:15:09 - INFO - codeparrot_training - Step 30625: {'lr': 0.00017547404373931507, 'samples': 15680512, 'steps': 30625, 'loss/train': 2.618983030319214} 02/25/2022 17:15:15 - INFO - codeparrot_training - Step 30626: {'lr': 0.0001754584253818763, 'samples': 15681024, 'steps': 30626, 'loss/train': 1.4781371355056763} 02/25/2022 17:15:18 - INFO - codeparrot_training - Step 30627: {'lr': 0.00017544280734374995, 'samples': 15681536, 'steps': 30627, 'loss/train': 1.9581143856048584} 02/25/2022 17:15:25 - INFO - codeparrot_training - Step 30628: {'lr': 0.00017542718962500298, 'samples': 15682048, 'steps': 30628, 'loss/train': 1.5257078409194946} 02/25/2022 17:15:28 - INFO - codeparrot_training - Step 30629: {'lr': 0.00017541157222570228, 'samples': 15682560, 'steps': 30629, 'loss/train': 1.7244051694869995} 02/25/2022 17:15:34 - INFO - codeparrot_training - Step 30630: {'lr': 0.00017539595514591466, 'samples': 15683072, 'steps': 30630, 'loss/train': 2.757523536682129} 02/25/2022 17:15:37 - INFO - codeparrot_training - Step 30631: {'lr': 0.00017538033838570711, 'samples': 15683584, 'steps': 30631, 'loss/train': 0.046574871987104416} 02/25/2022 17:15:43 - INFO - codeparrot_training - Step 30632: {'lr': 0.00017536472194514647, 'samples': 15684096, 'steps': 30632, 'loss/train': 1.671276569366455} 02/25/2022 17:15:48 - INFO - codeparrot_training - Step 30633: {'lr': 0.0001753491058242997, 'samples': 15684608, 'steps': 30633, 'loss/train': 2.4795539379119873} 02/25/2022 17:15:52 - INFO - codeparrot_training - Step 30634: {'lr': 0.0001753334900232336, 'samples': 15685120, 'steps': 30634, 'loss/train': 2.11317777633667} 02/25/2022 17:15:58 - INFO - codeparrot_training - Step 30635: {'lr': 0.00017531787454201517, 'samples': 15685632, 'steps': 30635, 'loss/train': 1.6909476518630981} 02/25/2022 17:16:01 - INFO - codeparrot_training - Step 30636: {'lr': 0.0001753022593807112, 'samples': 15686144, 'steps': 30636, 'loss/train': 2.088340997695923} 02/25/2022 17:16:04 - INFO - codeparrot_training - Step 30637: {'lr': 0.00017528664453938863, 'samples': 15686656, 'steps': 30637, 'loss/train': 2.3470187187194824} 02/25/2022 17:16:11 - INFO - codeparrot_training - Step 30638: {'lr': 0.00017527103001811447, 'samples': 15687168, 'steps': 30638, 'loss/train': 0.9760718941688538} 02/25/2022 17:16:14 - INFO - codeparrot_training - Step 30639: {'lr': 0.0001752554158169554, 'samples': 15687680, 'steps': 30639, 'loss/train': 1.9636294841766357} 02/25/2022 17:16:20 - INFO - codeparrot_training - Step 30640: {'lr': 0.00017523980193597836, 'samples': 15688192, 'steps': 30640, 'loss/train': 1.869028925895691} 02/25/2022 17:16:23 - INFO - codeparrot_training - Step 30641: {'lr': 0.00017522418837525027, 'samples': 15688704, 'steps': 30641, 'loss/train': 2.451420545578003} 02/25/2022 17:16:29 - INFO - codeparrot_training - Step 30642: {'lr': 0.00017520857513483813, 'samples': 15689216, 'steps': 30642, 'loss/train': 2.1372947692871094} 02/25/2022 17:16:32 - INFO - codeparrot_training - Step 30643: {'lr': 0.00017519296221480863, 'samples': 15689728, 'steps': 30643, 'loss/train': 2.002358913421631} 02/25/2022 17:16:38 - INFO - codeparrot_training - Step 30644: {'lr': 0.0001751773496152287, 'samples': 15690240, 'steps': 30644, 'loss/train': 1.2552012205123901} 02/25/2022 17:16:41 - INFO - codeparrot_training - Step 30645: {'lr': 0.0001751617373361652, 'samples': 15690752, 'steps': 30645, 'loss/train': 2.1874241828918457} 02/25/2022 17:16:47 - INFO - codeparrot_training - Step 30646: {'lr': 0.0001751461253776851, 'samples': 15691264, 'steps': 30646, 'loss/train': 0.030398182570934296} 02/25/2022 17:16:50 - INFO - codeparrot_training - Step 30647: {'lr': 0.00017513051373985533, 'samples': 15691776, 'steps': 30647, 'loss/train': 1.8538587093353271} 02/25/2022 17:16:56 - INFO - codeparrot_training - Step 30648: {'lr': 0.00017511490242274255, 'samples': 15692288, 'steps': 30648, 'loss/train': 3.050234317779541} 02/25/2022 17:17:01 - INFO - codeparrot_training - Step 30649: {'lr': 0.00017509929142641377, 'samples': 15692800, 'steps': 30649, 'loss/train': 1.6641796827316284} 02/25/2022 17:17:04 - INFO - codeparrot_training - Step 30650: {'lr': 0.00017508368075093582, 'samples': 15693312, 'steps': 30650, 'loss/train': 2.202838897705078} 02/25/2022 17:17:10 - INFO - codeparrot_training - Step 30651: {'lr': 0.00017506807039637573, 'samples': 15693824, 'steps': 30651, 'loss/train': 0.5823962092399597} 02/25/2022 17:17:13 - INFO - codeparrot_training - Step 30652: {'lr': 0.0001750524603628001, 'samples': 15694336, 'steps': 30652, 'loss/train': 1.0130858421325684} 02/25/2022 17:17:20 - INFO - codeparrot_training - Step 30653: {'lr': 0.0001750368506502759, 'samples': 15694848, 'steps': 30653, 'loss/train': 2.7226009368896484} 02/25/2022 17:17:23 - INFO - codeparrot_training - Step 30654: {'lr': 0.0001750212412588701, 'samples': 15695360, 'steps': 30654, 'loss/train': 0.23124033212661743} 02/25/2022 17:17:29 - INFO - codeparrot_training - Step 30655: {'lr': 0.00017500563218864957, 'samples': 15695872, 'steps': 30655, 'loss/train': 0.8997575044631958} 02/25/2022 17:17:32 - INFO - codeparrot_training - Step 30656: {'lr': 0.00017499002343968097, 'samples': 15696384, 'steps': 30656, 'loss/train': 2.005222797393799} 02/25/2022 17:17:38 - INFO - codeparrot_training - Step 30657: {'lr': 0.00017497441501203133, 'samples': 15696896, 'steps': 30657, 'loss/train': 1.5784491300582886} 02/25/2022 17:17:41 - INFO - codeparrot_training - Step 30658: {'lr': 0.00017495880690576742, 'samples': 15697408, 'steps': 30658, 'loss/train': 0.8188974261283875} 02/25/2022 17:17:47 - INFO - codeparrot_training - Step 30659: {'lr': 0.00017494319912095618, 'samples': 15697920, 'steps': 30659, 'loss/train': 1.558914303779602} 02/25/2022 17:17:50 - INFO - codeparrot_training - Step 30660: {'lr': 0.00017492759165766458, 'samples': 15698432, 'steps': 30660, 'loss/train': 2.6022419929504395} 02/25/2022 17:17:56 - INFO - codeparrot_training - Step 30661: {'lr': 0.0001749119845159592, 'samples': 15698944, 'steps': 30661, 'loss/train': 2.226147174835205} 02/25/2022 17:17:59 - INFO - codeparrot_training - Step 30662: {'lr': 0.00017489637769590706, 'samples': 15699456, 'steps': 30662, 'loss/train': 1.746221899986267} 02/25/2022 17:18:05 - INFO - codeparrot_training - Step 30663: {'lr': 0.00017488077119757496, 'samples': 15699968, 'steps': 30663, 'loss/train': 1.8763507604599} 02/25/2022 17:18:09 - INFO - codeparrot_training - Step 30664: {'lr': 0.0001748651650210299, 'samples': 15700480, 'steps': 30664, 'loss/train': 1.6448935270309448} 02/25/2022 17:18:14 - INFO - codeparrot_training - Step 30665: {'lr': 0.00017484955916633848, 'samples': 15700992, 'steps': 30665, 'loss/train': 0.02377876080572605} 02/25/2022 17:18:18 - INFO - codeparrot_training - Step 30666: {'lr': 0.00017483395363356775, 'samples': 15701504, 'steps': 30666, 'loss/train': 1.5679619312286377} 02/25/2022 17:18:23 - INFO - codeparrot_training - Step 30667: {'lr': 0.00017481834842278444, 'samples': 15702016, 'steps': 30667, 'loss/train': 0.10999450832605362} 02/25/2022 17:18:27 - INFO - codeparrot_training - Step 30668: {'lr': 0.00017480274353405558, 'samples': 15702528, 'steps': 30668, 'loss/train': 0.04276614636182785} 02/25/2022 17:18:32 - INFO - codeparrot_training - Step 30669: {'lr': 0.0001747871389674478, 'samples': 15703040, 'steps': 30669, 'loss/train': 1.3618367910385132} 02/25/2022 17:18:36 - INFO - codeparrot_training - Step 30670: {'lr': 0.00017477153472302798, 'samples': 15703552, 'steps': 30670, 'loss/train': 2.252397298812866} 02/25/2022 17:18:41 - INFO - codeparrot_training - Step 30671: {'lr': 0.0001747559308008631, 'samples': 15704064, 'steps': 30671, 'loss/train': 1.4648576974868774} 02/25/2022 17:18:45 - INFO - codeparrot_training - Step 30672: {'lr': 0.0001747403272010199, 'samples': 15704576, 'steps': 30672, 'loss/train': 2.3063623905181885} 02/25/2022 17:18:51 - INFO - codeparrot_training - Step 30673: {'lr': 0.00017472472392356526, 'samples': 15705088, 'steps': 30673, 'loss/train': 0.09646202623844147} 02/25/2022 17:18:54 - INFO - codeparrot_training - Step 30674: {'lr': 0.00017470912096856594, 'samples': 15705600, 'steps': 30674, 'loss/train': 2.003654718399048} 02/25/2022 17:19:00 - INFO - codeparrot_training - Step 30675: {'lr': 0.00017469351833608894, 'samples': 15706112, 'steps': 30675, 'loss/train': 1.76346755027771} 02/25/2022 17:19:04 - INFO - codeparrot_training - Step 30676: {'lr': 0.00017467791602620086, 'samples': 15706624, 'steps': 30676, 'loss/train': 2.195906400680542} 02/25/2022 17:19:09 - INFO - codeparrot_training - Step 30677: {'lr': 0.00017466231403896878, 'samples': 15707136, 'steps': 30677, 'loss/train': 1.9850581884384155} 02/25/2022 17:19:13 - INFO - codeparrot_training - Step 30678: {'lr': 0.00017464671237445942, 'samples': 15707648, 'steps': 30678, 'loss/train': 1.6075907945632935} 02/25/2022 17:19:18 - INFO - codeparrot_training - Step 30679: {'lr': 0.00017463111103273958, 'samples': 15708160, 'steps': 30679, 'loss/train': 2.813119888305664} 02/25/2022 17:19:21 - INFO - codeparrot_training - Step 30680: {'lr': 0.0001746155100138761, 'samples': 15708672, 'steps': 30680, 'loss/train': 1.9931228160858154} 02/25/2022 17:19:27 - INFO - codeparrot_training - Step 30681: {'lr': 0.0001745999093179359, 'samples': 15709184, 'steps': 30681, 'loss/train': 2.007575273513794} 02/25/2022 17:19:30 - INFO - codeparrot_training - Step 30682: {'lr': 0.00017458430894498573, 'samples': 15709696, 'steps': 30682, 'loss/train': 2.521340847015381} 02/25/2022 17:19:36 - INFO - codeparrot_training - Step 30683: {'lr': 0.00017456870889509242, 'samples': 15710208, 'steps': 30683, 'loss/train': 1.8616303205490112} 02/25/2022 17:19:39 - INFO - codeparrot_training - Step 30684: {'lr': 0.00017455310916832275, 'samples': 15710720, 'steps': 30684, 'loss/train': 0.9796485304832458} 02/25/2022 17:19:45 - INFO - codeparrot_training - Step 30685: {'lr': 0.00017453750976474363, 'samples': 15711232, 'steps': 30685, 'loss/train': 0.5382975935935974} 02/25/2022 17:19:48 - INFO - codeparrot_training - Step 30686: {'lr': 0.00017452191068442196, 'samples': 15711744, 'steps': 30686, 'loss/train': 1.6235179901123047} 02/25/2022 17:19:54 - INFO - codeparrot_training - Step 30687: {'lr': 0.0001745063119274244, 'samples': 15712256, 'steps': 30687, 'loss/train': 2.1804840564727783} 02/25/2022 17:19:57 - INFO - codeparrot_training - Step 30688: {'lr': 0.00017449071349381776, 'samples': 15712768, 'steps': 30688, 'loss/train': 1.2717500925064087} 02/25/2022 17:20:04 - INFO - codeparrot_training - Step 30689: {'lr': 0.00017447511538366895, 'samples': 15713280, 'steps': 30689, 'loss/train': 1.976916790008545} 02/25/2022 17:20:07 - INFO - codeparrot_training - Step 30690: {'lr': 0.00017445951759704486, 'samples': 15713792, 'steps': 30690, 'loss/train': 1.6629647016525269} 02/25/2022 17:20:13 - INFO - codeparrot_training - Step 30691: {'lr': 0.00017444392013401213, 'samples': 15714304, 'steps': 30691, 'loss/train': 1.923917531967163} 02/25/2022 17:20:18 - INFO - codeparrot_training - Step 30692: {'lr': 0.00017442832299463762, 'samples': 15714816, 'steps': 30692, 'loss/train': 1.995125412940979} 02/25/2022 17:20:21 - INFO - codeparrot_training - Step 30693: {'lr': 0.00017441272617898824, 'samples': 15715328, 'steps': 30693, 'loss/train': 1.0246213674545288} 02/25/2022 17:20:27 - INFO - codeparrot_training - Step 30694: {'lr': 0.00017439712968713068, 'samples': 15715840, 'steps': 30694, 'loss/train': 1.8903381824493408} 02/25/2022 17:20:31 - INFO - codeparrot_training - Step 30695: {'lr': 0.00017438153351913193, 'samples': 15716352, 'steps': 30695, 'loss/train': 0.8537416458129883} 02/25/2022 17:20:36 - INFO - codeparrot_training - Step 30696: {'lr': 0.00017436593767505859, 'samples': 15716864, 'steps': 30696, 'loss/train': 1.5773813724517822} 02/25/2022 17:20:40 - INFO - codeparrot_training - Step 30697: {'lr': 0.00017435034215497754, 'samples': 15717376, 'steps': 30697, 'loss/train': 1.8662402629852295} 02/25/2022 17:20:46 - INFO - codeparrot_training - Step 30698: {'lr': 0.0001743347469589556, 'samples': 15717888, 'steps': 30698, 'loss/train': 1.4936827421188354} 02/25/2022 17:20:49 - INFO - codeparrot_training - Step 30699: {'lr': 0.00017431915208705967, 'samples': 15718400, 'steps': 30699, 'loss/train': 2.2123234272003174} 02/25/2022 17:20:55 - INFO - codeparrot_training - Step 30700: {'lr': 0.0001743035575393564, 'samples': 15718912, 'steps': 30700, 'loss/train': 1.4433770179748535} 02/25/2022 17:20:58 - INFO - codeparrot_training - Step 30701: {'lr': 0.00017428796331591263, 'samples': 15719424, 'steps': 30701, 'loss/train': 2.867765188217163} 02/25/2022 17:21:04 - INFO - codeparrot_training - Step 30702: {'lr': 0.00017427236941679517, 'samples': 15719936, 'steps': 30702, 'loss/train': 1.2954092025756836} 02/25/2022 17:21:07 - INFO - codeparrot_training - Step 30703: {'lr': 0.00017425677584207095, 'samples': 15720448, 'steps': 30703, 'loss/train': 2.0587944984436035} 02/25/2022 17:21:13 - INFO - codeparrot_training - Step 30704: {'lr': 0.00017424118259180656, 'samples': 15720960, 'steps': 30704, 'loss/train': 1.3188984394073486} 02/25/2022 17:21:16 - INFO - codeparrot_training - Step 30705: {'lr': 0.00017422558966606887, 'samples': 15721472, 'steps': 30705, 'loss/train': 1.73975670337677} 02/25/2022 17:21:22 - INFO - codeparrot_training - Step 30706: {'lr': 0.0001742099970649247, 'samples': 15721984, 'steps': 30706, 'loss/train': 1.804073452949524} 02/25/2022 17:21:25 - INFO - codeparrot_training - Step 30707: {'lr': 0.00017419440478844082, 'samples': 15722496, 'steps': 30707, 'loss/train': 1.6832637786865234} 02/25/2022 17:21:31 - INFO - codeparrot_training - Step 30708: {'lr': 0.00017417881283668412, 'samples': 15723008, 'steps': 30708, 'loss/train': 2.0426106452941895} 02/25/2022 17:21:34 - INFO - codeparrot_training - Step 30709: {'lr': 0.00017416322120972127, 'samples': 15723520, 'steps': 30709, 'loss/train': 0.1720254123210907} 02/25/2022 17:21:40 - INFO - codeparrot_training - Step 30710: {'lr': 0.00017414762990761908, 'samples': 15724032, 'steps': 30710, 'loss/train': 1.8328558206558228} 02/25/2022 17:21:44 - INFO - codeparrot_training - Step 30711: {'lr': 0.0001741320389304444, 'samples': 15724544, 'steps': 30711, 'loss/train': 1.8271600008010864} 02/25/2022 17:21:49 - INFO - codeparrot_training - Step 30712: {'lr': 0.00017411644827826394, 'samples': 15725056, 'steps': 30712, 'loss/train': 1.6123133897781372} 02/25/2022 17:21:53 - INFO - codeparrot_training - Step 30713: {'lr': 0.0001741008579511445, 'samples': 15725568, 'steps': 30713, 'loss/train': 1.570552110671997} 02/25/2022 17:21:58 - INFO - codeparrot_training - Step 30714: {'lr': 0.00017408526794915292, 'samples': 15726080, 'steps': 30714, 'loss/train': 1.0803930759429932} 02/25/2022 17:22:02 - INFO - codeparrot_training - Step 30715: {'lr': 0.00017406967827235598, 'samples': 15726592, 'steps': 30715, 'loss/train': 2.446274995803833} 02/25/2022 17:22:07 - INFO - codeparrot_training - Step 30716: {'lr': 0.0001740540889208204, 'samples': 15727104, 'steps': 30716, 'loss/train': 1.8607929944992065} 02/25/2022 17:22:11 - INFO - codeparrot_training - Step 30717: {'lr': 0.00017403849989461296, 'samples': 15727616, 'steps': 30717, 'loss/train': 1.765053153038025} 02/25/2022 17:22:16 - INFO - codeparrot_training - Step 30718: {'lr': 0.00017402291119380047, 'samples': 15728128, 'steps': 30718, 'loss/train': 2.0131499767303467} 02/25/2022 17:22:20 - INFO - codeparrot_training - Step 30719: {'lr': 0.0001740073228184497, 'samples': 15728640, 'steps': 30719, 'loss/train': 1.424175500869751} 02/25/2022 17:22:26 - INFO - codeparrot_training - Step 30720: {'lr': 0.00017399173476862742, 'samples': 15729152, 'steps': 30720, 'loss/train': 3.1434195041656494} 02/25/2022 17:22:29 - INFO - codeparrot_training - Step 30721: {'lr': 0.00017397614704440045, 'samples': 15729664, 'steps': 30721, 'loss/train': 0.7791910171508789} 02/25/2022 17:22:35 - INFO - codeparrot_training - Step 30722: {'lr': 0.00017396055964583558, 'samples': 15730176, 'steps': 30722, 'loss/train': 1.2293877601623535} 02/25/2022 17:22:38 - INFO - codeparrot_training - Step 30723: {'lr': 0.00017394497257299942, 'samples': 15730688, 'steps': 30723, 'loss/train': 2.2874598503112793} 02/25/2022 17:22:44 - INFO - codeparrot_training - Step 30724: {'lr': 0.00017392938582595885, 'samples': 15731200, 'steps': 30724, 'loss/train': 1.5270161628723145} 02/25/2022 17:22:47 - INFO - codeparrot_training - Step 30725: {'lr': 0.00017391379940478075, 'samples': 15731712, 'steps': 30725, 'loss/train': 2.097254991531372} 02/25/2022 17:22:53 - INFO - codeparrot_training - Step 30726: {'lr': 0.00017389821330953165, 'samples': 15732224, 'steps': 30726, 'loss/train': 2.503765821456909} 02/25/2022 17:22:56 - INFO - codeparrot_training - Step 30727: {'lr': 0.0001738826275402785, 'samples': 15732736, 'steps': 30727, 'loss/train': 3.621640920639038} 02/25/2022 17:23:02 - INFO - codeparrot_training - Step 30728: {'lr': 0.00017386704209708794, 'samples': 15733248, 'steps': 30728, 'loss/train': 2.242464303970337} 02/25/2022 17:23:05 - INFO - codeparrot_training - Step 30729: {'lr': 0.0001738514569800269, 'samples': 15733760, 'steps': 30729, 'loss/train': 2.215031147003174} 02/25/2022 17:23:11 - INFO - codeparrot_training - Step 30730: {'lr': 0.00017383587218916196, 'samples': 15734272, 'steps': 30730, 'loss/train': 1.9182744026184082} 02/25/2022 17:23:15 - INFO - codeparrot_training - Step 30731: {'lr': 0.00017382028772455996, 'samples': 15734784, 'steps': 30731, 'loss/train': 1.1504204273223877} 02/25/2022 17:23:20 - INFO - codeparrot_training - Step 30732: {'lr': 0.00017380470358628763, 'samples': 15735296, 'steps': 30732, 'loss/train': 1.8763405084609985} 02/25/2022 17:23:24 - INFO - codeparrot_training - Step 30733: {'lr': 0.0001737891197744118, 'samples': 15735808, 'steps': 30733, 'loss/train': 1.8437106609344482} 02/25/2022 17:23:30 - INFO - codeparrot_training - Step 30734: {'lr': 0.00017377353628899927, 'samples': 15736320, 'steps': 30734, 'loss/train': 1.661285400390625} 02/25/2022 17:23:34 - INFO - codeparrot_training - Step 30735: {'lr': 0.0001737579531301166, 'samples': 15736832, 'steps': 30735, 'loss/train': 1.0417697429656982} 02/25/2022 17:23:39 - INFO - codeparrot_training - Step 30736: {'lr': 0.00017374237029783062, 'samples': 15737344, 'steps': 30736, 'loss/train': 1.5833008289337158} 02/25/2022 17:23:43 - INFO - codeparrot_training - Step 30737: {'lr': 0.00017372678779220815, 'samples': 15737856, 'steps': 30737, 'loss/train': 1.7285505533218384} 02/25/2022 17:23:48 - INFO - codeparrot_training - Step 30738: {'lr': 0.00017371120561331598, 'samples': 15738368, 'steps': 30738, 'loss/train': 1.145124912261963} 02/25/2022 17:23:52 - INFO - codeparrot_training - Step 30739: {'lr': 0.00017369562376122073, 'samples': 15738880, 'steps': 30739, 'loss/train': 2.554356098175049} 02/25/2022 17:23:57 - INFO - codeparrot_training - Step 30740: {'lr': 0.00017368004223598912, 'samples': 15739392, 'steps': 30740, 'loss/train': 2.5362050533294678} 02/25/2022 17:24:01 - INFO - codeparrot_training - Step 30741: {'lr': 0.000173664461037688, 'samples': 15739904, 'steps': 30741, 'loss/train': 2.315133810043335} 02/25/2022 17:24:06 - INFO - codeparrot_training - Step 30742: {'lr': 0.00017364888016638413, 'samples': 15740416, 'steps': 30742, 'loss/train': 0.9000276327133179} 02/25/2022 17:24:10 - INFO - codeparrot_training - Step 30743: {'lr': 0.00017363329962214428, 'samples': 15740928, 'steps': 30743, 'loss/train': 1.5901893377304077} 02/25/2022 17:24:15 - INFO - codeparrot_training - Step 30744: {'lr': 0.00017361771940503502, 'samples': 15741440, 'steps': 30744, 'loss/train': 2.456793785095215} 02/25/2022 17:24:19 - INFO - codeparrot_training - Step 30745: {'lr': 0.00017360213951512322, 'samples': 15741952, 'steps': 30745, 'loss/train': 1.951997995376587} 02/25/2022 17:24:26 - INFO - codeparrot_training - Step 30746: {'lr': 0.00017358655995247557, 'samples': 15742464, 'steps': 30746, 'loss/train': 1.5600237846374512} 02/25/2022 17:24:29 - INFO - codeparrot_training - Step 30747: {'lr': 0.00017357098071715894, 'samples': 15742976, 'steps': 30747, 'loss/train': 1.66391921043396} 02/25/2022 17:24:35 - INFO - codeparrot_training - Step 30748: {'lr': 0.00017355540180923987, 'samples': 15743488, 'steps': 30748, 'loss/train': 2.0510079860687256} 02/25/2022 17:24:40 - INFO - codeparrot_training - Step 30749: {'lr': 0.0001735398232287852, 'samples': 15744000, 'steps': 30749, 'loss/train': 2.2184791564941406} 02/25/2022 17:24:44 - INFO - codeparrot_training - Step 30750: {'lr': 0.0001735242449758616, 'samples': 15744512, 'steps': 30750, 'loss/train': 2.3929343223571777} 02/25/2022 17:24:49 - INFO - codeparrot_training - Step 30751: {'lr': 0.00017350866705053596, 'samples': 15745024, 'steps': 30751, 'loss/train': 2.2873752117156982} 02/25/2022 17:24:53 - INFO - codeparrot_training - Step 30752: {'lr': 0.00017349308945287484, 'samples': 15745536, 'steps': 30752, 'loss/train': 1.360449194908142} 02/25/2022 17:24:58 - INFO - codeparrot_training - Step 30753: {'lr': 0.000173477512182945, 'samples': 15746048, 'steps': 30753, 'loss/train': 2.048659324645996} 02/25/2022 17:25:02 - INFO - codeparrot_training - Step 30754: {'lr': 0.00017346193524081323, 'samples': 15746560, 'steps': 30754, 'loss/train': 1.7202531099319458} 02/25/2022 17:25:08 - INFO - codeparrot_training - Step 30755: {'lr': 0.00017344635862654624, 'samples': 15747072, 'steps': 30755, 'loss/train': 0.8473336696624756} 02/25/2022 17:25:12 - INFO - codeparrot_training - Step 30756: {'lr': 0.00017343078234021076, 'samples': 15747584, 'steps': 30756, 'loss/train': 1.283828854560852} 02/25/2022 17:25:17 - INFO - codeparrot_training - Step 30757: {'lr': 0.00017341520638187346, 'samples': 15748096, 'steps': 30757, 'loss/train': 1.289860725402832} 02/25/2022 17:25:21 - INFO - codeparrot_training - Step 30758: {'lr': 0.00017339963075160112, 'samples': 15748608, 'steps': 30758, 'loss/train': 2.3887338638305664} 02/25/2022 17:25:26 - INFO - codeparrot_training - Step 30759: {'lr': 0.0001733840554494604, 'samples': 15749120, 'steps': 30759, 'loss/train': 0.19671623408794403} 02/25/2022 17:25:30 - INFO - codeparrot_training - Step 30760: {'lr': 0.00017336848047551812, 'samples': 15749632, 'steps': 30760, 'loss/train': 1.5675932168960571} 02/25/2022 17:25:35 - INFO - codeparrot_training - Step 30761: {'lr': 0.00017335290582984087, 'samples': 15750144, 'steps': 30761, 'loss/train': 1.2926170825958252} 02/25/2022 17:25:39 - INFO - codeparrot_training - Step 30762: {'lr': 0.00017333733151249549, 'samples': 15750656, 'steps': 30762, 'loss/train': 1.3450815677642822} 02/25/2022 17:25:44 - INFO - codeparrot_training - Step 30763: {'lr': 0.00017332175752354857, 'samples': 15751168, 'steps': 30763, 'loss/train': 0.8815481066703796} 02/25/2022 17:25:48 - INFO - codeparrot_training - Step 30764: {'lr': 0.00017330618386306697, 'samples': 15751680, 'steps': 30764, 'loss/train': 2.398009777069092} 02/25/2022 17:25:55 - INFO - codeparrot_training - Step 30765: {'lr': 0.00017329061053111724, 'samples': 15752192, 'steps': 30765, 'loss/train': 2.5809693336486816} 02/25/2022 17:25:58 - INFO - codeparrot_training - Step 30766: {'lr': 0.00017327503752776624, 'samples': 15752704, 'steps': 30766, 'loss/train': 1.4935481548309326} 02/25/2022 17:26:04 - INFO - codeparrot_training - Step 30767: {'lr': 0.00017325946485308058, 'samples': 15753216, 'steps': 30767, 'loss/train': 2.256352663040161} 02/25/2022 17:26:07 - INFO - codeparrot_training - Step 30768: {'lr': 0.00017324389250712702, 'samples': 15753728, 'steps': 30768, 'loss/train': 1.2551031112670898} 02/25/2022 17:26:13 - INFO - codeparrot_training - Step 30769: {'lr': 0.00017322832048997233, 'samples': 15754240, 'steps': 30769, 'loss/train': 1.4574698209762573} 02/25/2022 17:26:16 - INFO - codeparrot_training - Step 30770: {'lr': 0.00017321274880168302, 'samples': 15754752, 'steps': 30770, 'loss/train': 0.5906045436859131} 02/25/2022 17:26:22 - INFO - codeparrot_training - Step 30771: {'lr': 0.00017319717744232593, 'samples': 15755264, 'steps': 30771, 'loss/train': 1.4669440984725952} 02/25/2022 17:26:25 - INFO - codeparrot_training - Step 30772: {'lr': 0.00017318160641196777, 'samples': 15755776, 'steps': 30772, 'loss/train': 1.480870008468628} 02/25/2022 17:26:31 - INFO - codeparrot_training - Step 30773: {'lr': 0.0001731660357106753, 'samples': 15756288, 'steps': 30773, 'loss/train': 0.9543995261192322} 02/25/2022 17:26:34 - INFO - codeparrot_training - Step 30774: {'lr': 0.000173150465338515, 'samples': 15756800, 'steps': 30774, 'loss/train': 1.2802789211273193} 02/25/2022 17:26:39 - INFO - codeparrot_training - Step 30775: {'lr': 0.00017313489529555373, 'samples': 15757312, 'steps': 30775, 'loss/train': 2.2093520164489746} 02/25/2022 17:26:43 - INFO - codeparrot_training - Step 30776: {'lr': 0.00017311932558185817, 'samples': 15757824, 'steps': 30776, 'loss/train': 1.8110259771347046} 02/25/2022 17:26:48 - INFO - codeparrot_training - Step 30777: {'lr': 0.00017310375619749508, 'samples': 15758336, 'steps': 30777, 'loss/train': 1.2443310022354126} 02/25/2022 17:26:52 - INFO - codeparrot_training - Step 30778: {'lr': 0.000173088187142531, 'samples': 15758848, 'steps': 30778, 'loss/train': 1.9007090330123901} 02/25/2022 17:26:58 - INFO - codeparrot_training - Step 30779: {'lr': 0.0001730726184170327, 'samples': 15759360, 'steps': 30779, 'loss/train': 1.0733305215835571} 02/25/2022 17:27:01 - INFO - codeparrot_training - Step 30780: {'lr': 0.00017305705002106686, 'samples': 15759872, 'steps': 30780, 'loss/train': 1.1906702518463135} 02/25/2022 17:27:07 - INFO - codeparrot_training - Step 30781: {'lr': 0.00017304148195470021, 'samples': 15760384, 'steps': 30781, 'loss/train': 1.8607711791992188} 02/25/2022 17:27:10 - INFO - codeparrot_training - Step 30782: {'lr': 0.00017302591421799953, 'samples': 15760896, 'steps': 30782, 'loss/train': 1.4264580011367798} 02/25/2022 17:27:16 - INFO - codeparrot_training - Step 30783: {'lr': 0.00017301034681103128, 'samples': 15761408, 'steps': 30783, 'loss/train': 0.6306841969490051} 02/25/2022 17:27:19 - INFO - codeparrot_training - Step 30784: {'lr': 0.00017299477973386224, 'samples': 15761920, 'steps': 30784, 'loss/train': 1.4932963848114014} 02/25/2022 17:27:25 - INFO - codeparrot_training - Step 30785: {'lr': 0.0001729792129865591, 'samples': 15762432, 'steps': 30785, 'loss/train': 2.1149187088012695} 02/25/2022 17:27:29 - INFO - codeparrot_training - Step 30786: {'lr': 0.00017296364656918867, 'samples': 15762944, 'steps': 30786, 'loss/train': 2.4017672538757324} 02/25/2022 17:27:34 - INFO - codeparrot_training - Step 30787: {'lr': 0.00017294808048181742, 'samples': 15763456, 'steps': 30787, 'loss/train': 1.8204617500305176} 02/25/2022 17:27:38 - INFO - codeparrot_training - Step 30788: {'lr': 0.00017293251472451216, 'samples': 15763968, 'steps': 30788, 'loss/train': 1.8793182373046875} 02/25/2022 17:27:43 - INFO - codeparrot_training - Step 30789: {'lr': 0.00017291694929733948, 'samples': 15764480, 'steps': 30789, 'loss/train': 0.851350724697113} 02/25/2022 17:27:47 - INFO - codeparrot_training - Step 30790: {'lr': 0.00017290138420036623, 'samples': 15764992, 'steps': 30790, 'loss/train': 2.3421406745910645} 02/25/2022 17:27:53 - INFO - codeparrot_training - Step 30791: {'lr': 0.00017288581943365888, 'samples': 15765504, 'steps': 30791, 'loss/train': 0.8024457097053528} 02/25/2022 17:27:56 - INFO - codeparrot_training - Step 30792: {'lr': 0.00017287025499728418, 'samples': 15766016, 'steps': 30792, 'loss/train': 0.7061970233917236} 02/25/2022 17:28:02 - INFO - codeparrot_training - Step 30793: {'lr': 0.00017285469089130883, 'samples': 15766528, 'steps': 30793, 'loss/train': 0.47067826986312866} 02/25/2022 17:28:05 - INFO - codeparrot_training - Step 30794: {'lr': 0.00017283912711579946, 'samples': 15767040, 'steps': 30794, 'loss/train': 2.7015745639801025} 02/25/2022 17:28:11 - INFO - codeparrot_training - Step 30795: {'lr': 0.00017282356367082292, 'samples': 15767552, 'steps': 30795, 'loss/train': 1.3296995162963867} 02/25/2022 17:28:14 - INFO - codeparrot_training - Step 30796: {'lr': 0.00017280800055644559, 'samples': 15768064, 'steps': 30796, 'loss/train': 1.2685362100601196} 02/25/2022 17:28:20 - INFO - codeparrot_training - Step 30797: {'lr': 0.00017279243777273425, 'samples': 15768576, 'steps': 30797, 'loss/train': 1.469246745109558} 02/25/2022 17:28:23 - INFO - codeparrot_training - Step 30798: {'lr': 0.00017277687531975567, 'samples': 15769088, 'steps': 30798, 'loss/train': 1.4838075637817383} 02/25/2022 17:28:29 - INFO - codeparrot_training - Step 30799: {'lr': 0.00017276131319757644, 'samples': 15769600, 'steps': 30799, 'loss/train': 1.4600505828857422} 02/25/2022 17:28:32 - INFO - codeparrot_training - Step 30800: {'lr': 0.00017274575140626317, 'samples': 15770112, 'steps': 30800, 'loss/train': 0.3699166178703308} 02/25/2022 17:28:39 - INFO - codeparrot_training - Step 30801: {'lr': 0.00017273018994588257, 'samples': 15770624, 'steps': 30801, 'loss/train': 0.6264426112174988} 02/25/2022 17:28:42 - INFO - codeparrot_training - Step 30802: {'lr': 0.0001727146288165013, 'samples': 15771136, 'steps': 30802, 'loss/train': 1.7573096752166748} 02/25/2022 17:28:48 - INFO - codeparrot_training - Step 30803: {'lr': 0.00017269906801818603, 'samples': 15771648, 'steps': 30803, 'loss/train': 2.4332785606384277} 02/25/2022 17:28:51 - INFO - codeparrot_training - Step 30804: {'lr': 0.00017268350755100345, 'samples': 15772160, 'steps': 30804, 'loss/train': 1.8508580923080444} 02/25/2022 17:28:57 - INFO - codeparrot_training - Step 30805: {'lr': 0.00017266794741502013, 'samples': 15772672, 'steps': 30805, 'loss/train': 1.782822608947754} 02/25/2022 17:29:00 - INFO - codeparrot_training - Step 30806: {'lr': 0.00017265238761030277, 'samples': 15773184, 'steps': 30806, 'loss/train': 1.3476858139038086} 02/25/2022 17:29:06 - INFO - codeparrot_training - Step 30807: {'lr': 0.000172636828136918, 'samples': 15773696, 'steps': 30807, 'loss/train': 2.2343435287475586} 02/25/2022 17:29:09 - INFO - codeparrot_training - Step 30808: {'lr': 0.00017262126899493256, 'samples': 15774208, 'steps': 30808, 'loss/train': 0.7837716341018677} 02/25/2022 17:29:15 - INFO - codeparrot_training - Step 30809: {'lr': 0.000172605710184413, 'samples': 15774720, 'steps': 30809, 'loss/train': 2.136336326599121} 02/25/2022 17:29:18 - INFO - codeparrot_training - Step 30810: {'lr': 0.00017259015170542597, 'samples': 15775232, 'steps': 30810, 'loss/train': 1.6527882814407349} 02/25/2022 17:29:25 - INFO - codeparrot_training - Step 30811: {'lr': 0.00017257459355803816, 'samples': 15775744, 'steps': 30811, 'loss/train': 2.1016275882720947} 02/25/2022 17:29:28 - INFO - codeparrot_training - Step 30812: {'lr': 0.00017255903574231625, 'samples': 15776256, 'steps': 30812, 'loss/train': 2.0298404693603516} 02/25/2022 17:29:34 - INFO - codeparrot_training - Step 30813: {'lr': 0.00017254347825832684, 'samples': 15776768, 'steps': 30813, 'loss/train': 1.8232800960540771} 02/25/2022 17:29:37 - INFO - codeparrot_training - Step 30814: {'lr': 0.00017252792110613652, 'samples': 15777280, 'steps': 30814, 'loss/train': 8.336360931396484} 02/25/2022 17:29:43 - INFO - codeparrot_training - Step 30815: {'lr': 0.00017251236428581196, 'samples': 15777792, 'steps': 30815, 'loss/train': 1.1440787315368652} 02/25/2022 17:29:46 - INFO - codeparrot_training - Step 30816: {'lr': 0.00017249680779741987, 'samples': 15778304, 'steps': 30816, 'loss/train': 1.5922085046768188} 02/25/2022 17:29:52 - INFO - codeparrot_training - Step 30817: {'lr': 0.0001724812516410269, 'samples': 15778816, 'steps': 30817, 'loss/train': 1.9258041381835938} 02/25/2022 17:29:55 - INFO - codeparrot_training - Step 30818: {'lr': 0.00017246569581669957, 'samples': 15779328, 'steps': 30818, 'loss/train': 1.764691710472107} 02/25/2022 17:30:01 - INFO - codeparrot_training - Step 30819: {'lr': 0.00017245014032450456, 'samples': 15779840, 'steps': 30819, 'loss/train': 0.8716151118278503} 02/25/2022 17:30:05 - INFO - codeparrot_training - Step 30820: {'lr': 0.00017243458516450855, 'samples': 15780352, 'steps': 30820, 'loss/train': 2.1852333545684814} 02/25/2022 17:30:10 - INFO - codeparrot_training - Step 30821: {'lr': 0.00017241903033677822, 'samples': 15780864, 'steps': 30821, 'loss/train': 1.449008584022522} 02/25/2022 17:30:14 - INFO - codeparrot_training - Step 30822: {'lr': 0.00017240347584138006, 'samples': 15781376, 'steps': 30822, 'loss/train': 1.9347764253616333} 02/25/2022 17:30:19 - INFO - codeparrot_training - Step 30823: {'lr': 0.00017238792167838073, 'samples': 15781888, 'steps': 30823, 'loss/train': 2.350262403488159} 02/25/2022 17:30:25 - INFO - codeparrot_training - Step 30824: {'lr': 0.00017237236784784692, 'samples': 15782400, 'steps': 30824, 'loss/train': 1.3809736967086792} 02/25/2022 17:30:28 - INFO - codeparrot_training - Step 30825: {'lr': 0.00017235681434984532, 'samples': 15782912, 'steps': 30825, 'loss/train': 1.7339568138122559} 02/25/2022 17:30:34 - INFO - codeparrot_training - Step 30826: {'lr': 0.00017234126118444238, 'samples': 15783424, 'steps': 30826, 'loss/train': 1.257758617401123} 02/25/2022 17:30:38 - INFO - codeparrot_training - Step 30827: {'lr': 0.00017232570835170482, 'samples': 15783936, 'steps': 30827, 'loss/train': 1.5525656938552856} 02/25/2022 17:30:43 - INFO - codeparrot_training - Step 30828: {'lr': 0.00017231015585169928, 'samples': 15784448, 'steps': 30828, 'loss/train': 1.4200551509857178} 02/25/2022 17:30:47 - INFO - codeparrot_training - Step 30829: {'lr': 0.00017229460368449234, 'samples': 15784960, 'steps': 30829, 'loss/train': 1.3803961277008057} 02/25/2022 17:30:52 - INFO - codeparrot_training - Step 30830: {'lr': 0.00017227905185015075, 'samples': 15785472, 'steps': 30830, 'loss/train': 1.514911413192749} 02/25/2022 17:30:56 - INFO - codeparrot_training - Step 30831: {'lr': 0.00017226350034874096, 'samples': 15785984, 'steps': 30831, 'loss/train': 3.714461088180542} 02/25/2022 17:31:01 - INFO - codeparrot_training - Step 30832: {'lr': 0.0001722479491803296, 'samples': 15786496, 'steps': 30832, 'loss/train': 2.6190905570983887} 02/25/2022 17:31:05 - INFO - codeparrot_training - Step 30833: {'lr': 0.00017223239834498333, 'samples': 15787008, 'steps': 30833, 'loss/train': 1.3432352542877197} 02/25/2022 17:31:10 - INFO - codeparrot_training - Step 30834: {'lr': 0.0001722168478427689, 'samples': 15787520, 'steps': 30834, 'loss/train': 1.4192978143692017} 02/25/2022 17:31:14 - INFO - codeparrot_training - Step 30835: {'lr': 0.0001722012976737527, 'samples': 15788032, 'steps': 30835, 'loss/train': 1.192130446434021} 02/25/2022 17:31:20 - INFO - codeparrot_training - Step 30836: {'lr': 0.0001721857478380014, 'samples': 15788544, 'steps': 30836, 'loss/train': 2.4973342418670654} 02/25/2022 17:31:23 - INFO - codeparrot_training - Step 30837: {'lr': 0.00017217019833558168, 'samples': 15789056, 'steps': 30837, 'loss/train': 2.3958022594451904} 02/25/2022 17:31:29 - INFO - codeparrot_training - Step 30838: {'lr': 0.0001721546491665602, 'samples': 15789568, 'steps': 30838, 'loss/train': 0.4230477809906006} 02/25/2022 17:31:32 - INFO - codeparrot_training - Step 30839: {'lr': 0.0001721391003310034, 'samples': 15790080, 'steps': 30839, 'loss/train': 1.9959704875946045} 02/25/2022 17:31:38 - INFO - codeparrot_training - Step 30840: {'lr': 0.00017212355182897799, 'samples': 15790592, 'steps': 30840, 'loss/train': 2.0177571773529053} 02/25/2022 17:31:41 - INFO - codeparrot_training - Step 30841: {'lr': 0.0001721080036605505, 'samples': 15791104, 'steps': 30841, 'loss/train': 1.3669719696044922} 02/25/2022 17:31:47 - INFO - codeparrot_training - Step 30842: {'lr': 0.0001720924558257877, 'samples': 15791616, 'steps': 30842, 'loss/train': 1.751502275466919} 02/25/2022 17:31:50 - INFO - codeparrot_training - Step 30843: {'lr': 0.00017207690832475602, 'samples': 15792128, 'steps': 30843, 'loss/train': 2.3211817741394043} 02/25/2022 17:31:56 - INFO - codeparrot_training - Step 30844: {'lr': 0.0001720613611575221, 'samples': 15792640, 'steps': 30844, 'loss/train': 2.121504783630371} 02/25/2022 17:31:59 - INFO - codeparrot_training - Step 30845: {'lr': 0.00017204581432415255, 'samples': 15793152, 'steps': 30845, 'loss/train': 1.253933072090149} 02/25/2022 17:32:06 - INFO - codeparrot_training - Step 30846: {'lr': 0.00017203026782471403, 'samples': 15793664, 'steps': 30846, 'loss/train': 1.7700453996658325} 02/25/2022 17:32:09 - INFO - codeparrot_training - Step 30847: {'lr': 0.00017201472165927312, 'samples': 15794176, 'steps': 30847, 'loss/train': 0.7101238369941711} 02/25/2022 17:32:15 - INFO - codeparrot_training - Step 30848: {'lr': 0.00017199917582789631, 'samples': 15794688, 'steps': 30848, 'loss/train': 1.5261906385421753} 02/25/2022 17:32:18 - INFO - codeparrot_training - Step 30849: {'lr': 0.0001719836303306503, 'samples': 15795200, 'steps': 30849, 'loss/train': 1.2616701126098633} 02/25/2022 17:32:24 - INFO - codeparrot_training - Step 30850: {'lr': 0.00017196808516760159, 'samples': 15795712, 'steps': 30850, 'loss/train': 1.6685019731521606} 02/25/2022 17:32:27 - INFO - codeparrot_training - Step 30851: {'lr': 0.00017195254033881686, 'samples': 15796224, 'steps': 30851, 'loss/train': 3.0583178997039795} 02/25/2022 17:32:33 - INFO - codeparrot_training - Step 30852: {'lr': 0.00017193699584436268, 'samples': 15796736, 'steps': 30852, 'loss/train': 1.7429862022399902} 02/25/2022 17:32:36 - INFO - codeparrot_training - Step 30853: {'lr': 0.0001719214516843056, 'samples': 15797248, 'steps': 30853, 'loss/train': 1.7161043882369995} 02/25/2022 17:32:42 - INFO - codeparrot_training - Step 30854: {'lr': 0.00017190590785871225, 'samples': 15797760, 'steps': 30854, 'loss/train': 0.8770737648010254} 02/25/2022 17:32:45 - INFO - codeparrot_training - Step 30855: {'lr': 0.00017189036436764913, 'samples': 15798272, 'steps': 30855, 'loss/train': 0.0209540706127882} 02/25/2022 17:32:51 - INFO - codeparrot_training - Step 30856: {'lr': 0.00017187482121118304, 'samples': 15798784, 'steps': 30856, 'loss/train': 2.1271042823791504} 02/25/2022 17:32:55 - INFO - codeparrot_training - Step 30857: {'lr': 0.00017185927838938027, 'samples': 15799296, 'steps': 30857, 'loss/train': 0.4109536409378052} 02/25/2022 17:33:00 - INFO - codeparrot_training - Step 30858: {'lr': 0.00017184373590230752, 'samples': 15799808, 'steps': 30858, 'loss/train': 1.4277530908584595} 02/25/2022 17:33:04 - INFO - codeparrot_training - Step 30859: {'lr': 0.0001718281937500314, 'samples': 15800320, 'steps': 30859, 'loss/train': 1.2299360036849976} 02/25/2022 17:33:09 - INFO - codeparrot_training - Step 30860: {'lr': 0.00017181265193261865, 'samples': 15800832, 'steps': 30860, 'loss/train': 1.243561863899231} 02/25/2022 17:33:13 - INFO - codeparrot_training - Step 30861: {'lr': 0.00017179711045013546, 'samples': 15801344, 'steps': 30861, 'loss/train': 2.1546435356140137} 02/25/2022 17:33:19 - INFO - codeparrot_training - Step 30862: {'lr': 0.00017178156930264867, 'samples': 15801856, 'steps': 30862, 'loss/train': 1.813400387763977} 02/25/2022 17:33:23 - INFO - codeparrot_training - Step 30863: {'lr': 0.00017176602849022484, 'samples': 15802368, 'steps': 30863, 'loss/train': 1.7000378370285034} 02/25/2022 17:33:28 - INFO - codeparrot_training - Step 30864: {'lr': 0.00017175048801293041, 'samples': 15802880, 'steps': 30864, 'loss/train': 2.5158183574676514} 02/25/2022 17:33:32 - INFO - codeparrot_training - Step 30865: {'lr': 0.00017173494787083222, 'samples': 15803392, 'steps': 30865, 'loss/train': 0.9467108845710754} 02/25/2022 17:33:37 - INFO - codeparrot_training - Step 30866: {'lr': 0.00017171940806399652, 'samples': 15803904, 'steps': 30866, 'loss/train': 1.1335984468460083} 02/25/2022 17:33:41 - INFO - codeparrot_training - Step 30867: {'lr': 0.00017170386859249005, 'samples': 15804416, 'steps': 30867, 'loss/train': 1.988924503326416} 02/25/2022 17:33:47 - INFO - codeparrot_training - Step 30868: {'lr': 0.0001716883294563793, 'samples': 15804928, 'steps': 30868, 'loss/train': 2.3729653358459473} 02/25/2022 17:33:50 - INFO - codeparrot_training - Step 30869: {'lr': 0.00017167279065573104, 'samples': 15805440, 'steps': 30869, 'loss/train': 2.2912564277648926} 02/25/2022 17:33:56 - INFO - codeparrot_training - Step 30870: {'lr': 0.00017165725219061152, 'samples': 15805952, 'steps': 30870, 'loss/train': 1.8677479028701782} 02/25/2022 17:33:59 - INFO - codeparrot_training - Step 30871: {'lr': 0.00017164171406108749, 'samples': 15806464, 'steps': 30871, 'loss/train': 2.0796620845794678} 02/25/2022 17:34:06 - INFO - codeparrot_training - Step 30872: {'lr': 0.00017162617626722545, 'samples': 15806976, 'steps': 30872, 'loss/train': 2.152864694595337} 02/25/2022 17:34:09 - INFO - codeparrot_training - Step 30873: {'lr': 0.00017161063880909212, 'samples': 15807488, 'steps': 30873, 'loss/train': 2.406723976135254} 02/25/2022 17:34:15 - INFO - codeparrot_training - Step 30874: {'lr': 0.0001715951016867538, 'samples': 15808000, 'steps': 30874, 'loss/train': 2.6085996627807617} 02/25/2022 17:34:18 - INFO - codeparrot_training - Step 30875: {'lr': 0.00017157956490027715, 'samples': 15808512, 'steps': 30875, 'loss/train': 1.7170699834823608} 02/25/2022 17:34:24 - INFO - codeparrot_training - Step 30876: {'lr': 0.00017156402844972874, 'samples': 15809024, 'steps': 30876, 'loss/train': 0.03191710636019707} 02/25/2022 17:34:27 - INFO - codeparrot_training - Step 30877: {'lr': 0.00017154849233517517, 'samples': 15809536, 'steps': 30877, 'loss/train': 2.4711148738861084} 02/25/2022 17:34:33 - INFO - codeparrot_training - Step 30878: {'lr': 0.000171532956556683, 'samples': 15810048, 'steps': 30878, 'loss/train': 1.783501386642456} 02/25/2022 17:34:36 - INFO - codeparrot_training - Step 30879: {'lr': 0.00017151742111431867, 'samples': 15810560, 'steps': 30879, 'loss/train': 2.419020414352417} 02/25/2022 17:34:42 - INFO - codeparrot_training - Step 30880: {'lr': 0.00017150188600814875, 'samples': 15811072, 'steps': 30880, 'loss/train': 1.9733247756958008} 02/25/2022 17:34:45 - INFO - codeparrot_training - Step 30881: {'lr': 0.00017148635123823985, 'samples': 15811584, 'steps': 30881, 'loss/train': 0.3049890697002411} 02/25/2022 17:34:52 - INFO - codeparrot_training - Step 30882: {'lr': 0.0001714708168046586, 'samples': 15812096, 'steps': 30882, 'loss/train': 1.6930447816848755} 02/25/2022 17:34:55 - INFO - codeparrot_training - Step 30883: {'lr': 0.0001714552827074713, 'samples': 15812608, 'steps': 30883, 'loss/train': 2.405334711074829} 02/25/2022 17:35:01 - INFO - codeparrot_training - Step 30884: {'lr': 0.00017143974894674464, 'samples': 15813120, 'steps': 30884, 'loss/train': 2.5295002460479736} 02/25/2022 17:35:04 - INFO - codeparrot_training - Step 30885: {'lr': 0.00017142421552254517, 'samples': 15813632, 'steps': 30885, 'loss/train': 1.4440367221832275} 02/25/2022 17:35:10 - INFO - codeparrot_training - Step 30886: {'lr': 0.0001714086824349395, 'samples': 15814144, 'steps': 30886, 'loss/train': 1.6534053087234497} 02/25/2022 17:35:13 - INFO - codeparrot_training - Step 30887: {'lr': 0.00017139314968399395, 'samples': 15814656, 'steps': 30887, 'loss/train': 1.4282454252243042} 02/25/2022 17:35:19 - INFO - codeparrot_training - Step 30888: {'lr': 0.0001713776172697752, 'samples': 15815168, 'steps': 30888, 'loss/train': 0.889448344707489} 02/25/2022 17:35:22 - INFO - codeparrot_training - Step 30889: {'lr': 0.00017136208519234986, 'samples': 15815680, 'steps': 30889, 'loss/train': 1.6427979469299316} 02/25/2022 17:35:28 - INFO - codeparrot_training - Step 30890: {'lr': 0.0001713465534517843, 'samples': 15816192, 'steps': 30890, 'loss/train': 2.5924878120422363} 02/25/2022 17:35:31 - INFO - codeparrot_training - Step 30891: {'lr': 0.00017133102204814515, 'samples': 15816704, 'steps': 30891, 'loss/train': 1.9412628412246704} 02/25/2022 17:35:37 - INFO - codeparrot_training - Step 30892: {'lr': 0.0001713154909814989, 'samples': 15817216, 'steps': 30892, 'loss/train': 1.671139121055603} 02/25/2022 17:35:41 - INFO - codeparrot_training - Step 30893: {'lr': 0.00017129996025191217, 'samples': 15817728, 'steps': 30893, 'loss/train': 1.4168286323547363} 02/25/2022 17:35:46 - INFO - codeparrot_training - Step 30894: {'lr': 0.00017128442985945137, 'samples': 15818240, 'steps': 30894, 'loss/train': 2.154322862625122} 02/25/2022 17:35:50 - INFO - codeparrot_training - Step 30895: {'lr': 0.0001712688998041831, 'samples': 15818752, 'steps': 30895, 'loss/train': 1.8076807260513306} 02/25/2022 17:35:55 - INFO - codeparrot_training - Step 30896: {'lr': 0.00017125337008617387, 'samples': 15819264, 'steps': 30896, 'loss/train': 1.6764289140701294} 02/25/2022 17:35:59 - INFO - codeparrot_training - Step 30897: {'lr': 0.00017123784070549014, 'samples': 15819776, 'steps': 30897, 'loss/train': 2.3929100036621094} 02/25/2022 17:36:04 - INFO - codeparrot_training - Step 30898: {'lr': 0.00017122231166219847, 'samples': 15820288, 'steps': 30898, 'loss/train': 2.544685125350952} 02/25/2022 17:36:08 - INFO - codeparrot_training - Step 30899: {'lr': 0.0001712067829563655, 'samples': 15820800, 'steps': 30899, 'loss/train': 1.954620361328125} 02/25/2022 17:36:13 - INFO - codeparrot_training - Step 30900: {'lr': 0.00017119125458805765, 'samples': 15821312, 'steps': 30900, 'loss/train': 2.143437385559082} 02/25/2022 17:36:17 - INFO - codeparrot_training - Step 30901: {'lr': 0.00017117572655734136, 'samples': 15821824, 'steps': 30901, 'loss/train': 1.8220163583755493} 02/25/2022 17:36:22 - INFO - codeparrot_training - Step 30902: {'lr': 0.00017116019886428326, 'samples': 15822336, 'steps': 30902, 'loss/train': 1.794333815574646} 02/25/2022 17:36:28 - INFO - codeparrot_training - Step 30903: {'lr': 0.0001711446715089498, 'samples': 15822848, 'steps': 30903, 'loss/train': 1.8099684715270996} 02/25/2022 17:36:31 - INFO - codeparrot_training - Step 30904: {'lr': 0.00017112914449140764, 'samples': 15823360, 'steps': 30904, 'loss/train': 1.5666985511779785} 02/25/2022 17:36:37 - INFO - codeparrot_training - Step 30905: {'lr': 0.0001711136178117231, 'samples': 15823872, 'steps': 30905, 'loss/train': 1.0048363208770752} 02/25/2022 17:36:40 - INFO - codeparrot_training - Step 30906: {'lr': 0.00017109809146996274, 'samples': 15824384, 'steps': 30906, 'loss/train': 2.2085349559783936} 02/25/2022 17:36:46 - INFO - codeparrot_training - Step 30907: {'lr': 0.00017108256546619313, 'samples': 15824896, 'steps': 30907, 'loss/train': 1.774291753768921} 02/25/2022 17:36:50 - INFO - codeparrot_training - Step 30908: {'lr': 0.00017106703980048084, 'samples': 15825408, 'steps': 30908, 'loss/train': 1.3725571632385254} 02/25/2022 17:36:55 - INFO - codeparrot_training - Step 30909: {'lr': 0.00017105151447289218, 'samples': 15825920, 'steps': 30909, 'loss/train': 1.9665273427963257} 02/25/2022 17:36:59 - INFO - codeparrot_training - Step 30910: {'lr': 0.0001710359894834938, 'samples': 15826432, 'steps': 30910, 'loss/train': 2.044612169265747} 02/25/2022 17:37:04 - INFO - codeparrot_training - Step 30911: {'lr': 0.0001710204648323521, 'samples': 15826944, 'steps': 30911, 'loss/train': 2.160705089569092} 02/25/2022 17:37:08 - INFO - codeparrot_training - Step 30912: {'lr': 0.00017100494051953372, 'samples': 15827456, 'steps': 30912, 'loss/train': 2.6525821685791016} 02/25/2022 17:37:13 - INFO - codeparrot_training - Step 30913: {'lr': 0.00017098941654510513, 'samples': 15827968, 'steps': 30913, 'loss/train': 1.2803303003311157} 02/25/2022 17:37:17 - INFO - codeparrot_training - Step 30914: {'lr': 0.00017097389290913275, 'samples': 15828480, 'steps': 30914, 'loss/train': 1.8472870588302612} 02/25/2022 17:37:22 - INFO - codeparrot_training - Step 30915: {'lr': 0.00017095836961168308, 'samples': 15828992, 'steps': 30915, 'loss/train': 8.416533470153809} 02/25/2022 17:37:26 - INFO - codeparrot_training - Step 30916: {'lr': 0.00017094284665282266, 'samples': 15829504, 'steps': 30916, 'loss/train': 3.010324239730835} 02/25/2022 17:37:32 - INFO - codeparrot_training - Step 30917: {'lr': 0.0001709273240326181, 'samples': 15830016, 'steps': 30917, 'loss/train': 1.221379280090332} 02/25/2022 17:37:35 - INFO - codeparrot_training - Step 30918: {'lr': 0.00017091180175113566, 'samples': 15830528, 'steps': 30918, 'loss/train': 2.2346441745758057} 02/25/2022 17:37:41 - INFO - codeparrot_training - Step 30919: {'lr': 0.00017089627980844193, 'samples': 15831040, 'steps': 30919, 'loss/train': 0.8587213754653931} 02/25/2022 17:37:44 - INFO - codeparrot_training - Step 30920: {'lr': 0.00017088075820460348, 'samples': 15831552, 'steps': 30920, 'loss/train': 1.1487531661987305} 02/25/2022 17:37:50 - INFO - codeparrot_training - Step 30921: {'lr': 0.00017086523693968678, 'samples': 15832064, 'steps': 30921, 'loss/train': 1.5996801853179932} 02/25/2022 17:37:53 - INFO - codeparrot_training - Step 30922: {'lr': 0.0001708497160137582, 'samples': 15832576, 'steps': 30922, 'loss/train': 1.2667784690856934} 02/25/2022 17:37:59 - INFO - codeparrot_training - Step 30923: {'lr': 0.00017083419542688428, 'samples': 15833088, 'steps': 30923, 'loss/train': 0.20584948360919952} 02/25/2022 17:38:02 - INFO - codeparrot_training - Step 30924: {'lr': 0.00017081867517913156, 'samples': 15833600, 'steps': 30924, 'loss/train': 1.4849941730499268} 02/25/2022 17:38:08 - INFO - codeparrot_training - Step 30925: {'lr': 0.00017080315527056645, 'samples': 15834112, 'steps': 30925, 'loss/train': 1.971928596496582} 02/25/2022 17:38:11 - INFO - codeparrot_training - Step 30926: {'lr': 0.0001707876357012556, 'samples': 15834624, 'steps': 30926, 'loss/train': 0.6870985627174377} 02/25/2022 17:38:18 - INFO - codeparrot_training - Step 30927: {'lr': 0.00017077211647126527, 'samples': 15835136, 'steps': 30927, 'loss/train': 2.1315577030181885} 02/25/2022 17:38:21 - INFO - codeparrot_training - Step 30928: {'lr': 0.00017075659758066205, 'samples': 15835648, 'steps': 30928, 'loss/train': 0.5204908847808838} 02/25/2022 17:38:27 - INFO - codeparrot_training - Step 30929: {'lr': 0.00017074107902951245, 'samples': 15836160, 'steps': 30929, 'loss/train': 1.8238492012023926} 02/25/2022 17:38:30 - INFO - codeparrot_training - Step 30930: {'lr': 0.00017072556081788287, 'samples': 15836672, 'steps': 30930, 'loss/train': 2.8112919330596924} 02/25/2022 17:38:36 - INFO - codeparrot_training - Step 30931: {'lr': 0.0001707100429458398, 'samples': 15837184, 'steps': 30931, 'loss/train': 1.8597627878189087} 02/25/2022 17:38:39 - INFO - codeparrot_training - Step 30932: {'lr': 0.00017069452541344972, 'samples': 15837696, 'steps': 30932, 'loss/train': 3.509359121322632} 02/25/2022 17:38:45 - INFO - codeparrot_training - Step 30933: {'lr': 0.00017067900822077914, 'samples': 15838208, 'steps': 30933, 'loss/train': 2.451078176498413} 02/25/2022 17:38:48 - INFO - codeparrot_training - Step 30934: {'lr': 0.00017066349136789452, 'samples': 15838720, 'steps': 30934, 'loss/train': 0.6647193431854248} 02/25/2022 17:38:54 - INFO - codeparrot_training - Step 30935: {'lr': 0.00017064797485486226, 'samples': 15839232, 'steps': 30935, 'loss/train': 0.14275012910366058} 02/25/2022 17:38:57 - INFO - codeparrot_training - Step 30936: {'lr': 0.00017063245868174893, 'samples': 15839744, 'steps': 30936, 'loss/train': 1.5685423612594604} 02/25/2022 17:39:03 - INFO - codeparrot_training - Step 30937: {'lr': 0.00017061694284862088, 'samples': 15840256, 'steps': 30937, 'loss/train': 2.4612739086151123} 02/25/2022 17:39:06 - INFO - codeparrot_training - Step 30938: {'lr': 0.00017060142735554463, 'samples': 15840768, 'steps': 30938, 'loss/train': 2.703680992126465} 02/25/2022 17:39:13 - INFO - codeparrot_training - Step 30939: {'lr': 0.00017058591220258672, 'samples': 15841280, 'steps': 30939, 'loss/train': 1.4875937700271606} 02/25/2022 17:39:16 - INFO - codeparrot_training - Step 30940: {'lr': 0.00017057039738981356, 'samples': 15841792, 'steps': 30940, 'loss/train': 1.9587736129760742} 02/25/2022 17:39:22 - INFO - codeparrot_training - Step 30941: {'lr': 0.00017055488291729153, 'samples': 15842304, 'steps': 30941, 'loss/train': 2.813506841659546} 02/25/2022 17:39:25 - INFO - codeparrot_training - Step 30942: {'lr': 0.00017053936878508717, 'samples': 15842816, 'steps': 30942, 'loss/train': 0.8924906253814697} 02/25/2022 17:39:31 - INFO - codeparrot_training - Step 30943: {'lr': 0.000170523854993267, 'samples': 15843328, 'steps': 30943, 'loss/train': 2.669710636138916} 02/25/2022 17:39:34 - INFO - codeparrot_training - Step 30944: {'lr': 0.00017050834154189732, 'samples': 15843840, 'steps': 30944, 'loss/train': 2.0535473823547363} 02/25/2022 17:39:40 - INFO - codeparrot_training - Step 30945: {'lr': 0.00017049282843104464, 'samples': 15844352, 'steps': 30945, 'loss/train': 5.883098602294922} 02/25/2022 17:39:43 - INFO - codeparrot_training - Step 30946: {'lr': 0.00017047731566077545, 'samples': 15844864, 'steps': 30946, 'loss/train': 1.8980615139007568} 02/25/2022 17:39:49 - INFO - codeparrot_training - Step 30947: {'lr': 0.0001704618032311563, 'samples': 15845376, 'steps': 30947, 'loss/train': 1.850115418434143} 02/25/2022 17:39:53 - INFO - codeparrot_training - Step 30948: {'lr': 0.00017044629114225341, 'samples': 15845888, 'steps': 30948, 'loss/train': 0.18369010090827942} 02/25/2022 17:39:58 - INFO - codeparrot_training - Step 30949: {'lr': 0.00017043077939413339, 'samples': 15846400, 'steps': 30949, 'loss/train': 1.6518300771713257} 02/25/2022 17:40:02 - INFO - codeparrot_training - Step 30950: {'lr': 0.0001704152679868626, 'samples': 15846912, 'steps': 30950, 'loss/train': 1.1802380084991455} 02/25/2022 17:40:07 - INFO - codeparrot_training - Step 30951: {'lr': 0.00017039975692050753, 'samples': 15847424, 'steps': 30951, 'loss/train': 2.2024519443511963} 02/25/2022 17:40:11 - INFO - codeparrot_training - Step 30952: {'lr': 0.00017038424619513471, 'samples': 15847936, 'steps': 30952, 'loss/train': 1.9004453420639038} 02/25/2022 17:40:16 - INFO - codeparrot_training - Step 30953: {'lr': 0.00017036873581081045, 'samples': 15848448, 'steps': 30953, 'loss/train': 2.141775131225586} 02/25/2022 17:40:20 - INFO - codeparrot_training - Step 30954: {'lr': 0.00017035322576760122, 'samples': 15848960, 'steps': 30954, 'loss/train': 1.3196666240692139} 02/25/2022 17:40:26 - INFO - codeparrot_training - Step 30955: {'lr': 0.00017033771606557346, 'samples': 15849472, 'steps': 30955, 'loss/train': 0.9540612101554871} 02/25/2022 17:40:29 - INFO - codeparrot_training - Step 30956: {'lr': 0.00017032220670479376, 'samples': 15849984, 'steps': 30956, 'loss/train': 1.969914436340332} 02/25/2022 17:40:35 - INFO - codeparrot_training - Step 30957: {'lr': 0.0001703066976853283, 'samples': 15850496, 'steps': 30957, 'loss/train': 0.7234318852424622} 02/25/2022 17:40:38 - INFO - codeparrot_training - Step 30958: {'lr': 0.00017029118900724363, 'samples': 15851008, 'steps': 30958, 'loss/train': 1.6052916049957275} 02/25/2022 17:40:44 - INFO - codeparrot_training - Step 30959: {'lr': 0.00017027568067060623, 'samples': 15851520, 'steps': 30959, 'loss/train': 2.1360130310058594} 02/25/2022 17:40:47 - INFO - codeparrot_training - Step 30960: {'lr': 0.0001702601726754825, 'samples': 15852032, 'steps': 30960, 'loss/train': 3.0374183654785156} 02/25/2022 17:40:53 - INFO - codeparrot_training - Step 30961: {'lr': 0.00017024466502193892, 'samples': 15852544, 'steps': 30961, 'loss/train': 2.400423049926758} 02/25/2022 17:40:56 - INFO - codeparrot_training - Step 30962: {'lr': 0.00017022915771004182, 'samples': 15853056, 'steps': 30962, 'loss/train': 1.2397576570510864} 02/25/2022 17:41:02 - INFO - codeparrot_training - Step 30963: {'lr': 0.00017021365073985767, 'samples': 15853568, 'steps': 30963, 'loss/train': 2.007594347000122} 02/25/2022 17:41:05 - INFO - codeparrot_training - Step 30964: {'lr': 0.0001701981441114529, 'samples': 15854080, 'steps': 30964, 'loss/train': 1.2358102798461914} 02/25/2022 17:41:12 - INFO - codeparrot_training - Step 30965: {'lr': 0.000170182637824894, 'samples': 15854592, 'steps': 30965, 'loss/train': 1.9479891061782837} 02/25/2022 17:41:15 - INFO - codeparrot_training - Step 30966: {'lr': 0.0001701671318802473, 'samples': 15855104, 'steps': 30966, 'loss/train': 2.5740203857421875} 02/25/2022 17:41:21 - INFO - codeparrot_training - Step 30967: {'lr': 0.00017015162627757925, 'samples': 15855616, 'steps': 30967, 'loss/train': 1.6464446783065796} 02/25/2022 17:41:24 - INFO - codeparrot_training - Step 30968: {'lr': 0.00017013612101695623, 'samples': 15856128, 'steps': 30968, 'loss/train': 1.8661909103393555} 02/25/2022 17:41:30 - INFO - codeparrot_training - Step 30969: {'lr': 0.00017012061609844486, 'samples': 15856640, 'steps': 30969, 'loss/train': 1.9088335037231445} 02/25/2022 17:41:33 - INFO - codeparrot_training - Step 30970: {'lr': 0.00017010511152211127, 'samples': 15857152, 'steps': 30970, 'loss/train': 1.4334156513214111} 02/25/2022 17:41:39 - INFO - codeparrot_training - Step 30971: {'lr': 0.00017008960728802204, 'samples': 15857664, 'steps': 30971, 'loss/train': 2.141254425048828} 02/25/2022 17:41:42 - INFO - codeparrot_training - Step 30972: {'lr': 0.0001700741033962435, 'samples': 15858176, 'steps': 30972, 'loss/train': 1.1563220024108887} 02/25/2022 17:41:48 - INFO - codeparrot_training - Step 30973: {'lr': 0.00017005859984684223, 'samples': 15858688, 'steps': 30973, 'loss/train': 0.5578577518463135} 02/25/2022 17:41:51 - INFO - codeparrot_training - Step 30974: {'lr': 0.00017004309663988456, 'samples': 15859200, 'steps': 30974, 'loss/train': 1.5617010593414307} 02/25/2022 17:41:57 - INFO - codeparrot_training - Step 30975: {'lr': 0.00017002759377543675, 'samples': 15859712, 'steps': 30975, 'loss/train': 8.43612289428711} 02/25/2022 17:42:01 - INFO - codeparrot_training - Step 30976: {'lr': 0.00017001209125356543, 'samples': 15860224, 'steps': 30976, 'loss/train': 1.6823818683624268} 02/25/2022 17:42:06 - INFO - codeparrot_training - Step 30977: {'lr': 0.00016999658907433684, 'samples': 15860736, 'steps': 30977, 'loss/train': 1.3721647262573242} 02/25/2022 17:42:10 - INFO - codeparrot_training - Step 30978: {'lr': 0.0001699810872378175, 'samples': 15861248, 'steps': 30978, 'loss/train': 0.9308111667633057} 02/25/2022 17:42:16 - INFO - codeparrot_training - Step 30979: {'lr': 0.00016996558574407377, 'samples': 15861760, 'steps': 30979, 'loss/train': 1.0938304662704468} 02/25/2022 17:42:19 - INFO - codeparrot_training - Step 30980: {'lr': 0.00016995008459317208, 'samples': 15862272, 'steps': 30980, 'loss/train': 1.901100516319275} 02/25/2022 17:42:25 - INFO - codeparrot_training - Step 30981: {'lr': 0.00016993458378517876, 'samples': 15862784, 'steps': 30981, 'loss/train': 1.3357551097869873} 02/25/2022 17:42:28 - INFO - codeparrot_training - Step 30982: {'lr': 0.00016991908332016033, 'samples': 15863296, 'steps': 30982, 'loss/train': 1.954347014427185} 02/25/2022 17:42:34 - INFO - codeparrot_training - Step 30983: {'lr': 0.00016990358319818308, 'samples': 15863808, 'steps': 30983, 'loss/train': 1.605167031288147} 02/25/2022 17:42:37 - INFO - codeparrot_training - Step 30984: {'lr': 0.0001698880834193134, 'samples': 15864320, 'steps': 30984, 'loss/train': 0.49428653717041016} 02/25/2022 17:42:43 - INFO - codeparrot_training - Step 30985: {'lr': 0.00016987258398361776, 'samples': 15864832, 'steps': 30985, 'loss/train': 2.966676712036133} 02/25/2022 17:42:49 - INFO - codeparrot_training - Step 30986: {'lr': 0.0001698570848911625, 'samples': 15865344, 'steps': 30986, 'loss/train': 1.2860169410705566} 02/25/2022 17:42:52 - INFO - codeparrot_training - Step 30987: {'lr': 0.00016984158614201416, 'samples': 15865856, 'steps': 30987, 'loss/train': 0.14702078700065613} 02/25/2022 17:42:58 - INFO - codeparrot_training - Step 30988: {'lr': 0.0001698260877362389, 'samples': 15866368, 'steps': 30988, 'loss/train': 1.084276556968689} 02/25/2022 17:43:01 - INFO - codeparrot_training - Step 30989: {'lr': 0.00016981058967390321, 'samples': 15866880, 'steps': 30989, 'loss/train': 2.4190120697021484} 02/25/2022 17:43:07 - INFO - codeparrot_training - Step 30990: {'lr': 0.00016979509195507353, 'samples': 15867392, 'steps': 30990, 'loss/train': 0.3845204710960388} 02/25/2022 17:43:10 - INFO - codeparrot_training - Step 30991: {'lr': 0.0001697795945798163, 'samples': 15867904, 'steps': 30991, 'loss/train': 1.586069107055664} 02/25/2022 17:43:16 - INFO - codeparrot_training - Step 30992: {'lr': 0.00016976409754819767, 'samples': 15868416, 'steps': 30992, 'loss/train': 2.422858238220215} 02/25/2022 17:43:20 - INFO - codeparrot_training - Step 30993: {'lr': 0.00016974860086028422, 'samples': 15868928, 'steps': 30993, 'loss/train': 2.136409282684326} 02/25/2022 17:43:23 - INFO - codeparrot_training - Step 30994: {'lr': 0.00016973310451614225, 'samples': 15869440, 'steps': 30994, 'loss/train': 1.3699604272842407} 02/25/2022 17:43:29 - INFO - codeparrot_training - Step 30995: {'lr': 0.00016971760851583828, 'samples': 15869952, 'steps': 30995, 'loss/train': 2.20263409614563} 02/25/2022 17:43:33 - INFO - codeparrot_training - Step 30996: {'lr': 0.0001697021128594385, 'samples': 15870464, 'steps': 30996, 'loss/train': 1.7646503448486328} 02/25/2022 17:43:38 - INFO - codeparrot_training - Step 30997: {'lr': 0.00016968661754700933, 'samples': 15870976, 'steps': 30997, 'loss/train': 1.811256766319275} 02/25/2022 17:43:42 - INFO - codeparrot_training - Step 30998: {'lr': 0.00016967112257861721, 'samples': 15871488, 'steps': 30998, 'loss/train': 0.2525671422481537} 02/25/2022 17:43:47 - INFO - codeparrot_training - Step 30999: {'lr': 0.00016965562795432848, 'samples': 15872000, 'steps': 30999, 'loss/train': 2.7075023651123047} 02/25/2022 17:43:47 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 17:44:05 - WARNING - huggingface_hub.repository - Several commits (31) will be pushed upstream. 02/25/2022 17:44:05 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 17:44:38 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy af58905..59d2b90 floral-grass-11 -> floral-grass-11 02/25/2022 17:44:43 - INFO - codeparrot_training - Step 31000: {'lr': 0.00016964013367420965, 'samples': 15872512, 'steps': 31000, 'loss/train': 0.6951963305473328} 02/25/2022 17:44:49 - INFO - codeparrot_training - Step 31001: {'lr': 0.00016962463973832688, 'samples': 15873024, 'steps': 31001, 'loss/train': 1.6582273244857788} 02/25/2022 17:44:52 - INFO - codeparrot_training - Step 31002: {'lr': 0.0001696091461467466, 'samples': 15873536, 'steps': 31002, 'loss/train': 2.0780632495880127} 02/25/2022 17:44:58 - INFO - codeparrot_training - Step 31003: {'lr': 0.00016959365289953522, 'samples': 15874048, 'steps': 31003, 'loss/train': 2.472001791000366} 02/25/2022 17:45:01 - INFO - codeparrot_training - Step 31004: {'lr': 0.00016957815999675923, 'samples': 15874560, 'steps': 31004, 'loss/train': 0.43674641847610474} 02/25/2022 17:45:07 - INFO - codeparrot_training - Step 31005: {'lr': 0.00016956266743848472, 'samples': 15875072, 'steps': 31005, 'loss/train': 1.7334411144256592} 02/25/2022 17:45:10 - INFO - codeparrot_training - Step 31006: {'lr': 0.00016954717522477824, 'samples': 15875584, 'steps': 31006, 'loss/train': 2.1179020404815674} 02/25/2022 17:45:16 - INFO - codeparrot_training - Step 31007: {'lr': 0.00016953168335570606, 'samples': 15876096, 'steps': 31007, 'loss/train': 2.2971134185791016} 02/25/2022 17:45:19 - INFO - codeparrot_training - Step 31008: {'lr': 0.00016951619183133476, 'samples': 15876608, 'steps': 31008, 'loss/train': 1.6437169313430786} 02/25/2022 17:45:25 - INFO - codeparrot_training - Step 31009: {'lr': 0.0001695007006517304, 'samples': 15877120, 'steps': 31009, 'loss/train': 1.7899909019470215} 02/25/2022 17:45:28 - INFO - codeparrot_training - Step 31010: {'lr': 0.0001694852098169595, 'samples': 15877632, 'steps': 31010, 'loss/train': 2.6129343509674072} 02/25/2022 17:45:35 - INFO - codeparrot_training - Step 31011: {'lr': 0.00016946971932708837, 'samples': 15878144, 'steps': 31011, 'loss/train': 1.7559475898742676} 02/25/2022 17:45:38 - INFO - codeparrot_training - Step 31012: {'lr': 0.00016945422918218338, 'samples': 15878656, 'steps': 31012, 'loss/train': 1.1093412637710571} 02/25/2022 17:45:43 - INFO - codeparrot_training - Step 31013: {'lr': 0.00016943873938231105, 'samples': 15879168, 'steps': 31013, 'loss/train': 1.1784868240356445} 02/25/2022 17:45:47 - INFO - codeparrot_training - Step 31014: {'lr': 0.00016942324992753744, 'samples': 15879680, 'steps': 31014, 'loss/train': 1.3519665002822876} 02/25/2022 17:45:52 - INFO - codeparrot_training - Step 31015: {'lr': 0.00016940776081792908, 'samples': 15880192, 'steps': 31015, 'loss/train': 2.262308359146118} 02/25/2022 17:45:56 - INFO - codeparrot_training - Step 31016: {'lr': 0.0001693922720535523, 'samples': 15880704, 'steps': 31016, 'loss/train': 2.342503070831299} 02/25/2022 17:46:02 - INFO - codeparrot_training - Step 31017: {'lr': 0.00016937678363447342, 'samples': 15881216, 'steps': 31017, 'loss/train': 8.604876518249512} 02/25/2022 17:46:05 - INFO - codeparrot_training - Step 31018: {'lr': 0.00016936129556075876, 'samples': 15881728, 'steps': 31018, 'loss/train': 0.6884424686431885} 02/25/2022 17:46:11 - INFO - codeparrot_training - Step 31019: {'lr': 0.00016934580783247471, 'samples': 15882240, 'steps': 31019, 'loss/train': 0.8045560121536255} 02/25/2022 17:46:14 - INFO - codeparrot_training - Step 31020: {'lr': 0.00016933032044968765, 'samples': 15882752, 'steps': 31020, 'loss/train': 1.5906058549880981} 02/25/2022 17:46:20 - INFO - codeparrot_training - Step 31021: {'lr': 0.00016931483341246385, 'samples': 15883264, 'steps': 31021, 'loss/train': 2.1294853687286377} 02/25/2022 17:46:24 - INFO - codeparrot_training - Step 31022: {'lr': 0.00016929934672086971, 'samples': 15883776, 'steps': 31022, 'loss/train': 0.7153582572937012} 02/25/2022 17:46:29 - INFO - codeparrot_training - Step 31023: {'lr': 0.00016928386037497157, 'samples': 15884288, 'steps': 31023, 'loss/train': 2.059047222137451} 02/25/2022 17:46:33 - INFO - codeparrot_training - Step 31024: {'lr': 0.00016926837437483566, 'samples': 15884800, 'steps': 31024, 'loss/train': 1.5293407440185547} 02/25/2022 17:46:38 - INFO - codeparrot_training - Step 31025: {'lr': 0.0001692528887205284, 'samples': 15885312, 'steps': 31025, 'loss/train': 1.914209246635437} 02/25/2022 17:46:42 - INFO - codeparrot_training - Step 31026: {'lr': 0.0001692374034121162, 'samples': 15885824, 'steps': 31026, 'loss/train': 2.1056015491485596} 02/25/2022 17:46:48 - INFO - codeparrot_training - Step 31027: {'lr': 0.0001692219184496653, 'samples': 15886336, 'steps': 31027, 'loss/train': 1.5960326194763184} 02/25/2022 17:46:51 - INFO - codeparrot_training - Step 31028: {'lr': 0.00016920643383324201, 'samples': 15886848, 'steps': 31028, 'loss/train': 0.8852987289428711} 02/25/2022 17:46:56 - INFO - codeparrot_training - Step 31029: {'lr': 0.0001691909495629127, 'samples': 15887360, 'steps': 31029, 'loss/train': 2.0812950134277344} 02/25/2022 17:47:00 - INFO - codeparrot_training - Step 31030: {'lr': 0.00016917546563874374, 'samples': 15887872, 'steps': 31030, 'loss/train': 2.6640775203704834} 02/25/2022 17:47:06 - INFO - codeparrot_training - Step 31031: {'lr': 0.00016915998206080147, 'samples': 15888384, 'steps': 31031, 'loss/train': 1.7001160383224487} 02/25/2022 17:47:10 - INFO - codeparrot_training - Step 31032: {'lr': 0.00016914449882915206, 'samples': 15888896, 'steps': 31032, 'loss/train': 2.2505905628204346} 02/25/2022 17:47:15 - INFO - codeparrot_training - Step 31033: {'lr': 0.00016912901594386197, 'samples': 15889408, 'steps': 31033, 'loss/train': 1.3172106742858887} 02/25/2022 17:47:19 - INFO - codeparrot_training - Step 31034: {'lr': 0.00016911353340499748, 'samples': 15889920, 'steps': 31034, 'loss/train': 1.4276812076568604} 02/25/2022 17:47:24 - INFO - codeparrot_training - Step 31035: {'lr': 0.00016909805121262506, 'samples': 15890432, 'steps': 31035, 'loss/train': 0.8314815163612366} 02/25/2022 17:47:28 - INFO - codeparrot_training - Step 31036: {'lr': 0.00016908256936681076, 'samples': 15890944, 'steps': 31036, 'loss/train': 2.0956528186798096} 02/25/2022 17:47:33 - INFO - codeparrot_training - Step 31037: {'lr': 0.00016906708786762104, 'samples': 15891456, 'steps': 31037, 'loss/train': 1.2844158411026} 02/25/2022 17:47:37 - INFO - codeparrot_training - Step 31038: {'lr': 0.00016905160671512227, 'samples': 15891968, 'steps': 31038, 'loss/train': 1.5763933658599854} 02/25/2022 17:47:42 - INFO - codeparrot_training - Step 31039: {'lr': 0.00016903612590938075, 'samples': 15892480, 'steps': 31039, 'loss/train': 1.890817403793335} 02/25/2022 17:47:46 - INFO - codeparrot_training - Step 31040: {'lr': 0.0001690206454504627, 'samples': 15892992, 'steps': 31040, 'loss/train': 2.108973264694214} 02/25/2022 17:47:51 - INFO - codeparrot_training - Step 31041: {'lr': 0.00016900516533843447, 'samples': 15893504, 'steps': 31041, 'loss/train': 1.5586225986480713} 02/25/2022 17:47:55 - INFO - codeparrot_training - Step 31042: {'lr': 0.00016898968557336238, 'samples': 15894016, 'steps': 31042, 'loss/train': 2.1700501441955566} 02/25/2022 17:48:01 - INFO - codeparrot_training - Step 31043: {'lr': 0.00016897420615531289, 'samples': 15894528, 'steps': 31043, 'loss/train': 2.3337037563323975} 02/25/2022 17:48:04 - INFO - codeparrot_training - Step 31044: {'lr': 0.00016895872708435207, 'samples': 15895040, 'steps': 31044, 'loss/train': 2.0221059322357178} 02/25/2022 17:48:10 - INFO - codeparrot_training - Step 31045: {'lr': 0.00016894324836054628, 'samples': 15895552, 'steps': 31045, 'loss/train': 1.4884897470474243} 02/25/2022 17:48:13 - INFO - codeparrot_training - Step 31046: {'lr': 0.00016892776998396193, 'samples': 15896064, 'steps': 31046, 'loss/train': 1.1874433755874634} 02/25/2022 17:48:19 - INFO - codeparrot_training - Step 31047: {'lr': 0.00016891229195466524, 'samples': 15896576, 'steps': 31047, 'loss/train': 1.7352169752120972} 02/25/2022 17:48:24 - INFO - codeparrot_training - Step 31048: {'lr': 0.00016889681427272263, 'samples': 15897088, 'steps': 31048, 'loss/train': 1.9792522192001343} 02/25/2022 17:48:28 - INFO - codeparrot_training - Step 31049: {'lr': 0.00016888133693820023, 'samples': 15897600, 'steps': 31049, 'loss/train': 1.6084520816802979} 02/25/2022 17:48:33 - INFO - codeparrot_training - Step 31050: {'lr': 0.00016886585995116443, 'samples': 15898112, 'steps': 31050, 'loss/train': 1.772125482559204} 02/25/2022 17:48:37 - INFO - codeparrot_training - Step 31051: {'lr': 0.00016885038331168153, 'samples': 15898624, 'steps': 31051, 'loss/train': 2.145721912384033} 02/25/2022 17:48:43 - INFO - codeparrot_training - Step 31052: {'lr': 0.0001688349070198179, 'samples': 15899136, 'steps': 31052, 'loss/train': 2.179814338684082} 02/25/2022 17:48:46 - INFO - codeparrot_training - Step 31053: {'lr': 0.00016881943107563969, 'samples': 15899648, 'steps': 31053, 'loss/train': 2.506361961364746} 02/25/2022 17:48:49 - INFO - codeparrot_training - Step 31054: {'lr': 0.00016880395547921322, 'samples': 15900160, 'steps': 31054, 'loss/train': 1.9710701704025269} 02/25/2022 17:48:55 - INFO - codeparrot_training - Step 31055: {'lr': 0.00016878848023060486, 'samples': 15900672, 'steps': 31055, 'loss/train': 1.7726755142211914} 02/25/2022 17:48:58 - INFO - codeparrot_training - Step 31056: {'lr': 0.00016877300532988092, 'samples': 15901184, 'steps': 31056, 'loss/train': 1.1743453741073608} 02/25/2022 17:49:05 - INFO - codeparrot_training - Step 31057: {'lr': 0.0001687575307771076, 'samples': 15901696, 'steps': 31057, 'loss/train': 1.358697533607483} 02/25/2022 17:49:11 - INFO - codeparrot_training - Step 31058: {'lr': 0.00016874205657235115, 'samples': 15902208, 'steps': 31058, 'loss/train': 2.1262764930725098} 02/25/2022 17:49:14 - INFO - codeparrot_training - Step 31059: {'lr': 0.00016872658271567794, 'samples': 15902720, 'steps': 31059, 'loss/train': 1.8100907802581787} 02/25/2022 17:49:20 - INFO - codeparrot_training - Step 31060: {'lr': 0.00016871110920715432, 'samples': 15903232, 'steps': 31060, 'loss/train': 1.5029383897781372} 02/25/2022 17:49:23 - INFO - codeparrot_training - Step 31061: {'lr': 0.0001686956360468465, 'samples': 15903744, 'steps': 31061, 'loss/train': 1.6959291696548462} 02/25/2022 17:49:29 - INFO - codeparrot_training - Step 31062: {'lr': 0.00016868016323482067, 'samples': 15904256, 'steps': 31062, 'loss/train': 2.564382791519165} 02/25/2022 17:49:32 - INFO - codeparrot_training - Step 31063: {'lr': 0.0001686646907711432, 'samples': 15904768, 'steps': 31063, 'loss/train': 1.7366644144058228} 02/25/2022 17:49:38 - INFO - codeparrot_training - Step 31064: {'lr': 0.00016864921865588045, 'samples': 15905280, 'steps': 31064, 'loss/train': 2.0076510906219482} 02/25/2022 17:49:41 - INFO - codeparrot_training - Step 31065: {'lr': 0.0001686337468890986, 'samples': 15905792, 'steps': 31065, 'loss/train': 2.0645558834075928} 02/25/2022 17:49:47 - INFO - codeparrot_training - Step 31066: {'lr': 0.00016861827547086388, 'samples': 15906304, 'steps': 31066, 'loss/train': 2.4808075428009033} 02/25/2022 17:49:50 - INFO - codeparrot_training - Step 31067: {'lr': 0.0001686028044012427, 'samples': 15906816, 'steps': 31067, 'loss/train': 1.8953602313995361} 02/25/2022 17:49:56 - INFO - codeparrot_training - Step 31068: {'lr': 0.0001685873336803012, 'samples': 15907328, 'steps': 31068, 'loss/train': 1.5350229740142822} 02/25/2022 17:50:00 - INFO - codeparrot_training - Step 31069: {'lr': 0.00016857186330810573, 'samples': 15907840, 'steps': 31069, 'loss/train': 1.3350399732589722} 02/25/2022 17:50:05 - INFO - codeparrot_training - Step 31070: {'lr': 0.00016855639328472257, 'samples': 15908352, 'steps': 31070, 'loss/train': 1.7629777193069458} 02/25/2022 17:50:09 - INFO - codeparrot_training - Step 31071: {'lr': 0.00016854092361021797, 'samples': 15908864, 'steps': 31071, 'loss/train': 0.61308354139328} 02/25/2022 17:50:14 - INFO - codeparrot_training - Step 31072: {'lr': 0.0001685254542846581, 'samples': 15909376, 'steps': 31072, 'loss/train': 1.1078263521194458} 02/25/2022 17:50:18 - INFO - codeparrot_training - Step 31073: {'lr': 0.00016850998530810934, 'samples': 15909888, 'steps': 31073, 'loss/train': 1.7525444030761719} 02/25/2022 17:50:24 - INFO - codeparrot_training - Step 31074: {'lr': 0.00016849451668063807, 'samples': 15910400, 'steps': 31074, 'loss/train': 2.3823907375335693} 02/25/2022 17:50:27 - INFO - codeparrot_training - Step 31075: {'lr': 0.00016847904840231026, 'samples': 15910912, 'steps': 31075, 'loss/train': 3.845890522003174} 02/25/2022 17:50:33 - INFO - codeparrot_training - Step 31076: {'lr': 0.00016846358047319232, 'samples': 15911424, 'steps': 31076, 'loss/train': 1.3276636600494385} 02/25/2022 17:50:37 - INFO - codeparrot_training - Step 31077: {'lr': 0.0001684481128933505, 'samples': 15911936, 'steps': 31077, 'loss/train': 1.5635157823562622} 02/25/2022 17:50:43 - INFO - codeparrot_training - Step 31078: {'lr': 0.00016843264566285122, 'samples': 15912448, 'steps': 31078, 'loss/train': 0.15091058611869812} 02/25/2022 17:50:46 - INFO - codeparrot_training - Step 31079: {'lr': 0.00016841717878176044, 'samples': 15912960, 'steps': 31079, 'loss/train': 2.4018688201904297} 02/25/2022 17:50:51 - INFO - codeparrot_training - Step 31080: {'lr': 0.0001684017122501446, 'samples': 15913472, 'steps': 31080, 'loss/train': 0.5490176677703857} 02/25/2022 17:50:55 - INFO - codeparrot_training - Step 31081: {'lr': 0.0001683862460680699, 'samples': 15913984, 'steps': 31081, 'loss/train': 2.7552993297576904} 02/25/2022 17:51:01 - INFO - codeparrot_training - Step 31082: {'lr': 0.00016837078023560257, 'samples': 15914496, 'steps': 31082, 'loss/train': 1.705405592918396} 02/25/2022 17:51:04 - INFO - codeparrot_training - Step 31083: {'lr': 0.00016835531475280902, 'samples': 15915008, 'steps': 31083, 'loss/train': 2.1103529930114746} 02/25/2022 17:51:09 - INFO - codeparrot_training - Step 31084: {'lr': 0.0001683398496197553, 'samples': 15915520, 'steps': 31084, 'loss/train': 2.0248754024505615} 02/25/2022 17:51:13 - INFO - codeparrot_training - Step 31085: {'lr': 0.00016832438483650768, 'samples': 15916032, 'steps': 31085, 'loss/train': 1.7194762229919434} 02/25/2022 17:51:19 - INFO - codeparrot_training - Step 31086: {'lr': 0.0001683089204031325, 'samples': 15916544, 'steps': 31086, 'loss/train': 1.9844094514846802} 02/25/2022 17:51:22 - INFO - codeparrot_training - Step 31087: {'lr': 0.00016829345631969606, 'samples': 15917056, 'steps': 31087, 'loss/train': 1.8127073049545288} 02/25/2022 17:51:28 - INFO - codeparrot_training - Step 31088: {'lr': 0.00016827799258626442, 'samples': 15917568, 'steps': 31088, 'loss/train': 1.471313238143921} 02/25/2022 17:51:31 - INFO - codeparrot_training - Step 31089: {'lr': 0.0001682625292029039, 'samples': 15918080, 'steps': 31089, 'loss/train': 1.976351022720337} 02/25/2022 17:51:37 - INFO - codeparrot_training - Step 31090: {'lr': 0.00016824706616968074, 'samples': 15918592, 'steps': 31090, 'loss/train': 1.6865428686141968} 02/25/2022 17:51:40 - INFO - codeparrot_training - Step 31091: {'lr': 0.0001682316034866613, 'samples': 15919104, 'steps': 31091, 'loss/train': 2.200965642929077} 02/25/2022 17:51:46 - INFO - codeparrot_training - Step 31092: {'lr': 0.0001682161411539116, 'samples': 15919616, 'steps': 31092, 'loss/train': 0.759330153465271} 02/25/2022 17:51:50 - INFO - codeparrot_training - Step 31093: {'lr': 0.000168200679171498, 'samples': 15920128, 'steps': 31093, 'loss/train': 1.1216405630111694} 02/25/2022 17:51:56 - INFO - codeparrot_training - Step 31094: {'lr': 0.00016818521753948667, 'samples': 15920640, 'steps': 31094, 'loss/train': 1.7577195167541504} 02/25/2022 17:51:59 - INFO - codeparrot_training - Step 31095: {'lr': 0.00016816975625794396, 'samples': 15921152, 'steps': 31095, 'loss/train': 1.0197408199310303} 02/25/2022 17:52:05 - INFO - codeparrot_training - Step 31096: {'lr': 0.00016815429532693609, 'samples': 15921664, 'steps': 31096, 'loss/train': 1.8471494913101196} 02/25/2022 17:52:08 - INFO - codeparrot_training - Step 31097: {'lr': 0.0001681388347465291, 'samples': 15922176, 'steps': 31097, 'loss/train': 1.7075550556182861} 02/25/2022 17:52:14 - INFO - codeparrot_training - Step 31098: {'lr': 0.00016812337451678938, 'samples': 15922688, 'steps': 31098, 'loss/train': 1.3168261051177979} 02/25/2022 17:52:17 - INFO - codeparrot_training - Step 31099: {'lr': 0.00016810791463778312, 'samples': 15923200, 'steps': 31099, 'loss/train': 1.2030969858169556} 02/25/2022 17:52:23 - INFO - codeparrot_training - Step 31100: {'lr': 0.00016809245510957666, 'samples': 15923712, 'steps': 31100, 'loss/train': 1.1703046560287476} 02/25/2022 17:52:26 - INFO - codeparrot_training - Step 31101: {'lr': 0.000168076995932236, 'samples': 15924224, 'steps': 31101, 'loss/train': 2.145098924636841} 02/25/2022 17:52:32 - INFO - codeparrot_training - Step 31102: {'lr': 0.0001680615371058275, 'samples': 15924736, 'steps': 31102, 'loss/train': 0.5684425830841064} 02/25/2022 17:52:35 - INFO - codeparrot_training - Step 31103: {'lr': 0.00016804607863041732, 'samples': 15925248, 'steps': 31103, 'loss/train': 1.3232879638671875} 02/25/2022 17:52:41 - INFO - codeparrot_training - Step 31104: {'lr': 0.00016803062050607187, 'samples': 15925760, 'steps': 31104, 'loss/train': 1.423197627067566} 02/25/2022 17:52:45 - INFO - codeparrot_training - Step 31105: {'lr': 0.00016801516273285704, 'samples': 15926272, 'steps': 31105, 'loss/train': 1.9002509117126465} 02/25/2022 17:52:50 - INFO - codeparrot_training - Step 31106: {'lr': 0.00016799970531083928, 'samples': 15926784, 'steps': 31106, 'loss/train': 1.266247272491455} 02/25/2022 17:52:54 - INFO - codeparrot_training - Step 31107: {'lr': 0.0001679842482400848, 'samples': 15927296, 'steps': 31107, 'loss/train': 2.3030006885528564} 02/25/2022 17:52:59 - INFO - codeparrot_training - Step 31108: {'lr': 0.00016796879152065965, 'samples': 15927808, 'steps': 31108, 'loss/train': 2.0188961029052734} 02/25/2022 17:53:03 - INFO - codeparrot_training - Step 31109: {'lr': 0.0001679533351526302, 'samples': 15928320, 'steps': 31109, 'loss/train': 1.8462263345718384} 02/25/2022 17:53:08 - INFO - codeparrot_training - Step 31110: {'lr': 0.00016793787913606262, 'samples': 15928832, 'steps': 31110, 'loss/train': 2.0196850299835205} 02/25/2022 17:53:12 - INFO - codeparrot_training - Step 31111: {'lr': 0.00016792242347102312, 'samples': 15929344, 'steps': 31111, 'loss/train': 1.6382567882537842} 02/25/2022 17:53:17 - INFO - codeparrot_training - Step 31112: {'lr': 0.00016790696815757787, 'samples': 15929856, 'steps': 31112, 'loss/train': 1.5102472305297852} 02/25/2022 17:53:21 - INFO - codeparrot_training - Step 31113: {'lr': 0.00016789151319579314, 'samples': 15930368, 'steps': 31113, 'loss/train': 0.6794791221618652} 02/25/2022 17:53:27 - INFO - codeparrot_training - Step 31114: {'lr': 0.00016787605858573508, 'samples': 15930880, 'steps': 31114, 'loss/train': 1.8656983375549316} 02/25/2022 17:53:30 - INFO - codeparrot_training - Step 31115: {'lr': 0.00016786060432746986, 'samples': 15931392, 'steps': 31115, 'loss/train': 1.166294813156128} 02/25/2022 17:53:36 - INFO - codeparrot_training - Step 31116: {'lr': 0.00016784515042106375, 'samples': 15931904, 'steps': 31116, 'loss/train': 1.1743524074554443} 02/25/2022 17:53:39 - INFO - codeparrot_training - Step 31117: {'lr': 0.0001678296968665829, 'samples': 15932416, 'steps': 31117, 'loss/train': 1.6533632278442383} 02/25/2022 17:53:45 - INFO - codeparrot_training - Step 31118: {'lr': 0.00016781424366409364, 'samples': 15932928, 'steps': 31118, 'loss/train': 2.6918015480041504} 02/25/2022 17:53:48 - INFO - codeparrot_training - Step 31119: {'lr': 0.00016779879081366201, 'samples': 15933440, 'steps': 31119, 'loss/train': 0.9125258922576904} 02/25/2022 17:53:54 - INFO - codeparrot_training - Step 31120: {'lr': 0.00016778333831535418, 'samples': 15933952, 'steps': 31120, 'loss/train': 1.5292145013809204} 02/25/2022 17:53:57 - INFO - codeparrot_training - Step 31121: {'lr': 0.0001677678861692365, 'samples': 15934464, 'steps': 31121, 'loss/train': 3.9622302055358887} 02/25/2022 17:54:03 - INFO - codeparrot_training - Step 31122: {'lr': 0.00016775243437537512, 'samples': 15934976, 'steps': 31122, 'loss/train': 2.188889980316162} 02/25/2022 17:54:06 - INFO - codeparrot_training - Step 31123: {'lr': 0.00016773698293383617, 'samples': 15935488, 'steps': 31123, 'loss/train': 0.6540660262107849} 02/25/2022 17:54:13 - INFO - codeparrot_training - Step 31124: {'lr': 0.0001677215318446858, 'samples': 15936000, 'steps': 31124, 'loss/train': 1.6189193725585938} 02/25/2022 17:54:16 - INFO - codeparrot_training - Step 31125: {'lr': 0.00016770608110799028, 'samples': 15936512, 'steps': 31125, 'loss/train': 1.781721830368042} 02/25/2022 17:54:22 - INFO - codeparrot_training - Step 31126: {'lr': 0.00016769063072381586, 'samples': 15937024, 'steps': 31126, 'loss/train': 2.638826608657837} 02/25/2022 17:54:25 - INFO - codeparrot_training - Step 31127: {'lr': 0.00016767518069222856, 'samples': 15937536, 'steps': 31127, 'loss/train': 1.7111948728561401} 02/25/2022 17:54:31 - INFO - codeparrot_training - Step 31128: {'lr': 0.00016765973101329462, 'samples': 15938048, 'steps': 31128, 'loss/train': 1.5463006496429443} 02/25/2022 17:54:34 - INFO - codeparrot_training - Step 31129: {'lr': 0.0001676442816870803, 'samples': 15938560, 'steps': 31129, 'loss/train': 2.2380406856536865} 02/25/2022 17:54:40 - INFO - codeparrot_training - Step 31130: {'lr': 0.00016762883271365168, 'samples': 15939072, 'steps': 31130, 'loss/train': 1.604448676109314} 02/25/2022 17:54:43 - INFO - codeparrot_training - Step 31131: {'lr': 0.00016761338409307508, 'samples': 15939584, 'steps': 31131, 'loss/train': 1.9851630926132202} 02/25/2022 17:54:49 - INFO - codeparrot_training - Step 31132: {'lr': 0.00016759793582541652, 'samples': 15940096, 'steps': 31132, 'loss/train': 1.4262844324111938} 02/25/2022 17:54:52 - INFO - codeparrot_training - Step 31133: {'lr': 0.0001675824879107422, 'samples': 15940608, 'steps': 31133, 'loss/train': 1.7848807573318481} 02/25/2022 17:54:59 - INFO - codeparrot_training - Step 31134: {'lr': 0.00016756704034911834, 'samples': 15941120, 'steps': 31134, 'loss/train': 0.8311417698860168} 02/25/2022 17:55:02 - INFO - codeparrot_training - Step 31135: {'lr': 0.00016755159314061125, 'samples': 15941632, 'steps': 31135, 'loss/train': 2.46219801902771} 02/25/2022 17:55:08 - INFO - codeparrot_training - Step 31136: {'lr': 0.00016753614628528678, 'samples': 15942144, 'steps': 31136, 'loss/train': 1.780564308166504} 02/25/2022 17:55:11 - INFO - codeparrot_training - Step 31137: {'lr': 0.00016752069978321133, 'samples': 15942656, 'steps': 31137, 'loss/train': 1.6061509847640991} 02/25/2022 17:55:17 - INFO - codeparrot_training - Step 31138: {'lr': 0.000167505253634451, 'samples': 15943168, 'steps': 31138, 'loss/train': 1.6682584285736084} 02/25/2022 17:55:20 - INFO - codeparrot_training - Step 31139: {'lr': 0.00016748980783907208, 'samples': 15943680, 'steps': 31139, 'loss/train': 0.3699662387371063} 02/25/2022 17:55:25 - INFO - codeparrot_training - Step 31140: {'lr': 0.00016747436239714048, 'samples': 15944192, 'steps': 31140, 'loss/train': 1.5237154960632324} 02/25/2022 17:55:29 - INFO - codeparrot_training - Step 31141: {'lr': 0.00016745891730872255, 'samples': 15944704, 'steps': 31141, 'loss/train': 2.326125383377075} 02/25/2022 17:55:35 - INFO - codeparrot_training - Step 31142: {'lr': 0.00016744347257388438, 'samples': 15945216, 'steps': 31142, 'loss/train': 1.4061886072158813} 02/25/2022 17:55:38 - INFO - codeparrot_training - Step 31143: {'lr': 0.00016742802819269217, 'samples': 15945728, 'steps': 31143, 'loss/train': 1.7516627311706543} 02/25/2022 17:55:44 - INFO - codeparrot_training - Step 31144: {'lr': 0.00016741258416521216, 'samples': 15946240, 'steps': 31144, 'loss/train': 1.2640248537063599} 02/25/2022 17:55:49 - INFO - codeparrot_training - Step 31145: {'lr': 0.00016739714049151034, 'samples': 15946752, 'steps': 31145, 'loss/train': 1.9980137348175049} 02/25/2022 17:55:53 - INFO - codeparrot_training - Step 31146: {'lr': 0.0001673816971716529, 'samples': 15947264, 'steps': 31146, 'loss/train': 1.8235807418823242} 02/25/2022 17:55:59 - INFO - codeparrot_training - Step 31147: {'lr': 0.0001673662542057061, 'samples': 15947776, 'steps': 31147, 'loss/train': 1.3574975728988647} 02/25/2022 17:56:03 - INFO - codeparrot_training - Step 31148: {'lr': 0.00016735081159373604, 'samples': 15948288, 'steps': 31148, 'loss/train': 2.2431135177612305} 02/25/2022 17:56:08 - INFO - codeparrot_training - Step 31149: {'lr': 0.00016733536933580885, 'samples': 15948800, 'steps': 31149, 'loss/train': 1.345247745513916} 02/25/2022 17:56:12 - INFO - codeparrot_training - Step 31150: {'lr': 0.0001673199274319906, 'samples': 15949312, 'steps': 31150, 'loss/train': 0.8538775444030762} 02/25/2022 17:56:17 - INFO - codeparrot_training - Step 31151: {'lr': 0.00016730448588234766, 'samples': 15949824, 'steps': 31151, 'loss/train': 1.5928668975830078} 02/25/2022 17:56:21 - INFO - codeparrot_training - Step 31152: {'lr': 0.000167289044686946, 'samples': 15950336, 'steps': 31152, 'loss/train': 1.114976406097412} 02/25/2022 17:56:26 - INFO - codeparrot_training - Step 31153: {'lr': 0.00016727360384585178, 'samples': 15950848, 'steps': 31153, 'loss/train': 2.363546133041382} 02/25/2022 17:56:30 - INFO - codeparrot_training - Step 31154: {'lr': 0.0001672581633591312, 'samples': 15951360, 'steps': 31154, 'loss/train': 1.1886793375015259} 02/25/2022 17:56:36 - INFO - codeparrot_training - Step 31155: {'lr': 0.0001672427232268503, 'samples': 15951872, 'steps': 31155, 'loss/train': 2.035230875015259} 02/25/2022 17:56:39 - INFO - codeparrot_training - Step 31156: {'lr': 0.00016722728344907534, 'samples': 15952384, 'steps': 31156, 'loss/train': 1.4630767107009888} 02/25/2022 17:56:45 - INFO - codeparrot_training - Step 31157: {'lr': 0.00016721184402587245, 'samples': 15952896, 'steps': 31157, 'loss/train': 1.1839317083358765} 02/25/2022 17:56:49 - INFO - codeparrot_training - Step 31158: {'lr': 0.00016719640495730772, 'samples': 15953408, 'steps': 31158, 'loss/train': 1.7174227237701416} 02/25/2022 17:56:54 - INFO - codeparrot_training - Step 31159: {'lr': 0.00016718096624344724, 'samples': 15953920, 'steps': 31159, 'loss/train': 1.1421518325805664} 02/25/2022 17:56:58 - INFO - codeparrot_training - Step 31160: {'lr': 0.00016716552788435723, 'samples': 15954432, 'steps': 31160, 'loss/train': 1.5369526147842407} 02/25/2022 17:57:03 - INFO - codeparrot_training - Step 31161: {'lr': 0.0001671500898801039, 'samples': 15954944, 'steps': 31161, 'loss/train': 1.771902084350586} 02/25/2022 17:57:07 - INFO - codeparrot_training - Step 31162: {'lr': 0.0001671346522307532, 'samples': 15955456, 'steps': 31162, 'loss/train': 2.038527011871338} 02/25/2022 17:57:12 - INFO - codeparrot_training - Step 31163: {'lr': 0.00016711921493637127, 'samples': 15955968, 'steps': 31163, 'loss/train': 1.1051788330078125} 02/25/2022 17:57:16 - INFO - codeparrot_training - Step 31164: {'lr': 0.00016710377799702438, 'samples': 15956480, 'steps': 31164, 'loss/train': 1.2831100225448608} 02/25/2022 17:57:21 - INFO - codeparrot_training - Step 31165: {'lr': 0.00016708834141277863, 'samples': 15956992, 'steps': 31165, 'loss/train': 1.8582972288131714} 02/25/2022 17:57:25 - INFO - codeparrot_training - Step 31166: {'lr': 0.00016707290518369998, 'samples': 15957504, 'steps': 31166, 'loss/train': 1.6135523319244385} 02/25/2022 17:57:31 - INFO - codeparrot_training - Step 31167: {'lr': 0.00016705746930985473, 'samples': 15958016, 'steps': 31167, 'loss/train': 1.0429036617279053} 02/25/2022 17:57:34 - INFO - codeparrot_training - Step 31168: {'lr': 0.0001670420337913089, 'samples': 15958528, 'steps': 31168, 'loss/train': 1.5927619934082031} 02/25/2022 17:57:40 - INFO - codeparrot_training - Step 31169: {'lr': 0.00016702659862812867, 'samples': 15959040, 'steps': 31169, 'loss/train': 1.475658893585205} 02/25/2022 17:57:43 - INFO - codeparrot_training - Step 31170: {'lr': 0.00016701116382038023, 'samples': 15959552, 'steps': 31170, 'loss/train': 1.6317473649978638} 02/25/2022 17:57:49 - INFO - codeparrot_training - Step 31171: {'lr': 0.00016699572936812953, 'samples': 15960064, 'steps': 31171, 'loss/train': 1.924875020980835} 02/25/2022 17:57:52 - INFO - codeparrot_training - Step 31172: {'lr': 0.00016698029527144277, 'samples': 15960576, 'steps': 31172, 'loss/train': 1.1838500499725342} 02/25/2022 17:57:58 - INFO - codeparrot_training - Step 31173: {'lr': 0.00016696486153038606, 'samples': 15961088, 'steps': 31173, 'loss/train': 1.4967623949050903} 02/25/2022 17:58:01 - INFO - codeparrot_training - Step 31174: {'lr': 0.00016694942814502562, 'samples': 15961600, 'steps': 31174, 'loss/train': 1.145328164100647} 02/25/2022 17:58:07 - INFO - codeparrot_training - Step 31175: {'lr': 0.00016693399511542734, 'samples': 15962112, 'steps': 31175, 'loss/train': 2.7329673767089844} 02/25/2022 17:58:10 - INFO - codeparrot_training - Step 31176: {'lr': 0.0001669185624416575, 'samples': 15962624, 'steps': 31176, 'loss/train': 1.8396117687225342} 02/25/2022 17:58:16 - INFO - codeparrot_training - Step 31177: {'lr': 0.00016690313012378212, 'samples': 15963136, 'steps': 31177, 'loss/train': 1.646539568901062} 02/25/2022 17:58:20 - INFO - codeparrot_training - Step 31178: {'lr': 0.00016688769816186732, 'samples': 15963648, 'steps': 31178, 'loss/train': 1.458148717880249} 02/25/2022 17:58:25 - INFO - codeparrot_training - Step 31179: {'lr': 0.00016687226655597937, 'samples': 15964160, 'steps': 31179, 'loss/train': 2.8021912574768066} 02/25/2022 17:58:29 - INFO - codeparrot_training - Step 31180: {'lr': 0.00016685683530618411, 'samples': 15964672, 'steps': 31180, 'loss/train': 0.9809687733650208} 02/25/2022 17:58:34 - INFO - codeparrot_training - Step 31181: {'lr': 0.00016684140441254776, 'samples': 15965184, 'steps': 31181, 'loss/train': 0.04575040191411972} 02/25/2022 17:58:38 - INFO - codeparrot_training - Step 31182: {'lr': 0.00016682597387513643, 'samples': 15965696, 'steps': 31182, 'loss/train': 1.824276089668274} 02/25/2022 17:58:43 - INFO - codeparrot_training - Step 31183: {'lr': 0.00016681054369401636, 'samples': 15966208, 'steps': 31183, 'loss/train': 2.458483934402466} 02/25/2022 17:58:47 - INFO - codeparrot_training - Step 31184: {'lr': 0.00016679511386925337, 'samples': 15966720, 'steps': 31184, 'loss/train': 2.017644166946411} 02/25/2022 17:58:52 - INFO - codeparrot_training - Step 31185: {'lr': 0.00016677968440091366, 'samples': 15967232, 'steps': 31185, 'loss/train': 1.9451687335968018} 02/25/2022 17:58:56 - INFO - codeparrot_training - Step 31186: {'lr': 0.00016676425528906343, 'samples': 15967744, 'steps': 31186, 'loss/train': 1.6944403648376465} 02/25/2022 17:59:01 - INFO - codeparrot_training - Step 31187: {'lr': 0.00016674882653376877, 'samples': 15968256, 'steps': 31187, 'loss/train': 1.6143730878829956} 02/25/2022 17:59:05 - INFO - codeparrot_training - Step 31188: {'lr': 0.00016673339813509557, 'samples': 15968768, 'steps': 31188, 'loss/train': 2.245373010635376} 02/25/2022 17:59:11 - INFO - codeparrot_training - Step 31189: {'lr': 0.00016671797009311007, 'samples': 15969280, 'steps': 31189, 'loss/train': 1.9551475048065186} 02/25/2022 17:59:17 - INFO - codeparrot_training - Step 31190: {'lr': 0.00016670254240787834, 'samples': 15969792, 'steps': 31190, 'loss/train': 2.5374176502227783} 02/25/2022 17:59:20 - INFO - codeparrot_training - Step 31191: {'lr': 0.00016668711507946654, 'samples': 15970304, 'steps': 31191, 'loss/train': 1.3559492826461792} 02/25/2022 17:59:25 - INFO - codeparrot_training - Step 31192: {'lr': 0.0001666716881079407, 'samples': 15970816, 'steps': 31192, 'loss/train': 1.6970767974853516} 02/25/2022 17:59:29 - INFO - codeparrot_training - Step 31193: {'lr': 0.0001666562614933668, 'samples': 15971328, 'steps': 31193, 'loss/train': 0.76556396484375} 02/25/2022 17:59:35 - INFO - codeparrot_training - Step 31194: {'lr': 0.00016664083523581106, 'samples': 15971840, 'steps': 31194, 'loss/train': 0.8046711087226868} 02/25/2022 17:59:38 - INFO - codeparrot_training - Step 31195: {'lr': 0.0001666254093353395, 'samples': 15972352, 'steps': 31195, 'loss/train': 1.5861634016036987} 02/25/2022 17:59:43 - INFO - codeparrot_training - Step 31196: {'lr': 0.0001666099837920182, 'samples': 15972864, 'steps': 31196, 'loss/train': 1.3574225902557373} 02/25/2022 17:59:47 - INFO - codeparrot_training - Step 31197: {'lr': 0.00016659455860591323, 'samples': 15973376, 'steps': 31197, 'loss/train': 1.2641143798828125} 02/25/2022 17:59:52 - INFO - codeparrot_training - Step 31198: {'lr': 0.00016657913377709078, 'samples': 15973888, 'steps': 31198, 'loss/train': 1.5527794361114502} 02/25/2022 17:59:56 - INFO - codeparrot_training - Step 31199: {'lr': 0.0001665637093056167, 'samples': 15974400, 'steps': 31199, 'loss/train': 1.507619023323059} 02/25/2022 18:00:02 - INFO - codeparrot_training - Step 31200: {'lr': 0.00016654828519155728, 'samples': 15974912, 'steps': 31200, 'loss/train': 1.67177152633667} 02/25/2022 18:00:05 - INFO - codeparrot_training - Step 31201: {'lr': 0.00016653286143497852, 'samples': 15975424, 'steps': 31201, 'loss/train': 1.2310543060302734} 02/25/2022 18:00:11 - INFO - codeparrot_training - Step 31202: {'lr': 0.00016651743803594644, 'samples': 15975936, 'steps': 31202, 'loss/train': 1.9363353252410889} 02/25/2022 18:00:14 - INFO - codeparrot_training - Step 31203: {'lr': 0.00016650201499452714, 'samples': 15976448, 'steps': 31203, 'loss/train': 1.2398371696472168} 02/25/2022 18:00:18 - INFO - codeparrot_training - Step 31204: {'lr': 0.00016648659231078666, 'samples': 15976960, 'steps': 31204, 'loss/train': 0.5455965399742126} 02/25/2022 18:00:24 - INFO - codeparrot_training - Step 31205: {'lr': 0.00016647116998479123, 'samples': 15977472, 'steps': 31205, 'loss/train': 0.49342748522758484} 02/25/2022 18:00:28 - INFO - codeparrot_training - Step 31206: {'lr': 0.00016645574801660667, 'samples': 15977984, 'steps': 31206, 'loss/train': 2.770235061645508} 02/25/2022 18:00:33 - INFO - codeparrot_training - Step 31207: {'lr': 0.0001664403264062992, 'samples': 15978496, 'steps': 31207, 'loss/train': 2.9392635822296143} 02/25/2022 18:00:36 - INFO - codeparrot_training - Step 31208: {'lr': 0.0001664249051539348, 'samples': 15979008, 'steps': 31208, 'loss/train': 1.678147554397583} 02/25/2022 18:00:42 - INFO - codeparrot_training - Step 31209: {'lr': 0.00016640948425957966, 'samples': 15979520, 'steps': 31209, 'loss/train': 0.8570581674575806} 02/25/2022 18:00:46 - INFO - codeparrot_training - Step 31210: {'lr': 0.0001663940637232997, 'samples': 15980032, 'steps': 31210, 'loss/train': 1.4423526525497437} 02/25/2022 18:00:51 - INFO - codeparrot_training - Step 31211: {'lr': 0.00016637864354516097, 'samples': 15980544, 'steps': 31211, 'loss/train': 1.9763888120651245} 02/25/2022 18:00:55 - INFO - codeparrot_training - Step 31212: {'lr': 0.00016636322372522955, 'samples': 15981056, 'steps': 31212, 'loss/train': 1.3488616943359375} 02/25/2022 18:01:01 - INFO - codeparrot_training - Step 31213: {'lr': 0.0001663478042635717, 'samples': 15981568, 'steps': 31213, 'loss/train': 1.6343997716903687} 02/25/2022 18:01:04 - INFO - codeparrot_training - Step 31214: {'lr': 0.00016633238516025316, 'samples': 15982080, 'steps': 31214, 'loss/train': 1.4659315347671509} 02/25/2022 18:01:10 - INFO - codeparrot_training - Step 31215: {'lr': 0.0001663169664153401, 'samples': 15982592, 'steps': 31215, 'loss/train': 1.0421639680862427} 02/25/2022 18:01:15 - INFO - codeparrot_training - Step 31216: {'lr': 0.00016630154802889857, 'samples': 15983104, 'steps': 31216, 'loss/train': 3.022994041442871} 02/25/2022 18:01:19 - INFO - codeparrot_training - Step 31217: {'lr': 0.0001662861300009947, 'samples': 15983616, 'steps': 31217, 'loss/train': 0.4751278758049011} 02/25/2022 18:01:24 - INFO - codeparrot_training - Step 31218: {'lr': 0.0001662707123316945, 'samples': 15984128, 'steps': 31218, 'loss/train': 1.8457940816879272} 02/25/2022 18:01:28 - INFO - codeparrot_training - Step 31219: {'lr': 0.0001662552950210639, 'samples': 15984640, 'steps': 31219, 'loss/train': 1.4903556108474731} 02/25/2022 18:01:33 - INFO - codeparrot_training - Step 31220: {'lr': 0.00016623987806916902, 'samples': 15985152, 'steps': 31220, 'loss/train': 2.1030163764953613} 02/25/2022 18:01:37 - INFO - codeparrot_training - Step 31221: {'lr': 0.0001662244614760759, 'samples': 15985664, 'steps': 31221, 'loss/train': 2.705338954925537} 02/25/2022 18:01:42 - INFO - codeparrot_training - Step 31222: {'lr': 0.00016620904524185072, 'samples': 15986176, 'steps': 31222, 'loss/train': 1.33682119846344} 02/25/2022 18:01:46 - INFO - codeparrot_training - Step 31223: {'lr': 0.00016619362936655927, 'samples': 15986688, 'steps': 31223, 'loss/train': 0.3181931972503662} 02/25/2022 18:01:52 - INFO - codeparrot_training - Step 31224: {'lr': 0.0001661782138502677, 'samples': 15987200, 'steps': 31224, 'loss/train': 2.599428415298462} 02/25/2022 18:01:56 - INFO - codeparrot_training - Step 31225: {'lr': 0.00016616279869304201, 'samples': 15987712, 'steps': 31225, 'loss/train': 0.8624588251113892} 02/25/2022 18:02:01 - INFO - codeparrot_training - Step 31226: {'lr': 0.00016614738389494834, 'samples': 15988224, 'steps': 31226, 'loss/train': 1.3411211967468262} 02/25/2022 18:02:05 - INFO - codeparrot_training - Step 31227: {'lr': 0.00016613196945605268, 'samples': 15988736, 'steps': 31227, 'loss/train': 2.5998666286468506} 02/25/2022 18:02:10 - INFO - codeparrot_training - Step 31228: {'lr': 0.00016611655537642097, 'samples': 15989248, 'steps': 31228, 'loss/train': 1.2592430114746094} 02/25/2022 18:02:14 - INFO - codeparrot_training - Step 31229: {'lr': 0.00016610114165611928, 'samples': 15989760, 'steps': 31229, 'loss/train': 1.348429799079895} 02/25/2022 18:02:19 - INFO - codeparrot_training - Step 31230: {'lr': 0.00016608572829521367, 'samples': 15990272, 'steps': 31230, 'loss/train': 2.002760410308838} 02/25/2022 18:02:23 - INFO - codeparrot_training - Step 31231: {'lr': 0.00016607031529377025, 'samples': 15990784, 'steps': 31231, 'loss/train': 1.2538737058639526} 02/25/2022 18:02:28 - INFO - codeparrot_training - Step 31232: {'lr': 0.00016605490265185485, 'samples': 15991296, 'steps': 31232, 'loss/train': 1.1970288753509521} 02/25/2022 18:02:32 - INFO - codeparrot_training - Step 31233: {'lr': 0.0001660394903695336, 'samples': 15991808, 'steps': 31233, 'loss/train': 1.6322320699691772} 02/25/2022 18:02:37 - INFO - codeparrot_training - Step 31234: {'lr': 0.00016602407844687255, 'samples': 15992320, 'steps': 31234, 'loss/train': 2.013763427734375} 02/25/2022 18:02:41 - INFO - codeparrot_training - Step 31235: {'lr': 0.00016600866688393768, 'samples': 15992832, 'steps': 31235, 'loss/train': 1.3738502264022827} 02/25/2022 18:02:46 - INFO - codeparrot_training - Step 31236: {'lr': 0.000165993255680795, 'samples': 15993344, 'steps': 31236, 'loss/train': 1.6297922134399414} 02/25/2022 18:02:50 - INFO - codeparrot_training - Step 31237: {'lr': 0.00016597784483751048, 'samples': 15993856, 'steps': 31237, 'loss/train': 1.540042757987976} 02/25/2022 18:02:55 - INFO - codeparrot_training - Step 31238: {'lr': 0.00016596243435415027, 'samples': 15994368, 'steps': 31238, 'loss/train': 1.8797917366027832} 02/25/2022 18:02:59 - INFO - codeparrot_training - Step 31239: {'lr': 0.00016594702423078028, 'samples': 15994880, 'steps': 31239, 'loss/train': 1.9537409543991089} 02/25/2022 18:03:05 - INFO - codeparrot_training - Step 31240: {'lr': 0.00016593161446746657, 'samples': 15995392, 'steps': 31240, 'loss/train': 2.262458324432373} 02/25/2022 18:03:09 - INFO - codeparrot_training - Step 31241: {'lr': 0.00016591620506427513, 'samples': 15995904, 'steps': 31241, 'loss/train': 1.32154381275177} 02/25/2022 18:03:14 - INFO - codeparrot_training - Step 31242: {'lr': 0.00016590079602127192, 'samples': 15996416, 'steps': 31242, 'loss/train': 1.9989919662475586} 02/25/2022 18:03:18 - INFO - codeparrot_training - Step 31243: {'lr': 0.000165885387338523, 'samples': 15996928, 'steps': 31243, 'loss/train': 1.2249836921691895} 02/25/2022 18:03:23 - INFO - codeparrot_training - Step 31244: {'lr': 0.0001658699790160944, 'samples': 15997440, 'steps': 31244, 'loss/train': 1.271769642829895} 02/25/2022 18:03:26 - INFO - codeparrot_training - Step 31245: {'lr': 0.0001658545710540521, 'samples': 15997952, 'steps': 31245, 'loss/train': 1.6355676651000977} 02/25/2022 18:03:32 - INFO - codeparrot_training - Step 31246: {'lr': 0.00016583916345246208, 'samples': 15998464, 'steps': 31246, 'loss/train': 1.394234299659729} 02/25/2022 18:03:35 - INFO - codeparrot_training - Step 31247: {'lr': 0.00016582375621139033, 'samples': 15998976, 'steps': 31247, 'loss/train': 2.311051845550537} 02/25/2022 18:03:41 - INFO - codeparrot_training - Step 31248: {'lr': 0.000165808349330903, 'samples': 15999488, 'steps': 31248, 'loss/train': 2.1872732639312744} 02/25/2022 18:03:47 - INFO - codeparrot_training - Step 31249: {'lr': 0.00016579294281106587, 'samples': 16000000, 'steps': 31249, 'loss/train': 0.5892047882080078} 02/25/2022 18:03:51 - INFO - codeparrot_training - Step 31250: {'lr': 0.000165777536651945, 'samples': 16000512, 'steps': 31250, 'loss/train': 1.1025352478027344} 02/25/2022 18:03:56 - INFO - codeparrot_training - Step 31251: {'lr': 0.00016576213085360648, 'samples': 16001024, 'steps': 31251, 'loss/train': 1.4451408386230469} 02/25/2022 18:04:00 - INFO - codeparrot_training - Step 31252: {'lr': 0.00016574672541611618, 'samples': 16001536, 'steps': 31252, 'loss/train': 1.3800761699676514} 02/25/2022 18:04:05 - INFO - codeparrot_training - Step 31253: {'lr': 0.0001657313203395403, 'samples': 16002048, 'steps': 31253, 'loss/train': 1.1232661008834839} 02/25/2022 18:04:09 - INFO - codeparrot_training - Step 31254: {'lr': 0.00016571591562394457, 'samples': 16002560, 'steps': 31254, 'loss/train': 1.5900959968566895} 02/25/2022 18:04:14 - INFO - codeparrot_training - Step 31255: {'lr': 0.00016570051126939505, 'samples': 16003072, 'steps': 31255, 'loss/train': 1.1457480192184448} 02/25/2022 18:04:18 - INFO - codeparrot_training - Step 31256: {'lr': 0.0001656851072759578, 'samples': 16003584, 'steps': 31256, 'loss/train': 2.219452381134033} 02/25/2022 18:04:23 - INFO - codeparrot_training - Step 31257: {'lr': 0.00016566970364369893, 'samples': 16004096, 'steps': 31257, 'loss/train': 1.8960309028625488} 02/25/2022 18:04:27 - INFO - codeparrot_training - Step 31258: {'lr': 0.00016565430037268413, 'samples': 16004608, 'steps': 31258, 'loss/train': 1.7764841318130493} 02/25/2022 18:04:33 - INFO - codeparrot_training - Step 31259: {'lr': 0.0001656388974629795, 'samples': 16005120, 'steps': 31259, 'loss/train': 1.7179733514785767} 02/25/2022 18:04:36 - INFO - codeparrot_training - Step 31260: {'lr': 0.0001656234949146511, 'samples': 16005632, 'steps': 31260, 'loss/train': 1.6806409358978271} 02/25/2022 18:04:42 - INFO - codeparrot_training - Step 31261: {'lr': 0.00016560809272776492, 'samples': 16006144, 'steps': 31261, 'loss/train': 2.0297491550445557} 02/25/2022 18:04:45 - INFO - codeparrot_training - Step 31262: {'lr': 0.00016559269090238676, 'samples': 16006656, 'steps': 31262, 'loss/train': 0.8652766346931458} 02/25/2022 18:04:51 - INFO - codeparrot_training - Step 31263: {'lr': 0.00016557728943858275, 'samples': 16007168, 'steps': 31263, 'loss/train': 1.8180917501449585} 02/25/2022 18:04:54 - INFO - codeparrot_training - Step 31264: {'lr': 0.00016556188833641877, 'samples': 16007680, 'steps': 31264, 'loss/train': 1.8327325582504272} 02/25/2022 18:05:00 - INFO - codeparrot_training - Step 31265: {'lr': 0.00016554648759596093, 'samples': 16008192, 'steps': 31265, 'loss/train': 1.4832602739334106} 02/25/2022 18:05:03 - INFO - codeparrot_training - Step 31266: {'lr': 0.00016553108721727516, 'samples': 16008704, 'steps': 31266, 'loss/train': 1.712033987045288} 02/25/2022 18:05:09 - INFO - codeparrot_training - Step 31267: {'lr': 0.00016551568720042732, 'samples': 16009216, 'steps': 31267, 'loss/train': 0.8692848682403564} 02/25/2022 18:05:12 - INFO - codeparrot_training - Step 31268: {'lr': 0.00016550028754548342, 'samples': 16009728, 'steps': 31268, 'loss/train': 2.2852439880371094} 02/25/2022 18:05:18 - INFO - codeparrot_training - Step 31269: {'lr': 0.00016548488825250944, 'samples': 16010240, 'steps': 31269, 'loss/train': 1.9843685626983643} 02/25/2022 18:05:21 - INFO - codeparrot_training - Step 31270: {'lr': 0.00016546948932157153, 'samples': 16010752, 'steps': 31270, 'loss/train': 1.6326148509979248} 02/25/2022 18:05:27 - INFO - codeparrot_training - Step 31271: {'lr': 0.00016545409075273532, 'samples': 16011264, 'steps': 31271, 'loss/train': 1.526254415512085} 02/25/2022 18:05:31 - INFO - codeparrot_training - Step 31272: {'lr': 0.00016543869254606697, 'samples': 16011776, 'steps': 31272, 'loss/train': 1.3565102815628052} 02/25/2022 18:05:36 - INFO - codeparrot_training - Step 31273: {'lr': 0.0001654232947016324, 'samples': 16012288, 'steps': 31273, 'loss/train': 1.2636314630508423} 02/25/2022 18:05:40 - INFO - codeparrot_training - Step 31274: {'lr': 0.00016540789721949768, 'samples': 16012800, 'steps': 31274, 'loss/train': 1.8139207363128662} 02/25/2022 18:05:45 - INFO - codeparrot_training - Step 31275: {'lr': 0.00016539250009972855, 'samples': 16013312, 'steps': 31275, 'loss/train': 1.3705881834030151} 02/25/2022 18:05:49 - INFO - codeparrot_training - Step 31276: {'lr': 0.00016537710334239115, 'samples': 16013824, 'steps': 31276, 'loss/train': 1.38565194606781} 02/25/2022 18:05:54 - INFO - codeparrot_training - Step 31277: {'lr': 0.00016536170694755133, 'samples': 16014336, 'steps': 31277, 'loss/train': 0.8240061402320862} 02/25/2022 18:05:58 - INFO - codeparrot_training - Step 31278: {'lr': 0.0001653463109152751, 'samples': 16014848, 'steps': 31278, 'loss/train': 1.8872151374816895} 02/25/2022 18:06:03 - INFO - codeparrot_training - Step 31279: {'lr': 0.00016533091524562844, 'samples': 16015360, 'steps': 31279, 'loss/train': 1.7890673875808716} 02/25/2022 18:06:07 - INFO - codeparrot_training - Step 31280: {'lr': 0.00016531551993867715, 'samples': 16015872, 'steps': 31280, 'loss/train': 0.13869307935237885} 02/25/2022 18:06:13 - INFO - codeparrot_training - Step 31281: {'lr': 0.00016530012499448737, 'samples': 16016384, 'steps': 31281, 'loss/train': 0.7887718677520752} 02/25/2022 18:06:16 - INFO - codeparrot_training - Step 31282: {'lr': 0.00016528473041312492, 'samples': 16016896, 'steps': 31282, 'loss/train': 2.1967010498046875} 02/25/2022 18:06:22 - INFO - codeparrot_training - Step 31283: {'lr': 0.0001652693361946558, 'samples': 16017408, 'steps': 31283, 'loss/train': 1.0679566860198975} 02/25/2022 18:06:25 - INFO - codeparrot_training - Step 31284: {'lr': 0.00016525394233914592, 'samples': 16017920, 'steps': 31284, 'loss/train': 1.8274058103561401} 02/25/2022 18:06:31 - INFO - codeparrot_training - Step 31285: {'lr': 0.0001652385488466613, 'samples': 16018432, 'steps': 31285, 'loss/train': 2.4772987365722656} 02/25/2022 18:06:35 - INFO - codeparrot_training - Step 31286: {'lr': 0.00016522315571726773, 'samples': 16018944, 'steps': 31286, 'loss/train': 0.8367212414741516} 02/25/2022 18:06:40 - INFO - codeparrot_training - Step 31287: {'lr': 0.00016520776295103127, 'samples': 16019456, 'steps': 31287, 'loss/train': 1.7551041841506958} 02/25/2022 18:06:44 - INFO - codeparrot_training - Step 31288: {'lr': 0.00016519237054801793, 'samples': 16019968, 'steps': 31288, 'loss/train': 2.6851449012756348} 02/25/2022 18:06:49 - INFO - codeparrot_training - Step 31289: {'lr': 0.00016517697850829344, 'samples': 16020480, 'steps': 31289, 'loss/train': 1.8984678983688354} 02/25/2022 18:06:53 - INFO - codeparrot_training - Step 31290: {'lr': 0.00016516158683192383, 'samples': 16020992, 'steps': 31290, 'loss/train': 1.0831626653671265} 02/25/2022 18:06:58 - INFO - codeparrot_training - Step 31291: {'lr': 0.00016514619551897503, 'samples': 16021504, 'steps': 31291, 'loss/train': 0.8535380363464355} 02/25/2022 18:07:02 - INFO - codeparrot_training - Step 31292: {'lr': 0.00016513080456951313, 'samples': 16022016, 'steps': 31292, 'loss/train': 0.36174851655960083} 02/25/2022 18:07:08 - INFO - codeparrot_training - Step 31293: {'lr': 0.00016511541398360377, 'samples': 16022528, 'steps': 31293, 'loss/train': 2.327173948287964} 02/25/2022 18:07:11 - INFO - codeparrot_training - Step 31294: {'lr': 0.00016510002376131307, 'samples': 16023040, 'steps': 31294, 'loss/train': 2.3153064250946045} 02/25/2022 18:07:15 - INFO - codeparrot_training - Step 31295: {'lr': 0.00016508463390270685, 'samples': 16023552, 'steps': 31295, 'loss/train': 2.1993751525878906} 02/25/2022 18:07:22 - INFO - codeparrot_training - Step 31296: {'lr': 0.00016506924440785122, 'samples': 16024064, 'steps': 31296, 'loss/train': 1.4185014963150024} 02/25/2022 18:07:25 - INFO - codeparrot_training - Step 31297: {'lr': 0.00016505385527681188, 'samples': 16024576, 'steps': 31297, 'loss/train': 1.308010458946228} 02/25/2022 18:07:31 - INFO - codeparrot_training - Step 31298: {'lr': 0.00016503846650965488, 'samples': 16025088, 'steps': 31298, 'loss/train': 4.247918128967285} 02/25/2022 18:07:34 - INFO - codeparrot_training - Step 31299: {'lr': 0.00016502307810644607, 'samples': 16025600, 'steps': 31299, 'loss/train': 2.6074845790863037} 02/25/2022 18:07:40 - INFO - codeparrot_training - Step 31300: {'lr': 0.00016500769006725142, 'samples': 16026112, 'steps': 31300, 'loss/train': 0.8778741955757141} 02/25/2022 18:07:43 - INFO - codeparrot_training - Step 31301: {'lr': 0.00016499230239213695, 'samples': 16026624, 'steps': 31301, 'loss/train': 1.4500478506088257} 02/25/2022 18:07:49 - INFO - codeparrot_training - Step 31302: {'lr': 0.00016497691508116837, 'samples': 16027136, 'steps': 31302, 'loss/train': 1.902868390083313} 02/25/2022 18:07:52 - INFO - codeparrot_training - Step 31303: {'lr': 0.00016496152813441165, 'samples': 16027648, 'steps': 31303, 'loss/train': 1.194953203201294} 02/25/2022 18:07:58 - INFO - codeparrot_training - Step 31304: {'lr': 0.00016494614155193276, 'samples': 16028160, 'steps': 31304, 'loss/train': 1.6722511053085327} 02/25/2022 18:08:03 - INFO - codeparrot_training - Step 31305: {'lr': 0.0001649307553337977, 'samples': 16028672, 'steps': 31305, 'loss/train': 1.999638557434082} 02/25/2022 18:08:07 - INFO - codeparrot_training - Step 31306: {'lr': 0.00016491536948007218, 'samples': 16029184, 'steps': 31306, 'loss/train': 1.7298903465270996} 02/25/2022 18:08:13 - INFO - codeparrot_training - Step 31307: {'lr': 0.00016489998399082218, 'samples': 16029696, 'steps': 31307, 'loss/train': 1.7059166431427002} 02/25/2022 18:08:17 - INFO - codeparrot_training - Step 31308: {'lr': 0.00016488459886611362, 'samples': 16030208, 'steps': 31308, 'loss/train': 1.421174168586731} 02/25/2022 18:08:22 - INFO - codeparrot_training - Step 31309: {'lr': 0.0001648692141060125, 'samples': 16030720, 'steps': 31309, 'loss/train': 0.2620045840740204} 02/25/2022 18:08:26 - INFO - codeparrot_training - Step 31310: {'lr': 0.0001648538297105846, 'samples': 16031232, 'steps': 31310, 'loss/train': 1.2898287773132324} 02/25/2022 18:08:31 - INFO - codeparrot_training - Step 31311: {'lr': 0.00016483844567989582, 'samples': 16031744, 'steps': 31311, 'loss/train': 1.1318995952606201} 02/25/2022 18:08:35 - INFO - codeparrot_training - Step 31312: {'lr': 0.0001648230620140121, 'samples': 16032256, 'steps': 31312, 'loss/train': 2.052733898162842} 02/25/2022 18:08:40 - INFO - codeparrot_training - Step 31313: {'lr': 0.00016480767871299932, 'samples': 16032768, 'steps': 31313, 'loss/train': 2.0317342281341553} 02/25/2022 18:08:44 - INFO - codeparrot_training - Step 31314: {'lr': 0.00016479229577692352, 'samples': 16033280, 'steps': 31314, 'loss/train': 1.9304574728012085} 02/25/2022 18:08:49 - INFO - codeparrot_training - Step 31315: {'lr': 0.00016477691320585032, 'samples': 16033792, 'steps': 31315, 'loss/train': 2.079672336578369} 02/25/2022 18:08:53 - INFO - codeparrot_training - Step 31316: {'lr': 0.00016476153099984582, 'samples': 16034304, 'steps': 31316, 'loss/train': 1.5480804443359375} 02/25/2022 18:08:58 - INFO - codeparrot_training - Step 31317: {'lr': 0.0001647461491589758, 'samples': 16034816, 'steps': 31317, 'loss/train': 1.8649154901504517} 02/25/2022 18:09:02 - INFO - codeparrot_training - Step 31318: {'lr': 0.00016473076768330633, 'samples': 16035328, 'steps': 31318, 'loss/train': 4.1953277587890625} 02/25/2022 18:09:07 - INFO - codeparrot_training - Step 31319: {'lr': 0.0001647153865729031, 'samples': 16035840, 'steps': 31319, 'loss/train': 0.2611694633960724} 02/25/2022 18:09:11 - INFO - codeparrot_training - Step 31320: {'lr': 0.00016470000582783205, 'samples': 16036352, 'steps': 31320, 'loss/train': 1.3507462739944458} 02/25/2022 18:09:17 - INFO - codeparrot_training - Step 31321: {'lr': 0.00016468462544815912, 'samples': 16036864, 'steps': 31321, 'loss/train': 3.5554237365722656} 02/25/2022 18:09:21 - INFO - codeparrot_training - Step 31322: {'lr': 0.00016466924543395017, 'samples': 16037376, 'steps': 31322, 'loss/train': 2.0216739177703857} 02/25/2022 18:09:26 - INFO - codeparrot_training - Step 31323: {'lr': 0.00016465386578527107, 'samples': 16037888, 'steps': 31323, 'loss/train': 1.93572199344635} 02/25/2022 18:09:30 - INFO - codeparrot_training - Step 31324: {'lr': 0.00016463848650218765, 'samples': 16038400, 'steps': 31324, 'loss/train': 2.1210403442382812} 02/25/2022 18:09:35 - INFO - codeparrot_training - Step 31325: {'lr': 0.00016462310758476596, 'samples': 16038912, 'steps': 31325, 'loss/train': 3.008291482925415} 02/25/2022 18:09:39 - INFO - codeparrot_training - Step 31326: {'lr': 0.0001646077290330717, 'samples': 16039424, 'steps': 31326, 'loss/train': 0.7228884100914001} 02/25/2022 18:09:44 - INFO - codeparrot_training - Step 31327: {'lr': 0.00016459235084717084, 'samples': 16039936, 'steps': 31327, 'loss/train': 1.9843220710754395} 02/25/2022 18:09:47 - INFO - codeparrot_training - Step 31328: {'lr': 0.00016457697302712918, 'samples': 16040448, 'steps': 31328, 'loss/train': 2.1160051822662354} 02/25/2022 18:09:53 - INFO - codeparrot_training - Step 31329: {'lr': 0.0001645615955730127, 'samples': 16040960, 'steps': 31329, 'loss/train': 2.9800057411193848} 02/25/2022 18:09:56 - INFO - codeparrot_training - Step 31330: {'lr': 0.00016454621848488714, 'samples': 16041472, 'steps': 31330, 'loss/train': 1.9088553190231323} 02/25/2022 18:10:02 - INFO - codeparrot_training - Step 31331: {'lr': 0.00016453084176281855, 'samples': 16041984, 'steps': 31331, 'loss/train': 0.1931828409433365} 02/25/2022 18:10:05 - INFO - codeparrot_training - Step 31332: {'lr': 0.0001645154654068727, 'samples': 16042496, 'steps': 31332, 'loss/train': 1.522559642791748} 02/25/2022 18:10:12 - INFO - codeparrot_training - Step 31333: {'lr': 0.0001645000894171153, 'samples': 16043008, 'steps': 31333, 'loss/train': 1.8941537141799927} 02/25/2022 18:10:15 - INFO - codeparrot_training - Step 31334: {'lr': 0.00016448471379361247, 'samples': 16043520, 'steps': 31334, 'loss/train': 1.9524195194244385} 02/25/2022 18:10:21 - INFO - codeparrot_training - Step 31335: {'lr': 0.00016446933853642991, 'samples': 16044032, 'steps': 31335, 'loss/train': 1.8290770053863525} 02/25/2022 18:10:24 - INFO - codeparrot_training - Step 31336: {'lr': 0.0001644539636456337, 'samples': 16044544, 'steps': 31336, 'loss/train': 1.6625949144363403} 02/25/2022 18:10:30 - INFO - codeparrot_training - Step 31337: {'lr': 0.0001644385891212894, 'samples': 16045056, 'steps': 31337, 'loss/train': 1.8477418422698975} 02/25/2022 18:10:33 - INFO - codeparrot_training - Step 31338: {'lr': 0.000164423214963463, 'samples': 16045568, 'steps': 31338, 'loss/train': 0.9607930183410645} 02/25/2022 18:10:39 - INFO - codeparrot_training - Step 31339: {'lr': 0.00016440784117222046, 'samples': 16046080, 'steps': 31339, 'loss/train': 1.4734187126159668} 02/25/2022 18:10:42 - INFO - codeparrot_training - Step 31340: {'lr': 0.0001643924677476276, 'samples': 16046592, 'steps': 31340, 'loss/train': 1.2680000066757202} 02/25/2022 18:10:49 - INFO - codeparrot_training - Step 31341: {'lr': 0.0001643770946897501, 'samples': 16047104, 'steps': 31341, 'loss/train': 0.39174580574035645} 02/25/2022 18:10:52 - INFO - codeparrot_training - Step 31342: {'lr': 0.00016436172199865395, 'samples': 16047616, 'steps': 31342, 'loss/train': 0.2295610010623932} 02/25/2022 18:10:58 - INFO - codeparrot_training - Step 31343: {'lr': 0.00016434634967440502, 'samples': 16048128, 'steps': 31343, 'loss/train': 1.4248918294906616} 02/25/2022 18:11:01 - INFO - codeparrot_training - Step 31344: {'lr': 0.0001643309777170692, 'samples': 16048640, 'steps': 31344, 'loss/train': 1.4873435497283936} 02/25/2022 18:11:07 - INFO - codeparrot_training - Step 31345: {'lr': 0.00016431560612671216, 'samples': 16049152, 'steps': 31345, 'loss/train': 1.9297418594360352} 02/25/2022 18:11:10 - INFO - codeparrot_training - Step 31346: {'lr': 0.0001643002349033999, 'samples': 16049664, 'steps': 31346, 'loss/train': 2.2651169300079346} 02/25/2022 18:11:16 - INFO - codeparrot_training - Step 31347: {'lr': 0.00016428486404719816, 'samples': 16050176, 'steps': 31347, 'loss/train': 1.835245132446289} 02/25/2022 18:11:19 - INFO - codeparrot_training - Step 31348: {'lr': 0.00016426949355817288, 'samples': 16050688, 'steps': 31348, 'loss/train': 1.778789758682251} 02/25/2022 18:11:25 - INFO - codeparrot_training - Step 31349: {'lr': 0.00016425412343638998, 'samples': 16051200, 'steps': 31349, 'loss/train': 2.3909201622009277} 02/25/2022 18:11:28 - INFO - codeparrot_training - Step 31350: {'lr': 0.00016423875368191506, 'samples': 16051712, 'steps': 31350, 'loss/train': 2.505403518676758} 02/25/2022 18:11:35 - INFO - codeparrot_training - Step 31351: {'lr': 0.0001642233842948141, 'samples': 16052224, 'steps': 31351, 'loss/train': 1.5558738708496094} 02/25/2022 18:11:38 - INFO - codeparrot_training - Step 31352: {'lr': 0.00016420801527515294, 'samples': 16052736, 'steps': 31352, 'loss/train': 2.3194472789764404} 02/25/2022 18:11:44 - INFO - codeparrot_training - Step 31353: {'lr': 0.00016419264662299749, 'samples': 16053248, 'steps': 31353, 'loss/train': 2.3688805103302} 02/25/2022 18:11:47 - INFO - codeparrot_training - Step 31354: {'lr': 0.0001641772783384134, 'samples': 16053760, 'steps': 31354, 'loss/train': 0.040782712399959564} 02/25/2022 18:11:53 - INFO - codeparrot_training - Step 31355: {'lr': 0.00016416191042146656, 'samples': 16054272, 'steps': 31355, 'loss/train': 1.2057888507843018} 02/25/2022 18:11:56 - INFO - codeparrot_training - Step 31356: {'lr': 0.00016414654287222287, 'samples': 16054784, 'steps': 31356, 'loss/train': 1.92635977268219} 02/25/2022 18:12:02 - INFO - codeparrot_training - Step 31357: {'lr': 0.00016413117569074822, 'samples': 16055296, 'steps': 31357, 'loss/train': 1.574373722076416} 02/25/2022 18:12:05 - INFO - codeparrot_training - Step 31358: {'lr': 0.00016411580887710825, 'samples': 16055808, 'steps': 31358, 'loss/train': 2.6870310306549072} 02/25/2022 18:12:11 - INFO - codeparrot_training - Step 31359: {'lr': 0.0001641004424313689, 'samples': 16056320, 'steps': 31359, 'loss/train': 2.15073823928833} 02/25/2022 18:12:14 - INFO - codeparrot_training - Step 31360: {'lr': 0.00016408507635359598, 'samples': 16056832, 'steps': 31360, 'loss/train': 2.0025622844696045} 02/25/2022 18:12:20 - INFO - codeparrot_training - Step 31361: {'lr': 0.0001640697106438553, 'samples': 16057344, 'steps': 31361, 'loss/train': 2.7186341285705566} 02/25/2022 18:12:24 - INFO - codeparrot_training - Step 31362: {'lr': 0.00016405434530221278, 'samples': 16057856, 'steps': 31362, 'loss/train': 1.7994041442871094} 02/25/2022 18:12:29 - INFO - codeparrot_training - Step 31363: {'lr': 0.00016403898032873405, 'samples': 16058368, 'steps': 31363, 'loss/train': 1.107377052307129} 02/25/2022 18:12:33 - INFO - codeparrot_training - Step 31364: {'lr': 0.00016402361572348507, 'samples': 16058880, 'steps': 31364, 'loss/train': 0.35256680846214294} 02/25/2022 18:12:38 - INFO - codeparrot_training - Step 31365: {'lr': 0.00016400825148653165, 'samples': 16059392, 'steps': 31365, 'loss/train': 1.1622322797775269} 02/25/2022 18:12:42 - INFO - codeparrot_training - Step 31366: {'lr': 0.00016399288761793956, 'samples': 16059904, 'steps': 31366, 'loss/train': 1.2829182147979736} 02/25/2022 18:12:48 - INFO - codeparrot_training - Step 31367: {'lr': 0.0001639775241177746, 'samples': 16060416, 'steps': 31367, 'loss/train': 1.6835365295410156} 02/25/2022 18:12:51 - INFO - codeparrot_training - Step 31368: {'lr': 0.0001639621609861026, 'samples': 16060928, 'steps': 31368, 'loss/train': 2.2982916831970215} 02/25/2022 18:12:57 - INFO - codeparrot_training - Step 31369: {'lr': 0.0001639467982229894, 'samples': 16061440, 'steps': 31369, 'loss/train': 2.454707622528076} 02/25/2022 18:13:00 - INFO - codeparrot_training - Step 31370: {'lr': 0.00016393143582850084, 'samples': 16061952, 'steps': 31370, 'loss/train': 2.0608131885528564} 02/25/2022 18:13:06 - INFO - codeparrot_training - Step 31371: {'lr': 0.00016391607380270262, 'samples': 16062464, 'steps': 31371, 'loss/train': 2.734260082244873} 02/25/2022 18:13:09 - INFO - codeparrot_training - Step 31372: {'lr': 0.00016390071214566067, 'samples': 16062976, 'steps': 31372, 'loss/train': 2.8478658199310303} 02/25/2022 18:13:15 - INFO - codeparrot_training - Step 31373: {'lr': 0.00016388535085744063, 'samples': 16063488, 'steps': 31373, 'loss/train': 2.019195556640625} 02/25/2022 18:13:19 - INFO - codeparrot_training - Step 31374: {'lr': 0.00016386998993810843, 'samples': 16064000, 'steps': 31374, 'loss/train': 2.0332553386688232} 02/25/2022 18:13:24 - INFO - codeparrot_training - Step 31375: {'lr': 0.0001638546293877299, 'samples': 16064512, 'steps': 31375, 'loss/train': 1.3325759172439575} 02/25/2022 18:13:28 - INFO - codeparrot_training - Step 31376: {'lr': 0.00016383926920637078, 'samples': 16065024, 'steps': 31376, 'loss/train': 1.8542050123214722} 02/25/2022 18:13:34 - INFO - codeparrot_training - Step 31377: {'lr': 0.00016382390939409684, 'samples': 16065536, 'steps': 31377, 'loss/train': 0.6890619993209839} 02/25/2022 18:13:38 - INFO - codeparrot_training - Step 31378: {'lr': 0.00016380854995097388, 'samples': 16066048, 'steps': 31378, 'loss/train': 0.48866522312164307} 02/25/2022 18:13:43 - INFO - codeparrot_training - Step 31379: {'lr': 0.00016379319087706784, 'samples': 16066560, 'steps': 31379, 'loss/train': 1.8394485712051392} 02/25/2022 18:13:47 - INFO - codeparrot_training - Step 31380: {'lr': 0.00016377783217244428, 'samples': 16067072, 'steps': 31380, 'loss/train': 1.0201244354248047} 02/25/2022 18:13:52 - INFO - codeparrot_training - Step 31381: {'lr': 0.0001637624738371691, 'samples': 16067584, 'steps': 31381, 'loss/train': 1.376876711845398} 02/25/2022 18:13:56 - INFO - codeparrot_training - Step 31382: {'lr': 0.00016374711587130813, 'samples': 16068096, 'steps': 31382, 'loss/train': 1.0548646450042725} 02/25/2022 18:14:01 - INFO - codeparrot_training - Step 31383: {'lr': 0.0001637317582749272, 'samples': 16068608, 'steps': 31383, 'loss/train': 1.7436797618865967} 02/25/2022 18:14:05 - INFO - codeparrot_training - Step 31384: {'lr': 0.00016371640104809193, 'samples': 16069120, 'steps': 31384, 'loss/train': 1.2995902299880981} 02/25/2022 18:14:10 - INFO - codeparrot_training - Step 31385: {'lr': 0.00016370104419086823, 'samples': 16069632, 'steps': 31385, 'loss/train': 1.227754831314087} 02/25/2022 18:14:14 - INFO - codeparrot_training - Step 31386: {'lr': 0.0001636856877033218, 'samples': 16070144, 'steps': 31386, 'loss/train': 1.9077762365341187} 02/25/2022 18:14:19 - INFO - codeparrot_training - Step 31387: {'lr': 0.00016367033158551852, 'samples': 16070656, 'steps': 31387, 'loss/train': 0.8147440552711487} 02/25/2022 18:14:23 - INFO - codeparrot_training - Step 31388: {'lr': 0.00016365497583752423, 'samples': 16071168, 'steps': 31388, 'loss/train': 1.2274079322814941} 02/25/2022 18:14:29 - INFO - codeparrot_training - Step 31389: {'lr': 0.00016363962045940449, 'samples': 16071680, 'steps': 31389, 'loss/train': 2.4627585411071777} 02/25/2022 18:14:33 - INFO - codeparrot_training - Step 31390: {'lr': 0.00016362426545122517, 'samples': 16072192, 'steps': 31390, 'loss/train': 1.0315213203430176} 02/25/2022 18:14:38 - INFO - codeparrot_training - Step 31391: {'lr': 0.0001636089108130521, 'samples': 16072704, 'steps': 31391, 'loss/train': 1.514729380607605} 02/25/2022 18:14:42 - INFO - codeparrot_training - Step 31392: {'lr': 0.00016359355654495112, 'samples': 16073216, 'steps': 31392, 'loss/train': 2.878905773162842} 02/25/2022 18:14:47 - INFO - codeparrot_training - Step 31393: {'lr': 0.0001635782026469878, 'samples': 16073728, 'steps': 31393, 'loss/train': 2.0227460861206055} 02/25/2022 18:14:51 - INFO - codeparrot_training - Step 31394: {'lr': 0.00016356284911922802, 'samples': 16074240, 'steps': 31394, 'loss/train': 1.5283938646316528} 02/25/2022 18:14:56 - INFO - codeparrot_training - Step 31395: {'lr': 0.00016354749596173758, 'samples': 16074752, 'steps': 31395, 'loss/train': 2.339381694793701} 02/25/2022 18:15:00 - INFO - codeparrot_training - Step 31396: {'lr': 0.00016353214317458214, 'samples': 16075264, 'steps': 31396, 'loss/train': 1.9438034296035767} 02/25/2022 18:15:05 - INFO - codeparrot_training - Step 31397: {'lr': 0.0001635167907578277, 'samples': 16075776, 'steps': 31397, 'loss/train': 0.17732861638069153} 02/25/2022 18:15:09 - INFO - codeparrot_training - Step 31398: {'lr': 0.00016350143871153976, 'samples': 16076288, 'steps': 31398, 'loss/train': 1.9937018156051636} 02/25/2022 18:15:14 - INFO - codeparrot_training - Step 31399: {'lr': 0.00016348608703578416, 'samples': 16076800, 'steps': 31399, 'loss/train': 1.9904096126556396} 02/25/2022 18:15:18 - INFO - codeparrot_training - Step 31400: {'lr': 0.0001634707357306267, 'samples': 16077312, 'steps': 31400, 'loss/train': 1.1059869527816772} 02/25/2022 18:15:23 - INFO - codeparrot_training - Step 31401: {'lr': 0.00016345538479613326, 'samples': 16077824, 'steps': 31401, 'loss/train': 2.338977336883545} 02/25/2022 18:15:27 - INFO - codeparrot_training - Step 31402: {'lr': 0.00016344003423236937, 'samples': 16078336, 'steps': 31402, 'loss/train': 1.0222047567367554} 02/25/2022 18:15:33 - INFO - codeparrot_training - Step 31403: {'lr': 0.00016342468403940087, 'samples': 16078848, 'steps': 31403, 'loss/train': 1.3300868272781372} 02/25/2022 18:15:36 - INFO - codeparrot_training - Step 31404: {'lr': 0.00016340933421729354, 'samples': 16079360, 'steps': 31404, 'loss/train': 2.7973134517669678} 02/25/2022 18:15:42 - INFO - codeparrot_training - Step 31405: {'lr': 0.00016339398476611323, 'samples': 16079872, 'steps': 31405, 'loss/train': 0.03932815417647362} 02/25/2022 18:15:45 - INFO - codeparrot_training - Step 31406: {'lr': 0.00016337863568592547, 'samples': 16080384, 'steps': 31406, 'loss/train': 1.5377815961837769} 02/25/2022 18:15:51 - INFO - codeparrot_training - Step 31407: {'lr': 0.00016336328697679614, 'samples': 16080896, 'steps': 31407, 'loss/train': 1.4611084461212158} 02/25/2022 18:15:55 - INFO - codeparrot_training - Step 31408: {'lr': 0.00016334793863879098, 'samples': 16081408, 'steps': 31408, 'loss/train': 2.497556209564209} 02/25/2022 18:16:00 - INFO - codeparrot_training - Step 31409: {'lr': 0.00016333259067197582, 'samples': 16081920, 'steps': 31409, 'loss/train': 1.7908082008361816} 02/25/2022 18:16:04 - INFO - codeparrot_training - Step 31410: {'lr': 0.00016331724307641626, 'samples': 16082432, 'steps': 31410, 'loss/train': 1.9612091779708862} 02/25/2022 18:16:09 - INFO - codeparrot_training - Step 31411: {'lr': 0.00016330189585217809, 'samples': 16082944, 'steps': 31411, 'loss/train': 1.506575107574463} 02/25/2022 18:16:13 - INFO - codeparrot_training - Step 31412: {'lr': 0.0001632865489993271, 'samples': 16083456, 'steps': 31412, 'loss/train': 0.2556702494621277} 02/25/2022 18:16:18 - INFO - codeparrot_training - Step 31413: {'lr': 0.00016327120251792898, 'samples': 16083968, 'steps': 31413, 'loss/train': 1.2654744386672974} 02/25/2022 18:16:22 - INFO - codeparrot_training - Step 31414: {'lr': 0.00016325585640804952, 'samples': 16084480, 'steps': 31414, 'loss/train': 2.5947704315185547} 02/25/2022 18:16:28 - INFO - codeparrot_training - Step 31415: {'lr': 0.00016324051066975438, 'samples': 16084992, 'steps': 31415, 'loss/train': 3.0848538875579834} 02/25/2022 18:16:31 - INFO - codeparrot_training - Step 31416: {'lr': 0.00016322516530310939, 'samples': 16085504, 'steps': 31416, 'loss/train': 2.0074641704559326} 02/25/2022 18:16:37 - INFO - codeparrot_training - Step 31417: {'lr': 0.00016320982030818015, 'samples': 16086016, 'steps': 31417, 'loss/train': 2.3168277740478516} 02/25/2022 18:16:40 - INFO - codeparrot_training - Step 31418: {'lr': 0.0001631944756850326, 'samples': 16086528, 'steps': 31418, 'loss/train': 1.6213756799697876} 02/25/2022 18:16:46 - INFO - codeparrot_training - Step 31419: {'lr': 0.0001631791314337323, 'samples': 16087040, 'steps': 31419, 'loss/train': 2.0430426597595215} 02/25/2022 18:16:49 - INFO - codeparrot_training - Step 31420: {'lr': 0.00016316378755434503, 'samples': 16087552, 'steps': 31420, 'loss/train': 0.6122271418571472} 02/25/2022 18:16:55 - INFO - codeparrot_training - Step 31421: {'lr': 0.00016314844404693646, 'samples': 16088064, 'steps': 31421, 'loss/train': 2.4425461292266846} 02/25/2022 18:16:58 - INFO - codeparrot_training - Step 31422: {'lr': 0.0001631331009115724, 'samples': 16088576, 'steps': 31422, 'loss/train': 0.983812689781189} 02/25/2022 18:17:04 - INFO - codeparrot_training - Step 31423: {'lr': 0.0001631177581483187, 'samples': 16089088, 'steps': 31423, 'loss/train': 2.4030473232269287} 02/25/2022 18:17:07 - INFO - codeparrot_training - Step 31424: {'lr': 0.00016310241575724077, 'samples': 16089600, 'steps': 31424, 'loss/train': 1.9232796430587769} 02/25/2022 18:17:13 - INFO - codeparrot_training - Step 31425: {'lr': 0.00016308707373840453, 'samples': 16090112, 'steps': 31425, 'loss/train': 1.9096519947052002} 02/25/2022 18:17:17 - INFO - codeparrot_training - Step 31426: {'lr': 0.00016307173209187569, 'samples': 16090624, 'steps': 31426, 'loss/train': 1.6862930059432983} 02/25/2022 18:17:23 - INFO - codeparrot_training - Step 31427: {'lr': 0.00016305639081772, 'samples': 16091136, 'steps': 31427, 'loss/train': 0.8484938144683838} 02/25/2022 18:17:26 - INFO - codeparrot_training - Step 31428: {'lr': 0.00016304104991600305, 'samples': 16091648, 'steps': 31428, 'loss/train': 1.4951131343841553} 02/25/2022 18:17:32 - INFO - codeparrot_training - Step 31429: {'lr': 0.00016302570938679062, 'samples': 16092160, 'steps': 31429, 'loss/train': 2.608872652053833} 02/25/2022 18:17:35 - INFO - codeparrot_training - Step 31430: {'lr': 0.00016301036923014844, 'samples': 16092672, 'steps': 31430, 'loss/train': 1.015969157218933} 02/25/2022 18:17:40 - INFO - codeparrot_training - Step 31431: {'lr': 0.0001629950294461423, 'samples': 16093184, 'steps': 31431, 'loss/train': 0.4371148645877838} 02/25/2022 18:17:46 - INFO - codeparrot_training - Step 31432: {'lr': 0.00016297969003483775, 'samples': 16093696, 'steps': 31432, 'loss/train': 2.5788638591766357} 02/25/2022 18:17:49 - INFO - codeparrot_training - Step 31433: {'lr': 0.00016296435099630058, 'samples': 16094208, 'steps': 31433, 'loss/train': 2.3890678882598877} 02/25/2022 18:17:56 - INFO - codeparrot_training - Step 31434: {'lr': 0.00016294901233059647, 'samples': 16094720, 'steps': 31434, 'loss/train': 1.4159319400787354} 02/25/2022 18:17:59 - INFO - codeparrot_training - Step 31435: {'lr': 0.00016293367403779116, 'samples': 16095232, 'steps': 31435, 'loss/train': 2.2135400772094727} 02/25/2022 18:18:05 - INFO - codeparrot_training - Step 31436: {'lr': 0.00016291833611795046, 'samples': 16095744, 'steps': 31436, 'loss/train': 2.3922173976898193} 02/25/2022 18:18:08 - INFO - codeparrot_training - Step 31437: {'lr': 0.00016290299857113982, 'samples': 16096256, 'steps': 31437, 'loss/train': 1.6987357139587402} 02/25/2022 18:18:13 - INFO - codeparrot_training - Step 31438: {'lr': 0.00016288766139742512, 'samples': 16096768, 'steps': 31438, 'loss/train': 1.840745449066162} 02/25/2022 18:18:17 - INFO - codeparrot_training - Step 31439: {'lr': 0.00016287232459687197, 'samples': 16097280, 'steps': 31439, 'loss/train': 1.9439425468444824} 02/25/2022 18:18:22 - INFO - codeparrot_training - Step 31440: {'lr': 0.00016285698816954626, 'samples': 16097792, 'steps': 31440, 'loss/train': 2.0155298709869385} 02/25/2022 18:18:26 - INFO - codeparrot_training - Step 31441: {'lr': 0.0001628416521155134, 'samples': 16098304, 'steps': 31441, 'loss/train': 1.8807417154312134} 02/25/2022 18:18:32 - INFO - codeparrot_training - Step 31442: {'lr': 0.00016282631643483924, 'samples': 16098816, 'steps': 31442, 'loss/train': 1.6496045589447021} 02/25/2022 18:18:35 - INFO - codeparrot_training - Step 31443: {'lr': 0.00016281098112758949, 'samples': 16099328, 'steps': 31443, 'loss/train': 0.7996290326118469} 02/25/2022 18:18:41 - INFO - codeparrot_training - Step 31444: {'lr': 0.0001627956461938298, 'samples': 16099840, 'steps': 31444, 'loss/train': 0.5370369553565979} 02/25/2022 18:18:44 - INFO - codeparrot_training - Step 31445: {'lr': 0.00016278031163362595, 'samples': 16100352, 'steps': 31445, 'loss/train': 1.209165096282959} 02/25/2022 18:18:50 - INFO - codeparrot_training - Step 31446: {'lr': 0.00016276497744704347, 'samples': 16100864, 'steps': 31446, 'loss/train': 2.0173287391662598} 02/25/2022 18:18:53 - INFO - codeparrot_training - Step 31447: {'lr': 0.00016274964363414813, 'samples': 16101376, 'steps': 31447, 'loss/train': 0.4982140064239502} 02/25/2022 18:18:59 - INFO - codeparrot_training - Step 31448: {'lr': 0.00016273431019500558, 'samples': 16101888, 'steps': 31448, 'loss/train': 1.2553842067718506} 02/25/2022 18:19:02 - INFO - codeparrot_training - Step 31449: {'lr': 0.00016271897712968165, 'samples': 16102400, 'steps': 31449, 'loss/train': 0.6253923177719116} 02/25/2022 18:19:09 - INFO - codeparrot_training - Step 31450: {'lr': 0.00016270364443824183, 'samples': 16102912, 'steps': 31450, 'loss/train': 2.9799771308898926} 02/25/2022 18:19:12 - INFO - codeparrot_training - Step 31451: {'lr': 0.00016268831212075182, 'samples': 16103424, 'steps': 31451, 'loss/train': 1.1534920930862427} 02/25/2022 18:19:18 - INFO - codeparrot_training - Step 31452: {'lr': 0.00016267298017727744, 'samples': 16103936, 'steps': 31452, 'loss/train': 1.5033506155014038} 02/25/2022 18:19:21 - INFO - codeparrot_training - Step 31453: {'lr': 0.00016265764860788428, 'samples': 16104448, 'steps': 31453, 'loss/train': 1.624739170074463} 02/25/2022 18:19:27 - INFO - codeparrot_training - Step 31454: {'lr': 0.000162642317412638, 'samples': 16104960, 'steps': 31454, 'loss/train': 2.84039306640625} 02/25/2022 18:19:30 - INFO - codeparrot_training - Step 31455: {'lr': 0.00016262698659160428, 'samples': 16105472, 'steps': 31455, 'loss/train': 1.781067132949829} 02/25/2022 18:19:36 - INFO - codeparrot_training - Step 31456: {'lr': 0.00016261165614484885, 'samples': 16105984, 'steps': 31456, 'loss/train': 2.2916932106018066} 02/25/2022 18:19:39 - INFO - codeparrot_training - Step 31457: {'lr': 0.00016259632607243727, 'samples': 16106496, 'steps': 31457, 'loss/train': 1.4510235786437988} 02/25/2022 18:19:45 - INFO - codeparrot_training - Step 31458: {'lr': 0.00016258099637443536, 'samples': 16107008, 'steps': 31458, 'loss/train': 0.9987894296646118} 02/25/2022 18:19:48 - INFO - codeparrot_training - Step 31459: {'lr': 0.00016256566705090864, 'samples': 16107520, 'steps': 31459, 'loss/train': 2.0802817344665527} 02/25/2022 18:19:54 - INFO - codeparrot_training - Step 31460: {'lr': 0.00016255033810192284, 'samples': 16108032, 'steps': 31460, 'loss/train': 1.784799575805664} 02/25/2022 18:19:58 - INFO - codeparrot_training - Step 31461: {'lr': 0.00016253500952754363, 'samples': 16108544, 'steps': 31461, 'loss/train': 2.1759583950042725} 02/25/2022 18:20:03 - INFO - codeparrot_training - Step 31462: {'lr': 0.00016251968132783667, 'samples': 16109056, 'steps': 31462, 'loss/train': 1.330784559249878} 02/25/2022 18:20:07 - INFO - codeparrot_training - Step 31463: {'lr': 0.00016250435350286766, 'samples': 16109568, 'steps': 31463, 'loss/train': 2.2426869869232178} 02/25/2022 18:20:12 - INFO - codeparrot_training - Step 31464: {'lr': 0.00016248902605270216, 'samples': 16110080, 'steps': 31464, 'loss/train': 2.0271759033203125} 02/25/2022 18:20:16 - INFO - codeparrot_training - Step 31465: {'lr': 0.00016247369897740588, 'samples': 16110592, 'steps': 31465, 'loss/train': 1.1936595439910889} 02/25/2022 18:20:22 - INFO - codeparrot_training - Step 31466: {'lr': 0.00016245837227704457, 'samples': 16111104, 'steps': 31466, 'loss/train': 2.5051722526550293} 02/25/2022 18:20:25 - INFO - codeparrot_training - Step 31467: {'lr': 0.0001624430459516837, 'samples': 16111616, 'steps': 31467, 'loss/train': 1.6761746406555176} 02/25/2022 18:20:31 - INFO - codeparrot_training - Step 31468: {'lr': 0.000162427720001389, 'samples': 16112128, 'steps': 31468, 'loss/train': 0.8441166281700134} 02/25/2022 18:20:34 - INFO - codeparrot_training - Step 31469: {'lr': 0.00016241239442622618, 'samples': 16112640, 'steps': 31469, 'loss/train': 1.2163292169570923} 02/25/2022 18:20:41 - INFO - codeparrot_training - Step 31470: {'lr': 0.00016239706922626082, 'samples': 16113152, 'steps': 31470, 'loss/train': 1.8924944400787354} 02/25/2022 18:20:44 - INFO - codeparrot_training - Step 31471: {'lr': 0.00016238174440155872, 'samples': 16113664, 'steps': 31471, 'loss/train': 0.9663808345794678} 02/25/2022 18:20:49 - INFO - codeparrot_training - Step 31472: {'lr': 0.0001623664199521853, 'samples': 16114176, 'steps': 31472, 'loss/train': 2.631345748901367} 02/25/2022 18:20:53 - INFO - codeparrot_training - Step 31473: {'lr': 0.00016235109587820628, 'samples': 16114688, 'steps': 31473, 'loss/train': 1.7658894062042236} 02/25/2022 18:20:58 - INFO - codeparrot_training - Step 31474: {'lr': 0.00016233577217968738, 'samples': 16115200, 'steps': 31474, 'loss/train': 2.008976936340332} 02/25/2022 18:21:02 - INFO - codeparrot_training - Step 31475: {'lr': 0.00016232044885669428, 'samples': 16115712, 'steps': 31475, 'loss/train': 2.281757354736328} 02/25/2022 18:21:08 - INFO - codeparrot_training - Step 31476: {'lr': 0.00016230512590929242, 'samples': 16116224, 'steps': 31476, 'loss/train': 1.5633317232131958} 02/25/2022 18:21:11 - INFO - codeparrot_training - Step 31477: {'lr': 0.00016228980333754756, 'samples': 16116736, 'steps': 31477, 'loss/train': 0.8711519837379456} 02/25/2022 18:21:17 - INFO - codeparrot_training - Step 31478: {'lr': 0.00016227448114152532, 'samples': 16117248, 'steps': 31478, 'loss/train': 0.7697080373764038} 02/25/2022 18:21:20 - INFO - codeparrot_training - Step 31479: {'lr': 0.00016225915932129148, 'samples': 16117760, 'steps': 31479, 'loss/train': 1.5787307024002075} 02/25/2022 18:21:26 - INFO - codeparrot_training - Step 31480: {'lr': 0.00016224383787691142, 'samples': 16118272, 'steps': 31480, 'loss/train': 1.3579158782958984} 02/25/2022 18:21:30 - INFO - codeparrot_training - Step 31481: {'lr': 0.00016222851680845087, 'samples': 16118784, 'steps': 31481, 'loss/train': 2.413877248764038} 02/25/2022 18:21:35 - INFO - codeparrot_training - Step 31482: {'lr': 0.0001622131961159755, 'samples': 16119296, 'steps': 31482, 'loss/train': 0.7387328147888184} 02/25/2022 18:21:39 - INFO - codeparrot_training - Step 31483: {'lr': 0.00016219787579955096, 'samples': 16119808, 'steps': 31483, 'loss/train': 1.0056289434432983} 02/25/2022 18:21:44 - INFO - codeparrot_training - Step 31484: {'lr': 0.0001621825558592429, 'samples': 16120320, 'steps': 31484, 'loss/train': 1.7875761985778809} 02/25/2022 18:21:48 - INFO - codeparrot_training - Step 31485: {'lr': 0.0001621672362951168, 'samples': 16120832, 'steps': 31485, 'loss/train': 1.2262378931045532} 02/25/2022 18:21:53 - INFO - codeparrot_training - Step 31486: {'lr': 0.00016215191710723838, 'samples': 16121344, 'steps': 31486, 'loss/train': 1.5009336471557617} 02/25/2022 18:21:57 - INFO - codeparrot_training - Step 31487: {'lr': 0.00016213659829567324, 'samples': 16121856, 'steps': 31487, 'loss/train': 1.916374683380127} 02/25/2022 18:22:02 - INFO - codeparrot_training - Step 31488: {'lr': 0.00016212127986048707, 'samples': 16122368, 'steps': 31488, 'loss/train': 2.270653009414673} 02/25/2022 18:22:06 - INFO - codeparrot_training - Step 31489: {'lr': 0.0001621059618017454, 'samples': 16122880, 'steps': 31489, 'loss/train': 1.6212043762207031} 02/25/2022 18:22:11 - INFO - codeparrot_training - Step 31490: {'lr': 0.00016209064411951384, 'samples': 16123392, 'steps': 31490, 'loss/train': 1.4815562963485718} 02/25/2022 18:22:15 - INFO - codeparrot_training - Step 31491: {'lr': 0.00016207532681385806, 'samples': 16123904, 'steps': 31491, 'loss/train': 2.0144875049591064} 02/25/2022 18:22:20 - INFO - codeparrot_training - Step 31492: {'lr': 0.0001620600098848437, 'samples': 16124416, 'steps': 31492, 'loss/train': 1.268331527709961} 02/25/2022 18:22:24 - INFO - codeparrot_training - Step 31493: {'lr': 0.00016204469333253634, 'samples': 16124928, 'steps': 31493, 'loss/train': 1.5123624801635742} 02/25/2022 18:22:29 - INFO - codeparrot_training - Step 31494: {'lr': 0.00016202937715700157, 'samples': 16125440, 'steps': 31494, 'loss/train': 1.5088878870010376} 02/25/2022 18:22:33 - INFO - codeparrot_training - Step 31495: {'lr': 0.00016201406135830498, 'samples': 16125952, 'steps': 31495, 'loss/train': 2.3948800563812256} 02/25/2022 18:22:39 - INFO - codeparrot_training - Step 31496: {'lr': 0.00016199874593651227, 'samples': 16126464, 'steps': 31496, 'loss/train': 1.7384083271026611} 02/25/2022 18:22:42 - INFO - codeparrot_training - Step 31497: {'lr': 0.00016198343089168899, 'samples': 16126976, 'steps': 31497, 'loss/train': 2.2056689262390137} 02/25/2022 18:22:48 - INFO - codeparrot_training - Step 31498: {'lr': 0.00016196811622390067, 'samples': 16127488, 'steps': 31498, 'loss/train': 2.7153568267822266} 02/25/2022 18:22:51 - INFO - codeparrot_training - Step 31499: {'lr': 0.00016195280193321304, 'samples': 16128000, 'steps': 31499, 'loss/train': 1.393386721611023} 02/25/2022 18:22:57 - INFO - codeparrot_training - Step 31500: {'lr': 0.00016193748801969163, 'samples': 16128512, 'steps': 31500, 'loss/train': 1.6365031003952026} 02/25/2022 18:23:00 - INFO - codeparrot_training - Step 31501: {'lr': 0.0001619221744834021, 'samples': 16129024, 'steps': 31501, 'loss/train': 2.2532615661621094} 02/25/2022 18:23:06 - INFO - codeparrot_training - Step 31502: {'lr': 0.00016190686132440995, 'samples': 16129536, 'steps': 31502, 'loss/train': 1.7475076913833618} 02/25/2022 18:23:09 - INFO - codeparrot_training - Step 31503: {'lr': 0.00016189154854278089, 'samples': 16130048, 'steps': 31503, 'loss/train': 2.1678028106689453} 02/25/2022 18:23:15 - INFO - codeparrot_training - Step 31504: {'lr': 0.00016187623613858038, 'samples': 16130560, 'steps': 31504, 'loss/train': 1.0537303686141968} 02/25/2022 18:23:18 - INFO - codeparrot_training - Step 31505: {'lr': 0.00016186092411187413, 'samples': 16131072, 'steps': 31505, 'loss/train': 1.9704115390777588} 02/25/2022 18:23:24 - INFO - codeparrot_training - Step 31506: {'lr': 0.00016184561246272778, 'samples': 16131584, 'steps': 31506, 'loss/train': 1.3763046264648438} 02/25/2022 18:23:28 - INFO - codeparrot_training - Step 31507: {'lr': 0.00016183030119120673, 'samples': 16132096, 'steps': 31507, 'loss/train': 3.0875656604766846} 02/25/2022 18:23:33 - INFO - codeparrot_training - Step 31508: {'lr': 0.0001618149902973767, 'samples': 16132608, 'steps': 31508, 'loss/train': 1.5790398120880127} 02/25/2022 18:23:37 - INFO - codeparrot_training - Step 31509: {'lr': 0.00016179967978130317, 'samples': 16133120, 'steps': 31509, 'loss/train': 1.7692136764526367} 02/25/2022 18:23:42 - INFO - codeparrot_training - Step 31510: {'lr': 0.00016178436964305197, 'samples': 16133632, 'steps': 31510, 'loss/train': 1.2031954526901245} 02/25/2022 18:23:46 - INFO - codeparrot_training - Step 31511: {'lr': 0.0001617690598826884, 'samples': 16134144, 'steps': 31511, 'loss/train': 1.9572832584381104} 02/25/2022 18:23:51 - INFO - codeparrot_training - Step 31512: {'lr': 0.00016175375050027818, 'samples': 16134656, 'steps': 31512, 'loss/train': 1.5301895141601562} 02/25/2022 18:23:55 - INFO - codeparrot_training - Step 31513: {'lr': 0.00016173844149588684, 'samples': 16135168, 'steps': 31513, 'loss/train': 1.6924278736114502} 02/25/2022 18:24:00 - INFO - codeparrot_training - Step 31514: {'lr': 0.00016172313286958012, 'samples': 16135680, 'steps': 31514, 'loss/train': 1.571447730064392} 02/25/2022 18:24:04 - INFO - codeparrot_training - Step 31515: {'lr': 0.00016170782462142337, 'samples': 16136192, 'steps': 31515, 'loss/train': 1.613728404045105} 02/25/2022 18:24:10 - INFO - codeparrot_training - Step 31516: {'lr': 0.00016169251675148222, 'samples': 16136704, 'steps': 31516, 'loss/train': 1.9871997833251953} 02/25/2022 18:24:14 - INFO - codeparrot_training - Step 31517: {'lr': 0.00016167720925982232, 'samples': 16137216, 'steps': 31517, 'loss/train': 1.9205714464187622} 02/25/2022 18:24:19 - INFO - codeparrot_training - Step 31518: {'lr': 0.00016166190214650923, 'samples': 16137728, 'steps': 31518, 'loss/train': 1.247193455696106} 02/25/2022 18:24:23 - INFO - codeparrot_training - Step 31519: {'lr': 0.0001616465954116086, 'samples': 16138240, 'steps': 31519, 'loss/train': 1.2244220972061157} 02/25/2022 18:24:28 - INFO - codeparrot_training - Step 31520: {'lr': 0.00016163128905518576, 'samples': 16138752, 'steps': 31520, 'loss/train': 1.8758896589279175} 02/25/2022 18:24:31 - INFO - codeparrot_training - Step 31521: {'lr': 0.0001616159830773064, 'samples': 16139264, 'steps': 31521, 'loss/train': 2.3563461303710938} 02/25/2022 18:24:37 - INFO - codeparrot_training - Step 31522: {'lr': 0.00016160067747803618, 'samples': 16139776, 'steps': 31522, 'loss/train': 2.1000547409057617} 02/25/2022 18:24:40 - INFO - codeparrot_training - Step 31523: {'lr': 0.00016158537225744064, 'samples': 16140288, 'steps': 31523, 'loss/train': 2.04306697845459} 02/25/2022 18:24:46 - INFO - codeparrot_training - Step 31524: {'lr': 0.0001615700674155852, 'samples': 16140800, 'steps': 31524, 'loss/train': 2.198699951171875} 02/25/2022 18:24:49 - INFO - codeparrot_training - Step 31525: {'lr': 0.00016155476295253552, 'samples': 16141312, 'steps': 31525, 'loss/train': 1.1795037984848022} 02/25/2022 18:24:56 - INFO - codeparrot_training - Step 31526: {'lr': 0.00016153945886835714, 'samples': 16141824, 'steps': 31526, 'loss/train': 1.513329029083252} 02/25/2022 18:24:59 - INFO - codeparrot_training - Step 31527: {'lr': 0.0001615241551631157, 'samples': 16142336, 'steps': 31527, 'loss/train': 1.8119617700576782} 02/25/2022 18:25:05 - INFO - codeparrot_training - Step 31528: {'lr': 0.00016150885183687663, 'samples': 16142848, 'steps': 31528, 'loss/train': 1.5955504179000854} 02/25/2022 18:25:10 - INFO - codeparrot_training - Step 31529: {'lr': 0.00016149354888970552, 'samples': 16143360, 'steps': 31529, 'loss/train': 1.851017951965332} 02/25/2022 18:25:14 - INFO - codeparrot_training - Step 31530: {'lr': 0.00016147824632166793, 'samples': 16143872, 'steps': 31530, 'loss/train': 2.9062321186065674} 02/25/2022 18:25:19 - INFO - codeparrot_training - Step 31531: {'lr': 0.00016146294413282945, 'samples': 16144384, 'steps': 31531, 'loss/train': 2.50898814201355} 02/25/2022 18:25:23 - INFO - codeparrot_training - Step 31532: {'lr': 0.0001614476423232557, 'samples': 16144896, 'steps': 31532, 'loss/train': 1.60396146774292} 02/25/2022 18:25:29 - INFO - codeparrot_training - Step 31533: {'lr': 0.00016143234089301202, 'samples': 16145408, 'steps': 31533, 'loss/train': 1.9290982484817505} 02/25/2022 18:25:32 - INFO - codeparrot_training - Step 31534: {'lr': 0.00016141703984216406, 'samples': 16145920, 'steps': 31534, 'loss/train': 2.207388162612915} 02/25/2022 18:25:36 - INFO - codeparrot_training - Step 31535: {'lr': 0.00016140173917077738, 'samples': 16146432, 'steps': 31535, 'loss/train': 2.1177854537963867} 02/25/2022 18:25:41 - INFO - codeparrot_training - Step 31536: {'lr': 0.00016138643887891763, 'samples': 16146944, 'steps': 31536, 'loss/train': 2.578043222427368} 02/25/2022 18:25:45 - INFO - codeparrot_training - Step 31537: {'lr': 0.00016137113896665011, 'samples': 16147456, 'steps': 31537, 'loss/train': 0.777461051940918} 02/25/2022 18:25:50 - INFO - codeparrot_training - Step 31538: {'lr': 0.00016135583943404052, 'samples': 16147968, 'steps': 31538, 'loss/train': 0.9611170291900635} 02/25/2022 18:25:54 - INFO - codeparrot_training - Step 31539: {'lr': 0.00016134054028115442, 'samples': 16148480, 'steps': 31539, 'loss/train': 1.5688283443450928} 02/25/2022 18:25:59 - INFO - codeparrot_training - Step 31540: {'lr': 0.00016132524150805726, 'samples': 16148992, 'steps': 31540, 'loss/train': 1.6839861869812012} 02/25/2022 18:26:03 - INFO - codeparrot_training - Step 31541: {'lr': 0.00016130994311481462, 'samples': 16149504, 'steps': 31541, 'loss/train': 1.8979129791259766} 02/25/2022 18:26:09 - INFO - codeparrot_training - Step 31542: {'lr': 0.00016129464510149195, 'samples': 16150016, 'steps': 31542, 'loss/train': 1.0789316892623901} 02/25/2022 18:26:13 - INFO - codeparrot_training - Step 31543: {'lr': 0.00016127934746815493, 'samples': 16150528, 'steps': 31543, 'loss/train': 1.0093352794647217} 02/25/2022 18:26:18 - INFO - codeparrot_training - Step 31544: {'lr': 0.00016126405021486896, 'samples': 16151040, 'steps': 31544, 'loss/train': 2.8517873287200928} 02/25/2022 18:26:22 - INFO - codeparrot_training - Step 31545: {'lr': 0.0001612487533416997, 'samples': 16151552, 'steps': 31545, 'loss/train': 1.5938973426818848} 02/25/2022 18:26:27 - INFO - codeparrot_training - Step 31546: {'lr': 0.00016123345684871255, 'samples': 16152064, 'steps': 31546, 'loss/train': 1.7922611236572266} 02/25/2022 18:26:31 - INFO - codeparrot_training - Step 31547: {'lr': 0.00016121816073597306, 'samples': 16152576, 'steps': 31547, 'loss/train': 1.693039059638977} 02/25/2022 18:26:36 - INFO - codeparrot_training - Step 31548: {'lr': 0.00016120286500354678, 'samples': 16153088, 'steps': 31548, 'loss/train': 2.9492082595825195} 02/25/2022 18:26:40 - INFO - codeparrot_training - Step 31549: {'lr': 0.00016118756965149928, 'samples': 16153600, 'steps': 31549, 'loss/train': 2.2102174758911133} 02/25/2022 18:26:45 - INFO - codeparrot_training - Step 31550: {'lr': 0.00016117227467989602, 'samples': 16154112, 'steps': 31550, 'loss/train': 1.0934679508209229} 02/25/2022 18:26:49 - INFO - codeparrot_training - Step 31551: {'lr': 0.00016115698008880251, 'samples': 16154624, 'steps': 31551, 'loss/train': 1.2507045269012451} 02/25/2022 18:26:54 - INFO - codeparrot_training - Step 31552: {'lr': 0.00016114168587828427, 'samples': 16155136, 'steps': 31552, 'loss/train': 1.506596326828003} 02/25/2022 18:26:58 - INFO - codeparrot_training - Step 31553: {'lr': 0.0001611263920484068, 'samples': 16155648, 'steps': 31553, 'loss/train': 2.0190672874450684} 02/25/2022 18:27:04 - INFO - codeparrot_training - Step 31554: {'lr': 0.0001611110985992358, 'samples': 16156160, 'steps': 31554, 'loss/train': 2.8666253089904785} 02/25/2022 18:27:07 - INFO - codeparrot_training - Step 31555: {'lr': 0.00016109580553083647, 'samples': 16156672, 'steps': 31555, 'loss/train': 1.9381645917892456} 02/25/2022 18:27:13 - INFO - codeparrot_training - Step 31556: {'lr': 0.00016108051284327452, 'samples': 16157184, 'steps': 31556, 'loss/train': 1.572195053100586} 02/25/2022 18:27:16 - INFO - codeparrot_training - Step 31557: {'lr': 0.0001610652205366154, 'samples': 16157696, 'steps': 31557, 'loss/train': 1.6516542434692383} 02/25/2022 18:27:22 - INFO - codeparrot_training - Step 31558: {'lr': 0.00016104992861092475, 'samples': 16158208, 'steps': 31558, 'loss/train': 1.7021163702011108} 02/25/2022 18:27:25 - INFO - codeparrot_training - Step 31559: {'lr': 0.00016103463706626783, 'samples': 16158720, 'steps': 31559, 'loss/train': 1.7618345022201538} 02/25/2022 18:27:31 - INFO - codeparrot_training - Step 31560: {'lr': 0.0001610193459027103, 'samples': 16159232, 'steps': 31560, 'loss/train': 2.249755382537842} 02/25/2022 18:27:37 - INFO - codeparrot_training - Step 31561: {'lr': 0.00016100405512031762, 'samples': 16159744, 'steps': 31561, 'loss/train': 2.2130277156829834} 02/25/2022 18:27:40 - INFO - codeparrot_training - Step 31562: {'lr': 0.00016098876471915547, 'samples': 16160256, 'steps': 31562, 'loss/train': 1.775108814239502} 02/25/2022 18:27:46 - INFO - codeparrot_training - Step 31563: {'lr': 0.00016097347469928904, 'samples': 16160768, 'steps': 31563, 'loss/train': 2.009779930114746} 02/25/2022 18:27:50 - INFO - codeparrot_training - Step 31564: {'lr': 0.00016095818506078397, 'samples': 16161280, 'steps': 31564, 'loss/train': 1.1456416845321655} 02/25/2022 18:27:55 - INFO - codeparrot_training - Step 31565: {'lr': 0.00016094289580370576, 'samples': 16161792, 'steps': 31565, 'loss/train': 3.4641997814178467} 02/25/2022 18:27:59 - INFO - codeparrot_training - Step 31566: {'lr': 0.00016092760692811993, 'samples': 16162304, 'steps': 31566, 'loss/train': 2.318700075149536} 02/25/2022 18:28:04 - INFO - codeparrot_training - Step 31567: {'lr': 0.000160912318434092, 'samples': 16162816, 'steps': 31567, 'loss/train': 2.204702854156494} 02/25/2022 18:28:08 - INFO - codeparrot_training - Step 31568: {'lr': 0.00016089703032168734, 'samples': 16163328, 'steps': 31568, 'loss/train': 1.3072080612182617} 02/25/2022 18:28:13 - INFO - codeparrot_training - Step 31569: {'lr': 0.00016088174259097152, 'samples': 16163840, 'steps': 31569, 'loss/train': 0.9428216218948364} 02/25/2022 18:28:17 - INFO - codeparrot_training - Step 31570: {'lr': 0.00016086645524200998, 'samples': 16164352, 'steps': 31570, 'loss/train': 0.3764359652996063} 02/25/2022 18:28:22 - INFO - codeparrot_training - Step 31571: {'lr': 0.0001608511682748684, 'samples': 16164864, 'steps': 31571, 'loss/train': 1.476344108581543} 02/25/2022 18:28:26 - INFO - codeparrot_training - Step 31572: {'lr': 0.00016083588168961196, 'samples': 16165376, 'steps': 31572, 'loss/train': 1.2195860147476196} 02/25/2022 18:28:32 - INFO - codeparrot_training - Step 31573: {'lr': 0.0001608205954863063, 'samples': 16165888, 'steps': 31573, 'loss/train': 3.0609970092773438} 02/25/2022 18:28:35 - INFO - codeparrot_training - Step 31574: {'lr': 0.0001608053096650169, 'samples': 16166400, 'steps': 31574, 'loss/train': 0.3347488045692444} 02/25/2022 18:28:41 - INFO - codeparrot_training - Step 31575: {'lr': 0.00016079002422580936, 'samples': 16166912, 'steps': 31575, 'loss/train': 1.8849388360977173} 02/25/2022 18:28:45 - INFO - codeparrot_training - Step 31576: {'lr': 0.0001607747391687489, 'samples': 16167424, 'steps': 31576, 'loss/train': 0.7627377510070801} 02/25/2022 18:28:50 - INFO - codeparrot_training - Step 31577: {'lr': 0.0001607594544939011, 'samples': 16167936, 'steps': 31577, 'loss/train': 1.2533965110778809} 02/25/2022 18:28:54 - INFO - codeparrot_training - Step 31578: {'lr': 0.00016074417020133148, 'samples': 16168448, 'steps': 31578, 'loss/train': 1.9179221391677856} 02/25/2022 18:28:59 - INFO - codeparrot_training - Step 31579: {'lr': 0.00016072888629110554, 'samples': 16168960, 'steps': 31579, 'loss/train': 1.9761630296707153} 02/25/2022 18:29:03 - INFO - codeparrot_training - Step 31580: {'lr': 0.00016071360276328874, 'samples': 16169472, 'steps': 31580, 'loss/train': 1.5991648435592651} 02/25/2022 18:29:08 - INFO - codeparrot_training - Step 31581: {'lr': 0.0001606983196179464, 'samples': 16169984, 'steps': 31581, 'loss/train': 1.6161880493164062} 02/25/2022 18:29:11 - INFO - codeparrot_training - Step 31582: {'lr': 0.00016068303685514413, 'samples': 16170496, 'steps': 31582, 'loss/train': 1.508562445640564} 02/25/2022 18:29:17 - INFO - codeparrot_training - Step 31583: {'lr': 0.00016066775447494744, 'samples': 16171008, 'steps': 31583, 'loss/train': 1.707594394683838} 02/25/2022 18:29:20 - INFO - codeparrot_training - Step 31584: {'lr': 0.0001606524724774217, 'samples': 16171520, 'steps': 31584, 'loss/train': 2.6192173957824707} 02/25/2022 18:29:26 - INFO - codeparrot_training - Step 31585: {'lr': 0.00016063719086263235, 'samples': 16172032, 'steps': 31585, 'loss/train': 1.3823825120925903} 02/25/2022 18:29:29 - INFO - codeparrot_training - Step 31586: {'lr': 0.00016062190963064496, 'samples': 16172544, 'steps': 31586, 'loss/train': 1.595389485359192} 02/25/2022 18:29:35 - INFO - codeparrot_training - Step 31587: {'lr': 0.00016060662878152488, 'samples': 16173056, 'steps': 31587, 'loss/train': 1.7498188018798828} 02/25/2022 18:29:38 - INFO - codeparrot_training - Step 31588: {'lr': 0.00016059134831533767, 'samples': 16173568, 'steps': 31588, 'loss/train': 1.6914992332458496} 02/25/2022 18:29:45 - INFO - codeparrot_training - Step 31589: {'lr': 0.00016057606823214867, 'samples': 16174080, 'steps': 31589, 'loss/train': 1.4717464447021484} 02/25/2022 18:29:49 - INFO - codeparrot_training - Step 31590: {'lr': 0.00016056078853202344, 'samples': 16174592, 'steps': 31590, 'loss/train': 1.5320957899093628} 02/25/2022 18:29:54 - INFO - codeparrot_training - Step 31591: {'lr': 0.00016054550921502735, 'samples': 16175104, 'steps': 31591, 'loss/train': 1.8790576457977295} 02/25/2022 18:29:57 - INFO - codeparrot_training - Step 31592: {'lr': 0.00016053023028122587, 'samples': 16175616, 'steps': 31592, 'loss/train': 1.9117008447647095} 02/25/2022 18:30:03 - INFO - codeparrot_training - Step 31593: {'lr': 0.00016051495173068458, 'samples': 16176128, 'steps': 31593, 'loss/train': 2.3490333557128906} 02/25/2022 18:30:07 - INFO - codeparrot_training - Step 31594: {'lr': 0.0001604996735634688, 'samples': 16176640, 'steps': 31594, 'loss/train': 1.184212565422058} 02/25/2022 18:30:12 - INFO - codeparrot_training - Step 31595: {'lr': 0.0001604843957796439, 'samples': 16177152, 'steps': 31595, 'loss/train': 2.103588104248047} 02/25/2022 18:30:16 - INFO - codeparrot_training - Step 31596: {'lr': 0.0001604691183792755, 'samples': 16177664, 'steps': 31596, 'loss/train': 1.6357944011688232} 02/25/2022 18:30:21 - INFO - codeparrot_training - Step 31597: {'lr': 0.00016045384136242902, 'samples': 16178176, 'steps': 31597, 'loss/train': 1.512934923171997} 02/25/2022 18:30:25 - INFO - codeparrot_training - Step 31598: {'lr': 0.00016043856472916976, 'samples': 16178688, 'steps': 31598, 'loss/train': 1.5081853866577148} 02/25/2022 18:30:32 - INFO - codeparrot_training - Step 31599: {'lr': 0.00016042328847956324, 'samples': 16179200, 'steps': 31599, 'loss/train': 2.2163279056549072} 02/25/2022 18:30:35 - INFO - codeparrot_training - Step 31600: {'lr': 0.00016040801261367493, 'samples': 16179712, 'steps': 31600, 'loss/train': 0.700271725654602} 02/25/2022 18:30:41 - INFO - codeparrot_training - Step 31601: {'lr': 0.00016039273713157023, 'samples': 16180224, 'steps': 31601, 'loss/train': 1.7624937295913696} 02/25/2022 18:30:44 - INFO - codeparrot_training - Step 31602: {'lr': 0.00016037746203331472, 'samples': 16180736, 'steps': 31602, 'loss/train': 0.7927564978599548} 02/25/2022 18:30:50 - INFO - codeparrot_training - Step 31603: {'lr': 0.00016036218731897356, 'samples': 16181248, 'steps': 31603, 'loss/train': 2.5134196281433105} 02/25/2022 18:30:53 - INFO - codeparrot_training - Step 31604: {'lr': 0.00016034691298861238, 'samples': 16181760, 'steps': 31604, 'loss/train': 1.9081127643585205} 02/25/2022 18:30:58 - INFO - codeparrot_training - Step 31605: {'lr': 0.00016033163904229648, 'samples': 16182272, 'steps': 31605, 'loss/train': 2.1580874919891357} 02/25/2022 18:31:02 - INFO - codeparrot_training - Step 31606: {'lr': 0.00016031636548009153, 'samples': 16182784, 'steps': 31606, 'loss/train': 1.7244893312454224} 02/25/2022 18:31:08 - INFO - codeparrot_training - Step 31607: {'lr': 0.00016030109230206265, 'samples': 16183296, 'steps': 31607, 'loss/train': 1.2768585681915283} 02/25/2022 18:31:11 - INFO - codeparrot_training - Step 31608: {'lr': 0.00016028581950827542, 'samples': 16183808, 'steps': 31608, 'loss/train': 0.056688010692596436} 02/25/2022 18:31:17 - INFO - codeparrot_training - Step 31609: {'lr': 0.00016027054709879529, 'samples': 16184320, 'steps': 31609, 'loss/train': 1.9658455848693848} 02/25/2022 18:31:21 - INFO - codeparrot_training - Step 31610: {'lr': 0.0001602552750736877, 'samples': 16184832, 'steps': 31610, 'loss/train': 1.5458626747131348} 02/25/2022 18:31:26 - INFO - codeparrot_training - Step 31611: {'lr': 0.00016024000343301792, 'samples': 16185344, 'steps': 31611, 'loss/train': 1.6033605337142944} 02/25/2022 18:31:30 - INFO - codeparrot_training - Step 31612: {'lr': 0.0001602247321768515, 'samples': 16185856, 'steps': 31612, 'loss/train': 1.0204944610595703} 02/25/2022 18:31:35 - INFO - codeparrot_training - Step 31613: {'lr': 0.00016020946130525376, 'samples': 16186368, 'steps': 31613, 'loss/train': 2.261725425720215} 02/25/2022 18:31:39 - INFO - codeparrot_training - Step 31614: {'lr': 0.00016019419081829024, 'samples': 16186880, 'steps': 31614, 'loss/train': 1.6130387783050537} 02/25/2022 18:31:44 - INFO - codeparrot_training - Step 31615: {'lr': 0.00016017892071602637, 'samples': 16187392, 'steps': 31615, 'loss/train': 1.3018862009048462} 02/25/2022 18:31:48 - INFO - codeparrot_training - Step 31616: {'lr': 0.00016016365099852736, 'samples': 16187904, 'steps': 31616, 'loss/train': 1.9669933319091797} 02/25/2022 18:31:53 - INFO - codeparrot_training - Step 31617: {'lr': 0.00016014838166585873, 'samples': 16188416, 'steps': 31617, 'loss/train': 1.4091639518737793} 02/25/2022 18:31:57 - INFO - codeparrot_training - Step 31618: {'lr': 0.00016013311271808594, 'samples': 16188928, 'steps': 31618, 'loss/train': 1.8106836080551147} 02/25/2022 18:32:03 - INFO - codeparrot_training - Step 31619: {'lr': 0.00016011784415527445, 'samples': 16189440, 'steps': 31619, 'loss/train': 1.875355839729309} 02/25/2022 18:32:06 - INFO - codeparrot_training - Step 31620: {'lr': 0.00016010257597748944, 'samples': 16189952, 'steps': 31620, 'loss/train': 1.0046404600143433} 02/25/2022 18:32:12 - INFO - codeparrot_training - Step 31621: {'lr': 0.0001600873081847965, 'samples': 16190464, 'steps': 31621, 'loss/train': 1.0814626216888428} 02/25/2022 18:32:15 - INFO - codeparrot_training - Step 31622: {'lr': 0.00016007204077726094, 'samples': 16190976, 'steps': 31622, 'loss/train': 2.0401828289031982} 02/25/2022 18:32:21 - INFO - codeparrot_training - Step 31623: {'lr': 0.00016005677375494835, 'samples': 16191488, 'steps': 31623, 'loss/train': 1.9461112022399902} 02/25/2022 18:32:24 - INFO - codeparrot_training - Step 31624: {'lr': 0.00016004150711792383, 'samples': 16192000, 'steps': 31624, 'loss/train': 0.0671166405081749} 02/25/2022 18:32:30 - INFO - codeparrot_training - Step 31625: {'lr': 0.00016002624086625296, 'samples': 16192512, 'steps': 31625, 'loss/train': 2.3351097106933594} 02/25/2022 18:32:33 - INFO - codeparrot_training - Step 31626: {'lr': 0.00016001097500000113, 'samples': 16193024, 'steps': 31626, 'loss/train': 1.3996005058288574} 02/25/2022 18:32:39 - INFO - codeparrot_training - Step 31627: {'lr': 0.0001599957095192337, 'samples': 16193536, 'steps': 31627, 'loss/train': 2.0448317527770996} 02/25/2022 18:32:42 - INFO - codeparrot_training - Step 31628: {'lr': 0.0001599804444240161, 'samples': 16194048, 'steps': 31628, 'loss/train': 2.384078025817871} 02/25/2022 18:32:48 - INFO - codeparrot_training - Step 31629: {'lr': 0.00015996517971441363, 'samples': 16194560, 'steps': 31629, 'loss/train': 1.5013035535812378} 02/25/2022 18:32:51 - INFO - codeparrot_training - Step 31630: {'lr': 0.0001599499153904918, 'samples': 16195072, 'steps': 31630, 'loss/train': 1.3970874547958374} 02/25/2022 18:32:57 - INFO - codeparrot_training - Step 31631: {'lr': 0.0001599346514523159, 'samples': 16195584, 'steps': 31631, 'loss/train': 0.038425154983997345} 02/25/2022 18:33:00 - INFO - codeparrot_training - Step 31632: {'lr': 0.00015991938789995137, 'samples': 16196096, 'steps': 31632, 'loss/train': 2.1713266372680664} 02/25/2022 18:33:06 - INFO - codeparrot_training - Step 31633: {'lr': 0.00015990412473346354, 'samples': 16196608, 'steps': 31633, 'loss/train': 1.9602012634277344} 02/25/2022 18:33:09 - INFO - codeparrot_training - Step 31634: {'lr': 0.0001598888619529179, 'samples': 16197120, 'steps': 31634, 'loss/train': 0.5772212147712708} 02/25/2022 18:33:15 - INFO - codeparrot_training - Step 31635: {'lr': 0.00015987359955837972, 'samples': 16197632, 'steps': 31635, 'loss/train': 1.5683538913726807} 02/25/2022 18:33:19 - INFO - codeparrot_training - Step 31636: {'lr': 0.00015985833754991442, 'samples': 16198144, 'steps': 31636, 'loss/train': 0.8712681531906128} 02/25/2022 18:33:24 - INFO - codeparrot_training - Step 31637: {'lr': 0.0001598430759275874, 'samples': 16198656, 'steps': 31637, 'loss/train': 1.6394400596618652} 02/25/2022 18:33:30 - INFO - codeparrot_training - Step 31638: {'lr': 0.000159827814691464, 'samples': 16199168, 'steps': 31638, 'loss/train': 2.8779196739196777} 02/25/2022 18:33:33 - INFO - codeparrot_training - Step 31639: {'lr': 0.0001598125538416096, 'samples': 16199680, 'steps': 31639, 'loss/train': 0.8959330320358276} 02/25/2022 18:33:39 - INFO - codeparrot_training - Step 31640: {'lr': 0.00015979729337808955, 'samples': 16200192, 'steps': 31640, 'loss/train': 1.1992782354354858} 02/25/2022 18:33:43 - INFO - codeparrot_training - Step 31641: {'lr': 0.00015978203330096935, 'samples': 16200704, 'steps': 31641, 'loss/train': 2.318474769592285} 02/25/2022 18:33:46 - INFO - codeparrot_training - Step 31642: {'lr': 0.00015976677361031422, 'samples': 16201216, 'steps': 31642, 'loss/train': 1.9080734252929688} 02/25/2022 18:33:51 - INFO - codeparrot_training - Step 31643: {'lr': 0.00015975151430618954, 'samples': 16201728, 'steps': 31643, 'loss/train': 1.6840846538543701} 02/25/2022 18:33:58 - INFO - codeparrot_training - Step 31644: {'lr': 0.00015973625538866072, 'samples': 16202240, 'steps': 31644, 'loss/train': 2.330674886703491} 02/25/2022 18:34:01 - INFO - codeparrot_training - Step 31645: {'lr': 0.00015972099685779322, 'samples': 16202752, 'steps': 31645, 'loss/train': 1.6319199800491333} 02/25/2022 18:34:07 - INFO - codeparrot_training - Step 31646: {'lr': 0.0001597057387136522, 'samples': 16203264, 'steps': 31646, 'loss/train': 1.3397094011306763} 02/25/2022 18:34:10 - INFO - codeparrot_training - Step 31647: {'lr': 0.00015969048095630314, 'samples': 16203776, 'steps': 31647, 'loss/train': 1.62205171585083} 02/25/2022 18:34:16 - INFO - codeparrot_training - Step 31648: {'lr': 0.00015967522358581137, 'samples': 16204288, 'steps': 31648, 'loss/train': 0.8132610321044922} 02/25/2022 18:34:19 - INFO - codeparrot_training - Step 31649: {'lr': 0.00015965996660224235, 'samples': 16204800, 'steps': 31649, 'loss/train': 1.024165153503418} 02/25/2022 18:34:25 - INFO - codeparrot_training - Step 31650: {'lr': 0.00015964471000566127, 'samples': 16205312, 'steps': 31650, 'loss/train': 2.103924036026001} 02/25/2022 18:34:28 - INFO - codeparrot_training - Step 31651: {'lr': 0.0001596294537961336, 'samples': 16205824, 'steps': 31651, 'loss/train': 2.021979808807373} 02/25/2022 18:34:34 - INFO - codeparrot_training - Step 31652: {'lr': 0.00015961419797372455, 'samples': 16206336, 'steps': 31652, 'loss/train': 0.940769374370575} 02/25/2022 18:34:37 - INFO - codeparrot_training - Step 31653: {'lr': 0.00015959894253849967, 'samples': 16206848, 'steps': 31653, 'loss/train': 2.1182608604431152} 02/25/2022 18:34:43 - INFO - codeparrot_training - Step 31654: {'lr': 0.00015958368749052426, 'samples': 16207360, 'steps': 31654, 'loss/train': 1.481539011001587} 02/25/2022 18:34:47 - INFO - codeparrot_training - Step 31655: {'lr': 0.00015956843282986354, 'samples': 16207872, 'steps': 31655, 'loss/train': 2.612964391708374} 02/25/2022 18:34:52 - INFO - codeparrot_training - Step 31656: {'lr': 0.00015955317855658298, 'samples': 16208384, 'steps': 31656, 'loss/train': 1.4114587306976318} 02/25/2022 18:34:56 - INFO - codeparrot_training - Step 31657: {'lr': 0.00015953792467074778, 'samples': 16208896, 'steps': 31657, 'loss/train': 1.0481922626495361} 02/25/2022 18:35:02 - INFO - codeparrot_training - Step 31658: {'lr': 0.0001595226711724236, 'samples': 16209408, 'steps': 31658, 'loss/train': 1.016542911529541} 02/25/2022 18:35:06 - INFO - codeparrot_training - Step 31659: {'lr': 0.0001595074180616754, 'samples': 16209920, 'steps': 31659, 'loss/train': 2.3109352588653564} 02/25/2022 18:35:09 - INFO - codeparrot_training - Step 31660: {'lr': 0.0001594921653385687, 'samples': 16210432, 'steps': 31660, 'loss/train': 2.574134111404419} 02/25/2022 18:35:14 - INFO - codeparrot_training - Step 31661: {'lr': 0.0001594769130031688, 'samples': 16210944, 'steps': 31661, 'loss/train': 2.1193525791168213} 02/25/2022 18:35:18 - INFO - codeparrot_training - Step 31662: {'lr': 0.0001594616610555411, 'samples': 16211456, 'steps': 31662, 'loss/train': 2.043381452560425} 02/25/2022 18:35:24 - INFO - codeparrot_training - Step 31663: {'lr': 0.000159446409495751, 'samples': 16211968, 'steps': 31663, 'loss/train': 1.0986430644989014} 02/25/2022 18:35:27 - INFO - codeparrot_training - Step 31664: {'lr': 0.0001594311583238636, 'samples': 16212480, 'steps': 31664, 'loss/train': 1.4999703168869019} 02/25/2022 18:35:33 - INFO - codeparrot_training - Step 31665: {'lr': 0.00015941590753994434, 'samples': 16212992, 'steps': 31665, 'loss/train': 8.629182815551758} 02/25/2022 18:35:36 - INFO - codeparrot_training - Step 31666: {'lr': 0.0001594006571440586, 'samples': 16213504, 'steps': 31666, 'loss/train': 1.6565648317337036} 02/25/2022 18:35:42 - INFO - codeparrot_training - Step 31667: {'lr': 0.00015938540713627178, 'samples': 16214016, 'steps': 31667, 'loss/train': 1.8821903467178345} 02/25/2022 18:35:46 - INFO - codeparrot_training - Step 31668: {'lr': 0.00015937015751664902, 'samples': 16214528, 'steps': 31668, 'loss/train': 1.2966912984848022} 02/25/2022 18:35:51 - INFO - codeparrot_training - Step 31669: {'lr': 0.00015935490828525566, 'samples': 16215040, 'steps': 31669, 'loss/train': 2.6232807636260986} 02/25/2022 18:35:55 - INFO - codeparrot_training - Step 31670: {'lr': 0.0001593396594421572, 'samples': 16215552, 'steps': 31670, 'loss/train': 3.301279067993164} 02/25/2022 18:36:00 - INFO - codeparrot_training - Step 31671: {'lr': 0.00015932441098741884, 'samples': 16216064, 'steps': 31671, 'loss/train': 2.928297281265259} 02/25/2022 18:36:04 - INFO - codeparrot_training - Step 31672: {'lr': 0.00015930916292110583, 'samples': 16216576, 'steps': 31672, 'loss/train': 1.9819281101226807} 02/25/2022 18:36:09 - INFO - codeparrot_training - Step 31673: {'lr': 0.0001592939152432836, 'samples': 16217088, 'steps': 31673, 'loss/train': 1.8660078048706055} 02/25/2022 18:36:13 - INFO - codeparrot_training - Step 31674: {'lr': 0.00015927866795401746, 'samples': 16217600, 'steps': 31674, 'loss/train': 9.75402545928955} 02/25/2022 18:36:18 - INFO - codeparrot_training - Step 31675: {'lr': 0.00015926342105337268, 'samples': 16218112, 'steps': 31675, 'loss/train': 1.33323073387146} 02/25/2022 18:36:22 - INFO - codeparrot_training - Step 31676: {'lr': 0.00015924817454141462, 'samples': 16218624, 'steps': 31676, 'loss/train': 2.3078315258026123} 02/25/2022 18:36:27 - INFO - codeparrot_training - Step 31677: {'lr': 0.0001592329284182086, 'samples': 16219136, 'steps': 31677, 'loss/train': 1.7296061515808105} 02/25/2022 18:36:31 - INFO - codeparrot_training - Step 31678: {'lr': 0.0001592176826838198, 'samples': 16219648, 'steps': 31678, 'loss/train': 1.7072107791900635} 02/25/2022 18:36:36 - INFO - codeparrot_training - Step 31679: {'lr': 0.00015920243733831362, 'samples': 16220160, 'steps': 31679, 'loss/train': 0.7049659490585327} 02/25/2022 18:36:40 - INFO - codeparrot_training - Step 31680: {'lr': 0.00015918719238175544, 'samples': 16220672, 'steps': 31680, 'loss/train': 2.18479061126709} 02/25/2022 18:36:46 - INFO - codeparrot_training - Step 31681: {'lr': 0.0001591719478142105, 'samples': 16221184, 'steps': 31681, 'loss/train': 1.0697076320648193} 02/25/2022 18:36:49 - INFO - codeparrot_training - Step 31682: {'lr': 0.00015915670363574402, 'samples': 16221696, 'steps': 31682, 'loss/train': 1.19609534740448} 02/25/2022 18:36:55 - INFO - codeparrot_training - Step 31683: {'lr': 0.00015914145984642137, 'samples': 16222208, 'steps': 31683, 'loss/train': 1.8251248598098755} 02/25/2022 18:36:58 - INFO - codeparrot_training - Step 31684: {'lr': 0.00015912621644630797, 'samples': 16222720, 'steps': 31684, 'loss/train': 1.3311402797698975} 02/25/2022 18:37:04 - INFO - codeparrot_training - Step 31685: {'lr': 0.0001591109734354689, 'samples': 16223232, 'steps': 31685, 'loss/train': 0.0562140978872776} 02/25/2022 18:37:07 - INFO - codeparrot_training - Step 31686: {'lr': 0.00015909573081396956, 'samples': 16223744, 'steps': 31686, 'loss/train': 0.07935921847820282} 02/25/2022 18:37:13 - INFO - codeparrot_training - Step 31687: {'lr': 0.00015908048858187524, 'samples': 16224256, 'steps': 31687, 'loss/train': 1.1900615692138672} 02/25/2022 18:37:17 - INFO - codeparrot_training - Step 31688: {'lr': 0.00015906524673925125, 'samples': 16224768, 'steps': 31688, 'loss/train': 1.9659942388534546} 02/25/2022 18:37:22 - INFO - codeparrot_training - Step 31689: {'lr': 0.00015905000528616296, 'samples': 16225280, 'steps': 31689, 'loss/train': 1.279463768005371} 02/25/2022 18:37:26 - INFO - codeparrot_training - Step 31690: {'lr': 0.00015903476422267547, 'samples': 16225792, 'steps': 31690, 'loss/train': 2.544485569000244} 02/25/2022 18:37:31 - INFO - codeparrot_training - Step 31691: {'lr': 0.00015901952354885413, 'samples': 16226304, 'steps': 31691, 'loss/train': 0.4169861674308777} 02/25/2022 18:37:35 - INFO - codeparrot_training - Step 31692: {'lr': 0.0001590042832647643, 'samples': 16226816, 'steps': 31692, 'loss/train': 1.5177192687988281} 02/25/2022 18:37:41 - INFO - codeparrot_training - Step 31693: {'lr': 0.0001589890433704713, 'samples': 16227328, 'steps': 31693, 'loss/train': 2.1188693046569824} 02/25/2022 18:37:45 - INFO - codeparrot_training - Step 31694: {'lr': 0.00015897380386604027, 'samples': 16227840, 'steps': 31694, 'loss/train': 2.1663472652435303} 02/25/2022 18:37:51 - INFO - codeparrot_training - Step 31695: {'lr': 0.00015895856475153658, 'samples': 16228352, 'steps': 31695, 'loss/train': 1.648962140083313} 02/25/2022 18:37:54 - INFO - codeparrot_training - Step 31696: {'lr': 0.00015894332602702543, 'samples': 16228864, 'steps': 31696, 'loss/train': 1.9686440229415894} 02/25/2022 18:38:00 - INFO - codeparrot_training - Step 31697: {'lr': 0.0001589280876925723, 'samples': 16229376, 'steps': 31697, 'loss/train': 1.9946438074111938} 02/25/2022 18:38:03 - INFO - codeparrot_training - Step 31698: {'lr': 0.00015891284974824222, 'samples': 16229888, 'steps': 31698, 'loss/train': 1.570151686668396} 02/25/2022 18:38:09 - INFO - codeparrot_training - Step 31699: {'lr': 0.00015889761219410053, 'samples': 16230400, 'steps': 31699, 'loss/train': 2.239841938018799} 02/25/2022 18:38:12 - INFO - codeparrot_training - Step 31700: {'lr': 0.0001588823750302126, 'samples': 16230912, 'steps': 31700, 'loss/train': 1.6195874214172363} 02/25/2022 18:38:17 - INFO - codeparrot_training - Step 31701: {'lr': 0.00015886713825664358, 'samples': 16231424, 'steps': 31701, 'loss/train': 2.39121413230896} 02/25/2022 18:38:21 - INFO - codeparrot_training - Step 31702: {'lr': 0.00015885190187345894, 'samples': 16231936, 'steps': 31702, 'loss/train': 1.454297423362732} 02/25/2022 18:38:27 - INFO - codeparrot_training - Step 31703: {'lr': 0.0001588366658807237, 'samples': 16232448, 'steps': 31703, 'loss/train': 1.4006729125976562} 02/25/2022 18:38:31 - INFO - codeparrot_training - Step 31704: {'lr': 0.00015882143027850325, 'samples': 16232960, 'steps': 31704, 'loss/train': 1.6190322637557983} 02/25/2022 18:38:36 - INFO - codeparrot_training - Step 31705: {'lr': 0.00015880619506686283, 'samples': 16233472, 'steps': 31705, 'loss/train': 0.7933560609817505} 02/25/2022 18:38:40 - INFO - codeparrot_training - Step 31706: {'lr': 0.0001587909602458678, 'samples': 16233984, 'steps': 31706, 'loss/train': 8.767669677734375} 02/25/2022 18:38:45 - INFO - codeparrot_training - Step 31707: {'lr': 0.00015877572581558325, 'samples': 16234496, 'steps': 31707, 'loss/train': 1.5217331647872925} 02/25/2022 18:38:49 - INFO - codeparrot_training - Step 31708: {'lr': 0.0001587604917760745, 'samples': 16235008, 'steps': 31708, 'loss/train': 2.2403440475463867} 02/25/2022 18:38:54 - INFO - codeparrot_training - Step 31709: {'lr': 0.00015874525812740686, 'samples': 16235520, 'steps': 31709, 'loss/train': 0.8287572860717773} 02/25/2022 18:38:58 - INFO - codeparrot_training - Step 31710: {'lr': 0.00015873002486964562, 'samples': 16236032, 'steps': 31710, 'loss/train': 2.5754573345184326} 02/25/2022 18:39:03 - INFO - codeparrot_training - Step 31711: {'lr': 0.00015871479200285594, 'samples': 16236544, 'steps': 31711, 'loss/train': 0.15411376953125} 02/25/2022 18:39:07 - INFO - codeparrot_training - Step 31712: {'lr': 0.00015869955952710308, 'samples': 16237056, 'steps': 31712, 'loss/train': 1.8623647689819336} 02/25/2022 18:39:13 - INFO - codeparrot_training - Step 31713: {'lr': 0.00015868432744245225, 'samples': 16237568, 'steps': 31713, 'loss/train': 0.5052101016044617} 02/25/2022 18:39:16 - INFO - codeparrot_training - Step 31714: {'lr': 0.00015866909574896887, 'samples': 16238080, 'steps': 31714, 'loss/train': 1.67640221118927} 02/25/2022 18:39:22 - INFO - codeparrot_training - Step 31715: {'lr': 0.0001586538644467181, 'samples': 16238592, 'steps': 31715, 'loss/train': 1.3188472986221313} 02/25/2022 18:39:25 - INFO - codeparrot_training - Step 31716: {'lr': 0.00015863863353576512, 'samples': 16239104, 'steps': 31716, 'loss/train': 1.0435198545455933} 02/25/2022 18:39:31 - INFO - codeparrot_training - Step 31717: {'lr': 0.00015862340301617522, 'samples': 16239616, 'steps': 31717, 'loss/train': 1.3111354112625122} 02/25/2022 18:39:34 - INFO - codeparrot_training - Step 31718: {'lr': 0.00015860817288801364, 'samples': 16240128, 'steps': 31718, 'loss/train': 1.2448370456695557} 02/25/2022 18:39:40 - INFO - codeparrot_training - Step 31719: {'lr': 0.00015859294315134564, 'samples': 16240640, 'steps': 31719, 'loss/train': 2.3127501010894775} 02/25/2022 18:39:43 - INFO - codeparrot_training - Step 31720: {'lr': 0.00015857771380623642, 'samples': 16241152, 'steps': 31720, 'loss/train': 1.96444571018219} 02/25/2022 18:39:49 - INFO - codeparrot_training - Step 31721: {'lr': 0.00015856248485275132, 'samples': 16241664, 'steps': 31721, 'loss/train': 1.3329789638519287} 02/25/2022 18:39:52 - INFO - codeparrot_training - Step 31722: {'lr': 0.00015854725629095543, 'samples': 16242176, 'steps': 31722, 'loss/train': 1.278171420097351} 02/25/2022 18:39:59 - INFO - codeparrot_training - Step 31723: {'lr': 0.00015853202812091406, 'samples': 16242688, 'steps': 31723, 'loss/train': 0.5689896941184998} 02/25/2022 18:40:02 - INFO - codeparrot_training - Step 31724: {'lr': 0.0001585168003426925, 'samples': 16243200, 'steps': 31724, 'loss/train': 2.6579689979553223} 02/25/2022 18:40:08 - INFO - codeparrot_training - Step 31725: {'lr': 0.00015850157295635585, 'samples': 16243712, 'steps': 31725, 'loss/train': 1.4840666055679321} 02/25/2022 18:40:11 - INFO - codeparrot_training - Step 31726: {'lr': 0.0001584863459619694, 'samples': 16244224, 'steps': 31726, 'loss/train': 1.70193612575531} 02/25/2022 18:40:17 - INFO - codeparrot_training - Step 31727: {'lr': 0.0001584711193595984, 'samples': 16244736, 'steps': 31727, 'loss/train': 2.2464041709899902} 02/25/2022 18:40:20 - INFO - codeparrot_training - Step 31728: {'lr': 0.00015845589314930815, 'samples': 16245248, 'steps': 31728, 'loss/train': 3.001596450805664} 02/25/2022 18:40:25 - INFO - codeparrot_training - Step 31729: {'lr': 0.0001584406673311637, 'samples': 16245760, 'steps': 31729, 'loss/train': 2.058844804763794} 02/25/2022 18:40:29 - INFO - codeparrot_training - Step 31730: {'lr': 0.0001584254419052303, 'samples': 16246272, 'steps': 31730, 'loss/train': 1.9087986946105957} 02/25/2022 18:40:34 - INFO - codeparrot_training - Step 31731: {'lr': 0.0001584102168715733, 'samples': 16246784, 'steps': 31731, 'loss/train': 2.9825191497802734} 02/25/2022 18:40:38 - INFO - codeparrot_training - Step 31732: {'lr': 0.0001583949922302579, 'samples': 16247296, 'steps': 31732, 'loss/train': 1.3170325756072998} 02/25/2022 18:40:44 - INFO - codeparrot_training - Step 31733: {'lr': 0.00015837976798134916, 'samples': 16247808, 'steps': 31733, 'loss/train': 2.037655830383301} 02/25/2022 18:40:47 - INFO - codeparrot_training - Step 31734: {'lr': 0.00015836454412491248, 'samples': 16248320, 'steps': 31734, 'loss/train': 1.617730975151062} 02/25/2022 18:40:53 - INFO - codeparrot_training - Step 31735: {'lr': 0.00015834932066101292, 'samples': 16248832, 'steps': 31735, 'loss/train': 1.6753456592559814} 02/25/2022 18:40:56 - INFO - codeparrot_training - Step 31736: {'lr': 0.0001583340975897158, 'samples': 16249344, 'steps': 31736, 'loss/train': 2.129549503326416} 02/25/2022 18:41:02 - INFO - codeparrot_training - Step 31737: {'lr': 0.00015831887491108638, 'samples': 16249856, 'steps': 31737, 'loss/train': 1.9992746114730835} 02/25/2022 18:41:05 - INFO - codeparrot_training - Step 31738: {'lr': 0.0001583036526251897, 'samples': 16250368, 'steps': 31738, 'loss/train': 0.10700304806232452} 02/25/2022 18:41:11 - INFO - codeparrot_training - Step 31739: {'lr': 0.00015828843073209103, 'samples': 16250880, 'steps': 31739, 'loss/train': 2.000544786453247} 02/25/2022 18:41:14 - INFO - codeparrot_training - Step 31740: {'lr': 0.00015827320923185562, 'samples': 16251392, 'steps': 31740, 'loss/train': 1.155184030532837} 02/25/2022 18:41:20 - INFO - codeparrot_training - Step 31741: {'lr': 0.0001582579881245488, 'samples': 16251904, 'steps': 31741, 'loss/train': 1.5773965120315552} 02/25/2022 18:41:23 - INFO - codeparrot_training - Step 31742: {'lr': 0.00015824276741023544, 'samples': 16252416, 'steps': 31742, 'loss/train': 1.9481098651885986} 02/25/2022 18:41:29 - INFO - codeparrot_training - Step 31743: {'lr': 0.00015822754708898096, 'samples': 16252928, 'steps': 31743, 'loss/train': 1.9054780006408691} 02/25/2022 18:41:33 - INFO - codeparrot_training - Step 31744: {'lr': 0.00015821232716085054, 'samples': 16253440, 'steps': 31744, 'loss/train': 1.865513801574707} 02/25/2022 18:41:38 - INFO - codeparrot_training - Step 31745: {'lr': 0.00015819710762590946, 'samples': 16253952, 'steps': 31745, 'loss/train': 1.294236183166504} 02/25/2022 18:41:41 - INFO - codeparrot_training - Step 31746: {'lr': 0.00015818188848422274, 'samples': 16254464, 'steps': 31746, 'loss/train': 1.5553377866744995} 02/25/2022 18:41:47 - INFO - codeparrot_training - Step 31747: {'lr': 0.00015816666973585564, 'samples': 16254976, 'steps': 31747, 'loss/train': 2.326873779296875} 02/25/2022 18:41:51 - INFO - codeparrot_training - Step 31748: {'lr': 0.00015815145138087336, 'samples': 16255488, 'steps': 31748, 'loss/train': 1.71562922000885} 02/25/2022 18:41:57 - INFO - codeparrot_training - Step 31749: {'lr': 0.0001581362334193411, 'samples': 16256000, 'steps': 31749, 'loss/train': 2.299311637878418} 02/25/2022 18:42:00 - INFO - codeparrot_training - Step 31750: {'lr': 0.00015812101585132416, 'samples': 16256512, 'steps': 31750, 'loss/train': 1.9052296876907349} 02/25/2022 18:42:06 - INFO - codeparrot_training - Step 31751: {'lr': 0.00015810579867688748, 'samples': 16257024, 'steps': 31751, 'loss/train': 0.9062367081642151} 02/25/2022 18:42:11 - INFO - codeparrot_training - Step 31752: {'lr': 0.00015809058189609644, 'samples': 16257536, 'steps': 31752, 'loss/train': 0.9063776731491089} 02/25/2022 18:42:15 - INFO - codeparrot_training - Step 31753: {'lr': 0.00015807536550901607, 'samples': 16258048, 'steps': 31753, 'loss/train': 0.7376357913017273} 02/25/2022 18:42:20 - INFO - codeparrot_training - Step 31754: {'lr': 0.0001580601495157118, 'samples': 16258560, 'steps': 31754, 'loss/train': 1.2490119934082031} 02/25/2022 18:42:24 - INFO - codeparrot_training - Step 31755: {'lr': 0.00015804493391624857, 'samples': 16259072, 'steps': 31755, 'loss/train': 2.912867307662964} 02/25/2022 18:42:27 - INFO - codeparrot_training - Step 31756: {'lr': 0.00015802971871069164, 'samples': 16259584, 'steps': 31756, 'loss/train': 2.0779659748077393} 02/25/2022 18:42:33 - INFO - codeparrot_training - Step 31757: {'lr': 0.00015801450389910622, 'samples': 16260096, 'steps': 31757, 'loss/train': 1.400538444519043} 02/25/2022 18:42:36 - INFO - codeparrot_training - Step 31758: {'lr': 0.00015799928948155746, 'samples': 16260608, 'steps': 31758, 'loss/train': 1.424329400062561} 02/25/2022 18:42:42 - INFO - codeparrot_training - Step 31759: {'lr': 0.0001579840754581105, 'samples': 16261120, 'steps': 31759, 'loss/train': 2.344059944152832} 02/25/2022 18:42:46 - INFO - codeparrot_training - Step 31760: {'lr': 0.00015796886182883053, 'samples': 16261632, 'steps': 31760, 'loss/train': 0.9455052614212036} 02/25/2022 18:42:52 - INFO - codeparrot_training - Step 31761: {'lr': 0.00015795364859378275, 'samples': 16262144, 'steps': 31761, 'loss/train': 1.7467272281646729} 02/25/2022 18:42:55 - INFO - codeparrot_training - Step 31762: {'lr': 0.00015793843575303228, 'samples': 16262656, 'steps': 31762, 'loss/train': 0.9557493329048157} 02/25/2022 18:43:01 - INFO - codeparrot_training - Step 31763: {'lr': 0.0001579232233066444, 'samples': 16263168, 'steps': 31763, 'loss/train': 1.8905915021896362} 02/25/2022 18:43:04 - INFO - codeparrot_training - Step 31764: {'lr': 0.00015790801125468418, 'samples': 16263680, 'steps': 31764, 'loss/train': 2.09942889213562} 02/25/2022 18:43:09 - INFO - codeparrot_training - Step 31765: {'lr': 0.00015789279959721675, 'samples': 16264192, 'steps': 31765, 'loss/train': 2.7502262592315674} 02/25/2022 18:43:13 - INFO - codeparrot_training - Step 31766: {'lr': 0.0001578775883343073, 'samples': 16264704, 'steps': 31766, 'loss/train': 1.4074475765228271} 02/25/2022 18:43:18 - INFO - codeparrot_training - Step 31767: {'lr': 0.00015786237746602105, 'samples': 16265216, 'steps': 31767, 'loss/train': 1.8503464460372925} 02/25/2022 18:43:22 - INFO - codeparrot_training - Step 31768: {'lr': 0.00015784716699242317, 'samples': 16265728, 'steps': 31768, 'loss/train': 2.2772557735443115} 02/25/2022 18:43:28 - INFO - codeparrot_training - Step 31769: {'lr': 0.00015783195691357866, 'samples': 16266240, 'steps': 31769, 'loss/train': 1.5083093643188477} 02/25/2022 18:43:31 - INFO - codeparrot_training - Step 31770: {'lr': 0.0001578167472295528, 'samples': 16266752, 'steps': 31770, 'loss/train': 1.286312222480774} 02/25/2022 18:43:37 - INFO - codeparrot_training - Step 31771: {'lr': 0.00015780153794041075, 'samples': 16267264, 'steps': 31771, 'loss/train': 1.2470710277557373} 02/25/2022 18:43:40 - INFO - codeparrot_training - Step 31772: {'lr': 0.0001577863290462177, 'samples': 16267776, 'steps': 31772, 'loss/train': 0.790630042552948} 02/25/2022 18:43:46 - INFO - codeparrot_training - Step 31773: {'lr': 0.00015777112054703862, 'samples': 16268288, 'steps': 31773, 'loss/train': 1.1662843227386475} 02/25/2022 18:43:52 - INFO - codeparrot_training - Step 31774: {'lr': 0.0001577559124429388, 'samples': 16268800, 'steps': 31774, 'loss/train': 0.5784724354743958} 02/25/2022 18:43:55 - INFO - codeparrot_training - Step 31775: {'lr': 0.0001577407047339834, 'samples': 16269312, 'steps': 31775, 'loss/train': 1.2439956665039062} 02/25/2022 18:44:00 - INFO - codeparrot_training - Step 31776: {'lr': 0.00015772549742023757, 'samples': 16269824, 'steps': 31776, 'loss/train': 1.5234769582748413} 02/25/2022 18:44:04 - INFO - codeparrot_training - Step 31777: {'lr': 0.00015771029050176634, 'samples': 16270336, 'steps': 31777, 'loss/train': 2.2709524631500244} 02/25/2022 18:44:09 - INFO - codeparrot_training - Step 31778: {'lr': 0.00015769508397863485, 'samples': 16270848, 'steps': 31778, 'loss/train': 1.2925472259521484} 02/25/2022 18:44:13 - INFO - codeparrot_training - Step 31779: {'lr': 0.00015767987785090837, 'samples': 16271360, 'steps': 31779, 'loss/train': 1.24466073513031} 02/25/2022 18:44:18 - INFO - codeparrot_training - Step 31780: {'lr': 0.0001576646721186521, 'samples': 16271872, 'steps': 31780, 'loss/train': 0.8119126558303833} 02/25/2022 18:44:22 - INFO - codeparrot_training - Step 31781: {'lr': 0.00015764946678193092, 'samples': 16272384, 'steps': 31781, 'loss/train': 1.8556114435195923} 02/25/2022 18:44:27 - INFO - codeparrot_training - Step 31782: {'lr': 0.00015763426184081008, 'samples': 16272896, 'steps': 31782, 'loss/train': 0.774508535861969} 02/25/2022 18:44:31 - INFO - codeparrot_training - Step 31783: {'lr': 0.00015761905729535478, 'samples': 16273408, 'steps': 31783, 'loss/train': 1.894516944885254} 02/25/2022 18:44:38 - INFO - codeparrot_training - Step 31784: {'lr': 0.00015760385314563007, 'samples': 16273920, 'steps': 31784, 'loss/train': 2.374850034713745} 02/25/2022 18:44:41 - INFO - codeparrot_training - Step 31785: {'lr': 0.00015758864939170126, 'samples': 16274432, 'steps': 31785, 'loss/train': 1.2760549783706665} 02/25/2022 18:44:47 - INFO - codeparrot_training - Step 31786: {'lr': 0.0001575734460336332, 'samples': 16274944, 'steps': 31786, 'loss/train': 2.0600438117980957} 02/25/2022 18:44:51 - INFO - codeparrot_training - Step 31787: {'lr': 0.00015755824307149113, 'samples': 16275456, 'steps': 31787, 'loss/train': 2.0675594806671143} 02/25/2022 18:44:56 - INFO - codeparrot_training - Step 31788: {'lr': 0.00015754304050534022, 'samples': 16275968, 'steps': 31788, 'loss/train': 1.8726649284362793} 02/25/2022 18:45:00 - INFO - codeparrot_training - Step 31789: {'lr': 0.00015752783833524565, 'samples': 16276480, 'steps': 31789, 'loss/train': 2.101017951965332} 02/25/2022 18:45:05 - INFO - codeparrot_training - Step 31790: {'lr': 0.00015751263656127234, 'samples': 16276992, 'steps': 31790, 'loss/train': 1.1056339740753174} 02/25/2022 18:45:09 - INFO - codeparrot_training - Step 31791: {'lr': 0.0001574974351834856, 'samples': 16277504, 'steps': 31791, 'loss/train': 1.7317559719085693} 02/25/2022 18:45:14 - INFO - codeparrot_training - Step 31792: {'lr': 0.00015748223420195042, 'samples': 16278016, 'steps': 31792, 'loss/train': 1.4771580696105957} 02/25/2022 18:45:17 - INFO - codeparrot_training - Step 31793: {'lr': 0.00015746703361673207, 'samples': 16278528, 'steps': 31793, 'loss/train': 1.4978071451187134} 02/25/2022 18:45:24 - INFO - codeparrot_training - Step 31794: {'lr': 0.00015745183342789546, 'samples': 16279040, 'steps': 31794, 'loss/train': 1.4167848825454712} 02/25/2022 18:45:28 - INFO - codeparrot_training - Step 31795: {'lr': 0.0001574366336355058, 'samples': 16279552, 'steps': 31795, 'loss/train': 0.8790462613105774} 02/25/2022 18:45:33 - INFO - codeparrot_training - Step 31796: {'lr': 0.00015742143423962823, 'samples': 16280064, 'steps': 31796, 'loss/train': 1.5654898881912231} 02/25/2022 18:45:37 - INFO - codeparrot_training - Step 31797: {'lr': 0.00015740623524032794, 'samples': 16280576, 'steps': 31797, 'loss/train': 1.3439390659332275} 02/25/2022 18:45:42 - INFO - codeparrot_training - Step 31798: {'lr': 0.00015739103663766984, 'samples': 16281088, 'steps': 31798, 'loss/train': 0.5737715363502502} 02/25/2022 18:45:45 - INFO - codeparrot_training - Step 31799: {'lr': 0.00015737583843171914, 'samples': 16281600, 'steps': 31799, 'loss/train': 1.4663745164871216} 02/25/2022 18:45:51 - INFO - codeparrot_training - Step 31800: {'lr': 0.00015736064062254093, 'samples': 16282112, 'steps': 31800, 'loss/train': 1.7405738830566406} 02/25/2022 18:45:55 - INFO - codeparrot_training - Step 31801: {'lr': 0.00015734544321020037, 'samples': 16282624, 'steps': 31801, 'loss/train': 1.870730996131897} 02/25/2022 18:46:00 - INFO - codeparrot_training - Step 31802: {'lr': 0.0001573302461947625, 'samples': 16283136, 'steps': 31802, 'loss/train': 2.4010608196258545} 02/25/2022 18:46:03 - INFO - codeparrot_training - Step 31803: {'lr': 0.0001573150495762924, 'samples': 16283648, 'steps': 31803, 'loss/train': 2.220834493637085} 02/25/2022 18:46:10 - INFO - codeparrot_training - Step 31804: {'lr': 0.0001572998533548552, 'samples': 16284160, 'steps': 31804, 'loss/train': 2.6818807125091553} 02/25/2022 18:46:14 - INFO - codeparrot_training - Step 31805: {'lr': 0.00015728465753051598, 'samples': 16284672, 'steps': 31805, 'loss/train': 1.5541250705718994} 02/25/2022 18:46:19 - INFO - codeparrot_training - Step 31806: {'lr': 0.0001572694621033399, 'samples': 16285184, 'steps': 31806, 'loss/train': 1.9580897092819214} 02/25/2022 18:46:23 - INFO - codeparrot_training - Step 31807: {'lr': 0.00015725426707339195, 'samples': 16285696, 'steps': 31807, 'loss/train': 2.7889318466186523} 02/25/2022 18:46:28 - INFO - codeparrot_training - Step 31808: {'lr': 0.0001572390724407373, 'samples': 16286208, 'steps': 31808, 'loss/train': 1.3352032899856567} 02/25/2022 18:46:32 - INFO - codeparrot_training - Step 31809: {'lr': 0.00015722387820544097, 'samples': 16286720, 'steps': 31809, 'loss/train': 1.9415841102600098} 02/25/2022 18:46:37 - INFO - codeparrot_training - Step 31810: {'lr': 0.00015720868436756808, 'samples': 16287232, 'steps': 31810, 'loss/train': 1.6297929286956787} 02/25/2022 18:46:41 - INFO - codeparrot_training - Step 31811: {'lr': 0.00015719349092718383, 'samples': 16287744, 'steps': 31811, 'loss/train': 1.4024025201797485} 02/25/2022 18:46:46 - INFO - codeparrot_training - Step 31812: {'lr': 0.00015717829788435307, 'samples': 16288256, 'steps': 31812, 'loss/train': 2.3132739067077637} 02/25/2022 18:46:50 - INFO - codeparrot_training - Step 31813: {'lr': 0.00015716310523914104, 'samples': 16288768, 'steps': 31813, 'loss/train': 1.8912065029144287} 02/25/2022 18:46:55 - INFO - codeparrot_training - Step 31814: {'lr': 0.00015714791299161276, 'samples': 16289280, 'steps': 31814, 'loss/train': 1.8868147134780884} 02/25/2022 18:47:02 - INFO - codeparrot_training - Step 31815: {'lr': 0.00015713272114183346, 'samples': 16289792, 'steps': 31815, 'loss/train': 0.5052993893623352} 02/25/2022 18:47:05 - INFO - codeparrot_training - Step 31816: {'lr': 0.00015711752968986797, 'samples': 16290304, 'steps': 31816, 'loss/train': 1.3647977113723755} 02/25/2022 18:47:11 - INFO - codeparrot_training - Step 31817: {'lr': 0.00015710233863578153, 'samples': 16290816, 'steps': 31817, 'loss/train': 2.4201929569244385} 02/25/2022 18:47:14 - INFO - codeparrot_training - Step 31818: {'lr': 0.00015708714797963913, 'samples': 16291328, 'steps': 31818, 'loss/train': 0.685491144657135} 02/25/2022 18:47:20 - INFO - codeparrot_training - Step 31819: {'lr': 0.0001570719577215059, 'samples': 16291840, 'steps': 31819, 'loss/train': 1.572088360786438} 02/25/2022 18:47:23 - INFO - codeparrot_training - Step 31820: {'lr': 0.00015705676786144702, 'samples': 16292352, 'steps': 31820, 'loss/train': 2.3007888793945312} 02/25/2022 18:47:29 - INFO - codeparrot_training - Step 31821: {'lr': 0.00015704157839952732, 'samples': 16292864, 'steps': 31821, 'loss/train': 0.7093051671981812} 02/25/2022 18:47:32 - INFO - codeparrot_training - Step 31822: {'lr': 0.000157026389335812, 'samples': 16293376, 'steps': 31822, 'loss/train': 2.0013861656188965} 02/25/2022 18:47:38 - INFO - codeparrot_training - Step 31823: {'lr': 0.00015701120067036607, 'samples': 16293888, 'steps': 31823, 'loss/train': 0.8477587699890137} 02/25/2022 18:47:41 - INFO - codeparrot_training - Step 31824: {'lr': 0.00015699601240325473, 'samples': 16294400, 'steps': 31824, 'loss/train': 1.998630166053772} 02/25/2022 18:47:47 - INFO - codeparrot_training - Step 31825: {'lr': 0.00015698082453454283, 'samples': 16294912, 'steps': 31825, 'loss/train': 0.9047789573669434} 02/25/2022 18:47:50 - INFO - codeparrot_training - Step 31826: {'lr': 0.00015696563706429556, 'samples': 16295424, 'steps': 31826, 'loss/train': 2.0305511951446533} 02/25/2022 18:47:56 - INFO - codeparrot_training - Step 31827: {'lr': 0.00015695044999257797, 'samples': 16295936, 'steps': 31827, 'loss/train': 1.7714321613311768} 02/25/2022 18:47:59 - INFO - codeparrot_training - Step 31828: {'lr': 0.0001569352633194552, 'samples': 16296448, 'steps': 31828, 'loss/train': 2.3439695835113525} 02/25/2022 18:48:03 - INFO - codeparrot_training - Step 31829: {'lr': 0.00015692007704499209, 'samples': 16296960, 'steps': 31829, 'loss/train': 1.3583475351333618} 02/25/2022 18:48:10 - INFO - codeparrot_training - Step 31830: {'lr': 0.00015690489116925382, 'samples': 16297472, 'steps': 31830, 'loss/train': 2.480048656463623} 02/25/2022 18:48:13 - INFO - codeparrot_training - Step 31831: {'lr': 0.00015688970569230544, 'samples': 16297984, 'steps': 31831, 'loss/train': 0.7590587735176086} 02/25/2022 18:48:17 - INFO - codeparrot_training - Step 31832: {'lr': 0.000156874520614212, 'samples': 16298496, 'steps': 31832, 'loss/train': 1.787411093711853} 02/25/2022 18:48:22 - INFO - codeparrot_training - Step 31833: {'lr': 0.00015685933593503865, 'samples': 16299008, 'steps': 31833, 'loss/train': 1.5617295503616333} 02/25/2022 18:48:28 - INFO - codeparrot_training - Step 31834: {'lr': 0.00015684415165485022, 'samples': 16299520, 'steps': 31834, 'loss/train': 1.5835037231445312} 02/25/2022 18:48:31 - INFO - codeparrot_training - Step 31835: {'lr': 0.00015682896777371186, 'samples': 16300032, 'steps': 31835, 'loss/train': 2.327470541000366} 02/25/2022 18:48:37 - INFO - codeparrot_training - Step 31836: {'lr': 0.00015681378429168865, 'samples': 16300544, 'steps': 31836, 'loss/train': 1.2230695486068726} 02/25/2022 18:48:40 - INFO - codeparrot_training - Step 31837: {'lr': 0.00015679860120884566, 'samples': 16301056, 'steps': 31837, 'loss/train': 1.0115549564361572} 02/25/2022 18:48:44 - INFO - codeparrot_training - Step 31838: {'lr': 0.00015678341852524776, 'samples': 16301568, 'steps': 31838, 'loss/train': 1.486220359802246} 02/25/2022 18:48:49 - INFO - codeparrot_training - Step 31839: {'lr': 0.00015676823624096012, 'samples': 16302080, 'steps': 31839, 'loss/train': 2.473174571990967} 02/25/2022 18:48:56 - INFO - codeparrot_training - Step 31840: {'lr': 0.00015675305435604775, 'samples': 16302592, 'steps': 31840, 'loss/train': 1.6202027797698975} 02/25/2022 18:48:59 - INFO - codeparrot_training - Step 31841: {'lr': 0.0001567378728705758, 'samples': 16303104, 'steps': 31841, 'loss/train': 1.4797914028167725} 02/25/2022 18:49:05 - INFO - codeparrot_training - Step 31842: {'lr': 0.00015672269178460904, 'samples': 16303616, 'steps': 31842, 'loss/train': 2.4769647121429443} 02/25/2022 18:49:08 - INFO - codeparrot_training - Step 31843: {'lr': 0.00015670751109821268, 'samples': 16304128, 'steps': 31843, 'loss/train': 1.3690781593322754} 02/25/2022 18:49:14 - INFO - codeparrot_training - Step 31844: {'lr': 0.0001566923308114518, 'samples': 16304640, 'steps': 31844, 'loss/train': 1.8152060508728027} 02/25/2022 18:49:17 - INFO - codeparrot_training - Step 31845: {'lr': 0.0001566771509243912, 'samples': 16305152, 'steps': 31845, 'loss/train': 2.0704500675201416} 02/25/2022 18:49:23 - INFO - codeparrot_training - Step 31846: {'lr': 0.0001566619714370962, 'samples': 16305664, 'steps': 31846, 'loss/train': 1.2604831457138062} 02/25/2022 18:49:26 - INFO - codeparrot_training - Step 31847: {'lr': 0.00015664679234963158, 'samples': 16306176, 'steps': 31847, 'loss/train': 2.9137635231018066} 02/25/2022 18:49:32 - INFO - codeparrot_training - Step 31848: {'lr': 0.00015663161366206253, 'samples': 16306688, 'steps': 31848, 'loss/train': 1.4659334421157837} 02/25/2022 18:49:35 - INFO - codeparrot_training - Step 31849: {'lr': 0.00015661643537445393, 'samples': 16307200, 'steps': 31849, 'loss/train': 1.1687768697738647} 02/25/2022 18:49:41 - INFO - codeparrot_training - Step 31850: {'lr': 0.00015660125748687094, 'samples': 16307712, 'steps': 31850, 'loss/train': 2.955305576324463} 02/25/2022 18:49:44 - INFO - codeparrot_training - Step 31851: {'lr': 0.00015658607999937847, 'samples': 16308224, 'steps': 31851, 'loss/train': 2.2020745277404785} 02/25/2022 18:49:51 - INFO - codeparrot_training - Step 31852: {'lr': 0.00015657090291204157, 'samples': 16308736, 'steps': 31852, 'loss/train': 1.3224120140075684} 02/25/2022 18:49:54 - INFO - codeparrot_training - Step 31853: {'lr': 0.00015655572622492523, 'samples': 16309248, 'steps': 31853, 'loss/train': 0.5819587707519531} 02/25/2022 18:50:00 - INFO - codeparrot_training - Step 31854: {'lr': 0.00015654054993809454, 'samples': 16309760, 'steps': 31854, 'loss/train': 1.0389347076416016} 02/25/2022 18:50:03 - INFO - codeparrot_training - Step 31855: {'lr': 0.00015652537405161444, 'samples': 16310272, 'steps': 31855, 'loss/train': 1.2572869062423706} 02/25/2022 18:50:09 - INFO - codeparrot_training - Step 31856: {'lr': 0.00015651019856554994, 'samples': 16310784, 'steps': 31856, 'loss/train': 1.8519079685211182} 02/25/2022 18:50:12 - INFO - codeparrot_training - Step 31857: {'lr': 0.00015649502347996603, 'samples': 16311296, 'steps': 31857, 'loss/train': 1.3159236907958984} 02/25/2022 18:50:18 - INFO - codeparrot_training - Step 31858: {'lr': 0.00015647984879492777, 'samples': 16311808, 'steps': 31858, 'loss/train': 1.7824357748031616} 02/25/2022 18:50:21 - INFO - codeparrot_training - Step 31859: {'lr': 0.00015646467451050023, 'samples': 16312320, 'steps': 31859, 'loss/train': 8.676250457763672} 02/25/2022 18:50:27 - INFO - codeparrot_training - Step 31860: {'lr': 0.00015644950062674822, 'samples': 16312832, 'steps': 31860, 'loss/train': 2.1044960021972656} 02/25/2022 18:50:30 - INFO - codeparrot_training - Step 31861: {'lr': 0.00015643432714373685, 'samples': 16313344, 'steps': 31861, 'loss/train': 1.0503605604171753} 02/25/2022 18:50:36 - INFO - codeparrot_training - Step 31862: {'lr': 0.00015641915406153112, 'samples': 16313856, 'steps': 31862, 'loss/train': 0.8284381628036499} 02/25/2022 18:50:40 - INFO - codeparrot_training - Step 31863: {'lr': 0.00015640398138019613, 'samples': 16314368, 'steps': 31863, 'loss/train': 1.6933575868606567} 02/25/2022 18:50:45 - INFO - codeparrot_training - Step 31864: {'lr': 0.00015638880909979664, 'samples': 16314880, 'steps': 31864, 'loss/train': 1.7500853538513184} 02/25/2022 18:50:49 - INFO - codeparrot_training - Step 31865: {'lr': 0.00015637363722039777, 'samples': 16315392, 'steps': 31865, 'loss/train': 1.5358532667160034} 02/25/2022 18:50:54 - INFO - codeparrot_training - Step 31866: {'lr': 0.00015635846574206449, 'samples': 16315904, 'steps': 31866, 'loss/train': 2.0121545791625977} 02/25/2022 18:50:58 - INFO - codeparrot_training - Step 31867: {'lr': 0.00015634329466486182, 'samples': 16316416, 'steps': 31867, 'loss/train': 2.1975982189178467} 02/25/2022 18:51:03 - INFO - codeparrot_training - Step 31868: {'lr': 0.00015632812398885487, 'samples': 16316928, 'steps': 31868, 'loss/train': 1.0573875904083252} 02/25/2022 18:51:07 - INFO - codeparrot_training - Step 31869: {'lr': 0.00015631295371410837, 'samples': 16317440, 'steps': 31869, 'loss/train': 1.326983094215393} 02/25/2022 18:51:12 - INFO - codeparrot_training - Step 31870: {'lr': 0.0001562977838406874, 'samples': 16317952, 'steps': 31870, 'loss/train': 2.1941564083099365} 02/25/2022 18:51:16 - INFO - codeparrot_training - Step 31871: {'lr': 0.000156282614368657, 'samples': 16318464, 'steps': 31871, 'loss/train': 1.5488618612289429} 02/25/2022 18:51:21 - INFO - codeparrot_training - Step 31872: {'lr': 0.00015626744529808223, 'samples': 16318976, 'steps': 31872, 'loss/train': 1.1792395114898682} 02/25/2022 18:51:25 - INFO - codeparrot_training - Step 31873: {'lr': 0.00015625227662902782, 'samples': 16319488, 'steps': 31873, 'loss/train': 1.9254331588745117} 02/25/2022 18:51:30 - INFO - codeparrot_training - Step 31874: {'lr': 0.00015623710836155892, 'samples': 16320000, 'steps': 31874, 'loss/train': 1.8640425205230713} 02/25/2022 18:51:34 - INFO - codeparrot_training - Step 31875: {'lr': 0.00015622194049574048, 'samples': 16320512, 'steps': 31875, 'loss/train': 0.4795461595058441} 02/25/2022 18:51:40 - INFO - codeparrot_training - Step 31876: {'lr': 0.00015620677303163755, 'samples': 16321024, 'steps': 31876, 'loss/train': 1.900381326675415} 02/25/2022 18:51:44 - INFO - codeparrot_training - Step 31877: {'lr': 0.00015619160596931492, 'samples': 16321536, 'steps': 31877, 'loss/train': 0.43327316641807556} 02/25/2022 18:51:49 - INFO - codeparrot_training - Step 31878: {'lr': 0.00015617643930883768, 'samples': 16322048, 'steps': 31878, 'loss/train': 1.862421989440918} 02/25/2022 18:51:53 - INFO - codeparrot_training - Step 31879: {'lr': 0.00015616127305027077, 'samples': 16322560, 'steps': 31879, 'loss/train': 1.6216498613357544} 02/25/2022 18:51:58 - INFO - codeparrot_training - Step 31880: {'lr': 0.0001561461071936792, 'samples': 16323072, 'steps': 31880, 'loss/train': 2.3397865295410156} 02/25/2022 18:52:02 - INFO - codeparrot_training - Step 31881: {'lr': 0.000156130941739128, 'samples': 16323584, 'steps': 31881, 'loss/train': 2.486037254333496} 02/25/2022 18:52:07 - INFO - codeparrot_training - Step 31882: {'lr': 0.0001561157766866819, 'samples': 16324096, 'steps': 31882, 'loss/train': 0.6346574425697327} 02/25/2022 18:52:11 - INFO - codeparrot_training - Step 31883: {'lr': 0.000156100612036406, 'samples': 16324608, 'steps': 31883, 'loss/train': 0.8969175219535828} 02/25/2022 18:52:16 - INFO - codeparrot_training - Step 31884: {'lr': 0.00015608544778836534, 'samples': 16325120, 'steps': 31884, 'loss/train': 1.0674989223480225} 02/25/2022 18:52:20 - INFO - codeparrot_training - Step 31885: {'lr': 0.0001560702839426248, 'samples': 16325632, 'steps': 31885, 'loss/train': 1.240315556526184} 02/25/2022 18:52:26 - INFO - codeparrot_training - Step 31886: {'lr': 0.00015605512049924927, 'samples': 16326144, 'steps': 31886, 'loss/train': 0.6611651182174683} 02/25/2022 18:52:30 - INFO - codeparrot_training - Step 31887: {'lr': 0.0001560399574583038, 'samples': 16326656, 'steps': 31887, 'loss/train': 1.142682671546936} 02/25/2022 18:52:35 - INFO - codeparrot_training - Step 31888: {'lr': 0.00015602479481985332, 'samples': 16327168, 'steps': 31888, 'loss/train': 1.334628701210022} 02/25/2022 18:52:39 - INFO - codeparrot_training - Step 31889: {'lr': 0.00015600963258396284, 'samples': 16327680, 'steps': 31889, 'loss/train': 1.9676766395568848} 02/25/2022 18:52:44 - INFO - codeparrot_training - Step 31890: {'lr': 0.00015599447075069714, 'samples': 16328192, 'steps': 31890, 'loss/train': 1.572779893875122} 02/25/2022 18:52:50 - INFO - codeparrot_training - Step 31891: {'lr': 0.0001559793093201213, 'samples': 16328704, 'steps': 31891, 'loss/train': 2.2284061908721924} 02/25/2022 18:52:53 - INFO - codeparrot_training - Step 31892: {'lr': 0.0001559641482923003, 'samples': 16329216, 'steps': 31892, 'loss/train': 1.5746623277664185} 02/25/2022 18:52:59 - INFO - codeparrot_training - Step 31893: {'lr': 0.00015594898766729898, 'samples': 16329728, 'steps': 31893, 'loss/train': 1.9416314363479614} 02/25/2022 18:53:02 - INFO - codeparrot_training - Step 31894: {'lr': 0.00015593382744518236, 'samples': 16330240, 'steps': 31894, 'loss/train': 1.0915865898132324} 02/25/2022 18:53:08 - INFO - codeparrot_training - Step 31895: {'lr': 0.00015591866762601538, 'samples': 16330752, 'steps': 31895, 'loss/train': 2.1597647666931152} 02/25/2022 18:53:11 - INFO - codeparrot_training - Step 31896: {'lr': 0.0001559035082098629, 'samples': 16331264, 'steps': 31896, 'loss/train': 1.8096113204956055} 02/25/2022 18:53:18 - INFO - codeparrot_training - Step 31897: {'lr': 0.00015588834919678991, 'samples': 16331776, 'steps': 31897, 'loss/train': 0.5843462347984314} 02/25/2022 18:53:21 - INFO - codeparrot_training - Step 31898: {'lr': 0.0001558731905868614, 'samples': 16332288, 'steps': 31898, 'loss/train': 1.7148973941802979} 02/25/2022 18:53:26 - INFO - codeparrot_training - Step 31899: {'lr': 0.00015585803238014223, 'samples': 16332800, 'steps': 31899, 'loss/train': 1.8135031461715698} 02/25/2022 18:53:30 - INFO - codeparrot_training - Step 31900: {'lr': 0.00015584287457669733, 'samples': 16333312, 'steps': 31900, 'loss/train': 1.3195728063583374} 02/25/2022 18:53:35 - INFO - codeparrot_training - Step 31901: {'lr': 0.00015582771717659167, 'samples': 16333824, 'steps': 31901, 'loss/train': 1.6785763502120972} 02/25/2022 18:53:39 - INFO - codeparrot_training - Step 31902: {'lr': 0.00015581256017989023, 'samples': 16334336, 'steps': 31902, 'loss/train': 0.8784569501876831} 02/25/2022 18:53:44 - INFO - codeparrot_training - Step 31903: {'lr': 0.00015579740358665777, 'samples': 16334848, 'steps': 31903, 'loss/train': 1.6951404809951782} 02/25/2022 18:53:48 - INFO - codeparrot_training - Step 31904: {'lr': 0.00015578224739695937, 'samples': 16335360, 'steps': 31904, 'loss/train': 1.5694926977157593} 02/25/2022 18:53:54 - INFO - codeparrot_training - Step 31905: {'lr': 0.0001557670916108599, 'samples': 16335872, 'steps': 31905, 'loss/train': 2.8908092975616455} 02/25/2022 18:53:57 - INFO - codeparrot_training - Step 31906: {'lr': 0.00015575193622842425, 'samples': 16336384, 'steps': 31906, 'loss/train': 0.05563532933592796} 02/25/2022 18:54:03 - INFO - codeparrot_training - Step 31907: {'lr': 0.00015573678124971747, 'samples': 16336896, 'steps': 31907, 'loss/train': 1.988373041152954} 02/25/2022 18:54:07 - INFO - codeparrot_training - Step 31908: {'lr': 0.00015572162667480433, 'samples': 16337408, 'steps': 31908, 'loss/train': 2.0349233150482178} 02/25/2022 18:54:12 - INFO - codeparrot_training - Step 31909: {'lr': 0.00015570647250374978, 'samples': 16337920, 'steps': 31909, 'loss/train': 0.9201633334159851} 02/25/2022 18:54:16 - INFO - codeparrot_training - Step 31910: {'lr': 0.00015569131873661873, 'samples': 16338432, 'steps': 31910, 'loss/train': 2.3175086975097656} 02/25/2022 18:54:21 - INFO - codeparrot_training - Step 31911: {'lr': 0.0001556761653734763, 'samples': 16338944, 'steps': 31911, 'loss/train': 1.4183411598205566} 02/25/2022 18:54:25 - INFO - codeparrot_training - Step 31912: {'lr': 0.00015566101241438706, 'samples': 16339456, 'steps': 31912, 'loss/train': 1.8293401002883911} 02/25/2022 18:54:30 - INFO - codeparrot_training - Step 31913: {'lr': 0.0001556458598594161, 'samples': 16339968, 'steps': 31913, 'loss/train': 1.4242357015609741} 02/25/2022 18:54:34 - INFO - codeparrot_training - Step 31914: {'lr': 0.00015563070770862835, 'samples': 16340480, 'steps': 31914, 'loss/train': 2.116787910461426} 02/25/2022 18:54:39 - INFO - codeparrot_training - Step 31915: {'lr': 0.00015561555596208876, 'samples': 16340992, 'steps': 31915, 'loss/train': 2.064756155014038} 02/25/2022 18:54:42 - INFO - codeparrot_training - Step 31916: {'lr': 0.00015560040461986204, 'samples': 16341504, 'steps': 31916, 'loss/train': 1.232796549797058} 02/25/2022 18:54:48 - INFO - codeparrot_training - Step 31917: {'lr': 0.00015558525368201327, 'samples': 16342016, 'steps': 31917, 'loss/train': 2.007220983505249} 02/25/2022 18:54:51 - INFO - codeparrot_training - Step 31918: {'lr': 0.00015557010314860723, 'samples': 16342528, 'steps': 31918, 'loss/train': 1.057302713394165} 02/25/2022 18:54:57 - INFO - codeparrot_training - Step 31919: {'lr': 0.0001555549530197089, 'samples': 16343040, 'steps': 31919, 'loss/train': 1.8840572834014893} 02/25/2022 18:55:00 - INFO - codeparrot_training - Step 31920: {'lr': 0.00015553980329538326, 'samples': 16343552, 'steps': 31920, 'loss/train': 2.9769749641418457} 02/25/2022 18:55:06 - INFO - codeparrot_training - Step 31921: {'lr': 0.00015552465397569503, 'samples': 16344064, 'steps': 31921, 'loss/train': 1.6371641159057617} 02/25/2022 18:55:09 - INFO - codeparrot_training - Step 31922: {'lr': 0.00015550950506070915, 'samples': 16344576, 'steps': 31922, 'loss/train': 1.532838225364685} 02/25/2022 18:55:16 - INFO - codeparrot_training - Step 31923: {'lr': 0.00015549435655049057, 'samples': 16345088, 'steps': 31923, 'loss/train': 0.9875187277793884} 02/25/2022 18:55:19 - INFO - codeparrot_training - Step 31924: {'lr': 0.00015547920844510428, 'samples': 16345600, 'steps': 31924, 'loss/train': 1.5907374620437622} 02/25/2022 18:55:25 - INFO - codeparrot_training - Step 31925: {'lr': 0.0001554640607446149, 'samples': 16346112, 'steps': 31925, 'loss/train': 2.4162392616271973} 02/25/2022 18:55:28 - INFO - codeparrot_training - Step 31926: {'lr': 0.00015544891344908748, 'samples': 16346624, 'steps': 31926, 'loss/train': 2.298130989074707} 02/25/2022 18:55:34 - INFO - codeparrot_training - Step 31927: {'lr': 0.00015543376655858688, 'samples': 16347136, 'steps': 31927, 'loss/train': 1.5891966819763184} 02/25/2022 18:55:37 - INFO - codeparrot_training - Step 31928: {'lr': 0.00015541862007317807, 'samples': 16347648, 'steps': 31928, 'loss/train': 2.1017534732818604} 02/25/2022 18:55:42 - INFO - codeparrot_training - Step 31929: {'lr': 0.00015540347399292586, 'samples': 16348160, 'steps': 31929, 'loss/train': 1.8119382858276367} 02/25/2022 18:55:46 - INFO - codeparrot_training - Step 31930: {'lr': 0.00015538832831789507, 'samples': 16348672, 'steps': 31930, 'loss/train': 1.8796244859695435} 02/25/2022 18:55:51 - INFO - codeparrot_training - Step 31931: {'lr': 0.00015537318304815063, 'samples': 16349184, 'steps': 31931, 'loss/train': 0.5294276475906372} 02/25/2022 18:55:55 - INFO - codeparrot_training - Step 31932: {'lr': 0.00015535803818375748, 'samples': 16349696, 'steps': 31932, 'loss/train': 2.3740463256835938} 02/25/2022 18:56:01 - INFO - codeparrot_training - Step 31933: {'lr': 0.00015534289372478045, 'samples': 16350208, 'steps': 31933, 'loss/train': 0.30523836612701416} 02/25/2022 18:56:05 - INFO - codeparrot_training - Step 31934: {'lr': 0.00015532774967128439, 'samples': 16350720, 'steps': 31934, 'loss/train': 1.152420163154602} 02/25/2022 18:56:10 - INFO - codeparrot_training - Step 31935: {'lr': 0.0001553126060233342, 'samples': 16351232, 'steps': 31935, 'loss/train': 2.621337413787842} 02/25/2022 18:56:14 - INFO - codeparrot_training - Step 31936: {'lr': 0.00015529746278099467, 'samples': 16351744, 'steps': 31936, 'loss/train': 1.2015964984893799} 02/25/2022 18:56:19 - INFO - codeparrot_training - Step 31937: {'lr': 0.00015528231994433082, 'samples': 16352256, 'steps': 31937, 'loss/train': 0.0820598155260086} 02/25/2022 18:56:23 - INFO - codeparrot_training - Step 31938: {'lr': 0.00015526717751340735, 'samples': 16352768, 'steps': 31938, 'loss/train': 2.08146595954895} 02/25/2022 18:56:28 - INFO - codeparrot_training - Step 31939: {'lr': 0.00015525203548828932, 'samples': 16353280, 'steps': 31939, 'loss/train': 2.4808058738708496} 02/25/2022 18:56:32 - INFO - codeparrot_training - Step 31940: {'lr': 0.0001552368938690414, 'samples': 16353792, 'steps': 31940, 'loss/train': 1.5145026445388794} 02/25/2022 18:56:37 - INFO - codeparrot_training - Step 31941: {'lr': 0.00015522175265572852, 'samples': 16354304, 'steps': 31941, 'loss/train': 0.05511852353811264} 02/25/2022 18:56:41 - INFO - codeparrot_training - Step 31942: {'lr': 0.00015520661184841567, 'samples': 16354816, 'steps': 31942, 'loss/train': 1.681165337562561} 02/25/2022 18:56:48 - INFO - codeparrot_training - Step 31943: {'lr': 0.0001551914714471675, 'samples': 16355328, 'steps': 31943, 'loss/train': 2.3063130378723145} 02/25/2022 18:56:51 - INFO - codeparrot_training - Step 31944: {'lr': 0.00015517633145204895, 'samples': 16355840, 'steps': 31944, 'loss/train': 1.2548444271087646} 02/25/2022 18:56:57 - INFO - codeparrot_training - Step 31945: {'lr': 0.00015516119186312488, 'samples': 16356352, 'steps': 31945, 'loss/train': 1.2488266229629517} 02/25/2022 18:57:00 - INFO - codeparrot_training - Step 31946: {'lr': 0.00015514605268046028, 'samples': 16356864, 'steps': 31946, 'loss/train': 1.8956669569015503} 02/25/2022 18:57:06 - INFO - codeparrot_training - Step 31947: {'lr': 0.00015513091390411977, 'samples': 16357376, 'steps': 31947, 'loss/train': 1.4030108451843262} 02/25/2022 18:57:09 - INFO - codeparrot_training - Step 31948: {'lr': 0.00015511577553416826, 'samples': 16357888, 'steps': 31948, 'loss/train': 2.4183309078216553} 02/25/2022 18:57:15 - INFO - codeparrot_training - Step 31949: {'lr': 0.00015510063757067066, 'samples': 16358400, 'steps': 31949, 'loss/train': 2.086169719696045} 02/25/2022 18:57:20 - INFO - codeparrot_training - Step 31950: {'lr': 0.0001550855000136919, 'samples': 16358912, 'steps': 31950, 'loss/train': 1.7187268733978271} 02/25/2022 18:57:23 - INFO - codeparrot_training - Step 31951: {'lr': 0.0001550703628632966, 'samples': 16359424, 'steps': 31951, 'loss/train': 1.9003520011901855} 02/25/2022 18:57:30 - INFO - codeparrot_training - Step 31952: {'lr': 0.00015505522611954976, 'samples': 16359936, 'steps': 31952, 'loss/train': 1.767127275466919} 02/25/2022 18:57:34 - INFO - codeparrot_training - Step 31953: {'lr': 0.00015504008978251614, 'samples': 16360448, 'steps': 31953, 'loss/train': 0.5519925355911255} 02/25/2022 18:57:39 - INFO - codeparrot_training - Step 31954: {'lr': 0.00015502495385226066, 'samples': 16360960, 'steps': 31954, 'loss/train': 2.0983614921569824} 02/25/2022 18:57:43 - INFO - codeparrot_training - Step 31955: {'lr': 0.00015500981832884819, 'samples': 16361472, 'steps': 31955, 'loss/train': 1.8321764469146729} 02/25/2022 18:57:48 - INFO - codeparrot_training - Step 31956: {'lr': 0.00015499468321234343, 'samples': 16361984, 'steps': 31956, 'loss/train': 1.2970038652420044} 02/25/2022 18:57:52 - INFO - codeparrot_training - Step 31957: {'lr': 0.00015497954850281125, 'samples': 16362496, 'steps': 31957, 'loss/train': 2.4311492443084717} 02/25/2022 18:57:57 - INFO - codeparrot_training - Step 31958: {'lr': 0.00015496441420031652, 'samples': 16363008, 'steps': 31958, 'loss/train': 1.999210238456726} 02/25/2022 18:58:00 - INFO - codeparrot_training - Step 31959: {'lr': 0.00015494928030492417, 'samples': 16363520, 'steps': 31959, 'loss/train': 0.7683048248291016} 02/25/2022 18:58:06 - INFO - codeparrot_training - Step 31960: {'lr': 0.00015493414681669882, 'samples': 16364032, 'steps': 31960, 'loss/train': 1.3272879123687744} 02/25/2022 18:58:10 - INFO - codeparrot_training - Step 31961: {'lr': 0.0001549190137357054, 'samples': 16364544, 'steps': 31961, 'loss/train': 1.0985316038131714} 02/25/2022 18:58:15 - INFO - codeparrot_training - Step 31962: {'lr': 0.0001549038810620087, 'samples': 16365056, 'steps': 31962, 'loss/train': 1.879543662071228} 02/25/2022 18:58:19 - INFO - codeparrot_training - Step 31963: {'lr': 0.0001548887487956737, 'samples': 16365568, 'steps': 31963, 'loss/train': 1.8775640726089478} 02/25/2022 18:58:24 - INFO - codeparrot_training - Step 31964: {'lr': 0.000154873616936765, 'samples': 16366080, 'steps': 31964, 'loss/train': 2.780747175216675} 02/25/2022 18:58:27 - INFO - codeparrot_training - Step 31965: {'lr': 0.00015485848548534752, 'samples': 16366592, 'steps': 31965, 'loss/train': 1.7711981534957886} 02/25/2022 18:58:33 - INFO - codeparrot_training - Step 31966: {'lr': 0.00015484335444148606, 'samples': 16367104, 'steps': 31966, 'loss/train': 2.0191216468811035} 02/25/2022 18:58:36 - INFO - codeparrot_training - Step 31967: {'lr': 0.00015482822380524547, 'samples': 16367616, 'steps': 31967, 'loss/train': 1.8072110414505005} 02/25/2022 18:58:43 - INFO - codeparrot_training - Step 31968: {'lr': 0.00015481309357669065, 'samples': 16368128, 'steps': 31968, 'loss/train': 1.5220261812210083} 02/25/2022 18:58:46 - INFO - codeparrot_training - Step 31969: {'lr': 0.00015479796375588624, 'samples': 16368640, 'steps': 31969, 'loss/train': 2.519467353820801} 02/25/2022 18:58:52 - INFO - codeparrot_training - Step 31970: {'lr': 0.00015478283434289708, 'samples': 16369152, 'steps': 31970, 'loss/train': 1.261850118637085} 02/25/2022 18:58:55 - INFO - codeparrot_training - Step 31971: {'lr': 0.00015476770533778806, 'samples': 16369664, 'steps': 31971, 'loss/train': 1.8854120969772339} 02/25/2022 18:59:00 - INFO - codeparrot_training - Step 31972: {'lr': 0.000154752576740624, 'samples': 16370176, 'steps': 31972, 'loss/train': 1.5968600511550903} 02/25/2022 18:59:04 - INFO - codeparrot_training - Step 31973: {'lr': 0.0001547374485514696, 'samples': 16370688, 'steps': 31973, 'loss/train': 0.3152819871902466} 02/25/2022 18:59:10 - INFO - codeparrot_training - Step 31974: {'lr': 0.00015472232077038973, 'samples': 16371200, 'steps': 31974, 'loss/train': 1.1850675344467163} 02/25/2022 18:59:13 - INFO - codeparrot_training - Step 31975: {'lr': 0.00015470719339744917, 'samples': 16371712, 'steps': 31975, 'loss/train': 1.9640557765960693} 02/25/2022 18:59:18 - INFO - codeparrot_training - Step 31976: {'lr': 0.00015469206643271274, 'samples': 16372224, 'steps': 31976, 'loss/train': 2.8242547512054443} 02/25/2022 18:59:22 - INFO - codeparrot_training - Step 31977: {'lr': 0.00015467693987624527, 'samples': 16372736, 'steps': 31977, 'loss/train': 1.5759389400482178} 02/25/2022 18:59:28 - INFO - codeparrot_training - Step 31978: {'lr': 0.00015466181372811148, 'samples': 16373248, 'steps': 31978, 'loss/train': 1.1964836120605469} 02/25/2022 18:59:32 - INFO - codeparrot_training - Step 31979: {'lr': 0.0001546466879883762, 'samples': 16373760, 'steps': 31979, 'loss/train': 1.551389217376709} 02/25/2022 18:59:37 - INFO - codeparrot_training - Step 31980: {'lr': 0.00015463156265710428, 'samples': 16374272, 'steps': 31980, 'loss/train': 1.5725407600402832} 02/25/2022 18:59:41 - INFO - codeparrot_training - Step 31981: {'lr': 0.00015461643773436045, 'samples': 16374784, 'steps': 31981, 'loss/train': 1.3243732452392578} 02/25/2022 18:59:46 - INFO - codeparrot_training - Step 31982: {'lr': 0.00015460131322020954, 'samples': 16375296, 'steps': 31982, 'loss/train': 0.8847343325614929} 02/25/2022 18:59:50 - INFO - codeparrot_training - Step 31983: {'lr': 0.00015458618911471622, 'samples': 16375808, 'steps': 31983, 'loss/train': 1.9870924949645996} 02/25/2022 18:59:55 - INFO - codeparrot_training - Step 31984: {'lr': 0.00015457106541794541, 'samples': 16376320, 'steps': 31984, 'loss/train': 1.9988124370574951} 02/25/2022 18:59:59 - INFO - codeparrot_training - Step 31985: {'lr': 0.00015455594212996187, 'samples': 16376832, 'steps': 31985, 'loss/train': 3.596651077270508} 02/25/2022 19:00:04 - INFO - codeparrot_training - Step 31986: {'lr': 0.0001545408192508304, 'samples': 16377344, 'steps': 31986, 'loss/train': 1.3388316631317139} 02/25/2022 19:00:08 - INFO - codeparrot_training - Step 31987: {'lr': 0.00015452569678061568, 'samples': 16377856, 'steps': 31987, 'loss/train': 2.265784978866577} 02/25/2022 19:00:14 - INFO - codeparrot_training - Step 31988: {'lr': 0.00015451057471938258, 'samples': 16378368, 'steps': 31988, 'loss/train': 1.4022538661956787} 02/25/2022 19:00:18 - INFO - codeparrot_training - Step 31989: {'lr': 0.0001544954530671958, 'samples': 16378880, 'steps': 31989, 'loss/train': 0.9224123954772949} 02/25/2022 19:00:23 - INFO - codeparrot_training - Step 31990: {'lr': 0.0001544803318241203, 'samples': 16379392, 'steps': 31990, 'loss/train': 1.9672799110412598} 02/25/2022 19:00:27 - INFO - codeparrot_training - Step 31991: {'lr': 0.00015446521099022066, 'samples': 16379904, 'steps': 31991, 'loss/train': 2.0309693813323975} 02/25/2022 19:00:32 - INFO - codeparrot_training - Step 31992: {'lr': 0.00015445009056556168, 'samples': 16380416, 'steps': 31992, 'loss/train': 2.064556360244751} 02/25/2022 19:00:36 - INFO - codeparrot_training - Step 31993: {'lr': 0.00015443497055020821, 'samples': 16380928, 'steps': 31993, 'loss/train': 1.3361923694610596} 02/25/2022 19:00:41 - INFO - codeparrot_training - Step 31994: {'lr': 0.00015441985094422506, 'samples': 16381440, 'steps': 31994, 'loss/train': 2.0053908824920654} 02/25/2022 19:00:45 - INFO - codeparrot_training - Step 31995: {'lr': 0.00015440473174767682, 'samples': 16381952, 'steps': 31995, 'loss/train': 0.2217193841934204} 02/25/2022 19:00:50 - INFO - codeparrot_training - Step 31996: {'lr': 0.00015438961296062833, 'samples': 16382464, 'steps': 31996, 'loss/train': 1.7954959869384766} 02/25/2022 19:00:54 - INFO - codeparrot_training - Step 31997: {'lr': 0.0001543744945831444, 'samples': 16382976, 'steps': 31997, 'loss/train': 1.6838881969451904} 02/25/2022 19:00:59 - INFO - codeparrot_training - Step 31998: {'lr': 0.0001543593766152899, 'samples': 16383488, 'steps': 31998, 'loss/train': 1.7603875398635864} 02/25/2022 19:01:03 - INFO - codeparrot_training - Step 31999: {'lr': 0.00015434425905712934, 'samples': 16384000, 'steps': 31999, 'loss/train': 1.2924124002456665} 02/25/2022 19:01:03 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 19:01:20 - WARNING - huggingface_hub.repository - Several commits (32) will be pushed upstream. 02/25/2022 19:01:20 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 19:01:55 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 59d2b90..06c1d4e floral-grass-11 -> floral-grass-11 02/25/2022 19:02:02 - INFO - codeparrot_training - Step 32000: {'lr': 0.00015432914190872756, 'samples': 16384512, 'steps': 32000, 'loss/train': 0.3606870770454407} 02/25/2022 19:02:05 - INFO - codeparrot_training - Step 32001: {'lr': 0.00015431402517014943, 'samples': 16385024, 'steps': 32001, 'loss/train': 1.5795643329620361} 02/25/2022 19:02:11 - INFO - codeparrot_training - Step 32002: {'lr': 0.0001542989088414596, 'samples': 16385536, 'steps': 32002, 'loss/train': 1.7262461185455322} 02/25/2022 19:02:14 - INFO - codeparrot_training - Step 32003: {'lr': 0.00015428379292272294, 'samples': 16386048, 'steps': 32003, 'loss/train': 0.9702414870262146} 02/25/2022 19:02:21 - INFO - codeparrot_training - Step 32004: {'lr': 0.000154268677414004, 'samples': 16386560, 'steps': 32004, 'loss/train': 3.348331928253174} 02/25/2022 19:02:24 - INFO - codeparrot_training - Step 32005: {'lr': 0.0001542535623153677, 'samples': 16387072, 'steps': 32005, 'loss/train': 0.5520191192626953} 02/25/2022 19:02:30 - INFO - codeparrot_training - Step 32006: {'lr': 0.0001542384476268787, 'samples': 16387584, 'steps': 32006, 'loss/train': 1.9785206317901611} 02/25/2022 19:02:33 - INFO - codeparrot_training - Step 32007: {'lr': 0.00015422333334860189, 'samples': 16388096, 'steps': 32007, 'loss/train': 1.1689119338989258} 02/25/2022 19:02:39 - INFO - codeparrot_training - Step 32008: {'lr': 0.00015420821948060182, 'samples': 16388608, 'steps': 32008, 'loss/train': 0.2037752866744995} 02/25/2022 19:02:42 - INFO - codeparrot_training - Step 32009: {'lr': 0.00015419310602294328, 'samples': 16389120, 'steps': 32009, 'loss/train': 1.7252893447875977} 02/25/2022 19:02:48 - INFO - codeparrot_training - Step 32010: {'lr': 0.0001541779929756911, 'samples': 16389632, 'steps': 32010, 'loss/train': 1.2176986932754517} 02/25/2022 19:02:51 - INFO - codeparrot_training - Step 32011: {'lr': 0.00015416288033891007, 'samples': 16390144, 'steps': 32011, 'loss/train': 1.7501472234725952} 02/25/2022 19:02:57 - INFO - codeparrot_training - Step 32012: {'lr': 0.00015414776811266471, 'samples': 16390656, 'steps': 32012, 'loss/train': 1.0003036260604858} 02/25/2022 19:03:00 - INFO - codeparrot_training - Step 32013: {'lr': 0.0001541326562970199, 'samples': 16391168, 'steps': 32013, 'loss/train': 2.307699680328369} 02/25/2022 19:03:07 - INFO - codeparrot_training - Step 32014: {'lr': 0.0001541175448920403, 'samples': 16391680, 'steps': 32014, 'loss/train': 1.3219987154006958} 02/25/2022 19:03:10 - INFO - codeparrot_training - Step 32015: {'lr': 0.00015410243389779078, 'samples': 16392192, 'steps': 32015, 'loss/train': 0.480421781539917} 02/25/2022 19:03:16 - INFO - codeparrot_training - Step 32016: {'lr': 0.00015408732331433596, 'samples': 16392704, 'steps': 32016, 'loss/train': 2.009598731994629} 02/25/2022 19:03:19 - INFO - codeparrot_training - Step 32017: {'lr': 0.00015407221314174056, 'samples': 16393216, 'steps': 32017, 'loss/train': 2.344987392425537} 02/25/2022 19:03:25 - INFO - codeparrot_training - Step 32018: {'lr': 0.00015405710338006935, 'samples': 16393728, 'steps': 32018, 'loss/train': 1.5353116989135742} 02/25/2022 19:03:28 - INFO - codeparrot_training - Step 32019: {'lr': 0.00015404199402938707, 'samples': 16394240, 'steps': 32019, 'loss/train': 1.5851848125457764} 02/25/2022 19:03:34 - INFO - codeparrot_training - Step 32020: {'lr': 0.0001540268850897584, 'samples': 16394752, 'steps': 32020, 'loss/train': 2.090320348739624} 02/25/2022 19:03:37 - INFO - codeparrot_training - Step 32021: {'lr': 0.00015401177656124804, 'samples': 16395264, 'steps': 32021, 'loss/train': 1.3740935325622559} 02/25/2022 19:03:43 - INFO - codeparrot_training - Step 32022: {'lr': 0.00015399666844392081, 'samples': 16395776, 'steps': 32022, 'loss/train': 2.2387239933013916} 02/25/2022 19:03:46 - INFO - codeparrot_training - Step 32023: {'lr': 0.00015398156073784133, 'samples': 16396288, 'steps': 32023, 'loss/train': 1.4275959730148315} 02/25/2022 19:03:53 - INFO - codeparrot_training - Step 32024: {'lr': 0.00015396645344307438, 'samples': 16396800, 'steps': 32024, 'loss/train': 1.3657249212265015} 02/25/2022 19:03:56 - INFO - codeparrot_training - Step 32025: {'lr': 0.0001539513465596846, 'samples': 16397312, 'steps': 32025, 'loss/train': 1.8009713888168335} 02/25/2022 19:04:01 - INFO - codeparrot_training - Step 32026: {'lr': 0.00015393624008773685, 'samples': 16397824, 'steps': 32026, 'loss/train': 1.7081345319747925} 02/25/2022 19:04:05 - INFO - codeparrot_training - Step 32027: {'lr': 0.00015392113402729567, 'samples': 16398336, 'steps': 32027, 'loss/train': 1.5810519456863403} 02/25/2022 19:04:11 - INFO - codeparrot_training - Step 32028: {'lr': 0.00015390602837842582, 'samples': 16398848, 'steps': 32028, 'loss/train': 1.6448781490325928} 02/25/2022 19:04:14 - INFO - codeparrot_training - Step 32029: {'lr': 0.00015389092314119213, 'samples': 16399360, 'steps': 32029, 'loss/train': 1.7197331190109253} 02/25/2022 19:04:20 - INFO - codeparrot_training - Step 32030: {'lr': 0.00015387581831565912, 'samples': 16399872, 'steps': 32030, 'loss/train': 1.612053632736206} 02/25/2022 19:04:23 - INFO - codeparrot_training - Step 32031: {'lr': 0.0001538607139018916, 'samples': 16400384, 'steps': 32031, 'loss/train': 1.8617560863494873} 02/25/2022 19:04:29 - INFO - codeparrot_training - Step 32032: {'lr': 0.00015384560989995422, 'samples': 16400896, 'steps': 32032, 'loss/train': 0.7518028616905212} 02/25/2022 19:04:32 - INFO - codeparrot_training - Step 32033: {'lr': 0.00015383050630991187, 'samples': 16401408, 'steps': 32033, 'loss/train': 0.9034668207168579} 02/25/2022 19:04:38 - INFO - codeparrot_training - Step 32034: {'lr': 0.00015381540313182893, 'samples': 16401920, 'steps': 32034, 'loss/train': 2.0993776321411133} 02/25/2022 19:04:42 - INFO - codeparrot_training - Step 32035: {'lr': 0.00015380030036577029, 'samples': 16402432, 'steps': 32035, 'loss/train': 2.0694406032562256} 02/25/2022 19:04:47 - INFO - codeparrot_training - Step 32036: {'lr': 0.0001537851980118006, 'samples': 16402944, 'steps': 32036, 'loss/train': 0.4352928400039673} 02/25/2022 19:04:53 - INFO - codeparrot_training - Step 32037: {'lr': 0.0001537700960699846, 'samples': 16403456, 'steps': 32037, 'loss/train': 1.75059974193573} 02/25/2022 19:04:56 - INFO - codeparrot_training - Step 32038: {'lr': 0.00015375499454038703, 'samples': 16403968, 'steps': 32038, 'loss/train': 1.6791043281555176} 02/25/2022 19:05:02 - INFO - codeparrot_training - Step 32039: {'lr': 0.00015373989342307238, 'samples': 16404480, 'steps': 32039, 'loss/train': 0.8867302536964417} 02/25/2022 19:05:05 - INFO - codeparrot_training - Step 32040: {'lr': 0.0001537247927181055, 'samples': 16404992, 'steps': 32040, 'loss/train': 1.4566854238510132} 02/25/2022 19:05:11 - INFO - codeparrot_training - Step 32041: {'lr': 0.00015370969242555103, 'samples': 16405504, 'steps': 32041, 'loss/train': 1.7527397871017456} 02/25/2022 19:05:14 - INFO - codeparrot_training - Step 32042: {'lr': 0.00015369459254547374, 'samples': 16406016, 'steps': 32042, 'loss/train': 1.806497573852539} 02/25/2022 19:05:19 - INFO - codeparrot_training - Step 32043: {'lr': 0.00015367949307793817, 'samples': 16406528, 'steps': 32043, 'loss/train': 2.0428524017333984} 02/25/2022 19:05:23 - INFO - codeparrot_training - Step 32044: {'lr': 0.00015366439402300902, 'samples': 16407040, 'steps': 32044, 'loss/train': 1.1290767192840576} 02/25/2022 19:05:29 - INFO - codeparrot_training - Step 32045: {'lr': 0.00015364929538075106, 'samples': 16407552, 'steps': 32045, 'loss/train': 0.6243045926094055} 02/25/2022 19:05:32 - INFO - codeparrot_training - Step 32046: {'lr': 0.000153634197151229, 'samples': 16408064, 'steps': 32046, 'loss/train': 2.4429209232330322} 02/25/2022 19:05:38 - INFO - codeparrot_training - Step 32047: {'lr': 0.00015361909933450736, 'samples': 16408576, 'steps': 32047, 'loss/train': 1.0027731657028198} 02/25/2022 19:05:41 - INFO - codeparrot_training - Step 32048: {'lr': 0.00015360400193065087, 'samples': 16409088, 'steps': 32048, 'loss/train': 1.831935167312622} 02/25/2022 19:05:47 - INFO - codeparrot_training - Step 32049: {'lr': 0.00015358890493972425, 'samples': 16409600, 'steps': 32049, 'loss/train': 1.6197808980941772} 02/25/2022 19:05:50 - INFO - codeparrot_training - Step 32050: {'lr': 0.00015357380836179214, 'samples': 16410112, 'steps': 32050, 'loss/train': 1.7030571699142456} 02/25/2022 19:05:56 - INFO - codeparrot_training - Step 32051: {'lr': 0.00015355871219691931, 'samples': 16410624, 'steps': 32051, 'loss/train': 1.4633952379226685} 02/25/2022 19:06:00 - INFO - codeparrot_training - Step 32052: {'lr': 0.00015354361644517024, 'samples': 16411136, 'steps': 32052, 'loss/train': 1.4353697299957275} 02/25/2022 19:06:05 - INFO - codeparrot_training - Step 32053: {'lr': 0.00015352852110660966, 'samples': 16411648, 'steps': 32053, 'loss/train': 2.0888679027557373} 02/25/2022 19:06:09 - INFO - codeparrot_training - Step 32054: {'lr': 0.00015351342618130228, 'samples': 16412160, 'steps': 32054, 'loss/train': 2.393428087234497} 02/25/2022 19:06:14 - INFO - codeparrot_training - Step 32055: {'lr': 0.00015349833166931288, 'samples': 16412672, 'steps': 32055, 'loss/train': 2.9036004543304443} 02/25/2022 19:06:18 - INFO - codeparrot_training - Step 32056: {'lr': 0.00015348323757070583, 'samples': 16413184, 'steps': 32056, 'loss/train': 2.3367366790771484} 02/25/2022 19:06:23 - INFO - codeparrot_training - Step 32057: {'lr': 0.00015346814388554598, 'samples': 16413696, 'steps': 32057, 'loss/train': 1.2342102527618408} 02/25/2022 19:06:27 - INFO - codeparrot_training - Step 32058: {'lr': 0.0001534530506138979, 'samples': 16414208, 'steps': 32058, 'loss/train': 0.8562390804290771} 02/25/2022 19:06:32 - INFO - codeparrot_training - Step 32059: {'lr': 0.00015343795775582647, 'samples': 16414720, 'steps': 32059, 'loss/train': 1.606260895729065} 02/25/2022 19:06:36 - INFO - codeparrot_training - Step 32060: {'lr': 0.00015342286531139603, 'samples': 16415232, 'steps': 32060, 'loss/train': 1.2808316946029663} 02/25/2022 19:06:42 - INFO - codeparrot_training - Step 32061: {'lr': 0.00015340777328067134, 'samples': 16415744, 'steps': 32061, 'loss/train': 2.294900417327881} 02/25/2022 19:06:46 - INFO - codeparrot_training - Step 32062: {'lr': 0.00015339268166371717, 'samples': 16416256, 'steps': 32062, 'loss/train': 0.21338191628456116} 02/25/2022 19:06:51 - INFO - codeparrot_training - Step 32063: {'lr': 0.000153377590460598, 'samples': 16416768, 'steps': 32063, 'loss/train': 1.3271257877349854} 02/25/2022 19:06:54 - INFO - codeparrot_training - Step 32064: {'lr': 0.00015336249967137861, 'samples': 16417280, 'steps': 32064, 'loss/train': 1.2165783643722534} 02/25/2022 19:07:00 - INFO - codeparrot_training - Step 32065: {'lr': 0.00015334740929612357, 'samples': 16417792, 'steps': 32065, 'loss/train': 2.2226016521453857} 02/25/2022 19:07:04 - INFO - codeparrot_training - Step 32066: {'lr': 0.00015333231933489756, 'samples': 16418304, 'steps': 32066, 'loss/train': 0.8278035521507263} 02/25/2022 19:07:09 - INFO - codeparrot_training - Step 32067: {'lr': 0.00015331722978776513, 'samples': 16418816, 'steps': 32067, 'loss/train': 1.4431467056274414} 02/25/2022 19:07:12 - INFO - codeparrot_training - Step 32068: {'lr': 0.00015330214065479103, 'samples': 16419328, 'steps': 32068, 'loss/train': 2.0595474243164062} 02/25/2022 19:07:18 - INFO - codeparrot_training - Step 32069: {'lr': 0.00015328705193603986, 'samples': 16419840, 'steps': 32069, 'loss/train': 1.831860065460205} 02/25/2022 19:07:22 - INFO - codeparrot_training - Step 32070: {'lr': 0.00015327196363157624, 'samples': 16420352, 'steps': 32070, 'loss/train': 1.6886903047561646} 02/25/2022 19:07:28 - INFO - codeparrot_training - Step 32071: {'lr': 0.0001532568757414648, 'samples': 16420864, 'steps': 32071, 'loss/train': 2.956587791442871} 02/25/2022 19:07:31 - INFO - codeparrot_training - Step 32072: {'lr': 0.0001532417882657702, 'samples': 16421376, 'steps': 32072, 'loss/train': 1.4410006999969482} 02/25/2022 19:07:37 - INFO - codeparrot_training - Step 32073: {'lr': 0.00015322670120455705, 'samples': 16421888, 'steps': 32073, 'loss/train': 1.8332265615463257} 02/25/2022 19:07:40 - INFO - codeparrot_training - Step 32074: {'lr': 0.00015321161455789, 'samples': 16422400, 'steps': 32074, 'loss/train': 1.5969840288162231} 02/25/2022 19:07:46 - INFO - codeparrot_training - Step 32075: {'lr': 0.0001531965283258336, 'samples': 16422912, 'steps': 32075, 'loss/train': 1.5386989116668701} 02/25/2022 19:07:50 - INFO - codeparrot_training - Step 32076: {'lr': 0.00015318144250845255, 'samples': 16423424, 'steps': 32076, 'loss/train': 0.8107405304908752} 02/25/2022 19:07:55 - INFO - codeparrot_training - Step 32077: {'lr': 0.00015316635710581157, 'samples': 16423936, 'steps': 32077, 'loss/train': 2.2198052406311035} 02/25/2022 19:07:58 - INFO - codeparrot_training - Step 32078: {'lr': 0.00015315127211797507, 'samples': 16424448, 'steps': 32078, 'loss/train': 1.8126347064971924} 02/25/2022 19:08:04 - INFO - codeparrot_training - Step 32079: {'lr': 0.00015313618754500774, 'samples': 16424960, 'steps': 32079, 'loss/train': 2.4541759490966797} 02/25/2022 19:08:07 - INFO - codeparrot_training - Step 32080: {'lr': 0.00015312110338697426, 'samples': 16425472, 'steps': 32080, 'loss/train': 2.4049551486968994} 02/25/2022 19:08:13 - INFO - codeparrot_training - Step 32081: {'lr': 0.00015310601964393927, 'samples': 16425984, 'steps': 32081, 'loss/train': 1.7723838090896606} 02/25/2022 19:08:17 - INFO - codeparrot_training - Step 32082: {'lr': 0.00015309093631596726, 'samples': 16426496, 'steps': 32082, 'loss/train': 1.9514926671981812} 02/25/2022 19:08:22 - INFO - codeparrot_training - Step 32083: {'lr': 0.0001530758534031229, 'samples': 16427008, 'steps': 32083, 'loss/train': 0.7206658124923706} 02/25/2022 19:08:26 - INFO - codeparrot_training - Step 32084: {'lr': 0.00015306077090547078, 'samples': 16427520, 'steps': 32084, 'loss/train': 1.3232718706130981} 02/25/2022 19:08:32 - INFO - codeparrot_training - Step 32085: {'lr': 0.00015304568882307557, 'samples': 16428032, 'steps': 32085, 'loss/train': 2.8251843452453613} 02/25/2022 19:08:35 - INFO - codeparrot_training - Step 32086: {'lr': 0.00015303060715600192, 'samples': 16428544, 'steps': 32086, 'loss/train': 1.1703813076019287} 02/25/2022 19:08:41 - INFO - codeparrot_training - Step 32087: {'lr': 0.0001530155259043143, 'samples': 16429056, 'steps': 32087, 'loss/train': 2.422834873199463} 02/25/2022 19:08:44 - INFO - codeparrot_training - Step 32088: {'lr': 0.00015300044506807731, 'samples': 16429568, 'steps': 32088, 'loss/train': 0.5209026336669922} 02/25/2022 19:08:50 - INFO - codeparrot_training - Step 32089: {'lr': 0.00015298536464735562, 'samples': 16430080, 'steps': 32089, 'loss/train': 1.880300760269165} 02/25/2022 19:08:53 - INFO - codeparrot_training - Step 32090: {'lr': 0.00015297028464221395, 'samples': 16430592, 'steps': 32090, 'loss/train': 1.9999984502792358} 02/25/2022 19:08:59 - INFO - codeparrot_training - Step 32091: {'lr': 0.00015295520505271664, 'samples': 16431104, 'steps': 32091, 'loss/train': 1.5997023582458496} 02/25/2022 19:09:02 - INFO - codeparrot_training - Step 32092: {'lr': 0.00015294012587892847, 'samples': 16431616, 'steps': 32092, 'loss/train': 2.377359390258789} 02/25/2022 19:09:08 - INFO - codeparrot_training - Step 32093: {'lr': 0.00015292504712091393, 'samples': 16432128, 'steps': 32093, 'loss/train': 2.70540714263916} 02/25/2022 19:09:11 - INFO - codeparrot_training - Step 32094: {'lr': 0.0001529099687787378, 'samples': 16432640, 'steps': 32094, 'loss/train': 2.134967803955078} 02/25/2022 19:09:17 - INFO - codeparrot_training - Step 32095: {'lr': 0.0001528948908524644, 'samples': 16433152, 'steps': 32095, 'loss/train': 3.335530996322632} 02/25/2022 19:09:21 - INFO - codeparrot_training - Step 32096: {'lr': 0.00015287981334215851, 'samples': 16433664, 'steps': 32096, 'loss/train': 1.8269751071929932} 02/25/2022 19:09:26 - INFO - codeparrot_training - Step 32097: {'lr': 0.0001528647362478846, 'samples': 16434176, 'steps': 32097, 'loss/train': 1.2229468822479248} 02/25/2022 19:09:30 - INFO - codeparrot_training - Step 32098: {'lr': 0.00015284965956970732, 'samples': 16434688, 'steps': 32098, 'loss/train': 0.839073121547699} 02/25/2022 19:09:35 - INFO - codeparrot_training - Step 32099: {'lr': 0.00015283458330769137, 'samples': 16435200, 'steps': 32099, 'loss/train': 1.888695240020752} 02/25/2022 19:09:39 - INFO - codeparrot_training - Step 32100: {'lr': 0.0001528195074619011, 'samples': 16435712, 'steps': 32100, 'loss/train': 2.032506227493286} 02/25/2022 19:09:44 - INFO - codeparrot_training - Step 32101: {'lr': 0.00015280443203240124, 'samples': 16436224, 'steps': 32101, 'loss/train': 1.1756312847137451} 02/25/2022 19:09:48 - INFO - codeparrot_training - Step 32102: {'lr': 0.00015278935701925633, 'samples': 16436736, 'steps': 32102, 'loss/train': 2.283989667892456} 02/25/2022 19:09:53 - INFO - codeparrot_training - Step 32103: {'lr': 0.00015277428242253097, 'samples': 16437248, 'steps': 32103, 'loss/train': 1.487247347831726} 02/25/2022 19:09:57 - INFO - codeparrot_training - Step 32104: {'lr': 0.00015275920824228968, 'samples': 16437760, 'steps': 32104, 'loss/train': 0.944449782371521} 02/25/2022 19:10:02 - INFO - codeparrot_training - Step 32105: {'lr': 0.00015274413447859703, 'samples': 16438272, 'steps': 32105, 'loss/train': 1.5960766077041626} 02/25/2022 19:10:06 - INFO - codeparrot_training - Step 32106: {'lr': 0.00015272906113151768, 'samples': 16438784, 'steps': 32106, 'loss/train': 1.8329920768737793} 02/25/2022 19:10:12 - INFO - codeparrot_training - Step 32107: {'lr': 0.00015271398820111614, 'samples': 16439296, 'steps': 32107, 'loss/train': 1.9043269157409668} 02/25/2022 19:10:16 - INFO - codeparrot_training - Step 32108: {'lr': 0.00015269891568745698, 'samples': 16439808, 'steps': 32108, 'loss/train': 1.1749900579452515} 02/25/2022 19:10:21 - INFO - codeparrot_training - Step 32109: {'lr': 0.0001526838435906048, 'samples': 16440320, 'steps': 32109, 'loss/train': 1.8395562171936035} 02/25/2022 19:10:24 - INFO - codeparrot_training - Step 32110: {'lr': 0.000152668771910624, 'samples': 16440832, 'steps': 32110, 'loss/train': 1.3838906288146973} 02/25/2022 19:10:30 - INFO - codeparrot_training - Step 32111: {'lr': 0.0001526537006475794, 'samples': 16441344, 'steps': 32111, 'loss/train': 2.6590213775634766} 02/25/2022 19:10:33 - INFO - codeparrot_training - Step 32112: {'lr': 0.00015263862980153538, 'samples': 16441856, 'steps': 32112, 'loss/train': 0.9427378177642822} 02/25/2022 19:10:39 - INFO - codeparrot_training - Step 32113: {'lr': 0.0001526235593725566, 'samples': 16442368, 'steps': 32113, 'loss/train': 1.7967900037765503} 02/25/2022 19:10:45 - INFO - codeparrot_training - Step 32114: {'lr': 0.00015260848936070754, 'samples': 16442880, 'steps': 32114, 'loss/train': 1.6786161661148071} 02/25/2022 19:10:48 - INFO - codeparrot_training - Step 32115: {'lr': 0.00015259341976605275, 'samples': 16443392, 'steps': 32115, 'loss/train': 2.338987112045288} 02/25/2022 19:10:52 - INFO - codeparrot_training - Step 32116: {'lr': 0.0001525783505886569, 'samples': 16443904, 'steps': 32116, 'loss/train': 1.472335934638977} 02/25/2022 19:10:59 - INFO - codeparrot_training - Step 32117: {'lr': 0.00015256328182858446, 'samples': 16444416, 'steps': 32117, 'loss/train': 1.4607925415039062} 02/25/2022 19:11:02 - INFO - codeparrot_training - Step 32118: {'lr': 0.00015254821348589993, 'samples': 16444928, 'steps': 32118, 'loss/train': 1.111583948135376} 02/25/2022 19:11:08 - INFO - codeparrot_training - Step 32119: {'lr': 0.00015253314556066787, 'samples': 16445440, 'steps': 32119, 'loss/train': 1.794385552406311} 02/25/2022 19:11:11 - INFO - codeparrot_training - Step 32120: {'lr': 0.000152518078052953, 'samples': 16445952, 'steps': 32120, 'loss/train': 1.9882535934448242} 02/25/2022 19:11:17 - INFO - codeparrot_training - Step 32121: {'lr': 0.00015250301096281967, 'samples': 16446464, 'steps': 32121, 'loss/train': 8.773414611816406} 02/25/2022 19:11:20 - INFO - codeparrot_training - Step 32122: {'lr': 0.0001524879442903324, 'samples': 16446976, 'steps': 32122, 'loss/train': 2.1362810134887695} 02/25/2022 19:11:24 - INFO - codeparrot_training - Step 32123: {'lr': 0.00015247287803555584, 'samples': 16447488, 'steps': 32123, 'loss/train': 0.4721432328224182} 02/25/2022 19:11:29 - INFO - codeparrot_training - Step 32124: {'lr': 0.00015245781219855453, 'samples': 16448000, 'steps': 32124, 'loss/train': 0.6360771059989929} 02/25/2022 19:11:35 - INFO - codeparrot_training - Step 32125: {'lr': 0.00015244274677939307, 'samples': 16448512, 'steps': 32125, 'loss/train': 1.5727765560150146} 02/25/2022 19:11:38 - INFO - codeparrot_training - Step 32126: {'lr': 0.00015242768177813577, 'samples': 16449024, 'steps': 32126, 'loss/train': 2.066434621810913} 02/25/2022 19:11:45 - INFO - codeparrot_training - Step 32127: {'lr': 0.00015241261719484733, 'samples': 16449536, 'steps': 32127, 'loss/train': 1.9476970434188843} 02/25/2022 19:11:48 - INFO - codeparrot_training - Step 32128: {'lr': 0.00015239755302959225, 'samples': 16450048, 'steps': 32128, 'loss/train': 1.1157804727554321} 02/25/2022 19:11:53 - INFO - codeparrot_training - Step 32129: {'lr': 0.00015238248928243516, 'samples': 16450560, 'steps': 32129, 'loss/train': 1.0420268774032593} 02/25/2022 19:11:57 - INFO - codeparrot_training - Step 32130: {'lr': 0.0001523674259534404, 'samples': 16451072, 'steps': 32130, 'loss/train': 1.4700485467910767} 02/25/2022 19:12:02 - INFO - codeparrot_training - Step 32131: {'lr': 0.00015235236304267255, 'samples': 16451584, 'steps': 32131, 'loss/train': 1.53701651096344} 02/25/2022 19:12:06 - INFO - codeparrot_training - Step 32132: {'lr': 0.00015233730055019617, 'samples': 16452096, 'steps': 32132, 'loss/train': 1.4403636455535889} 02/25/2022 19:12:12 - INFO - codeparrot_training - Step 32133: {'lr': 0.00015232223847607592, 'samples': 16452608, 'steps': 32133, 'loss/train': 2.6908533573150635} 02/25/2022 19:12:15 - INFO - codeparrot_training - Step 32134: {'lr': 0.0001523071768203761, 'samples': 16453120, 'steps': 32134, 'loss/train': 1.4578200578689575} 02/25/2022 19:12:21 - INFO - codeparrot_training - Step 32135: {'lr': 0.00015229211558316126, 'samples': 16453632, 'steps': 32135, 'loss/train': 1.8918706178665161} 02/25/2022 19:12:24 - INFO - codeparrot_training - Step 32136: {'lr': 0.000152277054764496, 'samples': 16454144, 'steps': 32136, 'loss/train': 1.830186128616333} 02/25/2022 19:12:30 - INFO - codeparrot_training - Step 32137: {'lr': 0.00015226199436444476, 'samples': 16454656, 'steps': 32137, 'loss/train': 1.0561946630477905} 02/25/2022 19:12:33 - INFO - codeparrot_training - Step 32138: {'lr': 0.00015224693438307228, 'samples': 16455168, 'steps': 32138, 'loss/train': 0.8843085765838623} 02/25/2022 19:12:39 - INFO - codeparrot_training - Step 32139: {'lr': 0.00015223187482044274, 'samples': 16455680, 'steps': 32139, 'loss/train': 3.4875128269195557} 02/25/2022 19:12:42 - INFO - codeparrot_training - Step 32140: {'lr': 0.00015221681567662083, 'samples': 16456192, 'steps': 32140, 'loss/train': 0.8410459756851196} 02/25/2022 19:12:48 - INFO - codeparrot_training - Step 32141: {'lr': 0.000152201756951671, 'samples': 16456704, 'steps': 32141, 'loss/train': 1.2555756568908691} 02/25/2022 19:12:51 - INFO - codeparrot_training - Step 32142: {'lr': 0.00015218669864565794, 'samples': 16457216, 'steps': 32142, 'loss/train': 1.162388801574707} 02/25/2022 19:12:58 - INFO - codeparrot_training - Step 32143: {'lr': 0.00015217164075864584, 'samples': 16457728, 'steps': 32143, 'loss/train': 1.9992378950119019} 02/25/2022 19:13:01 - INFO - codeparrot_training - Step 32144: {'lr': 0.0001521565832906994, 'samples': 16458240, 'steps': 32144, 'loss/train': 0.14400219917297363} 02/25/2022 19:13:07 - INFO - codeparrot_training - Step 32145: {'lr': 0.0001521415262418831, 'samples': 16458752, 'steps': 32145, 'loss/train': 1.5225850343704224} 02/25/2022 19:13:10 - INFO - codeparrot_training - Step 32146: {'lr': 0.00015212646961226144, 'samples': 16459264, 'steps': 32146, 'loss/train': 1.5470852851867676} 02/25/2022 19:13:16 - INFO - codeparrot_training - Step 32147: {'lr': 0.00015211141340189894, 'samples': 16459776, 'steps': 32147, 'loss/train': 1.4106649160385132} 02/25/2022 19:13:19 - INFO - codeparrot_training - Step 32148: {'lr': 0.00015209635761085998, 'samples': 16460288, 'steps': 32148, 'loss/train': 2.2167809009552} 02/25/2022 19:13:24 - INFO - codeparrot_training - Step 32149: {'lr': 0.00015208130223920923, 'samples': 16460800, 'steps': 32149, 'loss/train': 1.9636964797973633} 02/25/2022 19:13:28 - INFO - codeparrot_training - Step 32150: {'lr': 0.000152066247287011, 'samples': 16461312, 'steps': 32150, 'loss/train': 1.0256893634796143} 02/25/2022 19:13:33 - INFO - codeparrot_training - Step 32151: {'lr': 0.00015205119275432994, 'samples': 16461824, 'steps': 32151, 'loss/train': 4.954133033752441} 02/25/2022 19:13:37 - INFO - codeparrot_training - Step 32152: {'lr': 0.00015203613864123042, 'samples': 16462336, 'steps': 32152, 'loss/train': 1.060043215751648} 02/25/2022 19:13:43 - INFO - codeparrot_training - Step 32153: {'lr': 0.00015202108494777702, 'samples': 16462848, 'steps': 32153, 'loss/train': 0.5198764801025391} 02/25/2022 19:13:47 - INFO - codeparrot_training - Step 32154: {'lr': 0.00015200603167403416, 'samples': 16463360, 'steps': 32154, 'loss/train': 2.2902448177337646} 02/25/2022 19:13:52 - INFO - codeparrot_training - Step 32155: {'lr': 0.00015199097882006635, 'samples': 16463872, 'steps': 32155, 'loss/train': 2.0287246704101562} 02/25/2022 19:13:56 - INFO - codeparrot_training - Step 32156: {'lr': 0.000151975926385938, 'samples': 16464384, 'steps': 32156, 'loss/train': 2.140232563018799} 02/25/2022 19:14:01 - INFO - codeparrot_training - Step 32157: {'lr': 0.00015196087437171373, 'samples': 16464896, 'steps': 32157, 'loss/train': 2.3248045444488525} 02/25/2022 19:14:05 - INFO - codeparrot_training - Step 32158: {'lr': 0.0001519458227774579, 'samples': 16465408, 'steps': 32158, 'loss/train': 0.9733873605728149} 02/25/2022 19:14:10 - INFO - codeparrot_training - Step 32159: {'lr': 0.00015193077160323505, 'samples': 16465920, 'steps': 32159, 'loss/train': 2.111279249191284} 02/25/2022 19:14:14 - INFO - codeparrot_training - Step 32160: {'lr': 0.0001519157208491097, 'samples': 16466432, 'steps': 32160, 'loss/train': 0.050616730004549026} 02/25/2022 19:14:19 - INFO - codeparrot_training - Step 32161: {'lr': 0.00015190067051514618, 'samples': 16466944, 'steps': 32161, 'loss/train': 2.2860357761383057} 02/25/2022 19:14:23 - INFO - codeparrot_training - Step 32162: {'lr': 0.000151885620601409, 'samples': 16467456, 'steps': 32162, 'loss/train': 0.4523526132106781} 02/25/2022 19:14:29 - INFO - codeparrot_training - Step 32163: {'lr': 0.0001518705711079627, 'samples': 16467968, 'steps': 32163, 'loss/train': 3.2783427238464355} 02/25/2022 19:14:32 - INFO - codeparrot_training - Step 32164: {'lr': 0.0001518555220348718, 'samples': 16468480, 'steps': 32164, 'loss/train': 2.176258087158203} 02/25/2022 19:14:38 - INFO - codeparrot_training - Step 32165: {'lr': 0.00015184047338220058, 'samples': 16468992, 'steps': 32165, 'loss/train': 0.49233490228652954} 02/25/2022 19:14:41 - INFO - codeparrot_training - Step 32166: {'lr': 0.00015182542515001363, 'samples': 16469504, 'steps': 32166, 'loss/train': 0.6116815209388733} 02/25/2022 19:14:47 - INFO - codeparrot_training - Step 32167: {'lr': 0.00015181037733837533, 'samples': 16470016, 'steps': 32167, 'loss/train': 1.5458869934082031} 02/25/2022 19:14:50 - INFO - codeparrot_training - Step 32168: {'lr': 0.00015179532994735034, 'samples': 16470528, 'steps': 32168, 'loss/train': 1.0361545085906982} 02/25/2022 19:14:56 - INFO - codeparrot_training - Step 32169: {'lr': 0.00015178028297700285, 'samples': 16471040, 'steps': 32169, 'loss/train': 2.213805675506592} 02/25/2022 19:14:59 - INFO - codeparrot_training - Step 32170: {'lr': 0.00015176523642739746, 'samples': 16471552, 'steps': 32170, 'loss/train': 0.6268891096115112} 02/25/2022 19:15:05 - INFO - codeparrot_training - Step 32171: {'lr': 0.0001517501902985986, 'samples': 16472064, 'steps': 32171, 'loss/train': 1.7239582538604736} 02/25/2022 19:15:08 - INFO - codeparrot_training - Step 32172: {'lr': 0.00015173514459067072, 'samples': 16472576, 'steps': 32172, 'loss/train': 1.4961861371994019} 02/25/2022 19:15:14 - INFO - codeparrot_training - Step 32173: {'lr': 0.00015172009930367836, 'samples': 16473088, 'steps': 32173, 'loss/train': 2.1971781253814697} 02/25/2022 19:15:17 - INFO - codeparrot_training - Step 32174: {'lr': 0.0001517050544376858, 'samples': 16473600, 'steps': 32174, 'loss/train': 1.9610031843185425} 02/25/2022 19:15:23 - INFO - codeparrot_training - Step 32175: {'lr': 0.00015169000999275758, 'samples': 16474112, 'steps': 32175, 'loss/train': 0.16922281682491302} 02/25/2022 19:15:26 - INFO - codeparrot_training - Step 32176: {'lr': 0.00015167496596895814, 'samples': 16474624, 'steps': 32176, 'loss/train': 1.3845646381378174} 02/25/2022 19:15:32 - INFO - codeparrot_training - Step 32177: {'lr': 0.00015165992236635203, 'samples': 16475136, 'steps': 32177, 'loss/train': 1.2828192710876465} 02/25/2022 19:15:35 - INFO - codeparrot_training - Step 32178: {'lr': 0.00015164487918500346, 'samples': 16475648, 'steps': 32178, 'loss/train': 1.4181867837905884} 02/25/2022 19:15:42 - INFO - codeparrot_training - Step 32179: {'lr': 0.00015162983642497702, 'samples': 16476160, 'steps': 32179, 'loss/train': 2.120661497116089} 02/25/2022 19:15:45 - INFO - codeparrot_training - Step 32180: {'lr': 0.00015161479408633713, 'samples': 16476672, 'steps': 32180, 'loss/train': 2.622908353805542} 02/25/2022 19:15:50 - INFO - codeparrot_training - Step 32181: {'lr': 0.00015159975216914832, 'samples': 16477184, 'steps': 32181, 'loss/train': 2.021345376968384} 02/25/2022 19:15:54 - INFO - codeparrot_training - Step 32182: {'lr': 0.0001515847106734748, 'samples': 16477696, 'steps': 32182, 'loss/train': 2.836031675338745} 02/25/2022 19:15:59 - INFO - codeparrot_training - Step 32183: {'lr': 0.00015156966959938118, 'samples': 16478208, 'steps': 32183, 'loss/train': 1.6484124660491943} 02/25/2022 19:16:03 - INFO - codeparrot_training - Step 32184: {'lr': 0.0001515546289469318, 'samples': 16478720, 'steps': 32184, 'loss/train': 1.2992713451385498} 02/25/2022 19:16:08 - INFO - codeparrot_training - Step 32185: {'lr': 0.0001515395887161911, 'samples': 16479232, 'steps': 32185, 'loss/train': 1.5771814584732056} 02/25/2022 19:16:14 - INFO - codeparrot_training - Step 32186: {'lr': 0.0001515245489072237, 'samples': 16479744, 'steps': 32186, 'loss/train': 1.7997910976409912} 02/25/2022 19:16:17 - INFO - codeparrot_training - Step 32187: {'lr': 0.00015150950952009374, 'samples': 16480256, 'steps': 32187, 'loss/train': 2.053393602371216} 02/25/2022 19:16:24 - INFO - codeparrot_training - Step 32188: {'lr': 0.00015149447055486575, 'samples': 16480768, 'steps': 32188, 'loss/train': 2.0865182876586914} 02/25/2022 19:16:27 - INFO - codeparrot_training - Step 32189: {'lr': 0.0001514794320116042, 'samples': 16481280, 'steps': 32189, 'loss/train': 0.32376283407211304} 02/25/2022 19:16:33 - INFO - codeparrot_training - Step 32190: {'lr': 0.00015146439389037355, 'samples': 16481792, 'steps': 32190, 'loss/train': 1.196940302848816} 02/25/2022 19:16:36 - INFO - codeparrot_training - Step 32191: {'lr': 0.00015144935619123805, 'samples': 16482304, 'steps': 32191, 'loss/train': 1.2933422327041626} 02/25/2022 19:16:42 - INFO - codeparrot_training - Step 32192: {'lr': 0.00015143431891426223, 'samples': 16482816, 'steps': 32192, 'loss/train': 2.626477003097534} 02/25/2022 19:16:45 - INFO - codeparrot_training - Step 32193: {'lr': 0.00015141928205951056, 'samples': 16483328, 'steps': 32193, 'loss/train': 3.6751291751861572} 02/25/2022 19:16:51 - INFO - codeparrot_training - Step 32194: {'lr': 0.0001514042456270473, 'samples': 16483840, 'steps': 32194, 'loss/train': 2.1206066608428955} 02/25/2022 19:16:54 - INFO - codeparrot_training - Step 32195: {'lr': 0.000151389209616937, 'samples': 16484352, 'steps': 32195, 'loss/train': 1.5654696226119995} 02/25/2022 19:17:00 - INFO - codeparrot_training - Step 32196: {'lr': 0.00015137417402924397, 'samples': 16484864, 'steps': 32196, 'loss/train': 2.389921188354492} 02/25/2022 19:17:03 - INFO - codeparrot_training - Step 32197: {'lr': 0.00015135913886403268, 'samples': 16485376, 'steps': 32197, 'loss/train': 2.205422878265381} 02/25/2022 19:17:09 - INFO - codeparrot_training - Step 32198: {'lr': 0.0001513441041213675, 'samples': 16485888, 'steps': 32198, 'loss/train': 1.6165269613265991} 02/25/2022 19:17:13 - INFO - codeparrot_training - Step 32199: {'lr': 0.0001513290698013129, 'samples': 16486400, 'steps': 32199, 'loss/train': 1.6551381349563599} 02/25/2022 19:17:18 - INFO - codeparrot_training - Step 32200: {'lr': 0.00015131403590393322, 'samples': 16486912, 'steps': 32200, 'loss/train': 2.567089557647705} 02/25/2022 19:17:22 - INFO - codeparrot_training - Step 32201: {'lr': 0.00015129900242929285, 'samples': 16487424, 'steps': 32201, 'loss/train': 2.299525499343872} 02/25/2022 19:17:27 - INFO - codeparrot_training - Step 32202: {'lr': 0.0001512839693774562, 'samples': 16487936, 'steps': 32202, 'loss/train': 1.3168189525604248} 02/25/2022 19:17:31 - INFO - codeparrot_training - Step 32203: {'lr': 0.00015126893674848773, 'samples': 16488448, 'steps': 32203, 'loss/train': 1.5388538837432861} 02/25/2022 19:17:36 - INFO - codeparrot_training - Step 32204: {'lr': 0.00015125390454245177, 'samples': 16488960, 'steps': 32204, 'loss/train': 0.9252519011497498} 02/25/2022 19:17:40 - INFO - codeparrot_training - Step 32205: {'lr': 0.00015123887275941266, 'samples': 16489472, 'steps': 32205, 'loss/train': 1.110875129699707} 02/25/2022 19:17:45 - INFO - codeparrot_training - Step 32206: {'lr': 0.0001512238413994349, 'samples': 16489984, 'steps': 32206, 'loss/train': 1.1377098560333252} 02/25/2022 19:17:49 - INFO - codeparrot_training - Step 32207: {'lr': 0.00015120881046258285, 'samples': 16490496, 'steps': 32207, 'loss/train': 2.8585071563720703} 02/25/2022 19:17:55 - INFO - codeparrot_training - Step 32208: {'lr': 0.00015119377994892095, 'samples': 16491008, 'steps': 32208, 'loss/train': 1.7140177488327026} 02/25/2022 19:17:59 - INFO - codeparrot_training - Step 32209: {'lr': 0.00015117874985851344, 'samples': 16491520, 'steps': 32209, 'loss/train': 1.2388315200805664} 02/25/2022 19:18:04 - INFO - codeparrot_training - Step 32210: {'lr': 0.00015116372019142478, 'samples': 16492032, 'steps': 32210, 'loss/train': 1.3309440612792969} 02/25/2022 19:18:08 - INFO - codeparrot_training - Step 32211: {'lr': 0.00015114869094771932, 'samples': 16492544, 'steps': 32211, 'loss/train': 1.8389467000961304} 02/25/2022 19:18:13 - INFO - codeparrot_training - Step 32212: {'lr': 0.00015113366212746166, 'samples': 16493056, 'steps': 32212, 'loss/train': 0.8844029307365417} 02/25/2022 19:18:17 - INFO - codeparrot_training - Step 32213: {'lr': 0.00015111863373071583, 'samples': 16493568, 'steps': 32213, 'loss/train': 0.40649211406707764} 02/25/2022 19:18:23 - INFO - codeparrot_training - Step 32214: {'lr': 0.0001511036057575464, 'samples': 16494080, 'steps': 32214, 'loss/train': 2.34197735786438} 02/25/2022 19:18:26 - INFO - codeparrot_training - Step 32215: {'lr': 0.00015108857820801772, 'samples': 16494592, 'steps': 32215, 'loss/train': 2.0781137943267822} 02/25/2022 19:18:32 - INFO - codeparrot_training - Step 32216: {'lr': 0.00015107355108219425, 'samples': 16495104, 'steps': 32216, 'loss/train': 1.3165700435638428} 02/25/2022 19:18:35 - INFO - codeparrot_training - Step 32217: {'lr': 0.0001510585243801402, 'samples': 16495616, 'steps': 32217, 'loss/train': 1.8521404266357422} 02/25/2022 19:18:41 - INFO - codeparrot_training - Step 32218: {'lr': 0.00015104349810192, 'samples': 16496128, 'steps': 32218, 'loss/train': 1.7540336847305298} 02/25/2022 19:18:44 - INFO - codeparrot_training - Step 32219: {'lr': 0.00015102847224759803, 'samples': 16496640, 'steps': 32219, 'loss/train': 1.5529921054840088} 02/25/2022 19:18:50 - INFO - codeparrot_training - Step 32220: {'lr': 0.00015101344681723867, 'samples': 16497152, 'steps': 32220, 'loss/train': 1.6578809022903442} 02/25/2022 19:18:53 - INFO - codeparrot_training - Step 32221: {'lr': 0.00015099842181090633, 'samples': 16497664, 'steps': 32221, 'loss/train': 1.8058582544326782} 02/25/2022 19:18:59 - INFO - codeparrot_training - Step 32222: {'lr': 0.00015098339722866529, 'samples': 16498176, 'steps': 32222, 'loss/train': 1.709394097328186} 02/25/2022 19:19:02 - INFO - codeparrot_training - Step 32223: {'lr': 0.00015096837307057988, 'samples': 16498688, 'steps': 32223, 'loss/train': 1.873930811882019} 02/25/2022 19:19:09 - INFO - codeparrot_training - Step 32224: {'lr': 0.0001509533493367145, 'samples': 16499200, 'steps': 32224, 'loss/train': 0.7976589798927307} 02/25/2022 19:19:12 - INFO - codeparrot_training - Step 32225: {'lr': 0.00015093832602713363, 'samples': 16499712, 'steps': 32225, 'loss/train': 2.253554582595825} 02/25/2022 19:19:18 - INFO - codeparrot_training - Step 32226: {'lr': 0.00015092330314190144, 'samples': 16500224, 'steps': 32226, 'loss/train': 2.6590070724487305} 02/25/2022 19:19:21 - INFO - codeparrot_training - Step 32227: {'lr': 0.00015090828068108238, 'samples': 16500736, 'steps': 32227, 'loss/train': 1.500365972518921} 02/25/2022 19:19:27 - INFO - codeparrot_training - Step 32228: {'lr': 0.00015089325864474075, 'samples': 16501248, 'steps': 32228, 'loss/train': 2.505763053894043} 02/25/2022 19:19:30 - INFO - codeparrot_training - Step 32229: {'lr': 0.000150878237032941, 'samples': 16501760, 'steps': 32229, 'loss/train': 2.8648841381073} 02/25/2022 19:19:36 - INFO - codeparrot_training - Step 32230: {'lr': 0.00015086321584574736, 'samples': 16502272, 'steps': 32230, 'loss/train': 2.0625882148742676} 02/25/2022 19:19:39 - INFO - codeparrot_training - Step 32231: {'lr': 0.0001508481950832242, 'samples': 16502784, 'steps': 32231, 'loss/train': 1.1973849534988403} 02/25/2022 19:19:45 - INFO - codeparrot_training - Step 32232: {'lr': 0.0001508331747454359, 'samples': 16503296, 'steps': 32232, 'loss/train': 1.9053893089294434} 02/25/2022 19:19:48 - INFO - codeparrot_training - Step 32233: {'lr': 0.00015081815483244682, 'samples': 16503808, 'steps': 32233, 'loss/train': 2.1886916160583496} 02/25/2022 19:19:54 - INFO - codeparrot_training - Step 32234: {'lr': 0.00015080313534432128, 'samples': 16504320, 'steps': 32234, 'loss/train': 1.807803750038147} 02/25/2022 19:19:58 - INFO - codeparrot_training - Step 32235: {'lr': 0.00015078811628112354, 'samples': 16504832, 'steps': 32235, 'loss/train': 1.881060004234314} 02/25/2022 19:20:03 - INFO - codeparrot_training - Step 32236: {'lr': 0.00015077309764291804, 'samples': 16505344, 'steps': 32236, 'loss/train': 2.407815933227539} 02/25/2022 19:20:06 - INFO - codeparrot_training - Step 32237: {'lr': 0.00015075807942976913, 'samples': 16505856, 'steps': 32237, 'loss/train': 1.120932936668396} 02/25/2022 19:20:12 - INFO - codeparrot_training - Step 32238: {'lr': 0.00015074306164174106, 'samples': 16506368, 'steps': 32238, 'loss/train': 2.388240337371826} 02/25/2022 19:20:16 - INFO - codeparrot_training - Step 32239: {'lr': 0.00015072804427889819, 'samples': 16506880, 'steps': 32239, 'loss/train': 2.1095659732818604} 02/25/2022 19:20:23 - INFO - codeparrot_training - Step 32240: {'lr': 0.00015071302734130488, 'samples': 16507392, 'steps': 32240, 'loss/train': 1.3532465696334839} 02/25/2022 19:20:26 - INFO - codeparrot_training - Step 32241: {'lr': 0.0001506980108290254, 'samples': 16507904, 'steps': 32241, 'loss/train': 0.46470528841018677} 02/25/2022 19:20:32 - INFO - codeparrot_training - Step 32242: {'lr': 0.0001506829947421241, 'samples': 16508416, 'steps': 32242, 'loss/train': 1.6249240636825562} 02/25/2022 19:20:35 - INFO - codeparrot_training - Step 32243: {'lr': 0.00015066797908066536, 'samples': 16508928, 'steps': 32243, 'loss/train': 1.2109366655349731} 02/25/2022 19:20:41 - INFO - codeparrot_training - Step 32244: {'lr': 0.00015065296384471344, 'samples': 16509440, 'steps': 32244, 'loss/train': 1.9693979024887085} 02/25/2022 19:20:44 - INFO - codeparrot_training - Step 32245: {'lr': 0.00015063794903433266, 'samples': 16509952, 'steps': 32245, 'loss/train': 1.5340776443481445} 02/25/2022 19:20:50 - INFO - codeparrot_training - Step 32246: {'lr': 0.00015062293464958735, 'samples': 16510464, 'steps': 32246, 'loss/train': 1.1609259843826294} 02/25/2022 19:20:53 - INFO - codeparrot_training - Step 32247: {'lr': 0.0001506079206905419, 'samples': 16510976, 'steps': 32247, 'loss/train': 1.5968750715255737} 02/25/2022 19:20:59 - INFO - codeparrot_training - Step 32248: {'lr': 0.00015059290715726048, 'samples': 16511488, 'steps': 32248, 'loss/train': 1.9594534635543823} 02/25/2022 19:21:02 - INFO - codeparrot_training - Step 32249: {'lr': 0.00015057789404980749, 'samples': 16512000, 'steps': 32249, 'loss/train': 0.9509637355804443} 02/25/2022 19:21:09 - INFO - codeparrot_training - Step 32250: {'lr': 0.00015056288136824724, 'samples': 16512512, 'steps': 32250, 'loss/train': 1.2636862993240356} 02/25/2022 19:21:13 - INFO - codeparrot_training - Step 32251: {'lr': 0.0001505478691126441, 'samples': 16513024, 'steps': 32251, 'loss/train': 0.9818252921104431} 02/25/2022 19:21:18 - INFO - codeparrot_training - Step 32252: {'lr': 0.00015053285728306224, 'samples': 16513536, 'steps': 32252, 'loss/train': 1.9109362363815308} 02/25/2022 19:21:22 - INFO - codeparrot_training - Step 32253: {'lr': 0.000150517845879566, 'samples': 16514048, 'steps': 32253, 'loss/train': 1.8378657102584839} 02/25/2022 19:21:27 - INFO - codeparrot_training - Step 32254: {'lr': 0.00015050283490221974, 'samples': 16514560, 'steps': 32254, 'loss/train': 1.2148056030273438} 02/25/2022 19:21:31 - INFO - codeparrot_training - Step 32255: {'lr': 0.00015048782435108776, 'samples': 16515072, 'steps': 32255, 'loss/train': 2.220336437225342} 02/25/2022 19:21:36 - INFO - codeparrot_training - Step 32256: {'lr': 0.00015047281422623441, 'samples': 16515584, 'steps': 32256, 'loss/train': 1.813839316368103} 02/25/2022 19:21:40 - INFO - codeparrot_training - Step 32257: {'lr': 0.00015045780452772385, 'samples': 16516096, 'steps': 32257, 'loss/train': 1.558762788772583} 02/25/2022 19:21:45 - INFO - codeparrot_training - Step 32258: {'lr': 0.0001504427952556204, 'samples': 16516608, 'steps': 32258, 'loss/train': 1.2582591772079468} 02/25/2022 19:21:49 - INFO - codeparrot_training - Step 32259: {'lr': 0.00015042778640998844, 'samples': 16517120, 'steps': 32259, 'loss/train': 2.369511365890503} 02/25/2022 19:21:55 - INFO - codeparrot_training - Step 32260: {'lr': 0.00015041277799089233, 'samples': 16517632, 'steps': 32260, 'loss/train': 1.5628235340118408} 02/25/2022 19:21:58 - INFO - codeparrot_training - Step 32261: {'lr': 0.00015039776999839613, 'samples': 16518144, 'steps': 32261, 'loss/train': 3.29669189453125} 02/25/2022 19:22:04 - INFO - codeparrot_training - Step 32262: {'lr': 0.00015038276243256428, 'samples': 16518656, 'steps': 32262, 'loss/train': 1.3060795068740845} 02/25/2022 19:22:07 - INFO - codeparrot_training - Step 32263: {'lr': 0.00015036775529346104, 'samples': 16519168, 'steps': 32263, 'loss/train': 1.205307126045227} 02/25/2022 19:22:13 - INFO - codeparrot_training - Step 32264: {'lr': 0.0001503527485811508, 'samples': 16519680, 'steps': 32264, 'loss/train': 2.0570287704467773} 02/25/2022 19:22:16 - INFO - codeparrot_training - Step 32265: {'lr': 0.0001503377422956976, 'samples': 16520192, 'steps': 32265, 'loss/train': 0.6859361529350281} 02/25/2022 19:22:22 - INFO - codeparrot_training - Step 32266: {'lr': 0.00015032273643716593, 'samples': 16520704, 'steps': 32266, 'loss/train': 1.232322096824646} 02/25/2022 19:22:25 - INFO - codeparrot_training - Step 32267: {'lr': 0.00015030773100561996, 'samples': 16521216, 'steps': 32267, 'loss/train': 2.1694436073303223} 02/25/2022 19:22:31 - INFO - codeparrot_training - Step 32268: {'lr': 0.00015029272600112402, 'samples': 16521728, 'steps': 32268, 'loss/train': 1.8544132709503174} 02/25/2022 19:22:34 - INFO - codeparrot_training - Step 32269: {'lr': 0.0001502777214237425, 'samples': 16522240, 'steps': 32269, 'loss/train': 1.9719080924987793} 02/25/2022 19:22:40 - INFO - codeparrot_training - Step 32270: {'lr': 0.00015026271727353947, 'samples': 16522752, 'steps': 32270, 'loss/train': 2.713430643081665} 02/25/2022 19:22:44 - INFO - codeparrot_training - Step 32271: {'lr': 0.00015024771355057925, 'samples': 16523264, 'steps': 32271, 'loss/train': 2.079793930053711} 02/25/2022 19:22:49 - INFO - codeparrot_training - Step 32272: {'lr': 0.0001502327102549262, 'samples': 16523776, 'steps': 32272, 'loss/train': 2.1581075191497803} 02/25/2022 19:22:53 - INFO - codeparrot_training - Step 32273: {'lr': 0.00015021770738664458, 'samples': 16524288, 'steps': 32273, 'loss/train': 1.2612062692642212} 02/25/2022 19:22:58 - INFO - codeparrot_training - Step 32274: {'lr': 0.00015020270494579854, 'samples': 16524800, 'steps': 32274, 'loss/train': 1.3858479261398315} 02/25/2022 19:23:02 - INFO - codeparrot_training - Step 32275: {'lr': 0.00015018770293245243, 'samples': 16525312, 'steps': 32275, 'loss/train': 0.9788023829460144} 02/25/2022 19:23:07 - INFO - codeparrot_training - Step 32276: {'lr': 0.0001501727013466705, 'samples': 16525824, 'steps': 32276, 'loss/train': 2.2804222106933594} 02/25/2022 19:23:11 - INFO - codeparrot_training - Step 32277: {'lr': 0.00015015770018851713, 'samples': 16526336, 'steps': 32277, 'loss/train': 1.738651990890503} 02/25/2022 19:23:16 - INFO - codeparrot_training - Step 32278: {'lr': 0.00015014269945805636, 'samples': 16526848, 'steps': 32278, 'loss/train': 0.4127237796783447} 02/25/2022 19:23:20 - INFO - codeparrot_training - Step 32279: {'lr': 0.00015012769915535257, 'samples': 16527360, 'steps': 32279, 'loss/train': 1.448751449584961} 02/25/2022 19:23:26 - INFO - codeparrot_training - Step 32280: {'lr': 0.00015011269928047002, 'samples': 16527872, 'steps': 32280, 'loss/train': 2.204979419708252} 02/25/2022 19:23:29 - INFO - codeparrot_training - Step 32281: {'lr': 0.00015009769983347293, 'samples': 16528384, 'steps': 32281, 'loss/train': 0.8856463432312012} 02/25/2022 19:23:35 - INFO - codeparrot_training - Step 32282: {'lr': 0.0001500827008144256, 'samples': 16528896, 'steps': 32282, 'loss/train': 2.540271043777466} 02/25/2022 19:23:38 - INFO - codeparrot_training - Step 32283: {'lr': 0.0001500677022233922, 'samples': 16529408, 'steps': 32283, 'loss/train': 2.3061041831970215} 02/25/2022 19:23:44 - INFO - codeparrot_training - Step 32284: {'lr': 0.00015005270406043714, 'samples': 16529920, 'steps': 32284, 'loss/train': 2.043139934539795} 02/25/2022 19:23:47 - INFO - codeparrot_training - Step 32285: {'lr': 0.00015003770632562448, 'samples': 16530432, 'steps': 32285, 'loss/train': 1.3319199085235596} 02/25/2022 19:23:53 - INFO - codeparrot_training - Step 32286: {'lr': 0.00015002270901901855, 'samples': 16530944, 'steps': 32286, 'loss/train': 0.8908756971359253} 02/25/2022 19:23:58 - INFO - codeparrot_training - Step 32287: {'lr': 0.00015000771214068362, 'samples': 16531456, 'steps': 32287, 'loss/train': 2.5564520359039307} 02/25/2022 19:24:02 - INFO - codeparrot_training - Step 32288: {'lr': 0.00014999271569068385, 'samples': 16531968, 'steps': 32288, 'loss/train': 2.058134078979492} 02/25/2022 19:24:08 - INFO - codeparrot_training - Step 32289: {'lr': 0.00014997771966908353, 'samples': 16532480, 'steps': 32289, 'loss/train': 1.7867285013198853} 02/25/2022 19:24:11 - INFO - codeparrot_training - Step 32290: {'lr': 0.00014996272407594694, 'samples': 16532992, 'steps': 32290, 'loss/train': 1.731552243232727} 02/25/2022 19:24:17 - INFO - codeparrot_training - Step 32291: {'lr': 0.0001499477289113383, 'samples': 16533504, 'steps': 32291, 'loss/train': 0.7351446151733398} 02/25/2022 19:24:20 - INFO - codeparrot_training - Step 32292: {'lr': 0.00014993273417532172, 'samples': 16534016, 'steps': 32292, 'loss/train': 2.1299233436584473} 02/25/2022 19:24:26 - INFO - codeparrot_training - Step 32293: {'lr': 0.00014991773986796158, 'samples': 16534528, 'steps': 32293, 'loss/train': 1.0895709991455078} 02/25/2022 19:24:29 - INFO - codeparrot_training - Step 32294: {'lr': 0.000149902745989322, 'samples': 16535040, 'steps': 32294, 'loss/train': 2.260838747024536} 02/25/2022 19:24:36 - INFO - codeparrot_training - Step 32295: {'lr': 0.00014988775253946745, 'samples': 16535552, 'steps': 32295, 'loss/train': 1.7025607824325562} 02/25/2022 19:24:39 - INFO - codeparrot_training - Step 32296: {'lr': 0.00014987275951846185, 'samples': 16536064, 'steps': 32296, 'loss/train': 2.390096426010132} 02/25/2022 19:24:45 - INFO - codeparrot_training - Step 32297: {'lr': 0.00014985776692636954, 'samples': 16536576, 'steps': 32297, 'loss/train': 2.3033804893493652} 02/25/2022 19:24:48 - INFO - codeparrot_training - Step 32298: {'lr': 0.0001498427747632548, 'samples': 16537088, 'steps': 32298, 'loss/train': 2.121675968170166} 02/25/2022 19:24:54 - INFO - codeparrot_training - Step 32299: {'lr': 0.0001498277830291819, 'samples': 16537600, 'steps': 32299, 'loss/train': 1.3470861911773682} 02/25/2022 19:24:57 - INFO - codeparrot_training - Step 32300: {'lr': 0.00014981279172421482, 'samples': 16538112, 'steps': 32300, 'loss/train': 1.252536416053772} 02/25/2022 19:25:03 - INFO - codeparrot_training - Step 32301: {'lr': 0.00014979780084841792, 'samples': 16538624, 'steps': 32301, 'loss/train': 0.7695764899253845} 02/25/2022 19:25:06 - INFO - codeparrot_training - Step 32302: {'lr': 0.00014978281040185548, 'samples': 16539136, 'steps': 32302, 'loss/train': 1.116349220275879} 02/25/2022 19:25:10 - INFO - codeparrot_training - Step 32303: {'lr': 0.00014976782038459164, 'samples': 16539648, 'steps': 32303, 'loss/train': 1.879724383354187} 02/25/2022 19:25:15 - INFO - codeparrot_training - Step 32304: {'lr': 0.00014975283079669073, 'samples': 16540160, 'steps': 32304, 'loss/train': 2.2061822414398193} 02/25/2022 19:25:19 - INFO - codeparrot_training - Step 32305: {'lr': 0.00014973784163821674, 'samples': 16540672, 'steps': 32305, 'loss/train': 2.3807365894317627} 02/25/2022 19:25:25 - INFO - codeparrot_training - Step 32306: {'lr': 0.00014972285290923402, 'samples': 16541184, 'steps': 32306, 'loss/train': 1.610237717628479} 02/25/2022 19:25:31 - INFO - codeparrot_training - Step 32307: {'lr': 0.00014970786460980672, 'samples': 16541696, 'steps': 32307, 'loss/train': 1.5615085363388062} 02/25/2022 19:25:34 - INFO - codeparrot_training - Step 32308: {'lr': 0.00014969287673999927, 'samples': 16542208, 'steps': 32308, 'loss/train': 2.7214581966400146} 02/25/2022 19:25:40 - INFO - codeparrot_training - Step 32309: {'lr': 0.0001496778892998755, 'samples': 16542720, 'steps': 32309, 'loss/train': 1.3820781707763672} 02/25/2022 19:25:43 - INFO - codeparrot_training - Step 32310: {'lr': 0.00014966290228949982, 'samples': 16543232, 'steps': 32310, 'loss/train': 2.3556127548217773} 02/25/2022 19:25:49 - INFO - codeparrot_training - Step 32311: {'lr': 0.00014964791570893642, 'samples': 16543744, 'steps': 32311, 'loss/train': 0.9765664339065552} 02/25/2022 19:25:52 - INFO - codeparrot_training - Step 32312: {'lr': 0.0001496329295582496, 'samples': 16544256, 'steps': 32312, 'loss/train': 1.417349100112915} 02/25/2022 19:25:57 - INFO - codeparrot_training - Step 32313: {'lr': 0.00014961794383750327, 'samples': 16544768, 'steps': 32313, 'loss/train': 1.7288947105407715} 02/25/2022 19:26:01 - INFO - codeparrot_training - Step 32314: {'lr': 0.00014960295854676186, 'samples': 16545280, 'steps': 32314, 'loss/train': 2.018951177597046} 02/25/2022 19:26:07 - INFO - codeparrot_training - Step 32315: {'lr': 0.00014958797368608945, 'samples': 16545792, 'steps': 32315, 'loss/train': 2.0905697345733643} 02/25/2022 19:26:11 - INFO - codeparrot_training - Step 32316: {'lr': 0.0001495729892555503, 'samples': 16546304, 'steps': 32316, 'loss/train': 2.4227921962738037} 02/25/2022 19:26:16 - INFO - codeparrot_training - Step 32317: {'lr': 0.0001495580052552087, 'samples': 16546816, 'steps': 32317, 'loss/train': 1.0265288352966309} 02/25/2022 19:26:20 - INFO - codeparrot_training - Step 32318: {'lr': 0.00014954302168512857, 'samples': 16547328, 'steps': 32318, 'loss/train': 2.15868878364563} 02/25/2022 19:26:26 - INFO - codeparrot_training - Step 32319: {'lr': 0.00014952803854537429, 'samples': 16547840, 'steps': 32319, 'loss/train': 1.7687561511993408} 02/25/2022 19:26:29 - INFO - codeparrot_training - Step 32320: {'lr': 0.00014951305583600999, 'samples': 16548352, 'steps': 32320, 'loss/train': 2.0340535640716553} 02/25/2022 19:26:35 - INFO - codeparrot_training - Step 32321: {'lr': 0.00014949807355709986, 'samples': 16548864, 'steps': 32321, 'loss/train': 2.162201166152954} 02/25/2022 19:26:38 - INFO - codeparrot_training - Step 32322: {'lr': 0.000149483091708708, 'samples': 16549376, 'steps': 32322, 'loss/train': 1.4235886335372925} 02/25/2022 19:26:44 - INFO - codeparrot_training - Step 32323: {'lr': 0.0001494681102908987, 'samples': 16549888, 'steps': 32323, 'loss/train': 0.40066954493522644} 02/25/2022 19:26:47 - INFO - codeparrot_training - Step 32324: {'lr': 0.00014945312930373611, 'samples': 16550400, 'steps': 32324, 'loss/train': 2.9656054973602295} 02/25/2022 19:26:53 - INFO - codeparrot_training - Step 32325: {'lr': 0.0001494381487472844, 'samples': 16550912, 'steps': 32325, 'loss/train': 2.722982883453369} 02/25/2022 19:26:56 - INFO - codeparrot_training - Step 32326: {'lr': 0.00014942316862160768, 'samples': 16551424, 'steps': 32326, 'loss/train': 2.050794839859009} 02/25/2022 19:27:02 - INFO - codeparrot_training - Step 32327: {'lr': 0.00014940818892677021, 'samples': 16551936, 'steps': 32327, 'loss/train': 1.385240912437439} 02/25/2022 19:27:06 - INFO - codeparrot_training - Step 32328: {'lr': 0.00014939320966283608, 'samples': 16552448, 'steps': 32328, 'loss/train': 2.0653727054595947} 02/25/2022 19:27:11 - INFO - codeparrot_training - Step 32329: {'lr': 0.00014937823082986952, 'samples': 16552960, 'steps': 32329, 'loss/train': 2.039252519607544} 02/25/2022 19:27:15 - INFO - codeparrot_training - Step 32330: {'lr': 0.00014936325242793466, 'samples': 16553472, 'steps': 32330, 'loss/train': 1.0067226886749268} 02/25/2022 19:27:20 - INFO - codeparrot_training - Step 32331: {'lr': 0.0001493482744570957, 'samples': 16553984, 'steps': 32331, 'loss/train': 1.7187745571136475} 02/25/2022 19:27:24 - INFO - codeparrot_training - Step 32332: {'lr': 0.0001493332969174167, 'samples': 16554496, 'steps': 32332, 'loss/train': 1.872113823890686} 02/25/2022 19:27:29 - INFO - codeparrot_training - Step 32333: {'lr': 0.00014931831980896193, 'samples': 16555008, 'steps': 32333, 'loss/train': 2.166632652282715} 02/25/2022 19:27:33 - INFO - codeparrot_training - Step 32334: {'lr': 0.0001493033431317956, 'samples': 16555520, 'steps': 32334, 'loss/train': 2.0489113330841064} 02/25/2022 19:27:38 - INFO - codeparrot_training - Step 32335: {'lr': 0.00014928836688598164, 'samples': 16556032, 'steps': 32335, 'loss/train': 1.5888949632644653} 02/25/2022 19:27:42 - INFO - codeparrot_training - Step 32336: {'lr': 0.00014927339107158436, 'samples': 16556544, 'steps': 32336, 'loss/train': 1.2976455688476562} 02/25/2022 19:27:47 - INFO - codeparrot_training - Step 32337: {'lr': 0.00014925841568866788, 'samples': 16557056, 'steps': 32337, 'loss/train': 1.5581536293029785} 02/25/2022 19:27:51 - INFO - codeparrot_training - Step 32338: {'lr': 0.00014924344073729648, 'samples': 16557568, 'steps': 32338, 'loss/train': 2.0754854679107666} 02/25/2022 19:27:56 - INFO - codeparrot_training - Step 32339: {'lr': 0.00014922846621753406, 'samples': 16558080, 'steps': 32339, 'loss/train': 2.1283252239227295} 02/25/2022 19:28:00 - INFO - codeparrot_training - Step 32340: {'lr': 0.0001492134921294449, 'samples': 16558592, 'steps': 32340, 'loss/train': 1.738769769668579} 02/25/2022 19:28:06 - INFO - codeparrot_training - Step 32341: {'lr': 0.00014919851847309312, 'samples': 16559104, 'steps': 32341, 'loss/train': 1.997680902481079} 02/25/2022 19:28:10 - INFO - codeparrot_training - Step 32342: {'lr': 0.00014918354524854293, 'samples': 16559616, 'steps': 32342, 'loss/train': 1.2354999780654907} 02/25/2022 19:28:15 - INFO - codeparrot_training - Step 32343: {'lr': 0.00014916857245585847, 'samples': 16560128, 'steps': 32343, 'loss/train': 0.13435079157352448} 02/25/2022 19:28:18 - INFO - codeparrot_training - Step 32344: {'lr': 0.00014915360009510375, 'samples': 16560640, 'steps': 32344, 'loss/train': 2.8112382888793945} 02/25/2022 19:28:24 - INFO - codeparrot_training - Step 32345: {'lr': 0.000149138628166343, 'samples': 16561152, 'steps': 32345, 'loss/train': 0.7388100028038025} 02/25/2022 19:28:27 - INFO - codeparrot_training - Step 32346: {'lr': 0.00014912365666964032, 'samples': 16561664, 'steps': 32346, 'loss/train': 2.7711784839630127} 02/25/2022 19:28:33 - INFO - codeparrot_training - Step 32347: {'lr': 0.00014910868560505996, 'samples': 16562176, 'steps': 32347, 'loss/train': 0.7430292963981628} 02/25/2022 19:28:37 - INFO - codeparrot_training - Step 32348: {'lr': 0.00014909371497266583, 'samples': 16562688, 'steps': 32348, 'loss/train': 1.8972395658493042} 02/25/2022 19:28:42 - INFO - codeparrot_training - Step 32349: {'lr': 0.00014907874477252222, 'samples': 16563200, 'steps': 32349, 'loss/train': 0.7243121862411499} 02/25/2022 19:28:46 - INFO - codeparrot_training - Step 32350: {'lr': 0.0001490637750046932, 'samples': 16563712, 'steps': 32350, 'loss/train': 1.9064499139785767} 02/25/2022 19:28:52 - INFO - codeparrot_training - Step 32351: {'lr': 0.00014904880566924295, 'samples': 16564224, 'steps': 32351, 'loss/train': 1.4657062292099} 02/25/2022 19:28:55 - INFO - codeparrot_training - Step 32352: {'lr': 0.00014903383676623564, 'samples': 16564736, 'steps': 32352, 'loss/train': 2.2063562870025635} 02/25/2022 19:29:01 - INFO - codeparrot_training - Step 32353: {'lr': 0.00014901886829573523, 'samples': 16565248, 'steps': 32353, 'loss/train': 1.4842668771743774} 02/25/2022 19:29:04 - INFO - codeparrot_training - Step 32354: {'lr': 0.0001490039002578059, 'samples': 16565760, 'steps': 32354, 'loss/train': 1.9563828706741333} 02/25/2022 19:29:10 - INFO - codeparrot_training - Step 32355: {'lr': 0.0001489889326525118, 'samples': 16566272, 'steps': 32355, 'loss/train': 2.100186586380005} 02/25/2022 19:29:13 - INFO - codeparrot_training - Step 32356: {'lr': 0.00014897396547991712, 'samples': 16566784, 'steps': 32356, 'loss/train': 2.0756497383117676} 02/25/2022 19:29:19 - INFO - codeparrot_training - Step 32357: {'lr': 0.0001489589987400858, 'samples': 16567296, 'steps': 32357, 'loss/train': 1.6552541255950928} 02/25/2022 19:29:22 - INFO - codeparrot_training - Step 32358: {'lr': 0.00014894403243308207, 'samples': 16567808, 'steps': 32358, 'loss/train': 1.8560599088668823} 02/25/2022 19:29:28 - INFO - codeparrot_training - Step 32359: {'lr': 0.00014892906655897, 'samples': 16568320, 'steps': 32359, 'loss/train': 0.39049747586250305} 02/25/2022 19:29:34 - INFO - codeparrot_training - Step 32360: {'lr': 0.0001489141011178138, 'samples': 16568832, 'steps': 32360, 'loss/train': 2.875079870223999} 02/25/2022 19:29:38 - INFO - codeparrot_training - Step 32361: {'lr': 0.00014889913610967743, 'samples': 16569344, 'steps': 32361, 'loss/train': 1.2644240856170654} 02/25/2022 19:29:41 - INFO - codeparrot_training - Step 32362: {'lr': 0.00014888417153462503, 'samples': 16569856, 'steps': 32362, 'loss/train': 1.3506760597229004} 02/25/2022 19:29:47 - INFO - codeparrot_training - Step 32363: {'lr': 0.00014886920739272071, 'samples': 16570368, 'steps': 32363, 'loss/train': 1.1997145414352417} 02/25/2022 19:29:51 - INFO - codeparrot_training - Step 32364: {'lr': 0.00014885424368402868, 'samples': 16570880, 'steps': 32364, 'loss/train': 3.3429555892944336} 02/25/2022 19:29:56 - INFO - codeparrot_training - Step 32365: {'lr': 0.00014883928040861294, 'samples': 16571392, 'steps': 32365, 'loss/train': 1.9886585474014282} 02/25/2022 19:30:00 - INFO - codeparrot_training - Step 32366: {'lr': 0.00014882431756653756, 'samples': 16571904, 'steps': 32366, 'loss/train': 2.280878782272339} 02/25/2022 19:30:05 - INFO - codeparrot_training - Step 32367: {'lr': 0.0001488093551578667, 'samples': 16572416, 'steps': 32367, 'loss/train': 2.1996753215789795} 02/25/2022 19:30:09 - INFO - codeparrot_training - Step 32368: {'lr': 0.0001487943931826644, 'samples': 16572928, 'steps': 32368, 'loss/train': 0.8394864201545715} 02/25/2022 19:30:14 - INFO - codeparrot_training - Step 32369: {'lr': 0.00014877943164099484, 'samples': 16573440, 'steps': 32369, 'loss/train': 0.8564398288726807} 02/25/2022 19:30:18 - INFO - codeparrot_training - Step 32370: {'lr': 0.000148764470532922, 'samples': 16573952, 'steps': 32370, 'loss/train': 1.2653359174728394} 02/25/2022 19:30:24 - INFO - codeparrot_training - Step 32371: {'lr': 0.00014874950985851003, 'samples': 16574464, 'steps': 32371, 'loss/train': 1.6284235715866089} 02/25/2022 19:30:27 - INFO - codeparrot_training - Step 32372: {'lr': 0.00014873454961782304, 'samples': 16574976, 'steps': 32372, 'loss/train': 2.074125289916992} 02/25/2022 19:30:33 - INFO - codeparrot_training - Step 32373: {'lr': 0.0001487195898109251, 'samples': 16575488, 'steps': 32373, 'loss/train': 2.0571651458740234} 02/25/2022 19:30:37 - INFO - codeparrot_training - Step 32374: {'lr': 0.00014870463043788025, 'samples': 16576000, 'steps': 32374, 'loss/train': 1.0151761770248413} 02/25/2022 19:30:42 - INFO - codeparrot_training - Step 32375: {'lr': 0.00014868967149875257, 'samples': 16576512, 'steps': 32375, 'loss/train': 0.9607380628585815} 02/25/2022 19:30:46 - INFO - codeparrot_training - Step 32376: {'lr': 0.0001486747129936062, 'samples': 16577024, 'steps': 32376, 'loss/train': 3.0268239974975586} 02/25/2022 19:30:51 - INFO - codeparrot_training - Step 32377: {'lr': 0.0001486597549225051, 'samples': 16577536, 'steps': 32377, 'loss/train': 1.6138197183609009} 02/25/2022 19:30:55 - INFO - codeparrot_training - Step 32378: {'lr': 0.00014864479728551362, 'samples': 16578048, 'steps': 32378, 'loss/train': 1.1942315101623535} 02/25/2022 19:31:00 - INFO - codeparrot_training - Step 32379: {'lr': 0.00014862984008269547, 'samples': 16578560, 'steps': 32379, 'loss/train': 0.864029586315155} 02/25/2022 19:31:04 - INFO - codeparrot_training - Step 32380: {'lr': 0.00014861488331411492, 'samples': 16579072, 'steps': 32380, 'loss/train': 2.648854970932007} 02/25/2022 19:31:09 - INFO - codeparrot_training - Step 32381: {'lr': 0.00014859992697983604, 'samples': 16579584, 'steps': 32381, 'loss/train': 0.8249788880348206} 02/25/2022 19:31:13 - INFO - codeparrot_training - Step 32382: {'lr': 0.00014858497107992296, 'samples': 16580096, 'steps': 32382, 'loss/train': 0.701622724533081} 02/25/2022 19:31:18 - INFO - codeparrot_training - Step 32383: {'lr': 0.0001485700156144396, 'samples': 16580608, 'steps': 32383, 'loss/train': 1.7681553363800049} 02/25/2022 19:31:22 - INFO - codeparrot_training - Step 32384: {'lr': 0.00014855506058345002, 'samples': 16581120, 'steps': 32384, 'loss/train': 1.8353049755096436} 02/25/2022 19:31:27 - INFO - codeparrot_training - Step 32385: {'lr': 0.00014854010598701838, 'samples': 16581632, 'steps': 32385, 'loss/train': 1.6558737754821777} 02/25/2022 19:31:31 - INFO - codeparrot_training - Step 32386: {'lr': 0.0001485251518252088, 'samples': 16582144, 'steps': 32386, 'loss/train': 1.4918975830078125} 02/25/2022 19:31:36 - INFO - codeparrot_training - Step 32387: {'lr': 0.00014851019809808516, 'samples': 16582656, 'steps': 32387, 'loss/train': 2.743694305419922} 02/25/2022 19:31:40 - INFO - codeparrot_training - Step 32388: {'lr': 0.0001484952448057116, 'samples': 16583168, 'steps': 32388, 'loss/train': 1.87287437915802} 02/25/2022 19:31:46 - INFO - codeparrot_training - Step 32389: {'lr': 0.0001484802919481522, 'samples': 16583680, 'steps': 32389, 'loss/train': 2.2721259593963623} 02/25/2022 19:31:50 - INFO - codeparrot_training - Step 32390: {'lr': 0.00014846533952547094, 'samples': 16584192, 'steps': 32390, 'loss/train': 2.5826575756073} 02/25/2022 19:31:55 - INFO - codeparrot_training - Step 32391: {'lr': 0.00014845038753773208, 'samples': 16584704, 'steps': 32391, 'loss/train': 1.1010130643844604} 02/25/2022 19:31:59 - INFO - codeparrot_training - Step 32392: {'lr': 0.00014843543598499936, 'samples': 16585216, 'steps': 32392, 'loss/train': 1.7832626104354858} 02/25/2022 19:32:04 - INFO - codeparrot_training - Step 32393: {'lr': 0.00014842048486733703, 'samples': 16585728, 'steps': 32393, 'loss/train': 1.3486442565917969} 02/25/2022 19:32:08 - INFO - codeparrot_training - Step 32394: {'lr': 0.00014840553418480907, 'samples': 16586240, 'steps': 32394, 'loss/train': 1.0889757871627808} 02/25/2022 19:32:13 - INFO - codeparrot_training - Step 32395: {'lr': 0.00014839058393747965, 'samples': 16586752, 'steps': 32395, 'loss/train': 1.8366035223007202} 02/25/2022 19:32:16 - INFO - codeparrot_training - Step 32396: {'lr': 0.0001483756341254126, 'samples': 16587264, 'steps': 32396, 'loss/train': 3.02665638923645} 02/25/2022 19:32:22 - INFO - codeparrot_training - Step 32397: {'lr': 0.00014836068474867204, 'samples': 16587776, 'steps': 32397, 'loss/train': 2.0333051681518555} 02/25/2022 19:32:25 - INFO - codeparrot_training - Step 32398: {'lr': 0.00014834573580732209, 'samples': 16588288, 'steps': 32398, 'loss/train': 2.0648722648620605} 02/25/2022 19:32:32 - INFO - codeparrot_training - Step 32399: {'lr': 0.00014833078730142675, 'samples': 16588800, 'steps': 32399, 'loss/train': 2.195464611053467} 02/25/2022 19:32:35 - INFO - codeparrot_training - Step 32400: {'lr': 0.00014831583923105, 'samples': 16589312, 'steps': 32400, 'loss/train': 1.9659101963043213} 02/25/2022 19:32:41 - INFO - codeparrot_training - Step 32401: {'lr': 0.00014830089159625586, 'samples': 16589824, 'steps': 32401, 'loss/train': 0.2595829665660858} 02/25/2022 19:32:44 - INFO - codeparrot_training - Step 32402: {'lr': 0.00014828594439710843, 'samples': 16590336, 'steps': 32402, 'loss/train': 2.886333465576172} 02/25/2022 19:32:50 - INFO - codeparrot_training - Step 32403: {'lr': 0.0001482709976336717, 'samples': 16590848, 'steps': 32403, 'loss/train': 0.9266557693481445} 02/25/2022 19:32:53 - INFO - codeparrot_training - Step 32404: {'lr': 0.00014825605130600983, 'samples': 16591360, 'steps': 32404, 'loss/train': 1.7129405736923218} 02/25/2022 19:32:59 - INFO - codeparrot_training - Step 32405: {'lr': 0.00014824110541418665, 'samples': 16591872, 'steps': 32405, 'loss/train': 0.6770980358123779} 02/25/2022 19:33:02 - INFO - codeparrot_training - Step 32406: {'lr': 0.00014822615995826622, 'samples': 16592384, 'steps': 32406, 'loss/train': 2.1913487911224365} 02/25/2022 19:33:08 - INFO - codeparrot_training - Step 32407: {'lr': 0.00014821121493831268, 'samples': 16592896, 'steps': 32407, 'loss/train': 0.9003936052322388} 02/25/2022 19:33:11 - INFO - codeparrot_training - Step 32408: {'lr': 0.00014819627035439, 'samples': 16593408, 'steps': 32408, 'loss/train': 1.6318864822387695} 02/25/2022 19:33:17 - INFO - codeparrot_training - Step 32409: {'lr': 0.00014818132620656206, 'samples': 16593920, 'steps': 32409, 'loss/train': 1.6203229427337646} 02/25/2022 19:33:21 - INFO - codeparrot_training - Step 32410: {'lr': 0.000148166382494893, 'samples': 16594432, 'steps': 32410, 'loss/train': 1.3255537748336792} 02/25/2022 19:33:26 - INFO - codeparrot_training - Step 32411: {'lr': 0.0001481514392194469, 'samples': 16594944, 'steps': 32411, 'loss/train': 2.2070844173431396} 02/25/2022 19:33:30 - INFO - codeparrot_training - Step 32412: {'lr': 0.00014813649638028764, 'samples': 16595456, 'steps': 32412, 'loss/train': 2.1397461891174316} 02/25/2022 19:33:35 - INFO - codeparrot_training - Step 32413: {'lr': 0.00014812155397747931, 'samples': 16595968, 'steps': 32413, 'loss/train': 1.8519595861434937} 02/25/2022 19:33:39 - INFO - codeparrot_training - Step 32414: {'lr': 0.0001481066120110859, 'samples': 16596480, 'steps': 32414, 'loss/train': 0.7094731330871582} 02/25/2022 19:33:44 - INFO - codeparrot_training - Step 32415: {'lr': 0.00014809167048117139, 'samples': 16596992, 'steps': 32415, 'loss/train': 1.8885993957519531} 02/25/2022 19:33:48 - INFO - codeparrot_training - Step 32416: {'lr': 0.00014807672938779975, 'samples': 16597504, 'steps': 32416, 'loss/train': 1.122735857963562} 02/25/2022 19:33:53 - INFO - codeparrot_training - Step 32417: {'lr': 0.00014806178873103516, 'samples': 16598016, 'steps': 32417, 'loss/train': 1.1100540161132812} 02/25/2022 19:33:57 - INFO - codeparrot_training - Step 32418: {'lr': 0.00014804684851094145, 'samples': 16598528, 'steps': 32418, 'loss/train': 1.671758770942688} 02/25/2022 19:34:03 - INFO - codeparrot_training - Step 32419: {'lr': 0.0001480319087275826, 'samples': 16599040, 'steps': 32419, 'loss/train': 1.9955915212631226} 02/25/2022 19:34:07 - INFO - codeparrot_training - Step 32420: {'lr': 0.00014801696938102272, 'samples': 16599552, 'steps': 32420, 'loss/train': 1.784430742263794} 02/25/2022 19:34:12 - INFO - codeparrot_training - Step 32421: {'lr': 0.00014800203047132576, 'samples': 16600064, 'steps': 32421, 'loss/train': 1.9255657196044922} 02/25/2022 19:34:16 - INFO - codeparrot_training - Step 32422: {'lr': 0.00014798709199855575, 'samples': 16600576, 'steps': 32422, 'loss/train': 2.7534384727478027} 02/25/2022 19:34:21 - INFO - codeparrot_training - Step 32423: {'lr': 0.00014797215396277657, 'samples': 16601088, 'steps': 32423, 'loss/train': 1.1202454566955566} 02/25/2022 19:34:24 - INFO - codeparrot_training - Step 32424: {'lr': 0.00014795721636405232, 'samples': 16601600, 'steps': 32424, 'loss/train': 1.6647528409957886} 02/25/2022 19:34:30 - INFO - codeparrot_training - Step 32425: {'lr': 0.00014794227920244697, 'samples': 16602112, 'steps': 32425, 'loss/train': 1.5797253847122192} 02/25/2022 19:34:34 - INFO - codeparrot_training - Step 32426: {'lr': 0.00014792734247802452, 'samples': 16602624, 'steps': 32426, 'loss/train': 1.0034444332122803} 02/25/2022 19:34:39 - INFO - codeparrot_training - Step 32427: {'lr': 0.00014791240619084888, 'samples': 16603136, 'steps': 32427, 'loss/train': 1.3269704580307007} 02/25/2022 19:34:43 - INFO - codeparrot_training - Step 32428: {'lr': 0.00014789747034098407, 'samples': 16603648, 'steps': 32428, 'loss/train': 2.880244493484497} 02/25/2022 19:34:49 - INFO - codeparrot_training - Step 32429: {'lr': 0.00014788253492849403, 'samples': 16604160, 'steps': 32429, 'loss/train': 2.087341785430908} 02/25/2022 19:34:52 - INFO - codeparrot_training - Step 32430: {'lr': 0.0001478675999534429, 'samples': 16604672, 'steps': 32430, 'loss/train': 2.6799347400665283} 02/25/2022 19:34:58 - INFO - codeparrot_training - Step 32431: {'lr': 0.00014785266541589444, 'samples': 16605184, 'steps': 32431, 'loss/train': 2.979008197784424} 02/25/2022 19:35:01 - INFO - codeparrot_training - Step 32432: {'lr': 0.00014783773131591278, 'samples': 16605696, 'steps': 32432, 'loss/train': 1.423139214515686} 02/25/2022 19:35:07 - INFO - codeparrot_training - Step 32433: {'lr': 0.00014782279765356178, 'samples': 16606208, 'steps': 32433, 'loss/train': 0.9647095203399658} 02/25/2022 19:35:10 - INFO - codeparrot_training - Step 32434: {'lr': 0.0001478078644289056, 'samples': 16606720, 'steps': 32434, 'loss/train': 0.5865122675895691} 02/25/2022 19:35:16 - INFO - codeparrot_training - Step 32435: {'lr': 0.00014779293164200798, 'samples': 16607232, 'steps': 32435, 'loss/train': 1.6733436584472656} 02/25/2022 19:35:19 - INFO - codeparrot_training - Step 32436: {'lr': 0.00014777799929293294, 'samples': 16607744, 'steps': 32436, 'loss/train': 1.204079508781433} 02/25/2022 19:35:25 - INFO - codeparrot_training - Step 32437: {'lr': 0.00014776306738174453, 'samples': 16608256, 'steps': 32437, 'loss/train': 2.0726795196533203} 02/25/2022 19:35:30 - INFO - codeparrot_training - Step 32438: {'lr': 0.00014774813590850665, 'samples': 16608768, 'steps': 32438, 'loss/train': 2.0674962997436523} 02/25/2022 19:35:34 - INFO - codeparrot_training - Step 32439: {'lr': 0.00014773320487328342, 'samples': 16609280, 'steps': 32439, 'loss/train': 1.2218797206878662} 02/25/2022 19:35:37 - INFO - codeparrot_training - Step 32440: {'lr': 0.00014771827427613856, 'samples': 16609792, 'steps': 32440, 'loss/train': 2.6540238857269287} 02/25/2022 19:35:43 - INFO - codeparrot_training - Step 32441: {'lr': 0.00014770334411713613, 'samples': 16610304, 'steps': 32441, 'loss/train': 1.991631269454956} 02/25/2022 19:35:48 - INFO - codeparrot_training - Step 32442: {'lr': 0.0001476884143963401, 'samples': 16610816, 'steps': 32442, 'loss/train': 1.7428946495056152} 02/25/2022 19:35:52 - INFO - codeparrot_training - Step 32443: {'lr': 0.0001476734851138145, 'samples': 16611328, 'steps': 32443, 'loss/train': 1.5309555530548096} 02/25/2022 19:35:58 - INFO - codeparrot_training - Step 32444: {'lr': 0.0001476585562696231, 'samples': 16611840, 'steps': 32444, 'loss/train': 1.5594590902328491} 02/25/2022 19:36:01 - INFO - codeparrot_training - Step 32445: {'lr': 0.00014764362786382995, 'samples': 16612352, 'steps': 32445, 'loss/train': 1.4648518562316895} 02/25/2022 19:36:07 - INFO - codeparrot_training - Step 32446: {'lr': 0.00014762869989649898, 'samples': 16612864, 'steps': 32446, 'loss/train': 0.8124240636825562} 02/25/2022 19:36:10 - INFO - codeparrot_training - Step 32447: {'lr': 0.0001476137723676943, 'samples': 16613376, 'steps': 32447, 'loss/train': 1.1763348579406738} 02/25/2022 19:36:16 - INFO - codeparrot_training - Step 32448: {'lr': 0.00014759884527747958, 'samples': 16613888, 'steps': 32448, 'loss/train': 1.5468896627426147} 02/25/2022 19:36:19 - INFO - codeparrot_training - Step 32449: {'lr': 0.0001475839186259189, 'samples': 16614400, 'steps': 32449, 'loss/train': 0.6370042562484741} 02/25/2022 19:36:25 - INFO - codeparrot_training - Step 32450: {'lr': 0.00014756899241307614, 'samples': 16614912, 'steps': 32450, 'loss/train': 1.643971562385559} 02/25/2022 19:36:28 - INFO - codeparrot_training - Step 32451: {'lr': 0.0001475540666390154, 'samples': 16615424, 'steps': 32451, 'loss/train': 2.157505750656128} 02/25/2022 19:36:32 - INFO - codeparrot_training - Step 32452: {'lr': 0.00014753914130380047, 'samples': 16615936, 'steps': 32452, 'loss/train': 3.6336867809295654} 02/25/2022 19:36:37 - INFO - codeparrot_training - Step 32453: {'lr': 0.0001475242164074953, 'samples': 16616448, 'steps': 32453, 'loss/train': 3.3105990886688232} 02/25/2022 19:36:41 - INFO - codeparrot_training - Step 32454: {'lr': 0.00014750929195016385, 'samples': 16616960, 'steps': 32454, 'loss/train': 0.29292500019073486} 02/25/2022 19:36:47 - INFO - codeparrot_training - Step 32455: {'lr': 0.0001474943679318701, 'samples': 16617472, 'steps': 32455, 'loss/train': 0.9915158152580261} 02/25/2022 19:36:51 - INFO - codeparrot_training - Step 32456: {'lr': 0.0001474794443526779, 'samples': 16617984, 'steps': 32456, 'loss/train': 2.117870807647705} 02/25/2022 19:36:56 - INFO - codeparrot_training - Step 32457: {'lr': 0.0001474645212126512, 'samples': 16618496, 'steps': 32457, 'loss/train': 1.7741111516952515} 02/25/2022 19:37:02 - INFO - codeparrot_training - Step 32458: {'lr': 0.00014744959851185397, 'samples': 16619008, 'steps': 32458, 'loss/train': 0.20407609641551971} 02/25/2022 19:37:05 - INFO - codeparrot_training - Step 32459: {'lr': 0.00014743467625035001, 'samples': 16619520, 'steps': 32459, 'loss/train': 0.8214003443717957} 02/25/2022 19:37:11 - INFO - codeparrot_training - Step 32460: {'lr': 0.00014741975442820335, 'samples': 16620032, 'steps': 32460, 'loss/train': 2.6957285404205322} 02/25/2022 19:37:14 - INFO - codeparrot_training - Step 32461: {'lr': 0.00014740483304547794, 'samples': 16620544, 'steps': 32461, 'loss/train': 2.6454899311065674} 02/25/2022 19:37:20 - INFO - codeparrot_training - Step 32462: {'lr': 0.00014738991210223767, 'samples': 16621056, 'steps': 32462, 'loss/train': 2.3153529167175293} 02/25/2022 19:37:23 - INFO - codeparrot_training - Step 32463: {'lr': 0.00014737499159854633, 'samples': 16621568, 'steps': 32463, 'loss/train': 1.7571243047714233} 02/25/2022 19:37:29 - INFO - codeparrot_training - Step 32464: {'lr': 0.00014736007153446802, 'samples': 16622080, 'steps': 32464, 'loss/train': 1.3341004848480225} 02/25/2022 19:37:32 - INFO - codeparrot_training - Step 32465: {'lr': 0.0001473451519100666, 'samples': 16622592, 'steps': 32465, 'loss/train': 1.1712263822555542} 02/25/2022 19:37:39 - INFO - codeparrot_training - Step 32466: {'lr': 0.00014733023272540586, 'samples': 16623104, 'steps': 32466, 'loss/train': 0.3323075473308563} 02/25/2022 19:37:42 - INFO - codeparrot_training - Step 32467: {'lr': 0.0001473153139805498, 'samples': 16623616, 'steps': 32467, 'loss/train': 2.4661805629730225} 02/25/2022 19:37:48 - INFO - codeparrot_training - Step 32468: {'lr': 0.00014730039567556239, 'samples': 16624128, 'steps': 32468, 'loss/train': 1.6095824241638184} 02/25/2022 19:37:51 - INFO - codeparrot_training - Step 32469: {'lr': 0.00014728547781050753, 'samples': 16624640, 'steps': 32469, 'loss/train': 1.788891315460205} 02/25/2022 19:37:57 - INFO - codeparrot_training - Step 32470: {'lr': 0.00014727056038544895, 'samples': 16625152, 'steps': 32470, 'loss/train': 1.8447391986846924} 02/25/2022 19:38:00 - INFO - codeparrot_training - Step 32471: {'lr': 0.0001472556434004507, 'samples': 16625664, 'steps': 32471, 'loss/train': 1.5776461362838745} 02/25/2022 19:38:06 - INFO - codeparrot_training - Step 32472: {'lr': 0.00014724072685557666, 'samples': 16626176, 'steps': 32472, 'loss/train': 2.286524772644043} 02/25/2022 19:38:09 - INFO - codeparrot_training - Step 32473: {'lr': 0.00014722581075089067, 'samples': 16626688, 'steps': 32473, 'loss/train': 2.2687315940856934} 02/25/2022 19:38:15 - INFO - codeparrot_training - Step 32474: {'lr': 0.0001472108950864568, 'samples': 16627200, 'steps': 32474, 'loss/train': 1.6679648160934448} 02/25/2022 19:38:19 - INFO - codeparrot_training - Step 32475: {'lr': 0.0001471959798623387, 'samples': 16627712, 'steps': 32475, 'loss/train': 2.1277120113372803} 02/25/2022 19:38:25 - INFO - codeparrot_training - Step 32476: {'lr': 0.0001471810650786004, 'samples': 16628224, 'steps': 32476, 'loss/train': 2.6028218269348145} 02/25/2022 19:38:28 - INFO - codeparrot_training - Step 32477: {'lr': 0.00014716615073530575, 'samples': 16628736, 'steps': 32477, 'loss/train': 1.2286666631698608} 02/25/2022 19:38:32 - INFO - codeparrot_training - Step 32478: {'lr': 0.00014715123683251878, 'samples': 16629248, 'steps': 32478, 'loss/train': 0.06970261037349701} 02/25/2022 19:38:37 - INFO - codeparrot_training - Step 32479: {'lr': 0.00014713632337030313, 'samples': 16629760, 'steps': 32479, 'loss/train': 1.4311376810073853} 02/25/2022 19:38:41 - INFO - codeparrot_training - Step 32480: {'lr': 0.0001471214103487228, 'samples': 16630272, 'steps': 32480, 'loss/train': 1.5347286462783813} 02/25/2022 19:38:46 - INFO - codeparrot_training - Step 32481: {'lr': 0.0001471064977678417, 'samples': 16630784, 'steps': 32481, 'loss/train': 1.024117112159729} 02/25/2022 19:38:50 - INFO - codeparrot_training - Step 32482: {'lr': 0.0001470915856277238, 'samples': 16631296, 'steps': 32482, 'loss/train': 0.7031924724578857} 02/25/2022 19:38:55 - INFO - codeparrot_training - Step 32483: {'lr': 0.00014707667392843278, 'samples': 16631808, 'steps': 32483, 'loss/train': 1.365310549736023} 02/25/2022 19:38:59 - INFO - codeparrot_training - Step 32484: {'lr': 0.00014706176267003258, 'samples': 16632320, 'steps': 32484, 'loss/train': 2.31606125831604} 02/25/2022 19:39:04 - INFO - codeparrot_training - Step 32485: {'lr': 0.0001470468518525871, 'samples': 16632832, 'steps': 32485, 'loss/train': 1.7600618600845337} 02/25/2022 19:39:08 - INFO - codeparrot_training - Step 32486: {'lr': 0.0001470319414761602, 'samples': 16633344, 'steps': 32486, 'loss/train': 0.2055341750383377} 02/25/2022 19:39:13 - INFO - codeparrot_training - Step 32487: {'lr': 0.0001470170315408159, 'samples': 16633856, 'steps': 32487, 'loss/train': 1.2472333908081055} 02/25/2022 19:39:17 - INFO - codeparrot_training - Step 32488: {'lr': 0.00014700212204661785, 'samples': 16634368, 'steps': 32488, 'loss/train': 1.270909070968628} 02/25/2022 19:39:22 - INFO - codeparrot_training - Step 32489: {'lr': 0.00014698721299362996, 'samples': 16634880, 'steps': 32489, 'loss/train': 1.7824101448059082} 02/25/2022 19:39:26 - INFO - codeparrot_training - Step 32490: {'lr': 0.00014697230438191617, 'samples': 16635392, 'steps': 32490, 'loss/train': 1.016020655632019} 02/25/2022 19:39:33 - INFO - codeparrot_training - Step 32491: {'lr': 0.00014695739621154038, 'samples': 16635904, 'steps': 32491, 'loss/train': 1.9061532020568848} 02/25/2022 19:39:36 - INFO - codeparrot_training - Step 32492: {'lr': 0.0001469424884825663, 'samples': 16636416, 'steps': 32492, 'loss/train': 1.249210000038147} 02/25/2022 19:39:42 - INFO - codeparrot_training - Step 32493: {'lr': 0.00014692758119505789, 'samples': 16636928, 'steps': 32493, 'loss/train': 1.5328309535980225} 02/25/2022 19:39:45 - INFO - codeparrot_training - Step 32494: {'lr': 0.000146912674349079, 'samples': 16637440, 'steps': 32494, 'loss/train': 2.420698881149292} 02/25/2022 19:39:51 - INFO - codeparrot_training - Step 32495: {'lr': 0.00014689776794469357, 'samples': 16637952, 'steps': 32495, 'loss/train': 1.5119743347167969} 02/25/2022 19:39:56 - INFO - codeparrot_training - Step 32496: {'lr': 0.00014688286198196525, 'samples': 16638464, 'steps': 32496, 'loss/train': 1.5790690183639526} 02/25/2022 19:40:00 - INFO - codeparrot_training - Step 32497: {'lr': 0.00014686795646095803, 'samples': 16638976, 'steps': 32497, 'loss/train': 1.986587405204773} 02/25/2022 19:40:05 - INFO - codeparrot_training - Step 32498: {'lr': 0.00014685305138173574, 'samples': 16639488, 'steps': 32498, 'loss/train': 1.8447834253311157} 02/25/2022 19:40:09 - INFO - codeparrot_training - Step 32499: {'lr': 0.00014683814674436218, 'samples': 16640000, 'steps': 32499, 'loss/train': 0.35245779156684875} 02/25/2022 19:40:14 - INFO - codeparrot_training - Step 32500: {'lr': 0.00014682324254890135, 'samples': 16640512, 'steps': 32500, 'loss/train': 1.0489423274993896} 02/25/2022 19:40:18 - INFO - codeparrot_training - Step 32501: {'lr': 0.00014680833879541689, 'samples': 16641024, 'steps': 32501, 'loss/train': 1.6987041234970093} 02/25/2022 19:40:24 - INFO - codeparrot_training - Step 32502: {'lr': 0.00014679343548397282, 'samples': 16641536, 'steps': 32502, 'loss/train': 0.5247048735618591} 02/25/2022 19:40:28 - INFO - codeparrot_training - Step 32503: {'lr': 0.00014677853261463281, 'samples': 16642048, 'steps': 32503, 'loss/train': 1.193146824836731} 02/25/2022 19:40:33 - INFO - codeparrot_training - Step 32504: {'lr': 0.00014676363018746087, 'samples': 16642560, 'steps': 32504, 'loss/train': 1.7393361330032349} 02/25/2022 19:40:37 - INFO - codeparrot_training - Step 32505: {'lr': 0.00014674872820252076, 'samples': 16643072, 'steps': 32505, 'loss/train': 0.5128758549690247} 02/25/2022 19:40:42 - INFO - codeparrot_training - Step 32506: {'lr': 0.00014673382665987626, 'samples': 16643584, 'steps': 32506, 'loss/train': 1.4535635709762573} 02/25/2022 19:40:46 - INFO - codeparrot_training - Step 32507: {'lr': 0.00014671892555959124, 'samples': 16644096, 'steps': 32507, 'loss/train': 2.1673965454101562} 02/25/2022 19:40:51 - INFO - codeparrot_training - Step 32508: {'lr': 0.0001467040249017296, 'samples': 16644608, 'steps': 32508, 'loss/train': 2.990795850753784} 02/25/2022 19:40:55 - INFO - codeparrot_training - Step 32509: {'lr': 0.00014668912468635513, 'samples': 16645120, 'steps': 32509, 'loss/train': 1.3850171566009521} 02/25/2022 19:41:00 - INFO - codeparrot_training - Step 32510: {'lr': 0.00014667422491353158, 'samples': 16645632, 'steps': 32510, 'loss/train': 1.5463197231292725} 02/25/2022 19:41:04 - INFO - codeparrot_training - Step 32511: {'lr': 0.00014665932558332285, 'samples': 16646144, 'steps': 32511, 'loss/train': 1.9332659244537354} 02/25/2022 19:41:10 - INFO - codeparrot_training - Step 32512: {'lr': 0.00014664442669579277, 'samples': 16646656, 'steps': 32512, 'loss/train': 2.481074571609497} 02/25/2022 19:41:14 - INFO - codeparrot_training - Step 32513: {'lr': 0.00014662952825100523, 'samples': 16647168, 'steps': 32513, 'loss/train': 1.76753568649292} 02/25/2022 19:41:19 - INFO - codeparrot_training - Step 32514: {'lr': 0.00014661463024902386, 'samples': 16647680, 'steps': 32514, 'loss/train': 1.5623332262039185} 02/25/2022 19:41:23 - INFO - codeparrot_training - Step 32515: {'lr': 0.0001465997326899126, 'samples': 16648192, 'steps': 32515, 'loss/train': 2.2548162937164307} 02/25/2022 19:41:28 - INFO - codeparrot_training - Step 32516: {'lr': 0.00014658483557373523, 'samples': 16648704, 'steps': 32516, 'loss/train': 1.596663475036621} 02/25/2022 19:41:31 - INFO - codeparrot_training - Step 32517: {'lr': 0.0001465699389005557, 'samples': 16649216, 'steps': 32517, 'loss/train': 1.362250804901123} 02/25/2022 19:41:37 - INFO - codeparrot_training - Step 32518: {'lr': 0.0001465550426704376, 'samples': 16649728, 'steps': 32518, 'loss/train': 1.934340476989746} 02/25/2022 19:41:40 - INFO - codeparrot_training - Step 32519: {'lr': 0.00014654014688344485, 'samples': 16650240, 'steps': 32519, 'loss/train': 2.4947545528411865} 02/25/2022 19:41:46 - INFO - codeparrot_training - Step 32520: {'lr': 0.0001465252515396413, 'samples': 16650752, 'steps': 32520, 'loss/train': 1.3492956161499023} 02/25/2022 19:41:49 - INFO - codeparrot_training - Step 32521: {'lr': 0.0001465103566390907, 'samples': 16651264, 'steps': 32521, 'loss/train': 1.0873219966888428} 02/25/2022 19:41:55 - INFO - codeparrot_training - Step 32522: {'lr': 0.00014649546218185695, 'samples': 16651776, 'steps': 32522, 'loss/train': 0.9952855706214905} 02/25/2022 19:41:58 - INFO - codeparrot_training - Step 32523: {'lr': 0.00014648056816800366, 'samples': 16652288, 'steps': 32523, 'loss/train': 1.4225622415542603} 02/25/2022 19:42:04 - INFO - codeparrot_training - Step 32524: {'lr': 0.00014646567459759475, 'samples': 16652800, 'steps': 32524, 'loss/train': 1.1771849393844604} 02/25/2022 19:42:07 - INFO - codeparrot_training - Step 32525: {'lr': 0.00014645078147069402, 'samples': 16653312, 'steps': 32525, 'loss/train': 1.9480422735214233} 02/25/2022 19:42:14 - INFO - codeparrot_training - Step 32526: {'lr': 0.0001464358887873654, 'samples': 16653824, 'steps': 32526, 'loss/train': 1.8456637859344482} 02/25/2022 19:42:17 - INFO - codeparrot_training - Step 32527: {'lr': 0.00014642099654767237, 'samples': 16654336, 'steps': 32527, 'loss/train': 1.543684482574463} 02/25/2022 19:42:23 - INFO - codeparrot_training - Step 32528: {'lr': 0.00014640610475167898, 'samples': 16654848, 'steps': 32528, 'loss/train': 1.0154361724853516} 02/25/2022 19:42:26 - INFO - codeparrot_training - Step 32529: {'lr': 0.00014639121339944888, 'samples': 16655360, 'steps': 32529, 'loss/train': 2.256618022918701} 02/25/2022 19:42:32 - INFO - codeparrot_training - Step 32530: {'lr': 0.00014637632249104608, 'samples': 16655872, 'steps': 32530, 'loss/train': 8.19294548034668} 02/25/2022 19:42:35 - INFO - codeparrot_training - Step 32531: {'lr': 0.00014636143202653406, 'samples': 16656384, 'steps': 32531, 'loss/train': 1.152690052986145} 02/25/2022 19:42:41 - INFO - codeparrot_training - Step 32532: {'lr': 0.0001463465420059768, 'samples': 16656896, 'steps': 32532, 'loss/train': 1.9793477058410645} 02/25/2022 19:42:44 - INFO - codeparrot_training - Step 32533: {'lr': 0.00014633165242943804, 'samples': 16657408, 'steps': 32533, 'loss/train': 1.5656373500823975} 02/25/2022 19:42:50 - INFO - codeparrot_training - Step 32534: {'lr': 0.00014631676329698152, 'samples': 16657920, 'steps': 32534, 'loss/train': 1.5136991739273071} 02/25/2022 19:42:53 - INFO - codeparrot_training - Step 32535: {'lr': 0.00014630187460867118, 'samples': 16658432, 'steps': 32535, 'loss/train': 2.1637156009674072} 02/25/2022 19:42:59 - INFO - codeparrot_training - Step 32536: {'lr': 0.0001462869863645706, 'samples': 16658944, 'steps': 32536, 'loss/train': 2.0958502292633057} 02/25/2022 19:43:02 - INFO - codeparrot_training - Step 32537: {'lr': 0.0001462720985647436, 'samples': 16659456, 'steps': 32537, 'loss/train': 2.6352405548095703} 02/25/2022 19:43:08 - INFO - codeparrot_training - Step 32538: {'lr': 0.00014625721120925406, 'samples': 16659968, 'steps': 32538, 'loss/train': 2.3502702713012695} 02/25/2022 19:43:12 - INFO - codeparrot_training - Step 32539: {'lr': 0.0001462423242981657, 'samples': 16660480, 'steps': 32539, 'loss/train': 1.1097592115402222} 02/25/2022 19:43:17 - INFO - codeparrot_training - Step 32540: {'lr': 0.0001462274378315422, 'samples': 16660992, 'steps': 32540, 'loss/train': 1.4944705963134766} 02/25/2022 19:43:21 - INFO - codeparrot_training - Step 32541: {'lr': 0.00014621255180944742, 'samples': 16661504, 'steps': 32541, 'loss/train': 0.9297956228256226} 02/25/2022 19:43:26 - INFO - codeparrot_training - Step 32542: {'lr': 0.0001461976662319452, 'samples': 16662016, 'steps': 32542, 'loss/train': 1.0841641426086426} 02/25/2022 19:43:30 - INFO - codeparrot_training - Step 32543: {'lr': 0.00014618278109909916, 'samples': 16662528, 'steps': 32543, 'loss/train': 2.223261833190918} 02/25/2022 19:43:35 - INFO - codeparrot_training - Step 32544: {'lr': 0.0001461678964109731, 'samples': 16663040, 'steps': 32544, 'loss/train': 2.282331943511963} 02/25/2022 19:43:39 - INFO - codeparrot_training - Step 32545: {'lr': 0.00014615301216763083, 'samples': 16663552, 'steps': 32545, 'loss/train': 1.8736226558685303} 02/25/2022 19:43:44 - INFO - codeparrot_training - Step 32546: {'lr': 0.00014613812836913604, 'samples': 16664064, 'steps': 32546, 'loss/train': 1.63163423538208} 02/25/2022 19:43:48 - INFO - codeparrot_training - Step 32547: {'lr': 0.0001461232450155525, 'samples': 16664576, 'steps': 32547, 'loss/train': 2.2463481426239014} 02/25/2022 19:43:54 - INFO - codeparrot_training - Step 32548: {'lr': 0.0001461083621069441, 'samples': 16665088, 'steps': 32548, 'loss/train': 2.52325177192688} 02/25/2022 19:43:57 - INFO - codeparrot_training - Step 32549: {'lr': 0.00014609347964337447, 'samples': 16665600, 'steps': 32549, 'loss/train': 1.5620390176773071} 02/25/2022 19:44:03 - INFO - codeparrot_training - Step 32550: {'lr': 0.00014607859762490733, 'samples': 16666112, 'steps': 32550, 'loss/train': 2.4591023921966553} 02/25/2022 19:44:06 - INFO - codeparrot_training - Step 32551: {'lr': 0.00014606371605160645, 'samples': 16666624, 'steps': 32551, 'loss/train': 1.6290680170059204} 02/25/2022 19:44:12 - INFO - codeparrot_training - Step 32552: {'lr': 0.0001460488349235357, 'samples': 16667136, 'steps': 32552, 'loss/train': 0.312151163816452} 02/25/2022 19:44:15 - INFO - codeparrot_training - Step 32553: {'lr': 0.00014603395424075868, 'samples': 16667648, 'steps': 32553, 'loss/train': 2.084946632385254} 02/25/2022 19:44:21 - INFO - codeparrot_training - Step 32554: {'lr': 0.00014601907400333918, 'samples': 16668160, 'steps': 32554, 'loss/train': 1.6850183010101318} 02/25/2022 19:44:24 - INFO - codeparrot_training - Step 32555: {'lr': 0.00014600419421134092, 'samples': 16668672, 'steps': 32555, 'loss/train': 1.7700989246368408} 02/25/2022 19:44:30 - INFO - codeparrot_training - Step 32556: {'lr': 0.0001459893148648278, 'samples': 16669184, 'steps': 32556, 'loss/train': 1.63833749294281} 02/25/2022 19:44:33 - INFO - codeparrot_training - Step 32557: {'lr': 0.0001459744359638633, 'samples': 16669696, 'steps': 32557, 'loss/train': 3.292646884918213} 02/25/2022 19:44:40 - INFO - codeparrot_training - Step 32558: {'lr': 0.00014595955750851126, 'samples': 16670208, 'steps': 32558, 'loss/train': 2.00081205368042} 02/25/2022 19:44:45 - INFO - codeparrot_training - Step 32559: {'lr': 0.00014594467949883552, 'samples': 16670720, 'steps': 32559, 'loss/train': 0.06072288751602173} 02/25/2022 19:44:49 - INFO - codeparrot_training - Step 32560: {'lr': 0.00014592980193489974, 'samples': 16671232, 'steps': 32560, 'loss/train': 2.46502947807312} 02/25/2022 19:44:52 - INFO - codeparrot_training - Step 32561: {'lr': 0.00014591492481676765, 'samples': 16671744, 'steps': 32561, 'loss/train': 3.4827399253845215} 02/25/2022 19:44:58 - INFO - codeparrot_training - Step 32562: {'lr': 0.00014590004814450287, 'samples': 16672256, 'steps': 32562, 'loss/train': 2.4513795375823975} 02/25/2022 19:45:03 - INFO - codeparrot_training - Step 32563: {'lr': 0.0001458851719181693, 'samples': 16672768, 'steps': 32563, 'loss/train': 2.3921098709106445} 02/25/2022 19:45:07 - INFO - codeparrot_training - Step 32564: {'lr': 0.00014587029613783063, 'samples': 16673280, 'steps': 32564, 'loss/train': 2.1346254348754883} 02/25/2022 19:45:12 - INFO - codeparrot_training - Step 32565: {'lr': 0.00014585542080355053, 'samples': 16673792, 'steps': 32565, 'loss/train': 4.798409461975098} 02/25/2022 19:45:16 - INFO - codeparrot_training - Step 32566: {'lr': 0.00014584054591539264, 'samples': 16674304, 'steps': 32566, 'loss/train': 1.7423176765441895} 02/25/2022 19:45:19 - INFO - codeparrot_training - Step 32567: {'lr': 0.00014582567147342085, 'samples': 16674816, 'steps': 32567, 'loss/train': 2.152841329574585} 02/25/2022 19:45:25 - INFO - codeparrot_training - Step 32568: {'lr': 0.00014581079747769886, 'samples': 16675328, 'steps': 32568, 'loss/train': 0.5792051553726196} 02/25/2022 19:45:29 - INFO - codeparrot_training - Step 32569: {'lr': 0.00014579592392829015, 'samples': 16675840, 'steps': 32569, 'loss/train': 1.7976208925247192} 02/25/2022 19:45:34 - INFO - codeparrot_training - Step 32570: {'lr': 0.0001457810508252588, 'samples': 16676352, 'steps': 32570, 'loss/train': 2.345736503601074} 02/25/2022 19:45:40 - INFO - codeparrot_training - Step 32571: {'lr': 0.00014576617816866827, 'samples': 16676864, 'steps': 32571, 'loss/train': 2.0118987560272217} 02/25/2022 19:45:43 - INFO - codeparrot_training - Step 32572: {'lr': 0.00014575130595858237, 'samples': 16677376, 'steps': 32572, 'loss/train': 1.2332576513290405} 02/25/2022 19:45:49 - INFO - codeparrot_training - Step 32573: {'lr': 0.00014573643419506466, 'samples': 16677888, 'steps': 32573, 'loss/train': 1.5530755519866943} 02/25/2022 19:45:52 - INFO - codeparrot_training - Step 32574: {'lr': 0.00014572156287817918, 'samples': 16678400, 'steps': 32574, 'loss/train': 1.4895217418670654} 02/25/2022 19:45:58 - INFO - codeparrot_training - Step 32575: {'lr': 0.0001457066920079892, 'samples': 16678912, 'steps': 32575, 'loss/train': 0.6072016954421997} 02/25/2022 19:46:01 - INFO - codeparrot_training - Step 32576: {'lr': 0.00014569182158455873, 'samples': 16679424, 'steps': 32576, 'loss/train': 1.2757741212844849} 02/25/2022 19:46:07 - INFO - codeparrot_training - Step 32577: {'lr': 0.00014567695160795128, 'samples': 16679936, 'steps': 32577, 'loss/train': 1.829365611076355} 02/25/2022 19:46:10 - INFO - codeparrot_training - Step 32578: {'lr': 0.00014566208207823084, 'samples': 16680448, 'steps': 32578, 'loss/train': 1.710878849029541} 02/25/2022 19:46:16 - INFO - codeparrot_training - Step 32579: {'lr': 0.00014564721299546068, 'samples': 16680960, 'steps': 32579, 'loss/train': 0.3618599474430084} 02/25/2022 19:46:19 - INFO - codeparrot_training - Step 32580: {'lr': 0.00014563234435970487, 'samples': 16681472, 'steps': 32580, 'loss/train': 2.0613863468170166} 02/25/2022 19:46:25 - INFO - codeparrot_training - Step 32581: {'lr': 0.00014561747617102683, 'samples': 16681984, 'steps': 32581, 'loss/train': 2.2673733234405518} 02/25/2022 19:46:29 - INFO - codeparrot_training - Step 32582: {'lr': 0.00014560260842949048, 'samples': 16682496, 'steps': 32582, 'loss/train': 2.2760581970214844} 02/25/2022 19:46:32 - INFO - codeparrot_training - Step 32583: {'lr': 0.00014558774113515943, 'samples': 16683008, 'steps': 32583, 'loss/train': 1.891520619392395} 02/25/2022 19:46:39 - INFO - codeparrot_training - Step 32584: {'lr': 0.00014557287428809733, 'samples': 16683520, 'steps': 32584, 'loss/train': 1.7041987180709839} 02/25/2022 19:46:42 - INFO - codeparrot_training - Step 32585: {'lr': 0.00014555800788836775, 'samples': 16684032, 'steps': 32585, 'loss/train': 2.355768918991089} 02/25/2022 19:46:48 - INFO - codeparrot_training - Step 32586: {'lr': 0.00014554314193603463, 'samples': 16684544, 'steps': 32586, 'loss/train': 1.9019343852996826} 02/25/2022 19:46:51 - INFO - codeparrot_training - Step 32587: {'lr': 0.0001455282764311615, 'samples': 16685056, 'steps': 32587, 'loss/train': 1.5980372428894043} 02/25/2022 19:46:56 - INFO - codeparrot_training - Step 32588: {'lr': 0.00014551341137381208, 'samples': 16685568, 'steps': 32588, 'loss/train': 2.3350040912628174} 02/25/2022 19:47:00 - INFO - codeparrot_training - Step 32589: {'lr': 0.00014549854676404992, 'samples': 16686080, 'steps': 32589, 'loss/train': 2.4348292350769043} 02/25/2022 19:47:06 - INFO - codeparrot_training - Step 32590: {'lr': 0.00014548368260193888, 'samples': 16686592, 'steps': 32590, 'loss/train': 0.8817393183708191} 02/25/2022 19:47:09 - INFO - codeparrot_training - Step 32591: {'lr': 0.0001454688188875426, 'samples': 16687104, 'steps': 32591, 'loss/train': 1.4493517875671387} 02/25/2022 19:47:14 - INFO - codeparrot_training - Step 32592: {'lr': 0.00014545395562092468, 'samples': 16687616, 'steps': 32592, 'loss/train': 1.8948256969451904} 02/25/2022 19:47:21 - INFO - codeparrot_training - Step 32593: {'lr': 0.00014543909280214873, 'samples': 16688128, 'steps': 32593, 'loss/train': 0.7754912376403809} 02/25/2022 19:47:24 - INFO - codeparrot_training - Step 32594: {'lr': 0.0001454242304312786, 'samples': 16688640, 'steps': 32594, 'loss/train': 1.6801598072052002} 02/25/2022 19:47:29 - INFO - codeparrot_training - Step 32595: {'lr': 0.00014540936850837775, 'samples': 16689152, 'steps': 32595, 'loss/train': 1.9028607606887817} 02/25/2022 19:47:33 - INFO - codeparrot_training - Step 32596: {'lr': 0.00014539450703351015, 'samples': 16689664, 'steps': 32596, 'loss/train': 0.9734052419662476} 02/25/2022 19:47:38 - INFO - codeparrot_training - Step 32597: {'lr': 0.00014537964600673907, 'samples': 16690176, 'steps': 32597, 'loss/train': 2.524299144744873} 02/25/2022 19:47:42 - INFO - codeparrot_training - Step 32598: {'lr': 0.00014536478542812847, 'samples': 16690688, 'steps': 32598, 'loss/train': 1.862568974494934} 02/25/2022 19:47:48 - INFO - codeparrot_training - Step 32599: {'lr': 0.00014534992529774182, 'samples': 16691200, 'steps': 32599, 'loss/train': 0.26445573568344116} 02/25/2022 19:47:51 - INFO - codeparrot_training - Step 32600: {'lr': 0.00014533506561564306, 'samples': 16691712, 'steps': 32600, 'loss/train': 1.1494972705841064} 02/25/2022 19:47:56 - INFO - codeparrot_training - Step 32601: {'lr': 0.0001453202063818954, 'samples': 16692224, 'steps': 32601, 'loss/train': 1.9411702156066895} 02/25/2022 19:48:00 - INFO - codeparrot_training - Step 32602: {'lr': 0.0001453053475965629, 'samples': 16692736, 'steps': 32602, 'loss/train': 1.9713162183761597} 02/25/2022 19:48:06 - INFO - codeparrot_training - Step 32603: {'lr': 0.00014529048925970888, 'samples': 16693248, 'steps': 32603, 'loss/train': 2.7197704315185547} 02/25/2022 19:48:10 - INFO - codeparrot_training - Step 32604: {'lr': 0.0001452756313713974, 'samples': 16693760, 'steps': 32604, 'loss/train': 1.9436630010604858} 02/25/2022 19:48:15 - INFO - codeparrot_training - Step 32605: {'lr': 0.00014526077393169166, 'samples': 16694272, 'steps': 32605, 'loss/train': 2.1888067722320557} 02/25/2022 19:48:19 - INFO - codeparrot_training - Step 32606: {'lr': 0.00014524591694065558, 'samples': 16694784, 'steps': 32606, 'loss/train': 1.585457444190979} 02/25/2022 19:48:24 - INFO - codeparrot_training - Step 32607: {'lr': 0.00014523106039835277, 'samples': 16695296, 'steps': 32607, 'loss/train': 1.813966989517212} 02/25/2022 19:48:28 - INFO - codeparrot_training - Step 32608: {'lr': 0.00014521620430484668, 'samples': 16695808, 'steps': 32608, 'loss/train': 2.113527774810791} 02/25/2022 19:48:33 - INFO - codeparrot_training - Step 32609: {'lr': 0.00014520134866020124, 'samples': 16696320, 'steps': 32609, 'loss/train': 1.8030260801315308} 02/25/2022 19:48:37 - INFO - codeparrot_training - Step 32610: {'lr': 0.00014518649346447992, 'samples': 16696832, 'steps': 32610, 'loss/train': 1.4430670738220215} 02/25/2022 19:48:42 - INFO - codeparrot_training - Step 32611: {'lr': 0.00014517163871774637, 'samples': 16697344, 'steps': 32611, 'loss/train': 1.356073021888733} 02/25/2022 19:48:46 - INFO - codeparrot_training - Step 32612: {'lr': 0.00014515678442006416, 'samples': 16697856, 'steps': 32612, 'loss/train': 2.759971857070923} 02/25/2022 19:48:51 - INFO - codeparrot_training - Step 32613: {'lr': 0.0001451419305714971, 'samples': 16698368, 'steps': 32613, 'loss/train': 1.8395285606384277} 02/25/2022 19:48:55 - INFO - codeparrot_training - Step 32614: {'lr': 0.00014512707717210868, 'samples': 16698880, 'steps': 32614, 'loss/train': 1.7062286138534546} 02/25/2022 19:49:01 - INFO - codeparrot_training - Step 32615: {'lr': 0.0001451122242219626, 'samples': 16699392, 'steps': 32615, 'loss/train': 2.3146731853485107} 02/25/2022 19:49:04 - INFO - codeparrot_training - Step 32616: {'lr': 0.00014509737172112232, 'samples': 16699904, 'steps': 32616, 'loss/train': 1.404025912284851} 02/25/2022 19:49:10 - INFO - codeparrot_training - Step 32617: {'lr': 0.0001450825196696517, 'samples': 16700416, 'steps': 32617, 'loss/train': 1.3700178861618042} 02/25/2022 19:49:13 - INFO - codeparrot_training - Step 32618: {'lr': 0.00014506766806761422, 'samples': 16700928, 'steps': 32618, 'loss/train': 2.280946969985962} 02/25/2022 19:49:19 - INFO - codeparrot_training - Step 32619: {'lr': 0.00014505281691507354, 'samples': 16701440, 'steps': 32619, 'loss/train': 1.6797267198562622} 02/25/2022 19:49:22 - INFO - codeparrot_training - Step 32620: {'lr': 0.0001450379662120932, 'samples': 16701952, 'steps': 32620, 'loss/train': 1.0425136089324951} 02/25/2022 19:49:28 - INFO - codeparrot_training - Step 32621: {'lr': 0.00014502311595873695, 'samples': 16702464, 'steps': 32621, 'loss/train': 0.9134673476219177} 02/25/2022 19:49:31 - INFO - codeparrot_training - Step 32622: {'lr': 0.00014500826615506838, 'samples': 16702976, 'steps': 32622, 'loss/train': 1.7135289907455444} 02/25/2022 19:49:37 - INFO - codeparrot_training - Step 32623: {'lr': 0.00014499341680115102, 'samples': 16703488, 'steps': 32623, 'loss/train': 1.8786330223083496} 02/25/2022 19:49:40 - INFO - codeparrot_training - Step 32624: {'lr': 0.00014497856789704843, 'samples': 16704000, 'steps': 32624, 'loss/train': 1.5957815647125244} 02/25/2022 19:49:46 - INFO - codeparrot_training - Step 32625: {'lr': 0.00014496371944282442, 'samples': 16704512, 'steps': 32625, 'loss/train': 1.5695611238479614} 02/25/2022 19:49:49 - INFO - codeparrot_training - Step 32626: {'lr': 0.00014494887143854247, 'samples': 16705024, 'steps': 32626, 'loss/train': 1.8321589231491089} 02/25/2022 19:49:55 - INFO - codeparrot_training - Step 32627: {'lr': 0.0001449340238842662, 'samples': 16705536, 'steps': 32627, 'loss/train': 2.1155900955200195} 02/25/2022 19:49:58 - INFO - codeparrot_training - Step 32628: {'lr': 0.0001449191767800591, 'samples': 16706048, 'steps': 32628, 'loss/train': 2.2852180004119873} 02/25/2022 19:50:04 - INFO - codeparrot_training - Step 32629: {'lr': 0.000144904330125985, 'samples': 16706560, 'steps': 32629, 'loss/train': 1.7796056270599365} 02/25/2022 19:50:08 - INFO - codeparrot_training - Step 32630: {'lr': 0.00014488948392210724, 'samples': 16707072, 'steps': 32630, 'loss/train': 2.3164689540863037} 02/25/2022 19:50:13 - INFO - codeparrot_training - Step 32631: {'lr': 0.00014487463816848978, 'samples': 16707584, 'steps': 32631, 'loss/train': 2.30336594581604} 02/25/2022 19:50:17 - INFO - codeparrot_training - Step 32632: {'lr': 0.00014485979286519578, 'samples': 16708096, 'steps': 32632, 'loss/train': 0.7707721590995789} 02/25/2022 19:50:22 - INFO - codeparrot_training - Step 32633: {'lr': 0.00014484494801228915, 'samples': 16708608, 'steps': 32633, 'loss/train': 2.6576788425445557} 02/25/2022 19:50:26 - INFO - codeparrot_training - Step 32634: {'lr': 0.00014483010360983329, 'samples': 16709120, 'steps': 32634, 'loss/train': 1.0598995685577393} 02/25/2022 19:50:31 - INFO - codeparrot_training - Step 32635: {'lr': 0.00014481525965789205, 'samples': 16709632, 'steps': 32635, 'loss/train': 1.0487781763076782} 02/25/2022 19:50:35 - INFO - codeparrot_training - Step 32636: {'lr': 0.00014480041615652864, 'samples': 16710144, 'steps': 32636, 'loss/train': 1.2969876527786255} 02/25/2022 19:50:40 - INFO - codeparrot_training - Step 32637: {'lr': 0.00014478557310580696, 'samples': 16710656, 'steps': 32637, 'loss/train': 0.21595679223537445} 02/25/2022 19:50:44 - INFO - codeparrot_training - Step 32638: {'lr': 0.00014477073050579034, 'samples': 16711168, 'steps': 32638, 'loss/train': 2.559903383255005} 02/25/2022 19:50:50 - INFO - codeparrot_training - Step 32639: {'lr': 0.00014475588835654275, 'samples': 16711680, 'steps': 32639, 'loss/train': 1.2127933502197266} 02/25/2022 19:50:54 - INFO - codeparrot_training - Step 32640: {'lr': 0.00014474104665812727, 'samples': 16712192, 'steps': 32640, 'loss/train': 2.8421521186828613} 02/25/2022 19:50:59 - INFO - codeparrot_training - Step 32641: {'lr': 0.00014472620541060782, 'samples': 16712704, 'steps': 32641, 'loss/train': 1.2105140686035156} 02/25/2022 19:51:03 - INFO - codeparrot_training - Step 32642: {'lr': 0.00014471136461404778, 'samples': 16713216, 'steps': 32642, 'loss/train': 2.8805480003356934} 02/25/2022 19:51:08 - INFO - codeparrot_training - Step 32643: {'lr': 0.00014469652426851094, 'samples': 16713728, 'steps': 32643, 'loss/train': 1.6034473180770874} 02/25/2022 19:51:12 - INFO - codeparrot_training - Step 32644: {'lr': 0.00014468168437406075, 'samples': 16714240, 'steps': 32644, 'loss/train': 0.9617477655410767} 02/25/2022 19:51:17 - INFO - codeparrot_training - Step 32645: {'lr': 0.00014466684493076077, 'samples': 16714752, 'steps': 32645, 'loss/train': 1.3194661140441895} 02/25/2022 19:51:21 - INFO - codeparrot_training - Step 32646: {'lr': 0.0001446520059386745, 'samples': 16715264, 'steps': 32646, 'loss/train': 1.4790725708007812} 02/25/2022 19:51:26 - INFO - codeparrot_training - Step 32647: {'lr': 0.00014463716739786565, 'samples': 16715776, 'steps': 32647, 'loss/train': 0.9794721007347107} 02/25/2022 19:51:30 - INFO - codeparrot_training - Step 32648: {'lr': 0.00014462232930839776, 'samples': 16716288, 'steps': 32648, 'loss/train': 1.3576139211654663} 02/25/2022 19:51:36 - INFO - codeparrot_training - Step 32649: {'lr': 0.00014460749167033432, 'samples': 16716800, 'steps': 32649, 'loss/train': 0.7192365527153015} 02/25/2022 19:51:40 - INFO - codeparrot_training - Step 32650: {'lr': 0.00014459265448373883, 'samples': 16717312, 'steps': 32650, 'loss/train': 2.836864709854126} 02/25/2022 19:51:45 - INFO - codeparrot_training - Step 32651: {'lr': 0.00014457781774867504, 'samples': 16717824, 'steps': 32651, 'loss/train': 0.8483403325080872} 02/25/2022 19:51:49 - INFO - codeparrot_training - Step 32652: {'lr': 0.0001445629814652064, 'samples': 16718336, 'steps': 32652, 'loss/train': 2.5078163146972656} 02/25/2022 19:51:54 - INFO - codeparrot_training - Step 32653: {'lr': 0.00014454814563339647, 'samples': 16718848, 'steps': 32653, 'loss/train': 1.23050057888031} 02/25/2022 19:51:57 - INFO - codeparrot_training - Step 32654: {'lr': 0.0001445333102533088, 'samples': 16719360, 'steps': 32654, 'loss/train': 2.6685192584991455} 02/25/2022 19:52:03 - INFO - codeparrot_training - Step 32655: {'lr': 0.00014451847532500684, 'samples': 16719872, 'steps': 32655, 'loss/train': 1.9983913898468018} 02/25/2022 19:52:06 - INFO - codeparrot_training - Step 32656: {'lr': 0.00014450364084855433, 'samples': 16720384, 'steps': 32656, 'loss/train': 1.031639814376831} 02/25/2022 19:52:12 - INFO - codeparrot_training - Step 32657: {'lr': 0.0001444888068240147, 'samples': 16720896, 'steps': 32657, 'loss/train': 1.767969012260437} 02/25/2022 19:52:15 - INFO - codeparrot_training - Step 32658: {'lr': 0.00014447397325145157, 'samples': 16721408, 'steps': 32658, 'loss/train': 0.7126955389976501} 02/25/2022 19:52:21 - INFO - codeparrot_training - Step 32659: {'lr': 0.0001444591401309283, 'samples': 16721920, 'steps': 32659, 'loss/train': 2.202270984649658} 02/25/2022 19:52:27 - INFO - codeparrot_training - Step 32660: {'lr': 0.00014444430746250866, 'samples': 16722432, 'steps': 32660, 'loss/train': 0.9081114530563354} 02/25/2022 19:52:31 - INFO - codeparrot_training - Step 32661: {'lr': 0.0001444294752462561, 'samples': 16722944, 'steps': 32661, 'loss/train': 1.5721038579940796} 02/25/2022 19:52:36 - INFO - codeparrot_training - Step 32662: {'lr': 0.00014441464348223415, 'samples': 16723456, 'steps': 32662, 'loss/train': 1.8994137048721313} 02/25/2022 19:52:40 - INFO - codeparrot_training - Step 32663: {'lr': 0.00014439981217050625, 'samples': 16723968, 'steps': 32663, 'loss/train': 2.639317512512207} 02/25/2022 19:52:43 - INFO - codeparrot_training - Step 32664: {'lr': 0.0001443849813111361, 'samples': 16724480, 'steps': 32664, 'loss/train': 2.0904910564422607} 02/25/2022 19:52:49 - INFO - codeparrot_training - Step 32665: {'lr': 0.00014437015090418715, 'samples': 16724992, 'steps': 32665, 'loss/train': 1.0360610485076904} 02/25/2022 19:52:52 - INFO - codeparrot_training - Step 32666: {'lr': 0.00014435532094972292, 'samples': 16725504, 'steps': 32666, 'loss/train': 0.766703724861145} 02/25/2022 19:52:58 - INFO - codeparrot_training - Step 32667: {'lr': 0.00014434049144780686, 'samples': 16726016, 'steps': 32667, 'loss/train': 1.8390617370605469} 02/25/2022 19:53:03 - INFO - codeparrot_training - Step 32668: {'lr': 0.0001443256623985027, 'samples': 16726528, 'steps': 32668, 'loss/train': 2.619417428970337} 02/25/2022 19:53:07 - INFO - codeparrot_training - Step 32669: {'lr': 0.00014431083380187377, 'samples': 16727040, 'steps': 32669, 'loss/train': 1.761723518371582} 02/25/2022 19:53:13 - INFO - codeparrot_training - Step 32670: {'lr': 0.00014429600565798385, 'samples': 16727552, 'steps': 32670, 'loss/train': 1.877333164215088} 02/25/2022 19:53:16 - INFO - codeparrot_training - Step 32671: {'lr': 0.00014428117796689606, 'samples': 16728064, 'steps': 32671, 'loss/train': 0.5953896045684814} 02/25/2022 19:53:20 - INFO - codeparrot_training - Step 32672: {'lr': 0.00014426635072867423, 'samples': 16728576, 'steps': 32672, 'loss/train': 0.05128796398639679} 02/25/2022 19:53:25 - INFO - codeparrot_training - Step 32673: {'lr': 0.00014425152394338168, 'samples': 16729088, 'steps': 32673, 'loss/train': 1.9015426635742188} 02/25/2022 19:53:28 - INFO - codeparrot_training - Step 32674: {'lr': 0.00014423669761108222, 'samples': 16729600, 'steps': 32674, 'loss/train': 1.4324573278427124} 02/25/2022 19:53:35 - INFO - codeparrot_training - Step 32675: {'lr': 0.000144221871731839, 'samples': 16730112, 'steps': 32675, 'loss/train': 1.5122604370117188} 02/25/2022 19:53:38 - INFO - codeparrot_training - Step 32676: {'lr': 0.00014420704630571573, 'samples': 16730624, 'steps': 32676, 'loss/train': 2.054382562637329} 02/25/2022 19:53:44 - INFO - codeparrot_training - Step 32677: {'lr': 0.00014419222133277586, 'samples': 16731136, 'steps': 32677, 'loss/train': 0.923971951007843} 02/25/2022 19:53:47 - INFO - codeparrot_training - Step 32678: {'lr': 0.00014417739681308296, 'samples': 16731648, 'steps': 32678, 'loss/train': 2.005750894546509} 02/25/2022 19:53:53 - INFO - codeparrot_training - Step 32679: {'lr': 0.0001441625727467005, 'samples': 16732160, 'steps': 32679, 'loss/train': 0.8481064438819885} 02/25/2022 19:53:56 - INFO - codeparrot_training - Step 32680: {'lr': 0.000144147749133692, 'samples': 16732672, 'steps': 32680, 'loss/train': 2.0719780921936035} 02/25/2022 19:54:02 - INFO - codeparrot_training - Step 32681: {'lr': 0.00014413292597412078, 'samples': 16733184, 'steps': 32681, 'loss/train': 1.6500420570373535} 02/25/2022 19:54:06 - INFO - codeparrot_training - Step 32682: {'lr': 0.00014411810326805065, 'samples': 16733696, 'steps': 32682, 'loss/train': 0.3047272264957428} 02/25/2022 19:54:11 - INFO - codeparrot_training - Step 32683: {'lr': 0.00014410328101554487, 'samples': 16734208, 'steps': 32683, 'loss/train': 1.6220779418945312} 02/25/2022 19:54:15 - INFO - codeparrot_training - Step 32684: {'lr': 0.00014408845921666706, 'samples': 16734720, 'steps': 32684, 'loss/train': 1.9734541177749634} 02/25/2022 19:54:20 - INFO - codeparrot_training - Step 32685: {'lr': 0.00014407363787148053, 'samples': 16735232, 'steps': 32685, 'loss/train': 1.9033491611480713} 02/25/2022 19:54:24 - INFO - codeparrot_training - Step 32686: {'lr': 0.000144058816980049, 'samples': 16735744, 'steps': 32686, 'loss/train': 1.803678274154663} 02/25/2022 19:54:30 - INFO - codeparrot_training - Step 32687: {'lr': 0.0001440439965424359, 'samples': 16736256, 'steps': 32687, 'loss/train': 1.5141916275024414} 02/25/2022 19:54:33 - INFO - codeparrot_training - Step 32688: {'lr': 0.00014402917655870464, 'samples': 16736768, 'steps': 32688, 'loss/train': 0.7240457534790039} 02/25/2022 19:54:39 - INFO - codeparrot_training - Step 32689: {'lr': 0.00014401435702891867, 'samples': 16737280, 'steps': 32689, 'loss/train': 1.481002926826477} 02/25/2022 19:54:42 - INFO - codeparrot_training - Step 32690: {'lr': 0.0001439995379531416, 'samples': 16737792, 'steps': 32690, 'loss/train': 1.670120358467102} 02/25/2022 19:54:48 - INFO - codeparrot_training - Step 32691: {'lr': 0.00014398471933143686, 'samples': 16738304, 'steps': 32691, 'loss/train': 1.852657675743103} 02/25/2022 19:54:51 - INFO - codeparrot_training - Step 32692: {'lr': 0.00014396990116386792, 'samples': 16738816, 'steps': 32692, 'loss/train': 2.0642690658569336} 02/25/2022 19:54:57 - INFO - codeparrot_training - Step 32693: {'lr': 0.0001439550834504982, 'samples': 16739328, 'steps': 32693, 'loss/train': 2.485919713973999} 02/25/2022 19:55:00 - INFO - codeparrot_training - Step 32694: {'lr': 0.00014394026619139128, 'samples': 16739840, 'steps': 32694, 'loss/train': 0.9345873594284058} 02/25/2022 19:55:06 - INFO - codeparrot_training - Step 32695: {'lr': 0.0001439254493866106, 'samples': 16740352, 'steps': 32695, 'loss/train': 1.3722742795944214} 02/25/2022 19:55:09 - INFO - codeparrot_training - Step 32696: {'lr': 0.0001439106330362196, 'samples': 16740864, 'steps': 32696, 'loss/train': 2.4172208309173584} 02/25/2022 19:55:15 - INFO - codeparrot_training - Step 32697: {'lr': 0.00014389581714028166, 'samples': 16741376, 'steps': 32697, 'loss/train': 1.4686102867126465} 02/25/2022 19:55:19 - INFO - codeparrot_training - Step 32698: {'lr': 0.00014388100169886045, 'samples': 16741888, 'steps': 32698, 'loss/train': 1.51478910446167} 02/25/2022 19:55:24 - INFO - codeparrot_training - Step 32699: {'lr': 0.00014386618671201933, 'samples': 16742400, 'steps': 32699, 'loss/train': 1.7136051654815674} 02/25/2022 19:55:28 - INFO - codeparrot_training - Step 32700: {'lr': 0.00014385137217982178, 'samples': 16742912, 'steps': 32700, 'loss/train': 0.48612454533576965} 02/25/2022 19:55:33 - INFO - codeparrot_training - Step 32701: {'lr': 0.00014383655810233125, 'samples': 16743424, 'steps': 32701, 'loss/train': 1.5959501266479492} 02/25/2022 19:55:37 - INFO - codeparrot_training - Step 32702: {'lr': 0.00014382174447961105, 'samples': 16743936, 'steps': 32702, 'loss/train': 1.6617835760116577} 02/25/2022 19:55:42 - INFO - codeparrot_training - Step 32703: {'lr': 0.00014380693131172493, 'samples': 16744448, 'steps': 32703, 'loss/train': 2.0757088661193848} 02/25/2022 19:55:46 - INFO - codeparrot_training - Step 32704: {'lr': 0.0001437921185987361, 'samples': 16744960, 'steps': 32704, 'loss/train': 2.3541574478149414} 02/25/2022 19:55:51 - INFO - codeparrot_training - Step 32705: {'lr': 0.00014377730634070827, 'samples': 16745472, 'steps': 32705, 'loss/train': 1.3640838861465454} 02/25/2022 19:55:55 - INFO - codeparrot_training - Step 32706: {'lr': 0.00014376249453770454, 'samples': 16745984, 'steps': 32706, 'loss/train': 1.4453153610229492} 02/25/2022 19:56:01 - INFO - codeparrot_training - Step 32707: {'lr': 0.00014374768318978865, 'samples': 16746496, 'steps': 32707, 'loss/train': 0.6854455471038818} 02/25/2022 19:56:06 - INFO - codeparrot_training - Step 32708: {'lr': 0.00014373287229702388, 'samples': 16747008, 'steps': 32708, 'loss/train': 1.5794174671173096} 02/25/2022 19:56:10 - INFO - codeparrot_training - Step 32709: {'lr': 0.0001437180618594739, 'samples': 16747520, 'steps': 32709, 'loss/train': 2.129223108291626} 02/25/2022 19:56:15 - INFO - codeparrot_training - Step 32710: {'lr': 0.00014370325187720178, 'samples': 16748032, 'steps': 32710, 'loss/train': 2.1325604915618896} 02/25/2022 19:56:19 - INFO - codeparrot_training - Step 32711: {'lr': 0.00014368844235027135, 'samples': 16748544, 'steps': 32711, 'loss/train': 1.5141938924789429} 02/25/2022 19:56:24 - INFO - codeparrot_training - Step 32712: {'lr': 0.00014367363327874572, 'samples': 16749056, 'steps': 32712, 'loss/train': 1.8307890892028809} 02/25/2022 19:56:28 - INFO - codeparrot_training - Step 32713: {'lr': 0.0001436588246626887, 'samples': 16749568, 'steps': 32713, 'loss/train': 1.2590118646621704} 02/25/2022 19:56:33 - INFO - codeparrot_training - Step 32714: {'lr': 0.0001436440165021633, 'samples': 16750080, 'steps': 32714, 'loss/train': 1.5183470249176025} 02/25/2022 19:56:37 - INFO - codeparrot_training - Step 32715: {'lr': 0.00014362920879723324, 'samples': 16750592, 'steps': 32715, 'loss/train': 2.2522971630096436} 02/25/2022 19:56:42 - INFO - codeparrot_training - Step 32716: {'lr': 0.0001436144015479618, 'samples': 16751104, 'steps': 32716, 'loss/train': 1.300566554069519} 02/25/2022 19:56:46 - INFO - codeparrot_training - Step 32717: {'lr': 0.00014359959475441258, 'samples': 16751616, 'steps': 32717, 'loss/train': 1.5501042604446411} 02/25/2022 19:56:51 - INFO - codeparrot_training - Step 32718: {'lr': 0.0001435847884166489, 'samples': 16752128, 'steps': 32718, 'loss/train': 1.1983247995376587} 02/25/2022 19:56:55 - INFO - codeparrot_training - Step 32719: {'lr': 0.0001435699825347342, 'samples': 16752640, 'steps': 32719, 'loss/train': 1.5109435319900513} 02/25/2022 19:57:00 - INFO - codeparrot_training - Step 32720: {'lr': 0.00014355517710873183, 'samples': 16753152, 'steps': 32720, 'loss/train': 1.50523841381073} 02/25/2022 19:57:04 - INFO - codeparrot_training - Step 32721: {'lr': 0.00014354037213870535, 'samples': 16753664, 'steps': 32721, 'loss/train': 1.6309013366699219} 02/25/2022 19:57:11 - INFO - codeparrot_training - Step 32722: {'lr': 0.0001435255676247181, 'samples': 16754176, 'steps': 32722, 'loss/train': 1.374910593032837} 02/25/2022 19:57:14 - INFO - codeparrot_training - Step 32723: {'lr': 0.0001435107635668335, 'samples': 16754688, 'steps': 32723, 'loss/train': 2.370889186859131} 02/25/2022 19:57:20 - INFO - codeparrot_training - Step 32724: {'lr': 0.00014349595996511493, 'samples': 16755200, 'steps': 32724, 'loss/train': 1.9662412405014038} 02/25/2022 19:57:23 - INFO - codeparrot_training - Step 32725: {'lr': 0.00014348115681962593, 'samples': 16755712, 'steps': 32725, 'loss/train': 0.6598528027534485} 02/25/2022 19:57:29 - INFO - codeparrot_training - Step 32726: {'lr': 0.00014346635413042968, 'samples': 16756224, 'steps': 32726, 'loss/train': 2.1078903675079346} 02/25/2022 19:57:32 - INFO - codeparrot_training - Step 32727: {'lr': 0.00014345155189759003, 'samples': 16756736, 'steps': 32727, 'loss/train': 2.4211621284484863} 02/25/2022 19:57:38 - INFO - codeparrot_training - Step 32728: {'lr': 0.00014343675012116984, 'samples': 16757248, 'steps': 32728, 'loss/train': 1.2968860864639282} 02/25/2022 19:57:41 - INFO - codeparrot_training - Step 32729: {'lr': 0.0001434219488012329, 'samples': 16757760, 'steps': 32729, 'loss/train': 2.9250118732452393} 02/25/2022 19:57:47 - INFO - codeparrot_training - Step 32730: {'lr': 0.00014340714793784233, 'samples': 16758272, 'steps': 32730, 'loss/train': 1.5965840816497803} 02/25/2022 19:57:50 - INFO - codeparrot_training - Step 32731: {'lr': 0.00014339234753106195, 'samples': 16758784, 'steps': 32731, 'loss/train': 1.2012847661972046} 02/25/2022 19:57:56 - INFO - codeparrot_training - Step 32732: {'lr': 0.00014337754758095468, 'samples': 16759296, 'steps': 32732, 'loss/train': 2.1962292194366455} 02/25/2022 19:58:00 - INFO - codeparrot_training - Step 32733: {'lr': 0.00014336274808758427, 'samples': 16759808, 'steps': 32733, 'loss/train': 2.402853488922119} 02/25/2022 19:58:05 - INFO - codeparrot_training - Step 32734: {'lr': 0.00014334794905101396, 'samples': 16760320, 'steps': 32734, 'loss/train': 2.621136426925659} 02/25/2022 19:58:09 - INFO - codeparrot_training - Step 32735: {'lr': 0.00014333315047130712, 'samples': 16760832, 'steps': 32735, 'loss/train': 1.6332625150680542} 02/25/2022 19:58:14 - INFO - codeparrot_training - Step 32736: {'lr': 0.00014331835234852717, 'samples': 16761344, 'steps': 32736, 'loss/train': 2.4094061851501465} 02/25/2022 19:58:18 - INFO - codeparrot_training - Step 32737: {'lr': 0.0001433035546827376, 'samples': 16761856, 'steps': 32737, 'loss/train': 2.4797909259796143} 02/25/2022 19:58:23 - INFO - codeparrot_training - Step 32738: {'lr': 0.00014328875747400172, 'samples': 16762368, 'steps': 32738, 'loss/train': 1.9192359447479248} 02/25/2022 19:58:27 - INFO - codeparrot_training - Step 32739: {'lr': 0.0001432739607223828, 'samples': 16762880, 'steps': 32739, 'loss/train': 2.141568660736084} 02/25/2022 19:58:32 - INFO - codeparrot_training - Step 32740: {'lr': 0.00014325916442794445, 'samples': 16763392, 'steps': 32740, 'loss/train': 2.0738911628723145} 02/25/2022 19:58:36 - INFO - codeparrot_training - Step 32741: {'lr': 0.00014324436859074996, 'samples': 16763904, 'steps': 32741, 'loss/train': 3.309900999069214} 02/25/2022 19:58:42 - INFO - codeparrot_training - Step 32742: {'lr': 0.0001432295732108627, 'samples': 16764416, 'steps': 32742, 'loss/train': 1.4107065200805664} 02/25/2022 19:58:45 - INFO - codeparrot_training - Step 32743: {'lr': 0.00014321477828834595, 'samples': 16764928, 'steps': 32743, 'loss/train': 0.8776313066482544} 02/25/2022 19:58:51 - INFO - codeparrot_training - Step 32744: {'lr': 0.00014319998382326328, 'samples': 16765440, 'steps': 32744, 'loss/train': 1.152510166168213} 02/25/2022 19:58:54 - INFO - codeparrot_training - Step 32745: {'lr': 0.00014318518981567792, 'samples': 16765952, 'steps': 32745, 'loss/train': 1.1839841604232788} 02/25/2022 19:59:00 - INFO - codeparrot_training - Step 32746: {'lr': 0.00014317039626565335, 'samples': 16766464, 'steps': 32746, 'loss/train': 1.3747538328170776} 02/25/2022 19:59:04 - INFO - codeparrot_training - Step 32747: {'lr': 0.00014315560317325275, 'samples': 16766976, 'steps': 32747, 'loss/train': 1.5950841903686523} 02/25/2022 19:59:09 - INFO - codeparrot_training - Step 32748: {'lr': 0.00014314081053853984, 'samples': 16767488, 'steps': 32748, 'loss/train': 1.55852210521698} 02/25/2022 19:59:13 - INFO - codeparrot_training - Step 32749: {'lr': 0.00014312601836157752, 'samples': 16768000, 'steps': 32749, 'loss/train': 2.1252944469451904} 02/25/2022 19:59:18 - INFO - codeparrot_training - Step 32750: {'lr': 0.00014311122664242953, 'samples': 16768512, 'steps': 32750, 'loss/train': 1.4113901853561401} 02/25/2022 19:59:22 - INFO - codeparrot_training - Step 32751: {'lr': 0.00014309643538115902, 'samples': 16769024, 'steps': 32751, 'loss/train': 2.2619760036468506} 02/25/2022 19:59:28 - INFO - codeparrot_training - Step 32752: {'lr': 0.00014308164457782951, 'samples': 16769536, 'steps': 32752, 'loss/train': 1.678704023361206} 02/25/2022 19:59:31 - INFO - codeparrot_training - Step 32753: {'lr': 0.0001430668542325043, 'samples': 16770048, 'steps': 32753, 'loss/train': 0.7408945560455322} 02/25/2022 19:59:37 - INFO - codeparrot_training - Step 32754: {'lr': 0.0001430520643452467, 'samples': 16770560, 'steps': 32754, 'loss/train': 2.108042001724243} 02/25/2022 19:59:40 - INFO - codeparrot_training - Step 32755: {'lr': 0.00014303727491612003, 'samples': 16771072, 'steps': 32755, 'loss/train': 2.2555551528930664} 02/25/2022 19:59:46 - INFO - codeparrot_training - Step 32756: {'lr': 0.0001430224859451878, 'samples': 16771584, 'steps': 32756, 'loss/train': 2.203788995742798} 02/25/2022 19:59:49 - INFO - codeparrot_training - Step 32757: {'lr': 0.0001430076974325133, 'samples': 16772096, 'steps': 32757, 'loss/train': 1.6723406314849854} 02/25/2022 19:59:55 - INFO - codeparrot_training - Step 32758: {'lr': 0.00014299290937815979, 'samples': 16772608, 'steps': 32758, 'loss/train': 1.8768762350082397} 02/25/2022 19:59:58 - INFO - codeparrot_training - Step 32759: {'lr': 0.0001429781217821906, 'samples': 16773120, 'steps': 32759, 'loss/train': 1.6370434761047363} 02/25/2022 20:00:04 - INFO - codeparrot_training - Step 32760: {'lr': 0.00014296333464466928, 'samples': 16773632, 'steps': 32760, 'loss/train': 1.656333088874817} 02/25/2022 20:00:07 - INFO - codeparrot_training - Step 32761: {'lr': 0.00014294854796565904, 'samples': 16774144, 'steps': 32761, 'loss/train': 2.308772563934326} 02/25/2022 20:00:13 - INFO - codeparrot_training - Step 32762: {'lr': 0.00014293376174522322, 'samples': 16774656, 'steps': 32762, 'loss/train': 1.7694058418273926} 02/25/2022 20:00:16 - INFO - codeparrot_training - Step 32763: {'lr': 0.00014291897598342506, 'samples': 16775168, 'steps': 32763, 'loss/train': 0.5264816880226135} 02/25/2022 20:00:22 - INFO - codeparrot_training - Step 32764: {'lr': 0.00014290419068032812, 'samples': 16775680, 'steps': 32764, 'loss/train': 2.178091287612915} 02/25/2022 20:00:25 - INFO - codeparrot_training - Step 32765: {'lr': 0.0001428894058359955, 'samples': 16776192, 'steps': 32765, 'loss/train': 1.227644681930542} 02/25/2022 20:00:31 - INFO - codeparrot_training - Step 32766: {'lr': 0.00014287462145049084, 'samples': 16776704, 'steps': 32766, 'loss/train': 1.404919147491455} 02/25/2022 20:00:34 - INFO - codeparrot_training - Step 32767: {'lr': 0.00014285983752387714, 'samples': 16777216, 'steps': 32767, 'loss/train': 1.2115883827209473} 02/25/2022 20:00:40 - INFO - codeparrot_training - Step 32768: {'lr': 0.00014284505405621795, 'samples': 16777728, 'steps': 32768, 'loss/train': 0.5569152235984802} 02/25/2022 20:00:44 - INFO - codeparrot_training - Step 32769: {'lr': 0.0001428302710475764, 'samples': 16778240, 'steps': 32769, 'loss/train': 1.3821772336959839} 02/25/2022 20:00:49 - INFO - codeparrot_training - Step 32770: {'lr': 0.00014281548849801617, 'samples': 16778752, 'steps': 32770, 'loss/train': 0.4669029712677002} 02/25/2022 20:00:55 - INFO - codeparrot_training - Step 32771: {'lr': 0.00014280070640760011, 'samples': 16779264, 'steps': 32771, 'loss/train': 0.9726061224937439} 02/25/2022 20:00:58 - INFO - codeparrot_training - Step 32772: {'lr': 0.00014278592477639195, 'samples': 16779776, 'steps': 32772, 'loss/train': 1.459582805633545} 02/25/2022 20:01:04 - INFO - codeparrot_training - Step 32773: {'lr': 0.0001427711436044547, 'samples': 16780288, 'steps': 32773, 'loss/train': 3.1283576488494873} 02/25/2022 20:01:07 - INFO - codeparrot_training - Step 32774: {'lr': 0.000142756362891852, 'samples': 16780800, 'steps': 32774, 'loss/train': 0.8975827693939209} 02/25/2022 20:01:13 - INFO - codeparrot_training - Step 32775: {'lr': 0.00014274158263864685, 'samples': 16781312, 'steps': 32775, 'loss/train': 2.010094404220581} 02/25/2022 20:01:16 - INFO - codeparrot_training - Step 32776: {'lr': 0.00014272680284490276, 'samples': 16781824, 'steps': 32776, 'loss/train': 1.9739956855773926} 02/25/2022 20:01:23 - INFO - codeparrot_training - Step 32777: {'lr': 0.00014271202351068287, 'samples': 16782336, 'steps': 32777, 'loss/train': 1.9123178720474243} 02/25/2022 20:01:27 - INFO - codeparrot_training - Step 32778: {'lr': 0.00014269724463605074, 'samples': 16782848, 'steps': 32778, 'loss/train': 1.6754976511001587} 02/25/2022 20:01:32 - INFO - codeparrot_training - Step 32779: {'lr': 0.00014268246622106952, 'samples': 16783360, 'steps': 32779, 'loss/train': 2.1960582733154297} 02/25/2022 20:01:36 - INFO - codeparrot_training - Step 32780: {'lr': 0.00014266768826580255, 'samples': 16783872, 'steps': 32780, 'loss/train': 1.1884649991989136} 02/25/2022 20:01:41 - INFO - codeparrot_training - Step 32781: {'lr': 0.00014265291077031312, 'samples': 16784384, 'steps': 32781, 'loss/train': 1.93251633644104} 02/25/2022 20:01:45 - INFO - codeparrot_training - Step 32782: {'lr': 0.00014263813373466445, 'samples': 16784896, 'steps': 32782, 'loss/train': 1.337907075881958} 02/25/2022 20:01:50 - INFO - codeparrot_training - Step 32783: {'lr': 0.00014262335715891999, 'samples': 16785408, 'steps': 32783, 'loss/train': 0.6538954377174377} 02/25/2022 20:01:54 - INFO - codeparrot_training - Step 32784: {'lr': 0.00014260858104314298, 'samples': 16785920, 'steps': 32784, 'loss/train': 2.4194469451904297} 02/25/2022 20:01:59 - INFO - codeparrot_training - Step 32785: {'lr': 0.0001425938053873967, 'samples': 16786432, 'steps': 32785, 'loss/train': 0.6877073049545288} 02/25/2022 20:02:03 - INFO - codeparrot_training - Step 32786: {'lr': 0.00014257903019174436, 'samples': 16786944, 'steps': 32786, 'loss/train': 2.034318447113037} 02/25/2022 20:02:09 - INFO - codeparrot_training - Step 32787: {'lr': 0.00014256425545624947, 'samples': 16787456, 'steps': 32787, 'loss/train': 0.8257609009742737} 02/25/2022 20:02:13 - INFO - codeparrot_training - Step 32788: {'lr': 0.00014254948118097517, 'samples': 16787968, 'steps': 32788, 'loss/train': 1.5751572847366333} 02/25/2022 20:02:18 - INFO - codeparrot_training - Step 32789: {'lr': 0.00014253470736598478, 'samples': 16788480, 'steps': 32789, 'loss/train': 1.5892513990402222} 02/25/2022 20:02:22 - INFO - codeparrot_training - Step 32790: {'lr': 0.00014251993401134146, 'samples': 16788992, 'steps': 32790, 'loss/train': 1.7945740222930908} 02/25/2022 20:02:27 - INFO - codeparrot_training - Step 32791: {'lr': 0.00014250516111710877, 'samples': 16789504, 'steps': 32791, 'loss/train': 0.5350008010864258} 02/25/2022 20:02:31 - INFO - codeparrot_training - Step 32792: {'lr': 0.0001424903886833498, 'samples': 16790016, 'steps': 32792, 'loss/train': 2.1410818099975586} 02/25/2022 20:02:36 - INFO - codeparrot_training - Step 32793: {'lr': 0.00014247561671012785, 'samples': 16790528, 'steps': 32793, 'loss/train': 1.8263975381851196} 02/25/2022 20:02:40 - INFO - codeparrot_training - Step 32794: {'lr': 0.00014246084519750613, 'samples': 16791040, 'steps': 32794, 'loss/train': 1.4393744468688965} 02/25/2022 20:02:45 - INFO - codeparrot_training - Step 32795: {'lr': 0.0001424460741455481, 'samples': 16791552, 'steps': 32795, 'loss/train': 2.066148281097412} 02/25/2022 20:02:49 - INFO - codeparrot_training - Step 32796: {'lr': 0.00014243130355431693, 'samples': 16792064, 'steps': 32796, 'loss/train': 1.7517191171646118} 02/25/2022 20:02:56 - INFO - codeparrot_training - Step 32797: {'lr': 0.0001424165334238759, 'samples': 16792576, 'steps': 32797, 'loss/train': 1.7475212812423706} 02/25/2022 20:02:59 - INFO - codeparrot_training - Step 32798: {'lr': 0.0001424017637542882, 'samples': 16793088, 'steps': 32798, 'loss/train': 0.4339800775051117} 02/25/2022 20:03:05 - INFO - codeparrot_training - Step 32799: {'lr': 0.00014238699454561726, 'samples': 16793600, 'steps': 32799, 'loss/train': 1.0771853923797607} 02/25/2022 20:03:08 - INFO - codeparrot_training - Step 32800: {'lr': 0.00014237222579792616, 'samples': 16794112, 'steps': 32800, 'loss/train': 0.9307793378829956} 02/25/2022 20:03:14 - INFO - codeparrot_training - Step 32801: {'lr': 0.00014235745751127847, 'samples': 16794624, 'steps': 32801, 'loss/train': 2.038780927658081} 02/25/2022 20:03:17 - INFO - codeparrot_training - Step 32802: {'lr': 0.00014234268968573707, 'samples': 16795136, 'steps': 32802, 'loss/train': 1.3817161321640015} 02/25/2022 20:03:23 - INFO - codeparrot_training - Step 32803: {'lr': 0.00014232792232136548, 'samples': 16795648, 'steps': 32803, 'loss/train': 1.59222412109375} 02/25/2022 20:03:26 - INFO - codeparrot_training - Step 32804: {'lr': 0.00014231315541822682, 'samples': 16796160, 'steps': 32804, 'loss/train': 1.6140148639678955} 02/25/2022 20:03:32 - INFO - codeparrot_training - Step 32805: {'lr': 0.0001422983889763846, 'samples': 16796672, 'steps': 32805, 'loss/train': 0.5027738213539124} 02/25/2022 20:03:35 - INFO - codeparrot_training - Step 32806: {'lr': 0.00014228362299590163, 'samples': 16797184, 'steps': 32806, 'loss/train': 2.640164613723755} 02/25/2022 20:03:39 - INFO - codeparrot_training - Step 32807: {'lr': 0.00014226885747684154, 'samples': 16797696, 'steps': 32807, 'loss/train': 8.759848594665527} 02/25/2022 20:03:45 - INFO - codeparrot_training - Step 32808: {'lr': 0.00014225409241926739, 'samples': 16798208, 'steps': 32808, 'loss/train': 2.3352572917938232} 02/25/2022 20:03:49 - INFO - codeparrot_training - Step 32809: {'lr': 0.00014223932782324268, 'samples': 16798720, 'steps': 32809, 'loss/train': 1.123358130455017} 02/25/2022 20:03:54 - INFO - codeparrot_training - Step 32810: {'lr': 0.00014222456368883024, 'samples': 16799232, 'steps': 32810, 'loss/train': 1.2572427988052368} 02/25/2022 20:03:58 - INFO - codeparrot_training - Step 32811: {'lr': 0.00014220980001609365, 'samples': 16799744, 'steps': 32811, 'loss/train': 2.2566707134246826} 02/25/2022 20:04:04 - INFO - codeparrot_training - Step 32812: {'lr': 0.000142195036805096, 'samples': 16800256, 'steps': 32812, 'loss/train': 1.9251874685287476} 02/25/2022 20:04:07 - INFO - codeparrot_training - Step 32813: {'lr': 0.0001421802740559006, 'samples': 16800768, 'steps': 32813, 'loss/train': 0.09611745923757553} 02/25/2022 20:04:13 - INFO - codeparrot_training - Step 32814: {'lr': 0.00014216551176857072, 'samples': 16801280, 'steps': 32814, 'loss/train': 1.021488904953003} 02/25/2022 20:04:16 - INFO - codeparrot_training - Step 32815: {'lr': 0.00014215074994316956, 'samples': 16801792, 'steps': 32815, 'loss/train': 1.315518856048584} 02/25/2022 20:04:22 - INFO - codeparrot_training - Step 32816: {'lr': 0.00014213598857976023, 'samples': 16802304, 'steps': 32816, 'loss/train': 2.037801504135132} 02/25/2022 20:04:27 - INFO - codeparrot_training - Step 32817: {'lr': 0.00014212122767840615, 'samples': 16802816, 'steps': 32817, 'loss/train': 0.9158303141593933} 02/25/2022 20:04:31 - INFO - codeparrot_training - Step 32818: {'lr': 0.0001421064672391705, 'samples': 16803328, 'steps': 32818, 'loss/train': 2.670583963394165} 02/25/2022 20:04:37 - INFO - codeparrot_training - Step 32819: {'lr': 0.00014209170726211647, 'samples': 16803840, 'steps': 32819, 'loss/train': 1.5096617937088013} 02/25/2022 20:04:41 - INFO - codeparrot_training - Step 32820: {'lr': 0.00014207694774730722, 'samples': 16804352, 'steps': 32820, 'loss/train': 1.8073617219924927} 02/25/2022 20:04:46 - INFO - codeparrot_training - Step 32821: {'lr': 0.00014206218869480613, 'samples': 16804864, 'steps': 32821, 'loss/train': 0.5628405809402466} 02/25/2022 20:04:50 - INFO - codeparrot_training - Step 32822: {'lr': 0.0001420474301046764, 'samples': 16805376, 'steps': 32822, 'loss/train': 1.4855765104293823} 02/25/2022 20:04:55 - INFO - codeparrot_training - Step 32823: {'lr': 0.00014203267197698116, 'samples': 16805888, 'steps': 32823, 'loss/train': 1.5590139627456665} 02/25/2022 20:04:59 - INFO - codeparrot_training - Step 32824: {'lr': 0.00014201791431178357, 'samples': 16806400, 'steps': 32824, 'loss/train': 1.2142512798309326} 02/25/2022 20:05:04 - INFO - codeparrot_training - Step 32825: {'lr': 0.00014200315710914704, 'samples': 16806912, 'steps': 32825, 'loss/train': 0.8279922604560852} 02/25/2022 20:05:08 - INFO - codeparrot_training - Step 32826: {'lr': 0.00014198840036913472, 'samples': 16807424, 'steps': 32826, 'loss/train': 0.8543905019760132} 02/25/2022 20:05:13 - INFO - codeparrot_training - Step 32827: {'lr': 0.00014197364409180977, 'samples': 16807936, 'steps': 32827, 'loss/train': 1.1598032712936401} 02/25/2022 20:05:17 - INFO - codeparrot_training - Step 32828: {'lr': 0.00014195888827723535, 'samples': 16808448, 'steps': 32828, 'loss/train': 1.534693717956543} 02/25/2022 20:05:22 - INFO - codeparrot_training - Step 32829: {'lr': 0.00014194413292547482, 'samples': 16808960, 'steps': 32829, 'loss/train': 1.9712468385696411} 02/25/2022 20:05:25 - INFO - codeparrot_training - Step 32830: {'lr': 0.00014192937803659135, 'samples': 16809472, 'steps': 32830, 'loss/train': 1.5513536930084229} 02/25/2022 20:05:31 - INFO - codeparrot_training - Step 32831: {'lr': 0.00014191462361064805, 'samples': 16809984, 'steps': 32831, 'loss/train': 2.1370108127593994} 02/25/2022 20:05:34 - INFO - codeparrot_training - Step 32832: {'lr': 0.00014189986964770825, 'samples': 16810496, 'steps': 32832, 'loss/train': 1.3149220943450928} 02/25/2022 20:05:40 - INFO - codeparrot_training - Step 32833: {'lr': 0.0001418851161478349, 'samples': 16811008, 'steps': 32833, 'loss/train': 2.668180227279663} 02/25/2022 20:05:44 - INFO - codeparrot_training - Step 32834: {'lr': 0.00014187036311109154, 'samples': 16811520, 'steps': 32834, 'loss/train': 1.194145917892456} 02/25/2022 20:05:50 - INFO - codeparrot_training - Step 32835: {'lr': 0.0001418556105375411, 'samples': 16812032, 'steps': 32835, 'loss/train': 2.2563109397888184} 02/25/2022 20:05:53 - INFO - codeparrot_training - Step 32836: {'lr': 0.00014184085842724708, 'samples': 16812544, 'steps': 32836, 'loss/train': 1.9797940254211426} 02/25/2022 20:05:59 - INFO - codeparrot_training - Step 32837: {'lr': 0.0001418261067802723, 'samples': 16813056, 'steps': 32837, 'loss/train': 2.155033826828003} 02/25/2022 20:06:02 - INFO - codeparrot_training - Step 32838: {'lr': 0.00014181135559668018, 'samples': 16813568, 'steps': 32838, 'loss/train': 0.993283748626709} 02/25/2022 20:06:08 - INFO - codeparrot_training - Step 32839: {'lr': 0.0001417966048765338, 'samples': 16814080, 'steps': 32839, 'loss/train': 1.9749411344528198} 02/25/2022 20:06:11 - INFO - codeparrot_training - Step 32840: {'lr': 0.0001417818546198966, 'samples': 16814592, 'steps': 32840, 'loss/train': 1.293125033378601} 02/25/2022 20:06:17 - INFO - codeparrot_training - Step 32841: {'lr': 0.00014176710482683135, 'samples': 16815104, 'steps': 32841, 'loss/train': 2.366556406021118} 02/25/2022 20:06:20 - INFO - codeparrot_training - Step 32842: {'lr': 0.0001417523554974016, 'samples': 16815616, 'steps': 32842, 'loss/train': 2.071415424346924} 02/25/2022 20:06:26 - INFO - codeparrot_training - Step 32843: {'lr': 0.00014173760663167023, 'samples': 16816128, 'steps': 32843, 'loss/train': 1.687054991722107} 02/25/2022 20:06:30 - INFO - codeparrot_training - Step 32844: {'lr': 0.00014172285822970082, 'samples': 16816640, 'steps': 32844, 'loss/train': 1.7060917615890503} 02/25/2022 20:06:35 - INFO - codeparrot_training - Step 32845: {'lr': 0.00014170811029155613, 'samples': 16817152, 'steps': 32845, 'loss/train': 2.132366418838501} 02/25/2022 20:06:39 - INFO - codeparrot_training - Step 32846: {'lr': 0.00014169336281729956, 'samples': 16817664, 'steps': 32846, 'loss/train': 1.6334660053253174} 02/25/2022 20:06:44 - INFO - codeparrot_training - Step 32847: {'lr': 0.0001416786158069941, 'samples': 16818176, 'steps': 32847, 'loss/train': 1.469781756401062} 02/25/2022 20:06:48 - INFO - codeparrot_training - Step 32848: {'lr': 0.0001416638692607032, 'samples': 16818688, 'steps': 32848, 'loss/train': 2.0682122707366943} 02/25/2022 20:06:53 - INFO - codeparrot_training - Step 32849: {'lr': 0.00014164912317848988, 'samples': 16819200, 'steps': 32849, 'loss/train': 1.5972638130187988} 02/25/2022 20:06:57 - INFO - codeparrot_training - Step 32850: {'lr': 0.0001416343775604173, 'samples': 16819712, 'steps': 32850, 'loss/train': 0.36140212416648865} 02/25/2022 20:07:02 - INFO - codeparrot_training - Step 32851: {'lr': 0.00014161963240654857, 'samples': 16820224, 'steps': 32851, 'loss/train': 2.0866732597351074} 02/25/2022 20:07:06 - INFO - codeparrot_training - Step 32852: {'lr': 0.000141604887716947, 'samples': 16820736, 'steps': 32852, 'loss/train': 1.1813610792160034} 02/25/2022 20:07:11 - INFO - codeparrot_training - Step 32853: {'lr': 0.00014159014349167566, 'samples': 16821248, 'steps': 32853, 'loss/train': 1.0654414892196655} 02/25/2022 20:07:15 - INFO - codeparrot_training - Step 32854: {'lr': 0.0001415753997307977, 'samples': 16821760, 'steps': 32854, 'loss/train': 1.992229700088501} 02/25/2022 20:07:21 - INFO - codeparrot_training - Step 32855: {'lr': 0.00014156065643437628, 'samples': 16822272, 'steps': 32855, 'loss/train': 1.2218401432037354} 02/25/2022 20:07:24 - INFO - codeparrot_training - Step 32856: {'lr': 0.0001415459136024746, 'samples': 16822784, 'steps': 32856, 'loss/train': 1.2177947759628296} 02/25/2022 20:07:30 - INFO - codeparrot_training - Step 32857: {'lr': 0.0001415311712351558, 'samples': 16823296, 'steps': 32857, 'loss/train': 1.235691785812378} 02/25/2022 20:07:33 - INFO - codeparrot_training - Step 32858: {'lr': 0.00014151642933248304, 'samples': 16823808, 'steps': 32858, 'loss/train': 2.2234504222869873} 02/25/2022 20:07:39 - INFO - codeparrot_training - Step 32859: {'lr': 0.00014150168789451933, 'samples': 16824320, 'steps': 32859, 'loss/train': 1.8202614784240723} 02/25/2022 20:07:42 - INFO - codeparrot_training - Step 32860: {'lr': 0.00014148694692132804, 'samples': 16824832, 'steps': 32860, 'loss/train': 2.070849657058716} 02/25/2022 20:07:48 - INFO - codeparrot_training - Step 32861: {'lr': 0.00014147220641297213, 'samples': 16825344, 'steps': 32861, 'loss/train': 0.8616870045661926} 02/25/2022 20:07:51 - INFO - codeparrot_training - Step 32862: {'lr': 0.000141457466369515, 'samples': 16825856, 'steps': 32862, 'loss/train': 0.6561073660850525} 02/25/2022 20:07:57 - INFO - codeparrot_training - Step 32863: {'lr': 0.00014144272679101944, 'samples': 16826368, 'steps': 32863, 'loss/train': 1.1476329565048218} 02/25/2022 20:08:00 - INFO - codeparrot_training - Step 32864: {'lr': 0.00014142798767754886, 'samples': 16826880, 'steps': 32864, 'loss/train': 1.9029687643051147} 02/25/2022 20:08:06 - INFO - codeparrot_training - Step 32865: {'lr': 0.0001414132490291663, 'samples': 16827392, 'steps': 32865, 'loss/train': 1.0413790941238403} 02/25/2022 20:08:09 - INFO - codeparrot_training - Step 32866: {'lr': 0.0001413985108459349, 'samples': 16827904, 'steps': 32866, 'loss/train': 1.1820979118347168} 02/25/2022 20:08:15 - INFO - codeparrot_training - Step 32867: {'lr': 0.00014138377312791772, 'samples': 16828416, 'steps': 32867, 'loss/train': 2.109445095062256} 02/25/2022 20:08:18 - INFO - codeparrot_training - Step 32868: {'lr': 0.00014136903587517804, 'samples': 16828928, 'steps': 32868, 'loss/train': 1.9847474098205566} 02/25/2022 20:08:24 - INFO - codeparrot_training - Step 32869: {'lr': 0.0001413542990877789, 'samples': 16829440, 'steps': 32869, 'loss/train': 1.7672747373580933} 02/25/2022 20:08:28 - INFO - codeparrot_training - Step 32870: {'lr': 0.0001413395627657835, 'samples': 16829952, 'steps': 32870, 'loss/train': 1.5557104349136353} 02/25/2022 20:08:34 - INFO - codeparrot_training - Step 32871: {'lr': 0.00014132482690925476, 'samples': 16830464, 'steps': 32871, 'loss/train': 1.9749342203140259} 02/25/2022 20:08:39 - INFO - codeparrot_training - Step 32872: {'lr': 0.00014131009151825607, 'samples': 16830976, 'steps': 32872, 'loss/train': 1.632596492767334} 02/25/2022 20:08:42 - INFO - codeparrot_training - Step 32873: {'lr': 0.00014129535659285046, 'samples': 16831488, 'steps': 32873, 'loss/train': 1.1747798919677734} 02/25/2022 20:08:48 - INFO - codeparrot_training - Step 32874: {'lr': 0.0001412806221331009, 'samples': 16832000, 'steps': 32874, 'loss/train': 0.8990028500556946} 02/25/2022 20:08:52 - INFO - codeparrot_training - Step 32875: {'lr': 0.0001412658881390707, 'samples': 16832512, 'steps': 32875, 'loss/train': 2.813483238220215} 02/25/2022 20:08:57 - INFO - codeparrot_training - Step 32876: {'lr': 0.00014125115461082293, 'samples': 16833024, 'steps': 32876, 'loss/train': 1.6831914186477661} 02/25/2022 20:09:00 - INFO - codeparrot_training - Step 32877: {'lr': 0.00014123642154842072, 'samples': 16833536, 'steps': 32877, 'loss/train': 2.7684366703033447} 02/25/2022 20:09:07 - INFO - codeparrot_training - Step 32878: {'lr': 0.00014122168895192702, 'samples': 16834048, 'steps': 32878, 'loss/train': 1.8163495063781738} 02/25/2022 20:09:10 - INFO - codeparrot_training - Step 32879: {'lr': 0.00014120695682140529, 'samples': 16834560, 'steps': 32879, 'loss/train': 1.284104347229004} 02/25/2022 20:09:16 - INFO - codeparrot_training - Step 32880: {'lr': 0.00014119222515691815, 'samples': 16835072, 'steps': 32880, 'loss/train': 0.9736930131912231} 02/25/2022 20:09:19 - INFO - codeparrot_training - Step 32881: {'lr': 0.0001411774939585291, 'samples': 16835584, 'steps': 32881, 'loss/train': 1.817723274230957} 02/25/2022 20:09:25 - INFO - codeparrot_training - Step 32882: {'lr': 0.000141162763226301, 'samples': 16836096, 'steps': 32882, 'loss/train': 0.9431028366088867} 02/25/2022 20:09:28 - INFO - codeparrot_training - Step 32883: {'lr': 0.00014114803296029726, 'samples': 16836608, 'steps': 32883, 'loss/train': 1.243801236152649} 02/25/2022 20:09:34 - INFO - codeparrot_training - Step 32884: {'lr': 0.0001411333031605806, 'samples': 16837120, 'steps': 32884, 'loss/train': 1.7301125526428223} 02/25/2022 20:09:37 - INFO - codeparrot_training - Step 32885: {'lr': 0.0001411185738272144, 'samples': 16837632, 'steps': 32885, 'loss/train': 1.5202199220657349} 02/25/2022 20:09:43 - INFO - codeparrot_training - Step 32886: {'lr': 0.00014110384496026157, 'samples': 16838144, 'steps': 32886, 'loss/train': 0.9737198352813721} 02/25/2022 20:09:46 - INFO - codeparrot_training - Step 32887: {'lr': 0.00014108911655978535, 'samples': 16838656, 'steps': 32887, 'loss/train': 1.0447450876235962} 02/25/2022 20:09:50 - INFO - codeparrot_training - Step 32888: {'lr': 0.00014107438862584883, 'samples': 16839168, 'steps': 32888, 'loss/train': 1.9728931188583374} 02/25/2022 20:09:56 - INFO - codeparrot_training - Step 32889: {'lr': 0.00014105966115851497, 'samples': 16839680, 'steps': 32889, 'loss/train': 8.358400344848633} 02/25/2022 20:09:59 - INFO - codeparrot_training - Step 32890: {'lr': 0.00014104493415784686, 'samples': 16840192, 'steps': 32890, 'loss/train': 1.3235503435134888} 02/25/2022 20:10:05 - INFO - codeparrot_training - Step 32891: {'lr': 0.00014103020762390778, 'samples': 16840704, 'steps': 32891, 'loss/train': 3.6748952865600586} 02/25/2022 20:10:09 - INFO - codeparrot_training - Step 32892: {'lr': 0.00014101548155676064, 'samples': 16841216, 'steps': 32892, 'loss/train': 0.6811104416847229} 02/25/2022 20:10:14 - INFO - codeparrot_training - Step 32893: {'lr': 0.0001410007559564686, 'samples': 16841728, 'steps': 32893, 'loss/train': 1.6587862968444824} 02/25/2022 20:10:17 - INFO - codeparrot_training - Step 32894: {'lr': 0.0001409860308230946, 'samples': 16842240, 'steps': 32894, 'loss/train': 2.457082748413086} 02/25/2022 20:10:23 - INFO - codeparrot_training - Step 32895: {'lr': 0.00014097130615670195, 'samples': 16842752, 'steps': 32895, 'loss/train': 0.04452995955944061} 02/25/2022 20:10:27 - INFO - codeparrot_training - Step 32896: {'lr': 0.0001409565819573535, 'samples': 16843264, 'steps': 32896, 'loss/train': 0.7368571162223816} 02/25/2022 20:10:32 - INFO - codeparrot_training - Step 32897: {'lr': 0.0001409418582251126, 'samples': 16843776, 'steps': 32897, 'loss/train': 0.04284963756799698} 02/25/2022 20:10:36 - INFO - codeparrot_training - Step 32898: {'lr': 0.00014092713496004198, 'samples': 16844288, 'steps': 32898, 'loss/train': 2.2399790287017822} 02/25/2022 20:10:41 - INFO - codeparrot_training - Step 32899: {'lr': 0.00014091241216220496, 'samples': 16844800, 'steps': 32899, 'loss/train': 1.9340670108795166} 02/25/2022 20:10:45 - INFO - codeparrot_training - Step 32900: {'lr': 0.00014089768983166444, 'samples': 16845312, 'steps': 32900, 'loss/train': 0.8082156181335449} 02/25/2022 20:10:51 - INFO - codeparrot_training - Step 32901: {'lr': 0.00014088296796848377, 'samples': 16845824, 'steps': 32901, 'loss/train': 2.0896153450012207} 02/25/2022 20:10:55 - INFO - codeparrot_training - Step 32902: {'lr': 0.00014086824657272558, 'samples': 16846336, 'steps': 32902, 'loss/train': 1.1744898557662964} 02/25/2022 20:11:00 - INFO - codeparrot_training - Step 32903: {'lr': 0.0001408535256444533, 'samples': 16846848, 'steps': 32903, 'loss/train': 3.229671001434326} 02/25/2022 20:11:04 - INFO - codeparrot_training - Step 32904: {'lr': 0.00014083880518372976, 'samples': 16847360, 'steps': 32904, 'loss/train': 1.036189317703247} 02/25/2022 20:11:09 - INFO - codeparrot_training - Step 32905: {'lr': 0.00014082408519061835, 'samples': 16847872, 'steps': 32905, 'loss/train': 2.1869750022888184} 02/25/2022 20:11:13 - INFO - codeparrot_training - Step 32906: {'lr': 0.00014080936566518166, 'samples': 16848384, 'steps': 32906, 'loss/train': 1.297544240951538} 02/25/2022 20:11:18 - INFO - codeparrot_training - Step 32907: {'lr': 0.00014079464660748305, 'samples': 16848896, 'steps': 32907, 'loss/train': 1.8179373741149902} 02/25/2022 20:11:22 - INFO - codeparrot_training - Step 32908: {'lr': 0.00014077992801758544, 'samples': 16849408, 'steps': 32908, 'loss/train': 1.12723970413208} 02/25/2022 20:11:27 - INFO - codeparrot_training - Step 32909: {'lr': 0.00014076520989555197, 'samples': 16849920, 'steps': 32909, 'loss/train': 1.931674599647522} 02/25/2022 20:11:31 - INFO - codeparrot_training - Step 32910: {'lr': 0.00014075049224144572, 'samples': 16850432, 'steps': 32910, 'loss/train': 1.2654401063919067} 02/25/2022 20:11:37 - INFO - codeparrot_training - Step 32911: {'lr': 0.00014073577505532964, 'samples': 16850944, 'steps': 32911, 'loss/train': 1.4919214248657227} 02/25/2022 20:11:40 - INFO - codeparrot_training - Step 32912: {'lr': 0.00014072105833726683, 'samples': 16851456, 'steps': 32912, 'loss/train': 1.9069550037384033} 02/25/2022 20:11:46 - INFO - codeparrot_training - Step 32913: {'lr': 0.00014070634208732019, 'samples': 16851968, 'steps': 32913, 'loss/train': 2.5669286251068115} 02/25/2022 20:11:49 - INFO - codeparrot_training - Step 32914: {'lr': 0.000140691626305553, 'samples': 16852480, 'steps': 32914, 'loss/train': 2.8780205249786377} 02/25/2022 20:11:55 - INFO - codeparrot_training - Step 32915: {'lr': 0.00014067691099202813, 'samples': 16852992, 'steps': 32915, 'loss/train': 1.3380640745162964} 02/25/2022 20:11:59 - INFO - codeparrot_training - Step 32916: {'lr': 0.0001406621961468087, 'samples': 16853504, 'steps': 32916, 'loss/train': 3.0907726287841797} 02/25/2022 20:12:04 - INFO - codeparrot_training - Step 32917: {'lr': 0.00014064748176995757, 'samples': 16854016, 'steps': 32917, 'loss/train': 1.5749415159225464} 02/25/2022 20:12:08 - INFO - codeparrot_training - Step 32918: {'lr': 0.000140632767861538, 'samples': 16854528, 'steps': 32918, 'loss/train': 1.6800742149353027} 02/25/2022 20:12:13 - INFO - codeparrot_training - Step 32919: {'lr': 0.00014061805442161296, 'samples': 16855040, 'steps': 32919, 'loss/train': 1.8837531805038452} 02/25/2022 20:12:17 - INFO - codeparrot_training - Step 32920: {'lr': 0.00014060334145024543, 'samples': 16855552, 'steps': 32920, 'loss/train': 1.7325338125228882} 02/25/2022 20:12:22 - INFO - codeparrot_training - Step 32921: {'lr': 0.00014058862894749836, 'samples': 16856064, 'steps': 32921, 'loss/train': 0.8436089158058167} 02/25/2022 20:12:26 - INFO - codeparrot_training - Step 32922: {'lr': 0.00014057391691343492, 'samples': 16856576, 'steps': 32922, 'loss/train': 2.62870192527771} 02/25/2022 20:12:31 - INFO - codeparrot_training - Step 32923: {'lr': 0.0001405592053481181, 'samples': 16857088, 'steps': 32923, 'loss/train': 1.4654548168182373} 02/25/2022 20:12:35 - INFO - codeparrot_training - Step 32924: {'lr': 0.0001405444942516109, 'samples': 16857600, 'steps': 32924, 'loss/train': 1.807477593421936} 02/25/2022 20:12:40 - INFO - codeparrot_training - Step 32925: {'lr': 0.00014052978362397622, 'samples': 16858112, 'steps': 32925, 'loss/train': 1.6300159692764282} 02/25/2022 20:12:44 - INFO - codeparrot_training - Step 32926: {'lr': 0.00014051507346527728, 'samples': 16858624, 'steps': 32926, 'loss/train': 1.9050114154815674} 02/25/2022 20:12:50 - INFO - codeparrot_training - Step 32927: {'lr': 0.00014050036377557702, 'samples': 16859136, 'steps': 32927, 'loss/train': 1.1461583375930786} 02/25/2022 20:12:53 - INFO - codeparrot_training - Step 32928: {'lr': 0.0001404856545549384, 'samples': 16859648, 'steps': 32928, 'loss/train': 1.7911757230758667} 02/25/2022 20:12:59 - INFO - codeparrot_training - Step 32929: {'lr': 0.0001404709458034244, 'samples': 16860160, 'steps': 32929, 'loss/train': 2.375532388687134} 02/25/2022 20:13:02 - INFO - codeparrot_training - Step 32930: {'lr': 0.00014045623752109815, 'samples': 16860672, 'steps': 32930, 'loss/train': 1.6886895895004272} 02/25/2022 20:13:08 - INFO - codeparrot_training - Step 32931: {'lr': 0.00014044152970802264, 'samples': 16861184, 'steps': 32931, 'loss/train': 2.639129638671875} 02/25/2022 20:13:11 - INFO - codeparrot_training - Step 32932: {'lr': 0.0001404268223642608, 'samples': 16861696, 'steps': 32932, 'loss/train': 1.8888440132141113} 02/25/2022 20:13:17 - INFO - codeparrot_training - Step 32933: {'lr': 0.00014041211548987553, 'samples': 16862208, 'steps': 32933, 'loss/train': 1.4811986684799194} 02/25/2022 20:13:20 - INFO - codeparrot_training - Step 32934: {'lr': 0.0001403974090849301, 'samples': 16862720, 'steps': 32934, 'loss/train': 1.5700603723526} 02/25/2022 20:13:26 - INFO - codeparrot_training - Step 32935: {'lr': 0.00014038270314948727, 'samples': 16863232, 'steps': 32935, 'loss/train': 1.983089566230774} 02/25/2022 20:13:29 - INFO - codeparrot_training - Step 32936: {'lr': 0.0001403679976836103, 'samples': 16863744, 'steps': 32936, 'loss/train': 2.0529675483703613} 02/25/2022 20:13:36 - INFO - codeparrot_training - Step 32937: {'lr': 0.00014035329268736186, 'samples': 16864256, 'steps': 32937, 'loss/train': 1.0987694263458252} 02/25/2022 20:13:39 - INFO - codeparrot_training - Step 32938: {'lr': 0.00014033858816080516, 'samples': 16864768, 'steps': 32938, 'loss/train': 1.6088453531265259} 02/25/2022 20:13:45 - INFO - codeparrot_training - Step 32939: {'lr': 0.00014032388410400304, 'samples': 16865280, 'steps': 32939, 'loss/train': 1.8725632429122925} 02/25/2022 20:13:48 - INFO - codeparrot_training - Step 32940: {'lr': 0.00014030918051701876, 'samples': 16865792, 'steps': 32940, 'loss/train': 2.402594566345215} 02/25/2022 20:13:54 - INFO - codeparrot_training - Step 32941: {'lr': 0.00014029447739991496, 'samples': 16866304, 'steps': 32941, 'loss/train': 1.5512654781341553} 02/25/2022 20:13:57 - INFO - codeparrot_training - Step 32942: {'lr': 0.00014027977475275483, 'samples': 16866816, 'steps': 32942, 'loss/train': 0.975470781326294} 02/25/2022 20:14:03 - INFO - codeparrot_training - Step 32943: {'lr': 0.00014026507257560125, 'samples': 16867328, 'steps': 32943, 'loss/train': 1.9085592031478882} 02/25/2022 20:14:06 - INFO - codeparrot_training - Step 32944: {'lr': 0.00014025037086851734, 'samples': 16867840, 'steps': 32944, 'loss/train': 1.1147199869155884} 02/25/2022 20:14:13 - INFO - codeparrot_training - Step 32945: {'lr': 0.00014023566963156596, 'samples': 16868352, 'steps': 32945, 'loss/train': 2.1558568477630615} 02/25/2022 20:14:16 - INFO - codeparrot_training - Step 32946: {'lr': 0.00014022096886481017, 'samples': 16868864, 'steps': 32946, 'loss/train': 2.259960651397705} 02/25/2022 20:14:20 - INFO - codeparrot_training - Step 32947: {'lr': 0.00014020626856831275, 'samples': 16869376, 'steps': 32947, 'loss/train': 1.7982772588729858} 02/25/2022 20:14:26 - INFO - codeparrot_training - Step 32948: {'lr': 0.00014019156874213695, 'samples': 16869888, 'steps': 32948, 'loss/train': 1.83175790309906} 02/25/2022 20:14:29 - INFO - codeparrot_training - Step 32949: {'lr': 0.00014017686938634554, 'samples': 16870400, 'steps': 32949, 'loss/train': 2.2320189476013184} 02/25/2022 20:14:35 - INFO - codeparrot_training - Step 32950: {'lr': 0.0001401621705010016, 'samples': 16870912, 'steps': 32950, 'loss/train': 1.8464229106903076} 02/25/2022 20:14:38 - INFO - codeparrot_training - Step 32951: {'lr': 0.00014014747208616795, 'samples': 16871424, 'steps': 32951, 'loss/train': 1.6401088237762451} 02/25/2022 20:14:44 - INFO - codeparrot_training - Step 32952: {'lr': 0.00014013277414190773, 'samples': 16871936, 'steps': 32952, 'loss/train': 0.16312041878700256} 02/25/2022 20:14:48 - INFO - codeparrot_training - Step 32953: {'lr': 0.0001401180766682838, 'samples': 16872448, 'steps': 32953, 'loss/train': 2.1603598594665527} 02/25/2022 20:14:53 - INFO - codeparrot_training - Step 32954: {'lr': 0.0001401033796653592, 'samples': 16872960, 'steps': 32954, 'loss/train': 2.182615041732788} 02/25/2022 20:14:57 - INFO - codeparrot_training - Step 32955: {'lr': 0.00014008868313319668, 'samples': 16873472, 'steps': 32955, 'loss/train': 1.686958909034729} 02/25/2022 20:15:02 - INFO - codeparrot_training - Step 32956: {'lr': 0.00014007398707185941, 'samples': 16873984, 'steps': 32956, 'loss/train': 2.1222081184387207} 02/25/2022 20:15:05 - INFO - codeparrot_training - Step 32957: {'lr': 0.00014005929148141035, 'samples': 16874496, 'steps': 32957, 'loss/train': 2.1863038539886475} 02/25/2022 20:15:12 - INFO - codeparrot_training - Step 32958: {'lr': 0.0001400445963619123, 'samples': 16875008, 'steps': 32958, 'loss/train': 2.584641933441162} 02/25/2022 20:15:16 - INFO - codeparrot_training - Step 32959: {'lr': 0.00014002990171342833, 'samples': 16875520, 'steps': 32959, 'loss/train': 1.0032398700714111} 02/25/2022 20:15:21 - INFO - codeparrot_training - Step 32960: {'lr': 0.0001400152075360212, 'samples': 16876032, 'steps': 32960, 'loss/train': 1.5014384984970093} 02/25/2022 20:15:25 - INFO - codeparrot_training - Step 32961: {'lr': 0.00014000051382975415, 'samples': 16876544, 'steps': 32961, 'loss/train': 1.2810039520263672} 02/25/2022 20:15:30 - INFO - codeparrot_training - Step 32962: {'lr': 0.00013998582059468996, 'samples': 16877056, 'steps': 32962, 'loss/train': 2.713489055633545} 02/25/2022 20:15:34 - INFO - codeparrot_training - Step 32963: {'lr': 0.00013997112783089156, 'samples': 16877568, 'steps': 32963, 'loss/train': 1.9830602407455444} 02/25/2022 20:15:39 - INFO - codeparrot_training - Step 32964: {'lr': 0.0001399564355384218, 'samples': 16878080, 'steps': 32964, 'loss/train': 1.4365499019622803} 02/25/2022 20:15:43 - INFO - codeparrot_training - Step 32965: {'lr': 0.00013994174371734385, 'samples': 16878592, 'steps': 32965, 'loss/train': 1.7163954973220825} 02/25/2022 20:15:48 - INFO - codeparrot_training - Step 32966: {'lr': 0.00013992705236772052, 'samples': 16879104, 'steps': 32966, 'loss/train': 2.427286386489868} 02/25/2022 20:15:51 - INFO - codeparrot_training - Step 32967: {'lr': 0.00013991236148961473, 'samples': 16879616, 'steps': 32967, 'loss/train': 1.2211790084838867} 02/25/2022 20:15:57 - INFO - codeparrot_training - Step 32968: {'lr': 0.00013989767108308932, 'samples': 16880128, 'steps': 32968, 'loss/train': 1.1106135845184326} 02/25/2022 20:16:01 - INFO - codeparrot_training - Step 32969: {'lr': 0.00013988298114820747, 'samples': 16880640, 'steps': 32969, 'loss/train': 0.5665706396102905} 02/25/2022 20:16:06 - INFO - codeparrot_training - Step 32970: {'lr': 0.00013986829168503184, 'samples': 16881152, 'steps': 32970, 'loss/train': 1.8018169403076172} 02/25/2022 20:16:10 - INFO - codeparrot_training - Step 32971: {'lr': 0.00013985360269362567, 'samples': 16881664, 'steps': 32971, 'loss/train': 2.2217936515808105} 02/25/2022 20:16:15 - INFO - codeparrot_training - Step 32972: {'lr': 0.00013983891417405147, 'samples': 16882176, 'steps': 32972, 'loss/train': 0.04800500348210335} 02/25/2022 20:16:21 - INFO - codeparrot_training - Step 32973: {'lr': 0.00013982422612637252, 'samples': 16882688, 'steps': 32973, 'loss/train': 1.3335580825805664} 02/25/2022 20:16:25 - INFO - codeparrot_training - Step 32974: {'lr': 0.0001398095385506515, 'samples': 16883200, 'steps': 32974, 'loss/train': 1.90020751953125} 02/25/2022 20:16:30 - INFO - codeparrot_training - Step 32975: {'lr': 0.0001397948514469516, 'samples': 16883712, 'steps': 32975, 'loss/train': 1.9262281656265259} 02/25/2022 20:16:34 - INFO - codeparrot_training - Step 32976: {'lr': 0.0001397801648153354, 'samples': 16884224, 'steps': 32976, 'loss/train': 2.8658032417297363} 02/25/2022 20:16:39 - INFO - codeparrot_training - Step 32977: {'lr': 0.00013976547865586603, 'samples': 16884736, 'steps': 32977, 'loss/train': 0.4628658592700958} 02/25/2022 20:16:43 - INFO - codeparrot_training - Step 32978: {'lr': 0.00013975079296860626, 'samples': 16885248, 'steps': 32978, 'loss/train': 2.484701156616211} 02/25/2022 20:16:48 - INFO - codeparrot_training - Step 32979: {'lr': 0.00013973610775361932, 'samples': 16885760, 'steps': 32979, 'loss/train': 1.6312931776046753} 02/25/2022 20:16:52 - INFO - codeparrot_training - Step 32980: {'lr': 0.00013972142301096763, 'samples': 16886272, 'steps': 32980, 'loss/train': 1.5839051008224487} 02/25/2022 20:16:57 - INFO - codeparrot_training - Step 32981: {'lr': 0.00013970673874071448, 'samples': 16886784, 'steps': 32981, 'loss/train': 2.063230276107788} 02/25/2022 20:17:01 - INFO - codeparrot_training - Step 32982: {'lr': 0.00013969205494292254, 'samples': 16887296, 'steps': 32982, 'loss/train': 1.2427608966827393} 02/25/2022 20:17:07 - INFO - codeparrot_training - Step 32983: {'lr': 0.00013967737161765486, 'samples': 16887808, 'steps': 32983, 'loss/train': 0.2305237203836441} 02/25/2022 20:17:10 - INFO - codeparrot_training - Step 32984: {'lr': 0.00013966268876497434, 'samples': 16888320, 'steps': 32984, 'loss/train': 2.375281572341919} 02/25/2022 20:17:16 - INFO - codeparrot_training - Step 32985: {'lr': 0.00013964800638494385, 'samples': 16888832, 'steps': 32985, 'loss/train': 1.9276641607284546} 02/25/2022 20:17:19 - INFO - codeparrot_training - Step 32986: {'lr': 0.00013963332447762612, 'samples': 16889344, 'steps': 32986, 'loss/train': 1.2853120565414429} 02/25/2022 20:17:25 - INFO - codeparrot_training - Step 32987: {'lr': 0.00013961864304308427, 'samples': 16889856, 'steps': 32987, 'loss/train': 2.719148874282837} 02/25/2022 20:17:28 - INFO - codeparrot_training - Step 32988: {'lr': 0.0001396039620813811, 'samples': 16890368, 'steps': 32988, 'loss/train': 1.9276344776153564} 02/25/2022 20:17:34 - INFO - codeparrot_training - Step 32989: {'lr': 0.00013958928159257954, 'samples': 16890880, 'steps': 32989, 'loss/train': 1.7885204553604126} 02/25/2022 20:17:37 - INFO - codeparrot_training - Step 32990: {'lr': 0.0001395746015767423, 'samples': 16891392, 'steps': 32990, 'loss/train': 1.6885907649993896} 02/25/2022 20:17:43 - INFO - codeparrot_training - Step 32991: {'lr': 0.00013955992203393253, 'samples': 16891904, 'steps': 32991, 'loss/train': 1.5914902687072754} 02/25/2022 20:17:46 - INFO - codeparrot_training - Step 32992: {'lr': 0.000139545242964213, 'samples': 16892416, 'steps': 32992, 'loss/train': 1.9701309204101562} 02/25/2022 20:17:52 - INFO - codeparrot_training - Step 32993: {'lr': 0.00013953056436764654, 'samples': 16892928, 'steps': 32993, 'loss/train': 1.6841686964035034} 02/25/2022 20:17:56 - INFO - codeparrot_training - Step 32994: {'lr': 0.000139515886244296, 'samples': 16893440, 'steps': 32994, 'loss/train': 1.2654300928115845} 02/25/2022 20:18:01 - INFO - codeparrot_training - Step 32995: {'lr': 0.00013950120859422438, 'samples': 16893952, 'steps': 32995, 'loss/train': 1.7574375867843628} 02/25/2022 20:18:05 - INFO - codeparrot_training - Step 32996: {'lr': 0.0001394865314174945, 'samples': 16894464, 'steps': 32996, 'loss/train': 1.605767011642456} 02/25/2022 20:18:11 - INFO - codeparrot_training - Step 32997: {'lr': 0.00013947185471416927, 'samples': 16894976, 'steps': 32997, 'loss/train': 1.5172046422958374} 02/25/2022 20:18:14 - INFO - codeparrot_training - Step 32998: {'lr': 0.0001394571784843114, 'samples': 16895488, 'steps': 32998, 'loss/train': 2.2089178562164307} 02/25/2022 20:18:20 - INFO - codeparrot_training - Step 32999: {'lr': 0.00013944250272798393, 'samples': 16896000, 'steps': 32999, 'loss/train': 1.2068666219711304} 02/25/2022 20:18:20 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 20:18:38 - WARNING - huggingface_hub.repository - Several commits (33) will be pushed upstream. 02/25/2022 20:18:38 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 20:19:11 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 06c1d4e..6b82659 floral-grass-11 -> floral-grass-11 02/25/2022 20:19:16 - INFO - codeparrot_training - Step 33000: {'lr': 0.00013942782744524973, 'samples': 16896512, 'steps': 33000, 'loss/train': 1.1297675371170044} 02/25/2022 20:19:22 - INFO - codeparrot_training - Step 33001: {'lr': 0.00013941315263617157, 'samples': 16897024, 'steps': 33001, 'loss/train': 2.33842134475708} 02/25/2022 20:19:25 - INFO - codeparrot_training - Step 33002: {'lr': 0.00013939847830081226, 'samples': 16897536, 'steps': 33002, 'loss/train': 1.0919899940490723} 02/25/2022 20:19:32 - INFO - codeparrot_training - Step 33003: {'lr': 0.00013938380443923487, 'samples': 16898048, 'steps': 33003, 'loss/train': 1.9448785781860352} 02/25/2022 20:19:36 - INFO - codeparrot_training - Step 33004: {'lr': 0.00013936913105150212, 'samples': 16898560, 'steps': 33004, 'loss/train': 1.6131045818328857} 02/25/2022 20:19:41 - INFO - codeparrot_training - Step 33005: {'lr': 0.00013935445813767678, 'samples': 16899072, 'steps': 33005, 'loss/train': 2.596254587173462} 02/25/2022 20:19:45 - INFO - codeparrot_training - Step 33006: {'lr': 0.000139339785697822, 'samples': 16899584, 'steps': 33006, 'loss/train': 1.629105806350708} 02/25/2022 20:19:50 - INFO - codeparrot_training - Step 33007: {'lr': 0.00013932511373200023, 'samples': 16900096, 'steps': 33007, 'loss/train': 1.2458577156066895} 02/25/2022 20:19:53 - INFO - codeparrot_training - Step 33008: {'lr': 0.00013931044224027467, 'samples': 16900608, 'steps': 33008, 'loss/train': 1.3389471769332886} 02/25/2022 20:19:59 - INFO - codeparrot_training - Step 33009: {'lr': 0.0001392957712227079, 'samples': 16901120, 'steps': 33009, 'loss/train': 1.5443185567855835} 02/25/2022 20:20:02 - INFO - codeparrot_training - Step 33010: {'lr': 0.00013928110067936312, 'samples': 16901632, 'steps': 33010, 'loss/train': 1.9592325687408447} 02/25/2022 20:20:08 - INFO - codeparrot_training - Step 33011: {'lr': 0.0001392664306103027, 'samples': 16902144, 'steps': 33011, 'loss/train': 1.7590411901474} 02/25/2022 20:20:12 - INFO - codeparrot_training - Step 33012: {'lr': 0.00013925176101558986, 'samples': 16902656, 'steps': 33012, 'loss/train': 1.8600107431411743} 02/25/2022 20:20:17 - INFO - codeparrot_training - Step 33013: {'lr': 0.0001392370918952872, 'samples': 16903168, 'steps': 33013, 'loss/train': 2.094806432723999} 02/25/2022 20:20:21 - INFO - codeparrot_training - Step 33014: {'lr': 0.00013922242324945788, 'samples': 16903680, 'steps': 33014, 'loss/train': 2.209496021270752} 02/25/2022 20:20:26 - INFO - codeparrot_training - Step 33015: {'lr': 0.0001392077550781643, 'samples': 16904192, 'steps': 33015, 'loss/train': 1.3058743476867676} 02/25/2022 20:20:30 - INFO - codeparrot_training - Step 33016: {'lr': 0.00013919308738146962, 'samples': 16904704, 'steps': 33016, 'loss/train': 0.8081715106964111} 02/25/2022 20:20:35 - INFO - codeparrot_training - Step 33017: {'lr': 0.0001391784201594365, 'samples': 16905216, 'steps': 33017, 'loss/train': 1.964890956878662} 02/25/2022 20:20:39 - INFO - codeparrot_training - Step 33018: {'lr': 0.00013916375341212794, 'samples': 16905728, 'steps': 33018, 'loss/train': 1.5021774768829346} 02/25/2022 20:20:45 - INFO - codeparrot_training - Step 33019: {'lr': 0.00013914908713960665, 'samples': 16906240, 'steps': 33019, 'loss/train': 1.1030876636505127} 02/25/2022 20:20:48 - INFO - codeparrot_training - Step 33020: {'lr': 0.00013913442134193545, 'samples': 16906752, 'steps': 33020, 'loss/train': 2.2400050163269043} 02/25/2022 20:20:54 - INFO - codeparrot_training - Step 33021: {'lr': 0.00013911975601917708, 'samples': 16907264, 'steps': 33021, 'loss/train': 1.0865991115570068} 02/25/2022 20:20:57 - INFO - codeparrot_training - Step 33022: {'lr': 0.00013910509117139462, 'samples': 16907776, 'steps': 33022, 'loss/train': 2.0021352767944336} 02/25/2022 20:21:03 - INFO - codeparrot_training - Step 33023: {'lr': 0.0001390904267986507, 'samples': 16908288, 'steps': 33023, 'loss/train': 2.7526328563690186} 02/25/2022 20:21:07 - INFO - codeparrot_training - Step 33024: {'lr': 0.0001390757629010082, 'samples': 16908800, 'steps': 33024, 'loss/train': 8.040966033935547} 02/25/2022 20:21:12 - INFO - codeparrot_training - Step 33025: {'lr': 0.00013906109947852977, 'samples': 16909312, 'steps': 33025, 'loss/train': 1.3483787775039673} 02/25/2022 20:21:16 - INFO - codeparrot_training - Step 33026: {'lr': 0.00013904643653127847, 'samples': 16909824, 'steps': 33026, 'loss/train': 2.066166400909424} 02/25/2022 20:21:21 - INFO - codeparrot_training - Step 33027: {'lr': 0.00013903177405931701, 'samples': 16910336, 'steps': 33027, 'loss/train': 1.4834327697753906} 02/25/2022 20:21:24 - INFO - codeparrot_training - Step 33028: {'lr': 0.00013901711206270822, 'samples': 16910848, 'steps': 33028, 'loss/train': 1.009364128112793} 02/25/2022 20:21:31 - INFO - codeparrot_training - Step 33029: {'lr': 0.00013900245054151476, 'samples': 16911360, 'steps': 33029, 'loss/train': 1.5293819904327393} 02/25/2022 20:21:34 - INFO - codeparrot_training - Step 33030: {'lr': 0.00013898778949579966, 'samples': 16911872, 'steps': 33030, 'loss/train': 1.4307610988616943} 02/25/2022 20:21:40 - INFO - codeparrot_training - Step 33031: {'lr': 0.0001389731289256255, 'samples': 16912384, 'steps': 33031, 'loss/train': 1.7180373668670654} 02/25/2022 20:21:43 - INFO - codeparrot_training - Step 33032: {'lr': 0.0001389584688310554, 'samples': 16912896, 'steps': 33032, 'loss/train': 2.271662950515747} 02/25/2022 20:21:49 - INFO - codeparrot_training - Step 33033: {'lr': 0.00013894380921215178, 'samples': 16913408, 'steps': 33033, 'loss/train': 1.338437795639038} 02/25/2022 20:21:52 - INFO - codeparrot_training - Step 33034: {'lr': 0.0001389291500689777, 'samples': 16913920, 'steps': 33034, 'loss/train': 0.5751528739929199} 02/25/2022 20:21:58 - INFO - codeparrot_training - Step 33035: {'lr': 0.0001389144914015958, 'samples': 16914432, 'steps': 33035, 'loss/train': 1.5223861932754517} 02/25/2022 20:22:01 - INFO - codeparrot_training - Step 33036: {'lr': 0.00013889983321006916, 'samples': 16914944, 'steps': 33036, 'loss/train': 1.6493160724639893} 02/25/2022 20:22:07 - INFO - codeparrot_training - Step 33037: {'lr': 0.0001388851754944601, 'samples': 16915456, 'steps': 33037, 'loss/train': 1.842376470565796} 02/25/2022 20:22:10 - INFO - codeparrot_training - Step 33038: {'lr': 0.0001388705182548318, 'samples': 16915968, 'steps': 33038, 'loss/train': 1.6146433353424072} 02/25/2022 20:22:16 - INFO - codeparrot_training - Step 33039: {'lr': 0.0001388558614912469, 'samples': 16916480, 'steps': 33039, 'loss/train': 0.2921536862850189} 02/25/2022 20:22:20 - INFO - codeparrot_training - Step 33040: {'lr': 0.00013884120520376818, 'samples': 16916992, 'steps': 33040, 'loss/train': 2.442641258239746} 02/25/2022 20:22:25 - INFO - codeparrot_training - Step 33041: {'lr': 0.00013882654939245835, 'samples': 16917504, 'steps': 33041, 'loss/train': 1.280112862586975} 02/25/2022 20:22:29 - INFO - codeparrot_training - Step 33042: {'lr': 0.00013881189405738043, 'samples': 16918016, 'steps': 33042, 'loss/train': 3.812242269515991} 02/25/2022 20:22:34 - INFO - codeparrot_training - Step 33043: {'lr': 0.00013879723919859698, 'samples': 16918528, 'steps': 33043, 'loss/train': 2.054389715194702} 02/25/2022 20:22:38 - INFO - codeparrot_training - Step 33044: {'lr': 0.00013878258481617078, 'samples': 16919040, 'steps': 33044, 'loss/train': 2.501282215118408} 02/25/2022 20:22:43 - INFO - codeparrot_training - Step 33045: {'lr': 0.0001387679309101648, 'samples': 16919552, 'steps': 33045, 'loss/train': 1.7748994827270508} 02/25/2022 20:22:47 - INFO - codeparrot_training - Step 33046: {'lr': 0.00013875327748064166, 'samples': 16920064, 'steps': 33046, 'loss/train': 3.001192569732666} 02/25/2022 20:22:52 - INFO - codeparrot_training - Step 33047: {'lr': 0.0001387386245276641, 'samples': 16920576, 'steps': 33047, 'loss/train': 2.1625325679779053} 02/25/2022 20:22:56 - INFO - codeparrot_training - Step 33048: {'lr': 0.00013872397205129495, 'samples': 16921088, 'steps': 33048, 'loss/train': 1.4520047903060913} 02/25/2022 20:23:02 - INFO - codeparrot_training - Step 33049: {'lr': 0.00013870932005159698, 'samples': 16921600, 'steps': 33049, 'loss/train': 2.028303623199463} 02/25/2022 20:23:06 - INFO - codeparrot_training - Step 33050: {'lr': 0.000138694668528633, 'samples': 16922112, 'steps': 33050, 'loss/train': 1.6190156936645508} 02/25/2022 20:23:11 - INFO - codeparrot_training - Step 33051: {'lr': 0.00013868001748246566, 'samples': 16922624, 'steps': 33051, 'loss/train': 1.5195121765136719} 02/25/2022 20:23:15 - INFO - codeparrot_training - Step 33052: {'lr': 0.00013866536691315774, 'samples': 16923136, 'steps': 33052, 'loss/train': 1.4691048860549927} 02/25/2022 20:23:20 - INFO - codeparrot_training - Step 33053: {'lr': 0.00013865071682077207, 'samples': 16923648, 'steps': 33053, 'loss/train': 1.4723105430603027} 02/25/2022 20:23:24 - INFO - codeparrot_training - Step 33054: {'lr': 0.00013863606720537143, 'samples': 16924160, 'steps': 33054, 'loss/train': 0.8526555299758911} 02/25/2022 20:23:29 - INFO - codeparrot_training - Step 33055: {'lr': 0.00013862141806701846, 'samples': 16924672, 'steps': 33055, 'loss/train': 2.0327413082122803} 02/25/2022 20:23:33 - INFO - codeparrot_training - Step 33056: {'lr': 0.00013860676940577593, 'samples': 16925184, 'steps': 33056, 'loss/train': 2.1846323013305664} 02/25/2022 20:23:38 - INFO - codeparrot_training - Step 33057: {'lr': 0.00013859212122170668, 'samples': 16925696, 'steps': 33057, 'loss/train': 1.57632315158844} 02/25/2022 20:23:42 - INFO - codeparrot_training - Step 33058: {'lr': 0.00013857747351487344, 'samples': 16926208, 'steps': 33058, 'loss/train': 3.3128674030303955} 02/25/2022 20:23:47 - INFO - codeparrot_training - Step 33059: {'lr': 0.0001385628262853389, 'samples': 16926720, 'steps': 33059, 'loss/train': 2.1297178268432617} 02/25/2022 20:23:51 - INFO - codeparrot_training - Step 33060: {'lr': 0.00013854817953316573, 'samples': 16927232, 'steps': 33060, 'loss/train': 1.578182339668274} 02/25/2022 20:23:56 - INFO - codeparrot_training - Step 33061: {'lr': 0.0001385335332584169, 'samples': 16927744, 'steps': 33061, 'loss/train': 1.3774527311325073} 02/25/2022 20:24:00 - INFO - codeparrot_training - Step 33062: {'lr': 0.00013851888746115498, 'samples': 16928256, 'steps': 33062, 'loss/train': 1.1669151782989502} 02/25/2022 20:24:05 - INFO - codeparrot_training - Step 33063: {'lr': 0.00013850424214144276, 'samples': 16928768, 'steps': 33063, 'loss/train': 1.1518981456756592} 02/25/2022 20:24:09 - INFO - codeparrot_training - Step 33064: {'lr': 0.00013848959729934285, 'samples': 16929280, 'steps': 33064, 'loss/train': 1.6680086851119995} 02/25/2022 20:24:15 - INFO - codeparrot_training - Step 33065: {'lr': 0.00013847495293491818, 'samples': 16929792, 'steps': 33065, 'loss/train': 0.9807769060134888} 02/25/2022 20:24:19 - INFO - codeparrot_training - Step 33066: {'lr': 0.00013846030904823137, 'samples': 16930304, 'steps': 33066, 'loss/train': 0.9813781380653381} 02/25/2022 20:24:25 - INFO - codeparrot_training - Step 33067: {'lr': 0.0001384456656393453, 'samples': 16930816, 'steps': 33067, 'loss/train': 2.031611919403076} 02/25/2022 20:24:28 - INFO - codeparrot_training - Step 33068: {'lr': 0.00013843102270832242, 'samples': 16931328, 'steps': 33068, 'loss/train': 0.8346250653266907} 02/25/2022 20:24:34 - INFO - codeparrot_training - Step 33069: {'lr': 0.0001384163802552257, 'samples': 16931840, 'steps': 33069, 'loss/train': 1.6836495399475098} 02/25/2022 20:24:37 - INFO - codeparrot_training - Step 33070: {'lr': 0.0001384017382801177, 'samples': 16932352, 'steps': 33070, 'loss/train': 2.083275556564331} 02/25/2022 20:24:43 - INFO - codeparrot_training - Step 33071: {'lr': 0.00013838709678306137, 'samples': 16932864, 'steps': 33071, 'loss/train': 0.08037525415420532} 02/25/2022 20:24:46 - INFO - codeparrot_training - Step 33072: {'lr': 0.00013837245576411912, 'samples': 16933376, 'steps': 33072, 'loss/train': 1.5169219970703125} 02/25/2022 20:24:52 - INFO - codeparrot_training - Step 33073: {'lr': 0.0001383578152233539, 'samples': 16933888, 'steps': 33073, 'loss/train': 2.193718671798706} 02/25/2022 20:24:55 - INFO - codeparrot_training - Step 33074: {'lr': 0.00013834317516082825, 'samples': 16934400, 'steps': 33074, 'loss/train': 1.8317298889160156} 02/25/2022 20:25:02 - INFO - codeparrot_training - Step 33075: {'lr': 0.00013832853557660518, 'samples': 16934912, 'steps': 33075, 'loss/train': 2.081847667694092} 02/25/2022 20:25:05 - INFO - codeparrot_training - Step 33076: {'lr': 0.00013831389647074702, 'samples': 16935424, 'steps': 33076, 'loss/train': 1.608643651008606} 02/25/2022 20:25:11 - INFO - codeparrot_training - Step 33077: {'lr': 0.00013829925784331675, 'samples': 16935936, 'steps': 33077, 'loss/train': 2.0918939113616943} 02/25/2022 20:25:14 - INFO - codeparrot_training - Step 33078: {'lr': 0.0001382846196943769, 'samples': 16936448, 'steps': 33078, 'loss/train': 1.0892910957336426} 02/25/2022 20:25:20 - INFO - codeparrot_training - Step 33079: {'lr': 0.00013826998202399037, 'samples': 16936960, 'steps': 33079, 'loss/train': 0.45494207739830017} 02/25/2022 20:25:23 - INFO - codeparrot_training - Step 33080: {'lr': 0.00013825534483221974, 'samples': 16937472, 'steps': 33080, 'loss/train': 1.1691253185272217} 02/25/2022 20:25:29 - INFO - codeparrot_training - Step 33081: {'lr': 0.00013824070811912774, 'samples': 16937984, 'steps': 33081, 'loss/train': 1.1718149185180664} 02/25/2022 20:25:32 - INFO - codeparrot_training - Step 33082: {'lr': 0.00013822607188477697, 'samples': 16938496, 'steps': 33082, 'loss/train': 1.8045306205749512} 02/25/2022 20:25:38 - INFO - codeparrot_training - Step 33083: {'lr': 0.00013821143612923026, 'samples': 16939008, 'steps': 33083, 'loss/train': 2.6224429607391357} 02/25/2022 20:25:41 - INFO - codeparrot_training - Step 33084: {'lr': 0.0001381968008525503, 'samples': 16939520, 'steps': 33084, 'loss/train': 1.9999854564666748} 02/25/2022 20:25:48 - INFO - codeparrot_training - Step 33085: {'lr': 0.00013818216605479978, 'samples': 16940032, 'steps': 33085, 'loss/train': 2.1719233989715576} 02/25/2022 20:25:51 - INFO - codeparrot_training - Step 33086: {'lr': 0.0001381675317360413, 'samples': 16940544, 'steps': 33086, 'loss/train': 7.925590991973877} 02/25/2022 20:25:57 - INFO - codeparrot_training - Step 33087: {'lr': 0.00013815289789633752, 'samples': 16941056, 'steps': 33087, 'loss/train': 1.7833501100540161} 02/25/2022 20:26:00 - INFO - codeparrot_training - Step 33088: {'lr': 0.00013813826453575135, 'samples': 16941568, 'steps': 33088, 'loss/train': 2.3182373046875} 02/25/2022 20:26:06 - INFO - codeparrot_training - Step 33089: {'lr': 0.00013812363165434526, 'samples': 16942080, 'steps': 33089, 'loss/train': 2.175645589828491} 02/25/2022 20:26:09 - INFO - codeparrot_training - Step 33090: {'lr': 0.00013810899925218207, 'samples': 16942592, 'steps': 33090, 'loss/train': 2.6579976081848145} 02/25/2022 20:26:15 - INFO - codeparrot_training - Step 33091: {'lr': 0.00013809436732932428, 'samples': 16943104, 'steps': 33091, 'loss/train': 1.1017791032791138} 02/25/2022 20:26:18 - INFO - codeparrot_training - Step 33092: {'lr': 0.0001380797358858348, 'samples': 16943616, 'steps': 33092, 'loss/train': 1.3910006284713745} 02/25/2022 20:26:24 - INFO - codeparrot_training - Step 33093: {'lr': 0.00013806510492177614, 'samples': 16944128, 'steps': 33093, 'loss/train': 2.0862746238708496} 02/25/2022 20:26:27 - INFO - codeparrot_training - Step 33094: {'lr': 0.00013805047443721108, 'samples': 16944640, 'steps': 33094, 'loss/train': 1.7025426626205444} 02/25/2022 20:26:33 - INFO - codeparrot_training - Step 33095: {'lr': 0.0001380358444322021, 'samples': 16945152, 'steps': 33095, 'loss/train': 2.4947752952575684} 02/25/2022 20:26:37 - INFO - codeparrot_training - Step 33096: {'lr': 0.00013802121490681213, 'samples': 16945664, 'steps': 33096, 'loss/train': 2.175145149230957} 02/25/2022 20:26:42 - INFO - codeparrot_training - Step 33097: {'lr': 0.0001380065858611037, 'samples': 16946176, 'steps': 33097, 'loss/train': 2.319929361343384} 02/25/2022 20:26:46 - INFO - codeparrot_training - Step 33098: {'lr': 0.00013799195729513952, 'samples': 16946688, 'steps': 33098, 'loss/train': 0.7369640469551086} 02/25/2022 20:26:52 - INFO - codeparrot_training - Step 33099: {'lr': 0.00013797732920898216, 'samples': 16947200, 'steps': 33099, 'loss/train': 2.231600761413574} 02/25/2022 20:26:56 - INFO - codeparrot_training - Step 33100: {'lr': 0.0001379627016026944, 'samples': 16947712, 'steps': 33100, 'loss/train': 2.0435006618499756} 02/25/2022 20:26:59 - INFO - codeparrot_training - Step 33101: {'lr': 0.00013794807447633874, 'samples': 16948224, 'steps': 33101, 'loss/train': 0.7494664192199707} 02/25/2022 20:27:05 - INFO - codeparrot_training - Step 33102: {'lr': 0.00013793344782997817, 'samples': 16948736, 'steps': 33102, 'loss/train': 0.8188365697860718} 02/25/2022 20:27:08 - INFO - codeparrot_training - Step 33103: {'lr': 0.00013791882166367493, 'samples': 16949248, 'steps': 33103, 'loss/train': 0.4193016588687897} 02/25/2022 20:27:14 - INFO - codeparrot_training - Step 33104: {'lr': 0.000137904195977492, 'samples': 16949760, 'steps': 33104, 'loss/train': 2.408766269683838} 02/25/2022 20:27:17 - INFO - codeparrot_training - Step 33105: {'lr': 0.00013788957077149174, 'samples': 16950272, 'steps': 33105, 'loss/train': 0.5733808875083923} 02/25/2022 20:27:23 - INFO - codeparrot_training - Step 33106: {'lr': 0.00013787494604573718, 'samples': 16950784, 'steps': 33106, 'loss/train': 1.0324296951293945} 02/25/2022 20:27:26 - INFO - codeparrot_training - Step 33107: {'lr': 0.00013786032180029053, 'samples': 16951296, 'steps': 33107, 'loss/train': 2.0375800132751465} 02/25/2022 20:27:32 - INFO - codeparrot_training - Step 33108: {'lr': 0.0001378456980352148, 'samples': 16951808, 'steps': 33108, 'loss/train': 0.04046846181154251} 02/25/2022 20:27:37 - INFO - codeparrot_training - Step 33109: {'lr': 0.00013783107475057234, 'samples': 16952320, 'steps': 33109, 'loss/train': 1.7546592950820923} 02/25/2022 20:27:41 - INFO - codeparrot_training - Step 33110: {'lr': 0.00013781645194642616, 'samples': 16952832, 'steps': 33110, 'loss/train': 2.344900608062744} 02/25/2022 20:27:47 - INFO - codeparrot_training - Step 33111: {'lr': 0.0001378018296228385, 'samples': 16953344, 'steps': 33111, 'loss/train': 1.6780171394348145} 02/25/2022 20:27:50 - INFO - codeparrot_training - Step 33112: {'lr': 0.00013778720777987225, 'samples': 16953856, 'steps': 33112, 'loss/train': 1.6987971067428589} 02/25/2022 20:27:54 - INFO - codeparrot_training - Step 33113: {'lr': 0.00013777258641758982, 'samples': 16954368, 'steps': 33113, 'loss/train': 1.5293858051300049} 02/25/2022 20:28:00 - INFO - codeparrot_training - Step 33114: {'lr': 0.00013775796553605417, 'samples': 16954880, 'steps': 33114, 'loss/train': 1.8258339166641235} 02/25/2022 20:28:04 - INFO - codeparrot_training - Step 33115: {'lr': 0.00013774334513532772, 'samples': 16955392, 'steps': 33115, 'loss/train': 2.187664270401001} 02/25/2022 20:28:09 - INFO - codeparrot_training - Step 33116: {'lr': 0.00013772872521547314, 'samples': 16955904, 'steps': 33116, 'loss/train': 1.9151933193206787} 02/25/2022 20:28:12 - INFO - codeparrot_training - Step 33117: {'lr': 0.00013771410577655298, 'samples': 16956416, 'steps': 33117, 'loss/train': 0.0994546115398407} 02/25/2022 20:28:18 - INFO - codeparrot_training - Step 33118: {'lr': 0.00013769948681862997, 'samples': 16956928, 'steps': 33118, 'loss/train': 1.7644505500793457} 02/25/2022 20:28:22 - INFO - codeparrot_training - Step 33119: {'lr': 0.00013768486834176677, 'samples': 16957440, 'steps': 33119, 'loss/train': 0.7357638478279114} 02/25/2022 20:28:27 - INFO - codeparrot_training - Step 33120: {'lr': 0.0001376702503460259, 'samples': 16957952, 'steps': 33120, 'loss/train': 1.7830103635787964} 02/25/2022 20:28:30 - INFO - codeparrot_training - Step 33121: {'lr': 0.00013765563283146993, 'samples': 16958464, 'steps': 33121, 'loss/train': 0.5785823464393616} 02/25/2022 20:28:36 - INFO - codeparrot_training - Step 33122: {'lr': 0.0001376410157981616, 'samples': 16958976, 'steps': 33122, 'loss/train': 1.0229709148406982} 02/25/2022 20:28:39 - INFO - codeparrot_training - Step 33123: {'lr': 0.00013762639924616351, 'samples': 16959488, 'steps': 33123, 'loss/train': 0.8258456587791443} 02/25/2022 20:28:46 - INFO - codeparrot_training - Step 33124: {'lr': 0.00013761178317553825, 'samples': 16960000, 'steps': 33124, 'loss/train': 2.104133129119873} 02/25/2022 20:28:49 - INFO - codeparrot_training - Step 33125: {'lr': 0.00013759716758634833, 'samples': 16960512, 'steps': 33125, 'loss/train': 1.05113685131073} 02/25/2022 20:28:55 - INFO - codeparrot_training - Step 33126: {'lr': 0.00013758255247865652, 'samples': 16961024, 'steps': 33126, 'loss/train': 1.6678348779678345} 02/25/2022 20:28:58 - INFO - codeparrot_training - Step 33127: {'lr': 0.0001375679378525254, 'samples': 16961536, 'steps': 33127, 'loss/train': 1.6824684143066406} 02/25/2022 20:29:04 - INFO - codeparrot_training - Step 33128: {'lr': 0.0001375533237080175, 'samples': 16962048, 'steps': 33128, 'loss/train': 2.2287282943725586} 02/25/2022 20:29:07 - INFO - codeparrot_training - Step 33129: {'lr': 0.0001375387100451954, 'samples': 16962560, 'steps': 33129, 'loss/train': 2.9126675128936768} 02/25/2022 20:29:13 - INFO - codeparrot_training - Step 33130: {'lr': 0.00013752409686412182, 'samples': 16963072, 'steps': 33130, 'loss/train': 1.0279017686843872} 02/25/2022 20:29:16 - INFO - codeparrot_training - Step 33131: {'lr': 0.0001375094841648593, 'samples': 16963584, 'steps': 33131, 'loss/train': 1.4462863206863403} 02/25/2022 20:29:22 - INFO - codeparrot_training - Step 33132: {'lr': 0.0001374948719474704, 'samples': 16964096, 'steps': 33132, 'loss/train': 0.4972841441631317} 02/25/2022 20:29:25 - INFO - codeparrot_training - Step 33133: {'lr': 0.0001374802602120177, 'samples': 16964608, 'steps': 33133, 'loss/train': 1.721798062324524} 02/25/2022 20:29:32 - INFO - codeparrot_training - Step 33134: {'lr': 0.0001374656489585639, 'samples': 16965120, 'steps': 33134, 'loss/train': 0.14702363312244415} 02/25/2022 20:29:35 - INFO - codeparrot_training - Step 33135: {'lr': 0.00013745103818717152, 'samples': 16965632, 'steps': 33135, 'loss/train': 1.9318891763687134} 02/25/2022 20:29:41 - INFO - codeparrot_training - Step 33136: {'lr': 0.00013743642789790318, 'samples': 16966144, 'steps': 33136, 'loss/train': 0.4734211266040802} 02/25/2022 20:29:44 - INFO - codeparrot_training - Step 33137: {'lr': 0.00013742181809082144, 'samples': 16966656, 'steps': 33137, 'loss/train': 1.9879918098449707} 02/25/2022 20:29:48 - INFO - codeparrot_training - Step 33138: {'lr': 0.00013740720876598877, 'samples': 16967168, 'steps': 33138, 'loss/train': 2.406527280807495} 02/25/2022 20:29:53 - INFO - codeparrot_training - Step 33139: {'lr': 0.000137392599923468, 'samples': 16967680, 'steps': 33139, 'loss/train': 2.1891274452209473} 02/25/2022 20:29:59 - INFO - codeparrot_training - Step 33140: {'lr': 0.00013737799156332144, 'samples': 16968192, 'steps': 33140, 'loss/train': 1.9293526411056519} 02/25/2022 20:30:02 - INFO - codeparrot_training - Step 33141: {'lr': 0.00013736338368561197, 'samples': 16968704, 'steps': 33141, 'loss/train': 2.776402473449707} 02/25/2022 20:30:08 - INFO - codeparrot_training - Step 33142: {'lr': 0.00013734877629040183, 'samples': 16969216, 'steps': 33142, 'loss/train': 1.84799063205719} 02/25/2022 20:30:12 - INFO - codeparrot_training - Step 33143: {'lr': 0.0001373341693777539, 'samples': 16969728, 'steps': 33143, 'loss/train': 1.8786133527755737} 02/25/2022 20:30:17 - INFO - codeparrot_training - Step 33144: {'lr': 0.00013731956294773046, 'samples': 16970240, 'steps': 33144, 'loss/train': 1.5956374406814575} 02/25/2022 20:30:21 - INFO - codeparrot_training - Step 33145: {'lr': 0.00013730495700039444, 'samples': 16970752, 'steps': 33145, 'loss/train': 0.39798957109451294} 02/25/2022 20:30:26 - INFO - codeparrot_training - Step 33146: {'lr': 0.000137290351535808, 'samples': 16971264, 'steps': 33146, 'loss/train': 2.0345547199249268} 02/25/2022 20:30:30 - INFO - codeparrot_training - Step 33147: {'lr': 0.00013727574655403402, 'samples': 16971776, 'steps': 33147, 'loss/train': 2.4439237117767334} 02/25/2022 20:30:35 - INFO - codeparrot_training - Step 33148: {'lr': 0.00013726114205513484, 'samples': 16972288, 'steps': 33148, 'loss/train': 1.4308420419692993} 02/25/2022 20:30:39 - INFO - codeparrot_training - Step 33149: {'lr': 0.00013724653803917338, 'samples': 16972800, 'steps': 33149, 'loss/train': 0.8693727254867554} 02/25/2022 20:30:44 - INFO - codeparrot_training - Step 33150: {'lr': 0.00013723193450621167, 'samples': 16973312, 'steps': 33150, 'loss/train': 1.1945158243179321} 02/25/2022 20:30:48 - INFO - codeparrot_training - Step 33151: {'lr': 0.00013721733145631264, 'samples': 16973824, 'steps': 33151, 'loss/train': 1.2456892728805542} 02/25/2022 20:30:53 - INFO - codeparrot_training - Step 33152: {'lr': 0.0001372027288895387, 'samples': 16974336, 'steps': 33152, 'loss/train': 8.444211959838867} 02/25/2022 20:30:57 - INFO - codeparrot_training - Step 33153: {'lr': 0.00013718812680595254, 'samples': 16974848, 'steps': 33153, 'loss/train': 1.0180572271347046} 02/25/2022 20:31:02 - INFO - codeparrot_training - Step 33154: {'lr': 0.00013717352520561664, 'samples': 16975360, 'steps': 33154, 'loss/train': 0.6758490204811096} 02/25/2022 20:31:06 - INFO - codeparrot_training - Step 33155: {'lr': 0.00013715892408859348, 'samples': 16975872, 'steps': 33155, 'loss/train': 2.796698808670044} 02/25/2022 20:31:11 - INFO - codeparrot_training - Step 33156: {'lr': 0.0001371443234549456, 'samples': 16976384, 'steps': 33156, 'loss/train': 4.349503040313721} 02/25/2022 20:31:15 - INFO - codeparrot_training - Step 33157: {'lr': 0.0001371297233047357, 'samples': 16976896, 'steps': 33157, 'loss/train': 1.1091523170471191} 02/25/2022 20:31:21 - INFO - codeparrot_training - Step 33158: {'lr': 0.0001371151236380262, 'samples': 16977408, 'steps': 33158, 'loss/train': 2.1146535873413086} 02/25/2022 20:31:24 - INFO - codeparrot_training - Step 33159: {'lr': 0.0001371005244548797, 'samples': 16977920, 'steps': 33159, 'loss/train': 0.3008519113063812} 02/25/2022 20:31:30 - INFO - codeparrot_training - Step 33160: {'lr': 0.00013708592575535858, 'samples': 16978432, 'steps': 33160, 'loss/train': 2.463362455368042} 02/25/2022 20:31:33 - INFO - codeparrot_training - Step 33161: {'lr': 0.00013707132753952562, 'samples': 16978944, 'steps': 33161, 'loss/train': 1.5532010793685913} 02/25/2022 20:31:39 - INFO - codeparrot_training - Step 33162: {'lr': 0.0001370567298074431, 'samples': 16979456, 'steps': 33162, 'loss/train': 2.880884885787964} 02/25/2022 20:31:42 - INFO - codeparrot_training - Step 33163: {'lr': 0.0001370421325591739, 'samples': 16979968, 'steps': 33163, 'loss/train': 2.054656744003296} 02/25/2022 20:31:48 - INFO - codeparrot_training - Step 33164: {'lr': 0.00013702753579478017, 'samples': 16980480, 'steps': 33164, 'loss/train': 1.75658118724823} 02/25/2022 20:31:51 - INFO - codeparrot_training - Step 33165: {'lr': 0.00013701293951432465, 'samples': 16980992, 'steps': 33165, 'loss/train': 2.1616644859313965} 02/25/2022 20:31:57 - INFO - codeparrot_training - Step 33166: {'lr': 0.00013699834371786975, 'samples': 16981504, 'steps': 33166, 'loss/train': 3.6250452995300293} 02/25/2022 20:32:00 - INFO - codeparrot_training - Step 33167: {'lr': 0.00013698374840547827, 'samples': 16982016, 'steps': 33167, 'loss/train': 0.735996425151825} 02/25/2022 20:32:06 - INFO - codeparrot_training - Step 33168: {'lr': 0.0001369691535772123, 'samples': 16982528, 'steps': 33168, 'loss/train': 1.6416651010513306} 02/25/2022 20:32:09 - INFO - codeparrot_training - Step 33169: {'lr': 0.0001369545592331347, 'samples': 16983040, 'steps': 33169, 'loss/train': 1.380927562713623} 02/25/2022 20:32:15 - INFO - codeparrot_training - Step 33170: {'lr': 0.0001369399653733079, 'samples': 16983552, 'steps': 33170, 'loss/train': 0.5697680711746216} 02/25/2022 20:32:19 - INFO - codeparrot_training - Step 33171: {'lr': 0.00013692537199779434, 'samples': 16984064, 'steps': 33171, 'loss/train': 1.2204687595367432} 02/25/2022 20:32:24 - INFO - codeparrot_training - Step 33172: {'lr': 0.0001369107791066565, 'samples': 16984576, 'steps': 33172, 'loss/train': 1.3063995838165283} 02/25/2022 20:32:28 - INFO - codeparrot_training - Step 33173: {'lr': 0.00013689618669995707, 'samples': 16985088, 'steps': 33173, 'loss/train': 1.1784683465957642} 02/25/2022 20:32:33 - INFO - codeparrot_training - Step 33174: {'lr': 0.00013688159477775848, 'samples': 16985600, 'steps': 33174, 'loss/train': 1.8335788249969482} 02/25/2022 20:32:37 - INFO - codeparrot_training - Step 33175: {'lr': 0.0001368670033401231, 'samples': 16986112, 'steps': 33175, 'loss/train': 2.501119375228882} 02/25/2022 20:32:42 - INFO - codeparrot_training - Step 33176: {'lr': 0.00013685241238711366, 'samples': 16986624, 'steps': 33176, 'loss/train': 1.634509801864624} 02/25/2022 20:32:46 - INFO - codeparrot_training - Step 33177: {'lr': 0.00013683782191879253, 'samples': 16987136, 'steps': 33177, 'loss/train': 2.0425705909729004} 02/25/2022 20:32:51 - INFO - codeparrot_training - Step 33178: {'lr': 0.00013682323193522225, 'samples': 16987648, 'steps': 33178, 'loss/train': 1.3294026851654053} 02/25/2022 20:32:55 - INFO - codeparrot_training - Step 33179: {'lr': 0.00013680864243646523, 'samples': 16988160, 'steps': 33179, 'loss/train': 0.7168804407119751} 02/25/2022 20:33:01 - INFO - codeparrot_training - Step 33180: {'lr': 0.00013679405342258412, 'samples': 16988672, 'steps': 33180, 'loss/train': 1.2534846067428589} 02/25/2022 20:33:05 - INFO - codeparrot_training - Step 33181: {'lr': 0.00013677946489364134, 'samples': 16989184, 'steps': 33181, 'loss/train': 2.8118321895599365} 02/25/2022 20:33:10 - INFO - codeparrot_training - Step 33182: {'lr': 0.0001367648768496994, 'samples': 16989696, 'steps': 33182, 'loss/train': 0.6932120323181152} 02/25/2022 20:33:14 - INFO - codeparrot_training - Step 33183: {'lr': 0.00013675028929082067, 'samples': 16990208, 'steps': 33183, 'loss/train': 1.7817246913909912} 02/25/2022 20:33:19 - INFO - codeparrot_training - Step 33184: {'lr': 0.00013673570221706794, 'samples': 16990720, 'steps': 33184, 'loss/train': 1.319392204284668} 02/25/2022 20:33:23 - INFO - codeparrot_training - Step 33185: {'lr': 0.0001367211156285033, 'samples': 16991232, 'steps': 33185, 'loss/train': 1.7152141332626343} 02/25/2022 20:33:28 - INFO - codeparrot_training - Step 33186: {'lr': 0.00013670652952518949, 'samples': 16991744, 'steps': 33186, 'loss/train': 1.6319048404693604} 02/25/2022 20:33:32 - INFO - codeparrot_training - Step 33187: {'lr': 0.00013669194390718886, 'samples': 16992256, 'steps': 33187, 'loss/train': 1.635172963142395} 02/25/2022 20:33:37 - INFO - codeparrot_training - Step 33188: {'lr': 0.00013667735877456405, 'samples': 16992768, 'steps': 33188, 'loss/train': 1.7197566032409668} 02/25/2022 20:33:41 - INFO - codeparrot_training - Step 33189: {'lr': 0.00013666277412737748, 'samples': 16993280, 'steps': 33189, 'loss/train': 1.944015383720398} 02/25/2022 20:33:47 - INFO - codeparrot_training - Step 33190: {'lr': 0.0001366481899656916, 'samples': 16993792, 'steps': 33190, 'loss/train': 1.2589294910430908} 02/25/2022 20:33:51 - INFO - codeparrot_training - Step 33191: {'lr': 0.00013663360628956875, 'samples': 16994304, 'steps': 33191, 'loss/train': 1.6134530305862427} 02/25/2022 20:33:56 - INFO - codeparrot_training - Step 33192: {'lr': 0.00013661902309907166, 'samples': 16994816, 'steps': 33192, 'loss/train': 2.4372260570526123} 02/25/2022 20:34:00 - INFO - codeparrot_training - Step 33193: {'lr': 0.00013660444039426264, 'samples': 16995328, 'steps': 33193, 'loss/train': 2.484558582305908} 02/25/2022 20:34:05 - INFO - codeparrot_training - Step 33194: {'lr': 0.0001365898581752042, 'samples': 16995840, 'steps': 33194, 'loss/train': 0.4016493558883667} 02/25/2022 20:34:09 - INFO - codeparrot_training - Step 33195: {'lr': 0.0001365752764419587, 'samples': 16996352, 'steps': 33195, 'loss/train': 0.8811741471290588} 02/25/2022 20:34:14 - INFO - codeparrot_training - Step 33196: {'lr': 0.0001365606951945888, 'samples': 16996864, 'steps': 33196, 'loss/train': 2.482672929763794} 02/25/2022 20:34:18 - INFO - codeparrot_training - Step 33197: {'lr': 0.0001365461144331569, 'samples': 16997376, 'steps': 33197, 'loss/train': 0.10973562300205231} 02/25/2022 20:34:23 - INFO - codeparrot_training - Step 33198: {'lr': 0.00013653153415772536, 'samples': 16997888, 'steps': 33198, 'loss/train': 1.4116977453231812} 02/25/2022 20:34:27 - INFO - codeparrot_training - Step 33199: {'lr': 0.00013651695436835666, 'samples': 16998400, 'steps': 33199, 'loss/train': 0.6756554841995239} 02/25/2022 20:34:32 - INFO - codeparrot_training - Step 33200: {'lr': 0.00013650237506511331, 'samples': 16998912, 'steps': 33200, 'loss/train': 0.7271454334259033} 02/25/2022 20:34:36 - INFO - codeparrot_training - Step 33201: {'lr': 0.0001364877962480577, 'samples': 16999424, 'steps': 33201, 'loss/train': 3.139294147491455} 02/25/2022 20:34:41 - INFO - codeparrot_training - Step 33202: {'lr': 0.00013647321791725254, 'samples': 16999936, 'steps': 33202, 'loss/train': 1.9846982955932617} 02/25/2022 20:34:47 - INFO - codeparrot_training - Step 33203: {'lr': 0.00013645864007275982, 'samples': 17000448, 'steps': 33203, 'loss/train': 1.9110839366912842} 02/25/2022 20:34:50 - INFO - codeparrot_training - Step 33204: {'lr': 0.00013644406271464234, 'samples': 17000960, 'steps': 33204, 'loss/train': 0.6278583407402039} 02/25/2022 20:34:56 - INFO - codeparrot_training - Step 33205: {'lr': 0.00013642948584296238, 'samples': 17001472, 'steps': 33205, 'loss/train': 1.556687593460083} 02/25/2022 20:35:00 - INFO - codeparrot_training - Step 33206: {'lr': 0.0001364149094577826, 'samples': 17001984, 'steps': 33206, 'loss/train': 2.284637451171875} 02/25/2022 20:35:05 - INFO - codeparrot_training - Step 33207: {'lr': 0.00013640033355916508, 'samples': 17002496, 'steps': 33207, 'loss/train': 0.92034912109375} 02/25/2022 20:35:09 - INFO - codeparrot_training - Step 33208: {'lr': 0.00013638575814717258, 'samples': 17003008, 'steps': 33208, 'loss/train': 0.9885058999061584} 02/25/2022 20:35:14 - INFO - codeparrot_training - Step 33209: {'lr': 0.00013637118322186732, 'samples': 17003520, 'steps': 33209, 'loss/train': 0.9888502359390259} 02/25/2022 20:35:18 - INFO - codeparrot_training - Step 33210: {'lr': 0.0001363566087833119, 'samples': 17004032, 'steps': 33210, 'loss/train': 1.9248392581939697} 02/25/2022 20:35:23 - INFO - codeparrot_training - Step 33211: {'lr': 0.00013634203483156872, 'samples': 17004544, 'steps': 33211, 'loss/train': 2.1137912273406982} 02/25/2022 20:35:27 - INFO - codeparrot_training - Step 33212: {'lr': 0.00013632746136670016, 'samples': 17005056, 'steps': 33212, 'loss/train': 1.825897455215454} 02/25/2022 20:35:32 - INFO - codeparrot_training - Step 33213: {'lr': 0.00013631288838876853, 'samples': 17005568, 'steps': 33213, 'loss/train': 1.3221735954284668} 02/25/2022 20:35:36 - INFO - codeparrot_training - Step 33214: {'lr': 0.0001362983158978365, 'samples': 17006080, 'steps': 33214, 'loss/train': 1.058686375617981} 02/25/2022 20:35:42 - INFO - codeparrot_training - Step 33215: {'lr': 0.0001362837438939664, 'samples': 17006592, 'steps': 33215, 'loss/train': 2.1753990650177} 02/25/2022 20:35:46 - INFO - codeparrot_training - Step 33216: {'lr': 0.0001362691723772206, 'samples': 17007104, 'steps': 33216, 'loss/train': 2.5073959827423096} 02/25/2022 20:35:51 - INFO - codeparrot_training - Step 33217: {'lr': 0.0001362546013476616, 'samples': 17007616, 'steps': 33217, 'loss/train': 1.8683409690856934} 02/25/2022 20:35:55 - INFO - codeparrot_training - Step 33218: {'lr': 0.00013624003080535163, 'samples': 17008128, 'steps': 33218, 'loss/train': 1.471221685409546} 02/25/2022 20:36:00 - INFO - codeparrot_training - Step 33219: {'lr': 0.00013622546075035336, 'samples': 17008640, 'steps': 33219, 'loss/train': 1.8549385070800781} 02/25/2022 20:36:04 - INFO - codeparrot_training - Step 33220: {'lr': 0.0001362108911827291, 'samples': 17009152, 'steps': 33220, 'loss/train': 1.9270386695861816} 02/25/2022 20:36:09 - INFO - codeparrot_training - Step 33221: {'lr': 0.00013619632210254125, 'samples': 17009664, 'steps': 33221, 'loss/train': 1.8155683279037476} 02/25/2022 20:36:13 - INFO - codeparrot_training - Step 33222: {'lr': 0.00013618175350985206, 'samples': 17010176, 'steps': 33222, 'loss/train': 0.5815594792366028} 02/25/2022 20:36:18 - INFO - codeparrot_training - Step 33223: {'lr': 0.00013616718540472426, 'samples': 17010688, 'steps': 33223, 'loss/train': 2.2447140216827393} 02/25/2022 20:36:22 - INFO - codeparrot_training - Step 33224: {'lr': 0.00013615261778722007, 'samples': 17011200, 'steps': 33224, 'loss/train': 1.4483602046966553} 02/25/2022 20:36:28 - INFO - codeparrot_training - Step 33225: {'lr': 0.00013613805065740191, 'samples': 17011712, 'steps': 33225, 'loss/train': 1.7261251211166382} 02/25/2022 20:36:31 - INFO - codeparrot_training - Step 33226: {'lr': 0.00013612348401533208, 'samples': 17012224, 'steps': 33226, 'loss/train': 2.24259090423584} 02/25/2022 20:36:37 - INFO - codeparrot_training - Step 33227: {'lr': 0.00013610891786107322, 'samples': 17012736, 'steps': 33227, 'loss/train': 1.4078744649887085} 02/25/2022 20:36:40 - INFO - codeparrot_training - Step 33228: {'lr': 0.00013609435219468757, 'samples': 17013248, 'steps': 33228, 'loss/train': 1.9138964414596558} 02/25/2022 20:36:46 - INFO - codeparrot_training - Step 33229: {'lr': 0.00013607978701623755, 'samples': 17013760, 'steps': 33229, 'loss/train': 0.8258975744247437} 02/25/2022 20:36:49 - INFO - codeparrot_training - Step 33230: {'lr': 0.00013606522232578543, 'samples': 17014272, 'steps': 33230, 'loss/train': 1.6112148761749268} 02/25/2022 20:36:55 - INFO - codeparrot_training - Step 33231: {'lr': 0.00013605065812339384, 'samples': 17014784, 'steps': 33231, 'loss/train': 1.1645756959915161} 02/25/2022 20:36:58 - INFO - codeparrot_training - Step 33232: {'lr': 0.00013603609440912507, 'samples': 17015296, 'steps': 33232, 'loss/train': 1.3849626779556274} 02/25/2022 20:37:04 - INFO - codeparrot_training - Step 33233: {'lr': 0.00013602153118304146, 'samples': 17015808, 'steps': 33233, 'loss/train': 1.0744203329086304} 02/25/2022 20:37:07 - INFO - codeparrot_training - Step 33234: {'lr': 0.00013600696844520528, 'samples': 17016320, 'steps': 33234, 'loss/train': 1.4633487462997437} 02/25/2022 20:37:14 - INFO - codeparrot_training - Step 33235: {'lr': 0.0001359924061956792, 'samples': 17016832, 'steps': 33235, 'loss/train': 1.346463680267334} 02/25/2022 20:37:17 - INFO - codeparrot_training - Step 33236: {'lr': 0.00013597784443452533, 'samples': 17017344, 'steps': 33236, 'loss/train': 2.6140599250793457} 02/25/2022 20:37:23 - INFO - codeparrot_training - Step 33237: {'lr': 0.00013596328316180638, 'samples': 17017856, 'steps': 33237, 'loss/train': 2.1065449714660645} 02/25/2022 20:37:26 - INFO - codeparrot_training - Step 33238: {'lr': 0.00013594872237758428, 'samples': 17018368, 'steps': 33238, 'loss/train': 1.4545146226882935} 02/25/2022 20:37:32 - INFO - codeparrot_training - Step 33239: {'lr': 0.00013593416208192177, 'samples': 17018880, 'steps': 33239, 'loss/train': 0.8511667251586914} 02/25/2022 20:37:35 - INFO - codeparrot_training - Step 33240: {'lr': 0.00013591960227488098, 'samples': 17019392, 'steps': 33240, 'loss/train': 0.42749881744384766} 02/25/2022 20:37:41 - INFO - codeparrot_training - Step 33241: {'lr': 0.00013590504295652462, 'samples': 17019904, 'steps': 33241, 'loss/train': 1.4729161262512207} 02/25/2022 20:37:44 - INFO - codeparrot_training - Step 33242: {'lr': 0.00013589048412691458, 'samples': 17020416, 'steps': 33242, 'loss/train': 1.772336483001709} 02/25/2022 20:37:50 - INFO - codeparrot_training - Step 33243: {'lr': 0.0001358759257861136, 'samples': 17020928, 'steps': 33243, 'loss/train': 1.1782811880111694} 02/25/2022 20:37:53 - INFO - codeparrot_training - Step 33244: {'lr': 0.0001358613679341838, 'samples': 17021440, 'steps': 33244, 'loss/train': 2.6781327724456787} 02/25/2022 20:37:59 - INFO - codeparrot_training - Step 33245: {'lr': 0.00013584681057118787, 'samples': 17021952, 'steps': 33245, 'loss/train': 1.1321033239364624} 02/25/2022 20:38:02 - INFO - codeparrot_training - Step 33246: {'lr': 0.00013583225369718778, 'samples': 17022464, 'steps': 33246, 'loss/train': 2.751540422439575} 02/25/2022 20:38:08 - INFO - codeparrot_training - Step 33247: {'lr': 0.0001358176973122461, 'samples': 17022976, 'steps': 33247, 'loss/train': 0.070212721824646} 02/25/2022 20:38:11 - INFO - codeparrot_training - Step 33248: {'lr': 0.00013580314141642508, 'samples': 17023488, 'steps': 33248, 'loss/train': 1.7476670742034912} 02/25/2022 20:38:17 - INFO - codeparrot_training - Step 33249: {'lr': 0.00013578858600978727, 'samples': 17024000, 'steps': 33249, 'loss/train': 1.8255046606063843} 02/25/2022 20:38:20 - INFO - codeparrot_training - Step 33250: {'lr': 0.00013577403109239484, 'samples': 17024512, 'steps': 33250, 'loss/train': 0.021790722385048866} 02/25/2022 20:38:27 - INFO - codeparrot_training - Step 33251: {'lr': 0.0001357594766643102, 'samples': 17025024, 'steps': 33251, 'loss/train': 2.06384539604187} 02/25/2022 20:38:30 - INFO - codeparrot_training - Step 33252: {'lr': 0.00013574492272559565, 'samples': 17025536, 'steps': 33252, 'loss/train': 2.180598258972168} 02/25/2022 20:38:36 - INFO - codeparrot_training - Step 33253: {'lr': 0.0001357303692763136, 'samples': 17026048, 'steps': 33253, 'loss/train': 1.595044732093811} 02/25/2022 20:38:39 - INFO - codeparrot_training - Step 33254: {'lr': 0.00013571581631652638, 'samples': 17026560, 'steps': 33254, 'loss/train': 0.8401834964752197} 02/25/2022 20:38:45 - INFO - codeparrot_training - Step 33255: {'lr': 0.00013570126384629633, 'samples': 17027072, 'steps': 33255, 'loss/train': 0.8191642165184021} 02/25/2022 20:38:48 - INFO - codeparrot_training - Step 33256: {'lr': 0.00013568671186568566, 'samples': 17027584, 'steps': 33256, 'loss/train': 2.2237672805786133} 02/25/2022 20:38:54 - INFO - codeparrot_training - Step 33257: {'lr': 0.0001356721603747569, 'samples': 17028096, 'steps': 33257, 'loss/train': 1.6509199142456055} 02/25/2022 20:38:57 - INFO - codeparrot_training - Step 33258: {'lr': 0.0001356576093735723, 'samples': 17028608, 'steps': 33258, 'loss/train': 1.8999238014221191} 02/25/2022 20:39:02 - INFO - codeparrot_training - Step 33259: {'lr': 0.00013564305886219424, 'samples': 17029120, 'steps': 33259, 'loss/train': 1.1543700695037842} 02/25/2022 20:39:06 - INFO - codeparrot_training - Step 33260: {'lr': 0.00013562850884068486, 'samples': 17029632, 'steps': 33260, 'loss/train': 0.4423770606517792} 02/25/2022 20:39:12 - INFO - codeparrot_training - Step 33261: {'lr': 0.00013561395930910674, 'samples': 17030144, 'steps': 33261, 'loss/train': 1.4676690101623535} 02/25/2022 20:39:16 - INFO - codeparrot_training - Step 33262: {'lr': 0.00013559941026752213, 'samples': 17030656, 'steps': 33262, 'loss/train': 1.2208727598190308} 02/25/2022 20:39:21 - INFO - codeparrot_training - Step 33263: {'lr': 0.00013558486171599327, 'samples': 17031168, 'steps': 33263, 'loss/train': 1.3609212636947632} 02/25/2022 20:39:24 - INFO - codeparrot_training - Step 33264: {'lr': 0.00013557031365458256, 'samples': 17031680, 'steps': 33264, 'loss/train': 1.631249189376831} 02/25/2022 20:39:30 - INFO - codeparrot_training - Step 33265: {'lr': 0.00013555576608335218, 'samples': 17032192, 'steps': 33265, 'loss/train': 1.5071861743927002} 02/25/2022 20:39:33 - INFO - codeparrot_training - Step 33266: {'lr': 0.00013554121900236465, 'samples': 17032704, 'steps': 33266, 'loss/train': 1.347674012184143} 02/25/2022 20:39:39 - INFO - codeparrot_training - Step 33267: {'lr': 0.00013552667241168223, 'samples': 17033216, 'steps': 33267, 'loss/train': 1.5077143907546997} 02/25/2022 20:39:43 - INFO - codeparrot_training - Step 33268: {'lr': 0.00013551212631136717, 'samples': 17033728, 'steps': 33268, 'loss/train': 2.0944697856903076} 02/25/2022 20:39:48 - INFO - codeparrot_training - Step 33269: {'lr': 0.0001354975807014817, 'samples': 17034240, 'steps': 33269, 'loss/train': 0.8736912608146667} 02/25/2022 20:39:52 - INFO - codeparrot_training - Step 33270: {'lr': 0.00013548303558208835, 'samples': 17034752, 'steps': 33270, 'loss/train': 2.776947259902954} 02/25/2022 20:39:58 - INFO - codeparrot_training - Step 33271: {'lr': 0.0001354684909532492, 'samples': 17035264, 'steps': 33271, 'loss/train': 1.0085524320602417} 02/25/2022 20:40:01 - INFO - codeparrot_training - Step 33272: {'lr': 0.00013545394681502689, 'samples': 17035776, 'steps': 33272, 'loss/train': 2.1047728061676025} 02/25/2022 20:40:07 - INFO - codeparrot_training - Step 33273: {'lr': 0.0001354394031674833, 'samples': 17036288, 'steps': 33273, 'loss/train': 1.6534430980682373} 02/25/2022 20:40:11 - INFO - codeparrot_training - Step 33274: {'lr': 0.000135424860010681, 'samples': 17036800, 'steps': 33274, 'loss/train': 3.855248212814331} 02/25/2022 20:40:17 - INFO - codeparrot_training - Step 33275: {'lr': 0.00013541031734468211, 'samples': 17037312, 'steps': 33275, 'loss/train': 1.2242116928100586} 02/25/2022 20:40:20 - INFO - codeparrot_training - Step 33276: {'lr': 0.00013539577516954925, 'samples': 17037824, 'steps': 33276, 'loss/train': 1.2425414323806763} 02/25/2022 20:40:26 - INFO - codeparrot_training - Step 33277: {'lr': 0.00013538123348534435, 'samples': 17038336, 'steps': 33277, 'loss/train': 1.962565541267395} 02/25/2022 20:40:29 - INFO - codeparrot_training - Step 33278: {'lr': 0.0001353666922921299, 'samples': 17038848, 'steps': 33278, 'loss/train': 1.3285073041915894} 02/25/2022 20:40:35 - INFO - codeparrot_training - Step 33279: {'lr': 0.00013535215158996805, 'samples': 17039360, 'steps': 33279, 'loss/train': 1.932297706604004} 02/25/2022 20:40:38 - INFO - codeparrot_training - Step 33280: {'lr': 0.00013533761137892136, 'samples': 17039872, 'steps': 33280, 'loss/train': 1.4528483152389526} 02/25/2022 20:40:44 - INFO - codeparrot_training - Step 33281: {'lr': 0.00013532307165905176, 'samples': 17040384, 'steps': 33281, 'loss/train': 2.297314405441284} 02/25/2022 20:40:48 - INFO - codeparrot_training - Step 33282: {'lr': 0.0001353085324304218, 'samples': 17040896, 'steps': 33282, 'loss/train': 1.5145633220672607} 02/25/2022 20:40:53 - INFO - codeparrot_training - Step 33283: {'lr': 0.00013529399369309353, 'samples': 17041408, 'steps': 33283, 'loss/train': 2.052212715148926} 02/25/2022 20:40:59 - INFO - codeparrot_training - Step 33284: {'lr': 0.0001352794554471295, 'samples': 17041920, 'steps': 33284, 'loss/train': 2.34114408493042} 02/25/2022 20:41:02 - INFO - codeparrot_training - Step 33285: {'lr': 0.00013526491769259185, 'samples': 17042432, 'steps': 33285, 'loss/train': 2.142836332321167} 02/25/2022 20:41:08 - INFO - codeparrot_training - Step 33286: {'lr': 0.00013525038042954288, 'samples': 17042944, 'steps': 33286, 'loss/train': 2.1459133625030518} 02/25/2022 20:41:11 - INFO - codeparrot_training - Step 33287: {'lr': 0.00013523584365804473, 'samples': 17043456, 'steps': 33287, 'loss/train': 2.2856240272521973} 02/25/2022 20:41:17 - INFO - codeparrot_training - Step 33288: {'lr': 0.00013522130737815988, 'samples': 17043968, 'steps': 33288, 'loss/train': 1.3184434175491333} 02/25/2022 20:41:20 - INFO - codeparrot_training - Step 33289: {'lr': 0.00013520677158995048, 'samples': 17044480, 'steps': 33289, 'loss/train': 2.2861545085906982} 02/25/2022 20:41:26 - INFO - codeparrot_training - Step 33290: {'lr': 0.00013519223629347889, 'samples': 17044992, 'steps': 33290, 'loss/train': 0.9833929538726807} 02/25/2022 20:41:29 - INFO - codeparrot_training - Step 33291: {'lr': 0.00013517770148880714, 'samples': 17045504, 'steps': 33291, 'loss/train': 1.1772944927215576} 02/25/2022 20:41:35 - INFO - codeparrot_training - Step 33292: {'lr': 0.00013516316717599776, 'samples': 17046016, 'steps': 33292, 'loss/train': 1.5721584558486938} 02/25/2022 20:41:38 - INFO - codeparrot_training - Step 33293: {'lr': 0.00013514863335511293, 'samples': 17046528, 'steps': 33293, 'loss/train': 1.83499014377594} 02/25/2022 20:41:44 - INFO - codeparrot_training - Step 33294: {'lr': 0.00013513410002621487, 'samples': 17047040, 'steps': 33294, 'loss/train': 1.3630462884902954} 02/25/2022 20:41:47 - INFO - codeparrot_training - Step 33295: {'lr': 0.00013511956718936574, 'samples': 17047552, 'steps': 33295, 'loss/train': 2.10624098777771} 02/25/2022 20:41:53 - INFO - codeparrot_training - Step 33296: {'lr': 0.00013510503484462805, 'samples': 17048064, 'steps': 33296, 'loss/train': 0.9538619518280029} 02/25/2022 20:41:57 - INFO - codeparrot_training - Step 33297: {'lr': 0.0001350905029920639, 'samples': 17048576, 'steps': 33297, 'loss/train': 1.5796390771865845} 02/25/2022 20:42:02 - INFO - codeparrot_training - Step 33298: {'lr': 0.0001350759716317355, 'samples': 17049088, 'steps': 33298, 'loss/train': 1.9168552160263062} 02/25/2022 20:42:06 - INFO - codeparrot_training - Step 33299: {'lr': 0.0001350614407637051, 'samples': 17049600, 'steps': 33299, 'loss/train': 2.140123128890991} 02/25/2022 20:42:11 - INFO - codeparrot_training - Step 33300: {'lr': 0.00013504691038803505, 'samples': 17050112, 'steps': 33300, 'loss/train': 0.7203414440155029} 02/25/2022 20:42:15 - INFO - codeparrot_training - Step 33301: {'lr': 0.00013503238050478755, 'samples': 17050624, 'steps': 33301, 'loss/train': 2.3966708183288574} 02/25/2022 20:42:20 - INFO - codeparrot_training - Step 33302: {'lr': 0.00013501785111402482, 'samples': 17051136, 'steps': 33302, 'loss/train': 1.9984310865402222} 02/25/2022 20:42:24 - INFO - codeparrot_training - Step 33303: {'lr': 0.00013500332221580902, 'samples': 17051648, 'steps': 33303, 'loss/train': 1.9739950895309448} 02/25/2022 20:42:29 - INFO - codeparrot_training - Step 33304: {'lr': 0.00013498879381020255, 'samples': 17052160, 'steps': 33304, 'loss/train': 1.1981984376907349} 02/25/2022 20:42:33 - INFO - codeparrot_training - Step 33305: {'lr': 0.00013497426589726758, 'samples': 17052672, 'steps': 33305, 'loss/train': 1.3749852180480957} 02/25/2022 20:42:39 - INFO - codeparrot_training - Step 33306: {'lr': 0.00013495973847706634, 'samples': 17053184, 'steps': 33306, 'loss/train': 0.14559879899024963} 02/25/2022 20:42:42 - INFO - codeparrot_training - Step 33307: {'lr': 0.00013494521154966093, 'samples': 17053696, 'steps': 33307, 'loss/train': 1.5454174280166626} 02/25/2022 20:42:48 - INFO - codeparrot_training - Step 33308: {'lr': 0.0001349306851151138, 'samples': 17054208, 'steps': 33308, 'loss/train': 0.7908239364624023} 02/25/2022 20:42:51 - INFO - codeparrot_training - Step 33309: {'lr': 0.0001349161591734871, 'samples': 17054720, 'steps': 33309, 'loss/train': 2.9263968467712402} 02/25/2022 20:42:57 - INFO - codeparrot_training - Step 33310: {'lr': 0.0001349016337248429, 'samples': 17055232, 'steps': 33310, 'loss/train': 2.235459566116333} 02/25/2022 20:43:00 - INFO - codeparrot_training - Step 33311: {'lr': 0.00013488710876924376, 'samples': 17055744, 'steps': 33311, 'loss/train': 1.7963942289352417} 02/25/2022 20:43:06 - INFO - codeparrot_training - Step 33312: {'lr': 0.00013487258430675152, 'samples': 17056256, 'steps': 33312, 'loss/train': 0.7760738134384155} 02/25/2022 20:43:09 - INFO - codeparrot_training - Step 33313: {'lr': 0.00013485806033742865, 'samples': 17056768, 'steps': 33313, 'loss/train': 0.8822535276412964} 02/25/2022 20:43:15 - INFO - codeparrot_training - Step 33314: {'lr': 0.00013484353686133717, 'samples': 17057280, 'steps': 33314, 'loss/train': 1.7650905847549438} 02/25/2022 20:43:18 - INFO - codeparrot_training - Step 33315: {'lr': 0.00013482901387853967, 'samples': 17057792, 'steps': 33315, 'loss/train': 2.11855411529541} 02/25/2022 20:43:25 - INFO - codeparrot_training - Step 33316: {'lr': 0.00013481449138909784, 'samples': 17058304, 'steps': 33316, 'loss/train': 1.1121846437454224} 02/25/2022 20:43:29 - INFO - codeparrot_training - Step 33317: {'lr': 0.00013479996939307428, 'samples': 17058816, 'steps': 33317, 'loss/train': 2.221791982650757} 02/25/2022 20:43:34 - INFO - codeparrot_training - Step 33318: {'lr': 0.00013478544789053098, 'samples': 17059328, 'steps': 33318, 'loss/train': 2.2506325244903564} 02/25/2022 20:43:38 - INFO - codeparrot_training - Step 33319: {'lr': 0.00013477092688153033, 'samples': 17059840, 'steps': 33319, 'loss/train': 1.9401451349258423} 02/25/2022 20:43:43 - INFO - codeparrot_training - Step 33320: {'lr': 0.00013475640636613446, 'samples': 17060352, 'steps': 33320, 'loss/train': 2.0966241359710693} 02/25/2022 20:43:47 - INFO - codeparrot_training - Step 33321: {'lr': 0.00013474188634440552, 'samples': 17060864, 'steps': 33321, 'loss/train': 1.7538509368896484} 02/25/2022 20:43:53 - INFO - codeparrot_training - Step 33322: {'lr': 0.00013472736681640563, 'samples': 17061376, 'steps': 33322, 'loss/train': 0.045642364770174026} 02/25/2022 20:43:56 - INFO - codeparrot_training - Step 33323: {'lr': 0.00013471284778219722, 'samples': 17061888, 'steps': 33323, 'loss/train': 2.314727306365967} 02/25/2022 20:44:02 - INFO - codeparrot_training - Step 33324: {'lr': 0.00013469832924184238, 'samples': 17062400, 'steps': 33324, 'loss/train': 2.310760259628296} 02/25/2022 20:44:05 - INFO - codeparrot_training - Step 33325: {'lr': 0.00013468381119540325, 'samples': 17062912, 'steps': 33325, 'loss/train': 0.10972163081169128} 02/25/2022 20:44:12 - INFO - codeparrot_training - Step 33326: {'lr': 0.000134669293642942, 'samples': 17063424, 'steps': 33326, 'loss/train': 1.5418028831481934} 02/25/2022 20:44:16 - INFO - codeparrot_training - Step 33327: {'lr': 0.00013465477658452092, 'samples': 17063936, 'steps': 33327, 'loss/train': 1.0209107398986816} 02/25/2022 20:44:21 - INFO - codeparrot_training - Step 33328: {'lr': 0.0001346402600202022, 'samples': 17064448, 'steps': 33328, 'loss/train': 2.125606060028076} 02/25/2022 20:44:25 - INFO - codeparrot_training - Step 33329: {'lr': 0.00013462574395004792, 'samples': 17064960, 'steps': 33329, 'loss/train': 1.652251958847046} 02/25/2022 20:44:30 - INFO - codeparrot_training - Step 33330: {'lr': 0.00013461122837412026, 'samples': 17065472, 'steps': 33330, 'loss/train': 1.472922682762146} 02/25/2022 20:44:34 - INFO - codeparrot_training - Step 33331: {'lr': 0.00013459671329248154, 'samples': 17065984, 'steps': 33331, 'loss/train': 0.14456412196159363} 02/25/2022 20:44:39 - INFO - codeparrot_training - Step 33332: {'lr': 0.00013458219870519377, 'samples': 17066496, 'steps': 33332, 'loss/train': 1.5198042392730713} 02/25/2022 20:44:43 - INFO - codeparrot_training - Step 33333: {'lr': 0.00013456768461231938, 'samples': 17067008, 'steps': 33333, 'loss/train': 2.345118522644043} 02/25/2022 20:44:48 - INFO - codeparrot_training - Step 33334: {'lr': 0.00013455317101392024, 'samples': 17067520, 'steps': 33334, 'loss/train': 1.6569749116897583} 02/25/2022 20:44:52 - INFO - codeparrot_training - Step 33335: {'lr': 0.0001345386579100587, 'samples': 17068032, 'steps': 33335, 'loss/train': 2.184098243713379} 02/25/2022 20:44:58 - INFO - codeparrot_training - Step 33336: {'lr': 0.0001345241453007968, 'samples': 17068544, 'steps': 33336, 'loss/train': 1.4604190587997437} 02/25/2022 20:45:02 - INFO - codeparrot_training - Step 33337: {'lr': 0.000134509633186197, 'samples': 17069056, 'steps': 33337, 'loss/train': 2.704925298690796} 02/25/2022 20:45:07 - INFO - codeparrot_training - Step 33338: {'lr': 0.00013449512156632105, 'samples': 17069568, 'steps': 33338, 'loss/train': 1.7093887329101562} 02/25/2022 20:45:11 - INFO - codeparrot_training - Step 33339: {'lr': 0.00013448061044123144, 'samples': 17070080, 'steps': 33339, 'loss/train': 1.4775205850601196} 02/25/2022 20:45:16 - INFO - codeparrot_training - Step 33340: {'lr': 0.0001344660998109901, 'samples': 17070592, 'steps': 33340, 'loss/train': 1.6309698820114136} 02/25/2022 20:45:20 - INFO - codeparrot_training - Step 33341: {'lr': 0.0001344515896756595, 'samples': 17071104, 'steps': 33341, 'loss/train': 1.8075356483459473} 02/25/2022 20:45:25 - INFO - codeparrot_training - Step 33342: {'lr': 0.00013443708003530138, 'samples': 17071616, 'steps': 33342, 'loss/train': 1.0016676187515259} 02/25/2022 20:45:29 - INFO - codeparrot_training - Step 33343: {'lr': 0.00013442257088997823, 'samples': 17072128, 'steps': 33343, 'loss/train': 1.301164150238037} 02/25/2022 20:45:34 - INFO - codeparrot_training - Step 33344: {'lr': 0.0001344080622397521, 'samples': 17072640, 'steps': 33344, 'loss/train': 1.7645366191864014} 02/25/2022 20:45:38 - INFO - codeparrot_training - Step 33345: {'lr': 0.00013439355408468502, 'samples': 17073152, 'steps': 33345, 'loss/train': 1.2862515449523926} 02/25/2022 20:45:43 - INFO - codeparrot_training - Step 33346: {'lr': 0.00013437904642483932, 'samples': 17073664, 'steps': 33346, 'loss/train': 2.600736141204834} 02/25/2022 20:45:47 - INFO - codeparrot_training - Step 33347: {'lr': 0.00013436453926027713, 'samples': 17074176, 'steps': 33347, 'loss/train': 1.1010692119598389} 02/25/2022 20:45:53 - INFO - codeparrot_training - Step 33348: {'lr': 0.00013435003259106048, 'samples': 17074688, 'steps': 33348, 'loss/train': 1.8911545276641846} 02/25/2022 20:45:56 - INFO - codeparrot_training - Step 33349: {'lr': 0.0001343355264172515, 'samples': 17075200, 'steps': 33349, 'loss/train': 1.6415482759475708} 02/25/2022 20:46:02 - INFO - codeparrot_training - Step 33350: {'lr': 0.0001343210207389125, 'samples': 17075712, 'steps': 33350, 'loss/train': 1.190313458442688} 02/25/2022 20:46:05 - INFO - codeparrot_training - Step 33351: {'lr': 0.00013430651555610548, 'samples': 17076224, 'steps': 33351, 'loss/train': 2.041227340698242} 02/25/2022 20:46:11 - INFO - codeparrot_training - Step 33352: {'lr': 0.00013429201086889264, 'samples': 17076736, 'steps': 33352, 'loss/train': 1.9330180883407593} 02/25/2022 20:46:14 - INFO - codeparrot_training - Step 33353: {'lr': 0.00013427750667733597, 'samples': 17077248, 'steps': 33353, 'loss/train': 1.4775339365005493} 02/25/2022 20:46:20 - INFO - codeparrot_training - Step 33354: {'lr': 0.00013426300298149783, 'samples': 17077760, 'steps': 33354, 'loss/train': 2.110402822494507} 02/25/2022 20:46:23 - INFO - codeparrot_training - Step 33355: {'lr': 0.00013424849978144022, 'samples': 17078272, 'steps': 33355, 'loss/train': 1.1555075645446777} 02/25/2022 20:46:29 - INFO - codeparrot_training - Step 33356: {'lr': 0.00013423399707722527, 'samples': 17078784, 'steps': 33356, 'loss/train': 1.735811710357666} 02/25/2022 20:46:32 - INFO - codeparrot_training - Step 33357: {'lr': 0.00013421949486891504, 'samples': 17079296, 'steps': 33357, 'loss/train': 3.14457631111145} 02/25/2022 20:46:38 - INFO - codeparrot_training - Step 33358: {'lr': 0.00013420499315657184, 'samples': 17079808, 'steps': 33358, 'loss/train': 1.6877319812774658} 02/25/2022 20:46:41 - INFO - codeparrot_training - Step 33359: {'lr': 0.00013419049194025764, 'samples': 17080320, 'steps': 33359, 'loss/train': 1.126420259475708} 02/25/2022 20:46:47 - INFO - codeparrot_training - Step 33360: {'lr': 0.00013417599122003462, 'samples': 17080832, 'steps': 33360, 'loss/train': 0.7469651699066162} 02/25/2022 20:46:50 - INFO - codeparrot_training - Step 33361: {'lr': 0.0001341614909959648, 'samples': 17081344, 'steps': 33361, 'loss/train': 1.1103416681289673} 02/25/2022 20:46:57 - INFO - codeparrot_training - Step 33362: {'lr': 0.00013414699126811043, 'samples': 17081856, 'steps': 33362, 'loss/train': 2.720579147338867} 02/25/2022 20:47:00 - INFO - codeparrot_training - Step 33363: {'lr': 0.0001341324920365336, 'samples': 17082368, 'steps': 33363, 'loss/train': 1.9618735313415527} 02/25/2022 20:47:06 - INFO - codeparrot_training - Step 33364: {'lr': 0.00013411799330129633, 'samples': 17082880, 'steps': 33364, 'loss/train': 2.219003200531006} 02/25/2022 20:47:09 - INFO - codeparrot_training - Step 33365: {'lr': 0.00013410349506246073, 'samples': 17083392, 'steps': 33365, 'loss/train': 1.4462428092956543} 02/25/2022 20:47:15 - INFO - codeparrot_training - Step 33366: {'lr': 0.00013408899732008902, 'samples': 17083904, 'steps': 33366, 'loss/train': 1.0601718425750732} 02/25/2022 20:47:18 - INFO - codeparrot_training - Step 33367: {'lr': 0.00013407450007424317, 'samples': 17084416, 'steps': 33367, 'loss/train': 8.165153503417969} 02/25/2022 20:47:24 - INFO - codeparrot_training - Step 33368: {'lr': 0.00013406000332498552, 'samples': 17084928, 'steps': 33368, 'loss/train': 1.720931887626648} 02/25/2022 20:47:27 - INFO - codeparrot_training - Step 33369: {'lr': 0.00013404550707237783, 'samples': 17085440, 'steps': 33369, 'loss/train': 1.3504197597503662} 02/25/2022 20:47:33 - INFO - codeparrot_training - Step 33370: {'lr': 0.00013403101131648245, 'samples': 17085952, 'steps': 33370, 'loss/train': 1.1937004327774048} 02/25/2022 20:47:36 - INFO - codeparrot_training - Step 33371: {'lr': 0.00013401651605736133, 'samples': 17086464, 'steps': 33371, 'loss/train': 3.131883144378662} 02/25/2022 20:47:42 - INFO - codeparrot_training - Step 33372: {'lr': 0.00013400202129507677, 'samples': 17086976, 'steps': 33372, 'loss/train': 0.11477970331907272} 02/25/2022 20:47:46 - INFO - codeparrot_training - Step 33373: {'lr': 0.00013398752702969055, 'samples': 17087488, 'steps': 33373, 'loss/train': 1.726823091506958} 02/25/2022 20:47:51 - INFO - codeparrot_training - Step 33374: {'lr': 0.00013397303326126504, 'samples': 17088000, 'steps': 33374, 'loss/train': 0.8116297125816345} 02/25/2022 20:47:55 - INFO - codeparrot_training - Step 33375: {'lr': 0.00013395853998986206, 'samples': 17088512, 'steps': 33375, 'loss/train': 1.0939476490020752} 02/25/2022 20:48:00 - INFO - codeparrot_training - Step 33376: {'lr': 0.0001339440472155441, 'samples': 17089024, 'steps': 33376, 'loss/train': 1.861535906791687} 02/25/2022 20:48:04 - INFO - codeparrot_training - Step 33377: {'lr': 0.00013392955493837279, 'samples': 17089536, 'steps': 33377, 'loss/train': 1.3148138523101807} 02/25/2022 20:48:10 - INFO - codeparrot_training - Step 33378: {'lr': 0.00013391506315841046, 'samples': 17090048, 'steps': 33378, 'loss/train': 0.2312527298927307} 02/25/2022 20:48:13 - INFO - codeparrot_training - Step 33379: {'lr': 0.0001339005718757191, 'samples': 17090560, 'steps': 33379, 'loss/train': 1.6715458631515503} 02/25/2022 20:48:19 - INFO - codeparrot_training - Step 33380: {'lr': 0.00013388608109036085, 'samples': 17091072, 'steps': 33380, 'loss/train': 2.4183127880096436} 02/25/2022 20:48:22 - INFO - codeparrot_training - Step 33381: {'lr': 0.00013387159080239781, 'samples': 17091584, 'steps': 33381, 'loss/train': 0.658007025718689} 02/25/2022 20:48:29 - INFO - codeparrot_training - Step 33382: {'lr': 0.00013385710101189197, 'samples': 17092096, 'steps': 33382, 'loss/train': 1.1469815969467163} 02/25/2022 20:48:32 - INFO - codeparrot_training - Step 33383: {'lr': 0.00013384261171890539, 'samples': 17092608, 'steps': 33383, 'loss/train': 1.5197700262069702} 02/25/2022 20:48:38 - INFO - codeparrot_training - Step 33384: {'lr': 0.00013382812292350022, 'samples': 17093120, 'steps': 33384, 'loss/train': 1.2824183702468872} 02/25/2022 20:48:41 - INFO - codeparrot_training - Step 33385: {'lr': 0.00013381363462573848, 'samples': 17093632, 'steps': 33385, 'loss/train': 0.02857513353228569} 02/25/2022 20:48:47 - INFO - codeparrot_training - Step 33386: {'lr': 0.0001337991468256823, 'samples': 17094144, 'steps': 33386, 'loss/train': 1.6670411825180054} 02/25/2022 20:48:50 - INFO - codeparrot_training - Step 33387: {'lr': 0.0001337846595233935, 'samples': 17094656, 'steps': 33387, 'loss/train': 1.1680760383605957} 02/25/2022 20:48:56 - INFO - codeparrot_training - Step 33388: {'lr': 0.00013377017271893444, 'samples': 17095168, 'steps': 33388, 'loss/train': 1.1745519638061523} 02/25/2022 20:48:59 - INFO - codeparrot_training - Step 33389: {'lr': 0.00013375568641236707, 'samples': 17095680, 'steps': 33389, 'loss/train': 1.0801969766616821} 02/25/2022 20:49:05 - INFO - codeparrot_training - Step 33390: {'lr': 0.0001337412006037534, 'samples': 17096192, 'steps': 33390, 'loss/train': 2.1362245082855225} 02/25/2022 20:49:08 - INFO - codeparrot_training - Step 33391: {'lr': 0.00013372671529315544, 'samples': 17096704, 'steps': 33391, 'loss/train': 1.218953013420105} 02/25/2022 20:49:14 - INFO - codeparrot_training - Step 33392: {'lr': 0.00013371223048063541, 'samples': 17097216, 'steps': 33392, 'loss/train': 1.3273156881332397} 02/25/2022 20:49:17 - INFO - codeparrot_training - Step 33393: {'lr': 0.00013369774616625525, 'samples': 17097728, 'steps': 33393, 'loss/train': 1.6569035053253174} 02/25/2022 20:49:24 - INFO - codeparrot_training - Step 33394: {'lr': 0.000133683262350077, 'samples': 17098240, 'steps': 33394, 'loss/train': 1.5534240007400513} 02/25/2022 20:49:27 - INFO - codeparrot_training - Step 33395: {'lr': 0.00013366877903216273, 'samples': 17098752, 'steps': 33395, 'loss/train': 2.520770311355591} 02/25/2022 20:49:33 - INFO - codeparrot_training - Step 33396: {'lr': 0.0001336542962125744, 'samples': 17099264, 'steps': 33396, 'loss/train': 2.2042407989501953} 02/25/2022 20:49:36 - INFO - codeparrot_training - Step 33397: {'lr': 0.0001336398138913742, 'samples': 17099776, 'steps': 33397, 'loss/train': 1.7514128684997559} 02/25/2022 20:49:42 - INFO - codeparrot_training - Step 33398: {'lr': 0.0001336253320686241, 'samples': 17100288, 'steps': 33398, 'loss/train': 1.811836838722229} 02/25/2022 20:49:45 - INFO - codeparrot_training - Step 33399: {'lr': 0.0001336108507443861, 'samples': 17100800, 'steps': 33399, 'loss/train': 0.68487948179245} 02/25/2022 20:49:51 - INFO - codeparrot_training - Step 33400: {'lr': 0.00013359636991872216, 'samples': 17101312, 'steps': 33400, 'loss/train': 2.6286821365356445} 02/25/2022 20:49:56 - INFO - codeparrot_training - Step 33401: {'lr': 0.00013358188959169452, 'samples': 17101824, 'steps': 33401, 'loss/train': 1.5611859560012817} 02/25/2022 20:50:00 - INFO - codeparrot_training - Step 33402: {'lr': 0.00013356740976336507, 'samples': 17102336, 'steps': 33402, 'loss/train': 1.8862221240997314} 02/25/2022 20:50:05 - INFO - codeparrot_training - Step 33403: {'lr': 0.0001335529304337959, 'samples': 17102848, 'steps': 33403, 'loss/train': 2.2275445461273193} 02/25/2022 20:50:09 - INFO - codeparrot_training - Step 33404: {'lr': 0.0001335384516030489, 'samples': 17103360, 'steps': 33404, 'loss/train': 1.8185585737228394} 02/25/2022 20:50:14 - INFO - codeparrot_training - Step 33405: {'lr': 0.00013352397327118625, 'samples': 17103872, 'steps': 33405, 'loss/train': 1.7012587785720825} 02/25/2022 20:50:18 - INFO - codeparrot_training - Step 33406: {'lr': 0.00013350949543826986, 'samples': 17104384, 'steps': 33406, 'loss/train': 0.965317964553833} 02/25/2022 20:50:24 - INFO - codeparrot_training - Step 33407: {'lr': 0.000133495018104362, 'samples': 17104896, 'steps': 33407, 'loss/train': 2.268836736679077} 02/25/2022 20:50:27 - INFO - codeparrot_training - Step 33408: {'lr': 0.00013348054126952427, 'samples': 17105408, 'steps': 33408, 'loss/train': 1.6885693073272705} 02/25/2022 20:50:33 - INFO - codeparrot_training - Step 33409: {'lr': 0.000133466064933819, 'samples': 17105920, 'steps': 33409, 'loss/train': 1.1213762760162354} 02/25/2022 20:50:37 - INFO - codeparrot_training - Step 33410: {'lr': 0.00013345158909730802, 'samples': 17106432, 'steps': 33410, 'loss/train': 2.3392863273620605} 02/25/2022 20:50:42 - INFO - codeparrot_training - Step 33411: {'lr': 0.00013343711376005363, 'samples': 17106944, 'steps': 33411, 'loss/train': 1.5026854276657104} 02/25/2022 20:50:45 - INFO - codeparrot_training - Step 33412: {'lr': 0.0001334226389221174, 'samples': 17107456, 'steps': 33412, 'loss/train': 1.0822640657424927} 02/25/2022 20:50:51 - INFO - codeparrot_training - Step 33413: {'lr': 0.00013340816458356168, 'samples': 17107968, 'steps': 33413, 'loss/train': 1.0653711557388306} 02/25/2022 20:50:54 - INFO - codeparrot_training - Step 33414: {'lr': 0.00013339369074444825, 'samples': 17108480, 'steps': 33414, 'loss/train': 2.0920493602752686} 02/25/2022 20:51:00 - INFO - codeparrot_training - Step 33415: {'lr': 0.00013337921740483945, 'samples': 17108992, 'steps': 33415, 'loss/train': 3.0703022480010986} 02/25/2022 20:51:03 - INFO - codeparrot_training - Step 33416: {'lr': 0.00013336474456479685, 'samples': 17109504, 'steps': 33416, 'loss/train': 2.0900466442108154} 02/25/2022 20:51:10 - INFO - codeparrot_training - Step 33417: {'lr': 0.00013335027222438273, 'samples': 17110016, 'steps': 33417, 'loss/train': 1.3814113140106201} 02/25/2022 20:51:13 - INFO - codeparrot_training - Step 33418: {'lr': 0.00013333580038365892, 'samples': 17110528, 'steps': 33418, 'loss/train': 2.204141616821289} 02/25/2022 20:51:19 - INFO - codeparrot_training - Step 33419: {'lr': 0.00013332132904268757, 'samples': 17111040, 'steps': 33419, 'loss/train': 0.8567627668380737} 02/25/2022 20:51:22 - INFO - codeparrot_training - Step 33420: {'lr': 0.0001333068582015306, 'samples': 17111552, 'steps': 33420, 'loss/train': 2.138132333755493} 02/25/2022 20:51:28 - INFO - codeparrot_training - Step 33421: {'lr': 0.00013329238786025002, 'samples': 17112064, 'steps': 33421, 'loss/train': 2.030482292175293} 02/25/2022 20:51:31 - INFO - codeparrot_training - Step 33422: {'lr': 0.0001332779180189077, 'samples': 17112576, 'steps': 33422, 'loss/train': 2.1376850605010986} 02/25/2022 20:51:37 - INFO - codeparrot_training - Step 33423: {'lr': 0.00013326344867756578, 'samples': 17113088, 'steps': 33423, 'loss/train': 1.081028938293457} 02/25/2022 20:51:40 - INFO - codeparrot_training - Step 33424: {'lr': 0.0001332489798362862, 'samples': 17113600, 'steps': 33424, 'loss/train': 2.4660117626190186} 02/25/2022 20:51:46 - INFO - codeparrot_training - Step 33425: {'lr': 0.00013323451149513092, 'samples': 17114112, 'steps': 33425, 'loss/train': 1.915223240852356} 02/25/2022 20:51:49 - INFO - codeparrot_training - Step 33426: {'lr': 0.0001332200436541618, 'samples': 17114624, 'steps': 33426, 'loss/train': 1.6988694667816162} 02/25/2022 20:51:55 - INFO - codeparrot_training - Step 33427: {'lr': 0.00013320557631344104, 'samples': 17115136, 'steps': 33427, 'loss/train': 1.6248199939727783} 02/25/2022 20:51:58 - INFO - codeparrot_training - Step 33428: {'lr': 0.00013319110947303047, 'samples': 17115648, 'steps': 33428, 'loss/train': 1.44893217086792} 02/25/2022 20:52:04 - INFO - codeparrot_training - Step 33429: {'lr': 0.00013317664313299216, 'samples': 17116160, 'steps': 33429, 'loss/train': 1.6165456771850586} 02/25/2022 20:52:08 - INFO - codeparrot_training - Step 33430: {'lr': 0.00013316217729338785, 'samples': 17116672, 'steps': 33430, 'loss/train': 0.09326335042715073} 02/25/2022 20:52:13 - INFO - codeparrot_training - Step 33431: {'lr': 0.00013314771195427984, 'samples': 17117184, 'steps': 33431, 'loss/train': 1.220788598060608} 02/25/2022 20:52:17 - INFO - codeparrot_training - Step 33432: {'lr': 0.00013313324711572988, 'samples': 17117696, 'steps': 33432, 'loss/train': 1.998371958732605} 02/25/2022 20:52:22 - INFO - codeparrot_training - Step 33433: {'lr': 0.00013311878277779999, 'samples': 17118208, 'steps': 33433, 'loss/train': 1.4682552814483643} 02/25/2022 20:52:26 - INFO - codeparrot_training - Step 33434: {'lr': 0.00013310431894055202, 'samples': 17118720, 'steps': 33434, 'loss/train': 0.6298272013664246} 02/25/2022 20:52:31 - INFO - codeparrot_training - Step 33435: {'lr': 0.00013308985560404817, 'samples': 17119232, 'steps': 33435, 'loss/train': 2.096583127975464} 02/25/2022 20:52:35 - INFO - codeparrot_training - Step 33436: {'lr': 0.00013307539276835017, 'samples': 17119744, 'steps': 33436, 'loss/train': 0.9860432147979736} 02/25/2022 20:52:40 - INFO - codeparrot_training - Step 33437: {'lr': 0.00013306093043352014, 'samples': 17120256, 'steps': 33437, 'loss/train': 1.5159927606582642} 02/25/2022 20:52:44 - INFO - codeparrot_training - Step 33438: {'lr': 0.00013304646859961983, 'samples': 17120768, 'steps': 33438, 'loss/train': 2.343961238861084} 02/25/2022 20:52:50 - INFO - codeparrot_training - Step 33439: {'lr': 0.0001330320072667114, 'samples': 17121280, 'steps': 33439, 'loss/train': 1.0366311073303223} 02/25/2022 20:52:54 - INFO - codeparrot_training - Step 33440: {'lr': 0.0001330175464348567, 'samples': 17121792, 'steps': 33440, 'loss/train': 2.18906569480896} 02/25/2022 20:52:59 - INFO - codeparrot_training - Step 33441: {'lr': 0.0001330030861041176, 'samples': 17122304, 'steps': 33441, 'loss/train': 2.3539037704467773} 02/25/2022 20:53:02 - INFO - codeparrot_training - Step 33442: {'lr': 0.0001329886262745563, 'samples': 17122816, 'steps': 33442, 'loss/train': 1.4886040687561035} 02/25/2022 20:53:08 - INFO - codeparrot_training - Step 33443: {'lr': 0.0001329741669462344, 'samples': 17123328, 'steps': 33443, 'loss/train': 1.362316608428955} 02/25/2022 20:53:11 - INFO - codeparrot_training - Step 33444: {'lr': 0.0001329597081192141, 'samples': 17123840, 'steps': 33444, 'loss/train': 1.4776610136032104} 02/25/2022 20:53:17 - INFO - codeparrot_training - Step 33445: {'lr': 0.00013294524979355716, 'samples': 17124352, 'steps': 33445, 'loss/train': 1.8137837648391724} 02/25/2022 20:53:20 - INFO - codeparrot_training - Step 33446: {'lr': 0.00013293079196932578, 'samples': 17124864, 'steps': 33446, 'loss/train': 1.231760859489441} 02/25/2022 20:53:26 - INFO - codeparrot_training - Step 33447: {'lr': 0.00013291633464658154, 'samples': 17125376, 'steps': 33447, 'loss/train': 1.477229356765747} 02/25/2022 20:53:30 - INFO - codeparrot_training - Step 33448: {'lr': 0.00013290187782538662, 'samples': 17125888, 'steps': 33448, 'loss/train': 1.839331865310669} 02/25/2022 20:53:35 - INFO - codeparrot_training - Step 33449: {'lr': 0.0001328874215058028, 'samples': 17126400, 'steps': 33449, 'loss/train': 2.200927257537842} 02/25/2022 20:53:39 - INFO - codeparrot_training - Step 33450: {'lr': 0.00013287296568789225, 'samples': 17126912, 'steps': 33450, 'loss/train': 1.520546317100525} 02/25/2022 20:53:44 - INFO - codeparrot_training - Step 33451: {'lr': 0.00013285851037171654, 'samples': 17127424, 'steps': 33451, 'loss/train': 2.7835500240325928} 02/25/2022 20:53:48 - INFO - codeparrot_training - Step 33452: {'lr': 0.00013284405555733785, 'samples': 17127936, 'steps': 33452, 'loss/train': 2.5864064693450928} 02/25/2022 20:53:53 - INFO - codeparrot_training - Step 33453: {'lr': 0.00013282960124481793, 'samples': 17128448, 'steps': 33453, 'loss/train': 2.0206918716430664} 02/25/2022 20:53:57 - INFO - codeparrot_training - Step 33454: {'lr': 0.0001328151474342189, 'samples': 17128960, 'steps': 33454, 'loss/train': 1.7264811992645264} 02/25/2022 20:54:03 - INFO - codeparrot_training - Step 33455: {'lr': 0.0001328006941256026, 'samples': 17129472, 'steps': 33455, 'loss/train': 2.172121524810791} 02/25/2022 20:54:07 - INFO - codeparrot_training - Step 33456: {'lr': 0.00013278624131903086, 'samples': 17129984, 'steps': 33456, 'loss/train': 1.7189058065414429} 02/25/2022 20:54:12 - INFO - codeparrot_training - Step 33457: {'lr': 0.00013277178901456558, 'samples': 17130496, 'steps': 33457, 'loss/train': 1.3464924097061157} 02/25/2022 20:54:15 - INFO - codeparrot_training - Step 33458: {'lr': 0.0001327573372122688, 'samples': 17131008, 'steps': 33458, 'loss/train': 1.9749528169631958} 02/25/2022 20:54:21 - INFO - codeparrot_training - Step 33459: {'lr': 0.00013274288591220241, 'samples': 17131520, 'steps': 33459, 'loss/train': 0.8323876857757568} 02/25/2022 20:54:24 - INFO - codeparrot_training - Step 33460: {'lr': 0.00013272843511442822, 'samples': 17132032, 'steps': 33460, 'loss/train': 1.2661223411560059} 02/25/2022 20:54:30 - INFO - codeparrot_training - Step 33461: {'lr': 0.00013271398481900807, 'samples': 17132544, 'steps': 33461, 'loss/train': 2.106449604034424} 02/25/2022 20:54:33 - INFO - codeparrot_training - Step 33462: {'lr': 0.00013269953502600406, 'samples': 17133056, 'steps': 33462, 'loss/train': 1.5974721908569336} 02/25/2022 20:54:39 - INFO - codeparrot_training - Step 33463: {'lr': 0.00013268508573547805, 'samples': 17133568, 'steps': 33463, 'loss/train': 8.676044464111328} 02/25/2022 20:54:42 - INFO - codeparrot_training - Step 33464: {'lr': 0.0001326706369474918, 'samples': 17134080, 'steps': 33464, 'loss/train': 2.2265353202819824} 02/25/2022 20:54:49 - INFO - codeparrot_training - Step 33465: {'lr': 0.00013265618866210727, 'samples': 17134592, 'steps': 33465, 'loss/train': 1.908743143081665} 02/25/2022 20:54:52 - INFO - codeparrot_training - Step 33466: {'lr': 0.00013264174087938642, 'samples': 17135104, 'steps': 33466, 'loss/train': 2.4594006538391113} 02/25/2022 20:54:58 - INFO - codeparrot_training - Step 33467: {'lr': 0.00013262729359939102, 'samples': 17135616, 'steps': 33467, 'loss/train': 1.798116683959961} 02/25/2022 20:55:01 - INFO - codeparrot_training - Step 33468: {'lr': 0.0001326128468221832, 'samples': 17136128, 'steps': 33468, 'loss/train': 2.2674553394317627} 02/25/2022 20:55:07 - INFO - codeparrot_training - Step 33469: {'lr': 0.00013259840054782445, 'samples': 17136640, 'steps': 33469, 'loss/train': 1.3644338846206665} 02/25/2022 20:55:10 - INFO - codeparrot_training - Step 33470: {'lr': 0.000132583954776377, 'samples': 17137152, 'steps': 33470, 'loss/train': 1.0547000169754028} 02/25/2022 20:55:16 - INFO - codeparrot_training - Step 33471: {'lr': 0.0001325695095079025, 'samples': 17137664, 'steps': 33471, 'loss/train': 1.966468334197998} 02/25/2022 20:55:19 - INFO - codeparrot_training - Step 33472: {'lr': 0.00013255506474246312, 'samples': 17138176, 'steps': 33472, 'loss/train': 0.9320878982543945} 02/25/2022 20:55:25 - INFO - codeparrot_training - Step 33473: {'lr': 0.00013254062048012038, 'samples': 17138688, 'steps': 33473, 'loss/train': 1.82168710231781} 02/25/2022 20:55:28 - INFO - codeparrot_training - Step 33474: {'lr': 0.00013252617672093639, 'samples': 17139200, 'steps': 33474, 'loss/train': 1.3933069705963135} 02/25/2022 20:55:34 - INFO - codeparrot_training - Step 33475: {'lr': 0.00013251173346497298, 'samples': 17139712, 'steps': 33475, 'loss/train': 1.382158637046814} 02/25/2022 20:55:38 - INFO - codeparrot_training - Step 33476: {'lr': 0.0001324972907122919, 'samples': 17140224, 'steps': 33476, 'loss/train': 1.639510989189148} 02/25/2022 20:55:43 - INFO - codeparrot_training - Step 33477: {'lr': 0.0001324828484629552, 'samples': 17140736, 'steps': 33477, 'loss/train': 0.8593530058860779} 02/25/2022 20:55:47 - INFO - codeparrot_training - Step 33478: {'lr': 0.00013246840671702465, 'samples': 17141248, 'steps': 33478, 'loss/train': 2.601548910140991} 02/25/2022 20:55:52 - INFO - codeparrot_training - Step 33479: {'lr': 0.00013245396547456213, 'samples': 17141760, 'steps': 33479, 'loss/train': 1.5943324565887451} 02/25/2022 20:55:56 - INFO - codeparrot_training - Step 33480: {'lr': 0.00013243952473562942, 'samples': 17142272, 'steps': 33480, 'loss/train': 1.4353269338607788} 02/25/2022 20:56:01 - INFO - codeparrot_training - Step 33481: {'lr': 0.00013242508450028853, 'samples': 17142784, 'steps': 33481, 'loss/train': 0.19118362665176392} 02/25/2022 20:56:05 - INFO - codeparrot_training - Step 33482: {'lr': 0.00013241064476860126, 'samples': 17143296, 'steps': 33482, 'loss/train': 1.215999722480774} 02/25/2022 20:56:10 - INFO - codeparrot_training - Step 33483: {'lr': 0.00013239620554062945, 'samples': 17143808, 'steps': 33483, 'loss/train': 1.8257240056991577} 02/25/2022 20:56:14 - INFO - codeparrot_training - Step 33484: {'lr': 0.00013238176681643486, 'samples': 17144320, 'steps': 33484, 'loss/train': 2.3788487911224365} 02/25/2022 20:56:20 - INFO - codeparrot_training - Step 33485: {'lr': 0.00013236732859607953, 'samples': 17144832, 'steps': 33485, 'loss/train': 1.8935496807098389} 02/25/2022 20:56:23 - INFO - codeparrot_training - Step 33486: {'lr': 0.0001323528908796252, 'samples': 17145344, 'steps': 33486, 'loss/train': 2.148883581161499} 02/25/2022 20:56:29 - INFO - codeparrot_training - Step 33487: {'lr': 0.00013233845366713372, 'samples': 17145856, 'steps': 33487, 'loss/train': 1.2525650262832642} 02/25/2022 20:56:32 - INFO - codeparrot_training - Step 33488: {'lr': 0.00013232401695866685, 'samples': 17146368, 'steps': 33488, 'loss/train': 1.4463725090026855} 02/25/2022 20:56:38 - INFO - codeparrot_training - Step 33489: {'lr': 0.00013230958075428662, 'samples': 17146880, 'steps': 33489, 'loss/train': 1.3483624458312988} 02/25/2022 20:56:41 - INFO - codeparrot_training - Step 33490: {'lr': 0.00013229514505405478, 'samples': 17147392, 'steps': 33490, 'loss/train': 2.2947733402252197} 02/25/2022 20:56:47 - INFO - codeparrot_training - Step 33491: {'lr': 0.00013228070985803317, 'samples': 17147904, 'steps': 33491, 'loss/train': 2.344132423400879} 02/25/2022 20:56:50 - INFO - codeparrot_training - Step 33492: {'lr': 0.00013226627516628353, 'samples': 17148416, 'steps': 33492, 'loss/train': 0.8506695628166199} 02/25/2022 20:56:56 - INFO - codeparrot_training - Step 33493: {'lr': 0.00013225184097886785, 'samples': 17148928, 'steps': 33493, 'loss/train': 2.137237787246704} 02/25/2022 20:56:59 - INFO - codeparrot_training - Step 33494: {'lr': 0.00013223740729584793, 'samples': 17149440, 'steps': 33494, 'loss/train': 1.5528056621551514} 02/25/2022 20:57:05 - INFO - codeparrot_training - Step 33495: {'lr': 0.00013222297411728554, 'samples': 17149952, 'steps': 33495, 'loss/train': 1.4121397733688354} 02/25/2022 20:57:08 - INFO - codeparrot_training - Step 33496: {'lr': 0.00013220854144324243, 'samples': 17150464, 'steps': 33496, 'loss/train': 1.4583725929260254} 02/25/2022 20:57:14 - INFO - codeparrot_training - Step 33497: {'lr': 0.00013219410927378062, 'samples': 17150976, 'steps': 33497, 'loss/train': 1.376801609992981} 02/25/2022 20:57:17 - INFO - codeparrot_training - Step 33498: {'lr': 0.00013217967760896187, 'samples': 17151488, 'steps': 33498, 'loss/train': 1.8423397541046143} 02/25/2022 20:57:23 - INFO - codeparrot_training - Step 33499: {'lr': 0.0001321652464488479, 'samples': 17152000, 'steps': 33499, 'loss/train': 0.0896710753440857} 02/25/2022 20:57:26 - INFO - codeparrot_training - Step 33500: {'lr': 0.00013215081579350058, 'samples': 17152512, 'steps': 33500, 'loss/train': 1.9780653715133667} 02/25/2022 20:57:33 - INFO - codeparrot_training - Step 33501: {'lr': 0.0001321363856429818, 'samples': 17153024, 'steps': 33501, 'loss/train': 1.430535078048706} 02/25/2022 20:57:37 - INFO - codeparrot_training - Step 33502: {'lr': 0.00013212195599735324, 'samples': 17153536, 'steps': 33502, 'loss/train': 0.815642774105072} 02/25/2022 20:57:42 - INFO - codeparrot_training - Step 33503: {'lr': 0.00013210752685667698, 'samples': 17154048, 'steps': 33503, 'loss/train': 0.4819067120552063} 02/25/2022 20:57:46 - INFO - codeparrot_training - Step 33504: {'lr': 0.00013209309822101445, 'samples': 17154560, 'steps': 33504, 'loss/train': 1.550699234008789} 02/25/2022 20:57:51 - INFO - codeparrot_training - Step 33505: {'lr': 0.00013207867009042774, 'samples': 17155072, 'steps': 33505, 'loss/train': 0.050010789185762405} 02/25/2022 20:57:55 - INFO - codeparrot_training - Step 33506: {'lr': 0.00013206424246497845, 'samples': 17155584, 'steps': 33506, 'loss/train': 2.055128335952759} 02/25/2022 20:58:00 - INFO - codeparrot_training - Step 33507: {'lr': 0.0001320498153447287, 'samples': 17156096, 'steps': 33507, 'loss/train': 1.144755244255066} 02/25/2022 20:58:06 - INFO - codeparrot_training - Step 33508: {'lr': 0.00013203538872973988, 'samples': 17156608, 'steps': 33508, 'loss/train': 1.7556042671203613} 02/25/2022 20:58:09 - INFO - codeparrot_training - Step 33509: {'lr': 0.00013202096262007412, 'samples': 17157120, 'steps': 33509, 'loss/train': 0.9784806966781616} 02/25/2022 20:58:13 - INFO - codeparrot_training - Step 33510: {'lr': 0.00013200653701579297, 'samples': 17157632, 'steps': 33510, 'loss/train': 2.0428335666656494} 02/25/2022 20:58:20 - INFO - codeparrot_training - Step 33511: {'lr': 0.00013199211191695858, 'samples': 17158144, 'steps': 33511, 'loss/train': 1.5901081562042236} 02/25/2022 20:58:25 - INFO - codeparrot_training - Step 33512: {'lr': 0.0001319776873236323, 'samples': 17158656, 'steps': 33512, 'loss/train': 2.3524787425994873} 02/25/2022 20:58:28 - INFO - codeparrot_training - Step 33513: {'lr': 0.0001319632632358762, 'samples': 17159168, 'steps': 33513, 'loss/train': 1.2886229753494263} 02/25/2022 20:58:34 - INFO - codeparrot_training - Step 33514: {'lr': 0.00013194883965375194, 'samples': 17159680, 'steps': 33514, 'loss/train': 2.036590337753296} 02/25/2022 20:58:38 - INFO - codeparrot_training - Step 33515: {'lr': 0.0001319344165773214, 'samples': 17160192, 'steps': 33515, 'loss/train': 1.3453980684280396} 02/25/2022 20:58:43 - INFO - codeparrot_training - Step 33516: {'lr': 0.0001319199940066464, 'samples': 17160704, 'steps': 33516, 'loss/train': 1.2420707941055298} 02/25/2022 20:58:46 - INFO - codeparrot_training - Step 33517: {'lr': 0.00013190557194178864, 'samples': 17161216, 'steps': 33517, 'loss/train': 1.217384934425354} 02/25/2022 20:58:52 - INFO - codeparrot_training - Step 33518: {'lr': 0.0001318911503828098, 'samples': 17161728, 'steps': 33518, 'loss/train': 1.4960331916809082} 02/25/2022 20:58:55 - INFO - codeparrot_training - Step 33519: {'lr': 0.00013187672932977185, 'samples': 17162240, 'steps': 33519, 'loss/train': 1.9133321046829224} 02/25/2022 20:59:02 - INFO - codeparrot_training - Step 33520: {'lr': 0.00013186230878273653, 'samples': 17162752, 'steps': 33520, 'loss/train': 2.5817580223083496} 02/25/2022 20:59:06 - INFO - codeparrot_training - Step 33521: {'lr': 0.0001318478887417655, 'samples': 17163264, 'steps': 33521, 'loss/train': 2.210131883621216} 02/25/2022 20:59:11 - INFO - codeparrot_training - Step 33522: {'lr': 0.00013183346920692067, 'samples': 17163776, 'steps': 33522, 'loss/train': 1.3452799320220947} 02/25/2022 20:59:15 - INFO - codeparrot_training - Step 33523: {'lr': 0.00013181905017826357, 'samples': 17164288, 'steps': 33523, 'loss/train': 1.9927799701690674} 02/25/2022 20:59:20 - INFO - codeparrot_training - Step 33524: {'lr': 0.00013180463165585627, 'samples': 17164800, 'steps': 33524, 'loss/train': 0.8942004442214966} 02/25/2022 20:59:24 - INFO - codeparrot_training - Step 33525: {'lr': 0.0001317902136397604, 'samples': 17165312, 'steps': 33525, 'loss/train': 1.6530109643936157} 02/25/2022 20:59:29 - INFO - codeparrot_training - Step 33526: {'lr': 0.0001317757961300377, 'samples': 17165824, 'steps': 33526, 'loss/train': 1.8047325611114502} 02/25/2022 20:59:33 - INFO - codeparrot_training - Step 33527: {'lr': 0.00013176137912674984, 'samples': 17166336, 'steps': 33527, 'loss/train': 1.8546675443649292} 02/25/2022 20:59:38 - INFO - codeparrot_training - Step 33528: {'lr': 0.00013174696262995884, 'samples': 17166848, 'steps': 33528, 'loss/train': 2.9564239978790283} 02/25/2022 20:59:41 - INFO - codeparrot_training - Step 33529: {'lr': 0.00013173254663972623, 'samples': 17167360, 'steps': 33529, 'loss/train': 1.8618310689926147} 02/25/2022 20:59:48 - INFO - codeparrot_training - Step 33530: {'lr': 0.00013171813115611387, 'samples': 17167872, 'steps': 33530, 'loss/train': 1.823944330215454} 02/25/2022 20:59:51 - INFO - codeparrot_training - Step 33531: {'lr': 0.0001317037161791834, 'samples': 17168384, 'steps': 33531, 'loss/train': 1.596149206161499} 02/25/2022 20:59:57 - INFO - codeparrot_training - Step 33532: {'lr': 0.00013168930170899673, 'samples': 17168896, 'steps': 33532, 'loss/train': 2.4969022274017334} 02/25/2022 21:00:00 - INFO - codeparrot_training - Step 33533: {'lr': 0.00013167488774561552, 'samples': 17169408, 'steps': 33533, 'loss/train': 0.9365255236625671} 02/25/2022 21:00:06 - INFO - codeparrot_training - Step 33534: {'lr': 0.0001316604742891015, 'samples': 17169920, 'steps': 33534, 'loss/train': 0.6567209362983704} 02/25/2022 21:00:09 - INFO - codeparrot_training - Step 33535: {'lr': 0.00013164606133951633, 'samples': 17170432, 'steps': 33535, 'loss/train': 2.3283274173736572} 02/25/2022 21:00:15 - INFO - codeparrot_training - Step 33536: {'lr': 0.00013163164889692198, 'samples': 17170944, 'steps': 33536, 'loss/train': 3.23056960105896} 02/25/2022 21:00:18 - INFO - codeparrot_training - Step 33537: {'lr': 0.00013161723696137996, 'samples': 17171456, 'steps': 33537, 'loss/train': 1.2566150426864624} 02/25/2022 21:00:24 - INFO - codeparrot_training - Step 33538: {'lr': 0.0001316028255329523, 'samples': 17171968, 'steps': 33538, 'loss/train': 1.3540847301483154} 02/25/2022 21:00:27 - INFO - codeparrot_training - Step 33539: {'lr': 0.00013158841461170033, 'samples': 17172480, 'steps': 33539, 'loss/train': 0.04504428058862686} 02/25/2022 21:00:33 - INFO - codeparrot_training - Step 33540: {'lr': 0.00013157400419768611, 'samples': 17172992, 'steps': 33540, 'loss/train': 2.397364377975464} 02/25/2022 21:00:36 - INFO - codeparrot_training - Step 33541: {'lr': 0.00013155959429097115, 'samples': 17173504, 'steps': 33541, 'loss/train': 0.8878464102745056} 02/25/2022 21:00:42 - INFO - codeparrot_training - Step 33542: {'lr': 0.00013154518489161743, 'samples': 17174016, 'steps': 33542, 'loss/train': 2.290039300918579} 02/25/2022 21:00:45 - INFO - codeparrot_training - Step 33543: {'lr': 0.0001315307759996864, 'samples': 17174528, 'steps': 33543, 'loss/train': 2.1497087478637695} 02/25/2022 21:00:51 - INFO - codeparrot_training - Step 33544: {'lr': 0.00013151636761523993, 'samples': 17175040, 'steps': 33544, 'loss/train': 2.8595519065856934} 02/25/2022 21:00:55 - INFO - codeparrot_training - Step 33545: {'lr': 0.00013150195973833968, 'samples': 17175552, 'steps': 33545, 'loss/train': 2.3550238609313965} 02/25/2022 21:01:01 - INFO - codeparrot_training - Step 33546: {'lr': 0.0001314875523690476, 'samples': 17176064, 'steps': 33546, 'loss/train': 2.2303035259246826} 02/25/2022 21:01:04 - INFO - codeparrot_training - Step 33547: {'lr': 0.00013147314550742502, 'samples': 17176576, 'steps': 33547, 'loss/train': 1.4370452165603638} 02/25/2022 21:01:10 - INFO - codeparrot_training - Step 33548: {'lr': 0.0001314587391535339, 'samples': 17177088, 'steps': 33548, 'loss/train': 1.084684133529663} 02/25/2022 21:01:13 - INFO - codeparrot_training - Step 33549: {'lr': 0.00013144433330743583, 'samples': 17177600, 'steps': 33549, 'loss/train': 1.811046838760376} 02/25/2022 21:01:19 - INFO - codeparrot_training - Step 33550: {'lr': 0.00013142992796919266, 'samples': 17178112, 'steps': 33550, 'loss/train': 0.8219271302223206} 02/25/2022 21:01:22 - INFO - codeparrot_training - Step 33551: {'lr': 0.00013141552313886606, 'samples': 17178624, 'steps': 33551, 'loss/train': 0.722046971321106} 02/25/2022 21:01:28 - INFO - codeparrot_training - Step 33552: {'lr': 0.00013140111881651773, 'samples': 17179136, 'steps': 33552, 'loss/train': 1.7162096500396729} 02/25/2022 21:01:31 - INFO - codeparrot_training - Step 33553: {'lr': 0.00013138671500220923, 'samples': 17179648, 'steps': 33553, 'loss/train': 1.9073817729949951} 02/25/2022 21:01:37 - INFO - codeparrot_training - Step 33554: {'lr': 0.00013137231169600244, 'samples': 17180160, 'steps': 33554, 'loss/train': 1.2484458684921265} 02/25/2022 21:01:40 - INFO - codeparrot_training - Step 33555: {'lr': 0.00013135790889795903, 'samples': 17180672, 'steps': 33555, 'loss/train': 1.7559723854064941} 02/25/2022 21:01:46 - INFO - codeparrot_training - Step 33556: {'lr': 0.00013134350660814066, 'samples': 17181184, 'steps': 33556, 'loss/train': 2.2459285259246826} 02/25/2022 21:01:50 - INFO - codeparrot_training - Step 33557: {'lr': 0.00013132910482660892, 'samples': 17181696, 'steps': 33557, 'loss/train': 2.262085437774658} 02/25/2022 21:01:55 - INFO - codeparrot_training - Step 33558: {'lr': 0.00013131470355342572, 'samples': 17182208, 'steps': 33558, 'loss/train': 1.6074315309524536} 02/25/2022 21:01:59 - INFO - codeparrot_training - Step 33559: {'lr': 0.00013130030278865268, 'samples': 17182720, 'steps': 33559, 'loss/train': 1.5473116636276245} 02/25/2022 21:02:04 - INFO - codeparrot_training - Step 33560: {'lr': 0.0001312859025323514, 'samples': 17183232, 'steps': 33560, 'loss/train': 1.608634352684021} 02/25/2022 21:02:08 - INFO - codeparrot_training - Step 33561: {'lr': 0.00013127150278458355, 'samples': 17183744, 'steps': 33561, 'loss/train': 2.049781560897827} 02/25/2022 21:02:13 - INFO - codeparrot_training - Step 33562: {'lr': 0.00013125710354541095, 'samples': 17184256, 'steps': 33562, 'loss/train': 1.3477325439453125} 02/25/2022 21:02:19 - INFO - codeparrot_training - Step 33563: {'lr': 0.00013124270481489525, 'samples': 17184768, 'steps': 33563, 'loss/train': 2.2567009925842285} 02/25/2022 21:02:22 - INFO - codeparrot_training - Step 33564: {'lr': 0.00013122830659309808, 'samples': 17185280, 'steps': 33564, 'loss/train': 1.0545082092285156} 02/25/2022 21:02:28 - INFO - codeparrot_training - Step 33565: {'lr': 0.000131213908880081, 'samples': 17185792, 'steps': 33565, 'loss/train': 1.6416910886764526} 02/25/2022 21:02:31 - INFO - codeparrot_training - Step 33566: {'lr': 0.00013119951167590594, 'samples': 17186304, 'steps': 33566, 'loss/train': 1.9051103591918945} 02/25/2022 21:02:38 - INFO - codeparrot_training - Step 33567: {'lr': 0.00013118511498063447, 'samples': 17186816, 'steps': 33567, 'loss/train': 0.4766671359539032} 02/25/2022 21:02:41 - INFO - codeparrot_training - Step 33568: {'lr': 0.00013117071879432818, 'samples': 17187328, 'steps': 33568, 'loss/train': 1.9449372291564941} 02/25/2022 21:02:47 - INFO - codeparrot_training - Step 33569: {'lr': 0.00013115632311704883, 'samples': 17187840, 'steps': 33569, 'loss/train': 1.583652138710022} 02/25/2022 21:02:50 - INFO - codeparrot_training - Step 33570: {'lr': 0.00013114192794885798, 'samples': 17188352, 'steps': 33570, 'loss/train': 1.9384287595748901} 02/25/2022 21:02:56 - INFO - codeparrot_training - Step 33571: {'lr': 0.00013112753328981748, 'samples': 17188864, 'steps': 33571, 'loss/train': 1.3240394592285156} 02/25/2022 21:02:59 - INFO - codeparrot_training - Step 33572: {'lr': 0.0001311131391399888, 'samples': 17189376, 'steps': 33572, 'loss/train': 0.7764307856559753} 02/25/2022 21:03:05 - INFO - codeparrot_training - Step 33573: {'lr': 0.00013109874549943373, 'samples': 17189888, 'steps': 33573, 'loss/train': 2.430089235305786} 02/25/2022 21:03:08 - INFO - codeparrot_training - Step 33574: {'lr': 0.0001310843523682138, 'samples': 17190400, 'steps': 33574, 'loss/train': 1.2520678043365479} 02/25/2022 21:03:14 - INFO - codeparrot_training - Step 33575: {'lr': 0.00013106995974639083, 'samples': 17190912, 'steps': 33575, 'loss/train': 2.716520071029663} 02/25/2022 21:03:17 - INFO - codeparrot_training - Step 33576: {'lr': 0.00013105556763402627, 'samples': 17191424, 'steps': 33576, 'loss/train': 1.3092294931411743} 02/25/2022 21:03:23 - INFO - codeparrot_training - Step 33577: {'lr': 0.0001310411760311821, 'samples': 17191936, 'steps': 33577, 'loss/train': 0.8175541162490845} 02/25/2022 21:03:26 - INFO - codeparrot_training - Step 33578: {'lr': 0.00013102678493791958, 'samples': 17192448, 'steps': 33578, 'loss/train': 1.7693620920181274} 02/25/2022 21:03:32 - INFO - codeparrot_training - Step 33579: {'lr': 0.0001310123943543006, 'samples': 17192960, 'steps': 33579, 'loss/train': 1.3821778297424316} 02/25/2022 21:03:35 - INFO - codeparrot_training - Step 33580: {'lr': 0.00013099800428038665, 'samples': 17193472, 'steps': 33580, 'loss/train': 2.9223263263702393} 02/25/2022 21:03:42 - INFO - codeparrot_training - Step 33581: {'lr': 0.00013098361471623968, 'samples': 17193984, 'steps': 33581, 'loss/train': 0.9644354581832886} 02/25/2022 21:03:45 - INFO - codeparrot_training - Step 33582: {'lr': 0.00013096922566192088, 'samples': 17194496, 'steps': 33582, 'loss/train': 1.4427651166915894} 02/25/2022 21:03:51 - INFO - codeparrot_training - Step 33583: {'lr': 0.00013095483711749226, 'samples': 17195008, 'steps': 33583, 'loss/train': 2.0543999671936035} 02/25/2022 21:03:54 - INFO - codeparrot_training - Step 33584: {'lr': 0.0001309404490830152, 'samples': 17195520, 'steps': 33584, 'loss/train': 0.7783902883529663} 02/25/2022 21:04:00 - INFO - codeparrot_training - Step 33585: {'lr': 0.00013092606155855154, 'samples': 17196032, 'steps': 33585, 'loss/train': 1.1725890636444092} 02/25/2022 21:04:04 - INFO - codeparrot_training - Step 33586: {'lr': 0.00013091167454416284, 'samples': 17196544, 'steps': 33586, 'loss/train': 1.6601735353469849} 02/25/2022 21:04:09 - INFO - codeparrot_training - Step 33587: {'lr': 0.00013089728803991068, 'samples': 17197056, 'steps': 33587, 'loss/train': 0.08882290869951248} 02/25/2022 21:04:13 - INFO - codeparrot_training - Step 33588: {'lr': 0.00013088290204585666, 'samples': 17197568, 'steps': 33588, 'loss/train': 1.130894660949707} 02/25/2022 21:04:18 - INFO - codeparrot_training - Step 33589: {'lr': 0.00013086851656206254, 'samples': 17198080, 'steps': 33589, 'loss/train': 1.8784723281860352} 02/25/2022 21:04:22 - INFO - codeparrot_training - Step 33590: {'lr': 0.00013085413158858987, 'samples': 17198592, 'steps': 33590, 'loss/train': 2.4832186698913574} 02/25/2022 21:04:27 - INFO - codeparrot_training - Step 33591: {'lr': 0.00013083974712550025, 'samples': 17199104, 'steps': 33591, 'loss/train': 1.8793790340423584} 02/25/2022 21:04:31 - INFO - codeparrot_training - Step 33592: {'lr': 0.00013082536317285526, 'samples': 17199616, 'steps': 33592, 'loss/train': 2.6193909645080566} 02/25/2022 21:04:37 - INFO - codeparrot_training - Step 33593: {'lr': 0.00013081097973071662, 'samples': 17200128, 'steps': 33593, 'loss/train': 1.3323309421539307} 02/25/2022 21:04:41 - INFO - codeparrot_training - Step 33594: {'lr': 0.00013079659679914594, 'samples': 17200640, 'steps': 33594, 'loss/train': 1.5587894916534424} 02/25/2022 21:04:46 - INFO - codeparrot_training - Step 33595: {'lr': 0.00013078221437820475, 'samples': 17201152, 'steps': 33595, 'loss/train': 1.6539360284805298} 02/25/2022 21:04:50 - INFO - codeparrot_training - Step 33596: {'lr': 0.00013076783246795463, 'samples': 17201664, 'steps': 33596, 'loss/train': 1.6523311138153076} 02/25/2022 21:04:55 - INFO - codeparrot_training - Step 33597: {'lr': 0.00013075345106845733, 'samples': 17202176, 'steps': 33597, 'loss/train': 1.4353982210159302} 02/25/2022 21:04:59 - INFO - codeparrot_training - Step 33598: {'lr': 0.0001307390701797743, 'samples': 17202688, 'steps': 33598, 'loss/train': 1.3447364568710327} 02/25/2022 21:05:04 - INFO - codeparrot_training - Step 33599: {'lr': 0.00013072468980196742, 'samples': 17203200, 'steps': 33599, 'loss/train': 1.6747239828109741} 02/25/2022 21:05:08 - INFO - codeparrot_training - Step 33600: {'lr': 0.00013071030993509788, 'samples': 17203712, 'steps': 33600, 'loss/train': 1.739306926727295} 02/25/2022 21:05:13 - INFO - codeparrot_training - Step 33601: {'lr': 0.0001306959305792276, 'samples': 17204224, 'steps': 33601, 'loss/train': 0.9007031917572021} 02/25/2022 21:05:17 - INFO - codeparrot_training - Step 33602: {'lr': 0.00013068155173441809, 'samples': 17204736, 'steps': 33602, 'loss/train': 1.5792921781539917} 02/25/2022 21:05:23 - INFO - codeparrot_training - Step 33603: {'lr': 0.0001306671734007309, 'samples': 17205248, 'steps': 33603, 'loss/train': 2.1220898628234863} 02/25/2022 21:05:27 - INFO - codeparrot_training - Step 33604: {'lr': 0.00013065279557822757, 'samples': 17205760, 'steps': 33604, 'loss/train': 0.8287297487258911} 02/25/2022 21:05:32 - INFO - codeparrot_training - Step 33605: {'lr': 0.00013063841826696984, 'samples': 17206272, 'steps': 33605, 'loss/train': 1.906198263168335} 02/25/2022 21:05:36 - INFO - codeparrot_training - Step 33606: {'lr': 0.00013062404146701922, 'samples': 17206784, 'steps': 33606, 'loss/train': 1.7420077323913574} 02/25/2022 21:05:41 - INFO - codeparrot_training - Step 33607: {'lr': 0.0001306096651784373, 'samples': 17207296, 'steps': 33607, 'loss/train': 1.4945111274719238} 02/25/2022 21:05:45 - INFO - codeparrot_training - Step 33608: {'lr': 0.00013059528940128563, 'samples': 17207808, 'steps': 33608, 'loss/train': 2.7044003009796143} 02/25/2022 21:05:50 - INFO - codeparrot_training - Step 33609: {'lr': 0.00013058091413562585, 'samples': 17208320, 'steps': 33609, 'loss/train': 1.260374903678894} 02/25/2022 21:05:54 - INFO - codeparrot_training - Step 33610: {'lr': 0.00013056653938151955, 'samples': 17208832, 'steps': 33610, 'loss/train': 1.229081630706787} 02/25/2022 21:05:59 - INFO - codeparrot_training - Step 33611: {'lr': 0.0001305521651390282, 'samples': 17209344, 'steps': 33611, 'loss/train': 1.8966976404190063} 02/25/2022 21:06:03 - INFO - codeparrot_training - Step 33612: {'lr': 0.00013053779140821347, 'samples': 17209856, 'steps': 33612, 'loss/train': 1.5888992547988892} 02/25/2022 21:06:09 - INFO - codeparrot_training - Step 33613: {'lr': 0.00013052341818913692, 'samples': 17210368, 'steps': 33613, 'loss/train': 1.3090122938156128} 02/25/2022 21:06:12 - INFO - codeparrot_training - Step 33614: {'lr': 0.00013050904548186015, 'samples': 17210880, 'steps': 33614, 'loss/train': 1.8943570852279663} 02/25/2022 21:06:18 - INFO - codeparrot_training - Step 33615: {'lr': 0.00013049467328644457, 'samples': 17211392, 'steps': 33615, 'loss/train': 1.418304681777954} 02/25/2022 21:06:21 - INFO - codeparrot_training - Step 33616: {'lr': 0.00013048030160295196, 'samples': 17211904, 'steps': 33616, 'loss/train': 2.0662589073181152} 02/25/2022 21:06:27 - INFO - codeparrot_training - Step 33617: {'lr': 0.00013046593043144377, 'samples': 17212416, 'steps': 33617, 'loss/train': 2.159876823425293} 02/25/2022 21:06:30 - INFO - codeparrot_training - Step 33618: {'lr': 0.0001304515597719816, 'samples': 17212928, 'steps': 33618, 'loss/train': 0.46588242053985596} 02/25/2022 21:06:36 - INFO - codeparrot_training - Step 33619: {'lr': 0.00013043718962462688, 'samples': 17213440, 'steps': 33619, 'loss/train': 0.7992615103721619} 02/25/2022 21:06:39 - INFO - codeparrot_training - Step 33620: {'lr': 0.0001304228199894415, 'samples': 17213952, 'steps': 33620, 'loss/train': 1.5776960849761963} 02/25/2022 21:06:45 - INFO - codeparrot_training - Step 33621: {'lr': 0.00013040845086648655, 'samples': 17214464, 'steps': 33621, 'loss/train': 1.5923707485198975} 02/25/2022 21:06:48 - INFO - codeparrot_training - Step 33622: {'lr': 0.00013039408225582394, 'samples': 17214976, 'steps': 33622, 'loss/train': 2.7310385704040527} 02/25/2022 21:06:54 - INFO - codeparrot_training - Step 33623: {'lr': 0.000130379714157515, 'samples': 17215488, 'steps': 33623, 'loss/train': 0.5016627311706543} 02/25/2022 21:06:57 - INFO - codeparrot_training - Step 33624: {'lr': 0.00013036534657162146, 'samples': 17216000, 'steps': 33624, 'loss/train': 4.217424392700195} 02/25/2022 21:07:03 - INFO - codeparrot_training - Step 33625: {'lr': 0.00013035097949820485, 'samples': 17216512, 'steps': 33625, 'loss/train': 2.36342191696167} 02/25/2022 21:07:06 - INFO - codeparrot_training - Step 33626: {'lr': 0.00013033661293732657, 'samples': 17217024, 'steps': 33626, 'loss/train': 2.4237160682678223} 02/25/2022 21:07:12 - INFO - codeparrot_training - Step 33627: {'lr': 0.0001303222468890482, 'samples': 17217536, 'steps': 33627, 'loss/train': 2.6913466453552246} 02/25/2022 21:07:15 - INFO - codeparrot_training - Step 33628: {'lr': 0.00013030788135343142, 'samples': 17218048, 'steps': 33628, 'loss/train': 1.904963493347168} 02/25/2022 21:07:22 - INFO - codeparrot_training - Step 33629: {'lr': 0.00013029351633053765, 'samples': 17218560, 'steps': 33629, 'loss/train': 1.7030514478683472} 02/25/2022 21:07:25 - INFO - codeparrot_training - Step 33630: {'lr': 0.00013027915182042843, 'samples': 17219072, 'steps': 33630, 'loss/train': 2.454561710357666} 02/25/2022 21:07:31 - INFO - codeparrot_training - Step 33631: {'lr': 0.00013026478782316526, 'samples': 17219584, 'steps': 33631, 'loss/train': 2.4349617958068848} 02/25/2022 21:07:34 - INFO - codeparrot_training - Step 33632: {'lr': 0.00013025042433880977, 'samples': 17220096, 'steps': 33632, 'loss/train': 1.519382119178772} 02/25/2022 21:07:40 - INFO - codeparrot_training - Step 33633: {'lr': 0.00013023606136742345, 'samples': 17220608, 'steps': 33633, 'loss/train': 2.088820457458496} 02/25/2022 21:07:43 - INFO - codeparrot_training - Step 33634: {'lr': 0.00013022169890906783, 'samples': 17221120, 'steps': 33634, 'loss/train': 2.268474817276001} 02/25/2022 21:07:49 - INFO - codeparrot_training - Step 33635: {'lr': 0.00013020733696380427, 'samples': 17221632, 'steps': 33635, 'loss/train': 0.02734479494392872} 02/25/2022 21:07:52 - INFO - codeparrot_training - Step 33636: {'lr': 0.00013019297553169457, 'samples': 17222144, 'steps': 33636, 'loss/train': 2.41414475440979} 02/25/2022 21:07:58 - INFO - codeparrot_training - Step 33637: {'lr': 0.00013017861461280005, 'samples': 17222656, 'steps': 33637, 'loss/train': 1.1709057092666626} 02/25/2022 21:08:01 - INFO - codeparrot_training - Step 33638: {'lr': 0.00013016425420718248, 'samples': 17223168, 'steps': 33638, 'loss/train': 0.536399781703949} 02/25/2022 21:08:08 - INFO - codeparrot_training - Step 33639: {'lr': 0.00013014989431490298, 'samples': 17223680, 'steps': 33639, 'loss/train': 1.8530877828598022} 02/25/2022 21:08:11 - INFO - codeparrot_training - Step 33640: {'lr': 0.00013013553493602337, 'samples': 17224192, 'steps': 33640, 'loss/train': 1.8192707300186157} 02/25/2022 21:08:17 - INFO - codeparrot_training - Step 33641: {'lr': 0.00013012117607060498, 'samples': 17224704, 'steps': 33641, 'loss/train': 2.9733150005340576} 02/25/2022 21:08:20 - INFO - codeparrot_training - Step 33642: {'lr': 0.0001301068177187096, 'samples': 17225216, 'steps': 33642, 'loss/train': 2.9503421783447266} 02/25/2022 21:08:26 - INFO - codeparrot_training - Step 33643: {'lr': 0.00013009245988039836, 'samples': 17225728, 'steps': 33643, 'loss/train': 0.7980608940124512} 02/25/2022 21:08:29 - INFO - codeparrot_training - Step 33644: {'lr': 0.00013007810255573303, 'samples': 17226240, 'steps': 33644, 'loss/train': 1.8964685201644897} 02/25/2022 21:08:35 - INFO - codeparrot_training - Step 33645: {'lr': 0.0001300637457447749, 'samples': 17226752, 'steps': 33645, 'loss/train': 2.0529398918151855} 02/25/2022 21:08:38 - INFO - codeparrot_training - Step 33646: {'lr': 0.00013004938944758575, 'samples': 17227264, 'steps': 33646, 'loss/train': 1.1015039682388306} 02/25/2022 21:08:44 - INFO - codeparrot_training - Step 33647: {'lr': 0.00013003503366422692, 'samples': 17227776, 'steps': 33647, 'loss/train': 1.6781864166259766} 02/25/2022 21:08:47 - INFO - codeparrot_training - Step 33648: {'lr': 0.0001300206783947599, 'samples': 17228288, 'steps': 33648, 'loss/train': 0.8616465926170349} 02/25/2022 21:08:53 - INFO - codeparrot_training - Step 33649: {'lr': 0.00013000632363924618, 'samples': 17228800, 'steps': 33649, 'loss/train': 0.7489467859268188} 02/25/2022 21:08:57 - INFO - codeparrot_training - Step 33650: {'lr': 0.00012999196939774722, 'samples': 17229312, 'steps': 33650, 'loss/train': 0.5407328605651855} 02/25/2022 21:09:02 - INFO - codeparrot_training - Step 33651: {'lr': 0.0001299776156703246, 'samples': 17229824, 'steps': 33651, 'loss/train': 1.9846183061599731} 02/25/2022 21:09:06 - INFO - codeparrot_training - Step 33652: {'lr': 0.00012996326245703977, 'samples': 17230336, 'steps': 33652, 'loss/train': 1.2445614337921143} 02/25/2022 21:09:11 - INFO - codeparrot_training - Step 33653: {'lr': 0.0001299489097579542, 'samples': 17230848, 'steps': 33653, 'loss/train': 1.4827134609222412} 02/25/2022 21:09:15 - INFO - codeparrot_training - Step 33654: {'lr': 0.00012993455757312933, 'samples': 17231360, 'steps': 33654, 'loss/train': 2.2110941410064697} 02/25/2022 21:09:20 - INFO - codeparrot_training - Step 33655: {'lr': 0.00012992020590262677, 'samples': 17231872, 'steps': 33655, 'loss/train': 1.152616262435913} 02/25/2022 21:09:26 - INFO - codeparrot_training - Step 33656: {'lr': 0.0001299058547465079, 'samples': 17232384, 'steps': 33656, 'loss/train': 1.5722112655639648} 02/25/2022 21:09:29 - INFO - codeparrot_training - Step 33657: {'lr': 0.00012989150410483422, 'samples': 17232896, 'steps': 33657, 'loss/train': 2.3757901191711426} 02/25/2022 21:09:35 - INFO - codeparrot_training - Step 33658: {'lr': 0.00012987715397766707, 'samples': 17233408, 'steps': 33658, 'loss/train': 1.5098944902420044} 02/25/2022 21:09:38 - INFO - codeparrot_training - Step 33659: {'lr': 0.0001298628043650682, 'samples': 17233920, 'steps': 33659, 'loss/train': 0.9366226196289062} 02/25/2022 21:09:44 - INFO - codeparrot_training - Step 33660: {'lr': 0.00012984845526709893, 'samples': 17234432, 'steps': 33660, 'loss/train': 1.382674217224121} 02/25/2022 21:09:48 - INFO - codeparrot_training - Step 33661: {'lr': 0.00012983410668382066, 'samples': 17234944, 'steps': 33661, 'loss/train': 2.12652325630188} 02/25/2022 21:09:53 - INFO - codeparrot_training - Step 33662: {'lr': 0.0001298197586152949, 'samples': 17235456, 'steps': 33662, 'loss/train': 1.0333904027938843} 02/25/2022 21:09:57 - INFO - codeparrot_training - Step 33663: {'lr': 0.0001298054110615832, 'samples': 17235968, 'steps': 33663, 'loss/train': 2.891796350479126} 02/25/2022 21:10:02 - INFO - codeparrot_training - Step 33664: {'lr': 0.00012979106402274693, 'samples': 17236480, 'steps': 33664, 'loss/train': 2.807631015777588} 02/25/2022 21:10:06 - INFO - codeparrot_training - Step 33665: {'lr': 0.00012977671749884762, 'samples': 17236992, 'steps': 33665, 'loss/train': 1.7277345657348633} 02/25/2022 21:10:11 - INFO - codeparrot_training - Step 33666: {'lr': 0.0001297623714899465, 'samples': 17237504, 'steps': 33666, 'loss/train': 1.4725676774978638} 02/25/2022 21:10:15 - INFO - codeparrot_training - Step 33667: {'lr': 0.0001297480259961054, 'samples': 17238016, 'steps': 33667, 'loss/train': 1.6230171918869019} 02/25/2022 21:10:20 - INFO - codeparrot_training - Step 33668: {'lr': 0.0001297336810173855, 'samples': 17238528, 'steps': 33668, 'loss/train': 1.5123738050460815} 02/25/2022 21:10:24 - INFO - codeparrot_training - Step 33669: {'lr': 0.00012971933655384836, 'samples': 17239040, 'steps': 33669, 'loss/train': 1.4499177932739258} 02/25/2022 21:10:29 - INFO - codeparrot_training - Step 33670: {'lr': 0.00012970499260555525, 'samples': 17239552, 'steps': 33670, 'loss/train': 1.8074325323104858} 02/25/2022 21:10:33 - INFO - codeparrot_training - Step 33671: {'lr': 0.0001296906491725679, 'samples': 17240064, 'steps': 33671, 'loss/train': 2.260063409805298} 02/25/2022 21:10:38 - INFO - codeparrot_training - Step 33672: {'lr': 0.00012967630625494746, 'samples': 17240576, 'steps': 33672, 'loss/train': 1.3195381164550781} 02/25/2022 21:10:42 - INFO - codeparrot_training - Step 33673: {'lr': 0.00012966196385275574, 'samples': 17241088, 'steps': 33673, 'loss/train': 1.5441840887069702} 02/25/2022 21:10:48 - INFO - codeparrot_training - Step 33674: {'lr': 0.00012964762196605376, 'samples': 17241600, 'steps': 33674, 'loss/train': 1.3310438394546509} 02/25/2022 21:10:52 - INFO - codeparrot_training - Step 33675: {'lr': 0.00012963328059490325, 'samples': 17242112, 'steps': 33675, 'loss/train': 1.2927652597427368} 02/25/2022 21:10:57 - INFO - codeparrot_training - Step 33676: {'lr': 0.00012961893973936544, 'samples': 17242624, 'steps': 33676, 'loss/train': 1.964985728263855} 02/25/2022 21:11:01 - INFO - codeparrot_training - Step 33677: {'lr': 0.00012960459939950205, 'samples': 17243136, 'steps': 33677, 'loss/train': 0.8210015296936035} 02/25/2022 21:11:07 - INFO - codeparrot_training - Step 33678: {'lr': 0.00012959025957537415, 'samples': 17243648, 'steps': 33678, 'loss/train': 1.57382333278656} 02/25/2022 21:11:10 - INFO - codeparrot_training - Step 33679: {'lr': 0.00012957592026704344, 'samples': 17244160, 'steps': 33679, 'loss/train': 1.5136228799819946} 02/25/2022 21:11:15 - INFO - codeparrot_training - Step 33680: {'lr': 0.00012956158147457115, 'samples': 17244672, 'steps': 33680, 'loss/train': 1.345003604888916} 02/25/2022 21:11:19 - INFO - codeparrot_training - Step 33681: {'lr': 0.00012954724319801897, 'samples': 17245184, 'steps': 33681, 'loss/train': 2.4613077640533447} 02/25/2022 21:11:24 - INFO - codeparrot_training - Step 33682: {'lr': 0.000129532905437448, 'samples': 17245696, 'steps': 33682, 'loss/train': 1.4208694696426392} 02/25/2022 21:11:28 - INFO - codeparrot_training - Step 33683: {'lr': 0.0001295185681929199, 'samples': 17246208, 'steps': 33683, 'loss/train': 1.334010362625122} 02/25/2022 21:11:34 - INFO - codeparrot_training - Step 33684: {'lr': 0.00012950423146449586, 'samples': 17246720, 'steps': 33684, 'loss/train': 2.0325515270233154} 02/25/2022 21:11:38 - INFO - codeparrot_training - Step 33685: {'lr': 0.0001294898952522376, 'samples': 17247232, 'steps': 33685, 'loss/train': 0.7957848906517029} 02/25/2022 21:11:43 - INFO - codeparrot_training - Step 33686: {'lr': 0.00012947555955620626, 'samples': 17247744, 'steps': 33686, 'loss/train': 2.5259854793548584} 02/25/2022 21:11:47 - INFO - codeparrot_training - Step 33687: {'lr': 0.00012946122437646347, 'samples': 17248256, 'steps': 33687, 'loss/train': 2.0723929405212402} 02/25/2022 21:11:52 - INFO - codeparrot_training - Step 33688: {'lr': 0.00012944688971307033, 'samples': 17248768, 'steps': 33688, 'loss/train': 1.531112790107727} 02/25/2022 21:11:56 - INFO - codeparrot_training - Step 33689: {'lr': 0.00012943255556608857, 'samples': 17249280, 'steps': 33689, 'loss/train': 1.1501189470291138} 02/25/2022 21:12:01 - INFO - codeparrot_training - Step 33690: {'lr': 0.00012941822193557944, 'samples': 17249792, 'steps': 33690, 'loss/train': 2.2206125259399414} 02/25/2022 21:12:04 - INFO - codeparrot_training - Step 33691: {'lr': 0.00012940388882160436, 'samples': 17250304, 'steps': 33691, 'loss/train': 1.06920325756073} 02/25/2022 21:12:10 - INFO - codeparrot_training - Step 33692: {'lr': 0.00012938955622422466, 'samples': 17250816, 'steps': 33692, 'loss/train': 1.8763850927352905} 02/25/2022 21:12:13 - INFO - codeparrot_training - Step 33693: {'lr': 0.0001293752241435019, 'samples': 17251328, 'steps': 33693, 'loss/train': 2.325950860977173} 02/25/2022 21:12:19 - INFO - codeparrot_training - Step 33694: {'lr': 0.00012936089257949734, 'samples': 17251840, 'steps': 33694, 'loss/train': 2.7652454376220703} 02/25/2022 21:12:23 - INFO - codeparrot_training - Step 33695: {'lr': 0.0001293465615322724, 'samples': 17252352, 'steps': 33695, 'loss/train': 1.1024342775344849} 02/25/2022 21:12:29 - INFO - codeparrot_training - Step 33696: {'lr': 0.00012933223100188842, 'samples': 17252864, 'steps': 33696, 'loss/train': 1.908313512802124} 02/25/2022 21:12:32 - INFO - codeparrot_training - Step 33697: {'lr': 0.00012931790098840695, 'samples': 17253376, 'steps': 33697, 'loss/train': 0.09740322083234787} 02/25/2022 21:12:38 - INFO - codeparrot_training - Step 33698: {'lr': 0.00012930357149188926, 'samples': 17253888, 'steps': 33698, 'loss/train': 1.3092392683029175} 02/25/2022 21:12:41 - INFO - codeparrot_training - Step 33699: {'lr': 0.0001292892425123967, 'samples': 17254400, 'steps': 33699, 'loss/train': 0.8420547842979431} 02/25/2022 21:12:47 - INFO - codeparrot_training - Step 33700: {'lr': 0.00012927491404999075, 'samples': 17254912, 'steps': 33700, 'loss/train': 0.7378376126289368} 02/25/2022 21:12:50 - INFO - codeparrot_training - Step 33701: {'lr': 0.00012926058610473264, 'samples': 17255424, 'steps': 33701, 'loss/train': 0.9803818464279175} 02/25/2022 21:12:56 - INFO - codeparrot_training - Step 33702: {'lr': 0.00012924625867668388, 'samples': 17255936, 'steps': 33702, 'loss/train': 2.5155279636383057} 02/25/2022 21:12:59 - INFO - codeparrot_training - Step 33703: {'lr': 0.00012923193176590586, 'samples': 17256448, 'steps': 33703, 'loss/train': 1.3950456380844116} 02/25/2022 21:13:05 - INFO - codeparrot_training - Step 33704: {'lr': 0.00012921760537245986, 'samples': 17256960, 'steps': 33704, 'loss/train': 1.1246124505996704} 02/25/2022 21:13:08 - INFO - codeparrot_training - Step 33705: {'lr': 0.0001292032794964072, 'samples': 17257472, 'steps': 33705, 'loss/train': 1.4105521440505981} 02/25/2022 21:13:15 - INFO - codeparrot_training - Step 33706: {'lr': 0.00012918895413780945, 'samples': 17257984, 'steps': 33706, 'loss/train': 2.2826855182647705} 02/25/2022 21:13:18 - INFO - codeparrot_training - Step 33707: {'lr': 0.00012917462929672773, 'samples': 17258496, 'steps': 33707, 'loss/train': 2.267845869064331} 02/25/2022 21:13:24 - INFO - codeparrot_training - Step 33708: {'lr': 0.00012916030497322375, 'samples': 17259008, 'steps': 33708, 'loss/train': 2.367537260055542} 02/25/2022 21:13:27 - INFO - codeparrot_training - Step 33709: {'lr': 0.00012914598116735846, 'samples': 17259520, 'steps': 33709, 'loss/train': 2.224104881286621} 02/25/2022 21:13:33 - INFO - codeparrot_training - Step 33710: {'lr': 0.0001291316578791935, 'samples': 17260032, 'steps': 33710, 'loss/train': 1.4223227500915527} 02/25/2022 21:13:36 - INFO - codeparrot_training - Step 33711: {'lr': 0.00012911733510879005, 'samples': 17260544, 'steps': 33711, 'loss/train': 0.31277206540107727} 02/25/2022 21:13:42 - INFO - codeparrot_training - Step 33712: {'lr': 0.00012910301285620974, 'samples': 17261056, 'steps': 33712, 'loss/train': 1.3715959787368774} 02/25/2022 21:13:45 - INFO - codeparrot_training - Step 33713: {'lr': 0.0001290886911215135, 'samples': 17261568, 'steps': 33713, 'loss/train': 1.9319833517074585} 02/25/2022 21:13:51 - INFO - codeparrot_training - Step 33714: {'lr': 0.00012907436990476306, 'samples': 17262080, 'steps': 33714, 'loss/train': 2.2283267974853516} 02/25/2022 21:13:54 - INFO - codeparrot_training - Step 33715: {'lr': 0.00012906004920601948, 'samples': 17262592, 'steps': 33715, 'loss/train': 0.7467585206031799} 02/25/2022 21:14:00 - INFO - codeparrot_training - Step 33716: {'lr': 0.0001290457290253445, 'samples': 17263104, 'steps': 33716, 'loss/train': 1.4538532495498657} 02/25/2022 21:14:04 - INFO - codeparrot_training - Step 33717: {'lr': 0.00012903140936279897, 'samples': 17263616, 'steps': 33717, 'loss/train': 1.1254184246063232} 02/25/2022 21:14:07 - INFO - codeparrot_training - Step 33718: {'lr': 0.0001290170902184446, 'samples': 17264128, 'steps': 33718, 'loss/train': 2.270923614501953} 02/25/2022 21:14:13 - INFO - codeparrot_training - Step 33719: {'lr': 0.00012900277159234248, 'samples': 17264640, 'steps': 33719, 'loss/train': 3.5176162719726562} 02/25/2022 21:14:16 - INFO - codeparrot_training - Step 33720: {'lr': 0.00012898845348455418, 'samples': 17265152, 'steps': 33720, 'loss/train': 2.921058416366577} 02/25/2022 21:14:22 - INFO - codeparrot_training - Step 33721: {'lr': 0.00012897413589514089, 'samples': 17265664, 'steps': 33721, 'loss/train': 0.6855606436729431} 02/25/2022 21:14:26 - INFO - codeparrot_training - Step 33722: {'lr': 0.000128959818824164, 'samples': 17266176, 'steps': 33722, 'loss/train': 1.8103679418563843} 02/25/2022 21:14:31 - INFO - codeparrot_training - Step 33723: {'lr': 0.00012894550227168469, 'samples': 17266688, 'steps': 33723, 'loss/train': 2.340031623840332} 02/25/2022 21:14:37 - INFO - codeparrot_training - Step 33724: {'lr': 0.0001289311862377645, 'samples': 17267200, 'steps': 33724, 'loss/train': 1.982516884803772} 02/25/2022 21:14:40 - INFO - codeparrot_training - Step 33725: {'lr': 0.00012891687072246472, 'samples': 17267712, 'steps': 33725, 'loss/train': 1.9022910594940186} 02/25/2022 21:14:44 - INFO - codeparrot_training - Step 33726: {'lr': 0.00012890255572584657, 'samples': 17268224, 'steps': 33726, 'loss/train': 0.9450118541717529} 02/25/2022 21:14:49 - INFO - codeparrot_training - Step 33727: {'lr': 0.00012888824124797132, 'samples': 17268736, 'steps': 33727, 'loss/train': 1.738682508468628} 02/25/2022 21:14:55 - INFO - codeparrot_training - Step 33728: {'lr': 0.00012887392728890053, 'samples': 17269248, 'steps': 33728, 'loss/train': 1.5189931392669678} 02/25/2022 21:14:59 - INFO - codeparrot_training - Step 33729: {'lr': 0.0001288596138486953, 'samples': 17269760, 'steps': 33729, 'loss/train': 7.502089023590088} 02/25/2022 21:15:02 - INFO - codeparrot_training - Step 33730: {'lr': 0.0001288453009274171, 'samples': 17270272, 'steps': 33730, 'loss/train': 1.7766786813735962} 02/25/2022 21:15:08 - INFO - codeparrot_training - Step 33731: {'lr': 0.00012883098852512701, 'samples': 17270784, 'steps': 33731, 'loss/train': 2.1426002979278564} 02/25/2022 21:15:12 - INFO - codeparrot_training - Step 33732: {'lr': 0.00012881667664188659, 'samples': 17271296, 'steps': 33732, 'loss/train': 1.8388675451278687} 02/25/2022 21:15:17 - INFO - codeparrot_training - Step 33733: {'lr': 0.00012880236527775706, 'samples': 17271808, 'steps': 33733, 'loss/train': 2.0290133953094482} 02/25/2022 21:15:21 - INFO - codeparrot_training - Step 33734: {'lr': 0.00012878805443279973, 'samples': 17272320, 'steps': 33734, 'loss/train': 0.8609355092048645} 02/25/2022 21:15:26 - INFO - codeparrot_training - Step 33735: {'lr': 0.00012877374410707576, 'samples': 17272832, 'steps': 33735, 'loss/train': 1.5654685497283936} 02/25/2022 21:15:32 - INFO - codeparrot_training - Step 33736: {'lr': 0.00012875943430064668, 'samples': 17273344, 'steps': 33736, 'loss/train': 1.158992052078247} 02/25/2022 21:15:35 - INFO - codeparrot_training - Step 33737: {'lr': 0.00012874512501357367, 'samples': 17273856, 'steps': 33737, 'loss/train': 0.7392899990081787} 02/25/2022 21:15:41 - INFO - codeparrot_training - Step 33738: {'lr': 0.00012873081624591807, 'samples': 17274368, 'steps': 33738, 'loss/train': 1.5286954641342163} 02/25/2022 21:15:44 - INFO - codeparrot_training - Step 33739: {'lr': 0.00012871650799774103, 'samples': 17274880, 'steps': 33739, 'loss/train': 2.1024341583251953} 02/25/2022 21:15:49 - INFO - codeparrot_training - Step 33740: {'lr': 0.00012870220026910405, 'samples': 17275392, 'steps': 33740, 'loss/train': 1.8214421272277832} 02/25/2022 21:15:53 - INFO - codeparrot_training - Step 33741: {'lr': 0.00012868789306006833, 'samples': 17275904, 'steps': 33741, 'loss/train': 1.8990613222122192} 02/25/2022 21:15:59 - INFO - codeparrot_training - Step 33742: {'lr': 0.0001286735863706951, 'samples': 17276416, 'steps': 33742, 'loss/train': 0.8461995124816895} 02/25/2022 21:16:02 - INFO - codeparrot_training - Step 33743: {'lr': 0.00012865928020104576, 'samples': 17276928, 'steps': 33743, 'loss/train': 1.3933275938034058} 02/25/2022 21:16:08 - INFO - codeparrot_training - Step 33744: {'lr': 0.0001286449745511815, 'samples': 17277440, 'steps': 33744, 'loss/train': 0.628835141658783} 02/25/2022 21:16:12 - INFO - codeparrot_training - Step 33745: {'lr': 0.0001286306694211637, 'samples': 17277952, 'steps': 33745, 'loss/train': 1.5539894104003906} 02/25/2022 21:16:17 - INFO - codeparrot_training - Step 33746: {'lr': 0.00012861636481105343, 'samples': 17278464, 'steps': 33746, 'loss/train': 1.9513862133026123} 02/25/2022 21:16:21 - INFO - codeparrot_training - Step 33747: {'lr': 0.00012860206072091236, 'samples': 17278976, 'steps': 33747, 'loss/train': 2.8285558223724365} 02/25/2022 21:16:26 - INFO - codeparrot_training - Step 33748: {'lr': 0.00012858775715080125, 'samples': 17279488, 'steps': 33748, 'loss/train': 1.750298261642456} 02/25/2022 21:16:29 - INFO - codeparrot_training - Step 33749: {'lr': 0.0001285734541007818, 'samples': 17280000, 'steps': 33749, 'loss/train': 2.391174554824829} 02/25/2022 21:16:35 - INFO - codeparrot_training - Step 33750: {'lr': 0.00012855915157091496, 'samples': 17280512, 'steps': 33750, 'loss/train': 2.1921160221099854} 02/25/2022 21:16:38 - INFO - codeparrot_training - Step 33751: {'lr': 0.0001285448495612624, 'samples': 17281024, 'steps': 33751, 'loss/train': 1.680588722229004} 02/25/2022 21:16:45 - INFO - codeparrot_training - Step 33752: {'lr': 0.00012853054807188488, 'samples': 17281536, 'steps': 33752, 'loss/train': 1.7051770687103271} 02/25/2022 21:16:48 - INFO - codeparrot_training - Step 33753: {'lr': 0.00012851624710284403, 'samples': 17282048, 'steps': 33753, 'loss/train': 1.9384405612945557} 02/25/2022 21:16:54 - INFO - codeparrot_training - Step 33754: {'lr': 0.00012850194665420096, 'samples': 17282560, 'steps': 33754, 'loss/train': 1.533313512802124} 02/25/2022 21:16:57 - INFO - codeparrot_training - Step 33755: {'lr': 0.00012848764672601705, 'samples': 17283072, 'steps': 33755, 'loss/train': 2.5330677032470703} 02/25/2022 21:17:03 - INFO - codeparrot_training - Step 33756: {'lr': 0.00012847334731835345, 'samples': 17283584, 'steps': 33756, 'loss/train': 1.350999355316162} 02/25/2022 21:17:06 - INFO - codeparrot_training - Step 33757: {'lr': 0.00012845904843127143, 'samples': 17284096, 'steps': 33757, 'loss/train': 1.5097826719284058} 02/25/2022 21:17:12 - INFO - codeparrot_training - Step 33758: {'lr': 0.0001284447500648322, 'samples': 17284608, 'steps': 33758, 'loss/train': 0.9145926237106323} 02/25/2022 21:17:15 - INFO - codeparrot_training - Step 33759: {'lr': 0.00012843045221909715, 'samples': 17285120, 'steps': 33759, 'loss/train': 1.5351028442382812} 02/25/2022 21:17:21 - INFO - codeparrot_training - Step 33760: {'lr': 0.0001284161548941274, 'samples': 17285632, 'steps': 33760, 'loss/train': 2.0386199951171875} 02/25/2022 21:17:24 - INFO - codeparrot_training - Step 33761: {'lr': 0.0001284018580899843, 'samples': 17286144, 'steps': 33761, 'loss/train': 1.9935725927352905} 02/25/2022 21:17:30 - INFO - codeparrot_training - Step 33762: {'lr': 0.00012838756180672887, 'samples': 17286656, 'steps': 33762, 'loss/train': 2.113556146621704} 02/25/2022 21:17:34 - INFO - codeparrot_training - Step 33763: {'lr': 0.00012837326604442262, 'samples': 17287168, 'steps': 33763, 'loss/train': 0.44728103280067444} 02/25/2022 21:17:39 - INFO - codeparrot_training - Step 33764: {'lr': 0.00012835897080312668, 'samples': 17287680, 'steps': 33764, 'loss/train': 1.968787670135498} 02/25/2022 21:17:43 - INFO - codeparrot_training - Step 33765: {'lr': 0.0001283446760829023, 'samples': 17288192, 'steps': 33765, 'loss/train': 2.889301061630249} 02/25/2022 21:17:48 - INFO - codeparrot_training - Step 33766: {'lr': 0.0001283303818838106, 'samples': 17288704, 'steps': 33766, 'loss/train': 2.0273549556732178} 02/25/2022 21:17:52 - INFO - codeparrot_training - Step 33767: {'lr': 0.00012831608820591301, 'samples': 17289216, 'steps': 33767, 'loss/train': 1.0190807580947876} 02/25/2022 21:17:57 - INFO - codeparrot_training - Step 33768: {'lr': 0.00012830179504927052, 'samples': 17289728, 'steps': 33768, 'loss/train': 2.6717264652252197} 02/25/2022 21:18:01 - INFO - codeparrot_training - Step 33769: {'lr': 0.00012828750241394474, 'samples': 17290240, 'steps': 33769, 'loss/train': 2.1391308307647705} 02/25/2022 21:18:06 - INFO - codeparrot_training - Step 33770: {'lr': 0.00012827321029999645, 'samples': 17290752, 'steps': 33770, 'loss/train': 1.4765336513519287} 02/25/2022 21:18:10 - INFO - codeparrot_training - Step 33771: {'lr': 0.00012825891870748716, 'samples': 17291264, 'steps': 33771, 'loss/train': 1.2519994974136353} 02/25/2022 21:18:15 - INFO - codeparrot_training - Step 33772: {'lr': 0.0001282446276364779, 'samples': 17291776, 'steps': 33772, 'loss/train': 1.9881829023361206} 02/25/2022 21:18:19 - INFO - codeparrot_training - Step 33773: {'lr': 0.0001282303370870302, 'samples': 17292288, 'steps': 33773, 'loss/train': 1.3524502515792847} 02/25/2022 21:18:24 - INFO - codeparrot_training - Step 33774: {'lr': 0.00012821604705920487, 'samples': 17292800, 'steps': 33774, 'loss/train': 1.6359270811080933} 02/25/2022 21:18:28 - INFO - codeparrot_training - Step 33775: {'lr': 0.0001282017575530634, 'samples': 17293312, 'steps': 33775, 'loss/train': 1.7404199838638306} 02/25/2022 21:18:33 - INFO - codeparrot_training - Step 33776: {'lr': 0.00012818746856866687, 'samples': 17293824, 'steps': 33776, 'loss/train': 1.450870156288147} 02/25/2022 21:18:37 - INFO - codeparrot_training - Step 33777: {'lr': 0.0001281731801060767, 'samples': 17294336, 'steps': 33777, 'loss/train': 2.3660786151885986} 02/25/2022 21:18:43 - INFO - codeparrot_training - Step 33778: {'lr': 0.0001281588921653538, 'samples': 17294848, 'steps': 33778, 'loss/train': 2.1349151134490967} 02/25/2022 21:18:48 - INFO - codeparrot_training - Step 33779: {'lr': 0.0001281446047465596, 'samples': 17295360, 'steps': 33779, 'loss/train': 1.1453312635421753} 02/25/2022 21:18:52 - INFO - codeparrot_training - Step 33780: {'lr': 0.00012813031784975518, 'samples': 17295872, 'steps': 33780, 'loss/train': 2.256417751312256} 02/25/2022 21:18:57 - INFO - codeparrot_training - Step 33781: {'lr': 0.0001281160314750017, 'samples': 17296384, 'steps': 33781, 'loss/train': 2.6236774921417236} 02/25/2022 21:19:01 - INFO - codeparrot_training - Step 33782: {'lr': 0.00012810174562236055, 'samples': 17296896, 'steps': 33782, 'loss/train': 1.2795182466506958} 02/25/2022 21:19:04 - INFO - codeparrot_training - Step 33783: {'lr': 0.00012808746029189277, 'samples': 17297408, 'steps': 33783, 'loss/train': 0.646183967590332} 02/25/2022 21:19:10 - INFO - codeparrot_training - Step 33784: {'lr': 0.00012807317548365965, 'samples': 17297920, 'steps': 33784, 'loss/train': 1.2222410440444946} 02/25/2022 21:19:13 - INFO - codeparrot_training - Step 33785: {'lr': 0.00012805889119772224, 'samples': 17298432, 'steps': 33785, 'loss/train': 1.362569808959961} 02/25/2022 21:19:19 - INFO - codeparrot_training - Step 33786: {'lr': 0.00012804460743414187, 'samples': 17298944, 'steps': 33786, 'loss/train': 2.362821102142334} 02/25/2022 21:19:22 - INFO - codeparrot_training - Step 33787: {'lr': 0.00012803032419297973, 'samples': 17299456, 'steps': 33787, 'loss/train': 2.095163345336914} 02/25/2022 21:19:29 - INFO - codeparrot_training - Step 33788: {'lr': 0.0001280160414742969, 'samples': 17299968, 'steps': 33788, 'loss/train': 0.22723883390426636} 02/25/2022 21:19:32 - INFO - codeparrot_training - Step 33789: {'lr': 0.00012800175927815455, 'samples': 17300480, 'steps': 33789, 'loss/train': 2.116652727127075} 02/25/2022 21:19:38 - INFO - codeparrot_training - Step 33790: {'lr': 0.00012798747760461399, 'samples': 17300992, 'steps': 33790, 'loss/train': 1.2462120056152344} 02/25/2022 21:19:41 - INFO - codeparrot_training - Step 33791: {'lr': 0.00012797319645373634, 'samples': 17301504, 'steps': 33791, 'loss/train': 1.3809499740600586} 02/25/2022 21:19:47 - INFO - codeparrot_training - Step 33792: {'lr': 0.0001279589158255828, 'samples': 17302016, 'steps': 33792, 'loss/train': 1.826533317565918} 02/25/2022 21:19:52 - INFO - codeparrot_training - Step 33793: {'lr': 0.00012794463572021438, 'samples': 17302528, 'steps': 33793, 'loss/train': 1.3947933912277222} 02/25/2022 21:19:56 - INFO - codeparrot_training - Step 33794: {'lr': 0.0001279303561376925, 'samples': 17303040, 'steps': 33794, 'loss/train': 4.1246232986450195} 02/25/2022 21:20:01 - INFO - codeparrot_training - Step 33795: {'lr': 0.00012791607707807824, 'samples': 17303552, 'steps': 33795, 'loss/train': 1.8450731039047241} 02/25/2022 21:20:05 - INFO - codeparrot_training - Step 33796: {'lr': 0.0001279017985414327, 'samples': 17304064, 'steps': 33796, 'loss/train': 1.660609245300293} 02/25/2022 21:20:08 - INFO - codeparrot_training - Step 33797: {'lr': 0.00012788752052781698, 'samples': 17304576, 'steps': 33797, 'loss/train': 8.734773635864258} 02/25/2022 21:20:15 - INFO - codeparrot_training - Step 33798: {'lr': 0.00012787324303729247, 'samples': 17305088, 'steps': 33798, 'loss/train': 0.7384844422340393} 02/25/2022 21:20:18 - INFO - codeparrot_training - Step 33799: {'lr': 0.0001278589660699202, 'samples': 17305600, 'steps': 33799, 'loss/train': 2.1188859939575195} 02/25/2022 21:20:23 - INFO - codeparrot_training - Step 33800: {'lr': 0.00012784468962576134, 'samples': 17306112, 'steps': 33800, 'loss/train': 1.429151177406311} 02/25/2022 21:20:29 - INFO - codeparrot_training - Step 33801: {'lr': 0.00012783041370487692, 'samples': 17306624, 'steps': 33801, 'loss/train': 2.3084049224853516} 02/25/2022 21:20:33 - INFO - codeparrot_training - Step 33802: {'lr': 0.00012781613830732834, 'samples': 17307136, 'steps': 33802, 'loss/train': 3.241088390350342} 02/25/2022 21:20:38 - INFO - codeparrot_training - Step 33803: {'lr': 0.00012780186343317652, 'samples': 17307648, 'steps': 33803, 'loss/train': 1.4482510089874268} 02/25/2022 21:20:42 - INFO - codeparrot_training - Step 33804: {'lr': 0.00012778758908248288, 'samples': 17308160, 'steps': 33804, 'loss/train': 1.6408255100250244} 02/25/2022 21:20:47 - INFO - codeparrot_training - Step 33805: {'lr': 0.00012777331525530827, 'samples': 17308672, 'steps': 33805, 'loss/train': 1.994948387145996} 02/25/2022 21:20:51 - INFO - codeparrot_training - Step 33806: {'lr': 0.00012775904195171402, 'samples': 17309184, 'steps': 33806, 'loss/train': 1.8792275190353394} 02/25/2022 21:20:57 - INFO - codeparrot_training - Step 33807: {'lr': 0.00012774476917176112, 'samples': 17309696, 'steps': 33807, 'loss/train': 0.7920739054679871} 02/25/2022 21:21:00 - INFO - codeparrot_training - Step 33808: {'lr': 0.00012773049691551102, 'samples': 17310208, 'steps': 33808, 'loss/train': 1.5104888677597046} 02/25/2022 21:21:06 - INFO - codeparrot_training - Step 33809: {'lr': 0.00012771622518302446, 'samples': 17310720, 'steps': 33809, 'loss/train': 1.0942180156707764} 02/25/2022 21:21:09 - INFO - codeparrot_training - Step 33810: {'lr': 0.00012770195397436283, 'samples': 17311232, 'steps': 33810, 'loss/train': 1.358215570449829} 02/25/2022 21:21:15 - INFO - codeparrot_training - Step 33811: {'lr': 0.0001276876832895871, 'samples': 17311744, 'steps': 33811, 'loss/train': 1.5940165519714355} 02/25/2022 21:21:18 - INFO - codeparrot_training - Step 33812: {'lr': 0.00012767341312875868, 'samples': 17312256, 'steps': 33812, 'loss/train': 1.5972968339920044} 02/25/2022 21:21:24 - INFO - codeparrot_training - Step 33813: {'lr': 0.00012765914349193835, 'samples': 17312768, 'steps': 33813, 'loss/train': 0.7212535738945007} 02/25/2022 21:21:27 - INFO - codeparrot_training - Step 33814: {'lr': 0.0001276448743791875, 'samples': 17313280, 'steps': 33814, 'loss/train': 1.6966056823730469} 02/25/2022 21:21:33 - INFO - codeparrot_training - Step 33815: {'lr': 0.00012763060579056707, 'samples': 17313792, 'steps': 33815, 'loss/train': 1.6988589763641357} 02/25/2022 21:21:36 - INFO - codeparrot_training - Step 33816: {'lr': 0.0001276163377261383, 'samples': 17314304, 'steps': 33816, 'loss/train': 1.5363017320632935} 02/25/2022 21:21:42 - INFO - codeparrot_training - Step 33817: {'lr': 0.00012760207018596234, 'samples': 17314816, 'steps': 33817, 'loss/train': 0.8881706595420837} 02/25/2022 21:21:45 - INFO - codeparrot_training - Step 33818: {'lr': 0.00012758780317010022, 'samples': 17315328, 'steps': 33818, 'loss/train': 0.9441114664077759} 02/25/2022 21:21:51 - INFO - codeparrot_training - Step 33819: {'lr': 0.000127573536678613, 'samples': 17315840, 'steps': 33819, 'loss/train': 1.4173080921173096} 02/25/2022 21:21:54 - INFO - codeparrot_training - Step 33820: {'lr': 0.000127559270711562, 'samples': 17316352, 'steps': 33820, 'loss/train': 1.4545091390609741} 02/25/2022 21:21:59 - INFO - codeparrot_training - Step 33821: {'lr': 0.00012754500526900814, 'samples': 17316864, 'steps': 33821, 'loss/train': 2.3621437549591064} 02/25/2022 21:22:03 - INFO - codeparrot_training - Step 33822: {'lr': 0.0001275307403510126, 'samples': 17317376, 'steps': 33822, 'loss/train': 1.1926594972610474} 02/25/2022 21:22:09 - INFO - codeparrot_training - Step 33823: {'lr': 0.00012751647595763644, 'samples': 17317888, 'steps': 33823, 'loss/train': 1.8042646646499634} 02/25/2022 21:22:13 - INFO - codeparrot_training - Step 33824: {'lr': 0.00012750221208894085, 'samples': 17318400, 'steps': 33824, 'loss/train': 1.7800565958023071} 02/25/2022 21:22:18 - INFO - codeparrot_training - Step 33825: {'lr': 0.0001274879487449869, 'samples': 17318912, 'steps': 33825, 'loss/train': 2.3981049060821533} 02/25/2022 21:22:22 - INFO - codeparrot_training - Step 33826: {'lr': 0.00012747368592583568, 'samples': 17319424, 'steps': 33826, 'loss/train': 1.627747893333435} 02/25/2022 21:22:27 - INFO - codeparrot_training - Step 33827: {'lr': 0.00012745942363154828, 'samples': 17319936, 'steps': 33827, 'loss/train': 1.4943040609359741} 02/25/2022 21:22:31 - INFO - codeparrot_training - Step 33828: {'lr': 0.0001274451618621857, 'samples': 17320448, 'steps': 33828, 'loss/train': 2.628154754638672} 02/25/2022 21:22:36 - INFO - codeparrot_training - Step 33829: {'lr': 0.00012743090061780922, 'samples': 17320960, 'steps': 33829, 'loss/train': 1.7411680221557617} 02/25/2022 21:22:40 - INFO - codeparrot_training - Step 33830: {'lr': 0.00012741663989847984, 'samples': 17321472, 'steps': 33830, 'loss/train': 1.3970381021499634} 02/25/2022 21:22:45 - INFO - codeparrot_training - Step 33831: {'lr': 0.00012740237970425866, 'samples': 17321984, 'steps': 33831, 'loss/train': 2.2303411960601807} 02/25/2022 21:22:49 - INFO - codeparrot_training - Step 33832: {'lr': 0.00012738812003520666, 'samples': 17322496, 'steps': 33832, 'loss/train': 1.0654678344726562} 02/25/2022 21:22:55 - INFO - codeparrot_training - Step 33833: {'lr': 0.00012737386089138513, 'samples': 17323008, 'steps': 33833, 'loss/train': 2.1928462982177734} 02/25/2022 21:22:58 - INFO - codeparrot_training - Step 33834: {'lr': 0.00012735960227285503, 'samples': 17323520, 'steps': 33834, 'loss/train': 2.309786558151245} 02/25/2022 21:23:04 - INFO - codeparrot_training - Step 33835: {'lr': 0.00012734534417967747, 'samples': 17324032, 'steps': 33835, 'loss/train': 2.4196839332580566} 02/25/2022 21:23:07 - INFO - codeparrot_training - Step 33836: {'lr': 0.0001273310866119134, 'samples': 17324544, 'steps': 33836, 'loss/train': 0.07057811319828033} 02/25/2022 21:23:13 - INFO - codeparrot_training - Step 33837: {'lr': 0.00012731682956962405, 'samples': 17325056, 'steps': 33837, 'loss/train': 0.9658114910125732} 02/25/2022 21:23:16 - INFO - codeparrot_training - Step 33838: {'lr': 0.00012730257305287052, 'samples': 17325568, 'steps': 33838, 'loss/train': 2.374498128890991} 02/25/2022 21:23:22 - INFO - codeparrot_training - Step 33839: {'lr': 0.00012728831706171375, 'samples': 17326080, 'steps': 33839, 'loss/train': 1.5761159658432007} 02/25/2022 21:23:25 - INFO - codeparrot_training - Step 33840: {'lr': 0.00012727406159621478, 'samples': 17326592, 'steps': 33840, 'loss/train': 2.206552028656006} 02/25/2022 21:23:31 - INFO - codeparrot_training - Step 33841: {'lr': 0.00012725980665643488, 'samples': 17327104, 'steps': 33841, 'loss/train': 1.8146603107452393} 02/25/2022 21:23:34 - INFO - codeparrot_training - Step 33842: {'lr': 0.0001272455522424349, 'samples': 17327616, 'steps': 33842, 'loss/train': 1.3796119689941406} 02/25/2022 21:23:40 - INFO - codeparrot_training - Step 33843: {'lr': 0.00012723129835427616, 'samples': 17328128, 'steps': 33843, 'loss/train': 2.133607864379883} 02/25/2022 21:23:44 - INFO - codeparrot_training - Step 33844: {'lr': 0.00012721704499201939, 'samples': 17328640, 'steps': 33844, 'loss/train': 0.23322130739688873} 02/25/2022 21:23:49 - INFO - codeparrot_training - Step 33845: {'lr': 0.0001272027921557259, 'samples': 17329152, 'steps': 33845, 'loss/train': 1.4369958639144897} 02/25/2022 21:23:53 - INFO - codeparrot_training - Step 33846: {'lr': 0.00012718853984545653, 'samples': 17329664, 'steps': 33846, 'loss/train': 2.2980916500091553} 02/25/2022 21:23:58 - INFO - codeparrot_training - Step 33847: {'lr': 0.00012717428806127267, 'samples': 17330176, 'steps': 33847, 'loss/train': 2.0317742824554443} 02/25/2022 21:24:02 - INFO - codeparrot_training - Step 33848: {'lr': 0.000127160036803235, 'samples': 17330688, 'steps': 33848, 'loss/train': 1.8091988563537598} 02/25/2022 21:24:07 - INFO - codeparrot_training - Step 33849: {'lr': 0.00012714578607140475, 'samples': 17331200, 'steps': 33849, 'loss/train': 2.528547525405884} 02/25/2022 21:24:11 - INFO - codeparrot_training - Step 33850: {'lr': 0.0001271315358658429, 'samples': 17331712, 'steps': 33850, 'loss/train': 1.903111457824707} 02/25/2022 21:24:16 - INFO - codeparrot_training - Step 33851: {'lr': 0.00012711728618661062, 'samples': 17332224, 'steps': 33851, 'loss/train': 1.893397331237793} 02/25/2022 21:24:20 - INFO - codeparrot_training - Step 33852: {'lr': 0.0001271030370337689, 'samples': 17332736, 'steps': 33852, 'loss/train': 0.6803359985351562} 02/25/2022 21:24:27 - INFO - codeparrot_training - Step 33853: {'lr': 0.0001270887884073787, 'samples': 17333248, 'steps': 33853, 'loss/train': 2.1163859367370605} 02/25/2022 21:24:30 - INFO - codeparrot_training - Step 33854: {'lr': 0.000127074540307501, 'samples': 17333760, 'steps': 33854, 'loss/train': 2.4561607837677} 02/25/2022 21:24:36 - INFO - codeparrot_training - Step 33855: {'lr': 0.00012706029273419706, 'samples': 17334272, 'steps': 33855, 'loss/train': 1.0953820943832397} 02/25/2022 21:24:39 - INFO - codeparrot_training - Step 33856: {'lr': 0.00012704604568752775, 'samples': 17334784, 'steps': 33856, 'loss/train': 2.2849254608154297} 02/25/2022 21:24:45 - INFO - codeparrot_training - Step 33857: {'lr': 0.0001270317991675542, 'samples': 17335296, 'steps': 33857, 'loss/train': 0.9441487193107605} 02/25/2022 21:24:48 - INFO - codeparrot_training - Step 33858: {'lr': 0.00012701755317433722, 'samples': 17335808, 'steps': 33858, 'loss/train': 1.4715889692306519} 02/25/2022 21:24:54 - INFO - codeparrot_training - Step 33859: {'lr': 0.0001270033077079381, 'samples': 17336320, 'steps': 33859, 'loss/train': 2.308405876159668} 02/25/2022 21:24:57 - INFO - codeparrot_training - Step 33860: {'lr': 0.00012698906276841776, 'samples': 17336832, 'steps': 33860, 'loss/train': 2.1189324855804443} 02/25/2022 21:25:03 - INFO - codeparrot_training - Step 33861: {'lr': 0.00012697481835583725, 'samples': 17337344, 'steps': 33861, 'loss/train': 2.567265748977661} 02/25/2022 21:25:06 - INFO - codeparrot_training - Step 33862: {'lr': 0.0001269605744702574, 'samples': 17337856, 'steps': 33862, 'loss/train': 1.2684749364852905} 02/25/2022 21:25:14 - INFO - codeparrot_training - Step 33863: {'lr': 0.00012694633111173952, 'samples': 17338368, 'steps': 33863, 'loss/train': 1.4620167016983032} 02/25/2022 21:25:17 - INFO - codeparrot_training - Step 33864: {'lr': 0.00012693208828034447, 'samples': 17338880, 'steps': 33864, 'loss/train': 1.5807520151138306} 02/25/2022 21:25:23 - INFO - codeparrot_training - Step 33865: {'lr': 0.00012691784597613327, 'samples': 17339392, 'steps': 33865, 'loss/train': 2.109790802001953} 02/25/2022 21:25:26 - INFO - codeparrot_training - Step 33866: {'lr': 0.00012690360419916681, 'samples': 17339904, 'steps': 33866, 'loss/train': 2.128892421722412} 02/25/2022 21:25:32 - INFO - codeparrot_training - Step 33867: {'lr': 0.00012688936294950637, 'samples': 17340416, 'steps': 33867, 'loss/train': 2.1829886436462402} 02/25/2022 21:25:35 - INFO - codeparrot_training - Step 33868: {'lr': 0.0001268751222272128, 'samples': 17340928, 'steps': 33868, 'loss/train': 1.2478166818618774} 02/25/2022 21:25:41 - INFO - codeparrot_training - Step 33869: {'lr': 0.00012686088203234708, 'samples': 17341440, 'steps': 33869, 'loss/train': 0.27779507637023926} 02/25/2022 21:25:44 - INFO - codeparrot_training - Step 33870: {'lr': 0.00012684664236497017, 'samples': 17341952, 'steps': 33870, 'loss/train': 2.0378453731536865} 02/25/2022 21:25:50 - INFO - codeparrot_training - Step 33871: {'lr': 0.00012683240322514322, 'samples': 17342464, 'steps': 33871, 'loss/train': 1.2678548097610474} 02/25/2022 21:25:53 - INFO - codeparrot_training - Step 33872: {'lr': 0.00012681816461292713, 'samples': 17342976, 'steps': 33872, 'loss/train': 2.1260597705841064} 02/25/2022 21:26:01 - INFO - codeparrot_training - Step 33873: {'lr': 0.00012680392652838295, 'samples': 17343488, 'steps': 33873, 'loss/train': 1.573125958442688} 02/25/2022 21:26:04 - INFO - codeparrot_training - Step 33874: {'lr': 0.00012678968897157162, 'samples': 17344000, 'steps': 33874, 'loss/train': 1.628826379776001} 02/25/2022 21:26:09 - INFO - codeparrot_training - Step 33875: {'lr': 0.00012677545194255402, 'samples': 17344512, 'steps': 33875, 'loss/train': 1.747269630432129} 02/25/2022 21:26:13 - INFO - codeparrot_training - Step 33876: {'lr': 0.00012676121544139135, 'samples': 17345024, 'steps': 33876, 'loss/train': 1.6083604097366333} 02/25/2022 21:26:19 - INFO - codeparrot_training - Step 33877: {'lr': 0.0001267469794681444, 'samples': 17345536, 'steps': 33877, 'loss/train': 1.6406654119491577} 02/25/2022 21:26:22 - INFO - codeparrot_training - Step 33878: {'lr': 0.00012673274402287448, 'samples': 17346048, 'steps': 33878, 'loss/train': 1.620491623878479} 02/25/2022 21:26:28 - INFO - codeparrot_training - Step 33879: {'lr': 0.0001267185091056421, 'samples': 17346560, 'steps': 33879, 'loss/train': 1.682497262954712} 02/25/2022 21:26:31 - INFO - codeparrot_training - Step 33880: {'lr': 0.00012670427471650863, 'samples': 17347072, 'steps': 33880, 'loss/train': 2.2391676902770996} 02/25/2022 21:26:37 - INFO - codeparrot_training - Step 33881: {'lr': 0.00012669004085553477, 'samples': 17347584, 'steps': 33881, 'loss/train': 1.7439417839050293} 02/25/2022 21:26:40 - INFO - codeparrot_training - Step 33882: {'lr': 0.00012667580752278183, 'samples': 17348096, 'steps': 33882, 'loss/train': 2.0875656604766846} 02/25/2022 21:26:46 - INFO - codeparrot_training - Step 33883: {'lr': 0.00012666157471831035, 'samples': 17348608, 'steps': 33883, 'loss/train': 0.9700318574905396} 02/25/2022 21:26:53 - INFO - codeparrot_training - Step 33884: {'lr': 0.00012664734244218165, 'samples': 17349120, 'steps': 33884, 'loss/train': 1.6929476261138916} 02/25/2022 21:26:56 - INFO - codeparrot_training - Step 33885: {'lr': 0.00012663311069445644, 'samples': 17349632, 'steps': 33885, 'loss/train': 4.503514766693115} 02/25/2022 21:27:02 - INFO - codeparrot_training - Step 33886: {'lr': 0.000126618879475196, 'samples': 17350144, 'steps': 33886, 'loss/train': 2.246424674987793} 02/25/2022 21:27:05 - INFO - codeparrot_training - Step 33887: {'lr': 0.00012660464878446094, 'samples': 17350656, 'steps': 33887, 'loss/train': 1.6511085033416748} 02/25/2022 21:27:11 - INFO - codeparrot_training - Step 33888: {'lr': 0.00012659041862231245, 'samples': 17351168, 'steps': 33888, 'loss/train': 1.7126740217208862} 02/25/2022 21:27:14 - INFO - codeparrot_training - Step 33889: {'lr': 0.00012657618898881135, 'samples': 17351680, 'steps': 33889, 'loss/train': 1.870279312133789} 02/25/2022 21:27:20 - INFO - codeparrot_training - Step 33890: {'lr': 0.00012656195988401876, 'samples': 17352192, 'steps': 33890, 'loss/train': 1.5617823600769043} 02/25/2022 21:27:23 - INFO - codeparrot_training - Step 33891: {'lr': 0.00012654773130799552, 'samples': 17352704, 'steps': 33891, 'loss/train': 1.8851540088653564} 02/25/2022 21:27:29 - INFO - codeparrot_training - Step 33892: {'lr': 0.0001265335032608026, 'samples': 17353216, 'steps': 33892, 'loss/train': 1.4389629364013672} 02/25/2022 21:27:33 - INFO - codeparrot_training - Step 33893: {'lr': 0.0001265192757425009, 'samples': 17353728, 'steps': 33893, 'loss/train': 1.3848462104797363} 02/25/2022 21:27:38 - INFO - codeparrot_training - Step 33894: {'lr': 0.00012650504875315145, 'samples': 17354240, 'steps': 33894, 'loss/train': 1.4005266427993774} 02/25/2022 21:27:41 - INFO - codeparrot_training - Step 33895: {'lr': 0.00012649082229281516, 'samples': 17354752, 'steps': 33895, 'loss/train': 1.4963746070861816} 02/25/2022 21:27:47 - INFO - codeparrot_training - Step 33896: {'lr': 0.00012647659636155298, 'samples': 17355264, 'steps': 33896, 'loss/train': 2.163606643676758} 02/25/2022 21:27:51 - INFO - codeparrot_training - Step 33897: {'lr': 0.00012646237095942576, 'samples': 17355776, 'steps': 33897, 'loss/train': 1.659933090209961} 02/25/2022 21:27:56 - INFO - codeparrot_training - Step 33898: {'lr': 0.00012644814608649458, 'samples': 17356288, 'steps': 33898, 'loss/train': 2.6481409072875977} 02/25/2022 21:28:00 - INFO - codeparrot_training - Step 33899: {'lr': 0.00012643392174282032, 'samples': 17356800, 'steps': 33899, 'loss/train': 0.5791870951652527} 02/25/2022 21:28:03 - INFO - codeparrot_training - Step 33900: {'lr': 0.00012641969792846392, 'samples': 17357312, 'steps': 33900, 'loss/train': 3.5239765644073486} 02/25/2022 21:28:10 - INFO - codeparrot_training - Step 33901: {'lr': 0.00012640547464348617, 'samples': 17357824, 'steps': 33901, 'loss/train': 2.549034833908081} 02/25/2022 21:28:13 - INFO - codeparrot_training - Step 33902: {'lr': 0.00012639125188794822, 'samples': 17358336, 'steps': 33902, 'loss/train': 2.302746534347534} 02/25/2022 21:28:19 - INFO - codeparrot_training - Step 33903: {'lr': 0.00012637702966191084, 'samples': 17358848, 'steps': 33903, 'loss/train': 1.5887818336486816} 02/25/2022 21:28:24 - INFO - codeparrot_training - Step 33904: {'lr': 0.00012636280796543515, 'samples': 17359360, 'steps': 33904, 'loss/train': 0.2531879246234894} 02/25/2022 21:28:28 - INFO - codeparrot_training - Step 33905: {'lr': 0.00012634858679858176, 'samples': 17359872, 'steps': 33905, 'loss/train': 1.86305832862854} 02/25/2022 21:28:33 - INFO - codeparrot_training - Step 33906: {'lr': 0.00012633436616141187, 'samples': 17360384, 'steps': 33906, 'loss/train': 1.60151207447052} 02/25/2022 21:28:37 - INFO - codeparrot_training - Step 33907: {'lr': 0.00012632014605398628, 'samples': 17360896, 'steps': 33907, 'loss/train': 1.1821651458740234} 02/25/2022 21:28:42 - INFO - codeparrot_training - Step 33908: {'lr': 0.0001263059264763659, 'samples': 17361408, 'steps': 33908, 'loss/train': 1.4744224548339844} 02/25/2022 21:28:46 - INFO - codeparrot_training - Step 33909: {'lr': 0.00012629170742861157, 'samples': 17361920, 'steps': 33909, 'loss/train': 1.404135823249817} 02/25/2022 21:28:52 - INFO - codeparrot_training - Step 33910: {'lr': 0.00012627748891078439, 'samples': 17362432, 'steps': 33910, 'loss/train': 1.8911937475204468} 02/25/2022 21:28:56 - INFO - codeparrot_training - Step 33911: {'lr': 0.00012626327092294515, 'samples': 17362944, 'steps': 33911, 'loss/train': 1.41278076171875} 02/25/2022 21:29:01 - INFO - codeparrot_training - Step 33912: {'lr': 0.00012624905346515463, 'samples': 17363456, 'steps': 33912, 'loss/train': 2.0132040977478027} 02/25/2022 21:29:05 - INFO - codeparrot_training - Step 33913: {'lr': 0.00012623483653747403, 'samples': 17363968, 'steps': 33913, 'loss/train': 2.729323387145996} 02/25/2022 21:29:10 - INFO - codeparrot_training - Step 33914: {'lr': 0.00012622062013996406, 'samples': 17364480, 'steps': 33914, 'loss/train': 0.7326216101646423} 02/25/2022 21:29:14 - INFO - codeparrot_training - Step 33915: {'lr': 0.00012620640427268566, 'samples': 17364992, 'steps': 33915, 'loss/train': 1.2955124378204346} 02/25/2022 21:29:19 - INFO - codeparrot_training - Step 33916: {'lr': 0.00012619218893569962, 'samples': 17365504, 'steps': 33916, 'loss/train': 1.2184885740280151} 02/25/2022 21:29:23 - INFO - codeparrot_training - Step 33917: {'lr': 0.000126177974129067, 'samples': 17366016, 'steps': 33917, 'loss/train': 1.6754100322723389} 02/25/2022 21:29:28 - INFO - codeparrot_training - Step 33918: {'lr': 0.00012616375985284863, 'samples': 17366528, 'steps': 33918, 'loss/train': 1.5852715969085693} 02/25/2022 21:29:32 - INFO - codeparrot_training - Step 33919: {'lr': 0.0001261495461071054, 'samples': 17367040, 'steps': 33919, 'loss/train': 2.148063898086548} 02/25/2022 21:29:38 - INFO - codeparrot_training - Step 33920: {'lr': 0.0001261353328918981, 'samples': 17367552, 'steps': 33920, 'loss/train': 1.9192047119140625} 02/25/2022 21:29:42 - INFO - codeparrot_training - Step 33921: {'lr': 0.0001261211202072878, 'samples': 17368064, 'steps': 33921, 'loss/train': 1.1880720853805542} 02/25/2022 21:29:47 - INFO - codeparrot_training - Step 33922: {'lr': 0.00012610690805333526, 'samples': 17368576, 'steps': 33922, 'loss/train': 1.4836456775665283} 02/25/2022 21:29:51 - INFO - codeparrot_training - Step 33923: {'lr': 0.0001260926964301014, 'samples': 17369088, 'steps': 33923, 'loss/train': 1.558600664138794} 02/25/2022 21:29:56 - INFO - codeparrot_training - Step 33924: {'lr': 0.00012607848533764698, 'samples': 17369600, 'steps': 33924, 'loss/train': 1.5515732765197754} 02/25/2022 21:30:00 - INFO - codeparrot_training - Step 33925: {'lr': 0.00012606427477603307, 'samples': 17370112, 'steps': 33925, 'loss/train': 1.622775912284851} 02/25/2022 21:30:06 - INFO - codeparrot_training - Step 33926: {'lr': 0.0001260500647453205, 'samples': 17370624, 'steps': 33926, 'loss/train': 0.9117430448532104} 02/25/2022 21:30:09 - INFO - codeparrot_training - Step 33927: {'lr': 0.00012603585524557004, 'samples': 17371136, 'steps': 33927, 'loss/train': 0.8037216663360596} 02/25/2022 21:30:15 - INFO - codeparrot_training - Step 33928: {'lr': 0.00012602164627684254, 'samples': 17371648, 'steps': 33928, 'loss/train': 1.9353673458099365} 02/25/2022 21:30:18 - INFO - codeparrot_training - Step 33929: {'lr': 0.000126007437839199, 'samples': 17372160, 'steps': 33929, 'loss/train': 1.9547228813171387} 02/25/2022 21:30:25 - INFO - codeparrot_training - Step 33930: {'lr': 0.0001259932299327003, 'samples': 17372672, 'steps': 33930, 'loss/train': 2.6104300022125244} 02/25/2022 21:30:28 - INFO - codeparrot_training - Step 33931: {'lr': 0.00012597902255740716, 'samples': 17373184, 'steps': 33931, 'loss/train': 0.9223966598510742} 02/25/2022 21:30:34 - INFO - codeparrot_training - Step 33932: {'lr': 0.00012596481571338042, 'samples': 17373696, 'steps': 33932, 'loss/train': 0.9656538963317871} 02/25/2022 21:30:37 - INFO - codeparrot_training - Step 33933: {'lr': 0.00012595060940068115, 'samples': 17374208, 'steps': 33933, 'loss/train': 1.844779133796692} 02/25/2022 21:30:43 - INFO - codeparrot_training - Step 33934: {'lr': 0.00012593640361937007, 'samples': 17374720, 'steps': 33934, 'loss/train': 1.202297568321228} 02/25/2022 21:30:46 - INFO - codeparrot_training - Step 33935: {'lr': 0.000125922198369508, 'samples': 17375232, 'steps': 33935, 'loss/train': 1.727132797241211} 02/25/2022 21:30:52 - INFO - codeparrot_training - Step 33936: {'lr': 0.0001259079936511558, 'samples': 17375744, 'steps': 33936, 'loss/train': 2.462247610092163} 02/25/2022 21:30:55 - INFO - codeparrot_training - Step 33937: {'lr': 0.0001258937894643744, 'samples': 17376256, 'steps': 33937, 'loss/train': 2.097827434539795} 02/25/2022 21:31:01 - INFO - codeparrot_training - Step 33938: {'lr': 0.00012587958580922453, 'samples': 17376768, 'steps': 33938, 'loss/train': 1.4040406942367554} 02/25/2022 21:31:04 - INFO - codeparrot_training - Step 33939: {'lr': 0.0001258653826857673, 'samples': 17377280, 'steps': 33939, 'loss/train': 3.076289415359497} 02/25/2022 21:31:10 - INFO - codeparrot_training - Step 33940: {'lr': 0.00012585118009406313, 'samples': 17377792, 'steps': 33940, 'loss/train': 1.3370944261550903} 02/25/2022 21:31:13 - INFO - codeparrot_training - Step 33941: {'lr': 0.00012583697803417317, 'samples': 17378304, 'steps': 33941, 'loss/train': 1.432740330696106} 02/25/2022 21:31:19 - INFO - codeparrot_training - Step 33942: {'lr': 0.0001258227765061581, 'samples': 17378816, 'steps': 33942, 'loss/train': 1.8163667917251587} 02/25/2022 21:31:22 - INFO - codeparrot_training - Step 33943: {'lr': 0.000125808575510079, 'samples': 17379328, 'steps': 33943, 'loss/train': 3.35750675201416} 02/25/2022 21:31:28 - INFO - codeparrot_training - Step 33944: {'lr': 0.00012579437504599638, 'samples': 17379840, 'steps': 33944, 'loss/train': 1.80765700340271} 02/25/2022 21:31:31 - INFO - codeparrot_training - Step 33945: {'lr': 0.00012578017511397126, 'samples': 17380352, 'steps': 33945, 'loss/train': 1.9311050176620483} 02/25/2022 21:31:38 - INFO - codeparrot_training - Step 33946: {'lr': 0.00012576597571406438, 'samples': 17380864, 'steps': 33946, 'loss/train': 1.7766954898834229} 02/25/2022 21:31:41 - INFO - codeparrot_training - Step 33947: {'lr': 0.00012575177684633675, 'samples': 17381376, 'steps': 33947, 'loss/train': 1.289225459098816} 02/25/2022 21:31:47 - INFO - codeparrot_training - Step 33948: {'lr': 0.0001257375785108489, 'samples': 17381888, 'steps': 33948, 'loss/train': 2.7758259773254395} 02/25/2022 21:31:50 - INFO - codeparrot_training - Step 33949: {'lr': 0.0001257233807076619, 'samples': 17382400, 'steps': 33949, 'loss/train': 1.406354546546936} 02/25/2022 21:31:56 - INFO - codeparrot_training - Step 33950: {'lr': 0.00012570918343683636, 'samples': 17382912, 'steps': 33950, 'loss/train': 1.9868816137313843} 02/25/2022 21:31:59 - INFO - codeparrot_training - Step 33951: {'lr': 0.00012569498669843333, 'samples': 17383424, 'steps': 33951, 'loss/train': 1.9443607330322266} 02/25/2022 21:32:05 - INFO - codeparrot_training - Step 33952: {'lr': 0.0001256807904925135, 'samples': 17383936, 'steps': 33952, 'loss/train': 1.1866079568862915} 02/25/2022 21:32:08 - INFO - codeparrot_training - Step 33953: {'lr': 0.00012566659481913766, 'samples': 17384448, 'steps': 33953, 'loss/train': 1.3708676099777222} 02/25/2022 21:32:14 - INFO - codeparrot_training - Step 33954: {'lr': 0.00012565239967836657, 'samples': 17384960, 'steps': 33954, 'loss/train': 1.6714774370193481} 02/25/2022 21:32:17 - INFO - codeparrot_training - Step 33955: {'lr': 0.00012563820507026122, 'samples': 17385472, 'steps': 33955, 'loss/train': 2.0897037982940674} 02/25/2022 21:32:24 - INFO - codeparrot_training - Step 33956: {'lr': 0.0001256240109948823, 'samples': 17385984, 'steps': 33956, 'loss/train': 1.468039631843567} 02/25/2022 21:32:28 - INFO - codeparrot_training - Step 33957: {'lr': 0.00012560981745229061, 'samples': 17386496, 'steps': 33957, 'loss/train': 1.7863770723342896} 02/25/2022 21:32:33 - INFO - codeparrot_training - Step 33958: {'lr': 0.000125595624442547, 'samples': 17387008, 'steps': 33958, 'loss/train': 1.113387107849121} 02/25/2022 21:32:37 - INFO - codeparrot_training - Step 33959: {'lr': 0.0001255814319657121, 'samples': 17387520, 'steps': 33959, 'loss/train': 1.5041394233703613} 02/25/2022 21:32:42 - INFO - codeparrot_training - Step 33960: {'lr': 0.00012556724002184697, 'samples': 17388032, 'steps': 33960, 'loss/train': 2.3950276374816895} 02/25/2022 21:32:46 - INFO - codeparrot_training - Step 33961: {'lr': 0.00012555304861101225, 'samples': 17388544, 'steps': 33961, 'loss/train': 1.5436636209487915} 02/25/2022 21:32:51 - INFO - codeparrot_training - Step 33962: {'lr': 0.00012553885773326873, 'samples': 17389056, 'steps': 33962, 'loss/train': 1.258392333984375} 02/25/2022 21:32:54 - INFO - codeparrot_training - Step 33963: {'lr': 0.00012552466738867719, 'samples': 17389568, 'steps': 33963, 'loss/train': 0.9962368607521057} 02/25/2022 21:33:00 - INFO - codeparrot_training - Step 33964: {'lr': 0.0001255104775772985, 'samples': 17390080, 'steps': 33964, 'loss/train': 1.6323280334472656} 02/25/2022 21:33:03 - INFO - codeparrot_training - Step 33965: {'lr': 0.00012549628829919342, 'samples': 17390592, 'steps': 33965, 'loss/train': 2.3690903186798096} 02/25/2022 21:33:11 - INFO - codeparrot_training - Step 33966: {'lr': 0.00012548209955442265, 'samples': 17391104, 'steps': 33966, 'loss/train': 2.231538772583008} 02/25/2022 21:33:14 - INFO - codeparrot_training - Step 33967: {'lr': 0.00012546791134304696, 'samples': 17391616, 'steps': 33967, 'loss/train': 2.023271322250366} 02/25/2022 21:33:20 - INFO - codeparrot_training - Step 33968: {'lr': 0.0001254537236651273, 'samples': 17392128, 'steps': 33968, 'loss/train': 1.204742670059204} 02/25/2022 21:33:25 - INFO - codeparrot_training - Step 33969: {'lr': 0.00012543953652072436, 'samples': 17392640, 'steps': 33969, 'loss/train': 0.6722031235694885} 02/25/2022 21:33:29 - INFO - codeparrot_training - Step 33970: {'lr': 0.00012542534990989885, 'samples': 17393152, 'steps': 33970, 'loss/train': 2.123962163925171} 02/25/2022 21:33:34 - INFO - codeparrot_training - Step 33971: {'lr': 0.0001254111638327115, 'samples': 17393664, 'steps': 33971, 'loss/train': 1.146430253982544} 02/25/2022 21:33:38 - INFO - codeparrot_training - Step 33972: {'lr': 0.00012539697828922332, 'samples': 17394176, 'steps': 33972, 'loss/train': 2.195979595184326} 02/25/2022 21:33:43 - INFO - codeparrot_training - Step 33973: {'lr': 0.00012538279327949475, 'samples': 17394688, 'steps': 33973, 'loss/train': 0.11000077426433563} 02/25/2022 21:33:47 - INFO - codeparrot_training - Step 33974: {'lr': 0.0001253686088035869, 'samples': 17395200, 'steps': 33974, 'loss/train': 1.9265820980072021} 02/25/2022 21:33:54 - INFO - codeparrot_training - Step 33975: {'lr': 0.00012535442486156023, 'samples': 17395712, 'steps': 33975, 'loss/train': 1.5206562280654907} 02/25/2022 21:33:57 - INFO - codeparrot_training - Step 33976: {'lr': 0.0001253402414534757, 'samples': 17396224, 'steps': 33976, 'loss/train': 1.1792259216308594} 02/25/2022 21:34:03 - INFO - codeparrot_training - Step 33977: {'lr': 0.0001253260585793939, 'samples': 17396736, 'steps': 33977, 'loss/train': 0.7375963926315308} 02/25/2022 21:34:06 - INFO - codeparrot_training - Step 33978: {'lr': 0.00012531187623937584, 'samples': 17397248, 'steps': 33978, 'loss/train': 2.053083658218384} 02/25/2022 21:34:12 - INFO - codeparrot_training - Step 33979: {'lr': 0.00012529769443348193, 'samples': 17397760, 'steps': 33979, 'loss/train': 2.0573718547821045} 02/25/2022 21:34:15 - INFO - codeparrot_training - Step 33980: {'lr': 0.0001252835131617732, 'samples': 17398272, 'steps': 33980, 'loss/train': 1.6052539348602295} 02/25/2022 21:34:21 - INFO - codeparrot_training - Step 33981: {'lr': 0.0001252693324243102, 'samples': 17398784, 'steps': 33981, 'loss/train': 1.567251443862915} 02/25/2022 21:34:24 - INFO - codeparrot_training - Step 33982: {'lr': 0.00012525515222115395, 'samples': 17399296, 'steps': 33982, 'loss/train': 1.0509614944458008} 02/25/2022 21:34:30 - INFO - codeparrot_training - Step 33983: {'lr': 0.00012524097255236483, 'samples': 17399808, 'steps': 33983, 'loss/train': 2.0293431282043457} 02/25/2022 21:34:33 - INFO - codeparrot_training - Step 33984: {'lr': 0.0001252267934180039, 'samples': 17400320, 'steps': 33984, 'loss/train': 1.4426987171173096} 02/25/2022 21:34:39 - INFO - codeparrot_training - Step 33985: {'lr': 0.00012521261481813163, 'samples': 17400832, 'steps': 33985, 'loss/train': 1.6442755460739136} 02/25/2022 21:34:42 - INFO - codeparrot_training - Step 33986: {'lr': 0.000125198436752809, 'samples': 17401344, 'steps': 33986, 'loss/train': 2.235156774520874} 02/25/2022 21:34:48 - INFO - codeparrot_training - Step 33987: {'lr': 0.00012518425922209664, 'samples': 17401856, 'steps': 33987, 'loss/train': 2.2326109409332275} 02/25/2022 21:34:51 - INFO - codeparrot_training - Step 33988: {'lr': 0.0001251700822260553, 'samples': 17402368, 'steps': 33988, 'loss/train': 1.9844497442245483} 02/25/2022 21:34:57 - INFO - codeparrot_training - Step 33989: {'lr': 0.00012515590576474555, 'samples': 17402880, 'steps': 33989, 'loss/train': 1.950125813484192} 02/25/2022 21:35:00 - INFO - codeparrot_training - Step 33990: {'lr': 0.0001251417298382284, 'samples': 17403392, 'steps': 33990, 'loss/train': 1.9854763746261597} 02/25/2022 21:35:07 - INFO - codeparrot_training - Step 33991: {'lr': 0.00012512755444656442, 'samples': 17403904, 'steps': 33991, 'loss/train': 2.283198356628418} 02/25/2022 21:35:10 - INFO - codeparrot_training - Step 33992: {'lr': 0.00012511337958981433, 'samples': 17404416, 'steps': 33992, 'loss/train': 2.3445448875427246} 02/25/2022 21:35:16 - INFO - codeparrot_training - Step 33993: {'lr': 0.00012509920526803878, 'samples': 17404928, 'steps': 33993, 'loss/train': 1.1401385068893433} 02/25/2022 21:35:19 - INFO - codeparrot_training - Step 33994: {'lr': 0.00012508503148129865, 'samples': 17405440, 'steps': 33994, 'loss/train': 2.314635992050171} 02/25/2022 21:35:25 - INFO - codeparrot_training - Step 33995: {'lr': 0.0001250708582296546, 'samples': 17405952, 'steps': 33995, 'loss/train': 2.2194597721099854} 02/25/2022 21:35:28 - INFO - codeparrot_training - Step 33996: {'lr': 0.00012505668551316735, 'samples': 17406464, 'steps': 33996, 'loss/train': 1.5003371238708496} 02/25/2022 21:35:34 - INFO - codeparrot_training - Step 33997: {'lr': 0.00012504251333189743, 'samples': 17406976, 'steps': 33997, 'loss/train': 1.3384119272232056} 02/25/2022 21:35:37 - INFO - codeparrot_training - Step 33998: {'lr': 0.00012502834168590582, 'samples': 17407488, 'steps': 33998, 'loss/train': 2.855187177658081} 02/25/2022 21:35:43 - INFO - codeparrot_training - Step 33999: {'lr': 0.00012501417057525317, 'samples': 17408000, 'steps': 33999, 'loss/train': 1.9884394407272339} 02/25/2022 21:35:43 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 21:36:00 - WARNING - huggingface_hub.repository - Several commits (34) will be pushed upstream. 02/25/2022 21:36:00 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 21:36:35 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 6b82659..6cccefd floral-grass-11 -> floral-grass-11 02/25/2022 21:36:39 - INFO - codeparrot_training - Step 34000: {'lr': 0.00012500000000000006, 'samples': 17408512, 'steps': 34000, 'loss/train': 1.4243096113204956} 02/25/2022 21:36:47 - INFO - codeparrot_training - Step 34001: {'lr': 0.00012498582996020715, 'samples': 17409024, 'steps': 34001, 'loss/train': 0.8388751745223999} 02/25/2022 21:36:50 - INFO - codeparrot_training - Step 34002: {'lr': 0.0001249716604559354, 'samples': 17409536, 'steps': 34002, 'loss/train': 1.6036185026168823} 02/25/2022 21:36:56 - INFO - codeparrot_training - Step 34003: {'lr': 0.00012495749148724533, 'samples': 17410048, 'steps': 34003, 'loss/train': 1.8250926733016968} 02/25/2022 21:36:59 - INFO - codeparrot_training - Step 34004: {'lr': 0.00012494332305419765, 'samples': 17410560, 'steps': 34004, 'loss/train': 2.837171792984009} 02/25/2022 21:37:05 - INFO - codeparrot_training - Step 34005: {'lr': 0.00012492915515685307, 'samples': 17411072, 'steps': 34005, 'loss/train': 1.3692988157272339} 02/25/2022 21:37:08 - INFO - codeparrot_training - Step 34006: {'lr': 0.00012491498779527218, 'samples': 17411584, 'steps': 34006, 'loss/train': 1.5175013542175293} 02/25/2022 21:37:14 - INFO - codeparrot_training - Step 34007: {'lr': 0.0001249008209695159, 'samples': 17412096, 'steps': 34007, 'loss/train': 0.8875550627708435} 02/25/2022 21:37:17 - INFO - codeparrot_training - Step 34008: {'lr': 0.00012488665467964472, 'samples': 17412608, 'steps': 34008, 'loss/train': 0.9341545701026917} 02/25/2022 21:37:23 - INFO - codeparrot_training - Step 34009: {'lr': 0.0001248724889257194, 'samples': 17413120, 'steps': 34009, 'loss/train': 2.840592384338379} 02/25/2022 21:37:26 - INFO - codeparrot_training - Step 34010: {'lr': 0.00012485832370780052, 'samples': 17413632, 'steps': 34010, 'loss/train': 1.1056419610977173} 02/25/2022 21:37:33 - INFO - codeparrot_training - Step 34011: {'lr': 0.00012484415902594892, 'samples': 17414144, 'steps': 34011, 'loss/train': 1.3155574798583984} 02/25/2022 21:37:36 - INFO - codeparrot_training - Step 34012: {'lr': 0.00012482999488022515, 'samples': 17414656, 'steps': 34012, 'loss/train': 1.75338613986969} 02/25/2022 21:37:42 - INFO - codeparrot_training - Step 34013: {'lr': 0.00012481583127069008, 'samples': 17415168, 'steps': 34013, 'loss/train': 0.644737184047699} 02/25/2022 21:37:45 - INFO - codeparrot_training - Step 34014: {'lr': 0.00012480166819740408, 'samples': 17415680, 'steps': 34014, 'loss/train': 1.4448992013931274} 02/25/2022 21:37:51 - INFO - codeparrot_training - Step 34015: {'lr': 0.00012478750566042804, 'samples': 17416192, 'steps': 34015, 'loss/train': 1.5743814706802368} 02/25/2022 21:37:54 - INFO - codeparrot_training - Step 34016: {'lr': 0.00012477334365982248, 'samples': 17416704, 'steps': 34016, 'loss/train': 0.9618198871612549} 02/25/2022 21:38:00 - INFO - codeparrot_training - Step 34017: {'lr': 0.00012475918219564836, 'samples': 17417216, 'steps': 34017, 'loss/train': 2.1817731857299805} 02/25/2022 21:38:03 - INFO - codeparrot_training - Step 34018: {'lr': 0.00012474502126796593, 'samples': 17417728, 'steps': 34018, 'loss/train': 1.1738075017929077} 02/25/2022 21:38:09 - INFO - codeparrot_training - Step 34019: {'lr': 0.00012473086087683616, 'samples': 17418240, 'steps': 34019, 'loss/train': 2.341294765472412} 02/25/2022 21:38:12 - INFO - codeparrot_training - Step 34020: {'lr': 0.00012471670102231952, 'samples': 17418752, 'steps': 34020, 'loss/train': 2.4172229766845703} 02/25/2022 21:38:18 - INFO - codeparrot_training - Step 34021: {'lr': 0.00012470254170447685, 'samples': 17419264, 'steps': 34021, 'loss/train': 1.9748722314834595} 02/25/2022 21:38:21 - INFO - codeparrot_training - Step 34022: {'lr': 0.00012468838292336868, 'samples': 17419776, 'steps': 34022, 'loss/train': 2.4044883251190186} 02/25/2022 21:38:28 - INFO - codeparrot_training - Step 34023: {'lr': 0.00012467422467905568, 'samples': 17420288, 'steps': 34023, 'loss/train': 2.3403255939483643} 02/25/2022 21:38:31 - INFO - codeparrot_training - Step 34024: {'lr': 0.00012466006697159842, 'samples': 17420800, 'steps': 34024, 'loss/train': 1.6199828386306763} 02/25/2022 21:38:37 - INFO - codeparrot_training - Step 34025: {'lr': 0.00012464590980105776, 'samples': 17421312, 'steps': 34025, 'loss/train': 2.4142913818359375} 02/25/2022 21:38:40 - INFO - codeparrot_training - Step 34026: {'lr': 0.00012463175316749419, 'samples': 17421824, 'steps': 34026, 'loss/train': 2.1441690921783447} 02/25/2022 21:38:46 - INFO - codeparrot_training - Step 34027: {'lr': 0.00012461759707096837, 'samples': 17422336, 'steps': 34027, 'loss/train': 2.3603148460388184} 02/25/2022 21:38:49 - INFO - codeparrot_training - Step 34028: {'lr': 0.00012460344151154088, 'samples': 17422848, 'steps': 34028, 'loss/train': 1.9076552391052246} 02/25/2022 21:38:55 - INFO - codeparrot_training - Step 34029: {'lr': 0.0001245892864892725, 'samples': 17423360, 'steps': 34029, 'loss/train': 0.810051441192627} 02/25/2022 21:38:58 - INFO - codeparrot_training - Step 34030: {'lr': 0.0001245751320042238, 'samples': 17423872, 'steps': 34030, 'loss/train': 2.6465001106262207} 02/25/2022 21:39:04 - INFO - codeparrot_training - Step 34031: {'lr': 0.00012456097805645542, 'samples': 17424384, 'steps': 34031, 'loss/train': 1.9230165481567383} 02/25/2022 21:39:07 - INFO - codeparrot_training - Step 34032: {'lr': 0.00012454682464602788, 'samples': 17424896, 'steps': 34032, 'loss/train': 1.5525074005126953} 02/25/2022 21:39:13 - INFO - codeparrot_training - Step 34033: {'lr': 0.000124532671773002, 'samples': 17425408, 'steps': 34033, 'loss/train': 1.8166000843048096} 02/25/2022 21:39:16 - INFO - codeparrot_training - Step 34034: {'lr': 0.00012451851943743821, 'samples': 17425920, 'steps': 34034, 'loss/train': 1.4815582036972046} 02/25/2022 21:39:22 - INFO - codeparrot_training - Step 34035: {'lr': 0.00012450436763939747, 'samples': 17426432, 'steps': 34035, 'loss/train': 1.8599796295166016} 02/25/2022 21:39:25 - INFO - codeparrot_training - Step 34036: {'lr': 0.00012449021637893993, 'samples': 17426944, 'steps': 34036, 'loss/train': 1.8132766485214233} 02/25/2022 21:39:32 - INFO - codeparrot_training - Step 34037: {'lr': 0.00012447606565612655, 'samples': 17427456, 'steps': 34037, 'loss/train': 2.1141202449798584} 02/25/2022 21:39:35 - INFO - codeparrot_training - Step 34038: {'lr': 0.00012446191547101787, 'samples': 17427968, 'steps': 34038, 'loss/train': 0.9231016635894775} 02/25/2022 21:39:41 - INFO - codeparrot_training - Step 34039: {'lr': 0.00012444776582367446, 'samples': 17428480, 'steps': 34039, 'loss/train': 0.9292402863502502} 02/25/2022 21:39:44 - INFO - codeparrot_training - Step 34040: {'lr': 0.00012443361671415687, 'samples': 17428992, 'steps': 34040, 'loss/train': 1.6624534130096436} 02/25/2022 21:39:50 - INFO - codeparrot_training - Step 34041: {'lr': 0.0001244194681425259, 'samples': 17429504, 'steps': 34041, 'loss/train': 2.331366539001465} 02/25/2022 21:39:53 - INFO - codeparrot_training - Step 34042: {'lr': 0.000124405320108842, 'samples': 17430016, 'steps': 34042, 'loss/train': 2.037729024887085} 02/25/2022 21:39:59 - INFO - codeparrot_training - Step 34043: {'lr': 0.00012439117261316588, 'samples': 17430528, 'steps': 34043, 'loss/train': 1.4858005046844482} 02/25/2022 21:40:02 - INFO - codeparrot_training - Step 34044: {'lr': 0.00012437702565555798, 'samples': 17431040, 'steps': 34044, 'loss/train': 1.946694254875183} 02/25/2022 21:40:08 - INFO - codeparrot_training - Step 34045: {'lr': 0.0001243628792360791, 'samples': 17431552, 'steps': 34045, 'loss/train': 1.1625895500183105} 02/25/2022 21:40:11 - INFO - codeparrot_training - Step 34046: {'lr': 0.00012434873335478977, 'samples': 17432064, 'steps': 34046, 'loss/train': 0.3992147743701935} 02/25/2022 21:40:18 - INFO - codeparrot_training - Step 34047: {'lr': 0.00012433458801175046, 'samples': 17432576, 'steps': 34047, 'loss/train': 1.2809768915176392} 02/25/2022 21:40:21 - INFO - codeparrot_training - Step 34048: {'lr': 0.00012432044320702197, 'samples': 17433088, 'steps': 34048, 'loss/train': 1.73301362991333} 02/25/2022 21:40:27 - INFO - codeparrot_training - Step 34049: {'lr': 0.0001243062989406648, 'samples': 17433600, 'steps': 34049, 'loss/train': 2.468609094619751} 02/25/2022 21:40:30 - INFO - codeparrot_training - Step 34050: {'lr': 0.00012429215521273955, 'samples': 17434112, 'steps': 34050, 'loss/train': 1.0004311800003052} 02/25/2022 21:40:36 - INFO - codeparrot_training - Step 34051: {'lr': 0.00012427801202330666, 'samples': 17434624, 'steps': 34051, 'loss/train': 1.8955143690109253} 02/25/2022 21:40:39 - INFO - codeparrot_training - Step 34052: {'lr': 0.00012426386937242705, 'samples': 17435136, 'steps': 34052, 'loss/train': 2.399646282196045} 02/25/2022 21:40:45 - INFO - codeparrot_training - Step 34053: {'lr': 0.0001242497272601609, 'samples': 17435648, 'steps': 34053, 'loss/train': 1.944251537322998} 02/25/2022 21:40:48 - INFO - codeparrot_training - Step 34054: {'lr': 0.0001242355856865691, 'samples': 17436160, 'steps': 34054, 'loss/train': 0.9422216415405273} 02/25/2022 21:40:54 - INFO - codeparrot_training - Step 34055: {'lr': 0.000124221444651712, 'samples': 17436672, 'steps': 34055, 'loss/train': 1.0910838842391968} 02/25/2022 21:40:57 - INFO - codeparrot_training - Step 34056: {'lr': 0.0001242073041556505, 'samples': 17437184, 'steps': 34056, 'loss/train': 1.9763319492340088} 02/25/2022 21:41:04 - INFO - codeparrot_training - Step 34057: {'lr': 0.00012419316419844475, 'samples': 17437696, 'steps': 34057, 'loss/train': 1.1914631128311157} 02/25/2022 21:41:07 - INFO - codeparrot_training - Step 34058: {'lr': 0.00012417902478015566, 'samples': 17438208, 'steps': 34058, 'loss/train': 1.6025583744049072} 02/25/2022 21:41:13 - INFO - codeparrot_training - Step 34059: {'lr': 0.00012416488590084358, 'samples': 17438720, 'steps': 34059, 'loss/train': 0.6146384477615356} 02/25/2022 21:41:16 - INFO - codeparrot_training - Step 34060: {'lr': 0.00012415074756056923, 'samples': 17439232, 'steps': 34060, 'loss/train': 1.5017462968826294} 02/25/2022 21:41:22 - INFO - codeparrot_training - Step 34061: {'lr': 0.00012413660975939316, 'samples': 17439744, 'steps': 34061, 'loss/train': 1.7307779788970947} 02/25/2022 21:41:25 - INFO - codeparrot_training - Step 34062: {'lr': 0.00012412247249737586, 'samples': 17440256, 'steps': 34062, 'loss/train': 1.2051664590835571} 02/25/2022 21:41:31 - INFO - codeparrot_training - Step 34063: {'lr': 0.00012410833577457785, 'samples': 17440768, 'steps': 34063, 'loss/train': 1.3565372228622437} 02/25/2022 21:41:35 - INFO - codeparrot_training - Step 34064: {'lr': 0.0001240941995910598, 'samples': 17441280, 'steps': 34064, 'loss/train': 2.0141043663024902} 02/25/2022 21:41:41 - INFO - codeparrot_training - Step 34065: {'lr': 0.00012408006394688226, 'samples': 17441792, 'steps': 34065, 'loss/train': 1.417724847793579} 02/25/2022 21:41:44 - INFO - codeparrot_training - Step 34066: {'lr': 0.0001240659288421057, 'samples': 17442304, 'steps': 34066, 'loss/train': 2.277214765548706} 02/25/2022 21:41:51 - INFO - codeparrot_training - Step 34067: {'lr': 0.00012405179427679068, 'samples': 17442816, 'steps': 34067, 'loss/train': 1.711771011352539} 02/25/2022 21:41:54 - INFO - codeparrot_training - Step 34068: {'lr': 0.00012403766025099783, 'samples': 17443328, 'steps': 34068, 'loss/train': 0.7656440138816833} 02/25/2022 21:42:00 - INFO - codeparrot_training - Step 34069: {'lr': 0.00012402352676478757, 'samples': 17443840, 'steps': 34069, 'loss/train': 2.886143684387207} 02/25/2022 21:42:03 - INFO - codeparrot_training - Step 34070: {'lr': 0.00012400939381822072, 'samples': 17444352, 'steps': 34070, 'loss/train': 1.0049716234207153} 02/25/2022 21:42:09 - INFO - codeparrot_training - Step 34071: {'lr': 0.00012399526141135739, 'samples': 17444864, 'steps': 34071, 'loss/train': 2.914724588394165} 02/25/2022 21:42:13 - INFO - codeparrot_training - Step 34072: {'lr': 0.00012398112954425846, 'samples': 17445376, 'steps': 34072, 'loss/train': 2.105363607406616} 02/25/2022 21:42:18 - INFO - codeparrot_training - Step 34073: {'lr': 0.0001239669982169843, 'samples': 17445888, 'steps': 34073, 'loss/train': 2.1937716007232666} 02/25/2022 21:42:22 - INFO - codeparrot_training - Step 34074: {'lr': 0.00012395286742959566, 'samples': 17446400, 'steps': 34074, 'loss/train': 0.2757807672023773} 02/25/2022 21:42:27 - INFO - codeparrot_training - Step 34075: {'lr': 0.0001239387371821527, 'samples': 17446912, 'steps': 34075, 'loss/train': 0.8021522760391235} 02/25/2022 21:42:31 - INFO - codeparrot_training - Step 34076: {'lr': 0.00012392460747471628, 'samples': 17447424, 'steps': 34076, 'loss/train': 0.6329545378684998} 02/25/2022 21:42:36 - INFO - codeparrot_training - Step 34077: {'lr': 0.00012391047830734672, 'samples': 17447936, 'steps': 34077, 'loss/train': 2.679614543914795} 02/25/2022 21:42:40 - INFO - codeparrot_training - Step 34078: {'lr': 0.0001238963496801048, 'samples': 17448448, 'steps': 34078, 'loss/train': 1.8143888711929321} 02/25/2022 21:42:45 - INFO - codeparrot_training - Step 34079: {'lr': 0.0001238822215930507, 'samples': 17448960, 'steps': 34079, 'loss/train': 1.9522573947906494} 02/25/2022 21:42:49 - INFO - codeparrot_training - Step 34080: {'lr': 0.00012386809404624523, 'samples': 17449472, 'steps': 34080, 'loss/train': 0.7795721888542175} 02/25/2022 21:42:54 - INFO - codeparrot_training - Step 34081: {'lr': 0.00012385396703974865, 'samples': 17449984, 'steps': 34081, 'loss/train': 2.5577166080474854} 02/25/2022 21:42:58 - INFO - codeparrot_training - Step 34082: {'lr': 0.00012383984057362174, 'samples': 17450496, 'steps': 34082, 'loss/train': 2.729487419128418} 02/25/2022 21:43:03 - INFO - codeparrot_training - Step 34083: {'lr': 0.00012382571464792492, 'samples': 17451008, 'steps': 34083, 'loss/train': 1.553989052772522} 02/25/2022 21:43:07 - INFO - codeparrot_training - Step 34084: {'lr': 0.00012381158926271863, 'samples': 17451520, 'steps': 34084, 'loss/train': 1.9220963716506958} 02/25/2022 21:43:13 - INFO - codeparrot_training - Step 34085: {'lr': 0.00012379746441806345, 'samples': 17452032, 'steps': 34085, 'loss/train': 1.4028511047363281} 02/25/2022 21:43:17 - INFO - codeparrot_training - Step 34086: {'lr': 0.00012378334011401977, 'samples': 17452544, 'steps': 34086, 'loss/train': 1.713929295539856} 02/25/2022 21:43:22 - INFO - codeparrot_training - Step 34087: {'lr': 0.00012376921635064825, 'samples': 17453056, 'steps': 34087, 'loss/train': 2.0141427516937256} 02/25/2022 21:43:26 - INFO - codeparrot_training - Step 34088: {'lr': 0.00012375509312800934, 'samples': 17453568, 'steps': 34088, 'loss/train': 1.5257792472839355} 02/25/2022 21:43:31 - INFO - codeparrot_training - Step 34089: {'lr': 0.0001237409704461635, 'samples': 17454080, 'steps': 34089, 'loss/train': 2.831057548522949} 02/25/2022 21:43:35 - INFO - codeparrot_training - Step 34090: {'lr': 0.00012372684830517117, 'samples': 17454592, 'steps': 34090, 'loss/train': 1.6461092233657837} 02/25/2022 21:43:40 - INFO - codeparrot_training - Step 34091: {'lr': 0.00012371272670509304, 'samples': 17455104, 'steps': 34091, 'loss/train': 1.003557801246643} 02/25/2022 21:43:44 - INFO - codeparrot_training - Step 34092: {'lr': 0.00012369860564598946, 'samples': 17455616, 'steps': 34092, 'loss/train': 1.6743559837341309} 02/25/2022 21:43:50 - INFO - codeparrot_training - Step 34093: {'lr': 0.00012368448512792095, 'samples': 17456128, 'steps': 34093, 'loss/train': 2.147613048553467} 02/25/2022 21:43:54 - INFO - codeparrot_training - Step 34094: {'lr': 0.00012367036515094793, 'samples': 17456640, 'steps': 34094, 'loss/train': 2.846097946166992} 02/25/2022 21:43:59 - INFO - codeparrot_training - Step 34095: {'lr': 0.00012365624571513103, 'samples': 17457152, 'steps': 34095, 'loss/train': 1.2241394519805908} 02/25/2022 21:44:03 - INFO - codeparrot_training - Step 34096: {'lr': 0.00012364212682053067, 'samples': 17457664, 'steps': 34096, 'loss/train': 2.141115188598633} 02/25/2022 21:44:08 - INFO - codeparrot_training - Step 34097: {'lr': 0.0001236280084672073, 'samples': 17458176, 'steps': 34097, 'loss/train': 2.5461888313293457} 02/25/2022 21:44:12 - INFO - codeparrot_training - Step 34098: {'lr': 0.0001236138906552213, 'samples': 17458688, 'steps': 34098, 'loss/train': 0.9235637187957764} 02/25/2022 21:44:17 - INFO - codeparrot_training - Step 34099: {'lr': 0.0001235997733846334, 'samples': 17459200, 'steps': 34099, 'loss/train': 1.9525666236877441} 02/25/2022 21:44:21 - INFO - codeparrot_training - Step 34100: {'lr': 0.0001235856566555039, 'samples': 17459712, 'steps': 34100, 'loss/train': 1.4226510524749756} 02/25/2022 21:44:27 - INFO - codeparrot_training - Step 34101: {'lr': 0.0001235715404678933, 'samples': 17460224, 'steps': 34101, 'loss/train': 1.412358283996582} 02/25/2022 21:44:30 - INFO - codeparrot_training - Step 34102: {'lr': 0.00012355742482186198, 'samples': 17460736, 'steps': 34102, 'loss/train': 1.0481382608413696} 02/25/2022 21:44:37 - INFO - codeparrot_training - Step 34103: {'lr': 0.00012354330971747063, 'samples': 17461248, 'steps': 34103, 'loss/train': 2.2695329189300537} 02/25/2022 21:44:41 - INFO - codeparrot_training - Step 34104: {'lr': 0.00012352919515477956, 'samples': 17461760, 'steps': 34104, 'loss/train': 1.5227618217468262} 02/25/2022 21:44:46 - INFO - codeparrot_training - Step 34105: {'lr': 0.0001235150811338493, 'samples': 17462272, 'steps': 34105, 'loss/train': 1.0407601594924927} 02/25/2022 21:44:50 - INFO - codeparrot_training - Step 34106: {'lr': 0.0001235009676547401, 'samples': 17462784, 'steps': 34106, 'loss/train': 1.8043757677078247} 02/25/2022 21:44:55 - INFO - codeparrot_training - Step 34107: {'lr': 0.00012348685471751277, 'samples': 17463296, 'steps': 34107, 'loss/train': 0.4909404218196869} 02/25/2022 21:44:59 - INFO - codeparrot_training - Step 34108: {'lr': 0.0001234727423222274, 'samples': 17463808, 'steps': 34108, 'loss/train': 1.8534376621246338} 02/25/2022 21:45:04 - INFO - codeparrot_training - Step 34109: {'lr': 0.00012345863046894488, 'samples': 17464320, 'steps': 34109, 'loss/train': 1.036161184310913} 02/25/2022 21:45:08 - INFO - codeparrot_training - Step 34110: {'lr': 0.00012344451915772517, 'samples': 17464832, 'steps': 34110, 'loss/train': 1.4591485261917114} 02/25/2022 21:45:13 - INFO - codeparrot_training - Step 34111: {'lr': 0.00012343040838862907, 'samples': 17465344, 'steps': 34111, 'loss/train': 1.9940106868743896} 02/25/2022 21:45:17 - INFO - codeparrot_training - Step 34112: {'lr': 0.0001234162981617168, 'samples': 17465856, 'steps': 34112, 'loss/train': 1.9499503374099731} 02/25/2022 21:45:22 - INFO - codeparrot_training - Step 34113: {'lr': 0.00012340218847704915, 'samples': 17466368, 'steps': 34113, 'loss/train': 1.8292899131774902} 02/25/2022 21:45:25 - INFO - codeparrot_training - Step 34114: {'lr': 0.0001233880793346861, 'samples': 17466880, 'steps': 34114, 'loss/train': 1.6427180767059326} 02/25/2022 21:45:33 - INFO - codeparrot_training - Step 34115: {'lr': 0.0001233739707346884, 'samples': 17467392, 'steps': 34115, 'loss/train': 0.5149872303009033} 02/25/2022 21:45:36 - INFO - codeparrot_training - Step 34116: {'lr': 0.00012335986267711632, 'samples': 17467904, 'steps': 34116, 'loss/train': 1.7666956186294556} 02/25/2022 21:45:42 - INFO - codeparrot_training - Step 34117: {'lr': 0.00012334575516203048, 'samples': 17468416, 'steps': 34117, 'loss/train': 1.8671932220458984} 02/25/2022 21:45:45 - INFO - codeparrot_training - Step 34118: {'lr': 0.0001233316481894912, 'samples': 17468928, 'steps': 34118, 'loss/train': 1.2974510192871094} 02/25/2022 21:45:51 - INFO - codeparrot_training - Step 34119: {'lr': 0.0001233175417595589, 'samples': 17469440, 'steps': 34119, 'loss/train': 0.8795772790908813} 02/25/2022 21:45:54 - INFO - codeparrot_training - Step 34120: {'lr': 0.00012330343587229394, 'samples': 17469952, 'steps': 34120, 'loss/train': 0.1629948765039444} 02/25/2022 21:46:00 - INFO - codeparrot_training - Step 34121: {'lr': 0.00012328933052775694, 'samples': 17470464, 'steps': 34121, 'loss/train': 1.5280299186706543} 02/25/2022 21:46:03 - INFO - codeparrot_training - Step 34122: {'lr': 0.00012327522572600823, 'samples': 17470976, 'steps': 34122, 'loss/train': 1.272597074508667} 02/25/2022 21:46:09 - INFO - codeparrot_training - Step 34123: {'lr': 0.00012326112146710815, 'samples': 17471488, 'steps': 34123, 'loss/train': 0.7454657554626465} 02/25/2022 21:46:14 - INFO - codeparrot_training - Step 34124: {'lr': 0.00012324701775111714, 'samples': 17472000, 'steps': 34124, 'loss/train': 2.791496753692627} 02/25/2022 21:46:18 - INFO - codeparrot_training - Step 34125: {'lr': 0.00012323291457809574, 'samples': 17472512, 'steps': 34125, 'loss/train': 1.6492421627044678} 02/25/2022 21:46:21 - INFO - codeparrot_training - Step 34126: {'lr': 0.00012321881194810426, 'samples': 17473024, 'steps': 34126, 'loss/train': 2.4348554611206055} 02/25/2022 21:46:27 - INFO - codeparrot_training - Step 34127: {'lr': 0.00012320470986120315, 'samples': 17473536, 'steps': 34127, 'loss/train': 2.2779111862182617} 02/25/2022 21:46:30 - INFO - codeparrot_training - Step 34128: {'lr': 0.00012319060831745272, 'samples': 17474048, 'steps': 34128, 'loss/train': 2.267850399017334} 02/25/2022 21:46:37 - INFO - codeparrot_training - Step 34129: {'lr': 0.0001231765073169135, 'samples': 17474560, 'steps': 34129, 'loss/train': 0.024210814386606216} 02/25/2022 21:46:41 - INFO - codeparrot_training - Step 34130: {'lr': 0.00012316240685964592, 'samples': 17475072, 'steps': 34130, 'loss/train': 1.3733785152435303} 02/25/2022 21:46:46 - INFO - codeparrot_training - Step 34131: {'lr': 0.00012314830694571028, 'samples': 17475584, 'steps': 34131, 'loss/train': 1.220655918121338} 02/25/2022 21:46:50 - INFO - codeparrot_training - Step 34132: {'lr': 0.000123134207575167, 'samples': 17476096, 'steps': 34132, 'loss/train': 2.375781536102295} 02/25/2022 21:46:55 - INFO - codeparrot_training - Step 34133: {'lr': 0.0001231201087480764, 'samples': 17476608, 'steps': 34133, 'loss/train': 0.9140170812606812} 02/25/2022 21:46:59 - INFO - codeparrot_training - Step 34134: {'lr': 0.00012310601046449906, 'samples': 17477120, 'steps': 34134, 'loss/train': 0.8746749758720398} 02/25/2022 21:47:04 - INFO - codeparrot_training - Step 34135: {'lr': 0.0001230919127244953, 'samples': 17477632, 'steps': 34135, 'loss/train': 1.2982467412948608} 02/25/2022 21:47:10 - INFO - codeparrot_training - Step 34136: {'lr': 0.0001230778155281255, 'samples': 17478144, 'steps': 34136, 'loss/train': 1.7328559160232544} 02/25/2022 21:47:13 - INFO - codeparrot_training - Step 34137: {'lr': 0.0001230637188754499, 'samples': 17478656, 'steps': 34137, 'loss/train': 2.060474395751953} 02/25/2022 21:47:16 - INFO - codeparrot_training - Step 34138: {'lr': 0.00012304962276652915, 'samples': 17479168, 'steps': 34138, 'loss/train': 2.488842487335205} 02/25/2022 21:47:22 - INFO - codeparrot_training - Step 34139: {'lr': 0.00012303552720142346, 'samples': 17479680, 'steps': 34139, 'loss/train': 2.090043067932129} 02/25/2022 21:47:29 - INFO - codeparrot_training - Step 34140: {'lr': 0.00012302143218019328, 'samples': 17480192, 'steps': 34140, 'loss/train': 1.0664960145950317} 02/25/2022 21:47:33 - INFO - codeparrot_training - Step 34141: {'lr': 0.00012300733770289889, 'samples': 17480704, 'steps': 34141, 'loss/train': 2.038123607635498} 02/25/2022 21:47:38 - INFO - codeparrot_training - Step 34142: {'lr': 0.0001229932437696008, 'samples': 17481216, 'steps': 34142, 'loss/train': 1.902867317199707} 02/25/2022 21:47:42 - INFO - codeparrot_training - Step 34143: {'lr': 0.00012297915038035926, 'samples': 17481728, 'steps': 34143, 'loss/train': 2.5757765769958496} 02/25/2022 21:47:47 - INFO - codeparrot_training - Step 34144: {'lr': 0.00012296505753523488, 'samples': 17482240, 'steps': 34144, 'loss/train': 1.0799211263656616} 02/25/2022 21:47:51 - INFO - codeparrot_training - Step 34145: {'lr': 0.00012295096523428766, 'samples': 17482752, 'steps': 34145, 'loss/train': 2.9951064586639404} 02/25/2022 21:47:56 - INFO - codeparrot_training - Step 34146: {'lr': 0.00012293687347757828, 'samples': 17483264, 'steps': 34146, 'loss/train': 1.6820772886276245} 02/25/2022 21:47:59 - INFO - codeparrot_training - Step 34147: {'lr': 0.00012292278226516687, 'samples': 17483776, 'steps': 34147, 'loss/train': 1.5084384679794312} 02/25/2022 21:48:05 - INFO - codeparrot_training - Step 34148: {'lr': 0.00012290869159711413, 'samples': 17484288, 'steps': 34148, 'loss/train': 1.3514103889465332} 02/25/2022 21:48:08 - INFO - codeparrot_training - Step 34149: {'lr': 0.00012289460147347998, 'samples': 17484800, 'steps': 34149, 'loss/train': 1.5390719175338745} 02/25/2022 21:48:15 - INFO - codeparrot_training - Step 34150: {'lr': 0.0001228805118943251, 'samples': 17485312, 'steps': 34150, 'loss/train': 1.820581078529358} 02/25/2022 21:48:19 - INFO - codeparrot_training - Step 34151: {'lr': 0.00012286642285970964, 'samples': 17485824, 'steps': 34151, 'loss/train': 2.831040382385254} 02/25/2022 21:48:24 - INFO - codeparrot_training - Step 34152: {'lr': 0.00012285233436969427, 'samples': 17486336, 'steps': 34152, 'loss/train': 1.1981221437454224} 02/25/2022 21:48:28 - INFO - codeparrot_training - Step 34153: {'lr': 0.0001228382464243389, 'samples': 17486848, 'steps': 34153, 'loss/train': 2.040548801422119} 02/25/2022 21:48:33 - INFO - codeparrot_training - Step 34154: {'lr': 0.00012282415902370423, 'samples': 17487360, 'steps': 34154, 'loss/train': 8.590916633605957} 02/25/2022 21:48:37 - INFO - codeparrot_training - Step 34155: {'lr': 0.00012281007216785033, 'samples': 17487872, 'steps': 34155, 'loss/train': 2.0622220039367676} 02/25/2022 21:48:42 - INFO - codeparrot_training - Step 34156: {'lr': 0.00012279598585683785, 'samples': 17488384, 'steps': 34156, 'loss/train': 1.634106159210205} 02/25/2022 21:48:46 - INFO - codeparrot_training - Step 34157: {'lr': 0.00012278190009072694, 'samples': 17488896, 'steps': 34157, 'loss/train': 1.7318800687789917} 02/25/2022 21:48:51 - INFO - codeparrot_training - Step 34158: {'lr': 0.000122767814869578, 'samples': 17489408, 'steps': 34158, 'loss/train': 1.5939161777496338} 02/25/2022 21:48:54 - INFO - codeparrot_training - Step 34159: {'lr': 0.00012275373019345124, 'samples': 17489920, 'steps': 34159, 'loss/train': 1.2504925727844238} 02/25/2022 21:49:01 - INFO - codeparrot_training - Step 34160: {'lr': 0.00012273964606240718, 'samples': 17490432, 'steps': 34160, 'loss/train': 2.122093677520752} 02/25/2022 21:49:05 - INFO - codeparrot_training - Step 34161: {'lr': 0.00012272556247650606, 'samples': 17490944, 'steps': 34161, 'loss/train': 2.037853717803955} 02/25/2022 21:49:10 - INFO - codeparrot_training - Step 34162: {'lr': 0.00012271147943580823, 'samples': 17491456, 'steps': 34162, 'loss/train': 1.8190076351165771} 02/25/2022 21:49:14 - INFO - codeparrot_training - Step 34163: {'lr': 0.00012269739694037392, 'samples': 17491968, 'steps': 34163, 'loss/train': 2.6141421794891357} 02/25/2022 21:49:20 - INFO - codeparrot_training - Step 34164: {'lr': 0.0001226833149902636, 'samples': 17492480, 'steps': 34164, 'loss/train': 2.3264358043670654} 02/25/2022 21:49:23 - INFO - codeparrot_training - Step 34165: {'lr': 0.00012266923358553755, 'samples': 17492992, 'steps': 34165, 'loss/train': 1.684666395187378} 02/25/2022 21:49:27 - INFO - codeparrot_training - Step 34166: {'lr': 0.00012265515272625606, 'samples': 17493504, 'steps': 34166, 'loss/train': 3.095144271850586} 02/25/2022 21:49:32 - INFO - codeparrot_training - Step 34167: {'lr': 0.00012264107241247938, 'samples': 17494016, 'steps': 34167, 'loss/train': 1.890511155128479} 02/25/2022 21:49:36 - INFO - codeparrot_training - Step 34168: {'lr': 0.000122626992644268, 'samples': 17494528, 'steps': 34168, 'loss/train': 2.21645450592041} 02/25/2022 21:49:41 - INFO - codeparrot_training - Step 34169: {'lr': 0.00012261291342168212, 'samples': 17495040, 'steps': 34169, 'loss/train': 2.033522129058838} 02/25/2022 21:49:44 - INFO - codeparrot_training - Step 34170: {'lr': 0.00012259883474478208, 'samples': 17495552, 'steps': 34170, 'loss/train': 2.2934815883636475} 02/25/2022 21:49:50 - INFO - codeparrot_training - Step 34171: {'lr': 0.00012258475661362808, 'samples': 17496064, 'steps': 34171, 'loss/train': 1.2459129095077515} 02/25/2022 21:49:54 - INFO - codeparrot_training - Step 34172: {'lr': 0.0001225706790282806, 'samples': 17496576, 'steps': 34172, 'loss/train': 0.9784200191497803} 02/25/2022 21:49:59 - INFO - codeparrot_training - Step 34173: {'lr': 0.0001225566019887999, 'samples': 17497088, 'steps': 34173, 'loss/train': 1.7030739784240723} 02/25/2022 21:50:03 - INFO - codeparrot_training - Step 34174: {'lr': 0.00012254252549524626, 'samples': 17497600, 'steps': 34174, 'loss/train': 1.4188021421432495} 02/25/2022 21:50:08 - INFO - codeparrot_training - Step 34175: {'lr': 0.00012252844954767982, 'samples': 17498112, 'steps': 34175, 'loss/train': 1.7169009447097778} 02/25/2022 21:50:11 - INFO - codeparrot_training - Step 34176: {'lr': 0.00012251437414616118, 'samples': 17498624, 'steps': 34176, 'loss/train': 1.7522066831588745} 02/25/2022 21:50:19 - INFO - codeparrot_training - Step 34177: {'lr': 0.00012250029929075044, 'samples': 17499136, 'steps': 34177, 'loss/train': 0.4757154583930969} 02/25/2022 21:50:22 - INFO - codeparrot_training - Step 34178: {'lr': 0.00012248622498150786, 'samples': 17499648, 'steps': 34178, 'loss/train': 1.1543480157852173} 02/25/2022 21:50:27 - INFO - codeparrot_training - Step 34179: {'lr': 0.0001224721512184939, 'samples': 17500160, 'steps': 34179, 'loss/train': 0.6923044323921204} 02/25/2022 21:50:31 - INFO - codeparrot_training - Step 34180: {'lr': 0.0001224580780017687, 'samples': 17500672, 'steps': 34180, 'loss/train': 2.0192697048187256} 02/25/2022 21:50:37 - INFO - codeparrot_training - Step 34181: {'lr': 0.00012244400533139264, 'samples': 17501184, 'steps': 34181, 'loss/train': 1.540331482887268} 02/25/2022 21:50:40 - INFO - codeparrot_training - Step 34182: {'lr': 0.00012242993320742585, 'samples': 17501696, 'steps': 34182, 'loss/train': 1.4672720432281494} 02/25/2022 21:50:45 - INFO - codeparrot_training - Step 34183: {'lr': 0.00012241586162992893, 'samples': 17502208, 'steps': 34183, 'loss/train': 1.8776568174362183} 02/25/2022 21:50:49 - INFO - codeparrot_training - Step 34184: {'lr': 0.00012240179059896171, 'samples': 17502720, 'steps': 34184, 'loss/train': 1.552356243133545} 02/25/2022 21:50:54 - INFO - codeparrot_training - Step 34185: {'lr': 0.00012238772011458486, 'samples': 17503232, 'steps': 34185, 'loss/train': 0.6263514161109924} 02/25/2022 21:50:58 - INFO - codeparrot_training - Step 34186: {'lr': 0.00012237365017685835, 'samples': 17503744, 'steps': 34186, 'loss/train': 1.6560841798782349} 02/25/2022 21:51:05 - INFO - codeparrot_training - Step 34187: {'lr': 0.0001223595807858428, 'samples': 17504256, 'steps': 34187, 'loss/train': 0.9910959601402283} 02/25/2022 21:51:08 - INFO - codeparrot_training - Step 34188: {'lr': 0.00012234551194159808, 'samples': 17504768, 'steps': 34188, 'loss/train': 1.9278590679168701} 02/25/2022 21:51:14 - INFO - codeparrot_training - Step 34189: {'lr': 0.00012233144364418475, 'samples': 17505280, 'steps': 34189, 'loss/train': 2.078850030899048} 02/25/2022 21:51:17 - INFO - codeparrot_training - Step 34190: {'lr': 0.0001223173758936629, 'samples': 17505792, 'steps': 34190, 'loss/train': 1.0764309167861938} 02/25/2022 21:51:23 - INFO - codeparrot_training - Step 34191: {'lr': 0.00012230330869009292, 'samples': 17506304, 'steps': 34191, 'loss/train': 1.2835830450057983} 02/25/2022 21:51:26 - INFO - codeparrot_training - Step 34192: {'lr': 0.00012228924203353507, 'samples': 17506816, 'steps': 34192, 'loss/train': 0.9146235585212708} 02/25/2022 21:51:32 - INFO - codeparrot_training - Step 34193: {'lr': 0.00012227517592404953, 'samples': 17507328, 'steps': 34193, 'loss/train': 1.3004958629608154} 02/25/2022 21:51:35 - INFO - codeparrot_training - Step 34194: {'lr': 0.00012226111036169648, 'samples': 17507840, 'steps': 34194, 'loss/train': 1.843943476676941} 02/25/2022 21:51:42 - INFO - codeparrot_training - Step 34195: {'lr': 0.00012224704534653636, 'samples': 17508352, 'steps': 34195, 'loss/train': 0.8542377948760986} 02/25/2022 21:51:45 - INFO - codeparrot_training - Step 34196: {'lr': 0.00012223298087862936, 'samples': 17508864, 'steps': 34196, 'loss/train': 3.116443157196045} 02/25/2022 21:51:51 - INFO - codeparrot_training - Step 34197: {'lr': 0.00012221891695803565, 'samples': 17509376, 'steps': 34197, 'loss/train': 0.0853521078824997} 02/25/2022 21:51:54 - INFO - codeparrot_training - Step 34198: {'lr': 0.00012220485358481548, 'samples': 17509888, 'steps': 34198, 'loss/train': 2.267965316772461} 02/25/2022 21:52:00 - INFO - codeparrot_training - Step 34199: {'lr': 0.00012219079075902921, 'samples': 17510400, 'steps': 34199, 'loss/train': 1.4474376440048218} 02/25/2022 21:52:03 - INFO - codeparrot_training - Step 34200: {'lr': 0.00012217672848073701, 'samples': 17510912, 'steps': 34200, 'loss/train': 0.03651644289493561} 02/25/2022 21:52:09 - INFO - codeparrot_training - Step 34201: {'lr': 0.0001221626667499991, 'samples': 17511424, 'steps': 34201, 'loss/train': 0.019949259236454964} 02/25/2022 21:52:12 - INFO - codeparrot_training - Step 34202: {'lr': 0.00012214860556687568, 'samples': 17511936, 'steps': 34202, 'loss/train': 2.273759365081787} 02/25/2022 21:52:18 - INFO - codeparrot_training - Step 34203: {'lr': 0.0001221345449314271, 'samples': 17512448, 'steps': 34203, 'loss/train': 2.4573237895965576} 02/25/2022 21:52:24 - INFO - codeparrot_training - Step 34204: {'lr': 0.00012212048484371344, 'samples': 17512960, 'steps': 34204, 'loss/train': 1.961094856262207} 02/25/2022 21:52:27 - INFO - codeparrot_training - Step 34205: {'lr': 0.0001221064253037952, 'samples': 17513472, 'steps': 34205, 'loss/train': 1.192826747894287} 02/25/2022 21:52:33 - INFO - codeparrot_training - Step 34206: {'lr': 0.00012209236631173227, 'samples': 17513984, 'steps': 34206, 'loss/train': 0.6642242074012756} 02/25/2022 21:52:36 - INFO - codeparrot_training - Step 34207: {'lr': 0.0001220783078675851, 'samples': 17514496, 'steps': 34207, 'loss/train': 2.4273128509521484} 02/25/2022 21:52:42 - INFO - codeparrot_training - Step 34208: {'lr': 0.00012206424997141371, 'samples': 17515008, 'steps': 34208, 'loss/train': 2.315173625946045} 02/25/2022 21:52:45 - INFO - codeparrot_training - Step 34209: {'lr': 0.00012205019262327868, 'samples': 17515520, 'steps': 34209, 'loss/train': 0.1208794116973877} 02/25/2022 21:52:51 - INFO - codeparrot_training - Step 34210: {'lr': 0.00012203613582323978, 'samples': 17516032, 'steps': 34210, 'loss/train': 0.37559983134269714} 02/25/2022 21:52:54 - INFO - codeparrot_training - Step 34211: {'lr': 0.00012202207957135755, 'samples': 17516544, 'steps': 34211, 'loss/train': 1.8713535070419312} 02/25/2022 21:53:01 - INFO - codeparrot_training - Step 34212: {'lr': 0.000122008023867692, 'samples': 17517056, 'steps': 34212, 'loss/train': 2.0705177783966064} 02/25/2022 21:53:04 - INFO - codeparrot_training - Step 34213: {'lr': 0.00012199396871230362, 'samples': 17517568, 'steps': 34213, 'loss/train': 1.9738880395889282} 02/25/2022 21:53:10 - INFO - codeparrot_training - Step 34214: {'lr': 0.00012197991410525225, 'samples': 17518080, 'steps': 34214, 'loss/train': 2.5504486560821533} 02/25/2022 21:53:13 - INFO - codeparrot_training - Step 34215: {'lr': 0.00012196586004659835, 'samples': 17518592, 'steps': 34215, 'loss/train': 1.540117621421814} 02/25/2022 21:53:19 - INFO - codeparrot_training - Step 34216: {'lr': 0.00012195180653640206, 'samples': 17519104, 'steps': 34216, 'loss/train': 1.542211651802063} 02/25/2022 21:53:22 - INFO - codeparrot_training - Step 34217: {'lr': 0.00012193775357472348, 'samples': 17519616, 'steps': 34217, 'loss/train': 0.9172749519348145} 02/25/2022 21:53:28 - INFO - codeparrot_training - Step 34218: {'lr': 0.00012192370116162299, 'samples': 17520128, 'steps': 34218, 'loss/train': 1.5847469568252563} 02/25/2022 21:53:31 - INFO - codeparrot_training - Step 34219: {'lr': 0.0001219096492971607, 'samples': 17520640, 'steps': 34219, 'loss/train': 2.484056234359741} 02/25/2022 21:53:37 - INFO - codeparrot_training - Step 34220: {'lr': 0.00012189559798139682, 'samples': 17521152, 'steps': 34220, 'loss/train': 2.0310730934143066} 02/25/2022 21:53:40 - INFO - codeparrot_training - Step 34221: {'lr': 0.0001218815472143914, 'samples': 17521664, 'steps': 34221, 'loss/train': 1.2230699062347412} 02/25/2022 21:53:47 - INFO - codeparrot_training - Step 34222: {'lr': 0.00012186749699620484, 'samples': 17522176, 'steps': 34222, 'loss/train': 1.8630653619766235} 02/25/2022 21:53:51 - INFO - codeparrot_training - Step 34223: {'lr': 0.00012185344732689723, 'samples': 17522688, 'steps': 34223, 'loss/train': 0.0635826364159584} 02/25/2022 21:53:56 - INFO - codeparrot_training - Step 34224: {'lr': 0.00012183939820652878, 'samples': 17523200, 'steps': 34224, 'loss/train': 0.801608681678772} 02/25/2022 21:54:00 - INFO - codeparrot_training - Step 34225: {'lr': 0.00012182534963515956, 'samples': 17523712, 'steps': 34225, 'loss/train': 0.1870885193347931} 02/25/2022 21:54:05 - INFO - codeparrot_training - Step 34226: {'lr': 0.00012181130161284992, 'samples': 17524224, 'steps': 34226, 'loss/train': 0.91904217004776} 02/25/2022 21:54:09 - INFO - codeparrot_training - Step 34227: {'lr': 0.00012179725413965997, 'samples': 17524736, 'steps': 34227, 'loss/train': 1.6117677688598633} 02/25/2022 21:54:14 - INFO - codeparrot_training - Step 34228: {'lr': 0.0001217832072156499, 'samples': 17525248, 'steps': 34228, 'loss/train': 0.7262179851531982} 02/25/2022 21:54:18 - INFO - codeparrot_training - Step 34229: {'lr': 0.00012176916084087971, 'samples': 17525760, 'steps': 34229, 'loss/train': 1.738732933998108} 02/25/2022 21:54:23 - INFO - codeparrot_training - Step 34230: {'lr': 0.00012175511501540984, 'samples': 17526272, 'steps': 34230, 'loss/train': 1.9498666524887085} 02/25/2022 21:54:27 - INFO - codeparrot_training - Step 34231: {'lr': 0.00012174106973930035, 'samples': 17526784, 'steps': 34231, 'loss/train': 0.5765262842178345} 02/25/2022 21:54:33 - INFO - codeparrot_training - Step 34232: {'lr': 0.00012172702501261138, 'samples': 17527296, 'steps': 34232, 'loss/train': 1.3165740966796875} 02/25/2022 21:54:37 - INFO - codeparrot_training - Step 34233: {'lr': 0.000121712980835403, 'samples': 17527808, 'steps': 34233, 'loss/train': 1.34809410572052} 02/25/2022 21:54:42 - INFO - codeparrot_training - Step 34234: {'lr': 0.00012169893720773556, 'samples': 17528320, 'steps': 34234, 'loss/train': 2.336918354034424} 02/25/2022 21:54:46 - INFO - codeparrot_training - Step 34235: {'lr': 0.00012168489412966916, 'samples': 17528832, 'steps': 34235, 'loss/train': 1.5499122142791748} 02/25/2022 21:54:51 - INFO - codeparrot_training - Step 34236: {'lr': 0.0001216708516012639, 'samples': 17529344, 'steps': 34236, 'loss/train': 1.5450172424316406} 02/25/2022 21:54:55 - INFO - codeparrot_training - Step 34237: {'lr': 0.00012165680962257988, 'samples': 17529856, 'steps': 34237, 'loss/train': 1.0320149660110474} 02/25/2022 21:55:00 - INFO - codeparrot_training - Step 34238: {'lr': 0.00012164276819367742, 'samples': 17530368, 'steps': 34238, 'loss/train': 1.023850917816162} 02/25/2022 21:55:04 - INFO - codeparrot_training - Step 34239: {'lr': 0.0001216287273146165, 'samples': 17530880, 'steps': 34239, 'loss/train': 1.1832797527313232} 02/25/2022 21:55:09 - INFO - codeparrot_training - Step 34240: {'lr': 0.00012161468698545755, 'samples': 17531392, 'steps': 34240, 'loss/train': 1.522051215171814} 02/25/2022 21:55:13 - INFO - codeparrot_training - Step 34241: {'lr': 0.00012160064720626029, 'samples': 17531904, 'steps': 34241, 'loss/train': 1.848900318145752} 02/25/2022 21:55:19 - INFO - codeparrot_training - Step 34242: {'lr': 0.0001215866079770852, 'samples': 17532416, 'steps': 34242, 'loss/train': 4.349810600280762} 02/25/2022 21:55:23 - INFO - codeparrot_training - Step 34243: {'lr': 0.00012157256929799223, 'samples': 17532928, 'steps': 34243, 'loss/train': 1.546586275100708} 02/25/2022 21:55:28 - INFO - codeparrot_training - Step 34244: {'lr': 0.00012155853116904178, 'samples': 17533440, 'steps': 34244, 'loss/train': 2.323444128036499} 02/25/2022 21:55:32 - INFO - codeparrot_training - Step 34245: {'lr': 0.00012154449359029357, 'samples': 17533952, 'steps': 34245, 'loss/train': 1.5288819074630737} 02/25/2022 21:55:37 - INFO - codeparrot_training - Step 34246: {'lr': 0.00012153045656180808, 'samples': 17534464, 'steps': 34246, 'loss/train': 1.9346307516098022} 02/25/2022 21:55:41 - INFO - codeparrot_training - Step 34247: {'lr': 0.00012151642008364524, 'samples': 17534976, 'steps': 34247, 'loss/train': 2.1254489421844482} 02/25/2022 21:55:46 - INFO - codeparrot_training - Step 34248: {'lr': 0.00012150238415586542, 'samples': 17535488, 'steps': 34248, 'loss/train': 1.7623099088668823} 02/25/2022 21:55:50 - INFO - codeparrot_training - Step 34249: {'lr': 0.00012148834877852841, 'samples': 17536000, 'steps': 34249, 'loss/train': 1.994359016418457} 02/25/2022 21:55:55 - INFO - codeparrot_training - Step 34250: {'lr': 0.0001214743139516946, 'samples': 17536512, 'steps': 34250, 'loss/train': 1.7585986852645874} 02/25/2022 21:55:59 - INFO - codeparrot_training - Step 34251: {'lr': 0.00012146027967542391, 'samples': 17537024, 'steps': 34251, 'loss/train': 1.7589515447616577} 02/25/2022 21:56:04 - INFO - codeparrot_training - Step 34252: {'lr': 0.00012144624594977669, 'samples': 17537536, 'steps': 34252, 'loss/train': 2.556988000869751} 02/25/2022 21:56:08 - INFO - codeparrot_training - Step 34253: {'lr': 0.00012143221277481289, 'samples': 17538048, 'steps': 34253, 'loss/train': 1.621940016746521} 02/25/2022 21:56:14 - INFO - codeparrot_training - Step 34254: {'lr': 0.0001214181801505927, 'samples': 17538560, 'steps': 34254, 'loss/train': 2.0679194927215576} 02/25/2022 21:56:17 - INFO - codeparrot_training - Step 34255: {'lr': 0.0001214041480771761, 'samples': 17539072, 'steps': 34255, 'loss/train': 2.2575337886810303} 02/25/2022 21:56:22 - INFO - codeparrot_training - Step 34256: {'lr': 0.00012139011655462338, 'samples': 17539584, 'steps': 34256, 'loss/train': 1.7101577520370483} 02/25/2022 21:56:26 - INFO - codeparrot_training - Step 34257: {'lr': 0.00012137608558299456, 'samples': 17540096, 'steps': 34257, 'loss/train': 0.5349771976470947} 02/25/2022 21:56:31 - INFO - codeparrot_training - Step 34258: {'lr': 0.00012136205516234974, 'samples': 17540608, 'steps': 34258, 'loss/train': 1.7355486154556274} 02/25/2022 21:56:35 - INFO - codeparrot_training - Step 34259: {'lr': 0.00012134802529274897, 'samples': 17541120, 'steps': 34259, 'loss/train': 0.9616206288337708} 02/25/2022 21:56:41 - INFO - codeparrot_training - Step 34260: {'lr': 0.00012133399597425246, 'samples': 17541632, 'steps': 34260, 'loss/train': 2.388927459716797} 02/25/2022 21:56:45 - INFO - codeparrot_training - Step 34261: {'lr': 0.00012131996720692026, 'samples': 17542144, 'steps': 34261, 'loss/train': 1.3599094152450562} 02/25/2022 21:56:50 - INFO - codeparrot_training - Step 34262: {'lr': 0.00012130593899081249, 'samples': 17542656, 'steps': 34262, 'loss/train': 2.1522176265716553} 02/25/2022 21:56:54 - INFO - codeparrot_training - Step 34263: {'lr': 0.00012129191132598919, 'samples': 17543168, 'steps': 34263, 'loss/train': 2.4992573261260986} 02/25/2022 21:56:59 - INFO - codeparrot_training - Step 34264: {'lr': 0.00012127788421251038, 'samples': 17543680, 'steps': 34264, 'loss/train': 2.077789068222046} 02/25/2022 21:57:03 - INFO - codeparrot_training - Step 34265: {'lr': 0.00012126385765043635, 'samples': 17544192, 'steps': 34265, 'loss/train': 1.3943564891815186} 02/25/2022 21:57:08 - INFO - codeparrot_training - Step 34266: {'lr': 0.00012124983163982706, 'samples': 17544704, 'steps': 34266, 'loss/train': 8.703642845153809} 02/25/2022 21:57:12 - INFO - codeparrot_training - Step 34267: {'lr': 0.00012123580618074262, 'samples': 17545216, 'steps': 34267, 'loss/train': 2.1163244247436523} 02/25/2022 21:57:17 - INFO - codeparrot_training - Step 34268: {'lr': 0.00012122178127324298, 'samples': 17545728, 'steps': 34268, 'loss/train': 1.6408638954162598} 02/25/2022 21:57:21 - INFO - codeparrot_training - Step 34269: {'lr': 0.00012120775691738845, 'samples': 17546240, 'steps': 34269, 'loss/train': 2.154496192932129} 02/25/2022 21:57:27 - INFO - codeparrot_training - Step 34270: {'lr': 0.000121193733113239, 'samples': 17546752, 'steps': 34270, 'loss/train': 2.46219539642334} 02/25/2022 21:57:31 - INFO - codeparrot_training - Step 34271: {'lr': 0.0001211797098608547, 'samples': 17547264, 'steps': 34271, 'loss/train': 1.1195275783538818} 02/25/2022 21:57:36 - INFO - codeparrot_training - Step 34272: {'lr': 0.0001211656871602955, 'samples': 17547776, 'steps': 34272, 'loss/train': 0.8301324844360352} 02/25/2022 21:57:40 - INFO - codeparrot_training - Step 34273: {'lr': 0.00012115166501162168, 'samples': 17548288, 'steps': 34273, 'loss/train': 0.9998868703842163} 02/25/2022 21:57:45 - INFO - codeparrot_training - Step 34274: {'lr': 0.00012113764341489323, 'samples': 17548800, 'steps': 34274, 'loss/train': 1.7557333707809448} 02/25/2022 21:57:49 - INFO - codeparrot_training - Step 34275: {'lr': 0.00012112362237017018, 'samples': 17549312, 'steps': 34275, 'loss/train': 2.2235615253448486} 02/25/2022 21:57:54 - INFO - codeparrot_training - Step 34276: {'lr': 0.00012110960187751255, 'samples': 17549824, 'steps': 34276, 'loss/train': 1.8932520151138306} 02/25/2022 21:57:58 - INFO - codeparrot_training - Step 34277: {'lr': 0.0001210955819369805, 'samples': 17550336, 'steps': 34277, 'loss/train': 3.0477845668792725} 02/25/2022 21:58:03 - INFO - codeparrot_training - Step 34278: {'lr': 0.000121081562548634, 'samples': 17550848, 'steps': 34278, 'loss/train': 1.579720377922058} 02/25/2022 21:58:07 - INFO - codeparrot_training - Step 34279: {'lr': 0.00012106754371253331, 'samples': 17551360, 'steps': 34279, 'loss/train': 2.263288974761963} 02/25/2022 21:58:13 - INFO - codeparrot_training - Step 34280: {'lr': 0.00012105352542873815, 'samples': 17551872, 'steps': 34280, 'loss/train': 1.1526834964752197} 02/25/2022 21:58:17 - INFO - codeparrot_training - Step 34281: {'lr': 0.00012103950769730881, 'samples': 17552384, 'steps': 34281, 'loss/train': 0.21607467532157898} 02/25/2022 21:58:22 - INFO - codeparrot_training - Step 34282: {'lr': 0.00012102549051830521, 'samples': 17552896, 'steps': 34282, 'loss/train': 1.692558765411377} 02/25/2022 21:58:28 - INFO - codeparrot_training - Step 34283: {'lr': 0.00012101147389178762, 'samples': 17553408, 'steps': 34283, 'loss/train': 1.4710392951965332} 02/25/2022 21:58:31 - INFO - codeparrot_training - Step 34284: {'lr': 0.00012099745781781574, 'samples': 17553920, 'steps': 34284, 'loss/train': 1.6302279233932495} 02/25/2022 21:58:37 - INFO - codeparrot_training - Step 34285: {'lr': 0.00012098344229644989, 'samples': 17554432, 'steps': 34285, 'loss/train': 2.4527428150177} 02/25/2022 21:58:40 - INFO - codeparrot_training - Step 34286: {'lr': 0.0001209694273277499, 'samples': 17554944, 'steps': 34286, 'loss/train': 0.20489583909511566} 02/25/2022 21:58:46 - INFO - codeparrot_training - Step 34287: {'lr': 0.000120955412911776, 'samples': 17555456, 'steps': 34287, 'loss/train': 2.0604114532470703} 02/25/2022 21:58:49 - INFO - codeparrot_training - Step 34288: {'lr': 0.00012094139904858814, 'samples': 17555968, 'steps': 34288, 'loss/train': 2.0850396156311035} 02/25/2022 21:58:55 - INFO - codeparrot_training - Step 34289: {'lr': 0.00012092738573824636, 'samples': 17556480, 'steps': 34289, 'loss/train': 2.975572109222412} 02/25/2022 21:58:59 - INFO - codeparrot_training - Step 34290: {'lr': 0.00012091337298081054, 'samples': 17556992, 'steps': 34290, 'loss/train': 1.118453025817871} 02/25/2022 21:59:04 - INFO - codeparrot_training - Step 34291: {'lr': 0.00012089936077634098, 'samples': 17557504, 'steps': 34291, 'loss/train': 2.4356935024261475} 02/25/2022 21:59:08 - INFO - codeparrot_training - Step 34292: {'lr': 0.00012088534912489754, 'samples': 17558016, 'steps': 34292, 'loss/train': 2.5965936183929443} 02/25/2022 21:59:13 - INFO - codeparrot_training - Step 34293: {'lr': 0.00012087133802654027, 'samples': 17558528, 'steps': 34293, 'loss/train': 1.1163548231124878} 02/25/2022 21:59:17 - INFO - codeparrot_training - Step 34294: {'lr': 0.00012085732748132908, 'samples': 17559040, 'steps': 34294, 'loss/train': 1.2215179204940796} 02/25/2022 21:59:22 - INFO - codeparrot_training - Step 34295: {'lr': 0.0001208433174893242, 'samples': 17559552, 'steps': 34295, 'loss/train': 1.8212321996688843} 02/25/2022 21:59:26 - INFO - codeparrot_training - Step 34296: {'lr': 0.00012082930805058553, 'samples': 17560064, 'steps': 34296, 'loss/train': 1.0686501264572144} 02/25/2022 21:59:31 - INFO - codeparrot_training - Step 34297: {'lr': 0.00012081529916517309, 'samples': 17560576, 'steps': 34297, 'loss/train': 1.660291075706482} 02/25/2022 21:59:35 - INFO - codeparrot_training - Step 34298: {'lr': 0.00012080129083314682, 'samples': 17561088, 'steps': 34298, 'loss/train': 2.1139681339263916} 02/25/2022 21:59:40 - INFO - codeparrot_training - Step 34299: {'lr': 0.00012078728305456685, 'samples': 17561600, 'steps': 34299, 'loss/train': 0.8264246582984924} 02/25/2022 21:59:44 - INFO - codeparrot_training - Step 34300: {'lr': 0.00012077327582949313, 'samples': 17562112, 'steps': 34300, 'loss/train': 1.4739696979522705} 02/25/2022 21:59:49 - INFO - codeparrot_training - Step 34301: {'lr': 0.00012075926915798568, 'samples': 17562624, 'steps': 34301, 'loss/train': 2.240708827972412} 02/25/2022 21:59:53 - INFO - codeparrot_training - Step 34302: {'lr': 0.00012074526304010438, 'samples': 17563136, 'steps': 34302, 'loss/train': 2.087946891784668} 02/25/2022 22:00:00 - INFO - codeparrot_training - Step 34303: {'lr': 0.00012073125747590943, 'samples': 17563648, 'steps': 34303, 'loss/train': 1.0485347509384155} 02/25/2022 22:00:03 - INFO - codeparrot_training - Step 34304: {'lr': 0.00012071725246546073, 'samples': 17564160, 'steps': 34304, 'loss/train': 0.7819700837135315} 02/25/2022 22:00:09 - INFO - codeparrot_training - Step 34305: {'lr': 0.00012070324800881826, 'samples': 17564672, 'steps': 34305, 'loss/train': 0.3822716772556305} 02/25/2022 22:00:12 - INFO - codeparrot_training - Step 34306: {'lr': 0.00012068924410604194, 'samples': 17565184, 'steps': 34306, 'loss/train': 1.613389492034912} 02/25/2022 22:00:18 - INFO - codeparrot_training - Step 34307: {'lr': 0.00012067524075719191, 'samples': 17565696, 'steps': 34307, 'loss/train': 1.0227255821228027} 02/25/2022 22:00:21 - INFO - codeparrot_training - Step 34308: {'lr': 0.00012066123796232808, 'samples': 17566208, 'steps': 34308, 'loss/train': 1.2361652851104736} 02/25/2022 22:00:27 - INFO - codeparrot_training - Step 34309: {'lr': 0.00012064723572151045, 'samples': 17566720, 'steps': 34309, 'loss/train': 0.935239851474762} 02/25/2022 22:00:30 - INFO - codeparrot_training - Step 34310: {'lr': 0.00012063323403479901, 'samples': 17567232, 'steps': 34310, 'loss/train': 1.5385384559631348} 02/25/2022 22:00:36 - INFO - codeparrot_training - Step 34311: {'lr': 0.0001206192329022536, 'samples': 17567744, 'steps': 34311, 'loss/train': 2.8794918060302734} 02/25/2022 22:00:40 - INFO - codeparrot_training - Step 34312: {'lr': 0.00012060523232393438, 'samples': 17568256, 'steps': 34312, 'loss/train': 0.7928222417831421} 02/25/2022 22:00:45 - INFO - codeparrot_training - Step 34313: {'lr': 0.00012059123229990121, 'samples': 17568768, 'steps': 34313, 'loss/train': 1.6463814973831177} 02/25/2022 22:00:49 - INFO - codeparrot_training - Step 34314: {'lr': 0.00012057723283021426, 'samples': 17569280, 'steps': 34314, 'loss/train': 1.168013095855713} 02/25/2022 22:00:56 - INFO - codeparrot_training - Step 34315: {'lr': 0.0001205632339149332, 'samples': 17569792, 'steps': 34315, 'loss/train': 1.4635422229766846} 02/25/2022 22:00:59 - INFO - codeparrot_training - Step 34316: {'lr': 0.0001205492355541182, 'samples': 17570304, 'steps': 34316, 'loss/train': 2.195033311843872} 02/25/2022 22:01:05 - INFO - codeparrot_training - Step 34317: {'lr': 0.0001205352377478291, 'samples': 17570816, 'steps': 34317, 'loss/train': 8.661101341247559} 02/25/2022 22:01:08 - INFO - codeparrot_training - Step 34318: {'lr': 0.00012052124049612611, 'samples': 17571328, 'steps': 34318, 'loss/train': 1.613338828086853} 02/25/2022 22:01:14 - INFO - codeparrot_training - Step 34319: {'lr': 0.00012050724379906883, 'samples': 17571840, 'steps': 34319, 'loss/train': 2.1264991760253906} 02/25/2022 22:01:17 - INFO - codeparrot_training - Step 34320: {'lr': 0.00012049324765671748, 'samples': 17572352, 'steps': 34320, 'loss/train': 1.6607111692428589} 02/25/2022 22:01:23 - INFO - codeparrot_training - Step 34321: {'lr': 0.00012047925206913185, 'samples': 17572864, 'steps': 34321, 'loss/train': 1.7710572481155396} 02/25/2022 22:01:26 - INFO - codeparrot_training - Step 34322: {'lr': 0.00012046525703637217, 'samples': 17573376, 'steps': 34322, 'loss/train': 2.024055242538452} 02/25/2022 22:01:32 - INFO - codeparrot_training - Step 34323: {'lr': 0.000120451262558498, 'samples': 17573888, 'steps': 34323, 'loss/train': 2.007153272628784} 02/25/2022 22:01:35 - INFO - codeparrot_training - Step 34324: {'lr': 0.00012043726863556959, 'samples': 17574400, 'steps': 34324, 'loss/train': 3.044768810272217} 02/25/2022 22:01:42 - INFO - codeparrot_training - Step 34325: {'lr': 0.00012042327526764668, 'samples': 17574912, 'steps': 34325, 'loss/train': 1.6636451482772827} 02/25/2022 22:01:45 - INFO - codeparrot_training - Step 34326: {'lr': 0.00012040928245478941, 'samples': 17575424, 'steps': 34326, 'loss/train': 1.5949853658676147} 02/25/2022 22:01:51 - INFO - codeparrot_training - Step 34327: {'lr': 0.00012039529019705762, 'samples': 17575936, 'steps': 34327, 'loss/train': 2.127155303955078} 02/25/2022 22:01:54 - INFO - codeparrot_training - Step 34328: {'lr': 0.00012038129849451124, 'samples': 17576448, 'steps': 34328, 'loss/train': 0.9926344752311707} 02/25/2022 22:02:00 - INFO - codeparrot_training - Step 34329: {'lr': 0.00012036730734721014, 'samples': 17576960, 'steps': 34329, 'loss/train': 2.0554497241973877} 02/25/2022 22:02:03 - INFO - codeparrot_training - Step 34330: {'lr': 0.00012035331675521442, 'samples': 17577472, 'steps': 34330, 'loss/train': 1.907273769378662} 02/25/2022 22:02:09 - INFO - codeparrot_training - Step 34331: {'lr': 0.00012033932671858394, 'samples': 17577984, 'steps': 34331, 'loss/train': 2.640911340713501} 02/25/2022 22:02:12 - INFO - codeparrot_training - Step 34332: {'lr': 0.00012032533723737859, 'samples': 17578496, 'steps': 34332, 'loss/train': 0.9879363775253296} 02/25/2022 22:02:18 - INFO - codeparrot_training - Step 34333: {'lr': 0.00012031134831165824, 'samples': 17579008, 'steps': 34333, 'loss/train': 1.335127592086792} 02/25/2022 22:02:21 - INFO - codeparrot_training - Step 34334: {'lr': 0.00012029735994148297, 'samples': 17579520, 'steps': 34334, 'loss/train': 0.9870880842208862} 02/25/2022 22:02:28 - INFO - codeparrot_training - Step 34335: {'lr': 0.00012028337212691254, 'samples': 17580032, 'steps': 34335, 'loss/train': 1.6604045629501343} 02/25/2022 22:02:31 - INFO - codeparrot_training - Step 34336: {'lr': 0.00012026938486800714, 'samples': 17580544, 'steps': 34336, 'loss/train': 2.1227593421936035} 02/25/2022 22:02:37 - INFO - codeparrot_training - Step 34337: {'lr': 0.00012025539816482634, 'samples': 17581056, 'steps': 34337, 'loss/train': 1.1126620769500732} 02/25/2022 22:02:40 - INFO - codeparrot_training - Step 34338: {'lr': 0.00012024141201743027, 'samples': 17581568, 'steps': 34338, 'loss/train': 0.6179150938987732} 02/25/2022 22:02:46 - INFO - codeparrot_training - Step 34339: {'lr': 0.00012022742642587872, 'samples': 17582080, 'steps': 34339, 'loss/train': 1.9983370304107666} 02/25/2022 22:02:49 - INFO - codeparrot_training - Step 34340: {'lr': 0.00012021344139023186, 'samples': 17582592, 'steps': 34340, 'loss/train': 1.8165253400802612} 02/25/2022 22:02:55 - INFO - codeparrot_training - Step 34341: {'lr': 0.00012019945691054918, 'samples': 17583104, 'steps': 34341, 'loss/train': 1.3494136333465576} 02/25/2022 22:02:58 - INFO - codeparrot_training - Step 34342: {'lr': 0.00012018547298689092, 'samples': 17583616, 'steps': 34342, 'loss/train': 2.2382593154907227} 02/25/2022 22:03:03 - INFO - codeparrot_training - Step 34343: {'lr': 0.0001201714896193169, 'samples': 17584128, 'steps': 34343, 'loss/train': 1.6111931800842285} 02/25/2022 22:03:07 - INFO - codeparrot_training - Step 34344: {'lr': 0.00012015750680788698, 'samples': 17584640, 'steps': 34344, 'loss/train': 1.4953608512878418} 02/25/2022 22:03:13 - INFO - codeparrot_training - Step 34345: {'lr': 0.00012014352455266094, 'samples': 17585152, 'steps': 34345, 'loss/train': 1.877808690071106} 02/25/2022 22:03:17 - INFO - codeparrot_training - Step 34346: {'lr': 0.00012012954285369895, 'samples': 17585664, 'steps': 34346, 'loss/train': 2.4097847938537598} 02/25/2022 22:03:22 - INFO - codeparrot_training - Step 34347: {'lr': 0.00012011556171106075, 'samples': 17586176, 'steps': 34347, 'loss/train': 0.13597442209720612} 02/25/2022 22:03:26 - INFO - codeparrot_training - Step 34348: {'lr': 0.00012010158112480612, 'samples': 17586688, 'steps': 34348, 'loss/train': 2.021965503692627} 02/25/2022 22:03:31 - INFO - codeparrot_training - Step 34349: {'lr': 0.0001200876010949952, 'samples': 17587200, 'steps': 34349, 'loss/train': 1.5330291986465454} 02/25/2022 22:03:35 - INFO - codeparrot_training - Step 34350: {'lr': 0.00012007362162168772, 'samples': 17587712, 'steps': 34350, 'loss/train': 2.091512680053711} 02/25/2022 22:03:40 - INFO - codeparrot_training - Step 34351: {'lr': 0.0001200596427049436, 'samples': 17588224, 'steps': 34351, 'loss/train': 1.4846479892730713} 02/25/2022 22:03:44 - INFO - codeparrot_training - Step 34352: {'lr': 0.00012004566434482261, 'samples': 17588736, 'steps': 34352, 'loss/train': 1.3995633125305176} 02/25/2022 22:03:49 - INFO - codeparrot_training - Step 34353: {'lr': 0.00012003168654138482, 'samples': 17589248, 'steps': 34353, 'loss/train': 0.061498068273067474} 02/25/2022 22:03:53 - INFO - codeparrot_training - Step 34354: {'lr': 0.00012001770929469003, 'samples': 17589760, 'steps': 34354, 'loss/train': 1.894551157951355} 02/25/2022 22:03:58 - INFO - codeparrot_training - Step 34355: {'lr': 0.00012000373260479807, 'samples': 17590272, 'steps': 34355, 'loss/train': 1.7367933988571167} 02/25/2022 22:04:02 - INFO - codeparrot_training - Step 34356: {'lr': 0.00011998975647176876, 'samples': 17590784, 'steps': 34356, 'loss/train': 1.460475206375122} 02/25/2022 22:04:07 - INFO - codeparrot_training - Step 34357: {'lr': 0.00011997578089566228, 'samples': 17591296, 'steps': 34357, 'loss/train': 1.8767181634902954} 02/25/2022 22:04:11 - INFO - codeparrot_training - Step 34358: {'lr': 0.00011996180587653801, 'samples': 17591808, 'steps': 34358, 'loss/train': 3.2255258560180664} 02/25/2022 22:04:16 - INFO - codeparrot_training - Step 34359: {'lr': 0.00011994783141445622, 'samples': 17592320, 'steps': 34359, 'loss/train': 1.0728782415390015} 02/25/2022 22:04:20 - INFO - codeparrot_training - Step 34360: {'lr': 0.00011993385750947649, 'samples': 17592832, 'steps': 34360, 'loss/train': 1.1422830820083618} 02/25/2022 22:04:26 - INFO - codeparrot_training - Step 34361: {'lr': 0.00011991988416165894, 'samples': 17593344, 'steps': 34361, 'loss/train': 1.5575486421585083} 02/25/2022 22:04:30 - INFO - codeparrot_training - Step 34362: {'lr': 0.00011990591137106329, 'samples': 17593856, 'steps': 34362, 'loss/train': 1.9952211380004883} 02/25/2022 22:04:35 - INFO - codeparrot_training - Step 34363: {'lr': 0.0001198919391377494, 'samples': 17594368, 'steps': 34363, 'loss/train': 1.8301459550857544} 02/25/2022 22:04:39 - INFO - codeparrot_training - Step 34364: {'lr': 0.00011987796746177704, 'samples': 17594880, 'steps': 34364, 'loss/train': 1.5398815870285034} 02/25/2022 22:04:44 - INFO - codeparrot_training - Step 34365: {'lr': 0.00011986399634320625, 'samples': 17595392, 'steps': 34365, 'loss/train': 2.0846877098083496} 02/25/2022 22:04:48 - INFO - codeparrot_training - Step 34366: {'lr': 0.00011985002578209678, 'samples': 17595904, 'steps': 34366, 'loss/train': 1.6592000722885132} 02/25/2022 22:04:53 - INFO - codeparrot_training - Step 34367: {'lr': 0.00011983605577850845, 'samples': 17596416, 'steps': 34367, 'loss/train': 0.9313715696334839} 02/25/2022 22:04:57 - INFO - codeparrot_training - Step 34368: {'lr': 0.00011982208633250105, 'samples': 17596928, 'steps': 34368, 'loss/train': 2.610107660293579} 02/25/2022 22:05:02 - INFO - codeparrot_training - Step 34369: {'lr': 0.00011980811744413458, 'samples': 17597440, 'steps': 34369, 'loss/train': 0.9194834232330322} 02/25/2022 22:05:06 - INFO - codeparrot_training - Step 34370: {'lr': 0.00011979414911346878, 'samples': 17597952, 'steps': 34370, 'loss/train': 1.716711163520813} 02/25/2022 22:05:12 - INFO - codeparrot_training - Step 34371: {'lr': 0.00011978018134056353, 'samples': 17598464, 'steps': 34371, 'loss/train': 0.9407782554626465} 02/25/2022 22:05:16 - INFO - codeparrot_training - Step 34372: {'lr': 0.00011976621412547853, 'samples': 17598976, 'steps': 34372, 'loss/train': 1.4251182079315186} 02/25/2022 22:05:21 - INFO - codeparrot_training - Step 34373: {'lr': 0.00011975224746827379, 'samples': 17599488, 'steps': 34373, 'loss/train': 2.514246702194214} 02/25/2022 22:05:25 - INFO - codeparrot_training - Step 34374: {'lr': 0.000119738281369009, 'samples': 17600000, 'steps': 34374, 'loss/train': 1.043410062789917} 02/25/2022 22:05:30 - INFO - codeparrot_training - Step 34375: {'lr': 0.0001197243158277442, 'samples': 17600512, 'steps': 34375, 'loss/train': 0.3327135443687439} 02/25/2022 22:05:34 - INFO - codeparrot_training - Step 34376: {'lr': 0.0001197103508445389, 'samples': 17601024, 'steps': 34376, 'loss/train': 1.5241411924362183} 02/25/2022 22:05:39 - INFO - codeparrot_training - Step 34377: {'lr': 0.00011969638641945319, 'samples': 17601536, 'steps': 34377, 'loss/train': 2.4000070095062256} 02/25/2022 22:05:43 - INFO - codeparrot_training - Step 34378: {'lr': 0.00011968242255254669, 'samples': 17602048, 'steps': 34378, 'loss/train': 1.7739942073822021} 02/25/2022 22:05:48 - INFO - codeparrot_training - Step 34379: {'lr': 0.00011966845924387948, 'samples': 17602560, 'steps': 34379, 'loss/train': 2.6915862560272217} 02/25/2022 22:05:55 - INFO - codeparrot_training - Step 34380: {'lr': 0.00011965449649351103, 'samples': 17603072, 'steps': 34380, 'loss/train': 1.2555384635925293} 02/25/2022 22:05:58 - INFO - codeparrot_training - Step 34381: {'lr': 0.0001196405343015014, 'samples': 17603584, 'steps': 34381, 'loss/train': 1.3932793140411377} 02/25/2022 22:06:04 - INFO - codeparrot_training - Step 34382: {'lr': 0.00011962657266791027, 'samples': 17604096, 'steps': 34382, 'loss/train': 0.9549170136451721} 02/25/2022 22:06:07 - INFO - codeparrot_training - Step 34383: {'lr': 0.00011961261159279768, 'samples': 17604608, 'steps': 34383, 'loss/train': 1.380858302116394} 02/25/2022 22:06:11 - INFO - codeparrot_training - Step 34384: {'lr': 0.00011959865107622306, 'samples': 17605120, 'steps': 34384, 'loss/train': 1.988647699356079} 02/25/2022 22:06:16 - INFO - codeparrot_training - Step 34385: {'lr': 0.00011958469111824649, 'samples': 17605632, 'steps': 34385, 'loss/train': 1.6716965436935425} 02/25/2022 22:06:20 - INFO - codeparrot_training - Step 34386: {'lr': 0.00011957073171892763, 'samples': 17606144, 'steps': 34386, 'loss/train': 1.7977254390716553} 02/25/2022 22:06:25 - INFO - codeparrot_training - Step 34387: {'lr': 0.00011955677287832642, 'samples': 17606656, 'steps': 34387, 'loss/train': 1.5108684301376343} 02/25/2022 22:06:29 - INFO - codeparrot_training - Step 34388: {'lr': 0.00011954281459650257, 'samples': 17607168, 'steps': 34388, 'loss/train': 2.588124990463257} 02/25/2022 22:06:36 - INFO - codeparrot_training - Step 34389: {'lr': 0.00011952885687351589, 'samples': 17607680, 'steps': 34389, 'loss/train': 2.1757864952087402} 02/25/2022 22:06:39 - INFO - codeparrot_training - Step 34390: {'lr': 0.00011951489970942614, 'samples': 17608192, 'steps': 34390, 'loss/train': 1.5195791721343994} 02/25/2022 22:06:45 - INFO - codeparrot_training - Step 34391: {'lr': 0.00011950094310429304, 'samples': 17608704, 'steps': 34391, 'loss/train': 1.5775372982025146} 02/25/2022 22:06:48 - INFO - codeparrot_training - Step 34392: {'lr': 0.00011948698705817657, 'samples': 17609216, 'steps': 34392, 'loss/train': 1.64582097530365} 02/25/2022 22:06:54 - INFO - codeparrot_training - Step 34393: {'lr': 0.0001194730315711364, 'samples': 17609728, 'steps': 34393, 'loss/train': 0.4548017382621765} 02/25/2022 22:06:57 - INFO - codeparrot_training - Step 34394: {'lr': 0.0001194590766432323, 'samples': 17610240, 'steps': 34394, 'loss/train': 3.014071226119995} 02/25/2022 22:07:03 - INFO - codeparrot_training - Step 34395: {'lr': 0.00011944512227452398, 'samples': 17610752, 'steps': 34395, 'loss/train': 1.0806884765625} 02/25/2022 22:07:08 - INFO - codeparrot_training - Step 34396: {'lr': 0.0001194311684650714, 'samples': 17611264, 'steps': 34396, 'loss/train': 1.7794002294540405} 02/25/2022 22:07:12 - INFO - codeparrot_training - Step 34397: {'lr': 0.00011941721521493424, 'samples': 17611776, 'steps': 34397, 'loss/train': 1.7507359981536865} 02/25/2022 22:07:15 - INFO - codeparrot_training - Step 34398: {'lr': 0.00011940326252417225, 'samples': 17612288, 'steps': 34398, 'loss/train': 1.9027122259140015} 02/25/2022 22:07:22 - INFO - codeparrot_training - Step 34399: {'lr': 0.00011938931039284511, 'samples': 17612800, 'steps': 34399, 'loss/train': 1.7898634672164917} 02/25/2022 22:07:25 - INFO - codeparrot_training - Step 34400: {'lr': 0.00011937535882101281, 'samples': 17613312, 'steps': 34400, 'loss/train': 1.5655403137207031} 02/25/2022 22:07:31 - INFO - codeparrot_training - Step 34401: {'lr': 0.00011936140780873497, 'samples': 17613824, 'steps': 34401, 'loss/train': 1.9456537961959839} 02/25/2022 22:07:34 - INFO - codeparrot_training - Step 34402: {'lr': 0.00011934745735607138, 'samples': 17614336, 'steps': 34402, 'loss/train': 1.0505439043045044} 02/25/2022 22:07:40 - INFO - codeparrot_training - Step 34403: {'lr': 0.00011933350746308172, 'samples': 17614848, 'steps': 34403, 'loss/train': 1.397525668144226} 02/25/2022 22:07:43 - INFO - codeparrot_training - Step 34404: {'lr': 0.00011931955812982593, 'samples': 17615360, 'steps': 34404, 'loss/train': 1.1197293996810913} 02/25/2022 22:07:49 - INFO - codeparrot_training - Step 34405: {'lr': 0.00011930560935636361, 'samples': 17615872, 'steps': 34405, 'loss/train': 1.566070795059204} 02/25/2022 22:07:52 - INFO - codeparrot_training - Step 34406: {'lr': 0.0001192916611427546, 'samples': 17616384, 'steps': 34406, 'loss/train': 2.6954689025878906} 02/25/2022 22:07:58 - INFO - codeparrot_training - Step 34407: {'lr': 0.0001192777134890585, 'samples': 17616896, 'steps': 34407, 'loss/train': 0.8532391786575317} 02/25/2022 22:08:01 - INFO - codeparrot_training - Step 34408: {'lr': 0.00011926376639533526, 'samples': 17617408, 'steps': 34408, 'loss/train': 2.053250551223755} 02/25/2022 22:08:07 - INFO - codeparrot_training - Step 34409: {'lr': 0.00011924981986164443, 'samples': 17617920, 'steps': 34409, 'loss/train': 2.6227874755859375} 02/25/2022 22:08:11 - INFO - codeparrot_training - Step 34410: {'lr': 0.00011923587388804605, 'samples': 17618432, 'steps': 34410, 'loss/train': 3.803899049758911} 02/25/2022 22:08:16 - INFO - codeparrot_training - Step 34411: {'lr': 0.00011922192847459948, 'samples': 17618944, 'steps': 34411, 'loss/train': 0.18625019490718842} 02/25/2022 22:08:20 - INFO - codeparrot_training - Step 34412: {'lr': 0.00011920798362136472, 'samples': 17619456, 'steps': 34412, 'loss/train': 0.25925758481025696} 02/25/2022 22:08:25 - INFO - codeparrot_training - Step 34413: {'lr': 0.00011919403932840136, 'samples': 17619968, 'steps': 34413, 'loss/train': 1.047576904296875} 02/25/2022 22:08:29 - INFO - codeparrot_training - Step 34414: {'lr': 0.00011918009559576937, 'samples': 17620480, 'steps': 34414, 'loss/train': 1.256148099899292} 02/25/2022 22:08:34 - INFO - codeparrot_training - Step 34415: {'lr': 0.00011916615242352814, 'samples': 17620992, 'steps': 34415, 'loss/train': 1.1419812440872192} 02/25/2022 22:08:38 - INFO - codeparrot_training - Step 34416: {'lr': 0.00011915220981173763, 'samples': 17621504, 'steps': 34416, 'loss/train': 1.4046618938446045} 02/25/2022 22:08:43 - INFO - codeparrot_training - Step 34417: {'lr': 0.00011913826776045741, 'samples': 17622016, 'steps': 34417, 'loss/train': 1.9058088064193726} 02/25/2022 22:08:47 - INFO - codeparrot_training - Step 34418: {'lr': 0.00011912432626974751, 'samples': 17622528, 'steps': 34418, 'loss/train': 1.6757359504699707} 02/25/2022 22:08:53 - INFO - codeparrot_training - Step 34419: {'lr': 0.00011911038533966726, 'samples': 17623040, 'steps': 34419, 'loss/train': 1.6711442470550537} 02/25/2022 22:08:57 - INFO - codeparrot_training - Step 34420: {'lr': 0.00011909644497027663, 'samples': 17623552, 'steps': 34420, 'loss/train': 1.3533649444580078} 02/25/2022 22:09:02 - INFO - codeparrot_training - Step 34421: {'lr': 0.00011908250516163516, 'samples': 17624064, 'steps': 34421, 'loss/train': 1.3884401321411133} 02/25/2022 22:09:06 - INFO - codeparrot_training - Step 34422: {'lr': 0.00011906856591380278, 'samples': 17624576, 'steps': 34422, 'loss/train': 0.41252240538597107} 02/25/2022 22:09:11 - INFO - codeparrot_training - Step 34423: {'lr': 0.0001190546272268391, 'samples': 17625088, 'steps': 34423, 'loss/train': 2.0497453212738037} 02/25/2022 22:09:15 - INFO - codeparrot_training - Step 34424: {'lr': 0.00011904068910080379, 'samples': 17625600, 'steps': 34424, 'loss/train': 0.7143895030021667} 02/25/2022 22:09:20 - INFO - codeparrot_training - Step 34425: {'lr': 0.00011902675153575651, 'samples': 17626112, 'steps': 34425, 'loss/train': 2.6272928714752197} 02/25/2022 22:09:24 - INFO - codeparrot_training - Step 34426: {'lr': 0.00011901281453175711, 'samples': 17626624, 'steps': 34426, 'loss/train': 1.0482078790664673} 02/25/2022 22:09:29 - INFO - codeparrot_training - Step 34427: {'lr': 0.00011899887808886522, 'samples': 17627136, 'steps': 34427, 'loss/train': 1.4988723993301392} 02/25/2022 22:09:32 - INFO - codeparrot_training - Step 34428: {'lr': 0.00011898494220714057, 'samples': 17627648, 'steps': 34428, 'loss/train': 0.9011014699935913} 02/25/2022 22:09:38 - INFO - codeparrot_training - Step 34429: {'lr': 0.0001189710068866427, 'samples': 17628160, 'steps': 34429, 'loss/train': 1.652367353439331} 02/25/2022 22:09:42 - INFO - codeparrot_training - Step 34430: {'lr': 0.00011895707212743154, 'samples': 17628672, 'steps': 34430, 'loss/train': 2.3249340057373047} 02/25/2022 22:09:48 - INFO - codeparrot_training - Step 34431: {'lr': 0.00011894313792956666, 'samples': 17629184, 'steps': 34431, 'loss/train': 1.4634212255477905} 02/25/2022 22:09:51 - INFO - codeparrot_training - Step 34432: {'lr': 0.00011892920429310778, 'samples': 17629696, 'steps': 34432, 'loss/train': 1.5412893295288086} 02/25/2022 22:09:57 - INFO - codeparrot_training - Step 34433: {'lr': 0.00011891527121811446, 'samples': 17630208, 'steps': 34433, 'loss/train': 2.243695020675659} 02/25/2022 22:10:01 - INFO - codeparrot_training - Step 34434: {'lr': 0.00011890133870464658, 'samples': 17630720, 'steps': 34434, 'loss/train': 1.4952956438064575} 02/25/2022 22:10:06 - INFO - codeparrot_training - Step 34435: {'lr': 0.00011888740675276377, 'samples': 17631232, 'steps': 34435, 'loss/train': 2.9891302585601807} 02/25/2022 22:10:10 - INFO - codeparrot_training - Step 34436: {'lr': 0.00011887347536252565, 'samples': 17631744, 'steps': 34436, 'loss/train': 1.4947948455810547} 02/25/2022 22:10:15 - INFO - codeparrot_training - Step 34437: {'lr': 0.00011885954453399192, 'samples': 17632256, 'steps': 34437, 'loss/train': 1.708880066871643} 02/25/2022 22:10:19 - INFO - codeparrot_training - Step 34438: {'lr': 0.00011884561426722218, 'samples': 17632768, 'steps': 34438, 'loss/train': 1.7759019136428833} 02/25/2022 22:10:24 - INFO - codeparrot_training - Step 34439: {'lr': 0.00011883168456227628, 'samples': 17633280, 'steps': 34439, 'loss/train': 1.7213183641433716} 02/25/2022 22:10:28 - INFO - codeparrot_training - Step 34440: {'lr': 0.00011881775541921378, 'samples': 17633792, 'steps': 34440, 'loss/train': 2.7807064056396484} 02/25/2022 22:10:33 - INFO - codeparrot_training - Step 34441: {'lr': 0.00011880382683809435, 'samples': 17634304, 'steps': 34441, 'loss/train': 2.012974262237549} 02/25/2022 22:10:37 - INFO - codeparrot_training - Step 34442: {'lr': 0.00011878989881897758, 'samples': 17634816, 'steps': 34442, 'loss/train': 1.3133208751678467} 02/25/2022 22:10:42 - INFO - codeparrot_training - Step 34443: {'lr': 0.00011877597136192334, 'samples': 17635328, 'steps': 34443, 'loss/train': 1.5778939723968506} 02/25/2022 22:10:46 - INFO - codeparrot_training - Step 34444: {'lr': 0.00011876204446699104, 'samples': 17635840, 'steps': 34444, 'loss/train': 2.15189266204834} 02/25/2022 22:10:52 - INFO - codeparrot_training - Step 34445: {'lr': 0.0001187481181342407, 'samples': 17636352, 'steps': 34445, 'loss/train': 1.7066816091537476} 02/25/2022 22:10:55 - INFO - codeparrot_training - Step 34446: {'lr': 0.0001187341923637315, 'samples': 17636864, 'steps': 34446, 'loss/train': 1.9746837615966797} 02/25/2022 22:11:01 - INFO - codeparrot_training - Step 34447: {'lr': 0.00011872026715552348, 'samples': 17637376, 'steps': 34447, 'loss/train': 1.8586677312850952} 02/25/2022 22:11:05 - INFO - codeparrot_training - Step 34448: {'lr': 0.00011870634250967604, 'samples': 17637888, 'steps': 34448, 'loss/train': 1.7036064863204956} 02/25/2022 22:11:10 - INFO - codeparrot_training - Step 34449: {'lr': 0.00011869241842624912, 'samples': 17638400, 'steps': 34449, 'loss/train': 0.5855467915534973} 02/25/2022 22:11:14 - INFO - codeparrot_training - Step 34450: {'lr': 0.00011867849490530199, 'samples': 17638912, 'steps': 34450, 'loss/train': 0.8826522827148438} 02/25/2022 22:11:19 - INFO - codeparrot_training - Step 34451: {'lr': 0.00011866457194689462, 'samples': 17639424, 'steps': 34451, 'loss/train': 1.8082689046859741} 02/25/2022 22:11:23 - INFO - codeparrot_training - Step 34452: {'lr': 0.0001186506495510864, 'samples': 17639936, 'steps': 34452, 'loss/train': 1.526912808418274} 02/25/2022 22:11:28 - INFO - codeparrot_training - Step 34453: {'lr': 0.00011863672771793727, 'samples': 17640448, 'steps': 34453, 'loss/train': 1.8211246728897095} 02/25/2022 22:11:32 - INFO - codeparrot_training - Step 34454: {'lr': 0.0001186228064475065, 'samples': 17640960, 'steps': 34454, 'loss/train': 1.8936933279037476} 02/25/2022 22:11:38 - INFO - codeparrot_training - Step 34455: {'lr': 0.00011860888573985401, 'samples': 17641472, 'steps': 34455, 'loss/train': 1.4745776653289795} 02/25/2022 22:11:41 - INFO - codeparrot_training - Step 34456: {'lr': 0.00011859496559503924, 'samples': 17641984, 'steps': 34456, 'loss/train': 1.9079605340957642} 02/25/2022 22:11:47 - INFO - codeparrot_training - Step 34457: {'lr': 0.000118581046013122, 'samples': 17642496, 'steps': 34457, 'loss/train': 0.6239862442016602} 02/25/2022 22:11:50 - INFO - codeparrot_training - Step 34458: {'lr': 0.00011856712699416183, 'samples': 17643008, 'steps': 34458, 'loss/train': 2.514531373977661} 02/25/2022 22:11:56 - INFO - codeparrot_training - Step 34459: {'lr': 0.00011855320853821835, 'samples': 17643520, 'steps': 34459, 'loss/train': 1.8926079273223877} 02/25/2022 22:11:59 - INFO - codeparrot_training - Step 34460: {'lr': 0.0001185392906453511, 'samples': 17644032, 'steps': 34460, 'loss/train': 1.8797709941864014} 02/25/2022 22:12:05 - INFO - codeparrot_training - Step 34461: {'lr': 0.00011852537331561988, 'samples': 17644544, 'steps': 34461, 'loss/train': 1.1673511266708374} 02/25/2022 22:12:08 - INFO - codeparrot_training - Step 34462: {'lr': 0.00011851145654908421, 'samples': 17645056, 'steps': 34462, 'loss/train': 1.8389017581939697} 02/25/2022 22:12:14 - INFO - codeparrot_training - Step 34463: {'lr': 0.0001184975403458037, 'samples': 17645568, 'steps': 34463, 'loss/train': 2.3040175437927246} 02/25/2022 22:12:17 - INFO - codeparrot_training - Step 34464: {'lr': 0.0001184836247058379, 'samples': 17646080, 'steps': 34464, 'loss/train': 1.7416069507598877} 02/25/2022 22:12:24 - INFO - codeparrot_training - Step 34465: {'lr': 0.00011846970962924655, 'samples': 17646592, 'steps': 34465, 'loss/train': 1.2095332145690918} 02/25/2022 22:12:27 - INFO - codeparrot_training - Step 34466: {'lr': 0.00011845579511608922, 'samples': 17647104, 'steps': 34466, 'loss/train': 1.3330409526824951} 02/25/2022 22:12:33 - INFO - codeparrot_training - Step 34467: {'lr': 0.00011844188116642548, 'samples': 17647616, 'steps': 34467, 'loss/train': 0.02262728475034237} 02/25/2022 22:12:36 - INFO - codeparrot_training - Step 34468: {'lr': 0.00011842796778031487, 'samples': 17648128, 'steps': 34468, 'loss/train': 1.7267838716506958} 02/25/2022 22:12:42 - INFO - codeparrot_training - Step 34469: {'lr': 0.00011841405495781715, 'samples': 17648640, 'steps': 34469, 'loss/train': 1.9224244356155396} 02/25/2022 22:12:45 - INFO - codeparrot_training - Step 34470: {'lr': 0.00011840014269899186, 'samples': 17649152, 'steps': 34470, 'loss/train': 0.3081327974796295} 02/25/2022 22:12:51 - INFO - codeparrot_training - Step 34471: {'lr': 0.00011838623100389853, 'samples': 17649664, 'steps': 34471, 'loss/train': 1.6698603630065918} 02/25/2022 22:12:54 - INFO - codeparrot_training - Step 34472: {'lr': 0.00011837231987259672, 'samples': 17650176, 'steps': 34472, 'loss/train': 2.2035703659057617} 02/25/2022 22:13:00 - INFO - codeparrot_training - Step 34473: {'lr': 0.00011835840930514618, 'samples': 17650688, 'steps': 34473, 'loss/train': 2.0116829872131348} 02/25/2022 22:13:03 - INFO - codeparrot_training - Step 34474: {'lr': 0.00011834449930160643, 'samples': 17651200, 'steps': 34474, 'loss/train': 1.978700876235962} 02/25/2022 22:13:09 - INFO - codeparrot_training - Step 34475: {'lr': 0.00011833058986203704, 'samples': 17651712, 'steps': 34475, 'loss/train': 2.1101579666137695} 02/25/2022 22:13:13 - INFO - codeparrot_training - Step 34476: {'lr': 0.00011831668098649751, 'samples': 17652224, 'steps': 34476, 'loss/train': 0.505397617816925} 02/25/2022 22:13:18 - INFO - codeparrot_training - Step 34477: {'lr': 0.00011830277267504758, 'samples': 17652736, 'steps': 34477, 'loss/train': 2.005589723587036} 02/25/2022 22:13:22 - INFO - codeparrot_training - Step 34478: {'lr': 0.00011828886492774674, 'samples': 17653248, 'steps': 34478, 'loss/train': 2.6046063899993896} 02/25/2022 22:13:27 - INFO - codeparrot_training - Step 34479: {'lr': 0.0001182749577446546, 'samples': 17653760, 'steps': 34479, 'loss/train': 1.8999263048171997} 02/25/2022 22:13:31 - INFO - codeparrot_training - Step 34480: {'lr': 0.0001182610511258306, 'samples': 17654272, 'steps': 34480, 'loss/train': 1.9366304874420166} 02/25/2022 22:13:36 - INFO - codeparrot_training - Step 34481: {'lr': 0.00011824714507133455, 'samples': 17654784, 'steps': 34481, 'loss/train': 1.6994245052337646} 02/25/2022 22:13:40 - INFO - codeparrot_training - Step 34482: {'lr': 0.00011823323958122586, 'samples': 17655296, 'steps': 34482, 'loss/train': 0.6163861155509949} 02/25/2022 22:13:45 - INFO - codeparrot_training - Step 34483: {'lr': 0.00011821933465556406, 'samples': 17655808, 'steps': 34483, 'loss/train': 1.8265265226364136} 02/25/2022 22:13:51 - INFO - codeparrot_training - Step 34484: {'lr': 0.00011820543029440887, 'samples': 17656320, 'steps': 34484, 'loss/train': 2.1469321250915527} 02/25/2022 22:13:54 - INFO - codeparrot_training - Step 34485: {'lr': 0.00011819152649781978, 'samples': 17656832, 'steps': 34485, 'loss/train': 1.2261356115341187} 02/25/2022 22:13:58 - INFO - codeparrot_training - Step 34486: {'lr': 0.00011817762326585631, 'samples': 17657344, 'steps': 34486, 'loss/train': 1.5056616067886353} 02/25/2022 22:14:03 - INFO - codeparrot_training - Step 34487: {'lr': 0.00011816372059857797, 'samples': 17657856, 'steps': 34487, 'loss/train': 1.2098602056503296} 02/25/2022 22:14:07 - INFO - codeparrot_training - Step 34488: {'lr': 0.00011814981849604459, 'samples': 17658368, 'steps': 34488, 'loss/train': 1.6687403917312622} 02/25/2022 22:14:12 - INFO - codeparrot_training - Step 34489: {'lr': 0.0001181359169583153, 'samples': 17658880, 'steps': 34489, 'loss/train': 2.331040859222412} 02/25/2022 22:14:16 - INFO - codeparrot_training - Step 34490: {'lr': 0.00011812201598544998, 'samples': 17659392, 'steps': 34490, 'loss/train': 2.013587474822998} 02/25/2022 22:14:22 - INFO - codeparrot_training - Step 34491: {'lr': 0.00011810811557750797, 'samples': 17659904, 'steps': 34491, 'loss/train': 8.733814239501953} 02/25/2022 22:14:25 - INFO - codeparrot_training - Step 34492: {'lr': 0.00011809421573454901, 'samples': 17660416, 'steps': 34492, 'loss/train': 1.5935531854629517} 02/25/2022 22:14:31 - INFO - codeparrot_training - Step 34493: {'lr': 0.00011808031645663253, 'samples': 17660928, 'steps': 34493, 'loss/train': 1.9496949911117554} 02/25/2022 22:14:34 - INFO - codeparrot_training - Step 34494: {'lr': 0.0001180664177438181, 'samples': 17661440, 'steps': 34494, 'loss/train': 1.5631678104400635} 02/25/2022 22:14:40 - INFO - codeparrot_training - Step 34495: {'lr': 0.00011805251959616517, 'samples': 17661952, 'steps': 34495, 'loss/train': 2.1765522956848145} 02/25/2022 22:14:43 - INFO - codeparrot_training - Step 34496: {'lr': 0.00011803862201373342, 'samples': 17662464, 'steps': 34496, 'loss/train': 1.3541117906570435} 02/25/2022 22:14:49 - INFO - codeparrot_training - Step 34497: {'lr': 0.0001180247249965823, 'samples': 17662976, 'steps': 34497, 'loss/train': 1.4002761840820312} 02/25/2022 22:14:52 - INFO - codeparrot_training - Step 34498: {'lr': 0.00011801082854477136, 'samples': 17663488, 'steps': 34498, 'loss/train': 1.5403581857681274} 02/25/2022 22:14:58 - INFO - codeparrot_training - Step 34499: {'lr': 0.00011799693265836003, 'samples': 17664000, 'steps': 34499, 'loss/train': 1.3468595743179321} 02/25/2022 22:15:01 - INFO - codeparrot_training - Step 34500: {'lr': 0.00011798303733740801, 'samples': 17664512, 'steps': 34500, 'loss/train': 3.1558167934417725} 02/25/2022 22:15:08 - INFO - codeparrot_training - Step 34501: {'lr': 0.00011796914258197475, 'samples': 17665024, 'steps': 34501, 'loss/train': 2.0643725395202637} 02/25/2022 22:15:11 - INFO - codeparrot_training - Step 34502: {'lr': 0.00011795524839211974, 'samples': 17665536, 'steps': 34502, 'loss/train': 1.1217124462127686} 02/25/2022 22:15:17 - INFO - codeparrot_training - Step 34503: {'lr': 0.00011794135476790243, 'samples': 17666048, 'steps': 34503, 'loss/train': 2.5557968616485596} 02/25/2022 22:15:20 - INFO - codeparrot_training - Step 34504: {'lr': 0.00011792746170938254, 'samples': 17666560, 'steps': 34504, 'loss/train': 1.7109984159469604} 02/25/2022 22:15:26 - INFO - codeparrot_training - Step 34505: {'lr': 0.00011791356921661936, 'samples': 17667072, 'steps': 34505, 'loss/train': 2.3901760578155518} 02/25/2022 22:15:29 - INFO - codeparrot_training - Step 34506: {'lr': 0.00011789967728967269, 'samples': 17667584, 'steps': 34506, 'loss/train': 0.5616239309310913} 02/25/2022 22:15:35 - INFO - codeparrot_training - Step 34507: {'lr': 0.00011788578592860166, 'samples': 17668096, 'steps': 34507, 'loss/train': 2.199445962905884} 02/25/2022 22:15:38 - INFO - codeparrot_training - Step 34508: {'lr': 0.00011787189513346607, 'samples': 17668608, 'steps': 34508, 'loss/train': 0.9049603939056396} 02/25/2022 22:15:44 - INFO - codeparrot_training - Step 34509: {'lr': 0.00011785800490432525, 'samples': 17669120, 'steps': 34509, 'loss/train': 1.4657635688781738} 02/25/2022 22:15:47 - INFO - codeparrot_training - Step 34510: {'lr': 0.00011784411524123895, 'samples': 17669632, 'steps': 34510, 'loss/train': 2.1990153789520264} 02/25/2022 22:15:53 - INFO - codeparrot_training - Step 34511: {'lr': 0.0001178302261442663, 'samples': 17670144, 'steps': 34511, 'loss/train': 1.5574162006378174} 02/25/2022 22:15:57 - INFO - codeparrot_training - Step 34512: {'lr': 0.00011781633761346707, 'samples': 17670656, 'steps': 34512, 'loss/train': 2.137705087661743} 02/25/2022 22:16:02 - INFO - codeparrot_training - Step 34513: {'lr': 0.00011780244964890063, 'samples': 17671168, 'steps': 34513, 'loss/train': 1.1875412464141846} 02/25/2022 22:16:06 - INFO - codeparrot_training - Step 34514: {'lr': 0.00011778856225062668, 'samples': 17671680, 'steps': 34514, 'loss/train': 1.5434225797653198} 02/25/2022 22:16:11 - INFO - codeparrot_training - Step 34515: {'lr': 0.00011777467541870438, 'samples': 17672192, 'steps': 34515, 'loss/train': 2.895737886428833} 02/25/2022 22:16:15 - INFO - codeparrot_training - Step 34516: {'lr': 0.00011776078915319343, 'samples': 17672704, 'steps': 34516, 'loss/train': 0.9499350190162659} 02/25/2022 22:16:20 - INFO - codeparrot_training - Step 34517: {'lr': 0.00011774690345415323, 'samples': 17673216, 'steps': 34517, 'loss/train': 1.9704746007919312} 02/25/2022 22:16:24 - INFO - codeparrot_training - Step 34518: {'lr': 0.00011773301832164338, 'samples': 17673728, 'steps': 34518, 'loss/train': 1.9685696363449097} 02/25/2022 22:16:29 - INFO - codeparrot_training - Step 34519: {'lr': 0.00011771913375572327, 'samples': 17674240, 'steps': 34519, 'loss/train': 1.28207528591156} 02/25/2022 22:16:33 - INFO - codeparrot_training - Step 34520: {'lr': 0.00011770524975645239, 'samples': 17674752, 'steps': 34520, 'loss/train': 1.3010730743408203} 02/25/2022 22:16:38 - INFO - codeparrot_training - Step 34521: {'lr': 0.00011769136632389019, 'samples': 17675264, 'steps': 34521, 'loss/train': 1.799863338470459} 02/25/2022 22:16:44 - INFO - codeparrot_training - Step 34522: {'lr': 0.00011767748345809609, 'samples': 17675776, 'steps': 34522, 'loss/train': 2.1449368000030518} 02/25/2022 22:16:48 - INFO - codeparrot_training - Step 34523: {'lr': 0.00011766360115912975, 'samples': 17676288, 'steps': 34523, 'loss/train': 2.3421308994293213} 02/25/2022 22:16:53 - INFO - codeparrot_training - Step 34524: {'lr': 0.0001176497194270505, 'samples': 17676800, 'steps': 34524, 'loss/train': 1.8416329622268677} 02/25/2022 22:16:57 - INFO - codeparrot_training - Step 34525: {'lr': 0.00011763583826191781, 'samples': 17677312, 'steps': 34525, 'loss/train': 2.042531728744507} 02/25/2022 22:17:00 - INFO - codeparrot_training - Step 34526: {'lr': 0.00011762195766379107, 'samples': 17677824, 'steps': 34526, 'loss/train': 8.58413314819336} 02/25/2022 22:17:06 - INFO - codeparrot_training - Step 34527: {'lr': 0.00011760807763272993, 'samples': 17678336, 'steps': 34527, 'loss/train': 1.0910606384277344} 02/25/2022 22:17:09 - INFO - codeparrot_training - Step 34528: {'lr': 0.00011759419816879374, 'samples': 17678848, 'steps': 34528, 'loss/train': 1.202926754951477} 02/25/2022 22:17:15 - INFO - codeparrot_training - Step 34529: {'lr': 0.00011758031927204199, 'samples': 17679360, 'steps': 34529, 'loss/train': 2.0078704357147217} 02/25/2022 22:17:19 - INFO - codeparrot_training - Step 34530: {'lr': 0.000117566440942534, 'samples': 17679872, 'steps': 34530, 'loss/train': 1.1683276891708374} 02/25/2022 22:17:24 - INFO - codeparrot_training - Step 34531: {'lr': 0.00011755256318032939, 'samples': 17680384, 'steps': 34531, 'loss/train': 2.735950469970703} 02/25/2022 22:17:28 - INFO - codeparrot_training - Step 34532: {'lr': 0.00011753868598548756, 'samples': 17680896, 'steps': 34532, 'loss/train': 1.4186333417892456} 02/25/2022 22:17:33 - INFO - codeparrot_training - Step 34533: {'lr': 0.00011752480935806795, 'samples': 17681408, 'steps': 34533, 'loss/train': 2.1476638317108154} 02/25/2022 22:17:37 - INFO - codeparrot_training - Step 34534: {'lr': 0.00011751093329812986, 'samples': 17681920, 'steps': 34534, 'loss/train': 1.9056293964385986} 02/25/2022 22:17:42 - INFO - codeparrot_training - Step 34535: {'lr': 0.000117497057805733, 'samples': 17682432, 'steps': 34535, 'loss/train': 0.7239101529121399} 02/25/2022 22:17:45 - INFO - codeparrot_training - Step 34536: {'lr': 0.00011748318288093664, 'samples': 17682944, 'steps': 34536, 'loss/train': 2.2706823348999023} 02/25/2022 22:17:52 - INFO - codeparrot_training - Step 34537: {'lr': 0.00011746930852380028, 'samples': 17683456, 'steps': 34537, 'loss/train': 1.4279333353042603} 02/25/2022 22:17:56 - INFO - codeparrot_training - Step 34538: {'lr': 0.0001174554347343832, 'samples': 17683968, 'steps': 34538, 'loss/train': 1.6547961235046387} 02/25/2022 22:18:01 - INFO - codeparrot_training - Step 34539: {'lr': 0.00011744156151274505, 'samples': 17684480, 'steps': 34539, 'loss/train': 1.0733511447906494} 02/25/2022 22:18:07 - INFO - codeparrot_training - Step 34540: {'lr': 0.00011742768885894515, 'samples': 17684992, 'steps': 34540, 'loss/train': 2.257333517074585} 02/25/2022 22:18:10 - INFO - codeparrot_training - Step 34541: {'lr': 0.00011741381677304294, 'samples': 17685504, 'steps': 34541, 'loss/train': 1.711201548576355} 02/25/2022 22:18:16 - INFO - codeparrot_training - Step 34542: {'lr': 0.00011739994525509772, 'samples': 17686016, 'steps': 34542, 'loss/train': 1.2990092039108276} 02/25/2022 22:18:19 - INFO - codeparrot_training - Step 34543: {'lr': 0.00011738607430516915, 'samples': 17686528, 'steps': 34543, 'loss/train': 0.5107671022415161} 02/25/2022 22:18:25 - INFO - codeparrot_training - Step 34544: {'lr': 0.00011737220392331644, 'samples': 17687040, 'steps': 34544, 'loss/train': 1.5232945680618286} 02/25/2022 22:18:28 - INFO - codeparrot_training - Step 34545: {'lr': 0.00011735833410959929, 'samples': 17687552, 'steps': 34545, 'loss/train': 2.2349820137023926} 02/25/2022 22:18:35 - INFO - codeparrot_training - Step 34546: {'lr': 0.00011734446486407671, 'samples': 17688064, 'steps': 34546, 'loss/train': 1.894779086112976} 02/25/2022 22:18:38 - INFO - codeparrot_training - Step 34547: {'lr': 0.00011733059618680841, 'samples': 17688576, 'steps': 34547, 'loss/train': 1.9319714307785034} 02/25/2022 22:18:44 - INFO - codeparrot_training - Step 34548: {'lr': 0.00011731672807785362, 'samples': 17689088, 'steps': 34548, 'loss/train': 1.4587323665618896} 02/25/2022 22:18:47 - INFO - codeparrot_training - Step 34549: {'lr': 0.00011730286053727201, 'samples': 17689600, 'steps': 34549, 'loss/train': 1.6989874839782715} 02/25/2022 22:18:53 - INFO - codeparrot_training - Step 34550: {'lr': 0.00011728899356512265, 'samples': 17690112, 'steps': 34550, 'loss/train': 1.7564151287078857} 02/25/2022 22:18:56 - INFO - codeparrot_training - Step 34551: {'lr': 0.00011727512716146519, 'samples': 17690624, 'steps': 34551, 'loss/train': 0.7209196090698242} 02/25/2022 22:19:02 - INFO - codeparrot_training - Step 34552: {'lr': 0.00011726126132635884, 'samples': 17691136, 'steps': 34552, 'loss/train': 1.901749610900879} 02/25/2022 22:19:05 - INFO - codeparrot_training - Step 34553: {'lr': 0.00011724739605986318, 'samples': 17691648, 'steps': 34553, 'loss/train': 1.8601866960525513} 02/25/2022 22:19:11 - INFO - codeparrot_training - Step 34554: {'lr': 0.00011723353136203752, 'samples': 17692160, 'steps': 34554, 'loss/train': 1.503511667251587} 02/25/2022 22:19:14 - INFO - codeparrot_training - Step 34555: {'lr': 0.00011721966723294128, 'samples': 17692672, 'steps': 34555, 'loss/train': 2.1501471996307373} 02/25/2022 22:19:19 - INFO - codeparrot_training - Step 34556: {'lr': 0.0001172058036726337, 'samples': 17693184, 'steps': 34556, 'loss/train': 2.0608112812042236} 02/25/2022 22:19:23 - INFO - codeparrot_training - Step 34557: {'lr': 0.00011719194068117439, 'samples': 17693696, 'steps': 34557, 'loss/train': 1.8894293308258057} 02/25/2022 22:19:30 - INFO - codeparrot_training - Step 34558: {'lr': 0.00011717807825862264, 'samples': 17694208, 'steps': 34558, 'loss/train': 2.048642635345459} 02/25/2022 22:19:33 - INFO - codeparrot_training - Step 34559: {'lr': 0.00011716421640503783, 'samples': 17694720, 'steps': 34559, 'loss/train': 1.0017175674438477} 02/25/2022 22:19:39 - INFO - codeparrot_training - Step 34560: {'lr': 0.00011715035512047925, 'samples': 17695232, 'steps': 34560, 'loss/train': 1.1716833114624023} 02/25/2022 22:19:42 - INFO - codeparrot_training - Step 34561: {'lr': 0.00011713649440500646, 'samples': 17695744, 'steps': 34561, 'loss/train': 1.7081788778305054} 02/25/2022 22:19:48 - INFO - codeparrot_training - Step 34562: {'lr': 0.00011712263425867872, 'samples': 17696256, 'steps': 34562, 'loss/train': 1.4829665422439575} 02/25/2022 22:19:51 - INFO - codeparrot_training - Step 34563: {'lr': 0.00011710877468155543, 'samples': 17696768, 'steps': 34563, 'loss/train': 1.4972842931747437} 02/25/2022 22:19:57 - INFO - codeparrot_training - Step 34564: {'lr': 0.00011709491567369584, 'samples': 17697280, 'steps': 34564, 'loss/train': 0.6604659557342529} 02/25/2022 22:20:00 - INFO - codeparrot_training - Step 34565: {'lr': 0.00011708105723515955, 'samples': 17697792, 'steps': 34565, 'loss/train': 1.2514607906341553} 02/25/2022 22:20:06 - INFO - codeparrot_training - Step 34566: {'lr': 0.0001170671993660058, 'samples': 17698304, 'steps': 34566, 'loss/train': 2.00887393951416} 02/25/2022 22:20:09 - INFO - codeparrot_training - Step 34567: {'lr': 0.00011705334206629396, 'samples': 17698816, 'steps': 34567, 'loss/train': 2.506592035293579} 02/25/2022 22:20:15 - INFO - codeparrot_training - Step 34568: {'lr': 0.00011703948533608339, 'samples': 17699328, 'steps': 34568, 'loss/train': 1.84834623336792} 02/25/2022 22:20:19 - INFO - codeparrot_training - Step 34569: {'lr': 0.00011702562917543332, 'samples': 17699840, 'steps': 34569, 'loss/train': 1.971256971359253} 02/25/2022 22:20:24 - INFO - codeparrot_training - Step 34570: {'lr': 0.00011701177358440334, 'samples': 17700352, 'steps': 34570, 'loss/train': 1.0615979433059692} 02/25/2022 22:20:28 - INFO - codeparrot_training - Step 34571: {'lr': 0.00011699791856305272, 'samples': 17700864, 'steps': 34571, 'loss/train': 2.1620712280273438} 02/25/2022 22:20:33 - INFO - codeparrot_training - Step 34572: {'lr': 0.00011698406411144074, 'samples': 17701376, 'steps': 34572, 'loss/train': 2.0336813926696777} 02/25/2022 22:20:37 - INFO - codeparrot_training - Step 34573: {'lr': 0.00011697021022962673, 'samples': 17701888, 'steps': 34573, 'loss/train': 1.7189326286315918} 02/25/2022 22:20:42 - INFO - codeparrot_training - Step 34574: {'lr': 0.00011695635691767018, 'samples': 17702400, 'steps': 34574, 'loss/train': 1.8887455463409424} 02/25/2022 22:20:46 - INFO - codeparrot_training - Step 34575: {'lr': 0.00011694250417563036, 'samples': 17702912, 'steps': 34575, 'loss/train': 2.6055195331573486} 02/25/2022 22:20:51 - INFO - codeparrot_training - Step 34576: {'lr': 0.00011692865200356659, 'samples': 17703424, 'steps': 34576, 'loss/train': 2.192671298980713} 02/25/2022 22:20:55 - INFO - codeparrot_training - Step 34577: {'lr': 0.00011691480040153812, 'samples': 17703936, 'steps': 34577, 'loss/train': 1.7293568849563599} 02/25/2022 22:21:01 - INFO - codeparrot_training - Step 34578: {'lr': 0.00011690094936960449, 'samples': 17704448, 'steps': 34578, 'loss/train': 1.6918243169784546} 02/25/2022 22:21:04 - INFO - codeparrot_training - Step 34579: {'lr': 0.00011688709890782484, 'samples': 17704960, 'steps': 34579, 'loss/train': 1.848863959312439} 02/25/2022 22:21:09 - INFO - codeparrot_training - Step 34580: {'lr': 0.00011687324901625879, 'samples': 17705472, 'steps': 34580, 'loss/train': 1.8401762247085571} 02/25/2022 22:21:13 - INFO - codeparrot_training - Step 34581: {'lr': 0.00011685939969496526, 'samples': 17705984, 'steps': 34581, 'loss/train': 2.2630977630615234} 02/25/2022 22:21:20 - INFO - codeparrot_training - Step 34582: {'lr': 0.00011684555094400388, 'samples': 17706496, 'steps': 34582, 'loss/train': 2.4304563999176025} 02/25/2022 22:21:23 - INFO - codeparrot_training - Step 34583: {'lr': 0.0001168317027634338, 'samples': 17707008, 'steps': 34583, 'loss/train': 1.4651545286178589} 02/25/2022 22:21:29 - INFO - codeparrot_training - Step 34584: {'lr': 0.00011681785515331458, 'samples': 17707520, 'steps': 34584, 'loss/train': 1.0978509187698364} 02/25/2022 22:21:32 - INFO - codeparrot_training - Step 34585: {'lr': 0.0001168040081137052, 'samples': 17708032, 'steps': 34585, 'loss/train': 1.6135120391845703} 02/25/2022 22:21:38 - INFO - codeparrot_training - Step 34586: {'lr': 0.00011679016164466528, 'samples': 17708544, 'steps': 34586, 'loss/train': 0.91129469871521} 02/25/2022 22:21:41 - INFO - codeparrot_training - Step 34587: {'lr': 0.00011677631574625389, 'samples': 17709056, 'steps': 34587, 'loss/train': 1.7425577640533447} 02/25/2022 22:21:47 - INFO - codeparrot_training - Step 34588: {'lr': 0.00011676247041853063, 'samples': 17709568, 'steps': 34588, 'loss/train': 1.8360234498977661} 02/25/2022 22:21:51 - INFO - codeparrot_training - Step 34589: {'lr': 0.00011674862566155448, 'samples': 17710080, 'steps': 34589, 'loss/train': 1.300140142440796} 02/25/2022 22:21:56 - INFO - codeparrot_training - Step 34590: {'lr': 0.00011673478147538497, 'samples': 17710592, 'steps': 34590, 'loss/train': 1.4613616466522217} 02/25/2022 22:22:00 - INFO - codeparrot_training - Step 34591: {'lr': 0.00011672093786008128, 'samples': 17711104, 'steps': 34591, 'loss/train': 0.02263191156089306} 02/25/2022 22:22:05 - INFO - codeparrot_training - Step 34592: {'lr': 0.00011670709481570285, 'samples': 17711616, 'steps': 34592, 'loss/train': 2.3177847862243652} 02/25/2022 22:22:09 - INFO - codeparrot_training - Step 34593: {'lr': 0.0001166932523423089, 'samples': 17712128, 'steps': 34593, 'loss/train': 2.3077192306518555} 02/25/2022 22:22:15 - INFO - codeparrot_training - Step 34594: {'lr': 0.00011667941043995873, 'samples': 17712640, 'steps': 34594, 'loss/train': 0.9222477078437805} 02/25/2022 22:22:19 - INFO - codeparrot_training - Step 34595: {'lr': 0.00011666556910871151, 'samples': 17713152, 'steps': 34595, 'loss/train': 2.0848236083984375} 02/25/2022 22:22:24 - INFO - codeparrot_training - Step 34596: {'lr': 0.00011665172834862678, 'samples': 17713664, 'steps': 34596, 'loss/train': 1.283647060394287} 02/25/2022 22:22:28 - INFO - codeparrot_training - Step 34597: {'lr': 0.0001166378881597637, 'samples': 17714176, 'steps': 34597, 'loss/train': 1.5610575675964355} 02/25/2022 22:22:33 - INFO - codeparrot_training - Step 34598: {'lr': 0.00011662404854218156, 'samples': 17714688, 'steps': 34598, 'loss/train': 1.3010587692260742} 02/25/2022 22:22:37 - INFO - codeparrot_training - Step 34599: {'lr': 0.00011661020949593954, 'samples': 17715200, 'steps': 34599, 'loss/train': 1.7915834188461304} 02/25/2022 22:22:42 - INFO - codeparrot_training - Step 34600: {'lr': 0.00011659637102109713, 'samples': 17715712, 'steps': 34600, 'loss/train': 2.0628342628479004} 02/25/2022 22:22:46 - INFO - codeparrot_training - Step 34601: {'lr': 0.0001165825331177135, 'samples': 17716224, 'steps': 34601, 'loss/train': 0.058083467185497284} 02/25/2022 22:22:52 - INFO - codeparrot_training - Step 34602: {'lr': 0.00011656869578584792, 'samples': 17716736, 'steps': 34602, 'loss/train': 2.5492336750030518} 02/25/2022 22:22:55 - INFO - codeparrot_training - Step 34603: {'lr': 0.0001165548590255596, 'samples': 17717248, 'steps': 34603, 'loss/train': 1.669998288154602} 02/25/2022 22:23:01 - INFO - codeparrot_training - Step 34604: {'lr': 0.00011654102283690798, 'samples': 17717760, 'steps': 34604, 'loss/train': 1.1803606748580933} 02/25/2022 22:23:04 - INFO - codeparrot_training - Step 34605: {'lr': 0.00011652718721995226, 'samples': 17718272, 'steps': 34605, 'loss/train': 2.2157907485961914} 02/25/2022 22:23:10 - INFO - codeparrot_training - Step 34606: {'lr': 0.00011651335217475167, 'samples': 17718784, 'steps': 34606, 'loss/train': 3.046508550643921} 02/25/2022 22:23:13 - INFO - codeparrot_training - Step 34607: {'lr': 0.00011649951770136538, 'samples': 17719296, 'steps': 34607, 'loss/train': 2.4108340740203857} 02/25/2022 22:23:19 - INFO - codeparrot_training - Step 34608: {'lr': 0.00011648568379985289, 'samples': 17719808, 'steps': 34608, 'loss/train': 1.2670422792434692} 02/25/2022 22:23:22 - INFO - codeparrot_training - Step 34609: {'lr': 0.00011647185047027334, 'samples': 17720320, 'steps': 34609, 'loss/train': 1.6605414152145386} 02/25/2022 22:23:28 - INFO - codeparrot_training - Step 34610: {'lr': 0.00011645801771268596, 'samples': 17720832, 'steps': 34610, 'loss/train': 0.9107034206390381} 02/25/2022 22:23:31 - INFO - codeparrot_training - Step 34611: {'lr': 0.00011644418552714994, 'samples': 17721344, 'steps': 34611, 'loss/train': 1.0117378234863281} 02/25/2022 22:23:37 - INFO - codeparrot_training - Step 34612: {'lr': 0.0001164303539137247, 'samples': 17721856, 'steps': 34612, 'loss/train': 2.227492332458496} 02/25/2022 22:23:41 - INFO - codeparrot_training - Step 34613: {'lr': 0.00011641652287246942, 'samples': 17722368, 'steps': 34613, 'loss/train': 1.8931323289871216} 02/25/2022 22:23:46 - INFO - codeparrot_training - Step 34614: {'lr': 0.00011640269240344326, 'samples': 17722880, 'steps': 34614, 'loss/train': 1.9377555847167969} 02/25/2022 22:23:50 - INFO - codeparrot_training - Step 34615: {'lr': 0.00011638886250670572, 'samples': 17723392, 'steps': 34615, 'loss/train': 1.1202654838562012} 02/25/2022 22:23:55 - INFO - codeparrot_training - Step 34616: {'lr': 0.00011637503318231568, 'samples': 17723904, 'steps': 34616, 'loss/train': 3.031440019607544} 02/25/2022 22:23:59 - INFO - codeparrot_training - Step 34617: {'lr': 0.00011636120443033265, 'samples': 17724416, 'steps': 34617, 'loss/train': 0.8041306734085083} 02/25/2022 22:24:04 - INFO - codeparrot_training - Step 34618: {'lr': 0.0001163473762508157, 'samples': 17724928, 'steps': 34618, 'loss/train': 2.0794122219085693} 02/25/2022 22:24:08 - INFO - codeparrot_training - Step 34619: {'lr': 0.00011633354864382437, 'samples': 17725440, 'steps': 34619, 'loss/train': 1.8044875860214233} 02/25/2022 22:24:13 - INFO - codeparrot_training - Step 34620: {'lr': 0.00011631972160941745, 'samples': 17725952, 'steps': 34620, 'loss/train': 0.1137242391705513} 02/25/2022 22:24:17 - INFO - codeparrot_training - Step 34621: {'lr': 0.00011630589514765449, 'samples': 17726464, 'steps': 34621, 'loss/train': 1.539244294166565} 02/25/2022 22:24:22 - INFO - codeparrot_training - Step 34622: {'lr': 0.00011629206925859454, 'samples': 17726976, 'steps': 34622, 'loss/train': 1.1114946603775024} 02/25/2022 22:24:26 - INFO - codeparrot_training - Step 34623: {'lr': 0.00011627824394229708, 'samples': 17727488, 'steps': 34623, 'loss/train': 2.2846133708953857} 02/25/2022 22:24:31 - INFO - codeparrot_training - Step 34624: {'lr': 0.000116264419198821, 'samples': 17728000, 'steps': 34624, 'loss/train': 1.5699506998062134} 02/25/2022 22:24:35 - INFO - codeparrot_training - Step 34625: {'lr': 0.00011625059502822574, 'samples': 17728512, 'steps': 34625, 'loss/train': 0.9542575478553772} 02/25/2022 22:24:40 - INFO - codeparrot_training - Step 34626: {'lr': 0.0001162367714305704, 'samples': 17729024, 'steps': 34626, 'loss/train': 1.475652813911438} 02/25/2022 22:24:43 - INFO - codeparrot_training - Step 34627: {'lr': 0.00011622294840591432, 'samples': 17729536, 'steps': 34627, 'loss/train': 1.5890648365020752} 02/25/2022 22:24:50 - INFO - codeparrot_training - Step 34628: {'lr': 0.00011620912595431668, 'samples': 17730048, 'steps': 34628, 'loss/train': 1.347261905670166} 02/25/2022 22:24:53 - INFO - codeparrot_training - Step 34629: {'lr': 0.00011619530407583665, 'samples': 17730560, 'steps': 34629, 'loss/train': 0.06105469539761543} 02/25/2022 22:24:59 - INFO - codeparrot_training - Step 34630: {'lr': 0.00011618148277053334, 'samples': 17731072, 'steps': 34630, 'loss/train': 1.517052412033081} 02/25/2022 22:25:02 - INFO - codeparrot_training - Step 34631: {'lr': 0.00011616766203846618, 'samples': 17731584, 'steps': 34631, 'loss/train': 1.7262964248657227} 02/25/2022 22:25:07 - INFO - codeparrot_training - Step 34632: {'lr': 0.00011615384187969424, 'samples': 17732096, 'steps': 34632, 'loss/train': 1.9341062307357788} 02/25/2022 22:25:11 - INFO - codeparrot_training - Step 34633: {'lr': 0.00011614002229427676, 'samples': 17732608, 'steps': 34633, 'loss/train': 1.9326890707015991} 02/25/2022 22:25:16 - INFO - codeparrot_training - Step 34634: {'lr': 0.00011612620328227283, 'samples': 17733120, 'steps': 34634, 'loss/train': 0.2123493254184723} 02/25/2022 22:25:22 - INFO - codeparrot_training - Step 34635: {'lr': 0.00011611238484374181, 'samples': 17733632, 'steps': 34635, 'loss/train': 2.0048444271087646} 02/25/2022 22:25:26 - INFO - codeparrot_training - Step 34636: {'lr': 0.00011609856697874283, 'samples': 17734144, 'steps': 34636, 'loss/train': 1.5604960918426514} 02/25/2022 22:25:29 - INFO - codeparrot_training - Step 34637: {'lr': 0.00011608474968733507, 'samples': 17734656, 'steps': 34637, 'loss/train': 1.8149012327194214} 02/25/2022 22:25:35 - INFO - codeparrot_training - Step 34638: {'lr': 0.00011607093296957763, 'samples': 17735168, 'steps': 34638, 'loss/train': 2.2279791831970215} 02/25/2022 22:25:38 - INFO - codeparrot_training - Step 34639: {'lr': 0.00011605711682552986, 'samples': 17735680, 'steps': 34639, 'loss/train': 2.379338502883911} 02/25/2022 22:25:45 - INFO - codeparrot_training - Step 34640: {'lr': 0.00011604330125525078, 'samples': 17736192, 'steps': 34640, 'loss/train': 1.7767943143844604} 02/25/2022 22:25:48 - INFO - codeparrot_training - Step 34641: {'lr': 0.00011602948625879986, 'samples': 17736704, 'steps': 34641, 'loss/train': 1.3086358308792114} 02/25/2022 22:25:53 - INFO - codeparrot_training - Step 34642: {'lr': 0.0001160156718362359, 'samples': 17737216, 'steps': 34642, 'loss/train': 1.9066476821899414} 02/25/2022 22:25:57 - INFO - codeparrot_training - Step 34643: {'lr': 0.00011600185798761833, 'samples': 17737728, 'steps': 34643, 'loss/train': 1.8030650615692139} 02/25/2022 22:26:03 - INFO - codeparrot_training - Step 34644: {'lr': 0.00011598804471300617, 'samples': 17738240, 'steps': 34644, 'loss/train': 2.8161821365356445} 02/25/2022 22:26:06 - INFO - codeparrot_training - Step 34645: {'lr': 0.00011597423201245888, 'samples': 17738752, 'steps': 34645, 'loss/train': 2.0014748573303223} 02/25/2022 22:26:11 - INFO - codeparrot_training - Step 34646: {'lr': 0.0001159604198860352, 'samples': 17739264, 'steps': 34646, 'loss/train': 1.6776434183120728} 02/25/2022 22:26:15 - INFO - codeparrot_training - Step 34647: {'lr': 0.00011594660833379464, 'samples': 17739776, 'steps': 34647, 'loss/train': 1.5947719812393188} 02/25/2022 22:26:20 - INFO - codeparrot_training - Step 34648: {'lr': 0.0001159327973557962, 'samples': 17740288, 'steps': 34648, 'loss/train': 1.512926697731018} 02/25/2022 22:26:24 - INFO - codeparrot_training - Step 34649: {'lr': 0.00011591898695209913, 'samples': 17740800, 'steps': 34649, 'loss/train': 2.541478395462036} 02/25/2022 22:26:30 - INFO - codeparrot_training - Step 34650: {'lr': 0.00011590517712276244, 'samples': 17741312, 'steps': 34650, 'loss/train': 2.7670252323150635} 02/25/2022 22:26:33 - INFO - codeparrot_training - Step 34651: {'lr': 0.00011589136786784548, 'samples': 17741824, 'steps': 34651, 'loss/train': 1.616425633430481} 02/25/2022 22:26:39 - INFO - codeparrot_training - Step 34652: {'lr': 0.0001158775591874073, 'samples': 17742336, 'steps': 34652, 'loss/train': 1.431565523147583} 02/25/2022 22:26:42 - INFO - codeparrot_training - Step 34653: {'lr': 0.00011586375108150702, 'samples': 17742848, 'steps': 34653, 'loss/train': 0.9353484511375427} 02/25/2022 22:26:48 - INFO - codeparrot_training - Step 34654: {'lr': 0.00011584994355020389, 'samples': 17743360, 'steps': 34654, 'loss/train': 1.4318761825561523} 02/25/2022 22:26:54 - INFO - codeparrot_training - Step 34655: {'lr': 0.00011583613659355702, 'samples': 17743872, 'steps': 34655, 'loss/train': 2.1980092525482178} 02/25/2022 22:26:57 - INFO - codeparrot_training - Step 34656: {'lr': 0.00011582233021162553, 'samples': 17744384, 'steps': 34656, 'loss/train': 2.157557487487793} 02/25/2022 22:27:03 - INFO - codeparrot_training - Step 34657: {'lr': 0.00011580852440446848, 'samples': 17744896, 'steps': 34657, 'loss/train': 1.8558359146118164} 02/25/2022 22:27:06 - INFO - codeparrot_training - Step 34658: {'lr': 0.00011579471917214519, 'samples': 17745408, 'steps': 34658, 'loss/train': 0.5897127389907837} 02/25/2022 22:27:12 - INFO - codeparrot_training - Step 34659: {'lr': 0.00011578091451471471, 'samples': 17745920, 'steps': 34659, 'loss/train': 1.6939104795455933} 02/25/2022 22:27:15 - INFO - codeparrot_training - Step 34660: {'lr': 0.00011576711043223617, 'samples': 17746432, 'steps': 34660, 'loss/train': 1.783022403717041} 02/25/2022 22:27:21 - INFO - codeparrot_training - Step 34661: {'lr': 0.00011575330692476863, 'samples': 17746944, 'steps': 34661, 'loss/train': 0.041671521961688995} 02/25/2022 22:27:24 - INFO - codeparrot_training - Step 34662: {'lr': 0.00011573950399237138, 'samples': 17747456, 'steps': 34662, 'loss/train': 1.9889122247695923} 02/25/2022 22:27:30 - INFO - codeparrot_training - Step 34663: {'lr': 0.00011572570163510346, 'samples': 17747968, 'steps': 34663, 'loss/train': 2.0284688472747803} 02/25/2022 22:27:33 - INFO - codeparrot_training - Step 34664: {'lr': 0.000115711899853024, 'samples': 17748480, 'steps': 34664, 'loss/train': 1.3339725732803345} 02/25/2022 22:27:39 - INFO - codeparrot_training - Step 34665: {'lr': 0.00011569809864619202, 'samples': 17748992, 'steps': 34665, 'loss/train': 1.8196130990982056} 02/25/2022 22:27:43 - INFO - codeparrot_training - Step 34666: {'lr': 0.00011568429801466684, 'samples': 17749504, 'steps': 34666, 'loss/train': 1.3567323684692383} 02/25/2022 22:27:48 - INFO - codeparrot_training - Step 34667: {'lr': 0.00011567049795850749, 'samples': 17750016, 'steps': 34667, 'loss/train': 2.4590413570404053} 02/25/2022 22:27:52 - INFO - codeparrot_training - Step 34668: {'lr': 0.00011565669847777305, 'samples': 17750528, 'steps': 34668, 'loss/train': 2.0494492053985596} 02/25/2022 22:27:57 - INFO - codeparrot_training - Step 34669: {'lr': 0.00011564289957252255, 'samples': 17751040, 'steps': 34669, 'loss/train': 0.3943789005279541} 02/25/2022 22:28:01 - INFO - codeparrot_training - Step 34670: {'lr': 0.0001156291012428153, 'samples': 17751552, 'steps': 34670, 'loss/train': 1.8019856214523315} 02/25/2022 22:28:06 - INFO - codeparrot_training - Step 34671: {'lr': 0.00011561530348871036, 'samples': 17752064, 'steps': 34671, 'loss/train': 1.2965737581253052} 02/25/2022 22:28:10 - INFO - codeparrot_training - Step 34672: {'lr': 0.00011560150631026675, 'samples': 17752576, 'steps': 34672, 'loss/train': 1.956367015838623} 02/25/2022 22:28:15 - INFO - codeparrot_training - Step 34673: {'lr': 0.00011558770970754351, 'samples': 17753088, 'steps': 34673, 'loss/train': 1.4930142164230347} 02/25/2022 22:28:19 - INFO - codeparrot_training - Step 34674: {'lr': 0.00011557391368059997, 'samples': 17753600, 'steps': 34674, 'loss/train': 1.1915210485458374} 02/25/2022 22:28:25 - INFO - codeparrot_training - Step 34675: {'lr': 0.00011556011822949497, 'samples': 17754112, 'steps': 34675, 'loss/train': 0.9868058562278748} 02/25/2022 22:28:28 - INFO - codeparrot_training - Step 34676: {'lr': 0.00011554632335428795, 'samples': 17754624, 'steps': 34676, 'loss/train': 2.591519832611084} 02/25/2022 22:28:34 - INFO - codeparrot_training - Step 34677: {'lr': 0.00011553252905503755, 'samples': 17755136, 'steps': 34677, 'loss/train': 1.2586069107055664} 02/25/2022 22:28:37 - INFO - codeparrot_training - Step 34678: {'lr': 0.00011551873533180323, 'samples': 17755648, 'steps': 34678, 'loss/train': 1.4948464632034302} 02/25/2022 22:28:43 - INFO - codeparrot_training - Step 34679: {'lr': 0.00011550494218464386, 'samples': 17756160, 'steps': 34679, 'loss/train': 2.0157244205474854} 02/25/2022 22:28:46 - INFO - codeparrot_training - Step 34680: {'lr': 0.00011549114961361879, 'samples': 17756672, 'steps': 34680, 'loss/train': 1.7757574319839478} 02/25/2022 22:28:52 - INFO - codeparrot_training - Step 34681: {'lr': 0.00011547735761878672, 'samples': 17757184, 'steps': 34681, 'loss/train': 2.003019094467163} 02/25/2022 22:28:56 - INFO - codeparrot_training - Step 34682: {'lr': 0.00011546356620020704, 'samples': 17757696, 'steps': 34682, 'loss/train': 1.2299540042877197} 02/25/2022 22:29:01 - INFO - codeparrot_training - Step 34683: {'lr': 0.00011544977535793863, 'samples': 17758208, 'steps': 34683, 'loss/train': 1.5901049375534058} 02/25/2022 22:29:05 - INFO - codeparrot_training - Step 34684: {'lr': 0.00011543598509204084, 'samples': 17758720, 'steps': 34684, 'loss/train': 2.0335536003112793} 02/25/2022 22:29:11 - INFO - codeparrot_training - Step 34685: {'lr': 0.00011542219540257237, 'samples': 17759232, 'steps': 34685, 'loss/train': 1.7262800931930542} 02/25/2022 22:29:14 - INFO - codeparrot_training - Step 34686: {'lr': 0.00011540840628959259, 'samples': 17759744, 'steps': 34686, 'loss/train': 1.1418932676315308} 02/25/2022 22:29:20 - INFO - codeparrot_training - Step 34687: {'lr': 0.00011539461775316035, 'samples': 17760256, 'steps': 34687, 'loss/train': 1.444786787033081} 02/25/2022 22:29:23 - INFO - codeparrot_training - Step 34688: {'lr': 0.00011538082979333495, 'samples': 17760768, 'steps': 34688, 'loss/train': 1.8199483156204224} 02/25/2022 22:29:29 - INFO - codeparrot_training - Step 34689: {'lr': 0.00011536704241017529, 'samples': 17761280, 'steps': 34689, 'loss/train': 1.733020544052124} 02/25/2022 22:29:32 - INFO - codeparrot_training - Step 34690: {'lr': 0.00011535325560374049, 'samples': 17761792, 'steps': 34690, 'loss/train': 1.4590998888015747} 02/25/2022 22:29:38 - INFO - codeparrot_training - Step 34691: {'lr': 0.00011533946937408948, 'samples': 17762304, 'steps': 34691, 'loss/train': 1.5719199180603027} 02/25/2022 22:29:42 - INFO - codeparrot_training - Step 34692: {'lr': 0.00011532568372128155, 'samples': 17762816, 'steps': 34692, 'loss/train': 1.649390459060669} 02/25/2022 22:29:47 - INFO - codeparrot_training - Step 34693: {'lr': 0.00011531189864537558, 'samples': 17763328, 'steps': 34693, 'loss/train': 2.069047451019287} 02/25/2022 22:29:50 - INFO - codeparrot_training - Step 34694: {'lr': 0.0001152981141464307, 'samples': 17763840, 'steps': 34694, 'loss/train': 1.9279448986053467} 02/25/2022 22:29:57 - INFO - codeparrot_training - Step 34695: {'lr': 0.0001152843302245059, 'samples': 17764352, 'steps': 34695, 'loss/train': 1.2418310642242432} 02/25/2022 22:30:00 - INFO - codeparrot_training - Step 34696: {'lr': 0.00011527054687966021, 'samples': 17764864, 'steps': 34696, 'loss/train': 0.9884203672409058} 02/25/2022 22:30:05 - INFO - codeparrot_training - Step 34697: {'lr': 0.00011525676411195277, 'samples': 17765376, 'steps': 34697, 'loss/train': 1.4402081966400146} 02/25/2022 22:30:09 - INFO - codeparrot_training - Step 34698: {'lr': 0.00011524298192144256, 'samples': 17765888, 'steps': 34698, 'loss/train': 1.8921897411346436} 02/25/2022 22:30:14 - INFO - codeparrot_training - Step 34699: {'lr': 0.00011522920030818865, 'samples': 17766400, 'steps': 34699, 'loss/train': 2.6496877670288086} 02/25/2022 22:30:18 - INFO - codeparrot_training - Step 34700: {'lr': 0.00011521541927224994, 'samples': 17766912, 'steps': 34700, 'loss/train': 1.0045610666275024} 02/25/2022 22:30:23 - INFO - codeparrot_training - Step 34701: {'lr': 0.00011520163881368568, 'samples': 17767424, 'steps': 34701, 'loss/train': 2.375739812850952} 02/25/2022 22:30:27 - INFO - codeparrot_training - Step 34702: {'lr': 0.0001151878589325548, 'samples': 17767936, 'steps': 34702, 'loss/train': 2.2931594848632812} 02/25/2022 22:30:32 - INFO - codeparrot_training - Step 34703: {'lr': 0.0001151740796289163, 'samples': 17768448, 'steps': 34703, 'loss/train': 2.1976895332336426} 02/25/2022 22:30:36 - INFO - codeparrot_training - Step 34704: {'lr': 0.00011516030090282914, 'samples': 17768960, 'steps': 34704, 'loss/train': 3.053875684738159} 02/25/2022 22:30:41 - INFO - codeparrot_training - Step 34705: {'lr': 0.00011514652275435255, 'samples': 17769472, 'steps': 34705, 'loss/train': 1.0335137844085693} 02/25/2022 22:30:45 - INFO - codeparrot_training - Step 34706: {'lr': 0.00011513274518354544, 'samples': 17769984, 'steps': 34706, 'loss/train': 2.4896655082702637} 02/25/2022 22:30:50 - INFO - codeparrot_training - Step 34707: {'lr': 0.0001151189681904668, 'samples': 17770496, 'steps': 34707, 'loss/train': 1.9550503492355347} 02/25/2022 22:30:54 - INFO - codeparrot_training - Step 34708: {'lr': 0.00011510519177517559, 'samples': 17771008, 'steps': 34708, 'loss/train': 0.1877146065235138} 02/25/2022 22:30:59 - INFO - codeparrot_training - Step 34709: {'lr': 0.00011509141593773098, 'samples': 17771520, 'steps': 34709, 'loss/train': 1.8265786170959473} 02/25/2022 22:31:03 - INFO - codeparrot_training - Step 34710: {'lr': 0.00011507764067819185, 'samples': 17772032, 'steps': 34710, 'loss/train': 1.7760233879089355} 02/25/2022 22:31:10 - INFO - codeparrot_training - Step 34711: {'lr': 0.00011506386599661741, 'samples': 17772544, 'steps': 34711, 'loss/train': 1.6839982271194458} 02/25/2022 22:31:13 - INFO - codeparrot_training - Step 34712: {'lr': 0.00011505009189306636, 'samples': 17773056, 'steps': 34712, 'loss/train': 1.2640389204025269} 02/25/2022 22:31:19 - INFO - codeparrot_training - Step 34713: {'lr': 0.00011503631836759795, 'samples': 17773568, 'steps': 34713, 'loss/train': 1.7982176542282104} 02/25/2022 22:31:22 - INFO - codeparrot_training - Step 34714: {'lr': 0.00011502254542027102, 'samples': 17774080, 'steps': 34714, 'loss/train': 1.723174810409546} 02/25/2022 22:31:28 - INFO - codeparrot_training - Step 34715: {'lr': 0.00011500877305114479, 'samples': 17774592, 'steps': 34715, 'loss/train': 1.2586579322814941} 02/25/2022 22:31:32 - INFO - codeparrot_training - Step 34716: {'lr': 0.00011499500126027798, 'samples': 17775104, 'steps': 34716, 'loss/train': 0.7825419902801514} 02/25/2022 22:31:37 - INFO - codeparrot_training - Step 34717: {'lr': 0.00011498123004772976, 'samples': 17775616, 'steps': 34717, 'loss/train': 1.3392101526260376} 02/25/2022 22:31:41 - INFO - codeparrot_training - Step 34718: {'lr': 0.00011496745941355902, 'samples': 17776128, 'steps': 34718, 'loss/train': 2.122418165206909} 02/25/2022 22:31:46 - INFO - codeparrot_training - Step 34719: {'lr': 0.00011495368935782497, 'samples': 17776640, 'steps': 34719, 'loss/train': 2.154677391052246} 02/25/2022 22:31:50 - INFO - codeparrot_training - Step 34720: {'lr': 0.00011493991988058625, 'samples': 17777152, 'steps': 34720, 'loss/train': 1.484230399131775} 02/25/2022 22:31:56 - INFO - codeparrot_training - Step 34721: {'lr': 0.00011492615098190213, 'samples': 17777664, 'steps': 34721, 'loss/train': 2.2542247772216797} 02/25/2022 22:32:00 - INFO - codeparrot_training - Step 34722: {'lr': 0.00011491238266183138, 'samples': 17778176, 'steps': 34722, 'loss/train': 2.3974907398223877} 02/25/2022 22:32:05 - INFO - codeparrot_training - Step 34723: {'lr': 0.00011489861492043319, 'samples': 17778688, 'steps': 34723, 'loss/train': 1.3617750406265259} 02/25/2022 22:32:09 - INFO - codeparrot_training - Step 34724: {'lr': 0.00011488484775776645, 'samples': 17779200, 'steps': 34724, 'loss/train': 2.085836887359619} 02/25/2022 22:32:14 - INFO - codeparrot_training - Step 34725: {'lr': 0.00011487108117389008, 'samples': 17779712, 'steps': 34725, 'loss/train': 1.6692780256271362} 02/25/2022 22:32:18 - INFO - codeparrot_training - Step 34726: {'lr': 0.00011485731516886303, 'samples': 17780224, 'steps': 34726, 'loss/train': 0.5327391028404236} 02/25/2022 22:32:23 - INFO - codeparrot_training - Step 34727: {'lr': 0.00011484354974274441, 'samples': 17780736, 'steps': 34727, 'loss/train': 1.2682796716690063} 02/25/2022 22:32:27 - INFO - codeparrot_training - Step 34728: {'lr': 0.00011482978489559312, 'samples': 17781248, 'steps': 34728, 'loss/train': 2.6547932624816895} 02/25/2022 22:32:32 - INFO - codeparrot_training - Step 34729: {'lr': 0.0001148160206274681, 'samples': 17781760, 'steps': 34729, 'loss/train': 1.0286450386047363} 02/25/2022 22:32:36 - INFO - codeparrot_training - Step 34730: {'lr': 0.00011480225693842822, 'samples': 17782272, 'steps': 34730, 'loss/train': 1.509169101715088} 02/25/2022 22:32:42 - INFO - codeparrot_training - Step 34731: {'lr': 0.00011478849382853263, 'samples': 17782784, 'steps': 34731, 'loss/train': 1.5795807838439941} 02/25/2022 22:32:45 - INFO - codeparrot_training - Step 34732: {'lr': 0.00011477473129784022, 'samples': 17783296, 'steps': 34732, 'loss/train': 1.7360645532608032} 02/25/2022 22:32:51 - INFO - codeparrot_training - Step 34733: {'lr': 0.00011476096934640992, 'samples': 17783808, 'steps': 34733, 'loss/train': 1.1455860137939453} 02/25/2022 22:32:54 - INFO - codeparrot_training - Step 34734: {'lr': 0.00011474720797430057, 'samples': 17784320, 'steps': 34734, 'loss/train': 1.6169661283493042} 02/25/2022 22:33:00 - INFO - codeparrot_training - Step 34735: {'lr': 0.00011473344718157133, 'samples': 17784832, 'steps': 34735, 'loss/train': 1.66375732421875} 02/25/2022 22:33:03 - INFO - codeparrot_training - Step 34736: {'lr': 0.00011471968696828106, 'samples': 17785344, 'steps': 34736, 'loss/train': 3.9603748321533203} 02/25/2022 22:33:09 - INFO - codeparrot_training - Step 34737: {'lr': 0.0001147059273344887, 'samples': 17785856, 'steps': 34737, 'loss/train': 1.7098735570907593} 02/25/2022 22:33:12 - INFO - codeparrot_training - Step 34738: {'lr': 0.00011469216828025309, 'samples': 17786368, 'steps': 34738, 'loss/train': 0.5596854090690613} 02/25/2022 22:33:18 - INFO - codeparrot_training - Step 34739: {'lr': 0.00011467840980563335, 'samples': 17786880, 'steps': 34739, 'loss/train': 2.477786064147949} 02/25/2022 22:33:21 - INFO - codeparrot_training - Step 34740: {'lr': 0.00011466465191068834, 'samples': 17787392, 'steps': 34740, 'loss/train': 0.8669219017028809} 02/25/2022 22:33:28 - INFO - codeparrot_training - Step 34741: {'lr': 0.00011465089459547698, 'samples': 17787904, 'steps': 34741, 'loss/train': 1.8605616092681885} 02/25/2022 22:33:31 - INFO - codeparrot_training - Step 34742: {'lr': 0.00011463713786005813, 'samples': 17788416, 'steps': 34742, 'loss/train': 1.132358193397522} 02/25/2022 22:33:37 - INFO - codeparrot_training - Step 34743: {'lr': 0.00011462338170449088, 'samples': 17788928, 'steps': 34743, 'loss/train': 1.6693097352981567} 02/25/2022 22:33:40 - INFO - codeparrot_training - Step 34744: {'lr': 0.00011460962612883408, 'samples': 17789440, 'steps': 34744, 'loss/train': 2.387767791748047} 02/25/2022 22:33:46 - INFO - codeparrot_training - Step 34745: {'lr': 0.00011459587113314665, 'samples': 17789952, 'steps': 34745, 'loss/train': 1.6871999502182007} 02/25/2022 22:33:49 - INFO - codeparrot_training - Step 34746: {'lr': 0.00011458211671748753, 'samples': 17790464, 'steps': 34746, 'loss/train': 2.442793607711792} 02/25/2022 22:33:55 - INFO - codeparrot_training - Step 34747: {'lr': 0.0001145683628819155, 'samples': 17790976, 'steps': 34747, 'loss/train': 1.7778996229171753} 02/25/2022 22:33:58 - INFO - codeparrot_training - Step 34748: {'lr': 0.0001145546096264897, 'samples': 17791488, 'steps': 34748, 'loss/train': 0.8230723142623901} 02/25/2022 22:34:04 - INFO - codeparrot_training - Step 34749: {'lr': 0.00011454085695126887, 'samples': 17792000, 'steps': 34749, 'loss/train': 2.565108299255371} 02/25/2022 22:34:07 - INFO - codeparrot_training - Step 34750: {'lr': 0.00011452710485631215, 'samples': 17792512, 'steps': 34750, 'loss/train': 1.8774287700653076} 02/25/2022 22:34:13 - INFO - codeparrot_training - Step 34751: {'lr': 0.00011451335334167815, 'samples': 17793024, 'steps': 34751, 'loss/train': 0.710401713848114} 02/25/2022 22:34:16 - INFO - codeparrot_training - Step 34752: {'lr': 0.00011449960240742596, 'samples': 17793536, 'steps': 34752, 'loss/train': 1.6455191373825073} 02/25/2022 22:34:22 - INFO - codeparrot_training - Step 34753: {'lr': 0.00011448585205361439, 'samples': 17794048, 'steps': 34753, 'loss/train': 0.8933877348899841} 02/25/2022 22:34:25 - INFO - codeparrot_training - Step 34754: {'lr': 0.00011447210228030258, 'samples': 17794560, 'steps': 34754, 'loss/train': 2.0767128467559814} 02/25/2022 22:34:31 - INFO - codeparrot_training - Step 34755: {'lr': 0.00011445835308754906, 'samples': 17795072, 'steps': 34755, 'loss/train': 1.4412498474121094} 02/25/2022 22:34:34 - INFO - codeparrot_training - Step 34756: {'lr': 0.00011444460447541302, 'samples': 17795584, 'steps': 34756, 'loss/train': 2.297098159790039} 02/25/2022 22:34:40 - INFO - codeparrot_training - Step 34757: {'lr': 0.00011443085644395315, 'samples': 17796096, 'steps': 34757, 'loss/train': 1.4687154293060303} 02/25/2022 22:34:44 - INFO - codeparrot_training - Step 34758: {'lr': 0.00011441710899322863, 'samples': 17796608, 'steps': 34758, 'loss/train': 0.5483139753341675} 02/25/2022 22:34:50 - INFO - codeparrot_training - Step 34759: {'lr': 0.00011440336212329796, 'samples': 17797120, 'steps': 34759, 'loss/train': 1.6640715599060059} 02/25/2022 22:34:53 - INFO - codeparrot_training - Step 34760: {'lr': 0.00011438961583422036, 'samples': 17797632, 'steps': 34760, 'loss/train': 1.5567317008972168} 02/25/2022 22:34:59 - INFO - codeparrot_training - Step 34761: {'lr': 0.00011437587012605446, 'samples': 17798144, 'steps': 34761, 'loss/train': 2.008934497833252} 02/25/2022 22:35:02 - INFO - codeparrot_training - Step 34762: {'lr': 0.00011436212499885937, 'samples': 17798656, 'steps': 34762, 'loss/train': 1.55619215965271} 02/25/2022 22:35:08 - INFO - codeparrot_training - Step 34763: {'lr': 0.00011434838045269388, 'samples': 17799168, 'steps': 34763, 'loss/train': 1.3606946468353271} 02/25/2022 22:35:11 - INFO - codeparrot_training - Step 34764: {'lr': 0.00011433463648761682, 'samples': 17799680, 'steps': 34764, 'loss/train': 1.671073317527771} 02/25/2022 22:35:17 - INFO - codeparrot_training - Step 34765: {'lr': 0.00011432089310368704, 'samples': 17800192, 'steps': 34765, 'loss/train': 2.27286434173584} 02/25/2022 22:35:20 - INFO - codeparrot_training - Step 34766: {'lr': 0.00011430715030096356, 'samples': 17800704, 'steps': 34766, 'loss/train': 1.3739622831344604} 02/25/2022 22:35:26 - INFO - codeparrot_training - Step 34767: {'lr': 0.00011429340807950514, 'samples': 17801216, 'steps': 34767, 'loss/train': 0.9082155227661133} 02/25/2022 22:35:29 - INFO - codeparrot_training - Step 34768: {'lr': 0.00011427966643937068, 'samples': 17801728, 'steps': 34768, 'loss/train': 2.0918383598327637} 02/25/2022 22:35:35 - INFO - codeparrot_training - Step 34769: {'lr': 0.00011426592538061895, 'samples': 17802240, 'steps': 34769, 'loss/train': 1.7435027360916138} 02/25/2022 22:35:39 - INFO - codeparrot_training - Step 34770: {'lr': 0.00011425218490330896, 'samples': 17802752, 'steps': 34770, 'loss/train': 0.36293432116508484} 02/25/2022 22:35:44 - INFO - codeparrot_training - Step 34771: {'lr': 0.00011423844500749944, 'samples': 17803264, 'steps': 34771, 'loss/train': 2.0686099529266357} 02/25/2022 22:35:48 - INFO - codeparrot_training - Step 34772: {'lr': 0.00011422470569324949, 'samples': 17803776, 'steps': 34772, 'loss/train': 0.8506045937538147} 02/25/2022 22:35:53 - INFO - codeparrot_training - Step 34773: {'lr': 0.00011421096696061758, 'samples': 17804288, 'steps': 34773, 'loss/train': 0.5906069874763489} 02/25/2022 22:35:57 - INFO - codeparrot_training - Step 34774: {'lr': 0.00011419722880966291, 'samples': 17804800, 'steps': 34774, 'loss/train': 1.8113291263580322} 02/25/2022 22:36:02 - INFO - codeparrot_training - Step 34775: {'lr': 0.00011418349124044405, 'samples': 17805312, 'steps': 34775, 'loss/train': 1.4793859720230103} 02/25/2022 22:36:06 - INFO - codeparrot_training - Step 34776: {'lr': 0.0001141697542530202, 'samples': 17805824, 'steps': 34776, 'loss/train': 1.440613031387329} 02/25/2022 22:36:12 - INFO - codeparrot_training - Step 34777: {'lr': 0.0001141560178474498, 'samples': 17806336, 'steps': 34777, 'loss/train': 1.905440092086792} 02/25/2022 22:36:15 - INFO - codeparrot_training - Step 34778: {'lr': 0.00011414228202379195, 'samples': 17806848, 'steps': 34778, 'loss/train': 1.4701566696166992} 02/25/2022 22:36:21 - INFO - codeparrot_training - Step 34779: {'lr': 0.00011412854678210546, 'samples': 17807360, 'steps': 34779, 'loss/train': 2.088491678237915} 02/25/2022 22:36:25 - INFO - codeparrot_training - Step 34780: {'lr': 0.00011411481212244912, 'samples': 17807872, 'steps': 34780, 'loss/train': 1.9860289096832275} 02/25/2022 22:36:30 - INFO - codeparrot_training - Step 34781: {'lr': 0.00011410107804488168, 'samples': 17808384, 'steps': 34781, 'loss/train': 0.901138961315155} 02/25/2022 22:36:34 - INFO - codeparrot_training - Step 34782: {'lr': 0.00011408734454946218, 'samples': 17808896, 'steps': 34782, 'loss/train': 2.3694934844970703} 02/25/2022 22:36:39 - INFO - codeparrot_training - Step 34783: {'lr': 0.00011407361163624933, 'samples': 17809408, 'steps': 34783, 'loss/train': 1.860734462738037} 02/25/2022 22:36:43 - INFO - codeparrot_training - Step 34784: {'lr': 0.00011405987930530184, 'samples': 17809920, 'steps': 34784, 'loss/train': 1.1248018741607666} 02/25/2022 22:36:48 - INFO - codeparrot_training - Step 34785: {'lr': 0.0001140461475566788, 'samples': 17810432, 'steps': 34785, 'loss/train': 1.4287890195846558} 02/25/2022 22:36:52 - INFO - codeparrot_training - Step 34786: {'lr': 0.00011403241639043887, 'samples': 17810944, 'steps': 34786, 'loss/train': 3.293961763381958} 02/25/2022 22:36:57 - INFO - codeparrot_training - Step 34787: {'lr': 0.0001140186858066409, 'samples': 17811456, 'steps': 34787, 'loss/train': 1.580237865447998} 02/25/2022 22:37:01 - INFO - codeparrot_training - Step 34788: {'lr': 0.00011400495580534362, 'samples': 17811968, 'steps': 34788, 'loss/train': 1.9166027307510376} 02/25/2022 22:37:07 - INFO - codeparrot_training - Step 34789: {'lr': 0.000113991226386606, 'samples': 17812480, 'steps': 34789, 'loss/train': 3.150430917739868} 02/25/2022 22:37:10 - INFO - codeparrot_training - Step 34790: {'lr': 0.00011397749755048678, 'samples': 17812992, 'steps': 34790, 'loss/train': 2.3032455444335938} 02/25/2022 22:37:16 - INFO - codeparrot_training - Step 34791: {'lr': 0.00011396376929704477, 'samples': 17813504, 'steps': 34791, 'loss/train': 1.786041021347046} 02/25/2022 22:37:19 - INFO - codeparrot_training - Step 34792: {'lr': 0.00011395004162633868, 'samples': 17814016, 'steps': 34792, 'loss/train': 1.5370835065841675} 02/25/2022 22:37:25 - INFO - codeparrot_training - Step 34793: {'lr': 0.00011393631453842763, 'samples': 17814528, 'steps': 34793, 'loss/train': 1.2027186155319214} 02/25/2022 22:37:28 - INFO - codeparrot_training - Step 34794: {'lr': 0.00011392258803336996, 'samples': 17815040, 'steps': 34794, 'loss/train': 1.5637550354003906} 02/25/2022 22:37:34 - INFO - codeparrot_training - Step 34795: {'lr': 0.00011390886211122484, 'samples': 17815552, 'steps': 34795, 'loss/train': 0.5180971026420593} 02/25/2022 22:37:38 - INFO - codeparrot_training - Step 34796: {'lr': 0.00011389513677205084, 'samples': 17816064, 'steps': 34796, 'loss/train': 0.08174548298120499} 02/25/2022 22:37:43 - INFO - codeparrot_training - Step 34797: {'lr': 0.00011388141201590696, 'samples': 17816576, 'steps': 34797, 'loss/train': 0.4705628454685211} 02/25/2022 22:37:47 - INFO - codeparrot_training - Step 34798: {'lr': 0.0001138676878428519, 'samples': 17817088, 'steps': 34798, 'loss/train': 0.3046160936355591} 02/25/2022 22:37:52 - INFO - codeparrot_training - Step 34799: {'lr': 0.00011385396425294439, 'samples': 17817600, 'steps': 34799, 'loss/train': 1.660393238067627} 02/25/2022 22:37:56 - INFO - codeparrot_training - Step 34800: {'lr': 0.00011384024124624323, 'samples': 17818112, 'steps': 34800, 'loss/train': 0.9168550968170166} 02/25/2022 22:38:01 - INFO - codeparrot_training - Step 34801: {'lr': 0.00011382651882280728, 'samples': 17818624, 'steps': 34801, 'loss/train': 1.9195979833602905} 02/25/2022 22:38:08 - INFO - codeparrot_training - Step 34802: {'lr': 0.00011381279698269533, 'samples': 17819136, 'steps': 34802, 'loss/train': 2.484697103500366} 02/25/2022 22:38:11 - INFO - codeparrot_training - Step 34803: {'lr': 0.00011379907572596609, 'samples': 17819648, 'steps': 34803, 'loss/train': 2.1058056354522705} 02/25/2022 22:38:17 - INFO - codeparrot_training - Step 34804: {'lr': 0.00011378535505267825, 'samples': 17820160, 'steps': 34804, 'loss/train': 1.8874750137329102} 02/25/2022 22:38:21 - INFO - codeparrot_training - Step 34805: {'lr': 0.00011377163496289084, 'samples': 17820672, 'steps': 34805, 'loss/train': 0.70106041431427} 02/25/2022 22:38:24 - INFO - codeparrot_training - Step 34806: {'lr': 0.00011375791545666245, 'samples': 17821184, 'steps': 34806, 'loss/train': 1.9727753400802612} 02/25/2022 22:38:30 - INFO - codeparrot_training - Step 34807: {'lr': 0.0001137441965340519, 'samples': 17821696, 'steps': 34807, 'loss/train': 1.4847525358200073} 02/25/2022 22:38:33 - INFO - codeparrot_training - Step 34808: {'lr': 0.00011373047819511783, 'samples': 17822208, 'steps': 34808, 'loss/train': 2.4324989318847656} 02/25/2022 22:38:39 - INFO - codeparrot_training - Step 34809: {'lr': 0.00011371676043991924, 'samples': 17822720, 'steps': 34809, 'loss/train': 2.353549003601074} 02/25/2022 22:38:42 - INFO - codeparrot_training - Step 34810: {'lr': 0.00011370304326851469, 'samples': 17823232, 'steps': 34810, 'loss/train': 1.6818206310272217} 02/25/2022 22:38:48 - INFO - codeparrot_training - Step 34811: {'lr': 0.00011368932668096318, 'samples': 17823744, 'steps': 34811, 'loss/train': 2.0532338619232178} 02/25/2022 22:38:51 - INFO - codeparrot_training - Step 34812: {'lr': 0.00011367561067732316, 'samples': 17824256, 'steps': 34812, 'loss/train': 2.611616373062134} 02/25/2022 22:38:57 - INFO - codeparrot_training - Step 34813: {'lr': 0.00011366189525765361, 'samples': 17824768, 'steps': 34813, 'loss/train': 2.118795871734619} 02/25/2022 22:39:00 - INFO - codeparrot_training - Step 34814: {'lr': 0.00011364818042201311, 'samples': 17825280, 'steps': 34814, 'loss/train': 2.0578081607818604} 02/25/2022 22:39:07 - INFO - codeparrot_training - Step 34815: {'lr': 0.00011363446617046069, 'samples': 17825792, 'steps': 34815, 'loss/train': 0.41846537590026855} 02/25/2022 22:39:10 - INFO - codeparrot_training - Step 34816: {'lr': 0.00011362075250305473, 'samples': 17826304, 'steps': 34816, 'loss/train': 1.268330454826355} 02/25/2022 22:39:16 - INFO - codeparrot_training - Step 34817: {'lr': 0.00011360703941985429, 'samples': 17826816, 'steps': 34817, 'loss/train': 2.0518436431884766} 02/25/2022 22:39:19 - INFO - codeparrot_training - Step 34818: {'lr': 0.00011359332692091784, 'samples': 17827328, 'steps': 34818, 'loss/train': 0.8262566328048706} 02/25/2022 22:39:25 - INFO - codeparrot_training - Step 34819: {'lr': 0.00011357961500630439, 'samples': 17827840, 'steps': 34819, 'loss/train': 0.8251410126686096} 02/25/2022 22:39:28 - INFO - codeparrot_training - Step 34820: {'lr': 0.00011356590367607253, 'samples': 17828352, 'steps': 34820, 'loss/train': 1.5320590734481812} 02/25/2022 22:39:34 - INFO - codeparrot_training - Step 34821: {'lr': 0.00011355219293028104, 'samples': 17828864, 'steps': 34821, 'loss/train': 2.3063979148864746} 02/25/2022 22:39:37 - INFO - codeparrot_training - Step 34822: {'lr': 0.0001135384827689885, 'samples': 17829376, 'steps': 34822, 'loss/train': 1.833272933959961} 02/25/2022 22:39:43 - INFO - codeparrot_training - Step 34823: {'lr': 0.00011352477319225391, 'samples': 17829888, 'steps': 34823, 'loss/train': 0.49192261695861816} 02/25/2022 22:39:46 - INFO - codeparrot_training - Step 34824: {'lr': 0.00011351106420013583, 'samples': 17830400, 'steps': 34824, 'loss/train': 1.8050167560577393} 02/25/2022 22:39:53 - INFO - codeparrot_training - Step 34825: {'lr': 0.000113497355792693, 'samples': 17830912, 'steps': 34825, 'loss/train': 1.7704142332077026} 02/25/2022 22:39:56 - INFO - codeparrot_training - Step 34826: {'lr': 0.00011348364796998417, 'samples': 17831424, 'steps': 34826, 'loss/train': 2.008364677429199} 02/25/2022 22:40:02 - INFO - codeparrot_training - Step 34827: {'lr': 0.00011346994073206792, 'samples': 17831936, 'steps': 34827, 'loss/train': 1.5049372911453247} 02/25/2022 22:40:06 - INFO - codeparrot_training - Step 34828: {'lr': 0.00011345623407900322, 'samples': 17832448, 'steps': 34828, 'loss/train': 2.518047332763672} 02/25/2022 22:40:11 - INFO - codeparrot_training - Step 34829: {'lr': 0.00011344252801084864, 'samples': 17832960, 'steps': 34829, 'loss/train': 2.15262508392334} 02/25/2022 22:40:14 - INFO - codeparrot_training - Step 34830: {'lr': 0.0001134288225276629, 'samples': 17833472, 'steps': 34830, 'loss/train': 1.8653783798217773} 02/25/2022 22:40:20 - INFO - codeparrot_training - Step 34831: {'lr': 0.00011341511762950466, 'samples': 17833984, 'steps': 34831, 'loss/train': 1.3218104839324951} 02/25/2022 22:40:26 - INFO - codeparrot_training - Step 34832: {'lr': 0.00011340141331643275, 'samples': 17834496, 'steps': 34832, 'loss/train': 2.3935933113098145} 02/25/2022 22:40:29 - INFO - codeparrot_training - Step 34833: {'lr': 0.00011338770958850581, 'samples': 17835008, 'steps': 34833, 'loss/train': 2.6254355907440186} 02/25/2022 22:40:35 - INFO - codeparrot_training - Step 34834: {'lr': 0.00011337400644578255, 'samples': 17835520, 'steps': 34834, 'loss/train': 2.383310556411743} 02/25/2022 22:40:39 - INFO - codeparrot_training - Step 34835: {'lr': 0.00011336030388832156, 'samples': 17836032, 'steps': 34835, 'loss/train': 1.4206146001815796} 02/25/2022 22:40:42 - INFO - codeparrot_training - Step 34836: {'lr': 0.00011334660191618174, 'samples': 17836544, 'steps': 34836, 'loss/train': 0.7846734523773193} 02/25/2022 22:40:48 - INFO - codeparrot_training - Step 34837: {'lr': 0.0001133329005294217, 'samples': 17837056, 'steps': 34837, 'loss/train': 0.32985156774520874} 02/25/2022 22:40:51 - INFO - codeparrot_training - Step 34838: {'lr': 0.0001133191997281001, 'samples': 17837568, 'steps': 34838, 'loss/train': 2.610053300857544} 02/25/2022 22:40:56 - INFO - codeparrot_training - Step 34839: {'lr': 0.00011330549951227556, 'samples': 17838080, 'steps': 34839, 'loss/train': 1.4487978219985962} 02/25/2022 22:41:00 - INFO - codeparrot_training - Step 34840: {'lr': 0.00011329179988200694, 'samples': 17838592, 'steps': 34840, 'loss/train': 1.882456660270691} 02/25/2022 22:41:06 - INFO - codeparrot_training - Step 34841: {'lr': 0.00011327810083735285, 'samples': 17839104, 'steps': 34841, 'loss/train': 1.6914174556732178} 02/25/2022 22:41:10 - INFO - codeparrot_training - Step 34842: {'lr': 0.00011326440237837196, 'samples': 17839616, 'steps': 34842, 'loss/train': 1.529268503189087} 02/25/2022 22:41:15 - INFO - codeparrot_training - Step 34843: {'lr': 0.00011325070450512285, 'samples': 17840128, 'steps': 34843, 'loss/train': 1.900531530380249} 02/25/2022 22:41:19 - INFO - codeparrot_training - Step 34844: {'lr': 0.00011323700721766439, 'samples': 17840640, 'steps': 34844, 'loss/train': 2.039630651473999} 02/25/2022 22:41:24 - INFO - codeparrot_training - Step 34845: {'lr': 0.00011322331051605509, 'samples': 17841152, 'steps': 34845, 'loss/train': 1.405947208404541} 02/25/2022 22:41:28 - INFO - codeparrot_training - Step 34846: {'lr': 0.00011320961440035386, 'samples': 17841664, 'steps': 34846, 'loss/train': 2.322772264480591} 02/25/2022 22:41:33 - INFO - codeparrot_training - Step 34847: {'lr': 0.00011319591887061903, 'samples': 17842176, 'steps': 34847, 'loss/train': 1.8700200319290161} 02/25/2022 22:41:37 - INFO - codeparrot_training - Step 34848: {'lr': 0.00011318222392690956, 'samples': 17842688, 'steps': 34848, 'loss/train': 2.925299644470215} 02/25/2022 22:41:42 - INFO - codeparrot_training - Step 34849: {'lr': 0.00011316852956928386, 'samples': 17843200, 'steps': 34849, 'loss/train': 1.8161739110946655} 02/25/2022 22:41:46 - INFO - codeparrot_training - Step 34850: {'lr': 0.00011315483579780094, 'samples': 17843712, 'steps': 34850, 'loss/train': 1.7367805242538452} 02/25/2022 22:41:52 - INFO - codeparrot_training - Step 34851: {'lr': 0.00011314114261251906, 'samples': 17844224, 'steps': 34851, 'loss/train': 2.5559041500091553} 02/25/2022 22:41:56 - INFO - codeparrot_training - Step 34852: {'lr': 0.00011312745001349717, 'samples': 17844736, 'steps': 34852, 'loss/train': 2.043379545211792} 02/25/2022 22:42:01 - INFO - codeparrot_training - Step 34853: {'lr': 0.00011311375800079371, 'samples': 17845248, 'steps': 34853, 'loss/train': 0.7614732384681702} 02/25/2022 22:42:05 - INFO - codeparrot_training - Step 34854: {'lr': 0.00011310006657446764, 'samples': 17845760, 'steps': 34854, 'loss/train': 0.9589431285858154} 02/25/2022 22:42:10 - INFO - codeparrot_training - Step 34855: {'lr': 0.00011308637573457725, 'samples': 17846272, 'steps': 34855, 'loss/train': 1.9918023347854614} 02/25/2022 22:42:14 - INFO - codeparrot_training - Step 34856: {'lr': 0.00011307268548118141, 'samples': 17846784, 'steps': 34856, 'loss/train': 0.6873029470443726} 02/25/2022 22:42:19 - INFO - codeparrot_training - Step 34857: {'lr': 0.00011305899581433862, 'samples': 17847296, 'steps': 34857, 'loss/train': 1.7289001941680908} 02/25/2022 22:42:23 - INFO - codeparrot_training - Step 34858: {'lr': 0.00011304530673410773, 'samples': 17847808, 'steps': 34858, 'loss/train': 1.7044283151626587} 02/25/2022 22:42:28 - INFO - codeparrot_training - Step 34859: {'lr': 0.00011303161824054723, 'samples': 17848320, 'steps': 34859, 'loss/train': 2.1292171478271484} 02/25/2022 22:42:32 - INFO - codeparrot_training - Step 34860: {'lr': 0.00011301793033371579, 'samples': 17848832, 'steps': 34860, 'loss/train': 1.9011460542678833} 02/25/2022 22:42:38 - INFO - codeparrot_training - Step 34861: {'lr': 0.00011300424301367196, 'samples': 17849344, 'steps': 34861, 'loss/train': 2.2141149044036865} 02/25/2022 22:42:41 - INFO - codeparrot_training - Step 34862: {'lr': 0.00011299055628047452, 'samples': 17849856, 'steps': 34862, 'loss/train': 1.8247283697128296} 02/25/2022 22:42:47 - INFO - codeparrot_training - Step 34863: {'lr': 0.00011297687013418206, 'samples': 17850368, 'steps': 34863, 'loss/train': 1.9151647090911865} 02/25/2022 22:42:50 - INFO - codeparrot_training - Step 34864: {'lr': 0.00011296318457485316, 'samples': 17850880, 'steps': 34864, 'loss/train': 2.1352217197418213} 02/25/2022 22:42:56 - INFO - codeparrot_training - Step 34865: {'lr': 0.00011294949960254639, 'samples': 17851392, 'steps': 34865, 'loss/train': 1.6991958618164062} 02/25/2022 22:42:59 - INFO - codeparrot_training - Step 34866: {'lr': 0.00011293581521732052, 'samples': 17851904, 'steps': 34866, 'loss/train': 1.3736205101013184} 02/25/2022 22:43:05 - INFO - codeparrot_training - Step 34867: {'lr': 0.00011292213141923407, 'samples': 17852416, 'steps': 34867, 'loss/train': 2.411273717880249} 02/25/2022 22:43:08 - INFO - codeparrot_training - Step 34868: {'lr': 0.00011290844820834572, 'samples': 17852928, 'steps': 34868, 'loss/train': 1.9816040992736816} 02/25/2022 22:43:14 - INFO - codeparrot_training - Step 34869: {'lr': 0.00011289476558471396, 'samples': 17853440, 'steps': 34869, 'loss/train': 2.042404890060425} 02/25/2022 22:43:17 - INFO - codeparrot_training - Step 34870: {'lr': 0.00011288108354839755, 'samples': 17853952, 'steps': 34870, 'loss/train': 2.1103172302246094} 02/25/2022 22:43:23 - INFO - codeparrot_training - Step 34871: {'lr': 0.00011286740209945507, 'samples': 17854464, 'steps': 34871, 'loss/train': 1.5293422937393188} 02/25/2022 22:43:27 - INFO - codeparrot_training - Step 34872: {'lr': 0.00011285372123794507, 'samples': 17854976, 'steps': 34872, 'loss/train': 0.029536200687289238} 02/25/2022 22:43:32 - INFO - codeparrot_training - Step 34873: {'lr': 0.00011284004096392619, 'samples': 17855488, 'steps': 34873, 'loss/train': 1.5277615785598755} 02/25/2022 22:43:36 - INFO - codeparrot_training - Step 34874: {'lr': 0.0001128263612774569, 'samples': 17856000, 'steps': 34874, 'loss/train': 2.2245821952819824} 02/25/2022 22:43:41 - INFO - codeparrot_training - Step 34875: {'lr': 0.00011281268217859606, 'samples': 17856512, 'steps': 34875, 'loss/train': 2.654569387435913} 02/25/2022 22:43:45 - INFO - codeparrot_training - Step 34876: {'lr': 0.00011279900366740209, 'samples': 17857024, 'steps': 34876, 'loss/train': 1.8947581052780151} 02/25/2022 22:43:50 - INFO - codeparrot_training - Step 34877: {'lr': 0.00011278532574393366, 'samples': 17857536, 'steps': 34877, 'loss/train': 2.1822595596313477} 02/25/2022 22:43:54 - INFO - codeparrot_training - Step 34878: {'lr': 0.0001127716484082492, 'samples': 17858048, 'steps': 34878, 'loss/train': 2.4381775856018066} 02/25/2022 22:43:59 - INFO - codeparrot_training - Step 34879: {'lr': 0.00011275797166040752, 'samples': 17858560, 'steps': 34879, 'loss/train': 2.6193206310272217} 02/25/2022 22:44:03 - INFO - codeparrot_training - Step 34880: {'lr': 0.00011274429550046702, 'samples': 17859072, 'steps': 34880, 'loss/train': 2.463113784790039} 02/25/2022 22:44:08 - INFO - codeparrot_training - Step 34881: {'lr': 0.00011273061992848657, 'samples': 17859584, 'steps': 34881, 'loss/train': 1.1435399055480957} 02/25/2022 22:44:12 - INFO - codeparrot_training - Step 34882: {'lr': 0.00011271694494452436, 'samples': 17860096, 'steps': 34882, 'loss/train': 1.7033058404922485} 02/25/2022 22:44:17 - INFO - codeparrot_training - Step 34883: {'lr': 0.00011270327054863927, 'samples': 17860608, 'steps': 34883, 'loss/train': 1.3589262962341309} 02/25/2022 22:44:21 - INFO - codeparrot_training - Step 34884: {'lr': 0.00011268959674088966, 'samples': 17861120, 'steps': 34884, 'loss/train': 1.501327395439148} 02/25/2022 22:44:26 - INFO - codeparrot_training - Step 34885: {'lr': 0.00011267592352133444, 'samples': 17861632, 'steps': 34885, 'loss/train': 2.5504422187805176} 02/25/2022 22:44:30 - INFO - codeparrot_training - Step 34886: {'lr': 0.00011266225089003173, 'samples': 17862144, 'steps': 34886, 'loss/train': 0.5927507281303406} 02/25/2022 22:44:36 - INFO - codeparrot_training - Step 34887: {'lr': 0.00011264857884704045, 'samples': 17862656, 'steps': 34887, 'loss/train': 2.2058801651000977} 02/25/2022 22:44:39 - INFO - codeparrot_training - Step 34888: {'lr': 0.00011263490739241894, 'samples': 17863168, 'steps': 34888, 'loss/train': 1.3723721504211426} 02/25/2022 22:44:45 - INFO - codeparrot_training - Step 34889: {'lr': 0.00011262123652622605, 'samples': 17863680, 'steps': 34889, 'loss/train': 1.5350292921066284} 02/25/2022 22:44:48 - INFO - codeparrot_training - Step 34890: {'lr': 0.00011260756624851997, 'samples': 17864192, 'steps': 34890, 'loss/train': 2.0353267192840576} 02/25/2022 22:44:54 - INFO - codeparrot_training - Step 34891: {'lr': 0.00011259389655935953, 'samples': 17864704, 'steps': 34891, 'loss/train': 2.067101240158081} 02/25/2022 22:44:57 - INFO - codeparrot_training - Step 34892: {'lr': 0.00011258022745880315, 'samples': 17865216, 'steps': 34892, 'loss/train': 2.164034128189087} 02/25/2022 22:45:03 - INFO - codeparrot_training - Step 34893: {'lr': 0.00011256655894690951, 'samples': 17865728, 'steps': 34893, 'loss/train': 0.8312376737594604} 02/25/2022 22:45:06 - INFO - codeparrot_training - Step 34894: {'lr': 0.00011255289102373708, 'samples': 17866240, 'steps': 34894, 'loss/train': 0.8810471296310425} 02/25/2022 22:45:12 - INFO - codeparrot_training - Step 34895: {'lr': 0.00011253922368934446, 'samples': 17866752, 'steps': 34895, 'loss/train': 2.441986560821533} 02/25/2022 22:45:15 - INFO - codeparrot_training - Step 34896: {'lr': 0.00011252555694379005, 'samples': 17867264, 'steps': 34896, 'loss/train': 2.1777114868164062} 02/25/2022 22:45:22 - INFO - codeparrot_training - Step 34897: {'lr': 0.00011251189078713255, 'samples': 17867776, 'steps': 34897, 'loss/train': 0.8940674066543579} 02/25/2022 22:45:25 - INFO - codeparrot_training - Step 34898: {'lr': 0.00011249822521943051, 'samples': 17868288, 'steps': 34898, 'loss/train': 2.1172637939453125} 02/25/2022 22:45:31 - INFO - codeparrot_training - Step 34899: {'lr': 0.00011248456024074239, 'samples': 17868800, 'steps': 34899, 'loss/train': 1.5558733940124512} 02/25/2022 22:45:34 - INFO - codeparrot_training - Step 34900: {'lr': 0.00011247089585112666, 'samples': 17869312, 'steps': 34900, 'loss/train': 1.7457056045532227} 02/25/2022 22:45:40 - INFO - codeparrot_training - Step 34901: {'lr': 0.000112457232050642, 'samples': 17869824, 'steps': 34901, 'loss/train': 2.31425142288208} 02/25/2022 22:45:43 - INFO - codeparrot_training - Step 34902: {'lr': 0.00011244356883934692, 'samples': 17870336, 'steps': 34902, 'loss/train': 2.0756053924560547} 02/25/2022 22:45:49 - INFO - codeparrot_training - Step 34903: {'lr': 0.0001124299062172999, 'samples': 17870848, 'steps': 34903, 'loss/train': 2.2840089797973633} 02/25/2022 22:45:52 - INFO - codeparrot_training - Step 34904: {'lr': 0.0001124162441845594, 'samples': 17871360, 'steps': 34904, 'loss/train': 1.2529807090759277} 02/25/2022 22:45:58 - INFO - codeparrot_training - Step 34905: {'lr': 0.0001124025827411841, 'samples': 17871872, 'steps': 34905, 'loss/train': 1.1864484548568726} 02/25/2022 22:46:01 - INFO - codeparrot_training - Step 34906: {'lr': 0.00011238892188723245, 'samples': 17872384, 'steps': 34906, 'loss/train': 1.8505655527114868} 02/25/2022 22:46:07 - INFO - codeparrot_training - Step 34907: {'lr': 0.00011237526162276296, 'samples': 17872896, 'steps': 34907, 'loss/train': 2.631504774093628} 02/25/2022 22:46:11 - INFO - codeparrot_training - Step 34908: {'lr': 0.00011236160194783404, 'samples': 17873408, 'steps': 34908, 'loss/train': 0.26429206132888794} 02/25/2022 22:46:16 - INFO - codeparrot_training - Step 34909: {'lr': 0.00011234794286250443, 'samples': 17873920, 'steps': 34909, 'loss/train': 1.490645408630371} 02/25/2022 22:46:20 - INFO - codeparrot_training - Step 34910: {'lr': 0.00011233428436683252, 'samples': 17874432, 'steps': 34910, 'loss/train': 1.4881706237792969} 02/25/2022 22:46:25 - INFO - codeparrot_training - Step 34911: {'lr': 0.00011232062646087682, 'samples': 17874944, 'steps': 34911, 'loss/train': 1.62655508518219} 02/25/2022 22:46:31 - INFO - codeparrot_training - Step 34912: {'lr': 0.00011230696914469573, 'samples': 17875456, 'steps': 34912, 'loss/train': 2.2832798957824707} 02/25/2022 22:46:34 - INFO - codeparrot_training - Step 34913: {'lr': 0.00011229331241834798, 'samples': 17875968, 'steps': 34913, 'loss/train': 1.8659603595733643} 02/25/2022 22:46:38 - INFO - codeparrot_training - Step 34914: {'lr': 0.00011227965628189194, 'samples': 17876480, 'steps': 34914, 'loss/train': 1.4666457176208496} 02/25/2022 22:46:43 - INFO - codeparrot_training - Step 34915: {'lr': 0.00011226600073538609, 'samples': 17876992, 'steps': 34915, 'loss/train': 0.9396272301673889} 02/25/2022 22:46:47 - INFO - codeparrot_training - Step 34916: {'lr': 0.0001122523457788889, 'samples': 17877504, 'steps': 34916, 'loss/train': 2.5893049240112305} 02/25/2022 22:46:52 - INFO - codeparrot_training - Step 34917: {'lr': 0.00011223869141245902, 'samples': 17878016, 'steps': 34917, 'loss/train': 0.9040366411209106} 02/25/2022 22:46:59 - INFO - codeparrot_training - Step 34918: {'lr': 0.00011222503763615483, 'samples': 17878528, 'steps': 34918, 'loss/train': 2.832584857940674} 02/25/2022 22:47:02 - INFO - codeparrot_training - Step 34919: {'lr': 0.00011221138445003473, 'samples': 17879040, 'steps': 34919, 'loss/train': 3.3160455226898193} 02/25/2022 22:47:06 - INFO - codeparrot_training - Step 34920: {'lr': 0.00011219773185415749, 'samples': 17879552, 'steps': 34920, 'loss/train': 2.2828028202056885} 02/25/2022 22:47:11 - INFO - codeparrot_training - Step 34921: {'lr': 0.00011218407984858126, 'samples': 17880064, 'steps': 34921, 'loss/train': 1.2578774690628052} 02/25/2022 22:47:15 - INFO - codeparrot_training - Step 34922: {'lr': 0.00011217042843336472, 'samples': 17880576, 'steps': 34922, 'loss/train': 1.8347923755645752} 02/25/2022 22:47:20 - INFO - codeparrot_training - Step 34923: {'lr': 0.00011215677760856622, 'samples': 17881088, 'steps': 34923, 'loss/train': 1.6547961235046387} 02/25/2022 22:47:24 - INFO - codeparrot_training - Step 34924: {'lr': 0.00011214312737424451, 'samples': 17881600, 'steps': 34924, 'loss/train': 1.731743335723877} 02/25/2022 22:47:29 - INFO - codeparrot_training - Step 34925: {'lr': 0.00011212947773045767, 'samples': 17882112, 'steps': 34925, 'loss/train': 0.5325666069984436} 02/25/2022 22:47:33 - INFO - codeparrot_training - Step 34926: {'lr': 0.00011211582867726447, 'samples': 17882624, 'steps': 34926, 'loss/train': 1.871329426765442} 02/25/2022 22:47:38 - INFO - codeparrot_training - Step 34927: {'lr': 0.00011210218021472318, 'samples': 17883136, 'steps': 34927, 'loss/train': 2.5138485431671143} 02/25/2022 22:47:42 - INFO - codeparrot_training - Step 34928: {'lr': 0.00011208853234289245, 'samples': 17883648, 'steps': 34928, 'loss/train': 1.7874820232391357} 02/25/2022 22:47:49 - INFO - codeparrot_training - Step 34929: {'lr': 0.00011207488506183067, 'samples': 17884160, 'steps': 34929, 'loss/train': 1.4505573511123657} 02/25/2022 22:47:52 - INFO - codeparrot_training - Step 34930: {'lr': 0.00011206123837159626, 'samples': 17884672, 'steps': 34930, 'loss/train': 1.51365327835083} 02/25/2022 22:47:58 - INFO - codeparrot_training - Step 34931: {'lr': 0.00011204759227224762, 'samples': 17885184, 'steps': 34931, 'loss/train': 2.6081128120422363} 02/25/2022 22:48:01 - INFO - codeparrot_training - Step 34932: {'lr': 0.0001120339467638434, 'samples': 17885696, 'steps': 34932, 'loss/train': 2.1144351959228516} 02/25/2022 22:48:07 - INFO - codeparrot_training - Step 34933: {'lr': 0.00011202030184644194, 'samples': 17886208, 'steps': 34933, 'loss/train': 2.384094476699829} 02/25/2022 22:48:10 - INFO - codeparrot_training - Step 34934: {'lr': 0.00011200665752010169, 'samples': 17886720, 'steps': 34934, 'loss/train': 2.0969252586364746} 02/25/2022 22:48:15 - INFO - codeparrot_training - Step 34935: {'lr': 0.000111993013784881, 'samples': 17887232, 'steps': 34935, 'loss/train': 0.46304118633270264} 02/25/2022 22:48:19 - INFO - codeparrot_training - Step 34936: {'lr': 0.00011197937064083852, 'samples': 17887744, 'steps': 34936, 'loss/train': 0.5004669427871704} 02/25/2022 22:48:24 - INFO - codeparrot_training - Step 34937: {'lr': 0.00011196572808803258, 'samples': 17888256, 'steps': 34937, 'loss/train': 2.3549468517303467} 02/25/2022 22:48:28 - INFO - codeparrot_training - Step 34938: {'lr': 0.00011195208612652163, 'samples': 17888768, 'steps': 34938, 'loss/train': 1.580697774887085} 02/25/2022 22:48:34 - INFO - codeparrot_training - Step 34939: {'lr': 0.00011193844475636403, 'samples': 17889280, 'steps': 34939, 'loss/train': 0.3215804398059845} 02/25/2022 22:48:38 - INFO - codeparrot_training - Step 34940: {'lr': 0.00011192480397761836, 'samples': 17889792, 'steps': 34940, 'loss/train': 1.3663384914398193} 02/25/2022 22:48:43 - INFO - codeparrot_training - Step 34941: {'lr': 0.00011191116379034291, 'samples': 17890304, 'steps': 34941, 'loss/train': 1.5452347993850708} 02/25/2022 22:48:47 - INFO - codeparrot_training - Step 34942: {'lr': 0.00011189752419459637, 'samples': 17890816, 'steps': 34942, 'loss/train': 1.679283857345581} 02/25/2022 22:48:52 - INFO - codeparrot_training - Step 34943: {'lr': 0.00011188388519043679, 'samples': 17891328, 'steps': 34943, 'loss/train': 1.6597061157226562} 02/25/2022 22:48:56 - INFO - codeparrot_training - Step 34944: {'lr': 0.0001118702467779229, 'samples': 17891840, 'steps': 34944, 'loss/train': 1.6378945112228394} 02/25/2022 22:49:01 - INFO - codeparrot_training - Step 34945: {'lr': 0.00011185660895711292, 'samples': 17892352, 'steps': 34945, 'loss/train': 1.4472392797470093} 02/25/2022 22:49:05 - INFO - codeparrot_training - Step 34946: {'lr': 0.00011184297172806554, 'samples': 17892864, 'steps': 34946, 'loss/train': 1.899036169052124} 02/25/2022 22:49:10 - INFO - codeparrot_training - Step 34947: {'lr': 0.00011182933509083881, 'samples': 17893376, 'steps': 34947, 'loss/train': 2.451996088027954} 02/25/2022 22:49:14 - INFO - codeparrot_training - Step 34948: {'lr': 0.00011181569904549144, 'samples': 17893888, 'steps': 34948, 'loss/train': 1.5890038013458252} 02/25/2022 22:49:19 - INFO - codeparrot_training - Step 34949: {'lr': 0.00011180206359208162, 'samples': 17894400, 'steps': 34949, 'loss/train': 3.1941628456115723} 02/25/2022 22:49:23 - INFO - codeparrot_training - Step 34950: {'lr': 0.00011178842873066808, 'samples': 17894912, 'steps': 34950, 'loss/train': 2.425978899002075} 02/25/2022 22:49:30 - INFO - codeparrot_training - Step 34951: {'lr': 0.00011177479446130884, 'samples': 17895424, 'steps': 34951, 'loss/train': 1.8784880638122559} 02/25/2022 22:49:33 - INFO - codeparrot_training - Step 34952: {'lr': 0.00011176116078406257, 'samples': 17895936, 'steps': 34952, 'loss/train': 1.4919337034225464} 02/25/2022 22:49:38 - INFO - codeparrot_training - Step 34953: {'lr': 0.00011174752769898757, 'samples': 17896448, 'steps': 34953, 'loss/train': 2.8363733291625977} 02/25/2022 22:49:42 - INFO - codeparrot_training - Step 34954: {'lr': 0.00011173389520614222, 'samples': 17896960, 'steps': 34954, 'loss/train': 1.3076075315475464} 02/25/2022 22:49:48 - INFO - codeparrot_training - Step 34955: {'lr': 0.00011172026330558501, 'samples': 17897472, 'steps': 34955, 'loss/train': 1.2258182764053345} 02/25/2022 22:49:51 - INFO - codeparrot_training - Step 34956: {'lr': 0.00011170663199737429, 'samples': 17897984, 'steps': 34956, 'loss/train': 1.2298986911773682} 02/25/2022 22:49:56 - INFO - codeparrot_training - Step 34957: {'lr': 0.00011169300128156848, 'samples': 17898496, 'steps': 34957, 'loss/train': 1.129446029663086} 02/25/2022 22:50:00 - INFO - codeparrot_training - Step 34958: {'lr': 0.00011167937115822583, 'samples': 17899008, 'steps': 34958, 'loss/train': 1.7038719654083252} 02/25/2022 22:50:07 - INFO - codeparrot_training - Step 34959: {'lr': 0.0001116657416274049, 'samples': 17899520, 'steps': 34959, 'loss/train': 2.1519057750701904} 02/25/2022 22:50:10 - INFO - codeparrot_training - Step 34960: {'lr': 0.000111652112689164, 'samples': 17900032, 'steps': 34960, 'loss/train': 0.7725964784622192} 02/25/2022 22:50:16 - INFO - codeparrot_training - Step 34961: {'lr': 0.00011163848434356156, 'samples': 17900544, 'steps': 34961, 'loss/train': 2.222297430038452} 02/25/2022 22:50:19 - INFO - codeparrot_training - Step 34962: {'lr': 0.0001116248565906558, 'samples': 17901056, 'steps': 34962, 'loss/train': 1.4858317375183105} 02/25/2022 22:50:25 - INFO - codeparrot_training - Step 34963: {'lr': 0.0001116112294305053, 'samples': 17901568, 'steps': 34963, 'loss/train': 1.138819932937622} 02/25/2022 22:50:28 - INFO - codeparrot_training - Step 34964: {'lr': 0.00011159760286316836, 'samples': 17902080, 'steps': 34964, 'loss/train': 1.6068347692489624} 02/25/2022 22:50:34 - INFO - codeparrot_training - Step 34965: {'lr': 0.00011158397688870333, 'samples': 17902592, 'steps': 34965, 'loss/train': 1.5345022678375244} 02/25/2022 22:50:38 - INFO - codeparrot_training - Step 34966: {'lr': 0.00011157035150716854, 'samples': 17903104, 'steps': 34966, 'loss/train': 1.7706342935562134} 02/25/2022 22:50:43 - INFO - codeparrot_training - Step 34967: {'lr': 0.00011155672671862247, 'samples': 17903616, 'steps': 34967, 'loss/train': 2.6073594093322754} 02/25/2022 22:50:47 - INFO - codeparrot_training - Step 34968: {'lr': 0.00011154310252312341, 'samples': 17904128, 'steps': 34968, 'loss/train': 2.2426705360412598} 02/25/2022 22:50:52 - INFO - codeparrot_training - Step 34969: {'lr': 0.00011152947892072976, 'samples': 17904640, 'steps': 34969, 'loss/train': 1.2264734506607056} 02/25/2022 22:50:56 - INFO - codeparrot_training - Step 34970: {'lr': 0.00011151585591149973, 'samples': 17905152, 'steps': 34970, 'loss/train': 8.609617233276367} 02/25/2022 22:51:03 - INFO - codeparrot_training - Step 34971: {'lr': 0.00011150223349549193, 'samples': 17905664, 'steps': 34971, 'loss/train': 2.39105486869812} 02/25/2022 22:51:06 - INFO - codeparrot_training - Step 34972: {'lr': 0.00011148861167276455, 'samples': 17906176, 'steps': 34972, 'loss/train': 1.4714531898498535} 02/25/2022 22:51:12 - INFO - codeparrot_training - Step 34973: {'lr': 0.000111474990443376, 'samples': 17906688, 'steps': 34973, 'loss/train': 0.8845553994178772} 02/25/2022 22:51:15 - INFO - codeparrot_training - Step 34974: {'lr': 0.0001114613698073845, 'samples': 17907200, 'steps': 34974, 'loss/train': 1.391049861907959} 02/25/2022 22:51:21 - INFO - codeparrot_training - Step 34975: {'lr': 0.00011144774976484859, 'samples': 17907712, 'steps': 34975, 'loss/train': 1.3737519979476929} 02/25/2022 22:51:24 - INFO - codeparrot_training - Step 34976: {'lr': 0.00011143413031582644, 'samples': 17908224, 'steps': 34976, 'loss/train': 1.7991303205490112} 02/25/2022 22:51:30 - INFO - codeparrot_training - Step 34977: {'lr': 0.00011142051146037666, 'samples': 17908736, 'steps': 34977, 'loss/train': 1.9011895656585693} 02/25/2022 22:51:33 - INFO - codeparrot_training - Step 34978: {'lr': 0.0001114068931985572, 'samples': 17909248, 'steps': 34978, 'loss/train': 2.7613868713378906} 02/25/2022 22:51:39 - INFO - codeparrot_training - Step 34979: {'lr': 0.00011139327553042672, 'samples': 17909760, 'steps': 34979, 'loss/train': 2.097351312637329} 02/25/2022 22:51:42 - INFO - codeparrot_training - Step 34980: {'lr': 0.00011137965845604333, 'samples': 17910272, 'steps': 34980, 'loss/train': 1.0668549537658691} 02/25/2022 22:51:48 - INFO - codeparrot_training - Step 34981: {'lr': 0.00011136604197546566, 'samples': 17910784, 'steps': 34981, 'loss/train': 1.2242546081542969} 02/25/2022 22:51:51 - INFO - codeparrot_training - Step 34982: {'lr': 0.00011135242608875166, 'samples': 17911296, 'steps': 34982, 'loss/train': 0.8575018644332886} 02/25/2022 22:51:57 - INFO - codeparrot_training - Step 34983: {'lr': 0.00011133881079595992, 'samples': 17911808, 'steps': 34983, 'loss/train': 2.0313644409179688} 02/25/2022 22:52:00 - INFO - codeparrot_training - Step 34984: {'lr': 0.0001113251960971486, 'samples': 17912320, 'steps': 34984, 'loss/train': 2.2020838260650635} 02/25/2022 22:52:06 - INFO - codeparrot_training - Step 34985: {'lr': 0.00011131158199237628, 'samples': 17912832, 'steps': 34985, 'loss/train': 0.9864962100982666} 02/25/2022 22:52:10 - INFO - codeparrot_training - Step 34986: {'lr': 0.00011129796848170095, 'samples': 17913344, 'steps': 34986, 'loss/train': 3.3161728382110596} 02/25/2022 22:52:17 - INFO - codeparrot_training - Step 34987: {'lr': 0.00011128435556518115, 'samples': 17913856, 'steps': 34987, 'loss/train': 1.9074406623840332} 02/25/2022 22:52:20 - INFO - codeparrot_training - Step 34988: {'lr': 0.00011127074324287504, 'samples': 17914368, 'steps': 34988, 'loss/train': 0.2186937928199768} 02/25/2022 22:52:26 - INFO - codeparrot_training - Step 34989: {'lr': 0.0001112571315148411, 'samples': 17914880, 'steps': 34989, 'loss/train': 2.095627784729004} 02/25/2022 22:52:29 - INFO - codeparrot_training - Step 34990: {'lr': 0.00011124352038113755, 'samples': 17915392, 'steps': 34990, 'loss/train': 1.5660802125930786} 02/25/2022 22:52:35 - INFO - codeparrot_training - Step 34991: {'lr': 0.00011122990984182268, 'samples': 17915904, 'steps': 34991, 'loss/train': 1.0819975137710571} 02/25/2022 22:52:38 - INFO - codeparrot_training - Step 34992: {'lr': 0.00011121629989695475, 'samples': 17916416, 'steps': 34992, 'loss/train': 1.5947767496109009} 02/25/2022 22:52:44 - INFO - codeparrot_training - Step 34993: {'lr': 0.00011120269054659221, 'samples': 17916928, 'steps': 34993, 'loss/train': 2.1122138500213623} 02/25/2022 22:52:47 - INFO - codeparrot_training - Step 34994: {'lr': 0.00011118908179079326, 'samples': 17917440, 'steps': 34994, 'loss/train': 1.865601658821106} 02/25/2022 22:52:54 - INFO - codeparrot_training - Step 34995: {'lr': 0.00011117547362961622, 'samples': 17917952, 'steps': 34995, 'loss/train': 1.8611263036727905} 02/25/2022 22:52:58 - INFO - codeparrot_training - Step 34996: {'lr': 0.0001111618660631193, 'samples': 17918464, 'steps': 34996, 'loss/train': 1.7628575563430786} 02/25/2022 22:53:03 - INFO - codeparrot_training - Step 34997: {'lr': 0.00011114825909136089, 'samples': 17918976, 'steps': 34997, 'loss/train': 1.7161481380462646} 02/25/2022 22:53:07 - INFO - codeparrot_training - Step 34998: {'lr': 0.00011113465271439929, 'samples': 17919488, 'steps': 34998, 'loss/train': 1.1114963293075562} 02/25/2022 22:53:12 - INFO - codeparrot_training - Step 34999: {'lr': 0.00011112104693229272, 'samples': 17920000, 'steps': 34999, 'loss/train': 2.1389760971069336} 02/25/2022 22:53:12 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/25/2022 22:53:30 - WARNING - huggingface_hub.repository - Several commits (35) will be pushed upstream. 02/25/2022 22:53:30 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/25/2022 22:54:05 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 6cccefd..aab1012 floral-grass-11 -> floral-grass-11 02/25/2022 22:54:10 - INFO - codeparrot_training - Step 35000: {'lr': 0.00011110744174509952, 'samples': 17920512, 'steps': 35000, 'loss/train': 1.108235478401184} 02/25/2022 22:54:16 - INFO - codeparrot_training - Step 35001: {'lr': 0.00011109383715287782, 'samples': 17921024, 'steps': 35001, 'loss/train': 1.5042431354522705} 02/25/2022 22:54:19 - INFO - codeparrot_training - Step 35002: {'lr': 0.00011108023315568611, 'samples': 17921536, 'steps': 35002, 'loss/train': 1.7382547855377197} 02/25/2022 22:54:25 - INFO - codeparrot_training - Step 35003: {'lr': 0.00011106662975358256, 'samples': 17922048, 'steps': 35003, 'loss/train': 1.6398383378982544} 02/25/2022 22:54:29 - INFO - codeparrot_training - Step 35004: {'lr': 0.00011105302694662544, 'samples': 17922560, 'steps': 35004, 'loss/train': 2.1558969020843506} 02/25/2022 22:54:34 - INFO - codeparrot_training - Step 35005: {'lr': 0.00011103942473487297, 'samples': 17923072, 'steps': 35005, 'loss/train': 2.3898162841796875} 02/25/2022 22:54:38 - INFO - codeparrot_training - Step 35006: {'lr': 0.00011102582311838358, 'samples': 17923584, 'steps': 35006, 'loss/train': 1.802404522895813} 02/25/2022 22:54:44 - INFO - codeparrot_training - Step 35007: {'lr': 0.0001110122220972154, 'samples': 17924096, 'steps': 35007, 'loss/train': 0.7885240912437439} 02/25/2022 22:54:48 - INFO - codeparrot_training - Step 35008: {'lr': 0.00011099862167142674, 'samples': 17924608, 'steps': 35008, 'loss/train': 1.8774694204330444} 02/25/2022 22:54:53 - INFO - codeparrot_training - Step 35009: {'lr': 0.00011098502184107575, 'samples': 17925120, 'steps': 35009, 'loss/train': 1.8468815088272095} 02/25/2022 22:54:57 - INFO - codeparrot_training - Step 35010: {'lr': 0.0001109714226062209, 'samples': 17925632, 'steps': 35010, 'loss/train': 1.0692946910858154} 02/25/2022 22:55:02 - INFO - codeparrot_training - Step 35011: {'lr': 0.0001109578239669203, 'samples': 17926144, 'steps': 35011, 'loss/train': 1.1396328210830688} 02/25/2022 22:55:06 - INFO - codeparrot_training - Step 35012: {'lr': 0.00011094422592323224, 'samples': 17926656, 'steps': 35012, 'loss/train': 1.6678696870803833} 02/25/2022 22:55:11 - INFO - codeparrot_training - Step 35013: {'lr': 0.0001109306284752149, 'samples': 17927168, 'steps': 35013, 'loss/train': 1.1875629425048828} 02/25/2022 22:55:15 - INFO - codeparrot_training - Step 35014: {'lr': 0.00011091703162292666, 'samples': 17927680, 'steps': 35014, 'loss/train': 2.338594436645508} 02/25/2022 22:55:20 - INFO - codeparrot_training - Step 35015: {'lr': 0.00011090343536642559, 'samples': 17928192, 'steps': 35015, 'loss/train': 1.9597567319869995} 02/25/2022 22:55:24 - INFO - codeparrot_training - Step 35016: {'lr': 0.00011088983970577021, 'samples': 17928704, 'steps': 35016, 'loss/train': 0.8792717456817627} 02/25/2022 22:55:29 - INFO - codeparrot_training - Step 35017: {'lr': 0.00011087624464101842, 'samples': 17929216, 'steps': 35017, 'loss/train': 1.7256443500518799} 02/25/2022 22:55:33 - INFO - codeparrot_training - Step 35018: {'lr': 0.00011086265017222869, 'samples': 17929728, 'steps': 35018, 'loss/train': 1.8258897066116333} 02/25/2022 22:55:38 - INFO - codeparrot_training - Step 35019: {'lr': 0.00011084905629945912, 'samples': 17930240, 'steps': 35019, 'loss/train': 1.9284745454788208} 02/25/2022 22:55:42 - INFO - codeparrot_training - Step 35020: {'lr': 0.00011083546302276818, 'samples': 17930752, 'steps': 35020, 'loss/train': 0.4454461932182312} 02/25/2022 22:55:49 - INFO - codeparrot_training - Step 35021: {'lr': 0.00011082187034221375, 'samples': 17931264, 'steps': 35021, 'loss/train': 0.7002458572387695} 02/25/2022 22:55:53 - INFO - codeparrot_training - Step 35022: {'lr': 0.0001108082782578543, 'samples': 17931776, 'steps': 35022, 'loss/train': 1.445369839668274} 02/25/2022 22:55:58 - INFO - codeparrot_training - Step 35023: {'lr': 0.00011079468676974793, 'samples': 17932288, 'steps': 35023, 'loss/train': 2.4348347187042236} 02/25/2022 22:56:02 - INFO - codeparrot_training - Step 35024: {'lr': 0.0001107810958779531, 'samples': 17932800, 'steps': 35024, 'loss/train': 2.172879934310913} 02/25/2022 22:56:07 - INFO - codeparrot_training - Step 35025: {'lr': 0.00011076750558252763, 'samples': 17933312, 'steps': 35025, 'loss/train': 1.809644103050232} 02/25/2022 22:56:11 - INFO - codeparrot_training - Step 35026: {'lr': 0.00011075391588353007, 'samples': 17933824, 'steps': 35026, 'loss/train': 1.2199939489364624} 02/25/2022 22:56:16 - INFO - codeparrot_training - Step 35027: {'lr': 0.00011074032678101842, 'samples': 17934336, 'steps': 35027, 'loss/train': 2.377208948135376} 02/25/2022 22:56:20 - INFO - codeparrot_training - Step 35028: {'lr': 0.00011072673827505109, 'samples': 17934848, 'steps': 35028, 'loss/train': 2.5669732093811035} 02/25/2022 22:56:25 - INFO - codeparrot_training - Step 35029: {'lr': 0.00011071315036568619, 'samples': 17935360, 'steps': 35029, 'loss/train': 1.5660430192947388} 02/25/2022 22:56:29 - INFO - codeparrot_training - Step 35030: {'lr': 0.00011069956305298193, 'samples': 17935872, 'steps': 35030, 'loss/train': 1.7470132112503052} 02/25/2022 22:56:34 - INFO - codeparrot_training - Step 35031: {'lr': 0.0001106859763369964, 'samples': 17936384, 'steps': 35031, 'loss/train': 8.209218978881836} 02/25/2022 22:56:38 - INFO - codeparrot_training - Step 35032: {'lr': 0.00011067239021778803, 'samples': 17936896, 'steps': 35032, 'loss/train': 1.5783425569534302} 02/25/2022 22:56:45 - INFO - codeparrot_training - Step 35033: {'lr': 0.00011065880469541487, 'samples': 17937408, 'steps': 35033, 'loss/train': 2.4365322589874268} 02/25/2022 22:56:48 - INFO - codeparrot_training - Step 35034: {'lr': 0.00011064521976993517, 'samples': 17937920, 'steps': 35034, 'loss/train': 0.506431519985199} 02/25/2022 22:56:54 - INFO - codeparrot_training - Step 35035: {'lr': 0.00011063163544140698, 'samples': 17938432, 'steps': 35035, 'loss/train': 1.369081735610962} 02/25/2022 22:56:57 - INFO - codeparrot_training - Step 35036: {'lr': 0.0001106180517098887, 'samples': 17938944, 'steps': 35036, 'loss/train': 0.97723788022995} 02/25/2022 22:57:03 - INFO - codeparrot_training - Step 35037: {'lr': 0.00011060446857543846, 'samples': 17939456, 'steps': 35037, 'loss/train': 2.5222973823547363} 02/25/2022 22:57:06 - INFO - codeparrot_training - Step 35038: {'lr': 0.00011059088603811438, 'samples': 17939968, 'steps': 35038, 'loss/train': 2.940106153488159} 02/25/2022 22:57:12 - INFO - codeparrot_training - Step 35039: {'lr': 0.0001105773040979746, 'samples': 17940480, 'steps': 35039, 'loss/train': 0.5782507061958313} 02/25/2022 22:57:15 - INFO - codeparrot_training - Step 35040: {'lr': 0.00011056372275507747, 'samples': 17940992, 'steps': 35040, 'loss/train': 1.2188520431518555} 02/25/2022 22:57:21 - INFO - codeparrot_training - Step 35041: {'lr': 0.00011055014200948107, 'samples': 17941504, 'steps': 35041, 'loss/train': 1.243159294128418} 02/25/2022 22:57:24 - INFO - codeparrot_training - Step 35042: {'lr': 0.00011053656186124358, 'samples': 17942016, 'steps': 35042, 'loss/train': 2.232343912124634} 02/25/2022 22:57:31 - INFO - codeparrot_training - Step 35043: {'lr': 0.0001105229823104231, 'samples': 17942528, 'steps': 35043, 'loss/train': 1.370757818222046} 02/25/2022 22:57:35 - INFO - codeparrot_training - Step 35044: {'lr': 0.00011050940335707796, 'samples': 17943040, 'steps': 35044, 'loss/train': 0.0707676112651825} 02/25/2022 22:57:40 - INFO - codeparrot_training - Step 35045: {'lr': 0.00011049582500126623, 'samples': 17943552, 'steps': 35045, 'loss/train': 1.1718299388885498} 02/25/2022 22:57:43 - INFO - codeparrot_training - Step 35046: {'lr': 0.0001104822472430461, 'samples': 17944064, 'steps': 35046, 'loss/train': 1.9143428802490234} 02/25/2022 22:57:49 - INFO - codeparrot_training - Step 35047: {'lr': 0.00011046867008247563, 'samples': 17944576, 'steps': 35047, 'loss/train': 1.6932884454727173} 02/25/2022 22:57:52 - INFO - codeparrot_training - Step 35048: {'lr': 0.00011045509351961314, 'samples': 17945088, 'steps': 35048, 'loss/train': 2.5862393379211426} 02/25/2022 22:57:58 - INFO - codeparrot_training - Step 35049: {'lr': 0.00011044151755451676, 'samples': 17945600, 'steps': 35049, 'loss/train': 1.5274930000305176} 02/25/2022 22:58:02 - INFO - codeparrot_training - Step 35050: {'lr': 0.00011042794218724447, 'samples': 17946112, 'steps': 35050, 'loss/train': 0.43550872802734375} 02/25/2022 22:58:08 - INFO - codeparrot_training - Step 35051: {'lr': 0.00011041436741785477, 'samples': 17946624, 'steps': 35051, 'loss/train': 1.5613032579421997} 02/25/2022 22:58:11 - INFO - codeparrot_training - Step 35052: {'lr': 0.0001104007932464054, 'samples': 17947136, 'steps': 35052, 'loss/train': 1.0032163858413696} 02/25/2022 22:58:15 - INFO - codeparrot_training - Step 35053: {'lr': 0.00011038721967295479, 'samples': 17947648, 'steps': 35053, 'loss/train': 1.9781179428100586} 02/25/2022 22:58:20 - INFO - codeparrot_training - Step 35054: {'lr': 0.00011037364669756092, 'samples': 17948160, 'steps': 35054, 'loss/train': 1.5532522201538086} 02/25/2022 22:58:24 - INFO - codeparrot_training - Step 35055: {'lr': 0.0001103600743202822, 'samples': 17948672, 'steps': 35055, 'loss/train': 1.190137505531311} 02/25/2022 22:58:29 - INFO - codeparrot_training - Step 35056: {'lr': 0.00011034650254117637, 'samples': 17949184, 'steps': 35056, 'loss/train': 1.8900024890899658} 02/25/2022 22:58:33 - INFO - codeparrot_training - Step 35057: {'lr': 0.0001103329313603019, 'samples': 17949696, 'steps': 35057, 'loss/train': 1.394480586051941} 02/25/2022 22:58:39 - INFO - codeparrot_training - Step 35058: {'lr': 0.00011031936077771667, 'samples': 17950208, 'steps': 35058, 'loss/train': 1.4922599792480469} 02/25/2022 22:58:42 - INFO - codeparrot_training - Step 35059: {'lr': 0.00011030579079347916, 'samples': 17950720, 'steps': 35059, 'loss/train': 2.0848727226257324} 02/25/2022 22:58:49 - INFO - codeparrot_training - Step 35060: {'lr': 0.0001102922214076471, 'samples': 17951232, 'steps': 35060, 'loss/train': 2.190354824066162} 02/25/2022 22:58:52 - INFO - codeparrot_training - Step 35061: {'lr': 0.0001102786526202789, 'samples': 17951744, 'steps': 35061, 'loss/train': 1.7906067371368408} 02/25/2022 22:58:58 - INFO - codeparrot_training - Step 35062: {'lr': 0.00011026508443143248, 'samples': 17952256, 'steps': 35062, 'loss/train': 2.0709264278411865} 02/25/2022 22:59:01 - INFO - codeparrot_training - Step 35063: {'lr': 0.00011025151684116616, 'samples': 17952768, 'steps': 35063, 'loss/train': 2.20731782913208} 02/25/2022 22:59:07 - INFO - codeparrot_training - Step 35064: {'lr': 0.00011023794984953797, 'samples': 17953280, 'steps': 35064, 'loss/train': 1.8016061782836914} 02/25/2022 22:59:11 - INFO - codeparrot_training - Step 35065: {'lr': 0.00011022438345660602, 'samples': 17953792, 'steps': 35065, 'loss/train': 1.1277140378952026} 02/25/2022 22:59:16 - INFO - codeparrot_training - Step 35066: {'lr': 0.00011021081766242836, 'samples': 17954304, 'steps': 35066, 'loss/train': 1.03036367893219} 02/25/2022 22:59:19 - INFO - codeparrot_training - Step 35067: {'lr': 0.00011019725246706322, 'samples': 17954816, 'steps': 35067, 'loss/train': 1.2721691131591797} 02/25/2022 22:59:26 - INFO - codeparrot_training - Step 35068: {'lr': 0.00011018368787056868, 'samples': 17955328, 'steps': 35068, 'loss/train': 2.0335490703582764} 02/25/2022 22:59:29 - INFO - codeparrot_training - Step 35069: {'lr': 0.00011017012387300282, 'samples': 17955840, 'steps': 35069, 'loss/train': 2.016598701477051} 02/25/2022 22:59:35 - INFO - codeparrot_training - Step 35070: {'lr': 0.00011015656047442366, 'samples': 17956352, 'steps': 35070, 'loss/train': 1.2222919464111328} 02/25/2022 22:59:39 - INFO - codeparrot_training - Step 35071: {'lr': 0.00011014299767488947, 'samples': 17956864, 'steps': 35071, 'loss/train': 1.6149224042892456} 02/25/2022 22:59:44 - INFO - codeparrot_training - Step 35072: {'lr': 0.00011012943547445828, 'samples': 17957376, 'steps': 35072, 'loss/train': 2.5101821422576904} 02/25/2022 22:59:48 - INFO - codeparrot_training - Step 35073: {'lr': 0.00011011587387318819, 'samples': 17957888, 'steps': 35073, 'loss/train': 2.234025239944458} 02/25/2022 22:59:53 - INFO - codeparrot_training - Step 35074: {'lr': 0.00011010231287113717, 'samples': 17958400, 'steps': 35074, 'loss/train': 1.6715167760849} 02/25/2022 22:59:57 - INFO - codeparrot_training - Step 35075: {'lr': 0.0001100887524683635, 'samples': 17958912, 'steps': 35075, 'loss/train': 1.2137010097503662} 02/25/2022 23:00:02 - INFO - codeparrot_training - Step 35076: {'lr': 0.0001100751926649251, 'samples': 17959424, 'steps': 35076, 'loss/train': 2.151000738143921} 02/25/2022 23:00:06 - INFO - codeparrot_training - Step 35077: {'lr': 0.00011006163346088035, 'samples': 17959936, 'steps': 35077, 'loss/train': 0.8322744369506836} 02/25/2022 23:00:11 - INFO - codeparrot_training - Step 35078: {'lr': 0.00011004807485628693, 'samples': 17960448, 'steps': 35078, 'loss/train': 1.9726406335830688} 02/25/2022 23:00:15 - INFO - codeparrot_training - Step 35079: {'lr': 0.00011003451685120319, 'samples': 17960960, 'steps': 35079, 'loss/train': 0.9069707989692688} 02/25/2022 23:00:22 - INFO - codeparrot_training - Step 35080: {'lr': 0.00011002095944568708, 'samples': 17961472, 'steps': 35080, 'loss/train': 0.8275983929634094} 02/25/2022 23:00:25 - INFO - codeparrot_training - Step 35081: {'lr': 0.00011000740263979689, 'samples': 17961984, 'steps': 35081, 'loss/train': 1.3482609987258911} 02/25/2022 23:00:30 - INFO - codeparrot_training - Step 35082: {'lr': 0.00010999384643359033, 'samples': 17962496, 'steps': 35082, 'loss/train': 1.01606285572052} 02/25/2022 23:00:34 - INFO - codeparrot_training - Step 35083: {'lr': 0.0001099802908271258, 'samples': 17963008, 'steps': 35083, 'loss/train': 1.1263706684112549} 02/25/2022 23:00:39 - INFO - codeparrot_training - Step 35084: {'lr': 0.00010996673582046124, 'samples': 17963520, 'steps': 35084, 'loss/train': 1.6917088031768799} 02/25/2022 23:00:43 - INFO - codeparrot_training - Step 35085: {'lr': 0.00010995318141365463, 'samples': 17964032, 'steps': 35085, 'loss/train': 2.3367300033569336} 02/25/2022 23:00:49 - INFO - codeparrot_training - Step 35086: {'lr': 0.00010993962760676421, 'samples': 17964544, 'steps': 35086, 'loss/train': 1.702036738395691} 02/25/2022 23:00:52 - INFO - codeparrot_training - Step 35087: {'lr': 0.00010992607439984791, 'samples': 17965056, 'steps': 35087, 'loss/train': 1.744388222694397} 02/25/2022 23:00:57 - INFO - codeparrot_training - Step 35088: {'lr': 0.00010991252179296387, 'samples': 17965568, 'steps': 35088, 'loss/train': 1.6348741054534912} 02/25/2022 23:01:01 - INFO - codeparrot_training - Step 35089: {'lr': 0.00010989896978617004, 'samples': 17966080, 'steps': 35089, 'loss/train': 1.9306329488754272} 02/25/2022 23:01:08 - INFO - codeparrot_training - Step 35090: {'lr': 0.00010988541837952457, 'samples': 17966592, 'steps': 35090, 'loss/train': 1.7279155254364014} 02/25/2022 23:01:11 - INFO - codeparrot_training - Step 35091: {'lr': 0.00010987186757308551, 'samples': 17967104, 'steps': 35091, 'loss/train': 1.4924261569976807} 02/25/2022 23:01:17 - INFO - codeparrot_training - Step 35092: {'lr': 0.00010985831736691085, 'samples': 17967616, 'steps': 35092, 'loss/train': 1.0858821868896484} 02/25/2022 23:01:20 - INFO - codeparrot_training - Step 35093: {'lr': 0.00010984476776105859, 'samples': 17968128, 'steps': 35093, 'loss/train': 2.742690324783325} 02/25/2022 23:01:26 - INFO - codeparrot_training - Step 35094: {'lr': 0.00010983121875558694, 'samples': 17968640, 'steps': 35094, 'loss/train': 1.5161410570144653} 02/25/2022 23:01:29 - INFO - codeparrot_training - Step 35095: {'lr': 0.00010981767035055382, 'samples': 17969152, 'steps': 35095, 'loss/train': 1.2386881113052368} 02/25/2022 23:01:35 - INFO - codeparrot_training - Step 35096: {'lr': 0.00010980412254601729, 'samples': 17969664, 'steps': 35096, 'loss/train': 0.5685620903968811} 02/25/2022 23:01:38 - INFO - codeparrot_training - Step 35097: {'lr': 0.00010979057534203529, 'samples': 17970176, 'steps': 35097, 'loss/train': 1.8363076448440552} 02/25/2022 23:01:44 - INFO - codeparrot_training - Step 35098: {'lr': 0.00010977702873866602, 'samples': 17970688, 'steps': 35098, 'loss/train': 2.5595669746398926} 02/25/2022 23:01:47 - INFO - codeparrot_training - Step 35099: {'lr': 0.00010976348273596745, 'samples': 17971200, 'steps': 35099, 'loss/train': 1.9575541019439697} 02/25/2022 23:01:53 - INFO - codeparrot_training - Step 35100: {'lr': 0.0001097499373339976, 'samples': 17971712, 'steps': 35100, 'loss/train': 1.759906530380249} 02/25/2022 23:01:56 - INFO - codeparrot_training - Step 35101: {'lr': 0.00010973639253281437, 'samples': 17972224, 'steps': 35101, 'loss/train': 0.87238609790802} 02/25/2022 23:02:02 - INFO - codeparrot_training - Step 35102: {'lr': 0.00010972284833247601, 'samples': 17972736, 'steps': 35102, 'loss/train': 1.4301810264587402} 02/25/2022 23:02:05 - INFO - codeparrot_training - Step 35103: {'lr': 0.00010970930473304041, 'samples': 17973248, 'steps': 35103, 'loss/train': 1.3741896152496338} 02/25/2022 23:02:11 - INFO - codeparrot_training - Step 35104: {'lr': 0.00010969576173456561, 'samples': 17973760, 'steps': 35104, 'loss/train': 2.065410852432251} 02/25/2022 23:02:14 - INFO - codeparrot_training - Step 35105: {'lr': 0.00010968221933710951, 'samples': 17974272, 'steps': 35105, 'loss/train': 2.2047057151794434} 02/25/2022 23:02:21 - INFO - codeparrot_training - Step 35106: {'lr': 0.00010966867754073032, 'samples': 17974784, 'steps': 35106, 'loss/train': 1.1051528453826904} 02/25/2022 23:02:24 - INFO - codeparrot_training - Step 35107: {'lr': 0.00010965513634548596, 'samples': 17975296, 'steps': 35107, 'loss/train': 1.1823266744613647} 02/25/2022 23:02:30 - INFO - codeparrot_training - Step 35108: {'lr': 0.00010964159575143445, 'samples': 17975808, 'steps': 35108, 'loss/train': 1.8810521364212036} 02/25/2022 23:02:33 - INFO - codeparrot_training - Step 35109: {'lr': 0.00010962805575863366, 'samples': 17976320, 'steps': 35109, 'loss/train': 1.592991828918457} 02/25/2022 23:02:39 - INFO - codeparrot_training - Step 35110: {'lr': 0.00010961451636714182, 'samples': 17976832, 'steps': 35110, 'loss/train': 2.4122071266174316} 02/25/2022 23:02:42 - INFO - codeparrot_training - Step 35111: {'lr': 0.0001096009775770167, 'samples': 17977344, 'steps': 35111, 'loss/train': 1.9894005060195923} 02/25/2022 23:02:47 - INFO - codeparrot_training - Step 35112: {'lr': 0.00010958743938831659, 'samples': 17977856, 'steps': 35112, 'loss/train': 1.088016152381897} 02/25/2022 23:02:51 - INFO - codeparrot_training - Step 35113: {'lr': 0.00010957390180109913, 'samples': 17978368, 'steps': 35113, 'loss/train': 0.5615712404251099} 02/25/2022 23:02:56 - INFO - codeparrot_training - Step 35114: {'lr': 0.00010956036481542258, 'samples': 17978880, 'steps': 35114, 'loss/train': 2.6407108306884766} 02/25/2022 23:03:00 - INFO - codeparrot_training - Step 35115: {'lr': 0.00010954682843134472, 'samples': 17979392, 'steps': 35115, 'loss/train': 0.21209581196308136} 02/25/2022 23:03:06 - INFO - codeparrot_training - Step 35116: {'lr': 0.00010953329264892386, 'samples': 17979904, 'steps': 35116, 'loss/train': 1.5682331323623657} 02/25/2022 23:03:10 - INFO - codeparrot_training - Step 35117: {'lr': 0.00010951975746821758, 'samples': 17980416, 'steps': 35117, 'loss/train': 1.9463366270065308} 02/25/2022 23:03:16 - INFO - codeparrot_training - Step 35118: {'lr': 0.00010950622288928413, 'samples': 17980928, 'steps': 35118, 'loss/train': 1.2538259029388428} 02/25/2022 23:03:19 - INFO - codeparrot_training - Step 35119: {'lr': 0.00010949268891218133, 'samples': 17981440, 'steps': 35119, 'loss/train': 2.1794042587280273} 02/25/2022 23:03:25 - INFO - codeparrot_training - Step 35120: {'lr': 0.0001094791555369674, 'samples': 17981952, 'steps': 35120, 'loss/train': 1.0278507471084595} 02/25/2022 23:03:28 - INFO - codeparrot_training - Step 35121: {'lr': 0.00010946562276370001, 'samples': 17982464, 'steps': 35121, 'loss/train': 1.0653961896896362} 02/25/2022 23:03:33 - INFO - codeparrot_training - Step 35122: {'lr': 0.00010945209059243733, 'samples': 17982976, 'steps': 35122, 'loss/train': 1.7465919256210327} 02/25/2022 23:03:37 - INFO - codeparrot_training - Step 35123: {'lr': 0.00010943855902323716, 'samples': 17983488, 'steps': 35123, 'loss/train': 1.8433876037597656} 02/25/2022 23:03:42 - INFO - codeparrot_training - Step 35124: {'lr': 0.00010942502805615767, 'samples': 17984000, 'steps': 35124, 'loss/train': 1.6061464548110962} 02/25/2022 23:03:46 - INFO - codeparrot_training - Step 35125: {'lr': 0.00010941149769125675, 'samples': 17984512, 'steps': 35125, 'loss/train': 1.7023378610610962} 02/25/2022 23:03:53 - INFO - codeparrot_training - Step 35126: {'lr': 0.00010939796792859231, 'samples': 17985024, 'steps': 35126, 'loss/train': 2.8829214572906494} 02/25/2022 23:03:56 - INFO - codeparrot_training - Step 35127: {'lr': 0.00010938443876822222, 'samples': 17985536, 'steps': 35127, 'loss/train': 1.5216530561447144} 02/25/2022 23:04:02 - INFO - codeparrot_training - Step 35128: {'lr': 0.00010937091021020464, 'samples': 17986048, 'steps': 35128, 'loss/train': 1.276953101158142} 02/25/2022 23:04:05 - INFO - codeparrot_training - Step 35129: {'lr': 0.00010935738225459743, 'samples': 17986560, 'steps': 35129, 'loss/train': 2.7647812366485596} 02/25/2022 23:04:10 - INFO - codeparrot_training - Step 35130: {'lr': 0.00010934385490145857, 'samples': 17987072, 'steps': 35130, 'loss/train': 1.482032060623169} 02/25/2022 23:04:14 - INFO - codeparrot_training - Step 35131: {'lr': 0.00010933032815084593, 'samples': 17987584, 'steps': 35131, 'loss/train': 1.6871857643127441} 02/25/2022 23:04:19 - INFO - codeparrot_training - Step 35132: {'lr': 0.00010931680200281741, 'samples': 17988096, 'steps': 35132, 'loss/train': 0.3812721073627472} 02/25/2022 23:04:23 - INFO - codeparrot_training - Step 35133: {'lr': 0.00010930327645743116, 'samples': 17988608, 'steps': 35133, 'loss/train': 1.7362761497497559} 02/25/2022 23:04:28 - INFO - codeparrot_training - Step 35134: {'lr': 0.00010928975151474496, 'samples': 17989120, 'steps': 35134, 'loss/train': 1.382032871246338} 02/25/2022 23:04:32 - INFO - codeparrot_training - Step 35135: {'lr': 0.0001092762271748168, 'samples': 17989632, 'steps': 35135, 'loss/train': 2.2625138759613037} 02/25/2022 23:04:39 - INFO - codeparrot_training - Step 35136: {'lr': 0.0001092627034377045, 'samples': 17990144, 'steps': 35136, 'loss/train': 2.0682082176208496} 02/25/2022 23:04:42 - INFO - codeparrot_training - Step 35137: {'lr': 0.0001092491803034662, 'samples': 17990656, 'steps': 35137, 'loss/train': 1.8072658777236938} 02/25/2022 23:04:47 - INFO - codeparrot_training - Step 35138: {'lr': 0.00010923565777215972, 'samples': 17991168, 'steps': 35138, 'loss/train': 0.4985397756099701} 02/25/2022 23:04:51 - INFO - codeparrot_training - Step 35139: {'lr': 0.00010922213584384296, 'samples': 17991680, 'steps': 35139, 'loss/train': 1.545931339263916} 02/25/2022 23:04:56 - INFO - codeparrot_training - Step 35140: {'lr': 0.00010920861451857378, 'samples': 17992192, 'steps': 35140, 'loss/train': 1.739908218383789} 02/25/2022 23:05:00 - INFO - codeparrot_training - Step 35141: {'lr': 0.0001091950937964103, 'samples': 17992704, 'steps': 35141, 'loss/train': 1.2667440176010132} 02/25/2022 23:05:05 - INFO - codeparrot_training - Step 35142: {'lr': 0.0001091815736774103, 'samples': 17993216, 'steps': 35142, 'loss/train': 0.4341658353805542} 02/25/2022 23:05:09 - INFO - codeparrot_training - Step 35143: {'lr': 0.00010916805416163176, 'samples': 17993728, 'steps': 35143, 'loss/train': 1.6816037893295288} 02/25/2022 23:05:14 - INFO - codeparrot_training - Step 35144: {'lr': 0.00010915453524913243, 'samples': 17994240, 'steps': 35144, 'loss/train': 1.0795174837112427} 02/25/2022 23:05:18 - INFO - codeparrot_training - Step 35145: {'lr': 0.00010914101693997044, 'samples': 17994752, 'steps': 35145, 'loss/train': 1.9348419904708862} 02/25/2022 23:05:23 - INFO - codeparrot_training - Step 35146: {'lr': 0.00010912749923420354, 'samples': 17995264, 'steps': 35146, 'loss/train': 1.5501378774642944} 02/25/2022 23:05:29 - INFO - codeparrot_training - Step 35147: {'lr': 0.00010911398213188986, 'samples': 17995776, 'steps': 35147, 'loss/train': 1.7865114212036133} 02/25/2022 23:05:32 - INFO - codeparrot_training - Step 35148: {'lr': 0.00010910046563308699, 'samples': 17996288, 'steps': 35148, 'loss/train': 1.5714385509490967} 02/25/2022 23:05:38 - INFO - codeparrot_training - Step 35149: {'lr': 0.00010908694973785305, 'samples': 17996800, 'steps': 35149, 'loss/train': 1.765008568763733} 02/25/2022 23:05:42 - INFO - codeparrot_training - Step 35150: {'lr': 0.00010907343444624579, 'samples': 17997312, 'steps': 35150, 'loss/train': 1.4733761548995972} 02/25/2022 23:05:48 - INFO - codeparrot_training - Step 35151: {'lr': 0.00010905991975832335, 'samples': 17997824, 'steps': 35151, 'loss/train': 2.5241777896881104} 02/25/2022 23:05:52 - INFO - codeparrot_training - Step 35152: {'lr': 0.00010904640567414331, 'samples': 17998336, 'steps': 35152, 'loss/train': 1.0644471645355225} 02/25/2022 23:05:58 - INFO - codeparrot_training - Step 35153: {'lr': 0.00010903289219376381, 'samples': 17998848, 'steps': 35153, 'loss/train': 2.402423620223999} 02/25/2022 23:06:01 - INFO - codeparrot_training - Step 35154: {'lr': 0.00010901937931724251, 'samples': 17999360, 'steps': 35154, 'loss/train': 2.248983144760132} 02/25/2022 23:06:05 - INFO - codeparrot_training - Step 35155: {'lr': 0.0001090058670446376, 'samples': 17999872, 'steps': 35155, 'loss/train': 2.2496323585510254} 02/25/2022 23:06:10 - INFO - codeparrot_training - Step 35156: {'lr': 0.00010899235537600663, 'samples': 18000384, 'steps': 35156, 'loss/train': 1.2339597940444946} 02/25/2022 23:06:14 - INFO - codeparrot_training - Step 35157: {'lr': 0.00010897884431140772, 'samples': 18000896, 'steps': 35157, 'loss/train': 2.013819694519043} 02/25/2022 23:06:19 - INFO - codeparrot_training - Step 35158: {'lr': 0.00010896533385089857, 'samples': 18001408, 'steps': 35158, 'loss/train': 1.181535243988037} 02/25/2022 23:06:23 - INFO - codeparrot_training - Step 35159: {'lr': 0.00010895182399453724, 'samples': 18001920, 'steps': 35159, 'loss/train': 1.4936811923980713} 02/25/2022 23:06:28 - INFO - codeparrot_training - Step 35160: {'lr': 0.0001089383147423815, 'samples': 18002432, 'steps': 35160, 'loss/train': 1.084436058998108} 02/25/2022 23:06:32 - INFO - codeparrot_training - Step 35161: {'lr': 0.00010892480609448924, 'samples': 18002944, 'steps': 35161, 'loss/train': 1.3056615591049194} 02/25/2022 23:06:37 - INFO - codeparrot_training - Step 35162: {'lr': 0.0001089112980509182, 'samples': 18003456, 'steps': 35162, 'loss/train': 1.7850764989852905} 02/25/2022 23:06:41 - INFO - codeparrot_training - Step 35163: {'lr': 0.00010889779061172647, 'samples': 18003968, 'steps': 35163, 'loss/train': 1.6812313795089722} 02/25/2022 23:06:47 - INFO - codeparrot_training - Step 35164: {'lr': 0.00010888428377697182, 'samples': 18004480, 'steps': 35164, 'loss/train': 1.410692811012268} 02/25/2022 23:06:51 - INFO - codeparrot_training - Step 35165: {'lr': 0.00010887077754671204, 'samples': 18004992, 'steps': 35165, 'loss/train': 1.505059838294983} 02/25/2022 23:06:56 - INFO - codeparrot_training - Step 35166: {'lr': 0.00010885727192100498, 'samples': 18005504, 'steps': 35166, 'loss/train': 1.4676064252853394} 02/25/2022 23:07:00 - INFO - codeparrot_training - Step 35167: {'lr': 0.00010884376689990864, 'samples': 18006016, 'steps': 35167, 'loss/train': 2.4696547985076904} 02/25/2022 23:07:06 - INFO - codeparrot_training - Step 35168: {'lr': 0.00010883026248348076, 'samples': 18006528, 'steps': 35168, 'loss/train': 1.7583203315734863} 02/25/2022 23:07:09 - INFO - codeparrot_training - Step 35169: {'lr': 0.0001088167586717792, 'samples': 18007040, 'steps': 35169, 'loss/train': 1.964342474937439} 02/25/2022 23:07:15 - INFO - codeparrot_training - Step 35170: {'lr': 0.00010880325546486174, 'samples': 18007552, 'steps': 35170, 'loss/train': 2.919741153717041} 02/25/2022 23:07:18 - INFO - codeparrot_training - Step 35171: {'lr': 0.00010878975286278641, 'samples': 18008064, 'steps': 35171, 'loss/train': 1.266263484954834} 02/25/2022 23:07:25 - INFO - codeparrot_training - Step 35172: {'lr': 0.00010877625086561091, 'samples': 18008576, 'steps': 35172, 'loss/train': 2.008727550506592} 02/25/2022 23:07:28 - INFO - codeparrot_training - Step 35173: {'lr': 0.00010876274947339312, 'samples': 18009088, 'steps': 35173, 'loss/train': 1.7658183574676514} 02/25/2022 23:07:34 - INFO - codeparrot_training - Step 35174: {'lr': 0.00010874924868619079, 'samples': 18009600, 'steps': 35174, 'loss/train': 1.285618782043457} 02/25/2022 23:07:37 - INFO - codeparrot_training - Step 35175: {'lr': 0.0001087357485040619, 'samples': 18010112, 'steps': 35175, 'loss/train': 2.5516717433929443} 02/25/2022 23:07:42 - INFO - codeparrot_training - Step 35176: {'lr': 0.0001087222489270642, 'samples': 18010624, 'steps': 35176, 'loss/train': 1.502387285232544} 02/25/2022 23:07:46 - INFO - codeparrot_training - Step 35177: {'lr': 0.00010870874995525553, 'samples': 18011136, 'steps': 35177, 'loss/train': 1.8763298988342285} 02/25/2022 23:07:51 - INFO - codeparrot_training - Step 35178: {'lr': 0.0001086952515886937, 'samples': 18011648, 'steps': 35178, 'loss/train': 1.0448024272918701} 02/25/2022 23:07:55 - INFO - codeparrot_training - Step 35179: {'lr': 0.00010868175382743648, 'samples': 18012160, 'steps': 35179, 'loss/train': 1.884372591972351} 02/25/2022 23:08:00 - INFO - codeparrot_training - Step 35180: {'lr': 0.00010866825667154182, 'samples': 18012672, 'steps': 35180, 'loss/train': 1.6680703163146973} 02/25/2022 23:08:04 - INFO - codeparrot_training - Step 35181: {'lr': 0.00010865476012106748, 'samples': 18013184, 'steps': 35181, 'loss/train': 2.2585673332214355} 02/25/2022 23:08:10 - INFO - codeparrot_training - Step 35182: {'lr': 0.00010864126417607126, 'samples': 18013696, 'steps': 35182, 'loss/train': 2.1255857944488525} 02/25/2022 23:08:14 - INFO - codeparrot_training - Step 35183: {'lr': 0.00010862776883661087, 'samples': 18014208, 'steps': 35183, 'loss/train': 1.3871996402740479} 02/25/2022 23:08:19 - INFO - codeparrot_training - Step 35184: {'lr': 0.00010861427410274435, 'samples': 18014720, 'steps': 35184, 'loss/train': 1.528178095817566} 02/25/2022 23:08:23 - INFO - codeparrot_training - Step 35185: {'lr': 0.00010860077997452927, 'samples': 18015232, 'steps': 35185, 'loss/train': 1.3357255458831787} 02/25/2022 23:08:28 - INFO - codeparrot_training - Step 35186: {'lr': 0.00010858728645202376, 'samples': 18015744, 'steps': 35186, 'loss/train': 1.7962992191314697} 02/25/2022 23:08:32 - INFO - codeparrot_training - Step 35187: {'lr': 0.0001085737935352852, 'samples': 18016256, 'steps': 35187, 'loss/train': 3.7248575687408447} 02/25/2022 23:08:37 - INFO - codeparrot_training - Step 35188: {'lr': 0.00010856030122437172, 'samples': 18016768, 'steps': 35188, 'loss/train': 2.344808578491211} 02/25/2022 23:08:41 - INFO - codeparrot_training - Step 35189: {'lr': 0.00010854680951934087, 'samples': 18017280, 'steps': 35189, 'loss/train': 1.3696507215499878} 02/25/2022 23:08:46 - INFO - codeparrot_training - Step 35190: {'lr': 0.0001085333184202508, 'samples': 18017792, 'steps': 35190, 'loss/train': 1.771506905555725} 02/25/2022 23:08:50 - INFO - codeparrot_training - Step 35191: {'lr': 0.00010851982792715886, 'samples': 18018304, 'steps': 35191, 'loss/train': 2.2414228916168213} 02/25/2022 23:08:55 - INFO - codeparrot_training - Step 35192: {'lr': 0.00010850633804012314, 'samples': 18018816, 'steps': 35192, 'loss/train': 0.7836082577705383} 02/25/2022 23:08:59 - INFO - codeparrot_training - Step 35193: {'lr': 0.00010849284875920126, 'samples': 18019328, 'steps': 35193, 'loss/train': 2.035017728805542} 02/25/2022 23:09:04 - INFO - codeparrot_training - Step 35194: {'lr': 0.00010847936008445116, 'samples': 18019840, 'steps': 35194, 'loss/train': 0.7475261688232422} 02/25/2022 23:09:08 - INFO - codeparrot_training - Step 35195: {'lr': 0.00010846587201593056, 'samples': 18020352, 'steps': 35195, 'loss/train': 0.2526957392692566} 02/25/2022 23:09:13 - INFO - codeparrot_training - Step 35196: {'lr': 0.00010845238455369722, 'samples': 18020864, 'steps': 35196, 'loss/train': 1.6900062561035156} 02/25/2022 23:09:17 - INFO - codeparrot_training - Step 35197: {'lr': 0.00010843889769780882, 'samples': 18021376, 'steps': 35197, 'loss/train': 1.048990249633789} 02/25/2022 23:09:24 - INFO - codeparrot_training - Step 35198: {'lr': 0.00010842541144832332, 'samples': 18021888, 'steps': 35198, 'loss/train': 0.8782660961151123} 02/25/2022 23:09:27 - INFO - codeparrot_training - Step 35199: {'lr': 0.00010841192580529841, 'samples': 18022400, 'steps': 35199, 'loss/train': 1.7634166479110718} 02/25/2022 23:09:33 - INFO - codeparrot_training - Step 35200: {'lr': 0.00010839844076879185, 'samples': 18022912, 'steps': 35200, 'loss/train': 0.7242740392684937} 02/25/2022 23:09:38 - INFO - codeparrot_training - Step 35201: {'lr': 0.0001083849563388613, 'samples': 18023424, 'steps': 35201, 'loss/train': 1.9755542278289795} 02/25/2022 23:09:42 - INFO - codeparrot_training - Step 35202: {'lr': 0.00010837147251556473, 'samples': 18023936, 'steps': 35202, 'loss/train': 1.863426923751831} 02/25/2022 23:09:47 - INFO - codeparrot_training - Step 35203: {'lr': 0.00010835798929895982, 'samples': 18024448, 'steps': 35203, 'loss/train': 1.0128402709960938} 02/25/2022 23:09:51 - INFO - codeparrot_training - Step 35204: {'lr': 0.00010834450668910428, 'samples': 18024960, 'steps': 35204, 'loss/train': 0.3819665312767029} 02/25/2022 23:09:56 - INFO - codeparrot_training - Step 35205: {'lr': 0.00010833102468605583, 'samples': 18025472, 'steps': 35205, 'loss/train': 0.9572913646697998} 02/25/2022 23:10:00 - INFO - codeparrot_training - Step 35206: {'lr': 0.00010831754328987235, 'samples': 18025984, 'steps': 35206, 'loss/train': 1.4223482608795166} 02/25/2022 23:10:05 - INFO - codeparrot_training - Step 35207: {'lr': 0.00010830406250061143, 'samples': 18026496, 'steps': 35207, 'loss/train': 1.9077860116958618} 02/25/2022 23:10:09 - INFO - codeparrot_training - Step 35208: {'lr': 0.00010829058231833111, 'samples': 18027008, 'steps': 35208, 'loss/train': 0.6413766145706177} 02/25/2022 23:10:15 - INFO - codeparrot_training - Step 35209: {'lr': 0.00010827710274308874, 'samples': 18027520, 'steps': 35209, 'loss/train': 3.6506991386413574} 02/25/2022 23:10:19 - INFO - codeparrot_training - Step 35210: {'lr': 0.00010826362377494236, 'samples': 18028032, 'steps': 35210, 'loss/train': 2.284710645675659} 02/25/2022 23:10:24 - INFO - codeparrot_training - Step 35211: {'lr': 0.00010825014541394961, 'samples': 18028544, 'steps': 35211, 'loss/train': 1.28914475440979} 02/25/2022 23:10:28 - INFO - codeparrot_training - Step 35212: {'lr': 0.00010823666766016824, 'samples': 18029056, 'steps': 35212, 'loss/train': 1.1726716756820679} 02/25/2022 23:10:33 - INFO - codeparrot_training - Step 35213: {'lr': 0.00010822319051365587, 'samples': 18029568, 'steps': 35213, 'loss/train': 1.8452595472335815} 02/25/2022 23:10:37 - INFO - codeparrot_training - Step 35214: {'lr': 0.00010820971397447044, 'samples': 18030080, 'steps': 35214, 'loss/train': 1.6830847263336182} 02/25/2022 23:10:42 - INFO - codeparrot_training - Step 35215: {'lr': 0.00010819623804266954, 'samples': 18030592, 'steps': 35215, 'loss/train': 1.6310985088348389} 02/25/2022 23:10:46 - INFO - codeparrot_training - Step 35216: {'lr': 0.00010818276271831093, 'samples': 18031104, 'steps': 35216, 'loss/train': 1.6741418838500977} 02/25/2022 23:10:51 - INFO - codeparrot_training - Step 35217: {'lr': 0.00010816928800145226, 'samples': 18031616, 'steps': 35217, 'loss/train': 2.0633130073547363} 02/25/2022 23:10:55 - INFO - codeparrot_training - Step 35218: {'lr': 0.0001081558138921514, 'samples': 18032128, 'steps': 35218, 'loss/train': 2.2960011959075928} 02/25/2022 23:11:01 - INFO - codeparrot_training - Step 35219: {'lr': 0.00010814234039046603, 'samples': 18032640, 'steps': 35219, 'loss/train': 0.6507323980331421} 02/25/2022 23:11:05 - INFO - codeparrot_training - Step 35220: {'lr': 0.0001081288674964537, 'samples': 18033152, 'steps': 35220, 'loss/train': 0.778766393661499} 02/25/2022 23:11:10 - INFO - codeparrot_training - Step 35221: {'lr': 0.00010811539521017236, 'samples': 18033664, 'steps': 35221, 'loss/train': 1.3828907012939453} 02/25/2022 23:11:14 - INFO - codeparrot_training - Step 35222: {'lr': 0.00010810192353167958, 'samples': 18034176, 'steps': 35222, 'loss/train': 2.4312429428100586} 02/25/2022 23:11:19 - INFO - codeparrot_training - Step 35223: {'lr': 0.00010808845246103313, 'samples': 18034688, 'steps': 35223, 'loss/train': 1.6656094789505005} 02/25/2022 23:11:23 - INFO - codeparrot_training - Step 35224: {'lr': 0.00010807498199829061, 'samples': 18035200, 'steps': 35224, 'loss/train': 1.657833218574524} 02/25/2022 23:11:28 - INFO - codeparrot_training - Step 35225: {'lr': 0.00010806151214350995, 'samples': 18035712, 'steps': 35225, 'loss/train': 0.8315533399581909} 02/25/2022 23:11:32 - INFO - codeparrot_training - Step 35226: {'lr': 0.00010804804289674855, 'samples': 18036224, 'steps': 35226, 'loss/train': 0.467655211687088} 02/25/2022 23:11:37 - INFO - codeparrot_training - Step 35227: {'lr': 0.00010803457425806435, 'samples': 18036736, 'steps': 35227, 'loss/train': 1.8076138496398926} 02/25/2022 23:11:41 - INFO - codeparrot_training - Step 35228: {'lr': 0.00010802110622751485, 'samples': 18037248, 'steps': 35228, 'loss/train': 1.5313631296157837} 02/25/2022 23:11:47 - INFO - codeparrot_training - Step 35229: {'lr': 0.00010800763880515802, 'samples': 18037760, 'steps': 35229, 'loss/train': 2.445680856704712} 02/25/2022 23:11:51 - INFO - codeparrot_training - Step 35230: {'lr': 0.00010799417199105121, 'samples': 18038272, 'steps': 35230, 'loss/train': 1.7447046041488647} 02/25/2022 23:11:56 - INFO - codeparrot_training - Step 35231: {'lr': 0.00010798070578525235, 'samples': 18038784, 'steps': 35231, 'loss/train': 1.5245343446731567} 02/25/2022 23:12:00 - INFO - codeparrot_training - Step 35232: {'lr': 0.00010796724018781898, 'samples': 18039296, 'steps': 35232, 'loss/train': 1.9911589622497559} 02/25/2022 23:12:05 - INFO - codeparrot_training - Step 35233: {'lr': 0.00010795377519880894, 'samples': 18039808, 'steps': 35233, 'loss/train': 0.8974743485450745} 02/25/2022 23:12:09 - INFO - codeparrot_training - Step 35234: {'lr': 0.00010794031081827982, 'samples': 18040320, 'steps': 35234, 'loss/train': 2.3746330738067627} 02/25/2022 23:12:14 - INFO - codeparrot_training - Step 35235: {'lr': 0.00010792684704628931, 'samples': 18040832, 'steps': 35235, 'loss/train': 0.8331019282341003} 02/25/2022 23:12:18 - INFO - codeparrot_training - Step 35236: {'lr': 0.00010791338388289495, 'samples': 18041344, 'steps': 35236, 'loss/train': 1.5934780836105347} 02/25/2022 23:12:24 - INFO - codeparrot_training - Step 35237: {'lr': 0.00010789992132815466, 'samples': 18041856, 'steps': 35237, 'loss/train': 1.2432156801223755} 02/25/2022 23:12:27 - INFO - codeparrot_training - Step 35238: {'lr': 0.00010788645938212596, 'samples': 18042368, 'steps': 35238, 'loss/train': 1.760582447052002} 02/25/2022 23:12:33 - INFO - codeparrot_training - Step 35239: {'lr': 0.00010787299804486656, 'samples': 18042880, 'steps': 35239, 'loss/train': 0.0737852081656456} 02/25/2022 23:12:36 - INFO - codeparrot_training - Step 35240: {'lr': 0.000107859537316434, 'samples': 18043392, 'steps': 35240, 'loss/train': 1.9848288297653198} 02/25/2022 23:12:41 - INFO - codeparrot_training - Step 35241: {'lr': 0.00010784607719688616, 'samples': 18043904, 'steps': 35241, 'loss/train': 0.8845615983009338} 02/25/2022 23:12:45 - INFO - codeparrot_training - Step 35242: {'lr': 0.00010783261768628047, 'samples': 18044416, 'steps': 35242, 'loss/train': 1.969267725944519} 02/25/2022 23:12:50 - INFO - codeparrot_training - Step 35243: {'lr': 0.00010781915878467491, 'samples': 18044928, 'steps': 35243, 'loss/train': 0.9654189944267273} 02/25/2022 23:12:54 - INFO - codeparrot_training - Step 35244: {'lr': 0.00010780570049212674, 'samples': 18045440, 'steps': 35244, 'loss/train': 1.4688910245895386} 02/25/2022 23:13:01 - INFO - codeparrot_training - Step 35245: {'lr': 0.00010779224280869387, 'samples': 18045952, 'steps': 35245, 'loss/train': 1.3934355974197388} 02/25/2022 23:13:04 - INFO - codeparrot_training - Step 35246: {'lr': 0.00010777878573443379, 'samples': 18046464, 'steps': 35246, 'loss/train': 1.5335108041763306} 02/25/2022 23:13:09 - INFO - codeparrot_training - Step 35247: {'lr': 0.00010776532926940441, 'samples': 18046976, 'steps': 35247, 'loss/train': 1.4982197284698486} 02/25/2022 23:13:13 - INFO - codeparrot_training - Step 35248: {'lr': 0.00010775187341366301, 'samples': 18047488, 'steps': 35248, 'loss/train': 1.4862370491027832} 02/25/2022 23:13:18 - INFO - codeparrot_training - Step 35249: {'lr': 0.00010773841816726754, 'samples': 18048000, 'steps': 35249, 'loss/train': 1.0844924449920654} 02/25/2022 23:13:22 - INFO - codeparrot_training - Step 35250: {'lr': 0.00010772496353027538, 'samples': 18048512, 'steps': 35250, 'loss/train': 1.5317862033843994} 02/25/2022 23:13:27 - INFO - codeparrot_training - Step 35251: {'lr': 0.00010771150950274453, 'samples': 18049024, 'steps': 35251, 'loss/train': 2.1743288040161133} 02/25/2022 23:13:31 - INFO - codeparrot_training - Step 35252: {'lr': 0.00010769805608473218, 'samples': 18049536, 'steps': 35252, 'loss/train': 0.7117295861244202} 02/25/2022 23:13:36 - INFO - codeparrot_training - Step 35253: {'lr': 0.00010768460327629625, 'samples': 18050048, 'steps': 35253, 'loss/train': 1.3391027450561523} 02/25/2022 23:13:40 - INFO - codeparrot_training - Step 35254: {'lr': 0.0001076711510774942, 'samples': 18050560, 'steps': 35254, 'loss/train': 1.0776550769805908} 02/25/2022 23:13:47 - INFO - codeparrot_training - Step 35255: {'lr': 0.00010765769948838386, 'samples': 18051072, 'steps': 35255, 'loss/train': 1.361800193786621} 02/25/2022 23:13:50 - INFO - codeparrot_training - Step 35256: {'lr': 0.00010764424850902271, 'samples': 18051584, 'steps': 35256, 'loss/train': 1.6386168003082275} 02/25/2022 23:13:56 - INFO - codeparrot_training - Step 35257: {'lr': 0.00010763079813946841, 'samples': 18052096, 'steps': 35257, 'loss/train': 1.8726242780685425} 02/25/2022 23:13:59 - INFO - codeparrot_training - Step 35258: {'lr': 0.00010761734837977857, 'samples': 18052608, 'steps': 35258, 'loss/train': 1.4635015726089478} 02/25/2022 23:14:05 - INFO - codeparrot_training - Step 35259: {'lr': 0.0001076038992300107, 'samples': 18053120, 'steps': 35259, 'loss/train': 1.6628472805023193} 02/25/2022 23:14:08 - INFO - codeparrot_training - Step 35260: {'lr': 0.00010759045069022261, 'samples': 18053632, 'steps': 35260, 'loss/train': 1.4462248086929321} 02/25/2022 23:14:14 - INFO - codeparrot_training - Step 35261: {'lr': 0.0001075770027604718, 'samples': 18054144, 'steps': 35261, 'loss/train': 2.0075807571411133} 02/25/2022 23:14:17 - INFO - codeparrot_training - Step 35262: {'lr': 0.00010756355544081589, 'samples': 18054656, 'steps': 35262, 'loss/train': 1.3023672103881836} 02/25/2022 23:14:23 - INFO - codeparrot_training - Step 35263: {'lr': 0.00010755010873131238, 'samples': 18055168, 'steps': 35263, 'loss/train': 2.8115971088409424} 02/25/2022 23:14:26 - INFO - codeparrot_training - Step 35264: {'lr': 0.00010753666263201906, 'samples': 18055680, 'steps': 35264, 'loss/train': 1.7383822202682495} 02/25/2022 23:14:33 - INFO - codeparrot_training - Step 35265: {'lr': 0.00010752321714299345, 'samples': 18056192, 'steps': 35265, 'loss/train': 0.48266932368278503} 02/25/2022 23:14:36 - INFO - codeparrot_training - Step 35266: {'lr': 0.00010750977226429315, 'samples': 18056704, 'steps': 35266, 'loss/train': 1.9738794565200806} 02/25/2022 23:14:42 - INFO - codeparrot_training - Step 35267: {'lr': 0.00010749632799597561, 'samples': 18057216, 'steps': 35267, 'loss/train': 1.268410563468933} 02/25/2022 23:14:45 - INFO - codeparrot_training - Step 35268: {'lr': 0.00010748288433809867, 'samples': 18057728, 'steps': 35268, 'loss/train': 1.5101932287216187} 02/25/2022 23:14:51 - INFO - codeparrot_training - Step 35269: {'lr': 0.00010746944129071981, 'samples': 18058240, 'steps': 35269, 'loss/train': 1.923219084739685} 02/25/2022 23:14:56 - INFO - codeparrot_training - Step 35270: {'lr': 0.00010745599885389656, 'samples': 18058752, 'steps': 35270, 'loss/train': 1.9769055843353271} 02/25/2022 23:15:00 - INFO - codeparrot_training - Step 35271: {'lr': 0.00010744255702768649, 'samples': 18059264, 'steps': 35271, 'loss/train': 1.3716689348220825} 02/25/2022 23:15:05 - INFO - codeparrot_training - Step 35272: {'lr': 0.00010742911581214734, 'samples': 18059776, 'steps': 35272, 'loss/train': 0.8333598375320435} 02/25/2022 23:15:09 - INFO - codeparrot_training - Step 35273: {'lr': 0.00010741567520733655, 'samples': 18060288, 'steps': 35273, 'loss/train': 1.40419602394104} 02/25/2022 23:15:15 - INFO - codeparrot_training - Step 35274: {'lr': 0.00010740223521331177, 'samples': 18060800, 'steps': 35274, 'loss/train': 2.644897937774658} 02/25/2022 23:15:19 - INFO - codeparrot_training - Step 35275: {'lr': 0.00010738879583013042, 'samples': 18061312, 'steps': 35275, 'loss/train': 0.8513860702514648} 02/25/2022 23:15:24 - INFO - codeparrot_training - Step 35276: {'lr': 0.00010737535705785028, 'samples': 18061824, 'steps': 35276, 'loss/train': 1.778937816619873} 02/25/2022 23:15:28 - INFO - codeparrot_training - Step 35277: {'lr': 0.00010736191889652883, 'samples': 18062336, 'steps': 35277, 'loss/train': 1.96829092502594} 02/25/2022 23:15:33 - INFO - codeparrot_training - Step 35278: {'lr': 0.00010734848134622362, 'samples': 18062848, 'steps': 35278, 'loss/train': 1.4290715456008911} 02/25/2022 23:15:37 - INFO - codeparrot_training - Step 35279: {'lr': 0.00010733504440699212, 'samples': 18063360, 'steps': 35279, 'loss/train': 1.3900266885757446} 02/25/2022 23:15:42 - INFO - codeparrot_training - Step 35280: {'lr': 0.0001073216080788921, 'samples': 18063872, 'steps': 35280, 'loss/train': 1.899182915687561} 02/25/2022 23:15:46 - INFO - codeparrot_training - Step 35281: {'lr': 0.0001073081723619809, 'samples': 18064384, 'steps': 35281, 'loss/train': 0.83708256483078} 02/25/2022 23:15:51 - INFO - codeparrot_training - Step 35282: {'lr': 0.00010729473725631636, 'samples': 18064896, 'steps': 35282, 'loss/train': 2.265195608139038} 02/25/2022 23:15:55 - INFO - codeparrot_training - Step 35283: {'lr': 0.00010728130276195569, 'samples': 18065408, 'steps': 35283, 'loss/train': 2.308539867401123} 02/25/2022 23:16:00 - INFO - codeparrot_training - Step 35284: {'lr': 0.00010726786887895667, 'samples': 18065920, 'steps': 35284, 'loss/train': 2.254373788833618} 02/25/2022 23:16:04 - INFO - codeparrot_training - Step 35285: {'lr': 0.0001072544356073767, 'samples': 18066432, 'steps': 35285, 'loss/train': 2.101296901702881} 02/25/2022 23:16:09 - INFO - codeparrot_training - Step 35286: {'lr': 0.00010724100294727357, 'samples': 18066944, 'steps': 35286, 'loss/train': 0.6933637857437134} 02/25/2022 23:16:13 - INFO - codeparrot_training - Step 35287: {'lr': 0.00010722757089870447, 'samples': 18067456, 'steps': 35287, 'loss/train': 2.2166860103607178} 02/25/2022 23:16:18 - INFO - codeparrot_training - Step 35288: {'lr': 0.00010721413946172722, 'samples': 18067968, 'steps': 35288, 'loss/train': 1.36432945728302} 02/25/2022 23:16:22 - INFO - codeparrot_training - Step 35289: {'lr': 0.00010720070863639916, 'samples': 18068480, 'steps': 35289, 'loss/train': 1.2467559576034546} 02/25/2022 23:16:29 - INFO - codeparrot_training - Step 35290: {'lr': 0.00010718727842277812, 'samples': 18068992, 'steps': 35290, 'loss/train': 1.0521498918533325} 02/25/2022 23:16:33 - INFO - codeparrot_training - Step 35291: {'lr': 0.00010717384882092124, 'samples': 18069504, 'steps': 35291, 'loss/train': 1.5979316234588623} 02/25/2022 23:16:38 - INFO - codeparrot_training - Step 35292: {'lr': 0.00010716041983088631, 'samples': 18070016, 'steps': 35292, 'loss/train': 2.316084384918213} 02/25/2022 23:16:42 - INFO - codeparrot_training - Step 35293: {'lr': 0.00010714699145273071, 'samples': 18070528, 'steps': 35293, 'loss/train': 0.8314520120620728} 02/25/2022 23:16:47 - INFO - codeparrot_training - Step 35294: {'lr': 0.00010713356368651212, 'samples': 18071040, 'steps': 35294, 'loss/train': 1.5896910429000854} 02/25/2022 23:16:51 - INFO - codeparrot_training - Step 35295: {'lr': 0.000107120136532288, 'samples': 18071552, 'steps': 35295, 'loss/train': 0.8118664622306824} 02/25/2022 23:16:56 - INFO - codeparrot_training - Step 35296: {'lr': 0.00010710670999011581, 'samples': 18072064, 'steps': 35296, 'loss/train': 1.6951003074645996} 02/25/2022 23:17:00 - INFO - codeparrot_training - Step 35297: {'lr': 0.00010709328406005303, 'samples': 18072576, 'steps': 35297, 'loss/train': 2.0402376651763916} 02/25/2022 23:17:05 - INFO - codeparrot_training - Step 35298: {'lr': 0.00010707985874215734, 'samples': 18073088, 'steps': 35298, 'loss/train': 1.452854871749878} 02/25/2022 23:17:09 - INFO - codeparrot_training - Step 35299: {'lr': 0.00010706643403648616, 'samples': 18073600, 'steps': 35299, 'loss/train': 2.5785157680511475} 02/25/2022 23:17:16 - INFO - codeparrot_training - Step 35300: {'lr': 0.00010705300994309697, 'samples': 18074112, 'steps': 35300, 'loss/train': 2.028458595275879} 02/25/2022 23:17:19 - INFO - codeparrot_training - Step 35301: {'lr': 0.00010703958646204722, 'samples': 18074624, 'steps': 35301, 'loss/train': 1.025451898574829} 02/25/2022 23:17:25 - INFO - codeparrot_training - Step 35302: {'lr': 0.00010702616359339457, 'samples': 18075136, 'steps': 35302, 'loss/train': 0.8433977961540222} 02/25/2022 23:17:28 - INFO - codeparrot_training - Step 35303: {'lr': 0.00010701274133719644, 'samples': 18075648, 'steps': 35303, 'loss/train': 2.3498735427856445} 02/25/2022 23:17:34 - INFO - codeparrot_training - Step 35304: {'lr': 0.00010699931969351032, 'samples': 18076160, 'steps': 35304, 'loss/train': 1.493093490600586} 02/25/2022 23:17:37 - INFO - codeparrot_training - Step 35305: {'lr': 0.00010698589866239361, 'samples': 18076672, 'steps': 35305, 'loss/train': 1.0323865413665771} 02/25/2022 23:17:43 - INFO - codeparrot_training - Step 35306: {'lr': 0.000106972478243904, 'samples': 18077184, 'steps': 35306, 'loss/train': 1.8889663219451904} 02/25/2022 23:17:46 - INFO - codeparrot_training - Step 35307: {'lr': 0.00010695905843809886, 'samples': 18077696, 'steps': 35307, 'loss/train': 1.72743558883667} 02/25/2022 23:17:52 - INFO - codeparrot_training - Step 35308: {'lr': 0.0001069456392450357, 'samples': 18078208, 'steps': 35308, 'loss/train': 1.4814544916152954} 02/25/2022 23:17:55 - INFO - codeparrot_training - Step 35309: {'lr': 0.00010693222066477199, 'samples': 18078720, 'steps': 35309, 'loss/train': 1.78554368019104} 02/25/2022 23:18:02 - INFO - codeparrot_training - Step 35310: {'lr': 0.00010691880269736514, 'samples': 18079232, 'steps': 35310, 'loss/train': 2.375300884246826} 02/25/2022 23:18:06 - INFO - codeparrot_training - Step 35311: {'lr': 0.00010690538534287278, 'samples': 18079744, 'steps': 35311, 'loss/train': 0.9333917498588562} 02/25/2022 23:18:11 - INFO - codeparrot_training - Step 35312: {'lr': 0.00010689196860135234, 'samples': 18080256, 'steps': 35312, 'loss/train': 1.6969847679138184} 02/25/2022 23:18:15 - INFO - codeparrot_training - Step 35313: {'lr': 0.00010687855247286124, 'samples': 18080768, 'steps': 35313, 'loss/train': 2.0370936393737793} 02/25/2022 23:18:20 - INFO - codeparrot_training - Step 35314: {'lr': 0.00010686513695745689, 'samples': 18081280, 'steps': 35314, 'loss/train': 2.5557234287261963} 02/25/2022 23:18:24 - INFO - codeparrot_training - Step 35315: {'lr': 0.00010685172205519695, 'samples': 18081792, 'steps': 35315, 'loss/train': 1.0665340423583984} 02/25/2022 23:18:29 - INFO - codeparrot_training - Step 35316: {'lr': 0.00010683830776613865, 'samples': 18082304, 'steps': 35316, 'loss/train': 1.908714771270752} 02/25/2022 23:18:33 - INFO - codeparrot_training - Step 35317: {'lr': 0.0001068248940903398, 'samples': 18082816, 'steps': 35317, 'loss/train': 0.5259528160095215} 02/25/2022 23:18:38 - INFO - codeparrot_training - Step 35318: {'lr': 0.00010681148102785743, 'samples': 18083328, 'steps': 35318, 'loss/train': 1.6868510246276855} 02/25/2022 23:18:42 - INFO - codeparrot_training - Step 35319: {'lr': 0.00010679806857874933, 'samples': 18083840, 'steps': 35319, 'loss/train': 2.5881948471069336} 02/25/2022 23:18:49 - INFO - codeparrot_training - Step 35320: {'lr': 0.00010678465674307273, 'samples': 18084352, 'steps': 35320, 'loss/train': 1.679875135421753} 02/25/2022 23:18:52 - INFO - codeparrot_training - Step 35321: {'lr': 0.00010677124552088535, 'samples': 18084864, 'steps': 35321, 'loss/train': 1.8995732069015503} 02/25/2022 23:18:58 - INFO - codeparrot_training - Step 35322: {'lr': 0.0001067578349122443, 'samples': 18085376, 'steps': 35322, 'loss/train': 1.9725558757781982} 02/25/2022 23:19:01 - INFO - codeparrot_training - Step 35323: {'lr': 0.00010674442491720729, 'samples': 18085888, 'steps': 35323, 'loss/train': 0.6953269243240356} 02/25/2022 23:19:07 - INFO - codeparrot_training - Step 35324: {'lr': 0.00010673101553583159, 'samples': 18086400, 'steps': 35324, 'loss/train': 1.7642712593078613} 02/25/2022 23:19:10 - INFO - codeparrot_training - Step 35325: {'lr': 0.0001067176067681749, 'samples': 18086912, 'steps': 35325, 'loss/train': 1.4429268836975098} 02/25/2022 23:19:16 - INFO - codeparrot_training - Step 35326: {'lr': 0.00010670419861429431, 'samples': 18087424, 'steps': 35326, 'loss/train': 0.9882174730300903} 02/25/2022 23:19:19 - INFO - codeparrot_training - Step 35327: {'lr': 0.00010669079107424753, 'samples': 18087936, 'steps': 35327, 'loss/train': 1.7968116998672485} 02/25/2022 23:19:25 - INFO - codeparrot_training - Step 35328: {'lr': 0.0001066773841480918, 'samples': 18088448, 'steps': 35328, 'loss/train': 1.209031581878662} 02/25/2022 23:19:28 - INFO - codeparrot_training - Step 35329: {'lr': 0.00010666397783588475, 'samples': 18088960, 'steps': 35329, 'loss/train': 2.0982604026794434} 02/25/2022 23:19:34 - INFO - codeparrot_training - Step 35330: {'lr': 0.00010665057213768367, 'samples': 18089472, 'steps': 35330, 'loss/train': 2.2163751125335693} 02/25/2022 23:19:37 - INFO - codeparrot_training - Step 35331: {'lr': 0.00010663716705354603, 'samples': 18089984, 'steps': 35331, 'loss/train': 0.7657140493392944} 02/25/2022 23:19:43 - INFO - codeparrot_training - Step 35332: {'lr': 0.00010662376258352919, 'samples': 18090496, 'steps': 35332, 'loss/train': 3.191768169403076} 02/25/2022 23:19:46 - INFO - codeparrot_training - Step 35333: {'lr': 0.00010661035872769067, 'samples': 18091008, 'steps': 35333, 'loss/train': 2.4975204467773438} 02/25/2022 23:19:52 - INFO - codeparrot_training - Step 35334: {'lr': 0.00010659695548608785, 'samples': 18091520, 'steps': 35334, 'loss/train': 1.4593830108642578} 02/25/2022 23:19:55 - INFO - codeparrot_training - Step 35335: {'lr': 0.00010658355285877816, 'samples': 18092032, 'steps': 35335, 'loss/train': 0.7651156187057495} 02/25/2022 23:20:02 - INFO - codeparrot_training - Step 35336: {'lr': 0.00010657015084581886, 'samples': 18092544, 'steps': 35336, 'loss/train': 1.025732159614563} 02/25/2022 23:20:06 - INFO - codeparrot_training - Step 35337: {'lr': 0.00010655674944726762, 'samples': 18093056, 'steps': 35337, 'loss/train': 1.3811231851577759} 02/25/2022 23:20:11 - INFO - codeparrot_training - Step 35338: {'lr': 0.00010654334866318171, 'samples': 18093568, 'steps': 35338, 'loss/train': 2.148015022277832} 02/25/2022 23:20:15 - INFO - codeparrot_training - Step 35339: {'lr': 0.00010652994849361852, 'samples': 18094080, 'steps': 35339, 'loss/train': 2.7251837253570557} 02/25/2022 23:20:20 - INFO - codeparrot_training - Step 35340: {'lr': 0.00010651654893863541, 'samples': 18094592, 'steps': 35340, 'loss/train': 1.3495391607284546} 02/25/2022 23:20:24 - INFO - codeparrot_training - Step 35341: {'lr': 0.00010650314999828992, 'samples': 18095104, 'steps': 35341, 'loss/train': 2.0667080879211426} 02/25/2022 23:20:29 - INFO - codeparrot_training - Step 35342: {'lr': 0.00010648975167263935, 'samples': 18095616, 'steps': 35342, 'loss/train': 1.8798643350601196} 02/25/2022 23:20:33 - INFO - codeparrot_training - Step 35343: {'lr': 0.00010647635396174116, 'samples': 18096128, 'steps': 35343, 'loss/train': 0.05614614486694336} 02/25/2022 23:20:38 - INFO - codeparrot_training - Step 35344: {'lr': 0.00010646295686565257, 'samples': 18096640, 'steps': 35344, 'loss/train': 1.8618308305740356} 02/25/2022 23:20:42 - INFO - codeparrot_training - Step 35345: {'lr': 0.00010644956038443118, 'samples': 18097152, 'steps': 35345, 'loss/train': 1.8949614763259888} 02/25/2022 23:20:49 - INFO - codeparrot_training - Step 35346: {'lr': 0.00010643616451813431, 'samples': 18097664, 'steps': 35346, 'loss/train': 1.681885838508606} 02/25/2022 23:20:52 - INFO - codeparrot_training - Step 35347: {'lr': 0.00010642276926681932, 'samples': 18098176, 'steps': 35347, 'loss/train': 0.7785062789916992} 02/25/2022 23:20:56 - INFO - codeparrot_training - Step 35348: {'lr': 0.00010640937463054351, 'samples': 18098688, 'steps': 35348, 'loss/train': 1.749366283416748} 02/25/2022 23:21:01 - INFO - codeparrot_training - Step 35349: {'lr': 0.00010639598060936442, 'samples': 18099200, 'steps': 35349, 'loss/train': 3.2186315059661865} 02/25/2022 23:21:05 - INFO - codeparrot_training - Step 35350: {'lr': 0.00010638258720333938, 'samples': 18099712, 'steps': 35350, 'loss/train': 1.393624186515808} 02/25/2022 23:21:10 - INFO - codeparrot_training - Step 35351: {'lr': 0.00010636919441252563, 'samples': 18100224, 'steps': 35351, 'loss/train': 1.5828620195388794} 02/25/2022 23:21:14 - INFO - codeparrot_training - Step 35352: {'lr': 0.00010635580223698071, 'samples': 18100736, 'steps': 35352, 'loss/train': 2.0323829650878906} 02/25/2022 23:21:19 - INFO - codeparrot_training - Step 35353: {'lr': 0.00010634241067676196, 'samples': 18101248, 'steps': 35353, 'loss/train': 1.068260908126831} 02/25/2022 23:21:23 - INFO - codeparrot_training - Step 35354: {'lr': 0.00010632901973192666, 'samples': 18101760, 'steps': 35354, 'loss/train': 2.039034128189087} 02/25/2022 23:21:28 - INFO - codeparrot_training - Step 35355: {'lr': 0.00010631562940253214, 'samples': 18102272, 'steps': 35355, 'loss/train': 1.6819913387298584} 02/25/2022 23:21:32 - INFO - codeparrot_training - Step 35356: {'lr': 0.00010630223968863606, 'samples': 18102784, 'steps': 35356, 'loss/train': 1.430169939994812} 02/25/2022 23:21:39 - INFO - codeparrot_training - Step 35357: {'lr': 0.00010628885059029534, 'samples': 18103296, 'steps': 35357, 'loss/train': 2.618424415588379} 02/25/2022 23:21:42 - INFO - codeparrot_training - Step 35358: {'lr': 0.00010627546210756763, 'samples': 18103808, 'steps': 35358, 'loss/train': 1.5210448503494263} 02/25/2022 23:21:48 - INFO - codeparrot_training - Step 35359: {'lr': 0.00010626207424051012, 'samples': 18104320, 'steps': 35359, 'loss/train': 2.4756479263305664} 02/25/2022 23:21:51 - INFO - codeparrot_training - Step 35360: {'lr': 0.00010624868698918044, 'samples': 18104832, 'steps': 35360, 'loss/train': 0.47095155715942383} 02/25/2022 23:21:57 - INFO - codeparrot_training - Step 35361: {'lr': 0.00010623530035363557, 'samples': 18105344, 'steps': 35361, 'loss/train': 1.2323811054229736} 02/25/2022 23:22:00 - INFO - codeparrot_training - Step 35362: {'lr': 0.00010622191433393308, 'samples': 18105856, 'steps': 35362, 'loss/train': 1.3965120315551758} 02/25/2022 23:22:06 - INFO - codeparrot_training - Step 35363: {'lr': 0.00010620852893013019, 'samples': 18106368, 'steps': 35363, 'loss/train': 2.1996281147003174} 02/25/2022 23:22:09 - INFO - codeparrot_training - Step 35364: {'lr': 0.00010619514414228437, 'samples': 18106880, 'steps': 35364, 'loss/train': 1.555686593055725} 02/25/2022 23:22:15 - INFO - codeparrot_training - Step 35365: {'lr': 0.0001061817599704529, 'samples': 18107392, 'steps': 35365, 'loss/train': 2.06492018699646} 02/25/2022 23:22:18 - INFO - codeparrot_training - Step 35366: {'lr': 0.00010616837641469309, 'samples': 18107904, 'steps': 35366, 'loss/train': 1.7434812784194946} 02/25/2022 23:22:25 - INFO - codeparrot_training - Step 35367: {'lr': 0.00010615499347506221, 'samples': 18108416, 'steps': 35367, 'loss/train': 1.6285113096237183} 02/25/2022 23:22:28 - INFO - codeparrot_training - Step 35368: {'lr': 0.00010614161115161777, 'samples': 18108928, 'steps': 35368, 'loss/train': 0.14356619119644165} 02/25/2022 23:22:34 - INFO - codeparrot_training - Step 35369: {'lr': 0.00010612822944441697, 'samples': 18109440, 'steps': 35369, 'loss/train': 1.6089484691619873} 02/25/2022 23:22:37 - INFO - codeparrot_training - Step 35370: {'lr': 0.00010611484835351715, 'samples': 18109952, 'steps': 35370, 'loss/train': 0.8704349994659424} 02/25/2022 23:22:43 - INFO - codeparrot_training - Step 35371: {'lr': 0.0001061014678789755, 'samples': 18110464, 'steps': 35371, 'loss/train': 0.9865365624427795} 02/25/2022 23:22:47 - INFO - codeparrot_training - Step 35372: {'lr': 0.00010608808802084963, 'samples': 18110976, 'steps': 35372, 'loss/train': 1.3249534368515015} 02/25/2022 23:22:52 - INFO - codeparrot_training - Step 35373: {'lr': 0.00010607470877919664, 'samples': 18111488, 'steps': 35373, 'loss/train': 1.9026554822921753} 02/25/2022 23:22:56 - INFO - codeparrot_training - Step 35374: {'lr': 0.00010606133015407393, 'samples': 18112000, 'steps': 35374, 'loss/train': 2.0561776161193848} 02/25/2022 23:23:01 - INFO - codeparrot_training - Step 35375: {'lr': 0.00010604795214553867, 'samples': 18112512, 'steps': 35375, 'loss/train': 1.7132539749145508} 02/25/2022 23:23:05 - INFO - codeparrot_training - Step 35376: {'lr': 0.00010603457475364836, 'samples': 18113024, 'steps': 35376, 'loss/train': 1.452529788017273} 02/25/2022 23:23:11 - INFO - codeparrot_training - Step 35377: {'lr': 0.00010602119797846013, 'samples': 18113536, 'steps': 35377, 'loss/train': 1.4385355710983276} 02/25/2022 23:23:14 - INFO - codeparrot_training - Step 35378: {'lr': 0.00010600782182003155, 'samples': 18114048, 'steps': 35378, 'loss/train': 2.2396726608276367} 02/25/2022 23:23:20 - INFO - codeparrot_training - Step 35379: {'lr': 0.00010599444627841956, 'samples': 18114560, 'steps': 35379, 'loss/train': 1.9014734029769897} 02/25/2022 23:23:23 - INFO - codeparrot_training - Step 35380: {'lr': 0.00010598107135368174, 'samples': 18115072, 'steps': 35380, 'loss/train': 1.8209419250488281} 02/25/2022 23:23:29 - INFO - codeparrot_training - Step 35381: {'lr': 0.00010596769704587517, 'samples': 18115584, 'steps': 35381, 'loss/train': 1.9149832725524902} 02/25/2022 23:23:32 - INFO - codeparrot_training - Step 35382: {'lr': 0.00010595432335505745, 'samples': 18116096, 'steps': 35382, 'loss/train': 1.4094489812850952} 02/25/2022 23:23:39 - INFO - codeparrot_training - Step 35383: {'lr': 0.00010594095028128544, 'samples': 18116608, 'steps': 35383, 'loss/train': 2.4266788959503174} 02/25/2022 23:23:42 - INFO - codeparrot_training - Step 35384: {'lr': 0.00010592757782461679, 'samples': 18117120, 'steps': 35384, 'loss/train': 1.7293343544006348} 02/25/2022 23:23:48 - INFO - codeparrot_training - Step 35385: {'lr': 0.00010591420598510856, 'samples': 18117632, 'steps': 35385, 'loss/train': 1.4500473737716675} 02/25/2022 23:23:51 - INFO - codeparrot_training - Step 35386: {'lr': 0.00010590083476281828, 'samples': 18118144, 'steps': 35386, 'loss/train': 1.6789897680282593} 02/25/2022 23:23:57 - INFO - codeparrot_training - Step 35387: {'lr': 0.0001058874641578029, 'samples': 18118656, 'steps': 35387, 'loss/train': 1.9555487632751465} 02/25/2022 23:24:00 - INFO - codeparrot_training - Step 35388: {'lr': 0.00010587409417011995, 'samples': 18119168, 'steps': 35388, 'loss/train': 0.8048198223114014} 02/25/2022 23:24:06 - INFO - codeparrot_training - Step 35389: {'lr': 0.00010586072479982661, 'samples': 18119680, 'steps': 35389, 'loss/train': 2.1249825954437256} 02/25/2022 23:24:10 - INFO - codeparrot_training - Step 35390: {'lr': 0.00010584735604698007, 'samples': 18120192, 'steps': 35390, 'loss/train': 0.6738342642784119} 02/25/2022 23:24:15 - INFO - codeparrot_training - Step 35391: {'lr': 0.00010583398791163779, 'samples': 18120704, 'steps': 35391, 'loss/train': 1.5148173570632935} 02/25/2022 23:24:19 - INFO - codeparrot_training - Step 35392: {'lr': 0.00010582062039385689, 'samples': 18121216, 'steps': 35392, 'loss/train': 2.119899272918701} 02/25/2022 23:24:25 - INFO - codeparrot_training - Step 35393: {'lr': 0.00010580725349369469, 'samples': 18121728, 'steps': 35393, 'loss/train': 1.3911043405532837} 02/25/2022 23:24:29 - INFO - codeparrot_training - Step 35394: {'lr': 0.00010579388721120833, 'samples': 18122240, 'steps': 35394, 'loss/train': 2.0923969745635986} 02/25/2022 23:24:34 - INFO - codeparrot_training - Step 35395: {'lr': 0.00010578052154645528, 'samples': 18122752, 'steps': 35395, 'loss/train': 2.4774179458618164} 02/25/2022 23:24:38 - INFO - codeparrot_training - Step 35396: {'lr': 0.00010576715649949268, 'samples': 18123264, 'steps': 35396, 'loss/train': 2.2661447525024414} 02/25/2022 23:24:43 - INFO - codeparrot_training - Step 35397: {'lr': 0.00010575379207037774, 'samples': 18123776, 'steps': 35397, 'loss/train': 2.8824970722198486} 02/25/2022 23:24:47 - INFO - codeparrot_training - Step 35398: {'lr': 0.00010574042825916771, 'samples': 18124288, 'steps': 35398, 'loss/train': 2.696951150894165} 02/25/2022 23:24:52 - INFO - codeparrot_training - Step 35399: {'lr': 0.00010572706506591995, 'samples': 18124800, 'steps': 35399, 'loss/train': 1.6601413488388062} 02/25/2022 23:24:56 - INFO - codeparrot_training - Step 35400: {'lr': 0.00010571370249069162, 'samples': 18125312, 'steps': 35400, 'loss/train': 2.7954585552215576} 02/25/2022 23:25:01 - INFO - codeparrot_training - Step 35401: {'lr': 0.00010570034053354, 'samples': 18125824, 'steps': 35401, 'loss/train': 2.2081997394561768} 02/25/2022 23:25:05 - INFO - codeparrot_training - Step 35402: {'lr': 0.0001056869791945222, 'samples': 18126336, 'steps': 35402, 'loss/train': 1.6273868083953857} 02/25/2022 23:25:11 - INFO - codeparrot_training - Step 35403: {'lr': 0.00010567361847369561, 'samples': 18126848, 'steps': 35403, 'loss/train': 1.9625701904296875} 02/25/2022 23:25:15 - INFO - codeparrot_training - Step 35404: {'lr': 0.00010566025837111746, 'samples': 18127360, 'steps': 35404, 'loss/train': 1.2657372951507568} 02/25/2022 23:25:20 - INFO - codeparrot_training - Step 35405: {'lr': 0.00010564689888684489, 'samples': 18127872, 'steps': 35405, 'loss/train': 0.6094422936439514} 02/25/2022 23:25:24 - INFO - codeparrot_training - Step 35406: {'lr': 0.00010563354002093511, 'samples': 18128384, 'steps': 35406, 'loss/train': 3.6658592224121094} 02/25/2022 23:25:29 - INFO - codeparrot_training - Step 35407: {'lr': 0.00010562018177344548, 'samples': 18128896, 'steps': 35407, 'loss/train': 1.5406415462493896} 02/25/2022 23:25:33 - INFO - codeparrot_training - Step 35408: {'lr': 0.00010560682414443315, 'samples': 18129408, 'steps': 35408, 'loss/train': 1.6650298833847046} 02/25/2022 23:25:38 - INFO - codeparrot_training - Step 35409: {'lr': 0.00010559346713395531, 'samples': 18129920, 'steps': 35409, 'loss/train': 1.2835291624069214} 02/25/2022 23:25:42 - INFO - codeparrot_training - Step 35410: {'lr': 0.00010558011074206914, 'samples': 18130432, 'steps': 35410, 'loss/train': 1.1159237623214722} 02/25/2022 23:25:47 - INFO - codeparrot_training - Step 35411: {'lr': 0.000105566754968832, 'samples': 18130944, 'steps': 35411, 'loss/train': 1.5116472244262695} 02/25/2022 23:25:51 - INFO - codeparrot_training - Step 35412: {'lr': 0.00010555339981430093, 'samples': 18131456, 'steps': 35412, 'loss/train': 1.252220869064331} 02/25/2022 23:25:57 - INFO - codeparrot_training - Step 35413: {'lr': 0.00010554004527853341, 'samples': 18131968, 'steps': 35413, 'loss/train': 1.7288554906845093} 02/25/2022 23:26:01 - INFO - codeparrot_training - Step 35414: {'lr': 0.0001055266913615863, 'samples': 18132480, 'steps': 35414, 'loss/train': 1.1541352272033691} 02/25/2022 23:26:06 - INFO - codeparrot_training - Step 35415: {'lr': 0.00010551333806351704, 'samples': 18132992, 'steps': 35415, 'loss/train': 2.1531450748443604} 02/25/2022 23:26:10 - INFO - codeparrot_training - Step 35416: {'lr': 0.00010549998538438265, 'samples': 18133504, 'steps': 35416, 'loss/train': 0.8569403290748596} 02/25/2022 23:26:16 - INFO - codeparrot_training - Step 35417: {'lr': 0.00010548663332424067, 'samples': 18134016, 'steps': 35417, 'loss/train': 3.468822717666626} 02/25/2022 23:26:19 - INFO - codeparrot_training - Step 35418: {'lr': 0.00010547328188314784, 'samples': 18134528, 'steps': 35418, 'loss/train': 0.6029434204101562} 02/25/2022 23:26:25 - INFO - codeparrot_training - Step 35419: {'lr': 0.0001054599310611617, 'samples': 18135040, 'steps': 35419, 'loss/train': 1.1743953227996826} 02/25/2022 23:26:28 - INFO - codeparrot_training - Step 35420: {'lr': 0.00010544658085833919, 'samples': 18135552, 'steps': 35420, 'loss/train': 0.552441418170929} 02/25/2022 23:26:34 - INFO - codeparrot_training - Step 35421: {'lr': 0.00010543323127473784, 'samples': 18136064, 'steps': 35421, 'loss/train': 2.759134292602539} 02/25/2022 23:26:37 - INFO - codeparrot_training - Step 35422: {'lr': 0.00010541988231041444, 'samples': 18136576, 'steps': 35422, 'loss/train': 0.5660808682441711} 02/25/2022 23:26:44 - INFO - codeparrot_training - Step 35423: {'lr': 0.00010540653396542643, 'samples': 18137088, 'steps': 35423, 'loss/train': 2.1836600303649902} 02/25/2022 23:26:47 - INFO - codeparrot_training - Step 35424: {'lr': 0.00010539318623983082, 'samples': 18137600, 'steps': 35424, 'loss/train': 1.5300904512405396} 02/25/2022 23:26:53 - INFO - codeparrot_training - Step 35425: {'lr': 0.00010537983913368499, 'samples': 18138112, 'steps': 35425, 'loss/train': 1.9864635467529297} 02/25/2022 23:26:56 - INFO - codeparrot_training - Step 35426: {'lr': 0.00010536649264704601, 'samples': 18138624, 'steps': 35426, 'loss/train': 1.4066071510314941} 02/25/2022 23:27:02 - INFO - codeparrot_training - Step 35427: {'lr': 0.00010535314677997101, 'samples': 18139136, 'steps': 35427, 'loss/train': 2.257549285888672} 02/25/2022 23:27:06 - INFO - codeparrot_training - Step 35428: {'lr': 0.00010533980153251716, 'samples': 18139648, 'steps': 35428, 'loss/train': 1.54520845413208} 02/25/2022 23:27:09 - INFO - codeparrot_training - Step 35429: {'lr': 0.00010532645690474172, 'samples': 18140160, 'steps': 35429, 'loss/train': 8.41403579711914} 02/25/2022 23:27:15 - INFO - codeparrot_training - Step 35430: {'lr': 0.0001053131128967018, 'samples': 18140672, 'steps': 35430, 'loss/train': 0.6244770288467407} 02/25/2022 23:27:18 - INFO - codeparrot_training - Step 35431: {'lr': 0.0001052997695084546, 'samples': 18141184, 'steps': 35431, 'loss/train': 1.1674411296844482} 02/25/2022 23:27:24 - INFO - codeparrot_training - Step 35432: {'lr': 0.00010528642674005712, 'samples': 18141696, 'steps': 35432, 'loss/train': 0.9195600152015686} 02/25/2022 23:27:27 - INFO - codeparrot_training - Step 35433: {'lr': 0.00010527308459156674, 'samples': 18142208, 'steps': 35433, 'loss/train': 2.0221030712127686} 02/25/2022 23:27:33 - INFO - codeparrot_training - Step 35434: {'lr': 0.00010525974306304048, 'samples': 18142720, 'steps': 35434, 'loss/train': 0.85420823097229} 02/25/2022 23:27:36 - INFO - codeparrot_training - Step 35435: {'lr': 0.00010524640215453555, 'samples': 18143232, 'steps': 35435, 'loss/train': 1.716320276260376} 02/25/2022 23:27:42 - INFO - codeparrot_training - Step 35436: {'lr': 0.00010523306186610906, 'samples': 18143744, 'steps': 35436, 'loss/train': 2.248636245727539} 02/25/2022 23:27:45 - INFO - codeparrot_training - Step 35437: {'lr': 0.00010521972219781805, 'samples': 18144256, 'steps': 35437, 'loss/train': 1.2615989446640015} 02/25/2022 23:27:50 - INFO - codeparrot_training - Step 35438: {'lr': 0.00010520638314971991, 'samples': 18144768, 'steps': 35438, 'loss/train': 1.4481772184371948} 02/25/2022 23:27:54 - INFO - codeparrot_training - Step 35439: {'lr': 0.00010519304472187164, 'samples': 18145280, 'steps': 35439, 'loss/train': 1.5319799184799194} 02/25/2022 23:28:01 - INFO - codeparrot_training - Step 35440: {'lr': 0.00010517970691433035, 'samples': 18145792, 'steps': 35440, 'loss/train': 1.281589388847351} 02/25/2022 23:28:04 - INFO - codeparrot_training - Step 35441: {'lr': 0.00010516636972715315, 'samples': 18146304, 'steps': 35441, 'loss/train': 1.3462690114974976} 02/25/2022 23:28:10 - INFO - codeparrot_training - Step 35442: {'lr': 0.00010515303316039732, 'samples': 18146816, 'steps': 35442, 'loss/train': 1.4810951948165894} 02/25/2022 23:28:13 - INFO - codeparrot_training - Step 35443: {'lr': 0.00010513969721411989, 'samples': 18147328, 'steps': 35443, 'loss/train': 1.9309208393096924} 02/25/2022 23:28:19 - INFO - codeparrot_training - Step 35444: {'lr': 0.00010512636188837801, 'samples': 18147840, 'steps': 35444, 'loss/train': 1.8213180303573608} 02/25/2022 23:28:22 - INFO - codeparrot_training - Step 35445: {'lr': 0.00010511302718322871, 'samples': 18148352, 'steps': 35445, 'loss/train': 1.5979044437408447} 02/25/2022 23:28:28 - INFO - codeparrot_training - Step 35446: {'lr': 0.00010509969309872927, 'samples': 18148864, 'steps': 35446, 'loss/train': 1.7691913843154907} 02/25/2022 23:28:31 - INFO - codeparrot_training - Step 35447: {'lr': 0.00010508635963493673, 'samples': 18149376, 'steps': 35447, 'loss/train': 1.524527668952942} 02/25/2022 23:28:37 - INFO - codeparrot_training - Step 35448: {'lr': 0.00010507302679190823, 'samples': 18149888, 'steps': 35448, 'loss/train': 3.2382192611694336} 02/25/2022 23:28:40 - INFO - codeparrot_training - Step 35449: {'lr': 0.00010505969456970074, 'samples': 18150400, 'steps': 35449, 'loss/train': 1.6202162504196167} 02/25/2022 23:28:47 - INFO - codeparrot_training - Step 35450: {'lr': 0.00010504636296837161, 'samples': 18150912, 'steps': 35450, 'loss/train': 1.23264479637146} 02/25/2022 23:28:50 - INFO - codeparrot_training - Step 35451: {'lr': 0.00010503303198797773, 'samples': 18151424, 'steps': 35451, 'loss/train': 2.352633476257324} 02/25/2022 23:28:56 - INFO - codeparrot_training - Step 35452: {'lr': 0.00010501970162857652, 'samples': 18151936, 'steps': 35452, 'loss/train': 1.6278376579284668} 02/25/2022 23:28:59 - INFO - codeparrot_training - Step 35453: {'lr': 0.00010500637189022466, 'samples': 18152448, 'steps': 35453, 'loss/train': 1.8849884271621704} 02/25/2022 23:29:05 - INFO - codeparrot_training - Step 35454: {'lr': 0.00010499304277297955, 'samples': 18152960, 'steps': 35454, 'loss/train': 0.927249550819397} 02/25/2022 23:29:08 - INFO - codeparrot_training - Step 35455: {'lr': 0.00010497971427689813, 'samples': 18153472, 'steps': 35455, 'loss/train': 0.3041379451751709} 02/25/2022 23:29:14 - INFO - codeparrot_training - Step 35456: {'lr': 0.00010496638640203774, 'samples': 18153984, 'steps': 35456, 'loss/train': 1.241897463798523} 02/25/2022 23:29:17 - INFO - codeparrot_training - Step 35457: {'lr': 0.0001049530591484551, 'samples': 18154496, 'steps': 35457, 'loss/train': 1.42458975315094} 02/25/2022 23:29:23 - INFO - codeparrot_training - Step 35458: {'lr': 0.00010493973251620761, 'samples': 18155008, 'steps': 35458, 'loss/train': 2.3158721923828125} 02/25/2022 23:29:26 - INFO - codeparrot_training - Step 35459: {'lr': 0.00010492640650535216, 'samples': 18155520, 'steps': 35459, 'loss/train': 1.0699232816696167} 02/25/2022 23:29:32 - INFO - codeparrot_training - Step 35460: {'lr': 0.00010491308111594597, 'samples': 18156032, 'steps': 35460, 'loss/train': 2.5548694133758545} 02/25/2022 23:29:35 - INFO - codeparrot_training - Step 35461: {'lr': 0.00010489975634804611, 'samples': 18156544, 'steps': 35461, 'loss/train': 1.7255939245224} 02/25/2022 23:29:41 - INFO - codeparrot_training - Step 35462: {'lr': 0.0001048864322017096, 'samples': 18157056, 'steps': 35462, 'loss/train': 1.5617070198059082} 02/25/2022 23:29:44 - INFO - codeparrot_training - Step 35463: {'lr': 0.00010487310867699345, 'samples': 18157568, 'steps': 35463, 'loss/train': 1.327529788017273} 02/25/2022 23:29:50 - INFO - codeparrot_training - Step 35464: {'lr': 0.00010485978577395492, 'samples': 18158080, 'steps': 35464, 'loss/train': 3.01853346824646} 02/25/2022 23:29:53 - INFO - codeparrot_training - Step 35465: {'lr': 0.00010484646349265095, 'samples': 18158592, 'steps': 35465, 'loss/train': 0.8646223545074463} 02/25/2022 23:30:01 - INFO - codeparrot_training - Step 35466: {'lr': 0.00010483314183313866, 'samples': 18159104, 'steps': 35466, 'loss/train': 1.061560034751892} 02/25/2022 23:30:04 - INFO - codeparrot_training - Step 35467: {'lr': 0.000104819820795475, 'samples': 18159616, 'steps': 35467, 'loss/train': 1.4646320343017578} 02/25/2022 23:30:10 - INFO - codeparrot_training - Step 35468: {'lr': 0.00010480650037971723, 'samples': 18160128, 'steps': 35468, 'loss/train': 2.607125759124756} 02/25/2022 23:30:13 - INFO - codeparrot_training - Step 35469: {'lr': 0.0001047931805859223, 'samples': 18160640, 'steps': 35469, 'loss/train': 1.151485800743103} 02/25/2022 23:30:19 - INFO - codeparrot_training - Step 35470: {'lr': 0.00010477986141414724, 'samples': 18161152, 'steps': 35470, 'loss/train': 0.38980814814567566} 02/25/2022 23:30:22 - INFO - codeparrot_training - Step 35471: {'lr': 0.00010476654286444912, 'samples': 18161664, 'steps': 35471, 'loss/train': 2.2634479999542236} 02/25/2022 23:30:28 - INFO - codeparrot_training - Step 35472: {'lr': 0.00010475322493688505, 'samples': 18162176, 'steps': 35472, 'loss/train': 2.209099531173706} 02/25/2022 23:30:31 - INFO - codeparrot_training - Step 35473: {'lr': 0.00010473990763151206, 'samples': 18162688, 'steps': 35473, 'loss/train': 2.261345386505127} 02/25/2022 23:30:37 - INFO - codeparrot_training - Step 35474: {'lr': 0.00010472659094838719, 'samples': 18163200, 'steps': 35474, 'loss/train': 2.2410130500793457} 02/25/2022 23:30:40 - INFO - codeparrot_training - Step 35475: {'lr': 0.00010471327488756738, 'samples': 18163712, 'steps': 35475, 'loss/train': 1.7829253673553467} 02/25/2022 23:30:47 - INFO - codeparrot_training - Step 35476: {'lr': 0.00010469995944910985, 'samples': 18164224, 'steps': 35476, 'loss/train': 1.9092180728912354} 02/25/2022 23:30:50 - INFO - codeparrot_training - Step 35477: {'lr': 0.00010468664463307157, 'samples': 18164736, 'steps': 35477, 'loss/train': 1.5434972047805786} 02/25/2022 23:30:56 - INFO - codeparrot_training - Step 35478: {'lr': 0.00010467333043950956, 'samples': 18165248, 'steps': 35478, 'loss/train': 1.085848093032837} 02/25/2022 23:30:59 - INFO - codeparrot_training - Step 35479: {'lr': 0.00010466001686848072, 'samples': 18165760, 'steps': 35479, 'loss/train': 2.576321601867676} 02/25/2022 23:31:05 - INFO - codeparrot_training - Step 35480: {'lr': 0.00010464670392004236, 'samples': 18166272, 'steps': 35480, 'loss/train': 2.178196668624878} 02/25/2022 23:31:08 - INFO - codeparrot_training - Step 35481: {'lr': 0.00010463339159425137, 'samples': 18166784, 'steps': 35481, 'loss/train': 1.6818640232086182} 02/25/2022 23:31:13 - INFO - codeparrot_training - Step 35482: {'lr': 0.00010462007989116473, 'samples': 18167296, 'steps': 35482, 'loss/train': 1.6242027282714844} 02/25/2022 23:31:17 - INFO - codeparrot_training - Step 35483: {'lr': 0.00010460676881083953, 'samples': 18167808, 'steps': 35483, 'loss/train': 2.3885984420776367} 02/25/2022 23:31:23 - INFO - codeparrot_training - Step 35484: {'lr': 0.00010459345835333267, 'samples': 18168320, 'steps': 35484, 'loss/train': 0.04929136484861374} 02/25/2022 23:31:26 - INFO - codeparrot_training - Step 35485: {'lr': 0.00010458014851870134, 'samples': 18168832, 'steps': 35485, 'loss/train': 1.3643733263015747} 02/25/2022 23:31:32 - INFO - codeparrot_training - Step 35486: {'lr': 0.00010456683930700243, 'samples': 18169344, 'steps': 35486, 'loss/train': 1.028764009475708} 02/25/2022 23:31:36 - INFO - codeparrot_training - Step 35487: {'lr': 0.00010455353071829315, 'samples': 18169856, 'steps': 35487, 'loss/train': 0.9290938377380371} 02/25/2022 23:31:41 - INFO - codeparrot_training - Step 35488: {'lr': 0.00010454022275263017, 'samples': 18170368, 'steps': 35488, 'loss/train': 2.4605071544647217} 02/25/2022 23:31:45 - INFO - codeparrot_training - Step 35489: {'lr': 0.00010452691541007078, 'samples': 18170880, 'steps': 35489, 'loss/train': 1.9446767568588257} 02/25/2022 23:31:50 - INFO - codeparrot_training - Step 35490: {'lr': 0.00010451360869067183, 'samples': 18171392, 'steps': 35490, 'loss/train': 0.18995735049247742} 02/25/2022 23:31:54 - INFO - codeparrot_training - Step 35491: {'lr': 0.00010450030259449053, 'samples': 18171904, 'steps': 35491, 'loss/train': 2.010329484939575} 02/25/2022 23:32:00 - INFO - codeparrot_training - Step 35492: {'lr': 0.00010448699712158357, 'samples': 18172416, 'steps': 35492, 'loss/train': 3.019463062286377} 02/25/2022 23:32:03 - INFO - codeparrot_training - Step 35493: {'lr': 0.0001044736922720082, 'samples': 18172928, 'steps': 35493, 'loss/train': 0.7595480680465698} 02/25/2022 23:32:08 - INFO - codeparrot_training - Step 35494: {'lr': 0.00010446038804582122, 'samples': 18173440, 'steps': 35494, 'loss/train': 2.158639430999756} 02/25/2022 23:32:12 - INFO - codeparrot_training - Step 35495: {'lr': 0.0001044470844430799, 'samples': 18173952, 'steps': 35495, 'loss/train': 2.691209077835083} 02/25/2022 23:32:18 - INFO - codeparrot_training - Step 35496: {'lr': 0.00010443378146384088, 'samples': 18174464, 'steps': 35496, 'loss/train': 1.186385154724121} 02/25/2022 23:32:21 - INFO - codeparrot_training - Step 35497: {'lr': 0.00010442047910816144, 'samples': 18174976, 'steps': 35497, 'loss/train': 1.7737675905227661} 02/25/2022 23:32:27 - INFO - codeparrot_training - Step 35498: {'lr': 0.00010440717737609831, 'samples': 18175488, 'steps': 35498, 'loss/train': 1.436472773551941} 02/25/2022 23:32:30 - INFO - codeparrot_training - Step 35499: {'lr': 0.00010439387626770869, 'samples': 18176000, 'steps': 35499, 'loss/train': 2.16960072517395} 02/25/2022 23:32:36 - INFO - codeparrot_training - Step 35500: {'lr': 0.0001043805757830495, 'samples': 18176512, 'steps': 35500, 'loss/train': 1.3573946952819824} 02/25/2022 23:32:39 - INFO - codeparrot_training - Step 35501: {'lr': 0.00010436727592217766, 'samples': 18177024, 'steps': 35501, 'loss/train': 1.2822918891906738} 02/25/2022 23:32:45 - INFO - codeparrot_training - Step 35502: {'lr': 0.00010435397668515007, 'samples': 18177536, 'steps': 35502, 'loss/train': 1.9201580286026} 02/25/2022 23:32:48 - INFO - codeparrot_training - Step 35503: {'lr': 0.00010434067807202391, 'samples': 18178048, 'steps': 35503, 'loss/train': 2.1618285179138184} 02/25/2022 23:32:54 - INFO - codeparrot_training - Step 35504: {'lr': 0.00010432738008285602, 'samples': 18178560, 'steps': 35504, 'loss/train': 1.8175995349884033} 02/25/2022 23:32:59 - INFO - codeparrot_training - Step 35505: {'lr': 0.00010431408271770338, 'samples': 18179072, 'steps': 35505, 'loss/train': 2.1174609661102295} 02/25/2022 23:33:03 - INFO - codeparrot_training - Step 35506: {'lr': 0.00010430078597662284, 'samples': 18179584, 'steps': 35506, 'loss/train': 2.102395534515381} 02/25/2022 23:33:08 - INFO - codeparrot_training - Step 35507: {'lr': 0.0001042874898596716, 'samples': 18180096, 'steps': 35507, 'loss/train': 1.876882553100586} 02/25/2022 23:33:12 - INFO - codeparrot_training - Step 35508: {'lr': 0.00010427419436690644, 'samples': 18180608, 'steps': 35508, 'loss/train': 1.7557848691940308} 02/25/2022 23:33:15 - INFO - codeparrot_training - Step 35509: {'lr': 0.00010426089949838438, 'samples': 18181120, 'steps': 35509, 'loss/train': 1.5969310998916626} 02/25/2022 23:33:21 - INFO - codeparrot_training - Step 35510: {'lr': 0.00010424760525416227, 'samples': 18181632, 'steps': 35510, 'loss/train': 2.2206850051879883} 02/25/2022 23:33:25 - INFO - codeparrot_training - Step 35511: {'lr': 0.0001042343116342972, 'samples': 18182144, 'steps': 35511, 'loss/train': 2.0760226249694824} 02/25/2022 23:33:31 - INFO - codeparrot_training - Step 35512: {'lr': 0.00010422101863884598, 'samples': 18182656, 'steps': 35512, 'loss/train': 1.0528029203414917} 02/25/2022 23:33:35 - INFO - codeparrot_training - Step 35513: {'lr': 0.00010420772626786582, 'samples': 18183168, 'steps': 35513, 'loss/train': 2.6579325199127197} 02/25/2022 23:33:40 - INFO - codeparrot_training - Step 35514: {'lr': 0.00010419443452141325, 'samples': 18183680, 'steps': 35514, 'loss/train': 1.6120131015777588} 02/25/2022 23:33:44 - INFO - codeparrot_training - Step 35515: {'lr': 0.00010418114339954554, 'samples': 18184192, 'steps': 35515, 'loss/train': 1.038468599319458} 02/25/2022 23:33:49 - INFO - codeparrot_training - Step 35516: {'lr': 0.00010416785290231951, 'samples': 18184704, 'steps': 35516, 'loss/train': 1.4201387166976929} 02/25/2022 23:33:52 - INFO - codeparrot_training - Step 35517: {'lr': 0.00010415456302979209, 'samples': 18185216, 'steps': 35517, 'loss/train': 2.7407658100128174} 02/25/2022 23:33:58 - INFO - codeparrot_training - Step 35518: {'lr': 0.00010414127378202009, 'samples': 18185728, 'steps': 35518, 'loss/train': 3.2500479221343994} 02/25/2022 23:34:01 - INFO - codeparrot_training - Step 35519: {'lr': 0.0001041279851590607, 'samples': 18186240, 'steps': 35519, 'loss/train': 0.6206653714179993} 02/25/2022 23:34:07 - INFO - codeparrot_training - Step 35520: {'lr': 0.00010411469716097066, 'samples': 18186752, 'steps': 35520, 'loss/train': 1.6798497438430786} 02/25/2022 23:34:10 - INFO - codeparrot_training - Step 35521: {'lr': 0.00010410140978780685, 'samples': 18187264, 'steps': 35521, 'loss/train': 1.5994287729263306} 02/25/2022 23:34:17 - INFO - codeparrot_training - Step 35522: {'lr': 0.00010408812303962639, 'samples': 18187776, 'steps': 35522, 'loss/train': 1.5160788297653198} 02/25/2022 23:34:20 - INFO - codeparrot_training - Step 35523: {'lr': 0.00010407483691648606, 'samples': 18188288, 'steps': 35523, 'loss/train': 0.4246375858783722} 02/25/2022 23:34:26 - INFO - codeparrot_training - Step 35524: {'lr': 0.00010406155141844279, 'samples': 18188800, 'steps': 35524, 'loss/train': 2.1877810955047607} 02/25/2022 23:34:29 - INFO - codeparrot_training - Step 35525: {'lr': 0.00010404826654555341, 'samples': 18189312, 'steps': 35525, 'loss/train': 0.7963857054710388} 02/25/2022 23:34:35 - INFO - codeparrot_training - Step 35526: {'lr': 0.00010403498229787497, 'samples': 18189824, 'steps': 35526, 'loss/train': 1.8368951082229614} 02/25/2022 23:34:38 - INFO - codeparrot_training - Step 35527: {'lr': 0.00010402169867546435, 'samples': 18190336, 'steps': 35527, 'loss/train': 1.940693736076355} 02/25/2022 23:34:44 - INFO - codeparrot_training - Step 35528: {'lr': 0.00010400841567837843, 'samples': 18190848, 'steps': 35528, 'loss/train': 1.2978272438049316} 02/25/2022 23:34:47 - INFO - codeparrot_training - Step 35529: {'lr': 0.00010399513330667398, 'samples': 18191360, 'steps': 35529, 'loss/train': 0.6375097632408142} 02/25/2022 23:34:52 - INFO - codeparrot_training - Step 35530: {'lr': 0.00010398185156040812, 'samples': 18191872, 'steps': 35530, 'loss/train': 2.410601854324341} 02/25/2022 23:34:56 - INFO - codeparrot_training - Step 35531: {'lr': 0.00010396857043963764, 'samples': 18192384, 'steps': 35531, 'loss/train': 2.1480660438537598} 02/25/2022 23:35:03 - INFO - codeparrot_training - Step 35532: {'lr': 0.00010395528994441944, 'samples': 18192896, 'steps': 35532, 'loss/train': 1.7839181423187256} 02/25/2022 23:35:07 - INFO - codeparrot_training - Step 35533: {'lr': 0.0001039420100748103, 'samples': 18193408, 'steps': 35533, 'loss/train': 0.6810885071754456} 02/25/2022 23:35:12 - INFO - codeparrot_training - Step 35534: {'lr': 0.00010392873083086731, 'samples': 18193920, 'steps': 35534, 'loss/train': 3.1297543048858643} 02/25/2022 23:35:16 - INFO - codeparrot_training - Step 35535: {'lr': 0.00010391545221264725, 'samples': 18194432, 'steps': 35535, 'loss/train': 0.7706844806671143} 02/25/2022 23:35:21 - INFO - codeparrot_training - Step 35536: {'lr': 0.00010390217422020701, 'samples': 18194944, 'steps': 35536, 'loss/train': 1.6830925941467285} 02/25/2022 23:35:25 - INFO - codeparrot_training - Step 35537: {'lr': 0.00010388889685360339, 'samples': 18195456, 'steps': 35537, 'loss/train': 0.2152283489704132} 02/25/2022 23:35:30 - INFO - codeparrot_training - Step 35538: {'lr': 0.00010387562011289343, 'samples': 18195968, 'steps': 35538, 'loss/train': 1.922654628753662} 02/25/2022 23:35:34 - INFO - codeparrot_training - Step 35539: {'lr': 0.00010386234399813388, 'samples': 18196480, 'steps': 35539, 'loss/train': 1.8588982820510864} 02/25/2022 23:35:39 - INFO - codeparrot_training - Step 35540: {'lr': 0.00010384906850938167, 'samples': 18196992, 'steps': 35540, 'loss/train': 2.6272544860839844} 02/25/2022 23:35:43 - INFO - codeparrot_training - Step 35541: {'lr': 0.00010383579364669354, 'samples': 18197504, 'steps': 35541, 'loss/train': 2.1056222915649414} 02/25/2022 23:35:49 - INFO - codeparrot_training - Step 35542: {'lr': 0.00010382251941012655, 'samples': 18198016, 'steps': 35542, 'loss/train': 1.777288794517517} 02/25/2022 23:35:52 - INFO - codeparrot_training - Step 35543: {'lr': 0.00010380924579973747, 'samples': 18198528, 'steps': 35543, 'loss/train': 1.0001877546310425} 02/25/2022 23:35:58 - INFO - codeparrot_training - Step 35544: {'lr': 0.00010379597281558314, 'samples': 18199040, 'steps': 35544, 'loss/train': 1.594598650932312} 02/25/2022 23:36:01 - INFO - codeparrot_training - Step 35545: {'lr': 0.0001037827004577204, 'samples': 18199552, 'steps': 35545, 'loss/train': 2.4374799728393555} 02/25/2022 23:36:07 - INFO - codeparrot_training - Step 35546: {'lr': 0.00010376942872620618, 'samples': 18200064, 'steps': 35546, 'loss/train': 1.3352670669555664} 02/25/2022 23:36:10 - INFO - codeparrot_training - Step 35547: {'lr': 0.00010375615762109725, 'samples': 18200576, 'steps': 35547, 'loss/train': 2.118468999862671} 02/25/2022 23:36:16 - INFO - codeparrot_training - Step 35548: {'lr': 0.00010374288714245064, 'samples': 18201088, 'steps': 35548, 'loss/train': 1.0579332113265991} 02/25/2022 23:36:19 - INFO - codeparrot_training - Step 35549: {'lr': 0.0001037296172903229, 'samples': 18201600, 'steps': 35549, 'loss/train': 2.0325934886932373} 02/25/2022 23:36:25 - INFO - codeparrot_training - Step 35550: {'lr': 0.00010371634806477112, 'samples': 18202112, 'steps': 35550, 'loss/train': 1.336356282234192} 02/25/2022 23:36:28 - INFO - codeparrot_training - Step 35551: {'lr': 0.00010370307946585197, 'samples': 18202624, 'steps': 35551, 'loss/train': 1.1712565422058105} 02/25/2022 23:36:34 - INFO - codeparrot_training - Step 35552: {'lr': 0.00010368981149362256, 'samples': 18203136, 'steps': 35552, 'loss/train': 2.8321943283081055} 02/25/2022 23:36:37 - INFO - codeparrot_training - Step 35553: {'lr': 0.00010367654414813935, 'samples': 18203648, 'steps': 35553, 'loss/train': 1.9090839624404907} 02/25/2022 23:36:43 - INFO - codeparrot_training - Step 35554: {'lr': 0.00010366327742945944, 'samples': 18204160, 'steps': 35554, 'loss/train': 0.8347397446632385} 02/25/2022 23:36:46 - INFO - codeparrot_training - Step 35555: {'lr': 0.00010365001133763948, 'samples': 18204672, 'steps': 35555, 'loss/train': 1.8070425987243652} 02/25/2022 23:36:52 - INFO - codeparrot_training - Step 35556: {'lr': 0.00010363674587273662, 'samples': 18205184, 'steps': 35556, 'loss/train': 1.78842294216156} 02/25/2022 23:36:55 - INFO - codeparrot_training - Step 35557: {'lr': 0.00010362348103480726, 'samples': 18205696, 'steps': 35557, 'loss/train': 0.8642957210540771} 02/25/2022 23:37:01 - INFO - codeparrot_training - Step 35558: {'lr': 0.00010361021682390854, 'samples': 18206208, 'steps': 35558, 'loss/train': 1.9542944431304932} 02/25/2022 23:37:05 - INFO - codeparrot_training - Step 35559: {'lr': 0.00010359695324009705, 'samples': 18206720, 'steps': 35559, 'loss/train': 0.6702874898910522} 02/25/2022 23:37:10 - INFO - codeparrot_training - Step 35560: {'lr': 0.00010358369028342985, 'samples': 18207232, 'steps': 35560, 'loss/train': 1.0588432550430298} 02/25/2022 23:37:14 - INFO - codeparrot_training - Step 35561: {'lr': 0.0001035704279539636, 'samples': 18207744, 'steps': 35561, 'loss/train': 1.6165931224822998} 02/25/2022 23:37:19 - INFO - codeparrot_training - Step 35562: {'lr': 0.00010355716625175513, 'samples': 18208256, 'steps': 35562, 'loss/train': 1.4471553564071655} 02/25/2022 23:37:23 - INFO - codeparrot_training - Step 35563: {'lr': 0.00010354390517686118, 'samples': 18208768, 'steps': 35563, 'loss/train': 2.2126753330230713} 02/25/2022 23:37:29 - INFO - codeparrot_training - Step 35564: {'lr': 0.00010353064472933873, 'samples': 18209280, 'steps': 35564, 'loss/train': 2.165555238723755} 02/25/2022 23:37:32 - INFO - codeparrot_training - Step 35565: {'lr': 0.00010351738490924451, 'samples': 18209792, 'steps': 35565, 'loss/train': 2.593285322189331} 02/25/2022 23:37:37 - INFO - codeparrot_training - Step 35566: {'lr': 0.00010350412571663525, 'samples': 18210304, 'steps': 35566, 'loss/train': 1.0473862886428833} 02/25/2022 23:37:41 - INFO - codeparrot_training - Step 35567: {'lr': 0.00010349086715156784, 'samples': 18210816, 'steps': 35567, 'loss/train': 1.6462459564208984} 02/25/2022 23:37:47 - INFO - codeparrot_training - Step 35568: {'lr': 0.00010347760921409894, 'samples': 18211328, 'steps': 35568, 'loss/train': 2.1878182888031006} 02/25/2022 23:37:50 - INFO - codeparrot_training - Step 35569: {'lr': 0.0001034643519042855, 'samples': 18211840, 'steps': 35569, 'loss/train': 1.4180275201797485} 02/25/2022 23:37:56 - INFO - codeparrot_training - Step 35570: {'lr': 0.00010345109522218426, 'samples': 18212352, 'steps': 35570, 'loss/train': 1.5729774236679077} 02/25/2022 23:37:59 - INFO - codeparrot_training - Step 35571: {'lr': 0.000103437839167852, 'samples': 18212864, 'steps': 35571, 'loss/train': 1.1597007513046265} 02/25/2022 23:38:05 - INFO - codeparrot_training - Step 35572: {'lr': 0.00010342458374134539, 'samples': 18213376, 'steps': 35572, 'loss/train': 2.199892520904541} 02/25/2022 23:38:08 - INFO - codeparrot_training - Step 35573: {'lr': 0.00010341132894272143, 'samples': 18213888, 'steps': 35573, 'loss/train': 1.2314226627349854} 02/25/2022 23:38:14 - INFO - codeparrot_training - Step 35574: {'lr': 0.00010339807477203678, 'samples': 18214400, 'steps': 35574, 'loss/train': 2.3173766136169434} 02/25/2022 23:38:17 - INFO - codeparrot_training - Step 35575: {'lr': 0.00010338482122934822, 'samples': 18214912, 'steps': 35575, 'loss/train': 1.3393115997314453} 02/25/2022 23:38:23 - INFO - codeparrot_training - Step 35576: {'lr': 0.00010337156831471245, 'samples': 18215424, 'steps': 35576, 'loss/train': 2.509542465209961} 02/25/2022 23:38:26 - INFO - codeparrot_training - Step 35577: {'lr': 0.0001033583160281864, 'samples': 18215936, 'steps': 35577, 'loss/train': 1.648756742477417} 02/25/2022 23:38:33 - INFO - codeparrot_training - Step 35578: {'lr': 0.00010334506436982679, 'samples': 18216448, 'steps': 35578, 'loss/train': 1.7115774154663086} 02/25/2022 23:38:37 - INFO - codeparrot_training - Step 35579: {'lr': 0.00010333181333969033, 'samples': 18216960, 'steps': 35579, 'loss/train': 1.1724892854690552} 02/25/2022 23:38:42 - INFO - codeparrot_training - Step 35580: {'lr': 0.00010331856293783373, 'samples': 18217472, 'steps': 35580, 'loss/train': 1.849133014678955} 02/25/2022 23:38:46 - INFO - codeparrot_training - Step 35581: {'lr': 0.00010330531316431391, 'samples': 18217984, 'steps': 35581, 'loss/train': 2.23974609375} 02/25/2022 23:38:51 - INFO - codeparrot_training - Step 35582: {'lr': 0.00010329206401918748, 'samples': 18218496, 'steps': 35582, 'loss/train': 1.8942471742630005} 02/25/2022 23:38:54 - INFO - codeparrot_training - Step 35583: {'lr': 0.00010327881550251142, 'samples': 18219008, 'steps': 35583, 'loss/train': 1.1225790977478027} 02/25/2022 23:39:00 - INFO - codeparrot_training - Step 35584: {'lr': 0.00010326556761434213, 'samples': 18219520, 'steps': 35584, 'loss/train': 2.266474723815918} 02/25/2022 23:39:03 - INFO - codeparrot_training - Step 35585: {'lr': 0.0001032523203547367, 'samples': 18220032, 'steps': 35585, 'loss/train': 1.83694326877594} 02/25/2022 23:39:09 - INFO - codeparrot_training - Step 35586: {'lr': 0.00010323907372375161, 'samples': 18220544, 'steps': 35586, 'loss/train': 1.8051525354385376} 02/25/2022 23:39:13 - INFO - codeparrot_training - Step 35587: {'lr': 0.00010322582772144388, 'samples': 18221056, 'steps': 35587, 'loss/train': 1.1442192792892456} 02/25/2022 23:39:18 - INFO - codeparrot_training - Step 35588: {'lr': 0.00010321258234786996, 'samples': 18221568, 'steps': 35588, 'loss/train': 1.355179786682129} 02/25/2022 23:39:22 - INFO - codeparrot_training - Step 35589: {'lr': 0.00010319933760308677, 'samples': 18222080, 'steps': 35589, 'loss/train': 0.8855079412460327} 02/25/2022 23:39:28 - INFO - codeparrot_training - Step 35590: {'lr': 0.00010318609348715096, 'samples': 18222592, 'steps': 35590, 'loss/train': 1.1857198476791382} 02/25/2022 23:39:31 - INFO - codeparrot_training - Step 35591: {'lr': 0.00010317285000011948, 'samples': 18223104, 'steps': 35591, 'loss/train': 2.095787286758423} 02/25/2022 23:39:37 - INFO - codeparrot_training - Step 35592: {'lr': 0.00010315960714204869, 'samples': 18223616, 'steps': 35592, 'loss/train': 1.4601294994354248} 02/25/2022 23:39:40 - INFO - codeparrot_training - Step 35593: {'lr': 0.00010314636491299562, 'samples': 18224128, 'steps': 35593, 'loss/train': 1.440686583518982} 02/25/2022 23:39:46 - INFO - codeparrot_training - Step 35594: {'lr': 0.00010313312331301677, 'samples': 18224640, 'steps': 35594, 'loss/train': 2.279529571533203} 02/25/2022 23:39:51 - INFO - codeparrot_training - Step 35595: {'lr': 0.00010311988234216912, 'samples': 18225152, 'steps': 35595, 'loss/train': 1.548601508140564} 02/25/2022 23:39:55 - INFO - codeparrot_training - Step 35596: {'lr': 0.00010310664200050923, 'samples': 18225664, 'steps': 35596, 'loss/train': 1.823628544807434} 02/25/2022 23:40:00 - INFO - codeparrot_training - Step 35597: {'lr': 0.00010309340228809383, 'samples': 18226176, 'steps': 35597, 'loss/train': 1.5251606702804565} 02/25/2022 23:40:04 - INFO - codeparrot_training - Step 35598: {'lr': 0.0001030801632049796, 'samples': 18226688, 'steps': 35598, 'loss/train': 1.4135679006576538} 02/25/2022 23:40:09 - INFO - codeparrot_training - Step 35599: {'lr': 0.00010306692475122334, 'samples': 18227200, 'steps': 35599, 'loss/train': 2.0083534717559814} 02/25/2022 23:40:13 - INFO - codeparrot_training - Step 35600: {'lr': 0.00010305368692688174, 'samples': 18227712, 'steps': 35600, 'loss/train': 0.426821768283844} 02/25/2022 23:40:18 - INFO - codeparrot_training - Step 35601: {'lr': 0.00010304044973201148, 'samples': 18228224, 'steps': 35601, 'loss/train': 1.6690462827682495} 02/25/2022 23:40:22 - INFO - codeparrot_training - Step 35602: {'lr': 0.00010302721316666919, 'samples': 18228736, 'steps': 35602, 'loss/train': 2.3260042667388916} 02/25/2022 23:40:28 - INFO - codeparrot_training - Step 35603: {'lr': 0.00010301397723091171, 'samples': 18229248, 'steps': 35603, 'loss/train': 1.3810107707977295} 02/25/2022 23:40:31 - INFO - codeparrot_training - Step 35604: {'lr': 0.00010300074192479573, 'samples': 18229760, 'steps': 35604, 'loss/train': 1.0180336236953735} 02/25/2022 23:40:37 - INFO - codeparrot_training - Step 35605: {'lr': 0.00010298750724837786, 'samples': 18230272, 'steps': 35605, 'loss/train': 2.1794779300689697} 02/25/2022 23:40:40 - INFO - codeparrot_training - Step 35606: {'lr': 0.0001029742732017147, 'samples': 18230784, 'steps': 35606, 'loss/train': 4.077032089233398} 02/25/2022 23:40:46 - INFO - codeparrot_training - Step 35607: {'lr': 0.00010296103978486321, 'samples': 18231296, 'steps': 35607, 'loss/train': 1.3326951265335083} 02/25/2022 23:40:49 - INFO - codeparrot_training - Step 35608: {'lr': 0.00010294780699787993, 'samples': 18231808, 'steps': 35608, 'loss/train': 2.435358762741089} 02/25/2022 23:40:55 - INFO - codeparrot_training - Step 35609: {'lr': 0.00010293457484082155, 'samples': 18232320, 'steps': 35609, 'loss/train': 2.8295412063598633} 02/25/2022 23:40:58 - INFO - codeparrot_training - Step 35610: {'lr': 0.00010292134331374465, 'samples': 18232832, 'steps': 35610, 'loss/train': 1.5999478101730347} 02/25/2022 23:41:04 - INFO - codeparrot_training - Step 35611: {'lr': 0.00010290811241670611, 'samples': 18233344, 'steps': 35611, 'loss/train': 1.3185818195343018} 02/25/2022 23:41:07 - INFO - codeparrot_training - Step 35612: {'lr': 0.0001028948821497625, 'samples': 18233856, 'steps': 35612, 'loss/train': 2.8693010807037354} 02/25/2022 23:41:13 - INFO - codeparrot_training - Step 35613: {'lr': 0.00010288165251297052, 'samples': 18234368, 'steps': 35613, 'loss/train': 2.090181350708008} 02/25/2022 23:41:16 - INFO - codeparrot_training - Step 35614: {'lr': 0.00010286842350638681, 'samples': 18234880, 'steps': 35614, 'loss/train': 0.7473689317703247} 02/25/2022 23:41:22 - INFO - codeparrot_training - Step 35615: {'lr': 0.00010285519513006796, 'samples': 18235392, 'steps': 35615, 'loss/train': 1.6938201189041138} 02/25/2022 23:41:26 - INFO - codeparrot_training - Step 35616: {'lr': 0.00010284196738407084, 'samples': 18235904, 'steps': 35616, 'loss/train': 1.5739587545394897} 02/25/2022 23:41:31 - INFO - codeparrot_training - Step 35617: {'lr': 0.00010282874026845191, 'samples': 18236416, 'steps': 35617, 'loss/train': 1.7888811826705933} 02/25/2022 23:41:35 - INFO - codeparrot_training - Step 35618: {'lr': 0.00010281551378326811, 'samples': 18236928, 'steps': 35618, 'loss/train': 0.931207001209259} 02/25/2022 23:41:40 - INFO - codeparrot_training - Step 35619: {'lr': 0.00010280228792857571, 'samples': 18237440, 'steps': 35619, 'loss/train': 2.0260813236236572} 02/25/2022 23:41:44 - INFO - codeparrot_training - Step 35620: {'lr': 0.00010278906270443168, 'samples': 18237952, 'steps': 35620, 'loss/train': 1.224518060684204} 02/25/2022 23:41:49 - INFO - codeparrot_training - Step 35621: {'lr': 0.00010277583811089247, 'samples': 18238464, 'steps': 35621, 'loss/train': 2.755197286605835} 02/25/2022 23:41:53 - INFO - codeparrot_training - Step 35622: {'lr': 0.000102762614148015, 'samples': 18238976, 'steps': 35622, 'loss/train': 1.1081726551055908} 02/25/2022 23:41:58 - INFO - codeparrot_training - Step 35623: {'lr': 0.00010274939081585552, 'samples': 18239488, 'steps': 35623, 'loss/train': 1.8059111833572388} 02/25/2022 23:42:02 - INFO - codeparrot_training - Step 35624: {'lr': 0.00010273616811447103, 'samples': 18240000, 'steps': 35624, 'loss/train': 0.6258696913719177} 02/25/2022 23:42:08 - INFO - codeparrot_training - Step 35625: {'lr': 0.00010272294604391791, 'samples': 18240512, 'steps': 35625, 'loss/train': 1.4776450395584106} 02/25/2022 23:42:11 - INFO - codeparrot_training - Step 35626: {'lr': 0.00010270972460425313, 'samples': 18241024, 'steps': 35626, 'loss/train': 1.4091289043426514} 02/25/2022 23:42:17 - INFO - codeparrot_training - Step 35627: {'lr': 0.00010269650379553294, 'samples': 18241536, 'steps': 35627, 'loss/train': 2.6466026306152344} 02/25/2022 23:42:20 - INFO - codeparrot_training - Step 35628: {'lr': 0.00010268328361781423, 'samples': 18242048, 'steps': 35628, 'loss/train': 1.0453681945800781} 02/25/2022 23:42:26 - INFO - codeparrot_training - Step 35629: {'lr': 0.00010267006407115346, 'samples': 18242560, 'steps': 35629, 'loss/train': 2.5418148040771484} 02/25/2022 23:42:30 - INFO - codeparrot_training - Step 35630: {'lr': 0.00010265684515560745, 'samples': 18243072, 'steps': 35630, 'loss/train': 8.621676445007324} 02/25/2022 23:42:33 - INFO - codeparrot_training - Step 35631: {'lr': 0.00010264362687123274, 'samples': 18243584, 'steps': 35631, 'loss/train': 3.1544809341430664} 02/25/2022 23:42:39 - INFO - codeparrot_training - Step 35632: {'lr': 0.00010263040921808592, 'samples': 18244096, 'steps': 35632, 'loss/train': 1.3971437215805054} 02/25/2022 23:42:44 - INFO - codeparrot_training - Step 35633: {'lr': 0.00010261719219622356, 'samples': 18244608, 'steps': 35633, 'loss/train': 1.9479739665985107} 02/25/2022 23:42:48 - INFO - codeparrot_training - Step 35634: {'lr': 0.00010260397580570244, 'samples': 18245120, 'steps': 35634, 'loss/train': 1.7317043542861938} 02/25/2022 23:42:54 - INFO - codeparrot_training - Step 35635: {'lr': 0.00010259076004657905, 'samples': 18245632, 'steps': 35635, 'loss/train': 2.0960140228271484} 02/25/2022 23:42:57 - INFO - codeparrot_training - Step 35636: {'lr': 0.00010257754491891009, 'samples': 18246144, 'steps': 35636, 'loss/train': 1.5212658643722534} 02/25/2022 23:43:03 - INFO - codeparrot_training - Step 35637: {'lr': 0.00010256433042275199, 'samples': 18246656, 'steps': 35637, 'loss/train': 2.1963188648223877} 02/25/2022 23:43:06 - INFO - codeparrot_training - Step 35638: {'lr': 0.00010255111655816157, 'samples': 18247168, 'steps': 35638, 'loss/train': 2.380225419998169} 02/25/2022 23:43:12 - INFO - codeparrot_training - Step 35639: {'lr': 0.00010253790332519536, 'samples': 18247680, 'steps': 35639, 'loss/train': 1.555729627609253} 02/25/2022 23:43:16 - INFO - codeparrot_training - Step 35640: {'lr': 0.00010252469072390994, 'samples': 18248192, 'steps': 35640, 'loss/train': 2.052687406539917} 02/25/2022 23:43:21 - INFO - codeparrot_training - Step 35641: {'lr': 0.00010251147875436185, 'samples': 18248704, 'steps': 35641, 'loss/train': 1.7617690563201904} 02/25/2022 23:43:25 - INFO - codeparrot_training - Step 35642: {'lr': 0.00010249826741660781, 'samples': 18249216, 'steps': 35642, 'loss/train': 2.667482852935791} 02/25/2022 23:43:30 - INFO - codeparrot_training - Step 35643: {'lr': 0.00010248505671070427, 'samples': 18249728, 'steps': 35643, 'loss/train': 1.1361576318740845} 02/25/2022 23:43:34 - INFO - codeparrot_training - Step 35644: {'lr': 0.00010247184663670812, 'samples': 18250240, 'steps': 35644, 'loss/train': 2.1625943183898926} 02/25/2022 23:43:39 - INFO - codeparrot_training - Step 35645: {'lr': 0.00010245863719467552, 'samples': 18250752, 'steps': 35645, 'loss/train': 1.7517374753952026} 02/25/2022 23:43:43 - INFO - codeparrot_training - Step 35646: {'lr': 0.00010244542838466337, 'samples': 18251264, 'steps': 35646, 'loss/train': 1.5311287641525269} 02/25/2022 23:43:48 - INFO - codeparrot_training - Step 35647: {'lr': 0.00010243222020672812, 'samples': 18251776, 'steps': 35647, 'loss/train': 2.803532600402832} 02/25/2022 23:43:52 - INFO - codeparrot_training - Step 35648: {'lr': 0.00010241901266092644, 'samples': 18252288, 'steps': 35648, 'loss/train': 2.301795482635498} 02/25/2022 23:43:57 - INFO - codeparrot_training - Step 35649: {'lr': 0.0001024058057473147, 'samples': 18252800, 'steps': 35649, 'loss/train': 1.5233286619186401} 02/25/2022 23:44:00 - INFO - codeparrot_training - Step 35650: {'lr': 0.00010239259946594976, 'samples': 18253312, 'steps': 35650, 'loss/train': 2.147953748703003} 02/25/2022 23:44:07 - INFO - codeparrot_training - Step 35651: {'lr': 0.00010237939381688801, 'samples': 18253824, 'steps': 35651, 'loss/train': 0.2026338279247284} 02/25/2022 23:44:10 - INFO - codeparrot_training - Step 35652: {'lr': 0.00010236618880018609, 'samples': 18254336, 'steps': 35652, 'loss/train': 1.5804471969604492} 02/25/2022 23:44:16 - INFO - codeparrot_training - Step 35653: {'lr': 0.00010235298441590044, 'samples': 18254848, 'steps': 35653, 'loss/train': 2.2368381023406982} 02/25/2022 23:44:19 - INFO - codeparrot_training - Step 35654: {'lr': 0.0001023397806640878, 'samples': 18255360, 'steps': 35654, 'loss/train': 1.8397973775863647} 02/25/2022 23:44:25 - INFO - codeparrot_training - Step 35655: {'lr': 0.00010232657754480467, 'samples': 18255872, 'steps': 35655, 'loss/train': 2.2479805946350098} 02/25/2022 23:44:28 - INFO - codeparrot_training - Step 35656: {'lr': 0.0001023133750581075, 'samples': 18256384, 'steps': 35656, 'loss/train': 1.7291384935379028} 02/25/2022 23:44:34 - INFO - codeparrot_training - Step 35657: {'lr': 0.00010230017320405302, 'samples': 18256896, 'steps': 35657, 'loss/train': 0.8899371027946472} 02/25/2022 23:44:37 - INFO - codeparrot_training - Step 35658: {'lr': 0.00010228697198269771, 'samples': 18257408, 'steps': 35658, 'loss/train': 0.680962860584259} 02/25/2022 23:44:43 - INFO - codeparrot_training - Step 35659: {'lr': 0.00010227377139409808, 'samples': 18257920, 'steps': 35659, 'loss/train': 2.230807304382324} 02/25/2022 23:44:47 - INFO - codeparrot_training - Step 35660: {'lr': 0.00010226057143831064, 'samples': 18258432, 'steps': 35660, 'loss/train': 2.2269339561462402} 02/25/2022 23:44:53 - INFO - codeparrot_training - Step 35661: {'lr': 0.00010224737211539217, 'samples': 18258944, 'steps': 35661, 'loss/train': 1.005497932434082} 02/25/2022 23:44:56 - INFO - codeparrot_training - Step 35662: {'lr': 0.00010223417342539884, 'samples': 18259456, 'steps': 35662, 'loss/train': 1.2220706939697266} 02/25/2022 23:45:02 - INFO - codeparrot_training - Step 35663: {'lr': 0.0001022209753683875, 'samples': 18259968, 'steps': 35663, 'loss/train': 1.8131171464920044} 02/25/2022 23:45:06 - INFO - codeparrot_training - Step 35664: {'lr': 0.00010220777794441449, 'samples': 18260480, 'steps': 35664, 'loss/train': 0.9779194593429565} 02/25/2022 23:45:11 - INFO - codeparrot_training - Step 35665: {'lr': 0.00010219458115353661, 'samples': 18260992, 'steps': 35665, 'loss/train': 2.191406011581421} 02/25/2022 23:45:15 - INFO - codeparrot_training - Step 35666: {'lr': 0.00010218138499581003, 'samples': 18261504, 'steps': 35666, 'loss/train': 1.5816563367843628} 02/25/2022 23:45:20 - INFO - codeparrot_training - Step 35667: {'lr': 0.00010216818947129155, 'samples': 18262016, 'steps': 35667, 'loss/train': 1.7287938594818115} 02/25/2022 23:45:24 - INFO - codeparrot_training - Step 35668: {'lr': 0.00010215499458003752, 'samples': 18262528, 'steps': 35668, 'loss/train': 2.4390480518341064} 02/25/2022 23:45:29 - INFO - codeparrot_training - Step 35669: {'lr': 0.00010214180032210463, 'samples': 18263040, 'steps': 35669, 'loss/train': 1.2931214570999146} 02/25/2022 23:45:33 - INFO - codeparrot_training - Step 35670: {'lr': 0.0001021286066975493, 'samples': 18263552, 'steps': 35670, 'loss/train': 1.8427566289901733} 02/25/2022 23:45:39 - INFO - codeparrot_training - Step 35671: {'lr': 0.0001021154137064281, 'samples': 18264064, 'steps': 35671, 'loss/train': 1.9134794473648071} 02/25/2022 23:45:42 - INFO - codeparrot_training - Step 35672: {'lr': 0.0001021022213487974, 'samples': 18264576, 'steps': 35672, 'loss/train': 1.5787392854690552} 02/25/2022 23:45:48 - INFO - codeparrot_training - Step 35673: {'lr': 0.00010208902962471391, 'samples': 18265088, 'steps': 35673, 'loss/train': 1.4939777851104736} 02/25/2022 23:45:51 - INFO - codeparrot_training - Step 35674: {'lr': 0.00010207583853423405, 'samples': 18265600, 'steps': 35674, 'loss/train': 1.661348819732666} 02/25/2022 23:45:57 - INFO - codeparrot_training - Step 35675: {'lr': 0.00010206264807741434, 'samples': 18266112, 'steps': 35675, 'loss/train': 1.9807336330413818} 02/25/2022 23:46:00 - INFO - codeparrot_training - Step 35676: {'lr': 0.00010204945825431114, 'samples': 18266624, 'steps': 35676, 'loss/train': 3.0842208862304688} 02/25/2022 23:46:06 - INFO - codeparrot_training - Step 35677: {'lr': 0.0001020362690649812, 'samples': 18267136, 'steps': 35677, 'loss/train': 1.785537838935852} 02/25/2022 23:46:10 - INFO - codeparrot_training - Step 35678: {'lr': 0.0001020230805094808, 'samples': 18267648, 'steps': 35678, 'loss/train': 1.5328142642974854} 02/25/2022 23:46:15 - INFO - codeparrot_training - Step 35679: {'lr': 0.00010200989258786674, 'samples': 18268160, 'steps': 35679, 'loss/train': 1.5087634325027466} 02/25/2022 23:46:18 - INFO - codeparrot_training - Step 35680: {'lr': 0.00010199670530019511, 'samples': 18268672, 'steps': 35680, 'loss/train': 1.4962003231048584} 02/25/2022 23:46:25 - INFO - codeparrot_training - Step 35681: {'lr': 0.00010198351864652269, 'samples': 18269184, 'steps': 35681, 'loss/train': 2.130740165710449} 02/25/2022 23:46:28 - INFO - codeparrot_training - Step 35682: {'lr': 0.00010197033262690577, 'samples': 18269696, 'steps': 35682, 'loss/train': 2.107067108154297} 02/25/2022 23:46:34 - INFO - codeparrot_training - Step 35683: {'lr': 0.00010195714724140115, 'samples': 18270208, 'steps': 35683, 'loss/train': 1.1449522972106934} 02/25/2022 23:46:37 - INFO - codeparrot_training - Step 35684: {'lr': 0.00010194396249006491, 'samples': 18270720, 'steps': 35684, 'loss/train': 2.2388737201690674} 02/25/2022 23:46:43 - INFO - codeparrot_training - Step 35685: {'lr': 0.00010193077837295381, 'samples': 18271232, 'steps': 35685, 'loss/train': 2.305042266845703} 02/25/2022 23:46:46 - INFO - codeparrot_training - Step 35686: {'lr': 0.00010191759489012415, 'samples': 18271744, 'steps': 35686, 'loss/train': 0.840194582939148} 02/25/2022 23:46:52 - INFO - codeparrot_training - Step 35687: {'lr': 0.00010190441204163267, 'samples': 18272256, 'steps': 35687, 'loss/train': 1.3436988592147827} 02/25/2022 23:46:55 - INFO - codeparrot_training - Step 35688: {'lr': 0.00010189122982753549, 'samples': 18272768, 'steps': 35688, 'loss/train': 0.8906335234642029} 02/25/2022 23:47:01 - INFO - codeparrot_training - Step 35689: {'lr': 0.00010187804824788932, 'samples': 18273280, 'steps': 35689, 'loss/train': 1.6520715951919556} 02/25/2022 23:47:04 - INFO - codeparrot_training - Step 35690: {'lr': 0.00010186486730275049, 'samples': 18273792, 'steps': 35690, 'loss/train': 1.9635087251663208} 02/25/2022 23:47:10 - INFO - codeparrot_training - Step 35691: {'lr': 0.00010185168699217561, 'samples': 18274304, 'steps': 35691, 'loss/train': 2.116649866104126} 02/25/2022 23:47:13 - INFO - codeparrot_training - Step 35692: {'lr': 0.00010183850731622108, 'samples': 18274816, 'steps': 35692, 'loss/train': 0.07796463370323181} 02/25/2022 23:47:19 - INFO - codeparrot_training - Step 35693: {'lr': 0.00010182532827494332, 'samples': 18275328, 'steps': 35693, 'loss/train': 0.04188638925552368} 02/25/2022 23:47:22 - INFO - codeparrot_training - Step 35694: {'lr': 0.00010181214986839879, 'samples': 18275840, 'steps': 35694, 'loss/train': 1.5360817909240723} 02/25/2022 23:47:28 - INFO - codeparrot_training - Step 35695: {'lr': 0.00010179897209664388, 'samples': 18276352, 'steps': 35695, 'loss/train': 2.1865015029907227} 02/25/2022 23:47:31 - INFO - codeparrot_training - Step 35696: {'lr': 0.0001017857949597352, 'samples': 18276864, 'steps': 35696, 'loss/train': 1.3727922439575195} 02/25/2022 23:47:38 - INFO - codeparrot_training - Step 35697: {'lr': 0.00010177261845772912, 'samples': 18277376, 'steps': 35697, 'loss/train': 0.8442731499671936} 02/25/2022 23:47:41 - INFO - codeparrot_training - Step 35698: {'lr': 0.00010175944259068207, 'samples': 18277888, 'steps': 35698, 'loss/train': 2.0390079021453857} 02/25/2022 23:47:47 - INFO - codeparrot_training - Step 35699: {'lr': 0.0001017462673586504, 'samples': 18278400, 'steps': 35699, 'loss/train': 2.4970924854278564} 02/25/2022 23:47:50 - INFO - codeparrot_training - Step 35700: {'lr': 0.00010173309276169074, 'samples': 18278912, 'steps': 35700, 'loss/train': 1.4240550994873047} 02/25/2022 23:47:56 - INFO - codeparrot_training - Step 35701: {'lr': 0.00010171991879985942, 'samples': 18279424, 'steps': 35701, 'loss/train': 0.8915626406669617} 02/25/2022 23:47:59 - INFO - codeparrot_training - Step 35702: {'lr': 0.00010170674547321291, 'samples': 18279936, 'steps': 35702, 'loss/train': 1.6054563522338867} 02/25/2022 23:48:05 - INFO - codeparrot_training - Step 35703: {'lr': 0.00010169357278180749, 'samples': 18280448, 'steps': 35703, 'loss/train': 2.487030029296875} 02/25/2022 23:48:08 - INFO - codeparrot_training - Step 35704: {'lr': 0.00010168040072569984, 'samples': 18280960, 'steps': 35704, 'loss/train': 1.916651725769043} 02/25/2022 23:48:14 - INFO - codeparrot_training - Step 35705: {'lr': 0.00010166722930494623, 'samples': 18281472, 'steps': 35705, 'loss/train': 1.5042604207992554} 02/25/2022 23:48:17 - INFO - codeparrot_training - Step 35706: {'lr': 0.0001016540585196031, 'samples': 18281984, 'steps': 35706, 'loss/train': 0.22606606781482697} 02/25/2022 23:48:23 - INFO - codeparrot_training - Step 35707: {'lr': 0.00010164088836972682, 'samples': 18282496, 'steps': 35707, 'loss/train': 1.6904124021530151} 02/25/2022 23:48:27 - INFO - codeparrot_training - Step 35708: {'lr': 0.00010162771885537392, 'samples': 18283008, 'steps': 35708, 'loss/train': 1.1760166883468628} 02/25/2022 23:48:32 - INFO - codeparrot_training - Step 35709: {'lr': 0.00010161454997660081, 'samples': 18283520, 'steps': 35709, 'loss/train': 1.9237961769104004} 02/25/2022 23:48:36 - INFO - codeparrot_training - Step 35710: {'lr': 0.00010160138173346379, 'samples': 18284032, 'steps': 35710, 'loss/train': 1.5310009717941284} 02/25/2022 23:48:41 - INFO - codeparrot_training - Step 35711: {'lr': 0.00010158821412601926, 'samples': 18284544, 'steps': 35711, 'loss/train': 1.596812129020691} 02/25/2022 23:48:45 - INFO - codeparrot_training - Step 35712: {'lr': 0.00010157504715432381, 'samples': 18285056, 'steps': 35712, 'loss/train': 2.7541723251342773} 02/25/2022 23:48:50 - INFO - codeparrot_training - Step 35713: {'lr': 0.0001015618808184337, 'samples': 18285568, 'steps': 35713, 'loss/train': 2.4386332035064697} 02/25/2022 23:48:54 - INFO - codeparrot_training - Step 35714: {'lr': 0.00010154871511840535, 'samples': 18286080, 'steps': 35714, 'loss/train': 1.5850285291671753} 02/25/2022 23:48:59 - INFO - codeparrot_training - Step 35715: {'lr': 0.0001015355500542951, 'samples': 18286592, 'steps': 35715, 'loss/train': 1.33933687210083} 02/25/2022 23:49:03 - INFO - codeparrot_training - Step 35716: {'lr': 0.00010152238562615951, 'samples': 18287104, 'steps': 35716, 'loss/train': 1.891749382019043} 02/25/2022 23:49:09 - INFO - codeparrot_training - Step 35717: {'lr': 0.00010150922183405475, 'samples': 18287616, 'steps': 35717, 'loss/train': 1.4227796792984009} 02/25/2022 23:49:12 - INFO - codeparrot_training - Step 35718: {'lr': 0.00010149605867803755, 'samples': 18288128, 'steps': 35718, 'loss/train': 1.137722373008728} 02/25/2022 23:49:18 - INFO - codeparrot_training - Step 35719: {'lr': 0.00010148289615816386, 'samples': 18288640, 'steps': 35719, 'loss/train': 0.9836340546607971} 02/25/2022 23:49:21 - INFO - codeparrot_training - Step 35720: {'lr': 0.00010146973427449039, 'samples': 18289152, 'steps': 35720, 'loss/train': 2.1329128742218018} 02/25/2022 23:49:27 - INFO - codeparrot_training - Step 35721: {'lr': 0.00010145657302707334, 'samples': 18289664, 'steps': 35721, 'loss/train': 2.545548439025879} 02/25/2022 23:49:30 - INFO - codeparrot_training - Step 35722: {'lr': 0.00010144341241596933, 'samples': 18290176, 'steps': 35722, 'loss/train': 1.2810003757476807} 02/25/2022 23:49:36 - INFO - codeparrot_training - Step 35723: {'lr': 0.00010143025244123439, 'samples': 18290688, 'steps': 35723, 'loss/train': 0.42229577898979187} 02/25/2022 23:49:39 - INFO - codeparrot_training - Step 35724: {'lr': 0.00010141709310292516, 'samples': 18291200, 'steps': 35724, 'loss/train': 1.2862200736999512} 02/25/2022 23:49:44 - INFO - codeparrot_training - Step 35725: {'lr': 0.00010140393440109783, 'samples': 18291712, 'steps': 35725, 'loss/train': 0.7427252531051636} 02/25/2022 23:49:48 - INFO - codeparrot_training - Step 35726: {'lr': 0.00010139077633580896, 'samples': 18292224, 'steps': 35726, 'loss/train': 1.8521952629089355} 02/25/2022 23:49:53 - INFO - codeparrot_training - Step 35727: {'lr': 0.00010137761890711481, 'samples': 18292736, 'steps': 35727, 'loss/train': 1.9859143495559692} 02/25/2022 23:49:57 - INFO - codeparrot_training - Step 35728: {'lr': 0.00010136446211507175, 'samples': 18293248, 'steps': 35728, 'loss/train': 2.207202672958374} 02/25/2022 23:50:02 - INFO - codeparrot_training - Step 35729: {'lr': 0.00010135130595973602, 'samples': 18293760, 'steps': 35729, 'loss/train': 2.231048345565796} 02/25/2022 23:50:06 - INFO - codeparrot_training - Step 35730: {'lr': 0.0001013381504411642, 'samples': 18294272, 'steps': 35730, 'loss/train': 1.4255168437957764} 02/25/2022 23:50:11 - INFO - codeparrot_training - Step 35731: {'lr': 0.00010132499555941257, 'samples': 18294784, 'steps': 35731, 'loss/train': 3.725872278213501} 02/25/2022 23:50:15 - INFO - codeparrot_training - Step 35732: {'lr': 0.00010131184131453741, 'samples': 18295296, 'steps': 35732, 'loss/train': 2.2639212608337402} 02/25/2022 23:50:21 - INFO - codeparrot_training - Step 35733: {'lr': 0.00010129868770659503, 'samples': 18295808, 'steps': 35733, 'loss/train': 2.052513837814331} 02/25/2022 23:50:24 - INFO - codeparrot_training - Step 35734: {'lr': 0.00010128553473564197, 'samples': 18296320, 'steps': 35734, 'loss/train': 0.46537500619888306} 02/25/2022 23:50:30 - INFO - codeparrot_training - Step 35735: {'lr': 0.00010127238240173443, 'samples': 18296832, 'steps': 35735, 'loss/train': 2.5970702171325684} 02/25/2022 23:50:35 - INFO - codeparrot_training - Step 35736: {'lr': 0.0001012592307049288, 'samples': 18297344, 'steps': 35736, 'loss/train': 2.2687795162200928} 02/25/2022 23:50:39 - INFO - codeparrot_training - Step 35737: {'lr': 0.00010124607964528127, 'samples': 18297856, 'steps': 35737, 'loss/train': 1.0605558156967163} 02/25/2022 23:50:44 - INFO - codeparrot_training - Step 35738: {'lr': 0.0001012329292228484, 'samples': 18298368, 'steps': 35738, 'loss/train': 2.531243324279785} 02/25/2022 23:50:48 - INFO - codeparrot_training - Step 35739: {'lr': 0.00010121977943768643, 'samples': 18298880, 'steps': 35739, 'loss/train': 1.646443486213684} 02/25/2022 23:50:53 - INFO - codeparrot_training - Step 35740: {'lr': 0.00010120663028985169, 'samples': 18299392, 'steps': 35740, 'loss/train': 0.739367663860321} 02/25/2022 23:50:57 - INFO - codeparrot_training - Step 35741: {'lr': 0.00010119348177940049, 'samples': 18299904, 'steps': 35741, 'loss/train': 1.4047653675079346} 02/25/2022 23:51:04 - INFO - codeparrot_training - Step 35742: {'lr': 0.00010118033390638906, 'samples': 18300416, 'steps': 35742, 'loss/train': 2.459845781326294} 02/25/2022 23:51:07 - INFO - codeparrot_training - Step 35743: {'lr': 0.00010116718667087394, 'samples': 18300928, 'steps': 35743, 'loss/train': 2.2868740558624268} 02/25/2022 23:51:11 - INFO - codeparrot_training - Step 35744: {'lr': 0.00010115404007291131, 'samples': 18301440, 'steps': 35744, 'loss/train': 1.2523095607757568} 02/25/2022 23:51:16 - INFO - codeparrot_training - Step 35745: {'lr': 0.00010114089411255754, 'samples': 18301952, 'steps': 35745, 'loss/train': 1.5059741735458374} 02/25/2022 23:51:20 - INFO - codeparrot_training - Step 35746: {'lr': 0.00010112774878986877, 'samples': 18302464, 'steps': 35746, 'loss/train': 1.524960994720459} 02/25/2022 23:51:25 - INFO - codeparrot_training - Step 35747: {'lr': 0.00010111460410490158, 'samples': 18302976, 'steps': 35747, 'loss/train': 0.3220663368701935} 02/25/2022 23:51:29 - INFO - codeparrot_training - Step 35748: {'lr': 0.00010110146005771215, 'samples': 18303488, 'steps': 35748, 'loss/train': 1.0911425352096558} 02/25/2022 23:51:34 - INFO - codeparrot_training - Step 35749: {'lr': 0.00010108831664835676, 'samples': 18304000, 'steps': 35749, 'loss/train': 0.6062576770782471} 02/25/2022 23:51:38 - INFO - codeparrot_training - Step 35750: {'lr': 0.00010107517387689166, 'samples': 18304512, 'steps': 35750, 'loss/train': 0.12423809617757797} 02/25/2022 23:51:43 - INFO - codeparrot_training - Step 35751: {'lr': 0.00010106203174337334, 'samples': 18305024, 'steps': 35751, 'loss/train': 1.6128947734832764} 02/25/2022 23:51:47 - INFO - codeparrot_training - Step 35752: {'lr': 0.00010104889024785785, 'samples': 18305536, 'steps': 35752, 'loss/train': 3.344428777694702} 02/25/2022 23:51:53 - INFO - codeparrot_training - Step 35753: {'lr': 0.00010103574939040184, 'samples': 18306048, 'steps': 35753, 'loss/train': 1.7074198722839355} 02/25/2022 23:51:56 - INFO - codeparrot_training - Step 35754: {'lr': 0.00010102260917106115, 'samples': 18306560, 'steps': 35754, 'loss/train': 2.4637317657470703} 02/25/2022 23:52:02 - INFO - codeparrot_training - Step 35755: {'lr': 0.00010100946958989244, 'samples': 18307072, 'steps': 35755, 'loss/train': 1.056294322013855} 02/25/2022 23:52:05 - INFO - codeparrot_training - Step 35756: {'lr': 0.0001009963306469517, 'samples': 18307584, 'steps': 35756, 'loss/train': 0.06776978075504303} 02/25/2022 23:52:11 - INFO - codeparrot_training - Step 35757: {'lr': 0.0001009831923422956, 'samples': 18308096, 'steps': 35757, 'loss/train': 1.934425950050354} 02/25/2022 23:52:14 - INFO - codeparrot_training - Step 35758: {'lr': 0.00010097005467598, 'samples': 18308608, 'steps': 35758, 'loss/train': 1.629753828048706} 02/25/2022 23:52:20 - INFO - codeparrot_training - Step 35759: {'lr': 0.00010095691764806142, 'samples': 18309120, 'steps': 35759, 'loss/train': 1.4768569469451904} 02/25/2022 23:52:23 - INFO - codeparrot_training - Step 35760: {'lr': 0.00010094378125859602, 'samples': 18309632, 'steps': 35760, 'loss/train': 1.0320724248886108} 02/25/2022 23:52:29 - INFO - codeparrot_training - Step 35761: {'lr': 0.00010093064550764031, 'samples': 18310144, 'steps': 35761, 'loss/train': 1.607743501663208} 02/25/2022 23:52:32 - INFO - codeparrot_training - Step 35762: {'lr': 0.00010091751039525019, 'samples': 18310656, 'steps': 35762, 'loss/train': 1.2424111366271973} 02/25/2022 23:52:38 - INFO - codeparrot_training - Step 35763: {'lr': 0.00010090437592148218, 'samples': 18311168, 'steps': 35763, 'loss/train': 2.506679058074951} 02/25/2022 23:52:41 - INFO - codeparrot_training - Step 35764: {'lr': 0.00010089124208639242, 'samples': 18311680, 'steps': 35764, 'loss/train': 1.5359445810317993} 02/25/2022 23:52:48 - INFO - codeparrot_training - Step 35765: {'lr': 0.00010087810889003732, 'samples': 18312192, 'steps': 35765, 'loss/train': 1.8538669347763062} 02/25/2022 23:52:51 - INFO - codeparrot_training - Step 35766: {'lr': 0.00010086497633247304, 'samples': 18312704, 'steps': 35766, 'loss/train': 1.8985295295715332} 02/25/2022 23:52:57 - INFO - codeparrot_training - Step 35767: {'lr': 0.00010085184441375586, 'samples': 18313216, 'steps': 35767, 'loss/train': 1.0780740976333618} 02/25/2022 23:53:00 - INFO - codeparrot_training - Step 35768: {'lr': 0.00010083871313394191, 'samples': 18313728, 'steps': 35768, 'loss/train': 1.8520212173461914} 02/25/2022 23:53:06 - INFO - codeparrot_training - Step 35769: {'lr': 0.00010082558249308766, 'samples': 18314240, 'steps': 35769, 'loss/train': 1.1098905801773071} 02/25/2022 23:53:09 - INFO - codeparrot_training - Step 35770: {'lr': 0.00010081245249124923, 'samples': 18314752, 'steps': 35770, 'loss/train': 1.4739210605621338} 02/25/2022 23:53:15 - INFO - codeparrot_training - Step 35771: {'lr': 0.0001007993231284829, 'samples': 18315264, 'steps': 35771, 'loss/train': 1.1698055267333984} 02/25/2022 23:53:18 - INFO - codeparrot_training - Step 35772: {'lr': 0.0001007861944048448, 'samples': 18315776, 'steps': 35772, 'loss/train': 2.637181520462036} 02/25/2022 23:53:24 - INFO - codeparrot_training - Step 35773: {'lr': 0.00010077306632039132, 'samples': 18316288, 'steps': 35773, 'loss/train': 2.039930820465088} 02/25/2022 23:53:27 - INFO - codeparrot_training - Step 35774: {'lr': 0.00010075993887517868, 'samples': 18316800, 'steps': 35774, 'loss/train': 1.1873440742492676} 02/25/2022 23:53:33 - INFO - codeparrot_training - Step 35775: {'lr': 0.00010074681206926306, 'samples': 18317312, 'steps': 35775, 'loss/train': 2.5722475051879883} 02/25/2022 23:53:36 - INFO - codeparrot_training - Step 35776: {'lr': 0.0001007336859027006, 'samples': 18317824, 'steps': 35776, 'loss/train': 2.2421562671661377} 02/25/2022 23:53:42 - INFO - codeparrot_training - Step 35777: {'lr': 0.00010072056037554772, 'samples': 18318336, 'steps': 35777, 'loss/train': 1.755372166633606} 02/25/2022 23:53:45 - INFO - codeparrot_training - Step 35778: {'lr': 0.00010070743548786057, 'samples': 18318848, 'steps': 35778, 'loss/train': 0.9578810334205627} 02/25/2022 23:53:52 - INFO - codeparrot_training - Step 35779: {'lr': 0.00010069431123969536, 'samples': 18319360, 'steps': 35779, 'loss/train': 2.17842698097229} 02/25/2022 23:53:55 - INFO - codeparrot_training - Step 35780: {'lr': 0.00010068118763110824, 'samples': 18319872, 'steps': 35780, 'loss/train': 3.354618549346924} 02/25/2022 23:54:01 - INFO - codeparrot_training - Step 35781: {'lr': 0.00010066806466215556, 'samples': 18320384, 'steps': 35781, 'loss/train': 1.9873430728912354} 02/25/2022 23:54:04 - INFO - codeparrot_training - Step 35782: {'lr': 0.0001006549423328935, 'samples': 18320896, 'steps': 35782, 'loss/train': 0.951806902885437} 02/25/2022 23:54:10 - INFO - codeparrot_training - Step 35783: {'lr': 0.0001006418206433782, 'samples': 18321408, 'steps': 35783, 'loss/train': 2.1334714889526367} 02/25/2022 23:54:13 - INFO - codeparrot_training - Step 35784: {'lr': 0.00010062869959366586, 'samples': 18321920, 'steps': 35784, 'loss/train': 0.36228030920028687} 02/25/2022 23:54:19 - INFO - codeparrot_training - Step 35785: {'lr': 0.0001006155791838128, 'samples': 18322432, 'steps': 35785, 'loss/train': 1.8500031232833862} 02/25/2022 23:54:22 - INFO - codeparrot_training - Step 35786: {'lr': 0.0001006024594138752, 'samples': 18322944, 'steps': 35786, 'loss/train': 2.4462132453918457} 02/25/2022 23:54:28 - INFO - codeparrot_training - Step 35787: {'lr': 0.00010058934028390912, 'samples': 18323456, 'steps': 35787, 'loss/train': 2.0870611667633057} 02/25/2022 23:54:31 - INFO - codeparrot_training - Step 35788: {'lr': 0.00010057622179397103, 'samples': 18323968, 'steps': 35788, 'loss/train': 1.8753844499588013} 02/25/2022 23:54:37 - INFO - codeparrot_training - Step 35789: {'lr': 0.0001005631039441168, 'samples': 18324480, 'steps': 35789, 'loss/train': 1.2309633493423462} 02/25/2022 23:54:43 - INFO - codeparrot_training - Step 35790: {'lr': 0.00010054998673440286, 'samples': 18324992, 'steps': 35790, 'loss/train': 1.9031633138656616} 02/25/2022 23:54:47 - INFO - codeparrot_training - Step 35791: {'lr': 0.00010053687016488522, 'samples': 18325504, 'steps': 35791, 'loss/train': 1.7304120063781738} 02/25/2022 23:54:52 - INFO - codeparrot_training - Step 35792: {'lr': 0.00010052375423562038, 'samples': 18326016, 'steps': 35792, 'loss/train': 1.9291925430297852} 02/25/2022 23:54:56 - INFO - codeparrot_training - Step 35793: {'lr': 0.00010051063894666409, 'samples': 18326528, 'steps': 35793, 'loss/train': 3.257054090499878} 02/25/2022 23:55:01 - INFO - codeparrot_training - Step 35794: {'lr': 0.00010049752429807288, 'samples': 18327040, 'steps': 35794, 'loss/train': 1.846071481704712} 02/25/2022 23:55:05 - INFO - codeparrot_training - Step 35795: {'lr': 0.00010048441028990272, 'samples': 18327552, 'steps': 35795, 'loss/train': 1.2694448232650757} 02/25/2022 23:55:10 - INFO - codeparrot_training - Step 35796: {'lr': 0.00010047129692221003, 'samples': 18328064, 'steps': 35796, 'loss/train': 2.1736929416656494} 02/25/2022 23:55:14 - INFO - codeparrot_training - Step 35797: {'lr': 0.00010045818419505062, 'samples': 18328576, 'steps': 35797, 'loss/train': 1.857295274734497} 02/25/2022 23:55:19 - INFO - codeparrot_training - Step 35798: {'lr': 0.00010044507210848097, 'samples': 18329088, 'steps': 35798, 'loss/train': 2.3312344551086426} 02/25/2022 23:55:23 - INFO - codeparrot_training - Step 35799: {'lr': 0.00010043196066255709, 'samples': 18329600, 'steps': 35799, 'loss/train': 1.7168599367141724} 02/25/2022 23:55:29 - INFO - codeparrot_training - Step 35800: {'lr': 0.00010041884985733524, 'samples': 18330112, 'steps': 35800, 'loss/train': 1.7620187997817993} 02/25/2022 23:55:33 - INFO - codeparrot_training - Step 35801: {'lr': 0.00010040573969287155, 'samples': 18330624, 'steps': 35801, 'loss/train': 1.7488982677459717} 02/25/2022 23:55:38 - INFO - codeparrot_training - Step 35802: {'lr': 0.00010039263016922218, 'samples': 18331136, 'steps': 35802, 'loss/train': 2.3864548206329346} 02/25/2022 23:55:41 - INFO - codeparrot_training - Step 35803: {'lr': 0.0001003795212864432, 'samples': 18331648, 'steps': 35803, 'loss/train': 1.5186033248901367} 02/25/2022 23:55:47 - INFO - codeparrot_training - Step 35804: {'lr': 0.0001003664130445909, 'samples': 18332160, 'steps': 35804, 'loss/train': 0.7373254299163818} 02/25/2022 23:55:50 - INFO - codeparrot_training - Step 35805: {'lr': 0.0001003533054437214, 'samples': 18332672, 'steps': 35805, 'loss/train': 2.089855670928955} 02/25/2022 23:55:56 - INFO - codeparrot_training - Step 35806: {'lr': 0.00010034019848389081, 'samples': 18333184, 'steps': 35806, 'loss/train': 1.377484917640686} 02/25/2022 23:55:59 - INFO - codeparrot_training - Step 35807: {'lr': 0.00010032709216515518, 'samples': 18333696, 'steps': 35807, 'loss/train': 1.7374728918075562} 02/25/2022 23:56:05 - INFO - codeparrot_training - Step 35808: {'lr': 0.00010031398648757087, 'samples': 18334208, 'steps': 35808, 'loss/train': 2.625351667404175} 02/25/2022 23:56:08 - INFO - codeparrot_training - Step 35809: {'lr': 0.00010030088145119392, 'samples': 18334720, 'steps': 35809, 'loss/train': 2.052825689315796} 02/25/2022 23:56:15 - INFO - codeparrot_training - Step 35810: {'lr': 0.00010028777705608045, 'samples': 18335232, 'steps': 35810, 'loss/train': 1.3945361375808716} 02/25/2022 23:56:19 - INFO - codeparrot_training - Step 35811: {'lr': 0.0001002746733022865, 'samples': 18335744, 'steps': 35811, 'loss/train': 0.5266461372375488} 02/25/2022 23:56:24 - INFO - codeparrot_training - Step 35812: {'lr': 0.00010026157018986839, 'samples': 18336256, 'steps': 35812, 'loss/train': 1.0185898542404175} 02/25/2022 23:56:27 - INFO - codeparrot_training - Step 35813: {'lr': 0.0001002484677188821, 'samples': 18336768, 'steps': 35813, 'loss/train': 1.2480287551879883} 02/25/2022 23:56:33 - INFO - codeparrot_training - Step 35814: {'lr': 0.00010023536588938401, 'samples': 18337280, 'steps': 35814, 'loss/train': 1.4092193841934204} 02/25/2022 23:56:36 - INFO - codeparrot_training - Step 35815: {'lr': 0.00010022226470142986, 'samples': 18337792, 'steps': 35815, 'loss/train': 2.3569858074188232} 02/25/2022 23:56:42 - INFO - codeparrot_training - Step 35816: {'lr': 0.00010020916415507605, 'samples': 18338304, 'steps': 35816, 'loss/train': 2.0819053649902344} 02/25/2022 23:56:45 - INFO - codeparrot_training - Step 35817: {'lr': 0.00010019606425037855, 'samples': 18338816, 'steps': 35817, 'loss/train': 1.334726333618164} 02/25/2022 23:56:51 - INFO - codeparrot_training - Step 35818: {'lr': 0.00010018296498739371, 'samples': 18339328, 'steps': 35818, 'loss/train': 1.0104502439498901} 02/25/2022 23:56:54 - INFO - codeparrot_training - Step 35819: {'lr': 0.00010016986636617731, 'samples': 18339840, 'steps': 35819, 'loss/train': 0.9667519330978394} 02/25/2022 23:57:00 - INFO - codeparrot_training - Step 35820: {'lr': 0.0001001567683867857, 'samples': 18340352, 'steps': 35820, 'loss/train': 2.3153321743011475} 02/25/2022 23:57:04 - INFO - codeparrot_training - Step 35821: {'lr': 0.00010014367104927491, 'samples': 18340864, 'steps': 35821, 'loss/train': 1.448493480682373} 02/25/2022 23:57:09 - INFO - codeparrot_training - Step 35822: {'lr': 0.00010013057435370104, 'samples': 18341376, 'steps': 35822, 'loss/train': 1.8445961475372314} 02/25/2022 23:57:13 - INFO - codeparrot_training - Step 35823: {'lr': 0.00010011747830012014, 'samples': 18341888, 'steps': 35823, 'loss/train': 0.7104392647743225} 02/25/2022 23:57:18 - INFO - codeparrot_training - Step 35824: {'lr': 0.00010010438288858845, 'samples': 18342400, 'steps': 35824, 'loss/train': 1.7712794542312622} 02/25/2022 23:57:21 - INFO - codeparrot_training - Step 35825: {'lr': 0.00010009128811916197, 'samples': 18342912, 'steps': 35825, 'loss/train': 1.2480518817901611} 02/25/2022 23:57:28 - INFO - codeparrot_training - Step 35826: {'lr': 0.00010007819399189674, 'samples': 18343424, 'steps': 35826, 'loss/train': 1.5449738502502441} 02/25/2022 23:57:32 - INFO - codeparrot_training - Step 35827: {'lr': 0.000100065100506849, 'samples': 18343936, 'steps': 35827, 'loss/train': 1.6615954637527466} 02/25/2022 23:57:37 - INFO - codeparrot_training - Step 35828: {'lr': 0.00010005200766407476, 'samples': 18344448, 'steps': 35828, 'loss/train': 2.4104394912719727} 02/25/2022 23:57:41 - INFO - codeparrot_training - Step 35829: {'lr': 0.00010003891546363012, 'samples': 18344960, 'steps': 35829, 'loss/train': 1.866317868232727} 02/25/2022 23:57:46 - INFO - codeparrot_training - Step 35830: {'lr': 0.00010002582390557105, 'samples': 18345472, 'steps': 35830, 'loss/train': 1.8755161762237549} 02/25/2022 23:57:50 - INFO - codeparrot_training - Step 35831: {'lr': 0.00010001273298995381, 'samples': 18345984, 'steps': 35831, 'loss/train': 0.6883083581924438} 02/25/2022 23:57:55 - INFO - codeparrot_training - Step 35832: {'lr': 9.99996427168344e-05, 'samples': 18346496, 'steps': 35832, 'loss/train': 1.780527949333191} 02/25/2022 23:57:59 - INFO - codeparrot_training - Step 35833: {'lr': 9.998655308626888e-05, 'samples': 18347008, 'steps': 35833, 'loss/train': 2.493551254272461} 02/25/2022 23:58:04 - INFO - codeparrot_training - Step 35834: {'lr': 9.997346409831326e-05, 'samples': 18347520, 'steps': 35834, 'loss/train': 1.2785578966140747} 02/25/2022 23:58:08 - INFO - codeparrot_training - Step 35835: {'lr': 9.996037575302375e-05, 'samples': 18348032, 'steps': 35835, 'loss/train': 3.2638285160064697} 02/25/2022 23:58:14 - INFO - codeparrot_training - Step 35836: {'lr': 9.994728805045635e-05, 'samples': 18348544, 'steps': 35836, 'loss/train': 1.2130745649337769} 02/25/2022 23:58:18 - INFO - codeparrot_training - Step 35837: {'lr': 9.993420099066717e-05, 'samples': 18349056, 'steps': 35837, 'loss/train': 2.0023648738861084} 02/25/2022 23:58:23 - INFO - codeparrot_training - Step 35838: {'lr': 9.992111457371208e-05, 'samples': 18349568, 'steps': 35838, 'loss/train': 2.543809413909912} 02/25/2022 23:58:29 - INFO - codeparrot_training - Step 35839: {'lr': 9.990802879964739e-05, 'samples': 18350080, 'steps': 35839, 'loss/train': 2.232409954071045} 02/25/2022 23:58:32 - INFO - codeparrot_training - Step 35840: {'lr': 9.989494366852902e-05, 'samples': 18350592, 'steps': 35840, 'loss/train': 3.030867338180542} 02/25/2022 23:58:36 - INFO - codeparrot_training - Step 35841: {'lr': 9.988185918041307e-05, 'samples': 18351104, 'steps': 35841, 'loss/train': 0.17709724605083466} 02/25/2022 23:58:41 - INFO - codeparrot_training - Step 35842: {'lr': 9.986877533535546e-05, 'samples': 18351616, 'steps': 35842, 'loss/train': 0.7245816588401794} 02/25/2022 23:58:48 - INFO - codeparrot_training - Step 35843: {'lr': 9.985569213341247e-05, 'samples': 18352128, 'steps': 35843, 'loss/train': 1.4777520895004272} 02/25/2022 23:58:51 - INFO - codeparrot_training - Step 35844: {'lr': 9.984260957463998e-05, 'samples': 18352640, 'steps': 35844, 'loss/train': 0.8061791062355042} 02/25/2022 23:58:54 - INFO - codeparrot_training - Step 35845: {'lr': 9.98295276590941e-05, 'samples': 18353152, 'steps': 35845, 'loss/train': 1.0470699071884155} 02/25/2022 23:59:00 - INFO - codeparrot_training - Step 35846: {'lr': 9.981644638683075e-05, 'samples': 18353664, 'steps': 35846, 'loss/train': 0.7400323152542114} 02/25/2022 23:59:06 - INFO - codeparrot_training - Step 35847: {'lr': 9.98033657579061e-05, 'samples': 18354176, 'steps': 35847, 'loss/train': 1.6745328903198242} 02/25/2022 23:59:09 - INFO - codeparrot_training - Step 35848: {'lr': 9.979028577237609e-05, 'samples': 18354688, 'steps': 35848, 'loss/train': 0.5688425302505493} 02/25/2022 23:59:15 - INFO - codeparrot_training - Step 35849: {'lr': 9.977720643029697e-05, 'samples': 18355200, 'steps': 35849, 'loss/train': 1.9610005617141724} 02/25/2022 23:59:18 - INFO - codeparrot_training - Step 35850: {'lr': 9.976412773172441e-05, 'samples': 18355712, 'steps': 35850, 'loss/train': 1.489272952079773} 02/25/2022 23:59:24 - INFO - codeparrot_training - Step 35851: {'lr': 9.975104967671473e-05, 'samples': 18356224, 'steps': 35851, 'loss/train': 0.8180992603302002} 02/25/2022 23:59:27 - INFO - codeparrot_training - Step 35852: {'lr': 9.973797226532372e-05, 'samples': 18356736, 'steps': 35852, 'loss/train': 0.04300791025161743} 02/25/2022 23:59:31 - INFO - codeparrot_training - Step 35853: {'lr': 9.972489549760772e-05, 'samples': 18357248, 'steps': 35853, 'loss/train': 0.9210607409477234} 02/25/2022 23:59:36 - INFO - codeparrot_training - Step 35854: {'lr': 9.971181937362239e-05, 'samples': 18357760, 'steps': 35854, 'loss/train': 2.4391281604766846} 02/25/2022 23:59:42 - INFO - codeparrot_training - Step 35855: {'lr': 9.969874389342396e-05, 'samples': 18358272, 'steps': 35855, 'loss/train': 1.3115605115890503} 02/25/2022 23:59:46 - INFO - codeparrot_training - Step 35856: {'lr': 9.968566905706832e-05, 'samples': 18358784, 'steps': 35856, 'loss/train': 2.1106553077697754} 02/25/2022 23:59:51 - INFO - codeparrot_training - Step 35857: {'lr': 9.967259486461172e-05, 'samples': 18359296, 'steps': 35857, 'loss/train': 3.029512882232666} 02/25/2022 23:59:55 - INFO - codeparrot_training - Step 35858: {'lr': 9.96595213161098e-05, 'samples': 18359808, 'steps': 35858, 'loss/train': 1.8139848709106445} 02/26/2022 00:00:00 - INFO - codeparrot_training - Step 35859: {'lr': 9.964644841161885e-05, 'samples': 18360320, 'steps': 35859, 'loss/train': 2.35768461227417} 02/26/2022 00:00:04 - INFO - codeparrot_training - Step 35860: {'lr': 9.963337615119466e-05, 'samples': 18360832, 'steps': 35860, 'loss/train': 1.69540274143219} 02/26/2022 00:00:09 - INFO - codeparrot_training - Step 35861: {'lr': 9.962030453489346e-05, 'samples': 18361344, 'steps': 35861, 'loss/train': 1.393372654914856} 02/26/2022 00:00:13 - INFO - codeparrot_training - Step 35862: {'lr': 9.960723356277113e-05, 'samples': 18361856, 'steps': 35862, 'loss/train': 2.2338507175445557} 02/26/2022 00:00:18 - INFO - codeparrot_training - Step 35863: {'lr': 9.959416323488363e-05, 'samples': 18362368, 'steps': 35863, 'loss/train': 0.9011073708534241} 02/26/2022 00:00:22 - INFO - codeparrot_training - Step 35864: {'lr': 9.958109355128688e-05, 'samples': 18362880, 'steps': 35864, 'loss/train': 4.287140846252441} 02/26/2022 00:00:25 - INFO - codeparrot_training - Step 35865: {'lr': 9.956802451203709e-05, 'samples': 18363392, 'steps': 35865, 'loss/train': 1.5284981727600098} 02/26/2022 00:00:31 - INFO - codeparrot_training - Step 35866: {'lr': 9.955495611719009e-05, 'samples': 18363904, 'steps': 35866, 'loss/train': 3.0469791889190674} 02/26/2022 00:00:34 - INFO - codeparrot_training - Step 35867: {'lr': 9.954188836680189e-05, 'samples': 18364416, 'steps': 35867, 'loss/train': 2.0611648559570312} 02/26/2022 00:00:40 - INFO - codeparrot_training - Step 35868: {'lr': 9.952882126092835e-05, 'samples': 18364928, 'steps': 35868, 'loss/train': 2.296166181564331} 02/26/2022 00:00:44 - INFO - codeparrot_training - Step 35869: {'lr': 9.951575479962569e-05, 'samples': 18365440, 'steps': 35869, 'loss/train': 2.1618359088897705} 02/26/2022 00:00:49 - INFO - codeparrot_training - Step 35870: {'lr': 9.950268898294972e-05, 'samples': 18365952, 'steps': 35870, 'loss/train': 1.6320720911026} 02/26/2022 00:00:53 - INFO - codeparrot_training - Step 35871: {'lr': 9.948962381095647e-05, 'samples': 18366464, 'steps': 35871, 'loss/train': 1.590567946434021} 02/26/2022 00:01:00 - INFO - codeparrot_training - Step 35872: {'lr': 9.947655928370186e-05, 'samples': 18366976, 'steps': 35872, 'loss/train': 0.12426839768886566} 02/26/2022 00:01:03 - INFO - codeparrot_training - Step 35873: {'lr': 9.94634954012418e-05, 'samples': 18367488, 'steps': 35873, 'loss/train': 0.4463624954223633} 02/26/2022 00:01:09 - INFO - codeparrot_training - Step 35874: {'lr': 9.945043216363239e-05, 'samples': 18368000, 'steps': 35874, 'loss/train': 1.449934959411621} 02/26/2022 00:01:12 - INFO - codeparrot_training - Step 35875: {'lr': 9.943736957092955e-05, 'samples': 18368512, 'steps': 35875, 'loss/train': 1.1564255952835083} 02/26/2022 00:01:18 - INFO - codeparrot_training - Step 35876: {'lr': 9.942430762318919e-05, 'samples': 18369024, 'steps': 35876, 'loss/train': 1.4119863510131836} 02/26/2022 00:01:21 - INFO - codeparrot_training - Step 35877: {'lr': 9.941124632046722e-05, 'samples': 18369536, 'steps': 35877, 'loss/train': 1.3367893695831299} 02/26/2022 00:01:27 - INFO - codeparrot_training - Step 35878: {'lr': 9.939818566281971e-05, 'samples': 18370048, 'steps': 35878, 'loss/train': 1.0903595685958862} 02/26/2022 00:01:30 - INFO - codeparrot_training - Step 35879: {'lr': 9.938512565030258e-05, 'samples': 18370560, 'steps': 35879, 'loss/train': 2.6931049823760986} 02/26/2022 00:01:36 - INFO - codeparrot_training - Step 35880: {'lr': 9.937206628297172e-05, 'samples': 18371072, 'steps': 35880, 'loss/train': 1.529502272605896} 02/26/2022 00:01:39 - INFO - codeparrot_training - Step 35881: {'lr': 9.935900756088304e-05, 'samples': 18371584, 'steps': 35881, 'loss/train': 0.259307861328125} 02/26/2022 00:01:46 - INFO - codeparrot_training - Step 35882: {'lr': 9.934594948409259e-05, 'samples': 18372096, 'steps': 35882, 'loss/train': 1.7002257108688354} 02/26/2022 00:01:50 - INFO - codeparrot_training - Step 35883: {'lr': 9.933289205265627e-05, 'samples': 18372608, 'steps': 35883, 'loss/train': 0.8881520628929138} 02/26/2022 00:01:55 - INFO - codeparrot_training - Step 35884: {'lr': 9.931983526662999e-05, 'samples': 18373120, 'steps': 35884, 'loss/train': 0.8898929357528687} 02/26/2022 00:01:58 - INFO - codeparrot_training - Step 35885: {'lr': 9.930677912606961e-05, 'samples': 18373632, 'steps': 35885, 'loss/train': 1.0810331106185913} 02/26/2022 00:02:04 - INFO - codeparrot_training - Step 35886: {'lr': 9.929372363103124e-05, 'samples': 18374144, 'steps': 35886, 'loss/train': 2.11832594871521} 02/26/2022 00:02:07 - INFO - codeparrot_training - Step 35887: {'lr': 9.928066878157061e-05, 'samples': 18374656, 'steps': 35887, 'loss/train': 0.5292294025421143} 02/26/2022 00:02:13 - INFO - codeparrot_training - Step 35888: {'lr': 9.926761457774389e-05, 'samples': 18375168, 'steps': 35888, 'loss/train': 0.9705906510353088} 02/26/2022 00:02:16 - INFO - codeparrot_training - Step 35889: {'lr': 9.92545610196067e-05, 'samples': 18375680, 'steps': 35889, 'loss/train': 2.221918821334839} 02/26/2022 00:02:23 - INFO - codeparrot_training - Step 35890: {'lr': 9.924150810721516e-05, 'samples': 18376192, 'steps': 35890, 'loss/train': 2.068521499633789} 02/26/2022 00:02:26 - INFO - codeparrot_training - Step 35891: {'lr': 9.922845584062506e-05, 'samples': 18376704, 'steps': 35891, 'loss/train': 0.6108127236366272} 02/26/2022 00:02:32 - INFO - codeparrot_training - Step 35892: {'lr': 9.921540421989256e-05, 'samples': 18377216, 'steps': 35892, 'loss/train': 3.333068609237671} 02/26/2022 00:02:36 - INFO - codeparrot_training - Step 35893: {'lr': 9.920235324507321e-05, 'samples': 18377728, 'steps': 35893, 'loss/train': 1.1117558479309082} 02/26/2022 00:02:41 - INFO - codeparrot_training - Step 35894: {'lr': 9.918930291622317e-05, 'samples': 18378240, 'steps': 35894, 'loss/train': 1.6782842874526978} 02/26/2022 00:02:45 - INFO - codeparrot_training - Step 35895: {'lr': 9.917625323339819e-05, 'samples': 18378752, 'steps': 35895, 'loss/train': 2.5834484100341797} 02/26/2022 00:02:50 - INFO - codeparrot_training - Step 35896: {'lr': 9.916320419665434e-05, 'samples': 18379264, 'steps': 35896, 'loss/train': 2.003950834274292} 02/26/2022 00:02:53 - INFO - codeparrot_training - Step 35897: {'lr': 9.915015580604741e-05, 'samples': 18379776, 'steps': 35897, 'loss/train': 1.8916932344436646} 02/26/2022 00:02:59 - INFO - codeparrot_training - Step 35898: {'lr': 9.913710806163334e-05, 'samples': 18380288, 'steps': 35898, 'loss/train': 2.0987513065338135} 02/26/2022 00:03:03 - INFO - codeparrot_training - Step 35899: {'lr': 9.912406096346787e-05, 'samples': 18380800, 'steps': 35899, 'loss/train': 1.5130975246429443} 02/26/2022 00:03:08 - INFO - codeparrot_training - Step 35900: {'lr': 9.911101451160715e-05, 'samples': 18381312, 'steps': 35900, 'loss/train': 0.556807816028595} 02/26/2022 00:03:11 - INFO - codeparrot_training - Step 35901: {'lr': 9.909796870610691e-05, 'samples': 18381824, 'steps': 35901, 'loss/train': 0.40556639432907104} 02/26/2022 00:03:18 - INFO - codeparrot_training - Step 35902: {'lr': 9.908492354702303e-05, 'samples': 18382336, 'steps': 35902, 'loss/train': 1.4766087532043457} 02/26/2022 00:03:21 - INFO - codeparrot_training - Step 35903: {'lr': 9.907187903441137e-05, 'samples': 18382848, 'steps': 35903, 'loss/train': 0.6808434128761292} 02/26/2022 00:03:27 - INFO - codeparrot_training - Step 35904: {'lr': 9.905883516832793e-05, 'samples': 18383360, 'steps': 35904, 'loss/train': 2.0826196670532227} 02/26/2022 00:03:30 - INFO - codeparrot_training - Step 35905: {'lr': 9.904579194882848e-05, 'samples': 18383872, 'steps': 35905, 'loss/train': 2.017265558242798} 02/26/2022 00:03:36 - INFO - codeparrot_training - Step 35906: {'lr': 9.903274937596896e-05, 'samples': 18384384, 'steps': 35906, 'loss/train': 0.5058174729347229} 02/26/2022 00:03:39 - INFO - codeparrot_training - Step 35907: {'lr': 9.90197074498051e-05, 'samples': 18384896, 'steps': 35907, 'loss/train': 1.1410186290740967} 02/26/2022 00:03:45 - INFO - codeparrot_training - Step 35908: {'lr': 9.900666617039294e-05, 'samples': 18385408, 'steps': 35908, 'loss/train': 1.5514171123504639} 02/26/2022 00:03:48 - INFO - codeparrot_training - Step 35909: {'lr': 9.899362553778832e-05, 'samples': 18385920, 'steps': 35909, 'loss/train': 0.8423264026641846} 02/26/2022 00:03:53 - INFO - codeparrot_training - Step 35910: {'lr': 9.898058555204703e-05, 'samples': 18386432, 'steps': 35910, 'loss/train': 2.1960935592651367} 02/26/2022 00:03:57 - INFO - codeparrot_training - Step 35911: {'lr': 9.896754621322487e-05, 'samples': 18386944, 'steps': 35911, 'loss/train': 2.146303176879883} 02/26/2022 00:04:03 - INFO - codeparrot_training - Step 35912: {'lr': 9.895450752137788e-05, 'samples': 18387456, 'steps': 35912, 'loss/train': 1.890219807624817} 02/26/2022 00:04:06 - INFO - codeparrot_training - Step 35913: {'lr': 9.894146947656182e-05, 'samples': 18387968, 'steps': 35913, 'loss/train': 2.2780332565307617} 02/26/2022 00:04:12 - INFO - codeparrot_training - Step 35914: {'lr': 9.892843207883254e-05, 'samples': 18388480, 'steps': 35914, 'loss/train': 1.6994646787643433} 02/26/2022 00:04:15 - INFO - codeparrot_training - Step 35915: {'lr': 9.891539532824582e-05, 'samples': 18388992, 'steps': 35915, 'loss/train': 1.5538941621780396} 02/26/2022 00:04:20 - INFO - codeparrot_training - Step 35916: {'lr': 9.89023592248576e-05, 'samples': 18389504, 'steps': 35916, 'loss/train': 1.7740012407302856} 02/26/2022 00:04:27 - INFO - codeparrot_training - Step 35917: {'lr': 9.888932376872373e-05, 'samples': 18390016, 'steps': 35917, 'loss/train': 1.4152382612228394} 02/26/2022 00:04:31 - INFO - codeparrot_training - Step 35918: {'lr': 9.887628895990003e-05, 'samples': 18390528, 'steps': 35918, 'loss/train': 2.090456485748291} 02/26/2022 00:04:34 - INFO - codeparrot_training - Step 35919: {'lr': 9.886325479844229e-05, 'samples': 18391040, 'steps': 35919, 'loss/train': 0.2509082555770874} 02/26/2022 00:04:40 - INFO - codeparrot_training - Step 35920: {'lr': 9.885022128440629e-05, 'samples': 18391552, 'steps': 35920, 'loss/train': 1.359215497970581} 02/26/2022 00:04:43 - INFO - codeparrot_training - Step 35921: {'lr': 9.883718841784806e-05, 'samples': 18392064, 'steps': 35921, 'loss/train': 1.5009562969207764} 02/26/2022 00:04:49 - INFO - codeparrot_training - Step 35922: {'lr': 9.88241561988232e-05, 'samples': 18392576, 'steps': 35922, 'loss/train': 0.68912273645401} 02/26/2022 00:04:52 - INFO - codeparrot_training - Step 35923: {'lr': 9.881112462738784e-05, 'samples': 18393088, 'steps': 35923, 'loss/train': 1.9958263635635376} 02/26/2022 00:04:58 - INFO - codeparrot_training - Step 35924: {'lr': 9.879809370359744e-05, 'samples': 18393600, 'steps': 35924, 'loss/train': 2.1054980754852295} 02/26/2022 00:05:01 - INFO - codeparrot_training - Step 35925: {'lr': 9.878506342750806e-05, 'samples': 18394112, 'steps': 35925, 'loss/train': 0.9638531804084778} 02/26/2022 00:05:07 - INFO - codeparrot_training - Step 35926: {'lr': 9.877203379917541e-05, 'samples': 18394624, 'steps': 35926, 'loss/train': 2.6006088256835938} 02/26/2022 00:05:10 - INFO - codeparrot_training - Step 35927: {'lr': 9.875900481865549e-05, 'samples': 18395136, 'steps': 35927, 'loss/train': 1.6636947393417358} 02/26/2022 00:05:16 - INFO - codeparrot_training - Step 35928: {'lr': 9.874597648600378e-05, 'samples': 18395648, 'steps': 35928, 'loss/train': 0.9224361777305603} 02/26/2022 00:05:20 - INFO - codeparrot_training - Step 35929: {'lr': 9.87329488012764e-05, 'samples': 18396160, 'steps': 35929, 'loss/train': 1.7337384223937988} 02/26/2022 00:05:25 - INFO - codeparrot_training - Step 35930: {'lr': 9.871992176452894e-05, 'samples': 18396672, 'steps': 35930, 'loss/train': 1.2812590599060059} 02/26/2022 00:05:29 - INFO - codeparrot_training - Step 35931: {'lr': 9.870689537581745e-05, 'samples': 18397184, 'steps': 35931, 'loss/train': 1.6524097919464111} 02/26/2022 00:05:34 - INFO - codeparrot_training - Step 35932: {'lr': 9.869386963519742e-05, 'samples': 18397696, 'steps': 35932, 'loss/train': 1.4401382207870483} 02/26/2022 00:05:38 - INFO - codeparrot_training - Step 35933: {'lr': 9.868084454272488e-05, 'samples': 18398208, 'steps': 35933, 'loss/train': 1.5436817407608032} 02/26/2022 00:05:43 - INFO - codeparrot_training - Step 35934: {'lr': 9.866782009845549e-05, 'samples': 18398720, 'steps': 35934, 'loss/train': 2.2125649452209473} 02/26/2022 00:05:47 - INFO - codeparrot_training - Step 35935: {'lr': 9.86547963024452e-05, 'samples': 18399232, 'steps': 35935, 'loss/train': 1.5236769914627075} 02/26/2022 00:05:53 - INFO - codeparrot_training - Step 35936: {'lr': 9.864177315474967e-05, 'samples': 18399744, 'steps': 35936, 'loss/train': 1.952535629272461} 02/26/2022 00:05:56 - INFO - codeparrot_training - Step 35937: {'lr': 9.862875065542476e-05, 'samples': 18400256, 'steps': 35937, 'loss/train': 2.332662582397461} 02/26/2022 00:06:02 - INFO - codeparrot_training - Step 35938: {'lr': 9.86157288045261e-05, 'samples': 18400768, 'steps': 35938, 'loss/train': 1.490973711013794} 02/26/2022 00:06:06 - INFO - codeparrot_training - Step 35939: {'lr': 9.860270760210968e-05, 'samples': 18401280, 'steps': 35939, 'loss/train': 1.6946594715118408} 02/26/2022 00:06:11 - INFO - codeparrot_training - Step 35940: {'lr': 9.858968704823117e-05, 'samples': 18401792, 'steps': 35940, 'loss/train': 1.5306942462921143} 02/26/2022 00:06:15 - INFO - codeparrot_training - Step 35941: {'lr': 9.857666714294641e-05, 'samples': 18402304, 'steps': 35941, 'loss/train': 2.0050270557403564} 02/26/2022 00:06:20 - INFO - codeparrot_training - Step 35942: {'lr': 9.8563647886311e-05, 'samples': 18402816, 'steps': 35942, 'loss/train': 1.5935672521591187} 02/26/2022 00:06:23 - INFO - codeparrot_training - Step 35943: {'lr': 9.855062927838093e-05, 'samples': 18403328, 'steps': 35943, 'loss/train': 1.4004801511764526} 02/26/2022 00:06:29 - INFO - codeparrot_training - Step 35944: {'lr': 9.85376113192118e-05, 'samples': 18403840, 'steps': 35944, 'loss/train': 0.8524960875511169} 02/26/2022 00:06:33 - INFO - codeparrot_training - Step 35945: {'lr': 9.85245940088596e-05, 'samples': 18404352, 'steps': 35945, 'loss/train': 1.702237844467163} 02/26/2022 00:06:38 - INFO - codeparrot_training - Step 35946: {'lr': 9.851157734737978e-05, 'samples': 18404864, 'steps': 35946, 'loss/train': 2.018848180770874} 02/26/2022 00:06:42 - INFO - codeparrot_training - Step 35947: {'lr': 9.849856133482834e-05, 'samples': 18405376, 'steps': 35947, 'loss/train': 1.7983595132827759} 02/26/2022 00:06:47 - INFO - codeparrot_training - Step 35948: {'lr': 9.848554597126088e-05, 'samples': 18405888, 'steps': 35948, 'loss/train': 2.495162010192871} 02/26/2022 00:06:51 - INFO - codeparrot_training - Step 35949: {'lr': 9.847253125673342e-05, 'samples': 18406400, 'steps': 35949, 'loss/train': 1.4930616617202759} 02/26/2022 00:06:56 - INFO - codeparrot_training - Step 35950: {'lr': 9.84595171913013e-05, 'samples': 18406912, 'steps': 35950, 'loss/train': 1.4348050355911255} 02/26/2022 00:07:00 - INFO - codeparrot_training - Step 35951: {'lr': 9.844650377502063e-05, 'samples': 18407424, 'steps': 35951, 'loss/train': 1.898951768875122} 02/26/2022 00:07:05 - INFO - codeparrot_training - Step 35952: {'lr': 9.843349100794696e-05, 'samples': 18407936, 'steps': 35952, 'loss/train': 1.524706244468689} 02/26/2022 00:07:09 - INFO - codeparrot_training - Step 35953: {'lr': 9.842047889013613e-05, 'samples': 18408448, 'steps': 35953, 'loss/train': 1.0840580463409424} 02/26/2022 00:07:15 - INFO - codeparrot_training - Step 35954: {'lr': 9.840746742164372e-05, 'samples': 18408960, 'steps': 35954, 'loss/train': 2.1585886478424072} 02/26/2022 00:07:18 - INFO - codeparrot_training - Step 35955: {'lr': 9.83944566025257e-05, 'samples': 18409472, 'steps': 35955, 'loss/train': 2.505445718765259} 02/26/2022 00:07:24 - INFO - codeparrot_training - Step 35956: {'lr': 9.838144643283767e-05, 'samples': 18409984, 'steps': 35956, 'loss/train': 1.520660400390625} 02/26/2022 00:07:27 - INFO - codeparrot_training - Step 35957: {'lr': 9.836843691263528e-05, 'samples': 18410496, 'steps': 35957, 'loss/train': 2.3834221363067627} 02/26/2022 00:07:33 - INFO - codeparrot_training - Step 35958: {'lr': 9.835542804197448e-05, 'samples': 18411008, 'steps': 35958, 'loss/train': 2.029059648513794} 02/26/2022 00:07:36 - INFO - codeparrot_training - Step 35959: {'lr': 9.834241982091082e-05, 'samples': 18411520, 'steps': 35959, 'loss/train': 1.187527060508728} 02/26/2022 00:07:42 - INFO - codeparrot_training - Step 35960: {'lr': 9.832941224950012e-05, 'samples': 18412032, 'steps': 35960, 'loss/train': 1.8057677745819092} 02/26/2022 00:07:45 - INFO - codeparrot_training - Step 35961: {'lr': 9.831640532779795e-05, 'samples': 18412544, 'steps': 35961, 'loss/train': 2.1662979125976562} 02/26/2022 00:07:51 - INFO - codeparrot_training - Step 35962: {'lr': 9.830339905586025e-05, 'samples': 18413056, 'steps': 35962, 'loss/train': 2.6305973529815674} 02/26/2022 00:07:54 - INFO - codeparrot_training - Step 35963: {'lr': 9.82903934337426e-05, 'samples': 18413568, 'steps': 35963, 'loss/train': 1.7157042026519775} 02/26/2022 00:08:01 - INFO - codeparrot_training - Step 35964: {'lr': 9.827738846150073e-05, 'samples': 18414080, 'steps': 35964, 'loss/train': 1.8077280521392822} 02/26/2022 00:08:04 - INFO - codeparrot_training - Step 35965: {'lr': 9.826438413919026e-05, 'samples': 18414592, 'steps': 35965, 'loss/train': 1.522325873374939} 02/26/2022 00:08:10 - INFO - codeparrot_training - Step 35966: {'lr': 9.825138046686716e-05, 'samples': 18415104, 'steps': 35966, 'loss/train': 2.0576868057250977} 02/26/2022 00:08:13 - INFO - codeparrot_training - Step 35967: {'lr': 9.82383774445868e-05, 'samples': 18415616, 'steps': 35967, 'loss/train': 1.4208050966262817} 02/26/2022 00:08:19 - INFO - codeparrot_training - Step 35968: {'lr': 9.822537507240512e-05, 'samples': 18416128, 'steps': 35968, 'loss/train': 3.1818759441375732} 02/26/2022 00:08:22 - INFO - codeparrot_training - Step 35969: {'lr': 9.821237335037763e-05, 'samples': 18416640, 'steps': 35969, 'loss/train': 1.4404741525650024} 02/26/2022 00:08:28 - INFO - codeparrot_training - Step 35970: {'lr': 9.819937227856027e-05, 'samples': 18417152, 'steps': 35970, 'loss/train': 2.3471426963806152} 02/26/2022 00:08:31 - INFO - codeparrot_training - Step 35971: {'lr': 9.818637185700855e-05, 'samples': 18417664, 'steps': 35971, 'loss/train': 0.8370423316955566} 02/26/2022 00:08:37 - INFO - codeparrot_training - Step 35972: {'lr': 9.817337208577823e-05, 'samples': 18418176, 'steps': 35972, 'loss/train': 1.9797887802124023} 02/26/2022 00:08:40 - INFO - codeparrot_training - Step 35973: {'lr': 9.816037296492489e-05, 'samples': 18418688, 'steps': 35973, 'loss/train': 1.2125232219696045} 02/26/2022 00:08:46 - INFO - codeparrot_training - Step 35974: {'lr': 9.814737449450437e-05, 'samples': 18419200, 'steps': 35974, 'loss/train': 1.9455817937850952} 02/26/2022 00:08:50 - INFO - codeparrot_training - Step 35975: {'lr': 9.813437667457229e-05, 'samples': 18419712, 'steps': 35975, 'loss/train': 2.243952751159668} 02/26/2022 00:08:55 - INFO - codeparrot_training - Step 35976: {'lr': 9.812137950518432e-05, 'samples': 18420224, 'steps': 35976, 'loss/train': 0.57295161485672} 02/26/2022 00:08:59 - INFO - codeparrot_training - Step 35977: {'lr': 9.810838298639605e-05, 'samples': 18420736, 'steps': 35977, 'loss/train': 1.1149544715881348} 02/26/2022 00:09:04 - INFO - codeparrot_training - Step 35978: {'lr': 9.80953871182633e-05, 'samples': 18421248, 'steps': 35978, 'loss/train': 1.590366005897522} 02/26/2022 00:09:08 - INFO - codeparrot_training - Step 35979: {'lr': 9.80823919008417e-05, 'samples': 18421760, 'steps': 35979, 'loss/train': 2.021833658218384} 02/26/2022 00:09:13 - INFO - codeparrot_training - Step 35980: {'lr': 9.806939733418687e-05, 'samples': 18422272, 'steps': 35980, 'loss/train': 1.6028356552124023} 02/26/2022 00:09:17 - INFO - codeparrot_training - Step 35981: {'lr': 9.805640341835442e-05, 'samples': 18422784, 'steps': 35981, 'loss/train': 1.319348931312561} 02/26/2022 00:09:22 - INFO - codeparrot_training - Step 35982: {'lr': 9.804341015340021e-05, 'samples': 18423296, 'steps': 35982, 'loss/train': 1.474488377571106} 02/26/2022 00:09:26 - INFO - codeparrot_training - Step 35983: {'lr': 9.803041753937967e-05, 'samples': 18423808, 'steps': 35983, 'loss/train': 2.3666207790374756} 02/26/2022 00:09:32 - INFO - codeparrot_training - Step 35984: {'lr': 9.801742557634872e-05, 'samples': 18424320, 'steps': 35984, 'loss/train': 1.7574007511138916} 02/26/2022 00:09:35 - INFO - codeparrot_training - Step 35985: {'lr': 9.80044342643627e-05, 'samples': 18424832, 'steps': 35985, 'loss/train': 1.3362573385238647} 02/26/2022 00:09:41 - INFO - codeparrot_training - Step 35986: {'lr': 9.799144360347751e-05, 'samples': 18425344, 'steps': 35986, 'loss/train': 1.7247356176376343} 02/26/2022 00:09:44 - INFO - codeparrot_training - Step 35987: {'lr': 9.79784535937486e-05, 'samples': 18425856, 'steps': 35987, 'loss/train': 2.2092444896698} 02/26/2022 00:09:50 - INFO - codeparrot_training - Step 35988: {'lr': 9.796546423523192e-05, 'samples': 18426368, 'steps': 35988, 'loss/train': 0.9692898988723755} 02/26/2022 00:09:53 - INFO - codeparrot_training - Step 35989: {'lr': 9.795247552798273e-05, 'samples': 18426880, 'steps': 35989, 'loss/train': 0.5010618567466736} 02/26/2022 00:09:59 - INFO - codeparrot_training - Step 35990: {'lr': 9.793948747205694e-05, 'samples': 18427392, 'steps': 35990, 'loss/train': 1.6082537174224854} 02/26/2022 00:10:04 - INFO - codeparrot_training - Step 35991: {'lr': 9.792650006751002e-05, 'samples': 18427904, 'steps': 35991, 'loss/train': 1.2630656957626343} 02/26/2022 00:10:08 - INFO - codeparrot_training - Step 35992: {'lr': 9.791351331439779e-05, 'samples': 18428416, 'steps': 35992, 'loss/train': 1.6579424142837524} 02/26/2022 00:10:13 - INFO - codeparrot_training - Step 35993: {'lr': 9.790052721277576e-05, 'samples': 18428928, 'steps': 35993, 'loss/train': 2.244624614715576} 02/26/2022 00:10:17 - INFO - codeparrot_training - Step 35994: {'lr': 9.788754176269957e-05, 'samples': 18429440, 'steps': 35994, 'loss/train': 1.616926670074463} 02/26/2022 00:10:22 - INFO - codeparrot_training - Step 35995: {'lr': 9.787455696422476e-05, 'samples': 18429952, 'steps': 35995, 'loss/train': 1.8336987495422363} 02/26/2022 00:10:26 - INFO - codeparrot_training - Step 35996: {'lr': 9.786157281740712e-05, 'samples': 18430464, 'steps': 35996, 'loss/train': 2.339150905609131} 02/26/2022 00:10:31 - INFO - codeparrot_training - Step 35997: {'lr': 9.784858932230222e-05, 'samples': 18430976, 'steps': 35997, 'loss/train': 1.881528615951538} 02/26/2022 00:10:35 - INFO - codeparrot_training - Step 35998: {'lr': 9.783560647896564e-05, 'samples': 18431488, 'steps': 35998, 'loss/train': 0.11420764774084091} 02/26/2022 00:10:41 - INFO - codeparrot_training - Step 35999: {'lr': 9.7822624287453e-05, 'samples': 18432000, 'steps': 35999, 'loss/train': 1.2246854305267334} 02/26/2022 00:10:41 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 00:10:59 - WARNING - huggingface_hub.repository - Several commits (36) will be pushed upstream. 02/26/2022 00:10:59 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 00:11:34 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy aab1012..f71b356 floral-grass-11 -> floral-grass-11 02/26/2022 00:11:39 - INFO - codeparrot_training - Step 36000: {'lr': 9.780964274781984e-05, 'samples': 18432512, 'steps': 36000, 'loss/train': 1.5604214668273926} 02/26/2022 00:11:44 - INFO - codeparrot_training - Step 36001: {'lr': 9.77966618601219e-05, 'samples': 18433024, 'steps': 36001, 'loss/train': 1.2772327661514282} 02/26/2022 00:11:48 - INFO - codeparrot_training - Step 36002: {'lr': 9.778368162441476e-05, 'samples': 18433536, 'steps': 36002, 'loss/train': 0.23902873694896698} 02/26/2022 00:11:55 - INFO - codeparrot_training - Step 36003: {'lr': 9.7770702040754e-05, 'samples': 18434048, 'steps': 36003, 'loss/train': 1.5439203977584839} 02/26/2022 00:11:58 - INFO - codeparrot_training - Step 36004: {'lr': 9.775772310919509e-05, 'samples': 18434560, 'steps': 36004, 'loss/train': 1.5662668943405151} 02/26/2022 00:12:04 - INFO - codeparrot_training - Step 36005: {'lr': 9.774474482979388e-05, 'samples': 18435072, 'steps': 36005, 'loss/train': 1.638809084892273} 02/26/2022 00:12:07 - INFO - codeparrot_training - Step 36006: {'lr': 9.773176720260582e-05, 'samples': 18435584, 'steps': 36006, 'loss/train': 0.26204827427864075} 02/26/2022 00:12:13 - INFO - codeparrot_training - Step 36007: {'lr': 9.771879022768651e-05, 'samples': 18436096, 'steps': 36007, 'loss/train': 1.5781837701797485} 02/26/2022 00:12:16 - INFO - codeparrot_training - Step 36008: {'lr': 9.770581390509148e-05, 'samples': 18436608, 'steps': 36008, 'loss/train': 2.4978692531585693} 02/26/2022 00:12:22 - INFO - codeparrot_training - Step 36009: {'lr': 9.769283823487646e-05, 'samples': 18437120, 'steps': 36009, 'loss/train': 2.2008216381073} 02/26/2022 00:12:25 - INFO - codeparrot_training - Step 36010: {'lr': 9.767986321709696e-05, 'samples': 18437632, 'steps': 36010, 'loss/train': 1.958353042602539} 02/26/2022 00:12:31 - INFO - codeparrot_training - Step 36011: {'lr': 9.766688885180854e-05, 'samples': 18438144, 'steps': 36011, 'loss/train': 1.2244277000427246} 02/26/2022 00:12:34 - INFO - codeparrot_training - Step 36012: {'lr': 9.765391513906673e-05, 'samples': 18438656, 'steps': 36012, 'loss/train': 1.9946836233139038} 02/26/2022 00:12:41 - INFO - codeparrot_training - Step 36013: {'lr': 9.764094207892726e-05, 'samples': 18439168, 'steps': 36013, 'loss/train': 1.1914106607437134} 02/26/2022 00:12:44 - INFO - codeparrot_training - Step 36014: {'lr': 9.76279696714456e-05, 'samples': 18439680, 'steps': 36014, 'loss/train': 1.0989683866500854} 02/26/2022 00:12:50 - INFO - codeparrot_training - Step 36015: {'lr': 9.761499791667736e-05, 'samples': 18440192, 'steps': 36015, 'loss/train': 1.4466044902801514} 02/26/2022 00:12:53 - INFO - codeparrot_training - Step 36016: {'lr': 9.760202681467798e-05, 'samples': 18440704, 'steps': 36016, 'loss/train': 0.8954662680625916} 02/26/2022 00:12:59 - INFO - codeparrot_training - Step 36017: {'lr': 9.75890563655032e-05, 'samples': 18441216, 'steps': 36017, 'loss/train': 1.3574097156524658} 02/26/2022 00:13:02 - INFO - codeparrot_training - Step 36018: {'lr': 9.757608656920844e-05, 'samples': 18441728, 'steps': 36018, 'loss/train': 1.735330581665039} 02/26/2022 00:13:08 - INFO - codeparrot_training - Step 36019: {'lr': 9.756311742584947e-05, 'samples': 18442240, 'steps': 36019, 'loss/train': 9.147744178771973} 02/26/2022 00:13:11 - INFO - codeparrot_training - Step 36020: {'lr': 9.755014893548156e-05, 'samples': 18442752, 'steps': 36020, 'loss/train': 0.40673404932022095} 02/26/2022 00:13:17 - INFO - codeparrot_training - Step 36021: {'lr': 9.753718109816049e-05, 'samples': 18443264, 'steps': 36021, 'loss/train': 2.0481255054473877} 02/26/2022 00:13:20 - INFO - codeparrot_training - Step 36022: {'lr': 9.752421391394162e-05, 'samples': 18443776, 'steps': 36022, 'loss/train': 1.4590117931365967} 02/26/2022 00:13:27 - INFO - codeparrot_training - Step 36023: {'lr': 9.751124738288078e-05, 'samples': 18444288, 'steps': 36023, 'loss/train': 1.916585087776184} 02/26/2022 00:13:30 - INFO - codeparrot_training - Step 36024: {'lr': 9.749828150503314e-05, 'samples': 18444800, 'steps': 36024, 'loss/train': 1.1615715026855469} 02/26/2022 00:13:36 - INFO - codeparrot_training - Step 36025: {'lr': 9.748531628045456e-05, 'samples': 18445312, 'steps': 36025, 'loss/train': 0.618524968624115} 02/26/2022 00:13:39 - INFO - codeparrot_training - Step 36026: {'lr': 9.747235170920036e-05, 'samples': 18445824, 'steps': 36026, 'loss/train': 0.7209410667419434} 02/26/2022 00:13:45 - INFO - codeparrot_training - Step 36027: {'lr': 9.745938779132634e-05, 'samples': 18446336, 'steps': 36027, 'loss/train': 1.6152046918869019} 02/26/2022 00:13:48 - INFO - codeparrot_training - Step 36028: {'lr': 9.744642452688767e-05, 'samples': 18446848, 'steps': 36028, 'loss/train': 1.5498669147491455} 02/26/2022 00:13:54 - INFO - codeparrot_training - Step 36029: {'lr': 9.743346191594022e-05, 'samples': 18447360, 'steps': 36029, 'loss/train': 0.20733459293842316} 02/26/2022 00:13:57 - INFO - codeparrot_training - Step 36030: {'lr': 9.742049995853925e-05, 'samples': 18447872, 'steps': 36030, 'loss/train': 1.7159196138381958} 02/26/2022 00:14:03 - INFO - codeparrot_training - Step 36031: {'lr': 9.740753865474048e-05, 'samples': 18448384, 'steps': 36031, 'loss/train': 1.3167070150375366} 02/26/2022 00:14:06 - INFO - codeparrot_training - Step 36032: {'lr': 9.739457800459939e-05, 'samples': 18448896, 'steps': 36032, 'loss/train': 1.5952965021133423} 02/26/2022 00:14:13 - INFO - codeparrot_training - Step 36033: {'lr': 9.738161800817145e-05, 'samples': 18449408, 'steps': 36033, 'loss/train': 2.0263400077819824} 02/26/2022 00:14:16 - INFO - codeparrot_training - Step 36034: {'lr': 9.736865866551211e-05, 'samples': 18449920, 'steps': 36034, 'loss/train': 1.3503931760787964} 02/26/2022 00:14:22 - INFO - codeparrot_training - Step 36035: {'lr': 9.735569997667706e-05, 'samples': 18450432, 'steps': 36035, 'loss/train': 3.332951307296753} 02/26/2022 00:14:25 - INFO - codeparrot_training - Step 36036: {'lr': 9.734274194172174e-05, 'samples': 18450944, 'steps': 36036, 'loss/train': 1.9804167747497559} 02/26/2022 00:14:31 - INFO - codeparrot_training - Step 36037: {'lr': 9.732978456070163e-05, 'samples': 18451456, 'steps': 36037, 'loss/train': 1.8700439929962158} 02/26/2022 00:14:34 - INFO - codeparrot_training - Step 36038: {'lr': 9.731682783367215e-05, 'samples': 18451968, 'steps': 36038, 'loss/train': 2.0220859050750732} 02/26/2022 00:14:40 - INFO - codeparrot_training - Step 36039: {'lr': 9.7303871760689e-05, 'samples': 18452480, 'steps': 36039, 'loss/train': 1.0624700784683228} 02/26/2022 00:14:45 - INFO - codeparrot_training - Step 36040: {'lr': 9.729091634180757e-05, 'samples': 18452992, 'steps': 36040, 'loss/train': 1.7062714099884033} 02/26/2022 00:14:49 - INFO - codeparrot_training - Step 36041: {'lr': 9.727796157708335e-05, 'samples': 18453504, 'steps': 36041, 'loss/train': 1.4486877918243408} 02/26/2022 00:14:55 - INFO - codeparrot_training - Step 36042: {'lr': 9.726500746657177e-05, 'samples': 18454016, 'steps': 36042, 'loss/train': 0.1205480694770813} 02/26/2022 00:14:58 - INFO - codeparrot_training - Step 36043: {'lr': 9.72520540103285e-05, 'samples': 18454528, 'steps': 36043, 'loss/train': 1.0018887519836426} 02/26/2022 00:15:02 - INFO - codeparrot_training - Step 36044: {'lr': 9.72391012084089e-05, 'samples': 18455040, 'steps': 36044, 'loss/train': 0.9732137322425842} 02/26/2022 00:15:07 - INFO - codeparrot_training - Step 36045: {'lr': 9.722614906086849e-05, 'samples': 18455552, 'steps': 36045, 'loss/train': 1.1737741231918335} 02/26/2022 00:15:12 - INFO - codeparrot_training - Step 36046: {'lr': 9.721319756776276e-05, 'samples': 18456064, 'steps': 36046, 'loss/train': 1.9837580919265747} 02/26/2022 00:15:16 - INFO - codeparrot_training - Step 36047: {'lr': 9.72002467291471e-05, 'samples': 18456576, 'steps': 36047, 'loss/train': 2.022613763809204} 02/26/2022 00:15:23 - INFO - codeparrot_training - Step 36048: {'lr': 9.718729654507713e-05, 'samples': 18457088, 'steps': 36048, 'loss/train': 0.04201718419790268} 02/26/2022 00:15:26 - INFO - codeparrot_training - Step 36049: {'lr': 9.717434701560827e-05, 'samples': 18457600, 'steps': 36049, 'loss/train': 0.8163706064224243} 02/26/2022 00:15:32 - INFO - codeparrot_training - Step 36050: {'lr': 9.716139814079594e-05, 'samples': 18458112, 'steps': 36050, 'loss/train': 1.975005030632019} 02/26/2022 00:15:35 - INFO - codeparrot_training - Step 36051: {'lr': 9.71484499206956e-05, 'samples': 18458624, 'steps': 36051, 'loss/train': 0.0918508991599083} 02/26/2022 00:15:41 - INFO - codeparrot_training - Step 36052: {'lr': 9.713550235536283e-05, 'samples': 18459136, 'steps': 36052, 'loss/train': 1.9509221315383911} 02/26/2022 00:15:44 - INFO - codeparrot_training - Step 36053: {'lr': 9.712255544485296e-05, 'samples': 18459648, 'steps': 36053, 'loss/train': 2.816897392272949} 02/26/2022 00:15:50 - INFO - codeparrot_training - Step 36054: {'lr': 9.710960918922171e-05, 'samples': 18460160, 'steps': 36054, 'loss/train': 2.4247937202453613} 02/26/2022 00:15:53 - INFO - codeparrot_training - Step 36055: {'lr': 9.709666358852414e-05, 'samples': 18460672, 'steps': 36055, 'loss/train': 1.8876968622207642} 02/26/2022 00:15:58 - INFO - codeparrot_training - Step 36056: {'lr': 9.708371864281601e-05, 'samples': 18461184, 'steps': 36056, 'loss/train': 2.1317458152770996} 02/26/2022 00:16:02 - INFO - codeparrot_training - Step 36057: {'lr': 9.707077435215258e-05, 'samples': 18461696, 'steps': 36057, 'loss/train': 1.6023669242858887} 02/26/2022 00:16:08 - INFO - codeparrot_training - Step 36058: {'lr': 9.705783071658958e-05, 'samples': 18462208, 'steps': 36058, 'loss/train': 0.7936185598373413} 02/26/2022 00:16:12 - INFO - codeparrot_training - Step 36059: {'lr': 9.704488773618209e-05, 'samples': 18462720, 'steps': 36059, 'loss/train': 1.46337890625} 02/26/2022 00:16:17 - INFO - codeparrot_training - Step 36060: {'lr': 9.703194541098581e-05, 'samples': 18463232, 'steps': 36060, 'loss/train': 0.6197915077209473} 02/26/2022 00:16:21 - INFO - codeparrot_training - Step 36061: {'lr': 9.701900374105602e-05, 'samples': 18463744, 'steps': 36061, 'loss/train': 2.0371103286743164} 02/26/2022 00:16:26 - INFO - codeparrot_training - Step 36062: {'lr': 9.700606272644841e-05, 'samples': 18464256, 'steps': 36062, 'loss/train': 0.45785778760910034} 02/26/2022 00:16:30 - INFO - codeparrot_training - Step 36063: {'lr': 9.699312236721808e-05, 'samples': 18464768, 'steps': 36063, 'loss/train': 1.5471307039260864} 02/26/2022 00:16:35 - INFO - codeparrot_training - Step 36064: {'lr': 9.698018266342071e-05, 'samples': 18465280, 'steps': 36064, 'loss/train': 1.319109559059143} 02/26/2022 00:16:39 - INFO - codeparrot_training - Step 36065: {'lr': 9.696724361511156e-05, 'samples': 18465792, 'steps': 36065, 'loss/train': 2.1715946197509766} 02/26/2022 00:16:44 - INFO - codeparrot_training - Step 36066: {'lr': 9.695430522234625e-05, 'samples': 18466304, 'steps': 36066, 'loss/train': 2.603968858718872} 02/26/2022 00:16:48 - INFO - codeparrot_training - Step 36067: {'lr': 9.694136748518007e-05, 'samples': 18466816, 'steps': 36067, 'loss/train': 0.927971363067627} 02/26/2022 00:16:53 - INFO - codeparrot_training - Step 36068: {'lr': 9.69284304036685e-05, 'samples': 18467328, 'steps': 36068, 'loss/train': 1.7105233669281006} 02/26/2022 00:16:57 - INFO - codeparrot_training - Step 36069: {'lr': 9.69154939778668e-05, 'samples': 18467840, 'steps': 36069, 'loss/train': 1.2192550897598267} 02/26/2022 00:17:03 - INFO - codeparrot_training - Step 36070: {'lr': 9.690255820783064e-05, 'samples': 18468352, 'steps': 36070, 'loss/train': 2.7139899730682373} 02/26/2022 00:17:06 - INFO - codeparrot_training - Step 36071: {'lr': 9.688962309361529e-05, 'samples': 18468864, 'steps': 36071, 'loss/train': 1.2359057664871216} 02/26/2022 00:17:12 - INFO - codeparrot_training - Step 36072: {'lr': 9.687668863527619e-05, 'samples': 18469376, 'steps': 36072, 'loss/train': 1.5974620580673218} 02/26/2022 00:17:16 - INFO - codeparrot_training - Step 36073: {'lr': 9.686375483286864e-05, 'samples': 18469888, 'steps': 36073, 'loss/train': 0.8579695820808411} 02/26/2022 00:17:21 - INFO - codeparrot_training - Step 36074: {'lr': 9.685082168644823e-05, 'samples': 18470400, 'steps': 36074, 'loss/train': 1.3160451650619507} 02/26/2022 00:17:25 - INFO - codeparrot_training - Step 36075: {'lr': 9.683788919607026e-05, 'samples': 18470912, 'steps': 36075, 'loss/train': 1.2578535079956055} 02/26/2022 00:17:30 - INFO - codeparrot_training - Step 36076: {'lr': 9.682495736179014e-05, 'samples': 18471424, 'steps': 36076, 'loss/train': 2.295518636703491} 02/26/2022 00:17:34 - INFO - codeparrot_training - Step 36077: {'lr': 9.681202618366317e-05, 'samples': 18471936, 'steps': 36077, 'loss/train': 1.4106355905532837} 02/26/2022 00:17:39 - INFO - codeparrot_training - Step 36078: {'lr': 9.679909566174494e-05, 'samples': 18472448, 'steps': 36078, 'loss/train': 1.4337424039840698} 02/26/2022 00:17:42 - INFO - codeparrot_training - Step 36079: {'lr': 9.678616579609073e-05, 'samples': 18472960, 'steps': 36079, 'loss/train': 1.8247777223587036} 02/26/2022 00:17:49 - INFO - codeparrot_training - Step 36080: {'lr': 9.677323658675594e-05, 'samples': 18473472, 'steps': 36080, 'loss/train': 1.671035885810852} 02/26/2022 00:17:52 - INFO - codeparrot_training - Step 36081: {'lr': 9.676030803379585e-05, 'samples': 18473984, 'steps': 36081, 'loss/train': 1.8489046096801758} 02/26/2022 00:17:58 - INFO - codeparrot_training - Step 36082: {'lr': 9.674738013726603e-05, 'samples': 18474496, 'steps': 36082, 'loss/train': 2.2339859008789062} 02/26/2022 00:18:01 - INFO - codeparrot_training - Step 36083: {'lr': 9.673445289722177e-05, 'samples': 18475008, 'steps': 36083, 'loss/train': 0.49212533235549927} 02/26/2022 00:18:07 - INFO - codeparrot_training - Step 36084: {'lr': 9.672152631371848e-05, 'samples': 18475520, 'steps': 36084, 'loss/train': 1.9349168539047241} 02/26/2022 00:18:10 - INFO - codeparrot_training - Step 36085: {'lr': 9.670860038681136e-05, 'samples': 18476032, 'steps': 36085, 'loss/train': 2.043856620788574} 02/26/2022 00:18:16 - INFO - codeparrot_training - Step 36086: {'lr': 9.669567511655602e-05, 'samples': 18476544, 'steps': 36086, 'loss/train': 2.254288911819458} 02/26/2022 00:18:19 - INFO - codeparrot_training - Step 36087: {'lr': 9.668275050300776e-05, 'samples': 18477056, 'steps': 36087, 'loss/train': 2.0351157188415527} 02/26/2022 00:18:25 - INFO - codeparrot_training - Step 36088: {'lr': 9.666982654622189e-05, 'samples': 18477568, 'steps': 36088, 'loss/train': 1.0550917387008667} 02/26/2022 00:18:28 - INFO - codeparrot_training - Step 36089: {'lr': 9.66569032462537e-05, 'samples': 18478080, 'steps': 36089, 'loss/train': 1.5271868705749512} 02/26/2022 00:18:35 - INFO - codeparrot_training - Step 36090: {'lr': 9.664398060315876e-05, 'samples': 18478592, 'steps': 36090, 'loss/train': 1.8364299535751343} 02/26/2022 00:18:38 - INFO - codeparrot_training - Step 36091: {'lr': 9.663105861699229e-05, 'samples': 18479104, 'steps': 36091, 'loss/train': 1.8867496252059937} 02/26/2022 00:18:43 - INFO - codeparrot_training - Step 36092: {'lr': 9.661813728780958e-05, 'samples': 18479616, 'steps': 36092, 'loss/train': 1.69874906539917} 02/26/2022 00:18:47 - INFO - codeparrot_training - Step 36093: {'lr': 9.660521661566615e-05, 'samples': 18480128, 'steps': 36093, 'loss/train': 1.8722460269927979} 02/26/2022 00:18:52 - INFO - codeparrot_training - Step 36094: {'lr': 9.659229660061728e-05, 'samples': 18480640, 'steps': 36094, 'loss/train': 1.763447880744934} 02/26/2022 00:18:56 - INFO - codeparrot_training - Step 36095: {'lr': 9.657937724271829e-05, 'samples': 18481152, 'steps': 36095, 'loss/train': 1.0011274814605713} 02/26/2022 00:19:01 - INFO - codeparrot_training - Step 36096: {'lr': 9.656645854202442e-05, 'samples': 18481664, 'steps': 36096, 'loss/train': 1.1618714332580566} 02/26/2022 00:19:05 - INFO - codeparrot_training - Step 36097: {'lr': 9.655354049859133e-05, 'samples': 18482176, 'steps': 36097, 'loss/train': 2.0890953540802} 02/26/2022 00:19:10 - INFO - codeparrot_training - Step 36098: {'lr': 9.654062311247397e-05, 'samples': 18482688, 'steps': 36098, 'loss/train': 1.7578837871551514} 02/26/2022 00:19:14 - INFO - codeparrot_training - Step 36099: {'lr': 9.652770638372792e-05, 'samples': 18483200, 'steps': 36099, 'loss/train': 1.8363592624664307} 02/26/2022 00:19:19 - INFO - codeparrot_training - Step 36100: {'lr': 9.651479031240836e-05, 'samples': 18483712, 'steps': 36100, 'loss/train': 2.721839666366577} 02/26/2022 00:19:23 - INFO - codeparrot_training - Step 36101: {'lr': 9.650187489857079e-05, 'samples': 18484224, 'steps': 36101, 'loss/train': 1.9285857677459717} 02/26/2022 00:19:28 - INFO - codeparrot_training - Step 36102: {'lr': 9.648896014227046e-05, 'samples': 18484736, 'steps': 36102, 'loss/train': 1.311247706413269} 02/26/2022 00:19:32 - INFO - codeparrot_training - Step 36103: {'lr': 9.647604604356267e-05, 'samples': 18485248, 'steps': 36103, 'loss/train': 1.5463550090789795} 02/26/2022 00:19:37 - INFO - codeparrot_training - Step 36104: {'lr': 9.646313260250267e-05, 'samples': 18485760, 'steps': 36104, 'loss/train': 2.2677383422851562} 02/26/2022 00:19:41 - INFO - codeparrot_training - Step 36105: {'lr': 9.645021981914592e-05, 'samples': 18486272, 'steps': 36105, 'loss/train': 2.1401143074035645} 02/26/2022 00:19:47 - INFO - codeparrot_training - Step 36106: {'lr': 9.64373076935477e-05, 'samples': 18486784, 'steps': 36106, 'loss/train': 3.245847225189209} 02/26/2022 00:19:51 - INFO - codeparrot_training - Step 36107: {'lr': 9.642439622576327e-05, 'samples': 18487296, 'steps': 36107, 'loss/train': 1.6988332271575928} 02/26/2022 00:19:56 - INFO - codeparrot_training - Step 36108: {'lr': 9.641148541584788e-05, 'samples': 18487808, 'steps': 36108, 'loss/train': 1.2002081871032715} 02/26/2022 00:20:00 - INFO - codeparrot_training - Step 36109: {'lr': 9.639857526385701e-05, 'samples': 18488320, 'steps': 36109, 'loss/train': 1.9421395063400269} 02/26/2022 00:20:05 - INFO - codeparrot_training - Step 36110: {'lr': 9.638566576984586e-05, 'samples': 18488832, 'steps': 36110, 'loss/train': 1.74909245967865} 02/26/2022 00:20:09 - INFO - codeparrot_training - Step 36111: {'lr': 9.637275693386974e-05, 'samples': 18489344, 'steps': 36111, 'loss/train': 2.5859227180480957} 02/26/2022 00:20:14 - INFO - codeparrot_training - Step 36112: {'lr': 9.635984875598389e-05, 'samples': 18489856, 'steps': 36112, 'loss/train': 2.328519821166992} 02/26/2022 00:20:18 - INFO - codeparrot_training - Step 36113: {'lr': 9.63469412362437e-05, 'samples': 18490368, 'steps': 36113, 'loss/train': 1.3176090717315674} 02/26/2022 00:20:23 - INFO - codeparrot_training - Step 36114: {'lr': 9.633403437470437e-05, 'samples': 18490880, 'steps': 36114, 'loss/train': 1.733176589012146} 02/26/2022 00:20:27 - INFO - codeparrot_training - Step 36115: {'lr': 9.632112817142141e-05, 'samples': 18491392, 'steps': 36115, 'loss/train': 1.5681864023208618} 02/26/2022 00:20:33 - INFO - codeparrot_training - Step 36116: {'lr': 9.630822262644976e-05, 'samples': 18491904, 'steps': 36116, 'loss/train': 1.6882433891296387} 02/26/2022 00:20:36 - INFO - codeparrot_training - Step 36117: {'lr': 9.629531773984496e-05, 'samples': 18492416, 'steps': 36117, 'loss/train': 1.7538223266601562} 02/26/2022 00:20:42 - INFO - codeparrot_training - Step 36118: {'lr': 9.628241351166214e-05, 'samples': 18492928, 'steps': 36118, 'loss/train': 1.1836498975753784} 02/26/2022 00:20:45 - INFO - codeparrot_training - Step 36119: {'lr': 9.626950994195682e-05, 'samples': 18493440, 'steps': 36119, 'loss/train': 1.354833960533142} 02/26/2022 00:20:51 - INFO - codeparrot_training - Step 36120: {'lr': 9.625660703078392e-05, 'samples': 18493952, 'steps': 36120, 'loss/train': 1.5595625638961792} 02/26/2022 00:20:54 - INFO - codeparrot_training - Step 36121: {'lr': 9.6243704778199e-05, 'samples': 18494464, 'steps': 36121, 'loss/train': 2.852039337158203} 02/26/2022 00:21:00 - INFO - codeparrot_training - Step 36122: {'lr': 9.623080318425715e-05, 'samples': 18494976, 'steps': 36122, 'loss/train': 2.465924024581909} 02/26/2022 00:21:03 - INFO - codeparrot_training - Step 36123: {'lr': 9.621790224901387e-05, 'samples': 18495488, 'steps': 36123, 'loss/train': 2.143178939819336} 02/26/2022 00:21:09 - INFO - codeparrot_training - Step 36124: {'lr': 9.620500197252408e-05, 'samples': 18496000, 'steps': 36124, 'loss/train': 1.795910120010376} 02/26/2022 00:21:12 - INFO - codeparrot_training - Step 36125: {'lr': 9.619210235484333e-05, 'samples': 18496512, 'steps': 36125, 'loss/train': 0.8504828214645386} 02/26/2022 00:21:18 - INFO - codeparrot_training - Step 36126: {'lr': 9.617920339602665e-05, 'samples': 18497024, 'steps': 36126, 'loss/train': 1.2204484939575195} 02/26/2022 00:21:22 - INFO - codeparrot_training - Step 36127: {'lr': 9.616630509612955e-05, 'samples': 18497536, 'steps': 36127, 'loss/train': 1.0036232471466064} 02/26/2022 00:21:27 - INFO - codeparrot_training - Step 36128: {'lr': 9.615340745520712e-05, 'samples': 18498048, 'steps': 36128, 'loss/train': 1.9192155599594116} 02/26/2022 00:21:33 - INFO - codeparrot_training - Step 36129: {'lr': 9.614051047331469e-05, 'samples': 18498560, 'steps': 36129, 'loss/train': 1.749711036682129} 02/26/2022 00:21:36 - INFO - codeparrot_training - Step 36130: {'lr': 9.612761415050741e-05, 'samples': 18499072, 'steps': 36130, 'loss/train': 1.9420716762542725} 02/26/2022 00:21:42 - INFO - codeparrot_training - Step 36131: {'lr': 9.61147184868405e-05, 'samples': 18499584, 'steps': 36131, 'loss/train': 0.5370197296142578} 02/26/2022 00:21:45 - INFO - codeparrot_training - Step 36132: {'lr': 9.610182348236934e-05, 'samples': 18500096, 'steps': 36132, 'loss/train': 7.3840813636779785} 02/26/2022 00:21:51 - INFO - codeparrot_training - Step 36133: {'lr': 9.608892913714912e-05, 'samples': 18500608, 'steps': 36133, 'loss/train': 0.36018943786621094} 02/26/2022 00:21:54 - INFO - codeparrot_training - Step 36134: {'lr': 9.607603545123505e-05, 'samples': 18501120, 'steps': 36134, 'loss/train': 1.453967809677124} 02/26/2022 00:22:00 - INFO - codeparrot_training - Step 36135: {'lr': 9.606314242468228e-05, 'samples': 18501632, 'steps': 36135, 'loss/train': 1.7087455987930298} 02/26/2022 00:22:04 - INFO - codeparrot_training - Step 36136: {'lr': 9.605025005754622e-05, 'samples': 18502144, 'steps': 36136, 'loss/train': 2.5946903228759766} 02/26/2022 00:22:09 - INFO - codeparrot_training - Step 36137: {'lr': 9.6037358349882e-05, 'samples': 18502656, 'steps': 36137, 'loss/train': 1.105545163154602} 02/26/2022 00:22:13 - INFO - codeparrot_training - Step 36138: {'lr': 9.602446730174485e-05, 'samples': 18503168, 'steps': 36138, 'loss/train': 3.0374724864959717} 02/26/2022 00:22:18 - INFO - codeparrot_training - Step 36139: {'lr': 9.601157691318991e-05, 'samples': 18503680, 'steps': 36139, 'loss/train': 0.5579282641410828} 02/26/2022 00:22:22 - INFO - codeparrot_training - Step 36140: {'lr': 9.599868718427256e-05, 'samples': 18504192, 'steps': 36140, 'loss/train': 2.259603977203369} 02/26/2022 00:22:27 - INFO - codeparrot_training - Step 36141: {'lr': 9.598579811504792e-05, 'samples': 18504704, 'steps': 36141, 'loss/train': 1.8042490482330322} 02/26/2022 00:22:31 - INFO - codeparrot_training - Step 36142: {'lr': 9.597290970557124e-05, 'samples': 18505216, 'steps': 36142, 'loss/train': 1.6345702409744263} 02/26/2022 00:22:36 - INFO - codeparrot_training - Step 36143: {'lr': 9.596002195589759e-05, 'samples': 18505728, 'steps': 36143, 'loss/train': 0.7388705015182495} 02/26/2022 00:22:40 - INFO - codeparrot_training - Step 36144: {'lr': 9.59471348660824e-05, 'samples': 18506240, 'steps': 36144, 'loss/train': 1.916875958442688} 02/26/2022 00:22:45 - INFO - codeparrot_training - Step 36145: {'lr': 9.593424843618076e-05, 'samples': 18506752, 'steps': 36145, 'loss/train': 2.252598524093628} 02/26/2022 00:22:49 - INFO - codeparrot_training - Step 36146: {'lr': 9.592136266624787e-05, 'samples': 18507264, 'steps': 36146, 'loss/train': 2.5656566619873047} 02/26/2022 00:22:54 - INFO - codeparrot_training - Step 36147: {'lr': 9.590847755633885e-05, 'samples': 18507776, 'steps': 36147, 'loss/train': 2.0132853984832764} 02/26/2022 00:22:58 - INFO - codeparrot_training - Step 36148: {'lr': 9.589559310650911e-05, 'samples': 18508288, 'steps': 36148, 'loss/train': 1.502086877822876} 02/26/2022 00:23:03 - INFO - codeparrot_training - Step 36149: {'lr': 9.588270931681367e-05, 'samples': 18508800, 'steps': 36149, 'loss/train': 1.199326992034912} 02/26/2022 00:23:07 - INFO - codeparrot_training - Step 36150: {'lr': 9.586982618730778e-05, 'samples': 18509312, 'steps': 36150, 'loss/train': 1.222579002380371} 02/26/2022 00:23:13 - INFO - codeparrot_training - Step 36151: {'lr': 9.585694371804654e-05, 'samples': 18509824, 'steps': 36151, 'loss/train': 1.4391850233078003} 02/26/2022 00:23:16 - INFO - codeparrot_training - Step 36152: {'lr': 9.584406190908527e-05, 'samples': 18510336, 'steps': 36152, 'loss/train': 0.9116042256355286} 02/26/2022 00:23:22 - INFO - codeparrot_training - Step 36153: {'lr': 9.5831180760479e-05, 'samples': 18510848, 'steps': 36153, 'loss/train': 2.355449914932251} 02/26/2022 00:23:25 - INFO - codeparrot_training - Step 36154: {'lr': 9.581830027228319e-05, 'samples': 18511360, 'steps': 36154, 'loss/train': 1.5806246995925903} 02/26/2022 00:23:31 - INFO - codeparrot_training - Step 36155: {'lr': 9.580542044455265e-05, 'samples': 18511872, 'steps': 36155, 'loss/train': 1.011007308959961} 02/26/2022 00:23:34 - INFO - codeparrot_training - Step 36156: {'lr': 9.579254127734279e-05, 'samples': 18512384, 'steps': 36156, 'loss/train': 1.5051230192184448} 02/26/2022 00:23:40 - INFO - codeparrot_training - Step 36157: {'lr': 9.577966277070865e-05, 'samples': 18512896, 'steps': 36157, 'loss/train': 0.9843266606330872} 02/26/2022 00:23:43 - INFO - codeparrot_training - Step 36158: {'lr': 9.576678492470564e-05, 'samples': 18513408, 'steps': 36158, 'loss/train': 0.6475926041603088} 02/26/2022 00:23:49 - INFO - codeparrot_training - Step 36159: {'lr': 9.575390773938858e-05, 'samples': 18513920, 'steps': 36159, 'loss/train': 1.8057377338409424} 02/26/2022 00:23:52 - INFO - codeparrot_training - Step 36160: {'lr': 9.574103121481287e-05, 'samples': 18514432, 'steps': 36160, 'loss/train': 1.9257701635360718} 02/26/2022 00:23:59 - INFO - codeparrot_training - Step 36161: {'lr': 9.572815535103351e-05, 'samples': 18514944, 'steps': 36161, 'loss/train': 3.0665485858917236} 02/26/2022 00:24:02 - INFO - codeparrot_training - Step 36162: {'lr': 9.571528014810585e-05, 'samples': 18515456, 'steps': 36162, 'loss/train': 2.3938939571380615} 02/26/2022 00:24:08 - INFO - codeparrot_training - Step 36163: {'lr': 9.570240560608492e-05, 'samples': 18515968, 'steps': 36163, 'loss/train': 1.0796701908111572} 02/26/2022 00:24:11 - INFO - codeparrot_training - Step 36164: {'lr': 9.568953172502589e-05, 'samples': 18516480, 'steps': 36164, 'loss/train': 0.8898747563362122} 02/26/2022 00:24:17 - INFO - codeparrot_training - Step 36165: {'lr': 9.567665850498386e-05, 'samples': 18516992, 'steps': 36165, 'loss/train': 1.6962101459503174} 02/26/2022 00:24:20 - INFO - codeparrot_training - Step 36166: {'lr': 9.566378594601408e-05, 'samples': 18517504, 'steps': 36166, 'loss/train': 2.162107467651367} 02/26/2022 00:24:26 - INFO - codeparrot_training - Step 36167: {'lr': 9.56509140481716e-05, 'samples': 18518016, 'steps': 36167, 'loss/train': 1.370734453201294} 02/26/2022 00:24:29 - INFO - codeparrot_training - Step 36168: {'lr': 9.563804281151164e-05, 'samples': 18518528, 'steps': 36168, 'loss/train': 2.5249743461608887} 02/26/2022 00:24:35 - INFO - codeparrot_training - Step 36169: {'lr': 9.562517223608918e-05, 'samples': 18519040, 'steps': 36169, 'loss/train': 2.1326749324798584} 02/26/2022 00:24:38 - INFO - codeparrot_training - Step 36170: {'lr': 9.561230232195959e-05, 'samples': 18519552, 'steps': 36170, 'loss/train': 2.1402714252471924} 02/26/2022 00:24:44 - INFO - codeparrot_training - Step 36171: {'lr': 9.559943306917781e-05, 'samples': 18520064, 'steps': 36171, 'loss/train': 2.423326015472412} 02/26/2022 00:24:47 - INFO - codeparrot_training - Step 36172: {'lr': 9.558656447779906e-05, 'samples': 18520576, 'steps': 36172, 'loss/train': 2.217533588409424} 02/26/2022 00:24:54 - INFO - codeparrot_training - Step 36173: {'lr': 9.557369654787836e-05, 'samples': 18521088, 'steps': 36173, 'loss/train': 2.359816789627075} 02/26/2022 00:24:57 - INFO - codeparrot_training - Step 36174: {'lr': 9.5560829279471e-05, 'samples': 18521600, 'steps': 36174, 'loss/train': 2.111583948135376} 02/26/2022 00:25:03 - INFO - codeparrot_training - Step 36175: {'lr': 9.5547962672632e-05, 'samples': 18522112, 'steps': 36175, 'loss/train': 1.4911047220230103} 02/26/2022 00:25:06 - INFO - codeparrot_training - Step 36176: {'lr': 9.553509672741645e-05, 'samples': 18522624, 'steps': 36176, 'loss/train': 2.1724941730499268} 02/26/2022 00:25:12 - INFO - codeparrot_training - Step 36177: {'lr': 9.552223144387954e-05, 'samples': 18523136, 'steps': 36177, 'loss/train': 2.2457659244537354} 02/26/2022 00:25:15 - INFO - codeparrot_training - Step 36178: {'lr': 9.550936682207622e-05, 'samples': 18523648, 'steps': 36178, 'loss/train': 1.2652034759521484} 02/26/2022 00:25:21 - INFO - codeparrot_training - Step 36179: {'lr': 9.549650286206183e-05, 'samples': 18524160, 'steps': 36179, 'loss/train': 1.9991176128387451} 02/26/2022 00:25:24 - INFO - codeparrot_training - Step 36180: {'lr': 9.548363956389133e-05, 'samples': 18524672, 'steps': 36180, 'loss/train': 1.9419400691986084} 02/26/2022 00:25:31 - INFO - codeparrot_training - Step 36181: {'lr': 9.547077692761987e-05, 'samples': 18525184, 'steps': 36181, 'loss/train': 1.9898419380187988} 02/26/2022 00:25:34 - INFO - codeparrot_training - Step 36182: {'lr': 9.545791495330247e-05, 'samples': 18525696, 'steps': 36182, 'loss/train': 0.9300118088722229} 02/26/2022 00:25:40 - INFO - codeparrot_training - Step 36183: {'lr': 9.544505364099435e-05, 'samples': 18526208, 'steps': 36183, 'loss/train': 1.7209088802337646} 02/26/2022 00:25:43 - INFO - codeparrot_training - Step 36184: {'lr': 9.543219299075057e-05, 'samples': 18526720, 'steps': 36184, 'loss/train': 1.4446091651916504} 02/26/2022 00:25:49 - INFO - codeparrot_training - Step 36185: {'lr': 9.541933300262614e-05, 'samples': 18527232, 'steps': 36185, 'loss/train': 1.3330631256103516} 02/26/2022 00:25:52 - INFO - codeparrot_training - Step 36186: {'lr': 9.540647367667618e-05, 'samples': 18527744, 'steps': 36186, 'loss/train': 1.9552959203720093} 02/26/2022 00:25:58 - INFO - codeparrot_training - Step 36187: {'lr': 9.539361501295585e-05, 'samples': 18528256, 'steps': 36187, 'loss/train': 2.862837076187134} 02/26/2022 00:26:01 - INFO - codeparrot_training - Step 36188: {'lr': 9.53807570115201e-05, 'samples': 18528768, 'steps': 36188, 'loss/train': 1.9843339920043945} 02/26/2022 00:26:07 - INFO - codeparrot_training - Step 36189: {'lr': 9.536789967242426e-05, 'samples': 18529280, 'steps': 36189, 'loss/train': 2.0549697875976562} 02/26/2022 00:26:10 - INFO - codeparrot_training - Step 36190: {'lr': 9.535504299572307e-05, 'samples': 18529792, 'steps': 36190, 'loss/train': 1.5688729286193848} 02/26/2022 00:26:16 - INFO - codeparrot_training - Step 36191: {'lr': 9.534218698147185e-05, 'samples': 18530304, 'steps': 36191, 'loss/train': 0.041098542511463165} 02/26/2022 00:26:20 - INFO - codeparrot_training - Step 36192: {'lr': 9.532933162972548e-05, 'samples': 18530816, 'steps': 36192, 'loss/train': 1.5200234651565552} 02/26/2022 00:26:25 - INFO - codeparrot_training - Step 36193: {'lr': 9.531647694053936e-05, 'samples': 18531328, 'steps': 36193, 'loss/train': 1.0544005632400513} 02/26/2022 00:26:28 - INFO - codeparrot_training - Step 36194: {'lr': 9.530362291396813e-05, 'samples': 18531840, 'steps': 36194, 'loss/train': 2.4101407527923584} 02/26/2022 00:26:34 - INFO - codeparrot_training - Step 36195: {'lr': 9.529076955006716e-05, 'samples': 18532352, 'steps': 36195, 'loss/train': 1.412487506866455} 02/26/2022 00:26:40 - INFO - codeparrot_training - Step 36196: {'lr': 9.527791684889131e-05, 'samples': 18532864, 'steps': 36196, 'loss/train': 1.501369595527649} 02/26/2022 00:26:44 - INFO - codeparrot_training - Step 36197: {'lr': 9.526506481049594e-05, 'samples': 18533376, 'steps': 36197, 'loss/train': 2.132791757583618} 02/26/2022 00:26:49 - INFO - codeparrot_training - Step 36198: {'lr': 9.525221343493568e-05, 'samples': 18533888, 'steps': 36198, 'loss/train': 0.909018874168396} 02/26/2022 00:26:53 - INFO - codeparrot_training - Step 36199: {'lr': 9.52393627222659e-05, 'samples': 18534400, 'steps': 36199, 'loss/train': 1.8276982307434082} 02/26/2022 00:26:58 - INFO - codeparrot_training - Step 36200: {'lr': 9.522651267254148e-05, 'samples': 18534912, 'steps': 36200, 'loss/train': 1.7221652269363403} 02/26/2022 00:27:02 - INFO - codeparrot_training - Step 36201: {'lr': 9.52136632858176e-05, 'samples': 18535424, 'steps': 36201, 'loss/train': 0.8048366904258728} 02/26/2022 00:27:07 - INFO - codeparrot_training - Step 36202: {'lr': 9.520081456214924e-05, 'samples': 18535936, 'steps': 36202, 'loss/train': 3.0710699558258057} 02/26/2022 00:27:11 - INFO - codeparrot_training - Step 36203: {'lr': 9.518796650159146e-05, 'samples': 18536448, 'steps': 36203, 'loss/train': 1.6428643465042114} 02/26/2022 00:27:16 - INFO - codeparrot_training - Step 36204: {'lr': 9.517511910419916e-05, 'samples': 18536960, 'steps': 36204, 'loss/train': 2.255718946456909} 02/26/2022 00:27:20 - INFO - codeparrot_training - Step 36205: {'lr': 9.516227237002758e-05, 'samples': 18537472, 'steps': 36205, 'loss/train': 0.04485290125012398} 02/26/2022 00:27:25 - INFO - codeparrot_training - Step 36206: {'lr': 9.514942629913165e-05, 'samples': 18537984, 'steps': 36206, 'loss/train': 1.8761175870895386} 02/26/2022 00:27:28 - INFO - codeparrot_training - Step 36207: {'lr': 9.513658089156643e-05, 'samples': 18538496, 'steps': 36207, 'loss/train': 1.4736067056655884} 02/26/2022 00:27:35 - INFO - codeparrot_training - Step 36208: {'lr': 9.512373614738681e-05, 'samples': 18539008, 'steps': 36208, 'loss/train': 1.3543826341629028} 02/26/2022 00:27:38 - INFO - codeparrot_training - Step 36209: {'lr': 9.511089206664802e-05, 'samples': 18539520, 'steps': 36209, 'loss/train': 2.117323637008667} 02/26/2022 00:27:44 - INFO - codeparrot_training - Step 36210: {'lr': 9.5098048649405e-05, 'samples': 18540032, 'steps': 36210, 'loss/train': 1.7350062131881714} 02/26/2022 00:27:47 - INFO - codeparrot_training - Step 36211: {'lr': 9.508520589571274e-05, 'samples': 18540544, 'steps': 36211, 'loss/train': 1.5164942741394043} 02/26/2022 00:27:53 - INFO - codeparrot_training - Step 36212: {'lr': 9.50723638056262e-05, 'samples': 18541056, 'steps': 36212, 'loss/train': 1.7740974426269531} 02/26/2022 00:27:56 - INFO - codeparrot_training - Step 36213: {'lr': 9.505952237920051e-05, 'samples': 18541568, 'steps': 36213, 'loss/train': 1.9600930213928223} 02/26/2022 00:28:02 - INFO - codeparrot_training - Step 36214: {'lr': 9.504668161649066e-05, 'samples': 18542080, 'steps': 36214, 'loss/train': 1.6778844594955444} 02/26/2022 00:28:05 - INFO - codeparrot_training - Step 36215: {'lr': 9.503384151755159e-05, 'samples': 18542592, 'steps': 36215, 'loss/train': 1.3575268983840942} 02/26/2022 00:28:11 - INFO - codeparrot_training - Step 36216: {'lr': 9.502100208243828e-05, 'samples': 18543104, 'steps': 36216, 'loss/train': 2.378061532974243} 02/26/2022 00:28:15 - INFO - codeparrot_training - Step 36217: {'lr': 9.500816331120585e-05, 'samples': 18543616, 'steps': 36217, 'loss/train': 1.0308067798614502} 02/26/2022 00:28:21 - INFO - codeparrot_training - Step 36218: {'lr': 9.499532520390925e-05, 'samples': 18544128, 'steps': 36218, 'loss/train': 1.7748454809188843} 02/26/2022 00:28:24 - INFO - codeparrot_training - Step 36219: {'lr': 9.498248776060345e-05, 'samples': 18544640, 'steps': 36219, 'loss/train': 1.5875684022903442} 02/26/2022 00:28:30 - INFO - codeparrot_training - Step 36220: {'lr': 9.496965098134335e-05, 'samples': 18545152, 'steps': 36220, 'loss/train': 1.957951545715332} 02/26/2022 00:28:33 - INFO - codeparrot_training - Step 36221: {'lr': 9.495681486618415e-05, 'samples': 18545664, 'steps': 36221, 'loss/train': 1.8799099922180176} 02/26/2022 00:28:39 - INFO - codeparrot_training - Step 36222: {'lr': 9.494397941518071e-05, 'samples': 18546176, 'steps': 36222, 'loss/train': 1.6717936992645264} 02/26/2022 00:28:43 - INFO - codeparrot_training - Step 36223: {'lr': 9.493114462838795e-05, 'samples': 18546688, 'steps': 36223, 'loss/train': 2.255244731903076} 02/26/2022 00:28:48 - INFO - codeparrot_training - Step 36224: {'lr': 9.491831050586108e-05, 'samples': 18547200, 'steps': 36224, 'loss/train': 2.0785927772521973} 02/26/2022 00:28:52 - INFO - codeparrot_training - Step 36225: {'lr': 9.490547704765476e-05, 'samples': 18547712, 'steps': 36225, 'loss/train': 1.7624056339263916} 02/26/2022 00:28:57 - INFO - codeparrot_training - Step 36226: {'lr': 9.489264425382421e-05, 'samples': 18548224, 'steps': 36226, 'loss/train': 0.7971744537353516} 02/26/2022 00:29:00 - INFO - codeparrot_training - Step 36227: {'lr': 9.487981212442423e-05, 'samples': 18548736, 'steps': 36227, 'loss/train': 2.3726744651794434} 02/26/2022 00:29:07 - INFO - codeparrot_training - Step 36228: {'lr': 9.486698065951008e-05, 'samples': 18549248, 'steps': 36228, 'loss/train': 1.748160481452942} 02/26/2022 00:29:10 - INFO - codeparrot_training - Step 36229: {'lr': 9.485414985913631e-05, 'samples': 18549760, 'steps': 36229, 'loss/train': 1.1634185314178467} 02/26/2022 00:29:16 - INFO - codeparrot_training - Step 36230: {'lr': 9.484131972335822e-05, 'samples': 18550272, 'steps': 36230, 'loss/train': 2.3919410705566406} 02/26/2022 00:29:19 - INFO - codeparrot_training - Step 36231: {'lr': 9.482849025223053e-05, 'samples': 18550784, 'steps': 36231, 'loss/train': 1.634752631187439} 02/26/2022 00:29:25 - INFO - codeparrot_training - Step 36232: {'lr': 9.481566144580853e-05, 'samples': 18551296, 'steps': 36232, 'loss/train': 1.9851620197296143} 02/26/2022 00:29:28 - INFO - codeparrot_training - Step 36233: {'lr': 9.480283330414675e-05, 'samples': 18551808, 'steps': 36233, 'loss/train': 1.6267575025558472} 02/26/2022 00:29:34 - INFO - codeparrot_training - Step 36234: {'lr': 9.479000582730044e-05, 'samples': 18552320, 'steps': 36234, 'loss/train': 1.9866682291030884} 02/26/2022 00:29:37 - INFO - codeparrot_training - Step 36235: {'lr': 9.47771790153244e-05, 'samples': 18552832, 'steps': 36235, 'loss/train': 0.9661385416984558} 02/26/2022 00:29:43 - INFO - codeparrot_training - Step 36236: {'lr': 9.476435286827371e-05, 'samples': 18553344, 'steps': 36236, 'loss/train': 1.5910093784332275} 02/26/2022 00:29:47 - INFO - codeparrot_training - Step 36237: {'lr': 9.475152738620322e-05, 'samples': 18553856, 'steps': 36237, 'loss/train': 3.230879068374634} 02/26/2022 00:29:52 - INFO - codeparrot_training - Step 36238: {'lr': 9.473870256916792e-05, 'samples': 18554368, 'steps': 36238, 'loss/train': 2.672098398208618} 02/26/2022 00:29:56 - INFO - codeparrot_training - Step 36239: {'lr': 9.47258784172226e-05, 'samples': 18554880, 'steps': 36239, 'loss/train': 1.815095067024231} 02/26/2022 00:30:01 - INFO - codeparrot_training - Step 36240: {'lr': 9.471305493042242e-05, 'samples': 18555392, 'steps': 36240, 'loss/train': 0.35066667199134827} 02/26/2022 00:30:05 - INFO - codeparrot_training - Step 36241: {'lr': 9.47002321088222e-05, 'samples': 18555904, 'steps': 36241, 'loss/train': 0.9147576093673706} 02/26/2022 00:30:10 - INFO - codeparrot_training - Step 36242: {'lr': 9.468740995247688e-05, 'samples': 18556416, 'steps': 36242, 'loss/train': 2.4108824729919434} 02/26/2022 00:30:14 - INFO - codeparrot_training - Step 36243: {'lr': 9.467458846144129e-05, 'samples': 18556928, 'steps': 36243, 'loss/train': 1.6959593296051025} 02/26/2022 00:30:20 - INFO - codeparrot_training - Step 36244: {'lr': 9.466176763577051e-05, 'samples': 18557440, 'steps': 36244, 'loss/train': 1.2872647047042847} 02/26/2022 00:30:24 - INFO - codeparrot_training - Step 36245: {'lr': 9.46489474755194e-05, 'samples': 18557952, 'steps': 36245, 'loss/train': 2.546006202697754} 02/26/2022 00:30:29 - INFO - codeparrot_training - Step 36246: {'lr': 9.463612798074286e-05, 'samples': 18558464, 'steps': 36246, 'loss/train': 2.254347324371338} 02/26/2022 00:30:33 - INFO - codeparrot_training - Step 36247: {'lr': 9.462330915149575e-05, 'samples': 18558976, 'steps': 36247, 'loss/train': 1.3925974369049072} 02/26/2022 00:30:38 - INFO - codeparrot_training - Step 36248: {'lr': 9.461049098783312e-05, 'samples': 18559488, 'steps': 36248, 'loss/train': 1.3173164129257202} 02/26/2022 00:30:42 - INFO - codeparrot_training - Step 36249: {'lr': 9.459767348980971e-05, 'samples': 18560000, 'steps': 36249, 'loss/train': 1.2947278022766113} 02/26/2022 00:30:47 - INFO - codeparrot_training - Step 36250: {'lr': 9.458485665748071e-05, 'samples': 18560512, 'steps': 36250, 'loss/train': 1.5315148830413818} 02/26/2022 00:30:51 - INFO - codeparrot_training - Step 36251: {'lr': 9.457204049090065e-05, 'samples': 18561024, 'steps': 36251, 'loss/train': 1.5976450443267822} 02/26/2022 00:30:56 - INFO - codeparrot_training - Step 36252: {'lr': 9.45592249901247e-05, 'samples': 18561536, 'steps': 36252, 'loss/train': 1.2676984071731567} 02/26/2022 00:31:00 - INFO - codeparrot_training - Step 36253: {'lr': 9.45464101552076e-05, 'samples': 18562048, 'steps': 36253, 'loss/train': 2.419299602508545} 02/26/2022 00:31:06 - INFO - codeparrot_training - Step 36254: {'lr': 9.453359598620448e-05, 'samples': 18562560, 'steps': 36254, 'loss/train': 0.5856847167015076} 02/26/2022 00:31:09 - INFO - codeparrot_training - Step 36255: {'lr': 9.452078248316989e-05, 'samples': 18563072, 'steps': 36255, 'loss/train': 1.4574521780014038} 02/26/2022 00:31:15 - INFO - codeparrot_training - Step 36256: {'lr': 9.450796964615902e-05, 'samples': 18563584, 'steps': 36256, 'loss/train': 1.208796501159668} 02/26/2022 00:31:18 - INFO - codeparrot_training - Step 36257: {'lr': 9.449515747522658e-05, 'samples': 18564096, 'steps': 36257, 'loss/train': 2.0554802417755127} 02/26/2022 00:31:24 - INFO - codeparrot_training - Step 36258: {'lr': 9.448234597042754e-05, 'samples': 18564608, 'steps': 36258, 'loss/train': 1.1945327520370483} 02/26/2022 00:31:27 - INFO - codeparrot_training - Step 36259: {'lr': 9.446953513181666e-05, 'samples': 18565120, 'steps': 36259, 'loss/train': 2.031766414642334} 02/26/2022 00:31:33 - INFO - codeparrot_training - Step 36260: {'lr': 9.445672495944899e-05, 'samples': 18565632, 'steps': 36260, 'loss/train': 1.1894862651824951} 02/26/2022 00:31:36 - INFO - codeparrot_training - Step 36261: {'lr': 9.44439154533793e-05, 'samples': 18566144, 'steps': 36261, 'loss/train': 1.7685667276382446} 02/26/2022 00:31:42 - INFO - codeparrot_training - Step 36262: {'lr': 9.443110661366242e-05, 'samples': 18566656, 'steps': 36262, 'loss/train': 2.4416990280151367} 02/26/2022 00:31:45 - INFO - codeparrot_training - Step 36263: {'lr': 9.441829844035335e-05, 'samples': 18567168, 'steps': 36263, 'loss/train': 2.1631531715393066} 02/26/2022 00:31:52 - INFO - codeparrot_training - Step 36264: {'lr': 9.440549093350689e-05, 'samples': 18567680, 'steps': 36264, 'loss/train': 1.252649188041687} 02/26/2022 00:31:57 - INFO - codeparrot_training - Step 36265: {'lr': 9.439268409317791e-05, 'samples': 18568192, 'steps': 36265, 'loss/train': 2.0173165798187256} 02/26/2022 00:32:01 - INFO - codeparrot_training - Step 36266: {'lr': 9.437987791942115e-05, 'samples': 18568704, 'steps': 36266, 'loss/train': 1.5092350244522095} 02/26/2022 00:32:06 - INFO - codeparrot_training - Step 36267: {'lr': 9.43670724122917e-05, 'samples': 18569216, 'steps': 36267, 'loss/train': 0.8754414916038513} 02/26/2022 00:32:10 - INFO - codeparrot_training - Step 36268: {'lr': 9.435426757184426e-05, 'samples': 18569728, 'steps': 36268, 'loss/train': 1.6685688495635986} 02/26/2022 00:32:16 - INFO - codeparrot_training - Step 36269: {'lr': 9.434146339813373e-05, 'samples': 18570240, 'steps': 36269, 'loss/train': 0.416937917470932} 02/26/2022 00:32:19 - INFO - codeparrot_training - Step 36270: {'lr': 9.432865989121487e-05, 'samples': 18570752, 'steps': 36270, 'loss/train': 2.1066877841949463} 02/26/2022 00:32:25 - INFO - codeparrot_training - Step 36271: {'lr': 9.431585705114268e-05, 'samples': 18571264, 'steps': 36271, 'loss/train': 1.7855589389801025} 02/26/2022 00:32:28 - INFO - codeparrot_training - Step 36272: {'lr': 9.430305487797191e-05, 'samples': 18571776, 'steps': 36272, 'loss/train': 1.211093544960022} 02/26/2022 00:32:34 - INFO - codeparrot_training - Step 36273: {'lr': 9.429025337175742e-05, 'samples': 18572288, 'steps': 36273, 'loss/train': 1.1095423698425293} 02/26/2022 00:32:38 - INFO - codeparrot_training - Step 36274: {'lr': 9.427745253255396e-05, 'samples': 18572800, 'steps': 36274, 'loss/train': 1.5531195402145386} 02/26/2022 00:32:43 - INFO - codeparrot_training - Step 36275: {'lr': 9.42646523604165e-05, 'samples': 18573312, 'steps': 36275, 'loss/train': 2.3040332794189453} 02/26/2022 00:32:47 - INFO - codeparrot_training - Step 36276: {'lr': 9.425185285539986e-05, 'samples': 18573824, 'steps': 36276, 'loss/train': 2.5518527030944824} 02/26/2022 00:32:53 - INFO - codeparrot_training - Step 36277: {'lr': 9.423905401755881e-05, 'samples': 18574336, 'steps': 36277, 'loss/train': 1.2939387559890747} 02/26/2022 00:32:56 - INFO - codeparrot_training - Step 36278: {'lr': 9.422625584694811e-05, 'samples': 18574848, 'steps': 36278, 'loss/train': 2.286909818649292} 02/26/2022 00:33:00 - INFO - codeparrot_training - Step 36279: {'lr': 9.421345834362274e-05, 'samples': 18575360, 'steps': 36279, 'loss/train': 2.0417141914367676} 02/26/2022 00:33:05 - INFO - codeparrot_training - Step 36280: {'lr': 9.420066150763748e-05, 'samples': 18575872, 'steps': 36280, 'loss/train': 2.0484418869018555} 02/26/2022 00:33:09 - INFO - codeparrot_training - Step 36281: {'lr': 9.418786533904708e-05, 'samples': 18576384, 'steps': 36281, 'loss/train': 1.7631926536560059} 02/26/2022 00:33:14 - INFO - codeparrot_training - Step 36282: {'lr': 9.41750698379063e-05, 'samples': 18576896, 'steps': 36282, 'loss/train': 1.8416720628738403} 02/26/2022 00:33:18 - INFO - codeparrot_training - Step 36283: {'lr': 9.416227500427015e-05, 'samples': 18577408, 'steps': 36283, 'loss/train': 2.316648006439209} 02/26/2022 00:33:23 - INFO - codeparrot_training - Step 36284: {'lr': 9.414948083819325e-05, 'samples': 18577920, 'steps': 36284, 'loss/train': 1.7622628211975098} 02/26/2022 00:33:27 - INFO - codeparrot_training - Step 36285: {'lr': 9.413668733973063e-05, 'samples': 18578432, 'steps': 36285, 'loss/train': 2.356827735900879} 02/26/2022 00:33:32 - INFO - codeparrot_training - Step 36286: {'lr': 9.412389450893679e-05, 'samples': 18578944, 'steps': 36286, 'loss/train': 1.0693385601043701} 02/26/2022 00:33:36 - INFO - codeparrot_training - Step 36287: {'lr': 9.411110234586679e-05, 'samples': 18579456, 'steps': 36287, 'loss/train': 2.68483567237854} 02/26/2022 00:33:41 - INFO - codeparrot_training - Step 36288: {'lr': 9.409831085057521e-05, 'samples': 18579968, 'steps': 36288, 'loss/train': 2.4817874431610107} 02/26/2022 00:33:45 - INFO - codeparrot_training - Step 36289: {'lr': 9.408552002311716e-05, 'samples': 18580480, 'steps': 36289, 'loss/train': 0.7966066598892212} 02/26/2022 00:33:51 - INFO - codeparrot_training - Step 36290: {'lr': 9.407272986354703e-05, 'samples': 18580992, 'steps': 36290, 'loss/train': 1.9825495481491089} 02/26/2022 00:33:54 - INFO - codeparrot_training - Step 36291: {'lr': 9.405994037191996e-05, 'samples': 18581504, 'steps': 36291, 'loss/train': 0.9061518907546997} 02/26/2022 00:34:00 - INFO - codeparrot_training - Step 36292: {'lr': 9.404715154829044e-05, 'samples': 18582016, 'steps': 36292, 'loss/train': 1.6406949758529663} 02/26/2022 00:34:03 - INFO - codeparrot_training - Step 36293: {'lr': 9.40343633927136e-05, 'samples': 18582528, 'steps': 36293, 'loss/train': 2.3080708980560303} 02/26/2022 00:34:09 - INFO - codeparrot_training - Step 36294: {'lr': 9.402157590524385e-05, 'samples': 18583040, 'steps': 36294, 'loss/train': 1.2550100088119507} 02/26/2022 00:34:12 - INFO - codeparrot_training - Step 36295: {'lr': 9.400878908593621e-05, 'samples': 18583552, 'steps': 36295, 'loss/train': 2.0305874347686768} 02/26/2022 00:34:18 - INFO - codeparrot_training - Step 36296: {'lr': 9.399600293484533e-05, 'samples': 18584064, 'steps': 36296, 'loss/train': 2.0343048572540283} 02/26/2022 00:34:21 - INFO - codeparrot_training - Step 36297: {'lr': 9.39832174520261e-05, 'samples': 18584576, 'steps': 36297, 'loss/train': 3.303342819213867} 02/26/2022 00:34:27 - INFO - codeparrot_training - Step 36298: {'lr': 9.397043263753324e-05, 'samples': 18585088, 'steps': 36298, 'loss/train': 1.1517834663391113} 02/26/2022 00:34:30 - INFO - codeparrot_training - Step 36299: {'lr': 9.39576484914215e-05, 'samples': 18585600, 'steps': 36299, 'loss/train': 1.7336556911468506} 02/26/2022 00:34:37 - INFO - codeparrot_training - Step 36300: {'lr': 9.394486501374555e-05, 'samples': 18586112, 'steps': 36300, 'loss/train': 2.379361867904663} 02/26/2022 00:34:40 - INFO - codeparrot_training - Step 36301: {'lr': 9.393208220456032e-05, 'samples': 18586624, 'steps': 36301, 'loss/train': 1.6731102466583252} 02/26/2022 00:34:46 - INFO - codeparrot_training - Step 36302: {'lr': 9.391930006392052e-05, 'samples': 18587136, 'steps': 36302, 'loss/train': 2.2060508728027344} 02/26/2022 00:34:49 - INFO - codeparrot_training - Step 36303: {'lr': 9.390651859188084e-05, 'samples': 18587648, 'steps': 36303, 'loss/train': 2.2408595085144043} 02/26/2022 00:34:55 - INFO - codeparrot_training - Step 36304: {'lr': 9.389373778849611e-05, 'samples': 18588160, 'steps': 36304, 'loss/train': 1.8166465759277344} 02/26/2022 00:34:58 - INFO - codeparrot_training - Step 36305: {'lr': 9.388095765382094e-05, 'samples': 18588672, 'steps': 36305, 'loss/train': 1.9427070617675781} 02/26/2022 00:35:04 - INFO - codeparrot_training - Step 36306: {'lr': 9.386817818791024e-05, 'samples': 18589184, 'steps': 36306, 'loss/train': 1.088804006576538} 02/26/2022 00:35:07 - INFO - codeparrot_training - Step 36307: {'lr': 9.385539939081872e-05, 'samples': 18589696, 'steps': 36307, 'loss/train': 1.6397230625152588} 02/26/2022 00:35:13 - INFO - codeparrot_training - Step 36308: {'lr': 9.384262126260107e-05, 'samples': 18590208, 'steps': 36308, 'loss/train': 2.6562230587005615} 02/26/2022 00:35:16 - INFO - codeparrot_training - Step 36309: {'lr': 9.382984380331194e-05, 'samples': 18590720, 'steps': 36309, 'loss/train': 1.7649052143096924} 02/26/2022 00:35:23 - INFO - codeparrot_training - Step 36310: {'lr': 9.381706701300627e-05, 'samples': 18591232, 'steps': 36310, 'loss/train': 0.11203889548778534} 02/26/2022 00:35:26 - INFO - codeparrot_training - Step 36311: {'lr': 9.38042908917387e-05, 'samples': 18591744, 'steps': 36311, 'loss/train': 1.8558820486068726} 02/26/2022 00:35:32 - INFO - codeparrot_training - Step 36312: {'lr': 9.379151543956397e-05, 'samples': 18592256, 'steps': 36312, 'loss/train': 2.5380797386169434} 02/26/2022 00:35:35 - INFO - codeparrot_training - Step 36313: {'lr': 9.377874065653665e-05, 'samples': 18592768, 'steps': 36313, 'loss/train': 0.9079977869987488} 02/26/2022 00:35:41 - INFO - codeparrot_training - Step 36314: {'lr': 9.376596654271172e-05, 'samples': 18593280, 'steps': 36314, 'loss/train': 2.2043566703796387} 02/26/2022 00:35:45 - INFO - codeparrot_training - Step 36315: {'lr': 9.375319309814375e-05, 'samples': 18593792, 'steps': 36315, 'loss/train': 2.645874500274658} 02/26/2022 00:35:50 - INFO - codeparrot_training - Step 36316: {'lr': 9.374042032288752e-05, 'samples': 18594304, 'steps': 36316, 'loss/train': 2.0600063800811768} 02/26/2022 00:35:54 - INFO - codeparrot_training - Step 36317: {'lr': 9.372764821699761e-05, 'samples': 18594816, 'steps': 36317, 'loss/train': 1.090671420097351} 02/26/2022 00:35:59 - INFO - codeparrot_training - Step 36318: {'lr': 9.371487678052892e-05, 'samples': 18595328, 'steps': 36318, 'loss/train': 0.9171093702316284} 02/26/2022 00:36:03 - INFO - codeparrot_training - Step 36319: {'lr': 9.370210601353598e-05, 'samples': 18595840, 'steps': 36319, 'loss/train': 1.3418842554092407} 02/26/2022 00:36:08 - INFO - codeparrot_training - Step 36320: {'lr': 9.368933591607378e-05, 'samples': 18596352, 'steps': 36320, 'loss/train': 3.4095373153686523} 02/26/2022 00:36:12 - INFO - codeparrot_training - Step 36321: {'lr': 9.367656648819665e-05, 'samples': 18596864, 'steps': 36321, 'loss/train': 2.2872121334075928} 02/26/2022 00:36:17 - INFO - codeparrot_training - Step 36322: {'lr': 9.366379772995954e-05, 'samples': 18597376, 'steps': 36322, 'loss/train': 2.281195640563965} 02/26/2022 00:36:21 - INFO - codeparrot_training - Step 36323: {'lr': 9.365102964141701e-05, 'samples': 18597888, 'steps': 36323, 'loss/train': 1.2672016620635986} 02/26/2022 00:36:26 - INFO - codeparrot_training - Step 36324: {'lr': 9.363826222262397e-05, 'samples': 18598400, 'steps': 36324, 'loss/train': 1.452994465827942} 02/26/2022 00:36:30 - INFO - codeparrot_training - Step 36325: {'lr': 9.362549547363483e-05, 'samples': 18598912, 'steps': 36325, 'loss/train': 2.0354881286621094} 02/26/2022 00:36:36 - INFO - codeparrot_training - Step 36326: {'lr': 9.36127293945045e-05, 'samples': 18599424, 'steps': 36326, 'loss/train': 1.5614584684371948} 02/26/2022 00:36:39 - INFO - codeparrot_training - Step 36327: {'lr': 9.359996398528745e-05, 'samples': 18599936, 'steps': 36327, 'loss/train': 1.912429928779602} 02/26/2022 00:36:45 - INFO - codeparrot_training - Step 36328: {'lr': 9.35871992460387e-05, 'samples': 18600448, 'steps': 36328, 'loss/train': 0.6544637680053711} 02/26/2022 00:36:48 - INFO - codeparrot_training - Step 36329: {'lr': 9.357443517681252e-05, 'samples': 18600960, 'steps': 36329, 'loss/train': 2.163517951965332} 02/26/2022 00:36:54 - INFO - codeparrot_training - Step 36330: {'lr': 9.356167177766389e-05, 'samples': 18601472, 'steps': 36330, 'loss/train': 1.611802577972412} 02/26/2022 00:36:57 - INFO - codeparrot_training - Step 36331: {'lr': 9.354890904864729e-05, 'samples': 18601984, 'steps': 36331, 'loss/train': 1.443282127380371} 02/26/2022 00:37:03 - INFO - codeparrot_training - Step 36332: {'lr': 9.353614698981761e-05, 'samples': 18602496, 'steps': 36332, 'loss/train': 1.5898233652114868} 02/26/2022 00:37:06 - INFO - codeparrot_training - Step 36333: {'lr': 9.352338560122934e-05, 'samples': 18603008, 'steps': 36333, 'loss/train': 0.5311668515205383} 02/26/2022 00:37:12 - INFO - codeparrot_training - Step 36334: {'lr': 9.351062488293724e-05, 'samples': 18603520, 'steps': 36334, 'loss/train': 1.609014868736267} 02/26/2022 00:37:16 - INFO - codeparrot_training - Step 36335: {'lr': 9.349786483499582e-05, 'samples': 18604032, 'steps': 36335, 'loss/train': 2.0934019088745117} 02/26/2022 00:37:22 - INFO - codeparrot_training - Step 36336: {'lr': 9.348510545745995e-05, 'samples': 18604544, 'steps': 36336, 'loss/train': 1.9578651189804077} 02/26/2022 00:37:25 - INFO - codeparrot_training - Step 36337: {'lr': 9.347234675038419e-05, 'samples': 18605056, 'steps': 36337, 'loss/train': 1.813740849494934} 02/26/2022 00:37:31 - INFO - codeparrot_training - Step 36338: {'lr': 9.345958871382318e-05, 'samples': 18605568, 'steps': 36338, 'loss/train': 1.6773390769958496} 02/26/2022 00:37:34 - INFO - codeparrot_training - Step 36339: {'lr': 9.344683134783149e-05, 'samples': 18606080, 'steps': 36339, 'loss/train': 0.47529885172843933} 02/26/2022 00:37:40 - INFO - codeparrot_training - Step 36340: {'lr': 9.343407465246398e-05, 'samples': 18606592, 'steps': 36340, 'loss/train': 1.8597060441970825} 02/26/2022 00:37:43 - INFO - codeparrot_training - Step 36341: {'lr': 9.342131862777516e-05, 'samples': 18607104, 'steps': 36341, 'loss/train': 2.439831256866455} 02/26/2022 00:37:49 - INFO - codeparrot_training - Step 36342: {'lr': 9.340856327381969e-05, 'samples': 18607616, 'steps': 36342, 'loss/train': 1.724426507949829} 02/26/2022 00:37:52 - INFO - codeparrot_training - Step 36343: {'lr': 9.339580859065214e-05, 'samples': 18608128, 'steps': 36343, 'loss/train': 1.9011871814727783} 02/26/2022 00:37:58 - INFO - codeparrot_training - Step 36344: {'lr': 9.33830545783273e-05, 'samples': 18608640, 'steps': 36344, 'loss/train': 1.2725675106048584} 02/26/2022 00:38:01 - INFO - codeparrot_training - Step 36345: {'lr': 9.337030123689972e-05, 'samples': 18609152, 'steps': 36345, 'loss/train': 1.6196274757385254} 02/26/2022 00:38:07 - INFO - codeparrot_training - Step 36346: {'lr': 9.335754856642404e-05, 'samples': 18609664, 'steps': 36346, 'loss/train': 2.038508892059326} 02/26/2022 00:38:11 - INFO - codeparrot_training - Step 36347: {'lr': 9.334479656695476e-05, 'samples': 18610176, 'steps': 36347, 'loss/train': 1.5932917594909668} 02/26/2022 00:38:16 - INFO - codeparrot_training - Step 36348: {'lr': 9.333204523854678e-05, 'samples': 18610688, 'steps': 36348, 'loss/train': 1.1246910095214844} 02/26/2022 00:38:20 - INFO - codeparrot_training - Step 36349: {'lr': 9.331929458125451e-05, 'samples': 18611200, 'steps': 36349, 'loss/train': 2.1565911769866943} 02/26/2022 00:38:25 - INFO - codeparrot_training - Step 36350: {'lr': 9.330654459513265e-05, 'samples': 18611712, 'steps': 36350, 'loss/train': 1.7111234664916992} 02/26/2022 00:38:31 - INFO - codeparrot_training - Step 36351: {'lr': 9.329379528023574e-05, 'samples': 18612224, 'steps': 36351, 'loss/train': 1.4370098114013672} 02/26/2022 00:38:34 - INFO - codeparrot_training - Step 36352: {'lr': 9.328104663661852e-05, 'samples': 18612736, 'steps': 36352, 'loss/train': 1.0549601316452026} 02/26/2022 00:38:40 - INFO - codeparrot_training - Step 36353: {'lr': 9.326829866433551e-05, 'samples': 18613248, 'steps': 36353, 'loss/train': 0.15012721717357635} 02/26/2022 00:38:43 - INFO - codeparrot_training - Step 36354: {'lr': 9.325555136344135e-05, 'samples': 18613760, 'steps': 36354, 'loss/train': 0.8335407972335815} 02/26/2022 00:38:50 - INFO - codeparrot_training - Step 36355: {'lr': 9.324280473399067e-05, 'samples': 18614272, 'steps': 36355, 'loss/train': 2.3158681392669678} 02/26/2022 00:38:53 - INFO - codeparrot_training - Step 36356: {'lr': 9.323005877603791e-05, 'samples': 18614784, 'steps': 36356, 'loss/train': 0.5626108050346375} 02/26/2022 00:38:59 - INFO - codeparrot_training - Step 36357: {'lr': 9.321731348963788e-05, 'samples': 18615296, 'steps': 36357, 'loss/train': 1.7125240564346313} 02/26/2022 00:39:02 - INFO - codeparrot_training - Step 36358: {'lr': 9.320456887484504e-05, 'samples': 18615808, 'steps': 36358, 'loss/train': 0.712852418422699} 02/26/2022 00:39:08 - INFO - codeparrot_training - Step 36359: {'lr': 9.319182493171419e-05, 'samples': 18616320, 'steps': 36359, 'loss/train': 1.7100962400436401} 02/26/2022 00:39:11 - INFO - codeparrot_training - Step 36360: {'lr': 9.317908166029961e-05, 'samples': 18616832, 'steps': 36360, 'loss/train': 1.8680890798568726} 02/26/2022 00:39:17 - INFO - codeparrot_training - Step 36361: {'lr': 9.316633906065613e-05, 'samples': 18617344, 'steps': 36361, 'loss/train': 0.40120282769203186} 02/26/2022 00:39:20 - INFO - codeparrot_training - Step 36362: {'lr': 9.315359713283817e-05, 'samples': 18617856, 'steps': 36362, 'loss/train': 2.3107306957244873} 02/26/2022 00:39:26 - INFO - codeparrot_training - Step 36363: {'lr': 9.314085587690058e-05, 'samples': 18618368, 'steps': 36363, 'loss/train': 1.7699942588806152} 02/26/2022 00:39:29 - INFO - codeparrot_training - Step 36364: {'lr': 9.312811529289755e-05, 'samples': 18618880, 'steps': 36364, 'loss/train': 2.184447765350342} 02/26/2022 00:39:35 - INFO - codeparrot_training - Step 36365: {'lr': 9.311537538088396e-05, 'samples': 18619392, 'steps': 36365, 'loss/train': 2.7705628871917725} 02/26/2022 00:39:38 - INFO - codeparrot_training - Step 36366: {'lr': 9.310263614091421e-05, 'samples': 18619904, 'steps': 36366, 'loss/train': 1.8837342262268066} 02/26/2022 00:39:44 - INFO - codeparrot_training - Step 36367: {'lr': 9.308989757304303e-05, 'samples': 18620416, 'steps': 36367, 'loss/train': 0.6485057473182678} 02/26/2022 00:39:47 - INFO - codeparrot_training - Step 36368: {'lr': 9.307715967732491e-05, 'samples': 18620928, 'steps': 36368, 'loss/train': 1.4307818412780762} 02/26/2022 00:39:53 - INFO - codeparrot_training - Step 36369: {'lr': 9.306442245381439e-05, 'samples': 18621440, 'steps': 36369, 'loss/train': 2.0255372524261475} 02/26/2022 00:39:56 - INFO - codeparrot_training - Step 36370: {'lr': 9.305168590256599e-05, 'samples': 18621952, 'steps': 36370, 'loss/train': 2.8514492511749268} 02/26/2022 00:40:03 - INFO - codeparrot_training - Step 36371: {'lr': 9.303895002363439e-05, 'samples': 18622464, 'steps': 36371, 'loss/train': 2.1975536346435547} 02/26/2022 00:40:06 - INFO - codeparrot_training - Step 36372: {'lr': 9.30262148170741e-05, 'samples': 18622976, 'steps': 36372, 'loss/train': 1.1395689249038696} 02/26/2022 00:40:12 - INFO - codeparrot_training - Step 36373: {'lr': 9.301348028293965e-05, 'samples': 18623488, 'steps': 36373, 'loss/train': 0.3569713234901428} 02/26/2022 00:40:15 - INFO - codeparrot_training - Step 36374: {'lr': 9.300074642128554e-05, 'samples': 18624000, 'steps': 36374, 'loss/train': 1.5336700677871704} 02/26/2022 00:40:21 - INFO - codeparrot_training - Step 36375: {'lr': 9.298801323216646e-05, 'samples': 18624512, 'steps': 36375, 'loss/train': 2.599266290664673} 02/26/2022 00:40:24 - INFO - codeparrot_training - Step 36376: {'lr': 9.297528071563685e-05, 'samples': 18625024, 'steps': 36376, 'loss/train': 1.296905755996704} 02/26/2022 00:40:30 - INFO - codeparrot_training - Step 36377: {'lr': 9.296254887175132e-05, 'samples': 18625536, 'steps': 36377, 'loss/train': 1.8133147954940796} 02/26/2022 00:40:33 - INFO - codeparrot_training - Step 36378: {'lr': 9.294981770056424e-05, 'samples': 18626048, 'steps': 36378, 'loss/train': 2.8033108711242676} 02/26/2022 00:40:39 - INFO - codeparrot_training - Step 36379: {'lr': 9.293708720213037e-05, 'samples': 18626560, 'steps': 36379, 'loss/train': 0.46975770592689514} 02/26/2022 00:40:42 - INFO - codeparrot_training - Step 36380: {'lr': 9.292435737650406e-05, 'samples': 18627072, 'steps': 36380, 'loss/train': 1.9406933784484863} 02/26/2022 00:40:48 - INFO - codeparrot_training - Step 36381: {'lr': 9.291162822374011e-05, 'samples': 18627584, 'steps': 36381, 'loss/train': 2.463886260986328} 02/26/2022 00:40:52 - INFO - codeparrot_training - Step 36382: {'lr': 9.289889974389268e-05, 'samples': 18628096, 'steps': 36382, 'loss/train': 1.7609314918518066} 02/26/2022 00:40:57 - INFO - codeparrot_training - Step 36383: {'lr': 9.288617193701654e-05, 'samples': 18628608, 'steps': 36383, 'loss/train': 2.3828985691070557} 02/26/2022 00:41:01 - INFO - codeparrot_training - Step 36384: {'lr': 9.287344480316617e-05, 'samples': 18629120, 'steps': 36384, 'loss/train': 1.5279086828231812} 02/26/2022 00:41:06 - INFO - codeparrot_training - Step 36385: {'lr': 9.28607183423961e-05, 'samples': 18629632, 'steps': 36385, 'loss/train': 2.533825159072876} 02/26/2022 00:41:10 - INFO - codeparrot_training - Step 36386: {'lr': 9.28479925547607e-05, 'samples': 18630144, 'steps': 36386, 'loss/train': 0.9208022356033325} 02/26/2022 00:41:16 - INFO - codeparrot_training - Step 36387: {'lr': 9.283526744031467e-05, 'samples': 18630656, 'steps': 36387, 'loss/train': 1.663558840751648} 02/26/2022 00:41:19 - INFO - codeparrot_training - Step 36388: {'lr': 9.282254299911247e-05, 'samples': 18631168, 'steps': 36388, 'loss/train': 3.1090292930603027} 02/26/2022 00:41:25 - INFO - codeparrot_training - Step 36389: {'lr': 9.28098192312086e-05, 'samples': 18631680, 'steps': 36389, 'loss/train': 1.2936574220657349} 02/26/2022 00:41:28 - INFO - codeparrot_training - Step 36390: {'lr': 9.279709613665743e-05, 'samples': 18632192, 'steps': 36390, 'loss/train': 1.8255300521850586} 02/26/2022 00:41:34 - INFO - codeparrot_training - Step 36391: {'lr': 9.278437371551368e-05, 'samples': 18632704, 'steps': 36391, 'loss/train': 1.5036951303482056} 02/26/2022 00:41:38 - INFO - codeparrot_training - Step 36392: {'lr': 9.277165196783177e-05, 'samples': 18633216, 'steps': 36392, 'loss/train': 2.361042022705078} 02/26/2022 00:41:43 - INFO - codeparrot_training - Step 36393: {'lr': 9.275893089366607e-05, 'samples': 18633728, 'steps': 36393, 'loss/train': 1.6003185510635376} 02/26/2022 00:41:47 - INFO - codeparrot_training - Step 36394: {'lr': 9.274621049307128e-05, 'samples': 18634240, 'steps': 36394, 'loss/train': 1.0404809713363647} 02/26/2022 00:41:52 - INFO - codeparrot_training - Step 36395: {'lr': 9.273349076610177e-05, 'samples': 18634752, 'steps': 36395, 'loss/train': 2.1912903785705566} 02/26/2022 00:41:55 - INFO - codeparrot_training - Step 36396: {'lr': 9.272077171281207e-05, 'samples': 18635264, 'steps': 36396, 'loss/train': 0.31599026918411255} 02/26/2022 00:42:01 - INFO - codeparrot_training - Step 36397: {'lr': 9.270805333325655e-05, 'samples': 18635776, 'steps': 36397, 'loss/train': 0.7981106042861938} 02/26/2022 00:42:04 - INFO - codeparrot_training - Step 36398: {'lr': 9.269533562748989e-05, 'samples': 18636288, 'steps': 36398, 'loss/train': 1.8272963762283325} 02/26/2022 00:42:10 - INFO - codeparrot_training - Step 36399: {'lr': 9.268261859556643e-05, 'samples': 18636800, 'steps': 36399, 'loss/train': 1.8302654027938843} 02/26/2022 00:42:13 - INFO - codeparrot_training - Step 36400: {'lr': 9.266990223754068e-05, 'samples': 18637312, 'steps': 36400, 'loss/train': 1.2698239088058472} 02/26/2022 00:42:19 - INFO - codeparrot_training - Step 36401: {'lr': 9.265718655346703e-05, 'samples': 18637824, 'steps': 36401, 'loss/train': 0.3177264630794525} 02/26/2022 00:42:22 - INFO - codeparrot_training - Step 36402: {'lr': 9.264447154340022e-05, 'samples': 18638336, 'steps': 36402, 'loss/train': 1.8804055452346802} 02/26/2022 00:42:29 - INFO - codeparrot_training - Step 36403: {'lr': 9.263175720739434e-05, 'samples': 18638848, 'steps': 36403, 'loss/train': 2.492642641067505} 02/26/2022 00:42:32 - INFO - codeparrot_training - Step 36404: {'lr': 9.261904354550413e-05, 'samples': 18639360, 'steps': 36404, 'loss/train': 1.6978983879089355} 02/26/2022 00:42:38 - INFO - codeparrot_training - Step 36405: {'lr': 9.260633055778389e-05, 'samples': 18639872, 'steps': 36405, 'loss/train': 1.8529647588729858} 02/26/2022 00:42:41 - INFO - codeparrot_training - Step 36406: {'lr': 9.259361824428822e-05, 'samples': 18640384, 'steps': 36406, 'loss/train': 1.9268027544021606} 02/26/2022 00:42:46 - INFO - codeparrot_training - Step 36407: {'lr': 9.258090660507152e-05, 'samples': 18640896, 'steps': 36407, 'loss/train': 1.407084584236145} 02/26/2022 00:42:50 - INFO - codeparrot_training - Step 36408: {'lr': 9.25681956401882e-05, 'samples': 18641408, 'steps': 36408, 'loss/train': 1.5694947242736816} 02/26/2022 00:42:55 - INFO - codeparrot_training - Step 36409: {'lr': 9.255548534969268e-05, 'samples': 18641920, 'steps': 36409, 'loss/train': 1.7173643112182617} 02/26/2022 00:42:59 - INFO - codeparrot_training - Step 36410: {'lr': 9.254277573363957e-05, 'samples': 18642432, 'steps': 36410, 'loss/train': 1.9942305088043213} 02/26/2022 00:43:04 - INFO - codeparrot_training - Step 36411: {'lr': 9.253006679208317e-05, 'samples': 18642944, 'steps': 36411, 'loss/train': 1.5869961977005005} 02/26/2022 00:43:08 - INFO - codeparrot_training - Step 36412: {'lr': 9.251735852507798e-05, 'samples': 18643456, 'steps': 36412, 'loss/train': 1.273870587348938} 02/26/2022 00:43:13 - INFO - codeparrot_training - Step 36413: {'lr': 9.250465093267834e-05, 'samples': 18643968, 'steps': 36413, 'loss/train': 1.7111687660217285} 02/26/2022 00:43:19 - INFO - codeparrot_training - Step 36414: {'lr': 9.249194401493884e-05, 'samples': 18644480, 'steps': 36414, 'loss/train': 1.0950794219970703} 02/26/2022 00:43:22 - INFO - codeparrot_training - Step 36415: {'lr': 9.247923777191386e-05, 'samples': 18644992, 'steps': 36415, 'loss/train': 1.7551685571670532} 02/26/2022 00:43:29 - INFO - codeparrot_training - Step 36416: {'lr': 9.246653220365778e-05, 'samples': 18645504, 'steps': 36416, 'loss/train': 1.7830021381378174} 02/26/2022 00:43:32 - INFO - codeparrot_training - Step 36417: {'lr': 9.245382731022497e-05, 'samples': 18646016, 'steps': 36417, 'loss/train': 1.981549859046936} 02/26/2022 00:43:38 - INFO - codeparrot_training - Step 36418: {'lr': 9.244112309167005e-05, 'samples': 18646528, 'steps': 36418, 'loss/train': 1.9376245737075806} 02/26/2022 00:43:41 - INFO - codeparrot_training - Step 36419: {'lr': 9.24284195480472e-05, 'samples': 18647040, 'steps': 36419, 'loss/train': 5.026699542999268} 02/26/2022 00:43:47 - INFO - codeparrot_training - Step 36420: {'lr': 9.241571667941117e-05, 'samples': 18647552, 'steps': 36420, 'loss/train': 1.808791995048523} 02/26/2022 00:43:50 - INFO - codeparrot_training - Step 36421: {'lr': 9.2403014485816e-05, 'samples': 18648064, 'steps': 36421, 'loss/train': 0.2704356610774994} 02/26/2022 00:43:56 - INFO - codeparrot_training - Step 36422: {'lr': 9.239031296731634e-05, 'samples': 18648576, 'steps': 36422, 'loss/train': 1.584717035293579} 02/26/2022 00:43:59 - INFO - codeparrot_training - Step 36423: {'lr': 9.237761212396648e-05, 'samples': 18649088, 'steps': 36423, 'loss/train': 1.7895535230636597} 02/26/2022 00:44:05 - INFO - codeparrot_training - Step 36424: {'lr': 9.2364911955821e-05, 'samples': 18649600, 'steps': 36424, 'loss/train': 1.4520173072814941} 02/26/2022 00:44:08 - INFO - codeparrot_training - Step 36425: {'lr': 9.235221246293405e-05, 'samples': 18650112, 'steps': 36425, 'loss/train': 1.1449350118637085} 02/26/2022 00:44:14 - INFO - codeparrot_training - Step 36426: {'lr': 9.233951364536025e-05, 'samples': 18650624, 'steps': 36426, 'loss/train': 1.5596548318862915} 02/26/2022 00:44:18 - INFO - codeparrot_training - Step 36427: {'lr': 9.23268155031538e-05, 'samples': 18651136, 'steps': 36427, 'loss/train': 2.1618642807006836} 02/26/2022 00:44:23 - INFO - codeparrot_training - Step 36428: {'lr': 9.23141180363693e-05, 'samples': 18651648, 'steps': 36428, 'loss/train': 0.4900175631046295} 02/26/2022 00:44:27 - INFO - codeparrot_training - Step 36429: {'lr': 9.230142124506105e-05, 'samples': 18652160, 'steps': 36429, 'loss/train': 1.1107112169265747} 02/26/2022 00:44:32 - INFO - codeparrot_training - Step 36430: {'lr': 9.228872512928344e-05, 'samples': 18652672, 'steps': 36430, 'loss/train': 1.955566644668579} 02/26/2022 00:44:36 - INFO - codeparrot_training - Step 36431: {'lr': 9.227602968909077e-05, 'samples': 18653184, 'steps': 36431, 'loss/train': 0.5552517175674438} 02/26/2022 00:44:41 - INFO - codeparrot_training - Step 36432: {'lr': 9.226333492453759e-05, 'samples': 18653696, 'steps': 36432, 'loss/train': 2.1072685718536377} 02/26/2022 00:44:45 - INFO - codeparrot_training - Step 36433: {'lr': 9.225064083567819e-05, 'samples': 18654208, 'steps': 36433, 'loss/train': 1.8738425970077515} 02/26/2022 00:44:50 - INFO - codeparrot_training - Step 36434: {'lr': 9.223794742256694e-05, 'samples': 18654720, 'steps': 36434, 'loss/train': 3.0810461044311523} 02/26/2022 00:44:54 - INFO - codeparrot_training - Step 36435: {'lr': 9.222525468525825e-05, 'samples': 18655232, 'steps': 36435, 'loss/train': 1.5632967948913574} 02/26/2022 00:44:59 - INFO - codeparrot_training - Step 36436: {'lr': 9.221256262380637e-05, 'samples': 18655744, 'steps': 36436, 'loss/train': 8.585833549499512} 02/26/2022 00:45:03 - INFO - codeparrot_training - Step 36437: {'lr': 9.219987123826587e-05, 'samples': 18656256, 'steps': 36437, 'loss/train': 2.015568733215332} 02/26/2022 00:45:09 - INFO - codeparrot_training - Step 36438: {'lr': 9.218718052869099e-05, 'samples': 18656768, 'steps': 36438, 'loss/train': 1.0608787536621094} 02/26/2022 00:45:12 - INFO - codeparrot_training - Step 36439: {'lr': 9.217449049513615e-05, 'samples': 18657280, 'steps': 36439, 'loss/train': 2.932211399078369} 02/26/2022 00:45:18 - INFO - codeparrot_training - Step 36440: {'lr': 9.216180113765556e-05, 'samples': 18657792, 'steps': 36440, 'loss/train': 1.40361487865448} 02/26/2022 00:45:21 - INFO - codeparrot_training - Step 36441: {'lr': 9.21491124563038e-05, 'samples': 18658304, 'steps': 36441, 'loss/train': 2.0718863010406494} 02/26/2022 00:45:27 - INFO - codeparrot_training - Step 36442: {'lr': 9.213642445113513e-05, 'samples': 18658816, 'steps': 36442, 'loss/train': 2.466803789138794} 02/26/2022 00:45:30 - INFO - codeparrot_training - Step 36443: {'lr': 9.212373712220388e-05, 'samples': 18659328, 'steps': 36443, 'loss/train': 1.2844488620758057} 02/26/2022 00:45:36 - INFO - codeparrot_training - Step 36444: {'lr': 9.21110504695643e-05, 'samples': 18659840, 'steps': 36444, 'loss/train': 1.6582192182540894} 02/26/2022 00:45:39 - INFO - codeparrot_training - Step 36445: {'lr': 9.209836449327095e-05, 'samples': 18660352, 'steps': 36445, 'loss/train': 2.2200734615325928} 02/26/2022 00:45:45 - INFO - codeparrot_training - Step 36446: {'lr': 9.208567919337806e-05, 'samples': 18660864, 'steps': 36446, 'loss/train': 1.795745611190796} 02/26/2022 00:45:48 - INFO - codeparrot_training - Step 36447: {'lr': 9.207299456993998e-05, 'samples': 18661376, 'steps': 36447, 'loss/train': 1.5260186195373535} 02/26/2022 00:45:55 - INFO - codeparrot_training - Step 36448: {'lr': 9.206031062301095e-05, 'samples': 18661888, 'steps': 36448, 'loss/train': 1.56020987033844} 02/26/2022 00:45:58 - INFO - codeparrot_training - Step 36449: {'lr': 9.204762735264552e-05, 'samples': 18662400, 'steps': 36449, 'loss/train': 1.3249993324279785} 02/26/2022 00:46:04 - INFO - codeparrot_training - Step 36450: {'lr': 9.203494475889787e-05, 'samples': 18662912, 'steps': 36450, 'loss/train': 0.8002824187278748} 02/26/2022 00:46:07 - INFO - codeparrot_training - Step 36451: {'lr': 9.202226284182236e-05, 'samples': 18663424, 'steps': 36451, 'loss/train': 1.8464992046356201} 02/26/2022 00:46:13 - INFO - codeparrot_training - Step 36452: {'lr': 9.200958160147322e-05, 'samples': 18663936, 'steps': 36452, 'loss/train': 1.8326163291931152} 02/26/2022 00:46:16 - INFO - codeparrot_training - Step 36453: {'lr': 9.199690103790495e-05, 'samples': 18664448, 'steps': 36453, 'loss/train': 2.0581140518188477} 02/26/2022 00:46:22 - INFO - codeparrot_training - Step 36454: {'lr': 9.19842211511717e-05, 'samples': 18664960, 'steps': 36454, 'loss/train': 1.4904792308807373} 02/26/2022 00:46:25 - INFO - codeparrot_training - Step 36455: {'lr': 9.197154194132807e-05, 'samples': 18665472, 'steps': 36455, 'loss/train': 2.4088711738586426} 02/26/2022 00:46:31 - INFO - codeparrot_training - Step 36456: {'lr': 9.195886340842797e-05, 'samples': 18665984, 'steps': 36456, 'loss/train': 1.2211743593215942} 02/26/2022 00:46:34 - INFO - codeparrot_training - Step 36457: {'lr': 9.194618555252601e-05, 'samples': 18666496, 'steps': 36457, 'loss/train': 1.9535833597183228} 02/26/2022 00:46:40 - INFO - codeparrot_training - Step 36458: {'lr': 9.193350837367631e-05, 'samples': 18667008, 'steps': 36458, 'loss/train': 2.4182329177856445} 02/26/2022 00:46:43 - INFO - codeparrot_training - Step 36459: {'lr': 9.192083187193345e-05, 'samples': 18667520, 'steps': 36459, 'loss/train': 2.333261013031006} 02/26/2022 00:46:49 - INFO - codeparrot_training - Step 36460: {'lr': 9.190815604735139e-05, 'samples': 18668032, 'steps': 36460, 'loss/train': 1.7774379253387451} 02/26/2022 00:46:52 - INFO - codeparrot_training - Step 36461: {'lr': 9.189548089998464e-05, 'samples': 18668544, 'steps': 36461, 'loss/train': 0.9777742028236389} 02/26/2022 00:46:58 - INFO - codeparrot_training - Step 36462: {'lr': 9.188280642988738e-05, 'samples': 18669056, 'steps': 36462, 'loss/train': 2.45379900932312} 02/26/2022 00:47:01 - INFO - codeparrot_training - Step 36463: {'lr': 9.187013263711417e-05, 'samples': 18669568, 'steps': 36463, 'loss/train': 2.6046814918518066} 02/26/2022 00:47:08 - INFO - codeparrot_training - Step 36464: {'lr': 9.185745952171889e-05, 'samples': 18670080, 'steps': 36464, 'loss/train': 1.002368688583374} 02/26/2022 00:47:11 - INFO - codeparrot_training - Step 36465: {'lr': 9.184478708375616e-05, 'samples': 18670592, 'steps': 36465, 'loss/train': 2.8651559352874756} 02/26/2022 00:47:17 - INFO - codeparrot_training - Step 36466: {'lr': 9.183211532328e-05, 'samples': 18671104, 'steps': 36466, 'loss/train': 1.9365592002868652} 02/26/2022 00:47:20 - INFO - codeparrot_training - Step 36467: {'lr': 9.181944424034497e-05, 'samples': 18671616, 'steps': 36467, 'loss/train': 1.973009467124939} 02/26/2022 00:47:26 - INFO - codeparrot_training - Step 36468: {'lr': 9.180677383500518e-05, 'samples': 18672128, 'steps': 36468, 'loss/train': 1.2666012048721313} 02/26/2022 00:47:31 - INFO - codeparrot_training - Step 36469: {'lr': 9.179410410731498e-05, 'samples': 18672640, 'steps': 36469, 'loss/train': 0.8861597180366516} 02/26/2022 00:47:35 - INFO - codeparrot_training - Step 36470: {'lr': 9.178143505732847e-05, 'samples': 18673152, 'steps': 36470, 'loss/train': 1.7103712558746338} 02/26/2022 00:47:40 - INFO - codeparrot_training - Step 36471: {'lr': 9.176876668510018e-05, 'samples': 18673664, 'steps': 36471, 'loss/train': 1.3528631925582886} 02/26/2022 00:47:44 - INFO - codeparrot_training - Step 36472: {'lr': 9.175609899068421e-05, 'samples': 18674176, 'steps': 36472, 'loss/train': 2.02816104888916} 02/26/2022 00:47:50 - INFO - codeparrot_training - Step 36473: {'lr': 9.17434319741349e-05, 'samples': 18674688, 'steps': 36473, 'loss/train': 0.8615384697914124} 02/26/2022 00:47:53 - INFO - codeparrot_training - Step 36474: {'lr': 9.173076563550636e-05, 'samples': 18675200, 'steps': 36474, 'loss/train': 2.8255438804626465} 02/26/2022 00:47:59 - INFO - codeparrot_training - Step 36475: {'lr': 9.171809997485305e-05, 'samples': 18675712, 'steps': 36475, 'loss/train': 2.03831148147583} 02/26/2022 00:48:02 - INFO - codeparrot_training - Step 36476: {'lr': 9.170543499222917e-05, 'samples': 18676224, 'steps': 36476, 'loss/train': 1.6320687532424927} 02/26/2022 00:48:08 - INFO - codeparrot_training - Step 36477: {'lr': 9.169277068768891e-05, 'samples': 18676736, 'steps': 36477, 'loss/train': 1.1638494729995728} 02/26/2022 00:48:11 - INFO - codeparrot_training - Step 36478: {'lr': 9.168010706128649e-05, 'samples': 18677248, 'steps': 36478, 'loss/train': 0.7139557003974915} 02/26/2022 00:48:17 - INFO - codeparrot_training - Step 36479: {'lr': 9.166744411307629e-05, 'samples': 18677760, 'steps': 36479, 'loss/train': 2.1932663917541504} 02/26/2022 00:48:20 - INFO - codeparrot_training - Step 36480: {'lr': 9.165478184311248e-05, 'samples': 18678272, 'steps': 36480, 'loss/train': 1.1974139213562012} 02/26/2022 00:48:26 - INFO - codeparrot_training - Step 36481: {'lr': 9.164212025144933e-05, 'samples': 18678784, 'steps': 36481, 'loss/train': 1.3745049238204956} 02/26/2022 00:48:29 - INFO - codeparrot_training - Step 36482: {'lr': 9.162945933814101e-05, 'samples': 18679296, 'steps': 36482, 'loss/train': 1.2544997930526733} 02/26/2022 00:48:36 - INFO - codeparrot_training - Step 36483: {'lr': 9.161679910324175e-05, 'samples': 18679808, 'steps': 36483, 'loss/train': 2.548205852508545} 02/26/2022 00:48:39 - INFO - codeparrot_training - Step 36484: {'lr': 9.160413954680591e-05, 'samples': 18680320, 'steps': 36484, 'loss/train': 1.778460144996643} 02/26/2022 00:48:45 - INFO - codeparrot_training - Step 36485: {'lr': 9.159148066888761e-05, 'samples': 18680832, 'steps': 36485, 'loss/train': 2.4942657947540283} 02/26/2022 00:48:48 - INFO - codeparrot_training - Step 36486: {'lr': 9.157882246954113e-05, 'samples': 18681344, 'steps': 36486, 'loss/train': 0.688693642616272} 02/26/2022 00:48:54 - INFO - codeparrot_training - Step 36487: {'lr': 9.156616494882059e-05, 'samples': 18681856, 'steps': 36487, 'loss/train': 0.7565783262252808} 02/26/2022 00:48:57 - INFO - codeparrot_training - Step 36488: {'lr': 9.155350810678037e-05, 'samples': 18682368, 'steps': 36488, 'loss/train': 1.8944106101989746} 02/26/2022 00:49:03 - INFO - codeparrot_training - Step 36489: {'lr': 9.154085194347453e-05, 'samples': 18682880, 'steps': 36489, 'loss/train': 0.5719866752624512} 02/26/2022 00:49:06 - INFO - codeparrot_training - Step 36490: {'lr': 9.152819645895752e-05, 'samples': 18683392, 'steps': 36490, 'loss/train': 1.1436283588409424} 02/26/2022 00:49:12 - INFO - codeparrot_training - Step 36491: {'lr': 9.151554165328324e-05, 'samples': 18683904, 'steps': 36491, 'loss/train': 2.089611053466797} 02/26/2022 00:49:15 - INFO - codeparrot_training - Step 36492: {'lr': 9.150288752650612e-05, 'samples': 18684416, 'steps': 36492, 'loss/train': 0.33290091156959534} 02/26/2022 00:49:21 - INFO - codeparrot_training - Step 36493: {'lr': 9.149023407868023e-05, 'samples': 18684928, 'steps': 36493, 'loss/train': 1.9095097780227661} 02/26/2022 00:49:25 - INFO - codeparrot_training - Step 36494: {'lr': 9.147758130986004e-05, 'samples': 18685440, 'steps': 36494, 'loss/train': 1.8613122701644897} 02/26/2022 00:49:30 - INFO - codeparrot_training - Step 36495: {'lr': 9.146492922009933e-05, 'samples': 18685952, 'steps': 36495, 'loss/train': 0.5094005465507507} 02/26/2022 00:49:34 - INFO - codeparrot_training - Step 36496: {'lr': 9.145227780945264e-05, 'samples': 18686464, 'steps': 36496, 'loss/train': 0.2982470393180847} 02/26/2022 00:49:40 - INFO - codeparrot_training - Step 36497: {'lr': 9.143962707797396e-05, 'samples': 18686976, 'steps': 36497, 'loss/train': 1.2405245304107666} 02/26/2022 00:49:43 - INFO - codeparrot_training - Step 36498: {'lr': 9.142697702571775e-05, 'samples': 18687488, 'steps': 36498, 'loss/train': 1.6004984378814697} 02/26/2022 00:49:49 - INFO - codeparrot_training - Step 36499: {'lr': 9.141432765273783e-05, 'samples': 18688000, 'steps': 36499, 'loss/train': 2.5078694820404053} 02/26/2022 00:49:52 - INFO - codeparrot_training - Step 36500: {'lr': 9.140167895908866e-05, 'samples': 18688512, 'steps': 36500, 'loss/train': 1.4973552227020264} 02/26/2022 00:49:58 - INFO - codeparrot_training - Step 36501: {'lr': 9.138903094482426e-05, 'samples': 18689024, 'steps': 36501, 'loss/train': 2.0895373821258545} 02/26/2022 00:50:01 - INFO - codeparrot_training - Step 36502: {'lr': 9.137638360999898e-05, 'samples': 18689536, 'steps': 36502, 'loss/train': 1.540652871131897} 02/26/2022 00:50:07 - INFO - codeparrot_training - Step 36503: {'lr': 9.136373695466687e-05, 'samples': 18690048, 'steps': 36503, 'loss/train': 1.1501940488815308} 02/26/2022 00:50:10 - INFO - codeparrot_training - Step 36504: {'lr': 9.135109097888217e-05, 'samples': 18690560, 'steps': 36504, 'loss/train': 2.350374698638916} 02/26/2022 00:50:16 - INFO - codeparrot_training - Step 36505: {'lr': 9.133844568269892e-05, 'samples': 18691072, 'steps': 36505, 'loss/train': 2.0685176849365234} 02/26/2022 00:50:19 - INFO - codeparrot_training - Step 36506: {'lr': 9.13258010661715e-05, 'samples': 18691584, 'steps': 36506, 'loss/train': 1.2391624450683594} 02/26/2022 00:50:26 - INFO - codeparrot_training - Step 36507: {'lr': 9.131315712935392e-05, 'samples': 18692096, 'steps': 36507, 'loss/train': 1.4086748361587524} 02/26/2022 00:50:29 - INFO - codeparrot_training - Step 36508: {'lr': 9.130051387230043e-05, 'samples': 18692608, 'steps': 36508, 'loss/train': 1.7807008028030396} 02/26/2022 00:50:34 - INFO - codeparrot_training - Step 36509: {'lr': 9.128787129506502e-05, 'samples': 18693120, 'steps': 36509, 'loss/train': 1.27669095993042} 02/26/2022 00:50:38 - INFO - codeparrot_training - Step 36510: {'lr': 9.127522939770208e-05, 'samples': 18693632, 'steps': 36510, 'loss/train': 1.4267017841339111} 02/26/2022 00:50:43 - INFO - codeparrot_training - Step 36511: {'lr': 9.126258818026565e-05, 'samples': 18694144, 'steps': 36511, 'loss/train': 1.5238736867904663} 02/26/2022 00:50:47 - INFO - codeparrot_training - Step 36512: {'lr': 9.124994764280989e-05, 'samples': 18694656, 'steps': 36512, 'loss/train': 1.1365036964416504} 02/26/2022 00:50:52 - INFO - codeparrot_training - Step 36513: {'lr': 9.123730778538886e-05, 'samples': 18695168, 'steps': 36513, 'loss/train': 1.2731398344039917} 02/26/2022 00:50:56 - INFO - codeparrot_training - Step 36514: {'lr': 9.122466860805687e-05, 'samples': 18695680, 'steps': 36514, 'loss/train': 1.5848681926727295} 02/26/2022 00:51:01 - INFO - codeparrot_training - Step 36515: {'lr': 9.121203011086799e-05, 'samples': 18696192, 'steps': 36515, 'loss/train': 2.597829818725586} 02/26/2022 00:51:05 - INFO - codeparrot_training - Step 36516: {'lr': 9.119939229387636e-05, 'samples': 18696704, 'steps': 36516, 'loss/train': 0.7429189085960388} 02/26/2022 00:51:10 - INFO - codeparrot_training - Step 36517: {'lr': 9.1186755157136e-05, 'samples': 18697216, 'steps': 36517, 'loss/train': 1.1402100324630737} 02/26/2022 00:51:14 - INFO - codeparrot_training - Step 36518: {'lr': 9.117411870070128e-05, 'samples': 18697728, 'steps': 36518, 'loss/train': 0.9129608869552612} 02/26/2022 00:51:20 - INFO - codeparrot_training - Step 36519: {'lr': 9.116148292462614e-05, 'samples': 18698240, 'steps': 36519, 'loss/train': 0.587259829044342} 02/26/2022 00:51:23 - INFO - codeparrot_training - Step 36520: {'lr': 9.114884782896482e-05, 'samples': 18698752, 'steps': 36520, 'loss/train': 1.1782442331314087} 02/26/2022 00:51:29 - INFO - codeparrot_training - Step 36521: {'lr': 9.113621341377129e-05, 'samples': 18699264, 'steps': 36521, 'loss/train': 2.414233684539795} 02/26/2022 00:51:32 - INFO - codeparrot_training - Step 36522: {'lr': 9.112357967909988e-05, 'samples': 18699776, 'steps': 36522, 'loss/train': 0.5109829306602478} 02/26/2022 00:51:38 - INFO - codeparrot_training - Step 36523: {'lr': 9.111094662500458e-05, 'samples': 18700288, 'steps': 36523, 'loss/train': 0.5842905640602112} 02/26/2022 00:51:41 - INFO - codeparrot_training - Step 36524: {'lr': 9.109831425153956e-05, 'samples': 18700800, 'steps': 36524, 'loss/train': 2.1260931491851807} 02/26/2022 00:51:46 - INFO - codeparrot_training - Step 36525: {'lr': 9.10856825587588e-05, 'samples': 18701312, 'steps': 36525, 'loss/train': 1.952371597290039} 02/26/2022 00:51:50 - INFO - codeparrot_training - Step 36526: {'lr': 9.107305154671658e-05, 'samples': 18701824, 'steps': 36526, 'loss/train': 1.3889007568359375} 02/26/2022 00:51:55 - INFO - codeparrot_training - Step 36527: {'lr': 9.106042121546698e-05, 'samples': 18702336, 'steps': 36527, 'loss/train': 1.3719687461853027} 02/26/2022 00:51:59 - INFO - codeparrot_training - Step 36528: {'lr': 9.104779156506395e-05, 'samples': 18702848, 'steps': 36528, 'loss/train': 2.0373573303222656} 02/26/2022 00:52:07 - INFO - codeparrot_training - Step 36529: {'lr': 9.10351625955619e-05, 'samples': 18703360, 'steps': 36529, 'loss/train': 1.4478174448013306} 02/26/2022 00:52:10 - INFO - codeparrot_training - Step 36530: {'lr': 9.102253430701458e-05, 'samples': 18703872, 'steps': 36530, 'loss/train': 1.9549286365509033} 02/26/2022 00:52:16 - INFO - codeparrot_training - Step 36531: {'lr': 9.100990669947629e-05, 'samples': 18704384, 'steps': 36531, 'loss/train': 1.2932859659194946} 02/26/2022 00:52:19 - INFO - codeparrot_training - Step 36532: {'lr': 9.099727977300101e-05, 'samples': 18704896, 'steps': 36532, 'loss/train': 0.42060935497283936} 02/26/2022 00:52:25 - INFO - codeparrot_training - Step 36533: {'lr': 9.098465352764308e-05, 'samples': 18705408, 'steps': 36533, 'loss/train': 2.394883871078491} 02/26/2022 00:52:28 - INFO - codeparrot_training - Step 36534: {'lr': 9.097202796345619e-05, 'samples': 18705920, 'steps': 36534, 'loss/train': 1.678720474243164} 02/26/2022 00:52:34 - INFO - codeparrot_training - Step 36535: {'lr': 9.095940308049474e-05, 'samples': 18706432, 'steps': 36535, 'loss/train': 2.3407063484191895} 02/26/2022 00:52:37 - INFO - codeparrot_training - Step 36536: {'lr': 9.094677887881264e-05, 'samples': 18706944, 'steps': 36536, 'loss/train': 2.2407546043395996} 02/26/2022 00:52:43 - INFO - codeparrot_training - Step 36537: {'lr': 9.093415535846411e-05, 'samples': 18707456, 'steps': 36537, 'loss/train': 1.5770275592803955} 02/26/2022 00:52:46 - INFO - codeparrot_training - Step 36538: {'lr': 9.092153251950314e-05, 'samples': 18707968, 'steps': 36538, 'loss/train': 1.9819976091384888} 02/26/2022 00:52:52 - INFO - codeparrot_training - Step 36539: {'lr': 9.090891036198381e-05, 'samples': 18708480, 'steps': 36539, 'loss/train': 2.4673635959625244} 02/26/2022 00:52:55 - INFO - codeparrot_training - Step 36540: {'lr': 9.089628888596011e-05, 'samples': 18708992, 'steps': 36540, 'loss/train': 1.967557668685913} 02/26/2022 00:53:02 - INFO - codeparrot_training - Step 36541: {'lr': 9.088366809148627e-05, 'samples': 18709504, 'steps': 36541, 'loss/train': 1.6730071306228638} 02/26/2022 00:53:06 - INFO - codeparrot_training - Step 36542: {'lr': 9.087104797861628e-05, 'samples': 18710016, 'steps': 36542, 'loss/train': 1.5909526348114014} 02/26/2022 00:53:11 - INFO - codeparrot_training - Step 36543: {'lr': 9.085842854740418e-05, 'samples': 18710528, 'steps': 36543, 'loss/train': 1.136153221130371} 02/26/2022 00:53:15 - INFO - codeparrot_training - Step 36544: {'lr': 9.084580979790396e-05, 'samples': 18711040, 'steps': 36544, 'loss/train': 1.3035250902175903} 02/26/2022 00:53:20 - INFO - codeparrot_training - Step 36545: {'lr': 9.083319173016986e-05, 'samples': 18711552, 'steps': 36545, 'loss/train': 1.2188856601715088} 02/26/2022 00:53:24 - INFO - codeparrot_training - Step 36546: {'lr': 9.082057434425578e-05, 'samples': 18712064, 'steps': 36546, 'loss/train': 0.9534711837768555} 02/26/2022 00:53:29 - INFO - codeparrot_training - Step 36547: {'lr': 9.080795764021585e-05, 'samples': 18712576, 'steps': 36547, 'loss/train': 2.044180393218994} 02/26/2022 00:53:33 - INFO - codeparrot_training - Step 36548: {'lr': 9.079534161810396e-05, 'samples': 18713088, 'steps': 36548, 'loss/train': 1.6012495756149292} 02/26/2022 00:53:39 - INFO - codeparrot_training - Step 36549: {'lr': 9.07827262779744e-05, 'samples': 18713600, 'steps': 36549, 'loss/train': 1.5406208038330078} 02/26/2022 00:53:42 - INFO - codeparrot_training - Step 36550: {'lr': 9.077011161988097e-05, 'samples': 18714112, 'steps': 36550, 'loss/train': 1.481308937072754} 02/26/2022 00:53:49 - INFO - codeparrot_training - Step 36551: {'lr': 9.0757497643878e-05, 'samples': 18714624, 'steps': 36551, 'loss/train': 2.161729574203491} 02/26/2022 00:53:55 - INFO - codeparrot_training - Step 36552: {'lr': 9.074488435001918e-05, 'samples': 18715136, 'steps': 36552, 'loss/train': 2.0254125595092773} 02/26/2022 00:53:58 - INFO - codeparrot_training - Step 36553: {'lr': 9.07322717383588e-05, 'samples': 18715648, 'steps': 36553, 'loss/train': 2.609917640686035} 02/26/2022 00:54:04 - INFO - codeparrot_training - Step 36554: {'lr': 9.071965980895069e-05, 'samples': 18716160, 'steps': 36554, 'loss/train': 2.134484052658081} 02/26/2022 00:54:07 - INFO - codeparrot_training - Step 36555: {'lr': 9.070704856184913e-05, 'samples': 18716672, 'steps': 36555, 'loss/train': 2.4393458366394043} 02/26/2022 00:54:13 - INFO - codeparrot_training - Step 36556: {'lr': 9.069443799710786e-05, 'samples': 18717184, 'steps': 36556, 'loss/train': 2.3811662197113037} 02/26/2022 00:54:16 - INFO - codeparrot_training - Step 36557: {'lr': 9.068182811478113e-05, 'samples': 18717696, 'steps': 36557, 'loss/train': 1.608520269393921} 02/26/2022 00:54:22 - INFO - codeparrot_training - Step 36558: {'lr': 9.066921891492275e-05, 'samples': 18718208, 'steps': 36558, 'loss/train': 0.580219030380249} 02/26/2022 00:54:25 - INFO - codeparrot_training - Step 36559: {'lr': 9.065661039758699e-05, 'samples': 18718720, 'steps': 36559, 'loss/train': 1.437660574913025} 02/26/2022 00:54:32 - INFO - codeparrot_training - Step 36560: {'lr': 9.064400256282756e-05, 'samples': 18719232, 'steps': 36560, 'loss/train': 1.2620149850845337} 02/26/2022 00:54:36 - INFO - codeparrot_training - Step 36561: {'lr': 9.063139541069872e-05, 'samples': 18719744, 'steps': 36561, 'loss/train': 0.777113676071167} 02/26/2022 00:54:41 - INFO - codeparrot_training - Step 36562: {'lr': 9.061878894125439e-05, 'samples': 18720256, 'steps': 36562, 'loss/train': 2.1638152599334717} 02/26/2022 00:54:45 - INFO - codeparrot_training - Step 36563: {'lr': 9.060618315454843e-05, 'samples': 18720768, 'steps': 36563, 'loss/train': 1.0005766153335571} 02/26/2022 00:54:50 - INFO - codeparrot_training - Step 36564: {'lr': 9.059357805063509e-05, 'samples': 18721280, 'steps': 36564, 'loss/train': 1.9134806394577026} 02/26/2022 00:54:54 - INFO - codeparrot_training - Step 36565: {'lr': 9.058097362956825e-05, 'samples': 18721792, 'steps': 36565, 'loss/train': 1.4803436994552612} 02/26/2022 00:54:59 - INFO - codeparrot_training - Step 36566: {'lr': 9.056836989140185e-05, 'samples': 18722304, 'steps': 36566, 'loss/train': 2.0330381393432617} 02/26/2022 00:55:03 - INFO - codeparrot_training - Step 36567: {'lr': 9.055576683618987e-05, 'samples': 18722816, 'steps': 36567, 'loss/train': 1.4669479131698608} 02/26/2022 00:55:08 - INFO - codeparrot_training - Step 36568: {'lr': 9.054316446398647e-05, 'samples': 18723328, 'steps': 36568, 'loss/train': 1.5934340953826904} 02/26/2022 00:55:12 - INFO - codeparrot_training - Step 36569: {'lr': 9.053056277484547e-05, 'samples': 18723840, 'steps': 36569, 'loss/train': 1.7027442455291748} 02/26/2022 00:55:20 - INFO - codeparrot_training - Step 36570: {'lr': 9.051796176882093e-05, 'samples': 18724352, 'steps': 36570, 'loss/train': 0.7654800415039062} 02/26/2022 00:55:23 - INFO - codeparrot_training - Step 36571: {'lr': 9.050536144596672e-05, 'samples': 18724864, 'steps': 36571, 'loss/train': 1.9175941944122314} 02/26/2022 00:55:29 - INFO - codeparrot_training - Step 36572: {'lr': 9.049276180633698e-05, 'samples': 18725376, 'steps': 36572, 'loss/train': 2.3417036533355713} 02/26/2022 00:55:32 - INFO - codeparrot_training - Step 36573: {'lr': 9.048016284998559e-05, 'samples': 18725888, 'steps': 36573, 'loss/train': 1.097623348236084} 02/26/2022 00:55:38 - INFO - codeparrot_training - Step 36574: {'lr': 9.046756457696653e-05, 'samples': 18726400, 'steps': 36574, 'loss/train': 1.2813806533813477} 02/26/2022 00:55:41 - INFO - codeparrot_training - Step 36575: {'lr': 9.045496698733368e-05, 'samples': 18726912, 'steps': 36575, 'loss/train': 2.026169538497925} 02/26/2022 00:55:47 - INFO - codeparrot_training - Step 36576: {'lr': 9.044237008114117e-05, 'samples': 18727424, 'steps': 36576, 'loss/train': 1.244621992111206} 02/26/2022 00:55:50 - INFO - codeparrot_training - Step 36577: {'lr': 9.042977385844289e-05, 'samples': 18727936, 'steps': 36577, 'loss/train': 0.2353709638118744} 02/26/2022 00:55:56 - INFO - codeparrot_training - Step 36578: {'lr': 9.04171783192928e-05, 'samples': 18728448, 'steps': 36578, 'loss/train': 2.1739213466644287} 02/26/2022 00:55:59 - INFO - codeparrot_training - Step 36579: {'lr': 9.040458346374475e-05, 'samples': 18728960, 'steps': 36579, 'loss/train': 2.2177698612213135} 02/26/2022 00:56:05 - INFO - codeparrot_training - Step 36580: {'lr': 9.039198929185286e-05, 'samples': 18729472, 'steps': 36580, 'loss/train': 1.8375288248062134} 02/26/2022 00:56:08 - INFO - codeparrot_training - Step 36581: {'lr': 9.037939580367102e-05, 'samples': 18729984, 'steps': 36581, 'loss/train': 2.5287203788757324} 02/26/2022 00:56:14 - INFO - codeparrot_training - Step 36582: {'lr': 9.036680299925315e-05, 'samples': 18730496, 'steps': 36582, 'loss/train': 1.8921520709991455} 02/26/2022 00:56:17 - INFO - codeparrot_training - Step 36583: {'lr': 9.035421087865314e-05, 'samples': 18731008, 'steps': 36583, 'loss/train': 1.2827168703079224} 02/26/2022 00:56:23 - INFO - codeparrot_training - Step 36584: {'lr': 9.034161944192506e-05, 'samples': 18731520, 'steps': 36584, 'loss/train': 2.0836844444274902} 02/26/2022 00:56:26 - INFO - codeparrot_training - Step 36585: {'lr': 9.032902868912274e-05, 'samples': 18732032, 'steps': 36585, 'loss/train': 2.06073260307312} 02/26/2022 00:56:33 - INFO - codeparrot_training - Step 36586: {'lr': 9.031643862030028e-05, 'samples': 18732544, 'steps': 36586, 'loss/train': 2.0865678787231445} 02/26/2022 00:56:37 - INFO - codeparrot_training - Step 36587: {'lr': 9.030384923551136e-05, 'samples': 18733056, 'steps': 36587, 'loss/train': 0.4919186234474182} 02/26/2022 00:56:42 - INFO - codeparrot_training - Step 36588: {'lr': 9.02912605348101e-05, 'samples': 18733568, 'steps': 36588, 'loss/train': 1.92879056930542} 02/26/2022 00:56:46 - INFO - codeparrot_training - Step 36589: {'lr': 9.027867251825031e-05, 'samples': 18734080, 'steps': 36589, 'loss/train': 1.7919151782989502} 02/26/2022 00:56:51 - INFO - codeparrot_training - Step 36590: {'lr': 9.026608518588612e-05, 'samples': 18734592, 'steps': 36590, 'loss/train': 1.3220763206481934} 02/26/2022 00:56:55 - INFO - codeparrot_training - Step 36591: {'lr': 9.025349853777113e-05, 'samples': 18735104, 'steps': 36591, 'loss/train': 0.5083885788917542} 02/26/2022 00:57:00 - INFO - codeparrot_training - Step 36592: {'lr': 9.02409125739595e-05, 'samples': 18735616, 'steps': 36592, 'loss/train': 1.3132753372192383} 02/26/2022 00:57:04 - INFO - codeparrot_training - Step 36593: {'lr': 9.0228327294505e-05, 'samples': 18736128, 'steps': 36593, 'loss/train': 0.5286920666694641} 02/26/2022 00:57:09 - INFO - codeparrot_training - Step 36594: {'lr': 9.021574269946179e-05, 'samples': 18736640, 'steps': 36594, 'loss/train': 0.41230759024620056} 02/26/2022 00:57:13 - INFO - codeparrot_training - Step 36595: {'lr': 9.020315878888344e-05, 'samples': 18737152, 'steps': 36595, 'loss/train': 0.9003009796142578} 02/26/2022 00:57:20 - INFO - codeparrot_training - Step 36596: {'lr': 9.019057556282406e-05, 'samples': 18737664, 'steps': 36596, 'loss/train': 1.450286626815796} 02/26/2022 00:57:24 - INFO - codeparrot_training - Step 36597: {'lr': 9.017799302133747e-05, 'samples': 18738176, 'steps': 36597, 'loss/train': 2.3623502254486084} 02/26/2022 00:57:29 - INFO - codeparrot_training - Step 36598: {'lr': 9.016541116447766e-05, 'samples': 18738688, 'steps': 36598, 'loss/train': 0.5349457263946533} 02/26/2022 00:57:33 - INFO - codeparrot_training - Step 36599: {'lr': 9.015282999229848e-05, 'samples': 18739200, 'steps': 36599, 'loss/train': 1.5680148601531982} 02/26/2022 00:57:38 - INFO - codeparrot_training - Step 36600: {'lr': 9.014024950485383e-05, 'samples': 18739712, 'steps': 36600, 'loss/train': 1.4445619583129883} 02/26/2022 00:57:41 - INFO - codeparrot_training - Step 36601: {'lr': 9.012766970219752e-05, 'samples': 18740224, 'steps': 36601, 'loss/train': 2.0936193466186523} 02/26/2022 00:57:47 - INFO - codeparrot_training - Step 36602: {'lr': 9.011509058438358e-05, 'samples': 18740736, 'steps': 36602, 'loss/train': 1.6281620264053345} 02/26/2022 00:57:51 - INFO - codeparrot_training - Step 36603: {'lr': 9.010251215146581e-05, 'samples': 18741248, 'steps': 36603, 'loss/train': 1.5327023267745972} 02/26/2022 00:57:56 - INFO - codeparrot_training - Step 36604: {'lr': 9.008993440349811e-05, 'samples': 18741760, 'steps': 36604, 'loss/train': 1.826502799987793} 02/26/2022 00:58:00 - INFO - codeparrot_training - Step 36605: {'lr': 9.007735734053426e-05, 'samples': 18742272, 'steps': 36605, 'loss/train': 0.7369265556335449} 02/26/2022 00:58:07 - INFO - codeparrot_training - Step 36606: {'lr': 9.006478096262836e-05, 'samples': 18742784, 'steps': 36606, 'loss/train': 1.9530997276306152} 02/26/2022 00:58:11 - INFO - codeparrot_training - Step 36607: {'lr': 9.00522052698341e-05, 'samples': 18743296, 'steps': 36607, 'loss/train': 1.1178377866744995} 02/26/2022 00:58:16 - INFO - codeparrot_training - Step 36608: {'lr': 9.003963026220543e-05, 'samples': 18743808, 'steps': 36608, 'loss/train': 1.75657057762146} 02/26/2022 00:58:20 - INFO - codeparrot_training - Step 36609: {'lr': 9.002705593979618e-05, 'samples': 18744320, 'steps': 36609, 'loss/train': 2.128976345062256} 02/26/2022 00:58:25 - INFO - codeparrot_training - Step 36610: {'lr': 9.001448230266015e-05, 'samples': 18744832, 'steps': 36610, 'loss/train': 1.520991325378418} 02/26/2022 00:58:28 - INFO - codeparrot_training - Step 36611: {'lr': 9.000190935085137e-05, 'samples': 18745344, 'steps': 36611, 'loss/train': 1.533866047859192} 02/26/2022 00:58:34 - INFO - codeparrot_training - Step 36612: {'lr': 8.998933708442359e-05, 'samples': 18745856, 'steps': 36612, 'loss/train': 2.0014543533325195} 02/26/2022 00:58:37 - INFO - codeparrot_training - Step 36613: {'lr': 8.997676550343067e-05, 'samples': 18746368, 'steps': 36613, 'loss/train': 1.734210729598999} 02/26/2022 00:58:43 - INFO - codeparrot_training - Step 36614: {'lr': 8.996419460792641e-05, 'samples': 18746880, 'steps': 36614, 'loss/train': 2.078361749649048} 02/26/2022 00:58:46 - INFO - codeparrot_training - Step 36615: {'lr': 8.995162439796478e-05, 'samples': 18747392, 'steps': 36615, 'loss/train': 0.3727642595767975} 02/26/2022 00:58:54 - INFO - codeparrot_training - Step 36616: {'lr': 8.99390548735996e-05, 'samples': 18747904, 'steps': 36616, 'loss/train': 1.3749072551727295} 02/26/2022 00:58:57 - INFO - codeparrot_training - Step 36617: {'lr': 8.992648603488468e-05, 'samples': 18748416, 'steps': 36617, 'loss/train': 0.5264908075332642} 02/26/2022 00:59:03 - INFO - codeparrot_training - Step 36618: {'lr': 8.991391788187376e-05, 'samples': 18748928, 'steps': 36618, 'loss/train': 2.08695125579834} 02/26/2022 00:59:06 - INFO - codeparrot_training - Step 36619: {'lr': 8.990135041462088e-05, 'samples': 18749440, 'steps': 36619, 'loss/train': 1.4376933574676514} 02/26/2022 00:59:12 - INFO - codeparrot_training - Step 36620: {'lr': 8.98887836331798e-05, 'samples': 18749952, 'steps': 36620, 'loss/train': 2.2350635528564453} 02/26/2022 00:59:15 - INFO - codeparrot_training - Step 36621: {'lr': 8.987621753760431e-05, 'samples': 18750464, 'steps': 36621, 'loss/train': 1.6088550090789795} 02/26/2022 00:59:21 - INFO - codeparrot_training - Step 36622: {'lr': 8.986365212794815e-05, 'samples': 18750976, 'steps': 36622, 'loss/train': 1.5629996061325073} 02/26/2022 00:59:24 - INFO - codeparrot_training - Step 36623: {'lr': 8.985108740426538e-05, 'samples': 18751488, 'steps': 36623, 'loss/train': 1.1186065673828125} 02/26/2022 00:59:30 - INFO - codeparrot_training - Step 36624: {'lr': 8.983852336660957e-05, 'samples': 18752000, 'steps': 36624, 'loss/train': 1.2251043319702148} 02/26/2022 00:59:33 - INFO - codeparrot_training - Step 36625: {'lr': 8.982596001503488e-05, 'samples': 18752512, 'steps': 36625, 'loss/train': 1.1691792011260986} 02/26/2022 00:59:39 - INFO - codeparrot_training - Step 36626: {'lr': 8.981339734959473e-05, 'samples': 18753024, 'steps': 36626, 'loss/train': 0.4061290919780731} 02/26/2022 00:59:43 - INFO - codeparrot_training - Step 36627: {'lr': 8.98008353703432e-05, 'samples': 18753536, 'steps': 36627, 'loss/train': 0.4246062934398651} 02/26/2022 00:59:48 - INFO - codeparrot_training - Step 36628: {'lr': 8.978827407733395e-05, 'samples': 18754048, 'steps': 36628, 'loss/train': 2.015063524246216} 02/26/2022 00:59:52 - INFO - codeparrot_training - Step 36629: {'lr': 8.977571347062102e-05, 'samples': 18754560, 'steps': 36629, 'loss/train': 1.1896742582321167} 02/26/2022 00:59:55 - INFO - codeparrot_training - Step 36630: {'lr': 8.97631535502579e-05, 'samples': 18755072, 'steps': 36630, 'loss/train': 1.0434242486953735} 02/26/2022 01:00:01 - INFO - codeparrot_training - Step 36631: {'lr': 8.975059431629862e-05, 'samples': 18755584, 'steps': 36631, 'loss/train': 0.6546241044998169} 02/26/2022 01:00:08 - INFO - codeparrot_training - Step 36632: {'lr': 8.973803576879683e-05, 'samples': 18756096, 'steps': 36632, 'loss/train': 2.605837106704712} 02/26/2022 01:00:11 - INFO - codeparrot_training - Step 36633: {'lr': 8.972547790780658e-05, 'samples': 18756608, 'steps': 36633, 'loss/train': 1.5168198347091675} 02/26/2022 01:00:17 - INFO - codeparrot_training - Step 36634: {'lr': 8.971292073338133e-05, 'samples': 18757120, 'steps': 36634, 'loss/train': 0.7329345941543579} 02/26/2022 01:00:20 - INFO - codeparrot_training - Step 36635: {'lr': 8.970036424557509e-05, 'samples': 18757632, 'steps': 36635, 'loss/train': 1.1443909406661987} 02/26/2022 01:00:26 - INFO - codeparrot_training - Step 36636: {'lr': 8.968780844444152e-05, 'samples': 18758144, 'steps': 36636, 'loss/train': 0.39262306690216064} 02/26/2022 01:00:29 - INFO - codeparrot_training - Step 36637: {'lr': 8.967525333003455e-05, 'samples': 18758656, 'steps': 36637, 'loss/train': 1.4619505405426025} 02/26/2022 01:00:35 - INFO - codeparrot_training - Step 36638: {'lr': 8.966269890240788e-05, 'samples': 18759168, 'steps': 36638, 'loss/train': 2.2366981506347656} 02/26/2022 01:00:38 - INFO - codeparrot_training - Step 36639: {'lr': 8.965014516161532e-05, 'samples': 18759680, 'steps': 36639, 'loss/train': 1.7728203535079956} 02/26/2022 01:00:44 - INFO - codeparrot_training - Step 36640: {'lr': 8.963759210771053e-05, 'samples': 18760192, 'steps': 36640, 'loss/train': 2.0320136547088623} 02/26/2022 01:00:47 - INFO - codeparrot_training - Step 36641: {'lr': 8.962503974074743e-05, 'samples': 18760704, 'steps': 36641, 'loss/train': 2.209712266921997} 02/26/2022 01:00:55 - INFO - codeparrot_training - Step 36642: {'lr': 8.961248806077976e-05, 'samples': 18761216, 'steps': 36642, 'loss/train': 1.7144718170166016} 02/26/2022 01:00:58 - INFO - codeparrot_training - Step 36643: {'lr': 8.959993706786126e-05, 'samples': 18761728, 'steps': 36643, 'loss/train': 1.9846676588058472} 02/26/2022 01:01:04 - INFO - codeparrot_training - Step 36644: {'lr': 8.958738676204562e-05, 'samples': 18762240, 'steps': 36644, 'loss/train': 2.2778046131134033} 02/26/2022 01:01:07 - INFO - codeparrot_training - Step 36645: {'lr': 8.957483714338674e-05, 'samples': 18762752, 'steps': 36645, 'loss/train': 1.777929425239563} 02/26/2022 01:01:13 - INFO - codeparrot_training - Step 36646: {'lr': 8.956228821193835e-05, 'samples': 18763264, 'steps': 36646, 'loss/train': 1.7810192108154297} 02/26/2022 01:01:16 - INFO - codeparrot_training - Step 36647: {'lr': 8.954973996775415e-05, 'samples': 18763776, 'steps': 36647, 'loss/train': 1.401321291923523} 02/26/2022 01:01:22 - INFO - codeparrot_training - Step 36648: {'lr': 8.953719241088781e-05, 'samples': 18764288, 'steps': 36648, 'loss/train': 1.5986754894256592} 02/26/2022 01:01:25 - INFO - codeparrot_training - Step 36649: {'lr': 8.952464554139331e-05, 'samples': 18764800, 'steps': 36649, 'loss/train': 1.8167191743850708} 02/26/2022 01:01:31 - INFO - codeparrot_training - Step 36650: {'lr': 8.951209935932425e-05, 'samples': 18765312, 'steps': 36650, 'loss/train': 0.5065308809280396} 02/26/2022 01:01:34 - INFO - codeparrot_training - Step 36651: {'lr': 8.949955386473441e-05, 'samples': 18765824, 'steps': 36651, 'loss/train': 1.5306487083435059} 02/26/2022 01:01:41 - INFO - codeparrot_training - Step 36652: {'lr': 8.94870090576774e-05, 'samples': 18766336, 'steps': 36652, 'loss/train': 2.083111524581909} 02/26/2022 01:01:45 - INFO - codeparrot_training - Step 36653: {'lr': 8.947446493820719e-05, 'samples': 18766848, 'steps': 36653, 'loss/train': 2.302685499191284} 02/26/2022 01:01:50 - INFO - codeparrot_training - Step 36654: {'lr': 8.946192150637739e-05, 'samples': 18767360, 'steps': 36654, 'loss/train': 2.792943000793457} 02/26/2022 01:01:54 - INFO - codeparrot_training - Step 36655: {'lr': 8.944937876224173e-05, 'samples': 18767872, 'steps': 36655, 'loss/train': 1.550115942955017} 02/26/2022 01:01:59 - INFO - codeparrot_training - Step 36656: {'lr': 8.94368367058539e-05, 'samples': 18768384, 'steps': 36656, 'loss/train': 0.9503949880599976} 02/26/2022 01:02:03 - INFO - codeparrot_training - Step 36657: {'lr': 8.942429533726773e-05, 'samples': 18768896, 'steps': 36657, 'loss/train': 2.2719216346740723} 02/26/2022 01:02:08 - INFO - codeparrot_training - Step 36658: {'lr': 8.941175465653692e-05, 'samples': 18769408, 'steps': 36658, 'loss/train': 1.3227165937423706} 02/26/2022 01:02:12 - INFO - codeparrot_training - Step 36659: {'lr': 8.939921466371503e-05, 'samples': 18769920, 'steps': 36659, 'loss/train': 2.4320244789123535} 02/26/2022 01:02:17 - INFO - codeparrot_training - Step 36660: {'lr': 8.938667535885611e-05, 'samples': 18770432, 'steps': 36660, 'loss/train': 1.371164321899414} 02/26/2022 01:02:21 - INFO - codeparrot_training - Step 36661: {'lr': 8.937413674201351e-05, 'samples': 18770944, 'steps': 36661, 'loss/train': 2.8989126682281494} 02/26/2022 01:02:26 - INFO - codeparrot_training - Step 36662: {'lr': 8.936159881324118e-05, 'samples': 18771456, 'steps': 36662, 'loss/train': 1.17244291305542} 02/26/2022 01:02:30 - INFO - codeparrot_training - Step 36663: {'lr': 8.934906157259265e-05, 'samples': 18771968, 'steps': 36663, 'loss/train': 2.791363477706909} 02/26/2022 01:02:37 - INFO - codeparrot_training - Step 36664: {'lr': 8.93365250201219e-05, 'samples': 18772480, 'steps': 36664, 'loss/train': 1.1157643795013428} 02/26/2022 01:02:40 - INFO - codeparrot_training - Step 36665: {'lr': 8.932398915588232e-05, 'samples': 18772992, 'steps': 36665, 'loss/train': 1.331910252571106} 02/26/2022 01:02:46 - INFO - codeparrot_training - Step 36666: {'lr': 8.931145397992781e-05, 'samples': 18773504, 'steps': 36666, 'loss/train': 2.3801474571228027} 02/26/2022 01:02:50 - INFO - codeparrot_training - Step 36667: {'lr': 8.929891949231195e-05, 'samples': 18774016, 'steps': 36667, 'loss/train': 0.6971074938774109} 02/26/2022 01:02:55 - INFO - codeparrot_training - Step 36668: {'lr': 8.928638569308862e-05, 'samples': 18774528, 'steps': 36668, 'loss/train': 1.557456374168396} 02/26/2022 01:02:59 - INFO - codeparrot_training - Step 36669: {'lr': 8.927385258231124e-05, 'samples': 18775040, 'steps': 36669, 'loss/train': 1.8067389726638794} 02/26/2022 01:03:04 - INFO - codeparrot_training - Step 36670: {'lr': 8.926132016003371e-05, 'samples': 18775552, 'steps': 36670, 'loss/train': 2.0618882179260254} 02/26/2022 01:03:08 - INFO - codeparrot_training - Step 36671: {'lr': 8.924878842630957e-05, 'samples': 18776064, 'steps': 36671, 'loss/train': 2.746779441833496} 02/26/2022 01:03:13 - INFO - codeparrot_training - Step 36672: {'lr': 8.923625738119265e-05, 'samples': 18776576, 'steps': 36672, 'loss/train': 2.2812204360961914} 02/26/2022 01:03:17 - INFO - codeparrot_training - Step 36673: {'lr': 8.922372702473655e-05, 'samples': 18777088, 'steps': 36673, 'loss/train': 2.902531385421753} 02/26/2022 01:03:22 - INFO - codeparrot_training - Step 36674: {'lr': 8.921119735699496e-05, 'samples': 18777600, 'steps': 36674, 'loss/train': 2.0630908012390137} 02/26/2022 01:03:26 - INFO - codeparrot_training - Step 36675: {'lr': 8.919866837802146e-05, 'samples': 18778112, 'steps': 36675, 'loss/train': 1.6360176801681519} 02/26/2022 01:03:31 - INFO - codeparrot_training - Step 36676: {'lr': 8.91861400878699e-05, 'samples': 18778624, 'steps': 36676, 'loss/train': 2.1341984272003174} 02/26/2022 01:03:35 - INFO - codeparrot_training - Step 36677: {'lr': 8.917361248659384e-05, 'samples': 18779136, 'steps': 36677, 'loss/train': 1.956908106803894} 02/26/2022 01:03:42 - INFO - codeparrot_training - Step 36678: {'lr': 8.916108557424696e-05, 'samples': 18779648, 'steps': 36678, 'loss/train': 2.20556378364563} 02/26/2022 01:03:45 - INFO - codeparrot_training - Step 36679: {'lr': 8.914855935088281e-05, 'samples': 18780160, 'steps': 36679, 'loss/train': 0.5835480093955994} 02/26/2022 01:03:51 - INFO - codeparrot_training - Step 36680: {'lr': 8.913603381655528e-05, 'samples': 18780672, 'steps': 36680, 'loss/train': 1.3446823358535767} 02/26/2022 01:03:54 - INFO - codeparrot_training - Step 36681: {'lr': 8.912350897131788e-05, 'samples': 18781184, 'steps': 36681, 'loss/train': 1.9582816362380981} 02/26/2022 01:04:00 - INFO - codeparrot_training - Step 36682: {'lr': 8.911098481522429e-05, 'samples': 18781696, 'steps': 36682, 'loss/train': 2.449564218521118} 02/26/2022 01:04:03 - INFO - codeparrot_training - Step 36683: {'lr': 8.909846134832805e-05, 'samples': 18782208, 'steps': 36683, 'loss/train': 1.4189239740371704} 02/26/2022 01:04:09 - INFO - codeparrot_training - Step 36684: {'lr': 8.908593857068301e-05, 'samples': 18782720, 'steps': 36684, 'loss/train': 1.3596614599227905} 02/26/2022 01:04:12 - INFO - codeparrot_training - Step 36685: {'lr': 8.907341648234262e-05, 'samples': 18783232, 'steps': 36685, 'loss/train': 2.7265517711639404} 02/26/2022 01:04:18 - INFO - codeparrot_training - Step 36686: {'lr': 8.906089508336079e-05, 'samples': 18783744, 'steps': 36686, 'loss/train': 2.148716926574707} 02/26/2022 01:04:21 - INFO - codeparrot_training - Step 36687: {'lr': 8.904837437379082e-05, 'samples': 18784256, 'steps': 36687, 'loss/train': 0.6689602136611938} 02/26/2022 01:04:29 - INFO - codeparrot_training - Step 36688: {'lr': 8.903585435368658e-05, 'samples': 18784768, 'steps': 36688, 'loss/train': 1.9057968854904175} 02/26/2022 01:04:32 - INFO - codeparrot_training - Step 36689: {'lr': 8.902333502310154e-05, 'samples': 18785280, 'steps': 36689, 'loss/train': 0.8908526301383972} 02/26/2022 01:04:38 - INFO - codeparrot_training - Step 36690: {'lr': 8.90108163820896e-05, 'samples': 18785792, 'steps': 36690, 'loss/train': 1.8692876100540161} 02/26/2022 01:04:41 - INFO - codeparrot_training - Step 36691: {'lr': 8.899829843070404e-05, 'samples': 18786304, 'steps': 36691, 'loss/train': 1.663933515548706} 02/26/2022 01:04:47 - INFO - codeparrot_training - Step 36692: {'lr': 8.89857811689987e-05, 'samples': 18786816, 'steps': 36692, 'loss/train': 0.9016863107681274} 02/26/2022 01:04:50 - INFO - codeparrot_training - Step 36693: {'lr': 8.897326459702715e-05, 'samples': 18787328, 'steps': 36693, 'loss/train': 2.07425856590271} 02/26/2022 01:04:56 - INFO - codeparrot_training - Step 36694: {'lr': 8.896074871484295e-05, 'samples': 18787840, 'steps': 36694, 'loss/train': 0.817865788936615} 02/26/2022 01:04:59 - INFO - codeparrot_training - Step 36695: {'lr': 8.894823352249982e-05, 'samples': 18788352, 'steps': 36695, 'loss/train': 1.2361661195755005} 02/26/2022 01:05:05 - INFO - codeparrot_training - Step 36696: {'lr': 8.893571902005132e-05, 'samples': 18788864, 'steps': 36696, 'loss/train': 1.9286699295043945} 02/26/2022 01:05:08 - INFO - codeparrot_training - Step 36697: {'lr': 8.892320520755107e-05, 'samples': 18789376, 'steps': 36697, 'loss/train': 1.6497316360473633} 02/26/2022 01:05:14 - INFO - codeparrot_training - Step 36698: {'lr': 8.891069208505256e-05, 'samples': 18789888, 'steps': 36698, 'loss/train': 1.2243497371673584} 02/26/2022 01:05:17 - INFO - codeparrot_training - Step 36699: {'lr': 8.88981796526096e-05, 'samples': 18790400, 'steps': 36699, 'loss/train': 2.0028512477874756} 02/26/2022 01:05:23 - INFO - codeparrot_training - Step 36700: {'lr': 8.88856679102757e-05, 'samples': 18790912, 'steps': 36700, 'loss/train': 1.0738894939422607} 02/26/2022 01:05:27 - INFO - codeparrot_training - Step 36701: {'lr': 8.887315685810437e-05, 'samples': 18791424, 'steps': 36701, 'loss/train': 0.9110260605812073} 02/26/2022 01:05:32 - INFO - codeparrot_training - Step 36702: {'lr': 8.886064649614924e-05, 'samples': 18791936, 'steps': 36702, 'loss/train': 1.0963810682296753} 02/26/2022 01:05:36 - INFO - codeparrot_training - Step 36703: {'lr': 8.884813682446402e-05, 'samples': 18792448, 'steps': 36703, 'loss/train': 1.3647912740707397} 02/26/2022 01:05:41 - INFO - codeparrot_training - Step 36704: {'lr': 8.88356278431022e-05, 'samples': 18792960, 'steps': 36704, 'loss/train': 1.3876621723175049} 02/26/2022 01:05:45 - INFO - codeparrot_training - Step 36705: {'lr': 8.882311955211736e-05, 'samples': 18793472, 'steps': 36705, 'loss/train': 0.4826810657978058} 02/26/2022 01:05:50 - INFO - codeparrot_training - Step 36706: {'lr': 8.881061195156303e-05, 'samples': 18793984, 'steps': 36706, 'loss/train': 3.0438857078552246} 02/26/2022 01:05:56 - INFO - codeparrot_training - Step 36707: {'lr': 8.879810504149293e-05, 'samples': 18794496, 'steps': 36707, 'loss/train': 1.1649420261383057} 02/26/2022 01:05:59 - INFO - codeparrot_training - Step 36708: {'lr': 8.878559882196058e-05, 'samples': 18795008, 'steps': 36708, 'loss/train': 1.2484266757965088} 02/26/2022 01:06:06 - INFO - codeparrot_training - Step 36709: {'lr': 8.87730932930195e-05, 'samples': 18795520, 'steps': 36709, 'loss/train': 2.1707355976104736} 02/26/2022 01:06:09 - INFO - codeparrot_training - Step 36710: {'lr': 8.876058845472323e-05, 'samples': 18796032, 'steps': 36710, 'loss/train': 2.2836008071899414} 02/26/2022 01:06:15 - INFO - codeparrot_training - Step 36711: {'lr': 8.874808430712547e-05, 'samples': 18796544, 'steps': 36711, 'loss/train': 0.06388260424137115} 02/26/2022 01:06:18 - INFO - codeparrot_training - Step 36712: {'lr': 8.873558085027975e-05, 'samples': 18797056, 'steps': 36712, 'loss/train': 1.3427910804748535} 02/26/2022 01:06:24 - INFO - codeparrot_training - Step 36713: {'lr': 8.872307808423954e-05, 'samples': 18797568, 'steps': 36713, 'loss/train': 1.4716030359268188} 02/26/2022 01:06:27 - INFO - codeparrot_training - Step 36714: {'lr': 8.871057600905841e-05, 'samples': 18798080, 'steps': 36714, 'loss/train': 1.0129289627075195} 02/26/2022 01:06:33 - INFO - codeparrot_training - Step 36715: {'lr': 8.869807462479001e-05, 'samples': 18798592, 'steps': 36715, 'loss/train': 1.4471043348312378} 02/26/2022 01:06:36 - INFO - codeparrot_training - Step 36716: {'lr': 8.868557393148787e-05, 'samples': 18799104, 'steps': 36716, 'loss/train': 2.0507829189300537} 02/26/2022 01:06:42 - INFO - codeparrot_training - Step 36717: {'lr': 8.867307392920551e-05, 'samples': 18799616, 'steps': 36717, 'loss/train': 0.945280134677887} 02/26/2022 01:06:45 - INFO - codeparrot_training - Step 36718: {'lr': 8.866057461799637e-05, 'samples': 18800128, 'steps': 36718, 'loss/train': 1.4961342811584473} 02/26/2022 01:06:51 - INFO - codeparrot_training - Step 36719: {'lr': 8.864807599791419e-05, 'samples': 18800640, 'steps': 36719, 'loss/train': 2.179086446762085} 02/26/2022 01:06:54 - INFO - codeparrot_training - Step 36720: {'lr': 8.863557806901232e-05, 'samples': 18801152, 'steps': 36720, 'loss/train': 0.6030438542366028} 02/26/2022 01:07:00 - INFO - codeparrot_training - Step 36721: {'lr': 8.862308083134457e-05, 'samples': 18801664, 'steps': 36721, 'loss/train': 3.1088778972625732} 02/26/2022 01:07:03 - INFO - codeparrot_training - Step 36722: {'lr': 8.861058428496413e-05, 'samples': 18802176, 'steps': 36722, 'loss/train': 1.686672329902649} 02/26/2022 01:07:10 - INFO - codeparrot_training - Step 36723: {'lr': 8.859808842992478e-05, 'samples': 18802688, 'steps': 36723, 'loss/train': 1.7433375120162964} 02/26/2022 01:07:13 - INFO - codeparrot_training - Step 36724: {'lr': 8.858559326627991e-05, 'samples': 18803200, 'steps': 36724, 'loss/train': 1.9560959339141846} 02/26/2022 01:07:19 - INFO - codeparrot_training - Step 36725: {'lr': 8.857309879408323e-05, 'samples': 18803712, 'steps': 36725, 'loss/train': 0.5793426632881165} 02/26/2022 01:07:23 - INFO - codeparrot_training - Step 36726: {'lr': 8.8560605013388e-05, 'samples': 18804224, 'steps': 36726, 'loss/train': 1.6196162700653076} 02/26/2022 01:07:28 - INFO - codeparrot_training - Step 36727: {'lr': 8.854811192424797e-05, 'samples': 18804736, 'steps': 36727, 'loss/train': 2.195920467376709} 02/26/2022 01:07:32 - INFO - codeparrot_training - Step 36728: {'lr': 8.853561952671646e-05, 'samples': 18805248, 'steps': 36728, 'loss/train': 2.598487615585327} 02/26/2022 01:07:37 - INFO - codeparrot_training - Step 36729: {'lr': 8.852312782084723e-05, 'samples': 18805760, 'steps': 36729, 'loss/train': 1.9568482637405396} 02/26/2022 01:07:41 - INFO - codeparrot_training - Step 36730: {'lr': 8.85106368066935e-05, 'samples': 18806272, 'steps': 36730, 'loss/train': 2.3036105632781982} 02/26/2022 01:07:46 - INFO - codeparrot_training - Step 36731: {'lr': 8.849814648430901e-05, 'samples': 18806784, 'steps': 36731, 'loss/train': 1.8940600156784058} 02/26/2022 01:07:50 - INFO - codeparrot_training - Step 36732: {'lr': 8.84856568537471e-05, 'samples': 18807296, 'steps': 36732, 'loss/train': 0.8471187949180603} 02/26/2022 01:07:55 - INFO - codeparrot_training - Step 36733: {'lr': 8.84731679150614e-05, 'samples': 18807808, 'steps': 36733, 'loss/train': 1.2096806764602661} 02/26/2022 01:07:59 - INFO - codeparrot_training - Step 36734: {'lr': 8.84606796683054e-05, 'samples': 18808320, 'steps': 36734, 'loss/train': 1.26979660987854} 02/26/2022 01:08:05 - INFO - codeparrot_training - Step 36735: {'lr': 8.844819211353252e-05, 'samples': 18808832, 'steps': 36735, 'loss/train': 0.9659348726272583} 02/26/2022 01:08:09 - INFO - codeparrot_training - Step 36736: {'lr': 8.843570525079623e-05, 'samples': 18809344, 'steps': 36736, 'loss/train': 1.104225754737854} 02/26/2022 01:08:14 - INFO - codeparrot_training - Step 36737: {'lr': 8.842321908015017e-05, 'samples': 18809856, 'steps': 36737, 'loss/train': 2.7877025604248047} 02/26/2022 01:08:18 - INFO - codeparrot_training - Step 36738: {'lr': 8.841073360164769e-05, 'samples': 18810368, 'steps': 36738, 'loss/train': 1.5737988948822021} 02/26/2022 01:08:23 - INFO - codeparrot_training - Step 36739: {'lr': 8.839824881534234e-05, 'samples': 18810880, 'steps': 36739, 'loss/train': 0.966422975063324} 02/26/2022 01:08:27 - INFO - codeparrot_training - Step 36740: {'lr': 8.838576472128757e-05, 'samples': 18811392, 'steps': 36740, 'loss/train': 1.9636882543563843} 02/26/2022 01:08:32 - INFO - codeparrot_training - Step 36741: {'lr': 8.837328131953679e-05, 'samples': 18811904, 'steps': 36741, 'loss/train': 1.583749771118164} 02/26/2022 01:08:36 - INFO - codeparrot_training - Step 36742: {'lr': 8.836079861014365e-05, 'samples': 18812416, 'steps': 36742, 'loss/train': 2.1364831924438477} 02/26/2022 01:08:41 - INFO - codeparrot_training - Step 36743: {'lr': 8.83483165931615e-05, 'samples': 18812928, 'steps': 36743, 'loss/train': 1.0170072317123413} 02/26/2022 01:08:45 - INFO - codeparrot_training - Step 36744: {'lr': 8.833583526864383e-05, 'samples': 18813440, 'steps': 36744, 'loss/train': 1.1249419450759888} 02/26/2022 01:08:52 - INFO - codeparrot_training - Step 36745: {'lr': 8.832335463664404e-05, 'samples': 18813952, 'steps': 36745, 'loss/train': 1.7335306406021118} 02/26/2022 01:08:55 - INFO - codeparrot_training - Step 36746: {'lr': 8.831087469721571e-05, 'samples': 18814464, 'steps': 36746, 'loss/train': 1.799824595451355} 02/26/2022 01:09:01 - INFO - codeparrot_training - Step 36747: {'lr': 8.82983954504123e-05, 'samples': 18814976, 'steps': 36747, 'loss/train': 1.6073668003082275} 02/26/2022 01:09:04 - INFO - codeparrot_training - Step 36748: {'lr': 8.828591689628718e-05, 'samples': 18815488, 'steps': 36748, 'loss/train': 1.3555233478546143} 02/26/2022 01:09:10 - INFO - codeparrot_training - Step 36749: {'lr': 8.827343903489377e-05, 'samples': 18816000, 'steps': 36749, 'loss/train': 1.284003734588623} 02/26/2022 01:09:13 - INFO - codeparrot_training - Step 36750: {'lr': 8.826096186628569e-05, 'samples': 18816512, 'steps': 36750, 'loss/train': 1.3173454999923706} 02/26/2022 01:09:19 - INFO - codeparrot_training - Step 36751: {'lr': 8.824848539051627e-05, 'samples': 18817024, 'steps': 36751, 'loss/train': 1.413901448249817} 02/26/2022 01:09:22 - INFO - codeparrot_training - Step 36752: {'lr': 8.8236009607639e-05, 'samples': 18817536, 'steps': 36752, 'loss/train': 1.5600212812423706} 02/26/2022 01:09:27 - INFO - codeparrot_training - Step 36753: {'lr': 8.822353451770718e-05, 'samples': 18818048, 'steps': 36753, 'loss/train': 0.9365037083625793} 02/26/2022 01:09:31 - INFO - codeparrot_training - Step 36754: {'lr': 8.821106012077449e-05, 'samples': 18818560, 'steps': 36754, 'loss/train': 1.6110575199127197} 02/26/2022 01:09:37 - INFO - codeparrot_training - Step 36755: {'lr': 8.819858641689413e-05, 'samples': 18819072, 'steps': 36755, 'loss/train': 0.7587934136390686} 02/26/2022 01:09:41 - INFO - codeparrot_training - Step 36756: {'lr': 8.818611340611985e-05, 'samples': 18819584, 'steps': 36756, 'loss/train': 2.1823596954345703} 02/26/2022 01:09:46 - INFO - codeparrot_training - Step 36757: {'lr': 8.817364108850468e-05, 'samples': 18820096, 'steps': 36757, 'loss/train': 0.9404318928718567} 02/26/2022 01:09:50 - INFO - codeparrot_training - Step 36758: {'lr': 8.816116946410232e-05, 'samples': 18820608, 'steps': 36758, 'loss/train': 3.9288594722747803} 02/26/2022 01:09:55 - INFO - codeparrot_training - Step 36759: {'lr': 8.814869853296609e-05, 'samples': 18821120, 'steps': 36759, 'loss/train': 1.5254809856414795} 02/26/2022 01:09:59 - INFO - codeparrot_training - Step 36760: {'lr': 8.813622829514956e-05, 'samples': 18821632, 'steps': 36760, 'loss/train': 1.4608752727508545} 02/26/2022 01:10:04 - INFO - codeparrot_training - Step 36761: {'lr': 8.812375875070588e-05, 'samples': 18822144, 'steps': 36761, 'loss/train': 1.009140968322754} 02/26/2022 01:10:08 - INFO - codeparrot_training - Step 36762: {'lr': 8.81112898996887e-05, 'samples': 18822656, 'steps': 36762, 'loss/train': 1.6667149066925049} 02/26/2022 01:10:13 - INFO - codeparrot_training - Step 36763: {'lr': 8.809882174215125e-05, 'samples': 18823168, 'steps': 36763, 'loss/train': 1.8655192852020264} 02/26/2022 01:10:17 - INFO - codeparrot_training - Step 36764: {'lr': 8.808635427814723e-05, 'samples': 18823680, 'steps': 36764, 'loss/train': 1.089234471321106} 02/26/2022 01:10:22 - INFO - codeparrot_training - Step 36765: {'lr': 8.807388750772968e-05, 'samples': 18824192, 'steps': 36765, 'loss/train': 1.9756888151168823} 02/26/2022 01:10:26 - INFO - codeparrot_training - Step 36766: {'lr': 8.806142143095228e-05, 'samples': 18824704, 'steps': 36766, 'loss/train': 1.5609052181243896} 02/26/2022 01:10:31 - INFO - codeparrot_training - Step 36767: {'lr': 8.80489560478682e-05, 'samples': 18825216, 'steps': 36767, 'loss/train': 1.8127546310424805} 02/26/2022 01:10:35 - INFO - codeparrot_training - Step 36768: {'lr': 8.803649135853109e-05, 'samples': 18825728, 'steps': 36768, 'loss/train': 1.8154958486557007} 02/26/2022 01:10:40 - INFO - codeparrot_training - Step 36769: {'lr': 8.802402736299419e-05, 'samples': 18826240, 'steps': 36769, 'loss/train': 1.4463987350463867} 02/26/2022 01:10:47 - INFO - codeparrot_training - Step 36770: {'lr': 8.801156406131095e-05, 'samples': 18826752, 'steps': 36770, 'loss/train': 0.7040348649024963} 02/26/2022 01:10:50 - INFO - codeparrot_training - Step 36771: {'lr': 8.799910145353466e-05, 'samples': 18827264, 'steps': 36771, 'loss/train': 0.79520183801651} 02/26/2022 01:10:54 - INFO - codeparrot_training - Step 36772: {'lr': 8.798663953971884e-05, 'samples': 18827776, 'steps': 36772, 'loss/train': 1.9044753313064575} 02/26/2022 01:10:59 - INFO - codeparrot_training - Step 36773: {'lr': 8.79741783199168e-05, 'samples': 18828288, 'steps': 36773, 'loss/train': 1.3131247758865356} 02/26/2022 01:11:03 - INFO - codeparrot_training - Step 36774: {'lr': 8.796171779418194e-05, 'samples': 18828800, 'steps': 36774, 'loss/train': 1.9512815475463867} 02/26/2022 01:11:09 - INFO - codeparrot_training - Step 36775: {'lr': 8.794925796256756e-05, 'samples': 18829312, 'steps': 36775, 'loss/train': 1.364993929862976} 02/26/2022 01:11:12 - INFO - codeparrot_training - Step 36776: {'lr': 8.793679882512717e-05, 'samples': 18829824, 'steps': 36776, 'loss/train': 0.08782120048999786} 02/26/2022 01:11:18 - INFO - codeparrot_training - Step 36777: {'lr': 8.792434038191408e-05, 'samples': 18830336, 'steps': 36777, 'loss/train': 1.3393462896347046} 02/26/2022 01:11:21 - INFO - codeparrot_training - Step 36778: {'lr': 8.791188263298164e-05, 'samples': 18830848, 'steps': 36778, 'loss/train': 1.8954521417617798} 02/26/2022 01:11:26 - INFO - codeparrot_training - Step 36779: {'lr': 8.789942557838316e-05, 'samples': 18831360, 'steps': 36779, 'loss/train': 1.1856162548065186} 02/26/2022 01:11:33 - INFO - codeparrot_training - Step 36780: {'lr': 8.788696921817213e-05, 'samples': 18831872, 'steps': 36780, 'loss/train': 2.0256221294403076} 02/26/2022 01:11:37 - INFO - codeparrot_training - Step 36781: {'lr': 8.787451355240186e-05, 'samples': 18832384, 'steps': 36781, 'loss/train': 1.7585585117340088} 02/26/2022 01:11:40 - INFO - codeparrot_training - Step 36782: {'lr': 8.78620585811257e-05, 'samples': 18832896, 'steps': 36782, 'loss/train': 2.2875008583068848} 02/26/2022 01:11:46 - INFO - codeparrot_training - Step 36783: {'lr': 8.784960430439689e-05, 'samples': 18833408, 'steps': 36783, 'loss/train': 2.3009443283081055} 02/26/2022 01:11:49 - INFO - codeparrot_training - Step 36784: {'lr': 8.7837150722269e-05, 'samples': 18833920, 'steps': 36784, 'loss/train': 3.1083948612213135} 02/26/2022 01:11:55 - INFO - codeparrot_training - Step 36785: {'lr': 8.782469783479524e-05, 'samples': 18834432, 'steps': 36785, 'loss/train': 1.133135199546814} 02/26/2022 01:11:58 - INFO - codeparrot_training - Step 36786: {'lr': 8.7812245642029e-05, 'samples': 18834944, 'steps': 36786, 'loss/train': 2.467921733856201} 02/26/2022 01:12:04 - INFO - codeparrot_training - Step 36787: {'lr': 8.779979414402359e-05, 'samples': 18835456, 'steps': 36787, 'loss/train': 2.1750290393829346} 02/26/2022 01:12:07 - INFO - codeparrot_training - Step 36788: {'lr': 8.778734334083226e-05, 'samples': 18835968, 'steps': 36788, 'loss/train': 2.211529016494751} 02/26/2022 01:12:13 - INFO - codeparrot_training - Step 36789: {'lr': 8.777489323250854e-05, 'samples': 18836480, 'steps': 36789, 'loss/train': 1.9000332355499268} 02/26/2022 01:12:16 - INFO - codeparrot_training - Step 36790: {'lr': 8.776244381910567e-05, 'samples': 18836992, 'steps': 36790, 'loss/train': 1.1563711166381836} 02/26/2022 01:12:22 - INFO - codeparrot_training - Step 36791: {'lr': 8.774999510067697e-05, 'samples': 18837504, 'steps': 36791, 'loss/train': 1.0664424896240234} 02/26/2022 01:12:26 - INFO - codeparrot_training - Step 36792: {'lr': 8.773754707727568e-05, 'samples': 18838016, 'steps': 36792, 'loss/train': 1.1581929922103882} 02/26/2022 01:12:31 - INFO - codeparrot_training - Step 36793: {'lr': 8.772509974895529e-05, 'samples': 18838528, 'steps': 36793, 'loss/train': 2.2678403854370117} 02/26/2022 01:12:35 - INFO - codeparrot_training - Step 36794: {'lr': 8.771265311576898e-05, 'samples': 18839040, 'steps': 36794, 'loss/train': 2.21399188041687} 02/26/2022 01:12:40 - INFO - codeparrot_training - Step 36795: {'lr': 8.77002071777703e-05, 'samples': 18839552, 'steps': 36795, 'loss/train': 1.5858330726623535} 02/26/2022 01:12:44 - INFO - codeparrot_training - Step 36796: {'lr': 8.768776193501221e-05, 'samples': 18840064, 'steps': 36796, 'loss/train': 1.632371425628662} 02/26/2022 01:12:49 - INFO - codeparrot_training - Step 36797: {'lr': 8.767531738754831e-05, 'samples': 18840576, 'steps': 36797, 'loss/train': 1.3312007188796997} 02/26/2022 01:12:53 - INFO - codeparrot_training - Step 36798: {'lr': 8.766287353543174e-05, 'samples': 18841088, 'steps': 36798, 'loss/train': 0.6778756976127625} 02/26/2022 01:12:58 - INFO - codeparrot_training - Step 36799: {'lr': 8.765043037871601e-05, 'samples': 18841600, 'steps': 36799, 'loss/train': 2.2973077297210693} 02/26/2022 01:13:02 - INFO - codeparrot_training - Step 36800: {'lr': 8.763798791745412e-05, 'samples': 18842112, 'steps': 36800, 'loss/train': 1.3411606550216675} 02/26/2022 01:13:08 - INFO - codeparrot_training - Step 36801: {'lr': 8.762554615169959e-05, 'samples': 18842624, 'steps': 36801, 'loss/train': 1.8979246616363525} 02/26/2022 01:13:11 - INFO - codeparrot_training - Step 36802: {'lr': 8.761310508150563e-05, 'samples': 18843136, 'steps': 36802, 'loss/train': 2.303459405899048} 02/26/2022 01:13:17 - INFO - codeparrot_training - Step 36803: {'lr': 8.760066470692563e-05, 'samples': 18843648, 'steps': 36803, 'loss/train': 1.838030219078064} 02/26/2022 01:13:20 - INFO - codeparrot_training - Step 36804: {'lr': 8.758822502801281e-05, 'samples': 18844160, 'steps': 36804, 'loss/train': 2.050367593765259} 02/26/2022 01:13:26 - INFO - codeparrot_training - Step 36805: {'lr': 8.757578604482045e-05, 'samples': 18844672, 'steps': 36805, 'loss/train': 1.5041393041610718} 02/26/2022 01:13:30 - INFO - codeparrot_training - Step 36806: {'lr': 8.756334775740176e-05, 'samples': 18845184, 'steps': 36806, 'loss/train': 2.2302026748657227} 02/26/2022 01:13:35 - INFO - codeparrot_training - Step 36807: {'lr': 8.75509101658102e-05, 'samples': 18845696, 'steps': 36807, 'loss/train': 1.511582612991333} 02/26/2022 01:13:39 - INFO - codeparrot_training - Step 36808: {'lr': 8.753847327009897e-05, 'samples': 18846208, 'steps': 36808, 'loss/train': 0.8627186417579651} 02/26/2022 01:13:44 - INFO - codeparrot_training - Step 36809: {'lr': 8.752603707032133e-05, 'samples': 18846720, 'steps': 36809, 'loss/train': 1.1052284240722656} 02/26/2022 01:13:48 - INFO - codeparrot_training - Step 36810: {'lr': 8.751360156653049e-05, 'samples': 18847232, 'steps': 36810, 'loss/train': 1.7874600887298584} 02/26/2022 01:13:53 - INFO - codeparrot_training - Step 36811: {'lr': 8.750116675877984e-05, 'samples': 18847744, 'steps': 36811, 'loss/train': 2.403961658477783} 02/26/2022 01:13:57 - INFO - codeparrot_training - Step 36812: {'lr': 8.748873264712259e-05, 'samples': 18848256, 'steps': 36812, 'loss/train': 0.9178921580314636} 02/26/2022 01:14:00 - INFO - codeparrot_training - Step 36813: {'lr': 8.747629923161205e-05, 'samples': 18848768, 'steps': 36813, 'loss/train': 1.6454524993896484} 02/26/2022 01:14:06 - INFO - codeparrot_training - Step 36814: {'lr': 8.746386651230131e-05, 'samples': 18849280, 'steps': 36814, 'loss/train': 0.9535394906997681} 02/26/2022 01:14:11 - INFO - codeparrot_training - Step 36815: {'lr': 8.745143448924383e-05, 'samples': 18849792, 'steps': 36815, 'loss/train': 1.245305061340332} 02/26/2022 01:14:15 - INFO - codeparrot_training - Step 36816: {'lr': 8.743900316249273e-05, 'samples': 18850304, 'steps': 36816, 'loss/train': 1.4459348917007446} 02/26/2022 01:14:21 - INFO - codeparrot_training - Step 36817: {'lr': 8.742657253210151e-05, 'samples': 18850816, 'steps': 36817, 'loss/train': 2.023543357849121} 02/26/2022 01:14:24 - INFO - codeparrot_training - Step 36818: {'lr': 8.741414259812302e-05, 'samples': 18851328, 'steps': 36818, 'loss/train': 1.181449055671692} 02/26/2022 01:14:30 - INFO - codeparrot_training - Step 36819: {'lr': 8.740171336061082e-05, 'samples': 18851840, 'steps': 36819, 'loss/train': 1.7929778099060059} 02/26/2022 01:14:33 - INFO - codeparrot_training - Step 36820: {'lr': 8.738928481961806e-05, 'samples': 18852352, 'steps': 36820, 'loss/train': 1.7656954526901245} 02/26/2022 01:14:37 - INFO - codeparrot_training - Step 36821: {'lr': 8.737685697519796e-05, 'samples': 18852864, 'steps': 36821, 'loss/train': 2.5808839797973633} 02/26/2022 01:14:42 - INFO - codeparrot_training - Step 36822: {'lr': 8.736442982740366e-05, 'samples': 18853376, 'steps': 36822, 'loss/train': 1.9175353050231934} 02/26/2022 01:14:48 - INFO - codeparrot_training - Step 36823: {'lr': 8.73520033762886e-05, 'samples': 18853888, 'steps': 36823, 'loss/train': 1.0916659832000732} 02/26/2022 01:14:51 - INFO - codeparrot_training - Step 36824: {'lr': 8.733957762190593e-05, 'samples': 18854400, 'steps': 36824, 'loss/train': 2.4224956035614014} 02/26/2022 01:14:57 - INFO - codeparrot_training - Step 36825: {'lr': 8.732715256430884e-05, 'samples': 18854912, 'steps': 36825, 'loss/train': 0.5221654772758484} 02/26/2022 01:15:00 - INFO - codeparrot_training - Step 36826: {'lr': 8.731472820355049e-05, 'samples': 18855424, 'steps': 36826, 'loss/train': 0.7488956451416016} 02/26/2022 01:15:07 - INFO - codeparrot_training - Step 36827: {'lr': 8.730230453968427e-05, 'samples': 18855936, 'steps': 36827, 'loss/train': 1.9354008436203003} 02/26/2022 01:15:10 - INFO - codeparrot_training - Step 36828: {'lr': 8.728988157276333e-05, 'samples': 18856448, 'steps': 36828, 'loss/train': 1.7492464780807495} 02/26/2022 01:15:16 - INFO - codeparrot_training - Step 36829: {'lr': 8.727745930284075e-05, 'samples': 18856960, 'steps': 36829, 'loss/train': 2.198659896850586} 02/26/2022 01:15:19 - INFO - codeparrot_training - Step 36830: {'lr': 8.726503772997e-05, 'samples': 18857472, 'steps': 36830, 'loss/train': 0.5789958834648132} 02/26/2022 01:15:24 - INFO - codeparrot_training - Step 36831: {'lr': 8.725261685420411e-05, 'samples': 18857984, 'steps': 36831, 'loss/train': 2.080350875854492} 02/26/2022 01:15:28 - INFO - codeparrot_training - Step 36832: {'lr': 8.724019667559635e-05, 'samples': 18858496, 'steps': 36832, 'loss/train': 0.3867916166782379} 02/26/2022 01:15:34 - INFO - codeparrot_training - Step 36833: {'lr': 8.722777719419984e-05, 'samples': 18859008, 'steps': 36833, 'loss/train': 2.3656091690063477} 02/26/2022 01:15:37 - INFO - codeparrot_training - Step 36834: {'lr': 8.721535841006798e-05, 'samples': 18859520, 'steps': 36834, 'loss/train': 1.1518471240997314} 02/26/2022 01:15:43 - INFO - codeparrot_training - Step 36835: {'lr': 8.720294032325368e-05, 'samples': 18860032, 'steps': 36835, 'loss/train': 2.0127668380737305} 02/26/2022 01:15:46 - INFO - codeparrot_training - Step 36836: {'lr': 8.719052293381036e-05, 'samples': 18860544, 'steps': 36836, 'loss/train': 2.148756265640259} 02/26/2022 01:15:52 - INFO - codeparrot_training - Step 36837: {'lr': 8.717810624179109e-05, 'samples': 18861056, 'steps': 36837, 'loss/train': 1.2076683044433594} 02/26/2022 01:15:56 - INFO - codeparrot_training - Step 36838: {'lr': 8.716569024724925e-05, 'samples': 18861568, 'steps': 36838, 'loss/train': 3.046046257019043} 02/26/2022 01:16:01 - INFO - codeparrot_training - Step 36839: {'lr': 8.715327495023773e-05, 'samples': 18862080, 'steps': 36839, 'loss/train': 1.2962663173675537} 02/26/2022 01:16:05 - INFO - codeparrot_training - Step 36840: {'lr': 8.714086035080996e-05, 'samples': 18862592, 'steps': 36840, 'loss/train': 1.513439416885376} 02/26/2022 01:16:11 - INFO - codeparrot_training - Step 36841: {'lr': 8.712844644901893e-05, 'samples': 18863104, 'steps': 36841, 'loss/train': 1.9734536409378052} 02/26/2022 01:16:14 - INFO - codeparrot_training - Step 36842: {'lr': 8.711603324491802e-05, 'samples': 18863616, 'steps': 36842, 'loss/train': 1.1246708631515503} 02/26/2022 01:16:20 - INFO - codeparrot_training - Step 36843: {'lr': 8.71036207385603e-05, 'samples': 18864128, 'steps': 36843, 'loss/train': 1.082332968711853} 02/26/2022 01:16:23 - INFO - codeparrot_training - Step 36844: {'lr': 8.709120892999895e-05, 'samples': 18864640, 'steps': 36844, 'loss/train': 2.752357244491577} 02/26/2022 01:16:29 - INFO - codeparrot_training - Step 36845: {'lr': 8.707879781928702e-05, 'samples': 18865152, 'steps': 36845, 'loss/train': 1.7152692079544067} 02/26/2022 01:16:32 - INFO - codeparrot_training - Step 36846: {'lr': 8.70663874064779e-05, 'samples': 18865664, 'steps': 36846, 'loss/train': 1.0590592622756958} 02/26/2022 01:16:39 - INFO - codeparrot_training - Step 36847: {'lr': 8.705397769162462e-05, 'samples': 18866176, 'steps': 36847, 'loss/train': 2.0598013401031494} 02/26/2022 01:16:42 - INFO - codeparrot_training - Step 36848: {'lr': 8.704156867478036e-05, 'samples': 18866688, 'steps': 36848, 'loss/train': 1.2578845024108887} 02/26/2022 01:16:48 - INFO - codeparrot_training - Step 36849: {'lr': 8.70291603559982e-05, 'samples': 18867200, 'steps': 36849, 'loss/train': 1.9242167472839355} 02/26/2022 01:16:51 - INFO - codeparrot_training - Step 36850: {'lr': 8.701675273533145e-05, 'samples': 18867712, 'steps': 36850, 'loss/train': 2.022484302520752} 02/26/2022 01:16:57 - INFO - codeparrot_training - Step 36851: {'lr': 8.700434581283309e-05, 'samples': 18868224, 'steps': 36851, 'loss/train': 1.3832817077636719} 02/26/2022 01:17:00 - INFO - codeparrot_training - Step 36852: {'lr': 8.699193958855653e-05, 'samples': 18868736, 'steps': 36852, 'loss/train': 1.7992764711380005} 02/26/2022 01:17:06 - INFO - codeparrot_training - Step 36853: {'lr': 8.697953406255458e-05, 'samples': 18869248, 'steps': 36853, 'loss/train': 0.9926881194114685} 02/26/2022 01:17:09 - INFO - codeparrot_training - Step 36854: {'lr': 8.696712923488059e-05, 'samples': 18869760, 'steps': 36854, 'loss/train': 2.096977472305298} 02/26/2022 01:17:15 - INFO - codeparrot_training - Step 36855: {'lr': 8.69547251055876e-05, 'samples': 18870272, 'steps': 36855, 'loss/train': 2.0804827213287354} 02/26/2022 01:17:18 - INFO - codeparrot_training - Step 36856: {'lr': 8.694232167472896e-05, 'samples': 18870784, 'steps': 36856, 'loss/train': 1.3874249458312988} 02/26/2022 01:17:24 - INFO - codeparrot_training - Step 36857: {'lr': 8.692991894235744e-05, 'samples': 18871296, 'steps': 36857, 'loss/train': 2.0656490325927734} 02/26/2022 01:17:27 - INFO - codeparrot_training - Step 36858: {'lr': 8.691751690852647e-05, 'samples': 18871808, 'steps': 36858, 'loss/train': 1.6641432046890259} 02/26/2022 01:17:33 - INFO - codeparrot_training - Step 36859: {'lr': 8.690511557328901e-05, 'samples': 18872320, 'steps': 36859, 'loss/train': 2.2094168663024902} 02/26/2022 01:17:37 - INFO - codeparrot_training - Step 36860: {'lr': 8.689271493669836e-05, 'samples': 18872832, 'steps': 36860, 'loss/train': 3.599497079849243} 02/26/2022 01:17:40 - INFO - codeparrot_training - Step 36861: {'lr': 8.688031499880741e-05, 'samples': 18873344, 'steps': 36861, 'loss/train': 1.0254249572753906} 02/26/2022 01:17:46 - INFO - codeparrot_training - Step 36862: {'lr': 8.686791575966946e-05, 'samples': 18873856, 'steps': 36862, 'loss/train': 4.905345916748047} 02/26/2022 01:17:49 - INFO - codeparrot_training - Step 36863: {'lr': 8.685551721933749e-05, 'samples': 18874368, 'steps': 36863, 'loss/train': 1.6642621755599976} 02/26/2022 01:17:55 - INFO - codeparrot_training - Step 36864: {'lr': 8.684311937786474e-05, 'samples': 18874880, 'steps': 36864, 'loss/train': 2.1865453720092773} 02/26/2022 01:17:59 - INFO - codeparrot_training - Step 36865: {'lr': 8.683072223530427e-05, 'samples': 18875392, 'steps': 36865, 'loss/train': 0.7889824509620667} 02/26/2022 01:18:04 - INFO - codeparrot_training - Step 36866: {'lr': 8.681832579170917e-05, 'samples': 18875904, 'steps': 36866, 'loss/train': 1.3934391736984253} 02/26/2022 01:18:08 - INFO - codeparrot_training - Step 36867: {'lr': 8.680593004713255e-05, 'samples': 18876416, 'steps': 36867, 'loss/train': 1.8681923151016235} 02/26/2022 01:18:13 - INFO - codeparrot_training - Step 36868: {'lr': 8.679353500162743e-05, 'samples': 18876928, 'steps': 36868, 'loss/train': 0.5904476642608643} 02/26/2022 01:18:17 - INFO - codeparrot_training - Step 36869: {'lr': 8.678114065524706e-05, 'samples': 18877440, 'steps': 36869, 'loss/train': 1.245560646057129} 02/26/2022 01:18:22 - INFO - codeparrot_training - Step 36870: {'lr': 8.676874700804443e-05, 'samples': 18877952, 'steps': 36870, 'loss/train': 2.2014451026916504} 02/26/2022 01:18:26 - INFO - codeparrot_training - Step 36871: {'lr': 8.675635406007268e-05, 'samples': 18878464, 'steps': 36871, 'loss/train': 1.8092398643493652} 02/26/2022 01:18:31 - INFO - codeparrot_training - Step 36872: {'lr': 8.67439618113848e-05, 'samples': 18878976, 'steps': 36872, 'loss/train': 2.4203686714172363} 02/26/2022 01:18:35 - INFO - codeparrot_training - Step 36873: {'lr': 8.673157026203401e-05, 'samples': 18879488, 'steps': 36873, 'loss/train': 1.5392042398452759} 02/26/2022 01:18:41 - INFO - codeparrot_training - Step 36874: {'lr': 8.671917941207335e-05, 'samples': 18880000, 'steps': 36874, 'loss/train': 1.6053953170776367} 02/26/2022 01:18:44 - INFO - codeparrot_training - Step 36875: {'lr': 8.670678926155587e-05, 'samples': 18880512, 'steps': 36875, 'loss/train': 2.9327821731567383} 02/26/2022 01:18:50 - INFO - codeparrot_training - Step 36876: {'lr': 8.669439981053457e-05, 'samples': 18881024, 'steps': 36876, 'loss/train': 1.974166750907898} 02/26/2022 01:18:54 - INFO - codeparrot_training - Step 36877: {'lr': 8.668201105906268e-05, 'samples': 18881536, 'steps': 36877, 'loss/train': 0.8197054862976074} 02/26/2022 01:19:00 - INFO - codeparrot_training - Step 36878: {'lr': 8.666962300719319e-05, 'samples': 18882048, 'steps': 36878, 'loss/train': 1.4822155237197876} 02/26/2022 01:19:03 - INFO - codeparrot_training - Step 36879: {'lr': 8.665723565497918e-05, 'samples': 18882560, 'steps': 36879, 'loss/train': 1.7086328268051147} 02/26/2022 01:19:09 - INFO - codeparrot_training - Step 36880: {'lr': 8.664484900247364e-05, 'samples': 18883072, 'steps': 36880, 'loss/train': 1.2913063764572144} 02/26/2022 01:19:12 - INFO - codeparrot_training - Step 36881: {'lr': 8.663246304972976e-05, 'samples': 18883584, 'steps': 36881, 'loss/train': 1.814562201499939} 02/26/2022 01:19:18 - INFO - codeparrot_training - Step 36882: {'lr': 8.662007779680053e-05, 'samples': 18884096, 'steps': 36882, 'loss/train': 2.63313889503479} 02/26/2022 01:19:21 - INFO - codeparrot_training - Step 36883: {'lr': 8.660769324373902e-05, 'samples': 18884608, 'steps': 36883, 'loss/train': 2.911949634552002} 02/26/2022 01:19:27 - INFO - codeparrot_training - Step 36884: {'lr': 8.659530939059818e-05, 'samples': 18885120, 'steps': 36884, 'loss/train': 1.2410475015640259} 02/26/2022 01:19:31 - INFO - codeparrot_training - Step 36885: {'lr': 8.658292623743124e-05, 'samples': 18885632, 'steps': 36885, 'loss/train': 2.073704242706299} 02/26/2022 01:19:36 - INFO - codeparrot_training - Step 36886: {'lr': 8.657054378429113e-05, 'samples': 18886144, 'steps': 36886, 'loss/train': 1.582241177558899} 02/26/2022 01:19:40 - INFO - codeparrot_training - Step 36887: {'lr': 8.655816203123096e-05, 'samples': 18886656, 'steps': 36887, 'loss/train': 1.8512474298477173} 02/26/2022 01:19:45 - INFO - codeparrot_training - Step 36888: {'lr': 8.65457809783036e-05, 'samples': 18887168, 'steps': 36888, 'loss/train': 0.6571885943412781} 02/26/2022 01:19:49 - INFO - codeparrot_training - Step 36889: {'lr': 8.653340062556231e-05, 'samples': 18887680, 'steps': 36889, 'loss/train': 1.2929080724716187} 02/26/2022 01:19:55 - INFO - codeparrot_training - Step 36890: {'lr': 8.652102097305994e-05, 'samples': 18888192, 'steps': 36890, 'loss/train': 0.9289174675941467} 02/26/2022 01:19:58 - INFO - codeparrot_training - Step 36891: {'lr': 8.650864202084977e-05, 'samples': 18888704, 'steps': 36891, 'loss/train': 1.818631887435913} 02/26/2022 01:20:01 - INFO - codeparrot_training - Step 36892: {'lr': 8.649626376898451e-05, 'samples': 18889216, 'steps': 36892, 'loss/train': 0.7039647102355957} 02/26/2022 01:20:07 - INFO - codeparrot_training - Step 36893: {'lr': 8.64838862175174e-05, 'samples': 18889728, 'steps': 36893, 'loss/train': 1.8207097053527832} 02/26/2022 01:20:11 - INFO - codeparrot_training - Step 36894: {'lr': 8.647150936650134e-05, 'samples': 18890240, 'steps': 36894, 'loss/train': 2.6992299556732178} 02/26/2022 01:20:16 - INFO - codeparrot_training - Step 36895: {'lr': 8.645913321598956e-05, 'samples': 18890752, 'steps': 36895, 'loss/train': 1.1495805978775024} 02/26/2022 01:20:20 - INFO - codeparrot_training - Step 36896: {'lr': 8.644675776603475e-05, 'samples': 18891264, 'steps': 36896, 'loss/train': 0.7655107378959656} 02/26/2022 01:20:26 - INFO - codeparrot_training - Step 36897: {'lr': 8.643438301669019e-05, 'samples': 18891776, 'steps': 36897, 'loss/train': 2.3657946586608887} 02/26/2022 01:20:29 - INFO - codeparrot_training - Step 36898: {'lr': 8.64220089680087e-05, 'samples': 18892288, 'steps': 36898, 'loss/train': 1.6852672100067139} 02/26/2022 01:20:35 - INFO - codeparrot_training - Step 36899: {'lr': 8.640963562004356e-05, 'samples': 18892800, 'steps': 36899, 'loss/train': 2.520089864730835} 02/26/2022 01:20:38 - INFO - codeparrot_training - Step 36900: {'lr': 8.639726297284742e-05, 'samples': 18893312, 'steps': 36900, 'loss/train': 1.196007251739502} 02/26/2022 01:20:44 - INFO - codeparrot_training - Step 36901: {'lr': 8.638489102647354e-05, 'samples': 18893824, 'steps': 36901, 'loss/train': 2.1488497257232666} 02/26/2022 01:20:47 - INFO - codeparrot_training - Step 36902: {'lr': 8.637251978097476e-05, 'samples': 18894336, 'steps': 36902, 'loss/train': 2.2404894828796387} 02/26/2022 01:20:53 - INFO - codeparrot_training - Step 36903: {'lr': 8.636014923640423e-05, 'samples': 18894848, 'steps': 36903, 'loss/train': 2.3863933086395264} 02/26/2022 01:20:56 - INFO - codeparrot_training - Step 36904: {'lr': 8.634777939281485e-05, 'samples': 18895360, 'steps': 36904, 'loss/train': 1.0175422430038452} 02/26/2022 01:21:02 - INFO - codeparrot_training - Step 36905: {'lr': 8.633541025025962e-05, 'samples': 18895872, 'steps': 36905, 'loss/train': 1.083966612815857} 02/26/2022 01:21:05 - INFO - codeparrot_training - Step 36906: {'lr': 8.632304180879145e-05, 'samples': 18896384, 'steps': 36906, 'loss/train': 1.6595673561096191} 02/26/2022 01:21:11 - INFO - codeparrot_training - Step 36907: {'lr': 8.631067406846346e-05, 'samples': 18896896, 'steps': 36907, 'loss/train': 1.9447115659713745} 02/26/2022 01:21:14 - INFO - codeparrot_training - Step 36908: {'lr': 8.629830702932856e-05, 'samples': 18897408, 'steps': 36908, 'loss/train': 1.941929578781128} 02/26/2022 01:21:20 - INFO - codeparrot_training - Step 36909: {'lr': 8.628594069143975e-05, 'samples': 18897920, 'steps': 36909, 'loss/train': 1.0319911241531372} 02/26/2022 01:21:26 - INFO - codeparrot_training - Step 36910: {'lr': 8.62735750548499e-05, 'samples': 18898432, 'steps': 36910, 'loss/train': 1.553390383720398} 02/26/2022 01:21:29 - INFO - codeparrot_training - Step 36911: {'lr': 8.626121011961216e-05, 'samples': 18898944, 'steps': 36911, 'loss/train': 0.03581497445702553} 02/26/2022 01:21:35 - INFO - codeparrot_training - Step 36912: {'lr': 8.624884588577938e-05, 'samples': 18899456, 'steps': 36912, 'loss/train': 1.3789877891540527} 02/26/2022 01:21:38 - INFO - codeparrot_training - Step 36913: {'lr': 8.623648235340457e-05, 'samples': 18899968, 'steps': 36913, 'loss/train': 1.325390338897705} 02/26/2022 01:21:44 - INFO - codeparrot_training - Step 36914: {'lr': 8.622411952254055e-05, 'samples': 18900480, 'steps': 36914, 'loss/train': 1.1682175397872925} 02/26/2022 01:21:47 - INFO - codeparrot_training - Step 36915: {'lr': 8.621175739324052e-05, 'samples': 18900992, 'steps': 36915, 'loss/train': 1.6988774538040161} 02/26/2022 01:21:53 - INFO - codeparrot_training - Step 36916: {'lr': 8.619939596555729e-05, 'samples': 18901504, 'steps': 36916, 'loss/train': 1.4303603172302246} 02/26/2022 01:21:56 - INFO - codeparrot_training - Step 36917: {'lr': 8.618703523954385e-05, 'samples': 18902016, 'steps': 36917, 'loss/train': 0.629348635673523} 02/26/2022 01:22:02 - INFO - codeparrot_training - Step 36918: {'lr': 8.617467521525313e-05, 'samples': 18902528, 'steps': 36918, 'loss/train': 2.0985968112945557} 02/26/2022 01:22:05 - INFO - codeparrot_training - Step 36919: {'lr': 8.616231589273799e-05, 'samples': 18903040, 'steps': 36919, 'loss/train': 1.2436678409576416} 02/26/2022 01:22:12 - INFO - codeparrot_training - Step 36920: {'lr': 8.614995727205155e-05, 'samples': 18903552, 'steps': 36920, 'loss/train': 1.4784159660339355} 02/26/2022 01:22:15 - INFO - codeparrot_training - Step 36921: {'lr': 8.613759935324667e-05, 'samples': 18904064, 'steps': 36921, 'loss/train': 1.9755879640579224} 02/26/2022 01:22:20 - INFO - codeparrot_training - Step 36922: {'lr': 8.612524213637626e-05, 'samples': 18904576, 'steps': 36922, 'loss/train': 1.7446744441986084} 02/26/2022 01:22:24 - INFO - codeparrot_training - Step 36923: {'lr': 8.611288562149322e-05, 'samples': 18905088, 'steps': 36923, 'loss/train': 1.886895775794983} 02/26/2022 01:22:30 - INFO - codeparrot_training - Step 36924: {'lr': 8.610052980865062e-05, 'samples': 18905600, 'steps': 36924, 'loss/train': 2.2142117023468018} 02/26/2022 01:22:33 - INFO - codeparrot_training - Step 36925: {'lr': 8.608817469790123e-05, 'samples': 18906112, 'steps': 36925, 'loss/train': 2.1308860778808594} 02/26/2022 01:22:39 - INFO - codeparrot_training - Step 36926: {'lr': 8.607582028929819e-05, 'samples': 18906624, 'steps': 36926, 'loss/train': 1.9334596395492554} 02/26/2022 01:22:42 - INFO - codeparrot_training - Step 36927: {'lr': 8.606346658289413e-05, 'samples': 18907136, 'steps': 36927, 'loss/train': 1.5525330305099487} 02/26/2022 01:22:48 - INFO - codeparrot_training - Step 36928: {'lr': 8.60511135787422e-05, 'samples': 18907648, 'steps': 36928, 'loss/train': 0.7826526761054993} 02/26/2022 01:22:51 - INFO - codeparrot_training - Step 36929: {'lr': 8.603876127689517e-05, 'samples': 18908160, 'steps': 36929, 'loss/train': 1.7909104824066162} 02/26/2022 01:22:57 - INFO - codeparrot_training - Step 36930: {'lr': 8.602640967740618e-05, 'samples': 18908672, 'steps': 36930, 'loss/train': 3.16336989402771} 02/26/2022 01:23:00 - INFO - codeparrot_training - Step 36931: {'lr': 8.601405878032783e-05, 'samples': 18909184, 'steps': 36931, 'loss/train': 1.7063947916030884} 02/26/2022 01:23:06 - INFO - codeparrot_training - Step 36932: {'lr': 8.600170858571326e-05, 'samples': 18909696, 'steps': 36932, 'loss/train': 1.1706607341766357} 02/26/2022 01:23:10 - INFO - codeparrot_training - Step 36933: {'lr': 8.598935909361521e-05, 'samples': 18910208, 'steps': 36933, 'loss/train': 1.4866220951080322} 02/26/2022 01:23:15 - INFO - codeparrot_training - Step 36934: {'lr': 8.597701030408683e-05, 'samples': 18910720, 'steps': 36934, 'loss/train': 1.1321985721588135} 02/26/2022 01:23:19 - INFO - codeparrot_training - Step 36935: {'lr': 8.596466221718069e-05, 'samples': 18911232, 'steps': 36935, 'loss/train': 1.9462562799453735} 02/26/2022 01:23:24 - INFO - codeparrot_training - Step 36936: {'lr': 8.595231483294993e-05, 'samples': 18911744, 'steps': 36936, 'loss/train': 1.0940117835998535} 02/26/2022 01:23:28 - INFO - codeparrot_training - Step 36937: {'lr': 8.593996815144728e-05, 'samples': 18912256, 'steps': 36937, 'loss/train': 2.583690643310547} 02/26/2022 01:23:33 - INFO - codeparrot_training - Step 36938: {'lr': 8.59276221727258e-05, 'samples': 18912768, 'steps': 36938, 'loss/train': 1.597976565361023} 02/26/2022 01:23:37 - INFO - codeparrot_training - Step 36939: {'lr': 8.591527689683828e-05, 'samples': 18913280, 'steps': 36939, 'loss/train': 1.8345625400543213} 02/26/2022 01:23:42 - INFO - codeparrot_training - Step 36940: {'lr': 8.590293232383762e-05, 'samples': 18913792, 'steps': 36940, 'loss/train': 1.2770590782165527} 02/26/2022 01:23:45 - INFO - codeparrot_training - Step 36941: {'lr': 8.58905884537766e-05, 'samples': 18914304, 'steps': 36941, 'loss/train': 0.027937322854995728} 02/26/2022 01:23:51 - INFO - codeparrot_training - Step 36942: {'lr': 8.587824528670826e-05, 'samples': 18914816, 'steps': 36942, 'loss/train': 2.7899630069732666} 02/26/2022 01:23:54 - INFO - codeparrot_training - Step 36943: {'lr': 8.586590282268542e-05, 'samples': 18915328, 'steps': 36943, 'loss/train': 1.5152673721313477} 02/26/2022 01:24:00 - INFO - codeparrot_training - Step 36944: {'lr': 8.585356106176093e-05, 'samples': 18915840, 'steps': 36944, 'loss/train': 1.5155937671661377} 02/26/2022 01:24:03 - INFO - codeparrot_training - Step 36945: {'lr': 8.584122000398758e-05, 'samples': 18916352, 'steps': 36945, 'loss/train': 1.8368003368377686} 02/26/2022 01:24:10 - INFO - codeparrot_training - Step 36946: {'lr': 8.58288796494184e-05, 'samples': 18916864, 'steps': 36946, 'loss/train': 1.3719526529312134} 02/26/2022 01:24:13 - INFO - codeparrot_training - Step 36947: {'lr': 8.581653999810617e-05, 'samples': 18917376, 'steps': 36947, 'loss/train': 1.2114650011062622} 02/26/2022 01:24:19 - INFO - codeparrot_training - Step 36948: {'lr': 8.580420105010375e-05, 'samples': 18917888, 'steps': 36948, 'loss/train': 2.3056693077087402} 02/26/2022 01:24:22 - INFO - codeparrot_training - Step 36949: {'lr': 8.579186280546392e-05, 'samples': 18918400, 'steps': 36949, 'loss/train': 1.5895808935165405} 02/26/2022 01:24:28 - INFO - codeparrot_training - Step 36950: {'lr': 8.577952526423969e-05, 'samples': 18918912, 'steps': 36950, 'loss/train': 1.3941559791564941} 02/26/2022 01:24:31 - INFO - codeparrot_training - Step 36951: {'lr': 8.576718842648379e-05, 'samples': 18919424, 'steps': 36951, 'loss/train': 1.3350311517715454} 02/26/2022 01:24:37 - INFO - codeparrot_training - Step 36952: {'lr': 8.575485229224914e-05, 'samples': 18919936, 'steps': 36952, 'loss/train': 0.5477108359336853} 02/26/2022 01:24:40 - INFO - codeparrot_training - Step 36953: {'lr': 8.574251686158846e-05, 'samples': 18920448, 'steps': 36953, 'loss/train': 1.8919907808303833} 02/26/2022 01:24:46 - INFO - codeparrot_training - Step 36954: {'lr': 8.573018213455477e-05, 'samples': 18920960, 'steps': 36954, 'loss/train': 2.1822316646575928} 02/26/2022 01:24:49 - INFO - codeparrot_training - Step 36955: {'lr': 8.571784811120081e-05, 'samples': 18921472, 'steps': 36955, 'loss/train': 2.1003291606903076} 02/26/2022 01:24:55 - INFO - codeparrot_training - Step 36956: {'lr': 8.570551479157942e-05, 'samples': 18921984, 'steps': 36956, 'loss/train': 2.2563424110412598} 02/26/2022 01:24:58 - INFO - codeparrot_training - Step 36957: {'lr': 8.569318217574334e-05, 'samples': 18922496, 'steps': 36957, 'loss/train': 1.9402010440826416} 02/26/2022 01:25:04 - INFO - codeparrot_training - Step 36958: {'lr': 8.568085026374561e-05, 'samples': 18923008, 'steps': 36958, 'loss/train': 1.490246295928955} 02/26/2022 01:25:08 - INFO - codeparrot_training - Step 36959: {'lr': 8.56685190556389e-05, 'samples': 18923520, 'steps': 36959, 'loss/train': 1.853089451789856} 02/26/2022 01:25:13 - INFO - codeparrot_training - Step 36960: {'lr': 8.565618855147602e-05, 'samples': 18924032, 'steps': 36960, 'loss/train': 1.6487221717834473} 02/26/2022 01:25:17 - INFO - codeparrot_training - Step 36961: {'lr': 8.564385875130992e-05, 'samples': 18924544, 'steps': 36961, 'loss/train': 0.9966661930084229} 02/26/2022 01:25:22 - INFO - codeparrot_training - Step 36962: {'lr': 8.563152965519336e-05, 'samples': 18925056, 'steps': 36962, 'loss/train': 1.129966378211975} 02/26/2022 01:25:26 - INFO - codeparrot_training - Step 36963: {'lr': 8.561920126317909e-05, 'samples': 18925568, 'steps': 36963, 'loss/train': 0.6994410157203674} 02/26/2022 01:25:31 - INFO - codeparrot_training - Step 36964: {'lr': 8.560687357531993e-05, 'samples': 18926080, 'steps': 36964, 'loss/train': 1.9713150262832642} 02/26/2022 01:25:35 - INFO - codeparrot_training - Step 36965: {'lr': 8.559454659166888e-05, 'samples': 18926592, 'steps': 36965, 'loss/train': 0.9345439672470093} 02/26/2022 01:25:40 - INFO - codeparrot_training - Step 36966: {'lr': 8.558222031227838e-05, 'samples': 18927104, 'steps': 36966, 'loss/train': 2.3028151988983154} 02/26/2022 01:25:47 - INFO - codeparrot_training - Step 36967: {'lr': 8.556989473720156e-05, 'samples': 18927616, 'steps': 36967, 'loss/train': 2.0832371711730957} 02/26/2022 01:25:50 - INFO - codeparrot_training - Step 36968: {'lr': 8.555756986649099e-05, 'samples': 18928128, 'steps': 36968, 'loss/train': 1.4604074954986572} 02/26/2022 01:25:53 - INFO - codeparrot_training - Step 36969: {'lr': 8.554524570019978e-05, 'samples': 18928640, 'steps': 36969, 'loss/train': 1.4824813604354858} 02/26/2022 01:25:59 - INFO - codeparrot_training - Step 36970: {'lr': 8.553292223838033e-05, 'samples': 18929152, 'steps': 36970, 'loss/train': 1.9281446933746338} 02/26/2022 01:26:05 - INFO - codeparrot_training - Step 36971: {'lr': 8.552059948108567e-05, 'samples': 18929664, 'steps': 36971, 'loss/train': 1.9268885850906372} 02/26/2022 01:26:08 - INFO - codeparrot_training - Step 36972: {'lr': 8.550827742836851e-05, 'samples': 18930176, 'steps': 36972, 'loss/train': 1.6829442977905273} 02/26/2022 01:26:14 - INFO - codeparrot_training - Step 36973: {'lr': 8.549595608028168e-05, 'samples': 18930688, 'steps': 36973, 'loss/train': 1.2781786918640137} 02/26/2022 01:26:17 - INFO - codeparrot_training - Step 36974: {'lr': 8.548363543687798e-05, 'samples': 18931200, 'steps': 36974, 'loss/train': 2.119123697280884} 02/26/2022 01:26:23 - INFO - codeparrot_training - Step 36975: {'lr': 8.547131549821014e-05, 'samples': 18931712, 'steps': 36975, 'loss/train': 0.3199837803840637} 02/26/2022 01:26:26 - INFO - codeparrot_training - Step 36976: {'lr': 8.545899626433085e-05, 'samples': 18932224, 'steps': 36976, 'loss/train': 2.2008097171783447} 02/26/2022 01:26:32 - INFO - codeparrot_training - Step 36977: {'lr': 8.544667773529305e-05, 'samples': 18932736, 'steps': 36977, 'loss/train': 1.7141129970550537} 02/26/2022 01:26:36 - INFO - codeparrot_training - Step 36978: {'lr': 8.543435991114942e-05, 'samples': 18933248, 'steps': 36978, 'loss/train': 1.878537654876709} 02/26/2022 01:26:41 - INFO - codeparrot_training - Step 36979: {'lr': 8.542204279195276e-05, 'samples': 18933760, 'steps': 36979, 'loss/train': 1.8371224403381348} 02/26/2022 01:26:45 - INFO - codeparrot_training - Step 36980: {'lr': 8.540972637775571e-05, 'samples': 18934272, 'steps': 36980, 'loss/train': 1.2525502443313599} 02/26/2022 01:26:50 - INFO - codeparrot_training - Step 36981: {'lr': 8.539741066861123e-05, 'samples': 18934784, 'steps': 36981, 'loss/train': 1.5267858505249023} 02/26/2022 01:26:54 - INFO - codeparrot_training - Step 36982: {'lr': 8.538509566457198e-05, 'samples': 18935296, 'steps': 36982, 'loss/train': 1.4636080265045166} 02/26/2022 01:26:59 - INFO - codeparrot_training - Step 36983: {'lr': 8.537278136569068e-05, 'samples': 18935808, 'steps': 36983, 'loss/train': 1.436458706855774} 02/26/2022 01:27:03 - INFO - codeparrot_training - Step 36984: {'lr': 8.536046777202003e-05, 'samples': 18936320, 'steps': 36984, 'loss/train': 1.3397808074951172} 02/26/2022 01:27:08 - INFO - codeparrot_training - Step 36985: {'lr': 8.534815488361297e-05, 'samples': 18936832, 'steps': 36985, 'loss/train': 1.1611132621765137} 02/26/2022 01:27:12 - INFO - codeparrot_training - Step 36986: {'lr': 8.533584270052203e-05, 'samples': 18937344, 'steps': 36986, 'loss/train': 1.9433910846710205} 02/26/2022 01:27:17 - INFO - codeparrot_training - Step 36987: {'lr': 8.53235312228002e-05, 'samples': 18937856, 'steps': 36987, 'loss/train': 1.7604655027389526} 02/26/2022 01:27:21 - INFO - codeparrot_training - Step 36988: {'lr': 8.531122045049993e-05, 'samples': 18938368, 'steps': 36988, 'loss/train': 2.3688790798187256} 02/26/2022 01:27:27 - INFO - codeparrot_training - Step 36989: {'lr': 8.529891038367418e-05, 'samples': 18938880, 'steps': 36989, 'loss/train': 0.8820657730102539} 02/26/2022 01:27:30 - INFO - codeparrot_training - Step 36990: {'lr': 8.528660102237551e-05, 'samples': 18939392, 'steps': 36990, 'loss/train': 1.9192264080047607} 02/26/2022 01:27:34 - INFO - codeparrot_training - Step 36991: {'lr': 8.52742923666569e-05, 'samples': 18939904, 'steps': 36991, 'loss/train': 1.7725179195404053} 02/26/2022 01:27:39 - INFO - codeparrot_training - Step 36992: {'lr': 8.526198441657077e-05, 'samples': 18940416, 'steps': 36992, 'loss/train': 2.414793014526367} 02/26/2022 01:27:43 - INFO - codeparrot_training - Step 36993: {'lr': 8.524967717217006e-05, 'samples': 18940928, 'steps': 36993, 'loss/train': 2.566577434539795} 02/26/2022 01:27:49 - INFO - codeparrot_training - Step 36994: {'lr': 8.523737063350733e-05, 'samples': 18941440, 'steps': 36994, 'loss/train': 1.655949354171753} 02/26/2022 01:27:53 - INFO - codeparrot_training - Step 36995: {'lr': 8.522506480063555e-05, 'samples': 18941952, 'steps': 36995, 'loss/train': 1.919616937637329} 02/26/2022 01:27:58 - INFO - codeparrot_training - Step 36996: {'lr': 8.521275967360711e-05, 'samples': 18942464, 'steps': 36996, 'loss/train': 2.190977096557617} 02/26/2022 01:28:02 - INFO - codeparrot_training - Step 36997: {'lr': 8.520045525247497e-05, 'samples': 18942976, 'steps': 36997, 'loss/train': 0.4945140480995178} 02/26/2022 01:28:07 - INFO - codeparrot_training - Step 36998: {'lr': 8.518815153729173e-05, 'samples': 18943488, 'steps': 36998, 'loss/train': 2.316406011581421} 02/26/2022 01:28:11 - INFO - codeparrot_training - Step 36999: {'lr': 8.517584852811005e-05, 'samples': 18944000, 'steps': 36999, 'loss/train': 1.4211903810501099} 02/26/2022 01:28:11 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 01:28:29 - WARNING - huggingface_hub.repository - Several commits (37) will be pushed upstream. 02/26/2022 01:28:29 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 01:29:05 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy f71b356..333cdce floral-grass-11 -> floral-grass-11 02/26/2022 01:29:13 - INFO - codeparrot_training - Step 37000: {'lr': 8.516354622498279e-05, 'samples': 18944512, 'steps': 37000, 'loss/train': 1.0448174476623535} 02/26/2022 01:29:16 - INFO - codeparrot_training - Step 37001: {'lr': 8.515124462796254e-05, 'samples': 18945024, 'steps': 37001, 'loss/train': 2.14421010017395} 02/26/2022 01:29:22 - INFO - codeparrot_training - Step 37002: {'lr': 8.513894373710201e-05, 'samples': 18945536, 'steps': 37002, 'loss/train': 1.1430444717407227} 02/26/2022 01:29:25 - INFO - codeparrot_training - Step 37003: {'lr': 8.512664355245384e-05, 'samples': 18946048, 'steps': 37003, 'loss/train': 0.11870545148849487} 02/26/2022 01:29:31 - INFO - codeparrot_training - Step 37004: {'lr': 8.511434407407082e-05, 'samples': 18946560, 'steps': 37004, 'loss/train': 0.5594982504844666} 02/26/2022 01:29:35 - INFO - codeparrot_training - Step 37005: {'lr': 8.510204530200561e-05, 'samples': 18947072, 'steps': 37005, 'loss/train': 1.4826196432113647} 02/26/2022 01:29:40 - INFO - codeparrot_training - Step 37006: {'lr': 8.508974723631088e-05, 'samples': 18947584, 'steps': 37006, 'loss/train': 2.9788272380828857} 02/26/2022 01:29:44 - INFO - codeparrot_training - Step 37007: {'lr': 8.507744987703923e-05, 'samples': 18948096, 'steps': 37007, 'loss/train': 0.9081933498382568} 02/26/2022 01:29:49 - INFO - codeparrot_training - Step 37008: {'lr': 8.506515322424349e-05, 'samples': 18948608, 'steps': 37008, 'loss/train': 2.1243278980255127} 02/26/2022 01:29:53 - INFO - codeparrot_training - Step 37009: {'lr': 8.505285727797627e-05, 'samples': 18949120, 'steps': 37009, 'loss/train': 2.4234466552734375} 02/26/2022 01:29:58 - INFO - codeparrot_training - Step 37010: {'lr': 8.504056203829022e-05, 'samples': 18949632, 'steps': 37010, 'loss/train': 0.6121158599853516} 02/26/2022 01:30:02 - INFO - codeparrot_training - Step 37011: {'lr': 8.502826750523793e-05, 'samples': 18950144, 'steps': 37011, 'loss/train': 1.7025648355484009} 02/26/2022 01:30:07 - INFO - codeparrot_training - Step 37012: {'lr': 8.501597367887226e-05, 'samples': 18950656, 'steps': 37012, 'loss/train': 1.742186188697815} 02/26/2022 01:30:11 - INFO - codeparrot_training - Step 37013: {'lr': 8.500368055924576e-05, 'samples': 18951168, 'steps': 37013, 'loss/train': 2.372593879699707} 02/26/2022 01:30:17 - INFO - codeparrot_training - Step 37014: {'lr': 8.49913881464111e-05, 'samples': 18951680, 'steps': 37014, 'loss/train': 1.6525758504867554} 02/26/2022 01:30:22 - INFO - codeparrot_training - Step 37015: {'lr': 8.497909644042088e-05, 'samples': 18952192, 'steps': 37015, 'loss/train': 0.9016113877296448} 02/26/2022 01:30:26 - INFO - codeparrot_training - Step 37016: {'lr': 8.496680544132788e-05, 'samples': 18952704, 'steps': 37016, 'loss/train': 1.5941678285598755} 02/26/2022 01:30:31 - INFO - codeparrot_training - Step 37017: {'lr': 8.495451514918465e-05, 'samples': 18953216, 'steps': 37017, 'loss/train': 0.6739307045936584} 02/26/2022 01:30:35 - INFO - codeparrot_training - Step 37018: {'lr': 8.494222556404391e-05, 'samples': 18953728, 'steps': 37018, 'loss/train': 1.3178893327713013} 02/26/2022 01:30:41 - INFO - codeparrot_training - Step 37019: {'lr': 8.492993668595817e-05, 'samples': 18954240, 'steps': 37019, 'loss/train': 1.880487322807312} 02/26/2022 01:30:44 - INFO - codeparrot_training - Step 37020: {'lr': 8.491764851498024e-05, 'samples': 18954752, 'steps': 37020, 'loss/train': 1.421849012374878} 02/26/2022 01:30:49 - INFO - codeparrot_training - Step 37021: {'lr': 8.490536105116259e-05, 'samples': 18955264, 'steps': 37021, 'loss/train': 1.0910184383392334} 02/26/2022 01:30:53 - INFO - codeparrot_training - Step 37022: {'lr': 8.489307429455814e-05, 'samples': 18955776, 'steps': 37022, 'loss/train': 1.6827292442321777} 02/26/2022 01:30:59 - INFO - codeparrot_training - Step 37023: {'lr': 8.488078824521914e-05, 'samples': 18956288, 'steps': 37023, 'loss/train': 1.7838222980499268} 02/26/2022 01:31:03 - INFO - codeparrot_training - Step 37024: {'lr': 8.486850290319852e-05, 'samples': 18956800, 'steps': 37024, 'loss/train': 2.0251317024230957} 02/26/2022 01:31:08 - INFO - codeparrot_training - Step 37025: {'lr': 8.485621826854869e-05, 'samples': 18957312, 'steps': 37025, 'loss/train': 3.424682855606079} 02/26/2022 01:31:12 - INFO - codeparrot_training - Step 37026: {'lr': 8.484393434132259e-05, 'samples': 18957824, 'steps': 37026, 'loss/train': 1.7839360237121582} 02/26/2022 01:31:17 - INFO - codeparrot_training - Step 37027: {'lr': 8.483165112157243e-05, 'samples': 18958336, 'steps': 37027, 'loss/train': 2.322451591491699} 02/26/2022 01:31:21 - INFO - codeparrot_training - Step 37028: {'lr': 8.481936860935113e-05, 'samples': 18958848, 'steps': 37028, 'loss/train': 1.2318921089172363} 02/26/2022 01:31:27 - INFO - codeparrot_training - Step 37029: {'lr': 8.480708680471111e-05, 'samples': 18959360, 'steps': 37029, 'loss/train': 1.5701014995574951} 02/26/2022 01:31:30 - INFO - codeparrot_training - Step 37030: {'lr': 8.479480570770525e-05, 'samples': 18959872, 'steps': 37030, 'loss/train': 1.7174086570739746} 02/26/2022 01:31:34 - INFO - codeparrot_training - Step 37031: {'lr': 8.478252531838581e-05, 'samples': 18960384, 'steps': 37031, 'loss/train': 0.49307626485824585} 02/26/2022 01:31:39 - INFO - codeparrot_training - Step 37032: {'lr': 8.477024563680566e-05, 'samples': 18960896, 'steps': 37032, 'loss/train': 0.8573851585388184} 02/26/2022 01:31:43 - INFO - codeparrot_training - Step 37033: {'lr': 8.475796666301721e-05, 'samples': 18961408, 'steps': 37033, 'loss/train': 1.7240451574325562} 02/26/2022 01:31:48 - INFO - codeparrot_training - Step 37034: {'lr': 8.474568839707328e-05, 'samples': 18961920, 'steps': 37034, 'loss/train': 2.503028392791748} 02/26/2022 01:31:52 - INFO - codeparrot_training - Step 37035: {'lr': 8.473341083902634e-05, 'samples': 18962432, 'steps': 37035, 'loss/train': 1.8472881317138672} 02/26/2022 01:31:57 - INFO - codeparrot_training - Step 37036: {'lr': 8.472113398892894e-05, 'samples': 18962944, 'steps': 37036, 'loss/train': 2.088653326034546} 02/26/2022 01:32:01 - INFO - codeparrot_training - Step 37037: {'lr': 8.470885784683369e-05, 'samples': 18963456, 'steps': 37037, 'loss/train': 0.43235892057418823} 02/26/2022 01:32:06 - INFO - codeparrot_training - Step 37038: {'lr': 8.469658241279326e-05, 'samples': 18963968, 'steps': 37038, 'loss/train': 1.8186694383621216} 02/26/2022 01:32:10 - INFO - codeparrot_training - Step 37039: {'lr': 8.46843076868602e-05, 'samples': 18964480, 'steps': 37039, 'loss/train': 1.039555549621582} 02/26/2022 01:32:15 - INFO - codeparrot_training - Step 37040: {'lr': 8.467203366908707e-05, 'samples': 18964992, 'steps': 37040, 'loss/train': 1.0407304763793945} 02/26/2022 01:32:19 - INFO - codeparrot_training - Step 37041: {'lr': 8.465976035952635e-05, 'samples': 18965504, 'steps': 37041, 'loss/train': 2.1811790466308594} 02/26/2022 01:32:25 - INFO - codeparrot_training - Step 37042: {'lr': 8.464748775823081e-05, 'samples': 18966016, 'steps': 37042, 'loss/train': 2.341484546661377} 02/26/2022 01:32:28 - INFO - codeparrot_training - Step 37043: {'lr': 8.463521586525294e-05, 'samples': 18966528, 'steps': 37043, 'loss/train': 1.6124290227890015} 02/26/2022 01:32:34 - INFO - codeparrot_training - Step 37044: {'lr': 8.462294468064526e-05, 'samples': 18967040, 'steps': 37044, 'loss/train': 0.9585081934928894} 02/26/2022 01:32:37 - INFO - codeparrot_training - Step 37045: {'lr': 8.461067420446041e-05, 'samples': 18967552, 'steps': 37045, 'loss/train': 1.5336618423461914} 02/26/2022 01:32:43 - INFO - codeparrot_training - Step 37046: {'lr': 8.459840443675082e-05, 'samples': 18968064, 'steps': 37046, 'loss/train': 2.356720447540283} 02/26/2022 01:32:46 - INFO - codeparrot_training - Step 37047: {'lr': 8.458613537756923e-05, 'samples': 18968576, 'steps': 37047, 'loss/train': 2.3372151851654053} 02/26/2022 01:32:52 - INFO - codeparrot_training - Step 37048: {'lr': 8.457386702696809e-05, 'samples': 18969088, 'steps': 37048, 'loss/train': 0.9617457985877991} 02/26/2022 01:32:57 - INFO - codeparrot_training - Step 37049: {'lr': 8.4561599385e-05, 'samples': 18969600, 'steps': 37049, 'loss/train': 1.6585125923156738} 02/26/2022 01:33:01 - INFO - codeparrot_training - Step 37050: {'lr': 8.45493324517174e-05, 'samples': 18970112, 'steps': 37050, 'loss/train': 1.6518136262893677} 02/26/2022 01:33:08 - INFO - codeparrot_training - Step 37051: {'lr': 8.4537066227173e-05, 'samples': 18970624, 'steps': 37051, 'loss/train': 1.6124365329742432} 02/26/2022 01:33:11 - INFO - codeparrot_training - Step 37052: {'lr': 8.452480071141927e-05, 'samples': 18971136, 'steps': 37052, 'loss/train': 0.1726987361907959} 02/26/2022 01:33:17 - INFO - codeparrot_training - Step 37053: {'lr': 8.451253590450875e-05, 'samples': 18971648, 'steps': 37053, 'loss/train': 2.2282145023345947} 02/26/2022 01:33:20 - INFO - codeparrot_training - Step 37054: {'lr': 8.450027180649389e-05, 'samples': 18972160, 'steps': 37054, 'loss/train': 2.0039875507354736} 02/26/2022 01:33:26 - INFO - codeparrot_training - Step 37055: {'lr': 8.448800841742741e-05, 'samples': 18972672, 'steps': 37055, 'loss/train': 1.7905770540237427} 02/26/2022 01:33:29 - INFO - codeparrot_training - Step 37056: {'lr': 8.447574573736172e-05, 'samples': 18973184, 'steps': 37056, 'loss/train': 1.8973604440689087} 02/26/2022 01:33:35 - INFO - codeparrot_training - Step 37057: {'lr': 8.446348376634937e-05, 'samples': 18973696, 'steps': 37057, 'loss/train': 2.0184977054595947} 02/26/2022 01:33:38 - INFO - codeparrot_training - Step 37058: {'lr': 8.445122250444284e-05, 'samples': 18974208, 'steps': 37058, 'loss/train': 1.7296326160430908} 02/26/2022 01:33:44 - INFO - codeparrot_training - Step 37059: {'lr': 8.44389619516948e-05, 'samples': 18974720, 'steps': 37059, 'loss/train': 1.8135051727294922} 02/26/2022 01:33:47 - INFO - codeparrot_training - Step 37060: {'lr': 8.442670210815756e-05, 'samples': 18975232, 'steps': 37060, 'loss/train': 2.1664235591888428} 02/26/2022 01:33:53 - INFO - codeparrot_training - Step 37061: {'lr': 8.441444297388393e-05, 'samples': 18975744, 'steps': 37061, 'loss/train': 1.0071851015090942} 02/26/2022 01:33:56 - INFO - codeparrot_training - Step 37062: {'lr': 8.440218454892606e-05, 'samples': 18976256, 'steps': 37062, 'loss/train': 2.84360408782959} 02/26/2022 01:34:02 - INFO - codeparrot_training - Step 37063: {'lr': 8.438992683333677e-05, 'samples': 18976768, 'steps': 37063, 'loss/train': 1.911067247390747} 02/26/2022 01:34:05 - INFO - codeparrot_training - Step 37064: {'lr': 8.437766982716835e-05, 'samples': 18977280, 'steps': 37064, 'loss/train': 1.918104887008667} 02/26/2022 01:34:12 - INFO - codeparrot_training - Step 37065: {'lr': 8.436541353047356e-05, 'samples': 18977792, 'steps': 37065, 'loss/train': 1.3108104467391968} 02/26/2022 01:34:16 - INFO - codeparrot_training - Step 37066: {'lr': 8.435315794330456e-05, 'samples': 18978304, 'steps': 37066, 'loss/train': 0.42056167125701904} 02/26/2022 01:34:21 - INFO - codeparrot_training - Step 37067: {'lr': 8.434090306571415e-05, 'samples': 18978816, 'steps': 37067, 'loss/train': 1.9125841856002808} 02/26/2022 01:34:25 - INFO - codeparrot_training - Step 37068: {'lr': 8.43286488977546e-05, 'samples': 18979328, 'steps': 37068, 'loss/train': 3.784729242324829} 02/26/2022 01:34:30 - INFO - codeparrot_training - Step 37069: {'lr': 8.43163954394786e-05, 'samples': 18979840, 'steps': 37069, 'loss/train': 1.6972370147705078} 02/26/2022 01:34:34 - INFO - codeparrot_training - Step 37070: {'lr': 8.430414269093856e-05, 'samples': 18980352, 'steps': 37070, 'loss/train': 1.8246432542800903} 02/26/2022 01:34:40 - INFO - codeparrot_training - Step 37071: {'lr': 8.429189065218695e-05, 'samples': 18980864, 'steps': 37071, 'loss/train': 1.7695115804672241} 02/26/2022 01:34:43 - INFO - codeparrot_training - Step 37072: {'lr': 8.42796393232762e-05, 'samples': 18981376, 'steps': 37072, 'loss/train': 1.4122822284698486} 02/26/2022 01:34:48 - INFO - codeparrot_training - Step 37073: {'lr': 8.42673887042589e-05, 'samples': 18981888, 'steps': 37073, 'loss/train': 1.8022700548171997} 02/26/2022 01:34:52 - INFO - codeparrot_training - Step 37074: {'lr': 8.425513879518751e-05, 'samples': 18982400, 'steps': 37074, 'loss/train': 2.4020347595214844} 02/26/2022 01:34:59 - INFO - codeparrot_training - Step 37075: {'lr': 8.424288959611445e-05, 'samples': 18982912, 'steps': 37075, 'loss/train': 2.6910476684570312} 02/26/2022 01:35:02 - INFO - codeparrot_training - Step 37076: {'lr': 8.423064110709216e-05, 'samples': 18983424, 'steps': 37076, 'loss/train': 1.259127140045166} 02/26/2022 01:35:08 - INFO - codeparrot_training - Step 37077: {'lr': 8.421839332817324e-05, 'samples': 18983936, 'steps': 37077, 'loss/train': 2.249293565750122} 02/26/2022 01:35:11 - INFO - codeparrot_training - Step 37078: {'lr': 8.420614625941009e-05, 'samples': 18984448, 'steps': 37078, 'loss/train': 1.480204463005066} 02/26/2022 01:35:17 - INFO - codeparrot_training - Step 37079: {'lr': 8.419389990085515e-05, 'samples': 18984960, 'steps': 37079, 'loss/train': 1.628015160560608} 02/26/2022 01:35:20 - INFO - codeparrot_training - Step 37080: {'lr': 8.41816542525608e-05, 'samples': 18985472, 'steps': 37080, 'loss/train': 1.1605656147003174} 02/26/2022 01:35:26 - INFO - codeparrot_training - Step 37081: {'lr': 8.41694093145797e-05, 'samples': 18985984, 'steps': 37081, 'loss/train': 0.7357677817344666} 02/26/2022 01:35:29 - INFO - codeparrot_training - Step 37082: {'lr': 8.415716508696416e-05, 'samples': 18986496, 'steps': 37082, 'loss/train': 2.522886037826538} 02/26/2022 01:35:35 - INFO - codeparrot_training - Step 37083: {'lr': 8.414492156976669e-05, 'samples': 18987008, 'steps': 37083, 'loss/train': 1.1562961339950562} 02/26/2022 01:35:38 - INFO - codeparrot_training - Step 37084: {'lr': 8.413267876303964e-05, 'samples': 18987520, 'steps': 37084, 'loss/train': 1.4469730854034424} 02/26/2022 01:35:45 - INFO - codeparrot_training - Step 37085: {'lr': 8.412043666683559e-05, 'samples': 18988032, 'steps': 37085, 'loss/train': 1.6758081912994385} 02/26/2022 01:35:48 - INFO - codeparrot_training - Step 37086: {'lr': 8.410819528120692e-05, 'samples': 18988544, 'steps': 37086, 'loss/train': 1.744964599609375} 02/26/2022 01:35:54 - INFO - codeparrot_training - Step 37087: {'lr': 8.409595460620605e-05, 'samples': 18989056, 'steps': 37087, 'loss/train': 0.8033421635627747} 02/26/2022 01:35:58 - INFO - codeparrot_training - Step 37088: {'lr': 8.408371464188536e-05, 'samples': 18989568, 'steps': 37088, 'loss/train': 1.8627506494522095} 02/26/2022 01:36:03 - INFO - codeparrot_training - Step 37089: {'lr': 8.407147538829743e-05, 'samples': 18990080, 'steps': 37089, 'loss/train': 2.1073427200317383} 02/26/2022 01:36:07 - INFO - codeparrot_training - Step 37090: {'lr': 8.405923684549464e-05, 'samples': 18990592, 'steps': 37090, 'loss/train': 1.9563995599746704} 02/26/2022 01:36:12 - INFO - codeparrot_training - Step 37091: {'lr': 8.404699901352936e-05, 'samples': 18991104, 'steps': 37091, 'loss/train': 1.3462523221969604} 02/26/2022 01:36:16 - INFO - codeparrot_training - Step 37092: {'lr': 8.403476189245404e-05, 'samples': 18991616, 'steps': 37092, 'loss/train': 1.350356101989746} 02/26/2022 01:36:21 - INFO - codeparrot_training - Step 37093: {'lr': 8.4022525482321e-05, 'samples': 18992128, 'steps': 37093, 'loss/train': 1.7961091995239258} 02/26/2022 01:36:25 - INFO - codeparrot_training - Step 37094: {'lr': 8.401028978318287e-05, 'samples': 18992640, 'steps': 37094, 'loss/train': 2.33398175239563} 02/26/2022 01:36:31 - INFO - codeparrot_training - Step 37095: {'lr': 8.399805479509188e-05, 'samples': 18993152, 'steps': 37095, 'loss/train': 1.6561747789382935} 02/26/2022 01:36:35 - INFO - codeparrot_training - Step 37096: {'lr': 8.398582051810064e-05, 'samples': 18993664, 'steps': 37096, 'loss/train': 2.04085111618042} 02/26/2022 01:36:40 - INFO - codeparrot_training - Step 37097: {'lr': 8.39735869522613e-05, 'samples': 18994176, 'steps': 37097, 'loss/train': 0.5589407086372375} 02/26/2022 01:36:44 - INFO - codeparrot_training - Step 37098: {'lr': 8.396135409762643e-05, 'samples': 18994688, 'steps': 37098, 'loss/train': 1.7441991567611694} 02/26/2022 01:36:49 - INFO - codeparrot_training - Step 37099: {'lr': 8.394912195424834e-05, 'samples': 18995200, 'steps': 37099, 'loss/train': 1.0083661079406738} 02/26/2022 01:36:53 - INFO - codeparrot_training - Step 37100: {'lr': 8.393689052217964e-05, 'samples': 18995712, 'steps': 37100, 'loss/train': 2.9657833576202393} 02/26/2022 01:36:58 - INFO - codeparrot_training - Step 37101: {'lr': 8.392465980147243e-05, 'samples': 18996224, 'steps': 37101, 'loss/train': 1.6664283275604248} 02/26/2022 01:37:02 - INFO - codeparrot_training - Step 37102: {'lr': 8.391242979217931e-05, 'samples': 18996736, 'steps': 37102, 'loss/train': 0.24867655336856842} 02/26/2022 01:37:07 - INFO - codeparrot_training - Step 37103: {'lr': 8.390020049435252e-05, 'samples': 18997248, 'steps': 37103, 'loss/train': 1.1057223081588745} 02/26/2022 01:37:11 - INFO - codeparrot_training - Step 37104: {'lr': 8.388797190804468e-05, 'samples': 18997760, 'steps': 37104, 'loss/train': 2.328911066055298} 02/26/2022 01:37:16 - INFO - codeparrot_training - Step 37105: {'lr': 8.387574403330786e-05, 'samples': 18998272, 'steps': 37105, 'loss/train': 1.9053987264633179} 02/26/2022 01:37:22 - INFO - codeparrot_training - Step 37106: {'lr': 8.38635168701947e-05, 'samples': 18998784, 'steps': 37106, 'loss/train': 1.6752281188964844} 02/26/2022 01:37:25 - INFO - codeparrot_training - Step 37107: {'lr': 8.385129041875738e-05, 'samples': 18999296, 'steps': 37107, 'loss/train': 1.4451733827590942} 02/26/2022 01:37:31 - INFO - codeparrot_training - Step 37108: {'lr': 8.383906467904848e-05, 'samples': 18999808, 'steps': 37108, 'loss/train': 0.29254552721977234} 02/26/2022 01:37:34 - INFO - codeparrot_training - Step 37109: {'lr': 8.382683965112026e-05, 'samples': 19000320, 'steps': 37109, 'loss/train': 1.0294314622879028} 02/26/2022 01:37:41 - INFO - codeparrot_training - Step 37110: {'lr': 8.381461533502507e-05, 'samples': 19000832, 'steps': 37110, 'loss/train': 2.5762734413146973} 02/26/2022 01:37:45 - INFO - codeparrot_training - Step 37111: {'lr': 8.380239173081525e-05, 'samples': 19001344, 'steps': 37111, 'loss/train': 1.4563014507293701} 02/26/2022 01:37:50 - INFO - codeparrot_training - Step 37112: {'lr': 8.379016883854327e-05, 'samples': 19001856, 'steps': 37112, 'loss/train': 1.1990443468093872} 02/26/2022 01:37:54 - INFO - codeparrot_training - Step 37113: {'lr': 8.377794665826145e-05, 'samples': 19002368, 'steps': 37113, 'loss/train': 2.0943968296051025} 02/26/2022 01:37:59 - INFO - codeparrot_training - Step 37114: {'lr': 8.37657251900221e-05, 'samples': 19002880, 'steps': 37114, 'loss/train': 0.9344554543495178} 02/26/2022 01:38:03 - INFO - codeparrot_training - Step 37115: {'lr': 8.37535044338775e-05, 'samples': 19003392, 'steps': 37115, 'loss/train': 1.2807188034057617} 02/26/2022 01:38:08 - INFO - codeparrot_training - Step 37116: {'lr': 8.374128438988021e-05, 'samples': 19003904, 'steps': 37116, 'loss/train': 1.7306482791900635} 02/26/2022 01:38:12 - INFO - codeparrot_training - Step 37117: {'lr': 8.372906505808236e-05, 'samples': 19004416, 'steps': 37117, 'loss/train': 1.8291391134262085} 02/26/2022 01:38:17 - INFO - codeparrot_training - Step 37118: {'lr': 8.371684643853658e-05, 'samples': 19004928, 'steps': 37118, 'loss/train': 0.6260422468185425} 02/26/2022 01:38:21 - INFO - codeparrot_training - Step 37119: {'lr': 8.370462853129488e-05, 'samples': 19005440, 'steps': 37119, 'loss/train': 2.3483242988586426} 02/26/2022 01:38:28 - INFO - codeparrot_training - Step 37120: {'lr': 8.369241133640981e-05, 'samples': 19005952, 'steps': 37120, 'loss/train': 1.5019205808639526} 02/26/2022 01:38:31 - INFO - codeparrot_training - Step 37121: {'lr': 8.368019485393355e-05, 'samples': 19006464, 'steps': 37121, 'loss/train': 1.6771647930145264} 02/26/2022 01:38:37 - INFO - codeparrot_training - Step 37122: {'lr': 8.366797908391871e-05, 'samples': 19006976, 'steps': 37122, 'loss/train': 2.6337990760803223} 02/26/2022 01:38:41 - INFO - codeparrot_training - Step 37123: {'lr': 8.365576402641726e-05, 'samples': 19007488, 'steps': 37123, 'loss/train': 1.7889511585235596} 02/26/2022 01:38:46 - INFO - codeparrot_training - Step 37124: {'lr': 8.364354968148177e-05, 'samples': 19008000, 'steps': 37124, 'loss/train': 1.8414673805236816} 02/26/2022 01:38:50 - INFO - codeparrot_training - Step 37125: {'lr': 8.363133604916447e-05, 'samples': 19008512, 'steps': 37125, 'loss/train': 1.8435741662979126} 02/26/2022 01:38:55 - INFO - codeparrot_training - Step 37126: {'lr': 8.361912312951772e-05, 'samples': 19009024, 'steps': 37126, 'loss/train': 1.9183146953582764} 02/26/2022 01:38:59 - INFO - codeparrot_training - Step 37127: {'lr': 8.360691092259373e-05, 'samples': 19009536, 'steps': 37127, 'loss/train': 1.0897890329360962} 02/26/2022 01:39:04 - INFO - codeparrot_training - Step 37128: {'lr': 8.3594699428445e-05, 'samples': 19010048, 'steps': 37128, 'loss/train': 2.2542364597320557} 02/26/2022 01:39:08 - INFO - codeparrot_training - Step 37129: {'lr': 8.358248864712373e-05, 'samples': 19010560, 'steps': 37129, 'loss/train': 1.9320359230041504} 02/26/2022 01:39:15 - INFO - codeparrot_training - Step 37130: {'lr': 8.357027857868212e-05, 'samples': 19011072, 'steps': 37130, 'loss/train': 2.602360963821411} 02/26/2022 01:39:18 - INFO - codeparrot_training - Step 37131: {'lr': 8.355806922317271e-05, 'samples': 19011584, 'steps': 37131, 'loss/train': 1.882552146911621} 02/26/2022 01:39:24 - INFO - codeparrot_training - Step 37132: {'lr': 8.354586058064766e-05, 'samples': 19012096, 'steps': 37132, 'loss/train': 1.3717589378356934} 02/26/2022 01:39:27 - INFO - codeparrot_training - Step 37133: {'lr': 8.353365265115931e-05, 'samples': 19012608, 'steps': 37133, 'loss/train': 0.8835031986236572} 02/26/2022 01:39:33 - INFO - codeparrot_training - Step 37134: {'lr': 8.352144543475981e-05, 'samples': 19013120, 'steps': 37134, 'loss/train': 1.2702559232711792} 02/26/2022 01:39:36 - INFO - codeparrot_training - Step 37135: {'lr': 8.35092389315017e-05, 'samples': 19013632, 'steps': 37135, 'loss/train': 2.1608073711395264} 02/26/2022 01:39:42 - INFO - codeparrot_training - Step 37136: {'lr': 8.349703314143711e-05, 'samples': 19014144, 'steps': 37136, 'loss/train': 0.49818217754364014} 02/26/2022 01:39:46 - INFO - codeparrot_training - Step 37137: {'lr': 8.348482806461838e-05, 'samples': 19014656, 'steps': 37137, 'loss/train': 2.246999740600586} 02/26/2022 01:39:49 - INFO - codeparrot_training - Step 37138: {'lr': 8.347262370109768e-05, 'samples': 19015168, 'steps': 37138, 'loss/train': 1.8392107486724854} 02/26/2022 01:39:55 - INFO - codeparrot_training - Step 37139: {'lr': 8.346042005092754e-05, 'samples': 19015680, 'steps': 37139, 'loss/train': 2.2710347175598145} 02/26/2022 01:39:58 - INFO - codeparrot_training - Step 37140: {'lr': 8.344821711415995e-05, 'samples': 19016192, 'steps': 37140, 'loss/train': 4.144713401794434} 02/26/2022 01:40:04 - INFO - codeparrot_training - Step 37141: {'lr': 8.343601489084738e-05, 'samples': 19016704, 'steps': 37141, 'loss/train': 1.1207575798034668} 02/26/2022 01:40:07 - INFO - codeparrot_training - Step 37142: {'lr': 8.342381338104191e-05, 'samples': 19017216, 'steps': 37142, 'loss/train': 1.3024550676345825} 02/26/2022 01:40:14 - INFO - codeparrot_training - Step 37143: {'lr': 8.341161258479605e-05, 'samples': 19017728, 'steps': 37143, 'loss/train': 1.420929193496704} 02/26/2022 01:40:18 - INFO - codeparrot_training - Step 37144: {'lr': 8.339941250216194e-05, 'samples': 19018240, 'steps': 37144, 'loss/train': 1.2910969257354736} 02/26/2022 01:40:23 - INFO - codeparrot_training - Step 37145: {'lr': 8.338721313319183e-05, 'samples': 19018752, 'steps': 37145, 'loss/train': 1.5712597370147705} 02/26/2022 01:40:27 - INFO - codeparrot_training - Step 37146: {'lr': 8.337501447793791e-05, 'samples': 19019264, 'steps': 37146, 'loss/train': 1.859776496887207} 02/26/2022 01:40:32 - INFO - codeparrot_training - Step 37147: {'lr': 8.336281653645261e-05, 'samples': 19019776, 'steps': 37147, 'loss/train': 0.035972096025943756} 02/26/2022 01:40:36 - INFO - codeparrot_training - Step 37148: {'lr': 8.33506193087881e-05, 'samples': 19020288, 'steps': 37148, 'loss/train': 1.9353622198104858} 02/26/2022 01:40:41 - INFO - codeparrot_training - Step 37149: {'lr': 8.333842279499663e-05, 'samples': 19020800, 'steps': 37149, 'loss/train': 0.588671088218689} 02/26/2022 01:40:45 - INFO - codeparrot_training - Step 37150: {'lr': 8.332622699513032e-05, 'samples': 19021312, 'steps': 37150, 'loss/train': 2.456263780593872} 02/26/2022 01:40:50 - INFO - codeparrot_training - Step 37151: {'lr': 8.331403190924164e-05, 'samples': 19021824, 'steps': 37151, 'loss/train': 1.5952991247177124} 02/26/2022 01:40:54 - INFO - codeparrot_training - Step 37152: {'lr': 8.330183753738269e-05, 'samples': 19022336, 'steps': 37152, 'loss/train': 2.2287089824676514} 02/26/2022 01:40:59 - INFO - codeparrot_training - Step 37153: {'lr': 8.328964387960574e-05, 'samples': 19022848, 'steps': 37153, 'loss/train': 1.2397838830947876} 02/26/2022 01:41:03 - INFO - codeparrot_training - Step 37154: {'lr': 8.327745093596295e-05, 'samples': 19023360, 'steps': 37154, 'loss/train': 3.152423143386841} 02/26/2022 01:41:08 - INFO - codeparrot_training - Step 37155: {'lr': 8.32652587065067e-05, 'samples': 19023872, 'steps': 37155, 'loss/train': 1.4935261011123657} 02/26/2022 01:41:12 - INFO - codeparrot_training - Step 37156: {'lr': 8.325306719128906e-05, 'samples': 19024384, 'steps': 37156, 'loss/train': 1.400776743888855} 02/26/2022 01:41:18 - INFO - codeparrot_training - Step 37157: {'lr': 8.324087639036248e-05, 'samples': 19024896, 'steps': 37157, 'loss/train': 1.7071540355682373} 02/26/2022 01:41:24 - INFO - codeparrot_training - Step 37158: {'lr': 8.322868630377889e-05, 'samples': 19025408, 'steps': 37158, 'loss/train': 1.6731566190719604} 02/26/2022 01:41:27 - INFO - codeparrot_training - Step 37159: {'lr': 8.321649693159072e-05, 'samples': 19025920, 'steps': 37159, 'loss/train': 0.8045114874839783} 02/26/2022 01:41:33 - INFO - codeparrot_training - Step 37160: {'lr': 8.320430827385004e-05, 'samples': 19026432, 'steps': 37160, 'loss/train': 1.7433161735534668} 02/26/2022 01:41:36 - INFO - codeparrot_training - Step 37161: {'lr': 8.31921203306093e-05, 'samples': 19026944, 'steps': 37161, 'loss/train': 0.38707539439201355} 02/26/2022 01:41:42 - INFO - codeparrot_training - Step 37162: {'lr': 8.317993310192041e-05, 'samples': 19027456, 'steps': 37162, 'loss/train': 1.9996532201766968} 02/26/2022 01:41:45 - INFO - codeparrot_training - Step 37163: {'lr': 8.316774658783577e-05, 'samples': 19027968, 'steps': 37163, 'loss/train': 1.295071005821228} 02/26/2022 01:41:51 - INFO - codeparrot_training - Step 37164: {'lr': 8.315556078840744e-05, 'samples': 19028480, 'steps': 37164, 'loss/train': 1.472206950187683} 02/26/2022 01:41:54 - INFO - codeparrot_training - Step 37165: {'lr': 8.314337570368787e-05, 'samples': 19028992, 'steps': 37165, 'loss/train': 1.3874530792236328} 02/26/2022 01:42:01 - INFO - codeparrot_training - Step 37166: {'lr': 8.313119133372896e-05, 'samples': 19029504, 'steps': 37166, 'loss/train': 1.5878446102142334} 02/26/2022 01:42:04 - INFO - codeparrot_training - Step 37167: {'lr': 8.31190076785831e-05, 'samples': 19030016, 'steps': 37167, 'loss/train': 2.5487284660339355} 02/26/2022 01:42:10 - INFO - codeparrot_training - Step 37168: {'lr': 8.310682473830236e-05, 'samples': 19030528, 'steps': 37168, 'loss/train': 1.359023928642273} 02/26/2022 01:42:13 - INFO - codeparrot_training - Step 37169: {'lr': 8.309464251293905e-05, 'samples': 19031040, 'steps': 37169, 'loss/train': 1.8767603635787964} 02/26/2022 01:42:19 - INFO - codeparrot_training - Step 37170: {'lr': 8.308246100254529e-05, 'samples': 19031552, 'steps': 37170, 'loss/train': 1.4964475631713867} 02/26/2022 01:42:22 - INFO - codeparrot_training - Step 37171: {'lr': 8.307028020717328e-05, 'samples': 19032064, 'steps': 37171, 'loss/train': 0.9305400848388672} 02/26/2022 01:42:28 - INFO - codeparrot_training - Step 37172: {'lr': 8.305810012687518e-05, 'samples': 19032576, 'steps': 37172, 'loss/train': 2.2013516426086426} 02/26/2022 01:42:31 - INFO - codeparrot_training - Step 37173: {'lr': 8.304592076170309e-05, 'samples': 19033088, 'steps': 37173, 'loss/train': 1.4422435760498047} 02/26/2022 01:42:37 - INFO - codeparrot_training - Step 37174: {'lr': 8.303374211170933e-05, 'samples': 19033600, 'steps': 37174, 'loss/train': 1.7583857774734497} 02/26/2022 01:42:40 - INFO - codeparrot_training - Step 37175: {'lr': 8.302156417694601e-05, 'samples': 19034112, 'steps': 37175, 'loss/train': 1.3351573944091797} 02/26/2022 01:42:46 - INFO - codeparrot_training - Step 37176: {'lr': 8.300938695746529e-05, 'samples': 19034624, 'steps': 37176, 'loss/train': 1.1454607248306274} 02/26/2022 01:42:49 - INFO - codeparrot_training - Step 37177: {'lr': 8.299721045331923e-05, 'samples': 19035136, 'steps': 37177, 'loss/train': 1.6435911655426025} 02/26/2022 01:42:56 - INFO - codeparrot_training - Step 37178: {'lr': 8.29850346645602e-05, 'samples': 19035648, 'steps': 37178, 'loss/train': 1.7206287384033203} 02/26/2022 01:43:00 - INFO - codeparrot_training - Step 37179: {'lr': 8.297285959124023e-05, 'samples': 19036160, 'steps': 37179, 'loss/train': 0.9274430871009827} 02/26/2022 01:43:05 - INFO - codeparrot_training - Step 37180: {'lr': 8.296068523341152e-05, 'samples': 19036672, 'steps': 37180, 'loss/train': 1.7284140586853027} 02/26/2022 01:43:09 - INFO - codeparrot_training - Step 37181: {'lr': 8.294851159112606e-05, 'samples': 19037184, 'steps': 37181, 'loss/train': 1.8588786125183105} 02/26/2022 01:43:15 - INFO - codeparrot_training - Step 37182: {'lr': 8.293633866443625e-05, 'samples': 19037696, 'steps': 37182, 'loss/train': 2.407111644744873} 02/26/2022 01:43:18 - INFO - codeparrot_training - Step 37183: {'lr': 8.292416645339413e-05, 'samples': 19038208, 'steps': 37183, 'loss/train': 2.17269229888916} 02/26/2022 01:43:23 - INFO - codeparrot_training - Step 37184: {'lr': 8.29119949580518e-05, 'samples': 19038720, 'steps': 37184, 'loss/train': 0.7622824907302856} 02/26/2022 01:43:27 - INFO - codeparrot_training - Step 37185: {'lr': 8.289982417846134e-05, 'samples': 19039232, 'steps': 37185, 'loss/train': 0.922652542591095} 02/26/2022 01:43:33 - INFO - codeparrot_training - Step 37186: {'lr': 8.288765411467508e-05, 'samples': 19039744, 'steps': 37186, 'loss/train': 2.1901113986968994} 02/26/2022 01:43:36 - INFO - codeparrot_training - Step 37187: {'lr': 8.287548476674502e-05, 'samples': 19040256, 'steps': 37187, 'loss/train': 1.3653470277786255} 02/26/2022 01:43:43 - INFO - codeparrot_training - Step 37188: {'lr': 8.286331613472334e-05, 'samples': 19040768, 'steps': 37188, 'loss/train': 2.1037349700927734} 02/26/2022 01:43:47 - INFO - codeparrot_training - Step 37189: {'lr': 8.285114821866204e-05, 'samples': 19041280, 'steps': 37189, 'loss/train': 1.6034964323043823} 02/26/2022 01:43:52 - INFO - codeparrot_training - Step 37190: {'lr': 8.283898101861345e-05, 'samples': 19041792, 'steps': 37190, 'loss/train': 1.3340699672698975} 02/26/2022 01:43:56 - INFO - codeparrot_training - Step 37191: {'lr': 8.282681453462948e-05, 'samples': 19042304, 'steps': 37191, 'loss/train': 1.359729528427124} 02/26/2022 01:44:01 - INFO - codeparrot_training - Step 37192: {'lr': 8.281464876676253e-05, 'samples': 19042816, 'steps': 37192, 'loss/train': 2.09037446975708} 02/26/2022 01:44:05 - INFO - codeparrot_training - Step 37193: {'lr': 8.280248371506438e-05, 'samples': 19043328, 'steps': 37193, 'loss/train': 1.3655472993850708} 02/26/2022 01:44:10 - INFO - codeparrot_training - Step 37194: {'lr': 8.279031937958736e-05, 'samples': 19043840, 'steps': 37194, 'loss/train': 2.0710277557373047} 02/26/2022 01:44:14 - INFO - codeparrot_training - Step 37195: {'lr': 8.277815576038345e-05, 'samples': 19044352, 'steps': 37195, 'loss/train': 1.6258600950241089} 02/26/2022 01:44:19 - INFO - codeparrot_training - Step 37196: {'lr': 8.276599285750499e-05, 'samples': 19044864, 'steps': 37196, 'loss/train': 1.4014722108840942} 02/26/2022 01:44:23 - INFO - codeparrot_training - Step 37197: {'lr': 8.275383067100376e-05, 'samples': 19045376, 'steps': 37197, 'loss/train': 1.31800377368927} 02/26/2022 01:44:28 - INFO - codeparrot_training - Step 37198: {'lr': 8.274166920093207e-05, 'samples': 19045888, 'steps': 37198, 'loss/train': 1.0466171503067017} 02/26/2022 01:44:32 - INFO - codeparrot_training - Step 37199: {'lr': 8.272950844734186e-05, 'samples': 19046400, 'steps': 37199, 'loss/train': 1.8293547630310059} 02/26/2022 01:44:37 - INFO - codeparrot_training - Step 37200: {'lr': 8.271734841028553e-05, 'samples': 19046912, 'steps': 37200, 'loss/train': 1.7393453121185303} 02/26/2022 01:44:41 - INFO - codeparrot_training - Step 37201: {'lr': 8.270518908981476e-05, 'samples': 19047424, 'steps': 37201, 'loss/train': 2.0252933502197266} 02/26/2022 01:44:46 - INFO - codeparrot_training - Step 37202: {'lr': 8.269303048598195e-05, 'samples': 19047936, 'steps': 37202, 'loss/train': 1.845220923423767} 02/26/2022 01:44:50 - INFO - codeparrot_training - Step 37203: {'lr': 8.268087259883897e-05, 'samples': 19048448, 'steps': 37203, 'loss/train': 1.7153112888336182} 02/26/2022 01:44:56 - INFO - codeparrot_training - Step 37204: {'lr': 8.266871542843809e-05, 'samples': 19048960, 'steps': 37204, 'loss/train': 1.5581583976745605} 02/26/2022 01:45:00 - INFO - codeparrot_training - Step 37205: {'lr': 8.26565589748313e-05, 'samples': 19049472, 'steps': 37205, 'loss/train': 2.1535165309906006} 02/26/2022 01:45:05 - INFO - codeparrot_training - Step 37206: {'lr': 8.264440323807065e-05, 'samples': 19049984, 'steps': 37206, 'loss/train': 1.6526063680648804} 02/26/2022 01:45:09 - INFO - codeparrot_training - Step 37207: {'lr': 8.26322482182082e-05, 'samples': 19050496, 'steps': 37207, 'loss/train': 1.8197089433670044} 02/26/2022 01:45:14 - INFO - codeparrot_training - Step 37208: {'lr': 8.262009391529609e-05, 'samples': 19051008, 'steps': 37208, 'loss/train': 0.7985866665840149} 02/26/2022 01:45:18 - INFO - codeparrot_training - Step 37209: {'lr': 8.260794032938637e-05, 'samples': 19051520, 'steps': 37209, 'loss/train': 1.6394752264022827} 02/26/2022 01:45:23 - INFO - codeparrot_training - Step 37210: {'lr': 8.259578746053109e-05, 'samples': 19052032, 'steps': 37210, 'loss/train': 1.7461438179016113} 02/26/2022 01:45:27 - INFO - codeparrot_training - Step 37211: {'lr': 8.258363530878222e-05, 'samples': 19052544, 'steps': 37211, 'loss/train': 1.7207915782928467} 02/26/2022 01:45:32 - INFO - codeparrot_training - Step 37212: {'lr': 8.257148387419194e-05, 'samples': 19053056, 'steps': 37212, 'loss/train': 1.7985799312591553} 02/26/2022 01:45:36 - INFO - codeparrot_training - Step 37213: {'lr': 8.25593331568123e-05, 'samples': 19053568, 'steps': 37213, 'loss/train': 1.9586187601089478} 02/26/2022 01:45:42 - INFO - codeparrot_training - Step 37214: {'lr': 8.254718315669529e-05, 'samples': 19054080, 'steps': 37214, 'loss/train': 2.7502715587615967} 02/26/2022 01:45:46 - INFO - codeparrot_training - Step 37215: {'lr': 8.253503387389289e-05, 'samples': 19054592, 'steps': 37215, 'loss/train': 1.0251836776733398} 02/26/2022 01:45:51 - INFO - codeparrot_training - Step 37216: {'lr': 8.252288530845733e-05, 'samples': 19055104, 'steps': 37216, 'loss/train': 1.1509389877319336} 02/26/2022 01:45:55 - INFO - codeparrot_training - Step 37217: {'lr': 8.251073746044053e-05, 'samples': 19055616, 'steps': 37217, 'loss/train': 1.3730127811431885} 02/26/2022 01:46:00 - INFO - codeparrot_training - Step 37218: {'lr': 8.249859032989454e-05, 'samples': 19056128, 'steps': 37218, 'loss/train': 0.052121300250291824} 02/26/2022 01:46:04 - INFO - codeparrot_training - Step 37219: {'lr': 8.248644391687136e-05, 'samples': 19056640, 'steps': 37219, 'loss/train': 2.6654324531555176} 02/26/2022 01:46:09 - INFO - codeparrot_training - Step 37220: {'lr': 8.247429822142311e-05, 'samples': 19057152, 'steps': 37220, 'loss/train': 0.9788916707038879} 02/26/2022 01:46:13 - INFO - codeparrot_training - Step 37221: {'lr': 8.246215324360177e-05, 'samples': 19057664, 'steps': 37221, 'loss/train': 1.7591363191604614} 02/26/2022 01:46:18 - INFO - codeparrot_training - Step 37222: {'lr': 8.245000898345936e-05, 'samples': 19058176, 'steps': 37222, 'loss/train': 0.528789222240448} 02/26/2022 01:46:22 - INFO - codeparrot_training - Step 37223: {'lr': 8.243786544104792e-05, 'samples': 19058688, 'steps': 37223, 'loss/train': 0.2821579873561859} 02/26/2022 01:46:28 - INFO - codeparrot_training - Step 37224: {'lr': 8.242572261641939e-05, 'samples': 19059200, 'steps': 37224, 'loss/train': 3.1132941246032715} 02/26/2022 01:46:32 - INFO - codeparrot_training - Step 37225: {'lr': 8.241358050962591e-05, 'samples': 19059712, 'steps': 37225, 'loss/train': 0.6978120803833008} 02/26/2022 01:46:37 - INFO - codeparrot_training - Step 37226: {'lr': 8.240143912071937e-05, 'samples': 19060224, 'steps': 37226, 'loss/train': 1.2478735446929932} 02/26/2022 01:46:41 - INFO - codeparrot_training - Step 37227: {'lr': 8.238929844975201e-05, 'samples': 19060736, 'steps': 37227, 'loss/train': 2.6061248779296875} 02/26/2022 01:46:46 - INFO - codeparrot_training - Step 37228: {'lr': 8.237715849677552e-05, 'samples': 19061248, 'steps': 37228, 'loss/train': 0.9349194169044495} 02/26/2022 01:46:50 - INFO - codeparrot_training - Step 37229: {'lr': 8.236501926184212e-05, 'samples': 19061760, 'steps': 37229, 'loss/train': 0.5367432832717896} 02/26/2022 01:46:55 - INFO - codeparrot_training - Step 37230: {'lr': 8.235288074500366e-05, 'samples': 19062272, 'steps': 37230, 'loss/train': 1.963213562965393} 02/26/2022 01:46:59 - INFO - codeparrot_training - Step 37231: {'lr': 8.234074294631241e-05, 'samples': 19062784, 'steps': 37231, 'loss/train': 2.2465946674346924} 02/26/2022 01:47:04 - INFO - codeparrot_training - Step 37232: {'lr': 8.232860586582e-05, 'samples': 19063296, 'steps': 37232, 'loss/train': 0.9786977171897888} 02/26/2022 01:47:08 - INFO - codeparrot_training - Step 37233: {'lr': 8.231646950357868e-05, 'samples': 19063808, 'steps': 37233, 'loss/train': 2.409968852996826} 02/26/2022 01:47:14 - INFO - codeparrot_training - Step 37234: {'lr': 8.230433385964028e-05, 'samples': 19064320, 'steps': 37234, 'loss/train': 1.1553808450698853} 02/26/2022 01:47:18 - INFO - codeparrot_training - Step 37235: {'lr': 8.229219893405704e-05, 'samples': 19064832, 'steps': 37235, 'loss/train': 1.9721091985702515} 02/26/2022 01:47:23 - INFO - codeparrot_training - Step 37236: {'lr': 8.22800647268806e-05, 'samples': 19065344, 'steps': 37236, 'loss/train': 1.7504615783691406} 02/26/2022 01:47:27 - INFO - codeparrot_training - Step 37237: {'lr': 8.22679312381632e-05, 'samples': 19065856, 'steps': 37237, 'loss/train': 1.0252844095230103} 02/26/2022 01:47:32 - INFO - codeparrot_training - Step 37238: {'lr': 8.22557984679566e-05, 'samples': 19066368, 'steps': 37238, 'loss/train': 1.9307035207748413} 02/26/2022 01:47:36 - INFO - codeparrot_training - Step 37239: {'lr': 8.224366641631301e-05, 'samples': 19066880, 'steps': 37239, 'loss/train': 2.4583001136779785} 02/26/2022 01:47:41 - INFO - codeparrot_training - Step 37240: {'lr': 8.223153508328429e-05, 'samples': 19067392, 'steps': 37240, 'loss/train': 1.5907905101776123} 02/26/2022 01:47:45 - INFO - codeparrot_training - Step 37241: {'lr': 8.221940446892236e-05, 'samples': 19067904, 'steps': 37241, 'loss/train': 0.6465444564819336} 02/26/2022 01:47:50 - INFO - codeparrot_training - Step 37242: {'lr': 8.220727457327917e-05, 'samples': 19068416, 'steps': 37242, 'loss/train': 1.1750423908233643} 02/26/2022 01:47:54 - INFO - codeparrot_training - Step 37243: {'lr': 8.219514539640682e-05, 'samples': 19068928, 'steps': 37243, 'loss/train': 2.238308906555176} 02/26/2022 01:47:59 - INFO - codeparrot_training - Step 37244: {'lr': 8.218301693835719e-05, 'samples': 19069440, 'steps': 37244, 'loss/train': 1.9969607591629028} 02/26/2022 01:48:03 - INFO - codeparrot_training - Step 37245: {'lr': 8.217088919918223e-05, 'samples': 19069952, 'steps': 37245, 'loss/train': 1.8884025812149048} 02/26/2022 01:48:08 - INFO - codeparrot_training - Step 37246: {'lr': 8.215876217893379e-05, 'samples': 19070464, 'steps': 37246, 'loss/train': 1.062110424041748} 02/26/2022 01:48:12 - INFO - codeparrot_training - Step 37247: {'lr': 8.2146635877664e-05, 'samples': 19070976, 'steps': 37247, 'loss/train': 1.6500635147094727} 02/26/2022 01:48:18 - INFO - codeparrot_training - Step 37248: {'lr': 8.213451029542471e-05, 'samples': 19071488, 'steps': 37248, 'loss/train': 0.808623194694519} 02/26/2022 01:48:22 - INFO - codeparrot_training - Step 37249: {'lr': 8.212238543226791e-05, 'samples': 19072000, 'steps': 37249, 'loss/train': 2.1174252033233643} 02/26/2022 01:48:27 - INFO - codeparrot_training - Step 37250: {'lr': 8.211026128824539e-05, 'samples': 19072512, 'steps': 37250, 'loss/train': 1.0835115909576416} 02/26/2022 01:48:31 - INFO - codeparrot_training - Step 37251: {'lr': 8.209813786340927e-05, 'samples': 19073024, 'steps': 37251, 'loss/train': 1.1820204257965088} 02/26/2022 01:48:36 - INFO - codeparrot_training - Step 37252: {'lr': 8.208601515781137e-05, 'samples': 19073536, 'steps': 37252, 'loss/train': 2.7489683628082275} 02/26/2022 01:48:40 - INFO - codeparrot_training - Step 37253: {'lr': 8.207389317150382e-05, 'samples': 19074048, 'steps': 37253, 'loss/train': 2.659041404724121} 02/26/2022 01:48:45 - INFO - codeparrot_training - Step 37254: {'lr': 8.20617719045382e-05, 'samples': 19074560, 'steps': 37254, 'loss/train': 1.430891752243042} 02/26/2022 01:48:49 - INFO - codeparrot_training - Step 37255: {'lr': 8.204965135696674e-05, 'samples': 19075072, 'steps': 37255, 'loss/train': 1.758187174797058} 02/26/2022 01:48:54 - INFO - codeparrot_training - Step 37256: {'lr': 8.203753152884122e-05, 'samples': 19075584, 'steps': 37256, 'loss/train': 1.899909496307373} 02/26/2022 01:48:58 - INFO - codeparrot_training - Step 37257: {'lr': 8.202541242021358e-05, 'samples': 19076096, 'steps': 37257, 'loss/train': 1.2651687860488892} 02/26/2022 01:49:04 - INFO - codeparrot_training - Step 37258: {'lr': 8.201329403113566e-05, 'samples': 19076608, 'steps': 37258, 'loss/train': 1.2146570682525635} 02/26/2022 01:49:07 - INFO - codeparrot_training - Step 37259: {'lr': 8.200117636165952e-05, 'samples': 19077120, 'steps': 37259, 'loss/train': 1.2071138620376587} 02/26/2022 01:49:13 - INFO - codeparrot_training - Step 37260: {'lr': 8.198905941183701e-05, 'samples': 19077632, 'steps': 37260, 'loss/train': 0.5950890779495239} 02/26/2022 01:49:17 - INFO - codeparrot_training - Step 37261: {'lr': 8.197694318172e-05, 'samples': 19078144, 'steps': 37261, 'loss/train': 1.4595544338226318} 02/26/2022 01:49:22 - INFO - codeparrot_training - Step 37262: {'lr': 8.196482767136035e-05, 'samples': 19078656, 'steps': 37262, 'loss/train': 1.4657697677612305} 02/26/2022 01:49:26 - INFO - codeparrot_training - Step 37263: {'lr': 8.195271288081008e-05, 'samples': 19079168, 'steps': 37263, 'loss/train': 1.6008533239364624} 02/26/2022 01:49:31 - INFO - codeparrot_training - Step 37264: {'lr': 8.194059881012105e-05, 'samples': 19079680, 'steps': 37264, 'loss/train': 0.9014922380447388} 02/26/2022 01:49:35 - INFO - codeparrot_training - Step 37265: {'lr': 8.192848545934504e-05, 'samples': 19080192, 'steps': 37265, 'loss/train': 1.6199990510940552} 02/26/2022 01:49:40 - INFO - codeparrot_training - Step 37266: {'lr': 8.191637282853412e-05, 'samples': 19080704, 'steps': 37266, 'loss/train': 1.428519606590271} 02/26/2022 01:49:44 - INFO - codeparrot_training - Step 37267: {'lr': 8.190426091774006e-05, 'samples': 19081216, 'steps': 37267, 'loss/train': 1.2228596210479736} 02/26/2022 01:49:50 - INFO - codeparrot_training - Step 37268: {'lr': 8.189214972701478e-05, 'samples': 19081728, 'steps': 37268, 'loss/train': 1.8246833086013794} 02/26/2022 01:49:53 - INFO - codeparrot_training - Step 37269: {'lr': 8.188003925641008e-05, 'samples': 19082240, 'steps': 37269, 'loss/train': 0.9061700701713562} 02/26/2022 01:50:00 - INFO - codeparrot_training - Step 37270: {'lr': 8.186792950597807e-05, 'samples': 19082752, 'steps': 37270, 'loss/train': 1.2816135883331299} 02/26/2022 01:50:03 - INFO - codeparrot_training - Step 37271: {'lr': 8.185582047577029e-05, 'samples': 19083264, 'steps': 37271, 'loss/train': 1.4242651462554932} 02/26/2022 01:50:08 - INFO - codeparrot_training - Step 37272: {'lr': 8.184371216583888e-05, 'samples': 19083776, 'steps': 37272, 'loss/train': 1.7851136922836304} 02/26/2022 01:50:12 - INFO - codeparrot_training - Step 37273: {'lr': 8.18316045762355e-05, 'samples': 19084288, 'steps': 37273, 'loss/train': 2.043785572052002} 02/26/2022 01:50:18 - INFO - codeparrot_training - Step 37274: {'lr': 8.181949770701233e-05, 'samples': 19084800, 'steps': 37274, 'loss/train': 2.125408411026001} 02/26/2022 01:50:21 - INFO - codeparrot_training - Step 37275: {'lr': 8.180739155822084e-05, 'samples': 19085312, 'steps': 37275, 'loss/train': 1.7213927507400513} 02/26/2022 01:50:27 - INFO - codeparrot_training - Step 37276: {'lr': 8.179528612991316e-05, 'samples': 19085824, 'steps': 37276, 'loss/train': 2.7691538333892822} 02/26/2022 01:50:30 - INFO - codeparrot_training - Step 37277: {'lr': 8.178318142214097e-05, 'samples': 19086336, 'steps': 37277, 'loss/train': 1.141066312789917} 02/26/2022 01:50:36 - INFO - codeparrot_training - Step 37278: {'lr': 8.177107743495632e-05, 'samples': 19086848, 'steps': 37278, 'loss/train': 2.1283280849456787} 02/26/2022 01:50:39 - INFO - codeparrot_training - Step 37279: {'lr': 8.175897416841096e-05, 'samples': 19087360, 'steps': 37279, 'loss/train': 0.5556949973106384} 02/26/2022 01:50:46 - INFO - codeparrot_training - Step 37280: {'lr': 8.174687162255672e-05, 'samples': 19087872, 'steps': 37280, 'loss/train': 0.8359056711196899} 02/26/2022 01:50:50 - INFO - codeparrot_training - Step 37281: {'lr': 8.173476979744538e-05, 'samples': 19088384, 'steps': 37281, 'loss/train': 0.21715198457241058} 02/26/2022 01:50:55 - INFO - codeparrot_training - Step 37282: {'lr': 8.172266869312892e-05, 'samples': 19088896, 'steps': 37282, 'loss/train': 1.745500922203064} 02/26/2022 01:51:01 - INFO - codeparrot_training - Step 37283: {'lr': 8.171056830965911e-05, 'samples': 19089408, 'steps': 37283, 'loss/train': 2.3627512454986572} 02/26/2022 01:51:04 - INFO - codeparrot_training - Step 37284: {'lr': 8.169846864708783e-05, 'samples': 19089920, 'steps': 37284, 'loss/train': 0.6568889021873474} 02/26/2022 01:51:10 - INFO - codeparrot_training - Step 37285: {'lr': 8.168636970546675e-05, 'samples': 19090432, 'steps': 37285, 'loss/train': 1.143664002418518} 02/26/2022 01:51:13 - INFO - codeparrot_training - Step 37286: {'lr': 8.167427148484791e-05, 'samples': 19090944, 'steps': 37286, 'loss/train': 1.9979159832000732} 02/26/2022 01:51:19 - INFO - codeparrot_training - Step 37287: {'lr': 8.166217398528295e-05, 'samples': 19091456, 'steps': 37287, 'loss/train': 1.9059408903121948} 02/26/2022 01:51:22 - INFO - codeparrot_training - Step 37288: {'lr': 8.165007720682396e-05, 'samples': 19091968, 'steps': 37288, 'loss/train': 0.2532229721546173} 02/26/2022 01:51:28 - INFO - codeparrot_training - Step 37289: {'lr': 8.163798114952242e-05, 'samples': 19092480, 'steps': 37289, 'loss/train': 2.5167734622955322} 02/26/2022 01:51:31 - INFO - codeparrot_training - Step 37290: {'lr': 8.162588581343036e-05, 'samples': 19092992, 'steps': 37290, 'loss/train': 1.857791543006897} 02/26/2022 01:51:37 - INFO - codeparrot_training - Step 37291: {'lr': 8.16137911985995e-05, 'samples': 19093504, 'steps': 37291, 'loss/train': 0.4462227523326874} 02/26/2022 01:51:40 - INFO - codeparrot_training - Step 37292: {'lr': 8.160169730508182e-05, 'samples': 19094016, 'steps': 37292, 'loss/train': 1.4387286901474} 02/26/2022 01:51:46 - INFO - codeparrot_training - Step 37293: {'lr': 8.158960413292884e-05, 'samples': 19094528, 'steps': 37293, 'loss/train': 1.322270154953003} 02/26/2022 01:51:49 - INFO - codeparrot_training - Step 37294: {'lr': 8.157751168219262e-05, 'samples': 19095040, 'steps': 37294, 'loss/train': 3.5032103061676025} 02/26/2022 01:51:56 - INFO - codeparrot_training - Step 37295: {'lr': 8.156541995292477e-05, 'samples': 19095552, 'steps': 37295, 'loss/train': 1.9252204895019531} 02/26/2022 01:52:00 - INFO - codeparrot_training - Step 37296: {'lr': 8.155332894517734e-05, 'samples': 19096064, 'steps': 37296, 'loss/train': 1.2596262693405151} 02/26/2022 01:52:05 - INFO - codeparrot_training - Step 37297: {'lr': 8.154123865900179e-05, 'samples': 19096576, 'steps': 37297, 'loss/train': 1.458268404006958} 02/26/2022 01:52:09 - INFO - codeparrot_training - Step 37298: {'lr': 8.152914909445016e-05, 'samples': 19097088, 'steps': 37298, 'loss/train': 2.333829402923584} 02/26/2022 01:52:14 - INFO - codeparrot_training - Step 37299: {'lr': 8.15170602515741e-05, 'samples': 19097600, 'steps': 37299, 'loss/train': 2.772134304046631} 02/26/2022 01:52:18 - INFO - codeparrot_training - Step 37300: {'lr': 8.150497213042551e-05, 'samples': 19098112, 'steps': 37300, 'loss/train': 2.3670170307159424} 02/26/2022 01:52:23 - INFO - codeparrot_training - Step 37301: {'lr': 8.149288473105616e-05, 'samples': 19098624, 'steps': 37301, 'loss/train': 2.0371432304382324} 02/26/2022 01:52:27 - INFO - codeparrot_training - Step 37302: {'lr': 8.148079805351775e-05, 'samples': 19099136, 'steps': 37302, 'loss/train': 1.5605242252349854} 02/26/2022 01:52:33 - INFO - codeparrot_training - Step 37303: {'lr': 8.146871209786208e-05, 'samples': 19099648, 'steps': 37303, 'loss/train': 1.6081106662750244} 02/26/2022 01:52:36 - INFO - codeparrot_training - Step 37304: {'lr': 8.145662686414085e-05, 'samples': 19100160, 'steps': 37304, 'loss/train': 1.1677380800247192} 02/26/2022 01:52:40 - INFO - codeparrot_training - Step 37305: {'lr': 8.144454235240603e-05, 'samples': 19100672, 'steps': 37305, 'loss/train': 2.038557291030884} 02/26/2022 01:52:47 - INFO - codeparrot_training - Step 37306: {'lr': 8.143245856270922e-05, 'samples': 19101184, 'steps': 37306, 'loss/train': 2.5647871494293213} 02/26/2022 01:52:50 - INFO - codeparrot_training - Step 37307: {'lr': 8.142037549510229e-05, 'samples': 19101696, 'steps': 37307, 'loss/train': 1.7239336967468262} 02/26/2022 01:52:56 - INFO - codeparrot_training - Step 37308: {'lr': 8.14082931496368e-05, 'samples': 19102208, 'steps': 37308, 'loss/train': 2.8829071521759033} 02/26/2022 01:52:59 - INFO - codeparrot_training - Step 37309: {'lr': 8.139621152636476e-05, 'samples': 19102720, 'steps': 37309, 'loss/train': 2.682356357574463} 02/26/2022 01:53:05 - INFO - codeparrot_training - Step 37310: {'lr': 8.138413062533781e-05, 'samples': 19103232, 'steps': 37310, 'loss/train': 1.5462348461151123} 02/26/2022 01:53:08 - INFO - codeparrot_training - Step 37311: {'lr': 8.13720504466077e-05, 'samples': 19103744, 'steps': 37311, 'loss/train': 2.713484048843384} 02/26/2022 01:53:14 - INFO - codeparrot_training - Step 37312: {'lr': 8.13599709902261e-05, 'samples': 19104256, 'steps': 37312, 'loss/train': 1.5032546520233154} 02/26/2022 01:53:17 - INFO - codeparrot_training - Step 37313: {'lr': 8.134789225624491e-05, 'samples': 19104768, 'steps': 37313, 'loss/train': 1.992012619972229} 02/26/2022 01:53:23 - INFO - codeparrot_training - Step 37314: {'lr': 8.133581424471581e-05, 'samples': 19105280, 'steps': 37314, 'loss/train': 1.5557827949523926} 02/26/2022 01:53:26 - INFO - codeparrot_training - Step 37315: {'lr': 8.132373695569053e-05, 'samples': 19105792, 'steps': 37315, 'loss/train': 1.7976577281951904} 02/26/2022 01:53:33 - INFO - codeparrot_training - Step 37316: {'lr': 8.131166038922072e-05, 'samples': 19106304, 'steps': 37316, 'loss/train': 1.9578083753585815} 02/26/2022 01:53:36 - INFO - codeparrot_training - Step 37317: {'lr': 8.129958454535824e-05, 'samples': 19106816, 'steps': 37317, 'loss/train': 1.3418536186218262} 02/26/2022 01:53:41 - INFO - codeparrot_training - Step 37318: {'lr': 8.12875094241548e-05, 'samples': 19107328, 'steps': 37318, 'loss/train': 1.7534079551696777} 02/26/2022 01:53:45 - INFO - codeparrot_training - Step 37319: {'lr': 8.12754350256621e-05, 'samples': 19107840, 'steps': 37319, 'loss/train': 2.507652759552002} 02/26/2022 01:53:51 - INFO - codeparrot_training - Step 37320: {'lr': 8.126336134993176e-05, 'samples': 19108352, 'steps': 37320, 'loss/train': 0.5858209133148193} 02/26/2022 01:53:54 - INFO - codeparrot_training - Step 37321: {'lr': 8.12512883970157e-05, 'samples': 19108864, 'steps': 37321, 'loss/train': 2.189741373062134} 02/26/2022 01:54:00 - INFO - codeparrot_training - Step 37322: {'lr': 8.123921616696553e-05, 'samples': 19109376, 'steps': 37322, 'loss/train': 2.0077922344207764} 02/26/2022 01:54:03 - INFO - codeparrot_training - Step 37323: {'lr': 8.122714465983297e-05, 'samples': 19109888, 'steps': 37323, 'loss/train': 1.1128926277160645} 02/26/2022 01:54:08 - INFO - codeparrot_training - Step 37324: {'lr': 8.121507387566962e-05, 'samples': 19110400, 'steps': 37324, 'loss/train': 0.3049682676792145} 02/26/2022 01:54:12 - INFO - codeparrot_training - Step 37325: {'lr': 8.120300381452742e-05, 'samples': 19110912, 'steps': 37325, 'loss/train': 2.042713165283203} 02/26/2022 01:54:18 - INFO - codeparrot_training - Step 37326: {'lr': 8.119093447645787e-05, 'samples': 19111424, 'steps': 37326, 'loss/train': 1.5282267332077026} 02/26/2022 01:54:21 - INFO - codeparrot_training - Step 37327: {'lr': 8.117886586151287e-05, 'samples': 19111936, 'steps': 37327, 'loss/train': 0.17124544084072113} 02/26/2022 01:54:27 - INFO - codeparrot_training - Step 37328: {'lr': 8.116679796974389e-05, 'samples': 19112448, 'steps': 37328, 'loss/train': 2.7843339443206787} 02/26/2022 01:54:31 - INFO - codeparrot_training - Step 37329: {'lr': 8.115473080120281e-05, 'samples': 19112960, 'steps': 37329, 'loss/train': 0.0338941290974617} 02/26/2022 01:54:36 - INFO - codeparrot_training - Step 37330: {'lr': 8.114266435594117e-05, 'samples': 19113472, 'steps': 37330, 'loss/train': 1.403425931930542} 02/26/2022 01:54:39 - INFO - codeparrot_training - Step 37331: {'lr': 8.113059863401087e-05, 'samples': 19113984, 'steps': 37331, 'loss/train': 1.2984315156936646} 02/26/2022 01:54:45 - INFO - codeparrot_training - Step 37332: {'lr': 8.111853363546331e-05, 'samples': 19114496, 'steps': 37332, 'loss/train': 1.5057514905929565} 02/26/2022 01:54:48 - INFO - codeparrot_training - Step 37333: {'lr': 8.110646936035044e-05, 'samples': 19115008, 'steps': 37333, 'loss/train': 1.6936734914779663} 02/26/2022 01:54:54 - INFO - codeparrot_training - Step 37334: {'lr': 8.109440580872371e-05, 'samples': 19115520, 'steps': 37334, 'loss/train': 1.7782814502716064} 02/26/2022 01:54:57 - INFO - codeparrot_training - Step 37335: {'lr': 8.108234298063502e-05, 'samples': 19116032, 'steps': 37335, 'loss/train': 2.1618940830230713} 02/26/2022 01:55:03 - INFO - codeparrot_training - Step 37336: {'lr': 8.107028087613592e-05, 'samples': 19116544, 'steps': 37336, 'loss/train': 1.412340760231018} 02/26/2022 01:55:06 - INFO - codeparrot_training - Step 37337: {'lr': 8.105821949527808e-05, 'samples': 19117056, 'steps': 37337, 'loss/train': 0.7058124542236328} 02/26/2022 01:55:12 - INFO - codeparrot_training - Step 37338: {'lr': 8.104615883811311e-05, 'samples': 19117568, 'steps': 37338, 'loss/train': 0.9833065867424011} 02/26/2022 01:55:15 - INFO - codeparrot_training - Step 37339: {'lr': 8.103409890469285e-05, 'samples': 19118080, 'steps': 37339, 'loss/train': 1.828080654144287} 02/26/2022 01:55:21 - INFO - codeparrot_training - Step 37340: {'lr': 8.102203969506886e-05, 'samples': 19118592, 'steps': 37340, 'loss/train': 1.2084861993789673} 02/26/2022 01:55:24 - INFO - codeparrot_training - Step 37341: {'lr': 8.100998120929275e-05, 'samples': 19119104, 'steps': 37341, 'loss/train': 2.0466437339782715} 02/26/2022 01:55:31 - INFO - codeparrot_training - Step 37342: {'lr': 8.099792344741619e-05, 'samples': 19119616, 'steps': 37342, 'loss/train': 2.5775063037872314} 02/26/2022 01:55:34 - INFO - codeparrot_training - Step 37343: {'lr': 8.09858664094909e-05, 'samples': 19120128, 'steps': 37343, 'loss/train': 1.6297249794006348} 02/26/2022 01:55:40 - INFO - codeparrot_training - Step 37344: {'lr': 8.09738100955685e-05, 'samples': 19120640, 'steps': 37344, 'loss/train': 1.064758539199829} 02/26/2022 01:55:43 - INFO - codeparrot_training - Step 37345: {'lr': 8.096175450570065e-05, 'samples': 19121152, 'steps': 37345, 'loss/train': 1.4299209117889404} 02/26/2022 01:55:49 - INFO - codeparrot_training - Step 37346: {'lr': 8.094969963993887e-05, 'samples': 19121664, 'steps': 37346, 'loss/train': 2.241816282272339} 02/26/2022 01:55:53 - INFO - codeparrot_training - Step 37347: {'lr': 8.093764549833499e-05, 'samples': 19122176, 'steps': 37347, 'loss/train': 1.838773488998413} 02/26/2022 01:55:58 - INFO - codeparrot_training - Step 37348: {'lr': 8.092559208094055e-05, 'samples': 19122688, 'steps': 37348, 'loss/train': 2.0230047702789307} 02/26/2022 01:56:02 - INFO - codeparrot_training - Step 37349: {'lr': 8.091353938780718e-05, 'samples': 19123200, 'steps': 37349, 'loss/train': 1.7273041009902954} 02/26/2022 01:56:07 - INFO - codeparrot_training - Step 37350: {'lr': 8.090148741898651e-05, 'samples': 19123712, 'steps': 37350, 'loss/train': 1.973228931427002} 02/26/2022 01:56:11 - INFO - codeparrot_training - Step 37351: {'lr': 8.088943617453009e-05, 'samples': 19124224, 'steps': 37351, 'loss/train': 1.6383157968521118} 02/26/2022 01:56:18 - INFO - codeparrot_training - Step 37352: {'lr': 8.087738565448974e-05, 'samples': 19124736, 'steps': 37352, 'loss/train': 2.2831029891967773} 02/26/2022 01:56:21 - INFO - codeparrot_training - Step 37353: {'lr': 8.086533585891692e-05, 'samples': 19125248, 'steps': 37353, 'loss/train': 0.05401711165904999} 02/26/2022 01:56:27 - INFO - codeparrot_training - Step 37354: {'lr': 8.085328678786333e-05, 'samples': 19125760, 'steps': 37354, 'loss/train': 2.198316812515259} 02/26/2022 01:56:30 - INFO - codeparrot_training - Step 37355: {'lr': 8.084123844138047e-05, 'samples': 19126272, 'steps': 37355, 'loss/train': 0.8088244795799255} 02/26/2022 01:56:36 - INFO - codeparrot_training - Step 37356: {'lr': 8.082919081952009e-05, 'samples': 19126784, 'steps': 37356, 'loss/train': 1.7159335613250732} 02/26/2022 01:56:39 - INFO - codeparrot_training - Step 37357: {'lr': 8.081714392233375e-05, 'samples': 19127296, 'steps': 37357, 'loss/train': 1.7257295846939087} 02/26/2022 01:56:45 - INFO - codeparrot_training - Step 37358: {'lr': 8.080509774987305e-05, 'samples': 19127808, 'steps': 37358, 'loss/train': 1.7765673398971558} 02/26/2022 01:56:48 - INFO - codeparrot_training - Step 37359: {'lr': 8.079305230218947e-05, 'samples': 19128320, 'steps': 37359, 'loss/train': 1.519286870956421} 02/26/2022 01:56:53 - INFO - codeparrot_training - Step 37360: {'lr': 8.078100757933485e-05, 'samples': 19128832, 'steps': 37360, 'loss/train': 1.1347962617874146} 02/26/2022 01:56:57 - INFO - codeparrot_training - Step 37361: {'lr': 8.076896358136055e-05, 'samples': 19129344, 'steps': 37361, 'loss/train': 1.7465343475341797} 02/26/2022 01:57:03 - INFO - codeparrot_training - Step 37362: {'lr': 8.075692030831847e-05, 'samples': 19129856, 'steps': 37362, 'loss/train': 1.900414228439331} 02/26/2022 01:57:07 - INFO - codeparrot_training - Step 37363: {'lr': 8.074487776025979e-05, 'samples': 19130368, 'steps': 37363, 'loss/train': 1.4611653089523315} 02/26/2022 01:57:12 - INFO - codeparrot_training - Step 37364: {'lr': 8.073283593723644e-05, 'samples': 19130880, 'steps': 37364, 'loss/train': 2.3062353134155273} 02/26/2022 01:57:15 - INFO - codeparrot_training - Step 37365: {'lr': 8.072079483929975e-05, 'samples': 19131392, 'steps': 37365, 'loss/train': 1.746525526046753} 02/26/2022 01:57:21 - INFO - codeparrot_training - Step 37366: {'lr': 8.070875446650162e-05, 'samples': 19131904, 'steps': 37366, 'loss/train': 2.2053158283233643} 02/26/2022 01:57:24 - INFO - codeparrot_training - Step 37367: {'lr': 8.069671481889323e-05, 'samples': 19132416, 'steps': 37367, 'loss/train': 1.556070327758789} 02/26/2022 01:57:30 - INFO - codeparrot_training - Step 37368: {'lr': 8.068467589652648e-05, 'samples': 19132928, 'steps': 37368, 'loss/train': 2.6845715045928955} 02/26/2022 01:57:33 - INFO - codeparrot_training - Step 37369: {'lr': 8.067263769945271e-05, 'samples': 19133440, 'steps': 37369, 'loss/train': 1.8082600831985474} 02/26/2022 01:57:39 - INFO - codeparrot_training - Step 37370: {'lr': 8.066060022772376e-05, 'samples': 19133952, 'steps': 37370, 'loss/train': 1.0379939079284668} 02/26/2022 01:57:42 - INFO - codeparrot_training - Step 37371: {'lr': 8.064856348139085e-05, 'samples': 19134464, 'steps': 37371, 'loss/train': 1.405431866645813} 02/26/2022 01:57:49 - INFO - codeparrot_training - Step 37372: {'lr': 8.063652746050582e-05, 'samples': 19134976, 'steps': 37372, 'loss/train': 2.5331814289093018} 02/26/2022 01:57:52 - INFO - codeparrot_training - Step 37373: {'lr': 8.062449216512003e-05, 'samples': 19135488, 'steps': 37373, 'loss/train': 2.1231932640075684} 02/26/2022 01:57:59 - INFO - codeparrot_training - Step 37374: {'lr': 8.061245759528521e-05, 'samples': 19136000, 'steps': 37374, 'loss/train': 2.263923168182373} 02/26/2022 01:58:02 - INFO - codeparrot_training - Step 37375: {'lr': 8.060042375105284e-05, 'samples': 19136512, 'steps': 37375, 'loss/train': 2.292685031890869} 02/26/2022 01:58:06 - INFO - codeparrot_training - Step 37376: {'lr': 8.058839063247447e-05, 'samples': 19137024, 'steps': 37376, 'loss/train': 2.2224104404449463} 02/26/2022 01:58:11 - INFO - codeparrot_training - Step 37377: {'lr': 8.057635823960155e-05, 'samples': 19137536, 'steps': 37377, 'loss/train': 1.2549443244934082} 02/26/2022 01:58:15 - INFO - codeparrot_training - Step 37378: {'lr': 8.05643265724858e-05, 'samples': 19138048, 'steps': 37378, 'loss/train': 1.2091478109359741} 02/26/2022 01:58:20 - INFO - codeparrot_training - Step 37379: {'lr': 8.055229563117869e-05, 'samples': 19138560, 'steps': 37379, 'loss/train': 1.694042444229126} 02/26/2022 01:58:24 - INFO - codeparrot_training - Step 37380: {'lr': 8.054026541573172e-05, 'samples': 19139072, 'steps': 37380, 'loss/train': 1.5892401933670044} 02/26/2022 01:58:29 - INFO - codeparrot_training - Step 37381: {'lr': 8.052823592619637e-05, 'samples': 19139584, 'steps': 37381, 'loss/train': 2.951171398162842} 02/26/2022 01:58:33 - INFO - codeparrot_training - Step 37382: {'lr': 8.051620716262433e-05, 'samples': 19140096, 'steps': 37382, 'loss/train': 1.8875781297683716} 02/26/2022 01:58:38 - INFO - codeparrot_training - Step 37383: {'lr': 8.050417912506702e-05, 'samples': 19140608, 'steps': 37383, 'loss/train': 1.9337056875228882} 02/26/2022 01:58:42 - INFO - codeparrot_training - Step 37384: {'lr': 8.049215181357602e-05, 'samples': 19141120, 'steps': 37384, 'loss/train': 2.507456064224243} 02/26/2022 01:58:47 - INFO - codeparrot_training - Step 37385: {'lr': 8.048012522820273e-05, 'samples': 19141632, 'steps': 37385, 'loss/train': 0.5179448127746582} 02/26/2022 01:58:51 - INFO - codeparrot_training - Step 37386: {'lr': 8.046809936899879e-05, 'samples': 19142144, 'steps': 37386, 'loss/train': 1.0809290409088135} 02/26/2022 01:58:56 - INFO - codeparrot_training - Step 37387: {'lr': 8.045607423601572e-05, 'samples': 19142656, 'steps': 37387, 'loss/train': 1.7661043405532837} 02/26/2022 01:59:00 - INFO - codeparrot_training - Step 37388: {'lr': 8.044404982930498e-05, 'samples': 19143168, 'steps': 37388, 'loss/train': 0.7763864994049072} 02/26/2022 01:59:05 - INFO - codeparrot_training - Step 37389: {'lr': 8.043202614891801e-05, 'samples': 19143680, 'steps': 37389, 'loss/train': 0.527213454246521} 02/26/2022 01:59:09 - INFO - codeparrot_training - Step 37390: {'lr': 8.04200031949065e-05, 'samples': 19144192, 'steps': 37390, 'loss/train': 1.1823621988296509} 02/26/2022 01:59:15 - INFO - codeparrot_training - Step 37391: {'lr': 8.040798096732182e-05, 'samples': 19144704, 'steps': 37391, 'loss/train': 1.87386155128479} 02/26/2022 01:59:19 - INFO - codeparrot_training - Step 37392: {'lr': 8.039595946621551e-05, 'samples': 19145216, 'steps': 37392, 'loss/train': 0.8751384019851685} 02/26/2022 01:59:24 - INFO - codeparrot_training - Step 37393: {'lr': 8.038393869163896e-05, 'samples': 19145728, 'steps': 37393, 'loss/train': 1.0981802940368652} 02/26/2022 01:59:27 - INFO - codeparrot_training - Step 37394: {'lr': 8.037191864364385e-05, 'samples': 19146240, 'steps': 37394, 'loss/train': 1.4065470695495605} 02/26/2022 01:59:33 - INFO - codeparrot_training - Step 37395: {'lr': 8.035989932228158e-05, 'samples': 19146752, 'steps': 37395, 'loss/train': 1.101318597793579} 02/26/2022 01:59:37 - INFO - codeparrot_training - Step 37396: {'lr': 8.034788072760354e-05, 'samples': 19147264, 'steps': 37396, 'loss/train': 0.996615469455719} 02/26/2022 01:59:42 - INFO - codeparrot_training - Step 37397: {'lr': 8.033586285966148e-05, 'samples': 19147776, 'steps': 37397, 'loss/train': 0.8456780910491943} 02/26/2022 01:59:45 - INFO - codeparrot_training - Step 37398: {'lr': 8.032384571850656e-05, 'samples': 19148288, 'steps': 37398, 'loss/train': 1.9835975170135498} 02/26/2022 01:59:52 - INFO - codeparrot_training - Step 37399: {'lr': 8.031182930419048e-05, 'samples': 19148800, 'steps': 37399, 'loss/train': 2.304450035095215} 02/26/2022 01:59:55 - INFO - codeparrot_training - Step 37400: {'lr': 8.029981361676455e-05, 'samples': 19149312, 'steps': 37400, 'loss/train': 1.8341145515441895} 02/26/2022 02:00:01 - INFO - codeparrot_training - Step 37401: {'lr': 8.02877986562805e-05, 'samples': 19149824, 'steps': 37401, 'loss/train': 1.2928545475006104} 02/26/2022 02:00:04 - INFO - codeparrot_training - Step 37402: {'lr': 8.027578442278946e-05, 'samples': 19150336, 'steps': 37402, 'loss/train': 1.7776765823364258} 02/26/2022 02:00:10 - INFO - codeparrot_training - Step 37403: {'lr': 8.026377091634316e-05, 'samples': 19150848, 'steps': 37403, 'loss/train': 1.263899803161621} 02/26/2022 02:00:13 - INFO - codeparrot_training - Step 37404: {'lr': 8.025175813699287e-05, 'samples': 19151360, 'steps': 37404, 'loss/train': 1.4260401725769043} 02/26/2022 02:00:19 - INFO - codeparrot_training - Step 37405: {'lr': 8.023974608479032e-05, 'samples': 19151872, 'steps': 37405, 'loss/train': 1.4067226648330688} 02/26/2022 02:00:22 - INFO - codeparrot_training - Step 37406: {'lr': 8.022773475978665e-05, 'samples': 19152384, 'steps': 37406, 'loss/train': 2.385246992111206} 02/26/2022 02:00:28 - INFO - codeparrot_training - Step 37407: {'lr': 8.021572416203351e-05, 'samples': 19152896, 'steps': 37407, 'loss/train': 3.092381715774536} 02/26/2022 02:00:31 - INFO - codeparrot_training - Step 37408: {'lr': 8.020371429158221e-05, 'samples': 19153408, 'steps': 37408, 'loss/train': 2.0908305644989014} 02/26/2022 02:00:37 - INFO - codeparrot_training - Step 37409: {'lr': 8.01917051484844e-05, 'samples': 19153920, 'steps': 37409, 'loss/train': 1.589678168296814} 02/26/2022 02:00:41 - INFO - codeparrot_training - Step 37410: {'lr': 8.017969673279138e-05, 'samples': 19154432, 'steps': 37410, 'loss/train': 1.7165073156356812} 02/26/2022 02:00:46 - INFO - codeparrot_training - Step 37411: {'lr': 8.01676890445546e-05, 'samples': 19154944, 'steps': 37411, 'loss/train': 2.5316507816314697} 02/26/2022 02:00:50 - INFO - codeparrot_training - Step 37412: {'lr': 8.015568208382545e-05, 'samples': 19155456, 'steps': 37412, 'loss/train': 1.8428373336791992} 02/26/2022 02:00:55 - INFO - codeparrot_training - Step 37413: {'lr': 8.014367585065549e-05, 'samples': 19155968, 'steps': 37413, 'loss/train': 2.3472867012023926} 02/26/2022 02:00:59 - INFO - codeparrot_training - Step 37414: {'lr': 8.01316703450961e-05, 'samples': 19156480, 'steps': 37414, 'loss/train': 1.9508153200149536} 02/26/2022 02:01:04 - INFO - codeparrot_training - Step 37415: {'lr': 8.011966556719871e-05, 'samples': 19156992, 'steps': 37415, 'loss/train': 1.6075329780578613} 02/26/2022 02:01:08 - INFO - codeparrot_training - Step 37416: {'lr': 8.010766151701463e-05, 'samples': 19157504, 'steps': 37416, 'loss/train': 2.5764198303222656} 02/26/2022 02:01:13 - INFO - codeparrot_training - Step 37417: {'lr': 8.009565819459546e-05, 'samples': 19158016, 'steps': 37417, 'loss/train': 2.0447890758514404} 02/26/2022 02:01:17 - INFO - codeparrot_training - Step 37418: {'lr': 8.008365559999256e-05, 'samples': 19158528, 'steps': 37418, 'loss/train': 1.3697654008865356} 02/26/2022 02:01:22 - INFO - codeparrot_training - Step 37419: {'lr': 8.00716537332573e-05, 'samples': 19159040, 'steps': 37419, 'loss/train': 1.7474851608276367} 02/26/2022 02:01:26 - INFO - codeparrot_training - Step 37420: {'lr': 8.005965259444107e-05, 'samples': 19159552, 'steps': 37420, 'loss/train': 2.0200257301330566} 02/26/2022 02:01:32 - INFO - codeparrot_training - Step 37421: {'lr': 8.004765218359537e-05, 'samples': 19160064, 'steps': 37421, 'loss/train': 1.8225852251052856} 02/26/2022 02:01:36 - INFO - codeparrot_training - Step 37422: {'lr': 8.003565250077149e-05, 'samples': 19160576, 'steps': 37422, 'loss/train': 1.8792427778244019} 02/26/2022 02:01:41 - INFO - codeparrot_training - Step 37423: {'lr': 8.00236535460211e-05, 'samples': 19161088, 'steps': 37423, 'loss/train': 1.1113975048065186} 02/26/2022 02:01:44 - INFO - codeparrot_training - Step 37424: {'lr': 8.001165531939519e-05, 'samples': 19161600, 'steps': 37424, 'loss/train': 1.3480619192123413} 02/26/2022 02:01:50 - INFO - codeparrot_training - Step 37425: {'lr': 7.999965782094548e-05, 'samples': 19162112, 'steps': 37425, 'loss/train': 2.600320339202881} 02/26/2022 02:01:53 - INFO - codeparrot_training - Step 37426: {'lr': 7.998766105072316e-05, 'samples': 19162624, 'steps': 37426, 'loss/train': 2.0850894451141357} 02/26/2022 02:01:59 - INFO - codeparrot_training - Step 37427: {'lr': 7.99756650087799e-05, 'samples': 19163136, 'steps': 37427, 'loss/train': 2.140472650527954} 02/26/2022 02:02:02 - INFO - codeparrot_training - Step 37428: {'lr': 7.996366969516672e-05, 'samples': 19163648, 'steps': 37428, 'loss/train': 1.5940027236938477} 02/26/2022 02:02:08 - INFO - codeparrot_training - Step 37429: {'lr': 7.995167510993529e-05, 'samples': 19164160, 'steps': 37429, 'loss/train': 1.5432043075561523} 02/26/2022 02:02:11 - INFO - codeparrot_training - Step 37430: {'lr': 7.993968125313686e-05, 'samples': 19164672, 'steps': 37430, 'loss/train': 1.2817788124084473} 02/26/2022 02:02:18 - INFO - codeparrot_training - Step 37431: {'lr': 7.992768812482284e-05, 'samples': 19165184, 'steps': 37431, 'loss/train': 1.318253755569458} 02/26/2022 02:02:21 - INFO - codeparrot_training - Step 37432: {'lr': 7.991569572504453e-05, 'samples': 19165696, 'steps': 37432, 'loss/train': 1.844303011894226} 02/26/2022 02:02:27 - INFO - codeparrot_training - Step 37433: {'lr': 7.990370405385348e-05, 'samples': 19166208, 'steps': 37433, 'loss/train': 1.5911951065063477} 02/26/2022 02:02:30 - INFO - codeparrot_training - Step 37434: {'lr': 7.989171311130092e-05, 'samples': 19166720, 'steps': 37434, 'loss/train': 1.4388668537139893} 02/26/2022 02:02:36 - INFO - codeparrot_training - Step 37435: {'lr': 7.987972289743814e-05, 'samples': 19167232, 'steps': 37435, 'loss/train': 0.15204450488090515} 02/26/2022 02:02:39 - INFO - codeparrot_training - Step 37436: {'lr': 7.986773341231673e-05, 'samples': 19167744, 'steps': 37436, 'loss/train': 1.7728265523910522} 02/26/2022 02:02:45 - INFO - codeparrot_training - Step 37437: {'lr': 7.985574465598791e-05, 'samples': 19168256, 'steps': 37437, 'loss/train': 1.8559874296188354} 02/26/2022 02:02:48 - INFO - codeparrot_training - Step 37438: {'lr': 7.984375662850304e-05, 'samples': 19168768, 'steps': 37438, 'loss/train': 1.6557707786560059} 02/26/2022 02:02:54 - INFO - codeparrot_training - Step 37439: {'lr': 7.983176932991343e-05, 'samples': 19169280, 'steps': 37439, 'loss/train': 2.3579235076904297} 02/26/2022 02:02:57 - INFO - codeparrot_training - Step 37440: {'lr': 7.981978276027054e-05, 'samples': 19169792, 'steps': 37440, 'loss/train': 2.2395734786987305} 02/26/2022 02:03:03 - INFO - codeparrot_training - Step 37441: {'lr': 7.98077969196257e-05, 'samples': 19170304, 'steps': 37441, 'loss/train': 2.1372907161712646} 02/26/2022 02:03:06 - INFO - codeparrot_training - Step 37442: {'lr': 7.97958118080302e-05, 'samples': 19170816, 'steps': 37442, 'loss/train': 1.090391993522644} 02/26/2022 02:03:12 - INFO - codeparrot_training - Step 37443: {'lr': 7.97838274255353e-05, 'samples': 19171328, 'steps': 37443, 'loss/train': 1.7838338613510132} 02/26/2022 02:03:15 - INFO - codeparrot_training - Step 37444: {'lr': 7.977184377219252e-05, 'samples': 19171840, 'steps': 37444, 'loss/train': 2.1696956157684326} 02/26/2022 02:03:20 - INFO - codeparrot_training - Step 37445: {'lr': 7.975986084805311e-05, 'samples': 19172352, 'steps': 37445, 'loss/train': 3.2940948009490967} 02/26/2022 02:03:24 - INFO - codeparrot_training - Step 37446: {'lr': 7.97478786531684e-05, 'samples': 19172864, 'steps': 37446, 'loss/train': 1.753911018371582} 02/26/2022 02:03:30 - INFO - codeparrot_training - Step 37447: {'lr': 7.973589718758964e-05, 'samples': 19173376, 'steps': 37447, 'loss/train': 2.3564867973327637} 02/26/2022 02:03:33 - INFO - codeparrot_training - Step 37448: {'lr': 7.972391645136831e-05, 'samples': 19173888, 'steps': 37448, 'loss/train': 1.500327467918396} 02/26/2022 02:03:39 - INFO - codeparrot_training - Step 37449: {'lr': 7.971193644455566e-05, 'samples': 19174400, 'steps': 37449, 'loss/train': 1.5810664892196655} 02/26/2022 02:03:42 - INFO - codeparrot_training - Step 37450: {'lr': 7.9699957167203e-05, 'samples': 19174912, 'steps': 37450, 'loss/train': 1.7077873945236206} 02/26/2022 02:03:48 - INFO - codeparrot_training - Step 37451: {'lr': 7.968797861936153e-05, 'samples': 19175424, 'steps': 37451, 'loss/train': 1.5553562641143799} 02/26/2022 02:03:51 - INFO - codeparrot_training - Step 37452: {'lr': 7.96760008010828e-05, 'samples': 19175936, 'steps': 37452, 'loss/train': 1.8128317594528198} 02/26/2022 02:03:57 - INFO - codeparrot_training - Step 37453: {'lr': 7.966402371241799e-05, 'samples': 19176448, 'steps': 37453, 'loss/train': 2.0748281478881836} 02/26/2022 02:04:00 - INFO - codeparrot_training - Step 37454: {'lr': 7.965204735341842e-05, 'samples': 19176960, 'steps': 37454, 'loss/train': 1.9976259469985962} 02/26/2022 02:04:06 - INFO - codeparrot_training - Step 37455: {'lr': 7.964007172413529e-05, 'samples': 19177472, 'steps': 37455, 'loss/train': 1.3991798162460327} 02/26/2022 02:04:09 - INFO - codeparrot_training - Step 37456: {'lr': 7.962809682462008e-05, 'samples': 19177984, 'steps': 37456, 'loss/train': 2.845388889312744} 02/26/2022 02:04:16 - INFO - codeparrot_training - Step 37457: {'lr': 7.961612265492393e-05, 'samples': 19178496, 'steps': 37457, 'loss/train': 1.387342095375061} 02/26/2022 02:04:19 - INFO - codeparrot_training - Step 37458: {'lr': 7.960414921509837e-05, 'samples': 19179008, 'steps': 37458, 'loss/train': 2.001033067703247} 02/26/2022 02:04:25 - INFO - codeparrot_training - Step 37459: {'lr': 7.959217650519438e-05, 'samples': 19179520, 'steps': 37459, 'loss/train': 2.0248870849609375} 02/26/2022 02:04:28 - INFO - codeparrot_training - Step 37460: {'lr': 7.958020452526346e-05, 'samples': 19180032, 'steps': 37460, 'loss/train': 2.51249098777771} 02/26/2022 02:04:34 - INFO - codeparrot_training - Step 37461: {'lr': 7.956823327535675e-05, 'samples': 19180544, 'steps': 37461, 'loss/train': 2.244447946548462} 02/26/2022 02:04:37 - INFO - codeparrot_training - Step 37462: {'lr': 7.955626275552577e-05, 'samples': 19181056, 'steps': 37462, 'loss/train': 2.141691207885742} 02/26/2022 02:04:43 - INFO - codeparrot_training - Step 37463: {'lr': 7.954429296582149e-05, 'samples': 19181568, 'steps': 37463, 'loss/train': 1.6009907722473145} 02/26/2022 02:04:46 - INFO - codeparrot_training - Step 37464: {'lr': 7.953232390629539e-05, 'samples': 19182080, 'steps': 37464, 'loss/train': 1.098544716835022} 02/26/2022 02:04:52 - INFO - codeparrot_training - Step 37465: {'lr': 7.952035557699863e-05, 'samples': 19182592, 'steps': 37465, 'loss/train': 1.5164607763290405} 02/26/2022 02:04:55 - INFO - codeparrot_training - Step 37466: {'lr': 7.950838797798268e-05, 'samples': 19183104, 'steps': 37466, 'loss/train': 1.4035630226135254} 02/26/2022 02:05:02 - INFO - codeparrot_training - Step 37467: {'lr': 7.94964211092985e-05, 'samples': 19183616, 'steps': 37467, 'loss/train': 1.9000409841537476} 02/26/2022 02:05:05 - INFO - codeparrot_training - Step 37468: {'lr': 7.94844549709976e-05, 'samples': 19184128, 'steps': 37468, 'loss/train': 0.5668824315071106} 02/26/2022 02:05:11 - INFO - codeparrot_training - Step 37469: {'lr': 7.947248956313106e-05, 'samples': 19184640, 'steps': 37469, 'loss/train': 1.2484667301177979} 02/26/2022 02:05:14 - INFO - codeparrot_training - Step 37470: {'lr': 7.946052488575032e-05, 'samples': 19185152, 'steps': 37470, 'loss/train': 1.3246175050735474} 02/26/2022 02:05:20 - INFO - codeparrot_training - Step 37471: {'lr': 7.94485609389065e-05, 'samples': 19185664, 'steps': 37471, 'loss/train': 0.7225742936134338} 02/26/2022 02:05:24 - INFO - codeparrot_training - Step 37472: {'lr': 7.94365977226509e-05, 'samples': 19186176, 'steps': 37472, 'loss/train': 0.3974103331565857} 02/26/2022 02:05:29 - INFO - codeparrot_training - Step 37473: {'lr': 7.942463523703472e-05, 'samples': 19186688, 'steps': 37473, 'loss/train': 1.729839563369751} 02/26/2022 02:05:33 - INFO - codeparrot_training - Step 37474: {'lr': 7.941267348210929e-05, 'samples': 19187200, 'steps': 37474, 'loss/train': 1.6778931617736816} 02/26/2022 02:05:38 - INFO - codeparrot_training - Step 37475: {'lr': 7.940071245792579e-05, 'samples': 19187712, 'steps': 37475, 'loss/train': 1.685838222503662} 02/26/2022 02:05:44 - INFO - codeparrot_training - Step 37476: {'lr': 7.938875216453548e-05, 'samples': 19188224, 'steps': 37476, 'loss/train': 1.5195034742355347} 02/26/2022 02:05:47 - INFO - codeparrot_training - Step 37477: {'lr': 7.937679260198949e-05, 'samples': 19188736, 'steps': 37477, 'loss/train': 1.64725923538208} 02/26/2022 02:05:51 - INFO - codeparrot_training - Step 37478: {'lr': 7.936483377033924e-05, 'samples': 19189248, 'steps': 37478, 'loss/train': 1.3167613744735718} 02/26/2022 02:05:56 - INFO - codeparrot_training - Step 37479: {'lr': 7.935287566963584e-05, 'samples': 19189760, 'steps': 37479, 'loss/train': 1.630751132965088} 02/26/2022 02:06:02 - INFO - codeparrot_training - Step 37480: {'lr': 7.934091829993053e-05, 'samples': 19190272, 'steps': 37480, 'loss/train': 1.701886534690857} 02/26/2022 02:06:06 - INFO - codeparrot_training - Step 37481: {'lr': 7.932896166127456e-05, 'samples': 19190784, 'steps': 37481, 'loss/train': 1.539083480834961} 02/26/2022 02:06:09 - INFO - codeparrot_training - Step 37482: {'lr': 7.931700575371903e-05, 'samples': 19191296, 'steps': 37482, 'loss/train': 1.2061444520950317} 02/26/2022 02:06:15 - INFO - codeparrot_training - Step 37483: {'lr': 7.930505057731533e-05, 'samples': 19191808, 'steps': 37483, 'loss/train': 1.1710108518600464} 02/26/2022 02:06:20 - INFO - codeparrot_training - Step 37484: {'lr': 7.929309613211457e-05, 'samples': 19192320, 'steps': 37484, 'loss/train': 2.493664503097534} 02/26/2022 02:06:24 - INFO - codeparrot_training - Step 37485: {'lr': 7.928114241816803e-05, 'samples': 19192832, 'steps': 37485, 'loss/train': 1.649251103401184} 02/26/2022 02:06:29 - INFO - codeparrot_training - Step 37486: {'lr': 7.926918943552674e-05, 'samples': 19193344, 'steps': 37486, 'loss/train': 2.026777744293213} 02/26/2022 02:06:33 - INFO - codeparrot_training - Step 37487: {'lr': 7.925723718424213e-05, 'samples': 19193856, 'steps': 37487, 'loss/train': 1.492714762687683} 02/26/2022 02:06:38 - INFO - codeparrot_training - Step 37488: {'lr': 7.924528566436531e-05, 'samples': 19194368, 'steps': 37488, 'loss/train': 1.8701484203338623} 02/26/2022 02:06:42 - INFO - codeparrot_training - Step 37489: {'lr': 7.923333487594744e-05, 'samples': 19194880, 'steps': 37489, 'loss/train': 1.11849844455719} 02/26/2022 02:06:47 - INFO - codeparrot_training - Step 37490: {'lr': 7.922138481903968e-05, 'samples': 19195392, 'steps': 37490, 'loss/train': 2.541264057159424} 02/26/2022 02:06:51 - INFO - codeparrot_training - Step 37491: {'lr': 7.920943549369336e-05, 'samples': 19195904, 'steps': 37491, 'loss/train': 1.1719616651535034} 02/26/2022 02:06:57 - INFO - codeparrot_training - Step 37492: {'lr': 7.919748689995949e-05, 'samples': 19196416, 'steps': 37492, 'loss/train': 1.9301249980926514} 02/26/2022 02:07:00 - INFO - codeparrot_training - Step 37493: {'lr': 7.918553903788953e-05, 'samples': 19196928, 'steps': 37493, 'loss/train': 2.1314516067504883} 02/26/2022 02:07:06 - INFO - codeparrot_training - Step 37494: {'lr': 7.917359190753428e-05, 'samples': 19197440, 'steps': 37494, 'loss/train': 2.2338268756866455} 02/26/2022 02:07:09 - INFO - codeparrot_training - Step 37495: {'lr': 7.916164550894526e-05, 'samples': 19197952, 'steps': 37495, 'loss/train': 1.6606563329696655} 02/26/2022 02:07:15 - INFO - codeparrot_training - Step 37496: {'lr': 7.914969984217337e-05, 'samples': 19198464, 'steps': 37496, 'loss/train': 0.9926959276199341} 02/26/2022 02:07:18 - INFO - codeparrot_training - Step 37497: {'lr': 7.91377549072701e-05, 'samples': 19198976, 'steps': 37497, 'loss/train': 1.6473690271377563} 02/26/2022 02:07:24 - INFO - codeparrot_training - Step 37498: {'lr': 7.912581070428627e-05, 'samples': 19199488, 'steps': 37498, 'loss/train': 0.8211421370506287} 02/26/2022 02:07:27 - INFO - codeparrot_training - Step 37499: {'lr': 7.91138672332733e-05, 'samples': 19200000, 'steps': 37499, 'loss/train': 1.2416177988052368} 02/26/2022 02:07:33 - INFO - codeparrot_training - Step 37500: {'lr': 7.910192449428217e-05, 'samples': 19200512, 'steps': 37500, 'loss/train': 1.3083006143569946} 02/26/2022 02:07:36 - INFO - codeparrot_training - Step 37501: {'lr': 7.908998248736427e-05, 'samples': 19201024, 'steps': 37501, 'loss/train': 2.146533489227295} 02/26/2022 02:07:43 - INFO - codeparrot_training - Step 37502: {'lr': 7.907804121257048e-05, 'samples': 19201536, 'steps': 37502, 'loss/train': 1.65902841091156} 02/26/2022 02:07:46 - INFO - codeparrot_training - Step 37503: {'lr': 7.906610066995216e-05, 'samples': 19202048, 'steps': 37503, 'loss/train': 1.2601209878921509} 02/26/2022 02:07:52 - INFO - codeparrot_training - Step 37504: {'lr': 7.90541608595603e-05, 'samples': 19202560, 'steps': 37504, 'loss/train': 0.16535258293151855} 02/26/2022 02:07:55 - INFO - codeparrot_training - Step 37505: {'lr': 7.904222178144621e-05, 'samples': 19203072, 'steps': 37505, 'loss/train': 1.8236409425735474} 02/26/2022 02:08:01 - INFO - codeparrot_training - Step 37506: {'lr': 7.903028343566096e-05, 'samples': 19203584, 'steps': 37506, 'loss/train': 1.45333731174469} 02/26/2022 02:08:04 - INFO - codeparrot_training - Step 37507: {'lr': 7.90183458222557e-05, 'samples': 19204096, 'steps': 37507, 'loss/train': 1.0002244710922241} 02/26/2022 02:08:10 - INFO - codeparrot_training - Step 37508: {'lr': 7.900640894128147e-05, 'samples': 19204608, 'steps': 37508, 'loss/train': 1.2659326791763306} 02/26/2022 02:08:13 - INFO - codeparrot_training - Step 37509: {'lr': 7.899447279278956e-05, 'samples': 19205120, 'steps': 37509, 'loss/train': 0.9429946541786194} 02/26/2022 02:08:19 - INFO - codeparrot_training - Step 37510: {'lr': 7.898253737683103e-05, 'samples': 19205632, 'steps': 37510, 'loss/train': 1.514399766921997} 02/26/2022 02:08:22 - INFO - codeparrot_training - Step 37511: {'lr': 7.897060269345702e-05, 'samples': 19206144, 'steps': 37511, 'loss/train': 0.5376790761947632} 02/26/2022 02:08:29 - INFO - codeparrot_training - Step 37512: {'lr': 7.895866874271854e-05, 'samples': 19206656, 'steps': 37512, 'loss/train': 2.072449207305908} 02/26/2022 02:08:32 - INFO - codeparrot_training - Step 37513: {'lr': 7.894673552466689e-05, 'samples': 19207168, 'steps': 37513, 'loss/train': 1.615744709968567} 02/26/2022 02:08:38 - INFO - codeparrot_training - Step 37514: {'lr': 7.893480303935313e-05, 'samples': 19207680, 'steps': 37514, 'loss/train': 1.763232946395874} 02/26/2022 02:08:41 - INFO - codeparrot_training - Step 37515: {'lr': 7.892287128682834e-05, 'samples': 19208192, 'steps': 37515, 'loss/train': 1.7276393175125122} 02/26/2022 02:08:47 - INFO - codeparrot_training - Step 37516: {'lr': 7.891094026714357e-05, 'samples': 19208704, 'steps': 37516, 'loss/train': 0.6663673520088196} 02/26/2022 02:08:50 - INFO - codeparrot_training - Step 37517: {'lr': 7.889900998035007e-05, 'samples': 19209216, 'steps': 37517, 'loss/train': 2.6416876316070557} 02/26/2022 02:08:56 - INFO - codeparrot_training - Step 37518: {'lr': 7.888708042649892e-05, 'samples': 19209728, 'steps': 37518, 'loss/train': 1.9058363437652588} 02/26/2022 02:08:59 - INFO - codeparrot_training - Step 37519: {'lr': 7.887515160564116e-05, 'samples': 19210240, 'steps': 37519, 'loss/train': 1.670300006866455} 02/26/2022 02:09:05 - INFO - codeparrot_training - Step 37520: {'lr': 7.886322351782782e-05, 'samples': 19210752, 'steps': 37520, 'loss/train': 0.38533204793930054} 02/26/2022 02:09:08 - INFO - codeparrot_training - Step 37521: {'lr': 7.885129616311018e-05, 'samples': 19211264, 'steps': 37521, 'loss/train': 0.9540395736694336} 02/26/2022 02:09:14 - INFO - codeparrot_training - Step 37522: {'lr': 7.883936954153922e-05, 'samples': 19211776, 'steps': 37522, 'loss/train': 1.4791840314865112} 02/26/2022 02:09:17 - INFO - codeparrot_training - Step 37523: {'lr': 7.882744365316608e-05, 'samples': 19212288, 'steps': 37523, 'loss/train': 1.3955113887786865} 02/26/2022 02:09:23 - INFO - codeparrot_training - Step 37524: {'lr': 7.881551849804175e-05, 'samples': 19212800, 'steps': 37524, 'loss/train': 2.339679718017578} 02/26/2022 02:09:26 - INFO - codeparrot_training - Step 37525: {'lr': 7.880359407621743e-05, 'samples': 19213312, 'steps': 37525, 'loss/train': 1.4471436738967896} 02/26/2022 02:09:32 - INFO - codeparrot_training - Step 37526: {'lr': 7.879167038774415e-05, 'samples': 19213824, 'steps': 37526, 'loss/train': 0.8532547950744629} 02/26/2022 02:09:35 - INFO - codeparrot_training - Step 37527: {'lr': 7.877974743267302e-05, 'samples': 19214336, 'steps': 37527, 'loss/train': 1.6491354703903198} 02/26/2022 02:09:41 - INFO - codeparrot_training - Step 37528: {'lr': 7.876782521105505e-05, 'samples': 19214848, 'steps': 37528, 'loss/train': 1.108742594718933} 02/26/2022 02:09:45 - INFO - codeparrot_training - Step 37529: {'lr': 7.875590372294128e-05, 'samples': 19215360, 'steps': 37529, 'loss/train': 3.012669801712036} 02/26/2022 02:09:50 - INFO - codeparrot_training - Step 37530: {'lr': 7.874398296838292e-05, 'samples': 19215872, 'steps': 37530, 'loss/train': 1.7740246057510376} 02/26/2022 02:09:54 - INFO - codeparrot_training - Step 37531: {'lr': 7.873206294743088e-05, 'samples': 19216384, 'steps': 37531, 'loss/train': 1.08603036403656} 02/26/2022 02:10:00 - INFO - codeparrot_training - Step 37532: {'lr': 7.872014366013647e-05, 'samples': 19216896, 'steps': 37532, 'loss/train': 1.7007575035095215} 02/26/2022 02:10:03 - INFO - codeparrot_training - Step 37533: {'lr': 7.87082251065504e-05, 'samples': 19217408, 'steps': 37533, 'loss/train': 0.8413349390029907} 02/26/2022 02:10:09 - INFO - codeparrot_training - Step 37534: {'lr': 7.8696307286724e-05, 'samples': 19217920, 'steps': 37534, 'loss/train': 2.635847330093384} 02/26/2022 02:10:12 - INFO - codeparrot_training - Step 37535: {'lr': 7.868439020070811e-05, 'samples': 19218432, 'steps': 37535, 'loss/train': 2.3311822414398193} 02/26/2022 02:10:18 - INFO - codeparrot_training - Step 37536: {'lr': 7.867247384855409e-05, 'samples': 19218944, 'steps': 37536, 'loss/train': 1.5154963731765747} 02/26/2022 02:10:21 - INFO - codeparrot_training - Step 37537: {'lr': 7.866055823031262e-05, 'samples': 19219456, 'steps': 37537, 'loss/train': 1.4162359237670898} 02/26/2022 02:10:27 - INFO - codeparrot_training - Step 37538: {'lr': 7.864864334603497e-05, 'samples': 19219968, 'steps': 37538, 'loss/train': 0.8373293876647949} 02/26/2022 02:10:31 - INFO - codeparrot_training - Step 37539: {'lr': 7.863672919577208e-05, 'samples': 19220480, 'steps': 37539, 'loss/train': 2.602374792098999} 02/26/2022 02:10:36 - INFO - codeparrot_training - Step 37540: {'lr': 7.862481577957517e-05, 'samples': 19220992, 'steps': 37540, 'loss/train': 1.1344972848892212} 02/26/2022 02:10:40 - INFO - codeparrot_training - Step 37541: {'lr': 7.861290309749494e-05, 'samples': 19221504, 'steps': 37541, 'loss/train': 2.5414211750030518} 02/26/2022 02:10:45 - INFO - codeparrot_training - Step 37542: {'lr': 7.860099114958275e-05, 'samples': 19222016, 'steps': 37542, 'loss/train': 2.643895387649536} 02/26/2022 02:10:49 - INFO - codeparrot_training - Step 37543: {'lr': 7.858907993588937e-05, 'samples': 19222528, 'steps': 37543, 'loss/train': 2.0918405055999756} 02/26/2022 02:10:54 - INFO - codeparrot_training - Step 37544: {'lr': 7.857716945646603e-05, 'samples': 19223040, 'steps': 37544, 'loss/train': 2.726212501525879} 02/26/2022 02:10:58 - INFO - codeparrot_training - Step 37545: {'lr': 7.856525971136367e-05, 'samples': 19223552, 'steps': 37545, 'loss/train': 0.4558207392692566} 02/26/2022 02:11:03 - INFO - codeparrot_training - Step 37546: {'lr': 7.855335070063329e-05, 'samples': 19224064, 'steps': 37546, 'loss/train': 0.9627923965454102} 02/26/2022 02:11:07 - INFO - codeparrot_training - Step 37547: {'lr': 7.854144242432585e-05, 'samples': 19224576, 'steps': 37547, 'loss/train': 2.3472740650177} 02/26/2022 02:11:13 - INFO - codeparrot_training - Step 37548: {'lr': 7.852953488249248e-05, 'samples': 19225088, 'steps': 37548, 'loss/train': 2.2475218772888184} 02/26/2022 02:11:16 - INFO - codeparrot_training - Step 37549: {'lr': 7.851762807518414e-05, 'samples': 19225600, 'steps': 37549, 'loss/train': 1.0498695373535156} 02/26/2022 02:11:22 - INFO - codeparrot_training - Step 37550: {'lr': 7.850572200245185e-05, 'samples': 19226112, 'steps': 37550, 'loss/train': 1.9693659543991089} 02/26/2022 02:11:25 - INFO - codeparrot_training - Step 37551: {'lr': 7.849381666434652e-05, 'samples': 19226624, 'steps': 37551, 'loss/train': 1.588321328163147} 02/26/2022 02:11:31 - INFO - codeparrot_training - Step 37552: {'lr': 7.848191206091926e-05, 'samples': 19227136, 'steps': 37552, 'loss/train': 0.9630889892578125} 02/26/2022 02:11:34 - INFO - codeparrot_training - Step 37553: {'lr': 7.847000819222098e-05, 'samples': 19227648, 'steps': 37553, 'loss/train': 2.2406740188598633} 02/26/2022 02:11:40 - INFO - codeparrot_training - Step 37554: {'lr': 7.845810505830286e-05, 'samples': 19228160, 'steps': 37554, 'loss/train': 1.8263521194458008} 02/26/2022 02:11:43 - INFO - codeparrot_training - Step 37555: {'lr': 7.84462026592156e-05, 'samples': 19228672, 'steps': 37555, 'loss/train': 1.951887607574463} 02/26/2022 02:11:49 - INFO - codeparrot_training - Step 37556: {'lr': 7.84343009950104e-05, 'samples': 19229184, 'steps': 37556, 'loss/train': 1.4607607126235962} 02/26/2022 02:11:52 - INFO - codeparrot_training - Step 37557: {'lr': 7.842240006573811e-05, 'samples': 19229696, 'steps': 37557, 'loss/train': 2.2483668327331543} 02/26/2022 02:11:58 - INFO - codeparrot_training - Step 37558: {'lr': 7.841049987144994e-05, 'samples': 19230208, 'steps': 37558, 'loss/train': 1.1881259679794312} 02/26/2022 02:12:02 - INFO - codeparrot_training - Step 37559: {'lr': 7.839860041219651e-05, 'samples': 19230720, 'steps': 37559, 'loss/train': 1.3830498456954956} 02/26/2022 02:12:07 - INFO - codeparrot_training - Step 37560: {'lr': 7.838670168802908e-05, 'samples': 19231232, 'steps': 37560, 'loss/train': 0.6501631736755371} 02/26/2022 02:12:11 - INFO - codeparrot_training - Step 37561: {'lr': 7.837480369899855e-05, 'samples': 19231744, 'steps': 37561, 'loss/train': 1.4882956743240356} 02/26/2022 02:12:16 - INFO - codeparrot_training - Step 37562: {'lr': 7.836290644515584e-05, 'samples': 19232256, 'steps': 37562, 'loss/train': 2.1507835388183594} 02/26/2022 02:12:20 - INFO - codeparrot_training - Step 37563: {'lr': 7.835100992655186e-05, 'samples': 19232768, 'steps': 37563, 'loss/train': 0.5640310049057007} 02/26/2022 02:12:25 - INFO - codeparrot_training - Step 37564: {'lr': 7.833911414323775e-05, 'samples': 19233280, 'steps': 37564, 'loss/train': 1.0250979661941528} 02/26/2022 02:12:29 - INFO - codeparrot_training - Step 37565: {'lr': 7.832721909526436e-05, 'samples': 19233792, 'steps': 37565, 'loss/train': 2.854789972305298} 02/26/2022 02:12:34 - INFO - codeparrot_training - Step 37566: {'lr': 7.831532478268256e-05, 'samples': 19234304, 'steps': 37566, 'loss/train': 1.098199725151062} 02/26/2022 02:12:38 - INFO - codeparrot_training - Step 37567: {'lr': 7.830343120554345e-05, 'samples': 19234816, 'steps': 37567, 'loss/train': 2.814739465713501} 02/26/2022 02:12:43 - INFO - codeparrot_training - Step 37568: {'lr': 7.829153836389796e-05, 'samples': 19235328, 'steps': 37568, 'loss/train': 1.360663652420044} 02/26/2022 02:12:47 - INFO - codeparrot_training - Step 37569: {'lr': 7.827964625779699e-05, 'samples': 19235840, 'steps': 37569, 'loss/train': 2.1116974353790283} 02/26/2022 02:12:52 - INFO - codeparrot_training - Step 37570: {'lr': 7.82677548872914e-05, 'samples': 19236352, 'steps': 37570, 'loss/train': 1.5010920763015747} 02/26/2022 02:12:56 - INFO - codeparrot_training - Step 37571: {'lr': 7.825586425243228e-05, 'samples': 19236864, 'steps': 37571, 'loss/train': 1.6926275491714478} 02/26/2022 02:13:02 - INFO - codeparrot_training - Step 37572: {'lr': 7.824397435327051e-05, 'samples': 19237376, 'steps': 37572, 'loss/train': 3.410578966140747} 02/26/2022 02:13:05 - INFO - codeparrot_training - Step 37573: {'lr': 7.823208518985706e-05, 'samples': 19237888, 'steps': 37573, 'loss/train': 1.316833257675171} 02/26/2022 02:13:10 - INFO - codeparrot_training - Step 37574: {'lr': 7.822019676224268e-05, 'samples': 19238400, 'steps': 37574, 'loss/train': 1.5929675102233887} 02/26/2022 02:13:14 - INFO - codeparrot_training - Step 37575: {'lr': 7.82083090704786e-05, 'samples': 19238912, 'steps': 37575, 'loss/train': 1.9950404167175293} 02/26/2022 02:13:20 - INFO - codeparrot_training - Step 37576: {'lr': 7.819642211461545e-05, 'samples': 19239424, 'steps': 37576, 'loss/train': 0.31540194153785706} 02/26/2022 02:13:24 - INFO - codeparrot_training - Step 37577: {'lr': 7.818453589470431e-05, 'samples': 19239936, 'steps': 37577, 'loss/train': 1.7159521579742432} 02/26/2022 02:13:29 - INFO - codeparrot_training - Step 37578: {'lr': 7.817265041079599e-05, 'samples': 19240448, 'steps': 37578, 'loss/train': 1.6884583234786987} 02/26/2022 02:13:33 - INFO - codeparrot_training - Step 37579: {'lr': 7.816076566294159e-05, 'samples': 19240960, 'steps': 37579, 'loss/train': 1.075323462486267} 02/26/2022 02:13:38 - INFO - codeparrot_training - Step 37580: {'lr': 7.814888165119186e-05, 'samples': 19241472, 'steps': 37580, 'loss/train': 1.7058758735656738} 02/26/2022 02:13:44 - INFO - codeparrot_training - Step 37581: {'lr': 7.813699837559777e-05, 'samples': 19241984, 'steps': 37581, 'loss/train': 0.7966915369033813} 02/26/2022 02:13:47 - INFO - codeparrot_training - Step 37582: {'lr': 7.812511583621012e-05, 'samples': 19242496, 'steps': 37582, 'loss/train': 2.2373738288879395} 02/26/2022 02:13:53 - INFO - codeparrot_training - Step 37583: {'lr': 7.811323403307999e-05, 'samples': 19243008, 'steps': 37583, 'loss/train': 2.661898374557495} 02/26/2022 02:13:56 - INFO - codeparrot_training - Step 37584: {'lr': 7.810135296625817e-05, 'samples': 19243520, 'steps': 37584, 'loss/train': 0.9169530868530273} 02/26/2022 02:14:03 - INFO - codeparrot_training - Step 37585: {'lr': 7.80894726357956e-05, 'samples': 19244032, 'steps': 37585, 'loss/train': 1.4878981113433838} 02/26/2022 02:14:06 - INFO - codeparrot_training - Step 37586: {'lr': 7.807759304174302e-05, 'samples': 19244544, 'steps': 37586, 'loss/train': 2.100100040435791} 02/26/2022 02:14:12 - INFO - codeparrot_training - Step 37587: {'lr': 7.806571418415156e-05, 'samples': 19245056, 'steps': 37587, 'loss/train': 0.7248440980911255} 02/26/2022 02:14:15 - INFO - codeparrot_training - Step 37588: {'lr': 7.805383606307196e-05, 'samples': 19245568, 'steps': 37588, 'loss/train': 1.9430586099624634} 02/26/2022 02:14:21 - INFO - codeparrot_training - Step 37589: {'lr': 7.804195867855513e-05, 'samples': 19246080, 'steps': 37589, 'loss/train': 2.1219544410705566} 02/26/2022 02:14:24 - INFO - codeparrot_training - Step 37590: {'lr': 7.803008203065187e-05, 'samples': 19246592, 'steps': 37590, 'loss/train': 2.662882089614868} 02/26/2022 02:14:30 - INFO - codeparrot_training - Step 37591: {'lr': 7.801820611941322e-05, 'samples': 19247104, 'steps': 37591, 'loss/train': 2.174276828765869} 02/26/2022 02:14:33 - INFO - codeparrot_training - Step 37592: {'lr': 7.800633094488987e-05, 'samples': 19247616, 'steps': 37592, 'loss/train': 1.1945936679840088} 02/26/2022 02:14:39 - INFO - codeparrot_training - Step 37593: {'lr': 7.799445650713294e-05, 'samples': 19248128, 'steps': 37593, 'loss/train': 1.6402393579483032} 02/26/2022 02:14:42 - INFO - codeparrot_training - Step 37594: {'lr': 7.7982582806193e-05, 'samples': 19248640, 'steps': 37594, 'loss/train': 1.4843868017196655} 02/26/2022 02:14:48 - INFO - codeparrot_training - Step 37595: {'lr': 7.797070984212112e-05, 'samples': 19249152, 'steps': 37595, 'loss/train': 2.216923475265503} 02/26/2022 02:14:52 - INFO - codeparrot_training - Step 37596: {'lr': 7.795883761496803e-05, 'samples': 19249664, 'steps': 37596, 'loss/train': 0.7403322458267212} 02/26/2022 02:14:59 - INFO - codeparrot_training - Step 37597: {'lr': 7.794696612478478e-05, 'samples': 19250176, 'steps': 37597, 'loss/train': 1.7882583141326904} 02/26/2022 02:15:02 - INFO - codeparrot_training - Step 37598: {'lr': 7.793509537162197e-05, 'samples': 19250688, 'steps': 37598, 'loss/train': 2.614980697631836} 02/26/2022 02:15:08 - INFO - codeparrot_training - Step 37599: {'lr': 7.792322535553062e-05, 'samples': 19251200, 'steps': 37599, 'loss/train': 1.7173736095428467} 02/26/2022 02:15:11 - INFO - codeparrot_training - Step 37600: {'lr': 7.791135607656147e-05, 'samples': 19251712, 'steps': 37600, 'loss/train': 3.3310883045196533} 02/26/2022 02:15:17 - INFO - codeparrot_training - Step 37601: {'lr': 7.78994875347655e-05, 'samples': 19252224, 'steps': 37601, 'loss/train': 1.8627843856811523} 02/26/2022 02:15:20 - INFO - codeparrot_training - Step 37602: {'lr': 7.788761973019348e-05, 'samples': 19252736, 'steps': 37602, 'loss/train': 1.2182908058166504} 02/26/2022 02:15:26 - INFO - codeparrot_training - Step 37603: {'lr': 7.787575266289626e-05, 'samples': 19253248, 'steps': 37603, 'loss/train': 0.721784770488739} 02/26/2022 02:15:29 - INFO - codeparrot_training - Step 37604: {'lr': 7.786388633292457e-05, 'samples': 19253760, 'steps': 37604, 'loss/train': 2.258483648300171} 02/26/2022 02:15:35 - INFO - codeparrot_training - Step 37605: {'lr': 7.78520207403294e-05, 'samples': 19254272, 'steps': 37605, 'loss/train': 2.0320520401000977} 02/26/2022 02:15:38 - INFO - codeparrot_training - Step 37606: {'lr': 7.784015588516152e-05, 'samples': 19254784, 'steps': 37606, 'loss/train': 0.6851996183395386} 02/26/2022 02:15:45 - INFO - codeparrot_training - Step 37607: {'lr': 7.782829176747174e-05, 'samples': 19255296, 'steps': 37607, 'loss/train': 1.8158432245254517} 02/26/2022 02:15:48 - INFO - codeparrot_training - Step 37608: {'lr': 7.78164283873109e-05, 'samples': 19255808, 'steps': 37608, 'loss/train': 1.988140344619751} 02/26/2022 02:15:54 - INFO - codeparrot_training - Step 37609: {'lr': 7.780456574472971e-05, 'samples': 19256320, 'steps': 37609, 'loss/train': 0.06278786808252335} 02/26/2022 02:15:57 - INFO - codeparrot_training - Step 37610: {'lr': 7.779270383977916e-05, 'samples': 19256832, 'steps': 37610, 'loss/train': 1.2416313886642456} 02/26/2022 02:16:03 - INFO - codeparrot_training - Step 37611: {'lr': 7.778084267251001e-05, 'samples': 19257344, 'steps': 37611, 'loss/train': 1.9865598678588867} 02/26/2022 02:16:06 - INFO - codeparrot_training - Step 37612: {'lr': 7.776898224297302e-05, 'samples': 19257856, 'steps': 37612, 'loss/train': 1.4167735576629639} 02/26/2022 02:16:12 - INFO - codeparrot_training - Step 37613: {'lr': 7.775712255121895e-05, 'samples': 19258368, 'steps': 37613, 'loss/train': 1.3296536207199097} 02/26/2022 02:16:15 - INFO - codeparrot_training - Step 37614: {'lr': 7.774526359729876e-05, 'samples': 19258880, 'steps': 37614, 'loss/train': 1.7219594717025757} 02/26/2022 02:16:21 - INFO - codeparrot_training - Step 37615: {'lr': 7.773340538126314e-05, 'samples': 19259392, 'steps': 37615, 'loss/train': 0.7995109558105469} 02/26/2022 02:16:24 - INFO - codeparrot_training - Step 37616: {'lr': 7.772154790316294e-05, 'samples': 19259904, 'steps': 37616, 'loss/train': 1.7734417915344238} 02/26/2022 02:16:30 - INFO - codeparrot_training - Step 37617: {'lr': 7.770969116304883e-05, 'samples': 19260416, 'steps': 37617, 'loss/train': 2.0931289196014404} 02/26/2022 02:16:33 - INFO - codeparrot_training - Step 37618: {'lr': 7.769783516097179e-05, 'samples': 19260928, 'steps': 37618, 'loss/train': 1.3326913118362427} 02/26/2022 02:16:40 - INFO - codeparrot_training - Step 37619: {'lr': 7.768597989698248e-05, 'samples': 19261440, 'steps': 37619, 'loss/train': 0.1903793215751648} 02/26/2022 02:16:43 - INFO - codeparrot_training - Step 37620: {'lr': 7.767412537113175e-05, 'samples': 19261952, 'steps': 37620, 'loss/train': 1.813303828239441} 02/26/2022 02:16:49 - INFO - codeparrot_training - Step 37621: {'lr': 7.766227158347025e-05, 'samples': 19262464, 'steps': 37621, 'loss/train': 3.632634162902832} 02/26/2022 02:16:52 - INFO - codeparrot_training - Step 37622: {'lr': 7.765041853404892e-05, 'samples': 19262976, 'steps': 37622, 'loss/train': 3.0631906986236572} 02/26/2022 02:16:58 - INFO - codeparrot_training - Step 37623: {'lr': 7.763856622291848e-05, 'samples': 19263488, 'steps': 37623, 'loss/train': 2.079645872116089} 02/26/2022 02:17:01 - INFO - codeparrot_training - Step 37624: {'lr': 7.762671465012972e-05, 'samples': 19264000, 'steps': 37624, 'loss/train': 1.6525745391845703} 02/26/2022 02:17:07 - INFO - codeparrot_training - Step 37625: {'lr': 7.761486381573326e-05, 'samples': 19264512, 'steps': 37625, 'loss/train': 1.8114641904830933} 02/26/2022 02:17:10 - INFO - codeparrot_training - Step 37626: {'lr': 7.76030137197801e-05, 'samples': 19265024, 'steps': 37626, 'loss/train': 0.35486334562301636} 02/26/2022 02:17:16 - INFO - codeparrot_training - Step 37627: {'lr': 7.759116436232077e-05, 'samples': 19265536, 'steps': 37627, 'loss/train': 1.0562305450439453} 02/26/2022 02:17:19 - INFO - codeparrot_training - Step 37628: {'lr': 7.757931574340635e-05, 'samples': 19266048, 'steps': 37628, 'loss/train': 1.7527967691421509} 02/26/2022 02:17:26 - INFO - codeparrot_training - Step 37629: {'lr': 7.756746786308719e-05, 'samples': 19266560, 'steps': 37629, 'loss/train': 1.2501088380813599} 02/26/2022 02:17:29 - INFO - codeparrot_training - Step 37630: {'lr': 7.755562072141434e-05, 'samples': 19267072, 'steps': 37630, 'loss/train': 1.4705219268798828} 02/26/2022 02:17:35 - INFO - codeparrot_training - Step 37631: {'lr': 7.75437743184384e-05, 'samples': 19267584, 'steps': 37631, 'loss/train': 1.6251333951950073} 02/26/2022 02:17:38 - INFO - codeparrot_training - Step 37632: {'lr': 7.75319286542103e-05, 'samples': 19268096, 'steps': 37632, 'loss/train': 2.0825035572052} 02/26/2022 02:17:44 - INFO - codeparrot_training - Step 37633: {'lr': 7.75200837287805e-05, 'samples': 19268608, 'steps': 37633, 'loss/train': 2.6459314823150635} 02/26/2022 02:17:48 - INFO - codeparrot_training - Step 37634: {'lr': 7.750823954219996e-05, 'samples': 19269120, 'steps': 37634, 'loss/train': 1.6984012126922607} 02/26/2022 02:17:53 - INFO - codeparrot_training - Step 37635: {'lr': 7.749639609451928e-05, 'samples': 19269632, 'steps': 37635, 'loss/train': 1.7171841859817505} 02/26/2022 02:17:57 - INFO - codeparrot_training - Step 37636: {'lr': 7.748455338578941e-05, 'samples': 19270144, 'steps': 37636, 'loss/train': 1.8325073719024658} 02/26/2022 02:18:02 - INFO - codeparrot_training - Step 37637: {'lr': 7.747271141606079e-05, 'samples': 19270656, 'steps': 37637, 'loss/train': 1.9868876934051514} 02/26/2022 02:18:06 - INFO - codeparrot_training - Step 37638: {'lr': 7.746087018538437e-05, 'samples': 19271168, 'steps': 37638, 'loss/train': 0.4312187135219574} 02/26/2022 02:18:11 - INFO - codeparrot_training - Step 37639: {'lr': 7.744902969381071e-05, 'samples': 19271680, 'steps': 37639, 'loss/train': 1.9650650024414062} 02/26/2022 02:18:15 - INFO - codeparrot_training - Step 37640: {'lr': 7.743718994139071e-05, 'samples': 19272192, 'steps': 37640, 'loss/train': 1.4446587562561035} 02/26/2022 02:18:20 - INFO - codeparrot_training - Step 37641: {'lr': 7.742535092817499e-05, 'samples': 19272704, 'steps': 37641, 'loss/train': 1.2791903018951416} 02/26/2022 02:18:27 - INFO - codeparrot_training - Step 37642: {'lr': 7.741351265421426e-05, 'samples': 19273216, 'steps': 37642, 'loss/train': 1.136134386062622} 02/26/2022 02:18:30 - INFO - codeparrot_training - Step 37643: {'lr': 7.740167511955915e-05, 'samples': 19273728, 'steps': 37643, 'loss/train': 0.8914410471916199} 02/26/2022 02:18:37 - INFO - codeparrot_training - Step 37644: {'lr': 7.738983832426056e-05, 'samples': 19274240, 'steps': 37644, 'loss/train': 1.6500110626220703} 02/26/2022 02:18:40 - INFO - codeparrot_training - Step 37645: {'lr': 7.737800226836905e-05, 'samples': 19274752, 'steps': 37645, 'loss/train': 1.1622395515441895} 02/26/2022 02:18:43 - INFO - codeparrot_training - Step 37646: {'lr': 7.736616695193541e-05, 'samples': 19275264, 'steps': 37646, 'loss/train': 0.6352105736732483} 02/26/2022 02:18:49 - INFO - codeparrot_training - Step 37647: {'lr': 7.735433237501018e-05, 'samples': 19275776, 'steps': 37647, 'loss/train': 1.77727472782135} 02/26/2022 02:18:52 - INFO - codeparrot_training - Step 37648: {'lr': 7.734249853764428e-05, 'samples': 19276288, 'steps': 37648, 'loss/train': 1.2371327877044678} 02/26/2022 02:18:58 - INFO - codeparrot_training - Step 37649: {'lr': 7.733066543988828e-05, 'samples': 19276800, 'steps': 37649, 'loss/train': 0.880501389503479} 02/26/2022 02:19:02 - INFO - codeparrot_training - Step 37650: {'lr': 7.731883308179289e-05, 'samples': 19277312, 'steps': 37650, 'loss/train': 1.0649651288986206} 02/26/2022 02:19:07 - INFO - codeparrot_training - Step 37651: {'lr': 7.730700146340868e-05, 'samples': 19277824, 'steps': 37651, 'loss/train': 1.6635488271713257} 02/26/2022 02:19:10 - INFO - codeparrot_training - Step 37652: {'lr': 7.729517058478653e-05, 'samples': 19278336, 'steps': 37652, 'loss/train': 1.3024067878723145} 02/26/2022 02:19:17 - INFO - codeparrot_training - Step 37653: {'lr': 7.728334044597704e-05, 'samples': 19278848, 'steps': 37653, 'loss/train': 2.0576956272125244} 02/26/2022 02:19:20 - INFO - codeparrot_training - Step 37654: {'lr': 7.727151104703086e-05, 'samples': 19279360, 'steps': 37654, 'loss/train': 1.7574636936187744} 02/26/2022 02:19:26 - INFO - codeparrot_training - Step 37655: {'lr': 7.725968238799869e-05, 'samples': 19279872, 'steps': 37655, 'loss/train': 2.5031111240386963} 02/26/2022 02:19:29 - INFO - codeparrot_training - Step 37656: {'lr': 7.724785446893112e-05, 'samples': 19280384, 'steps': 37656, 'loss/train': 1.6008076667785645} 02/26/2022 02:19:35 - INFO - codeparrot_training - Step 37657: {'lr': 7.723602728987894e-05, 'samples': 19280896, 'steps': 37657, 'loss/train': 1.8773912191390991} 02/26/2022 02:19:38 - INFO - codeparrot_training - Step 37658: {'lr': 7.722420085089277e-05, 'samples': 19281408, 'steps': 37658, 'loss/train': 2.6788976192474365} 02/26/2022 02:19:44 - INFO - codeparrot_training - Step 37659: {'lr': 7.721237515202329e-05, 'samples': 19281920, 'steps': 37659, 'loss/train': 0.5612935423851013} 02/26/2022 02:19:47 - INFO - codeparrot_training - Step 37660: {'lr': 7.7200550193321e-05, 'samples': 19282432, 'steps': 37660, 'loss/train': 2.197359085083008} 02/26/2022 02:19:53 - INFO - codeparrot_training - Step 37661: {'lr': 7.718872597483679e-05, 'samples': 19282944, 'steps': 37661, 'loss/train': 0.9181150197982788} 02/26/2022 02:19:56 - INFO - codeparrot_training - Step 37662: {'lr': 7.717690249662113e-05, 'samples': 19283456, 'steps': 37662, 'loss/train': 2.171093463897705} 02/26/2022 02:20:02 - INFO - codeparrot_training - Step 37663: {'lr': 7.716507975872489e-05, 'samples': 19283968, 'steps': 37663, 'loss/train': 1.2959262132644653} 02/26/2022 02:20:05 - INFO - codeparrot_training - Step 37664: {'lr': 7.715325776119841e-05, 'samples': 19284480, 'steps': 37664, 'loss/train': 1.7885850667953491} 02/26/2022 02:20:12 - INFO - codeparrot_training - Step 37665: {'lr': 7.714143650409258e-05, 'samples': 19284992, 'steps': 37665, 'loss/train': 1.8924858570098877} 02/26/2022 02:20:15 - INFO - codeparrot_training - Step 37666: {'lr': 7.712961598745785e-05, 'samples': 19285504, 'steps': 37666, 'loss/train': 1.885291337966919} 02/26/2022 02:20:21 - INFO - codeparrot_training - Step 37667: {'lr': 7.711779621134512e-05, 'samples': 19286016, 'steps': 37667, 'loss/train': 2.95695424079895} 02/26/2022 02:20:24 - INFO - codeparrot_training - Step 37668: {'lr': 7.71059771758047e-05, 'samples': 19286528, 'steps': 37668, 'loss/train': 1.8408385515213013} 02/26/2022 02:20:30 - INFO - codeparrot_training - Step 37669: {'lr': 7.709415888088744e-05, 'samples': 19287040, 'steps': 37669, 'loss/train': 0.85841304063797} 02/26/2022 02:20:34 - INFO - codeparrot_training - Step 37670: {'lr': 7.708234132664382e-05, 'samples': 19287552, 'steps': 37670, 'loss/train': 1.4242287874221802} 02/26/2022 02:20:39 - INFO - codeparrot_training - Step 37671: {'lr': 7.707052451312471e-05, 'samples': 19288064, 'steps': 37671, 'loss/train': 1.4721338748931885} 02/26/2022 02:20:43 - INFO - codeparrot_training - Step 37672: {'lr': 7.705870844038041e-05, 'samples': 19288576, 'steps': 37672, 'loss/train': 0.7704534530639648} 02/26/2022 02:20:48 - INFO - codeparrot_training - Step 37673: {'lr': 7.704689310846175e-05, 'samples': 19289088, 'steps': 37673, 'loss/train': 2.886709213256836} 02/26/2022 02:20:52 - INFO - codeparrot_training - Step 37674: {'lr': 7.70350785174192e-05, 'samples': 19289600, 'steps': 37674, 'loss/train': 0.5283398628234863} 02/26/2022 02:20:58 - INFO - codeparrot_training - Step 37675: {'lr': 7.702326466730356e-05, 'samples': 19290112, 'steps': 37675, 'loss/train': 1.98975670337677} 02/26/2022 02:21:02 - INFO - codeparrot_training - Step 37676: {'lr': 7.70114515581653e-05, 'samples': 19290624, 'steps': 37676, 'loss/train': 2.1917145252227783} 02/26/2022 02:21:07 - INFO - codeparrot_training - Step 37677: {'lr': 7.699963919005506e-05, 'samples': 19291136, 'steps': 37677, 'loss/train': 0.47659945487976074} 02/26/2022 02:21:11 - INFO - codeparrot_training - Step 37678: {'lr': 7.698782756302336e-05, 'samples': 19291648, 'steps': 37678, 'loss/train': 0.7841382026672363} 02/26/2022 02:21:16 - INFO - codeparrot_training - Step 37679: {'lr': 7.697601667712092e-05, 'samples': 19292160, 'steps': 37679, 'loss/train': 3.004986524581909} 02/26/2022 02:21:20 - INFO - codeparrot_training - Step 37680: {'lr': 7.696420653239832e-05, 'samples': 19292672, 'steps': 37680, 'loss/train': 0.6561324596405029} 02/26/2022 02:21:25 - INFO - codeparrot_training - Step 37681: {'lr': 7.695239712890609e-05, 'samples': 19293184, 'steps': 37681, 'loss/train': 2.267157793045044} 02/26/2022 02:21:29 - INFO - codeparrot_training - Step 37682: {'lr': 7.694058846669477e-05, 'samples': 19293696, 'steps': 37682, 'loss/train': 2.2618095874786377} 02/26/2022 02:21:34 - INFO - codeparrot_training - Step 37683: {'lr': 7.692878054581512e-05, 'samples': 19294208, 'steps': 37683, 'loss/train': 2.078975200653076} 02/26/2022 02:21:38 - INFO - codeparrot_training - Step 37684: {'lr': 7.69169733663176e-05, 'samples': 19294720, 'steps': 37684, 'loss/train': 1.5002537965774536} 02/26/2022 02:21:44 - INFO - codeparrot_training - Step 37685: {'lr': 7.690516692825278e-05, 'samples': 19295232, 'steps': 37685, 'loss/train': 2.2105298042297363} 02/26/2022 02:21:47 - INFO - codeparrot_training - Step 37686: {'lr': 7.68933612316712e-05, 'samples': 19295744, 'steps': 37686, 'loss/train': 2.2338786125183105} 02/26/2022 02:21:53 - INFO - codeparrot_training - Step 37687: {'lr': 7.688155627662357e-05, 'samples': 19296256, 'steps': 37687, 'loss/train': 2.2466061115264893} 02/26/2022 02:21:56 - INFO - codeparrot_training - Step 37688: {'lr': 7.68697520631604e-05, 'samples': 19296768, 'steps': 37688, 'loss/train': 1.1804167032241821} 02/26/2022 02:22:02 - INFO - codeparrot_training - Step 37689: {'lr': 7.685794859133222e-05, 'samples': 19297280, 'steps': 37689, 'loss/train': 0.8269443511962891} 02/26/2022 02:22:05 - INFO - codeparrot_training - Step 37690: {'lr': 7.684614586118954e-05, 'samples': 19297792, 'steps': 37690, 'loss/train': 1.299662709236145} 02/26/2022 02:22:12 - INFO - codeparrot_training - Step 37691: {'lr': 7.683434387278308e-05, 'samples': 19298304, 'steps': 37691, 'loss/train': 0.918686032295227} 02/26/2022 02:22:15 - INFO - codeparrot_training - Step 37692: {'lr': 7.682254262616326e-05, 'samples': 19298816, 'steps': 37692, 'loss/train': 2.050319194793701} 02/26/2022 02:22:21 - INFO - codeparrot_training - Step 37693: {'lr': 7.68107421213807e-05, 'samples': 19299328, 'steps': 37693, 'loss/train': 1.658482551574707} 02/26/2022 02:22:24 - INFO - codeparrot_training - Step 37694: {'lr': 7.679894235848586e-05, 'samples': 19299840, 'steps': 37694, 'loss/train': 1.7333959341049194} 02/26/2022 02:22:30 - INFO - codeparrot_training - Step 37695: {'lr': 7.678714333752942e-05, 'samples': 19300352, 'steps': 37695, 'loss/train': 1.998561143875122} 02/26/2022 02:22:33 - INFO - codeparrot_training - Step 37696: {'lr': 7.677534505856188e-05, 'samples': 19300864, 'steps': 37696, 'loss/train': 0.7671431303024292} 02/26/2022 02:22:39 - INFO - codeparrot_training - Step 37697: {'lr': 7.676354752163373e-05, 'samples': 19301376, 'steps': 37697, 'loss/train': 2.2496800422668457} 02/26/2022 02:22:42 - INFO - codeparrot_training - Step 37698: {'lr': 7.675175072679546e-05, 'samples': 19301888, 'steps': 37698, 'loss/train': 2.202409505844116} 02/26/2022 02:22:48 - INFO - codeparrot_training - Step 37699: {'lr': 7.673995467409775e-05, 'samples': 19302400, 'steps': 37699, 'loss/train': 0.9044036269187927} 02/26/2022 02:22:51 - INFO - codeparrot_training - Step 37700: {'lr': 7.672815936359106e-05, 'samples': 19302912, 'steps': 37700, 'loss/train': 1.8131860494613647} 02/26/2022 02:22:58 - INFO - codeparrot_training - Step 37701: {'lr': 7.671636479532584e-05, 'samples': 19303424, 'steps': 37701, 'loss/train': 1.8823350667953491} 02/26/2022 02:23:01 - INFO - codeparrot_training - Step 37702: {'lr': 7.670457096935279e-05, 'samples': 19303936, 'steps': 37702, 'loss/train': 1.5045844316482544} 02/26/2022 02:23:07 - INFO - codeparrot_training - Step 37703: {'lr': 7.66927778857223e-05, 'samples': 19304448, 'steps': 37703, 'loss/train': 1.8278326988220215} 02/26/2022 02:23:10 - INFO - codeparrot_training - Step 37704: {'lr': 7.668098554448493e-05, 'samples': 19304960, 'steps': 37704, 'loss/train': 2.500993251800537} 02/26/2022 02:23:16 - INFO - codeparrot_training - Step 37705: {'lr': 7.666919394569111e-05, 'samples': 19305472, 'steps': 37705, 'loss/train': 0.6172981858253479} 02/26/2022 02:23:19 - INFO - codeparrot_training - Step 37706: {'lr': 7.665740308939156e-05, 'samples': 19305984, 'steps': 37706, 'loss/train': 1.8391003608703613} 02/26/2022 02:23:25 - INFO - codeparrot_training - Step 37707: {'lr': 7.664561297563653e-05, 'samples': 19306496, 'steps': 37707, 'loss/train': 0.7069413661956787} 02/26/2022 02:23:28 - INFO - codeparrot_training - Step 37708: {'lr': 7.663382360447672e-05, 'samples': 19307008, 'steps': 37708, 'loss/train': 1.7706489562988281} 02/26/2022 02:23:34 - INFO - codeparrot_training - Step 37709: {'lr': 7.662203497596246e-05, 'samples': 19307520, 'steps': 37709, 'loss/train': 1.852295160293579} 02/26/2022 02:23:37 - INFO - codeparrot_training - Step 37710: {'lr': 7.661024709014445e-05, 'samples': 19308032, 'steps': 37710, 'loss/train': 1.259118914604187} 02/26/2022 02:23:44 - INFO - codeparrot_training - Step 37711: {'lr': 7.659845994707304e-05, 'samples': 19308544, 'steps': 37711, 'loss/train': 1.4487167596817017} 02/26/2022 02:23:47 - INFO - codeparrot_training - Step 37712: {'lr': 7.65866735467988e-05, 'samples': 19309056, 'steps': 37712, 'loss/train': 0.6487449407577515} 02/26/2022 02:23:53 - INFO - codeparrot_training - Step 37713: {'lr': 7.657488788937211e-05, 'samples': 19309568, 'steps': 37713, 'loss/train': 2.396380662918091} 02/26/2022 02:23:56 - INFO - codeparrot_training - Step 37714: {'lr': 7.65631029748436e-05, 'samples': 19310080, 'steps': 37714, 'loss/train': 1.8077212572097778} 02/26/2022 02:24:02 - INFO - codeparrot_training - Step 37715: {'lr': 7.655131880326369e-05, 'samples': 19310592, 'steps': 37715, 'loss/train': 1.7396163940429688} 02/26/2022 02:24:05 - INFO - codeparrot_training - Step 37716: {'lr': 7.653953537468287e-05, 'samples': 19311104, 'steps': 37716, 'loss/train': 1.4670031070709229} 02/26/2022 02:24:11 - INFO - codeparrot_training - Step 37717: {'lr': 7.652775268915149e-05, 'samples': 19311616, 'steps': 37717, 'loss/train': 1.8817789554595947} 02/26/2022 02:24:14 - INFO - codeparrot_training - Step 37718: {'lr': 7.651597074672023e-05, 'samples': 19312128, 'steps': 37718, 'loss/train': 0.9007397890090942} 02/26/2022 02:24:20 - INFO - codeparrot_training - Step 37719: {'lr': 7.650418954743946e-05, 'samples': 19312640, 'steps': 37719, 'loss/train': 1.0696829557418823} 02/26/2022 02:24:23 - INFO - codeparrot_training - Step 37720: {'lr': 7.649240909135965e-05, 'samples': 19313152, 'steps': 37720, 'loss/train': 1.502488613128662} 02/26/2022 02:24:29 - INFO - codeparrot_training - Step 37721: {'lr': 7.648062937853117e-05, 'samples': 19313664, 'steps': 37721, 'loss/train': 1.6990318298339844} 02/26/2022 02:24:32 - INFO - codeparrot_training - Step 37722: {'lr': 7.646885040900466e-05, 'samples': 19314176, 'steps': 37722, 'loss/train': 2.4838478565216064} 02/26/2022 02:24:38 - INFO - codeparrot_training - Step 37723: {'lr': 7.645707218283044e-05, 'samples': 19314688, 'steps': 37723, 'loss/train': 1.113002896308899} 02/26/2022 02:24:41 - INFO - codeparrot_training - Step 37724: {'lr': 7.644529470005917e-05, 'samples': 19315200, 'steps': 37724, 'loss/train': 1.4262864589691162} 02/26/2022 02:24:49 - INFO - codeparrot_training - Step 37725: {'lr': 7.643351796074097e-05, 'samples': 19315712, 'steps': 37725, 'loss/train': 1.3646433353424072} 02/26/2022 02:24:52 - INFO - codeparrot_training - Step 37726: {'lr': 7.642174196492654e-05, 'samples': 19316224, 'steps': 37726, 'loss/train': 0.8107653856277466} 02/26/2022 02:24:57 - INFO - codeparrot_training - Step 37727: {'lr': 7.64099667126662e-05, 'samples': 19316736, 'steps': 37727, 'loss/train': 0.8023613691329956} 02/26/2022 02:25:01 - INFO - codeparrot_training - Step 37728: {'lr': 7.63981922040106e-05, 'samples': 19317248, 'steps': 37728, 'loss/train': 1.1433221101760864} 02/26/2022 02:25:06 - INFO - codeparrot_training - Step 37729: {'lr': 7.638641843900984e-05, 'samples': 19317760, 'steps': 37729, 'loss/train': 1.3264213800430298} 02/26/2022 02:25:12 - INFO - codeparrot_training - Step 37730: {'lr': 7.63746454177146e-05, 'samples': 19318272, 'steps': 37730, 'loss/train': 1.3353325128555298} 02/26/2022 02:25:16 - INFO - codeparrot_training - Step 37731: {'lr': 7.63628731401752e-05, 'samples': 19318784, 'steps': 37731, 'loss/train': 1.8617967367172241} 02/26/2022 02:25:19 - INFO - codeparrot_training - Step 37732: {'lr': 7.635110160644227e-05, 'samples': 19319296, 'steps': 37732, 'loss/train': 1.646790623664856} 02/26/2022 02:25:25 - INFO - codeparrot_training - Step 37733: {'lr': 7.633933081656592e-05, 'samples': 19319808, 'steps': 37733, 'loss/train': 1.2764478921890259} 02/26/2022 02:25:28 - INFO - codeparrot_training - Step 37734: {'lr': 7.63275607705968e-05, 'samples': 19320320, 'steps': 37734, 'loss/train': 1.2741814851760864} 02/26/2022 02:25:34 - INFO - codeparrot_training - Step 37735: {'lr': 7.631579146858525e-05, 'samples': 19320832, 'steps': 37735, 'loss/train': 1.9060357809066772} 02/26/2022 02:25:37 - INFO - codeparrot_training - Step 37736: {'lr': 7.630402291058164e-05, 'samples': 19321344, 'steps': 37736, 'loss/train': 1.738125205039978} 02/26/2022 02:25:44 - INFO - codeparrot_training - Step 37737: {'lr': 7.62922550966365e-05, 'samples': 19321856, 'steps': 37737, 'loss/train': 2.2915990352630615} 02/26/2022 02:25:48 - INFO - codeparrot_training - Step 37738: {'lr': 7.628048802680021e-05, 'samples': 19322368, 'steps': 37738, 'loss/train': 1.769662857055664} 02/26/2022 02:25:53 - INFO - codeparrot_training - Step 37739: {'lr': 7.626872170112315e-05, 'samples': 19322880, 'steps': 37739, 'loss/train': 1.2634891271591187} 02/26/2022 02:25:57 - INFO - codeparrot_training - Step 37740: {'lr': 7.625695611965564e-05, 'samples': 19323392, 'steps': 37740, 'loss/train': 0.36607810854911804} 02/26/2022 02:26:02 - INFO - codeparrot_training - Step 37741: {'lr': 7.624519128244823e-05, 'samples': 19323904, 'steps': 37741, 'loss/train': 1.370645523071289} 02/26/2022 02:26:06 - INFO - codeparrot_training - Step 37742: {'lr': 7.623342718955123e-05, 'samples': 19324416, 'steps': 37742, 'loss/train': 1.4751585721969604} 02/26/2022 02:26:11 - INFO - codeparrot_training - Step 37743: {'lr': 7.622166384101506e-05, 'samples': 19324928, 'steps': 37743, 'loss/train': 1.82878577709198} 02/26/2022 02:26:15 - INFO - codeparrot_training - Step 37744: {'lr': 7.620990123689004e-05, 'samples': 19325440, 'steps': 37744, 'loss/train': 1.614038348197937} 02/26/2022 02:26:20 - INFO - codeparrot_training - Step 37745: {'lr': 7.619813937722667e-05, 'samples': 19325952, 'steps': 37745, 'loss/train': 1.630193829536438} 02/26/2022 02:26:23 - INFO - codeparrot_training - Step 37746: {'lr': 7.61863782620753e-05, 'samples': 19326464, 'steps': 37746, 'loss/train': 1.2569080591201782} 02/26/2022 02:26:30 - INFO - codeparrot_training - Step 37747: {'lr': 7.617461789148627e-05, 'samples': 19326976, 'steps': 37747, 'loss/train': 0.7265498638153076} 02/26/2022 02:26:33 - INFO - codeparrot_training - Step 37748: {'lr': 7.616285826550995e-05, 'samples': 19327488, 'steps': 37748, 'loss/train': 2.1613547801971436} 02/26/2022 02:26:39 - INFO - codeparrot_training - Step 37749: {'lr': 7.615109938419679e-05, 'samples': 19328000, 'steps': 37749, 'loss/train': 1.4838377237319946} 02/26/2022 02:26:42 - INFO - codeparrot_training - Step 37750: {'lr': 7.613934124759711e-05, 'samples': 19328512, 'steps': 37750, 'loss/train': 1.4032288789749146} 02/26/2022 02:26:48 - INFO - codeparrot_training - Step 37751: {'lr': 7.612758385576133e-05, 'samples': 19329024, 'steps': 37751, 'loss/train': 1.520467758178711} 02/26/2022 02:26:51 - INFO - codeparrot_training - Step 37752: {'lr': 7.611582720873963e-05, 'samples': 19329536, 'steps': 37752, 'loss/train': 0.9275469779968262} 02/26/2022 02:26:57 - INFO - codeparrot_training - Step 37753: {'lr': 7.610407130658264e-05, 'samples': 19330048, 'steps': 37753, 'loss/train': 1.8000197410583496} 02/26/2022 02:27:00 - INFO - codeparrot_training - Step 37754: {'lr': 7.609231614934057e-05, 'samples': 19330560, 'steps': 37754, 'loss/train': 2.0067830085754395} 02/26/2022 02:27:06 - INFO - codeparrot_training - Step 37755: {'lr': 7.60805617370638e-05, 'samples': 19331072, 'steps': 37755, 'loss/train': 1.082345962524414} 02/26/2022 02:27:09 - INFO - codeparrot_training - Step 37756: {'lr': 7.60688080698026e-05, 'samples': 19331584, 'steps': 37756, 'loss/train': 1.7399219274520874} 02/26/2022 02:27:16 - INFO - codeparrot_training - Step 37757: {'lr': 7.605705514760746e-05, 'samples': 19332096, 'steps': 37757, 'loss/train': 1.2099108695983887} 02/26/2022 02:27:20 - INFO - codeparrot_training - Step 37758: {'lr': 7.604530297052866e-05, 'samples': 19332608, 'steps': 37758, 'loss/train': 1.954397201538086} 02/26/2022 02:27:25 - INFO - codeparrot_training - Step 37759: {'lr': 7.603355153861657e-05, 'samples': 19333120, 'steps': 37759, 'loss/train': 2.264927864074707} 02/26/2022 02:27:29 - INFO - codeparrot_training - Step 37760: {'lr': 7.602180085192142e-05, 'samples': 19333632, 'steps': 37760, 'loss/train': 2.3893911838531494} 02/26/2022 02:27:35 - INFO - codeparrot_training - Step 37761: {'lr': 7.60100509104937e-05, 'samples': 19334144, 'steps': 37761, 'loss/train': 1.7324268817901611} 02/26/2022 02:27:38 - INFO - codeparrot_training - Step 37762: {'lr': 7.59983017143836e-05, 'samples': 19334656, 'steps': 37762, 'loss/train': 0.6631697416305542} 02/26/2022 02:27:44 - INFO - codeparrot_training - Step 37763: {'lr': 7.598655326364167e-05, 'samples': 19335168, 'steps': 37763, 'loss/train': 1.6210612058639526} 02/26/2022 02:27:47 - INFO - codeparrot_training - Step 37764: {'lr': 7.597480555831796e-05, 'samples': 19335680, 'steps': 37764, 'loss/train': 0.8646954298019409} 02/26/2022 02:27:53 - INFO - codeparrot_training - Step 37765: {'lr': 7.5963058598463e-05, 'samples': 19336192, 'steps': 37765, 'loss/train': 2.4253830909729004} 02/26/2022 02:27:56 - INFO - codeparrot_training - Step 37766: {'lr': 7.595131238412693e-05, 'samples': 19336704, 'steps': 37766, 'loss/train': 1.2177038192749023} 02/26/2022 02:28:02 - INFO - codeparrot_training - Step 37767: {'lr': 7.593956691536032e-05, 'samples': 19337216, 'steps': 37767, 'loss/train': 2.2936832904815674} 02/26/2022 02:28:05 - INFO - codeparrot_training - Step 37768: {'lr': 7.592782219221323e-05, 'samples': 19337728, 'steps': 37768, 'loss/train': 0.20163801312446594} 02/26/2022 02:28:11 - INFO - codeparrot_training - Step 37769: {'lr': 7.591607821473612e-05, 'samples': 19338240, 'steps': 37769, 'loss/train': 2.099280595779419} 02/26/2022 02:28:14 - INFO - codeparrot_training - Step 37770: {'lr': 7.590433498297916e-05, 'samples': 19338752, 'steps': 37770, 'loss/train': 1.8411799669265747} 02/26/2022 02:28:20 - INFO - codeparrot_training - Step 37771: {'lr': 7.589259249699285e-05, 'samples': 19339264, 'steps': 37771, 'loss/train': 2.408233404159546} 02/26/2022 02:28:23 - INFO - codeparrot_training - Step 37772: {'lr': 7.588085075682738e-05, 'samples': 19339776, 'steps': 37772, 'loss/train': 1.186763048171997} 02/26/2022 02:28:30 - INFO - codeparrot_training - Step 37773: {'lr': 7.586910976253306e-05, 'samples': 19340288, 'steps': 37773, 'loss/train': 1.0824908018112183} 02/26/2022 02:28:35 - INFO - codeparrot_training - Step 37774: {'lr': 7.585736951416012e-05, 'samples': 19340800, 'steps': 37774, 'loss/train': 1.772605538368225} 02/26/2022 02:28:39 - INFO - codeparrot_training - Step 37775: {'lr': 7.584563001175895e-05, 'samples': 19341312, 'steps': 37775, 'loss/train': 1.5589966773986816} 02/26/2022 02:28:44 - INFO - codeparrot_training - Step 37776: {'lr': 7.583389125537982e-05, 'samples': 19341824, 'steps': 37776, 'loss/train': 1.0156264305114746} 02/26/2022 02:28:48 - INFO - codeparrot_training - Step 37777: {'lr': 7.5822153245073e-05, 'samples': 19342336, 'steps': 37777, 'loss/train': 2.102470874786377} 02/26/2022 02:28:53 - INFO - codeparrot_training - Step 37778: {'lr': 7.58104159808887e-05, 'samples': 19342848, 'steps': 37778, 'loss/train': 1.3758769035339355} 02/26/2022 02:28:57 - INFO - codeparrot_training - Step 37779: {'lr': 7.579867946287733e-05, 'samples': 19343360, 'steps': 37779, 'loss/train': 1.672695279121399} 02/26/2022 02:29:02 - INFO - codeparrot_training - Step 37780: {'lr': 7.57869436910891e-05, 'samples': 19343872, 'steps': 37780, 'loss/train': 1.244953989982605} 02/26/2022 02:29:06 - INFO - codeparrot_training - Step 37781: {'lr': 7.577520866557428e-05, 'samples': 19344384, 'steps': 37781, 'loss/train': 2.3688089847564697} 02/26/2022 02:29:12 - INFO - codeparrot_training - Step 37782: {'lr': 7.57634743863831e-05, 'samples': 19344896, 'steps': 37782, 'loss/train': 1.2336266040802002} 02/26/2022 02:29:16 - INFO - codeparrot_training - Step 37783: {'lr': 7.57517408535659e-05, 'samples': 19345408, 'steps': 37783, 'loss/train': 1.7861788272857666} 02/26/2022 02:29:21 - INFO - codeparrot_training - Step 37784: {'lr': 7.574000806717294e-05, 'samples': 19345920, 'steps': 37784, 'loss/train': 2.7549006938934326} 02/26/2022 02:29:25 - INFO - codeparrot_training - Step 37785: {'lr': 7.572827602725443e-05, 'samples': 19346432, 'steps': 37785, 'loss/train': 2.124006748199463} 02/26/2022 02:29:30 - INFO - codeparrot_training - Step 37786: {'lr': 7.571654473386067e-05, 'samples': 19346944, 'steps': 37786, 'loss/train': 2.021242141723633} 02/26/2022 02:29:34 - INFO - codeparrot_training - Step 37787: {'lr': 7.570481418704181e-05, 'samples': 19347456, 'steps': 37787, 'loss/train': 2.0368247032165527} 02/26/2022 02:29:40 - INFO - codeparrot_training - Step 37788: {'lr': 7.569308438684824e-05, 'samples': 19347968, 'steps': 37788, 'loss/train': 2.628750801086426} 02/26/2022 02:29:43 - INFO - codeparrot_training - Step 37789: {'lr': 7.568135533333018e-05, 'samples': 19348480, 'steps': 37789, 'loss/train': 1.37528395652771} 02/26/2022 02:29:49 - INFO - codeparrot_training - Step 37790: {'lr': 7.566962702653781e-05, 'samples': 19348992, 'steps': 37790, 'loss/train': 1.5383455753326416} 02/26/2022 02:29:52 - INFO - codeparrot_training - Step 37791: {'lr': 7.565789946652135e-05, 'samples': 19349504, 'steps': 37791, 'loss/train': 1.5342360734939575} 02/26/2022 02:29:58 - INFO - codeparrot_training - Step 37792: {'lr': 7.564617265333115e-05, 'samples': 19350016, 'steps': 37792, 'loss/train': 2.489480972290039} 02/26/2022 02:30:02 - INFO - codeparrot_training - Step 37793: {'lr': 7.563444658701741e-05, 'samples': 19350528, 'steps': 37793, 'loss/train': 1.9754176139831543} 02/26/2022 02:30:07 - INFO - codeparrot_training - Step 37794: {'lr': 7.562272126763029e-05, 'samples': 19351040, 'steps': 37794, 'loss/train': 2.429378032684326} 02/26/2022 02:30:11 - INFO - codeparrot_training - Step 37795: {'lr': 7.561099669522001e-05, 'samples': 19351552, 'steps': 37795, 'loss/train': 2.4170801639556885} 02/26/2022 02:30:16 - INFO - codeparrot_training - Step 37796: {'lr': 7.559927286983692e-05, 'samples': 19352064, 'steps': 37796, 'loss/train': 2.2970399856567383} 02/26/2022 02:30:20 - INFO - codeparrot_training - Step 37797: {'lr': 7.558754979153107e-05, 'samples': 19352576, 'steps': 37797, 'loss/train': 1.7543789148330688} 02/26/2022 02:30:25 - INFO - codeparrot_training - Step 37798: {'lr': 7.557582746035297e-05, 'samples': 19353088, 'steps': 37798, 'loss/train': 1.4067046642303467} 02/26/2022 02:30:29 - INFO - codeparrot_training - Step 37799: {'lr': 7.556410587635245e-05, 'samples': 19353600, 'steps': 37799, 'loss/train': 1.6838634014129639} 02/26/2022 02:30:34 - INFO - codeparrot_training - Step 37800: {'lr': 7.555238503958001e-05, 'samples': 19354112, 'steps': 37800, 'loss/train': 0.7755517363548279} 02/26/2022 02:30:38 - INFO - codeparrot_training - Step 37801: {'lr': 7.554066495008566e-05, 'samples': 19354624, 'steps': 37801, 'loss/train': 1.8027853965759277} 02/26/2022 02:30:43 - INFO - codeparrot_training - Step 37802: {'lr': 7.552894560791987e-05, 'samples': 19355136, 'steps': 37802, 'loss/train': 1.1575514078140259} 02/26/2022 02:30:47 - INFO - codeparrot_training - Step 37803: {'lr': 7.551722701313251e-05, 'samples': 19355648, 'steps': 37803, 'loss/train': 2.10512375831604} 02/26/2022 02:30:53 - INFO - codeparrot_training - Step 37804: {'lr': 7.550550916577404e-05, 'samples': 19356160, 'steps': 37804, 'loss/train': 1.02147376537323} 02/26/2022 02:30:56 - INFO - codeparrot_training - Step 37805: {'lr': 7.549379206589446e-05, 'samples': 19356672, 'steps': 37805, 'loss/train': 0.7600228190422058} 02/26/2022 02:31:02 - INFO - codeparrot_training - Step 37806: {'lr': 7.548207571354426e-05, 'samples': 19357184, 'steps': 37806, 'loss/train': 1.8661831617355347} 02/26/2022 02:31:05 - INFO - codeparrot_training - Step 37807: {'lr': 7.547036010877323e-05, 'samples': 19357696, 'steps': 37807, 'loss/train': 1.5922367572784424} 02/26/2022 02:31:11 - INFO - codeparrot_training - Step 37808: {'lr': 7.545864525163188e-05, 'samples': 19358208, 'steps': 37808, 'loss/train': 2.172541379928589} 02/26/2022 02:31:14 - INFO - codeparrot_training - Step 37809: {'lr': 7.544693114217016e-05, 'samples': 19358720, 'steps': 37809, 'loss/train': 1.2649165391921997} 02/26/2022 02:31:20 - INFO - codeparrot_training - Step 37810: {'lr': 7.543521778043843e-05, 'samples': 19359232, 'steps': 37810, 'loss/train': 1.7339946031570435} 02/26/2022 02:31:23 - INFO - codeparrot_training - Step 37811: {'lr': 7.542350516648683e-05, 'samples': 19359744, 'steps': 37811, 'loss/train': 1.2232105731964111} 02/26/2022 02:31:29 - INFO - codeparrot_training - Step 37812: {'lr': 7.541179330036549e-05, 'samples': 19360256, 'steps': 37812, 'loss/train': 0.4428284168243408} 02/26/2022 02:31:32 - INFO - codeparrot_training - Step 37813: {'lr': 7.540008218212448e-05, 'samples': 19360768, 'steps': 37813, 'loss/train': 2.109645128250122} 02/26/2022 02:31:38 - INFO - codeparrot_training - Step 37814: {'lr': 7.538837181181418e-05, 'samples': 19361280, 'steps': 37814, 'loss/train': 1.935546636581421} 02/26/2022 02:31:41 - INFO - codeparrot_training - Step 37815: {'lr': 7.537666218948466e-05, 'samples': 19361792, 'steps': 37815, 'loss/train': 2.405400514602661} 02/26/2022 02:31:47 - INFO - codeparrot_training - Step 37816: {'lr': 7.536495331518606e-05, 'samples': 19362304, 'steps': 37816, 'loss/train': 1.3233942985534668} 02/26/2022 02:31:50 - INFO - codeparrot_training - Step 37817: {'lr': 7.535324518896846e-05, 'samples': 19362816, 'steps': 37817, 'loss/train': 2.205601453781128} 02/26/2022 02:31:57 - INFO - codeparrot_training - Step 37818: {'lr': 7.534153781088218e-05, 'samples': 19363328, 'steps': 37818, 'loss/train': 2.071200132369995} 02/26/2022 02:32:00 - INFO - codeparrot_training - Step 37819: {'lr': 7.53298311809773e-05, 'samples': 19363840, 'steps': 37819, 'loss/train': 2.0881972312927246} 02/26/2022 02:32:06 - INFO - codeparrot_training - Step 37820: {'lr': 7.531812529930399e-05, 'samples': 19364352, 'steps': 37820, 'loss/train': 2.171382188796997} 02/26/2022 02:32:09 - INFO - codeparrot_training - Step 37821: {'lr': 7.530642016591225e-05, 'samples': 19364864, 'steps': 37821, 'loss/train': 0.10827948153018951} 02/26/2022 02:32:15 - INFO - codeparrot_training - Step 37822: {'lr': 7.52947157808524e-05, 'samples': 19365376, 'steps': 37822, 'loss/train': 1.5215182304382324} 02/26/2022 02:32:18 - INFO - codeparrot_training - Step 37823: {'lr': 7.528301214417452e-05, 'samples': 19365888, 'steps': 37823, 'loss/train': 1.8233108520507812} 02/26/2022 02:32:24 - INFO - codeparrot_training - Step 37824: {'lr': 7.527130925592876e-05, 'samples': 19366400, 'steps': 37824, 'loss/train': 1.275500774383545} 02/26/2022 02:32:27 - INFO - codeparrot_training - Step 37825: {'lr': 7.525960711616514e-05, 'samples': 19366912, 'steps': 37825, 'loss/train': 1.5067020654678345} 02/26/2022 02:32:33 - INFO - codeparrot_training - Step 37826: {'lr': 7.524790572493396e-05, 'samples': 19367424, 'steps': 37826, 'loss/train': 0.7007200121879578} 02/26/2022 02:32:36 - INFO - codeparrot_training - Step 37827: {'lr': 7.523620508228524e-05, 'samples': 19367936, 'steps': 37827, 'loss/train': 8.622401237487793} 02/26/2022 02:32:43 - INFO - codeparrot_training - Step 37828: {'lr': 7.522450518826912e-05, 'samples': 19368448, 'steps': 37828, 'loss/train': 1.0185248851776123} 02/26/2022 02:32:46 - INFO - codeparrot_training - Step 37829: {'lr': 7.521280604293568e-05, 'samples': 19368960, 'steps': 37829, 'loss/train': 1.9488813877105713} 02/26/2022 02:32:52 - INFO - codeparrot_training - Step 37830: {'lr': 7.520110764633511e-05, 'samples': 19369472, 'steps': 37830, 'loss/train': 0.44856682419776917} 02/26/2022 02:32:55 - INFO - codeparrot_training - Step 37831: {'lr': 7.518940999851753e-05, 'samples': 19369984, 'steps': 37831, 'loss/train': 1.596879243850708} 02/26/2022 02:33:01 - INFO - codeparrot_training - Step 37832: {'lr': 7.517771309953292e-05, 'samples': 19370496, 'steps': 37832, 'loss/train': 2.1883862018585205} 02/26/2022 02:33:04 - INFO - codeparrot_training - Step 37833: {'lr': 7.516601694943162e-05, 'samples': 19371008, 'steps': 37833, 'loss/train': 1.9536049365997314} 02/26/2022 02:33:10 - INFO - codeparrot_training - Step 37834: {'lr': 7.515432154826346e-05, 'samples': 19371520, 'steps': 37834, 'loss/train': 2.092209577560425} 02/26/2022 02:33:13 - INFO - codeparrot_training - Step 37835: {'lr': 7.51426268960787e-05, 'samples': 19372032, 'steps': 37835, 'loss/train': 1.6627986431121826} 02/26/2022 02:33:19 - INFO - codeparrot_training - Step 37836: {'lr': 7.513093299292736e-05, 'samples': 19372544, 'steps': 37836, 'loss/train': 2.3882265090942383} 02/26/2022 02:33:22 - INFO - codeparrot_training - Step 37837: {'lr': 7.511923983885971e-05, 'samples': 19373056, 'steps': 37837, 'loss/train': 0.8633599877357483} 02/26/2022 02:33:28 - INFO - codeparrot_training - Step 37838: {'lr': 7.510754743392555e-05, 'samples': 19373568, 'steps': 37838, 'loss/train': 1.60392427444458} 02/26/2022 02:33:31 - INFO - codeparrot_training - Step 37839: {'lr': 7.509585577817524e-05, 'samples': 19374080, 'steps': 37839, 'loss/train': 1.9865578413009644} 02/26/2022 02:33:38 - INFO - codeparrot_training - Step 37840: {'lr': 7.508416487165862e-05, 'samples': 19374592, 'steps': 37840, 'loss/train': 2.0998520851135254} 02/26/2022 02:33:41 - INFO - codeparrot_training - Step 37841: {'lr': 7.507247471442607e-05, 'samples': 19375104, 'steps': 37841, 'loss/train': 2.170912981033325} 02/26/2022 02:33:47 - INFO - codeparrot_training - Step 37842: {'lr': 7.506078530652736e-05, 'samples': 19375616, 'steps': 37842, 'loss/train': 1.8825217485427856} 02/26/2022 02:33:50 - INFO - codeparrot_training - Step 37843: {'lr': 7.504909664801273e-05, 'samples': 19376128, 'steps': 37843, 'loss/train': 1.8935712575912476} 02/26/2022 02:33:56 - INFO - codeparrot_training - Step 37844: {'lr': 7.503740873893217e-05, 'samples': 19376640, 'steps': 37844, 'loss/train': 1.1253705024719238} 02/26/2022 02:33:59 - INFO - codeparrot_training - Step 37845: {'lr': 7.502572157933585e-05, 'samples': 19377152, 'steps': 37845, 'loss/train': 0.06850094348192215} 02/26/2022 02:34:05 - INFO - codeparrot_training - Step 37846: {'lr': 7.50140351692738e-05, 'samples': 19377664, 'steps': 37846, 'loss/train': 0.3980892300605774} 02/26/2022 02:34:08 - INFO - codeparrot_training - Step 37847: {'lr': 7.500234950879603e-05, 'samples': 19378176, 'steps': 37847, 'loss/train': 0.7753595113754272} 02/26/2022 02:34:14 - INFO - codeparrot_training - Step 37848: {'lr': 7.499066459795258e-05, 'samples': 19378688, 'steps': 37848, 'loss/train': 0.8248132467269897} 02/26/2022 02:34:17 - INFO - codeparrot_training - Step 37849: {'lr': 7.49789804367936e-05, 'samples': 19379200, 'steps': 37849, 'loss/train': 1.2214226722717285} 02/26/2022 02:34:24 - INFO - codeparrot_training - Step 37850: {'lr': 7.496729702536912e-05, 'samples': 19379712, 'steps': 37850, 'loss/train': 1.7030028104782104} 02/26/2022 02:34:27 - INFO - codeparrot_training - Step 37851: {'lr': 7.495561436372916e-05, 'samples': 19380224, 'steps': 37851, 'loss/train': 2.2792766094207764} 02/26/2022 02:34:32 - INFO - codeparrot_training - Step 37852: {'lr': 7.494393245192366e-05, 'samples': 19380736, 'steps': 37852, 'loss/train': 2.000772714614868} 02/26/2022 02:34:36 - INFO - codeparrot_training - Step 37853: {'lr': 7.493225129000286e-05, 'samples': 19381248, 'steps': 37853, 'loss/train': 1.6239792108535767} 02/26/2022 02:34:41 - INFO - codeparrot_training - Step 37854: {'lr': 7.492057087801671e-05, 'samples': 19381760, 'steps': 37854, 'loss/train': 1.1903672218322754} 02/26/2022 02:34:45 - INFO - codeparrot_training - Step 37855: {'lr': 7.490889121601523e-05, 'samples': 19382272, 'steps': 37855, 'loss/train': 1.0262376070022583} 02/26/2022 02:34:50 - INFO - codeparrot_training - Step 37856: {'lr': 7.489721230404842e-05, 'samples': 19382784, 'steps': 37856, 'loss/train': 0.7008168697357178} 02/26/2022 02:34:54 - INFO - codeparrot_training - Step 37857: {'lr': 7.48855341421664e-05, 'samples': 19383296, 'steps': 37857, 'loss/train': 1.4924386739730835} 02/26/2022 02:34:59 - INFO - codeparrot_training - Step 37858: {'lr': 7.487385673041904e-05, 'samples': 19383808, 'steps': 37858, 'loss/train': 1.9464226961135864} 02/26/2022 02:35:03 - INFO - codeparrot_training - Step 37859: {'lr': 7.486218006885668e-05, 'samples': 19384320, 'steps': 37859, 'loss/train': 1.0982117652893066} 02/26/2022 02:35:08 - INFO - codeparrot_training - Step 37860: {'lr': 7.485050415752892e-05, 'samples': 19384832, 'steps': 37860, 'loss/train': 2.342677593231201} 02/26/2022 02:35:12 - INFO - codeparrot_training - Step 37861: {'lr': 7.483882899648611e-05, 'samples': 19385344, 'steps': 37861, 'loss/train': 1.7296215295791626} 02/26/2022 02:35:17 - INFO - codeparrot_training - Step 37862: {'lr': 7.482715458577802e-05, 'samples': 19385856, 'steps': 37862, 'loss/train': 1.3387377262115479} 02/26/2022 02:35:21 - INFO - codeparrot_training - Step 37863: {'lr': 7.481548092545496e-05, 'samples': 19386368, 'steps': 37863, 'loss/train': 2.5989415645599365} 02/26/2022 02:35:27 - INFO - codeparrot_training - Step 37864: {'lr': 7.480380801556657e-05, 'samples': 19386880, 'steps': 37864, 'loss/train': 2.5908684730529785} 02/26/2022 02:35:30 - INFO - codeparrot_training - Step 37865: {'lr': 7.479213585616315e-05, 'samples': 19387392, 'steps': 37865, 'loss/train': 2.1272919178009033} 02/26/2022 02:35:36 - INFO - codeparrot_training - Step 37866: {'lr': 7.478046444729456e-05, 'samples': 19387904, 'steps': 37866, 'loss/train': 1.7978525161743164} 02/26/2022 02:35:40 - INFO - codeparrot_training - Step 37867: {'lr': 7.476879378901075e-05, 'samples': 19388416, 'steps': 37867, 'loss/train': 3.0519626140594482} 02/26/2022 02:35:45 - INFO - codeparrot_training - Step 37868: {'lr': 7.475712388136185e-05, 'samples': 19388928, 'steps': 37868, 'loss/train': 1.9512221813201904} 02/26/2022 02:35:49 - INFO - codeparrot_training - Step 37869: {'lr': 7.474545472439781e-05, 'samples': 19389440, 'steps': 37869, 'loss/train': 0.9405859112739563} 02/26/2022 02:35:54 - INFO - codeparrot_training - Step 37870: {'lr': 7.473378631816857e-05, 'samples': 19389952, 'steps': 37870, 'loss/train': 3.8095788955688477} 02/26/2022 02:35:58 - INFO - codeparrot_training - Step 37871: {'lr': 7.472211866272405e-05, 'samples': 19390464, 'steps': 37871, 'loss/train': 2.5191195011138916} 02/26/2022 02:36:03 - INFO - codeparrot_training - Step 37872: {'lr': 7.471045175811442e-05, 'samples': 19390976, 'steps': 37872, 'loss/train': 1.3548946380615234} 02/26/2022 02:36:07 - INFO - codeparrot_training - Step 37873: {'lr': 7.469878560438956e-05, 'samples': 19391488, 'steps': 37873, 'loss/train': 1.8850998878479004} 02/26/2022 02:36:13 - INFO - codeparrot_training - Step 37874: {'lr': 7.46871202015994e-05, 'samples': 19392000, 'steps': 37874, 'loss/train': 1.2202781438827515} 02/26/2022 02:36:16 - INFO - codeparrot_training - Step 37875: {'lr': 7.46754555497939e-05, 'samples': 19392512, 'steps': 37875, 'loss/train': 1.0568088293075562} 02/26/2022 02:36:22 - INFO - codeparrot_training - Step 37876: {'lr': 7.466379164902315e-05, 'samples': 19393024, 'steps': 37876, 'loss/train': 1.5135680437088013} 02/26/2022 02:36:25 - INFO - codeparrot_training - Step 37877: {'lr': 7.465212849933706e-05, 'samples': 19393536, 'steps': 37877, 'loss/train': 2.227048397064209} 02/26/2022 02:36:31 - INFO - codeparrot_training - Step 37878: {'lr': 7.464046610078553e-05, 'samples': 19394048, 'steps': 37878, 'loss/train': 1.6113837957382202} 02/26/2022 02:36:35 - INFO - codeparrot_training - Step 37879: {'lr': 7.462880445341852e-05, 'samples': 19394560, 'steps': 37879, 'loss/train': 0.5117477774620056} 02/26/2022 02:36:40 - INFO - codeparrot_training - Step 37880: {'lr': 7.461714355728607e-05, 'samples': 19395072, 'steps': 37880, 'loss/train': 1.7949235439300537} 02/26/2022 02:36:44 - INFO - codeparrot_training - Step 37881: {'lr': 7.460548341243811e-05, 'samples': 19395584, 'steps': 37881, 'loss/train': 0.3060033917427063} 02/26/2022 02:36:49 - INFO - codeparrot_training - Step 37882: {'lr': 7.459382401892456e-05, 'samples': 19396096, 'steps': 37882, 'loss/train': 0.9207753539085388} 02/26/2022 02:36:52 - INFO - codeparrot_training - Step 37883: {'lr': 7.458216537679529e-05, 'samples': 19396608, 'steps': 37883, 'loss/train': 2.17743182182312} 02/26/2022 02:36:58 - INFO - codeparrot_training - Step 37884: {'lr': 7.45705074861004e-05, 'samples': 19397120, 'steps': 37884, 'loss/train': 2.0331780910491943} 02/26/2022 02:37:01 - INFO - codeparrot_training - Step 37885: {'lr': 7.455885034688974e-05, 'samples': 19397632, 'steps': 37885, 'loss/train': 1.174649715423584} 02/26/2022 02:37:08 - INFO - codeparrot_training - Step 37886: {'lr': 7.454719395921328e-05, 'samples': 19398144, 'steps': 37886, 'loss/train': 0.8099073171615601} 02/26/2022 02:37:11 - INFO - codeparrot_training - Step 37887: {'lr': 7.453553832312085e-05, 'samples': 19398656, 'steps': 37887, 'loss/train': 1.5410692691802979} 02/26/2022 02:37:17 - INFO - codeparrot_training - Step 37888: {'lr': 7.452388343866251e-05, 'samples': 19399168, 'steps': 37888, 'loss/train': 1.5804073810577393} 02/26/2022 02:37:20 - INFO - codeparrot_training - Step 37889: {'lr': 7.451222930588817e-05, 'samples': 19399680, 'steps': 37889, 'loss/train': 0.32616671919822693} 02/26/2022 02:37:26 - INFO - codeparrot_training - Step 37890: {'lr': 7.450057592484768e-05, 'samples': 19400192, 'steps': 37890, 'loss/train': 2.31925630569458} 02/26/2022 02:37:29 - INFO - codeparrot_training - Step 37891: {'lr': 7.448892329559092e-05, 'samples': 19400704, 'steps': 37891, 'loss/train': 0.8049443364143372} 02/26/2022 02:37:35 - INFO - codeparrot_training - Step 37892: {'lr': 7.447727141816798e-05, 'samples': 19401216, 'steps': 37892, 'loss/train': 1.657503604888916} 02/26/2022 02:37:38 - INFO - codeparrot_training - Step 37893: {'lr': 7.44656202926286e-05, 'samples': 19401728, 'steps': 37893, 'loss/train': 1.398797869682312} 02/26/2022 02:37:44 - INFO - codeparrot_training - Step 37894: {'lr': 7.445396991902289e-05, 'samples': 19402240, 'steps': 37894, 'loss/train': 0.8342295289039612} 02/26/2022 02:37:50 - INFO - codeparrot_training - Step 37895: {'lr': 7.44423202974005e-05, 'samples': 19402752, 'steps': 37895, 'loss/train': 1.3070570230484009} 02/26/2022 02:37:53 - INFO - codeparrot_training - Step 37896: {'lr': 7.443067142781155e-05, 'samples': 19403264, 'steps': 37896, 'loss/train': 0.8890364766120911} 02/26/2022 02:37:59 - INFO - codeparrot_training - Step 37897: {'lr': 7.441902331030576e-05, 'samples': 19403776, 'steps': 37897, 'loss/train': 3.535931348800659} 02/26/2022 02:38:02 - INFO - codeparrot_training - Step 37898: {'lr': 7.440737594493327e-05, 'samples': 19404288, 'steps': 37898, 'loss/train': 1.6163150072097778} 02/26/2022 02:38:06 - INFO - codeparrot_training - Step 37899: {'lr': 7.439572933174366e-05, 'samples': 19404800, 'steps': 37899, 'loss/train': 2.1310107707977295} 02/26/2022 02:38:11 - INFO - codeparrot_training - Step 37900: {'lr': 7.43840834707871e-05, 'samples': 19405312, 'steps': 37900, 'loss/train': 0.8816977739334106} 02/26/2022 02:38:17 - INFO - codeparrot_training - Step 37901: {'lr': 7.437243836211324e-05, 'samples': 19405824, 'steps': 37901, 'loss/train': 2.5828487873077393} 02/26/2022 02:38:21 - INFO - codeparrot_training - Step 37902: {'lr': 7.436079400577228e-05, 'samples': 19406336, 'steps': 37902, 'loss/train': 1.9444694519042969} 02/26/2022 02:38:24 - INFO - codeparrot_training - Step 37903: {'lr': 7.434915040181373e-05, 'samples': 19406848, 'steps': 37903, 'loss/train': 1.9368579387664795} 02/26/2022 02:38:30 - INFO - codeparrot_training - Step 37904: {'lr': 7.433750755028773e-05, 'samples': 19407360, 'steps': 37904, 'loss/train': 1.4657213687896729} 02/26/2022 02:38:33 - INFO - codeparrot_training - Step 37905: {'lr': 7.432586545124399e-05, 'samples': 19407872, 'steps': 37905, 'loss/train': 2.437809705734253} 02/26/2022 02:38:38 - INFO - codeparrot_training - Step 37906: {'lr': 7.431422410473254e-05, 'samples': 19408384, 'steps': 37906, 'loss/train': 2.106069803237915} 02/26/2022 02:38:42 - INFO - codeparrot_training - Step 37907: {'lr': 7.430258351080316e-05, 'samples': 19408896, 'steps': 37907, 'loss/train': 0.8649512529373169} 02/26/2022 02:38:47 - INFO - codeparrot_training - Step 37908: {'lr': 7.429094366950573e-05, 'samples': 19409408, 'steps': 37908, 'loss/train': 0.7029323577880859} 02/26/2022 02:38:51 - INFO - codeparrot_training - Step 37909: {'lr': 7.427930458089002e-05, 'samples': 19409920, 'steps': 37909, 'loss/train': 2.547236680984497} 02/26/2022 02:38:56 - INFO - codeparrot_training - Step 37910: {'lr': 7.426766624500602e-05, 'samples': 19410432, 'steps': 37910, 'loss/train': 1.298637866973877} 02/26/2022 02:39:00 - INFO - codeparrot_training - Step 37911: {'lr': 7.42560286619036e-05, 'samples': 19410944, 'steps': 37911, 'loss/train': 3.5130388736724854} 02/26/2022 02:39:07 - INFO - codeparrot_training - Step 37912: {'lr': 7.42443918316325e-05, 'samples': 19411456, 'steps': 37912, 'loss/train': 1.2513548135757446} 02/26/2022 02:39:10 - INFO - codeparrot_training - Step 37913: {'lr': 7.423275575424262e-05, 'samples': 19411968, 'steps': 37913, 'loss/train': 1.935681700706482} 02/26/2022 02:39:16 - INFO - codeparrot_training - Step 37914: {'lr': 7.422112042978373e-05, 'samples': 19412480, 'steps': 37914, 'loss/train': 0.6941214799880981} 02/26/2022 02:39:19 - INFO - codeparrot_training - Step 37915: {'lr': 7.420948585830584e-05, 'samples': 19412992, 'steps': 37915, 'loss/train': 2.090805768966675} 02/26/2022 02:39:25 - INFO - codeparrot_training - Step 37916: {'lr': 7.419785203985868e-05, 'samples': 19413504, 'steps': 37916, 'loss/train': 3.969656229019165} 02/26/2022 02:39:28 - INFO - codeparrot_training - Step 37917: {'lr': 7.41862189744921e-05, 'samples': 19414016, 'steps': 37917, 'loss/train': 0.5874213576316833} 02/26/2022 02:39:34 - INFO - codeparrot_training - Step 37918: {'lr': 7.417458666225586e-05, 'samples': 19414528, 'steps': 37918, 'loss/train': 1.4047911167144775} 02/26/2022 02:39:37 - INFO - codeparrot_training - Step 37919: {'lr': 7.416295510319992e-05, 'samples': 19415040, 'steps': 37919, 'loss/train': 1.738824725151062} 02/26/2022 02:39:43 - INFO - codeparrot_training - Step 37920: {'lr': 7.415132429737406e-05, 'samples': 19415552, 'steps': 37920, 'loss/train': 2.721315622329712} 02/26/2022 02:39:46 - INFO - codeparrot_training - Step 37921: {'lr': 7.413969424482808e-05, 'samples': 19416064, 'steps': 37921, 'loss/train': 1.6940052509307861} 02/26/2022 02:39:53 - INFO - codeparrot_training - Step 37922: {'lr': 7.412806494561175e-05, 'samples': 19416576, 'steps': 37922, 'loss/train': 0.5788756012916565} 02/26/2022 02:39:56 - INFO - codeparrot_training - Step 37923: {'lr': 7.4116436399775e-05, 'samples': 19417088, 'steps': 37923, 'loss/train': 1.7050844430923462} 02/26/2022 02:40:02 - INFO - codeparrot_training - Step 37924: {'lr': 7.41048086073676e-05, 'samples': 19417600, 'steps': 37924, 'loss/train': 1.9376347064971924} 02/26/2022 02:40:05 - INFO - codeparrot_training - Step 37925: {'lr': 7.409318156843933e-05, 'samples': 19418112, 'steps': 37925, 'loss/train': 1.453546166419983} 02/26/2022 02:40:11 - INFO - codeparrot_training - Step 37926: {'lr': 7.408155528303995e-05, 'samples': 19418624, 'steps': 37926, 'loss/train': 1.5567615032196045} 02/26/2022 02:40:14 - INFO - codeparrot_training - Step 37927: {'lr': 7.40699297512194e-05, 'samples': 19419136, 'steps': 37927, 'loss/train': 1.801559567451477} 02/26/2022 02:40:20 - INFO - codeparrot_training - Step 37928: {'lr': 7.405830497302732e-05, 'samples': 19419648, 'steps': 37928, 'loss/train': 1.6155083179473877} 02/26/2022 02:40:24 - INFO - codeparrot_training - Step 37929: {'lr': 7.404668094851378e-05, 'samples': 19420160, 'steps': 37929, 'loss/train': 1.728258490562439} 02/26/2022 02:40:29 - INFO - codeparrot_training - Step 37930: {'lr': 7.40350576777282e-05, 'samples': 19420672, 'steps': 37930, 'loss/train': 1.5946191549301147} 02/26/2022 02:40:33 - INFO - codeparrot_training - Step 37931: {'lr': 7.402343516072063e-05, 'samples': 19421184, 'steps': 37931, 'loss/train': 1.9840030670166016} 02/26/2022 02:40:39 - INFO - codeparrot_training - Step 37932: {'lr': 7.401181339754074e-05, 'samples': 19421696, 'steps': 37932, 'loss/train': 0.9495724439620972} 02/26/2022 02:40:42 - INFO - codeparrot_training - Step 37933: {'lr': 7.400019238823849e-05, 'samples': 19422208, 'steps': 37933, 'loss/train': 1.8813179731369019} 02/26/2022 02:40:48 - INFO - codeparrot_training - Step 37934: {'lr': 7.39885721328634e-05, 'samples': 19422720, 'steps': 37934, 'loss/train': 0.4933006465435028} 02/26/2022 02:40:51 - INFO - codeparrot_training - Step 37935: {'lr': 7.397695263146542e-05, 'samples': 19423232, 'steps': 37935, 'loss/train': 2.935905933380127} 02/26/2022 02:40:57 - INFO - codeparrot_training - Step 37936: {'lr': 7.396533388409421e-05, 'samples': 19423744, 'steps': 37936, 'loss/train': 1.4275503158569336} 02/26/2022 02:41:00 - INFO - codeparrot_training - Step 37937: {'lr': 7.395371589079977e-05, 'samples': 19424256, 'steps': 37937, 'loss/train': 2.7695472240448} 02/26/2022 02:41:06 - INFO - codeparrot_training - Step 37938: {'lr': 7.394209865163157e-05, 'samples': 19424768, 'steps': 37938, 'loss/train': 2.240042209625244} 02/26/2022 02:41:09 - INFO - codeparrot_training - Step 37939: {'lr': 7.393048216663959e-05, 'samples': 19425280, 'steps': 37939, 'loss/train': 0.4156077802181244} 02/26/2022 02:41:14 - INFO - codeparrot_training - Step 37940: {'lr': 7.391886643587343e-05, 'samples': 19425792, 'steps': 37940, 'loss/train': 1.6348804235458374} 02/26/2022 02:41:21 - INFO - codeparrot_training - Step 37941: {'lr': 7.390725145938302e-05, 'samples': 19426304, 'steps': 37941, 'loss/train': 2.4212088584899902} 02/26/2022 02:41:24 - INFO - codeparrot_training - Step 37942: {'lr': 7.389563723721801e-05, 'samples': 19426816, 'steps': 37942, 'loss/train': 1.5889395475387573} 02/26/2022 02:41:29 - INFO - codeparrot_training - Step 37943: {'lr': 7.388402376942818e-05, 'samples': 19427328, 'steps': 37943, 'loss/train': 1.5325666666030884} 02/26/2022 02:41:33 - INFO - codeparrot_training - Step 37944: {'lr': 7.387241105606321e-05, 'samples': 19427840, 'steps': 37944, 'loss/train': 1.6129833459854126} 02/26/2022 02:41:38 - INFO - codeparrot_training - Step 37945: {'lr': 7.386079909717297e-05, 'samples': 19428352, 'steps': 37945, 'loss/train': 2.318693161010742} 02/26/2022 02:41:42 - INFO - codeparrot_training - Step 37946: {'lr': 7.384918789280715e-05, 'samples': 19428864, 'steps': 37946, 'loss/train': 1.3313490152359009} 02/26/2022 02:41:48 - INFO - codeparrot_training - Step 37947: {'lr': 7.383757744301545e-05, 'samples': 19429376, 'steps': 37947, 'loss/train': 1.7612372636795044} 02/26/2022 02:41:51 - INFO - codeparrot_training - Step 37948: {'lr': 7.382596774784756e-05, 'samples': 19429888, 'steps': 37948, 'loss/train': 2.1369247436523438} 02/26/2022 02:41:57 - INFO - codeparrot_training - Step 37949: {'lr': 7.381435880735334e-05, 'samples': 19430400, 'steps': 37949, 'loss/train': 1.8385356664657593} 02/26/2022 02:42:00 - INFO - codeparrot_training - Step 37950: {'lr': 7.38027506215825e-05, 'samples': 19430912, 'steps': 37950, 'loss/train': 2.474994421005249} 02/26/2022 02:42:06 - INFO - codeparrot_training - Step 37951: {'lr': 7.379114319058471e-05, 'samples': 19431424, 'steps': 37951, 'loss/train': 0.9505074620246887} 02/26/2022 02:42:09 - INFO - codeparrot_training - Step 37952: {'lr': 7.377953651440964e-05, 'samples': 19431936, 'steps': 37952, 'loss/train': 1.0113667249679565} 02/26/2022 02:42:15 - INFO - codeparrot_training - Step 37953: {'lr': 7.376793059310713e-05, 'samples': 19432448, 'steps': 37953, 'loss/train': 1.8876501321792603} 02/26/2022 02:42:18 - INFO - codeparrot_training - Step 37954: {'lr': 7.375632542672688e-05, 'samples': 19432960, 'steps': 37954, 'loss/train': 2.267726421356201} 02/26/2022 02:42:24 - INFO - codeparrot_training - Step 37955: {'lr': 7.374472101531854e-05, 'samples': 19433472, 'steps': 37955, 'loss/train': 1.8480150699615479} 02/26/2022 02:42:28 - INFO - codeparrot_training - Step 37956: {'lr': 7.373311735893178e-05, 'samples': 19433984, 'steps': 37956, 'loss/train': 2.2423174381256104} 02/26/2022 02:42:34 - INFO - codeparrot_training - Step 37957: {'lr': 7.372151445761646e-05, 'samples': 19434496, 'steps': 37957, 'loss/train': 0.9533421993255615} 02/26/2022 02:42:37 - INFO - codeparrot_training - Step 37958: {'lr': 7.370991231142222e-05, 'samples': 19435008, 'steps': 37958, 'loss/train': 1.7286713123321533} 02/26/2022 02:42:41 - INFO - codeparrot_training - Step 37959: {'lr': 7.369831092039869e-05, 'samples': 19435520, 'steps': 37959, 'loss/train': 0.22948433458805084} 02/26/2022 02:42:46 - INFO - codeparrot_training - Step 37960: {'lr': 7.368671028459564e-05, 'samples': 19436032, 'steps': 37960, 'loss/train': 2.0543406009674072} 02/26/2022 02:42:52 - INFO - codeparrot_training - Step 37961: {'lr': 7.367511040406266e-05, 'samples': 19436544, 'steps': 37961, 'loss/train': 1.7658185958862305} 02/26/2022 02:42:55 - INFO - codeparrot_training - Step 37962: {'lr': 7.366351127884957e-05, 'samples': 19437056, 'steps': 37962, 'loss/train': 0.878489077091217} 02/26/2022 02:42:59 - INFO - codeparrot_training - Step 37963: {'lr': 7.365191290900603e-05, 'samples': 19437568, 'steps': 37963, 'loss/train': 2.6228792667388916} 02/26/2022 02:43:04 - INFO - codeparrot_training - Step 37964: {'lr': 7.364031529458171e-05, 'samples': 19438080, 'steps': 37964, 'loss/train': 1.7497022151947021} 02/26/2022 02:43:08 - INFO - codeparrot_training - Step 37965: {'lr': 7.362871843562616e-05, 'samples': 19438592, 'steps': 37965, 'loss/train': 2.4399523735046387} 02/26/2022 02:43:13 - INFO - codeparrot_training - Step 37966: {'lr': 7.361712233218928e-05, 'samples': 19439104, 'steps': 37966, 'loss/train': 1.2031710147857666} 02/26/2022 02:43:17 - INFO - codeparrot_training - Step 37967: {'lr': 7.360552698432055e-05, 'samples': 19439616, 'steps': 37967, 'loss/train': 0.023997321724891663} 02/26/2022 02:43:23 - INFO - codeparrot_training - Step 37968: {'lr': 7.359393239206991e-05, 'samples': 19440128, 'steps': 37968, 'loss/train': 2.0471606254577637} 02/26/2022 02:43:27 - INFO - codeparrot_training - Step 37969: {'lr': 7.358233855548668e-05, 'samples': 19440640, 'steps': 37969, 'loss/train': 2.9501149654388428} 02/26/2022 02:43:32 - INFO - codeparrot_training - Step 37970: {'lr': 7.357074547462076e-05, 'samples': 19441152, 'steps': 37970, 'loss/train': 1.5002484321594238} 02/26/2022 02:43:36 - INFO - codeparrot_training - Step 37971: {'lr': 7.355915314952169e-05, 'samples': 19441664, 'steps': 37971, 'loss/train': 1.643397569656372} 02/26/2022 02:43:42 - INFO - codeparrot_training - Step 37972: {'lr': 7.354756158023931e-05, 'samples': 19442176, 'steps': 37972, 'loss/train': 0.3950980007648468} 02/26/2022 02:43:45 - INFO - codeparrot_training - Step 37973: {'lr': 7.3535970766823e-05, 'samples': 19442688, 'steps': 37973, 'loss/train': 1.9400101900100708} 02/26/2022 02:43:51 - INFO - codeparrot_training - Step 37974: {'lr': 7.352438070932262e-05, 'samples': 19443200, 'steps': 37974, 'loss/train': 2.783689260482788} 02/26/2022 02:43:54 - INFO - codeparrot_training - Step 37975: {'lr': 7.351279140778772e-05, 'samples': 19443712, 'steps': 37975, 'loss/train': 1.078197956085205} 02/26/2022 02:43:59 - INFO - codeparrot_training - Step 37976: {'lr': 7.350120286226803e-05, 'samples': 19444224, 'steps': 37976, 'loss/train': 2.2455835342407227} 02/26/2022 02:44:03 - INFO - codeparrot_training - Step 37977: {'lr': 7.348961507281316e-05, 'samples': 19444736, 'steps': 37977, 'loss/train': 1.970812201499939} 02/26/2022 02:44:09 - INFO - codeparrot_training - Step 37978: {'lr': 7.347802803947271e-05, 'samples': 19445248, 'steps': 37978, 'loss/train': 1.2487633228302002} 02/26/2022 02:44:13 - INFO - codeparrot_training - Step 37979: {'lr': 7.346644176229625e-05, 'samples': 19445760, 'steps': 37979, 'loss/train': 1.3906623125076294} 02/26/2022 02:44:18 - INFO - codeparrot_training - Step 37980: {'lr': 7.34548562413336e-05, 'samples': 19446272, 'steps': 37980, 'loss/train': 0.7225399017333984} 02/26/2022 02:44:24 - INFO - codeparrot_training - Step 37981: {'lr': 7.344327147663427e-05, 'samples': 19446784, 'steps': 37981, 'loss/train': 2.2374472618103027} 02/26/2022 02:44:27 - INFO - codeparrot_training - Step 37982: {'lr': 7.34316874682479e-05, 'samples': 19447296, 'steps': 37982, 'loss/train': 2.211420774459839} 02/26/2022 02:44:33 - INFO - codeparrot_training - Step 37983: {'lr': 7.342010421622405e-05, 'samples': 19447808, 'steps': 37983, 'loss/train': 1.4731392860412598} 02/26/2022 02:44:36 - INFO - codeparrot_training - Step 37984: {'lr': 7.34085217206125e-05, 'samples': 19448320, 'steps': 37984, 'loss/train': 1.5622638463974} 02/26/2022 02:44:41 - INFO - codeparrot_training - Step 37985: {'lr': 7.339693998146274e-05, 'samples': 19448832, 'steps': 37985, 'loss/train': 1.8136826753616333} 02/26/2022 02:44:45 - INFO - codeparrot_training - Step 37986: {'lr': 7.338535899882439e-05, 'samples': 19449344, 'steps': 37986, 'loss/train': 2.1253254413604736} 02/26/2022 02:44:51 - INFO - codeparrot_training - Step 37987: {'lr': 7.337377877274703e-05, 'samples': 19449856, 'steps': 37987, 'loss/train': 0.3222852349281311} 02/26/2022 02:44:54 - INFO - codeparrot_training - Step 37988: {'lr': 7.336219930328042e-05, 'samples': 19450368, 'steps': 37988, 'loss/train': 0.4808877110481262} 02/26/2022 02:45:00 - INFO - codeparrot_training - Step 37989: {'lr': 7.335062059047396e-05, 'samples': 19450880, 'steps': 37989, 'loss/train': 0.1336582601070404} 02/26/2022 02:45:04 - INFO - codeparrot_training - Step 37990: {'lr': 7.333904263437748e-05, 'samples': 19451392, 'steps': 37990, 'loss/train': 2.6639249324798584} 02/26/2022 02:45:09 - INFO - codeparrot_training - Step 37991: {'lr': 7.332746543504032e-05, 'samples': 19451904, 'steps': 37991, 'loss/train': 1.6894506216049194} 02/26/2022 02:45:13 - INFO - codeparrot_training - Step 37992: {'lr': 7.331588899251227e-05, 'samples': 19452416, 'steps': 37992, 'loss/train': 2.1091763973236084} 02/26/2022 02:45:18 - INFO - codeparrot_training - Step 37993: {'lr': 7.330431330684285e-05, 'samples': 19452928, 'steps': 37993, 'loss/train': 2.9126014709472656} 02/26/2022 02:45:22 - INFO - codeparrot_training - Step 37994: {'lr': 7.329273837808162e-05, 'samples': 19453440, 'steps': 37994, 'loss/train': 2.5056447982788086} 02/26/2022 02:45:27 - INFO - codeparrot_training - Step 37995: {'lr': 7.328116420627811e-05, 'samples': 19453952, 'steps': 37995, 'loss/train': 0.940467119216919} 02/26/2022 02:45:30 - INFO - codeparrot_training - Step 37996: {'lr': 7.32695907914821e-05, 'samples': 19454464, 'steps': 37996, 'loss/train': 0.4037728011608124} 02/26/2022 02:45:36 - INFO - codeparrot_training - Step 37997: {'lr': 7.3258018133743e-05, 'samples': 19454976, 'steps': 37997, 'loss/train': 0.9487450122833252} 02/26/2022 02:45:39 - INFO - codeparrot_training - Step 37998: {'lr': 7.324644623311046e-05, 'samples': 19455488, 'steps': 37998, 'loss/train': 0.07264846563339233} 02/26/2022 02:45:45 - INFO - codeparrot_training - Step 37999: {'lr': 7.323487508963392e-05, 'samples': 19456000, 'steps': 37999, 'loss/train': 1.960349678993225} 02/26/2022 02:45:45 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 02:46:03 - WARNING - huggingface_hub.repository - Several commits (38) will be pushed upstream. 02/26/2022 02:46:03 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 02:46:37 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 333cdce..e53c3e1 floral-grass-11 -> floral-grass-11 02/26/2022 02:46:42 - INFO - codeparrot_training - Step 38000: {'lr': 7.322330470336314e-05, 'samples': 19456512, 'steps': 38000, 'loss/train': 1.6377456188201904} 02/26/2022 02:46:48 - INFO - codeparrot_training - Step 38001: {'lr': 7.321173507434758e-05, 'samples': 19457024, 'steps': 38001, 'loss/train': 0.6367076635360718} 02/26/2022 02:46:51 - INFO - codeparrot_training - Step 38002: {'lr': 7.320016620263673e-05, 'samples': 19457536, 'steps': 38002, 'loss/train': 2.099254608154297} 02/26/2022 02:46:57 - INFO - codeparrot_training - Step 38003: {'lr': 7.318859808828032e-05, 'samples': 19458048, 'steps': 38003, 'loss/train': 1.628173589706421} 02/26/2022 02:47:00 - INFO - codeparrot_training - Step 38004: {'lr': 7.317703073132778e-05, 'samples': 19458560, 'steps': 38004, 'loss/train': 2.3905887603759766} 02/26/2022 02:47:06 - INFO - codeparrot_training - Step 38005: {'lr': 7.316546413182874e-05, 'samples': 19459072, 'steps': 38005, 'loss/train': 0.27225345373153687} 02/26/2022 02:47:09 - INFO - codeparrot_training - Step 38006: {'lr': 7.315389828983257e-05, 'samples': 19459584, 'steps': 38006, 'loss/train': 0.3845939636230469} 02/26/2022 02:47:15 - INFO - codeparrot_training - Step 38007: {'lr': 7.314233320538907e-05, 'samples': 19460096, 'steps': 38007, 'loss/train': 0.7665659785270691} 02/26/2022 02:47:19 - INFO - codeparrot_training - Step 38008: {'lr': 7.313076887854764e-05, 'samples': 19460608, 'steps': 38008, 'loss/train': 1.9438731670379639} 02/26/2022 02:47:24 - INFO - codeparrot_training - Step 38009: {'lr': 7.311920530935781e-05, 'samples': 19461120, 'steps': 38009, 'loss/train': 1.8758307695388794} 02/26/2022 02:47:28 - INFO - codeparrot_training - Step 38010: {'lr': 7.310764249786908e-05, 'samples': 19461632, 'steps': 38010, 'loss/train': 1.646323561668396} 02/26/2022 02:47:33 - INFO - codeparrot_training - Step 38011: {'lr': 7.309608044413121e-05, 'samples': 19462144, 'steps': 38011, 'loss/train': 2.3616750240325928} 02/26/2022 02:47:37 - INFO - codeparrot_training - Step 38012: {'lr': 7.30845191481934e-05, 'samples': 19462656, 'steps': 38012, 'loss/train': 1.0274697542190552} 02/26/2022 02:47:43 - INFO - codeparrot_training - Step 38013: {'lr': 7.307295861010538e-05, 'samples': 19463168, 'steps': 38013, 'loss/train': 0.612236738204956} 02/26/2022 02:47:46 - INFO - codeparrot_training - Step 38014: {'lr': 7.306139882991656e-05, 'samples': 19463680, 'steps': 38014, 'loss/train': 0.9534772038459778} 02/26/2022 02:47:52 - INFO - codeparrot_training - Step 38015: {'lr': 7.30498398076766e-05, 'samples': 19464192, 'steps': 38015, 'loss/train': 1.9720512628555298} 02/26/2022 02:47:55 - INFO - codeparrot_training - Step 38016: {'lr': 7.303828154343494e-05, 'samples': 19464704, 'steps': 38016, 'loss/train': 0.32893434166908264} 02/26/2022 02:48:01 - INFO - codeparrot_training - Step 38017: {'lr': 7.302672403724112e-05, 'samples': 19465216, 'steps': 38017, 'loss/train': 1.5718871355056763} 02/26/2022 02:48:04 - INFO - codeparrot_training - Step 38018: {'lr': 7.301516728914451e-05, 'samples': 19465728, 'steps': 38018, 'loss/train': 2.3168623447418213} 02/26/2022 02:48:10 - INFO - codeparrot_training - Step 38019: {'lr': 7.300361129919481e-05, 'samples': 19466240, 'steps': 38019, 'loss/train': 1.5415875911712646} 02/26/2022 02:48:13 - INFO - codeparrot_training - Step 38020: {'lr': 7.299205606744142e-05, 'samples': 19466752, 'steps': 38020, 'loss/train': 1.5314196348190308} 02/26/2022 02:48:19 - INFO - codeparrot_training - Step 38021: {'lr': 7.298050159393388e-05, 'samples': 19467264, 'steps': 38021, 'loss/train': 1.9744246006011963} 02/26/2022 02:48:22 - INFO - codeparrot_training - Step 38022: {'lr': 7.296894787872158e-05, 'samples': 19467776, 'steps': 38022, 'loss/train': 1.652613639831543} 02/26/2022 02:48:28 - INFO - codeparrot_training - Step 38023: {'lr': 7.295739492185416e-05, 'samples': 19468288, 'steps': 38023, 'loss/train': 2.4441065788269043} 02/26/2022 02:48:32 - INFO - codeparrot_training - Step 38024: {'lr': 7.294584272338103e-05, 'samples': 19468800, 'steps': 38024, 'loss/train': 1.5685055255889893} 02/26/2022 02:48:37 - INFO - codeparrot_training - Step 38025: {'lr': 7.293429128335172e-05, 'samples': 19469312, 'steps': 38025, 'loss/train': 2.0573067665100098} 02/26/2022 02:48:41 - INFO - codeparrot_training - Step 38026: {'lr': 7.292274060181558e-05, 'samples': 19469824, 'steps': 38026, 'loss/train': 2.105271339416504} 02/26/2022 02:48:46 - INFO - codeparrot_training - Step 38027: {'lr': 7.29111906788223e-05, 'samples': 19470336, 'steps': 38027, 'loss/train': 1.6482328176498413} 02/26/2022 02:48:50 - INFO - codeparrot_training - Step 38028: {'lr': 7.289964151442113e-05, 'samples': 19470848, 'steps': 38028, 'loss/train': 0.9010985493659973} 02/26/2022 02:48:55 - INFO - codeparrot_training - Step 38029: {'lr': 7.288809310866182e-05, 'samples': 19471360, 'steps': 38029, 'loss/train': 2.1914548873901367} 02/26/2022 02:48:59 - INFO - codeparrot_training - Step 38030: {'lr': 7.287654546159356e-05, 'samples': 19471872, 'steps': 38030, 'loss/train': 1.8573434352874756} 02/26/2022 02:49:04 - INFO - codeparrot_training - Step 38031: {'lr': 7.286499857326597e-05, 'samples': 19472384, 'steps': 38031, 'loss/train': 1.8783557415008545} 02/26/2022 02:49:08 - INFO - codeparrot_training - Step 38032: {'lr': 7.285345244372843e-05, 'samples': 19472896, 'steps': 38032, 'loss/train': 2.015120029449463} 02/26/2022 02:49:13 - INFO - codeparrot_training - Step 38033: {'lr': 7.28419070730306e-05, 'samples': 19473408, 'steps': 38033, 'loss/train': 2.113590955734253} 02/26/2022 02:49:16 - INFO - codeparrot_training - Step 38034: {'lr': 7.283036246122163e-05, 'samples': 19473920, 'steps': 38034, 'loss/train': 2.720353364944458} 02/26/2022 02:49:22 - INFO - codeparrot_training - Step 38035: {'lr': 7.281881860835121e-05, 'samples': 19474432, 'steps': 38035, 'loss/train': 1.833739995956421} 02/26/2022 02:49:26 - INFO - codeparrot_training - Step 38036: {'lr': 7.280727551446862e-05, 'samples': 19474944, 'steps': 38036, 'loss/train': 1.8961317539215088} 02/26/2022 02:49:32 - INFO - codeparrot_training - Step 38037: {'lr': 7.279573317962348e-05, 'samples': 19475456, 'steps': 38037, 'loss/train': 1.1372878551483154} 02/26/2022 02:49:35 - INFO - codeparrot_training - Step 38038: {'lr': 7.278419160386515e-05, 'samples': 19475968, 'steps': 38038, 'loss/train': 1.3634076118469238} 02/26/2022 02:49:41 - INFO - codeparrot_training - Step 38039: {'lr': 7.277265078724308e-05, 'samples': 19476480, 'steps': 38039, 'loss/train': 0.1734483689069748} 02/26/2022 02:49:45 - INFO - codeparrot_training - Step 38040: {'lr': 7.276111072980662e-05, 'samples': 19476992, 'steps': 38040, 'loss/train': 0.5034950971603394} 02/26/2022 02:49:50 - INFO - codeparrot_training - Step 38041: {'lr': 7.274957143160535e-05, 'samples': 19477504, 'steps': 38041, 'loss/train': 1.6803172826766968} 02/26/2022 02:49:54 - INFO - codeparrot_training - Step 38042: {'lr': 7.273803289268862e-05, 'samples': 19478016, 'steps': 38042, 'loss/train': 1.9188061952590942} 02/26/2022 02:49:59 - INFO - codeparrot_training - Step 38043: {'lr': 7.272649511310589e-05, 'samples': 19478528, 'steps': 38043, 'loss/train': 1.6629756689071655} 02/26/2022 02:50:03 - INFO - codeparrot_training - Step 38044: {'lr': 7.271495809290657e-05, 'samples': 19479040, 'steps': 38044, 'loss/train': 1.9798567295074463} 02/26/2022 02:50:08 - INFO - codeparrot_training - Step 38045: {'lr': 7.270342183213999e-05, 'samples': 19479552, 'steps': 38045, 'loss/train': 0.5005486011505127} 02/26/2022 02:50:12 - INFO - codeparrot_training - Step 38046: {'lr': 7.269188633085571e-05, 'samples': 19480064, 'steps': 38046, 'loss/train': 1.2825769186019897} 02/26/2022 02:50:18 - INFO - codeparrot_training - Step 38047: {'lr': 7.26803515891031e-05, 'samples': 19480576, 'steps': 38047, 'loss/train': 1.5931419134140015} 02/26/2022 02:50:21 - INFO - codeparrot_training - Step 38048: {'lr': 7.266881760693158e-05, 'samples': 19481088, 'steps': 38048, 'loss/train': 0.5860764980316162} 02/26/2022 02:50:27 - INFO - codeparrot_training - Step 38049: {'lr': 7.265728438439046e-05, 'samples': 19481600, 'steps': 38049, 'loss/train': 2.6361136436462402} 02/26/2022 02:50:31 - INFO - codeparrot_training - Step 38050: {'lr': 7.264575192152925e-05, 'samples': 19482112, 'steps': 38050, 'loss/train': 2.3184337615966797} 02/26/2022 02:50:36 - INFO - codeparrot_training - Step 38051: {'lr': 7.263422021839736e-05, 'samples': 19482624, 'steps': 38051, 'loss/train': 1.7024630308151245} 02/26/2022 02:50:40 - INFO - codeparrot_training - Step 38052: {'lr': 7.262268927504414e-05, 'samples': 19483136, 'steps': 38052, 'loss/train': 1.4867494106292725} 02/26/2022 02:50:45 - INFO - codeparrot_training - Step 38053: {'lr': 7.261115909151894e-05, 'samples': 19483648, 'steps': 38053, 'loss/train': 0.991300106048584} 02/26/2022 02:50:49 - INFO - codeparrot_training - Step 38054: {'lr': 7.259962966787126e-05, 'samples': 19484160, 'steps': 38054, 'loss/train': 1.9065449237823486} 02/26/2022 02:50:55 - INFO - codeparrot_training - Step 38055: {'lr': 7.258810100415044e-05, 'samples': 19484672, 'steps': 38055, 'loss/train': 1.3514381647109985} 02/26/2022 02:50:58 - INFO - codeparrot_training - Step 38056: {'lr': 7.257657310040585e-05, 'samples': 19485184, 'steps': 38056, 'loss/train': 0.7219937443733215} 02/26/2022 02:51:04 - INFO - codeparrot_training - Step 38057: {'lr': 7.256504595668684e-05, 'samples': 19485696, 'steps': 38057, 'loss/train': 1.651576042175293} 02/26/2022 02:51:07 - INFO - codeparrot_training - Step 38058: {'lr': 7.255351957304287e-05, 'samples': 19486208, 'steps': 38058, 'loss/train': 2.153428554534912} 02/26/2022 02:51:13 - INFO - codeparrot_training - Step 38059: {'lr': 7.25419939495233e-05, 'samples': 19486720, 'steps': 38059, 'loss/train': 2.180460214614868} 02/26/2022 02:51:17 - INFO - codeparrot_training - Step 38060: {'lr': 7.253046908617747e-05, 'samples': 19487232, 'steps': 38060, 'loss/train': 1.2783384323120117} 02/26/2022 02:51:22 - INFO - codeparrot_training - Step 38061: {'lr': 7.25189449830547e-05, 'samples': 19487744, 'steps': 38061, 'loss/train': 1.4480818510055542} 02/26/2022 02:51:26 - INFO - codeparrot_training - Step 38062: {'lr': 7.250742164020446e-05, 'samples': 19488256, 'steps': 38062, 'loss/train': 1.1692036390304565} 02/26/2022 02:51:31 - INFO - codeparrot_training - Step 38063: {'lr': 7.249589905767601e-05, 'samples': 19488768, 'steps': 38063, 'loss/train': 1.20078706741333} 02/26/2022 02:51:35 - INFO - codeparrot_training - Step 38064: {'lr': 7.24843772355189e-05, 'samples': 19489280, 'steps': 38064, 'loss/train': 1.283164143562317} 02/26/2022 02:51:40 - INFO - codeparrot_training - Step 38065: {'lr': 7.247285617378221e-05, 'samples': 19489792, 'steps': 38065, 'loss/train': 1.3847582340240479} 02/26/2022 02:51:44 - INFO - codeparrot_training - Step 38066: {'lr': 7.246133587251549e-05, 'samples': 19490304, 'steps': 38066, 'loss/train': 1.312401533126831} 02/26/2022 02:51:49 - INFO - codeparrot_training - Step 38067: {'lr': 7.244981633176798e-05, 'samples': 19490816, 'steps': 38067, 'loss/train': 1.566505789756775} 02/26/2022 02:51:53 - INFO - codeparrot_training - Step 38068: {'lr': 7.243829755158923e-05, 'samples': 19491328, 'steps': 38068, 'loss/train': 1.360145092010498} 02/26/2022 02:51:58 - INFO - codeparrot_training - Step 38069: {'lr': 7.242677953202828e-05, 'samples': 19491840, 'steps': 38069, 'loss/train': 0.4492625892162323} 02/26/2022 02:52:02 - INFO - codeparrot_training - Step 38070: {'lr': 7.241526227313469e-05, 'samples': 19492352, 'steps': 38070, 'loss/train': 1.487532377243042} 02/26/2022 02:52:08 - INFO - codeparrot_training - Step 38071: {'lr': 7.240374577495765e-05, 'samples': 19492864, 'steps': 38071, 'loss/train': 1.5095921754837036} 02/26/2022 02:52:12 - INFO - codeparrot_training - Step 38072: {'lr': 7.239223003754672e-05, 'samples': 19493376, 'steps': 38072, 'loss/train': 1.449380874633789} 02/26/2022 02:52:17 - INFO - codeparrot_training - Step 38073: {'lr': 7.238071506095093e-05, 'samples': 19493888, 'steps': 38073, 'loss/train': 1.179553747177124} 02/26/2022 02:52:21 - INFO - codeparrot_training - Step 38074: {'lr': 7.236920084521981e-05, 'samples': 19494400, 'steps': 38074, 'loss/train': 2.299309015274048} 02/26/2022 02:52:26 - INFO - codeparrot_training - Step 38075: {'lr': 7.235768739040257e-05, 'samples': 19494912, 'steps': 38075, 'loss/train': 1.764549732208252} 02/26/2022 02:52:30 - INFO - codeparrot_training - Step 38076: {'lr': 7.234617469654866e-05, 'samples': 19495424, 'steps': 38076, 'loss/train': 1.7865841388702393} 02/26/2022 02:52:35 - INFO - codeparrot_training - Step 38077: {'lr': 7.233466276370735e-05, 'samples': 19495936, 'steps': 38077, 'loss/train': 1.5146217346191406} 02/26/2022 02:52:38 - INFO - codeparrot_training - Step 38078: {'lr': 7.232315159192793e-05, 'samples': 19496448, 'steps': 38078, 'loss/train': 2.66728138923645} 02/26/2022 02:52:44 - INFO - codeparrot_training - Step 38079: {'lr': 7.23116411812596e-05, 'samples': 19496960, 'steps': 38079, 'loss/train': 0.6679148077964783} 02/26/2022 02:52:48 - INFO - codeparrot_training - Step 38080: {'lr': 7.230013153175189e-05, 'samples': 19497472, 'steps': 38080, 'loss/train': 2.0153379440307617} 02/26/2022 02:52:53 - INFO - codeparrot_training - Step 38081: {'lr': 7.228862264345395e-05, 'samples': 19497984, 'steps': 38081, 'loss/train': 0.03077789396047592} 02/26/2022 02:52:57 - INFO - codeparrot_training - Step 38082: {'lr': 7.227711451641516e-05, 'samples': 19498496, 'steps': 38082, 'loss/train': 1.6509605646133423} 02/26/2022 02:53:02 - INFO - codeparrot_training - Step 38083: {'lr': 7.226560715068467e-05, 'samples': 19499008, 'steps': 38083, 'loss/train': 2.0321590900421143} 02/26/2022 02:53:06 - INFO - codeparrot_training - Step 38084: {'lr': 7.225410054631199e-05, 'samples': 19499520, 'steps': 38084, 'loss/train': 2.4848997592926025} 02/26/2022 02:53:12 - INFO - codeparrot_training - Step 38085: {'lr': 7.224259470334629e-05, 'samples': 19500032, 'steps': 38085, 'loss/train': 1.8340669870376587} 02/26/2022 02:53:16 - INFO - codeparrot_training - Step 38086: {'lr': 7.223108962183685e-05, 'samples': 19500544, 'steps': 38086, 'loss/train': 2.246068239212036} 02/26/2022 02:53:21 - INFO - codeparrot_training - Step 38087: {'lr': 7.221958530183293e-05, 'samples': 19501056, 'steps': 38087, 'loss/train': 0.7005888223648071} 02/26/2022 02:53:25 - INFO - codeparrot_training - Step 38088: {'lr': 7.22080817433839e-05, 'samples': 19501568, 'steps': 38088, 'loss/train': 0.1733948439359665} 02/26/2022 02:53:30 - INFO - codeparrot_training - Step 38089: {'lr': 7.2196578946539e-05, 'samples': 19502080, 'steps': 38089, 'loss/train': 2.106456756591797} 02/26/2022 02:53:34 - INFO - codeparrot_training - Step 38090: {'lr': 7.218507691134752e-05, 'samples': 19502592, 'steps': 38090, 'loss/train': 2.003767728805542} 02/26/2022 02:53:39 - INFO - codeparrot_training - Step 38091: {'lr': 7.21735756378587e-05, 'samples': 19503104, 'steps': 38091, 'loss/train': 1.810054063796997} 02/26/2022 02:53:43 - INFO - codeparrot_training - Step 38092: {'lr': 7.216207512612172e-05, 'samples': 19503616, 'steps': 38092, 'loss/train': 1.4031343460083008} 02/26/2022 02:53:48 - INFO - codeparrot_training - Step 38093: {'lr': 7.215057537618602e-05, 'samples': 19504128, 'steps': 38093, 'loss/train': 1.9758297204971313} 02/26/2022 02:53:52 - INFO - codeparrot_training - Step 38094: {'lr': 7.21390763881008e-05, 'samples': 19504640, 'steps': 38094, 'loss/train': 1.3060182332992554} 02/26/2022 02:53:57 - INFO - codeparrot_training - Step 38095: {'lr': 7.212757816191528e-05, 'samples': 19505152, 'steps': 38095, 'loss/train': 1.9382715225219727} 02/26/2022 02:54:01 - INFO - codeparrot_training - Step 38096: {'lr': 7.211608069767867e-05, 'samples': 19505664, 'steps': 38096, 'loss/train': 2.108659505844116} 02/26/2022 02:54:07 - INFO - codeparrot_training - Step 38097: {'lr': 7.210458399544037e-05, 'samples': 19506176, 'steps': 38097, 'loss/train': 2.769747257232666} 02/26/2022 02:54:11 - INFO - codeparrot_training - Step 38098: {'lr': 7.209308805524944e-05, 'samples': 19506688, 'steps': 38098, 'loss/train': 0.6523610353469849} 02/26/2022 02:54:16 - INFO - codeparrot_training - Step 38099: {'lr': 7.20815928771554e-05, 'samples': 19507200, 'steps': 38099, 'loss/train': 1.8880488872528076} 02/26/2022 02:54:20 - INFO - codeparrot_training - Step 38100: {'lr': 7.207009846120718e-05, 'samples': 19507712, 'steps': 38100, 'loss/train': 0.9422760009765625} 02/26/2022 02:54:25 - INFO - codeparrot_training - Step 38101: {'lr': 7.205860480745422e-05, 'samples': 19508224, 'steps': 38101, 'loss/train': 2.408179759979248} 02/26/2022 02:54:29 - INFO - codeparrot_training - Step 38102: {'lr': 7.204711191594562e-05, 'samples': 19508736, 'steps': 38102, 'loss/train': 2.15340518951416} 02/26/2022 02:54:34 - INFO - codeparrot_training - Step 38103: {'lr': 7.203561978673085e-05, 'samples': 19509248, 'steps': 38103, 'loss/train': 1.6576495170593262} 02/26/2022 02:54:38 - INFO - codeparrot_training - Step 38104: {'lr': 7.202412841985879e-05, 'samples': 19509760, 'steps': 38104, 'loss/train': 2.358712673187256} 02/26/2022 02:54:43 - INFO - codeparrot_training - Step 38105: {'lr': 7.201263781537895e-05, 'samples': 19510272, 'steps': 38105, 'loss/train': 1.4066082239151} 02/26/2022 02:54:47 - INFO - codeparrot_training - Step 38106: {'lr': 7.200114797334037e-05, 'samples': 19510784, 'steps': 38106, 'loss/train': 1.8579227924346924} 02/26/2022 02:54:53 - INFO - codeparrot_training - Step 38107: {'lr': 7.198965889379253e-05, 'samples': 19511296, 'steps': 38107, 'loss/train': 1.561664342880249} 02/26/2022 02:54:57 - INFO - codeparrot_training - Step 38108: {'lr': 7.197817057678427e-05, 'samples': 19511808, 'steps': 38108, 'loss/train': 1.1147043704986572} 02/26/2022 02:55:02 - INFO - codeparrot_training - Step 38109: {'lr': 7.196668302236509e-05, 'samples': 19512320, 'steps': 38109, 'loss/train': 2.020216941833496} 02/26/2022 02:55:06 - INFO - codeparrot_training - Step 38110: {'lr': 7.195519623058405e-05, 'samples': 19512832, 'steps': 38110, 'loss/train': 2.7649996280670166} 02/26/2022 02:55:11 - INFO - codeparrot_training - Step 38111: {'lr': 7.194371020149048e-05, 'samples': 19513344, 'steps': 38111, 'loss/train': 1.8314980268478394} 02/26/2022 02:55:15 - INFO - codeparrot_training - Step 38112: {'lr': 7.193222493513352e-05, 'samples': 19513856, 'steps': 38112, 'loss/train': 2.3585622310638428} 02/26/2022 02:55:20 - INFO - codeparrot_training - Step 38113: {'lr': 7.192074043156235e-05, 'samples': 19514368, 'steps': 38113, 'loss/train': 1.9599764347076416} 02/26/2022 02:55:24 - INFO - codeparrot_training - Step 38114: {'lr': 7.190925669082613e-05, 'samples': 19514880, 'steps': 38114, 'loss/train': 1.68540358543396} 02/26/2022 02:55:29 - INFO - codeparrot_training - Step 38115: {'lr': 7.189777371297415e-05, 'samples': 19515392, 'steps': 38115, 'loss/train': 1.7949540615081787} 02/26/2022 02:55:35 - INFO - codeparrot_training - Step 38116: {'lr': 7.188629149805556e-05, 'samples': 19515904, 'steps': 38116, 'loss/train': 1.2822867631912231} 02/26/2022 02:55:38 - INFO - codeparrot_training - Step 38117: {'lr': 7.187481004611956e-05, 'samples': 19516416, 'steps': 38117, 'loss/train': 0.1690545529127121} 02/26/2022 02:55:44 - INFO - codeparrot_training - Step 38118: {'lr': 7.18633293572152e-05, 'samples': 19516928, 'steps': 38118, 'loss/train': 1.5017328262329102} 02/26/2022 02:55:47 - INFO - codeparrot_training - Step 38119: {'lr': 7.18518494313919e-05, 'samples': 19517440, 'steps': 38119, 'loss/train': 1.8454328775405884} 02/26/2022 02:55:54 - INFO - codeparrot_training - Step 38120: {'lr': 7.184037026869867e-05, 'samples': 19517952, 'steps': 38120, 'loss/train': 1.5571293830871582} 02/26/2022 02:55:57 - INFO - codeparrot_training - Step 38121: {'lr': 7.182889186918474e-05, 'samples': 19518464, 'steps': 38121, 'loss/train': 2.341313600540161} 02/26/2022 02:56:01 - INFO - codeparrot_training - Step 38122: {'lr': 7.181741423289917e-05, 'samples': 19518976, 'steps': 38122, 'loss/train': 2.2340619564056396} 02/26/2022 02:56:06 - INFO - codeparrot_training - Step 38123: {'lr': 7.180593735989129e-05, 'samples': 19519488, 'steps': 38123, 'loss/train': 1.768855333328247} 02/26/2022 02:56:10 - INFO - codeparrot_training - Step 38124: {'lr': 7.179446125021022e-05, 'samples': 19520000, 'steps': 38124, 'loss/train': 2.415536880493164} 02/26/2022 02:56:15 - INFO - codeparrot_training - Step 38125: {'lr': 7.178298590390509e-05, 'samples': 19520512, 'steps': 38125, 'loss/train': 2.193103075027466} 02/26/2022 02:56:19 - INFO - codeparrot_training - Step 38126: {'lr': 7.177151132102496e-05, 'samples': 19521024, 'steps': 38126, 'loss/train': 1.5276339054107666} 02/26/2022 02:56:25 - INFO - codeparrot_training - Step 38127: {'lr': 7.176003750161916e-05, 'samples': 19521536, 'steps': 38127, 'loss/train': 1.4621673822402954} 02/26/2022 02:56:28 - INFO - codeparrot_training - Step 38128: {'lr': 7.174856444573676e-05, 'samples': 19522048, 'steps': 38128, 'loss/train': 1.3434168100357056} 02/26/2022 02:56:32 - INFO - codeparrot_training - Step 38129: {'lr': 7.173709215342695e-05, 'samples': 19522560, 'steps': 38129, 'loss/train': 0.39635080099105835} 02/26/2022 02:56:37 - INFO - codeparrot_training - Step 38130: {'lr': 7.17256206247387e-05, 'samples': 19523072, 'steps': 38130, 'loss/train': 2.165515899658203} 02/26/2022 02:56:41 - INFO - codeparrot_training - Step 38131: {'lr': 7.171414985972138e-05, 'samples': 19523584, 'steps': 38131, 'loss/train': 1.5523850917816162} 02/26/2022 02:56:48 - INFO - codeparrot_training - Step 38132: {'lr': 7.170267985842405e-05, 'samples': 19524096, 'steps': 38132, 'loss/train': 1.6655939817428589} 02/26/2022 02:56:52 - INFO - codeparrot_training - Step 38133: {'lr': 7.169121062089578e-05, 'samples': 19524608, 'steps': 38133, 'loss/train': 1.7885385751724243} 02/26/2022 02:56:57 - INFO - codeparrot_training - Step 38134: {'lr': 7.16797421471857e-05, 'samples': 19525120, 'steps': 38134, 'loss/train': 1.0705183744430542} 02/26/2022 02:57:03 - INFO - codeparrot_training - Step 38135: {'lr': 7.166827443734306e-05, 'samples': 19525632, 'steps': 38135, 'loss/train': 1.5479942560195923} 02/26/2022 02:57:06 - INFO - codeparrot_training - Step 38136: {'lr': 7.165680749141689e-05, 'samples': 19526144, 'steps': 38136, 'loss/train': 2.13727068901062} 02/26/2022 02:57:12 - INFO - codeparrot_training - Step 38137: {'lr': 7.164534130945624e-05, 'samples': 19526656, 'steps': 38137, 'loss/train': 0.5930886268615723} 02/26/2022 02:57:15 - INFO - codeparrot_training - Step 38138: {'lr': 7.16338758915105e-05, 'samples': 19527168, 'steps': 38138, 'loss/train': 0.3744605779647827} 02/26/2022 02:57:21 - INFO - codeparrot_training - Step 38139: {'lr': 7.162241123762841e-05, 'samples': 19527680, 'steps': 38139, 'loss/train': 0.6163443326950073} 02/26/2022 02:57:24 - INFO - codeparrot_training - Step 38140: {'lr': 7.16109473478594e-05, 'samples': 19528192, 'steps': 38140, 'loss/train': 2.636695384979248} 02/26/2022 02:57:32 - INFO - codeparrot_training - Step 38141: {'lr': 7.159948422225232e-05, 'samples': 19528704, 'steps': 38141, 'loss/train': 2.381225109100342} 02/26/2022 02:57:35 - INFO - codeparrot_training - Step 38142: {'lr': 7.158802186085656e-05, 'samples': 19529216, 'steps': 38142, 'loss/train': 1.7457891702651978} 02/26/2022 02:57:41 - INFO - codeparrot_training - Step 38143: {'lr': 7.157656026372092e-05, 'samples': 19529728, 'steps': 38143, 'loss/train': 1.2105522155761719} 02/26/2022 02:57:44 - INFO - codeparrot_training - Step 38144: {'lr': 7.156509943089471e-05, 'samples': 19530240, 'steps': 38144, 'loss/train': 1.0127432346343994} 02/26/2022 02:57:50 - INFO - codeparrot_training - Step 38145: {'lr': 7.155363936242686e-05, 'samples': 19530752, 'steps': 38145, 'loss/train': 1.8298380374908447} 02/26/2022 02:57:53 - INFO - codeparrot_training - Step 38146: {'lr': 7.154218005836668e-05, 'samples': 19531264, 'steps': 38146, 'loss/train': 1.9088324308395386} 02/26/2022 02:57:59 - INFO - codeparrot_training - Step 38147: {'lr': 7.153072151876311e-05, 'samples': 19531776, 'steps': 38147, 'loss/train': 0.6606325507164001} 02/26/2022 02:58:02 - INFO - codeparrot_training - Step 38148: {'lr': 7.151926374366524e-05, 'samples': 19532288, 'steps': 38148, 'loss/train': 1.7188962697982788} 02/26/2022 02:58:08 - INFO - codeparrot_training - Step 38149: {'lr': 7.150780673312212e-05, 'samples': 19532800, 'steps': 38149, 'loss/train': 2.3273098468780518} 02/26/2022 02:58:11 - INFO - codeparrot_training - Step 38150: {'lr': 7.149635048718294e-05, 'samples': 19533312, 'steps': 38150, 'loss/train': 2.488645315170288} 02/26/2022 02:58:17 - INFO - codeparrot_training - Step 38151: {'lr': 7.148489500589673e-05, 'samples': 19533824, 'steps': 38151, 'loss/train': 3.1085617542266846} 02/26/2022 02:58:20 - INFO - codeparrot_training - Step 38152: {'lr': 7.147344028931252e-05, 'samples': 19534336, 'steps': 38152, 'loss/train': 2.4928834438323975} 02/26/2022 02:58:24 - INFO - codeparrot_training - Step 38153: {'lr': 7.146198633747936e-05, 'samples': 19534848, 'steps': 38153, 'loss/train': 1.8379136323928833} 02/26/2022 02:58:31 - INFO - codeparrot_training - Step 38154: {'lr': 7.145053315044639e-05, 'samples': 19535360, 'steps': 38154, 'loss/train': 1.7272346019744873} 02/26/2022 02:58:34 - INFO - codeparrot_training - Step 38155: {'lr': 7.143908072826269e-05, 'samples': 19535872, 'steps': 38155, 'loss/train': 0.97761070728302} 02/26/2022 02:58:40 - INFO - codeparrot_training - Step 38156: {'lr': 7.142762907097721e-05, 'samples': 19536384, 'steps': 38156, 'loss/train': 1.1927820444107056} 02/26/2022 02:58:46 - INFO - codeparrot_training - Step 38157: {'lr': 7.141617817863905e-05, 'samples': 19536896, 'steps': 38157, 'loss/train': 1.2962172031402588} 02/26/2022 02:58:49 - INFO - codeparrot_training - Step 38158: {'lr': 7.14047280512973e-05, 'samples': 19537408, 'steps': 38158, 'loss/train': 1.986090898513794} 02/26/2022 02:58:53 - INFO - codeparrot_training - Step 38159: {'lr': 7.139327868900094e-05, 'samples': 19537920, 'steps': 38159, 'loss/train': 1.7715954780578613} 02/26/2022 02:58:58 - INFO - codeparrot_training - Step 38160: {'lr': 7.138183009179921e-05, 'samples': 19538432, 'steps': 38160, 'loss/train': 0.8882639408111572} 02/26/2022 02:59:04 - INFO - codeparrot_training - Step 38161: {'lr': 7.137038225974085e-05, 'samples': 19538944, 'steps': 38161, 'loss/train': 1.5299500226974487} 02/26/2022 02:59:07 - INFO - codeparrot_training - Step 38162: {'lr': 7.135893519287512e-05, 'samples': 19539456, 'steps': 38162, 'loss/train': 0.7520651817321777} 02/26/2022 02:59:15 - INFO - codeparrot_training - Step 38163: {'lr': 7.13474888912509e-05, 'samples': 19539968, 'steps': 38163, 'loss/train': 1.6187598705291748} 02/26/2022 02:59:18 - INFO - codeparrot_training - Step 38164: {'lr': 7.133604335491747e-05, 'samples': 19540480, 'steps': 38164, 'loss/train': 0.7635990381240845} 02/26/2022 02:59:24 - INFO - codeparrot_training - Step 38165: {'lr': 7.132459858392357e-05, 'samples': 19540992, 'steps': 38165, 'loss/train': 1.7196338176727295} 02/26/2022 02:59:27 - INFO - codeparrot_training - Step 38166: {'lr': 7.131315457831841e-05, 'samples': 19541504, 'steps': 38166, 'loss/train': 2.479658365249634} 02/26/2022 02:59:33 - INFO - codeparrot_training - Step 38167: {'lr': 7.130171133815086e-05, 'samples': 19542016, 'steps': 38167, 'loss/train': 1.700913667678833} 02/26/2022 02:59:36 - INFO - codeparrot_training - Step 38168: {'lr': 7.12902688634702e-05, 'samples': 19542528, 'steps': 38168, 'loss/train': 3.29341721534729} 02/26/2022 02:59:42 - INFO - codeparrot_training - Step 38169: {'lr': 7.127882715432515e-05, 'samples': 19543040, 'steps': 38169, 'loss/train': 1.369329571723938} 02/26/2022 02:59:45 - INFO - codeparrot_training - Step 38170: {'lr': 7.126738621076493e-05, 'samples': 19543552, 'steps': 38170, 'loss/train': 0.4311719834804535} 02/26/2022 02:59:51 - INFO - codeparrot_training - Step 38171: {'lr': 7.125594603283845e-05, 'samples': 19544064, 'steps': 38171, 'loss/train': 1.105877161026001} 02/26/2022 02:59:54 - INFO - codeparrot_training - Step 38172: {'lr': 7.124450662059468e-05, 'samples': 19544576, 'steps': 38172, 'loss/train': 1.7536799907684326} 02/26/2022 03:00:00 - INFO - codeparrot_training - Step 38173: {'lr': 7.123306797408274e-05, 'samples': 19545088, 'steps': 38173, 'loss/train': 1.8302478790283203} 02/26/2022 03:00:03 - INFO - codeparrot_training - Step 38174: {'lr': 7.122163009335159e-05, 'samples': 19545600, 'steps': 38174, 'loss/train': 0.6642072796821594} 02/26/2022 03:00:09 - INFO - codeparrot_training - Step 38175: {'lr': 7.121019297845019e-05, 'samples': 19546112, 'steps': 38175, 'loss/train': 1.7727161645889282} 02/26/2022 03:00:12 - INFO - codeparrot_training - Step 38176: {'lr': 7.119875662942748e-05, 'samples': 19546624, 'steps': 38176, 'loss/train': 1.3386083841323853} 02/26/2022 03:00:20 - INFO - codeparrot_training - Step 38177: {'lr': 7.11873210463326e-05, 'samples': 19547136, 'steps': 38177, 'loss/train': 1.9967362880706787} 02/26/2022 03:00:23 - INFO - codeparrot_training - Step 38178: {'lr': 7.117588622921448e-05, 'samples': 19547648, 'steps': 38178, 'loss/train': 1.3029136657714844} 02/26/2022 03:00:27 - INFO - codeparrot_training - Step 38179: {'lr': 7.116445217812207e-05, 'samples': 19548160, 'steps': 38179, 'loss/train': 1.8939669132232666} 02/26/2022 03:00:32 - INFO - codeparrot_training - Step 38180: {'lr': 7.115301889310427e-05, 'samples': 19548672, 'steps': 38180, 'loss/train': 1.8544868230819702} 02/26/2022 03:00:36 - INFO - codeparrot_training - Step 38181: {'lr': 7.114158637421023e-05, 'samples': 19549184, 'steps': 38181, 'loss/train': 1.6866735219955444} 02/26/2022 03:00:41 - INFO - codeparrot_training - Step 38182: {'lr': 7.113015462148884e-05, 'samples': 19549696, 'steps': 38182, 'loss/train': 1.9695546627044678} 02/26/2022 03:00:45 - INFO - codeparrot_training - Step 38183: {'lr': 7.111872363498905e-05, 'samples': 19550208, 'steps': 38183, 'loss/train': 3.7168359756469727} 02/26/2022 03:00:50 - INFO - codeparrot_training - Step 38184: {'lr': 7.11072934147598e-05, 'samples': 19550720, 'steps': 38184, 'loss/train': 1.8615844249725342} 02/26/2022 03:00:54 - INFO - codeparrot_training - Step 38185: {'lr': 7.109586396085016e-05, 'samples': 19551232, 'steps': 38185, 'loss/train': 2.2218596935272217} 02/26/2022 03:00:59 - INFO - codeparrot_training - Step 38186: {'lr': 7.108443527330904e-05, 'samples': 19551744, 'steps': 38186, 'loss/train': 1.7533131837844849} 02/26/2022 03:01:03 - INFO - codeparrot_training - Step 38187: {'lr': 7.107300735218539e-05, 'samples': 19552256, 'steps': 38187, 'loss/train': 1.4288514852523804} 02/26/2022 03:01:08 - INFO - codeparrot_training - Step 38188: {'lr': 7.106158019752806e-05, 'samples': 19552768, 'steps': 38188, 'loss/train': 1.6369614601135254} 02/26/2022 03:01:12 - INFO - codeparrot_training - Step 38189: {'lr': 7.10501538093862e-05, 'samples': 19553280, 'steps': 38189, 'loss/train': 1.2787474393844604} 02/26/2022 03:01:19 - INFO - codeparrot_training - Step 38190: {'lr': 7.103872818780866e-05, 'samples': 19553792, 'steps': 38190, 'loss/train': 1.8749148845672607} 02/26/2022 03:01:22 - INFO - codeparrot_training - Step 38191: {'lr': 7.102730333284438e-05, 'samples': 19554304, 'steps': 38191, 'loss/train': 2.030489444732666} 02/26/2022 03:01:28 - INFO - codeparrot_training - Step 38192: {'lr': 7.10158792445422e-05, 'samples': 19554816, 'steps': 38192, 'loss/train': 1.023307204246521} 02/26/2022 03:01:34 - INFO - codeparrot_training - Step 38193: {'lr': 7.100445592295126e-05, 'samples': 19555328, 'steps': 38193, 'loss/train': 1.9115947484970093} 02/26/2022 03:01:37 - INFO - codeparrot_training - Step 38194: {'lr': 7.099303336812032e-05, 'samples': 19555840, 'steps': 38194, 'loss/train': 1.884045124053955} 02/26/2022 03:01:43 - INFO - codeparrot_training - Step 38195: {'lr': 7.098161158009853e-05, 'samples': 19556352, 'steps': 38195, 'loss/train': 1.5298393964767456} 02/26/2022 03:01:46 - INFO - codeparrot_training - Step 38196: {'lr': 7.097019055893452e-05, 'samples': 19556864, 'steps': 38196, 'loss/train': 1.0842851400375366} 02/26/2022 03:01:52 - INFO - codeparrot_training - Step 38197: {'lr': 7.095877030467743e-05, 'samples': 19557376, 'steps': 38197, 'loss/train': 0.9352651238441467} 02/26/2022 03:01:55 - INFO - codeparrot_training - Step 38198: {'lr': 7.094735081737605e-05, 'samples': 19557888, 'steps': 38198, 'loss/train': 1.6735531091690063} 02/26/2022 03:02:02 - INFO - codeparrot_training - Step 38199: {'lr': 7.093593209707952e-05, 'samples': 19558400, 'steps': 38199, 'loss/train': 0.9866325259208679} 02/26/2022 03:02:06 - INFO - codeparrot_training - Step 38200: {'lr': 7.092451414383643e-05, 'samples': 19558912, 'steps': 38200, 'loss/train': 0.46933799982070923} 02/26/2022 03:02:12 - INFO - codeparrot_training - Step 38201: {'lr': 7.091309695769596e-05, 'samples': 19559424, 'steps': 38201, 'loss/train': 2.1038448810577393} 02/26/2022 03:02:15 - INFO - codeparrot_training - Step 38202: {'lr': 7.090168053870683e-05, 'samples': 19559936, 'steps': 38202, 'loss/train': 2.31260085105896} 02/26/2022 03:02:19 - INFO - codeparrot_training - Step 38203: {'lr': 7.089026488691817e-05, 'samples': 19560448, 'steps': 38203, 'loss/train': 2.04152250289917} 02/26/2022 03:02:24 - INFO - codeparrot_training - Step 38204: {'lr': 7.087885000237859e-05, 'samples': 19560960, 'steps': 38204, 'loss/train': 1.5692404508590698} 02/26/2022 03:02:28 - INFO - codeparrot_training - Step 38205: {'lr': 7.086743588513723e-05, 'samples': 19561472, 'steps': 38205, 'loss/train': 1.3013707399368286} 02/26/2022 03:02:33 - INFO - codeparrot_training - Step 38206: {'lr': 7.085602253524281e-05, 'samples': 19561984, 'steps': 38206, 'loss/train': 1.498496174812317} 02/26/2022 03:02:37 - INFO - codeparrot_training - Step 38207: {'lr': 7.084460995274439e-05, 'samples': 19562496, 'steps': 38207, 'loss/train': 1.7957921028137207} 02/26/2022 03:02:42 - INFO - codeparrot_training - Step 38208: {'lr': 7.083319813769076e-05, 'samples': 19563008, 'steps': 38208, 'loss/train': 0.141988143324852} 02/26/2022 03:02:46 - INFO - codeparrot_training - Step 38209: {'lr': 7.082178709013082e-05, 'samples': 19563520, 'steps': 38209, 'loss/train': 1.4009586572647095} 02/26/2022 03:02:53 - INFO - codeparrot_training - Step 38210: {'lr': 7.081037681011338e-05, 'samples': 19564032, 'steps': 38210, 'loss/train': 2.7622272968292236} 02/26/2022 03:02:56 - INFO - codeparrot_training - Step 38211: {'lr': 7.079896729768745e-05, 'samples': 19564544, 'steps': 38211, 'loss/train': 1.6070525646209717} 02/26/2022 03:03:02 - INFO - codeparrot_training - Step 38212: {'lr': 7.078755855290183e-05, 'samples': 19565056, 'steps': 38212, 'loss/train': 1.5549302101135254} 02/26/2022 03:03:05 - INFO - codeparrot_training - Step 38213: {'lr': 7.077615057580545e-05, 'samples': 19565568, 'steps': 38213, 'loss/train': 0.098085917532444} 02/26/2022 03:03:11 - INFO - codeparrot_training - Step 38214: {'lr': 7.0764743366447e-05, 'samples': 19566080, 'steps': 38214, 'loss/train': 2.2977380752563477} 02/26/2022 03:03:14 - INFO - codeparrot_training - Step 38215: {'lr': 7.075333692487559e-05, 'samples': 19566592, 'steps': 38215, 'loss/train': 1.908610463142395} 02/26/2022 03:03:20 - INFO - codeparrot_training - Step 38216: {'lr': 7.074193125113996e-05, 'samples': 19567104, 'steps': 38216, 'loss/train': 1.382575273513794} 02/26/2022 03:03:23 - INFO - codeparrot_training - Step 38217: {'lr': 7.073052634528898e-05, 'samples': 19567616, 'steps': 38217, 'loss/train': 2.211209774017334} 02/26/2022 03:03:29 - INFO - codeparrot_training - Step 38218: {'lr': 7.071912220737147e-05, 'samples': 19568128, 'steps': 38218, 'loss/train': 1.5111160278320312} 02/26/2022 03:03:32 - INFO - codeparrot_training - Step 38219: {'lr': 7.070771883743626e-05, 'samples': 19568640, 'steps': 38219, 'loss/train': 2.004314422607422} 02/26/2022 03:03:40 - INFO - codeparrot_training - Step 38220: {'lr': 7.069631623553232e-05, 'samples': 19569152, 'steps': 38220, 'loss/train': 0.437398761510849} 02/26/2022 03:03:43 - INFO - codeparrot_training - Step 38221: {'lr': 7.068491440170843e-05, 'samples': 19569664, 'steps': 38221, 'loss/train': 1.8562313318252563} 02/26/2022 03:03:49 - INFO - codeparrot_training - Step 38222: {'lr': 7.067351333601343e-05, 'samples': 19570176, 'steps': 38222, 'loss/train': 1.1736711263656616} 02/26/2022 03:03:52 - INFO - codeparrot_training - Step 38223: {'lr': 7.06621130384961e-05, 'samples': 19570688, 'steps': 38223, 'loss/train': 1.7102985382080078} 02/26/2022 03:03:58 - INFO - codeparrot_training - Step 38224: {'lr': 7.065071350920538e-05, 'samples': 19571200, 'steps': 38224, 'loss/train': 0.5994188785552979} 02/26/2022 03:04:01 - INFO - codeparrot_training - Step 38225: {'lr': 7.063931474819005e-05, 'samples': 19571712, 'steps': 38225, 'loss/train': 2.1661248207092285} 02/26/2022 03:04:07 - INFO - codeparrot_training - Step 38226: {'lr': 7.062791675549896e-05, 'samples': 19572224, 'steps': 38226, 'loss/train': 2.3120572566986084} 02/26/2022 03:04:10 - INFO - codeparrot_training - Step 38227: {'lr': 7.061651953118084e-05, 'samples': 19572736, 'steps': 38227, 'loss/train': 2.0443308353424072} 02/26/2022 03:04:16 - INFO - codeparrot_training - Step 38228: {'lr': 7.060512307528466e-05, 'samples': 19573248, 'steps': 38228, 'loss/train': 1.6328598260879517} 02/26/2022 03:04:19 - INFO - codeparrot_training - Step 38229: {'lr': 7.059372738785915e-05, 'samples': 19573760, 'steps': 38229, 'loss/train': 3.0696194171905518} 02/26/2022 03:04:25 - INFO - codeparrot_training - Step 38230: {'lr': 7.058233246895315e-05, 'samples': 19574272, 'steps': 38230, 'loss/train': 1.9886451959609985} 02/26/2022 03:04:28 - INFO - codeparrot_training - Step 38231: {'lr': 7.057093831861542e-05, 'samples': 19574784, 'steps': 38231, 'loss/train': 0.9862726926803589} 02/26/2022 03:04:34 - INFO - codeparrot_training - Step 38232: {'lr': 7.055954493689487e-05, 'samples': 19575296, 'steps': 38232, 'loss/train': 2.040212392807007} 02/26/2022 03:04:37 - INFO - codeparrot_training - Step 38233: {'lr': 7.054815232384018e-05, 'samples': 19575808, 'steps': 38233, 'loss/train': 1.2924765348434448} 02/26/2022 03:04:43 - INFO - codeparrot_training - Step 38234: {'lr': 7.053676047950037e-05, 'samples': 19576320, 'steps': 38234, 'loss/train': 2.6697871685028076} 02/26/2022 03:04:46 - INFO - codeparrot_training - Step 38235: {'lr': 7.052536940392393e-05, 'samples': 19576832, 'steps': 38235, 'loss/train': 1.3289813995361328} 02/26/2022 03:04:54 - INFO - codeparrot_training - Step 38236: {'lr': 7.051397909715992e-05, 'samples': 19577344, 'steps': 38236, 'loss/train': 1.9227101802825928} 02/26/2022 03:04:57 - INFO - codeparrot_training - Step 38237: {'lr': 7.050258955925692e-05, 'samples': 19577856, 'steps': 38237, 'loss/train': 1.0109972953796387} 02/26/2022 03:05:02 - INFO - codeparrot_training - Step 38238: {'lr': 7.049120079026403e-05, 'samples': 19578368, 'steps': 38238, 'loss/train': 0.5722323656082153} 02/26/2022 03:05:06 - INFO - codeparrot_training - Step 38239: {'lr': 7.047981279022964e-05, 'samples': 19578880, 'steps': 38239, 'loss/train': 2.2172272205352783} 02/26/2022 03:05:11 - INFO - codeparrot_training - Step 38240: {'lr': 7.046842555920283e-05, 'samples': 19579392, 'steps': 38240, 'loss/train': 2.937234878540039} 02/26/2022 03:05:15 - INFO - codeparrot_training - Step 38241: {'lr': 7.04570390972322e-05, 'samples': 19579904, 'steps': 38241, 'loss/train': 0.5131740570068359} 02/26/2022 03:05:20 - INFO - codeparrot_training - Step 38242: {'lr': 7.044565340436664e-05, 'samples': 19580416, 'steps': 38242, 'loss/train': 1.65739107131958} 02/26/2022 03:05:24 - INFO - codeparrot_training - Step 38243: {'lr': 7.043426848065493e-05, 'samples': 19580928, 'steps': 38243, 'loss/train': 1.4585521221160889} 02/26/2022 03:05:29 - INFO - codeparrot_training - Step 38244: {'lr': 7.042288432614577e-05, 'samples': 19581440, 'steps': 38244, 'loss/train': 1.4854989051818848} 02/26/2022 03:05:33 - INFO - codeparrot_training - Step 38245: {'lr': 7.041150094088788e-05, 'samples': 19581952, 'steps': 38245, 'loss/train': 1.4963589906692505} 02/26/2022 03:05:40 - INFO - codeparrot_training - Step 38246: {'lr': 7.040011832493021e-05, 'samples': 19582464, 'steps': 38246, 'loss/train': 1.304911494255066} 02/26/2022 03:05:43 - INFO - codeparrot_training - Step 38247: {'lr': 7.038873647832137e-05, 'samples': 19582976, 'steps': 38247, 'loss/train': 0.5257362723350525} 02/26/2022 03:05:49 - INFO - codeparrot_training - Step 38248: {'lr': 7.037735540111015e-05, 'samples': 19583488, 'steps': 38248, 'loss/train': 2.09340763092041} 02/26/2022 03:05:52 - INFO - codeparrot_training - Step 38249: {'lr': 7.036597509334521e-05, 'samples': 19584000, 'steps': 38249, 'loss/train': 1.0398221015930176} 02/26/2022 03:05:58 - INFO - codeparrot_training - Step 38250: {'lr': 7.035459555507548e-05, 'samples': 19584512, 'steps': 38250, 'loss/train': 0.8158255219459534} 02/26/2022 03:06:01 - INFO - codeparrot_training - Step 38251: {'lr': 7.034321678634964e-05, 'samples': 19585024, 'steps': 38251, 'loss/train': 0.887293815612793} 02/26/2022 03:06:07 - INFO - codeparrot_training - Step 38252: {'lr': 7.033183878721639e-05, 'samples': 19585536, 'steps': 38252, 'loss/train': 1.9078387022018433} 02/26/2022 03:06:12 - INFO - codeparrot_training - Step 38253: {'lr': 7.032046155772442e-05, 'samples': 19586048, 'steps': 38253, 'loss/train': 2.415226697921753} 02/26/2022 03:06:16 - INFO - codeparrot_training - Step 38254: {'lr': 7.030908509792261e-05, 'samples': 19586560, 'steps': 38254, 'loss/train': 0.4588976204395294} 02/26/2022 03:06:23 - INFO - codeparrot_training - Step 38255: {'lr': 7.029770940785963e-05, 'samples': 19587072, 'steps': 38255, 'loss/train': 2.852640151977539} 02/26/2022 03:06:26 - INFO - codeparrot_training - Step 38256: {'lr': 7.028633448758421e-05, 'samples': 19587584, 'steps': 38256, 'loss/train': 1.6801319122314453} 02/26/2022 03:06:32 - INFO - codeparrot_training - Step 38257: {'lr': 7.027496033714498e-05, 'samples': 19588096, 'steps': 38257, 'loss/train': 1.0426446199417114} 02/26/2022 03:06:35 - INFO - codeparrot_training - Step 38258: {'lr': 7.026358695659082e-05, 'samples': 19588608, 'steps': 38258, 'loss/train': 2.1895763874053955} 02/26/2022 03:06:41 - INFO - codeparrot_training - Step 38259: {'lr': 7.025221434597037e-05, 'samples': 19589120, 'steps': 38259, 'loss/train': 1.5701831579208374} 02/26/2022 03:06:45 - INFO - codeparrot_training - Step 38260: {'lr': 7.024084250533239e-05, 'samples': 19589632, 'steps': 38260, 'loss/train': 0.7637211084365845} 02/26/2022 03:06:50 - INFO - codeparrot_training - Step 38261: {'lr': 7.022947143472547e-05, 'samples': 19590144, 'steps': 38261, 'loss/train': 1.3565819263458252} 02/26/2022 03:06:54 - INFO - codeparrot_training - Step 38262: {'lr': 7.02181011341985e-05, 'samples': 19590656, 'steps': 38262, 'loss/train': 2.545463800430298} 02/26/2022 03:06:59 - INFO - codeparrot_training - Step 38263: {'lr': 7.020673160380006e-05, 'samples': 19591168, 'steps': 38263, 'loss/train': 2.026808500289917} 02/26/2022 03:07:03 - INFO - codeparrot_training - Step 38264: {'lr': 7.019536284357891e-05, 'samples': 19591680, 'steps': 38264, 'loss/train': 1.3249893188476562} 02/26/2022 03:07:10 - INFO - codeparrot_training - Step 38265: {'lr': 7.018399485358367e-05, 'samples': 19592192, 'steps': 38265, 'loss/train': 1.671152114868164} 02/26/2022 03:07:13 - INFO - codeparrot_training - Step 38266: {'lr': 7.017262763386317e-05, 'samples': 19592704, 'steps': 38266, 'loss/train': 2.67478084564209} 02/26/2022 03:07:19 - INFO - codeparrot_training - Step 38267: {'lr': 7.016126118446603e-05, 'samples': 19593216, 'steps': 38267, 'loss/train': 1.779833197593689} 02/26/2022 03:07:22 - INFO - codeparrot_training - Step 38268: {'lr': 7.014989550544085e-05, 'samples': 19593728, 'steps': 38268, 'loss/train': 1.939889669418335} 02/26/2022 03:07:28 - INFO - codeparrot_training - Step 38269: {'lr': 7.013853059683655e-05, 'samples': 19594240, 'steps': 38269, 'loss/train': 2.837552785873413} 02/26/2022 03:07:32 - INFO - codeparrot_training - Step 38270: {'lr': 7.012716645870157e-05, 'samples': 19594752, 'steps': 38270, 'loss/train': 1.5576125383377075} 02/26/2022 03:07:37 - INFO - codeparrot_training - Step 38271: {'lr': 7.011580309108472e-05, 'samples': 19595264, 'steps': 38271, 'loss/train': 1.4813158512115479} 02/26/2022 03:07:41 - INFO - codeparrot_training - Step 38272: {'lr': 7.01044404940346e-05, 'samples': 19595776, 'steps': 38272, 'loss/train': 2.7505342960357666} 02/26/2022 03:07:46 - INFO - codeparrot_training - Step 38273: {'lr': 7.009307866760008e-05, 'samples': 19596288, 'steps': 38273, 'loss/train': 1.6960562467575073} 02/26/2022 03:07:50 - INFO - codeparrot_training - Step 38274: {'lr': 7.008171761182955e-05, 'samples': 19596800, 'steps': 38274, 'loss/train': 2.1555190086364746} 02/26/2022 03:07:55 - INFO - codeparrot_training - Step 38275: {'lr': 7.007035732677186e-05, 'samples': 19597312, 'steps': 38275, 'loss/train': 2.4800174236297607} 02/26/2022 03:07:59 - INFO - codeparrot_training - Step 38276: {'lr': 7.005899781247557e-05, 'samples': 19597824, 'steps': 38276, 'loss/train': 2.0562477111816406} 02/26/2022 03:08:04 - INFO - codeparrot_training - Step 38277: {'lr': 7.004763906898953e-05, 'samples': 19598336, 'steps': 38277, 'loss/train': 1.5696412324905396} 02/26/2022 03:08:08 - INFO - codeparrot_training - Step 38278: {'lr': 7.003628109636212e-05, 'samples': 19598848, 'steps': 38278, 'loss/train': 1.393474817276001} 02/26/2022 03:08:13 - INFO - codeparrot_training - Step 38279: {'lr': 7.00249238946422e-05, 'samples': 19599360, 'steps': 38279, 'loss/train': 1.925711989402771} 02/26/2022 03:08:17 - INFO - codeparrot_training - Step 38280: {'lr': 7.001356746387829e-05, 'samples': 19599872, 'steps': 38280, 'loss/train': 1.9334263801574707} 02/26/2022 03:08:22 - INFO - codeparrot_training - Step 38281: {'lr': 7.00022118041192e-05, 'samples': 19600384, 'steps': 38281, 'loss/train': 1.6700961589813232} 02/26/2022 03:08:26 - INFO - codeparrot_training - Step 38282: {'lr': 6.999085691541346e-05, 'samples': 19600896, 'steps': 38282, 'loss/train': 1.2895236015319824} 02/26/2022 03:08:33 - INFO - codeparrot_training - Step 38283: {'lr': 6.997950279780974e-05, 'samples': 19601408, 'steps': 38283, 'loss/train': 1.5613279342651367} 02/26/2022 03:08:36 - INFO - codeparrot_training - Step 38284: {'lr': 6.996814945135657e-05, 'samples': 19601920, 'steps': 38284, 'loss/train': 2.217952251434326} 02/26/2022 03:08:42 - INFO - codeparrot_training - Step 38285: {'lr': 6.995679687610276e-05, 'samples': 19602432, 'steps': 38285, 'loss/train': 2.027120351791382} 02/26/2022 03:08:45 - INFO - codeparrot_training - Step 38286: {'lr': 6.994544507209688e-05, 'samples': 19602944, 'steps': 38286, 'loss/train': 1.4986342191696167} 02/26/2022 03:08:51 - INFO - codeparrot_training - Step 38287: {'lr': 6.993409403938752e-05, 'samples': 19603456, 'steps': 38287, 'loss/train': 1.6712936162948608} 02/26/2022 03:08:54 - INFO - codeparrot_training - Step 38288: {'lr': 6.992274377802327e-05, 'samples': 19603968, 'steps': 38288, 'loss/train': 3.0750985145568848} 02/26/2022 03:09:00 - INFO - codeparrot_training - Step 38289: {'lr': 6.991139428805288e-05, 'samples': 19604480, 'steps': 38289, 'loss/train': 1.3052424192428589} 02/26/2022 03:09:03 - INFO - codeparrot_training - Step 38290: {'lr': 6.990004556952487e-05, 'samples': 19604992, 'steps': 38290, 'loss/train': 1.5871858596801758} 02/26/2022 03:09:09 - INFO - codeparrot_training - Step 38291: {'lr': 6.98886976224879e-05, 'samples': 19605504, 'steps': 38291, 'loss/train': 1.7583281993865967} 02/26/2022 03:09:12 - INFO - codeparrot_training - Step 38292: {'lr': 6.987735044699045e-05, 'samples': 19606016, 'steps': 38292, 'loss/train': 2.274233102798462} 02/26/2022 03:09:20 - INFO - codeparrot_training - Step 38293: {'lr': 6.986600404308132e-05, 'samples': 19606528, 'steps': 38293, 'loss/train': 2.299293041229248} 02/26/2022 03:09:23 - INFO - codeparrot_training - Step 38294: {'lr': 6.985465841080899e-05, 'samples': 19607040, 'steps': 38294, 'loss/train': 1.7122441530227661} 02/26/2022 03:09:29 - INFO - codeparrot_training - Step 38295: {'lr': 6.984331355022222e-05, 'samples': 19607552, 'steps': 38295, 'loss/train': 1.8874236345291138} 02/26/2022 03:09:32 - INFO - codeparrot_training - Step 38296: {'lr': 6.983196946136935e-05, 'samples': 19608064, 'steps': 38296, 'loss/train': 1.9548982381820679} 02/26/2022 03:09:38 - INFO - codeparrot_training - Step 38297: {'lr': 6.982062614429918e-05, 'samples': 19608576, 'steps': 38297, 'loss/train': 3.8726401329040527} 02/26/2022 03:09:41 - INFO - codeparrot_training - Step 38298: {'lr': 6.980928359906025e-05, 'samples': 19609088, 'steps': 38298, 'loss/train': 0.9677563905715942} 02/26/2022 03:09:47 - INFO - codeparrot_training - Step 38299: {'lr': 6.979794182570112e-05, 'samples': 19609600, 'steps': 38299, 'loss/train': 1.4185266494750977} 02/26/2022 03:09:50 - INFO - codeparrot_training - Step 38300: {'lr': 6.97866008242703e-05, 'samples': 19610112, 'steps': 38300, 'loss/train': 1.8968249559402466} 02/26/2022 03:09:56 - INFO - codeparrot_training - Step 38301: {'lr': 6.977526059481653e-05, 'samples': 19610624, 'steps': 38301, 'loss/train': 7.962368488311768} 02/26/2022 03:10:00 - INFO - codeparrot_training - Step 38302: {'lr': 6.976392113738833e-05, 'samples': 19611136, 'steps': 38302, 'loss/train': 2.0363941192626953} 02/26/2022 03:10:05 - INFO - codeparrot_training - Step 38303: {'lr': 6.97525824520342e-05, 'samples': 19611648, 'steps': 38303, 'loss/train': 1.7641922235488892} 02/26/2022 03:10:09 - INFO - codeparrot_training - Step 38304: {'lr': 6.974124453880282e-05, 'samples': 19612160, 'steps': 38304, 'loss/train': 1.8338292837142944} 02/26/2022 03:10:14 - INFO - codeparrot_training - Step 38305: {'lr': 6.972990739774274e-05, 'samples': 19612672, 'steps': 38305, 'loss/train': 1.7249993085861206} 02/26/2022 03:10:18 - INFO - codeparrot_training - Step 38306: {'lr': 6.971857102890247e-05, 'samples': 19613184, 'steps': 38306, 'loss/train': 1.5775387287139893} 02/26/2022 03:10:23 - INFO - codeparrot_training - Step 38307: {'lr': 6.97072354323305e-05, 'samples': 19613696, 'steps': 38307, 'loss/train': 1.1379903554916382} 02/26/2022 03:10:27 - INFO - codeparrot_training - Step 38308: {'lr': 6.969590060807563e-05, 'samples': 19614208, 'steps': 38308, 'loss/train': 1.80392324924469} 02/26/2022 03:10:32 - INFO - codeparrot_training - Step 38309: {'lr': 6.968456655618621e-05, 'samples': 19614720, 'steps': 38309, 'loss/train': 1.4866101741790771} 02/26/2022 03:10:36 - INFO - codeparrot_training - Step 38310: {'lr': 6.967323327671085e-05, 'samples': 19615232, 'steps': 38310, 'loss/train': 0.8231110572814941} 02/26/2022 03:10:41 - INFO - codeparrot_training - Step 38311: {'lr': 6.966190076969806e-05, 'samples': 19615744, 'steps': 38311, 'loss/train': 1.5180293321609497} 02/26/2022 03:10:45 - INFO - codeparrot_training - Step 38312: {'lr': 6.965056903519648e-05, 'samples': 19616256, 'steps': 38312, 'loss/train': 1.631626844406128} 02/26/2022 03:10:51 - INFO - codeparrot_training - Step 38313: {'lr': 6.96392380732546e-05, 'samples': 19616768, 'steps': 38313, 'loss/train': 1.9530813694000244} 02/26/2022 03:10:56 - INFO - codeparrot_training - Step 38314: {'lr': 6.962790788392093e-05, 'samples': 19617280, 'steps': 38314, 'loss/train': 1.4769620895385742} 02/26/2022 03:11:00 - INFO - codeparrot_training - Step 38315: {'lr': 6.961657846724395e-05, 'samples': 19617792, 'steps': 38315, 'loss/train': 2.2006146907806396} 02/26/2022 03:11:05 - INFO - codeparrot_training - Step 38316: {'lr': 6.960524982327238e-05, 'samples': 19618304, 'steps': 38316, 'loss/train': 1.8449044227600098} 02/26/2022 03:11:09 - INFO - codeparrot_training - Step 38317: {'lr': 6.959392195205461e-05, 'samples': 19618816, 'steps': 38317, 'loss/train': 1.8796827793121338} 02/26/2022 03:11:12 - INFO - codeparrot_training - Step 38318: {'lr': 6.958259485363922e-05, 'samples': 19619328, 'steps': 38318, 'loss/train': 0.7383489608764648} 02/26/2022 03:11:18 - INFO - codeparrot_training - Step 38319: {'lr': 6.95712685280746e-05, 'samples': 19619840, 'steps': 38319, 'loss/train': 1.6463539600372314} 02/26/2022 03:11:21 - INFO - codeparrot_training - Step 38320: {'lr': 6.955994297540946e-05, 'samples': 19620352, 'steps': 38320, 'loss/train': 2.559006690979004} 02/26/2022 03:11:27 - INFO - codeparrot_training - Step 38321: {'lr': 6.954861819569225e-05, 'samples': 19620864, 'steps': 38321, 'loss/train': 1.0560564994812012} 02/26/2022 03:11:30 - INFO - codeparrot_training - Step 38322: {'lr': 6.953729418897142e-05, 'samples': 19621376, 'steps': 38322, 'loss/train': 1.7350120544433594} 02/26/2022 03:11:36 - INFO - codeparrot_training - Step 38323: {'lr': 6.952597095529548e-05, 'samples': 19621888, 'steps': 38323, 'loss/train': 0.9742646217346191} 02/26/2022 03:11:42 - INFO - codeparrot_training - Step 38324: {'lr': 6.9514648494713e-05, 'samples': 19622400, 'steps': 38324, 'loss/train': 2.3117151260375977} 02/26/2022 03:11:45 - INFO - codeparrot_training - Step 38325: {'lr': 6.95033268072725e-05, 'samples': 19622912, 'steps': 38325, 'loss/train': 1.4532585144042969} 02/26/2022 03:11:51 - INFO - codeparrot_training - Step 38326: {'lr': 6.949200589302243e-05, 'samples': 19623424, 'steps': 38326, 'loss/train': 1.5678054094314575} 02/26/2022 03:11:54 - INFO - codeparrot_training - Step 38327: {'lr': 6.948068575201118e-05, 'samples': 19623936, 'steps': 38327, 'loss/train': 0.8305355906486511} 02/26/2022 03:12:01 - INFO - codeparrot_training - Step 38328: {'lr': 6.946936638428747e-05, 'samples': 19624448, 'steps': 38328, 'loss/train': 1.9675689935684204} 02/26/2022 03:12:04 - INFO - codeparrot_training - Step 38329: {'lr': 6.945804778989959e-05, 'samples': 19624960, 'steps': 38329, 'loss/train': 1.5720771551132202} 02/26/2022 03:12:08 - INFO - codeparrot_training - Step 38330: {'lr': 6.944672996889623e-05, 'samples': 19625472, 'steps': 38330, 'loss/train': 2.2868762016296387} 02/26/2022 03:12:13 - INFO - codeparrot_training - Step 38331: {'lr': 6.94354129213256e-05, 'samples': 19625984, 'steps': 38331, 'loss/train': 1.3972021341323853} 02/26/2022 03:12:17 - INFO - codeparrot_training - Step 38332: {'lr': 6.94240966472364e-05, 'samples': 19626496, 'steps': 38332, 'loss/train': 1.7493994235992432} 02/26/2022 03:12:22 - INFO - codeparrot_training - Step 38333: {'lr': 6.941278114667696e-05, 'samples': 19627008, 'steps': 38333, 'loss/train': 1.1633567810058594} 02/26/2022 03:12:26 - INFO - codeparrot_training - Step 38334: {'lr': 6.940146641969599e-05, 'samples': 19627520, 'steps': 38334, 'loss/train': 0.05693844333291054} 02/26/2022 03:12:31 - INFO - codeparrot_training - Step 38335: {'lr': 6.939015246634162e-05, 'samples': 19628032, 'steps': 38335, 'loss/train': 1.7946480512619019} 02/26/2022 03:12:35 - INFO - codeparrot_training - Step 38336: {'lr': 6.937883928666256e-05, 'samples': 19628544, 'steps': 38336, 'loss/train': 2.596149444580078} 02/26/2022 03:12:40 - INFO - codeparrot_training - Step 38337: {'lr': 6.936752688070713e-05, 'samples': 19629056, 'steps': 38337, 'loss/train': 1.6972674131393433} 02/26/2022 03:12:44 - INFO - codeparrot_training - Step 38338: {'lr': 6.9356215248524e-05, 'samples': 19629568, 'steps': 38338, 'loss/train': 2.2823193073272705} 02/26/2022 03:12:51 - INFO - codeparrot_training - Step 38339: {'lr': 6.934490439016136e-05, 'samples': 19630080, 'steps': 38339, 'loss/train': 1.9745047092437744} 02/26/2022 03:12:54 - INFO - codeparrot_training - Step 38340: {'lr': 6.933359430566785e-05, 'samples': 19630592, 'steps': 38340, 'loss/train': 2.5512940883636475} 02/26/2022 03:13:00 - INFO - codeparrot_training - Step 38341: {'lr': 6.932228499509177e-05, 'samples': 19631104, 'steps': 38341, 'loss/train': 1.4923121929168701} 02/26/2022 03:13:03 - INFO - codeparrot_training - Step 38342: {'lr': 6.931097645848172e-05, 'samples': 19631616, 'steps': 38342, 'loss/train': 1.9151735305786133} 02/26/2022 03:13:09 - INFO - codeparrot_training - Step 38343: {'lr': 6.929966869588609e-05, 'samples': 19632128, 'steps': 38343, 'loss/train': 2.281074047088623} 02/26/2022 03:13:12 - INFO - codeparrot_training - Step 38344: {'lr': 6.928836170735328e-05, 'samples': 19632640, 'steps': 38344, 'loss/train': 1.8603179454803467} 02/26/2022 03:13:18 - INFO - codeparrot_training - Step 38345: {'lr': 6.927705549293167e-05, 'samples': 19633152, 'steps': 38345, 'loss/train': 1.8410143852233887} 02/26/2022 03:13:21 - INFO - codeparrot_training - Step 38346: {'lr': 6.926575005266986e-05, 'samples': 19633664, 'steps': 38346, 'loss/train': 2.4325764179229736} 02/26/2022 03:13:27 - INFO - codeparrot_training - Step 38347: {'lr': 6.925444538661618e-05, 'samples': 19634176, 'steps': 38347, 'loss/train': 1.8347617387771606} 02/26/2022 03:13:30 - INFO - codeparrot_training - Step 38348: {'lr': 6.924314149481905e-05, 'samples': 19634688, 'steps': 38348, 'loss/train': 1.1782866716384888} 02/26/2022 03:13:37 - INFO - codeparrot_training - Step 38349: {'lr': 6.923183837732694e-05, 'samples': 19635200, 'steps': 38349, 'loss/train': 1.8457951545715332} 02/26/2022 03:13:40 - INFO - codeparrot_training - Step 38350: {'lr': 6.922053603418812e-05, 'samples': 19635712, 'steps': 38350, 'loss/train': 2.0619657039642334} 02/26/2022 03:13:46 - INFO - codeparrot_training - Step 38351: {'lr': 6.920923446545121e-05, 'samples': 19636224, 'steps': 38351, 'loss/train': 2.1893832683563232} 02/26/2022 03:13:49 - INFO - codeparrot_training - Step 38352: {'lr': 6.919793367116453e-05, 'samples': 19636736, 'steps': 38352, 'loss/train': 1.5401129722595215} 02/26/2022 03:13:55 - INFO - codeparrot_training - Step 38353: {'lr': 6.918663365137651e-05, 'samples': 19637248, 'steps': 38353, 'loss/train': 0.9590820074081421} 02/26/2022 03:13:58 - INFO - codeparrot_training - Step 38354: {'lr': 6.917533440613543e-05, 'samples': 19637760, 'steps': 38354, 'loss/train': 1.5230873823165894} 02/26/2022 03:14:04 - INFO - codeparrot_training - Step 38355: {'lr': 6.91640359354899e-05, 'samples': 19638272, 'steps': 38355, 'loss/train': 1.3341290950775146} 02/26/2022 03:14:07 - INFO - codeparrot_training - Step 38356: {'lr': 6.91527382394882e-05, 'samples': 19638784, 'steps': 38356, 'loss/train': 1.6489259004592896} 02/26/2022 03:14:13 - INFO - codeparrot_training - Step 38357: {'lr': 6.914144131817876e-05, 'samples': 19639296, 'steps': 38357, 'loss/train': 1.5889368057250977} 02/26/2022 03:14:16 - INFO - codeparrot_training - Step 38358: {'lr': 6.913014517160986e-05, 'samples': 19639808, 'steps': 38358, 'loss/train': 2.0269086360931396} 02/26/2022 03:14:23 - INFO - codeparrot_training - Step 38359: {'lr': 6.911884979983007e-05, 'samples': 19640320, 'steps': 38359, 'loss/train': 2.6325814723968506} 02/26/2022 03:14:26 - INFO - codeparrot_training - Step 38360: {'lr': 6.91075552028877e-05, 'samples': 19640832, 'steps': 38360, 'loss/train': 1.1975805759429932} 02/26/2022 03:14:31 - INFO - codeparrot_training - Step 38361: {'lr': 6.909626138083114e-05, 'samples': 19641344, 'steps': 38361, 'loss/train': 1.8384859561920166} 02/26/2022 03:14:35 - INFO - codeparrot_training - Step 38362: {'lr': 6.908496833370865e-05, 'samples': 19641856, 'steps': 38362, 'loss/train': 3.091197967529297} 02/26/2022 03:14:41 - INFO - codeparrot_training - Step 38363: {'lr': 6.907367606156878e-05, 'samples': 19642368, 'steps': 38363, 'loss/train': 0.1851411908864975} 02/26/2022 03:14:44 - INFO - codeparrot_training - Step 38364: {'lr': 6.906238456445977e-05, 'samples': 19642880, 'steps': 38364, 'loss/train': 2.2467594146728516} 02/26/2022 03:14:50 - INFO - codeparrot_training - Step 38365: {'lr': 6.905109384243019e-05, 'samples': 19643392, 'steps': 38365, 'loss/train': 1.6727991104125977} 02/26/2022 03:14:53 - INFO - codeparrot_training - Step 38366: {'lr': 6.903980389552814e-05, 'samples': 19643904, 'steps': 38366, 'loss/train': 2.3676037788391113} 02/26/2022 03:14:59 - INFO - codeparrot_training - Step 38367: {'lr': 6.902851472380215e-05, 'samples': 19644416, 'steps': 38367, 'loss/train': 1.8005988597869873} 02/26/2022 03:15:02 - INFO - codeparrot_training - Step 38368: {'lr': 6.901722632730048e-05, 'samples': 19644928, 'steps': 38368, 'loss/train': 1.013195514678955} 02/26/2022 03:15:08 - INFO - codeparrot_training - Step 38369: {'lr': 6.900593870607169e-05, 'samples': 19645440, 'steps': 38369, 'loss/train': 1.7053675651550293} 02/26/2022 03:15:12 - INFO - codeparrot_training - Step 38370: {'lr': 6.899465186016385e-05, 'samples': 19645952, 'steps': 38370, 'loss/train': 0.2534477710723877} 02/26/2022 03:15:17 - INFO - codeparrot_training - Step 38371: {'lr': 6.898336578962549e-05, 'samples': 19646464, 'steps': 38371, 'loss/train': 0.800690233707428} 02/26/2022 03:15:21 - INFO - codeparrot_training - Step 38372: {'lr': 6.897208049450488e-05, 'samples': 19646976, 'steps': 38372, 'loss/train': 0.7948448061943054} 02/26/2022 03:15:26 - INFO - codeparrot_training - Step 38373: {'lr': 6.896079597485053e-05, 'samples': 19647488, 'steps': 38373, 'loss/train': 1.2068716287612915} 02/26/2022 03:15:30 - INFO - codeparrot_training - Step 38374: {'lr': 6.894951223071046e-05, 'samples': 19648000, 'steps': 38374, 'loss/train': 0.7735446691513062} 02/26/2022 03:15:35 - INFO - codeparrot_training - Step 38375: {'lr': 6.893822926213331e-05, 'samples': 19648512, 'steps': 38375, 'loss/train': 1.6190383434295654} 02/26/2022 03:15:39 - INFO - codeparrot_training - Step 38376: {'lr': 6.892694706916719e-05, 'samples': 19649024, 'steps': 38376, 'loss/train': 0.7150295972824097} 02/26/2022 03:15:44 - INFO - codeparrot_training - Step 38377: {'lr': 6.89156656518606e-05, 'samples': 19649536, 'steps': 38377, 'loss/train': 1.3908824920654297} 02/26/2022 03:15:48 - INFO - codeparrot_training - Step 38378: {'lr': 6.890438501026183e-05, 'samples': 19650048, 'steps': 38378, 'loss/train': 2.4899752140045166} 02/26/2022 03:15:53 - INFO - codeparrot_training - Step 38379: {'lr': 6.889310514441915e-05, 'samples': 19650560, 'steps': 38379, 'loss/train': 0.7606871724128723} 02/26/2022 03:15:57 - INFO - codeparrot_training - Step 38380: {'lr': 6.888182605438082e-05, 'samples': 19651072, 'steps': 38380, 'loss/train': 0.705439567565918} 02/26/2022 03:16:03 - INFO - codeparrot_training - Step 38381: {'lr': 6.887054774019533e-05, 'samples': 19651584, 'steps': 38381, 'loss/train': 4.246715545654297} 02/26/2022 03:16:07 - INFO - codeparrot_training - Step 38382: {'lr': 6.885927020191088e-05, 'samples': 19652096, 'steps': 38382, 'loss/train': 8.242523193359375} 02/26/2022 03:16:12 - INFO - codeparrot_training - Step 38383: {'lr': 6.884799343957579e-05, 'samples': 19652608, 'steps': 38383, 'loss/train': 2.509084463119507} 02/26/2022 03:16:16 - INFO - codeparrot_training - Step 38384: {'lr': 6.883671745323833e-05, 'samples': 19653120, 'steps': 38384, 'loss/train': 1.7395373582839966} 02/26/2022 03:16:22 - INFO - codeparrot_training - Step 38385: {'lr': 6.88254422429469e-05, 'samples': 19653632, 'steps': 38385, 'loss/train': 1.5212839841842651} 02/26/2022 03:16:25 - INFO - codeparrot_training - Step 38386: {'lr': 6.881416780874975e-05, 'samples': 19654144, 'steps': 38386, 'loss/train': 0.922737717628479} 02/26/2022 03:16:31 - INFO - codeparrot_training - Step 38387: {'lr': 6.880289415069516e-05, 'samples': 19654656, 'steps': 38387, 'loss/train': 1.8092321157455444} 02/26/2022 03:16:34 - INFO - codeparrot_training - Step 38388: {'lr': 6.879162126883137e-05, 'samples': 19655168, 'steps': 38388, 'loss/train': 1.829150676727295} 02/26/2022 03:16:40 - INFO - codeparrot_training - Step 38389: {'lr': 6.878034916320683e-05, 'samples': 19655680, 'steps': 38389, 'loss/train': 2.053861379623413} 02/26/2022 03:16:44 - INFO - codeparrot_training - Step 38390: {'lr': 6.876907783386971e-05, 'samples': 19656192, 'steps': 38390, 'loss/train': 1.4137603044509888} 02/26/2022 03:16:49 - INFO - codeparrot_training - Step 38391: {'lr': 6.875780728086831e-05, 'samples': 19656704, 'steps': 38391, 'loss/train': 1.724225640296936} 02/26/2022 03:16:53 - INFO - codeparrot_training - Step 38392: {'lr': 6.874653750425083e-05, 'samples': 19657216, 'steps': 38392, 'loss/train': 2.4909112453460693} 02/26/2022 03:16:58 - INFO - codeparrot_training - Step 38393: {'lr': 6.873526850406573e-05, 'samples': 19657728, 'steps': 38393, 'loss/train': 0.9760668277740479} 02/26/2022 03:17:02 - INFO - codeparrot_training - Step 38394: {'lr': 6.872400028036116e-05, 'samples': 19658240, 'steps': 38394, 'loss/train': 2.6618876457214355} 02/26/2022 03:17:08 - INFO - codeparrot_training - Step 38395: {'lr': 6.871273283318541e-05, 'samples': 19658752, 'steps': 38395, 'loss/train': 2.468435049057007} 02/26/2022 03:17:11 - INFO - codeparrot_training - Step 38396: {'lr': 6.870146616258677e-05, 'samples': 19659264, 'steps': 38396, 'loss/train': 1.029135823249817} 02/26/2022 03:17:17 - INFO - codeparrot_training - Step 38397: {'lr': 6.869020026861339e-05, 'samples': 19659776, 'steps': 38397, 'loss/train': 2.4733827114105225} 02/26/2022 03:17:20 - INFO - codeparrot_training - Step 38398: {'lr': 6.867893515131368e-05, 'samples': 19660288, 'steps': 38398, 'loss/train': 0.9131062030792236} 02/26/2022 03:17:26 - INFO - codeparrot_training - Step 38399: {'lr': 6.866767081073586e-05, 'samples': 19660800, 'steps': 38399, 'loss/train': 1.0162612199783325} 02/26/2022 03:17:29 - INFO - codeparrot_training - Step 38400: {'lr': 6.865640724692814e-05, 'samples': 19661312, 'steps': 38400, 'loss/train': 1.7749122381210327} 02/26/2022 03:17:35 - INFO - codeparrot_training - Step 38401: {'lr': 6.864514445993875e-05, 'samples': 19661824, 'steps': 38401, 'loss/train': 1.9093453884124756} 02/26/2022 03:17:38 - INFO - codeparrot_training - Step 38402: {'lr': 6.863388244981602e-05, 'samples': 19662336, 'steps': 38402, 'loss/train': 2.0899407863616943} 02/26/2022 03:17:44 - INFO - codeparrot_training - Step 38403: {'lr': 6.862262121660806e-05, 'samples': 19662848, 'steps': 38403, 'loss/train': 2.282686233520508} 02/26/2022 03:17:47 - INFO - codeparrot_training - Step 38404: {'lr': 6.861136076036337e-05, 'samples': 19663360, 'steps': 38404, 'loss/train': 2.10834002494812} 02/26/2022 03:17:51 - INFO - codeparrot_training - Step 38405: {'lr': 6.860010108112985e-05, 'samples': 19663872, 'steps': 38405, 'loss/train': 2.057204008102417} 02/26/2022 03:17:57 - INFO - codeparrot_training - Step 38406: {'lr': 6.8588842178956e-05, 'samples': 19664384, 'steps': 38406, 'loss/train': 1.6700900793075562} 02/26/2022 03:18:01 - INFO - codeparrot_training - Step 38407: {'lr': 6.857758405388984e-05, 'samples': 19664896, 'steps': 38407, 'loss/train': 2.088494300842285} 02/26/2022 03:18:06 - INFO - codeparrot_training - Step 38408: {'lr': 6.856632670597988e-05, 'samples': 19665408, 'steps': 38408, 'loss/train': 1.6609251499176025} 02/26/2022 03:18:10 - INFO - codeparrot_training - Step 38409: {'lr': 6.855507013527398e-05, 'samples': 19665920, 'steps': 38409, 'loss/train': 1.5301356315612793} 02/26/2022 03:18:15 - INFO - codeparrot_training - Step 38410: {'lr': 6.854381434182063e-05, 'samples': 19666432, 'steps': 38410, 'loss/train': 1.9033960103988647} 02/26/2022 03:18:19 - INFO - codeparrot_training - Step 38411: {'lr': 6.853255932566788e-05, 'samples': 19666944, 'steps': 38411, 'loss/train': 1.6113755702972412} 02/26/2022 03:18:24 - INFO - codeparrot_training - Step 38412: {'lr': 6.85213050868641e-05, 'samples': 19667456, 'steps': 38412, 'loss/train': 2.4330015182495117} 02/26/2022 03:18:30 - INFO - codeparrot_training - Step 38413: {'lr': 6.851005162545746e-05, 'samples': 19667968, 'steps': 38413, 'loss/train': 2.0863029956817627} 02/26/2022 03:18:33 - INFO - codeparrot_training - Step 38414: {'lr': 6.849879894149607e-05, 'samples': 19668480, 'steps': 38414, 'loss/train': 1.26804518699646} 02/26/2022 03:18:39 - INFO - codeparrot_training - Step 38415: {'lr': 6.848754703502815e-05, 'samples': 19668992, 'steps': 38415, 'loss/train': 0.8444098234176636} 02/26/2022 03:18:42 - INFO - codeparrot_training - Step 38416: {'lr': 6.847629590610202e-05, 'samples': 19669504, 'steps': 38416, 'loss/train': 0.7023478746414185} 02/26/2022 03:18:49 - INFO - codeparrot_training - Step 38417: {'lr': 6.846504555476579e-05, 'samples': 19670016, 'steps': 38417, 'loss/train': 1.8845778703689575} 02/26/2022 03:18:52 - INFO - codeparrot_training - Step 38418: {'lr': 6.845379598106766e-05, 'samples': 19670528, 'steps': 38418, 'loss/train': 2.440739393234253} 02/26/2022 03:18:58 - INFO - codeparrot_training - Step 38419: {'lr': 6.844254718505574e-05, 'samples': 19671040, 'steps': 38419, 'loss/train': 1.5760228633880615} 02/26/2022 03:19:01 - INFO - codeparrot_training - Step 38420: {'lr': 6.84312991667784e-05, 'samples': 19671552, 'steps': 38420, 'loss/train': 1.8081988096237183} 02/26/2022 03:19:07 - INFO - codeparrot_training - Step 38421: {'lr': 6.84200519262837e-05, 'samples': 19672064, 'steps': 38421, 'loss/train': 1.5852195024490356} 02/26/2022 03:19:10 - INFO - codeparrot_training - Step 38422: {'lr': 6.840880546361983e-05, 'samples': 19672576, 'steps': 38422, 'loss/train': 2.1679470539093018} 02/26/2022 03:19:16 - INFO - codeparrot_training - Step 38423: {'lr': 6.839755977883491e-05, 'samples': 19673088, 'steps': 38423, 'loss/train': 1.893385410308838} 02/26/2022 03:19:19 - INFO - codeparrot_training - Step 38424: {'lr': 6.838631487197728e-05, 'samples': 19673600, 'steps': 38424, 'loss/train': 1.641092300415039} 02/26/2022 03:19:24 - INFO - codeparrot_training - Step 38425: {'lr': 6.837507074309488e-05, 'samples': 19674112, 'steps': 38425, 'loss/train': 1.9038786888122559} 02/26/2022 03:19:28 - INFO - codeparrot_training - Step 38426: {'lr': 6.836382739223618e-05, 'samples': 19674624, 'steps': 38426, 'loss/train': 1.2102866172790527} 02/26/2022 03:19:34 - INFO - codeparrot_training - Step 38427: {'lr': 6.835258481944904e-05, 'samples': 19675136, 'steps': 38427, 'loss/train': 2.333507537841797} 02/26/2022 03:19:38 - INFO - codeparrot_training - Step 38428: {'lr': 6.834134302478179e-05, 'samples': 19675648, 'steps': 38428, 'loss/train': 2.0154917240142822} 02/26/2022 03:19:43 - INFO - codeparrot_training - Step 38429: {'lr': 6.833010200828255e-05, 'samples': 19676160, 'steps': 38429, 'loss/train': 0.6300233602523804} 02/26/2022 03:19:46 - INFO - codeparrot_training - Step 38430: {'lr': 6.831886176999947e-05, 'samples': 19676672, 'steps': 38430, 'loss/train': 1.9117789268493652} 02/26/2022 03:19:52 - INFO - codeparrot_training - Step 38431: {'lr': 6.830762230998064e-05, 'samples': 19677184, 'steps': 38431, 'loss/train': 1.068660855293274} 02/26/2022 03:19:55 - INFO - codeparrot_training - Step 38432: {'lr': 6.829638362827431e-05, 'samples': 19677696, 'steps': 38432, 'loss/train': 2.11419415473938} 02/26/2022 03:20:01 - INFO - codeparrot_training - Step 38433: {'lr': 6.828514572492859e-05, 'samples': 19678208, 'steps': 38433, 'loss/train': 1.270612359046936} 02/26/2022 03:20:04 - INFO - codeparrot_training - Step 38434: {'lr': 6.827390859999158e-05, 'samples': 19678720, 'steps': 38434, 'loss/train': 1.7337863445281982} 02/26/2022 03:20:10 - INFO - codeparrot_training - Step 38435: {'lr': 6.826267225351141e-05, 'samples': 19679232, 'steps': 38435, 'loss/train': 1.9830979108810425} 02/26/2022 03:20:13 - INFO - codeparrot_training - Step 38436: {'lr': 6.825143668553626e-05, 'samples': 19679744, 'steps': 38436, 'loss/train': 2.1001436710357666} 02/26/2022 03:20:19 - INFO - codeparrot_training - Step 38437: {'lr': 6.82402018961143e-05, 'samples': 19680256, 'steps': 38437, 'loss/train': 1.7360002994537354} 02/26/2022 03:20:22 - INFO - codeparrot_training - Step 38438: {'lr': 6.82289678852935e-05, 'samples': 19680768, 'steps': 38438, 'loss/train': 1.8273926973342896} 02/26/2022 03:20:28 - INFO - codeparrot_training - Step 38439: {'lr': 6.821773465312216e-05, 'samples': 19681280, 'steps': 38439, 'loss/train': 1.8329812288284302} 02/26/2022 03:20:31 - INFO - codeparrot_training - Step 38440: {'lr': 6.820650219964833e-05, 'samples': 19681792, 'steps': 38440, 'loss/train': 0.7322726249694824} 02/26/2022 03:20:37 - INFO - codeparrot_training - Step 38441: {'lr': 6.819527052492011e-05, 'samples': 19682304, 'steps': 38441, 'loss/train': 0.5114310383796692} 02/26/2022 03:20:41 - INFO - codeparrot_training - Step 38442: {'lr': 6.818403962898553e-05, 'samples': 19682816, 'steps': 38442, 'loss/train': 1.5200272798538208} 02/26/2022 03:20:46 - INFO - codeparrot_training - Step 38443: {'lr': 6.817280951189298e-05, 'samples': 19683328, 'steps': 38443, 'loss/train': 1.7517212629318237} 02/26/2022 03:20:50 - INFO - codeparrot_training - Step 38444: {'lr': 6.81615801736902e-05, 'samples': 19683840, 'steps': 38444, 'loss/train': 0.21692833304405212} 02/26/2022 03:20:55 - INFO - codeparrot_training - Step 38445: {'lr': 6.815035161442554e-05, 'samples': 19684352, 'steps': 38445, 'loss/train': 1.6718298196792603} 02/26/2022 03:20:59 - INFO - codeparrot_training - Step 38446: {'lr': 6.813912383414698e-05, 'samples': 19684864, 'steps': 38446, 'loss/train': 2.0770835876464844} 02/26/2022 03:21:04 - INFO - codeparrot_training - Step 38447: {'lr': 6.81278968329028e-05, 'samples': 19685376, 'steps': 38447, 'loss/train': 0.7157784104347229} 02/26/2022 03:21:08 - INFO - codeparrot_training - Step 38448: {'lr': 6.811667061074082e-05, 'samples': 19685888, 'steps': 38448, 'loss/train': 0.8628271818161011} 02/26/2022 03:21:13 - INFO - codeparrot_training - Step 38449: {'lr': 6.810544516770934e-05, 'samples': 19686400, 'steps': 38449, 'loss/train': 1.2255158424377441} 02/26/2022 03:21:19 - INFO - codeparrot_training - Step 38450: {'lr': 6.809422050385628e-05, 'samples': 19686912, 'steps': 38450, 'loss/train': 2.3205976486206055} 02/26/2022 03:21:22 - INFO - codeparrot_training - Step 38451: {'lr': 6.80829966192299e-05, 'samples': 19687424, 'steps': 38451, 'loss/train': 1.3617819547653198} 02/26/2022 03:21:29 - INFO - codeparrot_training - Step 38452: {'lr': 6.807177351387823e-05, 'samples': 19687936, 'steps': 38452, 'loss/train': 2.2382023334503174} 02/26/2022 03:21:32 - INFO - codeparrot_training - Step 38453: {'lr': 6.806055118784926e-05, 'samples': 19688448, 'steps': 38453, 'loss/train': 2.1060197353363037} 02/26/2022 03:21:37 - INFO - codeparrot_training - Step 38454: {'lr': 6.804932964119109e-05, 'samples': 19688960, 'steps': 38454, 'loss/train': 2.0379011631011963} 02/26/2022 03:21:41 - INFO - codeparrot_training - Step 38455: {'lr': 6.803810887395187e-05, 'samples': 19689472, 'steps': 38455, 'loss/train': 2.00443696975708} 02/26/2022 03:21:46 - INFO - codeparrot_training - Step 38456: {'lr': 6.802688888617962e-05, 'samples': 19689984, 'steps': 38456, 'loss/train': 2.0512964725494385} 02/26/2022 03:21:50 - INFO - codeparrot_training - Step 38457: {'lr': 6.801566967792239e-05, 'samples': 19690496, 'steps': 38457, 'loss/train': 1.3247833251953125} 02/26/2022 03:21:56 - INFO - codeparrot_training - Step 38458: {'lr': 6.800445124922816e-05, 'samples': 19691008, 'steps': 38458, 'loss/train': 0.9699744582176208} 02/26/2022 03:21:59 - INFO - codeparrot_training - Step 38459: {'lr': 6.799323360014514e-05, 'samples': 19691520, 'steps': 38459, 'loss/train': 1.271215796470642} 02/26/2022 03:22:02 - INFO - codeparrot_training - Step 38460: {'lr': 6.798201673072127e-05, 'samples': 19692032, 'steps': 38460, 'loss/train': 1.1692957878112793} 02/26/2022 03:22:08 - INFO - codeparrot_training - Step 38461: {'lr': 6.797080064100478e-05, 'samples': 19692544, 'steps': 38461, 'loss/train': 3.000379800796509} 02/26/2022 03:22:14 - INFO - codeparrot_training - Step 38462: {'lr': 6.79595853310434e-05, 'samples': 19693056, 'steps': 38462, 'loss/train': 2.173128843307495} 02/26/2022 03:22:18 - INFO - codeparrot_training - Step 38463: {'lr': 6.794837080088545e-05, 'samples': 19693568, 'steps': 38463, 'loss/train': 1.9213277101516724} 02/26/2022 03:22:21 - INFO - codeparrot_training - Step 38464: {'lr': 6.793715705057882e-05, 'samples': 19694080, 'steps': 38464, 'loss/train': 2.5000038146972656} 02/26/2022 03:22:27 - INFO - codeparrot_training - Step 38465: {'lr': 6.792594408017172e-05, 'samples': 19694592, 'steps': 38465, 'loss/train': 2.22418212890625} 02/26/2022 03:22:30 - INFO - codeparrot_training - Step 38466: {'lr': 6.791473188971192e-05, 'samples': 19695104, 'steps': 38466, 'loss/train': 2.945513963699341} 02/26/2022 03:22:36 - INFO - codeparrot_training - Step 38467: {'lr': 6.790352047924766e-05, 'samples': 19695616, 'steps': 38467, 'loss/train': 1.5484987497329712} 02/26/2022 03:22:39 - INFO - codeparrot_training - Step 38468: {'lr': 6.789230984882683e-05, 'samples': 19696128, 'steps': 38468, 'loss/train': 2.054885149002075} 02/26/2022 03:22:45 - INFO - codeparrot_training - Step 38469: {'lr': 6.788109999849765e-05, 'samples': 19696640, 'steps': 38469, 'loss/train': 1.6869285106658936} 02/26/2022 03:22:48 - INFO - codeparrot_training - Step 38470: {'lr': 6.786989092830786e-05, 'samples': 19697152, 'steps': 38470, 'loss/train': 1.016160249710083} 02/26/2022 03:22:54 - INFO - codeparrot_training - Step 38471: {'lr': 6.785868263830572e-05, 'samples': 19697664, 'steps': 38471, 'loss/train': 1.838807463645935} 02/26/2022 03:22:57 - INFO - codeparrot_training - Step 38472: {'lr': 6.784747512853905e-05, 'samples': 19698176, 'steps': 38472, 'loss/train': 1.3608500957489014} 02/26/2022 03:23:04 - INFO - codeparrot_training - Step 38473: {'lr': 6.783626839905605e-05, 'samples': 19698688, 'steps': 38473, 'loss/train': 2.1875054836273193} 02/26/2022 03:23:07 - INFO - codeparrot_training - Step 38474: {'lr': 6.782506244990463e-05, 'samples': 19699200, 'steps': 38474, 'loss/train': 1.6835252046585083} 02/26/2022 03:23:13 - INFO - codeparrot_training - Step 38475: {'lr': 6.78138572811328e-05, 'samples': 19699712, 'steps': 38475, 'loss/train': 0.9680857062339783} 02/26/2022 03:23:16 - INFO - codeparrot_training - Step 38476: {'lr': 6.780265289278852e-05, 'samples': 19700224, 'steps': 38476, 'loss/train': 1.007615089416504} 02/26/2022 03:23:22 - INFO - codeparrot_training - Step 38477: {'lr': 6.779144928491979e-05, 'samples': 19700736, 'steps': 38477, 'loss/train': 0.8837943077087402} 02/26/2022 03:23:25 - INFO - codeparrot_training - Step 38478: {'lr': 6.778024645757469e-05, 'samples': 19701248, 'steps': 38478, 'loss/train': 2.108769655227661} 02/26/2022 03:23:31 - INFO - codeparrot_training - Step 38479: {'lr': 6.776904441080115e-05, 'samples': 19701760, 'steps': 38479, 'loss/train': 2.174123525619507} 02/26/2022 03:23:34 - INFO - codeparrot_training - Step 38480: {'lr': 6.775784314464717e-05, 'samples': 19702272, 'steps': 38480, 'loss/train': 1.1795350313186646} 02/26/2022 03:23:40 - INFO - codeparrot_training - Step 38481: {'lr': 6.774664265916064e-05, 'samples': 19702784, 'steps': 38481, 'loss/train': 2.4828836917877197} 02/26/2022 03:23:43 - INFO - codeparrot_training - Step 38482: {'lr': 6.773544295438968e-05, 'samples': 19703296, 'steps': 38482, 'loss/train': 1.465942621231079} 02/26/2022 03:23:49 - INFO - codeparrot_training - Step 38483: {'lr': 6.772424403038224e-05, 'samples': 19703808, 'steps': 38483, 'loss/train': 2.299922466278076} 02/26/2022 03:23:54 - INFO - codeparrot_training - Step 38484: {'lr': 6.771304588718624e-05, 'samples': 19704320, 'steps': 38484, 'loss/train': 0.9538729786872864} 02/26/2022 03:23:58 - INFO - codeparrot_training - Step 38485: {'lr': 6.77018485248496e-05, 'samples': 19704832, 'steps': 38485, 'loss/train': 0.2449597269296646} 02/26/2022 03:24:04 - INFO - codeparrot_training - Step 38486: {'lr': 6.769065194342042e-05, 'samples': 19705344, 'steps': 38486, 'loss/train': 0.9421229362487793} 02/26/2022 03:24:07 - INFO - codeparrot_training - Step 38487: {'lr': 6.76794561429466e-05, 'samples': 19705856, 'steps': 38487, 'loss/train': 1.514542579650879} 02/26/2022 03:24:13 - INFO - codeparrot_training - Step 38488: {'lr': 6.76682611234761e-05, 'samples': 19706368, 'steps': 38488, 'loss/train': 0.4306022524833679} 02/26/2022 03:24:17 - INFO - codeparrot_training - Step 38489: {'lr': 6.76570668850568e-05, 'samples': 19706880, 'steps': 38489, 'loss/train': 1.4481408596038818} 02/26/2022 03:24:20 - INFO - codeparrot_training - Step 38490: {'lr': 6.764587342773679e-05, 'samples': 19707392, 'steps': 38490, 'loss/train': 1.9651048183441162} 02/26/2022 03:24:26 - INFO - codeparrot_training - Step 38491: {'lr': 6.763468075156395e-05, 'samples': 19707904, 'steps': 38491, 'loss/train': 2.8414618968963623} 02/26/2022 03:24:29 - INFO - codeparrot_training - Step 38492: {'lr': 6.762348885658626e-05, 'samples': 19708416, 'steps': 38492, 'loss/train': 2.0153722763061523} 02/26/2022 03:24:34 - INFO - codeparrot_training - Step 38493: {'lr': 6.761229774285152e-05, 'samples': 19708928, 'steps': 38493, 'loss/train': 2.105882167816162} 02/26/2022 03:24:38 - INFO - codeparrot_training - Step 38494: {'lr': 6.760110741040787e-05, 'samples': 19709440, 'steps': 38494, 'loss/train': 2.6483542919158936} 02/26/2022 03:24:43 - INFO - codeparrot_training - Step 38495: {'lr': 6.758991785930319e-05, 'samples': 19709952, 'steps': 38495, 'loss/train': 2.1674225330352783} 02/26/2022 03:24:47 - INFO - codeparrot_training - Step 38496: {'lr': 6.757872908958535e-05, 'samples': 19710464, 'steps': 38496, 'loss/train': 0.3592943549156189} 02/26/2022 03:24:52 - INFO - codeparrot_training - Step 38497: {'lr': 6.756754110130223e-05, 'samples': 19710976, 'steps': 38497, 'loss/train': 1.817550539970398} 02/26/2022 03:24:56 - INFO - codeparrot_training - Step 38498: {'lr': 6.755635389450191e-05, 'samples': 19711488, 'steps': 38498, 'loss/train': 1.4259891510009766} 02/26/2022 03:25:02 - INFO - codeparrot_training - Step 38499: {'lr': 6.754516746923217e-05, 'samples': 19712000, 'steps': 38499, 'loss/train': 1.8990557193756104} 02/26/2022 03:25:06 - INFO - codeparrot_training - Step 38500: {'lr': 6.753398182554116e-05, 'samples': 19712512, 'steps': 38500, 'loss/train': 1.7087346315383911} 02/26/2022 03:25:11 - INFO - codeparrot_training - Step 38501: {'lr': 6.752279696347646e-05, 'samples': 19713024, 'steps': 38501, 'loss/train': 2.697360038757324} 02/26/2022 03:25:15 - INFO - codeparrot_training - Step 38502: {'lr': 6.751161288308624e-05, 'samples': 19713536, 'steps': 38502, 'loss/train': 2.1709935665130615} 02/26/2022 03:25:20 - INFO - codeparrot_training - Step 38503: {'lr': 6.750042958441824e-05, 'samples': 19714048, 'steps': 38503, 'loss/train': 1.0028451681137085} 02/26/2022 03:25:24 - INFO - codeparrot_training - Step 38504: {'lr': 6.748924706752061e-05, 'samples': 19714560, 'steps': 38504, 'loss/train': 2.146336793899536} 02/26/2022 03:25:29 - INFO - codeparrot_training - Step 38505: {'lr': 6.747806533244093e-05, 'samples': 19715072, 'steps': 38505, 'loss/train': 1.8683357238769531} 02/26/2022 03:25:33 - INFO - codeparrot_training - Step 38506: {'lr': 6.746688437922736e-05, 'samples': 19715584, 'steps': 38506, 'loss/train': 1.897711157798767} 02/26/2022 03:25:38 - INFO - codeparrot_training - Step 38507: {'lr': 6.74557042079276e-05, 'samples': 19716096, 'steps': 38507, 'loss/train': 1.9326940774917603} 02/26/2022 03:25:42 - INFO - codeparrot_training - Step 38508: {'lr': 6.744452481858979e-05, 'samples': 19716608, 'steps': 38508, 'loss/train': 1.3787177801132202} 02/26/2022 03:25:49 - INFO - codeparrot_training - Step 38509: {'lr': 6.743334621126151e-05, 'samples': 19717120, 'steps': 38509, 'loss/train': 2.2313151359558105} 02/26/2022 03:25:52 - INFO - codeparrot_training - Step 38510: {'lr': 6.742216838599091e-05, 'samples': 19717632, 'steps': 38510, 'loss/train': 0.6717408895492554} 02/26/2022 03:25:58 - INFO - codeparrot_training - Step 38511: {'lr': 6.741099134282567e-05, 'samples': 19718144, 'steps': 38511, 'loss/train': 1.9280356168746948} 02/26/2022 03:26:01 - INFO - codeparrot_training - Step 38512: {'lr': 6.739981508181383e-05, 'samples': 19718656, 'steps': 38512, 'loss/train': 1.2840335369110107} 02/26/2022 03:26:07 - INFO - codeparrot_training - Step 38513: {'lr': 6.738863960300324e-05, 'samples': 19719168, 'steps': 38513, 'loss/train': 1.8356081247329712} 02/26/2022 03:26:10 - INFO - codeparrot_training - Step 38514: {'lr': 6.73774649064417e-05, 'samples': 19719680, 'steps': 38514, 'loss/train': 1.203269124031067} 02/26/2022 03:26:16 - INFO - codeparrot_training - Step 38515: {'lr': 6.736629099217703e-05, 'samples': 19720192, 'steps': 38515, 'loss/train': 2.4249298572540283} 02/26/2022 03:26:19 - INFO - codeparrot_training - Step 38516: {'lr': 6.735511786025725e-05, 'samples': 19720704, 'steps': 38516, 'loss/train': 1.4649714231491089} 02/26/2022 03:26:25 - INFO - codeparrot_training - Step 38517: {'lr': 6.734394551073017e-05, 'samples': 19721216, 'steps': 38517, 'loss/train': 2.0018768310546875} 02/26/2022 03:26:28 - INFO - codeparrot_training - Step 38518: {'lr': 6.733277394364363e-05, 'samples': 19721728, 'steps': 38518, 'loss/train': 0.41589176654815674} 02/26/2022 03:26:34 - INFO - codeparrot_training - Step 38519: {'lr': 6.732160315904538e-05, 'samples': 19722240, 'steps': 38519, 'loss/train': 1.5262006521224976} 02/26/2022 03:26:40 - INFO - codeparrot_training - Step 38520: {'lr': 6.731043315698346e-05, 'samples': 19722752, 'steps': 38520, 'loss/train': 1.461841344833374} 02/26/2022 03:26:44 - INFO - codeparrot_training - Step 38521: {'lr': 6.729926393750566e-05, 'samples': 19723264, 'steps': 38521, 'loss/train': 1.5198719501495361} 02/26/2022 03:26:47 - INFO - codeparrot_training - Step 38522: {'lr': 6.728809550065976e-05, 'samples': 19723776, 'steps': 38522, 'loss/train': 0.3158608376979828} 02/26/2022 03:26:52 - INFO - codeparrot_training - Step 38523: {'lr': 6.727692784649367e-05, 'samples': 19724288, 'steps': 38523, 'loss/train': 1.0442678928375244} 02/26/2022 03:26:56 - INFO - codeparrot_training - Step 38524: {'lr': 6.72657609750551e-05, 'samples': 19724800, 'steps': 38524, 'loss/train': 1.4605982303619385} 02/26/2022 03:27:01 - INFO - codeparrot_training - Step 38525: {'lr': 6.72545948863921e-05, 'samples': 19725312, 'steps': 38525, 'loss/train': 2.4296481609344482} 02/26/2022 03:27:05 - INFO - codeparrot_training - Step 38526: {'lr': 6.724342958055236e-05, 'samples': 19725824, 'steps': 38526, 'loss/train': 1.6514232158660889} 02/26/2022 03:27:10 - INFO - codeparrot_training - Step 38527: {'lr': 6.723226505758373e-05, 'samples': 19726336, 'steps': 38527, 'loss/train': 2.0844454765319824} 02/26/2022 03:27:14 - INFO - codeparrot_training - Step 38528: {'lr': 6.722110131753398e-05, 'samples': 19726848, 'steps': 38528, 'loss/train': 1.9466315507888794} 02/26/2022 03:27:20 - INFO - codeparrot_training - Step 38529: {'lr': 6.720993836045105e-05, 'samples': 19727360, 'steps': 38529, 'loss/train': 1.6696027517318726} 02/26/2022 03:27:25 - INFO - codeparrot_training - Step 38530: {'lr': 6.719877618638273e-05, 'samples': 19727872, 'steps': 38530, 'loss/train': 0.8651958107948303} 02/26/2022 03:27:28 - INFO - codeparrot_training - Step 38531: {'lr': 6.71876147953768e-05, 'samples': 19728384, 'steps': 38531, 'loss/train': 1.2277593612670898} 02/26/2022 03:27:34 - INFO - codeparrot_training - Step 38532: {'lr': 6.717645418748097e-05, 'samples': 19728896, 'steps': 38532, 'loss/train': 2.0817205905914307} 02/26/2022 03:27:37 - INFO - codeparrot_training - Step 38533: {'lr': 6.716529436274327e-05, 'samples': 19729408, 'steps': 38533, 'loss/train': 1.5316346883773804} 02/26/2022 03:27:44 - INFO - codeparrot_training - Step 38534: {'lr': 6.715413532121131e-05, 'samples': 19729920, 'steps': 38534, 'loss/train': 0.212492898106575} 02/26/2022 03:27:47 - INFO - codeparrot_training - Step 38535: {'lr': 6.714297706293313e-05, 'samples': 19730432, 'steps': 38535, 'loss/train': 2.2703354358673096} 02/26/2022 03:27:52 - INFO - codeparrot_training - Step 38536: {'lr': 6.71318195879562e-05, 'samples': 19730944, 'steps': 38536, 'loss/train': 2.4443178176879883} 02/26/2022 03:27:56 - INFO - codeparrot_training - Step 38537: {'lr': 6.712066289632856e-05, 'samples': 19731456, 'steps': 38537, 'loss/train': 2.447617530822754} 02/26/2022 03:28:01 - INFO - codeparrot_training - Step 38538: {'lr': 6.710950698809789e-05, 'samples': 19731968, 'steps': 38538, 'loss/train': 1.8363568782806396} 02/26/2022 03:28:05 - INFO - codeparrot_training - Step 38539: {'lr': 6.709835186331214e-05, 'samples': 19732480, 'steps': 38539, 'loss/train': 2.056351900100708} 02/26/2022 03:28:11 - INFO - codeparrot_training - Step 38540: {'lr': 6.708719752201883e-05, 'samples': 19732992, 'steps': 38540, 'loss/train': 1.7019981145858765} 02/26/2022 03:28:14 - INFO - codeparrot_training - Step 38541: {'lr': 6.707604396426595e-05, 'samples': 19733504, 'steps': 38541, 'loss/train': 1.4125553369522095} 02/26/2022 03:28:19 - INFO - codeparrot_training - Step 38542: {'lr': 6.706489119010115e-05, 'samples': 19734016, 'steps': 38542, 'loss/train': 0.6963955163955688} 02/26/2022 03:28:23 - INFO - codeparrot_training - Step 38543: {'lr': 6.705373919957241e-05, 'samples': 19734528, 'steps': 38543, 'loss/train': 2.0790679454803467} 02/26/2022 03:28:29 - INFO - codeparrot_training - Step 38544: {'lr': 6.704258799272722e-05, 'samples': 19735040, 'steps': 38544, 'loss/train': 0.8199334144592285} 02/26/2022 03:28:33 - INFO - codeparrot_training - Step 38545: {'lr': 6.703143756961355e-05, 'samples': 19735552, 'steps': 38545, 'loss/train': 1.4792839288711548} 02/26/2022 03:28:38 - INFO - codeparrot_training - Step 38546: {'lr': 6.702028793027906e-05, 'samples': 19736064, 'steps': 38546, 'loss/train': 1.3455629348754883} 02/26/2022 03:28:42 - INFO - codeparrot_training - Step 38547: {'lr': 6.70091390747716e-05, 'samples': 19736576, 'steps': 38547, 'loss/train': 0.7645719051361084} 02/26/2022 03:28:47 - INFO - codeparrot_training - Step 38548: {'lr': 6.699799100313888e-05, 'samples': 19737088, 'steps': 38548, 'loss/train': 1.2810925245285034} 02/26/2022 03:28:51 - INFO - codeparrot_training - Step 38549: {'lr': 6.698684371542866e-05, 'samples': 19737600, 'steps': 38549, 'loss/train': 1.9323935508728027} 02/26/2022 03:28:56 - INFO - codeparrot_training - Step 38550: {'lr': 6.697569721168861e-05, 'samples': 19738112, 'steps': 38550, 'loss/train': 1.6603186130523682} 02/26/2022 03:29:00 - INFO - codeparrot_training - Step 38551: {'lr': 6.696455149196664e-05, 'samples': 19738624, 'steps': 38551, 'loss/train': 2.3496146202087402} 02/26/2022 03:29:05 - INFO - codeparrot_training - Step 38552: {'lr': 6.695340655631041e-05, 'samples': 19739136, 'steps': 38552, 'loss/train': 1.5190728902816772} 02/26/2022 03:29:09 - INFO - codeparrot_training - Step 38553: {'lr': 6.694226240476767e-05, 'samples': 19739648, 'steps': 38553, 'loss/train': 1.7923954725265503} 02/26/2022 03:29:15 - INFO - codeparrot_training - Step 38554: {'lr': 6.693111903738605e-05, 'samples': 19740160, 'steps': 38554, 'loss/train': 0.9450716376304626} 02/26/2022 03:29:18 - INFO - codeparrot_training - Step 38555: {'lr': 6.691997645421347e-05, 'samples': 19740672, 'steps': 38555, 'loss/train': 1.521878719329834} 02/26/2022 03:29:24 - INFO - codeparrot_training - Step 38556: {'lr': 6.69088346552976e-05, 'samples': 19741184, 'steps': 38556, 'loss/train': 1.3934801816940308} 02/26/2022 03:29:27 - INFO - codeparrot_training - Step 38557: {'lr': 6.689769364068613e-05, 'samples': 19741696, 'steps': 38557, 'loss/train': 1.8056319952011108} 02/26/2022 03:29:33 - INFO - codeparrot_training - Step 38558: {'lr': 6.68865534104267e-05, 'samples': 19742208, 'steps': 38558, 'loss/train': 2.134234666824341} 02/26/2022 03:29:36 - INFO - codeparrot_training - Step 38559: {'lr': 6.687541396456723e-05, 'samples': 19742720, 'steps': 38559, 'loss/train': 2.627307653427124} 02/26/2022 03:29:42 - INFO - codeparrot_training - Step 38560: {'lr': 6.686427530315534e-05, 'samples': 19743232, 'steps': 38560, 'loss/train': 1.8709713220596313} 02/26/2022 03:29:45 - INFO - codeparrot_training - Step 38561: {'lr': 6.685313742623872e-05, 'samples': 19743744, 'steps': 38561, 'loss/train': 2.4307448863983154} 02/26/2022 03:29:51 - INFO - codeparrot_training - Step 38562: {'lr': 6.684200033386506e-05, 'samples': 19744256, 'steps': 38562, 'loss/train': 1.7083773612976074} 02/26/2022 03:29:54 - INFO - codeparrot_training - Step 38563: {'lr': 6.683086402608215e-05, 'samples': 19744768, 'steps': 38563, 'loss/train': 0.7015232443809509} 02/26/2022 03:30:01 - INFO - codeparrot_training - Step 38564: {'lr': 6.681972850293769e-05, 'samples': 19745280, 'steps': 38564, 'loss/train': 1.6064817905426025} 02/26/2022 03:30:04 - INFO - codeparrot_training - Step 38565: {'lr': 6.680859376447934e-05, 'samples': 19745792, 'steps': 38565, 'loss/train': 2.768766403198242} 02/26/2022 03:30:10 - INFO - codeparrot_training - Step 38566: {'lr': 6.679745981075472e-05, 'samples': 19746304, 'steps': 38566, 'loss/train': 1.3902474641799927} 02/26/2022 03:30:13 - INFO - codeparrot_training - Step 38567: {'lr': 6.67863266418117e-05, 'samples': 19746816, 'steps': 38567, 'loss/train': 1.518554925918579} 02/26/2022 03:30:19 - INFO - codeparrot_training - Step 38568: {'lr': 6.677519425769784e-05, 'samples': 19747328, 'steps': 38568, 'loss/train': 2.504375696182251} 02/26/2022 03:30:22 - INFO - codeparrot_training - Step 38569: {'lr': 6.676406265846083e-05, 'samples': 19747840, 'steps': 38569, 'loss/train': 1.2691808938980103} 02/26/2022 03:30:28 - INFO - codeparrot_training - Step 38570: {'lr': 6.675293184414846e-05, 'samples': 19748352, 'steps': 38570, 'loss/train': 1.733219861984253} 02/26/2022 03:30:31 - INFO - codeparrot_training - Step 38571: {'lr': 6.674180181480835e-05, 'samples': 19748864, 'steps': 38571, 'loss/train': 1.622839331626892} 02/26/2022 03:30:37 - INFO - codeparrot_training - Step 38572: {'lr': 6.673067257048815e-05, 'samples': 19749376, 'steps': 38572, 'loss/train': 1.3758234977722168} 02/26/2022 03:30:40 - INFO - codeparrot_training - Step 38573: {'lr': 6.671954411123549e-05, 'samples': 19749888, 'steps': 38573, 'loss/train': 1.2200340032577515} 02/26/2022 03:30:46 - INFO - codeparrot_training - Step 38574: {'lr': 6.670841643709824e-05, 'samples': 19750400, 'steps': 38574, 'loss/train': 1.8557848930358887} 02/26/2022 03:30:49 - INFO - codeparrot_training - Step 38575: {'lr': 6.669728954812379e-05, 'samples': 19750912, 'steps': 38575, 'loss/train': 1.7742339372634888} 02/26/2022 03:30:55 - INFO - codeparrot_training - Step 38576: {'lr': 6.668616344436005e-05, 'samples': 19751424, 'steps': 38576, 'loss/train': 0.9452046751976013} 02/26/2022 03:30:58 - INFO - codeparrot_training - Step 38577: {'lr': 6.667503812585446e-05, 'samples': 19751936, 'steps': 38577, 'loss/train': 1.2546623945236206} 02/26/2022 03:31:04 - INFO - codeparrot_training - Step 38578: {'lr': 6.666391359265497e-05, 'samples': 19752448, 'steps': 38578, 'loss/train': 2.090447187423706} 02/26/2022 03:31:07 - INFO - codeparrot_training - Step 38579: {'lr': 6.665278984480889e-05, 'samples': 19752960, 'steps': 38579, 'loss/train': 2.8809638023376465} 02/26/2022 03:31:13 - INFO - codeparrot_training - Step 38580: {'lr': 6.664166688236411e-05, 'samples': 19753472, 'steps': 38580, 'loss/train': 1.667805790901184} 02/26/2022 03:31:17 - INFO - codeparrot_training - Step 38581: {'lr': 6.663054470536816e-05, 'samples': 19753984, 'steps': 38581, 'loss/train': 1.2165825366973877} 02/26/2022 03:31:22 - INFO - codeparrot_training - Step 38582: {'lr': 6.66194233138688e-05, 'samples': 19754496, 'steps': 38582, 'loss/train': 2.2430546283721924} 02/26/2022 03:31:26 - INFO - codeparrot_training - Step 38583: {'lr': 6.66083027079136e-05, 'samples': 19755008, 'steps': 38583, 'loss/train': 0.7722062468528748} 02/26/2022 03:31:31 - INFO - codeparrot_training - Step 38584: {'lr': 6.659718288755018e-05, 'samples': 19755520, 'steps': 38584, 'loss/train': 1.7943100929260254} 02/26/2022 03:31:35 - INFO - codeparrot_training - Step 38585: {'lr': 6.658606385282612e-05, 'samples': 19756032, 'steps': 38585, 'loss/train': 1.3602079153060913} 02/26/2022 03:31:40 - INFO - codeparrot_training - Step 38586: {'lr': 6.657494560378922e-05, 'samples': 19756544, 'steps': 38586, 'loss/train': 1.8367716073989868} 02/26/2022 03:31:46 - INFO - codeparrot_training - Step 38587: {'lr': 6.656382814048698e-05, 'samples': 19757056, 'steps': 38587, 'loss/train': 1.418237566947937} 02/26/2022 03:31:49 - INFO - codeparrot_training - Step 38588: {'lr': 6.655271146296707e-05, 'samples': 19757568, 'steps': 38588, 'loss/train': 2.427154779434204} 02/26/2022 03:31:56 - INFO - codeparrot_training - Step 38589: {'lr': 6.654159557127701e-05, 'samples': 19758080, 'steps': 38589, 'loss/train': 0.9327219724655151} 02/26/2022 03:31:59 - INFO - codeparrot_training - Step 38590: {'lr': 6.653048046546456e-05, 'samples': 19758592, 'steps': 38590, 'loss/train': 1.7165138721466064} 02/26/2022 03:32:05 - INFO - codeparrot_training - Step 38591: {'lr': 6.65193661455773e-05, 'samples': 19759104, 'steps': 38591, 'loss/train': 2.1705565452575684} 02/26/2022 03:32:08 - INFO - codeparrot_training - Step 38592: {'lr': 6.650825261166282e-05, 'samples': 19759616, 'steps': 38592, 'loss/train': 1.8768551349639893} 02/26/2022 03:32:14 - INFO - codeparrot_training - Step 38593: {'lr': 6.649713986376862e-05, 'samples': 19760128, 'steps': 38593, 'loss/train': 1.4631894826889038} 02/26/2022 03:32:17 - INFO - codeparrot_training - Step 38594: {'lr': 6.64860279019425e-05, 'samples': 19760640, 'steps': 38594, 'loss/train': 2.925553560256958} 02/26/2022 03:32:23 - INFO - codeparrot_training - Step 38595: {'lr': 6.647491672623187e-05, 'samples': 19761152, 'steps': 38595, 'loss/train': 0.7812713980674744} 02/26/2022 03:32:26 - INFO - codeparrot_training - Step 38596: {'lr': 6.646380633668459e-05, 'samples': 19761664, 'steps': 38596, 'loss/train': 1.0813846588134766} 02/26/2022 03:32:32 - INFO - codeparrot_training - Step 38597: {'lr': 6.645269673334792e-05, 'samples': 19762176, 'steps': 38597, 'loss/train': 2.621788740158081} 02/26/2022 03:32:35 - INFO - codeparrot_training - Step 38598: {'lr': 6.64415879162697e-05, 'samples': 19762688, 'steps': 38598, 'loss/train': 1.9680814743041992} 02/26/2022 03:32:41 - INFO - codeparrot_training - Step 38599: {'lr': 6.643047988549733e-05, 'samples': 19763200, 'steps': 38599, 'loss/train': 1.3169771432876587} 02/26/2022 03:32:44 - INFO - codeparrot_training - Step 38600: {'lr': 6.641937264107867e-05, 'samples': 19763712, 'steps': 38600, 'loss/train': 3.1705527305603027} 02/26/2022 03:32:50 - INFO - codeparrot_training - Step 38601: {'lr': 6.640826618306098e-05, 'samples': 19764224, 'steps': 38601, 'loss/train': 1.2211713790893555} 02/26/2022 03:32:54 - INFO - codeparrot_training - Step 38602: {'lr': 6.639716051149203e-05, 'samples': 19764736, 'steps': 38602, 'loss/train': 1.803682565689087} 02/26/2022 03:32:59 - INFO - codeparrot_training - Step 38603: {'lr': 6.638605562641928e-05, 'samples': 19765248, 'steps': 38603, 'loss/train': 2.1569159030914307} 02/26/2022 03:33:03 - INFO - codeparrot_training - Step 38604: {'lr': 6.637495152789052e-05, 'samples': 19765760, 'steps': 38604, 'loss/train': 1.285002589225769} 02/26/2022 03:33:08 - INFO - codeparrot_training - Step 38605: {'lr': 6.636384821595302e-05, 'samples': 19766272, 'steps': 38605, 'loss/train': 1.8323359489440918} 02/26/2022 03:33:12 - INFO - codeparrot_training - Step 38606: {'lr': 6.635274569065452e-05, 'samples': 19766784, 'steps': 38606, 'loss/train': 1.6491860151290894} 02/26/2022 03:33:17 - INFO - codeparrot_training - Step 38607: {'lr': 6.634164395204259e-05, 'samples': 19767296, 'steps': 38607, 'loss/train': 2.204594373703003} 02/26/2022 03:33:21 - INFO - codeparrot_training - Step 38608: {'lr': 6.633054300016465e-05, 'samples': 19767808, 'steps': 38608, 'loss/train': 1.0186845064163208} 02/26/2022 03:33:26 - INFO - codeparrot_training - Step 38609: {'lr': 6.631944283506838e-05, 'samples': 19768320, 'steps': 38609, 'loss/train': 0.9944896101951599} 02/26/2022 03:33:30 - INFO - codeparrot_training - Step 38610: {'lr': 6.630834345680135e-05, 'samples': 19768832, 'steps': 38610, 'loss/train': 1.1949081420898438} 02/26/2022 03:33:36 - INFO - codeparrot_training - Step 38611: {'lr': 6.6297244865411e-05, 'samples': 19769344, 'steps': 38611, 'loss/train': 2.241675853729248} 02/26/2022 03:33:39 - INFO - codeparrot_training - Step 38612: {'lr': 6.628614706094488e-05, 'samples': 19769856, 'steps': 38612, 'loss/train': 1.6453428268432617} 02/26/2022 03:33:45 - INFO - codeparrot_training - Step 38613: {'lr': 6.627505004345061e-05, 'samples': 19770368, 'steps': 38613, 'loss/train': 0.45854154229164124} 02/26/2022 03:33:48 - INFO - codeparrot_training - Step 38614: {'lr': 6.62639538129757e-05, 'samples': 19770880, 'steps': 38614, 'loss/train': 1.6465308666229248} 02/26/2022 03:33:52 - INFO - codeparrot_training - Step 38615: {'lr': 6.625285836956769e-05, 'samples': 19771392, 'steps': 38615, 'loss/train': 1.4277374744415283} 02/26/2022 03:33:58 - INFO - codeparrot_training - Step 38616: {'lr': 6.624176371327401e-05, 'samples': 19771904, 'steps': 38616, 'loss/train': 1.0375969409942627} 02/26/2022 03:34:03 - INFO - codeparrot_training - Step 38617: {'lr': 6.62306698441423e-05, 'samples': 19772416, 'steps': 38617, 'loss/train': 1.7115790843963623} 02/26/2022 03:34:07 - INFO - codeparrot_training - Step 38618: {'lr': 6.62195767622201e-05, 'samples': 19772928, 'steps': 38618, 'loss/train': 2.1650514602661133} 02/26/2022 03:34:12 - INFO - codeparrot_training - Step 38619: {'lr': 6.620848446755485e-05, 'samples': 19773440, 'steps': 38619, 'loss/train': 1.8237920999526978} 02/26/2022 03:34:16 - INFO - codeparrot_training - Step 38620: {'lr': 6.6197392960194e-05, 'samples': 19773952, 'steps': 38620, 'loss/train': 0.21208587288856506} 02/26/2022 03:34:21 - INFO - codeparrot_training - Step 38621: {'lr': 6.618630224018526e-05, 'samples': 19774464, 'steps': 38621, 'loss/train': 1.5871200561523438} 02/26/2022 03:34:25 - INFO - codeparrot_training - Step 38622: {'lr': 6.617521230757603e-05, 'samples': 19774976, 'steps': 38622, 'loss/train': 2.3343801498413086} 02/26/2022 03:34:30 - INFO - codeparrot_training - Step 38623: {'lr': 6.616412316241379e-05, 'samples': 19775488, 'steps': 38623, 'loss/train': 1.4226933717727661} 02/26/2022 03:34:34 - INFO - codeparrot_training - Step 38624: {'lr': 6.615303480474601e-05, 'samples': 19776000, 'steps': 38624, 'loss/train': 1.3573405742645264} 02/26/2022 03:34:39 - INFO - codeparrot_training - Step 38625: {'lr': 6.614194723462036e-05, 'samples': 19776512, 'steps': 38625, 'loss/train': 2.7144830226898193} 02/26/2022 03:34:43 - INFO - codeparrot_training - Step 38626: {'lr': 6.61308604520842e-05, 'samples': 19777024, 'steps': 38626, 'loss/train': 0.4123234152793884} 02/26/2022 03:34:49 - INFO - codeparrot_training - Step 38627: {'lr': 6.611977445718504e-05, 'samples': 19777536, 'steps': 38627, 'loss/train': 1.7759041786193848} 02/26/2022 03:34:52 - INFO - codeparrot_training - Step 38628: {'lr': 6.610868924997033e-05, 'samples': 19778048, 'steps': 38628, 'loss/train': 2.022170305252075} 02/26/2022 03:34:58 - INFO - codeparrot_training - Step 38629: {'lr': 6.609760483048766e-05, 'samples': 19778560, 'steps': 38629, 'loss/train': 2.1668646335601807} 02/26/2022 03:35:01 - INFO - codeparrot_training - Step 38630: {'lr': 6.608652119878436e-05, 'samples': 19779072, 'steps': 38630, 'loss/train': 1.4865700006484985} 02/26/2022 03:35:07 - INFO - codeparrot_training - Step 38631: {'lr': 6.607543835490818e-05, 'samples': 19779584, 'steps': 38631, 'loss/train': 2.052323818206787} 02/26/2022 03:35:10 - INFO - codeparrot_training - Step 38632: {'lr': 6.606435629890628e-05, 'samples': 19780096, 'steps': 38632, 'loss/train': 1.1872520446777344} 02/26/2022 03:35:16 - INFO - codeparrot_training - Step 38633: {'lr': 6.60532750308263e-05, 'samples': 19780608, 'steps': 38633, 'loss/train': 1.5328195095062256} 02/26/2022 03:35:19 - INFO - codeparrot_training - Step 38634: {'lr': 6.604219455071567e-05, 'samples': 19781120, 'steps': 38634, 'loss/train': 2.6282432079315186} 02/26/2022 03:35:25 - INFO - codeparrot_training - Step 38635: {'lr': 6.603111485862195e-05, 'samples': 19781632, 'steps': 38635, 'loss/train': 1.7048457860946655} 02/26/2022 03:35:28 - INFO - codeparrot_training - Step 38636: {'lr': 6.60200359545924e-05, 'samples': 19782144, 'steps': 38636, 'loss/train': 1.3628875017166138} 02/26/2022 03:35:35 - INFO - codeparrot_training - Step 38637: {'lr': 6.600895783867469e-05, 'samples': 19782656, 'steps': 38637, 'loss/train': 1.6989284753799438} 02/26/2022 03:35:38 - INFO - codeparrot_training - Step 38638: {'lr': 6.599788051091606e-05, 'samples': 19783168, 'steps': 38638, 'loss/train': 1.4103751182556152} 02/26/2022 03:35:43 - INFO - codeparrot_training - Step 38639: {'lr': 6.598680397136428e-05, 'samples': 19783680, 'steps': 38639, 'loss/train': 1.1829496622085571} 02/26/2022 03:35:47 - INFO - codeparrot_training - Step 38640: {'lr': 6.597572822006642e-05, 'samples': 19784192, 'steps': 38640, 'loss/train': 0.5303569436073303} 02/26/2022 03:35:52 - INFO - codeparrot_training - Step 38641: {'lr': 6.596465325707019e-05, 'samples': 19784704, 'steps': 38641, 'loss/train': 0.8118969202041626} 02/26/2022 03:35:56 - INFO - codeparrot_training - Step 38642: {'lr': 6.595357908242289e-05, 'samples': 19785216, 'steps': 38642, 'loss/train': 1.6326771974563599} 02/26/2022 03:36:02 - INFO - codeparrot_training - Step 38643: {'lr': 6.594250569617204e-05, 'samples': 19785728, 'steps': 38643, 'loss/train': 2.5719337463378906} 02/26/2022 03:36:05 - INFO - codeparrot_training - Step 38644: {'lr': 6.59314330983651e-05, 'samples': 19786240, 'steps': 38644, 'loss/train': 0.7581230401992798} 02/26/2022 03:36:11 - INFO - codeparrot_training - Step 38645: {'lr': 6.592036128904944e-05, 'samples': 19786752, 'steps': 38645, 'loss/train': 0.7914933562278748} 02/26/2022 03:36:14 - INFO - codeparrot_training - Step 38646: {'lr': 6.590929026827241e-05, 'samples': 19787264, 'steps': 38646, 'loss/train': 1.3831466436386108} 02/26/2022 03:36:21 - INFO - codeparrot_training - Step 38647: {'lr': 6.589822003608162e-05, 'samples': 19787776, 'steps': 38647, 'loss/train': 2.1867780685424805} 02/26/2022 03:36:24 - INFO - codeparrot_training - Step 38648: {'lr': 6.58871505925244e-05, 'samples': 19788288, 'steps': 38648, 'loss/train': 2.6343843936920166} 02/26/2022 03:36:30 - INFO - codeparrot_training - Step 38649: {'lr': 6.587608193764816e-05, 'samples': 19788800, 'steps': 38649, 'loss/train': 3.540714740753174} 02/26/2022 03:36:33 - INFO - codeparrot_training - Step 38650: {'lr': 6.586501407150022e-05, 'samples': 19789312, 'steps': 38650, 'loss/train': 0.4508006274700165} 02/26/2022 03:36:39 - INFO - codeparrot_training - Step 38651: {'lr': 6.58539469941282e-05, 'samples': 19789824, 'steps': 38651, 'loss/train': 2.018789052963257} 02/26/2022 03:36:42 - INFO - codeparrot_training - Step 38652: {'lr': 6.584288070557937e-05, 'samples': 19790336, 'steps': 38652, 'loss/train': 3.4998269081115723} 02/26/2022 03:36:48 - INFO - codeparrot_training - Step 38653: {'lr': 6.583181520590117e-05, 'samples': 19790848, 'steps': 38653, 'loss/train': 1.9903764724731445} 02/26/2022 03:36:52 - INFO - codeparrot_training - Step 38654: {'lr': 6.5820750495141e-05, 'samples': 19791360, 'steps': 38654, 'loss/train': 1.8902126550674438} 02/26/2022 03:36:57 - INFO - codeparrot_training - Step 38655: {'lr': 6.580968657334616e-05, 'samples': 19791872, 'steps': 38655, 'loss/train': 2.3107526302337646} 02/26/2022 03:37:01 - INFO - codeparrot_training - Step 38656: {'lr': 6.579862344056422e-05, 'samples': 19792384, 'steps': 38656, 'loss/train': 1.2620282173156738} 02/26/2022 03:37:07 - INFO - codeparrot_training - Step 38657: {'lr': 6.57875610968425e-05, 'samples': 19792896, 'steps': 38657, 'loss/train': 2.3737082481384277} 02/26/2022 03:37:10 - INFO - codeparrot_training - Step 38658: {'lr': 6.577649954222836e-05, 'samples': 19793408, 'steps': 38658, 'loss/train': 1.1538290977478027} 02/26/2022 03:37:16 - INFO - codeparrot_training - Step 38659: {'lr': 6.576543877676913e-05, 'samples': 19793920, 'steps': 38659, 'loss/train': 1.7753148078918457} 02/26/2022 03:37:19 - INFO - codeparrot_training - Step 38660: {'lr': 6.575437880051233e-05, 'samples': 19794432, 'steps': 38660, 'loss/train': 1.6466093063354492} 02/26/2022 03:37:25 - INFO - codeparrot_training - Step 38661: {'lr': 6.574331961350527e-05, 'samples': 19794944, 'steps': 38661, 'loss/train': 2.5419349670410156} 02/26/2022 03:37:28 - INFO - codeparrot_training - Step 38662: {'lr': 6.573226121579532e-05, 'samples': 19795456, 'steps': 38662, 'loss/train': 0.5497141480445862} 02/26/2022 03:37:36 - INFO - codeparrot_training - Step 38663: {'lr': 6.572120360742977e-05, 'samples': 19795968, 'steps': 38663, 'loss/train': 1.6821372509002686} 02/26/2022 03:37:40 - INFO - codeparrot_training - Step 38664: {'lr': 6.571014678845617e-05, 'samples': 19796480, 'steps': 38664, 'loss/train': 2.0252134799957275} 02/26/2022 03:37:45 - INFO - codeparrot_training - Step 38665: {'lr': 6.569909075892175e-05, 'samples': 19796992, 'steps': 38665, 'loss/train': 1.8512320518493652} 02/26/2022 03:37:49 - INFO - codeparrot_training - Step 38666: {'lr': 6.568803551887392e-05, 'samples': 19797504, 'steps': 38666, 'loss/train': 1.83315110206604} 02/26/2022 03:37:54 - INFO - codeparrot_training - Step 38667: {'lr': 6.567698106835992e-05, 'samples': 19798016, 'steps': 38667, 'loss/train': 1.727878451347351} 02/26/2022 03:37:58 - INFO - codeparrot_training - Step 38668: {'lr': 6.566592740742733e-05, 'samples': 19798528, 'steps': 38668, 'loss/train': 1.367841124534607} 02/26/2022 03:38:03 - INFO - codeparrot_training - Step 38669: {'lr': 6.565487453612327e-05, 'samples': 19799040, 'steps': 38669, 'loss/train': 1.0969486236572266} 02/26/2022 03:38:07 - INFO - codeparrot_training - Step 38670: {'lr': 6.564382245449532e-05, 'samples': 19799552, 'steps': 38670, 'loss/train': 2.2816412448883057} 02/26/2022 03:38:12 - INFO - codeparrot_training - Step 38671: {'lr': 6.563277116259056e-05, 'samples': 19800064, 'steps': 38671, 'loss/train': 1.8742378950119019} 02/26/2022 03:38:16 - INFO - codeparrot_training - Step 38672: {'lr': 6.562172066045655e-05, 'samples': 19800576, 'steps': 38672, 'loss/train': 0.9214587807655334} 02/26/2022 03:38:23 - INFO - codeparrot_training - Step 38673: {'lr': 6.561067094814046e-05, 'samples': 19801088, 'steps': 38673, 'loss/train': 1.4929498434066772} 02/26/2022 03:38:27 - INFO - codeparrot_training - Step 38674: {'lr': 6.559962202568986e-05, 'samples': 19801600, 'steps': 38674, 'loss/train': 1.0426241159439087} 02/26/2022 03:38:32 - INFO - codeparrot_training - Step 38675: {'lr': 6.558857389315179e-05, 'samples': 19802112, 'steps': 38675, 'loss/train': 1.5233253240585327} 02/26/2022 03:38:36 - INFO - codeparrot_training - Step 38676: {'lr': 6.557752655057376e-05, 'samples': 19802624, 'steps': 38676, 'loss/train': 2.088609457015991} 02/26/2022 03:38:41 - INFO - codeparrot_training - Step 38677: {'lr': 6.556647999800298e-05, 'samples': 19803136, 'steps': 38677, 'loss/train': 2.1287267208099365} 02/26/2022 03:38:45 - INFO - codeparrot_training - Step 38678: {'lr': 6.55554342354869e-05, 'samples': 19803648, 'steps': 38678, 'loss/train': 1.4118505716323853} 02/26/2022 03:38:50 - INFO - codeparrot_training - Step 38679: {'lr': 6.554438926307279e-05, 'samples': 19804160, 'steps': 38679, 'loss/train': 2.1947858333587646} 02/26/2022 03:38:54 - INFO - codeparrot_training - Step 38680: {'lr': 6.553334508080793e-05, 'samples': 19804672, 'steps': 38680, 'loss/train': 1.3291409015655518} 02/26/2022 03:38:59 - INFO - codeparrot_training - Step 38681: {'lr': 6.55223016887396e-05, 'samples': 19805184, 'steps': 38681, 'loss/train': 1.111159086227417} 02/26/2022 03:39:03 - INFO - codeparrot_training - Step 38682: {'lr': 6.551125908691519e-05, 'samples': 19805696, 'steps': 38682, 'loss/train': 1.48469078540802} 02/26/2022 03:39:10 - INFO - codeparrot_training - Step 38683: {'lr': 6.550021727538196e-05, 'samples': 19806208, 'steps': 38683, 'loss/train': 1.2997806072235107} 02/26/2022 03:39:13 - INFO - codeparrot_training - Step 38684: {'lr': 6.548917625418727e-05, 'samples': 19806720, 'steps': 38684, 'loss/train': 1.4496510028839111} 02/26/2022 03:39:19 - INFO - codeparrot_training - Step 38685: {'lr': 6.547813602337824e-05, 'samples': 19807232, 'steps': 38685, 'loss/train': 2.0880749225616455} 02/26/2022 03:39:23 - INFO - codeparrot_training - Step 38686: {'lr': 6.546709658300234e-05, 'samples': 19807744, 'steps': 38686, 'loss/train': 1.9977792501449585} 02/26/2022 03:39:28 - INFO - codeparrot_training - Step 38687: {'lr': 6.545605793310686e-05, 'samples': 19808256, 'steps': 38687, 'loss/train': 2.467320680618286} 02/26/2022 03:39:31 - INFO - codeparrot_training - Step 38688: {'lr': 6.544502007373898e-05, 'samples': 19808768, 'steps': 38688, 'loss/train': 3.4030721187591553} 02/26/2022 03:39:37 - INFO - codeparrot_training - Step 38689: {'lr': 6.543398300494597e-05, 'samples': 19809280, 'steps': 38689, 'loss/train': 0.6142285466194153} 02/26/2022 03:39:41 - INFO - codeparrot_training - Step 38690: {'lr': 6.542294672677526e-05, 'samples': 19809792, 'steps': 38690, 'loss/train': 2.4969725608825684} 02/26/2022 03:39:46 - INFO - codeparrot_training - Step 38691: {'lr': 6.541191123927401e-05, 'samples': 19810304, 'steps': 38691, 'loss/train': 1.7564287185668945} 02/26/2022 03:39:50 - INFO - codeparrot_training - Step 38692: {'lr': 6.540087654248956e-05, 'samples': 19810816, 'steps': 38692, 'loss/train': 2.1521451473236084} 02/26/2022 03:39:57 - INFO - codeparrot_training - Step 38693: {'lr': 6.538984263646902e-05, 'samples': 19811328, 'steps': 38693, 'loss/train': 1.4548927545547485} 02/26/2022 03:40:00 - INFO - codeparrot_training - Step 38694: {'lr': 6.537880952125985e-05, 'samples': 19811840, 'steps': 38694, 'loss/train': 1.353824496269226} 02/26/2022 03:40:06 - INFO - codeparrot_training - Step 38695: {'lr': 6.536777719690926e-05, 'samples': 19812352, 'steps': 38695, 'loss/train': 2.5916013717651367} 02/26/2022 03:40:09 - INFO - codeparrot_training - Step 38696: {'lr': 6.535674566346448e-05, 'samples': 19812864, 'steps': 38696, 'loss/train': 0.9589845538139343} 02/26/2022 03:40:15 - INFO - codeparrot_training - Step 38697: {'lr': 6.534571492097269e-05, 'samples': 19813376, 'steps': 38697, 'loss/train': 1.6400796175003052} 02/26/2022 03:40:18 - INFO - codeparrot_training - Step 38698: {'lr': 6.53346849694813e-05, 'samples': 19813888, 'steps': 38698, 'loss/train': 2.89313006401062} 02/26/2022 03:40:24 - INFO - codeparrot_training - Step 38699: {'lr': 6.532365580903748e-05, 'samples': 19814400, 'steps': 38699, 'loss/train': 2.332109212875366} 02/26/2022 03:40:27 - INFO - codeparrot_training - Step 38700: {'lr': 6.531262743968849e-05, 'samples': 19814912, 'steps': 38700, 'loss/train': 1.483180284500122} 02/26/2022 03:40:33 - INFO - codeparrot_training - Step 38701: {'lr': 6.530159986148154e-05, 'samples': 19815424, 'steps': 38701, 'loss/train': 0.9306517243385315} 02/26/2022 03:40:36 - INFO - codeparrot_training - Step 38702: {'lr': 6.529057307446381e-05, 'samples': 19815936, 'steps': 38702, 'loss/train': 2.1577022075653076} 02/26/2022 03:40:42 - INFO - codeparrot_training - Step 38703: {'lr': 6.527954707868273e-05, 'samples': 19816448, 'steps': 38703, 'loss/train': 1.1545403003692627} 02/26/2022 03:40:45 - INFO - codeparrot_training - Step 38704: {'lr': 6.52685218741853e-05, 'samples': 19816960, 'steps': 38704, 'loss/train': 2.672415018081665} 02/26/2022 03:40:51 - INFO - codeparrot_training - Step 38705: {'lr': 6.525749746101902e-05, 'samples': 19817472, 'steps': 38705, 'loss/train': 1.1697603464126587} 02/26/2022 03:40:56 - INFO - codeparrot_training - Step 38706: {'lr': 6.52464738392308e-05, 'samples': 19817984, 'steps': 38706, 'loss/train': 2.028974771499634} 02/26/2022 03:41:00 - INFO - codeparrot_training - Step 38707: {'lr': 6.52354510088681e-05, 'samples': 19818496, 'steps': 38707, 'loss/train': 2.1986300945281982} 02/26/2022 03:41:07 - INFO - codeparrot_training - Step 38708: {'lr': 6.522442896997801e-05, 'samples': 19819008, 'steps': 38708, 'loss/train': 1.463073492050171} 02/26/2022 03:41:11 - INFO - codeparrot_training - Step 38709: {'lr': 6.521340772260791e-05, 'samples': 19819520, 'steps': 38709, 'loss/train': 1.7746306657791138} 02/26/2022 03:41:16 - INFO - codeparrot_training - Step 38710: {'lr': 6.520238726680477e-05, 'samples': 19820032, 'steps': 38710, 'loss/train': 1.5199174880981445} 02/26/2022 03:41:20 - INFO - codeparrot_training - Step 38711: {'lr': 6.519136760261598e-05, 'samples': 19820544, 'steps': 38711, 'loss/train': 1.934628963470459} 02/26/2022 03:41:25 - INFO - codeparrot_training - Step 38712: {'lr': 6.518034873008863e-05, 'samples': 19821056, 'steps': 38712, 'loss/train': 3.012913942337036} 02/26/2022 03:41:29 - INFO - codeparrot_training - Step 38713: {'lr': 6.516933064927012e-05, 'samples': 19821568, 'steps': 38713, 'loss/train': 2.005026340484619} 02/26/2022 03:41:34 - INFO - codeparrot_training - Step 38714: {'lr': 6.515831336020736e-05, 'samples': 19822080, 'steps': 38714, 'loss/train': 1.9883646965026855} 02/26/2022 03:41:37 - INFO - codeparrot_training - Step 38715: {'lr': 6.514729686294776e-05, 'samples': 19822592, 'steps': 38715, 'loss/train': 1.7896002531051636} 02/26/2022 03:41:43 - INFO - codeparrot_training - Step 38716: {'lr': 6.513628115753837e-05, 'samples': 19823104, 'steps': 38716, 'loss/train': 0.18262216448783875} 02/26/2022 03:41:46 - INFO - codeparrot_training - Step 38717: {'lr': 6.512526624402651e-05, 'samples': 19823616, 'steps': 38717, 'loss/train': 1.976986050605774} 02/26/2022 03:41:54 - INFO - codeparrot_training - Step 38718: {'lr': 6.511425212245933e-05, 'samples': 19824128, 'steps': 38718, 'loss/train': 2.043869733810425} 02/26/2022 03:41:58 - INFO - codeparrot_training - Step 38719: {'lr': 6.510323879288399e-05, 'samples': 19824640, 'steps': 38719, 'loss/train': 1.5984045267105103} 02/26/2022 03:42:01 - INFO - codeparrot_training - Step 38720: {'lr': 6.509222625534755e-05, 'samples': 19825152, 'steps': 38720, 'loss/train': 1.6781357526779175} 02/26/2022 03:42:07 - INFO - codeparrot_training - Step 38721: {'lr': 6.50812145098974e-05, 'samples': 19825664, 'steps': 38721, 'loss/train': 1.5842525959014893} 02/26/2022 03:42:10 - INFO - codeparrot_training - Step 38722: {'lr': 6.507020355658061e-05, 'samples': 19826176, 'steps': 38722, 'loss/train': 0.9933396577835083} 02/26/2022 03:42:16 - INFO - codeparrot_training - Step 38723: {'lr': 6.505919339544435e-05, 'samples': 19826688, 'steps': 38723, 'loss/train': 1.615172028541565} 02/26/2022 03:42:19 - INFO - codeparrot_training - Step 38724: {'lr': 6.50481840265357e-05, 'samples': 19827200, 'steps': 38724, 'loss/train': 0.9877386093139648} 02/26/2022 03:42:25 - INFO - codeparrot_training - Step 38725: {'lr': 6.503717544990198e-05, 'samples': 19827712, 'steps': 38725, 'loss/train': 2.0139565467834473} 02/26/2022 03:42:28 - INFO - codeparrot_training - Step 38726: {'lr': 6.502616766559019e-05, 'samples': 19828224, 'steps': 38726, 'loss/train': 1.6699538230895996} 02/26/2022 03:42:34 - INFO - codeparrot_training - Step 38727: {'lr': 6.501516067364771e-05, 'samples': 19828736, 'steps': 38727, 'loss/train': 1.4177770614624023} 02/26/2022 03:42:37 - INFO - codeparrot_training - Step 38728: {'lr': 6.500415447412139e-05, 'samples': 19829248, 'steps': 38728, 'loss/train': 1.5772100687026978} 02/26/2022 03:42:45 - INFO - codeparrot_training - Step 38729: {'lr': 6.499314906705862e-05, 'samples': 19829760, 'steps': 38729, 'loss/train': 1.6493568420410156} 02/26/2022 03:42:48 - INFO - codeparrot_training - Step 38730: {'lr': 6.498214445250636e-05, 'samples': 19830272, 'steps': 38730, 'loss/train': 0.5779489874839783} 02/26/2022 03:42:54 - INFO - codeparrot_training - Step 38731: {'lr': 6.497114063051198e-05, 'samples': 19830784, 'steps': 38731, 'loss/train': 2.2452361583709717} 02/26/2022 03:42:57 - INFO - codeparrot_training - Step 38732: {'lr': 6.496013760112235e-05, 'samples': 19831296, 'steps': 38732, 'loss/train': 1.1919969320297241} 02/26/2022 03:43:03 - INFO - codeparrot_training - Step 38733: {'lr': 6.494913536438483e-05, 'samples': 19831808, 'steps': 38733, 'loss/train': 2.563559055328369} 02/26/2022 03:43:06 - INFO - codeparrot_training - Step 38734: {'lr': 6.493813392034642e-05, 'samples': 19832320, 'steps': 38734, 'loss/train': 1.249554991722107} 02/26/2022 03:43:12 - INFO - codeparrot_training - Step 38735: {'lr': 6.492713326905428e-05, 'samples': 19832832, 'steps': 38735, 'loss/train': 2.5328257083892822} 02/26/2022 03:43:15 - INFO - codeparrot_training - Step 38736: {'lr': 6.491613341055547e-05, 'samples': 19833344, 'steps': 38736, 'loss/train': 1.0427772998809814} 02/26/2022 03:43:21 - INFO - codeparrot_training - Step 38737: {'lr': 6.490513434489726e-05, 'samples': 19833856, 'steps': 38737, 'loss/train': 2.762497901916504} 02/26/2022 03:43:24 - INFO - codeparrot_training - Step 38738: {'lr': 6.489413607212666e-05, 'samples': 19834368, 'steps': 38738, 'loss/train': 3.5796849727630615} 02/26/2022 03:43:31 - INFO - codeparrot_training - Step 38739: {'lr': 6.488313859229073e-05, 'samples': 19834880, 'steps': 38739, 'loss/train': 1.401711344718933} 02/26/2022 03:43:35 - INFO - codeparrot_training - Step 38740: {'lr': 6.487214190543675e-05, 'samples': 19835392, 'steps': 38740, 'loss/train': 0.5993425846099854} 02/26/2022 03:43:40 - INFO - codeparrot_training - Step 38741: {'lr': 6.486114601161172e-05, 'samples': 19835904, 'steps': 38741, 'loss/train': 1.3723808526992798} 02/26/2022 03:43:44 - INFO - codeparrot_training - Step 38742: {'lr': 6.485015091086275e-05, 'samples': 19836416, 'steps': 38742, 'loss/train': 1.9689902067184448} 02/26/2022 03:43:49 - INFO - codeparrot_training - Step 38743: {'lr': 6.483915660323686e-05, 'samples': 19836928, 'steps': 38743, 'loss/train': 0.9907351136207581} 02/26/2022 03:43:53 - INFO - codeparrot_training - Step 38744: {'lr': 6.482816308878129e-05, 'samples': 19837440, 'steps': 38744, 'loss/train': 1.8297984600067139} 02/26/2022 03:43:58 - INFO - codeparrot_training - Step 38745: {'lr': 6.481717036754308e-05, 'samples': 19837952, 'steps': 38745, 'loss/train': 2.0965633392333984} 02/26/2022 03:44:01 - INFO - codeparrot_training - Step 38746: {'lr': 6.480617843956932e-05, 'samples': 19838464, 'steps': 38746, 'loss/train': 1.1998233795166016} 02/26/2022 03:44:07 - INFO - codeparrot_training - Step 38747: {'lr': 6.479518730490702e-05, 'samples': 19838976, 'steps': 38747, 'loss/train': 1.9072761535644531} 02/26/2022 03:44:10 - INFO - codeparrot_training - Step 38748: {'lr': 6.478419696360346e-05, 'samples': 19839488, 'steps': 38748, 'loss/train': 0.18662342429161072} 02/26/2022 03:44:16 - INFO - codeparrot_training - Step 38749: {'lr': 6.477320741570545e-05, 'samples': 19840000, 'steps': 38749, 'loss/train': 0.6318190097808838} 02/26/2022 03:44:19 - INFO - codeparrot_training - Step 38750: {'lr': 6.476221866126028e-05, 'samples': 19840512, 'steps': 38750, 'loss/train': 1.8449482917785645} 02/26/2022 03:44:27 - INFO - codeparrot_training - Step 38751: {'lr': 6.475123070031485e-05, 'samples': 19841024, 'steps': 38751, 'loss/train': 1.4020644426345825} 02/26/2022 03:44:30 - INFO - codeparrot_training - Step 38752: {'lr': 6.474024353291641e-05, 'samples': 19841536, 'steps': 38752, 'loss/train': 1.1708887815475464} 02/26/2022 03:44:36 - INFO - codeparrot_training - Step 38753: {'lr': 6.472925715911195e-05, 'samples': 19842048, 'steps': 38753, 'loss/train': 1.0398430824279785} 02/26/2022 03:44:39 - INFO - codeparrot_training - Step 38754: {'lr': 6.471827157894852e-05, 'samples': 19842560, 'steps': 38754, 'loss/train': 0.9458532333374023} 02/26/2022 03:44:45 - INFO - codeparrot_training - Step 38755: {'lr': 6.470728679247309e-05, 'samples': 19843072, 'steps': 38755, 'loss/train': 1.241991639137268} 02/26/2022 03:44:48 - INFO - codeparrot_training - Step 38756: {'lr': 6.46963027997329e-05, 'samples': 19843584, 'steps': 38756, 'loss/train': 1.2338367700576782} 02/26/2022 03:44:54 - INFO - codeparrot_training - Step 38757: {'lr': 6.468531960077489e-05, 'samples': 19844096, 'steps': 38757, 'loss/train': 2.082691192626953} 02/26/2022 03:44:57 - INFO - codeparrot_training - Step 38758: {'lr': 6.467433719564613e-05, 'samples': 19844608, 'steps': 38758, 'loss/train': 1.4974420070648193} 02/26/2022 03:45:03 - INFO - codeparrot_training - Step 38759: {'lr': 6.466335558439362e-05, 'samples': 19845120, 'steps': 38759, 'loss/train': 2.7987849712371826} 02/26/2022 03:45:06 - INFO - codeparrot_training - Step 38760: {'lr': 6.465237476706449e-05, 'samples': 19845632, 'steps': 38760, 'loss/train': 2.9598820209503174} 02/26/2022 03:45:12 - INFO - codeparrot_training - Step 38761: {'lr': 6.464139474370573e-05, 'samples': 19846144, 'steps': 38761, 'loss/train': 1.4517033100128174} 02/26/2022 03:45:16 - INFO - codeparrot_training - Step 38762: {'lr': 6.463041551436438e-05, 'samples': 19846656, 'steps': 38762, 'loss/train': 0.6874664425849915} 02/26/2022 03:45:21 - INFO - codeparrot_training - Step 38763: {'lr': 6.46194370790874e-05, 'samples': 19847168, 'steps': 38763, 'loss/train': 2.9032158851623535} 02/26/2022 03:45:25 - INFO - codeparrot_training - Step 38764: {'lr': 6.460845943792195e-05, 'samples': 19847680, 'steps': 38764, 'loss/train': 1.7167894840240479} 02/26/2022 03:45:32 - INFO - codeparrot_training - Step 38765: {'lr': 6.459748259091493e-05, 'samples': 19848192, 'steps': 38765, 'loss/train': 1.9749704599380493} 02/26/2022 03:45:35 - INFO - codeparrot_training - Step 38766: {'lr': 6.458650653811357e-05, 'samples': 19848704, 'steps': 38766, 'loss/train': 0.9659585356712341} 02/26/2022 03:45:41 - INFO - codeparrot_training - Step 38767: {'lr': 6.457553127956458e-05, 'samples': 19849216, 'steps': 38767, 'loss/train': 1.6163103580474854} 02/26/2022 03:45:44 - INFO - codeparrot_training - Step 38768: {'lr': 6.456455681531522e-05, 'samples': 19849728, 'steps': 38768, 'loss/train': 2.5894775390625} 02/26/2022 03:45:50 - INFO - codeparrot_training - Step 38769: {'lr': 6.455358314541235e-05, 'samples': 19850240, 'steps': 38769, 'loss/train': 1.1043782234191895} 02/26/2022 03:45:53 - INFO - codeparrot_training - Step 38770: {'lr': 6.454261026990315e-05, 'samples': 19850752, 'steps': 38770, 'loss/train': 1.5259125232696533} 02/26/2022 03:45:59 - INFO - codeparrot_training - Step 38771: {'lr': 6.453163818883443e-05, 'samples': 19851264, 'steps': 38771, 'loss/train': 1.2306572198867798} 02/26/2022 03:46:03 - INFO - codeparrot_training - Step 38772: {'lr': 6.452066690225331e-05, 'samples': 19851776, 'steps': 38772, 'loss/train': 1.5919548273086548} 02/26/2022 03:46:08 - INFO - codeparrot_training - Step 38773: {'lr': 6.450969641020671e-05, 'samples': 19852288, 'steps': 38773, 'loss/train': 2.2383909225463867} 02/26/2022 03:46:12 - INFO - codeparrot_training - Step 38774: {'lr': 6.449872671274179e-05, 'samples': 19852800, 'steps': 38774, 'loss/train': 0.7657734751701355} 02/26/2022 03:46:19 - INFO - codeparrot_training - Step 38775: {'lr': 6.448775780990529e-05, 'samples': 19853312, 'steps': 38775, 'loss/train': 0.9641997218132019} 02/26/2022 03:46:22 - INFO - codeparrot_training - Step 38776: {'lr': 6.44767897017444e-05, 'samples': 19853824, 'steps': 38776, 'loss/train': 2.39081072807312} 02/26/2022 03:46:28 - INFO - codeparrot_training - Step 38777: {'lr': 6.446582238830597e-05, 'samples': 19854336, 'steps': 38777, 'loss/train': 2.127990484237671} 02/26/2022 03:46:31 - INFO - codeparrot_training - Step 38778: {'lr': 6.445485586963709e-05, 'samples': 19854848, 'steps': 38778, 'loss/train': 2.1040003299713135} 02/26/2022 03:46:37 - INFO - codeparrot_training - Step 38779: {'lr': 6.444389014578473e-05, 'samples': 19855360, 'steps': 38779, 'loss/train': 8.964900970458984} 02/26/2022 03:46:40 - INFO - codeparrot_training - Step 38780: {'lr': 6.443292521679578e-05, 'samples': 19855872, 'steps': 38780, 'loss/train': 0.41744714975357056} 02/26/2022 03:46:46 - INFO - codeparrot_training - Step 38781: {'lr': 6.442196108271727e-05, 'samples': 19856384, 'steps': 38781, 'loss/train': 1.8386110067367554} 02/26/2022 03:46:50 - INFO - codeparrot_training - Step 38782: {'lr': 6.44109977435961e-05, 'samples': 19856896, 'steps': 38782, 'loss/train': 1.5573471784591675} 02/26/2022 03:46:55 - INFO - codeparrot_training - Step 38783: {'lr': 6.440003519947935e-05, 'samples': 19857408, 'steps': 38783, 'loss/train': 1.8448214530944824} 02/26/2022 03:46:59 - INFO - codeparrot_training - Step 38784: {'lr': 6.438907345041389e-05, 'samples': 19857920, 'steps': 38784, 'loss/train': 1.2363420724868774} 02/26/2022 03:47:06 - INFO - codeparrot_training - Step 38785: {'lr': 6.437811249644673e-05, 'samples': 19858432, 'steps': 38785, 'loss/train': 1.7648439407348633} 02/26/2022 03:47:09 - INFO - codeparrot_training - Step 38786: {'lr': 6.43671523376247e-05, 'samples': 19858944, 'steps': 38786, 'loss/train': 1.922467589378357} 02/26/2022 03:47:15 - INFO - codeparrot_training - Step 38787: {'lr': 6.435619297399492e-05, 'samples': 19859456, 'steps': 38787, 'loss/train': 0.22969470918178558} 02/26/2022 03:47:18 - INFO - codeparrot_training - Step 38788: {'lr': 6.434523440560427e-05, 'samples': 19859968, 'steps': 38788, 'loss/train': 1.9244303703308105} 02/26/2022 03:47:24 - INFO - codeparrot_training - Step 38789: {'lr': 6.433427663249969e-05, 'samples': 19860480, 'steps': 38789, 'loss/train': 0.6102625131607056} 02/26/2022 03:47:27 - INFO - codeparrot_training - Step 38790: {'lr': 6.432331965472801e-05, 'samples': 19860992, 'steps': 38790, 'loss/train': 1.2116776704788208} 02/26/2022 03:47:33 - INFO - codeparrot_training - Step 38791: {'lr': 6.431236347233638e-05, 'samples': 19861504, 'steps': 38791, 'loss/train': 1.4490152597427368} 02/26/2022 03:47:36 - INFO - codeparrot_training - Step 38792: {'lr': 6.43014080853716e-05, 'samples': 19862016, 'steps': 38792, 'loss/train': 2.329697370529175} 02/26/2022 03:47:42 - INFO - codeparrot_training - Step 38793: {'lr': 6.429045349388063e-05, 'samples': 19862528, 'steps': 38793, 'loss/train': 1.5675824880599976} 02/26/2022 03:47:45 - INFO - codeparrot_training - Step 38794: {'lr': 6.42794996979103e-05, 'samples': 19863040, 'steps': 38794, 'loss/train': 1.5963491201400757} 02/26/2022 03:47:51 - INFO - codeparrot_training - Step 38795: {'lr': 6.426854669750773e-05, 'samples': 19863552, 'steps': 38795, 'loss/train': 2.4156441688537598} 02/26/2022 03:47:54 - INFO - codeparrot_training - Step 38796: {'lr': 6.42575944927197e-05, 'samples': 19864064, 'steps': 38796, 'loss/train': 0.5700355768203735} 02/26/2022 03:48:01 - INFO - codeparrot_training - Step 38797: {'lr': 6.424664308359315e-05, 'samples': 19864576, 'steps': 38797, 'loss/train': 1.6825120449066162} 02/26/2022 03:48:05 - INFO - codeparrot_training - Step 38798: {'lr': 6.423569247017493e-05, 'samples': 19865088, 'steps': 38798, 'loss/train': 1.7825416326522827} 02/26/2022 03:48:10 - INFO - codeparrot_training - Step 38799: {'lr': 6.42247426525121e-05, 'samples': 19865600, 'steps': 38799, 'loss/train': 1.7090396881103516} 02/26/2022 03:48:14 - INFO - codeparrot_training - Step 38800: {'lr': 6.421379363065141e-05, 'samples': 19866112, 'steps': 38800, 'loss/train': 1.4217395782470703} 02/26/2022 03:48:20 - INFO - codeparrot_training - Step 38801: {'lr': 6.420284540463998e-05, 'samples': 19866624, 'steps': 38801, 'loss/train': 1.579977035522461} 02/26/2022 03:48:23 - INFO - codeparrot_training - Step 38802: {'lr': 6.41918979745244e-05, 'samples': 19867136, 'steps': 38802, 'loss/train': 2.0598113536834717} 02/26/2022 03:48:29 - INFO - codeparrot_training - Step 38803: {'lr': 6.418095134035182e-05, 'samples': 19867648, 'steps': 38803, 'loss/train': 1.0589210987091064} 02/26/2022 03:48:32 - INFO - codeparrot_training - Step 38804: {'lr': 6.417000550216896e-05, 'samples': 19868160, 'steps': 38804, 'loss/train': 0.8834738731384277} 02/26/2022 03:48:37 - INFO - codeparrot_training - Step 38805: {'lr': 6.415906046002293e-05, 'samples': 19868672, 'steps': 38805, 'loss/train': 1.4800336360931396} 02/26/2022 03:48:41 - INFO - codeparrot_training - Step 38806: {'lr': 6.414811621396033e-05, 'samples': 19869184, 'steps': 38806, 'loss/train': 1.513084888458252} 02/26/2022 03:48:48 - INFO - codeparrot_training - Step 38807: {'lr': 6.413717276402828e-05, 'samples': 19869696, 'steps': 38807, 'loss/train': 1.9747583866119385} 02/26/2022 03:48:52 - INFO - codeparrot_training - Step 38808: {'lr': 6.412623011027346e-05, 'samples': 19870208, 'steps': 38808, 'loss/train': 0.9450200796127319} 02/26/2022 03:48:57 - INFO - codeparrot_training - Step 38809: {'lr': 6.411528825274301e-05, 'samples': 19870720, 'steps': 38809, 'loss/train': 1.3941991329193115} 02/26/2022 03:49:01 - INFO - codeparrot_training - Step 38810: {'lr': 6.410434719148347e-05, 'samples': 19871232, 'steps': 38810, 'loss/train': 1.262963891029358} 02/26/2022 03:49:06 - INFO - codeparrot_training - Step 38811: {'lr': 6.409340692654198e-05, 'samples': 19871744, 'steps': 38811, 'loss/train': 2.3142611980438232} 02/26/2022 03:49:10 - INFO - codeparrot_training - Step 38812: {'lr': 6.408246745796523e-05, 'samples': 19872256, 'steps': 38812, 'loss/train': 0.08465367555618286} 02/26/2022 03:49:15 - INFO - codeparrot_training - Step 38813: {'lr': 6.407152878580023e-05, 'samples': 19872768, 'steps': 38813, 'loss/train': 2.168731689453125} 02/26/2022 03:49:19 - INFO - codeparrot_training - Step 38814: {'lr': 6.406059091009373e-05, 'samples': 19873280, 'steps': 38814, 'loss/train': 1.2257672548294067} 02/26/2022 03:49:24 - INFO - codeparrot_training - Step 38815: {'lr': 6.404965383089265e-05, 'samples': 19873792, 'steps': 38815, 'loss/train': 0.5894840359687805} 02/26/2022 03:49:28 - INFO - codeparrot_training - Step 38816: {'lr': 6.403871754824372e-05, 'samples': 19874304, 'steps': 38816, 'loss/train': 0.60886150598526} 02/26/2022 03:49:33 - INFO - codeparrot_training - Step 38817: {'lr': 6.402778206219395e-05, 'samples': 19874816, 'steps': 38817, 'loss/train': 2.5555009841918945} 02/26/2022 03:49:37 - INFO - codeparrot_training - Step 38818: {'lr': 6.401684737279009e-05, 'samples': 19875328, 'steps': 38818, 'loss/train': 2.7031188011169434} 02/26/2022 03:49:42 - INFO - codeparrot_training - Step 38819: {'lr': 6.400591348007904e-05, 'samples': 19875840, 'steps': 38819, 'loss/train': 1.8502089977264404} 02/26/2022 03:49:46 - INFO - codeparrot_training - Step 38820: {'lr': 6.399498038410747e-05, 'samples': 19876352, 'steps': 38820, 'loss/train': 1.0791562795639038} 02/26/2022 03:49:51 - INFO - codeparrot_training - Step 38821: {'lr': 6.398404808492245e-05, 'samples': 19876864, 'steps': 38821, 'loss/train': 1.3779817819595337} 02/26/2022 03:49:55 - INFO - codeparrot_training - Step 38822: {'lr': 6.397311658257071e-05, 'samples': 19877376, 'steps': 38822, 'loss/train': 1.263892650604248} 02/26/2022 03:50:02 - INFO - codeparrot_training - Step 38823: {'lr': 6.396218587709904e-05, 'samples': 19877888, 'steps': 38823, 'loss/train': 1.3508083820343018} 02/26/2022 03:50:05 - INFO - codeparrot_training - Step 38824: {'lr': 6.395125596855425e-05, 'samples': 19878400, 'steps': 38824, 'loss/train': 0.11727475374937057} 02/26/2022 03:50:11 - INFO - codeparrot_training - Step 38825: {'lr': 6.394032685698325e-05, 'samples': 19878912, 'steps': 38825, 'loss/train': 1.6723921298980713} 02/26/2022 03:50:14 - INFO - codeparrot_training - Step 38826: {'lr': 6.392939854243282e-05, 'samples': 19879424, 'steps': 38826, 'loss/train': 2.1133294105529785} 02/26/2022 03:50:20 - INFO - codeparrot_training - Step 38827: {'lr': 6.391847102494976e-05, 'samples': 19879936, 'steps': 38827, 'loss/train': 0.07957497984170914} 02/26/2022 03:50:23 - INFO - codeparrot_training - Step 38828: {'lr': 6.390754430458081e-05, 'samples': 19880448, 'steps': 38828, 'loss/train': 2.9327425956726074} 02/26/2022 03:50:29 - INFO - codeparrot_training - Step 38829: {'lr': 6.389661838137292e-05, 'samples': 19880960, 'steps': 38829, 'loss/train': 1.6273353099822998} 02/26/2022 03:50:32 - INFO - codeparrot_training - Step 38830: {'lr': 6.388569325537283e-05, 'samples': 19881472, 'steps': 38830, 'loss/train': 1.8934136629104614} 02/26/2022 03:50:38 - INFO - codeparrot_training - Step 38831: {'lr': 6.387476892662733e-05, 'samples': 19881984, 'steps': 38831, 'loss/train': 1.2015880346298218} 02/26/2022 03:50:41 - INFO - codeparrot_training - Step 38832: {'lr': 6.38638453951832e-05, 'samples': 19882496, 'steps': 38832, 'loss/train': 1.5802578926086426} 02/26/2022 03:50:48 - INFO - codeparrot_training - Step 38833: {'lr': 6.385292266108719e-05, 'samples': 19883008, 'steps': 38833, 'loss/train': 1.0243046283721924} 02/26/2022 03:50:52 - INFO - codeparrot_training - Step 38834: {'lr': 6.384200072438623e-05, 'samples': 19883520, 'steps': 38834, 'loss/train': 1.5294610261917114} 02/26/2022 03:50:57 - INFO - codeparrot_training - Step 38835: {'lr': 6.383107958512696e-05, 'samples': 19884032, 'steps': 38835, 'loss/train': 1.7044965028762817} 02/26/2022 03:51:01 - INFO - codeparrot_training - Step 38836: {'lr': 6.382015924335636e-05, 'samples': 19884544, 'steps': 38836, 'loss/train': 1.6806957721710205} 02/26/2022 03:51:06 - INFO - codeparrot_training - Step 38837: {'lr': 6.380923969912094e-05, 'samples': 19885056, 'steps': 38837, 'loss/train': 2.737531900405884} 02/26/2022 03:51:10 - INFO - codeparrot_training - Step 38838: {'lr': 6.379832095246767e-05, 'samples': 19885568, 'steps': 38838, 'loss/train': 1.864533543586731} 02/26/2022 03:51:15 - INFO - codeparrot_training - Step 38839: {'lr': 6.378740300344324e-05, 'samples': 19886080, 'steps': 38839, 'loss/train': 1.286806583404541} 02/26/2022 03:51:19 - INFO - codeparrot_training - Step 38840: {'lr': 6.377648585209455e-05, 'samples': 19886592, 'steps': 38840, 'loss/train': 0.8436279892921448} 02/26/2022 03:51:24 - INFO - codeparrot_training - Step 38841: {'lr': 6.376556949846813e-05, 'samples': 19887104, 'steps': 38841, 'loss/train': 2.421673536300659} 02/26/2022 03:51:28 - INFO - codeparrot_training - Step 38842: {'lr': 6.375465394261096e-05, 'samples': 19887616, 'steps': 38842, 'loss/train': 1.671796441078186} 02/26/2022 03:51:34 - INFO - codeparrot_training - Step 38843: {'lr': 6.374373918456964e-05, 'samples': 19888128, 'steps': 38843, 'loss/train': 2.339667558670044} 02/26/2022 03:51:38 - INFO - codeparrot_training - Step 38844: {'lr': 6.37328252243911e-05, 'samples': 19888640, 'steps': 38844, 'loss/train': 2.1658670902252197} 02/26/2022 03:51:43 - INFO - codeparrot_training - Step 38845: {'lr': 6.372191206212188e-05, 'samples': 19889152, 'steps': 38845, 'loss/train': 2.45666241645813} 02/26/2022 03:51:48 - INFO - codeparrot_training - Step 38846: {'lr': 6.371099969780891e-05, 'samples': 19889664, 'steps': 38846, 'loss/train': 2.5400333404541016} 02/26/2022 03:51:52 - INFO - codeparrot_training - Step 38847: {'lr': 6.37000881314988e-05, 'samples': 19890176, 'steps': 38847, 'loss/train': 2.4450693130493164} 02/26/2022 03:51:57 - INFO - codeparrot_training - Step 38848: {'lr': 6.368917736323843e-05, 'samples': 19890688, 'steps': 38848, 'loss/train': 1.0211149454116821} 02/26/2022 03:52:01 - INFO - codeparrot_training - Step 38849: {'lr': 6.367826739307445e-05, 'samples': 19891200, 'steps': 38849, 'loss/train': 1.6614383459091187} 02/26/2022 03:52:06 - INFO - codeparrot_training - Step 38850: {'lr': 6.36673582210536e-05, 'samples': 19891712, 'steps': 38850, 'loss/train': 1.6359202861785889} 02/26/2022 03:52:10 - INFO - codeparrot_training - Step 38851: {'lr': 6.365644984722254e-05, 'samples': 19892224, 'steps': 38851, 'loss/train': 1.525418996810913} 02/26/2022 03:52:15 - INFO - codeparrot_training - Step 38852: {'lr': 6.364554227162819e-05, 'samples': 19892736, 'steps': 38852, 'loss/train': 2.3382010459899902} 02/26/2022 03:52:19 - INFO - codeparrot_training - Step 38853: {'lr': 6.363463549431714e-05, 'samples': 19893248, 'steps': 38853, 'loss/train': 1.3372362852096558} 02/26/2022 03:52:25 - INFO - codeparrot_training - Step 38854: {'lr': 6.362372951533613e-05, 'samples': 19893760, 'steps': 38854, 'loss/train': 1.1186985969543457} 02/26/2022 03:52:28 - INFO - codeparrot_training - Step 38855: {'lr': 6.361282433473181e-05, 'samples': 19894272, 'steps': 38855, 'loss/train': 1.403040885925293} 02/26/2022 03:52:33 - INFO - codeparrot_training - Step 38856: {'lr': 6.360191995255104e-05, 'samples': 19894784, 'steps': 38856, 'loss/train': 1.9209054708480835} 02/26/2022 03:52:37 - INFO - codeparrot_training - Step 38857: {'lr': 6.359101636884045e-05, 'samples': 19895296, 'steps': 38857, 'loss/train': 2.063040018081665} 02/26/2022 03:52:43 - INFO - codeparrot_training - Step 38858: {'lr': 6.358011358364676e-05, 'samples': 19895808, 'steps': 38858, 'loss/train': 1.859063982963562} 02/26/2022 03:52:47 - INFO - codeparrot_training - Step 38859: {'lr': 6.356921159701662e-05, 'samples': 19896320, 'steps': 38859, 'loss/train': 1.6190849542617798} 02/26/2022 03:52:52 - INFO - codeparrot_training - Step 38860: {'lr': 6.355831040899682e-05, 'samples': 19896832, 'steps': 38860, 'loss/train': 1.7135318517684937} 02/26/2022 03:52:55 - INFO - codeparrot_training - Step 38861: {'lr': 6.354741001963396e-05, 'samples': 19897344, 'steps': 38861, 'loss/train': 0.18433037400245667} 02/26/2022 03:53:01 - INFO - codeparrot_training - Step 38862: {'lr': 6.353651042897493e-05, 'samples': 19897856, 'steps': 38862, 'loss/train': 1.3983616828918457} 02/26/2022 03:53:05 - INFO - codeparrot_training - Step 38863: {'lr': 6.352561163706613e-05, 'samples': 19898368, 'steps': 38863, 'loss/train': 1.2397453784942627} 02/26/2022 03:53:10 - INFO - codeparrot_training - Step 38864: {'lr': 6.351471364395448e-05, 'samples': 19898880, 'steps': 38864, 'loss/train': 1.0877991914749146} 02/26/2022 03:53:14 - INFO - codeparrot_training - Step 38865: {'lr': 6.350381644968659e-05, 'samples': 19899392, 'steps': 38865, 'loss/train': 1.620319128036499} 02/26/2022 03:53:19 - INFO - codeparrot_training - Step 38866: {'lr': 6.34929200543091e-05, 'samples': 19899904, 'steps': 38866, 'loss/train': 2.918816089630127} 02/26/2022 03:53:23 - INFO - codeparrot_training - Step 38867: {'lr': 6.348202445786866e-05, 'samples': 19900416, 'steps': 38867, 'loss/train': 1.7316852807998657} 02/26/2022 03:53:29 - INFO - codeparrot_training - Step 38868: {'lr': 6.347112966041208e-05, 'samples': 19900928, 'steps': 38868, 'loss/train': 0.8470566868782043} 02/26/2022 03:53:32 - INFO - codeparrot_training - Step 38869: {'lr': 6.346023566198594e-05, 'samples': 19901440, 'steps': 38869, 'loss/train': 1.8697400093078613} 02/26/2022 03:53:38 - INFO - codeparrot_training - Step 38870: {'lr': 6.344934246263692e-05, 'samples': 19901952, 'steps': 38870, 'loss/train': 2.8419699668884277} 02/26/2022 03:53:41 - INFO - codeparrot_training - Step 38871: {'lr': 6.343845006241161e-05, 'samples': 19902464, 'steps': 38871, 'loss/train': 3.0880727767944336} 02/26/2022 03:53:47 - INFO - codeparrot_training - Step 38872: {'lr': 6.342755846135684e-05, 'samples': 19902976, 'steps': 38872, 'loss/train': 1.8982683420181274} 02/26/2022 03:53:50 - INFO - codeparrot_training - Step 38873: {'lr': 6.341666765951912e-05, 'samples': 19903488, 'steps': 38873, 'loss/train': 1.6093734502792358} 02/26/2022 03:53:56 - INFO - codeparrot_training - Step 38874: {'lr': 6.340577765694511e-05, 'samples': 19904000, 'steps': 38874, 'loss/train': 1.942063331604004} 02/26/2022 03:53:59 - INFO - codeparrot_training - Step 38875: {'lr': 6.339488845368155e-05, 'samples': 19904512, 'steps': 38875, 'loss/train': 1.6383899450302124} 02/26/2022 03:54:05 - INFO - codeparrot_training - Step 38876: {'lr': 6.338400004977505e-05, 'samples': 19905024, 'steps': 38876, 'loss/train': 1.677234411239624} 02/26/2022 03:54:08 - INFO - codeparrot_training - Step 38877: {'lr': 6.337311244527221e-05, 'samples': 19905536, 'steps': 38877, 'loss/train': 1.2173211574554443} 02/26/2022 03:54:15 - INFO - codeparrot_training - Step 38878: {'lr': 6.336222564021965e-05, 'samples': 19906048, 'steps': 38878, 'loss/train': 1.966973066329956} 02/26/2022 03:54:18 - INFO - codeparrot_training - Step 38879: {'lr': 6.33513396346642e-05, 'samples': 19906560, 'steps': 38879, 'loss/train': 1.4125511646270752} 02/26/2022 03:54:24 - INFO - codeparrot_training - Step 38880: {'lr': 6.334045442865219e-05, 'samples': 19907072, 'steps': 38880, 'loss/train': 1.6864938735961914} 02/26/2022 03:54:27 - INFO - codeparrot_training - Step 38881: {'lr': 6.332957002223047e-05, 'samples': 19907584, 'steps': 38881, 'loss/train': 2.026339054107666} 02/26/2022 03:54:33 - INFO - codeparrot_training - Step 38882: {'lr': 6.331868641544553e-05, 'samples': 19908096, 'steps': 38882, 'loss/train': 2.478119373321533} 02/26/2022 03:54:37 - INFO - codeparrot_training - Step 38883: {'lr': 6.330780360834418e-05, 'samples': 19908608, 'steps': 38883, 'loss/train': 1.8960620164871216} 02/26/2022 03:54:42 - INFO - codeparrot_training - Step 38884: {'lr': 6.329692160097281e-05, 'samples': 19909120, 'steps': 38884, 'loss/train': 1.472038984298706} 02/26/2022 03:54:46 - INFO - codeparrot_training - Step 38885: {'lr': 6.328604039337818e-05, 'samples': 19909632, 'steps': 38885, 'loss/train': 1.2072947025299072} 02/26/2022 03:54:51 - INFO - codeparrot_training - Step 38886: {'lr': 6.327515998560681e-05, 'samples': 19910144, 'steps': 38886, 'loss/train': 2.57529878616333} 02/26/2022 03:54:55 - INFO - codeparrot_training - Step 38887: {'lr': 6.326428037770545e-05, 'samples': 19910656, 'steps': 38887, 'loss/train': 2.057734251022339} 02/26/2022 03:55:01 - INFO - codeparrot_training - Step 38888: {'lr': 6.325340156972059e-05, 'samples': 19911168, 'steps': 38888, 'loss/train': 1.392457365989685} 02/26/2022 03:55:04 - INFO - codeparrot_training - Step 38889: {'lr': 6.324252356169888e-05, 'samples': 19911680, 'steps': 38889, 'loss/train': 1.569298505783081} 02/26/2022 03:55:10 - INFO - codeparrot_training - Step 38890: {'lr': 6.32316463536868e-05, 'samples': 19912192, 'steps': 38890, 'loss/train': 1.3328437805175781} 02/26/2022 03:55:13 - INFO - codeparrot_training - Step 38891: {'lr': 6.322076994573115e-05, 'samples': 19912704, 'steps': 38891, 'loss/train': 1.5142974853515625} 02/26/2022 03:55:19 - INFO - codeparrot_training - Step 38892: {'lr': 6.320989433787838e-05, 'samples': 19913216, 'steps': 38892, 'loss/train': 1.4834496974945068} 02/26/2022 03:55:22 - INFO - codeparrot_training - Step 38893: {'lr': 6.319901953017515e-05, 'samples': 19913728, 'steps': 38893, 'loss/train': 1.5543599128723145} 02/26/2022 03:55:28 - INFO - codeparrot_training - Step 38894: {'lr': 6.318814552266788e-05, 'samples': 19914240, 'steps': 38894, 'loss/train': 1.092104196548462} 02/26/2022 03:55:33 - INFO - codeparrot_training - Step 38895: {'lr': 6.317727231540338e-05, 'samples': 19914752, 'steps': 38895, 'loss/train': 1.6777210235595703} 02/26/2022 03:55:37 - INFO - codeparrot_training - Step 38896: {'lr': 6.316639990842804e-05, 'samples': 19915264, 'steps': 38896, 'loss/train': 2.6742794513702393} 02/26/2022 03:55:42 - INFO - codeparrot_training - Step 38897: {'lr': 6.315552830178867e-05, 'samples': 19915776, 'steps': 38897, 'loss/train': 1.963529109954834} 02/26/2022 03:55:46 - INFO - codeparrot_training - Step 38898: {'lr': 6.314465749553152e-05, 'samples': 19916288, 'steps': 38898, 'loss/train': 2.05869722366333} 02/26/2022 03:55:51 - INFO - codeparrot_training - Step 38899: {'lr': 6.313378748970344e-05, 'samples': 19916800, 'steps': 38899, 'loss/train': 0.3517754077911377} 02/26/2022 03:55:55 - INFO - codeparrot_training - Step 38900: {'lr': 6.312291828435076e-05, 'samples': 19917312, 'steps': 38900, 'loss/train': 2.547058343887329} 02/26/2022 03:56:01 - INFO - codeparrot_training - Step 38901: {'lr': 6.311204987952033e-05, 'samples': 19917824, 'steps': 38901, 'loss/train': 2.3607535362243652} 02/26/2022 03:56:04 - INFO - codeparrot_training - Step 38902: {'lr': 6.310118227525838e-05, 'samples': 19918336, 'steps': 38902, 'loss/train': 2.202324628829956} 02/26/2022 03:56:08 - INFO - codeparrot_training - Step 38903: {'lr': 6.30903154716117e-05, 'samples': 19918848, 'steps': 38903, 'loss/train': 2.181614875793457} 02/26/2022 03:56:14 - INFO - codeparrot_training - Step 38904: {'lr': 6.307944946862668e-05, 'samples': 19919360, 'steps': 38904, 'loss/train': 1.1483254432678223} 02/26/2022 03:56:20 - INFO - codeparrot_training - Step 38905: {'lr': 6.306858426635006e-05, 'samples': 19919872, 'steps': 38905, 'loss/train': 3.3665950298309326} 02/26/2022 03:56:23 - INFO - codeparrot_training - Step 38906: {'lr': 6.305771986482816e-05, 'samples': 19920384, 'steps': 38906, 'loss/train': 1.4355310201644897} 02/26/2022 03:56:29 - INFO - codeparrot_training - Step 38907: {'lr': 6.304685626410769e-05, 'samples': 19920896, 'steps': 38907, 'loss/train': 1.51449453830719} 02/26/2022 03:56:33 - INFO - codeparrot_training - Step 38908: {'lr': 6.303599346423505e-05, 'samples': 19921408, 'steps': 38908, 'loss/train': 3.7690417766571045} 02/26/2022 03:56:36 - INFO - codeparrot_training - Step 38909: {'lr': 6.302513146525693e-05, 'samples': 19921920, 'steps': 38909, 'loss/train': 1.0974311828613281} 02/26/2022 03:56:42 - INFO - codeparrot_training - Step 38910: {'lr': 6.301427026721975e-05, 'samples': 19922432, 'steps': 38910, 'loss/train': 1.3523627519607544} 02/26/2022 03:56:45 - INFO - codeparrot_training - Step 38911: {'lr': 6.300340987017009e-05, 'samples': 19922944, 'steps': 38911, 'loss/train': 1.7942637205123901} 02/26/2022 03:56:51 - INFO - codeparrot_training - Step 38912: {'lr': 6.299255027415443e-05, 'samples': 19923456, 'steps': 38912, 'loss/train': 7.154319763183594} 02/26/2022 03:56:54 - INFO - codeparrot_training - Step 38913: {'lr': 6.298169147921923e-05, 'samples': 19923968, 'steps': 38913, 'loss/train': 1.2122001647949219} 02/26/2022 03:57:01 - INFO - codeparrot_training - Step 38914: {'lr': 6.297083348541113e-05, 'samples': 19924480, 'steps': 38914, 'loss/train': 1.4609757661819458} 02/26/2022 03:57:04 - INFO - codeparrot_training - Step 38915: {'lr': 6.295997629277661e-05, 'samples': 19924992, 'steps': 38915, 'loss/train': 2.0608909130096436} 02/26/2022 03:57:10 - INFO - codeparrot_training - Step 38916: {'lr': 6.294911990136218e-05, 'samples': 19925504, 'steps': 38916, 'loss/train': 2.5247342586517334} 02/26/2022 03:57:13 - INFO - codeparrot_training - Step 38917: {'lr': 6.293826431121422e-05, 'samples': 19926016, 'steps': 38917, 'loss/train': 2.2232210636138916} 02/26/2022 03:57:19 - INFO - codeparrot_training - Step 38918: {'lr': 6.292740952237943e-05, 'samples': 19926528, 'steps': 38918, 'loss/train': 1.9103516340255737} 02/26/2022 03:57:24 - INFO - codeparrot_training - Step 38919: {'lr': 6.291655553490422e-05, 'samples': 19927040, 'steps': 38919, 'loss/train': 2.00417423248291} 02/26/2022 03:57:28 - INFO - codeparrot_training - Step 38920: {'lr': 6.290570234883505e-05, 'samples': 19927552, 'steps': 38920, 'loss/train': 0.45651569962501526} 02/26/2022 03:57:33 - INFO - codeparrot_training - Step 38921: {'lr': 6.28948499642184e-05, 'samples': 19928064, 'steps': 38921, 'loss/train': 2.1316864490509033} 02/26/2022 03:57:37 - INFO - codeparrot_training - Step 38922: {'lr': 6.288399838110087e-05, 'samples': 19928576, 'steps': 38922, 'loss/train': 1.490264892578125} 02/26/2022 03:57:43 - INFO - codeparrot_training - Step 38923: {'lr': 6.287314759952886e-05, 'samples': 19929088, 'steps': 38923, 'loss/train': 3.206657886505127} 02/26/2022 03:57:46 - INFO - codeparrot_training - Step 38924: {'lr': 6.286229761954887e-05, 'samples': 19929600, 'steps': 38924, 'loss/train': 1.715929627418518} 02/26/2022 03:57:52 - INFO - codeparrot_training - Step 38925: {'lr': 6.28514484412073e-05, 'samples': 19930112, 'steps': 38925, 'loss/train': 1.8856327533721924} 02/26/2022 03:57:55 - INFO - codeparrot_training - Step 38926: {'lr': 6.284060006455078e-05, 'samples': 19930624, 'steps': 38926, 'loss/train': 1.176323413848877} 02/26/2022 03:58:01 - INFO - codeparrot_training - Step 38927: {'lr': 6.28297524896257e-05, 'samples': 19931136, 'steps': 38927, 'loss/train': 1.6909675598144531} 02/26/2022 03:58:04 - INFO - codeparrot_training - Step 38928: {'lr': 6.281890571647852e-05, 'samples': 19931648, 'steps': 38928, 'loss/train': 2.348320960998535} 02/26/2022 03:58:10 - INFO - codeparrot_training - Step 38929: {'lr': 6.280805974515563e-05, 'samples': 19932160, 'steps': 38929, 'loss/train': 1.5610156059265137} 02/26/2022 03:58:13 - INFO - codeparrot_training - Step 38930: {'lr': 6.279721457570367e-05, 'samples': 19932672, 'steps': 38930, 'loss/train': 2.2120895385742188} 02/26/2022 03:58:19 - INFO - codeparrot_training - Step 38931: {'lr': 6.278637020816896e-05, 'samples': 19933184, 'steps': 38931, 'loss/train': 1.5010899305343628} 02/26/2022 03:58:22 - INFO - codeparrot_training - Step 38932: {'lr': 6.277552664259803e-05, 'samples': 19933696, 'steps': 38932, 'loss/train': 1.4336180686950684} 02/26/2022 03:58:29 - INFO - codeparrot_training - Step 38933: {'lr': 6.276468387903719e-05, 'samples': 19934208, 'steps': 38933, 'loss/train': 2.6342320442199707} 02/26/2022 03:58:32 - INFO - codeparrot_training - Step 38934: {'lr': 6.275384191753311e-05, 'samples': 19934720, 'steps': 38934, 'loss/train': 1.445743203163147} 02/26/2022 03:58:37 - INFO - codeparrot_training - Step 38935: {'lr': 6.274300075813199e-05, 'samples': 19935232, 'steps': 38935, 'loss/train': 1.3488141298294067} 02/26/2022 03:58:41 - INFO - codeparrot_training - Step 38936: {'lr': 6.273216040088056e-05, 'samples': 19935744, 'steps': 38936, 'loss/train': 1.3567650318145752} 02/26/2022 03:58:46 - INFO - codeparrot_training - Step 38937: {'lr': 6.272132084582496e-05, 'samples': 19936256, 'steps': 38937, 'loss/train': 1.5761371850967407} 02/26/2022 03:58:50 - INFO - codeparrot_training - Step 38938: {'lr': 6.27104820930118e-05, 'samples': 19936768, 'steps': 38938, 'loss/train': 2.4720654487609863} 02/26/2022 03:58:56 - INFO - codeparrot_training - Step 38939: {'lr': 6.269964414248741e-05, 'samples': 19937280, 'steps': 38939, 'loss/train': 1.7867281436920166} 02/26/2022 03:59:00 - INFO - codeparrot_training - Step 38940: {'lr': 6.26888069942984e-05, 'samples': 19937792, 'steps': 38940, 'loss/train': 1.887107253074646} 02/26/2022 03:59:03 - INFO - codeparrot_training - Step 38941: {'lr': 6.267797064849093e-05, 'samples': 19938304, 'steps': 38941, 'loss/train': 2.723895788192749} 02/26/2022 03:59:09 - INFO - codeparrot_training - Step 38942: {'lr': 6.266713510511163e-05, 'samples': 19938816, 'steps': 38942, 'loss/train': 1.1761444807052612} 02/26/2022 03:59:12 - INFO - codeparrot_training - Step 38943: {'lr': 6.265630036420675e-05, 'samples': 19939328, 'steps': 38943, 'loss/train': 2.514967918395996} 02/26/2022 03:59:18 - INFO - codeparrot_training - Step 38944: {'lr': 6.264546642582289e-05, 'samples': 19939840, 'steps': 38944, 'loss/train': 1.5851999521255493} 02/26/2022 03:59:21 - INFO - codeparrot_training - Step 38945: {'lr': 6.263463329000632e-05, 'samples': 19940352, 'steps': 38945, 'loss/train': 2.5215721130371094} 02/26/2022 03:59:27 - INFO - codeparrot_training - Step 38946: {'lr': 6.26238009568035e-05, 'samples': 19940864, 'steps': 38946, 'loss/train': 2.454012632369995} 02/26/2022 03:59:31 - INFO - codeparrot_training - Step 38947: {'lr': 6.261296942626077e-05, 'samples': 19941376, 'steps': 38947, 'loss/train': 1.1757171154022217} 02/26/2022 03:59:36 - INFO - codeparrot_training - Step 38948: {'lr': 6.260213869842462e-05, 'samples': 19941888, 'steps': 38948, 'loss/train': 1.0435585975646973} 02/26/2022 03:59:40 - INFO - codeparrot_training - Step 38949: {'lr': 6.259130877334143e-05, 'samples': 19942400, 'steps': 38949, 'loss/train': 1.7368721961975098} 02/26/2022 03:59:45 - INFO - codeparrot_training - Step 38950: {'lr': 6.258047965105753e-05, 'samples': 19942912, 'steps': 38950, 'loss/train': 1.3974124193191528} 02/26/2022 03:59:49 - INFO - codeparrot_training - Step 38951: {'lr': 6.25696513316193e-05, 'samples': 19943424, 'steps': 38951, 'loss/train': 1.3515630960464478} 02/26/2022 03:59:55 - INFO - codeparrot_training - Step 38952: {'lr': 6.255882381507324e-05, 'samples': 19943936, 'steps': 38952, 'loss/train': 1.0100599527359009} 02/26/2022 03:59:58 - INFO - codeparrot_training - Step 38953: {'lr': 6.254799710146564e-05, 'samples': 19944448, 'steps': 38953, 'loss/train': 2.2196245193481445} 02/26/2022 04:00:04 - INFO - codeparrot_training - Step 38954: {'lr': 6.25371711908429e-05, 'samples': 19944960, 'steps': 38954, 'loss/train': 1.175177812576294} 02/26/2022 04:00:07 - INFO - codeparrot_training - Step 38955: {'lr': 6.252634608325134e-05, 'samples': 19945472, 'steps': 38955, 'loss/train': 1.796678066253662} 02/26/2022 04:00:11 - INFO - codeparrot_training - Step 38956: {'lr': 6.251552177873742e-05, 'samples': 19945984, 'steps': 38956, 'loss/train': 2.1559677124023438} 02/26/2022 04:00:16 - INFO - codeparrot_training - Step 38957: {'lr': 6.250469827734751e-05, 'samples': 19946496, 'steps': 38957, 'loss/train': 1.7717361450195312} 02/26/2022 04:00:20 - INFO - codeparrot_training - Step 38958: {'lr': 6.249387557912792e-05, 'samples': 19947008, 'steps': 38958, 'loss/train': 2.8185548782348633} 02/26/2022 04:00:25 - INFO - codeparrot_training - Step 38959: {'lr': 6.248305368412502e-05, 'samples': 19947520, 'steps': 38959, 'loss/train': 1.7751221656799316} 02/26/2022 04:00:29 - INFO - codeparrot_training - Step 38960: {'lr': 6.24722325923851e-05, 'samples': 19948032, 'steps': 38960, 'loss/train': 1.2177268266677856} 02/26/2022 04:00:36 - INFO - codeparrot_training - Step 38961: {'lr': 6.246141230395469e-05, 'samples': 19948544, 'steps': 38961, 'loss/train': 2.254542827606201} 02/26/2022 04:00:39 - INFO - codeparrot_training - Step 38962: {'lr': 6.245059281888002e-05, 'samples': 19949056, 'steps': 38962, 'loss/train': 1.8975316286087036} 02/26/2022 04:00:45 - INFO - codeparrot_training - Step 38963: {'lr': 6.243977413720744e-05, 'samples': 19949568, 'steps': 38963, 'loss/train': 1.8330730199813843} 02/26/2022 04:00:48 - INFO - codeparrot_training - Step 38964: {'lr': 6.242895625898326e-05, 'samples': 19950080, 'steps': 38964, 'loss/train': 1.3657501935958862} 02/26/2022 04:00:54 - INFO - codeparrot_training - Step 38965: {'lr': 6.241813918425396e-05, 'samples': 19950592, 'steps': 38965, 'loss/train': 1.5496937036514282} 02/26/2022 04:00:57 - INFO - codeparrot_training - Step 38966: {'lr': 6.240732291306575e-05, 'samples': 19951104, 'steps': 38966, 'loss/train': 1.0134525299072266} 02/26/2022 04:01:03 - INFO - codeparrot_training - Step 38967: {'lr': 6.239650744546502e-05, 'samples': 19951616, 'steps': 38967, 'loss/train': 2.5222418308258057} 02/26/2022 04:01:06 - INFO - codeparrot_training - Step 38968: {'lr': 6.2385692781498e-05, 'samples': 19952128, 'steps': 38968, 'loss/train': 1.680293321609497} 02/26/2022 04:01:12 - INFO - codeparrot_training - Step 38969: {'lr': 6.237487892121118e-05, 'samples': 19952640, 'steps': 38969, 'loss/train': 0.2331373542547226} 02/26/2022 04:01:15 - INFO - codeparrot_training - Step 38970: {'lr': 6.236406586465074e-05, 'samples': 19953152, 'steps': 38970, 'loss/train': 0.7589865922927856} 02/26/2022 04:01:22 - INFO - codeparrot_training - Step 38971: {'lr': 6.235325361186316e-05, 'samples': 19953664, 'steps': 38971, 'loss/train': 1.479731798171997} 02/26/2022 04:01:27 - INFO - codeparrot_training - Step 38972: {'lr': 6.234244216289456e-05, 'samples': 19954176, 'steps': 38972, 'loss/train': 1.619925618171692} 02/26/2022 04:01:31 - INFO - codeparrot_training - Step 38973: {'lr': 6.233163151779143e-05, 'samples': 19954688, 'steps': 38973, 'loss/train': 1.6573783159255981} 02/26/2022 04:01:34 - INFO - codeparrot_training - Step 38974: {'lr': 6.232082167659989e-05, 'samples': 19955200, 'steps': 38974, 'loss/train': 1.2507054805755615} 02/26/2022 04:01:40 - INFO - codeparrot_training - Step 38975: {'lr': 6.23100126393665e-05, 'samples': 19955712, 'steps': 38975, 'loss/train': 0.04629538953304291} 02/26/2022 04:01:45 - INFO - codeparrot_training - Step 38976: {'lr': 6.229920440613729e-05, 'samples': 19956224, 'steps': 38976, 'loss/train': 0.3713911771774292} 02/26/2022 04:01:49 - INFO - codeparrot_training - Step 38977: {'lr': 6.228839697695876e-05, 'samples': 19956736, 'steps': 38977, 'loss/train': 2.4896554946899414} 02/26/2022 04:01:54 - INFO - codeparrot_training - Step 38978: {'lr': 6.227759035187706e-05, 'samples': 19957248, 'steps': 38978, 'loss/train': 2.1101582050323486} 02/26/2022 04:01:58 - INFO - codeparrot_training - Step 38979: {'lr': 6.226678453093867e-05, 'samples': 19957760, 'steps': 38979, 'loss/train': 2.0928194522857666} 02/26/2022 04:02:03 - INFO - codeparrot_training - Step 38980: {'lr': 6.225597951418965e-05, 'samples': 19958272, 'steps': 38980, 'loss/train': 2.3729329109191895} 02/26/2022 04:02:07 - INFO - codeparrot_training - Step 38981: {'lr': 6.224517530167645e-05, 'samples': 19958784, 'steps': 38981, 'loss/train': 1.4953956604003906} 02/26/2022 04:02:13 - INFO - codeparrot_training - Step 38982: {'lr': 6.223437189344525e-05, 'samples': 19959296, 'steps': 38982, 'loss/train': 1.1345945596694946} 02/26/2022 04:02:16 - INFO - codeparrot_training - Step 38983: {'lr': 6.222356928954243e-05, 'samples': 19959808, 'steps': 38983, 'loss/train': 2.1563832759857178} 02/26/2022 04:02:22 - INFO - codeparrot_training - Step 38984: {'lr': 6.221276749001423e-05, 'samples': 19960320, 'steps': 38984, 'loss/train': 1.7778867483139038} 02/26/2022 04:02:25 - INFO - codeparrot_training - Step 38985: {'lr': 6.220196649490689e-05, 'samples': 19960832, 'steps': 38985, 'loss/train': 1.9760615825653076} 02/26/2022 04:02:31 - INFO - codeparrot_training - Step 38986: {'lr': 6.219116630426664e-05, 'samples': 19961344, 'steps': 38986, 'loss/train': 2.253070831298828} 02/26/2022 04:02:34 - INFO - codeparrot_training - Step 38987: {'lr': 6.218036691813986e-05, 'samples': 19961856, 'steps': 38987, 'loss/train': 1.7850995063781738} 02/26/2022 04:02:40 - INFO - codeparrot_training - Step 38988: {'lr': 6.216956833657276e-05, 'samples': 19962368, 'steps': 38988, 'loss/train': 2.5457990169525146} 02/26/2022 04:02:43 - INFO - codeparrot_training - Step 38989: {'lr': 6.215877055961156e-05, 'samples': 19962880, 'steps': 38989, 'loss/train': 0.038928911089897156} 02/26/2022 04:02:49 - INFO - codeparrot_training - Step 38990: {'lr': 6.214797358730248e-05, 'samples': 19963392, 'steps': 38990, 'loss/train': 2.327338457107544} 02/26/2022 04:02:52 - INFO - codeparrot_training - Step 38991: {'lr': 6.213717741969191e-05, 'samples': 19963904, 'steps': 38991, 'loss/train': 1.9290884733200073} 02/26/2022 04:02:58 - INFO - codeparrot_training - Step 38992: {'lr': 6.2126382056826e-05, 'samples': 19964416, 'steps': 38992, 'loss/train': 0.7063797116279602} 02/26/2022 04:03:02 - INFO - codeparrot_training - Step 38993: {'lr': 6.211558749875105e-05, 'samples': 19964928, 'steps': 38993, 'loss/train': 1.392071008682251} 02/26/2022 04:03:07 - INFO - codeparrot_training - Step 38994: {'lr': 6.210479374551314e-05, 'samples': 19965440, 'steps': 38994, 'loss/train': 1.7028536796569824} 02/26/2022 04:03:11 - INFO - codeparrot_training - Step 38995: {'lr': 6.209400079715874e-05, 'samples': 19965952, 'steps': 38995, 'loss/train': 1.382046103477478} 02/26/2022 04:03:16 - INFO - codeparrot_training - Step 38996: {'lr': 6.208320865373396e-05, 'samples': 19966464, 'steps': 38996, 'loss/train': 1.1038291454315186} 02/26/2022 04:03:20 - INFO - codeparrot_training - Step 38997: {'lr': 6.207241731528507e-05, 'samples': 19966976, 'steps': 38997, 'loss/train': 1.7524518966674805} 02/26/2022 04:03:25 - INFO - codeparrot_training - Step 38998: {'lr': 6.206162678185817e-05, 'samples': 19967488, 'steps': 38998, 'loss/train': 1.4853224754333496} 02/26/2022 04:03:29 - INFO - codeparrot_training - Step 38999: {'lr': 6.205083705349965e-05, 'samples': 19968000, 'steps': 38999, 'loss/train': 1.7108924388885498} 02/26/2022 04:03:29 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 04:03:47 - WARNING - huggingface_hub.repository - Several commits (39) will be pushed upstream. 02/26/2022 04:03:47 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 04:04:24 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy e53c3e1..d2106d5 floral-grass-11 -> floral-grass-11 02/26/2022 04:04:30 - INFO - codeparrot_training - Step 39000: {'lr': 6.204004813025568e-05, 'samples': 19968512, 'steps': 39000, 'loss/train': 1.6747545003890991} 02/26/2022 04:04:34 - INFO - codeparrot_training - Step 39001: {'lr': 6.202926001217246e-05, 'samples': 19969024, 'steps': 39001, 'loss/train': 2.013767957687378} 02/26/2022 04:04:40 - INFO - codeparrot_training - Step 39002: {'lr': 6.201847269929612e-05, 'samples': 19969536, 'steps': 39002, 'loss/train': 1.7280240058898926} 02/26/2022 04:04:43 - INFO - codeparrot_training - Step 39003: {'lr': 6.200768619167301e-05, 'samples': 19970048, 'steps': 39003, 'loss/train': 1.1755962371826172} 02/26/2022 04:04:48 - INFO - codeparrot_training - Step 39004: {'lr': 6.199690048934933e-05, 'samples': 19970560, 'steps': 39004, 'loss/train': 2.298581123352051} 02/26/2022 04:04:52 - INFO - codeparrot_training - Step 39005: {'lr': 6.198611559237111e-05, 'samples': 19971072, 'steps': 39005, 'loss/train': 2.570070266723633} 02/26/2022 04:04:57 - INFO - codeparrot_training - Step 39006: {'lr': 6.197533150078483e-05, 'samples': 19971584, 'steps': 39006, 'loss/train': 0.574073851108551} 02/26/2022 04:05:01 - INFO - codeparrot_training - Step 39007: {'lr': 6.19645482146364e-05, 'samples': 19972096, 'steps': 39007, 'loss/train': 1.7723450660705566} 02/26/2022 04:05:08 - INFO - codeparrot_training - Step 39008: {'lr': 6.195376573397218e-05, 'samples': 19972608, 'steps': 39008, 'loss/train': 2.274547576904297} 02/26/2022 04:05:11 - INFO - codeparrot_training - Step 39009: {'lr': 6.194298405883825e-05, 'samples': 19973120, 'steps': 39009, 'loss/train': 1.4609389305114746} 02/26/2022 04:05:17 - INFO - codeparrot_training - Step 39010: {'lr': 6.193220318928102e-05, 'samples': 19973632, 'steps': 39010, 'loss/train': 2.611638069152832} 02/26/2022 04:05:20 - INFO - codeparrot_training - Step 39011: {'lr': 6.192142312534634e-05, 'samples': 19974144, 'steps': 39011, 'loss/train': 0.7639550566673279} 02/26/2022 04:05:26 - INFO - codeparrot_training - Step 39012: {'lr': 6.191064386708067e-05, 'samples': 19974656, 'steps': 39012, 'loss/train': 2.3598923683166504} 02/26/2022 04:05:29 - INFO - codeparrot_training - Step 39013: {'lr': 6.189986541453e-05, 'samples': 19975168, 'steps': 39013, 'loss/train': 2.697425603866577} 02/26/2022 04:05:35 - INFO - codeparrot_training - Step 39014: {'lr': 6.188908776774068e-05, 'samples': 19975680, 'steps': 39014, 'loss/train': 1.5055482387542725} 02/26/2022 04:05:38 - INFO - codeparrot_training - Step 39015: {'lr': 6.187831092675868e-05, 'samples': 19976192, 'steps': 39015, 'loss/train': 2.531839609146118} 02/26/2022 04:05:44 - INFO - codeparrot_training - Step 39016: {'lr': 6.186753489163033e-05, 'samples': 19976704, 'steps': 39016, 'loss/train': 1.9931358098983765} 02/26/2022 04:05:47 - INFO - codeparrot_training - Step 39017: {'lr': 6.185675966240161e-05, 'samples': 19977216, 'steps': 39017, 'loss/train': 1.963324785232544} 02/26/2022 04:05:54 - INFO - codeparrot_training - Step 39018: {'lr': 6.184598523911889e-05, 'samples': 19977728, 'steps': 39018, 'loss/train': 1.312903642654419} 02/26/2022 04:05:57 - INFO - codeparrot_training - Step 39019: {'lr': 6.183521162182824e-05, 'samples': 19978240, 'steps': 39019, 'loss/train': 2.381755828857422} 02/26/2022 04:06:03 - INFO - codeparrot_training - Step 39020: {'lr': 6.182443881057576e-05, 'samples': 19978752, 'steps': 39020, 'loss/train': 2.02602219581604} 02/26/2022 04:06:06 - INFO - codeparrot_training - Step 39021: {'lr': 6.181366680540756e-05, 'samples': 19979264, 'steps': 39021, 'loss/train': 1.2970889806747437} 02/26/2022 04:06:12 - INFO - codeparrot_training - Step 39022: {'lr': 6.180289560636995e-05, 'samples': 19979776, 'steps': 39022, 'loss/train': 3.331634521484375} 02/26/2022 04:06:15 - INFO - codeparrot_training - Step 39023: {'lr': 6.179212521350896e-05, 'samples': 19980288, 'steps': 39023, 'loss/train': 1.0699663162231445} 02/26/2022 04:06:21 - INFO - codeparrot_training - Step 39024: {'lr': 6.178135562687076e-05, 'samples': 19980800, 'steps': 39024, 'loss/train': 2.4809956550598145} 02/26/2022 04:06:24 - INFO - codeparrot_training - Step 39025: {'lr': 6.177058684650139e-05, 'samples': 19981312, 'steps': 39025, 'loss/train': 1.640254259109497} 02/26/2022 04:06:30 - INFO - codeparrot_training - Step 39026: {'lr': 6.175981887244711e-05, 'samples': 19981824, 'steps': 39026, 'loss/train': 0.6718084812164307} 02/26/2022 04:06:33 - INFO - codeparrot_training - Step 39027: {'lr': 6.174905170475401e-05, 'samples': 19982336, 'steps': 39027, 'loss/train': 1.9465161561965942} 02/26/2022 04:06:40 - INFO - codeparrot_training - Step 39028: {'lr': 6.17382853434682e-05, 'samples': 19982848, 'steps': 39028, 'loss/train': 1.8279520273208618} 02/26/2022 04:06:44 - INFO - codeparrot_training - Step 39029: {'lr': 6.172751978863572e-05, 'samples': 19983360, 'steps': 39029, 'loss/train': 2.2108936309814453} 02/26/2022 04:06:49 - INFO - codeparrot_training - Step 39030: {'lr': 6.171675504030283e-05, 'samples': 19983872, 'steps': 39030, 'loss/train': 1.1732641458511353} 02/26/2022 04:06:53 - INFO - codeparrot_training - Step 39031: {'lr': 6.170599109851549e-05, 'samples': 19984384, 'steps': 39031, 'loss/train': 1.1962976455688477} 02/26/2022 04:06:58 - INFO - codeparrot_training - Step 39032: {'lr': 6.169522796332005e-05, 'samples': 19984896, 'steps': 39032, 'loss/train': 1.1763854026794434} 02/26/2022 04:07:01 - INFO - codeparrot_training - Step 39033: {'lr': 6.168446563476232e-05, 'samples': 19985408, 'steps': 39033, 'loss/train': 1.3611056804656982} 02/26/2022 04:07:07 - INFO - codeparrot_training - Step 39034: {'lr': 6.167370411288862e-05, 'samples': 19985920, 'steps': 39034, 'loss/train': 1.4905256032943726} 02/26/2022 04:07:10 - INFO - codeparrot_training - Step 39035: {'lr': 6.166294339774489e-05, 'samples': 19986432, 'steps': 39035, 'loss/train': 1.3192150592803955} 02/26/2022 04:07:16 - INFO - codeparrot_training - Step 39036: {'lr': 6.165218348937746e-05, 'samples': 19986944, 'steps': 39036, 'loss/train': 2.604614734649658} 02/26/2022 04:07:19 - INFO - codeparrot_training - Step 39037: {'lr': 6.16414243878321e-05, 'samples': 19987456, 'steps': 39037, 'loss/train': 2.1776161193847656} 02/26/2022 04:07:26 - INFO - codeparrot_training - Step 39038: {'lr': 6.163066609315515e-05, 'samples': 19987968, 'steps': 39038, 'loss/train': 2.0572500228881836} 02/26/2022 04:07:29 - INFO - codeparrot_training - Step 39039: {'lr': 6.161990860539263e-05, 'samples': 19988480, 'steps': 39039, 'loss/train': 2.081935405731201} 02/26/2022 04:07:35 - INFO - codeparrot_training - Step 39040: {'lr': 6.160915192459057e-05, 'samples': 19988992, 'steps': 39040, 'loss/train': 1.36495041847229} 02/26/2022 04:07:38 - INFO - codeparrot_training - Step 39041: {'lr': 6.159839605079504e-05, 'samples': 19989504, 'steps': 39041, 'loss/train': 1.8949660062789917} 02/26/2022 04:07:44 - INFO - codeparrot_training - Step 39042: {'lr': 6.158764098405223e-05, 'samples': 19990016, 'steps': 39042, 'loss/train': 1.2063840627670288} 02/26/2022 04:07:47 - INFO - codeparrot_training - Step 39043: {'lr': 6.157688672440811e-05, 'samples': 19990528, 'steps': 39043, 'loss/train': 1.7869585752487183} 02/26/2022 04:07:53 - INFO - codeparrot_training - Step 39044: {'lr': 6.156613327190874e-05, 'samples': 19991040, 'steps': 39044, 'loss/train': 1.2570009231567383} 02/26/2022 04:07:58 - INFO - codeparrot_training - Step 39045: {'lr': 6.155538062660027e-05, 'samples': 19991552, 'steps': 39045, 'loss/train': 1.178591012954712} 02/26/2022 04:08:02 - INFO - codeparrot_training - Step 39046: {'lr': 6.154462878852874e-05, 'samples': 19992064, 'steps': 39046, 'loss/train': 1.2912369966506958} 02/26/2022 04:08:07 - INFO - codeparrot_training - Step 39047: {'lr': 6.153387775774013e-05, 'samples': 19992576, 'steps': 39047, 'loss/train': 0.7543942332267761} 02/26/2022 04:08:10 - INFO - codeparrot_training - Step 39048: {'lr': 6.152312753428049e-05, 'samples': 19993088, 'steps': 39048, 'loss/train': 1.7838596105575562} 02/26/2022 04:08:16 - INFO - codeparrot_training - Step 39049: {'lr': 6.151237811819602e-05, 'samples': 19993600, 'steps': 39049, 'loss/train': 1.394836664199829} 02/26/2022 04:08:20 - INFO - codeparrot_training - Step 39050: {'lr': 6.150162950953264e-05, 'samples': 19994112, 'steps': 39050, 'loss/train': 1.4068769216537476} 02/26/2022 04:08:25 - INFO - codeparrot_training - Step 39051: {'lr': 6.149088170833644e-05, 'samples': 19994624, 'steps': 39051, 'loss/train': 1.8832015991210938} 02/26/2022 04:08:28 - INFO - codeparrot_training - Step 39052: {'lr': 6.148013471465339e-05, 'samples': 19995136, 'steps': 39052, 'loss/train': 1.757644534111023} 02/26/2022 04:08:35 - INFO - codeparrot_training - Step 39053: {'lr': 6.146938852852962e-05, 'samples': 19995648, 'steps': 39053, 'loss/train': 0.8891740441322327} 02/26/2022 04:08:38 - INFO - codeparrot_training - Step 39054: {'lr': 6.145864315001115e-05, 'samples': 19996160, 'steps': 39054, 'loss/train': 0.5175991058349609} 02/26/2022 04:08:44 - INFO - codeparrot_training - Step 39055: {'lr': 6.144789857914395e-05, 'samples': 19996672, 'steps': 39055, 'loss/train': 2.0041868686676025} 02/26/2022 04:08:47 - INFO - codeparrot_training - Step 39056: {'lr': 6.143715481597403e-05, 'samples': 19997184, 'steps': 39056, 'loss/train': 0.5072973370552063} 02/26/2022 04:08:53 - INFO - codeparrot_training - Step 39057: {'lr': 6.142641186054754e-05, 'samples': 19997696, 'steps': 39057, 'loss/train': 2.1547584533691406} 02/26/2022 04:08:56 - INFO - codeparrot_training - Step 39058: {'lr': 6.141566971291043e-05, 'samples': 19998208, 'steps': 39058, 'loss/train': 0.5618374943733215} 02/26/2022 04:09:02 - INFO - codeparrot_training - Step 39059: {'lr': 6.140492837310868e-05, 'samples': 19998720, 'steps': 39059, 'loss/train': 1.1311376094818115} 02/26/2022 04:09:05 - INFO - codeparrot_training - Step 39060: {'lr': 6.139418784118827e-05, 'samples': 19999232, 'steps': 39060, 'loss/train': 2.414172887802124} 02/26/2022 04:09:11 - INFO - codeparrot_training - Step 39061: {'lr': 6.138344811719537e-05, 'samples': 19999744, 'steps': 39061, 'loss/train': 1.0803378820419312} 02/26/2022 04:09:14 - INFO - codeparrot_training - Step 39062: {'lr': 6.137270920117585e-05, 'samples': 20000256, 'steps': 39062, 'loss/train': 1.6616593599319458} 02/26/2022 04:09:20 - INFO - codeparrot_training - Step 39063: {'lr': 6.136197109317576e-05, 'samples': 20000768, 'steps': 39063, 'loss/train': 2.315521001815796} 02/26/2022 04:09:23 - INFO - codeparrot_training - Step 39064: {'lr': 6.135123379324104e-05, 'samples': 20001280, 'steps': 39064, 'loss/train': 1.2328269481658936} 02/26/2022 04:09:29 - INFO - codeparrot_training - Step 39065: {'lr': 6.134049730141777e-05, 'samples': 20001792, 'steps': 39065, 'loss/train': 2.0177531242370605} 02/26/2022 04:09:33 - INFO - codeparrot_training - Step 39066: {'lr': 6.132976161775184e-05, 'samples': 20002304, 'steps': 39066, 'loss/train': 2.2230148315429688} 02/26/2022 04:09:38 - INFO - codeparrot_training - Step 39067: {'lr': 6.131902674228945e-05, 'samples': 20002816, 'steps': 39067, 'loss/train': 2.1659538745880127} 02/26/2022 04:09:42 - INFO - codeparrot_training - Step 39068: {'lr': 6.130829267507629e-05, 'samples': 20003328, 'steps': 39068, 'loss/train': 1.5368413925170898} 02/26/2022 04:09:47 - INFO - codeparrot_training - Step 39069: {'lr': 6.129755941615856e-05, 'samples': 20003840, 'steps': 39069, 'loss/train': 1.3174980878829956} 02/26/2022 04:09:51 - INFO - codeparrot_training - Step 39070: {'lr': 6.128682696558208e-05, 'samples': 20004352, 'steps': 39070, 'loss/train': 3.0974884033203125} 02/26/2022 04:09:56 - INFO - codeparrot_training - Step 39071: {'lr': 6.127609532339306e-05, 'samples': 20004864, 'steps': 39071, 'loss/train': 1.7564198970794678} 02/26/2022 04:10:00 - INFO - codeparrot_training - Step 39072: {'lr': 6.126536448963718e-05, 'samples': 20005376, 'steps': 39072, 'loss/train': 1.2487951517105103} 02/26/2022 04:10:05 - INFO - codeparrot_training - Step 39073: {'lr': 6.125463446436064e-05, 'samples': 20005888, 'steps': 39073, 'loss/train': 2.3441574573516846} 02/26/2022 04:10:09 - INFO - codeparrot_training - Step 39074: {'lr': 6.124390524760922e-05, 'samples': 20006400, 'steps': 39074, 'loss/train': 0.940067708492279} 02/26/2022 04:10:15 - INFO - codeparrot_training - Step 39075: {'lr': 6.123317683942914e-05, 'samples': 20006912, 'steps': 39075, 'loss/train': 0.6586630940437317} 02/26/2022 04:10:19 - INFO - codeparrot_training - Step 39076: {'lr': 6.1222449239866e-05, 'samples': 20007424, 'steps': 39076, 'loss/train': 2.0088090896606445} 02/26/2022 04:10:24 - INFO - codeparrot_training - Step 39077: {'lr': 6.121172244896606e-05, 'samples': 20007936, 'steps': 39077, 'loss/train': 1.6931359767913818} 02/26/2022 04:10:27 - INFO - codeparrot_training - Step 39078: {'lr': 6.120099646677507e-05, 'samples': 20008448, 'steps': 39078, 'loss/train': 0.8467195630073547} 02/26/2022 04:10:33 - INFO - codeparrot_training - Step 39079: {'lr': 6.119027129333913e-05, 'samples': 20008960, 'steps': 39079, 'loss/train': 0.9533653259277344} 02/26/2022 04:10:37 - INFO - codeparrot_training - Step 39080: {'lr': 6.117954692870411e-05, 'samples': 20009472, 'steps': 39080, 'loss/train': 2.099364757537842} 02/26/2022 04:10:42 - INFO - codeparrot_training - Step 39081: {'lr': 6.116882337291596e-05, 'samples': 20009984, 'steps': 39081, 'loss/train': 0.2943916320800781} 02/26/2022 04:10:46 - INFO - codeparrot_training - Step 39082: {'lr': 6.115810062602054e-05, 'samples': 20010496, 'steps': 39082, 'loss/train': 1.418568730354309} 02/26/2022 04:10:51 - INFO - codeparrot_training - Step 39083: {'lr': 6.114737868806391e-05, 'samples': 20011008, 'steps': 39083, 'loss/train': 1.548271656036377} 02/26/2022 04:10:55 - INFO - codeparrot_training - Step 39084: {'lr': 6.113665755909195e-05, 'samples': 20011520, 'steps': 39084, 'loss/train': 1.1802722215652466} 02/26/2022 04:11:01 - INFO - codeparrot_training - Step 39085: {'lr': 6.112593723915058e-05, 'samples': 20012032, 'steps': 39085, 'loss/train': 1.373261570930481} 02/26/2022 04:11:04 - INFO - codeparrot_training - Step 39086: {'lr': 6.11152177282857e-05, 'samples': 20012544, 'steps': 39086, 'loss/train': 2.176295518875122} 02/26/2022 04:11:10 - INFO - codeparrot_training - Step 39087: {'lr': 6.110449902654322e-05, 'samples': 20013056, 'steps': 39087, 'loss/train': 0.8322294354438782} 02/26/2022 04:11:13 - INFO - codeparrot_training - Step 39088: {'lr': 6.109378113396913e-05, 'samples': 20013568, 'steps': 39088, 'loss/train': 1.8421934843063354} 02/26/2022 04:11:19 - INFO - codeparrot_training - Step 39089: {'lr': 6.108306405060931e-05, 'samples': 20014080, 'steps': 39089, 'loss/train': 1.0519932508468628} 02/26/2022 04:11:22 - INFO - codeparrot_training - Step 39090: {'lr': 6.107234777650963e-05, 'samples': 20014592, 'steps': 39090, 'loss/train': 1.990525245666504} 02/26/2022 04:11:28 - INFO - codeparrot_training - Step 39091: {'lr': 6.1061632311716e-05, 'samples': 20015104, 'steps': 39091, 'loss/train': 1.5158153772354126} 02/26/2022 04:11:31 - INFO - codeparrot_training - Step 39092: {'lr': 6.105091765627435e-05, 'samples': 20015616, 'steps': 39092, 'loss/train': 1.4494901895523071} 02/26/2022 04:11:37 - INFO - codeparrot_training - Step 39093: {'lr': 6.104020381023063e-05, 'samples': 20016128, 'steps': 39093, 'loss/train': 2.059077024459839} 02/26/2022 04:11:40 - INFO - codeparrot_training - Step 39094: {'lr': 6.1029490773630626e-05, 'samples': 20016640, 'steps': 39094, 'loss/train': 2.198873519897461} 02/26/2022 04:11:46 - INFO - codeparrot_training - Step 39095: {'lr': 6.1018778546520246e-05, 'samples': 20017152, 'steps': 39095, 'loss/train': 5.161421775817871} 02/26/2022 04:11:49 - INFO - codeparrot_training - Step 39096: {'lr': 6.100806712894544e-05, 'samples': 20017664, 'steps': 39096, 'loss/train': 1.4846274852752686} 02/26/2022 04:11:55 - INFO - codeparrot_training - Step 39097: {'lr': 6.099735652095209e-05, 'samples': 20018176, 'steps': 39097, 'loss/train': 0.45562854409217834} 02/26/2022 04:11:58 - INFO - codeparrot_training - Step 39098: {'lr': 6.0986646722586056e-05, 'samples': 20018688, 'steps': 39098, 'loss/train': 1.5801427364349365} 02/26/2022 04:12:04 - INFO - codeparrot_training - Step 39099: {'lr': 6.0975937733893125e-05, 'samples': 20019200, 'steps': 39099, 'loss/train': 1.6328351497650146} 02/26/2022 04:12:08 - INFO - codeparrot_training - Step 39100: {'lr': 6.0965229554919313e-05, 'samples': 20019712, 'steps': 39100, 'loss/train': 2.513152837753296} 02/26/2022 04:12:14 - INFO - codeparrot_training - Step 39101: {'lr': 6.09545221857104e-05, 'samples': 20020224, 'steps': 39101, 'loss/train': 2.6437675952911377} 02/26/2022 04:12:17 - INFO - codeparrot_training - Step 39102: {'lr': 6.09438156263124e-05, 'samples': 20020736, 'steps': 39102, 'loss/train': 3.066319704055786} 02/26/2022 04:12:22 - INFO - codeparrot_training - Step 39103: {'lr': 6.093310987677092e-05, 'samples': 20021248, 'steps': 39103, 'loss/train': 1.15779447555542} 02/26/2022 04:12:26 - INFO - codeparrot_training - Step 39104: {'lr': 6.0922404937132054e-05, 'samples': 20021760, 'steps': 39104, 'loss/train': 2.651045560836792} 02/26/2022 04:12:31 - INFO - codeparrot_training - Step 39105: {'lr': 6.0911700807441463e-05, 'samples': 20022272, 'steps': 39105, 'loss/train': 1.0412274599075317} 02/26/2022 04:12:37 - INFO - codeparrot_training - Step 39106: {'lr': 6.090099748774527e-05, 'samples': 20022784, 'steps': 39106, 'loss/train': 2.4487788677215576} 02/26/2022 04:12:40 - INFO - codeparrot_training - Step 39107: {'lr': 6.0890294978089014e-05, 'samples': 20023296, 'steps': 39107, 'loss/train': 1.915837049484253} 02/26/2022 04:12:46 - INFO - codeparrot_training - Step 39108: {'lr': 6.0879593278518744e-05, 'samples': 20023808, 'steps': 39108, 'loss/train': 2.437455654144287} 02/26/2022 04:12:49 - INFO - codeparrot_training - Step 39109: {'lr': 6.086889238908017e-05, 'samples': 20024320, 'steps': 39109, 'loss/train': 1.421085238456726} 02/26/2022 04:12:56 - INFO - codeparrot_training - Step 39110: {'lr': 6.085819230981934e-05, 'samples': 20024832, 'steps': 39110, 'loss/train': 1.716515302658081} 02/26/2022 04:12:59 - INFO - codeparrot_training - Step 39111: {'lr': 6.0847493040781855e-05, 'samples': 20025344, 'steps': 39111, 'loss/train': 2.38794207572937} 02/26/2022 04:13:05 - INFO - codeparrot_training - Step 39112: {'lr': 6.083679458201369e-05, 'samples': 20025856, 'steps': 39112, 'loss/train': 1.523942232131958} 02/26/2022 04:13:08 - INFO - codeparrot_training - Step 39113: {'lr': 6.082609693356056e-05, 'samples': 20026368, 'steps': 39113, 'loss/train': 0.8572320938110352} 02/26/2022 04:13:12 - INFO - codeparrot_training - Step 39114: {'lr': 6.0815400095468433e-05, 'samples': 20026880, 'steps': 39114, 'loss/train': 1.425020456314087} 02/26/2022 04:13:17 - INFO - codeparrot_training - Step 39115: {'lr': 6.080470406778305e-05, 'samples': 20027392, 'steps': 39115, 'loss/train': 1.7392770051956177} 02/26/2022 04:13:23 - INFO - codeparrot_training - Step 39116: {'lr': 6.079400885055025e-05, 'samples': 20027904, 'steps': 39116, 'loss/train': 1.9410251379013062} 02/26/2022 04:13:26 - INFO - codeparrot_training - Step 39117: {'lr': 6.078331444381577e-05, 'samples': 20028416, 'steps': 39117, 'loss/train': 0.9213608503341675} 02/26/2022 04:13:32 - INFO - codeparrot_training - Step 39118: {'lr': 6.0772620847625555e-05, 'samples': 20028928, 'steps': 39118, 'loss/train': 2.0220468044281006} 02/26/2022 04:13:35 - INFO - codeparrot_training - Step 39119: {'lr': 6.0761928062025324e-05, 'samples': 20029440, 'steps': 39119, 'loss/train': 1.5902371406555176} 02/26/2022 04:13:42 - INFO - codeparrot_training - Step 39120: {'lr': 6.075123608706093e-05, 'samples': 20029952, 'steps': 39120, 'loss/train': 2.356743097305298} 02/26/2022 04:13:45 - INFO - codeparrot_training - Step 39121: {'lr': 6.0740544922778055e-05, 'samples': 20030464, 'steps': 39121, 'loss/train': 1.110445499420166} 02/26/2022 04:13:48 - INFO - codeparrot_training - Step 39122: {'lr': 6.07298545692227e-05, 'samples': 20030976, 'steps': 39122, 'loss/train': 1.3796019554138184} 02/26/2022 04:13:54 - INFO - codeparrot_training - Step 39123: {'lr': 6.071916502644051e-05, 'samples': 20031488, 'steps': 39123, 'loss/train': 1.4902094602584839} 02/26/2022 04:13:57 - INFO - codeparrot_training - Step 39124: {'lr': 6.070847629447732e-05, 'samples': 20032000, 'steps': 39124, 'loss/train': 2.1751444339752197} 02/26/2022 04:14:03 - INFO - codeparrot_training - Step 39125: {'lr': 6.069778837337886e-05, 'samples': 20032512, 'steps': 39125, 'loss/train': 1.48564875125885} 02/26/2022 04:14:09 - INFO - codeparrot_training - Step 39126: {'lr': 6.068710126319102e-05, 'samples': 20033024, 'steps': 39126, 'loss/train': 1.5054903030395508} 02/26/2022 04:14:12 - INFO - codeparrot_training - Step 39127: {'lr': 6.067641496395951e-05, 'samples': 20033536, 'steps': 39127, 'loss/train': 1.8296412229537964} 02/26/2022 04:14:18 - INFO - codeparrot_training - Step 39128: {'lr': 6.066572947573015e-05, 'samples': 20034048, 'steps': 39128, 'loss/train': 1.9717140197753906} 02/26/2022 04:14:21 - INFO - codeparrot_training - Step 39129: {'lr': 6.065504479854858e-05, 'samples': 20034560, 'steps': 39129, 'loss/train': 2.417285680770874} 02/26/2022 04:14:27 - INFO - codeparrot_training - Step 39130: {'lr': 6.064436093246078e-05, 'samples': 20035072, 'steps': 39130, 'loss/train': 2.8860702514648438} 02/26/2022 04:14:31 - INFO - codeparrot_training - Step 39131: {'lr': 6.06336778775124e-05, 'samples': 20035584, 'steps': 39131, 'loss/train': 1.8716504573822021} 02/26/2022 04:14:36 - INFO - codeparrot_training - Step 39132: {'lr': 6.062299563374921e-05, 'samples': 20036096, 'steps': 39132, 'loss/train': 2.9836840629577637} 02/26/2022 04:14:40 - INFO - codeparrot_training - Step 39133: {'lr': 6.061231420121688e-05, 'samples': 20036608, 'steps': 39133, 'loss/train': 1.7031925916671753} 02/26/2022 04:14:45 - INFO - codeparrot_training - Step 39134: {'lr': 6.0601633579961375e-05, 'samples': 20037120, 'steps': 39134, 'loss/train': 2.1716651916503906} 02/26/2022 04:14:49 - INFO - codeparrot_training - Step 39135: {'lr': 6.0590953770028296e-05, 'samples': 20037632, 'steps': 39135, 'loss/train': 2.3985800743103027} 02/26/2022 04:14:54 - INFO - codeparrot_training - Step 39136: {'lr': 6.058027477146344e-05, 'samples': 20038144, 'steps': 39136, 'loss/train': 2.794325351715088} 02/26/2022 04:14:58 - INFO - codeparrot_training - Step 39137: {'lr': 6.056959658431252e-05, 'samples': 20038656, 'steps': 39137, 'loss/train': 1.6470996141433716} 02/26/2022 04:15:03 - INFO - codeparrot_training - Step 39138: {'lr': 6.0558919208621246e-05, 'samples': 20039168, 'steps': 39138, 'loss/train': 1.3356456756591797} 02/26/2022 04:15:07 - INFO - codeparrot_training - Step 39139: {'lr': 6.054824264443548e-05, 'samples': 20039680, 'steps': 39139, 'loss/train': 2.127650737762451} 02/26/2022 04:15:12 - INFO - codeparrot_training - Step 39140: {'lr': 6.0537566891800815e-05, 'samples': 20040192, 'steps': 39140, 'loss/train': 1.527857780456543} 02/26/2022 04:15:16 - INFO - codeparrot_training - Step 39141: {'lr': 6.052689195076319e-05, 'samples': 20040704, 'steps': 39141, 'loss/train': 1.3834233283996582} 02/26/2022 04:15:22 - INFO - codeparrot_training - Step 39142: {'lr': 6.051621782136804e-05, 'samples': 20041216, 'steps': 39142, 'loss/train': 0.9246218800544739} 02/26/2022 04:15:25 - INFO - codeparrot_training - Step 39143: {'lr': 6.050554450366133e-05, 'samples': 20041728, 'steps': 39143, 'loss/train': 2.255917549133301} 02/26/2022 04:15:31 - INFO - codeparrot_training - Step 39144: {'lr': 6.0494871997688636e-05, 'samples': 20042240, 'steps': 39144, 'loss/train': 1.8913518190383911} 02/26/2022 04:15:34 - INFO - codeparrot_training - Step 39145: {'lr': 6.0484200303495855e-05, 'samples': 20042752, 'steps': 39145, 'loss/train': 1.3118107318878174} 02/26/2022 04:15:40 - INFO - codeparrot_training - Step 39146: {'lr': 6.0473529421128466e-05, 'samples': 20043264, 'steps': 39146, 'loss/train': 0.20709088444709778} 02/26/2022 04:15:44 - INFO - codeparrot_training - Step 39147: {'lr': 6.046285935063234e-05, 'samples': 20043776, 'steps': 39147, 'loss/train': 0.9929444193840027} 02/26/2022 04:15:49 - INFO - codeparrot_training - Step 39148: {'lr': 6.045219009205308e-05, 'samples': 20044288, 'steps': 39148, 'loss/train': 1.2968947887420654} 02/26/2022 04:15:53 - INFO - codeparrot_training - Step 39149: {'lr': 6.044152164543659e-05, 'samples': 20044800, 'steps': 39149, 'loss/train': 1.6446397304534912} 02/26/2022 04:15:58 - INFO - codeparrot_training - Step 39150: {'lr': 6.043085401082829e-05, 'samples': 20045312, 'steps': 39150, 'loss/train': 1.9716802835464478} 02/26/2022 04:16:02 - INFO - codeparrot_training - Step 39151: {'lr': 6.042018718827408e-05, 'samples': 20045824, 'steps': 39151, 'loss/train': 2.2467217445373535} 02/26/2022 04:16:07 - INFO - codeparrot_training - Step 39152: {'lr': 6.040952117781953e-05, 'samples': 20046336, 'steps': 39152, 'loss/train': 1.6879045963287354} 02/26/2022 04:16:11 - INFO - codeparrot_training - Step 39153: {'lr': 6.039885597951045e-05, 'samples': 20046848, 'steps': 39153, 'loss/train': 1.7481259107589722} 02/26/2022 04:16:17 - INFO - codeparrot_training - Step 39154: {'lr': 6.038819159339248e-05, 'samples': 20047360, 'steps': 39154, 'loss/train': 1.6514978408813477} 02/26/2022 04:16:20 - INFO - codeparrot_training - Step 39155: {'lr': 6.037752801951127e-05, 'samples': 20047872, 'steps': 39155, 'loss/train': 1.8021762371063232} 02/26/2022 04:16:26 - INFO - codeparrot_training - Step 39156: {'lr': 6.036686525791246e-05, 'samples': 20048384, 'steps': 39156, 'loss/train': 2.1468286514282227} 02/26/2022 04:16:29 - INFO - codeparrot_training - Step 39157: {'lr': 6.035620330864183e-05, 'samples': 20048896, 'steps': 39157, 'loss/train': 1.4882481098175049} 02/26/2022 04:16:35 - INFO - codeparrot_training - Step 39158: {'lr': 6.034554217174504e-05, 'samples': 20049408, 'steps': 39158, 'loss/train': 1.5330560207366943} 02/26/2022 04:16:38 - INFO - codeparrot_training - Step 39159: {'lr': 6.033488184726771e-05, 'samples': 20049920, 'steps': 39159, 'loss/train': 1.8926385641098022} 02/26/2022 04:16:44 - INFO - codeparrot_training - Step 39160: {'lr': 6.0324222335255445e-05, 'samples': 20050432, 'steps': 39160, 'loss/train': 2.3933095932006836} 02/26/2022 04:16:48 - INFO - codeparrot_training - Step 39161: {'lr': 6.0313563635754074e-05, 'samples': 20050944, 'steps': 39161, 'loss/train': 1.404457926750183} 02/26/2022 04:16:53 - INFO - codeparrot_training - Step 39162: {'lr': 6.030290574880909e-05, 'samples': 20051456, 'steps': 39162, 'loss/train': 1.5727367401123047} 02/26/2022 04:16:56 - INFO - codeparrot_training - Step 39163: {'lr': 6.029224867446634e-05, 'samples': 20051968, 'steps': 39163, 'loss/train': 1.8429498672485352} 02/26/2022 04:17:02 - INFO - codeparrot_training - Step 39164: {'lr': 6.028159241277123e-05, 'samples': 20052480, 'steps': 39164, 'loss/train': 1.6968461275100708} 02/26/2022 04:17:05 - INFO - codeparrot_training - Step 39165: {'lr': 6.0270936963769616e-05, 'samples': 20052992, 'steps': 39165, 'loss/train': 2.4597506523132324} 02/26/2022 04:17:12 - INFO - codeparrot_training - Step 39166: {'lr': 6.026028232750696e-05, 'samples': 20053504, 'steps': 39166, 'loss/train': 2.032080888748169} 02/26/2022 04:17:15 - INFO - codeparrot_training - Step 39167: {'lr': 6.024962850402918e-05, 'samples': 20054016, 'steps': 39167, 'loss/train': 1.5263077020645142} 02/26/2022 04:17:21 - INFO - codeparrot_training - Step 39168: {'lr': 6.023897549338159e-05, 'samples': 20054528, 'steps': 39168, 'loss/train': 1.8026885986328125} 02/26/2022 04:17:24 - INFO - codeparrot_training - Step 39169: {'lr': 6.022832329561004e-05, 'samples': 20055040, 'steps': 39169, 'loss/train': 1.3546422719955444} 02/26/2022 04:17:30 - INFO - codeparrot_training - Step 39170: {'lr': 6.0217671910760106e-05, 'samples': 20055552, 'steps': 39170, 'loss/train': 1.0742846727371216} 02/26/2022 04:17:34 - INFO - codeparrot_training - Step 39171: {'lr': 6.0207021338877396e-05, 'samples': 20056064, 'steps': 39171, 'loss/train': 1.150915503501892} 02/26/2022 04:17:39 - INFO - codeparrot_training - Step 39172: {'lr': 6.0196371580007456e-05, 'samples': 20056576, 'steps': 39172, 'loss/train': 0.050631504505872726} 02/26/2022 04:17:43 - INFO - codeparrot_training - Step 39173: {'lr': 6.018572263419608e-05, 'samples': 20057088, 'steps': 39173, 'loss/train': 1.9853448867797852} 02/26/2022 04:17:48 - INFO - codeparrot_training - Step 39174: {'lr': 6.0175074501488765e-05, 'samples': 20057600, 'steps': 39174, 'loss/train': 1.737609624862671} 02/26/2022 04:17:52 - INFO - codeparrot_training - Step 39175: {'lr': 6.01644271819311e-05, 'samples': 20058112, 'steps': 39175, 'loss/train': 1.288604736328125} 02/26/2022 04:17:57 - INFO - codeparrot_training - Step 39176: {'lr': 6.015378067556884e-05, 'samples': 20058624, 'steps': 39176, 'loss/train': 1.0448052883148193} 02/26/2022 04:18:01 - INFO - codeparrot_training - Step 39177: {'lr': 6.014313498244747e-05, 'samples': 20059136, 'steps': 39177, 'loss/train': 1.7478086948394775} 02/26/2022 04:18:07 - INFO - codeparrot_training - Step 39178: {'lr': 6.0132490102612654e-05, 'samples': 20059648, 'steps': 39178, 'loss/train': 2.2179555892944336} 02/26/2022 04:18:11 - INFO - codeparrot_training - Step 39179: {'lr': 6.012184603610987e-05, 'samples': 20060160, 'steps': 39179, 'loss/train': 0.7077801823616028} 02/26/2022 04:18:14 - INFO - codeparrot_training - Step 39180: {'lr': 6.0111202782984876e-05, 'samples': 20060672, 'steps': 39180, 'loss/train': 1.6781692504882812} 02/26/2022 04:18:21 - INFO - codeparrot_training - Step 39181: {'lr': 6.0100560343283205e-05, 'samples': 20061184, 'steps': 39181, 'loss/train': 1.0841268301010132} 02/26/2022 04:18:25 - INFO - codeparrot_training - Step 39182: {'lr': 6.008991871705041e-05, 'samples': 20061696, 'steps': 39182, 'loss/train': 1.9873597621917725} 02/26/2022 04:18:30 - INFO - codeparrot_training - Step 39183: {'lr': 6.0079277904332066e-05, 'samples': 20062208, 'steps': 39183, 'loss/train': 2.6531574726104736} 02/26/2022 04:18:34 - INFO - codeparrot_training - Step 39184: {'lr': 6.0068637905173916e-05, 'samples': 20062720, 'steps': 39184, 'loss/train': 1.067253589630127} 02/26/2022 04:18:39 - INFO - codeparrot_training - Step 39185: {'lr': 6.0057998719621256e-05, 'samples': 20063232, 'steps': 39185, 'loss/train': 1.839909553527832} 02/26/2022 04:18:43 - INFO - codeparrot_training - Step 39186: {'lr': 6.0047360347719874e-05, 'samples': 20063744, 'steps': 39186, 'loss/train': 1.4952328205108643} 02/26/2022 04:18:48 - INFO - codeparrot_training - Step 39187: {'lr': 6.0036722789515244e-05, 'samples': 20064256, 'steps': 39187, 'loss/train': 1.8996855020523071} 02/26/2022 04:18:52 - INFO - codeparrot_training - Step 39188: {'lr': 6.0026086045053025e-05, 'samples': 20064768, 'steps': 39188, 'loss/train': 1.3404773473739624} 02/26/2022 04:18:57 - INFO - codeparrot_training - Step 39189: {'lr': 6.001545011437875e-05, 'samples': 20065280, 'steps': 39189, 'loss/train': 1.9518004655838013} 02/26/2022 04:19:01 - INFO - codeparrot_training - Step 39190: {'lr': 6.000481499753793e-05, 'samples': 20065792, 'steps': 39190, 'loss/train': 1.6301372051239014} 02/26/2022 04:19:07 - INFO - codeparrot_training - Step 39191: {'lr': 5.99941806945761e-05, 'samples': 20066304, 'steps': 39191, 'loss/train': 1.8701907396316528} 02/26/2022 04:19:11 - INFO - codeparrot_training - Step 39192: {'lr': 5.998354720553895e-05, 'samples': 20066816, 'steps': 39192, 'loss/train': 1.5031712055206299} 02/26/2022 04:19:16 - INFO - codeparrot_training - Step 39193: {'lr': 5.9972914530471935e-05, 'samples': 20067328, 'steps': 39193, 'loss/train': 1.1715377569198608} 02/26/2022 04:19:20 - INFO - codeparrot_training - Step 39194: {'lr': 5.996228266942061e-05, 'samples': 20067840, 'steps': 39194, 'loss/train': 2.02754545211792} 02/26/2022 04:19:25 - INFO - codeparrot_training - Step 39195: {'lr': 5.9951651622430435e-05, 'samples': 20068352, 'steps': 39195, 'loss/train': 2.533637046813965} 02/26/2022 04:19:29 - INFO - codeparrot_training - Step 39196: {'lr': 5.994102138954713e-05, 'samples': 20068864, 'steps': 39196, 'loss/train': 1.5374023914337158} 02/26/2022 04:19:34 - INFO - codeparrot_training - Step 39197: {'lr': 5.9930391970816153e-05, 'samples': 20069376, 'steps': 39197, 'loss/train': 2.2290003299713135} 02/26/2022 04:19:38 - INFO - codeparrot_training - Step 39198: {'lr': 5.991976336628299e-05, 'samples': 20069888, 'steps': 39198, 'loss/train': 1.6066607236862183} 02/26/2022 04:19:43 - INFO - codeparrot_training - Step 39199: {'lr': 5.9909135575993146e-05, 'samples': 20070400, 'steps': 39199, 'loss/train': 1.2327711582183838} 02/26/2022 04:19:47 - INFO - codeparrot_training - Step 39200: {'lr': 5.989850859999227e-05, 'samples': 20070912, 'steps': 39200, 'loss/train': 0.9524955153465271} 02/26/2022 04:19:52 - INFO - codeparrot_training - Step 39201: {'lr': 5.988788243832577e-05, 'samples': 20071424, 'steps': 39201, 'loss/train': 1.780321717262268} 02/26/2022 04:19:56 - INFO - codeparrot_training - Step 39202: {'lr': 5.987725709103933e-05, 'samples': 20071936, 'steps': 39202, 'loss/train': 2.4739410877227783} 02/26/2022 04:20:02 - INFO - codeparrot_training - Step 39203: {'lr': 5.986663255817823e-05, 'samples': 20072448, 'steps': 39203, 'loss/train': 2.1084718704223633} 02/26/2022 04:20:08 - INFO - codeparrot_training - Step 39204: {'lr': 5.9856008839788145e-05, 'samples': 20072960, 'steps': 39204, 'loss/train': 2.607719659805298} 02/26/2022 04:20:11 - INFO - codeparrot_training - Step 39205: {'lr': 5.9845385935914486e-05, 'samples': 20073472, 'steps': 39205, 'loss/train': 1.553194284439087} 02/26/2022 04:20:17 - INFO - codeparrot_training - Step 39206: {'lr': 5.983476384660297e-05, 'samples': 20073984, 'steps': 39206, 'loss/train': 0.08067571371793747} 02/26/2022 04:20:20 - INFO - codeparrot_training - Step 39207: {'lr': 5.9824142571898774e-05, 'samples': 20074496, 'steps': 39207, 'loss/train': 0.5256961584091187} 02/26/2022 04:20:26 - INFO - codeparrot_training - Step 39208: {'lr': 5.981352211184765e-05, 'samples': 20075008, 'steps': 39208, 'loss/train': 2.068943738937378} 02/26/2022 04:20:29 - INFO - codeparrot_training - Step 39209: {'lr': 5.980290246649495e-05, 'samples': 20075520, 'steps': 39209, 'loss/train': 2.2667617797851562} 02/26/2022 04:20:33 - INFO - codeparrot_training - Step 39210: {'lr': 5.979228363588626e-05, 'samples': 20076032, 'steps': 39210, 'loss/train': 2.1234488487243652} 02/26/2022 04:20:39 - INFO - codeparrot_training - Step 39211: {'lr': 5.9781665620067035e-05, 'samples': 20076544, 'steps': 39211, 'loss/train': 1.2851656675338745} 02/26/2022 04:20:42 - INFO - codeparrot_training - Step 39212: {'lr': 5.977104841908276e-05, 'samples': 20077056, 'steps': 39212, 'loss/train': 2.129596471786499} 02/26/2022 04:20:49 - INFO - codeparrot_training - Step 39213: {'lr': 5.976043203297885e-05, 'samples': 20077568, 'steps': 39213, 'loss/train': 0.2620987892150879} 02/26/2022 04:20:52 - INFO - codeparrot_training - Step 39214: {'lr': 5.9749816461800884e-05, 'samples': 20078080, 'steps': 39214, 'loss/train': 2.2618589401245117} 02/26/2022 04:20:57 - INFO - codeparrot_training - Step 39215: {'lr': 5.9739201705594323e-05, 'samples': 20078592, 'steps': 39215, 'loss/train': 1.6342507600784302} 02/26/2022 04:21:01 - INFO - codeparrot_training - Step 39216: {'lr': 5.97285877644046e-05, 'samples': 20079104, 'steps': 39216, 'loss/train': 1.6585915088653564} 02/26/2022 04:21:07 - INFO - codeparrot_training - Step 39217: {'lr': 5.971797463827716e-05, 'samples': 20079616, 'steps': 39217, 'loss/train': 0.9379733800888062} 02/26/2022 04:21:10 - INFO - codeparrot_training - Step 39218: {'lr': 5.970736232725746e-05, 'samples': 20080128, 'steps': 39218, 'loss/train': 1.3840649127960205} 02/26/2022 04:21:16 - INFO - codeparrot_training - Step 39219: {'lr': 5.969675083139106e-05, 'samples': 20080640, 'steps': 39219, 'loss/train': 2.2475056648254395} 02/26/2022 04:21:19 - INFO - codeparrot_training - Step 39220: {'lr': 5.968614015072335e-05, 'samples': 20081152, 'steps': 39220, 'loss/train': 2.007514476776123} 02/26/2022 04:21:25 - INFO - codeparrot_training - Step 39221: {'lr': 5.967553028529976e-05, 'samples': 20081664, 'steps': 39221, 'loss/train': 1.9549225568771362} 02/26/2022 04:21:28 - INFO - codeparrot_training - Step 39222: {'lr': 5.966492123516573e-05, 'samples': 20082176, 'steps': 39222, 'loss/train': 1.4100981950759888} 02/26/2022 04:21:35 - INFO - codeparrot_training - Step 39223: {'lr': 5.965431300036678e-05, 'samples': 20082688, 'steps': 39223, 'loss/train': 2.825514316558838} 02/26/2022 04:21:38 - INFO - codeparrot_training - Step 39224: {'lr': 5.964370558094831e-05, 'samples': 20083200, 'steps': 39224, 'loss/train': 2.0286624431610107} 02/26/2022 04:21:44 - INFO - codeparrot_training - Step 39225: {'lr': 5.963309897695579e-05, 'samples': 20083712, 'steps': 39225, 'loss/train': 1.9217307567596436} 02/26/2022 04:21:49 - INFO - codeparrot_training - Step 39226: {'lr': 5.962249318843455e-05, 'samples': 20084224, 'steps': 39226, 'loss/train': 0.926033616065979} 02/26/2022 04:21:53 - INFO - codeparrot_training - Step 39227: {'lr': 5.961188821543015e-05, 'samples': 20084736, 'steps': 39227, 'loss/train': 2.020341634750366} 02/26/2022 04:21:56 - INFO - codeparrot_training - Step 39228: {'lr': 5.9601284057987954e-05, 'samples': 20085248, 'steps': 39228, 'loss/train': 1.7333258390426636} 02/26/2022 04:22:02 - INFO - codeparrot_training - Step 39229: {'lr': 5.9590680716153425e-05, 'samples': 20085760, 'steps': 39229, 'loss/train': 1.9270201921463013} 02/26/2022 04:22:07 - INFO - codeparrot_training - Step 39230: {'lr': 5.958007818997188e-05, 'samples': 20086272, 'steps': 39230, 'loss/train': 1.4925305843353271} 02/26/2022 04:22:11 - INFO - codeparrot_training - Step 39231: {'lr': 5.956947647948888e-05, 'samples': 20086784, 'steps': 39231, 'loss/train': 1.597577452659607} 02/26/2022 04:22:16 - INFO - codeparrot_training - Step 39232: {'lr': 5.9558875584749786e-05, 'samples': 20087296, 'steps': 39232, 'loss/train': 1.6095819473266602} 02/26/2022 04:22:20 - INFO - codeparrot_training - Step 39233: {'lr': 5.95482755058e-05, 'samples': 20087808, 'steps': 39233, 'loss/train': 2.490502119064331} 02/26/2022 04:22:25 - INFO - codeparrot_training - Step 39234: {'lr': 5.953767624268483e-05, 'samples': 20088320, 'steps': 39234, 'loss/train': 2.119033098220825} 02/26/2022 04:22:29 - INFO - codeparrot_training - Step 39235: {'lr': 5.952707779544986e-05, 'samples': 20088832, 'steps': 39235, 'loss/train': 1.9056191444396973} 02/26/2022 04:22:35 - INFO - codeparrot_training - Step 39236: {'lr': 5.951648016414035e-05, 'samples': 20089344, 'steps': 39236, 'loss/train': 1.6329952478408813} 02/26/2022 04:22:38 - INFO - codeparrot_training - Step 39237: {'lr': 5.950588334880189e-05, 'samples': 20089856, 'steps': 39237, 'loss/train': 0.4335310757160187} 02/26/2022 04:22:42 - INFO - codeparrot_training - Step 39238: {'lr': 5.949528734947959e-05, 'samples': 20090368, 'steps': 39238, 'loss/train': 1.3485746383666992} 02/26/2022 04:22:49 - INFO - codeparrot_training - Step 39239: {'lr': 5.948469216621904e-05, 'samples': 20090880, 'steps': 39239, 'loss/train': 1.9692628383636475} 02/26/2022 04:22:52 - INFO - codeparrot_training - Step 39240: {'lr': 5.947409779906554e-05, 'samples': 20091392, 'steps': 39240, 'loss/train': 1.220459222793579} 02/26/2022 04:22:57 - INFO - codeparrot_training - Step 39241: {'lr': 5.9463504248064625e-05, 'samples': 20091904, 'steps': 39241, 'loss/train': 1.8257173299789429} 02/26/2022 04:23:01 - INFO - codeparrot_training - Step 39242: {'lr': 5.94529115132614e-05, 'samples': 20092416, 'steps': 39242, 'loss/train': 1.7604314088821411} 02/26/2022 04:23:07 - INFO - codeparrot_training - Step 39243: {'lr': 5.9442319594701504e-05, 'samples': 20092928, 'steps': 39243, 'loss/train': 1.203141689300537} 02/26/2022 04:23:10 - INFO - codeparrot_training - Step 39244: {'lr': 5.943172849243012e-05, 'samples': 20093440, 'steps': 39244, 'loss/train': 2.061610221862793} 02/26/2022 04:23:16 - INFO - codeparrot_training - Step 39245: {'lr': 5.9421138206492806e-05, 'samples': 20093952, 'steps': 39245, 'loss/train': 1.656941294670105} 02/26/2022 04:23:19 - INFO - codeparrot_training - Step 39246: {'lr': 5.941054873693472e-05, 'samples': 20094464, 'steps': 39246, 'loss/train': 1.434607744216919} 02/26/2022 04:23:25 - INFO - codeparrot_training - Step 39247: {'lr': 5.9399960083801385e-05, 'samples': 20094976, 'steps': 39247, 'loss/train': 2.743180751800537} 02/26/2022 04:23:28 - INFO - codeparrot_training - Step 39248: {'lr': 5.9389372247138004e-05, 'samples': 20095488, 'steps': 39248, 'loss/train': 0.854601263999939} 02/26/2022 04:23:35 - INFO - codeparrot_training - Step 39249: {'lr': 5.937878522699011e-05, 'samples': 20096000, 'steps': 39249, 'loss/train': 1.7251155376434326} 02/26/2022 04:23:38 - INFO - codeparrot_training - Step 39250: {'lr': 5.936819902340299e-05, 'samples': 20096512, 'steps': 39250, 'loss/train': 1.9975084066390991} 02/26/2022 04:23:43 - INFO - codeparrot_training - Step 39251: {'lr': 5.935761363642195e-05, 'samples': 20097024, 'steps': 39251, 'loss/train': 1.8870348930358887} 02/26/2022 04:23:47 - INFO - codeparrot_training - Step 39252: {'lr': 5.934702906609229e-05, 'samples': 20097536, 'steps': 39252, 'loss/train': 2.1907081604003906} 02/26/2022 04:23:52 - INFO - codeparrot_training - Step 39253: {'lr': 5.933644531245952e-05, 'samples': 20098048, 'steps': 39253, 'loss/train': 0.7847607135772705} 02/26/2022 04:23:56 - INFO - codeparrot_training - Step 39254: {'lr': 5.9325862375568835e-05, 'samples': 20098560, 'steps': 39254, 'loss/train': 1.7179933786392212} 02/26/2022 04:24:01 - INFO - codeparrot_training - Step 39255: {'lr': 5.931528025546565e-05, 'samples': 20099072, 'steps': 39255, 'loss/train': 1.0662907361984253} 02/26/2022 04:24:05 - INFO - codeparrot_training - Step 39256: {'lr': 5.9304698952195147e-05, 'samples': 20099584, 'steps': 39256, 'loss/train': 1.7633732557296753} 02/26/2022 04:24:10 - INFO - codeparrot_training - Step 39257: {'lr': 5.9294118465802847e-05, 'samples': 20100096, 'steps': 39257, 'loss/train': 0.9880204796791077} 02/26/2022 04:24:14 - INFO - codeparrot_training - Step 39258: {'lr': 5.9283538796334e-05, 'samples': 20100608, 'steps': 39258, 'loss/train': 1.5551226139068604} 02/26/2022 04:24:21 - INFO - codeparrot_training - Step 39259: {'lr': 5.9272959943833886e-05, 'samples': 20101120, 'steps': 39259, 'loss/train': 1.6984381675720215} 02/26/2022 04:24:26 - INFO - codeparrot_training - Step 39260: {'lr': 5.926238190834779e-05, 'samples': 20101632, 'steps': 39260, 'loss/train': 1.2085760831832886} 02/26/2022 04:24:30 - INFO - codeparrot_training - Step 39261: {'lr': 5.925180468992117e-05, 'samples': 20102144, 'steps': 39261, 'loss/train': 1.8080426454544067} 02/26/2022 04:24:35 - INFO - codeparrot_training - Step 39262: {'lr': 5.924122828859921e-05, 'samples': 20102656, 'steps': 39262, 'loss/train': 3.404402494430542} 02/26/2022 04:24:39 - INFO - codeparrot_training - Step 39263: {'lr': 5.923065270442729e-05, 'samples': 20103168, 'steps': 39263, 'loss/train': 1.4390474557876587} 02/26/2022 04:24:44 - INFO - codeparrot_training - Step 39264: {'lr': 5.922007793745068e-05, 'samples': 20103680, 'steps': 39264, 'loss/train': 2.132646322250366} 02/26/2022 04:24:48 - INFO - codeparrot_training - Step 39265: {'lr': 5.9209503987714574e-05, 'samples': 20104192, 'steps': 39265, 'loss/train': 1.136685848236084} 02/26/2022 04:24:53 - INFO - codeparrot_training - Step 39266: {'lr': 5.919893085526448e-05, 'samples': 20104704, 'steps': 39266, 'loss/train': 2.5582213401794434} 02/26/2022 04:24:57 - INFO - codeparrot_training - Step 39267: {'lr': 5.918835854014557e-05, 'samples': 20105216, 'steps': 39267, 'loss/train': 1.631758451461792} 02/26/2022 04:25:02 - INFO - codeparrot_training - Step 39268: {'lr': 5.917778704240312e-05, 'samples': 20105728, 'steps': 39268, 'loss/train': 1.1914089918136597} 02/26/2022 04:25:06 - INFO - codeparrot_training - Step 39269: {'lr': 5.916721636208239e-05, 'samples': 20106240, 'steps': 39269, 'loss/train': 0.9935445189476013} 02/26/2022 04:25:11 - INFO - codeparrot_training - Step 39270: {'lr': 5.915664649922875e-05, 'samples': 20106752, 'steps': 39270, 'loss/train': 0.8829196691513062} 02/26/2022 04:25:15 - INFO - codeparrot_training - Step 39271: {'lr': 5.914607745388739e-05, 'samples': 20107264, 'steps': 39271, 'loss/train': 0.5542932748794556} 02/26/2022 04:25:22 - INFO - codeparrot_training - Step 39272: {'lr': 5.913550922610378e-05, 'samples': 20107776, 'steps': 39272, 'loss/train': 2.1349236965179443} 02/26/2022 04:25:25 - INFO - codeparrot_training - Step 39273: {'lr': 5.912494181592287e-05, 'samples': 20108288, 'steps': 39273, 'loss/train': 1.980897307395935} 02/26/2022 04:25:31 - INFO - codeparrot_training - Step 39274: {'lr': 5.911437522339019e-05, 'samples': 20108800, 'steps': 39274, 'loss/train': 1.7240952253341675} 02/26/2022 04:25:34 - INFO - codeparrot_training - Step 39275: {'lr': 5.910380944855087e-05, 'samples': 20109312, 'steps': 39275, 'loss/train': 1.6117587089538574} 02/26/2022 04:25:40 - INFO - codeparrot_training - Step 39276: {'lr': 5.909324449145032e-05, 'samples': 20109824, 'steps': 39276, 'loss/train': 2.7535886764526367} 02/26/2022 04:25:43 - INFO - codeparrot_training - Step 39277: {'lr': 5.9082680352133546e-05, 'samples': 20110336, 'steps': 39277, 'loss/train': 2.589470863342285} 02/26/2022 04:25:49 - INFO - codeparrot_training - Step 39278: {'lr': 5.907211703064605e-05, 'samples': 20110848, 'steps': 39278, 'loss/train': 0.8415930271148682} 02/26/2022 04:25:53 - INFO - codeparrot_training - Step 39279: {'lr': 5.9061554527032876e-05, 'samples': 20111360, 'steps': 39279, 'loss/train': 7.042550086975098} 02/26/2022 04:25:56 - INFO - codeparrot_training - Step 39280: {'lr': 5.905099284133952e-05, 'samples': 20111872, 'steps': 39280, 'loss/train': 1.2340879440307617} 02/26/2022 04:26:02 - INFO - codeparrot_training - Step 39281: {'lr': 5.9040431973610945e-05, 'samples': 20112384, 'steps': 39281, 'loss/train': 2.5818324089050293} 02/26/2022 04:26:05 - INFO - codeparrot_training - Step 39282: {'lr': 5.9029871923892584e-05, 'samples': 20112896, 'steps': 39282, 'loss/train': 2.3087260723114014} 02/26/2022 04:26:11 - INFO - codeparrot_training - Step 39283: {'lr': 5.901931269222954e-05, 'samples': 20113408, 'steps': 39283, 'loss/train': 3.098461866378784} 02/26/2022 04:26:14 - INFO - codeparrot_training - Step 39284: {'lr': 5.9008754278667196e-05, 'samples': 20113920, 'steps': 39284, 'loss/train': 0.7448910474777222} 02/26/2022 04:26:21 - INFO - codeparrot_training - Step 39285: {'lr': 5.899819668325071e-05, 'samples': 20114432, 'steps': 39285, 'loss/train': 2.2561213970184326} 02/26/2022 04:26:24 - INFO - codeparrot_training - Step 39286: {'lr': 5.898763990602529e-05, 'samples': 20114944, 'steps': 39286, 'loss/train': 1.7975585460662842} 02/26/2022 04:26:30 - INFO - codeparrot_training - Step 39287: {'lr': 5.8977083947036105e-05, 'samples': 20115456, 'steps': 39287, 'loss/train': 1.4839191436767578} 02/26/2022 04:26:33 - INFO - codeparrot_training - Step 39288: {'lr': 5.8966528806328496e-05, 'samples': 20115968, 'steps': 39288, 'loss/train': 2.5699148178100586} 02/26/2022 04:26:39 - INFO - codeparrot_training - Step 39289: {'lr': 5.89559744839476e-05, 'samples': 20116480, 'steps': 39289, 'loss/train': 2.296220541000366} 02/26/2022 04:26:44 - INFO - codeparrot_training - Step 39290: {'lr': 5.8945420979938686e-05, 'samples': 20116992, 'steps': 39290, 'loss/train': 1.9628002643585205} 02/26/2022 04:26:48 - INFO - codeparrot_training - Step 39291: {'lr': 5.8934868294346825e-05, 'samples': 20117504, 'steps': 39291, 'loss/train': 0.9136307239532471} 02/26/2022 04:26:53 - INFO - codeparrot_training - Step 39292: {'lr': 5.8924316427217375e-05, 'samples': 20118016, 'steps': 39292, 'loss/train': 1.4834866523742676} 02/26/2022 04:26:57 - INFO - codeparrot_training - Step 39293: {'lr': 5.891376537859553e-05, 'samples': 20118528, 'steps': 39293, 'loss/train': 8.36025619506836} 02/26/2022 04:27:00 - INFO - codeparrot_training - Step 39294: {'lr': 5.89032151485264e-05, 'samples': 20119040, 'steps': 39294, 'loss/train': 1.8752689361572266} 02/26/2022 04:27:07 - INFO - codeparrot_training - Step 39295: {'lr': 5.889266573705515e-05, 'samples': 20119552, 'steps': 39295, 'loss/train': 1.0926858186721802} 02/26/2022 04:27:10 - INFO - codeparrot_training - Step 39296: {'lr': 5.8882117144227115e-05, 'samples': 20120064, 'steps': 39296, 'loss/train': 0.7730099558830261} 02/26/2022 04:27:16 - INFO - codeparrot_training - Step 39297: {'lr': 5.887156937008739e-05, 'samples': 20120576, 'steps': 39297, 'loss/train': 0.5640190839767456} 02/26/2022 04:27:22 - INFO - codeparrot_training - Step 39298: {'lr': 5.886102241468119e-05, 'samples': 20121088, 'steps': 39298, 'loss/train': 1.6185075044631958} 02/26/2022 04:27:25 - INFO - codeparrot_training - Step 39299: {'lr': 5.885047627805359e-05, 'samples': 20121600, 'steps': 39299, 'loss/train': 2.024411678314209} 02/26/2022 04:27:31 - INFO - codeparrot_training - Step 39300: {'lr': 5.883993096024992e-05, 'samples': 20122112, 'steps': 39300, 'loss/train': 2.4282212257385254} 02/26/2022 04:27:34 - INFO - codeparrot_training - Step 39301: {'lr': 5.88293864613153e-05, 'samples': 20122624, 'steps': 39301, 'loss/train': 2.111492395401001} 02/26/2022 04:27:40 - INFO - codeparrot_training - Step 39302: {'lr': 5.881884278129487e-05, 'samples': 20123136, 'steps': 39302, 'loss/train': 1.8113347291946411} 02/26/2022 04:27:43 - INFO - codeparrot_training - Step 39303: {'lr': 5.880829992023376e-05, 'samples': 20123648, 'steps': 39303, 'loss/train': 1.0204998254776} 02/26/2022 04:27:50 - INFO - codeparrot_training - Step 39304: {'lr': 5.8797757878177234e-05, 'samples': 20124160, 'steps': 39304, 'loss/train': 1.738308310508728} 02/26/2022 04:27:53 - INFO - codeparrot_training - Step 39305: {'lr': 5.8787216655170426e-05, 'samples': 20124672, 'steps': 39305, 'loss/train': 2.6062076091766357} 02/26/2022 04:27:59 - INFO - codeparrot_training - Step 39306: {'lr': 5.877667625125843e-05, 'samples': 20125184, 'steps': 39306, 'loss/train': 2.0921237468719482} 02/26/2022 04:28:02 - INFO - codeparrot_training - Step 39307: {'lr': 5.876613666648639e-05, 'samples': 20125696, 'steps': 39307, 'loss/train': 2.2125446796417236} 02/26/2022 04:28:08 - INFO - codeparrot_training - Step 39308: {'lr': 5.875559790089957e-05, 'samples': 20126208, 'steps': 39308, 'loss/train': 0.8944153785705566} 02/26/2022 04:28:11 - INFO - codeparrot_training - Step 39309: {'lr': 5.874505995454302e-05, 'samples': 20126720, 'steps': 39309, 'loss/train': 2.05277681350708} 02/26/2022 04:28:17 - INFO - codeparrot_training - Step 39310: {'lr': 5.8734522827461836e-05, 'samples': 20127232, 'steps': 39310, 'loss/train': 2.3895952701568604} 02/26/2022 04:28:20 - INFO - codeparrot_training - Step 39311: {'lr': 5.872398651970137e-05, 'samples': 20127744, 'steps': 39311, 'loss/train': 2.25655460357666} 02/26/2022 04:28:26 - INFO - codeparrot_training - Step 39312: {'lr': 5.871345103130646e-05, 'samples': 20128256, 'steps': 39312, 'loss/train': 1.6194692850112915} 02/26/2022 04:28:29 - INFO - codeparrot_training - Step 39313: {'lr': 5.870291636232247e-05, 'samples': 20128768, 'steps': 39313, 'loss/train': 1.6633578538894653} 02/26/2022 04:28:36 - INFO - codeparrot_training - Step 39314: {'lr': 5.869238251279438e-05, 'samples': 20129280, 'steps': 39314, 'loss/train': 1.8502283096313477} 02/26/2022 04:28:39 - INFO - codeparrot_training - Step 39315: {'lr': 5.8681849482767504e-05, 'samples': 20129792, 'steps': 39315, 'loss/train': 2.0411758422851562} 02/26/2022 04:28:45 - INFO - codeparrot_training - Step 39316: {'lr': 5.8671317272286664e-05, 'samples': 20130304, 'steps': 39316, 'loss/train': 2.1014180183410645} 02/26/2022 04:28:48 - INFO - codeparrot_training - Step 39317: {'lr': 5.8660785881397255e-05, 'samples': 20130816, 'steps': 39317, 'loss/train': 1.5701884031295776} 02/26/2022 04:28:54 - INFO - codeparrot_training - Step 39318: {'lr': 5.8650255310144204e-05, 'samples': 20131328, 'steps': 39318, 'loss/train': 1.7019730806350708} 02/26/2022 04:28:57 - INFO - codeparrot_training - Step 39319: {'lr': 5.8639725558572776e-05, 'samples': 20131840, 'steps': 39319, 'loss/train': 1.6158745288848877} 02/26/2022 04:29:03 - INFO - codeparrot_training - Step 39320: {'lr': 5.8629196626728e-05, 'samples': 20132352, 'steps': 39320, 'loss/train': 0.9566727876663208} 02/26/2022 04:29:06 - INFO - codeparrot_training - Step 39321: {'lr': 5.861866851465494e-05, 'samples': 20132864, 'steps': 39321, 'loss/train': 0.9225515723228455} 02/26/2022 04:29:12 - INFO - codeparrot_training - Step 39322: {'lr': 5.8608141222398717e-05, 'samples': 20133376, 'steps': 39322, 'loss/train': 1.4531960487365723} 02/26/2022 04:29:15 - INFO - codeparrot_training - Step 39323: {'lr': 5.859761475000447e-05, 'samples': 20133888, 'steps': 39323, 'loss/train': 1.919107437133789} 02/26/2022 04:29:21 - INFO - codeparrot_training - Step 39324: {'lr': 5.858708909751731e-05, 'samples': 20134400, 'steps': 39324, 'loss/train': 1.379542589187622} 02/26/2022 04:29:24 - INFO - codeparrot_training - Step 39325: {'lr': 5.857656426498223e-05, 'samples': 20134912, 'steps': 39325, 'loss/train': 0.8056464195251465} 02/26/2022 04:29:30 - INFO - codeparrot_training - Step 39326: {'lr': 5.856604025244433e-05, 'samples': 20135424, 'steps': 39326, 'loss/train': 1.9633475542068481} 02/26/2022 04:29:33 - INFO - codeparrot_training - Step 39327: {'lr': 5.8555517059948796e-05, 'samples': 20135936, 'steps': 39327, 'loss/train': 2.098339557647705} 02/26/2022 04:29:39 - INFO - codeparrot_training - Step 39328: {'lr': 5.854499468754063e-05, 'samples': 20136448, 'steps': 39328, 'loss/train': 1.830745816230774} 02/26/2022 04:29:42 - INFO - codeparrot_training - Step 39329: {'lr': 5.8534473135264886e-05, 'samples': 20136960, 'steps': 39329, 'loss/train': 2.612426519393921} 02/26/2022 04:29:47 - INFO - codeparrot_training - Step 39330: {'lr': 5.852395240316663e-05, 'samples': 20137472, 'steps': 39330, 'loss/train': 1.74697744846344} 02/26/2022 04:29:51 - INFO - codeparrot_training - Step 39331: {'lr': 5.851343249129101e-05, 'samples': 20137984, 'steps': 39331, 'loss/train': 1.3093727827072144} 02/26/2022 04:29:57 - INFO - codeparrot_training - Step 39332: {'lr': 5.850291339968297e-05, 'samples': 20138496, 'steps': 39332, 'loss/train': 0.5128688216209412} 02/26/2022 04:30:01 - INFO - codeparrot_training - Step 39333: {'lr': 5.84923951283878e-05, 'samples': 20139008, 'steps': 39333, 'loss/train': 1.4556185007095337} 02/26/2022 04:30:06 - INFO - codeparrot_training - Step 39334: {'lr': 5.848187767745025e-05, 'samples': 20139520, 'steps': 39334, 'loss/train': 2.0599207878112793} 02/26/2022 04:30:10 - INFO - codeparrot_training - Step 39335: {'lr': 5.847136104691558e-05, 'samples': 20140032, 'steps': 39335, 'loss/train': 0.5377969145774841} 02/26/2022 04:30:15 - INFO - codeparrot_training - Step 39336: {'lr': 5.8460845236828736e-05, 'samples': 20140544, 'steps': 39336, 'loss/train': 2.2448954582214355} 02/26/2022 04:30:19 - INFO - codeparrot_training - Step 39337: {'lr': 5.8450330247234925e-05, 'samples': 20141056, 'steps': 39337, 'loss/train': 1.9493519067764282} 02/26/2022 04:30:24 - INFO - codeparrot_training - Step 39338: {'lr': 5.8439816078178945e-05, 'samples': 20141568, 'steps': 39338, 'loss/train': 2.260248899459839} 02/26/2022 04:30:28 - INFO - codeparrot_training - Step 39339: {'lr': 5.8429302729706035e-05, 'samples': 20142080, 'steps': 39339, 'loss/train': 1.9130685329437256} 02/26/2022 04:30:35 - INFO - codeparrot_training - Step 39340: {'lr': 5.841879020186111e-05, 'samples': 20142592, 'steps': 39340, 'loss/train': 1.8006200790405273} 02/26/2022 04:30:38 - INFO - codeparrot_training - Step 39341: {'lr': 5.840827849468938e-05, 'samples': 20143104, 'steps': 39341, 'loss/train': 1.5545932054519653} 02/26/2022 04:30:44 - INFO - codeparrot_training - Step 39342: {'lr': 5.839776760823559e-05, 'samples': 20143616, 'steps': 39342, 'loss/train': 1.3660191297531128} 02/26/2022 04:30:47 - INFO - codeparrot_training - Step 39343: {'lr': 5.838725754254501e-05, 'samples': 20144128, 'steps': 39343, 'loss/train': 1.8794853687286377} 02/26/2022 04:30:53 - INFO - codeparrot_training - Step 39344: {'lr': 5.837674829766257e-05, 'samples': 20144640, 'steps': 39344, 'loss/train': 0.9432389140129089} 02/26/2022 04:30:56 - INFO - codeparrot_training - Step 39345: {'lr': 5.8366239873633235e-05, 'samples': 20145152, 'steps': 39345, 'loss/train': 1.6982903480529785} 02/26/2022 04:31:02 - INFO - codeparrot_training - Step 39346: {'lr': 5.835573227050214e-05, 'samples': 20145664, 'steps': 39346, 'loss/train': 1.7196199893951416} 02/26/2022 04:31:07 - INFO - codeparrot_training - Step 39347: {'lr': 5.834522548831422e-05, 'samples': 20146176, 'steps': 39347, 'loss/train': 0.8640785813331604} 02/26/2022 04:31:11 - INFO - codeparrot_training - Step 39348: {'lr': 5.8334719527114527e-05, 'samples': 20146688, 'steps': 39348, 'loss/train': 0.711805522441864} 02/26/2022 04:31:17 - INFO - codeparrot_training - Step 39349: {'lr': 5.832421438694793e-05, 'samples': 20147200, 'steps': 39349, 'loss/train': 1.9360721111297607} 02/26/2022 04:31:21 - INFO - codeparrot_training - Step 39350: {'lr': 5.831371006785963e-05, 'samples': 20147712, 'steps': 39350, 'loss/train': 1.4671344757080078} 02/26/2022 04:31:26 - INFO - codeparrot_training - Step 39351: {'lr': 5.830320656989452e-05, 'samples': 20148224, 'steps': 39351, 'loss/train': 0.11848215758800507} 02/26/2022 04:31:30 - INFO - codeparrot_training - Step 39352: {'lr': 5.829270389309763e-05, 'samples': 20148736, 'steps': 39352, 'loss/train': 0.18754246830940247} 02/26/2022 04:31:35 - INFO - codeparrot_training - Step 39353: {'lr': 5.82822020375138e-05, 'samples': 20149248, 'steps': 39353, 'loss/train': 1.3263994455337524} 02/26/2022 04:31:39 - INFO - codeparrot_training - Step 39354: {'lr': 5.8271701003188235e-05, 'samples': 20149760, 'steps': 39354, 'loss/train': 1.3562870025634766} 02/26/2022 04:31:44 - INFO - codeparrot_training - Step 39355: {'lr': 5.8261200790165836e-05, 'samples': 20150272, 'steps': 39355, 'loss/train': 1.4552899599075317} 02/26/2022 04:31:48 - INFO - codeparrot_training - Step 39356: {'lr': 5.825070139849156e-05, 'samples': 20150784, 'steps': 39356, 'loss/train': 0.9698016047477722} 02/26/2022 04:31:53 - INFO - codeparrot_training - Step 39357: {'lr': 5.824020282821033e-05, 'samples': 20151296, 'steps': 39357, 'loss/train': 2.542808771133423} 02/26/2022 04:31:57 - INFO - codeparrot_training - Step 39358: {'lr': 5.822970507936723e-05, 'samples': 20151808, 'steps': 39358, 'loss/train': 0.5101855397224426} 02/26/2022 04:32:02 - INFO - codeparrot_training - Step 39359: {'lr': 5.821920815200718e-05, 'samples': 20152320, 'steps': 39359, 'loss/train': 2.373020887374878} 02/26/2022 04:32:06 - INFO - codeparrot_training - Step 39360: {'lr': 5.8208712046175144e-05, 'samples': 20152832, 'steps': 39360, 'loss/train': 1.9011986255645752} 02/26/2022 04:32:12 - INFO - codeparrot_training - Step 39361: {'lr': 5.819821676191603e-05, 'samples': 20153344, 'steps': 39361, 'loss/train': 2.2148470878601074} 02/26/2022 04:32:15 - INFO - codeparrot_training - Step 39362: {'lr': 5.8187722299274906e-05, 'samples': 20153856, 'steps': 39362, 'loss/train': 1.9350521564483643} 02/26/2022 04:32:21 - INFO - codeparrot_training - Step 39363: {'lr': 5.817722865829669e-05, 'samples': 20154368, 'steps': 39363, 'loss/train': 1.215533971786499} 02/26/2022 04:32:24 - INFO - codeparrot_training - Step 39364: {'lr': 5.816673583902632e-05, 'samples': 20154880, 'steps': 39364, 'loss/train': 0.38346776366233826} 02/26/2022 04:32:30 - INFO - codeparrot_training - Step 39365: {'lr': 5.815624384150867e-05, 'samples': 20155392, 'steps': 39365, 'loss/train': 2.0341482162475586} 02/26/2022 04:32:33 - INFO - codeparrot_training - Step 39366: {'lr': 5.8145752665788785e-05, 'samples': 20155904, 'steps': 39366, 'loss/train': 1.2874284982681274} 02/26/2022 04:32:39 - INFO - codeparrot_training - Step 39367: {'lr': 5.813526231191155e-05, 'samples': 20156416, 'steps': 39367, 'loss/train': 1.7480957508087158} 02/26/2022 04:32:42 - INFO - codeparrot_training - Step 39368: {'lr': 5.812477277992204e-05, 'samples': 20156928, 'steps': 39368, 'loss/train': 0.8566776514053345} 02/26/2022 04:32:48 - INFO - codeparrot_training - Step 39369: {'lr': 5.8114284069864934e-05, 'samples': 20157440, 'steps': 39369, 'loss/train': 0.7865034341812134} 02/26/2022 04:32:51 - INFO - codeparrot_training - Step 39370: {'lr': 5.810379618178541e-05, 'samples': 20157952, 'steps': 39370, 'loss/train': 1.5552150011062622} 02/26/2022 04:32:58 - INFO - codeparrot_training - Step 39371: {'lr': 5.809330911572819e-05, 'samples': 20158464, 'steps': 39371, 'loss/train': 1.130629301071167} 02/26/2022 04:33:01 - INFO - codeparrot_training - Step 39372: {'lr': 5.808282287173844e-05, 'samples': 20158976, 'steps': 39372, 'loss/train': 1.0659348964691162} 02/26/2022 04:33:07 - INFO - codeparrot_training - Step 39373: {'lr': 5.807233744986079e-05, 'samples': 20159488, 'steps': 39373, 'loss/train': 0.5839967131614685} 02/26/2022 04:33:10 - INFO - codeparrot_training - Step 39374: {'lr': 5.8061852850140375e-05, 'samples': 20160000, 'steps': 39374, 'loss/train': 0.6007506251335144} 02/26/2022 04:33:16 - INFO - codeparrot_training - Step 39375: {'lr': 5.805136907262198e-05, 'samples': 20160512, 'steps': 39375, 'loss/train': 0.9831951260566711} 02/26/2022 04:33:19 - INFO - codeparrot_training - Step 39376: {'lr': 5.80408861173507e-05, 'samples': 20161024, 'steps': 39376, 'loss/train': 1.2300870418548584} 02/26/2022 04:33:25 - INFO - codeparrot_training - Step 39377: {'lr': 5.8030403984371175e-05, 'samples': 20161536, 'steps': 39377, 'loss/train': 1.255632996559143} 02/26/2022 04:33:28 - INFO - codeparrot_training - Step 39378: {'lr': 5.801992267372849e-05, 'samples': 20162048, 'steps': 39378, 'loss/train': 1.1090407371520996} 02/26/2022 04:33:34 - INFO - codeparrot_training - Step 39379: {'lr': 5.800944218546744e-05, 'samples': 20162560, 'steps': 39379, 'loss/train': 2.11875319480896} 02/26/2022 04:33:37 - INFO - codeparrot_training - Step 39380: {'lr': 5.799896251963305e-05, 'samples': 20163072, 'steps': 39380, 'loss/train': 0.767371416091919} 02/26/2022 04:33:43 - INFO - codeparrot_training - Step 39381: {'lr': 5.798848367627016e-05, 'samples': 20163584, 'steps': 39381, 'loss/train': 2.4251301288604736} 02/26/2022 04:33:46 - INFO - codeparrot_training - Step 39382: {'lr': 5.7978005655423605e-05, 'samples': 20164096, 'steps': 39382, 'loss/train': 1.3876231908798218} 02/26/2022 04:33:52 - INFO - codeparrot_training - Step 39383: {'lr': 5.7967528457138234e-05, 'samples': 20164608, 'steps': 39383, 'loss/train': 1.9034079313278198} 02/26/2022 04:33:55 - INFO - codeparrot_training - Step 39384: {'lr': 5.7957052081459074e-05, 'samples': 20165120, 'steps': 39384, 'loss/train': 1.6483125686645508} 02/26/2022 04:34:01 - INFO - codeparrot_training - Step 39385: {'lr': 5.794657652843091e-05, 'samples': 20165632, 'steps': 39385, 'loss/train': 1.4535356760025024} 02/26/2022 04:34:04 - INFO - codeparrot_training - Step 39386: {'lr': 5.793610179809863e-05, 'samples': 20166144, 'steps': 39386, 'loss/train': 1.7032256126403809} 02/26/2022 04:34:11 - INFO - codeparrot_training - Step 39387: {'lr': 5.7925627890507036e-05, 'samples': 20166656, 'steps': 39387, 'loss/train': 1.9047123193740845} 02/26/2022 04:34:14 - INFO - codeparrot_training - Step 39388: {'lr': 5.791515480570111e-05, 'samples': 20167168, 'steps': 39388, 'loss/train': 1.7393555641174316} 02/26/2022 04:34:20 - INFO - codeparrot_training - Step 39389: {'lr': 5.790468254372569e-05, 'samples': 20167680, 'steps': 39389, 'loss/train': 1.853602409362793} 02/26/2022 04:34:23 - INFO - codeparrot_training - Step 39390: {'lr': 5.7894211104625614e-05, 'samples': 20168192, 'steps': 39390, 'loss/train': 2.9697265625} 02/26/2022 04:34:29 - INFO - codeparrot_training - Step 39391: {'lr': 5.788374048844566e-05, 'samples': 20168704, 'steps': 39391, 'loss/train': 1.748284101486206} 02/26/2022 04:34:32 - INFO - codeparrot_training - Step 39392: {'lr': 5.787327069523085e-05, 'samples': 20169216, 'steps': 39392, 'loss/train': 1.6234177350997925} 02/26/2022 04:34:38 - INFO - codeparrot_training - Step 39393: {'lr': 5.786280172502592e-05, 'samples': 20169728, 'steps': 39393, 'loss/train': 1.5879952907562256} 02/26/2022 04:34:41 - INFO - codeparrot_training - Step 39394: {'lr': 5.7852333577875725e-05, 'samples': 20170240, 'steps': 39394, 'loss/train': 1.169845700263977} 02/26/2022 04:34:47 - INFO - codeparrot_training - Step 39395: {'lr': 5.7841866253825145e-05, 'samples': 20170752, 'steps': 39395, 'loss/train': 0.5379805564880371} 02/26/2022 04:34:50 - INFO - codeparrot_training - Step 39396: {'lr': 5.783139975291893e-05, 'samples': 20171264, 'steps': 39396, 'loss/train': 1.8642330169677734} 02/26/2022 04:34:57 - INFO - codeparrot_training - Step 39397: {'lr': 5.7820934075202005e-05, 'samples': 20171776, 'steps': 39397, 'loss/train': 2.3852691650390625} 02/26/2022 04:35:00 - INFO - codeparrot_training - Step 39398: {'lr': 5.781046922071922e-05, 'samples': 20172288, 'steps': 39398, 'loss/train': 2.105473518371582} 02/26/2022 04:35:06 - INFO - codeparrot_training - Step 39399: {'lr': 5.780000518951534e-05, 'samples': 20172800, 'steps': 39399, 'loss/train': 2.3800384998321533} 02/26/2022 04:35:11 - INFO - codeparrot_training - Step 39400: {'lr': 5.7789541981635134e-05, 'samples': 20173312, 'steps': 39400, 'loss/train': 2.753495216369629} 02/26/2022 04:35:15 - INFO - codeparrot_training - Step 39401: {'lr': 5.777907959712356e-05, 'samples': 20173824, 'steps': 39401, 'loss/train': 0.9524871110916138} 02/26/2022 04:35:20 - INFO - codeparrot_training - Step 39402: {'lr': 5.776861803602537e-05, 'samples': 20174336, 'steps': 39402, 'loss/train': 2.660076379776001} 02/26/2022 04:35:24 - INFO - codeparrot_training - Step 39403: {'lr': 5.77581572983854e-05, 'samples': 20174848, 'steps': 39403, 'loss/train': 2.8996992111206055} 02/26/2022 04:35:29 - INFO - codeparrot_training - Step 39404: {'lr': 5.774769738424837e-05, 'samples': 20175360, 'steps': 39404, 'loss/train': 1.8910436630249023} 02/26/2022 04:35:33 - INFO - codeparrot_training - Step 39405: {'lr': 5.7737238293659205e-05, 'samples': 20175872, 'steps': 39405, 'loss/train': 1.8484967947006226} 02/26/2022 04:35:40 - INFO - codeparrot_training - Step 39406: {'lr': 5.7726780026662594e-05, 'samples': 20176384, 'steps': 39406, 'loss/train': 2.237628936767578} 02/26/2022 04:35:43 - INFO - codeparrot_training - Step 39407: {'lr': 5.771632258330356e-05, 'samples': 20176896, 'steps': 39407, 'loss/train': 2.1191487312316895} 02/26/2022 04:35:49 - INFO - codeparrot_training - Step 39408: {'lr': 5.770586596362659e-05, 'samples': 20177408, 'steps': 39408, 'loss/train': 2.1390039920806885} 02/26/2022 04:35:52 - INFO - codeparrot_training - Step 39409: {'lr': 5.769541016767671e-05, 'samples': 20177920, 'steps': 39409, 'loss/train': 3.7390949726104736} 02/26/2022 04:35:58 - INFO - codeparrot_training - Step 39410: {'lr': 5.768495519549857e-05, 'samples': 20178432, 'steps': 39410, 'loss/train': 2.4449284076690674} 02/26/2022 04:36:01 - INFO - codeparrot_training - Step 39411: {'lr': 5.767450104713715e-05, 'samples': 20178944, 'steps': 39411, 'loss/train': 2.2156307697296143} 02/26/2022 04:36:07 - INFO - codeparrot_training - Step 39412: {'lr': 5.766404772263692e-05, 'samples': 20179456, 'steps': 39412, 'loss/train': 1.139458417892456} 02/26/2022 04:36:11 - INFO - codeparrot_training - Step 39413: {'lr': 5.7653595222042944e-05, 'samples': 20179968, 'steps': 39413, 'loss/train': 2.341097354888916} 02/26/2022 04:36:16 - INFO - codeparrot_training - Step 39414: {'lr': 5.764314354539982e-05, 'samples': 20180480, 'steps': 39414, 'loss/train': 0.0832589790225029} 02/26/2022 04:36:20 - INFO - codeparrot_training - Step 39415: {'lr': 5.763269269275251e-05, 'samples': 20180992, 'steps': 39415, 'loss/train': 1.841733694076538} 02/26/2022 04:36:27 - INFO - codeparrot_training - Step 39416: {'lr': 5.762224266414554e-05, 'samples': 20181504, 'steps': 39416, 'loss/train': 2.0044827461242676} 02/26/2022 04:36:30 - INFO - codeparrot_training - Step 39417: {'lr': 5.761179345962383e-05, 'samples': 20182016, 'steps': 39417, 'loss/train': 2.052302360534668} 02/26/2022 04:36:36 - INFO - codeparrot_training - Step 39418: {'lr': 5.760134507923206e-05, 'samples': 20182528, 'steps': 39418, 'loss/train': 1.4190049171447754} 02/26/2022 04:36:39 - INFO - codeparrot_training - Step 39419: {'lr': 5.7590897523015106e-05, 'samples': 20183040, 'steps': 39419, 'loss/train': 2.581770896911621} 02/26/2022 04:36:45 - INFO - codeparrot_training - Step 39420: {'lr': 5.758045079101765e-05, 'samples': 20183552, 'steps': 39420, 'loss/train': 1.4611321687698364} 02/26/2022 04:36:48 - INFO - codeparrot_training - Step 39421: {'lr': 5.757000488328446e-05, 'samples': 20184064, 'steps': 39421, 'loss/train': 2.043133020401001} 02/26/2022 04:36:54 - INFO - codeparrot_training - Step 39422: {'lr': 5.755955979986016e-05, 'samples': 20184576, 'steps': 39422, 'loss/train': 2.5207080841064453} 02/26/2022 04:36:57 - INFO - codeparrot_training - Step 39423: {'lr': 5.7549115540789705e-05, 'samples': 20185088, 'steps': 39423, 'loss/train': 1.8859179019927979} 02/26/2022 04:37:03 - INFO - codeparrot_training - Step 39424: {'lr': 5.753867210611771e-05, 'samples': 20185600, 'steps': 39424, 'loss/train': 0.813764750957489} 02/26/2022 04:37:06 - INFO - codeparrot_training - Step 39425: {'lr': 5.752822949588893e-05, 'samples': 20186112, 'steps': 39425, 'loss/train': 1.2927632331848145} 02/26/2022 04:37:12 - INFO - codeparrot_training - Step 39426: {'lr': 5.751778771014801e-05, 'samples': 20186624, 'steps': 39426, 'loss/train': 1.23225998878479} 02/26/2022 04:37:15 - INFO - codeparrot_training - Step 39427: {'lr': 5.750734674893987e-05, 'samples': 20187136, 'steps': 39427, 'loss/train': 1.6800318956375122} 02/26/2022 04:37:21 - INFO - codeparrot_training - Step 39428: {'lr': 5.749690661230914e-05, 'samples': 20187648, 'steps': 39428, 'loss/train': 2.3167762756347656} 02/26/2022 04:37:24 - INFO - codeparrot_training - Step 39429: {'lr': 5.748646730030049e-05, 'samples': 20188160, 'steps': 39429, 'loss/train': 0.8933626413345337} 02/26/2022 04:37:30 - INFO - codeparrot_training - Step 39430: {'lr': 5.747602881295866e-05, 'samples': 20188672, 'steps': 39430, 'loss/train': 2.0880558490753174} 02/26/2022 04:37:33 - INFO - codeparrot_training - Step 39431: {'lr': 5.746559115032843e-05, 'samples': 20189184, 'steps': 39431, 'loss/train': 1.8017241954803467} 02/26/2022 04:37:40 - INFO - codeparrot_training - Step 39432: {'lr': 5.745515431245449e-05, 'samples': 20189696, 'steps': 39432, 'loss/train': 1.7788697481155396} 02/26/2022 04:37:44 - INFO - codeparrot_training - Step 39433: {'lr': 5.7444718299381496e-05, 'samples': 20190208, 'steps': 39433, 'loss/train': 1.2854468822479248} 02/26/2022 04:37:49 - INFO - codeparrot_training - Step 39434: {'lr': 5.743428311115415e-05, 'samples': 20190720, 'steps': 39434, 'loss/train': 1.5779608488082886} 02/26/2022 04:37:53 - INFO - codeparrot_training - Step 39435: {'lr': 5.7423848747817256e-05, 'samples': 20191232, 'steps': 39435, 'loss/train': 1.9080621004104614} 02/26/2022 04:37:58 - INFO - codeparrot_training - Step 39436: {'lr': 5.741341520941543e-05, 'samples': 20191744, 'steps': 39436, 'loss/train': 1.552306890487671} 02/26/2022 04:38:02 - INFO - codeparrot_training - Step 39437: {'lr': 5.740298249599338e-05, 'samples': 20192256, 'steps': 39437, 'loss/train': 1.0564757585525513} 02/26/2022 04:38:07 - INFO - codeparrot_training - Step 39438: {'lr': 5.739255060759574e-05, 'samples': 20192768, 'steps': 39438, 'loss/train': 1.4067105054855347} 02/26/2022 04:38:11 - INFO - codeparrot_training - Step 39439: {'lr': 5.738211954426731e-05, 'samples': 20193280, 'steps': 39439, 'loss/train': 2.3647356033325195} 02/26/2022 04:38:16 - INFO - codeparrot_training - Step 39440: {'lr': 5.737168930605272e-05, 'samples': 20193792, 'steps': 39440, 'loss/train': 2.066711664199829} 02/26/2022 04:38:20 - INFO - codeparrot_training - Step 39441: {'lr': 5.736125989299659e-05, 'samples': 20194304, 'steps': 39441, 'loss/train': 1.963661551475525} 02/26/2022 04:38:27 - INFO - codeparrot_training - Step 39442: {'lr': 5.7350831305143766e-05, 'samples': 20194816, 'steps': 39442, 'loss/train': 0.5052897334098816} 02/26/2022 04:38:30 - INFO - codeparrot_training - Step 39443: {'lr': 5.7340403542538696e-05, 'samples': 20195328, 'steps': 39443, 'loss/train': 1.3866177797317505} 02/26/2022 04:38:36 - INFO - codeparrot_training - Step 39444: {'lr': 5.7329976605226206e-05, 'samples': 20195840, 'steps': 39444, 'loss/train': 2.0600945949554443} 02/26/2022 04:38:39 - INFO - codeparrot_training - Step 39445: {'lr': 5.7319550493250856e-05, 'samples': 20196352, 'steps': 39445, 'loss/train': 1.8887900114059448} 02/26/2022 04:38:45 - INFO - codeparrot_training - Step 39446: {'lr': 5.730912520665751e-05, 'samples': 20196864, 'steps': 39446, 'loss/train': 1.9036179780960083} 02/26/2022 04:38:48 - INFO - codeparrot_training - Step 39447: {'lr': 5.729870074549057e-05, 'samples': 20197376, 'steps': 39447, 'loss/train': 1.8594902753829956} 02/26/2022 04:38:54 - INFO - codeparrot_training - Step 39448: {'lr': 5.728827710979484e-05, 'samples': 20197888, 'steps': 39448, 'loss/train': 1.9182264804840088} 02/26/2022 04:38:57 - INFO - codeparrot_training - Step 39449: {'lr': 5.7277854299614876e-05, 'samples': 20198400, 'steps': 39449, 'loss/train': 1.698870062828064} 02/26/2022 04:39:03 - INFO - codeparrot_training - Step 39450: {'lr': 5.7267432314995535e-05, 'samples': 20198912, 'steps': 39450, 'loss/train': 2.3757519721984863} 02/26/2022 04:39:06 - INFO - codeparrot_training - Step 39451: {'lr': 5.725701115598117e-05, 'samples': 20199424, 'steps': 39451, 'loss/train': 0.23331354558467865} 02/26/2022 04:39:13 - INFO - codeparrot_training - Step 39452: {'lr': 5.7246590822616654e-05, 'samples': 20199936, 'steps': 39452, 'loss/train': 0.9359232187271118} 02/26/2022 04:39:16 - INFO - codeparrot_training - Step 39453: {'lr': 5.723617131494646e-05, 'samples': 20200448, 'steps': 39453, 'loss/train': 3.2734756469726562} 02/26/2022 04:39:22 - INFO - codeparrot_training - Step 39454: {'lr': 5.7225752633015354e-05, 'samples': 20200960, 'steps': 39454, 'loss/train': 2.4681668281555176} 02/26/2022 04:39:25 - INFO - codeparrot_training - Step 39455: {'lr': 5.721533477686791e-05, 'samples': 20201472, 'steps': 39455, 'loss/train': 2.2098896503448486} 02/26/2022 04:39:31 - INFO - codeparrot_training - Step 39456: {'lr': 5.720491774654876e-05, 'samples': 20201984, 'steps': 39456, 'loss/train': 2.2053308486938477} 02/26/2022 04:39:34 - INFO - codeparrot_training - Step 39457: {'lr': 5.7194501542102454e-05, 'samples': 20202496, 'steps': 39457, 'loss/train': 0.9811351895332336} 02/26/2022 04:39:40 - INFO - codeparrot_training - Step 39458: {'lr': 5.718408616357376e-05, 'samples': 20203008, 'steps': 39458, 'loss/train': 1.184763789176941} 02/26/2022 04:39:43 - INFO - codeparrot_training - Step 39459: {'lr': 5.717367161100723e-05, 'samples': 20203520, 'steps': 39459, 'loss/train': 1.5487140417099} 02/26/2022 04:39:49 - INFO - codeparrot_training - Step 39460: {'lr': 5.716325788444743e-05, 'samples': 20204032, 'steps': 39460, 'loss/train': 2.138392686843872} 02/26/2022 04:39:52 - INFO - codeparrot_training - Step 39461: {'lr': 5.7152844983938936e-05, 'samples': 20204544, 'steps': 39461, 'loss/train': 1.5722036361694336} 02/26/2022 04:39:58 - INFO - codeparrot_training - Step 39462: {'lr': 5.71424329095265e-05, 'samples': 20205056, 'steps': 39462, 'loss/train': 1.5240744352340698} 02/26/2022 04:40:01 - INFO - codeparrot_training - Step 39463: {'lr': 5.713202166125464e-05, 'samples': 20205568, 'steps': 39463, 'loss/train': 2.2927021980285645} 02/26/2022 04:40:07 - INFO - codeparrot_training - Step 39464: {'lr': 5.7121611239167954e-05, 'samples': 20206080, 'steps': 39464, 'loss/train': 0.7775042057037354} 02/26/2022 04:40:10 - INFO - codeparrot_training - Step 39465: {'lr': 5.711120164331096e-05, 'samples': 20206592, 'steps': 39465, 'loss/train': 1.3991249799728394} 02/26/2022 04:40:16 - INFO - codeparrot_training - Step 39466: {'lr': 5.710079287372841e-05, 'samples': 20207104, 'steps': 39466, 'loss/train': 2.436664342880249} 02/26/2022 04:40:20 - INFO - codeparrot_training - Step 39467: {'lr': 5.709038493046473e-05, 'samples': 20207616, 'steps': 39467, 'loss/train': 1.2326374053955078} 02/26/2022 04:40:26 - INFO - codeparrot_training - Step 39468: {'lr': 5.707997781356472e-05, 'samples': 20208128, 'steps': 39468, 'loss/train': 1.2545090913772583} 02/26/2022 04:40:30 - INFO - codeparrot_training - Step 39469: {'lr': 5.706957152307271e-05, 'samples': 20208640, 'steps': 39469, 'loss/train': 1.5603325366973877} 02/26/2022 04:40:35 - INFO - codeparrot_training - Step 39470: {'lr': 5.705916605903347e-05, 'samples': 20209152, 'steps': 39470, 'loss/train': 0.7118330597877502} 02/26/2022 04:40:39 - INFO - codeparrot_training - Step 39471: {'lr': 5.70487614214914e-05, 'samples': 20209664, 'steps': 39471, 'loss/train': 2.116262674331665} 02/26/2022 04:40:44 - INFO - codeparrot_training - Step 39472: {'lr': 5.703835761049131e-05, 'samples': 20210176, 'steps': 39472, 'loss/train': 1.6070528030395508} 02/26/2022 04:40:48 - INFO - codeparrot_training - Step 39473: {'lr': 5.702795462607749e-05, 'samples': 20210688, 'steps': 39473, 'loss/train': 1.1939600706100464} 02/26/2022 04:40:53 - INFO - codeparrot_training - Step 39474: {'lr': 5.701755246829471e-05, 'samples': 20211200, 'steps': 39474, 'loss/train': 2.164790391921997} 02/26/2022 04:40:57 - INFO - codeparrot_training - Step 39475: {'lr': 5.700715113718746e-05, 'samples': 20211712, 'steps': 39475, 'loss/train': 1.5403966903686523} 02/26/2022 04:41:02 - INFO - codeparrot_training - Step 39476: {'lr': 5.6996750632800215e-05, 'samples': 20212224, 'steps': 39476, 'loss/train': 2.2468080520629883} 02/26/2022 04:41:06 - INFO - codeparrot_training - Step 39477: {'lr': 5.698635095517768e-05, 'samples': 20212736, 'steps': 39477, 'loss/train': 2.535079002380371} 02/26/2022 04:41:13 - INFO - codeparrot_training - Step 39478: {'lr': 5.697595210436432e-05, 'samples': 20213248, 'steps': 39478, 'loss/train': 0.07792261987924576} 02/26/2022 04:41:16 - INFO - codeparrot_training - Step 39479: {'lr': 5.696555408040469e-05, 'samples': 20213760, 'steps': 39479, 'loss/train': 0.05075787007808685} 02/26/2022 04:41:22 - INFO - codeparrot_training - Step 39480: {'lr': 5.695515688334327e-05, 'samples': 20214272, 'steps': 39480, 'loss/train': 1.418874740600586} 02/26/2022 04:41:25 - INFO - codeparrot_training - Step 39481: {'lr': 5.6944760513224725e-05, 'samples': 20214784, 'steps': 39481, 'loss/train': 1.6823254823684692} 02/26/2022 04:41:31 - INFO - codeparrot_training - Step 39482: {'lr': 5.693436497009352e-05, 'samples': 20215296, 'steps': 39482, 'loss/train': 1.9608285427093506} 02/26/2022 04:41:34 - INFO - codeparrot_training - Step 39483: {'lr': 5.6923970253994176e-05, 'samples': 20215808, 'steps': 39483, 'loss/train': 1.87309730052948} 02/26/2022 04:41:40 - INFO - codeparrot_training - Step 39484: {'lr': 5.691357636497119e-05, 'samples': 20216320, 'steps': 39484, 'loss/train': 1.6646419763565063} 02/26/2022 04:41:43 - INFO - codeparrot_training - Step 39485: {'lr': 5.69031833030692e-05, 'samples': 20216832, 'steps': 39485, 'loss/train': 1.9159107208251953} 02/26/2022 04:41:49 - INFO - codeparrot_training - Step 39486: {'lr': 5.689279106833264e-05, 'samples': 20217344, 'steps': 39486, 'loss/train': 1.1292990446090698} 02/26/2022 04:41:52 - INFO - codeparrot_training - Step 39487: {'lr': 5.688239966080605e-05, 'samples': 20217856, 'steps': 39487, 'loss/train': 2.1154253482818604} 02/26/2022 04:41:59 - INFO - codeparrot_training - Step 39488: {'lr': 5.6872009080533885e-05, 'samples': 20218368, 'steps': 39488, 'loss/train': 1.952418327331543} 02/26/2022 04:42:02 - INFO - codeparrot_training - Step 39489: {'lr': 5.686161932756076e-05, 'samples': 20218880, 'steps': 39489, 'loss/train': 1.7327882051467896} 02/26/2022 04:42:08 - INFO - codeparrot_training - Step 39490: {'lr': 5.685123040193113e-05, 'samples': 20219392, 'steps': 39490, 'loss/train': 9.895872116088867} 02/26/2022 04:42:11 - INFO - codeparrot_training - Step 39491: {'lr': 5.684084230368952e-05, 'samples': 20219904, 'steps': 39491, 'loss/train': 2.5402824878692627} 02/26/2022 04:42:17 - INFO - codeparrot_training - Step 39492: {'lr': 5.6830455032880316e-05, 'samples': 20220416, 'steps': 39492, 'loss/train': 2.063387632369995} 02/26/2022 04:42:20 - INFO - codeparrot_training - Step 39493: {'lr': 5.6820068589548166e-05, 'samples': 20220928, 'steps': 39493, 'loss/train': 1.442240834236145} 02/26/2022 04:42:26 - INFO - codeparrot_training - Step 39494: {'lr': 5.680968297373754e-05, 'samples': 20221440, 'steps': 39494, 'loss/train': 2.397127628326416} 02/26/2022 04:42:29 - INFO - codeparrot_training - Step 39495: {'lr': 5.679929818549284e-05, 'samples': 20221952, 'steps': 39495, 'loss/train': 1.9233158826828003} 02/26/2022 04:42:35 - INFO - codeparrot_training - Step 39496: {'lr': 5.678891422485854e-05, 'samples': 20222464, 'steps': 39496, 'loss/train': 2.2014942169189453} 02/26/2022 04:42:38 - INFO - codeparrot_training - Step 39497: {'lr': 5.677853109187927e-05, 'samples': 20222976, 'steps': 39497, 'loss/train': 1.0298023223876953} 02/26/2022 04:42:45 - INFO - codeparrot_training - Step 39498: {'lr': 5.676814878659942e-05, 'samples': 20223488, 'steps': 39498, 'loss/train': 0.9240416288375854} 02/26/2022 04:42:48 - INFO - codeparrot_training - Step 39499: {'lr': 5.675776730906343e-05, 'samples': 20224000, 'steps': 39499, 'loss/train': 0.5989491939544678} 02/26/2022 04:42:54 - INFO - codeparrot_training - Step 39500: {'lr': 5.6747386659315755e-05, 'samples': 20224512, 'steps': 39500, 'loss/train': 1.4479424953460693} 02/26/2022 04:42:57 - INFO - codeparrot_training - Step 39501: {'lr': 5.673700683740099e-05, 'samples': 20225024, 'steps': 39501, 'loss/train': 1.8506078720092773} 02/26/2022 04:43:03 - INFO - codeparrot_training - Step 39502: {'lr': 5.672662784336344e-05, 'samples': 20225536, 'steps': 39502, 'loss/train': 2.0931456089019775} 02/26/2022 04:43:06 - INFO - codeparrot_training - Step 39503: {'lr': 5.6716249677247786e-05, 'samples': 20226048, 'steps': 39503, 'loss/train': 0.44452348351478577} 02/26/2022 04:43:12 - INFO - codeparrot_training - Step 39504: {'lr': 5.670587233909819e-05, 'samples': 20226560, 'steps': 39504, 'loss/train': 1.2021042108535767} 02/26/2022 04:43:15 - INFO - codeparrot_training - Step 39505: {'lr': 5.669549582895933e-05, 'samples': 20227072, 'steps': 39505, 'loss/train': 1.1801081895828247} 02/26/2022 04:43:21 - INFO - codeparrot_training - Step 39506: {'lr': 5.6685120146875545e-05, 'samples': 20227584, 'steps': 39506, 'loss/train': 1.5373822450637817} 02/26/2022 04:43:24 - INFO - codeparrot_training - Step 39507: {'lr': 5.6674745292891434e-05, 'samples': 20228096, 'steps': 39507, 'loss/train': 1.8907417058944702} 02/26/2022 04:43:31 - INFO - codeparrot_training - Step 39508: {'lr': 5.666437126705118e-05, 'samples': 20228608, 'steps': 39508, 'loss/train': 1.6787999868392944} 02/26/2022 04:43:34 - INFO - codeparrot_training - Step 39509: {'lr': 5.665399806939947e-05, 'samples': 20229120, 'steps': 39509, 'loss/train': 0.24926146864891052} 02/26/2022 04:43:40 - INFO - codeparrot_training - Step 39510: {'lr': 5.664362569998055e-05, 'samples': 20229632, 'steps': 39510, 'loss/train': 0.7358347177505493} 02/26/2022 04:43:44 - INFO - codeparrot_training - Step 39511: {'lr': 5.6633254158839074e-05, 'samples': 20230144, 'steps': 39511, 'loss/train': 1.1676466464996338} 02/26/2022 04:43:49 - INFO - codeparrot_training - Step 39512: {'lr': 5.662288344601921e-05, 'samples': 20230656, 'steps': 39512, 'loss/train': 1.6555445194244385} 02/26/2022 04:43:53 - INFO - codeparrot_training - Step 39513: {'lr': 5.6612513561565574e-05, 'samples': 20231168, 'steps': 39513, 'loss/train': 1.7136192321777344} 02/26/2022 04:43:58 - INFO - codeparrot_training - Step 39514: {'lr': 5.6602144505522455e-05, 'samples': 20231680, 'steps': 39514, 'loss/train': 2.512848138809204} 02/26/2022 04:44:02 - INFO - codeparrot_training - Step 39515: {'lr': 5.6591776277934383e-05, 'samples': 20232192, 'steps': 39515, 'loss/train': 1.4572277069091797} 02/26/2022 04:44:07 - INFO - codeparrot_training - Step 39516: {'lr': 5.6581408878845745e-05, 'samples': 20232704, 'steps': 39516, 'loss/train': 1.6925320625305176} 02/26/2022 04:44:13 - INFO - codeparrot_training - Step 39517: {'lr': 5.657104230830096e-05, 'samples': 20233216, 'steps': 39517, 'loss/train': 1.6385324001312256} 02/26/2022 04:44:16 - INFO - codeparrot_training - Step 39518: {'lr': 5.656067656634431e-05, 'samples': 20233728, 'steps': 39518, 'loss/train': 1.2513892650604248} 02/26/2022 04:44:22 - INFO - codeparrot_training - Step 39519: {'lr': 5.655031165302035e-05, 'samples': 20234240, 'steps': 39519, 'loss/train': 1.9805467128753662} 02/26/2022 04:44:25 - INFO - codeparrot_training - Step 39520: {'lr': 5.6539947568373466e-05, 'samples': 20234752, 'steps': 39520, 'loss/train': 1.527952790260315} 02/26/2022 04:44:31 - INFO - codeparrot_training - Step 39521: {'lr': 5.652958431244801e-05, 'samples': 20235264, 'steps': 39521, 'loss/train': 2.454334259033203} 02/26/2022 04:44:34 - INFO - codeparrot_training - Step 39522: {'lr': 5.6519221885288353e-05, 'samples': 20235776, 'steps': 39522, 'loss/train': 2.6586287021636963} 02/26/2022 04:44:41 - INFO - codeparrot_training - Step 39523: {'lr': 5.650886028693888e-05, 'samples': 20236288, 'steps': 39523, 'loss/train': 1.1758193969726562} 02/26/2022 04:44:45 - INFO - codeparrot_training - Step 39524: {'lr': 5.6498499517444044e-05, 'samples': 20236800, 'steps': 39524, 'loss/train': 1.6551343202590942} 02/26/2022 04:44:50 - INFO - codeparrot_training - Step 39525: {'lr': 5.6488139576848226e-05, 'samples': 20237312, 'steps': 39525, 'loss/train': 0.9471426010131836} 02/26/2022 04:44:54 - INFO - codeparrot_training - Step 39526: {'lr': 5.647778046519575e-05, 'samples': 20237824, 'steps': 39526, 'loss/train': 2.598798990249634} 02/26/2022 04:44:59 - INFO - codeparrot_training - Step 39527: {'lr': 5.646742218253095e-05, 'samples': 20238336, 'steps': 39527, 'loss/train': 1.8875553607940674} 02/26/2022 04:45:03 - INFO - codeparrot_training - Step 39528: {'lr': 5.645706472889833e-05, 'samples': 20238848, 'steps': 39528, 'loss/train': 0.7463963627815247} 02/26/2022 04:45:08 - INFO - codeparrot_training - Step 39529: {'lr': 5.6446708104342185e-05, 'samples': 20239360, 'steps': 39529, 'loss/train': 1.8360371589660645} 02/26/2022 04:45:12 - INFO - codeparrot_training - Step 39530: {'lr': 5.643635230890687e-05, 'samples': 20239872, 'steps': 39530, 'loss/train': 1.836422085762024} 02/26/2022 04:45:17 - INFO - codeparrot_training - Step 39531: {'lr': 5.64259973426367e-05, 'samples': 20240384, 'steps': 39531, 'loss/train': 2.352200508117676} 02/26/2022 04:45:21 - INFO - codeparrot_training - Step 39532: {'lr': 5.641564320557616e-05, 'samples': 20240896, 'steps': 39532, 'loss/train': 1.03348970413208} 02/26/2022 04:45:27 - INFO - codeparrot_training - Step 39533: {'lr': 5.6405289897769526e-05, 'samples': 20241408, 'steps': 39533, 'loss/train': 0.4009478986263275} 02/26/2022 04:45:31 - INFO - codeparrot_training - Step 39534: {'lr': 5.639493741926119e-05, 'samples': 20241920, 'steps': 39534, 'loss/train': 1.4194415807724} 02/26/2022 04:45:36 - INFO - codeparrot_training - Step 39535: {'lr': 5.638458577009537e-05, 'samples': 20242432, 'steps': 39535, 'loss/train': 2.4732158184051514} 02/26/2022 04:45:40 - INFO - codeparrot_training - Step 39536: {'lr': 5.637423495031657e-05, 'samples': 20242944, 'steps': 39536, 'loss/train': 1.328683853149414} 02/26/2022 04:45:45 - INFO - codeparrot_training - Step 39537: {'lr': 5.6363884959968996e-05, 'samples': 20243456, 'steps': 39537, 'loss/train': 2.2324206829071045} 02/26/2022 04:45:49 - INFO - codeparrot_training - Step 39538: {'lr': 5.635353579909719e-05, 'samples': 20243968, 'steps': 39538, 'loss/train': 0.16979217529296875} 02/26/2022 04:45:54 - INFO - codeparrot_training - Step 39539: {'lr': 5.6343187467745225e-05, 'samples': 20244480, 'steps': 39539, 'loss/train': 2.948667287826538} 02/26/2022 04:45:58 - INFO - codeparrot_training - Step 39540: {'lr': 5.633283996595762e-05, 'samples': 20244992, 'steps': 39540, 'loss/train': 1.7116411924362183} 02/26/2022 04:46:03 - INFO - codeparrot_training - Step 39541: {'lr': 5.6322493293778533e-05, 'samples': 20245504, 'steps': 39541, 'loss/train': 1.39324951171875} 02/26/2022 04:46:07 - INFO - codeparrot_training - Step 39542: {'lr': 5.631214745125254e-05, 'samples': 20246016, 'steps': 39542, 'loss/train': 1.6892502307891846} 02/26/2022 04:46:13 - INFO - codeparrot_training - Step 39543: {'lr': 5.63018024384237e-05, 'samples': 20246528, 'steps': 39543, 'loss/train': 1.9009425640106201} 02/26/2022 04:46:17 - INFO - codeparrot_training - Step 39544: {'lr': 5.629145825533646e-05, 'samples': 20247040, 'steps': 39544, 'loss/train': 0.9991382360458374} 02/26/2022 04:46:23 - INFO - codeparrot_training - Step 39545: {'lr': 5.628111490203505e-05, 'samples': 20247552, 'steps': 39545, 'loss/train': 1.6263071298599243} 02/26/2022 04:46:26 - INFO - codeparrot_training - Step 39546: {'lr': 5.6270772378563954e-05, 'samples': 20248064, 'steps': 39546, 'loss/train': 2.1895391941070557} 02/26/2022 04:46:32 - INFO - codeparrot_training - Step 39547: {'lr': 5.626043068496722e-05, 'samples': 20248576, 'steps': 39547, 'loss/train': 2.2050204277038574} 02/26/2022 04:46:35 - INFO - codeparrot_training - Step 39548: {'lr': 5.6250089821289375e-05, 'samples': 20249088, 'steps': 39548, 'loss/train': 1.783073902130127} 02/26/2022 04:46:41 - INFO - codeparrot_training - Step 39549: {'lr': 5.623974978757454e-05, 'samples': 20249600, 'steps': 39549, 'loss/train': 2.015637159347534} 02/26/2022 04:46:44 - INFO - codeparrot_training - Step 39550: {'lr': 5.622941058386716e-05, 'samples': 20250112, 'steps': 39550, 'loss/train': 2.2001965045928955} 02/26/2022 04:46:50 - INFO - codeparrot_training - Step 39551: {'lr': 5.621907221021144e-05, 'samples': 20250624, 'steps': 39551, 'loss/train': 0.774245023727417} 02/26/2022 04:46:54 - INFO - codeparrot_training - Step 39552: {'lr': 5.6208734666651686e-05, 'samples': 20251136, 'steps': 39552, 'loss/train': 1.522621989250183} 02/26/2022 04:46:59 - INFO - codeparrot_training - Step 39553: {'lr': 5.6198397953232114e-05, 'samples': 20251648, 'steps': 39553, 'loss/train': 1.6214483976364136} 02/26/2022 04:47:03 - INFO - codeparrot_training - Step 39554: {'lr': 5.618806206999716e-05, 'samples': 20252160, 'steps': 39554, 'loss/train': 1.1430423259735107} 02/26/2022 04:47:08 - INFO - codeparrot_training - Step 39555: {'lr': 5.6177727016990976e-05, 'samples': 20252672, 'steps': 39555, 'loss/train': 0.7819818258285522} 02/26/2022 04:47:12 - INFO - codeparrot_training - Step 39556: {'lr': 5.616739279425787e-05, 'samples': 20253184, 'steps': 39556, 'loss/train': 1.6150356531143188} 02/26/2022 04:47:17 - INFO - codeparrot_training - Step 39557: {'lr': 5.6157059401842045e-05, 'samples': 20253696, 'steps': 39557, 'loss/train': 1.6539686918258667} 02/26/2022 04:47:21 - INFO - codeparrot_training - Step 39558: {'lr': 5.614672683978789e-05, 'samples': 20254208, 'steps': 39558, 'loss/train': 1.4277534484863281} 02/26/2022 04:47:28 - INFO - codeparrot_training - Step 39559: {'lr': 5.6136395108139585e-05, 'samples': 20254720, 'steps': 39559, 'loss/train': 1.8054852485656738} 02/26/2022 04:47:31 - INFO - codeparrot_training - Step 39560: {'lr': 5.612606420694141e-05, 'samples': 20255232, 'steps': 39560, 'loss/train': 1.9222033023834229} 02/26/2022 04:47:37 - INFO - codeparrot_training - Step 39561: {'lr': 5.6115734136237554e-05, 'samples': 20255744, 'steps': 39561, 'loss/train': 2.471508026123047} 02/26/2022 04:47:40 - INFO - codeparrot_training - Step 39562: {'lr': 5.610540489607241e-05, 'samples': 20256256, 'steps': 39562, 'loss/train': 0.47053101658821106} 02/26/2022 04:47:45 - INFO - codeparrot_training - Step 39563: {'lr': 5.609507648649012e-05, 'samples': 20256768, 'steps': 39563, 'loss/train': 2.0279135704040527} 02/26/2022 04:47:49 - INFO - codeparrot_training - Step 39564: {'lr': 5.608474890753495e-05, 'samples': 20257280, 'steps': 39564, 'loss/train': 3.046379566192627} 02/26/2022 04:47:55 - INFO - codeparrot_training - Step 39565: {'lr': 5.60744221592511e-05, 'samples': 20257792, 'steps': 39565, 'loss/train': 1.2683528661727905} 02/26/2022 04:47:58 - INFO - codeparrot_training - Step 39566: {'lr': 5.606409624168291e-05, 'samples': 20258304, 'steps': 39566, 'loss/train': 1.5101704597473145} 02/26/2022 04:48:04 - INFO - codeparrot_training - Step 39567: {'lr': 5.6053771154874536e-05, 'samples': 20258816, 'steps': 39567, 'loss/train': 1.7682124376296997} 02/26/2022 04:48:07 - INFO - codeparrot_training - Step 39568: {'lr': 5.60434468988702e-05, 'samples': 20259328, 'steps': 39568, 'loss/train': 1.6780132055282593} 02/26/2022 04:48:14 - INFO - codeparrot_training - Step 39569: {'lr': 5.6033123473714174e-05, 'samples': 20259840, 'steps': 39569, 'loss/train': 1.5915560722351074} 02/26/2022 04:48:17 - INFO - codeparrot_training - Step 39570: {'lr': 5.602280087945058e-05, 'samples': 20260352, 'steps': 39570, 'loss/train': 1.4811367988586426} 02/26/2022 04:48:23 - INFO - codeparrot_training - Step 39571: {'lr': 5.601247911612381e-05, 'samples': 20260864, 'steps': 39571, 'loss/train': 2.1106085777282715} 02/26/2022 04:48:26 - INFO - codeparrot_training - Step 39572: {'lr': 5.6002158183777936e-05, 'samples': 20261376, 'steps': 39572, 'loss/train': 2.3033339977264404} 02/26/2022 04:48:32 - INFO - codeparrot_training - Step 39573: {'lr': 5.5991838082457253e-05, 'samples': 20261888, 'steps': 39573, 'loss/train': 0.6138990521430969} 02/26/2022 04:48:35 - INFO - codeparrot_training - Step 39574: {'lr': 5.598151881220584e-05, 'samples': 20262400, 'steps': 39574, 'loss/train': 1.1860277652740479} 02/26/2022 04:48:41 - INFO - codeparrot_training - Step 39575: {'lr': 5.5971200373068076e-05, 'samples': 20262912, 'steps': 39575, 'loss/train': 1.53574800491333} 02/26/2022 04:48:44 - INFO - codeparrot_training - Step 39576: {'lr': 5.5960882765087984e-05, 'samples': 20263424, 'steps': 39576, 'loss/train': 1.9003180265426636} 02/26/2022 04:48:50 - INFO - codeparrot_training - Step 39577: {'lr': 5.595056598831002e-05, 'samples': 20263936, 'steps': 39577, 'loss/train': 0.5860280990600586} 02/26/2022 04:48:53 - INFO - codeparrot_training - Step 39578: {'lr': 5.594025004277806e-05, 'samples': 20264448, 'steps': 39578, 'loss/train': 2.6915645599365234} 02/26/2022 04:48:59 - INFO - codeparrot_training - Step 39579: {'lr': 5.592993492853654e-05, 'samples': 20264960, 'steps': 39579, 'loss/train': 2.1316583156585693} 02/26/2022 04:49:03 - INFO - codeparrot_training - Step 39580: {'lr': 5.5919620645629445e-05, 'samples': 20265472, 'steps': 39580, 'loss/train': 5.451935768127441} 02/26/2022 04:49:09 - INFO - codeparrot_training - Step 39581: {'lr': 5.5909307194101255e-05, 'samples': 20265984, 'steps': 39581, 'loss/train': 1.6974302530288696} 02/26/2022 04:49:13 - INFO - codeparrot_training - Step 39582: {'lr': 5.589899457399578e-05, 'samples': 20266496, 'steps': 39582, 'loss/train': 3.507108449935913} 02/26/2022 04:49:18 - INFO - codeparrot_training - Step 39583: {'lr': 5.588868278535747e-05, 'samples': 20267008, 'steps': 39583, 'loss/train': 1.8848713636398315} 02/26/2022 04:49:22 - INFO - codeparrot_training - Step 39584: {'lr': 5.587837182823033e-05, 'samples': 20267520, 'steps': 39584, 'loss/train': 1.5594751834869385} 02/26/2022 04:49:27 - INFO - codeparrot_training - Step 39585: {'lr': 5.5868061702658676e-05, 'samples': 20268032, 'steps': 39585, 'loss/train': 2.2697362899780273} 02/26/2022 04:49:31 - INFO - codeparrot_training - Step 39586: {'lr': 5.5857752408686595e-05, 'samples': 20268544, 'steps': 39586, 'loss/train': 1.8169310092926025} 02/26/2022 04:49:36 - INFO - codeparrot_training - Step 39587: {'lr': 5.584744394635827e-05, 'samples': 20269056, 'steps': 39587, 'loss/train': 0.729655921459198} 02/26/2022 04:49:40 - INFO - codeparrot_training - Step 39588: {'lr': 5.583713631571777e-05, 'samples': 20269568, 'steps': 39588, 'loss/train': 1.6893389225006104} 02/26/2022 04:49:45 - INFO - codeparrot_training - Step 39589: {'lr': 5.58268295168094e-05, 'samples': 20270080, 'steps': 39589, 'loss/train': 1.3779979944229126} 02/26/2022 04:49:49 - INFO - codeparrot_training - Step 39590: {'lr': 5.5816523549677226e-05, 'samples': 20270592, 'steps': 39590, 'loss/train': 1.834897518157959} 02/26/2022 04:49:55 - INFO - codeparrot_training - Step 39591: {'lr': 5.58062184143654e-05, 'samples': 20271104, 'steps': 39591, 'loss/train': 1.8546406030654907} 02/26/2022 04:49:59 - INFO - codeparrot_training - Step 39592: {'lr': 5.5795914110918e-05, 'samples': 20271616, 'steps': 39592, 'loss/train': 1.0686100721359253} 02/26/2022 04:50:04 - INFO - codeparrot_training - Step 39593: {'lr': 5.5785610639379314e-05, 'samples': 20272128, 'steps': 39593, 'loss/train': 2.0364582538604736} 02/26/2022 04:50:08 - INFO - codeparrot_training - Step 39594: {'lr': 5.5775307999793415e-05, 'samples': 20272640, 'steps': 39594, 'loss/train': 2.130596399307251} 02/26/2022 04:50:13 - INFO - codeparrot_training - Step 39595: {'lr': 5.5765006192204395e-05, 'samples': 20273152, 'steps': 39595, 'loss/train': 0.9701781272888184} 02/26/2022 04:50:17 - INFO - codeparrot_training - Step 39596: {'lr': 5.5754705216656375e-05, 'samples': 20273664, 'steps': 39596, 'loss/train': 1.6545461416244507} 02/26/2022 04:50:22 - INFO - codeparrot_training - Step 39597: {'lr': 5.574440507319356e-05, 'samples': 20274176, 'steps': 39597, 'loss/train': 1.3062081336975098} 02/26/2022 04:50:26 - INFO - codeparrot_training - Step 39598: {'lr': 5.573410576185997e-05, 'samples': 20274688, 'steps': 39598, 'loss/train': 2.392810106277466} 02/26/2022 04:50:31 - INFO - codeparrot_training - Step 39599: {'lr': 5.572380728269993e-05, 'samples': 20275200, 'steps': 39599, 'loss/train': 7.505510330200195} 02/26/2022 04:50:35 - INFO - codeparrot_training - Step 39600: {'lr': 5.5713509635757274e-05, 'samples': 20275712, 'steps': 39600, 'loss/train': 2.972777843475342} 02/26/2022 04:50:41 - INFO - codeparrot_training - Step 39601: {'lr': 5.570321282107632e-05, 'samples': 20276224, 'steps': 39601, 'loss/train': 2.9780702590942383} 02/26/2022 04:50:45 - INFO - codeparrot_training - Step 39602: {'lr': 5.569291683870109e-05, 'samples': 20276736, 'steps': 39602, 'loss/train': 2.163320541381836} 02/26/2022 04:50:50 - INFO - codeparrot_training - Step 39603: {'lr': 5.568262168867572e-05, 'samples': 20277248, 'steps': 39603, 'loss/train': 2.3610901832580566} 02/26/2022 04:50:54 - INFO - codeparrot_training - Step 39604: {'lr': 5.567232737104422e-05, 'samples': 20277760, 'steps': 39604, 'loss/train': 1.554967999458313} 02/26/2022 04:50:59 - INFO - codeparrot_training - Step 39605: {'lr': 5.566203388585081e-05, 'samples': 20278272, 'steps': 39605, 'loss/train': 1.5985158681869507} 02/26/2022 04:51:03 - INFO - codeparrot_training - Step 39606: {'lr': 5.565174123313957e-05, 'samples': 20278784, 'steps': 39606, 'loss/train': 2.751771926879883} 02/26/2022 04:51:08 - INFO - codeparrot_training - Step 39607: {'lr': 5.564144941295454e-05, 'samples': 20279296, 'steps': 39607, 'loss/train': 1.0058128833770752} 02/26/2022 04:51:12 - INFO - codeparrot_training - Step 39608: {'lr': 5.563115842533978e-05, 'samples': 20279808, 'steps': 39608, 'loss/train': 2.8748834133148193} 02/26/2022 04:51:17 - INFO - codeparrot_training - Step 39609: {'lr': 5.562086827033946e-05, 'samples': 20280320, 'steps': 39609, 'loss/train': 1.484361171722412} 02/26/2022 04:51:21 - INFO - codeparrot_training - Step 39610: {'lr': 5.561057894799762e-05, 'samples': 20280832, 'steps': 39610, 'loss/train': 1.8666799068450928} 02/26/2022 04:51:26 - INFO - codeparrot_training - Step 39611: {'lr': 5.5600290458358296e-05, 'samples': 20281344, 'steps': 39611, 'loss/train': 1.147526741027832} 02/26/2022 04:51:30 - INFO - codeparrot_training - Step 39612: {'lr': 5.5590002801465645e-05, 'samples': 20281856, 'steps': 39612, 'loss/train': 0.5306660532951355} 02/26/2022 04:51:35 - INFO - codeparrot_training - Step 39613: {'lr': 5.557971597736369e-05, 'samples': 20282368, 'steps': 39613, 'loss/train': 2.113041400909424} 02/26/2022 04:51:38 - INFO - codeparrot_training - Step 39614: {'lr': 5.556942998609649e-05, 'samples': 20282880, 'steps': 39614, 'loss/train': 1.471741795539856} 02/26/2022 04:51:44 - INFO - codeparrot_training - Step 39615: {'lr': 5.555914482770805e-05, 'samples': 20283392, 'steps': 39615, 'loss/train': 1.3288798332214355} 02/26/2022 04:51:47 - INFO - codeparrot_training - Step 39616: {'lr': 5.554886050224256e-05, 'samples': 20283904, 'steps': 39616, 'loss/train': 0.5740442872047424} 02/26/2022 04:51:55 - INFO - codeparrot_training - Step 39617: {'lr': 5.553857700974402e-05, 'samples': 20284416, 'steps': 39617, 'loss/train': 0.944951593875885} 02/26/2022 04:51:58 - INFO - codeparrot_training - Step 39618: {'lr': 5.5528294350256443e-05, 'samples': 20284928, 'steps': 39618, 'loss/train': 2.2058258056640625} 02/26/2022 04:52:04 - INFO - codeparrot_training - Step 39619: {'lr': 5.551801252382385e-05, 'samples': 20285440, 'steps': 39619, 'loss/train': 1.2330358028411865} 02/26/2022 04:52:07 - INFO - codeparrot_training - Step 39620: {'lr': 5.550773153049046e-05, 'samples': 20285952, 'steps': 39620, 'loss/train': 0.3771727681159973} 02/26/2022 04:52:13 - INFO - codeparrot_training - Step 39621: {'lr': 5.549745137030007e-05, 'samples': 20286464, 'steps': 39621, 'loss/train': 1.0573660135269165} 02/26/2022 04:52:16 - INFO - codeparrot_training - Step 39622: {'lr': 5.54871720432969e-05, 'samples': 20286976, 'steps': 39622, 'loss/train': 1.3388012647628784} 02/26/2022 04:52:22 - INFO - codeparrot_training - Step 39623: {'lr': 5.547689354952484e-05, 'samples': 20287488, 'steps': 39623, 'loss/train': 0.615705668926239} 02/26/2022 04:52:25 - INFO - codeparrot_training - Step 39624: {'lr': 5.546661588902807e-05, 'samples': 20288000, 'steps': 39624, 'loss/train': 0.7831544280052185} 02/26/2022 04:52:31 - INFO - codeparrot_training - Step 39625: {'lr': 5.5456339061850576e-05, 'samples': 20288512, 'steps': 39625, 'loss/train': 1.7920844554901123} 02/26/2022 04:52:34 - INFO - codeparrot_training - Step 39626: {'lr': 5.544606306803632e-05, 'samples': 20289024, 'steps': 39626, 'loss/train': 2.372608184814453} 02/26/2022 04:52:41 - INFO - codeparrot_training - Step 39627: {'lr': 5.543578790762929e-05, 'samples': 20289536, 'steps': 39627, 'loss/train': 0.5385619401931763} 02/26/2022 04:52:45 - INFO - codeparrot_training - Step 39628: {'lr': 5.5425513580673634e-05, 'samples': 20290048, 'steps': 39628, 'loss/train': 2.880880832672119} 02/26/2022 04:52:50 - INFO - codeparrot_training - Step 39629: {'lr': 5.54152400872133e-05, 'samples': 20290560, 'steps': 39629, 'loss/train': 2.581942319869995} 02/26/2022 04:52:56 - INFO - codeparrot_training - Step 39630: {'lr': 5.540496742729229e-05, 'samples': 20291072, 'steps': 39630, 'loss/train': 1.8050655126571655} 02/26/2022 04:52:59 - INFO - codeparrot_training - Step 39631: {'lr': 5.53946956009545e-05, 'samples': 20291584, 'steps': 39631, 'loss/train': 1.3622560501098633} 02/26/2022 04:53:05 - INFO - codeparrot_training - Step 39632: {'lr': 5.5384424608244165e-05, 'samples': 20292096, 'steps': 39632, 'loss/train': 1.0495922565460205} 02/26/2022 04:53:08 - INFO - codeparrot_training - Step 39633: {'lr': 5.537415444920513e-05, 'samples': 20292608, 'steps': 39633, 'loss/train': 1.2890435457229614} 02/26/2022 04:53:14 - INFO - codeparrot_training - Step 39634: {'lr': 5.536388512388141e-05, 'samples': 20293120, 'steps': 39634, 'loss/train': 1.8021255731582642} 02/26/2022 04:53:17 - INFO - codeparrot_training - Step 39635: {'lr': 5.535361663231694e-05, 'samples': 20293632, 'steps': 39635, 'loss/train': 2.2562716007232666} 02/26/2022 04:53:21 - INFO - codeparrot_training - Step 39636: {'lr': 5.534334897455584e-05, 'samples': 20294144, 'steps': 39636, 'loss/train': 1.6376821994781494} 02/26/2022 04:53:27 - INFO - codeparrot_training - Step 39637: {'lr': 5.5333082150641975e-05, 'samples': 20294656, 'steps': 39637, 'loss/train': 1.8238276243209839} 02/26/2022 04:53:33 - INFO - codeparrot_training - Step 39638: {'lr': 5.5322816160619485e-05, 'samples': 20295168, 'steps': 39638, 'loss/train': 0.7472890615463257} 02/26/2022 04:53:37 - INFO - codeparrot_training - Step 39639: {'lr': 5.5312551004532116e-05, 'samples': 20295680, 'steps': 39639, 'loss/train': 1.3971142768859863} 02/26/2022 04:53:42 - INFO - codeparrot_training - Step 39640: {'lr': 5.5302286682424016e-05, 'samples': 20296192, 'steps': 39640, 'loss/train': 3.2534430027008057} 02/26/2022 04:53:45 - INFO - codeparrot_training - Step 39641: {'lr': 5.529202319433901e-05, 'samples': 20296704, 'steps': 39641, 'loss/train': 0.8869723677635193} 02/26/2022 04:53:51 - INFO - codeparrot_training - Step 39642: {'lr': 5.52817605403213e-05, 'samples': 20297216, 'steps': 39642, 'loss/train': 1.8182017803192139} 02/26/2022 04:53:55 - INFO - codeparrot_training - Step 39643: {'lr': 5.527149872041456e-05, 'samples': 20297728, 'steps': 39643, 'loss/train': 1.95842707157135} 02/26/2022 04:54:00 - INFO - codeparrot_training - Step 39644: {'lr': 5.526123773466296e-05, 'samples': 20298240, 'steps': 39644, 'loss/train': 2.507134437561035} 02/26/2022 04:54:04 - INFO - codeparrot_training - Step 39645: {'lr': 5.5250977583110316e-05, 'samples': 20298752, 'steps': 39645, 'loss/train': 0.7056419253349304} 02/26/2022 04:54:10 - INFO - codeparrot_training - Step 39646: {'lr': 5.524071826580071e-05, 'samples': 20299264, 'steps': 39646, 'loss/train': 1.2730189561843872} 02/26/2022 04:54:14 - INFO - codeparrot_training - Step 39647: {'lr': 5.523045978277802e-05, 'samples': 20299776, 'steps': 39647, 'loss/train': 1.396292805671692} 02/26/2022 04:54:19 - INFO - codeparrot_training - Step 39648: {'lr': 5.5220202134086185e-05, 'samples': 20300288, 'steps': 39648, 'loss/train': 2.679823637008667} 02/26/2022 04:54:23 - INFO - codeparrot_training - Step 39649: {'lr': 5.5209945319769166e-05, 'samples': 20300800, 'steps': 39649, 'loss/train': 0.8680575489997864} 02/26/2022 04:54:29 - INFO - codeparrot_training - Step 39650: {'lr': 5.519968933987082e-05, 'samples': 20301312, 'steps': 39650, 'loss/train': 1.508697509765625} 02/26/2022 04:54:32 - INFO - codeparrot_training - Step 39651: {'lr': 5.5189434194435204e-05, 'samples': 20301824, 'steps': 39651, 'loss/train': 1.804937720298767} 02/26/2022 04:54:36 - INFO - codeparrot_training - Step 39652: {'lr': 5.517917988350621e-05, 'samples': 20302336, 'steps': 39652, 'loss/train': 2.26228666305542} 02/26/2022 04:54:41 - INFO - codeparrot_training - Step 39653: {'lr': 5.516892640712776e-05, 'samples': 20302848, 'steps': 39653, 'loss/train': 1.994885802268982} 02/26/2022 04:54:45 - INFO - codeparrot_training - Step 39654: {'lr': 5.515867376534367e-05, 'samples': 20303360, 'steps': 39654, 'loss/train': 1.053203821182251} 02/26/2022 04:54:50 - INFO - codeparrot_training - Step 39655: {'lr': 5.5148421958198025e-05, 'samples': 20303872, 'steps': 39655, 'loss/train': 0.5563961863517761} 02/26/2022 04:54:54 - INFO - codeparrot_training - Step 39656: {'lr': 5.513817098573465e-05, 'samples': 20304384, 'steps': 39656, 'loss/train': 1.6732758283615112} 02/26/2022 04:54:59 - INFO - codeparrot_training - Step 39657: {'lr': 5.512792084799751e-05, 'samples': 20304896, 'steps': 39657, 'loss/train': 1.5255635976791382} 02/26/2022 04:55:05 - INFO - codeparrot_training - Step 39658: {'lr': 5.511767154503039e-05, 'samples': 20305408, 'steps': 39658, 'loss/train': 1.8469328880310059} 02/26/2022 04:55:08 - INFO - codeparrot_training - Step 39659: {'lr': 5.510742307687736e-05, 'samples': 20305920, 'steps': 39659, 'loss/train': 1.7309544086456299} 02/26/2022 04:55:14 - INFO - codeparrot_training - Step 39660: {'lr': 5.509717544358223e-05, 'samples': 20306432, 'steps': 39660, 'loss/train': 2.007723808288574} 02/26/2022 04:55:17 - INFO - codeparrot_training - Step 39661: {'lr': 5.508692864518891e-05, 'samples': 20306944, 'steps': 39661, 'loss/train': 1.743770718574524} 02/26/2022 04:55:21 - INFO - codeparrot_training - Step 39662: {'lr': 5.507668268174126e-05, 'samples': 20307456, 'steps': 39662, 'loss/train': 2.5784928798675537} 02/26/2022 04:55:28 - INFO - codeparrot_training - Step 39663: {'lr': 5.506643755328325e-05, 'samples': 20307968, 'steps': 39663, 'loss/train': 0.10176370292901993} 02/26/2022 04:55:33 - INFO - codeparrot_training - Step 39664: {'lr': 5.505619325985872e-05, 'samples': 20308480, 'steps': 39664, 'loss/train': 1.9804339408874512} 02/26/2022 04:55:37 - INFO - codeparrot_training - Step 39665: {'lr': 5.504594980151156e-05, 'samples': 20308992, 'steps': 39665, 'loss/train': 1.0873655080795288} 02/26/2022 04:55:40 - INFO - codeparrot_training - Step 39666: {'lr': 5.50357071782856e-05, 'samples': 20309504, 'steps': 39666, 'loss/train': 1.1225903034210205} 02/26/2022 04:55:46 - INFO - codeparrot_training - Step 39667: {'lr': 5.5025465390224794e-05, 'samples': 20310016, 'steps': 39667, 'loss/train': 1.82794189453125} 02/26/2022 04:55:49 - INFO - codeparrot_training - Step 39668: {'lr': 5.5015224437373005e-05, 'samples': 20310528, 'steps': 39668, 'loss/train': 4.054574966430664} 02/26/2022 04:55:55 - INFO - codeparrot_training - Step 39669: {'lr': 5.5004984319774076e-05, 'samples': 20311040, 'steps': 39669, 'loss/train': 1.000671625137329} 02/26/2022 04:55:58 - INFO - codeparrot_training - Step 39670: {'lr': 5.4994745037471816e-05, 'samples': 20311552, 'steps': 39670, 'loss/train': 1.4288913011550903} 02/26/2022 04:56:04 - INFO - codeparrot_training - Step 39671: {'lr': 5.498450659051018e-05, 'samples': 20312064, 'steps': 39671, 'loss/train': 1.8031368255615234} 02/26/2022 04:56:07 - INFO - codeparrot_training - Step 39672: {'lr': 5.497426897893296e-05, 'samples': 20312576, 'steps': 39672, 'loss/train': 2.184058904647827} 02/26/2022 04:56:14 - INFO - codeparrot_training - Step 39673: {'lr': 5.496403220278418e-05, 'samples': 20313088, 'steps': 39673, 'loss/train': 2.5792062282562256} 02/26/2022 04:56:20 - INFO - codeparrot_training - Step 39674: {'lr': 5.495379626210742e-05, 'samples': 20313600, 'steps': 39674, 'loss/train': 1.1711337566375732} 02/26/2022 04:56:23 - INFO - codeparrot_training - Step 39675: {'lr': 5.494356115694671e-05, 'samples': 20314112, 'steps': 39675, 'loss/train': 1.5987637042999268} 02/26/2022 04:56:29 - INFO - codeparrot_training - Step 39676: {'lr': 5.493332688734579e-05, 'samples': 20314624, 'steps': 39676, 'loss/train': 0.1176011860370636} 02/26/2022 04:56:33 - INFO - codeparrot_training - Step 39677: {'lr': 5.49230934533487e-05, 'samples': 20315136, 'steps': 39677, 'loss/train': 1.889054298400879} 02/26/2022 04:56:38 - INFO - codeparrot_training - Step 39678: {'lr': 5.4912860854999e-05, 'samples': 20315648, 'steps': 39678, 'loss/train': 2.9106810092926025} 02/26/2022 04:56:42 - INFO - codeparrot_training - Step 39679: {'lr': 5.49026290923407e-05, 'samples': 20316160, 'steps': 39679, 'loss/train': 2.006866216659546} 02/26/2022 04:56:47 - INFO - codeparrot_training - Step 39680: {'lr': 5.489239816541755e-05, 'samples': 20316672, 'steps': 39680, 'loss/train': 1.660393238067627} 02/26/2022 04:56:51 - INFO - codeparrot_training - Step 39681: {'lr': 5.4882168074273514e-05, 'samples': 20317184, 'steps': 39681, 'loss/train': 1.5633084774017334} 02/26/2022 04:56:54 - INFO - codeparrot_training - Step 39682: {'lr': 5.4871938818952195e-05, 'samples': 20317696, 'steps': 39682, 'loss/train': 1.2776139974594116} 02/26/2022 04:57:01 - INFO - codeparrot_training - Step 39683: {'lr': 5.486171039949761e-05, 'samples': 20318208, 'steps': 39683, 'loss/train': 2.2751355171203613} 02/26/2022 04:57:04 - INFO - codeparrot_training - Step 39684: {'lr': 5.4851482815953424e-05, 'samples': 20318720, 'steps': 39684, 'loss/train': 1.9149181842803955} 02/26/2022 04:57:10 - INFO - codeparrot_training - Step 39685: {'lr': 5.484125606836357e-05, 'samples': 20319232, 'steps': 39685, 'loss/train': 1.3627824783325195} 02/26/2022 04:57:13 - INFO - codeparrot_training - Step 39686: {'lr': 5.483103015677182e-05, 'samples': 20319744, 'steps': 39686, 'loss/train': 0.23077812790870667} 02/26/2022 04:57:19 - INFO - codeparrot_training - Step 39687: {'lr': 5.4820805081221944e-05, 'samples': 20320256, 'steps': 39687, 'loss/train': 2.1500403881073} 02/26/2022 04:57:24 - INFO - codeparrot_training - Step 39688: {'lr': 5.481058084175772e-05, 'samples': 20320768, 'steps': 39688, 'loss/train': 1.5449376106262207} 02/26/2022 04:57:28 - INFO - codeparrot_training - Step 39689: {'lr': 5.480035743842304e-05, 'samples': 20321280, 'steps': 39689, 'loss/train': 1.32004714012146} 02/26/2022 04:57:33 - INFO - codeparrot_training - Step 39690: {'lr': 5.4790134871261636e-05, 'samples': 20321792, 'steps': 39690, 'loss/train': 1.6051506996154785} 02/26/2022 04:57:37 - INFO - codeparrot_training - Step 39691: {'lr': 5.4779913140317326e-05, 'samples': 20322304, 'steps': 39691, 'loss/train': 0.8411818742752075} 02/26/2022 04:57:43 - INFO - codeparrot_training - Step 39692: {'lr': 5.476969224563383e-05, 'samples': 20322816, 'steps': 39692, 'loss/train': 1.7682324647903442} 02/26/2022 04:57:47 - INFO - codeparrot_training - Step 39693: {'lr': 5.4759472187255024e-05, 'samples': 20323328, 'steps': 39693, 'loss/train': 0.4707164466381073} 02/26/2022 04:57:52 - INFO - codeparrot_training - Step 39694: {'lr': 5.474925296522462e-05, 'samples': 20323840, 'steps': 39694, 'loss/train': 2.2049062252044678} 02/26/2022 04:57:56 - INFO - codeparrot_training - Step 39695: {'lr': 5.473903457958645e-05, 'samples': 20324352, 'steps': 39695, 'loss/train': 0.8677700161933899} 02/26/2022 04:58:01 - INFO - codeparrot_training - Step 39696: {'lr': 5.472881703038418e-05, 'samples': 20324864, 'steps': 39696, 'loss/train': 1.4744956493377686} 02/26/2022 04:58:05 - INFO - codeparrot_training - Step 39697: {'lr': 5.471860031766171e-05, 'samples': 20325376, 'steps': 39697, 'loss/train': 1.0862455368041992} 02/26/2022 04:58:08 - INFO - codeparrot_training - Step 39698: {'lr': 5.470838444146275e-05, 'samples': 20325888, 'steps': 39698, 'loss/train': 0.1650933474302292} 02/26/2022 04:58:14 - INFO - codeparrot_training - Step 39699: {'lr': 5.469816940183109e-05, 'samples': 20326400, 'steps': 39699, 'loss/train': 1.0252324342727661} 02/26/2022 04:58:19 - INFO - codeparrot_training - Step 39700: {'lr': 5.468795519881042e-05, 'samples': 20326912, 'steps': 39700, 'loss/train': 2.9577829837799072} 02/26/2022 04:58:23 - INFO - codeparrot_training - Step 39701: {'lr': 5.467774183244448e-05, 'samples': 20327424, 'steps': 39701, 'loss/train': 0.9262147545814514} 02/26/2022 04:58:30 - INFO - codeparrot_training - Step 39702: {'lr': 5.466752930277713e-05, 'samples': 20327936, 'steps': 39702, 'loss/train': 1.2280688285827637} 02/26/2022 04:58:34 - INFO - codeparrot_training - Step 39703: {'lr': 5.465731760985207e-05, 'samples': 20328448, 'steps': 39703, 'loss/train': 1.3474022150039673} 02/26/2022 04:58:39 - INFO - codeparrot_training - Step 39704: {'lr': 5.4647106753713014e-05, 'samples': 20328960, 'steps': 39704, 'loss/train': 0.9308189153671265} 02/26/2022 04:58:43 - INFO - codeparrot_training - Step 39705: {'lr': 5.463689673440367e-05, 'samples': 20329472, 'steps': 39705, 'loss/train': 1.720345139503479} 02/26/2022 04:58:48 - INFO - codeparrot_training - Step 39706: {'lr': 5.4626687551967896e-05, 'samples': 20329984, 'steps': 39706, 'loss/train': 1.885561227798462} 02/26/2022 04:58:52 - INFO - codeparrot_training - Step 39707: {'lr': 5.461647920644927e-05, 'samples': 20330496, 'steps': 39707, 'loss/train': 2.401057481765747} 02/26/2022 04:58:57 - INFO - codeparrot_training - Step 39708: {'lr': 5.4606271697891757e-05, 'samples': 20331008, 'steps': 39708, 'loss/train': 0.7946843504905701} 02/26/2022 04:59:01 - INFO - codeparrot_training - Step 39709: {'lr': 5.459606502633879e-05, 'samples': 20331520, 'steps': 39709, 'loss/train': 1.0909266471862793} 02/26/2022 04:59:06 - INFO - codeparrot_training - Step 39710: {'lr': 5.4585859191834284e-05, 'samples': 20332032, 'steps': 39710, 'loss/train': 0.4135775864124298} 02/26/2022 04:59:10 - INFO - codeparrot_training - Step 39711: {'lr': 5.457565419442187e-05, 'samples': 20332544, 'steps': 39711, 'loss/train': 2.2503304481506348} 02/26/2022 04:59:15 - INFO - codeparrot_training - Step 39712: {'lr': 5.4565450034145424e-05, 'samples': 20333056, 'steps': 39712, 'loss/train': 0.9864827990531921} 02/26/2022 04:59:18 - INFO - codeparrot_training - Step 39713: {'lr': 5.455524671104839e-05, 'samples': 20333568, 'steps': 39713, 'loss/train': 0.39865031838417053} 02/26/2022 04:59:26 - INFO - codeparrot_training - Step 39714: {'lr': 5.454504422517467e-05, 'samples': 20334080, 'steps': 39714, 'loss/train': 2.3310444355010986} 02/26/2022 04:59:29 - INFO - codeparrot_training - Step 39715: {'lr': 5.453484257656788e-05, 'samples': 20334592, 'steps': 39715, 'loss/train': 1.0801059007644653} 02/26/2022 04:59:35 - INFO - codeparrot_training - Step 39716: {'lr': 5.452464176527189e-05, 'samples': 20335104, 'steps': 39716, 'loss/train': 1.2765918970108032} 02/26/2022 04:59:38 - INFO - codeparrot_training - Step 39717: {'lr': 5.451444179133011e-05, 'samples': 20335616, 'steps': 39717, 'loss/train': 2.189959764480591} 02/26/2022 04:59:44 - INFO - codeparrot_training - Step 39718: {'lr': 5.4504242654786454e-05, 'samples': 20336128, 'steps': 39718, 'loss/train': 0.14243502914905548} 02/26/2022 04:59:47 - INFO - codeparrot_training - Step 39719: {'lr': 5.44940443556845e-05, 'samples': 20336640, 'steps': 39719, 'loss/train': 1.277060627937317} 02/26/2022 04:59:53 - INFO - codeparrot_training - Step 39720: {'lr': 5.4483846894068044e-05, 'samples': 20337152, 'steps': 39720, 'loss/train': 0.9580551385879517} 02/26/2022 04:59:56 - INFO - codeparrot_training - Step 39721: {'lr': 5.447365026998069e-05, 'samples': 20337664, 'steps': 39721, 'loss/train': 2.065368413925171} 02/26/2022 05:00:02 - INFO - codeparrot_training - Step 39722: {'lr': 5.4463454483466124e-05, 'samples': 20338176, 'steps': 39722, 'loss/train': 1.8422590494155884} 02/26/2022 05:00:05 - INFO - codeparrot_training - Step 39723: {'lr': 5.445325953456798e-05, 'samples': 20338688, 'steps': 39723, 'loss/train': 1.1569839715957642} 02/26/2022 05:00:13 - INFO - codeparrot_training - Step 39724: {'lr': 5.444306542333005e-05, 'samples': 20339200, 'steps': 39724, 'loss/train': 2.0523934364318848} 02/26/2022 05:00:16 - INFO - codeparrot_training - Step 39725: {'lr': 5.443287214979592e-05, 'samples': 20339712, 'steps': 39725, 'loss/train': 1.6092201471328735} 02/26/2022 05:00:22 - INFO - codeparrot_training - Step 39726: {'lr': 5.442267971400927e-05, 'samples': 20340224, 'steps': 39726, 'loss/train': 2.3666648864746094} 02/26/2022 05:00:25 - INFO - codeparrot_training - Step 39727: {'lr': 5.441248811601368e-05, 'samples': 20340736, 'steps': 39727, 'loss/train': 2.1323652267456055} 02/26/2022 05:00:30 - INFO - codeparrot_training - Step 39728: {'lr': 5.440229735585297e-05, 'samples': 20341248, 'steps': 39728, 'loss/train': 2.1372671127319336} 02/26/2022 05:00:34 - INFO - codeparrot_training - Step 39729: {'lr': 5.4392107433570704e-05, 'samples': 20341760, 'steps': 39729, 'loss/train': 1.4354342222213745} 02/26/2022 05:00:40 - INFO - codeparrot_training - Step 39730: {'lr': 5.4381918349210526e-05, 'samples': 20342272, 'steps': 39730, 'loss/train': 1.911924123764038} 02/26/2022 05:00:43 - INFO - codeparrot_training - Step 39731: {'lr': 5.437173010281604e-05, 'samples': 20342784, 'steps': 39731, 'loss/train': 2.8752458095550537} 02/26/2022 05:00:49 - INFO - codeparrot_training - Step 39732: {'lr': 5.4361542694430996e-05, 'samples': 20343296, 'steps': 39732, 'loss/train': 3.1536624431610107} 02/26/2022 05:00:52 - INFO - codeparrot_training - Step 39733: {'lr': 5.4351356124099e-05, 'samples': 20343808, 'steps': 39733, 'loss/train': 2.3629064559936523} 02/26/2022 05:00:59 - INFO - codeparrot_training - Step 39734: {'lr': 5.434117039186368e-05, 'samples': 20344320, 'steps': 39734, 'loss/train': 0.9162015914916992} 02/26/2022 05:01:03 - INFO - codeparrot_training - Step 39735: {'lr': 5.4330985497768545e-05, 'samples': 20344832, 'steps': 39735, 'loss/train': 1.1805990934371948} 02/26/2022 05:01:08 - INFO - codeparrot_training - Step 39736: {'lr': 5.432080144185744e-05, 'samples': 20345344, 'steps': 39736, 'loss/train': 1.0604205131530762} 02/26/2022 05:01:12 - INFO - codeparrot_training - Step 39737: {'lr': 5.431061822417388e-05, 'samples': 20345856, 'steps': 39737, 'loss/train': 2.219500780105591} 02/26/2022 05:01:17 - INFO - codeparrot_training - Step 39738: {'lr': 5.430043584476149e-05, 'samples': 20346368, 'steps': 39738, 'loss/train': 1.6543880701065063} 02/26/2022 05:01:21 - INFO - codeparrot_training - Step 39739: {'lr': 5.42902543036638e-05, 'samples': 20346880, 'steps': 39739, 'loss/train': 0.33507290482521057} 02/26/2022 05:01:27 - INFO - codeparrot_training - Step 39740: {'lr': 5.4280073600924626e-05, 'samples': 20347392, 'steps': 39740, 'loss/train': 8.76893424987793} 02/26/2022 05:01:30 - INFO - codeparrot_training - Step 39741: {'lr': 5.426989373658744e-05, 'samples': 20347904, 'steps': 39741, 'loss/train': 3.0817201137542725} 02/26/2022 05:01:36 - INFO - codeparrot_training - Step 39742: {'lr': 5.425971471069582e-05, 'samples': 20348416, 'steps': 39742, 'loss/train': 1.2712726593017578} 02/26/2022 05:01:39 - INFO - codeparrot_training - Step 39743: {'lr': 5.424953652329348e-05, 'samples': 20348928, 'steps': 39743, 'loss/train': 2.2463932037353516} 02/26/2022 05:01:46 - INFO - codeparrot_training - Step 39744: {'lr': 5.423935917442399e-05, 'samples': 20349440, 'steps': 39744, 'loss/train': 2.4513189792633057} 02/26/2022 05:01:50 - INFO - codeparrot_training - Step 39745: {'lr': 5.4229182664130904e-05, 'samples': 20349952, 'steps': 39745, 'loss/train': 1.9745986461639404} 02/26/2022 05:01:55 - INFO - codeparrot_training - Step 39746: {'lr': 5.42190069924578e-05, 'samples': 20350464, 'steps': 39746, 'loss/train': 1.25392746925354} 02/26/2022 05:01:59 - INFO - codeparrot_training - Step 39747: {'lr': 5.420883215944841e-05, 'samples': 20350976, 'steps': 39747, 'loss/train': 1.656606912612915} 02/26/2022 05:02:04 - INFO - codeparrot_training - Step 39748: {'lr': 5.419865816514607e-05, 'samples': 20351488, 'steps': 39748, 'loss/train': 2.42645525932312} 02/26/2022 05:02:08 - INFO - codeparrot_training - Step 39749: {'lr': 5.418848500959458e-05, 'samples': 20352000, 'steps': 39749, 'loss/train': 2.0131170749664307} 02/26/2022 05:02:13 - INFO - codeparrot_training - Step 39750: {'lr': 5.4178312692837405e-05, 'samples': 20352512, 'steps': 39750, 'loss/train': 1.3732068538665771} 02/26/2022 05:02:17 - INFO - codeparrot_training - Step 39751: {'lr': 5.416814121491825e-05, 'samples': 20353024, 'steps': 39751, 'loss/train': 1.518707275390625} 02/26/2022 05:02:22 - INFO - codeparrot_training - Step 39752: {'lr': 5.4157970575880486e-05, 'samples': 20353536, 'steps': 39752, 'loss/train': 1.727295160293579} 02/26/2022 05:02:26 - INFO - codeparrot_training - Step 39753: {'lr': 5.414780077576786e-05, 'samples': 20354048, 'steps': 39753, 'loss/train': 1.3572252988815308} 02/26/2022 05:02:31 - INFO - codeparrot_training - Step 39754: {'lr': 5.413763181462378e-05, 'samples': 20354560, 'steps': 39754, 'loss/train': 1.683567762374878} 02/26/2022 05:02:35 - INFO - codeparrot_training - Step 39755: {'lr': 5.412746369249197e-05, 'samples': 20355072, 'steps': 39755, 'loss/train': 1.8298965692520142} 02/26/2022 05:02:40 - INFO - codeparrot_training - Step 39756: {'lr': 5.41172964094159e-05, 'samples': 20355584, 'steps': 39756, 'loss/train': 1.5596152544021606} 02/26/2022 05:02:44 - INFO - codeparrot_training - Step 39757: {'lr': 5.410712996543915e-05, 'samples': 20356096, 'steps': 39757, 'loss/train': 2.074249505996704} 02/26/2022 05:02:49 - INFO - codeparrot_training - Step 39758: {'lr': 5.409696436060516e-05, 'samples': 20356608, 'steps': 39758, 'loss/train': 1.6563329696655273} 02/26/2022 05:02:53 - INFO - codeparrot_training - Step 39759: {'lr': 5.408679959495766e-05, 'samples': 20357120, 'steps': 39759, 'loss/train': 1.7762566804885864} 02/26/2022 05:03:00 - INFO - codeparrot_training - Step 39760: {'lr': 5.407663566854007e-05, 'samples': 20357632, 'steps': 39760, 'loss/train': 1.9043956995010376} 02/26/2022 05:03:03 - INFO - codeparrot_training - Step 39761: {'lr': 5.406647258139599e-05, 'samples': 20358144, 'steps': 39761, 'loss/train': 2.357501983642578} 02/26/2022 05:03:09 - INFO - codeparrot_training - Step 39762: {'lr': 5.405631033356884e-05, 'samples': 20358656, 'steps': 39762, 'loss/train': 0.3489668667316437} 02/26/2022 05:03:12 - INFO - codeparrot_training - Step 39763: {'lr': 5.4046148925102316e-05, 'samples': 20359168, 'steps': 39763, 'loss/train': 1.900965929031372} 02/26/2022 05:03:18 - INFO - codeparrot_training - Step 39764: {'lr': 5.4035988356039874e-05, 'samples': 20359680, 'steps': 39764, 'loss/train': 1.793460488319397} 02/26/2022 05:03:21 - INFO - codeparrot_training - Step 39765: {'lr': 5.402582862642502e-05, 'samples': 20360192, 'steps': 39765, 'loss/train': 1.7831315994262695} 02/26/2022 05:03:27 - INFO - codeparrot_training - Step 39766: {'lr': 5.4015669736301205e-05, 'samples': 20360704, 'steps': 39766, 'loss/train': 0.8137457370758057} 02/26/2022 05:03:30 - INFO - codeparrot_training - Step 39767: {'lr': 5.4005511685712134e-05, 'samples': 20361216, 'steps': 39767, 'loss/train': 1.6040894985198975} 02/26/2022 05:03:36 - INFO - codeparrot_training - Step 39768: {'lr': 5.399535447470111e-05, 'samples': 20361728, 'steps': 39768, 'loss/train': 1.795889973640442} 02/26/2022 05:03:39 - INFO - codeparrot_training - Step 39769: {'lr': 5.3985198103311925e-05, 'samples': 20362240, 'steps': 39769, 'loss/train': 1.8011457920074463} 02/26/2022 05:03:47 - INFO - codeparrot_training - Step 39770: {'lr': 5.397504257158775e-05, 'samples': 20362752, 'steps': 39770, 'loss/train': 2.3464174270629883} 02/26/2022 05:03:50 - INFO - codeparrot_training - Step 39771: {'lr': 5.3964887879572294e-05, 'samples': 20363264, 'steps': 39771, 'loss/train': 0.9428917169570923} 02/26/2022 05:03:55 - INFO - codeparrot_training - Step 39772: {'lr': 5.3954734027308966e-05, 'samples': 20363776, 'steps': 39772, 'loss/train': 2.6956887245178223} 02/26/2022 05:04:01 - INFO - codeparrot_training - Step 39773: {'lr': 5.3944581014841454e-05, 'samples': 20364288, 'steps': 39773, 'loss/train': 2.0494346618652344} 02/26/2022 05:04:04 - INFO - codeparrot_training - Step 39774: {'lr': 5.393442884221294e-05, 'samples': 20364800, 'steps': 39774, 'loss/train': 2.162763833999634} 02/26/2022 05:04:10 - INFO - codeparrot_training - Step 39775: {'lr': 5.3924277509467155e-05, 'samples': 20365312, 'steps': 39775, 'loss/train': 2.193805456161499} 02/26/2022 05:04:13 - INFO - codeparrot_training - Step 39776: {'lr': 5.391412701664744e-05, 'samples': 20365824, 'steps': 39776, 'loss/train': 2.1867339611053467} 02/26/2022 05:04:19 - INFO - codeparrot_training - Step 39777: {'lr': 5.390397736379749e-05, 'samples': 20366336, 'steps': 39777, 'loss/train': 1.8728506565093994} 02/26/2022 05:04:22 - INFO - codeparrot_training - Step 39778: {'lr': 5.389382855096048e-05, 'samples': 20366848, 'steps': 39778, 'loss/train': 0.4670170843601227} 02/26/2022 05:04:28 - INFO - codeparrot_training - Step 39779: {'lr': 5.388368057818011e-05, 'samples': 20367360, 'steps': 39779, 'loss/train': 2.3256709575653076} 02/26/2022 05:04:31 - INFO - codeparrot_training - Step 39780: {'lr': 5.387353344549978e-05, 'samples': 20367872, 'steps': 39780, 'loss/train': 1.5783545970916748} 02/26/2022 05:04:39 - INFO - codeparrot_training - Step 39781: {'lr': 5.3863387152962876e-05, 'samples': 20368384, 'steps': 39781, 'loss/train': 0.5226054787635803} 02/26/2022 05:04:42 - INFO - codeparrot_training - Step 39782: {'lr': 5.385324170061301e-05, 'samples': 20368896, 'steps': 39782, 'loss/train': 2.1844303607940674} 02/26/2022 05:04:48 - INFO - codeparrot_training - Step 39783: {'lr': 5.384309708849358e-05, 'samples': 20369408, 'steps': 39783, 'loss/train': 1.4209096431732178} 02/26/2022 05:04:51 - INFO - codeparrot_training - Step 39784: {'lr': 5.383295331664803e-05, 'samples': 20369920, 'steps': 39784, 'loss/train': 2.0249762535095215} 02/26/2022 05:04:57 - INFO - codeparrot_training - Step 39785: {'lr': 5.382281038511977e-05, 'samples': 20370432, 'steps': 39785, 'loss/train': 1.4781056642532349} 02/26/2022 05:05:00 - INFO - codeparrot_training - Step 39786: {'lr': 5.381266829395234e-05, 'samples': 20370944, 'steps': 39786, 'loss/train': 1.2523202896118164} 02/26/2022 05:05:05 - INFO - codeparrot_training - Step 39787: {'lr': 5.3802527043189166e-05, 'samples': 20371456, 'steps': 39787, 'loss/train': 2.046996593475342} 02/26/2022 05:05:09 - INFO - codeparrot_training - Step 39788: {'lr': 5.379238663287364e-05, 'samples': 20371968, 'steps': 39788, 'loss/train': 1.1402511596679688} 02/26/2022 05:05:14 - INFO - codeparrot_training - Step 39789: {'lr': 5.3782247063049185e-05, 'samples': 20372480, 'steps': 39789, 'loss/train': 2.7387282848358154} 02/26/2022 05:05:18 - INFO - codeparrot_training - Step 39790: {'lr': 5.3772108333759314e-05, 'samples': 20372992, 'steps': 39790, 'loss/train': 1.1888301372528076} 02/26/2022 05:05:25 - INFO - codeparrot_training - Step 39791: {'lr': 5.376197044504744e-05, 'samples': 20373504, 'steps': 39791, 'loss/train': 1.516459584236145} 02/26/2022 05:05:29 - INFO - codeparrot_training - Step 39792: {'lr': 5.375183339695697e-05, 'samples': 20374016, 'steps': 39792, 'loss/train': 2.8965492248535156} 02/26/2022 05:05:34 - INFO - codeparrot_training - Step 39793: {'lr': 5.374169718953126e-05, 'samples': 20374528, 'steps': 39793, 'loss/train': 1.5712018013000488} 02/26/2022 05:05:38 - INFO - codeparrot_training - Step 39794: {'lr': 5.373156182281388e-05, 'samples': 20375040, 'steps': 39794, 'loss/train': 2.4157755374908447} 02/26/2022 05:05:43 - INFO - codeparrot_training - Step 39795: {'lr': 5.372142729684815e-05, 'samples': 20375552, 'steps': 39795, 'loss/train': 1.7076926231384277} 02/26/2022 05:05:47 - INFO - codeparrot_training - Step 39796: {'lr': 5.371129361167751e-05, 'samples': 20376064, 'steps': 39796, 'loss/train': 2.3673014640808105} 02/26/2022 05:05:52 - INFO - codeparrot_training - Step 39797: {'lr': 5.37011607673453e-05, 'samples': 20376576, 'steps': 39797, 'loss/train': 2.1790261268615723} 02/26/2022 05:05:56 - INFO - codeparrot_training - Step 39798: {'lr': 5.369102876389503e-05, 'samples': 20377088, 'steps': 39798, 'loss/train': 2.948676824569702} 02/26/2022 05:06:01 - INFO - codeparrot_training - Step 39799: {'lr': 5.368089760137007e-05, 'samples': 20377600, 'steps': 39799, 'loss/train': 0.550044059753418} 02/26/2022 05:06:05 - INFO - codeparrot_training - Step 39800: {'lr': 5.367076727981382e-05, 'samples': 20378112, 'steps': 39800, 'loss/train': 2.8461899757385254} 02/26/2022 05:06:13 - INFO - codeparrot_training - Step 39801: {'lr': 5.366063779926958e-05, 'samples': 20378624, 'steps': 39801, 'loss/train': 3.3497750759124756} 02/26/2022 05:06:16 - INFO - codeparrot_training - Step 39802: {'lr': 5.3650509159780896e-05, 'samples': 20379136, 'steps': 39802, 'loss/train': 1.2837826013565063} 02/26/2022 05:06:20 - INFO - codeparrot_training - Step 39803: {'lr': 5.3640381361391015e-05, 'samples': 20379648, 'steps': 39803, 'loss/train': 2.0867919921875} 02/26/2022 05:06:25 - INFO - codeparrot_training - Step 39804: {'lr': 5.363025440414352e-05, 'samples': 20380160, 'steps': 39804, 'loss/train': 1.4675344228744507} 02/26/2022 05:06:28 - INFO - codeparrot_training - Step 39805: {'lr': 5.362012828808152e-05, 'samples': 20380672, 'steps': 39805, 'loss/train': 1.41089928150177} 02/26/2022 05:06:34 - INFO - codeparrot_training - Step 39806: {'lr': 5.361000301324859e-05, 'samples': 20381184, 'steps': 39806, 'loss/train': 2.1080501079559326} 02/26/2022 05:06:37 - INFO - codeparrot_training - Step 39807: {'lr': 5.359987857968801e-05, 'samples': 20381696, 'steps': 39807, 'loss/train': 1.9267653226852417} 02/26/2022 05:06:43 - INFO - codeparrot_training - Step 39808: {'lr': 5.358975498744331e-05, 'samples': 20382208, 'steps': 39808, 'loss/train': 2.1931066513061523} 02/26/2022 05:06:46 - INFO - codeparrot_training - Step 39809: {'lr': 5.35796322365576e-05, 'samples': 20382720, 'steps': 39809, 'loss/train': 1.9634737968444824} 02/26/2022 05:06:52 - INFO - codeparrot_training - Step 39810: {'lr': 5.356951032707441e-05, 'samples': 20383232, 'steps': 39810, 'loss/train': 1.4330230951309204} 02/26/2022 05:06:58 - INFO - codeparrot_training - Step 39811: {'lr': 5.3559389259037025e-05, 'samples': 20383744, 'steps': 39811, 'loss/train': 2.0131702423095703} 02/26/2022 05:07:01 - INFO - codeparrot_training - Step 39812: {'lr': 5.3549269032488966e-05, 'samples': 20384256, 'steps': 39812, 'loss/train': 1.5433257818222046} 02/26/2022 05:07:04 - INFO - codeparrot_training - Step 39813: {'lr': 5.353914964747333e-05, 'samples': 20384768, 'steps': 39813, 'loss/train': 1.7816805839538574} 02/26/2022 05:07:10 - INFO - codeparrot_training - Step 39814: {'lr': 5.352903110403368e-05, 'samples': 20385280, 'steps': 39814, 'loss/train': 0.8600420355796814} 02/26/2022 05:07:14 - INFO - codeparrot_training - Step 39815: {'lr': 5.3518913402213184e-05, 'samples': 20385792, 'steps': 39815, 'loss/train': 2.6810121536254883} 02/26/2022 05:07:21 - INFO - codeparrot_training - Step 39816: {'lr': 5.350879654205534e-05, 'samples': 20386304, 'steps': 39816, 'loss/train': 2.2573606967926025} 02/26/2022 05:07:24 - INFO - codeparrot_training - Step 39817: {'lr': 5.3498680523603406e-05, 'samples': 20386816, 'steps': 39817, 'loss/train': 1.8959447145462036} 02/26/2022 05:07:30 - INFO - codeparrot_training - Step 39818: {'lr': 5.348856534690075e-05, 'samples': 20387328, 'steps': 39818, 'loss/train': 1.5887534618377686} 02/26/2022 05:07:36 - INFO - codeparrot_training - Step 39819: {'lr': 5.347845101199061e-05, 'samples': 20387840, 'steps': 39819, 'loss/train': 2.023820400238037} 02/26/2022 05:07:39 - INFO - codeparrot_training - Step 39820: {'lr': 5.3468337518916445e-05, 'samples': 20388352, 'steps': 39820, 'loss/train': 1.3859535455703735} 02/26/2022 05:07:43 - INFO - codeparrot_training - Step 39821: {'lr': 5.345822486772153e-05, 'samples': 20388864, 'steps': 39821, 'loss/train': 0.5249124765396118} 02/26/2022 05:07:48 - INFO - codeparrot_training - Step 39822: {'lr': 5.344811305844915e-05, 'samples': 20389376, 'steps': 39822, 'loss/train': 1.6404755115509033} 02/26/2022 05:07:52 - INFO - codeparrot_training - Step 39823: {'lr': 5.343800209114258e-05, 'samples': 20389888, 'steps': 39823, 'loss/train': 1.5057337284088135} 02/26/2022 05:07:57 - INFO - codeparrot_training - Step 39824: {'lr': 5.342789196584527e-05, 'samples': 20390400, 'steps': 39824, 'loss/train': 0.6853026151657104} 02/26/2022 05:08:01 - INFO - codeparrot_training - Step 39825: {'lr': 5.3417782682600446e-05, 'samples': 20390912, 'steps': 39825, 'loss/train': 1.3691399097442627} 02/26/2022 05:08:06 - INFO - codeparrot_training - Step 39826: {'lr': 5.340767424145143e-05, 'samples': 20391424, 'steps': 39826, 'loss/train': 1.446277141571045} 02/26/2022 05:08:10 - INFO - codeparrot_training - Step 39827: {'lr': 5.3397566642441496e-05, 'samples': 20391936, 'steps': 39827, 'loss/train': 0.6511457562446594} 02/26/2022 05:08:17 - INFO - codeparrot_training - Step 39828: {'lr': 5.3387459885613886e-05, 'samples': 20392448, 'steps': 39828, 'loss/train': 1.8427156209945679} 02/26/2022 05:08:20 - INFO - codeparrot_training - Step 39829: {'lr': 5.337735397101207e-05, 'samples': 20392960, 'steps': 39829, 'loss/train': 1.8660967350006104} 02/26/2022 05:08:26 - INFO - codeparrot_training - Step 39830: {'lr': 5.336724889867919e-05, 'samples': 20393472, 'steps': 39830, 'loss/train': 2.286456823348999} 02/26/2022 05:08:29 - INFO - codeparrot_training - Step 39831: {'lr': 5.3357144668658594e-05, 'samples': 20393984, 'steps': 39831, 'loss/train': 1.088028073310852} 02/26/2022 05:08:35 - INFO - codeparrot_training - Step 39832: {'lr': 5.334704128099349e-05, 'samples': 20394496, 'steps': 39832, 'loss/train': 1.042543649673462} 02/26/2022 05:08:38 - INFO - codeparrot_training - Step 39833: {'lr': 5.333693873572729e-05, 'samples': 20395008, 'steps': 39833, 'loss/train': 1.624452829360962} 02/26/2022 05:08:44 - INFO - codeparrot_training - Step 39834: {'lr': 5.332683703290317e-05, 'samples': 20395520, 'steps': 39834, 'loss/train': 2.2484521865844727} 02/26/2022 05:08:47 - INFO - codeparrot_training - Step 39835: {'lr': 5.3316736172564466e-05, 'samples': 20396032, 'steps': 39835, 'loss/train': 2.118586301803589} 02/26/2022 05:08:53 - INFO - codeparrot_training - Step 39836: {'lr': 5.330663615475431e-05, 'samples': 20396544, 'steps': 39836, 'loss/train': 2.0278921127319336} 02/26/2022 05:08:56 - INFO - codeparrot_training - Step 39837: {'lr': 5.329653697951614e-05, 'samples': 20397056, 'steps': 39837, 'loss/train': 2.142737865447998} 02/26/2022 05:09:04 - INFO - codeparrot_training - Step 39838: {'lr': 5.328643864689317e-05, 'samples': 20397568, 'steps': 39838, 'loss/train': 0.24186043441295624} 02/26/2022 05:09:07 - INFO - codeparrot_training - Step 39839: {'lr': 5.327634115692862e-05, 'samples': 20398080, 'steps': 39839, 'loss/train': 1.9649680852890015} 02/26/2022 05:09:13 - INFO - codeparrot_training - Step 39840: {'lr': 5.326624450966569e-05, 'samples': 20398592, 'steps': 39840, 'loss/train': 1.2368621826171875} 02/26/2022 05:09:16 - INFO - codeparrot_training - Step 39841: {'lr': 5.325614870514775e-05, 'samples': 20399104, 'steps': 39841, 'loss/train': 1.774900197982788} 02/26/2022 05:09:22 - INFO - codeparrot_training - Step 39842: {'lr': 5.3246053743417924e-05, 'samples': 20399616, 'steps': 39842, 'loss/train': 1.5819013118743896} 02/26/2022 05:09:25 - INFO - codeparrot_training - Step 39843: {'lr': 5.323595962451969e-05, 'samples': 20400128, 'steps': 39843, 'loss/train': 1.3393691778182983} 02/26/2022 05:09:31 - INFO - codeparrot_training - Step 39844: {'lr': 5.3225866348495994e-05, 'samples': 20400640, 'steps': 39844, 'loss/train': 1.6038541793823242} 02/26/2022 05:09:34 - INFO - codeparrot_training - Step 39845: {'lr': 5.321577391539026e-05, 'samples': 20401152, 'steps': 39845, 'loss/train': 0.7268335819244385} 02/26/2022 05:09:40 - INFO - codeparrot_training - Step 39846: {'lr': 5.320568232524558e-05, 'samples': 20401664, 'steps': 39846, 'loss/train': 2.3017501831054688} 02/26/2022 05:09:43 - INFO - codeparrot_training - Step 39847: {'lr': 5.319559157810541e-05, 'samples': 20402176, 'steps': 39847, 'loss/train': 1.9190421104431152} 02/26/2022 05:09:49 - INFO - codeparrot_training - Step 39848: {'lr': 5.31855016740127e-05, 'samples': 20402688, 'steps': 39848, 'loss/train': 0.5984655022621155} 02/26/2022 05:09:52 - INFO - codeparrot_training - Step 39849: {'lr': 5.3175412613010895e-05, 'samples': 20403200, 'steps': 39849, 'loss/train': 1.1121525764465332} 02/26/2022 05:09:58 - INFO - codeparrot_training - Step 39850: {'lr': 5.316532439514304e-05, 'samples': 20403712, 'steps': 39850, 'loss/train': 1.1569085121154785} 02/26/2022 05:10:01 - INFO - codeparrot_training - Step 39851: {'lr': 5.3155237020452495e-05, 'samples': 20404224, 'steps': 39851, 'loss/train': 1.4745173454284668} 02/26/2022 05:10:07 - INFO - codeparrot_training - Step 39852: {'lr': 5.3145150488982414e-05, 'samples': 20404736, 'steps': 39852, 'loss/train': 2.1704354286193848} 02/26/2022 05:10:10 - INFO - codeparrot_training - Step 39853: {'lr': 5.3135064800775984e-05, 'samples': 20405248, 'steps': 39853, 'loss/train': 1.8462575674057007} 02/26/2022 05:10:18 - INFO - codeparrot_training - Step 39854: {'lr': 5.3124979955876396e-05, 'samples': 20405760, 'steps': 39854, 'loss/train': 2.3372702598571777} 02/26/2022 05:10:21 - INFO - codeparrot_training - Step 39855: {'lr': 5.311489595432692e-05, 'samples': 20406272, 'steps': 39855, 'loss/train': 0.8868804574012756} 02/26/2022 05:10:27 - INFO - codeparrot_training - Step 39856: {'lr': 5.310481279617071e-05, 'samples': 20406784, 'steps': 39856, 'loss/train': 2.015747308731079} 02/26/2022 05:10:30 - INFO - codeparrot_training - Step 39857: {'lr': 5.309473048145097e-05, 'samples': 20407296, 'steps': 39857, 'loss/train': 1.9868159294128418} 02/26/2022 05:10:36 - INFO - codeparrot_training - Step 39858: {'lr': 5.308464901021079e-05, 'samples': 20407808, 'steps': 39858, 'loss/train': 1.6486444473266602} 02/26/2022 05:10:39 - INFO - codeparrot_training - Step 39859: {'lr': 5.3074568382493524e-05, 'samples': 20408320, 'steps': 39859, 'loss/train': 1.1142995357513428} 02/26/2022 05:10:45 - INFO - codeparrot_training - Step 39860: {'lr': 5.3064488598342285e-05, 'samples': 20408832, 'steps': 39860, 'loss/train': 1.499235987663269} 02/26/2022 05:10:48 - INFO - codeparrot_training - Step 39861: {'lr': 5.3054409657800253e-05, 'samples': 20409344, 'steps': 39861, 'loss/train': 1.0868237018585205} 02/26/2022 05:10:54 - INFO - codeparrot_training - Step 39862: {'lr': 5.304433156091051e-05, 'samples': 20409856, 'steps': 39862, 'loss/train': 1.3696208000183105} 02/26/2022 05:11:01 - INFO - codeparrot_training - Step 39863: {'lr': 5.3034254307716386e-05, 'samples': 20410368, 'steps': 39863, 'loss/train': 2.6534674167633057} 02/26/2022 05:11:04 - INFO - codeparrot_training - Step 39864: {'lr': 5.3024177898260975e-05, 'samples': 20410880, 'steps': 39864, 'loss/train': 2.387263059616089} 02/26/2022 05:11:10 - INFO - codeparrot_training - Step 39865: {'lr': 5.3014102332587436e-05, 'samples': 20411392, 'steps': 39865, 'loss/train': 0.7505513429641724} 02/26/2022 05:11:13 - INFO - codeparrot_training - Step 39866: {'lr': 5.300402761073886e-05, 'samples': 20411904, 'steps': 39866, 'loss/train': 2.1400444507598877} 02/26/2022 05:11:20 - INFO - codeparrot_training - Step 39867: {'lr': 5.299395373275853e-05, 'samples': 20412416, 'steps': 39867, 'loss/train': 1.2390244007110596} 02/26/2022 05:11:23 - INFO - codeparrot_training - Step 39868: {'lr': 5.2983880698689565e-05, 'samples': 20412928, 'steps': 39868, 'loss/train': 1.8390448093414307} 02/26/2022 05:11:26 - INFO - codeparrot_training - Step 39869: {'lr': 5.297380850857508e-05, 'samples': 20413440, 'steps': 39869, 'loss/train': 1.5511404275894165} 02/26/2022 05:11:32 - INFO - codeparrot_training - Step 39870: {'lr': 5.29637371624582e-05, 'samples': 20413952, 'steps': 39870, 'loss/train': 0.892142117023468} 02/26/2022 05:11:35 - INFO - codeparrot_training - Step 39871: {'lr': 5.295366666038215e-05, 'samples': 20414464, 'steps': 39871, 'loss/train': 1.0398935079574585} 02/26/2022 05:11:41 - INFO - codeparrot_training - Step 39872: {'lr': 5.294359700239001e-05, 'samples': 20414976, 'steps': 39872, 'loss/train': 1.8959516286849976} 02/26/2022 05:11:44 - INFO - codeparrot_training - Step 39873: {'lr': 5.293352818852496e-05, 'samples': 20415488, 'steps': 39873, 'loss/train': 1.190548300743103} 02/26/2022 05:11:51 - INFO - codeparrot_training - Step 39874: {'lr': 5.2923460218830074e-05, 'samples': 20416000, 'steps': 39874, 'loss/train': 1.663425326347351} 02/26/2022 05:11:54 - INFO - codeparrot_training - Step 39875: {'lr': 5.2913393093348447e-05, 'samples': 20416512, 'steps': 39875, 'loss/train': 2.3053946495056152} 02/26/2022 05:12:00 - INFO - codeparrot_training - Step 39876: {'lr': 5.2903326812123314e-05, 'samples': 20417024, 'steps': 39876, 'loss/train': 2.026392936706543} 02/26/2022 05:12:03 - INFO - codeparrot_training - Step 39877: {'lr': 5.289326137519771e-05, 'samples': 20417536, 'steps': 39877, 'loss/train': 1.4735510349273682} 02/26/2022 05:12:09 - INFO - codeparrot_training - Step 39878: {'lr': 5.288319678261488e-05, 'samples': 20418048, 'steps': 39878, 'loss/train': 1.9659228324890137} 02/26/2022 05:12:12 - INFO - codeparrot_training - Step 39879: {'lr': 5.287313303441774e-05, 'samples': 20418560, 'steps': 39879, 'loss/train': 2.25028133392334} 02/26/2022 05:12:18 - INFO - codeparrot_training - Step 39880: {'lr': 5.286307013064956e-05, 'samples': 20419072, 'steps': 39880, 'loss/train': 1.6067264080047607} 02/26/2022 05:12:21 - INFO - codeparrot_training - Step 39881: {'lr': 5.285300807135332e-05, 'samples': 20419584, 'steps': 39881, 'loss/train': 1.510894775390625} 02/26/2022 05:12:27 - INFO - codeparrot_training - Step 39882: {'lr': 5.284294685657234e-05, 'samples': 20420096, 'steps': 39882, 'loss/train': 2.0477187633514404} 02/26/2022 05:12:30 - INFO - codeparrot_training - Step 39883: {'lr': 5.283288648634943e-05, 'samples': 20420608, 'steps': 39883, 'loss/train': 1.5893006324768066} 02/26/2022 05:12:36 - INFO - codeparrot_training - Step 39884: {'lr': 5.282282696072788e-05, 'samples': 20421120, 'steps': 39884, 'loss/train': 1.1562637090682983} 02/26/2022 05:12:40 - INFO - codeparrot_training - Step 39885: {'lr': 5.281276827975068e-05, 'samples': 20421632, 'steps': 39885, 'loss/train': 1.087403655052185} 02/26/2022 05:12:45 - INFO - codeparrot_training - Step 39886: {'lr': 5.280271044346108e-05, 'samples': 20422144, 'steps': 39886, 'loss/train': 2.4682846069335938} 02/26/2022 05:12:49 - INFO - codeparrot_training - Step 39887: {'lr': 5.279265345190193e-05, 'samples': 20422656, 'steps': 39887, 'loss/train': 1.378037691116333} 02/26/2022 05:12:54 - INFO - codeparrot_training - Step 39888: {'lr': 5.2782597305116504e-05, 'samples': 20423168, 'steps': 39888, 'loss/train': 1.7448939085006714} 02/26/2022 05:12:58 - INFO - codeparrot_training - Step 39889: {'lr': 5.277254200314774e-05, 'samples': 20423680, 'steps': 39889, 'loss/train': 1.8756383657455444} 02/26/2022 05:13:03 - INFO - codeparrot_training - Step 39890: {'lr': 5.2762487546038854e-05, 'samples': 20424192, 'steps': 39890, 'loss/train': 2.6077356338500977} 02/26/2022 05:13:07 - INFO - codeparrot_training - Step 39891: {'lr': 5.275243393383281e-05, 'samples': 20424704, 'steps': 39891, 'loss/train': 1.5853463411331177} 02/26/2022 05:13:12 - INFO - codeparrot_training - Step 39892: {'lr': 5.2742381166572735e-05, 'samples': 20425216, 'steps': 39892, 'loss/train': 1.5997183322906494} 02/26/2022 05:13:16 - INFO - codeparrot_training - Step 39893: {'lr': 5.273232924430158e-05, 'samples': 20425728, 'steps': 39893, 'loss/train': 2.2245750427246094} 02/26/2022 05:13:22 - INFO - codeparrot_training - Step 39894: {'lr': 5.272227816706257e-05, 'samples': 20426240, 'steps': 39894, 'loss/train': 2.6236751079559326} 02/26/2022 05:13:25 - INFO - codeparrot_training - Step 39895: {'lr': 5.271222793489866e-05, 'samples': 20426752, 'steps': 39895, 'loss/train': 1.755178451538086} 02/26/2022 05:13:31 - INFO - codeparrot_training - Step 39896: {'lr': 5.270217854785292e-05, 'samples': 20427264, 'steps': 39896, 'loss/train': 1.8808153867721558} 02/26/2022 05:13:35 - INFO - codeparrot_training - Step 39897: {'lr': 5.269213000596834e-05, 'samples': 20427776, 'steps': 39897, 'loss/train': 1.2185394763946533} 02/26/2022 05:13:40 - INFO - codeparrot_training - Step 39898: {'lr': 5.268208230928809e-05, 'samples': 20428288, 'steps': 39898, 'loss/train': 1.1425402164459229} 02/26/2022 05:13:44 - INFO - codeparrot_training - Step 39899: {'lr': 5.267203545785515e-05, 'samples': 20428800, 'steps': 39899, 'loss/train': 2.320765733718872} 02/26/2022 05:13:49 - INFO - codeparrot_training - Step 39900: {'lr': 5.266198945171252e-05, 'samples': 20429312, 'steps': 39900, 'loss/train': 1.8524881601333618} 02/26/2022 05:13:53 - INFO - codeparrot_training - Step 39901: {'lr': 5.2651944290903224e-05, 'samples': 20429824, 'steps': 39901, 'loss/train': 2.2199766635894775} 02/26/2022 05:13:58 - INFO - codeparrot_training - Step 39902: {'lr': 5.264189997547042e-05, 'samples': 20430336, 'steps': 39902, 'loss/train': 1.52739679813385} 02/26/2022 05:14:02 - INFO - codeparrot_training - Step 39903: {'lr': 5.2631856505456957e-05, 'samples': 20430848, 'steps': 39903, 'loss/train': 1.385076642036438} 02/26/2022 05:14:07 - INFO - codeparrot_training - Step 39904: {'lr': 5.262181388090609e-05, 'samples': 20431360, 'steps': 39904, 'loss/train': 1.4582237005233765} 02/26/2022 05:14:11 - INFO - codeparrot_training - Step 39905: {'lr': 5.261177210186055e-05, 'samples': 20431872, 'steps': 39905, 'loss/train': 3.2900726795196533} 02/26/2022 05:14:16 - INFO - codeparrot_training - Step 39906: {'lr': 5.260173116836356e-05, 'samples': 20432384, 'steps': 39906, 'loss/train': 2.355900287628174} 02/26/2022 05:14:20 - INFO - codeparrot_training - Step 39907: {'lr': 5.259169108045811e-05, 'samples': 20432896, 'steps': 39907, 'loss/train': 0.424358606338501} 02/26/2022 05:14:25 - INFO - codeparrot_training - Step 39908: {'lr': 5.2581651838187136e-05, 'samples': 20433408, 'steps': 39908, 'loss/train': 1.6566977500915527} 02/26/2022 05:14:29 - INFO - codeparrot_training - Step 39909: {'lr': 5.257161344159364e-05, 'samples': 20433920, 'steps': 39909, 'loss/train': 2.0595741271972656} 02/26/2022 05:14:35 - INFO - codeparrot_training - Step 39910: {'lr': 5.256157589072069e-05, 'samples': 20434432, 'steps': 39910, 'loss/train': 1.082891583442688} 02/26/2022 05:14:38 - INFO - codeparrot_training - Step 39911: {'lr': 5.2551539185611285e-05, 'samples': 20434944, 'steps': 39911, 'loss/train': 1.0988398790359497} 02/26/2022 05:14:45 - INFO - codeparrot_training - Step 39912: {'lr': 5.254150332630833e-05, 'samples': 20435456, 'steps': 39912, 'loss/train': 0.2989959418773651} 02/26/2022 05:14:48 - INFO - codeparrot_training - Step 39913: {'lr': 5.2531468312854935e-05, 'samples': 20435968, 'steps': 39913, 'loss/train': 1.5791479349136353} 02/26/2022 05:14:52 - INFO - codeparrot_training - Step 39914: {'lr': 5.252143414529401e-05, 'samples': 20436480, 'steps': 39914, 'loss/train': 0.8847836256027222} 02/26/2022 05:14:57 - INFO - codeparrot_training - Step 39915: {'lr': 5.251140082366857e-05, 'samples': 20436992, 'steps': 39915, 'loss/train': 0.7409181594848633} 02/26/2022 05:15:01 - INFO - codeparrot_training - Step 39916: {'lr': 5.2501368348021495e-05, 'samples': 20437504, 'steps': 39916, 'loss/train': 0.7507182955741882} 02/26/2022 05:15:06 - INFO - codeparrot_training - Step 39917: {'lr': 5.2491336718395926e-05, 'samples': 20438016, 'steps': 39917, 'loss/train': 0.03427466005086899} 02/26/2022 05:15:10 - INFO - codeparrot_training - Step 39918: {'lr': 5.2481305934834745e-05, 'samples': 20438528, 'steps': 39918, 'loss/train': 0.845059335231781} 02/26/2022 05:15:15 - INFO - codeparrot_training - Step 39919: {'lr': 5.247127599738094e-05, 'samples': 20439040, 'steps': 39919, 'loss/train': 3.348127841949463} 02/26/2022 05:15:19 - INFO - codeparrot_training - Step 39920: {'lr': 5.2461246906077396e-05, 'samples': 20439552, 'steps': 39920, 'loss/train': 2.5653140544891357} 02/26/2022 05:15:25 - INFO - codeparrot_training - Step 39921: {'lr': 5.245121866096722e-05, 'samples': 20440064, 'steps': 39921, 'loss/train': 0.905393123626709} 02/26/2022 05:15:28 - INFO - codeparrot_training - Step 39922: {'lr': 5.244119126209329e-05, 'samples': 20440576, 'steps': 39922, 'loss/train': 1.8390642404556274} 02/26/2022 05:15:34 - INFO - codeparrot_training - Step 39923: {'lr': 5.243116470949855e-05, 'samples': 20441088, 'steps': 39923, 'loss/train': 2.752764940261841} 02/26/2022 05:15:38 - INFO - codeparrot_training - Step 39924: {'lr': 5.24211390032259e-05, 'samples': 20441600, 'steps': 39924, 'loss/train': 1.442712426185608} 02/26/2022 05:15:43 - INFO - codeparrot_training - Step 39925: {'lr': 5.241111414331845e-05, 'samples': 20442112, 'steps': 39925, 'loss/train': 1.36506187915802} 02/26/2022 05:15:47 - INFO - codeparrot_training - Step 39926: {'lr': 5.2401090129819e-05, 'samples': 20442624, 'steps': 39926, 'loss/train': 1.8800948858261108} 02/26/2022 05:15:52 - INFO - codeparrot_training - Step 39927: {'lr': 5.2391066962770575e-05, 'samples': 20443136, 'steps': 39927, 'loss/train': 1.5546599626541138} 02/26/2022 05:15:56 - INFO - codeparrot_training - Step 39928: {'lr': 5.238104464221599e-05, 'samples': 20443648, 'steps': 39928, 'loss/train': 2.213608503341675} 02/26/2022 05:16:01 - INFO - codeparrot_training - Step 39929: {'lr': 5.237102316819831e-05, 'samples': 20444160, 'steps': 39929, 'loss/train': 2.7968904972076416} 02/26/2022 05:16:05 - INFO - codeparrot_training - Step 39930: {'lr': 5.236100254076043e-05, 'samples': 20444672, 'steps': 39930, 'loss/train': 1.476136326789856} 02/26/2022 05:16:12 - INFO - codeparrot_training - Step 39931: {'lr': 5.2350982759945266e-05, 'samples': 20445184, 'steps': 39931, 'loss/train': 1.0488280057907104} 02/26/2022 05:16:15 - INFO - codeparrot_training - Step 39932: {'lr': 5.234096382579565e-05, 'samples': 20445696, 'steps': 39932, 'loss/train': 1.8170726299285889} 02/26/2022 05:16:21 - INFO - codeparrot_training - Step 39933: {'lr': 5.233094573835464e-05, 'samples': 20446208, 'steps': 39933, 'loss/train': 1.8595818281173706} 02/26/2022 05:16:24 - INFO - codeparrot_training - Step 39934: {'lr': 5.232092849766509e-05, 'samples': 20446720, 'steps': 39934, 'loss/train': 1.4527502059936523} 02/26/2022 05:16:30 - INFO - codeparrot_training - Step 39935: {'lr': 5.2310912103769925e-05, 'samples': 20447232, 'steps': 39935, 'loss/train': 1.8730189800262451} 02/26/2022 05:16:33 - INFO - codeparrot_training - Step 39936: {'lr': 5.2300896556711966e-05, 'samples': 20447744, 'steps': 39936, 'loss/train': 1.5099983215332031} 02/26/2022 05:16:39 - INFO - codeparrot_training - Step 39937: {'lr': 5.229088185653427e-05, 'samples': 20448256, 'steps': 39937, 'loss/train': 1.108970284461975} 02/26/2022 05:16:42 - INFO - codeparrot_training - Step 39938: {'lr': 5.228086800327958e-05, 'samples': 20448768, 'steps': 39938, 'loss/train': 2.069016218185425} 02/26/2022 05:16:48 - INFO - codeparrot_training - Step 39939: {'lr': 5.2270854996991025e-05, 'samples': 20449280, 'steps': 39939, 'loss/train': 1.4826850891113281} 02/26/2022 05:16:51 - INFO - codeparrot_training - Step 39940: {'lr': 5.226084283771118e-05, 'samples': 20449792, 'steps': 39940, 'loss/train': 0.9875159859657288} 02/26/2022 05:16:57 - INFO - codeparrot_training - Step 39941: {'lr': 5.2250831525483156e-05, 'samples': 20450304, 'steps': 39941, 'loss/train': 1.9545085430145264} 02/26/2022 05:17:01 - INFO - codeparrot_training - Step 39942: {'lr': 5.224082106034972e-05, 'samples': 20450816, 'steps': 39942, 'loss/train': 1.7497848272323608} 02/26/2022 05:17:06 - INFO - codeparrot_training - Step 39943: {'lr': 5.223081144235395e-05, 'samples': 20451328, 'steps': 39943, 'loss/train': 1.555019497871399} 02/26/2022 05:17:10 - INFO - codeparrot_training - Step 39944: {'lr': 5.2220802671538446e-05, 'samples': 20451840, 'steps': 39944, 'loss/train': 0.8500522971153259} 02/26/2022 05:17:15 - INFO - codeparrot_training - Step 39945: {'lr': 5.2210794747946296e-05, 'samples': 20452352, 'steps': 39945, 'loss/train': 1.1895655393600464} 02/26/2022 05:17:19 - INFO - codeparrot_training - Step 39946: {'lr': 5.220078767162023e-05, 'samples': 20452864, 'steps': 39946, 'loss/train': 1.3385703563690186} 02/26/2022 05:17:25 - INFO - codeparrot_training - Step 39947: {'lr': 5.219078144260331e-05, 'samples': 20453376, 'steps': 39947, 'loss/train': 1.9422162771224976} 02/26/2022 05:17:29 - INFO - codeparrot_training - Step 39948: {'lr': 5.2180776060938126e-05, 'samples': 20453888, 'steps': 39948, 'loss/train': 1.648809552192688} 02/26/2022 05:17:34 - INFO - codeparrot_training - Step 39949: {'lr': 5.217077152666777e-05, 'samples': 20454400, 'steps': 39949, 'loss/train': 0.9245172142982483} 02/26/2022 05:17:40 - INFO - codeparrot_training - Step 39950: {'lr': 5.216076783983492e-05, 'samples': 20454912, 'steps': 39950, 'loss/train': 1.8229225873947144} 02/26/2022 05:17:43 - INFO - codeparrot_training - Step 39951: {'lr': 5.2150765000482605e-05, 'samples': 20455424, 'steps': 39951, 'loss/train': 0.5943360328674316} 02/26/2022 05:17:47 - INFO - codeparrot_training - Step 39952: {'lr': 5.2140763008653586e-05, 'samples': 20455936, 'steps': 39952, 'loss/train': 0.992263674736023} 02/26/2022 05:17:52 - INFO - codeparrot_training - Step 39953: {'lr': 5.21307618643907e-05, 'samples': 20456448, 'steps': 39953, 'loss/train': 1.3918200731277466} 02/26/2022 05:17:58 - INFO - codeparrot_training - Step 39954: {'lr': 5.212076156773673e-05, 'samples': 20456960, 'steps': 39954, 'loss/train': 1.640095591545105} 02/26/2022 05:18:01 - INFO - codeparrot_training - Step 39955: {'lr': 5.2110762118734665e-05, 'samples': 20457472, 'steps': 39955, 'loss/train': 1.9625239372253418} 02/26/2022 05:18:08 - INFO - codeparrot_training - Step 39956: {'lr': 5.210076351742726e-05, 'samples': 20457984, 'steps': 39956, 'loss/train': 0.7800009846687317} 02/26/2022 05:18:11 - INFO - codeparrot_training - Step 39957: {'lr': 5.209076576385735e-05, 'samples': 20458496, 'steps': 39957, 'loss/train': 1.860217809677124} 02/26/2022 05:18:17 - INFO - codeparrot_training - Step 39958: {'lr': 5.208076885806773e-05, 'samples': 20459008, 'steps': 39958, 'loss/train': 0.9180641174316406} 02/26/2022 05:18:20 - INFO - codeparrot_training - Step 39959: {'lr': 5.2070772800101196e-05, 'samples': 20459520, 'steps': 39959, 'loss/train': 2.3119800090789795} 02/26/2022 05:18:26 - INFO - codeparrot_training - Step 39960: {'lr': 5.206077759000069e-05, 'samples': 20460032, 'steps': 39960, 'loss/train': 1.6303881406784058} 02/26/2022 05:18:29 - INFO - codeparrot_training - Step 39961: {'lr': 5.2050783227808955e-05, 'samples': 20460544, 'steps': 39961, 'loss/train': 1.512158751487732} 02/26/2022 05:18:34 - INFO - codeparrot_training - Step 39962: {'lr': 5.204078971356882e-05, 'samples': 20461056, 'steps': 39962, 'loss/train': 0.6630454659461975} 02/26/2022 05:18:38 - INFO - codeparrot_training - Step 39963: {'lr': 5.203079704732303e-05, 'samples': 20461568, 'steps': 39963, 'loss/train': 2.811910390853882} 02/26/2022 05:18:43 - INFO - codeparrot_training - Step 39964: {'lr': 5.202080522911448e-05, 'samples': 20462080, 'steps': 39964, 'loss/train': 1.0674400329589844} 02/26/2022 05:18:47 - INFO - codeparrot_training - Step 39965: {'lr': 5.201081425898593e-05, 'samples': 20462592, 'steps': 39965, 'loss/train': 1.4454385042190552} 02/26/2022 05:18:53 - INFO - codeparrot_training - Step 39966: {'lr': 5.200082413698021e-05, 'samples': 20463104, 'steps': 39966, 'loss/train': 1.005170464515686} 02/26/2022 05:18:57 - INFO - codeparrot_training - Step 39967: {'lr': 5.199083486314002e-05, 'samples': 20463616, 'steps': 39967, 'loss/train': 2.5360870361328125} 02/26/2022 05:19:02 - INFO - codeparrot_training - Step 39968: {'lr': 5.198084643750825e-05, 'samples': 20464128, 'steps': 39968, 'loss/train': 1.8909685611724854} 02/26/2022 05:19:06 - INFO - codeparrot_training - Step 39969: {'lr': 5.19708588601277e-05, 'samples': 20464640, 'steps': 39969, 'loss/train': 2.437023162841797} 02/26/2022 05:19:11 - INFO - codeparrot_training - Step 39970: {'lr': 5.196087213104109e-05, 'samples': 20465152, 'steps': 39970, 'loss/train': 1.450929045677185} 02/26/2022 05:19:15 - INFO - codeparrot_training - Step 39971: {'lr': 5.1950886250291154e-05, 'samples': 20465664, 'steps': 39971, 'loss/train': 1.8635060787200928} 02/26/2022 05:19:20 - INFO - codeparrot_training - Step 39972: {'lr': 5.194090121792078e-05, 'samples': 20466176, 'steps': 39972, 'loss/train': 1.9153963327407837} 02/26/2022 05:19:24 - INFO - codeparrot_training - Step 39973: {'lr': 5.193091703397265e-05, 'samples': 20466688, 'steps': 39973, 'loss/train': 1.155573844909668} 02/26/2022 05:19:29 - INFO - codeparrot_training - Step 39974: {'lr': 5.192093369848971e-05, 'samples': 20467200, 'steps': 39974, 'loss/train': 2.5826423168182373} 02/26/2022 05:19:33 - INFO - codeparrot_training - Step 39975: {'lr': 5.191095121151446e-05, 'samples': 20467712, 'steps': 39975, 'loss/train': 2.0458474159240723} 02/26/2022 05:19:38 - INFO - codeparrot_training - Step 39976: {'lr': 5.1900969573089846e-05, 'samples': 20468224, 'steps': 39976, 'loss/train': 1.9865025281906128} 02/26/2022 05:19:42 - INFO - codeparrot_training - Step 39977: {'lr': 5.189098878325851e-05, 'samples': 20468736, 'steps': 39977, 'loss/train': 0.6222102046012878} 02/26/2022 05:19:47 - INFO - codeparrot_training - Step 39978: {'lr': 5.188100884206343e-05, 'samples': 20469248, 'steps': 39978, 'loss/train': 2.0489501953125} 02/26/2022 05:19:51 - INFO - codeparrot_training - Step 39979: {'lr': 5.187102974954705e-05, 'samples': 20469760, 'steps': 39979, 'loss/train': 2.1365416049957275} 02/26/2022 05:19:57 - INFO - codeparrot_training - Step 39980: {'lr': 5.1861051505752324e-05, 'samples': 20470272, 'steps': 39980, 'loss/train': 1.1082971096038818} 02/26/2022 05:20:00 - INFO - codeparrot_training - Step 39981: {'lr': 5.1851074110721875e-05, 'samples': 20470784, 'steps': 39981, 'loss/train': 1.1782704591751099} 02/26/2022 05:20:06 - INFO - codeparrot_training - Step 39982: {'lr': 5.184109756449865e-05, 'samples': 20471296, 'steps': 39982, 'loss/train': 1.077696681022644} 02/26/2022 05:20:10 - INFO - codeparrot_training - Step 39983: {'lr': 5.1831121867125074e-05, 'samples': 20471808, 'steps': 39983, 'loss/train': 1.541429877281189} 02/26/2022 05:20:15 - INFO - codeparrot_training - Step 39984: {'lr': 5.182114701864416e-05, 'samples': 20472320, 'steps': 39984, 'loss/train': 1.0684746503829956} 02/26/2022 05:20:19 - INFO - codeparrot_training - Step 39985: {'lr': 5.1811173019098434e-05, 'samples': 20472832, 'steps': 39985, 'loss/train': 1.7460517883300781} 02/26/2022 05:20:24 - INFO - codeparrot_training - Step 39986: {'lr': 5.180119986853077e-05, 'samples': 20473344, 'steps': 39986, 'loss/train': 2.0925605297088623} 02/26/2022 05:20:28 - INFO - codeparrot_training - Step 39987: {'lr': 5.1791227566983845e-05, 'samples': 20473856, 'steps': 39987, 'loss/train': 1.1940850019454956} 02/26/2022 05:20:33 - INFO - codeparrot_training - Step 39988: {'lr': 5.178125611450035e-05, 'samples': 20474368, 'steps': 39988, 'loss/train': 1.445656657218933} 02/26/2022 05:20:37 - INFO - codeparrot_training - Step 39989: {'lr': 5.177128551112295e-05, 'samples': 20474880, 'steps': 39989, 'loss/train': 1.4053744077682495} 02/26/2022 05:20:42 - INFO - codeparrot_training - Step 39990: {'lr': 5.176131575689447e-05, 'samples': 20475392, 'steps': 39990, 'loss/train': 1.9748722314834595} 02/26/2022 05:20:45 - INFO - codeparrot_training - Step 39991: {'lr': 5.175134685185759e-05, 'samples': 20475904, 'steps': 39991, 'loss/train': 1.6935545206069946} 02/26/2022 05:20:52 - INFO - codeparrot_training - Step 39992: {'lr': 5.174137879605498e-05, 'samples': 20476416, 'steps': 39992, 'loss/train': 1.3617392778396606} 02/26/2022 05:20:55 - INFO - codeparrot_training - Step 39993: {'lr': 5.1731411589529294e-05, 'samples': 20476928, 'steps': 39993, 'loss/train': 2.605715036392212} 02/26/2022 05:20:59 - INFO - codeparrot_training - Step 39994: {'lr': 5.1721445232323335e-05, 'samples': 20477440, 'steps': 39994, 'loss/train': 1.5338573455810547} 02/26/2022 05:21:04 - INFO - codeparrot_training - Step 39995: {'lr': 5.171147972447973e-05, 'samples': 20477952, 'steps': 39995, 'loss/train': 1.700262427330017} 02/26/2022 05:21:08 - INFO - codeparrot_training - Step 39996: {'lr': 5.170151506604121e-05, 'samples': 20478464, 'steps': 39996, 'loss/train': 1.9522618055343628} 02/26/2022 05:21:14 - INFO - codeparrot_training - Step 39997: {'lr': 5.169155125705036e-05, 'samples': 20478976, 'steps': 39997, 'loss/train': 0.7937091588973999} 02/26/2022 05:21:17 - INFO - codeparrot_training - Step 39998: {'lr': 5.1681588297549986e-05, 'samples': 20479488, 'steps': 39998, 'loss/train': 1.0893386602401733} 02/26/2022 05:21:23 - INFO - codeparrot_training - Step 39999: {'lr': 5.1671626187582723e-05, 'samples': 20480000, 'steps': 39999, 'loss/train': 2.366724729537964} 02/26/2022 05:21:23 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 05:21:41 - WARNING - huggingface_hub.repository - Several commits (40) will be pushed upstream. 02/26/2022 05:21:41 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 05:22:15 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy d2106d5..e25c2d7 floral-grass-11 -> floral-grass-11 02/26/2022 05:22:22 - INFO - codeparrot_training - Step 40000: {'lr': 5.1661664927191235e-05, 'samples': 20480512, 'steps': 40000, 'loss/train': 2.0706076622009277} 02/26/2022 05:22:25 - INFO - codeparrot_training - Step 40001: {'lr': 5.165170451641815e-05, 'samples': 20481024, 'steps': 40001, 'loss/train': 2.458712577819824} 02/26/2022 05:22:32 - INFO - codeparrot_training - Step 40002: {'lr': 5.164174495530621e-05, 'samples': 20481536, 'steps': 40002, 'loss/train': 2.227252244949341} 02/26/2022 05:22:35 - INFO - codeparrot_training - Step 40003: {'lr': 5.163178624389808e-05, 'samples': 20482048, 'steps': 40003, 'loss/train': 2.0496666431427} 02/26/2022 05:22:39 - INFO - codeparrot_training - Step 40004: {'lr': 5.162182838223639e-05, 'samples': 20482560, 'steps': 40004, 'loss/train': 1.467111349105835} 02/26/2022 05:22:44 - INFO - codeparrot_training - Step 40005: {'lr': 5.1611871370363774e-05, 'samples': 20483072, 'steps': 40005, 'loss/train': 1.0251861810684204} 02/26/2022 05:22:48 - INFO - codeparrot_training - Step 40006: {'lr': 5.160191520832283e-05, 'samples': 20483584, 'steps': 40006, 'loss/train': 0.993036687374115} 02/26/2022 05:22:53 - INFO - codeparrot_training - Step 40007: {'lr': 5.159195989615639e-05, 'samples': 20484096, 'steps': 40007, 'loss/train': 0.9503242373466492} 02/26/2022 05:22:57 - INFO - codeparrot_training - Step 40008: {'lr': 5.1582005433906954e-05, 'samples': 20484608, 'steps': 40008, 'loss/train': 2.1146693229675293} 02/26/2022 05:23:02 - INFO - codeparrot_training - Step 40009: {'lr': 5.157205182161723e-05, 'samples': 20485120, 'steps': 40009, 'loss/train': 1.5174983739852905} 02/26/2022 05:23:08 - INFO - codeparrot_training - Step 40010: {'lr': 5.1562099059329734e-05, 'samples': 20485632, 'steps': 40010, 'loss/train': 2.2172343730926514} 02/26/2022 05:23:11 - INFO - codeparrot_training - Step 40011: {'lr': 5.155214714708728e-05, 'samples': 20486144, 'steps': 40011, 'loss/train': 1.5754109621047974} 02/26/2022 05:23:17 - INFO - codeparrot_training - Step 40012: {'lr': 5.154219608493235e-05, 'samples': 20486656, 'steps': 40012, 'loss/train': 1.9428919553756714} 02/26/2022 05:23:21 - INFO - codeparrot_training - Step 40013: {'lr': 5.1532245872907755e-05, 'samples': 20487168, 'steps': 40013, 'loss/train': 1.9903252124786377} 02/26/2022 05:23:26 - INFO - codeparrot_training - Step 40014: {'lr': 5.152229651105586e-05, 'samples': 20487680, 'steps': 40014, 'loss/train': 1.218920111656189} 02/26/2022 05:23:30 - INFO - codeparrot_training - Step 40015: {'lr': 5.151234799941951e-05, 'samples': 20488192, 'steps': 40015, 'loss/train': 0.5999239087104797} 02/26/2022 05:23:35 - INFO - codeparrot_training - Step 40016: {'lr': 5.1502400338041156e-05, 'samples': 20488704, 'steps': 40016, 'loss/train': 1.3848007917404175} 02/26/2022 05:23:39 - INFO - codeparrot_training - Step 40017: {'lr': 5.1492453526963605e-05, 'samples': 20489216, 'steps': 40017, 'loss/train': 1.176537275314331} 02/26/2022 05:23:44 - INFO - codeparrot_training - Step 40018: {'lr': 5.148250756622924e-05, 'samples': 20489728, 'steps': 40018, 'loss/train': 0.8552457690238953} 02/26/2022 05:23:48 - INFO - codeparrot_training - Step 40019: {'lr': 5.147256245588083e-05, 'samples': 20490240, 'steps': 40019, 'loss/train': 0.6426879167556763} 02/26/2022 05:23:54 - INFO - codeparrot_training - Step 40020: {'lr': 5.1462618195960876e-05, 'samples': 20490752, 'steps': 40020, 'loss/train': 2.151014566421509} 02/26/2022 05:23:57 - INFO - codeparrot_training - Step 40021: {'lr': 5.145267478651208e-05, 'samples': 20491264, 'steps': 40021, 'loss/train': 0.49274951219558716} 02/26/2022 05:24:03 - INFO - codeparrot_training - Step 40022: {'lr': 5.144273222757698e-05, 'samples': 20491776, 'steps': 40022, 'loss/train': 2.256671190261841} 02/26/2022 05:24:07 - INFO - codeparrot_training - Step 40023: {'lr': 5.143279051919816e-05, 'samples': 20492288, 'steps': 40023, 'loss/train': 1.3380202054977417} 02/26/2022 05:24:12 - INFO - codeparrot_training - Step 40024: {'lr': 5.142284966141814e-05, 'samples': 20492800, 'steps': 40024, 'loss/train': 0.06912179291248322} 02/26/2022 05:24:16 - INFO - codeparrot_training - Step 40025: {'lr': 5.1412909654279665e-05, 'samples': 20493312, 'steps': 40025, 'loss/train': 1.2770986557006836} 02/26/2022 05:24:21 - INFO - codeparrot_training - Step 40026: {'lr': 5.140297049782519e-05, 'samples': 20493824, 'steps': 40026, 'loss/train': 1.455533504486084} 02/26/2022 05:24:25 - INFO - codeparrot_training - Step 40027: {'lr': 5.139303219209737e-05, 'samples': 20494336, 'steps': 40027, 'loss/train': 2.0087220668792725} 02/26/2022 05:24:30 - INFO - codeparrot_training - Step 40028: {'lr': 5.1383094737138645e-05, 'samples': 20494848, 'steps': 40028, 'loss/train': 0.9133704304695129} 02/26/2022 05:24:34 - INFO - codeparrot_training - Step 40029: {'lr': 5.137315813299173e-05, 'samples': 20495360, 'steps': 40029, 'loss/train': 1.5029116868972778} 02/26/2022 05:24:39 - INFO - codeparrot_training - Step 40030: {'lr': 5.136322237969918e-05, 'samples': 20495872, 'steps': 40030, 'loss/train': 0.5857676863670349} 02/26/2022 05:24:43 - INFO - codeparrot_training - Step 40031: {'lr': 5.135328747730347e-05, 'samples': 20496384, 'steps': 40031, 'loss/train': 2.138592481613159} 02/26/2022 05:24:48 - INFO - codeparrot_training - Step 40032: {'lr': 5.134335342584715e-05, 'samples': 20496896, 'steps': 40032, 'loss/train': 1.4250074625015259} 02/26/2022 05:24:52 - INFO - codeparrot_training - Step 40033: {'lr': 5.133342022537288e-05, 'samples': 20497408, 'steps': 40033, 'loss/train': 1.1370712518692017} 02/26/2022 05:24:57 - INFO - codeparrot_training - Step 40034: {'lr': 5.132348787592309e-05, 'samples': 20497920, 'steps': 40034, 'loss/train': 0.7452365756034851} 02/26/2022 05:25:01 - INFO - codeparrot_training - Step 40035: {'lr': 5.131355637754051e-05, 'samples': 20498432, 'steps': 40035, 'loss/train': 1.8102015256881714} 02/26/2022 05:25:07 - INFO - codeparrot_training - Step 40036: {'lr': 5.130362573026745e-05, 'samples': 20498944, 'steps': 40036, 'loss/train': 1.469778060913086} 02/26/2022 05:25:10 - INFO - codeparrot_training - Step 40037: {'lr': 5.129369593414662e-05, 'samples': 20499456, 'steps': 40037, 'loss/train': 2.7152087688446045} 02/26/2022 05:25:16 - INFO - codeparrot_training - Step 40038: {'lr': 5.128376698922052e-05, 'samples': 20499968, 'steps': 40038, 'loss/train': 2.024228572845459} 02/26/2022 05:25:20 - INFO - codeparrot_training - Step 40039: {'lr': 5.1273838895531636e-05, 'samples': 20500480, 'steps': 40039, 'loss/train': 1.8824213743209839} 02/26/2022 05:25:25 - INFO - codeparrot_training - Step 40040: {'lr': 5.12639116531225e-05, 'samples': 20500992, 'steps': 40040, 'loss/train': 1.4320727586746216} 02/26/2022 05:25:29 - INFO - codeparrot_training - Step 40041: {'lr': 5.125398526203567e-05, 'samples': 20501504, 'steps': 40041, 'loss/train': 2.0509164333343506} 02/26/2022 05:25:34 - INFO - codeparrot_training - Step 40042: {'lr': 5.124405972231369e-05, 'samples': 20502016, 'steps': 40042, 'loss/train': 1.5183888673782349} 02/26/2022 05:25:38 - INFO - codeparrot_training - Step 40043: {'lr': 5.123413503399904e-05, 'samples': 20502528, 'steps': 40043, 'loss/train': 1.6124614477157593} 02/26/2022 05:25:43 - INFO - codeparrot_training - Step 40044: {'lr': 5.122421119713419e-05, 'samples': 20503040, 'steps': 40044, 'loss/train': 1.0150903463363647} 02/26/2022 05:25:47 - INFO - codeparrot_training - Step 40045: {'lr': 5.121428821176174e-05, 'samples': 20503552, 'steps': 40045, 'loss/train': 2.369607448577881} 02/26/2022 05:25:52 - INFO - codeparrot_training - Step 40046: {'lr': 5.12043660779242e-05, 'samples': 20504064, 'steps': 40046, 'loss/train': 2.4435932636260986} 02/26/2022 05:25:56 - INFO - codeparrot_training - Step 40047: {'lr': 5.119444479566393e-05, 'samples': 20504576, 'steps': 40047, 'loss/train': 2.146171808242798} 02/26/2022 05:26:02 - INFO - codeparrot_training - Step 40048: {'lr': 5.118452436502361e-05, 'samples': 20505088, 'steps': 40048, 'loss/train': 1.655824899673462} 02/26/2022 05:26:06 - INFO - codeparrot_training - Step 40049: {'lr': 5.1174604786045634e-05, 'samples': 20505600, 'steps': 40049, 'loss/train': 2.0627388954162598} 02/26/2022 05:26:11 - INFO - codeparrot_training - Step 40050: {'lr': 5.116468605877253e-05, 'samples': 20506112, 'steps': 40050, 'loss/train': 1.1675745248794556} 02/26/2022 05:26:14 - INFO - codeparrot_training - Step 40051: {'lr': 5.1154768183246714e-05, 'samples': 20506624, 'steps': 40051, 'loss/train': 1.636649250984192} 02/26/2022 05:26:20 - INFO - codeparrot_training - Step 40052: {'lr': 5.1144851159510844e-05, 'samples': 20507136, 'steps': 40052, 'loss/train': 0.8711511492729187} 02/26/2022 05:26:24 - INFO - codeparrot_training - Step 40053: {'lr': 5.113493498760718e-05, 'samples': 20507648, 'steps': 40053, 'loss/train': 1.1125317811965942} 02/26/2022 05:26:29 - INFO - codeparrot_training - Step 40054: {'lr': 5.112501966757837e-05, 'samples': 20508160, 'steps': 40054, 'loss/train': 1.3584465980529785} 02/26/2022 05:26:33 - INFO - codeparrot_training - Step 40055: {'lr': 5.111510519946675e-05, 'samples': 20508672, 'steps': 40055, 'loss/train': 1.6737310886383057} 02/26/2022 05:26:38 - INFO - codeparrot_training - Step 40056: {'lr': 5.1105191583315e-05, 'samples': 20509184, 'steps': 40056, 'loss/train': 1.2760004997253418} 02/26/2022 05:26:42 - INFO - codeparrot_training - Step 40057: {'lr': 5.109527881916531e-05, 'samples': 20509696, 'steps': 40057, 'loss/train': 1.088344693183899} 02/26/2022 05:26:47 - INFO - codeparrot_training - Step 40058: {'lr': 5.1085366907060394e-05, 'samples': 20510208, 'steps': 40058, 'loss/train': 0.9741630554199219} 02/26/2022 05:26:51 - INFO - codeparrot_training - Step 40059: {'lr': 5.107545584704249e-05, 'samples': 20510720, 'steps': 40059, 'loss/train': 1.393269419670105} 02/26/2022 05:26:57 - INFO - codeparrot_training - Step 40060: {'lr': 5.106554563915425e-05, 'samples': 20511232, 'steps': 40060, 'loss/train': 1.1468664407730103} 02/26/2022 05:27:00 - INFO - codeparrot_training - Step 40061: {'lr': 5.1055636283438066e-05, 'samples': 20511744, 'steps': 40061, 'loss/train': 1.1084208488464355} 02/26/2022 05:27:06 - INFO - codeparrot_training - Step 40062: {'lr': 5.104572777993632e-05, 'samples': 20512256, 'steps': 40062, 'loss/train': 2.5606679916381836} 02/26/2022 05:27:09 - INFO - codeparrot_training - Step 40063: {'lr': 5.103582012869148e-05, 'samples': 20512768, 'steps': 40063, 'loss/train': 1.2763450145721436} 02/26/2022 05:27:15 - INFO - codeparrot_training - Step 40064: {'lr': 5.102591332974604e-05, 'samples': 20513280, 'steps': 40064, 'loss/train': 1.368504524230957} 02/26/2022 05:27:18 - INFO - codeparrot_training - Step 40065: {'lr': 5.101600738314244e-05, 'samples': 20513792, 'steps': 40065, 'loss/train': 1.5549052953720093} 02/26/2022 05:27:24 - INFO - codeparrot_training - Step 40066: {'lr': 5.100610228892305e-05, 'samples': 20514304, 'steps': 40066, 'loss/train': 1.6941736936569214} 02/26/2022 05:27:27 - INFO - codeparrot_training - Step 40067: {'lr': 5.099619804713027e-05, 'samples': 20514816, 'steps': 40067, 'loss/train': 2.4828040599823} 02/26/2022 05:27:33 - INFO - codeparrot_training - Step 40068: {'lr': 5.098629465780666e-05, 'samples': 20515328, 'steps': 40068, 'loss/train': 1.2498379945755005} 02/26/2022 05:27:36 - INFO - codeparrot_training - Step 40069: {'lr': 5.09763921209945e-05, 'samples': 20515840, 'steps': 40069, 'loss/train': 2.242964506149292} 02/26/2022 05:27:42 - INFO - codeparrot_training - Step 40070: {'lr': 5.09664904367364e-05, 'samples': 20516352, 'steps': 40070, 'loss/train': 2.03818941116333} 02/26/2022 05:27:46 - INFO - codeparrot_training - Step 40071: {'lr': 5.0956589605074546e-05, 'samples': 20516864, 'steps': 40071, 'loss/train': 0.3864399492740631} 02/26/2022 05:27:51 - INFO - codeparrot_training - Step 40072: {'lr': 5.0946689626051505e-05, 'samples': 20517376, 'steps': 40072, 'loss/train': 1.2123653888702393} 02/26/2022 05:27:57 - INFO - codeparrot_training - Step 40073: {'lr': 5.093679049970959e-05, 'samples': 20517888, 'steps': 40073, 'loss/train': 1.380172848701477} 02/26/2022 05:28:00 - INFO - codeparrot_training - Step 40074: {'lr': 5.092689222609137e-05, 'samples': 20518400, 'steps': 40074, 'loss/train': 2.345250368118286} 02/26/2022 05:28:06 - INFO - codeparrot_training - Step 40075: {'lr': 5.091699480523901e-05, 'samples': 20518912, 'steps': 40075, 'loss/train': 1.3724156618118286} 02/26/2022 05:28:09 - INFO - codeparrot_training - Step 40076: {'lr': 5.0907098237195084e-05, 'samples': 20519424, 'steps': 40076, 'loss/train': 1.8128714561462402} 02/26/2022 05:28:15 - INFO - codeparrot_training - Step 40077: {'lr': 5.0897202522001865e-05, 'samples': 20519936, 'steps': 40077, 'loss/train': 1.265244483947754} 02/26/2022 05:28:19 - INFO - codeparrot_training - Step 40078: {'lr': 5.088730765970195e-05, 'samples': 20520448, 'steps': 40078, 'loss/train': 1.8832745552062988} 02/26/2022 05:28:24 - INFO - codeparrot_training - Step 40079: {'lr': 5.087741365033746e-05, 'samples': 20520960, 'steps': 40079, 'loss/train': 2.4139490127563477} 02/26/2022 05:28:28 - INFO - codeparrot_training - Step 40080: {'lr': 5.086752049395094e-05, 'samples': 20521472, 'steps': 40080, 'loss/train': 0.6943578720092773} 02/26/2022 05:28:33 - INFO - codeparrot_training - Step 40081: {'lr': 5.085762819058468e-05, 'samples': 20521984, 'steps': 40081, 'loss/train': 2.4370670318603516} 02/26/2022 05:28:37 - INFO - codeparrot_training - Step 40082: {'lr': 5.084773674028118e-05, 'samples': 20522496, 'steps': 40082, 'loss/train': 1.584202527999878} 02/26/2022 05:28:42 - INFO - codeparrot_training - Step 40083: {'lr': 5.083784614308273e-05, 'samples': 20523008, 'steps': 40083, 'loss/train': 1.9010438919067383} 02/26/2022 05:28:46 - INFO - codeparrot_training - Step 40084: {'lr': 5.082795639903171e-05, 'samples': 20523520, 'steps': 40084, 'loss/train': 1.6217018365859985} 02/26/2022 05:28:52 - INFO - codeparrot_training - Step 40085: {'lr': 5.0818067508170475e-05, 'samples': 20524032, 'steps': 40085, 'loss/train': 0.5566498041152954} 02/26/2022 05:28:55 - INFO - codeparrot_training - Step 40086: {'lr': 5.080817947054131e-05, 'samples': 20524544, 'steps': 40086, 'loss/train': 1.7606621980667114} 02/26/2022 05:29:01 - INFO - codeparrot_training - Step 40087: {'lr': 5.079829228618674e-05, 'samples': 20525056, 'steps': 40087, 'loss/train': 1.311846137046814} 02/26/2022 05:29:04 - INFO - codeparrot_training - Step 40088: {'lr': 5.078840595514902e-05, 'samples': 20525568, 'steps': 40088, 'loss/train': 1.570685863494873} 02/26/2022 05:29:10 - INFO - codeparrot_training - Step 40089: {'lr': 5.077852047747053e-05, 'samples': 20526080, 'steps': 40089, 'loss/train': 2.0077085494995117} 02/26/2022 05:29:13 - INFO - codeparrot_training - Step 40090: {'lr': 5.076863585319352e-05, 'samples': 20526592, 'steps': 40090, 'loss/train': 1.116776943206787} 02/26/2022 05:29:19 - INFO - codeparrot_training - Step 40091: {'lr': 5.07587520823605e-05, 'samples': 20527104, 'steps': 40091, 'loss/train': 2.1811821460723877} 02/26/2022 05:29:22 - INFO - codeparrot_training - Step 40092: {'lr': 5.07488691650137e-05, 'samples': 20527616, 'steps': 40092, 'loss/train': 1.6807693243026733} 02/26/2022 05:29:28 - INFO - codeparrot_training - Step 40093: {'lr': 5.0738987101195454e-05, 'samples': 20528128, 'steps': 40093, 'loss/train': 1.837868332862854} 02/26/2022 05:29:31 - INFO - codeparrot_training - Step 40094: {'lr': 5.0729105890948064e-05, 'samples': 20528640, 'steps': 40094, 'loss/train': 1.634913682937622} 02/26/2022 05:29:37 - INFO - codeparrot_training - Step 40095: {'lr': 5.071922553431399e-05, 'samples': 20529152, 'steps': 40095, 'loss/train': 2.005528688430786} 02/26/2022 05:29:41 - INFO - codeparrot_training - Step 40096: {'lr': 5.070934603133548e-05, 'samples': 20529664, 'steps': 40096, 'loss/train': 2.3615410327911377} 02/26/2022 05:29:46 - INFO - codeparrot_training - Step 40097: {'lr': 5.069946738205483e-05, 'samples': 20530176, 'steps': 40097, 'loss/train': 0.8500372171401978} 02/26/2022 05:29:50 - INFO - codeparrot_training - Step 40098: {'lr': 5.068958958651432e-05, 'samples': 20530688, 'steps': 40098, 'loss/train': 1.7104334831237793} 02/26/2022 05:29:55 - INFO - codeparrot_training - Step 40099: {'lr': 5.067971264475638e-05, 'samples': 20531200, 'steps': 40099, 'loss/train': 0.045369166880846024} 02/26/2022 05:29:59 - INFO - codeparrot_training - Step 40100: {'lr': 5.066983655682325e-05, 'samples': 20531712, 'steps': 40100, 'loss/train': 1.5760809183120728} 02/26/2022 05:30:05 - INFO - codeparrot_training - Step 40101: {'lr': 5.065996132275724e-05, 'samples': 20532224, 'steps': 40101, 'loss/train': 2.4488492012023926} 02/26/2022 05:30:08 - INFO - codeparrot_training - Step 40102: {'lr': 5.065008694260062e-05, 'samples': 20532736, 'steps': 40102, 'loss/train': 1.4471408128738403} 02/26/2022 05:30:14 - INFO - codeparrot_training - Step 40103: {'lr': 5.06402134163958e-05, 'samples': 20533248, 'steps': 40103, 'loss/train': 0.8726754188537598} 02/26/2022 05:30:17 - INFO - codeparrot_training - Step 40104: {'lr': 5.063034074418496e-05, 'samples': 20533760, 'steps': 40104, 'loss/train': 1.3631744384765625} 02/26/2022 05:30:23 - INFO - codeparrot_training - Step 40105: {'lr': 5.062046892601047e-05, 'samples': 20534272, 'steps': 40105, 'loss/train': 1.4810874462127686} 02/26/2022 05:30:27 - INFO - codeparrot_training - Step 40106: {'lr': 5.06105979619145e-05, 'samples': 20534784, 'steps': 40106, 'loss/train': 1.344446063041687} 02/26/2022 05:30:32 - INFO - codeparrot_training - Step 40107: {'lr': 5.0600727851939476e-05, 'samples': 20535296, 'steps': 40107, 'loss/train': 1.884863018989563} 02/26/2022 05:30:36 - INFO - codeparrot_training - Step 40108: {'lr': 5.0590858596127554e-05, 'samples': 20535808, 'steps': 40108, 'loss/train': 1.4678008556365967} 02/26/2022 05:30:41 - INFO - codeparrot_training - Step 40109: {'lr': 5.0580990194521226e-05, 'samples': 20536320, 'steps': 40109, 'loss/train': 2.0973196029663086} 02/26/2022 05:30:45 - INFO - codeparrot_training - Step 40110: {'lr': 5.057112264716246e-05, 'samples': 20536832, 'steps': 40110, 'loss/train': 1.0496724843978882} 02/26/2022 05:30:50 - INFO - codeparrot_training - Step 40111: {'lr': 5.056125595409375e-05, 'samples': 20537344, 'steps': 40111, 'loss/train': 0.8334508538246155} 02/26/2022 05:30:54 - INFO - codeparrot_training - Step 40112: {'lr': 5.0551390115357225e-05, 'samples': 20537856, 'steps': 40112, 'loss/train': 2.1235551834106445} 02/26/2022 05:30:59 - INFO - codeparrot_training - Step 40113: {'lr': 5.054152513099533e-05, 'samples': 20538368, 'steps': 40113, 'loss/train': 0.5583958029747009} 02/26/2022 05:31:03 - INFO - codeparrot_training - Step 40114: {'lr': 5.0531661001050114e-05, 'samples': 20538880, 'steps': 40114, 'loss/train': 1.6286840438842773} 02/26/2022 05:31:09 - INFO - codeparrot_training - Step 40115: {'lr': 5.052179772556395e-05, 'samples': 20539392, 'steps': 40115, 'loss/train': 1.1722922325134277} 02/26/2022 05:31:12 - INFO - codeparrot_training - Step 40116: {'lr': 5.051193530457901e-05, 'samples': 20539904, 'steps': 40116, 'loss/train': 1.9946843385696411} 02/26/2022 05:31:18 - INFO - codeparrot_training - Step 40117: {'lr': 5.050207373813767e-05, 'samples': 20540416, 'steps': 40117, 'loss/train': 1.5008091926574707} 02/26/2022 05:31:21 - INFO - codeparrot_training - Step 40118: {'lr': 5.04922130262821e-05, 'samples': 20540928, 'steps': 40118, 'loss/train': 1.550695538520813} 02/26/2022 05:31:27 - INFO - codeparrot_training - Step 40119: {'lr': 5.048235316905453e-05, 'samples': 20541440, 'steps': 40119, 'loss/train': 2.4589316844940186} 02/26/2022 05:31:30 - INFO - codeparrot_training - Step 40120: {'lr': 5.047249416649713e-05, 'samples': 20541952, 'steps': 40120, 'loss/train': 1.2552516460418701} 02/26/2022 05:31:36 - INFO - codeparrot_training - Step 40121: {'lr': 5.046263601865228e-05, 'samples': 20542464, 'steps': 40121, 'loss/train': 1.1930433511734009} 02/26/2022 05:31:39 - INFO - codeparrot_training - Step 40122: {'lr': 5.045277872556214e-05, 'samples': 20542976, 'steps': 40122, 'loss/train': 2.5336406230926514} 02/26/2022 05:31:45 - INFO - codeparrot_training - Step 40123: {'lr': 5.044292228726893e-05, 'samples': 20543488, 'steps': 40123, 'loss/train': 2.752393960952759} 02/26/2022 05:31:48 - INFO - codeparrot_training - Step 40124: {'lr': 5.043306670381481e-05, 'samples': 20544000, 'steps': 40124, 'loss/train': 1.7778352499008179} 02/26/2022 05:31:55 - INFO - codeparrot_training - Step 40125: {'lr': 5.0423211975242136e-05, 'samples': 20544512, 'steps': 40125, 'loss/train': 2.417404890060425} 02/26/2022 05:31:58 - INFO - codeparrot_training - Step 40126: {'lr': 5.041335810159303e-05, 'samples': 20545024, 'steps': 40126, 'loss/train': 1.8781704902648926} 02/26/2022 05:32:03 - INFO - codeparrot_training - Step 40127: {'lr': 5.040350508290975e-05, 'samples': 20545536, 'steps': 40127, 'loss/train': 1.4455132484436035} 02/26/2022 05:32:07 - INFO - codeparrot_training - Step 40128: {'lr': 5.0393652919234395e-05, 'samples': 20546048, 'steps': 40128, 'loss/train': 0.7860716581344604} 02/26/2022 05:32:12 - INFO - codeparrot_training - Step 40129: {'lr': 5.0383801610609294e-05, 'samples': 20546560, 'steps': 40129, 'loss/train': 1.3679593801498413} 02/26/2022 05:32:16 - INFO - codeparrot_training - Step 40130: {'lr': 5.037395115707663e-05, 'samples': 20547072, 'steps': 40130, 'loss/train': 2.0277838706970215} 02/26/2022 05:32:22 - INFO - codeparrot_training - Step 40131: {'lr': 5.036410155867857e-05, 'samples': 20547584, 'steps': 40131, 'loss/train': 1.784590244293213} 02/26/2022 05:32:25 - INFO - codeparrot_training - Step 40132: {'lr': 5.0354252815457275e-05, 'samples': 20548096, 'steps': 40132, 'loss/train': 2.3224823474884033} 02/26/2022 05:32:31 - INFO - codeparrot_training - Step 40133: {'lr': 5.034440492745493e-05, 'samples': 20548608, 'steps': 40133, 'loss/train': 1.8877272605895996} 02/26/2022 05:32:34 - INFO - codeparrot_training - Step 40134: {'lr': 5.0334557894713834e-05, 'samples': 20549120, 'steps': 40134, 'loss/train': 1.1653307676315308} 02/26/2022 05:32:40 - INFO - codeparrot_training - Step 40135: {'lr': 5.0324711717276064e-05, 'samples': 20549632, 'steps': 40135, 'loss/train': 1.8135045766830444} 02/26/2022 05:32:43 - INFO - codeparrot_training - Step 40136: {'lr': 5.031486639518385e-05, 'samples': 20550144, 'steps': 40136, 'loss/train': 1.6274279356002808} 02/26/2022 05:32:49 - INFO - codeparrot_training - Step 40137: {'lr': 5.030502192847924e-05, 'samples': 20550656, 'steps': 40137, 'loss/train': 1.4808542728424072} 02/26/2022 05:32:52 - INFO - codeparrot_training - Step 40138: {'lr': 5.0295178317204595e-05, 'samples': 20551168, 'steps': 40138, 'loss/train': 1.9895950555801392} 02/26/2022 05:32:58 - INFO - codeparrot_training - Step 40139: {'lr': 5.028533556140199e-05, 'samples': 20551680, 'steps': 40139, 'loss/train': 2.356114387512207} 02/26/2022 05:33:01 - INFO - codeparrot_training - Step 40140: {'lr': 5.0275493661113584e-05, 'samples': 20552192, 'steps': 40140, 'loss/train': 1.7727941274642944} 02/26/2022 05:33:08 - INFO - codeparrot_training - Step 40141: {'lr': 5.0265652616381496e-05, 'samples': 20552704, 'steps': 40141, 'loss/train': 1.488487958908081} 02/26/2022 05:33:12 - INFO - codeparrot_training - Step 40142: {'lr': 5.025581242724797e-05, 'samples': 20553216, 'steps': 40142, 'loss/train': 1.5116512775421143} 02/26/2022 05:33:17 - INFO - codeparrot_training - Step 40143: {'lr': 5.024597309375506e-05, 'samples': 20553728, 'steps': 40143, 'loss/train': 2.096799850463867} 02/26/2022 05:33:21 - INFO - codeparrot_training - Step 40144: {'lr': 5.023613461594512e-05, 'samples': 20554240, 'steps': 40144, 'loss/train': 0.38335734605789185} 02/26/2022 05:33:26 - INFO - codeparrot_training - Step 40145: {'lr': 5.0226296993859996e-05, 'samples': 20554752, 'steps': 40145, 'loss/train': 1.1693456172943115} 02/26/2022 05:33:30 - INFO - codeparrot_training - Step 40146: {'lr': 5.021646022754206e-05, 'samples': 20555264, 'steps': 40146, 'loss/train': 0.5619250535964966} 02/26/2022 05:33:35 - INFO - codeparrot_training - Step 40147: {'lr': 5.020662431703329e-05, 'samples': 20555776, 'steps': 40147, 'loss/train': 1.6746019124984741} 02/26/2022 05:33:39 - INFO - codeparrot_training - Step 40148: {'lr': 5.0196789262376055e-05, 'samples': 20556288, 'steps': 40148, 'loss/train': 1.7113332748413086} 02/26/2022 05:33:44 - INFO - codeparrot_training - Step 40149: {'lr': 5.018695506361218e-05, 'samples': 20556800, 'steps': 40149, 'loss/train': 2.1674680709838867} 02/26/2022 05:33:48 - INFO - codeparrot_training - Step 40150: {'lr': 5.017712172078401e-05, 'samples': 20557312, 'steps': 40150, 'loss/train': 2.3900294303894043} 02/26/2022 05:33:55 - INFO - codeparrot_training - Step 40151: {'lr': 5.0167289233933546e-05, 'samples': 20557824, 'steps': 40151, 'loss/train': 0.1731039136648178} 02/26/2022 05:33:58 - INFO - codeparrot_training - Step 40152: {'lr': 5.0157457603103055e-05, 'samples': 20558336, 'steps': 40152, 'loss/train': 1.2805061340332031} 02/26/2022 05:34:04 - INFO - codeparrot_training - Step 40153: {'lr': 5.0147626828334456e-05, 'samples': 20558848, 'steps': 40153, 'loss/train': 2.149085283279419} 02/26/2022 05:34:07 - INFO - codeparrot_training - Step 40154: {'lr': 5.0137796909670046e-05, 'samples': 20559360, 'steps': 40154, 'loss/train': 1.5128902196884155} 02/26/2022 05:34:13 - INFO - codeparrot_training - Step 40155: {'lr': 5.0127967847151764e-05, 'samples': 20559872, 'steps': 40155, 'loss/train': 1.9556902647018433} 02/26/2022 05:34:16 - INFO - codeparrot_training - Step 40156: {'lr': 5.011813964082187e-05, 'samples': 20560384, 'steps': 40156, 'loss/train': 2.357943296432495} 02/26/2022 05:34:22 - INFO - codeparrot_training - Step 40157: {'lr': 5.010831229072238e-05, 'samples': 20560896, 'steps': 40157, 'loss/train': 0.8816267848014832} 02/26/2022 05:34:25 - INFO - codeparrot_training - Step 40158: {'lr': 5.0098485796895435e-05, 'samples': 20561408, 'steps': 40158, 'loss/train': 1.043492317199707} 02/26/2022 05:34:31 - INFO - codeparrot_training - Step 40159: {'lr': 5.0088660159383016e-05, 'samples': 20561920, 'steps': 40159, 'loss/train': 0.5299811959266663} 02/26/2022 05:34:34 - INFO - codeparrot_training - Step 40160: {'lr': 5.007883537822736e-05, 'samples': 20562432, 'steps': 40160, 'loss/train': 2.145418167114258} 02/26/2022 05:34:40 - INFO - codeparrot_training - Step 40161: {'lr': 5.006901145347051e-05, 'samples': 20562944, 'steps': 40161, 'loss/train': 0.6495620012283325} 02/26/2022 05:34:43 - INFO - codeparrot_training - Step 40162: {'lr': 5.0059188385154525e-05, 'samples': 20563456, 'steps': 40162, 'loss/train': 2.6692028045654297} 02/26/2022 05:34:50 - INFO - codeparrot_training - Step 40163: {'lr': 5.0049366173321395e-05, 'samples': 20563968, 'steps': 40163, 'loss/train': 1.9270213842391968} 02/26/2022 05:34:54 - INFO - codeparrot_training - Step 40164: {'lr': 5.0039544818013396e-05, 'samples': 20564480, 'steps': 40164, 'loss/train': 1.7030454874038696} 02/26/2022 05:34:59 - INFO - codeparrot_training - Step 40165: {'lr': 5.0029724319272466e-05, 'samples': 20564992, 'steps': 40165, 'loss/train': 2.9641973972320557} 02/26/2022 05:35:05 - INFO - codeparrot_training - Step 40166: {'lr': 5.001990467714071e-05, 'samples': 20565504, 'steps': 40166, 'loss/train': 2.1683390140533447} 02/26/2022 05:35:08 - INFO - codeparrot_training - Step 40167: {'lr': 5.0010085891660115e-05, 'samples': 20566016, 'steps': 40167, 'loss/train': 1.439074993133545} 02/26/2022 05:35:12 - INFO - codeparrot_training - Step 40168: {'lr': 5.000026796287288e-05, 'samples': 20566528, 'steps': 40168, 'loss/train': 0.9136861562728882} 02/26/2022 05:35:17 - INFO - codeparrot_training - Step 40169: {'lr': 4.9990450890820964e-05, 'samples': 20567040, 'steps': 40169, 'loss/train': 1.5852373838424683} 02/26/2022 05:35:21 - INFO - codeparrot_training - Step 40170: {'lr': 4.998063467554648e-05, 'samples': 20567552, 'steps': 40170, 'loss/train': 1.5793037414550781} 02/26/2022 05:35:26 - INFO - codeparrot_training - Step 40171: {'lr': 4.9970819317091356e-05, 'samples': 20568064, 'steps': 40171, 'loss/train': 1.851469874382019} 02/26/2022 05:35:30 - INFO - codeparrot_training - Step 40172: {'lr': 4.996100481549781e-05, 'samples': 20568576, 'steps': 40172, 'loss/train': 1.5916630029678345} 02/26/2022 05:35:36 - INFO - codeparrot_training - Step 40173: {'lr': 4.995119117080779e-05, 'samples': 20569088, 'steps': 40173, 'loss/train': 1.2014036178588867} 02/26/2022 05:35:39 - INFO - codeparrot_training - Step 40174: {'lr': 4.994137838306334e-05, 'samples': 20569600, 'steps': 40174, 'loss/train': 0.8751251101493835} 02/26/2022 05:35:45 - INFO - codeparrot_training - Step 40175: {'lr': 4.993156645230643e-05, 'samples': 20570112, 'steps': 40175, 'loss/train': 2.2376980781555176} 02/26/2022 05:35:48 - INFO - codeparrot_training - Step 40176: {'lr': 4.992175537857924e-05, 'samples': 20570624, 'steps': 40176, 'loss/train': 2.159022092819214} 02/26/2022 05:35:54 - INFO - codeparrot_training - Step 40177: {'lr': 4.991194516192371e-05, 'samples': 20571136, 'steps': 40177, 'loss/train': 2.358538866043091} 02/26/2022 05:35:57 - INFO - codeparrot_training - Step 40178: {'lr': 4.990213580238179e-05, 'samples': 20571648, 'steps': 40178, 'loss/train': 0.9790414571762085} 02/26/2022 05:36:03 - INFO - codeparrot_training - Step 40179: {'lr': 4.9892327299995655e-05, 'samples': 20572160, 'steps': 40179, 'loss/train': 2.0554122924804688} 02/26/2022 05:36:08 - INFO - codeparrot_training - Step 40180: {'lr': 4.988251965480728e-05, 'samples': 20572672, 'steps': 40180, 'loss/train': 2.4225683212280273} 02/26/2022 05:36:12 - INFO - codeparrot_training - Step 40181: {'lr': 4.98727128668586e-05, 'samples': 20573184, 'steps': 40181, 'loss/train': 1.674485206604004} 02/26/2022 05:36:18 - INFO - codeparrot_training - Step 40182: {'lr': 4.986290693619164e-05, 'samples': 20573696, 'steps': 40182, 'loss/train': 2.0279340744018555} 02/26/2022 05:36:21 - INFO - codeparrot_training - Step 40183: {'lr': 4.985310186284855e-05, 'samples': 20574208, 'steps': 40183, 'loss/train': 1.9559215307235718} 02/26/2022 05:36:26 - INFO - codeparrot_training - Step 40184: {'lr': 4.9843297646871096e-05, 'samples': 20574720, 'steps': 40184, 'loss/train': 0.44912201166152954} 02/26/2022 05:36:30 - INFO - codeparrot_training - Step 40185: {'lr': 4.9833494288301474e-05, 'samples': 20575232, 'steps': 40185, 'loss/train': 0.500665009021759} 02/26/2022 05:36:35 - INFO - codeparrot_training - Step 40186: {'lr': 4.982369178718152e-05, 'samples': 20575744, 'steps': 40186, 'loss/train': 1.400141716003418} 02/26/2022 05:36:39 - INFO - codeparrot_training - Step 40187: {'lr': 4.981389014355342e-05, 'samples': 20576256, 'steps': 40187, 'loss/train': 2.11801815032959} 02/26/2022 05:36:45 - INFO - codeparrot_training - Step 40188: {'lr': 4.9804089357458945e-05, 'samples': 20576768, 'steps': 40188, 'loss/train': 2.884381055831909} 02/26/2022 05:36:48 - INFO - codeparrot_training - Step 40189: {'lr': 4.979428942894024e-05, 'samples': 20577280, 'steps': 40189, 'loss/train': 1.8546353578567505} 02/26/2022 05:36:54 - INFO - codeparrot_training - Step 40190: {'lr': 4.978449035803917e-05, 'samples': 20577792, 'steps': 40190, 'loss/train': 1.9362952709197998} 02/26/2022 05:36:57 - INFO - codeparrot_training - Step 40191: {'lr': 4.9774692144797824e-05, 'samples': 20578304, 'steps': 40191, 'loss/train': 1.2781065702438354} 02/26/2022 05:37:03 - INFO - codeparrot_training - Step 40192: {'lr': 4.976489478925811e-05, 'samples': 20578816, 'steps': 40192, 'loss/train': 1.7723109722137451} 02/26/2022 05:37:06 - INFO - codeparrot_training - Step 40193: {'lr': 4.975509829146202e-05, 'samples': 20579328, 'steps': 40193, 'loss/train': 2.224621057510376} 02/26/2022 05:37:12 - INFO - codeparrot_training - Step 40194: {'lr': 4.974530265145144e-05, 'samples': 20579840, 'steps': 40194, 'loss/train': 2.2311601638793945} 02/26/2022 05:37:15 - INFO - codeparrot_training - Step 40195: {'lr': 4.973550786926845e-05, 'samples': 20580352, 'steps': 40195, 'loss/train': 1.803673505783081} 02/26/2022 05:37:21 - INFO - codeparrot_training - Step 40196: {'lr': 4.9725713944954956e-05, 'samples': 20580864, 'steps': 40196, 'loss/train': 1.989607810974121} 02/26/2022 05:37:24 - INFO - codeparrot_training - Step 40197: {'lr': 4.97159208785529e-05, 'samples': 20581376, 'steps': 40197, 'loss/train': 1.81013822555542} 02/26/2022 05:37:31 - INFO - codeparrot_training - Step 40198: {'lr': 4.970612867010418e-05, 'samples': 20581888, 'steps': 40198, 'loss/train': 1.4249409437179565} 02/26/2022 05:37:34 - INFO - codeparrot_training - Step 40199: {'lr': 4.969633731965087e-05, 'samples': 20582400, 'steps': 40199, 'loss/train': 1.727388858795166} 02/26/2022 05:37:40 - INFO - codeparrot_training - Step 40200: {'lr': 4.9686546827234865e-05, 'samples': 20582912, 'steps': 40200, 'loss/train': 1.0296684503555298} 02/26/2022 05:37:43 - INFO - codeparrot_training - Step 40201: {'lr': 4.967675719289807e-05, 'samples': 20583424, 'steps': 40201, 'loss/train': 2.1278061866760254} 02/26/2022 05:37:49 - INFO - codeparrot_training - Step 40202: {'lr': 4.966696841668239e-05, 'samples': 20583936, 'steps': 40202, 'loss/train': 2.240410804748535} 02/26/2022 05:37:52 - INFO - codeparrot_training - Step 40203: {'lr': 4.965718049862985e-05, 'samples': 20584448, 'steps': 40203, 'loss/train': 1.5604584217071533} 02/26/2022 05:37:58 - INFO - codeparrot_training - Step 40204: {'lr': 4.964739343878227e-05, 'samples': 20584960, 'steps': 40204, 'loss/train': 1.1971962451934814} 02/26/2022 05:38:01 - INFO - codeparrot_training - Step 40205: {'lr': 4.9637607237181775e-05, 'samples': 20585472, 'steps': 40205, 'loss/train': 0.9253528714179993} 02/26/2022 05:38:07 - INFO - codeparrot_training - Step 40206: {'lr': 4.962782189387e-05, 'samples': 20585984, 'steps': 40206, 'loss/train': 1.8904399871826172} 02/26/2022 05:38:10 - INFO - codeparrot_training - Step 40207: {'lr': 4.9618037408889095e-05, 'samples': 20586496, 'steps': 40207, 'loss/train': 1.267303705215454} 02/26/2022 05:38:17 - INFO - codeparrot_training - Step 40208: {'lr': 4.960825378228082e-05, 'samples': 20587008, 'steps': 40208, 'loss/train': 2.0993943214416504} 02/26/2022 05:38:20 - INFO - codeparrot_training - Step 40209: {'lr': 4.959847101408729e-05, 'samples': 20587520, 'steps': 40209, 'loss/train': 2.250439405441284} 02/26/2022 05:38:26 - INFO - codeparrot_training - Step 40210: {'lr': 4.9588689104350134e-05, 'samples': 20588032, 'steps': 40210, 'loss/train': 1.7145178318023682} 02/26/2022 05:38:29 - INFO - codeparrot_training - Step 40211: {'lr': 4.957890805311147e-05, 'samples': 20588544, 'steps': 40211, 'loss/train': 2.0793075561523438} 02/26/2022 05:38:35 - INFO - codeparrot_training - Step 40212: {'lr': 4.9569127860413066e-05, 'samples': 20589056, 'steps': 40212, 'loss/train': 1.5559533834457397} 02/26/2022 05:38:39 - INFO - codeparrot_training - Step 40213: {'lr': 4.955934852629698e-05, 'samples': 20589568, 'steps': 40213, 'loss/train': 2.2094738483428955} 02/26/2022 05:38:44 - INFO - codeparrot_training - Step 40214: {'lr': 4.9549570050804876e-05, 'samples': 20590080, 'steps': 40214, 'loss/train': 2.9797372817993164} 02/26/2022 05:38:48 - INFO - codeparrot_training - Step 40215: {'lr': 4.953979243397882e-05, 'samples': 20590592, 'steps': 40215, 'loss/train': 0.03249917924404144} 02/26/2022 05:38:53 - INFO - codeparrot_training - Step 40216: {'lr': 4.953001567586068e-05, 'samples': 20591104, 'steps': 40216, 'loss/train': 2.1189401149749756} 02/26/2022 05:38:57 - INFO - codeparrot_training - Step 40217: {'lr': 4.95202397764922e-05, 'samples': 20591616, 'steps': 40217, 'loss/train': 0.28025704622268677} 02/26/2022 05:39:03 - INFO - codeparrot_training - Step 40218: {'lr': 4.951046473591542e-05, 'samples': 20592128, 'steps': 40218, 'loss/train': 1.8474280834197998} 02/26/2022 05:39:06 - INFO - codeparrot_training - Step 40219: {'lr': 4.950069055417214e-05, 'samples': 20592640, 'steps': 40219, 'loss/train': 2.675217628479004} 02/26/2022 05:39:12 - INFO - codeparrot_training - Step 40220: {'lr': 4.949091723130425e-05, 'samples': 20593152, 'steps': 40220, 'loss/train': 2.324949026107788} 02/26/2022 05:39:15 - INFO - codeparrot_training - Step 40221: {'lr': 4.948114476735355e-05, 'samples': 20593664, 'steps': 40221, 'loss/train': 1.628584861755371} 02/26/2022 05:39:21 - INFO - codeparrot_training - Step 40222: {'lr': 4.9471373162362e-05, 'samples': 20594176, 'steps': 40222, 'loss/train': 1.1752337217330933} 02/26/2022 05:39:24 - INFO - codeparrot_training - Step 40223: {'lr': 4.9461602416371405e-05, 'samples': 20594688, 'steps': 40223, 'loss/train': 2.300471782684326} 02/26/2022 05:39:30 - INFO - codeparrot_training - Step 40224: {'lr': 4.945183252942362e-05, 'samples': 20595200, 'steps': 40224, 'loss/train': 2.1665046215057373} 02/26/2022 05:39:33 - INFO - codeparrot_training - Step 40225: {'lr': 4.944206350156047e-05, 'samples': 20595712, 'steps': 40225, 'loss/train': 2.160352945327759} 02/26/2022 05:39:39 - INFO - codeparrot_training - Step 40226: {'lr': 4.943229533282387e-05, 'samples': 20596224, 'steps': 40226, 'loss/train': 1.5185710191726685} 02/26/2022 05:39:43 - INFO - codeparrot_training - Step 40227: {'lr': 4.942252802325564e-05, 'samples': 20596736, 'steps': 40227, 'loss/train': 2.0852737426757812} 02/26/2022 05:39:48 - INFO - codeparrot_training - Step 40228: {'lr': 4.941276157289762e-05, 'samples': 20597248, 'steps': 40228, 'loss/train': 1.7140722274780273} 02/26/2022 05:39:52 - INFO - codeparrot_training - Step 40229: {'lr': 4.9402995981791573e-05, 'samples': 20597760, 'steps': 40229, 'loss/train': 2.2234158515930176} 02/26/2022 05:39:57 - INFO - codeparrot_training - Step 40230: {'lr': 4.939323124997944e-05, 'samples': 20598272, 'steps': 40230, 'loss/train': 1.0340170860290527} 02/26/2022 05:40:00 - INFO - codeparrot_training - Step 40231: {'lr': 4.938346737750302e-05, 'samples': 20598784, 'steps': 40231, 'loss/train': 1.3000550270080566} 02/26/2022 05:40:08 - INFO - codeparrot_training - Step 40232: {'lr': 4.9373704364404106e-05, 'samples': 20599296, 'steps': 40232, 'loss/train': 0.4428895115852356} 02/26/2022 05:40:12 - INFO - codeparrot_training - Step 40233: {'lr': 4.936394221072449e-05, 'samples': 20599808, 'steps': 40233, 'loss/train': 1.55039644241333} 02/26/2022 05:40:17 - INFO - codeparrot_training - Step 40234: {'lr': 4.935418091650609e-05, 'samples': 20600320, 'steps': 40234, 'loss/train': 1.3537769317626953} 02/26/2022 05:40:21 - INFO - codeparrot_training - Step 40235: {'lr': 4.934442048179069e-05, 'samples': 20600832, 'steps': 40235, 'loss/train': 1.1374977827072144} 02/26/2022 05:40:26 - INFO - codeparrot_training - Step 40236: {'lr': 4.933466090662006e-05, 'samples': 20601344, 'steps': 40236, 'loss/train': 2.037734270095825} 02/26/2022 05:40:30 - INFO - codeparrot_training - Step 40237: {'lr': 4.932490219103594e-05, 'samples': 20601856, 'steps': 40237, 'loss/train': 1.880330204963684} 02/26/2022 05:40:35 - INFO - codeparrot_training - Step 40238: {'lr': 4.93151443350803e-05, 'samples': 20602368, 'steps': 40238, 'loss/train': 2.1331236362457275} 02/26/2022 05:40:39 - INFO - codeparrot_training - Step 40239: {'lr': 4.9305387338794803e-05, 'samples': 20602880, 'steps': 40239, 'loss/train': 1.8540791273117065} 02/26/2022 05:40:44 - INFO - codeparrot_training - Step 40240: {'lr': 4.929563120222141e-05, 'samples': 20603392, 'steps': 40240, 'loss/train': 2.607050657272339} 02/26/2022 05:40:48 - INFO - codeparrot_training - Step 40241: {'lr': 4.928587592540168e-05, 'samples': 20603904, 'steps': 40241, 'loss/train': 1.510053277015686} 02/26/2022 05:40:55 - INFO - codeparrot_training - Step 40242: {'lr': 4.92761215083776e-05, 'samples': 20604416, 'steps': 40242, 'loss/train': 3.6378045082092285} 02/26/2022 05:40:59 - INFO - codeparrot_training - Step 40243: {'lr': 4.9266367951190796e-05, 'samples': 20604928, 'steps': 40243, 'loss/train': 0.13775387406349182} 02/26/2022 05:41:04 - INFO - codeparrot_training - Step 40244: {'lr': 4.925661525388328e-05, 'samples': 20605440, 'steps': 40244, 'loss/train': 1.8081549406051636} 02/26/2022 05:41:08 - INFO - codeparrot_training - Step 40245: {'lr': 4.924686341649653e-05, 'samples': 20605952, 'steps': 40245, 'loss/train': 2.084050178527832} 02/26/2022 05:41:13 - INFO - codeparrot_training - Step 40246: {'lr': 4.9237112439072555e-05, 'samples': 20606464, 'steps': 40246, 'loss/train': 1.5899858474731445} 02/26/2022 05:41:17 - INFO - codeparrot_training - Step 40247: {'lr': 4.9227362321652954e-05, 'samples': 20606976, 'steps': 40247, 'loss/train': 0.9158003926277161} 02/26/2022 05:41:22 - INFO - codeparrot_training - Step 40248: {'lr': 4.9217613064279714e-05, 'samples': 20607488, 'steps': 40248, 'loss/train': 1.5574148893356323} 02/26/2022 05:41:26 - INFO - codeparrot_training - Step 40249: {'lr': 4.920786466699434e-05, 'samples': 20608000, 'steps': 40249, 'loss/train': 2.0158865451812744} 02/26/2022 05:41:31 - INFO - codeparrot_training - Step 40250: {'lr': 4.919811712983879e-05, 'samples': 20608512, 'steps': 40250, 'loss/train': 1.5035561323165894} 02/26/2022 05:41:35 - INFO - codeparrot_training - Step 40251: {'lr': 4.918837045285468e-05, 'samples': 20609024, 'steps': 40251, 'loss/train': 2.6910500526428223} 02/26/2022 05:41:42 - INFO - codeparrot_training - Step 40252: {'lr': 4.917862463608389e-05, 'samples': 20609536, 'steps': 40252, 'loss/train': 2.20007061958313} 02/26/2022 05:41:45 - INFO - codeparrot_training - Step 40253: {'lr': 4.9168879679568095e-05, 'samples': 20610048, 'steps': 40253, 'loss/train': 1.6671972274780273} 02/26/2022 05:41:51 - INFO - codeparrot_training - Step 40254: {'lr': 4.915913558334906e-05, 'samples': 20610560, 'steps': 40254, 'loss/train': 2.116628646850586} 02/26/2022 05:41:55 - INFO - codeparrot_training - Step 40255: {'lr': 4.914939234746843e-05, 'samples': 20611072, 'steps': 40255, 'loss/train': 2.0490267276763916} 02/26/2022 05:42:00 - INFO - codeparrot_training - Step 40256: {'lr': 4.91396499719681e-05, 'samples': 20611584, 'steps': 40256, 'loss/train': 1.2523854970932007} 02/26/2022 05:42:04 - INFO - codeparrot_training - Step 40257: {'lr': 4.912990845688972e-05, 'samples': 20612096, 'steps': 40257, 'loss/train': 1.367032766342163} 02/26/2022 05:42:09 - INFO - codeparrot_training - Step 40258: {'lr': 4.912016780227504e-05, 'samples': 20612608, 'steps': 40258, 'loss/train': 1.9894146919250488} 02/26/2022 05:42:13 - INFO - codeparrot_training - Step 40259: {'lr': 4.91104280081657e-05, 'samples': 20613120, 'steps': 40259, 'loss/train': 2.3943426609039307} 02/26/2022 05:42:18 - INFO - codeparrot_training - Step 40260: {'lr': 4.910068907460358e-05, 'samples': 20613632, 'steps': 40260, 'loss/train': 2.610714912414551} 02/26/2022 05:42:22 - INFO - codeparrot_training - Step 40261: {'lr': 4.90909510016303e-05, 'samples': 20614144, 'steps': 40261, 'loss/train': 2.104128360748291} 02/26/2022 05:42:27 - INFO - codeparrot_training - Step 40262: {'lr': 4.908121378928759e-05, 'samples': 20614656, 'steps': 40262, 'loss/train': 0.2344641238451004} 02/26/2022 05:42:31 - INFO - codeparrot_training - Step 40263: {'lr': 4.907147743761717e-05, 'samples': 20615168, 'steps': 40263, 'loss/train': 1.5972188711166382} 02/26/2022 05:42:38 - INFO - codeparrot_training - Step 40264: {'lr': 4.906174194666069e-05, 'samples': 20615680, 'steps': 40264, 'loss/train': 1.8211365938186646} 02/26/2022 05:42:42 - INFO - codeparrot_training - Step 40265: {'lr': 4.905200731645992e-05, 'samples': 20616192, 'steps': 40265, 'loss/train': 1.0623358488082886} 02/26/2022 05:42:47 - INFO - codeparrot_training - Step 40266: {'lr': 4.904227354705659e-05, 'samples': 20616704, 'steps': 40266, 'loss/train': 2.013404130935669} 02/26/2022 05:42:51 - INFO - codeparrot_training - Step 40267: {'lr': 4.903254063849233e-05, 'samples': 20617216, 'steps': 40267, 'loss/train': 1.8484675884246826} 02/26/2022 05:42:56 - INFO - codeparrot_training - Step 40268: {'lr': 4.902280859080876e-05, 'samples': 20617728, 'steps': 40268, 'loss/train': 2.3819262981414795} 02/26/2022 05:43:00 - INFO - codeparrot_training - Step 40269: {'lr': 4.901307740404776e-05, 'samples': 20618240, 'steps': 40269, 'loss/train': 1.7495012283325195} 02/26/2022 05:43:05 - INFO - codeparrot_training - Step 40270: {'lr': 4.900334707825091e-05, 'samples': 20618752, 'steps': 40270, 'loss/train': 1.4493780136108398} 02/26/2022 05:43:09 - INFO - codeparrot_training - Step 40271: {'lr': 4.899361761345988e-05, 'samples': 20619264, 'steps': 40271, 'loss/train': 0.7620633244514465} 02/26/2022 05:43:14 - INFO - codeparrot_training - Step 40272: {'lr': 4.898388900971634e-05, 'samples': 20619776, 'steps': 40272, 'loss/train': 1.3191192150115967} 02/26/2022 05:43:18 - INFO - codeparrot_training - Step 40273: {'lr': 4.897416126706203e-05, 'samples': 20620288, 'steps': 40273, 'loss/train': 1.7160799503326416} 02/26/2022 05:43:23 - INFO - codeparrot_training - Step 40274: {'lr': 4.8964434385538574e-05, 'samples': 20620800, 'steps': 40274, 'loss/train': 1.4717628955841064} 02/26/2022 05:43:27 - INFO - codeparrot_training - Step 40275: {'lr': 4.8954708365187646e-05, 'samples': 20621312, 'steps': 40275, 'loss/train': 2.196061849594116} 02/26/2022 05:43:32 - INFO - codeparrot_training - Step 40276: {'lr': 4.894498320605084e-05, 'samples': 20621824, 'steps': 40276, 'loss/train': 1.260366439819336} 02/26/2022 05:43:36 - INFO - codeparrot_training - Step 40277: {'lr': 4.893525890816997e-05, 'samples': 20622336, 'steps': 40277, 'loss/train': 1.4206507205963135} 02/26/2022 05:43:43 - INFO - codeparrot_training - Step 40278: {'lr': 4.8925535471586515e-05, 'samples': 20622848, 'steps': 40278, 'loss/train': 0.799635112285614} 02/26/2022 05:43:46 - INFO - codeparrot_training - Step 40279: {'lr': 4.8915812896342363e-05, 'samples': 20623360, 'steps': 40279, 'loss/train': 4.539821624755859} 02/26/2022 05:43:52 - INFO - codeparrot_training - Step 40280: {'lr': 4.890609118247888e-05, 'samples': 20623872, 'steps': 40280, 'loss/train': 2.939774990081787} 02/26/2022 05:43:55 - INFO - codeparrot_training - Step 40281: {'lr': 4.889637033003794e-05, 'samples': 20624384, 'steps': 40281, 'loss/train': 0.7319871783256531} 02/26/2022 05:44:01 - INFO - codeparrot_training - Step 40282: {'lr': 4.888665033906098e-05, 'samples': 20624896, 'steps': 40282, 'loss/train': 1.9546502828598022} 02/26/2022 05:44:04 - INFO - codeparrot_training - Step 40283: {'lr': 4.887693120958991e-05, 'samples': 20625408, 'steps': 40283, 'loss/train': 1.8348294496536255} 02/26/2022 05:44:10 - INFO - codeparrot_training - Step 40284: {'lr': 4.8867212941666075e-05, 'samples': 20625920, 'steps': 40284, 'loss/train': 1.8929299116134644} 02/26/2022 05:44:13 - INFO - codeparrot_training - Step 40285: {'lr': 4.885749553533128e-05, 'samples': 20626432, 'steps': 40285, 'loss/train': 1.8761297464370728} 02/26/2022 05:44:19 - INFO - codeparrot_training - Step 40286: {'lr': 4.8847778990627065e-05, 'samples': 20626944, 'steps': 40286, 'loss/train': 1.6109845638275146} 02/26/2022 05:44:22 - INFO - codeparrot_training - Step 40287: {'lr': 4.883806330759513e-05, 'samples': 20627456, 'steps': 40287, 'loss/train': 1.7778126001358032} 02/26/2022 05:44:29 - INFO - codeparrot_training - Step 40288: {'lr': 4.882834848627707e-05, 'samples': 20627968, 'steps': 40288, 'loss/train': 1.9402728080749512} 02/26/2022 05:44:33 - INFO - codeparrot_training - Step 40289: {'lr': 4.8818634526714475e-05, 'samples': 20628480, 'steps': 40289, 'loss/train': 1.888139009475708} 02/26/2022 05:44:39 - INFO - codeparrot_training - Step 40290: {'lr': 4.880892142894891e-05, 'samples': 20628992, 'steps': 40290, 'loss/train': 1.9593098163604736} 02/26/2022 05:44:42 - INFO - codeparrot_training - Step 40291: {'lr': 4.879920919302211e-05, 'samples': 20629504, 'steps': 40291, 'loss/train': 2.0112409591674805} 02/26/2022 05:44:47 - INFO - codeparrot_training - Step 40292: {'lr': 4.87894978189756e-05, 'samples': 20630016, 'steps': 40292, 'loss/train': 1.7236138582229614} 02/26/2022 05:44:51 - INFO - codeparrot_training - Step 40293: {'lr': 4.877978730685098e-05, 'samples': 20630528, 'steps': 40293, 'loss/train': 2.652998208999634} 02/26/2022 05:44:57 - INFO - codeparrot_training - Step 40294: {'lr': 4.877007765668981e-05, 'samples': 20631040, 'steps': 40294, 'loss/train': 2.97902774810791} 02/26/2022 05:45:00 - INFO - codeparrot_training - Step 40295: {'lr': 4.87603688685338e-05, 'samples': 20631552, 'steps': 40295, 'loss/train': 0.0480986088514328} 02/26/2022 05:45:06 - INFO - codeparrot_training - Step 40296: {'lr': 4.875066094242445e-05, 'samples': 20632064, 'steps': 40296, 'loss/train': 1.9318784475326538} 02/26/2022 05:45:13 - INFO - codeparrot_training - Step 40297: {'lr': 4.874095387840338e-05, 'samples': 20632576, 'steps': 40297, 'loss/train': 1.6873944997787476} 02/26/2022 05:45:16 - INFO - codeparrot_training - Step 40298: {'lr': 4.8731247676512076e-05, 'samples': 20633088, 'steps': 40298, 'loss/train': 2.169782876968384} 02/26/2022 05:45:22 - INFO - codeparrot_training - Step 40299: {'lr': 4.8721542336792266e-05, 'samples': 20633600, 'steps': 40299, 'loss/train': 1.590254783630371} 02/26/2022 05:45:25 - INFO - codeparrot_training - Step 40300: {'lr': 4.871183785928546e-05, 'samples': 20634112, 'steps': 40300, 'loss/train': 0.49390989542007446} 02/26/2022 05:45:31 - INFO - codeparrot_training - Step 40301: {'lr': 4.8702134244033216e-05, 'samples': 20634624, 'steps': 40301, 'loss/train': 1.6930512189865112} 02/26/2022 05:45:34 - INFO - codeparrot_training - Step 40302: {'lr': 4.8692431491077075e-05, 'samples': 20635136, 'steps': 40302, 'loss/train': 1.3124794960021973} 02/26/2022 05:45:40 - INFO - codeparrot_training - Step 40303: {'lr': 4.868272960045866e-05, 'samples': 20635648, 'steps': 40303, 'loss/train': 2.376016139984131} 02/26/2022 05:45:43 - INFO - codeparrot_training - Step 40304: {'lr': 4.867302857221953e-05, 'samples': 20636160, 'steps': 40304, 'loss/train': 2.35194730758667} 02/26/2022 05:45:49 - INFO - codeparrot_training - Step 40305: {'lr': 4.8663328406401226e-05, 'samples': 20636672, 'steps': 40305, 'loss/train': 1.8355700969696045} 02/26/2022 05:45:52 - INFO - codeparrot_training - Step 40306: {'lr': 4.8653629103045214e-05, 'samples': 20637184, 'steps': 40306, 'loss/train': 2.4623522758483887} 02/26/2022 05:46:00 - INFO - codeparrot_training - Step 40307: {'lr': 4.8643930662193166e-05, 'samples': 20637696, 'steps': 40307, 'loss/train': 1.516063928604126} 02/26/2022 05:46:03 - INFO - codeparrot_training - Step 40308: {'lr': 4.863423308388659e-05, 'samples': 20638208, 'steps': 40308, 'loss/train': 1.8128044605255127} 02/26/2022 05:46:09 - INFO - codeparrot_training - Step 40309: {'lr': 4.862453636816705e-05, 'samples': 20638720, 'steps': 40309, 'loss/train': 1.3424099683761597} 02/26/2022 05:46:12 - INFO - codeparrot_training - Step 40310: {'lr': 4.861484051507603e-05, 'samples': 20639232, 'steps': 40310, 'loss/train': 1.806500792503357} 02/26/2022 05:46:18 - INFO - codeparrot_training - Step 40311: {'lr': 4.8605145524655e-05, 'samples': 20639744, 'steps': 40311, 'loss/train': 1.4328997135162354} 02/26/2022 05:46:21 - INFO - codeparrot_training - Step 40312: {'lr': 4.8595451396945685e-05, 'samples': 20640256, 'steps': 40312, 'loss/train': 0.896848738193512} 02/26/2022 05:46:27 - INFO - codeparrot_training - Step 40313: {'lr': 4.85857581319894e-05, 'samples': 20640768, 'steps': 40313, 'loss/train': 3.0436997413635254} 02/26/2022 05:46:30 - INFO - codeparrot_training - Step 40314: {'lr': 4.857606572982792e-05, 'samples': 20641280, 'steps': 40314, 'loss/train': 1.6188586950302124} 02/26/2022 05:46:36 - INFO - codeparrot_training - Step 40315: {'lr': 4.856637419050247e-05, 'samples': 20641792, 'steps': 40315, 'loss/train': 1.3449711799621582} 02/26/2022 05:46:39 - INFO - codeparrot_training - Step 40316: {'lr': 4.855668351405479e-05, 'samples': 20642304, 'steps': 40316, 'loss/train': 1.1660408973693848} 02/26/2022 05:46:45 - INFO - codeparrot_training - Step 40317: {'lr': 4.854699370052626e-05, 'samples': 20642816, 'steps': 40317, 'loss/train': 1.2835907936096191} 02/26/2022 05:46:48 - INFO - codeparrot_training - Step 40318: {'lr': 4.8537304749958544e-05, 'samples': 20643328, 'steps': 40318, 'loss/train': 2.2396318912506104} 02/26/2022 05:46:54 - INFO - codeparrot_training - Step 40319: {'lr': 4.8527616662392914e-05, 'samples': 20643840, 'steps': 40319, 'loss/train': 2.143279790878296} 02/26/2022 05:46:57 - INFO - codeparrot_training - Step 40320: {'lr': 4.851792943787109e-05, 'samples': 20644352, 'steps': 40320, 'loss/train': 1.4868887662887573} 02/26/2022 05:47:03 - INFO - codeparrot_training - Step 40321: {'lr': 4.850824307643439e-05, 'samples': 20644864, 'steps': 40321, 'loss/train': 1.2822750806808472} 02/26/2022 05:47:06 - INFO - codeparrot_training - Step 40322: {'lr': 4.849855757812455e-05, 'samples': 20645376, 'steps': 40322, 'loss/train': 2.142866611480713} 02/26/2022 05:47:13 - INFO - codeparrot_training - Step 40323: {'lr': 4.848887294298274e-05, 'samples': 20645888, 'steps': 40323, 'loss/train': 1.1699045896530151} 02/26/2022 05:47:17 - INFO - codeparrot_training - Step 40324: {'lr': 4.8479189171050706e-05, 'samples': 20646400, 'steps': 40324, 'loss/train': 0.7557324767112732} 02/26/2022 05:47:22 - INFO - codeparrot_training - Step 40325: {'lr': 4.846950626236976e-05, 'samples': 20646912, 'steps': 40325, 'loss/train': 1.879967212677002} 02/26/2022 05:47:26 - INFO - codeparrot_training - Step 40326: {'lr': 4.84598242169815e-05, 'samples': 20647424, 'steps': 40326, 'loss/train': 2.1193058490753174} 02/26/2022 05:47:31 - INFO - codeparrot_training - Step 40327: {'lr': 4.845014303492739e-05, 'samples': 20647936, 'steps': 40327, 'loss/train': 0.20236289501190186} 02/26/2022 05:47:35 - INFO - codeparrot_training - Step 40328: {'lr': 4.844046271624886e-05, 'samples': 20648448, 'steps': 40328, 'loss/train': 1.8047990798950195} 02/26/2022 05:47:41 - INFO - codeparrot_training - Step 40329: {'lr': 4.843078326098732e-05, 'samples': 20648960, 'steps': 40329, 'loss/train': 1.8903441429138184} 02/26/2022 05:47:44 - INFO - codeparrot_training - Step 40330: {'lr': 4.842110466918434e-05, 'samples': 20649472, 'steps': 40330, 'loss/train': 3.083993673324585} 02/26/2022 05:47:50 - INFO - codeparrot_training - Step 40331: {'lr': 4.841142694088138e-05, 'samples': 20649984, 'steps': 40331, 'loss/train': 1.4143931865692139} 02/26/2022 05:47:53 - INFO - codeparrot_training - Step 40332: {'lr': 4.8401750076119827e-05, 'samples': 20650496, 'steps': 40332, 'loss/train': 1.3540617227554321} 02/26/2022 05:47:58 - INFO - codeparrot_training - Step 40333: {'lr': 4.839207407494109e-05, 'samples': 20651008, 'steps': 40333, 'loss/train': 2.186516523361206} 02/26/2022 05:48:02 - INFO - codeparrot_training - Step 40334: {'lr': 4.838239893738677e-05, 'samples': 20651520, 'steps': 40334, 'loss/train': 1.4937856197357178} 02/26/2022 05:48:09 - INFO - codeparrot_training - Step 40335: {'lr': 4.837272466349818e-05, 'samples': 20652032, 'steps': 40335, 'loss/train': 1.642652153968811} 02/26/2022 05:48:12 - INFO - codeparrot_training - Step 40336: {'lr': 4.836305125331694e-05, 'samples': 20652544, 'steps': 40336, 'loss/train': 1.775629997253418} 02/26/2022 05:48:18 - INFO - codeparrot_training - Step 40337: {'lr': 4.835337870688422e-05, 'samples': 20653056, 'steps': 40337, 'loss/train': 1.382092833518982} 02/26/2022 05:48:24 - INFO - codeparrot_training - Step 40338: {'lr': 4.834370702424165e-05, 'samples': 20653568, 'steps': 40338, 'loss/train': 2.4745702743530273} 02/26/2022 05:48:27 - INFO - codeparrot_training - Step 40339: {'lr': 4.833403620543056e-05, 'samples': 20654080, 'steps': 40339, 'loss/train': 1.7895512580871582} 02/26/2022 05:48:33 - INFO - codeparrot_training - Step 40340: {'lr': 4.8324366250492553e-05, 'samples': 20654592, 'steps': 40340, 'loss/train': 1.7502992153167725} 02/26/2022 05:48:36 - INFO - codeparrot_training - Step 40341: {'lr': 4.8314697159468764e-05, 'samples': 20655104, 'steps': 40341, 'loss/train': 1.076488971710205} 02/26/2022 05:48:41 - INFO - codeparrot_training - Step 40342: {'lr': 4.8305028932400854e-05, 'samples': 20655616, 'steps': 40342, 'loss/train': 1.9662400484085083} 02/26/2022 05:48:45 - INFO - codeparrot_training - Step 40343: {'lr': 4.829536156933018e-05, 'samples': 20656128, 'steps': 40343, 'loss/train': 1.3128210306167603} 02/26/2022 05:48:52 - INFO - codeparrot_training - Step 40344: {'lr': 4.8285695070298094e-05, 'samples': 20656640, 'steps': 40344, 'loss/train': 1.4712445735931396} 02/26/2022 05:48:56 - INFO - codeparrot_training - Step 40345: {'lr': 4.827602943534598e-05, 'samples': 20657152, 'steps': 40345, 'loss/train': 1.8644556999206543} 02/26/2022 05:49:01 - INFO - codeparrot_training - Step 40346: {'lr': 4.826636466451537e-05, 'samples': 20657664, 'steps': 40346, 'loss/train': 1.7470310926437378} 02/26/2022 05:49:05 - INFO - codeparrot_training - Step 40347: {'lr': 4.82567007578476e-05, 'samples': 20658176, 'steps': 40347, 'loss/train': 2.2780463695526123} 02/26/2022 05:49:11 - INFO - codeparrot_training - Step 40348: {'lr': 4.8247037715384e-05, 'samples': 20658688, 'steps': 40348, 'loss/train': 2.12211537361145} 02/26/2022 05:49:14 - INFO - codeparrot_training - Step 40349: {'lr': 4.823737553716609e-05, 'samples': 20659200, 'steps': 40349, 'loss/train': 5.402233123779297} 02/26/2022 05:49:18 - INFO - codeparrot_training - Step 40350: {'lr': 4.8227714223235185e-05, 'samples': 20659712, 'steps': 40350, 'loss/train': 8.754063606262207} 02/26/2022 05:49:23 - INFO - codeparrot_training - Step 40351: {'lr': 4.82180537736327e-05, 'samples': 20660224, 'steps': 40351, 'loss/train': 1.3048845529556274} 02/26/2022 05:49:27 - INFO - codeparrot_training - Step 40352: {'lr': 4.820839418839992e-05, 'samples': 20660736, 'steps': 40352, 'loss/train': 1.3804601430892944} 02/26/2022 05:49:32 - INFO - codeparrot_training - Step 40353: {'lr': 4.819873546757836e-05, 'samples': 20661248, 'steps': 40353, 'loss/train': 1.3691766262054443} 02/26/2022 05:49:36 - INFO - codeparrot_training - Step 40354: {'lr': 4.818907761120936e-05, 'samples': 20661760, 'steps': 40354, 'loss/train': 1.8946551084518433} 02/26/2022 05:49:43 - INFO - codeparrot_training - Step 40355: {'lr': 4.817942061933425e-05, 'samples': 20662272, 'steps': 40355, 'loss/train': 2.6421525478363037} 02/26/2022 05:49:46 - INFO - codeparrot_training - Step 40356: {'lr': 4.816976449199437e-05, 'samples': 20662784, 'steps': 40356, 'loss/train': 2.05658221244812} 02/26/2022 05:49:52 - INFO - codeparrot_training - Step 40357: {'lr': 4.816010922923125e-05, 'samples': 20663296, 'steps': 40357, 'loss/train': 1.6093169450759888} 02/26/2022 05:49:55 - INFO - codeparrot_training - Step 40358: {'lr': 4.8150454831086014e-05, 'samples': 20663808, 'steps': 40358, 'loss/train': 1.247710108757019} 02/26/2022 05:50:01 - INFO - codeparrot_training - Step 40359: {'lr': 4.8140801297600175e-05, 'samples': 20664320, 'steps': 40359, 'loss/train': 1.6580345630645752} 02/26/2022 05:50:04 - INFO - codeparrot_training - Step 40360: {'lr': 4.8131148628815016e-05, 'samples': 20664832, 'steps': 40360, 'loss/train': 2.1830527782440186} 02/26/2022 05:50:10 - INFO - codeparrot_training - Step 40361: {'lr': 4.812149682477196e-05, 'samples': 20665344, 'steps': 40361, 'loss/train': 2.115399122238159} 02/26/2022 05:50:13 - INFO - codeparrot_training - Step 40362: {'lr': 4.811184588551232e-05, 'samples': 20665856, 'steps': 40362, 'loss/train': 0.8488881587982178} 02/26/2022 05:50:19 - INFO - codeparrot_training - Step 40363: {'lr': 4.810219581107739e-05, 'samples': 20666368, 'steps': 40363, 'loss/train': 2.124912738800049} 02/26/2022 05:50:26 - INFO - codeparrot_training - Step 40364: {'lr': 4.809254660150852e-05, 'samples': 20666880, 'steps': 40364, 'loss/train': 2.0002939701080322} 02/26/2022 05:50:30 - INFO - codeparrot_training - Step 40365: {'lr': 4.808289825684711e-05, 'samples': 20667392, 'steps': 40365, 'loss/train': 1.3772315979003906} 02/26/2022 05:50:35 - INFO - codeparrot_training - Step 40366: {'lr': 4.8073250777134456e-05, 'samples': 20667904, 'steps': 40366, 'loss/train': 2.5757248401641846} 02/26/2022 05:50:39 - INFO - codeparrot_training - Step 40367: {'lr': 4.8063604162411875e-05, 'samples': 20668416, 'steps': 40367, 'loss/train': 1.1028145551681519} 02/26/2022 05:50:44 - INFO - codeparrot_training - Step 40368: {'lr': 4.8053958412720615e-05, 'samples': 20668928, 'steps': 40368, 'loss/train': 0.9846283197402954} 02/26/2022 05:50:48 - INFO - codeparrot_training - Step 40369: {'lr': 4.804431352810215e-05, 'samples': 20669440, 'steps': 40369, 'loss/train': 0.6975376009941101} 02/26/2022 05:50:54 - INFO - codeparrot_training - Step 40370: {'lr': 4.8034669508597726e-05, 'samples': 20669952, 'steps': 40370, 'loss/train': 1.8684865236282349} 02/26/2022 05:50:57 - INFO - codeparrot_training - Step 40371: {'lr': 4.802502635424863e-05, 'samples': 20670464, 'steps': 40371, 'loss/train': 2.1115877628326416} 02/26/2022 05:51:00 - INFO - codeparrot_training - Step 40372: {'lr': 4.801538406509612e-05, 'samples': 20670976, 'steps': 40372, 'loss/train': 1.3481011390686035} 02/26/2022 05:51:06 - INFO - codeparrot_training - Step 40373: {'lr': 4.8005742641181656e-05, 'samples': 20671488, 'steps': 40373, 'loss/train': 1.3121333122253418} 02/26/2022 05:51:09 - INFO - codeparrot_training - Step 40374: {'lr': 4.799610208254635e-05, 'samples': 20672000, 'steps': 40374, 'loss/train': 1.856300711631775} 02/26/2022 05:51:15 - INFO - codeparrot_training - Step 40375: {'lr': 4.7986462389231766e-05, 'samples': 20672512, 'steps': 40375, 'loss/train': 1.6852102279663086} 02/26/2022 05:51:19 - INFO - codeparrot_training - Step 40376: {'lr': 4.797682356127886e-05, 'samples': 20673024, 'steps': 40376, 'loss/train': 3.256267547607422} 02/26/2022 05:51:24 - INFO - codeparrot_training - Step 40377: {'lr': 4.79671855987292e-05, 'samples': 20673536, 'steps': 40377, 'loss/train': 1.380620002746582} 02/26/2022 05:51:27 - INFO - codeparrot_training - Step 40378: {'lr': 4.795754850162387e-05, 'samples': 20674048, 'steps': 40378, 'loss/train': 1.87704336643219} 02/26/2022 05:51:33 - INFO - codeparrot_training - Step 40379: {'lr': 4.794791227000439e-05, 'samples': 20674560, 'steps': 40379, 'loss/train': 0.9133546948432922} 02/26/2022 05:51:36 - INFO - codeparrot_training - Step 40380: {'lr': 4.7938276903911754e-05, 'samples': 20675072, 'steps': 40380, 'loss/train': 1.7998042106628418} 02/26/2022 05:51:44 - INFO - codeparrot_training - Step 40381: {'lr': 4.7928642403387456e-05, 'samples': 20675584, 'steps': 40381, 'loss/train': 1.4639952182769775} 02/26/2022 05:51:47 - INFO - codeparrot_training - Step 40382: {'lr': 4.791900876847261e-05, 'samples': 20676096, 'steps': 40382, 'loss/train': 1.407915472984314} 02/26/2022 05:51:53 - INFO - codeparrot_training - Step 40383: {'lr': 4.790937599920872e-05, 'samples': 20676608, 'steps': 40383, 'loss/train': 1.1989779472351074} 02/26/2022 05:51:56 - INFO - codeparrot_training - Step 40384: {'lr': 4.789974409563674e-05, 'samples': 20677120, 'steps': 40384, 'loss/train': 1.5867364406585693} 02/26/2022 05:52:02 - INFO - codeparrot_training - Step 40385: {'lr': 4.789011305779814e-05, 'samples': 20677632, 'steps': 40385, 'loss/train': 1.241894245147705} 02/26/2022 05:52:05 - INFO - codeparrot_training - Step 40386: {'lr': 4.788048288573407e-05, 'samples': 20678144, 'steps': 40386, 'loss/train': 0.5858150720596313} 02/26/2022 05:52:11 - INFO - codeparrot_training - Step 40387: {'lr': 4.7870853579485896e-05, 'samples': 20678656, 'steps': 40387, 'loss/train': 1.0563205480575562} 02/26/2022 05:52:14 - INFO - codeparrot_training - Step 40388: {'lr': 4.7861225139094774e-05, 'samples': 20679168, 'steps': 40388, 'loss/train': 1.2035837173461914} 02/26/2022 05:52:20 - INFO - codeparrot_training - Step 40389: {'lr': 4.7851597564602e-05, 'samples': 20679680, 'steps': 40389, 'loss/train': 1.666856288909912} 02/26/2022 05:52:23 - INFO - codeparrot_training - Step 40390: {'lr': 4.78419708560488e-05, 'samples': 20680192, 'steps': 40390, 'loss/train': 1.10679030418396} 02/26/2022 05:52:30 - INFO - codeparrot_training - Step 40391: {'lr': 4.783234501347633e-05, 'samples': 20680704, 'steps': 40391, 'loss/train': 0.051086753606796265} 02/26/2022 05:52:34 - INFO - codeparrot_training - Step 40392: {'lr': 4.7822720036925975e-05, 'samples': 20681216, 'steps': 40392, 'loss/train': 1.8546780347824097} 02/26/2022 05:52:39 - INFO - codeparrot_training - Step 40393: {'lr': 4.781309592643887e-05, 'samples': 20681728, 'steps': 40393, 'loss/train': 2.0365428924560547} 02/26/2022 05:52:45 - INFO - codeparrot_training - Step 40394: {'lr': 4.780347268205626e-05, 'samples': 20682240, 'steps': 40394, 'loss/train': 1.6743618249893188} 02/26/2022 05:52:48 - INFO - codeparrot_training - Step 40395: {'lr': 4.7793850303819334e-05, 'samples': 20682752, 'steps': 40395, 'loss/train': 0.9641406536102295} 02/26/2022 05:52:52 - INFO - codeparrot_training - Step 40396: {'lr': 4.7784228791769386e-05, 'samples': 20683264, 'steps': 40396, 'loss/train': 1.098073124885559} 02/26/2022 05:52:57 - INFO - codeparrot_training - Step 40397: {'lr': 4.777460814594758e-05, 'samples': 20683776, 'steps': 40397, 'loss/train': 0.37153273820877075} 02/26/2022 05:53:03 - INFO - codeparrot_training - Step 40398: {'lr': 4.776498836639515e-05, 'samples': 20684288, 'steps': 40398, 'loss/train': 1.2333284616470337} 02/26/2022 05:53:06 - INFO - codeparrot_training - Step 40399: {'lr': 4.775536945315323e-05, 'samples': 20684800, 'steps': 40399, 'loss/train': 0.9079501032829285} 02/26/2022 05:53:13 - INFO - codeparrot_training - Step 40400: {'lr': 4.7745751406263163e-05, 'samples': 20685312, 'steps': 40400, 'loss/train': 1.0404052734375} 02/26/2022 05:53:16 - INFO - codeparrot_training - Step 40401: {'lr': 4.773613422576606e-05, 'samples': 20685824, 'steps': 40401, 'loss/train': 1.9159491062164307} 02/26/2022 05:53:22 - INFO - codeparrot_training - Step 40402: {'lr': 4.7726517911703124e-05, 'samples': 20686336, 'steps': 40402, 'loss/train': 1.39814031124115} 02/26/2022 05:53:25 - INFO - codeparrot_training - Step 40403: {'lr': 4.7716902464115504e-05, 'samples': 20686848, 'steps': 40403, 'loss/train': 2.0115244388580322} 02/26/2022 05:53:31 - INFO - codeparrot_training - Step 40404: {'lr': 4.770728788304451e-05, 'samples': 20687360, 'steps': 40404, 'loss/train': 0.49342501163482666} 02/26/2022 05:53:34 - INFO - codeparrot_training - Step 40405: {'lr': 4.769767416853127e-05, 'samples': 20687872, 'steps': 40405, 'loss/train': 2.1932361125946045} 02/26/2022 05:53:40 - INFO - codeparrot_training - Step 40406: {'lr': 4.768806132061693e-05, 'samples': 20688384, 'steps': 40406, 'loss/train': 2.0098977088928223} 02/26/2022 05:53:43 - INFO - codeparrot_training - Step 40407: {'lr': 4.767844933934265e-05, 'samples': 20688896, 'steps': 40407, 'loss/train': 1.7678189277648926} 02/26/2022 05:53:47 - INFO - codeparrot_training - Step 40408: {'lr': 4.7668838224749714e-05, 'samples': 20689408, 'steps': 40408, 'loss/train': 1.4884628057479858} 02/26/2022 05:53:53 - INFO - codeparrot_training - Step 40409: {'lr': 4.7659227976879134e-05, 'samples': 20689920, 'steps': 40409, 'loss/train': 1.2438020706176758} 02/26/2022 05:53:56 - INFO - codeparrot_training - Step 40410: {'lr': 4.764961859577233e-05, 'samples': 20690432, 'steps': 40410, 'loss/train': 2.1484482288360596} 02/26/2022 05:54:02 - INFO - codeparrot_training - Step 40411: {'lr': 4.764001008147015e-05, 'samples': 20690944, 'steps': 40411, 'loss/train': 0.8591733574867249} 02/26/2022 05:54:05 - INFO - codeparrot_training - Step 40412: {'lr': 4.7630402434014006e-05, 'samples': 20691456, 'steps': 40412, 'loss/train': 1.1374257802963257} 02/26/2022 05:54:11 - INFO - codeparrot_training - Step 40413: {'lr': 4.762079565344488e-05, 'samples': 20691968, 'steps': 40413, 'loss/train': 0.7427809238433838} 02/26/2022 05:54:15 - INFO - codeparrot_training - Step 40414: {'lr': 4.761118973980413e-05, 'samples': 20692480, 'steps': 40414, 'loss/train': 0.22924913465976715} 02/26/2022 05:54:20 - INFO - codeparrot_training - Step 40415: {'lr': 4.7601584693132636e-05, 'samples': 20692992, 'steps': 40415, 'loss/train': 1.5678691864013672} 02/26/2022 05:54:24 - INFO - codeparrot_training - Step 40416: {'lr': 4.759198051347177e-05, 'samples': 20693504, 'steps': 40416, 'loss/train': 1.2099350690841675} 02/26/2022 05:54:29 - INFO - codeparrot_training - Step 40417: {'lr': 4.7582377200862504e-05, 'samples': 20694016, 'steps': 40417, 'loss/train': 1.4026752710342407} 02/26/2022 05:54:33 - INFO - codeparrot_training - Step 40418: {'lr': 4.757277475534621e-05, 'samples': 20694528, 'steps': 40418, 'loss/train': 1.8972070217132568} 02/26/2022 05:54:38 - INFO - codeparrot_training - Step 40419: {'lr': 4.7563173176963734e-05, 'samples': 20695040, 'steps': 40419, 'loss/train': 2.4594573974609375} 02/26/2022 05:54:42 - INFO - codeparrot_training - Step 40420: {'lr': 4.7553572465756396e-05, 'samples': 20695552, 'steps': 40420, 'loss/train': 1.88540518283844} 02/26/2022 05:54:47 - INFO - codeparrot_training - Step 40421: {'lr': 4.7543972621765213e-05, 'samples': 20696064, 'steps': 40421, 'loss/train': 0.7198459506034851} 02/26/2022 05:54:51 - INFO - codeparrot_training - Step 40422: {'lr': 4.7534373645031435e-05, 'samples': 20696576, 'steps': 40422, 'loss/train': 1.031506896018982} 02/26/2022 05:54:56 - INFO - codeparrot_training - Step 40423: {'lr': 4.752477553559612e-05, 'samples': 20697088, 'steps': 40423, 'loss/train': 2.101296901702881} 02/26/2022 05:55:00 - INFO - codeparrot_training - Step 40424: {'lr': 4.7515178293500354e-05, 'samples': 20697600, 'steps': 40424, 'loss/train': 1.5176054239273071} 02/26/2022 05:55:05 - INFO - codeparrot_training - Step 40425: {'lr': 4.7505581918785206e-05, 'samples': 20698112, 'steps': 40425, 'loss/train': 1.1849312782287598} 02/26/2022 05:55:09 - INFO - codeparrot_training - Step 40426: {'lr': 4.7495986411491915e-05, 'samples': 20698624, 'steps': 40426, 'loss/train': 2.3570902347564697} 02/26/2022 05:55:15 - INFO - codeparrot_training - Step 40427: {'lr': 4.7486391771661505e-05, 'samples': 20699136, 'steps': 40427, 'loss/train': 1.1125259399414062} 02/26/2022 05:55:18 - INFO - codeparrot_training - Step 40428: {'lr': 4.74767979993351e-05, 'samples': 20699648, 'steps': 40428, 'loss/train': 2.4073944091796875} 02/26/2022 05:55:24 - INFO - codeparrot_training - Step 40429: {'lr': 4.746720509455371e-05, 'samples': 20700160, 'steps': 40429, 'loss/train': 2.1372411251068115} 02/26/2022 05:55:27 - INFO - codeparrot_training - Step 40430: {'lr': 4.745761305735857e-05, 'samples': 20700672, 'steps': 40430, 'loss/train': 1.3745031356811523} 02/26/2022 05:55:33 - INFO - codeparrot_training - Step 40431: {'lr': 4.744802188779071e-05, 'samples': 20701184, 'steps': 40431, 'loss/train': 1.794533371925354} 02/26/2022 05:55:37 - INFO - codeparrot_training - Step 40432: {'lr': 4.74384315858912e-05, 'samples': 20701696, 'steps': 40432, 'loss/train': 1.7580296993255615} 02/26/2022 05:55:42 - INFO - codeparrot_training - Step 40433: {'lr': 4.742884215170107e-05, 'samples': 20702208, 'steps': 40433, 'loss/train': 1.7651078701019287} 02/26/2022 05:55:46 - INFO - codeparrot_training - Step 40434: {'lr': 4.74192535852615e-05, 'samples': 20702720, 'steps': 40434, 'loss/train': 0.9584648013114929} 02/26/2022 05:55:51 - INFO - codeparrot_training - Step 40435: {'lr': 4.740966588661355e-05, 'samples': 20703232, 'steps': 40435, 'loss/train': 1.8396974802017212} 02/26/2022 05:55:55 - INFO - codeparrot_training - Step 40436: {'lr': 4.740007905579824e-05, 'samples': 20703744, 'steps': 40436, 'loss/train': 2.1046106815338135} 02/26/2022 05:56:00 - INFO - codeparrot_training - Step 40437: {'lr': 4.739049309285667e-05, 'samples': 20704256, 'steps': 40437, 'loss/train': 2.063457727432251} 02/26/2022 05:56:04 - INFO - codeparrot_training - Step 40438: {'lr': 4.738090799782982e-05, 'samples': 20704768, 'steps': 40438, 'loss/train': 2.6092586517333984} 02/26/2022 05:56:10 - INFO - codeparrot_training - Step 40439: {'lr': 4.737132377075889e-05, 'samples': 20705280, 'steps': 40439, 'loss/train': 1.60270094871521} 02/26/2022 05:56:13 - INFO - codeparrot_training - Step 40440: {'lr': 4.736174041168487e-05, 'samples': 20705792, 'steps': 40440, 'loss/train': 1.2395126819610596} 02/26/2022 05:56:19 - INFO - codeparrot_training - Step 40441: {'lr': 4.735215792064882e-05, 'samples': 20706304, 'steps': 40441, 'loss/train': 1.4732745885849} 02/26/2022 05:56:22 - INFO - codeparrot_training - Step 40442: {'lr': 4.734257629769168e-05, 'samples': 20706816, 'steps': 40442, 'loss/train': 1.563485860824585} 02/26/2022 05:56:28 - INFO - codeparrot_training - Step 40443: {'lr': 4.733299554285467e-05, 'samples': 20707328, 'steps': 40443, 'loss/train': 2.1985204219818115} 02/26/2022 05:56:31 - INFO - codeparrot_training - Step 40444: {'lr': 4.73234156561787e-05, 'samples': 20707840, 'steps': 40444, 'loss/train': 0.5319942831993103} 02/26/2022 05:56:37 - INFO - codeparrot_training - Step 40445: {'lr': 4.7313836637704996e-05, 'samples': 20708352, 'steps': 40445, 'loss/train': 2.4166722297668457} 02/26/2022 05:56:40 - INFO - codeparrot_training - Step 40446: {'lr': 4.73042584874743e-05, 'samples': 20708864, 'steps': 40446, 'loss/train': 2.171186923980713} 02/26/2022 05:56:46 - INFO - codeparrot_training - Step 40447: {'lr': 4.729468120552788e-05, 'samples': 20709376, 'steps': 40447, 'loss/train': 1.713104248046875} 02/26/2022 05:56:49 - INFO - codeparrot_training - Step 40448: {'lr': 4.7285104791906617e-05, 'samples': 20709888, 'steps': 40448, 'loss/train': 1.9176017045974731} 02/26/2022 05:56:56 - INFO - codeparrot_training - Step 40449: {'lr': 4.727552924665171e-05, 'samples': 20710400, 'steps': 40449, 'loss/train': 1.979875922203064} 02/26/2022 05:57:01 - INFO - codeparrot_training - Step 40450: {'lr': 4.7265954569803955e-05, 'samples': 20710912, 'steps': 40450, 'loss/train': 3.5115139484405518} 02/26/2022 05:57:05 - INFO - codeparrot_training - Step 40451: {'lr': 4.725638076140451e-05, 'samples': 20711424, 'steps': 40451, 'loss/train': 1.6537821292877197} 02/26/2022 05:57:10 - INFO - codeparrot_training - Step 40452: {'lr': 4.7246807821494317e-05, 'samples': 20711936, 'steps': 40452, 'loss/train': 1.6634182929992676} 02/26/2022 05:57:14 - INFO - codeparrot_training - Step 40453: {'lr': 4.723723575011454e-05, 'samples': 20712448, 'steps': 40453, 'loss/train': 1.9879093170166016} 02/26/2022 05:57:19 - INFO - codeparrot_training - Step 40454: {'lr': 4.7227664547305924e-05, 'samples': 20712960, 'steps': 40454, 'loss/train': 1.7678773403167725} 02/26/2022 05:57:23 - INFO - codeparrot_training - Step 40455: {'lr': 4.721809421310966e-05, 'samples': 20713472, 'steps': 40455, 'loss/train': 1.6923011541366577} 02/26/2022 05:57:28 - INFO - codeparrot_training - Step 40456: {'lr': 4.720852474756665e-05, 'samples': 20713984, 'steps': 40456, 'loss/train': 1.6868033409118652} 02/26/2022 05:57:32 - INFO - codeparrot_training - Step 40457: {'lr': 4.719895615071798e-05, 'samples': 20714496, 'steps': 40457, 'loss/train': 1.614898920059204} 02/26/2022 05:57:35 - INFO - codeparrot_training - Step 40458: {'lr': 4.718938842260459e-05, 'samples': 20715008, 'steps': 40458, 'loss/train': 1.5221036672592163} 02/26/2022 05:57:41 - INFO - codeparrot_training - Step 40459: {'lr': 4.7179821563267446e-05, 'samples': 20715520, 'steps': 40459, 'loss/train': 1.9841527938842773} 02/26/2022 05:57:45 - INFO - codeparrot_training - Step 40460: {'lr': 4.7170255572747485e-05, 'samples': 20716032, 'steps': 40460, 'loss/train': 1.0592180490493774} 02/26/2022 05:57:50 - INFO - codeparrot_training - Step 40461: {'lr': 4.7160690451085814e-05, 'samples': 20716544, 'steps': 40461, 'loss/train': 1.9526796340942383} 02/26/2022 05:57:54 - INFO - codeparrot_training - Step 40462: {'lr': 4.715112619832335e-05, 'samples': 20717056, 'steps': 40462, 'loss/train': 0.7734876871109009} 02/26/2022 05:57:59 - INFO - codeparrot_training - Step 40463: {'lr': 4.714156281450102e-05, 'samples': 20717568, 'steps': 40463, 'loss/train': 1.2418992519378662} 02/26/2022 05:58:06 - INFO - codeparrot_training - Step 40464: {'lr': 4.7132000299659774e-05, 'samples': 20718080, 'steps': 40464, 'loss/train': 1.4783622026443481} 02/26/2022 05:58:09 - INFO - codeparrot_training - Step 40465: {'lr': 4.712243865384067e-05, 'samples': 20718592, 'steps': 40465, 'loss/train': 1.2277157306671143} 02/26/2022 05:58:13 - INFO - codeparrot_training - Step 40466: {'lr': 4.7112877877084624e-05, 'samples': 20719104, 'steps': 40466, 'loss/train': 0.5284483432769775} 02/26/2022 05:58:18 - INFO - codeparrot_training - Step 40467: {'lr': 4.7103317969432596e-05, 'samples': 20719616, 'steps': 40467, 'loss/train': 1.923523187637329} 02/26/2022 05:58:22 - INFO - codeparrot_training - Step 40468: {'lr': 4.709375893092546e-05, 'samples': 20720128, 'steps': 40468, 'loss/train': 1.4221651554107666} 02/26/2022 05:58:27 - INFO - codeparrot_training - Step 40469: {'lr': 4.708420076160427e-05, 'samples': 20720640, 'steps': 40469, 'loss/train': 1.3970476388931274} 02/26/2022 05:58:31 - INFO - codeparrot_training - Step 40470: {'lr': 4.707464346150997e-05, 'samples': 20721152, 'steps': 40470, 'loss/train': 1.513027548789978} 02/26/2022 05:58:36 - INFO - codeparrot_training - Step 40471: {'lr': 4.706508703068343e-05, 'samples': 20721664, 'steps': 40471, 'loss/train': 1.8742296695709229} 02/26/2022 05:58:40 - INFO - codeparrot_training - Step 40472: {'lr': 4.705553146916558e-05, 'samples': 20722176, 'steps': 40472, 'loss/train': 1.8326905965805054} 02/26/2022 05:58:46 - INFO - codeparrot_training - Step 40473: {'lr': 4.704597677699743e-05, 'samples': 20722688, 'steps': 40473, 'loss/train': 1.5975911617279053} 02/26/2022 05:58:50 - INFO - codeparrot_training - Step 40474: {'lr': 4.7036422954219894e-05, 'samples': 20723200, 'steps': 40474, 'loss/train': 0.06833074986934662} 02/26/2022 05:58:55 - INFO - codeparrot_training - Step 40475: {'lr': 4.702687000087385e-05, 'samples': 20723712, 'steps': 40475, 'loss/train': 1.5726796388626099} 02/26/2022 05:58:59 - INFO - codeparrot_training - Step 40476: {'lr': 4.70173179170002e-05, 'samples': 20724224, 'steps': 40476, 'loss/train': 1.153774380683899} 02/26/2022 05:59:04 - INFO - codeparrot_training - Step 40477: {'lr': 4.700776670263995e-05, 'samples': 20724736, 'steps': 40477, 'loss/train': 2.6383652687072754} 02/26/2022 05:59:08 - INFO - codeparrot_training - Step 40478: {'lr': 4.699821635783399e-05, 'samples': 20725248, 'steps': 40478, 'loss/train': 1.3899964094161987} 02/26/2022 05:59:13 - INFO - codeparrot_training - Step 40479: {'lr': 4.698866688262321e-05, 'samples': 20725760, 'steps': 40479, 'loss/train': 0.6454765200614929} 02/26/2022 05:59:17 - INFO - codeparrot_training - Step 40480: {'lr': 4.6979118277048426e-05, 'samples': 20726272, 'steps': 40480, 'loss/train': 1.4977879524230957} 02/26/2022 05:59:22 - INFO - codeparrot_training - Step 40481: {'lr': 4.6969570541150725e-05, 'samples': 20726784, 'steps': 40481, 'loss/train': 2.570521831512451} 02/26/2022 05:59:26 - INFO - codeparrot_training - Step 40482: {'lr': 4.69600236749709e-05, 'samples': 20727296, 'steps': 40482, 'loss/train': 0.8650080561637878} 02/26/2022 05:59:32 - INFO - codeparrot_training - Step 40483: {'lr': 4.695047767854982e-05, 'samples': 20727808, 'steps': 40483, 'loss/train': 1.0418413877487183} 02/26/2022 05:59:35 - INFO - codeparrot_training - Step 40484: {'lr': 4.694093255192847e-05, 'samples': 20728320, 'steps': 40484, 'loss/train': 1.1293816566467285} 02/26/2022 05:59:41 - INFO - codeparrot_training - Step 40485: {'lr': 4.693138829514768e-05, 'samples': 20728832, 'steps': 40485, 'loss/train': 1.1596565246582031} 02/26/2022 05:59:44 - INFO - codeparrot_training - Step 40486: {'lr': 4.6921844908248326e-05, 'samples': 20729344, 'steps': 40486, 'loss/train': 2.1983063220977783} 02/26/2022 05:59:50 - INFO - codeparrot_training - Step 40487: {'lr': 4.691230239127126e-05, 'samples': 20729856, 'steps': 40487, 'loss/train': 2.0347418785095215} 02/26/2022 05:59:53 - INFO - codeparrot_training - Step 40488: {'lr': 4.690276074425753e-05, 'samples': 20730368, 'steps': 40488, 'loss/train': 0.9864697456359863} 02/26/2022 05:59:59 - INFO - codeparrot_training - Step 40489: {'lr': 4.689321996724777e-05, 'samples': 20730880, 'steps': 40489, 'loss/train': 1.7229087352752686} 02/26/2022 06:00:02 - INFO - codeparrot_training - Step 40490: {'lr': 4.688368006028299e-05, 'samples': 20731392, 'steps': 40490, 'loss/train': 2.7362678050994873} 02/26/2022 06:00:08 - INFO - codeparrot_training - Step 40491: {'lr': 4.687414102340398e-05, 'samples': 20731904, 'steps': 40491, 'loss/train': 0.0761817991733551} 02/26/2022 06:00:11 - INFO - codeparrot_training - Step 40492: {'lr': 4.686460285665173e-05, 'samples': 20732416, 'steps': 40492, 'loss/train': 1.157053828239441} 02/26/2022 06:00:17 - INFO - codeparrot_training - Step 40493: {'lr': 4.6855065560066996e-05, 'samples': 20732928, 'steps': 40493, 'loss/train': 1.4385110139846802} 02/26/2022 06:00:20 - INFO - codeparrot_training - Step 40494: {'lr': 4.684552913369067e-05, 'samples': 20733440, 'steps': 40494, 'loss/train': 1.5717777013778687} 02/26/2022 06:00:26 - INFO - codeparrot_training - Step 40495: {'lr': 4.683599357756352e-05, 'samples': 20733952, 'steps': 40495, 'loss/train': 1.8374850749969482} 02/26/2022 06:00:30 - INFO - codeparrot_training - Step 40496: {'lr': 4.6826458891726513e-05, 'samples': 20734464, 'steps': 40496, 'loss/train': 0.260576069355011} 02/26/2022 06:00:35 - INFO - codeparrot_training - Step 40497: {'lr': 4.6816925076220454e-05, 'samples': 20734976, 'steps': 40497, 'loss/train': 0.3610433340072632} 02/26/2022 06:00:39 - INFO - codeparrot_training - Step 40498: {'lr': 4.6807392131086175e-05, 'samples': 20735488, 'steps': 40498, 'loss/train': 6.080682277679443} 02/26/2022 06:00:44 - INFO - codeparrot_training - Step 40499: {'lr': 4.679786005636444e-05, 'samples': 20736000, 'steps': 40499, 'loss/train': 0.0848667174577713} 02/26/2022 06:00:48 - INFO - codeparrot_training - Step 40500: {'lr': 4.6788328852096216e-05, 'samples': 20736512, 'steps': 40500, 'loss/train': 1.1825543642044067} 02/26/2022 06:00:53 - INFO - codeparrot_training - Step 40501: {'lr': 4.677879851832226e-05, 'samples': 20737024, 'steps': 40501, 'loss/train': 2.076835870742798} 02/26/2022 06:00:57 - INFO - codeparrot_training - Step 40502: {'lr': 4.676926905508339e-05, 'samples': 20737536, 'steps': 40502, 'loss/train': 1.152147889137268} 02/26/2022 06:01:02 - INFO - codeparrot_training - Step 40503: {'lr': 4.675974046242037e-05, 'samples': 20738048, 'steps': 40503, 'loss/train': 1.346023440361023} 02/26/2022 06:01:06 - INFO - codeparrot_training - Step 40504: {'lr': 4.675021274037416e-05, 'samples': 20738560, 'steps': 40504, 'loss/train': 1.8701070547103882} 02/26/2022 06:01:12 - INFO - codeparrot_training - Step 40505: {'lr': 4.674068588898545e-05, 'samples': 20739072, 'steps': 40505, 'loss/train': 0.7971094250679016} 02/26/2022 06:01:16 - INFO - codeparrot_training - Step 40506: {'lr': 4.6731159908295214e-05, 'samples': 20739584, 'steps': 40506, 'loss/train': 1.0658234357833862} 02/26/2022 06:01:21 - INFO - codeparrot_training - Step 40507: {'lr': 4.672163479834401e-05, 'samples': 20740096, 'steps': 40507, 'loss/train': 1.5536221265792847} 02/26/2022 06:01:25 - INFO - codeparrot_training - Step 40508: {'lr': 4.671211055917285e-05, 'samples': 20740608, 'steps': 40508, 'loss/train': 2.74406361579895} 02/26/2022 06:01:30 - INFO - codeparrot_training - Step 40509: {'lr': 4.67025871908224e-05, 'samples': 20741120, 'steps': 40509, 'loss/train': 1.5952783823013306} 02/26/2022 06:01:34 - INFO - codeparrot_training - Step 40510: {'lr': 4.669306469333362e-05, 'samples': 20741632, 'steps': 40510, 'loss/train': 2.4164438247680664} 02/26/2022 06:01:39 - INFO - codeparrot_training - Step 40511: {'lr': 4.6683543066747076e-05, 'samples': 20742144, 'steps': 40511, 'loss/train': 1.5712264776229858} 02/26/2022 06:01:43 - INFO - codeparrot_training - Step 40512: {'lr': 4.667402231110374e-05, 'samples': 20742656, 'steps': 40512, 'loss/train': 1.611841082572937} 02/26/2022 06:01:48 - INFO - codeparrot_training - Step 40513: {'lr': 4.6664502426444266e-05, 'samples': 20743168, 'steps': 40513, 'loss/train': 2.0886714458465576} 02/26/2022 06:01:52 - INFO - codeparrot_training - Step 40514: {'lr': 4.665498341280963e-05, 'samples': 20743680, 'steps': 40514, 'loss/train': 2.2555325031280518} 02/26/2022 06:01:57 - INFO - codeparrot_training - Step 40515: {'lr': 4.664546527024033e-05, 'samples': 20744192, 'steps': 40515, 'loss/train': 0.9298934936523438} 02/26/2022 06:02:01 - INFO - codeparrot_training - Step 40516: {'lr': 4.663594799877735e-05, 'samples': 20744704, 'steps': 40516, 'loss/train': 1.6815675497055054} 02/26/2022 06:02:07 - INFO - codeparrot_training - Step 40517: {'lr': 4.662643159846133e-05, 'samples': 20745216, 'steps': 40517, 'loss/train': 2.065317153930664} 02/26/2022 06:02:11 - INFO - codeparrot_training - Step 40518: {'lr': 4.6616916069333185e-05, 'samples': 20745728, 'steps': 40518, 'loss/train': 3.2296125888824463} 02/26/2022 06:02:16 - INFO - codeparrot_training - Step 40519: {'lr': 4.6607401411433575e-05, 'samples': 20746240, 'steps': 40519, 'loss/train': 2.0922257900238037} 02/26/2022 06:02:20 - INFO - codeparrot_training - Step 40520: {'lr': 4.659788762480327e-05, 'samples': 20746752, 'steps': 40520, 'loss/train': 1.606555461883545} 02/26/2022 06:02:25 - INFO - codeparrot_training - Step 40521: {'lr': 4.6588374709483014e-05, 'samples': 20747264, 'steps': 40521, 'loss/train': 2.468932867050171} 02/26/2022 06:02:29 - INFO - codeparrot_training - Step 40522: {'lr': 4.6578862665513534e-05, 'samples': 20747776, 'steps': 40522, 'loss/train': 2.378203868865967} 02/26/2022 06:02:34 - INFO - codeparrot_training - Step 40523: {'lr': 4.6569351492935676e-05, 'samples': 20748288, 'steps': 40523, 'loss/train': 1.4238327741622925} 02/26/2022 06:02:38 - INFO - codeparrot_training - Step 40524: {'lr': 4.6559841191790106e-05, 'samples': 20748800, 'steps': 40524, 'loss/train': 2.251905918121338} 02/26/2022 06:02:43 - INFO - codeparrot_training - Step 40525: {'lr': 4.655033176211757e-05, 'samples': 20749312, 'steps': 40525, 'loss/train': 1.4019033908843994} 02/26/2022 06:02:47 - INFO - codeparrot_training - Step 40526: {'lr': 4.654082320395875e-05, 'samples': 20749824, 'steps': 40526, 'loss/train': 1.7638624906539917} 02/26/2022 06:02:52 - INFO - codeparrot_training - Step 40527: {'lr': 4.653131551735451e-05, 'samples': 20750336, 'steps': 40527, 'loss/train': 1.9404500722885132} 02/26/2022 06:02:56 - INFO - codeparrot_training - Step 40528: {'lr': 4.652180870234551e-05, 'samples': 20750848, 'steps': 40528, 'loss/train': 1.8244398832321167} 02/26/2022 06:03:02 - INFO - codeparrot_training - Step 40529: {'lr': 4.6512302758972445e-05, 'samples': 20751360, 'steps': 40529, 'loss/train': 1.048850178718567} 02/26/2022 06:03:08 - INFO - codeparrot_training - Step 40530: {'lr': 4.6502797687276025e-05, 'samples': 20751872, 'steps': 40530, 'loss/train': 1.8237043619155884} 02/26/2022 06:03:11 - INFO - codeparrot_training - Step 40531: {'lr': 4.649329348729706e-05, 'samples': 20752384, 'steps': 40531, 'loss/train': 0.8328178524971008} 02/26/2022 06:03:15 - INFO - codeparrot_training - Step 40532: {'lr': 4.648379015907619e-05, 'samples': 20752896, 'steps': 40532, 'loss/train': 0.9888048768043518} 02/26/2022 06:03:20 - INFO - codeparrot_training - Step 40533: {'lr': 4.647428770265416e-05, 'samples': 20753408, 'steps': 40533, 'loss/train': 0.6856012940406799} 02/26/2022 06:03:26 - INFO - codeparrot_training - Step 40534: {'lr': 4.646478611807156e-05, 'samples': 20753920, 'steps': 40534, 'loss/train': 2.7985596656799316} 02/26/2022 06:03:29 - INFO - codeparrot_training - Step 40535: {'lr': 4.645528540536928e-05, 'samples': 20754432, 'steps': 40535, 'loss/train': 1.9000080823898315} 02/26/2022 06:03:35 - INFO - codeparrot_training - Step 40536: {'lr': 4.644578556458792e-05, 'samples': 20754944, 'steps': 40536, 'loss/train': 2.3702783584594727} 02/26/2022 06:03:38 - INFO - codeparrot_training - Step 40537: {'lr': 4.643628659576818e-05, 'samples': 20755456, 'steps': 40537, 'loss/train': 1.701398491859436} 02/26/2022 06:03:44 - INFO - codeparrot_training - Step 40538: {'lr': 4.642678849895068e-05, 'samples': 20755968, 'steps': 40538, 'loss/train': 2.4501144886016846} 02/26/2022 06:03:47 - INFO - codeparrot_training - Step 40539: {'lr': 4.641729127417624e-05, 'samples': 20756480, 'steps': 40539, 'loss/train': 0.7281582951545715} 02/26/2022 06:03:53 - INFO - codeparrot_training - Step 40540: {'lr': 4.640779492148547e-05, 'samples': 20756992, 'steps': 40540, 'loss/train': 0.8816997408866882} 02/26/2022 06:03:57 - INFO - codeparrot_training - Step 40541: {'lr': 4.639829944091905e-05, 'samples': 20757504, 'steps': 40541, 'loss/train': 2.528623580932617} 02/26/2022 06:04:02 - INFO - codeparrot_training - Step 40542: {'lr': 4.6388804832517645e-05, 'samples': 20758016, 'steps': 40542, 'loss/train': 1.9930883646011353} 02/26/2022 06:04:06 - INFO - codeparrot_training - Step 40543: {'lr': 4.6379311096321986e-05, 'samples': 20758528, 'steps': 40543, 'loss/train': 1.8037618398666382} 02/26/2022 06:04:11 - INFO - codeparrot_training - Step 40544: {'lr': 4.636981823237263e-05, 'samples': 20759040, 'steps': 40544, 'loss/train': 1.8401269912719727} 02/26/2022 06:04:15 - INFO - codeparrot_training - Step 40545: {'lr': 4.6360326240710435e-05, 'samples': 20759552, 'steps': 40545, 'loss/train': 1.8866084814071655} 02/26/2022 06:04:20 - INFO - codeparrot_training - Step 40546: {'lr': 4.6350835121375836e-05, 'samples': 20760064, 'steps': 40546, 'loss/train': 1.5532060861587524} 02/26/2022 06:04:24 - INFO - codeparrot_training - Step 40547: {'lr': 4.634134487440964e-05, 'samples': 20760576, 'steps': 40547, 'loss/train': 1.2604817152023315} 02/26/2022 06:04:29 - INFO - codeparrot_training - Step 40548: {'lr': 4.633185549985239e-05, 'samples': 20761088, 'steps': 40548, 'loss/train': 1.6716194152832031} 02/26/2022 06:04:33 - INFO - codeparrot_training - Step 40549: {'lr': 4.632236699774492e-05, 'samples': 20761600, 'steps': 40549, 'loss/train': 2.2486982345581055} 02/26/2022 06:04:38 - INFO - codeparrot_training - Step 40550: {'lr': 4.6312879368127645e-05, 'samples': 20762112, 'steps': 40550, 'loss/train': 0.9296627044677734} 02/26/2022 06:04:42 - INFO - codeparrot_training - Step 40551: {'lr': 4.630339261104138e-05, 'samples': 20762624, 'steps': 40551, 'loss/train': 1.1658196449279785} 02/26/2022 06:04:48 - INFO - codeparrot_training - Step 40552: {'lr': 4.629390672652661e-05, 'samples': 20763136, 'steps': 40552, 'loss/train': 0.9665114283561707} 02/26/2022 06:04:51 - INFO - codeparrot_training - Step 40553: {'lr': 4.628442171462416e-05, 'samples': 20763648, 'steps': 40553, 'loss/train': 0.6147657036781311} 02/26/2022 06:04:57 - INFO - codeparrot_training - Step 40554: {'lr': 4.627493757537454e-05, 'samples': 20764160, 'steps': 40554, 'loss/train': 1.8051081895828247} 02/26/2022 06:05:00 - INFO - codeparrot_training - Step 40555: {'lr': 4.62654543088184e-05, 'samples': 20764672, 'steps': 40555, 'loss/train': 1.947871446609497} 02/26/2022 06:05:06 - INFO - codeparrot_training - Step 40556: {'lr': 4.625597191499631e-05, 'samples': 20765184, 'steps': 40556, 'loss/train': 1.5963685512542725} 02/26/2022 06:05:09 - INFO - codeparrot_training - Step 40557: {'lr': 4.6246490393948984e-05, 'samples': 20765696, 'steps': 40557, 'loss/train': 1.5938471555709839} 02/26/2022 06:05:15 - INFO - codeparrot_training - Step 40558: {'lr': 4.6237009745717015e-05, 'samples': 20766208, 'steps': 40558, 'loss/train': 5.4820146560668945} 02/26/2022 06:05:18 - INFO - codeparrot_training - Step 40559: {'lr': 4.6227529970340965e-05, 'samples': 20766720, 'steps': 40559, 'loss/train': 2.2246623039245605} 02/26/2022 06:05:24 - INFO - codeparrot_training - Step 40560: {'lr': 4.621805106786142e-05, 'samples': 20767232, 'steps': 40560, 'loss/train': 1.1664386987686157} 02/26/2022 06:05:27 - INFO - codeparrot_training - Step 40561: {'lr': 4.62085730383191e-05, 'samples': 20767744, 'steps': 40561, 'loss/train': 0.8692867159843445} 02/26/2022 06:05:34 - INFO - codeparrot_training - Step 40562: {'lr': 4.619909588175455e-05, 'samples': 20768256, 'steps': 40562, 'loss/train': 1.664820909500122} 02/26/2022 06:05:37 - INFO - codeparrot_training - Step 40563: {'lr': 4.618961959820836e-05, 'samples': 20768768, 'steps': 40563, 'loss/train': 1.2737410068511963} 02/26/2022 06:05:42 - INFO - codeparrot_training - Step 40564: {'lr': 4.618014418772104e-05, 'samples': 20769280, 'steps': 40564, 'loss/train': 1.7915550470352173} 02/26/2022 06:05:46 - INFO - codeparrot_training - Step 40565: {'lr': 4.617066965033334e-05, 'samples': 20769792, 'steps': 40565, 'loss/train': 1.1754978895187378} 02/26/2022 06:05:52 - INFO - codeparrot_training - Step 40566: {'lr': 4.6161195986085787e-05, 'samples': 20770304, 'steps': 40566, 'loss/train': 1.2529526948928833} 02/26/2022 06:05:55 - INFO - codeparrot_training - Step 40567: {'lr': 4.6151723195018925e-05, 'samples': 20770816, 'steps': 40567, 'loss/train': 2.1452362537384033} 02/26/2022 06:06:01 - INFO - codeparrot_training - Step 40568: {'lr': 4.614225127717334e-05, 'samples': 20771328, 'steps': 40568, 'loss/train': 1.314723253250122} 02/26/2022 06:06:04 - INFO - codeparrot_training - Step 40569: {'lr': 4.6132780232589574e-05, 'samples': 20771840, 'steps': 40569, 'loss/train': 1.7965604066848755} 02/26/2022 06:06:09 - INFO - codeparrot_training - Step 40570: {'lr': 4.612331006130829e-05, 'samples': 20772352, 'steps': 40570, 'loss/train': 0.4550262987613678} 02/26/2022 06:06:13 - INFO - codeparrot_training - Step 40571: {'lr': 4.611384076337003e-05, 'samples': 20772864, 'steps': 40571, 'loss/train': 2.019219398498535} 02/26/2022 06:06:18 - INFO - codeparrot_training - Step 40572: {'lr': 4.6104372338815315e-05, 'samples': 20773376, 'steps': 40572, 'loss/train': 2.0042831897735596} 02/26/2022 06:06:22 - INFO - codeparrot_training - Step 40573: {'lr': 4.609490478768466e-05, 'samples': 20773888, 'steps': 40573, 'loss/train': 1.4293339252471924} 02/26/2022 06:06:27 - INFO - codeparrot_training - Step 40574: {'lr': 4.608543811001875e-05, 'samples': 20774400, 'steps': 40574, 'loss/train': 1.8459911346435547} 02/26/2022 06:06:31 - INFO - codeparrot_training - Step 40575: {'lr': 4.607597230585808e-05, 'samples': 20774912, 'steps': 40575, 'loss/train': 1.714013695716858} 02/26/2022 06:06:37 - INFO - codeparrot_training - Step 40576: {'lr': 4.606650737524321e-05, 'samples': 20775424, 'steps': 40576, 'loss/train': 1.852400302886963} 02/26/2022 06:06:41 - INFO - codeparrot_training - Step 40577: {'lr': 4.6057043318214585e-05, 'samples': 20775936, 'steps': 40577, 'loss/train': 1.2687244415283203} 02/26/2022 06:06:46 - INFO - codeparrot_training - Step 40578: {'lr': 4.604758013481289e-05, 'samples': 20776448, 'steps': 40578, 'loss/train': 0.03738047555088997} 02/26/2022 06:06:50 - INFO - codeparrot_training - Step 40579: {'lr': 4.6038117825078526e-05, 'samples': 20776960, 'steps': 40579, 'loss/train': 2.158421754837036} 02/26/2022 06:06:55 - INFO - codeparrot_training - Step 40580: {'lr': 4.6028656389052236e-05, 'samples': 20777472, 'steps': 40580, 'loss/train': 2.041349172592163} 02/26/2022 06:06:59 - INFO - codeparrot_training - Step 40581: {'lr': 4.60191958267743e-05, 'samples': 20777984, 'steps': 40581, 'loss/train': 1.8198038339614868} 02/26/2022 06:07:04 - INFO - codeparrot_training - Step 40582: {'lr': 4.6009736138285406e-05, 'samples': 20778496, 'steps': 40582, 'loss/train': 3.0360329151153564} 02/26/2022 06:07:08 - INFO - codeparrot_training - Step 40583: {'lr': 4.6000277323625964e-05, 'samples': 20779008, 'steps': 40583, 'loss/train': 1.5629966259002686} 02/26/2022 06:07:13 - INFO - codeparrot_training - Step 40584: {'lr': 4.59908193828367e-05, 'samples': 20779520, 'steps': 40584, 'loss/train': 1.730132818222046} 02/26/2022 06:07:17 - INFO - codeparrot_training - Step 40585: {'lr': 4.598136231595784e-05, 'samples': 20780032, 'steps': 40585, 'loss/train': 1.7915922403335571} 02/26/2022 06:07:22 - INFO - codeparrot_training - Step 40586: {'lr': 4.5971906123030095e-05, 'samples': 20780544, 'steps': 40586, 'loss/train': 2.3261685371398926} 02/26/2022 06:07:26 - INFO - codeparrot_training - Step 40587: {'lr': 4.596245080409386e-05, 'samples': 20781056, 'steps': 40587, 'loss/train': 0.7651938199996948} 02/26/2022 06:07:32 - INFO - codeparrot_training - Step 40588: {'lr': 4.595299635918984e-05, 'samples': 20781568, 'steps': 40588, 'loss/train': 1.8871077299118042} 02/26/2022 06:07:35 - INFO - codeparrot_training - Step 40589: {'lr': 4.594354278835827e-05, 'samples': 20782080, 'steps': 40589, 'loss/train': 1.2763116359710693} 02/26/2022 06:07:41 - INFO - codeparrot_training - Step 40590: {'lr': 4.593409009163982e-05, 'samples': 20782592, 'steps': 40590, 'loss/train': 1.8814716339111328} 02/26/2022 06:07:45 - INFO - codeparrot_training - Step 40591: {'lr': 4.5924638269074864e-05, 'samples': 20783104, 'steps': 40591, 'loss/train': 1.482607364654541} 02/26/2022 06:07:50 - INFO - codeparrot_training - Step 40592: {'lr': 4.5915187320704016e-05, 'samples': 20783616, 'steps': 40592, 'loss/train': 1.1465833187103271} 02/26/2022 06:07:54 - INFO - codeparrot_training - Step 40593: {'lr': 4.590573724656771e-05, 'samples': 20784128, 'steps': 40593, 'loss/train': 1.9058486223220825} 02/26/2022 06:07:59 - INFO - codeparrot_training - Step 40594: {'lr': 4.5896288046706396e-05, 'samples': 20784640, 'steps': 40594, 'loss/train': 1.572718620300293} 02/26/2022 06:08:03 - INFO - codeparrot_training - Step 40595: {'lr': 4.5886839721160533e-05, 'samples': 20785152, 'steps': 40595, 'loss/train': 1.552043080329895} 02/26/2022 06:08:08 - INFO - codeparrot_training - Step 40596: {'lr': 4.5877392269970705e-05, 'samples': 20785664, 'steps': 40596, 'loss/train': 0.4863733649253845} 02/26/2022 06:08:12 - INFO - codeparrot_training - Step 40597: {'lr': 4.5867945693177285e-05, 'samples': 20786176, 'steps': 40597, 'loss/train': 2.2209222316741943} 02/26/2022 06:08:18 - INFO - codeparrot_training - Step 40598: {'lr': 4.585849999082079e-05, 'samples': 20786688, 'steps': 40598, 'loss/train': 2.3790736198425293} 02/26/2022 06:08:21 - INFO - codeparrot_training - Step 40599: {'lr': 4.5849055162941605e-05, 'samples': 20787200, 'steps': 40599, 'loss/train': 0.9918031692504883} 02/26/2022 06:08:27 - INFO - codeparrot_training - Step 40600: {'lr': 4.5839611209580275e-05, 'samples': 20787712, 'steps': 40600, 'loss/train': 2.91489315032959} 02/26/2022 06:08:30 - INFO - codeparrot_training - Step 40601: {'lr': 4.5830168130777225e-05, 'samples': 20788224, 'steps': 40601, 'loss/train': 0.4606277048587799} 02/26/2022 06:08:36 - INFO - codeparrot_training - Step 40602: {'lr': 4.5820725926572934e-05, 'samples': 20788736, 'steps': 40602, 'loss/train': 1.4059597253799438} 02/26/2022 06:08:40 - INFO - codeparrot_training - Step 40603: {'lr': 4.581128459700773e-05, 'samples': 20789248, 'steps': 40603, 'loss/train': 2.222914695739746} 02/26/2022 06:08:43 - INFO - codeparrot_training - Step 40604: {'lr': 4.5801844142122214e-05, 'samples': 20789760, 'steps': 40604, 'loss/train': 3.2750346660614014} 02/26/2022 06:08:49 - INFO - codeparrot_training - Step 40605: {'lr': 4.579240456195677e-05, 'samples': 20790272, 'steps': 40605, 'loss/train': 0.3980376124382019} 02/26/2022 06:08:54 - INFO - codeparrot_training - Step 40606: {'lr': 4.578296585655181e-05, 'samples': 20790784, 'steps': 40606, 'loss/train': 1.935340166091919} 02/26/2022 06:08:58 - INFO - codeparrot_training - Step 40607: {'lr': 4.57735280259477e-05, 'samples': 20791296, 'steps': 40607, 'loss/train': 2.185854911804199} 02/26/2022 06:09:03 - INFO - codeparrot_training - Step 40608: {'lr': 4.5764091070185035e-05, 'samples': 20791808, 'steps': 40608, 'loss/train': 1.781665325164795} 02/26/2022 06:09:07 - INFO - codeparrot_training - Step 40609: {'lr': 4.5754654989304143e-05, 'samples': 20792320, 'steps': 40609, 'loss/train': 1.6466495990753174} 02/26/2022 06:09:12 - INFO - codeparrot_training - Step 40610: {'lr': 4.574521978334545e-05, 'samples': 20792832, 'steps': 40610, 'loss/train': 2.1167585849761963} 02/26/2022 06:09:15 - INFO - codeparrot_training - Step 40611: {'lr': 4.573578545234933e-05, 'samples': 20793344, 'steps': 40611, 'loss/train': 0.7844367027282715} 02/26/2022 06:09:21 - INFO - codeparrot_training - Step 40612: {'lr': 4.572635199635627e-05, 'samples': 20793856, 'steps': 40612, 'loss/train': 1.0828701257705688} 02/26/2022 06:09:24 - INFO - codeparrot_training - Step 40613: {'lr': 4.5716919415406677e-05, 'samples': 20794368, 'steps': 40613, 'loss/train': 1.0152088403701782} 02/26/2022 06:09:31 - INFO - codeparrot_training - Step 40614: {'lr': 4.570748770954089e-05, 'samples': 20794880, 'steps': 40614, 'loss/train': 2.0802741050720215} 02/26/2022 06:09:35 - INFO - codeparrot_training - Step 40615: {'lr': 4.569805687879946e-05, 'samples': 20795392, 'steps': 40615, 'loss/train': 1.5632034540176392} 02/26/2022 06:09:40 - INFO - codeparrot_training - Step 40616: {'lr': 4.5688626923222564e-05, 'samples': 20795904, 'steps': 40616, 'loss/train': 1.374606728553772} 02/26/2022 06:09:44 - INFO - codeparrot_training - Step 40617: {'lr': 4.5679197842850766e-05, 'samples': 20796416, 'steps': 40617, 'loss/train': 1.8106578588485718} 02/26/2022 06:09:49 - INFO - codeparrot_training - Step 40618: {'lr': 4.5669769637724346e-05, 'samples': 20796928, 'steps': 40618, 'loss/train': 1.1967626810073853} 02/26/2022 06:09:53 - INFO - codeparrot_training - Step 40619: {'lr': 4.566034230788388e-05, 'samples': 20797440, 'steps': 40619, 'loss/train': 2.0095179080963135} 02/26/2022 06:09:58 - INFO - codeparrot_training - Step 40620: {'lr': 4.56509158533695e-05, 'samples': 20797952, 'steps': 40620, 'loss/train': 0.8096858859062195} 02/26/2022 06:10:02 - INFO - codeparrot_training - Step 40621: {'lr': 4.5641490274221766e-05, 'samples': 20798464, 'steps': 40621, 'loss/train': 2.2266359329223633} 02/26/2022 06:10:07 - INFO - codeparrot_training - Step 40622: {'lr': 4.5632065570480916e-05, 'samples': 20798976, 'steps': 40622, 'loss/train': 1.4248467683792114} 02/26/2022 06:10:11 - INFO - codeparrot_training - Step 40623: {'lr': 4.562264174218755e-05, 'samples': 20799488, 'steps': 40623, 'loss/train': 2.23813796043396} 02/26/2022 06:10:17 - INFO - codeparrot_training - Step 40624: {'lr': 4.5613218789381765e-05, 'samples': 20800000, 'steps': 40624, 'loss/train': 2.587691307067871} 02/26/2022 06:10:21 - INFO - codeparrot_training - Step 40625: {'lr': 4.56037967121041e-05, 'samples': 20800512, 'steps': 40625, 'loss/train': 1.147760272026062} 02/26/2022 06:10:26 - INFO - codeparrot_training - Step 40626: {'lr': 4.559437551039483e-05, 'samples': 20801024, 'steps': 40626, 'loss/train': 1.448170781135559} 02/26/2022 06:10:30 - INFO - codeparrot_training - Step 40627: {'lr': 4.5584955184294395e-05, 'samples': 20801536, 'steps': 40627, 'loss/train': 2.316469669342041} 02/26/2022 06:10:36 - INFO - codeparrot_training - Step 40628: {'lr': 4.55755357338431e-05, 'samples': 20802048, 'steps': 40628, 'loss/train': 0.6476761698722839} 02/26/2022 06:10:39 - INFO - codeparrot_training - Step 40629: {'lr': 4.5566117159081296e-05, 'samples': 20802560, 'steps': 40629, 'loss/train': 2.0663208961486816} 02/26/2022 06:10:44 - INFO - codeparrot_training - Step 40630: {'lr': 4.5556699460049283e-05, 'samples': 20803072, 'steps': 40630, 'loss/train': 1.1277186870574951} 02/26/2022 06:10:48 - INFO - codeparrot_training - Step 40631: {'lr': 4.5547282636787505e-05, 'samples': 20803584, 'steps': 40631, 'loss/train': 2.1589252948760986} 02/26/2022 06:10:53 - INFO - codeparrot_training - Step 40632: {'lr': 4.553786668933624e-05, 'samples': 20804096, 'steps': 40632, 'loss/train': 2.241276264190674} 02/26/2022 06:10:57 - INFO - codeparrot_training - Step 40633: {'lr': 4.552845161773583e-05, 'samples': 20804608, 'steps': 40633, 'loss/train': 1.279417872428894} 02/26/2022 06:11:03 - INFO - codeparrot_training - Step 40634: {'lr': 4.551903742202657e-05, 'samples': 20805120, 'steps': 40634, 'loss/train': 2.058643341064453} 02/26/2022 06:11:06 - INFO - codeparrot_training - Step 40635: {'lr': 4.5509624102248845e-05, 'samples': 20805632, 'steps': 40635, 'loss/train': 0.7822037935256958} 02/26/2022 06:11:12 - INFO - codeparrot_training - Step 40636: {'lr': 4.550021165844298e-05, 'samples': 20806144, 'steps': 40636, 'loss/train': 1.352022409439087} 02/26/2022 06:11:15 - INFO - codeparrot_training - Step 40637: {'lr': 4.5490800090649264e-05, 'samples': 20806656, 'steps': 40637, 'loss/train': 1.8850128650665283} 02/26/2022 06:11:21 - INFO - codeparrot_training - Step 40638: {'lr': 4.5481389398907976e-05, 'samples': 20807168, 'steps': 40638, 'loss/train': 2.1383614540100098} 02/26/2022 06:11:24 - INFO - codeparrot_training - Step 40639: {'lr': 4.547197958325952e-05, 'samples': 20807680, 'steps': 40639, 'loss/train': 1.354177713394165} 02/26/2022 06:11:30 - INFO - codeparrot_training - Step 40640: {'lr': 4.54625706437441e-05, 'samples': 20808192, 'steps': 40640, 'loss/train': 2.9157562255859375} 02/26/2022 06:11:33 - INFO - codeparrot_training - Step 40641: {'lr': 4.545316258040222e-05, 'samples': 20808704, 'steps': 40641, 'loss/train': 1.508209228515625} 02/26/2022 06:11:39 - INFO - codeparrot_training - Step 40642: {'lr': 4.544375539327389e-05, 'samples': 20809216, 'steps': 40642, 'loss/train': 1.7163926362991333} 02/26/2022 06:11:42 - INFO - codeparrot_training - Step 40643: {'lr': 4.543434908239963e-05, 'samples': 20809728, 'steps': 40643, 'loss/train': 1.3316614627838135} 02/26/2022 06:11:49 - INFO - codeparrot_training - Step 40644: {'lr': 4.54249436478196e-05, 'samples': 20810240, 'steps': 40644, 'loss/train': 1.2661103010177612} 02/26/2022 06:11:52 - INFO - codeparrot_training - Step 40645: {'lr': 4.5415539089574296e-05, 'samples': 20810752, 'steps': 40645, 'loss/train': 2.5089473724365234} 02/26/2022 06:11:58 - INFO - codeparrot_training - Step 40646: {'lr': 4.540613540770372e-05, 'samples': 20811264, 'steps': 40646, 'loss/train': 1.5070173740386963} 02/26/2022 06:12:01 - INFO - codeparrot_training - Step 40647: {'lr': 4.539673260224833e-05, 'samples': 20811776, 'steps': 40647, 'loss/train': 1.2447699308395386} 02/26/2022 06:12:07 - INFO - codeparrot_training - Step 40648: {'lr': 4.5387330673248404e-05, 'samples': 20812288, 'steps': 40648, 'loss/train': 1.0220937728881836} 02/26/2022 06:12:10 - INFO - codeparrot_training - Step 40649: {'lr': 4.537792962074416e-05, 'samples': 20812800, 'steps': 40649, 'loss/train': 0.5093560218811035} 02/26/2022 06:12:16 - INFO - codeparrot_training - Step 40650: {'lr': 4.5368529444775846e-05, 'samples': 20813312, 'steps': 40650, 'loss/train': 1.0306366682052612} 02/26/2022 06:12:19 - INFO - codeparrot_training - Step 40651: {'lr': 4.535913014538384e-05, 'samples': 20813824, 'steps': 40651, 'loss/train': 1.6842639446258545} 02/26/2022 06:12:25 - INFO - codeparrot_training - Step 40652: {'lr': 4.534973172260831e-05, 'samples': 20814336, 'steps': 40652, 'loss/train': 1.7435857057571411} 02/26/2022 06:12:29 - INFO - codeparrot_training - Step 40653: {'lr': 4.53403341764895e-05, 'samples': 20814848, 'steps': 40653, 'loss/train': 2.4779903888702393} 02/26/2022 06:12:34 - INFO - codeparrot_training - Step 40654: {'lr': 4.5330937507067795e-05, 'samples': 20815360, 'steps': 40654, 'loss/train': 1.433372974395752} 02/26/2022 06:12:38 - INFO - codeparrot_training - Step 40655: {'lr': 4.532154171438333e-05, 'samples': 20815872, 'steps': 40655, 'loss/train': 1.8924967050552368} 02/26/2022 06:12:43 - INFO - codeparrot_training - Step 40656: {'lr': 4.531214679847639e-05, 'samples': 20816384, 'steps': 40656, 'loss/train': 0.5280120968818665} 02/26/2022 06:12:47 - INFO - codeparrot_training - Step 40657: {'lr': 4.5302752759387166e-05, 'samples': 20816896, 'steps': 40657, 'loss/train': 2.0612668991088867} 02/26/2022 06:12:52 - INFO - codeparrot_training - Step 40658: {'lr': 4.529335959715602e-05, 'samples': 20817408, 'steps': 40658, 'loss/train': 1.9786375761032104} 02/26/2022 06:12:56 - INFO - codeparrot_training - Step 40659: {'lr': 4.528396731182313e-05, 'samples': 20817920, 'steps': 40659, 'loss/train': 3.2058160305023193} 02/26/2022 06:13:03 - INFO - codeparrot_training - Step 40660: {'lr': 4.5274575903428706e-05, 'samples': 20818432, 'steps': 40660, 'loss/train': 2.008648157119751} 02/26/2022 06:13:06 - INFO - codeparrot_training - Step 40661: {'lr': 4.526518537201291e-05, 'samples': 20818944, 'steps': 40661, 'loss/train': 0.45992931723594666} 02/26/2022 06:13:12 - INFO - codeparrot_training - Step 40662: {'lr': 4.525579571761612e-05, 'samples': 20819456, 'steps': 40662, 'loss/train': 1.9788191318511963} 02/26/2022 06:13:15 - INFO - codeparrot_training - Step 40663: {'lr': 4.524640694027851e-05, 'samples': 20819968, 'steps': 40663, 'loss/train': 1.5528110265731812} 02/26/2022 06:13:21 - INFO - codeparrot_training - Step 40664: {'lr': 4.523701904004027e-05, 'samples': 20820480, 'steps': 40664, 'loss/train': 1.4869941473007202} 02/26/2022 06:13:24 - INFO - codeparrot_training - Step 40665: {'lr': 4.522763201694152e-05, 'samples': 20820992, 'steps': 40665, 'loss/train': 1.5976141691207886} 02/26/2022 06:13:29 - INFO - codeparrot_training - Step 40666: {'lr': 4.521824587102269e-05, 'samples': 20821504, 'steps': 40666, 'loss/train': 1.2229297161102295} 02/26/2022 06:13:33 - INFO - codeparrot_training - Step 40667: {'lr': 4.520886060232382e-05, 'samples': 20822016, 'steps': 40667, 'loss/train': 1.2224538326263428} 02/26/2022 06:13:38 - INFO - codeparrot_training - Step 40668: {'lr': 4.519947621088519e-05, 'samples': 20822528, 'steps': 40668, 'loss/train': 1.6681156158447266} 02/26/2022 06:13:42 - INFO - codeparrot_training - Step 40669: {'lr': 4.51900926967469e-05, 'samples': 20823040, 'steps': 40669, 'loss/train': 1.7371987104415894} 02/26/2022 06:13:48 - INFO - codeparrot_training - Step 40670: {'lr': 4.518071005994931e-05, 'samples': 20823552, 'steps': 40670, 'loss/train': 2.11722469329834} 02/26/2022 06:13:52 - INFO - codeparrot_training - Step 40671: {'lr': 4.517132830053247e-05, 'samples': 20824064, 'steps': 40671, 'loss/train': 2.6994738578796387} 02/26/2022 06:13:57 - INFO - codeparrot_training - Step 40672: {'lr': 4.516194741853666e-05, 'samples': 20824576, 'steps': 40672, 'loss/train': 2.239804267883301} 02/26/2022 06:14:01 - INFO - codeparrot_training - Step 40673: {'lr': 4.515256741400195e-05, 'samples': 20825088, 'steps': 40673, 'loss/train': 1.952246904373169} 02/26/2022 06:14:06 - INFO - codeparrot_training - Step 40674: {'lr': 4.514318828696864e-05, 'samples': 20825600, 'steps': 40674, 'loss/train': 1.093126893043518} 02/26/2022 06:14:10 - INFO - codeparrot_training - Step 40675: {'lr': 4.513381003747682e-05, 'samples': 20826112, 'steps': 40675, 'loss/train': 2.2619130611419678} 02/26/2022 06:14:15 - INFO - codeparrot_training - Step 40676: {'lr': 4.5124432665566816e-05, 'samples': 20826624, 'steps': 40676, 'loss/train': 1.5883575677871704} 02/26/2022 06:14:19 - INFO - codeparrot_training - Step 40677: {'lr': 4.511505617127859e-05, 'samples': 20827136, 'steps': 40677, 'loss/train': 2.1372218132019043} 02/26/2022 06:14:24 - INFO - codeparrot_training - Step 40678: {'lr': 4.5105680554652444e-05, 'samples': 20827648, 'steps': 40678, 'loss/train': 0.2200467884540558} 02/26/2022 06:14:28 - INFO - codeparrot_training - Step 40679: {'lr': 4.509630581572844e-05, 'samples': 20828160, 'steps': 40679, 'loss/train': 1.5311400890350342} 02/26/2022 06:14:34 - INFO - codeparrot_training - Step 40680: {'lr': 4.5086931954546936e-05, 'samples': 20828672, 'steps': 40680, 'loss/train': 1.8666329383850098} 02/26/2022 06:14:38 - INFO - codeparrot_training - Step 40681: {'lr': 4.507755897114779e-05, 'samples': 20829184, 'steps': 40681, 'loss/train': 0.5599453449249268} 02/26/2022 06:14:43 - INFO - codeparrot_training - Step 40682: {'lr': 4.50681868655714e-05, 'samples': 20829696, 'steps': 40682, 'loss/train': 1.1868550777435303} 02/26/2022 06:14:47 - INFO - codeparrot_training - Step 40683: {'lr': 4.505881563785777e-05, 'samples': 20830208, 'steps': 40683, 'loss/train': 2.5260062217712402} 02/26/2022 06:14:52 - INFO - codeparrot_training - Step 40684: {'lr': 4.504944528804722e-05, 'samples': 20830720, 'steps': 40684, 'loss/train': 1.2936135530471802} 02/26/2022 06:14:56 - INFO - codeparrot_training - Step 40685: {'lr': 4.504007581617961e-05, 'samples': 20831232, 'steps': 40685, 'loss/train': 1.6963977813720703} 02/26/2022 06:15:01 - INFO - codeparrot_training - Step 40686: {'lr': 4.503070722229533e-05, 'samples': 20831744, 'steps': 40686, 'loss/train': 0.7575234770774841} 02/26/2022 06:15:05 - INFO - codeparrot_training - Step 40687: {'lr': 4.5021339506434333e-05, 'samples': 20832256, 'steps': 40687, 'loss/train': 2.60663104057312} 02/26/2022 06:15:10 - INFO - codeparrot_training - Step 40688: {'lr': 4.501197266863691e-05, 'samples': 20832768, 'steps': 40688, 'loss/train': 0.827700674533844} 02/26/2022 06:15:14 - INFO - codeparrot_training - Step 40689: {'lr': 4.50026067089431e-05, 'samples': 20833280, 'steps': 40689, 'loss/train': 1.858444333076477} 02/26/2022 06:15:20 - INFO - codeparrot_training - Step 40690: {'lr': 4.4993241627393015e-05, 'samples': 20833792, 'steps': 40690, 'loss/train': 0.8139358758926392} 02/26/2022 06:15:24 - INFO - codeparrot_training - Step 40691: {'lr': 4.498387742402674e-05, 'samples': 20834304, 'steps': 40691, 'loss/train': 1.3843085765838623} 02/26/2022 06:15:29 - INFO - codeparrot_training - Step 40692: {'lr': 4.4974514098884476e-05, 'samples': 20834816, 'steps': 40692, 'loss/train': 1.636030912399292} 02/26/2022 06:15:33 - INFO - codeparrot_training - Step 40693: {'lr': 4.496515165200632e-05, 'samples': 20835328, 'steps': 40693, 'loss/train': 0.40758177638053894} 02/26/2022 06:15:38 - INFO - codeparrot_training - Step 40694: {'lr': 4.495579008343234e-05, 'samples': 20835840, 'steps': 40694, 'loss/train': 1.7806816101074219} 02/26/2022 06:15:42 - INFO - codeparrot_training - Step 40695: {'lr': 4.494642939320262e-05, 'samples': 20836352, 'steps': 40695, 'loss/train': 2.207427740097046} 02/26/2022 06:15:47 - INFO - codeparrot_training - Step 40696: {'lr': 4.493706958135724e-05, 'samples': 20836864, 'steps': 40696, 'loss/train': 0.799826443195343} 02/26/2022 06:15:51 - INFO - codeparrot_training - Step 40697: {'lr': 4.492771064793641e-05, 'samples': 20837376, 'steps': 40697, 'loss/train': 2.2678768634796143} 02/26/2022 06:15:56 - INFO - codeparrot_training - Step 40698: {'lr': 4.4918352592980135e-05, 'samples': 20837888, 'steps': 40698, 'loss/train': 1.369477391242981} 02/26/2022 06:16:00 - INFO - codeparrot_training - Step 40699: {'lr': 4.490899541652854e-05, 'samples': 20838400, 'steps': 40699, 'loss/train': 0.9359695315361023} 02/26/2022 06:16:05 - INFO - codeparrot_training - Step 40700: {'lr': 4.4899639118621604e-05, 'samples': 20838912, 'steps': 40700, 'loss/train': 2.0658695697784424} 02/26/2022 06:16:09 - INFO - codeparrot_training - Step 40701: {'lr': 4.489028369929954e-05, 'samples': 20839424, 'steps': 40701, 'loss/train': 1.4029302597045898} 02/26/2022 06:16:14 - INFO - codeparrot_training - Step 40702: {'lr': 4.48809291586024e-05, 'samples': 20839936, 'steps': 40702, 'loss/train': 0.7578195929527283} 02/26/2022 06:16:18 - INFO - codeparrot_training - Step 40703: {'lr': 4.4871575496570196e-05, 'samples': 20840448, 'steps': 40703, 'loss/train': 0.972649097442627} 02/26/2022 06:16:23 - INFO - codeparrot_training - Step 40704: {'lr': 4.4862222713242974e-05, 'samples': 20840960, 'steps': 40704, 'loss/train': 1.363482117652893} 02/26/2022 06:16:27 - INFO - codeparrot_training - Step 40705: {'lr': 4.4852870808660935e-05, 'samples': 20841472, 'steps': 40705, 'loss/train': 1.8993037939071655} 02/26/2022 06:16:33 - INFO - codeparrot_training - Step 40706: {'lr': 4.4843519782864025e-05, 'samples': 20841984, 'steps': 40706, 'loss/train': 1.552755355834961} 02/26/2022 06:16:36 - INFO - codeparrot_training - Step 40707: {'lr': 4.483416963589235e-05, 'samples': 20842496, 'steps': 40707, 'loss/train': 2.6266660690307617} 02/26/2022 06:16:42 - INFO - codeparrot_training - Step 40708: {'lr': 4.482482036778587e-05, 'samples': 20843008, 'steps': 40708, 'loss/train': 1.4920932054519653} 02/26/2022 06:16:45 - INFO - codeparrot_training - Step 40709: {'lr': 4.481547197858479e-05, 'samples': 20843520, 'steps': 40709, 'loss/train': 3.0304181575775146} 02/26/2022 06:16:51 - INFO - codeparrot_training - Step 40710: {'lr': 4.480612446832899e-05, 'samples': 20844032, 'steps': 40710, 'loss/train': 1.311439037322998} 02/26/2022 06:16:54 - INFO - codeparrot_training - Step 40711: {'lr': 4.4796777837058736e-05, 'samples': 20844544, 'steps': 40711, 'loss/train': 1.6831220388412476} 02/26/2022 06:17:00 - INFO - codeparrot_training - Step 40712: {'lr': 4.4787432084813814e-05, 'samples': 20845056, 'steps': 40712, 'loss/train': 1.8098589181900024} 02/26/2022 06:17:03 - INFO - codeparrot_training - Step 40713: {'lr': 4.47780872116344e-05, 'samples': 20845568, 'steps': 40713, 'loss/train': 0.9848004579544067} 02/26/2022 06:17:09 - INFO - codeparrot_training - Step 40714: {'lr': 4.476874321756047e-05, 'samples': 20846080, 'steps': 40714, 'loss/train': 1.3946572542190552} 02/26/2022 06:17:12 - INFO - codeparrot_training - Step 40715: {'lr': 4.475940010263219e-05, 'samples': 20846592, 'steps': 40715, 'loss/train': 2.43644380569458} 02/26/2022 06:17:19 - INFO - codeparrot_training - Step 40716: {'lr': 4.475005786688932e-05, 'samples': 20847104, 'steps': 40716, 'loss/train': 1.3501920700073242} 02/26/2022 06:17:22 - INFO - codeparrot_training - Step 40717: {'lr': 4.474071651037209e-05, 'samples': 20847616, 'steps': 40717, 'loss/train': 1.3525301218032837} 02/26/2022 06:17:28 - INFO - codeparrot_training - Step 40718: {'lr': 4.4731376033120416e-05, 'samples': 20848128, 'steps': 40718, 'loss/train': 1.175323724746704} 02/26/2022 06:17:31 - INFO - codeparrot_training - Step 40719: {'lr': 4.4722036435174474e-05, 'samples': 20848640, 'steps': 40719, 'loss/train': 1.742234706878662} 02/26/2022 06:17:37 - INFO - codeparrot_training - Step 40720: {'lr': 4.471269771657399e-05, 'samples': 20849152, 'steps': 40720, 'loss/train': 2.0646002292633057} 02/26/2022 06:17:40 - INFO - codeparrot_training - Step 40721: {'lr': 4.470335987735921e-05, 'samples': 20849664, 'steps': 40721, 'loss/train': 2.9056432247161865} 02/26/2022 06:17:46 - INFO - codeparrot_training - Step 40722: {'lr': 4.469402291756999e-05, 'samples': 20850176, 'steps': 40722, 'loss/train': 1.9018305540084839} 02/26/2022 06:17:49 - INFO - codeparrot_training - Step 40723: {'lr': 4.468468683724644e-05, 'samples': 20850688, 'steps': 40723, 'loss/train': 1.7955676317214966} 02/26/2022 06:17:55 - INFO - codeparrot_training - Step 40724: {'lr': 4.4675351636428466e-05, 'samples': 20851200, 'steps': 40724, 'loss/train': 2.1126933097839355} 02/26/2022 06:17:58 - INFO - codeparrot_training - Step 40725: {'lr': 4.466601731515613e-05, 'samples': 20851712, 'steps': 40725, 'loss/train': 0.5542493462562561} 02/26/2022 06:18:04 - INFO - codeparrot_training - Step 40726: {'lr': 4.46566838734693e-05, 'samples': 20852224, 'steps': 40726, 'loss/train': 2.639909267425537} 02/26/2022 06:18:08 - INFO - codeparrot_training - Step 40727: {'lr': 4.4647351311408066e-05, 'samples': 20852736, 'steps': 40727, 'loss/train': 0.20054863393306732} 02/26/2022 06:18:13 - INFO - codeparrot_training - Step 40728: {'lr': 4.46380196290124e-05, 'samples': 20853248, 'steps': 40728, 'loss/train': 1.1855874061584473} 02/26/2022 06:18:17 - INFO - codeparrot_training - Step 40729: {'lr': 4.462868882632226e-05, 'samples': 20853760, 'steps': 40729, 'loss/train': 2.0065133571624756} 02/26/2022 06:18:22 - INFO - codeparrot_training - Step 40730: {'lr': 4.461935890337751e-05, 'samples': 20854272, 'steps': 40730, 'loss/train': 1.4884397983551025} 02/26/2022 06:18:26 - INFO - codeparrot_training - Step 40731: {'lr': 4.461002986021831e-05, 'samples': 20854784, 'steps': 40731, 'loss/train': 2.6291041374206543} 02/26/2022 06:18:31 - INFO - codeparrot_training - Step 40732: {'lr': 4.460070169688449e-05, 'samples': 20855296, 'steps': 40732, 'loss/train': 2.4353983402252197} 02/26/2022 06:18:35 - INFO - codeparrot_training - Step 40733: {'lr': 4.459137441341607e-05, 'samples': 20855808, 'steps': 40733, 'loss/train': 1.6389790773391724} 02/26/2022 06:18:41 - INFO - codeparrot_training - Step 40734: {'lr': 4.458204800985288e-05, 'samples': 20856320, 'steps': 40734, 'loss/train': 1.2082011699676514} 02/26/2022 06:18:44 - INFO - codeparrot_training - Step 40735: {'lr': 4.457272248623506e-05, 'samples': 20856832, 'steps': 40735, 'loss/train': 1.411363124847412} 02/26/2022 06:18:50 - INFO - codeparrot_training - Step 40736: {'lr': 4.456339784260246e-05, 'samples': 20857344, 'steps': 40736, 'loss/train': 1.9699811935424805} 02/26/2022 06:18:54 - INFO - codeparrot_training - Step 40737: {'lr': 4.4554074078995034e-05, 'samples': 20857856, 'steps': 40737, 'loss/train': 0.40651679039001465} 02/26/2022 06:18:59 - INFO - codeparrot_training - Step 40738: {'lr': 4.454475119545265e-05, 'samples': 20858368, 'steps': 40738, 'loss/train': 1.990043044090271} 02/26/2022 06:19:03 - INFO - codeparrot_training - Step 40739: {'lr': 4.453542919201536e-05, 'samples': 20858880, 'steps': 40739, 'loss/train': 1.5787826776504517} 02/26/2022 06:19:08 - INFO - codeparrot_training - Step 40740: {'lr': 4.4526108068723065e-05, 'samples': 20859392, 'steps': 40740, 'loss/train': 0.30250421166419983} 02/26/2022 06:19:12 - INFO - codeparrot_training - Step 40741: {'lr': 4.451678782561569e-05, 'samples': 20859904, 'steps': 40741, 'loss/train': 1.5993998050689697} 02/26/2022 06:19:17 - INFO - codeparrot_training - Step 40742: {'lr': 4.450746846273307e-05, 'samples': 20860416, 'steps': 40742, 'loss/train': 2.2453672885894775} 02/26/2022 06:19:21 - INFO - codeparrot_training - Step 40743: {'lr': 4.4498149980115255e-05, 'samples': 20860928, 'steps': 40743, 'loss/train': 1.0461004972457886} 02/26/2022 06:19:27 - INFO - codeparrot_training - Step 40744: {'lr': 4.448883237780213e-05, 'samples': 20861440, 'steps': 40744, 'loss/train': 0.5661978721618652} 02/26/2022 06:19:30 - INFO - codeparrot_training - Step 40745: {'lr': 4.4479515655833583e-05, 'samples': 20861952, 'steps': 40745, 'loss/train': 1.9539790153503418} 02/26/2022 06:19:35 - INFO - codeparrot_training - Step 40746: {'lr': 4.4470199814249544e-05, 'samples': 20862464, 'steps': 40746, 'loss/train': 2.5866692066192627} 02/26/2022 06:19:39 - INFO - codeparrot_training - Step 40747: {'lr': 4.4460884853089825e-05, 'samples': 20862976, 'steps': 40747, 'loss/train': 2.005302667617798} 02/26/2022 06:19:45 - INFO - codeparrot_training - Step 40748: {'lr': 4.4451570772394475e-05, 'samples': 20863488, 'steps': 40748, 'loss/train': 1.261811375617981} 02/26/2022 06:19:48 - INFO - codeparrot_training - Step 40749: {'lr': 4.444225757220327e-05, 'samples': 20864000, 'steps': 40749, 'loss/train': 1.837463617324829} 02/26/2022 06:19:54 - INFO - codeparrot_training - Step 40750: {'lr': 4.4432945252556286e-05, 'samples': 20864512, 'steps': 40750, 'loss/train': 2.495427131652832} 02/26/2022 06:19:57 - INFO - codeparrot_training - Step 40751: {'lr': 4.4423633813493156e-05, 'samples': 20865024, 'steps': 40751, 'loss/train': 2.406679630279541} 02/26/2022 06:20:03 - INFO - codeparrot_training - Step 40752: {'lr': 4.441432325505398e-05, 'samples': 20865536, 'steps': 40752, 'loss/train': 1.4487475156784058} 02/26/2022 06:20:07 - INFO - codeparrot_training - Step 40753: {'lr': 4.4405013577278495e-05, 'samples': 20866048, 'steps': 40753, 'loss/train': 1.676520824432373} 02/26/2022 06:20:13 - INFO - codeparrot_training - Step 40754: {'lr': 4.439570478020677e-05, 'samples': 20866560, 'steps': 40754, 'loss/train': 1.4793343544006348} 02/26/2022 06:20:16 - INFO - codeparrot_training - Step 40755: {'lr': 4.438639686387841e-05, 'samples': 20867072, 'steps': 40755, 'loss/train': 2.4943671226501465} 02/26/2022 06:20:24 - INFO - codeparrot_training - Step 40756: {'lr': 4.4377089828333525e-05, 'samples': 20867584, 'steps': 40756, 'loss/train': 2.4255130290985107} 02/26/2022 06:20:27 - INFO - codeparrot_training - Step 40757: {'lr': 4.436778367361183e-05, 'samples': 20868096, 'steps': 40757, 'loss/train': 1.2586151361465454} 02/26/2022 06:20:33 - INFO - codeparrot_training - Step 40758: {'lr': 4.4358478399753376e-05, 'samples': 20868608, 'steps': 40758, 'loss/train': 2.4424874782562256} 02/26/2022 06:20:36 - INFO - codeparrot_training - Step 40759: {'lr': 4.434917400679778e-05, 'samples': 20869120, 'steps': 40759, 'loss/train': 0.986875057220459} 02/26/2022 06:20:42 - INFO - codeparrot_training - Step 40760: {'lr': 4.433987049478508e-05, 'samples': 20869632, 'steps': 40760, 'loss/train': 0.9369402527809143} 02/26/2022 06:20:45 - INFO - codeparrot_training - Step 40761: {'lr': 4.4330567863755e-05, 'samples': 20870144, 'steps': 40761, 'loss/train': 2.6631221771240234} 02/26/2022 06:20:51 - INFO - codeparrot_training - Step 40762: {'lr': 4.432126611374754e-05, 'samples': 20870656, 'steps': 40762, 'loss/train': 1.162684679031372} 02/26/2022 06:20:54 - INFO - codeparrot_training - Step 40763: {'lr': 4.4311965244802465e-05, 'samples': 20871168, 'steps': 40763, 'loss/train': 1.3134104013442993} 02/26/2022 06:21:00 - INFO - codeparrot_training - Step 40764: {'lr': 4.430266525695961e-05, 'samples': 20871680, 'steps': 40764, 'loss/train': 2.3260109424591064} 02/26/2022 06:21:03 - INFO - codeparrot_training - Step 40765: {'lr': 4.429336615025875e-05, 'samples': 20872192, 'steps': 40765, 'loss/train': 1.045145034790039} 02/26/2022 06:21:11 - INFO - codeparrot_training - Step 40766: {'lr': 4.428406792473988e-05, 'samples': 20872704, 'steps': 40766, 'loss/train': 1.8641666173934937} 02/26/2022 06:21:14 - INFO - codeparrot_training - Step 40767: {'lr': 4.427477058044271e-05, 'samples': 20873216, 'steps': 40767, 'loss/train': 2.126070261001587} 02/26/2022 06:21:20 - INFO - codeparrot_training - Step 40768: {'lr': 4.426547411740711e-05, 'samples': 20873728, 'steps': 40768, 'loss/train': 0.6071975231170654} 02/26/2022 06:21:24 - INFO - codeparrot_training - Step 40769: {'lr': 4.425617853567282e-05, 'samples': 20874240, 'steps': 40769, 'loss/train': 1.2382758855819702} 02/26/2022 06:21:27 - INFO - codeparrot_training - Step 40770: {'lr': 4.424688383527981e-05, 'samples': 20874752, 'steps': 40770, 'loss/train': 2.252837657928467} 02/26/2022 06:21:33 - INFO - codeparrot_training - Step 40771: {'lr': 4.423759001626773e-05, 'samples': 20875264, 'steps': 40771, 'loss/train': 0.39704710245132446} 02/26/2022 06:21:36 - INFO - codeparrot_training - Step 40772: {'lr': 4.4228297078676625e-05, 'samples': 20875776, 'steps': 40772, 'loss/train': 1.6951396465301514} 02/26/2022 06:21:42 - INFO - codeparrot_training - Step 40773: {'lr': 4.421900502254603e-05, 'samples': 20876288, 'steps': 40773, 'loss/train': 1.009183406829834} 02/26/2022 06:21:45 - INFO - codeparrot_training - Step 40774: {'lr': 4.4209713847915915e-05, 'samples': 20876800, 'steps': 40774, 'loss/train': 2.698479175567627} 02/26/2022 06:21:51 - INFO - codeparrot_training - Step 40775: {'lr': 4.420042355482601e-05, 'samples': 20877312, 'steps': 40775, 'loss/train': 1.1100869178771973} 02/26/2022 06:21:54 - INFO - codeparrot_training - Step 40776: {'lr': 4.4191134143316244e-05, 'samples': 20877824, 'steps': 40776, 'loss/train': 2.363595724105835} 02/26/2022 06:22:02 - INFO - codeparrot_training - Step 40777: {'lr': 4.418184561342617e-05, 'samples': 20878336, 'steps': 40777, 'loss/train': 1.3484740257263184} 02/26/2022 06:22:05 - INFO - codeparrot_training - Step 40778: {'lr': 4.417255796519581e-05, 'samples': 20878848, 'steps': 40778, 'loss/train': 1.744852066040039} 02/26/2022 06:22:11 - INFO - codeparrot_training - Step 40779: {'lr': 4.416327119866484e-05, 'samples': 20879360, 'steps': 40779, 'loss/train': 1.7797449827194214} 02/26/2022 06:22:14 - INFO - codeparrot_training - Step 40780: {'lr': 4.415398531387305e-05, 'samples': 20879872, 'steps': 40780, 'loss/train': 1.4905694723129272} 02/26/2022 06:22:20 - INFO - codeparrot_training - Step 40781: {'lr': 4.4144700310860154e-05, 'samples': 20880384, 'steps': 40781, 'loss/train': 1.4019602537155151} 02/26/2022 06:22:25 - INFO - codeparrot_training - Step 40782: {'lr': 4.4135416189666074e-05, 'samples': 20880896, 'steps': 40782, 'loss/train': 1.5768240690231323} 02/26/2022 06:22:29 - INFO - codeparrot_training - Step 40783: {'lr': 4.41261329503305e-05, 'samples': 20881408, 'steps': 40783, 'loss/train': 1.9067399501800537} 02/26/2022 06:22:35 - INFO - codeparrot_training - Step 40784: {'lr': 4.411685059289314e-05, 'samples': 20881920, 'steps': 40784, 'loss/train': 1.9370423555374146} 02/26/2022 06:22:38 - INFO - codeparrot_training - Step 40785: {'lr': 4.410756911739386e-05, 'samples': 20882432, 'steps': 40785, 'loss/train': 2.0855767726898193} 02/26/2022 06:22:45 - INFO - codeparrot_training - Step 40786: {'lr': 4.409828852387238e-05, 'samples': 20882944, 'steps': 40786, 'loss/train': 1.1551172733306885} 02/26/2022 06:22:49 - INFO - codeparrot_training - Step 40787: {'lr': 4.408900881236846e-05, 'samples': 20883456, 'steps': 40787, 'loss/train': 1.6368690729141235} 02/26/2022 06:22:54 - INFO - codeparrot_training - Step 40788: {'lr': 4.407972998292178e-05, 'samples': 20883968, 'steps': 40788, 'loss/train': 1.2518243789672852} 02/26/2022 06:22:58 - INFO - codeparrot_training - Step 40789: {'lr': 4.4070452035572226e-05, 'samples': 20884480, 'steps': 40789, 'loss/train': 0.770409345626831} 02/26/2022 06:23:04 - INFO - codeparrot_training - Step 40790: {'lr': 4.406117497035944e-05, 'samples': 20884992, 'steps': 40790, 'loss/train': 1.7084684371948242} 02/26/2022 06:23:07 - INFO - codeparrot_training - Step 40791: {'lr': 4.4051898787323215e-05, 'samples': 20885504, 'steps': 40791, 'loss/train': 0.6682916283607483} 02/26/2022 06:23:13 - INFO - codeparrot_training - Step 40792: {'lr': 4.404262348650317e-05, 'samples': 20886016, 'steps': 40792, 'loss/train': 1.2871159315109253} 02/26/2022 06:23:16 - INFO - codeparrot_training - Step 40793: {'lr': 4.4033349067939264e-05, 'samples': 20886528, 'steps': 40793, 'loss/train': 1.8911652565002441} 02/26/2022 06:23:22 - INFO - codeparrot_training - Step 40794: {'lr': 4.4024075531670955e-05, 'samples': 20887040, 'steps': 40794, 'loss/train': 1.4692856073379517} 02/26/2022 06:23:25 - INFO - codeparrot_training - Step 40795: {'lr': 4.401480287773818e-05, 'samples': 20887552, 'steps': 40795, 'loss/train': 2.4328737258911133} 02/26/2022 06:23:32 - INFO - codeparrot_training - Step 40796: {'lr': 4.4005531106180495e-05, 'samples': 20888064, 'steps': 40796, 'loss/train': 0.5882765650749207} 02/26/2022 06:23:36 - INFO - codeparrot_training - Step 40797: {'lr': 4.399626021703779e-05, 'samples': 20888576, 'steps': 40797, 'loss/train': 0.7727590799331665} 02/26/2022 06:23:41 - INFO - codeparrot_training - Step 40798: {'lr': 4.3986990210349686e-05, 'samples': 20889088, 'steps': 40798, 'loss/train': 1.250331997871399} 02/26/2022 06:23:45 - INFO - codeparrot_training - Step 40799: {'lr': 4.397772108615592e-05, 'samples': 20889600, 'steps': 40799, 'loss/train': 0.5298184156417847} 02/26/2022 06:23:50 - INFO - codeparrot_training - Step 40800: {'lr': 4.3968452844496076e-05, 'samples': 20890112, 'steps': 40800, 'loss/train': 0.34215831756591797} 02/26/2022 06:23:54 - INFO - codeparrot_training - Step 40801: {'lr': 4.395918548541003e-05, 'samples': 20890624, 'steps': 40801, 'loss/train': 0.5815662741661072} 02/26/2022 06:23:59 - INFO - codeparrot_training - Step 40802: {'lr': 4.394991900893744e-05, 'samples': 20891136, 'steps': 40802, 'loss/train': 1.3196675777435303} 02/26/2022 06:24:03 - INFO - codeparrot_training - Step 40803: {'lr': 4.3940653415117934e-05, 'samples': 20891648, 'steps': 40803, 'loss/train': 1.2838267087936401} 02/26/2022 06:24:08 - INFO - codeparrot_training - Step 40804: {'lr': 4.3931388703991185e-05, 'samples': 20892160, 'steps': 40804, 'loss/train': 0.6708071827888489} 02/26/2022 06:24:12 - INFO - codeparrot_training - Step 40805: {'lr': 4.392212487559699e-05, 'samples': 20892672, 'steps': 40805, 'loss/train': 1.4132819175720215} 02/26/2022 06:24:18 - INFO - codeparrot_training - Step 40806: {'lr': 4.391286192997499e-05, 'samples': 20893184, 'steps': 40806, 'loss/train': 1.7606216669082642} 02/26/2022 06:24:21 - INFO - codeparrot_training - Step 40807: {'lr': 4.390359986716483e-05, 'samples': 20893696, 'steps': 40807, 'loss/train': 1.7435531616210938} 02/26/2022 06:24:27 - INFO - codeparrot_training - Step 40808: {'lr': 4.389433868720616e-05, 'samples': 20894208, 'steps': 40808, 'loss/train': 2.457872152328491} 02/26/2022 06:24:30 - INFO - codeparrot_training - Step 40809: {'lr': 4.388507839013875e-05, 'samples': 20894720, 'steps': 40809, 'loss/train': 2.2805721759796143} 02/26/2022 06:24:36 - INFO - codeparrot_training - Step 40810: {'lr': 4.387581897600215e-05, 'samples': 20895232, 'steps': 40810, 'loss/train': 1.54039466381073} 02/26/2022 06:24:39 - INFO - codeparrot_training - Step 40811: {'lr': 4.386656044483622e-05, 'samples': 20895744, 'steps': 40811, 'loss/train': 1.3677804470062256} 02/26/2022 06:24:47 - INFO - codeparrot_training - Step 40812: {'lr': 4.385730279668038e-05, 'samples': 20896256, 'steps': 40812, 'loss/train': 1.6924481391906738} 02/26/2022 06:24:50 - INFO - codeparrot_training - Step 40813: {'lr': 4.384804603157444e-05, 'samples': 20896768, 'steps': 40813, 'loss/train': 1.5936757326126099} 02/26/2022 06:24:56 - INFO - codeparrot_training - Step 40814: {'lr': 4.383879014955794e-05, 'samples': 20897280, 'steps': 40814, 'loss/train': 1.200735330581665} 02/26/2022 06:24:59 - INFO - codeparrot_training - Step 40815: {'lr': 4.3829535150670756e-05, 'samples': 20897792, 'steps': 40815, 'loss/train': 1.162596583366394} 02/26/2022 06:25:05 - INFO - codeparrot_training - Step 40816: {'lr': 4.3820281034952224e-05, 'samples': 20898304, 'steps': 40816, 'loss/train': 1.3583128452301025} 02/26/2022 06:25:08 - INFO - codeparrot_training - Step 40817: {'lr': 4.38110278024422e-05, 'samples': 20898816, 'steps': 40817, 'loss/train': 0.8648126721382141} 02/26/2022 06:25:14 - INFO - codeparrot_training - Step 40818: {'lr': 4.38017754531802e-05, 'samples': 20899328, 'steps': 40818, 'loss/train': 1.525160789489746} 02/26/2022 06:25:18 - INFO - codeparrot_training - Step 40819: {'lr': 4.379252398720601e-05, 'samples': 20899840, 'steps': 40819, 'loss/train': 1.1953896284103394} 02/26/2022 06:25:23 - INFO - codeparrot_training - Step 40820: {'lr': 4.378327340455915e-05, 'samples': 20900352, 'steps': 40820, 'loss/train': 1.8413482904434204} 02/26/2022 06:25:27 - INFO - codeparrot_training - Step 40821: {'lr': 4.377402370527925e-05, 'samples': 20900864, 'steps': 40821, 'loss/train': 0.7372773289680481} 02/26/2022 06:25:34 - INFO - codeparrot_training - Step 40822: {'lr': 4.376477488940592e-05, 'samples': 20901376, 'steps': 40822, 'loss/train': 1.7024935483932495} 02/26/2022 06:25:37 - INFO - codeparrot_training - Step 40823: {'lr': 4.375552695697885e-05, 'samples': 20901888, 'steps': 40823, 'loss/train': 1.5873041152954102} 02/26/2022 06:25:43 - INFO - codeparrot_training - Step 40824: {'lr': 4.374627990803759e-05, 'samples': 20902400, 'steps': 40824, 'loss/train': 1.8936855792999268} 02/26/2022 06:25:46 - INFO - codeparrot_training - Step 40825: {'lr': 4.3737033742621807e-05, 'samples': 20902912, 'steps': 40825, 'loss/train': 1.7708910703659058} 02/26/2022 06:25:52 - INFO - codeparrot_training - Step 40826: {'lr': 4.372778846077105e-05, 'samples': 20903424, 'steps': 40826, 'loss/train': 1.6903440952301025} 02/26/2022 06:25:55 - INFO - codeparrot_training - Step 40827: {'lr': 4.37185440625249e-05, 'samples': 20903936, 'steps': 40827, 'loss/train': 1.6371549367904663} 02/26/2022 06:26:01 - INFO - codeparrot_training - Step 40828: {'lr': 4.3709300547923074e-05, 'samples': 20904448, 'steps': 40828, 'loss/train': 0.6397308111190796} 02/26/2022 06:26:04 - INFO - codeparrot_training - Step 40829: {'lr': 4.37000579170051e-05, 'samples': 20904960, 'steps': 40829, 'loss/train': 0.825473427772522} 02/26/2022 06:26:10 - INFO - codeparrot_training - Step 40830: {'lr': 4.369081616981058e-05, 'samples': 20905472, 'steps': 40830, 'loss/train': 1.4700621366500854} 02/26/2022 06:26:13 - INFO - codeparrot_training - Step 40831: {'lr': 4.3681575306379004e-05, 'samples': 20905984, 'steps': 40831, 'loss/train': 2.26362943649292} 02/26/2022 06:26:19 - INFO - codeparrot_training - Step 40832: {'lr': 4.367233532675011e-05, 'samples': 20906496, 'steps': 40832, 'loss/train': 1.8314299583435059} 02/26/2022 06:26:22 - INFO - codeparrot_training - Step 40833: {'lr': 4.366309623096343e-05, 'samples': 20907008, 'steps': 40833, 'loss/train': 1.544396996498108} 02/26/2022 06:26:30 - INFO - codeparrot_training - Step 40834: {'lr': 4.365385801905852e-05, 'samples': 20907520, 'steps': 40834, 'loss/train': 1.4938440322875977} 02/26/2022 06:26:33 - INFO - codeparrot_training - Step 40835: {'lr': 4.364462069107489e-05, 'samples': 20908032, 'steps': 40835, 'loss/train': 2.0050759315490723} 02/26/2022 06:26:39 - INFO - codeparrot_training - Step 40836: {'lr': 4.3635384247052265e-05, 'samples': 20908544, 'steps': 40836, 'loss/train': 2.5611331462860107} 02/26/2022 06:26:42 - INFO - codeparrot_training - Step 40837: {'lr': 4.36261486870301e-05, 'samples': 20909056, 'steps': 40837, 'loss/train': 0.5659295916557312} 02/26/2022 06:26:48 - INFO - codeparrot_training - Step 40838: {'lr': 4.3616914011048004e-05, 'samples': 20909568, 'steps': 40838, 'loss/train': 1.9685627222061157} 02/26/2022 06:26:51 - INFO - codeparrot_training - Step 40839: {'lr': 4.360768021914543e-05, 'samples': 20910080, 'steps': 40839, 'loss/train': 1.1939127445220947} 02/26/2022 06:26:57 - INFO - codeparrot_training - Step 40840: {'lr': 4.359844731136209e-05, 'samples': 20910592, 'steps': 40840, 'loss/train': 1.7248051166534424} 02/26/2022 06:27:00 - INFO - codeparrot_training - Step 40841: {'lr': 4.3589215287737486e-05, 'samples': 20911104, 'steps': 40841, 'loss/train': 2.264495372772217} 02/26/2022 06:27:06 - INFO - codeparrot_training - Step 40842: {'lr': 4.35799841483111e-05, 'samples': 20911616, 'steps': 40842, 'loss/train': 1.7720999717712402} 02/26/2022 06:27:09 - INFO - codeparrot_training - Step 40843: {'lr': 4.357075389312248e-05, 'samples': 20912128, 'steps': 40843, 'loss/train': 1.3957901000976562} 02/26/2022 06:27:15 - INFO - codeparrot_training - Step 40844: {'lr': 4.356152452221127e-05, 'samples': 20912640, 'steps': 40844, 'loss/train': 1.2353039979934692} 02/26/2022 06:27:18 - INFO - codeparrot_training - Step 40845: {'lr': 4.3552296035616836e-05, 'samples': 20913152, 'steps': 40845, 'loss/train': 2.1513874530792236} 02/26/2022 06:27:24 - INFO - codeparrot_training - Step 40846: {'lr': 4.3543068433378965e-05, 'samples': 20913664, 'steps': 40846, 'loss/train': 0.04748530685901642} 02/26/2022 06:27:27 - INFO - codeparrot_training - Step 40847: {'lr': 4.3533841715536895e-05, 'samples': 20914176, 'steps': 40847, 'loss/train': 1.8627108335494995} 02/26/2022 06:27:35 - INFO - codeparrot_training - Step 40848: {'lr': 4.352461588213036e-05, 'samples': 20914688, 'steps': 40848, 'loss/train': 1.798604130744934} 02/26/2022 06:27:38 - INFO - codeparrot_training - Step 40849: {'lr': 4.3515390933198756e-05, 'samples': 20915200, 'steps': 40849, 'loss/train': 4.01839017868042} 02/26/2022 06:27:44 - INFO - codeparrot_training - Step 40850: {'lr': 4.350616686878175e-05, 'samples': 20915712, 'steps': 40850, 'loss/train': 0.030324609950184822} 02/26/2022 06:27:47 - INFO - codeparrot_training - Step 40851: {'lr': 4.349694368891865e-05, 'samples': 20916224, 'steps': 40851, 'loss/train': 1.670116901397705} 02/26/2022 06:27:53 - INFO - codeparrot_training - Step 40852: {'lr': 4.348772139364912e-05, 'samples': 20916736, 'steps': 40852, 'loss/train': 2.041996479034424} 02/26/2022 06:27:56 - INFO - codeparrot_training - Step 40853: {'lr': 4.3478499983012565e-05, 'samples': 20917248, 'steps': 40853, 'loss/train': 2.494936466217041} 02/26/2022 06:28:02 - INFO - codeparrot_training - Step 40854: {'lr': 4.346927945704865e-05, 'samples': 20917760, 'steps': 40854, 'loss/train': 2.2606112957000732} 02/26/2022 06:28:05 - INFO - codeparrot_training - Step 40855: {'lr': 4.346005981579665e-05, 'samples': 20918272, 'steps': 40855, 'loss/train': 1.6669801473617554} 02/26/2022 06:28:11 - INFO - codeparrot_training - Step 40856: {'lr': 4.345084105929622e-05, 'samples': 20918784, 'steps': 40856, 'loss/train': 2.2037057876586914} 02/26/2022 06:28:14 - INFO - codeparrot_training - Step 40857: {'lr': 4.3441623187586766e-05, 'samples': 20919296, 'steps': 40857, 'loss/train': 1.3628287315368652} 02/26/2022 06:28:22 - INFO - codeparrot_training - Step 40858: {'lr': 4.343240620070785e-05, 'samples': 20919808, 'steps': 40858, 'loss/train': 1.6588469743728638} 02/26/2022 06:28:25 - INFO - codeparrot_training - Step 40859: {'lr': 4.342319009869891e-05, 'samples': 20920320, 'steps': 40859, 'loss/train': 1.1480106115341187} 02/26/2022 06:28:30 - INFO - codeparrot_training - Step 40860: {'lr': 4.341397488159943e-05, 'samples': 20920832, 'steps': 40860, 'loss/train': 2.2994632720947266} 02/26/2022 06:28:34 - INFO - codeparrot_training - Step 40861: {'lr': 4.3404760549448804e-05, 'samples': 20921344, 'steps': 40861, 'loss/train': 1.8536527156829834} 02/26/2022 06:28:39 - INFO - codeparrot_training - Step 40862: {'lr': 4.339554710228666e-05, 'samples': 20921856, 'steps': 40862, 'loss/train': 2.179905652999878} 02/26/2022 06:28:43 - INFO - codeparrot_training - Step 40863: {'lr': 4.33863345401524e-05, 'samples': 20922368, 'steps': 40863, 'loss/train': 1.7461615800857544} 02/26/2022 06:28:48 - INFO - codeparrot_training - Step 40864: {'lr': 4.337712286308546e-05, 'samples': 20922880, 'steps': 40864, 'loss/train': 0.3026457726955414} 02/26/2022 06:28:52 - INFO - codeparrot_training - Step 40865: {'lr': 4.3367912071125256e-05, 'samples': 20923392, 'steps': 40865, 'loss/train': 1.9761264324188232} 02/26/2022 06:28:57 - INFO - codeparrot_training - Step 40866: {'lr': 4.335870216431137e-05, 'samples': 20923904, 'steps': 40866, 'loss/train': 2.746539831161499} 02/26/2022 06:29:01 - INFO - codeparrot_training - Step 40867: {'lr': 4.334949314268319e-05, 'samples': 20924416, 'steps': 40867, 'loss/train': 1.6884409189224243} 02/26/2022 06:29:08 - INFO - codeparrot_training - Step 40868: {'lr': 4.334028500628015e-05, 'samples': 20924928, 'steps': 40868, 'loss/train': 1.7547966241836548} 02/26/2022 06:29:11 - INFO - codeparrot_training - Step 40869: {'lr': 4.333107775514167e-05, 'samples': 20925440, 'steps': 40869, 'loss/train': 1.5080362558364868} 02/26/2022 06:29:17 - INFO - codeparrot_training - Step 40870: {'lr': 4.332187138930726e-05, 'samples': 20925952, 'steps': 40870, 'loss/train': 2.6367745399475098} 02/26/2022 06:29:20 - INFO - codeparrot_training - Step 40871: {'lr': 4.331266590881636e-05, 'samples': 20926464, 'steps': 40871, 'loss/train': 1.408315658569336} 02/26/2022 06:29:26 - INFO - codeparrot_training - Step 40872: {'lr': 4.330346131370835e-05, 'samples': 20926976, 'steps': 40872, 'loss/train': 1.6436982154846191} 02/26/2022 06:29:29 - INFO - codeparrot_training - Step 40873: {'lr': 4.329425760402267e-05, 'samples': 20927488, 'steps': 40873, 'loss/train': 1.768805980682373} 02/26/2022 06:29:35 - INFO - codeparrot_training - Step 40874: {'lr': 4.328505477979869e-05, 'samples': 20928000, 'steps': 40874, 'loss/train': 0.49441030621528625} 02/26/2022 06:29:38 - INFO - codeparrot_training - Step 40875: {'lr': 4.327585284107597e-05, 'samples': 20928512, 'steps': 40875, 'loss/train': 0.9649516344070435} 02/26/2022 06:29:44 - INFO - codeparrot_training - Step 40876: {'lr': 4.326665178789385e-05, 'samples': 20929024, 'steps': 40876, 'loss/train': 1.2714118957519531} 02/26/2022 06:29:47 - INFO - codeparrot_training - Step 40877: {'lr': 4.325745162029174e-05, 'samples': 20929536, 'steps': 40877, 'loss/train': 2.574646472930908} 02/26/2022 06:29:55 - INFO - codeparrot_training - Step 40878: {'lr': 4.324825233830901e-05, 'samples': 20930048, 'steps': 40878, 'loss/train': 1.4813597202301025} 02/26/2022 06:29:58 - INFO - codeparrot_training - Step 40879: {'lr': 4.323905394198519e-05, 'samples': 20930560, 'steps': 40879, 'loss/train': 1.8668663501739502} 02/26/2022 06:30:04 - INFO - codeparrot_training - Step 40880: {'lr': 4.3229856431359515e-05, 'samples': 20931072, 'steps': 40880, 'loss/train': 1.1853092908859253} 02/26/2022 06:30:07 - INFO - codeparrot_training - Step 40881: {'lr': 4.3220659806471635e-05, 'samples': 20931584, 'steps': 40881, 'loss/train': 2.443833112716675} 02/26/2022 06:30:13 - INFO - codeparrot_training - Step 40882: {'lr': 4.3211464067360654e-05, 'samples': 20932096, 'steps': 40882, 'loss/train': 1.5212701559066772} 02/26/2022 06:30:16 - INFO - codeparrot_training - Step 40883: {'lr': 4.3202269214066144e-05, 'samples': 20932608, 'steps': 40883, 'loss/train': 2.31856107711792} 02/26/2022 06:30:22 - INFO - codeparrot_training - Step 40884: {'lr': 4.319307524662738e-05, 'samples': 20933120, 'steps': 40884, 'loss/train': 2.0456039905548096} 02/26/2022 06:30:25 - INFO - codeparrot_training - Step 40885: {'lr': 4.318388216508395e-05, 'samples': 20933632, 'steps': 40885, 'loss/train': 1.1022002696990967} 02/26/2022 06:30:31 - INFO - codeparrot_training - Step 40886: {'lr': 4.317468996947499e-05, 'samples': 20934144, 'steps': 40886, 'loss/train': 0.9232212901115417} 02/26/2022 06:30:34 - INFO - codeparrot_training - Step 40887: {'lr': 4.3165498659840014e-05, 'samples': 20934656, 'steps': 40887, 'loss/train': 8.901434898376465} 02/26/2022 06:30:40 - INFO - codeparrot_training - Step 40888: {'lr': 4.3156308236218326e-05, 'samples': 20935168, 'steps': 40888, 'loss/train': 1.868941068649292} 02/26/2022 06:30:43 - INFO - codeparrot_training - Step 40889: {'lr': 4.314711869864946e-05, 'samples': 20935680, 'steps': 40889, 'loss/train': 1.9776207208633423} 02/26/2022 06:30:49 - INFO - codeparrot_training - Step 40890: {'lr': 4.3137930047172506e-05, 'samples': 20936192, 'steps': 40890, 'loss/train': 1.3485157489776611} 02/26/2022 06:30:52 - INFO - codeparrot_training - Step 40891: {'lr': 4.312874228182706e-05, 'samples': 20936704, 'steps': 40891, 'loss/train': 3.263789415359497} 02/26/2022 06:30:58 - INFO - codeparrot_training - Step 40892: {'lr': 4.3119555402652334e-05, 'samples': 20937216, 'steps': 40892, 'loss/train': 0.8976107835769653} 02/26/2022 06:31:01 - INFO - codeparrot_training - Step 40893: {'lr': 4.311036940968777e-05, 'samples': 20937728, 'steps': 40893, 'loss/train': 2.054450511932373} 02/26/2022 06:31:07 - INFO - codeparrot_training - Step 40894: {'lr': 4.310118430297272e-05, 'samples': 20938240, 'steps': 40894, 'loss/train': 2.3945393562316895} 02/26/2022 06:31:10 - INFO - codeparrot_training - Step 40895: {'lr': 4.309200008254649e-05, 'samples': 20938752, 'steps': 40895, 'loss/train': 1.1406890153884888} 02/26/2022 06:31:17 - INFO - codeparrot_training - Step 40896: {'lr': 4.308281674844836e-05, 'samples': 20939264, 'steps': 40896, 'loss/train': 1.7404903173446655} 02/26/2022 06:31:21 - INFO - codeparrot_training - Step 40897: {'lr': 4.307363430071781e-05, 'samples': 20939776, 'steps': 40897, 'loss/train': 2.1969399452209473} 02/26/2022 06:31:26 - INFO - codeparrot_training - Step 40898: {'lr': 4.3064452739394104e-05, 'samples': 20940288, 'steps': 40898, 'loss/train': 1.557886004447937} 02/26/2022 06:31:30 - INFO - codeparrot_training - Step 40899: {'lr': 4.3055272064516566e-05, 'samples': 20940800, 'steps': 40899, 'loss/train': 2.0844814777374268} 02/26/2022 06:31:35 - INFO - codeparrot_training - Step 40900: {'lr': 4.3046092276124466e-05, 'samples': 20941312, 'steps': 40900, 'loss/train': 1.954880714416504} 02/26/2022 06:31:41 - INFO - codeparrot_training - Step 40901: {'lr': 4.3036913374257276e-05, 'samples': 20941824, 'steps': 40901, 'loss/train': 1.1112169027328491} 02/26/2022 06:31:44 - INFO - codeparrot_training - Step 40902: {'lr': 4.302773535895421e-05, 'samples': 20942336, 'steps': 40902, 'loss/train': 1.0091986656188965} 02/26/2022 06:31:50 - INFO - codeparrot_training - Step 40903: {'lr': 4.3018558230254604e-05, 'samples': 20942848, 'steps': 40903, 'loss/train': 1.8580636978149414} 02/26/2022 06:31:53 - INFO - codeparrot_training - Step 40904: {'lr': 4.3009381988197707e-05, 'samples': 20943360, 'steps': 40904, 'loss/train': 1.3948968648910522} 02/26/2022 06:32:00 - INFO - codeparrot_training - Step 40905: {'lr': 4.3000206632822975e-05, 'samples': 20943872, 'steps': 40905, 'loss/train': 2.0072360038757324} 02/26/2022 06:32:04 - INFO - codeparrot_training - Step 40906: {'lr': 4.2991032164169605e-05, 'samples': 20944384, 'steps': 40906, 'loss/train': 2.0293049812316895} 02/26/2022 06:32:09 - INFO - codeparrot_training - Step 40907: {'lr': 4.298185858227693e-05, 'samples': 20944896, 'steps': 40907, 'loss/train': 1.7484631538391113} 02/26/2022 06:32:13 - INFO - codeparrot_training - Step 40908: {'lr': 4.2972685887184174e-05, 'samples': 20945408, 'steps': 40908, 'loss/train': 3.557563304901123} 02/26/2022 06:32:18 - INFO - codeparrot_training - Step 40909: {'lr': 4.296351407893076e-05, 'samples': 20945920, 'steps': 40909, 'loss/train': 2.3417558670043945} 02/26/2022 06:32:22 - INFO - codeparrot_training - Step 40910: {'lr': 4.2954343157555894e-05, 'samples': 20946432, 'steps': 40910, 'loss/train': 0.9013440012931824} 02/26/2022 06:32:27 - INFO - codeparrot_training - Step 40911: {'lr': 4.2945173123098895e-05, 'samples': 20946944, 'steps': 40911, 'loss/train': 2.2168171405792236} 02/26/2022 06:32:31 - INFO - codeparrot_training - Step 40912: {'lr': 4.293600397559896e-05, 'samples': 20947456, 'steps': 40912, 'loss/train': 0.04781728610396385} 02/26/2022 06:32:38 - INFO - codeparrot_training - Step 40913: {'lr': 4.292683571509548e-05, 'samples': 20947968, 'steps': 40913, 'loss/train': 2.33477783203125} 02/26/2022 06:32:42 - INFO - codeparrot_training - Step 40914: {'lr': 4.291766834162772e-05, 'samples': 20948480, 'steps': 40914, 'loss/train': 2.3277344703674316} 02/26/2022 06:32:47 - INFO - codeparrot_training - Step 40915: {'lr': 4.290850185523487e-05, 'samples': 20948992, 'steps': 40915, 'loss/train': 1.4409483671188354} 02/26/2022 06:32:51 - INFO - codeparrot_training - Step 40916: {'lr': 4.289933625595621e-05, 'samples': 20949504, 'steps': 40916, 'loss/train': 2.034222364425659} 02/26/2022 06:32:56 - INFO - codeparrot_training - Step 40917: {'lr': 4.289017154383107e-05, 'samples': 20950016, 'steps': 40917, 'loss/train': 1.5203218460083008} 02/26/2022 06:33:00 - INFO - codeparrot_training - Step 40918: {'lr': 4.288100771889869e-05, 'samples': 20950528, 'steps': 40918, 'loss/train': 2.0772271156311035} 02/26/2022 06:33:05 - INFO - codeparrot_training - Step 40919: {'lr': 4.287184478119824e-05, 'samples': 20951040, 'steps': 40919, 'loss/train': 1.5221832990646362} 02/26/2022 06:33:09 - INFO - codeparrot_training - Step 40920: {'lr': 4.2862682730769154e-05, 'samples': 20951552, 'steps': 40920, 'loss/train': 1.3872801065444946} 02/26/2022 06:33:14 - INFO - codeparrot_training - Step 40921: {'lr': 4.2853521567650435e-05, 'samples': 20952064, 'steps': 40921, 'loss/train': 1.5734838247299194} 02/26/2022 06:33:18 - INFO - codeparrot_training - Step 40922: {'lr': 4.2844361291881515e-05, 'samples': 20952576, 'steps': 40922, 'loss/train': 1.678175926208496} 02/26/2022 06:33:23 - INFO - codeparrot_training - Step 40923: {'lr': 4.283520190350151e-05, 'samples': 20953088, 'steps': 40923, 'loss/train': 1.7431775331497192} 02/26/2022 06:33:27 - INFO - codeparrot_training - Step 40924: {'lr': 4.282604340254984e-05, 'samples': 20953600, 'steps': 40924, 'loss/train': 1.6529545783996582} 02/26/2022 06:33:33 - INFO - codeparrot_training - Step 40925: {'lr': 4.281688578906551e-05, 'samples': 20954112, 'steps': 40925, 'loss/train': 1.1516813039779663} 02/26/2022 06:33:37 - INFO - codeparrot_training - Step 40926: {'lr': 4.280772906308789e-05, 'samples': 20954624, 'steps': 40926, 'loss/train': 0.6605219841003418} 02/26/2022 06:33:42 - INFO - codeparrot_training - Step 40927: {'lr': 4.279857322465611e-05, 'samples': 20955136, 'steps': 40927, 'loss/train': 0.5205278396606445} 02/26/2022 06:33:46 - INFO - codeparrot_training - Step 40928: {'lr': 4.278941827380953e-05, 'samples': 20955648, 'steps': 40928, 'loss/train': 1.8531779050827026} 02/26/2022 06:33:51 - INFO - codeparrot_training - Step 40929: {'lr': 4.2780264210587254e-05, 'samples': 20956160, 'steps': 40929, 'loss/train': 1.3304390907287598} 02/26/2022 06:33:55 - INFO - codeparrot_training - Step 40930: {'lr': 4.2771111035028564e-05, 'samples': 20956672, 'steps': 40930, 'loss/train': 1.2356091737747192} 02/26/2022 06:34:00 - INFO - codeparrot_training - Step 40931: {'lr': 4.276195874717253e-05, 'samples': 20957184, 'steps': 40931, 'loss/train': 0.5766004920005798} 02/26/2022 06:34:04 - INFO - codeparrot_training - Step 40932: {'lr': 4.275280734705855e-05, 'samples': 20957696, 'steps': 40932, 'loss/train': 1.6547880172729492} 02/26/2022 06:34:09 - INFO - codeparrot_training - Step 40933: {'lr': 4.274365683472572e-05, 'samples': 20958208, 'steps': 40933, 'loss/train': 2.637953042984009} 02/26/2022 06:34:13 - INFO - codeparrot_training - Step 40934: {'lr': 4.273450721021327e-05, 'samples': 20958720, 'steps': 40934, 'loss/train': 1.4736648797988892} 02/26/2022 06:34:18 - INFO - codeparrot_training - Step 40935: {'lr': 4.272535847356032e-05, 'samples': 20959232, 'steps': 40935, 'loss/train': 1.8897048234939575} 02/26/2022 06:34:22 - INFO - codeparrot_training - Step 40936: {'lr': 4.271621062480616e-05, 'samples': 20959744, 'steps': 40936, 'loss/train': 1.274722695350647} 02/26/2022 06:34:27 - INFO - codeparrot_training - Step 40937: {'lr': 4.2707063663989936e-05, 'samples': 20960256, 'steps': 40937, 'loss/train': 2.474787712097168} 02/26/2022 06:34:31 - INFO - codeparrot_training - Step 40938: {'lr': 4.269791759115082e-05, 'samples': 20960768, 'steps': 40938, 'loss/train': 0.8456509709358215} 02/26/2022 06:34:38 - INFO - codeparrot_training - Step 40939: {'lr': 4.268877240632796e-05, 'samples': 20961280, 'steps': 40939, 'loss/train': 2.729900360107422} 02/26/2022 06:34:41 - INFO - codeparrot_training - Step 40940: {'lr': 4.267962810956061e-05, 'samples': 20961792, 'steps': 40940, 'loss/train': 1.9282358884811401} 02/26/2022 06:34:47 - INFO - codeparrot_training - Step 40941: {'lr': 4.2670484700887855e-05, 'samples': 20962304, 'steps': 40941, 'loss/train': 1.52631676197052} 02/26/2022 06:34:50 - INFO - codeparrot_training - Step 40942: {'lr': 4.2661342180349035e-05, 'samples': 20962816, 'steps': 40942, 'loss/train': 1.077835202217102} 02/26/2022 06:34:56 - INFO - codeparrot_training - Step 40943: {'lr': 4.265220054798305e-05, 'samples': 20963328, 'steps': 40943, 'loss/train': 1.5397757291793823} 02/26/2022 06:34:59 - INFO - codeparrot_training - Step 40944: {'lr': 4.2643059803829266e-05, 'samples': 20963840, 'steps': 40944, 'loss/train': 1.5199321508407593} 02/26/2022 06:35:05 - INFO - codeparrot_training - Step 40945: {'lr': 4.2633919947926705e-05, 'samples': 20964352, 'steps': 40945, 'loss/train': 1.4621610641479492} 02/26/2022 06:35:08 - INFO - codeparrot_training - Step 40946: {'lr': 4.262478098031469e-05, 'samples': 20964864, 'steps': 40946, 'loss/train': 2.241328239440918} 02/26/2022 06:35:14 - INFO - codeparrot_training - Step 40947: {'lr': 4.261564290103217e-05, 'samples': 20965376, 'steps': 40947, 'loss/train': 2.9173591136932373} 02/26/2022 06:35:17 - INFO - codeparrot_training - Step 40948: {'lr': 4.260650571011845e-05, 'samples': 20965888, 'steps': 40948, 'loss/train': 1.6773098707199097} 02/26/2022 06:35:23 - INFO - codeparrot_training - Step 40949: {'lr': 4.2597369407612524e-05, 'samples': 20966400, 'steps': 40949, 'loss/train': 2.913886308670044} 02/26/2022 06:35:26 - INFO - codeparrot_training - Step 40950: {'lr': 4.258823399355374e-05, 'samples': 20966912, 'steps': 40950, 'loss/train': 2.2451696395874023} 02/26/2022 06:35:33 - INFO - codeparrot_training - Step 40951: {'lr': 4.257909946798097e-05, 'samples': 20967424, 'steps': 40951, 'loss/train': 1.7844198942184448} 02/26/2022 06:35:37 - INFO - codeparrot_training - Step 40952: {'lr': 4.256996583093356e-05, 'samples': 20967936, 'steps': 40952, 'loss/train': 0.5329646468162537} 02/26/2022 06:35:42 - INFO - codeparrot_training - Step 40953: {'lr': 4.256083308245051e-05, 'samples': 20968448, 'steps': 40953, 'loss/train': 2.287973403930664} 02/26/2022 06:35:45 - INFO - codeparrot_training - Step 40954: {'lr': 4.255170122257096e-05, 'samples': 20968960, 'steps': 40954, 'loss/train': 1.462609887123108} 02/26/2022 06:35:51 - INFO - codeparrot_training - Step 40955: {'lr': 4.2542570251334096e-05, 'samples': 20969472, 'steps': 40955, 'loss/train': 1.0726749897003174} 02/26/2022 06:35:55 - INFO - codeparrot_training - Step 40956: {'lr': 4.253344016877897e-05, 'samples': 20969984, 'steps': 40956, 'loss/train': 1.830863118171692} 02/26/2022 06:36:00 - INFO - codeparrot_training - Step 40957: {'lr': 4.252431097494472e-05, 'samples': 20970496, 'steps': 40957, 'loss/train': 1.4729727506637573} 02/26/2022 06:36:04 - INFO - codeparrot_training - Step 40958: {'lr': 4.2515182669870395e-05, 'samples': 20971008, 'steps': 40958, 'loss/train': 2.107224464416504} 02/26/2022 06:36:09 - INFO - codeparrot_training - Step 40959: {'lr': 4.2506055253595186e-05, 'samples': 20971520, 'steps': 40959, 'loss/train': 1.4418028593063354} 02/26/2022 06:36:13 - INFO - codeparrot_training - Step 40960: {'lr': 4.2496928726158154e-05, 'samples': 20972032, 'steps': 40960, 'loss/train': 1.3643921613693237} 02/26/2022 06:36:19 - INFO - codeparrot_training - Step 40961: {'lr': 4.248780308759839e-05, 'samples': 20972544, 'steps': 40961, 'loss/train': 1.6517362594604492} 02/26/2022 06:36:22 - INFO - codeparrot_training - Step 40962: {'lr': 4.2478678337954935e-05, 'samples': 20973056, 'steps': 40962, 'loss/train': 2.0048162937164307} 02/26/2022 06:36:28 - INFO - codeparrot_training - Step 40963: {'lr': 4.246955447726697e-05, 'samples': 20973568, 'steps': 40963, 'loss/train': 1.8953315019607544} 02/26/2022 06:36:31 - INFO - codeparrot_training - Step 40964: {'lr': 4.246043150557355e-05, 'samples': 20974080, 'steps': 40964, 'loss/train': 1.8628745079040527} 02/26/2022 06:36:37 - INFO - codeparrot_training - Step 40965: {'lr': 4.245130942291373e-05, 'samples': 20974592, 'steps': 40965, 'loss/train': 1.5999078750610352} 02/26/2022 06:36:41 - INFO - codeparrot_training - Step 40966: {'lr': 4.2442188229326526e-05, 'samples': 20975104, 'steps': 40966, 'loss/train': 1.1961758136749268} 02/26/2022 06:36:46 - INFO - codeparrot_training - Step 40967: {'lr': 4.243306792485116e-05, 'samples': 20975616, 'steps': 40967, 'loss/train': 1.8956518173217773} 02/26/2022 06:36:50 - INFO - codeparrot_training - Step 40968: {'lr': 4.242394850952661e-05, 'samples': 20976128, 'steps': 40968, 'loss/train': 1.1217402219772339} 02/26/2022 06:36:55 - INFO - codeparrot_training - Step 40969: {'lr': 4.241482998339197e-05, 'samples': 20976640, 'steps': 40969, 'loss/train': 1.3871036767959595} 02/26/2022 06:36:59 - INFO - codeparrot_training - Step 40970: {'lr': 4.240571234648621e-05, 'samples': 20977152, 'steps': 40970, 'loss/train': 1.8450934886932373} 02/26/2022 06:37:04 - INFO - codeparrot_training - Step 40971: {'lr': 4.2396595598848505e-05, 'samples': 20977664, 'steps': 40971, 'loss/train': 2.474783182144165} 02/26/2022 06:37:08 - INFO - codeparrot_training - Step 40972: {'lr': 4.238747974051787e-05, 'samples': 20978176, 'steps': 40972, 'loss/train': 1.4702836275100708} 02/26/2022 06:37:13 - INFO - codeparrot_training - Step 40973: {'lr': 4.2378364771533366e-05, 'samples': 20978688, 'steps': 40973, 'loss/train': 1.8752280473709106} 02/26/2022 06:37:17 - INFO - codeparrot_training - Step 40974: {'lr': 4.2369250691933955e-05, 'samples': 20979200, 'steps': 40974, 'loss/train': 1.3975797891616821} 02/26/2022 06:37:23 - INFO - codeparrot_training - Step 40975: {'lr': 4.236013750175879e-05, 'samples': 20979712, 'steps': 40975, 'loss/train': 1.2419483661651611} 02/26/2022 06:37:26 - INFO - codeparrot_training - Step 40976: {'lr': 4.2351025201046804e-05, 'samples': 20980224, 'steps': 40976, 'loss/train': 1.7040468454360962} 02/26/2022 06:37:32 - INFO - codeparrot_training - Step 40977: {'lr': 4.2341913789837235e-05, 'samples': 20980736, 'steps': 40977, 'loss/train': 1.8875755071640015} 02/26/2022 06:37:35 - INFO - codeparrot_training - Step 40978: {'lr': 4.2332803268168826e-05, 'samples': 20981248, 'steps': 40978, 'loss/train': 1.6545933485031128} 02/26/2022 06:37:41 - INFO - codeparrot_training - Step 40979: {'lr': 4.23236936360808e-05, 'samples': 20981760, 'steps': 40979, 'loss/train': 0.6189851760864258} 02/26/2022 06:37:44 - INFO - codeparrot_training - Step 40980: {'lr': 4.231458489361206e-05, 'samples': 20982272, 'steps': 40980, 'loss/train': 0.4759628176689148} 02/26/2022 06:37:50 - INFO - codeparrot_training - Step 40981: {'lr': 4.230547704080184e-05, 'samples': 20982784, 'steps': 40981, 'loss/train': 1.9501804113388062} 02/26/2022 06:37:53 - INFO - codeparrot_training - Step 40982: {'lr': 4.229637007768888e-05, 'samples': 20983296, 'steps': 40982, 'loss/train': 2.1409239768981934} 02/26/2022 06:37:59 - INFO - codeparrot_training - Step 40983: {'lr': 4.228726400431237e-05, 'samples': 20983808, 'steps': 40983, 'loss/train': 1.6045479774475098} 02/26/2022 06:38:02 - INFO - codeparrot_training - Step 40984: {'lr': 4.2278158820711205e-05, 'samples': 20984320, 'steps': 40984, 'loss/train': 0.621855616569519} 02/26/2022 06:38:08 - INFO - codeparrot_training - Step 40985: {'lr': 4.226905452692456e-05, 'samples': 20984832, 'steps': 40985, 'loss/train': 1.9347269535064697} 02/26/2022 06:38:11 - INFO - codeparrot_training - Step 40986: {'lr': 4.225995112299122e-05, 'samples': 20985344, 'steps': 40986, 'loss/train': 1.9254193305969238} 02/26/2022 06:38:18 - INFO - codeparrot_training - Step 40987: {'lr': 4.225084860895031e-05, 'samples': 20985856, 'steps': 40987, 'loss/train': 1.4470301866531372} 02/26/2022 06:38:21 - INFO - codeparrot_training - Step 40988: {'lr': 4.224174698484079e-05, 'samples': 20986368, 'steps': 40988, 'loss/train': 2.0797104835510254} 02/26/2022 06:38:27 - INFO - codeparrot_training - Step 40989: {'lr': 4.2232646250701666e-05, 'samples': 20986880, 'steps': 40989, 'loss/train': 2.3309998512268066} 02/26/2022 06:38:30 - INFO - codeparrot_training - Step 40990: {'lr': 4.2223546406571944e-05, 'samples': 20987392, 'steps': 40990, 'loss/train': 0.6763462424278259} 02/26/2022 06:38:36 - INFO - codeparrot_training - Step 40991: {'lr': 4.221444745249056e-05, 'samples': 20987904, 'steps': 40991, 'loss/train': 2.250633955001831} 02/26/2022 06:38:39 - INFO - codeparrot_training - Step 40992: {'lr': 4.220534938849643e-05, 'samples': 20988416, 'steps': 40992, 'loss/train': 1.4728803634643555} 02/26/2022 06:38:44 - INFO - codeparrot_training - Step 40993: {'lr': 4.219625221462869e-05, 'samples': 20988928, 'steps': 40993, 'loss/train': 2.3220465183258057} 02/26/2022 06:38:48 - INFO - codeparrot_training - Step 40994: {'lr': 4.21871559309262e-05, 'samples': 20989440, 'steps': 40994, 'loss/train': 1.1009777784347534} 02/26/2022 06:38:53 - INFO - codeparrot_training - Step 40995: {'lr': 4.217806053742795e-05, 'samples': 20989952, 'steps': 40995, 'loss/train': 0.7125465869903564} 02/26/2022 06:38:57 - INFO - codeparrot_training - Step 40996: {'lr': 4.216896603417286e-05, 'samples': 20990464, 'steps': 40996, 'loss/train': 1.5867514610290527} 02/26/2022 06:39:04 - INFO - codeparrot_training - Step 40997: {'lr': 4.215987242119995e-05, 'samples': 20990976, 'steps': 40997, 'loss/train': 0.9622074961662292} 02/26/2022 06:39:07 - INFO - codeparrot_training - Step 40998: {'lr': 4.2150779698548165e-05, 'samples': 20991488, 'steps': 40998, 'loss/train': 1.6610727310180664} 02/26/2022 06:39:13 - INFO - codeparrot_training - Step 40999: {'lr': 4.214168786625644e-05, 'samples': 20992000, 'steps': 40999, 'loss/train': 1.473726749420166} 02/26/2022 06:39:13 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 06:39:31 - WARNING - huggingface_hub.repository - Several commits (41) will be pushed upstream. 02/26/2022 06:39:31 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 06:40:06 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy e25c2d7..2427a89 floral-grass-11 -> floral-grass-11 02/26/2022 06:40:11 - INFO - codeparrot_training - Step 41000: {'lr': 4.213259692436367e-05, 'samples': 20992512, 'steps': 41000, 'loss/train': 2.428166151046753} 02/26/2022 06:40:17 - INFO - codeparrot_training - Step 41001: {'lr': 4.212350687290892e-05, 'samples': 20993024, 'steps': 41001, 'loss/train': 1.1785485744476318} 02/26/2022 06:40:20 - INFO - codeparrot_training - Step 41002: {'lr': 4.211441771193103e-05, 'samples': 20993536, 'steps': 41002, 'loss/train': 2.117093324661255} 02/26/2022 06:40:26 - INFO - codeparrot_training - Step 41003: {'lr': 4.210532944146897e-05, 'samples': 20994048, 'steps': 41003, 'loss/train': 2.542660713195801} 02/26/2022 06:40:29 - INFO - codeparrot_training - Step 41004: {'lr': 4.209624206156168e-05, 'samples': 20994560, 'steps': 41004, 'loss/train': 0.4339028596878052} 02/26/2022 06:40:35 - INFO - codeparrot_training - Step 41005: {'lr': 4.208715557224799e-05, 'samples': 20995072, 'steps': 41005, 'loss/train': 2.0721898078918457} 02/26/2022 06:40:38 - INFO - codeparrot_training - Step 41006: {'lr': 4.207806997356697e-05, 'samples': 20995584, 'steps': 41006, 'loss/train': 0.9623445868492126} 02/26/2022 06:40:45 - INFO - codeparrot_training - Step 41007: {'lr': 4.20689852655575e-05, 'samples': 20996096, 'steps': 41007, 'loss/train': 1.8240349292755127} 02/26/2022 06:40:48 - INFO - codeparrot_training - Step 41008: {'lr': 4.2059901448258445e-05, 'samples': 20996608, 'steps': 41008, 'loss/train': 1.4623123407363892} 02/26/2022 06:40:53 - INFO - codeparrot_training - Step 41009: {'lr': 4.20508185217087e-05, 'samples': 20997120, 'steps': 41009, 'loss/train': 0.836247444152832} 02/26/2022 06:40:57 - INFO - codeparrot_training - Step 41010: {'lr': 4.204173648594728e-05, 'samples': 20997632, 'steps': 41010, 'loss/train': 1.8821682929992676} 02/26/2022 06:41:02 - INFO - codeparrot_training - Step 41011: {'lr': 4.2032655341012996e-05, 'samples': 20998144, 'steps': 41011, 'loss/train': 2.2221384048461914} 02/26/2022 06:41:06 - INFO - codeparrot_training - Step 41012: {'lr': 4.202357508694482e-05, 'samples': 20998656, 'steps': 41012, 'loss/train': 2.626143455505371} 02/26/2022 06:41:11 - INFO - codeparrot_training - Step 41013: {'lr': 4.2014495723781524e-05, 'samples': 20999168, 'steps': 41013, 'loss/train': 1.1708762645721436} 02/26/2022 06:41:15 - INFO - codeparrot_training - Step 41014: {'lr': 4.200541725156215e-05, 'samples': 20999680, 'steps': 41014, 'loss/train': 0.8321356177330017} 02/26/2022 06:41:20 - INFO - codeparrot_training - Step 41015: {'lr': 4.199633967032546e-05, 'samples': 21000192, 'steps': 41015, 'loss/train': 1.5895556211471558} 02/26/2022 06:41:26 - INFO - codeparrot_training - Step 41016: {'lr': 4.198726298011052e-05, 'samples': 21000704, 'steps': 41016, 'loss/train': 1.6041032075881958} 02/26/2022 06:41:30 - INFO - codeparrot_training - Step 41017: {'lr': 4.1978187180955974e-05, 'samples': 21001216, 'steps': 41017, 'loss/train': 2.1447529792785645} 02/26/2022 06:41:35 - INFO - codeparrot_training - Step 41018: {'lr': 4.19691122729009e-05, 'samples': 21001728, 'steps': 41018, 'loss/train': 2.028188705444336} 02/26/2022 06:41:39 - INFO - codeparrot_training - Step 41019: {'lr': 4.196003825598402e-05, 'samples': 21002240, 'steps': 41019, 'loss/train': 3.0181140899658203} 02/26/2022 06:41:44 - INFO - codeparrot_training - Step 41020: {'lr': 4.1950965130244376e-05, 'samples': 21002752, 'steps': 41020, 'loss/train': 1.7762588262557983} 02/26/2022 06:41:48 - INFO - codeparrot_training - Step 41021: {'lr': 4.194189289572065e-05, 'samples': 21003264, 'steps': 41021, 'loss/train': 2.4327194690704346} 02/26/2022 06:41:54 - INFO - codeparrot_training - Step 41022: {'lr': 4.193282155245182e-05, 'samples': 21003776, 'steps': 41022, 'loss/train': 2.31196665763855} 02/26/2022 06:41:58 - INFO - codeparrot_training - Step 41023: {'lr': 4.192375110047667e-05, 'samples': 21004288, 'steps': 41023, 'loss/train': 1.6307587623596191} 02/26/2022 06:42:03 - INFO - codeparrot_training - Step 41024: {'lr': 4.191468153983419e-05, 'samples': 21004800, 'steps': 41024, 'loss/train': 0.5155414938926697} 02/26/2022 06:42:07 - INFO - codeparrot_training - Step 41025: {'lr': 4.190561287056305e-05, 'samples': 21005312, 'steps': 41025, 'loss/train': 1.6448475122451782} 02/26/2022 06:42:12 - INFO - codeparrot_training - Step 41026: {'lr': 4.1896545092702244e-05, 'samples': 21005824, 'steps': 41026, 'loss/train': 2.102656364440918} 02/26/2022 06:42:16 - INFO - codeparrot_training - Step 41027: {'lr': 4.188747820629049e-05, 'samples': 21006336, 'steps': 41027, 'loss/train': 0.018176639452576637} 02/26/2022 06:42:22 - INFO - codeparrot_training - Step 41028: {'lr': 4.187841221136676e-05, 'samples': 21006848, 'steps': 41028, 'loss/train': 2.1290993690490723} 02/26/2022 06:42:25 - INFO - codeparrot_training - Step 41029: {'lr': 4.186934710796983e-05, 'samples': 21007360, 'steps': 41029, 'loss/train': 0.5570828318595886} 02/26/2022 06:42:29 - INFO - codeparrot_training - Step 41030: {'lr': 4.1860282896138524e-05, 'samples': 21007872, 'steps': 41030, 'loss/train': 1.7246832847595215} 02/26/2022 06:42:34 - INFO - codeparrot_training - Step 41031: {'lr': 4.185121957591162e-05, 'samples': 21008384, 'steps': 41031, 'loss/train': 2.7483699321746826} 02/26/2022 06:42:38 - INFO - codeparrot_training - Step 41032: {'lr': 4.184215714732803e-05, 'samples': 21008896, 'steps': 41032, 'loss/train': 1.555988073348999} 02/26/2022 06:42:44 - INFO - codeparrot_training - Step 41033: {'lr': 4.183309561042658e-05, 'samples': 21009408, 'steps': 41033, 'loss/train': 1.2177904844284058} 02/26/2022 06:42:50 - INFO - codeparrot_training - Step 41034: {'lr': 4.1824034965246035e-05, 'samples': 21009920, 'steps': 41034, 'loss/train': 1.9417407512664795} 02/26/2022 06:42:53 - INFO - codeparrot_training - Step 41035: {'lr': 4.181497521182514e-05, 'samples': 21010432, 'steps': 41035, 'loss/train': 1.621850609779358} 02/26/2022 06:42:59 - INFO - codeparrot_training - Step 41036: {'lr': 4.180591635020287e-05, 'samples': 21010944, 'steps': 41036, 'loss/train': 1.9137673377990723} 02/26/2022 06:43:02 - INFO - codeparrot_training - Step 41037: {'lr': 4.179685838041794e-05, 'samples': 21011456, 'steps': 41037, 'loss/train': 2.069483518600464} 02/26/2022 06:43:08 - INFO - codeparrot_training - Step 41038: {'lr': 4.1787801302509146e-05, 'samples': 21011968, 'steps': 41038, 'loss/train': 0.8556708693504333} 02/26/2022 06:43:11 - INFO - codeparrot_training - Step 41039: {'lr': 4.177874511651525e-05, 'samples': 21012480, 'steps': 41039, 'loss/train': 3.206566095352173} 02/26/2022 06:43:17 - INFO - codeparrot_training - Step 41040: {'lr': 4.176968982247514e-05, 'samples': 21012992, 'steps': 41040, 'loss/train': 1.8868523836135864} 02/26/2022 06:43:20 - INFO - codeparrot_training - Step 41041: {'lr': 4.176063542042757e-05, 'samples': 21013504, 'steps': 41041, 'loss/train': 2.053576707839966} 02/26/2022 06:43:24 - INFO - codeparrot_training - Step 41042: {'lr': 4.175158191041128e-05, 'samples': 21014016, 'steps': 41042, 'loss/train': 8.96193790435791} 02/26/2022 06:43:30 - INFO - codeparrot_training - Step 41043: {'lr': 4.174252929246508e-05, 'samples': 21014528, 'steps': 41043, 'loss/train': 1.664588212966919} 02/26/2022 06:43:33 - INFO - codeparrot_training - Step 41044: {'lr': 4.173347756662779e-05, 'samples': 21015040, 'steps': 41044, 'loss/train': 2.246711015701294} 02/26/2022 06:43:39 - INFO - codeparrot_training - Step 41045: {'lr': 4.172442673293814e-05, 'samples': 21015552, 'steps': 41045, 'loss/train': 1.0297152996063232} 02/26/2022 06:43:42 - INFO - codeparrot_training - Step 41046: {'lr': 4.1715376791434916e-05, 'samples': 21016064, 'steps': 41046, 'loss/train': 1.606587290763855} 02/26/2022 06:43:48 - INFO - codeparrot_training - Step 41047: {'lr': 4.170632774215682e-05, 'samples': 21016576, 'steps': 41047, 'loss/train': 1.7754837274551392} 02/26/2022 06:43:51 - INFO - codeparrot_training - Step 41048: {'lr': 4.169727958514275e-05, 'samples': 21017088, 'steps': 41048, 'loss/train': 2.855811595916748} 02/26/2022 06:43:57 - INFO - codeparrot_training - Step 41049: {'lr': 4.16882323204314e-05, 'samples': 21017600, 'steps': 41049, 'loss/train': 1.797804355621338} 02/26/2022 06:44:03 - INFO - codeparrot_training - Step 41050: {'lr': 4.167918594806142e-05, 'samples': 21018112, 'steps': 41050, 'loss/train': 1.7380183935165405} 02/26/2022 06:44:06 - INFO - codeparrot_training - Step 41051: {'lr': 4.16701404680718e-05, 'samples': 21018624, 'steps': 41051, 'loss/train': 1.6924335956573486} 02/26/2022 06:44:10 - INFO - codeparrot_training - Step 41052: {'lr': 4.1661095880501014e-05, 'samples': 21019136, 'steps': 41052, 'loss/train': 1.2786972522735596} 02/26/2022 06:44:16 - INFO - codeparrot_training - Step 41053: {'lr': 4.1652052185388e-05, 'samples': 21019648, 'steps': 41053, 'loss/train': 1.8639967441558838} 02/26/2022 06:44:19 - INFO - codeparrot_training - Step 41054: {'lr': 4.164300938277138e-05, 'samples': 21020160, 'steps': 41054, 'loss/train': 0.5331993103027344} 02/26/2022 06:44:25 - INFO - codeparrot_training - Step 41055: {'lr': 4.1633967472690075e-05, 'samples': 21020672, 'steps': 41055, 'loss/train': 2.0341267585754395} 02/26/2022 06:44:28 - INFO - codeparrot_training - Step 41056: {'lr': 4.162492645518256e-05, 'samples': 21021184, 'steps': 41056, 'loss/train': 2.191824197769165} 02/26/2022 06:44:34 - INFO - codeparrot_training - Step 41057: {'lr': 4.161588633028776e-05, 'samples': 21021696, 'steps': 41057, 'loss/train': 1.7723459005355835} 02/26/2022 06:44:37 - INFO - codeparrot_training - Step 41058: {'lr': 4.160684709804427e-05, 'samples': 21022208, 'steps': 41058, 'loss/train': 1.047770380973816} 02/26/2022 06:44:43 - INFO - codeparrot_training - Step 41059: {'lr': 4.1597808758491006e-05, 'samples': 21022720, 'steps': 41059, 'loss/train': 1.1358656883239746} 02/26/2022 06:44:48 - INFO - codeparrot_training - Step 41060: {'lr': 4.1588771311666415e-05, 'samples': 21023232, 'steps': 41060, 'loss/train': 1.1257461309432983} 02/26/2022 06:44:52 - INFO - codeparrot_training - Step 41061: {'lr': 4.157973475760943e-05, 'samples': 21023744, 'steps': 41061, 'loss/train': 2.379237651824951} 02/26/2022 06:44:57 - INFO - codeparrot_training - Step 41062: {'lr': 4.1570699096358606e-05, 'samples': 21024256, 'steps': 41062, 'loss/train': 1.2789474725723267} 02/26/2022 06:45:01 - INFO - codeparrot_training - Step 41063: {'lr': 4.1561664327952774e-05, 'samples': 21024768, 'steps': 41063, 'loss/train': 1.6425762176513672} 02/26/2022 06:45:06 - INFO - codeparrot_training - Step 41064: {'lr': 4.15526304524306e-05, 'samples': 21025280, 'steps': 41064, 'loss/train': 2.6840994358062744} 02/26/2022 06:45:10 - INFO - codeparrot_training - Step 41065: {'lr': 4.1543597469830777e-05, 'samples': 21025792, 'steps': 41065, 'loss/train': 1.8795808553695679} 02/26/2022 06:45:15 - INFO - codeparrot_training - Step 41066: {'lr': 4.15345653801919e-05, 'samples': 21026304, 'steps': 41066, 'loss/train': 2.257610321044922} 02/26/2022 06:45:19 - INFO - codeparrot_training - Step 41067: {'lr': 4.1525534183552846e-05, 'samples': 21026816, 'steps': 41067, 'loss/train': 2.4211978912353516} 02/26/2022 06:45:25 - INFO - codeparrot_training - Step 41068: {'lr': 4.151650387995218e-05, 'samples': 21027328, 'steps': 41068, 'loss/train': 1.8978255987167358} 02/26/2022 06:45:29 - INFO - codeparrot_training - Step 41069: {'lr': 4.1507474469428625e-05, 'samples': 21027840, 'steps': 41069, 'loss/train': 1.6763325929641724} 02/26/2022 06:45:34 - INFO - codeparrot_training - Step 41070: {'lr': 4.149844595202076e-05, 'samples': 21028352, 'steps': 41070, 'loss/train': 1.2078220844268799} 02/26/2022 06:45:38 - INFO - codeparrot_training - Step 41071: {'lr': 4.1489418327767455e-05, 'samples': 21028864, 'steps': 41071, 'loss/train': 1.511915683746338} 02/26/2022 06:45:43 - INFO - codeparrot_training - Step 41072: {'lr': 4.148039159670722e-05, 'samples': 21029376, 'steps': 41072, 'loss/train': 1.2254576683044434} 02/26/2022 06:45:47 - INFO - codeparrot_training - Step 41073: {'lr': 4.14713657588788e-05, 'samples': 21029888, 'steps': 41073, 'loss/train': 1.8093395233154297} 02/26/2022 06:45:52 - INFO - codeparrot_training - Step 41074: {'lr': 4.1462340814320766e-05, 'samples': 21030400, 'steps': 41074, 'loss/train': 1.608319640159607} 02/26/2022 06:45:56 - INFO - codeparrot_training - Step 41075: {'lr': 4.1453316763071927e-05, 'samples': 21030912, 'steps': 41075, 'loss/train': 1.9729677438735962} 02/26/2022 06:46:01 - INFO - codeparrot_training - Step 41076: {'lr': 4.1444293605170766e-05, 'samples': 21031424, 'steps': 41076, 'loss/train': 1.6831915378570557} 02/26/2022 06:46:05 - INFO - codeparrot_training - Step 41077: {'lr': 4.143527134065617e-05, 'samples': 21031936, 'steps': 41077, 'loss/train': 1.9879294633865356} 02/26/2022 06:46:12 - INFO - codeparrot_training - Step 41078: {'lr': 4.142624996956651e-05, 'samples': 21032448, 'steps': 41078, 'loss/train': 5.675451278686523} 02/26/2022 06:46:15 - INFO - codeparrot_training - Step 41079: {'lr': 4.141722949194063e-05, 'samples': 21032960, 'steps': 41079, 'loss/train': 0.2515413165092468} 02/26/2022 06:46:21 - INFO - codeparrot_training - Step 41080: {'lr': 4.140820990781705e-05, 'samples': 21033472, 'steps': 41080, 'loss/train': 1.8496679067611694} 02/26/2022 06:46:24 - INFO - codeparrot_training - Step 41081: {'lr': 4.139919121723459e-05, 'samples': 21033984, 'steps': 41081, 'loss/train': 1.4130698442459106} 02/26/2022 06:46:30 - INFO - codeparrot_training - Step 41082: {'lr': 4.139017342023163e-05, 'samples': 21034496, 'steps': 41082, 'loss/train': 1.170663833618164} 02/26/2022 06:46:33 - INFO - codeparrot_training - Step 41083: {'lr': 4.1381156516846984e-05, 'samples': 21035008, 'steps': 41083, 'loss/train': 1.9587359428405762} 02/26/2022 06:46:39 - INFO - codeparrot_training - Step 41084: {'lr': 4.137214050711921e-05, 'samples': 21035520, 'steps': 41084, 'loss/train': 1.8764231204986572} 02/26/2022 06:46:42 - INFO - codeparrot_training - Step 41085: {'lr': 4.1363125391086915e-05, 'samples': 21036032, 'steps': 41085, 'loss/train': 0.9248769879341125} 02/26/2022 06:46:48 - INFO - codeparrot_training - Step 41086: {'lr': 4.135411116878876e-05, 'samples': 21036544, 'steps': 41086, 'loss/train': 1.3095935583114624} 02/26/2022 06:46:51 - INFO - codeparrot_training - Step 41087: {'lr': 4.1345097840263355e-05, 'samples': 21037056, 'steps': 41087, 'loss/train': 1.5368452072143555} 02/26/2022 06:46:58 - INFO - codeparrot_training - Step 41088: {'lr': 4.133608540554931e-05, 'samples': 21037568, 'steps': 41088, 'loss/train': 1.305829405784607} 02/26/2022 06:47:01 - INFO - codeparrot_training - Step 41089: {'lr': 4.132707386468512e-05, 'samples': 21038080, 'steps': 41089, 'loss/train': 1.9796684980392456} 02/26/2022 06:47:07 - INFO - codeparrot_training - Step 41090: {'lr': 4.131806321770959e-05, 'samples': 21038592, 'steps': 41090, 'loss/train': 1.7605584859848022} 02/26/2022 06:47:10 - INFO - codeparrot_training - Step 41091: {'lr': 4.1309053464661186e-05, 'samples': 21039104, 'steps': 41091, 'loss/train': 3.3215723037719727} 02/26/2022 06:47:16 - INFO - codeparrot_training - Step 41092: {'lr': 4.130004460557854e-05, 'samples': 21039616, 'steps': 41092, 'loss/train': 2.083867311477661} 02/26/2022 06:47:19 - INFO - codeparrot_training - Step 41093: {'lr': 4.129103664050016e-05, 'samples': 21040128, 'steps': 41093, 'loss/train': 2.3799123764038086} 02/26/2022 06:47:25 - INFO - codeparrot_training - Step 41094: {'lr': 4.128202956946478e-05, 'samples': 21040640, 'steps': 41094, 'loss/train': 1.8778314590454102} 02/26/2022 06:47:28 - INFO - codeparrot_training - Step 41095: {'lr': 4.1273023392510906e-05, 'samples': 21041152, 'steps': 41095, 'loss/train': 1.3559863567352295} 02/26/2022 06:47:34 - INFO - codeparrot_training - Step 41096: {'lr': 4.126401810967711e-05, 'samples': 21041664, 'steps': 41096, 'loss/train': 1.9984766244888306} 02/26/2022 06:47:37 - INFO - codeparrot_training - Step 41097: {'lr': 4.125501372100193e-05, 'samples': 21042176, 'steps': 41097, 'loss/train': 1.7497897148132324} 02/26/2022 06:47:44 - INFO - codeparrot_training - Step 41098: {'lr': 4.124601022652405e-05, 'samples': 21042688, 'steps': 41098, 'loss/train': 1.542076826095581} 02/26/2022 06:47:47 - INFO - codeparrot_training - Step 41099: {'lr': 4.123700762628196e-05, 'samples': 21043200, 'steps': 41099, 'loss/train': 1.0483652353286743} 02/26/2022 06:47:53 - INFO - codeparrot_training - Step 41100: {'lr': 4.122800592031425e-05, 'samples': 21043712, 'steps': 41100, 'loss/train': 1.0628455877304077} 02/26/2022 06:47:56 - INFO - codeparrot_training - Step 41101: {'lr': 4.121900510865942e-05, 'samples': 21044224, 'steps': 41101, 'loss/train': 0.1522289514541626} 02/26/2022 06:48:02 - INFO - codeparrot_training - Step 41102: {'lr': 4.121000519135615e-05, 'samples': 21044736, 'steps': 41102, 'loss/train': 1.7213683128356934} 02/26/2022 06:48:06 - INFO - codeparrot_training - Step 41103: {'lr': 4.120100616844289e-05, 'samples': 21045248, 'steps': 41103, 'loss/train': 1.6357460021972656} 02/26/2022 06:48:11 - INFO - codeparrot_training - Step 41104: {'lr': 4.1192008039958234e-05, 'samples': 21045760, 'steps': 41104, 'loss/train': 1.7782361507415771} 02/26/2022 06:48:15 - INFO - codeparrot_training - Step 41105: {'lr': 4.118301080594064e-05, 'samples': 21046272, 'steps': 41105, 'loss/train': 0.8308981657028198} 02/26/2022 06:48:21 - INFO - codeparrot_training - Step 41106: {'lr': 4.117401446642877e-05, 'samples': 21046784, 'steps': 41106, 'loss/train': 2.545792818069458} 02/26/2022 06:48:25 - INFO - codeparrot_training - Step 41107: {'lr': 4.116501902146113e-05, 'samples': 21047296, 'steps': 41107, 'loss/train': 2.341170072555542} 02/26/2022 06:48:28 - INFO - codeparrot_training - Step 41108: {'lr': 4.1156024471076245e-05, 'samples': 21047808, 'steps': 41108, 'loss/train': 1.9347031116485596} 02/26/2022 06:48:34 - INFO - codeparrot_training - Step 41109: {'lr': 4.114703081531254e-05, 'samples': 21048320, 'steps': 41109, 'loss/train': 1.3439594507217407} 02/26/2022 06:48:37 - INFO - codeparrot_training - Step 41110: {'lr': 4.1138038054208735e-05, 'samples': 21048832, 'steps': 41110, 'loss/train': 2.016472816467285} 02/26/2022 06:48:43 - INFO - codeparrot_training - Step 41111: {'lr': 4.112904618780317e-05, 'samples': 21049344, 'steps': 41111, 'loss/train': 1.6328643560409546} 02/26/2022 06:48:46 - INFO - codeparrot_training - Step 41112: {'lr': 4.11200552161346e-05, 'samples': 21049856, 'steps': 41112, 'loss/train': 0.8337056040763855} 02/26/2022 06:48:52 - INFO - codeparrot_training - Step 41113: {'lr': 4.111106513924123e-05, 'samples': 21050368, 'steps': 41113, 'loss/train': 2.245283603668213} 02/26/2022 06:48:56 - INFO - codeparrot_training - Step 41114: {'lr': 4.110207595716181e-05, 'samples': 21050880, 'steps': 41114, 'loss/train': 0.634177029132843} 02/26/2022 06:49:02 - INFO - codeparrot_training - Step 41115: {'lr': 4.1093087669934695e-05, 'samples': 21051392, 'steps': 41115, 'loss/train': 1.2829649448394775} 02/26/2022 06:49:05 - INFO - codeparrot_training - Step 41116: {'lr': 4.108410027759857e-05, 'samples': 21051904, 'steps': 41116, 'loss/train': 2.4708104133605957} 02/26/2022 06:49:11 - INFO - codeparrot_training - Step 41117: {'lr': 4.107511378019171e-05, 'samples': 21052416, 'steps': 41117, 'loss/train': 2.2025561332702637} 02/26/2022 06:49:14 - INFO - codeparrot_training - Step 41118: {'lr': 4.1066128177752764e-05, 'samples': 21052928, 'steps': 41118, 'loss/train': 1.4762216806411743} 02/26/2022 06:49:20 - INFO - codeparrot_training - Step 41119: {'lr': 4.1057143470320134e-05, 'samples': 21053440, 'steps': 41119, 'loss/train': 3.272343158721924} 02/26/2022 06:49:23 - INFO - codeparrot_training - Step 41120: {'lr': 4.104815965793249e-05, 'samples': 21053952, 'steps': 41120, 'loss/train': 1.0439702272415161} 02/26/2022 06:49:29 - INFO - codeparrot_training - Step 41121: {'lr': 4.1039176740628044e-05, 'samples': 21054464, 'steps': 41121, 'loss/train': 1.5443142652511597} 02/26/2022 06:49:33 - INFO - codeparrot_training - Step 41122: {'lr': 4.103019471844546e-05, 'samples': 21054976, 'steps': 41122, 'loss/train': 1.2620333433151245} 02/26/2022 06:49:38 - INFO - codeparrot_training - Step 41123: {'lr': 4.102121359142313e-05, 'samples': 21055488, 'steps': 41123, 'loss/train': 2.318340301513672} 02/26/2022 06:49:41 - INFO - codeparrot_training - Step 41124: {'lr': 4.101223335959961e-05, 'samples': 21056000, 'steps': 41124, 'loss/train': 0.699405312538147} 02/26/2022 06:49:48 - INFO - codeparrot_training - Step 41125: {'lr': 4.100325402301333e-05, 'samples': 21056512, 'steps': 41125, 'loss/train': 0.6871204972267151} 02/26/2022 06:49:51 - INFO - codeparrot_training - Step 41126: {'lr': 4.0994275581702736e-05, 'samples': 21057024, 'steps': 41126, 'loss/train': 2.4336421489715576} 02/26/2022 06:49:57 - INFO - codeparrot_training - Step 41127: {'lr': 4.098529803570625e-05, 'samples': 21057536, 'steps': 41127, 'loss/train': 2.231067180633545} 02/26/2022 06:50:00 - INFO - codeparrot_training - Step 41128: {'lr': 4.097632138506244e-05, 'samples': 21058048, 'steps': 41128, 'loss/train': 1.0571078062057495} 02/26/2022 06:50:06 - INFO - codeparrot_training - Step 41129: {'lr': 4.0967345629809694e-05, 'samples': 21058560, 'steps': 41129, 'loss/train': 0.187117800116539} 02/26/2022 06:50:09 - INFO - codeparrot_training - Step 41130: {'lr': 4.095837076998646e-05, 'samples': 21059072, 'steps': 41130, 'loss/train': 1.3468377590179443} 02/26/2022 06:50:15 - INFO - codeparrot_training - Step 41131: {'lr': 4.094939680563117e-05, 'samples': 21059584, 'steps': 41131, 'loss/train': 1.6845265626907349} 02/26/2022 06:50:18 - INFO - codeparrot_training - Step 41132: {'lr': 4.094042373678225e-05, 'samples': 21060096, 'steps': 41132, 'loss/train': 1.4617278575897217} 02/26/2022 06:50:24 - INFO - codeparrot_training - Step 41133: {'lr': 4.093145156347822e-05, 'samples': 21060608, 'steps': 41133, 'loss/train': 2.399596691131592} 02/26/2022 06:50:27 - INFO - codeparrot_training - Step 41134: {'lr': 4.092248028575746e-05, 'samples': 21061120, 'steps': 41134, 'loss/train': 2.0908172130584717} 02/26/2022 06:50:33 - INFO - codeparrot_training - Step 41135: {'lr': 4.091350990365841e-05, 'samples': 21061632, 'steps': 41135, 'loss/train': 0.8410221934318542} 02/26/2022 06:50:37 - INFO - codeparrot_training - Step 41136: {'lr': 4.090454041721942e-05, 'samples': 21062144, 'steps': 41136, 'loss/train': 1.5373375415802002} 02/26/2022 06:50:42 - INFO - codeparrot_training - Step 41137: {'lr': 4.0895571826479045e-05, 'samples': 21062656, 'steps': 41137, 'loss/train': 1.8123527765274048} 02/26/2022 06:50:46 - INFO - codeparrot_training - Step 41138: {'lr': 4.0886604131475655e-05, 'samples': 21063168, 'steps': 41138, 'loss/train': 1.6958438158035278} 02/26/2022 06:50:52 - INFO - codeparrot_training - Step 41139: {'lr': 4.087763733224767e-05, 'samples': 21063680, 'steps': 41139, 'loss/train': 1.895423173904419} 02/26/2022 06:50:55 - INFO - codeparrot_training - Step 41140: {'lr': 4.086867142883338e-05, 'samples': 21064192, 'steps': 41140, 'loss/train': 0.5312969088554382} 02/26/2022 06:51:01 - INFO - codeparrot_training - Step 41141: {'lr': 4.0859706421271376e-05, 'samples': 21064704, 'steps': 41141, 'loss/train': 2.028679847717285} 02/26/2022 06:51:04 - INFO - codeparrot_training - Step 41142: {'lr': 4.085074230959998e-05, 'samples': 21065216, 'steps': 41142, 'loss/train': 2.403795003890991} 02/26/2022 06:51:10 - INFO - codeparrot_training - Step 41143: {'lr': 4.084177909385758e-05, 'samples': 21065728, 'steps': 41143, 'loss/train': 1.3857519626617432} 02/26/2022 06:51:13 - INFO - codeparrot_training - Step 41144: {'lr': 4.083281677408254e-05, 'samples': 21066240, 'steps': 41144, 'loss/train': 1.2701811790466309} 02/26/2022 06:51:19 - INFO - codeparrot_training - Step 41145: {'lr': 4.0823855350313356e-05, 'samples': 21066752, 'steps': 41145, 'loss/train': 1.6163908243179321} 02/26/2022 06:51:23 - INFO - codeparrot_training - Step 41146: {'lr': 4.081489482258829e-05, 'samples': 21067264, 'steps': 41146, 'loss/train': 0.041384462267160416} 02/26/2022 06:51:28 - INFO - codeparrot_training - Step 41147: {'lr': 4.0805935190945916e-05, 'samples': 21067776, 'steps': 41147, 'loss/train': 2.0430448055267334} 02/26/2022 06:51:32 - INFO - codeparrot_training - Step 41148: {'lr': 4.079697645542435e-05, 'samples': 21068288, 'steps': 41148, 'loss/train': 2.744699716567993} 02/26/2022 06:51:37 - INFO - codeparrot_training - Step 41149: {'lr': 4.0788018616062186e-05, 'samples': 21068800, 'steps': 41149, 'loss/train': 2.1895534992218018} 02/26/2022 06:51:41 - INFO - codeparrot_training - Step 41150: {'lr': 4.077906167289766e-05, 'samples': 21069312, 'steps': 41150, 'loss/train': 2.1466238498687744} 02/26/2022 06:51:46 - INFO - codeparrot_training - Step 41151: {'lr': 4.0770105625969306e-05, 'samples': 21069824, 'steps': 41151, 'loss/train': 2.0979676246643066} 02/26/2022 06:51:50 - INFO - codeparrot_training - Step 41152: {'lr': 4.076115047531526e-05, 'samples': 21070336, 'steps': 41152, 'loss/train': 1.7607640027999878} 02/26/2022 06:51:56 - INFO - codeparrot_training - Step 41153: {'lr': 4.0752196220974065e-05, 'samples': 21070848, 'steps': 41153, 'loss/train': 1.2090706825256348} 02/26/2022 06:51:59 - INFO - codeparrot_training - Step 41154: {'lr': 4.0743242862983975e-05, 'samples': 21071360, 'steps': 41154, 'loss/train': 3.4084408283233643} 02/26/2022 06:52:05 - INFO - codeparrot_training - Step 41155: {'lr': 4.073429040138349e-05, 'samples': 21071872, 'steps': 41155, 'loss/train': 1.9405903816223145} 02/26/2022 06:52:09 - INFO - codeparrot_training - Step 41156: {'lr': 4.072533883621074e-05, 'samples': 21072384, 'steps': 41156, 'loss/train': 1.894167423248291} 02/26/2022 06:52:14 - INFO - codeparrot_training - Step 41157: {'lr': 4.071638816750425e-05, 'samples': 21072896, 'steps': 41157, 'loss/train': 2.133634328842163} 02/26/2022 06:52:18 - INFO - codeparrot_training - Step 41158: {'lr': 4.070743839530222e-05, 'samples': 21073408, 'steps': 41158, 'loss/train': 1.4407835006713867} 02/26/2022 06:52:23 - INFO - codeparrot_training - Step 41159: {'lr': 4.069848951964314e-05, 'samples': 21073920, 'steps': 41159, 'loss/train': 2.2803404331207275} 02/26/2022 06:52:26 - INFO - codeparrot_training - Step 41160: {'lr': 4.068954154056528e-05, 'samples': 21074432, 'steps': 41160, 'loss/train': 2.0037155151367188} 02/26/2022 06:52:32 - INFO - codeparrot_training - Step 41161: {'lr': 4.0680594458106956e-05, 'samples': 21074944, 'steps': 41161, 'loss/train': 2.6778759956359863} 02/26/2022 06:52:35 - INFO - codeparrot_training - Step 41162: {'lr': 4.067164827230643e-05, 'samples': 21075456, 'steps': 41162, 'loss/train': 1.4494116306304932} 02/26/2022 06:52:41 - INFO - codeparrot_training - Step 41163: {'lr': 4.066270298320215e-05, 'samples': 21075968, 'steps': 41163, 'loss/train': 1.5724364519119263} 02/26/2022 06:52:44 - INFO - codeparrot_training - Step 41164: {'lr': 4.0653758590832395e-05, 'samples': 21076480, 'steps': 41164, 'loss/train': 2.2255280017852783} 02/26/2022 06:52:50 - INFO - codeparrot_training - Step 41165: {'lr': 4.064481509523543e-05, 'samples': 21076992, 'steps': 41165, 'loss/train': 1.8004752397537231} 02/26/2022 06:52:53 - INFO - codeparrot_training - Step 41166: {'lr': 4.0635872496449566e-05, 'samples': 21077504, 'steps': 41166, 'loss/train': 1.2163745164871216} 02/26/2022 06:52:59 - INFO - codeparrot_training - Step 41167: {'lr': 4.062693079451321e-05, 'samples': 21078016, 'steps': 41167, 'loss/train': 1.349444031715393} 02/26/2022 06:53:03 - INFO - codeparrot_training - Step 41168: {'lr': 4.0617989989464586e-05, 'samples': 21078528, 'steps': 41168, 'loss/train': 1.8627488613128662} 02/26/2022 06:53:08 - INFO - codeparrot_training - Step 41169: {'lr': 4.0609050081342024e-05, 'samples': 21079040, 'steps': 41169, 'loss/train': 1.115132212638855} 02/26/2022 06:53:11 - INFO - codeparrot_training - Step 41170: {'lr': 4.0600111070183715e-05, 'samples': 21079552, 'steps': 41170, 'loss/train': 1.0146816968917847} 02/26/2022 06:53:18 - INFO - codeparrot_training - Step 41171: {'lr': 4.0591172956028126e-05, 'samples': 21080064, 'steps': 41171, 'loss/train': 1.741151213645935} 02/26/2022 06:53:21 - INFO - codeparrot_training - Step 41172: {'lr': 4.058223573891343e-05, 'samples': 21080576, 'steps': 41172, 'loss/train': 0.8500745892524719} 02/26/2022 06:53:27 - INFO - codeparrot_training - Step 41173: {'lr': 4.057329941887794e-05, 'samples': 21081088, 'steps': 41173, 'loss/train': 3.3174288272857666} 02/26/2022 06:53:30 - INFO - codeparrot_training - Step 41174: {'lr': 4.056436399595989e-05, 'samples': 21081600, 'steps': 41174, 'loss/train': 1.672269582748413} 02/26/2022 06:53:36 - INFO - codeparrot_training - Step 41175: {'lr': 4.0555429470197666e-05, 'samples': 21082112, 'steps': 41175, 'loss/train': 1.6578810214996338} 02/26/2022 06:53:39 - INFO - codeparrot_training - Step 41176: {'lr': 4.0546495841629475e-05, 'samples': 21082624, 'steps': 41176, 'loss/train': 1.4162365198135376} 02/26/2022 06:53:45 - INFO - codeparrot_training - Step 41177: {'lr': 4.053756311029358e-05, 'samples': 21083136, 'steps': 41177, 'loss/train': 1.2575935125350952} 02/26/2022 06:53:50 - INFO - codeparrot_training - Step 41178: {'lr': 4.052863127622825e-05, 'samples': 21083648, 'steps': 41178, 'loss/train': 2.2112715244293213} 02/26/2022 06:53:54 - INFO - codeparrot_training - Step 41179: {'lr': 4.0519700339471694e-05, 'samples': 21084160, 'steps': 41179, 'loss/train': 2.177452564239502} 02/26/2022 06:54:00 - INFO - codeparrot_training - Step 41180: {'lr': 4.0510770300062285e-05, 'samples': 21084672, 'steps': 41180, 'loss/train': 1.827593207359314} 02/26/2022 06:54:03 - INFO - codeparrot_training - Step 41181: {'lr': 4.0501841158038206e-05, 'samples': 21085184, 'steps': 41181, 'loss/train': 1.6736212968826294} 02/26/2022 06:54:09 - INFO - codeparrot_training - Step 41182: {'lr': 4.049291291343771e-05, 'samples': 21085696, 'steps': 41182, 'loss/train': 2.174193859100342} 02/26/2022 06:54:12 - INFO - codeparrot_training - Step 41183: {'lr': 4.048398556629901e-05, 'samples': 21086208, 'steps': 41183, 'loss/train': 1.22590172290802} 02/26/2022 06:54:18 - INFO - codeparrot_training - Step 41184: {'lr': 4.047505911666044e-05, 'samples': 21086720, 'steps': 41184, 'loss/train': 1.1328811645507812} 02/26/2022 06:54:21 - INFO - codeparrot_training - Step 41185: {'lr': 4.046613356456011e-05, 'samples': 21087232, 'steps': 41185, 'loss/train': 2.1060800552368164} 02/26/2022 06:54:27 - INFO - codeparrot_training - Step 41186: {'lr': 4.045720891003649e-05, 'samples': 21087744, 'steps': 41186, 'loss/train': 1.4242898225784302} 02/26/2022 06:54:30 - INFO - codeparrot_training - Step 41187: {'lr': 4.0448285153127486e-05, 'samples': 21088256, 'steps': 41187, 'loss/train': 3.2226996421813965} 02/26/2022 06:54:36 - INFO - codeparrot_training - Step 41188: {'lr': 4.0439362293871576e-05, 'samples': 21088768, 'steps': 41188, 'loss/train': 0.7156848311424255} 02/26/2022 06:54:39 - INFO - codeparrot_training - Step 41189: {'lr': 4.043044033230681e-05, 'samples': 21089280, 'steps': 41189, 'loss/train': 0.9726080298423767} 02/26/2022 06:54:45 - INFO - codeparrot_training - Step 41190: {'lr': 4.0421519268471624e-05, 'samples': 21089792, 'steps': 41190, 'loss/train': 1.661525845527649} 02/26/2022 06:54:48 - INFO - codeparrot_training - Step 41191: {'lr': 4.0412599102404e-05, 'samples': 21090304, 'steps': 41191, 'loss/train': 1.864445686340332} 02/26/2022 06:54:54 - INFO - codeparrot_training - Step 41192: {'lr': 4.040367983414228e-05, 'samples': 21090816, 'steps': 41192, 'loss/train': 2.1463263034820557} 02/26/2022 06:54:58 - INFO - codeparrot_training - Step 41193: {'lr': 4.039476146372459e-05, 'samples': 21091328, 'steps': 41193, 'loss/train': 1.0734728574752808} 02/26/2022 06:55:03 - INFO - codeparrot_training - Step 41194: {'lr': 4.038584399118925e-05, 'samples': 21091840, 'steps': 41194, 'loss/train': 1.345886468887329} 02/26/2022 06:55:07 - INFO - codeparrot_training - Step 41195: {'lr': 4.037692741657439e-05, 'samples': 21092352, 'steps': 41195, 'loss/train': 0.40856385231018066} 02/26/2022 06:55:12 - INFO - codeparrot_training - Step 41196: {'lr': 4.036801173991822e-05, 'samples': 21092864, 'steps': 41196, 'loss/train': 2.2371835708618164} 02/26/2022 06:55:16 - INFO - codeparrot_training - Step 41197: {'lr': 4.0359096961258864e-05, 'samples': 21093376, 'steps': 41197, 'loss/train': 1.3627609014511108} 02/26/2022 06:55:21 - INFO - codeparrot_training - Step 41198: {'lr': 4.0350183080634595e-05, 'samples': 21093888, 'steps': 41198, 'loss/train': 1.2348918914794922} 02/26/2022 06:55:25 - INFO - codeparrot_training - Step 41199: {'lr': 4.034127009808361e-05, 'samples': 21094400, 'steps': 41199, 'loss/train': 1.4219896793365479} 02/26/2022 06:55:30 - INFO - codeparrot_training - Step 41200: {'lr': 4.033235801364402e-05, 'samples': 21094912, 'steps': 41200, 'loss/train': 1.7162977457046509} 02/26/2022 06:55:34 - INFO - codeparrot_training - Step 41201: {'lr': 4.032344682735401e-05, 'samples': 21095424, 'steps': 41201, 'loss/train': 2.4636213779449463} 02/26/2022 06:55:39 - INFO - codeparrot_training - Step 41202: {'lr': 4.0314536539251796e-05, 'samples': 21095936, 'steps': 41202, 'loss/train': 1.429274559020996} 02/26/2022 06:55:43 - INFO - codeparrot_training - Step 41203: {'lr': 4.0305627149375524e-05, 'samples': 21096448, 'steps': 41203, 'loss/train': 1.6399781703948975} 02/26/2022 06:55:48 - INFO - codeparrot_training - Step 41204: {'lr': 4.029671865776338e-05, 'samples': 21096960, 'steps': 41204, 'loss/train': 1.4864203929901123} 02/26/2022 06:55:52 - INFO - codeparrot_training - Step 41205: {'lr': 4.028781106445345e-05, 'samples': 21097472, 'steps': 41205, 'loss/train': 1.7550987005233765} 02/26/2022 06:55:58 - INFO - codeparrot_training - Step 41206: {'lr': 4.0278904369483973e-05, 'samples': 21097984, 'steps': 41206, 'loss/train': 0.5205437541007996} 02/26/2022 06:56:02 - INFO - codeparrot_training - Step 41207: {'lr': 4.026999857289304e-05, 'samples': 21098496, 'steps': 41207, 'loss/train': 0.7449386715888977} 02/26/2022 06:56:07 - INFO - codeparrot_training - Step 41208: {'lr': 4.0261093674718975e-05, 'samples': 21099008, 'steps': 41208, 'loss/train': 0.026629606261849403} 02/26/2022 06:56:11 - INFO - codeparrot_training - Step 41209: {'lr': 4.025218967499963e-05, 'samples': 21099520, 'steps': 41209, 'loss/train': 1.4720178842544556} 02/26/2022 06:56:16 - INFO - codeparrot_training - Step 41210: {'lr': 4.024328657377338e-05, 'samples': 21100032, 'steps': 41210, 'loss/train': 2.110401153564453} 02/26/2022 06:56:20 - INFO - codeparrot_training - Step 41211: {'lr': 4.023438437107829e-05, 'samples': 21100544, 'steps': 41211, 'loss/train': 1.2986823320388794} 02/26/2022 06:56:25 - INFO - codeparrot_training - Step 41212: {'lr': 4.0225483066952476e-05, 'samples': 21101056, 'steps': 41212, 'loss/train': 1.8058794736862183} 02/26/2022 06:56:29 - INFO - codeparrot_training - Step 41213: {'lr': 4.021658266143402e-05, 'samples': 21101568, 'steps': 41213, 'loss/train': 2.134857416152954} 02/26/2022 06:56:34 - INFO - codeparrot_training - Step 41214: {'lr': 4.020768315456119e-05, 'samples': 21102080, 'steps': 41214, 'loss/train': 2.508793592453003} 02/26/2022 06:56:38 - INFO - codeparrot_training - Step 41215: {'lr': 4.019878454637202e-05, 'samples': 21102592, 'steps': 41215, 'loss/train': 1.2010328769683838} 02/26/2022 06:56:44 - INFO - codeparrot_training - Step 41216: {'lr': 4.018988683690461e-05, 'samples': 21103104, 'steps': 41216, 'loss/train': 1.67734956741333} 02/26/2022 06:56:47 - INFO - codeparrot_training - Step 41217: {'lr': 4.0180990026197075e-05, 'samples': 21103616, 'steps': 41217, 'loss/train': 1.0418322086334229} 02/26/2022 06:56:53 - INFO - codeparrot_training - Step 41218: {'lr': 4.01720941142876e-05, 'samples': 21104128, 'steps': 41218, 'loss/train': 2.3684756755828857} 02/26/2022 06:56:56 - INFO - codeparrot_training - Step 41219: {'lr': 4.016319910121424e-05, 'samples': 21104640, 'steps': 41219, 'loss/train': 1.4005308151245117} 02/26/2022 06:57:02 - INFO - codeparrot_training - Step 41220: {'lr': 4.0154304987015056e-05, 'samples': 21105152, 'steps': 41220, 'loss/train': 2.0739009380340576} 02/26/2022 06:57:05 - INFO - codeparrot_training - Step 41221: {'lr': 4.0145411771728256e-05, 'samples': 21105664, 'steps': 41221, 'loss/train': 2.0402870178222656} 02/26/2022 06:57:11 - INFO - codeparrot_training - Step 41222: {'lr': 4.013651945539187e-05, 'samples': 21106176, 'steps': 41222, 'loss/train': 1.9704837799072266} 02/26/2022 06:57:15 - INFO - codeparrot_training - Step 41223: {'lr': 4.012762803804401e-05, 'samples': 21106688, 'steps': 41223, 'loss/train': 1.009820818901062} 02/26/2022 06:57:20 - INFO - codeparrot_training - Step 41224: {'lr': 4.011873751972267e-05, 'samples': 21107200, 'steps': 41224, 'loss/train': 1.7188966274261475} 02/26/2022 06:57:24 - INFO - codeparrot_training - Step 41225: {'lr': 4.010984790046615e-05, 'samples': 21107712, 'steps': 41225, 'loss/train': 2.1077184677124023} 02/26/2022 06:57:30 - INFO - codeparrot_training - Step 41226: {'lr': 4.010095918031228e-05, 'samples': 21108224, 'steps': 41226, 'loss/train': 2.5008511543273926} 02/26/2022 06:57:33 - INFO - codeparrot_training - Step 41227: {'lr': 4.009207135929929e-05, 'samples': 21108736, 'steps': 41227, 'loss/train': 1.7562178373336792} 02/26/2022 06:57:39 - INFO - codeparrot_training - Step 41228: {'lr': 4.008318443746517e-05, 'samples': 21109248, 'steps': 41228, 'loss/train': 1.1937103271484375} 02/26/2022 06:57:42 - INFO - codeparrot_training - Step 41229: {'lr': 4.007429841484814e-05, 'samples': 21109760, 'steps': 41229, 'loss/train': 0.6701145172119141} 02/26/2022 06:57:48 - INFO - codeparrot_training - Step 41230: {'lr': 4.0065413291486026e-05, 'samples': 21110272, 'steps': 41230, 'loss/train': 1.6751470565795898} 02/26/2022 06:57:52 - INFO - codeparrot_training - Step 41231: {'lr': 4.0056529067417105e-05, 'samples': 21110784, 'steps': 41231, 'loss/train': 1.5667650699615479} 02/26/2022 06:57:55 - INFO - codeparrot_training - Step 41232: {'lr': 4.004764574267927e-05, 'samples': 21111296, 'steps': 41232, 'loss/train': 1.8498036861419678} 02/26/2022 06:58:01 - INFO - codeparrot_training - Step 41233: {'lr': 4.003876331731071e-05, 'samples': 21111808, 'steps': 41233, 'loss/train': 2.22662353515625} 02/26/2022 06:58:04 - INFO - codeparrot_training - Step 41234: {'lr': 4.0029881791349424e-05, 'samples': 21112320, 'steps': 41234, 'loss/train': 1.0748449563980103} 02/26/2022 06:58:10 - INFO - codeparrot_training - Step 41235: {'lr': 4.0021001164833456e-05, 'samples': 21112832, 'steps': 41235, 'loss/train': 0.7685216069221497} 02/26/2022 06:58:13 - INFO - codeparrot_training - Step 41236: {'lr': 4.001212143780078e-05, 'samples': 21113344, 'steps': 41236, 'loss/train': 1.6989455223083496} 02/26/2022 06:58:19 - INFO - codeparrot_training - Step 41237: {'lr': 4.000324261028956e-05, 'samples': 21113856, 'steps': 41237, 'loss/train': 0.07501186430454254} 02/26/2022 06:58:22 - INFO - codeparrot_training - Step 41238: {'lr': 3.999436468233778e-05, 'samples': 21114368, 'steps': 41238, 'loss/train': 2.3117334842681885} 02/26/2022 06:58:29 - INFO - codeparrot_training - Step 41239: {'lr': 3.9985487653983435e-05, 'samples': 21114880, 'steps': 41239, 'loss/train': 1.8053890466690063} 02/26/2022 06:58:32 - INFO - codeparrot_training - Step 41240: {'lr': 3.997661152526452e-05, 'samples': 21115392, 'steps': 41240, 'loss/train': 1.7694995403289795} 02/26/2022 06:58:38 - INFO - codeparrot_training - Step 41241: {'lr': 3.9967736296219164e-05, 'samples': 21115904, 'steps': 41241, 'loss/train': 0.1313379853963852} 02/26/2022 06:58:41 - INFO - codeparrot_training - Step 41242: {'lr': 3.9958861966885305e-05, 'samples': 21116416, 'steps': 41242, 'loss/train': 2.506037473678589} 02/26/2022 06:58:47 - INFO - codeparrot_training - Step 41243: {'lr': 3.994998853730108e-05, 'samples': 21116928, 'steps': 41243, 'loss/train': 1.5256882905960083} 02/26/2022 06:58:50 - INFO - codeparrot_training - Step 41244: {'lr': 3.994111600750427e-05, 'samples': 21117440, 'steps': 41244, 'loss/train': 2.7465505599975586} 02/26/2022 06:58:56 - INFO - codeparrot_training - Step 41245: {'lr': 3.993224437753309e-05, 'samples': 21117952, 'steps': 41245, 'loss/train': 1.1976745128631592} 02/26/2022 06:58:59 - INFO - codeparrot_training - Step 41246: {'lr': 3.992337364742543e-05, 'samples': 21118464, 'steps': 41246, 'loss/train': 0.7530762553215027} 02/26/2022 06:59:05 - INFO - codeparrot_training - Step 41247: {'lr': 3.991450381721945e-05, 'samples': 21118976, 'steps': 41247, 'loss/train': 1.2591278553009033} 02/26/2022 06:59:08 - INFO - codeparrot_training - Step 41248: {'lr': 3.99056348869529e-05, 'samples': 21119488, 'steps': 41248, 'loss/train': 1.8351786136627197} 02/26/2022 06:59:14 - INFO - codeparrot_training - Step 41249: {'lr': 3.9896766856663936e-05, 'samples': 21120000, 'steps': 41249, 'loss/train': 0.7252730131149292} 02/26/2022 06:59:17 - INFO - codeparrot_training - Step 41250: {'lr': 3.988789972639045e-05, 'samples': 21120512, 'steps': 41250, 'loss/train': 1.743416428565979} 02/26/2022 06:59:23 - INFO - codeparrot_training - Step 41251: {'lr': 3.987903349617064e-05, 'samples': 21121024, 'steps': 41251, 'loss/train': 2.4336726665496826} 02/26/2022 06:59:26 - INFO - codeparrot_training - Step 41252: {'lr': 3.987016816604219e-05, 'samples': 21121536, 'steps': 41252, 'loss/train': 1.9759290218353271} 02/26/2022 06:59:32 - INFO - codeparrot_training - Step 41253: {'lr': 3.986130373604327e-05, 'samples': 21122048, 'steps': 41253, 'loss/train': 1.8637784719467163} 02/26/2022 06:59:35 - INFO - codeparrot_training - Step 41254: {'lr': 3.985244020621173e-05, 'samples': 21122560, 'steps': 41254, 'loss/train': 1.6633825302124023} 02/26/2022 06:59:41 - INFO - codeparrot_training - Step 41255: {'lr': 3.984357757658569e-05, 'samples': 21123072, 'steps': 41255, 'loss/train': 1.6736080646514893} 02/26/2022 06:59:45 - INFO - codeparrot_training - Step 41256: {'lr': 3.9834715847203e-05, 'samples': 21123584, 'steps': 41256, 'loss/train': 2.252772569656372} 02/26/2022 06:59:50 - INFO - codeparrot_training - Step 41257: {'lr': 3.982585501810168e-05, 'samples': 21124096, 'steps': 41257, 'loss/train': 2.448148012161255} 02/26/2022 06:59:54 - INFO - codeparrot_training - Step 41258: {'lr': 3.981699508931966e-05, 'samples': 21124608, 'steps': 41258, 'loss/train': 1.6163902282714844} 02/26/2022 06:59:59 - INFO - codeparrot_training - Step 41259: {'lr': 3.980813606089481e-05, 'samples': 21125120, 'steps': 41259, 'loss/train': 2.9076037406921387} 02/26/2022 07:00:03 - INFO - codeparrot_training - Step 41260: {'lr': 3.979927793286522e-05, 'samples': 21125632, 'steps': 41260, 'loss/train': 2.1136412620544434} 02/26/2022 07:00:08 - INFO - codeparrot_training - Step 41261: {'lr': 3.979042070526881e-05, 'samples': 21126144, 'steps': 41261, 'loss/train': 2.231005907058716} 02/26/2022 07:00:12 - INFO - codeparrot_training - Step 41262: {'lr': 3.978156437814345e-05, 'samples': 21126656, 'steps': 41262, 'loss/train': 1.4403915405273438} 02/26/2022 07:00:17 - INFO - codeparrot_training - Step 41263: {'lr': 3.97727089515271e-05, 'samples': 21127168, 'steps': 41263, 'loss/train': 1.971002459526062} 02/26/2022 07:00:21 - INFO - codeparrot_training - Step 41264: {'lr': 3.976385442545774e-05, 'samples': 21127680, 'steps': 41264, 'loss/train': 1.0498101711273193} 02/26/2022 07:00:27 - INFO - codeparrot_training - Step 41265: {'lr': 3.975500079997327e-05, 'samples': 21128192, 'steps': 41265, 'loss/train': 2.522651195526123} 02/26/2022 07:00:30 - INFO - codeparrot_training - Step 41266: {'lr': 3.974614807511162e-05, 'samples': 21128704, 'steps': 41266, 'loss/train': 1.9527454376220703} 02/26/2022 07:00:36 - INFO - codeparrot_training - Step 41267: {'lr': 3.973729625091066e-05, 'samples': 21129216, 'steps': 41267, 'loss/train': 1.787041187286377} 02/26/2022 07:00:39 - INFO - codeparrot_training - Step 41268: {'lr': 3.972844532740841e-05, 'samples': 21129728, 'steps': 41268, 'loss/train': 1.775451898574829} 02/26/2022 07:00:45 - INFO - codeparrot_training - Step 41269: {'lr': 3.971959530464272e-05, 'samples': 21130240, 'steps': 41269, 'loss/train': 1.4558331966400146} 02/26/2022 07:00:49 - INFO - codeparrot_training - Step 41270: {'lr': 3.971074618265153e-05, 'samples': 21130752, 'steps': 41270, 'loss/train': 1.5109256505966187} 02/26/2022 07:00:55 - INFO - codeparrot_training - Step 41271: {'lr': 3.970189796147267e-05, 'samples': 21131264, 'steps': 41271, 'loss/train': 2.0873231887817383} 02/26/2022 07:00:58 - INFO - codeparrot_training - Step 41272: {'lr': 3.9693050641144145e-05, 'samples': 21131776, 'steps': 41272, 'loss/train': 1.465606451034546} 02/26/2022 07:01:02 - INFO - codeparrot_training - Step 41273: {'lr': 3.9684204221703826e-05, 'samples': 21132288, 'steps': 41273, 'loss/train': 2.0423688888549805} 02/26/2022 07:01:07 - INFO - codeparrot_training - Step 41274: {'lr': 3.967535870318958e-05, 'samples': 21132800, 'steps': 41274, 'loss/train': 2.768829584121704} 02/26/2022 07:01:11 - INFO - codeparrot_training - Step 41275: {'lr': 3.966651408563926e-05, 'samples': 21133312, 'steps': 41275, 'loss/train': 1.856337308883667} 02/26/2022 07:01:17 - INFO - codeparrot_training - Step 41276: {'lr': 3.965767036909085e-05, 'samples': 21133824, 'steps': 41276, 'loss/train': 1.9268488883972168} 02/26/2022 07:01:20 - INFO - codeparrot_training - Step 41277: {'lr': 3.9648827553582204e-05, 'samples': 21134336, 'steps': 41277, 'loss/train': 1.5484291315078735} 02/26/2022 07:01:26 - INFO - codeparrot_training - Step 41278: {'lr': 3.963998563915119e-05, 'samples': 21134848, 'steps': 41278, 'loss/train': 2.0642435550689697} 02/26/2022 07:01:29 - INFO - codeparrot_training - Step 41279: {'lr': 3.9631144625835614e-05, 'samples': 21135360, 'steps': 41279, 'loss/train': 2.2309176921844482} 02/26/2022 07:01:35 - INFO - codeparrot_training - Step 41280: {'lr': 3.962230451367349e-05, 'samples': 21135872, 'steps': 41280, 'loss/train': 1.3947393894195557} 02/26/2022 07:01:38 - INFO - codeparrot_training - Step 41281: {'lr': 3.961346530270252e-05, 'samples': 21136384, 'steps': 41281, 'loss/train': 2.2639381885528564} 02/26/2022 07:01:44 - INFO - codeparrot_training - Step 41282: {'lr': 3.96046269929608e-05, 'samples': 21136896, 'steps': 41282, 'loss/train': 1.8505698442459106} 02/26/2022 07:01:47 - INFO - codeparrot_training - Step 41283: {'lr': 3.9595789584485925e-05, 'samples': 21137408, 'steps': 41283, 'loss/train': 2.3372275829315186} 02/26/2022 07:01:55 - INFO - codeparrot_training - Step 41284: {'lr': 3.958695307731594e-05, 'samples': 21137920, 'steps': 41284, 'loss/train': 1.7233986854553223} 02/26/2022 07:01:58 - INFO - codeparrot_training - Step 41285: {'lr': 3.957811747148857e-05, 'samples': 21138432, 'steps': 41285, 'loss/train': 1.3501551151275635} 02/26/2022 07:02:04 - INFO - codeparrot_training - Step 41286: {'lr': 3.956928276704186e-05, 'samples': 21138944, 'steps': 41286, 'loss/train': 1.230549693107605} 02/26/2022 07:02:07 - INFO - codeparrot_training - Step 41287: {'lr': 3.9560448964013394e-05, 'samples': 21139456, 'steps': 41287, 'loss/train': 0.4091601073741913} 02/26/2022 07:02:13 - INFO - codeparrot_training - Step 41288: {'lr': 3.95516160624412e-05, 'samples': 21139968, 'steps': 41288, 'loss/train': 0.35843050479888916} 02/26/2022 07:02:16 - INFO - codeparrot_training - Step 41289: {'lr': 3.9542784062363004e-05, 'samples': 21140480, 'steps': 41289, 'loss/train': 0.051794614642858505} 02/26/2022 07:02:22 - INFO - codeparrot_training - Step 41290: {'lr': 3.9533952963816815e-05, 'samples': 21140992, 'steps': 41290, 'loss/train': 1.6985957622528076} 02/26/2022 07:02:25 - INFO - codeparrot_training - Step 41291: {'lr': 3.95251227668402e-05, 'samples': 21141504, 'steps': 41291, 'loss/train': 0.6630358099937439} 02/26/2022 07:02:31 - INFO - codeparrot_training - Step 41292: {'lr': 3.9516293471471204e-05, 'samples': 21142016, 'steps': 41292, 'loss/train': 0.9668679237365723} 02/26/2022 07:02:34 - INFO - codeparrot_training - Step 41293: {'lr': 3.9507465077747526e-05, 'samples': 21142528, 'steps': 41293, 'loss/train': 1.6144204139709473} 02/26/2022 07:02:42 - INFO - codeparrot_training - Step 41294: {'lr': 3.949863758570707e-05, 'samples': 21143040, 'steps': 41294, 'loss/train': 1.9905365705490112} 02/26/2022 07:02:45 - INFO - codeparrot_training - Step 41295: {'lr': 3.948981099538759e-05, 'samples': 21143552, 'steps': 41295, 'loss/train': 1.3369977474212646} 02/26/2022 07:02:51 - INFO - codeparrot_training - Step 41296: {'lr': 3.948098530682695e-05, 'samples': 21144064, 'steps': 41296, 'loss/train': 1.4646605253219604} 02/26/2022 07:02:54 - INFO - codeparrot_training - Step 41297: {'lr': 3.9472160520062836e-05, 'samples': 21144576, 'steps': 41297, 'loss/train': 1.7085332870483398} 02/26/2022 07:03:00 - INFO - codeparrot_training - Step 41298: {'lr': 3.946333663513321e-05, 'samples': 21145088, 'steps': 41298, 'loss/train': 1.3035117387771606} 02/26/2022 07:03:05 - INFO - codeparrot_training - Step 41299: {'lr': 3.945451365207581e-05, 'samples': 21145600, 'steps': 41299, 'loss/train': 2.3040616512298584} 02/26/2022 07:03:09 - INFO - codeparrot_training - Step 41300: {'lr': 3.944569157092839e-05, 'samples': 21146112, 'steps': 41300, 'loss/train': 0.05751051381230354} 02/26/2022 07:03:14 - INFO - codeparrot_training - Step 41301: {'lr': 3.9436870391728716e-05, 'samples': 21146624, 'steps': 41301, 'loss/train': 1.4843560457229614} 02/26/2022 07:03:18 - INFO - codeparrot_training - Step 41302: {'lr': 3.942805011451469e-05, 'samples': 21147136, 'steps': 41302, 'loss/train': 2.106691360473633} 02/26/2022 07:03:25 - INFO - codeparrot_training - Step 41303: {'lr': 3.941923073932405e-05, 'samples': 21147648, 'steps': 41303, 'loss/train': 1.0581002235412598} 02/26/2022 07:03:29 - INFO - codeparrot_training - Step 41304: {'lr': 3.941041226619455e-05, 'samples': 21148160, 'steps': 41304, 'loss/train': 1.5383330583572388} 02/26/2022 07:03:34 - INFO - codeparrot_training - Step 41305: {'lr': 3.9401594695163886e-05, 'samples': 21148672, 'steps': 41305, 'loss/train': 1.645405650138855} 02/26/2022 07:03:38 - INFO - codeparrot_training - Step 41306: {'lr': 3.939277802627e-05, 'samples': 21149184, 'steps': 41306, 'loss/train': 1.3650903701782227} 02/26/2022 07:03:43 - INFO - codeparrot_training - Step 41307: {'lr': 3.93839622595506e-05, 'samples': 21149696, 'steps': 41307, 'loss/train': 2.1440048217773438} 02/26/2022 07:03:47 - INFO - codeparrot_training - Step 41308: {'lr': 3.9375147395043384e-05, 'samples': 21150208, 'steps': 41308, 'loss/train': 0.9623491168022156} 02/26/2022 07:03:50 - INFO - codeparrot_training - Step 41309: {'lr': 3.93663334327862e-05, 'samples': 21150720, 'steps': 41309, 'loss/train': 1.1603679656982422} 02/26/2022 07:03:56 - INFO - codeparrot_training - Step 41310: {'lr': 3.935752037281667e-05, 'samples': 21151232, 'steps': 41310, 'loss/train': 0.5403485894203186} 02/26/2022 07:03:59 - INFO - codeparrot_training - Step 41311: {'lr': 3.9348708215172694e-05, 'samples': 21151744, 'steps': 41311, 'loss/train': 1.7832581996917725} 02/26/2022 07:04:05 - INFO - codeparrot_training - Step 41312: {'lr': 3.9339896959891985e-05, 'samples': 21152256, 'steps': 41312, 'loss/train': 2.46216082572937} 02/26/2022 07:04:12 - INFO - codeparrot_training - Step 41313: {'lr': 3.933108660701223e-05, 'samples': 21152768, 'steps': 41313, 'loss/train': 1.0073262453079224} 02/26/2022 07:04:16 - INFO - codeparrot_training - Step 41314: {'lr': 3.932227715657119e-05, 'samples': 21153280, 'steps': 41314, 'loss/train': 1.5049607753753662} 02/26/2022 07:04:21 - INFO - codeparrot_training - Step 41315: {'lr': 3.931346860860666e-05, 'samples': 21153792, 'steps': 41315, 'loss/train': 2.0568690299987793} 02/26/2022 07:04:25 - INFO - codeparrot_training - Step 41316: {'lr': 3.930466096315624e-05, 'samples': 21154304, 'steps': 41316, 'loss/train': 3.55737042427063} 02/26/2022 07:04:30 - INFO - codeparrot_training - Step 41317: {'lr': 3.9295854220257886e-05, 'samples': 21154816, 'steps': 41317, 'loss/train': 2.4277520179748535} 02/26/2022 07:04:34 - INFO - codeparrot_training - Step 41318: {'lr': 3.928704837994909e-05, 'samples': 21155328, 'steps': 41318, 'loss/train': 1.5596942901611328} 02/26/2022 07:04:39 - INFO - codeparrot_training - Step 41319: {'lr': 3.927824344226771e-05, 'samples': 21155840, 'steps': 41319, 'loss/train': 2.4670939445495605} 02/26/2022 07:04:43 - INFO - codeparrot_training - Step 41320: {'lr': 3.926943940725136e-05, 'samples': 21156352, 'steps': 41320, 'loss/train': 1.4914158582687378} 02/26/2022 07:04:48 - INFO - codeparrot_training - Step 41321: {'lr': 3.926063627493795e-05, 'samples': 21156864, 'steps': 41321, 'loss/train': 1.478821039199829} 02/26/2022 07:04:52 - INFO - codeparrot_training - Step 41322: {'lr': 3.925183404536492e-05, 'samples': 21157376, 'steps': 41322, 'loss/train': 1.0644373893737793} 02/26/2022 07:04:57 - INFO - codeparrot_training - Step 41323: {'lr': 3.924303271857019e-05, 'samples': 21157888, 'steps': 41323, 'loss/train': 1.0851390361785889} 02/26/2022 07:05:01 - INFO - codeparrot_training - Step 41324: {'lr': 3.923423229459133e-05, 'samples': 21158400, 'steps': 41324, 'loss/train': 1.8833128213882446} 02/26/2022 07:05:06 - INFO - codeparrot_training - Step 41325: {'lr': 3.9225432773466184e-05, 'samples': 21158912, 'steps': 41325, 'loss/train': 1.6743628978729248} 02/26/2022 07:05:10 - INFO - codeparrot_training - Step 41326: {'lr': 3.921663415523227e-05, 'samples': 21159424, 'steps': 41326, 'loss/train': 1.2833172082901} 02/26/2022 07:05:16 - INFO - codeparrot_training - Step 41327: {'lr': 3.92078364399274e-05, 'samples': 21159936, 'steps': 41327, 'loss/train': 1.1480040550231934} 02/26/2022 07:05:19 - INFO - codeparrot_training - Step 41328: {'lr': 3.919903962758917e-05, 'samples': 21160448, 'steps': 41328, 'loss/train': 1.569909930229187} 02/26/2022 07:05:22 - INFO - codeparrot_training - Step 41329: {'lr': 3.9190243718255385e-05, 'samples': 21160960, 'steps': 41329, 'loss/train': 0.3647741377353668} 02/26/2022 07:05:30 - INFO - codeparrot_training - Step 41330: {'lr': 3.9181448711963666e-05, 'samples': 21161472, 'steps': 41330, 'loss/train': 1.9107706546783447} 02/26/2022 07:05:33 - INFO - codeparrot_training - Step 41331: {'lr': 3.9172654608751635e-05, 'samples': 21161984, 'steps': 41331, 'loss/train': 2.3226635456085205} 02/26/2022 07:05:39 - INFO - codeparrot_training - Step 41332: {'lr': 3.9163861408656994e-05, 'samples': 21162496, 'steps': 41332, 'loss/train': 2.380817174911499} 02/26/2022 07:05:42 - INFO - codeparrot_training - Step 41333: {'lr': 3.9155069111717455e-05, 'samples': 21163008, 'steps': 41333, 'loss/train': 1.2203632593154907} 02/26/2022 07:05:48 - INFO - codeparrot_training - Step 41334: {'lr': 3.9146277717970664e-05, 'samples': 21163520, 'steps': 41334, 'loss/train': 0.7109491229057312} 02/26/2022 07:05:53 - INFO - codeparrot_training - Step 41335: {'lr': 3.913748722745425e-05, 'samples': 21164032, 'steps': 41335, 'loss/train': 3.091419219970703} 02/26/2022 07:05:57 - INFO - codeparrot_training - Step 41336: {'lr': 3.912869764020583e-05, 'samples': 21164544, 'steps': 41336, 'loss/train': 1.5819402933120728} 02/26/2022 07:06:02 - INFO - codeparrot_training - Step 41337: {'lr': 3.911990895626319e-05, 'samples': 21165056, 'steps': 41337, 'loss/train': 1.4858388900756836} 02/26/2022 07:06:06 - INFO - codeparrot_training - Step 41338: {'lr': 3.911112117566387e-05, 'samples': 21165568, 'steps': 41338, 'loss/train': 1.230688214302063} 02/26/2022 07:06:13 - INFO - codeparrot_training - Step 41339: {'lr': 3.910233429844556e-05, 'samples': 21166080, 'steps': 41339, 'loss/train': 1.9165751934051514} 02/26/2022 07:06:16 - INFO - codeparrot_training - Step 41340: {'lr': 3.909354832464582e-05, 'samples': 21166592, 'steps': 41340, 'loss/train': 2.2020111083984375} 02/26/2022 07:06:22 - INFO - codeparrot_training - Step 41341: {'lr': 3.9084763254302406e-05, 'samples': 21167104, 'steps': 41341, 'loss/train': 2.00919771194458} 02/26/2022 07:06:25 - INFO - codeparrot_training - Step 41342: {'lr': 3.907597908745289e-05, 'samples': 21167616, 'steps': 41342, 'loss/train': 2.7273736000061035} 02/26/2022 07:06:31 - INFO - codeparrot_training - Step 41343: {'lr': 3.90671958241349e-05, 'samples': 21168128, 'steps': 41343, 'loss/train': 1.630610466003418} 02/26/2022 07:06:34 - INFO - codeparrot_training - Step 41344: {'lr': 3.905841346438602e-05, 'samples': 21168640, 'steps': 41344, 'loss/train': 1.4005275964736938} 02/26/2022 07:06:40 - INFO - codeparrot_training - Step 41345: {'lr': 3.904963200824396e-05, 'samples': 21169152, 'steps': 41345, 'loss/train': 1.5225865840911865} 02/26/2022 07:06:43 - INFO - codeparrot_training - Step 41346: {'lr': 3.904085145574629e-05, 'samples': 21169664, 'steps': 41346, 'loss/train': 1.817198634147644} 02/26/2022 07:06:49 - INFO - codeparrot_training - Step 41347: {'lr': 3.9032071806930636e-05, 'samples': 21170176, 'steps': 41347, 'loss/train': 2.483788251876831} 02/26/2022 07:06:52 - INFO - codeparrot_training - Step 41348: {'lr': 3.902329306183453e-05, 'samples': 21170688, 'steps': 41348, 'loss/train': 1.8209511041641235} 02/26/2022 07:06:59 - INFO - codeparrot_training - Step 41349: {'lr': 3.9014515220495705e-05, 'samples': 21171200, 'steps': 41349, 'loss/train': 2.0334019660949707} 02/26/2022 07:07:03 - INFO - codeparrot_training - Step 41350: {'lr': 3.9005738282951704e-05, 'samples': 21171712, 'steps': 41350, 'loss/train': 2.1247360706329346} 02/26/2022 07:07:08 - INFO - codeparrot_training - Step 41351: {'lr': 3.899696224924007e-05, 'samples': 21172224, 'steps': 41351, 'loss/train': 1.9253935813903809} 02/26/2022 07:07:12 - INFO - codeparrot_training - Step 41352: {'lr': 3.89881871193985e-05, 'samples': 21172736, 'steps': 41352, 'loss/train': 1.7527488470077515} 02/26/2022 07:07:17 - INFO - codeparrot_training - Step 41353: {'lr': 3.8979412893464516e-05, 'samples': 21173248, 'steps': 41353, 'loss/train': 1.6387213468551636} 02/26/2022 07:07:21 - INFO - codeparrot_training - Step 41354: {'lr': 3.897063957147573e-05, 'samples': 21173760, 'steps': 41354, 'loss/train': 1.3716071844100952} 02/26/2022 07:07:26 - INFO - codeparrot_training - Step 41355: {'lr': 3.8961867153469665e-05, 'samples': 21174272, 'steps': 41355, 'loss/train': 2.152237892150879} 02/26/2022 07:07:30 - INFO - codeparrot_training - Step 41356: {'lr': 3.895309563948407e-05, 'samples': 21174784, 'steps': 41356, 'loss/train': 0.7340816259384155} 02/26/2022 07:07:35 - INFO - codeparrot_training - Step 41357: {'lr': 3.8944325029556274e-05, 'samples': 21175296, 'steps': 41357, 'loss/train': 0.9668439030647278} 02/26/2022 07:07:39 - INFO - codeparrot_training - Step 41358: {'lr': 3.893555532372403e-05, 'samples': 21175808, 'steps': 41358, 'loss/train': 0.7590364813804626} 02/26/2022 07:07:44 - INFO - codeparrot_training - Step 41359: {'lr': 3.89267865220248e-05, 'samples': 21176320, 'steps': 41359, 'loss/train': 1.936675786972046} 02/26/2022 07:07:48 - INFO - codeparrot_training - Step 41360: {'lr': 3.891801862449629e-05, 'samples': 21176832, 'steps': 41360, 'loss/train': 1.2729203701019287} 02/26/2022 07:07:55 - INFO - codeparrot_training - Step 41361: {'lr': 3.890925163117587e-05, 'samples': 21177344, 'steps': 41361, 'loss/train': 1.1969263553619385} 02/26/2022 07:07:59 - INFO - codeparrot_training - Step 41362: {'lr': 3.890048554210121e-05, 'samples': 21177856, 'steps': 41362, 'loss/train': 2.4606263637542725} 02/26/2022 07:08:02 - INFO - codeparrot_training - Step 41363: {'lr': 3.889172035730981e-05, 'samples': 21178368, 'steps': 41363, 'loss/train': 1.202028512954712} 02/26/2022 07:08:08 - INFO - codeparrot_training - Step 41364: {'lr': 3.8882956076839285e-05, 'samples': 21178880, 'steps': 41364, 'loss/train': 1.791395664215088} 02/26/2022 07:08:11 - INFO - codeparrot_training - Step 41365: {'lr': 3.887419270072715e-05, 'samples': 21179392, 'steps': 41365, 'loss/train': 1.1219316720962524} 02/26/2022 07:08:17 - INFO - codeparrot_training - Step 41366: {'lr': 3.886543022901093e-05, 'samples': 21179904, 'steps': 41366, 'loss/train': 1.7659783363342285} 02/26/2022 07:08:20 - INFO - codeparrot_training - Step 41367: {'lr': 3.88566686617281e-05, 'samples': 21180416, 'steps': 41367, 'loss/train': 2.1345179080963135} 02/26/2022 07:08:26 - INFO - codeparrot_training - Step 41368: {'lr': 3.8847907998916326e-05, 'samples': 21180928, 'steps': 41368, 'loss/train': 1.6918641328811646} 02/26/2022 07:08:29 - INFO - codeparrot_training - Step 41369: {'lr': 3.8839148240613075e-05, 'samples': 21181440, 'steps': 41369, 'loss/train': 2.7087740898132324} 02/26/2022 07:08:34 - INFO - codeparrot_training - Step 41370: {'lr': 3.883038938685585e-05, 'samples': 21181952, 'steps': 41370, 'loss/train': 1.932759404182434} 02/26/2022 07:08:38 - INFO - codeparrot_training - Step 41371: {'lr': 3.882163143768211e-05, 'samples': 21182464, 'steps': 41371, 'loss/train': 1.4846620559692383} 02/26/2022 07:08:43 - INFO - codeparrot_training - Step 41372: {'lr': 3.8812874393129524e-05, 'samples': 21182976, 'steps': 41372, 'loss/train': 1.9349771738052368} 02/26/2022 07:08:47 - INFO - codeparrot_training - Step 41373: {'lr': 3.8804118253235515e-05, 'samples': 21183488, 'steps': 41373, 'loss/train': 1.2578426599502563} 02/26/2022 07:08:52 - INFO - codeparrot_training - Step 41374: {'lr': 3.87953630180376e-05, 'samples': 21184000, 'steps': 41374, 'loss/train': 2.2274463176727295} 02/26/2022 07:09:00 - INFO - codeparrot_training - Step 41375: {'lr': 3.878660868757322e-05, 'samples': 21184512, 'steps': 41375, 'loss/train': 1.3964370489120483} 02/26/2022 07:09:03 - INFO - codeparrot_training - Step 41376: {'lr': 3.877785526188002e-05, 'samples': 21185024, 'steps': 41376, 'loss/train': 1.1004981994628906} 02/26/2022 07:09:09 - INFO - codeparrot_training - Step 41377: {'lr': 3.8769102740995346e-05, 'samples': 21185536, 'steps': 41377, 'loss/train': 1.451279640197754} 02/26/2022 07:09:12 - INFO - codeparrot_training - Step 41378: {'lr': 3.876035112495688e-05, 'samples': 21186048, 'steps': 41378, 'loss/train': 2.0096492767333984} 02/26/2022 07:09:18 - INFO - codeparrot_training - Step 41379: {'lr': 3.8751600413801877e-05, 'samples': 21186560, 'steps': 41379, 'loss/train': 1.3239006996154785} 02/26/2022 07:09:21 - INFO - codeparrot_training - Step 41380: {'lr': 3.874285060756799e-05, 'samples': 21187072, 'steps': 41380, 'loss/train': 1.61467444896698} 02/26/2022 07:09:27 - INFO - codeparrot_training - Step 41381: {'lr': 3.8734101706292604e-05, 'samples': 21187584, 'steps': 41381, 'loss/train': 1.8403139114379883} 02/26/2022 07:09:30 - INFO - codeparrot_training - Step 41382: {'lr': 3.8725353710013345e-05, 'samples': 21188096, 'steps': 41382, 'loss/train': 1.2576487064361572} 02/26/2022 07:09:36 - INFO - codeparrot_training - Step 41383: {'lr': 3.871660661876747e-05, 'samples': 21188608, 'steps': 41383, 'loss/train': 2.07369327545166} 02/26/2022 07:09:39 - INFO - codeparrot_training - Step 41384: {'lr': 3.870786043259264e-05, 'samples': 21189120, 'steps': 41384, 'loss/train': 1.441272497177124} 02/26/2022 07:09:47 - INFO - codeparrot_training - Step 41385: {'lr': 3.869911515152616e-05, 'samples': 21189632, 'steps': 41385, 'loss/train': 1.5089815855026245} 02/26/2022 07:09:50 - INFO - codeparrot_training - Step 41386: {'lr': 3.869037077560569e-05, 'samples': 21190144, 'steps': 41386, 'loss/train': 2.227071762084961} 02/26/2022 07:09:56 - INFO - codeparrot_training - Step 41387: {'lr': 3.868162730486849e-05, 'samples': 21190656, 'steps': 41387, 'loss/train': 1.7476576566696167} 02/26/2022 07:09:59 - INFO - codeparrot_training - Step 41388: {'lr': 3.8672884739352125e-05, 'samples': 21191168, 'steps': 41388, 'loss/train': 0.731699526309967} 02/26/2022 07:10:04 - INFO - codeparrot_training - Step 41389: {'lr': 3.8664143079094025e-05, 'samples': 21191680, 'steps': 41389, 'loss/train': 1.8979896306991577} 02/26/2022 07:10:08 - INFO - codeparrot_training - Step 41390: {'lr': 3.865540232413156e-05, 'samples': 21192192, 'steps': 41390, 'loss/train': 1.416177749633789} 02/26/2022 07:10:14 - INFO - codeparrot_training - Step 41391: {'lr': 3.864666247450233e-05, 'samples': 21192704, 'steps': 41391, 'loss/train': 1.3912047147750854} 02/26/2022 07:10:17 - INFO - codeparrot_training - Step 41392: {'lr': 3.8637923530243676e-05, 'samples': 21193216, 'steps': 41392, 'loss/train': 0.915386974811554} 02/26/2022 07:10:23 - INFO - codeparrot_training - Step 41393: {'lr': 3.8629185491393024e-05, 'samples': 21193728, 'steps': 41393, 'loss/train': 1.6362178325653076} 02/26/2022 07:10:26 - INFO - codeparrot_training - Step 41394: {'lr': 3.862044835798778e-05, 'samples': 21194240, 'steps': 41394, 'loss/train': 1.5786510705947876} 02/26/2022 07:10:34 - INFO - codeparrot_training - Step 41395: {'lr': 3.8611712130065454e-05, 'samples': 21194752, 'steps': 41395, 'loss/train': 2.140014410018921} 02/26/2022 07:10:37 - INFO - codeparrot_training - Step 41396: {'lr': 3.860297680766345e-05, 'samples': 21195264, 'steps': 41396, 'loss/train': 0.8324285745620728} 02/26/2022 07:10:43 - INFO - codeparrot_training - Step 41397: {'lr': 3.859424239081916e-05, 'samples': 21195776, 'steps': 41397, 'loss/train': 0.053798526525497437} 02/26/2022 07:10:46 - INFO - codeparrot_training - Step 41398: {'lr': 3.858550887956996e-05, 'samples': 21196288, 'steps': 41398, 'loss/train': 1.1822596788406372} 02/26/2022 07:10:52 - INFO - codeparrot_training - Step 41399: {'lr': 3.8576776273953366e-05, 'samples': 21196800, 'steps': 41399, 'loss/train': 2.7903246879577637} 02/26/2022 07:10:55 - INFO - codeparrot_training - Step 41400: {'lr': 3.85680445740067e-05, 'samples': 21197312, 'steps': 41400, 'loss/train': 0.7123786211013794} 02/26/2022 07:11:01 - INFO - codeparrot_training - Step 41401: {'lr': 3.855931377976743e-05, 'samples': 21197824, 'steps': 41401, 'loss/train': 2.029703140258789} 02/26/2022 07:11:04 - INFO - codeparrot_training - Step 41402: {'lr': 3.855058389127283e-05, 'samples': 21198336, 'steps': 41402, 'loss/train': 1.6012340784072876} 02/26/2022 07:11:08 - INFO - codeparrot_training - Step 41403: {'lr': 3.8541854908560464e-05, 'samples': 21198848, 'steps': 41403, 'loss/train': 1.6181527376174927} 02/26/2022 07:11:13 - INFO - codeparrot_training - Step 41404: {'lr': 3.853312683166765e-05, 'samples': 21199360, 'steps': 41404, 'loss/train': 0.7273722887039185} 02/26/2022 07:11:17 - INFO - codeparrot_training - Step 41405: {'lr': 3.852439966063176e-05, 'samples': 21199872, 'steps': 41405, 'loss/train': 2.57423996925354} 02/26/2022 07:11:22 - INFO - codeparrot_training - Step 41406: {'lr': 3.851567339549014e-05, 'samples': 21200384, 'steps': 41406, 'loss/train': 1.1345417499542236} 02/26/2022 07:11:26 - INFO - codeparrot_training - Step 41407: {'lr': 3.850694803628027e-05, 'samples': 21200896, 'steps': 41407, 'loss/train': 1.637454867362976} 02/26/2022 07:11:33 - INFO - codeparrot_training - Step 41408: {'lr': 3.8498223583039476e-05, 'samples': 21201408, 'steps': 41408, 'loss/train': 1.715193271636963} 02/26/2022 07:11:37 - INFO - codeparrot_training - Step 41409: {'lr': 3.8489500035805145e-05, 'samples': 21201920, 'steps': 41409, 'loss/train': 1.8838248252868652} 02/26/2022 07:11:42 - INFO - codeparrot_training - Step 41410: {'lr': 3.848077739461459e-05, 'samples': 21202432, 'steps': 41410, 'loss/train': 2.3701303005218506} 02/26/2022 07:11:45 - INFO - codeparrot_training - Step 41411: {'lr': 3.847205565950524e-05, 'samples': 21202944, 'steps': 41411, 'loss/train': 1.4260499477386475} 02/26/2022 07:11:51 - INFO - codeparrot_training - Step 41412: {'lr': 3.8463334830514397e-05, 'samples': 21203456, 'steps': 41412, 'loss/train': 1.1480962038040161} 02/26/2022 07:11:57 - INFO - codeparrot_training - Step 41413: {'lr': 3.845461490767957e-05, 'samples': 21203968, 'steps': 41413, 'loss/train': 2.6985738277435303} 02/26/2022 07:12:00 - INFO - codeparrot_training - Step 41414: {'lr': 3.844589589103789e-05, 'samples': 21204480, 'steps': 41414, 'loss/train': 1.3860639333724976} 02/26/2022 07:12:06 - INFO - codeparrot_training - Step 41415: {'lr': 3.843717778062686e-05, 'samples': 21204992, 'steps': 41415, 'loss/train': 2.0419063568115234} 02/26/2022 07:12:09 - INFO - codeparrot_training - Step 41416: {'lr': 3.842846057648375e-05, 'samples': 21205504, 'steps': 41416, 'loss/train': 2.8303675651550293} 02/26/2022 07:12:15 - INFO - codeparrot_training - Step 41417: {'lr': 3.841974427864603e-05, 'samples': 21206016, 'steps': 41417, 'loss/train': 1.1914162635803223} 02/26/2022 07:12:18 - INFO - codeparrot_training - Step 41418: {'lr': 3.841102888715081e-05, 'samples': 21206528, 'steps': 41418, 'loss/train': 1.8867751359939575} 02/26/2022 07:12:24 - INFO - codeparrot_training - Step 41419: {'lr': 3.840231440203565e-05, 'samples': 21207040, 'steps': 41419, 'loss/train': 1.3146039247512817} 02/26/2022 07:12:27 - INFO - codeparrot_training - Step 41420: {'lr': 3.8393600823337707e-05, 'samples': 21207552, 'steps': 41420, 'loss/train': 2.1114535331726074} 02/26/2022 07:12:34 - INFO - codeparrot_training - Step 41421: {'lr': 3.83848881510945e-05, 'samples': 21208064, 'steps': 41421, 'loss/train': 2.3817951679229736} 02/26/2022 07:12:38 - INFO - codeparrot_training - Step 41422: {'lr': 3.837617638534313e-05, 'samples': 21208576, 'steps': 41422, 'loss/train': 2.4134585857391357} 02/26/2022 07:12:43 - INFO - codeparrot_training - Step 41423: {'lr': 3.8367465526121084e-05, 'samples': 21209088, 'steps': 41423, 'loss/train': 1.5053220987319946} 02/26/2022 07:12:47 - INFO - codeparrot_training - Step 41424: {'lr': 3.835875557346552e-05, 'samples': 21209600, 'steps': 41424, 'loss/train': 1.760952353477478} 02/26/2022 07:12:52 - INFO - codeparrot_training - Step 41425: {'lr': 3.8350046527413944e-05, 'samples': 21210112, 'steps': 41425, 'loss/train': 1.8469682931900024} 02/26/2022 07:12:56 - INFO - codeparrot_training - Step 41426: {'lr': 3.834133838800355e-05, 'samples': 21210624, 'steps': 41426, 'loss/train': 1.6918387413024902} 02/26/2022 07:13:01 - INFO - codeparrot_training - Step 41427: {'lr': 3.833263115527163e-05, 'samples': 21211136, 'steps': 41427, 'loss/train': 1.2684355974197388} 02/26/2022 07:13:05 - INFO - codeparrot_training - Step 41428: {'lr': 3.832392482925548e-05, 'samples': 21211648, 'steps': 41428, 'loss/train': 2.085402011871338} 02/26/2022 07:13:10 - INFO - codeparrot_training - Step 41429: {'lr': 3.831521940999247e-05, 'samples': 21212160, 'steps': 41429, 'loss/train': 2.738940477371216} 02/26/2022 07:13:14 - INFO - codeparrot_training - Step 41430: {'lr': 3.830651489751985e-05, 'samples': 21212672, 'steps': 41430, 'loss/train': 0.6943312883377075} 02/26/2022 07:13:19 - INFO - codeparrot_training - Step 41431: {'lr': 3.8297811291874876e-05, 'samples': 21213184, 'steps': 41431, 'loss/train': 0.8365955948829651} 02/26/2022 07:13:23 - INFO - codeparrot_training - Step 41432: {'lr': 3.8289108593094815e-05, 'samples': 21213696, 'steps': 41432, 'loss/train': 1.7320804595947266} 02/26/2022 07:13:30 - INFO - codeparrot_training - Step 41433: {'lr': 3.828040680121703e-05, 'samples': 21214208, 'steps': 41433, 'loss/train': 1.596526861190796} 02/26/2022 07:13:33 - INFO - codeparrot_training - Step 41434: {'lr': 3.827170591627877e-05, 'samples': 21214720, 'steps': 41434, 'loss/train': 1.6446852684020996} 02/26/2022 07:13:39 - INFO - codeparrot_training - Step 41435: {'lr': 3.8263005938317305e-05, 'samples': 21215232, 'steps': 41435, 'loss/train': 1.9777724742889404} 02/26/2022 07:13:42 - INFO - codeparrot_training - Step 41436: {'lr': 3.8254306867369856e-05, 'samples': 21215744, 'steps': 41436, 'loss/train': 2.7991063594818115} 02/26/2022 07:13:48 - INFO - codeparrot_training - Step 41437: {'lr': 3.8245608703473684e-05, 'samples': 21216256, 'steps': 41437, 'loss/train': 1.8059407472610474} 02/26/2022 07:13:51 - INFO - codeparrot_training - Step 41438: {'lr': 3.823691144666613e-05, 'samples': 21216768, 'steps': 41438, 'loss/train': 1.3553271293640137} 02/26/2022 07:13:57 - INFO - codeparrot_training - Step 41439: {'lr': 3.82282150969844e-05, 'samples': 21217280, 'steps': 41439, 'loss/train': 2.0456159114837646} 02/26/2022 07:14:00 - INFO - codeparrot_training - Step 41440: {'lr': 3.821951965446577e-05, 'samples': 21217792, 'steps': 41440, 'loss/train': 2.8711917400360107} 02/26/2022 07:14:06 - INFO - codeparrot_training - Step 41441: {'lr': 3.821082511914739e-05, 'samples': 21218304, 'steps': 41441, 'loss/train': 1.0734338760375977} 02/26/2022 07:14:09 - INFO - codeparrot_training - Step 41442: {'lr': 3.820213149106666e-05, 'samples': 21218816, 'steps': 41442, 'loss/train': 0.9205026626586914} 02/26/2022 07:14:16 - INFO - codeparrot_training - Step 41443: {'lr': 3.8193438770260744e-05, 'samples': 21219328, 'steps': 41443, 'loss/train': 2.3073084354400635} 02/26/2022 07:14:20 - INFO - codeparrot_training - Step 41444: {'lr': 3.818474695676685e-05, 'samples': 21219840, 'steps': 41444, 'loss/train': 1.3472200632095337} 02/26/2022 07:14:25 - INFO - codeparrot_training - Step 41445: {'lr': 3.8176056050622214e-05, 'samples': 21220352, 'steps': 41445, 'loss/train': 1.788814902305603} 02/26/2022 07:14:29 - INFO - codeparrot_training - Step 41446: {'lr': 3.816736605186416e-05, 'samples': 21220864, 'steps': 41446, 'loss/train': 2.5812249183654785} 02/26/2022 07:14:34 - INFO - codeparrot_training - Step 41447: {'lr': 3.8158676960529825e-05, 'samples': 21221376, 'steps': 41447, 'loss/train': 1.7620497941970825} 02/26/2022 07:14:38 - INFO - codeparrot_training - Step 41448: {'lr': 3.8149988776656455e-05, 'samples': 21221888, 'steps': 41448, 'loss/train': 0.7768648266792297} 02/26/2022 07:14:43 - INFO - codeparrot_training - Step 41449: {'lr': 3.8141301500281196e-05, 'samples': 21222400, 'steps': 41449, 'loss/train': 1.3608688116073608} 02/26/2022 07:14:47 - INFO - codeparrot_training - Step 41450: {'lr': 3.8132615131441396e-05, 'samples': 21222912, 'steps': 41450, 'loss/train': 1.1408246755599976} 02/26/2022 07:14:52 - INFO - codeparrot_training - Step 41451: {'lr': 3.812392967017414e-05, 'samples': 21223424, 'steps': 41451, 'loss/train': 1.055207371711731} 02/26/2022 07:14:56 - INFO - codeparrot_training - Step 41452: {'lr': 3.8115245116516816e-05, 'samples': 21223936, 'steps': 41452, 'loss/train': 2.232064962387085} 02/26/2022 07:15:02 - INFO - codeparrot_training - Step 41453: {'lr': 3.810656147050637e-05, 'samples': 21224448, 'steps': 41453, 'loss/train': 1.3543843030929565} 02/26/2022 07:15:06 - INFO - codeparrot_training - Step 41454: {'lr': 3.8097878732180206e-05, 'samples': 21224960, 'steps': 41454, 'loss/train': 1.2970480918884277} 02/26/2022 07:15:11 - INFO - codeparrot_training - Step 41455: {'lr': 3.808919690157536e-05, 'samples': 21225472, 'steps': 41455, 'loss/train': 0.9547430276870728} 02/26/2022 07:15:15 - INFO - codeparrot_training - Step 41456: {'lr': 3.808051597872925e-05, 'samples': 21225984, 'steps': 41456, 'loss/train': 0.621574342250824} 02/26/2022 07:15:20 - INFO - codeparrot_training - Step 41457: {'lr': 3.8071835963678775e-05, 'samples': 21226496, 'steps': 41457, 'loss/train': 1.4541332721710205} 02/26/2022 07:15:24 - INFO - codeparrot_training - Step 41458: {'lr': 3.806315685646134e-05, 'samples': 21227008, 'steps': 41458, 'loss/train': 1.678794264793396} 02/26/2022 07:15:29 - INFO - codeparrot_training - Step 41459: {'lr': 3.805447865711398e-05, 'samples': 21227520, 'steps': 41459, 'loss/train': 1.676421880722046} 02/26/2022 07:15:33 - INFO - codeparrot_training - Step 41460: {'lr': 3.804580136567398e-05, 'samples': 21228032, 'steps': 41460, 'loss/train': 1.8396148681640625} 02/26/2022 07:15:38 - INFO - codeparrot_training - Step 41461: {'lr': 3.803712498217846e-05, 'samples': 21228544, 'steps': 41461, 'loss/train': 2.3146557807922363} 02/26/2022 07:15:42 - INFO - codeparrot_training - Step 41462: {'lr': 3.802844950666462e-05, 'samples': 21229056, 'steps': 41462, 'loss/train': 1.413842797279358} 02/26/2022 07:15:47 - INFO - codeparrot_training - Step 41463: {'lr': 3.80197749391695e-05, 'samples': 21229568, 'steps': 41463, 'loss/train': 1.1379414796829224} 02/26/2022 07:15:51 - INFO - codeparrot_training - Step 41464: {'lr': 3.8011101279730424e-05, 'samples': 21230080, 'steps': 41464, 'loss/train': 2.032546281814575} 02/26/2022 07:15:56 - INFO - codeparrot_training - Step 41465: {'lr': 3.800242852838448e-05, 'samples': 21230592, 'steps': 41465, 'loss/train': 1.5792001485824585} 02/26/2022 07:15:59 - INFO - codeparrot_training - Step 41466: {'lr': 3.799375668516883e-05, 'samples': 21231104, 'steps': 41466, 'loss/train': 1.4508529901504517} 02/26/2022 07:16:05 - INFO - codeparrot_training - Step 41467: {'lr': 3.798508575012052e-05, 'samples': 21231616, 'steps': 41467, 'loss/train': 1.586324691772461} 02/26/2022 07:16:08 - INFO - codeparrot_training - Step 41468: {'lr': 3.797641572327687e-05, 'samples': 21232128, 'steps': 41468, 'loss/train': 2.8964133262634277} 02/26/2022 07:16:15 - INFO - codeparrot_training - Step 41469: {'lr': 3.796774660467489e-05, 'samples': 21232640, 'steps': 41469, 'loss/train': 1.8242436647415161} 02/26/2022 07:16:18 - INFO - codeparrot_training - Step 41470: {'lr': 3.795907839435178e-05, 'samples': 21233152, 'steps': 41470, 'loss/train': 2.016366958618164} 02/26/2022 07:16:24 - INFO - codeparrot_training - Step 41471: {'lr': 3.7950411092344594e-05, 'samples': 21233664, 'steps': 41471, 'loss/train': 1.2429040670394897} 02/26/2022 07:16:27 - INFO - codeparrot_training - Step 41472: {'lr': 3.794174469869058e-05, 'samples': 21234176, 'steps': 41472, 'loss/train': 2.246628522872925} 02/26/2022 07:16:33 - INFO - codeparrot_training - Step 41473: {'lr': 3.7933079213426756e-05, 'samples': 21234688, 'steps': 41473, 'loss/train': 0.5318734645843506} 02/26/2022 07:16:36 - INFO - codeparrot_training - Step 41474: {'lr': 3.7924414636590326e-05, 'samples': 21235200, 'steps': 41474, 'loss/train': 0.15288417041301727} 02/26/2022 07:16:42 - INFO - codeparrot_training - Step 41475: {'lr': 3.791575096821828e-05, 'samples': 21235712, 'steps': 41475, 'loss/train': 1.821983814239502} 02/26/2022 07:16:45 - INFO - codeparrot_training - Step 41476: {'lr': 3.7907088208347874e-05, 'samples': 21236224, 'steps': 41476, 'loss/train': 2.4748477935791016} 02/26/2022 07:16:52 - INFO - codeparrot_training - Step 41477: {'lr': 3.7898426357016155e-05, 'samples': 21236736, 'steps': 41477, 'loss/train': 1.8321831226348877} 02/26/2022 07:16:55 - INFO - codeparrot_training - Step 41478: {'lr': 3.788976541426023e-05, 'samples': 21237248, 'steps': 41478, 'loss/train': 2.3670949935913086} 02/26/2022 07:16:59 - INFO - codeparrot_training - Step 41479: {'lr': 3.7881105380117134e-05, 'samples': 21237760, 'steps': 41479, 'loss/train': 2.263601064682007} 02/26/2022 07:17:05 - INFO - codeparrot_training - Step 41480: {'lr': 3.787244625462411e-05, 'samples': 21238272, 'steps': 41480, 'loss/train': 2.211785078048706} 02/26/2022 07:17:08 - INFO - codeparrot_training - Step 41481: {'lr': 3.786378803781812e-05, 'samples': 21238784, 'steps': 41481, 'loss/train': 1.4981693029403687} 02/26/2022 07:17:14 - INFO - codeparrot_training - Step 41482: {'lr': 3.7855130729736345e-05, 'samples': 21239296, 'steps': 41482, 'loss/train': 2.750112771987915} 02/26/2022 07:17:17 - INFO - codeparrot_training - Step 41483: {'lr': 3.7846474330415796e-05, 'samples': 21239808, 'steps': 41483, 'loss/train': 0.07967755198478699} 02/26/2022 07:17:23 - INFO - codeparrot_training - Step 41484: {'lr': 3.783781883989354e-05, 'samples': 21240320, 'steps': 41484, 'loss/train': 1.5107117891311646} 02/26/2022 07:17:26 - INFO - codeparrot_training - Step 41485: {'lr': 3.782916425820676e-05, 'samples': 21240832, 'steps': 41485, 'loss/train': 1.8773701190948486} 02/26/2022 07:17:32 - INFO - codeparrot_training - Step 41486: {'lr': 3.78205105853924e-05, 'samples': 21241344, 'steps': 41486, 'loss/train': 2.269402503967285} 02/26/2022 07:17:35 - INFO - codeparrot_training - Step 41487: {'lr': 3.781185782148775e-05, 'samples': 21241856, 'steps': 41487, 'loss/train': 1.4000815153121948} 02/26/2022 07:17:41 - INFO - codeparrot_training - Step 41488: {'lr': 3.7803205966529555e-05, 'samples': 21242368, 'steps': 41488, 'loss/train': 1.6193434000015259} 02/26/2022 07:17:44 - INFO - codeparrot_training - Step 41489: {'lr': 3.779455502055515e-05, 'samples': 21242880, 'steps': 41489, 'loss/train': 1.8121392726898193} 02/26/2022 07:17:51 - INFO - codeparrot_training - Step 41490: {'lr': 3.77859049836014e-05, 'samples': 21243392, 'steps': 41490, 'loss/train': 0.7978044152259827} 02/26/2022 07:17:54 - INFO - codeparrot_training - Step 41491: {'lr': 3.7777255855705574e-05, 'samples': 21243904, 'steps': 41491, 'loss/train': 2.133714437484741} 02/26/2022 07:18:00 - INFO - codeparrot_training - Step 41492: {'lr': 3.7768607636904485e-05, 'samples': 21244416, 'steps': 41492, 'loss/train': 2.0368926525115967} 02/26/2022 07:18:03 - INFO - codeparrot_training - Step 41493: {'lr': 3.775996032723533e-05, 'samples': 21244928, 'steps': 41493, 'loss/train': 1.4956350326538086} 02/26/2022 07:18:09 - INFO - codeparrot_training - Step 41494: {'lr': 3.775131392673506e-05, 'samples': 21245440, 'steps': 41494, 'loss/train': 1.3492099046707153} 02/26/2022 07:18:12 - INFO - codeparrot_training - Step 41495: {'lr': 3.774266843544089e-05, 'samples': 21245952, 'steps': 41495, 'loss/train': 2.071652412414551} 02/26/2022 07:18:18 - INFO - codeparrot_training - Step 41496: {'lr': 3.7734023853389616e-05, 'samples': 21246464, 'steps': 41496, 'loss/train': 2.0169267654418945} 02/26/2022 07:18:21 - INFO - codeparrot_training - Step 41497: {'lr': 3.7725380180618414e-05, 'samples': 21246976, 'steps': 41497, 'loss/train': 2.1661300659179688} 02/26/2022 07:18:27 - INFO - codeparrot_training - Step 41498: {'lr': 3.771673741716425e-05, 'samples': 21247488, 'steps': 41498, 'loss/train': 1.2068510055541992} 02/26/2022 07:18:30 - INFO - codeparrot_training - Step 41499: {'lr': 3.77080955630642e-05, 'samples': 21248000, 'steps': 41499, 'loss/train': 0.5352388024330139} 02/26/2022 07:18:36 - INFO - codeparrot_training - Step 41500: {'lr': 3.76994546183553e-05, 'samples': 21248512, 'steps': 41500, 'loss/train': 1.529056191444397} 02/26/2022 07:18:39 - INFO - codeparrot_training - Step 41501: {'lr': 3.7690814583074496e-05, 'samples': 21249024, 'steps': 41501, 'loss/train': 1.5785040855407715} 02/26/2022 07:18:45 - INFO - codeparrot_training - Step 41502: {'lr': 3.768217545725877e-05, 'samples': 21249536, 'steps': 41502, 'loss/train': 2.4972565174102783} 02/26/2022 07:18:49 - INFO - codeparrot_training - Step 41503: {'lr': 3.767353724094527e-05, 'samples': 21250048, 'steps': 41503, 'loss/train': 2.246795654296875} 02/26/2022 07:18:54 - INFO - codeparrot_training - Step 41504: {'lr': 3.766489993417088e-05, 'samples': 21250560, 'steps': 41504, 'loss/train': 1.4457815885543823} 02/26/2022 07:18:58 - INFO - codeparrot_training - Step 41505: {'lr': 3.765626353697266e-05, 'samples': 21251072, 'steps': 41505, 'loss/train': 0.6029650568962097} 02/26/2022 07:19:03 - INFO - codeparrot_training - Step 41506: {'lr': 3.764762804938751e-05, 'samples': 21251584, 'steps': 41506, 'loss/train': 1.3602945804595947} 02/26/2022 07:19:07 - INFO - codeparrot_training - Step 41507: {'lr': 3.763899347145255e-05, 'samples': 21252096, 'steps': 41507, 'loss/train': 1.9018856287002563} 02/26/2022 07:19:12 - INFO - codeparrot_training - Step 41508: {'lr': 3.76303598032047e-05, 'samples': 21252608, 'steps': 41508, 'loss/train': 0.3855888545513153} 02/26/2022 07:19:16 - INFO - codeparrot_training - Step 41509: {'lr': 3.7621727044680974e-05, 'samples': 21253120, 'steps': 41509, 'loss/train': 2.094248056411743} 02/26/2022 07:19:21 - INFO - codeparrot_training - Step 41510: {'lr': 3.761309519591827e-05, 'samples': 21253632, 'steps': 41510, 'loss/train': 1.4153131246566772} 02/26/2022 07:19:25 - INFO - codeparrot_training - Step 41511: {'lr': 3.760446425695368e-05, 'samples': 21254144, 'steps': 41511, 'loss/train': 1.2816791534423828} 02/26/2022 07:19:30 - INFO - codeparrot_training - Step 41512: {'lr': 3.759583422782406e-05, 'samples': 21254656, 'steps': 41512, 'loss/train': 0.9876706600189209} 02/26/2022 07:19:34 - INFO - codeparrot_training - Step 41513: {'lr': 3.7587205108566556e-05, 'samples': 21255168, 'steps': 41513, 'loss/train': 0.522168755531311} 02/26/2022 07:19:39 - INFO - codeparrot_training - Step 41514: {'lr': 3.7578576899217896e-05, 'samples': 21255680, 'steps': 41514, 'loss/train': 1.8396189212799072} 02/26/2022 07:19:43 - INFO - codeparrot_training - Step 41515: {'lr': 3.756994959981524e-05, 'samples': 21256192, 'steps': 41515, 'loss/train': 0.3136066198348999} 02/26/2022 07:19:49 - INFO - codeparrot_training - Step 41516: {'lr': 3.7561323210395434e-05, 'samples': 21256704, 'steps': 41516, 'loss/train': 2.4214723110198975} 02/26/2022 07:19:52 - INFO - codeparrot_training - Step 41517: {'lr': 3.755269773099548e-05, 'samples': 21257216, 'steps': 41517, 'loss/train': 2.0351264476776123} 02/26/2022 07:19:58 - INFO - codeparrot_training - Step 41518: {'lr': 3.754407316165226e-05, 'samples': 21257728, 'steps': 41518, 'loss/train': 1.9207797050476074} 02/26/2022 07:20:01 - INFO - codeparrot_training - Step 41519: {'lr': 3.7535449502402814e-05, 'samples': 21258240, 'steps': 41519, 'loss/train': 1.4459257125854492} 02/26/2022 07:20:07 - INFO - codeparrot_training - Step 41520: {'lr': 3.752682675328406e-05, 'samples': 21258752, 'steps': 41520, 'loss/train': 1.9852283000946045} 02/26/2022 07:20:10 - INFO - codeparrot_training - Step 41521: {'lr': 3.751820491433283e-05, 'samples': 21259264, 'steps': 41521, 'loss/train': 1.5598645210266113} 02/26/2022 07:20:16 - INFO - codeparrot_training - Step 41522: {'lr': 3.750958398558621e-05, 'samples': 21259776, 'steps': 41522, 'loss/train': 1.1238679885864258} 02/26/2022 07:20:19 - INFO - codeparrot_training - Step 41523: {'lr': 3.7500963967081054e-05, 'samples': 21260288, 'steps': 41523, 'loss/train': 2.8799936771392822} 02/26/2022 07:20:25 - INFO - codeparrot_training - Step 41524: {'lr': 3.7492344858854275e-05, 'samples': 21260800, 'steps': 41524, 'loss/train': 1.6685857772827148} 02/26/2022 07:20:28 - INFO - codeparrot_training - Step 41525: {'lr': 3.748372666094277e-05, 'samples': 21261312, 'steps': 41525, 'loss/train': 1.3893134593963623} 02/26/2022 07:20:35 - INFO - codeparrot_training - Step 41526: {'lr': 3.7475109373383556e-05, 'samples': 21261824, 'steps': 41526, 'loss/train': 1.341504454612732} 02/26/2022 07:20:40 - INFO - codeparrot_training - Step 41527: {'lr': 3.746649299621349e-05, 'samples': 21262336, 'steps': 41527, 'loss/train': 1.371767520904541} 02/26/2022 07:20:44 - INFO - codeparrot_training - Step 41528: {'lr': 3.745787752946947e-05, 'samples': 21262848, 'steps': 41528, 'loss/train': 1.3130661249160767} 02/26/2022 07:20:49 - INFO - codeparrot_training - Step 41529: {'lr': 3.744926297318838e-05, 'samples': 21263360, 'steps': 41529, 'loss/train': 2.644075870513916} 02/26/2022 07:20:53 - INFO - codeparrot_training - Step 41530: {'lr': 3.7440649327407205e-05, 'samples': 21263872, 'steps': 41530, 'loss/train': 1.7530325651168823} 02/26/2022 07:20:58 - INFO - codeparrot_training - Step 41531: {'lr': 3.743203659216277e-05, 'samples': 21264384, 'steps': 41531, 'loss/train': 1.6723353862762451} 02/26/2022 07:21:02 - INFO - codeparrot_training - Step 41532: {'lr': 3.7423424767492024e-05, 'samples': 21264896, 'steps': 41532, 'loss/train': 1.9476895332336426} 02/26/2022 07:21:07 - INFO - codeparrot_training - Step 41533: {'lr': 3.7414813853431734e-05, 'samples': 21265408, 'steps': 41533, 'loss/train': 1.081854224205017} 02/26/2022 07:21:11 - INFO - codeparrot_training - Step 41534: {'lr': 3.740620385001897e-05, 'samples': 21265920, 'steps': 41534, 'loss/train': 0.943854808807373} 02/26/2022 07:21:17 - INFO - codeparrot_training - Step 41535: {'lr': 3.739759475729049e-05, 'samples': 21266432, 'steps': 41535, 'loss/train': 2.0230562686920166} 02/26/2022 07:21:20 - INFO - codeparrot_training - Step 41536: {'lr': 3.738898657528322e-05, 'samples': 21266944, 'steps': 41536, 'loss/train': 1.4306625127792358} 02/26/2022 07:21:26 - INFO - codeparrot_training - Step 41537: {'lr': 3.738037930403399e-05, 'samples': 21267456, 'steps': 41537, 'loss/train': 0.8857371807098389} 02/26/2022 07:21:29 - INFO - codeparrot_training - Step 41538: {'lr': 3.737177294357971e-05, 'samples': 21267968, 'steps': 41538, 'loss/train': 1.7231508493423462} 02/26/2022 07:21:35 - INFO - codeparrot_training - Step 41539: {'lr': 3.736316749395727e-05, 'samples': 21268480, 'steps': 41539, 'loss/train': 1.501214623451233} 02/26/2022 07:21:38 - INFO - codeparrot_training - Step 41540: {'lr': 3.735456295520348e-05, 'samples': 21268992, 'steps': 41540, 'loss/train': 1.1314055919647217} 02/26/2022 07:21:44 - INFO - codeparrot_training - Step 41541: {'lr': 3.734595932735518e-05, 'samples': 21269504, 'steps': 41541, 'loss/train': 2.9809787273406982} 02/26/2022 07:21:47 - INFO - codeparrot_training - Step 41542: {'lr': 3.733735661044932e-05, 'samples': 21270016, 'steps': 41542, 'loss/train': 0.18138311803340912} 02/26/2022 07:21:53 - INFO - codeparrot_training - Step 41543: {'lr': 3.732875480452269e-05, 'samples': 21270528, 'steps': 41543, 'loss/train': 2.3136279582977295} 02/26/2022 07:21:56 - INFO - codeparrot_training - Step 41544: {'lr': 3.732015390961213e-05, 'samples': 21271040, 'steps': 41544, 'loss/train': 2.3509738445281982} 02/26/2022 07:22:02 - INFO - codeparrot_training - Step 41545: {'lr': 3.731155392575447e-05, 'samples': 21271552, 'steps': 41545, 'loss/train': 0.1376008242368698} 02/26/2022 07:22:05 - INFO - codeparrot_training - Step 41546: {'lr': 3.7302954852986626e-05, 'samples': 21272064, 'steps': 41546, 'loss/train': 2.445629358291626} 02/26/2022 07:22:11 - INFO - codeparrot_training - Step 41547: {'lr': 3.7294356691345316e-05, 'samples': 21272576, 'steps': 41547, 'loss/train': 1.8725117444992065} 02/26/2022 07:22:14 - INFO - codeparrot_training - Step 41548: {'lr': 3.728575944086757e-05, 'samples': 21273088, 'steps': 41548, 'loss/train': 0.10358990728855133} 02/26/2022 07:22:20 - INFO - codeparrot_training - Step 41549: {'lr': 3.7277163101589935e-05, 'samples': 21273600, 'steps': 41549, 'loss/train': 1.79083251953125} 02/26/2022 07:22:23 - INFO - codeparrot_training - Step 41550: {'lr': 3.726856767354947e-05, 'samples': 21274112, 'steps': 41550, 'loss/train': 1.5838149785995483} 02/26/2022 07:22:30 - INFO - codeparrot_training - Step 41551: {'lr': 3.725997315678287e-05, 'samples': 21274624, 'steps': 41551, 'loss/train': 0.5059872269630432} 02/26/2022 07:22:33 - INFO - codeparrot_training - Step 41552: {'lr': 3.725137955132707e-05, 'samples': 21275136, 'steps': 41552, 'loss/train': 1.694729208946228} 02/26/2022 07:22:39 - INFO - codeparrot_training - Step 41553: {'lr': 3.724278685721871e-05, 'samples': 21275648, 'steps': 41553, 'loss/train': 2.1793203353881836} 02/26/2022 07:22:42 - INFO - codeparrot_training - Step 41554: {'lr': 3.723419507449477e-05, 'samples': 21276160, 'steps': 41554, 'loss/train': 2.447911262512207} 02/26/2022 07:22:48 - INFO - codeparrot_training - Step 41555: {'lr': 3.7225604203191904e-05, 'samples': 21276672, 'steps': 41555, 'loss/train': 2.4053971767425537} 02/26/2022 07:22:51 - INFO - codeparrot_training - Step 41556: {'lr': 3.721701424334709e-05, 'samples': 21277184, 'steps': 41556, 'loss/train': 1.8048169612884521} 02/26/2022 07:22:57 - INFO - codeparrot_training - Step 41557: {'lr': 3.720842519499693e-05, 'samples': 21277696, 'steps': 41557, 'loss/train': 1.8096309900283813} 02/26/2022 07:23:00 - INFO - codeparrot_training - Step 41558: {'lr': 3.719983705817834e-05, 'samples': 21278208, 'steps': 41558, 'loss/train': 1.048996090888977} 02/26/2022 07:23:06 - INFO - codeparrot_training - Step 41559: {'lr': 3.719124983292802e-05, 'samples': 21278720, 'steps': 41559, 'loss/train': 0.7377447485923767} 02/26/2022 07:23:09 - INFO - codeparrot_training - Step 41560: {'lr': 3.718266351928287e-05, 'samples': 21279232, 'steps': 41560, 'loss/train': 1.9465328454971313} 02/26/2022 07:23:15 - INFO - codeparrot_training - Step 41561: {'lr': 3.717407811727963e-05, 'samples': 21279744, 'steps': 41561, 'loss/train': 2.427617073059082} 02/26/2022 07:23:19 - INFO - codeparrot_training - Step 41562: {'lr': 3.716549362695504e-05, 'samples': 21280256, 'steps': 41562, 'loss/train': 1.2041329145431519} 02/26/2022 07:23:24 - INFO - codeparrot_training - Step 41563: {'lr': 3.7156910048345846e-05, 'samples': 21280768, 'steps': 41563, 'loss/train': 1.2262251377105713} 02/26/2022 07:23:28 - INFO - codeparrot_training - Step 41564: {'lr': 3.7148327381488906e-05, 'samples': 21281280, 'steps': 41564, 'loss/train': 1.2153651714324951} 02/26/2022 07:23:33 - INFO - codeparrot_training - Step 41565: {'lr': 3.7139745626420976e-05, 'samples': 21281792, 'steps': 41565, 'loss/train': 1.364759922027588} 02/26/2022 07:23:37 - INFO - codeparrot_training - Step 41566: {'lr': 3.713116478317877e-05, 'samples': 21282304, 'steps': 41566, 'loss/train': 1.4748543500900269} 02/26/2022 07:23:42 - INFO - codeparrot_training - Step 41567: {'lr': 3.712258485179904e-05, 'samples': 21282816, 'steps': 41567, 'loss/train': 1.450737714767456} 02/26/2022 07:23:46 - INFO - codeparrot_training - Step 41568: {'lr': 3.7114005832318526e-05, 'samples': 21283328, 'steps': 41568, 'loss/train': 1.6766945123672485} 02/26/2022 07:23:51 - INFO - codeparrot_training - Step 41569: {'lr': 3.7105427724774074e-05, 'samples': 21283840, 'steps': 41569, 'loss/train': 0.7602059245109558} 02/26/2022 07:23:55 - INFO - codeparrot_training - Step 41570: {'lr': 3.709685052920234e-05, 'samples': 21284352, 'steps': 41570, 'loss/train': 1.4825658798217773} 02/26/2022 07:24:01 - INFO - codeparrot_training - Step 41571: {'lr': 3.708827424564012e-05, 'samples': 21284864, 'steps': 41571, 'loss/train': 0.09666864573955536} 02/26/2022 07:24:04 - INFO - codeparrot_training - Step 41572: {'lr': 3.707969887412405e-05, 'samples': 21285376, 'steps': 41572, 'loss/train': 2.673663854598999} 02/26/2022 07:24:10 - INFO - codeparrot_training - Step 41573: {'lr': 3.707112441469102e-05, 'samples': 21285888, 'steps': 41573, 'loss/train': 2.849592447280884} 02/26/2022 07:24:15 - INFO - codeparrot_training - Step 41574: {'lr': 3.706255086737767e-05, 'samples': 21286400, 'steps': 41574, 'loss/train': 2.052672863006592} 02/26/2022 07:24:19 - INFO - codeparrot_training - Step 41575: {'lr': 3.705397823222073e-05, 'samples': 21286912, 'steps': 41575, 'loss/train': 1.2658662796020508} 02/26/2022 07:24:25 - INFO - codeparrot_training - Step 41576: {'lr': 3.704540650925686e-05, 'samples': 21287424, 'steps': 41576, 'loss/train': 2.061177968978882} 02/26/2022 07:24:28 - INFO - codeparrot_training - Step 41577: {'lr': 3.7036835698522925e-05, 'samples': 21287936, 'steps': 41577, 'loss/train': 1.895163893699646} 02/26/2022 07:24:31 - INFO - codeparrot_training - Step 41578: {'lr': 3.7028265800055555e-05, 'samples': 21288448, 'steps': 41578, 'loss/train': 2.684748411178589} 02/26/2022 07:24:37 - INFO - codeparrot_training - Step 41579: {'lr': 3.701969681389147e-05, 'samples': 21288960, 'steps': 41579, 'loss/train': 0.38307473063468933} 02/26/2022 07:24:40 - INFO - codeparrot_training - Step 41580: {'lr': 3.7011128740067315e-05, 'samples': 21289472, 'steps': 41580, 'loss/train': 1.5973625183105469} 02/26/2022 07:24:46 - INFO - codeparrot_training - Step 41581: {'lr': 3.700256157861992e-05, 'samples': 21289984, 'steps': 41581, 'loss/train': 0.266282856464386} 02/26/2022 07:24:52 - INFO - codeparrot_training - Step 41582: {'lr': 3.6993995329585864e-05, 'samples': 21290496, 'steps': 41582, 'loss/train': 1.240788221359253} 02/26/2022 07:24:56 - INFO - codeparrot_training - Step 41583: {'lr': 3.6985429993002e-05, 'samples': 21291008, 'steps': 41583, 'loss/train': 2.1780056953430176} 02/26/2022 07:24:59 - INFO - codeparrot_training - Step 41584: {'lr': 3.697686556890481e-05, 'samples': 21291520, 'steps': 41584, 'loss/train': 2.1679952144622803} 02/26/2022 07:25:05 - INFO - codeparrot_training - Step 41585: {'lr': 3.696830205733112e-05, 'samples': 21292032, 'steps': 41585, 'loss/train': 1.212868332862854} 02/26/2022 07:25:10 - INFO - codeparrot_training - Step 41586: {'lr': 3.6959739458317545e-05, 'samples': 21292544, 'steps': 41586, 'loss/train': 2.5805232524871826} 02/26/2022 07:25:14 - INFO - codeparrot_training - Step 41587: {'lr': 3.695117777190088e-05, 'samples': 21293056, 'steps': 41587, 'loss/train': 1.4765158891677856} 02/26/2022 07:25:19 - INFO - codeparrot_training - Step 41588: {'lr': 3.694261699811763e-05, 'samples': 21293568, 'steps': 41588, 'loss/train': 1.8507057428359985} 02/26/2022 07:25:23 - INFO - codeparrot_training - Step 41589: {'lr': 3.693405713700462e-05, 'samples': 21294080, 'steps': 41589, 'loss/train': 1.4172418117523193} 02/26/2022 07:25:28 - INFO - codeparrot_training - Step 41590: {'lr': 3.692549818859839e-05, 'samples': 21294592, 'steps': 41590, 'loss/train': 1.536205768585205} 02/26/2022 07:25:32 - INFO - codeparrot_training - Step 41591: {'lr': 3.691694015293576e-05, 'samples': 21295104, 'steps': 41591, 'loss/train': 0.9635187387466431} 02/26/2022 07:25:38 - INFO - codeparrot_training - Step 41592: {'lr': 3.69083830300532e-05, 'samples': 21295616, 'steps': 41592, 'loss/train': 1.1006048917770386} 02/26/2022 07:25:41 - INFO - codeparrot_training - Step 41593: {'lr': 3.689982681998752e-05, 'samples': 21296128, 'steps': 41593, 'loss/train': 0.6185737252235413} 02/26/2022 07:25:47 - INFO - codeparrot_training - Step 41594: {'lr': 3.689127152277524e-05, 'samples': 21296640, 'steps': 41594, 'loss/train': 1.6980297565460205} 02/26/2022 07:25:50 - INFO - codeparrot_training - Step 41595: {'lr': 3.688271713845318e-05, 'samples': 21297152, 'steps': 41595, 'loss/train': 0.10553835332393646} 02/26/2022 07:25:56 - INFO - codeparrot_training - Step 41596: {'lr': 3.687416366705787e-05, 'samples': 21297664, 'steps': 41596, 'loss/train': 1.4553296566009521} 02/26/2022 07:25:59 - INFO - codeparrot_training - Step 41597: {'lr': 3.686561110862596e-05, 'samples': 21298176, 'steps': 41597, 'loss/train': 1.349586844444275} 02/26/2022 07:26:05 - INFO - codeparrot_training - Step 41598: {'lr': 3.6857059463194036e-05, 'samples': 21298688, 'steps': 41598, 'loss/train': 1.4395700693130493} 02/26/2022 07:26:08 - INFO - codeparrot_training - Step 41599: {'lr': 3.6848508730798844e-05, 'samples': 21299200, 'steps': 41599, 'loss/train': 1.942911148071289} 02/26/2022 07:26:14 - INFO - codeparrot_training - Step 41600: {'lr': 3.6839958911476953e-05, 'samples': 21299712, 'steps': 41600, 'loss/train': 1.5435937643051147} 02/26/2022 07:26:17 - INFO - codeparrot_training - Step 41601: {'lr': 3.683141000526502e-05, 'samples': 21300224, 'steps': 41601, 'loss/train': 1.9384915828704834} 02/26/2022 07:26:23 - INFO - codeparrot_training - Step 41602: {'lr': 3.682286201219956e-05, 'samples': 21300736, 'steps': 41602, 'loss/train': 1.2798683643341064} 02/26/2022 07:26:26 - INFO - codeparrot_training - Step 41603: {'lr': 3.681431493231732e-05, 'samples': 21301248, 'steps': 41603, 'loss/train': 1.2349262237548828} 02/26/2022 07:26:32 - INFO - codeparrot_training - Step 41604: {'lr': 3.6805768765654865e-05, 'samples': 21301760, 'steps': 41604, 'loss/train': 1.3032633066177368} 02/26/2022 07:26:35 - INFO - codeparrot_training - Step 41605: {'lr': 3.679722351224879e-05, 'samples': 21302272, 'steps': 41605, 'loss/train': 0.6310141086578369} 02/26/2022 07:26:41 - INFO - codeparrot_training - Step 41606: {'lr': 3.678867917213569e-05, 'samples': 21302784, 'steps': 41606, 'loss/train': 1.987351894378662} 02/26/2022 07:26:44 - INFO - codeparrot_training - Step 41607: {'lr': 3.678013574535222e-05, 'samples': 21303296, 'steps': 41607, 'loss/train': 3.2760982513427734} 02/26/2022 07:26:50 - INFO - codeparrot_training - Step 41608: {'lr': 3.6771593231934934e-05, 'samples': 21303808, 'steps': 41608, 'loss/train': 1.3467310667037964} 02/26/2022 07:26:54 - INFO - codeparrot_training - Step 41609: {'lr': 3.6763051631920426e-05, 'samples': 21304320, 'steps': 41609, 'loss/train': 0.9354490041732788} 02/26/2022 07:26:59 - INFO - codeparrot_training - Step 41610: {'lr': 3.6754510945345265e-05, 'samples': 21304832, 'steps': 41610, 'loss/train': 1.13763427734375} 02/26/2022 07:27:03 - INFO - codeparrot_training - Step 41611: {'lr': 3.6745971172246126e-05, 'samples': 21305344, 'steps': 41611, 'loss/train': 1.8417457342147827} 02/26/2022 07:27:08 - INFO - codeparrot_training - Step 41612: {'lr': 3.67374323126595e-05, 'samples': 21305856, 'steps': 41612, 'loss/train': 2.6171836853027344} 02/26/2022 07:27:12 - INFO - codeparrot_training - Step 41613: {'lr': 3.6728894366622026e-05, 'samples': 21306368, 'steps': 41613, 'loss/train': 1.8170359134674072} 02/26/2022 07:27:18 - INFO - codeparrot_training - Step 41614: {'lr': 3.672035733417023e-05, 'samples': 21306880, 'steps': 41614, 'loss/train': 0.8051912188529968} 02/26/2022 07:27:21 - INFO - codeparrot_training - Step 41615: {'lr': 3.671182121534064e-05, 'samples': 21307392, 'steps': 41615, 'loss/train': 2.0435030460357666} 02/26/2022 07:27:27 - INFO - codeparrot_training - Step 41616: {'lr': 3.670328601016995e-05, 'samples': 21307904, 'steps': 41616, 'loss/train': 1.3559987545013428} 02/26/2022 07:27:30 - INFO - codeparrot_training - Step 41617: {'lr': 3.669475171869457e-05, 'samples': 21308416, 'steps': 41617, 'loss/train': 1.9512840509414673} 02/26/2022 07:27:36 - INFO - codeparrot_training - Step 41618: {'lr': 3.668621834095129e-05, 'samples': 21308928, 'steps': 41618, 'loss/train': 1.828302025794983} 02/26/2022 07:27:40 - INFO - codeparrot_training - Step 41619: {'lr': 3.66776858769764e-05, 'samples': 21309440, 'steps': 41619, 'loss/train': 2.477667808532715} 02/26/2022 07:27:45 - INFO - codeparrot_training - Step 41620: {'lr': 3.6669154326806594e-05, 'samples': 21309952, 'steps': 41620, 'loss/train': 0.9591473937034607} 02/26/2022 07:27:49 - INFO - codeparrot_training - Step 41621: {'lr': 3.6660623690478344e-05, 'samples': 21310464, 'steps': 41621, 'loss/train': 1.0352134704589844} 02/26/2022 07:27:54 - INFO - codeparrot_training - Step 41622: {'lr': 3.665209396802838e-05, 'samples': 21310976, 'steps': 41622, 'loss/train': 2.271552324295044} 02/26/2022 07:27:58 - INFO - codeparrot_training - Step 41623: {'lr': 3.6643565159492963e-05, 'samples': 21311488, 'steps': 41623, 'loss/train': 1.494215965270996} 02/26/2022 07:28:03 - INFO - codeparrot_training - Step 41624: {'lr': 3.663503726490883e-05, 'samples': 21312000, 'steps': 41624, 'loss/train': 1.996005892753601} 02/26/2022 07:28:07 - INFO - codeparrot_training - Step 41625: {'lr': 3.662651028431238e-05, 'samples': 21312512, 'steps': 41625, 'loss/train': 1.11830472946167} 02/26/2022 07:28:12 - INFO - codeparrot_training - Step 41626: {'lr': 3.661798421774032e-05, 'samples': 21313024, 'steps': 41626, 'loss/train': 2.4957258701324463} 02/26/2022 07:28:16 - INFO - codeparrot_training - Step 41627: {'lr': 3.660945906522895e-05, 'samples': 21313536, 'steps': 41627, 'loss/train': 0.8284900784492493} 02/26/2022 07:28:22 - INFO - codeparrot_training - Step 41628: {'lr': 3.660093482681495e-05, 'samples': 21314048, 'steps': 41628, 'loss/train': 1.233595371246338} 02/26/2022 07:28:25 - INFO - codeparrot_training - Step 41629: {'lr': 3.659241150253473e-05, 'samples': 21314560, 'steps': 41629, 'loss/train': 1.0313550233840942} 02/26/2022 07:28:31 - INFO - codeparrot_training - Step 41630: {'lr': 3.658388909242491e-05, 'samples': 21315072, 'steps': 41630, 'loss/train': 2.564741849899292} 02/26/2022 07:28:34 - INFO - codeparrot_training - Step 41631: {'lr': 3.657536759652194e-05, 'samples': 21315584, 'steps': 41631, 'loss/train': 1.961073398590088} 02/26/2022 07:28:40 - INFO - codeparrot_training - Step 41632: {'lr': 3.656684701486235e-05, 'samples': 21316096, 'steps': 41632, 'loss/train': 2.1633036136627197} 02/26/2022 07:28:43 - INFO - codeparrot_training - Step 41633: {'lr': 3.655832734748252e-05, 'samples': 21316608, 'steps': 41633, 'loss/train': 1.0260932445526123} 02/26/2022 07:28:49 - INFO - codeparrot_training - Step 41634: {'lr': 3.6549808594419115e-05, 'samples': 21317120, 'steps': 41634, 'loss/train': 1.5241156816482544} 02/26/2022 07:28:53 - INFO - codeparrot_training - Step 41635: {'lr': 3.654129075570855e-05, 'samples': 21317632, 'steps': 41635, 'loss/train': 1.3140246868133545} 02/26/2022 07:28:58 - INFO - codeparrot_training - Step 41636: {'lr': 3.653277383138734e-05, 'samples': 21318144, 'steps': 41636, 'loss/train': 1.7592339515686035} 02/26/2022 07:29:02 - INFO - codeparrot_training - Step 41637: {'lr': 3.652425782149185e-05, 'samples': 21318656, 'steps': 41637, 'loss/train': 1.9927699565887451} 02/26/2022 07:29:07 - INFO - codeparrot_training - Step 41638: {'lr': 3.6515742726058734e-05, 'samples': 21319168, 'steps': 41638, 'loss/train': 1.6085721254348755} 02/26/2022 07:29:11 - INFO - codeparrot_training - Step 41639: {'lr': 3.650722854512437e-05, 'samples': 21319680, 'steps': 41639, 'loss/train': 0.17155827581882477} 02/26/2022 07:29:16 - INFO - codeparrot_training - Step 41640: {'lr': 3.6498715278725256e-05, 'samples': 21320192, 'steps': 41640, 'loss/train': 0.1545640230178833} 02/26/2022 07:29:20 - INFO - codeparrot_training - Step 41641: {'lr': 3.6490202926897784e-05, 'samples': 21320704, 'steps': 41641, 'loss/train': 1.3149207830429077} 02/26/2022 07:29:25 - INFO - codeparrot_training - Step 41642: {'lr': 3.648169148967856e-05, 'samples': 21321216, 'steps': 41642, 'loss/train': 1.4005154371261597} 02/26/2022 07:29:29 - INFO - codeparrot_training - Step 41643: {'lr': 3.647318096710389e-05, 'samples': 21321728, 'steps': 41643, 'loss/train': 0.5686456561088562} 02/26/2022 07:29:35 - INFO - codeparrot_training - Step 41644: {'lr': 3.646467135921044e-05, 'samples': 21322240, 'steps': 41644, 'loss/train': 0.6416675448417664} 02/26/2022 07:29:38 - INFO - codeparrot_training - Step 41645: {'lr': 3.645616266603441e-05, 'samples': 21322752, 'steps': 41645, 'loss/train': 2.630011558532715} 02/26/2022 07:29:44 - INFO - codeparrot_training - Step 41646: {'lr': 3.6447654887612453e-05, 'samples': 21323264, 'steps': 41646, 'loss/train': 2.4160327911376953} 02/26/2022 07:29:47 - INFO - codeparrot_training - Step 41647: {'lr': 3.6439148023980905e-05, 'samples': 21323776, 'steps': 41647, 'loss/train': 2.6682546138763428} 02/26/2022 07:29:53 - INFO - codeparrot_training - Step 41648: {'lr': 3.643064207517624e-05, 'samples': 21324288, 'steps': 41648, 'loss/train': 1.207150936126709} 02/26/2022 07:29:56 - INFO - codeparrot_training - Step 41649: {'lr': 3.6422137041234844e-05, 'samples': 21324800, 'steps': 41649, 'loss/train': 4.113279342651367} 02/26/2022 07:30:02 - INFO - codeparrot_training - Step 41650: {'lr': 3.641363292219324e-05, 'samples': 21325312, 'steps': 41650, 'loss/train': 0.5875191688537598} 02/26/2022 07:30:05 - INFO - codeparrot_training - Step 41651: {'lr': 3.640512971808782e-05, 'samples': 21325824, 'steps': 41651, 'loss/train': 2.2507638931274414} 02/26/2022 07:30:11 - INFO - codeparrot_training - Step 41652: {'lr': 3.639662742895497e-05, 'samples': 21326336, 'steps': 41652, 'loss/train': 1.445509672164917} 02/26/2022 07:30:15 - INFO - codeparrot_training - Step 41653: {'lr': 3.6388126054831135e-05, 'samples': 21326848, 'steps': 41653, 'loss/train': 1.7996963262557983} 02/26/2022 07:30:18 - INFO - codeparrot_training - Step 41654: {'lr': 3.637962559575275e-05, 'samples': 21327360, 'steps': 41654, 'loss/train': 0.9225202202796936} 02/26/2022 07:30:24 - INFO - codeparrot_training - Step 41655: {'lr': 3.637112605175624e-05, 'samples': 21327872, 'steps': 41655, 'loss/train': 1.8934005498886108} 02/26/2022 07:30:28 - INFO - codeparrot_training - Step 41656: {'lr': 3.6362627422877936e-05, 'samples': 21328384, 'steps': 41656, 'loss/train': 0.6428785920143127} 02/26/2022 07:30:33 - INFO - codeparrot_training - Step 41657: {'lr': 3.635412970915433e-05, 'samples': 21328896, 'steps': 41657, 'loss/train': 0.699927568435669} 02/26/2022 07:30:37 - INFO - codeparrot_training - Step 41658: {'lr': 3.634563291062182e-05, 'samples': 21329408, 'steps': 41658, 'loss/train': 1.4965465068817139} 02/26/2022 07:30:42 - INFO - codeparrot_training - Step 41659: {'lr': 3.633713702731678e-05, 'samples': 21329920, 'steps': 41659, 'loss/train': 2.6621553897857666} 02/26/2022 07:30:46 - INFO - codeparrot_training - Step 41660: {'lr': 3.6328642059275526e-05, 'samples': 21330432, 'steps': 41660, 'loss/train': 2.6097216606140137} 02/26/2022 07:30:51 - INFO - codeparrot_training - Step 41661: {'lr': 3.632014800653463e-05, 'samples': 21330944, 'steps': 41661, 'loss/train': 1.7781643867492676} 02/26/2022 07:30:55 - INFO - codeparrot_training - Step 41662: {'lr': 3.631165486913027e-05, 'samples': 21331456, 'steps': 41662, 'loss/train': 1.2566637992858887} 02/26/2022 07:31:00 - INFO - codeparrot_training - Step 41663: {'lr': 3.6303162647098956e-05, 'samples': 21331968, 'steps': 41663, 'loss/train': 1.3554425239562988} 02/26/2022 07:31:04 - INFO - codeparrot_training - Step 41664: {'lr': 3.6294671340477006e-05, 'samples': 21332480, 'steps': 41664, 'loss/train': 1.8000521659851074} 02/26/2022 07:31:10 - INFO - codeparrot_training - Step 41665: {'lr': 3.6286180949300944e-05, 'samples': 21332992, 'steps': 41665, 'loss/train': 1.0072382688522339} 02/26/2022 07:31:14 - INFO - codeparrot_training - Step 41666: {'lr': 3.627769147360691e-05, 'samples': 21333504, 'steps': 41666, 'loss/train': 2.0703353881835938} 02/26/2022 07:31:19 - INFO - codeparrot_training - Step 41667: {'lr': 3.626920291343141e-05, 'samples': 21334016, 'steps': 41667, 'loss/train': 2.1244912147521973} 02/26/2022 07:31:23 - INFO - codeparrot_training - Step 41668: {'lr': 3.626071526881072e-05, 'samples': 21334528, 'steps': 41668, 'loss/train': 2.1796233654022217} 02/26/2022 07:31:28 - INFO - codeparrot_training - Step 41669: {'lr': 3.625222853978133e-05, 'samples': 21335040, 'steps': 41669, 'loss/train': 1.580825924873352} 02/26/2022 07:31:32 - INFO - codeparrot_training - Step 41670: {'lr': 3.6243742726379536e-05, 'samples': 21335552, 'steps': 41670, 'loss/train': 1.9262468814849854} 02/26/2022 07:31:37 - INFO - codeparrot_training - Step 41671: {'lr': 3.623525782864165e-05, 'samples': 21336064, 'steps': 41671, 'loss/train': 1.7846593856811523} 02/26/2022 07:31:41 - INFO - codeparrot_training - Step 41672: {'lr': 3.6226773846604e-05, 'samples': 21336576, 'steps': 41672, 'loss/train': 0.3570833206176758} 02/26/2022 07:31:46 - INFO - codeparrot_training - Step 41673: {'lr': 3.6218290780303005e-05, 'samples': 21337088, 'steps': 41673, 'loss/train': 1.769381046295166} 02/26/2022 07:31:50 - INFO - codeparrot_training - Step 41674: {'lr': 3.6209808629774995e-05, 'samples': 21337600, 'steps': 41674, 'loss/train': 1.4203745126724243} 02/26/2022 07:31:56 - INFO - codeparrot_training - Step 41675: {'lr': 3.6201327395056244e-05, 'samples': 21338112, 'steps': 41675, 'loss/train': 1.7463834285736084} 02/26/2022 07:32:00 - INFO - codeparrot_training - Step 41676: {'lr': 3.619284707618309e-05, 'samples': 21338624, 'steps': 41676, 'loss/train': 2.070387125015259} 02/26/2022 07:32:05 - INFO - codeparrot_training - Step 41677: {'lr': 3.6184367673191946e-05, 'samples': 21339136, 'steps': 41677, 'loss/train': 0.94215327501297} 02/26/2022 07:32:09 - INFO - codeparrot_training - Step 41678: {'lr': 3.6175889186119e-05, 'samples': 21339648, 'steps': 41678, 'loss/train': 0.7357382774353027} 02/26/2022 07:32:14 - INFO - codeparrot_training - Step 41679: {'lr': 3.6167411615000786e-05, 'samples': 21340160, 'steps': 41679, 'loss/train': 1.9225282669067383} 02/26/2022 07:32:18 - INFO - codeparrot_training - Step 41680: {'lr': 3.615893495987335e-05, 'samples': 21340672, 'steps': 41680, 'loss/train': 0.03728966787457466} 02/26/2022 07:32:23 - INFO - codeparrot_training - Step 41681: {'lr': 3.6150459220773175e-05, 'samples': 21341184, 'steps': 41681, 'loss/train': 2.291738748550415} 02/26/2022 07:32:27 - INFO - codeparrot_training - Step 41682: {'lr': 3.61419843977365e-05, 'samples': 21341696, 'steps': 41682, 'loss/train': 1.8549846410751343} 02/26/2022 07:32:33 - INFO - codeparrot_training - Step 41683: {'lr': 3.613351049079977e-05, 'samples': 21342208, 'steps': 41683, 'loss/train': 1.6001858711242676} 02/26/2022 07:32:38 - INFO - codeparrot_training - Step 41684: {'lr': 3.612503749999904e-05, 'samples': 21342720, 'steps': 41684, 'loss/train': 1.2245246171951294} 02/26/2022 07:32:42 - INFO - codeparrot_training - Step 41685: {'lr': 3.6116565425370815e-05, 'samples': 21343232, 'steps': 41685, 'loss/train': 1.3882091045379639} 02/26/2022 07:32:47 - INFO - codeparrot_training - Step 41686: {'lr': 3.610809426695122e-05, 'samples': 21343744, 'steps': 41686, 'loss/train': 0.8962826132774353} 02/26/2022 07:32:51 - INFO - codeparrot_training - Step 41687: {'lr': 3.609962402477676e-05, 'samples': 21344256, 'steps': 41687, 'loss/train': 1.2971596717834473} 02/26/2022 07:32:56 - INFO - codeparrot_training - Step 41688: {'lr': 3.6091154698883495e-05, 'samples': 21344768, 'steps': 41688, 'loss/train': 1.413360357284546} 02/26/2022 07:33:00 - INFO - codeparrot_training - Step 41689: {'lr': 3.608268628930783e-05, 'samples': 21345280, 'steps': 41689, 'loss/train': 1.5986542701721191} 02/26/2022 07:33:06 - INFO - codeparrot_training - Step 41690: {'lr': 3.6074218796085966e-05, 'samples': 21345792, 'steps': 41690, 'loss/train': 1.52005934715271} 02/26/2022 07:33:09 - INFO - codeparrot_training - Step 41691: {'lr': 3.606575221925426e-05, 'samples': 21346304, 'steps': 41691, 'loss/train': 1.6590601205825806} 02/26/2022 07:33:15 - INFO - codeparrot_training - Step 41692: {'lr': 3.605728655884896e-05, 'samples': 21346816, 'steps': 41692, 'loss/train': 1.6791789531707764} 02/26/2022 07:33:18 - INFO - codeparrot_training - Step 41693: {'lr': 3.604882181490629e-05, 'samples': 21347328, 'steps': 41693, 'loss/train': 2.900899887084961} 02/26/2022 07:33:24 - INFO - codeparrot_training - Step 41694: {'lr': 3.604035798746255e-05, 'samples': 21347840, 'steps': 41694, 'loss/train': 2.037975311279297} 02/26/2022 07:33:27 - INFO - codeparrot_training - Step 41695: {'lr': 3.603189507655391e-05, 'samples': 21348352, 'steps': 41695, 'loss/train': 0.7879558801651001} 02/26/2022 07:33:33 - INFO - codeparrot_training - Step 41696: {'lr': 3.602343308221675e-05, 'samples': 21348864, 'steps': 41696, 'loss/train': 0.37129315733909607} 02/26/2022 07:33:36 - INFO - codeparrot_training - Step 41697: {'lr': 3.601497200448725e-05, 'samples': 21349376, 'steps': 41697, 'loss/train': 2.123992681503296} 02/26/2022 07:33:42 - INFO - codeparrot_training - Step 41698: {'lr': 3.6006511843401675e-05, 'samples': 21349888, 'steps': 41698, 'loss/train': 0.8518533706665039} 02/26/2022 07:33:45 - INFO - codeparrot_training - Step 41699: {'lr': 3.5998052598996186e-05, 'samples': 21350400, 'steps': 41699, 'loss/train': 2.4207725524902344} 02/26/2022 07:33:51 - INFO - codeparrot_training - Step 41700: {'lr': 3.5989594271307156e-05, 'samples': 21350912, 'steps': 41700, 'loss/train': 1.3484591245651245} 02/26/2022 07:33:55 - INFO - codeparrot_training - Step 41701: {'lr': 3.598113686037075e-05, 'samples': 21351424, 'steps': 41701, 'loss/train': 1.8877475261688232} 02/26/2022 07:34:00 - INFO - codeparrot_training - Step 41702: {'lr': 3.597268036622317e-05, 'samples': 21351936, 'steps': 41702, 'loss/train': 0.37708383798599243} 02/26/2022 07:34:04 - INFO - codeparrot_training - Step 41703: {'lr': 3.596422478890063e-05, 'samples': 21352448, 'steps': 41703, 'loss/train': 3.590773820877075} 02/26/2022 07:34:09 - INFO - codeparrot_training - Step 41704: {'lr': 3.5955770128439424e-05, 'samples': 21352960, 'steps': 41704, 'loss/train': 1.172747015953064} 02/26/2022 07:34:13 - INFO - codeparrot_training - Step 41705: {'lr': 3.594731638487575e-05, 'samples': 21353472, 'steps': 41705, 'loss/train': 2.263465166091919} 02/26/2022 07:34:18 - INFO - codeparrot_training - Step 41706: {'lr': 3.59388635582458e-05, 'samples': 21353984, 'steps': 41706, 'loss/train': 1.7478835582733154} 02/26/2022 07:34:22 - INFO - codeparrot_training - Step 41707: {'lr': 3.593041164858574e-05, 'samples': 21354496, 'steps': 41707, 'loss/train': 2.535195827484131} 02/26/2022 07:34:27 - INFO - codeparrot_training - Step 41708: {'lr': 3.592196065593184e-05, 'samples': 21355008, 'steps': 41708, 'loss/train': 1.4918303489685059} 02/26/2022 07:34:31 - INFO - codeparrot_training - Step 41709: {'lr': 3.591351058032033e-05, 'samples': 21355520, 'steps': 41709, 'loss/train': 1.8929601907730103} 02/26/2022 07:34:37 - INFO - codeparrot_training - Step 41710: {'lr': 3.5905061421787316e-05, 'samples': 21356032, 'steps': 41710, 'loss/train': 1.9910062551498413} 02/26/2022 07:34:40 - INFO - codeparrot_training - Step 41711: {'lr': 3.589661318036902e-05, 'samples': 21356544, 'steps': 41711, 'loss/train': 1.3086901903152466} 02/26/2022 07:34:46 - INFO - codeparrot_training - Step 41712: {'lr': 3.588816585610169e-05, 'samples': 21357056, 'steps': 41712, 'loss/train': 1.525822401046753} 02/26/2022 07:34:50 - INFO - codeparrot_training - Step 41713: {'lr': 3.587971944902146e-05, 'samples': 21357568, 'steps': 41713, 'loss/train': 1.63988196849823} 02/26/2022 07:34:55 - INFO - codeparrot_training - Step 41714: {'lr': 3.5871273959164534e-05, 'samples': 21358080, 'steps': 41714, 'loss/train': 2.04854416847229} 02/26/2022 07:34:59 - INFO - codeparrot_training - Step 41715: {'lr': 3.586282938656702e-05, 'samples': 21358592, 'steps': 41715, 'loss/train': 1.877226710319519} 02/26/2022 07:35:04 - INFO - codeparrot_training - Step 41716: {'lr': 3.585438573126518e-05, 'samples': 21359104, 'steps': 41716, 'loss/train': 1.7811907529830933} 02/26/2022 07:35:08 - INFO - codeparrot_training - Step 41717: {'lr': 3.5845942993295103e-05, 'samples': 21359616, 'steps': 41717, 'loss/train': 2.4954702854156494} 02/26/2022 07:35:13 - INFO - codeparrot_training - Step 41718: {'lr': 3.583750117269313e-05, 'samples': 21360128, 'steps': 41718, 'loss/train': 1.5006673336029053} 02/26/2022 07:35:17 - INFO - codeparrot_training - Step 41719: {'lr': 3.5829060269495204e-05, 'samples': 21360640, 'steps': 41719, 'loss/train': 1.727968692779541} 02/26/2022 07:35:24 - INFO - codeparrot_training - Step 41720: {'lr': 3.5820620283737615e-05, 'samples': 21361152, 'steps': 41720, 'loss/train': 1.1878557205200195} 02/26/2022 07:35:27 - INFO - codeparrot_training - Step 41721: {'lr': 3.58121812154564e-05, 'samples': 21361664, 'steps': 41721, 'loss/train': 0.7967495322227478} 02/26/2022 07:35:33 - INFO - codeparrot_training - Step 41722: {'lr': 3.580374306468795e-05, 'samples': 21362176, 'steps': 41722, 'loss/train': 1.8414092063903809} 02/26/2022 07:35:36 - INFO - codeparrot_training - Step 41723: {'lr': 3.5795305831468086e-05, 'samples': 21362688, 'steps': 41723, 'loss/train': 1.4105950593948364} 02/26/2022 07:35:42 - INFO - codeparrot_training - Step 41724: {'lr': 3.578686951583321e-05, 'samples': 21363200, 'steps': 41724, 'loss/train': 0.7590596675872803} 02/26/2022 07:35:45 - INFO - codeparrot_training - Step 41725: {'lr': 3.5778434117819284e-05, 'samples': 21363712, 'steps': 41725, 'loss/train': 2.1910815238952637} 02/26/2022 07:35:51 - INFO - codeparrot_training - Step 41726: {'lr': 3.576999963746258e-05, 'samples': 21364224, 'steps': 41726, 'loss/train': 2.05096697807312} 02/26/2022 07:35:54 - INFO - codeparrot_training - Step 41727: {'lr': 3.576156607479919e-05, 'samples': 21364736, 'steps': 41727, 'loss/train': 2.4411516189575195} 02/26/2022 07:36:00 - INFO - codeparrot_training - Step 41728: {'lr': 3.575313342986522e-05, 'samples': 21365248, 'steps': 41728, 'loss/train': 1.9012272357940674} 02/26/2022 07:36:03 - INFO - codeparrot_training - Step 41729: {'lr': 3.574470170269672e-05, 'samples': 21365760, 'steps': 41729, 'loss/train': 1.4051686525344849} 02/26/2022 07:36:09 - INFO - codeparrot_training - Step 41730: {'lr': 3.5736270893329955e-05, 'samples': 21366272, 'steps': 41730, 'loss/train': 1.2901688814163208} 02/26/2022 07:36:12 - INFO - codeparrot_training - Step 41731: {'lr': 3.572784100180096e-05, 'samples': 21366784, 'steps': 41731, 'loss/train': 1.1237056255340576} 02/26/2022 07:36:18 - INFO - codeparrot_training - Step 41732: {'lr': 3.571941202814588e-05, 'samples': 21367296, 'steps': 41732, 'loss/train': 2.126596212387085} 02/26/2022 07:36:21 - INFO - codeparrot_training - Step 41733: {'lr': 3.571098397240072e-05, 'samples': 21367808, 'steps': 41733, 'loss/train': 1.5708905458450317} 02/26/2022 07:36:27 - INFO - codeparrot_training - Step 41734: {'lr': 3.570255683460172e-05, 'samples': 21368320, 'steps': 41734, 'loss/train': 1.101157307624817} 02/26/2022 07:36:30 - INFO - codeparrot_training - Step 41735: {'lr': 3.569413061478491e-05, 'samples': 21368832, 'steps': 41735, 'loss/train': 2.1775007247924805} 02/26/2022 07:36:36 - INFO - codeparrot_training - Step 41736: {'lr': 3.56857053129864e-05, 'samples': 21369344, 'steps': 41736, 'loss/train': 2.413322687149048} 02/26/2022 07:36:40 - INFO - codeparrot_training - Step 41737: {'lr': 3.5677280929242214e-05, 'samples': 21369856, 'steps': 41737, 'loss/train': 2.2208516597747803} 02/26/2022 07:36:45 - INFO - codeparrot_training - Step 41738: {'lr': 3.566885746358858e-05, 'samples': 21370368, 'steps': 41738, 'loss/train': 1.4122841358184814} 02/26/2022 07:36:49 - INFO - codeparrot_training - Step 41739: {'lr': 3.56604349160615e-05, 'samples': 21370880, 'steps': 41739, 'loss/train': 1.4783846139907837} 02/26/2022 07:36:54 - INFO - codeparrot_training - Step 41740: {'lr': 3.5652013286697036e-05, 'samples': 21371392, 'steps': 41740, 'loss/train': 1.8731080293655396} 02/26/2022 07:36:58 - INFO - codeparrot_training - Step 41741: {'lr': 3.564359257553129e-05, 'samples': 21371904, 'steps': 41741, 'loss/train': 2.161520481109619} 02/26/2022 07:37:03 - INFO - codeparrot_training - Step 41742: {'lr': 3.563517278260025e-05, 'samples': 21372416, 'steps': 41742, 'loss/train': 1.2099199295043945} 02/26/2022 07:37:07 - INFO - codeparrot_training - Step 41743: {'lr': 3.562675390794015e-05, 'samples': 21372928, 'steps': 41743, 'loss/train': 2.234713315963745} 02/26/2022 07:37:12 - INFO - codeparrot_training - Step 41744: {'lr': 3.561833595158698e-05, 'samples': 21373440, 'steps': 41744, 'loss/train': 1.2447516918182373} 02/26/2022 07:37:15 - INFO - codeparrot_training - Step 41745: {'lr': 3.560991891357676e-05, 'samples': 21373952, 'steps': 41745, 'loss/train': 2.3983891010284424} 02/26/2022 07:37:22 - INFO - codeparrot_training - Step 41746: {'lr': 3.56015027939455e-05, 'samples': 21374464, 'steps': 41746, 'loss/train': 3.4015278816223145} 02/26/2022 07:37:25 - INFO - codeparrot_training - Step 41747: {'lr': 3.559308759272939e-05, 'samples': 21374976, 'steps': 41747, 'loss/train': 1.4039788246154785} 02/26/2022 07:37:31 - INFO - codeparrot_training - Step 41748: {'lr': 3.558467330996443e-05, 'samples': 21375488, 'steps': 41748, 'loss/train': 1.170341968536377} 02/26/2022 07:37:34 - INFO - codeparrot_training - Step 41749: {'lr': 3.557625994568664e-05, 'samples': 21376000, 'steps': 41749, 'loss/train': 2.9517874717712402} 02/26/2022 07:37:40 - INFO - codeparrot_training - Step 41750: {'lr': 3.5567847499932e-05, 'samples': 21376512, 'steps': 41750, 'loss/train': 1.657469391822815} 02/26/2022 07:37:43 - INFO - codeparrot_training - Step 41751: {'lr': 3.555943597273667e-05, 'samples': 21377024, 'steps': 41751, 'loss/train': 1.6412954330444336} 02/26/2022 07:37:49 - INFO - codeparrot_training - Step 41752: {'lr': 3.5551025364136544e-05, 'samples': 21377536, 'steps': 41752, 'loss/train': 2.201308488845825} 02/26/2022 07:37:52 - INFO - codeparrot_training - Step 41753: {'lr': 3.554261567416786e-05, 'samples': 21378048, 'steps': 41753, 'loss/train': 2.0442068576812744} 02/26/2022 07:37:58 - INFO - codeparrot_training - Step 41754: {'lr': 3.5534206902866404e-05, 'samples': 21378560, 'steps': 41754, 'loss/train': 1.6880371570587158} 02/26/2022 07:38:01 - INFO - codeparrot_training - Step 41755: {'lr': 3.552579905026837e-05, 'samples': 21379072, 'steps': 41755, 'loss/train': 1.3798959255218506} 02/26/2022 07:38:07 - INFO - codeparrot_training - Step 41756: {'lr': 3.551739211640964e-05, 'samples': 21379584, 'steps': 41756, 'loss/train': 1.8287084102630615} 02/26/2022 07:38:11 - INFO - codeparrot_training - Step 41757: {'lr': 3.5508986101326415e-05, 'samples': 21380096, 'steps': 41757, 'loss/train': 1.1338351964950562} 02/26/2022 07:38:16 - INFO - codeparrot_training - Step 41758: {'lr': 3.550058100505449e-05, 'samples': 21380608, 'steps': 41758, 'loss/train': 0.45335105061531067} 02/26/2022 07:38:20 - INFO - codeparrot_training - Step 41759: {'lr': 3.549217682762998e-05, 'samples': 21381120, 'steps': 41759, 'loss/train': 2.351400852203369} 02/26/2022 07:38:25 - INFO - codeparrot_training - Step 41760: {'lr': 3.548377356908886e-05, 'samples': 21381632, 'steps': 41760, 'loss/train': 1.165387749671936} 02/26/2022 07:38:29 - INFO - codeparrot_training - Step 41761: {'lr': 3.54753712294672e-05, 'samples': 21382144, 'steps': 41761, 'loss/train': 1.49990713596344} 02/26/2022 07:38:34 - INFO - codeparrot_training - Step 41762: {'lr': 3.546696980880085e-05, 'samples': 21382656, 'steps': 41762, 'loss/train': 2.260895252227783} 02/26/2022 07:38:38 - INFO - codeparrot_training - Step 41763: {'lr': 3.545856930712593e-05, 'samples': 21383168, 'steps': 41763, 'loss/train': 1.9449162483215332} 02/26/2022 07:38:43 - INFO - codeparrot_training - Step 41764: {'lr': 3.5450169724478315e-05, 'samples': 21383680, 'steps': 41764, 'loss/train': 0.7249825596809387} 02/26/2022 07:38:47 - INFO - codeparrot_training - Step 41765: {'lr': 3.5441771060894104e-05, 'samples': 21384192, 'steps': 41765, 'loss/train': 1.3583165407180786} 02/26/2022 07:38:52 - INFO - codeparrot_training - Step 41766: {'lr': 3.543337331640922e-05, 'samples': 21384704, 'steps': 41766, 'loss/train': 1.5960205793380737} 02/26/2022 07:38:56 - INFO - codeparrot_training - Step 41767: {'lr': 3.5424976491059635e-05, 'samples': 21385216, 'steps': 41767, 'loss/train': 2.4699699878692627} 02/26/2022 07:39:02 - INFO - codeparrot_training - Step 41768: {'lr': 3.541658058488126e-05, 'samples': 21385728, 'steps': 41768, 'loss/train': 0.7365851402282715} 02/26/2022 07:39:05 - INFO - codeparrot_training - Step 41769: {'lr': 3.540818559791017e-05, 'samples': 21386240, 'steps': 41769, 'loss/train': 2.187325954437256} 02/26/2022 07:39:11 - INFO - codeparrot_training - Step 41770: {'lr': 3.539979153018228e-05, 'samples': 21386752, 'steps': 41770, 'loss/train': 2.2277119159698486} 02/26/2022 07:39:14 - INFO - codeparrot_training - Step 41771: {'lr': 3.53913983817335e-05, 'samples': 21387264, 'steps': 41771, 'loss/train': 1.6255269050598145} 02/26/2022 07:39:20 - INFO - codeparrot_training - Step 41772: {'lr': 3.538300615259979e-05, 'samples': 21387776, 'steps': 41772, 'loss/train': 7.131189823150635} 02/26/2022 07:39:24 - INFO - codeparrot_training - Step 41773: {'lr': 3.537461484281721e-05, 'samples': 21388288, 'steps': 41773, 'loss/train': 2.6381707191467285} 02/26/2022 07:39:29 - INFO - codeparrot_training - Step 41774: {'lr': 3.53662244524216e-05, 'samples': 21388800, 'steps': 41774, 'loss/train': 1.708308458328247} 02/26/2022 07:39:33 - INFO - codeparrot_training - Step 41775: {'lr': 3.535783498144895e-05, 'samples': 21389312, 'steps': 41775, 'loss/train': 2.2258992195129395} 02/26/2022 07:39:38 - INFO - codeparrot_training - Step 41776: {'lr': 3.534944642993512e-05, 'samples': 21389824, 'steps': 41776, 'loss/train': 1.870118260383606} 02/26/2022 07:39:42 - INFO - codeparrot_training - Step 41777: {'lr': 3.534105879791616e-05, 'samples': 21390336, 'steps': 41777, 'loss/train': 2.0420596599578857} 02/26/2022 07:39:47 - INFO - codeparrot_training - Step 41778: {'lr': 3.5332672085427926e-05, 'samples': 21390848, 'steps': 41778, 'loss/train': 0.742152988910675} 02/26/2022 07:39:51 - INFO - codeparrot_training - Step 41779: {'lr': 3.532428629250636e-05, 'samples': 21391360, 'steps': 41779, 'loss/train': 1.2266942262649536} 02/26/2022 07:39:56 - INFO - codeparrot_training - Step 41780: {'lr': 3.5315901419187364e-05, 'samples': 21391872, 'steps': 41780, 'loss/train': 1.7314984798431396} 02/26/2022 07:40:00 - INFO - codeparrot_training - Step 41781: {'lr': 3.530751746550689e-05, 'samples': 21392384, 'steps': 41781, 'loss/train': 1.7999262809753418} 02/26/2022 07:40:06 - INFO - codeparrot_training - Step 41782: {'lr': 3.529913443150087e-05, 'samples': 21392896, 'steps': 41782, 'loss/train': 1.9495538473129272} 02/26/2022 07:40:10 - INFO - codeparrot_training - Step 41783: {'lr': 3.529075231720516e-05, 'samples': 21393408, 'steps': 41783, 'loss/train': 1.362709403038025} 02/26/2022 07:40:15 - INFO - codeparrot_training - Step 41784: {'lr': 3.528237112265564e-05, 'samples': 21393920, 'steps': 41784, 'loss/train': 1.2696821689605713} 02/26/2022 07:40:19 - INFO - codeparrot_training - Step 41785: {'lr': 3.527399084788832e-05, 'samples': 21394432, 'steps': 41785, 'loss/train': 1.9738383293151855} 02/26/2022 07:40:24 - INFO - codeparrot_training - Step 41786: {'lr': 3.526561149293906e-05, 'samples': 21394944, 'steps': 41786, 'loss/train': 1.3871738910675049} 02/26/2022 07:40:28 - INFO - codeparrot_training - Step 41787: {'lr': 3.5257233057843645e-05, 'samples': 21395456, 'steps': 41787, 'loss/train': 1.9540824890136719} 02/26/2022 07:40:33 - INFO - codeparrot_training - Step 41788: {'lr': 3.5248855542638184e-05, 'samples': 21395968, 'steps': 41788, 'loss/train': 1.787583827972412} 02/26/2022 07:40:37 - INFO - codeparrot_training - Step 41789: {'lr': 3.5240478947358297e-05, 'samples': 21396480, 'steps': 41789, 'loss/train': 0.7123517394065857} 02/26/2022 07:40:42 - INFO - codeparrot_training - Step 41790: {'lr': 3.52321032720401e-05, 'samples': 21396992, 'steps': 41790, 'loss/train': 0.757858395576477} 02/26/2022 07:40:46 - INFO - codeparrot_training - Step 41791: {'lr': 3.5223728516719286e-05, 'samples': 21397504, 'steps': 41791, 'loss/train': 2.3760859966278076} 02/26/2022 07:40:51 - INFO - codeparrot_training - Step 41792: {'lr': 3.521535468143197e-05, 'samples': 21398016, 'steps': 41792, 'loss/train': 1.1649057865142822} 02/26/2022 07:40:55 - INFO - codeparrot_training - Step 41793: {'lr': 3.520698176621373e-05, 'samples': 21398528, 'steps': 41793, 'loss/train': 0.3401975631713867} 02/26/2022 07:41:01 - INFO - codeparrot_training - Step 41794: {'lr': 3.519860977110065e-05, 'samples': 21399040, 'steps': 41794, 'loss/train': 2.129559278488159} 02/26/2022 07:41:04 - INFO - codeparrot_training - Step 41795: {'lr': 3.519023869612845e-05, 'samples': 21399552, 'steps': 41795, 'loss/train': 1.4806514978408813} 02/26/2022 07:41:10 - INFO - codeparrot_training - Step 41796: {'lr': 3.518186854133315e-05, 'samples': 21400064, 'steps': 41796, 'loss/train': 1.207190990447998} 02/26/2022 07:41:13 - INFO - codeparrot_training - Step 41797: {'lr': 3.517349930675043e-05, 'samples': 21400576, 'steps': 41797, 'loss/train': 2.6549975872039795} 02/26/2022 07:41:19 - INFO - codeparrot_training - Step 41798: {'lr': 3.5165130992416275e-05, 'samples': 21401088, 'steps': 41798, 'loss/train': 2.072608470916748} 02/26/2022 07:41:22 - INFO - codeparrot_training - Step 41799: {'lr': 3.515676359836642e-05, 'samples': 21401600, 'steps': 41799, 'loss/train': 0.135090172290802} 02/26/2022 07:41:28 - INFO - codeparrot_training - Step 41800: {'lr': 3.514839712463683e-05, 'samples': 21402112, 'steps': 41800, 'loss/train': 2.0122764110565186} 02/26/2022 07:41:32 - INFO - codeparrot_training - Step 41801: {'lr': 3.5140031571263276e-05, 'samples': 21402624, 'steps': 41801, 'loss/train': 1.4339104890823364} 02/26/2022 07:41:37 - INFO - codeparrot_training - Step 41802: {'lr': 3.5131666938281626e-05, 'samples': 21403136, 'steps': 41802, 'loss/train': 0.9168344140052795} 02/26/2022 07:41:40 - INFO - codeparrot_training - Step 41803: {'lr': 3.51233032257276e-05, 'samples': 21403648, 'steps': 41803, 'loss/train': 1.9832981824874878} 02/26/2022 07:41:46 - INFO - codeparrot_training - Step 41804: {'lr': 3.51149404336372e-05, 'samples': 21404160, 'steps': 41804, 'loss/train': 2.1262660026550293} 02/26/2022 07:41:50 - INFO - codeparrot_training - Step 41805: {'lr': 3.510657856204613e-05, 'samples': 21404672, 'steps': 41805, 'loss/train': 2.0327484607696533} 02/26/2022 07:41:56 - INFO - codeparrot_training - Step 41806: {'lr': 3.5098217610990244e-05, 'samples': 21405184, 'steps': 41806, 'loss/train': 1.4959701299667358} 02/26/2022 07:42:01 - INFO - codeparrot_training - Step 41807: {'lr': 3.5089857580505316e-05, 'samples': 21405696, 'steps': 41807, 'loss/train': 0.739684522151947} 02/26/2022 07:42:05 - INFO - codeparrot_training - Step 41808: {'lr': 3.508149847062725e-05, 'samples': 21406208, 'steps': 41808, 'loss/train': 1.3491206169128418} 02/26/2022 07:42:11 - INFO - codeparrot_training - Step 41809: {'lr': 3.5073140281391776e-05, 'samples': 21406720, 'steps': 41809, 'loss/train': 1.8846805095672607} 02/26/2022 07:42:15 - INFO - codeparrot_training - Step 41810: {'lr': 3.506478301283475e-05, 'samples': 21407232, 'steps': 41810, 'loss/train': 2.0595109462738037} 02/26/2022 07:42:20 - INFO - codeparrot_training - Step 41811: {'lr': 3.505642666499187e-05, 'samples': 21407744, 'steps': 41811, 'loss/train': 1.2388486862182617} 02/26/2022 07:42:24 - INFO - codeparrot_training - Step 41812: {'lr': 3.504807123789908e-05, 'samples': 21408256, 'steps': 41812, 'loss/train': 1.553773045539856} 02/26/2022 07:42:29 - INFO - codeparrot_training - Step 41813: {'lr': 3.503971673159204e-05, 'samples': 21408768, 'steps': 41813, 'loss/train': 1.818275809288025} 02/26/2022 07:42:33 - INFO - codeparrot_training - Step 41814: {'lr': 3.5031363146106695e-05, 'samples': 21409280, 'steps': 41814, 'loss/train': 2.6383988857269287} 02/26/2022 07:42:38 - INFO - codeparrot_training - Step 41815: {'lr': 3.50230104814786e-05, 'samples': 21409792, 'steps': 41815, 'loss/train': 1.3958686590194702} 02/26/2022 07:42:42 - INFO - codeparrot_training - Step 41816: {'lr': 3.501465873774376e-05, 'samples': 21410304, 'steps': 41816, 'loss/train': 2.5607266426086426} 02/26/2022 07:42:47 - INFO - codeparrot_training - Step 41817: {'lr': 3.500630791493778e-05, 'samples': 21410816, 'steps': 41817, 'loss/train': 2.1913537979125977} 02/26/2022 07:42:51 - INFO - codeparrot_training - Step 41818: {'lr': 3.499795801309663e-05, 'samples': 21411328, 'steps': 41818, 'loss/train': 2.8457465171813965} 02/26/2022 07:42:57 - INFO - codeparrot_training - Step 41819: {'lr': 3.498960903225584e-05, 'samples': 21411840, 'steps': 41819, 'loss/train': 1.5430333614349365} 02/26/2022 07:43:01 - INFO - codeparrot_training - Step 41820: {'lr': 3.4981260972451354e-05, 'samples': 21412352, 'steps': 41820, 'loss/train': 3.1928465366363525} 02/26/2022 07:43:06 - INFO - codeparrot_training - Step 41821: {'lr': 3.4972913833718864e-05, 'samples': 21412864, 'steps': 41821, 'loss/train': 1.643866777420044} 02/26/2022 07:43:10 - INFO - codeparrot_training - Step 41822: {'lr': 3.496456761609412e-05, 'samples': 21413376, 'steps': 41822, 'loss/train': 2.3708415031433105} 02/26/2022 07:43:15 - INFO - codeparrot_training - Step 41823: {'lr': 3.495622231961285e-05, 'samples': 21413888, 'steps': 41823, 'loss/train': 0.815904974937439} 02/26/2022 07:43:19 - INFO - codeparrot_training - Step 41824: {'lr': 3.494787794431087e-05, 'samples': 21414400, 'steps': 41824, 'loss/train': 1.3587661981582642} 02/26/2022 07:43:25 - INFO - codeparrot_training - Step 41825: {'lr': 3.493953449022391e-05, 'samples': 21414912, 'steps': 41825, 'loss/train': 2.3958585262298584} 02/26/2022 07:43:28 - INFO - codeparrot_training - Step 41826: {'lr': 3.493119195738764e-05, 'samples': 21415424, 'steps': 41826, 'loss/train': 2.8388452529907227} 02/26/2022 07:43:34 - INFO - codeparrot_training - Step 41827: {'lr': 3.492285034583792e-05, 'samples': 21415936, 'steps': 41827, 'loss/train': 1.6827870607376099} 02/26/2022 07:43:37 - INFO - codeparrot_training - Step 41828: {'lr': 3.491450965561041e-05, 'samples': 21416448, 'steps': 41828, 'loss/train': 1.5402657985687256} 02/26/2022 07:43:44 - INFO - codeparrot_training - Step 41829: {'lr': 3.490616988674083e-05, 'samples': 21416960, 'steps': 41829, 'loss/train': 1.8274123668670654} 02/26/2022 07:43:47 - INFO - codeparrot_training - Step 41830: {'lr': 3.489783103926486e-05, 'samples': 21417472, 'steps': 41830, 'loss/train': 1.3994922637939453} 02/26/2022 07:43:53 - INFO - codeparrot_training - Step 41831: {'lr': 3.488949311321835e-05, 'samples': 21417984, 'steps': 41831, 'loss/train': 1.6066696643829346} 02/26/2022 07:43:56 - INFO - codeparrot_training - Step 41832: {'lr': 3.488115610863693e-05, 'samples': 21418496, 'steps': 41832, 'loss/train': 1.883328914642334} 02/26/2022 07:44:00 - INFO - codeparrot_training - Step 41833: {'lr': 3.4872820025556364e-05, 'samples': 21419008, 'steps': 41833, 'loss/train': 1.6999083757400513} 02/26/2022 07:44:05 - INFO - codeparrot_training - Step 41834: {'lr': 3.4864484864012225e-05, 'samples': 21419520, 'steps': 41834, 'loss/train': 2.2516164779663086} 02/26/2022 07:44:09 - INFO - codeparrot_training - Step 41835: {'lr': 3.485615062404041e-05, 'samples': 21420032, 'steps': 41835, 'loss/train': 1.5778331756591797} 02/26/2022 07:44:14 - INFO - codeparrot_training - Step 41836: {'lr': 3.484781730567651e-05, 'samples': 21420544, 'steps': 41836, 'loss/train': 2.472184181213379} 02/26/2022 07:44:18 - INFO - codeparrot_training - Step 41837: {'lr': 3.483948490895625e-05, 'samples': 21421056, 'steps': 41837, 'loss/train': 1.3338651657104492} 02/26/2022 07:44:23 - INFO - codeparrot_training - Step 41838: {'lr': 3.483115343391527e-05, 'samples': 21421568, 'steps': 41838, 'loss/train': 1.1206046342849731} 02/26/2022 07:44:27 - INFO - codeparrot_training - Step 41839: {'lr': 3.4822822880589325e-05, 'samples': 21422080, 'steps': 41839, 'loss/train': 1.0529612302780151} 02/26/2022 07:44:33 - INFO - codeparrot_training - Step 41840: {'lr': 3.481449324901412e-05, 'samples': 21422592, 'steps': 41840, 'loss/train': 1.7585954666137695} 02/26/2022 07:44:37 - INFO - codeparrot_training - Step 41841: {'lr': 3.4806164539225274e-05, 'samples': 21423104, 'steps': 41841, 'loss/train': 1.643898367881775} 02/26/2022 07:44:42 - INFO - codeparrot_training - Step 41842: {'lr': 3.479783675125844e-05, 'samples': 21423616, 'steps': 41842, 'loss/train': 1.2547779083251953} 02/26/2022 07:44:46 - INFO - codeparrot_training - Step 41843: {'lr': 3.47895098851494e-05, 'samples': 21424128, 'steps': 41843, 'loss/train': 0.6766678094863892} 02/26/2022 07:44:51 - INFO - codeparrot_training - Step 41844: {'lr': 3.4781183940933744e-05, 'samples': 21424640, 'steps': 41844, 'loss/train': 1.009781002998352} 02/26/2022 07:44:55 - INFO - codeparrot_training - Step 41845: {'lr': 3.4772858918647176e-05, 'samples': 21425152, 'steps': 41845, 'loss/train': 1.7243602275848389} 02/26/2022 07:45:00 - INFO - codeparrot_training - Step 41846: {'lr': 3.476453481832528e-05, 'samples': 21425664, 'steps': 41846, 'loss/train': 1.997341275215149} 02/26/2022 07:45:04 - INFO - codeparrot_training - Step 41847: {'lr': 3.475621164000381e-05, 'samples': 21426176, 'steps': 41847, 'loss/train': 0.5192564129829407} 02/26/2022 07:45:09 - INFO - codeparrot_training - Step 41848: {'lr': 3.474788938371834e-05, 'samples': 21426688, 'steps': 41848, 'loss/train': 1.4176605939865112} 02/26/2022 07:45:13 - INFO - codeparrot_training - Step 41849: {'lr': 3.473956804950468e-05, 'samples': 21427200, 'steps': 41849, 'loss/train': 2.113344192504883} 02/26/2022 07:45:20 - INFO - codeparrot_training - Step 41850: {'lr': 3.473124763739824e-05, 'samples': 21427712, 'steps': 41850, 'loss/train': 2.7613272666931152} 02/26/2022 07:45:23 - INFO - codeparrot_training - Step 41851: {'lr': 3.472292814743483e-05, 'samples': 21428224, 'steps': 41851, 'loss/train': 1.2220218181610107} 02/26/2022 07:45:28 - INFO - codeparrot_training - Step 41852: {'lr': 3.4714609579649975e-05, 'samples': 21428736, 'steps': 41852, 'loss/train': 2.3839950561523438} 02/26/2022 07:45:32 - INFO - codeparrot_training - Step 41853: {'lr': 3.470629193407948e-05, 'samples': 21429248, 'steps': 41853, 'loss/train': 2.0858848094940186} 02/26/2022 07:45:38 - INFO - codeparrot_training - Step 41854: {'lr': 3.469797521075879e-05, 'samples': 21429760, 'steps': 41854, 'loss/train': 1.7080988883972168} 02/26/2022 07:45:41 - INFO - codeparrot_training - Step 41855: {'lr': 3.468965940972363e-05, 'samples': 21430272, 'steps': 41855, 'loss/train': 1.432517170906067} 02/26/2022 07:45:47 - INFO - codeparrot_training - Step 41856: {'lr': 3.468134453100957e-05, 'samples': 21430784, 'steps': 41856, 'loss/train': 0.24918366968631744} 02/26/2022 07:45:51 - INFO - codeparrot_training - Step 41857: {'lr': 3.4673030574652346e-05, 'samples': 21431296, 'steps': 41857, 'loss/train': 2.2553653717041016} 02/26/2022 07:45:54 - INFO - codeparrot_training - Step 41858: {'lr': 3.4664717540687395e-05, 'samples': 21431808, 'steps': 41858, 'loss/train': 1.7428793907165527} 02/26/2022 07:46:00 - INFO - codeparrot_training - Step 41859: {'lr': 3.465640542915044e-05, 'samples': 21432320, 'steps': 41859, 'loss/train': 2.9964118003845215} 02/26/2022 07:46:05 - INFO - codeparrot_training - Step 41860: {'lr': 3.464809424007703e-05, 'samples': 21432832, 'steps': 41860, 'loss/train': 2.4415841102600098} 02/26/2022 07:46:09 - INFO - codeparrot_training - Step 41861: {'lr': 3.463978397350287e-05, 'samples': 21433344, 'steps': 41861, 'loss/train': 1.4502977132797241} 02/26/2022 07:46:12 - INFO - codeparrot_training - Step 41862: {'lr': 3.463147462946348e-05, 'samples': 21433856, 'steps': 41862, 'loss/train': 1.4036734104156494} 02/26/2022 07:46:18 - INFO - codeparrot_training - Step 41863: {'lr': 3.462316620799447e-05, 'samples': 21434368, 'steps': 41863, 'loss/train': 1.6313860416412354} 02/26/2022 07:46:21 - INFO - codeparrot_training - Step 41864: {'lr': 3.461485870913137e-05, 'samples': 21434880, 'steps': 41864, 'loss/train': 2.0481882095336914} 02/26/2022 07:46:27 - INFO - codeparrot_training - Step 41865: {'lr': 3.460655213290986e-05, 'samples': 21435392, 'steps': 41865, 'loss/train': 1.6504580974578857} 02/26/2022 07:46:30 - INFO - codeparrot_training - Step 41866: {'lr': 3.459824647936552e-05, 'samples': 21435904, 'steps': 41866, 'loss/train': 1.0296905040740967} 02/26/2022 07:46:37 - INFO - codeparrot_training - Step 41867: {'lr': 3.458994174853386e-05, 'samples': 21436416, 'steps': 41867, 'loss/train': 2.121839761734009} 02/26/2022 07:46:41 - INFO - codeparrot_training - Step 41868: {'lr': 3.4581637940450465e-05, 'samples': 21436928, 'steps': 41868, 'loss/train': 2.0996737480163574} 02/26/2022 07:46:46 - INFO - codeparrot_training - Step 41869: {'lr': 3.457333505515095e-05, 'samples': 21437440, 'steps': 41869, 'loss/train': 1.441436767578125} 02/26/2022 07:46:50 - INFO - codeparrot_training - Step 41870: {'lr': 3.4565033092670895e-05, 'samples': 21437952, 'steps': 41870, 'loss/train': 0.4527837932109833} 02/26/2022 07:46:55 - INFO - codeparrot_training - Step 41871: {'lr': 3.45567320530458e-05, 'samples': 21438464, 'steps': 41871, 'loss/train': 1.5662294626235962} 02/26/2022 07:47:01 - INFO - codeparrot_training - Step 41872: {'lr': 3.454843193631127e-05, 'samples': 21438976, 'steps': 41872, 'loss/train': 2.2271969318389893} 02/26/2022 07:47:04 - INFO - codeparrot_training - Step 41873: {'lr': 3.4540132742502785e-05, 'samples': 21439488, 'steps': 41873, 'loss/train': 2.4509475231170654} 02/26/2022 07:47:10 - INFO - codeparrot_training - Step 41874: {'lr': 3.453183447165603e-05, 'samples': 21440000, 'steps': 41874, 'loss/train': 1.315755844116211} 02/26/2022 07:47:13 - INFO - codeparrot_training - Step 41875: {'lr': 3.452353712380646e-05, 'samples': 21440512, 'steps': 41875, 'loss/train': 2.3272464275360107} 02/26/2022 07:47:19 - INFO - codeparrot_training - Step 41876: {'lr': 3.451524069898962e-05, 'samples': 21441024, 'steps': 41876, 'loss/train': 1.9262360334396362} 02/26/2022 07:47:22 - INFO - codeparrot_training - Step 41877: {'lr': 3.4506945197241044e-05, 'samples': 21441536, 'steps': 41877, 'loss/train': 0.9264810681343079} 02/26/2022 07:47:29 - INFO - codeparrot_training - Step 41878: {'lr': 3.449865061859633e-05, 'samples': 21442048, 'steps': 41878, 'loss/train': 2.1967406272888184} 02/26/2022 07:47:33 - INFO - codeparrot_training - Step 41879: {'lr': 3.449035696309094e-05, 'samples': 21442560, 'steps': 41879, 'loss/train': 3.3565893173217773} 02/26/2022 07:47:38 - INFO - codeparrot_training - Step 41880: {'lr': 3.4482064230760474e-05, 'samples': 21443072, 'steps': 41880, 'loss/train': 0.9036532044410706} 02/26/2022 07:47:42 - INFO - codeparrot_training - Step 41881: {'lr': 3.447377242164032e-05, 'samples': 21443584, 'steps': 41881, 'loss/train': 1.2806609869003296} 02/26/2022 07:47:47 - INFO - codeparrot_training - Step 41882: {'lr': 3.4465481535766146e-05, 'samples': 21444096, 'steps': 41882, 'loss/train': 1.8677949905395508} 02/26/2022 07:47:51 - INFO - codeparrot_training - Step 41883: {'lr': 3.445719157317343e-05, 'samples': 21444608, 'steps': 41883, 'loss/train': 0.9789207577705383} 02/26/2022 07:47:56 - INFO - codeparrot_training - Step 41884: {'lr': 3.4448902533897654e-05, 'samples': 21445120, 'steps': 41884, 'loss/train': 1.8151228427886963} 02/26/2022 07:48:00 - INFO - codeparrot_training - Step 41885: {'lr': 3.444061441797428e-05, 'samples': 21445632, 'steps': 41885, 'loss/train': 1.6921089887619019} 02/26/2022 07:48:06 - INFO - codeparrot_training - Step 41886: {'lr': 3.44323272254389e-05, 'samples': 21446144, 'steps': 41886, 'loss/train': 2.488142728805542} 02/26/2022 07:48:10 - INFO - codeparrot_training - Step 41887: {'lr': 3.4424040956326954e-05, 'samples': 21446656, 'steps': 41887, 'loss/train': 1.718224048614502} 02/26/2022 07:48:15 - INFO - codeparrot_training - Step 41888: {'lr': 3.441575561067406e-05, 'samples': 21447168, 'steps': 41888, 'loss/train': 0.5454226136207581} 02/26/2022 07:48:19 - INFO - codeparrot_training - Step 41889: {'lr': 3.44074711885155e-05, 'samples': 21447680, 'steps': 41889, 'loss/train': 1.5699777603149414} 02/26/2022 07:48:24 - INFO - codeparrot_training - Step 41890: {'lr': 3.439918768988695e-05, 'samples': 21448192, 'steps': 41890, 'loss/train': 2.2480530738830566} 02/26/2022 07:48:28 - INFO - codeparrot_training - Step 41891: {'lr': 3.4390905114823736e-05, 'samples': 21448704, 'steps': 41891, 'loss/train': 1.9418888092041016} 02/26/2022 07:48:33 - INFO - codeparrot_training - Step 41892: {'lr': 3.438262346336155e-05, 'samples': 21449216, 'steps': 41892, 'loss/train': 1.5452089309692383} 02/26/2022 07:48:37 - INFO - codeparrot_training - Step 41893: {'lr': 3.437434273553561e-05, 'samples': 21449728, 'steps': 41893, 'loss/train': 1.7702889442443848} 02/26/2022 07:48:42 - INFO - codeparrot_training - Step 41894: {'lr': 3.4366062931381585e-05, 'samples': 21450240, 'steps': 41894, 'loss/train': 1.9535242319107056} 02/26/2022 07:48:45 - INFO - codeparrot_training - Step 41895: {'lr': 3.4357784050934816e-05, 'samples': 21450752, 'steps': 41895, 'loss/train': 2.064274787902832} 02/26/2022 07:48:51 - INFO - codeparrot_training - Step 41896: {'lr': 3.43495060942309e-05, 'samples': 21451264, 'steps': 41896, 'loss/train': 1.08790922164917} 02/26/2022 07:48:55 - INFO - codeparrot_training - Step 41897: {'lr': 3.434122906130521e-05, 'samples': 21451776, 'steps': 41897, 'loss/train': 1.9326579570770264} 02/26/2022 07:49:01 - INFO - codeparrot_training - Step 41898: {'lr': 3.433295295219321e-05, 'samples': 21452288, 'steps': 41898, 'loss/train': 2.16262149810791} 02/26/2022 07:49:05 - INFO - codeparrot_training - Step 41899: {'lr': 3.432467776693029e-05, 'samples': 21452800, 'steps': 41899, 'loss/train': 1.4047818183898926} 02/26/2022 07:49:10 - INFO - codeparrot_training - Step 41900: {'lr': 3.431640350555204e-05, 'samples': 21453312, 'steps': 41900, 'loss/train': 2.2607452869415283} 02/26/2022 07:49:14 - INFO - codeparrot_training - Step 41901: {'lr': 3.430813016809384e-05, 'samples': 21453824, 'steps': 41901, 'loss/train': 0.7123552560806274} 02/26/2022 07:49:19 - INFO - codeparrot_training - Step 41902: {'lr': 3.42998577545911e-05, 'samples': 21454336, 'steps': 41902, 'loss/train': 1.88678777217865} 02/26/2022 07:49:23 - INFO - codeparrot_training - Step 41903: {'lr': 3.429158626507922e-05, 'samples': 21454848, 'steps': 41903, 'loss/train': 1.1145275831222534} 02/26/2022 07:49:28 - INFO - codeparrot_training - Step 41904: {'lr': 3.4283315699593756e-05, 'samples': 21455360, 'steps': 41904, 'loss/train': 1.7459561824798584} 02/26/2022 07:49:32 - INFO - codeparrot_training - Step 41905: {'lr': 3.4275046058170054e-05, 'samples': 21455872, 'steps': 41905, 'loss/train': 1.6811538934707642} 02/26/2022 07:49:37 - INFO - codeparrot_training - Step 41906: {'lr': 3.426677734084358e-05, 'samples': 21456384, 'steps': 41906, 'loss/train': 1.0550471544265747} 02/26/2022 07:49:41 - INFO - codeparrot_training - Step 41907: {'lr': 3.425850954764967e-05, 'samples': 21456896, 'steps': 41907, 'loss/train': 1.8260973691940308} 02/26/2022 07:49:46 - INFO - codeparrot_training - Step 41908: {'lr': 3.425024267862384e-05, 'samples': 21457408, 'steps': 41908, 'loss/train': 1.8912158012390137} 02/26/2022 07:49:50 - INFO - codeparrot_training - Step 41909: {'lr': 3.424197673380147e-05, 'samples': 21457920, 'steps': 41909, 'loss/train': 2.2257046699523926} 02/26/2022 07:49:55 - INFO - codeparrot_training - Step 41910: {'lr': 3.423371171321796e-05, 'samples': 21458432, 'steps': 41910, 'loss/train': 2.71388578414917} 02/26/2022 07:49:59 - INFO - codeparrot_training - Step 41911: {'lr': 3.422544761690866e-05, 'samples': 21458944, 'steps': 41911, 'loss/train': 2.782745838165283} 02/26/2022 07:50:06 - INFO - codeparrot_training - Step 41912: {'lr': 3.421718444490907e-05, 'samples': 21459456, 'steps': 41912, 'loss/train': 2.449758291244507} 02/26/2022 07:50:09 - INFO - codeparrot_training - Step 41913: {'lr': 3.420892219725455e-05, 'samples': 21459968, 'steps': 41913, 'loss/train': 0.5024000406265259} 02/26/2022 07:50:15 - INFO - codeparrot_training - Step 41914: {'lr': 3.420066087398049e-05, 'samples': 21460480, 'steps': 41914, 'loss/train': 2.586467742919922} 02/26/2022 07:50:18 - INFO - codeparrot_training - Step 41915: {'lr': 3.419240047512223e-05, 'samples': 21460992, 'steps': 41915, 'loss/train': 1.3416094779968262} 02/26/2022 07:50:24 - INFO - codeparrot_training - Step 41916: {'lr': 3.418414100071524e-05, 'samples': 21461504, 'steps': 41916, 'loss/train': 1.6230154037475586} 02/26/2022 07:50:27 - INFO - codeparrot_training - Step 41917: {'lr': 3.417588245079487e-05, 'samples': 21462016, 'steps': 41917, 'loss/train': 2.0060741901397705} 02/26/2022 07:50:33 - INFO - codeparrot_training - Step 41918: {'lr': 3.416762482539648e-05, 'samples': 21462528, 'steps': 41918, 'loss/train': 1.5923959016799927} 02/26/2022 07:50:36 - INFO - codeparrot_training - Step 41919: {'lr': 3.4159368124555446e-05, 'samples': 21463040, 'steps': 41919, 'loss/train': 0.3508755564689636} 02/26/2022 07:50:42 - INFO - codeparrot_training - Step 41920: {'lr': 3.415111234830709e-05, 'samples': 21463552, 'steps': 41920, 'loss/train': 1.2949495315551758} 02/26/2022 07:50:45 - INFO - codeparrot_training - Step 41921: {'lr': 3.414285749668688e-05, 'samples': 21464064, 'steps': 41921, 'loss/train': 1.8322685956954956} 02/26/2022 07:50:52 - INFO - codeparrot_training - Step 41922: {'lr': 3.413460356973008e-05, 'samples': 21464576, 'steps': 41922, 'loss/train': 1.1594280004501343} 02/26/2022 07:50:56 - INFO - codeparrot_training - Step 41923: {'lr': 3.4126350567472205e-05, 'samples': 21465088, 'steps': 41923, 'loss/train': 1.1350784301757812} 02/26/2022 07:51:01 - INFO - codeparrot_training - Step 41924: {'lr': 3.41180984899484e-05, 'samples': 21465600, 'steps': 41924, 'loss/train': 0.9808856844902039} 02/26/2022 07:51:05 - INFO - codeparrot_training - Step 41925: {'lr': 3.410984733719413e-05, 'samples': 21466112, 'steps': 41925, 'loss/train': 1.1718684434890747} 02/26/2022 07:51:10 - INFO - codeparrot_training - Step 41926: {'lr': 3.410159710924468e-05, 'samples': 21466624, 'steps': 41926, 'loss/train': 1.5446263551712036} 02/26/2022 07:51:14 - INFO - codeparrot_training - Step 41927: {'lr': 3.409334780613557e-05, 'samples': 21467136, 'steps': 41927, 'loss/train': 1.1448591947555542} 02/26/2022 07:51:19 - INFO - codeparrot_training - Step 41928: {'lr': 3.408509942790186e-05, 'samples': 21467648, 'steps': 41928, 'loss/train': 1.2616558074951172} 02/26/2022 07:51:23 - INFO - codeparrot_training - Step 41929: {'lr': 3.407685197457908e-05, 'samples': 21468160, 'steps': 41929, 'loss/train': 1.2948966026306152} 02/26/2022 07:51:28 - INFO - codeparrot_training - Step 41930: {'lr': 3.4068605446202464e-05, 'samples': 21468672, 'steps': 41930, 'loss/train': 1.7025885581970215} 02/26/2022 07:51:32 - INFO - codeparrot_training - Step 41931: {'lr': 3.406035984280748e-05, 'samples': 21469184, 'steps': 41931, 'loss/train': 2.1909499168395996} 02/26/2022 07:51:37 - INFO - codeparrot_training - Step 41932: {'lr': 3.4052115164429214e-05, 'samples': 21469696, 'steps': 41932, 'loss/train': 1.654496192932129} 02/26/2022 07:51:41 - INFO - codeparrot_training - Step 41933: {'lr': 3.4043871411103214e-05, 'samples': 21470208, 'steps': 41933, 'loss/train': 1.3355154991149902} 02/26/2022 07:51:48 - INFO - codeparrot_training - Step 41934: {'lr': 3.40356285828646e-05, 'samples': 21470720, 'steps': 41934, 'loss/train': 1.9920549392700195} 02/26/2022 07:51:51 - INFO - codeparrot_training - Step 41935: {'lr': 3.402738667974883e-05, 'samples': 21471232, 'steps': 41935, 'loss/train': 1.2140552997589111} 02/26/2022 07:51:57 - INFO - codeparrot_training - Step 41936: {'lr': 3.401914570179118e-05, 'samples': 21471744, 'steps': 41936, 'loss/train': 1.9474793672561646} 02/26/2022 07:52:00 - INFO - codeparrot_training - Step 41937: {'lr': 3.4010905649026924e-05, 'samples': 21472256, 'steps': 41937, 'loss/train': 1.204066514968872} 02/26/2022 07:52:06 - INFO - codeparrot_training - Step 41938: {'lr': 3.400266652149131e-05, 'samples': 21472768, 'steps': 41938, 'loss/train': 1.4890319108963013} 02/26/2022 07:52:09 - INFO - codeparrot_training - Step 41939: {'lr': 3.3994428319219726e-05, 'samples': 21473280, 'steps': 41939, 'loss/train': 1.913551926612854} 02/26/2022 07:52:15 - INFO - codeparrot_training - Step 41940: {'lr': 3.3986191042247454e-05, 'samples': 21473792, 'steps': 41940, 'loss/train': 1.6753807067871094} 02/26/2022 07:52:18 - INFO - codeparrot_training - Step 41941: {'lr': 3.397795469060971e-05, 'samples': 21474304, 'steps': 41941, 'loss/train': 2.238720178604126} 02/26/2022 07:52:24 - INFO - codeparrot_training - Step 41942: {'lr': 3.396971926434178e-05, 'samples': 21474816, 'steps': 41942, 'loss/train': 1.5360300540924072} 02/26/2022 07:52:27 - INFO - codeparrot_training - Step 41943: {'lr': 3.3961484763479015e-05, 'samples': 21475328, 'steps': 41943, 'loss/train': 1.2779909372329712} 02/26/2022 07:52:34 - INFO - codeparrot_training - Step 41944: {'lr': 3.395325118805661e-05, 'samples': 21475840, 'steps': 41944, 'loss/train': 1.166858434677124} 02/26/2022 07:52:37 - INFO - codeparrot_training - Step 41945: {'lr': 3.394501853810999e-05, 'samples': 21476352, 'steps': 41945, 'loss/train': 1.1693891286849976} 02/26/2022 07:52:42 - INFO - codeparrot_training - Step 41946: {'lr': 3.393678681367418e-05, 'samples': 21476864, 'steps': 41946, 'loss/train': 0.8781642913818359} 02/26/2022 07:52:46 - INFO - codeparrot_training - Step 41947: {'lr': 3.392855601478462e-05, 'samples': 21477376, 'steps': 41947, 'loss/train': 1.5784435272216797} 02/26/2022 07:52:51 - INFO - codeparrot_training - Step 41948: {'lr': 3.392032614147647e-05, 'samples': 21477888, 'steps': 41948, 'loss/train': 1.5444133281707764} 02/26/2022 07:52:55 - INFO - codeparrot_training - Step 41949: {'lr': 3.391209719378513e-05, 'samples': 21478400, 'steps': 41949, 'loss/train': 1.3093897104263306} 02/26/2022 07:53:00 - INFO - codeparrot_training - Step 41950: {'lr': 3.390386917174565e-05, 'samples': 21478912, 'steps': 41950, 'loss/train': 1.2384485006332397} 02/26/2022 07:53:04 - INFO - codeparrot_training - Step 41951: {'lr': 3.3895642075393416e-05, 'samples': 21479424, 'steps': 41951, 'loss/train': 2.4142026901245117} 02/26/2022 07:53:09 - INFO - codeparrot_training - Step 41952: {'lr': 3.3887415904763655e-05, 'samples': 21479936, 'steps': 41952, 'loss/train': 1.362563133239746} 02/26/2022 07:53:13 - INFO - codeparrot_training - Step 41953: {'lr': 3.387919065989156e-05, 'samples': 21480448, 'steps': 41953, 'loss/train': 1.158187985420227} 02/26/2022 07:53:19 - INFO - codeparrot_training - Step 41954: {'lr': 3.3870966340812355e-05, 'samples': 21480960, 'steps': 41954, 'loss/train': 1.6217172145843506} 02/26/2022 07:53:23 - INFO - codeparrot_training - Step 41955: {'lr': 3.386274294756134e-05, 'samples': 21481472, 'steps': 41955, 'loss/train': 1.2747089862823486} 02/26/2022 07:53:29 - INFO - codeparrot_training - Step 41956: {'lr': 3.385452048017371e-05, 'samples': 21481984, 'steps': 41956, 'loss/train': 1.8110721111297607} 02/26/2022 07:53:32 - INFO - codeparrot_training - Step 41957: {'lr': 3.38462989386846e-05, 'samples': 21482496, 'steps': 41957, 'loss/train': 1.302634358406067} 02/26/2022 07:53:38 - INFO - codeparrot_training - Step 41958: {'lr': 3.383807832312938e-05, 'samples': 21483008, 'steps': 41958, 'loss/train': 1.5550804138183594} 02/26/2022 07:53:41 - INFO - codeparrot_training - Step 41959: {'lr': 3.382985863354321e-05, 'samples': 21483520, 'steps': 41959, 'loss/train': 1.358167290687561} 02/26/2022 07:53:47 - INFO - codeparrot_training - Step 41960: {'lr': 3.3821639869961257e-05, 'samples': 21484032, 'steps': 41960, 'loss/train': 1.4872336387634277} 02/26/2022 07:53:50 - INFO - codeparrot_training - Step 41961: {'lr': 3.381342203241872e-05, 'samples': 21484544, 'steps': 41961, 'loss/train': 1.9987504482269287} 02/26/2022 07:53:56 - INFO - codeparrot_training - Step 41962: {'lr': 3.380520512095086e-05, 'samples': 21485056, 'steps': 41962, 'loss/train': 2.2526466846466064} 02/26/2022 07:53:59 - INFO - codeparrot_training - Step 41963: {'lr': 3.37969891355929e-05, 'samples': 21485568, 'steps': 41963, 'loss/train': 2.975898265838623} 02/26/2022 07:54:05 - INFO - codeparrot_training - Step 41964: {'lr': 3.3788774076379956e-05, 'samples': 21486080, 'steps': 41964, 'loss/train': 1.1899324655532837} 02/26/2022 07:54:08 - INFO - codeparrot_training - Step 41965: {'lr': 3.3780559943347194e-05, 'samples': 21486592, 'steps': 41965, 'loss/train': 0.22895056009292603} 02/26/2022 07:54:14 - INFO - codeparrot_training - Step 41966: {'lr': 3.3772346736529985e-05, 'samples': 21487104, 'steps': 41966, 'loss/train': 0.9826042652130127} 02/26/2022 07:54:17 - INFO - codeparrot_training - Step 41967: {'lr': 3.376413445596324e-05, 'samples': 21487616, 'steps': 41967, 'loss/train': 1.99070143699646} 02/26/2022 07:54:23 - INFO - codeparrot_training - Step 41968: {'lr': 3.375592310168235e-05, 'samples': 21488128, 'steps': 41968, 'loss/train': 2.310971736907959} 02/26/2022 07:54:26 - INFO - codeparrot_training - Step 41969: {'lr': 3.374771267372237e-05, 'samples': 21488640, 'steps': 41969, 'loss/train': 1.9975749254226685} 02/26/2022 07:54:33 - INFO - codeparrot_training - Step 41970: {'lr': 3.373950317211857e-05, 'samples': 21489152, 'steps': 41970, 'loss/train': 0.6903871297836304} 02/26/2022 07:54:36 - INFO - codeparrot_training - Step 41971: {'lr': 3.37312945969061e-05, 'samples': 21489664, 'steps': 41971, 'loss/train': 1.5590122938156128} 02/26/2022 07:54:42 - INFO - codeparrot_training - Step 41972: {'lr': 3.3723086948120066e-05, 'samples': 21490176, 'steps': 41972, 'loss/train': 1.0656887292861938} 02/26/2022 07:54:45 - INFO - codeparrot_training - Step 41973: {'lr': 3.371488022579558e-05, 'samples': 21490688, 'steps': 41973, 'loss/train': 0.576991856098175} 02/26/2022 07:54:51 - INFO - codeparrot_training - Step 41974: {'lr': 3.370667442996794e-05, 'samples': 21491200, 'steps': 41974, 'loss/train': 0.9533164501190186} 02/26/2022 07:54:55 - INFO - codeparrot_training - Step 41975: {'lr': 3.369846956067224e-05, 'samples': 21491712, 'steps': 41975, 'loss/train': 1.4333529472351074} 02/26/2022 07:55:00 - INFO - codeparrot_training - Step 41976: {'lr': 3.3690265617943585e-05, 'samples': 21492224, 'steps': 41976, 'loss/train': 1.68449866771698} 02/26/2022 07:55:03 - INFO - codeparrot_training - Step 41977: {'lr': 3.368206260181711e-05, 'samples': 21492736, 'steps': 41977, 'loss/train': 1.1230305433273315} 02/26/2022 07:55:09 - INFO - codeparrot_training - Step 41978: {'lr': 3.367386051232804e-05, 'samples': 21493248, 'steps': 41978, 'loss/train': 0.6900342702865601} 02/26/2022 07:55:12 - INFO - codeparrot_training - Step 41979: {'lr': 3.366565934951146e-05, 'samples': 21493760, 'steps': 41979, 'loss/train': 1.8926395177841187} 02/26/2022 07:55:19 - INFO - codeparrot_training - Step 41980: {'lr': 3.365745911340248e-05, 'samples': 21494272, 'steps': 41980, 'loss/train': 1.38222336769104} 02/26/2022 07:55:22 - INFO - codeparrot_training - Step 41981: {'lr': 3.364925980403621e-05, 'samples': 21494784, 'steps': 41981, 'loss/train': 1.3039599657058716} 02/26/2022 07:55:28 - INFO - codeparrot_training - Step 41982: {'lr': 3.3641061421447876e-05, 'samples': 21495296, 'steps': 41982, 'loss/train': 1.8142094612121582} 02/26/2022 07:55:31 - INFO - codeparrot_training - Step 41983: {'lr': 3.3632863965672475e-05, 'samples': 21495808, 'steps': 41983, 'loss/train': 1.3767694234848022} 02/26/2022 07:55:37 - INFO - codeparrot_training - Step 41984: {'lr': 3.3624667436745305e-05, 'samples': 21496320, 'steps': 41984, 'loss/train': 3.4410488605499268} 02/26/2022 07:55:41 - INFO - codeparrot_training - Step 41985: {'lr': 3.3616471834701204e-05, 'samples': 21496832, 'steps': 41985, 'loss/train': 5.851458549499512} 02/26/2022 07:55:46 - INFO - codeparrot_training - Step 41986: {'lr': 3.360827715957551e-05, 'samples': 21497344, 'steps': 41986, 'loss/train': 1.6347200870513916} 02/26/2022 07:55:50 - INFO - codeparrot_training - Step 41987: {'lr': 3.3600083411403195e-05, 'samples': 21497856, 'steps': 41987, 'loss/train': 1.8506697416305542} 02/26/2022 07:55:55 - INFO - codeparrot_training - Step 41988: {'lr': 3.3591890590219495e-05, 'samples': 21498368, 'steps': 41988, 'loss/train': 1.601239800453186} 02/26/2022 07:55:59 - INFO - codeparrot_training - Step 41989: {'lr': 3.358369869605934e-05, 'samples': 21498880, 'steps': 41989, 'loss/train': 2.545621871948242} 02/26/2022 07:56:05 - INFO - codeparrot_training - Step 41990: {'lr': 3.3575507728957926e-05, 'samples': 21499392, 'steps': 41990, 'loss/train': 1.1321842670440674} 02/26/2022 07:56:08 - INFO - codeparrot_training - Step 41991: {'lr': 3.356731768895027e-05, 'samples': 21499904, 'steps': 41991, 'loss/train': 1.6649385690689087} 02/26/2022 07:56:14 - INFO - codeparrot_training - Step 41992: {'lr': 3.355912857607157e-05, 'samples': 21500416, 'steps': 41992, 'loss/train': 0.2660706043243408} 02/26/2022 07:56:18 - INFO - codeparrot_training - Step 41993: {'lr': 3.355094039035681e-05, 'samples': 21500928, 'steps': 41993, 'loss/train': 1.1076467037200928} 02/26/2022 07:56:23 - INFO - codeparrot_training - Step 41994: {'lr': 3.35427531318411e-05, 'samples': 21501440, 'steps': 41994, 'loss/train': 1.601947546005249} 02/26/2022 07:56:27 - INFO - codeparrot_training - Step 41995: {'lr': 3.3534566800559444e-05, 'samples': 21501952, 'steps': 41995, 'loss/train': 2.1484107971191406} 02/26/2022 07:56:32 - INFO - codeparrot_training - Step 41996: {'lr': 3.352638139654704e-05, 'samples': 21502464, 'steps': 41996, 'loss/train': 1.9688382148742676} 02/26/2022 07:56:36 - INFO - codeparrot_training - Step 41997: {'lr': 3.3518196919838874e-05, 'samples': 21502976, 'steps': 41997, 'loss/train': 1.4797433614730835} 02/26/2022 07:56:41 - INFO - codeparrot_training - Step 41998: {'lr': 3.351001337047e-05, 'samples': 21503488, 'steps': 41998, 'loss/train': 1.3362501859664917} 02/26/2022 07:56:45 - INFO - codeparrot_training - Step 41999: {'lr': 3.350183074847549e-05, 'samples': 21504000, 'steps': 41999, 'loss/train': 1.3207478523254395} 02/26/2022 07:56:45 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 07:57:03 - WARNING - huggingface_hub.repository - Several commits (42) will be pushed upstream. 02/26/2022 07:57:03 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 07:57:36 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 2427a89..2c9f21a floral-grass-11 -> floral-grass-11 02/26/2022 07:57:44 - INFO - codeparrot_training - Step 42000: {'lr': 3.3493649053890325e-05, 'samples': 21504512, 'steps': 42000, 'loss/train': 1.5494685173034668} 02/26/2022 07:57:48 - INFO - codeparrot_training - Step 42001: {'lr': 3.34854682867497e-05, 'samples': 21505024, 'steps': 42001, 'loss/train': 1.836612582206726} 02/26/2022 07:57:53 - INFO - codeparrot_training - Step 42002: {'lr': 3.347728844708855e-05, 'samples': 21505536, 'steps': 42002, 'loss/train': 1.8202868700027466} 02/26/2022 07:57:57 - INFO - codeparrot_training - Step 42003: {'lr': 3.3469109534941976e-05, 'samples': 21506048, 'steps': 42003, 'loss/train': 0.8949858546257019} 02/26/2022 07:58:02 - INFO - codeparrot_training - Step 42004: {'lr': 3.346093155034491e-05, 'samples': 21506560, 'steps': 42004, 'loss/train': 1.3837007284164429} 02/26/2022 07:58:06 - INFO - codeparrot_training - Step 42005: {'lr': 3.345275449333249e-05, 'samples': 21507072, 'steps': 42005, 'loss/train': 1.9728405475616455} 02/26/2022 07:58:11 - INFO - codeparrot_training - Step 42006: {'lr': 3.3444578363939715e-05, 'samples': 21507584, 'steps': 42006, 'loss/train': 1.2473104000091553} 02/26/2022 07:58:15 - INFO - codeparrot_training - Step 42007: {'lr': 3.343640316220162e-05, 'samples': 21508096, 'steps': 42007, 'loss/train': 1.5608479976654053} 02/26/2022 07:58:20 - INFO - codeparrot_training - Step 42008: {'lr': 3.342822888815314e-05, 'samples': 21508608, 'steps': 42008, 'loss/train': 0.278462678194046} 02/26/2022 07:58:24 - INFO - codeparrot_training - Step 42009: {'lr': 3.342005554182942e-05, 'samples': 21509120, 'steps': 42009, 'loss/train': 1.3350027799606323} 02/26/2022 07:58:30 - INFO - codeparrot_training - Step 42010: {'lr': 3.34118831232654e-05, 'samples': 21509632, 'steps': 42010, 'loss/train': 2.1366682052612305} 02/26/2022 07:58:33 - INFO - codeparrot_training - Step 42011: {'lr': 3.3403711632496114e-05, 'samples': 21510144, 'steps': 42011, 'loss/train': 1.8913192749023438} 02/26/2022 07:58:39 - INFO - codeparrot_training - Step 42012: {'lr': 3.3395541069556477e-05, 'samples': 21510656, 'steps': 42012, 'loss/train': 1.1241506338119507} 02/26/2022 07:58:42 - INFO - codeparrot_training - Step 42013: {'lr': 3.3387371434481646e-05, 'samples': 21511168, 'steps': 42013, 'loss/train': 1.008219599723816} 02/26/2022 07:58:48 - INFO - codeparrot_training - Step 42014: {'lr': 3.337920272730649e-05, 'samples': 21511680, 'steps': 42014, 'loss/train': 2.3886592388153076} 02/26/2022 07:58:51 - INFO - codeparrot_training - Step 42015: {'lr': 3.3371034948066064e-05, 'samples': 21512192, 'steps': 42015, 'loss/train': 2.2312192916870117} 02/26/2022 07:58:58 - INFO - codeparrot_training - Step 42016: {'lr': 3.336286809679531e-05, 'samples': 21512704, 'steps': 42016, 'loss/train': 1.4221895933151245} 02/26/2022 07:59:01 - INFO - codeparrot_training - Step 42017: {'lr': 3.335470217352926e-05, 'samples': 21513216, 'steps': 42017, 'loss/train': 1.9809162616729736} 02/26/2022 07:59:07 - INFO - codeparrot_training - Step 42018: {'lr': 3.3346537178302825e-05, 'samples': 21513728, 'steps': 42018, 'loss/train': 1.392099380493164} 02/26/2022 07:59:10 - INFO - codeparrot_training - Step 42019: {'lr': 3.333837311115115e-05, 'samples': 21514240, 'steps': 42019, 'loss/train': 2.8892767429351807} 02/26/2022 07:59:16 - INFO - codeparrot_training - Step 42020: {'lr': 3.3330209972108976e-05, 'samples': 21514752, 'steps': 42020, 'loss/train': 0.7633283138275146} 02/26/2022 07:59:19 - INFO - codeparrot_training - Step 42021: {'lr': 3.332204776121142e-05, 'samples': 21515264, 'steps': 42021, 'loss/train': 2.427926778793335} 02/26/2022 07:59:25 - INFO - codeparrot_training - Step 42022: {'lr': 3.331388647849337e-05, 'samples': 21515776, 'steps': 42022, 'loss/train': 0.5591949820518494} 02/26/2022 07:59:29 - INFO - codeparrot_training - Step 42023: {'lr': 3.3305726123989936e-05, 'samples': 21516288, 'steps': 42023, 'loss/train': 2.251450538635254} 02/26/2022 07:59:32 - INFO - codeparrot_training - Step 42024: {'lr': 3.329756669773584e-05, 'samples': 21516800, 'steps': 42024, 'loss/train': 2.8341686725616455} 02/26/2022 07:59:38 - INFO - codeparrot_training - Step 42025: {'lr': 3.3289408199766224e-05, 'samples': 21517312, 'steps': 42025, 'loss/train': 1.8091281652450562} 02/26/2022 07:59:41 - INFO - codeparrot_training - Step 42026: {'lr': 3.3281250630115886e-05, 'samples': 21517824, 'steps': 42026, 'loss/train': 2.607653856277466} 02/26/2022 07:59:48 - INFO - codeparrot_training - Step 42027: {'lr': 3.327309398882e-05, 'samples': 21518336, 'steps': 42027, 'loss/train': 1.6428308486938477} 02/26/2022 07:59:51 - INFO - codeparrot_training - Step 42028: {'lr': 3.326493827591323e-05, 'samples': 21518848, 'steps': 42028, 'loss/train': 0.2699722647666931} 02/26/2022 07:59:57 - INFO - codeparrot_training - Step 42029: {'lr': 3.325678349143071e-05, 'samples': 21519360, 'steps': 42029, 'loss/train': 1.4577146768569946} 02/26/2022 08:00:00 - INFO - codeparrot_training - Step 42030: {'lr': 3.324862963540726e-05, 'samples': 21519872, 'steps': 42030, 'loss/train': 1.332014560699463} 02/26/2022 08:00:06 - INFO - codeparrot_training - Step 42031: {'lr': 3.3240476707877884e-05, 'samples': 21520384, 'steps': 42031, 'loss/train': 1.5099050998687744} 02/26/2022 08:00:09 - INFO - codeparrot_training - Step 42032: {'lr': 3.323232470887749e-05, 'samples': 21520896, 'steps': 42032, 'loss/train': 1.7806147336959839} 02/26/2022 08:00:15 - INFO - codeparrot_training - Step 42033: {'lr': 3.322417363844099e-05, 'samples': 21521408, 'steps': 42033, 'loss/train': 1.884886384010315} 02/26/2022 08:00:18 - INFO - codeparrot_training - Step 42034: {'lr': 3.3216023496603245e-05, 'samples': 21521920, 'steps': 42034, 'loss/train': 0.9722924828529358} 02/26/2022 08:00:24 - INFO - codeparrot_training - Step 42035: {'lr': 3.3207874283399247e-05, 'samples': 21522432, 'steps': 42035, 'loss/train': 2.0101940631866455} 02/26/2022 08:00:27 - INFO - codeparrot_training - Step 42036: {'lr': 3.3199725998863896e-05, 'samples': 21522944, 'steps': 42036, 'loss/train': 1.547472596168518} 02/26/2022 08:00:34 - INFO - codeparrot_training - Step 42037: {'lr': 3.319157864303207e-05, 'samples': 21523456, 'steps': 42037, 'loss/train': 2.094116449356079} 02/26/2022 08:00:38 - INFO - codeparrot_training - Step 42038: {'lr': 3.318343221593864e-05, 'samples': 21523968, 'steps': 42038, 'loss/train': 2.825582981109619} 02/26/2022 08:00:43 - INFO - codeparrot_training - Step 42039: {'lr': 3.3175286717618574e-05, 'samples': 21524480, 'steps': 42039, 'loss/train': 2.123781442642212} 02/26/2022 08:00:47 - INFO - codeparrot_training - Step 42040: {'lr': 3.316714214810676e-05, 'samples': 21524992, 'steps': 42040, 'loss/train': 2.219564437866211} 02/26/2022 08:00:52 - INFO - codeparrot_training - Step 42041: {'lr': 3.315899850743803e-05, 'samples': 21525504, 'steps': 42041, 'loss/train': 1.5891379117965698} 02/26/2022 08:00:56 - INFO - codeparrot_training - Step 42042: {'lr': 3.315085579564725e-05, 'samples': 21526016, 'steps': 42042, 'loss/train': 1.7330372333526611} 02/26/2022 08:01:01 - INFO - codeparrot_training - Step 42043: {'lr': 3.314271401276939e-05, 'samples': 21526528, 'steps': 42043, 'loss/train': 2.413616180419922} 02/26/2022 08:01:05 - INFO - codeparrot_training - Step 42044: {'lr': 3.3134573158839276e-05, 'samples': 21527040, 'steps': 42044, 'loss/train': 2.170334577560425} 02/26/2022 08:01:10 - INFO - codeparrot_training - Step 42045: {'lr': 3.312643323389181e-05, 'samples': 21527552, 'steps': 42045, 'loss/train': 1.7984298467636108} 02/26/2022 08:01:13 - INFO - codeparrot_training - Step 42046: {'lr': 3.311829423796184e-05, 'samples': 21528064, 'steps': 42046, 'loss/train': 1.1373330354690552} 02/26/2022 08:01:19 - INFO - codeparrot_training - Step 42047: {'lr': 3.311015617108415e-05, 'samples': 21528576, 'steps': 42047, 'loss/train': 2.523714780807495} 02/26/2022 08:01:23 - INFO - codeparrot_training - Step 42048: {'lr': 3.310201903329377e-05, 'samples': 21529088, 'steps': 42048, 'loss/train': 1.9073712825775146} 02/26/2022 08:01:30 - INFO - codeparrot_training - Step 42049: {'lr': 3.3093882824625445e-05, 'samples': 21529600, 'steps': 42049, 'loss/train': 1.5528634786605835} 02/26/2022 08:01:33 - INFO - codeparrot_training - Step 42050: {'lr': 3.308574754511404e-05, 'samples': 21530112, 'steps': 42050, 'loss/train': 2.2407617568969727} 02/26/2022 08:01:39 - INFO - codeparrot_training - Step 42051: {'lr': 3.307761319479438e-05, 'samples': 21530624, 'steps': 42051, 'loss/train': 2.026862382888794} 02/26/2022 08:01:42 - INFO - codeparrot_training - Step 42052: {'lr': 3.30694797737014e-05, 'samples': 21531136, 'steps': 42052, 'loss/train': 0.8512178659439087} 02/26/2022 08:01:48 - INFO - codeparrot_training - Step 42053: {'lr': 3.306134728186983e-05, 'samples': 21531648, 'steps': 42053, 'loss/train': 0.7336413264274597} 02/26/2022 08:01:51 - INFO - codeparrot_training - Step 42054: {'lr': 3.3053215719334645e-05, 'samples': 21532160, 'steps': 42054, 'loss/train': 2.0124905109405518} 02/26/2022 08:01:57 - INFO - codeparrot_training - Step 42055: {'lr': 3.3045085086130524e-05, 'samples': 21532672, 'steps': 42055, 'loss/train': 1.8189895153045654} 02/26/2022 08:02:00 - INFO - codeparrot_training - Step 42056: {'lr': 3.30369553822924e-05, 'samples': 21533184, 'steps': 42056, 'loss/train': 1.744791865348816} 02/26/2022 08:02:06 - INFO - codeparrot_training - Step 42057: {'lr': 3.3028826607855e-05, 'samples': 21533696, 'steps': 42057, 'loss/train': 1.1750150918960571} 02/26/2022 08:02:09 - INFO - codeparrot_training - Step 42058: {'lr': 3.302069876285335e-05, 'samples': 21534208, 'steps': 42058, 'loss/train': 1.453096866607666} 02/26/2022 08:02:15 - INFO - codeparrot_training - Step 42059: {'lr': 3.3012571847322006e-05, 'samples': 21534720, 'steps': 42059, 'loss/train': 2.6410202980041504} 02/26/2022 08:02:18 - INFO - codeparrot_training - Step 42060: {'lr': 3.300444586129595e-05, 'samples': 21535232, 'steps': 42060, 'loss/train': 1.747113585472107} 02/26/2022 08:02:24 - INFO - codeparrot_training - Step 42061: {'lr': 3.299632080480988e-05, 'samples': 21535744, 'steps': 42061, 'loss/train': 1.5606273412704468} 02/26/2022 08:02:27 - INFO - codeparrot_training - Step 42062: {'lr': 3.298819667789882e-05, 'samples': 21536256, 'steps': 42062, 'loss/train': 1.8926509618759155} 02/26/2022 08:02:34 - INFO - codeparrot_training - Step 42063: {'lr': 3.298007348059726e-05, 'samples': 21536768, 'steps': 42063, 'loss/train': 2.08646559715271} 02/26/2022 08:02:37 - INFO - codeparrot_training - Step 42064: {'lr': 3.297195121294022e-05, 'samples': 21537280, 'steps': 42064, 'loss/train': 2.0213544368743896} 02/26/2022 08:02:43 - INFO - codeparrot_training - Step 42065: {'lr': 3.296382987496238e-05, 'samples': 21537792, 'steps': 42065, 'loss/train': 1.816352367401123} 02/26/2022 08:02:46 - INFO - codeparrot_training - Step 42066: {'lr': 3.2955709466698625e-05, 'samples': 21538304, 'steps': 42066, 'loss/train': 0.8643584847450256} 02/26/2022 08:02:52 - INFO - codeparrot_training - Step 42067: {'lr': 3.29475899881837e-05, 'samples': 21538816, 'steps': 42067, 'loss/train': 0.5341842174530029} 02/26/2022 08:02:55 - INFO - codeparrot_training - Step 42068: {'lr': 3.293947143945236e-05, 'samples': 21539328, 'steps': 42068, 'loss/train': 1.8572545051574707} 02/26/2022 08:03:01 - INFO - codeparrot_training - Step 42069: {'lr': 3.293135382053938e-05, 'samples': 21539840, 'steps': 42069, 'loss/train': 2.3421874046325684} 02/26/2022 08:03:06 - INFO - codeparrot_training - Step 42070: {'lr': 3.292323713147957e-05, 'samples': 21540352, 'steps': 42070, 'loss/train': 1.0097898244857788} 02/26/2022 08:03:10 - INFO - codeparrot_training - Step 42071: {'lr': 3.2915121372307704e-05, 'samples': 21540864, 'steps': 42071, 'loss/train': 2.0727267265319824} 02/26/2022 08:03:16 - INFO - codeparrot_training - Step 42072: {'lr': 3.290700654305853e-05, 'samples': 21541376, 'steps': 42072, 'loss/train': 0.837982714176178} 02/26/2022 08:03:20 - INFO - codeparrot_training - Step 42073: {'lr': 3.289889264376672e-05, 'samples': 21541888, 'steps': 42073, 'loss/train': 1.8117860555648804} 02/26/2022 08:03:25 - INFO - codeparrot_training - Step 42074: {'lr': 3.289077967446721e-05, 'samples': 21542400, 'steps': 42074, 'loss/train': 4.799282073974609} 02/26/2022 08:03:29 - INFO - codeparrot_training - Step 42075: {'lr': 3.288266763519462e-05, 'samples': 21542912, 'steps': 42075, 'loss/train': 2.1497035026550293} 02/26/2022 08:03:34 - INFO - codeparrot_training - Step 42076: {'lr': 3.287455652598378e-05, 'samples': 21543424, 'steps': 42076, 'loss/train': 1.7228047847747803} 02/26/2022 08:03:38 - INFO - codeparrot_training - Step 42077: {'lr': 3.28664463468693e-05, 'samples': 21543936, 'steps': 42077, 'loss/train': 2.276219367980957} 02/26/2022 08:03:43 - INFO - codeparrot_training - Step 42078: {'lr': 3.285833709788611e-05, 'samples': 21544448, 'steps': 42078, 'loss/train': 1.7032498121261597} 02/26/2022 08:03:47 - INFO - codeparrot_training - Step 42079: {'lr': 3.285022877906882e-05, 'samples': 21544960, 'steps': 42079, 'loss/train': 1.9602081775665283} 02/26/2022 08:03:52 - INFO - codeparrot_training - Step 42080: {'lr': 3.284212139045223e-05, 'samples': 21545472, 'steps': 42080, 'loss/train': 1.1926156282424927} 02/26/2022 08:03:56 - INFO - codeparrot_training - Step 42081: {'lr': 3.2834014932070945e-05, 'samples': 21545984, 'steps': 42081, 'loss/train': 0.9904447793960571} 02/26/2022 08:04:02 - INFO - codeparrot_training - Step 42082: {'lr': 3.2825909403959884e-05, 'samples': 21546496, 'steps': 42082, 'loss/train': 4.8680739402771} 02/26/2022 08:04:05 - INFO - codeparrot_training - Step 42083: {'lr': 3.281780480615362e-05, 'samples': 21547008, 'steps': 42083, 'loss/train': 1.8000751733779907} 02/26/2022 08:04:12 - INFO - codeparrot_training - Step 42084: {'lr': 3.280970113868695e-05, 'samples': 21547520, 'steps': 42084, 'loss/train': 1.3541169166564941} 02/26/2022 08:04:15 - INFO - codeparrot_training - Step 42085: {'lr': 3.280159840159447e-05, 'samples': 21548032, 'steps': 42085, 'loss/train': 3.1170871257781982} 02/26/2022 08:04:20 - INFO - codeparrot_training - Step 42086: {'lr': 3.279349659491104e-05, 'samples': 21548544, 'steps': 42086, 'loss/train': 1.6846998929977417} 02/26/2022 08:04:24 - INFO - codeparrot_training - Step 42087: {'lr': 3.278539571867131e-05, 'samples': 21549056, 'steps': 42087, 'loss/train': 0.8489043116569519} 02/26/2022 08:04:30 - INFO - codeparrot_training - Step 42088: {'lr': 3.277729577290997e-05, 'samples': 21549568, 'steps': 42088, 'loss/train': 2.935265064239502} 02/26/2022 08:04:33 - INFO - codeparrot_training - Step 42089: {'lr': 3.2769196757661654e-05, 'samples': 21550080, 'steps': 42089, 'loss/train': 1.5126219987869263} 02/26/2022 08:04:39 - INFO - codeparrot_training - Step 42090: {'lr': 3.276109867296117e-05, 'samples': 21550592, 'steps': 42090, 'loss/train': 2.5620248317718506} 02/26/2022 08:04:42 - INFO - codeparrot_training - Step 42091: {'lr': 3.2753001518843194e-05, 'samples': 21551104, 'steps': 42091, 'loss/train': 1.7641351222991943} 02/26/2022 08:04:48 - INFO - codeparrot_training - Step 42092: {'lr': 3.2744905295342295e-05, 'samples': 21551616, 'steps': 42092, 'loss/train': 2.3234059810638428} 02/26/2022 08:04:51 - INFO - codeparrot_training - Step 42093: {'lr': 3.273681000249329e-05, 'samples': 21552128, 'steps': 42093, 'loss/train': 1.7815600633621216} 02/26/2022 08:04:58 - INFO - codeparrot_training - Step 42094: {'lr': 3.2728715640330784e-05, 'samples': 21552640, 'steps': 42094, 'loss/train': 1.3718794584274292} 02/26/2022 08:05:01 - INFO - codeparrot_training - Step 42095: {'lr': 3.272062220888949e-05, 'samples': 21553152, 'steps': 42095, 'loss/train': 1.898031234741211} 02/26/2022 08:05:07 - INFO - codeparrot_training - Step 42096: {'lr': 3.271252970820401e-05, 'samples': 21553664, 'steps': 42096, 'loss/train': 1.3841553926467896} 02/26/2022 08:05:10 - INFO - codeparrot_training - Step 42097: {'lr': 3.2704438138309146e-05, 'samples': 21554176, 'steps': 42097, 'loss/train': 1.3685778379440308} 02/26/2022 08:05:16 - INFO - codeparrot_training - Step 42098: {'lr': 3.2696347499239385e-05, 'samples': 21554688, 'steps': 42098, 'loss/train': 2.179177761077881} 02/26/2022 08:05:20 - INFO - codeparrot_training - Step 42099: {'lr': 3.268825779102949e-05, 'samples': 21555200, 'steps': 42099, 'loss/train': 2.3120627403259277} 02/26/2022 08:05:23 - INFO - codeparrot_training - Step 42100: {'lr': 3.268016901371407e-05, 'samples': 21555712, 'steps': 42100, 'loss/train': 1.750062346458435} 02/26/2022 08:05:29 - INFO - codeparrot_training - Step 42101: {'lr': 3.267208116732784e-05, 'samples': 21556224, 'steps': 42101, 'loss/train': 2.200613021850586} 02/26/2022 08:05:32 - INFO - codeparrot_training - Step 42102: {'lr': 3.266399425190542e-05, 'samples': 21556736, 'steps': 42102, 'loss/train': 0.6815493106842041} 02/26/2022 08:05:38 - INFO - codeparrot_training - Step 42103: {'lr': 3.265590826748141e-05, 'samples': 21557248, 'steps': 42103, 'loss/train': 2.0527563095092773} 02/26/2022 08:05:41 - INFO - codeparrot_training - Step 42104: {'lr': 3.2647823214090436e-05, 'samples': 21557760, 'steps': 42104, 'loss/train': 2.652242660522461} 02/26/2022 08:05:47 - INFO - codeparrot_training - Step 42105: {'lr': 3.2639739091767236e-05, 'samples': 21558272, 'steps': 42105, 'loss/train': 2.2487101554870605} 02/26/2022 08:05:50 - INFO - codeparrot_training - Step 42106: {'lr': 3.263165590054634e-05, 'samples': 21558784, 'steps': 42106, 'loss/train': 1.612600564956665} 02/26/2022 08:05:56 - INFO - codeparrot_training - Step 42107: {'lr': 3.262357364046245e-05, 'samples': 21559296, 'steps': 42107, 'loss/train': 1.869500756263733} 02/26/2022 08:05:59 - INFO - codeparrot_training - Step 42108: {'lr': 3.261549231155006e-05, 'samples': 21559808, 'steps': 42108, 'loss/train': 0.8940420150756836} 02/26/2022 08:06:06 - INFO - codeparrot_training - Step 42109: {'lr': 3.260741191384395e-05, 'samples': 21560320, 'steps': 42109, 'loss/train': 1.62667977809906} 02/26/2022 08:06:09 - INFO - codeparrot_training - Step 42110: {'lr': 3.2599332447378644e-05, 'samples': 21560832, 'steps': 42110, 'loss/train': 1.635771632194519} 02/26/2022 08:06:15 - INFO - codeparrot_training - Step 42111: {'lr': 3.259125391218875e-05, 'samples': 21561344, 'steps': 42111, 'loss/train': 1.9833099842071533} 02/26/2022 08:06:18 - INFO - codeparrot_training - Step 42112: {'lr': 3.258317630830887e-05, 'samples': 21561856, 'steps': 42112, 'loss/train': 0.9248422384262085} 02/26/2022 08:06:24 - INFO - codeparrot_training - Step 42113: {'lr': 3.257509963577365e-05, 'samples': 21562368, 'steps': 42113, 'loss/train': 1.0906859636306763} 02/26/2022 08:06:27 - INFO - codeparrot_training - Step 42114: {'lr': 3.256702389461763e-05, 'samples': 21562880, 'steps': 42114, 'loss/train': 2.0703229904174805} 02/26/2022 08:06:33 - INFO - codeparrot_training - Step 42115: {'lr': 3.2558949084875525e-05, 'samples': 21563392, 'steps': 42115, 'loss/train': 2.565927267074585} 02/26/2022 08:06:36 - INFO - codeparrot_training - Step 42116: {'lr': 3.255087520658173e-05, 'samples': 21563904, 'steps': 42116, 'loss/train': 2.0509915351867676} 02/26/2022 08:06:42 - INFO - codeparrot_training - Step 42117: {'lr': 3.254280225977102e-05, 'samples': 21564416, 'steps': 42117, 'loss/train': 1.783235788345337} 02/26/2022 08:06:45 - INFO - codeparrot_training - Step 42118: {'lr': 3.253473024447781e-05, 'samples': 21564928, 'steps': 42118, 'loss/train': 0.44805601239204407} 02/26/2022 08:06:51 - INFO - codeparrot_training - Step 42119: {'lr': 3.2526659160736885e-05, 'samples': 21565440, 'steps': 42119, 'loss/train': 1.477900505065918} 02/26/2022 08:06:54 - INFO - codeparrot_training - Step 42120: {'lr': 3.251858900858259e-05, 'samples': 21565952, 'steps': 42120, 'loss/train': 1.5423990488052368} 02/26/2022 08:07:01 - INFO - codeparrot_training - Step 42121: {'lr': 3.2510519788049644e-05, 'samples': 21566464, 'steps': 42121, 'loss/train': 1.6041160821914673} 02/26/2022 08:07:04 - INFO - codeparrot_training - Step 42122: {'lr': 3.250245149917252e-05, 'samples': 21566976, 'steps': 42122, 'loss/train': 0.27017030119895935} 02/26/2022 08:07:10 - INFO - codeparrot_training - Step 42123: {'lr': 3.249438414198594e-05, 'samples': 21567488, 'steps': 42123, 'loss/train': 0.8072147965431213} 02/26/2022 08:07:13 - INFO - codeparrot_training - Step 42124: {'lr': 3.2486317716524246e-05, 'samples': 21568000, 'steps': 42124, 'loss/train': 2.299391031265259} 02/26/2022 08:07:19 - INFO - codeparrot_training - Step 42125: {'lr': 3.247825222282216e-05, 'samples': 21568512, 'steps': 42125, 'loss/train': 0.8135859370231628} 02/26/2022 08:07:22 - INFO - codeparrot_training - Step 42126: {'lr': 3.247018766091411e-05, 'samples': 21569024, 'steps': 42126, 'loss/train': 2.7928144931793213} 02/26/2022 08:07:28 - INFO - codeparrot_training - Step 42127: {'lr': 3.246212403083476e-05, 'samples': 21569536, 'steps': 42127, 'loss/train': 2.626312255859375} 02/26/2022 08:07:31 - INFO - codeparrot_training - Step 42128: {'lr': 3.245406133261858e-05, 'samples': 21570048, 'steps': 42128, 'loss/train': 1.3988430500030518} 02/26/2022 08:07:37 - INFO - codeparrot_training - Step 42129: {'lr': 3.244599956630015e-05, 'samples': 21570560, 'steps': 42129, 'loss/train': 1.2855547666549683} 02/26/2022 08:07:40 - INFO - codeparrot_training - Step 42130: {'lr': 3.243793873191395e-05, 'samples': 21571072, 'steps': 42130, 'loss/train': 1.954200029373169} 02/26/2022 08:07:47 - INFO - codeparrot_training - Step 42131: {'lr': 3.24298788294945e-05, 'samples': 21571584, 'steps': 42131, 'loss/train': 2.026719570159912} 02/26/2022 08:07:53 - INFO - codeparrot_training - Step 42132: {'lr': 3.2421819859076415e-05, 'samples': 21572096, 'steps': 42132, 'loss/train': 2.2114007472991943} 02/26/2022 08:07:56 - INFO - codeparrot_training - Step 42133: {'lr': 3.241376182069414e-05, 'samples': 21572608, 'steps': 42133, 'loss/train': 1.7515803575515747} 02/26/2022 08:08:02 - INFO - codeparrot_training - Step 42134: {'lr': 3.240570471438225e-05, 'samples': 21573120, 'steps': 42134, 'loss/train': 0.6081182360649109} 02/26/2022 08:08:05 - INFO - codeparrot_training - Step 42135: {'lr': 3.239764854017516e-05, 'samples': 21573632, 'steps': 42135, 'loss/train': 2.7466824054718018} 02/26/2022 08:08:11 - INFO - codeparrot_training - Step 42136: {'lr': 3.2389593298107506e-05, 'samples': 21574144, 'steps': 42136, 'loss/train': 0.5307530760765076} 02/26/2022 08:08:14 - INFO - codeparrot_training - Step 42137: {'lr': 3.238153898821372e-05, 'samples': 21574656, 'steps': 42137, 'loss/train': 2.0336854457855225} 02/26/2022 08:08:20 - INFO - codeparrot_training - Step 42138: {'lr': 3.237348561052833e-05, 'samples': 21575168, 'steps': 42138, 'loss/train': 1.5019077062606812} 02/26/2022 08:08:23 - INFO - codeparrot_training - Step 42139: {'lr': 3.2365433165085777e-05, 'samples': 21575680, 'steps': 42139, 'loss/train': 2.0933120250701904} 02/26/2022 08:08:30 - INFO - codeparrot_training - Step 42140: {'lr': 3.235738165192065e-05, 'samples': 21576192, 'steps': 42140, 'loss/train': 1.8026868104934692} 02/26/2022 08:08:33 - INFO - codeparrot_training - Step 42141: {'lr': 3.234933107106738e-05, 'samples': 21576704, 'steps': 42141, 'loss/train': 1.2884650230407715} 02/26/2022 08:08:39 - INFO - codeparrot_training - Step 42142: {'lr': 3.234128142256049e-05, 'samples': 21577216, 'steps': 42142, 'loss/train': 2.894014358520508} 02/26/2022 08:08:42 - INFO - codeparrot_training - Step 42143: {'lr': 3.2333232706434346e-05, 'samples': 21577728, 'steps': 42143, 'loss/train': 1.3358802795410156} 02/26/2022 08:08:48 - INFO - codeparrot_training - Step 42144: {'lr': 3.23251849227236e-05, 'samples': 21578240, 'steps': 42144, 'loss/train': 2.144378662109375} 02/26/2022 08:08:51 - INFO - codeparrot_training - Step 42145: {'lr': 3.231713807146264e-05, 'samples': 21578752, 'steps': 42145, 'loss/train': 2.0812461376190186} 02/26/2022 08:08:57 - INFO - codeparrot_training - Step 42146: {'lr': 3.2309092152685905e-05, 'samples': 21579264, 'steps': 42146, 'loss/train': 2.204704999923706} 02/26/2022 08:09:00 - INFO - codeparrot_training - Step 42147: {'lr': 3.230104716642787e-05, 'samples': 21579776, 'steps': 42147, 'loss/train': 1.767375111579895} 02/26/2022 08:09:06 - INFO - codeparrot_training - Step 42148: {'lr': 3.229300311272307e-05, 'samples': 21580288, 'steps': 42148, 'loss/train': 0.8443734645843506} 02/26/2022 08:09:09 - INFO - codeparrot_training - Step 42149: {'lr': 3.2284959991605914e-05, 'samples': 21580800, 'steps': 42149, 'loss/train': 0.5046215057373047} 02/26/2022 08:09:15 - INFO - codeparrot_training - Step 42150: {'lr': 3.227691780311087e-05, 'samples': 21581312, 'steps': 42150, 'loss/train': 0.4444860517978668} 02/26/2022 08:09:18 - INFO - codeparrot_training - Step 42151: {'lr': 3.226887654727231e-05, 'samples': 21581824, 'steps': 42151, 'loss/train': 1.469581127166748} 02/26/2022 08:09:24 - INFO - codeparrot_training - Step 42152: {'lr': 3.226083622412479e-05, 'samples': 21582336, 'steps': 42152, 'loss/train': 1.8099740743637085} 02/26/2022 08:09:27 - INFO - codeparrot_training - Step 42153: {'lr': 3.225279683370266e-05, 'samples': 21582848, 'steps': 42153, 'loss/train': 0.1667536199092865} 02/26/2022 08:09:33 - INFO - codeparrot_training - Step 42154: {'lr': 3.224475837604052e-05, 'samples': 21583360, 'steps': 42154, 'loss/train': 0.7521966099739075} 02/26/2022 08:09:36 - INFO - codeparrot_training - Step 42155: {'lr': 3.223672085117257e-05, 'samples': 21583872, 'steps': 42155, 'loss/train': 1.3273061513900757} 02/26/2022 08:09:43 - INFO - codeparrot_training - Step 42156: {'lr': 3.222868425913345e-05, 'samples': 21584384, 'steps': 42156, 'loss/train': 1.8184926509857178} 02/26/2022 08:09:46 - INFO - codeparrot_training - Step 42157: {'lr': 3.2220648599957407e-05, 'samples': 21584896, 'steps': 42157, 'loss/train': 1.391922116279602} 02/26/2022 08:09:52 - INFO - codeparrot_training - Step 42158: {'lr': 3.221261387367908e-05, 'samples': 21585408, 'steps': 42158, 'loss/train': 1.7123470306396484} 02/26/2022 08:09:55 - INFO - codeparrot_training - Step 42159: {'lr': 3.220458008033261e-05, 'samples': 21585920, 'steps': 42159, 'loss/train': 1.8236193656921387} 02/26/2022 08:10:01 - INFO - codeparrot_training - Step 42160: {'lr': 3.219654721995266e-05, 'samples': 21586432, 'steps': 42160, 'loss/train': 1.917496681213379} 02/26/2022 08:10:04 - INFO - codeparrot_training - Step 42161: {'lr': 3.218851529257347e-05, 'samples': 21586944, 'steps': 42161, 'loss/train': 1.730018973350525} 02/26/2022 08:10:10 - INFO - codeparrot_training - Step 42162: {'lr': 3.218048429822959e-05, 'samples': 21587456, 'steps': 42162, 'loss/train': 0.8438036441802979} 02/26/2022 08:10:13 - INFO - codeparrot_training - Step 42163: {'lr': 3.217245423695534e-05, 'samples': 21587968, 'steps': 42163, 'loss/train': 1.4133445024490356} 02/26/2022 08:10:19 - INFO - codeparrot_training - Step 42164: {'lr': 3.2164425108785114e-05, 'samples': 21588480, 'steps': 42164, 'loss/train': 2.0032968521118164} 02/26/2022 08:10:22 - INFO - codeparrot_training - Step 42165: {'lr': 3.215639691375327e-05, 'samples': 21588992, 'steps': 42165, 'loss/train': 1.2030644416809082} 02/26/2022 08:10:29 - INFO - codeparrot_training - Step 42166: {'lr': 3.21483696518943e-05, 'samples': 21589504, 'steps': 42166, 'loss/train': 1.6856492757797241} 02/26/2022 08:10:32 - INFO - codeparrot_training - Step 42167: {'lr': 3.214034332324256e-05, 'samples': 21590016, 'steps': 42167, 'loss/train': 0.993902862071991} 02/26/2022 08:10:38 - INFO - codeparrot_training - Step 42168: {'lr': 3.2132317927832386e-05, 'samples': 21590528, 'steps': 42168, 'loss/train': 1.6544512510299683} 02/26/2022 08:10:41 - INFO - codeparrot_training - Step 42169: {'lr': 3.2124293465698146e-05, 'samples': 21591040, 'steps': 42169, 'loss/train': 1.4478480815887451} 02/26/2022 08:10:47 - INFO - codeparrot_training - Step 42170: {'lr': 3.21162699368743e-05, 'samples': 21591552, 'steps': 42170, 'loss/train': 0.8067625761032104} 02/26/2022 08:10:50 - INFO - codeparrot_training - Step 42171: {'lr': 3.2108247341395166e-05, 'samples': 21592064, 'steps': 42171, 'loss/train': 1.3318606615066528} 02/26/2022 08:10:56 - INFO - codeparrot_training - Step 42172: {'lr': 3.2100225679295104e-05, 'samples': 21592576, 'steps': 42172, 'loss/train': 0.8684771656990051} 02/26/2022 08:10:59 - INFO - codeparrot_training - Step 42173: {'lr': 3.2092204950608436e-05, 'samples': 21593088, 'steps': 42173, 'loss/train': 1.346718192100525} 02/26/2022 08:11:05 - INFO - codeparrot_training - Step 42174: {'lr': 3.208418515536962e-05, 'samples': 21593600, 'steps': 42174, 'loss/train': 2.430318832397461} 02/26/2022 08:11:08 - INFO - codeparrot_training - Step 42175: {'lr': 3.2076166293612956e-05, 'samples': 21594112, 'steps': 42175, 'loss/train': 0.9546864628791809} 02/26/2022 08:11:15 - INFO - codeparrot_training - Step 42176: {'lr': 3.206814836537281e-05, 'samples': 21594624, 'steps': 42176, 'loss/train': 1.54837965965271} 02/26/2022 08:11:19 - INFO - codeparrot_training - Step 42177: {'lr': 3.2060131370683505e-05, 'samples': 21595136, 'steps': 42177, 'loss/train': 1.2214683294296265} 02/26/2022 08:11:24 - INFO - codeparrot_training - Step 42178: {'lr': 3.205211530957935e-05, 'samples': 21595648, 'steps': 42178, 'loss/train': 1.5218291282653809} 02/26/2022 08:11:28 - INFO - codeparrot_training - Step 42179: {'lr': 3.204410018209478e-05, 'samples': 21596160, 'steps': 42179, 'loss/train': 0.08363626897335052} 02/26/2022 08:11:33 - INFO - codeparrot_training - Step 42180: {'lr': 3.2036085988264045e-05, 'samples': 21596672, 'steps': 42180, 'loss/train': 1.9375556707382202} 02/26/2022 08:11:37 - INFO - codeparrot_training - Step 42181: {'lr': 3.2028072728121525e-05, 'samples': 21597184, 'steps': 42181, 'loss/train': 1.794095754623413} 02/26/2022 08:11:43 - INFO - codeparrot_training - Step 42182: {'lr': 3.20200604017015e-05, 'samples': 21597696, 'steps': 42182, 'loss/train': 1.295919418334961} 02/26/2022 08:11:46 - INFO - codeparrot_training - Step 42183: {'lr': 3.201204900903834e-05, 'samples': 21598208, 'steps': 42183, 'loss/train': 2.2694764137268066} 02/26/2022 08:11:50 - INFO - codeparrot_training - Step 42184: {'lr': 3.200403855016637e-05, 'samples': 21598720, 'steps': 42184, 'loss/train': 1.5445663928985596} 02/26/2022 08:11:55 - INFO - codeparrot_training - Step 42185: {'lr': 3.199602902511986e-05, 'samples': 21599232, 'steps': 42185, 'loss/train': 1.3078842163085938} 02/26/2022 08:11:59 - INFO - codeparrot_training - Step 42186: {'lr': 3.198802043393309e-05, 'samples': 21599744, 'steps': 42186, 'loss/train': 1.4613194465637207} 02/26/2022 08:12:05 - INFO - codeparrot_training - Step 42187: {'lr': 3.198001277664045e-05, 'samples': 21600256, 'steps': 42187, 'loss/train': 1.7340518236160278} 02/26/2022 08:12:09 - INFO - codeparrot_training - Step 42188: {'lr': 3.197200605327616e-05, 'samples': 21600768, 'steps': 42188, 'loss/train': 1.4510672092437744} 02/26/2022 08:12:14 - INFO - codeparrot_training - Step 42189: {'lr': 3.196400026387469e-05, 'samples': 21601280, 'steps': 42189, 'loss/train': 1.1401158571243286} 02/26/2022 08:12:18 - INFO - codeparrot_training - Step 42190: {'lr': 3.195599540847008e-05, 'samples': 21601792, 'steps': 42190, 'loss/train': 2.254307508468628} 02/26/2022 08:12:24 - INFO - codeparrot_training - Step 42191: {'lr': 3.1947991487096815e-05, 'samples': 21602304, 'steps': 42191, 'loss/train': 1.657935380935669} 02/26/2022 08:12:27 - INFO - codeparrot_training - Step 42192: {'lr': 3.1939988499789076e-05, 'samples': 21602816, 'steps': 42192, 'loss/train': 2.2772719860076904} 02/26/2022 08:12:33 - INFO - codeparrot_training - Step 42193: {'lr': 3.193198644658127e-05, 'samples': 21603328, 'steps': 42193, 'loss/train': 2.2732653617858887} 02/26/2022 08:12:36 - INFO - codeparrot_training - Step 42194: {'lr': 3.1923985327507485e-05, 'samples': 21603840, 'steps': 42194, 'loss/train': 2.5038034915924072} 02/26/2022 08:12:42 - INFO - codeparrot_training - Step 42195: {'lr': 3.191598514260213e-05, 'samples': 21604352, 'steps': 42195, 'loss/train': 2.12198805809021} 02/26/2022 08:12:45 - INFO - codeparrot_training - Step 42196: {'lr': 3.190798589189944e-05, 'samples': 21604864, 'steps': 42196, 'loss/train': 1.015647292137146} 02/26/2022 08:12:51 - INFO - codeparrot_training - Step 42197: {'lr': 3.189998757543378e-05, 'samples': 21605376, 'steps': 42197, 'loss/train': 1.8399434089660645} 02/26/2022 08:12:54 - INFO - codeparrot_training - Step 42198: {'lr': 3.189199019323918e-05, 'samples': 21605888, 'steps': 42198, 'loss/train': 1.6753349304199219} 02/26/2022 08:13:00 - INFO - codeparrot_training - Step 42199: {'lr': 3.1883993745350114e-05, 'samples': 21606400, 'steps': 42199, 'loss/train': 1.3343709707260132} 02/26/2022 08:13:03 - INFO - codeparrot_training - Step 42200: {'lr': 3.187599823180071e-05, 'samples': 21606912, 'steps': 42200, 'loss/train': 1.2284804582595825} 02/26/2022 08:13:08 - INFO - codeparrot_training - Step 42201: {'lr': 3.186800365262532e-05, 'samples': 21607424, 'steps': 42201, 'loss/train': 1.6069765090942383} 02/26/2022 08:13:12 - INFO - codeparrot_training - Step 42202: {'lr': 3.1860010007858125e-05, 'samples': 21607936, 'steps': 42202, 'loss/train': 1.5057530403137207} 02/26/2022 08:13:19 - INFO - codeparrot_training - Step 42203: {'lr': 3.18520172975334e-05, 'samples': 21608448, 'steps': 42203, 'loss/train': 2.3396847248077393} 02/26/2022 08:13:22 - INFO - codeparrot_training - Step 42204: {'lr': 3.184402552168528e-05, 'samples': 21608960, 'steps': 42204, 'loss/train': 1.5844547748565674} 02/26/2022 08:13:28 - INFO - codeparrot_training - Step 42205: {'lr': 3.183603468034815e-05, 'samples': 21609472, 'steps': 42205, 'loss/train': 0.2942953109741211} 02/26/2022 08:13:31 - INFO - codeparrot_training - Step 42206: {'lr': 3.1828044773556195e-05, 'samples': 21609984, 'steps': 42206, 'loss/train': 1.871066927909851} 02/26/2022 08:13:37 - INFO - codeparrot_training - Step 42207: {'lr': 3.182005580134359e-05, 'samples': 21610496, 'steps': 42207, 'loss/train': 1.0271105766296387} 02/26/2022 08:13:40 - INFO - codeparrot_training - Step 42208: {'lr': 3.181206776374454e-05, 'samples': 21611008, 'steps': 42208, 'loss/train': 1.138145923614502} 02/26/2022 08:13:46 - INFO - codeparrot_training - Step 42209: {'lr': 3.180408066079335e-05, 'samples': 21611520, 'steps': 42209, 'loss/train': 2.3307571411132812} 02/26/2022 08:13:49 - INFO - codeparrot_training - Step 42210: {'lr': 3.1796094492524216e-05, 'samples': 21612032, 'steps': 42210, 'loss/train': 1.8100149631500244} 02/26/2022 08:13:55 - INFO - codeparrot_training - Step 42211: {'lr': 3.1788109258971324e-05, 'samples': 21612544, 'steps': 42211, 'loss/train': 0.5624557137489319} 02/26/2022 08:13:58 - INFO - codeparrot_training - Step 42212: {'lr': 3.1780124960168824e-05, 'samples': 21613056, 'steps': 42212, 'loss/train': 1.4848436117172241} 02/26/2022 08:14:05 - INFO - codeparrot_training - Step 42213: {'lr': 3.1772141596151023e-05, 'samples': 21613568, 'steps': 42213, 'loss/train': 2.242154598236084} 02/26/2022 08:14:08 - INFO - codeparrot_training - Step 42214: {'lr': 3.1764159166952075e-05, 'samples': 21614080, 'steps': 42214, 'loss/train': 1.3728671073913574} 02/26/2022 08:14:14 - INFO - codeparrot_training - Step 42215: {'lr': 3.175617767260616e-05, 'samples': 21614592, 'steps': 42215, 'loss/train': 1.8052622079849243} 02/26/2022 08:14:17 - INFO - codeparrot_training - Step 42216: {'lr': 3.174819711314744e-05, 'samples': 21615104, 'steps': 42216, 'loss/train': 1.9396475553512573} 02/26/2022 08:14:23 - INFO - codeparrot_training - Step 42217: {'lr': 3.1740217488610195e-05, 'samples': 21615616, 'steps': 42217, 'loss/train': 1.9824506044387817} 02/26/2022 08:14:26 - INFO - codeparrot_training - Step 42218: {'lr': 3.173223879902856e-05, 'samples': 21616128, 'steps': 42218, 'loss/train': 1.5740658044815063} 02/26/2022 08:14:32 - INFO - codeparrot_training - Step 42219: {'lr': 3.172426104443671e-05, 'samples': 21616640, 'steps': 42219, 'loss/train': 1.2283740043640137} 02/26/2022 08:14:35 - INFO - codeparrot_training - Step 42220: {'lr': 3.1716284224868744e-05, 'samples': 21617152, 'steps': 42220, 'loss/train': 1.7913917303085327} 02/26/2022 08:14:41 - INFO - codeparrot_training - Step 42221: {'lr': 3.170830834035896e-05, 'samples': 21617664, 'steps': 42221, 'loss/train': 1.6830540895462036} 02/26/2022 08:14:44 - INFO - codeparrot_training - Step 42222: {'lr': 3.170033339094147e-05, 'samples': 21618176, 'steps': 42222, 'loss/train': 1.0630455017089844} 02/26/2022 08:14:51 - INFO - codeparrot_training - Step 42223: {'lr': 3.169235937665038e-05, 'samples': 21618688, 'steps': 42223, 'loss/train': 1.574013590812683} 02/26/2022 08:14:55 - INFO - codeparrot_training - Step 42224: {'lr': 3.168438629752002e-05, 'samples': 21619200, 'steps': 42224, 'loss/train': 2.3928468227386475} 02/26/2022 08:15:00 - INFO - codeparrot_training - Step 42225: {'lr': 3.16764141535843e-05, 'samples': 21619712, 'steps': 42225, 'loss/train': 1.7158324718475342} 02/26/2022 08:15:04 - INFO - codeparrot_training - Step 42226: {'lr': 3.1668442944877556e-05, 'samples': 21620224, 'steps': 42226, 'loss/train': 1.8108742237091064} 02/26/2022 08:15:09 - INFO - codeparrot_training - Step 42227: {'lr': 3.166047267143382e-05, 'samples': 21620736, 'steps': 42227, 'loss/train': 2.149655342102051} 02/26/2022 08:15:13 - INFO - codeparrot_training - Step 42228: {'lr': 3.16525033332874e-05, 'samples': 21621248, 'steps': 42228, 'loss/train': 1.1872278451919556} 02/26/2022 08:15:18 - INFO - codeparrot_training - Step 42229: {'lr': 3.164453493047223e-05, 'samples': 21621760, 'steps': 42229, 'loss/train': 1.3453013896942139} 02/26/2022 08:15:22 - INFO - codeparrot_training - Step 42230: {'lr': 3.163656746302257e-05, 'samples': 21622272, 'steps': 42230, 'loss/train': 2.3913726806640625} 02/26/2022 08:15:27 - INFO - codeparrot_training - Step 42231: {'lr': 3.162860093097247e-05, 'samples': 21622784, 'steps': 42231, 'loss/train': 2.323552131652832} 02/26/2022 08:15:31 - INFO - codeparrot_training - Step 42232: {'lr': 3.162063533435619e-05, 'samples': 21623296, 'steps': 42232, 'loss/train': 2.199755907058716} 02/26/2022 08:15:38 - INFO - codeparrot_training - Step 42233: {'lr': 3.161267067320769e-05, 'samples': 21623808, 'steps': 42233, 'loss/train': 1.5659087896347046} 02/26/2022 08:15:41 - INFO - codeparrot_training - Step 42234: {'lr': 3.1604706947561205e-05, 'samples': 21624320, 'steps': 42234, 'loss/train': 2.040925979614258} 02/26/2022 08:15:47 - INFO - codeparrot_training - Step 42235: {'lr': 3.159674415745073e-05, 'samples': 21624832, 'steps': 42235, 'loss/train': 3.170947551727295} 02/26/2022 08:15:50 - INFO - codeparrot_training - Step 42236: {'lr': 3.158878230291054e-05, 'samples': 21625344, 'steps': 42236, 'loss/train': 0.8010342121124268} 02/26/2022 08:15:56 - INFO - codeparrot_training - Step 42237: {'lr': 3.1580821383974636e-05, 'samples': 21625856, 'steps': 42237, 'loss/train': 2.8901357650756836} 02/26/2022 08:15:59 - INFO - codeparrot_training - Step 42238: {'lr': 3.1572861400677135e-05, 'samples': 21626368, 'steps': 42238, 'loss/train': 1.9376899003982544} 02/26/2022 08:16:05 - INFO - codeparrot_training - Step 42239: {'lr': 3.156490235305209e-05, 'samples': 21626880, 'steps': 42239, 'loss/train': 1.1256794929504395} 02/26/2022 08:16:08 - INFO - codeparrot_training - Step 42240: {'lr': 3.15569442411337e-05, 'samples': 21627392, 'steps': 42240, 'loss/train': 1.6314609050750732} 02/26/2022 08:16:14 - INFO - codeparrot_training - Step 42241: {'lr': 3.1548987064955985e-05, 'samples': 21627904, 'steps': 42241, 'loss/train': 0.919796884059906} 02/26/2022 08:16:17 - INFO - codeparrot_training - Step 42242: {'lr': 3.154103082455306e-05, 'samples': 21628416, 'steps': 42242, 'loss/train': 2.8674371242523193} 02/26/2022 08:16:23 - INFO - codeparrot_training - Step 42243: {'lr': 3.153307551995893e-05, 'samples': 21628928, 'steps': 42243, 'loss/train': 2.042776107788086} 02/26/2022 08:16:28 - INFO - codeparrot_training - Step 42244: {'lr': 3.1525121151207787e-05, 'samples': 21629440, 'steps': 42244, 'loss/train': 1.126320719718933} 02/26/2022 08:16:32 - INFO - codeparrot_training - Step 42245: {'lr': 3.151716771833366e-05, 'samples': 21629952, 'steps': 42245, 'loss/train': 1.825308918952942} 02/26/2022 08:16:37 - INFO - codeparrot_training - Step 42246: {'lr': 3.15092152213706e-05, 'samples': 21630464, 'steps': 42246, 'loss/train': 1.1401276588439941} 02/26/2022 08:16:41 - INFO - codeparrot_training - Step 42247: {'lr': 3.1501263660352606e-05, 'samples': 21630976, 'steps': 42247, 'loss/train': 0.5610069036483765} 02/26/2022 08:16:48 - INFO - codeparrot_training - Step 42248: {'lr': 3.1493313035313916e-05, 'samples': 21631488, 'steps': 42248, 'loss/train': 1.1515543460845947} 02/26/2022 08:16:51 - INFO - codeparrot_training - Step 42249: {'lr': 3.14853633462884e-05, 'samples': 21632000, 'steps': 42249, 'loss/train': 1.8861671686172485} 02/26/2022 08:16:57 - INFO - codeparrot_training - Step 42250: {'lr': 3.147741459331033e-05, 'samples': 21632512, 'steps': 42250, 'loss/train': 1.9794964790344238} 02/26/2022 08:17:00 - INFO - codeparrot_training - Step 42251: {'lr': 3.146946677641352e-05, 'samples': 21633024, 'steps': 42251, 'loss/train': 1.2931498289108276} 02/26/2022 08:17:06 - INFO - codeparrot_training - Step 42252: {'lr': 3.146151989563214e-05, 'samples': 21633536, 'steps': 42252, 'loss/train': 1.604335904121399} 02/26/2022 08:17:09 - INFO - codeparrot_training - Step 42253: {'lr': 3.145357395100021e-05, 'samples': 21634048, 'steps': 42253, 'loss/train': 1.6945403814315796} 02/26/2022 08:17:15 - INFO - codeparrot_training - Step 42254: {'lr': 3.144562894255185e-05, 'samples': 21634560, 'steps': 42254, 'loss/train': 1.5309677124023438} 02/26/2022 08:17:18 - INFO - codeparrot_training - Step 42255: {'lr': 3.143768487032092e-05, 'samples': 21635072, 'steps': 42255, 'loss/train': 2.0649807453155518} 02/26/2022 08:17:24 - INFO - codeparrot_training - Step 42256: {'lr': 3.14297417343416e-05, 'samples': 21635584, 'steps': 42256, 'loss/train': 2.199768304824829} 02/26/2022 08:17:27 - INFO - codeparrot_training - Step 42257: {'lr': 3.142179953464788e-05, 'samples': 21636096, 'steps': 42257, 'loss/train': 2.436216115951538} 02/26/2022 08:17:34 - INFO - codeparrot_training - Step 42258: {'lr': 3.1413858271273746e-05, 'samples': 21636608, 'steps': 42258, 'loss/train': 1.1742217540740967} 02/26/2022 08:17:38 - INFO - codeparrot_training - Step 42259: {'lr': 3.1405917944253183e-05, 'samples': 21637120, 'steps': 42259, 'loss/train': 2.3030474185943604} 02/26/2022 08:17:41 - INFO - codeparrot_training - Step 42260: {'lr': 3.139797855362031e-05, 'samples': 21637632, 'steps': 42260, 'loss/train': 2.2223446369171143} 02/26/2022 08:17:47 - INFO - codeparrot_training - Step 42261: {'lr': 3.139004009940907e-05, 'samples': 21638144, 'steps': 42261, 'loss/train': 1.2616757154464722} 02/26/2022 08:17:50 - INFO - codeparrot_training - Step 42262: {'lr': 3.138210258165344e-05, 'samples': 21638656, 'steps': 42262, 'loss/train': 0.5822426676750183} 02/26/2022 08:17:56 - INFO - codeparrot_training - Step 42263: {'lr': 3.137416600038753e-05, 'samples': 21639168, 'steps': 42263, 'loss/train': 2.199300527572632} 02/26/2022 08:17:59 - INFO - codeparrot_training - Step 42264: {'lr': 3.136623035564526e-05, 'samples': 21639680, 'steps': 42264, 'loss/train': 1.511427879333496} 02/26/2022 08:18:05 - INFO - codeparrot_training - Step 42265: {'lr': 3.135829564746062e-05, 'samples': 21640192, 'steps': 42265, 'loss/train': 1.935234546661377} 02/26/2022 08:18:08 - INFO - codeparrot_training - Step 42266: {'lr': 3.1350361875867565e-05, 'samples': 21640704, 'steps': 42266, 'loss/train': 1.5492324829101562} 02/26/2022 08:18:14 - INFO - codeparrot_training - Step 42267: {'lr': 3.1342429040900205e-05, 'samples': 21641216, 'steps': 42267, 'loss/train': 1.8779046535491943} 02/26/2022 08:18:17 - INFO - codeparrot_training - Step 42268: {'lr': 3.133449714259243e-05, 'samples': 21641728, 'steps': 42268, 'loss/train': 3.9783482551574707} 02/26/2022 08:18:24 - INFO - codeparrot_training - Step 42269: {'lr': 3.132656618097826e-05, 'samples': 21642240, 'steps': 42269, 'loss/train': 2.906923532485962} 02/26/2022 08:18:27 - INFO - codeparrot_training - Step 42270: {'lr': 3.131863615609157e-05, 'samples': 21642752, 'steps': 42270, 'loss/train': 1.0831663608551025} 02/26/2022 08:18:33 - INFO - codeparrot_training - Step 42271: {'lr': 3.1310707067966456e-05, 'samples': 21643264, 'steps': 42271, 'loss/train': 1.9712858200073242} 02/26/2022 08:18:36 - INFO - codeparrot_training - Step 42272: {'lr': 3.1302778916636824e-05, 'samples': 21643776, 'steps': 42272, 'loss/train': 2.589186906814575} 02/26/2022 08:18:42 - INFO - codeparrot_training - Step 42273: {'lr': 3.129485170213667e-05, 'samples': 21644288, 'steps': 42273, 'loss/train': 2.5730698108673096} 02/26/2022 08:18:45 - INFO - codeparrot_training - Step 42274: {'lr': 3.128692542449985e-05, 'samples': 21644800, 'steps': 42274, 'loss/train': 2.575571298599243} 02/26/2022 08:18:51 - INFO - codeparrot_training - Step 42275: {'lr': 3.127900008376044e-05, 'samples': 21645312, 'steps': 42275, 'loss/train': 2.1145694255828857} 02/26/2022 08:18:54 - INFO - codeparrot_training - Step 42276: {'lr': 3.1271075679952355e-05, 'samples': 21645824, 'steps': 42276, 'loss/train': 0.9241102337837219} 02/26/2022 08:19:00 - INFO - codeparrot_training - Step 42277: {'lr': 3.1263152213109516e-05, 'samples': 21646336, 'steps': 42277, 'loss/train': 0.840644896030426} 02/26/2022 08:19:03 - INFO - codeparrot_training - Step 42278: {'lr': 3.1255229683265826e-05, 'samples': 21646848, 'steps': 42278, 'loss/train': 0.11633437126874924} 02/26/2022 08:19:10 - INFO - codeparrot_training - Step 42279: {'lr': 3.124730809045534e-05, 'samples': 21647360, 'steps': 42279, 'loss/train': 1.53107750415802} 02/26/2022 08:19:13 - INFO - codeparrot_training - Step 42280: {'lr': 3.1239387434711914e-05, 'samples': 21647872, 'steps': 42280, 'loss/train': 1.4984701871871948} 02/26/2022 08:19:19 - INFO - codeparrot_training - Step 42281: {'lr': 3.1231467716069494e-05, 'samples': 21648384, 'steps': 42281, 'loss/train': 0.7035585641860962} 02/26/2022 08:19:22 - INFO - codeparrot_training - Step 42282: {'lr': 3.1223548934561945e-05, 'samples': 21648896, 'steps': 42282, 'loss/train': 2.0375263690948486} 02/26/2022 08:19:28 - INFO - codeparrot_training - Step 42283: {'lr': 3.12156310902233e-05, 'samples': 21649408, 'steps': 42283, 'loss/train': 4.35195779800415} 02/26/2022 08:19:31 - INFO - codeparrot_training - Step 42284: {'lr': 3.120771418308735e-05, 'samples': 21649920, 'steps': 42284, 'loss/train': 1.8416963815689087} 02/26/2022 08:19:37 - INFO - codeparrot_training - Step 42285: {'lr': 3.119979821318819e-05, 'samples': 21650432, 'steps': 42285, 'loss/train': 1.4881956577301025} 02/26/2022 08:19:40 - INFO - codeparrot_training - Step 42286: {'lr': 3.1191883180559515e-05, 'samples': 21650944, 'steps': 42286, 'loss/train': 1.1917601823806763} 02/26/2022 08:19:46 - INFO - codeparrot_training - Step 42287: {'lr': 3.118396908523538e-05, 'samples': 21651456, 'steps': 42287, 'loss/train': 2.698991060256958} 02/26/2022 08:19:49 - INFO - codeparrot_training - Step 42288: {'lr': 3.1176055927249595e-05, 'samples': 21651968, 'steps': 42288, 'loss/train': 2.216356039047241} 02/26/2022 08:19:55 - INFO - codeparrot_training - Step 42289: {'lr': 3.11681437066362e-05, 'samples': 21652480, 'steps': 42289, 'loss/train': 0.3686668574810028} 02/26/2022 08:19:58 - INFO - codeparrot_training - Step 42290: {'lr': 3.1160232423428885e-05, 'samples': 21652992, 'steps': 42290, 'loss/train': 2.4461677074432373} 02/26/2022 08:20:04 - INFO - codeparrot_training - Step 42291: {'lr': 3.1152322077661724e-05, 'samples': 21653504, 'steps': 42291, 'loss/train': 2.5734941959381104} 02/26/2022 08:20:07 - INFO - codeparrot_training - Step 42292: {'lr': 3.114441266936843e-05, 'samples': 21654016, 'steps': 42292, 'loss/train': 1.983603596687317} 02/26/2022 08:20:13 - INFO - codeparrot_training - Step 42293: {'lr': 3.113650419858311e-05, 'samples': 21654528, 'steps': 42293, 'loss/train': 2.3230643272399902} 02/26/2022 08:20:19 - INFO - codeparrot_training - Step 42294: {'lr': 3.11285966653394e-05, 'samples': 21655040, 'steps': 42294, 'loss/train': 1.3155059814453125} 02/26/2022 08:20:23 - INFO - codeparrot_training - Step 42295: {'lr': 3.1120690069671334e-05, 'samples': 21655552, 'steps': 42295, 'loss/train': 0.4166004955768585} 02/26/2022 08:20:27 - INFO - codeparrot_training - Step 42296: {'lr': 3.1112784411612667e-05, 'samples': 21656064, 'steps': 42296, 'loss/train': 7.708520412445068} 02/26/2022 08:20:32 - INFO - codeparrot_training - Step 42297: {'lr': 3.110487969119738e-05, 'samples': 21656576, 'steps': 42297, 'loss/train': 1.7013347148895264} 02/26/2022 08:20:36 - INFO - codeparrot_training - Step 42298: {'lr': 3.1096975908459304e-05, 'samples': 21657088, 'steps': 42298, 'loss/train': 2.0086581707000732} 02/26/2022 08:20:41 - INFO - codeparrot_training - Step 42299: {'lr': 3.108907306343226e-05, 'samples': 21657600, 'steps': 42299, 'loss/train': 2.2423789501190186} 02/26/2022 08:20:45 - INFO - codeparrot_training - Step 42300: {'lr': 3.108117115615005e-05, 'samples': 21658112, 'steps': 42300, 'loss/train': 0.8812868595123291} 02/26/2022 08:20:50 - INFO - codeparrot_training - Step 42301: {'lr': 3.1073270186646655e-05, 'samples': 21658624, 'steps': 42301, 'loss/train': 1.3461356163024902} 02/26/2022 08:20:54 - INFO - codeparrot_training - Step 42302: {'lr': 3.1065370154955835e-05, 'samples': 21659136, 'steps': 42302, 'loss/train': 1.2405699491500854} 02/26/2022 08:20:59 - INFO - codeparrot_training - Step 42303: {'lr': 3.105747106111148e-05, 'samples': 21659648, 'steps': 42303, 'loss/train': 1.9656100273132324} 02/26/2022 08:21:03 - INFO - codeparrot_training - Step 42304: {'lr': 3.1049572905147375e-05, 'samples': 21660160, 'steps': 42304, 'loss/train': 1.7684097290039062} 02/26/2022 08:21:08 - INFO - codeparrot_training - Step 42305: {'lr': 3.104167568709734e-05, 'samples': 21660672, 'steps': 42305, 'loss/train': 0.8035821318626404} 02/26/2022 08:21:11 - INFO - codeparrot_training - Step 42306: {'lr': 3.103377940699528e-05, 'samples': 21661184, 'steps': 42306, 'loss/train': 2.4385085105895996} 02/26/2022 08:21:18 - INFO - codeparrot_training - Step 42307: {'lr': 3.102588406487497e-05, 'samples': 21661696, 'steps': 42307, 'loss/train': 0.8234643340110779} 02/26/2022 08:21:22 - INFO - codeparrot_training - Step 42308: {'lr': 3.101798966077024e-05, 'samples': 21662208, 'steps': 42308, 'loss/train': 1.9941766262054443} 02/26/2022 08:21:27 - INFO - codeparrot_training - Step 42309: {'lr': 3.101009619471487e-05, 'samples': 21662720, 'steps': 42309, 'loss/train': 0.6015418767929077} 02/26/2022 08:21:33 - INFO - codeparrot_training - Step 42310: {'lr': 3.100220366674275e-05, 'samples': 21663232, 'steps': 42310, 'loss/train': 1.7579880952835083} 02/26/2022 08:21:36 - INFO - codeparrot_training - Step 42311: {'lr': 3.099431207688769e-05, 'samples': 21663744, 'steps': 42311, 'loss/train': 1.1578298807144165} 02/26/2022 08:21:42 - INFO - codeparrot_training - Step 42312: {'lr': 3.0986421425183403e-05, 'samples': 21664256, 'steps': 42312, 'loss/train': 1.647736668586731} 02/26/2022 08:21:45 - INFO - codeparrot_training - Step 42313: {'lr': 3.0978531711663706e-05, 'samples': 21664768, 'steps': 42313, 'loss/train': 2.0021708011627197} 02/26/2022 08:21:52 - INFO - codeparrot_training - Step 42314: {'lr': 3.097064293636251e-05, 'samples': 21665280, 'steps': 42314, 'loss/train': 1.823815941810608} 02/26/2022 08:21:55 - INFO - codeparrot_training - Step 42315: {'lr': 3.0962755099313486e-05, 'samples': 21665792, 'steps': 42315, 'loss/train': 2.05018949508667} 02/26/2022 08:22:01 - INFO - codeparrot_training - Step 42316: {'lr': 3.095486820055049e-05, 'samples': 21666304, 'steps': 42316, 'loss/train': 1.6738413572311401} 02/26/2022 08:22:05 - INFO - codeparrot_training - Step 42317: {'lr': 3.0946982240107255e-05, 'samples': 21666816, 'steps': 42317, 'loss/train': 1.5243638753890991} 02/26/2022 08:22:08 - INFO - codeparrot_training - Step 42318: {'lr': 3.09390972180176e-05, 'samples': 21667328, 'steps': 42318, 'loss/train': 0.03138976916670799} 02/26/2022 08:22:14 - INFO - codeparrot_training - Step 42319: {'lr': 3.093121313431527e-05, 'samples': 21667840, 'steps': 42319, 'loss/train': 0.18562248349189758} 02/26/2022 08:22:17 - INFO - codeparrot_training - Step 42320: {'lr': 3.092332998903416e-05, 'samples': 21668352, 'steps': 42320, 'loss/train': 1.573781132698059} 02/26/2022 08:22:23 - INFO - codeparrot_training - Step 42321: {'lr': 3.091544778220781e-05, 'samples': 21668864, 'steps': 42321, 'loss/train': 2.235450029373169} 02/26/2022 08:22:26 - INFO - codeparrot_training - Step 42322: {'lr': 3.090756651387019e-05, 'samples': 21669376, 'steps': 42322, 'loss/train': 1.8926372528076172} 02/26/2022 08:22:32 - INFO - codeparrot_training - Step 42323: {'lr': 3.089968618405492e-05, 'samples': 21669888, 'steps': 42323, 'loss/train': 1.7123184204101562} 02/26/2022 08:22:35 - INFO - codeparrot_training - Step 42324: {'lr': 3.089180679279596e-05, 'samples': 21670400, 'steps': 42324, 'loss/train': 1.0564671754837036} 02/26/2022 08:22:41 - INFO - codeparrot_training - Step 42325: {'lr': 3.088392834012679e-05, 'samples': 21670912, 'steps': 42325, 'loss/train': 1.7197434902191162} 02/26/2022 08:22:44 - INFO - codeparrot_training - Step 42326: {'lr': 3.087605082608136e-05, 'samples': 21671424, 'steps': 42326, 'loss/train': 1.489651083946228} 02/26/2022 08:22:50 - INFO - codeparrot_training - Step 42327: {'lr': 3.086817425069327e-05, 'samples': 21671936, 'steps': 42327, 'loss/train': 1.2963221073150635} 02/26/2022 08:22:55 - INFO - codeparrot_training - Step 42328: {'lr': 3.086029861399647e-05, 'samples': 21672448, 'steps': 42328, 'loss/train': 2.574021100997925} 02/26/2022 08:22:59 - INFO - codeparrot_training - Step 42329: {'lr': 3.085242391602447e-05, 'samples': 21672960, 'steps': 42329, 'loss/train': 1.0760599374771118} 02/26/2022 08:23:05 - INFO - codeparrot_training - Step 42330: {'lr': 3.084455015681112e-05, 'samples': 21673472, 'steps': 42330, 'loss/train': 1.3865455389022827} 02/26/2022 08:23:08 - INFO - codeparrot_training - Step 42331: {'lr': 3.083667733639009e-05, 'samples': 21673984, 'steps': 42331, 'loss/train': 1.760359287261963} 02/26/2022 08:23:14 - INFO - codeparrot_training - Step 42332: {'lr': 3.082880545479519e-05, 'samples': 21674496, 'steps': 42332, 'loss/train': 1.8878101110458374} 02/26/2022 08:23:17 - INFO - codeparrot_training - Step 42333: {'lr': 3.082093451206011e-05, 'samples': 21675008, 'steps': 42333, 'loss/train': 2.0384135246276855} 02/26/2022 08:23:25 - INFO - codeparrot_training - Step 42334: {'lr': 3.0813064508218555e-05, 'samples': 21675520, 'steps': 42334, 'loss/train': 0.2993811070919037} 02/26/2022 08:23:28 - INFO - codeparrot_training - Step 42335: {'lr': 3.080519544330415e-05, 'samples': 21676032, 'steps': 42335, 'loss/train': 2.2004995346069336} 02/26/2022 08:23:34 - INFO - codeparrot_training - Step 42336: {'lr': 3.0797327317350744e-05, 'samples': 21676544, 'steps': 42336, 'loss/train': 1.6180719137191772} 02/26/2022 08:23:38 - INFO - codeparrot_training - Step 42337: {'lr': 3.078946013039199e-05, 'samples': 21677056, 'steps': 42337, 'loss/train': 1.9453177452087402} 02/26/2022 08:23:43 - INFO - codeparrot_training - Step 42338: {'lr': 3.0781593882461575e-05, 'samples': 21677568, 'steps': 42338, 'loss/train': 1.849726915359497} 02/26/2022 08:23:46 - INFO - codeparrot_training - Step 42339: {'lr': 3.077372857359317e-05, 'samples': 21678080, 'steps': 42339, 'loss/train': 0.7992461919784546} 02/26/2022 08:23:52 - INFO - codeparrot_training - Step 42340: {'lr': 3.076586420382052e-05, 'samples': 21678592, 'steps': 42340, 'loss/train': 2.3807413578033447} 02/26/2022 08:23:55 - INFO - codeparrot_training - Step 42341: {'lr': 3.075800077317731e-05, 'samples': 21679104, 'steps': 42341, 'loss/train': 1.4038259983062744} 02/26/2022 08:24:01 - INFO - codeparrot_training - Step 42342: {'lr': 3.075013828169723e-05, 'samples': 21679616, 'steps': 42342, 'loss/train': 0.455833375453949} 02/26/2022 08:24:05 - INFO - codeparrot_training - Step 42343: {'lr': 3.074227672941385e-05, 'samples': 21680128, 'steps': 42343, 'loss/train': 2.2301933765411377} 02/26/2022 08:24:08 - INFO - codeparrot_training - Step 42344: {'lr': 3.0734416116360994e-05, 'samples': 21680640, 'steps': 42344, 'loss/train': 2.6922214031219482} 02/26/2022 08:24:16 - INFO - codeparrot_training - Step 42345: {'lr': 3.07265564425723e-05, 'samples': 21681152, 'steps': 42345, 'loss/train': 2.2362728118896484} 02/26/2022 08:24:20 - INFO - codeparrot_training - Step 42346: {'lr': 3.0718697708081386e-05, 'samples': 21681664, 'steps': 42346, 'loss/train': 1.0965604782104492} 02/26/2022 08:24:25 - INFO - codeparrot_training - Step 42347: {'lr': 3.0710839912921896e-05, 'samples': 21682176, 'steps': 42347, 'loss/train': 2.840318202972412} 02/26/2022 08:24:29 - INFO - codeparrot_training - Step 42348: {'lr': 3.070298305712757e-05, 'samples': 21682688, 'steps': 42348, 'loss/train': 0.7933400869369507} 02/26/2022 08:24:34 - INFO - codeparrot_training - Step 42349: {'lr': 3.0695127140732055e-05, 'samples': 21683200, 'steps': 42349, 'loss/train': 2.0706417560577393} 02/26/2022 08:24:37 - INFO - codeparrot_training - Step 42350: {'lr': 3.0687272163768986e-05, 'samples': 21683712, 'steps': 42350, 'loss/train': 2.36657452583313} 02/26/2022 08:24:43 - INFO - codeparrot_training - Step 42351: {'lr': 3.0679418126271923e-05, 'samples': 21684224, 'steps': 42351, 'loss/train': 1.8002938032150269} 02/26/2022 08:24:47 - INFO - codeparrot_training - Step 42352: {'lr': 3.067156502827465e-05, 'samples': 21684736, 'steps': 42352, 'loss/train': 1.7515136003494263} 02/26/2022 08:24:52 - INFO - codeparrot_training - Step 42353: {'lr': 3.066371286981076e-05, 'samples': 21685248, 'steps': 42353, 'loss/train': 1.3350982666015625} 02/26/2022 08:24:56 - INFO - codeparrot_training - Step 42354: {'lr': 3.065586165091386e-05, 'samples': 21685760, 'steps': 42354, 'loss/train': 1.6943905353546143} 02/26/2022 08:25:03 - INFO - codeparrot_training - Step 42355: {'lr': 3.064801137161763e-05, 'samples': 21686272, 'steps': 42355, 'loss/train': 1.2283161878585815} 02/26/2022 08:25:06 - INFO - codeparrot_training - Step 42356: {'lr': 3.064016203195558e-05, 'samples': 21686784, 'steps': 42356, 'loss/train': 2.336106777191162} 02/26/2022 08:25:12 - INFO - codeparrot_training - Step 42357: {'lr': 3.063231363196148e-05, 'samples': 21687296, 'steps': 42357, 'loss/train': 0.7059616446495056} 02/26/2022 08:25:15 - INFO - codeparrot_training - Step 42358: {'lr': 3.0624466171668864e-05, 'samples': 21687808, 'steps': 42358, 'loss/train': 1.272324800491333} 02/26/2022 08:25:21 - INFO - codeparrot_training - Step 42359: {'lr': 3.0616619651111455e-05, 'samples': 21688320, 'steps': 42359, 'loss/train': 1.4218025207519531} 02/26/2022 08:25:24 - INFO - codeparrot_training - Step 42360: {'lr': 3.06087740703227e-05, 'samples': 21688832, 'steps': 42360, 'loss/train': 1.0767157077789307} 02/26/2022 08:25:28 - INFO - codeparrot_training - Step 42361: {'lr': 3.0600929429336315e-05, 'samples': 21689344, 'steps': 42361, 'loss/train': 1.4596195220947266} 02/26/2022 08:25:33 - INFO - codeparrot_training - Step 42362: {'lr': 3.0593085728185865e-05, 'samples': 21689856, 'steps': 42362, 'loss/train': 1.6313985586166382} 02/26/2022 08:25:39 - INFO - codeparrot_training - Step 42363: {'lr': 3.0585242966905045e-05, 'samples': 21690368, 'steps': 42363, 'loss/train': 1.8777227401733398} 02/26/2022 08:25:42 - INFO - codeparrot_training - Step 42364: {'lr': 3.057740114552726e-05, 'samples': 21690880, 'steps': 42364, 'loss/train': 2.151515007019043} 02/26/2022 08:25:49 - INFO - codeparrot_training - Step 42365: {'lr': 3.05695602640863e-05, 'samples': 21691392, 'steps': 42365, 'loss/train': 1.2571848630905151} 02/26/2022 08:25:53 - INFO - codeparrot_training - Step 42366: {'lr': 3.056172032261559e-05, 'samples': 21691904, 'steps': 42366, 'loss/train': 1.419875979423523} 02/26/2022 08:25:58 - INFO - codeparrot_training - Step 42367: {'lr': 3.0553881321148836e-05, 'samples': 21692416, 'steps': 42367, 'loss/train': 1.7723785638809204} 02/26/2022 08:26:02 - INFO - codeparrot_training - Step 42368: {'lr': 3.05460432597196e-05, 'samples': 21692928, 'steps': 42368, 'loss/train': 1.3066699504852295} 02/26/2022 08:26:07 - INFO - codeparrot_training - Step 42369: {'lr': 3.053820613836142e-05, 'samples': 21693440, 'steps': 42369, 'loss/train': 0.8458231091499329} 02/26/2022 08:26:11 - INFO - codeparrot_training - Step 42370: {'lr': 3.053036995710781e-05, 'samples': 21693952, 'steps': 42370, 'loss/train': 1.9319634437561035} 02/26/2022 08:26:16 - INFO - codeparrot_training - Step 42371: {'lr': 3.0522534715992476e-05, 'samples': 21694464, 'steps': 42371, 'loss/train': 2.048814058303833} 02/26/2022 08:26:20 - INFO - codeparrot_training - Step 42372: {'lr': 3.051470041504889e-05, 'samples': 21694976, 'steps': 42372, 'loss/train': 1.852872371673584} 02/26/2022 08:26:25 - INFO - codeparrot_training - Step 42373: {'lr': 3.0506867054310662e-05, 'samples': 21695488, 'steps': 42373, 'loss/train': 1.510689377784729} 02/26/2022 08:26:29 - INFO - codeparrot_training - Step 42374: {'lr': 3.0499034633811234e-05, 'samples': 21696000, 'steps': 42374, 'loss/train': 2.0618083477020264} 02/26/2022 08:26:36 - INFO - codeparrot_training - Step 42375: {'lr': 3.0491203153584334e-05, 'samples': 21696512, 'steps': 42375, 'loss/train': 1.0150772333145142} 02/26/2022 08:26:39 - INFO - codeparrot_training - Step 42376: {'lr': 3.048337261366338e-05, 'samples': 21697024, 'steps': 42376, 'loss/train': 2.3928732872009277} 02/26/2022 08:26:45 - INFO - codeparrot_training - Step 42377: {'lr': 3.047554301408198e-05, 'samples': 21697536, 'steps': 42377, 'loss/train': 1.7462522983551025} 02/26/2022 08:26:49 - INFO - codeparrot_training - Step 42378: {'lr': 3.0467714354873582e-05, 'samples': 21698048, 'steps': 42378, 'loss/train': 2.236633777618408} 02/26/2022 08:26:52 - INFO - codeparrot_training - Step 42379: {'lr': 3.0459886636071827e-05, 'samples': 21698560, 'steps': 42379, 'loss/train': 0.3334674537181854} 02/26/2022 08:26:58 - INFO - codeparrot_training - Step 42380: {'lr': 3.0452059857710184e-05, 'samples': 21699072, 'steps': 42380, 'loss/train': 1.3294657468795776} 02/26/2022 08:27:01 - INFO - codeparrot_training - Step 42381: {'lr': 3.044423401982227e-05, 'samples': 21699584, 'steps': 42381, 'loss/train': 0.7406435012817383} 02/26/2022 08:27:07 - INFO - codeparrot_training - Step 42382: {'lr': 3.0436409122441472e-05, 'samples': 21700096, 'steps': 42382, 'loss/train': 1.9715319871902466} 02/26/2022 08:27:10 - INFO - codeparrot_training - Step 42383: {'lr': 3.0428585165601403e-05, 'samples': 21700608, 'steps': 42383, 'loss/train': 1.4250870943069458} 02/26/2022 08:27:16 - INFO - codeparrot_training - Step 42384: {'lr': 3.0420762149335562e-05, 'samples': 21701120, 'steps': 42384, 'loss/train': 0.9443029165267944} 02/26/2022 08:27:19 - INFO - codeparrot_training - Step 42385: {'lr': 3.0412940073677454e-05, 'samples': 21701632, 'steps': 42385, 'loss/train': 1.3233338594436646} 02/26/2022 08:27:26 - INFO - codeparrot_training - Step 42386: {'lr': 3.0405118938660547e-05, 'samples': 21702144, 'steps': 42386, 'loss/train': 2.2087249755859375} 02/26/2022 08:27:30 - INFO - codeparrot_training - Step 42387: {'lr': 3.0397298744318404e-05, 'samples': 21702656, 'steps': 42387, 'loss/train': 1.8795509338378906} 02/26/2022 08:27:35 - INFO - codeparrot_training - Step 42388: {'lr': 3.038947949068452e-05, 'samples': 21703168, 'steps': 42388, 'loss/train': 2.835408926010132} 02/26/2022 08:27:39 - INFO - codeparrot_training - Step 42389: {'lr': 3.038166117779237e-05, 'samples': 21703680, 'steps': 42389, 'loss/train': 1.5740962028503418} 02/26/2022 08:27:44 - INFO - codeparrot_training - Step 42390: {'lr': 3.03738438056754e-05, 'samples': 21704192, 'steps': 42390, 'loss/train': 1.1738770008087158} 02/26/2022 08:27:48 - INFO - codeparrot_training - Step 42391: {'lr': 3.0366027374367195e-05, 'samples': 21704704, 'steps': 42391, 'loss/train': 1.4663310050964355} 02/26/2022 08:27:53 - INFO - codeparrot_training - Step 42392: {'lr': 3.035821188390117e-05, 'samples': 21705216, 'steps': 42392, 'loss/train': 1.115710973739624} 02/26/2022 08:27:57 - INFO - codeparrot_training - Step 42393: {'lr': 3.0350397334310774e-05, 'samples': 21705728, 'steps': 42393, 'loss/train': 2.5159871578216553} 02/26/2022 08:28:02 - INFO - codeparrot_training - Step 42394: {'lr': 3.0342583725629592e-05, 'samples': 21706240, 'steps': 42394, 'loss/train': 1.4024797677993774} 02/26/2022 08:28:06 - INFO - codeparrot_training - Step 42395: {'lr': 3.0334771057891036e-05, 'samples': 21706752, 'steps': 42395, 'loss/train': 2.4171509742736816} 02/26/2022 08:28:11 - INFO - codeparrot_training - Step 42396: {'lr': 3.0326959331128558e-05, 'samples': 21707264, 'steps': 42396, 'loss/train': 1.453932285308838} 02/26/2022 08:28:17 - INFO - codeparrot_training - Step 42397: {'lr': 3.031914854537557e-05, 'samples': 21707776, 'steps': 42397, 'loss/train': 1.4566223621368408} 02/26/2022 08:28:20 - INFO - codeparrot_training - Step 42398: {'lr': 3.0311338700665657e-05, 'samples': 21708288, 'steps': 42398, 'loss/train': 1.837753176689148} 02/26/2022 08:28:26 - INFO - codeparrot_training - Step 42399: {'lr': 3.0303529797032214e-05, 'samples': 21708800, 'steps': 42399, 'loss/train': 2.5541043281555176} 02/26/2022 08:28:29 - INFO - codeparrot_training - Step 42400: {'lr': 3.0295721834508684e-05, 'samples': 21709312, 'steps': 42400, 'loss/train': 1.472905158996582} 02/26/2022 08:28:37 - INFO - codeparrot_training - Step 42401: {'lr': 3.0287914813128482e-05, 'samples': 21709824, 'steps': 42401, 'loss/train': 1.9961215257644653} 02/26/2022 08:28:40 - INFO - codeparrot_training - Step 42402: {'lr': 3.0280108732925165e-05, 'samples': 21710336, 'steps': 42402, 'loss/train': 1.735489010810852} 02/26/2022 08:28:46 - INFO - codeparrot_training - Step 42403: {'lr': 3.0272303593931987e-05, 'samples': 21710848, 'steps': 42403, 'loss/train': 0.4769146144390106} 02/26/2022 08:28:50 - INFO - codeparrot_training - Step 42404: {'lr': 3.026449939618256e-05, 'samples': 21711360, 'steps': 42404, 'loss/train': 2.2279601097106934} 02/26/2022 08:28:53 - INFO - codeparrot_training - Step 42405: {'lr': 3.0256696139710187e-05, 'samples': 21711872, 'steps': 42405, 'loss/train': 2.2474365234375} 02/26/2022 08:28:59 - INFO - codeparrot_training - Step 42406: {'lr': 3.0248893824548402e-05, 'samples': 21712384, 'steps': 42406, 'loss/train': 2.2075304985046387} 02/26/2022 08:29:02 - INFO - codeparrot_training - Step 42407: {'lr': 3.024109245073056e-05, 'samples': 21712896, 'steps': 42407, 'loss/train': 2.376878499984741} 02/26/2022 08:29:08 - INFO - codeparrot_training - Step 42408: {'lr': 3.0233292018290113e-05, 'samples': 21713408, 'steps': 42408, 'loss/train': 2.2249040603637695} 02/26/2022 08:29:11 - INFO - codeparrot_training - Step 42409: {'lr': 3.0225492527260424e-05, 'samples': 21713920, 'steps': 42409, 'loss/train': 2.0526816844940186} 02/26/2022 08:29:17 - INFO - codeparrot_training - Step 42410: {'lr': 3.021769397767496e-05, 'samples': 21714432, 'steps': 42410, 'loss/train': 8.584919929504395} 02/26/2022 08:29:20 - INFO - codeparrot_training - Step 42411: {'lr': 3.0209896369567142e-05, 'samples': 21714944, 'steps': 42411, 'loss/train': 1.173508644104004} 02/26/2022 08:29:26 - INFO - codeparrot_training - Step 42412: {'lr': 3.0202099702970303e-05, 'samples': 21715456, 'steps': 42412, 'loss/train': 0.9102339148521423} 02/26/2022 08:29:29 - INFO - codeparrot_training - Step 42413: {'lr': 3.0194303977917835e-05, 'samples': 21715968, 'steps': 42413, 'loss/train': 1.3327618837356567} 02/26/2022 08:29:37 - INFO - codeparrot_training - Step 42414: {'lr': 3.0186509194443214e-05, 'samples': 21716480, 'steps': 42414, 'loss/train': 2.7832436561584473} 02/26/2022 08:29:40 - INFO - codeparrot_training - Step 42415: {'lr': 3.0178715352579793e-05, 'samples': 21716992, 'steps': 42415, 'loss/train': 1.7575819492340088} 02/26/2022 08:29:46 - INFO - codeparrot_training - Step 42416: {'lr': 3.017092245236097e-05, 'samples': 21717504, 'steps': 42416, 'loss/train': 1.674803614616394} 02/26/2022 08:29:49 - INFO - codeparrot_training - Step 42417: {'lr': 3.016313049382005e-05, 'samples': 21718016, 'steps': 42417, 'loss/train': 1.9499880075454712} 02/26/2022 08:29:55 - INFO - codeparrot_training - Step 42418: {'lr': 3.0155339476990533e-05, 'samples': 21718528, 'steps': 42418, 'loss/train': 0.9251204133033752} 02/26/2022 08:29:58 - INFO - codeparrot_training - Step 42419: {'lr': 3.014754940190567e-05, 'samples': 21719040, 'steps': 42419, 'loss/train': 1.6484394073486328} 02/26/2022 08:30:04 - INFO - codeparrot_training - Step 42420: {'lr': 3.0139760268598992e-05, 'samples': 21719552, 'steps': 42420, 'loss/train': 2.097637176513672} 02/26/2022 08:30:07 - INFO - codeparrot_training - Step 42421: {'lr': 3.0131972077103694e-05, 'samples': 21720064, 'steps': 42421, 'loss/train': 1.2741641998291016} 02/26/2022 08:30:12 - INFO - codeparrot_training - Step 42422: {'lr': 3.0124184827453244e-05, 'samples': 21720576, 'steps': 42422, 'loss/train': 2.612875461578369} 02/26/2022 08:30:16 - INFO - codeparrot_training - Step 42423: {'lr': 3.01163985196809e-05, 'samples': 21721088, 'steps': 42423, 'loss/train': 1.790881872177124} 02/26/2022 08:30:23 - INFO - codeparrot_training - Step 42424: {'lr': 3.0108613153820215e-05, 'samples': 21721600, 'steps': 42424, 'loss/train': 2.145505428314209} 02/26/2022 08:30:27 - INFO - codeparrot_training - Step 42425: {'lr': 3.0100828729904274e-05, 'samples': 21722112, 'steps': 42425, 'loss/train': 2.0099036693573} 02/26/2022 08:30:32 - INFO - codeparrot_training - Step 42426: {'lr': 3.0093045247966633e-05, 'samples': 21722624, 'steps': 42426, 'loss/train': 1.2753727436065674} 02/26/2022 08:30:35 - INFO - codeparrot_training - Step 42427: {'lr': 3.008526270804049e-05, 'samples': 21723136, 'steps': 42427, 'loss/train': 1.5990394353866577} 02/26/2022 08:30:41 - INFO - codeparrot_training - Step 42428: {'lr': 3.0077481110159317e-05, 'samples': 21723648, 'steps': 42428, 'loss/train': 1.1991422176361084} 02/26/2022 08:30:45 - INFO - codeparrot_training - Step 42429: {'lr': 3.0069700454356392e-05, 'samples': 21724160, 'steps': 42429, 'loss/train': 1.2713897228240967} 02/26/2022 08:30:50 - INFO - codeparrot_training - Step 42430: {'lr': 3.0061920740664995e-05, 'samples': 21724672, 'steps': 42430, 'loss/train': 2.3636350631713867} 02/26/2022 08:30:53 - INFO - codeparrot_training - Step 42431: {'lr': 3.005414196911846e-05, 'samples': 21725184, 'steps': 42431, 'loss/train': 1.3054356575012207} 02/26/2022 08:30:59 - INFO - codeparrot_training - Step 42432: {'lr': 3.0046364139750203e-05, 'samples': 21725696, 'steps': 42432, 'loss/train': 1.9482861757278442} 02/26/2022 08:31:02 - INFO - codeparrot_training - Step 42433: {'lr': 3.003858725259348e-05, 'samples': 21726208, 'steps': 42433, 'loss/train': 1.681530475616455} 02/26/2022 08:31:10 - INFO - codeparrot_training - Step 42434: {'lr': 3.0030811307681595e-05, 'samples': 21726720, 'steps': 42434, 'loss/train': 0.826411783695221} 02/26/2022 08:31:13 - INFO - codeparrot_training - Step 42435: {'lr': 3.0023036305047853e-05, 'samples': 21727232, 'steps': 42435, 'loss/train': 1.230531096458435} 02/26/2022 08:31:19 - INFO - codeparrot_training - Step 42436: {'lr': 3.001526224472553e-05, 'samples': 21727744, 'steps': 42436, 'loss/train': 1.088001012802124} 02/26/2022 08:31:22 - INFO - codeparrot_training - Step 42437: {'lr': 3.000748912674803e-05, 'samples': 21728256, 'steps': 42437, 'loss/train': 1.0397390127182007} 02/26/2022 08:31:28 - INFO - codeparrot_training - Step 42438: {'lr': 2.9999716951148588e-05, 'samples': 21728768, 'steps': 42438, 'loss/train': 1.8594697713851929} 02/26/2022 08:31:31 - INFO - codeparrot_training - Step 42439: {'lr': 2.9991945717960518e-05, 'samples': 21729280, 'steps': 42439, 'loss/train': 1.3108189105987549} 02/26/2022 08:31:37 - INFO - codeparrot_training - Step 42440: {'lr': 2.9984175427217013e-05, 'samples': 21729792, 'steps': 42440, 'loss/train': 1.2480056285858154} 02/26/2022 08:31:40 - INFO - codeparrot_training - Step 42441: {'lr': 2.9976406078951497e-05, 'samples': 21730304, 'steps': 42441, 'loss/train': 1.4303146600723267} 02/26/2022 08:31:46 - INFO - codeparrot_training - Step 42442: {'lr': 2.9968637673197186e-05, 'samples': 21730816, 'steps': 42442, 'loss/train': 2.3805794715881348} 02/26/2022 08:31:51 - INFO - codeparrot_training - Step 42443: {'lr': 2.996087020998739e-05, 'samples': 21731328, 'steps': 42443, 'loss/train': 0.4328295886516571} 02/26/2022 08:31:55 - INFO - codeparrot_training - Step 42444: {'lr': 2.9953103689355275e-05, 'samples': 21731840, 'steps': 42444, 'loss/train': 1.1838759183883667} 02/26/2022 08:32:00 - INFO - codeparrot_training - Step 42445: {'lr': 2.9945338111334265e-05, 'samples': 21732352, 'steps': 42445, 'loss/train': 1.1324834823608398} 02/26/2022 08:32:03 - INFO - codeparrot_training - Step 42446: {'lr': 2.993757347595752e-05, 'samples': 21732864, 'steps': 42446, 'loss/train': 0.7343243956565857} 02/26/2022 08:32:09 - INFO - codeparrot_training - Step 42447: {'lr': 2.992980978325835e-05, 'samples': 21733376, 'steps': 42447, 'loss/train': 1.7889782190322876} 02/26/2022 08:32:12 - INFO - codeparrot_training - Step 42448: {'lr': 2.992204703326995e-05, 'samples': 21733888, 'steps': 42448, 'loss/train': 0.775888204574585} 02/26/2022 08:32:20 - INFO - codeparrot_training - Step 42449: {'lr': 2.9914285226025655e-05, 'samples': 21734400, 'steps': 42449, 'loss/train': 0.9927094578742981} 02/26/2022 08:32:23 - INFO - codeparrot_training - Step 42450: {'lr': 2.990652436155869e-05, 'samples': 21734912, 'steps': 42450, 'loss/train': 0.948738694190979} 02/26/2022 08:32:29 - INFO - codeparrot_training - Step 42451: {'lr': 2.9898764439902303e-05, 'samples': 21735424, 'steps': 42451, 'loss/train': 1.452772855758667} 02/26/2022 08:32:32 - INFO - codeparrot_training - Step 42452: {'lr': 2.9891005461089638e-05, 'samples': 21735936, 'steps': 42452, 'loss/train': 0.7828855514526367} 02/26/2022 08:32:36 - INFO - codeparrot_training - Step 42453: {'lr': 2.988324742515408e-05, 'samples': 21736448, 'steps': 42453, 'loss/train': 2.955240249633789} 02/26/2022 08:32:41 - INFO - codeparrot_training - Step 42454: {'lr': 2.9875490332128718e-05, 'samples': 21736960, 'steps': 42454, 'loss/train': 2.041384696960449} 02/26/2022 08:32:45 - INFO - codeparrot_training - Step 42455: {'lr': 2.9867734182046996e-05, 'samples': 21737472, 'steps': 42455, 'loss/train': 2.191662073135376} 02/26/2022 08:32:50 - INFO - codeparrot_training - Step 42456: {'lr': 2.9859978974941887e-05, 'samples': 21737984, 'steps': 42456, 'loss/train': 0.8403798937797546} 02/26/2022 08:32:56 - INFO - codeparrot_training - Step 42457: {'lr': 2.985222471084678e-05, 'samples': 21738496, 'steps': 42457, 'loss/train': 1.6946067810058594} 02/26/2022 08:32:59 - INFO - codeparrot_training - Step 42458: {'lr': 2.9844471389794763e-05, 'samples': 21739008, 'steps': 42458, 'loss/train': 1.6430662870407104} 02/26/2022 08:33:07 - INFO - codeparrot_training - Step 42459: {'lr': 2.983671901181925e-05, 'samples': 21739520, 'steps': 42459, 'loss/train': 2.090486764907837} 02/26/2022 08:33:10 - INFO - codeparrot_training - Step 42460: {'lr': 2.982896757695322e-05, 'samples': 21740032, 'steps': 42460, 'loss/train': 3.746598482131958} 02/26/2022 08:33:16 - INFO - codeparrot_training - Step 42461: {'lr': 2.9821217085230002e-05, 'samples': 21740544, 'steps': 42461, 'loss/train': 1.7544538974761963} 02/26/2022 08:33:19 - INFO - codeparrot_training - Step 42462: {'lr': 2.9813467536682764e-05, 'samples': 21741056, 'steps': 42462, 'loss/train': 1.0944174528121948} 02/26/2022 08:33:25 - INFO - codeparrot_training - Step 42463: {'lr': 2.9805718931344788e-05, 'samples': 21741568, 'steps': 42463, 'loss/train': 1.6205421686172485} 02/26/2022 08:33:28 - INFO - codeparrot_training - Step 42464: {'lr': 2.9797971269249103e-05, 'samples': 21742080, 'steps': 42464, 'loss/train': 0.6108075380325317} 02/26/2022 08:33:33 - INFO - codeparrot_training - Step 42465: {'lr': 2.979022455042904e-05, 'samples': 21742592, 'steps': 42465, 'loss/train': 1.625720739364624} 02/26/2022 08:33:37 - INFO - codeparrot_training - Step 42466: {'lr': 2.978247877491769e-05, 'samples': 21743104, 'steps': 42466, 'loss/train': 1.8870893716812134} 02/26/2022 08:33:42 - INFO - codeparrot_training - Step 42467: {'lr': 2.9774733942748295e-05, 'samples': 21743616, 'steps': 42467, 'loss/train': 1.022445559501648} 02/26/2022 08:33:46 - INFO - codeparrot_training - Step 42468: {'lr': 2.976699005395403e-05, 'samples': 21744128, 'steps': 42468, 'loss/train': 1.6900724172592163} 02/26/2022 08:33:53 - INFO - codeparrot_training - Step 42469: {'lr': 2.975924710856803e-05, 'samples': 21744640, 'steps': 42469, 'loss/train': 1.018027424812317} 02/26/2022 08:33:56 - INFO - codeparrot_training - Step 42470: {'lr': 2.9751505106623472e-05, 'samples': 21745152, 'steps': 42470, 'loss/train': 0.714512825012207} 02/26/2022 08:34:02 - INFO - codeparrot_training - Step 42471: {'lr': 2.9743764048153545e-05, 'samples': 21745664, 'steps': 42471, 'loss/train': 2.2454349994659424} 02/26/2022 08:34:06 - INFO - codeparrot_training - Step 42472: {'lr': 2.9736023933191388e-05, 'samples': 21746176, 'steps': 42472, 'loss/train': 1.4878127574920654} 02/26/2022 08:34:11 - INFO - codeparrot_training - Step 42473: {'lr': 2.9728284761770198e-05, 'samples': 21746688, 'steps': 42473, 'loss/train': 0.6094620227813721} 02/26/2022 08:34:15 - INFO - codeparrot_training - Step 42474: {'lr': 2.9720546533923004e-05, 'samples': 21747200, 'steps': 42474, 'loss/train': 1.5329066514968872} 02/26/2022 08:34:20 - INFO - codeparrot_training - Step 42475: {'lr': 2.9712809249683116e-05, 'samples': 21747712, 'steps': 42475, 'loss/train': 1.4630742073059082} 02/26/2022 08:34:24 - INFO - codeparrot_training - Step 42476: {'lr': 2.9705072909083587e-05, 'samples': 21748224, 'steps': 42476, 'loss/train': 1.295642614364624} 02/26/2022 08:34:29 - INFO - codeparrot_training - Step 42477: {'lr': 2.9697337512157586e-05, 'samples': 21748736, 'steps': 42477, 'loss/train': 2.0840306282043457} 02/26/2022 08:34:33 - INFO - codeparrot_training - Step 42478: {'lr': 2.96896030589382e-05, 'samples': 21749248, 'steps': 42478, 'loss/train': 2.4006764888763428} 02/26/2022 08:34:40 - INFO - codeparrot_training - Step 42479: {'lr': 2.9681869549458617e-05, 'samples': 21749760, 'steps': 42479, 'loss/train': 1.8804248571395874} 02/26/2022 08:34:43 - INFO - codeparrot_training - Step 42480: {'lr': 2.967413698375196e-05, 'samples': 21750272, 'steps': 42480, 'loss/train': 2.096904993057251} 02/26/2022 08:34:49 - INFO - codeparrot_training - Step 42481: {'lr': 2.9666405361851335e-05, 'samples': 21750784, 'steps': 42481, 'loss/train': 0.3582897484302521} 02/26/2022 08:34:52 - INFO - codeparrot_training - Step 42482: {'lr': 2.965867468378988e-05, 'samples': 21751296, 'steps': 42482, 'loss/train': 1.8962023258209229} 02/26/2022 08:34:58 - INFO - codeparrot_training - Step 42483: {'lr': 2.9650944949600632e-05, 'samples': 21751808, 'steps': 42483, 'loss/train': 1.4463261365890503} 02/26/2022 08:35:01 - INFO - codeparrot_training - Step 42484: {'lr': 2.964321615931684e-05, 'samples': 21752320, 'steps': 42484, 'loss/train': 3.009766101837158} 02/26/2022 08:35:07 - INFO - codeparrot_training - Step 42485: {'lr': 2.96354883129715e-05, 'samples': 21752832, 'steps': 42485, 'loss/train': 1.0305395126342773} 02/26/2022 08:35:10 - INFO - codeparrot_training - Step 42486: {'lr': 2.9627761410597782e-05, 'samples': 21753344, 'steps': 42486, 'loss/train': 1.8539543151855469} 02/26/2022 08:35:16 - INFO - codeparrot_training - Step 42487: {'lr': 2.962003545222869e-05, 'samples': 21753856, 'steps': 42487, 'loss/train': 2.012070655822754} 02/26/2022 08:35:19 - INFO - codeparrot_training - Step 42488: {'lr': 2.9612310437897472e-05, 'samples': 21754368, 'steps': 42488, 'loss/train': 0.9607176780700684} 02/26/2022 08:35:25 - INFO - codeparrot_training - Step 42489: {'lr': 2.960458636763705e-05, 'samples': 21754880, 'steps': 42489, 'loss/train': 1.3465746641159058} 02/26/2022 08:35:28 - INFO - codeparrot_training - Step 42490: {'lr': 2.9596863241480697e-05, 'samples': 21755392, 'steps': 42490, 'loss/train': 0.32310932874679565} 02/26/2022 08:35:34 - INFO - codeparrot_training - Step 42491: {'lr': 2.958914105946131e-05, 'samples': 21755904, 'steps': 42491, 'loss/train': 1.3257124423980713} 02/26/2022 08:35:38 - INFO - codeparrot_training - Step 42492: {'lr': 2.958141982161211e-05, 'samples': 21756416, 'steps': 42492, 'loss/train': 0.9360162019729614} 02/26/2022 08:35:43 - INFO - codeparrot_training - Step 42493: {'lr': 2.957369952796607e-05, 'samples': 21756928, 'steps': 42493, 'loss/train': 2.098499298095703} 02/26/2022 08:35:47 - INFO - codeparrot_training - Step 42494: {'lr': 2.9565980178556385e-05, 'samples': 21757440, 'steps': 42494, 'loss/train': 2.0931572914123535} 02/26/2022 08:35:54 - INFO - codeparrot_training - Step 42495: {'lr': 2.9558261773415946e-05, 'samples': 21757952, 'steps': 42495, 'loss/train': 1.463809847831726} 02/26/2022 08:35:57 - INFO - codeparrot_training - Step 42496: {'lr': 2.9550544312577975e-05, 'samples': 21758464, 'steps': 42496, 'loss/train': 1.48870849609375} 02/26/2022 08:36:03 - INFO - codeparrot_training - Step 42497: {'lr': 2.954282779607545e-05, 'samples': 21758976, 'steps': 42497, 'loss/train': 1.8429895639419556} 02/26/2022 08:36:06 - INFO - codeparrot_training - Step 42498: {'lr': 2.953511222394151e-05, 'samples': 21759488, 'steps': 42498, 'loss/train': 1.6548742055892944} 02/26/2022 08:36:12 - INFO - codeparrot_training - Step 42499: {'lr': 2.9527397596209066e-05, 'samples': 21760000, 'steps': 42499, 'loss/train': 2.055388927459717} 02/26/2022 08:36:15 - INFO - codeparrot_training - Step 42500: {'lr': 2.9519683912911265e-05, 'samples': 21760512, 'steps': 42500, 'loss/train': 1.2720177173614502} 02/26/2022 08:36:21 - INFO - codeparrot_training - Step 42501: {'lr': 2.951197117408111e-05, 'samples': 21761024, 'steps': 42501, 'loss/train': 0.6935843229293823} 02/26/2022 08:36:24 - INFO - codeparrot_training - Step 42502: {'lr': 2.950425937975168e-05, 'samples': 21761536, 'steps': 42502, 'loss/train': 0.7947617173194885} 02/26/2022 08:36:31 - INFO - codeparrot_training - Step 42503: {'lr': 2.949654852995601e-05, 'samples': 21762048, 'steps': 42503, 'loss/train': 2.2537436485290527} 02/26/2022 08:36:34 - INFO - codeparrot_training - Step 42504: {'lr': 2.9488838624727092e-05, 'samples': 21762560, 'steps': 42504, 'loss/train': 0.44738835096359253} 02/26/2022 08:36:37 - INFO - codeparrot_training - Step 42505: {'lr': 2.948112966409791e-05, 'samples': 21763072, 'steps': 42505, 'loss/train': 2.3323655128479004} 02/26/2022 08:36:45 - INFO - codeparrot_training - Step 42506: {'lr': 2.9473421648101627e-05, 'samples': 21763584, 'steps': 42506, 'loss/train': 0.7735550999641418} 02/26/2022 08:36:48 - INFO - codeparrot_training - Step 42507: {'lr': 2.9465714576771164e-05, 'samples': 21764096, 'steps': 42507, 'loss/train': 1.2459367513656616} 02/26/2022 08:36:54 - INFO - codeparrot_training - Step 42508: {'lr': 2.9458008450139546e-05, 'samples': 21764608, 'steps': 42508, 'loss/train': 2.6039538383483887} 02/26/2022 08:36:57 - INFO - codeparrot_training - Step 42509: {'lr': 2.945030326823975e-05, 'samples': 21765120, 'steps': 42509, 'loss/train': 1.5605965852737427} 02/26/2022 08:37:03 - INFO - codeparrot_training - Step 42510: {'lr': 2.9442599031104855e-05, 'samples': 21765632, 'steps': 42510, 'loss/train': 1.2071714401245117} 02/26/2022 08:37:06 - INFO - codeparrot_training - Step 42511: {'lr': 2.9434895738767843e-05, 'samples': 21766144, 'steps': 42511, 'loss/train': 1.2468230724334717} 02/26/2022 08:37:12 - INFO - codeparrot_training - Step 42512: {'lr': 2.942719339126171e-05, 'samples': 21766656, 'steps': 42512, 'loss/train': 1.1582939624786377} 02/26/2022 08:37:15 - INFO - codeparrot_training - Step 42513: {'lr': 2.9419491988619374e-05, 'samples': 21767168, 'steps': 42513, 'loss/train': 2.4870450496673584} 02/26/2022 08:37:21 - INFO - codeparrot_training - Step 42514: {'lr': 2.941179153087395e-05, 'samples': 21767680, 'steps': 42514, 'loss/train': 1.1085726022720337} 02/26/2022 08:37:24 - INFO - codeparrot_training - Step 42515: {'lr': 2.940409201805838e-05, 'samples': 21768192, 'steps': 42515, 'loss/train': 2.046430826187134} 02/26/2022 08:37:31 - INFO - codeparrot_training - Step 42516: {'lr': 2.9396393450205617e-05, 'samples': 21768704, 'steps': 42516, 'loss/train': 2.6757395267486572} 02/26/2022 08:37:37 - INFO - codeparrot_training - Step 42517: {'lr': 2.9388695827348598e-05, 'samples': 21769216, 'steps': 42517, 'loss/train': 1.5505820512771606} 02/26/2022 08:37:40 - INFO - codeparrot_training - Step 42518: {'lr': 2.9380999149520417e-05, 'samples': 21769728, 'steps': 42518, 'loss/train': 1.9852464199066162} 02/26/2022 08:37:44 - INFO - codeparrot_training - Step 42519: {'lr': 2.9373303416753983e-05, 'samples': 21770240, 'steps': 42519, 'loss/train': 1.8253002166748047} 02/26/2022 08:37:49 - INFO - codeparrot_training - Step 42520: {'lr': 2.9365608629082246e-05, 'samples': 21770752, 'steps': 42520, 'loss/train': 1.0893845558166504} 02/26/2022 08:37:55 - INFO - codeparrot_training - Step 42521: {'lr': 2.9357914786538153e-05, 'samples': 21771264, 'steps': 42521, 'loss/train': 2.7885308265686035} 02/26/2022 08:37:58 - INFO - codeparrot_training - Step 42522: {'lr': 2.9350221889154733e-05, 'samples': 21771776, 'steps': 42522, 'loss/train': 2.164459228515625} 02/26/2022 08:38:04 - INFO - codeparrot_training - Step 42523: {'lr': 2.93425299369649e-05, 'samples': 21772288, 'steps': 42523, 'loss/train': 1.3384541273117065} 02/26/2022 08:38:07 - INFO - codeparrot_training - Step 42524: {'lr': 2.933483893000158e-05, 'samples': 21772800, 'steps': 42524, 'loss/train': 1.1142393350601196} 02/26/2022 08:38:14 - INFO - codeparrot_training - Step 42525: {'lr': 2.932714886829771e-05, 'samples': 21773312, 'steps': 42525, 'loss/train': 1.8720353841781616} 02/26/2022 08:38:18 - INFO - codeparrot_training - Step 42526: {'lr': 2.9319459751886323e-05, 'samples': 21773824, 'steps': 42526, 'loss/train': 0.49738532304763794} 02/26/2022 08:38:23 - INFO - codeparrot_training - Step 42527: {'lr': 2.9311771580800284e-05, 'samples': 21774336, 'steps': 42527, 'loss/train': 1.9495829343795776} 02/26/2022 08:38:27 - INFO - codeparrot_training - Step 42528: {'lr': 2.930408435507248e-05, 'samples': 21774848, 'steps': 42528, 'loss/train': 1.5275200605392456} 02/26/2022 08:38:32 - INFO - codeparrot_training - Step 42529: {'lr': 2.9296398074736025e-05, 'samples': 21775360, 'steps': 42529, 'loss/train': 2.031266689300537} 02/26/2022 08:38:36 - INFO - codeparrot_training - Step 42530: {'lr': 2.9288712739823586e-05, 'samples': 21775872, 'steps': 42530, 'loss/train': 2.146622896194458} 02/26/2022 08:38:41 - INFO - codeparrot_training - Step 42531: {'lr': 2.9281028350368304e-05, 'samples': 21776384, 'steps': 42531, 'loss/train': 1.7802083492279053} 02/26/2022 08:38:45 - INFO - codeparrot_training - Step 42532: {'lr': 2.9273344906402933e-05, 'samples': 21776896, 'steps': 42532, 'loss/train': 0.5005853772163391} 02/26/2022 08:38:50 - INFO - codeparrot_training - Step 42533: {'lr': 2.926566240796058e-05, 'samples': 21777408, 'steps': 42533, 'loss/train': 2.0109593868255615} 02/26/2022 08:38:54 - INFO - codeparrot_training - Step 42534: {'lr': 2.9257980855073918e-05, 'samples': 21777920, 'steps': 42534, 'loss/train': 1.67629873752594} 02/26/2022 08:38:59 - INFO - codeparrot_training - Step 42535: {'lr': 2.925030024777603e-05, 'samples': 21778432, 'steps': 42535, 'loss/train': 0.08264874666929245} 02/26/2022 08:39:03 - INFO - codeparrot_training - Step 42536: {'lr': 2.9242620586099723e-05, 'samples': 21778944, 'steps': 42536, 'loss/train': 2.4621009826660156} 02/26/2022 08:39:08 - INFO - codeparrot_training - Step 42537: {'lr': 2.923494187007797e-05, 'samples': 21779456, 'steps': 42537, 'loss/train': 1.8159173727035522} 02/26/2022 08:39:12 - INFO - codeparrot_training - Step 42538: {'lr': 2.9227264099743634e-05, 'samples': 21779968, 'steps': 42538, 'loss/train': 1.6453264951705933} 02/26/2022 08:39:17 - INFO - codeparrot_training - Step 42539: {'lr': 2.9219587275129578e-05, 'samples': 21780480, 'steps': 42539, 'loss/train': 1.9699792861938477} 02/26/2022 08:39:21 - INFO - codeparrot_training - Step 42540: {'lr': 2.9211911396268664e-05, 'samples': 21780992, 'steps': 42540, 'loss/train': 2.738023281097412} 02/26/2022 08:39:28 - INFO - codeparrot_training - Step 42541: {'lr': 2.9204236463193868e-05, 'samples': 21781504, 'steps': 42541, 'loss/train': 2.03415584564209} 02/26/2022 08:39:31 - INFO - codeparrot_training - Step 42542: {'lr': 2.9196562475938022e-05, 'samples': 21782016, 'steps': 42542, 'loss/train': 1.325456976890564} 02/26/2022 08:39:37 - INFO - codeparrot_training - Step 42543: {'lr': 2.9188889434533988e-05, 'samples': 21782528, 'steps': 42543, 'loss/train': 1.7245045900344849} 02/26/2022 08:39:40 - INFO - codeparrot_training - Step 42544: {'lr': 2.9181217339014577e-05, 'samples': 21783040, 'steps': 42544, 'loss/train': 1.2913734912872314} 02/26/2022 08:39:46 - INFO - codeparrot_training - Step 42545: {'lr': 2.917354618941276e-05, 'samples': 21783552, 'steps': 42545, 'loss/train': 1.4345474243164062} 02/26/2022 08:39:49 - INFO - codeparrot_training - Step 42546: {'lr': 2.916587598576134e-05, 'samples': 21784064, 'steps': 42546, 'loss/train': 2.3989639282226562} 02/26/2022 08:39:55 - INFO - codeparrot_training - Step 42547: {'lr': 2.915820672809319e-05, 'samples': 21784576, 'steps': 42547, 'loss/train': 1.8241488933563232} 02/26/2022 08:39:58 - INFO - codeparrot_training - Step 42548: {'lr': 2.9150538416441135e-05, 'samples': 21785088, 'steps': 42548, 'loss/train': 1.5071810483932495} 02/26/2022 08:40:04 - INFO - codeparrot_training - Step 42549: {'lr': 2.914287105083807e-05, 'samples': 21785600, 'steps': 42549, 'loss/train': 1.6656205654144287} 02/26/2022 08:40:07 - INFO - codeparrot_training - Step 42550: {'lr': 2.9135204631316776e-05, 'samples': 21786112, 'steps': 42550, 'loss/train': 2.246217727661133} 02/26/2022 08:40:15 - INFO - codeparrot_training - Step 42551: {'lr': 2.912753915791022e-05, 'samples': 21786624, 'steps': 42551, 'loss/train': 1.5617811679840088} 02/26/2022 08:40:18 - INFO - codeparrot_training - Step 42552: {'lr': 2.911987463065105e-05, 'samples': 21787136, 'steps': 42552, 'loss/train': 1.8169063329696655} 02/26/2022 08:40:24 - INFO - codeparrot_training - Step 42553: {'lr': 2.9112211049572263e-05, 'samples': 21787648, 'steps': 42553, 'loss/train': 0.3276555836200714} 02/26/2022 08:40:27 - INFO - codeparrot_training - Step 42554: {'lr': 2.9104548414706555e-05, 'samples': 21788160, 'steps': 42554, 'loss/train': 1.6070188283920288} 02/26/2022 08:40:33 - INFO - codeparrot_training - Step 42555: {'lr': 2.9096886726086957e-05, 'samples': 21788672, 'steps': 42555, 'loss/train': 0.46988704800605774} 02/26/2022 08:40:36 - INFO - codeparrot_training - Step 42556: {'lr': 2.9089225983746028e-05, 'samples': 21789184, 'steps': 42556, 'loss/train': 0.986107349395752} 02/26/2022 08:40:42 - INFO - codeparrot_training - Step 42557: {'lr': 2.908156618771676e-05, 'samples': 21789696, 'steps': 42557, 'loss/train': 1.9635928869247437} 02/26/2022 08:40:45 - INFO - codeparrot_training - Step 42558: {'lr': 2.9073907338031864e-05, 'samples': 21790208, 'steps': 42558, 'loss/train': 2.283731698989868} 02/26/2022 08:40:51 - INFO - codeparrot_training - Step 42559: {'lr': 2.9066249434724275e-05, 'samples': 21790720, 'steps': 42559, 'loss/train': 1.7113900184631348} 02/26/2022 08:40:54 - INFO - codeparrot_training - Step 42560: {'lr': 2.9058592477826635e-05, 'samples': 21791232, 'steps': 42560, 'loss/train': 0.26361116766929626} 02/26/2022 08:41:02 - INFO - codeparrot_training - Step 42561: {'lr': 2.905093646737189e-05, 'samples': 21791744, 'steps': 42561, 'loss/train': 1.330338716506958} 02/26/2022 08:41:05 - INFO - codeparrot_training - Step 42562: {'lr': 2.9043281403392768e-05, 'samples': 21792256, 'steps': 42562, 'loss/train': 1.6504161357879639} 02/26/2022 08:41:11 - INFO - codeparrot_training - Step 42563: {'lr': 2.9035627285922017e-05, 'samples': 21792768, 'steps': 42563, 'loss/train': 1.2182122468948364} 02/26/2022 08:41:14 - INFO - codeparrot_training - Step 42564: {'lr': 2.9027974114992528e-05, 'samples': 21793280, 'steps': 42564, 'loss/train': 2.004009246826172} 02/26/2022 08:41:19 - INFO - codeparrot_training - Step 42565: {'lr': 2.9020321890637026e-05, 'samples': 21793792, 'steps': 42565, 'loss/train': 2.345698356628418} 02/26/2022 08:41:23 - INFO - codeparrot_training - Step 42566: {'lr': 2.9012670612888286e-05, 'samples': 21794304, 'steps': 42566, 'loss/train': 1.1215437650680542} 02/26/2022 08:41:29 - INFO - codeparrot_training - Step 42567: {'lr': 2.9005020281779065e-05, 'samples': 21794816, 'steps': 42567, 'loss/train': 0.8166053295135498} 02/26/2022 08:41:32 - INFO - codeparrot_training - Step 42568: {'lr': 2.8997370897342197e-05, 'samples': 21795328, 'steps': 42568, 'loss/train': 1.7318686246871948} 02/26/2022 08:41:37 - INFO - codeparrot_training - Step 42569: {'lr': 2.8989722459610402e-05, 'samples': 21795840, 'steps': 42569, 'loss/train': 1.941198468208313} 02/26/2022 08:41:41 - INFO - codeparrot_training - Step 42570: {'lr': 2.898207496861649e-05, 'samples': 21796352, 'steps': 42570, 'loss/train': 1.4996109008789062} 02/26/2022 08:41:46 - INFO - codeparrot_training - Step 42571: {'lr': 2.8974428424393127e-05, 'samples': 21796864, 'steps': 42571, 'loss/train': 2.7091126441955566} 02/26/2022 08:41:50 - INFO - codeparrot_training - Step 42572: {'lr': 2.896678282697318e-05, 'samples': 21797376, 'steps': 42572, 'loss/train': 2.3050708770751953} 02/26/2022 08:41:55 - INFO - codeparrot_training - Step 42573: {'lr': 2.8959138176389342e-05, 'samples': 21797888, 'steps': 42573, 'loss/train': 2.0634372234344482} 02/26/2022 08:41:59 - INFO - codeparrot_training - Step 42574: {'lr': 2.8951494472674365e-05, 'samples': 21798400, 'steps': 42574, 'loss/train': 1.5202707052230835} 02/26/2022 08:42:05 - INFO - codeparrot_training - Step 42575: {'lr': 2.8943851715860946e-05, 'samples': 21798912, 'steps': 42575, 'loss/train': 0.830349862575531} 02/26/2022 08:42:08 - INFO - codeparrot_training - Step 42576: {'lr': 2.893620990598192e-05, 'samples': 21799424, 'steps': 42576, 'loss/train': 1.5002254247665405} 02/26/2022 08:42:12 - INFO - codeparrot_training - Step 42577: {'lr': 2.892856904306998e-05, 'samples': 21799936, 'steps': 42577, 'loss/train': 0.6080021262168884} 02/26/2022 08:42:19 - INFO - codeparrot_training - Step 42578: {'lr': 2.8920929127157852e-05, 'samples': 21800448, 'steps': 42578, 'loss/train': 0.7348437309265137} 02/26/2022 08:42:23 - INFO - codeparrot_training - Step 42579: {'lr': 2.8913290158278232e-05, 'samples': 21800960, 'steps': 42579, 'loss/train': 1.9319316148757935} 02/26/2022 08:42:28 - INFO - codeparrot_training - Step 42580: {'lr': 2.89056521364639e-05, 'samples': 21801472, 'steps': 42580, 'loss/train': 2.409445285797119} 02/26/2022 08:42:32 - INFO - codeparrot_training - Step 42581: {'lr': 2.889801506174755e-05, 'samples': 21801984, 'steps': 42581, 'loss/train': 0.8004657030105591} 02/26/2022 08:42:37 - INFO - codeparrot_training - Step 42582: {'lr': 2.889037893416191e-05, 'samples': 21802496, 'steps': 42582, 'loss/train': 1.27082359790802} 02/26/2022 08:42:41 - INFO - codeparrot_training - Step 42583: {'lr': 2.8882743753739615e-05, 'samples': 21803008, 'steps': 42583, 'loss/train': 0.9892042279243469} 02/26/2022 08:42:46 - INFO - codeparrot_training - Step 42584: {'lr': 2.8875109520513505e-05, 'samples': 21803520, 'steps': 42584, 'loss/train': 0.99040687084198} 02/26/2022 08:42:50 - INFO - codeparrot_training - Step 42585: {'lr': 2.8867476234516134e-05, 'samples': 21804032, 'steps': 42585, 'loss/train': 2.163007974624634} 02/26/2022 08:42:55 - INFO - codeparrot_training - Step 42586: {'lr': 2.885984389578039e-05, 'samples': 21804544, 'steps': 42586, 'loss/train': 0.9899985194206238} 02/26/2022 08:42:59 - INFO - codeparrot_training - Step 42587: {'lr': 2.8852212504338752e-05, 'samples': 21805056, 'steps': 42587, 'loss/train': 1.6692215204238892} 02/26/2022 08:43:06 - INFO - codeparrot_training - Step 42588: {'lr': 2.8844582060224055e-05, 'samples': 21805568, 'steps': 42588, 'loss/train': 0.6595707535743713} 02/26/2022 08:43:09 - INFO - codeparrot_training - Step 42589: {'lr': 2.883695256346891e-05, 'samples': 21806080, 'steps': 42589, 'loss/train': 1.915860652923584} 02/26/2022 08:43:15 - INFO - codeparrot_training - Step 42590: {'lr': 2.8829324014106122e-05, 'samples': 21806592, 'steps': 42590, 'loss/train': 0.5317558646202087} 02/26/2022 08:43:19 - INFO - codeparrot_training - Step 42591: {'lr': 2.8821696412168168e-05, 'samples': 21807104, 'steps': 42591, 'loss/train': 1.5545148849487305} 02/26/2022 08:43:24 - INFO - codeparrot_training - Step 42592: {'lr': 2.881406975768791e-05, 'samples': 21807616, 'steps': 42592, 'loss/train': 2.173269033432007} 02/26/2022 08:43:28 - INFO - codeparrot_training - Step 42593: {'lr': 2.8806444050697876e-05, 'samples': 21808128, 'steps': 42593, 'loss/train': 2.067986488342285} 02/26/2022 08:43:34 - INFO - codeparrot_training - Step 42594: {'lr': 2.8798819291230905e-05, 'samples': 21808640, 'steps': 42594, 'loss/train': 0.9304419755935669} 02/26/2022 08:43:37 - INFO - codeparrot_training - Step 42595: {'lr': 2.8791195479319437e-05, 'samples': 21809152, 'steps': 42595, 'loss/train': 2.0435805320739746} 02/26/2022 08:43:43 - INFO - codeparrot_training - Step 42596: {'lr': 2.878357261499631e-05, 'samples': 21809664, 'steps': 42596, 'loss/train': 1.7889502048492432} 02/26/2022 08:43:46 - INFO - codeparrot_training - Step 42597: {'lr': 2.8775950698294084e-05, 'samples': 21810176, 'steps': 42597, 'loss/train': 1.5202713012695312} 02/26/2022 08:43:52 - INFO - codeparrot_training - Step 42598: {'lr': 2.876832972924545e-05, 'samples': 21810688, 'steps': 42598, 'loss/train': 0.14154919981956482} 02/26/2022 08:43:55 - INFO - codeparrot_training - Step 42599: {'lr': 2.876070970788308e-05, 'samples': 21811200, 'steps': 42599, 'loss/train': 1.1133460998535156} 02/26/2022 08:44:02 - INFO - codeparrot_training - Step 42600: {'lr': 2.875309063423956e-05, 'samples': 21811712, 'steps': 42600, 'loss/train': 1.5177854299545288} 02/26/2022 08:44:06 - INFO - codeparrot_training - Step 42601: {'lr': 2.87454725083475e-05, 'samples': 21812224, 'steps': 42601, 'loss/train': 1.5799976587295532} 02/26/2022 08:44:11 - INFO - codeparrot_training - Step 42602: {'lr': 2.873785533023962e-05, 'samples': 21812736, 'steps': 42602, 'loss/train': 8.020909309387207} 02/26/2022 08:44:15 - INFO - codeparrot_training - Step 42603: {'lr': 2.8730239099948514e-05, 'samples': 21813248, 'steps': 42603, 'loss/train': 1.5111581087112427} 02/26/2022 08:44:20 - INFO - codeparrot_training - Step 42604: {'lr': 2.8722623817506786e-05, 'samples': 21813760, 'steps': 42604, 'loss/train': 2.1292598247528076} 02/26/2022 08:44:24 - INFO - codeparrot_training - Step 42605: {'lr': 2.8715009482947056e-05, 'samples': 21814272, 'steps': 42605, 'loss/train': 1.3609613180160522} 02/26/2022 08:44:29 - INFO - codeparrot_training - Step 42606: {'lr': 2.870739609630199e-05, 'samples': 21814784, 'steps': 42606, 'loss/train': 1.2966855764389038} 02/26/2022 08:44:33 - INFO - codeparrot_training - Step 42607: {'lr': 2.8699783657604172e-05, 'samples': 21815296, 'steps': 42607, 'loss/train': 1.4713834524154663} 02/26/2022 08:44:38 - INFO - codeparrot_training - Step 42608: {'lr': 2.8692172166886215e-05, 'samples': 21815808, 'steps': 42608, 'loss/train': 0.7541775703430176} 02/26/2022 08:44:42 - INFO - codeparrot_training - Step 42609: {'lr': 2.8684561624180733e-05, 'samples': 21816320, 'steps': 42609, 'loss/train': 1.8889870643615723} 02/26/2022 08:44:49 - INFO - codeparrot_training - Step 42610: {'lr': 2.8676952029520225e-05, 'samples': 21816832, 'steps': 42610, 'loss/train': 1.6704007387161255} 02/26/2022 08:44:54 - INFO - codeparrot_training - Step 42611: {'lr': 2.8669343382937474e-05, 'samples': 21817344, 'steps': 42611, 'loss/train': 1.8519840240478516} 02/26/2022 08:44:58 - INFO - codeparrot_training - Step 42612: {'lr': 2.8661735684464952e-05, 'samples': 21817856, 'steps': 42612, 'loss/train': 2.4465653896331787} 02/26/2022 08:45:04 - INFO - codeparrot_training - Step 42613: {'lr': 2.865412893413527e-05, 'samples': 21818368, 'steps': 42613, 'loss/train': 1.7758008241653442} 02/26/2022 08:45:07 - INFO - codeparrot_training - Step 42614: {'lr': 2.864652313198096e-05, 'samples': 21818880, 'steps': 42614, 'loss/train': 2.6221158504486084} 02/26/2022 08:45:13 - INFO - codeparrot_training - Step 42615: {'lr': 2.863891827803472e-05, 'samples': 21819392, 'steps': 42615, 'loss/train': 1.1948521137237549} 02/26/2022 08:45:16 - INFO - codeparrot_training - Step 42616: {'lr': 2.863131437232905e-05, 'samples': 21819904, 'steps': 42616, 'loss/train': 1.8522579669952393} 02/26/2022 08:45:22 - INFO - codeparrot_training - Step 42617: {'lr': 2.862371141489653e-05, 'samples': 21820416, 'steps': 42617, 'loss/train': 1.9316136837005615} 02/26/2022 08:45:25 - INFO - codeparrot_training - Step 42618: {'lr': 2.8616109405769697e-05, 'samples': 21820928, 'steps': 42618, 'loss/train': 2.8018152713775635} 02/26/2022 08:45:29 - INFO - codeparrot_training - Step 42619: {'lr': 2.8608508344981216e-05, 'samples': 21821440, 'steps': 42619, 'loss/train': 0.1489679366350174} 02/26/2022 08:45:34 - INFO - codeparrot_training - Step 42620: {'lr': 2.860090823256359e-05, 'samples': 21821952, 'steps': 42620, 'loss/train': 1.258629322052002} 02/26/2022 08:45:38 - INFO - codeparrot_training - Step 42621: {'lr': 2.8593309068549344e-05, 'samples': 21822464, 'steps': 42621, 'loss/train': 1.5470376014709473} 02/26/2022 08:45:43 - INFO - codeparrot_training - Step 42622: {'lr': 2.8585710852971015e-05, 'samples': 21822976, 'steps': 42622, 'loss/train': 2.7234253883361816} 02/26/2022 08:45:47 - INFO - codeparrot_training - Step 42623: {'lr': 2.8578113585861264e-05, 'samples': 21823488, 'steps': 42623, 'loss/train': 1.8056012392044067} 02/26/2022 08:45:54 - INFO - codeparrot_training - Step 42624: {'lr': 2.8570517267252488e-05, 'samples': 21824000, 'steps': 42624, 'loss/train': 2.4880359172821045} 02/26/2022 08:45:58 - INFO - codeparrot_training - Step 42625: {'lr': 2.8562921897177408e-05, 'samples': 21824512, 'steps': 42625, 'loss/train': 1.9545180797576904} 02/26/2022 08:46:03 - INFO - codeparrot_training - Step 42626: {'lr': 2.8555327475668358e-05, 'samples': 21825024, 'steps': 42626, 'loss/train': 1.5446813106536865} 02/26/2022 08:46:06 - INFO - codeparrot_training - Step 42627: {'lr': 2.8547734002758035e-05, 'samples': 21825536, 'steps': 42627, 'loss/train': 1.217718482017517} 02/26/2022 08:46:12 - INFO - codeparrot_training - Step 42628: {'lr': 2.8540141478478832e-05, 'samples': 21826048, 'steps': 42628, 'loss/train': 1.8193233013153076} 02/26/2022 08:46:15 - INFO - codeparrot_training - Step 42629: {'lr': 2.853254990286347e-05, 'samples': 21826560, 'steps': 42629, 'loss/train': 1.8696283102035522} 02/26/2022 08:46:21 - INFO - codeparrot_training - Step 42630: {'lr': 2.85249592759442e-05, 'samples': 21827072, 'steps': 42630, 'loss/train': 2.6454615592956543} 02/26/2022 08:46:26 - INFO - codeparrot_training - Step 42631: {'lr': 2.851736959775375e-05, 'samples': 21827584, 'steps': 42631, 'loss/train': 1.6382313966751099} 02/26/2022 08:46:30 - INFO - codeparrot_training - Step 42632: {'lr': 2.8509780868324507e-05, 'samples': 21828096, 'steps': 42632, 'loss/train': 0.9552024006843567} 02/26/2022 08:46:37 - INFO - codeparrot_training - Step 42633: {'lr': 2.8502193087689144e-05, 'samples': 21828608, 'steps': 42633, 'loss/train': 2.2013673782348633} 02/26/2022 08:46:41 - INFO - codeparrot_training - Step 42634: {'lr': 2.8494606255879935e-05, 'samples': 21829120, 'steps': 42634, 'loss/train': 1.4669660329818726} 02/26/2022 08:46:46 - INFO - codeparrot_training - Step 42635: {'lr': 2.848702037292955e-05, 'samples': 21829632, 'steps': 42635, 'loss/train': 1.0452195405960083} 02/26/2022 08:46:49 - INFO - codeparrot_training - Step 42636: {'lr': 2.8479435438870382e-05, 'samples': 21830144, 'steps': 42636, 'loss/train': 1.7292696237564087} 02/26/2022 08:46:55 - INFO - codeparrot_training - Step 42637: {'lr': 2.8471851453735042e-05, 'samples': 21830656, 'steps': 42637, 'loss/train': 1.498539924621582} 02/26/2022 08:46:58 - INFO - codeparrot_training - Step 42638: {'lr': 2.8464268417555923e-05, 'samples': 21831168, 'steps': 42638, 'loss/train': 2.4578187465667725} 02/26/2022 08:47:04 - INFO - codeparrot_training - Step 42639: {'lr': 2.845668633036555e-05, 'samples': 21831680, 'steps': 42639, 'loss/train': 1.7947949171066284} 02/26/2022 08:47:08 - INFO - codeparrot_training - Step 42640: {'lr': 2.8449105192196318e-05, 'samples': 21832192, 'steps': 42640, 'loss/train': 2.132197380065918} 02/26/2022 08:47:13 - INFO - codeparrot_training - Step 42641: {'lr': 2.844152500308081e-05, 'samples': 21832704, 'steps': 42641, 'loss/train': 2.067821502685547} 02/26/2022 08:47:16 - INFO - codeparrot_training - Step 42642: {'lr': 2.8433945763051472e-05, 'samples': 21833216, 'steps': 42642, 'loss/train': 2.615954637527466} 02/26/2022 08:47:24 - INFO - codeparrot_training - Step 42643: {'lr': 2.842636747214075e-05, 'samples': 21833728, 'steps': 42643, 'loss/train': 0.8795211911201477} 02/26/2022 08:47:27 - INFO - codeparrot_training - Step 42644: {'lr': 2.8418790130381067e-05, 'samples': 21834240, 'steps': 42644, 'loss/train': 2.1950387954711914} 02/26/2022 08:47:33 - INFO - codeparrot_training - Step 42645: {'lr': 2.8411213737805e-05, 'samples': 21834752, 'steps': 42645, 'loss/train': 1.9486632347106934} 02/26/2022 08:47:36 - INFO - codeparrot_training - Step 42646: {'lr': 2.8403638294444896e-05, 'samples': 21835264, 'steps': 42646, 'loss/train': 2.1200315952301025} 02/26/2022 08:47:42 - INFO - codeparrot_training - Step 42647: {'lr': 2.8396063800333246e-05, 'samples': 21835776, 'steps': 42647, 'loss/train': 2.3895275592803955} 02/26/2022 08:47:45 - INFO - codeparrot_training - Step 42648: {'lr': 2.838849025550244e-05, 'samples': 21836288, 'steps': 42648, 'loss/train': 0.36600908637046814} 02/26/2022 08:47:51 - INFO - codeparrot_training - Step 42649: {'lr': 2.8380917659985045e-05, 'samples': 21836800, 'steps': 42649, 'loss/train': 1.7769306898117065} 02/26/2022 08:47:54 - INFO - codeparrot_training - Step 42650: {'lr': 2.8373346013813417e-05, 'samples': 21837312, 'steps': 42650, 'loss/train': 1.5565471649169922} 02/26/2022 08:48:00 - INFO - codeparrot_training - Step 42651: {'lr': 2.8365775317020004e-05, 'samples': 21837824, 'steps': 42651, 'loss/train': 1.6738334894180298} 02/26/2022 08:48:03 - INFO - codeparrot_training - Step 42652: {'lr': 2.8358205569637168e-05, 'samples': 21838336, 'steps': 42652, 'loss/train': 0.7688226699829102} 02/26/2022 08:48:11 - INFO - codeparrot_training - Step 42653: {'lr': 2.835063677169744e-05, 'samples': 21838848, 'steps': 42653, 'loss/train': 0.896828293800354} 02/26/2022 08:48:14 - INFO - codeparrot_training - Step 42654: {'lr': 2.834306892323324e-05, 'samples': 21839360, 'steps': 42654, 'loss/train': 1.6306073665618896} 02/26/2022 08:48:20 - INFO - codeparrot_training - Step 42655: {'lr': 2.8335502024276925e-05, 'samples': 21839872, 'steps': 42655, 'loss/train': 1.7563389539718628} 02/26/2022 08:48:23 - INFO - codeparrot_training - Step 42656: {'lr': 2.8327936074860865e-05, 'samples': 21840384, 'steps': 42656, 'loss/train': 1.9090511798858643} 02/26/2022 08:48:29 - INFO - codeparrot_training - Step 42657: {'lr': 2.8320371075017613e-05, 'samples': 21840896, 'steps': 42657, 'loss/train': 1.8975558280944824} 02/26/2022 08:48:32 - INFO - codeparrot_training - Step 42658: {'lr': 2.831280702477951e-05, 'samples': 21841408, 'steps': 42658, 'loss/train': 2.1529974937438965} 02/26/2022 08:48:38 - INFO - codeparrot_training - Step 42659: {'lr': 2.830524392417888e-05, 'samples': 21841920, 'steps': 42659, 'loss/train': 1.7760286331176758} 02/26/2022 08:48:41 - INFO - codeparrot_training - Step 42660: {'lr': 2.829768177324829e-05, 'samples': 21842432, 'steps': 42660, 'loss/train': 1.5817909240722656} 02/26/2022 08:48:46 - INFO - codeparrot_training - Step 42661: {'lr': 2.8290120572019933e-05, 'samples': 21842944, 'steps': 42661, 'loss/train': 1.4045246839523315} 02/26/2022 08:48:50 - INFO - codeparrot_training - Step 42662: {'lr': 2.828256032052634e-05, 'samples': 21843456, 'steps': 42662, 'loss/train': 1.3400585651397705} 02/26/2022 08:48:55 - INFO - codeparrot_training - Step 42663: {'lr': 2.8275001018799817e-05, 'samples': 21843968, 'steps': 42663, 'loss/train': 2.821943521499634} 02/26/2022 08:48:59 - INFO - codeparrot_training - Step 42664: {'lr': 2.8267442666872893e-05, 'samples': 21844480, 'steps': 42664, 'loss/train': 2.115737199783325} 02/26/2022 08:49:04 - INFO - codeparrot_training - Step 42665: {'lr': 2.825988526477771e-05, 'samples': 21844992, 'steps': 42665, 'loss/train': 0.9830888509750366} 02/26/2022 08:49:08 - INFO - codeparrot_training - Step 42666: {'lr': 2.825232881254686e-05, 'samples': 21845504, 'steps': 42666, 'loss/train': 0.5918101668357849} 02/26/2022 08:49:13 - INFO - codeparrot_training - Step 42667: {'lr': 2.8244773310212522e-05, 'samples': 21846016, 'steps': 42667, 'loss/train': 0.917984127998352} 02/26/2022 08:49:17 - INFO - codeparrot_training - Step 42668: {'lr': 2.8237218757807297e-05, 'samples': 21846528, 'steps': 42668, 'loss/train': 1.7263184785842896} 02/26/2022 08:49:22 - INFO - codeparrot_training - Step 42669: {'lr': 2.8229665155363294e-05, 'samples': 21847040, 'steps': 42669, 'loss/train': 1.574794888496399} 02/26/2022 08:49:26 - INFO - codeparrot_training - Step 42670: {'lr': 2.8222112502913037e-05, 'samples': 21847552, 'steps': 42670, 'loss/train': 2.032003879547119} 02/26/2022 08:49:33 - INFO - codeparrot_training - Step 42671: {'lr': 2.8214560800488788e-05, 'samples': 21848064, 'steps': 42671, 'loss/train': 2.283432960510254} 02/26/2022 08:49:36 - INFO - codeparrot_training - Step 42672: {'lr': 2.8207010048122954e-05, 'samples': 21848576, 'steps': 42672, 'loss/train': 1.1152524948120117} 02/26/2022 08:49:42 - INFO - codeparrot_training - Step 42673: {'lr': 2.819946024584791e-05, 'samples': 21849088, 'steps': 42673, 'loss/train': 1.4223766326904297} 02/26/2022 08:49:46 - INFO - codeparrot_training - Step 42674: {'lr': 2.8191911393695923e-05, 'samples': 21849600, 'steps': 42674, 'loss/train': 2.281524896621704} 02/26/2022 08:49:51 - INFO - codeparrot_training - Step 42675: {'lr': 2.8184363491699285e-05, 'samples': 21850112, 'steps': 42675, 'loss/train': 1.5728832483291626} 02/26/2022 08:49:55 - INFO - codeparrot_training - Step 42676: {'lr': 2.8176816539890488e-05, 'samples': 21850624, 'steps': 42676, 'loss/train': 0.5886314511299133} 02/26/2022 08:50:00 - INFO - codeparrot_training - Step 42677: {'lr': 2.8169270538301733e-05, 'samples': 21851136, 'steps': 42677, 'loss/train': 1.7444263696670532} 02/26/2022 08:50:04 - INFO - codeparrot_training - Step 42678: {'lr': 2.816172548696541e-05, 'samples': 21851648, 'steps': 42678, 'loss/train': 2.145519495010376} 02/26/2022 08:50:09 - INFO - codeparrot_training - Step 42679: {'lr': 2.8154181385913747e-05, 'samples': 21852160, 'steps': 42679, 'loss/train': 2.729649782180786} 02/26/2022 08:50:13 - INFO - codeparrot_training - Step 42680: {'lr': 2.8146638235179213e-05, 'samples': 21852672, 'steps': 42680, 'loss/train': 2.2197585105895996} 02/26/2022 08:50:20 - INFO - codeparrot_training - Step 42681: {'lr': 2.8139096034794005e-05, 'samples': 21853184, 'steps': 42681, 'loss/train': 2.3298866748809814} 02/26/2022 08:50:23 - INFO - codeparrot_training - Step 42682: {'lr': 2.813155478479046e-05, 'samples': 21853696, 'steps': 42682, 'loss/train': 0.1754862368106842} 02/26/2022 08:50:29 - INFO - codeparrot_training - Step 42683: {'lr': 2.8124014485200827e-05, 'samples': 21854208, 'steps': 42683, 'loss/train': 1.8918489217758179} 02/26/2022 08:50:32 - INFO - codeparrot_training - Step 42684: {'lr': 2.811647513605753e-05, 'samples': 21854720, 'steps': 42684, 'loss/train': 1.54283607006073} 02/26/2022 08:50:38 - INFO - codeparrot_training - Step 42685: {'lr': 2.810893673739273e-05, 'samples': 21855232, 'steps': 42685, 'loss/train': 1.3746970891952515} 02/26/2022 08:50:41 - INFO - codeparrot_training - Step 42686: {'lr': 2.810139928923891e-05, 'samples': 21855744, 'steps': 42686, 'loss/train': 1.8520458936691284} 02/26/2022 08:50:47 - INFO - codeparrot_training - Step 42687: {'lr': 2.809386279162812e-05, 'samples': 21856256, 'steps': 42687, 'loss/train': 1.4545599222183228} 02/26/2022 08:50:50 - INFO - codeparrot_training - Step 42688: {'lr': 2.8086327244592815e-05, 'samples': 21856768, 'steps': 42688, 'loss/train': 1.3848674297332764} 02/26/2022 08:50:56 - INFO - codeparrot_training - Step 42689: {'lr': 2.8078792648165154e-05, 'samples': 21857280, 'steps': 42689, 'loss/train': 1.182730793952942} 02/26/2022 08:51:00 - INFO - codeparrot_training - Step 42690: {'lr': 2.8071259002377585e-05, 'samples': 21857792, 'steps': 42690, 'loss/train': 1.8035038709640503} 02/26/2022 08:51:07 - INFO - codeparrot_training - Step 42691: {'lr': 2.8063726307262172e-05, 'samples': 21858304, 'steps': 42691, 'loss/train': 7.538046836853027} 02/26/2022 08:51:11 - INFO - codeparrot_training - Step 42692: {'lr': 2.8056194562851355e-05, 'samples': 21858816, 'steps': 42692, 'loss/train': 0.4667845368385315} 02/26/2022 08:51:14 - INFO - codeparrot_training - Step 42693: {'lr': 2.8048663769177308e-05, 'samples': 21859328, 'steps': 42693, 'loss/train': 1.344470739364624} 02/26/2022 08:51:19 - INFO - codeparrot_training - Step 42694: {'lr': 2.804113392627225e-05, 'samples': 21859840, 'steps': 42694, 'loss/train': 2.0444798469543457} 02/26/2022 08:51:23 - INFO - codeparrot_training - Step 42695: {'lr': 2.8033605034168546e-05, 'samples': 21860352, 'steps': 42695, 'loss/train': 2.03246808052063} 02/26/2022 08:51:29 - INFO - codeparrot_training - Step 42696: {'lr': 2.8026077092898396e-05, 'samples': 21860864, 'steps': 42696, 'loss/train': 1.15717613697052} 02/26/2022 08:51:32 - INFO - codeparrot_training - Step 42697: {'lr': 2.8018550102494046e-05, 'samples': 21861376, 'steps': 42697, 'loss/train': 5.0372796058654785} 02/26/2022 08:51:38 - INFO - codeparrot_training - Step 42698: {'lr': 2.8011024062987693e-05, 'samples': 21861888, 'steps': 42698, 'loss/train': 1.5328001976013184} 02/26/2022 08:51:41 - INFO - codeparrot_training - Step 42699: {'lr': 2.8003498974411678e-05, 'samples': 21862400, 'steps': 42699, 'loss/train': 1.55319082736969} 02/26/2022 08:51:46 - INFO - codeparrot_training - Step 42700: {'lr': 2.7995974836798194e-05, 'samples': 21862912, 'steps': 42700, 'loss/train': 1.9834460020065308} 02/26/2022 08:51:50 - INFO - codeparrot_training - Step 42701: {'lr': 2.7988451650179435e-05, 'samples': 21863424, 'steps': 42701, 'loss/train': 1.7467377185821533} 02/26/2022 08:51:57 - INFO - codeparrot_training - Step 42702: {'lr': 2.7980929414587602e-05, 'samples': 21863936, 'steps': 42702, 'loss/train': 1.1268304586410522} 02/26/2022 08:52:01 - INFO - codeparrot_training - Step 42703: {'lr': 2.797340813005503e-05, 'samples': 21864448, 'steps': 42703, 'loss/train': 1.7960563898086548} 02/26/2022 08:52:06 - INFO - codeparrot_training - Step 42704: {'lr': 2.796588779661388e-05, 'samples': 21864960, 'steps': 42704, 'loss/train': 2.186213970184326} 02/26/2022 08:52:10 - INFO - codeparrot_training - Step 42705: {'lr': 2.795836841429636e-05, 'samples': 21865472, 'steps': 42705, 'loss/train': 1.7804704904556274} 02/26/2022 08:52:15 - INFO - codeparrot_training - Step 42706: {'lr': 2.795084998313463e-05, 'samples': 21865984, 'steps': 42706, 'loss/train': 1.4118626117706299} 02/26/2022 08:52:19 - INFO - codeparrot_training - Step 42707: {'lr': 2.7943332503161e-05, 'samples': 21866496, 'steps': 42707, 'loss/train': 1.3047691583633423} 02/26/2022 08:52:24 - INFO - codeparrot_training - Step 42708: {'lr': 2.793581597440764e-05, 'samples': 21867008, 'steps': 42708, 'loss/train': 1.8832931518554688} 02/26/2022 08:52:28 - INFO - codeparrot_training - Step 42709: {'lr': 2.792830039690672e-05, 'samples': 21867520, 'steps': 42709, 'loss/train': 2.0773541927337646} 02/26/2022 08:52:33 - INFO - codeparrot_training - Step 42710: {'lr': 2.7920785770690377e-05, 'samples': 21868032, 'steps': 42710, 'loss/train': 2.4615700244903564} 02/26/2022 08:52:36 - INFO - codeparrot_training - Step 42711: {'lr': 2.791327209579095e-05, 'samples': 21868544, 'steps': 42711, 'loss/train': 1.280531406402588} 02/26/2022 08:52:42 - INFO - codeparrot_training - Step 42712: {'lr': 2.790575937224052e-05, 'samples': 21869056, 'steps': 42712, 'loss/train': 1.0784783363342285} 02/26/2022 08:52:45 - INFO - codeparrot_training - Step 42713: {'lr': 2.7898247600071284e-05, 'samples': 21869568, 'steps': 42713, 'loss/train': 2.4117431640625} 02/26/2022 08:52:51 - INFO - codeparrot_training - Step 42714: {'lr': 2.7890736779315417e-05, 'samples': 21870080, 'steps': 42714, 'loss/train': 2.470004081726074} 02/26/2022 08:52:54 - INFO - codeparrot_training - Step 42715: {'lr': 2.788322691000514e-05, 'samples': 21870592, 'steps': 42715, 'loss/train': 1.7888002395629883} 02/26/2022 08:53:02 - INFO - codeparrot_training - Step 42716: {'lr': 2.787571799217259e-05, 'samples': 21871104, 'steps': 42716, 'loss/train': 1.6819690465927124} 02/26/2022 08:53:05 - INFO - codeparrot_training - Step 42717: {'lr': 2.786821002584991e-05, 'samples': 21871616, 'steps': 42717, 'loss/train': 2.0524652004241943} 02/26/2022 08:53:11 - INFO - codeparrot_training - Step 42718: {'lr': 2.7860703011069244e-05, 'samples': 21872128, 'steps': 42718, 'loss/train': 1.5147196054458618} 02/26/2022 08:53:14 - INFO - codeparrot_training - Step 42719: {'lr': 2.785319694786287e-05, 'samples': 21872640, 'steps': 42719, 'loss/train': 2.29386568069458} 02/26/2022 08:53:20 - INFO - codeparrot_training - Step 42720: {'lr': 2.784569183626276e-05, 'samples': 21873152, 'steps': 42720, 'loss/train': 0.7137149572372437} 02/26/2022 08:53:23 - INFO - codeparrot_training - Step 42721: {'lr': 2.783818767630131e-05, 'samples': 21873664, 'steps': 42721, 'loss/train': 2.816556692123413} 02/26/2022 08:53:29 - INFO - codeparrot_training - Step 42722: {'lr': 2.7830684468010403e-05, 'samples': 21874176, 'steps': 42722, 'loss/train': 0.13773326575756073} 02/26/2022 08:53:33 - INFO - codeparrot_training - Step 42723: {'lr': 2.7823182211422326e-05, 'samples': 21874688, 'steps': 42723, 'loss/train': 1.8043315410614014} 02/26/2022 08:53:38 - INFO - codeparrot_training - Step 42724: {'lr': 2.7815680906569162e-05, 'samples': 21875200, 'steps': 42724, 'loss/train': 2.102477550506592} 02/26/2022 08:53:42 - INFO - codeparrot_training - Step 42725: {'lr': 2.7808180553483164e-05, 'samples': 21875712, 'steps': 42725, 'loss/train': 1.5142573118209839} 02/26/2022 08:53:47 - INFO - codeparrot_training - Step 42726: {'lr': 2.7800681152196277e-05, 'samples': 21876224, 'steps': 42726, 'loss/train': 1.1307779550552368} 02/26/2022 08:53:50 - INFO - codeparrot_training - Step 42727: {'lr': 2.7793182702740753e-05, 'samples': 21876736, 'steps': 42727, 'loss/train': 0.04525814577937126} 02/26/2022 08:53:58 - INFO - codeparrot_training - Step 42728: {'lr': 2.7785685205148625e-05, 'samples': 21877248, 'steps': 42728, 'loss/train': 1.863305926322937} 02/26/2022 08:54:01 - INFO - codeparrot_training - Step 42729: {'lr': 2.777818865945217e-05, 'samples': 21877760, 'steps': 42729, 'loss/train': 1.04934823513031} 02/26/2022 08:54:07 - INFO - codeparrot_training - Step 42730: {'lr': 2.7770693065683277e-05, 'samples': 21878272, 'steps': 42730, 'loss/train': 1.5525736808776855} 02/26/2022 08:54:10 - INFO - codeparrot_training - Step 42731: {'lr': 2.776319842387423e-05, 'samples': 21878784, 'steps': 42731, 'loss/train': 2.3730316162109375} 02/26/2022 08:54:16 - INFO - codeparrot_training - Step 42732: {'lr': 2.775570473405703e-05, 'samples': 21879296, 'steps': 42732, 'loss/train': 1.4984279870986938} 02/26/2022 08:54:19 - INFO - codeparrot_training - Step 42733: {'lr': 2.7748211996263845e-05, 'samples': 21879808, 'steps': 42733, 'loss/train': 0.18546174466609955} 02/26/2022 08:54:25 - INFO - codeparrot_training - Step 42734: {'lr': 2.774072021052676e-05, 'samples': 21880320, 'steps': 42734, 'loss/train': 1.2692204713821411} 02/26/2022 08:54:28 - INFO - codeparrot_training - Step 42735: {'lr': 2.773322937687786e-05, 'samples': 21880832, 'steps': 42735, 'loss/train': 1.0544992685317993} 02/26/2022 08:54:34 - INFO - codeparrot_training - Step 42736: {'lr': 2.772573949534918e-05, 'samples': 21881344, 'steps': 42736, 'loss/train': 2.1529440879821777} 02/26/2022 08:54:37 - INFO - codeparrot_training - Step 42737: {'lr': 2.771825056597291e-05, 'samples': 21881856, 'steps': 42737, 'loss/train': 1.9256763458251953} 02/26/2022 08:54:45 - INFO - codeparrot_training - Step 42738: {'lr': 2.7710762588781053e-05, 'samples': 21882368, 'steps': 42738, 'loss/train': 2.49788761138916} 02/26/2022 08:54:50 - INFO - codeparrot_training - Step 42739: {'lr': 2.7703275563805697e-05, 'samples': 21882880, 'steps': 42739, 'loss/train': 1.7890676259994507} 02/26/2022 08:54:54 - INFO - codeparrot_training - Step 42740: {'lr': 2.7695789491078925e-05, 'samples': 21883392, 'steps': 42740, 'loss/train': 3.541902542114258} 02/26/2022 08:54:57 - INFO - codeparrot_training - Step 42741: {'lr': 2.7688304370632773e-05, 'samples': 21883904, 'steps': 42741, 'loss/train': 0.9804078340530396} 02/26/2022 08:55:03 - INFO - codeparrot_training - Step 42742: {'lr': 2.7680820202499373e-05, 'samples': 21884416, 'steps': 42742, 'loss/train': 2.0154688358306885} 02/26/2022 08:55:08 - INFO - codeparrot_training - Step 42743: {'lr': 2.7673336986710733e-05, 'samples': 21884928, 'steps': 42743, 'loss/train': 1.085662841796875} 02/26/2022 08:55:12 - INFO - codeparrot_training - Step 42744: {'lr': 2.766585472329894e-05, 'samples': 21885440, 'steps': 42744, 'loss/train': 2.0492982864379883} 02/26/2022 08:55:17 - INFO - codeparrot_training - Step 42745: {'lr': 2.7658373412295962e-05, 'samples': 21885952, 'steps': 42745, 'loss/train': 1.88246750831604} 02/26/2022 08:55:21 - INFO - codeparrot_training - Step 42746: {'lr': 2.7650893053733972e-05, 'samples': 21886464, 'steps': 42746, 'loss/train': 2.7981507778167725} 02/26/2022 08:55:28 - INFO - codeparrot_training - Step 42747: {'lr': 2.7643413647644945e-05, 'samples': 21886976, 'steps': 42747, 'loss/train': 1.7878522872924805} 02/26/2022 08:55:31 - INFO - codeparrot_training - Step 42748: {'lr': 2.7635935194060933e-05, 'samples': 21887488, 'steps': 42748, 'loss/train': 1.369206428527832} 02/26/2022 08:55:35 - INFO - codeparrot_training - Step 42749: {'lr': 2.762845769301389e-05, 'samples': 21888000, 'steps': 42749, 'loss/train': 1.4847517013549805} 02/26/2022 08:55:40 - INFO - codeparrot_training - Step 42750: {'lr': 2.762098114453601e-05, 'samples': 21888512, 'steps': 42750, 'loss/train': 1.7825376987457275} 02/26/2022 08:55:46 - INFO - codeparrot_training - Step 42751: {'lr': 2.761350554865921e-05, 'samples': 21889024, 'steps': 42751, 'loss/train': 2.1239709854125977} 02/26/2022 08:55:49 - INFO - codeparrot_training - Step 42752: {'lr': 2.7606030905415552e-05, 'samples': 21889536, 'steps': 42752, 'loss/train': 1.5714926719665527} 02/26/2022 08:55:55 - INFO - codeparrot_training - Step 42753: {'lr': 2.7598557214836977e-05, 'samples': 21890048, 'steps': 42753, 'loss/train': 1.631894588470459} 02/26/2022 08:55:58 - INFO - codeparrot_training - Step 42754: {'lr': 2.759108447695563e-05, 'samples': 21890560, 'steps': 42754, 'loss/train': 1.4789563417434692} 02/26/2022 08:56:04 - INFO - codeparrot_training - Step 42755: {'lr': 2.7583612691803373e-05, 'samples': 21891072, 'steps': 42755, 'loss/train': 1.7448941469192505} 02/26/2022 08:56:07 - INFO - codeparrot_training - Step 42756: {'lr': 2.757614185941243e-05, 'samples': 21891584, 'steps': 42756, 'loss/train': 0.7392439842224121} 02/26/2022 08:56:13 - INFO - codeparrot_training - Step 42757: {'lr': 2.7568671979814554e-05, 'samples': 21892096, 'steps': 42757, 'loss/train': 1.2305270433425903} 02/26/2022 08:56:16 - INFO - codeparrot_training - Step 42758: {'lr': 2.7561203053041882e-05, 'samples': 21892608, 'steps': 42758, 'loss/train': 2.3158226013183594} 02/26/2022 08:56:22 - INFO - codeparrot_training - Step 42759: {'lr': 2.7553735079126368e-05, 'samples': 21893120, 'steps': 42759, 'loss/train': 0.6426398158073425} 02/26/2022 08:56:25 - INFO - codeparrot_training - Step 42760: {'lr': 2.7546268058100094e-05, 'samples': 21893632, 'steps': 42760, 'loss/train': 2.1408002376556396} 02/26/2022 08:56:31 - INFO - codeparrot_training - Step 42761: {'lr': 2.7538801989994894e-05, 'samples': 21894144, 'steps': 42761, 'loss/train': 0.8857948184013367} 02/26/2022 08:56:34 - INFO - codeparrot_training - Step 42762: {'lr': 2.7531336874842855e-05, 'samples': 21894656, 'steps': 42762, 'loss/train': 2.3472254276275635} 02/26/2022 08:56:41 - INFO - codeparrot_training - Step 42763: {'lr': 2.7523872712675896e-05, 'samples': 21895168, 'steps': 42763, 'loss/train': 1.4410910606384277} 02/26/2022 08:56:45 - INFO - codeparrot_training - Step 42764: {'lr': 2.751640950352613e-05, 'samples': 21895680, 'steps': 42764, 'loss/train': 0.9286631941795349} 02/26/2022 08:56:50 - INFO - codeparrot_training - Step 42765: {'lr': 2.7508947247425282e-05, 'samples': 21896192, 'steps': 42765, 'loss/train': 1.7183406352996826} 02/26/2022 08:56:54 - INFO - codeparrot_training - Step 42766: {'lr': 2.7501485944405547e-05, 'samples': 21896704, 'steps': 42766, 'loss/train': 2.250690460205078} 02/26/2022 08:56:59 - INFO - codeparrot_training - Step 42767: {'lr': 2.7494025594498707e-05, 'samples': 21897216, 'steps': 42767, 'loss/train': 2.3027243614196777} 02/26/2022 08:57:03 - INFO - codeparrot_training - Step 42768: {'lr': 2.7486566197736873e-05, 'samples': 21897728, 'steps': 42768, 'loss/train': 1.9800598621368408} 02/26/2022 08:57:08 - INFO - codeparrot_training - Step 42769: {'lr': 2.7479107754151937e-05, 'samples': 21898240, 'steps': 42769, 'loss/train': 1.6510326862335205} 02/26/2022 08:57:12 - INFO - codeparrot_training - Step 42770: {'lr': 2.7471650263775848e-05, 'samples': 21898752, 'steps': 42770, 'loss/train': 1.366326928138733} 02/26/2022 08:57:17 - INFO - codeparrot_training - Step 42771: {'lr': 2.7464193726640497e-05, 'samples': 21899264, 'steps': 42771, 'loss/train': 1.0782345533370972} 02/26/2022 08:57:21 - INFO - codeparrot_training - Step 42772: {'lr': 2.745673814277794e-05, 'samples': 21899776, 'steps': 42772, 'loss/train': 1.430353045463562} 02/26/2022 08:57:28 - INFO - codeparrot_training - Step 42773: {'lr': 2.744928351222001e-05, 'samples': 21900288, 'steps': 42773, 'loss/train': 1.9887620210647583} 02/26/2022 08:57:31 - INFO - codeparrot_training - Step 42774: {'lr': 2.744182983499871e-05, 'samples': 21900800, 'steps': 42774, 'loss/train': 1.608001947402954} 02/26/2022 08:57:37 - INFO - codeparrot_training - Step 42775: {'lr': 2.7434377111145908e-05, 'samples': 21901312, 'steps': 42775, 'loss/train': 1.6254725456237793} 02/26/2022 08:57:40 - INFO - codeparrot_training - Step 42776: {'lr': 2.7426925340693577e-05, 'samples': 21901824, 'steps': 42776, 'loss/train': 1.550337791442871} 02/26/2022 08:57:46 - INFO - codeparrot_training - Step 42777: {'lr': 2.7419474523673633e-05, 'samples': 21902336, 'steps': 42777, 'loss/train': 1.4018499851226807} 02/26/2022 08:57:49 - INFO - codeparrot_training - Step 42778: {'lr': 2.7412024660117997e-05, 'samples': 21902848, 'steps': 42778, 'loss/train': 1.440045952796936} 02/26/2022 08:57:55 - INFO - codeparrot_training - Step 42779: {'lr': 2.7404575750058503e-05, 'samples': 21903360, 'steps': 42779, 'loss/train': 1.0258502960205078} 02/26/2022 08:57:58 - INFO - codeparrot_training - Step 42780: {'lr': 2.7397127793527183e-05, 'samples': 21903872, 'steps': 42780, 'loss/train': 1.6109308004379272} 02/26/2022 08:58:04 - INFO - codeparrot_training - Step 42781: {'lr': 2.7389680790555872e-05, 'samples': 21904384, 'steps': 42781, 'loss/train': 2.0095551013946533} 02/26/2022 08:58:07 - INFO - codeparrot_training - Step 42782: {'lr': 2.738223474117649e-05, 'samples': 21904896, 'steps': 42782, 'loss/train': 2.636659622192383} 02/26/2022 08:58:15 - INFO - codeparrot_training - Step 42783: {'lr': 2.7374789645420895e-05, 'samples': 21905408, 'steps': 42783, 'loss/train': 0.7914487719535828} 02/26/2022 08:58:18 - INFO - codeparrot_training - Step 42784: {'lr': 2.736734550332104e-05, 'samples': 21905920, 'steps': 42784, 'loss/train': 2.005279064178467} 02/26/2022 08:58:24 - INFO - codeparrot_training - Step 42785: {'lr': 2.7359902314908758e-05, 'samples': 21906432, 'steps': 42785, 'loss/train': 2.5495150089263916} 02/26/2022 08:58:27 - INFO - codeparrot_training - Step 42786: {'lr': 2.7352460080215995e-05, 'samples': 21906944, 'steps': 42786, 'loss/train': 1.6117534637451172} 02/26/2022 08:58:33 - INFO - codeparrot_training - Step 42787: {'lr': 2.734501879927459e-05, 'samples': 21907456, 'steps': 42787, 'loss/train': 1.3697905540466309} 02/26/2022 08:58:36 - INFO - codeparrot_training - Step 42788: {'lr': 2.7337578472116348e-05, 'samples': 21907968, 'steps': 42788, 'loss/train': 1.1379101276397705} 02/26/2022 08:58:42 - INFO - codeparrot_training - Step 42789: {'lr': 2.73301390987733e-05, 'samples': 21908480, 'steps': 42789, 'loss/train': 2.1402475833892822} 02/26/2022 08:58:45 - INFO - codeparrot_training - Step 42790: {'lr': 2.7322700679277223e-05, 'samples': 21908992, 'steps': 42790, 'loss/train': 1.7058736085891724} 02/26/2022 08:58:51 - INFO - codeparrot_training - Step 42791: {'lr': 2.731526321365996e-05, 'samples': 21909504, 'steps': 42791, 'loss/train': 2.434065580368042} 02/26/2022 08:58:54 - INFO - codeparrot_training - Step 42792: {'lr': 2.7307826701953392e-05, 'samples': 21910016, 'steps': 42792, 'loss/train': 1.7840499877929688} 02/26/2022 08:59:00 - INFO - codeparrot_training - Step 42793: {'lr': 2.730039114418939e-05, 'samples': 21910528, 'steps': 42793, 'loss/train': 2.0112526416778564} 02/26/2022 08:59:03 - INFO - codeparrot_training - Step 42794: {'lr': 2.729295654039976e-05, 'samples': 21911040, 'steps': 42794, 'loss/train': 1.910519003868103} 02/26/2022 08:59:11 - INFO - codeparrot_training - Step 42795: {'lr': 2.7285522890616504e-05, 'samples': 21911552, 'steps': 42795, 'loss/train': 1.799712061882019} 02/26/2022 08:59:14 - INFO - codeparrot_training - Step 42796: {'lr': 2.7278090194871237e-05, 'samples': 21912064, 'steps': 42796, 'loss/train': 2.5501792430877686} 02/26/2022 08:59:19 - INFO - codeparrot_training - Step 42797: {'lr': 2.7270658453195957e-05, 'samples': 21912576, 'steps': 42797, 'loss/train': 2.504366636276245} 02/26/2022 08:59:23 - INFO - codeparrot_training - Step 42798: {'lr': 2.726322766562242e-05, 'samples': 21913088, 'steps': 42798, 'loss/train': 1.3743587732315063} 02/26/2022 08:59:28 - INFO - codeparrot_training - Step 42799: {'lr': 2.7255797832182572e-05, 'samples': 21913600, 'steps': 42799, 'loss/train': 1.4520469903945923} 02/26/2022 08:59:32 - INFO - codeparrot_training - Step 42800: {'lr': 2.7248368952908055e-05, 'samples': 21914112, 'steps': 42800, 'loss/train': 2.395886182785034} 02/26/2022 08:59:37 - INFO - codeparrot_training - Step 42801: {'lr': 2.724094102783084e-05, 'samples': 21914624, 'steps': 42801, 'loss/train': 1.3224307298660278} 02/26/2022 08:59:41 - INFO - codeparrot_training - Step 42802: {'lr': 2.7233514056982655e-05, 'samples': 21915136, 'steps': 42802, 'loss/train': 2.0678722858428955} 02/26/2022 08:59:46 - INFO - codeparrot_training - Step 42803: {'lr': 2.722608804039542e-05, 'samples': 21915648, 'steps': 42803, 'loss/train': 0.7412338256835938} 02/26/2022 08:59:52 - INFO - codeparrot_training - Step 42804: {'lr': 2.7218662978100854e-05, 'samples': 21916160, 'steps': 42804, 'loss/train': 1.9862208366394043} 02/26/2022 08:59:55 - INFO - codeparrot_training - Step 42805: {'lr': 2.7211238870130826e-05, 'samples': 21916672, 'steps': 42805, 'loss/train': 1.9224451780319214} 02/26/2022 08:59:59 - INFO - codeparrot_training - Step 42806: {'lr': 2.7203815716517062e-05, 'samples': 21917184, 'steps': 42806, 'loss/train': 1.2230496406555176} 02/26/2022 09:00:04 - INFO - codeparrot_training - Step 42807: {'lr': 2.7196393517291417e-05, 'samples': 21917696, 'steps': 42807, 'loss/train': 1.4456367492675781} 02/26/2022 09:00:12 - INFO - codeparrot_training - Step 42808: {'lr': 2.718897227248571e-05, 'samples': 21918208, 'steps': 42808, 'loss/train': 0.760595440864563} 02/26/2022 09:00:15 - INFO - codeparrot_training - Step 42809: {'lr': 2.718155198213168e-05, 'samples': 21918720, 'steps': 42809, 'loss/train': 1.357527256011963} 02/26/2022 09:00:19 - INFO - codeparrot_training - Step 42810: {'lr': 2.717413264626109e-05, 'samples': 21919232, 'steps': 42810, 'loss/train': 1.1176189184188843} 02/26/2022 09:00:24 - INFO - codeparrot_training - Step 42811: {'lr': 2.71667142649058e-05, 'samples': 21919744, 'steps': 42811, 'loss/train': 0.6460078954696655} 02/26/2022 09:00:28 - INFO - codeparrot_training - Step 42812: {'lr': 2.7159296838097565e-05, 'samples': 21920256, 'steps': 42812, 'loss/train': 3.394054889678955} 02/26/2022 09:00:33 - INFO - codeparrot_training - Step 42813: {'lr': 2.715188036586813e-05, 'samples': 21920768, 'steps': 42813, 'loss/train': 2.3883965015411377} 02/26/2022 09:00:37 - INFO - codeparrot_training - Step 42814: {'lr': 2.71444648482492e-05, 'samples': 21921280, 'steps': 42814, 'loss/train': 2.1934337615966797} 02/26/2022 09:00:42 - INFO - codeparrot_training - Step 42815: {'lr': 2.713705028527272e-05, 'samples': 21921792, 'steps': 42815, 'loss/train': 1.7336368560791016} 02/26/2022 09:00:46 - INFO - codeparrot_training - Step 42816: {'lr': 2.712963667697027e-05, 'samples': 21922304, 'steps': 42816, 'loss/train': 2.1103127002716064} 02/26/2022 09:00:51 - INFO - codeparrot_training - Step 42817: {'lr': 2.7122224023373778e-05, 'samples': 21922816, 'steps': 42817, 'loss/train': 1.9909189939498901} 02/26/2022 09:00:55 - INFO - codeparrot_training - Step 42818: {'lr': 2.7114812324514822e-05, 'samples': 21923328, 'steps': 42818, 'loss/train': 0.8142868876457214} 02/26/2022 09:01:02 - INFO - codeparrot_training - Step 42819: {'lr': 2.7107401580425295e-05, 'samples': 21923840, 'steps': 42819, 'loss/train': 1.324966311454773} 02/26/2022 09:01:06 - INFO - codeparrot_training - Step 42820: {'lr': 2.7099991791136867e-05, 'samples': 21924352, 'steps': 42820, 'loss/train': 1.7044597864151} 02/26/2022 09:01:11 - INFO - codeparrot_training - Step 42821: {'lr': 2.7092582956681294e-05, 'samples': 21924864, 'steps': 42821, 'loss/train': 1.8956921100616455} 02/26/2022 09:01:15 - INFO - codeparrot_training - Step 42822: {'lr': 2.708517507709027e-05, 'samples': 21925376, 'steps': 42822, 'loss/train': 0.36719071865081787} 02/26/2022 09:01:20 - INFO - codeparrot_training - Step 42823: {'lr': 2.7077768152395626e-05, 'samples': 21925888, 'steps': 42823, 'loss/train': 0.9787956476211548} 02/26/2022 09:01:23 - INFO - codeparrot_training - Step 42824: {'lr': 2.7070362182629038e-05, 'samples': 21926400, 'steps': 42824, 'loss/train': 1.921280026435852} 02/26/2022 09:01:29 - INFO - codeparrot_training - Step 42825: {'lr': 2.706295716782223e-05, 'samples': 21926912, 'steps': 42825, 'loss/train': 1.9951118230819702} 02/26/2022 09:01:33 - INFO - codeparrot_training - Step 42826: {'lr': 2.7055553108006866e-05, 'samples': 21927424, 'steps': 42826, 'loss/train': 1.4546281099319458} 02/26/2022 09:01:38 - INFO - codeparrot_training - Step 42827: {'lr': 2.7048150003214784e-05, 'samples': 21927936, 'steps': 42827, 'loss/train': 0.6301075220108032} 02/26/2022 09:01:42 - INFO - codeparrot_training - Step 42828: {'lr': 2.70407478534776e-05, 'samples': 21928448, 'steps': 42828, 'loss/train': 1.3864514827728271} 02/26/2022 09:01:47 - INFO - codeparrot_training - Step 42829: {'lr': 2.7033346658827034e-05, 'samples': 21928960, 'steps': 42829, 'loss/train': 0.6658378839492798} 02/26/2022 09:01:51 - INFO - codeparrot_training - Step 42830: {'lr': 2.7025946419294845e-05, 'samples': 21929472, 'steps': 42830, 'loss/train': 5.909066200256348} 02/26/2022 09:01:58 - INFO - codeparrot_training - Step 42831: {'lr': 2.7018547134912725e-05, 'samples': 21929984, 'steps': 42831, 'loss/train': 0.8054854273796082} 02/26/2022 09:02:01 - INFO - codeparrot_training - Step 42832: {'lr': 2.7011148805712316e-05, 'samples': 21930496, 'steps': 42832, 'loss/train': 1.828218936920166} 02/26/2022 09:02:07 - INFO - codeparrot_training - Step 42833: {'lr': 2.7003751431725314e-05, 'samples': 21931008, 'steps': 42833, 'loss/train': 0.6490828990936279} 02/26/2022 09:02:10 - INFO - codeparrot_training - Step 42834: {'lr': 2.6996355012983502e-05, 'samples': 21931520, 'steps': 42834, 'loss/train': 2.3491430282592773} 02/26/2022 09:02:16 - INFO - codeparrot_training - Step 42835: {'lr': 2.698895954951841e-05, 'samples': 21932032, 'steps': 42835, 'loss/train': 2.0695760250091553} 02/26/2022 09:02:20 - INFO - codeparrot_training - Step 42836: {'lr': 2.6981565041361873e-05, 'samples': 21932544, 'steps': 42836, 'loss/train': 1.6444177627563477} 02/26/2022 09:02:25 - INFO - codeparrot_training - Step 42837: {'lr': 2.697417148854542e-05, 'samples': 21933056, 'steps': 42837, 'loss/train': 1.3091180324554443} 02/26/2022 09:02:29 - INFO - codeparrot_training - Step 42838: {'lr': 2.6966778891100884e-05, 'samples': 21933568, 'steps': 42838, 'loss/train': 2.7320001125335693} 02/26/2022 09:02:34 - INFO - codeparrot_training - Step 42839: {'lr': 2.6959387249059775e-05, 'samples': 21934080, 'steps': 42839, 'loss/train': 1.8345000743865967} 02/26/2022 09:02:41 - INFO - codeparrot_training - Step 42840: {'lr': 2.6951996562453866e-05, 'samples': 21934592, 'steps': 42840, 'loss/train': 1.17380952835083} 02/26/2022 09:02:45 - INFO - codeparrot_training - Step 42841: {'lr': 2.6944606831314722e-05, 'samples': 21935104, 'steps': 42841, 'loss/train': 1.2819693088531494} 02/26/2022 09:02:50 - INFO - codeparrot_training - Step 42842: {'lr': 2.6937218055674116e-05, 'samples': 21935616, 'steps': 42842, 'loss/train': 0.9826082587242126} 02/26/2022 09:02:54 - INFO - codeparrot_training - Step 42843: {'lr': 2.6929830235563613e-05, 'samples': 21936128, 'steps': 42843, 'loss/train': 0.06444505602121353} 02/26/2022 09:02:59 - INFO - codeparrot_training - Step 42844: {'lr': 2.6922443371014904e-05, 'samples': 21936640, 'steps': 42844, 'loss/train': 2.2556002140045166} 02/26/2022 09:03:03 - INFO - codeparrot_training - Step 42845: {'lr': 2.6915057462059578e-05, 'samples': 21937152, 'steps': 42845, 'loss/train': 1.4239226579666138} 02/26/2022 09:03:08 - INFO - codeparrot_training - Step 42846: {'lr': 2.690767250872933e-05, 'samples': 21937664, 'steps': 42846, 'loss/train': 1.681216835975647} 02/26/2022 09:03:12 - INFO - codeparrot_training - Step 42847: {'lr': 2.6900288511055775e-05, 'samples': 21938176, 'steps': 42847, 'loss/train': 8.6160249710083} 02/26/2022 09:03:17 - INFO - codeparrot_training - Step 42848: {'lr': 2.6892905469070554e-05, 'samples': 21938688, 'steps': 42848, 'loss/train': 2.25268292427063} 02/26/2022 09:03:21 - INFO - codeparrot_training - Step 42849: {'lr': 2.6885523382805226e-05, 'samples': 21939200, 'steps': 42849, 'loss/train': 0.38531386852264404} 02/26/2022 09:03:27 - INFO - codeparrot_training - Step 42850: {'lr': 2.6878142252291515e-05, 'samples': 21939712, 'steps': 42850, 'loss/train': 0.6773548126220703} 02/26/2022 09:03:30 - INFO - codeparrot_training - Step 42851: {'lr': 2.687076207756095e-05, 'samples': 21940224, 'steps': 42851, 'loss/train': 1.1235246658325195} 02/26/2022 09:03:34 - INFO - codeparrot_training - Step 42852: {'lr': 2.6863382858645313e-05, 'samples': 21940736, 'steps': 42852, 'loss/train': 2.1405186653137207} 02/26/2022 09:03:39 - INFO - codeparrot_training - Step 42853: {'lr': 2.6856004595575966e-05, 'samples': 21941248, 'steps': 42853, 'loss/train': 2.213709831237793} 02/26/2022 09:03:43 - INFO - codeparrot_training - Step 42854: {'lr': 2.6848627288384665e-05, 'samples': 21941760, 'steps': 42854, 'loss/train': 1.7483052015304565} 02/26/2022 09:03:48 - INFO - codeparrot_training - Step 42855: {'lr': 2.684125093710296e-05, 'samples': 21942272, 'steps': 42855, 'loss/train': 1.7833229303359985} 02/26/2022 09:03:52 - INFO - codeparrot_training - Step 42856: {'lr': 2.6833875541762587e-05, 'samples': 21942784, 'steps': 42856, 'loss/train': 1.3186804056167603} 02/26/2022 09:03:59 - INFO - codeparrot_training - Step 42857: {'lr': 2.682650110239493e-05, 'samples': 21943296, 'steps': 42857, 'loss/train': 1.5045169591903687} 02/26/2022 09:04:02 - INFO - codeparrot_training - Step 42858: {'lr': 2.6819127619031687e-05, 'samples': 21943808, 'steps': 42858, 'loss/train': 1.8732916116714478} 02/26/2022 09:04:08 - INFO - codeparrot_training - Step 42859: {'lr': 2.6811755091704415e-05, 'samples': 21944320, 'steps': 42859, 'loss/train': 1.7660460472106934} 02/26/2022 09:04:11 - INFO - codeparrot_training - Step 42860: {'lr': 2.6804383520444812e-05, 'samples': 21944832, 'steps': 42860, 'loss/train': 1.4170451164245605} 02/26/2022 09:04:19 - INFO - codeparrot_training - Step 42861: {'lr': 2.6797012905284247e-05, 'samples': 21945344, 'steps': 42861, 'loss/train': 1.2842391729354858} 02/26/2022 09:04:22 - INFO - codeparrot_training - Step 42862: {'lr': 2.6789643246254463e-05, 'samples': 21945856, 'steps': 42862, 'loss/train': 2.146132707595825} 02/26/2022 09:04:28 - INFO - codeparrot_training - Step 42863: {'lr': 2.6782274543386914e-05, 'samples': 21946368, 'steps': 42863, 'loss/train': 2.29295015335083} 02/26/2022 09:04:31 - INFO - codeparrot_training - Step 42864: {'lr': 2.6774906796713295e-05, 'samples': 21946880, 'steps': 42864, 'loss/train': 0.8187737464904785} 02/26/2022 09:04:37 - INFO - codeparrot_training - Step 42865: {'lr': 2.6767540006265052e-05, 'samples': 21947392, 'steps': 42865, 'loss/train': 1.1048065423965454} 02/26/2022 09:04:40 - INFO - codeparrot_training - Step 42866: {'lr': 2.6760174172073826e-05, 'samples': 21947904, 'steps': 42866, 'loss/train': 3.348705530166626} 02/26/2022 09:04:46 - INFO - codeparrot_training - Step 42867: {'lr': 2.6752809294171094e-05, 'samples': 21948416, 'steps': 42867, 'loss/train': 1.3998491764068604} 02/26/2022 09:04:49 - INFO - codeparrot_training - Step 42868: {'lr': 2.674544537258841e-05, 'samples': 21948928, 'steps': 42868, 'loss/train': 2.9467873573303223} 02/26/2022 09:04:55 - INFO - codeparrot_training - Step 42869: {'lr': 2.6738082407357367e-05, 'samples': 21949440, 'steps': 42869, 'loss/train': 1.3641581535339355} 02/26/2022 09:04:58 - INFO - codeparrot_training - Step 42870: {'lr': 2.6730720398509516e-05, 'samples': 21949952, 'steps': 42870, 'loss/train': 1.601888656616211} 02/26/2022 09:05:04 - INFO - codeparrot_training - Step 42871: {'lr': 2.6723359346076338e-05, 'samples': 21950464, 'steps': 42871, 'loss/train': 1.9685685634613037} 02/26/2022 09:05:07 - INFO - codeparrot_training - Step 42872: {'lr': 2.6715999250089358e-05, 'samples': 21950976, 'steps': 42872, 'loss/train': 2.945636749267578} 02/26/2022 09:05:15 - INFO - codeparrot_training - Step 42873: {'lr': 2.6708640110580192e-05, 'samples': 21951488, 'steps': 42873, 'loss/train': 2.2861685752868652} 02/26/2022 09:05:19 - INFO - codeparrot_training - Step 42874: {'lr': 2.6701281927580284e-05, 'samples': 21952000, 'steps': 42874, 'loss/train': 0.08399217575788498} 02/26/2022 09:05:24 - INFO - codeparrot_training - Step 42875: {'lr': 2.6693924701121196e-05, 'samples': 21952512, 'steps': 42875, 'loss/train': 1.6448696851730347} 02/26/2022 09:05:27 - INFO - codeparrot_training - Step 42876: {'lr': 2.6686568431234376e-05, 'samples': 21953024, 'steps': 42876, 'loss/train': 1.7844494581222534} 02/26/2022 09:05:33 - INFO - codeparrot_training - Step 42877: {'lr': 2.6679213117951434e-05, 'samples': 21953536, 'steps': 42877, 'loss/train': 2.702091693878174} 02/26/2022 09:05:38 - INFO - codeparrot_training - Step 42878: {'lr': 2.6671858761303848e-05, 'samples': 21954048, 'steps': 42878, 'loss/train': 1.927246332168579} 02/26/2022 09:05:42 - INFO - codeparrot_training - Step 42879: {'lr': 2.666450536132309e-05, 'samples': 21954560, 'steps': 42879, 'loss/train': 1.785115361213684} 02/26/2022 09:05:47 - INFO - codeparrot_training - Step 42880: {'lr': 2.6657152918040605e-05, 'samples': 21955072, 'steps': 42880, 'loss/train': 2.2215871810913086} 02/26/2022 09:05:51 - INFO - codeparrot_training - Step 42881: {'lr': 2.6649801431488043e-05, 'samples': 21955584, 'steps': 42881, 'loss/train': 2.1950230598449707} 02/26/2022 09:05:58 - INFO - codeparrot_training - Step 42882: {'lr': 2.6642450901696785e-05, 'samples': 21956096, 'steps': 42882, 'loss/train': 1.02681303024292} 02/26/2022 09:06:02 - INFO - codeparrot_training - Step 42883: {'lr': 2.663510132869834e-05, 'samples': 21956608, 'steps': 42883, 'loss/train': 2.010436534881592} 02/26/2022 09:06:07 - INFO - codeparrot_training - Step 42884: {'lr': 2.6627752712524157e-05, 'samples': 21957120, 'steps': 42884, 'loss/train': 1.6045893430709839} 02/26/2022 09:06:11 - INFO - codeparrot_training - Step 42885: {'lr': 2.6620405053205816e-05, 'samples': 21957632, 'steps': 42885, 'loss/train': 0.755906343460083} 02/26/2022 09:06:16 - INFO - codeparrot_training - Step 42886: {'lr': 2.661305835077471e-05, 'samples': 21958144, 'steps': 42886, 'loss/train': 1.079380750656128} 02/26/2022 09:06:20 - INFO - codeparrot_training - Step 42887: {'lr': 2.6605712605262315e-05, 'samples': 21958656, 'steps': 42887, 'loss/train': 0.6011679768562317} 02/26/2022 09:06:25 - INFO - codeparrot_training - Step 42888: {'lr': 2.659836781670008e-05, 'samples': 21959168, 'steps': 42888, 'loss/train': 0.6199708580970764} 02/26/2022 09:06:29 - INFO - codeparrot_training - Step 42889: {'lr': 2.659102398511956e-05, 'samples': 21959680, 'steps': 42889, 'loss/train': 1.0654090642929077} 02/26/2022 09:06:34 - INFO - codeparrot_training - Step 42890: {'lr': 2.6583681110552093e-05, 'samples': 21960192, 'steps': 42890, 'loss/train': 0.5442640781402588} 02/26/2022 09:06:38 - INFO - codeparrot_training - Step 42891: {'lr': 2.6576339193029293e-05, 'samples': 21960704, 'steps': 42891, 'loss/train': 1.217951774597168} 02/26/2022 09:06:45 - INFO - codeparrot_training - Step 42892: {'lr': 2.656899823258241e-05, 'samples': 21961216, 'steps': 42892, 'loss/train': 1.5691418647766113} 02/26/2022 09:06:48 - INFO - codeparrot_training - Step 42893: {'lr': 2.6561658229243028e-05, 'samples': 21961728, 'steps': 42893, 'loss/train': 2.1433398723602295} 02/26/2022 09:06:54 - INFO - codeparrot_training - Step 42894: {'lr': 2.655431918304252e-05, 'samples': 21962240, 'steps': 42894, 'loss/train': 2.3932013511657715} 02/26/2022 09:06:57 - INFO - codeparrot_training - Step 42895: {'lr': 2.654698109401246e-05, 'samples': 21962752, 'steps': 42895, 'loss/train': 1.721458077430725} 02/26/2022 09:07:03 - INFO - codeparrot_training - Step 42896: {'lr': 2.6539643962184058e-05, 'samples': 21963264, 'steps': 42896, 'loss/train': 1.9035429954528809} 02/26/2022 09:07:06 - INFO - codeparrot_training - Step 42897: {'lr': 2.6532307787588943e-05, 'samples': 21963776, 'steps': 42897, 'loss/train': 2.4287753105163574} 02/26/2022 09:07:12 - INFO - codeparrot_training - Step 42898: {'lr': 2.6524972570258377e-05, 'samples': 21964288, 'steps': 42898, 'loss/train': 1.8739185333251953} 02/26/2022 09:07:15 - INFO - codeparrot_training - Step 42899: {'lr': 2.6517638310223996e-05, 'samples': 21964800, 'steps': 42899, 'loss/train': 1.8823384046554565} 02/26/2022 09:07:21 - INFO - codeparrot_training - Step 42900: {'lr': 2.6510305007516973e-05, 'samples': 21965312, 'steps': 42900, 'loss/train': 1.3579708337783813} 02/26/2022 09:07:24 - INFO - codeparrot_training - Step 42901: {'lr': 2.6502972662168888e-05, 'samples': 21965824, 'steps': 42901, 'loss/train': 2.5500195026397705} 02/26/2022 09:07:31 - INFO - codeparrot_training - Step 42902: {'lr': 2.6495641274211057e-05, 'samples': 21966336, 'steps': 42902, 'loss/train': 1.8145298957824707} 02/26/2022 09:07:35 - INFO - codeparrot_training - Step 42903: {'lr': 2.648831084367498e-05, 'samples': 21966848, 'steps': 42903, 'loss/train': 1.9289880990982056} 02/26/2022 09:07:40 - INFO - codeparrot_training - Step 42904: {'lr': 2.6480981370592016e-05, 'samples': 21967360, 'steps': 42904, 'loss/train': 2.3520002365112305} 02/26/2022 09:07:44 - INFO - codeparrot_training - Step 42905: {'lr': 2.6473652854993535e-05, 'samples': 21967872, 'steps': 42905, 'loss/train': 1.8150992393493652} 02/26/2022 09:07:49 - INFO - codeparrot_training - Step 42906: {'lr': 2.64663252969109e-05, 'samples': 21968384, 'steps': 42906, 'loss/train': 2.84921932220459} 02/26/2022 09:07:53 - INFO - codeparrot_training - Step 42907: {'lr': 2.6458998696375608e-05, 'samples': 21968896, 'steps': 42907, 'loss/train': 1.309278964996338} 02/26/2022 09:07:58 - INFO - codeparrot_training - Step 42908: {'lr': 2.6451673053418972e-05, 'samples': 21969408, 'steps': 42908, 'loss/train': 3.8604185581207275} 02/26/2022 09:08:02 - INFO - codeparrot_training - Step 42909: {'lr': 2.6444348368072385e-05, 'samples': 21969920, 'steps': 42909, 'loss/train': 1.825760841369629} 02/26/2022 09:08:07 - INFO - codeparrot_training - Step 42910: {'lr': 2.643702464036718e-05, 'samples': 21970432, 'steps': 42910, 'loss/train': 2.179658889770508} 02/26/2022 09:08:11 - INFO - codeparrot_training - Step 42911: {'lr': 2.6429701870334805e-05, 'samples': 21970944, 'steps': 42911, 'loss/train': 0.33878087997436523} 02/26/2022 09:08:16 - INFO - codeparrot_training - Step 42912: {'lr': 2.6422380058006597e-05, 'samples': 21971456, 'steps': 42912, 'loss/train': 1.795589804649353} 02/26/2022 09:08:20 - INFO - codeparrot_training - Step 42913: {'lr': 2.6415059203413944e-05, 'samples': 21971968, 'steps': 42913, 'loss/train': 0.8963598608970642} 02/26/2022 09:08:25 - INFO - codeparrot_training - Step 42914: {'lr': 2.6407739306588103e-05, 'samples': 21972480, 'steps': 42914, 'loss/train': 1.4238911867141724} 02/26/2022 09:08:29 - INFO - codeparrot_training - Step 42915: {'lr': 2.6400420367560573e-05, 'samples': 21972992, 'steps': 42915, 'loss/train': 1.0536459684371948} 02/26/2022 09:08:35 - INFO - codeparrot_training - Step 42916: {'lr': 2.639310238636264e-05, 'samples': 21973504, 'steps': 42916, 'loss/train': 1.2638682126998901} 02/26/2022 09:08:38 - INFO - codeparrot_training - Step 42917: {'lr': 2.638578536302566e-05, 'samples': 21974016, 'steps': 42917, 'loss/train': 0.14891071617603302} 02/26/2022 09:08:45 - INFO - codeparrot_training - Step 42918: {'lr': 2.6378469297580977e-05, 'samples': 21974528, 'steps': 42918, 'loss/train': 2.00984787940979} 02/26/2022 09:08:49 - INFO - codeparrot_training - Step 42919: {'lr': 2.6371154190059838e-05, 'samples': 21975040, 'steps': 42919, 'loss/train': 1.3349881172180176} 02/26/2022 09:08:54 - INFO - codeparrot_training - Step 42920: {'lr': 2.6363840040493748e-05, 'samples': 21975552, 'steps': 42920, 'loss/train': 2.069300889968872} 02/26/2022 09:08:58 - INFO - codeparrot_training - Step 42921: {'lr': 2.635652684891393e-05, 'samples': 21976064, 'steps': 42921, 'loss/train': 1.7318470478057861} 02/26/2022 09:09:03 - INFO - codeparrot_training - Step 42922: {'lr': 2.6349214615351754e-05, 'samples': 21976576, 'steps': 42922, 'loss/train': 2.430473804473877} 02/26/2022 09:09:07 - INFO - codeparrot_training - Step 42923: {'lr': 2.6341903339838462e-05, 'samples': 21977088, 'steps': 42923, 'loss/train': 6.047491073608398} 02/26/2022 09:09:12 - INFO - codeparrot_training - Step 42924: {'lr': 2.6334593022405513e-05, 'samples': 21977600, 'steps': 42924, 'loss/train': 0.8418083786964417} 02/26/2022 09:09:16 - INFO - codeparrot_training - Step 42925: {'lr': 2.6327283663084067e-05, 'samples': 21978112, 'steps': 42925, 'loss/train': 3.577162504196167} 02/26/2022 09:09:21 - INFO - codeparrot_training - Step 42926: {'lr': 2.6319975261905634e-05, 'samples': 21978624, 'steps': 42926, 'loss/train': 1.8548102378845215} 02/26/2022 09:09:25 - INFO - codeparrot_training - Step 42927: {'lr': 2.6312667818901292e-05, 'samples': 21979136, 'steps': 42927, 'loss/train': 2.1302738189697266} 02/26/2022 09:09:32 - INFO - codeparrot_training - Step 42928: {'lr': 2.6305361334102497e-05, 'samples': 21979648, 'steps': 42928, 'loss/train': 1.8433518409729004} 02/26/2022 09:09:36 - INFO - codeparrot_training - Step 42929: {'lr': 2.6298055807540465e-05, 'samples': 21980160, 'steps': 42929, 'loss/train': 1.9911000728607178} 02/26/2022 09:09:41 - INFO - codeparrot_training - Step 42930: {'lr': 2.6290751239246623e-05, 'samples': 21980672, 'steps': 42930, 'loss/train': 1.6666487455368042} 02/26/2022 09:09:45 - INFO - codeparrot_training - Step 42931: {'lr': 2.6283447629252054e-05, 'samples': 21981184, 'steps': 42931, 'loss/train': 1.787108063697815} 02/26/2022 09:09:50 - INFO - codeparrot_training - Step 42932: {'lr': 2.6276144977588234e-05, 'samples': 21981696, 'steps': 42932, 'loss/train': 2.0184366703033447} 02/26/2022 09:09:54 - INFO - codeparrot_training - Step 42933: {'lr': 2.6268843284286303e-05, 'samples': 21982208, 'steps': 42933, 'loss/train': 1.377815842628479} 02/26/2022 09:09:59 - INFO - codeparrot_training - Step 42934: {'lr': 2.6261542549377683e-05, 'samples': 21982720, 'steps': 42934, 'loss/train': 0.7126639485359192} 02/26/2022 09:10:03 - INFO - codeparrot_training - Step 42935: {'lr': 2.625424277289351e-05, 'samples': 21983232, 'steps': 42935, 'loss/train': 1.7294493913650513} 02/26/2022 09:10:08 - INFO - codeparrot_training - Step 42936: {'lr': 2.6246943954865126e-05, 'samples': 21983744, 'steps': 42936, 'loss/train': 0.9614233374595642} 02/26/2022 09:10:12 - INFO - codeparrot_training - Step 42937: {'lr': 2.6239646095323754e-05, 'samples': 21984256, 'steps': 42937, 'loss/train': 1.8178631067276} 02/26/2022 09:10:17 - INFO - codeparrot_training - Step 42938: {'lr': 2.623234919430076e-05, 'samples': 21984768, 'steps': 42938, 'loss/train': 1.7760825157165527} 02/26/2022 09:10:21 - INFO - codeparrot_training - Step 42939: {'lr': 2.622505325182728e-05, 'samples': 21985280, 'steps': 42939, 'loss/train': 1.433170199394226} 02/26/2022 09:10:28 - INFO - codeparrot_training - Step 42940: {'lr': 2.6217758267934655e-05, 'samples': 21985792, 'steps': 42940, 'loss/train': 3.0513675212860107} 02/26/2022 09:10:31 - INFO - codeparrot_training - Step 42941: {'lr': 2.6210464242654025e-05, 'samples': 21986304, 'steps': 42941, 'loss/train': 1.539406657218933} 02/26/2022 09:10:37 - INFO - codeparrot_training - Step 42942: {'lr': 2.6203171176016754e-05, 'samples': 21986816, 'steps': 42942, 'loss/train': 2.441770553588867} 02/26/2022 09:10:40 - INFO - codeparrot_training - Step 42943: {'lr': 2.619587906805404e-05, 'samples': 21987328, 'steps': 42943, 'loss/train': 2.55214262008667} 02/26/2022 09:10:46 - INFO - codeparrot_training - Step 42944: {'lr': 2.618858791879711e-05, 'samples': 21987840, 'steps': 42944, 'loss/train': 1.7590361833572388} 02/26/2022 09:10:49 - INFO - codeparrot_training - Step 42945: {'lr': 2.6181297728277154e-05, 'samples': 21988352, 'steps': 42945, 'loss/train': 2.0431292057037354} 02/26/2022 09:10:55 - INFO - codeparrot_training - Step 42946: {'lr': 2.6174008496525515e-05, 'samples': 21988864, 'steps': 42946, 'loss/train': 1.5398809909820557} 02/26/2022 09:10:58 - INFO - codeparrot_training - Step 42947: {'lr': 2.6166720223573333e-05, 'samples': 21989376, 'steps': 42947, 'loss/train': 1.4127390384674072} 02/26/2022 09:11:04 - INFO - codeparrot_training - Step 42948: {'lr': 2.615943290945183e-05, 'samples': 21989888, 'steps': 42948, 'loss/train': 2.0602078437805176} 02/26/2022 09:11:07 - INFO - codeparrot_training - Step 42949: {'lr': 2.6152146554192206e-05, 'samples': 21990400, 'steps': 42949, 'loss/train': 0.34139397740364075} 02/26/2022 09:11:15 - INFO - codeparrot_training - Step 42950: {'lr': 2.6144861157825773e-05, 'samples': 21990912, 'steps': 42950, 'loss/train': 2.181121349334717} 02/26/2022 09:11:18 - INFO - codeparrot_training - Step 42951: {'lr': 2.613757672038364e-05, 'samples': 21991424, 'steps': 42951, 'loss/train': 1.5091323852539062} 02/26/2022 09:11:24 - INFO - codeparrot_training - Step 42952: {'lr': 2.613029324189706e-05, 'samples': 21991936, 'steps': 42952, 'loss/train': 1.7841856479644775} 02/26/2022 09:11:27 - INFO - codeparrot_training - Step 42953: {'lr': 2.6123010722397173e-05, 'samples': 21992448, 'steps': 42953, 'loss/train': 2.0733344554901123} 02/26/2022 09:11:32 - INFO - codeparrot_training - Step 42954: {'lr': 2.6115729161915265e-05, 'samples': 21992960, 'steps': 42954, 'loss/train': 1.8553426265716553} 02/26/2022 09:11:36 - INFO - codeparrot_training - Step 42955: {'lr': 2.6108448560482474e-05, 'samples': 21993472, 'steps': 42955, 'loss/train': 1.8669122457504272} 02/26/2022 09:11:41 - INFO - codeparrot_training - Step 42956: {'lr': 2.6101168918130026e-05, 'samples': 21993984, 'steps': 42956, 'loss/train': 2.373201608657837} 02/26/2022 09:11:45 - INFO - codeparrot_training - Step 42957: {'lr': 2.609389023488898e-05, 'samples': 21994496, 'steps': 42957, 'loss/train': 1.4350141286849976} 02/26/2022 09:11:50 - INFO - codeparrot_training - Step 42958: {'lr': 2.6086612510790698e-05, 'samples': 21995008, 'steps': 42958, 'loss/train': 1.8349874019622803} 02/26/2022 09:11:54 - INFO - codeparrot_training - Step 42959: {'lr': 2.607933574586624e-05, 'samples': 21995520, 'steps': 42959, 'loss/train': 1.8287482261657715} 02/26/2022 09:11:59 - INFO - codeparrot_training - Step 42960: {'lr': 2.6072059940146776e-05, 'samples': 21996032, 'steps': 42960, 'loss/train': 2.688157320022583} 02/26/2022 09:12:03 - INFO - codeparrot_training - Step 42961: {'lr': 2.6064785093663528e-05, 'samples': 21996544, 'steps': 42961, 'loss/train': 1.2066864967346191} 02/26/2022 09:12:08 - INFO - codeparrot_training - Step 42962: {'lr': 2.6057511206447644e-05, 'samples': 21997056, 'steps': 42962, 'loss/train': 1.4768345355987549} 02/26/2022 09:12:14 - INFO - codeparrot_training - Step 42963: {'lr': 2.6050238278530285e-05, 'samples': 21997568, 'steps': 42963, 'loss/train': 0.8944421410560608} 02/26/2022 09:12:17 - INFO - codeparrot_training - Step 42964: {'lr': 2.6042966309942544e-05, 'samples': 21998080, 'steps': 42964, 'loss/train': 2.481175661087036} 02/26/2022 09:12:25 - INFO - codeparrot_training - Step 42965: {'lr': 2.6035695300715702e-05, 'samples': 21998592, 'steps': 42965, 'loss/train': 2.2580296993255615} 02/26/2022 09:12:28 - INFO - codeparrot_training - Step 42966: {'lr': 2.602842525088073e-05, 'samples': 21999104, 'steps': 42966, 'loss/train': 1.8494553565979004} 02/26/2022 09:12:34 - INFO - codeparrot_training - Step 42967: {'lr': 2.6021156160468935e-05, 'samples': 21999616, 'steps': 42967, 'loss/train': 0.03226064145565033} 02/26/2022 09:12:37 - INFO - codeparrot_training - Step 42968: {'lr': 2.6013888029511294e-05, 'samples': 22000128, 'steps': 42968, 'loss/train': 2.2895877361297607} 02/26/2022 09:12:43 - INFO - codeparrot_training - Step 42969: {'lr': 2.6006620858039176e-05, 'samples': 22000640, 'steps': 42969, 'loss/train': 1.4299345016479492} 02/26/2022 09:12:46 - INFO - codeparrot_training - Step 42970: {'lr': 2.599935464608344e-05, 'samples': 22001152, 'steps': 42970, 'loss/train': 1.790269374847412} 02/26/2022 09:12:50 - INFO - codeparrot_training - Step 42971: {'lr': 2.5992089393675396e-05, 'samples': 22001664, 'steps': 42971, 'loss/train': 1.8767997026443481} 02/26/2022 09:12:55 - INFO - codeparrot_training - Step 42972: {'lr': 2.598482510084607e-05, 'samples': 22002176, 'steps': 42972, 'loss/train': 1.2279059886932373} 02/26/2022 09:12:59 - INFO - codeparrot_training - Step 42973: {'lr': 2.5977561767626668e-05, 'samples': 22002688, 'steps': 42973, 'loss/train': 1.668615698814392} 02/26/2022 09:13:04 - INFO - codeparrot_training - Step 42974: {'lr': 2.597029939404827e-05, 'samples': 22003200, 'steps': 42974, 'loss/train': 2.2787258625030518} 02/26/2022 09:13:08 - INFO - codeparrot_training - Step 42975: {'lr': 2.5963037980141935e-05, 'samples': 22003712, 'steps': 42975, 'loss/train': 1.1480618715286255} 02/26/2022 09:13:15 - INFO - codeparrot_training - Step 42976: {'lr': 2.595577752593878e-05, 'samples': 22004224, 'steps': 42976, 'loss/train': 1.9394205808639526} 02/26/2022 09:13:18 - INFO - codeparrot_training - Step 42977: {'lr': 2.5948518031469998e-05, 'samples': 22004736, 'steps': 42977, 'loss/train': 1.4486584663391113} 02/26/2022 09:13:24 - INFO - codeparrot_training - Step 42978: {'lr': 2.5941259496766595e-05, 'samples': 22005248, 'steps': 42978, 'loss/train': 2.7966549396514893} 02/26/2022 09:13:28 - INFO - codeparrot_training - Step 42979: {'lr': 2.5934001921859684e-05, 'samples': 22005760, 'steps': 42979, 'loss/train': 1.8758902549743652} 02/26/2022 09:13:33 - INFO - codeparrot_training - Step 42980: {'lr': 2.5926745306780324e-05, 'samples': 22006272, 'steps': 42980, 'loss/train': 1.7142419815063477} 02/26/2022 09:13:37 - INFO - codeparrot_training - Step 42981: {'lr': 2.5919489651559686e-05, 'samples': 22006784, 'steps': 42981, 'loss/train': 1.8362321853637695} 02/26/2022 09:13:42 - INFO - codeparrot_training - Step 42982: {'lr': 2.5912234956228798e-05, 'samples': 22007296, 'steps': 42982, 'loss/train': 1.8216270208358765} 02/26/2022 09:13:46 - INFO - codeparrot_training - Step 42983: {'lr': 2.5904981220818747e-05, 'samples': 22007808, 'steps': 42983, 'loss/train': 1.2957565784454346} 02/26/2022 09:13:53 - INFO - codeparrot_training - Step 42984: {'lr': 2.5897728445360536e-05, 'samples': 22008320, 'steps': 42984, 'loss/train': 1.6118088960647583} 02/26/2022 09:13:56 - INFO - codeparrot_training - Step 42985: {'lr': 2.5890476629885334e-05, 'samples': 22008832, 'steps': 42985, 'loss/train': 2.1677441596984863} 02/26/2022 09:14:02 - INFO - codeparrot_training - Step 42986: {'lr': 2.5883225774424146e-05, 'samples': 22009344, 'steps': 42986, 'loss/train': 0.9376107454299927} 02/26/2022 09:14:05 - INFO - codeparrot_training - Step 42987: {'lr': 2.587597587900814e-05, 'samples': 22009856, 'steps': 42987, 'loss/train': 2.973043441772461} 02/26/2022 09:14:11 - INFO - codeparrot_training - Step 42988: {'lr': 2.586872694366818e-05, 'samples': 22010368, 'steps': 42988, 'loss/train': 1.6875289678573608} 02/26/2022 09:14:16 - INFO - codeparrot_training - Step 42989: {'lr': 2.5861478968435492e-05, 'samples': 22010880, 'steps': 42989, 'loss/train': 1.72090482711792} 02/26/2022 09:14:20 - INFO - codeparrot_training - Step 42990: {'lr': 2.5854231953340995e-05, 'samples': 22011392, 'steps': 42990, 'loss/train': 3.1360926628112793} 02/26/2022 09:14:26 - INFO - codeparrot_training - Step 42991: {'lr': 2.5846985898415888e-05, 'samples': 22011904, 'steps': 42991, 'loss/train': 2.655022382736206} 02/26/2022 09:14:29 - INFO - codeparrot_training - Step 42992: {'lr': 2.5839740803691032e-05, 'samples': 22012416, 'steps': 42992, 'loss/train': 0.8258548378944397} 02/26/2022 09:14:33 - INFO - codeparrot_training - Step 42993: {'lr': 2.5832496669197597e-05, 'samples': 22012928, 'steps': 42993, 'loss/train': 3.383549451828003} 02/26/2022 09:14:38 - INFO - codeparrot_training - Step 42994: {'lr': 2.5825253494966504e-05, 'samples': 22013440, 'steps': 42994, 'loss/train': 1.366349220275879} 02/26/2022 09:14:42 - INFO - codeparrot_training - Step 42995: {'lr': 2.5818011281028953e-05, 'samples': 22013952, 'steps': 42995, 'loss/train': 2.713393211364746} 02/26/2022 09:14:49 - INFO - codeparrot_training - Step 42996: {'lr': 2.5810770027415748e-05, 'samples': 22014464, 'steps': 42996, 'loss/train': 0.46251311898231506} 02/26/2022 09:14:52 - INFO - codeparrot_training - Step 42997: {'lr': 2.580352973415806e-05, 'samples': 22014976, 'steps': 42997, 'loss/train': 1.973315715789795} 02/26/2022 09:14:58 - INFO - codeparrot_training - Step 42998: {'lr': 2.5796290401286865e-05, 'samples': 22015488, 'steps': 42998, 'loss/train': 1.8694614171981812} 02/26/2022 09:15:01 - INFO - codeparrot_training - Step 42999: {'lr': 2.578905202883311e-05, 'samples': 22016000, 'steps': 42999, 'loss/train': 1.5767271518707275} 02/26/2022 09:15:01 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 09:15:19 - WARNING - huggingface_hub.repository - Several commits (43) will be pushed upstream. 02/26/2022 09:15:19 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 09:15:53 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 2c9f21a..d24e10b floral-grass-11 -> floral-grass-11 02/26/2022 09:16:00 - INFO - codeparrot_training - Step 43000: {'lr': 2.5781814616827938e-05, 'samples': 22016512, 'steps': 43000, 'loss/train': 1.3371886014938354} 02/26/2022 09:16:03 - INFO - codeparrot_training - Step 43001: {'lr': 2.5774578165302236e-05, 'samples': 22017024, 'steps': 43001, 'loss/train': 1.2176768779754639} 02/26/2022 09:16:09 - INFO - codeparrot_training - Step 43002: {'lr': 2.576734267428707e-05, 'samples': 22017536, 'steps': 43002, 'loss/train': 2.634197950363159} 02/26/2022 09:16:12 - INFO - codeparrot_training - Step 43003: {'lr': 2.5760108143813353e-05, 'samples': 22018048, 'steps': 43003, 'loss/train': 1.4724316596984863} 02/26/2022 09:16:18 - INFO - codeparrot_training - Step 43004: {'lr': 2.575287457391218e-05, 'samples': 22018560, 'steps': 43004, 'loss/train': 1.85142183303833} 02/26/2022 09:16:21 - INFO - codeparrot_training - Step 43005: {'lr': 2.5745641964614485e-05, 'samples': 22019072, 'steps': 43005, 'loss/train': 2.0375685691833496} 02/26/2022 09:16:27 - INFO - codeparrot_training - Step 43006: {'lr': 2.5738410315951256e-05, 'samples': 22019584, 'steps': 43006, 'loss/train': 0.5772395133972168} 02/26/2022 09:16:32 - INFO - codeparrot_training - Step 43007: {'lr': 2.5731179627953406e-05, 'samples': 22020096, 'steps': 43007, 'loss/train': 1.8390233516693115} 02/26/2022 09:16:36 - INFO - codeparrot_training - Step 43008: {'lr': 2.572394990065202e-05, 'samples': 22020608, 'steps': 43008, 'loss/train': 3.028017282485962} 02/26/2022 09:16:41 - INFO - codeparrot_training - Step 43009: {'lr': 2.5716721134078e-05, 'samples': 22021120, 'steps': 43009, 'loss/train': 1.6743240356445312} 02/26/2022 09:16:45 - INFO - codeparrot_training - Step 43010: {'lr': 2.5709493328262363e-05, 'samples': 22021632, 'steps': 43010, 'loss/train': 0.19287554919719696} 02/26/2022 09:16:52 - INFO - codeparrot_training - Step 43011: {'lr': 2.5702266483235952e-05, 'samples': 22022144, 'steps': 43011, 'loss/train': 1.4503531455993652} 02/26/2022 09:16:55 - INFO - codeparrot_training - Step 43012: {'lr': 2.5695040599029883e-05, 'samples': 22022656, 'steps': 43012, 'loss/train': 2.0527825355529785} 02/26/2022 09:17:01 - INFO - codeparrot_training - Step 43013: {'lr': 2.568781567567502e-05, 'samples': 22023168, 'steps': 43013, 'loss/train': 1.9433341026306152} 02/26/2022 09:17:04 - INFO - codeparrot_training - Step 43014: {'lr': 2.5680591713202333e-05, 'samples': 22023680, 'steps': 43014, 'loss/train': 0.9118227362632751} 02/26/2022 09:17:10 - INFO - codeparrot_training - Step 43015: {'lr': 2.567336871164272e-05, 'samples': 22024192, 'steps': 43015, 'loss/train': 1.5988863706588745} 02/26/2022 09:17:13 - INFO - codeparrot_training - Step 43016: {'lr': 2.5666146671027206e-05, 'samples': 22024704, 'steps': 43016, 'loss/train': 1.2842426300048828} 02/26/2022 09:17:19 - INFO - codeparrot_training - Step 43017: {'lr': 2.565892559138669e-05, 'samples': 22025216, 'steps': 43017, 'loss/train': 1.4712259769439697} 02/26/2022 09:17:22 - INFO - codeparrot_training - Step 43018: {'lr': 2.565170547275211e-05, 'samples': 22025728, 'steps': 43018, 'loss/train': 0.6036337018013} 02/26/2022 09:17:28 - INFO - codeparrot_training - Step 43019: {'lr': 2.5644486315154315e-05, 'samples': 22026240, 'steps': 43019, 'loss/train': 2.047445774078369} 02/26/2022 09:17:31 - INFO - codeparrot_training - Step 43020: {'lr': 2.563726811862435e-05, 'samples': 22026752, 'steps': 43020, 'loss/train': 2.0443203449249268} 02/26/2022 09:17:35 - INFO - codeparrot_training - Step 43021: {'lr': 2.5630050883193062e-05, 'samples': 22027264, 'steps': 43021, 'loss/train': 1.2457826137542725} 02/26/2022 09:17:42 - INFO - codeparrot_training - Step 43022: {'lr': 2.5622834608891475e-05, 'samples': 22027776, 'steps': 43022, 'loss/train': 0.8604428172111511} 02/26/2022 09:17:48 - INFO - codeparrot_training - Step 43023: {'lr': 2.5615619295750293e-05, 'samples': 22028288, 'steps': 43023, 'loss/train': 1.1581933498382568} 02/26/2022 09:17:51 - INFO - codeparrot_training - Step 43024: {'lr': 2.560840494380062e-05, 'samples': 22028800, 'steps': 43024, 'loss/train': 0.9329136610031128} 02/26/2022 09:17:57 - INFO - codeparrot_training - Step 43025: {'lr': 2.560119155307325e-05, 'samples': 22029312, 'steps': 43025, 'loss/train': 2.4660651683807373} 02/26/2022 09:18:00 - INFO - codeparrot_training - Step 43026: {'lr': 2.5593979123599205e-05, 'samples': 22029824, 'steps': 43026, 'loss/train': 0.03613194078207016} 02/26/2022 09:18:06 - INFO - codeparrot_training - Step 43027: {'lr': 2.558676765540921e-05, 'samples': 22030336, 'steps': 43027, 'loss/train': 2.0503616333007812} 02/26/2022 09:18:09 - INFO - codeparrot_training - Step 43028: {'lr': 2.5579557148534272e-05, 'samples': 22030848, 'steps': 43028, 'loss/train': 1.5886479616165161} 02/26/2022 09:18:15 - INFO - codeparrot_training - Step 43029: {'lr': 2.5572347603005196e-05, 'samples': 22031360, 'steps': 43029, 'loss/train': 1.7415753602981567} 02/26/2022 09:18:18 - INFO - codeparrot_training - Step 43030: {'lr': 2.5565139018853046e-05, 'samples': 22031872, 'steps': 43030, 'loss/train': 2.2385222911834717} 02/26/2022 09:18:25 - INFO - codeparrot_training - Step 43031: {'lr': 2.555793139610846e-05, 'samples': 22032384, 'steps': 43031, 'loss/train': 1.4883617162704468} 02/26/2022 09:18:29 - INFO - codeparrot_training - Step 43032: {'lr': 2.5550724734802466e-05, 'samples': 22032896, 'steps': 43032, 'loss/train': 1.266660213470459} 02/26/2022 09:18:34 - INFO - codeparrot_training - Step 43033: {'lr': 2.5543519034965878e-05, 'samples': 22033408, 'steps': 43033, 'loss/train': 1.63584566116333} 02/26/2022 09:18:38 - INFO - codeparrot_training - Step 43034: {'lr': 2.5536314296629613e-05, 'samples': 22033920, 'steps': 43034, 'loss/train': 1.8765552043914795} 02/26/2022 09:18:43 - INFO - codeparrot_training - Step 43035: {'lr': 2.5529110519824482e-05, 'samples': 22034432, 'steps': 43035, 'loss/train': 1.5633749961853027} 02/26/2022 09:18:47 - INFO - codeparrot_training - Step 43036: {'lr': 2.5521907704581405e-05, 'samples': 22034944, 'steps': 43036, 'loss/train': 2.941575765609741} 02/26/2022 09:18:52 - INFO - codeparrot_training - Step 43037: {'lr': 2.551470585093113e-05, 'samples': 22035456, 'steps': 43037, 'loss/train': 2.3214941024780273} 02/26/2022 09:18:56 - INFO - codeparrot_training - Step 43038: {'lr': 2.550750495890461e-05, 'samples': 22035968, 'steps': 43038, 'loss/train': 1.8605632781982422} 02/26/2022 09:19:01 - INFO - codeparrot_training - Step 43039: {'lr': 2.550030502853265e-05, 'samples': 22036480, 'steps': 43039, 'loss/train': 2.09035325050354} 02/26/2022 09:19:05 - INFO - codeparrot_training - Step 43040: {'lr': 2.5493106059846115e-05, 'samples': 22036992, 'steps': 43040, 'loss/train': 2.655060052871704} 02/26/2022 09:19:08 - INFO - codeparrot_training - Step 43041: {'lr': 2.5485908052875762e-05, 'samples': 22037504, 'steps': 43041, 'loss/train': 1.8697302341461182} 02/26/2022 09:19:16 - INFO - codeparrot_training - Step 43042: {'lr': 2.547871100765256e-05, 'samples': 22038016, 'steps': 43042, 'loss/train': 2.2321293354034424} 02/26/2022 09:19:19 - INFO - codeparrot_training - Step 43043: {'lr': 2.547151492420724e-05, 'samples': 22038528, 'steps': 43043, 'loss/train': 1.0352306365966797} 02/26/2022 09:19:25 - INFO - codeparrot_training - Step 43044: {'lr': 2.546431980257066e-05, 'samples': 22039040, 'steps': 43044, 'loss/train': 2.053722620010376} 02/26/2022 09:19:28 - INFO - codeparrot_training - Step 43045: {'lr': 2.5457125642773638e-05, 'samples': 22039552, 'steps': 43045, 'loss/train': 1.6874339580535889} 02/26/2022 09:19:34 - INFO - codeparrot_training - Step 43046: {'lr': 2.5449932444846944e-05, 'samples': 22040064, 'steps': 43046, 'loss/train': 1.937677025794983} 02/26/2022 09:19:37 - INFO - codeparrot_training - Step 43047: {'lr': 2.544274020882148e-05, 'samples': 22040576, 'steps': 43047, 'loss/train': 2.045397996902466} 02/26/2022 09:19:43 - INFO - codeparrot_training - Step 43048: {'lr': 2.543554893472802e-05, 'samples': 22041088, 'steps': 43048, 'loss/train': 1.6934360265731812} 02/26/2022 09:19:48 - INFO - codeparrot_training - Step 43049: {'lr': 2.5428358622597375e-05, 'samples': 22041600, 'steps': 43049, 'loss/train': 2.004577875137329} 02/26/2022 09:19:52 - INFO - codeparrot_training - Step 43050: {'lr': 2.542116927246024e-05, 'samples': 22042112, 'steps': 43050, 'loss/train': 0.7194834351539612} 02/26/2022 09:19:57 - INFO - codeparrot_training - Step 43051: {'lr': 2.5413980884347598e-05, 'samples': 22042624, 'steps': 43051, 'loss/train': 1.7380951642990112} 02/26/2022 09:20:01 - INFO - codeparrot_training - Step 43052: {'lr': 2.5406793458290113e-05, 'samples': 22043136, 'steps': 43052, 'loss/train': 2.375088691711426} 02/26/2022 09:20:04 - INFO - codeparrot_training - Step 43053: {'lr': 2.539960699431862e-05, 'samples': 22043648, 'steps': 43053, 'loss/train': 1.9548765420913696} 02/26/2022 09:20:10 - INFO - codeparrot_training - Step 43054: {'lr': 2.5392421492463846e-05, 'samples': 22044160, 'steps': 43054, 'loss/train': 1.586997389793396} 02/26/2022 09:20:13 - INFO - codeparrot_training - Step 43055: {'lr': 2.5385236952756687e-05, 'samples': 22044672, 'steps': 43055, 'loss/train': 1.9782105684280396} 02/26/2022 09:20:19 - INFO - codeparrot_training - Step 43056: {'lr': 2.537805337522783e-05, 'samples': 22045184, 'steps': 43056, 'loss/train': 1.9436240196228027} 02/26/2022 09:20:23 - INFO - codeparrot_training - Step 43057: {'lr': 2.5370870759908073e-05, 'samples': 22045696, 'steps': 43057, 'loss/train': 0.04014909639954567} 02/26/2022 09:20:30 - INFO - codeparrot_training - Step 43058: {'lr': 2.5363689106828096e-05, 'samples': 22046208, 'steps': 43058, 'loss/train': 1.643044114112854} 02/26/2022 09:20:33 - INFO - codeparrot_training - Step 43059: {'lr': 2.535650841601883e-05, 'samples': 22046720, 'steps': 43059, 'loss/train': 1.259970784187317} 02/26/2022 09:20:39 - INFO - codeparrot_training - Step 43060: {'lr': 2.5349328687510888e-05, 'samples': 22047232, 'steps': 43060, 'loss/train': 1.5826419591903687} 02/26/2022 09:20:42 - INFO - codeparrot_training - Step 43061: {'lr': 2.5342149921335182e-05, 'samples': 22047744, 'steps': 43061, 'loss/train': 1.0303560495376587} 02/26/2022 09:20:48 - INFO - codeparrot_training - Step 43062: {'lr': 2.533497211752228e-05, 'samples': 22048256, 'steps': 43062, 'loss/train': 2.214817762374878} 02/26/2022 09:20:53 - INFO - codeparrot_training - Step 43063: {'lr': 2.5327795276103072e-05, 'samples': 22048768, 'steps': 43063, 'loss/train': 1.8114770650863647} 02/26/2022 09:20:57 - INFO - codeparrot_training - Step 43064: {'lr': 2.5320619397108197e-05, 'samples': 22049280, 'steps': 43064, 'loss/train': 1.6828633546829224} 02/26/2022 09:21:03 - INFO - codeparrot_training - Step 43065: {'lr': 2.531344448056852e-05, 'samples': 22049792, 'steps': 43065, 'loss/train': 1.882368803024292} 02/26/2022 09:21:06 - INFO - codeparrot_training - Step 43066: {'lr': 2.530627052651463e-05, 'samples': 22050304, 'steps': 43066, 'loss/train': 1.5169315338134766} 02/26/2022 09:21:13 - INFO - codeparrot_training - Step 43067: {'lr': 2.529909753497736e-05, 'samples': 22050816, 'steps': 43067, 'loss/train': 0.9377289414405823} 02/26/2022 09:21:17 - INFO - codeparrot_training - Step 43068: {'lr': 2.5291925505987358e-05, 'samples': 22051328, 'steps': 43068, 'loss/train': 2.026712417602539} 02/26/2022 09:21:20 - INFO - codeparrot_training - Step 43069: {'lr': 2.5284754439575452e-05, 'samples': 22051840, 'steps': 43069, 'loss/train': 2.1036181449890137} 02/26/2022 09:21:26 - INFO - codeparrot_training - Step 43070: {'lr': 2.527758433577232e-05, 'samples': 22052352, 'steps': 43070, 'loss/train': 1.6851140260696411} 02/26/2022 09:21:31 - INFO - codeparrot_training - Step 43071: {'lr': 2.5270415194608627e-05, 'samples': 22052864, 'steps': 43071, 'loss/train': 1.0842164754867554} 02/26/2022 09:21:35 - INFO - codeparrot_training - Step 43072: {'lr': 2.526324701611507e-05, 'samples': 22053376, 'steps': 43072, 'loss/train': 1.1889334917068481} 02/26/2022 09:21:40 - INFO - codeparrot_training - Step 43073: {'lr': 2.525607980032246e-05, 'samples': 22053888, 'steps': 43073, 'loss/train': 1.4469133615493774} 02/26/2022 09:21:44 - INFO - codeparrot_training - Step 43074: {'lr': 2.5248913547261442e-05, 'samples': 22054400, 'steps': 43074, 'loss/train': 1.580714464187622} 02/26/2022 09:21:49 - INFO - codeparrot_training - Step 43075: {'lr': 2.524174825696271e-05, 'samples': 22054912, 'steps': 43075, 'loss/train': 0.7569447755813599} 02/26/2022 09:21:53 - INFO - codeparrot_training - Step 43076: {'lr': 2.5234583929456904e-05, 'samples': 22055424, 'steps': 43076, 'loss/train': 1.4094403982162476} 02/26/2022 09:22:00 - INFO - codeparrot_training - Step 43077: {'lr': 2.522742056477484e-05, 'samples': 22055936, 'steps': 43077, 'loss/train': 1.3814173936843872} 02/26/2022 09:22:03 - INFO - codeparrot_training - Step 43078: {'lr': 2.5220258162947097e-05, 'samples': 22056448, 'steps': 43078, 'loss/train': 1.5424473285675049} 02/26/2022 09:22:09 - INFO - codeparrot_training - Step 43079: {'lr': 2.521309672400443e-05, 'samples': 22056960, 'steps': 43079, 'loss/train': 1.7623589038848877} 02/26/2022 09:22:12 - INFO - codeparrot_training - Step 43080: {'lr': 2.52059362479774e-05, 'samples': 22057472, 'steps': 43080, 'loss/train': 1.5497207641601562} 02/26/2022 09:22:18 - INFO - codeparrot_training - Step 43081: {'lr': 2.5198776734896818e-05, 'samples': 22057984, 'steps': 43081, 'loss/train': 1.9506393671035767} 02/26/2022 09:22:21 - INFO - codeparrot_training - Step 43082: {'lr': 2.5191618184793292e-05, 'samples': 22058496, 'steps': 43082, 'loss/train': 1.6275516748428345} 02/26/2022 09:22:27 - INFO - codeparrot_training - Step 43083: {'lr': 2.5184460597697496e-05, 'samples': 22059008, 'steps': 43083, 'loss/train': 2.3514351844787598} 02/26/2022 09:22:30 - INFO - codeparrot_training - Step 43084: {'lr': 2.517730397364004e-05, 'samples': 22059520, 'steps': 43084, 'loss/train': 1.8151108026504517} 02/26/2022 09:22:36 - INFO - codeparrot_training - Step 43085: {'lr': 2.5170148312651658e-05, 'samples': 22060032, 'steps': 43085, 'loss/train': 2.272878885269165} 02/26/2022 09:22:39 - INFO - codeparrot_training - Step 43086: {'lr': 2.5162993614762985e-05, 'samples': 22060544, 'steps': 43086, 'loss/train': 0.44775155186653137} 02/26/2022 09:22:47 - INFO - codeparrot_training - Step 43087: {'lr': 2.5155839880004637e-05, 'samples': 22061056, 'steps': 43087, 'loss/train': 0.6676808595657349} 02/26/2022 09:22:50 - INFO - codeparrot_training - Step 43088: {'lr': 2.514868710840723e-05, 'samples': 22061568, 'steps': 43088, 'loss/train': 2.3862431049346924} 02/26/2022 09:22:56 - INFO - codeparrot_training - Step 43089: {'lr': 2.5141535300001518e-05, 'samples': 22062080, 'steps': 43089, 'loss/train': 1.9383883476257324} 02/26/2022 09:22:59 - INFO - codeparrot_training - Step 43090: {'lr': 2.5134384454818028e-05, 'samples': 22062592, 'steps': 43090, 'loss/train': 1.6923736333847046} 02/26/2022 09:23:05 - INFO - codeparrot_training - Step 43091: {'lr': 2.512723457288746e-05, 'samples': 22063104, 'steps': 43091, 'loss/train': 3.0778558254241943} 02/26/2022 09:23:08 - INFO - codeparrot_training - Step 43092: {'lr': 2.5120085654240428e-05, 'samples': 22063616, 'steps': 43092, 'loss/train': 2.2247273921966553} 02/26/2022 09:23:14 - INFO - codeparrot_training - Step 43093: {'lr': 2.5112937698907464e-05, 'samples': 22064128, 'steps': 43093, 'loss/train': 1.5931341648101807} 02/26/2022 09:23:17 - INFO - codeparrot_training - Step 43094: {'lr': 2.5105790706919317e-05, 'samples': 22064640, 'steps': 43094, 'loss/train': 1.8834272623062134} 02/26/2022 09:23:23 - INFO - codeparrot_training - Step 43095: {'lr': 2.5098644678306526e-05, 'samples': 22065152, 'steps': 43095, 'loss/train': 1.5963129997253418} 02/26/2022 09:23:26 - INFO - codeparrot_training - Step 43096: {'lr': 2.509149961309981e-05, 'samples': 22065664, 'steps': 43096, 'loss/train': 2.1256372928619385} 02/26/2022 09:23:33 - INFO - codeparrot_training - Step 43097: {'lr': 2.5084355511329592e-05, 'samples': 22066176, 'steps': 43097, 'loss/train': 1.1419918537139893} 02/26/2022 09:23:37 - INFO - codeparrot_training - Step 43098: {'lr': 2.5077212373026626e-05, 'samples': 22066688, 'steps': 43098, 'loss/train': 1.5744365453720093} 02/26/2022 09:23:42 - INFO - codeparrot_training - Step 43099: {'lr': 2.507007019822141e-05, 'samples': 22067200, 'steps': 43099, 'loss/train': 1.2878320217132568} 02/26/2022 09:23:46 - INFO - codeparrot_training - Step 43100: {'lr': 2.5062928986944677e-05, 'samples': 22067712, 'steps': 43100, 'loss/train': 3.3926568031311035} 02/26/2022 09:23:52 - INFO - codeparrot_training - Step 43101: {'lr': 2.505578873922684e-05, 'samples': 22068224, 'steps': 43101, 'loss/train': 1.2551594972610474} 02/26/2022 09:23:55 - INFO - codeparrot_training - Step 43102: {'lr': 2.5048649455098632e-05, 'samples': 22068736, 'steps': 43102, 'loss/train': 1.7071492671966553} 02/26/2022 09:24:01 - INFO - codeparrot_training - Step 43103: {'lr': 2.504151113459055e-05, 'samples': 22069248, 'steps': 43103, 'loss/train': 2.2507243156433105} 02/26/2022 09:24:04 - INFO - codeparrot_training - Step 43104: {'lr': 2.5034373777733267e-05, 'samples': 22069760, 'steps': 43104, 'loss/train': 0.9035642743110657} 02/26/2022 09:24:10 - INFO - codeparrot_training - Step 43105: {'lr': 2.50272373845572e-05, 'samples': 22070272, 'steps': 43105, 'loss/train': 2.2026631832122803} 02/26/2022 09:24:13 - INFO - codeparrot_training - Step 43106: {'lr': 2.5020101955093078e-05, 'samples': 22070784, 'steps': 43106, 'loss/train': 2.4340567588806152} 02/26/2022 09:24:19 - INFO - codeparrot_training - Step 43107: {'lr': 2.5012967489371346e-05, 'samples': 22071296, 'steps': 43107, 'loss/train': 2.2721855640411377} 02/26/2022 09:24:22 - INFO - codeparrot_training - Step 43108: {'lr': 2.500583398742265e-05, 'samples': 22071808, 'steps': 43108, 'loss/train': 1.805285096168518} 02/26/2022 09:24:28 - INFO - codeparrot_training - Step 43109: {'lr': 2.4998701449277545e-05, 'samples': 22072320, 'steps': 43109, 'loss/train': 2.4406607151031494} 02/26/2022 09:24:32 - INFO - codeparrot_training - Step 43110: {'lr': 2.4991569874966537e-05, 'samples': 22072832, 'steps': 43110, 'loss/train': 1.4582270383834839} 02/26/2022 09:24:37 - INFO - codeparrot_training - Step 43111: {'lr': 2.4984439264520154e-05, 'samples': 22073344, 'steps': 43111, 'loss/train': 1.9722706079483032} 02/26/2022 09:24:41 - INFO - codeparrot_training - Step 43112: {'lr': 2.497730961796904e-05, 'samples': 22073856, 'steps': 43112, 'loss/train': 1.6750661134719849} 02/26/2022 09:24:48 - INFO - codeparrot_training - Step 43113: {'lr': 2.4970180935343646e-05, 'samples': 22074368, 'steps': 43113, 'loss/train': 0.39469850063323975} 02/26/2022 09:24:51 - INFO - codeparrot_training - Step 43114: {'lr': 2.4963053216674582e-05, 'samples': 22074880, 'steps': 43114, 'loss/train': 2.681544780731201} 02/26/2022 09:24:57 - INFO - codeparrot_training - Step 43115: {'lr': 2.4955926461992268e-05, 'samples': 22075392, 'steps': 43115, 'loss/train': 1.2985799312591553} 02/26/2022 09:25:00 - INFO - codeparrot_training - Step 43116: {'lr': 2.4948800671327347e-05, 'samples': 22075904, 'steps': 43116, 'loss/train': 2.1947743892669678} 02/26/2022 09:25:06 - INFO - codeparrot_training - Step 43117: {'lr': 2.494167584471027e-05, 'samples': 22076416, 'steps': 43117, 'loss/train': 1.6213518381118774} 02/26/2022 09:25:09 - INFO - codeparrot_training - Step 43118: {'lr': 2.4934551982171704e-05, 'samples': 22076928, 'steps': 43118, 'loss/train': 1.6491273641586304} 02/26/2022 09:25:15 - INFO - codeparrot_training - Step 43119: {'lr': 2.492742908374193e-05, 'samples': 22077440, 'steps': 43119, 'loss/train': 1.3500185012817383} 02/26/2022 09:25:18 - INFO - codeparrot_training - Step 43120: {'lr': 2.492030714945162e-05, 'samples': 22077952, 'steps': 43120, 'loss/train': 1.6277194023132324} 02/26/2022 09:25:24 - INFO - codeparrot_training - Step 43121: {'lr': 2.491318617933122e-05, 'samples': 22078464, 'steps': 43121, 'loss/train': 2.232564687728882} 02/26/2022 09:25:27 - INFO - codeparrot_training - Step 43122: {'lr': 2.4906066173411347e-05, 'samples': 22078976, 'steps': 43122, 'loss/train': 1.4550281763076782} 02/26/2022 09:25:35 - INFO - codeparrot_training - Step 43123: {'lr': 2.4898947131722305e-05, 'samples': 22079488, 'steps': 43123, 'loss/train': 1.9753600358963013} 02/26/2022 09:25:38 - INFO - codeparrot_training - Step 43124: {'lr': 2.489182905429474e-05, 'samples': 22080000, 'steps': 43124, 'loss/train': 0.4699411690235138} 02/26/2022 09:25:44 - INFO - codeparrot_training - Step 43125: {'lr': 2.4884711941159127e-05, 'samples': 22080512, 'steps': 43125, 'loss/train': 1.6645084619522095} 02/26/2022 09:25:47 - INFO - codeparrot_training - Step 43126: {'lr': 2.4877595792345885e-05, 'samples': 22081024, 'steps': 43126, 'loss/train': 1.249341607093811} 02/26/2022 09:25:53 - INFO - codeparrot_training - Step 43127: {'lr': 2.487048060788552e-05, 'samples': 22081536, 'steps': 43127, 'loss/train': 1.398234486579895} 02/26/2022 09:25:56 - INFO - codeparrot_training - Step 43128: {'lr': 2.486336638780856e-05, 'samples': 22082048, 'steps': 43128, 'loss/train': 0.4382885992527008} 02/26/2022 09:26:02 - INFO - codeparrot_training - Step 43129: {'lr': 2.4856253132145484e-05, 'samples': 22082560, 'steps': 43129, 'loss/train': 3.2967722415924072} 02/26/2022 09:26:05 - INFO - codeparrot_training - Step 43130: {'lr': 2.4849140840926655e-05, 'samples': 22083072, 'steps': 43130, 'loss/train': 1.3725587129592896} 02/26/2022 09:26:11 - INFO - codeparrot_training - Step 43131: {'lr': 2.4842029514182656e-05, 'samples': 22083584, 'steps': 43131, 'loss/train': 1.3670753240585327} 02/26/2022 09:26:14 - INFO - codeparrot_training - Step 43132: {'lr': 2.4834919151943915e-05, 'samples': 22084096, 'steps': 43132, 'loss/train': 2.967434883117676} 02/26/2022 09:26:22 - INFO - codeparrot_training - Step 43133: {'lr': 2.4827809754240872e-05, 'samples': 22084608, 'steps': 43133, 'loss/train': 0.5620272159576416} 02/26/2022 09:26:25 - INFO - codeparrot_training - Step 43134: {'lr': 2.482070132110395e-05, 'samples': 22085120, 'steps': 43134, 'loss/train': 2.092459201812744} 02/26/2022 09:26:31 - INFO - codeparrot_training - Step 43135: {'lr': 2.481359385256371e-05, 'samples': 22085632, 'steps': 43135, 'loss/train': 2.5158591270446777} 02/26/2022 09:26:34 - INFO - codeparrot_training - Step 43136: {'lr': 2.4806487348650486e-05, 'samples': 22086144, 'steps': 43136, 'loss/train': 0.18821310997009277} 02/26/2022 09:26:40 - INFO - codeparrot_training - Step 43137: {'lr': 2.479938180939481e-05, 'samples': 22086656, 'steps': 43137, 'loss/train': 2.0100908279418945} 02/26/2022 09:26:43 - INFO - codeparrot_training - Step 43138: {'lr': 2.4792277234826993e-05, 'samples': 22087168, 'steps': 43138, 'loss/train': 2.117788791656494} 02/26/2022 09:26:49 - INFO - codeparrot_training - Step 43139: {'lr': 2.4785173624977647e-05, 'samples': 22087680, 'steps': 43139, 'loss/train': 0.8608589172363281} 02/26/2022 09:26:52 - INFO - codeparrot_training - Step 43140: {'lr': 2.4778070979877028e-05, 'samples': 22088192, 'steps': 43140, 'loss/train': 0.9559106826782227} 02/26/2022 09:26:58 - INFO - codeparrot_training - Step 43141: {'lr': 2.4770969299555664e-05, 'samples': 22088704, 'steps': 43141, 'loss/train': 2.077723741531372} 02/26/2022 09:27:01 - INFO - codeparrot_training - Step 43142: {'lr': 2.476386858404392e-05, 'samples': 22089216, 'steps': 43142, 'loss/train': 1.8173515796661377} 02/26/2022 09:27:07 - INFO - codeparrot_training - Step 43143: {'lr': 2.4756768833372273e-05, 'samples': 22089728, 'steps': 43143, 'loss/train': 2.4149539470672607} 02/26/2022 09:27:10 - INFO - codeparrot_training - Step 43144: {'lr': 2.4749670047571114e-05, 'samples': 22090240, 'steps': 43144, 'loss/train': 1.8064653873443604} 02/26/2022 09:27:17 - INFO - codeparrot_training - Step 43145: {'lr': 2.474257222667084e-05, 'samples': 22090752, 'steps': 43145, 'loss/train': 2.363929271697998} 02/26/2022 09:27:21 - INFO - codeparrot_training - Step 43146: {'lr': 2.473547537070181e-05, 'samples': 22091264, 'steps': 43146, 'loss/train': 1.602826476097107} 02/26/2022 09:27:27 - INFO - codeparrot_training - Step 43147: {'lr': 2.472837947969453e-05, 'samples': 22091776, 'steps': 43147, 'loss/train': 2.1847267150878906} 02/26/2022 09:27:30 - INFO - codeparrot_training - Step 43148: {'lr': 2.4721284553679335e-05, 'samples': 22092288, 'steps': 43148, 'loss/train': 0.6073406934738159} 02/26/2022 09:27:36 - INFO - codeparrot_training - Step 43149: {'lr': 2.471419059268662e-05, 'samples': 22092800, 'steps': 43149, 'loss/train': 1.8850690126419067} 02/26/2022 09:27:40 - INFO - codeparrot_training - Step 43150: {'lr': 2.470709759674672e-05, 'samples': 22093312, 'steps': 43150, 'loss/train': 0.9164625406265259} 02/26/2022 09:27:43 - INFO - codeparrot_training - Step 43151: {'lr': 2.4700005565890136e-05, 'samples': 22093824, 'steps': 43151, 'loss/train': 1.8063238859176636} 02/26/2022 09:27:48 - INFO - codeparrot_training - Step 43152: {'lr': 2.4692914500147186e-05, 'samples': 22094336, 'steps': 43152, 'loss/train': 0.8614669442176819} 02/26/2022 09:27:52 - INFO - codeparrot_training - Step 43153: {'lr': 2.4685824399548252e-05, 'samples': 22094848, 'steps': 43153, 'loss/train': 1.6871877908706665} 02/26/2022 09:27:58 - INFO - codeparrot_training - Step 43154: {'lr': 2.4678735264123647e-05, 'samples': 22095360, 'steps': 43154, 'loss/train': 0.99406898021698} 02/26/2022 09:28:01 - INFO - codeparrot_training - Step 43155: {'lr': 2.467164709390385e-05, 'samples': 22095872, 'steps': 43155, 'loss/train': 1.5107645988464355} 02/26/2022 09:28:07 - INFO - codeparrot_training - Step 43156: {'lr': 2.4664559888919108e-05, 'samples': 22096384, 'steps': 43156, 'loss/train': 1.2792789936065674} 02/26/2022 09:28:10 - INFO - codeparrot_training - Step 43157: {'lr': 2.465747364919993e-05, 'samples': 22096896, 'steps': 43157, 'loss/train': 0.5637739300727844} 02/26/2022 09:28:15 - INFO - codeparrot_training - Step 43158: {'lr': 2.4650388374776484e-05, 'samples': 22097408, 'steps': 43158, 'loss/train': 1.9193655252456665} 02/26/2022 09:28:23 - INFO - codeparrot_training - Step 43159: {'lr': 2.46433040656793e-05, 'samples': 22097920, 'steps': 43159, 'loss/train': 0.19971466064453125} 02/26/2022 09:28:26 - INFO - codeparrot_training - Step 43160: {'lr': 2.4636220721938552e-05, 'samples': 22098432, 'steps': 43160, 'loss/train': 1.2645370960235596} 02/26/2022 09:28:32 - INFO - codeparrot_training - Step 43161: {'lr': 2.4629138343584795e-05, 'samples': 22098944, 'steps': 43161, 'loss/train': 1.433838129043579} 02/26/2022 09:28:35 - INFO - codeparrot_training - Step 43162: {'lr': 2.4622056930648146e-05, 'samples': 22099456, 'steps': 43162, 'loss/train': 1.863355040550232} 02/26/2022 09:28:41 - INFO - codeparrot_training - Step 43163: {'lr': 2.461497648315908e-05, 'samples': 22099968, 'steps': 43163, 'loss/train': 2.170952796936035} 02/26/2022 09:28:44 - INFO - codeparrot_training - Step 43164: {'lr': 2.460789700114782e-05, 'samples': 22100480, 'steps': 43164, 'loss/train': 3.1801652908325195} 02/26/2022 09:28:50 - INFO - codeparrot_training - Step 43165: {'lr': 2.4600818484644906e-05, 'samples': 22100992, 'steps': 43165, 'loss/train': 1.2370730638504028} 02/26/2022 09:28:53 - INFO - codeparrot_training - Step 43166: {'lr': 2.4593740933680387e-05, 'samples': 22101504, 'steps': 43166, 'loss/train': 2.8357667922973633} 02/26/2022 09:28:59 - INFO - codeparrot_training - Step 43167: {'lr': 2.4586664348284743e-05, 'samples': 22102016, 'steps': 43167, 'loss/train': 1.5399960279464722} 02/26/2022 09:29:02 - INFO - codeparrot_training - Step 43168: {'lr': 2.4579588728488228e-05, 'samples': 22102528, 'steps': 43168, 'loss/train': 1.034315586090088} 02/26/2022 09:29:09 - INFO - codeparrot_training - Step 43169: {'lr': 2.4572514074321206e-05, 'samples': 22103040, 'steps': 43169, 'loss/train': 1.7540181875228882} 02/26/2022 09:29:13 - INFO - codeparrot_training - Step 43170: {'lr': 2.456544038581396e-05, 'samples': 22103552, 'steps': 43170, 'loss/train': 0.8240705132484436} 02/26/2022 09:29:18 - INFO - codeparrot_training - Step 43171: {'lr': 2.4558367662996796e-05, 'samples': 22104064, 'steps': 43171, 'loss/train': 1.6378718614578247} 02/26/2022 09:29:22 - INFO - codeparrot_training - Step 43172: {'lr': 2.4551295905899968e-05, 'samples': 22104576, 'steps': 43172, 'loss/train': 2.108032703399658} 02/26/2022 09:29:28 - INFO - codeparrot_training - Step 43173: {'lr': 2.4544225114553786e-05, 'samples': 22105088, 'steps': 43173, 'loss/train': 1.9060156345367432} 02/26/2022 09:29:31 - INFO - codeparrot_training - Step 43174: {'lr': 2.453715528898859e-05, 'samples': 22105600, 'steps': 43174, 'loss/train': 0.03589240834116936} 02/26/2022 09:29:37 - INFO - codeparrot_training - Step 43175: {'lr': 2.45300864292346e-05, 'samples': 22106112, 'steps': 43175, 'loss/train': 1.489375352859497} 02/26/2022 09:29:40 - INFO - codeparrot_training - Step 43176: {'lr': 2.4523018535322154e-05, 'samples': 22106624, 'steps': 43176, 'loss/train': 2.1516973972320557} 02/26/2022 09:29:46 - INFO - codeparrot_training - Step 43177: {'lr': 2.4515951607281457e-05, 'samples': 22107136, 'steps': 43177, 'loss/train': 1.1386784315109253} 02/26/2022 09:29:49 - INFO - codeparrot_training - Step 43178: {'lr': 2.4508885645142865e-05, 'samples': 22107648, 'steps': 43178, 'loss/train': 1.605818748474121} 02/26/2022 09:29:55 - INFO - codeparrot_training - Step 43179: {'lr': 2.4501820648936608e-05, 'samples': 22108160, 'steps': 43179, 'loss/train': 1.4051034450531006} 02/26/2022 09:29:58 - INFO - codeparrot_training - Step 43180: {'lr': 2.4494756618692937e-05, 'samples': 22108672, 'steps': 43180, 'loss/train': 1.4265937805175781} 02/26/2022 09:30:06 - INFO - codeparrot_training - Step 43181: {'lr': 2.4487693554442082e-05, 'samples': 22109184, 'steps': 43181, 'loss/train': 1.743186116218567} 02/26/2022 09:30:09 - INFO - codeparrot_training - Step 43182: {'lr': 2.4480631456214374e-05, 'samples': 22109696, 'steps': 43182, 'loss/train': 1.8575931787490845} 02/26/2022 09:30:15 - INFO - codeparrot_training - Step 43183: {'lr': 2.4473570324040047e-05, 'samples': 22110208, 'steps': 43183, 'loss/train': 2.446790933609009} 02/26/2022 09:30:18 - INFO - codeparrot_training - Step 43184: {'lr': 2.4466510157949318e-05, 'samples': 22110720, 'steps': 43184, 'loss/train': 2.6403350830078125} 02/26/2022 09:30:24 - INFO - codeparrot_training - Step 43185: {'lr': 2.445945095797242e-05, 'samples': 22111232, 'steps': 43185, 'loss/train': 1.784572958946228} 02/26/2022 09:30:27 - INFO - codeparrot_training - Step 43186: {'lr': 2.4452392724139656e-05, 'samples': 22111744, 'steps': 43186, 'loss/train': 1.5917832851409912} 02/26/2022 09:30:33 - INFO - codeparrot_training - Step 43187: {'lr': 2.44453354564812e-05, 'samples': 22112256, 'steps': 43187, 'loss/train': 1.6441023349761963} 02/26/2022 09:30:36 - INFO - codeparrot_training - Step 43188: {'lr': 2.4438279155027333e-05, 'samples': 22112768, 'steps': 43188, 'loss/train': 1.41212797164917} 02/26/2022 09:30:42 - INFO - codeparrot_training - Step 43189: {'lr': 2.4431223819808197e-05, 'samples': 22113280, 'steps': 43189, 'loss/train': 0.12104640156030655} 02/26/2022 09:30:45 - INFO - codeparrot_training - Step 43190: {'lr': 2.44241694508541e-05, 'samples': 22113792, 'steps': 43190, 'loss/train': 1.7831441164016724} 02/26/2022 09:30:51 - INFO - codeparrot_training - Step 43191: {'lr': 2.4417116048195213e-05, 'samples': 22114304, 'steps': 43191, 'loss/train': 1.0775939226150513} 02/26/2022 09:30:54 - INFO - codeparrot_training - Step 43192: {'lr': 2.441006361186185e-05, 'samples': 22114816, 'steps': 43192, 'loss/train': 0.931842565536499} 02/26/2022 09:31:00 - INFO - codeparrot_training - Step 43193: {'lr': 2.4403012141884063e-05, 'samples': 22115328, 'steps': 43193, 'loss/train': 1.607734203338623} 02/26/2022 09:31:03 - INFO - codeparrot_training - Step 43194: {'lr': 2.4395961638292162e-05, 'samples': 22115840, 'steps': 43194, 'loss/train': 1.892438292503357} 02/26/2022 09:31:11 - INFO - codeparrot_training - Step 43195: {'lr': 2.4388912101116295e-05, 'samples': 22116352, 'steps': 43195, 'loss/train': 2.179072618484497} 02/26/2022 09:31:14 - INFO - codeparrot_training - Step 43196: {'lr': 2.4381863530386766e-05, 'samples': 22116864, 'steps': 43196, 'loss/train': 1.7569949626922607} 02/26/2022 09:31:20 - INFO - codeparrot_training - Step 43197: {'lr': 2.437481592613361e-05, 'samples': 22117376, 'steps': 43197, 'loss/train': 1.214928150177002} 02/26/2022 09:31:23 - INFO - codeparrot_training - Step 43198: {'lr': 2.436776928838716e-05, 'samples': 22117888, 'steps': 43198, 'loss/train': 2.00553297996521} 02/26/2022 09:31:29 - INFO - codeparrot_training - Step 43199: {'lr': 2.4360723617177477e-05, 'samples': 22118400, 'steps': 43199, 'loss/train': 0.6758792400360107} 02/26/2022 09:31:32 - INFO - codeparrot_training - Step 43200: {'lr': 2.43536789125349e-05, 'samples': 22118912, 'steps': 43200, 'loss/train': 2.642744541168213} 02/26/2022 09:31:38 - INFO - codeparrot_training - Step 43201: {'lr': 2.43466351744894e-05, 'samples': 22119424, 'steps': 43201, 'loss/train': 1.9860972166061401} 02/26/2022 09:31:41 - INFO - codeparrot_training - Step 43202: {'lr': 2.433959240307135e-05, 'samples': 22119936, 'steps': 43202, 'loss/train': 1.9324525594711304} 02/26/2022 09:31:47 - INFO - codeparrot_training - Step 43203: {'lr': 2.4332550598310743e-05, 'samples': 22120448, 'steps': 43203, 'loss/train': 1.4562253952026367} 02/26/2022 09:31:50 - INFO - codeparrot_training - Step 43204: {'lr': 2.4325509760237895e-05, 'samples': 22120960, 'steps': 43204, 'loss/train': 2.2617857456207275} 02/26/2022 09:31:56 - INFO - codeparrot_training - Step 43205: {'lr': 2.4318469888882893e-05, 'samples': 22121472, 'steps': 43205, 'loss/train': 0.6300603747367859} 02/26/2022 09:31:59 - INFO - codeparrot_training - Step 43206: {'lr': 2.4311430984275935e-05, 'samples': 22121984, 'steps': 43206, 'loss/train': 1.659628987312317} 02/26/2022 09:32:06 - INFO - codeparrot_training - Step 43207: {'lr': 2.4304393046447077e-05, 'samples': 22122496, 'steps': 43207, 'loss/train': 0.9781558513641357} 02/26/2022 09:32:10 - INFO - codeparrot_training - Step 43208: {'lr': 2.4297356075426575e-05, 'samples': 22123008, 'steps': 43208, 'loss/train': 1.295369029045105} 02/26/2022 09:32:15 - INFO - codeparrot_training - Step 43209: {'lr': 2.4290320071244515e-05, 'samples': 22123520, 'steps': 43209, 'loss/train': 1.990005612373352} 02/26/2022 09:32:19 - INFO - codeparrot_training - Step 43210: {'lr': 2.4283285033931097e-05, 'samples': 22124032, 'steps': 43210, 'loss/train': 0.5798360705375671} 02/26/2022 09:32:24 - INFO - codeparrot_training - Step 43211: {'lr': 2.4276250963516322e-05, 'samples': 22124544, 'steps': 43211, 'loss/train': 1.4787448644638062} 02/26/2022 09:32:28 - INFO - codeparrot_training - Step 43212: {'lr': 2.4269217860030473e-05, 'samples': 22125056, 'steps': 43212, 'loss/train': 1.4362761974334717} 02/26/2022 09:32:33 - INFO - codeparrot_training - Step 43213: {'lr': 2.426218572350364e-05, 'samples': 22125568, 'steps': 43213, 'loss/train': 1.7281453609466553} 02/26/2022 09:32:37 - INFO - codeparrot_training - Step 43214: {'lr': 2.4255154553965902e-05, 'samples': 22126080, 'steps': 43214, 'loss/train': 1.4424231052398682} 02/26/2022 09:32:44 - INFO - codeparrot_training - Step 43215: {'lr': 2.4248124351447354e-05, 'samples': 22126592, 'steps': 43215, 'loss/train': 1.3509424924850464} 02/26/2022 09:32:48 - INFO - codeparrot_training - Step 43216: {'lr': 2.4241095115978218e-05, 'samples': 22127104, 'steps': 43216, 'loss/train': 1.6565086841583252} 02/26/2022 09:32:53 - INFO - codeparrot_training - Step 43217: {'lr': 2.4234066847588526e-05, 'samples': 22127616, 'steps': 43217, 'loss/train': 1.8656890392303467} 02/26/2022 09:32:56 - INFO - codeparrot_training - Step 43218: {'lr': 2.4227039546308394e-05, 'samples': 22128128, 'steps': 43218, 'loss/train': 2.865680456161499} 02/26/2022 09:33:02 - INFO - codeparrot_training - Step 43219: {'lr': 2.4220013212167907e-05, 'samples': 22128640, 'steps': 43219, 'loss/train': 2.386889934539795} 02/26/2022 09:33:06 - INFO - codeparrot_training - Step 43220: {'lr': 2.421298784519724e-05, 'samples': 22129152, 'steps': 43220, 'loss/train': 0.08196766674518585} 02/26/2022 09:33:11 - INFO - codeparrot_training - Step 43221: {'lr': 2.4205963445426414e-05, 'samples': 22129664, 'steps': 43221, 'loss/train': 1.9377459287643433} 02/26/2022 09:33:15 - INFO - codeparrot_training - Step 43222: {'lr': 2.4198940012885583e-05, 'samples': 22130176, 'steps': 43222, 'loss/train': 1.8808695077896118} 02/26/2022 09:33:20 - INFO - codeparrot_training - Step 43223: {'lr': 2.4191917547604748e-05, 'samples': 22130688, 'steps': 43223, 'loss/train': 1.5802565813064575} 02/26/2022 09:33:24 - INFO - codeparrot_training - Step 43224: {'lr': 2.4184896049614018e-05, 'samples': 22131200, 'steps': 43224, 'loss/train': 2.7260637283325195} 02/26/2022 09:33:29 - INFO - codeparrot_training - Step 43225: {'lr': 2.4177875518943516e-05, 'samples': 22131712, 'steps': 43225, 'loss/train': 3.524568557739258} 02/26/2022 09:33:33 - INFO - codeparrot_training - Step 43226: {'lr': 2.4170855955623265e-05, 'samples': 22132224, 'steps': 43226, 'loss/train': 1.9323642253875732} 02/26/2022 09:33:40 - INFO - codeparrot_training - Step 43227: {'lr': 2.4163837359683442e-05, 'samples': 22132736, 'steps': 43227, 'loss/train': 0.7207525372505188} 02/26/2022 09:33:43 - INFO - codeparrot_training - Step 43228: {'lr': 2.4156819731153938e-05, 'samples': 22133248, 'steps': 43228, 'loss/train': 1.1741726398468018} 02/26/2022 09:33:49 - INFO - codeparrot_training - Step 43229: {'lr': 2.414980307006495e-05, 'samples': 22133760, 'steps': 43229, 'loss/train': 2.3310253620147705} 02/26/2022 09:33:52 - INFO - codeparrot_training - Step 43230: {'lr': 2.414278737644643e-05, 'samples': 22134272, 'steps': 43230, 'loss/train': 1.921450138092041} 02/26/2022 09:33:58 - INFO - codeparrot_training - Step 43231: {'lr': 2.4135772650328597e-05, 'samples': 22134784, 'steps': 43231, 'loss/train': 1.8129687309265137} 02/26/2022 09:34:01 - INFO - codeparrot_training - Step 43232: {'lr': 2.412875889174129e-05, 'samples': 22135296, 'steps': 43232, 'loss/train': 2.041994571685791} 02/26/2022 09:34:07 - INFO - codeparrot_training - Step 43233: {'lr': 2.4121746100714713e-05, 'samples': 22135808, 'steps': 43233, 'loss/train': 2.628507614135742} 02/26/2022 09:34:10 - INFO - codeparrot_training - Step 43234: {'lr': 2.411473427727881e-05, 'samples': 22136320, 'steps': 43234, 'loss/train': 2.3726131916046143} 02/26/2022 09:34:16 - INFO - codeparrot_training - Step 43235: {'lr': 2.410772342146375e-05, 'samples': 22136832, 'steps': 43235, 'loss/train': 1.8973512649536133} 02/26/2022 09:34:19 - INFO - codeparrot_training - Step 43236: {'lr': 2.4100713533299374e-05, 'samples': 22137344, 'steps': 43236, 'loss/train': 1.4323657751083374} 02/26/2022 09:34:25 - INFO - codeparrot_training - Step 43237: {'lr': 2.4093704612815876e-05, 'samples': 22137856, 'steps': 43237, 'loss/train': 0.1868385225534439} 02/26/2022 09:34:28 - INFO - codeparrot_training - Step 43238: {'lr': 2.408669666004315e-05, 'samples': 22138368, 'steps': 43238, 'loss/train': 1.523546576499939} 02/26/2022 09:34:34 - INFO - codeparrot_training - Step 43239: {'lr': 2.4079689675011312e-05, 'samples': 22138880, 'steps': 43239, 'loss/train': 0.7568512558937073} 02/26/2022 09:34:38 - INFO - codeparrot_training - Step 43240: {'lr': 2.4072683657750365e-05, 'samples': 22139392, 'steps': 43240, 'loss/train': 1.989466905593872} 02/26/2022 09:34:43 - INFO - codeparrot_training - Step 43241: {'lr': 2.4065678608290313e-05, 'samples': 22139904, 'steps': 43241, 'loss/train': 1.2015511989593506} 02/26/2022 09:34:47 - INFO - codeparrot_training - Step 43242: {'lr': 2.4058674526661078e-05, 'samples': 22140416, 'steps': 43242, 'loss/train': 1.7314058542251587} 02/26/2022 09:34:54 - INFO - codeparrot_training - Step 43243: {'lr': 2.40516714128928e-05, 'samples': 22140928, 'steps': 43243, 'loss/train': 2.0399346351623535} 02/26/2022 09:34:57 - INFO - codeparrot_training - Step 43244: {'lr': 2.4044669267015402e-05, 'samples': 22141440, 'steps': 43244, 'loss/train': 1.6607333421707153} 02/26/2022 09:35:03 - INFO - codeparrot_training - Step 43245: {'lr': 2.4037668089058912e-05, 'samples': 22141952, 'steps': 43245, 'loss/train': 0.6854348182678223} 02/26/2022 09:35:06 - INFO - codeparrot_training - Step 43246: {'lr': 2.4030667879053225e-05, 'samples': 22142464, 'steps': 43246, 'loss/train': 1.3459415435791016} 02/26/2022 09:35:12 - INFO - codeparrot_training - Step 43247: {'lr': 2.4023668637028483e-05, 'samples': 22142976, 'steps': 43247, 'loss/train': 1.8875693082809448} 02/26/2022 09:35:15 - INFO - codeparrot_training - Step 43248: {'lr': 2.4016670363014582e-05, 'samples': 22143488, 'steps': 43248, 'loss/train': 1.2095513343811035} 02/26/2022 09:35:21 - INFO - codeparrot_training - Step 43249: {'lr': 2.4009673057041493e-05, 'samples': 22144000, 'steps': 43249, 'loss/train': 1.6082024574279785} 02/26/2022 09:35:24 - INFO - codeparrot_training - Step 43250: {'lr': 2.4002676719139166e-05, 'samples': 22144512, 'steps': 43250, 'loss/train': 1.9440619945526123} 02/26/2022 09:35:30 - INFO - codeparrot_training - Step 43251: {'lr': 2.399568134933766e-05, 'samples': 22145024, 'steps': 43251, 'loss/train': 1.5103626251220703} 02/26/2022 09:35:33 - INFO - codeparrot_training - Step 43252: {'lr': 2.3988686947666815e-05, 'samples': 22145536, 'steps': 43252, 'loss/train': 1.9341816902160645} 02/26/2022 09:35:41 - INFO - codeparrot_training - Step 43253: {'lr': 2.39816935141568e-05, 'samples': 22146048, 'steps': 43253, 'loss/train': 1.6588099002838135} 02/26/2022 09:35:44 - INFO - codeparrot_training - Step 43254: {'lr': 2.397470104883734e-05, 'samples': 22146560, 'steps': 43254, 'loss/train': 0.36438480019569397} 02/26/2022 09:35:50 - INFO - codeparrot_training - Step 43255: {'lr': 2.396770955173852e-05, 'samples': 22147072, 'steps': 43255, 'loss/train': 2.2409348487854004} 02/26/2022 09:35:53 - INFO - codeparrot_training - Step 43256: {'lr': 2.3960719022890264e-05, 'samples': 22147584, 'steps': 43256, 'loss/train': 1.9302009344100952} 02/26/2022 09:35:59 - INFO - codeparrot_training - Step 43257: {'lr': 2.3953729462322522e-05, 'samples': 22148096, 'steps': 43257, 'loss/train': 1.5525544881820679} 02/26/2022 09:36:02 - INFO - codeparrot_training - Step 43258: {'lr': 2.394674087006518e-05, 'samples': 22148608, 'steps': 43258, 'loss/train': 1.984934687614441} 02/26/2022 09:36:08 - INFO - codeparrot_training - Step 43259: {'lr': 2.3939753246148248e-05, 'samples': 22149120, 'steps': 43259, 'loss/train': 1.010594129562378} 02/26/2022 09:36:11 - INFO - codeparrot_training - Step 43260: {'lr': 2.3932766590601647e-05, 'samples': 22149632, 'steps': 43260, 'loss/train': 1.0503628253936768} 02/26/2022 09:36:17 - INFO - codeparrot_training - Step 43261: {'lr': 2.392578090345529e-05, 'samples': 22150144, 'steps': 43261, 'loss/train': 1.0467463731765747} 02/26/2022 09:36:20 - INFO - codeparrot_training - Step 43262: {'lr': 2.391879618473905e-05, 'samples': 22150656, 'steps': 43262, 'loss/train': 2.2039053440093994} 02/26/2022 09:36:28 - INFO - codeparrot_training - Step 43263: {'lr': 2.3911812434482928e-05, 'samples': 22151168, 'steps': 43263, 'loss/train': 2.0943071842193604} 02/26/2022 09:36:31 - INFO - codeparrot_training - Step 43264: {'lr': 2.390482965271684e-05, 'samples': 22151680, 'steps': 43264, 'loss/train': 1.0844749212265015} 02/26/2022 09:36:37 - INFO - codeparrot_training - Step 43265: {'lr': 2.3897847839470605e-05, 'samples': 22152192, 'steps': 43265, 'loss/train': 1.2232099771499634} 02/26/2022 09:36:40 - INFO - codeparrot_training - Step 43266: {'lr': 2.3890866994774245e-05, 'samples': 22152704, 'steps': 43266, 'loss/train': 0.8230526447296143} 02/26/2022 09:36:46 - INFO - codeparrot_training - Step 43267: {'lr': 2.3883887118657604e-05, 'samples': 22153216, 'steps': 43267, 'loss/train': 2.3939335346221924} 02/26/2022 09:36:49 - INFO - codeparrot_training - Step 43268: {'lr': 2.38769082111506e-05, 'samples': 22153728, 'steps': 43268, 'loss/train': 2.4508931636810303} 02/26/2022 09:36:55 - INFO - codeparrot_training - Step 43269: {'lr': 2.386993027228307e-05, 'samples': 22154240, 'steps': 43269, 'loss/train': 2.039569139480591} 02/26/2022 09:36:58 - INFO - codeparrot_training - Step 43270: {'lr': 2.386295330208507e-05, 'samples': 22154752, 'steps': 43270, 'loss/train': 1.6284921169281006} 02/26/2022 09:37:04 - INFO - codeparrot_training - Step 43271: {'lr': 2.385597730058625e-05, 'samples': 22155264, 'steps': 43271, 'loss/train': 1.267389178276062} 02/26/2022 09:37:07 - INFO - codeparrot_training - Step 43272: {'lr': 2.3849002267816667e-05, 'samples': 22155776, 'steps': 43272, 'loss/train': 0.8139084577560425} 02/26/2022 09:37:15 - INFO - codeparrot_training - Step 43273: {'lr': 2.3842028203806098e-05, 'samples': 22156288, 'steps': 43273, 'loss/train': 1.2592933177947998} 02/26/2022 09:37:18 - INFO - codeparrot_training - Step 43274: {'lr': 2.383505510858458e-05, 'samples': 22156800, 'steps': 43274, 'loss/train': 1.1139179468154907} 02/26/2022 09:37:24 - INFO - codeparrot_training - Step 43275: {'lr': 2.3828082982181754e-05, 'samples': 22157312, 'steps': 43275, 'loss/train': 1.969099521636963} 02/26/2022 09:37:27 - INFO - codeparrot_training - Step 43276: {'lr': 2.382111182462765e-05, 'samples': 22157824, 'steps': 43276, 'loss/train': 1.7925219535827637} 02/26/2022 09:37:33 - INFO - codeparrot_training - Step 43277: {'lr': 2.3814141635952054e-05, 'samples': 22158336, 'steps': 43277, 'loss/train': 2.275580644607544} 02/26/2022 09:37:36 - INFO - codeparrot_training - Step 43278: {'lr': 2.380717241618488e-05, 'samples': 22158848, 'steps': 43278, 'loss/train': 0.42239734530448914} 02/26/2022 09:37:42 - INFO - codeparrot_training - Step 43279: {'lr': 2.380020416535597e-05, 'samples': 22159360, 'steps': 43279, 'loss/train': 1.710526943206787} 02/26/2022 09:37:46 - INFO - codeparrot_training - Step 43280: {'lr': 2.379323688349516e-05, 'samples': 22159872, 'steps': 43280, 'loss/train': 0.8309153318405151} 02/26/2022 09:37:51 - INFO - codeparrot_training - Step 43281: {'lr': 2.3786270570632233e-05, 'samples': 22160384, 'steps': 43281, 'loss/train': 2.502852439880371} 02/26/2022 09:37:55 - INFO - codeparrot_training - Step 43282: {'lr': 2.377930522679714e-05, 'samples': 22160896, 'steps': 43282, 'loss/train': 0.5534853339195251} 02/26/2022 09:38:00 - INFO - codeparrot_training - Step 43283: {'lr': 2.377234085201968e-05, 'samples': 22161408, 'steps': 43283, 'loss/train': 0.08750396966934204} 02/26/2022 09:38:04 - INFO - codeparrot_training - Step 43284: {'lr': 2.376537744632967e-05, 'samples': 22161920, 'steps': 43284, 'loss/train': 0.8126357793807983} 02/26/2022 09:38:09 - INFO - codeparrot_training - Step 43285: {'lr': 2.3758415009756895e-05, 'samples': 22162432, 'steps': 43285, 'loss/train': 1.3534290790557861} 02/26/2022 09:38:13 - INFO - codeparrot_training - Step 43286: {'lr': 2.3751453542331296e-05, 'samples': 22162944, 'steps': 43286, 'loss/train': 1.7971128225326538} 02/26/2022 09:38:18 - INFO - codeparrot_training - Step 43287: {'lr': 2.3744493044082576e-05, 'samples': 22163456, 'steps': 43287, 'loss/train': 2.2954046726226807} 02/26/2022 09:38:22 - INFO - codeparrot_training - Step 43288: {'lr': 2.3737533515040682e-05, 'samples': 22163968, 'steps': 43288, 'loss/train': 1.7782648801803589} 02/26/2022 09:38:29 - INFO - codeparrot_training - Step 43289: {'lr': 2.3730574955235255e-05, 'samples': 22164480, 'steps': 43289, 'loss/train': 1.3095701932907104} 02/26/2022 09:38:32 - INFO - codeparrot_training - Step 43290: {'lr': 2.3723617364696247e-05, 'samples': 22164992, 'steps': 43290, 'loss/train': 2.352405548095703} 02/26/2022 09:38:38 - INFO - codeparrot_training - Step 43291: {'lr': 2.3716660743453384e-05, 'samples': 22165504, 'steps': 43291, 'loss/train': 1.0517081022262573} 02/26/2022 09:38:41 - INFO - codeparrot_training - Step 43292: {'lr': 2.3709705091536555e-05, 'samples': 22166016, 'steps': 43292, 'loss/train': 0.9705249667167664} 02/26/2022 09:38:47 - INFO - codeparrot_training - Step 43293: {'lr': 2.370275040897543e-05, 'samples': 22166528, 'steps': 43293, 'loss/train': 1.9242444038391113} 02/26/2022 09:38:50 - INFO - codeparrot_training - Step 43294: {'lr': 2.369579669579988e-05, 'samples': 22167040, 'steps': 43294, 'loss/train': 1.4593243598937988} 02/26/2022 09:38:56 - INFO - codeparrot_training - Step 43295: {'lr': 2.3688843952039656e-05, 'samples': 22167552, 'steps': 43295, 'loss/train': 2.6275806427001953} 02/26/2022 09:38:59 - INFO - codeparrot_training - Step 43296: {'lr': 2.3681892177724652e-05, 'samples': 22168064, 'steps': 43296, 'loss/train': 0.8268952369689941} 02/26/2022 09:39:05 - INFO - codeparrot_training - Step 43297: {'lr': 2.3674941372884478e-05, 'samples': 22168576, 'steps': 43297, 'loss/train': 1.5829060077667236} 02/26/2022 09:39:08 - INFO - codeparrot_training - Step 43298: {'lr': 2.3667991537549006e-05, 'samples': 22169088, 'steps': 43298, 'loss/train': 1.8122408390045166} 02/26/2022 09:39:15 - INFO - codeparrot_training - Step 43299: {'lr': 2.3661042671747957e-05, 'samples': 22169600, 'steps': 43299, 'loss/train': 1.3882919549942017} 02/26/2022 09:39:19 - INFO - codeparrot_training - Step 43300: {'lr': 2.365409477551117e-05, 'samples': 22170112, 'steps': 43300, 'loss/train': 1.8376020193099976} 02/26/2022 09:39:24 - INFO - codeparrot_training - Step 43301: {'lr': 2.364714784886837e-05, 'samples': 22170624, 'steps': 43301, 'loss/train': 2.0100438594818115} 02/26/2022 09:39:28 - INFO - codeparrot_training - Step 43302: {'lr': 2.3640201891849315e-05, 'samples': 22171136, 'steps': 43302, 'loss/train': 2.088324546813965} 02/26/2022 09:39:33 - INFO - codeparrot_training - Step 43303: {'lr': 2.3633256904483786e-05, 'samples': 22171648, 'steps': 43303, 'loss/train': 1.601377248764038} 02/26/2022 09:39:37 - INFO - codeparrot_training - Step 43304: {'lr': 2.3626312886801423e-05, 'samples': 22172160, 'steps': 43304, 'loss/train': 2.2252700328826904} 02/26/2022 09:39:43 - INFO - codeparrot_training - Step 43305: {'lr': 2.3619369838832118e-05, 'samples': 22172672, 'steps': 43305, 'loss/train': 3.0573503971099854} 02/26/2022 09:39:46 - INFO - codeparrot_training - Step 43306: {'lr': 2.3612427760605542e-05, 'samples': 22173184, 'steps': 43306, 'loss/train': 1.2639302015304565} 02/26/2022 09:39:52 - INFO - codeparrot_training - Step 43307: {'lr': 2.3605486652151426e-05, 'samples': 22173696, 'steps': 43307, 'loss/train': 1.4656106233596802} 02/26/2022 09:39:55 - INFO - codeparrot_training - Step 43308: {'lr': 2.359854651349949e-05, 'samples': 22174208, 'steps': 43308, 'loss/train': 2.149136543273926} 02/26/2022 09:40:02 - INFO - codeparrot_training - Step 43309: {'lr': 2.359160734467952e-05, 'samples': 22174720, 'steps': 43309, 'loss/train': 3.0191521644592285} 02/26/2022 09:40:06 - INFO - codeparrot_training - Step 43310: {'lr': 2.3584669145721217e-05, 'samples': 22175232, 'steps': 43310, 'loss/train': 1.396069049835205} 02/26/2022 09:40:11 - INFO - codeparrot_training - Step 43311: {'lr': 2.35777319166543e-05, 'samples': 22175744, 'steps': 43311, 'loss/train': 2.0735700130462646} 02/26/2022 09:40:15 - INFO - codeparrot_training - Step 43312: {'lr': 2.3570795657508442e-05, 'samples': 22176256, 'steps': 43312, 'loss/train': 2.6623117923736572} 02/26/2022 09:40:21 - INFO - codeparrot_training - Step 43313: {'lr': 2.356386036831343e-05, 'samples': 22176768, 'steps': 43313, 'loss/train': 2.631047487258911} 02/26/2022 09:40:24 - INFO - codeparrot_training - Step 43314: {'lr': 2.3556926049098952e-05, 'samples': 22177280, 'steps': 43314, 'loss/train': 1.9996322393417358} 02/26/2022 09:40:28 - INFO - codeparrot_training - Step 43315: {'lr': 2.354999269989469e-05, 'samples': 22177792, 'steps': 43315, 'loss/train': 8.571395874023438} 02/26/2022 09:40:33 - INFO - codeparrot_training - Step 43316: {'lr': 2.354306032073031e-05, 'samples': 22178304, 'steps': 43316, 'loss/train': 1.7630122900009155} 02/26/2022 09:40:37 - INFO - codeparrot_training - Step 43317: {'lr': 2.353612891163559e-05, 'samples': 22178816, 'steps': 43317, 'loss/train': 0.7591730356216431} 02/26/2022 09:40:42 - INFO - codeparrot_training - Step 43318: {'lr': 2.3529198472640206e-05, 'samples': 22179328, 'steps': 43318, 'loss/train': 1.7589277029037476} 02/26/2022 09:40:46 - INFO - codeparrot_training - Step 43319: {'lr': 2.3522269003773798e-05, 'samples': 22179840, 'steps': 43319, 'loss/train': 1.097955584526062} 02/26/2022 09:40:53 - INFO - codeparrot_training - Step 43320: {'lr': 2.3515340505066042e-05, 'samples': 22180352, 'steps': 43320, 'loss/train': 2.1457056999206543} 02/26/2022 09:40:56 - INFO - codeparrot_training - Step 43321: {'lr': 2.3508412976546683e-05, 'samples': 22180864, 'steps': 43321, 'loss/train': 1.7428361177444458} 02/26/2022 09:41:02 - INFO - codeparrot_training - Step 43322: {'lr': 2.350148641824537e-05, 'samples': 22181376, 'steps': 43322, 'loss/train': 2.8708314895629883} 02/26/2022 09:41:05 - INFO - codeparrot_training - Step 43323: {'lr': 2.3494560830191775e-05, 'samples': 22181888, 'steps': 43323, 'loss/train': 0.7016080617904663} 02/26/2022 09:41:11 - INFO - codeparrot_training - Step 43324: {'lr': 2.3487636212415508e-05, 'samples': 22182400, 'steps': 43324, 'loss/train': 1.868202567100525} 02/26/2022 09:41:14 - INFO - codeparrot_training - Step 43325: {'lr': 2.3480712564946327e-05, 'samples': 22182912, 'steps': 43325, 'loss/train': 1.9066132307052612} 02/26/2022 09:41:20 - INFO - codeparrot_training - Step 43326: {'lr': 2.347378988781379e-05, 'samples': 22183424, 'steps': 43326, 'loss/train': 3.118809700012207} 02/26/2022 09:41:23 - INFO - codeparrot_training - Step 43327: {'lr': 2.346686818104771e-05, 'samples': 22183936, 'steps': 43327, 'loss/train': 1.613738775253296} 02/26/2022 09:41:29 - INFO - codeparrot_training - Step 43328: {'lr': 2.3459947444677553e-05, 'samples': 22184448, 'steps': 43328, 'loss/train': 2.110063314437866} 02/26/2022 09:41:32 - INFO - codeparrot_training - Step 43329: {'lr': 2.3453027678733085e-05, 'samples': 22184960, 'steps': 43329, 'loss/train': 1.4017025232315063} 02/26/2022 09:41:40 - INFO - codeparrot_training - Step 43330: {'lr': 2.3446108883243834e-05, 'samples': 22185472, 'steps': 43330, 'loss/train': 2.2838540077209473} 02/26/2022 09:41:43 - INFO - codeparrot_training - Step 43331: {'lr': 2.3439191058239634e-05, 'samples': 22185984, 'steps': 43331, 'loss/train': 0.2765112817287445} 02/26/2022 09:41:49 - INFO - codeparrot_training - Step 43332: {'lr': 2.3432274203749908e-05, 'samples': 22186496, 'steps': 43332, 'loss/train': 1.5505725145339966} 02/26/2022 09:41:52 - INFO - codeparrot_training - Step 43333: {'lr': 2.3425358319804414e-05, 'samples': 22187008, 'steps': 43333, 'loss/train': 1.612720012664795} 02/26/2022 09:41:58 - INFO - codeparrot_training - Step 43334: {'lr': 2.3418443406432704e-05, 'samples': 22187520, 'steps': 43334, 'loss/train': 2.1533167362213135} 02/26/2022 09:42:01 - INFO - codeparrot_training - Step 43335: {'lr': 2.3411529463664484e-05, 'samples': 22188032, 'steps': 43335, 'loss/train': 2.0199599266052246} 02/26/2022 09:42:07 - INFO - codeparrot_training - Step 43336: {'lr': 2.3404616491529306e-05, 'samples': 22188544, 'steps': 43336, 'loss/train': 1.377267837524414} 02/26/2022 09:42:10 - INFO - codeparrot_training - Step 43337: {'lr': 2.3397704490056792e-05, 'samples': 22189056, 'steps': 43337, 'loss/train': 0.9962981343269348} 02/26/2022 09:42:16 - INFO - codeparrot_training - Step 43338: {'lr': 2.3390793459276528e-05, 'samples': 22189568, 'steps': 43338, 'loss/train': 8.646394729614258} 02/26/2022 09:42:19 - INFO - codeparrot_training - Step 43339: {'lr': 2.338388339921818e-05, 'samples': 22190080, 'steps': 43339, 'loss/train': 1.412485957145691} 02/26/2022 09:42:25 - INFO - codeparrot_training - Step 43340: {'lr': 2.3376974309911343e-05, 'samples': 22190592, 'steps': 43340, 'loss/train': 1.7455847263336182} 02/26/2022 09:42:28 - INFO - codeparrot_training - Step 43341: {'lr': 2.337006619138557e-05, 'samples': 22191104, 'steps': 43341, 'loss/train': 2.3395497798919678} 02/26/2022 09:42:34 - INFO - codeparrot_training - Step 43342: {'lr': 2.3363159043670425e-05, 'samples': 22191616, 'steps': 43342, 'loss/train': 0.5850828289985657} 02/26/2022 09:42:37 - INFO - codeparrot_training - Step 43343: {'lr': 2.3356252866795575e-05, 'samples': 22192128, 'steps': 43343, 'loss/train': 1.5073628425598145} 02/26/2022 09:42:43 - INFO - codeparrot_training - Step 43344: {'lr': 2.334934766079058e-05, 'samples': 22192640, 'steps': 43344, 'loss/train': 1.417761206626892} 02/26/2022 09:42:46 - INFO - codeparrot_training - Step 43345: {'lr': 2.3342443425685006e-05, 'samples': 22193152, 'steps': 43345, 'loss/train': 1.6512362957000732} 02/26/2022 09:42:54 - INFO - codeparrot_training - Step 43346: {'lr': 2.3335540161508407e-05, 'samples': 22193664, 'steps': 43346, 'loss/train': 1.7044094800949097} 02/26/2022 09:42:57 - INFO - codeparrot_training - Step 43347: {'lr': 2.3328637868290398e-05, 'samples': 22194176, 'steps': 43347, 'loss/train': 0.9756754636764526} 02/26/2022 09:43:03 - INFO - codeparrot_training - Step 43348: {'lr': 2.332173654606054e-05, 'samples': 22194688, 'steps': 43348, 'loss/train': 2.291933059692383} 02/26/2022 09:43:07 - INFO - codeparrot_training - Step 43349: {'lr': 2.3314836194848394e-05, 'samples': 22195200, 'steps': 43349, 'loss/train': 1.176311731338501} 02/26/2022 09:43:12 - INFO - codeparrot_training - Step 43350: {'lr': 2.330793681468352e-05, 'samples': 22195712, 'steps': 43350, 'loss/train': 1.9870012998580933} 02/26/2022 09:43:15 - INFO - codeparrot_training - Step 43351: {'lr': 2.3301038405595388e-05, 'samples': 22196224, 'steps': 43351, 'loss/train': 1.7480851411819458} 02/26/2022 09:43:21 - INFO - codeparrot_training - Step 43352: {'lr': 2.3294140967613675e-05, 'samples': 22196736, 'steps': 43352, 'loss/train': 2.084657669067383} 02/26/2022 09:43:24 - INFO - codeparrot_training - Step 43353: {'lr': 2.3287244500767884e-05, 'samples': 22197248, 'steps': 43353, 'loss/train': 0.8548081517219543} 02/26/2022 09:43:30 - INFO - codeparrot_training - Step 43354: {'lr': 2.328034900508752e-05, 'samples': 22197760, 'steps': 43354, 'loss/train': 0.9051329493522644} 02/26/2022 09:43:34 - INFO - codeparrot_training - Step 43355: {'lr': 2.327345448060214e-05, 'samples': 22198272, 'steps': 43355, 'loss/train': 0.10485494881868362} 02/26/2022 09:43:41 - INFO - codeparrot_training - Step 43356: {'lr': 2.3266560927341306e-05, 'samples': 22198784, 'steps': 43356, 'loss/train': 1.3880500793457031} 02/26/2022 09:43:44 - INFO - codeparrot_training - Step 43357: {'lr': 2.325966834533455e-05, 'samples': 22199296, 'steps': 43357, 'loss/train': 1.5446600914001465} 02/26/2022 09:43:50 - INFO - codeparrot_training - Step 43358: {'lr': 2.3252776734611377e-05, 'samples': 22199808, 'steps': 43358, 'loss/train': 2.0897037982940674} 02/26/2022 09:43:53 - INFO - codeparrot_training - Step 43359: {'lr': 2.324588609520123e-05, 'samples': 22200320, 'steps': 43359, 'loss/train': 0.7792462110519409} 02/26/2022 09:43:59 - INFO - codeparrot_training - Step 43360: {'lr': 2.3238996427133783e-05, 'samples': 22200832, 'steps': 43360, 'loss/train': 1.8640575408935547} 02/26/2022 09:44:02 - INFO - codeparrot_training - Step 43361: {'lr': 2.3232107730438408e-05, 'samples': 22201344, 'steps': 43361, 'loss/train': 1.979748010635376} 02/26/2022 09:44:08 - INFO - codeparrot_training - Step 43362: {'lr': 2.3225220005144763e-05, 'samples': 22201856, 'steps': 43362, 'loss/train': 1.7013325691223145} 02/26/2022 09:44:12 - INFO - codeparrot_training - Step 43363: {'lr': 2.3218333251282198e-05, 'samples': 22202368, 'steps': 43363, 'loss/train': 0.49709972739219666} 02/26/2022 09:44:17 - INFO - codeparrot_training - Step 43364: {'lr': 2.321144746888032e-05, 'samples': 22202880, 'steps': 43364, 'loss/train': 1.5686622858047485} 02/26/2022 09:44:20 - INFO - codeparrot_training - Step 43365: {'lr': 2.3204562657968526e-05, 'samples': 22203392, 'steps': 43365, 'loss/train': 0.9851326942443848} 02/26/2022 09:44:28 - INFO - codeparrot_training - Step 43366: {'lr': 2.3197678818576485e-05, 'samples': 22203904, 'steps': 43366, 'loss/train': 1.5127288103103638} 02/26/2022 09:44:31 - INFO - codeparrot_training - Step 43367: {'lr': 2.3190795950733478e-05, 'samples': 22204416, 'steps': 43367, 'loss/train': 2.011246919631958} 02/26/2022 09:44:37 - INFO - codeparrot_training - Step 43368: {'lr': 2.3183914054469096e-05, 'samples': 22204928, 'steps': 43368, 'loss/train': 0.7503435015678406} 02/26/2022 09:44:40 - INFO - codeparrot_training - Step 43369: {'lr': 2.3177033129812785e-05, 'samples': 22205440, 'steps': 43369, 'loss/train': 2.5692148208618164} 02/26/2022 09:44:46 - INFO - codeparrot_training - Step 43370: {'lr': 2.3170153176794135e-05, 'samples': 22205952, 'steps': 43370, 'loss/train': 0.31932225823402405} 02/26/2022 09:44:49 - INFO - codeparrot_training - Step 43371: {'lr': 2.3163274195442425e-05, 'samples': 22206464, 'steps': 43371, 'loss/train': 1.7933956384658813} 02/26/2022 09:44:55 - INFO - codeparrot_training - Step 43372: {'lr': 2.3156396185787244e-05, 'samples': 22206976, 'steps': 43372, 'loss/train': 1.6930683851242065} 02/26/2022 09:44:58 - INFO - codeparrot_training - Step 43373: {'lr': 2.314951914785801e-05, 'samples': 22207488, 'steps': 43373, 'loss/train': 2.106527090072632} 02/26/2022 09:45:04 - INFO - codeparrot_training - Step 43374: {'lr': 2.3142643081684234e-05, 'samples': 22208000, 'steps': 43374, 'loss/train': 0.9784080386161804} 02/26/2022 09:45:09 - INFO - codeparrot_training - Step 43375: {'lr': 2.313576798729536e-05, 'samples': 22208512, 'steps': 43375, 'loss/train': 1.622563123703003} 02/26/2022 09:45:13 - INFO - codeparrot_training - Step 43376: {'lr': 2.312889386472078e-05, 'samples': 22209024, 'steps': 43376, 'loss/train': 1.9017900228500366} 02/26/2022 09:45:19 - INFO - codeparrot_training - Step 43377: {'lr': 2.312202071398997e-05, 'samples': 22209536, 'steps': 43377, 'loss/train': 0.9812229871749878} 02/26/2022 09:45:22 - INFO - codeparrot_training - Step 43378: {'lr': 2.3115148535132414e-05, 'samples': 22210048, 'steps': 43378, 'loss/train': 0.97850501537323} 02/26/2022 09:45:28 - INFO - codeparrot_training - Step 43379: {'lr': 2.3108277328177522e-05, 'samples': 22210560, 'steps': 43379, 'loss/train': 1.0578759908676147} 02/26/2022 09:45:31 - INFO - codeparrot_training - Step 43380: {'lr': 2.3101407093154726e-05, 'samples': 22211072, 'steps': 43380, 'loss/train': 1.8145272731781006} 02/26/2022 09:45:35 - INFO - codeparrot_training - Step 43381: {'lr': 2.309453783009338e-05, 'samples': 22211584, 'steps': 43381, 'loss/train': 1.930153489112854} 02/26/2022 09:45:42 - INFO - codeparrot_training - Step 43382: {'lr': 2.3087669539023054e-05, 'samples': 22212096, 'steps': 43382, 'loss/train': 1.6122373342514038} 02/26/2022 09:45:45 - INFO - codeparrot_training - Step 43383: {'lr': 2.3080802219973108e-05, 'samples': 22212608, 'steps': 43383, 'loss/train': 1.7784168720245361} 02/26/2022 09:45:51 - INFO - codeparrot_training - Step 43384: {'lr': 2.3073935872972933e-05, 'samples': 22213120, 'steps': 43384, 'loss/train': 1.5295833349227905} 02/26/2022 09:45:54 - INFO - codeparrot_training - Step 43385: {'lr': 2.306707049805193e-05, 'samples': 22213632, 'steps': 43385, 'loss/train': 1.7486096620559692} 02/26/2022 09:46:01 - INFO - codeparrot_training - Step 43386: {'lr': 2.3060206095239568e-05, 'samples': 22214144, 'steps': 43386, 'loss/train': 2.384612798690796} 02/26/2022 09:46:05 - INFO - codeparrot_training - Step 43387: {'lr': 2.3053342664565215e-05, 'samples': 22214656, 'steps': 43387, 'loss/train': 2.0957627296447754} 02/26/2022 09:46:10 - INFO - codeparrot_training - Step 43388: {'lr': 2.304648020605829e-05, 'samples': 22215168, 'steps': 43388, 'loss/train': 2.064143657684326} 02/26/2022 09:46:13 - INFO - codeparrot_training - Step 43389: {'lr': 2.3039618719748135e-05, 'samples': 22215680, 'steps': 43389, 'loss/train': 1.608757734298706} 02/26/2022 09:46:19 - INFO - codeparrot_training - Step 43390: {'lr': 2.3032758205664224e-05, 'samples': 22216192, 'steps': 43390, 'loss/train': 2.877185106277466} 02/26/2022 09:46:23 - INFO - codeparrot_training - Step 43391: {'lr': 2.302589866383592e-05, 'samples': 22216704, 'steps': 43391, 'loss/train': 1.7679684162139893} 02/26/2022 09:46:28 - INFO - codeparrot_training - Step 43392: {'lr': 2.3019040094292565e-05, 'samples': 22217216, 'steps': 43392, 'loss/train': 0.6853064298629761} 02/26/2022 09:46:31 - INFO - codeparrot_training - Step 43393: {'lr': 2.301218249706355e-05, 'samples': 22217728, 'steps': 43393, 'loss/train': 0.793110191822052} 02/26/2022 09:46:37 - INFO - codeparrot_training - Step 43394: {'lr': 2.3005325872178296e-05, 'samples': 22218240, 'steps': 43394, 'loss/train': 0.7027353048324585} 02/26/2022 09:46:41 - INFO - codeparrot_training - Step 43395: {'lr': 2.299847021966617e-05, 'samples': 22218752, 'steps': 43395, 'loss/train': 0.8983511924743652} 02/26/2022 09:46:47 - INFO - codeparrot_training - Step 43396: {'lr': 2.299161553955645e-05, 'samples': 22219264, 'steps': 43396, 'loss/train': 0.4964800775051117} 02/26/2022 09:46:51 - INFO - codeparrot_training - Step 43397: {'lr': 2.298476183187867e-05, 'samples': 22219776, 'steps': 43397, 'loss/train': 1.0525320768356323} 02/26/2022 09:46:56 - INFO - codeparrot_training - Step 43398: {'lr': 2.2977909096661974e-05, 'samples': 22220288, 'steps': 43398, 'loss/train': 1.3047536611557007} 02/26/2022 09:47:00 - INFO - codeparrot_training - Step 43399: {'lr': 2.2971057333935895e-05, 'samples': 22220800, 'steps': 43399, 'loss/train': 1.5811887979507446} 02/26/2022 09:47:05 - INFO - codeparrot_training - Step 43400: {'lr': 2.296420654372966e-05, 'samples': 22221312, 'steps': 43400, 'loss/train': 1.232550859451294} 02/26/2022 09:47:08 - INFO - codeparrot_training - Step 43401: {'lr': 2.295735672607277e-05, 'samples': 22221824, 'steps': 43401, 'loss/train': 2.2581450939178467} 02/26/2022 09:47:14 - INFO - codeparrot_training - Step 43402: {'lr': 2.295050788099437e-05, 'samples': 22222336, 'steps': 43402, 'loss/train': 1.3651314973831177} 02/26/2022 09:47:17 - INFO - codeparrot_training - Step 43403: {'lr': 2.2943660008523963e-05, 'samples': 22222848, 'steps': 43403, 'loss/train': 1.9363576173782349} 02/26/2022 09:47:23 - INFO - codeparrot_training - Step 43404: {'lr': 2.2936813108690752e-05, 'samples': 22223360, 'steps': 43404, 'loss/train': 1.8281817436218262} 02/26/2022 09:47:26 - INFO - codeparrot_training - Step 43405: {'lr': 2.2929967181524235e-05, 'samples': 22223872, 'steps': 43405, 'loss/train': 1.103778600692749} 02/26/2022 09:47:32 - INFO - codeparrot_training - Step 43406: {'lr': 2.2923122227053532e-05, 'samples': 22224384, 'steps': 43406, 'loss/train': 2.184523820877075} 02/26/2022 09:47:36 - INFO - codeparrot_training - Step 43407: {'lr': 2.291627824530812e-05, 'samples': 22224896, 'steps': 43407, 'loss/train': 2.1645805835723877} 02/26/2022 09:47:42 - INFO - codeparrot_training - Step 43408: {'lr': 2.2909435236317222e-05, 'samples': 22225408, 'steps': 43408, 'loss/train': 1.6693423986434937} 02/26/2022 09:47:45 - INFO - codeparrot_training - Step 43409: {'lr': 2.2902593200110234e-05, 'samples': 22225920, 'steps': 43409, 'loss/train': 1.8765085935592651} 02/26/2022 09:47:51 - INFO - codeparrot_training - Step 43410: {'lr': 2.2895752136716434e-05, 'samples': 22226432, 'steps': 43410, 'loss/train': 0.6838909983634949} 02/26/2022 09:47:54 - INFO - codeparrot_training - Step 43411: {'lr': 2.288891204616511e-05, 'samples': 22226944, 'steps': 43411, 'loss/train': 1.6518410444259644} 02/26/2022 09:48:00 - INFO - codeparrot_training - Step 43412: {'lr': 2.2882072928485515e-05, 'samples': 22227456, 'steps': 43412, 'loss/train': 1.6013000011444092} 02/26/2022 09:48:03 - INFO - codeparrot_training - Step 43413: {'lr': 2.2875234783707067e-05, 'samples': 22227968, 'steps': 43413, 'loss/train': 3.5163512229919434} 02/26/2022 09:48:09 - INFO - codeparrot_training - Step 43414: {'lr': 2.2868397611858994e-05, 'samples': 22228480, 'steps': 43414, 'loss/train': 1.1242539882659912} 02/26/2022 09:48:12 - INFO - codeparrot_training - Step 43415: {'lr': 2.2861561412970576e-05, 'samples': 22228992, 'steps': 43415, 'loss/train': 2.2537989616394043} 02/26/2022 09:48:18 - INFO - codeparrot_training - Step 43416: {'lr': 2.2854726187071046e-05, 'samples': 22229504, 'steps': 43416, 'loss/train': 1.7038037776947021} 02/26/2022 09:48:22 - INFO - codeparrot_training - Step 43417: {'lr': 2.284789193418982e-05, 'samples': 22230016, 'steps': 43417, 'loss/train': 1.4832667112350464} 02/26/2022 09:48:27 - INFO - codeparrot_training - Step 43418: {'lr': 2.2841058654356068e-05, 'samples': 22230528, 'steps': 43418, 'loss/train': 3.860633611679077} 02/26/2022 09:48:31 - INFO - codeparrot_training - Step 43419: {'lr': 2.2834226347599103e-05, 'samples': 22231040, 'steps': 43419, 'loss/train': 0.5670018792152405} 02/26/2022 09:48:36 - INFO - codeparrot_training - Step 43420: {'lr': 2.2827395013948123e-05, 'samples': 22231552, 'steps': 43420, 'loss/train': 2.107569932937622} 02/26/2022 09:48:40 - INFO - codeparrot_training - Step 43421: {'lr': 2.282056465343249e-05, 'samples': 22232064, 'steps': 43421, 'loss/train': 1.589634656906128} 02/26/2022 09:48:45 - INFO - codeparrot_training - Step 43422: {'lr': 2.281373526608138e-05, 'samples': 22232576, 'steps': 43422, 'loss/train': 1.9231767654418945} 02/26/2022 09:48:49 - INFO - codeparrot_training - Step 43423: {'lr': 2.2806906851924186e-05, 'samples': 22233088, 'steps': 43423, 'loss/train': 0.6679226160049438} 02/26/2022 09:48:54 - INFO - codeparrot_training - Step 43424: {'lr': 2.2800079410989966e-05, 'samples': 22233600, 'steps': 43424, 'loss/train': 2.052579641342163} 02/26/2022 09:48:57 - INFO - codeparrot_training - Step 43425: {'lr': 2.2793252943308114e-05, 'samples': 22234112, 'steps': 43425, 'loss/train': 0.2066899687051773} 02/26/2022 09:49:03 - INFO - codeparrot_training - Step 43426: {'lr': 2.278642744890777e-05, 'samples': 22234624, 'steps': 43426, 'loss/train': 1.3800404071807861} 02/26/2022 09:49:07 - INFO - codeparrot_training - Step 43427: {'lr': 2.2779602927818305e-05, 'samples': 22235136, 'steps': 43427, 'loss/train': 2.0924065113067627} 02/26/2022 09:49:12 - INFO - codeparrot_training - Step 43428: {'lr': 2.2772779380068802e-05, 'samples': 22235648, 'steps': 43428, 'loss/train': 1.7433634996414185} 02/26/2022 09:49:15 - INFO - codeparrot_training - Step 43429: {'lr': 2.2765956805688576e-05, 'samples': 22236160, 'steps': 43429, 'loss/train': 2.530639410018921} 02/26/2022 09:49:21 - INFO - codeparrot_training - Step 43430: {'lr': 2.275913520470685e-05, 'samples': 22236672, 'steps': 43430, 'loss/train': 1.3204100131988525} 02/26/2022 09:49:24 - INFO - codeparrot_training - Step 43431: {'lr': 2.275231457715285e-05, 'samples': 22237184, 'steps': 43431, 'loss/train': 1.4688754081726074} 02/26/2022 09:49:31 - INFO - codeparrot_training - Step 43432: {'lr': 2.2745494923055694e-05, 'samples': 22237696, 'steps': 43432, 'loss/train': 1.9970577955245972} 02/26/2022 09:49:35 - INFO - codeparrot_training - Step 43433: {'lr': 2.2738676242444744e-05, 'samples': 22238208, 'steps': 43433, 'loss/train': 0.5037768483161926} 02/26/2022 09:49:40 - INFO - codeparrot_training - Step 43434: {'lr': 2.273185853534912e-05, 'samples': 22238720, 'steps': 43434, 'loss/train': 1.9678643941879272} 02/26/2022 09:49:44 - INFO - codeparrot_training - Step 43435: {'lr': 2.272504180179802e-05, 'samples': 22239232, 'steps': 43435, 'loss/train': 1.108302354812622} 02/26/2022 09:49:49 - INFO - codeparrot_training - Step 43436: {'lr': 2.2718226041820724e-05, 'samples': 22239744, 'steps': 43436, 'loss/train': 2.790809392929077} 02/26/2022 09:49:52 - INFO - codeparrot_training - Step 43437: {'lr': 2.2711411255446374e-05, 'samples': 22240256, 'steps': 43437, 'loss/train': 1.8792283535003662} 02/26/2022 09:49:58 - INFO - codeparrot_training - Step 43438: {'lr': 2.2704597442704173e-05, 'samples': 22240768, 'steps': 43438, 'loss/train': 1.9621168375015259} 02/26/2022 09:50:01 - INFO - codeparrot_training - Step 43439: {'lr': 2.2697784603623235e-05, 'samples': 22241280, 'steps': 43439, 'loss/train': 1.5062484741210938} 02/26/2022 09:50:07 - INFO - codeparrot_training - Step 43440: {'lr': 2.2690972738232866e-05, 'samples': 22241792, 'steps': 43440, 'loss/train': 0.8874250650405884} 02/26/2022 09:50:14 - INFO - codeparrot_training - Step 43441: {'lr': 2.268416184656219e-05, 'samples': 22242304, 'steps': 43441, 'loss/train': 2.6573238372802734} 02/26/2022 09:50:17 - INFO - codeparrot_training - Step 43442: {'lr': 2.26773519286404e-05, 'samples': 22242816, 'steps': 43442, 'loss/train': 2.785123825073242} 02/26/2022 09:50:23 - INFO - codeparrot_training - Step 43443: {'lr': 2.2670542984496583e-05, 'samples': 22243328, 'steps': 43443, 'loss/train': 1.7695703506469727} 02/26/2022 09:50:26 - INFO - codeparrot_training - Step 43444: {'lr': 2.2663735014160026e-05, 'samples': 22243840, 'steps': 43444, 'loss/train': 1.756841778755188} 02/26/2022 09:50:32 - INFO - codeparrot_training - Step 43445: {'lr': 2.265692801765984e-05, 'samples': 22244352, 'steps': 43445, 'loss/train': 1.8958699703216553} 02/26/2022 09:50:35 - INFO - codeparrot_training - Step 43446: {'lr': 2.2650121995025198e-05, 'samples': 22244864, 'steps': 43446, 'loss/train': 2.4778921604156494} 02/26/2022 09:50:41 - INFO - codeparrot_training - Step 43447: {'lr': 2.2643316946285186e-05, 'samples': 22245376, 'steps': 43447, 'loss/train': 1.6054095029830933} 02/26/2022 09:50:44 - INFO - codeparrot_training - Step 43448: {'lr': 2.263651287146906e-05, 'samples': 22245888, 'steps': 43448, 'loss/train': 1.0989842414855957} 02/26/2022 09:50:50 - INFO - codeparrot_training - Step 43449: {'lr': 2.2629709770605912e-05, 'samples': 22246400, 'steps': 43449, 'loss/train': 1.7166266441345215} 02/26/2022 09:50:53 - INFO - codeparrot_training - Step 43450: {'lr': 2.2622907643724903e-05, 'samples': 22246912, 'steps': 43450, 'loss/train': 1.7227504253387451} 02/26/2022 09:50:59 - INFO - codeparrot_training - Step 43451: {'lr': 2.2616106490855076e-05, 'samples': 22247424, 'steps': 43451, 'loss/train': 0.9361081123352051} 02/26/2022 09:51:02 - INFO - codeparrot_training - Step 43452: {'lr': 2.2609306312025735e-05, 'samples': 22247936, 'steps': 43452, 'loss/train': 8.648787498474121} 02/26/2022 09:51:09 - INFO - codeparrot_training - Step 43453: {'lr': 2.2602507107265886e-05, 'samples': 22248448, 'steps': 43453, 'loss/train': 2.183933973312378} 02/26/2022 09:51:12 - INFO - codeparrot_training - Step 43454: {'lr': 2.2595708876604727e-05, 'samples': 22248960, 'steps': 43454, 'loss/train': 1.903673529624939} 02/26/2022 09:51:18 - INFO - codeparrot_training - Step 43455: {'lr': 2.2588911620071262e-05, 'samples': 22249472, 'steps': 43455, 'loss/train': 1.6323210000991821} 02/26/2022 09:51:21 - INFO - codeparrot_training - Step 43456: {'lr': 2.2582115337694746e-05, 'samples': 22249984, 'steps': 43456, 'loss/train': 2.4872117042541504} 02/26/2022 09:51:27 - INFO - codeparrot_training - Step 43457: {'lr': 2.2575320029504186e-05, 'samples': 22250496, 'steps': 43457, 'loss/train': 2.29856276512146} 02/26/2022 09:51:30 - INFO - codeparrot_training - Step 43458: {'lr': 2.2568525695528834e-05, 'samples': 22251008, 'steps': 43458, 'loss/train': 2.756319284439087} 02/26/2022 09:51:36 - INFO - codeparrot_training - Step 43459: {'lr': 2.2561732335797615e-05, 'samples': 22251520, 'steps': 43459, 'loss/train': 1.7654610872268677} 02/26/2022 09:51:39 - INFO - codeparrot_training - Step 43460: {'lr': 2.255493995033975e-05, 'samples': 22252032, 'steps': 43460, 'loss/train': 1.4436110258102417} 02/26/2022 09:51:45 - INFO - codeparrot_training - Step 43461: {'lr': 2.254814853918427e-05, 'samples': 22252544, 'steps': 43461, 'loss/train': 1.920397400856018} 02/26/2022 09:51:48 - INFO - codeparrot_training - Step 43462: {'lr': 2.2541358102360383e-05, 'samples': 22253056, 'steps': 43462, 'loss/train': 2.2659895420074463} 02/26/2022 09:51:54 - INFO - codeparrot_training - Step 43463: {'lr': 2.2534568639897e-05, 'samples': 22253568, 'steps': 43463, 'loss/train': 0.08842244744300842} 02/26/2022 09:51:58 - INFO - codeparrot_training - Step 43464: {'lr': 2.252778015182333e-05, 'samples': 22254080, 'steps': 43464, 'loss/train': 3.660236358642578} 02/26/2022 09:52:03 - INFO - codeparrot_training - Step 43465: {'lr': 2.252099263816837e-05, 'samples': 22254592, 'steps': 43465, 'loss/train': 1.9817352294921875} 02/26/2022 09:52:07 - INFO - codeparrot_training - Step 43466: {'lr': 2.2514206098961348e-05, 'samples': 22255104, 'steps': 43466, 'loss/train': 1.95352041721344} 02/26/2022 09:52:12 - INFO - codeparrot_training - Step 43467: {'lr': 2.250742053423116e-05, 'samples': 22255616, 'steps': 43467, 'loss/train': 2.5268237590789795} 02/26/2022 09:52:16 - INFO - codeparrot_training - Step 43468: {'lr': 2.2500635944006976e-05, 'samples': 22256128, 'steps': 43468, 'loss/train': 2.38124418258667} 02/26/2022 09:52:21 - INFO - codeparrot_training - Step 43469: {'lr': 2.249385232831777e-05, 'samples': 22256640, 'steps': 43469, 'loss/train': 1.3344988822937012} 02/26/2022 09:52:25 - INFO - codeparrot_training - Step 43470: {'lr': 2.248706968719272e-05, 'samples': 22257152, 'steps': 43470, 'loss/train': 2.3130009174346924} 02/26/2022 09:52:30 - INFO - codeparrot_training - Step 43471: {'lr': 2.2480288020660793e-05, 'samples': 22257664, 'steps': 43471, 'loss/train': 1.9131823778152466} 02/26/2022 09:52:34 - INFO - codeparrot_training - Step 43472: {'lr': 2.2473507328751085e-05, 'samples': 22258176, 'steps': 43472, 'loss/train': 1.6652742624282837} 02/26/2022 09:52:39 - INFO - codeparrot_training - Step 43473: {'lr': 2.24667276114926e-05, 'samples': 22258688, 'steps': 43473, 'loss/train': 1.5640021562576294} 02/26/2022 09:52:43 - INFO - codeparrot_training - Step 43474: {'lr': 2.245994886891442e-05, 'samples': 22259200, 'steps': 43474, 'loss/train': 1.9226869344711304} 02/26/2022 09:52:49 - INFO - codeparrot_training - Step 43475: {'lr': 2.2453171101045584e-05, 'samples': 22259712, 'steps': 43475, 'loss/train': 2.1402430534362793} 02/26/2022 09:52:52 - INFO - codeparrot_training - Step 43476: {'lr': 2.244639430791509e-05, 'samples': 22260224, 'steps': 43476, 'loss/train': 1.4616705179214478} 02/26/2022 09:52:58 - INFO - codeparrot_training - Step 43477: {'lr': 2.243961848955195e-05, 'samples': 22260736, 'steps': 43477, 'loss/train': 0.03049241192638874} 02/26/2022 09:53:02 - INFO - codeparrot_training - Step 43478: {'lr': 2.2432843645985273e-05, 'samples': 22261248, 'steps': 43478, 'loss/train': 0.6343022584915161} 02/26/2022 09:53:07 - INFO - codeparrot_training - Step 43479: {'lr': 2.242606977724404e-05, 'samples': 22261760, 'steps': 43479, 'loss/train': 3.0201783180236816} 02/26/2022 09:53:11 - INFO - codeparrot_training - Step 43480: {'lr': 2.2419296883357254e-05, 'samples': 22262272, 'steps': 43480, 'loss/train': 0.8773584365844727} 02/26/2022 09:53:16 - INFO - codeparrot_training - Step 43481: {'lr': 2.2412524964353948e-05, 'samples': 22262784, 'steps': 43481, 'loss/train': 1.2619014978408813} 02/26/2022 09:53:20 - INFO - codeparrot_training - Step 43482: {'lr': 2.2405754020263043e-05, 'samples': 22263296, 'steps': 43482, 'loss/train': 1.7896931171417236} 02/26/2022 09:53:25 - INFO - codeparrot_training - Step 43483: {'lr': 2.2398984051113708e-05, 'samples': 22263808, 'steps': 43483, 'loss/train': 1.3970919847488403} 02/26/2022 09:53:29 - INFO - codeparrot_training - Step 43484: {'lr': 2.239221505693481e-05, 'samples': 22264320, 'steps': 43484, 'loss/train': 1.3139902353286743} 02/26/2022 09:53:34 - INFO - codeparrot_training - Step 43485: {'lr': 2.2385447037755412e-05, 'samples': 22264832, 'steps': 43485, 'loss/train': 0.5896218419075012} 02/26/2022 09:53:38 - INFO - codeparrot_training - Step 43486: {'lr': 2.2378679993604435e-05, 'samples': 22265344, 'steps': 43486, 'loss/train': 2.613023281097412} 02/26/2022 09:53:44 - INFO - codeparrot_training - Step 43487: {'lr': 2.237191392451096e-05, 'samples': 22265856, 'steps': 43487, 'loss/train': 2.0351693630218506} 02/26/2022 09:53:48 - INFO - codeparrot_training - Step 43488: {'lr': 2.2365148830503918e-05, 'samples': 22266368, 'steps': 43488, 'loss/train': 2.0091772079467773} 02/26/2022 09:53:51 - INFO - codeparrot_training - Step 43489: {'lr': 2.2358384711612307e-05, 'samples': 22266880, 'steps': 43489, 'loss/train': 1.3579537868499756} 02/26/2022 09:53:57 - INFO - codeparrot_training - Step 43490: {'lr': 2.2351621567865022e-05, 'samples': 22267392, 'steps': 43490, 'loss/train': 1.9796391725540161} 02/26/2022 09:54:01 - INFO - codeparrot_training - Step 43491: {'lr': 2.234485939929118e-05, 'samples': 22267904, 'steps': 43491, 'loss/train': 0.026413578540086746} 02/26/2022 09:54:06 - INFO - codeparrot_training - Step 43492: {'lr': 2.2338098205919615e-05, 'samples': 22268416, 'steps': 43492, 'loss/train': 0.3130205273628235} 02/26/2022 09:54:10 - INFO - codeparrot_training - Step 43493: {'lr': 2.2331337987779447e-05, 'samples': 22268928, 'steps': 43493, 'loss/train': 1.8519543409347534} 02/26/2022 09:54:15 - INFO - codeparrot_training - Step 43494: {'lr': 2.2324578744899432e-05, 'samples': 22269440, 'steps': 43494, 'loss/train': 1.9269509315490723} 02/26/2022 09:54:19 - INFO - codeparrot_training - Step 43495: {'lr': 2.2317820477308653e-05, 'samples': 22269952, 'steps': 43495, 'loss/train': 2.3492367267608643} 02/26/2022 09:54:24 - INFO - codeparrot_training - Step 43496: {'lr': 2.2311063185036007e-05, 'samples': 22270464, 'steps': 43496, 'loss/train': 1.2519398927688599} 02/26/2022 09:54:28 - INFO - codeparrot_training - Step 43497: {'lr': 2.230430686811058e-05, 'samples': 22270976, 'steps': 43497, 'loss/train': 2.450395107269287} 02/26/2022 09:54:35 - INFO - codeparrot_training - Step 43498: {'lr': 2.229755152656107e-05, 'samples': 22271488, 'steps': 43498, 'loss/train': 0.159735769033432} 02/26/2022 09:54:38 - INFO - codeparrot_training - Step 43499: {'lr': 2.2290797160416628e-05, 'samples': 22272000, 'steps': 43499, 'loss/train': 0.40374743938446045} 02/26/2022 09:54:44 - INFO - codeparrot_training - Step 43500: {'lr': 2.2284043769706025e-05, 'samples': 22272512, 'steps': 43500, 'loss/train': 1.3648148775100708} 02/26/2022 09:54:47 - INFO - codeparrot_training - Step 43501: {'lr': 2.2277291354458385e-05, 'samples': 22273024, 'steps': 43501, 'loss/train': 1.0472849607467651} 02/26/2022 09:54:53 - INFO - codeparrot_training - Step 43502: {'lr': 2.2270539914702433e-05, 'samples': 22273536, 'steps': 43502, 'loss/train': 1.6023541688919067} 02/26/2022 09:54:56 - INFO - codeparrot_training - Step 43503: {'lr': 2.2263789450467197e-05, 'samples': 22274048, 'steps': 43503, 'loss/train': 2.093118667602539} 02/26/2022 09:55:02 - INFO - codeparrot_training - Step 43504: {'lr': 2.225703996178155e-05, 'samples': 22274560, 'steps': 43504, 'loss/train': 0.5411143898963928} 02/26/2022 09:55:05 - INFO - codeparrot_training - Step 43505: {'lr': 2.2250291448674438e-05, 'samples': 22275072, 'steps': 43505, 'loss/train': 2.3279290199279785} 02/26/2022 09:55:11 - INFO - codeparrot_training - Step 43506: {'lr': 2.2243543911174808e-05, 'samples': 22275584, 'steps': 43506, 'loss/train': 1.6749930381774902} 02/26/2022 09:55:14 - INFO - codeparrot_training - Step 43507: {'lr': 2.2236797349311473e-05, 'samples': 22276096, 'steps': 43507, 'loss/train': 1.1816524267196655} 02/26/2022 09:55:20 - INFO - codeparrot_training - Step 43508: {'lr': 2.2230051763113353e-05, 'samples': 22276608, 'steps': 43508, 'loss/train': 1.7725175619125366} 02/26/2022 09:55:23 - INFO - codeparrot_training - Step 43509: {'lr': 2.2223307152609394e-05, 'samples': 22277120, 'steps': 43509, 'loss/train': 1.676793098449707} 02/26/2022 09:55:29 - INFO - codeparrot_training - Step 43510: {'lr': 2.2216563517828468e-05, 'samples': 22277632, 'steps': 43510, 'loss/train': 1.7410683631896973} 02/26/2022 09:55:32 - INFO - codeparrot_training - Step 43511: {'lr': 2.2209820858799468e-05, 'samples': 22278144, 'steps': 43511, 'loss/train': 1.521414041519165} 02/26/2022 09:55:38 - INFO - codeparrot_training - Step 43512: {'lr': 2.22030791755512e-05, 'samples': 22278656, 'steps': 43512, 'loss/train': 1.7858301401138306} 02/26/2022 09:55:41 - INFO - codeparrot_training - Step 43513: {'lr': 2.219633846811267e-05, 'samples': 22279168, 'steps': 43513, 'loss/train': 1.272739052772522} 02/26/2022 09:55:48 - INFO - codeparrot_training - Step 43514: {'lr': 2.218959873651269e-05, 'samples': 22279680, 'steps': 43514, 'loss/train': 1.957121729850769} 02/26/2022 09:55:51 - INFO - codeparrot_training - Step 43515: {'lr': 2.2182859980780122e-05, 'samples': 22280192, 'steps': 43515, 'loss/train': 1.0967785120010376} 02/26/2022 09:55:57 - INFO - codeparrot_training - Step 43516: {'lr': 2.2176122200943784e-05, 'samples': 22280704, 'steps': 43516, 'loss/train': 2.4115428924560547} 02/26/2022 09:56:00 - INFO - codeparrot_training - Step 43517: {'lr': 2.2169385397032644e-05, 'samples': 22281216, 'steps': 43517, 'loss/train': 1.3720903396606445} 02/26/2022 09:56:06 - INFO - codeparrot_training - Step 43518: {'lr': 2.2162649569075543e-05, 'samples': 22281728, 'steps': 43518, 'loss/train': 1.6678321361541748} 02/26/2022 09:56:09 - INFO - codeparrot_training - Step 43519: {'lr': 2.2155914717101265e-05, 'samples': 22282240, 'steps': 43519, 'loss/train': 1.8068166971206665} 02/26/2022 09:56:15 - INFO - codeparrot_training - Step 43520: {'lr': 2.214918084113868e-05, 'samples': 22282752, 'steps': 43520, 'loss/train': 2.0317234992980957} 02/26/2022 09:56:18 - INFO - codeparrot_training - Step 43521: {'lr': 2.2142447941216697e-05, 'samples': 22283264, 'steps': 43521, 'loss/train': 2.336134433746338} 02/26/2022 09:56:24 - INFO - codeparrot_training - Step 43522: {'lr': 2.213571601736411e-05, 'samples': 22283776, 'steps': 43522, 'loss/train': 0.7021390199661255} 02/26/2022 09:56:27 - INFO - codeparrot_training - Step 43523: {'lr': 2.2128985069609754e-05, 'samples': 22284288, 'steps': 43523, 'loss/train': 1.630194067955017} 02/26/2022 09:56:33 - INFO - codeparrot_training - Step 43524: {'lr': 2.2122255097982437e-05, 'samples': 22284800, 'steps': 43524, 'loss/train': 0.6716061234474182} 02/26/2022 09:56:37 - INFO - codeparrot_training - Step 43525: {'lr': 2.211552610251105e-05, 'samples': 22285312, 'steps': 43525, 'loss/train': 1.8806359767913818} 02/26/2022 09:56:42 - INFO - codeparrot_training - Step 43526: {'lr': 2.2108798083224384e-05, 'samples': 22285824, 'steps': 43526, 'loss/train': 1.6183557510375977} 02/26/2022 09:56:46 - INFO - codeparrot_training - Step 43527: {'lr': 2.21020710401513e-05, 'samples': 22286336, 'steps': 43527, 'loss/train': 1.3140428066253662} 02/26/2022 09:56:51 - INFO - codeparrot_training - Step 43528: {'lr': 2.2095344973320554e-05, 'samples': 22286848, 'steps': 43528, 'loss/train': 2.096261978149414} 02/26/2022 09:56:55 - INFO - codeparrot_training - Step 43529: {'lr': 2.2088619882760925e-05, 'samples': 22287360, 'steps': 43529, 'loss/train': 2.4248530864715576} 02/26/2022 09:57:00 - INFO - codeparrot_training - Step 43530: {'lr': 2.2081895768501337e-05, 'samples': 22287872, 'steps': 43530, 'loss/train': 0.021893871948122978} 02/26/2022 09:57:04 - INFO - codeparrot_training - Step 43531: {'lr': 2.2075172630570488e-05, 'samples': 22288384, 'steps': 43531, 'loss/train': 1.6067407131195068} 02/26/2022 09:57:09 - INFO - codeparrot_training - Step 43532: {'lr': 2.2068450468997302e-05, 'samples': 22288896, 'steps': 43532, 'loss/train': 0.09487932175397873} 02/26/2022 09:57:15 - INFO - codeparrot_training - Step 43533: {'lr': 2.2061729283810422e-05, 'samples': 22289408, 'steps': 43533, 'loss/train': 1.2157052755355835} 02/26/2022 09:57:19 - INFO - codeparrot_training - Step 43534: {'lr': 2.2055009075038767e-05, 'samples': 22289920, 'steps': 43534, 'loss/train': 2.417781352996826} 02/26/2022 09:57:24 - INFO - codeparrot_training - Step 43535: {'lr': 2.2048289842711013e-05, 'samples': 22290432, 'steps': 43535, 'loss/train': 1.621270775794983} 02/26/2022 09:57:28 - INFO - codeparrot_training - Step 43536: {'lr': 2.20415715868561e-05, 'samples': 22290944, 'steps': 43536, 'loss/train': 0.9649053812026978} 02/26/2022 09:57:33 - INFO - codeparrot_training - Step 43537: {'lr': 2.203485430750263e-05, 'samples': 22291456, 'steps': 43537, 'loss/train': 1.3556554317474365} 02/26/2022 09:57:37 - INFO - codeparrot_training - Step 43538: {'lr': 2.202813800467948e-05, 'samples': 22291968, 'steps': 43538, 'loss/train': 0.5001811385154724} 02/26/2022 09:57:42 - INFO - codeparrot_training - Step 43539: {'lr': 2.2021422678415365e-05, 'samples': 22292480, 'steps': 43539, 'loss/train': 1.3180254697799683} 02/26/2022 09:57:46 - INFO - codeparrot_training - Step 43540: {'lr': 2.201470832873917e-05, 'samples': 22292992, 'steps': 43540, 'loss/train': 2.034135103225708} 02/26/2022 09:57:51 - INFO - codeparrot_training - Step 43541: {'lr': 2.2007994955679482e-05, 'samples': 22293504, 'steps': 43541, 'loss/train': 1.5188831090927124} 02/26/2022 09:57:54 - INFO - codeparrot_training - Step 43542: {'lr': 2.2001282559265202e-05, 'samples': 22294016, 'steps': 43542, 'loss/train': 2.059890031814575} 02/26/2022 09:58:01 - INFO - codeparrot_training - Step 43543: {'lr': 2.1994571139524966e-05, 'samples': 22294528, 'steps': 43543, 'loss/train': 2.3666932582855225} 02/26/2022 09:58:04 - INFO - codeparrot_training - Step 43544: {'lr': 2.1987860696487644e-05, 'samples': 22295040, 'steps': 43544, 'loss/train': 2.2725865840911865} 02/26/2022 09:58:10 - INFO - codeparrot_training - Step 43545: {'lr': 2.1981151230181906e-05, 'samples': 22295552, 'steps': 43545, 'loss/train': 1.6586047410964966} 02/26/2022 09:58:13 - INFO - codeparrot_training - Step 43546: {'lr': 2.1974442740636535e-05, 'samples': 22296064, 'steps': 43546, 'loss/train': 1.552412509918213} 02/26/2022 09:58:19 - INFO - codeparrot_training - Step 43547: {'lr': 2.1967735227880203e-05, 'samples': 22296576, 'steps': 43547, 'loss/train': 2.3389294147491455} 02/26/2022 09:58:22 - INFO - codeparrot_training - Step 43548: {'lr': 2.196102869194172e-05, 'samples': 22297088, 'steps': 43548, 'loss/train': 2.385468006134033} 02/26/2022 09:58:28 - INFO - codeparrot_training - Step 43549: {'lr': 2.1954323132849786e-05, 'samples': 22297600, 'steps': 43549, 'loss/train': 2.4554502964019775} 02/26/2022 09:58:31 - INFO - codeparrot_training - Step 43550: {'lr': 2.1947618550633096e-05, 'samples': 22298112, 'steps': 43550, 'loss/train': 2.7939515113830566} 02/26/2022 09:58:37 - INFO - codeparrot_training - Step 43551: {'lr': 2.1940914945320384e-05, 'samples': 22298624, 'steps': 43551, 'loss/train': 1.6824169158935547} 02/26/2022 09:58:40 - INFO - codeparrot_training - Step 43552: {'lr': 2.19342123169404e-05, 'samples': 22299136, 'steps': 43552, 'loss/train': 0.776712954044342} 02/26/2022 09:58:46 - INFO - codeparrot_training - Step 43553: {'lr': 2.192751066552179e-05, 'samples': 22299648, 'steps': 43553, 'loss/train': 1.4674180746078491} 02/26/2022 09:58:49 - INFO - codeparrot_training - Step 43554: {'lr': 2.1920809991093387e-05, 'samples': 22300160, 'steps': 43554, 'loss/train': 2.2019076347351074} 02/26/2022 09:58:55 - INFO - codeparrot_training - Step 43555: {'lr': 2.1914110293683733e-05, 'samples': 22300672, 'steps': 43555, 'loss/train': 0.9122690558433533} 02/26/2022 09:58:58 - INFO - codeparrot_training - Step 43556: {'lr': 2.190741157332163e-05, 'samples': 22301184, 'steps': 43556, 'loss/train': 2.2633981704711914} 02/26/2022 09:59:02 - INFO - codeparrot_training - Step 43557: {'lr': 2.1900713830035724e-05, 'samples': 22301696, 'steps': 43557, 'loss/train': 1.1077462434768677} 02/26/2022 09:59:08 - INFO - codeparrot_training - Step 43558: {'lr': 2.18940170638548e-05, 'samples': 22302208, 'steps': 43558, 'loss/train': 1.3015111684799194} 02/26/2022 09:59:11 - INFO - codeparrot_training - Step 43559: {'lr': 2.1887321274807416e-05, 'samples': 22302720, 'steps': 43559, 'loss/train': 0.40784981846809387} 02/26/2022 09:59:17 - INFO - codeparrot_training - Step 43560: {'lr': 2.1880626462922326e-05, 'samples': 22303232, 'steps': 43560, 'loss/train': 1.8171428442001343} 02/26/2022 09:59:21 - INFO - codeparrot_training - Step 43561: {'lr': 2.1873932628228204e-05, 'samples': 22303744, 'steps': 43561, 'loss/train': 2.3191559314727783} 02/26/2022 09:59:26 - INFO - codeparrot_training - Step 43562: {'lr': 2.1867239770753723e-05, 'samples': 22304256, 'steps': 43562, 'loss/train': 1.5726062059402466} 02/26/2022 09:59:30 - INFO - codeparrot_training - Step 43563: {'lr': 2.186054789052752e-05, 'samples': 22304768, 'steps': 43563, 'loss/train': 0.5699591636657715} 02/26/2022 09:59:36 - INFO - codeparrot_training - Step 43564: {'lr': 2.1853856987578302e-05, 'samples': 22305280, 'steps': 43564, 'loss/train': 1.682510256767273} 02/26/2022 09:59:39 - INFO - codeparrot_training - Step 43565: {'lr': 2.1847167061934737e-05, 'samples': 22305792, 'steps': 43565, 'loss/train': 1.5512629747390747} 02/26/2022 09:59:45 - INFO - codeparrot_training - Step 43566: {'lr': 2.1840478113625413e-05, 'samples': 22306304, 'steps': 43566, 'loss/train': 1.0707979202270508} 02/26/2022 09:59:48 - INFO - codeparrot_training - Step 43567: {'lr': 2.1833790142679055e-05, 'samples': 22306816, 'steps': 43567, 'loss/train': 1.5064936876296997} 02/26/2022 09:59:54 - INFO - codeparrot_training - Step 43568: {'lr': 2.1827103149124312e-05, 'samples': 22307328, 'steps': 43568, 'loss/train': 0.9083361625671387} 02/26/2022 09:59:57 - INFO - codeparrot_training - Step 43569: {'lr': 2.1820417132989795e-05, 'samples': 22307840, 'steps': 43569, 'loss/train': 1.3410167694091797} 02/26/2022 10:00:03 - INFO - codeparrot_training - Step 43570: {'lr': 2.1813732094304124e-05, 'samples': 22308352, 'steps': 43570, 'loss/train': 2.022702217102051} 02/26/2022 10:00:07 - INFO - codeparrot_training - Step 43571: {'lr': 2.1807048033095994e-05, 'samples': 22308864, 'steps': 43571, 'loss/train': 2.2553858757019043} 02/26/2022 10:00:12 - INFO - codeparrot_training - Step 43572: {'lr': 2.1800364949394027e-05, 'samples': 22309376, 'steps': 43572, 'loss/train': 1.5829468965530396} 02/26/2022 10:00:16 - INFO - codeparrot_training - Step 43573: {'lr': 2.179368284322683e-05, 'samples': 22309888, 'steps': 43573, 'loss/train': 1.0391923189163208} 02/26/2022 10:00:22 - INFO - codeparrot_training - Step 43574: {'lr': 2.1787001714622974e-05, 'samples': 22310400, 'steps': 43574, 'loss/train': 2.8245155811309814} 02/26/2022 10:00:25 - INFO - codeparrot_training - Step 43575: {'lr': 2.1780321563611233e-05, 'samples': 22310912, 'steps': 43575, 'loss/train': 2.0118017196655273} 02/26/2022 10:00:31 - INFO - codeparrot_training - Step 43576: {'lr': 2.1773642390220062e-05, 'samples': 22311424, 'steps': 43576, 'loss/train': 1.3278862237930298} 02/26/2022 10:00:34 - INFO - codeparrot_training - Step 43577: {'lr': 2.1766964194478157e-05, 'samples': 22311936, 'steps': 43577, 'loss/train': 1.2484737634658813} 02/26/2022 10:00:40 - INFO - codeparrot_training - Step 43578: {'lr': 2.1760286976414053e-05, 'samples': 22312448, 'steps': 43578, 'loss/train': 2.48964786529541} 02/26/2022 10:00:43 - INFO - codeparrot_training - Step 43579: {'lr': 2.175361073605647e-05, 'samples': 22312960, 'steps': 43579, 'loss/train': 0.3831650912761688} 02/26/2022 10:00:49 - INFO - codeparrot_training - Step 43580: {'lr': 2.1746935473433927e-05, 'samples': 22313472, 'steps': 43580, 'loss/train': 1.5037912130355835} 02/26/2022 10:00:53 - INFO - codeparrot_training - Step 43581: {'lr': 2.1740261188575056e-05, 'samples': 22313984, 'steps': 43581, 'loss/train': 1.6030086278915405} 02/26/2022 10:00:59 - INFO - codeparrot_training - Step 43582: {'lr': 2.1733587881508366e-05, 'samples': 22314496, 'steps': 43582, 'loss/train': 1.0923399925231934} 02/26/2022 10:01:02 - INFO - codeparrot_training - Step 43583: {'lr': 2.1726915552262556e-05, 'samples': 22315008, 'steps': 43583, 'loss/train': 2.1169118881225586} 02/26/2022 10:01:08 - INFO - codeparrot_training - Step 43584: {'lr': 2.1720244200866128e-05, 'samples': 22315520, 'steps': 43584, 'loss/train': 2.002955436706543} 02/26/2022 10:01:11 - INFO - codeparrot_training - Step 43585: {'lr': 2.1713573827347706e-05, 'samples': 22316032, 'steps': 43585, 'loss/train': 1.750496745109558} 02/26/2022 10:01:17 - INFO - codeparrot_training - Step 43586: {'lr': 2.1706904431735812e-05, 'samples': 22316544, 'steps': 43586, 'loss/train': 1.5691057443618774} 02/26/2022 10:01:20 - INFO - codeparrot_training - Step 43587: {'lr': 2.170023601405907e-05, 'samples': 22317056, 'steps': 43587, 'loss/train': 1.8770387172698975} 02/26/2022 10:01:26 - INFO - codeparrot_training - Step 43588: {'lr': 2.169356857434604e-05, 'samples': 22317568, 'steps': 43588, 'loss/train': 2.1737616062164307} 02/26/2022 10:01:29 - INFO - codeparrot_training - Step 43589: {'lr': 2.168690211262525e-05, 'samples': 22318080, 'steps': 43589, 'loss/train': 1.8542524576187134} 02/26/2022 10:01:35 - INFO - codeparrot_training - Step 43590: {'lr': 2.1680236628925238e-05, 'samples': 22318592, 'steps': 43590, 'loss/train': 1.8732661008834839} 02/26/2022 10:01:39 - INFO - codeparrot_training - Step 43591: {'lr': 2.1673572123274643e-05, 'samples': 22319104, 'steps': 43591, 'loss/train': 2.0041987895965576} 02/26/2022 10:01:45 - INFO - codeparrot_training - Step 43592: {'lr': 2.1666908595701917e-05, 'samples': 22319616, 'steps': 43592, 'loss/train': 1.4893574714660645} 02/26/2022 10:01:48 - INFO - codeparrot_training - Step 43593: {'lr': 2.1660246046235733e-05, 'samples': 22320128, 'steps': 43593, 'loss/train': 0.6406717896461487} 02/26/2022 10:01:52 - INFO - codeparrot_training - Step 43594: {'lr': 2.165358447490448e-05, 'samples': 22320640, 'steps': 43594, 'loss/train': 0.5837734341621399} 02/26/2022 10:01:57 - INFO - codeparrot_training - Step 43595: {'lr': 2.164692388173678e-05, 'samples': 22321152, 'steps': 43595, 'loss/train': 2.337806224822998} 02/26/2022 10:02:03 - INFO - codeparrot_training - Step 43596: {'lr': 2.1640264266761106e-05, 'samples': 22321664, 'steps': 43596, 'loss/train': 3.754213571548462} 02/26/2022 10:02:06 - INFO - codeparrot_training - Step 43597: {'lr': 2.163360563000613e-05, 'samples': 22322176, 'steps': 43597, 'loss/train': 1.6627697944641113} 02/26/2022 10:02:10 - INFO - codeparrot_training - Step 43598: {'lr': 2.1626947971500193e-05, 'samples': 22322688, 'steps': 43598, 'loss/train': 0.7337480783462524} 02/26/2022 10:02:15 - INFO - codeparrot_training - Step 43599: {'lr': 2.162029129127191e-05, 'samples': 22323200, 'steps': 43599, 'loss/train': 2.1815218925476074} 02/26/2022 10:02:19 - INFO - codeparrot_training - Step 43600: {'lr': 2.1613635589349755e-05, 'samples': 22323712, 'steps': 43600, 'loss/train': 0.5456998348236084} 02/26/2022 10:02:24 - INFO - codeparrot_training - Step 43601: {'lr': 2.1606980865762293e-05, 'samples': 22324224, 'steps': 43601, 'loss/train': 1.6283646821975708} 02/26/2022 10:02:28 - INFO - codeparrot_training - Step 43602: {'lr': 2.1600327120538028e-05, 'samples': 22324736, 'steps': 43602, 'loss/train': 1.1674281358718872} 02/26/2022 10:02:33 - INFO - codeparrot_training - Step 43603: {'lr': 2.1593674353705433e-05, 'samples': 22325248, 'steps': 43603, 'loss/train': 1.5113232135772705} 02/26/2022 10:02:37 - INFO - codeparrot_training - Step 43604: {'lr': 2.1587022565292935e-05, 'samples': 22325760, 'steps': 43604, 'loss/train': 1.2222647666931152} 02/26/2022 10:02:42 - INFO - codeparrot_training - Step 43605: {'lr': 2.1580371755329174e-05, 'samples': 22326272, 'steps': 43605, 'loss/train': 2.115407705307007} 02/26/2022 10:02:48 - INFO - codeparrot_training - Step 43606: {'lr': 2.1573721923842577e-05, 'samples': 22326784, 'steps': 43606, 'loss/train': 1.6032240390777588} 02/26/2022 10:02:51 - INFO - codeparrot_training - Step 43607: {'lr': 2.1567073070861585e-05, 'samples': 22327296, 'steps': 43607, 'loss/train': 0.5503872632980347} 02/26/2022 10:02:58 - INFO - codeparrot_training - Step 43608: {'lr': 2.156042519641474e-05, 'samples': 22327808, 'steps': 43608, 'loss/train': 1.4683725833892822} 02/26/2022 10:03:01 - INFO - codeparrot_training - Step 43609: {'lr': 2.1553778300530458e-05, 'samples': 22328320, 'steps': 43609, 'loss/train': 1.0390137434005737} 02/26/2022 10:03:07 - INFO - codeparrot_training - Step 43610: {'lr': 2.1547132383237272e-05, 'samples': 22328832, 'steps': 43610, 'loss/train': 1.2063630819320679} 02/26/2022 10:03:10 - INFO - codeparrot_training - Step 43611: {'lr': 2.1540487444563663e-05, 'samples': 22329344, 'steps': 43611, 'loss/train': 2.2851593494415283} 02/26/2022 10:03:14 - INFO - codeparrot_training - Step 43612: {'lr': 2.1533843484538047e-05, 'samples': 22329856, 'steps': 43612, 'loss/train': 2.853872060775757} 02/26/2022 10:03:19 - INFO - codeparrot_training - Step 43613: {'lr': 2.152720050318885e-05, 'samples': 22330368, 'steps': 43613, 'loss/train': 2.212364435195923} 02/26/2022 10:03:23 - INFO - codeparrot_training - Step 43614: {'lr': 2.15205585005446e-05, 'samples': 22330880, 'steps': 43614, 'loss/train': 0.7002944946289062} 02/26/2022 10:03:28 - INFO - codeparrot_training - Step 43615: {'lr': 2.1513917476633755e-05, 'samples': 22331392, 'steps': 43615, 'loss/train': 0.7644915580749512} 02/26/2022 10:03:32 - INFO - codeparrot_training - Step 43616: {'lr': 2.150727743148473e-05, 'samples': 22331904, 'steps': 43616, 'loss/train': 1.8617172241210938} 02/26/2022 10:03:38 - INFO - codeparrot_training - Step 43617: {'lr': 2.1500638365125918e-05, 'samples': 22332416, 'steps': 43617, 'loss/train': 1.6511237621307373} 02/26/2022 10:03:41 - INFO - codeparrot_training - Step 43618: {'lr': 2.1494000277585885e-05, 'samples': 22332928, 'steps': 43618, 'loss/train': 0.6012080311775208} 02/26/2022 10:03:48 - INFO - codeparrot_training - Step 43619: {'lr': 2.1487363168892965e-05, 'samples': 22333440, 'steps': 43619, 'loss/train': 2.4701123237609863} 02/26/2022 10:03:51 - INFO - codeparrot_training - Step 43620: {'lr': 2.148072703907564e-05, 'samples': 22333952, 'steps': 43620, 'loss/train': 1.5888080596923828} 02/26/2022 10:03:56 - INFO - codeparrot_training - Step 43621: {'lr': 2.1474091888162265e-05, 'samples': 22334464, 'steps': 43621, 'loss/train': 1.536924123764038} 02/26/2022 10:04:00 - INFO - codeparrot_training - Step 43622: {'lr': 2.146745771618136e-05, 'samples': 22334976, 'steps': 43622, 'loss/train': 1.5509012937545776} 02/26/2022 10:04:05 - INFO - codeparrot_training - Step 43623: {'lr': 2.146082452316131e-05, 'samples': 22335488, 'steps': 43623, 'loss/train': 1.5179985761642456} 02/26/2022 10:04:09 - INFO - codeparrot_training - Step 43624: {'lr': 2.145419230913051e-05, 'samples': 22336000, 'steps': 43624, 'loss/train': 1.7410173416137695} 02/26/2022 10:04:15 - INFO - codeparrot_training - Step 43625: {'lr': 2.144756107411733e-05, 'samples': 22336512, 'steps': 43625, 'loss/train': 1.8667746782302856} 02/26/2022 10:04:18 - INFO - codeparrot_training - Step 43626: {'lr': 2.1440930818150295e-05, 'samples': 22337024, 'steps': 43626, 'loss/train': 1.4250669479370117} 02/26/2022 10:04:24 - INFO - codeparrot_training - Step 43627: {'lr': 2.1434301541257667e-05, 'samples': 22337536, 'steps': 43627, 'loss/train': 1.4384822845458984} 02/26/2022 10:04:27 - INFO - codeparrot_training - Step 43628: {'lr': 2.1427673243468004e-05, 'samples': 22338048, 'steps': 43628, 'loss/train': 1.5117721557617188} 02/26/2022 10:04:33 - INFO - codeparrot_training - Step 43629: {'lr': 2.1421045924809508e-05, 'samples': 22338560, 'steps': 43629, 'loss/train': 1.7281244993209839} 02/26/2022 10:04:37 - INFO - codeparrot_training - Step 43630: {'lr': 2.1414419585310735e-05, 'samples': 22339072, 'steps': 43630, 'loss/train': 1.583446741104126} 02/26/2022 10:04:42 - INFO - codeparrot_training - Step 43631: {'lr': 2.1407794224999942e-05, 'samples': 22339584, 'steps': 43631, 'loss/train': 1.3835880756378174} 02/26/2022 10:04:46 - INFO - codeparrot_training - Step 43632: {'lr': 2.1401169843905694e-05, 'samples': 22340096, 'steps': 43632, 'loss/train': 1.801163911819458} 02/26/2022 10:04:51 - INFO - codeparrot_training - Step 43633: {'lr': 2.139454644205613e-05, 'samples': 22340608, 'steps': 43633, 'loss/train': 2.3483221530914307} 02/26/2022 10:04:55 - INFO - codeparrot_training - Step 43634: {'lr': 2.138792401947978e-05, 'samples': 22341120, 'steps': 43634, 'loss/train': 2.1796042919158936} 02/26/2022 10:05:00 - INFO - codeparrot_training - Step 43635: {'lr': 2.1381302576204937e-05, 'samples': 22341632, 'steps': 43635, 'loss/train': 1.7966972589492798} 02/26/2022 10:05:04 - INFO - codeparrot_training - Step 43636: {'lr': 2.1374682112260097e-05, 'samples': 22342144, 'steps': 43636, 'loss/train': 1.9575128555297852} 02/26/2022 10:05:09 - INFO - codeparrot_training - Step 43637: {'lr': 2.1368062627673405e-05, 'samples': 22342656, 'steps': 43637, 'loss/train': 2.6049258708953857} 02/26/2022 10:05:13 - INFO - codeparrot_training - Step 43638: {'lr': 2.13614441224734e-05, 'samples': 22343168, 'steps': 43638, 'loss/train': 1.7215691804885864} 02/26/2022 10:05:19 - INFO - codeparrot_training - Step 43639: {'lr': 2.135482659668833e-05, 'samples': 22343680, 'steps': 43639, 'loss/train': 1.0972094535827637} 02/26/2022 10:05:22 - INFO - codeparrot_training - Step 43640: {'lr': 2.1348210050346596e-05, 'samples': 22344192, 'steps': 43640, 'loss/train': 1.209674596786499} 02/26/2022 10:05:28 - INFO - codeparrot_training - Step 43641: {'lr': 2.1341594483476556e-05, 'samples': 22344704, 'steps': 43641, 'loss/train': 1.6548192501068115} 02/26/2022 10:05:31 - INFO - codeparrot_training - Step 43642: {'lr': 2.13349798961065e-05, 'samples': 22345216, 'steps': 43642, 'loss/train': 1.3497061729431152} 02/26/2022 10:05:37 - INFO - codeparrot_training - Step 43643: {'lr': 2.1328366288264734e-05, 'samples': 22345728, 'steps': 43643, 'loss/train': 1.309401273727417} 02/26/2022 10:05:40 - INFO - codeparrot_training - Step 43644: {'lr': 2.132175365997968e-05, 'samples': 22346240, 'steps': 43644, 'loss/train': 1.948535442352295} 02/26/2022 10:05:46 - INFO - codeparrot_training - Step 43645: {'lr': 2.1315142011279626e-05, 'samples': 22346752, 'steps': 43645, 'loss/train': 1.441284418106079} 02/26/2022 10:05:49 - INFO - codeparrot_training - Step 43646: {'lr': 2.1308531342192904e-05, 'samples': 22347264, 'steps': 43646, 'loss/train': 1.3826335668563843} 02/26/2022 10:05:55 - INFO - codeparrot_training - Step 43647: {'lr': 2.1301921652747747e-05, 'samples': 22347776, 'steps': 43647, 'loss/train': 1.5711207389831543} 02/26/2022 10:05:58 - INFO - codeparrot_training - Step 43648: {'lr': 2.12953129429726e-05, 'samples': 22348288, 'steps': 43648, 'loss/train': 2.4665095806121826} 02/26/2022 10:06:04 - INFO - codeparrot_training - Step 43649: {'lr': 2.1288705212895692e-05, 'samples': 22348800, 'steps': 43649, 'loss/train': 2.434887170791626} 02/26/2022 10:06:07 - INFO - codeparrot_training - Step 43650: {'lr': 2.1282098462545362e-05, 'samples': 22349312, 'steps': 43650, 'loss/train': 1.5098457336425781} 02/26/2022 10:06:13 - INFO - codeparrot_training - Step 43651: {'lr': 2.127549269194984e-05, 'samples': 22349824, 'steps': 43651, 'loss/train': 1.7913058996200562} 02/26/2022 10:06:16 - INFO - codeparrot_training - Step 43652: {'lr': 2.126888790113754e-05, 'samples': 22350336, 'steps': 43652, 'loss/train': 0.6994002461433411} 02/26/2022 10:06:22 - INFO - codeparrot_training - Step 43653: {'lr': 2.126228409013667e-05, 'samples': 22350848, 'steps': 43653, 'loss/train': 1.750333309173584} 02/26/2022 10:06:26 - INFO - codeparrot_training - Step 43654: {'lr': 2.1255681258975562e-05, 'samples': 22351360, 'steps': 43654, 'loss/train': 1.9532947540283203} 02/26/2022 10:06:31 - INFO - codeparrot_training - Step 43655: {'lr': 2.1249079407682476e-05, 'samples': 22351872, 'steps': 43655, 'loss/train': 1.5574061870574951} 02/26/2022 10:06:35 - INFO - codeparrot_training - Step 43656: {'lr': 2.1242478536285638e-05, 'samples': 22352384, 'steps': 43656, 'loss/train': 2.4159815311431885} 02/26/2022 10:06:40 - INFO - codeparrot_training - Step 43657: {'lr': 2.1235878644813437e-05, 'samples': 22352896, 'steps': 43657, 'loss/train': 1.7896896600723267} 02/26/2022 10:06:44 - INFO - codeparrot_training - Step 43658: {'lr': 2.122927973329411e-05, 'samples': 22353408, 'steps': 43658, 'loss/train': 2.0833568572998047} 02/26/2022 10:06:50 - INFO - codeparrot_training - Step 43659: {'lr': 2.122268180175588e-05, 'samples': 22353920, 'steps': 43659, 'loss/train': 2.0339841842651367} 02/26/2022 10:06:53 - INFO - codeparrot_training - Step 43660: {'lr': 2.1216084850226998e-05, 'samples': 22354432, 'steps': 43660, 'loss/train': 1.7435036897659302} 02/26/2022 10:06:58 - INFO - codeparrot_training - Step 43661: {'lr': 2.1209488878735833e-05, 'samples': 22354944, 'steps': 43661, 'loss/train': 2.865453004837036} 02/26/2022 10:07:02 - INFO - codeparrot_training - Step 43662: {'lr': 2.1202893887310503e-05, 'samples': 22355456, 'steps': 43662, 'loss/train': 1.6757209300994873} 02/26/2022 10:07:08 - INFO - codeparrot_training - Step 43663: {'lr': 2.1196299875979403e-05, 'samples': 22355968, 'steps': 43663, 'loss/train': 1.7064749002456665} 02/26/2022 10:07:11 - INFO - codeparrot_training - Step 43664: {'lr': 2.1189706844770618e-05, 'samples': 22356480, 'steps': 43664, 'loss/train': 1.3960672616958618} 02/26/2022 10:07:18 - INFO - codeparrot_training - Step 43665: {'lr': 2.1183114793712517e-05, 'samples': 22356992, 'steps': 43665, 'loss/train': 1.3924715518951416} 02/26/2022 10:07:21 - INFO - codeparrot_training - Step 43666: {'lr': 2.1176523722833268e-05, 'samples': 22357504, 'steps': 43666, 'loss/train': 1.9153714179992676} 02/26/2022 10:07:27 - INFO - codeparrot_training - Step 43667: {'lr': 2.1169933632161186e-05, 'samples': 22358016, 'steps': 43667, 'loss/train': 2.349346876144409} 02/26/2022 10:07:30 - INFO - codeparrot_training - Step 43668: {'lr': 2.1163344521724385e-05, 'samples': 22358528, 'steps': 43668, 'loss/train': 2.1138269901275635} 02/26/2022 10:07:36 - INFO - codeparrot_training - Step 43669: {'lr': 2.1156756391551178e-05, 'samples': 22359040, 'steps': 43669, 'loss/train': 1.7477502822875977} 02/26/2022 10:07:39 - INFO - codeparrot_training - Step 43670: {'lr': 2.1150169241669736e-05, 'samples': 22359552, 'steps': 43670, 'loss/train': 2.5730812549591064} 02/26/2022 10:07:45 - INFO - codeparrot_training - Step 43671: {'lr': 2.1143583072108398e-05, 'samples': 22360064, 'steps': 43671, 'loss/train': 0.9646165370941162} 02/26/2022 10:07:48 - INFO - codeparrot_training - Step 43672: {'lr': 2.1136997882895166e-05, 'samples': 22360576, 'steps': 43672, 'loss/train': 1.162533164024353} 02/26/2022 10:07:54 - INFO - codeparrot_training - Step 43673: {'lr': 2.1130413674058407e-05, 'samples': 22361088, 'steps': 43673, 'loss/train': 1.509287714958191} 02/26/2022 10:07:57 - INFO - codeparrot_training - Step 43674: {'lr': 2.112383044562624e-05, 'samples': 22361600, 'steps': 43674, 'loss/train': 1.2716783285140991} 02/26/2022 10:08:04 - INFO - codeparrot_training - Step 43675: {'lr': 2.1117248197626976e-05, 'samples': 22362112, 'steps': 43675, 'loss/train': 2.0076851844787598} 02/26/2022 10:08:07 - INFO - codeparrot_training - Step 43676: {'lr': 2.111066693008873e-05, 'samples': 22362624, 'steps': 43676, 'loss/train': 1.412522315979004} 02/26/2022 10:08:13 - INFO - codeparrot_training - Step 43677: {'lr': 2.11040866430397e-05, 'samples': 22363136, 'steps': 43677, 'loss/train': 0.7914162874221802} 02/26/2022 10:08:16 - INFO - codeparrot_training - Step 43678: {'lr': 2.109750733650803e-05, 'samples': 22363648, 'steps': 43678, 'loss/train': 1.700441837310791} 02/26/2022 10:08:22 - INFO - codeparrot_training - Step 43679: {'lr': 2.109092901052201e-05, 'samples': 22364160, 'steps': 43679, 'loss/train': 1.3637765645980835} 02/26/2022 10:08:25 - INFO - codeparrot_training - Step 43680: {'lr': 2.1084351665109745e-05, 'samples': 22364672, 'steps': 43680, 'loss/train': 1.7694653272628784} 02/26/2022 10:08:31 - INFO - codeparrot_training - Step 43681: {'lr': 2.1077775300299446e-05, 'samples': 22365184, 'steps': 43681, 'loss/train': 1.0543186664581299} 02/26/2022 10:08:34 - INFO - codeparrot_training - Step 43682: {'lr': 2.107119991611922e-05, 'samples': 22365696, 'steps': 43682, 'loss/train': 3.1474647521972656} 02/26/2022 10:08:40 - INFO - codeparrot_training - Step 43683: {'lr': 2.106462551259733e-05, 'samples': 22366208, 'steps': 43683, 'loss/train': 2.513465166091919} 02/26/2022 10:08:43 - INFO - codeparrot_training - Step 43684: {'lr': 2.105805208976186e-05, 'samples': 22366720, 'steps': 43684, 'loss/train': 1.6875081062316895} 02/26/2022 10:08:49 - INFO - codeparrot_training - Step 43685: {'lr': 2.1051479647641035e-05, 'samples': 22367232, 'steps': 43685, 'loss/train': 1.676844596862793} 02/26/2022 10:08:53 - INFO - codeparrot_training - Step 43686: {'lr': 2.1044908186262895e-05, 'samples': 22367744, 'steps': 43686, 'loss/train': 3.0588529109954834} 02/26/2022 10:08:58 - INFO - codeparrot_training - Step 43687: {'lr': 2.1038337705655748e-05, 'samples': 22368256, 'steps': 43687, 'loss/train': 1.7073297500610352} 02/26/2022 10:09:02 - INFO - codeparrot_training - Step 43688: {'lr': 2.1031768205847624e-05, 'samples': 22368768, 'steps': 43688, 'loss/train': 1.9989266395568848} 02/26/2022 10:09:07 - INFO - codeparrot_training - Step 43689: {'lr': 2.1025199686866725e-05, 'samples': 22369280, 'steps': 43689, 'loss/train': 2.059293270111084} 02/26/2022 10:09:11 - INFO - codeparrot_training - Step 43690: {'lr': 2.1018632148741113e-05, 'samples': 22369792, 'steps': 43690, 'loss/train': 1.9883449077606201} 02/26/2022 10:09:16 - INFO - codeparrot_training - Step 43691: {'lr': 2.1012065591499014e-05, 'samples': 22370304, 'steps': 43691, 'loss/train': 0.9303855299949646} 02/26/2022 10:09:20 - INFO - codeparrot_training - Step 43692: {'lr': 2.1005500015168487e-05, 'samples': 22370816, 'steps': 43692, 'loss/train': 1.8114391565322876} 02/26/2022 10:09:25 - INFO - codeparrot_training - Step 43693: {'lr': 2.099893541977771e-05, 'samples': 22371328, 'steps': 43693, 'loss/train': 1.7076948881149292} 02/26/2022 10:09:29 - INFO - codeparrot_training - Step 43694: {'lr': 2.099237180535474e-05, 'samples': 22371840, 'steps': 43694, 'loss/train': 0.3772452473640442} 02/26/2022 10:09:34 - INFO - codeparrot_training - Step 43695: {'lr': 2.0985809171927772e-05, 'samples': 22372352, 'steps': 43695, 'loss/train': 1.763005018234253} 02/26/2022 10:09:40 - INFO - codeparrot_training - Step 43696: {'lr': 2.0979247519524876e-05, 'samples': 22372864, 'steps': 43696, 'loss/train': 1.1091537475585938} 02/26/2022 10:09:43 - INFO - codeparrot_training - Step 43697: {'lr': 2.0972686848174133e-05, 'samples': 22373376, 'steps': 43697, 'loss/train': 2.7304844856262207} 02/26/2022 10:09:47 - INFO - codeparrot_training - Step 43698: {'lr': 2.0966127157903665e-05, 'samples': 22373888, 'steps': 43698, 'loss/train': 1.645606517791748} 02/26/2022 10:09:52 - INFO - codeparrot_training - Step 43699: {'lr': 2.095956844874161e-05, 'samples': 22374400, 'steps': 43699, 'loss/train': 1.3573448657989502} 02/26/2022 10:09:56 - INFO - codeparrot_training - Step 43700: {'lr': 2.0953010720716037e-05, 'samples': 22374912, 'steps': 43700, 'loss/train': 1.6097822189331055} 02/26/2022 10:10:02 - INFO - codeparrot_training - Step 43701: {'lr': 2.0946453973855e-05, 'samples': 22375424, 'steps': 43701, 'loss/train': 1.5295767784118652} 02/26/2022 10:10:06 - INFO - codeparrot_training - Step 43702: {'lr': 2.0939898208186675e-05, 'samples': 22375936, 'steps': 43702, 'loss/train': 1.5707643032073975} 02/26/2022 10:10:12 - INFO - codeparrot_training - Step 43703: {'lr': 2.0933343423739066e-05, 'samples': 22376448, 'steps': 43703, 'loss/train': 1.3826990127563477} 02/26/2022 10:10:15 - INFO - codeparrot_training - Step 43704: {'lr': 2.092678962054029e-05, 'samples': 22376960, 'steps': 43704, 'loss/train': 2.2209463119506836} 02/26/2022 10:10:21 - INFO - codeparrot_training - Step 43705: {'lr': 2.092023679861835e-05, 'samples': 22377472, 'steps': 43705, 'loss/train': 1.704289436340332} 02/26/2022 10:10:24 - INFO - codeparrot_training - Step 43706: {'lr': 2.0913684958001472e-05, 'samples': 22377984, 'steps': 43706, 'loss/train': 1.503014087677002} 02/26/2022 10:10:30 - INFO - codeparrot_training - Step 43707: {'lr': 2.0907134098717555e-05, 'samples': 22378496, 'steps': 43707, 'loss/train': 1.75319242477417} 02/26/2022 10:10:34 - INFO - codeparrot_training - Step 43708: {'lr': 2.0900584220794744e-05, 'samples': 22379008, 'steps': 43708, 'loss/train': 2.362241744995117} 02/26/2022 10:10:39 - INFO - codeparrot_training - Step 43709: {'lr': 2.0894035324261067e-05, 'samples': 22379520, 'steps': 43709, 'loss/train': 2.19525146484375} 02/26/2022 10:10:43 - INFO - codeparrot_training - Step 43710: {'lr': 2.088748740914462e-05, 'samples': 22380032, 'steps': 43710, 'loss/train': 1.2324306964874268} 02/26/2022 10:10:49 - INFO - codeparrot_training - Step 43711: {'lr': 2.0880940475473425e-05, 'samples': 22380544, 'steps': 43711, 'loss/train': 1.3014642000198364} 02/26/2022 10:10:52 - INFO - codeparrot_training - Step 43712: {'lr': 2.0874394523275526e-05, 'samples': 22381056, 'steps': 43712, 'loss/train': 1.6936137676239014} 02/26/2022 10:10:58 - INFO - codeparrot_training - Step 43713: {'lr': 2.086784955257895e-05, 'samples': 22381568, 'steps': 43713, 'loss/train': 1.50624418258667} 02/26/2022 10:11:01 - INFO - codeparrot_training - Step 43714: {'lr': 2.0861305563411758e-05, 'samples': 22382080, 'steps': 43714, 'loss/train': 1.9275946617126465} 02/26/2022 10:11:07 - INFO - codeparrot_training - Step 43715: {'lr': 2.0854762555801988e-05, 'samples': 22382592, 'steps': 43715, 'loss/train': 1.298649549484253} 02/26/2022 10:11:10 - INFO - codeparrot_training - Step 43716: {'lr': 2.084822052977764e-05, 'samples': 22383104, 'steps': 43716, 'loss/train': 1.914510726928711} 02/26/2022 10:11:16 - INFO - codeparrot_training - Step 43717: {'lr': 2.084167948536672e-05, 'samples': 22383616, 'steps': 43717, 'loss/train': 2.3082385063171387} 02/26/2022 10:11:19 - INFO - codeparrot_training - Step 43718: {'lr': 2.083513942259732e-05, 'samples': 22384128, 'steps': 43718, 'loss/train': 2.0877203941345215} 02/26/2022 10:11:25 - INFO - codeparrot_training - Step 43719: {'lr': 2.0828600341497415e-05, 'samples': 22384640, 'steps': 43719, 'loss/train': 2.9534261226654053} 02/26/2022 10:11:28 - INFO - codeparrot_training - Step 43720: {'lr': 2.0822062242095014e-05, 'samples': 22385152, 'steps': 43720, 'loss/train': 0.9044831395149231} 02/26/2022 10:11:34 - INFO - codeparrot_training - Step 43721: {'lr': 2.0815525124418087e-05, 'samples': 22385664, 'steps': 43721, 'loss/train': 2.0298373699188232} 02/26/2022 10:11:38 - INFO - codeparrot_training - Step 43722: {'lr': 2.080898898849473e-05, 'samples': 22386176, 'steps': 43722, 'loss/train': 1.7623164653778076} 02/26/2022 10:11:43 - INFO - codeparrot_training - Step 43723: {'lr': 2.0802453834352834e-05, 'samples': 22386688, 'steps': 43723, 'loss/train': 1.785817265510559} 02/26/2022 10:11:47 - INFO - codeparrot_training - Step 43724: {'lr': 2.0795919662020518e-05, 'samples': 22387200, 'steps': 43724, 'loss/train': 1.8084228038787842} 02/26/2022 10:11:52 - INFO - codeparrot_training - Step 43725: {'lr': 2.0789386471525647e-05, 'samples': 22387712, 'steps': 43725, 'loss/train': 0.9158523082733154} 02/26/2022 10:11:56 - INFO - codeparrot_training - Step 43726: {'lr': 2.0782854262896277e-05, 'samples': 22388224, 'steps': 43726, 'loss/train': 1.596768856048584} 02/26/2022 10:12:01 - INFO - codeparrot_training - Step 43727: {'lr': 2.0776323036160337e-05, 'samples': 22388736, 'steps': 43727, 'loss/train': 1.9071859121322632} 02/26/2022 10:12:05 - INFO - codeparrot_training - Step 43728: {'lr': 2.0769792791345942e-05, 'samples': 22389248, 'steps': 43728, 'loss/train': 1.6885617971420288} 02/26/2022 10:12:10 - INFO - codeparrot_training - Step 43729: {'lr': 2.0763263528480875e-05, 'samples': 22389760, 'steps': 43729, 'loss/train': 1.3484243154525757} 02/26/2022 10:12:14 - INFO - codeparrot_training - Step 43730: {'lr': 2.0756735247593222e-05, 'samples': 22390272, 'steps': 43730, 'loss/train': 1.6374986171722412} 02/26/2022 10:12:20 - INFO - codeparrot_training - Step 43731: {'lr': 2.075020794871091e-05, 'samples': 22390784, 'steps': 43731, 'loss/train': 1.3780603408813477} 02/26/2022 10:12:24 - INFO - codeparrot_training - Step 43732: {'lr': 2.0743681631861965e-05, 'samples': 22391296, 'steps': 43732, 'loss/train': 1.3645790815353394} 02/26/2022 10:12:29 - INFO - codeparrot_training - Step 43733: {'lr': 2.0737156297074233e-05, 'samples': 22391808, 'steps': 43733, 'loss/train': 2.2477939128875732} 02/26/2022 10:12:33 - INFO - codeparrot_training - Step 43734: {'lr': 2.0730631944375745e-05, 'samples': 22392320, 'steps': 43734, 'loss/train': 0.0358416773378849} 02/26/2022 10:12:38 - INFO - codeparrot_training - Step 43735: {'lr': 2.072410857379445e-05, 'samples': 22392832, 'steps': 43735, 'loss/train': 0.8678162693977356} 02/26/2022 10:12:42 - INFO - codeparrot_training - Step 43736: {'lr': 2.071758618535821e-05, 'samples': 22393344, 'steps': 43736, 'loss/train': 1.9455173015594482} 02/26/2022 10:12:47 - INFO - codeparrot_training - Step 43737: {'lr': 2.0711064779095068e-05, 'samples': 22393856, 'steps': 43737, 'loss/train': 2.4433224201202393} 02/26/2022 10:12:51 - INFO - codeparrot_training - Step 43738: {'lr': 2.070454435503294e-05, 'samples': 22394368, 'steps': 43738, 'loss/train': 1.611151933670044} 02/26/2022 10:12:56 - INFO - codeparrot_training - Step 43739: {'lr': 2.069802491319969e-05, 'samples': 22394880, 'steps': 43739, 'loss/train': 1.200068712234497} 02/26/2022 10:13:00 - INFO - codeparrot_training - Step 43740: {'lr': 2.0691506453623277e-05, 'samples': 22395392, 'steps': 43740, 'loss/train': 1.3425055742263794} 02/26/2022 10:13:05 - INFO - codeparrot_training - Step 43741: {'lr': 2.0684988976331696e-05, 'samples': 22395904, 'steps': 43741, 'loss/train': 1.4413142204284668} 02/26/2022 10:13:09 - INFO - codeparrot_training - Step 43742: {'lr': 2.0678472481352793e-05, 'samples': 22396416, 'steps': 43742, 'loss/train': 1.354150652885437} 02/26/2022 10:13:14 - INFO - codeparrot_training - Step 43743: {'lr': 2.067195696871449e-05, 'samples': 22396928, 'steps': 43743, 'loss/train': 2.612377405166626} 02/26/2022 10:13:18 - INFO - codeparrot_training - Step 43744: {'lr': 2.066544243844465e-05, 'samples': 22397440, 'steps': 43744, 'loss/train': 0.7962895035743713} 02/26/2022 10:13:23 - INFO - codeparrot_training - Step 43745: {'lr': 2.0658928890571277e-05, 'samples': 22397952, 'steps': 43745, 'loss/train': 1.8486801385879517} 02/26/2022 10:13:26 - INFO - codeparrot_training - Step 43746: {'lr': 2.0652416325122243e-05, 'samples': 22398464, 'steps': 43746, 'loss/train': 1.0842667818069458} 02/26/2022 10:13:33 - INFO - codeparrot_training - Step 43747: {'lr': 2.064590474212544e-05, 'samples': 22398976, 'steps': 43747, 'loss/train': 0.8362195491790771} 02/26/2022 10:13:36 - INFO - codeparrot_training - Step 43748: {'lr': 2.0639394141608704e-05, 'samples': 22399488, 'steps': 43748, 'loss/train': 1.8332408666610718} 02/26/2022 10:13:42 - INFO - codeparrot_training - Step 43749: {'lr': 2.0632884523599986e-05, 'samples': 22400000, 'steps': 43749, 'loss/train': 2.9326047897338867} 02/26/2022 10:13:47 - INFO - codeparrot_training - Step 43750: {'lr': 2.0626375888127186e-05, 'samples': 22400512, 'steps': 43750, 'loss/train': 0.7321078181266785} 02/26/2022 10:13:51 - INFO - codeparrot_training - Step 43751: {'lr': 2.0619868235218164e-05, 'samples': 22401024, 'steps': 43751, 'loss/train': 2.223775863647461} 02/26/2022 10:13:56 - INFO - codeparrot_training - Step 43752: {'lr': 2.0613361564900735e-05, 'samples': 22401536, 'steps': 43752, 'loss/train': 1.9427775144577026} 02/26/2022 10:14:00 - INFO - codeparrot_training - Step 43753: {'lr': 2.0606855877202873e-05, 'samples': 22402048, 'steps': 43753, 'loss/train': 2.0571823120117188} 02/26/2022 10:14:06 - INFO - codeparrot_training - Step 43754: {'lr': 2.0600351172152392e-05, 'samples': 22402560, 'steps': 43754, 'loss/train': 1.2308913469314575} 02/26/2022 10:14:09 - INFO - codeparrot_training - Step 43755: {'lr': 2.059384744977716e-05, 'samples': 22403072, 'steps': 43755, 'loss/train': 1.4520541429519653} 02/26/2022 10:14:16 - INFO - codeparrot_training - Step 43756: {'lr': 2.0587344710105012e-05, 'samples': 22403584, 'steps': 43756, 'loss/train': 1.9432131052017212} 02/26/2022 10:14:20 - INFO - codeparrot_training - Step 43757: {'lr': 2.058084295316387e-05, 'samples': 22404096, 'steps': 43757, 'loss/train': 1.9516907930374146} 02/26/2022 10:14:25 - INFO - codeparrot_training - Step 43758: {'lr': 2.0574342178981547e-05, 'samples': 22404608, 'steps': 43758, 'loss/train': 2.027433156967163} 02/26/2022 10:14:29 - INFO - codeparrot_training - Step 43759: {'lr': 2.056784238758591e-05, 'samples': 22405120, 'steps': 43759, 'loss/train': 1.9764513969421387} 02/26/2022 10:14:34 - INFO - codeparrot_training - Step 43760: {'lr': 2.0561343579004716e-05, 'samples': 22405632, 'steps': 43760, 'loss/train': 2.2532684803009033} 02/26/2022 10:14:38 - INFO - codeparrot_training - Step 43761: {'lr': 2.055484575326591e-05, 'samples': 22406144, 'steps': 43761, 'loss/train': 0.8994282484054565} 02/26/2022 10:14:43 - INFO - codeparrot_training - Step 43762: {'lr': 2.054834891039728e-05, 'samples': 22406656, 'steps': 43762, 'loss/train': 1.210808277130127} 02/26/2022 10:14:47 - INFO - codeparrot_training - Step 43763: {'lr': 2.0541853050426716e-05, 'samples': 22407168, 'steps': 43763, 'loss/train': 1.6927560567855835} 02/26/2022 10:14:52 - INFO - codeparrot_training - Step 43764: {'lr': 2.0535358173381923e-05, 'samples': 22407680, 'steps': 43764, 'loss/train': 2.2049617767333984} 02/26/2022 10:14:56 - INFO - codeparrot_training - Step 43765: {'lr': 2.052886427929082e-05, 'samples': 22408192, 'steps': 43765, 'loss/train': 2.969558000564575} 02/26/2022 10:15:02 - INFO - codeparrot_training - Step 43766: {'lr': 2.0522371368181166e-05, 'samples': 22408704, 'steps': 43766, 'loss/train': 1.4443888664245605} 02/26/2022 10:15:05 - INFO - codeparrot_training - Step 43767: {'lr': 2.0515879440080905e-05, 'samples': 22409216, 'steps': 43767, 'loss/train': 2.103226661682129} 02/26/2022 10:15:11 - INFO - codeparrot_training - Step 43768: {'lr': 2.050938849501763e-05, 'samples': 22409728, 'steps': 43768, 'loss/train': 1.6351176500320435} 02/26/2022 10:15:15 - INFO - codeparrot_training - Step 43769: {'lr': 2.0502898533019315e-05, 'samples': 22410240, 'steps': 43769, 'loss/train': 1.7441072463989258} 02/26/2022 10:15:20 - INFO - codeparrot_training - Step 43770: {'lr': 2.0496409554113694e-05, 'samples': 22410752, 'steps': 43770, 'loss/train': 1.6881424188613892} 02/26/2022 10:15:23 - INFO - codeparrot_training - Step 43771: {'lr': 2.04899215583286e-05, 'samples': 22411264, 'steps': 43771, 'loss/train': 2.434337615966797} 02/26/2022 10:15:29 - INFO - codeparrot_training - Step 43772: {'lr': 2.0483434545691792e-05, 'samples': 22411776, 'steps': 43772, 'loss/train': 3.2211732864379883} 02/26/2022 10:15:33 - INFO - codeparrot_training - Step 43773: {'lr': 2.0476948516231082e-05, 'samples': 22412288, 'steps': 43773, 'loss/train': 1.474689245223999} 02/26/2022 10:15:38 - INFO - codeparrot_training - Step 43774: {'lr': 2.0470463469974193e-05, 'samples': 22412800, 'steps': 43774, 'loss/train': 2.4542195796966553} 02/26/2022 10:15:42 - INFO - codeparrot_training - Step 43775: {'lr': 2.0463979406949023e-05, 'samples': 22413312, 'steps': 43775, 'loss/train': 1.8150604963302612} 02/26/2022 10:15:47 - INFO - codeparrot_training - Step 43776: {'lr': 2.0457496327183245e-05, 'samples': 22413824, 'steps': 43776, 'loss/train': 1.1934990882873535} 02/26/2022 10:15:51 - INFO - codeparrot_training - Step 43777: {'lr': 2.0451014230704696e-05, 'samples': 22414336, 'steps': 43777, 'loss/train': 2.025015354156494} 02/26/2022 10:15:56 - INFO - codeparrot_training - Step 43778: {'lr': 2.044453311754105e-05, 'samples': 22414848, 'steps': 43778, 'loss/train': 1.4752459526062012} 02/26/2022 10:16:00 - INFO - codeparrot_training - Step 43779: {'lr': 2.043805298772017e-05, 'samples': 22415360, 'steps': 43779, 'loss/train': 2.145583391189575} 02/26/2022 10:16:05 - INFO - codeparrot_training - Step 43780: {'lr': 2.0431573841269812e-05, 'samples': 22415872, 'steps': 43780, 'loss/train': 1.887131690979004} 02/26/2022 10:16:08 - INFO - codeparrot_training - Step 43781: {'lr': 2.0425095678217653e-05, 'samples': 22416384, 'steps': 43781, 'loss/train': 1.146106243133545} 02/26/2022 10:16:15 - INFO - codeparrot_training - Step 43782: {'lr': 2.0418618498591474e-05, 'samples': 22416896, 'steps': 43782, 'loss/train': 2.37729811668396} 02/26/2022 10:16:19 - INFO - codeparrot_training - Step 43783: {'lr': 2.0412142302419055e-05, 'samples': 22417408, 'steps': 43783, 'loss/train': 1.6423615217208862} 02/26/2022 10:16:24 - INFO - codeparrot_training - Step 43784: {'lr': 2.040566708972813e-05, 'samples': 22417920, 'steps': 43784, 'loss/train': 2.652238607406616} 02/26/2022 10:16:28 - INFO - codeparrot_training - Step 43785: {'lr': 2.0399192860546397e-05, 'samples': 22418432, 'steps': 43785, 'loss/train': 2.1094307899475098} 02/26/2022 10:16:33 - INFO - codeparrot_training - Step 43786: {'lr': 2.0392719614901634e-05, 'samples': 22418944, 'steps': 43786, 'loss/train': 1.3287436962127686} 02/26/2022 10:16:37 - INFO - codeparrot_training - Step 43787: {'lr': 2.0386247352821518e-05, 'samples': 22419456, 'steps': 43787, 'loss/train': 0.2299940288066864} 02/26/2022 10:16:42 - INFO - codeparrot_training - Step 43788: {'lr': 2.0379776074333835e-05, 'samples': 22419968, 'steps': 43788, 'loss/train': 2.07014536857605} 02/26/2022 10:16:46 - INFO - codeparrot_training - Step 43789: {'lr': 2.0373305779466305e-05, 'samples': 22420480, 'steps': 43789, 'loss/train': 1.465870976448059} 02/26/2022 10:16:51 - INFO - codeparrot_training - Step 43790: {'lr': 2.036683646824658e-05, 'samples': 22420992, 'steps': 43790, 'loss/train': 1.9039493799209595} 02/26/2022 10:16:55 - INFO - codeparrot_training - Step 43791: {'lr': 2.0360368140702386e-05, 'samples': 22421504, 'steps': 43791, 'loss/train': 1.6106352806091309} 02/26/2022 10:17:02 - INFO - codeparrot_training - Step 43792: {'lr': 2.0353900796861504e-05, 'samples': 22422016, 'steps': 43792, 'loss/train': 2.552624464035034} 02/26/2022 10:17:05 - INFO - codeparrot_training - Step 43793: {'lr': 2.0347434436751605e-05, 'samples': 22422528, 'steps': 43793, 'loss/train': 2.0767154693603516} 02/26/2022 10:17:11 - INFO - codeparrot_training - Step 43794: {'lr': 2.034096906040034e-05, 'samples': 22423040, 'steps': 43794, 'loss/train': 1.2512662410736084} 02/26/2022 10:17:14 - INFO - codeparrot_training - Step 43795: {'lr': 2.0334504667835403e-05, 'samples': 22423552, 'steps': 43795, 'loss/train': 1.1248472929000854} 02/26/2022 10:17:19 - INFO - codeparrot_training - Step 43796: {'lr': 2.0328041259084578e-05, 'samples': 22424064, 'steps': 43796, 'loss/train': 2.4028737545013428} 02/26/2022 10:17:23 - INFO - codeparrot_training - Step 43797: {'lr': 2.0321578834175457e-05, 'samples': 22424576, 'steps': 43797, 'loss/train': 1.6904635429382324} 02/26/2022 10:17:28 - INFO - codeparrot_training - Step 43798: {'lr': 2.0315117393135822e-05, 'samples': 22425088, 'steps': 43798, 'loss/train': 1.0677117109298706} 02/26/2022 10:17:32 - INFO - codeparrot_training - Step 43799: {'lr': 2.0308656935993237e-05, 'samples': 22425600, 'steps': 43799, 'loss/train': 1.1503819227218628} 02/26/2022 10:17:37 - INFO - codeparrot_training - Step 43800: {'lr': 2.030219746277545e-05, 'samples': 22426112, 'steps': 43800, 'loss/train': 1.6439311504364014} 02/26/2022 10:17:41 - INFO - codeparrot_training - Step 43801: {'lr': 2.0295738973510057e-05, 'samples': 22426624, 'steps': 43801, 'loss/train': 1.3753182888031006} 02/26/2022 10:17:46 - INFO - codeparrot_training - Step 43802: {'lr': 2.0289281468224896e-05, 'samples': 22427136, 'steps': 43802, 'loss/train': 2.3771438598632812} 02/26/2022 10:17:50 - INFO - codeparrot_training - Step 43803: {'lr': 2.0282824946947416e-05, 'samples': 22427648, 'steps': 43803, 'loss/train': 1.1201207637786865} 02/26/2022 10:17:56 - INFO - codeparrot_training - Step 43804: {'lr': 2.0276369409705398e-05, 'samples': 22428160, 'steps': 43804, 'loss/train': 2.048480749130249} 02/26/2022 10:18:00 - INFO - codeparrot_training - Step 43805: {'lr': 2.0269914856526434e-05, 'samples': 22428672, 'steps': 43805, 'loss/train': 1.1324502229690552} 02/26/2022 10:18:05 - INFO - codeparrot_training - Step 43806: {'lr': 2.0263461287438306e-05, 'samples': 22429184, 'steps': 43806, 'loss/train': 2.0590100288391113} 02/26/2022 10:18:09 - INFO - codeparrot_training - Step 43807: {'lr': 2.0257008702468467e-05, 'samples': 22429696, 'steps': 43807, 'loss/train': 1.0743722915649414} 02/26/2022 10:18:14 - INFO - codeparrot_training - Step 43808: {'lr': 2.0250557101644697e-05, 'samples': 22430208, 'steps': 43808, 'loss/train': 1.3453279733657837} 02/26/2022 10:18:18 - INFO - codeparrot_training - Step 43809: {'lr': 2.024410648499453e-05, 'samples': 22430720, 'steps': 43809, 'loss/train': 4.067195415496826} 02/26/2022 10:18:23 - INFO - codeparrot_training - Step 43810: {'lr': 2.0237656852545692e-05, 'samples': 22431232, 'steps': 43810, 'loss/train': 2.2540249824523926} 02/26/2022 10:18:27 - INFO - codeparrot_training - Step 43811: {'lr': 2.0231208204325806e-05, 'samples': 22431744, 'steps': 43811, 'loss/train': 0.5296192765235901} 02/26/2022 10:18:32 - INFO - codeparrot_training - Step 43812: {'lr': 2.022476054036243e-05, 'samples': 22432256, 'steps': 43812, 'loss/train': 1.5596256256103516} 02/26/2022 10:18:36 - INFO - codeparrot_training - Step 43813: {'lr': 2.0218313860683207e-05, 'samples': 22432768, 'steps': 43813, 'loss/train': 1.3284732103347778} 02/26/2022 10:18:42 - INFO - codeparrot_training - Step 43814: {'lr': 2.0211868165315784e-05, 'samples': 22433280, 'steps': 43814, 'loss/train': 2.014993190765381} 02/26/2022 10:18:45 - INFO - codeparrot_training - Step 43815: {'lr': 2.020542345428775e-05, 'samples': 22433792, 'steps': 43815, 'loss/train': 0.9022366404533386} 02/26/2022 10:18:51 - INFO - codeparrot_training - Step 43816: {'lr': 2.0198979727626743e-05, 'samples': 22434304, 'steps': 43816, 'loss/train': 1.9093165397644043} 02/26/2022 10:18:54 - INFO - codeparrot_training - Step 43817: {'lr': 2.019253698536025e-05, 'samples': 22434816, 'steps': 43817, 'loss/train': 0.9826906323432922} 02/26/2022 10:19:00 - INFO - codeparrot_training - Step 43818: {'lr': 2.018609522751602e-05, 'samples': 22435328, 'steps': 43818, 'loss/train': 1.0435973405838013} 02/26/2022 10:19:03 - INFO - codeparrot_training - Step 43819: {'lr': 2.0179654454121594e-05, 'samples': 22435840, 'steps': 43819, 'loss/train': 1.8581339120864868} 02/26/2022 10:19:09 - INFO - codeparrot_training - Step 43820: {'lr': 2.0173214665204552e-05, 'samples': 22436352, 'steps': 43820, 'loss/train': 2.1677000522613525} 02/26/2022 10:19:12 - INFO - codeparrot_training - Step 43821: {'lr': 2.0166775860792435e-05, 'samples': 22436864, 'steps': 43821, 'loss/train': 1.340701937675476} 02/26/2022 10:19:18 - INFO - codeparrot_training - Step 43822: {'lr': 2.0160338040912912e-05, 'samples': 22437376, 'steps': 43822, 'loss/train': 0.30107632279396057} 02/26/2022 10:19:21 - INFO - codeparrot_training - Step 43823: {'lr': 2.0153901205593517e-05, 'samples': 22437888, 'steps': 43823, 'loss/train': 1.8809938430786133} 02/26/2022 10:19:27 - INFO - codeparrot_training - Step 43824: {'lr': 2.014746535486181e-05, 'samples': 22438400, 'steps': 43824, 'loss/train': 1.2629835605621338} 02/26/2022 10:19:30 - INFO - codeparrot_training - Step 43825: {'lr': 2.0141030488745354e-05, 'samples': 22438912, 'steps': 43825, 'loss/train': 2.004812240600586} 02/26/2022 10:19:36 - INFO - codeparrot_training - Step 43826: {'lr': 2.0134596607271765e-05, 'samples': 22439424, 'steps': 43826, 'loss/train': 2.4527204036712646} 02/26/2022 10:19:39 - INFO - codeparrot_training - Step 43827: {'lr': 2.0128163710468577e-05, 'samples': 22439936, 'steps': 43827, 'loss/train': 2.2179901599884033} 02/26/2022 10:19:45 - INFO - codeparrot_training - Step 43828: {'lr': 2.0121731798363324e-05, 'samples': 22440448, 'steps': 43828, 'loss/train': 2.125952959060669} 02/26/2022 10:19:49 - INFO - codeparrot_training - Step 43829: {'lr': 2.0115300870983567e-05, 'samples': 22440960, 'steps': 43829, 'loss/train': 1.1136139631271362} 02/26/2022 10:19:54 - INFO - codeparrot_training - Step 43830: {'lr': 2.0108870928356893e-05, 'samples': 22441472, 'steps': 43830, 'loss/train': 1.0993467569351196} 02/26/2022 10:19:58 - INFO - codeparrot_training - Step 43831: {'lr': 2.010244197051081e-05, 'samples': 22441984, 'steps': 43831, 'loss/train': 2.5315487384796143} 02/26/2022 10:20:03 - INFO - codeparrot_training - Step 43832: {'lr': 2.0096013997472823e-05, 'samples': 22442496, 'steps': 43832, 'loss/train': 2.315072536468506} 02/26/2022 10:20:07 - INFO - codeparrot_training - Step 43833: {'lr': 2.0089587009270606e-05, 'samples': 22443008, 'steps': 43833, 'loss/train': 1.696114182472229} 02/26/2022 10:20:12 - INFO - codeparrot_training - Step 43834: {'lr': 2.0083161005931522e-05, 'samples': 22443520, 'steps': 43834, 'loss/train': 1.0973819494247437} 02/26/2022 10:20:16 - INFO - codeparrot_training - Step 43835: {'lr': 2.007673598748319e-05, 'samples': 22444032, 'steps': 43835, 'loss/train': 1.7770588397979736} 02/26/2022 10:20:21 - INFO - codeparrot_training - Step 43836: {'lr': 2.0070311953953062e-05, 'samples': 22444544, 'steps': 43836, 'loss/train': 2.0600974559783936} 02/26/2022 10:20:25 - INFO - codeparrot_training - Step 43837: {'lr': 2.006388890536881e-05, 'samples': 22445056, 'steps': 43837, 'loss/train': 1.2696826457977295} 02/26/2022 10:20:30 - INFO - codeparrot_training - Step 43838: {'lr': 2.0057466841757744e-05, 'samples': 22445568, 'steps': 43838, 'loss/train': 1.8642808198928833} 02/26/2022 10:20:34 - INFO - codeparrot_training - Step 43839: {'lr': 2.0051045763147535e-05, 'samples': 22446080, 'steps': 43839, 'loss/train': 1.5182468891143799} 02/26/2022 10:20:40 - INFO - codeparrot_training - Step 43840: {'lr': 2.0044625669565582e-05, 'samples': 22446592, 'steps': 43840, 'loss/train': 1.523544192314148} 02/26/2022 10:20:44 - INFO - codeparrot_training - Step 43841: {'lr': 2.0038206561039524e-05, 'samples': 22447104, 'steps': 43841, 'loss/train': 2.371044874191284} 02/26/2022 10:20:49 - INFO - codeparrot_training - Step 43842: {'lr': 2.0031788437596676e-05, 'samples': 22447616, 'steps': 43842, 'loss/train': 0.7976179718971252} 02/26/2022 10:20:53 - INFO - codeparrot_training - Step 43843: {'lr': 2.0025371299264655e-05, 'samples': 22448128, 'steps': 43843, 'loss/train': 1.791491985321045} 02/26/2022 10:20:58 - INFO - codeparrot_training - Step 43844: {'lr': 2.0018955146070882e-05, 'samples': 22448640, 'steps': 43844, 'loss/train': 1.8317155838012695} 02/26/2022 10:21:02 - INFO - codeparrot_training - Step 43845: {'lr': 2.0012539978042922e-05, 'samples': 22449152, 'steps': 43845, 'loss/train': 0.15922223031520844} 02/26/2022 10:21:07 - INFO - codeparrot_training - Step 43846: {'lr': 2.000612579520822e-05, 'samples': 22449664, 'steps': 43846, 'loss/train': 1.7359062433242798} 02/26/2022 10:21:11 - INFO - codeparrot_training - Step 43847: {'lr': 1.9999712597594227e-05, 'samples': 22450176, 'steps': 43847, 'loss/train': 1.3494253158569336} 02/26/2022 10:21:16 - INFO - codeparrot_training - Step 43848: {'lr': 1.9993300385228396e-05, 'samples': 22450688, 'steps': 43848, 'loss/train': 1.2517229318618774} 02/26/2022 10:21:20 - INFO - codeparrot_training - Step 43849: {'lr': 1.998688915813829e-05, 'samples': 22451200, 'steps': 43849, 'loss/train': 2.7270681858062744} 02/26/2022 10:21:26 - INFO - codeparrot_training - Step 43850: {'lr': 1.9980478916351297e-05, 'samples': 22451712, 'steps': 43850, 'loss/train': 1.3634605407714844} 02/26/2022 10:21:29 - INFO - codeparrot_training - Step 43851: {'lr': 1.99740696598949e-05, 'samples': 22452224, 'steps': 43851, 'loss/train': 2.0398809909820557} 02/26/2022 10:21:35 - INFO - codeparrot_training - Step 43852: {'lr': 1.9967661388796493e-05, 'samples': 22452736, 'steps': 43852, 'loss/train': 1.5950738191604614} 02/26/2022 10:21:40 - INFO - codeparrot_training - Step 43853: {'lr': 1.9961254103083638e-05, 'samples': 22453248, 'steps': 43853, 'loss/train': 2.6485507488250732} 02/26/2022 10:21:44 - INFO - codeparrot_training - Step 43854: {'lr': 1.9954847802783698e-05, 'samples': 22453760, 'steps': 43854, 'loss/train': 2.2636215686798096} 02/26/2022 10:21:47 - INFO - codeparrot_training - Step 43855: {'lr': 1.9948442487924156e-05, 'samples': 22454272, 'steps': 43855, 'loss/train': 1.1294500827789307} 02/26/2022 10:21:53 - INFO - codeparrot_training - Step 43856: {'lr': 1.9942038158532405e-05, 'samples': 22454784, 'steps': 43856, 'loss/train': 1.655468225479126} 02/26/2022 10:21:58 - INFO - codeparrot_training - Step 43857: {'lr': 1.993563481463595e-05, 'samples': 22455296, 'steps': 43857, 'loss/train': 0.7238036394119263} 02/26/2022 10:22:02 - INFO - codeparrot_training - Step 43858: {'lr': 1.992923245626213e-05, 'samples': 22455808, 'steps': 43858, 'loss/train': 1.7627135515213013} 02/26/2022 10:22:08 - INFO - codeparrot_training - Step 43859: {'lr': 1.9922831083438535e-05, 'samples': 22456320, 'steps': 43859, 'loss/train': 0.20615874230861664} 02/26/2022 10:22:11 - INFO - codeparrot_training - Step 43860: {'lr': 1.9916430696192368e-05, 'samples': 22456832, 'steps': 43860, 'loss/train': 1.5028220415115356} 02/26/2022 10:22:17 - INFO - codeparrot_training - Step 43861: {'lr': 1.9910031294551183e-05, 'samples': 22457344, 'steps': 43861, 'loss/train': 2.1154861450195312} 02/26/2022 10:22:20 - INFO - codeparrot_training - Step 43862: {'lr': 1.9903632878542327e-05, 'samples': 22457856, 'steps': 43862, 'loss/train': 1.9556525945663452} 02/26/2022 10:22:26 - INFO - codeparrot_training - Step 43863: {'lr': 1.9897235448193358e-05, 'samples': 22458368, 'steps': 43863, 'loss/train': 1.813613772392273} 02/26/2022 10:22:29 - INFO - codeparrot_training - Step 43864: {'lr': 1.9890839003531446e-05, 'samples': 22458880, 'steps': 43864, 'loss/train': 2.005075454711914} 02/26/2022 10:22:35 - INFO - codeparrot_training - Step 43865: {'lr': 1.9884443544584156e-05, 'samples': 22459392, 'steps': 43865, 'loss/train': 1.1705480813980103} 02/26/2022 10:22:38 - INFO - codeparrot_training - Step 43866: {'lr': 1.9878049071378856e-05, 'samples': 22459904, 'steps': 43866, 'loss/train': 1.6140512228012085} 02/26/2022 10:22:44 - INFO - codeparrot_training - Step 43867: {'lr': 1.9871655583942883e-05, 'samples': 22460416, 'steps': 43867, 'loss/train': 1.451699137687683} 02/26/2022 10:22:47 - INFO - codeparrot_training - Step 43868: {'lr': 1.9865263082303687e-05, 'samples': 22460928, 'steps': 43868, 'loss/train': 1.4963024854660034} 02/26/2022 10:22:53 - INFO - codeparrot_training - Step 43869: {'lr': 1.9858871566488633e-05, 'samples': 22461440, 'steps': 43869, 'loss/train': 0.3987034857273102} 02/26/2022 10:22:56 - INFO - codeparrot_training - Step 43870: {'lr': 1.9852481036525123e-05, 'samples': 22461952, 'steps': 43870, 'loss/train': 1.4994840621948242} 02/26/2022 10:23:02 - INFO - codeparrot_training - Step 43871: {'lr': 1.9846091492440434e-05, 'samples': 22462464, 'steps': 43871, 'loss/train': 1.8571969270706177} 02/26/2022 10:23:05 - INFO - codeparrot_training - Step 43872: {'lr': 1.9839702934262045e-05, 'samples': 22462976, 'steps': 43872, 'loss/train': 0.23138749599456787} 02/26/2022 10:23:11 - INFO - codeparrot_training - Step 43873: {'lr': 1.9833315362017297e-05, 'samples': 22463488, 'steps': 43873, 'loss/train': 1.7051478624343872} 02/26/2022 10:23:15 - INFO - codeparrot_training - Step 43874: {'lr': 1.9826928775733527e-05, 'samples': 22464000, 'steps': 43874, 'loss/train': 1.7428762912750244} 02/26/2022 10:23:20 - INFO - codeparrot_training - Step 43875: {'lr': 1.9820543175438076e-05, 'samples': 22464512, 'steps': 43875, 'loss/train': 1.8086543083190918} 02/26/2022 10:23:24 - INFO - codeparrot_training - Step 43876: {'lr': 1.9814158561158368e-05, 'samples': 22465024, 'steps': 43876, 'loss/train': 2.5017173290252686} 02/26/2022 10:23:29 - INFO - codeparrot_training - Step 43877: {'lr': 1.980777493292171e-05, 'samples': 22465536, 'steps': 43877, 'loss/train': 0.5445043444633484} 02/26/2022 10:23:33 - INFO - codeparrot_training - Step 43878: {'lr': 1.9801392290755448e-05, 'samples': 22466048, 'steps': 43878, 'loss/train': 1.9074139595031738} 02/26/2022 10:23:38 - INFO - codeparrot_training - Step 43879: {'lr': 1.979501063468689e-05, 'samples': 22466560, 'steps': 43879, 'loss/train': 0.796567440032959} 02/26/2022 10:23:42 - INFO - codeparrot_training - Step 43880: {'lr': 1.9788629964743454e-05, 'samples': 22467072, 'steps': 43880, 'loss/train': 2.2069852352142334} 02/26/2022 10:23:47 - INFO - codeparrot_training - Step 43881: {'lr': 1.97822502809524e-05, 'samples': 22467584, 'steps': 43881, 'loss/train': 1.8154195547103882} 02/26/2022 10:23:51 - INFO - codeparrot_training - Step 43882: {'lr': 1.9775871583341127e-05, 'samples': 22468096, 'steps': 43882, 'loss/train': 1.9115105867385864} 02/26/2022 10:23:56 - INFO - codeparrot_training - Step 43883: {'lr': 1.976949387193683e-05, 'samples': 22468608, 'steps': 43883, 'loss/train': 2.654982328414917} 02/26/2022 10:24:00 - INFO - codeparrot_training - Step 43884: {'lr': 1.9763117146766985e-05, 'samples': 22469120, 'steps': 43884, 'loss/train': 2.313267230987549} 02/26/2022 10:24:06 - INFO - codeparrot_training - Step 43885: {'lr': 1.9756741407858825e-05, 'samples': 22469632, 'steps': 43885, 'loss/train': 1.5166712999343872} 02/26/2022 10:24:09 - INFO - codeparrot_training - Step 43886: {'lr': 1.9750366655239687e-05, 'samples': 22470144, 'steps': 43886, 'loss/train': 0.6908068060874939} 02/26/2022 10:24:15 - INFO - codeparrot_training - Step 43887: {'lr': 1.9743992888936803e-05, 'samples': 22470656, 'steps': 43887, 'loss/train': 0.4037047028541565} 02/26/2022 10:24:18 - INFO - codeparrot_training - Step 43888: {'lr': 1.973762010897759e-05, 'samples': 22471168, 'steps': 43888, 'loss/train': 2.2648537158966064} 02/26/2022 10:24:24 - INFO - codeparrot_training - Step 43889: {'lr': 1.9731248315389306e-05, 'samples': 22471680, 'steps': 43889, 'loss/train': 1.0088422298431396} 02/26/2022 10:24:27 - INFO - codeparrot_training - Step 43890: {'lr': 1.9724877508199235e-05, 'samples': 22472192, 'steps': 43890, 'loss/train': 1.3535374402999878} 02/26/2022 10:24:33 - INFO - codeparrot_training - Step 43891: {'lr': 1.9718507687434606e-05, 'samples': 22472704, 'steps': 43891, 'loss/train': 2.247473955154419} 02/26/2022 10:24:36 - INFO - codeparrot_training - Step 43892: {'lr': 1.971213885312284e-05, 'samples': 22473216, 'steps': 43892, 'loss/train': 2.134769916534424} 02/26/2022 10:24:42 - INFO - codeparrot_training - Step 43893: {'lr': 1.9705771005291084e-05, 'samples': 22473728, 'steps': 43893, 'loss/train': 2.107015609741211} 02/26/2022 10:24:46 - INFO - codeparrot_training - Step 43894: {'lr': 1.969940414396676e-05, 'samples': 22474240, 'steps': 43894, 'loss/train': 1.4998337030410767} 02/26/2022 10:24:52 - INFO - codeparrot_training - Step 43895: {'lr': 1.9693038269176984e-05, 'samples': 22474752, 'steps': 43895, 'loss/train': 1.8322309255599976} 02/26/2022 10:24:55 - INFO - codeparrot_training - Step 43896: {'lr': 1.968667338094915e-05, 'samples': 22475264, 'steps': 43896, 'loss/train': 2.7887208461761475} 02/26/2022 10:25:01 - INFO - codeparrot_training - Step 43897: {'lr': 1.9680309479310405e-05, 'samples': 22475776, 'steps': 43897, 'loss/train': 2.078357696533203} 02/26/2022 10:25:04 - INFO - codeparrot_training - Step 43898: {'lr': 1.9673946564288197e-05, 'samples': 22476288, 'steps': 43898, 'loss/train': 2.2084178924560547} 02/26/2022 10:25:10 - INFO - codeparrot_training - Step 43899: {'lr': 1.9667584635909562e-05, 'samples': 22476800, 'steps': 43899, 'loss/train': 1.5692343711853027} 02/26/2022 10:25:13 - INFO - codeparrot_training - Step 43900: {'lr': 1.9661223694201895e-05, 'samples': 22477312, 'steps': 43900, 'loss/train': 1.3374872207641602} 02/26/2022 10:25:19 - INFO - codeparrot_training - Step 43901: {'lr': 1.9654863739192367e-05, 'samples': 22477824, 'steps': 43901, 'loss/train': 0.5901874303817749} 02/26/2022 10:25:22 - INFO - codeparrot_training - Step 43902: {'lr': 1.964850477090835e-05, 'samples': 22478336, 'steps': 43902, 'loss/train': 1.7448763847351074} 02/26/2022 10:25:28 - INFO - codeparrot_training - Step 43903: {'lr': 1.9642146789376897e-05, 'samples': 22478848, 'steps': 43903, 'loss/train': 1.2017637491226196} 02/26/2022 10:25:31 - INFO - codeparrot_training - Step 43904: {'lr': 1.963578979462541e-05, 'samples': 22479360, 'steps': 43904, 'loss/train': 0.21360686421394348} 02/26/2022 10:25:37 - INFO - codeparrot_training - Step 43905: {'lr': 1.962943378668097e-05, 'samples': 22479872, 'steps': 43905, 'loss/train': 2.2890353202819824} 02/26/2022 10:25:41 - INFO - codeparrot_training - Step 43906: {'lr': 1.9623078765570958e-05, 'samples': 22480384, 'steps': 43906, 'loss/train': 1.4535874128341675} 02/26/2022 10:25:46 - INFO - codeparrot_training - Step 43907: {'lr': 1.9616724731322504e-05, 'samples': 22480896, 'steps': 43907, 'loss/train': 2.1506361961364746} 02/26/2022 10:25:50 - INFO - codeparrot_training - Step 43908: {'lr': 1.9610371683962874e-05, 'samples': 22481408, 'steps': 43908, 'loss/train': 2.1257059574127197} 02/26/2022 10:25:55 - INFO - codeparrot_training - Step 43909: {'lr': 1.960401962351918e-05, 'samples': 22481920, 'steps': 43909, 'loss/train': 1.7139891386032104} 02/26/2022 10:25:59 - INFO - codeparrot_training - Step 43910: {'lr': 1.9597668550018767e-05, 'samples': 22482432, 'steps': 43910, 'loss/train': 1.9152476787567139} 02/26/2022 10:26:04 - INFO - codeparrot_training - Step 43911: {'lr': 1.959131846348877e-05, 'samples': 22482944, 'steps': 43911, 'loss/train': 2.175318956375122} 02/26/2022 10:26:08 - INFO - codeparrot_training - Step 43912: {'lr': 1.958496936395643e-05, 'samples': 22483456, 'steps': 43912, 'loss/train': 1.6276425123214722} 02/26/2022 10:26:13 - INFO - codeparrot_training - Step 43913: {'lr': 1.957862125144888e-05, 'samples': 22483968, 'steps': 43913, 'loss/train': 1.1410492658615112} 02/26/2022 10:26:17 - INFO - codeparrot_training - Step 43914: {'lr': 1.9572274125993326e-05, 'samples': 22484480, 'steps': 43914, 'loss/train': 2.458616256713867} 02/26/2022 10:26:25 - INFO - codeparrot_training - Step 43915: {'lr': 1.9565927987617027e-05, 'samples': 22484992, 'steps': 43915, 'loss/train': 0.8795093894004822} 02/26/2022 10:26:28 - INFO - codeparrot_training - Step 43916: {'lr': 1.9559582836347094e-05, 'samples': 22485504, 'steps': 43916, 'loss/train': 2.233673572540283} 02/26/2022 10:26:32 - INFO - codeparrot_training - Step 43917: {'lr': 1.955323867221076e-05, 'samples': 22486016, 'steps': 43917, 'loss/train': 1.1261332035064697} 02/26/2022 10:26:37 - INFO - codeparrot_training - Step 43918: {'lr': 1.954689549523514e-05, 'samples': 22486528, 'steps': 43918, 'loss/train': 1.9638397693634033} 02/26/2022 10:26:41 - INFO - codeparrot_training - Step 43919: {'lr': 1.9540553305447462e-05, 'samples': 22487040, 'steps': 43919, 'loss/train': 2.019672155380249} 02/26/2022 10:26:46 - INFO - codeparrot_training - Step 43920: {'lr': 1.95342121028749e-05, 'samples': 22487552, 'steps': 43920, 'loss/train': 1.5649964809417725} 02/26/2022 10:26:50 - INFO - codeparrot_training - Step 43921: {'lr': 1.952787188754457e-05, 'samples': 22488064, 'steps': 43921, 'loss/train': 2.075366497039795} 02/26/2022 10:26:56 - INFO - codeparrot_training - Step 43922: {'lr': 1.9521532659483614e-05, 'samples': 22488576, 'steps': 43922, 'loss/train': 0.8432846665382385} 02/26/2022 10:26:59 - INFO - codeparrot_training - Step 43923: {'lr': 1.9515194418719268e-05, 'samples': 22489088, 'steps': 43923, 'loss/train': 1.9711413383483887} 02/26/2022 10:27:05 - INFO - codeparrot_training - Step 43924: {'lr': 1.950885716527867e-05, 'samples': 22489600, 'steps': 43924, 'loss/train': 1.4739127159118652} 02/26/2022 10:27:08 - INFO - codeparrot_training - Step 43925: {'lr': 1.9502520899188913e-05, 'samples': 22490112, 'steps': 43925, 'loss/train': 2.1115334033966064} 02/26/2022 10:27:15 - INFO - codeparrot_training - Step 43926: {'lr': 1.9496185620477113e-05, 'samples': 22490624, 'steps': 43926, 'loss/train': 2.2978007793426514} 02/26/2022 10:27:19 - INFO - codeparrot_training - Step 43927: {'lr': 1.9489851329170526e-05, 'samples': 22491136, 'steps': 43927, 'loss/train': 1.0188546180725098} 02/26/2022 10:27:24 - INFO - codeparrot_training - Step 43928: {'lr': 1.948351802529616e-05, 'samples': 22491648, 'steps': 43928, 'loss/train': 2.3698296546936035} 02/26/2022 10:27:28 - INFO - codeparrot_training - Step 43929: {'lr': 1.9477185708881297e-05, 'samples': 22492160, 'steps': 43929, 'loss/train': 1.9153876304626465} 02/26/2022 10:27:33 - INFO - codeparrot_training - Step 43930: {'lr': 1.9470854379952913e-05, 'samples': 22492672, 'steps': 43930, 'loss/train': 1.6623371839523315} 02/26/2022 10:27:37 - INFO - codeparrot_training - Step 43931: {'lr': 1.9464524038538185e-05, 'samples': 22493184, 'steps': 43931, 'loss/train': 1.2903441190719604} 02/26/2022 10:27:42 - INFO - codeparrot_training - Step 43932: {'lr': 1.9458194684664226e-05, 'samples': 22493696, 'steps': 43932, 'loss/train': 1.928707242012024} 02/26/2022 10:27:46 - INFO - codeparrot_training - Step 43933: {'lr': 1.9451866318358214e-05, 'samples': 22494208, 'steps': 43933, 'loss/train': 0.9612748026847839} 02/26/2022 10:27:51 - INFO - codeparrot_training - Step 43934: {'lr': 1.9445538939647122e-05, 'samples': 22494720, 'steps': 43934, 'loss/train': 1.7725878953933716} 02/26/2022 10:27:59 - INFO - codeparrot_training - Step 43935: {'lr': 1.9439212548558182e-05, 'samples': 22495232, 'steps': 43935, 'loss/train': 2.034044027328491} 02/26/2022 10:28:02 - INFO - codeparrot_training - Step 43936: {'lr': 1.94328871451184e-05, 'samples': 22495744, 'steps': 43936, 'loss/train': 1.49234139919281} 02/26/2022 10:28:08 - INFO - codeparrot_training - Step 43937: {'lr': 1.9426562729355003e-05, 'samples': 22496256, 'steps': 43937, 'loss/train': 1.8445290327072144} 02/26/2022 10:28:11 - INFO - codeparrot_training - Step 43938: {'lr': 1.9420239301294912e-05, 'samples': 22496768, 'steps': 43938, 'loss/train': 2.1561474800109863} 02/26/2022 10:28:17 - INFO - codeparrot_training - Step 43939: {'lr': 1.941391686096533e-05, 'samples': 22497280, 'steps': 43939, 'loss/train': 0.9168580174446106} 02/26/2022 10:28:20 - INFO - codeparrot_training - Step 43940: {'lr': 1.940759540839329e-05, 'samples': 22497792, 'steps': 43940, 'loss/train': 2.79150652885437} 02/26/2022 10:28:26 - INFO - codeparrot_training - Step 43941: {'lr': 1.9401274943605913e-05, 'samples': 22498304, 'steps': 43941, 'loss/train': 1.4240717887878418} 02/26/2022 10:28:29 - INFO - codeparrot_training - Step 43942: {'lr': 1.9394955466630254e-05, 'samples': 22498816, 'steps': 43942, 'loss/train': 0.761337399482727} 02/26/2022 10:28:35 - INFO - codeparrot_training - Step 43943: {'lr': 1.938863697749338e-05, 'samples': 22499328, 'steps': 43943, 'loss/train': 2.0061745643615723} 02/26/2022 10:28:38 - INFO - codeparrot_training - Step 43944: {'lr': 1.938231947622232e-05, 'samples': 22499840, 'steps': 43944, 'loss/train': 1.974061369895935} 02/26/2022 10:28:44 - INFO - codeparrot_training - Step 43945: {'lr': 1.9376002962844195e-05, 'samples': 22500352, 'steps': 43945, 'loss/train': 0.44279757142066956} 02/26/2022 10:28:47 - INFO - codeparrot_training - Step 43946: {'lr': 1.936968743738607e-05, 'samples': 22500864, 'steps': 43946, 'loss/train': 1.695180058479309} 02/26/2022 10:28:54 - INFO - codeparrot_training - Step 43947: {'lr': 1.9363372899874944e-05, 'samples': 22501376, 'steps': 43947, 'loss/train': 2.4326558113098145} 02/26/2022 10:28:58 - INFO - codeparrot_training - Step 43948: {'lr': 1.935705935033788e-05, 'samples': 22501888, 'steps': 43948, 'loss/train': 1.4776768684387207} 02/26/2022 10:29:03 - INFO - codeparrot_training - Step 43949: {'lr': 1.935074678880197e-05, 'samples': 22502400, 'steps': 43949, 'loss/train': 1.9536120891571045} 02/26/2022 10:29:07 - INFO - codeparrot_training - Step 43950: {'lr': 1.9344435215294193e-05, 'samples': 22502912, 'steps': 43950, 'loss/train': 1.8860721588134766} 02/26/2022 10:29:12 - INFO - codeparrot_training - Step 43951: {'lr': 1.9338124629841664e-05, 'samples': 22503424, 'steps': 43951, 'loss/train': 0.06276270002126694} 02/26/2022 10:29:16 - INFO - codeparrot_training - Step 43952: {'lr': 1.9331815032471277e-05, 'samples': 22503936, 'steps': 43952, 'loss/train': 1.4646512269973755} 02/26/2022 10:29:21 - INFO - codeparrot_training - Step 43953: {'lr': 1.932550642321021e-05, 'samples': 22504448, 'steps': 43953, 'loss/train': 1.6647932529449463} 02/26/2022 10:29:25 - INFO - codeparrot_training - Step 43954: {'lr': 1.9319198802085436e-05, 'samples': 22504960, 'steps': 43954, 'loss/train': 1.5435038805007935} 02/26/2022 10:29:32 - INFO - codeparrot_training - Step 43955: {'lr': 1.9312892169123962e-05, 'samples': 22505472, 'steps': 43955, 'loss/train': 1.6119916439056396} 02/26/2022 10:29:36 - INFO - codeparrot_training - Step 43956: {'lr': 1.930658652435274e-05, 'samples': 22505984, 'steps': 43956, 'loss/train': 2.1861534118652344} 02/26/2022 10:29:41 - INFO - codeparrot_training - Step 43957: {'lr': 1.930028186779892e-05, 'samples': 22506496, 'steps': 43957, 'loss/train': 1.6345672607421875} 02/26/2022 10:29:44 - INFO - codeparrot_training - Step 43958: {'lr': 1.9293978199489414e-05, 'samples': 22507008, 'steps': 43958, 'loss/train': 1.0778197050094604} 02/26/2022 10:29:50 - INFO - codeparrot_training - Step 43959: {'lr': 1.9287675519451266e-05, 'samples': 22507520, 'steps': 43959, 'loss/train': 1.6340985298156738} 02/26/2022 10:29:53 - INFO - codeparrot_training - Step 43960: {'lr': 1.9281373827711446e-05, 'samples': 22508032, 'steps': 43960, 'loss/train': 0.7536647915840149} 02/26/2022 10:29:59 - INFO - codeparrot_training - Step 43961: {'lr': 1.927507312429694e-05, 'samples': 22508544, 'steps': 43961, 'loss/train': 0.5697888135910034} 02/26/2022 10:30:03 - INFO - codeparrot_training - Step 43962: {'lr': 1.9268773409234773e-05, 'samples': 22509056, 'steps': 43962, 'loss/train': 2.946394205093384} 02/26/2022 10:30:08 - INFO - codeparrot_training - Step 43963: {'lr': 1.9262474682551905e-05, 'samples': 22509568, 'steps': 43963, 'loss/train': 1.8599181175231934} 02/26/2022 10:30:12 - INFO - codeparrot_training - Step 43964: {'lr': 1.9256176944275367e-05, 'samples': 22510080, 'steps': 43964, 'loss/train': 1.7560877799987793} 02/26/2022 10:30:17 - INFO - codeparrot_training - Step 43965: {'lr': 1.9249880194432024e-05, 'samples': 22510592, 'steps': 43965, 'loss/train': 2.406370162963867} 02/26/2022 10:30:21 - INFO - codeparrot_training - Step 43966: {'lr': 1.9243584433048965e-05, 'samples': 22511104, 'steps': 43966, 'loss/train': 2.0007286071777344} 02/26/2022 10:30:26 - INFO - codeparrot_training - Step 43967: {'lr': 1.923728966015309e-05, 'samples': 22511616, 'steps': 43967, 'loss/train': 1.6675264835357666} 02/26/2022 10:30:30 - INFO - codeparrot_training - Step 43968: {'lr': 1.9230995875771483e-05, 'samples': 22512128, 'steps': 43968, 'loss/train': 1.068717360496521} 02/26/2022 10:30:35 - INFO - codeparrot_training - Step 43969: {'lr': 1.9224703079930905e-05, 'samples': 22512640, 'steps': 43969, 'loss/train': 1.3827135562896729} 02/26/2022 10:30:39 - INFO - codeparrot_training - Step 43970: {'lr': 1.921841127265847e-05, 'samples': 22513152, 'steps': 43970, 'loss/train': 0.5793749690055847} 02/26/2022 10:30:46 - INFO - codeparrot_training - Step 43971: {'lr': 1.9212120453981047e-05, 'samples': 22513664, 'steps': 43971, 'loss/train': 1.5587520599365234} 02/26/2022 10:30:50 - INFO - codeparrot_training - Step 43972: {'lr': 1.9205830623925696e-05, 'samples': 22514176, 'steps': 43972, 'loss/train': 1.8233580589294434} 02/26/2022 10:30:55 - INFO - codeparrot_training - Step 43973: {'lr': 1.9199541782519174e-05, 'samples': 22514688, 'steps': 43973, 'loss/train': 0.9353955388069153} 02/26/2022 10:30:59 - INFO - codeparrot_training - Step 43974: {'lr': 1.91932539297886e-05, 'samples': 22515200, 'steps': 43974, 'loss/train': 1.533398985862732} 02/26/2022 10:31:04 - INFO - codeparrot_training - Step 43975: {'lr': 1.9186967065760786e-05, 'samples': 22515712, 'steps': 43975, 'loss/train': 2.2748358249664307} 02/26/2022 10:31:07 - INFO - codeparrot_training - Step 43976: {'lr': 1.9180681190462763e-05, 'samples': 22516224, 'steps': 43976, 'loss/train': 1.627120018005371} 02/26/2022 10:31:13 - INFO - codeparrot_training - Step 43977: {'lr': 1.9174396303921398e-05, 'samples': 22516736, 'steps': 43977, 'loss/train': 1.7510101795196533} 02/26/2022 10:31:17 - INFO - codeparrot_training - Step 43978: {'lr': 1.9168112406163618e-05, 'samples': 22517248, 'steps': 43978, 'loss/train': 0.9625411629676819} 02/26/2022 10:31:22 - INFO - codeparrot_training - Step 43979: {'lr': 1.9161829497216315e-05, 'samples': 22517760, 'steps': 43979, 'loss/train': 2.05888032913208} 02/26/2022 10:31:25 - INFO - codeparrot_training - Step 43980: {'lr': 1.915554757710647e-05, 'samples': 22518272, 'steps': 43980, 'loss/train': 1.0286204814910889} 02/26/2022 10:31:33 - INFO - codeparrot_training - Step 43981: {'lr': 1.9149266645860976e-05, 'samples': 22518784, 'steps': 43981, 'loss/train': 1.733374834060669} 02/26/2022 10:31:36 - INFO - codeparrot_training - Step 43982: {'lr': 1.91429867035067e-05, 'samples': 22519296, 'steps': 43982, 'loss/train': 2.0034515857696533} 02/26/2022 10:31:42 - INFO - codeparrot_training - Step 43983: {'lr': 1.9136707750070536e-05, 'samples': 22519808, 'steps': 43983, 'loss/train': 2.4939634799957275} 02/26/2022 10:31:45 - INFO - codeparrot_training - Step 43984: {'lr': 1.9130429785579438e-05, 'samples': 22520320, 'steps': 43984, 'loss/train': 0.7421519160270691} 02/26/2022 10:31:51 - INFO - codeparrot_training - Step 43985: {'lr': 1.9124152810060275e-05, 'samples': 22520832, 'steps': 43985, 'loss/train': 3.264716386795044} 02/26/2022 10:31:54 - INFO - codeparrot_training - Step 43986: {'lr': 1.911787682353991e-05, 'samples': 22521344, 'steps': 43986, 'loss/train': 1.2016572952270508} 02/26/2022 10:32:00 - INFO - codeparrot_training - Step 43987: {'lr': 1.9111601826045237e-05, 'samples': 22521856, 'steps': 43987, 'loss/train': 1.7218455076217651} 02/26/2022 10:32:03 - INFO - codeparrot_training - Step 43988: {'lr': 1.9105327817603186e-05, 'samples': 22522368, 'steps': 43988, 'loss/train': 1.9870600700378418} 02/26/2022 10:32:09 - INFO - codeparrot_training - Step 43989: {'lr': 1.9099054798240535e-05, 'samples': 22522880, 'steps': 43989, 'loss/train': 1.014692783355713} 02/26/2022 10:32:12 - INFO - codeparrot_training - Step 43990: {'lr': 1.909278276798429e-05, 'samples': 22523392, 'steps': 43990, 'loss/train': 1.6703882217407227} 02/26/2022 10:32:18 - INFO - codeparrot_training - Step 43991: {'lr': 1.9086511726861182e-05, 'samples': 22523904, 'steps': 43991, 'loss/train': 2.2936129570007324} 02/26/2022 10:32:21 - INFO - codeparrot_training - Step 43992: {'lr': 1.9080241674898162e-05, 'samples': 22524416, 'steps': 43992, 'loss/train': 1.3780274391174316} 02/26/2022 10:32:28 - INFO - codeparrot_training - Step 43993: {'lr': 1.9073972612122072e-05, 'samples': 22524928, 'steps': 43993, 'loss/train': 1.5659040212631226} 02/26/2022 10:32:32 - INFO - codeparrot_training - Step 43994: {'lr': 1.9067704538559745e-05, 'samples': 22525440, 'steps': 43994, 'loss/train': 1.0398471355438232} 02/26/2022 10:32:37 - INFO - codeparrot_training - Step 43995: {'lr': 1.9061437454238025e-05, 'samples': 22525952, 'steps': 43995, 'loss/train': 1.5446808338165283} 02/26/2022 10:32:41 - INFO - codeparrot_training - Step 43996: {'lr': 1.9055171359183777e-05, 'samples': 22526464, 'steps': 43996, 'loss/train': 2.550922393798828} 02/26/2022 10:32:46 - INFO - codeparrot_training - Step 43997: {'lr': 1.904890625342387e-05, 'samples': 22526976, 'steps': 43997, 'loss/train': 2.6649065017700195} 02/26/2022 10:32:50 - INFO - codeparrot_training - Step 43998: {'lr': 1.9042642136985117e-05, 'samples': 22527488, 'steps': 43998, 'loss/train': 1.2582118511199951} 02/26/2022 10:32:55 - INFO - codeparrot_training - Step 43999: {'lr': 1.90363790098943e-05, 'samples': 22528000, 'steps': 43999, 'loss/train': 1.9502559900283813} 02/26/2022 10:32:55 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 10:33:14 - WARNING - huggingface_hub.repository - Several commits (44) will be pushed upstream. 02/26/2022 10:33:14 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 10:33:49 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy d24e10b..84c8fb0 floral-grass-11 -> floral-grass-11 02/26/2022 10:33:53 - INFO - codeparrot_training - Step 44000: {'lr': 1.9030116872178316e-05, 'samples': 22528512, 'steps': 44000, 'loss/train': 2.3320531845092773} 02/26/2022 10:33:59 - INFO - codeparrot_training - Step 44001: {'lr': 1.9023855723864002e-05, 'samples': 22529024, 'steps': 44001, 'loss/train': 1.6663116216659546} 02/26/2022 10:34:02 - INFO - codeparrot_training - Step 44002: {'lr': 1.901759556497809e-05, 'samples': 22529536, 'steps': 44002, 'loss/train': 2.0551815032958984} 02/26/2022 10:34:10 - INFO - codeparrot_training - Step 44003: {'lr': 1.90113363955475e-05, 'samples': 22530048, 'steps': 44003, 'loss/train': 1.423163890838623} 02/26/2022 10:34:13 - INFO - codeparrot_training - Step 44004: {'lr': 1.900507821559899e-05, 'samples': 22530560, 'steps': 44004, 'loss/train': 1.4444445371627808} 02/26/2022 10:34:19 - INFO - codeparrot_training - Step 44005: {'lr': 1.8998821025159375e-05, 'samples': 22531072, 'steps': 44005, 'loss/train': 2.4846723079681396} 02/26/2022 10:34:22 - INFO - codeparrot_training - Step 44006: {'lr': 1.899256482425543e-05, 'samples': 22531584, 'steps': 44006, 'loss/train': 1.8023244142532349} 02/26/2022 10:34:28 - INFO - codeparrot_training - Step 44007: {'lr': 1.8986309612914006e-05, 'samples': 22532096, 'steps': 44007, 'loss/train': 0.2879928946495056} 02/26/2022 10:34:31 - INFO - codeparrot_training - Step 44008: {'lr': 1.8980055391161877e-05, 'samples': 22532608, 'steps': 44008, 'loss/train': 1.7266427278518677} 02/26/2022 10:34:37 - INFO - codeparrot_training - Step 44009: {'lr': 1.8973802159025862e-05, 'samples': 22533120, 'steps': 44009, 'loss/train': 2.2361907958984375} 02/26/2022 10:34:40 - INFO - codeparrot_training - Step 44010: {'lr': 1.8967549916532656e-05, 'samples': 22533632, 'steps': 44010, 'loss/train': 0.29530131816864014} 02/26/2022 10:34:46 - INFO - codeparrot_training - Step 44011: {'lr': 1.8961298663709185e-05, 'samples': 22534144, 'steps': 44011, 'loss/train': 1.5945879220962524} 02/26/2022 10:34:49 - INFO - codeparrot_training - Step 44012: {'lr': 1.895504840058207e-05, 'samples': 22534656, 'steps': 44012, 'loss/train': 1.5260460376739502} 02/26/2022 10:34:55 - INFO - codeparrot_training - Step 44013: {'lr': 1.894879912717823e-05, 'samples': 22535168, 'steps': 44013, 'loss/train': 1.5213029384613037} 02/26/2022 10:34:58 - INFO - codeparrot_training - Step 44014: {'lr': 1.894255084352428e-05, 'samples': 22535680, 'steps': 44014, 'loss/train': 3.3415493965148926} 02/26/2022 10:35:04 - INFO - codeparrot_training - Step 44015: {'lr': 1.893630354964715e-05, 'samples': 22536192, 'steps': 44015, 'loss/train': 3.2358953952789307} 02/26/2022 10:35:07 - INFO - codeparrot_training - Step 44016: {'lr': 1.8930057245573506e-05, 'samples': 22536704, 'steps': 44016, 'loss/train': 1.6172772645950317} 02/26/2022 10:35:13 - INFO - codeparrot_training - Step 44017: {'lr': 1.8923811931330137e-05, 'samples': 22537216, 'steps': 44017, 'loss/train': 0.9790546894073486} 02/26/2022 10:35:16 - INFO - codeparrot_training - Step 44018: {'lr': 1.891756760694374e-05, 'samples': 22537728, 'steps': 44018, 'loss/train': 2.160904884338379} 02/26/2022 10:35:24 - INFO - codeparrot_training - Step 44019: {'lr': 1.8911324272441133e-05, 'samples': 22538240, 'steps': 44019, 'loss/train': 1.056497573852539} 02/26/2022 10:35:27 - INFO - codeparrot_training - Step 44020: {'lr': 1.890508192784904e-05, 'samples': 22538752, 'steps': 44020, 'loss/train': 1.7328752279281616} 02/26/2022 10:35:33 - INFO - codeparrot_training - Step 44021: {'lr': 1.8898840573194214e-05, 'samples': 22539264, 'steps': 44021, 'loss/train': 1.2247378826141357} 02/26/2022 10:35:36 - INFO - codeparrot_training - Step 44022: {'lr': 1.889260020850331e-05, 'samples': 22539776, 'steps': 44022, 'loss/train': 3.085550546646118} 02/26/2022 10:35:42 - INFO - codeparrot_training - Step 44023: {'lr': 1.8886360833803156e-05, 'samples': 22540288, 'steps': 44023, 'loss/train': 1.4534519910812378} 02/26/2022 10:35:45 - INFO - codeparrot_training - Step 44024: {'lr': 1.8880122449120463e-05, 'samples': 22540800, 'steps': 44024, 'loss/train': 2.2866058349609375} 02/26/2022 10:35:51 - INFO - codeparrot_training - Step 44025: {'lr': 1.8873885054481926e-05, 'samples': 22541312, 'steps': 44025, 'loss/train': 1.5168508291244507} 02/26/2022 10:35:54 - INFO - codeparrot_training - Step 44026: {'lr': 1.886764864991425e-05, 'samples': 22541824, 'steps': 44026, 'loss/train': 2.1663818359375} 02/26/2022 10:36:00 - INFO - codeparrot_training - Step 44027: {'lr': 1.8861413235444213e-05, 'samples': 22542336, 'steps': 44027, 'loss/train': 1.4665101766586304} 02/26/2022 10:36:03 - INFO - codeparrot_training - Step 44028: {'lr': 1.885517881109841e-05, 'samples': 22542848, 'steps': 44028, 'loss/train': 2.0435917377471924} 02/26/2022 10:36:11 - INFO - codeparrot_training - Step 44029: {'lr': 1.8848945376903737e-05, 'samples': 22543360, 'steps': 44029, 'loss/train': 0.6146186590194702} 02/26/2022 10:36:15 - INFO - codeparrot_training - Step 44030: {'lr': 1.8842712932886697e-05, 'samples': 22543872, 'steps': 44030, 'loss/train': 2.2952003479003906} 02/26/2022 10:36:18 - INFO - codeparrot_training - Step 44031: {'lr': 1.883648147907413e-05, 'samples': 22544384, 'steps': 44031, 'loss/train': 1.5369566679000854} 02/26/2022 10:36:24 - INFO - codeparrot_training - Step 44032: {'lr': 1.88302510154926e-05, 'samples': 22544896, 'steps': 44032, 'loss/train': 1.8704774379730225} 02/26/2022 10:36:27 - INFO - codeparrot_training - Step 44033: {'lr': 1.8824021542168973e-05, 'samples': 22545408, 'steps': 44033, 'loss/train': 1.8428127765655518} 02/26/2022 10:36:33 - INFO - codeparrot_training - Step 44034: {'lr': 1.8817793059129756e-05, 'samples': 22545920, 'steps': 44034, 'loss/train': 2.1620123386383057} 02/26/2022 10:36:36 - INFO - codeparrot_training - Step 44035: {'lr': 1.8811565566401735e-05, 'samples': 22546432, 'steps': 44035, 'loss/train': 2.19956374168396} 02/26/2022 10:36:42 - INFO - codeparrot_training - Step 44036: {'lr': 1.8805339064011524e-05, 'samples': 22546944, 'steps': 44036, 'loss/train': 1.258091688156128} 02/26/2022 10:36:45 - INFO - codeparrot_training - Step 44037: {'lr': 1.8799113551985852e-05, 'samples': 22547456, 'steps': 44037, 'loss/train': 2.4753246307373047} 02/26/2022 10:36:53 - INFO - codeparrot_training - Step 44038: {'lr': 1.8792889030351367e-05, 'samples': 22547968, 'steps': 44038, 'loss/train': 1.0560418367385864} 02/26/2022 10:36:56 - INFO - codeparrot_training - Step 44039: {'lr': 1.8786665499134742e-05, 'samples': 22548480, 'steps': 44039, 'loss/train': 4.3891167640686035} 02/26/2022 10:37:02 - INFO - codeparrot_training - Step 44040: {'lr': 1.8780442958362566e-05, 'samples': 22548992, 'steps': 44040, 'loss/train': 1.6752862930297852} 02/26/2022 10:37:06 - INFO - codeparrot_training - Step 44041: {'lr': 1.877422140806162e-05, 'samples': 22549504, 'steps': 44041, 'loss/train': 0.7909420728683472} 02/26/2022 10:37:11 - INFO - codeparrot_training - Step 44042: {'lr': 1.8768000848258472e-05, 'samples': 22550016, 'steps': 44042, 'loss/train': 1.6012473106384277} 02/26/2022 10:37:15 - INFO - codeparrot_training - Step 44043: {'lr': 1.8761781278979763e-05, 'samples': 22550528, 'steps': 44043, 'loss/train': 1.874677062034607} 02/26/2022 10:37:20 - INFO - codeparrot_training - Step 44044: {'lr': 1.8755562700252193e-05, 'samples': 22551040, 'steps': 44044, 'loss/train': 1.1454931497573853} 02/26/2022 10:37:24 - INFO - codeparrot_training - Step 44045: {'lr': 1.87493451121023e-05, 'samples': 22551552, 'steps': 44045, 'loss/train': 1.2387669086456299} 02/26/2022 10:37:29 - INFO - codeparrot_training - Step 44046: {'lr': 1.874312851455681e-05, 'samples': 22552064, 'steps': 44046, 'loss/train': 2.014160394668579} 02/26/2022 10:37:33 - INFO - codeparrot_training - Step 44047: {'lr': 1.8736912907642345e-05, 'samples': 22552576, 'steps': 44047, 'loss/train': 1.901646614074707} 02/26/2022 10:37:39 - INFO - codeparrot_training - Step 44048: {'lr': 1.873069829138552e-05, 'samples': 22553088, 'steps': 44048, 'loss/train': 1.37613046169281} 02/26/2022 10:37:42 - INFO - codeparrot_training - Step 44049: {'lr': 1.872448466581289e-05, 'samples': 22553600, 'steps': 44049, 'loss/train': 1.799850583076477} 02/26/2022 10:37:49 - INFO - codeparrot_training - Step 44050: {'lr': 1.8718272030951167e-05, 'samples': 22554112, 'steps': 44050, 'loss/train': 1.1026482582092285} 02/26/2022 10:37:53 - INFO - codeparrot_training - Step 44051: {'lr': 1.8712060386826934e-05, 'samples': 22554624, 'steps': 44051, 'loss/train': 1.2876883745193481} 02/26/2022 10:37:58 - INFO - codeparrot_training - Step 44052: {'lr': 1.8705849733466812e-05, 'samples': 22555136, 'steps': 44052, 'loss/train': 1.4196891784667969} 02/26/2022 10:38:02 - INFO - codeparrot_training - Step 44053: {'lr': 1.8699640070897307e-05, 'samples': 22555648, 'steps': 44053, 'loss/train': 1.5514488220214844} 02/26/2022 10:38:07 - INFO - codeparrot_training - Step 44054: {'lr': 1.8693431399145172e-05, 'samples': 22556160, 'steps': 44054, 'loss/train': 2.8206305503845215} 02/26/2022 10:38:11 - INFO - codeparrot_training - Step 44055: {'lr': 1.8687223718236918e-05, 'samples': 22556672, 'steps': 44055, 'loss/train': 1.5404722690582275} 02/26/2022 10:38:16 - INFO - codeparrot_training - Step 44056: {'lr': 1.8681017028199132e-05, 'samples': 22557184, 'steps': 44056, 'loss/train': 2.372636556625366} 02/26/2022 10:38:20 - INFO - codeparrot_training - Step 44057: {'lr': 1.8674811329058378e-05, 'samples': 22557696, 'steps': 44057, 'loss/train': 1.7107274532318115} 02/26/2022 10:38:25 - INFO - codeparrot_training - Step 44058: {'lr': 1.8668606620841326e-05, 'samples': 22558208, 'steps': 44058, 'loss/train': 1.6024490594863892} 02/26/2022 10:38:29 - INFO - codeparrot_training - Step 44059: {'lr': 1.8662402903574516e-05, 'samples': 22558720, 'steps': 44059, 'loss/train': 0.5669743418693542} 02/26/2022 10:38:34 - INFO - codeparrot_training - Step 44060: {'lr': 1.8656200177284503e-05, 'samples': 22559232, 'steps': 44060, 'loss/train': 1.5877149105072021} 02/26/2022 10:38:38 - INFO - codeparrot_training - Step 44061: {'lr': 1.8649998441997856e-05, 'samples': 22559744, 'steps': 44061, 'loss/train': 1.8641678094863892} 02/26/2022 10:38:43 - INFO - codeparrot_training - Step 44062: {'lr': 1.8643797697741162e-05, 'samples': 22560256, 'steps': 44062, 'loss/train': 1.7686386108398438} 02/26/2022 10:38:47 - INFO - codeparrot_training - Step 44063: {'lr': 1.8637597944540952e-05, 'samples': 22560768, 'steps': 44063, 'loss/train': 1.1887909173965454} 02/26/2022 10:38:52 - INFO - codeparrot_training - Step 44064: {'lr': 1.8631399182423876e-05, 'samples': 22561280, 'steps': 44064, 'loss/train': 1.5086917877197266} 02/26/2022 10:38:56 - INFO - codeparrot_training - Step 44065: {'lr': 1.8625201411416325e-05, 'samples': 22561792, 'steps': 44065, 'loss/train': 1.4407860040664673} 02/26/2022 10:39:03 - INFO - codeparrot_training - Step 44066: {'lr': 1.8619004631545004e-05, 'samples': 22562304, 'steps': 44066, 'loss/train': 1.3726171255111694} 02/26/2022 10:39:07 - INFO - codeparrot_training - Step 44067: {'lr': 1.8612808842836337e-05, 'samples': 22562816, 'steps': 44067, 'loss/train': 1.6653813123703003} 02/26/2022 10:39:12 - INFO - codeparrot_training - Step 44068: {'lr': 1.8606614045317022e-05, 'samples': 22563328, 'steps': 44068, 'loss/train': 2.2861428260803223} 02/26/2022 10:39:16 - INFO - codeparrot_training - Step 44069: {'lr': 1.86004202390134e-05, 'samples': 22563840, 'steps': 44069, 'loss/train': 2.1283395290374756} 02/26/2022 10:39:21 - INFO - codeparrot_training - Step 44070: {'lr': 1.859422742395217e-05, 'samples': 22564352, 'steps': 44070, 'loss/train': 1.6871061325073242} 02/26/2022 10:39:25 - INFO - codeparrot_training - Step 44071: {'lr': 1.858803560015973e-05, 'samples': 22564864, 'steps': 44071, 'loss/train': 2.382408380508423} 02/26/2022 10:39:30 - INFO - codeparrot_training - Step 44072: {'lr': 1.8581844767662727e-05, 'samples': 22565376, 'steps': 44072, 'loss/train': 1.5408034324645996} 02/26/2022 10:39:34 - INFO - codeparrot_training - Step 44073: {'lr': 1.857565492648755e-05, 'samples': 22565888, 'steps': 44073, 'loss/train': 1.791016936302185} 02/26/2022 10:39:39 - INFO - codeparrot_training - Step 44074: {'lr': 1.8569466076660825e-05, 'samples': 22566400, 'steps': 44074, 'loss/train': 1.8555350303649902} 02/26/2022 10:39:43 - INFO - codeparrot_training - Step 44075: {'lr': 1.8563278218208995e-05, 'samples': 22566912, 'steps': 44075, 'loss/train': 1.4032213687896729} 02/26/2022 10:39:50 - INFO - codeparrot_training - Step 44076: {'lr': 1.85570913511586e-05, 'samples': 22567424, 'steps': 44076, 'loss/train': 3.2220664024353027} 02/26/2022 10:39:53 - INFO - codeparrot_training - Step 44077: {'lr': 1.855090547553617e-05, 'samples': 22567936, 'steps': 44077, 'loss/train': 2.0979886054992676} 02/26/2022 10:39:59 - INFO - codeparrot_training - Step 44078: {'lr': 1.8544720591368134e-05, 'samples': 22568448, 'steps': 44078, 'loss/train': 1.9048371315002441} 02/26/2022 10:40:02 - INFO - codeparrot_training - Step 44079: {'lr': 1.8538536698680995e-05, 'samples': 22568960, 'steps': 44079, 'loss/train': 2.847787380218506} 02/26/2022 10:40:08 - INFO - codeparrot_training - Step 44080: {'lr': 1.8532353797501317e-05, 'samples': 22569472, 'steps': 44080, 'loss/train': 1.921404480934143} 02/26/2022 10:40:12 - INFO - codeparrot_training - Step 44081: {'lr': 1.852617188785552e-05, 'samples': 22569984, 'steps': 44081, 'loss/train': 1.736572504043579} 02/26/2022 10:40:17 - INFO - codeparrot_training - Step 44082: {'lr': 1.8519990969770113e-05, 'samples': 22570496, 'steps': 44082, 'loss/train': 1.7315726280212402} 02/26/2022 10:40:20 - INFO - codeparrot_training - Step 44083: {'lr': 1.8513811043271517e-05, 'samples': 22571008, 'steps': 44083, 'loss/train': 1.828129768371582} 02/26/2022 10:40:26 - INFO - codeparrot_training - Step 44084: {'lr': 1.8507632108386268e-05, 'samples': 22571520, 'steps': 44084, 'loss/train': 1.5297503471374512} 02/26/2022 10:40:29 - INFO - codeparrot_training - Step 44085: {'lr': 1.8501454165140847e-05, 'samples': 22572032, 'steps': 44085, 'loss/train': 1.3314677476882935} 02/26/2022 10:40:37 - INFO - codeparrot_training - Step 44086: {'lr': 1.8495277213561674e-05, 'samples': 22572544, 'steps': 44086, 'loss/train': 0.475856214761734} 02/26/2022 10:40:40 - INFO - codeparrot_training - Step 44087: {'lr': 1.84891012536752e-05, 'samples': 22573056, 'steps': 44087, 'loss/train': 0.9418395757675171} 02/26/2022 10:40:46 - INFO - codeparrot_training - Step 44088: {'lr': 1.8482926285507905e-05, 'samples': 22573568, 'steps': 44088, 'loss/train': 4.145501136779785} 02/26/2022 10:40:49 - INFO - codeparrot_training - Step 44089: {'lr': 1.847675230908627e-05, 'samples': 22574080, 'steps': 44089, 'loss/train': 3.4087753295898438} 02/26/2022 10:40:55 - INFO - codeparrot_training - Step 44090: {'lr': 1.8470579324436714e-05, 'samples': 22574592, 'steps': 44090, 'loss/train': 1.4923803806304932} 02/26/2022 10:40:58 - INFO - codeparrot_training - Step 44091: {'lr': 1.8464407331585665e-05, 'samples': 22575104, 'steps': 44091, 'loss/train': 1.3637189865112305} 02/26/2022 10:41:04 - INFO - codeparrot_training - Step 44092: {'lr': 1.845823633055954e-05, 'samples': 22575616, 'steps': 44092, 'loss/train': 0.42889466881752014} 02/26/2022 10:41:07 - INFO - codeparrot_training - Step 44093: {'lr': 1.8452066321384825e-05, 'samples': 22576128, 'steps': 44093, 'loss/train': 0.9857776165008545} 02/26/2022 10:41:13 - INFO - codeparrot_training - Step 44094: {'lr': 1.8445897304087966e-05, 'samples': 22576640, 'steps': 44094, 'loss/train': 1.7164846658706665} 02/26/2022 10:41:17 - INFO - codeparrot_training - Step 44095: {'lr': 1.8439729278695332e-05, 'samples': 22577152, 'steps': 44095, 'loss/train': 1.7889143228530884} 02/26/2022 10:41:24 - INFO - codeparrot_training - Step 44096: {'lr': 1.843356224523335e-05, 'samples': 22577664, 'steps': 44096, 'loss/train': 1.4949747323989868} 02/26/2022 10:41:27 - INFO - codeparrot_training - Step 44097: {'lr': 1.8427396203728464e-05, 'samples': 22578176, 'steps': 44097, 'loss/train': 1.7959195375442505} 02/26/2022 10:41:33 - INFO - codeparrot_training - Step 44098: {'lr': 1.8421231154207075e-05, 'samples': 22578688, 'steps': 44098, 'loss/train': 2.374171257019043} 02/26/2022 10:41:36 - INFO - codeparrot_training - Step 44099: {'lr': 1.841506709669566e-05, 'samples': 22579200, 'steps': 44099, 'loss/train': 2.2732620239257812} 02/26/2022 10:41:42 - INFO - codeparrot_training - Step 44100: {'lr': 1.8408904031220476e-05, 'samples': 22579712, 'steps': 44100, 'loss/train': 2.139552354812622} 02/26/2022 10:41:45 - INFO - codeparrot_training - Step 44101: {'lr': 1.8402741957808056e-05, 'samples': 22580224, 'steps': 44101, 'loss/train': 2.0501139163970947} 02/26/2022 10:41:51 - INFO - codeparrot_training - Step 44102: {'lr': 1.8396580876484714e-05, 'samples': 22580736, 'steps': 44102, 'loss/train': 2.0415778160095215} 02/26/2022 10:41:54 - INFO - codeparrot_training - Step 44103: {'lr': 1.8390420787276955e-05, 'samples': 22581248, 'steps': 44103, 'loss/train': 1.4672152996063232} 02/26/2022 10:42:00 - INFO - codeparrot_training - Step 44104: {'lr': 1.8384261690211014e-05, 'samples': 22581760, 'steps': 44104, 'loss/train': 2.853710174560547} 02/26/2022 10:42:03 - INFO - codeparrot_training - Step 44105: {'lr': 1.8378103585313388e-05, 'samples': 22582272, 'steps': 44105, 'loss/train': 1.8300520181655884} 02/26/2022 10:42:09 - INFO - codeparrot_training - Step 44106: {'lr': 1.837194647261037e-05, 'samples': 22582784, 'steps': 44106, 'loss/train': 1.7034151554107666} 02/26/2022 10:42:12 - INFO - codeparrot_training - Step 44107: {'lr': 1.8365790352128458e-05, 'samples': 22583296, 'steps': 44107, 'loss/train': 8.701147079467773} 02/26/2022 10:42:18 - INFO - codeparrot_training - Step 44108: {'lr': 1.835963522389389e-05, 'samples': 22583808, 'steps': 44108, 'loss/train': 1.4534105062484741} 02/26/2022 10:42:21 - INFO - codeparrot_training - Step 44109: {'lr': 1.835348108793311e-05, 'samples': 22584320, 'steps': 44109, 'loss/train': 2.158106565475464} 02/26/2022 10:42:27 - INFO - codeparrot_training - Step 44110: {'lr': 1.8347327944272435e-05, 'samples': 22584832, 'steps': 44110, 'loss/train': 1.9505138397216797} 02/26/2022 10:42:30 - INFO - codeparrot_training - Step 44111: {'lr': 1.8341175792938287e-05, 'samples': 22585344, 'steps': 44111, 'loss/train': 0.720163881778717} 02/26/2022 10:42:38 - INFO - codeparrot_training - Step 44112: {'lr': 1.8335024633956977e-05, 'samples': 22585856, 'steps': 44112, 'loss/train': 1.6401547193527222} 02/26/2022 10:42:41 - INFO - codeparrot_training - Step 44113: {'lr': 1.8328874467354846e-05, 'samples': 22586368, 'steps': 44113, 'loss/train': 2.9826443195343018} 02/26/2022 10:42:47 - INFO - codeparrot_training - Step 44114: {'lr': 1.8322725293158233e-05, 'samples': 22586880, 'steps': 44114, 'loss/train': 1.5513803958892822} 02/26/2022 10:42:50 - INFO - codeparrot_training - Step 44115: {'lr': 1.8316577111393534e-05, 'samples': 22587392, 'steps': 44115, 'loss/train': 1.8280835151672363} 02/26/2022 10:42:56 - INFO - codeparrot_training - Step 44116: {'lr': 1.8310429922087035e-05, 'samples': 22587904, 'steps': 44116, 'loss/train': 3.183056592941284} 02/26/2022 10:42:59 - INFO - codeparrot_training - Step 44117: {'lr': 1.8304283725265104e-05, 'samples': 22588416, 'steps': 44117, 'loss/train': 2.2480101585388184} 02/26/2022 10:43:05 - INFO - codeparrot_training - Step 44118: {'lr': 1.8298138520953995e-05, 'samples': 22588928, 'steps': 44118, 'loss/train': 1.4812006950378418} 02/26/2022 10:43:08 - INFO - codeparrot_training - Step 44119: {'lr': 1.829199430918016e-05, 'samples': 22589440, 'steps': 44119, 'loss/train': 2.079787015914917} 02/26/2022 10:43:14 - INFO - codeparrot_training - Step 44120: {'lr': 1.8285851089969803e-05, 'samples': 22589952, 'steps': 44120, 'loss/train': 2.3262670040130615} 02/26/2022 10:43:17 - INFO - codeparrot_training - Step 44121: {'lr': 1.8279708863349315e-05, 'samples': 22590464, 'steps': 44121, 'loss/train': 2.307448387145996} 02/26/2022 10:43:24 - INFO - codeparrot_training - Step 44122: {'lr': 1.8273567629344928e-05, 'samples': 22590976, 'steps': 44122, 'loss/train': 3.5219831466674805} 02/26/2022 10:43:28 - INFO - codeparrot_training - Step 44123: {'lr': 1.826742738798301e-05, 'samples': 22591488, 'steps': 44123, 'loss/train': 2.1535136699676514} 02/26/2022 10:43:33 - INFO - codeparrot_training - Step 44124: {'lr': 1.8261288139289874e-05, 'samples': 22592000, 'steps': 44124, 'loss/train': 1.588499903678894} 02/26/2022 10:43:37 - INFO - codeparrot_training - Step 44125: {'lr': 1.82551498832918e-05, 'samples': 22592512, 'steps': 44125, 'loss/train': 2.4162540435791016} 02/26/2022 10:43:42 - INFO - codeparrot_training - Step 44126: {'lr': 1.824901262001502e-05, 'samples': 22593024, 'steps': 44126, 'loss/train': 2.277219772338867} 02/26/2022 10:43:46 - INFO - codeparrot_training - Step 44127: {'lr': 1.8242876349485932e-05, 'samples': 22593536, 'steps': 44127, 'loss/train': 1.1755133867263794} 02/26/2022 10:43:51 - INFO - codeparrot_training - Step 44128: {'lr': 1.823674107173076e-05, 'samples': 22594048, 'steps': 44128, 'loss/train': 0.45689237117767334} 02/26/2022 10:43:55 - INFO - codeparrot_training - Step 44129: {'lr': 1.8230606786775793e-05, 'samples': 22594560, 'steps': 44129, 'loss/train': 2.6282219886779785} 02/26/2022 10:44:01 - INFO - codeparrot_training - Step 44130: {'lr': 1.8224473494647286e-05, 'samples': 22595072, 'steps': 44130, 'loss/train': 1.2036617994308472} 02/26/2022 10:44:04 - INFO - codeparrot_training - Step 44131: {'lr': 1.8218341195371578e-05, 'samples': 22595584, 'steps': 44131, 'loss/train': 1.94817316532135} 02/26/2022 10:44:11 - INFO - codeparrot_training - Step 44132: {'lr': 1.8212209888974874e-05, 'samples': 22596096, 'steps': 44132, 'loss/train': 3.273123264312744} 02/26/2022 10:44:15 - INFO - codeparrot_training - Step 44133: {'lr': 1.8206079575483482e-05, 'samples': 22596608, 'steps': 44133, 'loss/train': 0.9376332759857178} 02/26/2022 10:44:20 - INFO - codeparrot_training - Step 44134: {'lr': 1.8199950254923603e-05, 'samples': 22597120, 'steps': 44134, 'loss/train': 2.055082321166992} 02/26/2022 10:44:24 - INFO - codeparrot_training - Step 44135: {'lr': 1.819382192732158e-05, 'samples': 22597632, 'steps': 44135, 'loss/train': 2.9164419174194336} 02/26/2022 10:44:29 - INFO - codeparrot_training - Step 44136: {'lr': 1.8187694592703586e-05, 'samples': 22598144, 'steps': 44136, 'loss/train': 1.5746794939041138} 02/26/2022 10:44:33 - INFO - codeparrot_training - Step 44137: {'lr': 1.8181568251095875e-05, 'samples': 22598656, 'steps': 44137, 'loss/train': 2.1007883548736572} 02/26/2022 10:44:38 - INFO - codeparrot_training - Step 44138: {'lr': 1.817544290252482e-05, 'samples': 22599168, 'steps': 44138, 'loss/train': 1.6616401672363281} 02/26/2022 10:44:42 - INFO - codeparrot_training - Step 44139: {'lr': 1.816931854701645e-05, 'samples': 22599680, 'steps': 44139, 'loss/train': 3.0018463134765625} 02/26/2022 10:44:47 - INFO - codeparrot_training - Step 44140: {'lr': 1.8163195184597138e-05, 'samples': 22600192, 'steps': 44140, 'loss/train': 1.820244550704956} 02/26/2022 10:44:51 - INFO - codeparrot_training - Step 44141: {'lr': 1.8157072815293053e-05, 'samples': 22600704, 'steps': 44141, 'loss/train': 1.670430064201355} 02/26/2022 10:44:58 - INFO - codeparrot_training - Step 44142: {'lr': 1.8150951439130535e-05, 'samples': 22601216, 'steps': 44142, 'loss/train': 0.9679425954818726} 02/26/2022 10:45:01 - INFO - codeparrot_training - Step 44143: {'lr': 1.8144831056135626e-05, 'samples': 22601728, 'steps': 44143, 'loss/train': 1.1052898168563843} 02/26/2022 10:45:07 - INFO - codeparrot_training - Step 44144: {'lr': 1.8138711666334683e-05, 'samples': 22602240, 'steps': 44144, 'loss/train': 1.3196940422058105} 02/26/2022 10:45:10 - INFO - codeparrot_training - Step 44145: {'lr': 1.8132593269753862e-05, 'samples': 22602752, 'steps': 44145, 'loss/train': 2.3234992027282715} 02/26/2022 10:45:16 - INFO - codeparrot_training - Step 44146: {'lr': 1.8126475866419385e-05, 'samples': 22603264, 'steps': 44146, 'loss/train': 1.489924430847168} 02/26/2022 10:45:19 - INFO - codeparrot_training - Step 44147: {'lr': 1.8120359456357484e-05, 'samples': 22603776, 'steps': 44147, 'loss/train': 1.3693950176239014} 02/26/2022 10:45:25 - INFO - codeparrot_training - Step 44148: {'lr': 1.811424403959433e-05, 'samples': 22604288, 'steps': 44148, 'loss/train': 1.1188582181930542} 02/26/2022 10:45:28 - INFO - codeparrot_training - Step 44149: {'lr': 1.8108129616156072e-05, 'samples': 22604800, 'steps': 44149, 'loss/train': 1.974057912826538} 02/26/2022 10:45:34 - INFO - codeparrot_training - Step 44150: {'lr': 1.8102016186068992e-05, 'samples': 22605312, 'steps': 44150, 'loss/train': 1.701819896697998} 02/26/2022 10:45:37 - INFO - codeparrot_training - Step 44151: {'lr': 1.809590374935924e-05, 'samples': 22605824, 'steps': 44151, 'loss/train': 1.7709394693374634} 02/26/2022 10:45:43 - INFO - codeparrot_training - Step 44152: {'lr': 1.808979230605301e-05, 'samples': 22606336, 'steps': 44152, 'loss/train': 1.6412721872329712} 02/26/2022 10:45:48 - INFO - codeparrot_training - Step 44153: {'lr': 1.80836818561764e-05, 'samples': 22606848, 'steps': 44153, 'loss/train': 1.0612002611160278} 02/26/2022 10:45:52 - INFO - codeparrot_training - Step 44154: {'lr': 1.8077572399755717e-05, 'samples': 22607360, 'steps': 44154, 'loss/train': 2.0660037994384766} 02/26/2022 10:45:55 - INFO - codeparrot_training - Step 44155: {'lr': 1.8071463936817078e-05, 'samples': 22607872, 'steps': 44155, 'loss/train': 1.5193873643875122} 02/26/2022 10:46:01 - INFO - codeparrot_training - Step 44156: {'lr': 1.8065356467386635e-05, 'samples': 22608384, 'steps': 44156, 'loss/train': 1.4883440732955933} 02/26/2022 10:46:04 - INFO - codeparrot_training - Step 44157: {'lr': 1.805924999149053e-05, 'samples': 22608896, 'steps': 44157, 'loss/train': 1.6660815477371216} 02/26/2022 10:46:12 - INFO - codeparrot_training - Step 44158: {'lr': 1.8053144509154963e-05, 'samples': 22609408, 'steps': 44158, 'loss/train': 1.0225225687026978} 02/26/2022 10:46:15 - INFO - codeparrot_training - Step 44159: {'lr': 1.804704002040605e-05, 'samples': 22609920, 'steps': 44159, 'loss/train': 2.331615686416626} 02/26/2022 10:46:21 - INFO - codeparrot_training - Step 44160: {'lr': 1.804093652527003e-05, 'samples': 22610432, 'steps': 44160, 'loss/train': 0.13454057276248932} 02/26/2022 10:46:26 - INFO - codeparrot_training - Step 44161: {'lr': 1.8034834023772927e-05, 'samples': 22610944, 'steps': 44161, 'loss/train': 1.656209945678711} 02/26/2022 10:46:30 - INFO - codeparrot_training - Step 44162: {'lr': 1.8028732515940948e-05, 'samples': 22611456, 'steps': 44162, 'loss/train': 1.6359809637069702} 02/26/2022 10:46:35 - INFO - codeparrot_training - Step 44163: {'lr': 1.802263200180021e-05, 'samples': 22611968, 'steps': 44163, 'loss/train': 1.8141642808914185} 02/26/2022 10:46:39 - INFO - codeparrot_training - Step 44164: {'lr': 1.801653248137691e-05, 'samples': 22612480, 'steps': 44164, 'loss/train': 0.5752533078193665} 02/26/2022 10:46:42 - INFO - codeparrot_training - Step 44165: {'lr': 1.8010433954697065e-05, 'samples': 22612992, 'steps': 44165, 'loss/train': 0.09879948198795319} 02/26/2022 10:46:48 - INFO - codeparrot_training - Step 44166: {'lr': 1.8004336421786864e-05, 'samples': 22613504, 'steps': 44166, 'loss/train': 2.1428003311157227} 02/26/2022 10:46:55 - INFO - codeparrot_training - Step 44167: {'lr': 1.7998239882672407e-05, 'samples': 22614016, 'steps': 44167, 'loss/train': 1.5720183849334717} 02/26/2022 10:46:59 - INFO - codeparrot_training - Step 44168: {'lr': 1.799214433737989e-05, 'samples': 22614528, 'steps': 44168, 'loss/train': 1.9422537088394165} 02/26/2022 10:47:04 - INFO - codeparrot_training - Step 44169: {'lr': 1.7986049785935266e-05, 'samples': 22615040, 'steps': 44169, 'loss/train': 1.6239545345306396} 02/26/2022 10:47:07 - INFO - codeparrot_training - Step 44170: {'lr': 1.7979956228364792e-05, 'samples': 22615552, 'steps': 44170, 'loss/train': 1.8635287284851074} 02/26/2022 10:47:13 - INFO - codeparrot_training - Step 44171: {'lr': 1.7973863664694502e-05, 'samples': 22616064, 'steps': 44171, 'loss/train': 1.7481529712677002} 02/26/2022 10:47:17 - INFO - codeparrot_training - Step 44172: {'lr': 1.7967772094950456e-05, 'samples': 22616576, 'steps': 44172, 'loss/train': 1.626937985420227} 02/26/2022 10:47:20 - INFO - codeparrot_training - Step 44173: {'lr': 1.7961681519158863e-05, 'samples': 22617088, 'steps': 44173, 'loss/train': 1.8340483903884888} 02/26/2022 10:47:26 - INFO - codeparrot_training - Step 44174: {'lr': 1.795559193734572e-05, 'samples': 22617600, 'steps': 44174, 'loss/train': 2.196171283721924} 02/26/2022 10:47:29 - INFO - codeparrot_training - Step 44175: {'lr': 1.7949503349537154e-05, 'samples': 22618112, 'steps': 44175, 'loss/train': 1.8305608034133911} 02/26/2022 10:47:35 - INFO - codeparrot_training - Step 44176: {'lr': 1.7943415755759167e-05, 'samples': 22618624, 'steps': 44176, 'loss/train': 0.08788692206144333} 02/26/2022 10:47:42 - INFO - codeparrot_training - Step 44177: {'lr': 1.793732915603796e-05, 'samples': 22619136, 'steps': 44177, 'loss/train': 2.0525529384613037} 02/26/2022 10:47:45 - INFO - codeparrot_training - Step 44178: {'lr': 1.7931243550399546e-05, 'samples': 22619648, 'steps': 44178, 'loss/train': 1.7399303913116455} 02/26/2022 10:47:51 - INFO - codeparrot_training - Step 44179: {'lr': 1.7925158938870005e-05, 'samples': 22620160, 'steps': 44179, 'loss/train': 1.8708152770996094} 02/26/2022 10:47:54 - INFO - codeparrot_training - Step 44180: {'lr': 1.7919075321475327e-05, 'samples': 22620672, 'steps': 44180, 'loss/train': 2.30462646484375} 02/26/2022 10:48:00 - INFO - codeparrot_training - Step 44181: {'lr': 1.7912992698241677e-05, 'samples': 22621184, 'steps': 44181, 'loss/train': 1.6517940759658813} 02/26/2022 10:48:03 - INFO - codeparrot_training - Step 44182: {'lr': 1.7906911069195094e-05, 'samples': 22621696, 'steps': 44182, 'loss/train': 1.805814504623413} 02/26/2022 10:48:09 - INFO - codeparrot_training - Step 44183: {'lr': 1.790083043436158e-05, 'samples': 22622208, 'steps': 44183, 'loss/train': 2.0057175159454346} 02/26/2022 10:48:12 - INFO - codeparrot_training - Step 44184: {'lr': 1.789475079376718e-05, 'samples': 22622720, 'steps': 44184, 'loss/train': 1.5925467014312744} 02/26/2022 10:48:18 - INFO - codeparrot_training - Step 44185: {'lr': 1.7888672147438028e-05, 'samples': 22623232, 'steps': 44185, 'loss/train': 2.3879640102386475} 02/26/2022 10:48:21 - INFO - codeparrot_training - Step 44186: {'lr': 1.7882594495400058e-05, 'samples': 22623744, 'steps': 44186, 'loss/train': 1.004015326499939} 02/26/2022 10:48:28 - INFO - codeparrot_training - Step 44187: {'lr': 1.787651783767938e-05, 'samples': 22624256, 'steps': 44187, 'loss/train': 2.11611270904541} 02/26/2022 10:48:32 - INFO - codeparrot_training - Step 44188: {'lr': 1.7870442174301948e-05, 'samples': 22624768, 'steps': 44188, 'loss/train': 0.9738661050796509} 02/26/2022 10:48:37 - INFO - codeparrot_training - Step 44189: {'lr': 1.7864367505293853e-05, 'samples': 22625280, 'steps': 44189, 'loss/train': 2.6405553817749023} 02/26/2022 10:48:41 - INFO - codeparrot_training - Step 44190: {'lr': 1.7858293830681104e-05, 'samples': 22625792, 'steps': 44190, 'loss/train': 2.2659103870391846} 02/26/2022 10:48:46 - INFO - codeparrot_training - Step 44191: {'lr': 1.7852221150489734e-05, 'samples': 22626304, 'steps': 44191, 'loss/train': 2.840360641479492} 02/26/2022 10:48:50 - INFO - codeparrot_training - Step 44192: {'lr': 1.7846149464745666e-05, 'samples': 22626816, 'steps': 44192, 'loss/train': 1.9779052734375} 02/26/2022 10:48:56 - INFO - codeparrot_training - Step 44193: {'lr': 1.7840078773475016e-05, 'samples': 22627328, 'steps': 44193, 'loss/train': 2.1348936557769775} 02/26/2022 10:48:59 - INFO - codeparrot_training - Step 44194: {'lr': 1.783400907670374e-05, 'samples': 22627840, 'steps': 44194, 'loss/train': 2.149169683456421} 02/26/2022 10:49:04 - INFO - codeparrot_training - Step 44195: {'lr': 1.7827940374457896e-05, 'samples': 22628352, 'steps': 44195, 'loss/train': 1.5813424587249756} 02/26/2022 10:49:08 - INFO - codeparrot_training - Step 44196: {'lr': 1.7821872666763356e-05, 'samples': 22628864, 'steps': 44196, 'loss/train': 1.56320059299469} 02/26/2022 10:49:14 - INFO - codeparrot_training - Step 44197: {'lr': 1.7815805953646237e-05, 'samples': 22629376, 'steps': 44197, 'loss/train': 0.8962571024894714} 02/26/2022 10:49:17 - INFO - codeparrot_training - Step 44198: {'lr': 1.780974023513246e-05, 'samples': 22629888, 'steps': 44198, 'loss/train': 1.8402847051620483} 02/26/2022 10:49:23 - INFO - codeparrot_training - Step 44199: {'lr': 1.780367551124809e-05, 'samples': 22630400, 'steps': 44199, 'loss/train': 2.4448423385620117} 02/26/2022 10:49:26 - INFO - codeparrot_training - Step 44200: {'lr': 1.779761178201894e-05, 'samples': 22630912, 'steps': 44200, 'loss/train': 2.2313296794891357} 02/26/2022 10:49:32 - INFO - codeparrot_training - Step 44201: {'lr': 1.7791549047471157e-05, 'samples': 22631424, 'steps': 44201, 'loss/train': 1.9121071100234985} 02/26/2022 10:49:35 - INFO - codeparrot_training - Step 44202: {'lr': 1.7785487307630578e-05, 'samples': 22631936, 'steps': 44202, 'loss/train': 1.7731974124908447} 02/26/2022 10:49:43 - INFO - codeparrot_training - Step 44203: {'lr': 1.7779426562523348e-05, 'samples': 22632448, 'steps': 44203, 'loss/train': 2.4096925258636475} 02/26/2022 10:49:47 - INFO - codeparrot_training - Step 44204: {'lr': 1.7773366812175202e-05, 'samples': 22632960, 'steps': 44204, 'loss/train': 1.598775029182434} 02/26/2022 10:49:52 - INFO - codeparrot_training - Step 44205: {'lr': 1.776730805661228e-05, 'samples': 22633472, 'steps': 44205, 'loss/train': 2.3445546627044678} 02/26/2022 10:49:55 - INFO - codeparrot_training - Step 44206: {'lr': 1.7761250295860425e-05, 'samples': 22633984, 'steps': 44206, 'loss/train': 1.988905429840088} 02/26/2022 10:50:01 - INFO - codeparrot_training - Step 44207: {'lr': 1.7755193529945645e-05, 'samples': 22634496, 'steps': 44207, 'loss/train': 1.450088620185852} 02/26/2022 10:50:04 - INFO - codeparrot_training - Step 44208: {'lr': 1.774913775889389e-05, 'samples': 22635008, 'steps': 44208, 'loss/train': 1.1929043531417847} 02/26/2022 10:50:10 - INFO - codeparrot_training - Step 44209: {'lr': 1.7743082982731085e-05, 'samples': 22635520, 'steps': 44209, 'loss/train': 2.1798148155212402} 02/26/2022 10:50:13 - INFO - codeparrot_training - Step 44210: {'lr': 1.7737029201483123e-05, 'samples': 22636032, 'steps': 44210, 'loss/train': 1.9419649839401245} 02/26/2022 10:50:19 - INFO - codeparrot_training - Step 44211: {'lr': 1.7730976415176016e-05, 'samples': 22636544, 'steps': 44211, 'loss/train': 2.2358546257019043} 02/26/2022 10:50:22 - INFO - codeparrot_training - Step 44212: {'lr': 1.7724924623835625e-05, 'samples': 22637056, 'steps': 44212, 'loss/train': 1.6401432752609253} 02/26/2022 10:50:30 - INFO - codeparrot_training - Step 44213: {'lr': 1.7718873827487937e-05, 'samples': 22637568, 'steps': 44213, 'loss/train': 1.1242268085479736} 02/26/2022 10:50:33 - INFO - codeparrot_training - Step 44214: {'lr': 1.771282402615876e-05, 'samples': 22638080, 'steps': 44214, 'loss/train': 0.7484341263771057} 02/26/2022 10:50:39 - INFO - codeparrot_training - Step 44215: {'lr': 1.7706775219874154e-05, 'samples': 22638592, 'steps': 44215, 'loss/train': 2.0688838958740234} 02/26/2022 10:50:42 - INFO - codeparrot_training - Step 44216: {'lr': 1.770072740865997e-05, 'samples': 22639104, 'steps': 44216, 'loss/train': 0.730161726474762} 02/26/2022 10:50:48 - INFO - codeparrot_training - Step 44217: {'lr': 1.769468059254206e-05, 'samples': 22639616, 'steps': 44217, 'loss/train': 2.212062120437622} 02/26/2022 10:50:51 - INFO - codeparrot_training - Step 44218: {'lr': 1.768863477154642e-05, 'samples': 22640128, 'steps': 44218, 'loss/train': 2.183596611022949} 02/26/2022 10:50:57 - INFO - codeparrot_training - Step 44219: {'lr': 1.768258994569885e-05, 'samples': 22640640, 'steps': 44219, 'loss/train': 0.9215399622917175} 02/26/2022 10:51:00 - INFO - codeparrot_training - Step 44220: {'lr': 1.767654611502531e-05, 'samples': 22641152, 'steps': 44220, 'loss/train': 1.0041439533233643} 02/26/2022 10:51:06 - INFO - codeparrot_training - Step 44221: {'lr': 1.7670503279551692e-05, 'samples': 22641664, 'steps': 44221, 'loss/train': 2.5924010276794434} 02/26/2022 10:51:09 - INFO - codeparrot_training - Step 44222: {'lr': 1.766446143930389e-05, 'samples': 22642176, 'steps': 44222, 'loss/train': 2.2745039463043213} 02/26/2022 10:51:15 - INFO - codeparrot_training - Step 44223: {'lr': 1.7658420594307696e-05, 'samples': 22642688, 'steps': 44223, 'loss/train': 2.3978307247161865} 02/26/2022 10:51:19 - INFO - codeparrot_training - Step 44224: {'lr': 1.7652380744589082e-05, 'samples': 22643200, 'steps': 44224, 'loss/train': 1.124882698059082} 02/26/2022 10:51:26 - INFO - codeparrot_training - Step 44225: {'lr': 1.7646341890173917e-05, 'samples': 22643712, 'steps': 44225, 'loss/train': 1.7671010494232178} 02/26/2022 10:51:29 - INFO - codeparrot_training - Step 44226: {'lr': 1.7640304031088015e-05, 'samples': 22644224, 'steps': 44226, 'loss/train': 0.977418839931488} 02/26/2022 10:51:34 - INFO - codeparrot_training - Step 44227: {'lr': 1.763426716735725e-05, 'samples': 22644736, 'steps': 44227, 'loss/train': 0.3391740322113037} 02/26/2022 10:51:38 - INFO - codeparrot_training - Step 44228: {'lr': 1.7628231299007536e-05, 'samples': 22645248, 'steps': 44228, 'loss/train': 2.1769115924835205} 02/26/2022 10:51:43 - INFO - codeparrot_training - Step 44229: {'lr': 1.762219642606466e-05, 'samples': 22645760, 'steps': 44229, 'loss/train': 1.501988172531128} 02/26/2022 10:51:47 - INFO - codeparrot_training - Step 44230: {'lr': 1.7616162548554554e-05, 'samples': 22646272, 'steps': 44230, 'loss/train': 2.473156213760376} 02/26/2022 10:51:52 - INFO - codeparrot_training - Step 44231: {'lr': 1.761012966650294e-05, 'samples': 22646784, 'steps': 44231, 'loss/train': 1.0713082551956177} 02/26/2022 10:51:56 - INFO - codeparrot_training - Step 44232: {'lr': 1.7604097779935796e-05, 'samples': 22647296, 'steps': 44232, 'loss/train': 2.0067172050476074} 02/26/2022 10:52:01 - INFO - codeparrot_training - Step 44233: {'lr': 1.7598066888878883e-05, 'samples': 22647808, 'steps': 44233, 'loss/train': 1.6761754751205444} 02/26/2022 10:52:05 - INFO - codeparrot_training - Step 44234: {'lr': 1.7592036993358094e-05, 'samples': 22648320, 'steps': 44234, 'loss/train': 1.9190446138381958} 02/26/2022 10:52:12 - INFO - codeparrot_training - Step 44235: {'lr': 1.7586008093399165e-05, 'samples': 22648832, 'steps': 44235, 'loss/train': 2.0201337337493896} 02/26/2022 10:52:15 - INFO - codeparrot_training - Step 44236: {'lr': 1.757998018902804e-05, 'samples': 22649344, 'steps': 44236, 'loss/train': 1.665717363357544} 02/26/2022 10:52:21 - INFO - codeparrot_training - Step 44237: {'lr': 1.7573953280270395e-05, 'samples': 22649856, 'steps': 44237, 'loss/train': 1.8131917715072632} 02/26/2022 10:52:25 - INFO - codeparrot_training - Step 44238: {'lr': 1.7567927367152237e-05, 'samples': 22650368, 'steps': 44238, 'loss/train': 2.357252359390259} 02/26/2022 10:52:30 - INFO - codeparrot_training - Step 44239: {'lr': 1.7561902449699184e-05, 'samples': 22650880, 'steps': 44239, 'loss/train': 1.7787820100784302} 02/26/2022 10:52:34 - INFO - codeparrot_training - Step 44240: {'lr': 1.7555878527937163e-05, 'samples': 22651392, 'steps': 44240, 'loss/train': 2.1327555179595947} 02/26/2022 10:52:39 - INFO - codeparrot_training - Step 44241: {'lr': 1.754985560189193e-05, 'samples': 22651904, 'steps': 44241, 'loss/train': 0.6350491642951965} 02/26/2022 10:52:43 - INFO - codeparrot_training - Step 44242: {'lr': 1.754383367158932e-05, 'samples': 22652416, 'steps': 44242, 'loss/train': 1.1296526193618774} 02/26/2022 10:52:48 - INFO - codeparrot_training - Step 44243: {'lr': 1.7537812737055124e-05, 'samples': 22652928, 'steps': 44243, 'loss/train': 1.479059100151062} 02/26/2022 10:52:52 - INFO - codeparrot_training - Step 44244: {'lr': 1.7531792798315128e-05, 'samples': 22653440, 'steps': 44244, 'loss/train': 1.8070179224014282} 02/26/2022 10:52:57 - INFO - codeparrot_training - Step 44245: {'lr': 1.7525773855395084e-05, 'samples': 22653952, 'steps': 44245, 'loss/train': 2.6320652961730957} 02/26/2022 10:53:01 - INFO - codeparrot_training - Step 44246: {'lr': 1.7519755908320834e-05, 'samples': 22654464, 'steps': 44246, 'loss/train': 0.6372883915901184} 02/26/2022 10:53:06 - INFO - codeparrot_training - Step 44247: {'lr': 1.7513738957118135e-05, 'samples': 22654976, 'steps': 44247, 'loss/train': 1.9252136945724487} 02/26/2022 10:53:10 - INFO - codeparrot_training - Step 44248: {'lr': 1.7507723001812747e-05, 'samples': 22655488, 'steps': 44248, 'loss/train': 1.052933692932129} 02/26/2022 10:53:15 - INFO - codeparrot_training - Step 44249: {'lr': 1.75017080424304e-05, 'samples': 22656000, 'steps': 44249, 'loss/train': 1.8560402393341064} 02/26/2022 10:53:19 - INFO - codeparrot_training - Step 44250: {'lr': 1.7495694078996982e-05, 'samples': 22656512, 'steps': 44250, 'loss/train': 0.8501306176185608} 02/26/2022 10:53:26 - INFO - codeparrot_training - Step 44251: {'lr': 1.7489681111538147e-05, 'samples': 22657024, 'steps': 44251, 'loss/train': 1.3095051050186157} 02/26/2022 10:53:30 - INFO - codeparrot_training - Step 44252: {'lr': 1.7483669140079705e-05, 'samples': 22657536, 'steps': 44252, 'loss/train': 1.2871676683425903} 02/26/2022 10:53:35 - INFO - codeparrot_training - Step 44253: {'lr': 1.7477658164647357e-05, 'samples': 22658048, 'steps': 44253, 'loss/train': 2.25771164894104} 02/26/2022 10:53:39 - INFO - codeparrot_training - Step 44254: {'lr': 1.747164818526692e-05, 'samples': 22658560, 'steps': 44254, 'loss/train': 1.1048643589019775} 02/26/2022 10:53:44 - INFO - codeparrot_training - Step 44255: {'lr': 1.746563920196409e-05, 'samples': 22659072, 'steps': 44255, 'loss/train': 1.9347702264785767} 02/26/2022 10:53:48 - INFO - codeparrot_training - Step 44256: {'lr': 1.7459631214764654e-05, 'samples': 22659584, 'steps': 44256, 'loss/train': 1.8936128616333008} 02/26/2022 10:53:53 - INFO - codeparrot_training - Step 44257: {'lr': 1.745362422369426e-05, 'samples': 22660096, 'steps': 44257, 'loss/train': 2.983276605606079} 02/26/2022 10:53:57 - INFO - codeparrot_training - Step 44258: {'lr': 1.7447618228778745e-05, 'samples': 22660608, 'steps': 44258, 'loss/train': 0.921862781047821} 02/26/2022 10:54:04 - INFO - codeparrot_training - Step 44259: {'lr': 1.7441613230043786e-05, 'samples': 22661120, 'steps': 44259, 'loss/train': 1.5292614698410034} 02/26/2022 10:54:07 - INFO - codeparrot_training - Step 44260: {'lr': 1.743560922751511e-05, 'samples': 22661632, 'steps': 44260, 'loss/train': 1.538718342781067} 02/26/2022 10:54:13 - INFO - codeparrot_training - Step 44261: {'lr': 1.7429606221218392e-05, 'samples': 22662144, 'steps': 44261, 'loss/train': 0.7069941759109497} 02/26/2022 10:54:16 - INFO - codeparrot_training - Step 44262: {'lr': 1.7423604211179444e-05, 'samples': 22662656, 'steps': 44262, 'loss/train': 1.9481362104415894} 02/26/2022 10:54:22 - INFO - codeparrot_training - Step 44263: {'lr': 1.7417603197423914e-05, 'samples': 22663168, 'steps': 44263, 'loss/train': 0.46221479773521423} 02/26/2022 10:54:25 - INFO - codeparrot_training - Step 44264: {'lr': 1.741160317997753e-05, 'samples': 22663680, 'steps': 44264, 'loss/train': 2.5421078205108643} 02/26/2022 10:54:31 - INFO - codeparrot_training - Step 44265: {'lr': 1.740560415886594e-05, 'samples': 22664192, 'steps': 44265, 'loss/train': 0.11504635959863663} 02/26/2022 10:54:34 - INFO - codeparrot_training - Step 44266: {'lr': 1.73996061341149e-05, 'samples': 22664704, 'steps': 44266, 'loss/train': 1.5536805391311646} 02/26/2022 10:54:40 - INFO - codeparrot_training - Step 44267: {'lr': 1.739360910575011e-05, 'samples': 22665216, 'steps': 44267, 'loss/train': 1.0991525650024414} 02/26/2022 10:54:43 - INFO - codeparrot_training - Step 44268: {'lr': 1.7387613073797215e-05, 'samples': 22665728, 'steps': 44268, 'loss/train': 1.2090414762496948} 02/26/2022 10:54:49 - INFO - codeparrot_training - Step 44269: {'lr': 1.7381618038281978e-05, 'samples': 22666240, 'steps': 44269, 'loss/train': 1.298271656036377} 02/26/2022 10:54:52 - INFO - codeparrot_training - Step 44270: {'lr': 1.7375623999229955e-05, 'samples': 22666752, 'steps': 44270, 'loss/train': 0.57925945520401} 02/26/2022 10:55:00 - INFO - codeparrot_training - Step 44271: {'lr': 1.736963095666691e-05, 'samples': 22667264, 'steps': 44271, 'loss/train': 0.42965155839920044} 02/26/2022 10:55:04 - INFO - codeparrot_training - Step 44272: {'lr': 1.7363638910618486e-05, 'samples': 22667776, 'steps': 44272, 'loss/train': 2.509251356124878} 02/26/2022 10:55:09 - INFO - codeparrot_training - Step 44273: {'lr': 1.7357647861110438e-05, 'samples': 22668288, 'steps': 44273, 'loss/train': 1.362303614616394} 02/26/2022 10:55:12 - INFO - codeparrot_training - Step 44274: {'lr': 1.7351657808168248e-05, 'samples': 22668800, 'steps': 44274, 'loss/train': 1.1205365657806396} 02/26/2022 10:55:18 - INFO - codeparrot_training - Step 44275: {'lr': 1.7345668751817755e-05, 'samples': 22669312, 'steps': 44275, 'loss/train': 2.245718240737915} 02/26/2022 10:55:21 - INFO - codeparrot_training - Step 44276: {'lr': 1.733968069208447e-05, 'samples': 22669824, 'steps': 44276, 'loss/train': 1.7272087335586548} 02/26/2022 10:55:27 - INFO - codeparrot_training - Step 44277: {'lr': 1.733369362899423e-05, 'samples': 22670336, 'steps': 44277, 'loss/train': 0.06671199947595596} 02/26/2022 10:55:31 - INFO - codeparrot_training - Step 44278: {'lr': 1.732770756257246e-05, 'samples': 22670848, 'steps': 44278, 'loss/train': 2.2394421100616455} 02/26/2022 10:55:36 - INFO - codeparrot_training - Step 44279: {'lr': 1.7321722492844972e-05, 'samples': 22671360, 'steps': 44279, 'loss/train': 1.0178898572921753} 02/26/2022 10:55:39 - INFO - codeparrot_training - Step 44280: {'lr': 1.7315738419837272e-05, 'samples': 22671872, 'steps': 44280, 'loss/train': 0.20388707518577576} 02/26/2022 10:55:47 - INFO - codeparrot_training - Step 44281: {'lr': 1.7309755343575122e-05, 'samples': 22672384, 'steps': 44281, 'loss/train': 1.9329198598861694} 02/26/2022 10:55:51 - INFO - codeparrot_training - Step 44282: {'lr': 1.730377326408408e-05, 'samples': 22672896, 'steps': 44282, 'loss/train': 2.291621685028076} 02/26/2022 10:55:54 - INFO - codeparrot_training - Step 44283: {'lr': 1.72977921813898e-05, 'samples': 22673408, 'steps': 44283, 'loss/train': 1.476567268371582} 02/26/2022 10:56:00 - INFO - codeparrot_training - Step 44284: {'lr': 1.7291812095517834e-05, 'samples': 22673920, 'steps': 44284, 'loss/train': 1.6869803667068481} 02/26/2022 10:56:03 - INFO - codeparrot_training - Step 44285: {'lr': 1.7285833006493917e-05, 'samples': 22674432, 'steps': 44285, 'loss/train': 2.1983492374420166} 02/26/2022 10:56:09 - INFO - codeparrot_training - Step 44286: {'lr': 1.7279854914343557e-05, 'samples': 22674944, 'steps': 44286, 'loss/train': 2.136845111846924} 02/26/2022 10:56:12 - INFO - codeparrot_training - Step 44287: {'lr': 1.7273877819092425e-05, 'samples': 22675456, 'steps': 44287, 'loss/train': 1.2943073511123657} 02/26/2022 10:56:18 - INFO - codeparrot_training - Step 44288: {'lr': 1.726790172076606e-05, 'samples': 22675968, 'steps': 44288, 'loss/train': 1.0539237260818481} 02/26/2022 10:56:21 - INFO - codeparrot_training - Step 44289: {'lr': 1.7261926619390162e-05, 'samples': 22676480, 'steps': 44289, 'loss/train': 0.6272141337394714} 02/26/2022 10:56:27 - INFO - codeparrot_training - Step 44290: {'lr': 1.7255952514990236e-05, 'samples': 22676992, 'steps': 44290, 'loss/train': 2.3690638542175293} 02/26/2022 10:56:30 - INFO - codeparrot_training - Step 44291: {'lr': 1.7249979407591904e-05, 'samples': 22677504, 'steps': 44291, 'loss/train': 1.5071882009506226} 02/26/2022 10:56:36 - INFO - codeparrot_training - Step 44292: {'lr': 1.7244007297220726e-05, 'samples': 22678016, 'steps': 44292, 'loss/train': 1.57838773727417} 02/26/2022 10:56:39 - INFO - codeparrot_training - Step 44293: {'lr': 1.723803618390235e-05, 'samples': 22678528, 'steps': 44293, 'loss/train': 1.4604501724243164} 02/26/2022 10:56:45 - INFO - codeparrot_training - Step 44294: {'lr': 1.7232066067662284e-05, 'samples': 22679040, 'steps': 44294, 'loss/train': 1.5234919786453247} 02/26/2022 10:56:48 - INFO - codeparrot_training - Step 44295: {'lr': 1.7226096948526227e-05, 'samples': 22679552, 'steps': 44295, 'loss/train': 1.8620376586914062} 02/26/2022 10:56:54 - INFO - codeparrot_training - Step 44296: {'lr': 1.722012882651955e-05, 'samples': 22680064, 'steps': 44296, 'loss/train': 1.6121175289154053} 02/26/2022 10:56:57 - INFO - codeparrot_training - Step 44297: {'lr': 1.721416170166798e-05, 'samples': 22680576, 'steps': 44297, 'loss/train': 1.5987181663513184} 02/26/2022 10:57:05 - INFO - codeparrot_training - Step 44298: {'lr': 1.7208195573996998e-05, 'samples': 22681088, 'steps': 44298, 'loss/train': 0.6864548325538635} 02/26/2022 10:57:08 - INFO - codeparrot_training - Step 44299: {'lr': 1.7202230443532223e-05, 'samples': 22681600, 'steps': 44299, 'loss/train': 0.8742496371269226} 02/26/2022 10:57:14 - INFO - codeparrot_training - Step 44300: {'lr': 1.719626631029911e-05, 'samples': 22682112, 'steps': 44300, 'loss/train': 2.184666156768799} 02/26/2022 10:57:19 - INFO - codeparrot_training - Step 44301: {'lr': 1.7190303174323323e-05, 'samples': 22682624, 'steps': 44301, 'loss/train': 2.066521406173706} 02/26/2022 10:57:23 - INFO - codeparrot_training - Step 44302: {'lr': 1.7184341035630325e-05, 'samples': 22683136, 'steps': 44302, 'loss/train': 1.77566397190094} 02/26/2022 10:57:28 - INFO - codeparrot_training - Step 44303: {'lr': 1.717837989424567e-05, 'samples': 22683648, 'steps': 44303, 'loss/train': 1.2108327150344849} 02/26/2022 10:57:32 - INFO - codeparrot_training - Step 44304: {'lr': 1.717241975019493e-05, 'samples': 22684160, 'steps': 44304, 'loss/train': 1.5419927835464478} 02/26/2022 10:57:37 - INFO - codeparrot_training - Step 44305: {'lr': 1.7166460603503632e-05, 'samples': 22684672, 'steps': 44305, 'loss/train': 1.5790438652038574} 02/26/2022 10:57:41 - INFO - codeparrot_training - Step 44306: {'lr': 1.716050245419726e-05, 'samples': 22685184, 'steps': 44306, 'loss/train': 1.9302430152893066} 02/26/2022 10:57:48 - INFO - codeparrot_training - Step 44307: {'lr': 1.7154545302301346e-05, 'samples': 22685696, 'steps': 44307, 'loss/train': 1.3340981006622314} 02/26/2022 10:57:52 - INFO - codeparrot_training - Step 44308: {'lr': 1.7148589147841426e-05, 'samples': 22686208, 'steps': 44308, 'loss/train': 1.9165781736373901} 02/26/2022 10:57:57 - INFO - codeparrot_training - Step 44309: {'lr': 1.7142633990843036e-05, 'samples': 22686720, 'steps': 44309, 'loss/train': 1.89263117313385} 02/26/2022 10:58:01 - INFO - codeparrot_training - Step 44310: {'lr': 1.7136679831331657e-05, 'samples': 22687232, 'steps': 44310, 'loss/train': 2.063504219055176} 02/26/2022 10:58:06 - INFO - codeparrot_training - Step 44311: {'lr': 1.713072666933277e-05, 'samples': 22687744, 'steps': 44311, 'loss/train': 1.4691232442855835} 02/26/2022 10:58:10 - INFO - codeparrot_training - Step 44312: {'lr': 1.7124774504871933e-05, 'samples': 22688256, 'steps': 44312, 'loss/train': 2.1377978324890137} 02/26/2022 10:58:15 - INFO - codeparrot_training - Step 44313: {'lr': 1.7118823337974625e-05, 'samples': 22688768, 'steps': 44313, 'loss/train': 1.274139404296875} 02/26/2022 10:58:19 - INFO - codeparrot_training - Step 44314: {'lr': 1.7112873168666333e-05, 'samples': 22689280, 'steps': 44314, 'loss/train': 1.5638521909713745} 02/26/2022 10:58:24 - INFO - codeparrot_training - Step 44315: {'lr': 1.7106923996972473e-05, 'samples': 22689792, 'steps': 44315, 'loss/train': 1.7805567979812622} 02/26/2022 10:58:28 - INFO - codeparrot_training - Step 44316: {'lr': 1.7100975822918667e-05, 'samples': 22690304, 'steps': 44316, 'loss/train': 1.4318487644195557} 02/26/2022 10:58:35 - INFO - codeparrot_training - Step 44317: {'lr': 1.7095028646530313e-05, 'samples': 22690816, 'steps': 44317, 'loss/train': 1.335247278213501} 02/26/2022 10:58:39 - INFO - codeparrot_training - Step 44318: {'lr': 1.708908246783292e-05, 'samples': 22691328, 'steps': 44318, 'loss/train': 0.7506019473075867} 02/26/2022 10:58:44 - INFO - codeparrot_training - Step 44319: {'lr': 1.7083137286851875e-05, 'samples': 22691840, 'steps': 44319, 'loss/train': 2.368936777114868} 02/26/2022 10:58:48 - INFO - codeparrot_training - Step 44320: {'lr': 1.7077193103612777e-05, 'samples': 22692352, 'steps': 44320, 'loss/train': 1.7087756395339966} 02/26/2022 10:58:53 - INFO - codeparrot_training - Step 44321: {'lr': 1.707124991814099e-05, 'samples': 22692864, 'steps': 44321, 'loss/train': 1.6384295225143433} 02/26/2022 10:58:57 - INFO - codeparrot_training - Step 44322: {'lr': 1.7065307730462027e-05, 'samples': 22693376, 'steps': 44322, 'loss/train': 2.0050432682037354} 02/26/2022 10:59:03 - INFO - codeparrot_training - Step 44323: {'lr': 1.7059366540601277e-05, 'samples': 22693888, 'steps': 44323, 'loss/train': 1.3051033020019531} 02/26/2022 10:59:06 - INFO - codeparrot_training - Step 44324: {'lr': 1.7053426348584283e-05, 'samples': 22694400, 'steps': 44324, 'loss/train': 1.8936129808425903} 02/26/2022 10:59:11 - INFO - codeparrot_training - Step 44325: {'lr': 1.7047487154436438e-05, 'samples': 22694912, 'steps': 44325, 'loss/train': 2.197650194168091} 02/26/2022 10:59:15 - INFO - codeparrot_training - Step 44326: {'lr': 1.7041548958183163e-05, 'samples': 22695424, 'steps': 44326, 'loss/train': 1.7790600061416626} 02/26/2022 10:59:22 - INFO - codeparrot_training - Step 44327: {'lr': 1.7035611759849913e-05, 'samples': 22695936, 'steps': 44327, 'loss/train': 1.5710827112197876} 02/26/2022 10:59:26 - INFO - codeparrot_training - Step 44328: {'lr': 1.7029675559462137e-05, 'samples': 22696448, 'steps': 44328, 'loss/train': 1.9791598320007324} 02/26/2022 10:59:31 - INFO - codeparrot_training - Step 44329: {'lr': 1.7023740357045235e-05, 'samples': 22696960, 'steps': 44329, 'loss/train': 1.5320208072662354} 02/26/2022 10:59:35 - INFO - codeparrot_training - Step 44330: {'lr': 1.7017806152624744e-05, 'samples': 22697472, 'steps': 44330, 'loss/train': 0.6081035137176514} 02/26/2022 10:59:40 - INFO - codeparrot_training - Step 44331: {'lr': 1.7011872946225888e-05, 'samples': 22697984, 'steps': 44331, 'loss/train': 1.1642194986343384} 02/26/2022 10:59:44 - INFO - codeparrot_training - Step 44332: {'lr': 1.700594073787423e-05, 'samples': 22698496, 'steps': 44332, 'loss/train': 1.7371748685836792} 02/26/2022 10:59:49 - INFO - codeparrot_training - Step 44333: {'lr': 1.700000952759509e-05, 'samples': 22699008, 'steps': 44333, 'loss/train': 2.411820411682129} 02/26/2022 10:59:53 - INFO - codeparrot_training - Step 44334: {'lr': 1.6994079315414022e-05, 'samples': 22699520, 'steps': 44334, 'loss/train': 1.6131850481033325} 02/26/2022 10:59:58 - INFO - codeparrot_training - Step 44335: {'lr': 1.6988150101356238e-05, 'samples': 22700032, 'steps': 44335, 'loss/train': 2.5622992515563965} 02/26/2022 11:00:02 - INFO - codeparrot_training - Step 44336: {'lr': 1.6982221885447263e-05, 'samples': 22700544, 'steps': 44336, 'loss/train': 2.8687050342559814} 02/26/2022 11:00:07 - INFO - codeparrot_training - Step 44337: {'lr': 1.6976294667712443e-05, 'samples': 22701056, 'steps': 44337, 'loss/train': 0.29597508907318115} 02/26/2022 11:00:11 - INFO - codeparrot_training - Step 44338: {'lr': 1.697036844817723e-05, 'samples': 22701568, 'steps': 44338, 'loss/train': 1.9202061891555786} 02/26/2022 11:00:16 - INFO - codeparrot_training - Step 44339: {'lr': 1.696444322686691e-05, 'samples': 22702080, 'steps': 44339, 'loss/train': 1.9780937433242798} 02/26/2022 11:00:20 - INFO - codeparrot_training - Step 44340: {'lr': 1.695851900380696e-05, 'samples': 22702592, 'steps': 44340, 'loss/train': 1.8543251752853394} 02/26/2022 11:00:25 - INFO - codeparrot_training - Step 44341: {'lr': 1.6952595779022667e-05, 'samples': 22703104, 'steps': 44341, 'loss/train': 1.5015946626663208} 02/26/2022 11:00:29 - INFO - codeparrot_training - Step 44342: {'lr': 1.694667355253948e-05, 'samples': 22703616, 'steps': 44342, 'loss/train': 1.5951582193374634} 02/26/2022 11:00:36 - INFO - codeparrot_training - Step 44343: {'lr': 1.6940752324382747e-05, 'samples': 22704128, 'steps': 44343, 'loss/train': 1.6642869710922241} 02/26/2022 11:00:40 - INFO - codeparrot_training - Step 44344: {'lr': 1.693483209457783e-05, 'samples': 22704640, 'steps': 44344, 'loss/train': 2.298598051071167} 02/26/2022 11:00:45 - INFO - codeparrot_training - Step 44345: {'lr': 1.6928912863150042e-05, 'samples': 22705152, 'steps': 44345, 'loss/train': 2.067037582397461} 02/26/2022 11:00:49 - INFO - codeparrot_training - Step 44346: {'lr': 1.692299463012484e-05, 'samples': 22705664, 'steps': 44346, 'loss/train': 1.6908059120178223} 02/26/2022 11:00:55 - INFO - codeparrot_training - Step 44347: {'lr': 1.69170773955275e-05, 'samples': 22706176, 'steps': 44347, 'loss/train': 1.96216881275177} 02/26/2022 11:00:58 - INFO - codeparrot_training - Step 44348: {'lr': 1.6911161159383403e-05, 'samples': 22706688, 'steps': 44348, 'loss/train': 0.8237321972846985} 02/26/2022 11:01:03 - INFO - codeparrot_training - Step 44349: {'lr': 1.6905245921717853e-05, 'samples': 22707200, 'steps': 44349, 'loss/train': 0.5964906811714172} 02/26/2022 11:01:07 - INFO - codeparrot_training - Step 44350: {'lr': 1.6899331682556195e-05, 'samples': 22707712, 'steps': 44350, 'loss/train': 1.7712651491165161} 02/26/2022 11:01:12 - INFO - codeparrot_training - Step 44351: {'lr': 1.6893418441923823e-05, 'samples': 22708224, 'steps': 44351, 'loss/train': 1.2285692691802979} 02/26/2022 11:01:16 - INFO - codeparrot_training - Step 44352: {'lr': 1.6887506199846026e-05, 'samples': 22708736, 'steps': 44352, 'loss/train': 1.5095158815383911} 02/26/2022 11:01:23 - INFO - codeparrot_training - Step 44353: {'lr': 1.688159495634811e-05, 'samples': 22709248, 'steps': 44353, 'loss/train': 1.5155912637710571} 02/26/2022 11:01:26 - INFO - codeparrot_training - Step 44354: {'lr': 1.6875684711455398e-05, 'samples': 22709760, 'steps': 44354, 'loss/train': 3.1292927265167236} 02/26/2022 11:01:32 - INFO - codeparrot_training - Step 44355: {'lr': 1.6869775465193277e-05, 'samples': 22710272, 'steps': 44355, 'loss/train': 0.6916528344154358} 02/26/2022 11:01:35 - INFO - codeparrot_training - Step 44356: {'lr': 1.686386721758698e-05, 'samples': 22710784, 'steps': 44356, 'loss/train': 1.5902073383331299} 02/26/2022 11:01:41 - INFO - codeparrot_training - Step 44357: {'lr': 1.685795996866188e-05, 'samples': 22711296, 'steps': 44357, 'loss/train': 2.023514747619629} 02/26/2022 11:01:44 - INFO - codeparrot_training - Step 44358: {'lr': 1.6852053718443174e-05, 'samples': 22711808, 'steps': 44358, 'loss/train': 2.0408995151519775} 02/26/2022 11:01:50 - INFO - codeparrot_training - Step 44359: {'lr': 1.6846148466956314e-05, 'samples': 22712320, 'steps': 44359, 'loss/train': 2.2770464420318604} 02/26/2022 11:01:53 - INFO - codeparrot_training - Step 44360: {'lr': 1.6840244214226503e-05, 'samples': 22712832, 'steps': 44360, 'loss/train': 1.7497460842132568} 02/26/2022 11:01:59 - INFO - codeparrot_training - Step 44361: {'lr': 1.6834340960279028e-05, 'samples': 22713344, 'steps': 44361, 'loss/train': 1.7307929992675781} 02/26/2022 11:02:02 - INFO - codeparrot_training - Step 44362: {'lr': 1.682843870513917e-05, 'samples': 22713856, 'steps': 44362, 'loss/train': 1.2015419006347656} 02/26/2022 11:02:08 - INFO - codeparrot_training - Step 44363: {'lr': 1.6822537448832305e-05, 'samples': 22714368, 'steps': 44363, 'loss/train': 1.3100426197052002} 02/26/2022 11:02:11 - INFO - codeparrot_training - Step 44364: {'lr': 1.6816637191383573e-05, 'samples': 22714880, 'steps': 44364, 'loss/train': 0.4053735136985779} 02/26/2022 11:02:19 - INFO - codeparrot_training - Step 44365: {'lr': 1.6810737932818428e-05, 'samples': 22715392, 'steps': 44365, 'loss/train': 2.5232043266296387} 02/26/2022 11:02:22 - INFO - codeparrot_training - Step 44366: {'lr': 1.6804839673161958e-05, 'samples': 22715904, 'steps': 44366, 'loss/train': 2.31803822517395} 02/26/2022 11:02:28 - INFO - codeparrot_training - Step 44367: {'lr': 1.679894241243954e-05, 'samples': 22716416, 'steps': 44367, 'loss/train': 0.9804304242134094} 02/26/2022 11:02:33 - INFO - codeparrot_training - Step 44368: {'lr': 1.679304615067634e-05, 'samples': 22716928, 'steps': 44368, 'loss/train': 1.317438006401062} 02/26/2022 11:02:37 - INFO - codeparrot_training - Step 44369: {'lr': 1.6787150887897785e-05, 'samples': 22717440, 'steps': 44369, 'loss/train': 1.6659883260726929} 02/26/2022 11:02:40 - INFO - codeparrot_training - Step 44370: {'lr': 1.6781256624128937e-05, 'samples': 22717952, 'steps': 44370, 'loss/train': 1.5732172727584839} 02/26/2022 11:02:46 - INFO - codeparrot_training - Step 44371: {'lr': 1.677536335939517e-05, 'samples': 22718464, 'steps': 44371, 'loss/train': 2.2136690616607666} 02/26/2022 11:02:49 - INFO - codeparrot_training - Step 44372: {'lr': 1.676947109372165e-05, 'samples': 22718976, 'steps': 44372, 'loss/train': 1.3972251415252686} 02/26/2022 11:02:55 - INFO - codeparrot_training - Step 44373: {'lr': 1.6763579827133723e-05, 'samples': 22719488, 'steps': 44373, 'loss/train': 2.0953214168548584} 02/26/2022 11:02:58 - INFO - codeparrot_training - Step 44374: {'lr': 1.675768955965648e-05, 'samples': 22720000, 'steps': 44374, 'loss/train': 1.6399335861206055} 02/26/2022 11:03:06 - INFO - codeparrot_training - Step 44375: {'lr': 1.675180029131529e-05, 'samples': 22720512, 'steps': 44375, 'loss/train': 1.7166416645050049} 02/26/2022 11:03:09 - INFO - codeparrot_training - Step 44376: {'lr': 1.674591202213527e-05, 'samples': 22721024, 'steps': 44376, 'loss/train': 3.4678473472595215} 02/26/2022 11:03:15 - INFO - codeparrot_training - Step 44377: {'lr': 1.674002475214176e-05, 'samples': 22721536, 'steps': 44377, 'loss/train': 0.4510103166103363} 02/26/2022 11:03:18 - INFO - codeparrot_training - Step 44378: {'lr': 1.673413848135988e-05, 'samples': 22722048, 'steps': 44378, 'loss/train': 2.1093909740448} 02/26/2022 11:03:24 - INFO - codeparrot_training - Step 44379: {'lr': 1.6728253209814915e-05, 'samples': 22722560, 'steps': 44379, 'loss/train': 1.158706784248352} 02/26/2022 11:03:27 - INFO - codeparrot_training - Step 44380: {'lr': 1.6722368937531984e-05, 'samples': 22723072, 'steps': 44380, 'loss/train': 1.2744829654693604} 02/26/2022 11:03:33 - INFO - codeparrot_training - Step 44381: {'lr': 1.6716485664536372e-05, 'samples': 22723584, 'steps': 44381, 'loss/train': 1.732061505317688} 02/26/2022 11:03:36 - INFO - codeparrot_training - Step 44382: {'lr': 1.6710603390853284e-05, 'samples': 22724096, 'steps': 44382, 'loss/train': 1.1910959482192993} 02/26/2022 11:03:41 - INFO - codeparrot_training - Step 44383: {'lr': 1.670472211650789e-05, 'samples': 22724608, 'steps': 44383, 'loss/train': 1.8050498962402344} 02/26/2022 11:03:47 - INFO - codeparrot_training - Step 44384: {'lr': 1.669884184152534e-05, 'samples': 22725120, 'steps': 44384, 'loss/train': 1.7572084665298462} 02/26/2022 11:03:50 - INFO - codeparrot_training - Step 44385: {'lr': 1.669296256593092e-05, 'samples': 22725632, 'steps': 44385, 'loss/train': 1.8314539194107056} 02/26/2022 11:03:56 - INFO - codeparrot_training - Step 44386: {'lr': 1.6687084289749772e-05, 'samples': 22726144, 'steps': 44386, 'loss/train': 1.5001260042190552} 02/26/2022 11:04:00 - INFO - codeparrot_training - Step 44387: {'lr': 1.6681207013007043e-05, 'samples': 22726656, 'steps': 44387, 'loss/train': 2.005467176437378} 02/26/2022 11:04:03 - INFO - codeparrot_training - Step 44388: {'lr': 1.667533073572791e-05, 'samples': 22727168, 'steps': 44388, 'loss/train': 0.6919127702713013} 02/26/2022 11:04:10 - INFO - codeparrot_training - Step 44389: {'lr': 1.6669455457937606e-05, 'samples': 22727680, 'steps': 44389, 'loss/train': 3.4618687629699707} 02/26/2022 11:04:14 - INFO - codeparrot_training - Step 44390: {'lr': 1.666358117966127e-05, 'samples': 22728192, 'steps': 44390, 'loss/train': 1.8993316888809204} 02/26/2022 11:04:19 - INFO - codeparrot_training - Step 44391: {'lr': 1.665770790092405e-05, 'samples': 22728704, 'steps': 44391, 'loss/train': 0.6506956219673157} 02/26/2022 11:04:23 - INFO - codeparrot_training - Step 44392: {'lr': 1.66518356217511e-05, 'samples': 22729216, 'steps': 44392, 'loss/train': 0.6920478940010071} 02/26/2022 11:04:29 - INFO - codeparrot_training - Step 44393: {'lr': 1.6645964342167613e-05, 'samples': 22729728, 'steps': 44393, 'loss/train': 2.802110195159912} 02/26/2022 11:04:32 - INFO - codeparrot_training - Step 44394: {'lr': 1.6640094062198714e-05, 'samples': 22730240, 'steps': 44394, 'loss/train': 1.5093233585357666} 02/26/2022 11:04:38 - INFO - codeparrot_training - Step 44395: {'lr': 1.6634224781869544e-05, 'samples': 22730752, 'steps': 44395, 'loss/train': 2.13940167427063} 02/26/2022 11:04:41 - INFO - codeparrot_training - Step 44396: {'lr': 1.6628356501205283e-05, 'samples': 22731264, 'steps': 44396, 'loss/train': 2.4444870948791504} 02/26/2022 11:04:47 - INFO - codeparrot_training - Step 44397: {'lr': 1.6622489220230964e-05, 'samples': 22731776, 'steps': 44397, 'loss/train': 0.8667925000190735} 02/26/2022 11:04:50 - INFO - codeparrot_training - Step 44398: {'lr': 1.661662293897187e-05, 'samples': 22732288, 'steps': 44398, 'loss/train': 1.5421849489212036} 02/26/2022 11:04:57 - INFO - codeparrot_training - Step 44399: {'lr': 1.661075765745304e-05, 'samples': 22732800, 'steps': 44399, 'loss/train': 1.5873042345046997} 02/26/2022 11:05:01 - INFO - codeparrot_training - Step 44400: {'lr': 1.6604893375699592e-05, 'samples': 22733312, 'steps': 44400, 'loss/train': 2.275908946990967} 02/26/2022 11:05:06 - INFO - codeparrot_training - Step 44401: {'lr': 1.6599030093736644e-05, 'samples': 22733824, 'steps': 44401, 'loss/train': 1.5064443349838257} 02/26/2022 11:05:10 - INFO - codeparrot_training - Step 44402: {'lr': 1.6593167811589398e-05, 'samples': 22734336, 'steps': 44402, 'loss/train': 0.9278638362884521} 02/26/2022 11:05:15 - INFO - codeparrot_training - Step 44403: {'lr': 1.6587306529282838e-05, 'samples': 22734848, 'steps': 44403, 'loss/train': 1.7654253244400024} 02/26/2022 11:05:19 - INFO - codeparrot_training - Step 44404: {'lr': 1.6581446246842243e-05, 'samples': 22735360, 'steps': 44404, 'loss/train': 2.624229907989502} 02/26/2022 11:05:24 - INFO - codeparrot_training - Step 44405: {'lr': 1.657558696429254e-05, 'samples': 22735872, 'steps': 44405, 'loss/train': 1.9898579120635986} 02/26/2022 11:05:28 - INFO - codeparrot_training - Step 44406: {'lr': 1.6569728681658904e-05, 'samples': 22736384, 'steps': 44406, 'loss/train': 1.9804118871688843} 02/26/2022 11:05:34 - INFO - codeparrot_training - Step 44407: {'lr': 1.6563871398966423e-05, 'samples': 22736896, 'steps': 44407, 'loss/train': 0.9044759273529053} 02/26/2022 11:05:37 - INFO - codeparrot_training - Step 44408: {'lr': 1.655801511624025e-05, 'samples': 22737408, 'steps': 44408, 'loss/train': 1.9096637964248657} 02/26/2022 11:05:45 - INFO - codeparrot_training - Step 44409: {'lr': 1.655215983350536e-05, 'samples': 22737920, 'steps': 44409, 'loss/train': 2.122568130493164} 02/26/2022 11:05:48 - INFO - codeparrot_training - Step 44410: {'lr': 1.65463055507869e-05, 'samples': 22738432, 'steps': 44410, 'loss/train': 0.9513373374938965} 02/26/2022 11:05:54 - INFO - codeparrot_training - Step 44411: {'lr': 1.6540452268109906e-05, 'samples': 22738944, 'steps': 44411, 'loss/train': 1.8396836519241333} 02/26/2022 11:05:57 - INFO - codeparrot_training - Step 44412: {'lr': 1.6534599985499527e-05, 'samples': 22739456, 'steps': 44412, 'loss/train': 0.6635790467262268} 02/26/2022 11:06:03 - INFO - codeparrot_training - Step 44413: {'lr': 1.6528748702980796e-05, 'samples': 22739968, 'steps': 44413, 'loss/train': 2.1034364700317383} 02/26/2022 11:06:06 - INFO - codeparrot_training - Step 44414: {'lr': 1.652289842057875e-05, 'samples': 22740480, 'steps': 44414, 'loss/train': 2.0697436332702637} 02/26/2022 11:06:12 - INFO - codeparrot_training - Step 44415: {'lr': 1.6517049138318446e-05, 'samples': 22740992, 'steps': 44415, 'loss/train': 1.4194016456604004} 02/26/2022 11:06:15 - INFO - codeparrot_training - Step 44416: {'lr': 1.6511200856225012e-05, 'samples': 22741504, 'steps': 44416, 'loss/train': 1.4980175495147705} 02/26/2022 11:06:21 - INFO - codeparrot_training - Step 44417: {'lr': 1.6505353574323423e-05, 'samples': 22742016, 'steps': 44417, 'loss/train': 0.16788151860237122} 02/26/2022 11:06:24 - INFO - codeparrot_training - Step 44418: {'lr': 1.6499507292638773e-05, 'samples': 22742528, 'steps': 44418, 'loss/train': 2.3986454010009766} 02/26/2022 11:06:30 - INFO - codeparrot_training - Step 44419: {'lr': 1.6493662011196065e-05, 'samples': 22743040, 'steps': 44419, 'loss/train': 2.2067408561706543} 02/26/2022 11:06:33 - INFO - codeparrot_training - Step 44420: {'lr': 1.6487817730020365e-05, 'samples': 22743552, 'steps': 44420, 'loss/train': 1.5402960777282715} 02/26/2022 11:06:39 - INFO - codeparrot_training - Step 44421: {'lr': 1.6481974449136738e-05, 'samples': 22744064, 'steps': 44421, 'loss/train': 1.4308686256408691} 02/26/2022 11:06:42 - INFO - codeparrot_training - Step 44422: {'lr': 1.6476132168570162e-05, 'samples': 22744576, 'steps': 44422, 'loss/train': 2.275926113128662} 02/26/2022 11:06:48 - INFO - codeparrot_training - Step 44423: {'lr': 1.6470290888345675e-05, 'samples': 22745088, 'steps': 44423, 'loss/train': 2.815519094467163} 02/26/2022 11:06:51 - INFO - codeparrot_training - Step 44424: {'lr': 1.6464450608488308e-05, 'samples': 22745600, 'steps': 44424, 'loss/train': 1.0224368572235107} 02/26/2022 11:06:57 - INFO - codeparrot_training - Step 44425: {'lr': 1.6458611329023073e-05, 'samples': 22746112, 'steps': 44425, 'loss/train': 1.6621652841567993} 02/26/2022 11:07:00 - INFO - codeparrot_training - Step 44426: {'lr': 1.6452773049975027e-05, 'samples': 22746624, 'steps': 44426, 'loss/train': 0.5357076525688171} 02/26/2022 11:07:07 - INFO - codeparrot_training - Step 44427: {'lr': 1.6446935771369103e-05, 'samples': 22747136, 'steps': 44427, 'loss/train': 1.5596338510513306} 02/26/2022 11:07:11 - INFO - codeparrot_training - Step 44428: {'lr': 1.644109949323036e-05, 'samples': 22747648, 'steps': 44428, 'loss/train': 2.2854785919189453} 02/26/2022 11:07:16 - INFO - codeparrot_training - Step 44429: {'lr': 1.6435264215583807e-05, 'samples': 22748160, 'steps': 44429, 'loss/train': 1.2561132907867432} 02/26/2022 11:07:20 - INFO - codeparrot_training - Step 44430: {'lr': 1.6429429938454394e-05, 'samples': 22748672, 'steps': 44430, 'loss/train': 2.0563724040985107} 02/26/2022 11:07:25 - INFO - codeparrot_training - Step 44431: {'lr': 1.6423596661867102e-05, 'samples': 22749184, 'steps': 44431, 'loss/train': 2.307969808578491} 02/26/2022 11:07:29 - INFO - codeparrot_training - Step 44432: {'lr': 1.6417764385846996e-05, 'samples': 22749696, 'steps': 44432, 'loss/train': 1.527620792388916} 02/26/2022 11:07:35 - INFO - codeparrot_training - Step 44433: {'lr': 1.6411933110419023e-05, 'samples': 22750208, 'steps': 44433, 'loss/train': 2.0716638565063477} 02/26/2022 11:07:38 - INFO - codeparrot_training - Step 44434: {'lr': 1.6406102835608144e-05, 'samples': 22750720, 'steps': 44434, 'loss/train': 2.2128257751464844} 02/26/2022 11:07:45 - INFO - codeparrot_training - Step 44435: {'lr': 1.6400273561439305e-05, 'samples': 22751232, 'steps': 44435, 'loss/train': 1.1949172019958496} 02/26/2022 11:07:48 - INFO - codeparrot_training - Step 44436: {'lr': 1.639444528793757e-05, 'samples': 22751744, 'steps': 44436, 'loss/train': 1.7842636108398438} 02/26/2022 11:07:54 - INFO - codeparrot_training - Step 44437: {'lr': 1.6388618015127836e-05, 'samples': 22752256, 'steps': 44437, 'loss/train': 3.123776435852051} 02/26/2022 11:07:57 - INFO - codeparrot_training - Step 44438: {'lr': 1.638279174303506e-05, 'samples': 22752768, 'steps': 44438, 'loss/train': 1.9284619092941284} 02/26/2022 11:08:05 - INFO - codeparrot_training - Step 44439: {'lr': 1.637696647168427e-05, 'samples': 22753280, 'steps': 44439, 'loss/train': 2.3993191719055176} 02/26/2022 11:08:09 - INFO - codeparrot_training - Step 44440: {'lr': 1.6371142201100338e-05, 'samples': 22753792, 'steps': 44440, 'loss/train': 1.1672440767288208} 02/26/2022 11:08:14 - INFO - codeparrot_training - Step 44441: {'lr': 1.636531893130827e-05, 'samples': 22754304, 'steps': 44441, 'loss/train': 1.2598628997802734} 02/26/2022 11:08:18 - INFO - codeparrot_training - Step 44442: {'lr': 1.635949666233294e-05, 'samples': 22754816, 'steps': 44442, 'loss/train': 1.899202585220337} 02/26/2022 11:08:23 - INFO - codeparrot_training - Step 44443: {'lr': 1.6353675394199434e-05, 'samples': 22755328, 'steps': 44443, 'loss/train': 1.6192600727081299} 02/26/2022 11:08:27 - INFO - codeparrot_training - Step 44444: {'lr': 1.6347855126932515e-05, 'samples': 22755840, 'steps': 44444, 'loss/train': 1.7620224952697754} 02/26/2022 11:08:32 - INFO - codeparrot_training - Step 44445: {'lr': 1.6342035860557213e-05, 'samples': 22756352, 'steps': 44445, 'loss/train': 1.531874179840088} 02/26/2022 11:08:36 - INFO - codeparrot_training - Step 44446: {'lr': 1.63362175950984e-05, 'samples': 22756864, 'steps': 44446, 'loss/train': 1.4054603576660156} 02/26/2022 11:08:41 - INFO - codeparrot_training - Step 44447: {'lr': 1.6330400330581112e-05, 'samples': 22757376, 'steps': 44447, 'loss/train': 0.9037963151931763} 02/26/2022 11:08:45 - INFO - codeparrot_training - Step 44448: {'lr': 1.6324584067030107e-05, 'samples': 22757888, 'steps': 44448, 'loss/train': 1.8330234289169312} 02/26/2022 11:08:52 - INFO - codeparrot_training - Step 44449: {'lr': 1.631876880447042e-05, 'samples': 22758400, 'steps': 44449, 'loss/train': 1.782422661781311} 02/26/2022 11:08:55 - INFO - codeparrot_training - Step 44450: {'lr': 1.6312954542926888e-05, 'samples': 22758912, 'steps': 44450, 'loss/train': 1.4474053382873535} 02/26/2022 11:09:01 - INFO - codeparrot_training - Step 44451: {'lr': 1.63071412824245e-05, 'samples': 22759424, 'steps': 44451, 'loss/train': 1.9806078672409058} 02/26/2022 11:09:04 - INFO - codeparrot_training - Step 44452: {'lr': 1.6301329022988114e-05, 'samples': 22759936, 'steps': 44452, 'loss/train': 2.3462915420532227} 02/26/2022 11:09:10 - INFO - codeparrot_training - Step 44453: {'lr': 1.6295517764642635e-05, 'samples': 22760448, 'steps': 44453, 'loss/train': 0.9762347936630249} 02/26/2022 11:09:13 - INFO - codeparrot_training - Step 44454: {'lr': 1.6289707507412905e-05, 'samples': 22760960, 'steps': 44454, 'loss/train': 1.8561501502990723} 02/26/2022 11:09:19 - INFO - codeparrot_training - Step 44455: {'lr': 1.628389825132387e-05, 'samples': 22761472, 'steps': 44455, 'loss/train': 1.9037665128707886} 02/26/2022 11:09:22 - INFO - codeparrot_training - Step 44456: {'lr': 1.627808999640043e-05, 'samples': 22761984, 'steps': 44456, 'loss/train': 1.3427125215530396} 02/26/2022 11:09:28 - INFO - codeparrot_training - Step 44457: {'lr': 1.6272282742667427e-05, 'samples': 22762496, 'steps': 44457, 'loss/train': 1.6749818325042725} 02/26/2022 11:09:31 - INFO - codeparrot_training - Step 44458: {'lr': 1.626647649014973e-05, 'samples': 22763008, 'steps': 44458, 'loss/train': 1.0923738479614258} 02/26/2022 11:09:39 - INFO - codeparrot_training - Step 44459: {'lr': 1.626067123887223e-05, 'samples': 22763520, 'steps': 44459, 'loss/train': 2.346987724304199} 02/26/2022 11:09:42 - INFO - codeparrot_training - Step 44460: {'lr': 1.6254866988859778e-05, 'samples': 22764032, 'steps': 44460, 'loss/train': 1.9795336723327637} 02/26/2022 11:09:48 - INFO - codeparrot_training - Step 44461: {'lr': 1.624906374013732e-05, 'samples': 22764544, 'steps': 44461, 'loss/train': 1.8764890432357788} 02/26/2022 11:09:51 - INFO - codeparrot_training - Step 44462: {'lr': 1.624326149272959e-05, 'samples': 22765056, 'steps': 44462, 'loss/train': 0.8494221568107605} 02/26/2022 11:09:57 - INFO - codeparrot_training - Step 44463: {'lr': 1.6237460246661533e-05, 'samples': 22765568, 'steps': 44463, 'loss/train': 2.170180320739746} 02/26/2022 11:10:00 - INFO - codeparrot_training - Step 44464: {'lr': 1.6231660001957916e-05, 'samples': 22766080, 'steps': 44464, 'loss/train': 1.8503493070602417} 02/26/2022 11:10:06 - INFO - codeparrot_training - Step 44465: {'lr': 1.6225860758643742e-05, 'samples': 22766592, 'steps': 44465, 'loss/train': 1.6686559915542603} 02/26/2022 11:10:10 - INFO - codeparrot_training - Step 44466: {'lr': 1.6220062516743627e-05, 'samples': 22767104, 'steps': 44466, 'loss/train': 1.4229120016098022} 02/26/2022 11:10:16 - INFO - codeparrot_training - Step 44467: {'lr': 1.6214265276282614e-05, 'samples': 22767616, 'steps': 44467, 'loss/train': 2.4440057277679443} 02/26/2022 11:10:19 - INFO - codeparrot_training - Step 44468: {'lr': 1.6208469037285402e-05, 'samples': 22768128, 'steps': 44468, 'loss/train': 0.8162800073623657} 02/26/2022 11:10:24 - INFO - codeparrot_training - Step 44469: {'lr': 1.620267379977694e-05, 'samples': 22768640, 'steps': 44469, 'loss/train': 2.914889097213745} 02/26/2022 11:10:28 - INFO - codeparrot_training - Step 44470: {'lr': 1.6196879563781908e-05, 'samples': 22769152, 'steps': 44470, 'loss/train': 1.7064741849899292} 02/26/2022 11:10:35 - INFO - codeparrot_training - Step 44471: {'lr': 1.6191086329325227e-05, 'samples': 22769664, 'steps': 44471, 'loss/train': 1.5800516605377197} 02/26/2022 11:10:39 - INFO - codeparrot_training - Step 44472: {'lr': 1.6185294096431685e-05, 'samples': 22770176, 'steps': 44472, 'loss/train': 0.9188938140869141} 02/26/2022 11:10:44 - INFO - codeparrot_training - Step 44473: {'lr': 1.6179502865126122e-05, 'samples': 22770688, 'steps': 44473, 'loss/train': 1.3499151468276978} 02/26/2022 11:10:48 - INFO - codeparrot_training - Step 44474: {'lr': 1.6173712635433323e-05, 'samples': 22771200, 'steps': 44474, 'loss/train': 0.5873643755912781} 02/26/2022 11:10:53 - INFO - codeparrot_training - Step 44475: {'lr': 1.6167923407378078e-05, 'samples': 22771712, 'steps': 44475, 'loss/train': 1.6316195726394653} 02/26/2022 11:10:57 - INFO - codeparrot_training - Step 44476: {'lr': 1.6162135180985223e-05, 'samples': 22772224, 'steps': 44476, 'loss/train': 1.1441177129745483} 02/26/2022 11:11:02 - INFO - codeparrot_training - Step 44477: {'lr': 1.615634795627946e-05, 'samples': 22772736, 'steps': 44477, 'loss/train': 1.8901101350784302} 02/26/2022 11:11:06 - INFO - codeparrot_training - Step 44478: {'lr': 1.6150561733285717e-05, 'samples': 22773248, 'steps': 44478, 'loss/train': 1.8721638917922974} 02/26/2022 11:11:11 - INFO - codeparrot_training - Step 44479: {'lr': 1.61447765120287e-05, 'samples': 22773760, 'steps': 44479, 'loss/train': 0.9614415764808655} 02/26/2022 11:11:15 - INFO - codeparrot_training - Step 44480: {'lr': 1.6138992292533183e-05, 'samples': 22774272, 'steps': 44480, 'loss/train': 1.1102651357650757} 02/26/2022 11:11:22 - INFO - codeparrot_training - Step 44481: {'lr': 1.613320907482396e-05, 'samples': 22774784, 'steps': 44481, 'loss/train': 2.107565402984619} 02/26/2022 11:11:25 - INFO - codeparrot_training - Step 44482: {'lr': 1.612742685892582e-05, 'samples': 22775296, 'steps': 44482, 'loss/train': 1.6693578958511353} 02/26/2022 11:11:31 - INFO - codeparrot_training - Step 44483: {'lr': 1.612164564486354e-05, 'samples': 22775808, 'steps': 44483, 'loss/train': 0.8629449009895325} 02/26/2022 11:11:34 - INFO - codeparrot_training - Step 44484: {'lr': 1.6115865432661858e-05, 'samples': 22776320, 'steps': 44484, 'loss/train': 1.665207028388977} 02/26/2022 11:11:40 - INFO - codeparrot_training - Step 44485: {'lr': 1.61100862223455e-05, 'samples': 22776832, 'steps': 44485, 'loss/train': 1.5447379350662231} 02/26/2022 11:11:43 - INFO - codeparrot_training - Step 44486: {'lr': 1.6104308013939306e-05, 'samples': 22777344, 'steps': 44486, 'loss/train': 1.5844171047210693} 02/26/2022 11:11:49 - INFO - codeparrot_training - Step 44487: {'lr': 1.609853080746798e-05, 'samples': 22777856, 'steps': 44487, 'loss/train': 1.3804858922958374} 02/26/2022 11:11:52 - INFO - codeparrot_training - Step 44488: {'lr': 1.6092754602956284e-05, 'samples': 22778368, 'steps': 44488, 'loss/train': 2.5782389640808105} 02/26/2022 11:11:58 - INFO - codeparrot_training - Step 44489: {'lr': 1.6086979400428913e-05, 'samples': 22778880, 'steps': 44489, 'loss/train': 0.08129153400659561} 02/26/2022 11:12:02 - INFO - codeparrot_training - Step 44490: {'lr': 1.6081205199910687e-05, 'samples': 22779392, 'steps': 44490, 'loss/train': 1.9636930227279663} 02/26/2022 11:12:07 - INFO - codeparrot_training - Step 44491: {'lr': 1.6075432001426305e-05, 'samples': 22779904, 'steps': 44491, 'loss/train': 0.9341439604759216} 02/26/2022 11:12:11 - INFO - codeparrot_training - Step 44492: {'lr': 1.606965980500047e-05, 'samples': 22780416, 'steps': 44492, 'loss/train': 1.3446927070617676} 02/26/2022 11:12:16 - INFO - codeparrot_training - Step 44493: {'lr': 1.606388861065791e-05, 'samples': 22780928, 'steps': 44493, 'loss/train': 1.7613171339035034} 02/26/2022 11:12:20 - INFO - codeparrot_training - Step 44494: {'lr': 1.6058118418423418e-05, 'samples': 22781440, 'steps': 44494, 'loss/train': 1.5892575979232788} 02/26/2022 11:12:27 - INFO - codeparrot_training - Step 44495: {'lr': 1.6052349228321634e-05, 'samples': 22781952, 'steps': 44495, 'loss/train': 1.4850115776062012} 02/26/2022 11:12:30 - INFO - codeparrot_training - Step 44496: {'lr': 1.6046581040377316e-05, 'samples': 22782464, 'steps': 44496, 'loss/train': 1.5176955461502075} 02/26/2022 11:12:36 - INFO - codeparrot_training - Step 44497: {'lr': 1.6040813854615115e-05, 'samples': 22782976, 'steps': 44497, 'loss/train': 2.4214513301849365} 02/26/2022 11:12:40 - INFO - codeparrot_training - Step 44498: {'lr': 1.6035047671059816e-05, 'samples': 22783488, 'steps': 44498, 'loss/train': 1.9264425039291382} 02/26/2022 11:12:45 - INFO - codeparrot_training - Step 44499: {'lr': 1.6029282489736036e-05, 'samples': 22784000, 'steps': 44499, 'loss/train': 0.5881751179695129} 02/26/2022 11:12:49 - INFO - codeparrot_training - Step 44500: {'lr': 1.6023518310668618e-05, 'samples': 22784512, 'steps': 44500, 'loss/train': 1.1519964933395386} 02/26/2022 11:12:54 - INFO - codeparrot_training - Step 44501: {'lr': 1.601775513388204e-05, 'samples': 22785024, 'steps': 44501, 'loss/train': 1.3802087306976318} 02/26/2022 11:12:58 - INFO - codeparrot_training - Step 44502: {'lr': 1.601199295940117e-05, 'samples': 22785536, 'steps': 44502, 'loss/train': 1.8513593673706055} 02/26/2022 11:13:03 - INFO - codeparrot_training - Step 44503: {'lr': 1.6006231787250552e-05, 'samples': 22786048, 'steps': 44503, 'loss/train': 1.5561649799346924} 02/26/2022 11:13:06 - INFO - codeparrot_training - Step 44504: {'lr': 1.600047161745505e-05, 'samples': 22786560, 'steps': 44504, 'loss/train': 1.5796668529510498} 02/26/2022 11:13:12 - INFO - codeparrot_training - Step 44505: {'lr': 1.5994712450039112e-05, 'samples': 22787072, 'steps': 44505, 'loss/train': 2.1288797855377197} 02/26/2022 11:13:16 - INFO - codeparrot_training - Step 44506: {'lr': 1.5988954285027586e-05, 'samples': 22787584, 'steps': 44506, 'loss/train': 1.2501821517944336} 02/26/2022 11:13:23 - INFO - codeparrot_training - Step 44507: {'lr': 1.5983197122445004e-05, 'samples': 22788096, 'steps': 44507, 'loss/train': 1.4690552949905396} 02/26/2022 11:13:26 - INFO - codeparrot_training - Step 44508: {'lr': 1.5977440962316208e-05, 'samples': 22788608, 'steps': 44508, 'loss/train': 0.3202507793903351} 02/26/2022 11:13:32 - INFO - codeparrot_training - Step 44509: {'lr': 1.5971685804665654e-05, 'samples': 22789120, 'steps': 44509, 'loss/train': 1.7949697971343994} 02/26/2022 11:13:35 - INFO - codeparrot_training - Step 44510: {'lr': 1.5965931649518122e-05, 'samples': 22789632, 'steps': 44510, 'loss/train': 1.3624011278152466} 02/26/2022 11:13:41 - INFO - codeparrot_training - Step 44511: {'lr': 1.596017849689818e-05, 'samples': 22790144, 'steps': 44511, 'loss/train': 1.8824430704116821} 02/26/2022 11:13:45 - INFO - codeparrot_training - Step 44512: {'lr': 1.5954426346830558e-05, 'samples': 22790656, 'steps': 44512, 'loss/train': 1.8574906587600708} 02/26/2022 11:13:50 - INFO - codeparrot_training - Step 44513: {'lr': 1.5948675199339873e-05, 'samples': 22791168, 'steps': 44513, 'loss/train': 2.2995009422302246} 02/26/2022 11:13:53 - INFO - codeparrot_training - Step 44514: {'lr': 1.5942925054450718e-05, 'samples': 22791680, 'steps': 44514, 'loss/train': 1.6055896282196045} 02/26/2022 11:13:59 - INFO - codeparrot_training - Step 44515: {'lr': 1.593717591218774e-05, 'samples': 22792192, 'steps': 44515, 'loss/train': 2.008331298828125} 02/26/2022 11:14:03 - INFO - codeparrot_training - Step 44516: {'lr': 1.5931427772575585e-05, 'samples': 22792704, 'steps': 44516, 'loss/train': 8.181538581848145} 02/26/2022 11:14:10 - INFO - codeparrot_training - Step 44517: {'lr': 1.592568063563887e-05, 'samples': 22793216, 'steps': 44517, 'loss/train': 0.44485682249069214} 02/26/2022 11:14:13 - INFO - codeparrot_training - Step 44518: {'lr': 1.591993450140225e-05, 'samples': 22793728, 'steps': 44518, 'loss/train': 1.7015061378479004} 02/26/2022 11:14:19 - INFO - codeparrot_training - Step 44519: {'lr': 1.5914189369890226e-05, 'samples': 22794240, 'steps': 44519, 'loss/train': 1.489059567451477} 02/26/2022 11:14:22 - INFO - codeparrot_training - Step 44520: {'lr': 1.5908445241127527e-05, 'samples': 22794752, 'steps': 44520, 'loss/train': 0.8103165626525879} 02/26/2022 11:14:28 - INFO - codeparrot_training - Step 44521: {'lr': 1.5902702115138752e-05, 'samples': 22795264, 'steps': 44521, 'loss/train': 2.0369951725006104} 02/26/2022 11:14:31 - INFO - codeparrot_training - Step 44522: {'lr': 1.5896959991948428e-05, 'samples': 22795776, 'steps': 44522, 'loss/train': 2.053330421447754} 02/26/2022 11:14:37 - INFO - codeparrot_training - Step 44523: {'lr': 1.589121887158121e-05, 'samples': 22796288, 'steps': 44523, 'loss/train': 2.026573419570923} 02/26/2022 11:14:40 - INFO - codeparrot_training - Step 44524: {'lr': 1.5885478754061627e-05, 'samples': 22796800, 'steps': 44524, 'loss/train': 0.892111599445343} 02/26/2022 11:14:46 - INFO - codeparrot_training - Step 44525: {'lr': 1.587973963941433e-05, 'samples': 22797312, 'steps': 44525, 'loss/train': 0.5852540731430054} 02/26/2022 11:14:49 - INFO - codeparrot_training - Step 44526: {'lr': 1.587400152766391e-05, 'samples': 22797824, 'steps': 44526, 'loss/train': 1.5767624378204346} 02/26/2022 11:14:56 - INFO - codeparrot_training - Step 44527: {'lr': 1.5868264418834933e-05, 'samples': 22798336, 'steps': 44527, 'loss/train': 1.603103518486023} 02/26/2022 11:15:00 - INFO - codeparrot_training - Step 44528: {'lr': 1.586252831295193e-05, 'samples': 22798848, 'steps': 44528, 'loss/train': 1.9427510499954224} 02/26/2022 11:15:05 - INFO - codeparrot_training - Step 44529: {'lr': 1.585679321003952e-05, 'samples': 22799360, 'steps': 44529, 'loss/train': 0.6830424666404724} 02/26/2022 11:15:09 - INFO - codeparrot_training - Step 44530: {'lr': 1.5851059110122275e-05, 'samples': 22799872, 'steps': 44530, 'loss/train': 1.0778858661651611} 02/26/2022 11:15:15 - INFO - codeparrot_training - Step 44531: {'lr': 1.584532601322475e-05, 'samples': 22800384, 'steps': 44531, 'loss/train': 1.698647141456604} 02/26/2022 11:15:18 - INFO - codeparrot_training - Step 44532: {'lr': 1.5839593919371425e-05, 'samples': 22800896, 'steps': 44532, 'loss/train': 0.4144231379032135} 02/26/2022 11:15:24 - INFO - codeparrot_training - Step 44533: {'lr': 1.583386282858701e-05, 'samples': 22801408, 'steps': 44533, 'loss/train': 1.755776286125183} 02/26/2022 11:15:27 - INFO - codeparrot_training - Step 44534: {'lr': 1.5828132740895896e-05, 'samples': 22801920, 'steps': 44534, 'loss/train': 1.4762258529663086} 02/26/2022 11:15:33 - INFO - codeparrot_training - Step 44535: {'lr': 1.5822403656322786e-05, 'samples': 22802432, 'steps': 44535, 'loss/train': 2.0760912895202637} 02/26/2022 11:15:36 - INFO - codeparrot_training - Step 44536: {'lr': 1.5816675574892082e-05, 'samples': 22802944, 'steps': 44536, 'loss/train': 2.3944971561431885} 02/26/2022 11:15:42 - INFO - codeparrot_training - Step 44537: {'lr': 1.5810948496628396e-05, 'samples': 22803456, 'steps': 44537, 'loss/train': 2.216747760772705} 02/26/2022 11:15:46 - INFO - codeparrot_training - Step 44538: {'lr': 1.5805222421556215e-05, 'samples': 22803968, 'steps': 44538, 'loss/train': 2.25126051902771} 02/26/2022 11:15:51 - INFO - codeparrot_training - Step 44539: {'lr': 1.579949734970018e-05, 'samples': 22804480, 'steps': 44539, 'loss/train': 2.663761854171753} 02/26/2022 11:15:55 - INFO - codeparrot_training - Step 44540: {'lr': 1.579377328108464e-05, 'samples': 22804992, 'steps': 44540, 'loss/train': 0.8956112265586853} 02/26/2022 11:16:02 - INFO - codeparrot_training - Step 44541: {'lr': 1.5788050215734235e-05, 'samples': 22805504, 'steps': 44541, 'loss/train': 2.7742488384246826} 02/26/2022 11:16:05 - INFO - codeparrot_training - Step 44542: {'lr': 1.5782328153673447e-05, 'samples': 22806016, 'steps': 44542, 'loss/train': 1.8063017129898071} 02/26/2022 11:16:11 - INFO - codeparrot_training - Step 44543: {'lr': 1.5776607094926843e-05, 'samples': 22806528, 'steps': 44543, 'loss/train': 2.1666834354400635} 02/26/2022 11:16:14 - INFO - codeparrot_training - Step 44544: {'lr': 1.5770887039518816e-05, 'samples': 22807040, 'steps': 44544, 'loss/train': 2.014113187789917} 02/26/2022 11:16:20 - INFO - codeparrot_training - Step 44545: {'lr': 1.576516798747396e-05, 'samples': 22807552, 'steps': 44545, 'loss/train': 0.9947926998138428} 02/26/2022 11:16:23 - INFO - codeparrot_training - Step 44546: {'lr': 1.5759449938816727e-05, 'samples': 22808064, 'steps': 44546, 'loss/train': 1.1864845752716064} 02/26/2022 11:16:29 - INFO - codeparrot_training - Step 44547: {'lr': 1.5753732893571653e-05, 'samples': 22808576, 'steps': 44547, 'loss/train': 1.9343748092651367} 02/26/2022 11:16:32 - INFO - codeparrot_training - Step 44548: {'lr': 1.5748016851763192e-05, 'samples': 22809088, 'steps': 44548, 'loss/train': 2.335641384124756} 02/26/2022 11:16:38 - INFO - codeparrot_training - Step 44549: {'lr': 1.5742301813415876e-05, 'samples': 22809600, 'steps': 44549, 'loss/train': 1.488487720489502} 02/26/2022 11:16:41 - INFO - codeparrot_training - Step 44550: {'lr': 1.5736587778554107e-05, 'samples': 22810112, 'steps': 44550, 'loss/train': 1.9591190814971924} 02/26/2022 11:16:47 - INFO - codeparrot_training - Step 44551: {'lr': 1.5730874747202418e-05, 'samples': 22810624, 'steps': 44551, 'loss/train': 1.547108769416809} 02/26/2022 11:16:50 - INFO - codeparrot_training - Step 44552: {'lr': 1.5725162719385315e-05, 'samples': 22811136, 'steps': 44552, 'loss/train': 1.6096739768981934} 02/26/2022 11:16:54 - INFO - codeparrot_training - Step 44553: {'lr': 1.57194516951272e-05, 'samples': 22811648, 'steps': 44553, 'loss/train': 2.193249464035034} 02/26/2022 11:17:01 - INFO - codeparrot_training - Step 44554: {'lr': 1.5713741674452526e-05, 'samples': 22812160, 'steps': 44554, 'loss/train': 1.7400882244110107} 02/26/2022 11:17:05 - INFO - codeparrot_training - Step 44555: {'lr': 1.5708032657385825e-05, 'samples': 22812672, 'steps': 44555, 'loss/train': 2.021022319793701} 02/26/2022 11:17:10 - INFO - codeparrot_training - Step 44556: {'lr': 1.5702324643951522e-05, 'samples': 22813184, 'steps': 44556, 'loss/train': 1.346304178237915} 02/26/2022 11:17:14 - INFO - codeparrot_training - Step 44557: {'lr': 1.5696617634174043e-05, 'samples': 22813696, 'steps': 44557, 'loss/train': 1.8561760187149048} 02/26/2022 11:17:19 - INFO - codeparrot_training - Step 44558: {'lr': 1.5690911628077843e-05, 'samples': 22814208, 'steps': 44558, 'loss/train': 1.738935947418213} 02/26/2022 11:17:23 - INFO - codeparrot_training - Step 44559: {'lr': 1.56852066256874e-05, 'samples': 22814720, 'steps': 44559, 'loss/train': 1.6938575506210327} 02/26/2022 11:17:28 - INFO - codeparrot_training - Step 44560: {'lr': 1.5679502627027137e-05, 'samples': 22815232, 'steps': 44560, 'loss/train': 2.8369622230529785} 02/26/2022 11:17:32 - INFO - codeparrot_training - Step 44561: {'lr': 1.5673799632121484e-05, 'samples': 22815744, 'steps': 44561, 'loss/train': 2.1059303283691406} 02/26/2022 11:17:37 - INFO - codeparrot_training - Step 44562: {'lr': 1.5668097640994804e-05, 'samples': 22816256, 'steps': 44562, 'loss/train': 1.4703333377838135} 02/26/2022 11:17:40 - INFO - codeparrot_training - Step 44563: {'lr': 1.566239665367164e-05, 'samples': 22816768, 'steps': 44563, 'loss/train': 1.5801500082015991} 02/26/2022 11:17:48 - INFO - codeparrot_training - Step 44564: {'lr': 1.565669667017636e-05, 'samples': 22817280, 'steps': 44564, 'loss/train': 2.679043769836426} 02/26/2022 11:17:51 - INFO - codeparrot_training - Step 44565: {'lr': 1.565099769053338e-05, 'samples': 22817792, 'steps': 44565, 'loss/train': 2.0830371379852295} 02/26/2022 11:17:57 - INFO - codeparrot_training - Step 44566: {'lr': 1.564529971476708e-05, 'samples': 22818304, 'steps': 44566, 'loss/train': 0.9572405219078064} 02/26/2022 11:18:01 - INFO - codeparrot_training - Step 44567: {'lr': 1.563960274290191e-05, 'samples': 22818816, 'steps': 44567, 'loss/train': 2.737130641937256} 02/26/2022 11:18:06 - INFO - codeparrot_training - Step 44568: {'lr': 1.563390677496229e-05, 'samples': 22819328, 'steps': 44568, 'loss/train': 1.5434027910232544} 02/26/2022 11:18:10 - INFO - codeparrot_training - Step 44569: {'lr': 1.5628211810972543e-05, 'samples': 22819840, 'steps': 44569, 'loss/train': 1.8749423027038574} 02/26/2022 11:18:15 - INFO - codeparrot_training - Step 44570: {'lr': 1.5622517850957168e-05, 'samples': 22820352, 'steps': 44570, 'loss/train': 1.623382329940796} 02/26/2022 11:18:19 - INFO - codeparrot_training - Step 44571: {'lr': 1.5616824894940485e-05, 'samples': 22820864, 'steps': 44571, 'loss/train': 1.5651400089263916} 02/26/2022 11:18:24 - INFO - codeparrot_training - Step 44572: {'lr': 1.5611132942946914e-05, 'samples': 22821376, 'steps': 44572, 'loss/train': 1.0098519325256348} 02/26/2022 11:18:28 - INFO - codeparrot_training - Step 44573: {'lr': 1.5605441995000774e-05, 'samples': 22821888, 'steps': 44573, 'loss/train': 1.1188361644744873} 02/26/2022 11:18:35 - INFO - codeparrot_training - Step 44574: {'lr': 1.559975205112657e-05, 'samples': 22822400, 'steps': 44574, 'loss/train': 0.8829925656318665} 02/26/2022 11:18:38 - INFO - codeparrot_training - Step 44575: {'lr': 1.5594063111348504e-05, 'samples': 22822912, 'steps': 44575, 'loss/train': 1.6833213567733765} 02/26/2022 11:18:44 - INFO - codeparrot_training - Step 44576: {'lr': 1.5588375175691116e-05, 'samples': 22823424, 'steps': 44576, 'loss/train': 0.3324960768222809} 02/26/2022 11:18:47 - INFO - codeparrot_training - Step 44577: {'lr': 1.5582688244178633e-05, 'samples': 22823936, 'steps': 44577, 'loss/train': 1.344488501548767} 02/26/2022 11:18:53 - INFO - codeparrot_training - Step 44578: {'lr': 1.5577002316835563e-05, 'samples': 22824448, 'steps': 44578, 'loss/train': 1.4971399307250977} 02/26/2022 11:18:57 - INFO - codeparrot_training - Step 44579: {'lr': 1.5571317393686085e-05, 'samples': 22824960, 'steps': 44579, 'loss/train': 1.6542730331420898} 02/26/2022 11:19:02 - INFO - codeparrot_training - Step 44580: {'lr': 1.5565633474754702e-05, 'samples': 22825472, 'steps': 44580, 'loss/train': 1.607014536857605} 02/26/2022 11:19:06 - INFO - codeparrot_training - Step 44581: {'lr': 1.555995056006565e-05, 'samples': 22825984, 'steps': 44581, 'loss/train': 2.12030029296875} 02/26/2022 11:19:11 - INFO - codeparrot_training - Step 44582: {'lr': 1.5554268649643378e-05, 'samples': 22826496, 'steps': 44582, 'loss/train': 0.6701213717460632} 02/26/2022 11:19:15 - INFO - codeparrot_training - Step 44583: {'lr': 1.5548587743512172e-05, 'samples': 22827008, 'steps': 44583, 'loss/train': 1.2296571731567383} 02/26/2022 11:19:22 - INFO - codeparrot_training - Step 44584: {'lr': 1.5542907841696347e-05, 'samples': 22827520, 'steps': 44584, 'loss/train': 1.1790920495986938} 02/26/2022 11:19:25 - INFO - codeparrot_training - Step 44585: {'lr': 1.5537228944220245e-05, 'samples': 22828032, 'steps': 44585, 'loss/train': 0.9743176102638245} 02/26/2022 11:19:31 - INFO - codeparrot_training - Step 44586: {'lr': 1.5531551051108235e-05, 'samples': 22828544, 'steps': 44586, 'loss/train': 1.2688344717025757} 02/26/2022 11:19:35 - INFO - codeparrot_training - Step 44587: {'lr': 1.55258741623846e-05, 'samples': 22829056, 'steps': 44587, 'loss/train': 1.8284367322921753} 02/26/2022 11:19:40 - INFO - codeparrot_training - Step 44588: {'lr': 1.552019827807369e-05, 'samples': 22829568, 'steps': 44588, 'loss/train': 0.9335541129112244} 02/26/2022 11:19:44 - INFO - codeparrot_training - Step 44589: {'lr': 1.5514523398199724e-05, 'samples': 22830080, 'steps': 44589, 'loss/train': 1.7284339666366577} 02/26/2022 11:19:50 - INFO - codeparrot_training - Step 44590: {'lr': 1.550884952278714e-05, 'samples': 22830592, 'steps': 44590, 'loss/train': 1.8680366277694702} 02/26/2022 11:19:54 - INFO - codeparrot_training - Step 44591: {'lr': 1.5503176651860183e-05, 'samples': 22831104, 'steps': 44591, 'loss/train': 1.9334073066711426} 02/26/2022 11:19:59 - INFO - codeparrot_training - Step 44592: {'lr': 1.5497504785443152e-05, 'samples': 22831616, 'steps': 44592, 'loss/train': 2.0922441482543945} 02/26/2022 11:20:03 - INFO - codeparrot_training - Step 44593: {'lr': 1.5491833923560294e-05, 'samples': 22832128, 'steps': 44593, 'loss/train': 2.21962571144104} 02/26/2022 11:20:08 - INFO - codeparrot_training - Step 44594: {'lr': 1.5486164066236014e-05, 'samples': 22832640, 'steps': 44594, 'loss/train': 1.5537835359573364} 02/26/2022 11:20:12 - INFO - codeparrot_training - Step 44595: {'lr': 1.548049521349451e-05, 'samples': 22833152, 'steps': 44595, 'loss/train': 1.1310774087905884} 02/26/2022 11:20:17 - INFO - codeparrot_training - Step 44596: {'lr': 1.5474827365360157e-05, 'samples': 22833664, 'steps': 44596, 'loss/train': 1.4799506664276123} 02/26/2022 11:20:21 - INFO - codeparrot_training - Step 44597: {'lr': 1.5469160521857095e-05, 'samples': 22834176, 'steps': 44597, 'loss/train': 1.1180918216705322} 02/26/2022 11:20:26 - INFO - codeparrot_training - Step 44598: {'lr': 1.5463494683009725e-05, 'samples': 22834688, 'steps': 44598, 'loss/train': 2.638831615447998} 02/26/2022 11:20:30 - INFO - codeparrot_training - Step 44599: {'lr': 1.5457829848842226e-05, 'samples': 22835200, 'steps': 44599, 'loss/train': 2.268336296081543} 02/26/2022 11:20:37 - INFO - codeparrot_training - Step 44600: {'lr': 1.5452166019378987e-05, 'samples': 22835712, 'steps': 44600, 'loss/train': 1.7282999753952026} 02/26/2022 11:20:41 - INFO - codeparrot_training - Step 44601: {'lr': 1.5446503194644136e-05, 'samples': 22836224, 'steps': 44601, 'loss/train': 1.4947994947433472} 02/26/2022 11:20:46 - INFO - codeparrot_training - Step 44602: {'lr': 1.544084137466198e-05, 'samples': 22836736, 'steps': 44602, 'loss/train': 1.7952890396118164} 02/26/2022 11:20:50 - INFO - codeparrot_training - Step 44603: {'lr': 1.543518055945678e-05, 'samples': 22837248, 'steps': 44603, 'loss/train': 2.8607635498046875} 02/26/2022 11:20:55 - INFO - codeparrot_training - Step 44604: {'lr': 1.542952074905285e-05, 'samples': 22837760, 'steps': 44604, 'loss/train': 2.0992369651794434} 02/26/2022 11:20:59 - INFO - codeparrot_training - Step 44605: {'lr': 1.5423861943474282e-05, 'samples': 22838272, 'steps': 44605, 'loss/train': 1.0394814014434814} 02/26/2022 11:21:04 - INFO - codeparrot_training - Step 44606: {'lr': 1.5418204142745447e-05, 'samples': 22838784, 'steps': 44606, 'loss/train': 1.0934089422225952} 02/26/2022 11:21:08 - INFO - codeparrot_training - Step 44607: {'lr': 1.5412547346890543e-05, 'samples': 22839296, 'steps': 44607, 'loss/train': 2.3242218494415283} 02/26/2022 11:21:13 - INFO - codeparrot_training - Step 44608: {'lr': 1.540689155593375e-05, 'samples': 22839808, 'steps': 44608, 'loss/train': 1.1360435485839844} 02/26/2022 11:21:17 - INFO - codeparrot_training - Step 44609: {'lr': 1.540123676989938e-05, 'samples': 22840320, 'steps': 44609, 'loss/train': 2.215571165084839} 02/26/2022 11:21:24 - INFO - codeparrot_training - Step 44610: {'lr': 1.539558298881161e-05, 'samples': 22840832, 'steps': 44610, 'loss/train': 1.9261587858200073} 02/26/2022 11:21:27 - INFO - codeparrot_training - Step 44611: {'lr': 1.5389930212694664e-05, 'samples': 22841344, 'steps': 44611, 'loss/train': 0.9890658259391785} 02/26/2022 11:21:33 - INFO - codeparrot_training - Step 44612: {'lr': 1.5384278441572754e-05, 'samples': 22841856, 'steps': 44612, 'loss/train': 2.372225761413574} 02/26/2022 11:21:36 - INFO - codeparrot_training - Step 44613: {'lr': 1.5378627675470103e-05, 'samples': 22842368, 'steps': 44613, 'loss/train': 1.7916009426116943} 02/26/2022 11:21:42 - INFO - codeparrot_training - Step 44614: {'lr': 1.537297791441092e-05, 'samples': 22842880, 'steps': 44614, 'loss/train': 1.7261452674865723} 02/26/2022 11:21:45 - INFO - codeparrot_training - Step 44615: {'lr': 1.5367329158419404e-05, 'samples': 22843392, 'steps': 44615, 'loss/train': 2.048445701599121} 02/26/2022 11:21:51 - INFO - codeparrot_training - Step 44616: {'lr': 1.53616814075197e-05, 'samples': 22843904, 'steps': 44616, 'loss/train': 0.3136724829673767} 02/26/2022 11:21:54 - INFO - codeparrot_training - Step 44617: {'lr': 1.53560346617361e-05, 'samples': 22844416, 'steps': 44617, 'loss/train': 1.753232479095459} 02/26/2022 11:22:00 - INFO - codeparrot_training - Step 44618: {'lr': 1.5350388921092723e-05, 'samples': 22844928, 'steps': 44618, 'loss/train': 1.9669389724731445} 02/26/2022 11:22:03 - INFO - codeparrot_training - Step 44619: {'lr': 1.5344744185613768e-05, 'samples': 22845440, 'steps': 44619, 'loss/train': 0.9677431583404541} 02/26/2022 11:22:11 - INFO - codeparrot_training - Step 44620: {'lr': 1.5339100455323384e-05, 'samples': 22845952, 'steps': 44620, 'loss/train': 1.199379324913025} 02/26/2022 11:22:14 - INFO - codeparrot_training - Step 44621: {'lr': 1.533345773024583e-05, 'samples': 22846464, 'steps': 44621, 'loss/train': 2.422481060028076} 02/26/2022 11:22:20 - INFO - codeparrot_training - Step 44622: {'lr': 1.5327816010405227e-05, 'samples': 22846976, 'steps': 44622, 'loss/train': 1.6680974960327148} 02/26/2022 11:22:23 - INFO - codeparrot_training - Step 44623: {'lr': 1.532217529582572e-05, 'samples': 22847488, 'steps': 44623, 'loss/train': 2.0155797004699707} 02/26/2022 11:22:29 - INFO - codeparrot_training - Step 44624: {'lr': 1.5316535586531482e-05, 'samples': 22848000, 'steps': 44624, 'loss/train': 0.9815978407859802} 02/26/2022 11:22:32 - INFO - codeparrot_training - Step 44625: {'lr': 1.531089688254672e-05, 'samples': 22848512, 'steps': 44625, 'loss/train': 2.4685962200164795} 02/26/2022 11:22:38 - INFO - codeparrot_training - Step 44626: {'lr': 1.5305259183895527e-05, 'samples': 22849024, 'steps': 44626, 'loss/train': 1.4522428512573242} 02/26/2022 11:22:41 - INFO - codeparrot_training - Step 44627: {'lr': 1.5299622490602098e-05, 'samples': 22849536, 'steps': 44627, 'loss/train': 1.910078763961792} 02/26/2022 11:22:47 - INFO - codeparrot_training - Step 44628: {'lr': 1.5293986802690533e-05, 'samples': 22850048, 'steps': 44628, 'loss/train': 1.6691898107528687} 02/26/2022 11:22:50 - INFO - codeparrot_training - Step 44629: {'lr': 1.5288352120185033e-05, 'samples': 22850560, 'steps': 44629, 'loss/train': 1.1679444313049316} 02/26/2022 11:22:58 - INFO - codeparrot_training - Step 44630: {'lr': 1.528271844310966e-05, 'samples': 22851072, 'steps': 44630, 'loss/train': 1.4711912870407104} 02/26/2022 11:23:01 - INFO - codeparrot_training - Step 44631: {'lr': 1.5277085771488676e-05, 'samples': 22851584, 'steps': 44631, 'loss/train': 1.7267378568649292} 02/26/2022 11:23:07 - INFO - codeparrot_training - Step 44632: {'lr': 1.5271454105346028e-05, 'samples': 22852096, 'steps': 44632, 'loss/train': 1.5025639533996582} 02/26/2022 11:23:10 - INFO - codeparrot_training - Step 44633: {'lr': 1.5265823444706007e-05, 'samples': 22852608, 'steps': 44633, 'loss/train': 8.783032417297363} 02/26/2022 11:23:16 - INFO - codeparrot_training - Step 44634: {'lr': 1.526019378959259e-05, 'samples': 22853120, 'steps': 44634, 'loss/train': 1.2788257598876953} 02/26/2022 11:23:19 - INFO - codeparrot_training - Step 44635: {'lr': 1.525456514003004e-05, 'samples': 22853632, 'steps': 44635, 'loss/train': 1.1093562841415405} 02/26/2022 11:23:25 - INFO - codeparrot_training - Step 44636: {'lr': 1.5248937496042337e-05, 'samples': 22854144, 'steps': 44636, 'loss/train': 1.7389986515045166} 02/26/2022 11:23:28 - INFO - codeparrot_training - Step 44637: {'lr': 1.5243310857653681e-05, 'samples': 22854656, 'steps': 44637, 'loss/train': 0.5567581057548523} 02/26/2022 11:23:34 - INFO - codeparrot_training - Step 44638: {'lr': 1.5237685224888081e-05, 'samples': 22855168, 'steps': 44638, 'loss/train': 2.364442825317383} 02/26/2022 11:23:37 - INFO - codeparrot_training - Step 44639: {'lr': 1.5232060597769797e-05, 'samples': 22855680, 'steps': 44639, 'loss/train': 1.1462093591690063} 02/26/2022 11:23:43 - INFO - codeparrot_training - Step 44640: {'lr': 1.5226436976322727e-05, 'samples': 22856192, 'steps': 44640, 'loss/train': 1.674525260925293} 02/26/2022 11:23:46 - INFO - codeparrot_training - Step 44641: {'lr': 1.5220814360571072e-05, 'samples': 22856704, 'steps': 44641, 'loss/train': 3.4201090335845947} 02/26/2022 11:23:52 - INFO - codeparrot_training - Step 44642: {'lr': 1.5215192750538897e-05, 'samples': 22857216, 'steps': 44642, 'loss/train': 1.2100610733032227} 02/26/2022 11:23:55 - INFO - codeparrot_training - Step 44643: {'lr': 1.5209572146250295e-05, 'samples': 22857728, 'steps': 44643, 'loss/train': 1.5242233276367188} 02/26/2022 11:24:00 - INFO - codeparrot_training - Step 44644: {'lr': 1.5203952547729328e-05, 'samples': 22858240, 'steps': 44644, 'loss/train': 1.9821656942367554} 02/26/2022 11:24:04 - INFO - codeparrot_training - Step 44645: {'lr': 1.5198333955000088e-05, 'samples': 22858752, 'steps': 44645, 'loss/train': 1.4836583137512207} 02/26/2022 11:24:11 - INFO - codeparrot_training - Step 44646: {'lr': 1.5192716368086585e-05, 'samples': 22859264, 'steps': 44646, 'loss/train': 1.4414249658584595} 02/26/2022 11:24:15 - INFO - codeparrot_training - Step 44647: {'lr': 1.5187099787012965e-05, 'samples': 22859776, 'steps': 44647, 'loss/train': 1.9191162586212158} 02/26/2022 11:24:20 - INFO - codeparrot_training - Step 44648: {'lr': 1.5181484211803238e-05, 'samples': 22860288, 'steps': 44648, 'loss/train': 1.0874743461608887} 02/26/2022 11:24:23 - INFO - codeparrot_training - Step 44649: {'lr': 1.5175869642481494e-05, 'samples': 22860800, 'steps': 44649, 'loss/train': 1.8841404914855957} 02/26/2022 11:24:29 - INFO - codeparrot_training - Step 44650: {'lr': 1.5170256079071687e-05, 'samples': 22861312, 'steps': 44650, 'loss/train': 1.0572062730789185} 02/26/2022 11:24:32 - INFO - codeparrot_training - Step 44651: {'lr': 1.5164643521597993e-05, 'samples': 22861824, 'steps': 44651, 'loss/train': 1.6270796060562134} 02/26/2022 11:24:38 - INFO - codeparrot_training - Step 44652: {'lr': 1.5159031970084391e-05, 'samples': 22862336, 'steps': 44652, 'loss/train': 0.7014657258987427} 02/26/2022 11:24:41 - INFO - codeparrot_training - Step 44653: {'lr': 1.5153421424554947e-05, 'samples': 22862848, 'steps': 44653, 'loss/train': 0.9855467677116394} 02/26/2022 11:24:47 - INFO - codeparrot_training - Step 44654: {'lr': 1.5147811885033641e-05, 'samples': 22863360, 'steps': 44654, 'loss/train': 1.1339077949523926} 02/26/2022 11:24:51 - INFO - codeparrot_training - Step 44655: {'lr': 1.514220335154451e-05, 'samples': 22863872, 'steps': 44655, 'loss/train': 1.8605103492736816} 02/26/2022 11:24:58 - INFO - codeparrot_training - Step 44656: {'lr': 1.5136595824111642e-05, 'samples': 22864384, 'steps': 44656, 'loss/train': 1.6077550649642944} 02/26/2022 11:25:01 - INFO - codeparrot_training - Step 44657: {'lr': 1.5130989302759025e-05, 'samples': 22864896, 'steps': 44657, 'loss/train': 2.056802749633789} 02/26/2022 11:25:07 - INFO - codeparrot_training - Step 44658: {'lr': 1.512538378751066e-05, 'samples': 22865408, 'steps': 44658, 'loss/train': 2.618234634399414} 02/26/2022 11:25:10 - INFO - codeparrot_training - Step 44659: {'lr': 1.5119779278390561e-05, 'samples': 22865920, 'steps': 44659, 'loss/train': 2.2130839824676514} 02/26/2022 11:25:16 - INFO - codeparrot_training - Step 44660: {'lr': 1.5114175775422761e-05, 'samples': 22866432, 'steps': 44660, 'loss/train': 1.6912882328033447} 02/26/2022 11:25:19 - INFO - codeparrot_training - Step 44661: {'lr': 1.5108573278631243e-05, 'samples': 22866944, 'steps': 44661, 'loss/train': 1.4367542266845703} 02/26/2022 11:25:25 - INFO - codeparrot_training - Step 44662: {'lr': 1.5102971788040014e-05, 'samples': 22867456, 'steps': 44662, 'loss/train': 1.1090903282165527} 02/26/2022 11:25:28 - INFO - codeparrot_training - Step 44663: {'lr': 1.509737130367303e-05, 'samples': 22867968, 'steps': 44663, 'loss/train': 1.3047922849655151} 02/26/2022 11:25:34 - INFO - codeparrot_training - Step 44664: {'lr': 1.509177182555435e-05, 'samples': 22868480, 'steps': 44664, 'loss/train': 1.6728254556655884} 02/26/2022 11:25:37 - INFO - codeparrot_training - Step 44665: {'lr': 1.5086173353707932e-05, 'samples': 22868992, 'steps': 44665, 'loss/train': 2.834684371948242} 02/26/2022 11:25:44 - INFO - codeparrot_training - Step 44666: {'lr': 1.5080575888157755e-05, 'samples': 22869504, 'steps': 44666, 'loss/train': 0.025383299216628075} 02/26/2022 11:25:48 - INFO - codeparrot_training - Step 44667: {'lr': 1.5074979428927743e-05, 'samples': 22870016, 'steps': 44667, 'loss/train': 2.3409533500671387} 02/26/2022 11:25:53 - INFO - codeparrot_training - Step 44668: {'lr': 1.506938397604199e-05, 'samples': 22870528, 'steps': 44668, 'loss/train': 2.192551374435425} 02/26/2022 11:25:57 - INFO - codeparrot_training - Step 44669: {'lr': 1.5063789529524336e-05, 'samples': 22871040, 'steps': 44669, 'loss/train': 1.894571304321289} 02/26/2022 11:26:02 - INFO - codeparrot_training - Step 44670: {'lr': 1.5058196089398901e-05, 'samples': 22871552, 'steps': 44670, 'loss/train': 1.6740198135375977} 02/26/2022 11:26:06 - INFO - codeparrot_training - Step 44671: {'lr': 1.5052603655689445e-05, 'samples': 22872064, 'steps': 44671, 'loss/train': 1.7244845628738403} 02/26/2022 11:26:11 - INFO - codeparrot_training - Step 44672: {'lr': 1.5047012228420088e-05, 'samples': 22872576, 'steps': 44672, 'loss/train': 1.27073073387146} 02/26/2022 11:26:15 - INFO - codeparrot_training - Step 44673: {'lr': 1.5041421807614697e-05, 'samples': 22873088, 'steps': 44673, 'loss/train': 1.600583791732788} 02/26/2022 11:26:20 - INFO - codeparrot_training - Step 44674: {'lr': 1.5035832393297311e-05, 'samples': 22873600, 'steps': 44674, 'loss/train': 0.4095900356769562} 02/26/2022 11:26:24 - INFO - codeparrot_training - Step 44675: {'lr': 1.5030243985491743e-05, 'samples': 22874112, 'steps': 44675, 'loss/train': 1.4403547048568726} 02/26/2022 11:26:31 - INFO - codeparrot_training - Step 44676: {'lr': 1.5024656584222029e-05, 'samples': 22874624, 'steps': 44676, 'loss/train': 1.2183821201324463} 02/26/2022 11:26:34 - INFO - codeparrot_training - Step 44677: {'lr': 1.5019070189512014e-05, 'samples': 22875136, 'steps': 44677, 'loss/train': 2.425318717956543} 02/26/2022 11:26:40 - INFO - codeparrot_training - Step 44678: {'lr': 1.5013484801385758e-05, 'samples': 22875648, 'steps': 44678, 'loss/train': 2.4563510417938232} 02/26/2022 11:26:44 - INFO - codeparrot_training - Step 44679: {'lr': 1.5007900419867104e-05, 'samples': 22876160, 'steps': 44679, 'loss/train': 1.4586529731750488} 02/26/2022 11:26:49 - INFO - codeparrot_training - Step 44680: {'lr': 1.5002317044979979e-05, 'samples': 22876672, 'steps': 44680, 'loss/train': 2.079988479614258} 02/26/2022 11:26:53 - INFO - codeparrot_training - Step 44681: {'lr': 1.499673467674828e-05, 'samples': 22877184, 'steps': 44681, 'loss/train': 0.6317312717437744} 02/26/2022 11:26:58 - INFO - codeparrot_training - Step 44682: {'lr': 1.4991153315195987e-05, 'samples': 22877696, 'steps': 44682, 'loss/train': 0.9572232961654663} 02/26/2022 11:27:02 - INFO - codeparrot_training - Step 44683: {'lr': 1.4985572960346971e-05, 'samples': 22878208, 'steps': 44683, 'loss/train': 2.1569111347198486} 02/26/2022 11:27:07 - INFO - codeparrot_training - Step 44684: {'lr': 1.497999361222513e-05, 'samples': 22878720, 'steps': 44684, 'loss/train': 1.4307478666305542} 02/26/2022 11:27:11 - INFO - codeparrot_training - Step 44685: {'lr': 1.497441527085433e-05, 'samples': 22879232, 'steps': 44685, 'loss/train': 1.980157732963562} 02/26/2022 11:27:16 - INFO - codeparrot_training - Step 44686: {'lr': 1.4968837936258528e-05, 'samples': 22879744, 'steps': 44686, 'loss/train': 1.2960748672485352} 02/26/2022 11:27:20 - INFO - codeparrot_training - Step 44687: {'lr': 1.496326160846162e-05, 'samples': 22880256, 'steps': 44687, 'loss/train': 0.7699154019355774} 02/26/2022 11:27:26 - INFO - codeparrot_training - Step 44688: {'lr': 1.4957686287487449e-05, 'samples': 22880768, 'steps': 44688, 'loss/train': 2.26187801361084} 02/26/2022 11:27:29 - INFO - codeparrot_training - Step 44689: {'lr': 1.4952111973359884e-05, 'samples': 22881280, 'steps': 44689, 'loss/train': 0.2561168670654297} 02/26/2022 11:27:33 - INFO - codeparrot_training - Step 44690: {'lr': 1.4946538666102877e-05, 'samples': 22881792, 'steps': 44690, 'loss/train': 1.0869572162628174} 02/26/2022 11:27:38 - INFO - codeparrot_training - Step 44691: {'lr': 1.4940966365740272e-05, 'samples': 22882304, 'steps': 44691, 'loss/train': 1.7562129497528076} 02/26/2022 11:27:42 - INFO - codeparrot_training - Step 44692: {'lr': 1.493539507229591e-05, 'samples': 22882816, 'steps': 44692, 'loss/train': 1.228003978729248} 02/26/2022 11:27:49 - INFO - codeparrot_training - Step 44693: {'lr': 1.4929824785793634e-05, 'samples': 22883328, 'steps': 44693, 'loss/train': 2.1903302669525146} 02/26/2022 11:27:53 - INFO - codeparrot_training - Step 44694: {'lr': 1.4924255506257395e-05, 'samples': 22883840, 'steps': 44694, 'loss/train': 0.9961408972740173} 02/26/2022 11:27:58 - INFO - codeparrot_training - Step 44695: {'lr': 1.491868723371098e-05, 'samples': 22884352, 'steps': 44695, 'loss/train': 2.0601065158843994} 02/26/2022 11:28:02 - INFO - codeparrot_training - Step 44696: {'lr': 1.4913119968178291e-05, 'samples': 22884864, 'steps': 44696, 'loss/train': 1.5648421049118042} 02/26/2022 11:28:07 - INFO - codeparrot_training - Step 44697: {'lr': 1.4907553709683108e-05, 'samples': 22885376, 'steps': 44697, 'loss/train': 1.8498815298080444} 02/26/2022 11:28:11 - INFO - codeparrot_training - Step 44698: {'lr': 1.4901988458249332e-05, 'samples': 22885888, 'steps': 44698, 'loss/train': 1.4192838668823242} 02/26/2022 11:28:16 - INFO - codeparrot_training - Step 44699: {'lr': 1.4896424213900777e-05, 'samples': 22886400, 'steps': 44699, 'loss/train': 1.2419383525848389} 02/26/2022 11:28:20 - INFO - codeparrot_training - Step 44700: {'lr': 1.4890860976661313e-05, 'samples': 22886912, 'steps': 44700, 'loss/train': 1.0514472723007202} 02/26/2022 11:28:25 - INFO - codeparrot_training - Step 44701: {'lr': 1.4885298746554727e-05, 'samples': 22887424, 'steps': 44701, 'loss/train': 1.8465126752853394} 02/26/2022 11:28:29 - INFO - codeparrot_training - Step 44702: {'lr': 1.487973752360483e-05, 'samples': 22887936, 'steps': 44702, 'loss/train': 1.3820991516113281} 02/26/2022 11:28:36 - INFO - codeparrot_training - Step 44703: {'lr': 1.4874177307835523e-05, 'samples': 22888448, 'steps': 44703, 'loss/train': 1.9503690004348755} 02/26/2022 11:28:40 - INFO - codeparrot_training - Step 44704: {'lr': 1.4868618099270536e-05, 'samples': 22888960, 'steps': 44704, 'loss/train': 1.6652113199234009} 02/26/2022 11:28:45 - INFO - codeparrot_training - Step 44705: {'lr': 1.4863059897933823e-05, 'samples': 22889472, 'steps': 44705, 'loss/train': 0.6835692524909973} 02/26/2022 11:28:49 - INFO - codeparrot_training - Step 44706: {'lr': 1.4857502703849002e-05, 'samples': 22889984, 'steps': 44706, 'loss/train': 1.8001538515090942} 02/26/2022 11:28:54 - INFO - codeparrot_training - Step 44707: {'lr': 1.4851946517039999e-05, 'samples': 22890496, 'steps': 44707, 'loss/train': 1.6719951629638672} 02/26/2022 11:28:58 - INFO - codeparrot_training - Step 44708: {'lr': 1.4846391337530574e-05, 'samples': 22891008, 'steps': 44708, 'loss/train': 2.054715156555176} 02/26/2022 11:29:04 - INFO - codeparrot_training - Step 44709: {'lr': 1.4840837165344595e-05, 'samples': 22891520, 'steps': 44709, 'loss/train': 0.8846263289451599} 02/26/2022 11:29:07 - INFO - codeparrot_training - Step 44710: {'lr': 1.483528400050574e-05, 'samples': 22892032, 'steps': 44710, 'loss/train': 1.3819191455841064} 02/26/2022 11:29:13 - INFO - codeparrot_training - Step 44711: {'lr': 1.4829731843037902e-05, 'samples': 22892544, 'steps': 44711, 'loss/train': 1.6048927307128906} 02/26/2022 11:29:16 - INFO - codeparrot_training - Step 44712: {'lr': 1.4824180692964762e-05, 'samples': 22893056, 'steps': 44712, 'loss/train': 1.6939247846603394} 02/26/2022 11:29:23 - INFO - codeparrot_training - Step 44713: {'lr': 1.4818630550310214e-05, 'samples': 22893568, 'steps': 44713, 'loss/train': 0.9921132326126099} 02/26/2022 11:29:27 - INFO - codeparrot_training - Step 44714: {'lr': 1.4813081415097934e-05, 'samples': 22894080, 'steps': 44714, 'loss/train': 1.8767247200012207} 02/26/2022 11:29:33 - INFO - codeparrot_training - Step 44715: {'lr': 1.4807533287351738e-05, 'samples': 22894592, 'steps': 44715, 'loss/train': 1.3981701135635376} 02/26/2022 11:29:37 - INFO - codeparrot_training - Step 44716: {'lr': 1.4801986167095382e-05, 'samples': 22895104, 'steps': 44716, 'loss/train': 1.323700189590454} 02/26/2022 11:29:40 - INFO - codeparrot_training - Step 44717: {'lr': 1.4796440054352656e-05, 'samples': 22895616, 'steps': 44717, 'loss/train': 3.6197593212127686} 02/26/2022 11:29:46 - INFO - codeparrot_training - Step 44718: {'lr': 1.4790894949147288e-05, 'samples': 22896128, 'steps': 44718, 'loss/train': 1.5982636213302612} 02/26/2022 11:29:49 - INFO - codeparrot_training - Step 44719: {'lr': 1.4785350851503037e-05, 'samples': 22896640, 'steps': 44719, 'loss/train': 0.8854488730430603} 02/26/2022 11:29:55 - INFO - codeparrot_training - Step 44720: {'lr': 1.4779807761443637e-05, 'samples': 22897152, 'steps': 44720, 'loss/train': 1.3149296045303345} 02/26/2022 11:29:58 - INFO - codeparrot_training - Step 44721: {'lr': 1.477426567899287e-05, 'samples': 22897664, 'steps': 44721, 'loss/train': 1.3728604316711426} 02/26/2022 11:30:04 - INFO - codeparrot_training - Step 44722: {'lr': 1.4768724604174444e-05, 'samples': 22898176, 'steps': 44722, 'loss/train': 2.3593132495880127} 02/26/2022 11:30:07 - INFO - codeparrot_training - Step 44723: {'lr': 1.4763184537012142e-05, 'samples': 22898688, 'steps': 44723, 'loss/train': 1.705065369606018} 02/26/2022 11:30:14 - INFO - codeparrot_training - Step 44724: {'lr': 1.4757645477529585e-05, 'samples': 22899200, 'steps': 44724, 'loss/train': 1.4917598962783813} 02/26/2022 11:30:18 - INFO - codeparrot_training - Step 44725: {'lr': 1.4752107425750643e-05, 'samples': 22899712, 'steps': 44725, 'loss/train': 2.5256097316741943} 02/26/2022 11:30:23 - INFO - codeparrot_training - Step 44726: {'lr': 1.4746570381698936e-05, 'samples': 22900224, 'steps': 44726, 'loss/train': 5.361443996429443} 02/26/2022 11:30:27 - INFO - codeparrot_training - Step 44727: {'lr': 1.474103434539828e-05, 'samples': 22900736, 'steps': 44727, 'loss/train': 1.583449125289917} 02/26/2022 11:30:32 - INFO - codeparrot_training - Step 44728: {'lr': 1.4735499316872265e-05, 'samples': 22901248, 'steps': 44728, 'loss/train': 1.9465041160583496} 02/26/2022 11:30:36 - INFO - codeparrot_training - Step 44729: {'lr': 1.4729965296144704e-05, 'samples': 22901760, 'steps': 44729, 'loss/train': 1.7254831790924072} 02/26/2022 11:30:41 - INFO - codeparrot_training - Step 44730: {'lr': 1.472443228323922e-05, 'samples': 22902272, 'steps': 44730, 'loss/train': 0.29090437293052673} 02/26/2022 11:30:45 - INFO - codeparrot_training - Step 44731: {'lr': 1.4718900278179652e-05, 'samples': 22902784, 'steps': 44731, 'loss/train': 1.8774993419647217} 02/26/2022 11:30:50 - INFO - codeparrot_training - Step 44732: {'lr': 1.4713369280989513e-05, 'samples': 22903296, 'steps': 44732, 'loss/train': 1.6809306144714355} 02/26/2022 11:30:54 - INFO - codeparrot_training - Step 44733: {'lr': 1.4707839291692616e-05, 'samples': 22903808, 'steps': 44733, 'loss/train': 2.260448694229126} 02/26/2022 11:31:01 - INFO - codeparrot_training - Step 44734: {'lr': 1.4702310310312633e-05, 'samples': 22904320, 'steps': 44734, 'loss/train': 2.3479745388031006} 02/26/2022 11:31:04 - INFO - codeparrot_training - Step 44735: {'lr': 1.4696782336873243e-05, 'samples': 22904832, 'steps': 44735, 'loss/train': 2.1346280574798584} 02/26/2022 11:31:10 - INFO - codeparrot_training - Step 44736: {'lr': 1.4691255371398094e-05, 'samples': 22905344, 'steps': 44736, 'loss/train': 1.2090847492218018} 02/26/2022 11:31:13 - INFO - codeparrot_training - Step 44737: {'lr': 1.4685729413910886e-05, 'samples': 22905856, 'steps': 44737, 'loss/train': 2.1135356426239014} 02/26/2022 11:31:19 - INFO - codeparrot_training - Step 44738: {'lr': 1.4680204464435326e-05, 'samples': 22906368, 'steps': 44738, 'loss/train': 0.7936621904373169} 02/26/2022 11:31:22 - INFO - codeparrot_training - Step 44739: {'lr': 1.4674680522995003e-05, 'samples': 22906880, 'steps': 44739, 'loss/train': 2.1766417026519775} 02/26/2022 11:31:28 - INFO - codeparrot_training - Step 44740: {'lr': 1.4669157589613652e-05, 'samples': 22907392, 'steps': 44740, 'loss/train': 2.0970025062561035} 02/26/2022 11:31:31 - INFO - codeparrot_training - Step 44741: {'lr': 1.4663635664314917e-05, 'samples': 22907904, 'steps': 44741, 'loss/train': 1.4811991453170776} 02/26/2022 11:31:37 - INFO - codeparrot_training - Step 44742: {'lr': 1.4658114747122448e-05, 'samples': 22908416, 'steps': 44742, 'loss/train': 1.4360178709030151} 02/26/2022 11:31:40 - INFO - codeparrot_training - Step 44743: {'lr': 1.4652594838059835e-05, 'samples': 22908928, 'steps': 44743, 'loss/train': 0.7907061576843262} 02/26/2022 11:31:46 - INFO - codeparrot_training - Step 44744: {'lr': 1.4647075937150811e-05, 'samples': 22909440, 'steps': 44744, 'loss/train': 0.9782797694206238} 02/26/2022 11:31:49 - INFO - codeparrot_training - Step 44745: {'lr': 1.4641558044418967e-05, 'samples': 22909952, 'steps': 44745, 'loss/train': 0.5114972591400146} 02/26/2022 11:31:55 - INFO - codeparrot_training - Step 44746: {'lr': 1.4636041159887981e-05, 'samples': 22910464, 'steps': 44746, 'loss/train': 1.0731189250946045} 02/26/2022 11:31:59 - INFO - codeparrot_training - Step 44747: {'lr': 1.4630525283581386e-05, 'samples': 22910976, 'steps': 44747, 'loss/train': 1.8338072299957275} 02/26/2022 11:32:04 - INFO - codeparrot_training - Step 44748: {'lr': 1.4625010415522999e-05, 'samples': 22911488, 'steps': 44748, 'loss/train': 1.7852144241333008} 02/26/2022 11:32:08 - INFO - codeparrot_training - Step 44749: {'lr': 1.4619496555736216e-05, 'samples': 22912000, 'steps': 44749, 'loss/train': 2.174665689468384} 02/26/2022 11:32:13 - INFO - codeparrot_training - Step 44750: {'lr': 1.4613983704244827e-05, 'samples': 22912512, 'steps': 44750, 'loss/train': 1.481002926826477} 02/26/2022 11:32:17 - INFO - codeparrot_training - Step 44751: {'lr': 1.4608471861072337e-05, 'samples': 22913024, 'steps': 44751, 'loss/train': 2.3604652881622314} 02/26/2022 11:32:24 - INFO - codeparrot_training - Step 44752: {'lr': 1.4602961026242479e-05, 'samples': 22913536, 'steps': 44752, 'loss/train': 2.2931647300720215} 02/26/2022 11:32:27 - INFO - codeparrot_training - Step 44753: {'lr': 1.4597451199778761e-05, 'samples': 22914048, 'steps': 44753, 'loss/train': 0.5501608848571777} 02/26/2022 11:32:33 - INFO - codeparrot_training - Step 44754: {'lr': 1.4591942381704803e-05, 'samples': 22914560, 'steps': 44754, 'loss/train': 1.8694555759429932} 02/26/2022 11:32:36 - INFO - codeparrot_training - Step 44755: {'lr': 1.45864345720442e-05, 'samples': 22915072, 'steps': 44755, 'loss/train': 2.1331048011779785} 02/26/2022 11:32:42 - INFO - codeparrot_training - Step 44756: {'lr': 1.4580927770820568e-05, 'samples': 22915584, 'steps': 44756, 'loss/train': 1.9914377927780151} 02/26/2022 11:32:45 - INFO - codeparrot_training - Step 44757: {'lr': 1.457542197805753e-05, 'samples': 22916096, 'steps': 44757, 'loss/train': 2.684089422225952} 02/26/2022 11:32:51 - INFO - codeparrot_training - Step 44758: {'lr': 1.4569917193778593e-05, 'samples': 22916608, 'steps': 44758, 'loss/train': 1.728378415107727} 02/26/2022 11:32:54 - INFO - codeparrot_training - Step 44759: {'lr': 1.4564413418007349e-05, 'samples': 22917120, 'steps': 44759, 'loss/train': 2.331244707107544} 02/26/2022 11:33:02 - INFO - codeparrot_training - Step 44760: {'lr': 1.455891065076742e-05, 'samples': 22917632, 'steps': 44760, 'loss/train': 2.3482542037963867} 02/26/2022 11:33:05 - INFO - codeparrot_training - Step 44761: {'lr': 1.455340889208237e-05, 'samples': 22918144, 'steps': 44761, 'loss/train': 0.6966246366500854} 02/26/2022 11:33:11 - INFO - codeparrot_training - Step 44762: {'lr': 1.4547908141975763e-05, 'samples': 22918656, 'steps': 44762, 'loss/train': 1.5532572269439697} 02/26/2022 11:33:14 - INFO - codeparrot_training - Step 44763: {'lr': 1.4542408400471108e-05, 'samples': 22919168, 'steps': 44763, 'loss/train': 1.6490448713302612} 02/26/2022 11:33:20 - INFO - codeparrot_training - Step 44764: {'lr': 1.4536909667592052e-05, 'samples': 22919680, 'steps': 44764, 'loss/train': 1.952669382095337} 02/26/2022 11:33:23 - INFO - codeparrot_training - Step 44765: {'lr': 1.453141194336205e-05, 'samples': 22920192, 'steps': 44765, 'loss/train': 2.195852756500244} 02/26/2022 11:33:29 - INFO - codeparrot_training - Step 44766: {'lr': 1.4525915227804804e-05, 'samples': 22920704, 'steps': 44766, 'loss/train': 1.8344160318374634} 02/26/2022 11:33:32 - INFO - codeparrot_training - Step 44767: {'lr': 1.4520419520943683e-05, 'samples': 22921216, 'steps': 44767, 'loss/train': 1.6205499172210693} 02/26/2022 11:33:38 - INFO - codeparrot_training - Step 44768: {'lr': 1.4514924822802367e-05, 'samples': 22921728, 'steps': 44768, 'loss/train': 0.5510485768318176} 02/26/2022 11:33:41 - INFO - codeparrot_training - Step 44769: {'lr': 1.4509431133404278e-05, 'samples': 22922240, 'steps': 44769, 'loss/train': 1.9467521905899048} 02/26/2022 11:33:49 - INFO - codeparrot_training - Step 44770: {'lr': 1.450393845277309e-05, 'samples': 22922752, 'steps': 44770, 'loss/train': 1.3639692068099976} 02/26/2022 11:33:52 - INFO - codeparrot_training - Step 44771: {'lr': 1.4498446780932179e-05, 'samples': 22923264, 'steps': 44771, 'loss/train': 1.977573275566101} 02/26/2022 11:33:58 - INFO - codeparrot_training - Step 44772: {'lr': 1.449295611790516e-05, 'samples': 22923776, 'steps': 44772, 'loss/train': 1.6996464729309082} 02/26/2022 11:34:01 - INFO - codeparrot_training - Step 44773: {'lr': 1.4487466463715516e-05, 'samples': 22924288, 'steps': 44773, 'loss/train': 1.724143624305725} 02/26/2022 11:34:07 - INFO - codeparrot_training - Step 44774: {'lr': 1.4481977818386866e-05, 'samples': 22924800, 'steps': 44774, 'loss/train': 1.9280657768249512} 02/26/2022 11:34:10 - INFO - codeparrot_training - Step 44775: {'lr': 1.4476490181942554e-05, 'samples': 22925312, 'steps': 44775, 'loss/train': 2.70141339302063} 02/26/2022 11:34:16 - INFO - codeparrot_training - Step 44776: {'lr': 1.4471003554406197e-05, 'samples': 22925824, 'steps': 44776, 'loss/train': 1.3111261129379272} 02/26/2022 11:34:19 - INFO - codeparrot_training - Step 44777: {'lr': 1.4465517935801254e-05, 'samples': 22926336, 'steps': 44777, 'loss/train': 2.184539794921875} 02/26/2022 11:34:25 - INFO - codeparrot_training - Step 44778: {'lr': 1.4460033326151257e-05, 'samples': 22926848, 'steps': 44778, 'loss/train': 2.2106947898864746} 02/26/2022 11:34:28 - INFO - codeparrot_training - Step 44779: {'lr': 1.4454549725479716e-05, 'samples': 22927360, 'steps': 44779, 'loss/train': 0.666465163230896} 02/26/2022 11:34:36 - INFO - codeparrot_training - Step 44780: {'lr': 1.4449067133810057e-05, 'samples': 22927872, 'steps': 44780, 'loss/train': 1.6523337364196777} 02/26/2022 11:34:39 - INFO - codeparrot_training - Step 44781: {'lr': 1.4443585551165817e-05, 'samples': 22928384, 'steps': 44781, 'loss/train': 1.522513508796692} 02/26/2022 11:34:45 - INFO - codeparrot_training - Step 44782: {'lr': 1.443810497757042e-05, 'samples': 22928896, 'steps': 44782, 'loss/train': 0.9574680328369141} 02/26/2022 11:34:49 - INFO - codeparrot_training - Step 44783: {'lr': 1.4432625413047407e-05, 'samples': 22929408, 'steps': 44783, 'loss/train': 3.3301377296447754} 02/26/2022 11:34:52 - INFO - codeparrot_training - Step 44784: {'lr': 1.4427146857620228e-05, 'samples': 22929920, 'steps': 44784, 'loss/train': 1.077704668045044} 02/26/2022 11:34:58 - INFO - codeparrot_training - Step 44785: {'lr': 1.4421669311312363e-05, 'samples': 22930432, 'steps': 44785, 'loss/train': 1.9786309003829956} 02/26/2022 11:35:02 - INFO - codeparrot_training - Step 44786: {'lr': 1.4416192774147213e-05, 'samples': 22930944, 'steps': 44786, 'loss/train': 2.3678464889526367} 02/26/2022 11:35:05 - INFO - codeparrot_training - Step 44787: {'lr': 1.4410717246148341e-05, 'samples': 22931456, 'steps': 44787, 'loss/train': 1.082016110420227} 02/26/2022 11:35:11 - INFO - codeparrot_training - Step 44788: {'lr': 1.4405242727339145e-05, 'samples': 22931968, 'steps': 44788, 'loss/train': 2.3670239448547363} 02/26/2022 11:35:14 - INFO - codeparrot_training - Step 44789: {'lr': 1.4399769217743052e-05, 'samples': 22932480, 'steps': 44789, 'loss/train': 2.005310297012329} 02/26/2022 11:35:20 - INFO - codeparrot_training - Step 44790: {'lr': 1.4394296717383516e-05, 'samples': 22932992, 'steps': 44790, 'loss/train': 1.7135089635849} 02/26/2022 11:35:23 - INFO - codeparrot_training - Step 44791: {'lr': 1.438882522628404e-05, 'samples': 22933504, 'steps': 44791, 'loss/train': 1.2183119058609009} 02/26/2022 11:35:29 - INFO - codeparrot_training - Step 44792: {'lr': 1.4383354744468031e-05, 'samples': 22934016, 'steps': 44792, 'loss/train': 1.9528683423995972} 02/26/2022 11:35:32 - INFO - codeparrot_training - Step 44793: {'lr': 1.4377885271958907e-05, 'samples': 22934528, 'steps': 44793, 'loss/train': 3.095869779586792} 02/26/2022 11:35:38 - INFO - codeparrot_training - Step 44794: {'lr': 1.4372416808780042e-05, 'samples': 22935040, 'steps': 44794, 'loss/train': 1.505162000656128} 02/26/2022 11:35:41 - INFO - codeparrot_training - Step 44795: {'lr': 1.4366949354955e-05, 'samples': 22935552, 'steps': 44795, 'loss/train': 2.209728479385376} 02/26/2022 11:35:47 - INFO - codeparrot_training - Step 44796: {'lr': 1.4361482910507095e-05, 'samples': 22936064, 'steps': 44796, 'loss/train': 1.2412135601043701} 02/26/2022 11:35:51 - INFO - codeparrot_training - Step 44797: {'lr': 1.4356017475459782e-05, 'samples': 22936576, 'steps': 44797, 'loss/train': 0.9276085495948792} 02/26/2022 11:35:58 - INFO - codeparrot_training - Step 44798: {'lr': 1.435055304983643e-05, 'samples': 22937088, 'steps': 44798, 'loss/train': 0.4544755220413208} 02/26/2022 11:36:01 - INFO - codeparrot_training - Step 44799: {'lr': 1.4345089633660519e-05, 'samples': 22937600, 'steps': 44799, 'loss/train': 1.5171109437942505} 02/26/2022 11:36:07 - INFO - codeparrot_training - Step 44800: {'lr': 1.4339627226955393e-05, 'samples': 22938112, 'steps': 44800, 'loss/train': 1.1359609365463257} 02/26/2022 11:36:10 - INFO - codeparrot_training - Step 44801: {'lr': 1.4334165829744534e-05, 'samples': 22938624, 'steps': 44801, 'loss/train': 0.08711253851652145} 02/26/2022 11:36:16 - INFO - codeparrot_training - Step 44802: {'lr': 1.4328705442051199e-05, 'samples': 22939136, 'steps': 44802, 'loss/train': 2.0219101905822754} 02/26/2022 11:36:19 - INFO - codeparrot_training - Step 44803: {'lr': 1.4323246063898898e-05, 'samples': 22939648, 'steps': 44803, 'loss/train': 1.6056190729141235} 02/26/2022 11:36:25 - INFO - codeparrot_training - Step 44804: {'lr': 1.4317787695310918e-05, 'samples': 22940160, 'steps': 44804, 'loss/train': 1.6616590023040771} 02/26/2022 11:36:28 - INFO - codeparrot_training - Step 44805: {'lr': 1.4312330336310797e-05, 'samples': 22940672, 'steps': 44805, 'loss/train': 2.397465229034424} 02/26/2022 11:36:34 - INFO - codeparrot_training - Step 44806: {'lr': 1.4306873986921737e-05, 'samples': 22941184, 'steps': 44806, 'loss/train': 2.635054111480713} 02/26/2022 11:36:37 - INFO - codeparrot_training - Step 44807: {'lr': 1.430141864716722e-05, 'samples': 22941696, 'steps': 44807, 'loss/train': 1.749933123588562} 02/26/2022 11:36:45 - INFO - codeparrot_training - Step 44808: {'lr': 1.429596431707053e-05, 'samples': 22942208, 'steps': 44808, 'loss/train': 2.2393569946289062} 02/26/2022 11:36:48 - INFO - codeparrot_training - Step 44809: {'lr': 1.4290510996655182e-05, 'samples': 22942720, 'steps': 44809, 'loss/train': 1.040702223777771} 02/26/2022 11:36:54 - INFO - codeparrot_training - Step 44810: {'lr': 1.4285058685944347e-05, 'samples': 22943232, 'steps': 44810, 'loss/train': 1.6356934309005737} 02/26/2022 11:36:57 - INFO - codeparrot_training - Step 44811: {'lr': 1.4279607384961535e-05, 'samples': 22943744, 'steps': 44811, 'loss/train': 2.3489394187927246} 02/26/2022 11:37:03 - INFO - codeparrot_training - Step 44812: {'lr': 1.4274157093729978e-05, 'samples': 22944256, 'steps': 44812, 'loss/train': 1.4525930881500244} 02/26/2022 11:37:06 - INFO - codeparrot_training - Step 44813: {'lr': 1.4268707812273102e-05, 'samples': 22944768, 'steps': 44813, 'loss/train': 1.2331892251968384} 02/26/2022 11:37:12 - INFO - codeparrot_training - Step 44814: {'lr': 1.426325954061425e-05, 'samples': 22945280, 'steps': 44814, 'loss/train': 1.5273914337158203} 02/26/2022 11:37:15 - INFO - codeparrot_training - Step 44815: {'lr': 1.4257812278776761e-05, 'samples': 22945792, 'steps': 44815, 'loss/train': 1.8169147968292236} 02/26/2022 11:37:21 - INFO - codeparrot_training - Step 44816: {'lr': 1.425236602678387e-05, 'samples': 22946304, 'steps': 44816, 'loss/train': 1.6976604461669922} 02/26/2022 11:37:24 - INFO - codeparrot_training - Step 44817: {'lr': 1.4246920784659029e-05, 'samples': 22946816, 'steps': 44817, 'loss/train': 2.6473615169525146} 02/26/2022 11:37:32 - INFO - codeparrot_training - Step 44818: {'lr': 1.4241476552425552e-05, 'samples': 22947328, 'steps': 44818, 'loss/train': 2.080000877380371} 02/26/2022 11:37:35 - INFO - codeparrot_training - Step 44819: {'lr': 1.4236033330106701e-05, 'samples': 22947840, 'steps': 44819, 'loss/train': 2.176170825958252} 02/26/2022 11:37:41 - INFO - codeparrot_training - Step 44820: {'lr': 1.4230591117725788e-05, 'samples': 22948352, 'steps': 44820, 'loss/train': 1.4614702463150024} 02/26/2022 11:37:44 - INFO - codeparrot_training - Step 44821: {'lr': 1.4225149915306184e-05, 'samples': 22948864, 'steps': 44821, 'loss/train': 1.2352185249328613} 02/26/2022 11:37:50 - INFO - codeparrot_training - Step 44822: {'lr': 1.4219709722871177e-05, 'samples': 22949376, 'steps': 44822, 'loss/train': 1.61871337890625} 02/26/2022 11:37:53 - INFO - codeparrot_training - Step 44823: {'lr': 1.421427054044408e-05, 'samples': 22949888, 'steps': 44823, 'loss/train': 0.9498617053031921} 02/26/2022 11:37:59 - INFO - codeparrot_training - Step 44824: {'lr': 1.4208832368048124e-05, 'samples': 22950400, 'steps': 44824, 'loss/train': 1.7852410078048706} 02/26/2022 11:38:02 - INFO - codeparrot_training - Step 44825: {'lr': 1.4203395205706681e-05, 'samples': 22950912, 'steps': 44825, 'loss/train': 2.3695123195648193} 02/26/2022 11:38:07 - INFO - codeparrot_training - Step 44826: {'lr': 1.4197959053443039e-05, 'samples': 22951424, 'steps': 44826, 'loss/train': 1.4653220176696777} 02/26/2022 11:38:11 - INFO - codeparrot_training - Step 44827: {'lr': 1.419252391128048e-05, 'samples': 22951936, 'steps': 44827, 'loss/train': 1.5267162322998047} 02/26/2022 11:38:18 - INFO - codeparrot_training - Step 44828: {'lr': 1.4187089779242212e-05, 'samples': 22952448, 'steps': 44828, 'loss/train': 1.3887460231781006} 02/26/2022 11:38:22 - INFO - codeparrot_training - Step 44829: {'lr': 1.4181656657351604e-05, 'samples': 22952960, 'steps': 44829, 'loss/train': 1.1708043813705444} 02/26/2022 11:38:27 - INFO - codeparrot_training - Step 44830: {'lr': 1.4176224545631917e-05, 'samples': 22953472, 'steps': 44830, 'loss/train': 0.2806812822818756} 02/26/2022 11:38:31 - INFO - codeparrot_training - Step 44831: {'lr': 1.4170793444106405e-05, 'samples': 22953984, 'steps': 44831, 'loss/train': 3.100749969482422} 02/26/2022 11:38:36 - INFO - codeparrot_training - Step 44832: {'lr': 1.4165363352798305e-05, 'samples': 22954496, 'steps': 44832, 'loss/train': 1.7939221858978271} 02/26/2022 11:38:40 - INFO - codeparrot_training - Step 44833: {'lr': 1.4159934271730873e-05, 'samples': 22955008, 'steps': 44833, 'loss/train': 1.3883085250854492} 02/26/2022 11:38:45 - INFO - codeparrot_training - Step 44834: {'lr': 1.4154506200927453e-05, 'samples': 22955520, 'steps': 44834, 'loss/train': 1.2299606800079346} 02/26/2022 11:38:49 - INFO - codeparrot_training - Step 44835: {'lr': 1.4149079140411163e-05, 'samples': 22956032, 'steps': 44835, 'loss/train': 2.4087371826171875} 02/26/2022 11:38:54 - INFO - codeparrot_training - Step 44836: {'lr': 1.4143653090205432e-05, 'samples': 22956544, 'steps': 44836, 'loss/train': 3.22493577003479} 02/26/2022 11:38:58 - INFO - codeparrot_training - Step 44837: {'lr': 1.4138228050333323e-05, 'samples': 22957056, 'steps': 44837, 'loss/train': 1.6801273822784424} 02/26/2022 11:39:03 - INFO - codeparrot_training - Step 44838: {'lr': 1.413280402081818e-05, 'samples': 22957568, 'steps': 44838, 'loss/train': 0.3348790109157562} 02/26/2022 11:39:07 - INFO - codeparrot_training - Step 44839: {'lr': 1.4127381001683176e-05, 'samples': 22958080, 'steps': 44839, 'loss/train': 1.442067265510559} 02/26/2022 11:39:12 - INFO - codeparrot_training - Step 44840: {'lr': 1.4121958992951628e-05, 'samples': 22958592, 'steps': 44840, 'loss/train': 2.6171531677246094} 02/26/2022 11:39:16 - INFO - codeparrot_training - Step 44841: {'lr': 1.4116537994646655e-05, 'samples': 22959104, 'steps': 44841, 'loss/train': 1.5465601682662964} 02/26/2022 11:39:21 - INFO - codeparrot_training - Step 44842: {'lr': 1.4111118006791573e-05, 'samples': 22959616, 'steps': 44842, 'loss/train': 1.819445013999939} 02/26/2022 11:39:25 - INFO - codeparrot_training - Step 44843: {'lr': 1.4105699029409502e-05, 'samples': 22960128, 'steps': 44843, 'loss/train': 1.1812183856964111} 02/26/2022 11:39:32 - INFO - codeparrot_training - Step 44844: {'lr': 1.4100281062523812e-05, 'samples': 22960640, 'steps': 44844, 'loss/train': 0.9964702129364014} 02/26/2022 11:39:35 - INFO - codeparrot_training - Step 44845: {'lr': 1.4094864106157512e-05, 'samples': 22961152, 'steps': 44845, 'loss/train': 2.3635566234588623} 02/26/2022 11:39:41 - INFO - codeparrot_training - Step 44846: {'lr': 1.4089448160333946e-05, 'samples': 22961664, 'steps': 44846, 'loss/train': 0.991076648235321} 02/26/2022 11:39:45 - INFO - codeparrot_training - Step 44847: {'lr': 1.4084033225076259e-05, 'samples': 22962176, 'steps': 44847, 'loss/train': 1.69138765335083} 02/26/2022 11:39:50 - INFO - codeparrot_training - Step 44848: {'lr': 1.4078619300407685e-05, 'samples': 22962688, 'steps': 44848, 'loss/train': 1.7338188886642456} 02/26/2022 11:39:53 - INFO - codeparrot_training - Step 44849: {'lr': 1.4073206386351372e-05, 'samples': 22963200, 'steps': 44849, 'loss/train': 1.3852802515029907} 02/26/2022 11:39:59 - INFO - codeparrot_training - Step 44850: {'lr': 1.406779448293055e-05, 'samples': 22963712, 'steps': 44850, 'loss/train': 2.29586124420166} 02/26/2022 11:40:02 - INFO - codeparrot_training - Step 44851: {'lr': 1.4062383590168342e-05, 'samples': 22964224, 'steps': 44851, 'loss/train': 2.5437958240509033} 02/26/2022 11:40:08 - INFO - codeparrot_training - Step 44852: {'lr': 1.4056973708088006e-05, 'samples': 22964736, 'steps': 44852, 'loss/train': 2.7314047813415527} 02/26/2022 11:40:12 - INFO - codeparrot_training - Step 44853: {'lr': 1.405156483671266e-05, 'samples': 22965248, 'steps': 44853, 'loss/train': 2.251418352127075} 02/26/2022 11:40:19 - INFO - codeparrot_training - Step 44854: {'lr': 1.4046156976065483e-05, 'samples': 22965760, 'steps': 44854, 'loss/train': 0.707883358001709} 02/26/2022 11:40:23 - INFO - codeparrot_training - Step 44855: {'lr': 1.4040750126169621e-05, 'samples': 22966272, 'steps': 44855, 'loss/train': 1.8748682737350464} 02/26/2022 11:40:29 - INFO - codeparrot_training - Step 44856: {'lr': 1.4035344287048307e-05, 'samples': 22966784, 'steps': 44856, 'loss/train': 1.1896917819976807} 02/26/2022 11:40:32 - INFO - codeparrot_training - Step 44857: {'lr': 1.4029939458724632e-05, 'samples': 22967296, 'steps': 44857, 'loss/train': 1.3778964281082153} 02/26/2022 11:40:38 - INFO - codeparrot_training - Step 44858: {'lr': 1.4024535641221775e-05, 'samples': 22967808, 'steps': 44858, 'loss/train': 1.0070184469223022} 02/26/2022 11:40:41 - INFO - codeparrot_training - Step 44859: {'lr': 1.4019132834562825e-05, 'samples': 22968320, 'steps': 44859, 'loss/train': 2.2810237407684326} 02/26/2022 11:40:47 - INFO - codeparrot_training - Step 44860: {'lr': 1.4013731038771043e-05, 'samples': 22968832, 'steps': 44860, 'loss/train': 2.036309242248535} 02/26/2022 11:40:50 - INFO - codeparrot_training - Step 44861: {'lr': 1.4008330253869438e-05, 'samples': 22969344, 'steps': 44861, 'loss/train': 0.9027878642082214} 02/26/2022 11:40:56 - INFO - codeparrot_training - Step 44862: {'lr': 1.4002930479881325e-05, 'samples': 22969856, 'steps': 44862, 'loss/train': 1.34867525100708} 02/26/2022 11:40:59 - INFO - codeparrot_training - Step 44863: {'lr': 1.3997531716829603e-05, 'samples': 22970368, 'steps': 44863, 'loss/train': 2.2102622985839844} 02/26/2022 11:41:07 - INFO - codeparrot_training - Step 44864: {'lr': 1.3992133964737585e-05, 'samples': 22970880, 'steps': 44864, 'loss/train': 1.5386334657669067} 02/26/2022 11:41:10 - INFO - codeparrot_training - Step 44865: {'lr': 1.398673722362831e-05, 'samples': 22971392, 'steps': 44865, 'loss/train': 1.4795500040054321} 02/26/2022 11:41:16 - INFO - codeparrot_training - Step 44866: {'lr': 1.3981341493524897e-05, 'samples': 22971904, 'steps': 44866, 'loss/train': 2.3604021072387695} 02/26/2022 11:41:19 - INFO - codeparrot_training - Step 44867: {'lr': 1.3975946774450437e-05, 'samples': 22972416, 'steps': 44867, 'loss/train': 1.7739827632904053} 02/26/2022 11:41:25 - INFO - codeparrot_training - Step 44868: {'lr': 1.3970553066428138e-05, 'samples': 22972928, 'steps': 44868, 'loss/train': 0.7035548686981201} 02/26/2022 11:41:28 - INFO - codeparrot_training - Step 44869: {'lr': 1.3965160369481034e-05, 'samples': 22973440, 'steps': 44869, 'loss/train': 1.1788643598556519} 02/26/2022 11:41:34 - INFO - codeparrot_training - Step 44870: {'lr': 1.3959768683632218e-05, 'samples': 22973952, 'steps': 44870, 'loss/train': 0.8870359659194946} 02/26/2022 11:41:37 - INFO - codeparrot_training - Step 44871: {'lr': 1.3954378008904755e-05, 'samples': 22974464, 'steps': 44871, 'loss/train': 0.2659461498260498} 02/26/2022 11:41:43 - INFO - codeparrot_training - Step 44872: {'lr': 1.394898834532185e-05, 'samples': 22974976, 'steps': 44872, 'loss/train': 0.4549286663532257} 02/26/2022 11:41:46 - INFO - codeparrot_training - Step 44873: {'lr': 1.394359969290651e-05, 'samples': 22975488, 'steps': 44873, 'loss/train': 1.9302061796188354} 02/26/2022 11:41:52 - INFO - codeparrot_training - Step 44874: {'lr': 1.3938212051681803e-05, 'samples': 22976000, 'steps': 44874, 'loss/train': 0.11666565388441086} 02/26/2022 11:41:55 - INFO - codeparrot_training - Step 44875: {'lr': 1.3932825421670875e-05, 'samples': 22976512, 'steps': 44875, 'loss/train': 1.7317155599594116} 02/26/2022 11:42:03 - INFO - codeparrot_training - Step 44876: {'lr': 1.3927439802896762e-05, 'samples': 22977024, 'steps': 44876, 'loss/train': 1.5664552450180054} 02/26/2022 11:42:06 - INFO - codeparrot_training - Step 44877: {'lr': 1.3922055195382532e-05, 'samples': 22977536, 'steps': 44877, 'loss/train': 1.6684882640838623} 02/26/2022 11:42:12 - INFO - codeparrot_training - Step 44878: {'lr': 1.3916671599151192e-05, 'samples': 22978048, 'steps': 44878, 'loss/train': 2.0447936058044434} 02/26/2022 11:42:15 - INFO - codeparrot_training - Step 44879: {'lr': 1.3911289014225976e-05, 'samples': 22978560, 'steps': 44879, 'loss/train': 1.8103251457214355} 02/26/2022 11:42:21 - INFO - codeparrot_training - Step 44880: {'lr': 1.390590744062975e-05, 'samples': 22979072, 'steps': 44880, 'loss/train': 2.3412225246429443} 02/26/2022 11:42:24 - INFO - codeparrot_training - Step 44881: {'lr': 1.3900526878385666e-05, 'samples': 22979584, 'steps': 44881, 'loss/train': 2.4388139247894287} 02/26/2022 11:42:30 - INFO - codeparrot_training - Step 44882: {'lr': 1.3895147327516733e-05, 'samples': 22980096, 'steps': 44882, 'loss/train': 1.9126757383346558} 02/26/2022 11:42:33 - INFO - codeparrot_training - Step 44883: {'lr': 1.3889768788046097e-05, 'samples': 22980608, 'steps': 44883, 'loss/train': 2.176185131072998} 02/26/2022 11:42:39 - INFO - codeparrot_training - Step 44884: {'lr': 1.3884391259996632e-05, 'samples': 22981120, 'steps': 44884, 'loss/train': 1.6940706968307495} 02/26/2022 11:42:42 - INFO - codeparrot_training - Step 44885: {'lr': 1.387901474339151e-05, 'samples': 22981632, 'steps': 44885, 'loss/train': 1.2070517539978027} 02/26/2022 11:42:48 - INFO - codeparrot_training - Step 44886: {'lr': 1.387363923825366e-05, 'samples': 22982144, 'steps': 44886, 'loss/train': 0.8737601041793823} 02/26/2022 11:42:51 - INFO - codeparrot_training - Step 44887: {'lr': 1.3868264744606229e-05, 'samples': 22982656, 'steps': 44887, 'loss/train': 1.9998855590820312} 02/26/2022 11:42:57 - INFO - codeparrot_training - Step 44888: {'lr': 1.3862891262472144e-05, 'samples': 22983168, 'steps': 44888, 'loss/train': 2.0377228260040283} 02/26/2022 11:43:00 - INFO - codeparrot_training - Step 44889: {'lr': 1.3857518791874441e-05, 'samples': 22983680, 'steps': 44889, 'loss/train': 0.19210028648376465} 02/26/2022 11:43:07 - INFO - codeparrot_training - Step 44890: {'lr': 1.385214733283613e-05, 'samples': 22984192, 'steps': 44890, 'loss/train': 0.9174827933311462} 02/26/2022 11:43:11 - INFO - codeparrot_training - Step 44891: {'lr': 1.3846776885380247e-05, 'samples': 22984704, 'steps': 44891, 'loss/train': 1.795182466506958} 02/26/2022 11:43:16 - INFO - codeparrot_training - Step 44892: {'lr': 1.3841407449529803e-05, 'samples': 22985216, 'steps': 44892, 'loss/train': 0.7177965044975281} 02/26/2022 11:43:20 - INFO - codeparrot_training - Step 44893: {'lr': 1.3836039025307778e-05, 'samples': 22985728, 'steps': 44893, 'loss/train': 1.2636419534683228} 02/26/2022 11:43:25 - INFO - codeparrot_training - Step 44894: {'lr': 1.3830671612737156e-05, 'samples': 22986240, 'steps': 44894, 'loss/train': 2.26444411277771} 02/26/2022 11:43:29 - INFO - codeparrot_training - Step 44895: {'lr': 1.3825305211840944e-05, 'samples': 22986752, 'steps': 44895, 'loss/train': 1.291749358177185} 02/26/2022 11:43:34 - INFO - codeparrot_training - Step 44896: {'lr': 1.3819939822642125e-05, 'samples': 22987264, 'steps': 44896, 'loss/train': 1.6828207969665527} 02/26/2022 11:43:38 - INFO - codeparrot_training - Step 44897: {'lr': 1.3814575445163735e-05, 'samples': 22987776, 'steps': 44897, 'loss/train': 2.33162522315979} 02/26/2022 11:43:43 - INFO - codeparrot_training - Step 44898: {'lr': 1.3809212079428645e-05, 'samples': 22988288, 'steps': 44898, 'loss/train': 2.1885838508605957} 02/26/2022 11:43:47 - INFO - codeparrot_training - Step 44899: {'lr': 1.3803849725459922e-05, 'samples': 22988800, 'steps': 44899, 'loss/train': 2.8827908039093018} 02/26/2022 11:43:54 - INFO - codeparrot_training - Step 44900: {'lr': 1.3798488383280488e-05, 'samples': 22989312, 'steps': 44900, 'loss/train': 1.083183765411377} 02/26/2022 11:43:57 - INFO - codeparrot_training - Step 44901: {'lr': 1.3793128052913384e-05, 'samples': 22989824, 'steps': 44901, 'loss/train': 1.840578556060791} 02/26/2022 11:44:03 - INFO - codeparrot_training - Step 44902: {'lr': 1.3787768734381451e-05, 'samples': 22990336, 'steps': 44902, 'loss/train': 1.6816685199737549} 02/26/2022 11:44:09 - INFO - codeparrot_training - Step 44903: {'lr': 1.3782410427707726e-05, 'samples': 22990848, 'steps': 44903, 'loss/train': 1.649601697921753} 02/26/2022 11:44:12 - INFO - codeparrot_training - Step 44904: {'lr': 1.3777053132915135e-05, 'samples': 22991360, 'steps': 44904, 'loss/train': 0.4503486454486847} 02/26/2022 11:44:15 - INFO - codeparrot_training - Step 44905: {'lr': 1.3771696850026688e-05, 'samples': 22991872, 'steps': 44905, 'loss/train': 1.1975904703140259} 02/26/2022 11:44:21 - INFO - codeparrot_training - Step 44906: {'lr': 1.3766341579065227e-05, 'samples': 22992384, 'steps': 44906, 'loss/train': 1.4130488634109497} 02/26/2022 11:44:26 - INFO - codeparrot_training - Step 44907: {'lr': 1.3760987320053791e-05, 'samples': 22992896, 'steps': 44907, 'loss/train': 1.5262269973754883} 02/26/2022 11:44:30 - INFO - codeparrot_training - Step 44908: {'lr': 1.3755634073015221e-05, 'samples': 22993408, 'steps': 44908, 'loss/train': 1.9153634309768677} 02/26/2022 11:44:36 - INFO - codeparrot_training - Step 44909: {'lr': 1.3750281837972528e-05, 'samples': 22993920, 'steps': 44909, 'loss/train': 2.522473096847534} 02/26/2022 11:44:39 - INFO - codeparrot_training - Step 44910: {'lr': 1.3744930614948608e-05, 'samples': 22994432, 'steps': 44910, 'loss/train': 0.5706866979598999} 02/26/2022 11:44:46 - INFO - codeparrot_training - Step 44911: {'lr': 1.3739580403966418e-05, 'samples': 22994944, 'steps': 44911, 'loss/train': 1.0238890647888184} 02/26/2022 11:44:50 - INFO - codeparrot_training - Step 44912: {'lr': 1.3734231205048826e-05, 'samples': 22995456, 'steps': 44912, 'loss/train': 1.5381544828414917} 02/26/2022 11:44:55 - INFO - codeparrot_training - Step 44913: {'lr': 1.3728883018218701e-05, 'samples': 22995968, 'steps': 44913, 'loss/train': 0.21254390478134155} 02/26/2022 11:44:59 - INFO - codeparrot_training - Step 44914: {'lr': 1.3723535843499086e-05, 'samples': 22996480, 'steps': 44914, 'loss/train': 2.314822196960449} 02/26/2022 11:45:05 - INFO - codeparrot_training - Step 44915: {'lr': 1.3718189680912819e-05, 'samples': 22996992, 'steps': 44915, 'loss/train': 1.8591642379760742} 02/26/2022 11:45:08 - INFO - codeparrot_training - Step 44916: {'lr': 1.3712844530482771e-05, 'samples': 22997504, 'steps': 44916, 'loss/train': 1.4261101484298706} 02/26/2022 11:45:11 - INFO - codeparrot_training - Step 44917: {'lr': 1.370750039223187e-05, 'samples': 22998016, 'steps': 44917, 'loss/train': 2.20125412940979} 02/26/2022 11:45:17 - INFO - codeparrot_training - Step 44918: {'lr': 1.3702157266183012e-05, 'samples': 22998528, 'steps': 44918, 'loss/train': 0.04204076528549194} 02/26/2022 11:45:23 - INFO - codeparrot_training - Step 44919: {'lr': 1.369681515235907e-05, 'samples': 22999040, 'steps': 44919, 'loss/train': 1.4420565366744995} 02/26/2022 11:45:26 - INFO - codeparrot_training - Step 44920: {'lr': 1.369147405078297e-05, 'samples': 22999552, 'steps': 44920, 'loss/train': 1.529695749282837} 02/26/2022 11:45:33 - INFO - codeparrot_training - Step 44921: {'lr': 1.3686133961477498e-05, 'samples': 23000064, 'steps': 44921, 'loss/train': 2.391444683074951} 02/26/2022 11:45:37 - INFO - codeparrot_training - Step 44922: {'lr': 1.3680794884465635e-05, 'samples': 23000576, 'steps': 44922, 'loss/train': 2.1998631954193115} 02/26/2022 11:45:40 - INFO - codeparrot_training - Step 44923: {'lr': 1.3675456819770227e-05, 'samples': 23001088, 'steps': 44923, 'loss/train': 1.3627636432647705} 02/26/2022 11:45:46 - INFO - codeparrot_training - Step 44924: {'lr': 1.3670119767414085e-05, 'samples': 23001600, 'steps': 44924, 'loss/train': 0.7539050579071045} 02/26/2022 11:45:52 - INFO - codeparrot_training - Step 44925: {'lr': 1.3664783727420083e-05, 'samples': 23002112, 'steps': 44925, 'loss/train': 1.9430809020996094} 02/26/2022 11:45:55 - INFO - codeparrot_training - Step 44926: {'lr': 1.3659448699811145e-05, 'samples': 23002624, 'steps': 44926, 'loss/train': 1.7160112857818604} 02/26/2022 11:46:01 - INFO - codeparrot_training - Step 44927: {'lr': 1.3654114684610085e-05, 'samples': 23003136, 'steps': 44927, 'loss/train': 1.6460176706314087} 02/26/2022 11:46:04 - INFO - codeparrot_training - Step 44928: {'lr': 1.364878168183975e-05, 'samples': 23003648, 'steps': 44928, 'loss/train': 1.4424197673797607} 02/26/2022 11:46:10 - INFO - codeparrot_training - Step 44929: {'lr': 1.3643449691522953e-05, 'samples': 23004160, 'steps': 44929, 'loss/train': 1.2292176485061646} 02/26/2022 11:46:13 - INFO - codeparrot_training - Step 44930: {'lr': 1.363811871368259e-05, 'samples': 23004672, 'steps': 44930, 'loss/train': 1.455814003944397} 02/26/2022 11:46:19 - INFO - codeparrot_training - Step 44931: {'lr': 1.3632788748341452e-05, 'samples': 23005184, 'steps': 44931, 'loss/train': 2.065498113632202} 02/26/2022 11:46:22 - INFO - codeparrot_training - Step 44932: {'lr': 1.3627459795522435e-05, 'samples': 23005696, 'steps': 44932, 'loss/train': 1.2602853775024414} 02/26/2022 11:46:28 - INFO - codeparrot_training - Step 44933: {'lr': 1.3622131855248243e-05, 'samples': 23006208, 'steps': 44933, 'loss/train': 1.2619924545288086} 02/26/2022 11:46:31 - INFO - codeparrot_training - Step 44934: {'lr': 1.361680492754186e-05, 'samples': 23006720, 'steps': 44934, 'loss/train': 1.9449198246002197} 02/26/2022 11:46:37 - INFO - codeparrot_training - Step 44935: {'lr': 1.3611479012425959e-05, 'samples': 23007232, 'steps': 44935, 'loss/train': 1.1407411098480225} 02/26/2022 11:46:40 - INFO - codeparrot_training - Step 44936: {'lr': 1.3606154109923497e-05, 'samples': 23007744, 'steps': 44936, 'loss/train': 1.7550444602966309} 02/26/2022 11:46:47 - INFO - codeparrot_training - Step 44937: {'lr': 1.3600830220057148e-05, 'samples': 23008256, 'steps': 44937, 'loss/train': 1.7893890142440796} 02/26/2022 11:46:51 - INFO - codeparrot_training - Step 44938: {'lr': 1.359550734284981e-05, 'samples': 23008768, 'steps': 44938, 'loss/train': 2.1626994609832764} 02/26/2022 11:46:56 - INFO - codeparrot_training - Step 44939: {'lr': 1.359018547832419e-05, 'samples': 23009280, 'steps': 44939, 'loss/train': 2.213036060333252} 02/26/2022 11:47:00 - INFO - codeparrot_training - Step 44940: {'lr': 1.3584864626503241e-05, 'samples': 23009792, 'steps': 44940, 'loss/train': 2.429441213607788} 02/26/2022 11:47:05 - INFO - codeparrot_training - Step 44941: {'lr': 1.3579544787409608e-05, 'samples': 23010304, 'steps': 44941, 'loss/train': 0.3529791235923767} 02/26/2022 11:47:09 - INFO - codeparrot_training - Step 44942: {'lr': 1.357422596106614e-05, 'samples': 23010816, 'steps': 44942, 'loss/train': 1.4843324422836304} 02/26/2022 11:47:14 - INFO - codeparrot_training - Step 44943: {'lr': 1.356890814749559e-05, 'samples': 23011328, 'steps': 44943, 'loss/train': 2.6002357006073} 02/26/2022 11:47:18 - INFO - codeparrot_training - Step 44944: {'lr': 1.3563591346720804e-05, 'samples': 23011840, 'steps': 44944, 'loss/train': 2.2776660919189453} 02/26/2022 11:47:23 - INFO - codeparrot_training - Step 44945: {'lr': 1.3558275558764488e-05, 'samples': 23012352, 'steps': 44945, 'loss/train': 2.0558323860168457} 02/26/2022 11:47:26 - INFO - codeparrot_training - Step 44946: {'lr': 1.3552960783649481e-05, 'samples': 23012864, 'steps': 44946, 'loss/train': 1.8615316152572632} 02/26/2022 11:47:34 - INFO - codeparrot_training - Step 44947: {'lr': 1.3547647021398435e-05, 'samples': 23013376, 'steps': 44947, 'loss/train': 1.7510186433792114} 02/26/2022 11:47:37 - INFO - codeparrot_training - Step 44948: {'lr': 1.3542334272034245e-05, 'samples': 23013888, 'steps': 44948, 'loss/train': 2.126037359237671} 02/26/2022 11:47:43 - INFO - codeparrot_training - Step 44949: {'lr': 1.3537022535579618e-05, 'samples': 23014400, 'steps': 44949, 'loss/train': 1.4729446172714233} 02/26/2022 11:47:46 - INFO - codeparrot_training - Step 44950: {'lr': 1.3531711812057284e-05, 'samples': 23014912, 'steps': 44950, 'loss/train': 1.7517380714416504} 02/26/2022 11:47:52 - INFO - codeparrot_training - Step 44951: {'lr': 1.3526402101489977e-05, 'samples': 23015424, 'steps': 44951, 'loss/train': 1.76023268699646} 02/26/2022 11:47:55 - INFO - codeparrot_training - Step 44952: {'lr': 1.3521093403900508e-05, 'samples': 23015936, 'steps': 44952, 'loss/train': 1.7382863759994507} 02/26/2022 11:48:01 - INFO - codeparrot_training - Step 44953: {'lr': 1.3515785719311586e-05, 'samples': 23016448, 'steps': 44953, 'loss/train': 1.0642966032028198} 02/26/2022 11:48:04 - INFO - codeparrot_training - Step 44954: {'lr': 1.3510479047745966e-05, 'samples': 23016960, 'steps': 44954, 'loss/train': 1.8943291902542114} 02/26/2022 11:48:10 - INFO - codeparrot_training - Step 44955: {'lr': 1.35051733892263e-05, 'samples': 23017472, 'steps': 44955, 'loss/train': 2.3570845127105713} 02/26/2022 11:48:13 - INFO - codeparrot_training - Step 44956: {'lr': 1.3499868743775401e-05, 'samples': 23017984, 'steps': 44956, 'loss/train': 1.9444695711135864} 02/26/2022 11:48:20 - INFO - codeparrot_training - Step 44957: {'lr': 1.3494565111416002e-05, 'samples': 23018496, 'steps': 44957, 'loss/train': 1.4855417013168335} 02/26/2022 11:48:24 - INFO - codeparrot_training - Step 44958: {'lr': 1.3489262492170778e-05, 'samples': 23019008, 'steps': 44958, 'loss/train': 0.9789879322052002} 02/26/2022 11:48:29 - INFO - codeparrot_training - Step 44959: {'lr': 1.3483960886062435e-05, 'samples': 23019520, 'steps': 44959, 'loss/train': 1.6357616186141968} 02/26/2022 11:48:33 - INFO - codeparrot_training - Step 44960: {'lr': 1.3478660293113675e-05, 'samples': 23020032, 'steps': 44960, 'loss/train': 1.8695590496063232} 02/26/2022 11:48:38 - INFO - codeparrot_training - Step 44961: {'lr': 1.347336071334726e-05, 'samples': 23020544, 'steps': 44961, 'loss/train': 2.5499942302703857} 02/26/2022 11:48:42 - INFO - codeparrot_training - Step 44962: {'lr': 1.3468062146785865e-05, 'samples': 23021056, 'steps': 44962, 'loss/train': 1.955837607383728} 02/26/2022 11:48:47 - INFO - codeparrot_training - Step 44963: {'lr': 1.3462764593452192e-05, 'samples': 23021568, 'steps': 44963, 'loss/train': 1.4898210763931274} 02/26/2022 11:48:51 - INFO - codeparrot_training - Step 44964: {'lr': 1.3457468053368893e-05, 'samples': 23022080, 'steps': 44964, 'loss/train': 2.160562753677368} 02/26/2022 11:48:57 - INFO - codeparrot_training - Step 44965: {'lr': 1.3452172526558726e-05, 'samples': 23022592, 'steps': 44965, 'loss/train': 1.4715397357940674} 02/26/2022 11:49:00 - INFO - codeparrot_training - Step 44966: {'lr': 1.344687801304431e-05, 'samples': 23023104, 'steps': 44966, 'loss/train': 1.6743314266204834} 02/26/2022 11:49:07 - INFO - codeparrot_training - Step 44967: {'lr': 1.3441584512848381e-05, 'samples': 23023616, 'steps': 44967, 'loss/train': 1.6812307834625244} 02/26/2022 11:49:10 - INFO - codeparrot_training - Step 44968: {'lr': 1.3436292025993558e-05, 'samples': 23024128, 'steps': 44968, 'loss/train': 1.242566466331482} 02/26/2022 11:49:16 - INFO - codeparrot_training - Step 44969: {'lr': 1.343100055250257e-05, 'samples': 23024640, 'steps': 44969, 'loss/train': 0.8106839656829834} 02/26/2022 11:49:19 - INFO - codeparrot_training - Step 44970: {'lr': 1.3425710092398013e-05, 'samples': 23025152, 'steps': 44970, 'loss/train': 1.2639392614364624} 02/26/2022 11:49:25 - INFO - codeparrot_training - Step 44971: {'lr': 1.3420420645702675e-05, 'samples': 23025664, 'steps': 44971, 'loss/train': 1.9180103540420532} 02/26/2022 11:49:28 - INFO - codeparrot_training - Step 44972: {'lr': 1.3415132212439062e-05, 'samples': 23026176, 'steps': 44972, 'loss/train': 2.346921443939209} 02/26/2022 11:49:34 - INFO - codeparrot_training - Step 44973: {'lr': 1.3409844792629939e-05, 'samples': 23026688, 'steps': 44973, 'loss/train': 1.6661607027053833} 02/26/2022 11:49:37 - INFO - codeparrot_training - Step 44974: {'lr': 1.3404558386297867e-05, 'samples': 23027200, 'steps': 44974, 'loss/train': 1.7475236654281616} 02/26/2022 11:49:43 - INFO - codeparrot_training - Step 44975: {'lr': 1.3399272993465605e-05, 'samples': 23027712, 'steps': 44975, 'loss/train': 2.166574716567993} 02/26/2022 11:49:47 - INFO - codeparrot_training - Step 44976: {'lr': 1.3393988614155666e-05, 'samples': 23028224, 'steps': 44976, 'loss/train': 0.2191791534423828} 02/26/2022 11:49:53 - INFO - codeparrot_training - Step 44977: {'lr': 1.3388705248390782e-05, 'samples': 23028736, 'steps': 44977, 'loss/train': 0.2064363956451416} 02/26/2022 11:49:56 - INFO - codeparrot_training - Step 44978: {'lr': 1.3383422896193515e-05, 'samples': 23029248, 'steps': 44978, 'loss/train': 1.958709955215454} 02/26/2022 11:50:02 - INFO - codeparrot_training - Step 44979: {'lr': 1.3378141557586599e-05, 'samples': 23029760, 'steps': 44979, 'loss/train': 0.9487249851226807} 02/26/2022 11:50:05 - INFO - codeparrot_training - Step 44980: {'lr': 1.3372861232592542e-05, 'samples': 23030272, 'steps': 44980, 'loss/train': 0.9992273449897766} 02/26/2022 11:50:11 - INFO - codeparrot_training - Step 44981: {'lr': 1.3367581921234023e-05, 'samples': 23030784, 'steps': 44981, 'loss/train': 0.7073620557785034} 02/26/2022 11:50:14 - INFO - codeparrot_training - Step 44982: {'lr': 1.3362303623533606e-05, 'samples': 23031296, 'steps': 44982, 'loss/train': 1.50754976272583} 02/26/2022 11:50:20 - INFO - codeparrot_training - Step 44983: {'lr': 1.3357026339513967e-05, 'samples': 23031808, 'steps': 44983, 'loss/train': 0.4032500982284546} 02/26/2022 11:50:23 - INFO - codeparrot_training - Step 44984: {'lr': 1.3351750069197699e-05, 'samples': 23032320, 'steps': 44984, 'loss/train': 2.0182955265045166} 02/26/2022 11:50:29 - INFO - codeparrot_training - Step 44985: {'lr': 1.3346474812607396e-05, 'samples': 23032832, 'steps': 44985, 'loss/train': 2.0755176544189453} 02/26/2022 11:50:34 - INFO - codeparrot_training - Step 44986: {'lr': 1.3341200569765621e-05, 'samples': 23033344, 'steps': 44986, 'loss/train': 2.204955577850342} 02/26/2022 11:50:38 - INFO - codeparrot_training - Step 44987: {'lr': 1.3335927340694997e-05, 'samples': 23033856, 'steps': 44987, 'loss/train': 1.1490033864974976} 02/26/2022 11:50:44 - INFO - codeparrot_training - Step 44988: {'lr': 1.3330655125418145e-05, 'samples': 23034368, 'steps': 44988, 'loss/train': 0.2702863812446594} 02/26/2022 11:50:47 - INFO - codeparrot_training - Step 44989: {'lr': 1.3325383923957601e-05, 'samples': 23034880, 'steps': 44989, 'loss/train': 2.4160664081573486} 02/26/2022 11:50:53 - INFO - codeparrot_training - Step 44990: {'lr': 1.3320113736335931e-05, 'samples': 23035392, 'steps': 44990, 'loss/train': 1.412811517715454} 02/26/2022 11:50:56 - INFO - codeparrot_training - Step 44991: {'lr': 1.3314844562575757e-05, 'samples': 23035904, 'steps': 44991, 'loss/train': 1.1590325832366943} 02/26/2022 11:51:02 - INFO - codeparrot_training - Step 44992: {'lr': 1.3309576402699641e-05, 'samples': 23036416, 'steps': 44992, 'loss/train': 1.7436206340789795} 02/26/2022 11:51:05 - INFO - codeparrot_training - Step 44993: {'lr': 1.3304309256730152e-05, 'samples': 23036928, 'steps': 44993, 'loss/train': 2.0711193084716797} 02/26/2022 11:51:11 - INFO - codeparrot_training - Step 44994: {'lr': 1.3299043124689798e-05, 'samples': 23037440, 'steps': 44994, 'loss/train': 1.8494958877563477} 02/26/2022 11:51:14 - INFO - codeparrot_training - Step 44995: {'lr': 1.3293778006601225e-05, 'samples': 23037952, 'steps': 44995, 'loss/train': 1.8960152864456177} 02/26/2022 11:51:20 - INFO - codeparrot_training - Step 44996: {'lr': 1.3288513902486921e-05, 'samples': 23038464, 'steps': 44996, 'loss/train': 1.916840672492981} 02/26/2022 11:51:23 - INFO - codeparrot_training - Step 44997: {'lr': 1.3283250812369474e-05, 'samples': 23038976, 'steps': 44997, 'loss/train': 1.4744246006011963} 02/26/2022 11:51:29 - INFO - codeparrot_training - Step 44998: {'lr': 1.3277988736271397e-05, 'samples': 23039488, 'steps': 44998, 'loss/train': 2.744082450866699} 02/26/2022 11:51:33 - INFO - codeparrot_training - Step 44999: {'lr': 1.327272767421525e-05, 'samples': 23040000, 'steps': 44999, 'loss/train': 2.3884806632995605} 02/26/2022 11:51:33 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 11:51:51 - WARNING - huggingface_hub.repository - Several commits (45) will be pushed upstream. 02/26/2022 11:51:51 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 11:52:41 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 84c8fb0..9a19140 floral-grass-11 -> floral-grass-11 02/26/2022 11:52:49 - INFO - codeparrot_training - Step 45000: {'lr': 1.3267467626223605e-05, 'samples': 23040512, 'steps': 45000, 'loss/train': 2.671565294265747} 02/26/2022 11:52:52 - INFO - codeparrot_training - Step 45001: {'lr': 1.3262208592318942e-05, 'samples': 23041024, 'steps': 45001, 'loss/train': 2.178778886795044} 02/26/2022 11:52:58 - INFO - codeparrot_training - Step 45002: {'lr': 1.3256950572523769e-05, 'samples': 23041536, 'steps': 45002, 'loss/train': 1.4157593250274658} 02/26/2022 11:53:01 - INFO - codeparrot_training - Step 45003: {'lr': 1.3251693566860679e-05, 'samples': 23042048, 'steps': 45003, 'loss/train': 1.25260591506958} 02/26/2022 11:53:07 - INFO - codeparrot_training - Step 45004: {'lr': 1.3246437575352154e-05, 'samples': 23042560, 'steps': 45004, 'loss/train': 2.4184486865997314} 02/26/2022 11:53:10 - INFO - codeparrot_training - Step 45005: {'lr': 1.3241182598020679e-05, 'samples': 23043072, 'steps': 45005, 'loss/train': 2.295644998550415} 02/26/2022 11:53:16 - INFO - codeparrot_training - Step 45006: {'lr': 1.3235928634888871e-05, 'samples': 23043584, 'steps': 45006, 'loss/train': 1.434892177581787} 02/26/2022 11:53:19 - INFO - codeparrot_training - Step 45007: {'lr': 1.3230675685979077e-05, 'samples': 23044096, 'steps': 45007, 'loss/train': 1.694848656654358} 02/26/2022 11:53:25 - INFO - codeparrot_training - Step 45008: {'lr': 1.3225423751313942e-05, 'samples': 23044608, 'steps': 45008, 'loss/train': 1.8372126817703247} 02/26/2022 11:53:28 - INFO - codeparrot_training - Step 45009: {'lr': 1.3220172830915866e-05, 'samples': 23045120, 'steps': 45009, 'loss/train': 1.0034587383270264} 02/26/2022 11:53:34 - INFO - codeparrot_training - Step 45010: {'lr': 1.3214922924807471e-05, 'samples': 23045632, 'steps': 45010, 'loss/train': 1.565590500831604} 02/26/2022 11:53:37 - INFO - codeparrot_training - Step 45011: {'lr': 1.3209674033011072e-05, 'samples': 23046144, 'steps': 45011, 'loss/train': 1.5464892387390137} 02/26/2022 11:53:43 - INFO - codeparrot_training - Step 45012: {'lr': 1.3204426155549259e-05, 'samples': 23046656, 'steps': 45012, 'loss/train': 0.4041423797607422} 02/26/2022 11:53:46 - INFO - codeparrot_training - Step 45013: {'lr': 1.3199179292444491e-05, 'samples': 23047168, 'steps': 45013, 'loss/train': 1.8522764444351196} 02/26/2022 11:53:52 - INFO - codeparrot_training - Step 45014: {'lr': 1.3193933443719274e-05, 'samples': 23047680, 'steps': 45014, 'loss/train': 0.8630092740058899} 02/26/2022 11:53:56 - INFO - codeparrot_training - Step 45015: {'lr': 1.3188688609396011e-05, 'samples': 23048192, 'steps': 45015, 'loss/train': 1.7922227382659912} 02/26/2022 11:54:01 - INFO - codeparrot_training - Step 45016: {'lr': 1.3183444789497262e-05, 'samples': 23048704, 'steps': 45016, 'loss/train': 1.0154199600219727} 02/26/2022 11:54:05 - INFO - codeparrot_training - Step 45017: {'lr': 1.3178201984045373e-05, 'samples': 23049216, 'steps': 45017, 'loss/train': 0.5893858671188354} 02/26/2022 11:54:10 - INFO - codeparrot_training - Step 45018: {'lr': 1.317296019306291e-05, 'samples': 23049728, 'steps': 45018, 'loss/train': 0.25592130422592163} 02/26/2022 11:54:14 - INFO - codeparrot_training - Step 45019: {'lr': 1.3167719416572272e-05, 'samples': 23050240, 'steps': 45019, 'loss/train': 1.3772392272949219} 02/26/2022 11:54:19 - INFO - codeparrot_training - Step 45020: {'lr': 1.3162479654595938e-05, 'samples': 23050752, 'steps': 45020, 'loss/train': 2.304441452026367} 02/26/2022 11:54:23 - INFO - codeparrot_training - Step 45021: {'lr': 1.315724090715631e-05, 'samples': 23051264, 'steps': 45021, 'loss/train': 1.3221042156219482} 02/26/2022 11:54:28 - INFO - codeparrot_training - Step 45022: {'lr': 1.3152003174275867e-05, 'samples': 23051776, 'steps': 45022, 'loss/train': 1.5573818683624268} 02/26/2022 11:54:32 - INFO - codeparrot_training - Step 45023: {'lr': 1.3146766455977066e-05, 'samples': 23052288, 'steps': 45023, 'loss/train': 1.6099047660827637} 02/26/2022 11:54:38 - INFO - codeparrot_training - Step 45024: {'lr': 1.3141530752282276e-05, 'samples': 23052800, 'steps': 45024, 'loss/train': 1.2864468097686768} 02/26/2022 11:54:42 - INFO - codeparrot_training - Step 45025: {'lr': 1.3136296063213954e-05, 'samples': 23053312, 'steps': 45025, 'loss/train': 1.70708429813385} 02/26/2022 11:54:47 - INFO - codeparrot_training - Step 45026: {'lr': 1.3131062388794523e-05, 'samples': 23053824, 'steps': 45026, 'loss/train': 1.533958911895752} 02/26/2022 11:54:51 - INFO - codeparrot_training - Step 45027: {'lr': 1.312582972904644e-05, 'samples': 23054336, 'steps': 45027, 'loss/train': 2.4968132972717285} 02/26/2022 11:54:56 - INFO - codeparrot_training - Step 45028: {'lr': 1.3120598083992074e-05, 'samples': 23054848, 'steps': 45028, 'loss/train': 2.053069591522217} 02/26/2022 11:55:00 - INFO - codeparrot_training - Step 45029: {'lr': 1.3115367453653799e-05, 'samples': 23055360, 'steps': 45029, 'loss/train': 1.4023457765579224} 02/26/2022 11:55:05 - INFO - codeparrot_training - Step 45030: {'lr': 1.311013783805412e-05, 'samples': 23055872, 'steps': 45030, 'loss/train': 1.0853075981140137} 02/26/2022 11:55:09 - INFO - codeparrot_training - Step 45031: {'lr': 1.3104909237215356e-05, 'samples': 23056384, 'steps': 45031, 'loss/train': 2.683966875076294} 02/26/2022 11:55:14 - INFO - codeparrot_training - Step 45032: {'lr': 1.3099681651160018e-05, 'samples': 23056896, 'steps': 45032, 'loss/train': 8.609257698059082} 02/26/2022 11:55:18 - INFO - codeparrot_training - Step 45033: {'lr': 1.3094455079910333e-05, 'samples': 23057408, 'steps': 45033, 'loss/train': 2.4009974002838135} 02/26/2022 11:55:24 - INFO - codeparrot_training - Step 45034: {'lr': 1.3089229523488816e-05, 'samples': 23057920, 'steps': 45034, 'loss/train': 0.835014820098877} 02/26/2022 11:55:28 - INFO - codeparrot_training - Step 45035: {'lr': 1.308400498191778e-05, 'samples': 23058432, 'steps': 45035, 'loss/train': 0.5543079376220703} 02/26/2022 11:55:33 - INFO - codeparrot_training - Step 45036: {'lr': 1.3078781455219708e-05, 'samples': 23058944, 'steps': 45036, 'loss/train': 2.108165979385376} 02/26/2022 11:55:37 - INFO - codeparrot_training - Step 45037: {'lr': 1.307355894341683e-05, 'samples': 23059456, 'steps': 45037, 'loss/train': 2.0037920475006104} 02/26/2022 11:55:42 - INFO - codeparrot_training - Step 45038: {'lr': 1.306833744653163e-05, 'samples': 23059968, 'steps': 45038, 'loss/train': 2.6731386184692383} 02/26/2022 11:55:46 - INFO - codeparrot_training - Step 45039: {'lr': 1.3063116964586425e-05, 'samples': 23060480, 'steps': 45039, 'loss/train': 2.332934856414795} 02/26/2022 11:55:51 - INFO - codeparrot_training - Step 45040: {'lr': 1.305789749760361e-05, 'samples': 23060992, 'steps': 45040, 'loss/train': 2.214817523956299} 02/26/2022 11:55:55 - INFO - codeparrot_training - Step 45041: {'lr': 1.3052679045605476e-05, 'samples': 23061504, 'steps': 45041, 'loss/train': 2.0601751804351807} 02/26/2022 11:56:00 - INFO - codeparrot_training - Step 45042: {'lr': 1.3047461608614475e-05, 'samples': 23062016, 'steps': 45042, 'loss/train': 1.4372636079788208} 02/26/2022 11:56:04 - INFO - codeparrot_training - Step 45043: {'lr': 1.3042245186652895e-05, 'samples': 23062528, 'steps': 45043, 'loss/train': 2.499464511871338} 02/26/2022 11:56:10 - INFO - codeparrot_training - Step 45044: {'lr': 1.3037029779743054e-05, 'samples': 23063040, 'steps': 45044, 'loss/train': 2.0009520053863525} 02/26/2022 11:56:13 - INFO - codeparrot_training - Step 45045: {'lr': 1.3031815387907375e-05, 'samples': 23063552, 'steps': 45045, 'loss/train': 1.4358735084533691} 02/26/2022 11:56:19 - INFO - codeparrot_training - Step 45046: {'lr': 1.3026602011168176e-05, 'samples': 23064064, 'steps': 45046, 'loss/train': 1.6239103078842163} 02/26/2022 11:56:22 - INFO - codeparrot_training - Step 45047: {'lr': 1.3021389649547743e-05, 'samples': 23064576, 'steps': 45047, 'loss/train': 1.9477182626724243} 02/26/2022 11:56:28 - INFO - codeparrot_training - Step 45048: {'lr': 1.3016178303068393e-05, 'samples': 23065088, 'steps': 45048, 'loss/train': 1.7775816917419434} 02/26/2022 11:56:31 - INFO - codeparrot_training - Step 45049: {'lr': 1.3010967971752524e-05, 'samples': 23065600, 'steps': 45049, 'loss/train': 0.8595528602600098} 02/26/2022 11:56:37 - INFO - codeparrot_training - Step 45050: {'lr': 1.3005758655622424e-05, 'samples': 23066112, 'steps': 45050, 'loss/train': 1.3412004709243774} 02/26/2022 11:56:40 - INFO - codeparrot_training - Step 45051: {'lr': 1.300055035470038e-05, 'samples': 23066624, 'steps': 45051, 'loss/train': 0.6212685704231262} 02/26/2022 11:56:46 - INFO - codeparrot_training - Step 45052: {'lr': 1.2995343069008708e-05, 'samples': 23067136, 'steps': 45052, 'loss/train': 1.173230528831482} 02/26/2022 11:56:49 - INFO - codeparrot_training - Step 45053: {'lr': 1.2990136798569752e-05, 'samples': 23067648, 'steps': 45053, 'loss/train': 1.4153577089309692} 02/26/2022 11:56:55 - INFO - codeparrot_training - Step 45054: {'lr': 1.2984931543405771e-05, 'samples': 23068160, 'steps': 45054, 'loss/train': 2.286909580230713} 02/26/2022 11:56:59 - INFO - codeparrot_training - Step 45055: {'lr': 1.2979727303539108e-05, 'samples': 23068672, 'steps': 45055, 'loss/train': 1.2688261270523071} 02/26/2022 11:57:04 - INFO - codeparrot_training - Step 45056: {'lr': 1.2974524078991995e-05, 'samples': 23069184, 'steps': 45056, 'loss/train': 2.0816214084625244} 02/26/2022 11:57:08 - INFO - codeparrot_training - Step 45057: {'lr': 1.2969321869786777e-05, 'samples': 23069696, 'steps': 45057, 'loss/train': 1.7652606964111328} 02/26/2022 11:57:14 - INFO - codeparrot_training - Step 45058: {'lr': 1.296412067594574e-05, 'samples': 23070208, 'steps': 45058, 'loss/train': 1.4285666942596436} 02/26/2022 11:57:17 - INFO - codeparrot_training - Step 45059: {'lr': 1.2958920497491144e-05, 'samples': 23070720, 'steps': 45059, 'loss/train': 0.5469043850898743} 02/26/2022 11:57:23 - INFO - codeparrot_training - Step 45060: {'lr': 1.2953721334445196e-05, 'samples': 23071232, 'steps': 45060, 'loss/train': 1.449249505996704} 02/26/2022 11:57:26 - INFO - codeparrot_training - Step 45061: {'lr': 1.294852318683029e-05, 'samples': 23071744, 'steps': 45061, 'loss/train': 1.708475947380066} 02/26/2022 11:57:32 - INFO - codeparrot_training - Step 45062: {'lr': 1.2943326054668663e-05, 'samples': 23072256, 'steps': 45062, 'loss/train': 2.487797737121582} 02/26/2022 11:57:35 - INFO - codeparrot_training - Step 45063: {'lr': 1.2938129937982519e-05, 'samples': 23072768, 'steps': 45063, 'loss/train': 2.5748367309570312} 02/26/2022 11:57:40 - INFO - codeparrot_training - Step 45064: {'lr': 1.2932934836794114e-05, 'samples': 23073280, 'steps': 45064, 'loss/train': 2.007824659347534} 02/26/2022 11:57:44 - INFO - codeparrot_training - Step 45065: {'lr': 1.2927740751125795e-05, 'samples': 23073792, 'steps': 45065, 'loss/train': 1.6610277891159058} 02/26/2022 11:57:50 - INFO - codeparrot_training - Step 45066: {'lr': 1.2922547680999708e-05, 'samples': 23074304, 'steps': 45066, 'loss/train': 2.152963638305664} 02/26/2022 11:57:53 - INFO - codeparrot_training - Step 45067: {'lr': 1.2917355626438227e-05, 'samples': 23074816, 'steps': 45067, 'loss/train': 1.9496700763702393} 02/26/2022 11:57:59 - INFO - codeparrot_training - Step 45068: {'lr': 1.2912164587463442e-05, 'samples': 23075328, 'steps': 45068, 'loss/train': 1.4794533252716064} 02/26/2022 11:58:02 - INFO - codeparrot_training - Step 45069: {'lr': 1.2906974564097674e-05, 'samples': 23075840, 'steps': 45069, 'loss/train': 1.267409324645996} 02/26/2022 11:58:08 - INFO - codeparrot_training - Step 45070: {'lr': 1.2901785556363122e-05, 'samples': 23076352, 'steps': 45070, 'loss/train': 0.6815871596336365} 02/26/2022 11:58:11 - INFO - codeparrot_training - Step 45071: {'lr': 1.2896597564282104e-05, 'samples': 23076864, 'steps': 45071, 'loss/train': 2.6047310829162598} 02/26/2022 11:58:18 - INFO - codeparrot_training - Step 45072: {'lr': 1.2891410587876712e-05, 'samples': 23077376, 'steps': 45072, 'loss/train': 2.4416463375091553} 02/26/2022 11:58:21 - INFO - codeparrot_training - Step 45073: {'lr': 1.2886224627169235e-05, 'samples': 23077888, 'steps': 45073, 'loss/train': 1.580350637435913} 02/26/2022 11:58:27 - INFO - codeparrot_training - Step 45074: {'lr': 1.2881039682181878e-05, 'samples': 23078400, 'steps': 45074, 'loss/train': 2.1410484313964844} 02/26/2022 11:58:30 - INFO - codeparrot_training - Step 45075: {'lr': 1.28758557529369e-05, 'samples': 23078912, 'steps': 45075, 'loss/train': 2.637892484664917} 02/26/2022 11:58:36 - INFO - codeparrot_training - Step 45076: {'lr': 1.2870672839456394e-05, 'samples': 23079424, 'steps': 45076, 'loss/train': 1.7528022527694702} 02/26/2022 11:58:40 - INFO - codeparrot_training - Step 45077: {'lr': 1.2865490941762676e-05, 'samples': 23079936, 'steps': 45077, 'loss/train': 2.098456621170044} 02/26/2022 11:58:46 - INFO - codeparrot_training - Step 45078: {'lr': 1.2860310059877838e-05, 'samples': 23080448, 'steps': 45078, 'loss/train': 2.971890926361084} 02/26/2022 11:58:49 - INFO - codeparrot_training - Step 45079: {'lr': 1.2855130193824171e-05, 'samples': 23080960, 'steps': 45079, 'loss/train': 1.1212985515594482} 02/26/2022 11:58:56 - INFO - codeparrot_training - Step 45080: {'lr': 1.284995134362385e-05, 'samples': 23081472, 'steps': 45080, 'loss/train': 1.6538738012313843} 02/26/2022 11:58:59 - INFO - codeparrot_training - Step 45081: {'lr': 1.2844773509299024e-05, 'samples': 23081984, 'steps': 45081, 'loss/train': 2.364436149597168} 02/26/2022 11:59:05 - INFO - codeparrot_training - Step 45082: {'lr': 1.2839596690871841e-05, 'samples': 23082496, 'steps': 45082, 'loss/train': 1.856343150138855} 02/26/2022 11:59:08 - INFO - codeparrot_training - Step 45083: {'lr': 1.2834420888364535e-05, 'samples': 23083008, 'steps': 45083, 'loss/train': 2.1124069690704346} 02/26/2022 11:59:14 - INFO - codeparrot_training - Step 45084: {'lr': 1.282924610179928e-05, 'samples': 23083520, 'steps': 45084, 'loss/train': 1.8365552425384521} 02/26/2022 11:59:17 - INFO - codeparrot_training - Step 45085: {'lr': 1.2824072331198227e-05, 'samples': 23084032, 'steps': 45085, 'loss/train': 1.9485359191894531} 02/26/2022 11:59:23 - INFO - codeparrot_training - Step 45086: {'lr': 1.2818899576583526e-05, 'samples': 23084544, 'steps': 45086, 'loss/train': 0.6901110410690308} 02/26/2022 11:59:26 - INFO - codeparrot_training - Step 45087: {'lr': 1.2813727837977324e-05, 'samples': 23085056, 'steps': 45087, 'loss/train': 1.8562246561050415} 02/26/2022 11:59:32 - INFO - codeparrot_training - Step 45088: {'lr': 1.2808557115401826e-05, 'samples': 23085568, 'steps': 45088, 'loss/train': 1.5247613191604614} 02/26/2022 11:59:35 - INFO - codeparrot_training - Step 45089: {'lr': 1.2803387408879153e-05, 'samples': 23086080, 'steps': 45089, 'loss/train': 2.16066837310791} 02/26/2022 11:59:42 - INFO - codeparrot_training - Step 45090: {'lr': 1.2798218718431453e-05, 'samples': 23086592, 'steps': 45090, 'loss/train': 1.5529731512069702} 02/26/2022 11:59:45 - INFO - codeparrot_training - Step 45091: {'lr': 1.2793051044080822e-05, 'samples': 23087104, 'steps': 45091, 'loss/train': 1.941236972808838} 02/26/2022 11:59:51 - INFO - codeparrot_training - Step 45092: {'lr': 1.278788438584949e-05, 'samples': 23087616, 'steps': 45092, 'loss/train': 5.634986877441406} 02/26/2022 11:59:54 - INFO - codeparrot_training - Step 45093: {'lr': 1.2782718743759525e-05, 'samples': 23088128, 'steps': 45093, 'loss/train': 2.2821693420410156} 02/26/2022 11:59:58 - INFO - codeparrot_training - Step 45094: {'lr': 1.2777554117833073e-05, 'samples': 23088640, 'steps': 45094, 'loss/train': 1.4865084886550903} 02/26/2022 12:00:04 - INFO - codeparrot_training - Step 45095: {'lr': 1.2772390508092203e-05, 'samples': 23089152, 'steps': 45095, 'loss/train': 1.6523545980453491} 02/26/2022 12:00:07 - INFO - codeparrot_training - Step 45096: {'lr': 1.2767227914559116e-05, 'samples': 23089664, 'steps': 45096, 'loss/train': 1.8405457735061646} 02/26/2022 12:00:13 - INFO - codeparrot_training - Step 45097: {'lr': 1.2762066337255907e-05, 'samples': 23090176, 'steps': 45097, 'loss/train': 2.533374547958374} 02/26/2022 12:00:16 - INFO - codeparrot_training - Step 45098: {'lr': 1.2756905776204698e-05, 'samples': 23090688, 'steps': 45098, 'loss/train': 0.31162989139556885} 02/26/2022 12:00:22 - INFO - codeparrot_training - Step 45099: {'lr': 1.2751746231427497e-05, 'samples': 23091200, 'steps': 45099, 'loss/train': 1.6668719053268433} 02/26/2022 12:00:27 - INFO - codeparrot_training - Step 45100: {'lr': 1.2746587702946538e-05, 'samples': 23091712, 'steps': 45100, 'loss/train': 2.486804485321045} 02/26/2022 12:00:31 - INFO - codeparrot_training - Step 45101: {'lr': 1.2741430190783804e-05, 'samples': 23092224, 'steps': 45101, 'loss/train': 2.341907262802124} 02/26/2022 12:00:34 - INFO - codeparrot_training - Step 45102: {'lr': 1.2736273694961553e-05, 'samples': 23092736, 'steps': 45102, 'loss/train': 2.0972723960876465} 02/26/2022 12:00:41 - INFO - codeparrot_training - Step 45103: {'lr': 1.2731118215501658e-05, 'samples': 23093248, 'steps': 45103, 'loss/train': 0.5704176425933838} 02/26/2022 12:00:44 - INFO - codeparrot_training - Step 45104: {'lr': 1.2725963752426379e-05, 'samples': 23093760, 'steps': 45104, 'loss/train': 0.40021708607673645} 02/26/2022 12:00:49 - INFO - codeparrot_training - Step 45105: {'lr': 1.2720810305757668e-05, 'samples': 23094272, 'steps': 45105, 'loss/train': 1.5304100513458252} 02/26/2022 12:00:53 - INFO - codeparrot_training - Step 45106: {'lr': 1.2715657875517733e-05, 'samples': 23094784, 'steps': 45106, 'loss/train': 2.3171169757843018} 02/26/2022 12:00:59 - INFO - codeparrot_training - Step 45107: {'lr': 1.27105064617285e-05, 'samples': 23095296, 'steps': 45107, 'loss/train': 2.572324752807617} 02/26/2022 12:01:04 - INFO - codeparrot_training - Step 45108: {'lr': 1.2705356064412172e-05, 'samples': 23095808, 'steps': 45108, 'loss/train': 1.3853480815887451} 02/26/2022 12:01:07 - INFO - codeparrot_training - Step 45109: {'lr': 1.2700206683590676e-05, 'samples': 23096320, 'steps': 45109, 'loss/train': 1.732122540473938} 02/26/2022 12:01:13 - INFO - codeparrot_training - Step 45110: {'lr': 1.2695058319286245e-05, 'samples': 23096832, 'steps': 45110, 'loss/train': 2.681973695755005} 02/26/2022 12:01:16 - INFO - codeparrot_training - Step 45111: {'lr': 1.2689910971520752e-05, 'samples': 23097344, 'steps': 45111, 'loss/train': 1.8421803712844849} 02/26/2022 12:01:22 - INFO - codeparrot_training - Step 45112: {'lr': 1.2684764640316343e-05, 'samples': 23097856, 'steps': 45112, 'loss/train': 0.5336638689041138} 02/26/2022 12:01:25 - INFO - codeparrot_training - Step 45113: {'lr': 1.267961932569503e-05, 'samples': 23098368, 'steps': 45113, 'loss/train': 1.389957070350647} 02/26/2022 12:01:31 - INFO - codeparrot_training - Step 45114: {'lr': 1.2674475027678906e-05, 'samples': 23098880, 'steps': 45114, 'loss/train': 1.7390813827514648} 02/26/2022 12:01:34 - INFO - codeparrot_training - Step 45115: {'lr': 1.266933174628998e-05, 'samples': 23099392, 'steps': 45115, 'loss/train': 1.7608330249786377} 02/26/2022 12:01:41 - INFO - codeparrot_training - Step 45116: {'lr': 1.2664189481550236e-05, 'samples': 23099904, 'steps': 45116, 'loss/train': 1.451500415802002} 02/26/2022 12:01:44 - INFO - codeparrot_training - Step 45117: {'lr': 1.265904823348174e-05, 'samples': 23100416, 'steps': 45117, 'loss/train': 2.1356723308563232} 02/26/2022 12:01:50 - INFO - codeparrot_training - Step 45118: {'lr': 1.2653908002106557e-05, 'samples': 23100928, 'steps': 45118, 'loss/train': 1.1042168140411377} 02/26/2022 12:01:53 - INFO - codeparrot_training - Step 45119: {'lr': 1.264876878744664e-05, 'samples': 23101440, 'steps': 45119, 'loss/train': 2.336280107498169} 02/26/2022 12:01:59 - INFO - codeparrot_training - Step 45120: {'lr': 1.2643630589524058e-05, 'samples': 23101952, 'steps': 45120, 'loss/train': 1.5264019966125488} 02/26/2022 12:02:02 - INFO - codeparrot_training - Step 45121: {'lr': 1.2638493408360735e-05, 'samples': 23102464, 'steps': 45121, 'loss/train': 1.165435791015625} 02/26/2022 12:02:08 - INFO - codeparrot_training - Step 45122: {'lr': 1.2633357243978793e-05, 'samples': 23102976, 'steps': 45122, 'loss/train': 2.081279754638672} 02/26/2022 12:02:11 - INFO - codeparrot_training - Step 45123: {'lr': 1.262822209640016e-05, 'samples': 23103488, 'steps': 45123, 'loss/train': 0.5477192997932434} 02/26/2022 12:02:17 - INFO - codeparrot_training - Step 45124: {'lr': 1.2623087965646846e-05, 'samples': 23104000, 'steps': 45124, 'loss/train': 1.7766499519348145} 02/26/2022 12:02:20 - INFO - codeparrot_training - Step 45125: {'lr': 1.2617954851740832e-05, 'samples': 23104512, 'steps': 45125, 'loss/train': 1.2019747495651245} 02/26/2022 12:02:26 - INFO - codeparrot_training - Step 45126: {'lr': 1.2612822754704129e-05, 'samples': 23105024, 'steps': 45126, 'loss/train': 1.513781189918518} 02/26/2022 12:02:29 - INFO - codeparrot_training - Step 45127: {'lr': 1.2607691674558747e-05, 'samples': 23105536, 'steps': 45127, 'loss/train': 0.7065152525901794} 02/26/2022 12:02:36 - INFO - codeparrot_training - Step 45128: {'lr': 1.2602561611326613e-05, 'samples': 23106048, 'steps': 45128, 'loss/train': 1.7598798274993896} 02/26/2022 12:02:39 - INFO - codeparrot_training - Step 45129: {'lr': 1.2597432565029682e-05, 'samples': 23106560, 'steps': 45129, 'loss/train': 0.29196837544441223} 02/26/2022 12:02:45 - INFO - codeparrot_training - Step 45130: {'lr': 1.2592304535690018e-05, 'samples': 23107072, 'steps': 45130, 'loss/train': 1.2039690017700195} 02/26/2022 12:02:48 - INFO - codeparrot_training - Step 45131: {'lr': 1.2587177523329524e-05, 'samples': 23107584, 'steps': 45131, 'loss/train': 1.0128332376480103} 02/26/2022 12:02:54 - INFO - codeparrot_training - Step 45132: {'lr': 1.2582051527970178e-05, 'samples': 23108096, 'steps': 45132, 'loss/train': 2.611274242401123} 02/26/2022 12:02:57 - INFO - codeparrot_training - Step 45133: {'lr': 1.257692654963391e-05, 'samples': 23108608, 'steps': 45133, 'loss/train': 2.5898587703704834} 02/26/2022 12:03:03 - INFO - codeparrot_training - Step 45134: {'lr': 1.2571802588342728e-05, 'samples': 23109120, 'steps': 45134, 'loss/train': 2.2823262214660645} 02/26/2022 12:03:06 - INFO - codeparrot_training - Step 45135: {'lr': 1.2566679644118534e-05, 'samples': 23109632, 'steps': 45135, 'loss/train': 1.3068331480026245} 02/26/2022 12:03:12 - INFO - codeparrot_training - Step 45136: {'lr': 1.2561557716983307e-05, 'samples': 23110144, 'steps': 45136, 'loss/train': 0.837346613407135} 02/26/2022 12:03:15 - INFO - codeparrot_training - Step 45137: {'lr': 1.2556436806958976e-05, 'samples': 23110656, 'steps': 45137, 'loss/train': 3.0111570358276367} 02/26/2022 12:03:19 - INFO - codeparrot_training - Step 45138: {'lr': 1.255131691406744e-05, 'samples': 23111168, 'steps': 45138, 'loss/train': 0.5239671468734741} 02/26/2022 12:03:25 - INFO - codeparrot_training - Step 45139: {'lr': 1.254619803833068e-05, 'samples': 23111680, 'steps': 45139, 'loss/train': 2.8919575214385986} 02/26/2022 12:03:28 - INFO - codeparrot_training - Step 45140: {'lr': 1.2541080179770569e-05, 'samples': 23112192, 'steps': 45140, 'loss/train': 1.582980751991272} 02/26/2022 12:03:34 - INFO - codeparrot_training - Step 45141: {'lr': 1.2535963338409173e-05, 'samples': 23112704, 'steps': 45141, 'loss/train': 0.9383519291877747} 02/26/2022 12:03:37 - INFO - codeparrot_training - Step 45142: {'lr': 1.2530847514268196e-05, 'samples': 23113216, 'steps': 45142, 'loss/train': 0.8160671591758728} 02/26/2022 12:03:43 - INFO - codeparrot_training - Step 45143: {'lr': 1.252573270736973e-05, 'samples': 23113728, 'steps': 45143, 'loss/train': 1.442929983139038} 02/26/2022 12:03:49 - INFO - codeparrot_training - Step 45144: {'lr': 1.2520618917735566e-05, 'samples': 23114240, 'steps': 45144, 'loss/train': 1.965590000152588} 02/26/2022 12:03:52 - INFO - codeparrot_training - Step 45145: {'lr': 1.2515506145387739e-05, 'samples': 23114752, 'steps': 45145, 'loss/train': 0.45528197288513184} 02/26/2022 12:03:58 - INFO - codeparrot_training - Step 45146: {'lr': 1.2510394390348012e-05, 'samples': 23115264, 'steps': 45146, 'loss/train': 2.4050965309143066} 02/26/2022 12:04:01 - INFO - codeparrot_training - Step 45147: {'lr': 1.2505283652638366e-05, 'samples': 23115776, 'steps': 45147, 'loss/train': 1.8518065214157104} 02/26/2022 12:04:07 - INFO - codeparrot_training - Step 45148: {'lr': 1.2500173932280646e-05, 'samples': 23116288, 'steps': 45148, 'loss/train': 2.4183380603790283} 02/26/2022 12:04:11 - INFO - codeparrot_training - Step 45149: {'lr': 1.249506522929683e-05, 'samples': 23116800, 'steps': 45149, 'loss/train': 2.1279654502868652} 02/26/2022 12:04:16 - INFO - codeparrot_training - Step 45150: {'lr': 1.2489957543708658e-05, 'samples': 23117312, 'steps': 45150, 'loss/train': 1.2020108699798584} 02/26/2022 12:04:20 - INFO - codeparrot_training - Step 45151: {'lr': 1.2484850875538135e-05, 'samples': 23117824, 'steps': 45151, 'loss/train': 1.4040043354034424} 02/26/2022 12:04:25 - INFO - codeparrot_training - Step 45152: {'lr': 1.2479745224807049e-05, 'samples': 23118336, 'steps': 45152, 'loss/train': 2.096092462539673} 02/26/2022 12:04:29 - INFO - codeparrot_training - Step 45153: {'lr': 1.2474640591537356e-05, 'samples': 23118848, 'steps': 45153, 'loss/train': 2.048578977584839} 02/26/2022 12:04:34 - INFO - codeparrot_training - Step 45154: {'lr': 1.24695369757509e-05, 'samples': 23119360, 'steps': 45154, 'loss/train': 1.5250835418701172} 02/26/2022 12:04:38 - INFO - codeparrot_training - Step 45155: {'lr': 1.2464434377469496e-05, 'samples': 23119872, 'steps': 45155, 'loss/train': 0.7809669375419617} 02/26/2022 12:04:43 - INFO - codeparrot_training - Step 45156: {'lr': 1.245933279671499e-05, 'samples': 23120384, 'steps': 45156, 'loss/train': 2.1010615825653076} 02/26/2022 12:04:47 - INFO - codeparrot_training - Step 45157: {'lr': 1.2454232233509332e-05, 'samples': 23120896, 'steps': 45157, 'loss/train': 1.7512787580490112} 02/26/2022 12:04:52 - INFO - codeparrot_training - Step 45158: {'lr': 1.2449132687874315e-05, 'samples': 23121408, 'steps': 45158, 'loss/train': 1.8734853267669678} 02/26/2022 12:04:56 - INFO - codeparrot_training - Step 45159: {'lr': 1.2444034159831752e-05, 'samples': 23121920, 'steps': 45159, 'loss/train': 1.7184488773345947} 02/26/2022 12:05:01 - INFO - codeparrot_training - Step 45160: {'lr': 1.2438936649403515e-05, 'samples': 23122432, 'steps': 45160, 'loss/train': 1.4589784145355225} 02/26/2022 12:05:05 - INFO - codeparrot_training - Step 45161: {'lr': 1.2433840156611448e-05, 'samples': 23122944, 'steps': 45161, 'loss/train': 2.0252017974853516} 02/26/2022 12:05:11 - INFO - codeparrot_training - Step 45162: {'lr': 1.242874468147734e-05, 'samples': 23123456, 'steps': 45162, 'loss/train': 1.729162573814392} 02/26/2022 12:05:14 - INFO - codeparrot_training - Step 45163: {'lr': 1.2423650224023143e-05, 'samples': 23123968, 'steps': 45163, 'loss/train': 2.562262535095215} 02/26/2022 12:05:20 - INFO - codeparrot_training - Step 45164: {'lr': 1.2418556784270508e-05, 'samples': 23124480, 'steps': 45164, 'loss/train': 1.890058994293213} 02/26/2022 12:05:23 - INFO - codeparrot_training - Step 45165: {'lr': 1.2413464362241362e-05, 'samples': 23124992, 'steps': 45165, 'loss/train': 0.8579510450363159} 02/26/2022 12:05:29 - INFO - codeparrot_training - Step 45166: {'lr': 1.2408372957957465e-05, 'samples': 23125504, 'steps': 45166, 'loss/train': 0.8872958421707153} 02/26/2022 12:05:32 - INFO - codeparrot_training - Step 45167: {'lr': 1.2403282571440716e-05, 'samples': 23126016, 'steps': 45167, 'loss/train': 1.5252865552902222} 02/26/2022 12:05:38 - INFO - codeparrot_training - Step 45168: {'lr': 1.2398193202712821e-05, 'samples': 23126528, 'steps': 45168, 'loss/train': 1.1240007877349854} 02/26/2022 12:05:41 - INFO - codeparrot_training - Step 45169: {'lr': 1.2393104851795623e-05, 'samples': 23127040, 'steps': 45169, 'loss/train': 1.2660431861877441} 02/26/2022 12:05:47 - INFO - codeparrot_training - Step 45170: {'lr': 1.2388017518710909e-05, 'samples': 23127552, 'steps': 45170, 'loss/train': 2.0403618812561035} 02/26/2022 12:05:50 - INFO - codeparrot_training - Step 45171: {'lr': 1.2382931203480497e-05, 'samples': 23128064, 'steps': 45171, 'loss/train': 1.2678524255752563} 02/26/2022 12:05:57 - INFO - codeparrot_training - Step 45172: {'lr': 1.2377845906126118e-05, 'samples': 23128576, 'steps': 45172, 'loss/train': 1.1919927597045898} 02/26/2022 12:06:00 - INFO - codeparrot_training - Step 45173: {'lr': 1.2372761626669616e-05, 'samples': 23129088, 'steps': 45173, 'loss/train': 1.4092915058135986} 02/26/2022 12:06:05 - INFO - codeparrot_training - Step 45174: {'lr': 1.2367678365132778e-05, 'samples': 23129600, 'steps': 45174, 'loss/train': 1.9317097663879395} 02/26/2022 12:06:09 - INFO - codeparrot_training - Step 45175: {'lr': 1.2362596121537285e-05, 'samples': 23130112, 'steps': 45175, 'loss/train': 2.4590182304382324} 02/26/2022 12:06:14 - INFO - codeparrot_training - Step 45176: {'lr': 1.2357514895905003e-05, 'samples': 23130624, 'steps': 45176, 'loss/train': 2.421562433242798} 02/26/2022 12:06:18 - INFO - codeparrot_training - Step 45177: {'lr': 1.2352434688257697e-05, 'samples': 23131136, 'steps': 45177, 'loss/train': 1.9065955877304077} 02/26/2022 12:06:23 - INFO - codeparrot_training - Step 45178: {'lr': 1.234735549861707e-05, 'samples': 23131648, 'steps': 45178, 'loss/train': 2.655139446258545} 02/26/2022 12:06:27 - INFO - codeparrot_training - Step 45179: {'lr': 1.234227732700488e-05, 'samples': 23132160, 'steps': 45179, 'loss/train': 0.8640618920326233} 02/26/2022 12:06:32 - INFO - codeparrot_training - Step 45180: {'lr': 1.2337200173442948e-05, 'samples': 23132672, 'steps': 45180, 'loss/train': 2.1679911613464355} 02/26/2022 12:06:36 - INFO - codeparrot_training - Step 45181: {'lr': 1.2332124037953003e-05, 'samples': 23133184, 'steps': 45181, 'loss/train': 1.7762260437011719} 02/26/2022 12:06:42 - INFO - codeparrot_training - Step 45182: {'lr': 1.2327048920556749e-05, 'samples': 23133696, 'steps': 45182, 'loss/train': 3.0028209686279297} 02/26/2022 12:06:45 - INFO - codeparrot_training - Step 45183: {'lr': 1.2321974821275923e-05, 'samples': 23134208, 'steps': 45183, 'loss/train': 0.267306387424469} 02/26/2022 12:06:51 - INFO - codeparrot_training - Step 45184: {'lr': 1.2316901740132336e-05, 'samples': 23134720, 'steps': 45184, 'loss/train': 2.4571051597595215} 02/26/2022 12:06:55 - INFO - codeparrot_training - Step 45185: {'lr': 1.2311829677147612e-05, 'samples': 23135232, 'steps': 45185, 'loss/train': 1.403219223022461} 02/26/2022 12:07:00 - INFO - codeparrot_training - Step 45186: {'lr': 1.2306758632343567e-05, 'samples': 23135744, 'steps': 45186, 'loss/train': 1.4392234086990356} 02/26/2022 12:07:04 - INFO - codeparrot_training - Step 45187: {'lr': 1.2301688605741878e-05, 'samples': 23136256, 'steps': 45187, 'loss/train': 1.6896017789840698} 02/26/2022 12:07:09 - INFO - codeparrot_training - Step 45188: {'lr': 1.2296619597364278e-05, 'samples': 23136768, 'steps': 45188, 'loss/train': 1.8735121488571167} 02/26/2022 12:07:13 - INFO - codeparrot_training - Step 45189: {'lr': 1.22915516072325e-05, 'samples': 23137280, 'steps': 45189, 'loss/train': 1.7878131866455078} 02/26/2022 12:07:18 - INFO - codeparrot_training - Step 45190: {'lr': 1.2286484635368244e-05, 'samples': 23137792, 'steps': 45190, 'loss/train': 0.8920719623565674} 02/26/2022 12:07:22 - INFO - codeparrot_training - Step 45191: {'lr': 1.2281418681793167e-05, 'samples': 23138304, 'steps': 45191, 'loss/train': 1.6619768142700195} 02/26/2022 12:07:27 - INFO - codeparrot_training - Step 45192: {'lr': 1.2276353746529023e-05, 'samples': 23138816, 'steps': 45192, 'loss/train': 1.7054188251495361} 02/26/2022 12:07:31 - INFO - codeparrot_training - Step 45193: {'lr': 1.2271289829597493e-05, 'samples': 23139328, 'steps': 45193, 'loss/train': 1.158315658569336} 02/26/2022 12:07:37 - INFO - codeparrot_training - Step 45194: {'lr': 1.226622693102028e-05, 'samples': 23139840, 'steps': 45194, 'loss/train': 1.4815367460250854} 02/26/2022 12:07:40 - INFO - codeparrot_training - Step 45195: {'lr': 1.2261165050819034e-05, 'samples': 23140352, 'steps': 45195, 'loss/train': 1.7711548805236816} 02/26/2022 12:07:46 - INFO - codeparrot_training - Step 45196: {'lr': 1.2256104189015516e-05, 'samples': 23140864, 'steps': 45196, 'loss/train': 1.2785003185272217} 02/26/2022 12:07:49 - INFO - codeparrot_training - Step 45197: {'lr': 1.2251044345631319e-05, 'samples': 23141376, 'steps': 45197, 'loss/train': 0.9468376040458679} 02/26/2022 12:07:55 - INFO - codeparrot_training - Step 45198: {'lr': 1.2245985520688175e-05, 'samples': 23141888, 'steps': 45198, 'loss/train': 1.7057024240493774} 02/26/2022 12:07:58 - INFO - codeparrot_training - Step 45199: {'lr': 1.2240927714207706e-05, 'samples': 23142400, 'steps': 45199, 'loss/train': 1.9232268333435059} 02/26/2022 12:08:04 - INFO - codeparrot_training - Step 45200: {'lr': 1.2235870926211617e-05, 'samples': 23142912, 'steps': 45200, 'loss/train': 1.2355166673660278} 02/26/2022 12:08:08 - INFO - codeparrot_training - Step 45201: {'lr': 1.2230815156721532e-05, 'samples': 23143424, 'steps': 45201, 'loss/train': 1.7763829231262207} 02/26/2022 12:08:14 - INFO - codeparrot_training - Step 45202: {'lr': 1.2225760405759207e-05, 'samples': 23143936, 'steps': 45202, 'loss/train': 2.1136059761047363} 02/26/2022 12:08:17 - INFO - codeparrot_training - Step 45203: {'lr': 1.2220706673346154e-05, 'samples': 23144448, 'steps': 45203, 'loss/train': 2.2113077640533447} 02/26/2022 12:08:21 - INFO - codeparrot_training - Step 45204: {'lr': 1.2215653959504136e-05, 'samples': 23144960, 'steps': 45204, 'loss/train': 1.0620841979980469} 02/26/2022 12:08:26 - INFO - codeparrot_training - Step 45205: {'lr': 1.221060226425469e-05, 'samples': 23145472, 'steps': 45205, 'loss/train': 2.7467172145843506} 02/26/2022 12:08:30 - INFO - codeparrot_training - Step 45206: {'lr': 1.2205551587619602e-05, 'samples': 23145984, 'steps': 45206, 'loss/train': 1.1453804969787598} 02/26/2022 12:08:35 - INFO - codeparrot_training - Step 45207: {'lr': 1.2200501929620356e-05, 'samples': 23146496, 'steps': 45207, 'loss/train': 1.5322825908660889} 02/26/2022 12:08:41 - INFO - codeparrot_training - Step 45208: {'lr': 1.2195453290278686e-05, 'samples': 23147008, 'steps': 45208, 'loss/train': 1.585524559020996} 02/26/2022 12:08:44 - INFO - codeparrot_training - Step 45209: {'lr': 1.2190405669616157e-05, 'samples': 23147520, 'steps': 45209, 'loss/train': 1.4892983436584473} 02/26/2022 12:08:50 - INFO - codeparrot_training - Step 45210: {'lr': 1.2185359067654417e-05, 'samples': 23148032, 'steps': 45210, 'loss/train': 1.5818986892700195} 02/26/2022 12:08:54 - INFO - codeparrot_training - Step 45211: {'lr': 1.2180313484415117e-05, 'samples': 23148544, 'steps': 45211, 'loss/train': 1.7678420543670654} 02/26/2022 12:08:59 - INFO - codeparrot_training - Step 45212: {'lr': 1.2175268919919823e-05, 'samples': 23149056, 'steps': 45212, 'loss/train': 3.5372023582458496} 02/26/2022 12:09:03 - INFO - codeparrot_training - Step 45213: {'lr': 1.217022537419013e-05, 'samples': 23149568, 'steps': 45213, 'loss/train': 1.703425645828247} 02/26/2022 12:09:08 - INFO - codeparrot_training - Step 45214: {'lr': 1.2165182847247713e-05, 'samples': 23150080, 'steps': 45214, 'loss/train': 0.3876635432243347} 02/26/2022 12:09:12 - INFO - codeparrot_training - Step 45215: {'lr': 1.216014133911411e-05, 'samples': 23150592, 'steps': 45215, 'loss/train': 1.9977920055389404} 02/26/2022 12:09:17 - INFO - codeparrot_training - Step 45216: {'lr': 1.2155100849810946e-05, 'samples': 23151104, 'steps': 45216, 'loss/train': 1.6331762075424194} 02/26/2022 12:09:21 - INFO - codeparrot_training - Step 45217: {'lr': 1.2150061379359812e-05, 'samples': 23151616, 'steps': 45217, 'loss/train': 1.7448798418045044} 02/26/2022 12:09:26 - INFO - codeparrot_training - Step 45218: {'lr': 1.2145022927782273e-05, 'samples': 23152128, 'steps': 45218, 'loss/train': 0.9834668636322021} 02/26/2022 12:09:30 - INFO - codeparrot_training - Step 45219: {'lr': 1.2139985495099926e-05, 'samples': 23152640, 'steps': 45219, 'loss/train': 2.9776532649993896} 02/26/2022 12:09:36 - INFO - codeparrot_training - Step 45220: {'lr': 1.213494908133439e-05, 'samples': 23153152, 'steps': 45220, 'loss/train': 1.8694608211517334} 02/26/2022 12:09:39 - INFO - codeparrot_training - Step 45221: {'lr': 1.2129913686507177e-05, 'samples': 23153664, 'steps': 45221, 'loss/train': 2.2419707775115967} 02/26/2022 12:09:45 - INFO - codeparrot_training - Step 45222: {'lr': 1.2124879310639852e-05, 'samples': 23154176, 'steps': 45222, 'loss/train': 1.7461345195770264} 02/26/2022 12:09:48 - INFO - codeparrot_training - Step 45223: {'lr': 1.2119845953754039e-05, 'samples': 23154688, 'steps': 45223, 'loss/train': 2.1057026386260986} 02/26/2022 12:09:54 - INFO - codeparrot_training - Step 45224: {'lr': 1.2114813615871273e-05, 'samples': 23155200, 'steps': 45224, 'loss/train': 1.709058165550232} 02/26/2022 12:09:57 - INFO - codeparrot_training - Step 45225: {'lr': 1.2109782297013122e-05, 'samples': 23155712, 'steps': 45225, 'loss/train': 2.0414555072784424} 02/26/2022 12:10:03 - INFO - codeparrot_training - Step 45226: {'lr': 1.210475199720107e-05, 'samples': 23156224, 'steps': 45226, 'loss/train': 1.4972212314605713} 02/26/2022 12:10:06 - INFO - codeparrot_training - Step 45227: {'lr': 1.2099722716456763e-05, 'samples': 23156736, 'steps': 45227, 'loss/train': 2.3076183795928955} 02/26/2022 12:10:12 - INFO - codeparrot_training - Step 45228: {'lr': 1.2094694454801713e-05, 'samples': 23157248, 'steps': 45228, 'loss/train': 2.0042624473571777} 02/26/2022 12:10:15 - INFO - codeparrot_training - Step 45229: {'lr': 1.2089667212257432e-05, 'samples': 23157760, 'steps': 45229, 'loss/train': 1.3523269891738892} 02/26/2022 12:10:21 - INFO - codeparrot_training - Step 45230: {'lr': 1.208464098884543e-05, 'samples': 23158272, 'steps': 45230, 'loss/train': 1.8856594562530518} 02/26/2022 12:10:25 - INFO - codeparrot_training - Step 45231: {'lr': 1.2079615784587328e-05, 'samples': 23158784, 'steps': 45231, 'loss/train': 0.09384184330701828} 02/26/2022 12:10:31 - INFO - codeparrot_training - Step 45232: {'lr': 1.2074591599504609e-05, 'samples': 23159296, 'steps': 45232, 'loss/train': 1.964783787727356} 02/26/2022 12:10:34 - INFO - codeparrot_training - Step 45233: {'lr': 1.2069568433618783e-05, 'samples': 23159808, 'steps': 45233, 'loss/train': 2.2132627964019775} 02/26/2022 12:10:40 - INFO - codeparrot_training - Step 45234: {'lr': 1.2064546286951361e-05, 'samples': 23160320, 'steps': 45234, 'loss/train': 2.107429265975952} 02/26/2022 12:10:43 - INFO - codeparrot_training - Step 45235: {'lr': 1.2059525159523882e-05, 'samples': 23160832, 'steps': 45235, 'loss/train': 2.46586275100708} 02/26/2022 12:10:49 - INFO - codeparrot_training - Step 45236: {'lr': 1.20545050513578e-05, 'samples': 23161344, 'steps': 45236, 'loss/train': 2.2400896549224854} 02/26/2022 12:10:52 - INFO - codeparrot_training - Step 45237: {'lr': 1.2049485962474738e-05, 'samples': 23161856, 'steps': 45237, 'loss/train': 2.2663466930389404} 02/26/2022 12:10:57 - INFO - codeparrot_training - Step 45238: {'lr': 1.2044467892896066e-05, 'samples': 23162368, 'steps': 45238, 'loss/train': 1.9039990901947021} 02/26/2022 12:11:01 - INFO - codeparrot_training - Step 45239: {'lr': 1.2039450842643352e-05, 'samples': 23162880, 'steps': 45239, 'loss/train': 0.6447505354881287} 02/26/2022 12:11:07 - INFO - codeparrot_training - Step 45240: {'lr': 1.2034434811738049e-05, 'samples': 23163392, 'steps': 45240, 'loss/train': 1.5896403789520264} 02/26/2022 12:11:11 - INFO - codeparrot_training - Step 45241: {'lr': 1.2029419800201725e-05, 'samples': 23163904, 'steps': 45241, 'loss/train': 0.8047797679901123} 02/26/2022 12:11:16 - INFO - codeparrot_training - Step 45242: {'lr': 1.2024405808055722e-05, 'samples': 23164416, 'steps': 45242, 'loss/train': 2.7478795051574707} 02/26/2022 12:11:19 - INFO - codeparrot_training - Step 45243: {'lr': 1.2019392835321663e-05, 'samples': 23164928, 'steps': 45243, 'loss/train': 1.7348052263259888} 02/26/2022 12:11:25 - INFO - codeparrot_training - Step 45244: {'lr': 1.2014380882020892e-05, 'samples': 23165440, 'steps': 45244, 'loss/train': 1.4122836589813232} 02/26/2022 12:11:28 - INFO - codeparrot_training - Step 45245: {'lr': 1.2009369948175031e-05, 'samples': 23165952, 'steps': 45245, 'loss/train': 2.3270492553710938} 02/26/2022 12:11:34 - INFO - codeparrot_training - Step 45246: {'lr': 1.2004360033805395e-05, 'samples': 23166464, 'steps': 45246, 'loss/train': 2.6379005908966064} 02/26/2022 12:11:37 - INFO - codeparrot_training - Step 45247: {'lr': 1.1999351138933524e-05, 'samples': 23166976, 'steps': 45247, 'loss/train': 2.834739923477173} 02/26/2022 12:11:43 - INFO - codeparrot_training - Step 45248: {'lr': 1.1994343263580843e-05, 'samples': 23167488, 'steps': 45248, 'loss/train': 1.8965092897415161} 02/26/2022 12:11:46 - INFO - codeparrot_training - Step 45249: {'lr': 1.1989336407768836e-05, 'samples': 23168000, 'steps': 45249, 'loss/train': 1.741094708442688} 02/26/2022 12:11:52 - INFO - codeparrot_training - Step 45250: {'lr': 1.198433057151893e-05, 'samples': 23168512, 'steps': 45250, 'loss/train': 1.092063069343567} 02/26/2022 12:11:55 - INFO - codeparrot_training - Step 45251: {'lr': 1.197932575485261e-05, 'samples': 23169024, 'steps': 45251, 'loss/train': 2.190891742706299} 02/26/2022 12:12:01 - INFO - codeparrot_training - Step 45252: {'lr': 1.1974321957791218e-05, 'samples': 23169536, 'steps': 45252, 'loss/train': 2.755958080291748} 02/26/2022 12:12:04 - INFO - codeparrot_training - Step 45253: {'lr': 1.1969319180356292e-05, 'samples': 23170048, 'steps': 45253, 'loss/train': 1.9685778617858887} 02/26/2022 12:12:10 - INFO - codeparrot_training - Step 45254: {'lr': 1.1964317422569204e-05, 'samples': 23170560, 'steps': 45254, 'loss/train': 1.8924490213394165} 02/26/2022 12:12:14 - INFO - codeparrot_training - Step 45255: {'lr': 1.195931668445141e-05, 'samples': 23171072, 'steps': 45255, 'loss/train': 2.355959415435791} 02/26/2022 12:12:20 - INFO - codeparrot_training - Step 45256: {'lr': 1.1954316966024282e-05, 'samples': 23171584, 'steps': 45256, 'loss/train': 1.4803780317306519} 02/26/2022 12:12:24 - INFO - codeparrot_training - Step 45257: {'lr': 1.19493182673093e-05, 'samples': 23172096, 'steps': 45257, 'loss/train': 1.2660396099090576} 02/26/2022 12:12:30 - INFO - codeparrot_training - Step 45258: {'lr': 1.194432058832784e-05, 'samples': 23172608, 'steps': 45258, 'loss/train': 1.8761061429977417} 02/26/2022 12:12:33 - INFO - codeparrot_training - Step 45259: {'lr': 1.1939323929101326e-05, 'samples': 23173120, 'steps': 45259, 'loss/train': 1.8708829879760742} 02/26/2022 12:12:39 - INFO - codeparrot_training - Step 45260: {'lr': 1.1934328289651131e-05, 'samples': 23173632, 'steps': 45260, 'loss/train': 2.60703444480896} 02/26/2022 12:12:42 - INFO - codeparrot_training - Step 45261: {'lr': 1.192933366999871e-05, 'samples': 23174144, 'steps': 45261, 'loss/train': 0.9393560290336609} 02/26/2022 12:12:48 - INFO - codeparrot_training - Step 45262: {'lr': 1.1924340070165434e-05, 'samples': 23174656, 'steps': 45262, 'loss/train': 0.1498311161994934} 02/26/2022 12:12:51 - INFO - codeparrot_training - Step 45263: {'lr': 1.1919347490172677e-05, 'samples': 23175168, 'steps': 45263, 'loss/train': 1.5746078491210938} 02/26/2022 12:12:57 - INFO - codeparrot_training - Step 45264: {'lr': 1.1914355930041836e-05, 'samples': 23175680, 'steps': 45264, 'loss/train': 1.025717854499817} 02/26/2022 12:13:00 - INFO - codeparrot_training - Step 45265: {'lr': 1.1909365389794258e-05, 'samples': 23176192, 'steps': 45265, 'loss/train': 2.2685678005218506} 02/26/2022 12:13:07 - INFO - codeparrot_training - Step 45266: {'lr': 1.1904375869451394e-05, 'samples': 23176704, 'steps': 45266, 'loss/train': 1.3533354997634888} 02/26/2022 12:13:10 - INFO - codeparrot_training - Step 45267: {'lr': 1.1899387369034592e-05, 'samples': 23177216, 'steps': 45267, 'loss/train': 1.3332960605621338} 02/26/2022 12:13:16 - INFO - codeparrot_training - Step 45268: {'lr': 1.1894399888565193e-05, 'samples': 23177728, 'steps': 45268, 'loss/train': 2.0228734016418457} 02/26/2022 12:13:19 - INFO - codeparrot_training - Step 45269: {'lr': 1.1889413428064543e-05, 'samples': 23178240, 'steps': 45269, 'loss/train': 1.8938934803009033} 02/26/2022 12:13:25 - INFO - codeparrot_training - Step 45270: {'lr': 1.1884427987554097e-05, 'samples': 23178752, 'steps': 45270, 'loss/train': 1.0400285720825195} 02/26/2022 12:13:30 - INFO - codeparrot_training - Step 45271: {'lr': 1.1879443567055088e-05, 'samples': 23179264, 'steps': 45271, 'loss/train': 1.9002655744552612} 02/26/2022 12:13:34 - INFO - codeparrot_training - Step 45272: {'lr': 1.1874460166589024e-05, 'samples': 23179776, 'steps': 45272, 'loss/train': 0.550462007522583} 02/26/2022 12:13:39 - INFO - codeparrot_training - Step 45273: {'lr': 1.1869477786177058e-05, 'samples': 23180288, 'steps': 45273, 'loss/train': 0.43645116686820984} 02/26/2022 12:13:43 - INFO - codeparrot_training - Step 45274: {'lr': 1.1864496425840699e-05, 'samples': 23180800, 'steps': 45274, 'loss/train': 1.7730686664581299} 02/26/2022 12:13:49 - INFO - codeparrot_training - Step 45275: {'lr': 1.185951608560118e-05, 'samples': 23181312, 'steps': 45275, 'loss/train': 2.263324737548828} 02/26/2022 12:13:52 - INFO - codeparrot_training - Step 45276: {'lr': 1.1854536765479929e-05, 'samples': 23181824, 'steps': 45276, 'loss/train': 1.4306524991989136} 02/26/2022 12:13:58 - INFO - codeparrot_training - Step 45277: {'lr': 1.184955846549815e-05, 'samples': 23182336, 'steps': 45277, 'loss/train': 1.6198973655700684} 02/26/2022 12:14:02 - INFO - codeparrot_training - Step 45278: {'lr': 1.18445811856773e-05, 'samples': 23182848, 'steps': 45278, 'loss/train': 2.242349624633789} 02/26/2022 12:14:05 - INFO - codeparrot_training - Step 45279: {'lr': 1.1839604926038584e-05, 'samples': 23183360, 'steps': 45279, 'loss/train': 2.072956085205078} 02/26/2022 12:14:11 - INFO - codeparrot_training - Step 45280: {'lr': 1.1834629686603455e-05, 'samples': 23183872, 'steps': 45280, 'loss/train': 1.8613369464874268} 02/26/2022 12:14:15 - INFO - codeparrot_training - Step 45281: {'lr': 1.1829655467393092e-05, 'samples': 23184384, 'steps': 45281, 'loss/train': 1.722774624824524} 02/26/2022 12:14:20 - INFO - codeparrot_training - Step 45282: {'lr': 1.1824682268428866e-05, 'samples': 23184896, 'steps': 45282, 'loss/train': 0.9641186594963074} 02/26/2022 12:14:24 - INFO - codeparrot_training - Step 45283: {'lr': 1.1819710089732067e-05, 'samples': 23185408, 'steps': 45283, 'loss/train': 1.4226659536361694} 02/26/2022 12:14:29 - INFO - codeparrot_training - Step 45284: {'lr': 1.181473893132401e-05, 'samples': 23185920, 'steps': 45284, 'loss/train': 1.9249154329299927} 02/26/2022 12:14:33 - INFO - codeparrot_training - Step 45285: {'lr': 1.1809768793225983e-05, 'samples': 23186432, 'steps': 45285, 'loss/train': 2.214202404022217} 02/26/2022 12:14:40 - INFO - codeparrot_training - Step 45286: {'lr': 1.1804799675459276e-05, 'samples': 23186944, 'steps': 45286, 'loss/train': 0.5878369212150574} 02/26/2022 12:14:43 - INFO - codeparrot_training - Step 45287: {'lr': 1.1799831578045151e-05, 'samples': 23187456, 'steps': 45287, 'loss/train': 0.21804076433181763} 02/26/2022 12:14:49 - INFO - codeparrot_training - Step 45288: {'lr': 1.179486450100492e-05, 'samples': 23187968, 'steps': 45288, 'loss/train': 2.0877740383148193} 02/26/2022 12:14:52 - INFO - codeparrot_training - Step 45289: {'lr': 1.1789898444359876e-05, 'samples': 23188480, 'steps': 45289, 'loss/train': 1.3582688570022583} 02/26/2022 12:14:58 - INFO - codeparrot_training - Step 45290: {'lr': 1.1784933408131248e-05, 'samples': 23188992, 'steps': 45290, 'loss/train': 1.643548846244812} 02/26/2022 12:15:01 - INFO - codeparrot_training - Step 45291: {'lr': 1.1779969392340273e-05, 'samples': 23189504, 'steps': 45291, 'loss/train': 1.3875877857208252} 02/26/2022 12:15:07 - INFO - codeparrot_training - Step 45292: {'lr': 1.177500639700832e-05, 'samples': 23190016, 'steps': 45292, 'loss/train': 1.0693222284317017} 02/26/2022 12:15:10 - INFO - codeparrot_training - Step 45293: {'lr': 1.1770044422156623e-05, 'samples': 23190528, 'steps': 45293, 'loss/train': 1.7935291528701782} 02/26/2022 12:15:16 - INFO - codeparrot_training - Step 45294: {'lr': 1.176508346780636e-05, 'samples': 23191040, 'steps': 45294, 'loss/train': 1.1440677642822266} 02/26/2022 12:15:19 - INFO - codeparrot_training - Step 45295: {'lr': 1.1760123533978817e-05, 'samples': 23191552, 'steps': 45295, 'loss/train': 2.1591925621032715} 02/26/2022 12:15:25 - INFO - codeparrot_training - Step 45296: {'lr': 1.1755164620695314e-05, 'samples': 23192064, 'steps': 45296, 'loss/train': 1.8017840385437012} 02/26/2022 12:15:28 - INFO - codeparrot_training - Step 45297: {'lr': 1.1750206727976998e-05, 'samples': 23192576, 'steps': 45297, 'loss/train': 2.583359479904175} 02/26/2022 12:15:34 - INFO - codeparrot_training - Step 45298: {'lr': 1.1745249855845158e-05, 'samples': 23193088, 'steps': 45298, 'loss/train': 1.352054476737976} 02/26/2022 12:15:37 - INFO - codeparrot_training - Step 45299: {'lr': 1.1740294004320973e-05, 'samples': 23193600, 'steps': 45299, 'loss/train': 0.40386995673179626} 02/26/2022 12:15:43 - INFO - codeparrot_training - Step 45300: {'lr': 1.1735339173425757e-05, 'samples': 23194112, 'steps': 45300, 'loss/train': 1.9061481952667236} 02/26/2022 12:15:46 - INFO - codeparrot_training - Step 45301: {'lr': 1.173038536318069e-05, 'samples': 23194624, 'steps': 45301, 'loss/train': 0.03340005502104759} 02/26/2022 12:15:53 - INFO - codeparrot_training - Step 45302: {'lr': 1.1725432573606975e-05, 'samples': 23195136, 'steps': 45302, 'loss/train': 2.3133888244628906} 02/26/2022 12:15:56 - INFO - codeparrot_training - Step 45303: {'lr': 1.172048080472582e-05, 'samples': 23195648, 'steps': 45303, 'loss/train': 1.8865264654159546} 02/26/2022 12:16:01 - INFO - codeparrot_training - Step 45304: {'lr': 1.171553005655851e-05, 'samples': 23196160, 'steps': 45304, 'loss/train': 1.6483254432678223} 02/26/2022 12:16:05 - INFO - codeparrot_training - Step 45305: {'lr': 1.1710580329126196e-05, 'samples': 23196672, 'steps': 45305, 'loss/train': 1.6172759532928467} 02/26/2022 12:16:11 - INFO - codeparrot_training - Step 45306: {'lr': 1.1705631622450086e-05, 'samples': 23197184, 'steps': 45306, 'loss/train': 0.6871628165245056} 02/26/2022 12:16:14 - INFO - codeparrot_training - Step 45307: {'lr': 1.1700683936551355e-05, 'samples': 23197696, 'steps': 45307, 'loss/train': 1.934171438217163} 02/26/2022 12:16:20 - INFO - codeparrot_training - Step 45308: {'lr': 1.1695737271451263e-05, 'samples': 23198208, 'steps': 45308, 'loss/train': 1.0775582790374756} 02/26/2022 12:16:23 - INFO - codeparrot_training - Step 45309: {'lr': 1.169079162717096e-05, 'samples': 23198720, 'steps': 45309, 'loss/train': 1.492814064025879} 02/26/2022 12:16:29 - INFO - codeparrot_training - Step 45310: {'lr': 1.1685847003731598e-05, 'samples': 23199232, 'steps': 45310, 'loss/train': 1.3731632232666016} 02/26/2022 12:16:32 - INFO - codeparrot_training - Step 45311: {'lr': 1.1680903401154463e-05, 'samples': 23199744, 'steps': 45311, 'loss/train': 1.9341137409210205} 02/26/2022 12:16:38 - INFO - codeparrot_training - Step 45312: {'lr': 1.1675960819460595e-05, 'samples': 23200256, 'steps': 45312, 'loss/train': 2.4144809246063232} 02/26/2022 12:16:41 - INFO - codeparrot_training - Step 45313: {'lr': 1.1671019258671285e-05, 'samples': 23200768, 'steps': 45313, 'loss/train': 1.560051441192627} 02/26/2022 12:16:48 - INFO - codeparrot_training - Step 45314: {'lr': 1.1666078718807593e-05, 'samples': 23201280, 'steps': 45314, 'loss/train': 1.2260018587112427} 02/26/2022 12:16:51 - INFO - codeparrot_training - Step 45315: {'lr': 1.1661139199890813e-05, 'samples': 23201792, 'steps': 45315, 'loss/train': 0.16942980885505676} 02/26/2022 12:16:56 - INFO - codeparrot_training - Step 45316: {'lr': 1.1656200701941955e-05, 'samples': 23202304, 'steps': 45316, 'loss/train': 1.7891305685043335} 02/26/2022 12:17:00 - INFO - codeparrot_training - Step 45317: {'lr': 1.1651263224982305e-05, 'samples': 23202816, 'steps': 45317, 'loss/train': 1.336639404296875} 02/26/2022 12:17:06 - INFO - codeparrot_training - Step 45318: {'lr': 1.1646326769032934e-05, 'samples': 23203328, 'steps': 45318, 'loss/train': 1.3226101398468018} 02/26/2022 12:17:09 - INFO - codeparrot_training - Step 45319: {'lr': 1.1641391334115015e-05, 'samples': 23203840, 'steps': 45319, 'loss/train': 1.901580572128296} 02/26/2022 12:17:14 - INFO - codeparrot_training - Step 45320: {'lr': 1.16364569202497e-05, 'samples': 23204352, 'steps': 45320, 'loss/train': 1.0646063089370728} 02/26/2022 12:17:18 - INFO - codeparrot_training - Step 45321: {'lr': 1.1631523527458109e-05, 'samples': 23204864, 'steps': 45321, 'loss/train': 1.7868127822875977} 02/26/2022 12:17:23 - INFO - codeparrot_training - Step 45322: {'lr': 1.1626591155761368e-05, 'samples': 23205376, 'steps': 45322, 'loss/train': 1.2724719047546387} 02/26/2022 12:17:27 - INFO - codeparrot_training - Step 45323: {'lr': 1.1621659805180623e-05, 'samples': 23205888, 'steps': 45323, 'loss/train': 1.95565664768219} 02/26/2022 12:17:33 - INFO - codeparrot_training - Step 45324: {'lr': 1.1616729475737025e-05, 'samples': 23206400, 'steps': 45324, 'loss/train': 1.5162220001220703} 02/26/2022 12:17:36 - INFO - codeparrot_training - Step 45325: {'lr': 1.1611800167451642e-05, 'samples': 23206912, 'steps': 45325, 'loss/train': 0.9925150871276855} 02/26/2022 12:17:42 - INFO - codeparrot_training - Step 45326: {'lr': 1.1606871880345592e-05, 'samples': 23207424, 'steps': 45326, 'loss/train': 0.8877344727516174} 02/26/2022 12:17:45 - INFO - codeparrot_training - Step 45327: {'lr': 1.160194461444003e-05, 'samples': 23207936, 'steps': 45327, 'loss/train': 1.5658241510391235} 02/26/2022 12:17:51 - INFO - codeparrot_training - Step 45328: {'lr': 1.1597018369756018e-05, 'samples': 23208448, 'steps': 45328, 'loss/train': 2.043818235397339} 02/26/2022 12:17:54 - INFO - codeparrot_training - Step 45329: {'lr': 1.1592093146314709e-05, 'samples': 23208960, 'steps': 45329, 'loss/train': 2.1026744842529297} 02/26/2022 12:18:00 - INFO - codeparrot_training - Step 45330: {'lr': 1.1587168944137112e-05, 'samples': 23209472, 'steps': 45330, 'loss/train': 2.0544748306274414} 02/26/2022 12:18:03 - INFO - codeparrot_training - Step 45331: {'lr': 1.1582245763244432e-05, 'samples': 23209984, 'steps': 45331, 'loss/train': 1.0296118259429932} 02/26/2022 12:18:09 - INFO - codeparrot_training - Step 45332: {'lr': 1.1577323603657652e-05, 'samples': 23210496, 'steps': 45332, 'loss/train': 0.8347918391227722} 02/26/2022 12:18:12 - INFO - codeparrot_training - Step 45333: {'lr': 1.157240246539798e-05, 'samples': 23211008, 'steps': 45333, 'loss/train': 1.4511092901229858} 02/26/2022 12:18:19 - INFO - codeparrot_training - Step 45334: {'lr': 1.156748234848637e-05, 'samples': 23211520, 'steps': 45334, 'loss/train': 1.2427624464035034} 02/26/2022 12:18:22 - INFO - codeparrot_training - Step 45335: {'lr': 1.1562563252943998e-05, 'samples': 23212032, 'steps': 45335, 'loss/train': 2.073885679244995} 02/26/2022 12:18:28 - INFO - codeparrot_training - Step 45336: {'lr': 1.1557645178791847e-05, 'samples': 23212544, 'steps': 45336, 'loss/train': 1.9329290390014648} 02/26/2022 12:18:31 - INFO - codeparrot_training - Step 45337: {'lr': 1.1552728126051098e-05, 'samples': 23213056, 'steps': 45337, 'loss/train': 1.5023045539855957} 02/26/2022 12:18:37 - INFO - codeparrot_training - Step 45338: {'lr': 1.1547812094742677e-05, 'samples': 23213568, 'steps': 45338, 'loss/train': 1.5433142185211182} 02/26/2022 12:18:40 - INFO - codeparrot_training - Step 45339: {'lr': 1.1542897084887732e-05, 'samples': 23214080, 'steps': 45339, 'loss/train': 1.8299283981323242} 02/26/2022 12:18:46 - INFO - codeparrot_training - Step 45340: {'lr': 1.1537983096507277e-05, 'samples': 23214592, 'steps': 45340, 'loss/train': 1.4210702180862427} 02/26/2022 12:18:49 - INFO - codeparrot_training - Step 45341: {'lr': 1.1533070129622458e-05, 'samples': 23215104, 'steps': 45341, 'loss/train': 1.679061770439148} 02/26/2022 12:18:55 - INFO - codeparrot_training - Step 45342: {'lr': 1.1528158184254151e-05, 'samples': 23215616, 'steps': 45342, 'loss/train': 2.0270400047302246} 02/26/2022 12:18:58 - INFO - codeparrot_training - Step 45343: {'lr': 1.1523247260423558e-05, 'samples': 23216128, 'steps': 45343, 'loss/train': 2.6155829429626465} 02/26/2022 12:19:04 - INFO - codeparrot_training - Step 45344: {'lr': 1.1518337358151636e-05, 'samples': 23216640, 'steps': 45344, 'loss/train': 2.6217074394226074} 02/26/2022 12:19:07 - INFO - codeparrot_training - Step 45345: {'lr': 1.1513428477459398e-05, 'samples': 23217152, 'steps': 45345, 'loss/train': 1.8489625453948975} 02/26/2022 12:19:13 - INFO - codeparrot_training - Step 45346: {'lr': 1.1508520618367934e-05, 'samples': 23217664, 'steps': 45346, 'loss/train': 2.1252923011779785} 02/26/2022 12:19:16 - INFO - codeparrot_training - Step 45347: {'lr': 1.1503613780898231e-05, 'samples': 23218176, 'steps': 45347, 'loss/train': 0.4666687250137329} 02/26/2022 12:19:23 - INFO - codeparrot_training - Step 45348: {'lr': 1.1498707965071325e-05, 'samples': 23218688, 'steps': 45348, 'loss/train': 2.7555248737335205} 02/26/2022 12:19:26 - INFO - codeparrot_training - Step 45349: {'lr': 1.1493803170908174e-05, 'samples': 23219200, 'steps': 45349, 'loss/train': 2.4711475372314453} 02/26/2022 12:19:32 - INFO - codeparrot_training - Step 45350: {'lr': 1.1488899398429897e-05, 'samples': 23219712, 'steps': 45350, 'loss/train': 1.7018671035766602} 02/26/2022 12:19:35 - INFO - codeparrot_training - Step 45351: {'lr': 1.1483996647657424e-05, 'samples': 23220224, 'steps': 45351, 'loss/train': 2.3630166053771973} 02/26/2022 12:19:41 - INFO - codeparrot_training - Step 45352: {'lr': 1.1479094918611765e-05, 'samples': 23220736, 'steps': 45352, 'loss/train': 1.1727098226547241} 02/26/2022 12:19:44 - INFO - codeparrot_training - Step 45353: {'lr': 1.1474194211313905e-05, 'samples': 23221248, 'steps': 45353, 'loss/train': 1.8439149856567383} 02/26/2022 12:19:50 - INFO - codeparrot_training - Step 45354: {'lr': 1.146929452578488e-05, 'samples': 23221760, 'steps': 45354, 'loss/train': 1.3220343589782715} 02/26/2022 12:19:53 - INFO - codeparrot_training - Step 45355: {'lr': 1.1464395862045645e-05, 'samples': 23222272, 'steps': 45355, 'loss/train': 2.2462637424468994} 02/26/2022 12:19:59 - INFO - codeparrot_training - Step 45356: {'lr': 1.1459498220117214e-05, 'samples': 23222784, 'steps': 45356, 'loss/train': 2.5816359519958496} 02/26/2022 12:20:02 - INFO - codeparrot_training - Step 45357: {'lr': 1.1454601600020515e-05, 'samples': 23223296, 'steps': 45357, 'loss/train': 2.331681966781616} 02/26/2022 12:20:08 - INFO - codeparrot_training - Step 45358: {'lr': 1.1449706001776583e-05, 'samples': 23223808, 'steps': 45358, 'loss/train': 0.09742573648691177} 02/26/2022 12:20:11 - INFO - codeparrot_training - Step 45359: {'lr': 1.144481142540635e-05, 'samples': 23224320, 'steps': 45359, 'loss/train': 0.7433115839958191} 02/26/2022 12:20:18 - INFO - codeparrot_training - Step 45360: {'lr': 1.1439917870930795e-05, 'samples': 23224832, 'steps': 45360, 'loss/train': 1.5361292362213135} 02/26/2022 12:20:21 - INFO - codeparrot_training - Step 45361: {'lr': 1.1435025338370875e-05, 'samples': 23225344, 'steps': 45361, 'loss/train': 2.9006142616271973} 02/26/2022 12:20:27 - INFO - codeparrot_training - Step 45362: {'lr': 1.1430133827747574e-05, 'samples': 23225856, 'steps': 45362, 'loss/train': 1.2615761756896973} 02/26/2022 12:20:30 - INFO - codeparrot_training - Step 45363: {'lr': 1.142524333908182e-05, 'samples': 23226368, 'steps': 45363, 'loss/train': 1.749088168144226} 02/26/2022 12:20:36 - INFO - codeparrot_training - Step 45364: {'lr': 1.1420353872394567e-05, 'samples': 23226880, 'steps': 45364, 'loss/train': 1.0706027746200562} 02/26/2022 12:20:39 - INFO - codeparrot_training - Step 45365: {'lr': 1.1415465427706717e-05, 'samples': 23227392, 'steps': 45365, 'loss/train': 1.099647045135498} 02/26/2022 12:20:45 - INFO - codeparrot_training - Step 45366: {'lr': 1.1410578005039308e-05, 'samples': 23227904, 'steps': 45366, 'loss/train': 1.9706231355667114} 02/26/2022 12:20:48 - INFO - codeparrot_training - Step 45367: {'lr': 1.1405691604413182e-05, 'samples': 23228416, 'steps': 45367, 'loss/train': 1.6307319402694702} 02/26/2022 12:20:54 - INFO - codeparrot_training - Step 45368: {'lr': 1.1400806225849352e-05, 'samples': 23228928, 'steps': 45368, 'loss/train': 1.7613496780395508} 02/26/2022 12:20:57 - INFO - codeparrot_training - Step 45369: {'lr': 1.1395921869368664e-05, 'samples': 23229440, 'steps': 45369, 'loss/train': 1.8023744821548462} 02/26/2022 12:21:04 - INFO - codeparrot_training - Step 45370: {'lr': 1.1391038534992126e-05, 'samples': 23229952, 'steps': 45370, 'loss/train': 3.5374059677124023} 02/26/2022 12:21:07 - INFO - codeparrot_training - Step 45371: {'lr': 1.1386156222740557e-05, 'samples': 23230464, 'steps': 45371, 'loss/train': 2.701000928878784} 02/26/2022 12:21:12 - INFO - codeparrot_training - Step 45372: {'lr': 1.1381274932634994e-05, 'samples': 23230976, 'steps': 45372, 'loss/train': 2.431722402572632} 02/26/2022 12:21:18 - INFO - codeparrot_training - Step 45373: {'lr': 1.1376394664696227e-05, 'samples': 23231488, 'steps': 45373, 'loss/train': 1.1495332717895508} 02/26/2022 12:21:22 - INFO - codeparrot_training - Step 45374: {'lr': 1.1371515418945239e-05, 'samples': 23232000, 'steps': 45374, 'loss/train': 2.1136183738708496} 02/26/2022 12:21:25 - INFO - codeparrot_training - Step 45375: {'lr': 1.1366637195402874e-05, 'samples': 23232512, 'steps': 45375, 'loss/train': 1.4459030628204346} 02/26/2022 12:21:30 - INFO - codeparrot_training - Step 45376: {'lr': 1.1361759994090115e-05, 'samples': 23233024, 'steps': 45376, 'loss/train': 1.4107810258865356} 02/26/2022 12:21:36 - INFO - codeparrot_training - Step 45377: {'lr': 1.1356883815027751e-05, 'samples': 23233536, 'steps': 45377, 'loss/train': 1.5117871761322021} 02/26/2022 12:21:39 - INFO - codeparrot_training - Step 45378: {'lr': 1.135200865823674e-05, 'samples': 23234048, 'steps': 45378, 'loss/train': 0.8988096714019775} 02/26/2022 12:21:46 - INFO - codeparrot_training - Step 45379: {'lr': 1.1347134523737923e-05, 'samples': 23234560, 'steps': 45379, 'loss/train': 1.5875827074050903} 02/26/2022 12:21:49 - INFO - codeparrot_training - Step 45380: {'lr': 1.134226141155223e-05, 'samples': 23235072, 'steps': 45380, 'loss/train': 2.5511093139648438} 02/26/2022 12:21:55 - INFO - codeparrot_training - Step 45381: {'lr': 1.1337389321700504e-05, 'samples': 23235584, 'steps': 45381, 'loss/train': 2.0994980335235596} 02/26/2022 12:21:58 - INFO - codeparrot_training - Step 45382: {'lr': 1.1332518254203616e-05, 'samples': 23236096, 'steps': 45382, 'loss/train': 1.953726053237915} 02/26/2022 12:22:04 - INFO - codeparrot_training - Step 45383: {'lr': 1.1327648209082413e-05, 'samples': 23236608, 'steps': 45383, 'loss/train': 1.3440583944320679} 02/26/2022 12:22:08 - INFO - codeparrot_training - Step 45384: {'lr': 1.1322779186357796e-05, 'samples': 23237120, 'steps': 45384, 'loss/train': 2.516601085662842} 02/26/2022 12:22:11 - INFO - codeparrot_training - Step 45385: {'lr': 1.1317911186050606e-05, 'samples': 23237632, 'steps': 45385, 'loss/train': 1.6852357387542725} 02/26/2022 12:22:17 - INFO - codeparrot_training - Step 45386: {'lr': 1.1313044208181717e-05, 'samples': 23238144, 'steps': 45386, 'loss/train': 1.2206635475158691} 02/26/2022 12:22:20 - INFO - codeparrot_training - Step 45387: {'lr': 1.1308178252771888e-05, 'samples': 23238656, 'steps': 45387, 'loss/train': 1.815065622329712} 02/26/2022 12:22:26 - INFO - codeparrot_training - Step 45388: {'lr': 1.1303313319842078e-05, 'samples': 23239168, 'steps': 45388, 'loss/train': 0.47971311211586} 02/26/2022 12:22:29 - INFO - codeparrot_training - Step 45389: {'lr': 1.1298449409413076e-05, 'samples': 23239680, 'steps': 45389, 'loss/train': 2.2196054458618164} 02/26/2022 12:22:35 - INFO - codeparrot_training - Step 45390: {'lr': 1.129358652150575e-05, 'samples': 23240192, 'steps': 45390, 'loss/train': 2.972101926803589} 02/26/2022 12:22:38 - INFO - codeparrot_training - Step 45391: {'lr': 1.1288724656140837e-05, 'samples': 23240704, 'steps': 45391, 'loss/train': 1.461857795715332} 02/26/2022 12:22:44 - INFO - codeparrot_training - Step 45392: {'lr': 1.1283863813339262e-05, 'samples': 23241216, 'steps': 45392, 'loss/train': 0.8047551512718201} 02/26/2022 12:22:47 - INFO - codeparrot_training - Step 45393: {'lr': 1.1279003993121817e-05, 'samples': 23241728, 'steps': 45393, 'loss/train': 0.6176438331604004} 02/26/2022 12:22:53 - INFO - codeparrot_training - Step 45394: {'lr': 1.1274145195509317e-05, 'samples': 23242240, 'steps': 45394, 'loss/train': 1.6496230363845825} 02/26/2022 12:22:56 - INFO - codeparrot_training - Step 45395: {'lr': 1.1269287420522578e-05, 'samples': 23242752, 'steps': 45395, 'loss/train': 0.8046233654022217} 02/26/2022 12:23:02 - INFO - codeparrot_training - Step 45396: {'lr': 1.1264430668182363e-05, 'samples': 23243264, 'steps': 45396, 'loss/train': 1.852026343345642} 02/26/2022 12:23:06 - INFO - codeparrot_training - Step 45397: {'lr': 1.125957493850957e-05, 'samples': 23243776, 'steps': 45397, 'loss/train': 1.6996564865112305} 02/26/2022 12:23:11 - INFO - codeparrot_training - Step 45398: {'lr': 1.1254720231524934e-05, 'samples': 23244288, 'steps': 45398, 'loss/train': 1.3712424039840698} 02/26/2022 12:23:15 - INFO - codeparrot_training - Step 45399: {'lr': 1.1249866547249272e-05, 'samples': 23244800, 'steps': 45399, 'loss/train': 2.1269595623016357} 02/26/2022 12:23:20 - INFO - codeparrot_training - Step 45400: {'lr': 1.1245013885703342e-05, 'samples': 23245312, 'steps': 45400, 'loss/train': 1.056168556213379} 02/26/2022 12:23:24 - INFO - codeparrot_training - Step 45401: {'lr': 1.1240162246907992e-05, 'samples': 23245824, 'steps': 45401, 'loss/train': 1.077668309211731} 02/26/2022 12:23:29 - INFO - codeparrot_training - Step 45402: {'lr': 1.1235311630883982e-05, 'samples': 23246336, 'steps': 45402, 'loss/train': 1.8941786289215088} 02/26/2022 12:23:33 - INFO - codeparrot_training - Step 45403: {'lr': 1.1230462037652046e-05, 'samples': 23246848, 'steps': 45403, 'loss/train': 1.5298796892166138} 02/26/2022 12:23:38 - INFO - codeparrot_training - Step 45404: {'lr': 1.1225613467232998e-05, 'samples': 23247360, 'steps': 45404, 'loss/train': 1.7032171487808228} 02/26/2022 12:23:45 - INFO - codeparrot_training - Step 45405: {'lr': 1.1220765919647602e-05, 'samples': 23247872, 'steps': 45405, 'loss/train': 1.6266518831253052} 02/26/2022 12:23:48 - INFO - codeparrot_training - Step 45406: {'lr': 1.1215919394916619e-05, 'samples': 23248384, 'steps': 45406, 'loss/train': 0.9353079795837402} 02/26/2022 12:23:54 - INFO - codeparrot_training - Step 45407: {'lr': 1.1211073893060863e-05, 'samples': 23248896, 'steps': 45407, 'loss/train': 1.487837314605713} 02/26/2022 12:23:57 - INFO - codeparrot_training - Step 45408: {'lr': 1.1206229414100988e-05, 'samples': 23249408, 'steps': 45408, 'loss/train': 0.3855515718460083} 02/26/2022 12:24:03 - INFO - codeparrot_training - Step 45409: {'lr': 1.1201385958057836e-05, 'samples': 23249920, 'steps': 45409, 'loss/train': 1.765134334564209} 02/26/2022 12:24:06 - INFO - codeparrot_training - Step 45410: {'lr': 1.1196543524952057e-05, 'samples': 23250432, 'steps': 45410, 'loss/train': 2.3218090534210205} 02/26/2022 12:24:12 - INFO - codeparrot_training - Step 45411: {'lr': 1.1191702114804552e-05, 'samples': 23250944, 'steps': 45411, 'loss/train': 0.8160605430603027} 02/26/2022 12:24:15 - INFO - codeparrot_training - Step 45412: {'lr': 1.1186861727635889e-05, 'samples': 23251456, 'steps': 45412, 'loss/train': 1.7722963094711304} 02/26/2022 12:24:19 - INFO - codeparrot_training - Step 45413: {'lr': 1.1182022363466909e-05, 'samples': 23251968, 'steps': 45413, 'loss/train': 0.650709867477417} 02/26/2022 12:24:24 - INFO - codeparrot_training - Step 45414: {'lr': 1.1177184022318293e-05, 'samples': 23252480, 'steps': 45414, 'loss/train': 1.802277684211731} 02/26/2022 12:24:31 - INFO - codeparrot_training - Step 45415: {'lr': 1.1172346704210857e-05, 'samples': 23252992, 'steps': 45415, 'loss/train': 3.045698881149292} 02/26/2022 12:24:35 - INFO - codeparrot_training - Step 45416: {'lr': 1.1167510409165166e-05, 'samples': 23253504, 'steps': 45416, 'loss/train': 2.619114875793457} 02/26/2022 12:24:38 - INFO - codeparrot_training - Step 45417: {'lr': 1.1162675137202067e-05, 'samples': 23254016, 'steps': 45417, 'loss/train': 0.9451178312301636} 02/26/2022 12:24:44 - INFO - codeparrot_training - Step 45418: {'lr': 1.115784088834218e-05, 'samples': 23254528, 'steps': 45418, 'loss/train': 2.0916895866394043} 02/26/2022 12:24:47 - INFO - codeparrot_training - Step 45419: {'lr': 1.1153007662606296e-05, 'samples': 23255040, 'steps': 45419, 'loss/train': 1.4889509677886963} 02/26/2022 12:24:53 - INFO - codeparrot_training - Step 45420: {'lr': 1.1148175460015092e-05, 'samples': 23255552, 'steps': 45420, 'loss/train': 0.9746044278144836} 02/26/2022 12:24:56 - INFO - codeparrot_training - Step 45421: {'lr': 1.1143344280589273e-05, 'samples': 23256064, 'steps': 45421, 'loss/train': 0.9515894055366516} 02/26/2022 12:25:02 - INFO - codeparrot_training - Step 45422: {'lr': 1.1138514124349463e-05, 'samples': 23256576, 'steps': 45422, 'loss/train': 1.6756755113601685} 02/26/2022 12:25:05 - INFO - codeparrot_training - Step 45423: {'lr': 1.113368499131645e-05, 'samples': 23257088, 'steps': 45423, 'loss/train': 3.002045154571533} 02/26/2022 12:25:10 - INFO - codeparrot_training - Step 45424: {'lr': 1.1128856881510913e-05, 'samples': 23257600, 'steps': 45424, 'loss/train': 2.4058916568756104} 02/26/2022 12:25:14 - INFO - codeparrot_training - Step 45425: {'lr': 1.1124029794953472e-05, 'samples': 23258112, 'steps': 45425, 'loss/train': 0.14879821240901947} 02/26/2022 12:25:20 - INFO - codeparrot_training - Step 45426: {'lr': 1.1119203731664807e-05, 'samples': 23258624, 'steps': 45426, 'loss/train': 0.17805123329162598} 02/26/2022 12:25:26 - INFO - codeparrot_training - Step 45427: {'lr': 1.1114378691665649e-05, 'samples': 23259136, 'steps': 45427, 'loss/train': 1.9532897472381592} 02/26/2022 12:25:29 - INFO - codeparrot_training - Step 45428: {'lr': 1.1109554674976651e-05, 'samples': 23259648, 'steps': 45428, 'loss/train': 1.526245355606079} 02/26/2022 12:25:34 - INFO - codeparrot_training - Step 45429: {'lr': 1.1104731681618463e-05, 'samples': 23260160, 'steps': 45429, 'loss/train': 1.6933094263076782} 02/26/2022 12:25:38 - INFO - codeparrot_training - Step 45430: {'lr': 1.1099909711611705e-05, 'samples': 23260672, 'steps': 45430, 'loss/train': 1.3095364570617676} 02/26/2022 12:25:43 - INFO - codeparrot_training - Step 45431: {'lr': 1.1095088764977112e-05, 'samples': 23261184, 'steps': 45431, 'loss/train': 1.7271593809127808} 02/26/2022 12:25:47 - INFO - codeparrot_training - Step 45432: {'lr': 1.1090268841735307e-05, 'samples': 23261696, 'steps': 45432, 'loss/train': 3.4255309104919434} 02/26/2022 12:25:52 - INFO - codeparrot_training - Step 45433: {'lr': 1.1085449941906912e-05, 'samples': 23262208, 'steps': 45433, 'loss/train': 1.8797121047973633} 02/26/2022 12:25:56 - INFO - codeparrot_training - Step 45434: {'lr': 1.1080632065512546e-05, 'samples': 23262720, 'steps': 45434, 'loss/train': 1.8912968635559082} 02/26/2022 12:26:02 - INFO - codeparrot_training - Step 45435: {'lr': 1.1075815212572948e-05, 'samples': 23263232, 'steps': 45435, 'loss/train': 1.7881375551223755} 02/26/2022 12:26:06 - INFO - codeparrot_training - Step 45436: {'lr': 1.1070999383108682e-05, 'samples': 23263744, 'steps': 45436, 'loss/train': 1.7056219577789307} 02/26/2022 12:26:11 - INFO - codeparrot_training - Step 45437: {'lr': 1.106618457714037e-05, 'samples': 23264256, 'steps': 45437, 'loss/train': 1.3320502042770386} 02/26/2022 12:26:15 - INFO - codeparrot_training - Step 45438: {'lr': 1.1061370794688635e-05, 'samples': 23264768, 'steps': 45438, 'loss/train': 2.526350736618042} 02/26/2022 12:26:20 - INFO - codeparrot_training - Step 45439: {'lr': 1.1056558035774156e-05, 'samples': 23265280, 'steps': 45439, 'loss/train': 1.2630500793457031} 02/26/2022 12:26:23 - INFO - codeparrot_training - Step 45440: {'lr': 1.105174630041747e-05, 'samples': 23265792, 'steps': 45440, 'loss/train': 2.1425564289093018} 02/26/2022 12:26:29 - INFO - codeparrot_training - Step 45441: {'lr': 1.104693558863923e-05, 'samples': 23266304, 'steps': 45441, 'loss/train': 2.606959104537964} 02/26/2022 12:26:33 - INFO - codeparrot_training - Step 45442: {'lr': 1.104212590046011e-05, 'samples': 23266816, 'steps': 45442, 'loss/train': 1.2670127153396606} 02/26/2022 12:26:38 - INFO - codeparrot_training - Step 45443: {'lr': 1.1037317235900568e-05, 'samples': 23267328, 'steps': 45443, 'loss/train': 2.846410036087036} 02/26/2022 12:26:41 - INFO - codeparrot_training - Step 45444: {'lr': 1.103250959498131e-05, 'samples': 23267840, 'steps': 45444, 'loss/train': 2.292482852935791} 02/26/2022 12:26:47 - INFO - codeparrot_training - Step 45445: {'lr': 1.1027702977722875e-05, 'samples': 23268352, 'steps': 45445, 'loss/train': 1.5529601573944092} 02/26/2022 12:26:50 - INFO - codeparrot_training - Step 45446: {'lr': 1.1022897384145941e-05, 'samples': 23268864, 'steps': 45446, 'loss/train': 1.0820204019546509} 02/26/2022 12:26:56 - INFO - codeparrot_training - Step 45447: {'lr': 1.1018092814270963e-05, 'samples': 23269376, 'steps': 45447, 'loss/train': 0.5086148977279663} 02/26/2022 12:27:00 - INFO - codeparrot_training - Step 45448: {'lr': 1.101328926811862e-05, 'samples': 23269888, 'steps': 45448, 'loss/train': 2.6104953289031982} 02/26/2022 12:27:05 - INFO - codeparrot_training - Step 45449: {'lr': 1.1008486745709423e-05, 'samples': 23270400, 'steps': 45449, 'loss/train': 1.473425030708313} 02/26/2022 12:27:09 - INFO - codeparrot_training - Step 45450: {'lr': 1.1003685247064076e-05, 'samples': 23270912, 'steps': 45450, 'loss/train': 1.9737762212753296} 02/26/2022 12:27:15 - INFO - codeparrot_training - Step 45451: {'lr': 1.0998884772202955e-05, 'samples': 23271424, 'steps': 45451, 'loss/train': 0.8413974642753601} 02/26/2022 12:27:18 - INFO - codeparrot_training - Step 45452: {'lr': 1.0994085321146763e-05, 'samples': 23271936, 'steps': 45452, 'loss/train': 2.181739330291748} 02/26/2022 12:27:24 - INFO - codeparrot_training - Step 45453: {'lr': 1.0989286893915984e-05, 'samples': 23272448, 'steps': 45453, 'loss/train': 1.670751929283142} 02/26/2022 12:27:27 - INFO - codeparrot_training - Step 45454: {'lr': 1.0984489490531242e-05, 'samples': 23272960, 'steps': 45454, 'loss/train': 1.9378257989883423} 02/26/2022 12:27:33 - INFO - codeparrot_training - Step 45455: {'lr': 1.0979693111013045e-05, 'samples': 23273472, 'steps': 45455, 'loss/train': 0.9125990867614746} 02/26/2022 12:27:36 - INFO - codeparrot_training - Step 45456: {'lr': 1.0974897755381935e-05, 'samples': 23273984, 'steps': 45456, 'loss/train': 2.0668399333953857} 02/26/2022 12:27:42 - INFO - codeparrot_training - Step 45457: {'lr': 1.097010342365845e-05, 'samples': 23274496, 'steps': 45457, 'loss/train': 1.9885073900222778} 02/26/2022 12:27:45 - INFO - codeparrot_training - Step 45458: {'lr': 1.0965310115863159e-05, 'samples': 23275008, 'steps': 45458, 'loss/train': 1.798528790473938} 02/26/2022 12:27:51 - INFO - codeparrot_training - Step 45459: {'lr': 1.096051783201657e-05, 'samples': 23275520, 'steps': 45459, 'loss/train': 2.220317840576172} 02/26/2022 12:27:54 - INFO - codeparrot_training - Step 45460: {'lr': 1.0955726572139252e-05, 'samples': 23276032, 'steps': 45460, 'loss/train': 1.0446014404296875} 02/26/2022 12:28:01 - INFO - codeparrot_training - Step 45461: {'lr': 1.0950936336251632e-05, 'samples': 23276544, 'steps': 45461, 'loss/train': 1.6791778802871704} 02/26/2022 12:28:04 - INFO - codeparrot_training - Step 45462: {'lr': 1.0946147124374361e-05, 'samples': 23277056, 'steps': 45462, 'loss/train': 1.4799556732177734} 02/26/2022 12:28:09 - INFO - codeparrot_training - Step 45463: {'lr': 1.094135893652784e-05, 'samples': 23277568, 'steps': 45463, 'loss/train': 1.6639615297317505} 02/26/2022 12:28:13 - INFO - codeparrot_training - Step 45464: {'lr': 1.0936571772732662e-05, 'samples': 23278080, 'steps': 45464, 'loss/train': 2.132904052734375} 02/26/2022 12:28:18 - INFO - codeparrot_training - Step 45465: {'lr': 1.0931785633009256e-05, 'samples': 23278592, 'steps': 45465, 'loss/train': 2.355464458465576} 02/26/2022 12:28:22 - INFO - codeparrot_training - Step 45466: {'lr': 1.0927000517378216e-05, 'samples': 23279104, 'steps': 45466, 'loss/train': 1.3249207735061646} 02/26/2022 12:28:27 - INFO - codeparrot_training - Step 45467: {'lr': 1.0922216425859943e-05, 'samples': 23279616, 'steps': 45467, 'loss/train': 1.9733928442001343} 02/26/2022 12:28:31 - INFO - codeparrot_training - Step 45468: {'lr': 1.0917433358475031e-05, 'samples': 23280128, 'steps': 45468, 'loss/train': 2.930763006210327} 02/26/2022 12:28:36 - INFO - codeparrot_training - Step 45469: {'lr': 1.0912651315243883e-05, 'samples': 23280640, 'steps': 45469, 'loss/train': 1.4898325204849243} 02/26/2022 12:28:40 - INFO - codeparrot_training - Step 45470: {'lr': 1.0907870296187034e-05, 'samples': 23281152, 'steps': 45470, 'loss/train': 1.352459192276001} 02/26/2022 12:28:46 - INFO - codeparrot_training - Step 45471: {'lr': 1.0903090301324914e-05, 'samples': 23281664, 'steps': 45471, 'loss/train': 2.132077693939209} 02/26/2022 12:28:49 - INFO - codeparrot_training - Step 45472: {'lr': 1.0898311330678117e-05, 'samples': 23282176, 'steps': 45472, 'loss/train': 1.977565050125122} 02/26/2022 12:28:55 - INFO - codeparrot_training - Step 45473: {'lr': 1.0893533384266936e-05, 'samples': 23282688, 'steps': 45473, 'loss/train': 1.3192912340164185} 02/26/2022 12:28:58 - INFO - codeparrot_training - Step 45474: {'lr': 1.0888756462111987e-05, 'samples': 23283200, 'steps': 45474, 'loss/train': 1.3958088159561157} 02/26/2022 12:29:04 - INFO - codeparrot_training - Step 45475: {'lr': 1.0883980564233676e-05, 'samples': 23283712, 'steps': 45475, 'loss/train': 2.227958917617798} 02/26/2022 12:29:07 - INFO - codeparrot_training - Step 45476: {'lr': 1.0879205690652428e-05, 'samples': 23284224, 'steps': 45476, 'loss/train': 1.4331892728805542} 02/26/2022 12:29:13 - INFO - codeparrot_training - Step 45477: {'lr': 1.0874431841388754e-05, 'samples': 23284736, 'steps': 45477, 'loss/train': 1.2371387481689453} 02/26/2022 12:29:16 - INFO - codeparrot_training - Step 45478: {'lr': 1.0869659016463113e-05, 'samples': 23285248, 'steps': 45478, 'loss/train': 2.2302229404449463} 02/26/2022 12:29:22 - INFO - codeparrot_training - Step 45479: {'lr': 1.0864887215895902e-05, 'samples': 23285760, 'steps': 45479, 'loss/train': 1.913045048713684} 02/26/2022 12:29:25 - INFO - codeparrot_training - Step 45480: {'lr': 1.086011643970755e-05, 'samples': 23286272, 'steps': 45480, 'loss/train': 1.6322394609451294} 02/26/2022 12:29:31 - INFO - codeparrot_training - Step 45481: {'lr': 1.0855346687918543e-05, 'samples': 23286784, 'steps': 45481, 'loss/train': 0.5191765427589417} 02/26/2022 12:29:35 - INFO - codeparrot_training - Step 45482: {'lr': 1.0850577960549307e-05, 'samples': 23287296, 'steps': 45482, 'loss/train': 0.7215321063995361} 02/26/2022 12:29:40 - INFO - codeparrot_training - Step 45483: {'lr': 1.0845810257620242e-05, 'samples': 23287808, 'steps': 45483, 'loss/train': 1.601241111755371} 02/26/2022 12:29:44 - INFO - codeparrot_training - Step 45484: {'lr': 1.0841043579151777e-05, 'samples': 23288320, 'steps': 45484, 'loss/train': 1.6214818954467773} 02/26/2022 12:29:51 - INFO - codeparrot_training - Step 45485: {'lr': 1.0836277925164368e-05, 'samples': 23288832, 'steps': 45485, 'loss/train': 2.210165500640869} 02/26/2022 12:29:54 - INFO - codeparrot_training - Step 45486: {'lr': 1.0831513295678385e-05, 'samples': 23289344, 'steps': 45486, 'loss/train': 1.294748067855835} 02/26/2022 12:30:00 - INFO - codeparrot_training - Step 45487: {'lr': 1.082674969071426e-05, 'samples': 23289856, 'steps': 45487, 'loss/train': 0.06890608370304108} 02/26/2022 12:30:03 - INFO - codeparrot_training - Step 45488: {'lr': 1.0821987110292364e-05, 'samples': 23290368, 'steps': 45488, 'loss/train': 1.5776761770248413} 02/26/2022 12:30:06 - INFO - codeparrot_training - Step 45489: {'lr': 1.081722555443318e-05, 'samples': 23290880, 'steps': 45489, 'loss/train': 2.2941017150878906} 02/26/2022 12:30:12 - INFO - codeparrot_training - Step 45490: {'lr': 1.0812465023157025e-05, 'samples': 23291392, 'steps': 45490, 'loss/train': 1.2923625707626343} 02/26/2022 12:30:15 - INFO - codeparrot_training - Step 45491: {'lr': 1.0807705516484328e-05, 'samples': 23291904, 'steps': 45491, 'loss/train': 1.4162800312042236} 02/26/2022 12:30:21 - INFO - codeparrot_training - Step 45492: {'lr': 1.0802947034435435e-05, 'samples': 23292416, 'steps': 45492, 'loss/train': 1.673720359802246} 02/26/2022 12:30:24 - INFO - codeparrot_training - Step 45493: {'lr': 1.0798189577030798e-05, 'samples': 23292928, 'steps': 45493, 'loss/train': 1.978043556213379} 02/26/2022 12:30:30 - INFO - codeparrot_training - Step 45494: {'lr': 1.0793433144290766e-05, 'samples': 23293440, 'steps': 45494, 'loss/train': 2.053358554840088} 02/26/2022 12:30:34 - INFO - codeparrot_training - Step 45495: {'lr': 1.0788677736235708e-05, 'samples': 23293952, 'steps': 45495, 'loss/train': 1.8812906742095947} 02/26/2022 12:30:40 - INFO - codeparrot_training - Step 45496: {'lr': 1.0783923352885943e-05, 'samples': 23294464, 'steps': 45496, 'loss/train': 2.3655803203582764} 02/26/2022 12:30:43 - INFO - codeparrot_training - Step 45497: {'lr': 1.0779169994261956e-05, 'samples': 23294976, 'steps': 45497, 'loss/train': 1.645861029624939} 02/26/2022 12:30:49 - INFO - codeparrot_training - Step 45498: {'lr': 1.0774417660384034e-05, 'samples': 23295488, 'steps': 45498, 'loss/train': 1.7130159139633179} 02/26/2022 12:30:52 - INFO - codeparrot_training - Step 45499: {'lr': 1.0769666351272522e-05, 'samples': 23296000, 'steps': 45499, 'loss/train': 2.355069160461426} 02/26/2022 12:30:58 - INFO - codeparrot_training - Step 45500: {'lr': 1.0764916066947795e-05, 'samples': 23296512, 'steps': 45500, 'loss/train': 0.8156676888465881} 02/26/2022 12:31:01 - INFO - codeparrot_training - Step 45501: {'lr': 1.0760166807430194e-05, 'samples': 23297024, 'steps': 45501, 'loss/train': 2.6515986919403076} 02/26/2022 12:31:07 - INFO - codeparrot_training - Step 45502: {'lr': 1.0755418572740067e-05, 'samples': 23297536, 'steps': 45502, 'loss/train': 0.25662875175476074} 02/26/2022 12:31:11 - INFO - codeparrot_training - Step 45503: {'lr': 1.075067136289784e-05, 'samples': 23298048, 'steps': 45503, 'loss/train': 1.7687629461288452} 02/26/2022 12:31:16 - INFO - codeparrot_training - Step 45504: {'lr': 1.0745925177923665e-05, 'samples': 23298560, 'steps': 45504, 'loss/train': 4.2185163497924805} 02/26/2022 12:31:19 - INFO - codeparrot_training - Step 45505: {'lr': 1.0741180017838026e-05, 'samples': 23299072, 'steps': 45505, 'loss/train': 2.6711742877960205} 02/26/2022 12:31:26 - INFO - codeparrot_training - Step 45506: {'lr': 1.0736435882661156e-05, 'samples': 23299584, 'steps': 45506, 'loss/train': 1.556153416633606} 02/26/2022 12:31:30 - INFO - codeparrot_training - Step 45507: {'lr': 1.073169277241351e-05, 'samples': 23300096, 'steps': 45507, 'loss/train': 1.7478265762329102} 02/26/2022 12:31:33 - INFO - codeparrot_training - Step 45508: {'lr': 1.0726950687115211e-05, 'samples': 23300608, 'steps': 45508, 'loss/train': 0.37199845910072327} 02/26/2022 12:31:39 - INFO - codeparrot_training - Step 45509: {'lr': 1.0722209626786744e-05, 'samples': 23301120, 'steps': 45509, 'loss/train': 0.9415773153305054} 02/26/2022 12:31:44 - INFO - codeparrot_training - Step 45510: {'lr': 1.0717469591448315e-05, 'samples': 23301632, 'steps': 45510, 'loss/train': 1.8271976709365845} 02/26/2022 12:31:48 - INFO - codeparrot_training - Step 45511: {'lr': 1.0712730581120322e-05, 'samples': 23302144, 'steps': 45511, 'loss/train': 1.9965388774871826} 02/26/2022 12:31:53 - INFO - codeparrot_training - Step 45512: {'lr': 1.0707992595822946e-05, 'samples': 23302656, 'steps': 45512, 'loss/train': 1.8485907316207886} 02/26/2022 12:31:57 - INFO - codeparrot_training - Step 45513: {'lr': 1.0703255635576559e-05, 'samples': 23303168, 'steps': 45513, 'loss/train': 1.9083240032196045} 02/26/2022 12:32:02 - INFO - codeparrot_training - Step 45514: {'lr': 1.0698519700401422e-05, 'samples': 23303680, 'steps': 45514, 'loss/train': 2.5984535217285156} 02/26/2022 12:32:06 - INFO - codeparrot_training - Step 45515: {'lr': 1.0693784790317878e-05, 'samples': 23304192, 'steps': 45515, 'loss/train': 1.425366759300232} 02/26/2022 12:32:11 - INFO - codeparrot_training - Step 45516: {'lr': 1.0689050905346165e-05, 'samples': 23304704, 'steps': 45516, 'loss/train': 1.091313362121582} 02/26/2022 12:32:15 - INFO - codeparrot_training - Step 45517: {'lr': 1.0684318045506597e-05, 'samples': 23305216, 'steps': 45517, 'loss/train': 1.4275914430618286} 02/26/2022 12:32:21 - INFO - codeparrot_training - Step 45518: {'lr': 1.0679586210819354e-05, 'samples': 23305728, 'steps': 45518, 'loss/train': 1.656777262687683} 02/26/2022 12:32:24 - INFO - codeparrot_training - Step 45519: {'lr': 1.0674855401304805e-05, 'samples': 23306240, 'steps': 45519, 'loss/train': 1.0106686353683472} 02/26/2022 12:32:30 - INFO - codeparrot_training - Step 45520: {'lr': 1.067012561698319e-05, 'samples': 23306752, 'steps': 45520, 'loss/train': 1.3933219909667969} 02/26/2022 12:32:33 - INFO - codeparrot_training - Step 45521: {'lr': 1.0665396857874764e-05, 'samples': 23307264, 'steps': 45521, 'loss/train': 1.9338077306747437} 02/26/2022 12:32:39 - INFO - codeparrot_training - Step 45522: {'lr': 1.0660669123999767e-05, 'samples': 23307776, 'steps': 45522, 'loss/train': 1.6024292707443237} 02/26/2022 12:32:42 - INFO - codeparrot_training - Step 45523: {'lr': 1.0655942415378456e-05, 'samples': 23308288, 'steps': 45523, 'loss/train': 1.117063045501709} 02/26/2022 12:32:48 - INFO - codeparrot_training - Step 45524: {'lr': 1.0651216732031094e-05, 'samples': 23308800, 'steps': 45524, 'loss/train': 1.8036937713623047} 02/26/2022 12:32:52 - INFO - codeparrot_training - Step 45525: {'lr': 1.0646492073977915e-05, 'samples': 23309312, 'steps': 45525, 'loss/train': 2.4579007625579834} 02/26/2022 12:32:55 - INFO - codeparrot_training - Step 45526: {'lr': 1.064176844123918e-05, 'samples': 23309824, 'steps': 45526, 'loss/train': 2.0174057483673096} 02/26/2022 12:33:01 - INFO - codeparrot_training - Step 45527: {'lr': 1.0637045833835069e-05, 'samples': 23310336, 'steps': 45527, 'loss/train': 0.6836249828338623} 02/26/2022 12:33:05 - INFO - codeparrot_training - Step 45528: {'lr': 1.0632324251785869e-05, 'samples': 23310848, 'steps': 45528, 'loss/train': 2.2893130779266357} 02/26/2022 12:33:10 - INFO - codeparrot_training - Step 45529: {'lr': 1.0627603695111787e-05, 'samples': 23311360, 'steps': 45529, 'loss/train': 0.9735177159309387} 02/26/2022 12:33:14 - INFO - codeparrot_training - Step 45530: {'lr': 1.0622884163833058e-05, 'samples': 23311872, 'steps': 45530, 'loss/train': 1.6222349405288696} 02/26/2022 12:33:19 - INFO - codeparrot_training - Step 45531: {'lr': 1.061816565796983e-05, 'samples': 23312384, 'steps': 45531, 'loss/train': 1.2825584411621094} 02/26/2022 12:33:23 - INFO - codeparrot_training - Step 45532: {'lr': 1.0613448177542423e-05, 'samples': 23312896, 'steps': 45532, 'loss/train': 1.4990627765655518} 02/26/2022 12:33:29 - INFO - codeparrot_training - Step 45533: {'lr': 1.0608731722570958e-05, 'samples': 23313408, 'steps': 45533, 'loss/train': 1.6389977931976318} 02/26/2022 12:33:32 - INFO - codeparrot_training - Step 45534: {'lr': 1.0604016293075696e-05, 'samples': 23313920, 'steps': 45534, 'loss/train': 1.8013746738433838} 02/26/2022 12:33:38 - INFO - codeparrot_training - Step 45535: {'lr': 1.0599301889076762e-05, 'samples': 23314432, 'steps': 45535, 'loss/train': 3.0061020851135254} 02/26/2022 12:33:41 - INFO - codeparrot_training - Step 45536: {'lr': 1.0594588510594445e-05, 'samples': 23314944, 'steps': 45536, 'loss/train': 0.9037851691246033} 02/26/2022 12:33:47 - INFO - codeparrot_training - Step 45537: {'lr': 1.0589876157648865e-05, 'samples': 23315456, 'steps': 45537, 'loss/train': 1.9245526790618896} 02/26/2022 12:33:50 - INFO - codeparrot_training - Step 45538: {'lr': 1.0585164830260285e-05, 'samples': 23315968, 'steps': 45538, 'loss/train': 1.0642341375350952} 02/26/2022 12:33:56 - INFO - codeparrot_training - Step 45539: {'lr': 1.0580454528448774e-05, 'samples': 23316480, 'steps': 45539, 'loss/train': 1.3503497838974} 02/26/2022 12:33:59 - INFO - codeparrot_training - Step 45540: {'lr': 1.057574525223462e-05, 'samples': 23316992, 'steps': 45540, 'loss/train': 1.8431942462921143} 02/26/2022 12:34:05 - INFO - codeparrot_training - Step 45541: {'lr': 1.057103700163789e-05, 'samples': 23317504, 'steps': 45541, 'loss/train': 2.5416059494018555} 02/26/2022 12:34:08 - INFO - codeparrot_training - Step 45542: {'lr': 1.0566329776678874e-05, 'samples': 23318016, 'steps': 45542, 'loss/train': 1.654799461364746} 02/26/2022 12:34:15 - INFO - codeparrot_training - Step 45543: {'lr': 1.0561623577377639e-05, 'samples': 23318528, 'steps': 45543, 'loss/train': 2.061863899230957} 02/26/2022 12:34:18 - INFO - codeparrot_training - Step 45544: {'lr': 1.055691840375439e-05, 'samples': 23319040, 'steps': 45544, 'loss/train': 7.775964736938477} 02/26/2022 12:34:24 - INFO - codeparrot_training - Step 45545: {'lr': 1.0552214255829224e-05, 'samples': 23319552, 'steps': 45545, 'loss/train': 2.3264169692993164} 02/26/2022 12:34:27 - INFO - codeparrot_training - Step 45546: {'lr': 1.0547511133622428e-05, 'samples': 23320064, 'steps': 45546, 'loss/train': 1.6504597663879395} 02/26/2022 12:34:33 - INFO - codeparrot_training - Step 45547: {'lr': 1.0542809037153989e-05, 'samples': 23320576, 'steps': 45547, 'loss/train': 1.3006246089935303} 02/26/2022 12:34:36 - INFO - codeparrot_training - Step 45548: {'lr': 1.0538107966444138e-05, 'samples': 23321088, 'steps': 45548, 'loss/train': 2.3742191791534424} 02/26/2022 12:34:42 - INFO - codeparrot_training - Step 45549: {'lr': 1.0533407921513e-05, 'samples': 23321600, 'steps': 45549, 'loss/train': 1.9453610181808472} 02/26/2022 12:34:45 - INFO - codeparrot_training - Step 45550: {'lr': 1.0528708902380696e-05, 'samples': 23322112, 'steps': 45550, 'loss/train': 1.8022223711013794} 02/26/2022 12:34:51 - INFO - codeparrot_training - Step 45551: {'lr': 1.0524010909067377e-05, 'samples': 23322624, 'steps': 45551, 'loss/train': 0.34935805201530457} 02/26/2022 12:34:54 - INFO - codeparrot_training - Step 45552: {'lr': 1.0519313941593139e-05, 'samples': 23323136, 'steps': 45552, 'loss/train': 1.9266070127487183} 02/26/2022 12:35:01 - INFO - codeparrot_training - Step 45553: {'lr': 1.0514617999978104e-05, 'samples': 23323648, 'steps': 45553, 'loss/train': 1.1075153350830078} 02/26/2022 12:35:05 - INFO - codeparrot_training - Step 45554: {'lr': 1.0509923084242423e-05, 'samples': 23324160, 'steps': 45554, 'loss/train': 2.091653823852539} 02/26/2022 12:35:10 - INFO - codeparrot_training - Step 45555: {'lr': 1.0505229194406191e-05, 'samples': 23324672, 'steps': 45555, 'loss/train': 1.7227206230163574} 02/26/2022 12:35:14 - INFO - codeparrot_training - Step 45556: {'lr': 1.0500536330489502e-05, 'samples': 23325184, 'steps': 45556, 'loss/train': 1.3503150939941406} 02/26/2022 12:35:19 - INFO - codeparrot_training - Step 45557: {'lr': 1.0495844492512451e-05, 'samples': 23325696, 'steps': 45557, 'loss/train': 2.2403905391693115} 02/26/2022 12:35:23 - INFO - codeparrot_training - Step 45558: {'lr': 1.0491153680495164e-05, 'samples': 23326208, 'steps': 45558, 'loss/train': 1.721691608428955} 02/26/2022 12:35:28 - INFO - codeparrot_training - Step 45559: {'lr': 1.0486463894457732e-05, 'samples': 23326720, 'steps': 45559, 'loss/train': 2.797123432159424} 02/26/2022 12:35:32 - INFO - codeparrot_training - Step 45560: {'lr': 1.0481775134420224e-05, 'samples': 23327232, 'steps': 45560, 'loss/train': 0.14029265940189362} 02/26/2022 12:35:37 - INFO - codeparrot_training - Step 45561: {'lr': 1.0477087400402707e-05, 'samples': 23327744, 'steps': 45561, 'loss/train': 2.30580997467041} 02/26/2022 12:35:41 - INFO - codeparrot_training - Step 45562: {'lr': 1.0472400692425333e-05, 'samples': 23328256, 'steps': 45562, 'loss/train': 2.5527184009552} 02/26/2022 12:35:46 - INFO - codeparrot_training - Step 45563: {'lr': 1.0467715010508139e-05, 'samples': 23328768, 'steps': 45563, 'loss/train': 1.0978443622589111} 02/26/2022 12:35:50 - INFO - codeparrot_training - Step 45564: {'lr': 1.0463030354671165e-05, 'samples': 23329280, 'steps': 45564, 'loss/train': 2.0104455947875977} 02/26/2022 12:35:56 - INFO - codeparrot_training - Step 45565: {'lr': 1.0458346724934509e-05, 'samples': 23329792, 'steps': 45565, 'loss/train': 1.9947469234466553} 02/26/2022 12:36:00 - INFO - codeparrot_training - Step 45566: {'lr': 1.0453664121318235e-05, 'samples': 23330304, 'steps': 45566, 'loss/train': 0.20785105228424072} 02/26/2022 12:36:05 - INFO - codeparrot_training - Step 45567: {'lr': 1.0448982543842412e-05, 'samples': 23330816, 'steps': 45567, 'loss/train': 1.6825751066207886} 02/26/2022 12:36:09 - INFO - codeparrot_training - Step 45568: {'lr': 1.0444301992527077e-05, 'samples': 23331328, 'steps': 45568, 'loss/train': 2.8616325855255127} 02/26/2022 12:36:14 - INFO - codeparrot_training - Step 45569: {'lr': 1.0439622467392302e-05, 'samples': 23331840, 'steps': 45569, 'loss/train': 2.049180030822754} 02/26/2022 12:36:18 - INFO - codeparrot_training - Step 45570: {'lr': 1.0434943968458066e-05, 'samples': 23332352, 'steps': 45570, 'loss/train': 1.535075306892395} 02/26/2022 12:36:23 - INFO - codeparrot_training - Step 45571: {'lr': 1.0430266495744495e-05, 'samples': 23332864, 'steps': 45571, 'loss/train': 1.9570480585098267} 02/26/2022 12:36:27 - INFO - codeparrot_training - Step 45572: {'lr': 1.04255900492716e-05, 'samples': 23333376, 'steps': 45572, 'loss/train': 0.7161442637443542} 02/26/2022 12:36:32 - INFO - codeparrot_training - Step 45573: {'lr': 1.0420914629059391e-05, 'samples': 23333888, 'steps': 45573, 'loss/train': 2.4145686626434326} 02/26/2022 12:36:36 - INFO - codeparrot_training - Step 45574: {'lr': 1.0416240235127882e-05, 'samples': 23334400, 'steps': 45574, 'loss/train': 1.8772698640823364} 02/26/2022 12:36:41 - INFO - codeparrot_training - Step 45575: {'lr': 1.0411566867497169e-05, 'samples': 23334912, 'steps': 45575, 'loss/train': 1.1102863550186157} 02/26/2022 12:36:45 - INFO - codeparrot_training - Step 45576: {'lr': 1.0406894526187177e-05, 'samples': 23335424, 'steps': 45576, 'loss/train': 1.2965295314788818} 02/26/2022 12:36:50 - INFO - codeparrot_training - Step 45577: {'lr': 1.0402223211218031e-05, 'samples': 23335936, 'steps': 45577, 'loss/train': 0.15410499274730682} 02/26/2022 12:36:54 - INFO - codeparrot_training - Step 45578: {'lr': 1.0397552922609632e-05, 'samples': 23336448, 'steps': 45578, 'loss/train': 1.8596950769424438} 02/26/2022 12:36:59 - INFO - codeparrot_training - Step 45579: {'lr': 1.0392883660382046e-05, 'samples': 23336960, 'steps': 45579, 'loss/train': 1.4166401624679565} 02/26/2022 12:37:03 - INFO - codeparrot_training - Step 45580: {'lr': 1.038821542455523e-05, 'samples': 23337472, 'steps': 45580, 'loss/train': 1.94240140914917} 02/26/2022 12:37:10 - INFO - codeparrot_training - Step 45581: {'lr': 1.0383548215149307e-05, 'samples': 23337984, 'steps': 45581, 'loss/train': 2.0994479656219482} 02/26/2022 12:37:13 - INFO - codeparrot_training - Step 45582: {'lr': 1.0378882032184095e-05, 'samples': 23338496, 'steps': 45582, 'loss/train': 3.729790210723877} 02/26/2022 12:37:19 - INFO - codeparrot_training - Step 45583: {'lr': 1.0374216875679688e-05, 'samples': 23339008, 'steps': 45583, 'loss/train': 1.7085816860198975} 02/26/2022 12:37:22 - INFO - codeparrot_training - Step 45584: {'lr': 1.0369552745656014e-05, 'samples': 23339520, 'steps': 45584, 'loss/train': 2.3823957443237305} 02/26/2022 12:37:28 - INFO - codeparrot_training - Step 45585: {'lr': 1.0364889642133141e-05, 'samples': 23340032, 'steps': 45585, 'loss/train': 2.9953062534332275} 02/26/2022 12:37:31 - INFO - codeparrot_training - Step 45586: {'lr': 1.036022756513097e-05, 'samples': 23340544, 'steps': 45586, 'loss/train': 1.8608744144439697} 02/26/2022 12:37:37 - INFO - codeparrot_training - Step 45587: {'lr': 1.0355566514669484e-05, 'samples': 23341056, 'steps': 45587, 'loss/train': 1.9176645278930664} 02/26/2022 12:37:40 - INFO - codeparrot_training - Step 45588: {'lr': 1.035090649076864e-05, 'samples': 23341568, 'steps': 45588, 'loss/train': 1.3425835371017456} 02/26/2022 12:37:46 - INFO - codeparrot_training - Step 45589: {'lr': 1.0346247493448425e-05, 'samples': 23342080, 'steps': 45589, 'loss/train': 0.6975095868110657} 02/26/2022 12:37:50 - INFO - codeparrot_training - Step 45590: {'lr': 1.034158952272879e-05, 'samples': 23342592, 'steps': 45590, 'loss/train': 2.5314528942108154} 02/26/2022 12:37:55 - INFO - codeparrot_training - Step 45591: {'lr': 1.0336932578629692e-05, 'samples': 23343104, 'steps': 45591, 'loss/train': 1.6060001850128174} 02/26/2022 12:37:59 - INFO - codeparrot_training - Step 45592: {'lr': 1.0332276661171064e-05, 'samples': 23343616, 'steps': 45592, 'loss/train': 3.0115532875061035} 02/26/2022 12:38:05 - INFO - codeparrot_training - Step 45593: {'lr': 1.0327621770372859e-05, 'samples': 23344128, 'steps': 45593, 'loss/train': 0.7314513325691223} 02/26/2022 12:38:08 - INFO - codeparrot_training - Step 45594: {'lr': 1.0322967906255032e-05, 'samples': 23344640, 'steps': 45594, 'loss/train': 1.1862751245498657} 02/26/2022 12:38:12 - INFO - codeparrot_training - Step 45595: {'lr': 1.0318315068837514e-05, 'samples': 23345152, 'steps': 45595, 'loss/train': 3.2045083045959473} 02/26/2022 12:38:18 - INFO - codeparrot_training - Step 45596: {'lr': 1.0313663258140177e-05, 'samples': 23345664, 'steps': 45596, 'loss/train': 1.443165898323059} 02/26/2022 12:38:21 - INFO - codeparrot_training - Step 45597: {'lr': 1.0309012474183032e-05, 'samples': 23346176, 'steps': 45597, 'loss/train': 1.5422154664993286} 02/26/2022 12:38:27 - INFO - codeparrot_training - Step 45598: {'lr': 1.0304362716985955e-05, 'samples': 23346688, 'steps': 45598, 'loss/train': 0.7026045918464661} 02/26/2022 12:38:30 - INFO - codeparrot_training - Step 45599: {'lr': 1.0299713986568926e-05, 'samples': 23347200, 'steps': 45599, 'loss/train': 2.0096731185913086} 02/26/2022 12:38:37 - INFO - codeparrot_training - Step 45600: {'lr': 1.0295066282951737e-05, 'samples': 23347712, 'steps': 45600, 'loss/train': 1.1956053972244263} 02/26/2022 12:38:40 - INFO - codeparrot_training - Step 45601: {'lr': 1.02904196061544e-05, 'samples': 23348224, 'steps': 45601, 'loss/train': 1.2708628177642822} 02/26/2022 12:38:46 - INFO - codeparrot_training - Step 45602: {'lr': 1.0285773956196814e-05, 'samples': 23348736, 'steps': 45602, 'loss/train': 1.9006245136260986} 02/26/2022 12:38:49 - INFO - codeparrot_training - Step 45603: {'lr': 1.0281129333098827e-05, 'samples': 23349248, 'steps': 45603, 'loss/train': 0.837859570980072} 02/26/2022 12:38:55 - INFO - codeparrot_training - Step 45604: {'lr': 1.0276485736880337e-05, 'samples': 23349760, 'steps': 45604, 'loss/train': 1.5647293329238892} 02/26/2022 12:38:58 - INFO - codeparrot_training - Step 45605: {'lr': 1.02718431675613e-05, 'samples': 23350272, 'steps': 45605, 'loss/train': 1.5073418617248535} 02/26/2022 12:39:04 - INFO - codeparrot_training - Step 45606: {'lr': 1.0267201625161566e-05, 'samples': 23350784, 'steps': 45606, 'loss/train': 1.3233749866485596} 02/26/2022 12:39:07 - INFO - codeparrot_training - Step 45607: {'lr': 1.0262561109701001e-05, 'samples': 23351296, 'steps': 45607, 'loss/train': 1.6218550205230713} 02/26/2022 12:39:13 - INFO - codeparrot_training - Step 45608: {'lr': 1.0257921621199484e-05, 'samples': 23351808, 'steps': 45608, 'loss/train': 1.9633405208587646} 02/26/2022 12:39:16 - INFO - codeparrot_training - Step 45609: {'lr': 1.0253283159676941e-05, 'samples': 23352320, 'steps': 45609, 'loss/train': 1.0450513362884521} 02/26/2022 12:39:22 - INFO - codeparrot_training - Step 45610: {'lr': 1.0248645725153189e-05, 'samples': 23352832, 'steps': 45610, 'loss/train': 1.9641263484954834} 02/26/2022 12:39:25 - INFO - codeparrot_training - Step 45611: {'lr': 1.0244009317648074e-05, 'samples': 23353344, 'steps': 45611, 'loss/train': 2.4277329444885254} 02/26/2022 12:39:32 - INFO - codeparrot_training - Step 45612: {'lr': 1.0239373937181551e-05, 'samples': 23353856, 'steps': 45612, 'loss/train': 1.9305319786071777} 02/26/2022 12:39:35 - INFO - codeparrot_training - Step 45613: {'lr': 1.0234739583773383e-05, 'samples': 23354368, 'steps': 45613, 'loss/train': 1.3441033363342285} 02/26/2022 12:39:41 - INFO - codeparrot_training - Step 45614: {'lr': 1.0230106257443472e-05, 'samples': 23354880, 'steps': 45614, 'loss/train': 1.438027024269104} 02/26/2022 12:39:46 - INFO - codeparrot_training - Step 45615: {'lr': 1.0225473958211634e-05, 'samples': 23355392, 'steps': 45615, 'loss/train': 1.5967894792556763} 02/26/2022 12:39:50 - INFO - codeparrot_training - Step 45616: {'lr': 1.0220842686097741e-05, 'samples': 23355904, 'steps': 45616, 'loss/train': 1.7690527439117432} 02/26/2022 12:39:55 - INFO - codeparrot_training - Step 45617: {'lr': 1.0216212441121615e-05, 'samples': 23356416, 'steps': 45617, 'loss/train': 0.6522355675697327} 02/26/2022 12:39:59 - INFO - codeparrot_training - Step 45618: {'lr': 1.0211583223303095e-05, 'samples': 23356928, 'steps': 45618, 'loss/train': 2.0137455463409424} 02/26/2022 12:40:04 - INFO - codeparrot_training - Step 45619: {'lr': 1.0206955032662001e-05, 'samples': 23357440, 'steps': 45619, 'loss/train': 0.7497518062591553} 02/26/2022 12:40:08 - INFO - codeparrot_training - Step 45620: {'lr': 1.0202327869218208e-05, 'samples': 23357952, 'steps': 45620, 'loss/train': 1.5751317739486694} 02/26/2022 12:40:13 - INFO - codeparrot_training - Step 45621: {'lr': 1.0197701732991448e-05, 'samples': 23358464, 'steps': 45621, 'loss/train': 1.2662396430969238} 02/26/2022 12:40:17 - INFO - codeparrot_training - Step 45622: {'lr': 1.0193076624001624e-05, 'samples': 23358976, 'steps': 45622, 'loss/train': 1.7119075059890747} 02/26/2022 12:40:22 - INFO - codeparrot_training - Step 45623: {'lr': 1.0188452542268468e-05, 'samples': 23359488, 'steps': 45623, 'loss/train': 1.2963422536849976} 02/26/2022 12:40:26 - INFO - codeparrot_training - Step 45624: {'lr': 1.018382948781188e-05, 'samples': 23360000, 'steps': 45624, 'loss/train': 1.2522579431533813} 02/26/2022 12:40:32 - INFO - codeparrot_training - Step 45625: {'lr': 1.0179207460651596e-05, 'samples': 23360512, 'steps': 45625, 'loss/train': 1.1295278072357178} 02/26/2022 12:40:35 - INFO - codeparrot_training - Step 45626: {'lr': 1.017458646080746e-05, 'samples': 23361024, 'steps': 45626, 'loss/train': 1.0604910850524902} 02/26/2022 12:40:41 - INFO - codeparrot_training - Step 45627: {'lr': 1.0169966488299181e-05, 'samples': 23361536, 'steps': 45627, 'loss/train': 2.5549004077911377} 02/26/2022 12:40:44 - INFO - codeparrot_training - Step 45628: {'lr': 1.0165347543146685e-05, 'samples': 23362048, 'steps': 45628, 'loss/train': 1.8073127269744873} 02/26/2022 12:40:50 - INFO - codeparrot_training - Step 45629: {'lr': 1.0160729625369653e-05, 'samples': 23362560, 'steps': 45629, 'loss/train': 0.17217500507831573} 02/26/2022 12:40:53 - INFO - codeparrot_training - Step 45630: {'lr': 1.0156112734987899e-05, 'samples': 23363072, 'steps': 45630, 'loss/train': 0.9125199317932129} 02/26/2022 12:40:59 - INFO - codeparrot_training - Step 45631: {'lr': 1.015149687202116e-05, 'samples': 23363584, 'steps': 45631, 'loss/train': 2.1482908725738525} 02/26/2022 12:41:02 - INFO - codeparrot_training - Step 45632: {'lr': 1.0146882036489307e-05, 'samples': 23364096, 'steps': 45632, 'loss/train': 1.8630468845367432} 02/26/2022 12:41:08 - INFO - codeparrot_training - Step 45633: {'lr': 1.0142268228412021e-05, 'samples': 23364608, 'steps': 45633, 'loss/train': 1.496239185333252} 02/26/2022 12:41:12 - INFO - codeparrot_training - Step 45634: {'lr': 1.0137655447809119e-05, 'samples': 23365120, 'steps': 45634, 'loss/train': 1.337036371231079} 02/26/2022 12:41:15 - INFO - codeparrot_training - Step 45635: {'lr': 1.0133043694700277e-05, 'samples': 23365632, 'steps': 45635, 'loss/train': 2.2077653408050537} 02/26/2022 12:41:21 - INFO - codeparrot_training - Step 45636: {'lr': 1.0128432969105373e-05, 'samples': 23366144, 'steps': 45636, 'loss/train': 1.3853659629821777} 02/26/2022 12:41:24 - INFO - codeparrot_training - Step 45637: {'lr': 1.0123823271044025e-05, 'samples': 23366656, 'steps': 45637, 'loss/train': 1.3224844932556152} 02/26/2022 12:41:31 - INFO - codeparrot_training - Step 45638: {'lr': 1.0119214600536135e-05, 'samples': 23367168, 'steps': 45638, 'loss/train': 1.3349225521087646} 02/26/2022 12:41:34 - INFO - codeparrot_training - Step 45639: {'lr': 1.0114606957601302e-05, 'samples': 23367680, 'steps': 45639, 'loss/train': 1.48549222946167} 02/26/2022 12:41:40 - INFO - codeparrot_training - Step 45640: {'lr': 1.011000034225934e-05, 'samples': 23368192, 'steps': 45640, 'loss/train': 2.0417563915252686} 02/26/2022 12:41:43 - INFO - codeparrot_training - Step 45641: {'lr': 1.010539475452993e-05, 'samples': 23368704, 'steps': 45641, 'loss/train': 0.6281805634498596} 02/26/2022 12:41:49 - INFO - codeparrot_training - Step 45642: {'lr': 1.0100790194432912e-05, 'samples': 23369216, 'steps': 45642, 'loss/train': 1.8049238920211792} 02/26/2022 12:41:52 - INFO - codeparrot_training - Step 45643: {'lr': 1.0096186661987861e-05, 'samples': 23369728, 'steps': 45643, 'loss/train': 0.6389544010162354} 02/26/2022 12:41:58 - INFO - codeparrot_training - Step 45644: {'lr': 1.009158415721459e-05, 'samples': 23370240, 'steps': 45644, 'loss/train': 1.5484758615493774} 02/26/2022 12:42:01 - INFO - codeparrot_training - Step 45645: {'lr': 1.0086982680132778e-05, 'samples': 23370752, 'steps': 45645, 'loss/train': 3.094593048095703} 02/26/2022 12:42:07 - INFO - codeparrot_training - Step 45646: {'lr': 1.008238223076216e-05, 'samples': 23371264, 'steps': 45646, 'loss/train': 2.4040956497192383} 02/26/2022 12:42:11 - INFO - codeparrot_training - Step 45647: {'lr': 1.0077782809122471e-05, 'samples': 23371776, 'steps': 45647, 'loss/train': 1.758974313735962} 02/26/2022 12:42:17 - INFO - codeparrot_training - Step 45648: {'lr': 1.0073184415233333e-05, 'samples': 23372288, 'steps': 45648, 'loss/train': 2.1329472064971924} 02/26/2022 12:42:21 - INFO - codeparrot_training - Step 45649: {'lr': 1.0068587049114507e-05, 'samples': 23372800, 'steps': 45649, 'loss/train': 0.2544361650943756} 02/26/2022 12:42:26 - INFO - codeparrot_training - Step 45650: {'lr': 1.0063990710785648e-05, 'samples': 23373312, 'steps': 45650, 'loss/train': 1.8752689361572266} 02/26/2022 12:42:30 - INFO - codeparrot_training - Step 45651: {'lr': 1.0059395400266485e-05, 'samples': 23373824, 'steps': 45651, 'loss/train': 2.842592477798462} 02/26/2022 12:42:35 - INFO - codeparrot_training - Step 45652: {'lr': 1.0054801117576672e-05, 'samples': 23374336, 'steps': 45652, 'loss/train': 1.9933407306671143} 02/26/2022 12:42:39 - INFO - codeparrot_training - Step 45653: {'lr': 1.0050207862735916e-05, 'samples': 23374848, 'steps': 45653, 'loss/train': 1.311305284500122} 02/26/2022 12:42:44 - INFO - codeparrot_training - Step 45654: {'lr': 1.0045615635763838e-05, 'samples': 23375360, 'steps': 45654, 'loss/train': 1.054903507232666} 02/26/2022 12:42:48 - INFO - codeparrot_training - Step 45655: {'lr': 1.0041024436680174e-05, 'samples': 23375872, 'steps': 45655, 'loss/train': 0.29673346877098083} 02/26/2022 12:42:53 - INFO - codeparrot_training - Step 45656: {'lr': 1.0036434265504574e-05, 'samples': 23376384, 'steps': 45656, 'loss/train': 2.4695589542388916} 02/26/2022 12:42:57 - INFO - codeparrot_training - Step 45657: {'lr': 1.0031845122256716e-05, 'samples': 23376896, 'steps': 45657, 'loss/train': 1.6470028162002563} 02/26/2022 12:43:02 - INFO - codeparrot_training - Step 45658: {'lr': 1.002725700695617e-05, 'samples': 23377408, 'steps': 45658, 'loss/train': 0.3379676043987274} 02/26/2022 12:43:06 - INFO - codeparrot_training - Step 45659: {'lr': 1.0022669919622695e-05, 'samples': 23377920, 'steps': 45659, 'loss/train': 2.492748498916626} 02/26/2022 12:43:11 - INFO - codeparrot_training - Step 45660: {'lr': 1.0018083860275889e-05, 'samples': 23378432, 'steps': 45660, 'loss/train': 1.9696780443191528} 02/26/2022 12:43:15 - INFO - codeparrot_training - Step 45661: {'lr': 1.001349882893543e-05, 'samples': 23378944, 'steps': 45661, 'loss/train': 1.0503180027008057} 02/26/2022 12:43:21 - INFO - codeparrot_training - Step 45662: {'lr': 1.0008914825620913e-05, 'samples': 23379456, 'steps': 45662, 'loss/train': 0.06641270220279694} 02/26/2022 12:43:24 - INFO - codeparrot_training - Step 45663: {'lr': 1.0004331850352017e-05, 'samples': 23379968, 'steps': 45663, 'loss/train': 2.2936954498291016} 02/26/2022 12:43:30 - INFO - codeparrot_training - Step 45664: {'lr': 9.999749903148337e-06, 'samples': 23380480, 'steps': 45664, 'loss/train': 0.952839732170105} 02/26/2022 12:43:33 - INFO - codeparrot_training - Step 45665: {'lr': 9.995168984029552e-06, 'samples': 23380992, 'steps': 45665, 'loss/train': 1.309648036956787} 02/26/2022 12:43:39 - INFO - codeparrot_training - Step 45666: {'lr': 9.9905890930152e-06, 'samples': 23381504, 'steps': 45666, 'loss/train': 2.489650011062622} 02/26/2022 12:43:42 - INFO - codeparrot_training - Step 45667: {'lr': 9.986010230124992e-06, 'samples': 23382016, 'steps': 45667, 'loss/train': 2.2035741806030273} 02/26/2022 12:43:48 - INFO - codeparrot_training - Step 45668: {'lr': 9.981432395378493e-06, 'samples': 23382528, 'steps': 45668, 'loss/train': 1.8708242177963257} 02/26/2022 12:43:51 - INFO - codeparrot_training - Step 45669: {'lr': 9.976855588795352e-06, 'samples': 23383040, 'steps': 45669, 'loss/train': 1.7897173166275024} 02/26/2022 12:43:57 - INFO - codeparrot_training - Step 45670: {'lr': 9.972279810395085e-06, 'samples': 23383552, 'steps': 45670, 'loss/train': 1.3311960697174072} 02/26/2022 12:44:00 - INFO - codeparrot_training - Step 45671: {'lr': 9.967705060197396e-06, 'samples': 23384064, 'steps': 45671, 'loss/train': 2.46360445022583} 02/26/2022 12:44:06 - INFO - codeparrot_training - Step 45672: {'lr': 9.963131338221798e-06, 'samples': 23384576, 'steps': 45672, 'loss/train': 1.1261680126190186} 02/26/2022 12:44:09 - INFO - codeparrot_training - Step 45673: {'lr': 9.958558644487998e-06, 'samples': 23385088, 'steps': 45673, 'loss/train': 1.9089025259017944} 02/26/2022 12:44:15 - INFO - codeparrot_training - Step 45674: {'lr': 9.953986979015451e-06, 'samples': 23385600, 'steps': 45674, 'loss/train': 1.5332258939743042} 02/26/2022 12:44:19 - INFO - codeparrot_training - Step 45675: {'lr': 9.949416341823836e-06, 'samples': 23386112, 'steps': 45675, 'loss/train': 1.4545354843139648} 02/26/2022 12:44:24 - INFO - codeparrot_training - Step 45676: {'lr': 9.944846732932667e-06, 'samples': 23386624, 'steps': 45676, 'loss/train': 2.3801071643829346} 02/26/2022 12:44:28 - INFO - codeparrot_training - Step 45677: {'lr': 9.940278152361593e-06, 'samples': 23387136, 'steps': 45677, 'loss/train': 2.2263152599334717} 02/26/2022 12:44:33 - INFO - codeparrot_training - Step 45678: {'lr': 9.935710600130071e-06, 'samples': 23387648, 'steps': 45678, 'loss/train': 1.2591277360916138} 02/26/2022 12:44:36 - INFO - codeparrot_training - Step 45679: {'lr': 9.93114407625778e-06, 'samples': 23388160, 'steps': 45679, 'loss/train': 1.8275260925292969} 02/26/2022 12:44:42 - INFO - codeparrot_training - Step 45680: {'lr': 9.926578580764234e-06, 'samples': 23388672, 'steps': 45680, 'loss/train': 0.9706999659538269} 02/26/2022 12:44:46 - INFO - codeparrot_training - Step 45681: {'lr': 9.922014113669025e-06, 'samples': 23389184, 'steps': 45681, 'loss/train': 2.613708019256592} 02/26/2022 12:44:51 - INFO - codeparrot_training - Step 45682: {'lr': 9.917450674991612e-06, 'samples': 23389696, 'steps': 45682, 'loss/train': 1.945132851600647} 02/26/2022 12:44:54 - INFO - codeparrot_training - Step 45683: {'lr': 9.912888264751645e-06, 'samples': 23390208, 'steps': 45683, 'loss/train': 1.57122802734375} 02/26/2022 12:45:01 - INFO - codeparrot_training - Step 45684: {'lr': 9.908326882968582e-06, 'samples': 23390720, 'steps': 45684, 'loss/train': 1.5135960578918457} 02/26/2022 12:45:04 - INFO - codeparrot_training - Step 45685: {'lr': 9.903766529662045e-06, 'samples': 23391232, 'steps': 45685, 'loss/train': 2.096989870071411} 02/26/2022 12:45:10 - INFO - codeparrot_training - Step 45686: {'lr': 9.899207204851546e-06, 'samples': 23391744, 'steps': 45686, 'loss/train': 2.0623619556427} 02/26/2022 12:45:13 - INFO - codeparrot_training - Step 45687: {'lr': 9.894648908556597e-06, 'samples': 23392256, 'steps': 45687, 'loss/train': 1.8422881364822388} 02/26/2022 12:45:19 - INFO - codeparrot_training - Step 45688: {'lr': 9.89009164079671e-06, 'samples': 23392768, 'steps': 45688, 'loss/train': 0.7527719736099243} 02/26/2022 12:45:22 - INFO - codeparrot_training - Step 45689: {'lr': 9.885535401591456e-06, 'samples': 23393280, 'steps': 45689, 'loss/train': 1.548490047454834} 02/26/2022 12:45:28 - INFO - codeparrot_training - Step 45690: {'lr': 9.880980190960315e-06, 'samples': 23393792, 'steps': 45690, 'loss/train': 1.3348819017410278} 02/26/2022 12:45:31 - INFO - codeparrot_training - Step 45691: {'lr': 9.876426008922801e-06, 'samples': 23394304, 'steps': 45691, 'loss/train': 2.257467746734619} 02/26/2022 12:45:37 - INFO - codeparrot_training - Step 45692: {'lr': 9.871872855498399e-06, 'samples': 23394816, 'steps': 45692, 'loss/train': 1.1500526666641235} 02/26/2022 12:45:40 - INFO - codeparrot_training - Step 45693: {'lr': 9.867320730706674e-06, 'samples': 23395328, 'steps': 45693, 'loss/train': 1.1437115669250488} 02/26/2022 12:45:47 - INFO - codeparrot_training - Step 45694: {'lr': 9.862769634567114e-06, 'samples': 23395840, 'steps': 45694, 'loss/train': 2.3052122592926025} 02/26/2022 12:45:50 - INFO - codeparrot_training - Step 45695: {'lr': 9.858219567099175e-06, 'samples': 23396352, 'steps': 45695, 'loss/train': 1.3616516590118408} 02/26/2022 12:45:56 - INFO - codeparrot_training - Step 45696: {'lr': 9.853670528322367e-06, 'samples': 23396864, 'steps': 45696, 'loss/train': 2.3466379642486572} 02/26/2022 12:45:59 - INFO - codeparrot_training - Step 45697: {'lr': 9.849122518256205e-06, 'samples': 23397376, 'steps': 45697, 'loss/train': 2.5476202964782715} 02/26/2022 12:46:05 - INFO - codeparrot_training - Step 45698: {'lr': 9.844575536920113e-06, 'samples': 23397888, 'steps': 45698, 'loss/train': 1.311321496963501} 02/26/2022 12:46:08 - INFO - codeparrot_training - Step 45699: {'lr': 9.840029584333637e-06, 'samples': 23398400, 'steps': 45699, 'loss/train': 1.832392930984497} 02/26/2022 12:46:14 - INFO - codeparrot_training - Step 45700: {'lr': 9.835484660516203e-06, 'samples': 23398912, 'steps': 45700, 'loss/train': 1.4693660736083984} 02/26/2022 12:46:17 - INFO - codeparrot_training - Step 45701: {'lr': 9.83094076548724e-06, 'samples': 23399424, 'steps': 45701, 'loss/train': 0.95565265417099} 02/26/2022 12:46:23 - INFO - codeparrot_training - Step 45702: {'lr': 9.826397899266315e-06, 'samples': 23399936, 'steps': 45702, 'loss/train': 1.0762507915496826} 02/26/2022 12:46:26 - INFO - codeparrot_training - Step 45703: {'lr': 9.821856061872858e-06, 'samples': 23400448, 'steps': 45703, 'loss/train': 2.089848041534424} 02/26/2022 12:46:32 - INFO - codeparrot_training - Step 45704: {'lr': 9.81731525332627e-06, 'samples': 23400960, 'steps': 45704, 'loss/train': 1.6372756958007812} 02/26/2022 12:46:35 - INFO - codeparrot_training - Step 45705: {'lr': 9.812775473646035e-06, 'samples': 23401472, 'steps': 45705, 'loss/train': 1.3960237503051758} 02/26/2022 12:46:41 - INFO - codeparrot_training - Step 45706: {'lr': 9.80823672285161e-06, 'samples': 23401984, 'steps': 45706, 'loss/train': 1.920245885848999} 02/26/2022 12:46:44 - INFO - codeparrot_training - Step 45707: {'lr': 9.803699000962397e-06, 'samples': 23402496, 'steps': 45707, 'loss/train': 1.6112788915634155} 02/26/2022 12:46:51 - INFO - codeparrot_training - Step 45708: {'lr': 9.799162307997933e-06, 'samples': 23403008, 'steps': 45708, 'loss/train': 2.238391399383545} 02/26/2022 12:46:54 - INFO - codeparrot_training - Step 45709: {'lr': 9.794626643977539e-06, 'samples': 23403520, 'steps': 45709, 'loss/train': 1.5660067796707153} 02/26/2022 12:47:00 - INFO - codeparrot_training - Step 45710: {'lr': 9.790092008920698e-06, 'samples': 23404032, 'steps': 45710, 'loss/train': 1.097868800163269} 02/26/2022 12:47:03 - INFO - codeparrot_training - Step 45711: {'lr': 9.785558402846811e-06, 'samples': 23404544, 'steps': 45711, 'loss/train': 2.1826376914978027} 02/26/2022 12:47:09 - INFO - codeparrot_training - Step 45712: {'lr': 9.781025825775392e-06, 'samples': 23405056, 'steps': 45712, 'loss/train': 0.9298344850540161} 02/26/2022 12:47:12 - INFO - codeparrot_training - Step 45713: {'lr': 9.7764942777257e-06, 'samples': 23405568, 'steps': 45713, 'loss/train': 1.4729042053222656} 02/26/2022 12:47:18 - INFO - codeparrot_training - Step 45714: {'lr': 9.771963758717251e-06, 'samples': 23406080, 'steps': 45714, 'loss/train': 1.8881902694702148} 02/26/2022 12:47:21 - INFO - codeparrot_training - Step 45715: {'lr': 9.767434268769414e-06, 'samples': 23406592, 'steps': 45715, 'loss/train': 1.3229073286056519} 02/26/2022 12:47:27 - INFO - codeparrot_training - Step 45716: {'lr': 9.762905807901651e-06, 'samples': 23407104, 'steps': 45716, 'loss/train': 1.5853618383407593} 02/26/2022 12:47:30 - INFO - codeparrot_training - Step 45717: {'lr': 9.758378376133275e-06, 'samples': 23407616, 'steps': 45717, 'loss/train': 2.241222381591797} 02/26/2022 12:47:36 - INFO - codeparrot_training - Step 45718: {'lr': 9.753851973483746e-06, 'samples': 23408128, 'steps': 45718, 'loss/train': 0.7458863258361816} 02/26/2022 12:47:39 - INFO - codeparrot_training - Step 45719: {'lr': 9.74932659997238e-06, 'samples': 23408640, 'steps': 45719, 'loss/train': 1.5619745254516602} 02/26/2022 12:47:45 - INFO - codeparrot_training - Step 45720: {'lr': 9.744802255618662e-06, 'samples': 23409152, 'steps': 45720, 'loss/train': 1.2937605381011963} 02/26/2022 12:47:49 - INFO - codeparrot_training - Step 45721: {'lr': 9.74027894044191e-06, 'samples': 23409664, 'steps': 45721, 'loss/train': 1.4370262622833252} 02/26/2022 12:47:54 - INFO - codeparrot_training - Step 45722: {'lr': 9.735756654461498e-06, 'samples': 23410176, 'steps': 45722, 'loss/train': 2.0643310546875} 02/26/2022 12:47:57 - INFO - codeparrot_training - Step 45723: {'lr': 9.731235397696797e-06, 'samples': 23410688, 'steps': 45723, 'loss/train': 1.572877287864685} 02/26/2022 12:48:03 - INFO - codeparrot_training - Step 45724: {'lr': 9.72671517016724e-06, 'samples': 23411200, 'steps': 45724, 'loss/train': 2.330118417739868} 02/26/2022 12:48:07 - INFO - codeparrot_training - Step 45725: {'lr': 9.722195971892112e-06, 'samples': 23411712, 'steps': 45725, 'loss/train': 0.7508022785186768} 02/26/2022 12:48:12 - INFO - codeparrot_training - Step 45726: {'lr': 9.71767780289079e-06, 'samples': 23412224, 'steps': 45726, 'loss/train': 1.4067645072937012} 02/26/2022 12:48:15 - INFO - codeparrot_training - Step 45727: {'lr': 9.713160663182646e-06, 'samples': 23412736, 'steps': 45727, 'loss/train': 1.3673279285430908} 02/26/2022 12:48:21 - INFO - codeparrot_training - Step 45728: {'lr': 9.708644552787028e-06, 'samples': 23413248, 'steps': 45728, 'loss/train': 1.1845699548721313} 02/26/2022 12:48:25 - INFO - codeparrot_training - Step 45729: {'lr': 9.704129471723277e-06, 'samples': 23413760, 'steps': 45729, 'loss/train': 1.5683904886245728} 02/26/2022 12:48:31 - INFO - codeparrot_training - Step 45730: {'lr': 9.699615420010744e-06, 'samples': 23414272, 'steps': 45730, 'loss/train': 1.3461216688156128} 02/26/2022 12:48:34 - INFO - codeparrot_training - Step 45731: {'lr': 9.695102397668715e-06, 'samples': 23414784, 'steps': 45731, 'loss/train': 2.922683000564575} 02/26/2022 12:48:40 - INFO - codeparrot_training - Step 45732: {'lr': 9.690590404716592e-06, 'samples': 23415296, 'steps': 45732, 'loss/train': 1.7593162059783936} 02/26/2022 12:48:43 - INFO - codeparrot_training - Step 45733: {'lr': 9.686079441173668e-06, 'samples': 23415808, 'steps': 45733, 'loss/train': 1.700556755065918} 02/26/2022 12:48:49 - INFO - codeparrot_training - Step 45734: {'lr': 9.681569507059285e-06, 'samples': 23416320, 'steps': 45734, 'loss/train': 1.8904430866241455} 02/26/2022 12:48:52 - INFO - codeparrot_training - Step 45735: {'lr': 9.677060602392707e-06, 'samples': 23416832, 'steps': 45735, 'loss/train': 2.1521337032318115} 02/26/2022 12:48:58 - INFO - codeparrot_training - Step 45736: {'lr': 9.672552727193306e-06, 'samples': 23417344, 'steps': 45736, 'loss/train': 0.817400336265564} 02/26/2022 12:49:01 - INFO - codeparrot_training - Step 45737: {'lr': 9.668045881480403e-06, 'samples': 23417856, 'steps': 45737, 'loss/train': 1.1943998336791992} 02/26/2022 12:49:07 - INFO - codeparrot_training - Step 45738: {'lr': 9.663540065273258e-06, 'samples': 23418368, 'steps': 45738, 'loss/train': 0.8992491364479065} 02/26/2022 12:49:10 - INFO - codeparrot_training - Step 45739: {'lr': 9.659035278591189e-06, 'samples': 23418880, 'steps': 45739, 'loss/train': 2.6421351432800293} 02/26/2022 12:49:16 - INFO - codeparrot_training - Step 45740: {'lr': 9.654531521453513e-06, 'samples': 23419392, 'steps': 45740, 'loss/train': 1.8610260486602783} 02/26/2022 12:49:20 - INFO - codeparrot_training - Step 45741: {'lr': 9.650028793879494e-06, 'samples': 23419904, 'steps': 45741, 'loss/train': 0.9102139472961426} 02/26/2022 12:49:25 - INFO - codeparrot_training - Step 45742: {'lr': 9.645527095888424e-06, 'samples': 23420416, 'steps': 45742, 'loss/train': 2.4732449054718018} 02/26/2022 12:49:29 - INFO - codeparrot_training - Step 45743: {'lr': 9.64102642749959e-06, 'samples': 23420928, 'steps': 45743, 'loss/train': 1.081050992012024} 02/26/2022 12:49:34 - INFO - codeparrot_training - Step 45744: {'lr': 9.636526788732308e-06, 'samples': 23421440, 'steps': 45744, 'loss/train': 3.349553108215332} 02/26/2022 12:49:38 - INFO - codeparrot_training - Step 45745: {'lr': 9.632028179605817e-06, 'samples': 23421952, 'steps': 45745, 'loss/train': 2.085125684738159} 02/26/2022 12:49:44 - INFO - codeparrot_training - Step 45746: {'lr': 9.627530600139351e-06, 'samples': 23422464, 'steps': 45746, 'loss/train': 2.2292213439941406} 02/26/2022 12:49:47 - INFO - codeparrot_training - Step 45747: {'lr': 9.62303405035228e-06, 'samples': 23422976, 'steps': 45747, 'loss/train': 1.5764613151550293} 02/26/2022 12:49:53 - INFO - codeparrot_training - Step 45748: {'lr': 9.618538530263731e-06, 'samples': 23423488, 'steps': 45748, 'loss/train': 1.3521808385849} 02/26/2022 12:49:56 - INFO - codeparrot_training - Step 45749: {'lr': 9.614044039893077e-06, 'samples': 23424000, 'steps': 45749, 'loss/train': 2.0272583961486816} 02/26/2022 12:50:02 - INFO - codeparrot_training - Step 45750: {'lr': 9.609550579259497e-06, 'samples': 23424512, 'steps': 45750, 'loss/train': 2.5384294986724854} 02/26/2022 12:50:05 - INFO - codeparrot_training - Step 45751: {'lr': 9.605058148382335e-06, 'samples': 23425024, 'steps': 45751, 'loss/train': 1.9197872877120972} 02/26/2022 12:50:10 - INFO - codeparrot_training - Step 45752: {'lr': 9.600566747280714e-06, 'samples': 23425536, 'steps': 45752, 'loss/train': 0.6177244186401367} 02/26/2022 12:50:14 - INFO - codeparrot_training - Step 45753: {'lr': 9.596076375973955e-06, 'samples': 23426048, 'steps': 45753, 'loss/train': 2.314452886581421} 02/26/2022 12:50:19 - INFO - codeparrot_training - Step 45754: {'lr': 9.591587034481237e-06, 'samples': 23426560, 'steps': 45754, 'loss/train': 1.8757658004760742} 02/26/2022 12:50:23 - INFO - codeparrot_training - Step 45755: {'lr': 9.587098722821847e-06, 'samples': 23427072, 'steps': 45755, 'loss/train': 1.6897307634353638} 02/26/2022 12:50:29 - INFO - codeparrot_training - Step 45756: {'lr': 9.582611441014965e-06, 'samples': 23427584, 'steps': 45756, 'loss/train': 1.489898920059204} 02/26/2022 12:50:33 - INFO - codeparrot_training - Step 45757: {'lr': 9.578125189079855e-06, 'samples': 23428096, 'steps': 45757, 'loss/train': 2.992708683013916} 02/26/2022 12:50:38 - INFO - codeparrot_training - Step 45758: {'lr': 9.573639967035697e-06, 'samples': 23428608, 'steps': 45758, 'loss/train': 1.3026658296585083} 02/26/2022 12:50:41 - INFO - codeparrot_training - Step 45759: {'lr': 9.56915577490175e-06, 'samples': 23429120, 'steps': 45759, 'loss/train': 2.2049973011016846} 02/26/2022 12:50:47 - INFO - codeparrot_training - Step 45760: {'lr': 9.564672612697168e-06, 'samples': 23429632, 'steps': 45760, 'loss/train': 1.5544084310531616} 02/26/2022 12:50:51 - INFO - codeparrot_training - Step 45761: {'lr': 9.56019048044121e-06, 'samples': 23430144, 'steps': 45761, 'loss/train': 1.6291383504867554} 02/26/2022 12:50:56 - INFO - codeparrot_training - Step 45762: {'lr': 9.555709378153004e-06, 'samples': 23430656, 'steps': 45762, 'loss/train': 1.9854035377502441} 02/26/2022 12:51:00 - INFO - codeparrot_training - Step 45763: {'lr': 9.55122930585181e-06, 'samples': 23431168, 'steps': 45763, 'loss/train': 1.219187617301941} 02/26/2022 12:51:05 - INFO - codeparrot_training - Step 45764: {'lr': 9.546750263556808e-06, 'samples': 23431680, 'steps': 45764, 'loss/train': 1.121265172958374} 02/26/2022 12:51:09 - INFO - codeparrot_training - Step 45765: {'lr': 9.542272251287176e-06, 'samples': 23432192, 'steps': 45765, 'loss/train': 1.8777953386306763} 02/26/2022 12:51:15 - INFO - codeparrot_training - Step 45766: {'lr': 9.537795269062066e-06, 'samples': 23432704, 'steps': 45766, 'loss/train': 0.7203315496444702} 02/26/2022 12:51:18 - INFO - codeparrot_training - Step 45767: {'lr': 9.53331931690074e-06, 'samples': 23433216, 'steps': 45767, 'loss/train': 1.976548433303833} 02/26/2022 12:51:24 - INFO - codeparrot_training - Step 45768: {'lr': 9.52884439482224e-06, 'samples': 23433728, 'steps': 45768, 'loss/train': 2.1859312057495117} 02/26/2022 12:51:27 - INFO - codeparrot_training - Step 45769: {'lr': 9.524370502845908e-06, 'samples': 23434240, 'steps': 45769, 'loss/train': 1.7748059034347534} 02/26/2022 12:51:33 - INFO - codeparrot_training - Step 45770: {'lr': 9.519897640990733e-06, 'samples': 23434752, 'steps': 45770, 'loss/train': 1.8329581022262573} 02/26/2022 12:51:36 - INFO - codeparrot_training - Step 45771: {'lr': 9.515425809276002e-06, 'samples': 23435264, 'steps': 45771, 'loss/train': 0.9880490899085999} 02/26/2022 12:51:42 - INFO - codeparrot_training - Step 45772: {'lr': 9.510955007720784e-06, 'samples': 23435776, 'steps': 45772, 'loss/train': 1.8825429677963257} 02/26/2022 12:51:45 - INFO - codeparrot_training - Step 45773: {'lr': 9.506485236344342e-06, 'samples': 23436288, 'steps': 45773, 'loss/train': 1.7602052688598633} 02/26/2022 12:51:51 - INFO - codeparrot_training - Step 45774: {'lr': 9.502016495165688e-06, 'samples': 23436800, 'steps': 45774, 'loss/train': 2.645533800125122} 02/26/2022 12:51:54 - INFO - codeparrot_training - Step 45775: {'lr': 9.497548784204058e-06, 'samples': 23437312, 'steps': 45775, 'loss/train': 2.595423698425293} 02/26/2022 12:52:01 - INFO - codeparrot_training - Step 45776: {'lr': 9.493082103478518e-06, 'samples': 23437824, 'steps': 45776, 'loss/train': 0.023891957476735115} 02/26/2022 12:52:04 - INFO - codeparrot_training - Step 45777: {'lr': 9.488616453008304e-06, 'samples': 23438336, 'steps': 45777, 'loss/train': 1.465423583984375} 02/26/2022 12:52:10 - INFO - codeparrot_training - Step 45778: {'lr': 9.484151832812454e-06, 'samples': 23438848, 'steps': 45778, 'loss/train': 1.2099231481552124} 02/26/2022 12:52:13 - INFO - codeparrot_training - Step 45779: {'lr': 9.479688242910122e-06, 'samples': 23439360, 'steps': 45779, 'loss/train': 2.161647081375122} 02/26/2022 12:52:19 - INFO - codeparrot_training - Step 45780: {'lr': 9.47522568332046e-06, 'samples': 23439872, 'steps': 45780, 'loss/train': 2.580091953277588} 02/26/2022 12:52:22 - INFO - codeparrot_training - Step 45781: {'lr': 9.470764154062505e-06, 'samples': 23440384, 'steps': 45781, 'loss/train': 2.033205270767212} 02/26/2022 12:52:28 - INFO - codeparrot_training - Step 45782: {'lr': 9.466303655155439e-06, 'samples': 23440896, 'steps': 45782, 'loss/train': 1.76809561252594} 02/26/2022 12:52:31 - INFO - codeparrot_training - Step 45783: {'lr': 9.461844186618385e-06, 'samples': 23441408, 'steps': 45783, 'loss/train': 1.7237430810928345} 02/26/2022 12:52:37 - INFO - codeparrot_training - Step 45784: {'lr': 9.457385748470382e-06, 'samples': 23441920, 'steps': 45784, 'loss/train': 1.134592056274414} 02/26/2022 12:52:41 - INFO - codeparrot_training - Step 45785: {'lr': 9.452928340730554e-06, 'samples': 23442432, 'steps': 45785, 'loss/train': 1.6695693731307983} 02/26/2022 12:52:47 - INFO - codeparrot_training - Step 45786: {'lr': 9.448471963417998e-06, 'samples': 23442944, 'steps': 45786, 'loss/train': 1.2193089723587036} 02/26/2022 12:52:50 - INFO - codeparrot_training - Step 45787: {'lr': 9.444016616551837e-06, 'samples': 23443456, 'steps': 45787, 'loss/train': 0.04601413011550903} 02/26/2022 12:52:56 - INFO - codeparrot_training - Step 45788: {'lr': 9.439562300151112e-06, 'samples': 23443968, 'steps': 45788, 'loss/train': 1.4977885484695435} 02/26/2022 12:52:59 - INFO - codeparrot_training - Step 45789: {'lr': 9.43510901423486e-06, 'samples': 23444480, 'steps': 45789, 'loss/train': 2.6571083068847656} 02/26/2022 12:53:05 - INFO - codeparrot_training - Step 45790: {'lr': 9.430656758822264e-06, 'samples': 23444992, 'steps': 45790, 'loss/train': 2.8447751998901367} 02/26/2022 12:53:08 - INFO - codeparrot_training - Step 45791: {'lr': 9.426205533932363e-06, 'samples': 23445504, 'steps': 45791, 'loss/train': 0.9111694693565369} 02/26/2022 12:53:14 - INFO - codeparrot_training - Step 45792: {'lr': 9.42175533958417e-06, 'samples': 23446016, 'steps': 45792, 'loss/train': 0.949516773223877} 02/26/2022 12:53:17 - INFO - codeparrot_training - Step 45793: {'lr': 9.41730617579678e-06, 'samples': 23446528, 'steps': 45793, 'loss/train': 0.7893884778022766} 02/26/2022 12:53:22 - INFO - codeparrot_training - Step 45794: {'lr': 9.41285804258929e-06, 'samples': 23447040, 'steps': 45794, 'loss/train': 1.7310644388198853} 02/26/2022 12:53:26 - INFO - codeparrot_training - Step 45795: {'lr': 9.40841093998071e-06, 'samples': 23447552, 'steps': 45795, 'loss/train': 1.2345225811004639} 02/26/2022 12:53:32 - INFO - codeparrot_training - Step 45796: {'lr': 9.403964867990111e-06, 'samples': 23448064, 'steps': 45796, 'loss/train': 0.19714412093162537} 02/26/2022 12:53:35 - INFO - codeparrot_training - Step 45797: {'lr': 9.399519826636478e-06, 'samples': 23448576, 'steps': 45797, 'loss/train': 2.241469621658325} 02/26/2022 12:53:40 - INFO - codeparrot_training - Step 45798: {'lr': 9.395075815938958e-06, 'samples': 23449088, 'steps': 45798, 'loss/train': 0.5774608850479126} 02/26/2022 12:53:44 - INFO - codeparrot_training - Step 45799: {'lr': 9.390632835916513e-06, 'samples': 23449600, 'steps': 45799, 'loss/train': 1.9616197347640991} 02/26/2022 12:53:50 - INFO - codeparrot_training - Step 45800: {'lr': 9.386190886588208e-06, 'samples': 23450112, 'steps': 45800, 'loss/train': 3.375455379486084} 02/26/2022 12:53:53 - INFO - codeparrot_training - Step 45801: {'lr': 9.381749967973002e-06, 'samples': 23450624, 'steps': 45801, 'loss/train': 2.8787474632263184} 02/26/2022 12:54:00 - INFO - codeparrot_training - Step 45802: {'lr': 9.377310080090047e-06, 'samples': 23451136, 'steps': 45802, 'loss/train': 2.1386520862579346} 02/26/2022 12:54:03 - INFO - codeparrot_training - Step 45803: {'lr': 9.372871222958213e-06, 'samples': 23451648, 'steps': 45803, 'loss/train': 1.2434087991714478} 02/26/2022 12:54:09 - INFO - codeparrot_training - Step 45804: {'lr': 9.368433396596681e-06, 'samples': 23452160, 'steps': 45804, 'loss/train': 1.6879916191101074} 02/26/2022 12:54:12 - INFO - codeparrot_training - Step 45805: {'lr': 9.363996601024299e-06, 'samples': 23452672, 'steps': 45805, 'loss/train': 0.04310918599367142} 02/26/2022 12:54:18 - INFO - codeparrot_training - Step 45806: {'lr': 9.359560836260189e-06, 'samples': 23453184, 'steps': 45806, 'loss/train': 1.6436880826950073} 02/26/2022 12:54:21 - INFO - codeparrot_training - Step 45807: {'lr': 9.35512610232328e-06, 'samples': 23453696, 'steps': 45807, 'loss/train': 2.1185781955718994} 02/26/2022 12:54:27 - INFO - codeparrot_training - Step 45808: {'lr': 9.35069239923264e-06, 'samples': 23454208, 'steps': 45808, 'loss/train': 1.0220075845718384} 02/26/2022 12:54:30 - INFO - codeparrot_training - Step 45809: {'lr': 9.346259727007173e-06, 'samples': 23454720, 'steps': 45809, 'loss/train': 1.7280954122543335} 02/26/2022 12:54:36 - INFO - codeparrot_training - Step 45810: {'lr': 9.341828085665915e-06, 'samples': 23455232, 'steps': 45810, 'loss/train': 1.114738941192627} 02/26/2022 12:54:39 - INFO - codeparrot_training - Step 45811: {'lr': 9.337397475227855e-06, 'samples': 23455744, 'steps': 45811, 'loss/train': 1.9941314458847046} 02/26/2022 12:54:45 - INFO - codeparrot_training - Step 45812: {'lr': 9.332967895712002e-06, 'samples': 23456256, 'steps': 45812, 'loss/train': 2.0828349590301514} 02/26/2022 12:54:49 - INFO - codeparrot_training - Step 45813: {'lr': 9.328539347137233e-06, 'samples': 23456768, 'steps': 45813, 'loss/train': 0.07467823475599289} 02/26/2022 12:54:54 - INFO - codeparrot_training - Step 45814: {'lr': 9.324111829522641e-06, 'samples': 23457280, 'steps': 45814, 'loss/train': 2.331038236618042} 02/26/2022 12:54:58 - INFO - codeparrot_training - Step 45815: {'lr': 9.319685342887074e-06, 'samples': 23457792, 'steps': 45815, 'loss/train': 2.171290636062622} 02/26/2022 12:55:03 - INFO - codeparrot_training - Step 45816: {'lr': 9.315259887249572e-06, 'samples': 23458304, 'steps': 45816, 'loss/train': 1.9826956987380981} 02/26/2022 12:55:07 - INFO - codeparrot_training - Step 45817: {'lr': 9.310835462629092e-06, 'samples': 23458816, 'steps': 45817, 'loss/train': 2.1105475425720215} 02/26/2022 12:55:12 - INFO - codeparrot_training - Step 45818: {'lr': 9.306412069044562e-06, 'samples': 23459328, 'steps': 45818, 'loss/train': 2.259345054626465} 02/26/2022 12:55:16 - INFO - codeparrot_training - Step 45819: {'lr': 9.301989706514886e-06, 'samples': 23459840, 'steps': 45819, 'loss/train': 2.4962852001190186} 02/26/2022 12:55:21 - INFO - codeparrot_training - Step 45820: {'lr': 9.2975683750591e-06, 'samples': 23460352, 'steps': 45820, 'loss/train': 2.236016273498535} 02/26/2022 12:55:25 - INFO - codeparrot_training - Step 45821: {'lr': 9.293148074696084e-06, 'samples': 23460864, 'steps': 45821, 'loss/train': 2.167633295059204} 02/26/2022 12:55:31 - INFO - codeparrot_training - Step 45822: {'lr': 9.288728805444818e-06, 'samples': 23461376, 'steps': 45822, 'loss/train': 2.3480963706970215} 02/26/2022 12:55:34 - INFO - codeparrot_training - Step 45823: {'lr': 9.28431056732415e-06, 'samples': 23461888, 'steps': 45823, 'loss/train': 1.847638487815857} 02/26/2022 12:55:40 - INFO - codeparrot_training - Step 45824: {'lr': 9.279893360353093e-06, 'samples': 23462400, 'steps': 45824, 'loss/train': 3.16902494430542} 02/26/2022 12:55:43 - INFO - codeparrot_training - Step 45825: {'lr': 9.275477184550546e-06, 'samples': 23462912, 'steps': 45825, 'loss/train': 1.3458139896392822} 02/26/2022 12:55:49 - INFO - codeparrot_training - Step 45826: {'lr': 9.271062039935412e-06, 'samples': 23463424, 'steps': 45826, 'loss/train': 0.8282353281974792} 02/26/2022 12:55:52 - INFO - codeparrot_training - Step 45827: {'lr': 9.266647926526595e-06, 'samples': 23463936, 'steps': 45827, 'loss/train': 1.0621026754379272} 02/26/2022 12:55:58 - INFO - codeparrot_training - Step 45828: {'lr': 9.26223484434302e-06, 'samples': 23464448, 'steps': 45828, 'loss/train': 2.1471261978149414} 02/26/2022 12:56:02 - INFO - codeparrot_training - Step 45829: {'lr': 9.257822793403591e-06, 'samples': 23464960, 'steps': 45829, 'loss/train': 2.161721706390381} 02/26/2022 12:56:05 - INFO - codeparrot_training - Step 45830: {'lr': 9.253411773727211e-06, 'samples': 23465472, 'steps': 45830, 'loss/train': 2.138962745666504} 02/26/2022 12:56:11 - INFO - codeparrot_training - Step 45831: {'lr': 9.24900178533275e-06, 'samples': 23465984, 'steps': 45831, 'loss/train': 0.8545023798942566} 02/26/2022 12:56:14 - INFO - codeparrot_training - Step 45832: {'lr': 9.244592828239112e-06, 'samples': 23466496, 'steps': 45832, 'loss/train': 1.726701021194458} 02/26/2022 12:56:21 - INFO - codeparrot_training - Step 45833: {'lr': 9.240184902465226e-06, 'samples': 23467008, 'steps': 45833, 'loss/train': 1.7529367208480835} 02/26/2022 12:56:24 - INFO - codeparrot_training - Step 45834: {'lr': 9.23577800802991e-06, 'samples': 23467520, 'steps': 45834, 'loss/train': 0.9927670955657959} 02/26/2022 12:56:30 - INFO - codeparrot_training - Step 45835: {'lr': 9.231372144952094e-06, 'samples': 23468032, 'steps': 45835, 'loss/train': 1.4131137132644653} 02/26/2022 12:56:33 - INFO - codeparrot_training - Step 45836: {'lr': 9.226967313250595e-06, 'samples': 23468544, 'steps': 45836, 'loss/train': 0.463163822889328} 02/26/2022 12:56:39 - INFO - codeparrot_training - Step 45837: {'lr': 9.222563512944343e-06, 'samples': 23469056, 'steps': 45837, 'loss/train': 1.06686532497406} 02/26/2022 12:56:42 - INFO - codeparrot_training - Step 45838: {'lr': 9.218160744052157e-06, 'samples': 23469568, 'steps': 45838, 'loss/train': 1.561614751815796} 02/26/2022 12:56:48 - INFO - codeparrot_training - Step 45839: {'lr': 9.213759006592936e-06, 'samples': 23470080, 'steps': 45839, 'loss/train': 2.109992265701294} 02/26/2022 12:56:51 - INFO - codeparrot_training - Step 45840: {'lr': 9.209358300585474e-06, 'samples': 23470592, 'steps': 45840, 'loss/train': 1.4622077941894531} 02/26/2022 12:56:57 - INFO - codeparrot_training - Step 45841: {'lr': 9.204958626048671e-06, 'samples': 23471104, 'steps': 45841, 'loss/train': 1.6549198627471924} 02/26/2022 12:57:00 - INFO - codeparrot_training - Step 45842: {'lr': 9.200559983001372e-06, 'samples': 23471616, 'steps': 45842, 'loss/train': 1.8383203744888306} 02/26/2022 12:57:07 - INFO - codeparrot_training - Step 45843: {'lr': 9.196162371462451e-06, 'samples': 23472128, 'steps': 45843, 'loss/train': 1.9243422746658325} 02/26/2022 12:57:10 - INFO - codeparrot_training - Step 45844: {'lr': 9.191765791450645e-06, 'samples': 23472640, 'steps': 45844, 'loss/train': 1.528111457824707} 02/26/2022 12:57:16 - INFO - codeparrot_training - Step 45845: {'lr': 9.187370242984883e-06, 'samples': 23473152, 'steps': 45845, 'loss/train': 0.7781782746315002} 02/26/2022 12:57:19 - INFO - codeparrot_training - Step 45846: {'lr': 9.182975726083925e-06, 'samples': 23473664, 'steps': 45846, 'loss/train': 1.097321629524231} 02/26/2022 12:57:25 - INFO - codeparrot_training - Step 45847: {'lr': 9.178582240766703e-06, 'samples': 23474176, 'steps': 45847, 'loss/train': 1.3071389198303223} 02/26/2022 12:57:28 - INFO - codeparrot_training - Step 45848: {'lr': 9.174189787051896e-06, 'samples': 23474688, 'steps': 45848, 'loss/train': 1.2330803871154785} 02/26/2022 12:57:34 - INFO - codeparrot_training - Step 45849: {'lr': 9.169798364958404e-06, 'samples': 23475200, 'steps': 45849, 'loss/train': 1.8518048524856567} 02/26/2022 12:57:37 - INFO - codeparrot_training - Step 45850: {'lr': 9.165407974505019e-06, 'samples': 23475712, 'steps': 45850, 'loss/train': 1.329521894454956} 02/26/2022 12:57:43 - INFO - codeparrot_training - Step 45851: {'lr': 9.161018615710587e-06, 'samples': 23476224, 'steps': 45851, 'loss/train': 1.731979489326477} 02/26/2022 12:57:46 - INFO - codeparrot_training - Step 45852: {'lr': 9.156630288593842e-06, 'samples': 23476736, 'steps': 45852, 'loss/train': 2.024141550064087} 02/26/2022 12:57:52 - INFO - codeparrot_training - Step 45853: {'lr': 9.152242993173632e-06, 'samples': 23477248, 'steps': 45853, 'loss/train': 2.0704832077026367} 02/26/2022 12:57:56 - INFO - codeparrot_training - Step 45854: {'lr': 9.147856729468717e-06, 'samples': 23477760, 'steps': 45854, 'loss/train': 1.230162501335144} 02/26/2022 12:58:01 - INFO - codeparrot_training - Step 45855: {'lr': 9.143471497497946e-06, 'samples': 23478272, 'steps': 45855, 'loss/train': 1.0638129711151123} 02/26/2022 12:58:05 - INFO - codeparrot_training - Step 45856: {'lr': 9.139087297280024e-06, 'samples': 23478784, 'steps': 45856, 'loss/train': 1.7369027137756348} 02/26/2022 12:58:10 - INFO - codeparrot_training - Step 45857: {'lr': 9.134704128833798e-06, 'samples': 23479296, 'steps': 45857, 'loss/train': 2.1756985187530518} 02/26/2022 12:58:14 - INFO - codeparrot_training - Step 45858: {'lr': 9.130321992177976e-06, 'samples': 23479808, 'steps': 45858, 'loss/train': 1.2524319887161255} 02/26/2022 12:58:20 - INFO - codeparrot_training - Step 45859: {'lr': 9.125940887331402e-06, 'samples': 23480320, 'steps': 45859, 'loss/train': 1.6620538234710693} 02/26/2022 12:58:23 - INFO - codeparrot_training - Step 45860: {'lr': 9.121560814312813e-06, 'samples': 23480832, 'steps': 45860, 'loss/train': 1.9836480617523193} 02/26/2022 12:58:29 - INFO - codeparrot_training - Step 45861: {'lr': 9.11718177314097e-06, 'samples': 23481344, 'steps': 45861, 'loss/train': 1.1807059049606323} 02/26/2022 12:58:33 - INFO - codeparrot_training - Step 45862: {'lr': 9.11280376383461e-06, 'samples': 23481856, 'steps': 45862, 'loss/train': 1.1456806659698486} 02/26/2022 12:58:38 - INFO - codeparrot_training - Step 45863: {'lr': 9.10842678641252e-06, 'samples': 23482368, 'steps': 45863, 'loss/train': 1.6775380373001099} 02/26/2022 12:58:41 - INFO - codeparrot_training - Step 45864: {'lr': 9.10405084089344e-06, 'samples': 23482880, 'steps': 45864, 'loss/train': 1.735140323638916} 02/26/2022 12:58:47 - INFO - codeparrot_training - Step 45865: {'lr': 9.099675927296131e-06, 'samples': 23483392, 'steps': 45865, 'loss/train': 1.1546891927719116} 02/26/2022 12:58:51 - INFO - codeparrot_training - Step 45866: {'lr': 9.09530204563927e-06, 'samples': 23483904, 'steps': 45866, 'loss/train': 0.9374484419822693} 02/26/2022 12:58:56 - INFO - codeparrot_training - Step 45867: {'lr': 9.090929195941678e-06, 'samples': 23484416, 'steps': 45867, 'loss/train': 1.7938917875289917} 02/26/2022 12:59:00 - INFO - codeparrot_training - Step 45868: {'lr': 9.08655737822206e-06, 'samples': 23484928, 'steps': 45868, 'loss/train': 1.1096880435943604} 02/26/2022 12:59:05 - INFO - codeparrot_training - Step 45869: {'lr': 9.082186592499097e-06, 'samples': 23485440, 'steps': 45869, 'loss/train': 2.4260993003845215} 02/26/2022 12:59:08 - INFO - codeparrot_training - Step 45870: {'lr': 9.077816838791552e-06, 'samples': 23485952, 'steps': 45870, 'loss/train': 0.9869159460067749} 02/26/2022 12:59:15 - INFO - codeparrot_training - Step 45871: {'lr': 9.073448117118132e-06, 'samples': 23486464, 'steps': 45871, 'loss/train': 2.30340576171875} 02/26/2022 12:59:18 - INFO - codeparrot_training - Step 45872: {'lr': 9.069080427497572e-06, 'samples': 23486976, 'steps': 45872, 'loss/train': 3.0496270656585693} 02/26/2022 12:59:24 - INFO - codeparrot_training - Step 45873: {'lr': 9.064713769948579e-06, 'samples': 23487488, 'steps': 45873, 'loss/train': 1.0741225481033325} 02/26/2022 12:59:27 - INFO - codeparrot_training - Step 45874: {'lr': 9.060348144489833e-06, 'samples': 23488000, 'steps': 45874, 'loss/train': 1.533676028251648} 02/26/2022 12:59:33 - INFO - codeparrot_training - Step 45875: {'lr': 9.05598355114004e-06, 'samples': 23488512, 'steps': 45875, 'loss/train': 2.1415016651153564} 02/26/2022 12:59:36 - INFO - codeparrot_training - Step 45876: {'lr': 9.051619989917909e-06, 'samples': 23489024, 'steps': 45876, 'loss/train': 0.37592703104019165} 02/26/2022 12:59:42 - INFO - codeparrot_training - Step 45877: {'lr': 9.047257460842118e-06, 'samples': 23489536, 'steps': 45877, 'loss/train': 0.7004016637802124} 02/26/2022 12:59:45 - INFO - codeparrot_training - Step 45878: {'lr': 9.04289596393143e-06, 'samples': 23490048, 'steps': 45878, 'loss/train': 1.8229241371154785} 02/26/2022 12:59:51 - INFO - codeparrot_training - Step 45879: {'lr': 9.038535499204387e-06, 'samples': 23490560, 'steps': 45879, 'loss/train': 1.2572519779205322} 02/26/2022 12:59:54 - INFO - codeparrot_training - Step 45880: {'lr': 9.034176066679779e-06, 'samples': 23491072, 'steps': 45880, 'loss/train': 2.00264835357666} 02/26/2022 13:00:00 - INFO - codeparrot_training - Step 45881: {'lr': 9.0298176663762e-06, 'samples': 23491584, 'steps': 45881, 'loss/train': 1.9450125694274902} 02/26/2022 13:00:04 - INFO - codeparrot_training - Step 45882: {'lr': 9.025460298312443e-06, 'samples': 23492096, 'steps': 45882, 'loss/train': 1.8951421976089478} 02/26/2022 13:00:09 - INFO - codeparrot_training - Step 45883: {'lr': 9.021103962507022e-06, 'samples': 23492608, 'steps': 45883, 'loss/train': 0.12006185203790665} 02/26/2022 13:00:13 - INFO - codeparrot_training - Step 45884: {'lr': 9.016748658978723e-06, 'samples': 23493120, 'steps': 45884, 'loss/train': 1.4472486972808838} 02/26/2022 13:00:18 - INFO - codeparrot_training - Step 45885: {'lr': 9.01239438774612e-06, 'samples': 23493632, 'steps': 45885, 'loss/train': 0.9986704587936401} 02/26/2022 13:00:22 - INFO - codeparrot_training - Step 45886: {'lr': 9.008041148827944e-06, 'samples': 23494144, 'steps': 45886, 'loss/train': 1.5340802669525146} 02/26/2022 13:00:27 - INFO - codeparrot_training - Step 45887: {'lr': 9.003688942242737e-06, 'samples': 23494656, 'steps': 45887, 'loss/train': 1.8583381175994873} 02/26/2022 13:00:31 - INFO - codeparrot_training - Step 45888: {'lr': 8.999337768009235e-06, 'samples': 23495168, 'steps': 45888, 'loss/train': 1.8377958536148071} 02/26/2022 13:00:36 - INFO - codeparrot_training - Step 45889: {'lr': 8.994987626146007e-06, 'samples': 23495680, 'steps': 45889, 'loss/train': 1.835382103919983} 02/26/2022 13:00:40 - INFO - codeparrot_training - Step 45890: {'lr': 8.990638516671757e-06, 'samples': 23496192, 'steps': 45890, 'loss/train': 2.2052927017211914} 02/26/2022 13:00:46 - INFO - codeparrot_training - Step 45891: {'lr': 8.986290439605083e-06, 'samples': 23496704, 'steps': 45891, 'loss/train': 1.654951572418213} 02/26/2022 13:00:50 - INFO - codeparrot_training - Step 45892: {'lr': 8.981943394964637e-06, 'samples': 23497216, 'steps': 45892, 'loss/train': 2.0534675121307373} 02/26/2022 13:00:55 - INFO - codeparrot_training - Step 45893: {'lr': 8.97759738276896e-06, 'samples': 23497728, 'steps': 45893, 'loss/train': 1.6069839000701904} 02/26/2022 13:00:59 - INFO - codeparrot_training - Step 45894: {'lr': 8.97325240303673e-06, 'samples': 23498240, 'steps': 45894, 'loss/train': 0.6512133479118347} 02/26/2022 13:01:04 - INFO - codeparrot_training - Step 45895: {'lr': 8.968908455786601e-06, 'samples': 23498752, 'steps': 45895, 'loss/train': 2.1880362033843994} 02/26/2022 13:01:08 - INFO - codeparrot_training - Step 45896: {'lr': 8.964565541037084e-06, 'samples': 23499264, 'steps': 45896, 'loss/train': 0.9493980407714844} 02/26/2022 13:01:13 - INFO - codeparrot_training - Step 45897: {'lr': 8.960223658806831e-06, 'samples': 23499776, 'steps': 45897, 'loss/train': 2.1147823333740234} 02/26/2022 13:01:17 - INFO - codeparrot_training - Step 45898: {'lr': 8.955882809114468e-06, 'samples': 23500288, 'steps': 45898, 'loss/train': 1.4750529527664185} 02/26/2022 13:01:22 - INFO - codeparrot_training - Step 45899: {'lr': 8.951542991978562e-06, 'samples': 23500800, 'steps': 45899, 'loss/train': 1.3232852220535278} 02/26/2022 13:01:28 - INFO - codeparrot_training - Step 45900: {'lr': 8.947204207417681e-06, 'samples': 23501312, 'steps': 45900, 'loss/train': 1.0434068441390991} 02/26/2022 13:01:31 - INFO - codeparrot_training - Step 45901: {'lr': 8.942866455450448e-06, 'samples': 23501824, 'steps': 45901, 'loss/train': 1.7676472663879395} 02/26/2022 13:01:37 - INFO - codeparrot_training - Step 45902: {'lr': 8.938529736095435e-06, 'samples': 23502336, 'steps': 45902, 'loss/train': 2.1735122203826904} 02/26/2022 13:01:40 - INFO - codeparrot_training - Step 45903: {'lr': 8.93419404937118e-06, 'samples': 23502848, 'steps': 45903, 'loss/train': 1.2070268392562866} 02/26/2022 13:01:46 - INFO - codeparrot_training - Step 45904: {'lr': 8.929859395296363e-06, 'samples': 23503360, 'steps': 45904, 'loss/train': 0.3817344605922699} 02/26/2022 13:01:50 - INFO - codeparrot_training - Step 45905: {'lr': 8.925525773889414e-06, 'samples': 23503872, 'steps': 45905, 'loss/train': 0.6035551428794861} 02/26/2022 13:01:55 - INFO - codeparrot_training - Step 45906: {'lr': 8.921193185168986e-06, 'samples': 23504384, 'steps': 45906, 'loss/train': 1.1104191541671753} 02/26/2022 13:01:59 - INFO - codeparrot_training - Step 45907: {'lr': 8.916861629153616e-06, 'samples': 23504896, 'steps': 45907, 'loss/train': 2.7395594120025635} 02/26/2022 13:02:04 - INFO - codeparrot_training - Step 45908: {'lr': 8.912531105861876e-06, 'samples': 23505408, 'steps': 45908, 'loss/train': 2.068023204803467} 02/26/2022 13:02:08 - INFO - codeparrot_training - Step 45909: {'lr': 8.90820161531225e-06, 'samples': 23505920, 'steps': 45909, 'loss/train': 1.781388521194458} 02/26/2022 13:02:13 - INFO - codeparrot_training - Step 45910: {'lr': 8.903873157523362e-06, 'samples': 23506432, 'steps': 45910, 'loss/train': 0.5364334583282471} 02/26/2022 13:02:17 - INFO - codeparrot_training - Step 45911: {'lr': 8.899545732513754e-06, 'samples': 23506944, 'steps': 45911, 'loss/train': 2.1767072677612305} 02/26/2022 13:02:22 - INFO - codeparrot_training - Step 45912: {'lr': 8.89521934030188e-06, 'samples': 23507456, 'steps': 45912, 'loss/train': 1.7564771175384521} 02/26/2022 13:02:26 - INFO - codeparrot_training - Step 45913: {'lr': 8.89089398090634e-06, 'samples': 23507968, 'steps': 45913, 'loss/train': 1.9760472774505615} 02/26/2022 13:02:31 - INFO - codeparrot_training - Step 45914: {'lr': 8.886569654345672e-06, 'samples': 23508480, 'steps': 45914, 'loss/train': 2.5871505737304688} 02/26/2022 13:02:35 - INFO - codeparrot_training - Step 45915: {'lr': 8.88224636063839e-06, 'samples': 23508992, 'steps': 45915, 'loss/train': 2.2539894580841064} 02/26/2022 13:02:41 - INFO - codeparrot_training - Step 45916: {'lr': 8.877924099802954e-06, 'samples': 23509504, 'steps': 45916, 'loss/train': 1.4650945663452148} 02/26/2022 13:02:44 - INFO - codeparrot_training - Step 45917: {'lr': 8.873602871857955e-06, 'samples': 23510016, 'steps': 45917, 'loss/train': 2.014577865600586} 02/26/2022 13:02:50 - INFO - codeparrot_training - Step 45918: {'lr': 8.869282676821882e-06, 'samples': 23510528, 'steps': 45918, 'loss/train': 1.9250085353851318} 02/26/2022 13:02:53 - INFO - codeparrot_training - Step 45919: {'lr': 8.86496351471322e-06, 'samples': 23511040, 'steps': 45919, 'loss/train': 1.820523738861084} 02/26/2022 13:02:59 - INFO - codeparrot_training - Step 45920: {'lr': 8.860645385550481e-06, 'samples': 23511552, 'steps': 45920, 'loss/train': 1.5361367464065552} 02/26/2022 13:03:02 - INFO - codeparrot_training - Step 45921: {'lr': 8.856328289352177e-06, 'samples': 23512064, 'steps': 45921, 'loss/train': 0.7257646918296814} 02/26/2022 13:03:08 - INFO - codeparrot_training - Step 45922: {'lr': 8.852012226136796e-06, 'samples': 23512576, 'steps': 45922, 'loss/train': 0.7444840669631958} 02/26/2022 13:03:11 - INFO - codeparrot_training - Step 45923: {'lr': 8.847697195922822e-06, 'samples': 23513088, 'steps': 45923, 'loss/train': 2.3029396533966064} 02/26/2022 13:03:17 - INFO - codeparrot_training - Step 45924: {'lr': 8.843383198728711e-06, 'samples': 23513600, 'steps': 45924, 'loss/train': 1.55702805519104} 02/26/2022 13:03:20 - INFO - codeparrot_training - Step 45925: {'lr': 8.839070234572977e-06, 'samples': 23514112, 'steps': 45925, 'loss/train': 1.0189182758331299} 02/26/2022 13:03:26 - INFO - codeparrot_training - Step 45926: {'lr': 8.834758303474133e-06, 'samples': 23514624, 'steps': 45926, 'loss/train': 2.2082135677337646} 02/26/2022 13:03:30 - INFO - codeparrot_training - Step 45927: {'lr': 8.830447405450582e-06, 'samples': 23515136, 'steps': 45927, 'loss/train': 1.8855373859405518} 02/26/2022 13:03:35 - INFO - codeparrot_training - Step 45928: {'lr': 8.826137540520778e-06, 'samples': 23515648, 'steps': 45928, 'loss/train': 2.0455801486968994} 02/26/2022 13:03:39 - INFO - codeparrot_training - Step 45929: {'lr': 8.821828708703238e-06, 'samples': 23516160, 'steps': 45929, 'loss/train': 1.8347651958465576} 02/26/2022 13:03:42 - INFO - codeparrot_training - Step 45930: {'lr': 8.817520910016447e-06, 'samples': 23516672, 'steps': 45930, 'loss/train': 1.90438711643219} 02/26/2022 13:03:48 - INFO - codeparrot_training - Step 45931: {'lr': 8.813214144478776e-06, 'samples': 23517184, 'steps': 45931, 'loss/train': 2.3967397212982178} 02/26/2022 13:03:53 - INFO - codeparrot_training - Step 45932: {'lr': 8.808908412108685e-06, 'samples': 23517696, 'steps': 45932, 'loss/train': 2.6964433193206787} 02/26/2022 13:03:57 - INFO - codeparrot_training - Step 45933: {'lr': 8.804603712924686e-06, 'samples': 23518208, 'steps': 45933, 'loss/train': 1.3708142042160034} 02/26/2022 13:04:02 - INFO - codeparrot_training - Step 45934: {'lr': 8.800300046945154e-06, 'samples': 23518720, 'steps': 45934, 'loss/train': 2.43684720993042} 02/26/2022 13:04:06 - INFO - codeparrot_training - Step 45935: {'lr': 8.795997414188573e-06, 'samples': 23519232, 'steps': 45935, 'loss/train': 0.43741047382354736} 02/26/2022 13:04:12 - INFO - codeparrot_training - Step 45936: {'lr': 8.791695814673289e-06, 'samples': 23519744, 'steps': 45936, 'loss/train': 2.609468460083008} 02/26/2022 13:04:16 - INFO - codeparrot_training - Step 45937: {'lr': 8.787395248417846e-06, 'samples': 23520256, 'steps': 45937, 'loss/train': 2.0758635997772217} 02/26/2022 13:04:21 - INFO - codeparrot_training - Step 45938: {'lr': 8.783095715440558e-06, 'samples': 23520768, 'steps': 45938, 'loss/train': 1.8781394958496094} 02/26/2022 13:04:24 - INFO - codeparrot_training - Step 45939: {'lr': 8.778797215759971e-06, 'samples': 23521280, 'steps': 45939, 'loss/train': 2.147948980331421} 02/26/2022 13:04:30 - INFO - codeparrot_training - Step 45940: {'lr': 8.774499749394344e-06, 'samples': 23521792, 'steps': 45940, 'loss/train': 2.4299027919769287} 02/26/2022 13:04:33 - INFO - codeparrot_training - Step 45941: {'lr': 8.77020331636219e-06, 'samples': 23522304, 'steps': 45941, 'loss/train': 1.6396194696426392} 02/26/2022 13:04:39 - INFO - codeparrot_training - Step 45942: {'lr': 8.765907916681887e-06, 'samples': 23522816, 'steps': 45942, 'loss/train': 2.0596086978912354} 02/26/2022 13:04:42 - INFO - codeparrot_training - Step 45943: {'lr': 8.761613550371861e-06, 'samples': 23523328, 'steps': 45943, 'loss/train': 0.8581370711326599} 02/26/2022 13:04:48 - INFO - codeparrot_training - Step 45944: {'lr': 8.757320217450432e-06, 'samples': 23523840, 'steps': 45944, 'loss/train': 1.881097674369812} 02/26/2022 13:04:51 - INFO - codeparrot_training - Step 45945: {'lr': 8.753027917936084e-06, 'samples': 23524352, 'steps': 45945, 'loss/train': 1.4257121086120605} 02/26/2022 13:04:57 - INFO - codeparrot_training - Step 45946: {'lr': 8.748736651847112e-06, 'samples': 23524864, 'steps': 45946, 'loss/train': 2.3621208667755127} 02/26/2022 13:05:01 - INFO - codeparrot_training - Step 45947: {'lr': 8.744446419201996e-06, 'samples': 23525376, 'steps': 45947, 'loss/train': 2.346371650695801} 02/26/2022 13:05:06 - INFO - codeparrot_training - Step 45948: {'lr': 8.74015722001903e-06, 'samples': 23525888, 'steps': 45948, 'loss/train': 1.1491217613220215} 02/26/2022 13:05:09 - INFO - codeparrot_training - Step 45949: {'lr': 8.735869054316642e-06, 'samples': 23526400, 'steps': 45949, 'loss/train': 2.288332939147949} 02/26/2022 13:05:15 - INFO - codeparrot_training - Step 45950: {'lr': 8.731581922113152e-06, 'samples': 23526912, 'steps': 45950, 'loss/train': 2.613100290298462} 02/26/2022 13:05:18 - INFO - codeparrot_training - Step 45951: {'lr': 8.72729582342699e-06, 'samples': 23527424, 'steps': 45951, 'loss/train': 0.8736313581466675} 02/26/2022 13:05:25 - INFO - codeparrot_training - Step 45952: {'lr': 8.723010758276446e-06, 'samples': 23527936, 'steps': 45952, 'loss/train': 2.044598340988159} 02/26/2022 13:05:28 - INFO - codeparrot_training - Step 45953: {'lr': 8.718726726679948e-06, 'samples': 23528448, 'steps': 45953, 'loss/train': 0.5576093792915344} 02/26/2022 13:05:34 - INFO - codeparrot_training - Step 45954: {'lr': 8.714443728655763e-06, 'samples': 23528960, 'steps': 45954, 'loss/train': 2.2565579414367676} 02/26/2022 13:05:37 - INFO - codeparrot_training - Step 45955: {'lr': 8.710161764222319e-06, 'samples': 23529472, 'steps': 45955, 'loss/train': 1.1954596042633057} 02/26/2022 13:05:43 - INFO - codeparrot_training - Step 45956: {'lr': 8.705880833397934e-06, 'samples': 23529984, 'steps': 45956, 'loss/train': 1.9215234518051147} 02/26/2022 13:05:46 - INFO - codeparrot_training - Step 45957: {'lr': 8.701600936200927e-06, 'samples': 23530496, 'steps': 45957, 'loss/train': 1.510150671005249} 02/26/2022 13:05:52 - INFO - codeparrot_training - Step 45958: {'lr': 8.697322072649617e-06, 'samples': 23531008, 'steps': 45958, 'loss/train': 1.6921449899673462} 02/26/2022 13:05:55 - INFO - codeparrot_training - Step 45959: {'lr': 8.69304424276235e-06, 'samples': 23531520, 'steps': 45959, 'loss/train': 2.34102725982666} 02/26/2022 13:06:01 - INFO - codeparrot_training - Step 45960: {'lr': 8.6887674465575e-06, 'samples': 23532032, 'steps': 45960, 'loss/train': 2.7821972370147705} 02/26/2022 13:06:04 - INFO - codeparrot_training - Step 45961: {'lr': 8.68449168405333e-06, 'samples': 23532544, 'steps': 45961, 'loss/train': 2.6380436420440674} 02/26/2022 13:06:10 - INFO - codeparrot_training - Step 45962: {'lr': 8.680216955268161e-06, 'samples': 23533056, 'steps': 45962, 'loss/train': 0.8853145241737366} 02/26/2022 13:06:14 - INFO - codeparrot_training - Step 45963: {'lr': 8.675943260220309e-06, 'samples': 23533568, 'steps': 45963, 'loss/train': 1.9960031509399414} 02/26/2022 13:06:19 - INFO - codeparrot_training - Step 45964: {'lr': 8.671670598928095e-06, 'samples': 23534080, 'steps': 45964, 'loss/train': 1.354811429977417} 02/26/2022 13:06:23 - INFO - codeparrot_training - Step 45965: {'lr': 8.667398971409834e-06, 'samples': 23534592, 'steps': 45965, 'loss/train': 2.2799642086029053} 02/26/2022 13:06:28 - INFO - codeparrot_training - Step 45966: {'lr': 8.663128377683794e-06, 'samples': 23535104, 'steps': 45966, 'loss/train': 1.219589114189148} 02/26/2022 13:06:32 - INFO - codeparrot_training - Step 45967: {'lr': 8.658858817768233e-06, 'samples': 23535616, 'steps': 45967, 'loss/train': 0.9523189663887024} 02/26/2022 13:06:37 - INFO - codeparrot_training - Step 45968: {'lr': 8.65459029168153e-06, 'samples': 23536128, 'steps': 45968, 'loss/train': 1.415152907371521} 02/26/2022 13:06:41 - INFO - codeparrot_training - Step 45969: {'lr': 8.650322799441945e-06, 'samples': 23536640, 'steps': 45969, 'loss/train': 1.439796805381775} 02/26/2022 13:06:46 - INFO - codeparrot_training - Step 45970: {'lr': 8.646056341067714e-06, 'samples': 23537152, 'steps': 45970, 'loss/train': 1.2613742351531982} 02/26/2022 13:06:50 - INFO - codeparrot_training - Step 45971: {'lr': 8.64179091657713e-06, 'samples': 23537664, 'steps': 45971, 'loss/train': 2.420267105102539} 02/26/2022 13:06:56 - INFO - codeparrot_training - Step 45972: {'lr': 8.637526525988481e-06, 'samples': 23538176, 'steps': 45972, 'loss/train': 1.1902494430541992} 02/26/2022 13:06:59 - INFO - codeparrot_training - Step 45973: {'lr': 8.633263169320005e-06, 'samples': 23538688, 'steps': 45973, 'loss/train': 1.8855726718902588} 02/26/2022 13:07:05 - INFO - codeparrot_training - Step 45974: {'lr': 8.629000846590045e-06, 'samples': 23539200, 'steps': 45974, 'loss/train': 1.6354668140411377} 02/26/2022 13:07:11 - INFO - codeparrot_training - Step 45975: {'lr': 8.624739557816758e-06, 'samples': 23539712, 'steps': 45975, 'loss/train': 2.1895813941955566} 02/26/2022 13:07:14 - INFO - codeparrot_training - Step 45976: {'lr': 8.620479303018458e-06, 'samples': 23540224, 'steps': 45976, 'loss/train': 1.7519341707229614} 02/26/2022 13:07:18 - INFO - codeparrot_training - Step 45977: {'lr': 8.616220082213355e-06, 'samples': 23540736, 'steps': 45977, 'loss/train': 2.413083553314209} 02/26/2022 13:07:23 - INFO - codeparrot_training - Step 45978: {'lr': 8.611961895419767e-06, 'samples': 23541248, 'steps': 45978, 'loss/train': 1.788680076599121} 02/26/2022 13:07:29 - INFO - codeparrot_training - Step 45979: {'lr': 8.60770474265582e-06, 'samples': 23541760, 'steps': 45979, 'loss/train': 2.295016050338745} 02/26/2022 13:07:32 - INFO - codeparrot_training - Step 45980: {'lr': 8.603448623939857e-06, 'samples': 23542272, 'steps': 45980, 'loss/train': 1.5782240629196167} 02/26/2022 13:07:39 - INFO - codeparrot_training - Step 45981: {'lr': 8.599193539290034e-06, 'samples': 23542784, 'steps': 45981, 'loss/train': 2.5019900798797607} 02/26/2022 13:07:42 - INFO - codeparrot_training - Step 45982: {'lr': 8.594939488724695e-06, 'samples': 23543296, 'steps': 45982, 'loss/train': 0.2256821244955063} 02/26/2022 13:07:46 - INFO - codeparrot_training - Step 45983: {'lr': 8.590686472261882e-06, 'samples': 23543808, 'steps': 45983, 'loss/train': 2.6255862712860107} 02/26/2022 13:07:51 - INFO - codeparrot_training - Step 45984: {'lr': 8.58643448991997e-06, 'samples': 23544320, 'steps': 45984, 'loss/train': 0.9441692233085632} 02/26/2022 13:07:55 - INFO - codeparrot_training - Step 45985: {'lr': 8.582183541717081e-06, 'samples': 23544832, 'steps': 45985, 'loss/train': 0.22554410994052887} 02/26/2022 13:08:00 - INFO - codeparrot_training - Step 45986: {'lr': 8.57793362767148e-06, 'samples': 23545344, 'steps': 45986, 'loss/train': 0.9938243627548218} 02/26/2022 13:08:04 - INFO - codeparrot_training - Step 45987: {'lr': 8.573684747801347e-06, 'samples': 23545856, 'steps': 45987, 'loss/train': 0.31699928641319275} 02/26/2022 13:08:09 - INFO - codeparrot_training - Step 45988: {'lr': 8.569436902124888e-06, 'samples': 23546368, 'steps': 45988, 'loss/train': 0.4721957743167877} 02/26/2022 13:08:13 - INFO - codeparrot_training - Step 45989: {'lr': 8.565190090660286e-06, 'samples': 23546880, 'steps': 45989, 'loss/train': 2.2545809745788574} 02/26/2022 13:08:18 - INFO - codeparrot_training - Step 45990: {'lr': 8.560944313425772e-06, 'samples': 23547392, 'steps': 45990, 'loss/train': 1.7034109830856323} 02/26/2022 13:08:24 - INFO - codeparrot_training - Step 45991: {'lr': 8.556699570439503e-06, 'samples': 23547904, 'steps': 45991, 'loss/train': 1.6170326471328735} 02/26/2022 13:08:27 - INFO - codeparrot_training - Step 45992: {'lr': 8.552455861719655e-06, 'samples': 23548416, 'steps': 45992, 'loss/train': 1.6514886617660522} 02/26/2022 13:08:33 - INFO - codeparrot_training - Step 45993: {'lr': 8.548213187284409e-06, 'samples': 23548928, 'steps': 45993, 'loss/train': 1.1597504615783691} 02/26/2022 13:08:36 - INFO - codeparrot_training - Step 45994: {'lr': 8.543971547151946e-06, 'samples': 23549440, 'steps': 45994, 'loss/train': 1.3295881748199463} 02/26/2022 13:08:42 - INFO - codeparrot_training - Step 45995: {'lr': 8.539730941340473e-06, 'samples': 23549952, 'steps': 45995, 'loss/train': 1.5939793586730957} 02/26/2022 13:08:45 - INFO - codeparrot_training - Step 45996: {'lr': 8.535491369868087e-06, 'samples': 23550464, 'steps': 45996, 'loss/train': 1.8474210500717163} 02/26/2022 13:08:51 - INFO - codeparrot_training - Step 45997: {'lr': 8.531252832752967e-06, 'samples': 23550976, 'steps': 45997, 'loss/train': 1.1621639728546143} 02/26/2022 13:08:55 - INFO - codeparrot_training - Step 45998: {'lr': 8.52701533001332e-06, 'samples': 23551488, 'steps': 45998, 'loss/train': 1.7301279306411743} 02/26/2022 13:09:00 - INFO - codeparrot_training - Step 45999: {'lr': 8.522778861667273e-06, 'samples': 23552000, 'steps': 45999, 'loss/train': 1.087491512298584} 02/26/2022 13:09:00 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 13:09:18 - WARNING - huggingface_hub.repository - Several commits (46) will be pushed upstream. 02/26/2022 13:09:18 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 13:09:53 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 9a19140..572893f floral-grass-11 -> floral-grass-11 02/26/2022 13:09:58 - INFO - codeparrot_training - Step 46000: {'lr': 8.51854342773295e-06, 'samples': 23552512, 'steps': 46000, 'loss/train': 1.445226788520813} 02/26/2022 13:10:03 - INFO - codeparrot_training - Step 46001: {'lr': 8.514309028228473e-06, 'samples': 23553024, 'steps': 46001, 'loss/train': 2.392515182495117} 02/26/2022 13:10:07 - INFO - codeparrot_training - Step 46002: {'lr': 8.51007566317205e-06, 'samples': 23553536, 'steps': 46002, 'loss/train': 1.8956668376922607} 02/26/2022 13:10:12 - INFO - codeparrot_training - Step 46003: {'lr': 8.505843332581781e-06, 'samples': 23554048, 'steps': 46003, 'loss/train': 1.446045160293579} 02/26/2022 13:10:16 - INFO - codeparrot_training - Step 46004: {'lr': 8.501612036475815e-06, 'samples': 23554560, 'steps': 46004, 'loss/train': 0.8679484128952026} 02/26/2022 13:10:21 - INFO - codeparrot_training - Step 46005: {'lr': 8.49738177487222e-06, 'samples': 23555072, 'steps': 46005, 'loss/train': 1.8368754386901855} 02/26/2022 13:10:25 - INFO - codeparrot_training - Step 46006: {'lr': 8.493152547789152e-06, 'samples': 23555584, 'steps': 46006, 'loss/train': 1.892811894416809} 02/26/2022 13:10:31 - INFO - codeparrot_training - Step 46007: {'lr': 8.48892435524476e-06, 'samples': 23556096, 'steps': 46007, 'loss/train': 1.4583792686462402} 02/26/2022 13:10:34 - INFO - codeparrot_training - Step 46008: {'lr': 8.484697197257142e-06, 'samples': 23556608, 'steps': 46008, 'loss/train': 0.33918002247810364} 02/26/2022 13:10:40 - INFO - codeparrot_training - Step 46009: {'lr': 8.480471073844365e-06, 'samples': 23557120, 'steps': 46009, 'loss/train': 1.828680157661438} 02/26/2022 13:10:43 - INFO - codeparrot_training - Step 46010: {'lr': 8.476245985024528e-06, 'samples': 23557632, 'steps': 46010, 'loss/train': 2.216860055923462} 02/26/2022 13:10:50 - INFO - codeparrot_training - Step 46011: {'lr': 8.47202193081581e-06, 'samples': 23558144, 'steps': 46011, 'loss/train': 1.7501882314682007} 02/26/2022 13:10:54 - INFO - codeparrot_training - Step 46012: {'lr': 8.467798911236224e-06, 'samples': 23558656, 'steps': 46012, 'loss/train': 1.0804424285888672} 02/26/2022 13:10:59 - INFO - codeparrot_training - Step 46013: {'lr': 8.463576926303922e-06, 'samples': 23559168, 'steps': 46013, 'loss/train': 0.7688332200050354} 02/26/2022 13:11:03 - INFO - codeparrot_training - Step 46014: {'lr': 8.459355976036947e-06, 'samples': 23559680, 'steps': 46014, 'loss/train': 0.9228298664093018} 02/26/2022 13:11:08 - INFO - codeparrot_training - Step 46015: {'lr': 8.455136060453395e-06, 'samples': 23560192, 'steps': 46015, 'loss/train': 2.0492424964904785} 02/26/2022 13:11:12 - INFO - codeparrot_training - Step 46016: {'lr': 8.450917179571306e-06, 'samples': 23560704, 'steps': 46016, 'loss/train': 0.7214325070381165} 02/26/2022 13:11:17 - INFO - codeparrot_training - Step 46017: {'lr': 8.446699333408859e-06, 'samples': 23561216, 'steps': 46017, 'loss/train': 2.676889657974243} 02/26/2022 13:11:21 - INFO - codeparrot_training - Step 46018: {'lr': 8.442482521983986e-06, 'samples': 23561728, 'steps': 46018, 'loss/train': 1.5434142351150513} 02/26/2022 13:11:27 - INFO - codeparrot_training - Step 46019: {'lr': 8.438266745314838e-06, 'samples': 23562240, 'steps': 46019, 'loss/train': 1.4020642042160034} 02/26/2022 13:11:30 - INFO - codeparrot_training - Step 46020: {'lr': 8.434052003419457e-06, 'samples': 23562752, 'steps': 46020, 'loss/train': 1.0194350481033325} 02/26/2022 13:11:36 - INFO - codeparrot_training - Step 46021: {'lr': 8.429838296315883e-06, 'samples': 23563264, 'steps': 46021, 'loss/train': 1.6184508800506592} 02/26/2022 13:11:40 - INFO - codeparrot_training - Step 46022: {'lr': 8.425625624022187e-06, 'samples': 23563776, 'steps': 46022, 'loss/train': 1.6239887475967407} 02/26/2022 13:11:45 - INFO - codeparrot_training - Step 46023: {'lr': 8.421413986556408e-06, 'samples': 23564288, 'steps': 46023, 'loss/train': 0.8545622229576111} 02/26/2022 13:11:49 - INFO - codeparrot_training - Step 46024: {'lr': 8.417203383936534e-06, 'samples': 23564800, 'steps': 46024, 'loss/train': 1.424389362335205} 02/26/2022 13:11:54 - INFO - codeparrot_training - Step 46025: {'lr': 8.412993816180687e-06, 'samples': 23565312, 'steps': 46025, 'loss/train': 2.0163636207580566} 02/26/2022 13:11:58 - INFO - codeparrot_training - Step 46026: {'lr': 8.40878528330688e-06, 'samples': 23565824, 'steps': 46026, 'loss/train': 1.8338297605514526} 02/26/2022 13:12:03 - INFO - codeparrot_training - Step 46027: {'lr': 8.4045777853331e-06, 'samples': 23566336, 'steps': 46027, 'loss/train': 2.553800582885742} 02/26/2022 13:12:07 - INFO - codeparrot_training - Step 46028: {'lr': 8.400371322277362e-06, 'samples': 23566848, 'steps': 46028, 'loss/train': 1.0743021965026855} 02/26/2022 13:12:12 - INFO - codeparrot_training - Step 46029: {'lr': 8.39616589415776e-06, 'samples': 23567360, 'steps': 46029, 'loss/train': 1.1663625240325928} 02/26/2022 13:12:16 - INFO - codeparrot_training - Step 46030: {'lr': 8.391961500992252e-06, 'samples': 23567872, 'steps': 46030, 'loss/train': 1.4144514799118042} 02/26/2022 13:12:21 - INFO - codeparrot_training - Step 46031: {'lr': 8.387758142798879e-06, 'samples': 23568384, 'steps': 46031, 'loss/train': 2.026470184326172} 02/26/2022 13:12:25 - INFO - codeparrot_training - Step 46032: {'lr': 8.383555819595601e-06, 'samples': 23568896, 'steps': 46032, 'loss/train': 1.2396329641342163} 02/26/2022 13:12:31 - INFO - codeparrot_training - Step 46033: {'lr': 8.379354531400457e-06, 'samples': 23569408, 'steps': 46033, 'loss/train': 2.0409650802612305} 02/26/2022 13:12:34 - INFO - codeparrot_training - Step 46034: {'lr': 8.375154278231434e-06, 'samples': 23569920, 'steps': 46034, 'loss/train': 1.0311235189437866} 02/26/2022 13:12:40 - INFO - codeparrot_training - Step 46035: {'lr': 8.370955060106573e-06, 'samples': 23570432, 'steps': 46035, 'loss/train': 1.600880742073059} 02/26/2022 13:12:43 - INFO - codeparrot_training - Step 46036: {'lr': 8.366756877043747e-06, 'samples': 23570944, 'steps': 46036, 'loss/train': 1.2611210346221924} 02/26/2022 13:12:49 - INFO - codeparrot_training - Step 46037: {'lr': 8.362559729061025e-06, 'samples': 23571456, 'steps': 46037, 'loss/train': 1.92186439037323} 02/26/2022 13:12:52 - INFO - codeparrot_training - Step 46038: {'lr': 8.358363616176396e-06, 'samples': 23571968, 'steps': 46038, 'loss/train': 1.3052363395690918} 02/26/2022 13:12:58 - INFO - codeparrot_training - Step 46039: {'lr': 8.354168538407787e-06, 'samples': 23572480, 'steps': 46039, 'loss/train': 1.0481584072113037} 02/26/2022 13:13:01 - INFO - codeparrot_training - Step 46040: {'lr': 8.349974495773182e-06, 'samples': 23572992, 'steps': 46040, 'loss/train': 1.6523873805999756} 02/26/2022 13:13:07 - INFO - codeparrot_training - Step 46041: {'lr': 8.34578148829057e-06, 'samples': 23573504, 'steps': 46041, 'loss/train': 2.1606557369232178} 02/26/2022 13:13:10 - INFO - codeparrot_training - Step 46042: {'lr': 8.341589515977882e-06, 'samples': 23574016, 'steps': 46042, 'loss/train': 1.7149804830551147} 02/26/2022 13:13:16 - INFO - codeparrot_training - Step 46043: {'lr': 8.3373985788531e-06, 'samples': 23574528, 'steps': 46043, 'loss/train': 2.0021350383758545} 02/26/2022 13:13:20 - INFO - codeparrot_training - Step 46044: {'lr': 8.333208676934156e-06, 'samples': 23575040, 'steps': 46044, 'loss/train': 1.8283710479736328} 02/26/2022 13:13:25 - INFO - codeparrot_training - Step 46045: {'lr': 8.329019810239008e-06, 'samples': 23575552, 'steps': 46045, 'loss/train': 1.4947823286056519} 02/26/2022 13:13:29 - INFO - codeparrot_training - Step 46046: {'lr': 8.324831978785614e-06, 'samples': 23576064, 'steps': 46046, 'loss/train': 4.915385723114014} 02/26/2022 13:13:34 - INFO - codeparrot_training - Step 46047: {'lr': 8.320645182591875e-06, 'samples': 23576576, 'steps': 46047, 'loss/train': 0.7544673681259155} 02/26/2022 13:13:38 - INFO - codeparrot_training - Step 46048: {'lr': 8.316459421675776e-06, 'samples': 23577088, 'steps': 46048, 'loss/train': 1.2011456489562988} 02/26/2022 13:13:43 - INFO - codeparrot_training - Step 46049: {'lr': 8.312274696055195e-06, 'samples': 23577600, 'steps': 46049, 'loss/train': 1.645440697669983} 02/26/2022 13:13:47 - INFO - codeparrot_training - Step 46050: {'lr': 8.308091005748114e-06, 'samples': 23578112, 'steps': 46050, 'loss/train': 1.871660828590393} 02/26/2022 13:13:52 - INFO - codeparrot_training - Step 46051: {'lr': 8.30390835077241e-06, 'samples': 23578624, 'steps': 46051, 'loss/train': 2.5569796562194824} 02/26/2022 13:13:56 - INFO - codeparrot_training - Step 46052: {'lr': 8.29972673114604e-06, 'samples': 23579136, 'steps': 46052, 'loss/train': 1.9057695865631104} 02/26/2022 13:14:02 - INFO - codeparrot_training - Step 46053: {'lr': 8.295546146886851e-06, 'samples': 23579648, 'steps': 46053, 'loss/train': 1.3864293098449707} 02/26/2022 13:14:05 - INFO - codeparrot_training - Step 46054: {'lr': 8.291366598012828e-06, 'samples': 23580160, 'steps': 46054, 'loss/train': 1.3967887163162231} 02/26/2022 13:14:11 - INFO - codeparrot_training - Step 46055: {'lr': 8.28718808454179e-06, 'samples': 23580672, 'steps': 46055, 'loss/train': 2.090911388397217} 02/26/2022 13:14:14 - INFO - codeparrot_training - Step 46056: {'lr': 8.283010606491752e-06, 'samples': 23581184, 'steps': 46056, 'loss/train': 1.7933493852615356} 02/26/2022 13:14:21 - INFO - codeparrot_training - Step 46057: {'lr': 8.278834163880505e-06, 'samples': 23581696, 'steps': 46057, 'loss/train': 2.135993719100952} 02/26/2022 13:14:24 - INFO - codeparrot_training - Step 46058: {'lr': 8.274658756725978e-06, 'samples': 23582208, 'steps': 46058, 'loss/train': 1.618568778038025} 02/26/2022 13:14:30 - INFO - codeparrot_training - Step 46059: {'lr': 8.270484385046046e-06, 'samples': 23582720, 'steps': 46059, 'loss/train': 1.7424323558807373} 02/26/2022 13:14:33 - INFO - codeparrot_training - Step 46060: {'lr': 8.26631104885861e-06, 'samples': 23583232, 'steps': 46060, 'loss/train': 2.4552671909332275} 02/26/2022 13:14:39 - INFO - codeparrot_training - Step 46061: {'lr': 8.262138748181547e-06, 'samples': 23583744, 'steps': 46061, 'loss/train': 1.168294906616211} 02/26/2022 13:14:42 - INFO - codeparrot_training - Step 46062: {'lr': 8.25796748303273e-06, 'samples': 23584256, 'steps': 46062, 'loss/train': 1.5917696952819824} 02/26/2022 13:14:48 - INFO - codeparrot_training - Step 46063: {'lr': 8.253797253429978e-06, 'samples': 23584768, 'steps': 46063, 'loss/train': 2.327592134475708} 02/26/2022 13:14:51 - INFO - codeparrot_training - Step 46064: {'lr': 8.249628059391251e-06, 'samples': 23585280, 'steps': 46064, 'loss/train': 0.6519719958305359} 02/26/2022 13:14:57 - INFO - codeparrot_training - Step 46065: {'lr': 8.24545990093431e-06, 'samples': 23585792, 'steps': 46065, 'loss/train': 1.8860714435577393} 02/26/2022 13:15:01 - INFO - codeparrot_training - Step 46066: {'lr': 8.241292778077086e-06, 'samples': 23586304, 'steps': 46066, 'loss/train': 2.0867278575897217} 02/26/2022 13:15:06 - INFO - codeparrot_training - Step 46067: {'lr': 8.23712669083737e-06, 'samples': 23586816, 'steps': 46067, 'loss/train': 0.8177598118782043} 02/26/2022 13:15:10 - INFO - codeparrot_training - Step 46068: {'lr': 8.232961639233038e-06, 'samples': 23587328, 'steps': 46068, 'loss/train': 2.596421718597412} 02/26/2022 13:15:16 - INFO - codeparrot_training - Step 46069: {'lr': 8.228797623281908e-06, 'samples': 23587840, 'steps': 46069, 'loss/train': 3.172916889190674} 02/26/2022 13:15:20 - INFO - codeparrot_training - Step 46070: {'lr': 8.22463464300191e-06, 'samples': 23588352, 'steps': 46070, 'loss/train': 1.8660672903060913} 02/26/2022 13:15:25 - INFO - codeparrot_training - Step 46071: {'lr': 8.220472698410753e-06, 'samples': 23588864, 'steps': 46071, 'loss/train': 1.1300395727157593} 02/26/2022 13:15:29 - INFO - codeparrot_training - Step 46072: {'lr': 8.21631178952631e-06, 'samples': 23589376, 'steps': 46072, 'loss/train': 3.1932754516601562} 02/26/2022 13:15:35 - INFO - codeparrot_training - Step 46073: {'lr': 8.212151916366428e-06, 'samples': 23589888, 'steps': 46073, 'loss/train': 2.163679599761963} 02/26/2022 13:15:38 - INFO - codeparrot_training - Step 46074: {'lr': 8.207993078948928e-06, 'samples': 23590400, 'steps': 46074, 'loss/train': 1.4854207038879395} 02/26/2022 13:15:44 - INFO - codeparrot_training - Step 46075: {'lr': 8.203835277291572e-06, 'samples': 23590912, 'steps': 46075, 'loss/train': 0.3289591073989868} 02/26/2022 13:15:47 - INFO - codeparrot_training - Step 46076: {'lr': 8.199678511412234e-06, 'samples': 23591424, 'steps': 46076, 'loss/train': 1.9764361381530762} 02/26/2022 13:15:53 - INFO - codeparrot_training - Step 46077: {'lr': 8.195522781328652e-06, 'samples': 23591936, 'steps': 46077, 'loss/train': 1.4759719371795654} 02/26/2022 13:15:56 - INFO - codeparrot_training - Step 46078: {'lr': 8.191368087058754e-06, 'samples': 23592448, 'steps': 46078, 'loss/train': 1.5361700057983398} 02/26/2022 13:16:02 - INFO - codeparrot_training - Step 46079: {'lr': 8.187214428620166e-06, 'samples': 23592960, 'steps': 46079, 'loss/train': 0.8220481276512146} 02/26/2022 13:16:06 - INFO - codeparrot_training - Step 46080: {'lr': 8.183061806030818e-06, 'samples': 23593472, 'steps': 46080, 'loss/train': 3.4895219802856445} 02/26/2022 13:16:11 - INFO - codeparrot_training - Step 46081: {'lr': 8.178910219308388e-06, 'samples': 23593984, 'steps': 46081, 'loss/train': 0.33559557795524597} 02/26/2022 13:16:15 - INFO - codeparrot_training - Step 46082: {'lr': 8.174759668470783e-06, 'samples': 23594496, 'steps': 46082, 'loss/train': 1.1306363344192505} 02/26/2022 13:16:20 - INFO - codeparrot_training - Step 46083: {'lr': 8.170610153535708e-06, 'samples': 23595008, 'steps': 46083, 'loss/train': 0.44970691204071045} 02/26/2022 13:16:24 - INFO - codeparrot_training - Step 46084: {'lr': 8.166461674520953e-06, 'samples': 23595520, 'steps': 46084, 'loss/train': 0.8695681691169739} 02/26/2022 13:16:29 - INFO - codeparrot_training - Step 46085: {'lr': 8.162314231444256e-06, 'samples': 23596032, 'steps': 46085, 'loss/train': 3.0058653354644775} 02/26/2022 13:16:33 - INFO - codeparrot_training - Step 46086: {'lr': 8.158167824323408e-06, 'samples': 23596544, 'steps': 46086, 'loss/train': 1.5750559568405151} 02/26/2022 13:16:39 - INFO - codeparrot_training - Step 46087: {'lr': 8.154022453176202e-06, 'samples': 23597056, 'steps': 46087, 'loss/train': 1.1741547584533691} 02/26/2022 13:16:42 - INFO - codeparrot_training - Step 46088: {'lr': 8.149878118020371e-06, 'samples': 23597568, 'steps': 46088, 'loss/train': 2.2009966373443604} 02/26/2022 13:16:48 - INFO - codeparrot_training - Step 46089: {'lr': 8.14573481887368e-06, 'samples': 23598080, 'steps': 46089, 'loss/train': 1.8772943019866943} 02/26/2022 13:16:52 - INFO - codeparrot_training - Step 46090: {'lr': 8.141592555753807e-06, 'samples': 23598592, 'steps': 46090, 'loss/train': 2.1989333629608154} 02/26/2022 13:16:57 - INFO - codeparrot_training - Step 46091: {'lr': 8.137451328678603e-06, 'samples': 23599104, 'steps': 46091, 'loss/train': 0.9876439571380615} 02/26/2022 13:17:01 - INFO - codeparrot_training - Step 46092: {'lr': 8.133311137665745e-06, 'samples': 23599616, 'steps': 46092, 'loss/train': 0.758363664150238} 02/26/2022 13:17:06 - INFO - codeparrot_training - Step 46093: {'lr': 8.129171982732996e-06, 'samples': 23600128, 'steps': 46093, 'loss/train': 0.9829350113868713} 02/26/2022 13:17:10 - INFO - codeparrot_training - Step 46094: {'lr': 8.125033863898013e-06, 'samples': 23600640, 'steps': 46094, 'loss/train': 1.7901220321655273} 02/26/2022 13:17:15 - INFO - codeparrot_training - Step 46095: {'lr': 8.120896781178639e-06, 'samples': 23601152, 'steps': 46095, 'loss/train': 2.7484991550445557} 02/26/2022 13:17:19 - INFO - codeparrot_training - Step 46096: {'lr': 8.116760734592526e-06, 'samples': 23601664, 'steps': 46096, 'loss/train': 1.8779226541519165} 02/26/2022 13:17:24 - INFO - codeparrot_training - Step 46097: {'lr': 8.112625724157412e-06, 'samples': 23602176, 'steps': 46097, 'loss/train': 2.378857374191284} 02/26/2022 13:17:28 - INFO - codeparrot_training - Step 46098: {'lr': 8.10849174989095e-06, 'samples': 23602688, 'steps': 46098, 'loss/train': 1.6186782121658325} 02/26/2022 13:17:33 - INFO - codeparrot_training - Step 46099: {'lr': 8.104358811810957e-06, 'samples': 23603200, 'steps': 46099, 'loss/train': 2.0097146034240723} 02/26/2022 13:17:36 - INFO - codeparrot_training - Step 46100: {'lr': 8.10022690993506e-06, 'samples': 23603712, 'steps': 46100, 'loss/train': 1.5551592111587524} 02/26/2022 13:17:42 - INFO - codeparrot_training - Step 46101: {'lr': 8.096096044280993e-06, 'samples': 23604224, 'steps': 46101, 'loss/train': 2.4417436122894287} 02/26/2022 13:17:45 - INFO - codeparrot_training - Step 46102: {'lr': 8.091966214866409e-06, 'samples': 23604736, 'steps': 46102, 'loss/train': 1.0414965152740479} 02/26/2022 13:17:51 - INFO - codeparrot_training - Step 46103: {'lr': 8.087837421709043e-06, 'samples': 23605248, 'steps': 46103, 'loss/train': 2.6877174377441406} 02/26/2022 13:17:54 - INFO - codeparrot_training - Step 46104: {'lr': 8.083709664826577e-06, 'samples': 23605760, 'steps': 46104, 'loss/train': 2.2320430278778076} 02/26/2022 13:18:01 - INFO - codeparrot_training - Step 46105: {'lr': 8.079582944236663e-06, 'samples': 23606272, 'steps': 46105, 'loss/train': 2.077634811401367} 02/26/2022 13:18:04 - INFO - codeparrot_training - Step 46106: {'lr': 8.075457259956981e-06, 'samples': 23606784, 'steps': 46106, 'loss/train': 2.1648178100585938} 02/26/2022 13:18:10 - INFO - codeparrot_training - Step 46107: {'lr': 8.071332612005266e-06, 'samples': 23607296, 'steps': 46107, 'loss/train': 1.6990742683410645} 02/26/2022 13:18:13 - INFO - codeparrot_training - Step 46108: {'lr': 8.067209000399089e-06, 'samples': 23607808, 'steps': 46108, 'loss/train': 0.46222829818725586} 02/26/2022 13:18:18 - INFO - codeparrot_training - Step 46109: {'lr': 8.06308642515624e-06, 'samples': 23608320, 'steps': 46109, 'loss/train': 0.5889194011688232} 02/26/2022 13:18:22 - INFO - codeparrot_training - Step 46110: {'lr': 8.058964886294234e-06, 'samples': 23608832, 'steps': 46110, 'loss/train': 1.3687952756881714} 02/26/2022 13:18:27 - INFO - codeparrot_training - Step 46111: {'lr': 8.054844383830834e-06, 'samples': 23609344, 'steps': 46111, 'loss/train': 1.207642674446106} 02/26/2022 13:18:33 - INFO - codeparrot_training - Step 46112: {'lr': 8.050724917783635e-06, 'samples': 23609856, 'steps': 46112, 'loss/train': 1.4088826179504395} 02/26/2022 13:18:36 - INFO - codeparrot_training - Step 46113: {'lr': 8.04660648817035e-06, 'samples': 23610368, 'steps': 46113, 'loss/train': 0.6511171460151672} 02/26/2022 13:18:43 - INFO - codeparrot_training - Step 46114: {'lr': 8.042489095008516e-06, 'samples': 23610880, 'steps': 46114, 'loss/train': 0.82984459400177} 02/26/2022 13:18:46 - INFO - codeparrot_training - Step 46115: {'lr': 8.03837273831584e-06, 'samples': 23611392, 'steps': 46115, 'loss/train': 1.4841128587722778} 02/26/2022 13:18:52 - INFO - codeparrot_training - Step 46116: {'lr': 8.034257418109924e-06, 'samples': 23611904, 'steps': 46116, 'loss/train': 2.2533814907073975} 02/26/2022 13:18:55 - INFO - codeparrot_training - Step 46117: {'lr': 8.030143134408474e-06, 'samples': 23612416, 'steps': 46117, 'loss/train': 2.6550283432006836} 02/26/2022 13:18:59 - INFO - codeparrot_training - Step 46118: {'lr': 8.02602988722903e-06, 'samples': 23612928, 'steps': 46118, 'loss/train': 1.1456936597824097} 02/26/2022 13:19:04 - INFO - codeparrot_training - Step 46119: {'lr': 8.021917676589247e-06, 'samples': 23613440, 'steps': 46119, 'loss/train': 3.120382308959961} 02/26/2022 13:19:08 - INFO - codeparrot_training - Step 46120: {'lr': 8.017806502506692e-06, 'samples': 23613952, 'steps': 46120, 'loss/train': 1.411298155784607} 02/26/2022 13:19:13 - INFO - codeparrot_training - Step 46121: {'lr': 8.013696364999047e-06, 'samples': 23614464, 'steps': 46121, 'loss/train': 1.811667561531067} 02/26/2022 13:19:19 - INFO - codeparrot_training - Step 46122: {'lr': 8.009587264083907e-06, 'samples': 23614976, 'steps': 46122, 'loss/train': 0.86668860912323} 02/26/2022 13:19:22 - INFO - codeparrot_training - Step 46123: {'lr': 8.005479199778842e-06, 'samples': 23615488, 'steps': 46123, 'loss/train': 1.4036000967025757} 02/26/2022 13:19:29 - INFO - codeparrot_training - Step 46124: {'lr': 8.001372172101422e-06, 'samples': 23616000, 'steps': 46124, 'loss/train': 1.817915678024292} 02/26/2022 13:19:32 - INFO - codeparrot_training - Step 46125: {'lr': 7.99726618106933e-06, 'samples': 23616512, 'steps': 46125, 'loss/train': 1.6431174278259277} 02/26/2022 13:19:38 - INFO - codeparrot_training - Step 46126: {'lr': 7.993161226700074e-06, 'samples': 23617024, 'steps': 46126, 'loss/train': 1.3536540269851685} 02/26/2022 13:19:41 - INFO - codeparrot_training - Step 46127: {'lr': 7.98905730901131e-06, 'samples': 23617536, 'steps': 46127, 'loss/train': 2.411864757537842} 02/26/2022 13:19:47 - INFO - codeparrot_training - Step 46128: {'lr': 7.984954428020524e-06, 'samples': 23618048, 'steps': 46128, 'loss/train': 1.6283249855041504} 02/26/2022 13:19:50 - INFO - codeparrot_training - Step 46129: {'lr': 7.980852583745368e-06, 'samples': 23618560, 'steps': 46129, 'loss/train': 2.173253059387207} 02/26/2022 13:19:56 - INFO - codeparrot_training - Step 46130: {'lr': 7.97675177620341e-06, 'samples': 23619072, 'steps': 46130, 'loss/train': 0.7825809717178345} 02/26/2022 13:19:59 - INFO - codeparrot_training - Step 46131: {'lr': 7.972652005412196e-06, 'samples': 23619584, 'steps': 46131, 'loss/train': 1.8228380680084229} 02/26/2022 13:20:05 - INFO - codeparrot_training - Step 46132: {'lr': 7.968553271389261e-06, 'samples': 23620096, 'steps': 46132, 'loss/train': 1.6632410287857056} 02/26/2022 13:20:08 - INFO - codeparrot_training - Step 46133: {'lr': 7.96445557415218e-06, 'samples': 23620608, 'steps': 46133, 'loss/train': 2.0599172115325928} 02/26/2022 13:20:15 - INFO - codeparrot_training - Step 46134: {'lr': 7.960358913718546e-06, 'samples': 23621120, 'steps': 46134, 'loss/train': 1.736124873161316} 02/26/2022 13:20:18 - INFO - codeparrot_training - Step 46135: {'lr': 7.956263290105875e-06, 'samples': 23621632, 'steps': 46135, 'loss/train': 2.062497854232788} 02/26/2022 13:20:24 - INFO - codeparrot_training - Step 46136: {'lr': 7.952168703331708e-06, 'samples': 23622144, 'steps': 46136, 'loss/train': 1.4588398933410645} 02/26/2022 13:20:28 - INFO - codeparrot_training - Step 46137: {'lr': 7.948075153413587e-06, 'samples': 23622656, 'steps': 46137, 'loss/train': 1.1146643161773682} 02/26/2022 13:20:33 - INFO - codeparrot_training - Step 46138: {'lr': 7.943982640369052e-06, 'samples': 23623168, 'steps': 46138, 'loss/train': 1.966606616973877} 02/26/2022 13:20:37 - INFO - codeparrot_training - Step 46139: {'lr': 7.939891164215645e-06, 'samples': 23623680, 'steps': 46139, 'loss/train': 1.8707178831100464} 02/26/2022 13:20:42 - INFO - codeparrot_training - Step 46140: {'lr': 7.935800724970883e-06, 'samples': 23624192, 'steps': 46140, 'loss/train': 1.5159659385681152} 02/26/2022 13:20:46 - INFO - codeparrot_training - Step 46141: {'lr': 7.931711322652274e-06, 'samples': 23624704, 'steps': 46141, 'loss/train': 0.6920977234840393} 02/26/2022 13:20:51 - INFO - codeparrot_training - Step 46142: {'lr': 7.927622957277364e-06, 'samples': 23625216, 'steps': 46142, 'loss/train': 1.78933584690094} 02/26/2022 13:20:54 - INFO - codeparrot_training - Step 46143: {'lr': 7.92353562886361e-06, 'samples': 23625728, 'steps': 46143, 'loss/train': 2.1544649600982666} 02/26/2022 13:21:00 - INFO - codeparrot_training - Step 46144: {'lr': 7.919449337428636e-06, 'samples': 23626240, 'steps': 46144, 'loss/train': 1.3264826536178589} 02/26/2022 13:21:03 - INFO - codeparrot_training - Step 46145: {'lr': 7.91536408298979e-06, 'samples': 23626752, 'steps': 46145, 'loss/train': 0.9895395040512085} 02/26/2022 13:21:09 - INFO - codeparrot_training - Step 46146: {'lr': 7.911279865564697e-06, 'samples': 23627264, 'steps': 46146, 'loss/train': 1.5277233123779297} 02/26/2022 13:21:12 - INFO - codeparrot_training - Step 46147: {'lr': 7.907196685170787e-06, 'samples': 23627776, 'steps': 46147, 'loss/train': 1.7551686763763428} 02/26/2022 13:21:18 - INFO - codeparrot_training - Step 46148: {'lr': 7.903114541825628e-06, 'samples': 23628288, 'steps': 46148, 'loss/train': 1.9766554832458496} 02/26/2022 13:21:22 - INFO - codeparrot_training - Step 46149: {'lr': 7.899033435546598e-06, 'samples': 23628800, 'steps': 46149, 'loss/train': 1.8132158517837524} 02/26/2022 13:21:28 - INFO - codeparrot_training - Step 46150: {'lr': 7.894953366351265e-06, 'samples': 23629312, 'steps': 46150, 'loss/train': 1.2718414068222046} 02/26/2022 13:21:31 - INFO - codeparrot_training - Step 46151: {'lr': 7.890874334257031e-06, 'samples': 23629824, 'steps': 46151, 'loss/train': 2.383247137069702} 02/26/2022 13:21:37 - INFO - codeparrot_training - Step 46152: {'lr': 7.886796339281466e-06, 'samples': 23630336, 'steps': 46152, 'loss/train': 2.277276039123535} 02/26/2022 13:21:40 - INFO - codeparrot_training - Step 46153: {'lr': 7.882719381441944e-06, 'samples': 23630848, 'steps': 46153, 'loss/train': 1.1779136657714844} 02/26/2022 13:21:46 - INFO - codeparrot_training - Step 46154: {'lr': 7.87864346075598e-06, 'samples': 23631360, 'steps': 46154, 'loss/train': 1.4246450662612915} 02/26/2022 13:21:49 - INFO - codeparrot_training - Step 46155: {'lr': 7.87456857724103e-06, 'samples': 23631872, 'steps': 46155, 'loss/train': 0.8637968897819519} 02/26/2022 13:21:55 - INFO - codeparrot_training - Step 46156: {'lr': 7.870494730914556e-06, 'samples': 23632384, 'steps': 46156, 'loss/train': 1.0921019315719604} 02/26/2022 13:21:58 - INFO - codeparrot_training - Step 46157: {'lr': 7.866421921793987e-06, 'samples': 23632896, 'steps': 46157, 'loss/train': 0.554743766784668} 02/26/2022 13:22:04 - INFO - codeparrot_training - Step 46158: {'lr': 7.862350149896779e-06, 'samples': 23633408, 'steps': 46158, 'loss/train': 1.5484811067581177} 02/26/2022 13:22:07 - INFO - codeparrot_training - Step 46159: {'lr': 7.858279415240365e-06, 'samples': 23633920, 'steps': 46159, 'loss/train': 1.6888887882232666} 02/26/2022 13:22:14 - INFO - codeparrot_training - Step 46160: {'lr': 7.854209717842232e-06, 'samples': 23634432, 'steps': 46160, 'loss/train': 1.7405825853347778} 02/26/2022 13:22:17 - INFO - codeparrot_training - Step 46161: {'lr': 7.85014105771975e-06, 'samples': 23634944, 'steps': 46161, 'loss/train': 1.9780075550079346} 02/26/2022 13:22:22 - INFO - codeparrot_training - Step 46162: {'lr': 7.846073434890383e-06, 'samples': 23635456, 'steps': 46162, 'loss/train': 1.694478988647461} 02/26/2022 13:22:26 - INFO - codeparrot_training - Step 46163: {'lr': 7.842006849371502e-06, 'samples': 23635968, 'steps': 46163, 'loss/train': 1.0776216983795166} 02/26/2022 13:22:31 - INFO - codeparrot_training - Step 46164: {'lr': 7.837941301180624e-06, 'samples': 23636480, 'steps': 46164, 'loss/train': 1.905545949935913} 02/26/2022 13:22:35 - INFO - codeparrot_training - Step 46165: {'lr': 7.833876790335093e-06, 'samples': 23636992, 'steps': 46165, 'loss/train': 2.163379192352295} 02/26/2022 13:22:41 - INFO - codeparrot_training - Step 46166: {'lr': 7.829813316852342e-06, 'samples': 23637504, 'steps': 46166, 'loss/train': 0.5601224303245544} 02/26/2022 13:22:44 - INFO - codeparrot_training - Step 46167: {'lr': 7.825750880749743e-06, 'samples': 23638016, 'steps': 46167, 'loss/train': 1.049468994140625} 02/26/2022 13:22:49 - INFO - codeparrot_training - Step 46168: {'lr': 7.821689482044758e-06, 'samples': 23638528, 'steps': 46168, 'loss/train': 0.643880307674408} 02/26/2022 13:22:53 - INFO - codeparrot_training - Step 46169: {'lr': 7.817629120754732e-06, 'samples': 23639040, 'steps': 46169, 'loss/train': 1.6984556913375854} 02/26/2022 13:22:59 - INFO - codeparrot_training - Step 46170: {'lr': 7.813569796897124e-06, 'samples': 23639552, 'steps': 46170, 'loss/train': 0.9831675887107849} 02/26/2022 13:23:03 - INFO - codeparrot_training - Step 46171: {'lr': 7.809511510489226e-06, 'samples': 23640064, 'steps': 46171, 'loss/train': 1.606002688407898} 02/26/2022 13:23:08 - INFO - codeparrot_training - Step 46172: {'lr': 7.805454261548495e-06, 'samples': 23640576, 'steps': 46172, 'loss/train': 2.493654251098633} 02/26/2022 13:23:12 - INFO - codeparrot_training - Step 46173: {'lr': 7.801398050092306e-06, 'samples': 23641088, 'steps': 46173, 'loss/train': 1.751905083656311} 02/26/2022 13:23:17 - INFO - codeparrot_training - Step 46174: {'lr': 7.797342876138037e-06, 'samples': 23641600, 'steps': 46174, 'loss/train': 2.019636869430542} 02/26/2022 13:23:21 - INFO - codeparrot_training - Step 46175: {'lr': 7.793288739703003e-06, 'samples': 23642112, 'steps': 46175, 'loss/train': 1.7807376384735107} 02/26/2022 13:23:27 - INFO - codeparrot_training - Step 46176: {'lr': 7.789235640804637e-06, 'samples': 23642624, 'steps': 46176, 'loss/train': 2.1343982219696045} 02/26/2022 13:23:30 - INFO - codeparrot_training - Step 46177: {'lr': 7.785183579460258e-06, 'samples': 23643136, 'steps': 46177, 'loss/train': 1.0486819744110107} 02/26/2022 13:23:34 - INFO - codeparrot_training - Step 46178: {'lr': 7.781132555687242e-06, 'samples': 23643648, 'steps': 46178, 'loss/train': 1.5688691139221191} 02/26/2022 13:23:39 - INFO - codeparrot_training - Step 46179: {'lr': 7.777082569502964e-06, 'samples': 23644160, 'steps': 46179, 'loss/train': 1.3218811750411987} 02/26/2022 13:23:43 - INFO - codeparrot_training - Step 46180: {'lr': 7.773033620924741e-06, 'samples': 23644672, 'steps': 46180, 'loss/train': 1.9818297624588013} 02/26/2022 13:23:49 - INFO - codeparrot_training - Step 46181: {'lr': 7.768985709969922e-06, 'samples': 23645184, 'steps': 46181, 'loss/train': 1.5427573919296265} 02/26/2022 13:23:52 - INFO - codeparrot_training - Step 46182: {'lr': 7.764938836655828e-06, 'samples': 23645696, 'steps': 46182, 'loss/train': 2.053532838821411} 02/26/2022 13:23:58 - INFO - codeparrot_training - Step 46183: {'lr': 7.760893000999858e-06, 'samples': 23646208, 'steps': 46183, 'loss/train': 0.8363004326820374} 02/26/2022 13:24:01 - INFO - codeparrot_training - Step 46184: {'lr': 7.756848203019279e-06, 'samples': 23646720, 'steps': 46184, 'loss/train': 2.4321486949920654} 02/26/2022 13:24:07 - INFO - codeparrot_training - Step 46185: {'lr': 7.752804442731437e-06, 'samples': 23647232, 'steps': 46185, 'loss/train': 1.4808934926986694} 02/26/2022 13:24:10 - INFO - codeparrot_training - Step 46186: {'lr': 7.748761720153652e-06, 'samples': 23647744, 'steps': 46186, 'loss/train': 2.048483371734619} 02/26/2022 13:24:16 - INFO - codeparrot_training - Step 46187: {'lr': 7.744720035303298e-06, 'samples': 23648256, 'steps': 46187, 'loss/train': 2.595313787460327} 02/26/2022 13:24:19 - INFO - codeparrot_training - Step 46188: {'lr': 7.740679388197586e-06, 'samples': 23648768, 'steps': 46188, 'loss/train': 0.47151291370391846} 02/26/2022 13:24:25 - INFO - codeparrot_training - Step 46189: {'lr': 7.736639778853888e-06, 'samples': 23649280, 'steps': 46189, 'loss/train': 0.3350508511066437} 02/26/2022 13:24:30 - INFO - codeparrot_training - Step 46190: {'lr': 7.732601207289497e-06, 'samples': 23649792, 'steps': 46190, 'loss/train': 2.1283323764801025} 02/26/2022 13:24:34 - INFO - codeparrot_training - Step 46191: {'lr': 7.728563673521732e-06, 'samples': 23650304, 'steps': 46191, 'loss/train': 2.0503880977630615} 02/26/2022 13:24:39 - INFO - codeparrot_training - Step 46192: {'lr': 7.724527177567858e-06, 'samples': 23650816, 'steps': 46192, 'loss/train': 2.233004093170166} 02/26/2022 13:24:43 - INFO - codeparrot_training - Step 46193: {'lr': 7.720491719445195e-06, 'samples': 23651328, 'steps': 46193, 'loss/train': 0.49077585339546204} 02/26/2022 13:24:48 - INFO - codeparrot_training - Step 46194: {'lr': 7.716457299170975e-06, 'samples': 23651840, 'steps': 46194, 'loss/train': 1.1692137718200684} 02/26/2022 13:24:52 - INFO - codeparrot_training - Step 46195: {'lr': 7.712423916762552e-06, 'samples': 23652352, 'steps': 46195, 'loss/train': 1.0794163942337036} 02/26/2022 13:24:58 - INFO - codeparrot_training - Step 46196: {'lr': 7.708391572237183e-06, 'samples': 23652864, 'steps': 46196, 'loss/train': 1.959424614906311} 02/26/2022 13:25:02 - INFO - codeparrot_training - Step 46197: {'lr': 7.70436026561211e-06, 'samples': 23653376, 'steps': 46197, 'loss/train': 1.2503868341445923} 02/26/2022 13:25:07 - INFO - codeparrot_training - Step 46198: {'lr': 7.700329996904593e-06, 'samples': 23653888, 'steps': 46198, 'loss/train': 2.1610403060913086} 02/26/2022 13:25:11 - INFO - codeparrot_training - Step 46199: {'lr': 7.696300766131952e-06, 'samples': 23654400, 'steps': 46199, 'loss/train': 1.25043523311615} 02/26/2022 13:25:16 - INFO - codeparrot_training - Step 46200: {'lr': 7.692272573311426e-06, 'samples': 23654912, 'steps': 46200, 'loss/train': 1.3419629335403442} 02/26/2022 13:25:20 - INFO - codeparrot_training - Step 46201: {'lr': 7.68824541846025e-06, 'samples': 23655424, 'steps': 46201, 'loss/train': 2.131420373916626} 02/26/2022 13:25:25 - INFO - codeparrot_training - Step 46202: {'lr': 7.684219301595658e-06, 'samples': 23655936, 'steps': 46202, 'loss/train': 2.698681116104126} 02/26/2022 13:25:28 - INFO - codeparrot_training - Step 46203: {'lr': 7.68019422273497e-06, 'samples': 23656448, 'steps': 46203, 'loss/train': 1.2755769491195679} 02/26/2022 13:25:34 - INFO - codeparrot_training - Step 46204: {'lr': 7.676170181895342e-06, 'samples': 23656960, 'steps': 46204, 'loss/train': 2.321380376815796} 02/26/2022 13:25:37 - INFO - codeparrot_training - Step 46205: {'lr': 7.67214717909412e-06, 'samples': 23657472, 'steps': 46205, 'loss/train': 1.3493386507034302} 02/26/2022 13:25:44 - INFO - codeparrot_training - Step 46206: {'lr': 7.668125214348427e-06, 'samples': 23657984, 'steps': 46206, 'loss/train': 1.7210662364959717} 02/26/2022 13:25:47 - INFO - codeparrot_training - Step 46207: {'lr': 7.664104287675555e-06, 'samples': 23658496, 'steps': 46207, 'loss/train': 1.959551215171814} 02/26/2022 13:25:52 - INFO - codeparrot_training - Step 46208: {'lr': 7.660084399092659e-06, 'samples': 23659008, 'steps': 46208, 'loss/train': 1.615695595741272} 02/26/2022 13:25:56 - INFO - codeparrot_training - Step 46209: {'lr': 7.656065548617086e-06, 'samples': 23659520, 'steps': 46209, 'loss/train': 1.3801510334014893} 02/26/2022 13:26:02 - INFO - codeparrot_training - Step 46210: {'lr': 7.652047736265933e-06, 'samples': 23660032, 'steps': 46210, 'loss/train': 1.4071534872055054} 02/26/2022 13:26:05 - INFO - codeparrot_training - Step 46211: {'lr': 7.648030962056463e-06, 'samples': 23660544, 'steps': 46211, 'loss/train': 1.8339293003082275} 02/26/2022 13:26:11 - INFO - codeparrot_training - Step 46212: {'lr': 7.644015226005857e-06, 'samples': 23661056, 'steps': 46212, 'loss/train': 0.5704971551895142} 02/26/2022 13:26:14 - INFO - codeparrot_training - Step 46213: {'lr': 7.64000052813138e-06, 'samples': 23661568, 'steps': 46213, 'loss/train': 1.8081496953964233} 02/26/2022 13:26:20 - INFO - codeparrot_training - Step 46214: {'lr': 7.635986868450129e-06, 'samples': 23662080, 'steps': 46214, 'loss/train': 1.4197722673416138} 02/26/2022 13:26:23 - INFO - codeparrot_training - Step 46215: {'lr': 7.631974246979367e-06, 'samples': 23662592, 'steps': 46215, 'loss/train': 1.6627098321914673} 02/26/2022 13:26:28 - INFO - codeparrot_training - Step 46216: {'lr': 7.627962663736277e-06, 'samples': 23663104, 'steps': 46216, 'loss/train': 2.077439308166504} 02/26/2022 13:26:32 - INFO - codeparrot_training - Step 46217: {'lr': 7.623952118738009e-06, 'samples': 23663616, 'steps': 46217, 'loss/train': 2.4660086631774902} 02/26/2022 13:26:38 - INFO - codeparrot_training - Step 46218: {'lr': 7.619942612001801e-06, 'samples': 23664128, 'steps': 46218, 'loss/train': 1.0642417669296265} 02/26/2022 13:26:42 - INFO - codeparrot_training - Step 46219: {'lr': 7.615934143544778e-06, 'samples': 23664640, 'steps': 46219, 'loss/train': 2.13999342918396} 02/26/2022 13:26:47 - INFO - codeparrot_training - Step 46220: {'lr': 7.611926713384121e-06, 'samples': 23665152, 'steps': 46220, 'loss/train': 1.95249342918396} 02/26/2022 13:26:51 - INFO - codeparrot_training - Step 46221: {'lr': 7.6079203215370095e-06, 'samples': 23665664, 'steps': 46221, 'loss/train': 1.711698055267334} 02/26/2022 13:26:56 - INFO - codeparrot_training - Step 46222: {'lr': 7.6039149680205975e-06, 'samples': 23666176, 'steps': 46222, 'loss/train': 1.4104186296463013} 02/26/2022 13:27:00 - INFO - codeparrot_training - Step 46223: {'lr': 7.599910652852066e-06, 'samples': 23666688, 'steps': 46223, 'loss/train': 2.186832904815674} 02/26/2022 13:27:05 - INFO - codeparrot_training - Step 46224: {'lr': 7.5959073760485105e-06, 'samples': 23667200, 'steps': 46224, 'loss/train': 2.075455665588379} 02/26/2022 13:27:09 - INFO - codeparrot_training - Step 46225: {'lr': 7.591905137627114e-06, 'samples': 23667712, 'steps': 46225, 'loss/train': 1.6174030303955078} 02/26/2022 13:27:14 - INFO - codeparrot_training - Step 46226: {'lr': 7.587903937605056e-06, 'samples': 23668224, 'steps': 46226, 'loss/train': 1.9038349390029907} 02/26/2022 13:27:18 - INFO - codeparrot_training - Step 46227: {'lr': 7.583903775999435e-06, 'samples': 23668736, 'steps': 46227, 'loss/train': 2.4301764965057373} 02/26/2022 13:27:24 - INFO - codeparrot_training - Step 46228: {'lr': 7.579904652827402e-06, 'samples': 23669248, 'steps': 46228, 'loss/train': 1.7941884994506836} 02/26/2022 13:27:27 - INFO - codeparrot_training - Step 46229: {'lr': 7.575906568106028e-06, 'samples': 23669760, 'steps': 46229, 'loss/train': 2.422224283218384} 02/26/2022 13:27:33 - INFO - codeparrot_training - Step 46230: {'lr': 7.571909521852549e-06, 'samples': 23670272, 'steps': 46230, 'loss/train': 1.4480489492416382} 02/26/2022 13:27:36 - INFO - codeparrot_training - Step 46231: {'lr': 7.5679135140840076e-06, 'samples': 23670784, 'steps': 46231, 'loss/train': 0.734429657459259} 02/26/2022 13:27:42 - INFO - codeparrot_training - Step 46232: {'lr': 7.563918544817555e-06, 'samples': 23671296, 'steps': 46232, 'loss/train': 3.3537869453430176} 02/26/2022 13:27:45 - INFO - codeparrot_training - Step 46233: {'lr': 7.559924614070263e-06, 'samples': 23671808, 'steps': 46233, 'loss/train': 0.9917415976524353} 02/26/2022 13:27:51 - INFO - codeparrot_training - Step 46234: {'lr': 7.555931721859311e-06, 'samples': 23672320, 'steps': 46234, 'loss/train': 1.5242668390274048} 02/26/2022 13:27:54 - INFO - codeparrot_training - Step 46235: {'lr': 7.55193986820174e-06, 'samples': 23672832, 'steps': 46235, 'loss/train': 1.3739936351776123} 02/26/2022 13:28:00 - INFO - codeparrot_training - Step 46236: {'lr': 7.547949053114705e-06, 'samples': 23673344, 'steps': 46236, 'loss/train': 2.2427783012390137} 02/26/2022 13:28:03 - INFO - codeparrot_training - Step 46237: {'lr': 7.54395927661522e-06, 'samples': 23673856, 'steps': 46237, 'loss/train': 2.017179250717163} 02/26/2022 13:28:09 - INFO - codeparrot_training - Step 46238: {'lr': 7.5399705387204365e-06, 'samples': 23674368, 'steps': 46238, 'loss/train': 0.41155701875686646} 02/26/2022 13:28:12 - INFO - codeparrot_training - Step 46239: {'lr': 7.535982839447425e-06, 'samples': 23674880, 'steps': 46239, 'loss/train': 0.9758650660514832} 02/26/2022 13:28:18 - INFO - codeparrot_training - Step 46240: {'lr': 7.53199617881331e-06, 'samples': 23675392, 'steps': 46240, 'loss/train': 1.9534239768981934} 02/26/2022 13:28:21 - INFO - codeparrot_training - Step 46241: {'lr': 7.528010556835108e-06, 'samples': 23675904, 'steps': 46241, 'loss/train': 1.3885554075241089} 02/26/2022 13:28:27 - INFO - codeparrot_training - Step 46242: {'lr': 7.5240259735299135e-06, 'samples': 23676416, 'steps': 46242, 'loss/train': 1.9461522102355957} 02/26/2022 13:28:30 - INFO - codeparrot_training - Step 46243: {'lr': 7.5200424289147695e-06, 'samples': 23676928, 'steps': 46243, 'loss/train': 2.7684545516967773} 02/26/2022 13:28:37 - INFO - codeparrot_training - Step 46244: {'lr': 7.516059923006829e-06, 'samples': 23677440, 'steps': 46244, 'loss/train': 2.2178447246551514} 02/26/2022 13:28:40 - INFO - codeparrot_training - Step 46245: {'lr': 7.512078455823052e-06, 'samples': 23677952, 'steps': 46245, 'loss/train': 1.61265230178833} 02/26/2022 13:28:46 - INFO - codeparrot_training - Step 46246: {'lr': 7.508098027380534e-06, 'samples': 23678464, 'steps': 46246, 'loss/train': 1.355047345161438} 02/26/2022 13:28:49 - INFO - codeparrot_training - Step 46247: {'lr': 7.504118637696317e-06, 'samples': 23678976, 'steps': 46247, 'loss/train': 1.4046990871429443} 02/26/2022 13:28:55 - INFO - codeparrot_training - Step 46248: {'lr': 7.5001402867874995e-06, 'samples': 23679488, 'steps': 46248, 'loss/train': 1.0004087686538696} 02/26/2022 13:28:58 - INFO - codeparrot_training - Step 46249: {'lr': 7.496162974671039e-06, 'samples': 23680000, 'steps': 46249, 'loss/train': 2.097203016281128} 02/26/2022 13:29:04 - INFO - codeparrot_training - Step 46250: {'lr': 7.4921867013640064e-06, 'samples': 23680512, 'steps': 46250, 'loss/train': 0.35356733202934265} 02/26/2022 13:29:07 - INFO - codeparrot_training - Step 46251: {'lr': 7.488211466883443e-06, 'samples': 23681024, 'steps': 46251, 'loss/train': 1.7939902544021606} 02/26/2022 13:29:13 - INFO - codeparrot_training - Step 46252: {'lr': 7.484237271246391e-06, 'samples': 23681536, 'steps': 46252, 'loss/train': 1.8993935585021973} 02/26/2022 13:29:16 - INFO - codeparrot_training - Step 46253: {'lr': 7.480264114469865e-06, 'samples': 23682048, 'steps': 46253, 'loss/train': 1.6327611207962036} 02/26/2022 13:29:22 - INFO - codeparrot_training - Step 46254: {'lr': 7.476291996570877e-06, 'samples': 23682560, 'steps': 46254, 'loss/train': 0.8919408321380615} 02/26/2022 13:29:26 - INFO - codeparrot_training - Step 46255: {'lr': 7.472320917566416e-06, 'samples': 23683072, 'steps': 46255, 'loss/train': 1.2883343696594238} 02/26/2022 13:29:31 - INFO - codeparrot_training - Step 46256: {'lr': 7.468350877473551e-06, 'samples': 23683584, 'steps': 46256, 'loss/train': 0.9894641637802124} 02/26/2022 13:29:35 - INFO - codeparrot_training - Step 46257: {'lr': 7.46438187630924e-06, 'samples': 23684096, 'steps': 46257, 'loss/train': 1.46388840675354} 02/26/2022 13:29:40 - INFO - codeparrot_training - Step 46258: {'lr': 7.460413914090525e-06, 'samples': 23684608, 'steps': 46258, 'loss/train': 0.6810938119888306} 02/26/2022 13:29:44 - INFO - codeparrot_training - Step 46259: {'lr': 7.456446990834337e-06, 'samples': 23685120, 'steps': 46259, 'loss/train': 1.031083106994629} 02/26/2022 13:29:49 - INFO - codeparrot_training - Step 46260: {'lr': 7.452481106557746e-06, 'samples': 23685632, 'steps': 46260, 'loss/train': 1.3681429624557495} 02/26/2022 13:29:53 - INFO - codeparrot_training - Step 46261: {'lr': 7.44851626127771e-06, 'samples': 23686144, 'steps': 46261, 'loss/train': 0.2631775140762329} 02/26/2022 13:29:58 - INFO - codeparrot_training - Step 46262: {'lr': 7.444552455011216e-06, 'samples': 23686656, 'steps': 46262, 'loss/train': 1.740143060684204} 02/26/2022 13:30:02 - INFO - codeparrot_training - Step 46263: {'lr': 7.440589687775251e-06, 'samples': 23687168, 'steps': 46263, 'loss/train': 1.8481998443603516} 02/26/2022 13:30:08 - INFO - codeparrot_training - Step 46264: {'lr': 7.436627959586744e-06, 'samples': 23687680, 'steps': 46264, 'loss/train': 3.0240542888641357} 02/26/2022 13:30:12 - INFO - codeparrot_training - Step 46265: {'lr': 7.432667270462712e-06, 'samples': 23688192, 'steps': 46265, 'loss/train': 1.746176838874817} 02/26/2022 13:30:17 - INFO - codeparrot_training - Step 46266: {'lr': 7.428707620420138e-06, 'samples': 23688704, 'steps': 46266, 'loss/train': 0.054960038512945175} 02/26/2022 13:30:21 - INFO - codeparrot_training - Step 46267: {'lr': 7.424749009475929e-06, 'samples': 23689216, 'steps': 46267, 'loss/train': 1.7184443473815918} 02/26/2022 13:30:26 - INFO - codeparrot_training - Step 46268: {'lr': 7.4207914376470395e-06, 'samples': 23689728, 'steps': 46268, 'loss/train': 1.4960073232650757} 02/26/2022 13:30:30 - INFO - codeparrot_training - Step 46269: {'lr': 7.4168349049504864e-06, 'samples': 23690240, 'steps': 46269, 'loss/train': 2.735445499420166} 02/26/2022 13:30:35 - INFO - codeparrot_training - Step 46270: {'lr': 7.4128794114032e-06, 'samples': 23690752, 'steps': 46270, 'loss/train': 2.057077407836914} 02/26/2022 13:30:39 - INFO - codeparrot_training - Step 46271: {'lr': 7.408924957022084e-06, 'samples': 23691264, 'steps': 46271, 'loss/train': 1.2970287799835205} 02/26/2022 13:30:44 - INFO - codeparrot_training - Step 46272: {'lr': 7.4049715418240955e-06, 'samples': 23691776, 'steps': 46272, 'loss/train': 1.7952051162719727} 02/26/2022 13:30:48 - INFO - codeparrot_training - Step 46273: {'lr': 7.401019165826195e-06, 'samples': 23692288, 'steps': 46273, 'loss/train': 1.2831530570983887} 02/26/2022 13:30:54 - INFO - codeparrot_training - Step 46274: {'lr': 7.397067829045284e-06, 'samples': 23692800, 'steps': 46274, 'loss/train': 1.8026372194290161} 02/26/2022 13:30:57 - INFO - codeparrot_training - Step 46275: {'lr': 7.393117531498322e-06, 'samples': 23693312, 'steps': 46275, 'loss/train': 2.117004632949829} 02/26/2022 13:31:03 - INFO - codeparrot_training - Step 46276: {'lr': 7.389168273202157e-06, 'samples': 23693824, 'steps': 46276, 'loss/train': 1.5674726963043213} 02/26/2022 13:31:06 - INFO - codeparrot_training - Step 46277: {'lr': 7.385220054173802e-06, 'samples': 23694336, 'steps': 46277, 'loss/train': 0.9597045183181763} 02/26/2022 13:31:12 - INFO - codeparrot_training - Step 46278: {'lr': 7.381272874430078e-06, 'samples': 23694848, 'steps': 46278, 'loss/train': 1.9084405899047852} 02/26/2022 13:31:15 - INFO - codeparrot_training - Step 46279: {'lr': 7.377326733988027e-06, 'samples': 23695360, 'steps': 46279, 'loss/train': 1.8546720743179321} 02/26/2022 13:31:21 - INFO - codeparrot_training - Step 46280: {'lr': 7.373381632864384e-06, 'samples': 23695872, 'steps': 46280, 'loss/train': 1.6214810609817505} 02/26/2022 13:31:24 - INFO - codeparrot_training - Step 46281: {'lr': 7.369437571076165e-06, 'samples': 23696384, 'steps': 46281, 'loss/train': 2.410975933074951} 02/26/2022 13:31:30 - INFO - codeparrot_training - Step 46282: {'lr': 7.365494548640217e-06, 'samples': 23696896, 'steps': 46282, 'loss/train': 1.1699903011322021} 02/26/2022 13:31:33 - INFO - codeparrot_training - Step 46283: {'lr': 7.361552565573499e-06, 'samples': 23697408, 'steps': 46283, 'loss/train': 1.9359527826309204} 02/26/2022 13:31:39 - INFO - codeparrot_training - Step 46284: {'lr': 7.357611621892774e-06, 'samples': 23697920, 'steps': 46284, 'loss/train': 1.2825313806533813} 02/26/2022 13:31:44 - INFO - codeparrot_training - Step 46285: {'lr': 7.353671717615029e-06, 'samples': 23698432, 'steps': 46285, 'loss/train': 0.47180798649787903} 02/26/2022 13:31:48 - INFO - codeparrot_training - Step 46286: {'lr': 7.349732852757085e-06, 'samples': 23698944, 'steps': 46286, 'loss/train': 1.7396972179412842} 02/26/2022 13:31:53 - INFO - codeparrot_training - Step 46287: {'lr': 7.345795027335844e-06, 'samples': 23699456, 'steps': 46287, 'loss/train': 2.2069430351257324} 02/26/2022 13:31:57 - INFO - codeparrot_training - Step 46288: {'lr': 7.341858241368182e-06, 'samples': 23699968, 'steps': 46288, 'loss/train': 1.7909590005874634} 02/26/2022 13:32:03 - INFO - codeparrot_training - Step 46289: {'lr': 7.3379224948709465e-06, 'samples': 23700480, 'steps': 46289, 'loss/train': 2.207418441772461} 02/26/2022 13:32:06 - INFO - codeparrot_training - Step 46290: {'lr': 7.333987787860958e-06, 'samples': 23700992, 'steps': 46290, 'loss/train': 0.9212024211883545} 02/26/2022 13:32:12 - INFO - codeparrot_training - Step 46291: {'lr': 7.330054120355146e-06, 'samples': 23701504, 'steps': 46291, 'loss/train': 2.1827926635742188} 02/26/2022 13:32:15 - INFO - codeparrot_training - Step 46292: {'lr': 7.32612149237033e-06, 'samples': 23702016, 'steps': 46292, 'loss/train': 0.8848879337310791} 02/26/2022 13:32:21 - INFO - codeparrot_training - Step 46293: {'lr': 7.322189903923332e-06, 'samples': 23702528, 'steps': 46293, 'loss/train': 2.041585922241211} 02/26/2022 13:32:24 - INFO - codeparrot_training - Step 46294: {'lr': 7.318259355030999e-06, 'samples': 23703040, 'steps': 46294, 'loss/train': 2.701075792312622} 02/26/2022 13:32:30 - INFO - codeparrot_training - Step 46295: {'lr': 7.314329845710205e-06, 'samples': 23703552, 'steps': 46295, 'loss/train': 1.5426957607269287} 02/26/2022 13:32:33 - INFO - codeparrot_training - Step 46296: {'lr': 7.310401375977771e-06, 'samples': 23704064, 'steps': 46296, 'loss/train': 1.7253940105438232} 02/26/2022 13:32:39 - INFO - codeparrot_training - Step 46297: {'lr': 7.306473945850517e-06, 'samples': 23704576, 'steps': 46297, 'loss/train': 0.8173562288284302} 02/26/2022 13:32:42 - INFO - codeparrot_training - Step 46298: {'lr': 7.302547555345234e-06, 'samples': 23705088, 'steps': 46298, 'loss/train': 0.8185781240463257} 02/26/2022 13:32:48 - INFO - codeparrot_training - Step 46299: {'lr': 7.2986222044788e-06, 'samples': 23705600, 'steps': 46299, 'loss/train': 2.2433414459228516} 02/26/2022 13:32:52 - INFO - codeparrot_training - Step 46300: {'lr': 7.294697893267976e-06, 'samples': 23706112, 'steps': 46300, 'loss/train': 1.1419854164123535} 02/26/2022 13:32:57 - INFO - codeparrot_training - Step 46301: {'lr': 7.29077462172964e-06, 'samples': 23706624, 'steps': 46301, 'loss/train': 4.863748073577881} 02/26/2022 13:33:01 - INFO - codeparrot_training - Step 46302: {'lr': 7.2868523898805e-06, 'samples': 23707136, 'steps': 46302, 'loss/train': 1.9545962810516357} 02/26/2022 13:33:06 - INFO - codeparrot_training - Step 46303: {'lr': 7.282931197737458e-06, 'samples': 23707648, 'steps': 46303, 'loss/train': 2.0479788780212402} 02/26/2022 13:33:10 - INFO - codeparrot_training - Step 46304: {'lr': 7.279011045317252e-06, 'samples': 23708160, 'steps': 46304, 'loss/train': 0.6766265034675598} 02/26/2022 13:33:15 - INFO - codeparrot_training - Step 46305: {'lr': 7.275091932636702e-06, 'samples': 23708672, 'steps': 46305, 'loss/train': 0.6496062874794006} 02/26/2022 13:33:19 - INFO - codeparrot_training - Step 46306: {'lr': 7.271173859712571e-06, 'samples': 23709184, 'steps': 46306, 'loss/train': 0.18399237096309662} 02/26/2022 13:33:25 - INFO - codeparrot_training - Step 46307: {'lr': 7.267256826561652e-06, 'samples': 23709696, 'steps': 46307, 'loss/train': 1.5856776237487793} 02/26/2022 13:33:29 - INFO - codeparrot_training - Step 46308: {'lr': 7.263340833200738e-06, 'samples': 23710208, 'steps': 46308, 'loss/train': 2.046168327331543} 02/26/2022 13:33:34 - INFO - codeparrot_training - Step 46309: {'lr': 7.25942587964662e-06, 'samples': 23710720, 'steps': 46309, 'loss/train': 1.9194923639297485} 02/26/2022 13:33:38 - INFO - codeparrot_training - Step 46310: {'lr': 7.255511965916006e-06, 'samples': 23711232, 'steps': 46310, 'loss/train': 1.292546272277832} 02/26/2022 13:33:43 - INFO - codeparrot_training - Step 46311: {'lr': 7.25159909202569e-06, 'samples': 23711744, 'steps': 46311, 'loss/train': 1.5153529644012451} 02/26/2022 13:33:47 - INFO - codeparrot_training - Step 46312: {'lr': 7.247687257992463e-06, 'samples': 23712256, 'steps': 46312, 'loss/train': 1.8743751049041748} 02/26/2022 13:33:52 - INFO - codeparrot_training - Step 46313: {'lr': 7.243776463833035e-06, 'samples': 23712768, 'steps': 46313, 'loss/train': 2.3574328422546387} 02/26/2022 13:33:56 - INFO - codeparrot_training - Step 46314: {'lr': 7.239866709564252e-06, 'samples': 23713280, 'steps': 46314, 'loss/train': 2.1100828647613525} 02/26/2022 13:34:01 - INFO - codeparrot_training - Step 46315: {'lr': 7.235957995202741e-06, 'samples': 23713792, 'steps': 46315, 'loss/train': 2.3595588207244873} 02/26/2022 13:34:05 - INFO - codeparrot_training - Step 46316: {'lr': 7.232050320765321e-06, 'samples': 23714304, 'steps': 46316, 'loss/train': 1.548946738243103} 02/26/2022 13:34:11 - INFO - codeparrot_training - Step 46317: {'lr': 7.2281436862686735e-06, 'samples': 23714816, 'steps': 46317, 'loss/train': 0.9009461998939514} 02/26/2022 13:34:14 - INFO - codeparrot_training - Step 46318: {'lr': 7.224238091729618e-06, 'samples': 23715328, 'steps': 46318, 'loss/train': 2.0814337730407715} 02/26/2022 13:34:21 - INFO - codeparrot_training - Step 46319: {'lr': 7.220333537164808e-06, 'samples': 23715840, 'steps': 46319, 'loss/train': 2.3470852375030518} 02/26/2022 13:34:24 - INFO - codeparrot_training - Step 46320: {'lr': 7.216430022591009e-06, 'samples': 23716352, 'steps': 46320, 'loss/train': 1.624616026878357} 02/26/2022 13:34:30 - INFO - codeparrot_training - Step 46321: {'lr': 7.2125275480249e-06, 'samples': 23716864, 'steps': 46321, 'loss/train': 1.901342749595642} 02/26/2022 13:34:33 - INFO - codeparrot_training - Step 46322: {'lr': 7.208626113483274e-06, 'samples': 23717376, 'steps': 46322, 'loss/train': 0.9478985071182251} 02/26/2022 13:34:39 - INFO - codeparrot_training - Step 46323: {'lr': 7.204725718982785e-06, 'samples': 23717888, 'steps': 46323, 'loss/train': 3.80117130279541} 02/26/2022 13:34:42 - INFO - codeparrot_training - Step 46324: {'lr': 7.200826364540142e-06, 'samples': 23718400, 'steps': 46324, 'loss/train': 1.3132497072219849} 02/26/2022 13:34:48 - INFO - codeparrot_training - Step 46325: {'lr': 7.196928050172052e-06, 'samples': 23718912, 'steps': 46325, 'loss/train': 2.0616514682769775} 02/26/2022 13:34:51 - INFO - codeparrot_training - Step 46326: {'lr': 7.193030775895254e-06, 'samples': 23719424, 'steps': 46326, 'loss/train': 1.5557838678359985} 02/26/2022 13:34:57 - INFO - codeparrot_training - Step 46327: {'lr': 7.1891345417263995e-06, 'samples': 23719936, 'steps': 46327, 'loss/train': 1.473690152168274} 02/26/2022 13:35:00 - INFO - codeparrot_training - Step 46328: {'lr': 7.185239347682199e-06, 'samples': 23720448, 'steps': 46328, 'loss/train': 1.9749585390090942} 02/26/2022 13:35:06 - INFO - codeparrot_training - Step 46329: {'lr': 7.181345193779304e-06, 'samples': 23720960, 'steps': 46329, 'loss/train': 0.7785266041755676} 02/26/2022 13:35:09 - INFO - codeparrot_training - Step 46330: {'lr': 7.177452080034452e-06, 'samples': 23721472, 'steps': 46330, 'loss/train': 0.4698942005634308} 02/26/2022 13:35:15 - INFO - codeparrot_training - Step 46331: {'lr': 7.173560006464269e-06, 'samples': 23721984, 'steps': 46331, 'loss/train': 2.16408109664917} 02/26/2022 13:35:19 - INFO - codeparrot_training - Step 46332: {'lr': 7.169668973085463e-06, 'samples': 23722496, 'steps': 46332, 'loss/train': 2.0029828548431396} 02/26/2022 13:35:24 - INFO - codeparrot_training - Step 46333: {'lr': 7.165778979914689e-06, 'samples': 23723008, 'steps': 46333, 'loss/train': 2.0213077068328857} 02/26/2022 13:35:28 - INFO - codeparrot_training - Step 46334: {'lr': 7.161890026968598e-06, 'samples': 23723520, 'steps': 46334, 'loss/train': 2.0433108806610107} 02/26/2022 13:35:33 - INFO - codeparrot_training - Step 46335: {'lr': 7.158002114263845e-06, 'samples': 23724032, 'steps': 46335, 'loss/train': 1.7285903692245483} 02/26/2022 13:35:37 - INFO - codeparrot_training - Step 46336: {'lr': 7.154115241817166e-06, 'samples': 23724544, 'steps': 46336, 'loss/train': 1.5634167194366455} 02/26/2022 13:35:42 - INFO - codeparrot_training - Step 46337: {'lr': 7.150229409645076e-06, 'samples': 23725056, 'steps': 46337, 'loss/train': 1.351056694984436} 02/26/2022 13:35:46 - INFO - codeparrot_training - Step 46338: {'lr': 7.146344617764311e-06, 'samples': 23725568, 'steps': 46338, 'loss/train': 2.9512109756469727} 02/26/2022 13:35:51 - INFO - codeparrot_training - Step 46339: {'lr': 7.142460866191497e-06, 'samples': 23726080, 'steps': 46339, 'loss/train': 2.253892660140991} 02/26/2022 13:35:55 - INFO - codeparrot_training - Step 46340: {'lr': 7.138578154943287e-06, 'samples': 23726592, 'steps': 46340, 'loss/train': 2.0414609909057617} 02/26/2022 13:36:00 - INFO - codeparrot_training - Step 46341: {'lr': 7.134696484036251e-06, 'samples': 23727104, 'steps': 46341, 'loss/train': 0.4035230875015259} 02/26/2022 13:36:04 - INFO - codeparrot_training - Step 46342: {'lr': 7.130815853487071e-06, 'samples': 23727616, 'steps': 46342, 'loss/train': 0.42983752489089966} 02/26/2022 13:36:10 - INFO - codeparrot_training - Step 46343: {'lr': 7.1269362633123725e-06, 'samples': 23728128, 'steps': 46343, 'loss/train': 2.3718860149383545} 02/26/2022 13:36:13 - INFO - codeparrot_training - Step 46344: {'lr': 7.1230577135287525e-06, 'samples': 23728640, 'steps': 46344, 'loss/train': 0.7256945967674255} 02/26/2022 13:36:20 - INFO - codeparrot_training - Step 46345: {'lr': 7.119180204152781e-06, 'samples': 23729152, 'steps': 46345, 'loss/train': 1.1530717611312866} 02/26/2022 13:36:23 - INFO - codeparrot_training - Step 46346: {'lr': 7.115303735201168e-06, 'samples': 23729664, 'steps': 46346, 'loss/train': 2.5768351554870605} 02/26/2022 13:36:29 - INFO - codeparrot_training - Step 46347: {'lr': 7.111428306690482e-06, 'samples': 23730176, 'steps': 46347, 'loss/train': 1.0652800798416138} 02/26/2022 13:36:33 - INFO - codeparrot_training - Step 46348: {'lr': 7.107553918637266e-06, 'samples': 23730688, 'steps': 46348, 'loss/train': 1.7955538034439087} 02/26/2022 13:36:38 - INFO - codeparrot_training - Step 46349: {'lr': 7.103680571058202e-06, 'samples': 23731200, 'steps': 46349, 'loss/train': 1.219046950340271} 02/26/2022 13:36:41 - INFO - codeparrot_training - Step 46350: {'lr': 7.099808263969832e-06, 'samples': 23731712, 'steps': 46350, 'loss/train': 2.186893939971924} 02/26/2022 13:36:47 - INFO - codeparrot_training - Step 46351: {'lr': 7.095936997388752e-06, 'samples': 23732224, 'steps': 46351, 'loss/train': 1.5777232646942139} 02/26/2022 13:36:51 - INFO - codeparrot_training - Step 46352: {'lr': 7.092066771331507e-06, 'samples': 23732736, 'steps': 46352, 'loss/train': 0.8324767351150513} 02/26/2022 13:36:56 - INFO - codeparrot_training - Step 46353: {'lr': 7.0881975858147764e-06, 'samples': 23733248, 'steps': 46353, 'loss/train': 0.571177065372467} 02/26/2022 13:37:00 - INFO - codeparrot_training - Step 46354: {'lr': 7.084329440855048e-06, 'samples': 23733760, 'steps': 46354, 'loss/train': 2.2714080810546875} 02/26/2022 13:37:06 - INFO - codeparrot_training - Step 46355: {'lr': 7.080462336468918e-06, 'samples': 23734272, 'steps': 46355, 'loss/train': 1.245651364326477} 02/26/2022 13:37:10 - INFO - codeparrot_training - Step 46356: {'lr': 7.076596272672931e-06, 'samples': 23734784, 'steps': 46356, 'loss/train': 1.6411590576171875} 02/26/2022 13:37:15 - INFO - codeparrot_training - Step 46357: {'lr': 7.072731249483738e-06, 'samples': 23735296, 'steps': 46357, 'loss/train': 1.6794277429580688} 02/26/2022 13:37:19 - INFO - codeparrot_training - Step 46358: {'lr': 7.068867266917745e-06, 'samples': 23735808, 'steps': 46358, 'loss/train': 2.4114811420440674} 02/26/2022 13:37:24 - INFO - codeparrot_training - Step 46359: {'lr': 7.065004324991631e-06, 'samples': 23736320, 'steps': 46359, 'loss/train': 0.2611892819404602} 02/26/2022 13:37:28 - INFO - codeparrot_training - Step 46360: {'lr': 7.061142423721856e-06, 'samples': 23736832, 'steps': 46360, 'loss/train': 1.831305980682373} 02/26/2022 13:37:33 - INFO - codeparrot_training - Step 46361: {'lr': 7.057281563125045e-06, 'samples': 23737344, 'steps': 46361, 'loss/train': 2.3734524250030518} 02/26/2022 13:37:37 - INFO - codeparrot_training - Step 46362: {'lr': 7.053421743217714e-06, 'samples': 23737856, 'steps': 46362, 'loss/train': 2.57153582572937} 02/26/2022 13:37:42 - INFO - codeparrot_training - Step 46363: {'lr': 7.049562964016349e-06, 'samples': 23738368, 'steps': 46363, 'loss/train': 1.5510889291763306} 02/26/2022 13:37:46 - INFO - codeparrot_training - Step 46364: {'lr': 7.045705225537491e-06, 'samples': 23738880, 'steps': 46364, 'loss/train': 0.845072865486145} 02/26/2022 13:37:52 - INFO - codeparrot_training - Step 46365: {'lr': 7.0418485277977115e-06, 'samples': 23739392, 'steps': 46365, 'loss/train': 1.5776699781417847} 02/26/2022 13:37:56 - INFO - codeparrot_training - Step 46366: {'lr': 7.037992870813525e-06, 'samples': 23739904, 'steps': 46366, 'loss/train': 1.4175328016281128} 02/26/2022 13:38:01 - INFO - codeparrot_training - Step 46367: {'lr': 7.034138254601391e-06, 'samples': 23740416, 'steps': 46367, 'loss/train': 1.720585823059082} 02/26/2022 13:38:05 - INFO - codeparrot_training - Step 46368: {'lr': 7.030284679177851e-06, 'samples': 23740928, 'steps': 46368, 'loss/train': 2.760316848754883} 02/26/2022 13:38:10 - INFO - codeparrot_training - Step 46369: {'lr': 7.0264321445594745e-06, 'samples': 23741440, 'steps': 46369, 'loss/train': 1.2613515853881836} 02/26/2022 13:38:14 - INFO - codeparrot_training - Step 46370: {'lr': 7.022580650762667e-06, 'samples': 23741952, 'steps': 46370, 'loss/train': 1.168245553970337} 02/26/2022 13:38:19 - INFO - codeparrot_training - Step 46371: {'lr': 7.0187301978039965e-06, 'samples': 23742464, 'steps': 46371, 'loss/train': 2.0983407497406006} 02/26/2022 13:38:23 - INFO - codeparrot_training - Step 46372: {'lr': 7.014880785699895e-06, 'samples': 23742976, 'steps': 46372, 'loss/train': 1.5561175346374512} 02/26/2022 13:38:28 - INFO - codeparrot_training - Step 46373: {'lr': 7.011032414466906e-06, 'samples': 23743488, 'steps': 46373, 'loss/train': 2.3228728771209717} 02/26/2022 13:38:32 - INFO - codeparrot_training - Step 46374: {'lr': 7.007185084121487e-06, 'samples': 23744000, 'steps': 46374, 'loss/train': 1.7513600587844849} 02/26/2022 13:38:37 - INFO - codeparrot_training - Step 46375: {'lr': 7.003338794680153e-06, 'samples': 23744512, 'steps': 46375, 'loss/train': 2.057473659515381} 02/26/2022 13:38:41 - INFO - codeparrot_training - Step 46376: {'lr': 6.999493546159336e-06, 'samples': 23745024, 'steps': 46376, 'loss/train': 0.9103937745094299} 02/26/2022 13:38:47 - INFO - codeparrot_training - Step 46377: {'lr': 6.995649338575521e-06, 'samples': 23745536, 'steps': 46377, 'loss/train': 1.9823440313339233} 02/26/2022 13:38:51 - INFO - codeparrot_training - Step 46378: {'lr': 6.9918061719451975e-06, 'samples': 23746048, 'steps': 46378, 'loss/train': 1.6942082643508911} 02/26/2022 13:38:56 - INFO - codeparrot_training - Step 46379: {'lr': 6.9879640462848225e-06, 'samples': 23746560, 'steps': 46379, 'loss/train': 1.7283331155776978} 02/26/2022 13:38:59 - INFO - codeparrot_training - Step 46380: {'lr': 6.984122961610828e-06, 'samples': 23747072, 'steps': 46380, 'loss/train': 2.2948594093322754} 02/26/2022 13:39:05 - INFO - codeparrot_training - Step 46381: {'lr': 6.9802829179396734e-06, 'samples': 23747584, 'steps': 46381, 'loss/train': 1.6163698434829712} 02/26/2022 13:39:09 - INFO - codeparrot_training - Step 46382: {'lr': 6.976443915287817e-06, 'samples': 23748096, 'steps': 46382, 'loss/train': 3.2901971340179443} 02/26/2022 13:39:14 - INFO - codeparrot_training - Step 46383: {'lr': 6.972605953671745e-06, 'samples': 23748608, 'steps': 46383, 'loss/train': 2.113461971282959} 02/26/2022 13:39:20 - INFO - codeparrot_training - Step 46384: {'lr': 6.968769033107836e-06, 'samples': 23749120, 'steps': 46384, 'loss/train': 2.248030424118042} 02/26/2022 13:39:23 - INFO - codeparrot_training - Step 46385: {'lr': 6.964933153612546e-06, 'samples': 23749632, 'steps': 46385, 'loss/train': 1.8545007705688477} 02/26/2022 13:39:29 - INFO - codeparrot_training - Step 46386: {'lr': 6.961098315202308e-06, 'samples': 23750144, 'steps': 46386, 'loss/train': 1.8316305875778198} 02/26/2022 13:39:33 - INFO - codeparrot_training - Step 46387: {'lr': 6.957264517893552e-06, 'samples': 23750656, 'steps': 46387, 'loss/train': 1.458087682723999} 02/26/2022 13:39:38 - INFO - codeparrot_training - Step 46388: {'lr': 6.953431761702711e-06, 'samples': 23751168, 'steps': 46388, 'loss/train': 1.8730438947677612} 02/26/2022 13:39:42 - INFO - codeparrot_training - Step 46389: {'lr': 6.949600046646187e-06, 'samples': 23751680, 'steps': 46389, 'loss/train': 1.6114581823349} 02/26/2022 13:39:47 - INFO - codeparrot_training - Step 46390: {'lr': 6.945769372740413e-06, 'samples': 23752192, 'steps': 46390, 'loss/train': 1.8649907112121582} 02/26/2022 13:39:51 - INFO - codeparrot_training - Step 46391: {'lr': 6.941939740001735e-06, 'samples': 23752704, 'steps': 46391, 'loss/train': 1.341841220855713} 02/26/2022 13:39:56 - INFO - codeparrot_training - Step 46392: {'lr': 6.938111148446668e-06, 'samples': 23753216, 'steps': 46392, 'loss/train': 2.4057626724243164} 02/26/2022 13:40:00 - INFO - codeparrot_training - Step 46393: {'lr': 6.9342835980915345e-06, 'samples': 23753728, 'steps': 46393, 'loss/train': 1.709973692893982} 02/26/2022 13:40:05 - INFO - codeparrot_training - Step 46394: {'lr': 6.930457088952735e-06, 'samples': 23754240, 'steps': 46394, 'loss/train': 1.5739259719848633} 02/26/2022 13:40:09 - INFO - codeparrot_training - Step 46395: {'lr': 6.926631621046647e-06, 'samples': 23754752, 'steps': 46395, 'loss/train': 1.7106016874313354} 02/26/2022 13:40:14 - INFO - codeparrot_training - Step 46396: {'lr': 6.92280719438973e-06, 'samples': 23755264, 'steps': 46396, 'loss/train': 2.3995444774627686} 02/26/2022 13:40:18 - INFO - codeparrot_training - Step 46397: {'lr': 6.918983808998331e-06, 'samples': 23755776, 'steps': 46397, 'loss/train': 0.45177268981933594} 02/26/2022 13:40:23 - INFO - codeparrot_training - Step 46398: {'lr': 6.915161464888797e-06, 'samples': 23756288, 'steps': 46398, 'loss/train': 1.8366248607635498} 02/26/2022 13:40:27 - INFO - codeparrot_training - Step 46399: {'lr': 6.911340162077506e-06, 'samples': 23756800, 'steps': 46399, 'loss/train': 1.68247652053833} 02/26/2022 13:40:32 - INFO - codeparrot_training - Step 46400: {'lr': 6.907519900580861e-06, 'samples': 23757312, 'steps': 46400, 'loss/train': 0.4413408637046814} 02/26/2022 13:40:36 - INFO - codeparrot_training - Step 46401: {'lr': 6.903700680415209e-06, 'samples': 23757824, 'steps': 46401, 'loss/train': 1.572052001953125} 02/26/2022 13:40:42 - INFO - codeparrot_training - Step 46402: {'lr': 6.899882501596927e-06, 'samples': 23758336, 'steps': 46402, 'loss/train': 1.7348308563232422} 02/26/2022 13:40:45 - INFO - codeparrot_training - Step 46403: {'lr': 6.896065364142307e-06, 'samples': 23758848, 'steps': 46403, 'loss/train': 0.9265848994255066} 02/26/2022 13:40:51 - INFO - codeparrot_training - Step 46404: {'lr': 6.892249268067807e-06, 'samples': 23759360, 'steps': 46404, 'loss/train': 1.5321924686431885} 02/26/2022 13:40:54 - INFO - codeparrot_training - Step 46405: {'lr': 6.888434213389694e-06, 'samples': 23759872, 'steps': 46405, 'loss/train': 1.8817692995071411} 02/26/2022 13:41:00 - INFO - codeparrot_training - Step 46406: {'lr': 6.8846202001243145e-06, 'samples': 23760384, 'steps': 46406, 'loss/train': 1.6436309814453125} 02/26/2022 13:41:03 - INFO - codeparrot_training - Step 46407: {'lr': 6.880807228288016e-06, 'samples': 23760896, 'steps': 46407, 'loss/train': 2.03767991065979} 02/26/2022 13:41:09 - INFO - codeparrot_training - Step 46408: {'lr': 6.876995297897176e-06, 'samples': 23761408, 'steps': 46408, 'loss/train': 1.8102270364761353} 02/26/2022 13:41:12 - INFO - codeparrot_training - Step 46409: {'lr': 6.873184408968059e-06, 'samples': 23761920, 'steps': 46409, 'loss/train': 1.194539189338684} 02/26/2022 13:41:18 - INFO - codeparrot_training - Step 46410: {'lr': 6.869374561517066e-06, 'samples': 23762432, 'steps': 46410, 'loss/train': 0.08907842636108398} 02/26/2022 13:41:21 - INFO - codeparrot_training - Step 46411: {'lr': 6.86556575556041e-06, 'samples': 23762944, 'steps': 46411, 'loss/train': 2.4222910404205322} 02/26/2022 13:41:27 - INFO - codeparrot_training - Step 46412: {'lr': 6.86175799111452e-06, 'samples': 23763456, 'steps': 46412, 'loss/train': 2.30281925201416} 02/26/2022 13:41:31 - INFO - codeparrot_training - Step 46413: {'lr': 6.857951268195605e-06, 'samples': 23763968, 'steps': 46413, 'loss/train': 1.2040692567825317} 02/26/2022 13:41:36 - INFO - codeparrot_training - Step 46414: {'lr': 6.854145586820071e-06, 'samples': 23764480, 'steps': 46414, 'loss/train': 2.3244926929473877} 02/26/2022 13:41:40 - INFO - codeparrot_training - Step 46415: {'lr': 6.850340947004124e-06, 'samples': 23764992, 'steps': 46415, 'loss/train': 1.1270174980163574} 02/26/2022 13:41:45 - INFO - codeparrot_training - Step 46416: {'lr': 6.846537348764114e-06, 'samples': 23765504, 'steps': 46416, 'loss/train': 2.4419026374816895} 02/26/2022 13:41:49 - INFO - codeparrot_training - Step 46417: {'lr': 6.842734792116334e-06, 'samples': 23766016, 'steps': 46417, 'loss/train': 1.7396389245986938} 02/26/2022 13:41:54 - INFO - codeparrot_training - Step 46418: {'lr': 6.838933277077103e-06, 'samples': 23766528, 'steps': 46418, 'loss/train': 2.305845022201538} 02/26/2022 13:41:58 - INFO - codeparrot_training - Step 46419: {'lr': 6.83513280366263e-06, 'samples': 23767040, 'steps': 46419, 'loss/train': 1.3567595481872559} 02/26/2022 13:42:03 - INFO - codeparrot_training - Step 46420: {'lr': 6.8313333718892365e-06, 'samples': 23767552, 'steps': 46420, 'loss/train': 2.64897084236145} 02/26/2022 13:42:07 - INFO - codeparrot_training - Step 46421: {'lr': 6.827534981773187e-06, 'samples': 23768064, 'steps': 46421, 'loss/train': 2.4512853622436523} 02/26/2022 13:42:13 - INFO - codeparrot_training - Step 46422: {'lr': 6.823737633330801e-06, 'samples': 23768576, 'steps': 46422, 'loss/train': 0.6194876432418823} 02/26/2022 13:42:16 - INFO - codeparrot_training - Step 46423: {'lr': 6.819941326578289e-06, 'samples': 23769088, 'steps': 46423, 'loss/train': 2.2124860286712646} 02/26/2022 13:42:22 - INFO - codeparrot_training - Step 46424: {'lr': 6.816146061531914e-06, 'samples': 23769600, 'steps': 46424, 'loss/train': 2.097661256790161} 02/26/2022 13:42:25 - INFO - codeparrot_training - Step 46425: {'lr': 6.812351838207942e-06, 'samples': 23770112, 'steps': 46425, 'loss/train': 0.35688576102256775} 02/26/2022 13:42:31 - INFO - codeparrot_training - Step 46426: {'lr': 6.808558656622665e-06, 'samples': 23770624, 'steps': 46426, 'loss/train': 1.4441722631454468} 02/26/2022 13:42:34 - INFO - codeparrot_training - Step 46427: {'lr': 6.804766516792293e-06, 'samples': 23771136, 'steps': 46427, 'loss/train': 1.9445942640304565} 02/26/2022 13:42:40 - INFO - codeparrot_training - Step 46428: {'lr': 6.800975418733063e-06, 'samples': 23771648, 'steps': 46428, 'loss/train': 1.642012596130371} 02/26/2022 13:42:43 - INFO - codeparrot_training - Step 46429: {'lr': 6.797185362461239e-06, 'samples': 23772160, 'steps': 46429, 'loss/train': 0.5557148456573486} 02/26/2022 13:42:49 - INFO - codeparrot_training - Step 46430: {'lr': 6.7933963479930585e-06, 'samples': 23772672, 'steps': 46430, 'loss/train': 0.997287929058075} 02/26/2022 13:42:52 - INFO - codeparrot_training - Step 46431: {'lr': 6.78960837534473e-06, 'samples': 23773184, 'steps': 46431, 'loss/train': 0.7129632234573364} 02/26/2022 13:42:58 - INFO - codeparrot_training - Step 46432: {'lr': 6.785821444532492e-06, 'samples': 23773696, 'steps': 46432, 'loss/train': 1.0941447019577026} 02/26/2022 13:43:02 - INFO - codeparrot_training - Step 46433: {'lr': 6.782035555572552e-06, 'samples': 23774208, 'steps': 46433, 'loss/train': 2.6145706176757812} 02/26/2022 13:43:07 - INFO - codeparrot_training - Step 46434: {'lr': 6.778250708481148e-06, 'samples': 23774720, 'steps': 46434, 'loss/train': 1.8168832063674927} 02/26/2022 13:43:11 - INFO - codeparrot_training - Step 46435: {'lr': 6.7744669032745175e-06, 'samples': 23775232, 'steps': 46435, 'loss/train': 2.020702600479126} 02/26/2022 13:43:16 - INFO - codeparrot_training - Step 46436: {'lr': 6.770684139968814e-06, 'samples': 23775744, 'steps': 46436, 'loss/train': 1.5341795682907104} 02/26/2022 13:43:20 - INFO - codeparrot_training - Step 46437: {'lr': 6.766902418580273e-06, 'samples': 23776256, 'steps': 46437, 'loss/train': 1.170082688331604} 02/26/2022 13:43:25 - INFO - codeparrot_training - Step 46438: {'lr': 6.76312173912505e-06, 'samples': 23776768, 'steps': 46438, 'loss/train': 0.9335953593254089} 02/26/2022 13:43:29 - INFO - codeparrot_training - Step 46439: {'lr': 6.759342101619409e-06, 'samples': 23777280, 'steps': 46439, 'loss/train': 2.61860990524292} 02/26/2022 13:43:34 - INFO - codeparrot_training - Step 46440: {'lr': 6.755563506079504e-06, 'samples': 23777792, 'steps': 46440, 'loss/train': 1.5336909294128418} 02/26/2022 13:43:38 - INFO - codeparrot_training - Step 46441: {'lr': 6.751785952521544e-06, 'samples': 23778304, 'steps': 46441, 'loss/train': 2.984496831893921} 02/26/2022 13:43:43 - INFO - codeparrot_training - Step 46442: {'lr': 6.748009440961655e-06, 'samples': 23778816, 'steps': 46442, 'loss/train': 1.4492080211639404} 02/26/2022 13:43:47 - INFO - codeparrot_training - Step 46443: {'lr': 6.744233971416075e-06, 'samples': 23779328, 'steps': 46443, 'loss/train': 1.7095447778701782} 02/26/2022 13:43:52 - INFO - codeparrot_training - Step 46444: {'lr': 6.740459543900956e-06, 'samples': 23779840, 'steps': 46444, 'loss/train': 1.5660254955291748} 02/26/2022 13:43:56 - INFO - codeparrot_training - Step 46445: {'lr': 6.736686158432481e-06, 'samples': 23780352, 'steps': 46445, 'loss/train': 1.6093417406082153} 02/26/2022 13:44:02 - INFO - codeparrot_training - Step 46446: {'lr': 6.732913815026748e-06, 'samples': 23780864, 'steps': 46446, 'loss/train': 1.5200092792510986} 02/26/2022 13:44:05 - INFO - codeparrot_training - Step 46447: {'lr': 6.72914251370002e-06, 'samples': 23781376, 'steps': 46447, 'loss/train': 3.6204516887664795} 02/26/2022 13:44:11 - INFO - codeparrot_training - Step 46448: {'lr': 6.725372254468343e-06, 'samples': 23781888, 'steps': 46448, 'loss/train': 2.036477565765381} 02/26/2022 13:44:17 - INFO - codeparrot_training - Step 46449: {'lr': 6.721603037348006e-06, 'samples': 23782400, 'steps': 46449, 'loss/train': 1.6362109184265137} 02/26/2022 13:44:20 - INFO - codeparrot_training - Step 46450: {'lr': 6.717834862354999e-06, 'samples': 23782912, 'steps': 46450, 'loss/train': 1.953580617904663} 02/26/2022 13:44:26 - INFO - codeparrot_training - Step 46451: {'lr': 6.7140677295055575e-06, 'samples': 23783424, 'steps': 46451, 'loss/train': 1.2113935947418213} 02/26/2022 13:44:29 - INFO - codeparrot_training - Step 46452: {'lr': 6.71030163881578e-06, 'samples': 23783936, 'steps': 46452, 'loss/train': 1.748292326927185} 02/26/2022 13:44:35 - INFO - codeparrot_training - Step 46453: {'lr': 6.706536590301876e-06, 'samples': 23784448, 'steps': 46453, 'loss/train': 1.4966521263122559} 02/26/2022 13:44:38 - INFO - codeparrot_training - Step 46454: {'lr': 6.70277258397986e-06, 'samples': 23784960, 'steps': 46454, 'loss/train': 1.5227965116500854} 02/26/2022 13:44:44 - INFO - codeparrot_training - Step 46455: {'lr': 6.699009619865914e-06, 'samples': 23785472, 'steps': 46455, 'loss/train': 1.4812519550323486} 02/26/2022 13:44:47 - INFO - codeparrot_training - Step 46456: {'lr': 6.695247697976164e-06, 'samples': 23785984, 'steps': 46456, 'loss/train': 1.727266788482666} 02/26/2022 13:44:51 - INFO - codeparrot_training - Step 46457: {'lr': 6.691486818326709e-06, 'samples': 23786496, 'steps': 46457, 'loss/train': 1.5667589902877808} 02/26/2022 13:44:57 - INFO - codeparrot_training - Step 46458: {'lr': 6.687726980933673e-06, 'samples': 23787008, 'steps': 46458, 'loss/train': 1.1811867952346802} 02/26/2022 13:45:00 - INFO - codeparrot_training - Step 46459: {'lr': 6.683968185813183e-06, 'samples': 23787520, 'steps': 46459, 'loss/train': 1.806470513343811} 02/26/2022 13:45:06 - INFO - codeparrot_training - Step 46460: {'lr': 6.680210432981254e-06, 'samples': 23788032, 'steps': 46460, 'loss/train': 2.3082308769226074} 02/26/2022 13:45:09 - INFO - codeparrot_training - Step 46461: {'lr': 6.676453722454068e-06, 'samples': 23788544, 'steps': 46461, 'loss/train': 2.1304683685302734} 02/26/2022 13:45:15 - INFO - codeparrot_training - Step 46462: {'lr': 6.672698054247695e-06, 'samples': 23789056, 'steps': 46462, 'loss/train': 2.252291440963745} 02/26/2022 13:45:20 - INFO - codeparrot_training - Step 46463: {'lr': 6.668943428378233e-06, 'samples': 23789568, 'steps': 46463, 'loss/train': 0.9569962024688721} 02/26/2022 13:45:24 - INFO - codeparrot_training - Step 46464: {'lr': 6.665189844861724e-06, 'samples': 23790080, 'steps': 46464, 'loss/train': 1.9051259756088257} 02/26/2022 13:45:29 - INFO - codeparrot_training - Step 46465: {'lr': 6.661437303714269e-06, 'samples': 23790592, 'steps': 46465, 'loss/train': 4.860052108764648} 02/26/2022 13:45:33 - INFO - codeparrot_training - Step 46466: {'lr': 6.657685804951991e-06, 'samples': 23791104, 'steps': 46466, 'loss/train': 1.9033058881759644} 02/26/2022 13:45:39 - INFO - codeparrot_training - Step 46467: {'lr': 6.653935348590878e-06, 'samples': 23791616, 'steps': 46467, 'loss/train': 2.4031310081481934} 02/26/2022 13:45:43 - INFO - codeparrot_training - Step 46468: {'lr': 6.650185934647029e-06, 'samples': 23792128, 'steps': 46468, 'loss/train': 1.2548633813858032} 02/26/2022 13:45:46 - INFO - codeparrot_training - Step 46469: {'lr': 6.646437563136543e-06, 'samples': 23792640, 'steps': 46469, 'loss/train': 1.8490608930587769} 02/26/2022 13:45:52 - INFO - codeparrot_training - Step 46470: {'lr': 6.642690234075432e-06, 'samples': 23793152, 'steps': 46470, 'loss/train': 2.378690004348755} 02/26/2022 13:45:55 - INFO - codeparrot_training - Step 46471: {'lr': 6.63894394747977e-06, 'samples': 23793664, 'steps': 46471, 'loss/train': 0.8607810735702515} 02/26/2022 13:46:01 - INFO - codeparrot_training - Step 46472: {'lr': 6.635198703365569e-06, 'samples': 23794176, 'steps': 46472, 'loss/train': 2.0531651973724365} 02/26/2022 13:46:04 - INFO - codeparrot_training - Step 46473: {'lr': 6.631454501748902e-06, 'samples': 23794688, 'steps': 46473, 'loss/train': 2.18971848487854} 02/26/2022 13:46:10 - INFO - codeparrot_training - Step 46474: {'lr': 6.627711342645837e-06, 'samples': 23795200, 'steps': 46474, 'loss/train': 1.747352123260498} 02/26/2022 13:46:13 - INFO - codeparrot_training - Step 46475: {'lr': 6.6239692260723624e-06, 'samples': 23795712, 'steps': 46475, 'loss/train': 1.4273356199264526} 02/26/2022 13:46:19 - INFO - codeparrot_training - Step 46476: {'lr': 6.620228152044494e-06, 'samples': 23796224, 'steps': 46476, 'loss/train': 1.5860884189605713} 02/26/2022 13:46:22 - INFO - codeparrot_training - Step 46477: {'lr': 6.616488120578329e-06, 'samples': 23796736, 'steps': 46477, 'loss/train': 1.7078667879104614} 02/26/2022 13:46:28 - INFO - codeparrot_training - Step 46478: {'lr': 6.612749131689827e-06, 'samples': 23797248, 'steps': 46478, 'loss/train': 2.1136834621429443} 02/26/2022 13:46:31 - INFO - codeparrot_training - Step 46479: {'lr': 6.609011185395031e-06, 'samples': 23797760, 'steps': 46479, 'loss/train': 3.049741506576538} 02/26/2022 13:46:38 - INFO - codeparrot_training - Step 46480: {'lr': 6.605274281709927e-06, 'samples': 23798272, 'steps': 46480, 'loss/train': 1.4769575595855713} 02/26/2022 13:46:41 - INFO - codeparrot_training - Step 46481: {'lr': 6.6015384206505605e-06, 'samples': 23798784, 'steps': 46481, 'loss/train': 1.2307194471359253} 02/26/2022 13:46:47 - INFO - codeparrot_training - Step 46482: {'lr': 6.597803602232916e-06, 'samples': 23799296, 'steps': 46482, 'loss/train': 1.357944130897522} 02/26/2022 13:46:50 - INFO - codeparrot_training - Step 46483: {'lr': 6.594069826472981e-06, 'samples': 23799808, 'steps': 46483, 'loss/train': 1.6806366443634033} 02/26/2022 13:46:56 - INFO - codeparrot_training - Step 46484: {'lr': 6.590337093386772e-06, 'samples': 23800320, 'steps': 46484, 'loss/train': 1.6275931596755981} 02/26/2022 13:46:59 - INFO - codeparrot_training - Step 46485: {'lr': 6.586605402990275e-06, 'samples': 23800832, 'steps': 46485, 'loss/train': 1.729817509651184} 02/26/2022 13:47:05 - INFO - codeparrot_training - Step 46486: {'lr': 6.582874755299451e-06, 'samples': 23801344, 'steps': 46486, 'loss/train': 0.7806394100189209} 02/26/2022 13:47:08 - INFO - codeparrot_training - Step 46487: {'lr': 6.579145150330284e-06, 'samples': 23801856, 'steps': 46487, 'loss/train': 0.62368243932724} 02/26/2022 13:47:14 - INFO - codeparrot_training - Step 46488: {'lr': 6.57541658809882e-06, 'samples': 23802368, 'steps': 46488, 'loss/train': 1.6624560356140137} 02/26/2022 13:47:17 - INFO - codeparrot_training - Step 46489: {'lr': 6.571689068620906e-06, 'samples': 23802880, 'steps': 46489, 'loss/train': 2.3215036392211914} 02/26/2022 13:47:23 - INFO - codeparrot_training - Step 46490: {'lr': 6.567962591912613e-06, 'samples': 23803392, 'steps': 46490, 'loss/train': 2.1143174171447754} 02/26/2022 13:47:26 - INFO - codeparrot_training - Step 46491: {'lr': 6.564237157989872e-06, 'samples': 23803904, 'steps': 46491, 'loss/train': 2.7534046173095703} 02/26/2022 13:47:32 - INFO - codeparrot_training - Step 46492: {'lr': 6.5605127668686425e-06, 'samples': 23804416, 'steps': 46492, 'loss/train': 2.5225470066070557} 02/26/2022 13:47:35 - INFO - codeparrot_training - Step 46493: {'lr': 6.556789418564857e-06, 'samples': 23804928, 'steps': 46493, 'loss/train': 1.5890918970108032} 02/26/2022 13:47:41 - INFO - codeparrot_training - Step 46494: {'lr': 6.553067113094502e-06, 'samples': 23805440, 'steps': 46494, 'loss/train': 2.932908535003662} 02/26/2022 13:47:45 - INFO - codeparrot_training - Step 46495: {'lr': 6.5493458504734815e-06, 'samples': 23805952, 'steps': 46495, 'loss/train': 1.5942814350128174} 02/26/2022 13:47:51 - INFO - codeparrot_training - Step 46496: {'lr': 6.545625630717783e-06, 'samples': 23806464, 'steps': 46496, 'loss/train': 1.4911079406738281} 02/26/2022 13:47:54 - INFO - codeparrot_training - Step 46497: {'lr': 6.541906453843311e-06, 'samples': 23806976, 'steps': 46497, 'loss/train': 0.730410099029541} 02/26/2022 13:47:59 - INFO - codeparrot_training - Step 46498: {'lr': 6.5381883198660234e-06, 'samples': 23807488, 'steps': 46498, 'loss/train': 1.453140139579773} 02/26/2022 13:48:03 - INFO - codeparrot_training - Step 46499: {'lr': 6.534471228801769e-06, 'samples': 23808000, 'steps': 46499, 'loss/train': 1.7993172407150269} 02/26/2022 13:48:09 - INFO - codeparrot_training - Step 46500: {'lr': 6.530755180666592e-06, 'samples': 23808512, 'steps': 46500, 'loss/train': 0.24077600240707397} 02/26/2022 13:48:12 - INFO - codeparrot_training - Step 46501: {'lr': 6.52704017547634e-06, 'samples': 23809024, 'steps': 46501, 'loss/train': 1.7888671159744263} 02/26/2022 13:48:18 - INFO - codeparrot_training - Step 46502: {'lr': 6.523326213246916e-06, 'samples': 23809536, 'steps': 46502, 'loss/train': 0.8614785075187683} 02/26/2022 13:48:21 - INFO - codeparrot_training - Step 46503: {'lr': 6.519613293994253e-06, 'samples': 23810048, 'steps': 46503, 'loss/train': 0.784758985042572} 02/26/2022 13:48:27 - INFO - codeparrot_training - Step 46504: {'lr': 6.515901417734255e-06, 'samples': 23810560, 'steps': 46504, 'loss/train': 1.494071125984192} 02/26/2022 13:48:31 - INFO - codeparrot_training - Step 46505: {'lr': 6.512190584482825e-06, 'samples': 23811072, 'steps': 46505, 'loss/train': 1.955809473991394} 02/26/2022 13:48:36 - INFO - codeparrot_training - Step 46506: {'lr': 6.508480794255895e-06, 'samples': 23811584, 'steps': 46506, 'loss/train': 1.6849819421768188} 02/26/2022 13:48:40 - INFO - codeparrot_training - Step 46507: {'lr': 6.5047720470692575e-06, 'samples': 23812096, 'steps': 46507, 'loss/train': 1.6729947328567505} 02/26/2022 13:48:45 - INFO - codeparrot_training - Step 46508: {'lr': 6.5010643429388724e-06, 'samples': 23812608, 'steps': 46508, 'loss/train': 1.6640431880950928} 02/26/2022 13:48:49 - INFO - codeparrot_training - Step 46509: {'lr': 6.497357681880589e-06, 'samples': 23813120, 'steps': 46509, 'loss/train': 1.933899998664856} 02/26/2022 13:48:55 - INFO - codeparrot_training - Step 46510: {'lr': 6.493652063910366e-06, 'samples': 23813632, 'steps': 46510, 'loss/train': 2.2764484882354736} 02/26/2022 13:48:58 - INFO - codeparrot_training - Step 46511: {'lr': 6.48994748904394e-06, 'samples': 23814144, 'steps': 46511, 'loss/train': 1.3437808752059937} 02/26/2022 13:49:04 - INFO - codeparrot_training - Step 46512: {'lr': 6.4862439572973e-06, 'samples': 23814656, 'steps': 46512, 'loss/train': 2.2047863006591797} 02/26/2022 13:49:07 - INFO - codeparrot_training - Step 46513: {'lr': 6.482541468686265e-06, 'samples': 23815168, 'steps': 46513, 'loss/train': 0.4005439281463623} 02/26/2022 13:49:13 - INFO - codeparrot_training - Step 46514: {'lr': 6.478840023226712e-06, 'samples': 23815680, 'steps': 46514, 'loss/train': 1.5261503458023071} 02/26/2022 13:49:16 - INFO - codeparrot_training - Step 46515: {'lr': 6.475139620934433e-06, 'samples': 23816192, 'steps': 46515, 'loss/train': 2.8783137798309326} 02/26/2022 13:49:22 - INFO - codeparrot_training - Step 46516: {'lr': 6.471440261825362e-06, 'samples': 23816704, 'steps': 46516, 'loss/train': 1.5007215738296509} 02/26/2022 13:49:26 - INFO - codeparrot_training - Step 46517: {'lr': 6.467741945915289e-06, 'samples': 23817216, 'steps': 46517, 'loss/train': 1.7192862033843994} 02/26/2022 13:49:31 - INFO - codeparrot_training - Step 46518: {'lr': 6.464044673220121e-06, 'samples': 23817728, 'steps': 46518, 'loss/train': 1.9812179803848267} 02/26/2022 13:49:35 - INFO - codeparrot_training - Step 46519: {'lr': 6.460348443755621e-06, 'samples': 23818240, 'steps': 46519, 'loss/train': 1.9676893949508667} 02/26/2022 13:49:40 - INFO - codeparrot_training - Step 46520: {'lr': 6.456653257537665e-06, 'samples': 23818752, 'steps': 46520, 'loss/train': 1.057741641998291} 02/26/2022 13:49:44 - INFO - codeparrot_training - Step 46521: {'lr': 6.452959114582102e-06, 'samples': 23819264, 'steps': 46521, 'loss/train': 1.8243581056594849} 02/26/2022 13:49:49 - INFO - codeparrot_training - Step 46522: {'lr': 6.44926601490467e-06, 'samples': 23819776, 'steps': 46522, 'loss/train': 2.279841184616089} 02/26/2022 13:49:53 - INFO - codeparrot_training - Step 46523: {'lr': 6.445573958521273e-06, 'samples': 23820288, 'steps': 46523, 'loss/train': 1.6697983741760254} 02/26/2022 13:49:58 - INFO - codeparrot_training - Step 46524: {'lr': 6.441882945447702e-06, 'samples': 23820800, 'steps': 46524, 'loss/train': 2.1427297592163086} 02/26/2022 13:50:02 - INFO - codeparrot_training - Step 46525: {'lr': 6.438192975699781e-06, 'samples': 23821312, 'steps': 46525, 'loss/train': 1.2807435989379883} 02/26/2022 13:50:07 - INFO - codeparrot_training - Step 46526: {'lr': 6.434504049293271e-06, 'samples': 23821824, 'steps': 46526, 'loss/train': 1.1216572523117065} 02/26/2022 13:50:11 - INFO - codeparrot_training - Step 46527: {'lr': 6.430816166244024e-06, 'samples': 23822336, 'steps': 46527, 'loss/train': 1.4811056852340698} 02/26/2022 13:50:18 - INFO - codeparrot_training - Step 46528: {'lr': 6.427129326567804e-06, 'samples': 23822848, 'steps': 46528, 'loss/train': 1.458317756652832} 02/26/2022 13:50:22 - INFO - codeparrot_training - Step 46529: {'lr': 6.42344353028046e-06, 'samples': 23823360, 'steps': 46529, 'loss/train': 0.5162584185600281} 02/26/2022 13:50:27 - INFO - codeparrot_training - Step 46530: {'lr': 6.419758777397672e-06, 'samples': 23823872, 'steps': 46530, 'loss/train': 1.766250729560852} 02/26/2022 13:50:31 - INFO - codeparrot_training - Step 46531: {'lr': 6.416075067935345e-06, 'samples': 23824384, 'steps': 46531, 'loss/train': 1.853021264076233} 02/26/2022 13:50:36 - INFO - codeparrot_training - Step 46532: {'lr': 6.4123924019091896e-06, 'samples': 23824896, 'steps': 46532, 'loss/train': 2.168959856033325} 02/26/2022 13:50:40 - INFO - codeparrot_training - Step 46533: {'lr': 6.408710779334998e-06, 'samples': 23825408, 'steps': 46533, 'loss/train': 1.7554508447647095} 02/26/2022 13:50:45 - INFO - codeparrot_training - Step 46534: {'lr': 6.4050302002285345e-06, 'samples': 23825920, 'steps': 46534, 'loss/train': 1.4360580444335938} 02/26/2022 13:50:49 - INFO - codeparrot_training - Step 46535: {'lr': 6.401350664605565e-06, 'samples': 23826432, 'steps': 46535, 'loss/train': 1.4463597536087036} 02/26/2022 13:50:54 - INFO - codeparrot_training - Step 46536: {'lr': 6.397672172481883e-06, 'samples': 23826944, 'steps': 46536, 'loss/train': 3.1432900428771973} 02/26/2022 13:50:58 - INFO - codeparrot_training - Step 46537: {'lr': 6.3939947238732255e-06, 'samples': 23827456, 'steps': 46537, 'loss/train': 2.0650248527526855} 02/26/2022 13:51:05 - INFO - codeparrot_training - Step 46538: {'lr': 6.3903183187953015e-06, 'samples': 23827968, 'steps': 46538, 'loss/train': 1.4216917753219604} 02/26/2022 13:51:08 - INFO - codeparrot_training - Step 46539: {'lr': 6.386642957263905e-06, 'samples': 23828480, 'steps': 46539, 'loss/train': 1.4810951948165894} 02/26/2022 13:51:14 - INFO - codeparrot_training - Step 46540: {'lr': 6.3829686392948005e-06, 'samples': 23828992, 'steps': 46540, 'loss/train': 3.243936777114868} 02/26/2022 13:51:18 - INFO - codeparrot_training - Step 46541: {'lr': 6.3792953649036975e-06, 'samples': 23829504, 'steps': 46541, 'loss/train': 1.8178826570510864} 02/26/2022 13:51:23 - INFO - codeparrot_training - Step 46542: {'lr': 6.375623134106307e-06, 'samples': 23830016, 'steps': 46542, 'loss/train': 2.0278468132019043} 02/26/2022 13:51:27 - INFO - codeparrot_training - Step 46543: {'lr': 6.371951946918419e-06, 'samples': 23830528, 'steps': 46543, 'loss/train': 1.3376773595809937} 02/26/2022 13:51:32 - INFO - codeparrot_training - Step 46544: {'lr': 6.368281803355691e-06, 'samples': 23831040, 'steps': 46544, 'loss/train': 0.41307809948921204} 02/26/2022 13:51:36 - INFO - codeparrot_training - Step 46545: {'lr': 6.364612703433942e-06, 'samples': 23831552, 'steps': 46545, 'loss/train': 1.3543895483016968} 02/26/2022 13:51:41 - INFO - codeparrot_training - Step 46546: {'lr': 6.360944647168798e-06, 'samples': 23832064, 'steps': 46546, 'loss/train': 0.6874322891235352} 02/26/2022 13:51:45 - INFO - codeparrot_training - Step 46547: {'lr': 6.357277634575998e-06, 'samples': 23832576, 'steps': 46547, 'loss/train': 0.3340718150138855} 02/26/2022 13:51:52 - INFO - codeparrot_training - Step 46548: {'lr': 6.35361166567125e-06, 'samples': 23833088, 'steps': 46548, 'loss/train': 1.3821195363998413} 02/26/2022 13:51:56 - INFO - codeparrot_training - Step 46549: {'lr': 6.349946740470292e-06, 'samples': 23833600, 'steps': 46549, 'loss/train': 2.128309965133667} 02/26/2022 13:52:01 - INFO - codeparrot_training - Step 46550: {'lr': 6.34628285898875e-06, 'samples': 23834112, 'steps': 46550, 'loss/train': 1.1764681339263916} 02/26/2022 13:52:05 - INFO - codeparrot_training - Step 46551: {'lr': 6.34262002124239e-06, 'samples': 23834624, 'steps': 46551, 'loss/train': 1.3445438146591187} 02/26/2022 13:52:10 - INFO - codeparrot_training - Step 46552: {'lr': 6.338958227246866e-06, 'samples': 23835136, 'steps': 46552, 'loss/train': 1.153898000717163} 02/26/2022 13:52:14 - INFO - codeparrot_training - Step 46553: {'lr': 6.335297477017887e-06, 'samples': 23835648, 'steps': 46553, 'loss/train': 1.9815213680267334} 02/26/2022 13:52:19 - INFO - codeparrot_training - Step 46554: {'lr': 6.331637770571108e-06, 'samples': 23836160, 'steps': 46554, 'loss/train': 1.7304085493087769} 02/26/2022 13:52:23 - INFO - codeparrot_training - Step 46555: {'lr': 6.327979107922238e-06, 'samples': 23836672, 'steps': 46555, 'loss/train': 2.0602192878723145} 02/26/2022 13:52:28 - INFO - codeparrot_training - Step 46556: {'lr': 6.324321489086904e-06, 'samples': 23837184, 'steps': 46556, 'loss/train': 1.6734329462051392} 02/26/2022 13:52:32 - INFO - codeparrot_training - Step 46557: {'lr': 6.3206649140808145e-06, 'samples': 23837696, 'steps': 46557, 'loss/train': 1.937697410583496} 02/26/2022 13:52:39 - INFO - codeparrot_training - Step 46558: {'lr': 6.317009382919625e-06, 'samples': 23838208, 'steps': 46558, 'loss/train': 1.1814743280410767} 02/26/2022 13:52:42 - INFO - codeparrot_training - Step 46559: {'lr': 6.313354895618989e-06, 'samples': 23838720, 'steps': 46559, 'loss/train': 1.5727505683898926} 02/26/2022 13:52:48 - INFO - codeparrot_training - Step 46560: {'lr': 6.309701452194561e-06, 'samples': 23839232, 'steps': 46560, 'loss/train': 1.6556941270828247} 02/26/2022 13:52:51 - INFO - codeparrot_training - Step 46561: {'lr': 6.306049052661994e-06, 'samples': 23839744, 'steps': 46561, 'loss/train': 1.7878711223602295} 02/26/2022 13:52:57 - INFO - codeparrot_training - Step 46562: {'lr': 6.302397697036916e-06, 'samples': 23840256, 'steps': 46562, 'loss/train': 2.382051944732666} 02/26/2022 13:53:00 - INFO - codeparrot_training - Step 46563: {'lr': 6.298747385335007e-06, 'samples': 23840768, 'steps': 46563, 'loss/train': 1.2224739789962769} 02/26/2022 13:53:06 - INFO - codeparrot_training - Step 46564: {'lr': 6.29509811757184e-06, 'samples': 23841280, 'steps': 46564, 'loss/train': 1.1466196775436401} 02/26/2022 13:53:11 - INFO - codeparrot_training - Step 46565: {'lr': 6.291449893763123e-06, 'samples': 23841792, 'steps': 46565, 'loss/train': 1.770437479019165} 02/26/2022 13:53:15 - INFO - codeparrot_training - Step 46566: {'lr': 6.287802713924456e-06, 'samples': 23842304, 'steps': 46566, 'loss/train': 1.4644834995269775} 02/26/2022 13:53:20 - INFO - codeparrot_training - Step 46567: {'lr': 6.284156578071437e-06, 'samples': 23842816, 'steps': 46567, 'loss/train': 1.7263942956924438} 02/26/2022 13:53:24 - INFO - codeparrot_training - Step 46568: {'lr': 6.28051148621972e-06, 'samples': 23843328, 'steps': 46568, 'loss/train': 1.174325942993164} 02/26/2022 13:53:29 - INFO - codeparrot_training - Step 46569: {'lr': 6.2768674383848755e-06, 'samples': 23843840, 'steps': 46569, 'loss/train': 1.6952998638153076} 02/26/2022 13:53:33 - INFO - codeparrot_training - Step 46570: {'lr': 6.273224434582559e-06, 'samples': 23844352, 'steps': 46570, 'loss/train': 0.7334764003753662} 02/26/2022 13:53:38 - INFO - codeparrot_training - Step 46571: {'lr': 6.26958247482834e-06, 'samples': 23844864, 'steps': 46571, 'loss/train': 1.1163926124572754} 02/26/2022 13:53:42 - INFO - codeparrot_training - Step 46572: {'lr': 6.2659415591378445e-06, 'samples': 23845376, 'steps': 46572, 'loss/train': 2.25423264503479} 02/26/2022 13:53:49 - INFO - codeparrot_training - Step 46573: {'lr': 6.2623016875266445e-06, 'samples': 23845888, 'steps': 46573, 'loss/train': 1.303412914276123} 02/26/2022 13:53:53 - INFO - codeparrot_training - Step 46574: {'lr': 6.258662860010394e-06, 'samples': 23846400, 'steps': 46574, 'loss/train': 2.097568988800049} 02/26/2022 13:53:58 - INFO - codeparrot_training - Step 46575: {'lr': 6.255025076604609e-06, 'samples': 23846912, 'steps': 46575, 'loss/train': 1.623725175857544} 02/26/2022 13:54:02 - INFO - codeparrot_training - Step 46576: {'lr': 6.251388337324887e-06, 'samples': 23847424, 'steps': 46576, 'loss/train': 0.6567850112915039} 02/26/2022 13:54:05 - INFO - codeparrot_training - Step 46577: {'lr': 6.247752642186827e-06, 'samples': 23847936, 'steps': 46577, 'loss/train': 2.17887806892395} 02/26/2022 13:54:11 - INFO - codeparrot_training - Step 46578: {'lr': 6.244117991206e-06, 'samples': 23848448, 'steps': 46578, 'loss/train': 1.637640357017517} 02/26/2022 13:54:14 - INFO - codeparrot_training - Step 46579: {'lr': 6.240484384397949e-06, 'samples': 23848960, 'steps': 46579, 'loss/train': 1.7565633058547974} 02/26/2022 13:54:20 - INFO - codeparrot_training - Step 46580: {'lr': 6.2368518217783e-06, 'samples': 23849472, 'steps': 46580, 'loss/train': 2.4432973861694336} 02/26/2022 13:54:23 - INFO - codeparrot_training - Step 46581: {'lr': 6.2332203033625415e-06, 'samples': 23849984, 'steps': 46581, 'loss/train': 0.6739732027053833} 02/26/2022 13:54:29 - INFO - codeparrot_training - Step 46582: {'lr': 6.229589829166271e-06, 'samples': 23850496, 'steps': 46582, 'loss/train': 2.023789167404175} 02/26/2022 13:54:32 - INFO - codeparrot_training - Step 46583: {'lr': 6.225960399205033e-06, 'samples': 23851008, 'steps': 46583, 'loss/train': 0.9450035691261292} 02/26/2022 13:54:40 - INFO - codeparrot_training - Step 46584: {'lr': 6.2223320134943976e-06, 'samples': 23851520, 'steps': 46584, 'loss/train': 1.9493979215621948} 02/26/2022 13:54:43 - INFO - codeparrot_training - Step 46585: {'lr': 6.2187046720498516e-06, 'samples': 23852032, 'steps': 46585, 'loss/train': 2.1133739948272705} 02/26/2022 13:54:49 - INFO - codeparrot_training - Step 46586: {'lr': 6.215078374886968e-06, 'samples': 23852544, 'steps': 46586, 'loss/train': 2.5516552925109863} 02/26/2022 13:54:52 - INFO - codeparrot_training - Step 46587: {'lr': 6.21145312202126e-06, 'samples': 23853056, 'steps': 46587, 'loss/train': 0.6385740041732788} 02/26/2022 13:54:58 - INFO - codeparrot_training - Step 46588: {'lr': 6.207828913468355e-06, 'samples': 23853568, 'steps': 46588, 'loss/train': 1.7570998668670654} 02/26/2022 13:55:01 - INFO - codeparrot_training - Step 46589: {'lr': 6.204205749243602e-06, 'samples': 23854080, 'steps': 46589, 'loss/train': 0.6544342637062073} 02/26/2022 13:55:07 - INFO - codeparrot_training - Step 46590: {'lr': 6.200583629362683e-06, 'samples': 23854592, 'steps': 46590, 'loss/train': 2.156109571456909} 02/26/2022 13:55:10 - INFO - codeparrot_training - Step 46591: {'lr': 6.196962553841001e-06, 'samples': 23855104, 'steps': 46591, 'loss/train': 0.5243734121322632} 02/26/2022 13:55:16 - INFO - codeparrot_training - Step 46592: {'lr': 6.1933425226941566e-06, 'samples': 23855616, 'steps': 46592, 'loss/train': 1.1431992053985596} 02/26/2022 13:55:20 - INFO - codeparrot_training - Step 46593: {'lr': 6.1897235359376084e-06, 'samples': 23856128, 'steps': 46593, 'loss/train': 0.17018328607082367} 02/26/2022 13:55:27 - INFO - codeparrot_training - Step 46594: {'lr': 6.186105593586844e-06, 'samples': 23856640, 'steps': 46594, 'loss/train': 2.735783338546753} 02/26/2022 13:55:31 - INFO - codeparrot_training - Step 46595: {'lr': 6.182488695657379e-06, 'samples': 23857152, 'steps': 46595, 'loss/train': 2.716249465942383} 02/26/2022 13:55:36 - INFO - codeparrot_training - Step 46596: {'lr': 6.178872842164729e-06, 'samples': 23857664, 'steps': 46596, 'loss/train': 0.9551397562026978} 02/26/2022 13:55:39 - INFO - codeparrot_training - Step 46597: {'lr': 6.175258033124381e-06, 'samples': 23858176, 'steps': 46597, 'loss/train': 1.4555304050445557} 02/26/2022 13:55:45 - INFO - codeparrot_training - Step 46598: {'lr': 6.171644268551796e-06, 'samples': 23858688, 'steps': 46598, 'loss/train': 1.7013649940490723} 02/26/2022 13:55:48 - INFO - codeparrot_training - Step 46599: {'lr': 6.168031548462433e-06, 'samples': 23859200, 'steps': 46599, 'loss/train': 1.5595775842666626} 02/26/2022 13:55:54 - INFO - codeparrot_training - Step 46600: {'lr': 6.164419872871835e-06, 'samples': 23859712, 'steps': 46600, 'loss/train': 1.9842448234558105} 02/26/2022 13:55:57 - INFO - codeparrot_training - Step 46601: {'lr': 6.1608092417954346e-06, 'samples': 23860224, 'steps': 46601, 'loss/train': 1.2747728824615479} 02/26/2022 13:56:03 - INFO - codeparrot_training - Step 46602: {'lr': 6.157199655248691e-06, 'samples': 23860736, 'steps': 46602, 'loss/train': 1.8581600189208984} 02/26/2022 13:56:06 - INFO - codeparrot_training - Step 46603: {'lr': 6.153591113247036e-06, 'samples': 23861248, 'steps': 46603, 'loss/train': 1.6494197845458984} 02/26/2022 13:56:12 - INFO - codeparrot_training - Step 46604: {'lr': 6.149983615806015e-06, 'samples': 23861760, 'steps': 46604, 'loss/train': 1.5091625452041626} 02/26/2022 13:56:15 - INFO - codeparrot_training - Step 46605: {'lr': 6.14637716294103e-06, 'samples': 23862272, 'steps': 46605, 'loss/train': 0.6269817352294922} 02/26/2022 13:56:23 - INFO - codeparrot_training - Step 46606: {'lr': 6.142771754667514e-06, 'samples': 23862784, 'steps': 46606, 'loss/train': 1.720922827720642} 02/26/2022 13:56:26 - INFO - codeparrot_training - Step 46607: {'lr': 6.139167391000927e-06, 'samples': 23863296, 'steps': 46607, 'loss/train': 1.9761912822723389} 02/26/2022 13:56:32 - INFO - codeparrot_training - Step 46608: {'lr': 6.135564071956729e-06, 'samples': 23863808, 'steps': 46608, 'loss/train': 1.5667837858200073} 02/26/2022 13:56:35 - INFO - codeparrot_training - Step 46609: {'lr': 6.1319617975503505e-06, 'samples': 23864320, 'steps': 46609, 'loss/train': 1.2838783264160156} 02/26/2022 13:56:41 - INFO - codeparrot_training - Step 46610: {'lr': 6.128360567797198e-06, 'samples': 23864832, 'steps': 46610, 'loss/train': 1.78585684299469} 02/26/2022 13:56:44 - INFO - codeparrot_training - Step 46611: {'lr': 6.124760382712674e-06, 'samples': 23865344, 'steps': 46611, 'loss/train': 1.6675465106964111} 02/26/2022 13:56:50 - INFO - codeparrot_training - Step 46612: {'lr': 6.121161242312268e-06, 'samples': 23865856, 'steps': 46612, 'loss/train': 2.134647846221924} 02/26/2022 13:56:53 - INFO - codeparrot_training - Step 46613: {'lr': 6.117563146611382e-06, 'samples': 23866368, 'steps': 46613, 'loss/train': 2.284396171569824} 02/26/2022 13:56:59 - INFO - codeparrot_training - Step 46614: {'lr': 6.113966095625395e-06, 'samples': 23866880, 'steps': 46614, 'loss/train': 1.4648711681365967} 02/26/2022 13:57:02 - INFO - codeparrot_training - Step 46615: {'lr': 6.110370089369766e-06, 'samples': 23867392, 'steps': 46615, 'loss/train': 2.062326431274414} 02/26/2022 13:57:08 - INFO - codeparrot_training - Step 46616: {'lr': 6.106775127859815e-06, 'samples': 23867904, 'steps': 46616, 'loss/train': 2.564812421798706} 02/26/2022 13:57:11 - INFO - codeparrot_training - Step 46617: {'lr': 6.103181211111031e-06, 'samples': 23868416, 'steps': 46617, 'loss/train': 1.0981943607330322} 02/26/2022 13:57:16 - INFO - codeparrot_training - Step 46618: {'lr': 6.099588339138762e-06, 'samples': 23868928, 'steps': 46618, 'loss/train': 1.1418571472167969} 02/26/2022 13:57:20 - INFO - codeparrot_training - Step 46619: {'lr': 6.09599651195844e-06, 'samples': 23869440, 'steps': 46619, 'loss/train': 1.1658334732055664} 02/26/2022 13:57:27 - INFO - codeparrot_training - Step 46620: {'lr': 6.092405729585359e-06, 'samples': 23869952, 'steps': 46620, 'loss/train': 1.969346046447754} 02/26/2022 13:57:30 - INFO - codeparrot_training - Step 46621: {'lr': 6.0888159920350075e-06, 'samples': 23870464, 'steps': 46621, 'loss/train': 2.5843257904052734} 02/26/2022 13:57:36 - INFO - codeparrot_training - Step 46622: {'lr': 6.085227299322676e-06, 'samples': 23870976, 'steps': 46622, 'loss/train': 1.7043955326080322} 02/26/2022 13:57:39 - INFO - codeparrot_training - Step 46623: {'lr': 6.081639651463855e-06, 'samples': 23871488, 'steps': 46623, 'loss/train': 1.9177359342575073} 02/26/2022 13:57:45 - INFO - codeparrot_training - Step 46624: {'lr': 6.0780530484737805e-06, 'samples': 23872000, 'steps': 46624, 'loss/train': 3.2926406860351562} 02/26/2022 13:57:48 - INFO - codeparrot_training - Step 46625: {'lr': 6.0744674903678855e-06, 'samples': 23872512, 'steps': 46625, 'loss/train': 1.9187465906143188} 02/26/2022 13:57:54 - INFO - codeparrot_training - Step 46626: {'lr': 6.07088297716149e-06, 'samples': 23873024, 'steps': 46626, 'loss/train': 1.8744899034500122} 02/26/2022 13:57:57 - INFO - codeparrot_training - Step 46627: {'lr': 6.06729950887e-06, 'samples': 23873536, 'steps': 46627, 'loss/train': 1.0220211744308472} 02/26/2022 13:58:03 - INFO - codeparrot_training - Step 46628: {'lr': 6.063717085508763e-06, 'samples': 23874048, 'steps': 46628, 'loss/train': 1.101572036743164} 02/26/2022 13:58:06 - INFO - codeparrot_training - Step 46629: {'lr': 6.060135707093073e-06, 'samples': 23874560, 'steps': 46629, 'loss/train': 1.3812395334243774} 02/26/2022 13:58:13 - INFO - codeparrot_training - Step 46630: {'lr': 6.056555373638306e-06, 'samples': 23875072, 'steps': 46630, 'loss/train': 1.8080193996429443} 02/26/2022 13:58:17 - INFO - codeparrot_training - Step 46631: {'lr': 6.052976085159783e-06, 'samples': 23875584, 'steps': 46631, 'loss/train': 0.9176732301712036} 02/26/2022 13:58:22 - INFO - codeparrot_training - Step 46632: {'lr': 6.049397841672882e-06, 'samples': 23876096, 'steps': 46632, 'loss/train': 1.2067385911941528} 02/26/2022 13:58:26 - INFO - codeparrot_training - Step 46633: {'lr': 6.045820643192895e-06, 'samples': 23876608, 'steps': 46633, 'loss/train': 1.5077482461929321} 02/26/2022 13:58:32 - INFO - codeparrot_training - Step 46634: {'lr': 6.042244489735116e-06, 'samples': 23877120, 'steps': 46634, 'loss/train': 2.1659486293792725} 02/26/2022 13:58:35 - INFO - codeparrot_training - Step 46635: {'lr': 6.038669381314921e-06, 'samples': 23877632, 'steps': 46635, 'loss/train': 1.721977710723877} 02/26/2022 13:58:41 - INFO - codeparrot_training - Step 46636: {'lr': 6.035095317947603e-06, 'samples': 23878144, 'steps': 46636, 'loss/train': 1.7914661169052124} 02/26/2022 13:58:44 - INFO - codeparrot_training - Step 46637: {'lr': 6.0315222996484565e-06, 'samples': 23878656, 'steps': 46637, 'loss/train': 1.6718448400497437} 02/26/2022 13:58:50 - INFO - codeparrot_training - Step 46638: {'lr': 6.027950326432774e-06, 'samples': 23879168, 'steps': 46638, 'loss/train': 1.4683914184570312} 02/26/2022 13:58:53 - INFO - codeparrot_training - Step 46639: {'lr': 6.024379398315933e-06, 'samples': 23879680, 'steps': 46639, 'loss/train': 0.6901495456695557} 02/26/2022 13:59:01 - INFO - codeparrot_training - Step 46640: {'lr': 6.020809515313141e-06, 'samples': 23880192, 'steps': 46640, 'loss/train': 1.9465771913528442} 02/26/2022 13:59:04 - INFO - codeparrot_training - Step 46641: {'lr': 6.0172406774397785e-06, 'samples': 23880704, 'steps': 46641, 'loss/train': 1.105635643005371} 02/26/2022 13:59:10 - INFO - codeparrot_training - Step 46642: {'lr': 6.013672884711025e-06, 'samples': 23881216, 'steps': 46642, 'loss/train': 1.9301466941833496} 02/26/2022 13:59:13 - INFO - codeparrot_training - Step 46643: {'lr': 6.0101061371422575e-06, 'samples': 23881728, 'steps': 46643, 'loss/train': 1.4662811756134033} 02/26/2022 13:59:18 - INFO - codeparrot_training - Step 46644: {'lr': 6.0065404347486864e-06, 'samples': 23882240, 'steps': 46644, 'loss/train': 1.681962251663208} 02/26/2022 13:59:22 - INFO - codeparrot_training - Step 46645: {'lr': 6.002975777545689e-06, 'samples': 23882752, 'steps': 46645, 'loss/train': 1.3840441703796387} 02/26/2022 13:59:28 - INFO - codeparrot_training - Step 46646: {'lr': 5.99941216554839e-06, 'samples': 23883264, 'steps': 46646, 'loss/train': 2.5264551639556885} 02/26/2022 13:59:31 - INFO - codeparrot_training - Step 46647: {'lr': 5.995849598772169e-06, 'samples': 23883776, 'steps': 46647, 'loss/train': 1.7247811555862427} 02/26/2022 13:59:37 - INFO - codeparrot_training - Step 46648: {'lr': 5.992288077232261e-06, 'samples': 23884288, 'steps': 46648, 'loss/train': 1.6359807252883911} 02/26/2022 13:59:40 - INFO - codeparrot_training - Step 46649: {'lr': 5.988727600943878e-06, 'samples': 23884800, 'steps': 46649, 'loss/train': 2.1761908531188965} 02/26/2022 13:59:46 - INFO - codeparrot_training - Step 46650: {'lr': 5.985168169922311e-06, 'samples': 23885312, 'steps': 46650, 'loss/train': 2.2371673583984375} 02/26/2022 13:59:49 - INFO - codeparrot_training - Step 46651: {'lr': 5.981609784182801e-06, 'samples': 23885824, 'steps': 46651, 'loss/train': 1.9049779176712036} 02/26/2022 13:59:56 - INFO - codeparrot_training - Step 46652: {'lr': 5.978052443740584e-06, 'samples': 23886336, 'steps': 46652, 'loss/train': 1.652034878730774} 02/26/2022 14:00:00 - INFO - codeparrot_training - Step 46653: {'lr': 5.974496148610897e-06, 'samples': 23886848, 'steps': 46653, 'loss/train': 2.0710084438323975} 02/26/2022 14:00:05 - INFO - codeparrot_training - Step 46654: {'lr': 5.970940898809008e-06, 'samples': 23887360, 'steps': 46654, 'loss/train': 1.9836301803588867} 02/26/2022 14:00:09 - INFO - codeparrot_training - Step 46655: {'lr': 5.9673866943501245e-06, 'samples': 23887872, 'steps': 46655, 'loss/train': 1.204217791557312} 02/26/2022 14:00:14 - INFO - codeparrot_training - Step 46656: {'lr': 5.963833535249458e-06, 'samples': 23888384, 'steps': 46656, 'loss/train': 1.724589228630066} 02/26/2022 14:00:18 - INFO - codeparrot_training - Step 46657: {'lr': 5.960281421522218e-06, 'samples': 23888896, 'steps': 46657, 'loss/train': 2.331415891647339} 02/26/2022 14:00:23 - INFO - codeparrot_training - Step 46658: {'lr': 5.95673035318367e-06, 'samples': 23889408, 'steps': 46658, 'loss/train': 0.16043223440647125} 02/26/2022 14:00:27 - INFO - codeparrot_training - Step 46659: {'lr': 5.95318033024897e-06, 'samples': 23889920, 'steps': 46659, 'loss/train': 3.100290060043335} 02/26/2022 14:00:32 - INFO - codeparrot_training - Step 46660: {'lr': 5.9496313527333814e-06, 'samples': 23890432, 'steps': 46660, 'loss/train': 0.701377272605896} 02/26/2022 14:00:36 - INFO - codeparrot_training - Step 46661: {'lr': 5.946083420652032e-06, 'samples': 23890944, 'steps': 46661, 'loss/train': 1.7472103834152222} 02/26/2022 14:00:41 - INFO - codeparrot_training - Step 46662: {'lr': 5.942536534020215e-06, 'samples': 23891456, 'steps': 46662, 'loss/train': 1.168807864189148} 02/26/2022 14:00:45 - INFO - codeparrot_training - Step 46663: {'lr': 5.938990692853058e-06, 'samples': 23891968, 'steps': 46663, 'loss/train': 3.1657886505126953} 02/26/2022 14:00:50 - INFO - codeparrot_training - Step 46664: {'lr': 5.93544589716577e-06, 'samples': 23892480, 'steps': 46664, 'loss/train': 2.089587688446045} 02/26/2022 14:00:54 - INFO - codeparrot_training - Step 46665: {'lr': 5.931902146973506e-06, 'samples': 23892992, 'steps': 46665, 'loss/train': 1.3023558855056763} 02/26/2022 14:00:59 - INFO - codeparrot_training - Step 46666: {'lr': 5.928359442291503e-06, 'samples': 23893504, 'steps': 46666, 'loss/train': 1.4478703737258911} 02/26/2022 14:01:03 - INFO - codeparrot_training - Step 46667: {'lr': 5.924817783134889e-06, 'samples': 23894016, 'steps': 46667, 'loss/train': 1.4270223379135132} 02/26/2022 14:01:10 - INFO - codeparrot_training - Step 46668: {'lr': 5.921277169518874e-06, 'samples': 23894528, 'steps': 46668, 'loss/train': 3.0140647888183594} 02/26/2022 14:01:13 - INFO - codeparrot_training - Step 46669: {'lr': 5.917737601458556e-06, 'samples': 23895040, 'steps': 46669, 'loss/train': 2.3817083835601807} 02/26/2022 14:01:19 - INFO - codeparrot_training - Step 46670: {'lr': 5.914199078969202e-06, 'samples': 23895552, 'steps': 46670, 'loss/train': 0.9312611818313599} 02/26/2022 14:01:22 - INFO - codeparrot_training - Step 46671: {'lr': 5.910661602065881e-06, 'samples': 23896064, 'steps': 46671, 'loss/train': 2.0018935203552246} 02/26/2022 14:01:28 - INFO - codeparrot_training - Step 46672: {'lr': 5.907125170763805e-06, 'samples': 23896576, 'steps': 46672, 'loss/train': 1.690332055091858} 02/26/2022 14:01:31 - INFO - codeparrot_training - Step 46673: {'lr': 5.903589785078073e-06, 'samples': 23897088, 'steps': 46673, 'loss/train': 1.9078359603881836} 02/26/2022 14:01:37 - INFO - codeparrot_training - Step 46674: {'lr': 5.900055445023839e-06, 'samples': 23897600, 'steps': 46674, 'loss/train': 1.2770737409591675} 02/26/2022 14:01:40 - INFO - codeparrot_training - Step 46675: {'lr': 5.8965221506162845e-06, 'samples': 23898112, 'steps': 46675, 'loss/train': 2.472773790359497} 02/26/2022 14:01:46 - INFO - codeparrot_training - Step 46676: {'lr': 5.892989901870538e-06, 'samples': 23898624, 'steps': 46676, 'loss/train': 2.0403401851654053} 02/26/2022 14:01:49 - INFO - codeparrot_training - Step 46677: {'lr': 5.889458698801642e-06, 'samples': 23899136, 'steps': 46677, 'loss/train': 2.0397324562072754} 02/26/2022 14:01:57 - INFO - codeparrot_training - Step 46678: {'lr': 5.885928541424862e-06, 'samples': 23899648, 'steps': 46678, 'loss/train': 2.402076244354248} 02/26/2022 14:02:00 - INFO - codeparrot_training - Step 46679: {'lr': 5.882399429755187e-06, 'samples': 23900160, 'steps': 46679, 'loss/train': 1.7445820569992065} 02/26/2022 14:02:06 - INFO - codeparrot_training - Step 46680: {'lr': 5.878871363807853e-06, 'samples': 23900672, 'steps': 46680, 'loss/train': 1.826703429222107} 02/26/2022 14:02:09 - INFO - codeparrot_training - Step 46681: {'lr': 5.875344343597877e-06, 'samples': 23901184, 'steps': 46681, 'loss/train': 2.3417375087738037} 02/26/2022 14:02:15 - INFO - codeparrot_training - Step 46682: {'lr': 5.871818369140442e-06, 'samples': 23901696, 'steps': 46682, 'loss/train': 1.435309886932373} 02/26/2022 14:02:18 - INFO - codeparrot_training - Step 46683: {'lr': 5.8682934404505615e-06, 'samples': 23902208, 'steps': 46683, 'loss/train': 2.4600021839141846} 02/26/2022 14:02:24 - INFO - codeparrot_training - Step 46684: {'lr': 5.8647695575434756e-06, 'samples': 23902720, 'steps': 46684, 'loss/train': 2.341658353805542} 02/26/2022 14:02:27 - INFO - codeparrot_training - Step 46685: {'lr': 5.861246720434115e-06, 'samples': 23903232, 'steps': 46685, 'loss/train': 1.5958586931228638} 02/26/2022 14:02:32 - INFO - codeparrot_training - Step 46686: {'lr': 5.857724929137692e-06, 'samples': 23903744, 'steps': 46686, 'loss/train': 1.2421454191207886} 02/26/2022 14:02:36 - INFO - codeparrot_training - Step 46687: {'lr': 5.854204183669221e-06, 'samples': 23904256, 'steps': 46687, 'loss/train': 2.580960750579834} 02/26/2022 14:02:43 - INFO - codeparrot_training - Step 46688: {'lr': 5.850684484043856e-06, 'samples': 23904768, 'steps': 46688, 'loss/train': 1.64923095703125} 02/26/2022 14:02:47 - INFO - codeparrot_training - Step 46689: {'lr': 5.847165830276613e-06, 'samples': 23905280, 'steps': 46689, 'loss/train': 1.640005111694336} 02/26/2022 14:02:52 - INFO - codeparrot_training - Step 46690: {'lr': 5.843648222382591e-06, 'samples': 23905792, 'steps': 46690, 'loss/train': 0.9614103436470032} 02/26/2022 14:02:56 - INFO - codeparrot_training - Step 46691: {'lr': 5.840131660376835e-06, 'samples': 23906304, 'steps': 46691, 'loss/train': 2.1645987033843994} 02/26/2022 14:03:02 - INFO - codeparrot_training - Step 46692: {'lr': 5.836616144274442e-06, 'samples': 23906816, 'steps': 46692, 'loss/train': 0.7547193765640259} 02/26/2022 14:03:05 - INFO - codeparrot_training - Step 46693: {'lr': 5.833101674090457e-06, 'samples': 23907328, 'steps': 46693, 'loss/train': 1.377374291419983} 02/26/2022 14:03:10 - INFO - codeparrot_training - Step 46694: {'lr': 5.829588249839923e-06, 'samples': 23907840, 'steps': 46694, 'loss/train': 0.6536068320274353} 02/26/2022 14:03:14 - INFO - codeparrot_training - Step 46695: {'lr': 5.826075871537911e-06, 'samples': 23908352, 'steps': 46695, 'loss/train': 0.4406853914260864} 02/26/2022 14:03:20 - INFO - codeparrot_training - Step 46696: {'lr': 5.8225645391994095e-06, 'samples': 23908864, 'steps': 46696, 'loss/train': 2.3112378120422363} 02/26/2022 14:03:23 - INFO - codeparrot_training - Step 46697: {'lr': 5.819054252839573e-06, 'samples': 23909376, 'steps': 46697, 'loss/train': 4.267125606536865} 02/26/2022 14:03:30 - INFO - codeparrot_training - Step 46698: {'lr': 5.815545012473334e-06, 'samples': 23909888, 'steps': 46698, 'loss/train': 1.966652274131775} 02/26/2022 14:03:34 - INFO - codeparrot_training - Step 46699: {'lr': 5.81203681811579e-06, 'samples': 23910400, 'steps': 46699, 'loss/train': 0.8440201878547668} 02/26/2022 14:03:39 - INFO - codeparrot_training - Step 46700: {'lr': 5.808529669781903e-06, 'samples': 23910912, 'steps': 46700, 'loss/train': 1.2726991176605225} 02/26/2022 14:03:43 - INFO - codeparrot_training - Step 46701: {'lr': 5.805023567486745e-06, 'samples': 23911424, 'steps': 46701, 'loss/train': 1.7687829732894897} 02/26/2022 14:03:49 - INFO - codeparrot_training - Step 46702: {'lr': 5.801518511245357e-06, 'samples': 23911936, 'steps': 46702, 'loss/train': 1.2757089138031006} 02/26/2022 14:03:52 - INFO - codeparrot_training - Step 46703: {'lr': 5.798014501072701e-06, 'samples': 23912448, 'steps': 46703, 'loss/train': 2.372622489929199} 02/26/2022 14:03:58 - INFO - codeparrot_training - Step 46704: {'lr': 5.794511536983793e-06, 'samples': 23912960, 'steps': 46704, 'loss/train': 1.5417133569717407} 02/26/2022 14:04:01 - INFO - codeparrot_training - Step 46705: {'lr': 5.791009618993676e-06, 'samples': 23913472, 'steps': 46705, 'loss/train': 2.2302517890930176} 02/26/2022 14:04:07 - INFO - codeparrot_training - Step 46706: {'lr': 5.7875087471173095e-06, 'samples': 23913984, 'steps': 46706, 'loss/train': 2.0182416439056396} 02/26/2022 14:04:10 - INFO - codeparrot_training - Step 46707: {'lr': 5.784008921369738e-06, 'samples': 23914496, 'steps': 46707, 'loss/train': 1.498546838760376} 02/26/2022 14:04:16 - INFO - codeparrot_training - Step 46708: {'lr': 5.780510141765893e-06, 'samples': 23915008, 'steps': 46708, 'loss/train': 1.0078778266906738} 02/26/2022 14:04:19 - INFO - codeparrot_training - Step 46709: {'lr': 5.777012408320792e-06, 'samples': 23915520, 'steps': 46709, 'loss/train': 1.6712477207183838} 02/26/2022 14:04:25 - INFO - codeparrot_training - Step 46710: {'lr': 5.7735157210494214e-06, 'samples': 23916032, 'steps': 46710, 'loss/train': 0.14980806410312653} 02/26/2022 14:04:28 - INFO - codeparrot_training - Step 46711: {'lr': 5.770020079966826e-06, 'samples': 23916544, 'steps': 46711, 'loss/train': 1.2432053089141846} 02/26/2022 14:04:33 - INFO - codeparrot_training - Step 46712: {'lr': 5.766525485087826e-06, 'samples': 23917056, 'steps': 46712, 'loss/train': 1.7230957746505737} 02/26/2022 14:04:37 - INFO - codeparrot_training - Step 46713: {'lr': 5.763031936427521e-06, 'samples': 23917568, 'steps': 46713, 'loss/train': 2.444340229034424} 02/26/2022 14:04:44 - INFO - codeparrot_training - Step 46714: {'lr': 5.759539434000816e-06, 'samples': 23918080, 'steps': 46714, 'loss/train': 2.7467739582061768} 02/26/2022 14:04:48 - INFO - codeparrot_training - Step 46715: {'lr': 5.756047977822726e-06, 'samples': 23918592, 'steps': 46715, 'loss/train': 2.4739832878112793} 02/26/2022 14:04:53 - INFO - codeparrot_training - Step 46716: {'lr': 5.752557567908129e-06, 'samples': 23919104, 'steps': 46716, 'loss/train': 2.059391975402832} 02/26/2022 14:04:57 - INFO - codeparrot_training - Step 46717: {'lr': 5.74906820427204e-06, 'samples': 23919616, 'steps': 46717, 'loss/train': 1.3544613122940063} 02/26/2022 14:05:02 - INFO - codeparrot_training - Step 46718: {'lr': 5.745579886929336e-06, 'samples': 23920128, 'steps': 46718, 'loss/train': 1.9531452655792236} 02/26/2022 14:05:06 - INFO - codeparrot_training - Step 46719: {'lr': 5.742092615895089e-06, 'samples': 23920640, 'steps': 46719, 'loss/train': 1.3006643056869507} 02/26/2022 14:05:11 - INFO - codeparrot_training - Step 46720: {'lr': 5.738606391184093e-06, 'samples': 23921152, 'steps': 46720, 'loss/train': 1.5238136053085327} 02/26/2022 14:05:14 - INFO - codeparrot_training - Step 46721: {'lr': 5.73512121281139e-06, 'samples': 23921664, 'steps': 46721, 'loss/train': 2.314146041870117} 02/26/2022 14:05:20 - INFO - codeparrot_training - Step 46722: {'lr': 5.7316370807918015e-06, 'samples': 23922176, 'steps': 46722, 'loss/train': 1.0148773193359375} 02/26/2022 14:05:23 - INFO - codeparrot_training - Step 46723: {'lr': 5.7281539951403726e-06, 'samples': 23922688, 'steps': 46723, 'loss/train': 1.2166365385055542} 02/26/2022 14:05:31 - INFO - codeparrot_training - Step 46724: {'lr': 5.724671955871951e-06, 'samples': 23923200, 'steps': 46724, 'loss/train': 1.886130928993225} 02/26/2022 14:05:34 - INFO - codeparrot_training - Step 46725: {'lr': 5.721190963001499e-06, 'samples': 23923712, 'steps': 46725, 'loss/train': 1.9318009614944458} 02/26/2022 14:05:40 - INFO - codeparrot_training - Step 46726: {'lr': 5.717711016543836e-06, 'samples': 23924224, 'steps': 46726, 'loss/train': 1.7105162143707275} 02/26/2022 14:05:43 - INFO - codeparrot_training - Step 46727: {'lr': 5.714232116513979e-06, 'samples': 23924736, 'steps': 46727, 'loss/train': 1.3184913396835327} 02/26/2022 14:05:49 - INFO - codeparrot_training - Step 46728: {'lr': 5.710754262926776e-06, 'samples': 23925248, 'steps': 46728, 'loss/train': 1.3114069700241089} 02/26/2022 14:05:52 - INFO - codeparrot_training - Step 46729: {'lr': 5.707277455797105e-06, 'samples': 23925760, 'steps': 46729, 'loss/train': 1.8186323642730713} 02/26/2022 14:05:58 - INFO - codeparrot_training - Step 46730: {'lr': 5.703801695139899e-06, 'samples': 23926272, 'steps': 46730, 'loss/train': 2.53918719291687} 02/26/2022 14:06:01 - INFO - codeparrot_training - Step 46731: {'lr': 5.7003269809700335e-06, 'samples': 23926784, 'steps': 46731, 'loss/train': 2.002119779586792} 02/26/2022 14:06:07 - INFO - codeparrot_training - Step 46732: {'lr': 5.696853313302414e-06, 'samples': 23927296, 'steps': 46732, 'loss/train': 1.4633177518844604} 02/26/2022 14:06:10 - INFO - codeparrot_training - Step 46733: {'lr': 5.693380692151889e-06, 'samples': 23927808, 'steps': 46733, 'loss/train': 2.9672770500183105} 02/26/2022 14:06:18 - INFO - codeparrot_training - Step 46734: {'lr': 5.689909117533309e-06, 'samples': 23928320, 'steps': 46734, 'loss/train': 1.9776338338851929} 02/26/2022 14:06:21 - INFO - codeparrot_training - Step 46735: {'lr': 5.686438589461606e-06, 'samples': 23928832, 'steps': 46735, 'loss/train': 1.1969789266586304} 02/26/2022 14:06:26 - INFO - codeparrot_training - Step 46736: {'lr': 5.68296910795163e-06, 'samples': 23929344, 'steps': 46736, 'loss/train': 1.754347562789917} 02/26/2022 14:06:30 - INFO - codeparrot_training - Step 46737: {'lr': 5.679500673018257e-06, 'samples': 23929856, 'steps': 46737, 'loss/train': 1.4279024600982666} 02/26/2022 14:06:36 - INFO - codeparrot_training - Step 46738: {'lr': 5.67603328467628e-06, 'samples': 23930368, 'steps': 46738, 'loss/train': 1.0800254344940186} 02/26/2022 14:06:39 - INFO - codeparrot_training - Step 46739: {'lr': 5.672566942940605e-06, 'samples': 23930880, 'steps': 46739, 'loss/train': 2.2174394130706787} 02/26/2022 14:06:45 - INFO - codeparrot_training - Step 46740: {'lr': 5.66910164782608e-06, 'samples': 23931392, 'steps': 46740, 'loss/train': 0.46291375160217285} 02/26/2022 14:06:48 - INFO - codeparrot_training - Step 46741: {'lr': 5.6656373993475565e-06, 'samples': 23931904, 'steps': 46741, 'loss/train': 1.7776920795440674} 02/26/2022 14:06:54 - INFO - codeparrot_training - Step 46742: {'lr': 5.662174197519826e-06, 'samples': 23932416, 'steps': 46742, 'loss/train': 1.7433170080184937} 02/26/2022 14:06:57 - INFO - codeparrot_training - Step 46743: {'lr': 5.658712042357766e-06, 'samples': 23932928, 'steps': 46743, 'loss/train': 1.5523154735565186} 02/26/2022 14:07:04 - INFO - codeparrot_training - Step 46744: {'lr': 5.655250933876199e-06, 'samples': 23933440, 'steps': 46744, 'loss/train': 2.678130626678467} 02/26/2022 14:07:08 - INFO - codeparrot_training - Step 46745: {'lr': 5.651790872089946e-06, 'samples': 23933952, 'steps': 46745, 'loss/train': 1.7714459896087646} 02/26/2022 14:07:13 - INFO - codeparrot_training - Step 46746: {'lr': 5.648331857013855e-06, 'samples': 23934464, 'steps': 46746, 'loss/train': 2.5082831382751465} 02/26/2022 14:07:17 - INFO - codeparrot_training - Step 46747: {'lr': 5.644873888662666e-06, 'samples': 23934976, 'steps': 46747, 'loss/train': 2.53240966796875} 02/26/2022 14:07:22 - INFO - codeparrot_training - Step 46748: {'lr': 5.641416967051283e-06, 'samples': 23935488, 'steps': 46748, 'loss/train': 1.7202811241149902} 02/26/2022 14:07:26 - INFO - codeparrot_training - Step 46749: {'lr': 5.637961092194471e-06, 'samples': 23936000, 'steps': 46749, 'loss/train': 1.3155291080474854} 02/26/2022 14:07:31 - INFO - codeparrot_training - Step 46750: {'lr': 5.634506264107053e-06, 'samples': 23936512, 'steps': 46750, 'loss/train': 2.2414345741271973} 02/26/2022 14:07:35 - INFO - codeparrot_training - Step 46751: {'lr': 5.631052482803795e-06, 'samples': 23937024, 'steps': 46751, 'loss/train': 2.1189589500427246} 02/26/2022 14:07:40 - INFO - codeparrot_training - Step 46752: {'lr': 5.627599748299516e-06, 'samples': 23937536, 'steps': 46752, 'loss/train': 2.3657515048980713} 02/26/2022 14:07:44 - INFO - codeparrot_training - Step 46753: {'lr': 5.624148060608986e-06, 'samples': 23938048, 'steps': 46753, 'loss/train': 1.374780535697937} 02/26/2022 14:07:49 - INFO - codeparrot_training - Step 46754: {'lr': 5.620697419747078e-06, 'samples': 23938560, 'steps': 46754, 'loss/train': 0.5056179165840149} 02/26/2022 14:07:53 - INFO - codeparrot_training - Step 46755: {'lr': 5.6172478257284224e-06, 'samples': 23939072, 'steps': 46755, 'loss/train': 0.8102360367774963} 02/26/2022 14:07:58 - INFO - codeparrot_training - Step 46756: {'lr': 5.613799278567922e-06, 'samples': 23939584, 'steps': 46756, 'loss/train': 2.0188512802124023} 02/26/2022 14:08:02 - INFO - codeparrot_training - Step 46757: {'lr': 5.610351778280259e-06, 'samples': 23940096, 'steps': 46757, 'loss/train': 1.7047706842422485} 02/26/2022 14:08:07 - INFO - codeparrot_training - Step 46758: {'lr': 5.606905324880313e-06, 'samples': 23940608, 'steps': 46758, 'loss/train': 1.7402641773223877} 02/26/2022 14:08:11 - INFO - codeparrot_training - Step 46759: {'lr': 5.603459918382736e-06, 'samples': 23941120, 'steps': 46759, 'loss/train': 2.0662107467651367} 02/26/2022 14:08:18 - INFO - codeparrot_training - Step 46760: {'lr': 5.600015558802352e-06, 'samples': 23941632, 'steps': 46760, 'loss/train': 2.454516887664795} 02/26/2022 14:08:21 - INFO - codeparrot_training - Step 46761: {'lr': 5.596572246153869e-06, 'samples': 23942144, 'steps': 46761, 'loss/train': 2.2911412715911865} 02/26/2022 14:08:27 - INFO - codeparrot_training - Step 46762: {'lr': 5.593129980452083e-06, 'samples': 23942656, 'steps': 46762, 'loss/train': 1.1962132453918457} 02/26/2022 14:08:30 - INFO - codeparrot_training - Step 46763: {'lr': 5.589688761711703e-06, 'samples': 23943168, 'steps': 46763, 'loss/train': 2.023801565170288} 02/26/2022 14:08:36 - INFO - codeparrot_training - Step 46764: {'lr': 5.586248589947523e-06, 'samples': 23943680, 'steps': 46764, 'loss/train': 2.108915328979492} 02/26/2022 14:08:39 - INFO - codeparrot_training - Step 46765: {'lr': 5.582809465174199e-06, 'samples': 23944192, 'steps': 46765, 'loss/train': 2.0662779808044434} 02/26/2022 14:08:45 - INFO - codeparrot_training - Step 46766: {'lr': 5.579371387406551e-06, 'samples': 23944704, 'steps': 46766, 'loss/train': 1.2213648557662964} 02/26/2022 14:08:50 - INFO - codeparrot_training - Step 46767: {'lr': 5.575934356659235e-06, 'samples': 23945216, 'steps': 46767, 'loss/train': 0.6193291544914246} 02/26/2022 14:08:54 - INFO - codeparrot_training - Step 46768: {'lr': 5.572498372947016e-06, 'samples': 23945728, 'steps': 46768, 'loss/train': 1.8453292846679688} 02/26/2022 14:09:01 - INFO - codeparrot_training - Step 46769: {'lr': 5.569063436284577e-06, 'samples': 23946240, 'steps': 46769, 'loss/train': 1.2371264696121216} 02/26/2022 14:09:04 - INFO - codeparrot_training - Step 46770: {'lr': 5.5656295466866845e-06, 'samples': 23946752, 'steps': 46770, 'loss/train': 2.5571675300598145} 02/26/2022 14:09:08 - INFO - codeparrot_training - Step 46771: {'lr': 5.5621967041679935e-06, 'samples': 23947264, 'steps': 46771, 'loss/train': 1.4545338153839111} 02/26/2022 14:09:13 - INFO - codeparrot_training - Step 46772: {'lr': 5.558764908743269e-06, 'samples': 23947776, 'steps': 46772, 'loss/train': 2.4007105827331543} 02/26/2022 14:09:17 - INFO - codeparrot_training - Step 46773: {'lr': 5.555334160427139e-06, 'samples': 23948288, 'steps': 46773, 'loss/train': 2.6095728874206543} 02/26/2022 14:09:22 - INFO - codeparrot_training - Step 46774: {'lr': 5.551904459234369e-06, 'samples': 23948800, 'steps': 46774, 'loss/train': 1.1040265560150146} 02/26/2022 14:09:26 - INFO - codeparrot_training - Step 46775: {'lr': 5.548475805179587e-06, 'samples': 23949312, 'steps': 46775, 'loss/train': 1.3597338199615479} 02/26/2022 14:09:31 - INFO - codeparrot_training - Step 46776: {'lr': 5.545048198277558e-06, 'samples': 23949824, 'steps': 46776, 'loss/train': 1.2872953414916992} 02/26/2022 14:09:35 - INFO - codeparrot_training - Step 46777: {'lr': 5.5416216385428816e-06, 'samples': 23950336, 'steps': 46777, 'loss/train': 2.621816635131836} 02/26/2022 14:09:40 - INFO - codeparrot_training - Step 46778: {'lr': 5.538196125990269e-06, 'samples': 23950848, 'steps': 46778, 'loss/train': 1.016654133796692} 02/26/2022 14:09:44 - INFO - codeparrot_training - Step 46779: {'lr': 5.53477166063443e-06, 'samples': 23951360, 'steps': 46779, 'loss/train': 1.1899681091308594} 02/26/2022 14:09:51 - INFO - codeparrot_training - Step 46780: {'lr': 5.531348242489992e-06, 'samples': 23951872, 'steps': 46780, 'loss/train': 2.40132212638855} 02/26/2022 14:09:54 - INFO - codeparrot_training - Step 46781: {'lr': 5.5279258715716096e-06, 'samples': 23952384, 'steps': 46781, 'loss/train': 1.9497724771499634} 02/26/2022 14:10:00 - INFO - codeparrot_training - Step 46782: {'lr': 5.524504547893994e-06, 'samples': 23952896, 'steps': 46782, 'loss/train': 1.4064050912857056} 02/26/2022 14:10:03 - INFO - codeparrot_training - Step 46783: {'lr': 5.521084271471744e-06, 'samples': 23953408, 'steps': 46783, 'loss/train': 1.335425853729248} 02/26/2022 14:10:09 - INFO - codeparrot_training - Step 46784: {'lr': 5.517665042319542e-06, 'samples': 23953920, 'steps': 46784, 'loss/train': 1.67551589012146} 02/26/2022 14:10:15 - INFO - codeparrot_training - Step 46785: {'lr': 5.5142468604520435e-06, 'samples': 23954432, 'steps': 46785, 'loss/train': 1.8045179843902588} 02/26/2022 14:10:18 - INFO - codeparrot_training - Step 46786: {'lr': 5.5108297258838755e-06, 'samples': 23954944, 'steps': 46786, 'loss/train': 1.6436898708343506} 02/26/2022 14:10:23 - INFO - codeparrot_training - Step 46787: {'lr': 5.507413638629693e-06, 'samples': 23955456, 'steps': 46787, 'loss/train': 1.7725476026535034} 02/26/2022 14:10:27 - INFO - codeparrot_training - Step 46788: {'lr': 5.5039985987040686e-06, 'samples': 23955968, 'steps': 46788, 'loss/train': 1.9082043170928955} 02/26/2022 14:10:34 - INFO - codeparrot_training - Step 46789: {'lr': 5.500584606121739e-06, 'samples': 23956480, 'steps': 46789, 'loss/train': 1.1810981035232544} 02/26/2022 14:10:38 - INFO - codeparrot_training - Step 46790: {'lr': 5.497171660897221e-06, 'samples': 23956992, 'steps': 46790, 'loss/train': 3.068063735961914} 02/26/2022 14:10:43 - INFO - codeparrot_training - Step 46791: {'lr': 5.493759763045225e-06, 'samples': 23957504, 'steps': 46791, 'loss/train': 1.1218230724334717} 02/26/2022 14:10:47 - INFO - codeparrot_training - Step 46792: {'lr': 5.490348912580268e-06, 'samples': 23958016, 'steps': 46792, 'loss/train': 1.0819718837738037} 02/26/2022 14:10:52 - INFO - codeparrot_training - Step 46793: {'lr': 5.486939109517086e-06, 'samples': 23958528, 'steps': 46793, 'loss/train': 1.4107733964920044} 02/26/2022 14:10:56 - INFO - codeparrot_training - Step 46794: {'lr': 5.483530353870142e-06, 'samples': 23959040, 'steps': 46794, 'loss/train': 0.8035626411437988} 02/26/2022 14:10:59 - INFO - codeparrot_training - Step 46795: {'lr': 5.480122645654145e-06, 'samples': 23959552, 'steps': 46795, 'loss/train': 2.2564501762390137} 02/26/2022 14:11:05 - INFO - codeparrot_training - Step 46796: {'lr': 5.47671598488364e-06, 'samples': 23960064, 'steps': 46796, 'loss/train': 0.7502437233924866} 02/26/2022 14:11:08 - INFO - codeparrot_training - Step 46797: {'lr': 5.4733103715732535e-06, 'samples': 23960576, 'steps': 46797, 'loss/train': 1.5987889766693115} 02/26/2022 14:11:14 - INFO - codeparrot_training - Step 46798: {'lr': 5.469905805737585e-06, 'samples': 23961088, 'steps': 46798, 'loss/train': 1.4029862880706787} 02/26/2022 14:11:17 - INFO - codeparrot_training - Step 46799: {'lr': 5.466502287391179e-06, 'samples': 23961600, 'steps': 46799, 'loss/train': 2.0328757762908936} 02/26/2022 14:11:23 - INFO - codeparrot_training - Step 46800: {'lr': 5.463099816548578e-06, 'samples': 23962112, 'steps': 46800, 'loss/train': 2.150216817855835} 02/26/2022 14:11:26 - INFO - codeparrot_training - Step 46801: {'lr': 5.459698393224466e-06, 'samples': 23962624, 'steps': 46801, 'loss/train': 2.3617630004882812} 02/26/2022 14:11:32 - INFO - codeparrot_training - Step 46802: {'lr': 5.456298017433331e-06, 'samples': 23963136, 'steps': 46802, 'loss/train': 1.958127737045288} 02/26/2022 14:11:35 - INFO - codeparrot_training - Step 46803: {'lr': 5.4528986891898e-06, 'samples': 23963648, 'steps': 46803, 'loss/train': 0.9350897669792175} 02/26/2022 14:11:41 - INFO - codeparrot_training - Step 46804: {'lr': 5.449500408508334e-06, 'samples': 23964160, 'steps': 46804, 'loss/train': 1.5976330041885376} 02/26/2022 14:11:44 - INFO - codeparrot_training - Step 46805: {'lr': 5.446103175403616e-06, 'samples': 23964672, 'steps': 46805, 'loss/train': 1.387372612953186} 02/26/2022 14:11:52 - INFO - codeparrot_training - Step 46806: {'lr': 5.442706989890106e-06, 'samples': 23965184, 'steps': 46806, 'loss/train': 0.8588217496871948} 02/26/2022 14:11:55 - INFO - codeparrot_training - Step 46807: {'lr': 5.439311851982404e-06, 'samples': 23965696, 'steps': 46807, 'loss/train': 2.1299164295196533} 02/26/2022 14:12:00 - INFO - codeparrot_training - Step 46808: {'lr': 5.435917761694998e-06, 'samples': 23966208, 'steps': 46808, 'loss/train': 1.5720378160476685} 02/26/2022 14:12:04 - INFO - codeparrot_training - Step 46809: {'lr': 5.432524719042487e-06, 'samples': 23966720, 'steps': 46809, 'loss/train': 1.6974263191223145} 02/26/2022 14:12:09 - INFO - codeparrot_training - Step 46810: {'lr': 5.42913272403936e-06, 'samples': 23967232, 'steps': 46810, 'loss/train': 3.2405030727386475} 02/26/2022 14:12:13 - INFO - codeparrot_training - Step 46811: {'lr': 5.425741776700216e-06, 'samples': 23967744, 'steps': 46811, 'loss/train': 0.40171346068382263} 02/26/2022 14:12:18 - INFO - codeparrot_training - Step 46812: {'lr': 5.4223518770394884e-06, 'samples': 23968256, 'steps': 46812, 'loss/train': 1.8758174180984497} 02/26/2022 14:12:22 - INFO - codeparrot_training - Step 46813: {'lr': 5.418963025071777e-06, 'samples': 23968768, 'steps': 46813, 'loss/train': 1.6103911399841309} 02/26/2022 14:12:27 - INFO - codeparrot_training - Step 46814: {'lr': 5.415575220811514e-06, 'samples': 23969280, 'steps': 46814, 'loss/train': 1.8378190994262695} 02/26/2022 14:12:31 - INFO - codeparrot_training - Step 46815: {'lr': 5.412188464273327e-06, 'samples': 23969792, 'steps': 46815, 'loss/train': 0.42845645546913147} 02/26/2022 14:12:39 - INFO - codeparrot_training - Step 46816: {'lr': 5.40880275547162e-06, 'samples': 23970304, 'steps': 46816, 'loss/train': 0.4071495234966278} 02/26/2022 14:12:42 - INFO - codeparrot_training - Step 46817: {'lr': 5.405418094420939e-06, 'samples': 23970816, 'steps': 46817, 'loss/train': 1.4022789001464844} 02/26/2022 14:12:48 - INFO - codeparrot_training - Step 46818: {'lr': 5.40203448113577e-06, 'samples': 23971328, 'steps': 46818, 'loss/train': 2.10654616355896} 02/26/2022 14:12:51 - INFO - codeparrot_training - Step 46819: {'lr': 5.398651915630631e-06, 'samples': 23971840, 'steps': 46819, 'loss/train': 1.2048009634017944} 02/26/2022 14:12:57 - INFO - codeparrot_training - Step 46820: {'lr': 5.39527039792001e-06, 'samples': 23972352, 'steps': 46820, 'loss/train': 0.7934987545013428} 02/26/2022 14:13:00 - INFO - codeparrot_training - Step 46821: {'lr': 5.391889928018368e-06, 'samples': 23972864, 'steps': 46821, 'loss/train': 1.607393503189087} 02/26/2022 14:13:06 - INFO - codeparrot_training - Step 46822: {'lr': 5.388510505940164e-06, 'samples': 23973376, 'steps': 46822, 'loss/train': 3.01770281791687} 02/26/2022 14:13:09 - INFO - codeparrot_training - Step 46823: {'lr': 5.385132131699943e-06, 'samples': 23973888, 'steps': 46823, 'loss/train': 1.8295824527740479} 02/26/2022 14:13:15 - INFO - codeparrot_training - Step 46824: {'lr': 5.381754805312139e-06, 'samples': 23974400, 'steps': 46824, 'loss/train': 1.118936538696289} 02/26/2022 14:13:19 - INFO - codeparrot_training - Step 46825: {'lr': 5.378378526791239e-06, 'samples': 23974912, 'steps': 46825, 'loss/train': 1.9161953926086426} 02/26/2022 14:13:26 - INFO - codeparrot_training - Step 46826: {'lr': 5.375003296151676e-06, 'samples': 23975424, 'steps': 46826, 'loss/train': 1.5617352724075317} 02/26/2022 14:13:29 - INFO - codeparrot_training - Step 46827: {'lr': 5.371629113407911e-06, 'samples': 23975936, 'steps': 46827, 'loss/train': 0.651698112487793} 02/26/2022 14:13:35 - INFO - codeparrot_training - Step 46828: {'lr': 5.3682559785744045e-06, 'samples': 23976448, 'steps': 46828, 'loss/train': 1.2936040163040161} 02/26/2022 14:13:38 - INFO - codeparrot_training - Step 46829: {'lr': 5.364883891665645e-06, 'samples': 23976960, 'steps': 46829, 'loss/train': 1.6932175159454346} 02/26/2022 14:13:44 - INFO - codeparrot_training - Step 46830: {'lr': 5.36151285269601e-06, 'samples': 23977472, 'steps': 46830, 'loss/train': 0.7406253218650818} 02/26/2022 14:13:47 - INFO - codeparrot_training - Step 46831: {'lr': 5.35814286167996e-06, 'samples': 23977984, 'steps': 46831, 'loss/train': 1.9916950464248657} 02/26/2022 14:13:53 - INFO - codeparrot_training - Step 46832: {'lr': 5.3547739186319836e-06, 'samples': 23978496, 'steps': 46832, 'loss/train': 0.9536425471305847} 02/26/2022 14:13:56 - INFO - codeparrot_training - Step 46833: {'lr': 5.351406023566457e-06, 'samples': 23979008, 'steps': 46833, 'loss/train': 1.3169853687286377} 02/26/2022 14:14:02 - INFO - codeparrot_training - Step 46834: {'lr': 5.348039176497816e-06, 'samples': 23979520, 'steps': 46834, 'loss/train': 2.296553373336792} 02/26/2022 14:14:06 - INFO - codeparrot_training - Step 46835: {'lr': 5.3446733774404895e-06, 'samples': 23980032, 'steps': 46835, 'loss/train': 2.886805534362793} 02/26/2022 14:14:13 - INFO - codeparrot_training - Step 46836: {'lr': 5.341308626408886e-06, 'samples': 23980544, 'steps': 46836, 'loss/train': 1.98550546169281} 02/26/2022 14:14:16 - INFO - codeparrot_training - Step 46837: {'lr': 5.337944923417465e-06, 'samples': 23981056, 'steps': 46837, 'loss/train': 2.3220250606536865} 02/26/2022 14:14:22 - INFO - codeparrot_training - Step 46838: {'lr': 5.334582268480575e-06, 'samples': 23981568, 'steps': 46838, 'loss/train': 2.136589288711548} 02/26/2022 14:14:25 - INFO - codeparrot_training - Step 46839: {'lr': 5.3312206616126235e-06, 'samples': 23982080, 'steps': 46839, 'loss/train': 2.316633462905884} 02/26/2022 14:14:31 - INFO - codeparrot_training - Step 46840: {'lr': 5.327860102828041e-06, 'samples': 23982592, 'steps': 46840, 'loss/train': 2.326310396194458} 02/26/2022 14:14:34 - INFO - codeparrot_training - Step 46841: {'lr': 5.324500592141235e-06, 'samples': 23983104, 'steps': 46841, 'loss/train': 1.8999918699264526} 02/26/2022 14:14:40 - INFO - codeparrot_training - Step 46842: {'lr': 5.321142129566581e-06, 'samples': 23983616, 'steps': 46842, 'loss/train': 0.9576194286346436} 02/26/2022 14:14:43 - INFO - codeparrot_training - Step 46843: {'lr': 5.317784715118429e-06, 'samples': 23984128, 'steps': 46843, 'loss/train': 1.371591329574585} 02/26/2022 14:14:49 - INFO - codeparrot_training - Step 46844: {'lr': 5.314428348811212e-06, 'samples': 23984640, 'steps': 46844, 'loss/train': 2.247483968734741} 02/26/2022 14:14:52 - INFO - codeparrot_training - Step 46845: {'lr': 5.311073030659252e-06, 'samples': 23985152, 'steps': 46845, 'loss/train': 2.291668176651001} 02/26/2022 14:14:58 - INFO - codeparrot_training - Step 46846: {'lr': 5.307718760677011e-06, 'samples': 23985664, 'steps': 46846, 'loss/train': 1.7852822542190552} 02/26/2022 14:15:01 - INFO - codeparrot_training - Step 46847: {'lr': 5.30436553887878e-06, 'samples': 23986176, 'steps': 46847, 'loss/train': 1.6136611700057983} 02/26/2022 14:15:07 - INFO - codeparrot_training - Step 46848: {'lr': 5.30101336527894e-06, 'samples': 23986688, 'steps': 46848, 'loss/train': 1.0306272506713867} 02/26/2022 14:15:10 - INFO - codeparrot_training - Step 46849: {'lr': 5.297662239891865e-06, 'samples': 23987200, 'steps': 46849, 'loss/train': 0.9531620740890503} 02/26/2022 14:15:16 - INFO - codeparrot_training - Step 46850: {'lr': 5.294312162731935e-06, 'samples': 23987712, 'steps': 46850, 'loss/train': 1.41732656955719} 02/26/2022 14:15:19 - INFO - codeparrot_training - Step 46851: {'lr': 5.290963133813415e-06, 'samples': 23988224, 'steps': 46851, 'loss/train': 1.2813444137573242} 02/26/2022 14:15:27 - INFO - codeparrot_training - Step 46852: {'lr': 5.287615153150737e-06, 'samples': 23988736, 'steps': 46852, 'loss/train': 2.7821009159088135} 02/26/2022 14:15:30 - INFO - codeparrot_training - Step 46853: {'lr': 5.284268220758171e-06, 'samples': 23989248, 'steps': 46853, 'loss/train': 2.277313709259033} 02/26/2022 14:15:36 - INFO - codeparrot_training - Step 46854: {'lr': 5.2809223366501734e-06, 'samples': 23989760, 'steps': 46854, 'loss/train': 2.1209797859191895} 02/26/2022 14:15:39 - INFO - codeparrot_training - Step 46855: {'lr': 5.27757750084093e-06, 'samples': 23990272, 'steps': 46855, 'loss/train': 2.236997127532959} 02/26/2022 14:15:44 - INFO - codeparrot_training - Step 46856: {'lr': 5.274233713344845e-06, 'samples': 23990784, 'steps': 46856, 'loss/train': 1.7585680484771729} 02/26/2022 14:15:48 - INFO - codeparrot_training - Step 46857: {'lr': 5.27089097417624e-06, 'samples': 23991296, 'steps': 46857, 'loss/train': 0.3344886600971222} 02/26/2022 14:15:54 - INFO - codeparrot_training - Step 46858: {'lr': 5.267549283349438e-06, 'samples': 23991808, 'steps': 46858, 'loss/train': 1.0785777568817139} 02/26/2022 14:15:57 - INFO - codeparrot_training - Step 46859: {'lr': 5.264208640878732e-06, 'samples': 23992320, 'steps': 46859, 'loss/train': 2.6864705085754395} 02/26/2022 14:16:03 - INFO - codeparrot_training - Step 46860: {'lr': 5.260869046778471e-06, 'samples': 23992832, 'steps': 46860, 'loss/train': 1.2688500881195068} 02/26/2022 14:16:06 - INFO - codeparrot_training - Step 46861: {'lr': 5.257530501062896e-06, 'samples': 23993344, 'steps': 46861, 'loss/train': 1.5397019386291504} 02/26/2022 14:16:12 - INFO - codeparrot_training - Step 46862: {'lr': 5.254193003746355e-06, 'samples': 23993856, 'steps': 46862, 'loss/train': 2.134564161300659} 02/26/2022 14:16:15 - INFO - codeparrot_training - Step 46863: {'lr': 5.250856554843142e-06, 'samples': 23994368, 'steps': 46863, 'loss/train': 1.7671452760696411} 02/26/2022 14:16:22 - INFO - codeparrot_training - Step 46864: {'lr': 5.247521154367552e-06, 'samples': 23994880, 'steps': 46864, 'loss/train': 1.6623950004577637} 02/26/2022 14:16:26 - INFO - codeparrot_training - Step 46865: {'lr': 5.24418680233385e-06, 'samples': 23995392, 'steps': 46865, 'loss/train': 1.0587592124938965} 02/26/2022 14:16:31 - INFO - codeparrot_training - Step 46866: {'lr': 5.2408534987563316e-06, 'samples': 23995904, 'steps': 46866, 'loss/train': 1.7487235069274902} 02/26/2022 14:16:35 - INFO - codeparrot_training - Step 46867: {'lr': 5.23752124364929e-06, 'samples': 23996416, 'steps': 46867, 'loss/train': 1.8403130769729614} 02/26/2022 14:16:40 - INFO - codeparrot_training - Step 46868: {'lr': 5.234190037026992e-06, 'samples': 23996928, 'steps': 46868, 'loss/train': 1.5571342706680298} 02/26/2022 14:16:44 - INFO - codeparrot_training - Step 46869: {'lr': 5.230859878903676e-06, 'samples': 23997440, 'steps': 46869, 'loss/train': 1.5209988355636597} 02/26/2022 14:16:49 - INFO - codeparrot_training - Step 46870: {'lr': 5.227530769293637e-06, 'samples': 23997952, 'steps': 46870, 'loss/train': 1.699662208557129} 02/26/2022 14:16:53 - INFO - codeparrot_training - Step 46871: {'lr': 5.224202708211168e-06, 'samples': 23998464, 'steps': 46871, 'loss/train': 2.245939016342163} 02/26/2022 14:16:58 - INFO - codeparrot_training - Step 46872: {'lr': 5.220875695670452e-06, 'samples': 23998976, 'steps': 46872, 'loss/train': 2.2990434169769287} 02/26/2022 14:17:02 - INFO - codeparrot_training - Step 46873: {'lr': 5.217549731685811e-06, 'samples': 23999488, 'steps': 46873, 'loss/train': 0.8813782334327698} 02/26/2022 14:17:09 - INFO - codeparrot_training - Step 46874: {'lr': 5.214224816271429e-06, 'samples': 24000000, 'steps': 46874, 'loss/train': 0.6537315249443054} 02/26/2022 14:17:13 - INFO - codeparrot_training - Step 46875: {'lr': 5.210900949441571e-06, 'samples': 24000512, 'steps': 46875, 'loss/train': 2.4819328784942627} 02/26/2022 14:17:18 - INFO - codeparrot_training - Step 46876: {'lr': 5.207578131210533e-06, 'samples': 24001024, 'steps': 46876, 'loss/train': 1.7477836608886719} 02/26/2022 14:17:22 - INFO - codeparrot_training - Step 46877: {'lr': 5.204256361592469e-06, 'samples': 24001536, 'steps': 46877, 'loss/train': 1.5514863729476929} 02/26/2022 14:17:27 - INFO - codeparrot_training - Step 46878: {'lr': 5.200935640601617e-06, 'samples': 24002048, 'steps': 46878, 'loss/train': 2.7574987411499023} 02/26/2022 14:17:31 - INFO - codeparrot_training - Step 46879: {'lr': 5.197615968252245e-06, 'samples': 24002560, 'steps': 46879, 'loss/train': 2.185171127319336} 02/26/2022 14:17:36 - INFO - codeparrot_training - Step 46880: {'lr': 5.194297344558535e-06, 'samples': 24003072, 'steps': 46880, 'loss/train': 0.9565368294715881} 02/26/2022 14:17:40 - INFO - codeparrot_training - Step 46881: {'lr': 5.190979769534782e-06, 'samples': 24003584, 'steps': 46881, 'loss/train': 1.2464548349380493} 02/26/2022 14:17:45 - INFO - codeparrot_training - Step 46882: {'lr': 5.1876632431950845e-06, 'samples': 24004096, 'steps': 46882, 'loss/train': 1.5799823999404907} 02/26/2022 14:17:49 - INFO - codeparrot_training - Step 46883: {'lr': 5.18434776555371e-06, 'samples': 24004608, 'steps': 46883, 'loss/train': 0.8450912833213806} 02/26/2022 14:17:56 - INFO - codeparrot_training - Step 46884: {'lr': 5.181033336624841e-06, 'samples': 24005120, 'steps': 46884, 'loss/train': 1.6226387023925781} 02/26/2022 14:17:59 - INFO - codeparrot_training - Step 46885: {'lr': 5.177719956422716e-06, 'samples': 24005632, 'steps': 46885, 'loss/train': 1.790230631828308} 02/26/2022 14:18:05 - INFO - codeparrot_training - Step 46886: {'lr': 5.174407624961463e-06, 'samples': 24006144, 'steps': 46886, 'loss/train': 1.0855509042739868} 02/26/2022 14:18:09 - INFO - codeparrot_training - Step 46887: {'lr': 5.171096342255349e-06, 'samples': 24006656, 'steps': 46887, 'loss/train': 2.2399508953094482} 02/26/2022 14:18:14 - INFO - codeparrot_training - Step 46888: {'lr': 5.167786108318473e-06, 'samples': 24007168, 'steps': 46888, 'loss/train': 1.3225986957550049} 02/26/2022 14:18:18 - INFO - codeparrot_training - Step 46889: {'lr': 5.164476923165101e-06, 'samples': 24007680, 'steps': 46889, 'loss/train': 1.6188924312591553} 02/26/2022 14:18:23 - INFO - codeparrot_training - Step 46890: {'lr': 5.161168786809334e-06, 'samples': 24008192, 'steps': 46890, 'loss/train': 1.9687080383300781} 02/26/2022 14:18:27 - INFO - codeparrot_training - Step 46891: {'lr': 5.157861699265382e-06, 'samples': 24008704, 'steps': 46891, 'loss/train': 1.893865942955017} 02/26/2022 14:18:32 - INFO - codeparrot_training - Step 46892: {'lr': 5.1545556605473724e-06, 'samples': 24009216, 'steps': 46892, 'loss/train': 1.6712392568588257} 02/26/2022 14:18:36 - INFO - codeparrot_training - Step 46893: {'lr': 5.151250670669544e-06, 'samples': 24009728, 'steps': 46893, 'loss/train': 2.116797924041748} 02/26/2022 14:18:41 - INFO - codeparrot_training - Step 46894: {'lr': 5.147946729645997e-06, 'samples': 24010240, 'steps': 46894, 'loss/train': 0.9294131398200989} 02/26/2022 14:18:45 - INFO - codeparrot_training - Step 46895: {'lr': 5.144643837490914e-06, 'samples': 24010752, 'steps': 46895, 'loss/train': 1.5480866432189941} 02/26/2022 14:18:50 - INFO - codeparrot_training - Step 46896: {'lr': 5.1413419942183955e-06, 'samples': 24011264, 'steps': 46896, 'loss/train': 1.5727174282073975} 02/26/2022 14:18:54 - INFO - codeparrot_training - Step 46897: {'lr': 5.138041199842653e-06, 'samples': 24011776, 'steps': 46897, 'loss/train': 2.2603986263275146} 02/26/2022 14:19:02 - INFO - codeparrot_training - Step 46898: {'lr': 5.134741454377756e-06, 'samples': 24012288, 'steps': 46898, 'loss/train': 1.9924979209899902} 02/26/2022 14:19:05 - INFO - codeparrot_training - Step 46899: {'lr': 5.13144275783789e-06, 'samples': 24012800, 'steps': 46899, 'loss/train': 1.704669713973999} 02/26/2022 14:19:11 - INFO - codeparrot_training - Step 46900: {'lr': 5.128145110237154e-06, 'samples': 24013312, 'steps': 46900, 'loss/train': 1.674017310142517} 02/26/2022 14:19:14 - INFO - codeparrot_training - Step 46901: {'lr': 5.124848511589703e-06, 'samples': 24013824, 'steps': 46901, 'loss/train': 1.8421716690063477} 02/26/2022 14:19:20 - INFO - codeparrot_training - Step 46902: {'lr': 5.1215529619096365e-06, 'samples': 24014336, 'steps': 46902, 'loss/train': 1.8025825023651123} 02/26/2022 14:19:23 - INFO - codeparrot_training - Step 46903: {'lr': 5.118258461211084e-06, 'samples': 24014848, 'steps': 46903, 'loss/train': 1.0603227615356445} 02/26/2022 14:19:29 - INFO - codeparrot_training - Step 46904: {'lr': 5.114965009508143e-06, 'samples': 24015360, 'steps': 46904, 'loss/train': 1.1741602420806885} 02/26/2022 14:19:32 - INFO - codeparrot_training - Step 46905: {'lr': 5.111672606814944e-06, 'samples': 24015872, 'steps': 46905, 'loss/train': 0.34088361263275146} 02/26/2022 14:19:38 - INFO - codeparrot_training - Step 46906: {'lr': 5.1083812531455835e-06, 'samples': 24016384, 'steps': 46906, 'loss/train': 2.0534825325012207} 02/26/2022 14:19:41 - INFO - codeparrot_training - Step 46907: {'lr': 5.105090948514163e-06, 'samples': 24016896, 'steps': 46907, 'loss/train': 2.3434557914733887} 02/26/2022 14:19:48 - INFO - codeparrot_training - Step 46908: {'lr': 5.101801692934727e-06, 'samples': 24017408, 'steps': 46908, 'loss/train': 0.6817491054534912} 02/26/2022 14:19:52 - INFO - codeparrot_training - Step 46909: {'lr': 5.098513486421458e-06, 'samples': 24017920, 'steps': 46909, 'loss/train': 2.259007453918457} 02/26/2022 14:19:57 - INFO - codeparrot_training - Step 46910: {'lr': 5.095226328988373e-06, 'samples': 24018432, 'steps': 46910, 'loss/train': 1.4829856157302856} 02/26/2022 14:20:01 - INFO - codeparrot_training - Step 46911: {'lr': 5.091940220649571e-06, 'samples': 24018944, 'steps': 46911, 'loss/train': 1.1623817682266235} 02/26/2022 14:20:06 - INFO - codeparrot_training - Step 46912: {'lr': 5.088655161419126e-06, 'samples': 24019456, 'steps': 46912, 'loss/train': 1.6048622131347656} 02/26/2022 14:20:10 - INFO - codeparrot_training - Step 46913: {'lr': 5.085371151311108e-06, 'samples': 24019968, 'steps': 46913, 'loss/train': 1.5661735534667969} 02/26/2022 14:20:15 - INFO - codeparrot_training - Step 46914: {'lr': 5.082088190339618e-06, 'samples': 24020480, 'steps': 46914, 'loss/train': 2.6421194076538086} 02/26/2022 14:20:19 - INFO - codeparrot_training - Step 46915: {'lr': 5.078806278518672e-06, 'samples': 24020992, 'steps': 46915, 'loss/train': 1.5811841487884521} 02/26/2022 14:20:24 - INFO - codeparrot_training - Step 46916: {'lr': 5.075525415862342e-06, 'samples': 24021504, 'steps': 46916, 'loss/train': 2.1828420162200928} 02/26/2022 14:20:28 - INFO - codeparrot_training - Step 46917: {'lr': 5.072245602384701e-06, 'samples': 24022016, 'steps': 46917, 'loss/train': 2.3859052658081055} 02/26/2022 14:20:33 - INFO - codeparrot_training - Step 46918: {'lr': 5.068966838099793e-06, 'samples': 24022528, 'steps': 46918, 'loss/train': 1.5189543962478638} 02/26/2022 14:20:37 - INFO - codeparrot_training - Step 46919: {'lr': 5.0656891230216325e-06, 'samples': 24023040, 'steps': 46919, 'loss/train': 1.4126026630401611} 02/26/2022 14:20:45 - INFO - codeparrot_training - Step 46920: {'lr': 5.0624124571643225e-06, 'samples': 24023552, 'steps': 46920, 'loss/train': 1.1939884424209595} 02/26/2022 14:20:48 - INFO - codeparrot_training - Step 46921: {'lr': 5.059136840541823e-06, 'samples': 24024064, 'steps': 46921, 'loss/train': 1.0683212280273438} 02/26/2022 14:20:53 - INFO - codeparrot_training - Step 46922: {'lr': 5.055862273168233e-06, 'samples': 24024576, 'steps': 46922, 'loss/train': 1.4167451858520508} 02/26/2022 14:20:57 - INFO - codeparrot_training - Step 46923: {'lr': 5.052588755057513e-06, 'samples': 24025088, 'steps': 46923, 'loss/train': 2.0334348678588867} 02/26/2022 14:21:03 - INFO - codeparrot_training - Step 46924: {'lr': 5.049316286223765e-06, 'samples': 24025600, 'steps': 46924, 'loss/train': 2.108639717102051} 02/26/2022 14:21:06 - INFO - codeparrot_training - Step 46925: {'lr': 5.046044866680949e-06, 'samples': 24026112, 'steps': 46925, 'loss/train': 1.158650517463684} 02/26/2022 14:21:12 - INFO - codeparrot_training - Step 46926: {'lr': 5.042774496443081e-06, 'samples': 24026624, 'steps': 46926, 'loss/train': 1.2206578254699707} 02/26/2022 14:21:15 - INFO - codeparrot_training - Step 46927: {'lr': 5.039505175524178e-06, 'samples': 24027136, 'steps': 46927, 'loss/train': 0.8434476852416992} 02/26/2022 14:21:21 - INFO - codeparrot_training - Step 46928: {'lr': 5.0362369039382845e-06, 'samples': 24027648, 'steps': 46928, 'loss/train': 2.3522393703460693} 02/26/2022 14:21:24 - INFO - codeparrot_training - Step 46929: {'lr': 5.032969681699362e-06, 'samples': 24028160, 'steps': 46929, 'loss/train': 2.226112127304077} 02/26/2022 14:21:31 - INFO - codeparrot_training - Step 46930: {'lr': 5.029703508821398e-06, 'samples': 24028672, 'steps': 46930, 'loss/train': 1.7182066440582275} 02/26/2022 14:21:35 - INFO - codeparrot_training - Step 46931: {'lr': 5.026438385318382e-06, 'samples': 24029184, 'steps': 46931, 'loss/train': 0.6680001020431519} 02/26/2022 14:21:40 - INFO - codeparrot_training - Step 46932: {'lr': 5.0231743112043595e-06, 'samples': 24029696, 'steps': 46932, 'loss/train': 1.7184449434280396} 02/26/2022 14:21:44 - INFO - codeparrot_training - Step 46933: {'lr': 5.019911286493234e-06, 'samples': 24030208, 'steps': 46933, 'loss/train': 1.6746971607208252} 02/26/2022 14:21:49 - INFO - codeparrot_training - Step 46934: {'lr': 5.01664931119905e-06, 'samples': 24030720, 'steps': 46934, 'loss/train': 1.3019837141036987} 02/26/2022 14:21:53 - INFO - codeparrot_training - Step 46935: {'lr': 5.013388385335715e-06, 'samples': 24031232, 'steps': 46935, 'loss/train': 1.9929417371749878} 02/26/2022 14:21:58 - INFO - codeparrot_training - Step 46936: {'lr': 5.0101285089172445e-06, 'samples': 24031744, 'steps': 46936, 'loss/train': 2.174039125442505} 02/26/2022 14:22:02 - INFO - codeparrot_training - Step 46937: {'lr': 5.006869681957599e-06, 'samples': 24032256, 'steps': 46937, 'loss/train': 2.351221799850464} 02/26/2022 14:22:07 - INFO - codeparrot_training - Step 46938: {'lr': 5.003611904470739e-06, 'samples': 24032768, 'steps': 46938, 'loss/train': 1.9174375534057617} 02/26/2022 14:22:11 - INFO - codeparrot_training - Step 46939: {'lr': 5.000355176470572e-06, 'samples': 24033280, 'steps': 46939, 'loss/train': 0.2696290910243988} 02/26/2022 14:22:17 - INFO - codeparrot_training - Step 46940: {'lr': 4.997099497971114e-06, 'samples': 24033792, 'steps': 46940, 'loss/train': 2.0405056476593018} 02/26/2022 14:22:20 - INFO - codeparrot_training - Step 46941: {'lr': 4.993844868986269e-06, 'samples': 24034304, 'steps': 46941, 'loss/train': 1.7250773906707764} 02/26/2022 14:22:26 - INFO - codeparrot_training - Step 46942: {'lr': 4.990591289530028e-06, 'samples': 24034816, 'steps': 46942, 'loss/train': 2.0224058628082275} 02/26/2022 14:22:29 - INFO - codeparrot_training - Step 46943: {'lr': 4.987338759616267e-06, 'samples': 24035328, 'steps': 46943, 'loss/train': 1.268270492553711} 02/26/2022 14:22:36 - INFO - codeparrot_training - Step 46944: {'lr': 4.984087279258948e-06, 'samples': 24035840, 'steps': 46944, 'loss/train': 1.3498919010162354} 02/26/2022 14:22:40 - INFO - codeparrot_training - Step 46945: {'lr': 4.980836848471976e-06, 'samples': 24036352, 'steps': 46945, 'loss/train': 0.8643433451652527} 02/26/2022 14:22:45 - INFO - codeparrot_training - Step 46946: {'lr': 4.9775874672693674e-06, 'samples': 24036864, 'steps': 46946, 'loss/train': 1.3721646070480347} 02/26/2022 14:22:49 - INFO - codeparrot_training - Step 46947: {'lr': 4.9743391356648905e-06, 'samples': 24037376, 'steps': 46947, 'loss/train': 2.0076985359191895} 02/26/2022 14:22:54 - INFO - codeparrot_training - Step 46948: {'lr': 4.971091853672588e-06, 'samples': 24037888, 'steps': 46948, 'loss/train': 1.3897405862808228} 02/26/2022 14:22:58 - INFO - codeparrot_training - Step 46949: {'lr': 4.967845621306283e-06, 'samples': 24038400, 'steps': 46949, 'loss/train': 2.2570042610168457} 02/26/2022 14:23:03 - INFO - codeparrot_training - Step 46950: {'lr': 4.964600438579991e-06, 'samples': 24038912, 'steps': 46950, 'loss/train': 1.2331100702285767} 02/26/2022 14:23:07 - INFO - codeparrot_training - Step 46951: {'lr': 4.961356305507481e-06, 'samples': 24039424, 'steps': 46951, 'loss/train': 2.7906792163848877} 02/26/2022 14:23:12 - INFO - codeparrot_training - Step 46952: {'lr': 4.958113222102739e-06, 'samples': 24039936, 'steps': 46952, 'loss/train': 1.2655057907104492} 02/26/2022 14:23:16 - INFO - codeparrot_training - Step 46953: {'lr': 4.954871188379617e-06, 'samples': 24040448, 'steps': 46953, 'loss/train': 1.2923851013183594} 02/26/2022 14:23:23 - INFO - codeparrot_training - Step 46954: {'lr': 4.95163020435202e-06, 'samples': 24040960, 'steps': 46954, 'loss/train': 3.0559093952178955} 02/26/2022 14:23:26 - INFO - codeparrot_training - Step 46955: {'lr': 4.948390270033826e-06, 'samples': 24041472, 'steps': 46955, 'loss/train': 1.5363434553146362} 02/26/2022 14:23:32 - INFO - codeparrot_training - Step 46956: {'lr': 4.945151385438912e-06, 'samples': 24041984, 'steps': 46956, 'loss/train': 1.6995373964309692} 02/26/2022 14:23:35 - INFO - codeparrot_training - Step 46957: {'lr': 4.941913550581184e-06, 'samples': 24042496, 'steps': 46957, 'loss/train': 0.6171913743019104} 02/26/2022 14:23:41 - INFO - codeparrot_training - Step 46958: {'lr': 4.938676765474437e-06, 'samples': 24043008, 'steps': 46958, 'loss/train': 1.4346436262130737} 02/26/2022 14:23:44 - INFO - codeparrot_training - Step 46959: {'lr': 4.935441030132631e-06, 'samples': 24043520, 'steps': 46959, 'loss/train': 1.4168815612792969} 02/26/2022 14:23:50 - INFO - codeparrot_training - Step 46960: {'lr': 4.932206344569562e-06, 'samples': 24044032, 'steps': 46960, 'loss/train': 1.081868052482605} 02/26/2022 14:23:53 - INFO - codeparrot_training - Step 46961: {'lr': 4.9289727087990784e-06, 'samples': 24044544, 'steps': 46961, 'loss/train': 1.7110190391540527} 02/26/2022 14:23:59 - INFO - codeparrot_training - Step 46962: {'lr': 4.92574012283506e-06, 'samples': 24045056, 'steps': 46962, 'loss/train': 0.7463676333427429} 02/26/2022 14:24:02 - INFO - codeparrot_training - Step 46963: {'lr': 4.9225085866913825e-06, 'samples': 24045568, 'steps': 46963, 'loss/train': 0.28492891788482666} 02/26/2022 14:24:08 - INFO - codeparrot_training - Step 46964: {'lr': 4.919278100381841e-06, 'samples': 24046080, 'steps': 46964, 'loss/train': 0.6663783192634583} 02/26/2022 14:24:11 - INFO - codeparrot_training - Step 46965: {'lr': 4.916048663920314e-06, 'samples': 24046592, 'steps': 46965, 'loss/train': 2.271629810333252} 02/26/2022 14:24:19 - INFO - codeparrot_training - Step 46966: {'lr': 4.912820277320568e-06, 'samples': 24047104, 'steps': 46966, 'loss/train': 1.2768833637237549} 02/26/2022 14:24:23 - INFO - codeparrot_training - Step 46967: {'lr': 4.9095929405965075e-06, 'samples': 24047616, 'steps': 46967, 'loss/train': 2.0954296588897705} 02/26/2022 14:24:28 - INFO - codeparrot_training - Step 46968: {'lr': 4.906366653761901e-06, 'samples': 24048128, 'steps': 46968, 'loss/train': 2.1196422576904297} 02/26/2022 14:24:31 - INFO - codeparrot_training - Step 46969: {'lr': 4.903141416830625e-06, 'samples': 24048640, 'steps': 46969, 'loss/train': 0.7138397693634033} 02/26/2022 14:24:37 - INFO - codeparrot_training - Step 46970: {'lr': 4.8999172298164195e-06, 'samples': 24049152, 'steps': 46970, 'loss/train': 2.2004547119140625} 02/26/2022 14:24:40 - INFO - codeparrot_training - Step 46971: {'lr': 4.896694092733189e-06, 'samples': 24049664, 'steps': 46971, 'loss/train': 2.0217840671539307} 02/26/2022 14:24:46 - INFO - codeparrot_training - Step 46972: {'lr': 4.893472005594674e-06, 'samples': 24050176, 'steps': 46972, 'loss/train': 2.4144418239593506} 02/26/2022 14:24:50 - INFO - codeparrot_training - Step 46973: {'lr': 4.890250968414694e-06, 'samples': 24050688, 'steps': 46973, 'loss/train': 1.228333830833435} 02/26/2022 14:24:55 - INFO - codeparrot_training - Step 46974: {'lr': 4.887030981207019e-06, 'samples': 24051200, 'steps': 46974, 'loss/train': 1.450463891029358} 02/26/2022 14:24:58 - INFO - codeparrot_training - Step 46975: {'lr': 4.883812043985497e-06, 'samples': 24051712, 'steps': 46975, 'loss/train': 1.1599994897842407} 02/26/2022 14:25:06 - INFO - codeparrot_training - Step 46976: {'lr': 4.880594156763896e-06, 'samples': 24052224, 'steps': 46976, 'loss/train': 1.5338329076766968} 02/26/2022 14:25:09 - INFO - codeparrot_training - Step 46977: {'lr': 4.877377319556009e-06, 'samples': 24052736, 'steps': 46977, 'loss/train': 2.1217143535614014} 02/26/2022 14:25:15 - INFO - codeparrot_training - Step 46978: {'lr': 4.874161532375576e-06, 'samples': 24053248, 'steps': 46978, 'loss/train': 1.2750530242919922} 02/26/2022 14:25:18 - INFO - codeparrot_training - Step 46979: {'lr': 4.8709467952364194e-06, 'samples': 24053760, 'steps': 46979, 'loss/train': 2.248365879058838} 02/26/2022 14:25:24 - INFO - codeparrot_training - Step 46980: {'lr': 4.86773310815225e-06, 'samples': 24054272, 'steps': 46980, 'loss/train': 2.3472487926483154} 02/26/2022 14:25:27 - INFO - codeparrot_training - Step 46981: {'lr': 4.864520471136919e-06, 'samples': 24054784, 'steps': 46981, 'loss/train': 1.949555516242981} 02/26/2022 14:25:33 - INFO - codeparrot_training - Step 46982: {'lr': 4.861308884204108e-06, 'samples': 24055296, 'steps': 46982, 'loss/train': 2.0906949043273926} 02/26/2022 14:25:36 - INFO - codeparrot_training - Step 46983: {'lr': 4.8580983473676396e-06, 'samples': 24055808, 'steps': 46983, 'loss/train': 2.1700401306152344} 02/26/2022 14:25:42 - INFO - codeparrot_training - Step 46984: {'lr': 4.8548888606411986e-06, 'samples': 24056320, 'steps': 46984, 'loss/train': 1.7726908922195435} 02/26/2022 14:25:45 - INFO - codeparrot_training - Step 46985: {'lr': 4.851680424038635e-06, 'samples': 24056832, 'steps': 46985, 'loss/train': 1.71482515335083} 02/26/2022 14:25:51 - INFO - codeparrot_training - Step 46986: {'lr': 4.848473037573575e-06, 'samples': 24057344, 'steps': 46986, 'loss/train': 1.1402982473373413} 02/26/2022 14:25:54 - INFO - codeparrot_training - Step 46987: {'lr': 4.845266701259815e-06, 'samples': 24057856, 'steps': 46987, 'loss/train': 2.1147561073303223} 02/26/2022 14:26:00 - INFO - codeparrot_training - Step 46988: {'lr': 4.842061415111093e-06, 'samples': 24058368, 'steps': 46988, 'loss/train': 1.3244916200637817} 02/26/2022 14:26:03 - INFO - codeparrot_training - Step 46989: {'lr': 4.838857179141148e-06, 'samples': 24058880, 'steps': 46989, 'loss/train': 1.4193236827850342} 02/26/2022 14:26:09 - INFO - codeparrot_training - Step 46990: {'lr': 4.835653993363693e-06, 'samples': 24059392, 'steps': 46990, 'loss/train': 2.154270648956299} 02/26/2022 14:26:12 - INFO - codeparrot_training - Step 46991: {'lr': 4.832451857792436e-06, 'samples': 24059904, 'steps': 46991, 'loss/train': 1.374183177947998} 02/26/2022 14:26:19 - INFO - codeparrot_training - Step 46992: {'lr': 4.829250772441091e-06, 'samples': 24060416, 'steps': 46992, 'loss/train': 0.976502001285553} 02/26/2022 14:26:23 - INFO - codeparrot_training - Step 46993: {'lr': 4.826050737323423e-06, 'samples': 24060928, 'steps': 46993, 'loss/train': 1.0698624849319458} 02/26/2022 14:26:28 - INFO - codeparrot_training - Step 46994: {'lr': 4.822851752453089e-06, 'samples': 24061440, 'steps': 46994, 'loss/train': 1.904506802558899} 02/26/2022 14:26:32 - INFO - codeparrot_training - Step 46995: {'lr': 4.8196538178438e-06, 'samples': 24061952, 'steps': 46995, 'loss/train': 2.0647518634796143} 02/26/2022 14:26:37 - INFO - codeparrot_training - Step 46996: {'lr': 4.816456933509239e-06, 'samples': 24062464, 'steps': 46996, 'loss/train': 1.026369571685791} 02/26/2022 14:26:41 - INFO - codeparrot_training - Step 46997: {'lr': 4.8132610994631455e-06, 'samples': 24062976, 'steps': 46997, 'loss/train': 2.110910415649414} 02/26/2022 14:26:46 - INFO - codeparrot_training - Step 46998: {'lr': 4.810066315719175e-06, 'samples': 24063488, 'steps': 46998, 'loss/train': 1.6281605958938599} 02/26/2022 14:26:52 - INFO - codeparrot_training - Step 46999: {'lr': 4.806872582291038e-06, 'samples': 24064000, 'steps': 46999, 'loss/train': 1.2316577434539795} 02/26/2022 14:26:52 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 14:27:10 - WARNING - huggingface_hub.repository - Several commits (47) will be pushed upstream. 02/26/2022 14:27:10 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 14:27:44 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 572893f..2e5c310 floral-grass-11 -> floral-grass-11 02/26/2022 14:27:48 - INFO - codeparrot_training - Step 47000: {'lr': 4.803679899192393e-06, 'samples': 24064512, 'steps': 47000, 'loss/train': 1.5690922737121582} 02/26/2022 14:27:56 - INFO - codeparrot_training - Step 47001: {'lr': 4.8004882664369195e-06, 'samples': 24065024, 'steps': 47001, 'loss/train': 2.004244804382324} 02/26/2022 14:27:59 - INFO - codeparrot_training - Step 47002: {'lr': 4.797297684038304e-06, 'samples': 24065536, 'steps': 47002, 'loss/train': 2.0392816066741943} 02/26/2022 14:28:03 - INFO - codeparrot_training - Step 47003: {'lr': 4.794108152010201e-06, 'samples': 24066048, 'steps': 47003, 'loss/train': 2.2683329582214355} 02/26/2022 14:28:08 - INFO - codeparrot_training - Step 47004: {'lr': 4.790919670366295e-06, 'samples': 24066560, 'steps': 47004, 'loss/train': 2.059736490249634} 02/26/2022 14:28:12 - INFO - codeparrot_training - Step 47005: {'lr': 4.787732239120185e-06, 'samples': 24067072, 'steps': 47005, 'loss/train': 0.6227854490280151} 02/26/2022 14:28:17 - INFO - codeparrot_training - Step 47006: {'lr': 4.784545858285611e-06, 'samples': 24067584, 'steps': 47006, 'loss/train': 2.4907712936401367} 02/26/2022 14:28:21 - INFO - codeparrot_training - Step 47007: {'lr': 4.781360527876144e-06, 'samples': 24068096, 'steps': 47007, 'loss/train': 1.6988836526870728} 02/26/2022 14:28:27 - INFO - codeparrot_training - Step 47008: {'lr': 4.778176247905497e-06, 'samples': 24068608, 'steps': 47008, 'loss/train': 1.662015676498413} 02/26/2022 14:28:30 - INFO - codeparrot_training - Step 47009: {'lr': 4.774993018387241e-06, 'samples': 24069120, 'steps': 47009, 'loss/train': 2.578217029571533} 02/26/2022 14:28:35 - INFO - codeparrot_training - Step 47010: {'lr': 4.7718108393350615e-06, 'samples': 24069632, 'steps': 47010, 'loss/train': 2.2534897327423096} 02/26/2022 14:28:39 - INFO - codeparrot_training - Step 47011: {'lr': 4.7686297107625846e-06, 'samples': 24070144, 'steps': 47011, 'loss/train': 0.869590699672699} 02/26/2022 14:28:46 - INFO - codeparrot_training - Step 47012: {'lr': 4.7654496326834105e-06, 'samples': 24070656, 'steps': 47012, 'loss/train': 1.3880791664123535} 02/26/2022 14:28:50 - INFO - codeparrot_training - Step 47013: {'lr': 4.762270605111169e-06, 'samples': 24071168, 'steps': 47013, 'loss/train': 1.8887298107147217} 02/26/2022 14:28:55 - INFO - codeparrot_training - Step 47014: {'lr': 4.759092628059514e-06, 'samples': 24071680, 'steps': 47014, 'loss/train': 0.9282025098800659} 02/26/2022 14:28:58 - INFO - codeparrot_training - Step 47015: {'lr': 4.755915701542019e-06, 'samples': 24072192, 'steps': 47015, 'loss/train': 2.086386203765869} 02/26/2022 14:29:04 - INFO - codeparrot_training - Step 47016: {'lr': 4.752739825572339e-06, 'samples': 24072704, 'steps': 47016, 'loss/train': 1.3414241075515747} 02/26/2022 14:29:07 - INFO - codeparrot_training - Step 47017: {'lr': 4.749565000164019e-06, 'samples': 24073216, 'steps': 47017, 'loss/train': 1.908897042274475} 02/26/2022 14:29:13 - INFO - codeparrot_training - Step 47018: {'lr': 4.746391225330687e-06, 'samples': 24073728, 'steps': 47018, 'loss/train': 0.7843557596206665} 02/26/2022 14:29:17 - INFO - codeparrot_training - Step 47019: {'lr': 4.743218501085916e-06, 'samples': 24074240, 'steps': 47019, 'loss/train': 1.6694109439849854} 02/26/2022 14:29:22 - INFO - codeparrot_training - Step 47020: {'lr': 4.74004682744339e-06, 'samples': 24074752, 'steps': 47020, 'loss/train': 3.3496742248535156} 02/26/2022 14:29:26 - INFO - codeparrot_training - Step 47021: {'lr': 4.73687620441654e-06, 'samples': 24075264, 'steps': 47021, 'loss/train': 1.4146969318389893} 02/26/2022 14:29:33 - INFO - codeparrot_training - Step 47022: {'lr': 4.73370663201908e-06, 'samples': 24075776, 'steps': 47022, 'loss/train': 0.36437684297561646} 02/26/2022 14:29:36 - INFO - codeparrot_training - Step 47023: {'lr': 4.730538110264526e-06, 'samples': 24076288, 'steps': 47023, 'loss/train': 2.077320098876953} 02/26/2022 14:29:42 - INFO - codeparrot_training - Step 47024: {'lr': 4.727370639166506e-06, 'samples': 24076800, 'steps': 47024, 'loss/train': 1.8466856479644775} 02/26/2022 14:29:45 - INFO - codeparrot_training - Step 47025: {'lr': 4.724204218738482e-06, 'samples': 24077312, 'steps': 47025, 'loss/train': 1.3524516820907593} 02/26/2022 14:29:51 - INFO - codeparrot_training - Step 47026: {'lr': 4.721038848994136e-06, 'samples': 24077824, 'steps': 47026, 'loss/train': 0.6863139271736145} 02/26/2022 14:29:57 - INFO - codeparrot_training - Step 47027: {'lr': 4.717874529946931e-06, 'samples': 24078336, 'steps': 47027, 'loss/train': 1.8413203954696655} 02/26/2022 14:30:00 - INFO - codeparrot_training - Step 47028: {'lr': 4.714711261610494e-06, 'samples': 24078848, 'steps': 47028, 'loss/train': 0.9946373105049133} 02/26/2022 14:30:05 - INFO - codeparrot_training - Step 47029: {'lr': 4.7115490439983425e-06, 'samples': 24079360, 'steps': 47029, 'loss/train': 2.519493818283081} 02/26/2022 14:30:09 - INFO - codeparrot_training - Step 47030: {'lr': 4.7083878771240205e-06, 'samples': 24079872, 'steps': 47030, 'loss/train': 1.7761424779891968} 02/26/2022 14:30:14 - INFO - codeparrot_training - Step 47031: {'lr': 4.705227761001074e-06, 'samples': 24080384, 'steps': 47031, 'loss/train': 2.5566325187683105} 02/26/2022 14:30:18 - INFO - codeparrot_training - Step 47032: {'lr': 4.702068695643047e-06, 'samples': 24080896, 'steps': 47032, 'loss/train': 2.4438281059265137} 02/26/2022 14:30:24 - INFO - codeparrot_training - Step 47033: {'lr': 4.698910681063484e-06, 'samples': 24081408, 'steps': 47033, 'loss/train': 1.466216802597046} 02/26/2022 14:30:27 - INFO - codeparrot_training - Step 47034: {'lr': 4.695753717275875e-06, 'samples': 24081920, 'steps': 47034, 'loss/train': 2.448518991470337} 02/26/2022 14:30:33 - INFO - codeparrot_training - Step 47035: {'lr': 4.6925978042937636e-06, 'samples': 24082432, 'steps': 47035, 'loss/train': 2.0654125213623047} 02/26/2022 14:30:36 - INFO - codeparrot_training - Step 47036: {'lr': 4.689442942130667e-06, 'samples': 24082944, 'steps': 47036, 'loss/train': 0.5978533029556274} 02/26/2022 14:30:40 - INFO - codeparrot_training - Step 47037: {'lr': 4.686289130800131e-06, 'samples': 24083456, 'steps': 47037, 'loss/train': 0.4005730450153351} 02/26/2022 14:30:47 - INFO - codeparrot_training - Step 47038: {'lr': 4.683136370315616e-06, 'samples': 24083968, 'steps': 47038, 'loss/train': 2.3211472034454346} 02/26/2022 14:30:52 - INFO - codeparrot_training - Step 47039: {'lr': 4.679984660690639e-06, 'samples': 24084480, 'steps': 47039, 'loss/train': 1.9712399244308472} 02/26/2022 14:30:56 - INFO - codeparrot_training - Step 47040: {'lr': 4.676834001938718e-06, 'samples': 24084992, 'steps': 47040, 'loss/train': 1.4720009565353394} 02/26/2022 14:31:01 - INFO - codeparrot_training - Step 47041: {'lr': 4.673684394073341e-06, 'samples': 24085504, 'steps': 47041, 'loss/train': 2.356607675552368} 02/26/2022 14:31:05 - INFO - codeparrot_training - Step 47042: {'lr': 4.6705358371079974e-06, 'samples': 24086016, 'steps': 47042, 'loss/train': 1.11809241771698} 02/26/2022 14:31:10 - INFO - codeparrot_training - Step 47043: {'lr': 4.6673883310561775e-06, 'samples': 24086528, 'steps': 47043, 'loss/train': 2.2559280395507812} 02/26/2022 14:31:14 - INFO - codeparrot_training - Step 47044: {'lr': 4.664241875931368e-06, 'samples': 24087040, 'steps': 47044, 'loss/train': 0.714023232460022} 02/26/2022 14:31:19 - INFO - codeparrot_training - Step 47045: {'lr': 4.661096471747034e-06, 'samples': 24087552, 'steps': 47045, 'loss/train': 1.7764296531677246} 02/26/2022 14:31:23 - INFO - codeparrot_training - Step 47046: {'lr': 4.6579521185166895e-06, 'samples': 24088064, 'steps': 47046, 'loss/train': 0.6463498473167419} 02/26/2022 14:31:29 - INFO - codeparrot_training - Step 47047: {'lr': 4.654808816253742e-06, 'samples': 24088576, 'steps': 47047, 'loss/train': 1.051945447921753} 02/26/2022 14:31:32 - INFO - codeparrot_training - Step 47048: {'lr': 4.651666564971679e-06, 'samples': 24089088, 'steps': 47048, 'loss/train': 1.0134605169296265} 02/26/2022 14:31:35 - INFO - codeparrot_training - Step 47049: {'lr': 4.648525364683992e-06, 'samples': 24089600, 'steps': 47049, 'loss/train': 2.4695560932159424} 02/26/2022 14:31:42 - INFO - codeparrot_training - Step 47050: {'lr': 4.645385215404113e-06, 'samples': 24090112, 'steps': 47050, 'loss/train': 1.4363354444503784} 02/26/2022 14:31:46 - INFO - codeparrot_training - Step 47051: {'lr': 4.642246117145505e-06, 'samples': 24090624, 'steps': 47051, 'loss/train': 2.5005080699920654} 02/26/2022 14:31:51 - INFO - codeparrot_training - Step 47052: {'lr': 4.6391080699215716e-06, 'samples': 24091136, 'steps': 47052, 'loss/train': 2.089142322540283} 02/26/2022 14:31:55 - INFO - codeparrot_training - Step 47053: {'lr': 4.635971073745832e-06, 'samples': 24091648, 'steps': 47053, 'loss/train': 1.2328163385391235} 02/26/2022 14:32:00 - INFO - codeparrot_training - Step 47054: {'lr': 4.632835128631635e-06, 'samples': 24092160, 'steps': 47054, 'loss/train': 2.4630727767944336} 02/26/2022 14:32:04 - INFO - codeparrot_training - Step 47055: {'lr': 4.629700234592498e-06, 'samples': 24092672, 'steps': 47055, 'loss/train': 1.7775522470474243} 02/26/2022 14:32:09 - INFO - codeparrot_training - Step 47056: {'lr': 4.626566391641774e-06, 'samples': 24093184, 'steps': 47056, 'loss/train': 1.4107571840286255} 02/26/2022 14:32:13 - INFO - codeparrot_training - Step 47057: {'lr': 4.6234335997929474e-06, 'samples': 24093696, 'steps': 47057, 'loss/train': 0.07298669964075089} 02/26/2022 14:32:18 - INFO - codeparrot_training - Step 47058: {'lr': 4.6203018590594e-06, 'samples': 24094208, 'steps': 47058, 'loss/train': 0.7180238366127014} 02/26/2022 14:32:22 - INFO - codeparrot_training - Step 47059: {'lr': 4.617171169454621e-06, 'samples': 24094720, 'steps': 47059, 'loss/train': 2.3873889446258545} 02/26/2022 14:32:28 - INFO - codeparrot_training - Step 47060: {'lr': 4.6140415309919026e-06, 'samples': 24095232, 'steps': 47060, 'loss/train': 2.0663018226623535} 02/26/2022 14:32:31 - INFO - codeparrot_training - Step 47061: {'lr': 4.610912943684709e-06, 'samples': 24095744, 'steps': 47061, 'loss/train': 2.0417423248291016} 02/26/2022 14:32:37 - INFO - codeparrot_training - Step 47062: {'lr': 4.607785407546444e-06, 'samples': 24096256, 'steps': 47062, 'loss/train': 0.7667291164398193} 02/26/2022 14:32:40 - INFO - codeparrot_training - Step 47063: {'lr': 4.604658922590543e-06, 'samples': 24096768, 'steps': 47063, 'loss/train': 2.317803144454956} 02/26/2022 14:32:46 - INFO - codeparrot_training - Step 47064: {'lr': 4.601533488830356e-06, 'samples': 24097280, 'steps': 47064, 'loss/train': 0.20103920996189117} 02/26/2022 14:32:49 - INFO - codeparrot_training - Step 47065: {'lr': 4.598409106279261e-06, 'samples': 24097792, 'steps': 47065, 'loss/train': 1.5555360317230225} 02/26/2022 14:32:55 - INFO - codeparrot_training - Step 47066: {'lr': 4.595285774950636e-06, 'samples': 24098304, 'steps': 47066, 'loss/train': 2.5949933528900146} 02/26/2022 14:32:58 - INFO - codeparrot_training - Step 47067: {'lr': 4.5921634948579145e-06, 'samples': 24098816, 'steps': 47067, 'loss/train': 1.6819952726364136} 02/26/2022 14:33:04 - INFO - codeparrot_training - Step 47068: {'lr': 4.589042266014448e-06, 'samples': 24099328, 'steps': 47068, 'loss/train': 1.6068577766418457} 02/26/2022 14:33:07 - INFO - codeparrot_training - Step 47069: {'lr': 4.585922088433558e-06, 'samples': 24099840, 'steps': 47069, 'loss/train': 1.8582770824432373} 02/26/2022 14:33:14 - INFO - codeparrot_training - Step 47070: {'lr': 4.5828029621286795e-06, 'samples': 24100352, 'steps': 47070, 'loss/train': 1.2109534740447998} 02/26/2022 14:33:17 - INFO - codeparrot_training - Step 47071: {'lr': 4.579684887113133e-06, 'samples': 24100864, 'steps': 47071, 'loss/train': 1.5589314699172974} 02/26/2022 14:33:23 - INFO - codeparrot_training - Step 47072: {'lr': 4.5765678634003e-06, 'samples': 24101376, 'steps': 47072, 'loss/train': 2.1774210929870605} 02/26/2022 14:33:26 - INFO - codeparrot_training - Step 47073: {'lr': 4.5734518910034995e-06, 'samples': 24101888, 'steps': 47073, 'loss/train': 2.6302433013916016} 02/26/2022 14:33:32 - INFO - codeparrot_training - Step 47074: {'lr': 4.570336969936112e-06, 'samples': 24102400, 'steps': 47074, 'loss/train': 2.142443895339966} 02/26/2022 14:33:35 - INFO - codeparrot_training - Step 47075: {'lr': 4.56722310021146e-06, 'samples': 24102912, 'steps': 47075, 'loss/train': 2.084932804107666} 02/26/2022 14:33:41 - INFO - codeparrot_training - Step 47076: {'lr': 4.564110281842892e-06, 'samples': 24103424, 'steps': 47076, 'loss/train': 1.6592644453048706} 02/26/2022 14:33:44 - INFO - codeparrot_training - Step 47077: {'lr': 4.560998514843761e-06, 'samples': 24103936, 'steps': 47077, 'loss/train': 0.32203760743141174} 02/26/2022 14:33:50 - INFO - codeparrot_training - Step 47078: {'lr': 4.557887799227334e-06, 'samples': 24104448, 'steps': 47078, 'loss/train': 0.6809161305427551} 02/26/2022 14:33:53 - INFO - codeparrot_training - Step 47079: {'lr': 4.554778135007015e-06, 'samples': 24104960, 'steps': 47079, 'loss/train': 1.2344796657562256} 02/26/2022 14:34:00 - INFO - codeparrot_training - Step 47080: {'lr': 4.551669522196072e-06, 'samples': 24105472, 'steps': 47080, 'loss/train': 1.651402473449707} 02/26/2022 14:34:04 - INFO - codeparrot_training - Step 47081: {'lr': 4.548561960807857e-06, 'samples': 24105984, 'steps': 47081, 'loss/train': 2.0178136825561523} 02/26/2022 14:34:09 - INFO - codeparrot_training - Step 47082: {'lr': 4.5454554508556345e-06, 'samples': 24106496, 'steps': 47082, 'loss/train': 1.548479437828064} 02/26/2022 14:34:12 - INFO - codeparrot_training - Step 47083: {'lr': 4.542349992352785e-06, 'samples': 24107008, 'steps': 47083, 'loss/train': 1.00944983959198} 02/26/2022 14:34:18 - INFO - codeparrot_training - Step 47084: {'lr': 4.539245585312546e-06, 'samples': 24107520, 'steps': 47084, 'loss/train': 1.1888713836669922} 02/26/2022 14:34:21 - INFO - codeparrot_training - Step 47085: {'lr': 4.536142229748213e-06, 'samples': 24108032, 'steps': 47085, 'loss/train': 1.8347615003585815} 02/26/2022 14:34:27 - INFO - codeparrot_training - Step 47086: {'lr': 4.53303992567311e-06, 'samples': 24108544, 'steps': 47086, 'loss/train': 2.052255392074585} 02/26/2022 14:34:30 - INFO - codeparrot_training - Step 47087: {'lr': 4.529938673100531e-06, 'samples': 24109056, 'steps': 47087, 'loss/train': 2.5584685802459717} 02/26/2022 14:34:36 - INFO - codeparrot_training - Step 47088: {'lr': 4.526838472043771e-06, 'samples': 24109568, 'steps': 47088, 'loss/train': 4.138465404510498} 02/26/2022 14:34:39 - INFO - codeparrot_training - Step 47089: {'lr': 4.523739322516041e-06, 'samples': 24110080, 'steps': 47089, 'loss/train': 1.7930083274841309} 02/26/2022 14:34:45 - INFO - codeparrot_training - Step 47090: {'lr': 4.5206412245306916e-06, 'samples': 24110592, 'steps': 47090, 'loss/train': 2.3545634746551514} 02/26/2022 14:34:48 - INFO - codeparrot_training - Step 47091: {'lr': 4.517544178100963e-06, 'samples': 24111104, 'steps': 47091, 'loss/train': 1.5640872716903687} 02/26/2022 14:34:54 - INFO - codeparrot_training - Step 47092: {'lr': 4.5144481832401215e-06, 'samples': 24111616, 'steps': 47092, 'loss/train': 1.0407648086547852} 02/26/2022 14:34:57 - INFO - codeparrot_training - Step 47093: {'lr': 4.511353239961436e-06, 'samples': 24112128, 'steps': 47093, 'loss/train': 1.559370517730713} 02/26/2022 14:35:03 - INFO - codeparrot_training - Step 47094: {'lr': 4.508259348278171e-06, 'samples': 24112640, 'steps': 47094, 'loss/train': 1.5072500705718994} 02/26/2022 14:35:06 - INFO - codeparrot_training - Step 47095: {'lr': 4.505166508203567e-06, 'samples': 24113152, 'steps': 47095, 'loss/train': 2.1189537048339844} 02/26/2022 14:35:13 - INFO - codeparrot_training - Step 47096: {'lr': 4.502074719750865e-06, 'samples': 24113664, 'steps': 47096, 'loss/train': 1.1228779554367065} 02/26/2022 14:35:16 - INFO - codeparrot_training - Step 47097: {'lr': 4.4989839829333305e-06, 'samples': 24114176, 'steps': 47097, 'loss/train': 2.733769178390503} 02/26/2022 14:35:22 - INFO - codeparrot_training - Step 47098: {'lr': 4.495894297764175e-06, 'samples': 24114688, 'steps': 47098, 'loss/train': 1.8858383893966675} 02/26/2022 14:35:25 - INFO - codeparrot_training - Step 47099: {'lr': 4.492805664256666e-06, 'samples': 24115200, 'steps': 47099, 'loss/train': 1.477362871170044} 02/26/2022 14:35:31 - INFO - codeparrot_training - Step 47100: {'lr': 4.489718082424044e-06, 'samples': 24115712, 'steps': 47100, 'loss/train': 1.7046114206314087} 02/26/2022 14:35:34 - INFO - codeparrot_training - Step 47101: {'lr': 4.486631552279463e-06, 'samples': 24116224, 'steps': 47101, 'loss/train': 1.1065185070037842} 02/26/2022 14:35:40 - INFO - codeparrot_training - Step 47102: {'lr': 4.48354607383622e-06, 'samples': 24116736, 'steps': 47102, 'loss/train': 2.017695426940918} 02/26/2022 14:35:43 - INFO - codeparrot_training - Step 47103: {'lr': 4.480461647107526e-06, 'samples': 24117248, 'steps': 47103, 'loss/train': 2.986078977584839} 02/26/2022 14:35:49 - INFO - codeparrot_training - Step 47104: {'lr': 4.477378272106564e-06, 'samples': 24117760, 'steps': 47104, 'loss/train': 1.5272072553634644} 02/26/2022 14:35:52 - INFO - codeparrot_training - Step 47105: {'lr': 4.474295948846519e-06, 'samples': 24118272, 'steps': 47105, 'loss/train': 1.681247591972351} 02/26/2022 14:35:58 - INFO - codeparrot_training - Step 47106: {'lr': 4.4712146773406295e-06, 'samples': 24118784, 'steps': 47106, 'loss/train': 1.3817273378372192} 02/26/2022 14:36:04 - INFO - codeparrot_training - Step 47107: {'lr': 4.468134457602136e-06, 'samples': 24119296, 'steps': 47107, 'loss/train': 1.3703384399414062} 02/26/2022 14:36:07 - INFO - codeparrot_training - Step 47108: {'lr': 4.465055289644166e-06, 'samples': 24119808, 'steps': 47108, 'loss/train': 1.4463871717453003} 02/26/2022 14:36:11 - INFO - codeparrot_training - Step 47109: {'lr': 4.461977173479903e-06, 'samples': 24120320, 'steps': 47109, 'loss/train': 1.2517337799072266} 02/26/2022 14:36:17 - INFO - codeparrot_training - Step 47110: {'lr': 4.458900109122588e-06, 'samples': 24120832, 'steps': 47110, 'loss/train': 0.4688619077205658} 02/26/2022 14:36:20 - INFO - codeparrot_training - Step 47111: {'lr': 4.455824096585376e-06, 'samples': 24121344, 'steps': 47111, 'loss/train': 1.6020241975784302} 02/26/2022 14:36:26 - INFO - codeparrot_training - Step 47112: {'lr': 4.452749135881479e-06, 'samples': 24121856, 'steps': 47112, 'loss/train': 2.4513211250305176} 02/26/2022 14:36:29 - INFO - codeparrot_training - Step 47113: {'lr': 4.449675227023969e-06, 'samples': 24122368, 'steps': 47113, 'loss/train': 1.5038057565689087} 02/26/2022 14:36:35 - INFO - codeparrot_training - Step 47114: {'lr': 4.446602370026115e-06, 'samples': 24122880, 'steps': 47114, 'loss/train': 1.8735201358795166} 02/26/2022 14:36:38 - INFO - codeparrot_training - Step 47115: {'lr': 4.443530564901043e-06, 'samples': 24123392, 'steps': 47115, 'loss/train': 2.096792221069336} 02/26/2022 14:36:45 - INFO - codeparrot_training - Step 47116: {'lr': 4.44045981166194e-06, 'samples': 24123904, 'steps': 47116, 'loss/train': 1.113234043121338} 02/26/2022 14:36:48 - INFO - codeparrot_training - Step 47117: {'lr': 4.437390110321876e-06, 'samples': 24124416, 'steps': 47117, 'loss/train': 2.3915045261383057} 02/26/2022 14:36:54 - INFO - codeparrot_training - Step 47118: {'lr': 4.434321460894092e-06, 'samples': 24124928, 'steps': 47118, 'loss/train': 1.185437560081482} 02/26/2022 14:36:57 - INFO - codeparrot_training - Step 47119: {'lr': 4.431253863391688e-06, 'samples': 24125440, 'steps': 47119, 'loss/train': 1.9552744626998901} 02/26/2022 14:37:03 - INFO - codeparrot_training - Step 47120: {'lr': 4.4281873178278475e-06, 'samples': 24125952, 'steps': 47120, 'loss/train': 0.7424376010894775} 02/26/2022 14:37:06 - INFO - codeparrot_training - Step 47121: {'lr': 4.425121824215644e-06, 'samples': 24126464, 'steps': 47121, 'loss/train': 1.4990954399108887} 02/26/2022 14:37:12 - INFO - codeparrot_training - Step 47122: {'lr': 4.422057382568234e-06, 'samples': 24126976, 'steps': 47122, 'loss/train': 1.594918131828308} 02/26/2022 14:37:15 - INFO - codeparrot_training - Step 47123: {'lr': 4.418993992898746e-06, 'samples': 24127488, 'steps': 47123, 'loss/train': 2.3720953464508057} 02/26/2022 14:37:21 - INFO - codeparrot_training - Step 47124: {'lr': 4.415931655220334e-06, 'samples': 24128000, 'steps': 47124, 'loss/train': 2.011507749557495} 02/26/2022 14:37:24 - INFO - codeparrot_training - Step 47125: {'lr': 4.412870369546074e-06, 'samples': 24128512, 'steps': 47125, 'loss/train': 1.8809527158737183} 02/26/2022 14:37:31 - INFO - codeparrot_training - Step 47126: {'lr': 4.40981013588912e-06, 'samples': 24129024, 'steps': 47126, 'loss/train': 0.7757999897003174} 02/26/2022 14:37:34 - INFO - codeparrot_training - Step 47127: {'lr': 4.406750954262517e-06, 'samples': 24129536, 'steps': 47127, 'loss/train': 2.9754371643066406} 02/26/2022 14:37:39 - INFO - codeparrot_training - Step 47128: {'lr': 4.403692824679423e-06, 'samples': 24130048, 'steps': 47128, 'loss/train': 0.17659859359264374} 02/26/2022 14:37:43 - INFO - codeparrot_training - Step 47129: {'lr': 4.400635747152937e-06, 'samples': 24130560, 'steps': 47129, 'loss/train': 1.472274899482727} 02/26/2022 14:37:48 - INFO - codeparrot_training - Step 47130: {'lr': 4.397579721696132e-06, 'samples': 24131072, 'steps': 47130, 'loss/train': 1.6382604837417603} 02/26/2022 14:37:52 - INFO - codeparrot_training - Step 47131: {'lr': 4.394524748322109e-06, 'samples': 24131584, 'steps': 47131, 'loss/train': 0.8831396102905273} 02/26/2022 14:37:57 - INFO - codeparrot_training - Step 47132: {'lr': 4.391470827043942e-06, 'samples': 24132096, 'steps': 47132, 'loss/train': 1.23939049243927} 02/26/2022 14:38:01 - INFO - codeparrot_training - Step 47133: {'lr': 4.388417957874729e-06, 'samples': 24132608, 'steps': 47133, 'loss/train': 2.3350114822387695} 02/26/2022 14:38:06 - INFO - codeparrot_training - Step 47134: {'lr': 4.385366140827546e-06, 'samples': 24133120, 'steps': 47134, 'loss/train': 1.7037105560302734} 02/26/2022 14:38:10 - INFO - codeparrot_training - Step 47135: {'lr': 4.382315375915463e-06, 'samples': 24133632, 'steps': 47135, 'loss/train': 1.7490791082382202} 02/26/2022 14:38:15 - INFO - codeparrot_training - Step 47136: {'lr': 4.379265663151527e-06, 'samples': 24134144, 'steps': 47136, 'loss/train': 0.4786655902862549} 02/26/2022 14:38:19 - INFO - codeparrot_training - Step 47137: {'lr': 4.376217002548866e-06, 'samples': 24134656, 'steps': 47137, 'loss/train': 2.0586559772491455} 02/26/2022 14:38:24 - INFO - codeparrot_training - Step 47138: {'lr': 4.373169394120469e-06, 'samples': 24135168, 'steps': 47138, 'loss/train': 1.3500629663467407} 02/26/2022 14:38:28 - INFO - codeparrot_training - Step 47139: {'lr': 4.370122837879437e-06, 'samples': 24135680, 'steps': 47139, 'loss/train': 0.6475369334220886} 02/26/2022 14:38:33 - INFO - codeparrot_training - Step 47140: {'lr': 4.367077333838787e-06, 'samples': 24136192, 'steps': 47140, 'loss/train': 2.9690399169921875} 02/26/2022 14:38:37 - INFO - codeparrot_training - Step 47141: {'lr': 4.364032882011593e-06, 'samples': 24136704, 'steps': 47141, 'loss/train': 1.6853876113891602} 02/26/2022 14:38:43 - INFO - codeparrot_training - Step 47142: {'lr': 4.360989482410871e-06, 'samples': 24137216, 'steps': 47142, 'loss/train': 1.6282105445861816} 02/26/2022 14:38:46 - INFO - codeparrot_training - Step 47143: {'lr': 4.357947135049667e-06, 'samples': 24137728, 'steps': 47143, 'loss/train': 2.1060256958007812} 02/26/2022 14:38:52 - INFO - codeparrot_training - Step 47144: {'lr': 4.354905839941026e-06, 'samples': 24138240, 'steps': 47144, 'loss/train': 1.1118172407150269} 02/26/2022 14:38:55 - INFO - codeparrot_training - Step 47145: {'lr': 4.351865597097965e-06, 'samples': 24138752, 'steps': 47145, 'loss/train': 1.6369283199310303} 02/26/2022 14:39:01 - INFO - codeparrot_training - Step 47146: {'lr': 4.348826406533501e-06, 'samples': 24139264, 'steps': 47146, 'loss/train': 1.6946613788604736} 02/26/2022 14:39:04 - INFO - codeparrot_training - Step 47147: {'lr': 4.3457882682607085e-06, 'samples': 24139776, 'steps': 47147, 'loss/train': 1.0469201803207397} 02/26/2022 14:39:10 - INFO - codeparrot_training - Step 47148: {'lr': 4.342751182292493e-06, 'samples': 24140288, 'steps': 47148, 'loss/train': 1.8560876846313477} 02/26/2022 14:39:13 - INFO - codeparrot_training - Step 47149: {'lr': 4.339715148641954e-06, 'samples': 24140800, 'steps': 47149, 'loss/train': 3.023561477661133} 02/26/2022 14:39:19 - INFO - codeparrot_training - Step 47150: {'lr': 4.336680167322055e-06, 'samples': 24141312, 'steps': 47150, 'loss/train': 1.7399812936782837} 02/26/2022 14:39:22 - INFO - codeparrot_training - Step 47151: {'lr': 4.333646238345868e-06, 'samples': 24141824, 'steps': 47151, 'loss/train': 1.856559157371521} 02/26/2022 14:39:29 - INFO - codeparrot_training - Step 47152: {'lr': 4.330613361726271e-06, 'samples': 24142336, 'steps': 47152, 'loss/train': 1.5507713556289673} 02/26/2022 14:39:32 - INFO - codeparrot_training - Step 47153: {'lr': 4.327581537476338e-06, 'samples': 24142848, 'steps': 47153, 'loss/train': 1.2864854335784912} 02/26/2022 14:39:38 - INFO - codeparrot_training - Step 47154: {'lr': 4.324550765609003e-06, 'samples': 24143360, 'steps': 47154, 'loss/train': 2.471898078918457} 02/26/2022 14:39:41 - INFO - codeparrot_training - Step 47155: {'lr': 4.321521046137339e-06, 'samples': 24143872, 'steps': 47155, 'loss/train': 2.5953712463378906} 02/26/2022 14:39:47 - INFO - codeparrot_training - Step 47156: {'lr': 4.318492379074224e-06, 'samples': 24144384, 'steps': 47156, 'loss/train': 1.9046624898910522} 02/26/2022 14:39:50 - INFO - codeparrot_training - Step 47157: {'lr': 4.315464764432703e-06, 'samples': 24144896, 'steps': 47157, 'loss/train': 1.9269174337387085} 02/26/2022 14:39:56 - INFO - codeparrot_training - Step 47158: {'lr': 4.3124382022256825e-06, 'samples': 24145408, 'steps': 47158, 'loss/train': 1.6814074516296387} 02/26/2022 14:39:59 - INFO - codeparrot_training - Step 47159: {'lr': 4.309412692466208e-06, 'samples': 24145920, 'steps': 47159, 'loss/train': 1.686313271522522} 02/26/2022 14:40:05 - INFO - codeparrot_training - Step 47160: {'lr': 4.306388235167158e-06, 'samples': 24146432, 'steps': 47160, 'loss/train': 1.508376955986023} 02/26/2022 14:40:09 - INFO - codeparrot_training - Step 47161: {'lr': 4.30336483034155e-06, 'samples': 24146944, 'steps': 47161, 'loss/train': 1.4307645559310913} 02/26/2022 14:40:12 - INFO - codeparrot_training - Step 47162: {'lr': 4.300342478002289e-06, 'samples': 24147456, 'steps': 47162, 'loss/train': 1.6573303937911987} 02/26/2022 14:40:18 - INFO - codeparrot_training - Step 47163: {'lr': 4.297321178162339e-06, 'samples': 24147968, 'steps': 47163, 'loss/train': 2.3991005420684814} 02/26/2022 14:40:22 - INFO - codeparrot_training - Step 47164: {'lr': 4.294300930834688e-06, 'samples': 24148480, 'steps': 47164, 'loss/train': 1.4439725875854492} 02/26/2022 14:40:27 - INFO - codeparrot_training - Step 47165: {'lr': 4.291281736032188e-06, 'samples': 24148992, 'steps': 47165, 'loss/train': 2.271268367767334} 02/26/2022 14:40:31 - INFO - codeparrot_training - Step 47166: {'lr': 4.288263593767827e-06, 'samples': 24149504, 'steps': 47166, 'loss/train': 2.306211233139038} 02/26/2022 14:40:36 - INFO - codeparrot_training - Step 47167: {'lr': 4.285246504054513e-06, 'samples': 24150016, 'steps': 47167, 'loss/train': 1.6509721279144287} 02/26/2022 14:40:40 - INFO - codeparrot_training - Step 47168: {'lr': 4.282230466905207e-06, 'samples': 24150528, 'steps': 47168, 'loss/train': 1.5485646724700928} 02/26/2022 14:40:45 - INFO - codeparrot_training - Step 47169: {'lr': 4.279215482332788e-06, 'samples': 24151040, 'steps': 47169, 'loss/train': 1.28037428855896} 02/26/2022 14:40:49 - INFO - codeparrot_training - Step 47170: {'lr': 4.2762015503501614e-06, 'samples': 24151552, 'steps': 47170, 'loss/train': 1.7375378608703613} 02/26/2022 14:40:54 - INFO - codeparrot_training - Step 47171: {'lr': 4.2731886709702904e-06, 'samples': 24152064, 'steps': 47171, 'loss/train': 2.593419313430786} 02/26/2022 14:40:58 - INFO - codeparrot_training - Step 47172: {'lr': 4.270176844206053e-06, 'samples': 24152576, 'steps': 47172, 'loss/train': 0.46203529834747314} 02/26/2022 14:41:04 - INFO - codeparrot_training - Step 47173: {'lr': 4.267166070070355e-06, 'samples': 24153088, 'steps': 47173, 'loss/train': 1.961051344871521} 02/26/2022 14:41:07 - INFO - codeparrot_training - Step 47174: {'lr': 4.264156348576048e-06, 'samples': 24153600, 'steps': 47174, 'loss/train': 1.4205868244171143} 02/26/2022 14:41:13 - INFO - codeparrot_training - Step 47175: {'lr': 4.261147679736094e-06, 'samples': 24154112, 'steps': 47175, 'loss/train': 2.190095901489258} 02/26/2022 14:41:16 - INFO - codeparrot_training - Step 47176: {'lr': 4.258140063563343e-06, 'samples': 24154624, 'steps': 47176, 'loss/train': 1.9713971614837646} 02/26/2022 14:41:22 - INFO - codeparrot_training - Step 47177: {'lr': 4.255133500070701e-06, 'samples': 24155136, 'steps': 47177, 'loss/train': 3.0334203243255615} 02/26/2022 14:41:25 - INFO - codeparrot_training - Step 47178: {'lr': 4.252127989271048e-06, 'samples': 24155648, 'steps': 47178, 'loss/train': 0.36063963174819946} 02/26/2022 14:41:31 - INFO - codeparrot_training - Step 47179: {'lr': 4.249123531177207e-06, 'samples': 24156160, 'steps': 47179, 'loss/train': 1.4529038667678833} 02/26/2022 14:41:35 - INFO - codeparrot_training - Step 47180: {'lr': 4.246120125802111e-06, 'samples': 24156672, 'steps': 47180, 'loss/train': 0.04210694879293442} 02/26/2022 14:41:40 - INFO - codeparrot_training - Step 47181: {'lr': 4.243117773158611e-06, 'samples': 24157184, 'steps': 47181, 'loss/train': 0.3954634368419647} 02/26/2022 14:41:44 - INFO - codeparrot_training - Step 47182: {'lr': 4.240116473259531e-06, 'samples': 24157696, 'steps': 47182, 'loss/train': 1.2419888973236084} 02/26/2022 14:41:49 - INFO - codeparrot_training - Step 47183: {'lr': 4.2371162261177484e-06, 'samples': 24158208, 'steps': 47183, 'loss/train': 1.7005199193954468} 02/26/2022 14:41:53 - INFO - codeparrot_training - Step 47184: {'lr': 4.234117031746143e-06, 'samples': 24158720, 'steps': 47184, 'loss/train': 0.935300350189209} 02/26/2022 14:41:58 - INFO - codeparrot_training - Step 47185: {'lr': 4.231118890157509e-06, 'samples': 24159232, 'steps': 47185, 'loss/train': 3.2207820415496826} 02/26/2022 14:42:02 - INFO - codeparrot_training - Step 47186: {'lr': 4.2281218013647815e-06, 'samples': 24159744, 'steps': 47186, 'loss/train': 1.2835636138916016} 02/26/2022 14:42:07 - INFO - codeparrot_training - Step 47187: {'lr': 4.2251257653806996e-06, 'samples': 24160256, 'steps': 47187, 'loss/train': 1.5822453498840332} 02/26/2022 14:42:11 - INFO - codeparrot_training - Step 47188: {'lr': 4.222130782218142e-06, 'samples': 24160768, 'steps': 47188, 'loss/train': 3.468564510345459} 02/26/2022 14:42:17 - INFO - codeparrot_training - Step 47189: {'lr': 4.219136851889905e-06, 'samples': 24161280, 'steps': 47189, 'loss/train': 2.1898415088653564} 02/26/2022 14:42:20 - INFO - codeparrot_training - Step 47190: {'lr': 4.216143974408892e-06, 'samples': 24161792, 'steps': 47190, 'loss/train': 0.965121865272522} 02/26/2022 14:42:26 - INFO - codeparrot_training - Step 47191: {'lr': 4.213152149787847e-06, 'samples': 24162304, 'steps': 47191, 'loss/train': 2.138892650604248} 02/26/2022 14:42:29 - INFO - codeparrot_training - Step 47192: {'lr': 4.210161378039618e-06, 'samples': 24162816, 'steps': 47192, 'loss/train': 1.96351957321167} 02/26/2022 14:42:35 - INFO - codeparrot_training - Step 47193: {'lr': 4.207171659177e-06, 'samples': 24163328, 'steps': 47193, 'loss/train': 2.7734086513519287} 02/26/2022 14:42:38 - INFO - codeparrot_training - Step 47194: {'lr': 4.204182993212819e-06, 'samples': 24163840, 'steps': 47194, 'loss/train': 1.9226479530334473} 02/26/2022 14:42:44 - INFO - codeparrot_training - Step 47195: {'lr': 4.201195380159867e-06, 'samples': 24164352, 'steps': 47195, 'loss/train': 1.737646460533142} 02/26/2022 14:42:47 - INFO - codeparrot_training - Step 47196: {'lr': 4.1982088200309685e-06, 'samples': 24164864, 'steps': 47196, 'loss/train': 1.0406370162963867} 02/26/2022 14:42:53 - INFO - codeparrot_training - Step 47197: {'lr': 4.195223312838836e-06, 'samples': 24165376, 'steps': 47197, 'loss/train': 1.7770664691925049} 02/26/2022 14:42:59 - INFO - codeparrot_training - Step 47198: {'lr': 4.192238858596375e-06, 'samples': 24165888, 'steps': 47198, 'loss/train': 1.632149338722229} 02/26/2022 14:43:03 - INFO - codeparrot_training - Step 47199: {'lr': 4.1892554573162715e-06, 'samples': 24166400, 'steps': 47199, 'loss/train': 1.4625093936920166} 02/26/2022 14:43:08 - INFO - codeparrot_training - Step 47200: {'lr': 4.186273109011374e-06, 'samples': 24166912, 'steps': 47200, 'loss/train': 2.194828987121582} 02/26/2022 14:43:11 - INFO - codeparrot_training - Step 47201: {'lr': 4.183291813694395e-06, 'samples': 24167424, 'steps': 47201, 'loss/train': 1.5172010660171509} 02/26/2022 14:43:17 - INFO - codeparrot_training - Step 47202: {'lr': 4.180311571378159e-06, 'samples': 24167936, 'steps': 47202, 'loss/train': 1.6183710098266602} 02/26/2022 14:43:21 - INFO - codeparrot_training - Step 47203: {'lr': 4.177332382075405e-06, 'samples': 24168448, 'steps': 47203, 'loss/train': 1.1115070581436157} 02/26/2022 14:43:24 - INFO - codeparrot_training - Step 47204: {'lr': 4.1743542457989005e-06, 'samples': 24168960, 'steps': 47204, 'loss/train': 1.5006262063980103} 02/26/2022 14:43:30 - INFO - codeparrot_training - Step 47205: {'lr': 4.171377162561385e-06, 'samples': 24169472, 'steps': 47205, 'loss/train': 2.4143946170806885} 02/26/2022 14:43:33 - INFO - codeparrot_training - Step 47206: {'lr': 4.168401132375654e-06, 'samples': 24169984, 'steps': 47206, 'loss/train': 0.43767157196998596} 02/26/2022 14:43:39 - INFO - codeparrot_training - Step 47207: {'lr': 4.16542615525442e-06, 'samples': 24170496, 'steps': 47207, 'loss/train': 1.577985167503357} 02/26/2022 14:43:42 - INFO - codeparrot_training - Step 47208: {'lr': 4.162452231210479e-06, 'samples': 24171008, 'steps': 47208, 'loss/train': 1.830161690711975} 02/26/2022 14:43:49 - INFO - codeparrot_training - Step 47209: {'lr': 4.159479360256485e-06, 'samples': 24171520, 'steps': 47209, 'loss/train': 0.8036623597145081} 02/26/2022 14:43:53 - INFO - codeparrot_training - Step 47210: {'lr': 4.156507542405235e-06, 'samples': 24172032, 'steps': 47210, 'loss/train': 0.3942053020000458} 02/26/2022 14:43:58 - INFO - codeparrot_training - Step 47211: {'lr': 4.153536777669442e-06, 'samples': 24172544, 'steps': 47211, 'loss/train': 1.0266683101654053} 02/26/2022 14:44:02 - INFO - codeparrot_training - Step 47212: {'lr': 4.150567066061845e-06, 'samples': 24173056, 'steps': 47212, 'loss/train': 0.4471285045146942} 02/26/2022 14:44:07 - INFO - codeparrot_training - Step 47213: {'lr': 4.147598407595127e-06, 'samples': 24173568, 'steps': 47213, 'loss/train': 2.4111576080322266} 02/26/2022 14:44:11 - INFO - codeparrot_training - Step 47214: {'lr': 4.144630802282057e-06, 'samples': 24174080, 'steps': 47214, 'loss/train': 1.5770294666290283} 02/26/2022 14:44:16 - INFO - codeparrot_training - Step 47215: {'lr': 4.141664250135291e-06, 'samples': 24174592, 'steps': 47215, 'loss/train': 2.1797947883605957} 02/26/2022 14:44:20 - INFO - codeparrot_training - Step 47216: {'lr': 4.138698751167597e-06, 'samples': 24175104, 'steps': 47216, 'loss/train': 1.363838791847229} 02/26/2022 14:44:25 - INFO - codeparrot_training - Step 47217: {'lr': 4.135734305391603e-06, 'samples': 24175616, 'steps': 47217, 'loss/train': 1.8961713314056396} 02/26/2022 14:44:29 - INFO - codeparrot_training - Step 47218: {'lr': 4.132770912820105e-06, 'samples': 24176128, 'steps': 47218, 'loss/train': 1.6389878988265991} 02/26/2022 14:44:34 - INFO - codeparrot_training - Step 47219: {'lr': 4.129808573465732e-06, 'samples': 24176640, 'steps': 47219, 'loss/train': 1.65084969997406} 02/26/2022 14:44:38 - INFO - codeparrot_training - Step 47220: {'lr': 4.12684728734114e-06, 'samples': 24177152, 'steps': 47220, 'loss/train': 1.2980422973632812} 02/26/2022 14:44:44 - INFO - codeparrot_training - Step 47221: {'lr': 4.123887054459124e-06, 'samples': 24177664, 'steps': 47221, 'loss/train': 1.327880859375} 02/26/2022 14:44:47 - INFO - codeparrot_training - Step 47222: {'lr': 4.120927874832259e-06, 'samples': 24178176, 'steps': 47222, 'loss/train': 0.8891405463218689} 02/26/2022 14:44:53 - INFO - codeparrot_training - Step 47223: {'lr': 4.117969748473282e-06, 'samples': 24178688, 'steps': 47223, 'loss/train': 1.5660713911056519} 02/26/2022 14:44:58 - INFO - codeparrot_training - Step 47224: {'lr': 4.115012675394825e-06, 'samples': 24179200, 'steps': 47224, 'loss/train': 1.7935417890548706} 02/26/2022 14:45:02 - INFO - codeparrot_training - Step 47225: {'lr': 4.112056655609625e-06, 'samples': 24179712, 'steps': 47225, 'loss/train': 2.003937244415283} 02/26/2022 14:45:07 - INFO - codeparrot_training - Step 47226: {'lr': 4.1091016891302566e-06, 'samples': 24180224, 'steps': 47226, 'loss/train': 0.07567253708839417} 02/26/2022 14:45:11 - INFO - codeparrot_training - Step 47227: {'lr': 4.106147775969432e-06, 'samples': 24180736, 'steps': 47227, 'loss/train': 1.8219544887542725} 02/26/2022 14:45:17 - INFO - codeparrot_training - Step 47228: {'lr': 4.10319491613978e-06, 'samples': 24181248, 'steps': 47228, 'loss/train': 0.9790166616439819} 02/26/2022 14:45:20 - INFO - codeparrot_training - Step 47229: {'lr': 4.100243109653984e-06, 'samples': 24181760, 'steps': 47229, 'loss/train': 1.9543795585632324} 02/26/2022 14:45:26 - INFO - codeparrot_training - Step 47230: {'lr': 4.097292356524618e-06, 'samples': 24182272, 'steps': 47230, 'loss/train': 1.9397372007369995} 02/26/2022 14:45:29 - INFO - codeparrot_training - Step 47231: {'lr': 4.0943426567644215e-06, 'samples': 24182784, 'steps': 47231, 'loss/train': 2.0798254013061523} 02/26/2022 14:45:35 - INFO - codeparrot_training - Step 47232: {'lr': 4.091394010385913e-06, 'samples': 24183296, 'steps': 47232, 'loss/train': 2.189432144165039} 02/26/2022 14:45:38 - INFO - codeparrot_training - Step 47233: {'lr': 4.088446417401831e-06, 'samples': 24183808, 'steps': 47233, 'loss/train': 2.1390795707702637} 02/26/2022 14:45:44 - INFO - codeparrot_training - Step 47234: {'lr': 4.0854998778247776e-06, 'samples': 24184320, 'steps': 47234, 'loss/train': 2.073695421218872} 02/26/2022 14:45:48 - INFO - codeparrot_training - Step 47235: {'lr': 4.0825543916673255e-06, 'samples': 24184832, 'steps': 47235, 'loss/train': 1.6629016399383545} 02/26/2022 14:45:53 - INFO - codeparrot_training - Step 47236: {'lr': 4.079609958942132e-06, 'samples': 24185344, 'steps': 47236, 'loss/train': 2.354749917984009} 02/26/2022 14:45:57 - INFO - codeparrot_training - Step 47237: {'lr': 4.076666579661797e-06, 'samples': 24185856, 'steps': 47237, 'loss/train': 1.5385057926177979} 02/26/2022 14:46:02 - INFO - codeparrot_training - Step 47238: {'lr': 4.073724253838923e-06, 'samples': 24186368, 'steps': 47238, 'loss/train': 0.4567966163158417} 02/26/2022 14:46:06 - INFO - codeparrot_training - Step 47239: {'lr': 4.070782981486166e-06, 'samples': 24186880, 'steps': 47239, 'loss/train': 2.5535647869110107} 02/26/2022 14:46:11 - INFO - codeparrot_training - Step 47240: {'lr': 4.067842762616014e-06, 'samples': 24187392, 'steps': 47240, 'loss/train': 2.1779215335845947} 02/26/2022 14:46:15 - INFO - codeparrot_training - Step 47241: {'lr': 4.064903597241182e-06, 'samples': 24187904, 'steps': 47241, 'loss/train': 1.413515567779541} 02/26/2022 14:46:20 - INFO - codeparrot_training - Step 47242: {'lr': 4.061965485374186e-06, 'samples': 24188416, 'steps': 47242, 'loss/train': 2.044708728790283} 02/26/2022 14:46:24 - INFO - codeparrot_training - Step 47243: {'lr': 4.059028427027683e-06, 'samples': 24188928, 'steps': 47243, 'loss/train': 1.5489630699157715} 02/26/2022 14:46:27 - INFO - codeparrot_training - Step 47244: {'lr': 4.056092422214164e-06, 'samples': 24189440, 'steps': 47244, 'loss/train': 1.5935860872268677} 02/26/2022 14:46:33 - INFO - codeparrot_training - Step 47245: {'lr': 4.053157470946256e-06, 'samples': 24189952, 'steps': 47245, 'loss/train': 3.0423436164855957} 02/26/2022 14:46:36 - INFO - codeparrot_training - Step 47246: {'lr': 4.050223573236506e-06, 'samples': 24190464, 'steps': 47246, 'loss/train': 2.2079644203186035} 02/26/2022 14:46:42 - INFO - codeparrot_training - Step 47247: {'lr': 4.047290729097542e-06, 'samples': 24190976, 'steps': 47247, 'loss/train': 0.03209612891077995} 02/26/2022 14:46:46 - INFO - codeparrot_training - Step 47248: {'lr': 4.0443589385418534e-06, 'samples': 24191488, 'steps': 47248, 'loss/train': 2.941185235977173} 02/26/2022 14:46:51 - INFO - codeparrot_training - Step 47249: {'lr': 4.041428201582042e-06, 'samples': 24192000, 'steps': 47249, 'loss/train': 1.6676666736602783} 02/26/2022 14:46:55 - INFO - codeparrot_training - Step 47250: {'lr': 4.038498518230627e-06, 'samples': 24192512, 'steps': 47250, 'loss/train': 2.11346173286438} 02/26/2022 14:47:01 - INFO - codeparrot_training - Step 47251: {'lr': 4.035569888500235e-06, 'samples': 24193024, 'steps': 47251, 'loss/train': 1.7145932912826538} 02/26/2022 14:47:06 - INFO - codeparrot_training - Step 47252: {'lr': 4.032642312403329e-06, 'samples': 24193536, 'steps': 47252, 'loss/train': 1.607780933380127} 02/26/2022 14:47:09 - INFO - codeparrot_training - Step 47253: {'lr': 4.029715789952482e-06, 'samples': 24194048, 'steps': 47253, 'loss/train': 1.9386937618255615} 02/26/2022 14:47:15 - INFO - codeparrot_training - Step 47254: {'lr': 4.026790321160212e-06, 'samples': 24194560, 'steps': 47254, 'loss/train': 1.0495073795318604} 02/26/2022 14:47:18 - INFO - codeparrot_training - Step 47255: {'lr': 4.023865906039093e-06, 'samples': 24195072, 'steps': 47255, 'loss/train': 1.0093233585357666} 02/26/2022 14:47:25 - INFO - codeparrot_training - Step 47256: {'lr': 4.020942544601614e-06, 'samples': 24195584, 'steps': 47256, 'loss/train': 1.9473412036895752} 02/26/2022 14:47:28 - INFO - codeparrot_training - Step 47257: {'lr': 4.01802023686032e-06, 'samples': 24196096, 'steps': 47257, 'loss/train': 1.7149568796157837} 02/26/2022 14:47:34 - INFO - codeparrot_training - Step 47258: {'lr': 4.0150989828277306e-06, 'samples': 24196608, 'steps': 47258, 'loss/train': 1.9411085844039917} 02/26/2022 14:47:37 - INFO - codeparrot_training - Step 47259: {'lr': 4.012178782516307e-06, 'samples': 24197120, 'steps': 47259, 'loss/train': 2.141460657119751} 02/26/2022 14:47:43 - INFO - codeparrot_training - Step 47260: {'lr': 4.009259635938623e-06, 'samples': 24197632, 'steps': 47260, 'loss/train': 2.2508716583251953} 02/26/2022 14:47:46 - INFO - codeparrot_training - Step 47261: {'lr': 4.006341543107167e-06, 'samples': 24198144, 'steps': 47261, 'loss/train': 2.378004789352417} 02/26/2022 14:47:52 - INFO - codeparrot_training - Step 47262: {'lr': 4.003424504034431e-06, 'samples': 24198656, 'steps': 47262, 'loss/train': 1.7187739610671997} 02/26/2022 14:47:55 - INFO - codeparrot_training - Step 47263: {'lr': 4.000508518732876e-06, 'samples': 24199168, 'steps': 47263, 'loss/train': 1.439517855644226} 02/26/2022 14:48:01 - INFO - codeparrot_training - Step 47264: {'lr': 3.997593587215076e-06, 'samples': 24199680, 'steps': 47264, 'loss/train': 1.0263742208480835} 02/26/2022 14:48:04 - INFO - codeparrot_training - Step 47265: {'lr': 3.994679709493437e-06, 'samples': 24200192, 'steps': 47265, 'loss/train': 1.2490408420562744} 02/26/2022 14:48:10 - INFO - codeparrot_training - Step 47266: {'lr': 3.991766885580505e-06, 'samples': 24200704, 'steps': 47266, 'loss/train': 1.748345136642456} 02/26/2022 14:48:13 - INFO - codeparrot_training - Step 47267: {'lr': 3.988855115488688e-06, 'samples': 24201216, 'steps': 47267, 'loss/train': 1.529651165008545} 02/26/2022 14:48:19 - INFO - codeparrot_training - Step 47268: {'lr': 3.9859443992305285e-06, 'samples': 24201728, 'steps': 47268, 'loss/train': 2.4793283939361572} 02/26/2022 14:48:22 - INFO - codeparrot_training - Step 47269: {'lr': 3.983034736818464e-06, 'samples': 24202240, 'steps': 47269, 'loss/train': 1.4896708726882935} 02/26/2022 14:48:28 - INFO - codeparrot_training - Step 47270: {'lr': 3.980126128264955e-06, 'samples': 24202752, 'steps': 47270, 'loss/train': 2.1946334838867188} 02/26/2022 14:48:32 - INFO - codeparrot_training - Step 47271: {'lr': 3.977218573582464e-06, 'samples': 24203264, 'steps': 47271, 'loss/train': 2.0485382080078125} 02/26/2022 14:48:37 - INFO - codeparrot_training - Step 47272: {'lr': 3.974312072783454e-06, 'samples': 24203776, 'steps': 47272, 'loss/train': 2.103181838989258} 02/26/2022 14:48:41 - INFO - codeparrot_training - Step 47273: {'lr': 3.971406625880358e-06, 'samples': 24204288, 'steps': 47273, 'loss/train': 2.0182337760925293} 02/26/2022 14:48:46 - INFO - codeparrot_training - Step 47274: {'lr': 3.9685022328856405e-06, 'samples': 24204800, 'steps': 47274, 'loss/train': 1.8431813716888428} 02/26/2022 14:48:50 - INFO - codeparrot_training - Step 47275: {'lr': 3.965598893811734e-06, 'samples': 24205312, 'steps': 47275, 'loss/train': 1.9366247653961182} 02/26/2022 14:48:55 - INFO - codeparrot_training - Step 47276: {'lr': 3.9626966086710735e-06, 'samples': 24205824, 'steps': 47276, 'loss/train': 2.3885579109191895} 02/26/2022 14:48:59 - INFO - codeparrot_training - Step 47277: {'lr': 3.959795377476095e-06, 'samples': 24206336, 'steps': 47277, 'loss/train': 1.496716022491455} 02/26/2022 14:49:04 - INFO - codeparrot_training - Step 47278: {'lr': 3.956895200239258e-06, 'samples': 24206848, 'steps': 47278, 'loss/train': 2.5304102897644043} 02/26/2022 14:49:08 - INFO - codeparrot_training - Step 47279: {'lr': 3.953996076972888e-06, 'samples': 24207360, 'steps': 47279, 'loss/train': 1.0460617542266846} 02/26/2022 14:49:13 - INFO - codeparrot_training - Step 47280: {'lr': 3.95109800768953e-06, 'samples': 24207872, 'steps': 47280, 'loss/train': 2.0970051288604736} 02/26/2022 14:49:17 - INFO - codeparrot_training - Step 47281: {'lr': 3.948200992401507e-06, 'samples': 24208384, 'steps': 47281, 'loss/train': 1.5946303606033325} 02/26/2022 14:49:23 - INFO - codeparrot_training - Step 47282: {'lr': 3.945305031121283e-06, 'samples': 24208896, 'steps': 47282, 'loss/train': 2.0065109729766846} 02/26/2022 14:49:26 - INFO - codeparrot_training - Step 47283: {'lr': 3.9424101238612065e-06, 'samples': 24209408, 'steps': 47283, 'loss/train': 0.2687365710735321} 02/26/2022 14:49:32 - INFO - codeparrot_training - Step 47284: {'lr': 3.939516270633742e-06, 'samples': 24209920, 'steps': 47284, 'loss/train': 2.01710844039917} 02/26/2022 14:49:35 - INFO - codeparrot_training - Step 47285: {'lr': 3.93662347145124e-06, 'samples': 24210432, 'steps': 47285, 'loss/train': 2.6557397842407227} 02/26/2022 14:49:41 - INFO - codeparrot_training - Step 47286: {'lr': 3.933731726326162e-06, 'samples': 24210944, 'steps': 47286, 'loss/train': 2.4320626258850098} 02/26/2022 14:49:44 - INFO - codeparrot_training - Step 47287: {'lr': 3.9308410352707765e-06, 'samples': 24211456, 'steps': 47287, 'loss/train': 1.5946294069290161} 02/26/2022 14:49:50 - INFO - codeparrot_training - Step 47288: {'lr': 3.927951398297547e-06, 'samples': 24211968, 'steps': 47288, 'loss/train': 1.4650859832763672} 02/26/2022 14:49:53 - INFO - codeparrot_training - Step 47289: {'lr': 3.92506281541885e-06, 'samples': 24212480, 'steps': 47289, 'loss/train': 1.8791393041610718} 02/26/2022 14:49:59 - INFO - codeparrot_training - Step 47290: {'lr': 3.922175286647067e-06, 'samples': 24212992, 'steps': 47290, 'loss/train': 2.2669014930725098} 02/26/2022 14:50:02 - INFO - codeparrot_training - Step 47291: {'lr': 3.919288811994492e-06, 'samples': 24213504, 'steps': 47291, 'loss/train': 2.48101806640625} 02/26/2022 14:50:09 - INFO - codeparrot_training - Step 47292: {'lr': 3.916403391473588e-06, 'samples': 24214016, 'steps': 47292, 'loss/train': 1.8488614559173584} 02/26/2022 14:50:12 - INFO - codeparrot_training - Step 47293: {'lr': 3.913519025096651e-06, 'samples': 24214528, 'steps': 47293, 'loss/train': 1.4907691478729248} 02/26/2022 14:50:17 - INFO - codeparrot_training - Step 47294: {'lr': 3.910635712876059e-06, 'samples': 24215040, 'steps': 47294, 'loss/train': 0.8014967441558838} 02/26/2022 14:50:21 - INFO - codeparrot_training - Step 47295: {'lr': 3.907753454824165e-06, 'samples': 24215552, 'steps': 47295, 'loss/train': 1.6446328163146973} 02/26/2022 14:50:26 - INFO - codeparrot_training - Step 47296: {'lr': 3.904872250953317e-06, 'samples': 24216064, 'steps': 47296, 'loss/train': 1.9052540063858032} 02/26/2022 14:50:30 - INFO - codeparrot_training - Step 47297: {'lr': 3.901992101275842e-06, 'samples': 24216576, 'steps': 47297, 'loss/train': 2.2618050575256348} 02/26/2022 14:50:36 - INFO - codeparrot_training - Step 47298: {'lr': 3.899113005804089e-06, 'samples': 24217088, 'steps': 47298, 'loss/train': 2.1732382774353027} 02/26/2022 14:50:39 - INFO - codeparrot_training - Step 47299: {'lr': 3.89623496455041e-06, 'samples': 24217600, 'steps': 47299, 'loss/train': 2.699371814727783} 02/26/2022 14:50:45 - INFO - codeparrot_training - Step 47300: {'lr': 3.893357977527101e-06, 'samples': 24218112, 'steps': 47300, 'loss/train': 1.1672178506851196} 02/26/2022 14:50:48 - INFO - codeparrot_training - Step 47301: {'lr': 3.890482044746485e-06, 'samples': 24218624, 'steps': 47301, 'loss/train': 1.9867523908615112} 02/26/2022 14:50:54 - INFO - codeparrot_training - Step 47302: {'lr': 3.887607166220914e-06, 'samples': 24219136, 'steps': 47302, 'loss/train': 2.301151752471924} 02/26/2022 14:50:58 - INFO - codeparrot_training - Step 47303: {'lr': 3.884733341962682e-06, 'samples': 24219648, 'steps': 47303, 'loss/train': 1.9328258037567139} 02/26/2022 14:51:03 - INFO - codeparrot_training - Step 47304: {'lr': 3.881860571984086e-06, 'samples': 24220160, 'steps': 47304, 'loss/train': 2.714582681655884} 02/26/2022 14:51:06 - INFO - codeparrot_training - Step 47305: {'lr': 3.87898885629745e-06, 'samples': 24220672, 'steps': 47305, 'loss/train': 1.4150865077972412} 02/26/2022 14:51:12 - INFO - codeparrot_training - Step 47306: {'lr': 3.876118194915096e-06, 'samples': 24221184, 'steps': 47306, 'loss/train': 1.7300724983215332} 02/26/2022 14:51:15 - INFO - codeparrot_training - Step 47307: {'lr': 3.873248587849293e-06, 'samples': 24221696, 'steps': 47307, 'loss/train': 1.9061883687973022} 02/26/2022 14:51:21 - INFO - codeparrot_training - Step 47308: {'lr': 3.870380035112336e-06, 'samples': 24222208, 'steps': 47308, 'loss/train': 1.6285836696624756} 02/26/2022 14:51:24 - INFO - codeparrot_training - Step 47309: {'lr': 3.867512536716522e-06, 'samples': 24222720, 'steps': 47309, 'loss/train': 2.1681153774261475} 02/26/2022 14:51:30 - INFO - codeparrot_training - Step 47310: {'lr': 3.86464609267409e-06, 'samples': 24223232, 'steps': 47310, 'loss/train': 2.4143123626708984} 02/26/2022 14:51:33 - INFO - codeparrot_training - Step 47311: {'lr': 3.861780702997392e-06, 'samples': 24223744, 'steps': 47311, 'loss/train': 0.46697837114334106} 02/26/2022 14:51:40 - INFO - codeparrot_training - Step 47312: {'lr': 3.858916367698667e-06, 'samples': 24224256, 'steps': 47312, 'loss/train': 2.148773193359375} 02/26/2022 14:51:43 - INFO - codeparrot_training - Step 47313: {'lr': 3.856053086790184e-06, 'samples': 24224768, 'steps': 47313, 'loss/train': 1.2401666641235352} 02/26/2022 14:51:49 - INFO - codeparrot_training - Step 47314: {'lr': 3.853190860284184e-06, 'samples': 24225280, 'steps': 47314, 'loss/train': 1.8470346927642822} 02/26/2022 14:51:52 - INFO - codeparrot_training - Step 47315: {'lr': 3.850329688192988e-06, 'samples': 24225792, 'steps': 47315, 'loss/train': 1.945963740348816} 02/26/2022 14:51:58 - INFO - codeparrot_training - Step 47316: {'lr': 3.8474695705287836e-06, 'samples': 24226304, 'steps': 47316, 'loss/train': 1.5298091173171997} 02/26/2022 14:52:01 - INFO - codeparrot_training - Step 47317: {'lr': 3.844610507303919e-06, 'samples': 24226816, 'steps': 47317, 'loss/train': 1.9373265504837036} 02/26/2022 14:52:07 - INFO - codeparrot_training - Step 47318: {'lr': 3.841752498530526e-06, 'samples': 24227328, 'steps': 47318, 'loss/train': 1.685444712638855} 02/26/2022 14:52:10 - INFO - codeparrot_training - Step 47319: {'lr': 3.838895544220927e-06, 'samples': 24227840, 'steps': 47319, 'loss/train': 1.9345566034317017} 02/26/2022 14:52:16 - INFO - codeparrot_training - Step 47320: {'lr': 3.836039644387307e-06, 'samples': 24228352, 'steps': 47320, 'loss/train': 1.5173789262771606} 02/26/2022 14:52:19 - INFO - codeparrot_training - Step 47321: {'lr': 3.833184799041961e-06, 'samples': 24228864, 'steps': 47321, 'loss/train': 0.48507097363471985} 02/26/2022 14:52:25 - INFO - codeparrot_training - Step 47322: {'lr': 3.830331008197046e-06, 'samples': 24229376, 'steps': 47322, 'loss/train': 1.4980988502502441} 02/26/2022 14:52:28 - INFO - codeparrot_training - Step 47323: {'lr': 3.8274782718648594e-06, 'samples': 24229888, 'steps': 47323, 'loss/train': 0.8690078258514404} 02/26/2022 14:52:34 - INFO - codeparrot_training - Step 47324: {'lr': 3.824626590057556e-06, 'samples': 24230400, 'steps': 47324, 'loss/train': 3.355264186859131} 02/26/2022 14:52:37 - INFO - codeparrot_training - Step 47325: {'lr': 3.8217759627874326e-06, 'samples': 24230912, 'steps': 47325, 'loss/train': 2.0121114253997803} 02/26/2022 14:52:43 - INFO - codeparrot_training - Step 47326: {'lr': 3.8189263900665905e-06, 'samples': 24231424, 'steps': 47326, 'loss/train': 1.4708033800125122} 02/26/2022 14:52:46 - INFO - codeparrot_training - Step 47327: {'lr': 3.816077871907325e-06, 'samples': 24231936, 'steps': 47327, 'loss/train': 2.1761159896850586} 02/26/2022 14:52:53 - INFO - codeparrot_training - Step 47328: {'lr': 3.8132304083217937e-06, 'samples': 24232448, 'steps': 47328, 'loss/train': 1.5536339282989502} 02/26/2022 14:52:56 - INFO - codeparrot_training - Step 47329: {'lr': 3.810383999322237e-06, 'samples': 24232960, 'steps': 47329, 'loss/train': 1.9731495380401611} 02/26/2022 14:53:02 - INFO - codeparrot_training - Step 47330: {'lr': 3.8075386449208103e-06, 'samples': 24233472, 'steps': 47330, 'loss/train': 1.7803375720977783} 02/26/2022 14:53:05 - INFO - codeparrot_training - Step 47331: {'lr': 3.8046943451297276e-06, 'samples': 24233984, 'steps': 47331, 'loss/train': 1.6086125373840332} 02/26/2022 14:53:11 - INFO - codeparrot_training - Step 47332: {'lr': 3.8018510999611454e-06, 'samples': 24234496, 'steps': 47332, 'loss/train': 2.4230523109436035} 02/26/2022 14:53:14 - INFO - codeparrot_training - Step 47333: {'lr': 3.7990089094272485e-06, 'samples': 24235008, 'steps': 47333, 'loss/train': 0.918489396572113} 02/26/2022 14:53:20 - INFO - codeparrot_training - Step 47334: {'lr': 3.796167773540221e-06, 'samples': 24235520, 'steps': 47334, 'loss/train': 1.5616077184677124} 02/26/2022 14:53:23 - INFO - codeparrot_training - Step 47335: {'lr': 3.7933276923122484e-06, 'samples': 24236032, 'steps': 47335, 'loss/train': 1.6945126056671143} 02/26/2022 14:53:29 - INFO - codeparrot_training - Step 47336: {'lr': 3.790488665755459e-06, 'samples': 24236544, 'steps': 47336, 'loss/train': 0.9781691431999207} 02/26/2022 14:53:32 - INFO - codeparrot_training - Step 47337: {'lr': 3.7876506938820386e-06, 'samples': 24237056, 'steps': 47337, 'loss/train': 2.12276291847229} 02/26/2022 14:53:39 - INFO - codeparrot_training - Step 47338: {'lr': 3.7848137767041713e-06, 'samples': 24237568, 'steps': 47338, 'loss/train': 1.6119569540023804} 02/26/2022 14:53:42 - INFO - codeparrot_training - Step 47339: {'lr': 3.7819779142339582e-06, 'samples': 24238080, 'steps': 47339, 'loss/train': 1.9657509326934814} 02/26/2022 14:53:48 - INFO - codeparrot_training - Step 47340: {'lr': 3.779143106483557e-06, 'samples': 24238592, 'steps': 47340, 'loss/train': 1.524492621421814} 02/26/2022 14:53:51 - INFO - codeparrot_training - Step 47341: {'lr': 3.7763093534651237e-06, 'samples': 24239104, 'steps': 47341, 'loss/train': 1.890255331993103} 02/26/2022 14:53:57 - INFO - codeparrot_training - Step 47342: {'lr': 3.7734766551908163e-06, 'samples': 24239616, 'steps': 47342, 'loss/train': 0.9133675694465637} 02/26/2022 14:54:00 - INFO - codeparrot_training - Step 47343: {'lr': 3.770645011672763e-06, 'samples': 24240128, 'steps': 47343, 'loss/train': 1.5039483308792114} 02/26/2022 14:54:06 - INFO - codeparrot_training - Step 47344: {'lr': 3.7678144229230382e-06, 'samples': 24240640, 'steps': 47344, 'loss/train': 2.1652932167053223} 02/26/2022 14:54:09 - INFO - codeparrot_training - Step 47345: {'lr': 3.764984888953826e-06, 'samples': 24241152, 'steps': 47345, 'loss/train': 1.8980461359024048} 02/26/2022 14:54:15 - INFO - codeparrot_training - Step 47346: {'lr': 3.762156409777229e-06, 'samples': 24241664, 'steps': 47346, 'loss/train': 0.4810251295566559} 02/26/2022 14:54:18 - INFO - codeparrot_training - Step 47347: {'lr': 3.759328985405375e-06, 'samples': 24242176, 'steps': 47347, 'loss/train': 2.0912744998931885} 02/26/2022 14:54:24 - INFO - codeparrot_training - Step 47348: {'lr': 3.756502615850338e-06, 'samples': 24242688, 'steps': 47348, 'loss/train': 1.7837584018707275} 02/26/2022 14:54:27 - INFO - codeparrot_training - Step 47349: {'lr': 3.7536773011242753e-06, 'samples': 24243200, 'steps': 47349, 'loss/train': 2.5813305377960205} 02/26/2022 14:54:34 - INFO - codeparrot_training - Step 47350: {'lr': 3.7508530412392883e-06, 'samples': 24243712, 'steps': 47350, 'loss/train': 1.704620361328125} 02/26/2022 14:54:39 - INFO - codeparrot_training - Step 47351: {'lr': 3.748029836207395e-06, 'samples': 24244224, 'steps': 47351, 'loss/train': 0.703568696975708} 02/26/2022 14:54:43 - INFO - codeparrot_training - Step 47352: {'lr': 3.745207686040808e-06, 'samples': 24244736, 'steps': 47352, 'loss/train': 0.9513370990753174} 02/26/2022 14:54:48 - INFO - codeparrot_training - Step 47353: {'lr': 3.7423865907515175e-06, 'samples': 24245248, 'steps': 47353, 'loss/train': 2.3073689937591553} 02/26/2022 14:54:52 - INFO - codeparrot_training - Step 47354: {'lr': 3.7395665503516806e-06, 'samples': 24245760, 'steps': 47354, 'loss/train': 1.8946318626403809} 02/26/2022 14:54:57 - INFO - codeparrot_training - Step 47355: {'lr': 3.7367475648533157e-06, 'samples': 24246272, 'steps': 47355, 'loss/train': 1.4713902473449707} 02/26/2022 14:55:01 - INFO - codeparrot_training - Step 47356: {'lr': 3.7339296342685515e-06, 'samples': 24246784, 'steps': 47356, 'loss/train': 1.4497612714767456} 02/26/2022 14:55:06 - INFO - codeparrot_training - Step 47357: {'lr': 3.7311127586094064e-06, 'samples': 24247296, 'steps': 47357, 'loss/train': 1.3966649770736694} 02/26/2022 14:55:10 - INFO - codeparrot_training - Step 47358: {'lr': 3.7282969378879816e-06, 'samples': 24247808, 'steps': 47358, 'loss/train': 0.34197771549224854} 02/26/2022 14:55:15 - INFO - codeparrot_training - Step 47359: {'lr': 3.7254821721163233e-06, 'samples': 24248320, 'steps': 47359, 'loss/train': 0.8078081011772156} 02/26/2022 14:55:19 - INFO - codeparrot_training - Step 47360: {'lr': 3.722668461306533e-06, 'samples': 24248832, 'steps': 47360, 'loss/train': 1.8701469898223877} 02/26/2022 14:55:24 - INFO - codeparrot_training - Step 47361: {'lr': 3.719855805470601e-06, 'samples': 24249344, 'steps': 47361, 'loss/train': 2.415700912475586} 02/26/2022 14:55:28 - INFO - codeparrot_training - Step 47362: {'lr': 3.7170442046206287e-06, 'samples': 24249856, 'steps': 47362, 'loss/train': 1.931935429573059} 02/26/2022 14:55:34 - INFO - codeparrot_training - Step 47363: {'lr': 3.7142336587686064e-06, 'samples': 24250368, 'steps': 47363, 'loss/train': 1.5630022287368774} 02/26/2022 14:55:37 - INFO - codeparrot_training - Step 47364: {'lr': 3.711424167926608e-06, 'samples': 24250880, 'steps': 47364, 'loss/train': 2.204237937927246} 02/26/2022 14:55:43 - INFO - codeparrot_training - Step 47365: {'lr': 3.7086157321066794e-06, 'samples': 24251392, 'steps': 47365, 'loss/train': 3.933102607727051} 02/26/2022 14:55:46 - INFO - codeparrot_training - Step 47366: {'lr': 3.7058083513208384e-06, 'samples': 24251904, 'steps': 47366, 'loss/train': 1.891339659690857} 02/26/2022 14:55:52 - INFO - codeparrot_training - Step 47367: {'lr': 3.703002025581076e-06, 'samples': 24252416, 'steps': 47367, 'loss/train': 2.610166072845459} 02/26/2022 14:55:55 - INFO - codeparrot_training - Step 47368: {'lr': 3.700196754899465e-06, 'samples': 24252928, 'steps': 47368, 'loss/train': 1.6646748781204224} 02/26/2022 14:56:01 - INFO - codeparrot_training - Step 47369: {'lr': 3.697392539287997e-06, 'samples': 24253440, 'steps': 47369, 'loss/train': 2.169846296310425} 02/26/2022 14:56:04 - INFO - codeparrot_training - Step 47370: {'lr': 3.694589378758689e-06, 'samples': 24253952, 'steps': 47370, 'loss/train': 1.9926788806915283} 02/26/2022 14:56:10 - INFO - codeparrot_training - Step 47371: {'lr': 3.69178727332356e-06, 'samples': 24254464, 'steps': 47371, 'loss/train': 0.7172079086303711} 02/26/2022 14:56:13 - INFO - codeparrot_training - Step 47372: {'lr': 3.6889862229946004e-06, 'samples': 24254976, 'steps': 47372, 'loss/train': 1.9406155347824097} 02/26/2022 14:56:19 - INFO - codeparrot_training - Step 47373: {'lr': 3.6861862277838e-06, 'samples': 24255488, 'steps': 47373, 'loss/train': 2.2167246341705322} 02/26/2022 14:56:23 - INFO - codeparrot_training - Step 47374: {'lr': 3.683387287703177e-06, 'samples': 24256000, 'steps': 47374, 'loss/train': 2.2148501873016357} 02/26/2022 14:56:28 - INFO - codeparrot_training - Step 47375: {'lr': 3.680589402764695e-06, 'samples': 24256512, 'steps': 47375, 'loss/train': 0.8398675322532654} 02/26/2022 14:56:32 - INFO - codeparrot_training - Step 47376: {'lr': 3.6777925729803707e-06, 'samples': 24257024, 'steps': 47376, 'loss/train': 2.2023701667785645} 02/26/2022 14:56:37 - INFO - codeparrot_training - Step 47377: {'lr': 3.67499679836214e-06, 'samples': 24257536, 'steps': 47377, 'loss/train': 1.3533493280410767} 02/26/2022 14:56:41 - INFO - codeparrot_training - Step 47378: {'lr': 3.672202078922049e-06, 'samples': 24258048, 'steps': 47378, 'loss/train': 1.2798556089401245} 02/26/2022 14:56:46 - INFO - codeparrot_training - Step 47379: {'lr': 3.6694084146719764e-06, 'samples': 24258560, 'steps': 47379, 'loss/train': 1.0382750034332275} 02/26/2022 14:56:50 - INFO - codeparrot_training - Step 47380: {'lr': 3.6666158056239683e-06, 'samples': 24259072, 'steps': 47380, 'loss/train': 1.899681806564331} 02/26/2022 14:56:55 - INFO - codeparrot_training - Step 47381: {'lr': 3.6638242517899323e-06, 'samples': 24259584, 'steps': 47381, 'loss/train': 1.4494026899337769} 02/26/2022 14:56:59 - INFO - codeparrot_training - Step 47382: {'lr': 3.661033753181886e-06, 'samples': 24260096, 'steps': 47382, 'loss/train': 1.8838310241699219} 02/26/2022 14:57:05 - INFO - codeparrot_training - Step 47383: {'lr': 3.6582443098117367e-06, 'samples': 24260608, 'steps': 47383, 'loss/train': 2.3559370040893555} 02/26/2022 14:57:08 - INFO - codeparrot_training - Step 47384: {'lr': 3.6554559216914475e-06, 'samples': 24261120, 'steps': 47384, 'loss/train': 1.1949645280838013} 02/26/2022 14:57:14 - INFO - codeparrot_training - Step 47385: {'lr': 3.6526685888329525e-06, 'samples': 24261632, 'steps': 47385, 'loss/train': 0.6307389736175537} 02/26/2022 14:57:17 - INFO - codeparrot_training - Step 47386: {'lr': 3.6498823112482428e-06, 'samples': 24262144, 'steps': 47386, 'loss/train': 1.106641411781311} 02/26/2022 14:57:23 - INFO - codeparrot_training - Step 47387: {'lr': 3.6470970889491417e-06, 'samples': 24262656, 'steps': 47387, 'loss/train': 1.0352836847305298} 02/26/2022 14:57:26 - INFO - codeparrot_training - Step 47388: {'lr': 3.6443129219476958e-06, 'samples': 24263168, 'steps': 47388, 'loss/train': 1.0688190460205078} 02/26/2022 14:57:32 - INFO - codeparrot_training - Step 47389: {'lr': 3.6415298102557838e-06, 'samples': 24263680, 'steps': 47389, 'loss/train': 1.5037307739257812} 02/26/2022 14:57:35 - INFO - codeparrot_training - Step 47390: {'lr': 3.6387477538853132e-06, 'samples': 24264192, 'steps': 47390, 'loss/train': 1.4985733032226562} 02/26/2022 14:57:41 - INFO - codeparrot_training - Step 47391: {'lr': 3.6359667528482464e-06, 'samples': 24264704, 'steps': 47391, 'loss/train': 1.855826735496521} 02/26/2022 14:57:44 - INFO - codeparrot_training - Step 47392: {'lr': 3.6331868071564634e-06, 'samples': 24265216, 'steps': 47392, 'loss/train': 1.3368381261825562} 02/26/2022 14:57:51 - INFO - codeparrot_training - Step 47393: {'lr': 3.6304079168218705e-06, 'samples': 24265728, 'steps': 47393, 'loss/train': 0.9868101477622986} 02/26/2022 14:57:54 - INFO - codeparrot_training - Step 47394: {'lr': 3.6276300818563477e-06, 'samples': 24266240, 'steps': 47394, 'loss/train': 2.121750831604004} 02/26/2022 14:57:59 - INFO - codeparrot_training - Step 47395: {'lr': 3.624853302271858e-06, 'samples': 24266752, 'steps': 47395, 'loss/train': 0.9207903146743774} 02/26/2022 14:58:03 - INFO - codeparrot_training - Step 47396: {'lr': 3.6220775780802794e-06, 'samples': 24267264, 'steps': 47396, 'loss/train': 1.5492602586746216} 02/26/2022 14:58:08 - INFO - codeparrot_training - Step 47397: {'lr': 3.619302909293465e-06, 'samples': 24267776, 'steps': 47397, 'loss/train': 1.608044981956482} 02/26/2022 14:58:12 - INFO - codeparrot_training - Step 47398: {'lr': 3.616529295923321e-06, 'samples': 24268288, 'steps': 47398, 'loss/train': 1.4074846506118774} 02/26/2022 14:58:18 - INFO - codeparrot_training - Step 47399: {'lr': 3.613756737981727e-06, 'samples': 24268800, 'steps': 47399, 'loss/train': 1.6231176853179932} 02/26/2022 14:58:21 - INFO - codeparrot_training - Step 47400: {'lr': 3.6109852354805626e-06, 'samples': 24269312, 'steps': 47400, 'loss/train': 1.6907190084457397} 02/26/2022 14:58:27 - INFO - codeparrot_training - Step 47401: {'lr': 3.608214788431735e-06, 'samples': 24269824, 'steps': 47401, 'loss/train': 1.3162966966629028} 02/26/2022 14:58:30 - INFO - codeparrot_training - Step 47402: {'lr': 3.6054453968470124e-06, 'samples': 24270336, 'steps': 47402, 'loss/train': 0.8982712626457214} 02/26/2022 14:58:36 - INFO - codeparrot_training - Step 47403: {'lr': 3.6026770607383853e-06, 'samples': 24270848, 'steps': 47403, 'loss/train': 1.4896438121795654} 02/26/2022 14:58:39 - INFO - codeparrot_training - Step 47404: {'lr': 3.599909780117622e-06, 'samples': 24271360, 'steps': 47404, 'loss/train': 1.3685739040374756} 02/26/2022 14:58:45 - INFO - codeparrot_training - Step 47405: {'lr': 3.5971435549966012e-06, 'samples': 24271872, 'steps': 47405, 'loss/train': 1.7780133485794067} 02/26/2022 14:58:48 - INFO - codeparrot_training - Step 47406: {'lr': 3.594378385387176e-06, 'samples': 24272384, 'steps': 47406, 'loss/train': 2.2715048789978027} 02/26/2022 14:58:54 - INFO - codeparrot_training - Step 47407: {'lr': 3.5916142713011967e-06, 'samples': 24272896, 'steps': 47407, 'loss/train': 2.9677579402923584} 02/26/2022 14:58:57 - INFO - codeparrot_training - Step 47408: {'lr': 3.588851212750488e-06, 'samples': 24273408, 'steps': 47408, 'loss/train': 2.1170194149017334} 02/26/2022 14:59:03 - INFO - codeparrot_training - Step 47409: {'lr': 3.586089209746929e-06, 'samples': 24273920, 'steps': 47409, 'loss/train': 1.920351505279541} 02/26/2022 14:59:07 - INFO - codeparrot_training - Step 47410: {'lr': 3.5833282623022877e-06, 'samples': 24274432, 'steps': 47410, 'loss/train': 2.446288585662842} 02/26/2022 14:59:12 - INFO - codeparrot_training - Step 47411: {'lr': 3.5805683704284165e-06, 'samples': 24274944, 'steps': 47411, 'loss/train': 1.3646293878555298} 02/26/2022 14:59:16 - INFO - codeparrot_training - Step 47412: {'lr': 3.577809534137139e-06, 'samples': 24275456, 'steps': 47412, 'loss/train': 2.390080690383911} 02/26/2022 14:59:21 - INFO - codeparrot_training - Step 47413: {'lr': 3.5750517534403336e-06, 'samples': 24275968, 'steps': 47413, 'loss/train': 0.7405255436897278} 02/26/2022 14:59:25 - INFO - codeparrot_training - Step 47414: {'lr': 3.572295028349687e-06, 'samples': 24276480, 'steps': 47414, 'loss/train': 1.668779969215393} 02/26/2022 14:59:30 - INFO - codeparrot_training - Step 47415: {'lr': 3.569539358877133e-06, 'samples': 24276992, 'steps': 47415, 'loss/train': 2.4658994674682617} 02/26/2022 14:59:34 - INFO - codeparrot_training - Step 47416: {'lr': 3.566784745034385e-06, 'samples': 24277504, 'steps': 47416, 'loss/train': 1.9553532600402832} 02/26/2022 14:59:39 - INFO - codeparrot_training - Step 47417: {'lr': 3.564031186833322e-06, 'samples': 24278016, 'steps': 47417, 'loss/train': 1.6932862997055054} 02/26/2022 14:59:43 - INFO - codeparrot_training - Step 47418: {'lr': 3.561278684285657e-06, 'samples': 24278528, 'steps': 47418, 'loss/train': 1.7158762216567993} 02/26/2022 14:59:49 - INFO - codeparrot_training - Step 47419: {'lr': 3.558527237403242e-06, 'samples': 24279040, 'steps': 47419, 'loss/train': 1.5551339387893677} 02/26/2022 14:59:52 - INFO - codeparrot_training - Step 47420: {'lr': 3.555776846197817e-06, 'samples': 24279552, 'steps': 47420, 'loss/train': 1.2387826442718506} 02/26/2022 14:59:58 - INFO - codeparrot_training - Step 47421: {'lr': 3.553027510681206e-06, 'samples': 24280064, 'steps': 47421, 'loss/train': 1.6238585710525513} 02/26/2022 15:00:02 - INFO - codeparrot_training - Step 47422: {'lr': 3.5502792308651776e-06, 'samples': 24280576, 'steps': 47422, 'loss/train': 1.372475266456604} 02/26/2022 15:00:07 - INFO - codeparrot_training - Step 47423: {'lr': 3.5475320067614726e-06, 'samples': 24281088, 'steps': 47423, 'loss/train': 1.4243037700653076} 02/26/2022 15:00:11 - INFO - codeparrot_training - Step 47424: {'lr': 3.5447858383818865e-06, 'samples': 24281600, 'steps': 47424, 'loss/train': 2.4198484420776367} 02/26/2022 15:00:16 - INFO - codeparrot_training - Step 47425: {'lr': 3.542040725738216e-06, 'samples': 24282112, 'steps': 47425, 'loss/train': 0.3040185868740082} 02/26/2022 15:00:20 - INFO - codeparrot_training - Step 47426: {'lr': 3.539296668842146e-06, 'samples': 24282624, 'steps': 47426, 'loss/train': 1.7414060831069946} 02/26/2022 15:00:25 - INFO - codeparrot_training - Step 47427: {'lr': 3.5365536677055e-06, 'samples': 24283136, 'steps': 47427, 'loss/train': 2.551719903945923} 02/26/2022 15:00:29 - INFO - codeparrot_training - Step 47428: {'lr': 3.5338117223399634e-06, 'samples': 24283648, 'steps': 47428, 'loss/train': 1.7943593263626099} 02/26/2022 15:00:35 - INFO - codeparrot_training - Step 47429: {'lr': 3.5310708327573326e-06, 'samples': 24284160, 'steps': 47429, 'loss/train': 1.3163498640060425} 02/26/2022 15:00:38 - INFO - codeparrot_training - Step 47430: {'lr': 3.528330998969348e-06, 'samples': 24284672, 'steps': 47430, 'loss/train': 3.230424404144287} 02/26/2022 15:00:45 - INFO - codeparrot_training - Step 47431: {'lr': 3.525592220987722e-06, 'samples': 24285184, 'steps': 47431, 'loss/train': 1.8927180767059326} 02/26/2022 15:00:48 - INFO - codeparrot_training - Step 47432: {'lr': 3.5228544988241682e-06, 'samples': 24285696, 'steps': 47432, 'loss/train': 1.9084442853927612} 02/26/2022 15:00:53 - INFO - codeparrot_training - Step 47433: {'lr': 3.520117832490455e-06, 'samples': 24286208, 'steps': 47433, 'loss/train': 1.9714888334274292} 02/26/2022 15:00:57 - INFO - codeparrot_training - Step 47434: {'lr': 3.517382221998294e-06, 'samples': 24286720, 'steps': 47434, 'loss/train': 1.7211190462112427} 02/26/2022 15:01:03 - INFO - codeparrot_training - Step 47435: {'lr': 3.5146476673593995e-06, 'samples': 24287232, 'steps': 47435, 'loss/train': 1.8167824745178223} 02/26/2022 15:01:06 - INFO - codeparrot_training - Step 47436: {'lr': 3.511914168585484e-06, 'samples': 24287744, 'steps': 47436, 'loss/train': 0.7277877926826477} 02/26/2022 15:01:11 - INFO - codeparrot_training - Step 47437: {'lr': 3.5091817256882595e-06, 'samples': 24288256, 'steps': 47437, 'loss/train': 1.646148920059204} 02/26/2022 15:01:15 - INFO - codeparrot_training - Step 47438: {'lr': 3.50645033867944e-06, 'samples': 24288768, 'steps': 47438, 'loss/train': 1.8899078369140625} 02/26/2022 15:01:20 - INFO - codeparrot_training - Step 47439: {'lr': 3.50372000757071e-06, 'samples': 24289280, 'steps': 47439, 'loss/train': 0.9157168865203857} 02/26/2022 15:01:24 - INFO - codeparrot_training - Step 47440: {'lr': 3.5009907323737824e-06, 'samples': 24289792, 'steps': 47440, 'loss/train': 1.067230224609375} 02/26/2022 15:01:30 - INFO - codeparrot_training - Step 47441: {'lr': 3.4982625131003144e-06, 'samples': 24290304, 'steps': 47441, 'loss/train': 2.1384260654449463} 02/26/2022 15:01:34 - INFO - codeparrot_training - Step 47442: {'lr': 3.4955353497620468e-06, 'samples': 24290816, 'steps': 47442, 'loss/train': 2.2401299476623535} 02/26/2022 15:01:39 - INFO - codeparrot_training - Step 47443: {'lr': 3.492809242370609e-06, 'samples': 24291328, 'steps': 47443, 'loss/train': 1.7456283569335938} 02/26/2022 15:01:43 - INFO - codeparrot_training - Step 47444: {'lr': 3.4900841909377145e-06, 'samples': 24291840, 'steps': 47444, 'loss/train': 1.1179039478302002} 02/26/2022 15:01:48 - INFO - codeparrot_training - Step 47445: {'lr': 3.4873601954749923e-06, 'samples': 24292352, 'steps': 47445, 'loss/train': 1.4707119464874268} 02/26/2022 15:01:52 - INFO - codeparrot_training - Step 47446: {'lr': 3.484637255994183e-06, 'samples': 24292864, 'steps': 47446, 'loss/train': 1.4192888736724854} 02/26/2022 15:01:57 - INFO - codeparrot_training - Step 47447: {'lr': 3.4819153725068886e-06, 'samples': 24293376, 'steps': 47447, 'loss/train': 1.6675523519515991} 02/26/2022 15:02:01 - INFO - codeparrot_training - Step 47448: {'lr': 3.479194545024822e-06, 'samples': 24293888, 'steps': 47448, 'loss/train': 1.9217102527618408} 02/26/2022 15:02:06 - INFO - codeparrot_training - Step 47449: {'lr': 3.476474773559557e-06, 'samples': 24294400, 'steps': 47449, 'loss/train': 0.9532285332679749} 02/26/2022 15:02:10 - INFO - codeparrot_training - Step 47450: {'lr': 3.4737560581228343e-06, 'samples': 24294912, 'steps': 47450, 'loss/train': 1.9057599306106567} 02/26/2022 15:02:16 - INFO - codeparrot_training - Step 47451: {'lr': 3.4710383987262283e-06, 'samples': 24295424, 'steps': 47451, 'loss/train': 2.1189582347869873} 02/26/2022 15:02:19 - INFO - codeparrot_training - Step 47452: {'lr': 3.468321795381452e-06, 'samples': 24295936, 'steps': 47452, 'loss/train': 1.730580449104309} 02/26/2022 15:02:25 - INFO - codeparrot_training - Step 47453: {'lr': 3.4656062481000504e-06, 'samples': 24296448, 'steps': 47453, 'loss/train': 1.8071969747543335} 02/26/2022 15:02:28 - INFO - codeparrot_training - Step 47454: {'lr': 3.462891756893738e-06, 'samples': 24296960, 'steps': 47454, 'loss/train': 1.624642252922058} 02/26/2022 15:02:34 - INFO - codeparrot_training - Step 47455: {'lr': 3.4601783217740877e-06, 'samples': 24297472, 'steps': 47455, 'loss/train': 1.4693812131881714} 02/26/2022 15:02:37 - INFO - codeparrot_training - Step 47456: {'lr': 3.457465942752813e-06, 'samples': 24297984, 'steps': 47456, 'loss/train': 1.4773683547973633} 02/26/2022 15:02:43 - INFO - codeparrot_training - Step 47457: {'lr': 3.454754619841405e-06, 'samples': 24298496, 'steps': 47457, 'loss/train': 2.012878179550171} 02/26/2022 15:02:46 - INFO - codeparrot_training - Step 47458: {'lr': 3.4520443530515754e-06, 'samples': 24299008, 'steps': 47458, 'loss/train': 1.4629181623458862} 02/26/2022 15:02:52 - INFO - codeparrot_training - Step 47459: {'lr': 3.4493351423948715e-06, 'samples': 24299520, 'steps': 47459, 'loss/train': 1.534316062927246} 02/26/2022 15:02:55 - INFO - codeparrot_training - Step 47460: {'lr': 3.4466269878829504e-06, 'samples': 24300032, 'steps': 47460, 'loss/train': 1.194318413734436} 02/26/2022 15:03:01 - INFO - codeparrot_training - Step 47461: {'lr': 3.443919889527386e-06, 'samples': 24300544, 'steps': 47461, 'loss/train': 1.2803740501403809} 02/26/2022 15:03:04 - INFO - codeparrot_training - Step 47462: {'lr': 3.44121384733978e-06, 'samples': 24301056, 'steps': 47462, 'loss/train': 2.1351022720336914} 02/26/2022 15:03:10 - INFO - codeparrot_training - Step 47463: {'lr': 3.4385088613317075e-06, 'samples': 24301568, 'steps': 47463, 'loss/train': 1.7526729106903076} 02/26/2022 15:03:13 - INFO - codeparrot_training - Step 47464: {'lr': 3.435804931514769e-06, 'samples': 24302080, 'steps': 47464, 'loss/train': 1.1786935329437256} 02/26/2022 15:03:19 - INFO - codeparrot_training - Step 47465: {'lr': 3.4331020579005666e-06, 'samples': 24302592, 'steps': 47465, 'loss/train': 2.1805975437164307} 02/26/2022 15:03:22 - INFO - codeparrot_training - Step 47466: {'lr': 3.430400240500675e-06, 'samples': 24303104, 'steps': 47466, 'loss/train': 2.414189577102661} 02/26/2022 15:03:29 - INFO - codeparrot_training - Step 47467: {'lr': 3.4276994793266127e-06, 'samples': 24303616, 'steps': 47467, 'loss/train': 1.058998703956604} 02/26/2022 15:03:32 - INFO - codeparrot_training - Step 47468: {'lr': 3.4249997743900083e-06, 'samples': 24304128, 'steps': 47468, 'loss/train': 3.5504348278045654} 02/26/2022 15:03:38 - INFO - codeparrot_training - Step 47469: {'lr': 3.4223011257024097e-06, 'samples': 24304640, 'steps': 47469, 'loss/train': 2.891859769821167} 02/26/2022 15:03:41 - INFO - codeparrot_training - Step 47470: {'lr': 3.4196035332753893e-06, 'samples': 24305152, 'steps': 47470, 'loss/train': 1.0789157152175903} 02/26/2022 15:03:47 - INFO - codeparrot_training - Step 47471: {'lr': 3.416906997120467e-06, 'samples': 24305664, 'steps': 47471, 'loss/train': 2.1937248706817627} 02/26/2022 15:03:50 - INFO - codeparrot_training - Step 47472: {'lr': 3.4142115172492437e-06, 'samples': 24306176, 'steps': 47472, 'loss/train': 1.5575580596923828} 02/26/2022 15:03:56 - INFO - codeparrot_training - Step 47473: {'lr': 3.411517093673211e-06, 'samples': 24306688, 'steps': 47473, 'loss/train': 1.9336330890655518} 02/26/2022 15:03:59 - INFO - codeparrot_training - Step 47474: {'lr': 3.40882372640397e-06, 'samples': 24307200, 'steps': 47474, 'loss/train': 2.114950180053711} 02/26/2022 15:04:05 - INFO - codeparrot_training - Step 47475: {'lr': 3.4061314154529843e-06, 'samples': 24307712, 'steps': 47475, 'loss/train': 1.6089240312576294} 02/26/2022 15:04:08 - INFO - codeparrot_training - Step 47476: {'lr': 3.4034401608318557e-06, 'samples': 24308224, 'steps': 47476, 'loss/train': 0.06756830215454102} 02/26/2022 15:04:15 - INFO - codeparrot_training - Step 47477: {'lr': 3.4007499625520746e-06, 'samples': 24308736, 'steps': 47477, 'loss/train': 2.412686586380005} 02/26/2022 15:04:18 - INFO - codeparrot_training - Step 47478: {'lr': 3.3980608206251873e-06, 'samples': 24309248, 'steps': 47478, 'loss/train': 2.7915639877319336} 02/26/2022 15:04:24 - INFO - codeparrot_training - Step 47479: {'lr': 3.3953727350627126e-06, 'samples': 24309760, 'steps': 47479, 'loss/train': 2.4462890625} 02/26/2022 15:04:27 - INFO - codeparrot_training - Step 47480: {'lr': 3.392685705876142e-06, 'samples': 24310272, 'steps': 47480, 'loss/train': 0.7830101251602173} 02/26/2022 15:04:33 - INFO - codeparrot_training - Step 47481: {'lr': 3.3899997330769927e-06, 'samples': 24310784, 'steps': 47481, 'loss/train': 1.7793724536895752} 02/26/2022 15:04:36 - INFO - codeparrot_training - Step 47482: {'lr': 3.387314816676784e-06, 'samples': 24311296, 'steps': 47482, 'loss/train': 1.7306865453720093} 02/26/2022 15:04:42 - INFO - codeparrot_training - Step 47483: {'lr': 3.384630956687007e-06, 'samples': 24311808, 'steps': 47483, 'loss/train': 1.7700717449188232} 02/26/2022 15:04:45 - INFO - codeparrot_training - Step 47484: {'lr': 3.381948153119152e-06, 'samples': 24312320, 'steps': 47484, 'loss/train': 2.3642008304595947} 02/26/2022 15:04:51 - INFO - codeparrot_training - Step 47485: {'lr': 3.3792664059847376e-06, 'samples': 24312832, 'steps': 47485, 'loss/train': 1.9936814308166504} 02/26/2022 15:04:54 - INFO - codeparrot_training - Step 47486: {'lr': 3.3765857152952274e-06, 'samples': 24313344, 'steps': 47486, 'loss/train': 1.6109830141067505} 02/26/2022 15:05:01 - INFO - codeparrot_training - Step 47487: {'lr': 3.373906081062139e-06, 'samples': 24313856, 'steps': 47487, 'loss/train': 1.959956407546997} 02/26/2022 15:05:04 - INFO - codeparrot_training - Step 47488: {'lr': 3.3712275032968807e-06, 'samples': 24314368, 'steps': 47488, 'loss/train': 0.8398148417472839} 02/26/2022 15:05:10 - INFO - codeparrot_training - Step 47489: {'lr': 3.368549982010999e-06, 'samples': 24314880, 'steps': 47489, 'loss/train': 1.0779094696044922} 02/26/2022 15:05:13 - INFO - codeparrot_training - Step 47490: {'lr': 3.3658735172159284e-06, 'samples': 24315392, 'steps': 47490, 'loss/train': 2.073573350906372} 02/26/2022 15:05:19 - INFO - codeparrot_training - Step 47491: {'lr': 3.36319810892316e-06, 'samples': 24315904, 'steps': 47491, 'loss/train': 0.5230470895767212} 02/26/2022 15:05:22 - INFO - codeparrot_training - Step 47492: {'lr': 3.360523757144102e-06, 'samples': 24316416, 'steps': 47492, 'loss/train': 1.7574613094329834} 02/26/2022 15:05:28 - INFO - codeparrot_training - Step 47493: {'lr': 3.3578504618902437e-06, 'samples': 24316928, 'steps': 47493, 'loss/train': 0.9887236952781677} 02/26/2022 15:05:31 - INFO - codeparrot_training - Step 47494: {'lr': 3.35517822317305e-06, 'samples': 24317440, 'steps': 47494, 'loss/train': 1.171176791191101} 02/26/2022 15:05:37 - INFO - codeparrot_training - Step 47495: {'lr': 3.352507041003955e-06, 'samples': 24317952, 'steps': 47495, 'loss/train': 1.6361877918243408} 02/26/2022 15:05:40 - INFO - codeparrot_training - Step 47496: {'lr': 3.349836915394394e-06, 'samples': 24318464, 'steps': 47496, 'loss/train': 1.488653540611267} 02/26/2022 15:05:46 - INFO - codeparrot_training - Step 47497: {'lr': 3.347167846355803e-06, 'samples': 24318976, 'steps': 47497, 'loss/train': 2.1035194396972656} 02/26/2022 15:05:49 - INFO - codeparrot_training - Step 47498: {'lr': 3.3444998338996448e-06, 'samples': 24319488, 'steps': 47498, 'loss/train': 2.71098256111145} 02/26/2022 15:05:55 - INFO - codeparrot_training - Step 47499: {'lr': 3.3418328780372986e-06, 'samples': 24320000, 'steps': 47499, 'loss/train': 0.03847160562872887} 02/26/2022 15:05:58 - INFO - codeparrot_training - Step 47500: {'lr': 3.339166978780256e-06, 'samples': 24320512, 'steps': 47500, 'loss/train': 1.979993224143982} 02/26/2022 15:06:04 - INFO - codeparrot_training - Step 47501: {'lr': 3.3365021361398962e-06, 'samples': 24321024, 'steps': 47501, 'loss/train': 1.8054109811782837} 02/26/2022 15:06:07 - INFO - codeparrot_training - Step 47502: {'lr': 3.3338383501275993e-06, 'samples': 24321536, 'steps': 47502, 'loss/train': 0.2787412106990814} 02/26/2022 15:06:14 - INFO - codeparrot_training - Step 47503: {'lr': 3.3311756207548284e-06, 'samples': 24322048, 'steps': 47503, 'loss/train': 2.277825355529785} 02/26/2022 15:06:17 - INFO - codeparrot_training - Step 47504: {'lr': 3.328513948032991e-06, 'samples': 24322560, 'steps': 47504, 'loss/train': 1.191369891166687} 02/26/2022 15:06:23 - INFO - codeparrot_training - Step 47505: {'lr': 3.325853331973466e-06, 'samples': 24323072, 'steps': 47505, 'loss/train': 1.222564697265625} 02/26/2022 15:06:26 - INFO - codeparrot_training - Step 47506: {'lr': 3.3231937725876347e-06, 'samples': 24323584, 'steps': 47506, 'loss/train': 0.8408288359642029} 02/26/2022 15:06:32 - INFO - codeparrot_training - Step 47507: {'lr': 3.3205352698869317e-06, 'samples': 24324096, 'steps': 47507, 'loss/train': 0.8858546018600464} 02/26/2022 15:06:35 - INFO - codeparrot_training - Step 47508: {'lr': 3.317877823882737e-06, 'samples': 24324608, 'steps': 47508, 'loss/train': 1.0575590133666992} 02/26/2022 15:06:41 - INFO - codeparrot_training - Step 47509: {'lr': 3.3152214345864294e-06, 'samples': 24325120, 'steps': 47509, 'loss/train': 1.9508386850357056} 02/26/2022 15:06:46 - INFO - codeparrot_training - Step 47510: {'lr': 3.3125661020093346e-06, 'samples': 24325632, 'steps': 47510, 'loss/train': 2.042740821838379} 02/26/2022 15:06:50 - INFO - codeparrot_training - Step 47511: {'lr': 3.3099118261629147e-06, 'samples': 24326144, 'steps': 47511, 'loss/train': 1.6537288427352905} 02/26/2022 15:06:57 - INFO - codeparrot_training - Step 47512: {'lr': 3.3072586070584664e-06, 'samples': 24326656, 'steps': 47512, 'loss/train': 2.5362508296966553} 02/26/2022 15:07:00 - INFO - codeparrot_training - Step 47513: {'lr': 3.304606444707453e-06, 'samples': 24327168, 'steps': 47513, 'loss/train': 1.4630937576293945} 02/26/2022 15:07:04 - INFO - codeparrot_training - Step 47514: {'lr': 3.3019553391211153e-06, 'samples': 24327680, 'steps': 47514, 'loss/train': 0.23679926991462708} 02/26/2022 15:07:09 - INFO - codeparrot_training - Step 47515: {'lr': 3.299305290310889e-06, 'samples': 24328192, 'steps': 47515, 'loss/train': 1.9297516345977783} 02/26/2022 15:07:12 - INFO - codeparrot_training - Step 47516: {'lr': 3.2966562982880977e-06, 'samples': 24328704, 'steps': 47516, 'loss/train': 1.1394567489624023} 02/26/2022 15:07:18 - INFO - codeparrot_training - Step 47517: {'lr': 3.294008363064094e-06, 'samples': 24329216, 'steps': 47517, 'loss/train': 2.362560987472534} 02/26/2022 15:07:22 - INFO - codeparrot_training - Step 47518: {'lr': 3.2913614846502015e-06, 'samples': 24329728, 'steps': 47518, 'loss/train': 1.7055608034133911} 02/26/2022 15:07:27 - INFO - codeparrot_training - Step 47519: {'lr': 3.288715663057801e-06, 'samples': 24330240, 'steps': 47519, 'loss/train': 0.8176570534706116} 02/26/2022 15:07:31 - INFO - codeparrot_training - Step 47520: {'lr': 3.286070898298188e-06, 'samples': 24330752, 'steps': 47520, 'loss/train': 1.3797684907913208} 02/26/2022 15:07:36 - INFO - codeparrot_training - Step 47521: {'lr': 3.2834271903826873e-06, 'samples': 24331264, 'steps': 47521, 'loss/train': 1.334601640701294} 02/26/2022 15:07:42 - INFO - codeparrot_training - Step 47522: {'lr': 3.280784539322679e-06, 'samples': 24331776, 'steps': 47522, 'loss/train': 2.0882368087768555} 02/26/2022 15:07:46 - INFO - codeparrot_training - Step 47523: {'lr': 3.2781429451294586e-06, 'samples': 24332288, 'steps': 47523, 'loss/train': 3.377593517303467} 02/26/2022 15:07:49 - INFO - codeparrot_training - Step 47524: {'lr': 3.2755024078142957e-06, 'samples': 24332800, 'steps': 47524, 'loss/train': 1.9391717910766602} 02/26/2022 15:07:55 - INFO - codeparrot_training - Step 47525: {'lr': 3.2728629273885424e-06, 'samples': 24333312, 'steps': 47525, 'loss/train': 2.326525926589966} 02/26/2022 15:07:58 - INFO - codeparrot_training - Step 47526: {'lr': 3.2702245038635227e-06, 'samples': 24333824, 'steps': 47526, 'loss/train': 1.3061593770980835} 02/26/2022 15:08:04 - INFO - codeparrot_training - Step 47527: {'lr': 3.267587137250505e-06, 'samples': 24334336, 'steps': 47527, 'loss/train': 1.1045658588409424} 02/26/2022 15:08:09 - INFO - codeparrot_training - Step 47528: {'lr': 3.2649508275607863e-06, 'samples': 24334848, 'steps': 47528, 'loss/train': 1.3623377084732056} 02/26/2022 15:08:13 - INFO - codeparrot_training - Step 47529: {'lr': 3.262315574805663e-06, 'samples': 24335360, 'steps': 47529, 'loss/train': 1.696001648902893} 02/26/2022 15:08:18 - INFO - codeparrot_training - Step 47530: {'lr': 3.2596813789964596e-06, 'samples': 24335872, 'steps': 47530, 'loss/train': 1.374510407447815} 02/26/2022 15:08:22 - INFO - codeparrot_training - Step 47531: {'lr': 3.257048240144417e-06, 'samples': 24336384, 'steps': 47531, 'loss/train': 0.8178138732910156} 02/26/2022 15:08:28 - INFO - codeparrot_training - Step 47532: {'lr': 3.254416158260831e-06, 'samples': 24336896, 'steps': 47532, 'loss/train': 1.0946320295333862} 02/26/2022 15:08:32 - INFO - codeparrot_training - Step 47533: {'lr': 3.2517851333569716e-06, 'samples': 24337408, 'steps': 47533, 'loss/train': 0.9300341606140137} 02/26/2022 15:08:37 - INFO - codeparrot_training - Step 47534: {'lr': 3.249155165444134e-06, 'samples': 24337920, 'steps': 47534, 'loss/train': 1.6627955436706543} 02/26/2022 15:08:41 - INFO - codeparrot_training - Step 47535: {'lr': 3.246526254533533e-06, 'samples': 24338432, 'steps': 47535, 'loss/train': 1.8643912076950073} 02/26/2022 15:08:46 - INFO - codeparrot_training - Step 47536: {'lr': 3.243898400636491e-06, 'samples': 24338944, 'steps': 47536, 'loss/train': 1.4736212491989136} 02/26/2022 15:08:50 - INFO - codeparrot_training - Step 47537: {'lr': 3.2412716037641955e-06, 'samples': 24339456, 'steps': 47537, 'loss/train': 2.1177971363067627} 02/26/2022 15:08:55 - INFO - codeparrot_training - Step 47538: {'lr': 3.2386458639279693e-06, 'samples': 24339968, 'steps': 47538, 'loss/train': 1.907954216003418} 02/26/2022 15:08:59 - INFO - codeparrot_training - Step 47539: {'lr': 3.236021181139026e-06, 'samples': 24340480, 'steps': 47539, 'loss/train': 0.431205689907074} 02/26/2022 15:09:02 - INFO - codeparrot_training - Step 47540: {'lr': 3.233397555408607e-06, 'samples': 24340992, 'steps': 47540, 'loss/train': 7.405642509460449} 02/26/2022 15:09:08 - INFO - codeparrot_training - Step 47541: {'lr': 3.230774986747953e-06, 'samples': 24341504, 'steps': 47541, 'loss/train': 1.1553869247436523} 02/26/2022 15:09:12 - INFO - codeparrot_training - Step 47542: {'lr': 3.228153475168305e-06, 'samples': 24342016, 'steps': 47542, 'loss/train': 2.4501562118530273} 02/26/2022 15:09:17 - INFO - codeparrot_training - Step 47543: {'lr': 3.225533020680904e-06, 'samples': 24342528, 'steps': 47543, 'loss/train': 2.49342679977417} 02/26/2022 15:09:21 - INFO - codeparrot_training - Step 47544: {'lr': 3.222913623296936e-06, 'samples': 24343040, 'steps': 47544, 'loss/train': 2.1924617290496826} 02/26/2022 15:09:27 - INFO - codeparrot_training - Step 47545: {'lr': 3.220295283027641e-06, 'samples': 24343552, 'steps': 47545, 'loss/train': 1.7320542335510254} 02/26/2022 15:09:30 - INFO - codeparrot_training - Step 47546: {'lr': 3.217677999884261e-06, 'samples': 24344064, 'steps': 47546, 'loss/train': 2.0045504570007324} 02/26/2022 15:09:36 - INFO - codeparrot_training - Step 47547: {'lr': 3.215061773877953e-06, 'samples': 24344576, 'steps': 47547, 'loss/train': 1.5592222213745117} 02/26/2022 15:09:40 - INFO - codeparrot_training - Step 47548: {'lr': 3.212446605020014e-06, 'samples': 24345088, 'steps': 47548, 'loss/train': 1.1184674501419067} 02/26/2022 15:09:43 - INFO - codeparrot_training - Step 47549: {'lr': 3.2098324933215182e-06, 'samples': 24345600, 'steps': 47549, 'loss/train': 1.0899333953857422} 02/26/2022 15:09:49 - INFO - codeparrot_training - Step 47550: {'lr': 3.20721943879379e-06, 'samples': 24346112, 'steps': 47550, 'loss/train': 1.733041524887085} 02/26/2022 15:09:52 - INFO - codeparrot_training - Step 47551: {'lr': 3.204607441447932e-06, 'samples': 24346624, 'steps': 47551, 'loss/train': 1.2579232454299927} 02/26/2022 15:09:58 - INFO - codeparrot_training - Step 47552: {'lr': 3.2019965012952125e-06, 'samples': 24347136, 'steps': 47552, 'loss/train': 0.6781299710273743} 02/26/2022 15:10:01 - INFO - codeparrot_training - Step 47553: {'lr': 3.1993866183467335e-06, 'samples': 24347648, 'steps': 47553, 'loss/train': 0.7907813191413879} 02/26/2022 15:10:07 - INFO - codeparrot_training - Step 47554: {'lr': 3.1967777926137363e-06, 'samples': 24348160, 'steps': 47554, 'loss/train': 0.8809179067611694} 02/26/2022 15:10:10 - INFO - codeparrot_training - Step 47555: {'lr': 3.194170024107351e-06, 'samples': 24348672, 'steps': 47555, 'loss/train': 1.2193326950073242} 02/26/2022 15:10:16 - INFO - codeparrot_training - Step 47556: {'lr': 3.191563312838819e-06, 'samples': 24349184, 'steps': 47556, 'loss/train': 1.982816457748413} 02/26/2022 15:10:19 - INFO - codeparrot_training - Step 47557: {'lr': 3.1889576588192136e-06, 'samples': 24349696, 'steps': 47557, 'loss/train': 0.85979163646698} 02/26/2022 15:10:25 - INFO - codeparrot_training - Step 47558: {'lr': 3.186353062059749e-06, 'samples': 24350208, 'steps': 47558, 'loss/train': 1.4743715524673462} 02/26/2022 15:10:28 - INFO - codeparrot_training - Step 47559: {'lr': 3.183749522571583e-06, 'samples': 24350720, 'steps': 47559, 'loss/train': 1.5385243892669678} 02/26/2022 15:10:34 - INFO - codeparrot_training - Step 47560: {'lr': 3.181147040365873e-06, 'samples': 24351232, 'steps': 47560, 'loss/train': 1.7087470293045044} 02/26/2022 15:10:38 - INFO - codeparrot_training - Step 47561: {'lr': 3.1785456154537485e-06, 'samples': 24351744, 'steps': 47561, 'loss/train': 1.294171929359436} 02/26/2022 15:10:43 - INFO - codeparrot_training - Step 47562: {'lr': 3.1759452478463404e-06, 'samples': 24352256, 'steps': 47562, 'loss/train': 1.5617821216583252} 02/26/2022 15:10:47 - INFO - codeparrot_training - Step 47563: {'lr': 3.1733459375548335e-06, 'samples': 24352768, 'steps': 47563, 'loss/train': 2.0581600666046143} 02/26/2022 15:10:52 - INFO - codeparrot_training - Step 47564: {'lr': 3.1707476845903025e-06, 'samples': 24353280, 'steps': 47564, 'loss/train': 1.7121607065200806} 02/26/2022 15:11:00 - INFO - codeparrot_training - Step 47565: {'lr': 3.168150488963961e-06, 'samples': 24353792, 'steps': 47565, 'loss/train': 8.073162078857422} 02/26/2022 15:11:03 - INFO - codeparrot_training - Step 47566: {'lr': 3.1655543506868557e-06, 'samples': 24354304, 'steps': 47566, 'loss/train': 1.8005565404891968} 02/26/2022 15:11:09 - INFO - codeparrot_training - Step 47567: {'lr': 3.1629592697701435e-06, 'samples': 24354816, 'steps': 47567, 'loss/train': 0.9688205122947693} 02/26/2022 15:11:12 - INFO - codeparrot_training - Step 47568: {'lr': 3.1603652462249e-06, 'samples': 24355328, 'steps': 47568, 'loss/train': 1.6241395473480225} 02/26/2022 15:11:18 - INFO - codeparrot_training - Step 47569: {'lr': 3.1577722800623098e-06, 'samples': 24355840, 'steps': 47569, 'loss/train': 2.0020203590393066} 02/26/2022 15:11:21 - INFO - codeparrot_training - Step 47570: {'lr': 3.1551803712934477e-06, 'samples': 24356352, 'steps': 47570, 'loss/train': 1.3297297954559326} 02/26/2022 15:11:27 - INFO - codeparrot_training - Step 47571: {'lr': 3.1525895199293886e-06, 'samples': 24356864, 'steps': 47571, 'loss/train': 1.98710298538208} 02/26/2022 15:11:30 - INFO - codeparrot_training - Step 47572: {'lr': 3.149999725981262e-06, 'samples': 24357376, 'steps': 47572, 'loss/train': 1.347033977508545} 02/26/2022 15:11:36 - INFO - codeparrot_training - Step 47573: {'lr': 3.1474109894601423e-06, 'samples': 24357888, 'steps': 47573, 'loss/train': 1.4820574522018433} 02/26/2022 15:11:39 - INFO - codeparrot_training - Step 47574: {'lr': 3.1448233103771596e-06, 'samples': 24358400, 'steps': 47574, 'loss/train': 1.6216740608215332} 02/26/2022 15:11:47 - INFO - codeparrot_training - Step 47575: {'lr': 3.142236688743333e-06, 'samples': 24358912, 'steps': 47575, 'loss/train': 2.0336694717407227} 02/26/2022 15:11:50 - INFO - codeparrot_training - Step 47576: {'lr': 3.1396511245697922e-06, 'samples': 24359424, 'steps': 47576, 'loss/train': 1.9226903915405273} 02/26/2022 15:11:55 - INFO - codeparrot_training - Step 47577: {'lr': 3.137066617867612e-06, 'samples': 24359936, 'steps': 47577, 'loss/train': 0.7686557173728943} 02/26/2022 15:11:59 - INFO - codeparrot_training - Step 47578: {'lr': 3.1344831686478393e-06, 'samples': 24360448, 'steps': 47578, 'loss/train': 1.770987629890442} 02/26/2022 15:12:04 - INFO - codeparrot_training - Step 47579: {'lr': 3.131900776921548e-06, 'samples': 24360960, 'steps': 47579, 'loss/train': 1.230187177658081} 02/26/2022 15:12:08 - INFO - codeparrot_training - Step 47580: {'lr': 3.129319442699785e-06, 'samples': 24361472, 'steps': 47580, 'loss/train': 1.4721728563308716} 02/26/2022 15:12:13 - INFO - codeparrot_training - Step 47581: {'lr': 3.126739165993653e-06, 'samples': 24361984, 'steps': 47581, 'loss/train': 1.4379863739013672} 02/26/2022 15:12:17 - INFO - codeparrot_training - Step 47582: {'lr': 3.12415994681417e-06, 'samples': 24362496, 'steps': 47582, 'loss/train': 3.3473212718963623} 02/26/2022 15:12:22 - INFO - codeparrot_training - Step 47583: {'lr': 3.121581785172439e-06, 'samples': 24363008, 'steps': 47583, 'loss/train': 1.0834342241287231} 02/26/2022 15:12:26 - INFO - codeparrot_training - Step 47584: {'lr': 3.119004681079396e-06, 'samples': 24363520, 'steps': 47584, 'loss/train': 1.609164834022522} 02/26/2022 15:12:33 - INFO - codeparrot_training - Step 47585: {'lr': 3.11642863454617e-06, 'samples': 24364032, 'steps': 47585, 'loss/train': 1.1637941598892212} 02/26/2022 15:12:36 - INFO - codeparrot_training - Step 47586: {'lr': 3.113853645583753e-06, 'samples': 24364544, 'steps': 47586, 'loss/train': 0.15011459589004517} 02/26/2022 15:12:42 - INFO - codeparrot_training - Step 47587: {'lr': 3.111279714203219e-06, 'samples': 24365056, 'steps': 47587, 'loss/train': 1.823533058166504} 02/26/2022 15:12:45 - INFO - codeparrot_training - Step 47588: {'lr': 3.1087068404155593e-06, 'samples': 24365568, 'steps': 47588, 'loss/train': 0.8395389318466187} 02/26/2022 15:12:51 - INFO - codeparrot_training - Step 47589: {'lr': 3.1061350242317933e-06, 'samples': 24366080, 'steps': 47589, 'loss/train': 2.761305093765259} 02/26/2022 15:12:54 - INFO - codeparrot_training - Step 47590: {'lr': 3.1035642656629393e-06, 'samples': 24366592, 'steps': 47590, 'loss/train': 2.197582721710205} 02/26/2022 15:13:00 - INFO - codeparrot_training - Step 47591: {'lr': 3.1009945647200443e-06, 'samples': 24367104, 'steps': 47591, 'loss/train': 1.3107774257659912} 02/26/2022 15:13:03 - INFO - codeparrot_training - Step 47592: {'lr': 3.098425921414072e-06, 'samples': 24367616, 'steps': 47592, 'loss/train': 2.3657925128936768} 02/26/2022 15:13:09 - INFO - codeparrot_training - Step 47593: {'lr': 3.095858335756041e-06, 'samples': 24368128, 'steps': 47593, 'loss/train': 1.4001954793930054} 02/26/2022 15:13:12 - INFO - codeparrot_training - Step 47594: {'lr': 3.0932918077569427e-06, 'samples': 24368640, 'steps': 47594, 'loss/train': 1.7272108793258667} 02/26/2022 15:13:18 - INFO - codeparrot_training - Step 47595: {'lr': 3.090726337427796e-06, 'samples': 24369152, 'steps': 47595, 'loss/train': 2.380660057067871} 02/26/2022 15:13:21 - INFO - codeparrot_training - Step 47596: {'lr': 3.0881619247795645e-06, 'samples': 24369664, 'steps': 47596, 'loss/train': 0.865058183670044} 02/26/2022 15:13:29 - INFO - codeparrot_training - Step 47597: {'lr': 3.085598569823267e-06, 'samples': 24370176, 'steps': 47597, 'loss/train': 1.7573013305664062} 02/26/2022 15:13:32 - INFO - codeparrot_training - Step 47598: {'lr': 3.083036272569839e-06, 'samples': 24370688, 'steps': 47598, 'loss/train': 1.7063276767730713} 02/26/2022 15:13:37 - INFO - codeparrot_training - Step 47599: {'lr': 3.080475033030272e-06, 'samples': 24371200, 'steps': 47599, 'loss/train': 1.7071788311004639} 02/26/2022 15:13:41 - INFO - codeparrot_training - Step 47600: {'lr': 3.077914851215585e-06, 'samples': 24371712, 'steps': 47600, 'loss/train': 1.8461300134658813} 02/26/2022 15:13:46 - INFO - codeparrot_training - Step 47601: {'lr': 3.0753557271366584e-06, 'samples': 24372224, 'steps': 47601, 'loss/train': 1.123298168182373} 02/26/2022 15:13:50 - INFO - codeparrot_training - Step 47602: {'lr': 3.0727976608045106e-06, 'samples': 24372736, 'steps': 47602, 'loss/train': 1.5736615657806396} 02/26/2022 15:13:55 - INFO - codeparrot_training - Step 47603: {'lr': 3.0702406522301053e-06, 'samples': 24373248, 'steps': 47603, 'loss/train': 2.1834020614624023} 02/26/2022 15:13:59 - INFO - codeparrot_training - Step 47604: {'lr': 3.067684701424378e-06, 'samples': 24373760, 'steps': 47604, 'loss/train': 1.6734988689422607} 02/26/2022 15:14:04 - INFO - codeparrot_training - Step 47605: {'lr': 3.0651298083982925e-06, 'samples': 24374272, 'steps': 47605, 'loss/train': 2.1561081409454346} 02/26/2022 15:14:08 - INFO - codeparrot_training - Step 47606: {'lr': 3.062575973162729e-06, 'samples': 24374784, 'steps': 47606, 'loss/train': 1.7029731273651123} 02/26/2022 15:14:13 - INFO - codeparrot_training - Step 47607: {'lr': 3.060023195728734e-06, 'samples': 24375296, 'steps': 47607, 'loss/train': 2.3150179386138916} 02/26/2022 15:14:17 - INFO - codeparrot_training - Step 47608: {'lr': 3.0574714761071596e-06, 'samples': 24375808, 'steps': 47608, 'loss/train': 1.840514898300171} 02/26/2022 15:14:22 - INFO - codeparrot_training - Step 47609: {'lr': 3.05492081430897e-06, 'samples': 24376320, 'steps': 47609, 'loss/train': 1.8019163608551025} 02/26/2022 15:14:26 - INFO - codeparrot_training - Step 47610: {'lr': 3.052371210345073e-06, 'samples': 24376832, 'steps': 47610, 'loss/train': 0.302370548248291} 02/26/2022 15:14:33 - INFO - codeparrot_training - Step 47611: {'lr': 3.0498226642264314e-06, 'samples': 24377344, 'steps': 47611, 'loss/train': 1.8133140802383423} 02/26/2022 15:14:36 - INFO - codeparrot_training - Step 47612: {'lr': 3.0472751759639263e-06, 'samples': 24377856, 'steps': 47612, 'loss/train': 0.9807099103927612} 02/26/2022 15:14:42 - INFO - codeparrot_training - Step 47613: {'lr': 3.044728745568465e-06, 'samples': 24378368, 'steps': 47613, 'loss/train': 1.9523282051086426} 02/26/2022 15:14:48 - INFO - codeparrot_training - Step 47614: {'lr': 3.042183373050983e-06, 'samples': 24378880, 'steps': 47614, 'loss/train': 2.3784940242767334} 02/26/2022 15:14:51 - INFO - codeparrot_training - Step 47615: {'lr': 3.0396390584223333e-06, 'samples': 24379392, 'steps': 47615, 'loss/train': 0.07598035782575607} 02/26/2022 15:14:57 - INFO - codeparrot_training - Step 47616: {'lr': 3.037095801693479e-06, 'samples': 24379904, 'steps': 47616, 'loss/train': 1.2915031909942627} 02/26/2022 15:15:00 - INFO - codeparrot_training - Step 47617: {'lr': 3.0345536028752728e-06, 'samples': 24380416, 'steps': 47617, 'loss/train': 1.849656105041504} 02/26/2022 15:15:05 - INFO - codeparrot_training - Step 47618: {'lr': 3.0320124619786506e-06, 'samples': 24380928, 'steps': 47618, 'loss/train': 2.1262218952178955} 02/26/2022 15:15:09 - INFO - codeparrot_training - Step 47619: {'lr': 3.0294723790144362e-06, 'samples': 24381440, 'steps': 47619, 'loss/train': 1.899937391281128} 02/26/2022 15:15:16 - INFO - codeparrot_training - Step 47620: {'lr': 3.026933353993566e-06, 'samples': 24381952, 'steps': 47620, 'loss/train': 1.5570577383041382} 02/26/2022 15:15:20 - INFO - codeparrot_training - Step 47621: {'lr': 3.0243953869268647e-06, 'samples': 24382464, 'steps': 47621, 'loss/train': 1.6037689447402954} 02/26/2022 15:15:25 - INFO - codeparrot_training - Step 47622: {'lr': 3.02185847782524e-06, 'samples': 24382976, 'steps': 47622, 'loss/train': 1.9914475679397583} 02/26/2022 15:15:29 - INFO - codeparrot_training - Step 47623: {'lr': 3.0193226266995444e-06, 'samples': 24383488, 'steps': 47623, 'loss/train': 2.3011107444763184} 02/26/2022 15:15:34 - INFO - codeparrot_training - Step 47624: {'lr': 3.016787833560658e-06, 'samples': 24384000, 'steps': 47624, 'loss/train': 0.1855044662952423} 02/26/2022 15:15:37 - INFO - codeparrot_training - Step 47625: {'lr': 3.0142540984194056e-06, 'samples': 24384512, 'steps': 47625, 'loss/train': 1.5040817260742188} 02/26/2022 15:15:43 - INFO - codeparrot_training - Step 47626: {'lr': 3.011721421286695e-06, 'samples': 24385024, 'steps': 47626, 'loss/train': 1.5098698139190674} 02/26/2022 15:15:46 - INFO - codeparrot_training - Step 47627: {'lr': 3.0091898021733233e-06, 'samples': 24385536, 'steps': 47627, 'loss/train': 1.8019421100616455} 02/26/2022 15:15:52 - INFO - codeparrot_training - Step 47628: {'lr': 3.0066592410901427e-06, 'samples': 24386048, 'steps': 47628, 'loss/train': 3.0033986568450928} 02/26/2022 15:15:55 - INFO - codeparrot_training - Step 47629: {'lr': 3.004129738048006e-06, 'samples': 24386560, 'steps': 47629, 'loss/train': 1.6713452339172363} 02/26/2022 15:16:01 - INFO - codeparrot_training - Step 47630: {'lr': 3.0016012930577652e-06, 'samples': 24387072, 'steps': 47630, 'loss/train': 1.8882120847702026} 02/26/2022 15:16:05 - INFO - codeparrot_training - Step 47631: {'lr': 2.9990739061302176e-06, 'samples': 24387584, 'steps': 47631, 'loss/train': 1.9239529371261597} 02/26/2022 15:16:12 - INFO - codeparrot_training - Step 47632: {'lr': 2.996547577276215e-06, 'samples': 24388096, 'steps': 47632, 'loss/train': 1.236438274383545} 02/26/2022 15:16:15 - INFO - codeparrot_training - Step 47633: {'lr': 2.994022306506555e-06, 'samples': 24388608, 'steps': 47633, 'loss/train': 1.224778175354004} 02/26/2022 15:16:21 - INFO - codeparrot_training - Step 47634: {'lr': 2.9914980938320903e-06, 'samples': 24389120, 'steps': 47634, 'loss/train': 2.2543954849243164} 02/26/2022 15:16:25 - INFO - codeparrot_training - Step 47635: {'lr': 2.988974939263617e-06, 'samples': 24389632, 'steps': 47635, 'loss/train': 2.205904006958008} 02/26/2022 15:16:30 - INFO - codeparrot_training - Step 47636: {'lr': 2.986452842811932e-06, 'samples': 24390144, 'steps': 47636, 'loss/train': 0.9505888223648071} 02/26/2022 15:16:34 - INFO - codeparrot_training - Step 47637: {'lr': 2.9839318044878326e-06, 'samples': 24390656, 'steps': 47637, 'loss/train': 1.8812248706817627} 02/26/2022 15:16:37 - INFO - codeparrot_training - Step 47638: {'lr': 2.981411824302144e-06, 'samples': 24391168, 'steps': 47638, 'loss/train': 2.115784168243408} 02/26/2022 15:16:43 - INFO - codeparrot_training - Step 47639: {'lr': 2.978892902265662e-06, 'samples': 24391680, 'steps': 47639, 'loss/train': 1.4282594919204712} 02/26/2022 15:16:46 - INFO - codeparrot_training - Step 47640: {'lr': 2.9763750383891565e-06, 'samples': 24392192, 'steps': 47640, 'loss/train': 0.3510097861289978} 02/26/2022 15:16:52 - INFO - codeparrot_training - Step 47641: {'lr': 2.973858232683424e-06, 'samples': 24392704, 'steps': 47641, 'loss/train': 2.1685428619384766} 02/26/2022 15:16:59 - INFO - codeparrot_training - Step 47642: {'lr': 2.9713424851592342e-06, 'samples': 24393216, 'steps': 47642, 'loss/train': 1.6736844778060913} 02/26/2022 15:17:02 - INFO - codeparrot_training - Step 47643: {'lr': 2.968827795827356e-06, 'samples': 24393728, 'steps': 47643, 'loss/train': 1.7606958150863647} 02/26/2022 15:17:08 - INFO - codeparrot_training - Step 47644: {'lr': 2.9663141646986137e-06, 'samples': 24394240, 'steps': 47644, 'loss/train': 0.2881993353366852} 02/26/2022 15:17:11 - INFO - codeparrot_training - Step 47645: {'lr': 2.9638015917837213e-06, 'samples': 24394752, 'steps': 47645, 'loss/train': 1.7967191934585571} 02/26/2022 15:17:17 - INFO - codeparrot_training - Step 47646: {'lr': 2.961290077093476e-06, 'samples': 24395264, 'steps': 47646, 'loss/train': 1.323774814605713} 02/26/2022 15:17:20 - INFO - codeparrot_training - Step 47647: {'lr': 2.9587796206386185e-06, 'samples': 24395776, 'steps': 47647, 'loss/train': 1.4970837831497192} 02/26/2022 15:17:26 - INFO - codeparrot_training - Step 47648: {'lr': 2.956270222429891e-06, 'samples': 24396288, 'steps': 47648, 'loss/train': 0.6243499517440796} 02/26/2022 15:17:29 - INFO - codeparrot_training - Step 47649: {'lr': 2.9537618824780345e-06, 'samples': 24396800, 'steps': 47649, 'loss/train': 1.6987028121948242} 02/26/2022 15:17:35 - INFO - codeparrot_training - Step 47650: {'lr': 2.9512546007938457e-06, 'samples': 24397312, 'steps': 47650, 'loss/train': 1.4116984605789185} 02/26/2022 15:17:38 - INFO - codeparrot_training - Step 47651: {'lr': 2.948748377388039e-06, 'samples': 24397824, 'steps': 47651, 'loss/train': 0.46655547618865967} 02/26/2022 15:17:46 - INFO - codeparrot_training - Step 47652: {'lr': 2.9462432122713545e-06, 'samples': 24398336, 'steps': 47652, 'loss/train': 1.8256189823150635} 02/26/2022 15:17:49 - INFO - codeparrot_training - Step 47653: {'lr': 2.9437391054544795e-06, 'samples': 24398848, 'steps': 47653, 'loss/train': 2.047633647918701} 02/26/2022 15:17:55 - INFO - codeparrot_training - Step 47654: {'lr': 2.94123605694821e-06, 'samples': 24399360, 'steps': 47654, 'loss/train': 0.8964255452156067} 02/26/2022 15:17:58 - INFO - codeparrot_training - Step 47655: {'lr': 2.938734066763232e-06, 'samples': 24399872, 'steps': 47655, 'loss/train': 2.083402395248413} 02/26/2022 15:18:04 - INFO - codeparrot_training - Step 47656: {'lr': 2.9362331349102323e-06, 'samples': 24400384, 'steps': 47656, 'loss/train': 1.3712934255599976} 02/26/2022 15:18:07 - INFO - codeparrot_training - Step 47657: {'lr': 2.933733261399979e-06, 'samples': 24400896, 'steps': 47657, 'loss/train': 2.032747507095337} 02/26/2022 15:18:13 - INFO - codeparrot_training - Step 47658: {'lr': 2.9312344462431863e-06, 'samples': 24401408, 'steps': 47658, 'loss/train': 0.589367151260376} 02/26/2022 15:18:16 - INFO - codeparrot_training - Step 47659: {'lr': 2.928736689450512e-06, 'samples': 24401920, 'steps': 47659, 'loss/train': 2.638061761856079} 02/26/2022 15:18:22 - INFO - codeparrot_training - Step 47660: {'lr': 2.92623999103267e-06, 'samples': 24402432, 'steps': 47660, 'loss/train': 1.7041376829147339} 02/26/2022 15:18:25 - INFO - codeparrot_training - Step 47661: {'lr': 2.9237443510003737e-06, 'samples': 24402944, 'steps': 47661, 'loss/train': 1.6518422365188599} 02/26/2022 15:18:31 - INFO - codeparrot_training - Step 47662: {'lr': 2.9212497693642814e-06, 'samples': 24403456, 'steps': 47662, 'loss/train': 0.4850243628025055} 02/26/2022 15:18:34 - INFO - codeparrot_training - Step 47663: {'lr': 2.918756246135107e-06, 'samples': 24403968, 'steps': 47663, 'loss/train': 1.7888721227645874} 02/26/2022 15:18:40 - INFO - codeparrot_training - Step 47664: {'lr': 2.9162637813235083e-06, 'samples': 24404480, 'steps': 47664, 'loss/train': 0.19029074907302856} 02/26/2022 15:18:43 - INFO - codeparrot_training - Step 47665: {'lr': 2.913772374940227e-06, 'samples': 24404992, 'steps': 47665, 'loss/train': 2.127103328704834} 02/26/2022 15:18:50 - INFO - codeparrot_training - Step 47666: {'lr': 2.9112820269958374e-06, 'samples': 24405504, 'steps': 47666, 'loss/train': 1.6891711950302124} 02/26/2022 15:18:54 - INFO - codeparrot_training - Step 47667: {'lr': 2.9087927375010816e-06, 'samples': 24406016, 'steps': 47667, 'loss/train': 2.296438217163086} 02/26/2022 15:19:00 - INFO - codeparrot_training - Step 47668: {'lr': 2.9063045064665894e-06, 'samples': 24406528, 'steps': 47668, 'loss/train': 1.5600039958953857} 02/26/2022 15:19:03 - INFO - codeparrot_training - Step 47669: {'lr': 2.9038173339030194e-06, 'samples': 24407040, 'steps': 47669, 'loss/train': 1.7696064710617065} 02/26/2022 15:19:07 - INFO - codeparrot_training - Step 47670: {'lr': 2.9013312198210573e-06, 'samples': 24407552, 'steps': 47670, 'loss/train': 2.2763867378234863} 02/26/2022 15:19:12 - INFO - codeparrot_training - Step 47671: {'lr': 2.898846164231306e-06, 'samples': 24408064, 'steps': 47671, 'loss/train': 1.164310336112976} 02/26/2022 15:19:16 - INFO - codeparrot_training - Step 47672: {'lr': 2.896362167144423e-06, 'samples': 24408576, 'steps': 47672, 'loss/train': 1.4535681009292603} 02/26/2022 15:19:21 - INFO - codeparrot_training - Step 47673: {'lr': 2.8938792285710946e-06, 'samples': 24409088, 'steps': 47673, 'loss/train': 1.094604253768921} 02/26/2022 15:19:25 - INFO - codeparrot_training - Step 47674: {'lr': 2.8913973485219236e-06, 'samples': 24409600, 'steps': 47674, 'loss/train': 0.3681040406227112} 02/26/2022 15:19:30 - INFO - codeparrot_training - Step 47675: {'lr': 2.8889165270075122e-06, 'samples': 24410112, 'steps': 47675, 'loss/train': 2.456069231033325} 02/26/2022 15:19:34 - INFO - codeparrot_training - Step 47676: {'lr': 2.8864367640385192e-06, 'samples': 24410624, 'steps': 47676, 'loss/train': 2.094545364379883} 02/26/2022 15:19:39 - INFO - codeparrot_training - Step 47677: {'lr': 2.8839580596255743e-06, 'samples': 24411136, 'steps': 47677, 'loss/train': 0.7565767765045166} 02/26/2022 15:19:43 - INFO - codeparrot_training - Step 47678: {'lr': 2.8814804137792803e-06, 'samples': 24411648, 'steps': 47678, 'loss/train': 1.3240689039230347} 02/26/2022 15:19:50 - INFO - codeparrot_training - Step 47679: {'lr': 2.879003826510268e-06, 'samples': 24412160, 'steps': 47679, 'loss/train': 1.258461356163025} 02/26/2022 15:19:54 - INFO - codeparrot_training - Step 47680: {'lr': 2.876528297829112e-06, 'samples': 24412672, 'steps': 47680, 'loss/train': 1.2884104251861572} 02/26/2022 15:19:59 - INFO - codeparrot_training - Step 47681: {'lr': 2.87405382774647e-06, 'samples': 24413184, 'steps': 47681, 'loss/train': 2.0182693004608154} 02/26/2022 15:20:03 - INFO - codeparrot_training - Step 47682: {'lr': 2.871580416272862e-06, 'samples': 24413696, 'steps': 47682, 'loss/train': 1.2920496463775635} 02/26/2022 15:20:08 - INFO - codeparrot_training - Step 47683: {'lr': 2.8691080634190013e-06, 'samples': 24414208, 'steps': 47683, 'loss/train': 1.6101053953170776} 02/26/2022 15:20:12 - INFO - codeparrot_training - Step 47684: {'lr': 2.866636769195352e-06, 'samples': 24414720, 'steps': 47684, 'loss/train': 0.8663586974143982} 02/26/2022 15:20:17 - INFO - codeparrot_training - Step 47685: {'lr': 2.864166533612572e-06, 'samples': 24415232, 'steps': 47685, 'loss/train': 2.719475269317627} 02/26/2022 15:20:21 - INFO - codeparrot_training - Step 47686: {'lr': 2.8616973566812087e-06, 'samples': 24415744, 'steps': 47686, 'loss/train': 2.119499921798706} 02/26/2022 15:20:26 - INFO - codeparrot_training - Step 47687: {'lr': 2.8592292384118923e-06, 'samples': 24416256, 'steps': 47687, 'loss/train': 1.5559251308441162} 02/26/2022 15:20:30 - INFO - codeparrot_training - Step 47688: {'lr': 2.8567621788151423e-06, 'samples': 24416768, 'steps': 47688, 'loss/train': 1.6287193298339844} 02/26/2022 15:20:37 - INFO - codeparrot_training - Step 47689: {'lr': 2.8542961779015606e-06, 'samples': 24417280, 'steps': 47689, 'loss/train': 1.0163534879684448} 02/26/2022 15:20:40 - INFO - codeparrot_training - Step 47690: {'lr': 2.851831235681668e-06, 'samples': 24417792, 'steps': 47690, 'loss/train': 1.458247423171997} 02/26/2022 15:20:46 - INFO - codeparrot_training - Step 47691: {'lr': 2.849367352166066e-06, 'samples': 24418304, 'steps': 47691, 'loss/train': 2.1798388957977295} 02/26/2022 15:20:49 - INFO - codeparrot_training - Step 47692: {'lr': 2.8469045273653294e-06, 'samples': 24418816, 'steps': 47692, 'loss/train': 1.5786243677139282} 02/26/2022 15:20:55 - INFO - codeparrot_training - Step 47693: {'lr': 2.8444427612899502e-06, 'samples': 24419328, 'steps': 47693, 'loss/train': 1.2595247030258179} 02/26/2022 15:20:58 - INFO - codeparrot_training - Step 47694: {'lr': 2.841982053950476e-06, 'samples': 24419840, 'steps': 47694, 'loss/train': 1.4161241054534912} 02/26/2022 15:21:04 - INFO - codeparrot_training - Step 47695: {'lr': 2.839522405357481e-06, 'samples': 24420352, 'steps': 47695, 'loss/train': 1.810314416885376} 02/26/2022 15:21:07 - INFO - codeparrot_training - Step 47696: {'lr': 2.8370638155215123e-06, 'samples': 24420864, 'steps': 47696, 'loss/train': 1.4527337551116943} 02/26/2022 15:21:13 - INFO - codeparrot_training - Step 47697: {'lr': 2.8346062844530617e-06, 'samples': 24421376, 'steps': 47697, 'loss/train': 1.8947196006774902} 02/26/2022 15:21:16 - INFO - codeparrot_training - Step 47698: {'lr': 2.8321498121626767e-06, 'samples': 24421888, 'steps': 47698, 'loss/train': 1.9455469846725464} 02/26/2022 15:21:24 - INFO - codeparrot_training - Step 47699: {'lr': 2.829694398660848e-06, 'samples': 24422400, 'steps': 47699, 'loss/train': 0.8469429016113281} 02/26/2022 15:21:27 - INFO - codeparrot_training - Step 47700: {'lr': 2.8272400439581513e-06, 'samples': 24422912, 'steps': 47700, 'loss/train': 0.3897291421890259} 02/26/2022 15:21:32 - INFO - codeparrot_training - Step 47701: {'lr': 2.824786748065078e-06, 'samples': 24423424, 'steps': 47701, 'loss/train': 1.4133137464523315} 02/26/2022 15:21:36 - INFO - codeparrot_training - Step 47702: {'lr': 2.8223345109921196e-06, 'samples': 24423936, 'steps': 47702, 'loss/train': 2.38584566116333} 02/26/2022 15:21:41 - INFO - codeparrot_training - Step 47703: {'lr': 2.8198833327497952e-06, 'samples': 24424448, 'steps': 47703, 'loss/train': 0.43464207649230957} 02/26/2022 15:21:45 - INFO - codeparrot_training - Step 47704: {'lr': 2.817433213348597e-06, 'samples': 24424960, 'steps': 47704, 'loss/train': 0.126841738820076} 02/26/2022 15:21:51 - INFO - codeparrot_training - Step 47705: {'lr': 2.8149841527990437e-06, 'samples': 24425472, 'steps': 47705, 'loss/train': 0.4895975887775421} 02/26/2022 15:21:54 - INFO - codeparrot_training - Step 47706: {'lr': 2.8125361511115997e-06, 'samples': 24425984, 'steps': 47706, 'loss/train': 2.1269664764404297} 02/26/2022 15:21:59 - INFO - codeparrot_training - Step 47707: {'lr': 2.8100892082967565e-06, 'samples': 24426496, 'steps': 47707, 'loss/train': 1.6531054973602295} 02/26/2022 15:22:03 - INFO - codeparrot_training - Step 47708: {'lr': 2.8076433243650056e-06, 'samples': 24427008, 'steps': 47708, 'loss/train': 1.432506799697876} 02/26/2022 15:22:08 - INFO - codeparrot_training - Step 47709: {'lr': 2.8051984993268107e-06, 'samples': 24427520, 'steps': 47709, 'loss/train': 0.6510240435600281} 02/26/2022 15:22:12 - INFO - codeparrot_training - Step 47710: {'lr': 2.802754733192664e-06, 'samples': 24428032, 'steps': 47710, 'loss/train': 1.3855977058410645} 02/26/2022 15:22:17 - INFO - codeparrot_training - Step 47711: {'lr': 2.8003120259730007e-06, 'samples': 24428544, 'steps': 47711, 'loss/train': 2.165266752243042} 02/26/2022 15:22:21 - INFO - codeparrot_training - Step 47712: {'lr': 2.7978703776783132e-06, 'samples': 24429056, 'steps': 47712, 'loss/train': 0.7595996856689453} 02/26/2022 15:22:26 - INFO - codeparrot_training - Step 47713: {'lr': 2.795429788319065e-06, 'samples': 24429568, 'steps': 47713, 'loss/train': 1.2099123001098633} 02/26/2022 15:22:30 - INFO - codeparrot_training - Step 47714: {'lr': 2.792990257905692e-06, 'samples': 24430080, 'steps': 47714, 'loss/train': 8.540417671203613} 02/26/2022 15:22:37 - INFO - codeparrot_training - Step 47715: {'lr': 2.790551786448631e-06, 'samples': 24430592, 'steps': 47715, 'loss/train': 0.41386866569519043} 02/26/2022 15:22:41 - INFO - codeparrot_training - Step 47716: {'lr': 2.7881143739583724e-06, 'samples': 24431104, 'steps': 47716, 'loss/train': 2.180248975753784} 02/26/2022 15:22:46 - INFO - codeparrot_training - Step 47717: {'lr': 2.7856780204453256e-06, 'samples': 24431616, 'steps': 47717, 'loss/train': 1.8266843557357788} 02/26/2022 15:22:50 - INFO - codeparrot_training - Step 47718: {'lr': 2.7832427259199546e-06, 'samples': 24432128, 'steps': 47718, 'loss/train': 0.36090385913848877} 02/26/2022 15:22:56 - INFO - codeparrot_training - Step 47719: {'lr': 2.780808490392639e-06, 'samples': 24432640, 'steps': 47719, 'loss/train': 1.6662006378173828} 02/26/2022 15:22:59 - INFO - codeparrot_training - Step 47720: {'lr': 2.778375313873871e-06, 'samples': 24433152, 'steps': 47720, 'loss/train': 1.184025764465332} 02/26/2022 15:23:05 - INFO - codeparrot_training - Step 47721: {'lr': 2.7759431963740035e-06, 'samples': 24433664, 'steps': 47721, 'loss/train': 2.0826425552368164} 02/26/2022 15:23:08 - INFO - codeparrot_training - Step 47722: {'lr': 2.7735121379035276e-06, 'samples': 24434176, 'steps': 47722, 'loss/train': 2.0559120178222656} 02/26/2022 15:23:14 - INFO - codeparrot_training - Step 47723: {'lr': 2.771082138472797e-06, 'samples': 24434688, 'steps': 47723, 'loss/train': 1.30818510055542} 02/26/2022 15:23:17 - INFO - codeparrot_training - Step 47724: {'lr': 2.7686531980922746e-06, 'samples': 24435200, 'steps': 47724, 'loss/train': 1.9541465044021606} 02/26/2022 15:23:25 - INFO - codeparrot_training - Step 47725: {'lr': 2.766225316772314e-06, 'samples': 24435712, 'steps': 47725, 'loss/train': 1.82325279712677} 02/26/2022 15:23:28 - INFO - codeparrot_training - Step 47726: {'lr': 2.7637984945233507e-06, 'samples': 24436224, 'steps': 47726, 'loss/train': 1.9219880104064941} 02/26/2022 15:23:34 - INFO - codeparrot_training - Step 47727: {'lr': 2.761372731355766e-06, 'samples': 24436736, 'steps': 47727, 'loss/train': 1.6483488082885742} 02/26/2022 15:23:37 - INFO - codeparrot_training - Step 47728: {'lr': 2.7589480272799395e-06, 'samples': 24437248, 'steps': 47728, 'loss/train': 0.16203279793262482} 02/26/2022 15:23:43 - INFO - codeparrot_training - Step 47729: {'lr': 2.7565243823062803e-06, 'samples': 24437760, 'steps': 47729, 'loss/train': 0.5907553434371948} 02/26/2022 15:23:47 - INFO - codeparrot_training - Step 47730: {'lr': 2.7541017964451685e-06, 'samples': 24438272, 'steps': 47730, 'loss/train': 1.9620444774627686} 02/26/2022 15:23:52 - INFO - codeparrot_training - Step 47731: {'lr': 2.7516802697069855e-06, 'samples': 24438784, 'steps': 47731, 'loss/train': 1.0339369773864746} 02/26/2022 15:23:56 - INFO - codeparrot_training - Step 47732: {'lr': 2.7492598021020833e-06, 'samples': 24439296, 'steps': 47732, 'loss/train': 1.3877111673355103} 02/26/2022 15:24:01 - INFO - codeparrot_training - Step 47733: {'lr': 2.7468403936408147e-06, 'samples': 24439808, 'steps': 47733, 'loss/train': 1.1138633489608765} 02/26/2022 15:24:05 - INFO - codeparrot_training - Step 47734: {'lr': 2.7444220443336165e-06, 'samples': 24440320, 'steps': 47734, 'loss/train': 1.8677597045898438} 02/26/2022 15:24:12 - INFO - codeparrot_training - Step 47735: {'lr': 2.7420047541907576e-06, 'samples': 24440832, 'steps': 47735, 'loss/train': 1.9140660762786865} 02/26/2022 15:24:16 - INFO - codeparrot_training - Step 47736: {'lr': 2.7395885232226748e-06, 'samples': 24441344, 'steps': 47736, 'loss/train': 1.9463571310043335} 02/26/2022 15:24:21 - INFO - codeparrot_training - Step 47737: {'lr': 2.7371733514396645e-06, 'samples': 24441856, 'steps': 47737, 'loss/train': 1.8360120058059692} 02/26/2022 15:24:24 - INFO - codeparrot_training - Step 47738: {'lr': 2.73475923885208e-06, 'samples': 24442368, 'steps': 47738, 'loss/train': 2.069361686706543} 02/26/2022 15:24:30 - INFO - codeparrot_training - Step 47739: {'lr': 2.7323461854702746e-06, 'samples': 24442880, 'steps': 47739, 'loss/train': 1.7338221073150635} 02/26/2022 15:24:34 - INFO - codeparrot_training - Step 47740: {'lr': 2.7299341913046006e-06, 'samples': 24443392, 'steps': 47740, 'loss/train': 2.8587684631347656} 02/26/2022 15:24:39 - INFO - codeparrot_training - Step 47741: {'lr': 2.727523256365355e-06, 'samples': 24443904, 'steps': 47741, 'loss/train': 2.726759433746338} 02/26/2022 15:24:43 - INFO - codeparrot_training - Step 47742: {'lr': 2.7251133806628916e-06, 'samples': 24444416, 'steps': 47742, 'loss/train': 2.069056510925293} 02/26/2022 15:24:48 - INFO - codeparrot_training - Step 47743: {'lr': 2.7227045642075064e-06, 'samples': 24444928, 'steps': 47743, 'loss/train': 0.9621003866195679} 02/26/2022 15:24:52 - INFO - codeparrot_training - Step 47744: {'lr': 2.7202968070095537e-06, 'samples': 24445440, 'steps': 47744, 'loss/train': 2.2205052375793457} 02/26/2022 15:24:59 - INFO - codeparrot_training - Step 47745: {'lr': 2.7178901090793296e-06, 'samples': 24445952, 'steps': 47745, 'loss/train': 1.2422583103179932} 02/26/2022 15:25:03 - INFO - codeparrot_training - Step 47746: {'lr': 2.715484470427132e-06, 'samples': 24446464, 'steps': 47746, 'loss/train': 2.436001777648926} 02/26/2022 15:25:08 - INFO - codeparrot_training - Step 47747: {'lr': 2.713079891063286e-06, 'samples': 24446976, 'steps': 47747, 'loss/train': 1.2731224298477173} 02/26/2022 15:25:12 - INFO - codeparrot_training - Step 47748: {'lr': 2.710676370998061e-06, 'samples': 24447488, 'steps': 47748, 'loss/train': 1.7909307479858398} 02/26/2022 15:25:17 - INFO - codeparrot_training - Step 47749: {'lr': 2.7082739102418097e-06, 'samples': 24448000, 'steps': 47749, 'loss/train': 3.53122878074646} 02/26/2022 15:25:21 - INFO - codeparrot_training - Step 47750: {'lr': 2.7058725088047465e-06, 'samples': 24448512, 'steps': 47750, 'loss/train': 0.3870745003223419} 02/26/2022 15:25:26 - INFO - codeparrot_training - Step 47751: {'lr': 2.7034721666972518e-06, 'samples': 24449024, 'steps': 47751, 'loss/train': 2.2625205516815186} 02/26/2022 15:25:30 - INFO - codeparrot_training - Step 47752: {'lr': 2.7010728839295118e-06, 'samples': 24449536, 'steps': 47752, 'loss/train': 1.1067169904708862} 02/26/2022 15:25:35 - INFO - codeparrot_training - Step 47753: {'lr': 2.6986746605118795e-06, 'samples': 24450048, 'steps': 47753, 'loss/train': 1.8368946313858032} 02/26/2022 15:25:39 - INFO - codeparrot_training - Step 47754: {'lr': 2.6962774964545687e-06, 'samples': 24450560, 'steps': 47754, 'loss/train': 1.2937439680099487} 02/26/2022 15:25:44 - INFO - codeparrot_training - Step 47755: {'lr': 2.693881391767905e-06, 'samples': 24451072, 'steps': 47755, 'loss/train': 1.3347302675247192} 02/26/2022 15:25:48 - INFO - codeparrot_training - Step 47756: {'lr': 2.691486346462102e-06, 'samples': 24451584, 'steps': 47756, 'loss/train': 1.0759286880493164} 02/26/2022 15:25:53 - INFO - codeparrot_training - Step 47757: {'lr': 2.689092360547485e-06, 'samples': 24452096, 'steps': 47757, 'loss/train': 0.9846163392066956} 02/26/2022 15:25:57 - INFO - codeparrot_training - Step 47758: {'lr': 2.68669943403424e-06, 'samples': 24452608, 'steps': 47758, 'loss/train': 0.17225030064582825} 02/26/2022 15:26:02 - INFO - codeparrot_training - Step 47759: {'lr': 2.6843075669326366e-06, 'samples': 24453120, 'steps': 47759, 'loss/train': 1.6041593551635742} 02/26/2022 15:26:06 - INFO - codeparrot_training - Step 47760: {'lr': 2.681916759252917e-06, 'samples': 24453632, 'steps': 47760, 'loss/train': 1.8171963691711426} 02/26/2022 15:26:13 - INFO - codeparrot_training - Step 47761: {'lr': 2.6795270110053784e-06, 'samples': 24454144, 'steps': 47761, 'loss/train': 1.549590826034546} 02/26/2022 15:26:16 - INFO - codeparrot_training - Step 47762: {'lr': 2.677138322200151e-06, 'samples': 24454656, 'steps': 47762, 'loss/train': 1.2060056924819946} 02/26/2022 15:26:22 - INFO - codeparrot_training - Step 47763: {'lr': 2.6747506928475882e-06, 'samples': 24455168, 'steps': 47763, 'loss/train': 2.1794633865356445} 02/26/2022 15:26:25 - INFO - codeparrot_training - Step 47764: {'lr': 2.672364122957793e-06, 'samples': 24455680, 'steps': 47764, 'loss/train': 1.4822766780853271} 02/26/2022 15:26:31 - INFO - codeparrot_training - Step 47765: {'lr': 2.6699786125411187e-06, 'samples': 24456192, 'steps': 47765, 'loss/train': 1.182654857635498} 02/26/2022 15:26:34 - INFO - codeparrot_training - Step 47766: {'lr': 2.667594161607695e-06, 'samples': 24456704, 'steps': 47766, 'loss/train': 1.8918431997299194} 02/26/2022 15:26:40 - INFO - codeparrot_training - Step 47767: {'lr': 2.665210770167764e-06, 'samples': 24457216, 'steps': 47767, 'loss/train': 2.267146587371826} 02/26/2022 15:26:43 - INFO - codeparrot_training - Step 47768: {'lr': 2.6628284382315125e-06, 'samples': 24457728, 'steps': 47768, 'loss/train': 0.549653172492981} 02/26/2022 15:26:49 - INFO - codeparrot_training - Step 47769: {'lr': 2.6604471658091823e-06, 'samples': 24458240, 'steps': 47769, 'loss/train': 2.4738872051239014} 02/26/2022 15:26:52 - INFO - codeparrot_training - Step 47770: {'lr': 2.658066952910959e-06, 'samples': 24458752, 'steps': 47770, 'loss/train': 1.3863539695739746} 02/26/2022 15:27:00 - INFO - codeparrot_training - Step 47771: {'lr': 2.6556877995470572e-06, 'samples': 24459264, 'steps': 47771, 'loss/train': 1.9759031534194946} 02/26/2022 15:27:03 - INFO - codeparrot_training - Step 47772: {'lr': 2.6533097057276077e-06, 'samples': 24459776, 'steps': 47772, 'loss/train': 1.876699686050415} 02/26/2022 15:27:09 - INFO - codeparrot_training - Step 47773: {'lr': 2.650932671462852e-06, 'samples': 24460288, 'steps': 47773, 'loss/train': 1.3217414617538452} 02/26/2022 15:27:12 - INFO - codeparrot_training - Step 47774: {'lr': 2.648556696762977e-06, 'samples': 24460800, 'steps': 47774, 'loss/train': 1.2455205917358398} 02/26/2022 15:27:18 - INFO - codeparrot_training - Step 47775: {'lr': 2.6461817816381407e-06, 'samples': 24461312, 'steps': 47775, 'loss/train': 1.4605668783187866} 02/26/2022 15:27:21 - INFO - codeparrot_training - Step 47776: {'lr': 2.6438079260984737e-06, 'samples': 24461824, 'steps': 47776, 'loss/train': 0.34825846552848816} 02/26/2022 15:27:27 - INFO - codeparrot_training - Step 47777: {'lr': 2.641435130154246e-06, 'samples': 24462336, 'steps': 47777, 'loss/train': 2.0794758796691895} 02/26/2022 15:27:30 - INFO - codeparrot_training - Step 47778: {'lr': 2.639063393815533e-06, 'samples': 24462848, 'steps': 47778, 'loss/train': 1.456945776939392} 02/26/2022 15:27:36 - INFO - codeparrot_training - Step 47779: {'lr': 2.636692717092548e-06, 'samples': 24463360, 'steps': 47779, 'loss/train': 2.4288454055786133} 02/26/2022 15:27:39 - INFO - codeparrot_training - Step 47780: {'lr': 2.634323099995395e-06, 'samples': 24463872, 'steps': 47780, 'loss/train': 2.041590452194214} 02/26/2022 15:27:46 - INFO - codeparrot_training - Step 47781: {'lr': 2.6319545425342593e-06, 'samples': 24464384, 'steps': 47781, 'loss/train': 2.0931224822998047} 02/26/2022 15:27:52 - INFO - codeparrot_training - Step 47782: {'lr': 2.6295870447193006e-06, 'samples': 24464896, 'steps': 47782, 'loss/train': 0.8226175308227539} 02/26/2022 15:27:55 - INFO - codeparrot_training - Step 47783: {'lr': 2.6272206065606486e-06, 'samples': 24465408, 'steps': 47783, 'loss/train': 1.8073890209197998} 02/26/2022 15:28:01 - INFO - codeparrot_training - Step 47784: {'lr': 2.6248552280684067e-06, 'samples': 24465920, 'steps': 47784, 'loss/train': 2.063500165939331} 02/26/2022 15:28:04 - INFO - codeparrot_training - Step 47785: {'lr': 2.6224909092527337e-06, 'samples': 24466432, 'steps': 47785, 'loss/train': 1.3700064420700073} 02/26/2022 15:28:10 - INFO - codeparrot_training - Step 47786: {'lr': 2.620127650123788e-06, 'samples': 24466944, 'steps': 47786, 'loss/train': 1.7232751846313477} 02/26/2022 15:28:13 - INFO - codeparrot_training - Step 47787: {'lr': 2.6177654506916447e-06, 'samples': 24467456, 'steps': 47787, 'loss/train': 2.575131416320801} 02/26/2022 15:28:19 - INFO - codeparrot_training - Step 47788: {'lr': 2.6154043109664626e-06, 'samples': 24467968, 'steps': 47788, 'loss/train': 3.261383056640625} 02/26/2022 15:28:22 - INFO - codeparrot_training - Step 47789: {'lr': 2.613044230958289e-06, 'samples': 24468480, 'steps': 47789, 'loss/train': 1.7220776081085205} 02/26/2022 15:28:28 - INFO - codeparrot_training - Step 47790: {'lr': 2.6106852106773106e-06, 'samples': 24468992, 'steps': 47790, 'loss/train': 1.222286343574524} 02/26/2022 15:28:31 - INFO - codeparrot_training - Step 47791: {'lr': 2.608327250133602e-06, 'samples': 24469504, 'steps': 47791, 'loss/train': 0.5067892074584961} 02/26/2022 15:28:38 - INFO - codeparrot_training - Step 47792: {'lr': 2.6059703493372665e-06, 'samples': 24470016, 'steps': 47792, 'loss/train': 1.6780601739883423} 02/26/2022 15:28:42 - INFO - codeparrot_training - Step 47793: {'lr': 2.6036145082983797e-06, 'samples': 24470528, 'steps': 47793, 'loss/train': 0.21104831993579865} 02/26/2022 15:28:47 - INFO - codeparrot_training - Step 47794: {'lr': 2.601259727027072e-06, 'samples': 24471040, 'steps': 47794, 'loss/train': 2.0346317291259766} 02/26/2022 15:28:51 - INFO - codeparrot_training - Step 47795: {'lr': 2.5989060055333913e-06, 'samples': 24471552, 'steps': 47795, 'loss/train': 0.6854222416877747} 02/26/2022 15:28:56 - INFO - codeparrot_training - Step 47796: {'lr': 2.596553343827468e-06, 'samples': 24472064, 'steps': 47796, 'loss/train': 1.9851348400115967} 02/26/2022 15:29:00 - INFO - codeparrot_training - Step 47797: {'lr': 2.594201741919322e-06, 'samples': 24472576, 'steps': 47797, 'loss/train': 1.6300476789474487} 02/26/2022 15:29:05 - INFO - codeparrot_training - Step 47798: {'lr': 2.5918511998190565e-06, 'samples': 24473088, 'steps': 47798, 'loss/train': 2.656505584716797} 02/26/2022 15:29:09 - INFO - codeparrot_training - Step 47799: {'lr': 2.5895017175367463e-06, 'samples': 24473600, 'steps': 47799, 'loss/train': 1.9577513933181763} 02/26/2022 15:29:12 - INFO - codeparrot_training - Step 47800: {'lr': 2.5871532950824397e-06, 'samples': 24474112, 'steps': 47800, 'loss/train': 1.0012179613113403} 02/26/2022 15:29:18 - INFO - codeparrot_training - Step 47801: {'lr': 2.584805932466239e-06, 'samples': 24474624, 'steps': 47801, 'loss/train': 1.2555547952651978} 02/26/2022 15:29:21 - INFO - codeparrot_training - Step 47802: {'lr': 2.5824596296981362e-06, 'samples': 24475136, 'steps': 47802, 'loss/train': 1.833644151687622} 02/26/2022 15:29:27 - INFO - codeparrot_training - Step 47803: {'lr': 2.580114386788207e-06, 'samples': 24475648, 'steps': 47803, 'loss/train': 0.8827839493751526} 02/26/2022 15:29:30 - INFO - codeparrot_training - Step 47804: {'lr': 2.5777702037465267e-06, 'samples': 24476160, 'steps': 47804, 'loss/train': 2.377434015274048} 02/26/2022 15:29:36 - INFO - codeparrot_training - Step 47805: {'lr': 2.5754270805830872e-06, 'samples': 24476672, 'steps': 47805, 'loss/train': 1.6889660358428955} 02/26/2022 15:29:43 - INFO - codeparrot_training - Step 47806: {'lr': 2.573085017307991e-06, 'samples': 24477184, 'steps': 47806, 'loss/train': 1.3994858264923096} 02/26/2022 15:29:47 - INFO - codeparrot_training - Step 47807: {'lr': 2.570744013931203e-06, 'samples': 24477696, 'steps': 47807, 'loss/train': 1.5294290781021118} 02/26/2022 15:29:52 - INFO - codeparrot_training - Step 47808: {'lr': 2.568404070462771e-06, 'samples': 24478208, 'steps': 47808, 'loss/train': 2.0072031021118164} 02/26/2022 15:29:56 - INFO - codeparrot_training - Step 47809: {'lr': 2.5660651869127415e-06, 'samples': 24478720, 'steps': 47809, 'loss/train': 2.090956926345825} 02/26/2022 15:30:01 - INFO - codeparrot_training - Step 47810: {'lr': 2.563727363291135e-06, 'samples': 24479232, 'steps': 47810, 'loss/train': 2.6381170749664307} 02/26/2022 15:30:04 - INFO - codeparrot_training - Step 47811: {'lr': 2.5613905996079156e-06, 'samples': 24479744, 'steps': 47811, 'loss/train': 0.9159024357795715} 02/26/2022 15:30:10 - INFO - codeparrot_training - Step 47812: {'lr': 2.5590548958731586e-06, 'samples': 24480256, 'steps': 47812, 'loss/train': 2.4449822902679443} 02/26/2022 15:30:14 - INFO - codeparrot_training - Step 47813: {'lr': 2.556720252096828e-06, 'samples': 24480768, 'steps': 47813, 'loss/train': 1.6399636268615723} 02/26/2022 15:30:19 - INFO - codeparrot_training - Step 47814: {'lr': 2.5543866682889437e-06, 'samples': 24481280, 'steps': 47814, 'loss/train': 1.7094807624816895} 02/26/2022 15:30:23 - INFO - codeparrot_training - Step 47815: {'lr': 2.5520541444594703e-06, 'samples': 24481792, 'steps': 47815, 'loss/train': 0.7401649951934814} 02/26/2022 15:30:30 - INFO - codeparrot_training - Step 47816: {'lr': 2.5497226806184548e-06, 'samples': 24482304, 'steps': 47816, 'loss/train': 0.8024858236312866} 02/26/2022 15:30:33 - INFO - codeparrot_training - Step 47817: {'lr': 2.5473922767758063e-06, 'samples': 24482816, 'steps': 47817, 'loss/train': 2.027305841445923} 02/26/2022 15:30:39 - INFO - codeparrot_training - Step 47818: {'lr': 2.5450629329416277e-06, 'samples': 24483328, 'steps': 47818, 'loss/train': 1.7449239492416382} 02/26/2022 15:30:42 - INFO - codeparrot_training - Step 47819: {'lr': 2.5427346491257717e-06, 'samples': 24483840, 'steps': 47819, 'loss/train': 1.380434274673462} 02/26/2022 15:30:48 - INFO - codeparrot_training - Step 47820: {'lr': 2.5404074253382592e-06, 'samples': 24484352, 'steps': 47820, 'loss/train': 0.8455139994621277} 02/26/2022 15:30:52 - INFO - codeparrot_training - Step 47821: {'lr': 2.5380812615890813e-06, 'samples': 24484864, 'steps': 47821, 'loss/train': 2.0565526485443115} 02/26/2022 15:30:57 - INFO - codeparrot_training - Step 47822: {'lr': 2.5357561578882026e-06, 'samples': 24485376, 'steps': 47822, 'loss/train': 2.365363597869873} 02/26/2022 15:31:01 - INFO - codeparrot_training - Step 47823: {'lr': 2.533432114245532e-06, 'samples': 24485888, 'steps': 47823, 'loss/train': 0.7811762690544128} 02/26/2022 15:31:06 - INFO - codeparrot_training - Step 47824: {'lr': 2.531109130671061e-06, 'samples': 24486400, 'steps': 47824, 'loss/train': 1.7682344913482666} 02/26/2022 15:31:10 - INFO - codeparrot_training - Step 47825: {'lr': 2.528787207174754e-06, 'samples': 24486912, 'steps': 47825, 'loss/train': 0.8853068351745605} 02/26/2022 15:31:15 - INFO - codeparrot_training - Step 47826: {'lr': 2.526466343766548e-06, 'samples': 24487424, 'steps': 47826, 'loss/train': 2.8831467628479004} 02/26/2022 15:31:19 - INFO - codeparrot_training - Step 47827: {'lr': 2.5241465404563512e-06, 'samples': 24487936, 'steps': 47827, 'loss/train': 1.5609880685806274} 02/26/2022 15:31:26 - INFO - codeparrot_training - Step 47828: {'lr': 2.5218277972541557e-06, 'samples': 24488448, 'steps': 47828, 'loss/train': 1.6217362880706787} 02/26/2022 15:31:29 - INFO - codeparrot_training - Step 47829: {'lr': 2.5195101141698706e-06, 'samples': 24488960, 'steps': 47829, 'loss/train': 1.2341086864471436} 02/26/2022 15:31:35 - INFO - codeparrot_training - Step 47830: {'lr': 2.5171934912133765e-06, 'samples': 24489472, 'steps': 47830, 'loss/train': 1.317106008529663} 02/26/2022 15:31:38 - INFO - codeparrot_training - Step 47831: {'lr': 2.514877928394693e-06, 'samples': 24489984, 'steps': 47831, 'loss/train': 1.6443696022033691} 02/26/2022 15:31:44 - INFO - codeparrot_training - Step 47832: {'lr': 2.512563425723646e-06, 'samples': 24490496, 'steps': 47832, 'loss/train': 1.7121108770370483} 02/26/2022 15:31:47 - INFO - codeparrot_training - Step 47833: {'lr': 2.5102499832102275e-06, 'samples': 24491008, 'steps': 47833, 'loss/train': 2.020407199859619} 02/26/2022 15:31:53 - INFO - codeparrot_training - Step 47834: {'lr': 2.5079376008642906e-06, 'samples': 24491520, 'steps': 47834, 'loss/train': 1.2551398277282715} 02/26/2022 15:31:56 - INFO - codeparrot_training - Step 47835: {'lr': 2.505626278695744e-06, 'samples': 24492032, 'steps': 47835, 'loss/train': 2.0137217044830322} 02/26/2022 15:32:02 - INFO - codeparrot_training - Step 47836: {'lr': 2.503316016714552e-06, 'samples': 24492544, 'steps': 47836, 'loss/train': 2.295029640197754} 02/26/2022 15:32:05 - INFO - codeparrot_training - Step 47837: {'lr': 2.50100681493054e-06, 'samples': 24493056, 'steps': 47837, 'loss/train': 1.4651015996932983} 02/26/2022 15:32:13 - INFO - codeparrot_training - Step 47838: {'lr': 2.4986986733536164e-06, 'samples': 24493568, 'steps': 47838, 'loss/train': 1.918484091758728} 02/26/2022 15:32:16 - INFO - codeparrot_training - Step 47839: {'lr': 2.496391591993663e-06, 'samples': 24494080, 'steps': 47839, 'loss/train': 1.6724934577941895} 02/26/2022 15:32:22 - INFO - codeparrot_training - Step 47840: {'lr': 2.494085570860616e-06, 'samples': 24494592, 'steps': 47840, 'loss/train': 1.6329292058944702} 02/26/2022 15:32:25 - INFO - codeparrot_training - Step 47841: {'lr': 2.4917806099643003e-06, 'samples': 24495104, 'steps': 47841, 'loss/train': 2.21370005607605} 02/26/2022 15:32:31 - INFO - codeparrot_training - Step 47842: {'lr': 2.4894767093145975e-06, 'samples': 24495616, 'steps': 47842, 'loss/train': 2.584339141845703} 02/26/2022 15:32:34 - INFO - codeparrot_training - Step 47843: {'lr': 2.4871738689213886e-06, 'samples': 24496128, 'steps': 47843, 'loss/train': 1.3005484342575073} 02/26/2022 15:32:39 - INFO - codeparrot_training - Step 47844: {'lr': 2.4848720887945263e-06, 'samples': 24496640, 'steps': 47844, 'loss/train': 1.1254960298538208} 02/26/2022 15:32:43 - INFO - codeparrot_training - Step 47845: {'lr': 2.482571368943892e-06, 'samples': 24497152, 'steps': 47845, 'loss/train': 0.4927946925163269} 02/26/2022 15:32:49 - INFO - codeparrot_training - Step 47846: {'lr': 2.4802717093792835e-06, 'samples': 24497664, 'steps': 47846, 'loss/train': 1.3502267599105835} 02/26/2022 15:32:52 - INFO - codeparrot_training - Step 47847: {'lr': 2.4779731101106372e-06, 'samples': 24498176, 'steps': 47847, 'loss/train': 0.7948471307754517} 02/26/2022 15:32:58 - INFO - codeparrot_training - Step 47848: {'lr': 2.4756755711477508e-06, 'samples': 24498688, 'steps': 47848, 'loss/train': 1.4448049068450928} 02/26/2022 15:33:01 - INFO - codeparrot_training - Step 47849: {'lr': 2.4733790925004773e-06, 'samples': 24499200, 'steps': 47849, 'loss/train': 2.192993640899658} 02/26/2022 15:33:07 - INFO - codeparrot_training - Step 47850: {'lr': 2.4710836741786145e-06, 'samples': 24499712, 'steps': 47850, 'loss/train': 2.073532819747925} 02/26/2022 15:33:10 - INFO - codeparrot_training - Step 47851: {'lr': 2.4687893161920717e-06, 'samples': 24500224, 'steps': 47851, 'loss/train': 1.7523607015609741} 02/26/2022 15:33:17 - INFO - codeparrot_training - Step 47852: {'lr': 2.466496018550618e-06, 'samples': 24500736, 'steps': 47852, 'loss/train': 2.029283285140991} 02/26/2022 15:33:21 - INFO - codeparrot_training - Step 47853: {'lr': 2.464203781264107e-06, 'samples': 24501248, 'steps': 47853, 'loss/train': 2.250612735748291} 02/26/2022 15:33:26 - INFO - codeparrot_training - Step 47854: {'lr': 2.461912604342309e-06, 'samples': 24501760, 'steps': 47854, 'loss/train': 1.9850369691848755} 02/26/2022 15:33:30 - INFO - codeparrot_training - Step 47855: {'lr': 2.4596224877951044e-06, 'samples': 24502272, 'steps': 47855, 'loss/train': 2.4468674659729004} 02/26/2022 15:33:35 - INFO - codeparrot_training - Step 47856: {'lr': 2.457333431632264e-06, 'samples': 24502784, 'steps': 47856, 'loss/train': 1.7805763483047485} 02/26/2022 15:33:39 - INFO - codeparrot_training - Step 47857: {'lr': 2.45504543586364e-06, 'samples': 24503296, 'steps': 47857, 'loss/train': 0.8310919404029846} 02/26/2022 15:33:45 - INFO - codeparrot_training - Step 47858: {'lr': 2.452758500498947e-06, 'samples': 24503808, 'steps': 47858, 'loss/train': 3.1714062690734863} 02/26/2022 15:33:48 - INFO - codeparrot_training - Step 47859: {'lr': 2.450472625548067e-06, 'samples': 24504320, 'steps': 47859, 'loss/train': 2.0159599781036377} 02/26/2022 15:33:54 - INFO - codeparrot_training - Step 47860: {'lr': 2.4481878110207134e-06, 'samples': 24504832, 'steps': 47860, 'loss/train': 2.3763718605041504} 02/26/2022 15:33:57 - INFO - codeparrot_training - Step 47861: {'lr': 2.44590405692674e-06, 'samples': 24505344, 'steps': 47861, 'loss/train': 3.1528851985931396} 02/26/2022 15:34:02 - INFO - codeparrot_training - Step 47862: {'lr': 2.443621363275916e-06, 'samples': 24505856, 'steps': 47862, 'loss/train': 1.9872863292694092} 02/26/2022 15:34:06 - INFO - codeparrot_training - Step 47863: {'lr': 2.4413397300780127e-06, 'samples': 24506368, 'steps': 47863, 'loss/train': 1.5886955261230469} 02/26/2022 15:34:13 - INFO - codeparrot_training - Step 47864: {'lr': 2.439059157342799e-06, 'samples': 24506880, 'steps': 47864, 'loss/train': 2.1267330646514893} 02/26/2022 15:34:17 - INFO - codeparrot_training - Step 47865: {'lr': 2.4367796450800174e-06, 'samples': 24507392, 'steps': 47865, 'loss/train': 2.1475813388824463} 02/26/2022 15:34:22 - INFO - codeparrot_training - Step 47866: {'lr': 2.434501193299493e-06, 'samples': 24507904, 'steps': 47866, 'loss/train': 1.3327138423919678} 02/26/2022 15:34:25 - INFO - codeparrot_training - Step 47867: {'lr': 2.432223802010969e-06, 'samples': 24508416, 'steps': 47867, 'loss/train': 1.3871217966079712} 02/26/2022 15:34:31 - INFO - codeparrot_training - Step 47868: {'lr': 2.429947471224131e-06, 'samples': 24508928, 'steps': 47868, 'loss/train': 2.093010902404785} 02/26/2022 15:34:34 - INFO - codeparrot_training - Step 47869: {'lr': 2.4276722009488328e-06, 'samples': 24509440, 'steps': 47869, 'loss/train': 0.7602604627609253} 02/26/2022 15:34:40 - INFO - codeparrot_training - Step 47870: {'lr': 2.425397991194761e-06, 'samples': 24509952, 'steps': 47870, 'loss/train': 1.63302743434906} 02/26/2022 15:34:43 - INFO - codeparrot_training - Step 47871: {'lr': 2.4231248419716856e-06, 'samples': 24510464, 'steps': 47871, 'loss/train': 2.0227210521698} 02/26/2022 15:34:49 - INFO - codeparrot_training - Step 47872: {'lr': 2.420852753289293e-06, 'samples': 24510976, 'steps': 47872, 'loss/train': 2.16332745552063} 02/26/2022 15:34:52 - INFO - codeparrot_training - Step 47873: {'lr': 2.418581725157382e-06, 'samples': 24511488, 'steps': 47873, 'loss/train': 0.573132336139679} 02/26/2022 15:35:00 - INFO - codeparrot_training - Step 47874: {'lr': 2.4163117575856376e-06, 'samples': 24512000, 'steps': 47874, 'loss/train': 1.449105978012085} 02/26/2022 15:35:03 - INFO - codeparrot_training - Step 47875: {'lr': 2.414042850583803e-06, 'samples': 24512512, 'steps': 47875, 'loss/train': 2.2010200023651123} 02/26/2022 15:35:09 - INFO - codeparrot_training - Step 47876: {'lr': 2.4117750041615926e-06, 'samples': 24513024, 'steps': 47876, 'loss/train': 1.3662270307540894} 02/26/2022 15:35:12 - INFO - codeparrot_training - Step 47877: {'lr': 2.409508218328693e-06, 'samples': 24513536, 'steps': 47877, 'loss/train': 0.9514808654785156} 02/26/2022 15:35:18 - INFO - codeparrot_training - Step 47878: {'lr': 2.4072424930948744e-06, 'samples': 24514048, 'steps': 47878, 'loss/train': 2.7840561866760254} 02/26/2022 15:35:21 - INFO - codeparrot_training - Step 47879: {'lr': 2.4049778284697953e-06, 'samples': 24514560, 'steps': 47879, 'loss/train': 1.0101137161254883} 02/26/2022 15:35:27 - INFO - codeparrot_training - Step 47880: {'lr': 2.4027142244631706e-06, 'samples': 24515072, 'steps': 47880, 'loss/train': 1.3735039234161377} 02/26/2022 15:35:30 - INFO - codeparrot_training - Step 47881: {'lr': 2.4004516810846864e-06, 'samples': 24515584, 'steps': 47881, 'loss/train': 1.14693284034729} 02/26/2022 15:35:36 - INFO - codeparrot_training - Step 47882: {'lr': 2.3981901983440855e-06, 'samples': 24516096, 'steps': 47882, 'loss/train': 1.460396647453308} 02/26/2022 15:35:39 - INFO - codeparrot_training - Step 47883: {'lr': 2.3959297762509714e-06, 'samples': 24516608, 'steps': 47883, 'loss/train': 1.0896514654159546} 02/26/2022 15:35:46 - INFO - codeparrot_training - Step 47884: {'lr': 2.3936704148151134e-06, 'samples': 24517120, 'steps': 47884, 'loss/train': 2.7399637699127197} 02/26/2022 15:35:50 - INFO - codeparrot_training - Step 47885: {'lr': 2.3914121140460877e-06, 'samples': 24517632, 'steps': 47885, 'loss/train': 2.3188326358795166} 02/26/2022 15:35:55 - INFO - codeparrot_training - Step 47886: {'lr': 2.3891548739536918e-06, 'samples': 24518144, 'steps': 47886, 'loss/train': 1.9701364040374756} 02/26/2022 15:35:59 - INFO - codeparrot_training - Step 47887: {'lr': 2.3868986945474735e-06, 'samples': 24518656, 'steps': 47887, 'loss/train': 0.6902536749839783} 02/26/2022 15:36:04 - INFO - codeparrot_training - Step 47888: {'lr': 2.3846435758372033e-06, 'samples': 24519168, 'steps': 47888, 'loss/train': 1.5973650217056274} 02/26/2022 15:36:08 - INFO - codeparrot_training - Step 47889: {'lr': 2.382389517832456e-06, 'samples': 24519680, 'steps': 47889, 'loss/train': 1.7861621379852295} 02/26/2022 15:36:13 - INFO - codeparrot_training - Step 47890: {'lr': 2.380136520542947e-06, 'samples': 24520192, 'steps': 47890, 'loss/train': 1.3621666431427002} 02/26/2022 15:36:17 - INFO - codeparrot_training - Step 47891: {'lr': 2.377884583978307e-06, 'samples': 24520704, 'steps': 47891, 'loss/train': 1.4045051336288452} 02/26/2022 15:36:22 - INFO - codeparrot_training - Step 47892: {'lr': 2.375633708148195e-06, 'samples': 24521216, 'steps': 47892, 'loss/train': 1.2694860696792603} 02/26/2022 15:36:26 - INFO - codeparrot_training - Step 47893: {'lr': 2.3733838930621865e-06, 'samples': 24521728, 'steps': 47893, 'loss/train': 2.413146495819092} 02/26/2022 15:36:31 - INFO - codeparrot_training - Step 47894: {'lr': 2.371135138730024e-06, 'samples': 24522240, 'steps': 47894, 'loss/train': 1.6571993827819824} 02/26/2022 15:36:37 - INFO - codeparrot_training - Step 47895: {'lr': 2.368887445161255e-06, 'samples': 24522752, 'steps': 47895, 'loss/train': 1.3963418006896973} 02/26/2022 15:36:40 - INFO - codeparrot_training - Step 47896: {'lr': 2.3666408123655393e-06, 'samples': 24523264, 'steps': 47896, 'loss/train': 8.758283615112305} 02/26/2022 15:36:44 - INFO - codeparrot_training - Step 47897: {'lr': 2.364395240352535e-06, 'samples': 24523776, 'steps': 47897, 'loss/train': 1.697182059288025} 02/26/2022 15:36:49 - INFO - codeparrot_training - Step 47898: {'lr': 2.36215072913179e-06, 'samples': 24524288, 'steps': 47898, 'loss/train': 1.6604324579238892} 02/26/2022 15:36:53 - INFO - codeparrot_training - Step 47899: {'lr': 2.3599072787129637e-06, 'samples': 24524800, 'steps': 47899, 'loss/train': 0.9347661137580872} 02/26/2022 15:37:00 - INFO - codeparrot_training - Step 47900: {'lr': 2.357664889105687e-06, 'samples': 24525312, 'steps': 47900, 'loss/train': 1.2365621328353882} 02/26/2022 15:37:04 - INFO - codeparrot_training - Step 47901: {'lr': 2.3554235603195362e-06, 'samples': 24525824, 'steps': 47901, 'loss/train': 0.9771762490272522} 02/26/2022 15:37:09 - INFO - codeparrot_training - Step 47902: {'lr': 2.3531832923640862e-06, 'samples': 24526336, 'steps': 47902, 'loss/train': 0.9857110381126404} 02/26/2022 15:37:13 - INFO - codeparrot_training - Step 47903: {'lr': 2.3509440852489683e-06, 'samples': 24526848, 'steps': 47903, 'loss/train': 1.839868426322937} 02/26/2022 15:37:18 - INFO - codeparrot_training - Step 47904: {'lr': 2.348705938983786e-06, 'samples': 24527360, 'steps': 47904, 'loss/train': 2.1556990146636963} 02/26/2022 15:37:24 - INFO - codeparrot_training - Step 47905: {'lr': 2.3464688535781155e-06, 'samples': 24527872, 'steps': 47905, 'loss/train': 1.6833938360214233} 02/26/2022 15:37:27 - INFO - codeparrot_training - Step 47906: {'lr': 2.3442328290415037e-06, 'samples': 24528384, 'steps': 47906, 'loss/train': 1.6223562955856323} 02/26/2022 15:37:33 - INFO - codeparrot_training - Step 47907: {'lr': 2.3419978653835826e-06, 'samples': 24528896, 'steps': 47907, 'loss/train': 1.238167643547058} 02/26/2022 15:37:36 - INFO - codeparrot_training - Step 47908: {'lr': 2.3397639626139002e-06, 'samples': 24529408, 'steps': 47908, 'loss/train': 1.3795874118804932} 02/26/2022 15:37:44 - INFO - codeparrot_training - Step 47909: {'lr': 2.3375311207420037e-06, 'samples': 24529920, 'steps': 47909, 'loss/train': 0.9087982773780823} 02/26/2022 15:37:47 - INFO - codeparrot_training - Step 47910: {'lr': 2.3352993397775247e-06, 'samples': 24530432, 'steps': 47910, 'loss/train': 1.6121654510498047} 02/26/2022 15:37:53 - INFO - codeparrot_training - Step 47911: {'lr': 2.3330686197299276e-06, 'samples': 24530944, 'steps': 47911, 'loss/train': 1.5290254354476929} 02/26/2022 15:37:56 - INFO - codeparrot_training - Step 47912: {'lr': 2.330838960608872e-06, 'samples': 24531456, 'steps': 47912, 'loss/train': 1.2620149850845337} 02/26/2022 15:38:02 - INFO - codeparrot_training - Step 47913: {'lr': 2.328610362423822e-06, 'samples': 24531968, 'steps': 47913, 'loss/train': 0.2835817039012909} 02/26/2022 15:38:05 - INFO - codeparrot_training - Step 47914: {'lr': 2.326382825184381e-06, 'samples': 24532480, 'steps': 47914, 'loss/train': 2.129727602005005} 02/26/2022 15:38:11 - INFO - codeparrot_training - Step 47915: {'lr': 2.3241563489000692e-06, 'samples': 24532992, 'steps': 47915, 'loss/train': 1.2792174816131592} 02/26/2022 15:38:14 - INFO - codeparrot_training - Step 47916: {'lr': 2.3219309335804073e-06, 'samples': 24533504, 'steps': 47916, 'loss/train': 0.8699167370796204} 02/26/2022 15:38:20 - INFO - codeparrot_training - Step 47917: {'lr': 2.31970657923497e-06, 'samples': 24534016, 'steps': 47917, 'loss/train': 2.230705976486206} 02/26/2022 15:38:23 - INFO - codeparrot_training - Step 47918: {'lr': 2.317483285873251e-06, 'samples': 24534528, 'steps': 47918, 'loss/train': 0.9881263375282288} 02/26/2022 15:38:30 - INFO - codeparrot_training - Step 47919: {'lr': 2.315261053504797e-06, 'samples': 24535040, 'steps': 47919, 'loss/train': 2.4287497997283936} 02/26/2022 15:38:34 - INFO - codeparrot_training - Step 47920: {'lr': 2.3130398821391008e-06, 'samples': 24535552, 'steps': 47920, 'loss/train': 2.214642286300659} 02/26/2022 15:38:39 - INFO - codeparrot_training - Step 47921: {'lr': 2.3108197717857105e-06, 'samples': 24536064, 'steps': 47921, 'loss/train': 1.6973623037338257} 02/26/2022 15:38:43 - INFO - codeparrot_training - Step 47922: {'lr': 2.3086007224540906e-06, 'samples': 24536576, 'steps': 47922, 'loss/train': 2.3015148639678955} 02/26/2022 15:38:48 - INFO - codeparrot_training - Step 47923: {'lr': 2.306382734153817e-06, 'samples': 24537088, 'steps': 47923, 'loss/train': 1.8384315967559814} 02/26/2022 15:38:52 - INFO - codeparrot_training - Step 47924: {'lr': 2.3041658068942984e-06, 'samples': 24537600, 'steps': 47924, 'loss/train': 2.262763023376465} 02/26/2022 15:38:57 - INFO - codeparrot_training - Step 47925: {'lr': 2.3019499406851106e-06, 'samples': 24538112, 'steps': 47925, 'loss/train': 1.6985430717468262} 02/26/2022 15:39:01 - INFO - codeparrot_training - Step 47926: {'lr': 2.299735135535691e-06, 'samples': 24538624, 'steps': 47926, 'loss/train': 1.980005145072937} 02/26/2022 15:39:06 - INFO - codeparrot_training - Step 47927: {'lr': 2.297521391455587e-06, 'samples': 24539136, 'steps': 47927, 'loss/train': 1.6654679775238037} 02/26/2022 15:39:10 - INFO - codeparrot_training - Step 47928: {'lr': 2.2953087084542355e-06, 'samples': 24539648, 'steps': 47928, 'loss/train': 1.1862330436706543} 02/26/2022 15:39:17 - INFO - codeparrot_training - Step 47929: {'lr': 2.2930970865411015e-06, 'samples': 24540160, 'steps': 47929, 'loss/train': 2.939847469329834} 02/26/2022 15:39:20 - INFO - codeparrot_training - Step 47930: {'lr': 2.290886525725705e-06, 'samples': 24540672, 'steps': 47930, 'loss/train': 2.215400457382202} 02/26/2022 15:39:26 - INFO - codeparrot_training - Step 47931: {'lr': 2.2886770260175104e-06, 'samples': 24541184, 'steps': 47931, 'loss/train': 1.4276045560836792} 02/26/2022 15:39:29 - INFO - codeparrot_training - Step 47932: {'lr': 2.2864685874259273e-06, 'samples': 24541696, 'steps': 47932, 'loss/train': 1.2141623497009277} 02/26/2022 15:39:35 - INFO - codeparrot_training - Step 47933: {'lr': 2.2842612099604753e-06, 'samples': 24542208, 'steps': 47933, 'loss/train': 1.406009316444397} 02/26/2022 15:39:38 - INFO - codeparrot_training - Step 47934: {'lr': 2.2820548936305918e-06, 'samples': 24542720, 'steps': 47934, 'loss/train': 1.1579920053482056} 02/26/2022 15:39:44 - INFO - codeparrot_training - Step 47935: {'lr': 2.2798496384457414e-06, 'samples': 24543232, 'steps': 47935, 'loss/train': 1.4007172584533691} 02/26/2022 15:39:47 - INFO - codeparrot_training - Step 47936: {'lr': 2.2776454444153326e-06, 'samples': 24543744, 'steps': 47936, 'loss/train': 1.006408452987671} 02/26/2022 15:39:53 - INFO - codeparrot_training - Step 47937: {'lr': 2.275442311548831e-06, 'samples': 24544256, 'steps': 47937, 'loss/train': 1.9216763973236084} 02/26/2022 15:39:56 - INFO - codeparrot_training - Step 47938: {'lr': 2.2732402398556727e-06, 'samples': 24544768, 'steps': 47938, 'loss/train': 0.3714328706264496} 02/26/2022 15:40:02 - INFO - codeparrot_training - Step 47939: {'lr': 2.2710392293452954e-06, 'samples': 24545280, 'steps': 47939, 'loss/train': 2.0875027179718018} 02/26/2022 15:40:05 - INFO - codeparrot_training - Step 47940: {'lr': 2.268839280027135e-06, 'samples': 24545792, 'steps': 47940, 'loss/train': 1.7391934394836426} 02/26/2022 15:40:11 - INFO - codeparrot_training - Step 47941: {'lr': 2.2666403919106016e-06, 'samples': 24546304, 'steps': 47941, 'loss/train': 2.399202585220337} 02/26/2022 15:40:14 - INFO - codeparrot_training - Step 47942: {'lr': 2.264442565005076e-06, 'samples': 24546816, 'steps': 47942, 'loss/train': 1.592761516571045} 02/26/2022 15:40:20 - INFO - codeparrot_training - Step 47943: {'lr': 2.262245799320051e-06, 'samples': 24547328, 'steps': 47943, 'loss/train': 0.9611049294471741} 02/26/2022 15:40:23 - INFO - codeparrot_training - Step 47944: {'lr': 2.2600500948648793e-06, 'samples': 24547840, 'steps': 47944, 'loss/train': 1.998374342918396} 02/26/2022 15:40:31 - INFO - codeparrot_training - Step 47945: {'lr': 2.257855451649027e-06, 'samples': 24548352, 'steps': 47945, 'loss/train': 1.0205129384994507} 02/26/2022 15:40:34 - INFO - codeparrot_training - Step 47946: {'lr': 2.2556618696818187e-06, 'samples': 24548864, 'steps': 47946, 'loss/train': 1.6476207971572876} 02/26/2022 15:40:40 - INFO - codeparrot_training - Step 47947: {'lr': 2.25346934897272e-06, 'samples': 24549376, 'steps': 47947, 'loss/train': 2.0553367137908936} 02/26/2022 15:40:43 - INFO - codeparrot_training - Step 47948: {'lr': 2.251277889531056e-06, 'samples': 24549888, 'steps': 47948, 'loss/train': 1.4443731307983398} 02/26/2022 15:40:49 - INFO - codeparrot_training - Step 47949: {'lr': 2.249087491366264e-06, 'samples': 24550400, 'steps': 47949, 'loss/train': 2.54288387298584} 02/26/2022 15:40:52 - INFO - codeparrot_training - Step 47950: {'lr': 2.2468981544876978e-06, 'samples': 24550912, 'steps': 47950, 'loss/train': 1.1472249031066895} 02/26/2022 15:40:58 - INFO - codeparrot_training - Step 47951: {'lr': 2.2447098789047658e-06, 'samples': 24551424, 'steps': 47951, 'loss/train': 1.5696851015090942} 02/26/2022 15:41:01 - INFO - codeparrot_training - Step 47952: {'lr': 2.2425226646268228e-06, 'samples': 24551936, 'steps': 47952, 'loss/train': 0.9127157926559448} 02/26/2022 15:41:07 - INFO - codeparrot_training - Step 47953: {'lr': 2.240336511663249e-06, 'samples': 24552448, 'steps': 47953, 'loss/train': 1.50718355178833} 02/26/2022 15:41:10 - INFO - codeparrot_training - Step 47954: {'lr': 2.238151420023371e-06, 'samples': 24552960, 'steps': 47954, 'loss/train': 1.2241694927215576} 02/26/2022 15:41:18 - INFO - codeparrot_training - Step 47955: {'lr': 2.235967389716598e-06, 'samples': 24553472, 'steps': 47955, 'loss/train': 1.1092032194137573} 02/26/2022 15:41:21 - INFO - codeparrot_training - Step 47956: {'lr': 2.233784420752255e-06, 'samples': 24553984, 'steps': 47956, 'loss/train': 1.4935420751571655} 02/26/2022 15:41:27 - INFO - codeparrot_training - Step 47957: {'lr': 2.2316025131397245e-06, 'samples': 24554496, 'steps': 47957, 'loss/train': 2.387575149536133} 02/26/2022 15:41:30 - INFO - codeparrot_training - Step 47958: {'lr': 2.2294216668883315e-06, 'samples': 24555008, 'steps': 47958, 'loss/train': 2.4424500465393066} 02/26/2022 15:41:36 - INFO - codeparrot_training - Step 47959: {'lr': 2.2272418820074024e-06, 'samples': 24555520, 'steps': 47959, 'loss/train': 1.5509511232376099} 02/26/2022 15:41:39 - INFO - codeparrot_training - Step 47960: {'lr': 2.2250631585063187e-06, 'samples': 24556032, 'steps': 47960, 'loss/train': 0.11174685508012772} 02/26/2022 15:41:45 - INFO - codeparrot_training - Step 47961: {'lr': 2.22288549639435e-06, 'samples': 24556544, 'steps': 47961, 'loss/train': 2.091057538986206} 02/26/2022 15:41:48 - INFO - codeparrot_training - Step 47962: {'lr': 2.2207088956808785e-06, 'samples': 24557056, 'steps': 47962, 'loss/train': 1.0130252838134766} 02/26/2022 15:41:54 - INFO - codeparrot_training - Step 47963: {'lr': 2.2185333563752296e-06, 'samples': 24557568, 'steps': 47963, 'loss/train': 3.0432074069976807} 02/26/2022 15:41:57 - INFO - codeparrot_training - Step 47964: {'lr': 2.2163588784867015e-06, 'samples': 24558080, 'steps': 47964, 'loss/train': 1.931697130203247} 02/26/2022 15:42:04 - INFO - codeparrot_training - Step 47965: {'lr': 2.2141854620245927e-06, 'samples': 24558592, 'steps': 47965, 'loss/train': 1.3123613595962524} 02/26/2022 15:42:08 - INFO - codeparrot_training - Step 47966: {'lr': 2.2120131069982563e-06, 'samples': 24559104, 'steps': 47966, 'loss/train': 1.4895377159118652} 02/26/2022 15:42:13 - INFO - codeparrot_training - Step 47967: {'lr': 2.209841813416935e-06, 'samples': 24559616, 'steps': 47967, 'loss/train': 0.5027288794517517} 02/26/2022 15:42:17 - INFO - codeparrot_training - Step 47968: {'lr': 2.2076715812900107e-06, 'samples': 24560128, 'steps': 47968, 'loss/train': 1.7456347942352295} 02/26/2022 15:42:22 - INFO - codeparrot_training - Step 47969: {'lr': 2.2055024106266976e-06, 'samples': 24560640, 'steps': 47969, 'loss/train': 2.1062068939208984} 02/26/2022 15:42:25 - INFO - codeparrot_training - Step 47970: {'lr': 2.2033343014363494e-06, 'samples': 24561152, 'steps': 47970, 'loss/train': 1.4630951881408691} 02/26/2022 15:42:31 - INFO - codeparrot_training - Step 47971: {'lr': 2.201167253728237e-06, 'samples': 24561664, 'steps': 47971, 'loss/train': 1.4736186265945435} 02/26/2022 15:42:35 - INFO - codeparrot_training - Step 47972: {'lr': 2.199001267511658e-06, 'samples': 24562176, 'steps': 47972, 'loss/train': 1.333685278892517} 02/26/2022 15:42:40 - INFO - codeparrot_training - Step 47973: {'lr': 2.196836342795827e-06, 'samples': 24562688, 'steps': 47973, 'loss/train': 0.21811817586421967} 02/26/2022 15:42:44 - INFO - codeparrot_training - Step 47974: {'lr': 2.1946724795900984e-06, 'samples': 24563200, 'steps': 47974, 'loss/train': 2.17622709274292} 02/26/2022 15:42:51 - INFO - codeparrot_training - Step 47975: {'lr': 2.1925096779036868e-06, 'samples': 24563712, 'steps': 47975, 'loss/train': 1.3610150814056396} 02/26/2022 15:42:55 - INFO - codeparrot_training - Step 47976: {'lr': 2.19034793774589e-06, 'samples': 24564224, 'steps': 47976, 'loss/train': 1.4638632535934448} 02/26/2022 15:43:00 - INFO - codeparrot_training - Step 47977: {'lr': 2.188187259125951e-06, 'samples': 24564736, 'steps': 47977, 'loss/train': 0.938526451587677} 02/26/2022 15:43:04 - INFO - codeparrot_training - Step 47978: {'lr': 2.1860276420531124e-06, 'samples': 24565248, 'steps': 47978, 'loss/train': 0.755862832069397} 02/26/2022 15:43:09 - INFO - codeparrot_training - Step 47979: {'lr': 2.183869086536672e-06, 'samples': 24565760, 'steps': 47979, 'loss/train': 2.3575170040130615} 02/26/2022 15:43:13 - INFO - codeparrot_training - Step 47980: {'lr': 2.181711592585817e-06, 'samples': 24566272, 'steps': 47980, 'loss/train': 1.8960074186325073} 02/26/2022 15:43:18 - INFO - codeparrot_training - Step 47981: {'lr': 2.1795551602098175e-06, 'samples': 24566784, 'steps': 47981, 'loss/train': 0.8626107573509216} 02/26/2022 15:43:22 - INFO - codeparrot_training - Step 47982: {'lr': 2.1773997894179444e-06, 'samples': 24567296, 'steps': 47982, 'loss/train': 1.5351321697235107} 02/26/2022 15:43:27 - INFO - codeparrot_training - Step 47983: {'lr': 2.1752454802193567e-06, 'samples': 24567808, 'steps': 47983, 'loss/train': 1.9386399984359741} 02/26/2022 15:43:31 - INFO - codeparrot_training - Step 47984: {'lr': 2.1730922326233804e-06, 'samples': 24568320, 'steps': 47984, 'loss/train': 1.9180781841278076} 02/26/2022 15:43:36 - INFO - codeparrot_training - Step 47985: {'lr': 2.1709400466391194e-06, 'samples': 24568832, 'steps': 47985, 'loss/train': 1.9001375436782837} 02/26/2022 15:43:42 - INFO - codeparrot_training - Step 47986: {'lr': 2.1687889222758995e-06, 'samples': 24569344, 'steps': 47986, 'loss/train': 1.4409270286560059} 02/26/2022 15:43:45 - INFO - codeparrot_training - Step 47987: {'lr': 2.166638859542852e-06, 'samples': 24569856, 'steps': 47987, 'loss/train': 0.44483739137649536} 02/26/2022 15:43:51 - INFO - codeparrot_training - Step 47988: {'lr': 2.1644898584492756e-06, 'samples': 24570368, 'steps': 47988, 'loss/train': 1.263789176940918} 02/26/2022 15:43:54 - INFO - codeparrot_training - Step 47989: {'lr': 2.1623419190043014e-06, 'samples': 24570880, 'steps': 47989, 'loss/train': 1.4638453722000122} 02/26/2022 15:44:01 - INFO - codeparrot_training - Step 47990: {'lr': 2.1601950412171723e-06, 'samples': 24571392, 'steps': 47990, 'loss/train': 0.4411675035953522} 02/26/2022 15:44:05 - INFO - codeparrot_training - Step 47991: {'lr': 2.158049225097047e-06, 'samples': 24571904, 'steps': 47991, 'loss/train': 2.0408377647399902} 02/26/2022 15:44:11 - INFO - codeparrot_training - Step 47992: {'lr': 2.1559044706531415e-06, 'samples': 24572416, 'steps': 47992, 'loss/train': 0.5840863585472107} 02/26/2022 15:44:14 - INFO - codeparrot_training - Step 47993: {'lr': 2.15376077789467e-06, 'samples': 24572928, 'steps': 47993, 'loss/train': 0.5253375172615051} 02/26/2022 15:44:20 - INFO - codeparrot_training - Step 47994: {'lr': 2.151618146830764e-06, 'samples': 24573440, 'steps': 47994, 'loss/train': 2.518958330154419} 02/26/2022 15:44:23 - INFO - codeparrot_training - Step 47995: {'lr': 2.149476577470638e-06, 'samples': 24573952, 'steps': 47995, 'loss/train': 1.8539888858795166} 02/26/2022 15:44:29 - INFO - codeparrot_training - Step 47996: {'lr': 2.1473360698234245e-06, 'samples': 24574464, 'steps': 47996, 'loss/train': 1.164291262626648} 02/26/2022 15:44:32 - INFO - codeparrot_training - Step 47997: {'lr': 2.1451966238983656e-06, 'samples': 24574976, 'steps': 47997, 'loss/train': 1.8377363681793213} 02/26/2022 15:44:38 - INFO - codeparrot_training - Step 47998: {'lr': 2.143058239704565e-06, 'samples': 24575488, 'steps': 47998, 'loss/train': 2.02597975730896} 02/26/2022 15:44:41 - INFO - codeparrot_training - Step 47999: {'lr': 2.140920917251182e-06, 'samples': 24576000, 'steps': 47999, 'loss/train': 1.8916027545928955} 02/26/2022 15:44:41 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 15:44:59 - WARNING - huggingface_hub.repository - Several commits (48) will be pushed upstream. 02/26/2022 15:44:59 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 15:45:34 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 2e5c310..8e26ae5 floral-grass-11 -> floral-grass-11 02/26/2022 15:45:43 - INFO - codeparrot_training - Step 48000: {'lr': 2.1387846565474044e-06, 'samples': 24576512, 'steps': 48000, 'loss/train': 0.4398959279060364} 02/26/2022 15:45:46 - INFO - codeparrot_training - Step 48001: {'lr': 2.136649457602363e-06, 'samples': 24577024, 'steps': 48001, 'loss/train': 2.441908359527588} 02/26/2022 15:45:52 - INFO - codeparrot_training - Step 48002: {'lr': 2.1345153204252175e-06, 'samples': 24577536, 'steps': 48002, 'loss/train': 1.3035922050476074} 02/26/2022 15:45:55 - INFO - codeparrot_training - Step 48003: {'lr': 2.1323822450250997e-06, 'samples': 24578048, 'steps': 48003, 'loss/train': 2.239109516143799} 02/26/2022 15:46:01 - INFO - codeparrot_training - Step 48004: {'lr': 2.1302502314111406e-06, 'samples': 24578560, 'steps': 48004, 'loss/train': 2.2362773418426514} 02/26/2022 15:46:04 - INFO - codeparrot_training - Step 48005: {'lr': 2.128119279592472e-06, 'samples': 24579072, 'steps': 48005, 'loss/train': 0.5483190417289734} 02/26/2022 15:46:10 - INFO - codeparrot_training - Step 48006: {'lr': 2.1259893895782533e-06, 'samples': 24579584, 'steps': 48006, 'loss/train': 1.788357138633728} 02/26/2022 15:46:13 - INFO - codeparrot_training - Step 48007: {'lr': 2.1238605613775887e-06, 'samples': 24580096, 'steps': 48007, 'loss/train': 1.2798937559127808} 02/26/2022 15:46:19 - INFO - codeparrot_training - Step 48008: {'lr': 2.121732794999581e-06, 'samples': 24580608, 'steps': 48008, 'loss/train': 1.9912223815917969} 02/26/2022 15:46:22 - INFO - codeparrot_training - Step 48009: {'lr': 2.1196060904533632e-06, 'samples': 24581120, 'steps': 48009, 'loss/train': 1.3062742948532104} 02/26/2022 15:46:29 - INFO - codeparrot_training - Step 48010: {'lr': 2.117480447748066e-06, 'samples': 24581632, 'steps': 48010, 'loss/train': 3.3507866859436035} 02/26/2022 15:46:33 - INFO - codeparrot_training - Step 48011: {'lr': 2.1153558668927375e-06, 'samples': 24582144, 'steps': 48011, 'loss/train': 1.4072787761688232} 02/26/2022 15:46:39 - INFO - codeparrot_training - Step 48012: {'lr': 2.1132323478965374e-06, 'samples': 24582656, 'steps': 48012, 'loss/train': 1.1705113649368286} 02/26/2022 15:46:42 - INFO - codeparrot_training - Step 48013: {'lr': 2.1111098907685144e-06, 'samples': 24583168, 'steps': 48013, 'loss/train': 0.9358199238777161} 02/26/2022 15:46:48 - INFO - codeparrot_training - Step 48014: {'lr': 2.1089884955177996e-06, 'samples': 24583680, 'steps': 48014, 'loss/train': 2.55591082572937} 02/26/2022 15:46:51 - INFO - codeparrot_training - Step 48015: {'lr': 2.1068681621534693e-06, 'samples': 24584192, 'steps': 48015, 'loss/train': 2.250408411026001} 02/26/2022 15:46:57 - INFO - codeparrot_training - Step 48016: {'lr': 2.1047488906845712e-06, 'samples': 24584704, 'steps': 48016, 'loss/train': 1.4187424182891846} 02/26/2022 15:47:00 - INFO - codeparrot_training - Step 48017: {'lr': 2.102630681120238e-06, 'samples': 24585216, 'steps': 48017, 'loss/train': 1.6364868879318237} 02/26/2022 15:47:06 - INFO - codeparrot_training - Step 48018: {'lr': 2.100513533469517e-06, 'samples': 24585728, 'steps': 48018, 'loss/train': 1.7462069988250732} 02/26/2022 15:47:09 - INFO - codeparrot_training - Step 48019: {'lr': 2.098397447741485e-06, 'samples': 24586240, 'steps': 48019, 'loss/train': 1.7926084995269775} 02/26/2022 15:47:15 - INFO - codeparrot_training - Step 48020: {'lr': 2.0962824239451893e-06, 'samples': 24586752, 'steps': 48020, 'loss/train': 1.2424486875534058} 02/26/2022 15:47:18 - INFO - codeparrot_training - Step 48021: {'lr': 2.094168462089707e-06, 'samples': 24587264, 'steps': 48021, 'loss/train': 2.1140058040618896} 02/26/2022 15:47:24 - INFO - codeparrot_training - Step 48022: {'lr': 2.0920555621840588e-06, 'samples': 24587776, 'steps': 48022, 'loss/train': 2.862093687057495} 02/26/2022 15:47:28 - INFO - codeparrot_training - Step 48023: {'lr': 2.0899437242373754e-06, 'samples': 24588288, 'steps': 48023, 'loss/train': 2.104945182800293} 02/26/2022 15:47:33 - INFO - codeparrot_training - Step 48024: {'lr': 2.0878329482586223e-06, 'samples': 24588800, 'steps': 48024, 'loss/train': 1.7665393352508545} 02/26/2022 15:47:37 - INFO - codeparrot_training - Step 48025: {'lr': 2.0857232342568754e-06, 'samples': 24589312, 'steps': 48025, 'loss/train': 1.6297985315322876} 02/26/2022 15:47:44 - INFO - codeparrot_training - Step 48026: {'lr': 2.083614582241156e-06, 'samples': 24589824, 'steps': 48026, 'loss/train': 1.2695255279541016} 02/26/2022 15:47:47 - INFO - codeparrot_training - Step 48027: {'lr': 2.081506992220539e-06, 'samples': 24590336, 'steps': 48027, 'loss/train': 1.9487683773040771} 02/26/2022 15:47:53 - INFO - codeparrot_training - Step 48028: {'lr': 2.07940046420399e-06, 'samples': 24590848, 'steps': 48028, 'loss/train': 1.2654321193695068} 02/26/2022 15:47:57 - INFO - codeparrot_training - Step 48029: {'lr': 2.0772949982005853e-06, 'samples': 24591360, 'steps': 48029, 'loss/train': 1.6626259088516235} 02/26/2022 15:48:02 - INFO - codeparrot_training - Step 48030: {'lr': 2.075190594219317e-06, 'samples': 24591872, 'steps': 48030, 'loss/train': 1.4501103162765503} 02/26/2022 15:48:06 - INFO - codeparrot_training - Step 48031: {'lr': 2.073087252269207e-06, 'samples': 24592384, 'steps': 48031, 'loss/train': 1.3885493278503418} 02/26/2022 15:48:11 - INFO - codeparrot_training - Step 48032: {'lr': 2.0709849723593023e-06, 'samples': 24592896, 'steps': 48032, 'loss/train': 1.429519772529602} 02/26/2022 15:48:15 - INFO - codeparrot_training - Step 48033: {'lr': 2.0688837544985407e-06, 'samples': 24593408, 'steps': 48033, 'loss/train': 2.0424609184265137} 02/26/2022 15:48:20 - INFO - codeparrot_training - Step 48034: {'lr': 2.066783598695943e-06, 'samples': 24593920, 'steps': 48034, 'loss/train': 1.3377666473388672} 02/26/2022 15:48:24 - INFO - codeparrot_training - Step 48035: {'lr': 2.064684504960529e-06, 'samples': 24594432, 'steps': 48035, 'loss/train': 2.2199840545654297} 02/26/2022 15:48:29 - INFO - codeparrot_training - Step 48036: {'lr': 2.0625864733012645e-06, 'samples': 24594944, 'steps': 48036, 'loss/train': 2.0081417560577393} 02/26/2022 15:48:33 - INFO - codeparrot_training - Step 48037: {'lr': 2.06048950372717e-06, 'samples': 24595456, 'steps': 48037, 'loss/train': 1.405119776725769} 02/26/2022 15:48:40 - INFO - codeparrot_training - Step 48038: {'lr': 2.05839359624721e-06, 'samples': 24595968, 'steps': 48038, 'loss/train': 1.1608302593231201} 02/26/2022 15:48:43 - INFO - codeparrot_training - Step 48039: {'lr': 2.05629875087035e-06, 'samples': 24596480, 'steps': 48039, 'loss/train': 2.286485433578491} 02/26/2022 15:48:49 - INFO - codeparrot_training - Step 48040: {'lr': 2.054204967605583e-06, 'samples': 24596992, 'steps': 48040, 'loss/train': 2.239549160003662} 02/26/2022 15:48:52 - INFO - codeparrot_training - Step 48041: {'lr': 2.0521122464618735e-06, 'samples': 24597504, 'steps': 48041, 'loss/train': 1.8396660089492798} 02/26/2022 15:48:58 - INFO - codeparrot_training - Step 48042: {'lr': 2.0500205874481593e-06, 'samples': 24598016, 'steps': 48042, 'loss/train': 1.7345612049102783} 02/26/2022 15:49:01 - INFO - codeparrot_training - Step 48043: {'lr': 2.047929990573433e-06, 'samples': 24598528, 'steps': 48043, 'loss/train': 0.44698697328567505} 02/26/2022 15:49:07 - INFO - codeparrot_training - Step 48044: {'lr': 2.0458404558466593e-06, 'samples': 24599040, 'steps': 48044, 'loss/train': 2.0572776794433594} 02/26/2022 15:49:10 - INFO - codeparrot_training - Step 48045: {'lr': 2.0437519832767483e-06, 'samples': 24599552, 'steps': 48045, 'loss/train': 1.5602202415466309} 02/26/2022 15:49:16 - INFO - codeparrot_training - Step 48046: {'lr': 2.0416645728726922e-06, 'samples': 24600064, 'steps': 48046, 'loss/train': 2.6575729846954346} 02/26/2022 15:49:19 - INFO - codeparrot_training - Step 48047: {'lr': 2.0395782246433737e-06, 'samples': 24600576, 'steps': 48047, 'loss/train': 0.8888649344444275} 02/26/2022 15:49:27 - INFO - codeparrot_training - Step 48048: {'lr': 2.0374929385978125e-06, 'samples': 24601088, 'steps': 48048, 'loss/train': 1.738717794418335} 02/26/2022 15:49:30 - INFO - codeparrot_training - Step 48049: {'lr': 2.035408714744863e-06, 'samples': 24601600, 'steps': 48049, 'loss/train': 0.8226458430290222} 02/26/2022 15:49:36 - INFO - codeparrot_training - Step 48050: {'lr': 2.0333255530934903e-06, 'samples': 24602112, 'steps': 48050, 'loss/train': 1.8816875219345093} 02/26/2022 15:49:39 - INFO - codeparrot_training - Step 48051: {'lr': 2.031243453652604e-06, 'samples': 24602624, 'steps': 48051, 'loss/train': 2.0343282222747803} 02/26/2022 15:49:45 - INFO - codeparrot_training - Step 48052: {'lr': 2.0291624164311686e-06, 'samples': 24603136, 'steps': 48052, 'loss/train': 1.7116533517837524} 02/26/2022 15:49:48 - INFO - codeparrot_training - Step 48053: {'lr': 2.0270824414380107e-06, 'samples': 24603648, 'steps': 48053, 'loss/train': 1.1586358547210693} 02/26/2022 15:49:54 - INFO - codeparrot_training - Step 48054: {'lr': 2.025003528682151e-06, 'samples': 24604160, 'steps': 48054, 'loss/train': 2.3007030487060547} 02/26/2022 15:49:57 - INFO - codeparrot_training - Step 48055: {'lr': 2.022925678172388e-06, 'samples': 24604672, 'steps': 48055, 'loss/train': 2.2319235801696777} 02/26/2022 15:50:03 - INFO - codeparrot_training - Step 48056: {'lr': 2.020848889917687e-06, 'samples': 24605184, 'steps': 48056, 'loss/train': 1.3865286111831665} 02/26/2022 15:50:06 - INFO - codeparrot_training - Step 48057: {'lr': 2.018773163926901e-06, 'samples': 24605696, 'steps': 48057, 'loss/train': 1.8625080585479736} 02/26/2022 15:50:14 - INFO - codeparrot_training - Step 48058: {'lr': 2.0166985002089967e-06, 'samples': 24606208, 'steps': 48058, 'loss/train': 1.929304599761963} 02/26/2022 15:50:17 - INFO - codeparrot_training - Step 48059: {'lr': 2.014624898772771e-06, 'samples': 24606720, 'steps': 48059, 'loss/train': 1.9097601175308228} 02/26/2022 15:50:23 - INFO - codeparrot_training - Step 48060: {'lr': 2.0125523596271623e-06, 'samples': 24607232, 'steps': 48060, 'loss/train': 1.6167644262313843} 02/26/2022 15:50:26 - INFO - codeparrot_training - Step 48061: {'lr': 2.010480882781024e-06, 'samples': 24607744, 'steps': 48061, 'loss/train': 1.1160197257995605} 02/26/2022 15:50:32 - INFO - codeparrot_training - Step 48062: {'lr': 2.0084104682432657e-06, 'samples': 24608256, 'steps': 48062, 'loss/train': 2.3234784603118896} 02/26/2022 15:50:36 - INFO - codeparrot_training - Step 48063: {'lr': 2.0063411160227143e-06, 'samples': 24608768, 'steps': 48063, 'loss/train': 3.278146982192993} 02/26/2022 15:50:41 - INFO - codeparrot_training - Step 48064: {'lr': 2.0042728261282504e-06, 'samples': 24609280, 'steps': 48064, 'loss/train': 1.760576605796814} 02/26/2022 15:50:45 - INFO - codeparrot_training - Step 48065: {'lr': 2.002205598568729e-06, 'samples': 24609792, 'steps': 48065, 'loss/train': 1.1821355819702148} 02/26/2022 15:50:50 - INFO - codeparrot_training - Step 48066: {'lr': 2.0001394333530322e-06, 'samples': 24610304, 'steps': 48066, 'loss/train': 0.029472487047314644} 02/26/2022 15:50:54 - INFO - codeparrot_training - Step 48067: {'lr': 1.998074330489957e-06, 'samples': 24610816, 'steps': 48067, 'loss/train': 2.4676690101623535} 02/26/2022 15:50:59 - INFO - codeparrot_training - Step 48068: {'lr': 1.9960102899884146e-06, 'samples': 24611328, 'steps': 48068, 'loss/train': 1.6439640522003174} 02/26/2022 15:51:03 - INFO - codeparrot_training - Step 48069: {'lr': 1.9939473118571748e-06, 'samples': 24611840, 'steps': 48069, 'loss/train': 2.1548399925231934} 02/26/2022 15:51:08 - INFO - codeparrot_training - Step 48070: {'lr': 1.9918853961051476e-06, 'samples': 24612352, 'steps': 48070, 'loss/train': 2.229962110519409} 02/26/2022 15:51:12 - INFO - codeparrot_training - Step 48071: {'lr': 1.989824542741131e-06, 'samples': 24612864, 'steps': 48071, 'loss/train': 1.1238858699798584} 02/26/2022 15:51:17 - INFO - codeparrot_training - Step 48072: {'lr': 1.9877647517739518e-06, 'samples': 24613376, 'steps': 48072, 'loss/train': 2.0451135635375977} 02/26/2022 15:51:21 - INFO - codeparrot_training - Step 48073: {'lr': 1.985706023212408e-06, 'samples': 24613888, 'steps': 48073, 'loss/train': 0.42863568663597107} 02/26/2022 15:51:28 - INFO - codeparrot_training - Step 48074: {'lr': 1.983648357065382e-06, 'samples': 24614400, 'steps': 48074, 'loss/train': 1.0816551446914673} 02/26/2022 15:51:32 - INFO - codeparrot_training - Step 48075: {'lr': 1.981591753341644e-06, 'samples': 24614912, 'steps': 48075, 'loss/train': 2.251687526702881} 02/26/2022 15:51:35 - INFO - codeparrot_training - Step 48076: {'lr': 1.9795362120499927e-06, 'samples': 24615424, 'steps': 48076, 'loss/train': 2.5053560733795166} 02/26/2022 15:51:41 - INFO - codeparrot_training - Step 48077: {'lr': 1.9774817331992546e-06, 'samples': 24615936, 'steps': 48077, 'loss/train': 1.9417693614959717} 02/26/2022 15:51:44 - INFO - codeparrot_training - Step 48078: {'lr': 1.9754283167982557e-06, 'samples': 24616448, 'steps': 48078, 'loss/train': 1.716273546218872} 02/26/2022 15:51:50 - INFO - codeparrot_training - Step 48079: {'lr': 1.973375962855767e-06, 'samples': 24616960, 'steps': 48079, 'loss/train': 1.5255002975463867} 02/26/2022 15:51:53 - INFO - codeparrot_training - Step 48080: {'lr': 1.9713246713805587e-06, 'samples': 24617472, 'steps': 48080, 'loss/train': 0.9694469571113586} 02/26/2022 15:51:59 - INFO - codeparrot_training - Step 48081: {'lr': 1.96927444238143e-06, 'samples': 24617984, 'steps': 48081, 'loss/train': 1.755071997642517} 02/26/2022 15:52:02 - INFO - codeparrot_training - Step 48082: {'lr': 1.9672252758671795e-06, 'samples': 24618496, 'steps': 48082, 'loss/train': 1.8848768472671509} 02/26/2022 15:52:09 - INFO - codeparrot_training - Step 48083: {'lr': 1.9651771718465774e-06, 'samples': 24619008, 'steps': 48083, 'loss/train': 1.168839931488037} 02/26/2022 15:52:13 - INFO - codeparrot_training - Step 48084: {'lr': 1.9631301303283944e-06, 'samples': 24619520, 'steps': 48084, 'loss/train': 1.6247429847717285} 02/26/2022 15:52:18 - INFO - codeparrot_training - Step 48085: {'lr': 1.9610841513214015e-06, 'samples': 24620032, 'steps': 48085, 'loss/train': 2.532076835632324} 02/26/2022 15:52:21 - INFO - codeparrot_training - Step 48086: {'lr': 1.95903923483437e-06, 'samples': 24620544, 'steps': 48086, 'loss/train': 1.5146517753601074} 02/26/2022 15:52:27 - INFO - codeparrot_training - Step 48087: {'lr': 1.956995380876042e-06, 'samples': 24621056, 'steps': 48087, 'loss/train': 2.0462818145751953} 02/26/2022 15:52:30 - INFO - codeparrot_training - Step 48088: {'lr': 1.954952589455189e-06, 'samples': 24621568, 'steps': 48088, 'loss/train': 0.31213557720184326} 02/26/2022 15:52:36 - INFO - codeparrot_training - Step 48089: {'lr': 1.9529108605805534e-06, 'samples': 24622080, 'steps': 48089, 'loss/train': 0.6354958415031433} 02/26/2022 15:52:39 - INFO - codeparrot_training - Step 48090: {'lr': 1.9508701942609064e-06, 'samples': 24622592, 'steps': 48090, 'loss/train': 1.7278589010238647} 02/26/2022 15:52:45 - INFO - codeparrot_training - Step 48091: {'lr': 1.948830590504935e-06, 'samples': 24623104, 'steps': 48091, 'loss/train': 1.6710814237594604} 02/26/2022 15:52:48 - INFO - codeparrot_training - Step 48092: {'lr': 1.946792049321411e-06, 'samples': 24623616, 'steps': 48092, 'loss/train': 1.0024529695510864} 02/26/2022 15:52:54 - INFO - codeparrot_training - Step 48093: {'lr': 1.944754570719076e-06, 'samples': 24624128, 'steps': 48093, 'loss/train': 1.8819817304611206} 02/26/2022 15:52:58 - INFO - codeparrot_training - Step 48094: {'lr': 1.942718154706646e-06, 'samples': 24624640, 'steps': 48094, 'loss/train': 2.11812424659729} 02/26/2022 15:53:03 - INFO - codeparrot_training - Step 48095: {'lr': 1.9406828012928647e-06, 'samples': 24625152, 'steps': 48095, 'loss/train': 1.1509935855865479} 02/26/2022 15:53:07 - INFO - codeparrot_training - Step 48096: {'lr': 1.9386485104863904e-06, 'samples': 24625664, 'steps': 48096, 'loss/train': 2.1935925483703613} 02/26/2022 15:53:12 - INFO - codeparrot_training - Step 48097: {'lr': 1.9366152822959948e-06, 'samples': 24626176, 'steps': 48097, 'loss/train': 1.5797220468521118} 02/26/2022 15:53:16 - INFO - codeparrot_training - Step 48098: {'lr': 1.934583116730365e-06, 'samples': 24626688, 'steps': 48098, 'loss/train': 2.4963557720184326} 02/26/2022 15:53:21 - INFO - codeparrot_training - Step 48099: {'lr': 1.9325520137982166e-06, 'samples': 24627200, 'steps': 48099, 'loss/train': 0.6718520522117615} 02/26/2022 15:53:27 - INFO - codeparrot_training - Step 48100: {'lr': 1.930521973508237e-06, 'samples': 24627712, 'steps': 48100, 'loss/train': 1.710747241973877} 02/26/2022 15:53:30 - INFO - codeparrot_training - Step 48101: {'lr': 1.928492995869141e-06, 'samples': 24628224, 'steps': 48101, 'loss/train': 1.610666036605835} 02/26/2022 15:53:36 - INFO - codeparrot_training - Step 48102: {'lr': 1.9264650808896167e-06, 'samples': 24628736, 'steps': 48102, 'loss/train': 1.7758033275604248} 02/26/2022 15:53:39 - INFO - codeparrot_training - Step 48103: {'lr': 1.9244382285783515e-06, 'samples': 24629248, 'steps': 48103, 'loss/train': 0.4534592032432556} 02/26/2022 15:53:43 - INFO - codeparrot_training - Step 48104: {'lr': 1.9224124389439767e-06, 'samples': 24629760, 'steps': 48104, 'loss/train': 1.828382134437561} 02/26/2022 15:53:49 - INFO - codeparrot_training - Step 48105: {'lr': 1.920387711995236e-06, 'samples': 24630272, 'steps': 48105, 'loss/train': 1.3766778707504272} 02/26/2022 15:53:53 - INFO - codeparrot_training - Step 48106: {'lr': 1.9183640477407894e-06, 'samples': 24630784, 'steps': 48106, 'loss/train': 0.21424119174480438} 02/26/2022 15:53:58 - INFO - codeparrot_training - Step 48107: {'lr': 1.9163414461892957e-06, 'samples': 24631296, 'steps': 48107, 'loss/train': 0.5620800256729126} 02/26/2022 15:54:02 - INFO - codeparrot_training - Step 48108: {'lr': 1.9143199073494154e-06, 'samples': 24631808, 'steps': 48108, 'loss/train': 2.1171562671661377} 02/26/2022 15:54:07 - INFO - codeparrot_training - Step 48109: {'lr': 1.912299431229808e-06, 'samples': 24632320, 'steps': 48109, 'loss/train': 1.9083143472671509} 02/26/2022 15:54:11 - INFO - codeparrot_training - Step 48110: {'lr': 1.9102800178391335e-06, 'samples': 24632832, 'steps': 48110, 'loss/train': 2.0866315364837646} 02/26/2022 15:54:16 - INFO - codeparrot_training - Step 48111: {'lr': 1.9082616671860508e-06, 'samples': 24633344, 'steps': 48111, 'loss/train': 2.3169713020324707} 02/26/2022 15:54:20 - INFO - codeparrot_training - Step 48112: {'lr': 1.9062443792791929e-06, 'samples': 24633856, 'steps': 48112, 'loss/train': 1.7829821109771729} 02/26/2022 15:54:25 - INFO - codeparrot_training - Step 48113: {'lr': 1.9042281541272188e-06, 'samples': 24634368, 'steps': 48113, 'loss/train': 1.5835213661193848} 02/26/2022 15:54:29 - INFO - codeparrot_training - Step 48114: {'lr': 1.9022129917387055e-06, 'samples': 24634880, 'steps': 48114, 'loss/train': 8.325510025024414} 02/26/2022 15:54:35 - INFO - codeparrot_training - Step 48115: {'lr': 1.9001988921223957e-06, 'samples': 24635392, 'steps': 48115, 'loss/train': 2.0862507820129395} 02/26/2022 15:54:38 - INFO - codeparrot_training - Step 48116: {'lr': 1.8981858552868104e-06, 'samples': 24635904, 'steps': 48116, 'loss/train': 1.3885626792907715} 02/26/2022 15:54:44 - INFO - codeparrot_training - Step 48117: {'lr': 1.8961738812406372e-06, 'samples': 24636416, 'steps': 48117, 'loss/train': 1.9017176628112793} 02/26/2022 15:54:49 - INFO - codeparrot_training - Step 48118: {'lr': 1.8941629699924522e-06, 'samples': 24636928, 'steps': 48118, 'loss/train': 1.3964951038360596} 02/26/2022 15:54:53 - INFO - codeparrot_training - Step 48119: {'lr': 1.8921531215509157e-06, 'samples': 24637440, 'steps': 48119, 'loss/train': 0.062121689319610596} 02/26/2022 15:54:58 - INFO - codeparrot_training - Step 48120: {'lr': 1.8901443359245762e-06, 'samples': 24637952, 'steps': 48120, 'loss/train': 1.5635122060775757} 02/26/2022 15:55:02 - INFO - codeparrot_training - Step 48121: {'lr': 1.8881366131220934e-06, 'samples': 24638464, 'steps': 48121, 'loss/train': 2.294280529022217} 02/26/2022 15:55:07 - INFO - codeparrot_training - Step 48122: {'lr': 1.8861299531520436e-06, 'samples': 24638976, 'steps': 48122, 'loss/train': 1.6215163469314575} 02/26/2022 15:55:11 - INFO - codeparrot_training - Step 48123: {'lr': 1.8841243560230591e-06, 'samples': 24639488, 'steps': 48123, 'loss/train': 2.1305184364318848} 02/26/2022 15:55:16 - INFO - codeparrot_training - Step 48124: {'lr': 1.8821198217436608e-06, 'samples': 24640000, 'steps': 48124, 'loss/train': 1.4907678365707397} 02/26/2022 15:55:20 - INFO - codeparrot_training - Step 48125: {'lr': 1.8801163503225082e-06, 'samples': 24640512, 'steps': 48125, 'loss/train': 0.9205184578895569} 02/26/2022 15:55:25 - INFO - codeparrot_training - Step 48126: {'lr': 1.8781139417681226e-06, 'samples': 24641024, 'steps': 48126, 'loss/train': 0.487720787525177} 02/26/2022 15:55:29 - INFO - codeparrot_training - Step 48127: {'lr': 1.876112596089108e-06, 'samples': 24641536, 'steps': 48127, 'loss/train': 5.531866073608398} 02/26/2022 15:55:34 - INFO - codeparrot_training - Step 48128: {'lr': 1.8741123132940685e-06, 'samples': 24642048, 'steps': 48128, 'loss/train': 1.314464807510376} 02/26/2022 15:55:38 - INFO - codeparrot_training - Step 48129: {'lr': 1.8721130933915253e-06, 'samples': 24642560, 'steps': 48129, 'loss/train': 0.5421221256256104} 02/26/2022 15:55:44 - INFO - codeparrot_training - Step 48130: {'lr': 1.870114936390055e-06, 'samples': 24643072, 'steps': 48130, 'loss/train': 1.758766531944275} 02/26/2022 15:55:48 - INFO - codeparrot_training - Step 48131: {'lr': 1.8681178422982336e-06, 'samples': 24643584, 'steps': 48131, 'loss/train': 2.070037603378296} 02/26/2022 15:55:53 - INFO - codeparrot_training - Step 48132: {'lr': 1.8661218111246102e-06, 'samples': 24644096, 'steps': 48132, 'loss/train': 1.9842796325683594} 02/26/2022 15:55:57 - INFO - codeparrot_training - Step 48133: {'lr': 1.8641268428777057e-06, 'samples': 24644608, 'steps': 48133, 'loss/train': 1.1908373832702637} 02/26/2022 15:56:03 - INFO - codeparrot_training - Step 48134: {'lr': 1.8621329375660968e-06, 'samples': 24645120, 'steps': 48134, 'loss/train': 1.1644662618637085} 02/26/2022 15:56:06 - INFO - codeparrot_training - Step 48135: {'lr': 1.8601400951983316e-06, 'samples': 24645632, 'steps': 48135, 'loss/train': 2.710977077484131} 02/26/2022 15:56:12 - INFO - codeparrot_training - Step 48136: {'lr': 1.8581483157829315e-06, 'samples': 24646144, 'steps': 48136, 'loss/train': 0.1924516260623932} 02/26/2022 15:56:15 - INFO - codeparrot_training - Step 48137: {'lr': 1.8561575993284175e-06, 'samples': 24646656, 'steps': 48137, 'loss/train': 2.6672000885009766} 02/26/2022 15:56:21 - INFO - codeparrot_training - Step 48138: {'lr': 1.854167945843338e-06, 'samples': 24647168, 'steps': 48138, 'loss/train': 1.7876639366149902} 02/26/2022 15:56:24 - INFO - codeparrot_training - Step 48139: {'lr': 1.8521793553361865e-06, 'samples': 24647680, 'steps': 48139, 'loss/train': 2.317499876022339} 02/26/2022 15:56:31 - INFO - codeparrot_training - Step 48140: {'lr': 1.8501918278155394e-06, 'samples': 24648192, 'steps': 48140, 'loss/train': 1.517192006111145} 02/26/2022 15:56:34 - INFO - codeparrot_training - Step 48141: {'lr': 1.848205363289862e-06, 'samples': 24648704, 'steps': 48141, 'loss/train': 0.8325765132904053} 02/26/2022 15:56:40 - INFO - codeparrot_training - Step 48142: {'lr': 1.8462199617676478e-06, 'samples': 24649216, 'steps': 48142, 'loss/train': 0.8042821884155273} 02/26/2022 15:56:43 - INFO - codeparrot_training - Step 48143: {'lr': 1.8442356232574453e-06, 'samples': 24649728, 'steps': 48143, 'loss/train': 1.657462477684021} 02/26/2022 15:56:48 - INFO - codeparrot_training - Step 48144: {'lr': 1.8422523477677477e-06, 'samples': 24650240, 'steps': 48144, 'loss/train': 0.9021700024604797} 02/26/2022 15:56:52 - INFO - codeparrot_training - Step 48145: {'lr': 1.8402701353070483e-06, 'samples': 24650752, 'steps': 48145, 'loss/train': 1.8635834455490112} 02/26/2022 15:56:58 - INFO - codeparrot_training - Step 48146: {'lr': 1.8382889858838403e-06, 'samples': 24651264, 'steps': 48146, 'loss/train': 2.2870888710021973} 02/26/2022 15:57:01 - INFO - codeparrot_training - Step 48147: {'lr': 1.8363088995065614e-06, 'samples': 24651776, 'steps': 48147, 'loss/train': 2.040375232696533} 02/26/2022 15:57:07 - INFO - codeparrot_training - Step 48148: {'lr': 1.8343298761837601e-06, 'samples': 24652288, 'steps': 48148, 'loss/train': 0.05891959369182587} 02/26/2022 15:57:10 - INFO - codeparrot_training - Step 48149: {'lr': 1.8323519159239022e-06, 'samples': 24652800, 'steps': 48149, 'loss/train': 2.134617328643799} 02/26/2022 15:57:16 - INFO - codeparrot_training - Step 48150: {'lr': 1.830375018735425e-06, 'samples': 24653312, 'steps': 48150, 'loss/train': 2.1146655082702637} 02/26/2022 15:57:20 - INFO - codeparrot_training - Step 48151: {'lr': 1.8283991846268221e-06, 'samples': 24653824, 'steps': 48151, 'loss/train': 0.5072862505912781} 02/26/2022 15:57:25 - INFO - codeparrot_training - Step 48152: {'lr': 1.826424413606559e-06, 'samples': 24654336, 'steps': 48152, 'loss/train': 1.6900368928909302} 02/26/2022 15:57:29 - INFO - codeparrot_training - Step 48153: {'lr': 1.8244507056830729e-06, 'samples': 24654848, 'steps': 48153, 'loss/train': 2.5450944900512695} 02/26/2022 15:57:34 - INFO - codeparrot_training - Step 48154: {'lr': 1.8224780608648572e-06, 'samples': 24655360, 'steps': 48154, 'loss/train': 1.4937857389450073} 02/26/2022 15:57:38 - INFO - codeparrot_training - Step 48155: {'lr': 1.820506479160322e-06, 'samples': 24655872, 'steps': 48155, 'loss/train': 1.6144745349884033} 02/26/2022 15:57:43 - INFO - codeparrot_training - Step 48156: {'lr': 1.8185359605779605e-06, 'samples': 24656384, 'steps': 48156, 'loss/train': 1.5675196647644043} 02/26/2022 15:57:47 - INFO - codeparrot_training - Step 48157: {'lr': 1.8165665051261548e-06, 'samples': 24656896, 'steps': 48157, 'loss/train': 1.4855611324310303} 02/26/2022 15:57:52 - INFO - codeparrot_training - Step 48158: {'lr': 1.8145981128133982e-06, 'samples': 24657408, 'steps': 48158, 'loss/train': 1.4053477048873901} 02/26/2022 15:57:56 - INFO - codeparrot_training - Step 48159: {'lr': 1.812630783648045e-06, 'samples': 24657920, 'steps': 48159, 'loss/train': 1.4907281398773193} 02/26/2022 15:58:02 - INFO - codeparrot_training - Step 48160: {'lr': 1.810664517638616e-06, 'samples': 24658432, 'steps': 48160, 'loss/train': 1.9743613004684448} 02/26/2022 15:58:05 - INFO - codeparrot_training - Step 48161: {'lr': 1.8086993147934661e-06, 'samples': 24658944, 'steps': 48161, 'loss/train': 2.0596156120300293} 02/26/2022 15:58:11 - INFO - codeparrot_training - Step 48162: {'lr': 1.8067351751210603e-06, 'samples': 24659456, 'steps': 48162, 'loss/train': 2.3782312870025635} 02/26/2022 15:58:14 - INFO - codeparrot_training - Step 48163: {'lr': 1.8047720986298089e-06, 'samples': 24659968, 'steps': 48163, 'loss/train': 1.170705795288086} 02/26/2022 15:58:20 - INFO - codeparrot_training - Step 48164: {'lr': 1.802810085328066e-06, 'samples': 24660480, 'steps': 48164, 'loss/train': 1.574622631072998} 02/26/2022 15:58:23 - INFO - codeparrot_training - Step 48165: {'lr': 1.8008491352242696e-06, 'samples': 24660992, 'steps': 48165, 'loss/train': 2.717625379562378} 02/26/2022 15:58:29 - INFO - codeparrot_training - Step 48166: {'lr': 1.798889248326857e-06, 'samples': 24661504, 'steps': 48166, 'loss/train': 1.588113784790039} 02/26/2022 15:58:32 - INFO - codeparrot_training - Step 48167: {'lr': 1.796930424644183e-06, 'samples': 24662016, 'steps': 48167, 'loss/train': 2.857224225997925} 02/26/2022 15:58:38 - INFO - codeparrot_training - Step 48168: {'lr': 1.7949726641846298e-06, 'samples': 24662528, 'steps': 48168, 'loss/train': 2.3585195541381836} 02/26/2022 15:58:41 - INFO - codeparrot_training - Step 48169: {'lr': 1.7930159669566072e-06, 'samples': 24663040, 'steps': 48169, 'loss/train': 1.2491806745529175} 02/26/2022 15:58:48 - INFO - codeparrot_training - Step 48170: {'lr': 1.7910603329684693e-06, 'samples': 24663552, 'steps': 48170, 'loss/train': 2.4087884426116943} 02/26/2022 15:58:51 - INFO - codeparrot_training - Step 48171: {'lr': 1.7891057622286543e-06, 'samples': 24664064, 'steps': 48171, 'loss/train': 1.4979145526885986} 02/26/2022 15:58:57 - INFO - codeparrot_training - Step 48172: {'lr': 1.787152254745461e-06, 'samples': 24664576, 'steps': 48172, 'loss/train': 1.2991231679916382} 02/26/2022 15:59:00 - INFO - codeparrot_training - Step 48173: {'lr': 1.7851998105272992e-06, 'samples': 24665088, 'steps': 48173, 'loss/train': 0.7050226926803589} 02/26/2022 15:59:06 - INFO - codeparrot_training - Step 48174: {'lr': 1.7832484295825236e-06, 'samples': 24665600, 'steps': 48174, 'loss/train': 1.8083151578903198} 02/26/2022 15:59:09 - INFO - codeparrot_training - Step 48175: {'lr': 1.781298111919516e-06, 'samples': 24666112, 'steps': 48175, 'loss/train': 1.4875048398971558} 02/26/2022 15:59:15 - INFO - codeparrot_training - Step 48176: {'lr': 1.7793488575466032e-06, 'samples': 24666624, 'steps': 48176, 'loss/train': 2.6878955364227295} 02/26/2022 15:59:18 - INFO - codeparrot_training - Step 48177: {'lr': 1.777400666472112e-06, 'samples': 24667136, 'steps': 48177, 'loss/train': 2.1547327041625977} 02/26/2022 15:59:24 - INFO - codeparrot_training - Step 48178: {'lr': 1.7754535387044246e-06, 'samples': 24667648, 'steps': 48178, 'loss/train': 1.327903389930725} 02/26/2022 15:59:28 - INFO - codeparrot_training - Step 48179: {'lr': 1.7735074742518954e-06, 'samples': 24668160, 'steps': 48179, 'loss/train': 2.301586389541626} 02/26/2022 15:59:31 - INFO - codeparrot_training - Step 48180: {'lr': 1.771562473122823e-06, 'samples': 24668672, 'steps': 48180, 'loss/train': 1.8878509998321533} 02/26/2022 15:59:37 - INFO - codeparrot_training - Step 48181: {'lr': 1.7696185353255624e-06, 'samples': 24669184, 'steps': 48181, 'loss/train': 0.4387229084968567} 02/26/2022 15:59:40 - INFO - codeparrot_training - Step 48182: {'lr': 1.7676756608684396e-06, 'samples': 24669696, 'steps': 48182, 'loss/train': 1.2409852743148804} 02/26/2022 15:59:46 - INFO - codeparrot_training - Step 48183: {'lr': 1.7657338497597542e-06, 'samples': 24670208, 'steps': 48183, 'loss/train': 1.322680950164795} 02/26/2022 15:59:49 - INFO - codeparrot_training - Step 48184: {'lr': 1.7637931020078602e-06, 'samples': 24670720, 'steps': 48184, 'loss/train': 1.5506795644760132} 02/26/2022 15:59:55 - INFO - codeparrot_training - Step 48185: {'lr': 1.7618534176210288e-06, 'samples': 24671232, 'steps': 48185, 'loss/train': 1.3698604106903076} 02/26/2022 16:00:01 - INFO - codeparrot_training - Step 48186: {'lr': 1.7599147966076146e-06, 'samples': 24671744, 'steps': 48186, 'loss/train': 1.5163310766220093} 02/26/2022 16:00:04 - INFO - codeparrot_training - Step 48187: {'lr': 1.7579772389758886e-06, 'samples': 24672256, 'steps': 48187, 'loss/train': 1.2335726022720337} 02/26/2022 16:00:10 - INFO - codeparrot_training - Step 48188: {'lr': 1.7560407447341497e-06, 'samples': 24672768, 'steps': 48188, 'loss/train': 1.682826280593872} 02/26/2022 16:00:13 - INFO - codeparrot_training - Step 48189: {'lr': 1.7541053138907525e-06, 'samples': 24673280, 'steps': 48189, 'loss/train': 1.3601185083389282} 02/26/2022 16:00:19 - INFO - codeparrot_training - Step 48190: {'lr': 1.7521709464539127e-06, 'samples': 24673792, 'steps': 48190, 'loss/train': 1.3802342414855957} 02/26/2022 16:00:22 - INFO - codeparrot_training - Step 48191: {'lr': 1.750237642431929e-06, 'samples': 24674304, 'steps': 48191, 'loss/train': 0.38658061623573303} 02/26/2022 16:00:28 - INFO - codeparrot_training - Step 48192: {'lr': 1.7483054018331002e-06, 'samples': 24674816, 'steps': 48192, 'loss/train': 1.097922921180725} 02/26/2022 16:00:32 - INFO - codeparrot_training - Step 48193: {'lr': 1.7463742246657533e-06, 'samples': 24675328, 'steps': 48193, 'loss/train': 1.7961775064468384} 02/26/2022 16:00:35 - INFO - codeparrot_training - Step 48194: {'lr': 1.7444441109380482e-06, 'samples': 24675840, 'steps': 48194, 'loss/train': 1.5271315574645996} 02/26/2022 16:00:41 - INFO - codeparrot_training - Step 48195: {'lr': 1.7425150606583395e-06, 'samples': 24676352, 'steps': 48195, 'loss/train': 1.4275624752044678} 02/26/2022 16:00:44 - INFO - codeparrot_training - Step 48196: {'lr': 1.7405870738348705e-06, 'samples': 24676864, 'steps': 48196, 'loss/train': 1.9668711423873901} 02/26/2022 16:00:50 - INFO - codeparrot_training - Step 48197: {'lr': 1.7386601504758848e-06, 'samples': 24677376, 'steps': 48197, 'loss/train': 1.1421281099319458} 02/26/2022 16:00:54 - INFO - codeparrot_training - Step 48198: {'lr': 1.7367342905896532e-06, 'samples': 24677888, 'steps': 48198, 'loss/train': 1.576008677482605} 02/26/2022 16:00:59 - INFO - codeparrot_training - Step 48199: {'lr': 1.7348094941844194e-06, 'samples': 24678400, 'steps': 48199, 'loss/train': 1.7285265922546387} 02/26/2022 16:01:03 - INFO - codeparrot_training - Step 48200: {'lr': 1.7328857612684267e-06, 'samples': 24678912, 'steps': 48200, 'loss/train': 1.9396990537643433} 02/26/2022 16:01:08 - INFO - codeparrot_training - Step 48201: {'lr': 1.7309630918499187e-06, 'samples': 24679424, 'steps': 48201, 'loss/train': 1.681591272354126} 02/26/2022 16:01:12 - INFO - codeparrot_training - Step 48202: {'lr': 1.7290414859371383e-06, 'samples': 24679936, 'steps': 48202, 'loss/train': 1.8649177551269531} 02/26/2022 16:01:17 - INFO - codeparrot_training - Step 48203: {'lr': 1.7271209435383017e-06, 'samples': 24680448, 'steps': 48203, 'loss/train': 2.9391820430755615} 02/26/2022 16:01:23 - INFO - codeparrot_training - Step 48204: {'lr': 1.7252014646616242e-06, 'samples': 24680960, 'steps': 48204, 'loss/train': 1.167412281036377} 02/26/2022 16:01:26 - INFO - codeparrot_training - Step 48205: {'lr': 1.7232830493153774e-06, 'samples': 24681472, 'steps': 48205, 'loss/train': 2.2361602783203125} 02/26/2022 16:01:32 - INFO - codeparrot_training - Step 48206: {'lr': 1.7213656975077485e-06, 'samples': 24681984, 'steps': 48206, 'loss/train': 0.8090372681617737} 02/26/2022 16:01:36 - INFO - codeparrot_training - Step 48207: {'lr': 1.719449409246926e-06, 'samples': 24682496, 'steps': 48207, 'loss/train': 2.1550731658935547} 02/26/2022 16:01:41 - INFO - codeparrot_training - Step 48208: {'lr': 1.717534184541153e-06, 'samples': 24683008, 'steps': 48208, 'loss/train': 2.4501538276672363} 02/26/2022 16:01:45 - INFO - codeparrot_training - Step 48209: {'lr': 1.7156200233986453e-06, 'samples': 24683520, 'steps': 48209, 'loss/train': 1.6562236547470093} 02/26/2022 16:01:50 - INFO - codeparrot_training - Step 48210: {'lr': 1.7137069258275629e-06, 'samples': 24684032, 'steps': 48210, 'loss/train': 2.0163137912750244} 02/26/2022 16:01:54 - INFO - codeparrot_training - Step 48211: {'lr': 1.7117948918361493e-06, 'samples': 24684544, 'steps': 48211, 'loss/train': 0.8923929929733276} 02/26/2022 16:01:59 - INFO - codeparrot_training - Step 48212: {'lr': 1.709883921432509e-06, 'samples': 24685056, 'steps': 48212, 'loss/train': 0.3566322326660156} 02/26/2022 16:02:03 - INFO - codeparrot_training - Step 48213: {'lr': 1.707974014624941e-06, 'samples': 24685568, 'steps': 48213, 'loss/train': 1.1037888526916504} 02/26/2022 16:02:08 - INFO - codeparrot_training - Step 48214: {'lr': 1.7060651714215503e-06, 'samples': 24686080, 'steps': 48214, 'loss/train': 2.328897476196289} 02/26/2022 16:02:12 - INFO - codeparrot_training - Step 48215: {'lr': 1.7041573918305241e-06, 'samples': 24686592, 'steps': 48215, 'loss/train': 2.175065755844116} 02/26/2022 16:02:17 - INFO - codeparrot_training - Step 48216: {'lr': 1.7022506758600509e-06, 'samples': 24687104, 'steps': 48216, 'loss/train': 1.1348466873168945} 02/26/2022 16:02:21 - INFO - codeparrot_training - Step 48217: {'lr': 1.7003450235182904e-06, 'samples': 24687616, 'steps': 48217, 'loss/train': 1.7976552248001099} 02/26/2022 16:02:27 - INFO - codeparrot_training - Step 48218: {'lr': 1.6984404348134308e-06, 'samples': 24688128, 'steps': 48218, 'loss/train': 1.7065542936325073} 02/26/2022 16:02:30 - INFO - codeparrot_training - Step 48219: {'lr': 1.6965369097535766e-06, 'samples': 24688640, 'steps': 48219, 'loss/train': 1.7317028045654297} 02/26/2022 16:02:36 - INFO - codeparrot_training - Step 48220: {'lr': 1.6946344483469155e-06, 'samples': 24689152, 'steps': 48220, 'loss/train': 1.4688822031021118} 02/26/2022 16:02:39 - INFO - codeparrot_training - Step 48221: {'lr': 1.692733050601608e-06, 'samples': 24689664, 'steps': 48221, 'loss/train': 1.8245819807052612} 02/26/2022 16:02:46 - INFO - codeparrot_training - Step 48222: {'lr': 1.6908327165257865e-06, 'samples': 24690176, 'steps': 48222, 'loss/train': 1.0730267763137817} 02/26/2022 16:02:50 - INFO - codeparrot_training - Step 48223: {'lr': 1.6889334461275829e-06, 'samples': 24690688, 'steps': 48223, 'loss/train': 1.7888169288635254} 02/26/2022 16:02:53 - INFO - codeparrot_training - Step 48224: {'lr': 1.6870352394151579e-06, 'samples': 24691200, 'steps': 48224, 'loss/train': 0.5856693387031555} 02/26/2022 16:02:59 - INFO - codeparrot_training - Step 48225: {'lr': 1.6851380963966156e-06, 'samples': 24691712, 'steps': 48225, 'loss/train': 1.3649907112121582} 02/26/2022 16:03:02 - INFO - codeparrot_training - Step 48226: {'lr': 1.6832420170800888e-06, 'samples': 24692224, 'steps': 48226, 'loss/train': 1.9108003377914429} 02/26/2022 16:03:08 - INFO - codeparrot_training - Step 48227: {'lr': 1.681347001473682e-06, 'samples': 24692736, 'steps': 48227, 'loss/train': 2.4020004272460938} 02/26/2022 16:03:11 - INFO - codeparrot_training - Step 48228: {'lr': 1.6794530495855831e-06, 'samples': 24693248, 'steps': 48228, 'loss/train': 0.6762791872024536} 02/26/2022 16:03:17 - INFO - codeparrot_training - Step 48229: {'lr': 1.6775601614238134e-06, 'samples': 24693760, 'steps': 48229, 'loss/train': 2.834538459777832} 02/26/2022 16:03:20 - INFO - codeparrot_training - Step 48230: {'lr': 1.675668336996533e-06, 'samples': 24694272, 'steps': 48230, 'loss/train': 2.110729217529297} 02/26/2022 16:03:26 - INFO - codeparrot_training - Step 48231: {'lr': 1.6737775763118468e-06, 'samples': 24694784, 'steps': 48231, 'loss/train': 1.7509702444076538} 02/26/2022 16:03:32 - INFO - codeparrot_training - Step 48232: {'lr': 1.671887879377859e-06, 'samples': 24695296, 'steps': 48232, 'loss/train': 0.037955548614263535} 02/26/2022 16:03:36 - INFO - codeparrot_training - Step 48233: {'lr': 1.6699992462026193e-06, 'samples': 24695808, 'steps': 48233, 'loss/train': 0.7673493027687073} 02/26/2022 16:03:41 - INFO - codeparrot_training - Step 48234: {'lr': 1.6681116767942871e-06, 'samples': 24696320, 'steps': 48234, 'loss/train': 1.8250377178192139} 02/26/2022 16:03:44 - INFO - codeparrot_training - Step 48235: {'lr': 1.6662251711608846e-06, 'samples': 24696832, 'steps': 48235, 'loss/train': 2.45802640914917} 02/26/2022 16:03:50 - INFO - codeparrot_training - Step 48236: {'lr': 1.6643397293105156e-06, 'samples': 24697344, 'steps': 48236, 'loss/train': 1.9320323467254639} 02/26/2022 16:03:53 - INFO - codeparrot_training - Step 48237: {'lr': 1.6624553512512574e-06, 'samples': 24697856, 'steps': 48237, 'loss/train': 1.4622244834899902} 02/26/2022 16:03:59 - INFO - codeparrot_training - Step 48238: {'lr': 1.6605720369912147e-06, 'samples': 24698368, 'steps': 48238, 'loss/train': 2.220015287399292} 02/26/2022 16:04:02 - INFO - codeparrot_training - Step 48239: {'lr': 1.658689786538381e-06, 'samples': 24698880, 'steps': 48239, 'loss/train': 0.7419174909591675} 02/26/2022 16:04:08 - INFO - codeparrot_training - Step 48240: {'lr': 1.6568085999008887e-06, 'samples': 24699392, 'steps': 48240, 'loss/train': 2.6873536109924316} 02/26/2022 16:04:11 - INFO - codeparrot_training - Step 48241: {'lr': 1.6549284770867867e-06, 'samples': 24699904, 'steps': 48241, 'loss/train': 1.902570128440857} 02/26/2022 16:04:18 - INFO - codeparrot_training - Step 48242: {'lr': 1.6530494181040968e-06, 'samples': 24700416, 'steps': 48242, 'loss/train': 2.783463954925537} 02/26/2022 16:04:21 - INFO - codeparrot_training - Step 48243: {'lr': 1.6511714229608677e-06, 'samples': 24700928, 'steps': 48243, 'loss/train': 1.3428328037261963} 02/26/2022 16:04:27 - INFO - codeparrot_training - Step 48244: {'lr': 1.6492944916651765e-06, 'samples': 24701440, 'steps': 48244, 'loss/train': 1.5959863662719727} 02/26/2022 16:04:30 - INFO - codeparrot_training - Step 48245: {'lr': 1.6474186242250445e-06, 'samples': 24701952, 'steps': 48245, 'loss/train': 1.367204189300537} 02/26/2022 16:04:36 - INFO - codeparrot_training - Step 48246: {'lr': 1.6455438206485207e-06, 'samples': 24702464, 'steps': 48246, 'loss/train': 2.2631332874298096} 02/26/2022 16:04:39 - INFO - codeparrot_training - Step 48247: {'lr': 1.6436700809435989e-06, 'samples': 24702976, 'steps': 48247, 'loss/train': 1.3153676986694336} 02/26/2022 16:04:45 - INFO - codeparrot_training - Step 48248: {'lr': 1.641797405118356e-06, 'samples': 24703488, 'steps': 48248, 'loss/train': 1.4450786113739014} 02/26/2022 16:04:48 - INFO - codeparrot_training - Step 48249: {'lr': 1.6399257931807855e-06, 'samples': 24704000, 'steps': 48249, 'loss/train': 1.1745692491531372} 02/26/2022 16:04:54 - INFO - codeparrot_training - Step 48250: {'lr': 1.6380552451389085e-06, 'samples': 24704512, 'steps': 48250, 'loss/train': 2.138521671295166} 02/26/2022 16:04:57 - INFO - codeparrot_training - Step 48251: {'lr': 1.6361857610007191e-06, 'samples': 24705024, 'steps': 48251, 'loss/train': 1.839398980140686} 02/26/2022 16:05:03 - INFO - codeparrot_training - Step 48252: {'lr': 1.6343173407742662e-06, 'samples': 24705536, 'steps': 48252, 'loss/train': 2.284489393234253} 02/26/2022 16:05:07 - INFO - codeparrot_training - Step 48253: {'lr': 1.6324499844675434e-06, 'samples': 24706048, 'steps': 48253, 'loss/train': 1.533995509147644} 02/26/2022 16:05:12 - INFO - codeparrot_training - Step 48254: {'lr': 1.6305836920885442e-06, 'samples': 24706560, 'steps': 48254, 'loss/train': 0.5098047852516174} 02/26/2022 16:05:16 - INFO - codeparrot_training - Step 48255: {'lr': 1.6287184636452345e-06, 'samples': 24707072, 'steps': 48255, 'loss/train': 2.63771915435791} 02/26/2022 16:05:21 - INFO - codeparrot_training - Step 48256: {'lr': 1.6268542991456358e-06, 'samples': 24707584, 'steps': 48256, 'loss/train': 1.6032116413116455} 02/26/2022 16:05:25 - INFO - codeparrot_training - Step 48257: {'lr': 1.624991198597714e-06, 'samples': 24708096, 'steps': 48257, 'loss/train': 1.1501610279083252} 02/26/2022 16:05:30 - INFO - codeparrot_training - Step 48258: {'lr': 1.62312916200949e-06, 'samples': 24708608, 'steps': 48258, 'loss/train': 2.0278096199035645} 02/26/2022 16:05:34 - INFO - codeparrot_training - Step 48259: {'lr': 1.6212681893888747e-06, 'samples': 24709120, 'steps': 48259, 'loss/train': 1.8049098253250122} 02/26/2022 16:05:39 - INFO - codeparrot_training - Step 48260: {'lr': 1.619408280743917e-06, 'samples': 24709632, 'steps': 48260, 'loss/train': 0.5946051478385925} 02/26/2022 16:05:42 - INFO - codeparrot_training - Step 48261: {'lr': 1.6175494360825272e-06, 'samples': 24710144, 'steps': 48261, 'loss/train': 1.7702735662460327} 02/26/2022 16:05:48 - INFO - codeparrot_training - Step 48262: {'lr': 1.615691655412671e-06, 'samples': 24710656, 'steps': 48262, 'loss/train': 1.2659212350845337} 02/26/2022 16:05:51 - INFO - codeparrot_training - Step 48263: {'lr': 1.6138349387423422e-06, 'samples': 24711168, 'steps': 48263, 'loss/train': 5.6597771644592285} 02/26/2022 16:05:57 - INFO - codeparrot_training - Step 48264: {'lr': 1.6119792860794513e-06, 'samples': 24711680, 'steps': 48264, 'loss/train': 1.6914167404174805} 02/26/2022 16:06:00 - INFO - codeparrot_training - Step 48265: {'lr': 1.6101246974319916e-06, 'samples': 24712192, 'steps': 48265, 'loss/train': 1.9811280965805054} 02/26/2022 16:06:07 - INFO - codeparrot_training - Step 48266: {'lr': 1.6082711728078735e-06, 'samples': 24712704, 'steps': 48266, 'loss/train': 2.077152729034424} 02/26/2022 16:06:11 - INFO - codeparrot_training - Step 48267: {'lr': 1.6064187122150353e-06, 'samples': 24713216, 'steps': 48267, 'loss/train': 1.2700115442276} 02/26/2022 16:06:16 - INFO - codeparrot_training - Step 48268: {'lr': 1.6045673156614427e-06, 'samples': 24713728, 'steps': 48268, 'loss/train': 1.4722819328308105} 02/26/2022 16:06:20 - INFO - codeparrot_training - Step 48269: {'lr': 1.602716983155006e-06, 'samples': 24714240, 'steps': 48269, 'loss/train': 1.5597320795059204} 02/26/2022 16:06:25 - INFO - codeparrot_training - Step 48270: {'lr': 1.6008677147036633e-06, 'samples': 24714752, 'steps': 48270, 'loss/train': 2.409360885620117} 02/26/2022 16:06:29 - INFO - codeparrot_training - Step 48271: {'lr': 1.5990195103153249e-06, 'samples': 24715264, 'steps': 48271, 'loss/train': 0.4325543940067291} 02/26/2022 16:06:34 - INFO - codeparrot_training - Step 48272: {'lr': 1.5971723699979013e-06, 'samples': 24715776, 'steps': 48272, 'loss/train': 2.0530693531036377} 02/26/2022 16:06:38 - INFO - codeparrot_training - Step 48273: {'lr': 1.5953262937593305e-06, 'samples': 24716288, 'steps': 48273, 'loss/train': 1.0317387580871582} 02/26/2022 16:06:43 - INFO - codeparrot_training - Step 48274: {'lr': 1.5934812816074673e-06, 'samples': 24716800, 'steps': 48274, 'loss/train': 1.731128215789795} 02/26/2022 16:06:47 - INFO - codeparrot_training - Step 48275: {'lr': 1.5916373335503054e-06, 'samples': 24717312, 'steps': 48275, 'loss/train': 1.4390676021575928} 02/26/2022 16:06:52 - INFO - codeparrot_training - Step 48276: {'lr': 1.589794449595644e-06, 'samples': 24717824, 'steps': 48276, 'loss/train': 0.07702524960041046} 02/26/2022 16:06:56 - INFO - codeparrot_training - Step 48277: {'lr': 1.587952629751449e-06, 'samples': 24718336, 'steps': 48277, 'loss/train': 1.5532214641571045} 02/26/2022 16:07:02 - INFO - codeparrot_training - Step 48278: {'lr': 1.5861118740255753e-06, 'samples': 24718848, 'steps': 48278, 'loss/train': 1.0452167987823486} 02/26/2022 16:07:06 - INFO - codeparrot_training - Step 48279: {'lr': 1.5842721824259332e-06, 'samples': 24719360, 'steps': 48279, 'loss/train': 1.191022276878357} 02/26/2022 16:07:11 - INFO - codeparrot_training - Step 48280: {'lr': 1.5824335549603774e-06, 'samples': 24719872, 'steps': 48280, 'loss/train': 1.511879324913025} 02/26/2022 16:07:15 - INFO - codeparrot_training - Step 48281: {'lr': 1.5805959916367906e-06, 'samples': 24720384, 'steps': 48281, 'loss/train': 1.860042929649353} 02/26/2022 16:07:20 - INFO - codeparrot_training - Step 48282: {'lr': 1.5787594924630556e-06, 'samples': 24720896, 'steps': 48282, 'loss/train': 2.0549185276031494} 02/26/2022 16:07:24 - INFO - codeparrot_training - Step 48283: {'lr': 1.5769240574470267e-06, 'samples': 24721408, 'steps': 48283, 'loss/train': 1.8378620147705078} 02/26/2022 16:07:30 - INFO - codeparrot_training - Step 48284: {'lr': 1.5750896865965592e-06, 'samples': 24721920, 'steps': 48284, 'loss/train': 3.098766565322876} 02/26/2022 16:07:33 - INFO - codeparrot_training - Step 48285: {'lr': 1.5732563799195632e-06, 'samples': 24722432, 'steps': 48285, 'loss/train': 0.4641614556312561} 02/26/2022 16:07:39 - INFO - codeparrot_training - Step 48286: {'lr': 1.5714241374238103e-06, 'samples': 24722944, 'steps': 48286, 'loss/train': 2.4651243686676025} 02/26/2022 16:07:42 - INFO - codeparrot_training - Step 48287: {'lr': 1.5695929591172109e-06, 'samples': 24723456, 'steps': 48287, 'loss/train': 1.2019466161727905} 02/26/2022 16:07:47 - INFO - codeparrot_training - Step 48288: {'lr': 1.567762845007592e-06, 'samples': 24723968, 'steps': 48288, 'loss/train': 1.4710183143615723} 02/26/2022 16:07:51 - INFO - codeparrot_training - Step 48289: {'lr': 1.5659337951028086e-06, 'samples': 24724480, 'steps': 48289, 'loss/train': 0.6994738578796387} 02/26/2022 16:07:58 - INFO - codeparrot_training - Step 48290: {'lr': 1.564105809410632e-06, 'samples': 24724992, 'steps': 48290, 'loss/train': 1.680452823638916} 02/26/2022 16:08:01 - INFO - codeparrot_training - Step 48291: {'lr': 1.5622788879389727e-06, 'samples': 24725504, 'steps': 48291, 'loss/train': 0.2028939127922058} 02/26/2022 16:08:07 - INFO - codeparrot_training - Step 48292: {'lr': 1.560453030695602e-06, 'samples': 24726016, 'steps': 48292, 'loss/train': 2.0981016159057617} 02/26/2022 16:08:10 - INFO - codeparrot_training - Step 48293: {'lr': 1.5586282376884032e-06, 'samples': 24726528, 'steps': 48293, 'loss/train': 0.9162625074386597} 02/26/2022 16:08:16 - INFO - codeparrot_training - Step 48294: {'lr': 1.5568045089250916e-06, 'samples': 24727040, 'steps': 48294, 'loss/train': 1.4696649312973022} 02/26/2022 16:08:19 - INFO - codeparrot_training - Step 48295: {'lr': 1.5549818444135777e-06, 'samples': 24727552, 'steps': 48295, 'loss/train': 1.9322148561477661} 02/26/2022 16:08:25 - INFO - codeparrot_training - Step 48296: {'lr': 1.5531602441616332e-06, 'samples': 24728064, 'steps': 48296, 'loss/train': 2.042065382003784} 02/26/2022 16:08:31 - INFO - codeparrot_training - Step 48297: {'lr': 1.5513397081770298e-06, 'samples': 24728576, 'steps': 48297, 'loss/train': 2.0719542503356934} 02/26/2022 16:08:35 - INFO - codeparrot_training - Step 48298: {'lr': 1.549520236467622e-06, 'samples': 24729088, 'steps': 48298, 'loss/train': 1.8613502979278564} 02/26/2022 16:08:40 - INFO - codeparrot_training - Step 48299: {'lr': 1.5477018290411815e-06, 'samples': 24729600, 'steps': 48299, 'loss/train': 1.3619030714035034} 02/26/2022 16:08:44 - INFO - codeparrot_training - Step 48300: {'lr': 1.54588448590548e-06, 'samples': 24730112, 'steps': 48300, 'loss/train': 1.9936413764953613} 02/26/2022 16:08:49 - INFO - codeparrot_training - Step 48301: {'lr': 1.5440682070683165e-06, 'samples': 24730624, 'steps': 48301, 'loss/train': 1.289435863494873} 02/26/2022 16:08:53 - INFO - codeparrot_training - Step 48302: {'lr': 1.5422529925374907e-06, 'samples': 24731136, 'steps': 48302, 'loss/train': 0.029236581176519394} 02/26/2022 16:08:58 - INFO - codeparrot_training - Step 48303: {'lr': 1.540438842320746e-06, 'samples': 24731648, 'steps': 48303, 'loss/train': 0.9637269377708435} 02/26/2022 16:09:02 - INFO - codeparrot_training - Step 48304: {'lr': 1.5386257564258543e-06, 'samples': 24732160, 'steps': 48304, 'loss/train': 1.585838794708252} 02/26/2022 16:09:07 - INFO - codeparrot_training - Step 48305: {'lr': 1.5368137348606148e-06, 'samples': 24732672, 'steps': 48305, 'loss/train': 2.0499584674835205} 02/26/2022 16:09:11 - INFO - codeparrot_training - Step 48306: {'lr': 1.5350027776327435e-06, 'samples': 24733184, 'steps': 48306, 'loss/train': 1.1607283353805542} 02/26/2022 16:09:16 - INFO - codeparrot_training - Step 48307: {'lr': 1.5331928847500399e-06, 'samples': 24733696, 'steps': 48307, 'loss/train': 1.7936731576919556} 02/26/2022 16:09:20 - INFO - codeparrot_training - Step 48308: {'lr': 1.5313840562202475e-06, 'samples': 24734208, 'steps': 48308, 'loss/train': 1.611100673675537} 02/26/2022 16:09:25 - INFO - codeparrot_training - Step 48309: {'lr': 1.5295762920511103e-06, 'samples': 24734720, 'steps': 48309, 'loss/train': 2.516822576522827} 02/26/2022 16:09:29 - INFO - codeparrot_training - Step 48310: {'lr': 1.5277695922503442e-06, 'samples': 24735232, 'steps': 48310, 'loss/train': 1.8794735670089722} 02/26/2022 16:09:34 - INFO - codeparrot_training - Step 48311: {'lr': 1.525963956825749e-06, 'samples': 24735744, 'steps': 48311, 'loss/train': 1.3278895616531372} 02/26/2022 16:09:38 - INFO - codeparrot_training - Step 48312: {'lr': 1.5241593857850122e-06, 'samples': 24736256, 'steps': 48312, 'loss/train': 1.3245714902877808} 02/26/2022 16:09:44 - INFO - codeparrot_training - Step 48313: {'lr': 1.5223558791358505e-06, 'samples': 24736768, 'steps': 48313, 'loss/train': 0.967051088809967} 02/26/2022 16:09:48 - INFO - codeparrot_training - Step 48314: {'lr': 1.5205534368860352e-06, 'samples': 24737280, 'steps': 48314, 'loss/train': 1.2493215799331665} 02/26/2022 16:09:53 - INFO - codeparrot_training - Step 48315: {'lr': 1.5187520590432823e-06, 'samples': 24737792, 'steps': 48315, 'loss/train': 1.688737392425537} 02/26/2022 16:09:56 - INFO - codeparrot_training - Step 48316: {'lr': 1.5169517456152803e-06, 'samples': 24738304, 'steps': 48316, 'loss/train': 2.0536041259765625} 02/26/2022 16:10:02 - INFO - codeparrot_training - Step 48317: {'lr': 1.515152496609745e-06, 'samples': 24738816, 'steps': 48317, 'loss/train': 1.2988651990890503} 02/26/2022 16:10:05 - INFO - codeparrot_training - Step 48318: {'lr': 1.5133543120344207e-06, 'samples': 24739328, 'steps': 48318, 'loss/train': 1.5435714721679688} 02/26/2022 16:10:11 - INFO - codeparrot_training - Step 48319: {'lr': 1.5115571918969396e-06, 'samples': 24739840, 'steps': 48319, 'loss/train': 1.995025396347046} 02/26/2022 16:10:14 - INFO - codeparrot_training - Step 48320: {'lr': 1.5097611362051012e-06, 'samples': 24740352, 'steps': 48320, 'loss/train': 0.7486575841903687} 02/26/2022 16:10:20 - INFO - codeparrot_training - Step 48321: {'lr': 1.5079661449664828e-06, 'samples': 24740864, 'steps': 48321, 'loss/train': 1.749605417251587} 02/26/2022 16:10:23 - INFO - codeparrot_training - Step 48322: {'lr': 1.506172218188856e-06, 'samples': 24741376, 'steps': 48322, 'loss/train': 1.3769954442977905} 02/26/2022 16:10:30 - INFO - codeparrot_training - Step 48323: {'lr': 1.5043793558798812e-06, 'samples': 24741888, 'steps': 48323, 'loss/train': 1.4700435400009155} 02/26/2022 16:10:33 - INFO - codeparrot_training - Step 48324: {'lr': 1.5025875580472748e-06, 'samples': 24742400, 'steps': 48324, 'loss/train': 2.265261650085449} 02/26/2022 16:10:39 - INFO - codeparrot_training - Step 48325: {'lr': 1.5007968246986137e-06, 'samples': 24742912, 'steps': 48325, 'loss/train': 3.0830962657928467} 02/26/2022 16:10:42 - INFO - codeparrot_training - Step 48326: {'lr': 1.4990071558416695e-06, 'samples': 24743424, 'steps': 48326, 'loss/train': 1.3435145616531372} 02/26/2022 16:10:48 - INFO - codeparrot_training - Step 48327: {'lr': 1.4972185514840752e-06, 'samples': 24743936, 'steps': 48327, 'loss/train': 0.5534811019897461} 02/26/2022 16:10:51 - INFO - codeparrot_training - Step 48328: {'lr': 1.4954310116334913e-06, 'samples': 24744448, 'steps': 48328, 'loss/train': 1.7710227966308594} 02/26/2022 16:10:57 - INFO - codeparrot_training - Step 48329: {'lr': 1.4936445362975504e-06, 'samples': 24744960, 'steps': 48329, 'loss/train': 1.5421433448791504} 02/26/2022 16:11:00 - INFO - codeparrot_training - Step 48330: {'lr': 1.491859125483913e-06, 'samples': 24745472, 'steps': 48330, 'loss/train': 0.7247592210769653} 02/26/2022 16:11:06 - INFO - codeparrot_training - Step 48331: {'lr': 1.4900747792002678e-06, 'samples': 24745984, 'steps': 48331, 'loss/train': 0.9581589102745056} 02/26/2022 16:11:09 - INFO - codeparrot_training - Step 48332: {'lr': 1.4882914974542195e-06, 'samples': 24746496, 'steps': 48332, 'loss/train': 0.7496263980865479} 02/26/2022 16:11:14 - INFO - codeparrot_training - Step 48333: {'lr': 1.4865092802534285e-06, 'samples': 24747008, 'steps': 48333, 'loss/train': 1.6404470205307007} 02/26/2022 16:11:18 - INFO - codeparrot_training - Step 48334: {'lr': 1.484728127605528e-06, 'samples': 24747520, 'steps': 48334, 'loss/train': 1.3127260208129883} 02/26/2022 16:11:24 - INFO - codeparrot_training - Step 48335: {'lr': 1.4829480395181228e-06, 'samples': 24748032, 'steps': 48335, 'loss/train': 1.2283155918121338} 02/26/2022 16:11:28 - INFO - codeparrot_training - Step 48336: {'lr': 1.4811690159988457e-06, 'samples': 24748544, 'steps': 48336, 'loss/train': 1.6740950345993042} 02/26/2022 16:11:33 - INFO - codeparrot_training - Step 48337: {'lr': 1.479391057055357e-06, 'samples': 24749056, 'steps': 48337, 'loss/train': 1.8351653814315796} 02/26/2022 16:11:37 - INFO - codeparrot_training - Step 48338: {'lr': 1.4776141626952344e-06, 'samples': 24749568, 'steps': 48338, 'loss/train': 1.1003621816635132} 02/26/2022 16:11:42 - INFO - codeparrot_training - Step 48339: {'lr': 1.4758383329260828e-06, 'samples': 24750080, 'steps': 48339, 'loss/train': 1.506985068321228} 02/26/2022 16:11:46 - INFO - codeparrot_training - Step 48340: {'lr': 1.4740635677555348e-06, 'samples': 24750592, 'steps': 48340, 'loss/train': 1.9745590686798096} 02/26/2022 16:11:51 - INFO - codeparrot_training - Step 48341: {'lr': 1.4722898671911678e-06, 'samples': 24751104, 'steps': 48341, 'loss/train': 1.2870476245880127} 02/26/2022 16:11:55 - INFO - codeparrot_training - Step 48342: {'lr': 1.4705172312406146e-06, 'samples': 24751616, 'steps': 48342, 'loss/train': 1.9093905687332153} 02/26/2022 16:12:00 - INFO - codeparrot_training - Step 48343: {'lr': 1.4687456599114245e-06, 'samples': 24752128, 'steps': 48343, 'loss/train': 2.1184017658233643} 02/26/2022 16:12:03 - INFO - codeparrot_training - Step 48344: {'lr': 1.4669751532112308e-06, 'samples': 24752640, 'steps': 48344, 'loss/train': 1.351211667060852} 02/26/2022 16:12:10 - INFO - codeparrot_training - Step 48345: {'lr': 1.4652057111476103e-06, 'samples': 24753152, 'steps': 48345, 'loss/train': 0.8802154660224915} 02/26/2022 16:12:13 - INFO - codeparrot_training - Step 48346: {'lr': 1.4634373337281125e-06, 'samples': 24753664, 'steps': 48346, 'loss/train': 1.505128026008606} 02/26/2022 16:12:19 - INFO - codeparrot_training - Step 48347: {'lr': 1.4616700209603151e-06, 'samples': 24754176, 'steps': 48347, 'loss/train': 2.3156750202178955} 02/26/2022 16:12:22 - INFO - codeparrot_training - Step 48348: {'lr': 1.4599037728518228e-06, 'samples': 24754688, 'steps': 48348, 'loss/train': 2.5019710063934326} 02/26/2022 16:12:28 - INFO - codeparrot_training - Step 48349: {'lr': 1.458138589410185e-06, 'samples': 24755200, 'steps': 48349, 'loss/train': 1.2664331197738647} 02/26/2022 16:12:31 - INFO - codeparrot_training - Step 48350: {'lr': 1.4563744706429517e-06, 'samples': 24755712, 'steps': 48350, 'loss/train': 1.1959731578826904} 02/26/2022 16:12:37 - INFO - codeparrot_training - Step 48351: {'lr': 1.4546114165576995e-06, 'samples': 24756224, 'steps': 48351, 'loss/train': 1.715791940689087} 02/26/2022 16:12:40 - INFO - codeparrot_training - Step 48352: {'lr': 1.4528494271619507e-06, 'samples': 24756736, 'steps': 48352, 'loss/train': 2.287785768508911} 02/26/2022 16:12:46 - INFO - codeparrot_training - Step 48353: {'lr': 1.4510885024632825e-06, 'samples': 24757248, 'steps': 48353, 'loss/train': 1.948127031326294} 02/26/2022 16:12:49 - INFO - codeparrot_training - Step 48354: {'lr': 1.4493286424692441e-06, 'samples': 24757760, 'steps': 48354, 'loss/train': 1.8993802070617676} 02/26/2022 16:12:54 - INFO - codeparrot_training - Step 48355: {'lr': 1.44756984718733e-06, 'samples': 24758272, 'steps': 48355, 'loss/train': 1.9417109489440918} 02/26/2022 16:12:58 - INFO - codeparrot_training - Step 48356: {'lr': 1.445812116625117e-06, 'samples': 24758784, 'steps': 48356, 'loss/train': 1.736548662185669} 02/26/2022 16:13:03 - INFO - codeparrot_training - Step 48357: {'lr': 1.4440554507901272e-06, 'samples': 24759296, 'steps': 48357, 'loss/train': 0.8327333331108093} 02/26/2022 16:13:09 - INFO - codeparrot_training - Step 48358: {'lr': 1.4422998496898543e-06, 'samples': 24759808, 'steps': 48358, 'loss/train': 2.761749744415283} 02/26/2022 16:13:12 - INFO - codeparrot_training - Step 48359: {'lr': 1.4405453133318757e-06, 'samples': 24760320, 'steps': 48359, 'loss/train': 0.676769495010376} 02/26/2022 16:13:19 - INFO - codeparrot_training - Step 48360: {'lr': 1.4387918417236578e-06, 'samples': 24760832, 'steps': 48360, 'loss/train': 1.3783200979232788} 02/26/2022 16:13:23 - INFO - codeparrot_training - Step 48361: {'lr': 1.4370394348727223e-06, 'samples': 24761344, 'steps': 48361, 'loss/train': 1.2662625312805176} 02/26/2022 16:13:28 - INFO - codeparrot_training - Step 48362: {'lr': 1.435288092786563e-06, 'samples': 24761856, 'steps': 48362, 'loss/train': 2.250563859939575} 02/26/2022 16:13:32 - INFO - codeparrot_training - Step 48363: {'lr': 1.4335378154727574e-06, 'samples': 24762368, 'steps': 48363, 'loss/train': 1.7857874631881714} 02/26/2022 16:13:37 - INFO - codeparrot_training - Step 48364: {'lr': 1.4317886029387162e-06, 'samples': 24762880, 'steps': 48364, 'loss/train': 2.1841542720794678} 02/26/2022 16:13:41 - INFO - codeparrot_training - Step 48365: {'lr': 1.430040455191961e-06, 'samples': 24763392, 'steps': 48365, 'loss/train': 1.6927770376205444} 02/26/2022 16:13:46 - INFO - codeparrot_training - Step 48366: {'lr': 1.4282933722399583e-06, 'samples': 24763904, 'steps': 48366, 'loss/train': 0.5043333768844604} 02/26/2022 16:13:50 - INFO - codeparrot_training - Step 48367: {'lr': 1.4265473540902574e-06, 'samples': 24764416, 'steps': 48367, 'loss/train': 1.06955087184906} 02/26/2022 16:13:56 - INFO - codeparrot_training - Step 48368: {'lr': 1.424802400750269e-06, 'samples': 24764928, 'steps': 48368, 'loss/train': 1.1917561292648315} 02/26/2022 16:14:00 - INFO - codeparrot_training - Step 48369: {'lr': 1.4230585122275152e-06, 'samples': 24765440, 'steps': 48369, 'loss/train': 1.95912504196167} 02/26/2022 16:14:03 - INFO - codeparrot_training - Step 48370: {'lr': 1.4213156885294343e-06, 'samples': 24765952, 'steps': 48370, 'loss/train': 1.4929018020629883} 02/26/2022 16:14:09 - INFO - codeparrot_training - Step 48371: {'lr': 1.4195739296635202e-06, 'samples': 24766464, 'steps': 48371, 'loss/train': 2.3770599365234375} 02/26/2022 16:14:13 - INFO - codeparrot_training - Step 48372: {'lr': 1.4178332356372114e-06, 'samples': 24766976, 'steps': 48372, 'loss/train': 2.1477420330047607} 02/26/2022 16:14:18 - INFO - codeparrot_training - Step 48373: {'lr': 1.4160936064579744e-06, 'samples': 24767488, 'steps': 48373, 'loss/train': 0.5264931321144104} 02/26/2022 16:14:22 - INFO - codeparrot_training - Step 48374: {'lr': 1.4143550421332196e-06, 'samples': 24768000, 'steps': 48374, 'loss/train': 1.6225192546844482} 02/26/2022 16:14:27 - INFO - codeparrot_training - Step 48375: {'lr': 1.412617542670469e-06, 'samples': 24768512, 'steps': 48375, 'loss/train': 1.0946015119552612} 02/26/2022 16:14:31 - INFO - codeparrot_training - Step 48376: {'lr': 1.4108811080771333e-06, 'samples': 24769024, 'steps': 48376, 'loss/train': 2.090301513671875} 02/26/2022 16:14:36 - INFO - codeparrot_training - Step 48377: {'lr': 1.409145738360651e-06, 'samples': 24769536, 'steps': 48377, 'loss/train': 1.943251371383667} 02/26/2022 16:14:40 - INFO - codeparrot_training - Step 48378: {'lr': 1.4074114335284327e-06, 'samples': 24770048, 'steps': 48378, 'loss/train': 2.4265153408050537} 02/26/2022 16:14:45 - INFO - codeparrot_training - Step 48379: {'lr': 1.4056781935879448e-06, 'samples': 24770560, 'steps': 48379, 'loss/train': 1.4212602376937866} 02/26/2022 16:14:48 - INFO - codeparrot_training - Step 48380: {'lr': 1.4039460185465703e-06, 'samples': 24771072, 'steps': 48380, 'loss/train': 0.567535400390625} 02/26/2022 16:14:55 - INFO - codeparrot_training - Step 48381: {'lr': 1.4022149084117753e-06, 'samples': 24771584, 'steps': 48381, 'loss/train': 2.491757869720459} 02/26/2022 16:14:58 - INFO - codeparrot_training - Step 48382: {'lr': 1.4004848631909428e-06, 'samples': 24772096, 'steps': 48382, 'loss/train': 0.42956382036209106} 02/26/2022 16:15:04 - INFO - codeparrot_training - Step 48383: {'lr': 1.3987558828914837e-06, 'samples': 24772608, 'steps': 48383, 'loss/train': 1.173322081565857} 02/26/2022 16:15:07 - INFO - codeparrot_training - Step 48384: {'lr': 1.3970279675208364e-06, 'samples': 24773120, 'steps': 48384, 'loss/train': 0.4197263717651367} 02/26/2022 16:15:13 - INFO - codeparrot_training - Step 48385: {'lr': 1.395301117086356e-06, 'samples': 24773632, 'steps': 48385, 'loss/train': 2.993868350982666} 02/26/2022 16:15:16 - INFO - codeparrot_training - Step 48386: {'lr': 1.3935753315954814e-06, 'samples': 24774144, 'steps': 48386, 'loss/train': 1.747635841369629} 02/26/2022 16:15:22 - INFO - codeparrot_training - Step 48387: {'lr': 1.391850611055595e-06, 'samples': 24774656, 'steps': 48387, 'loss/train': 3.0021746158599854} 02/26/2022 16:15:25 - INFO - codeparrot_training - Step 48388: {'lr': 1.3901269554740525e-06, 'samples': 24775168, 'steps': 48388, 'loss/train': 1.4070488214492798} 02/26/2022 16:15:31 - INFO - codeparrot_training - Step 48389: {'lr': 1.3884043648582922e-06, 'samples': 24775680, 'steps': 48389, 'loss/train': 2.9738705158233643} 02/26/2022 16:15:34 - INFO - codeparrot_training - Step 48390: {'lr': 1.3866828392156416e-06, 'samples': 24776192, 'steps': 48390, 'loss/train': 1.5986583232879639} 02/26/2022 16:15:40 - INFO - codeparrot_training - Step 48391: {'lr': 1.3849623785535114e-06, 'samples': 24776704, 'steps': 48391, 'loss/train': 1.838472604751587} 02/26/2022 16:15:44 - INFO - codeparrot_training - Step 48392: {'lr': 1.383242982879257e-06, 'samples': 24777216, 'steps': 48392, 'loss/train': 2.0443947315216064} 02/26/2022 16:15:49 - INFO - codeparrot_training - Step 48393: {'lr': 1.3815246522002334e-06, 'samples': 24777728, 'steps': 48393, 'loss/train': 1.369342565536499} 02/26/2022 16:15:53 - INFO - codeparrot_training - Step 48394: {'lr': 1.3798073865238236e-06, 'samples': 24778240, 'steps': 48394, 'loss/train': 1.599927544593811} 02/26/2022 16:15:58 - INFO - codeparrot_training - Step 48395: {'lr': 1.378091185857383e-06, 'samples': 24778752, 'steps': 48395, 'loss/train': 1.2066359519958496} 02/26/2022 16:16:02 - INFO - codeparrot_training - Step 48396: {'lr': 1.376376050208239e-06, 'samples': 24779264, 'steps': 48396, 'loss/train': 0.7172000408172607} 02/26/2022 16:16:07 - INFO - codeparrot_training - Step 48397: {'lr': 1.3746619795837467e-06, 'samples': 24779776, 'steps': 48397, 'loss/train': 2.030796527862549} 02/26/2022 16:16:11 - INFO - codeparrot_training - Step 48398: {'lr': 1.3729489739912893e-06, 'samples': 24780288, 'steps': 48398, 'loss/train': 1.5205879211425781} 02/26/2022 16:16:16 - INFO - codeparrot_training - Step 48399: {'lr': 1.3712370334381385e-06, 'samples': 24780800, 'steps': 48399, 'loss/train': 1.164409875869751} 02/26/2022 16:16:20 - INFO - codeparrot_training - Step 48400: {'lr': 1.3695261579316775e-06, 'samples': 24781312, 'steps': 48400, 'loss/train': 3.031507968902588} 02/26/2022 16:16:25 - INFO - codeparrot_training - Step 48401: {'lr': 1.367816347479206e-06, 'samples': 24781824, 'steps': 48401, 'loss/train': 2.8563151359558105} 02/26/2022 16:16:29 - INFO - codeparrot_training - Step 48402: {'lr': 1.3661076020880514e-06, 'samples': 24782336, 'steps': 48402, 'loss/train': 1.5785893201828003} 02/26/2022 16:16:34 - INFO - codeparrot_training - Step 48403: {'lr': 1.364399921765569e-06, 'samples': 24782848, 'steps': 48403, 'loss/train': 1.141401767730713} 02/26/2022 16:16:38 - INFO - codeparrot_training - Step 48404: {'lr': 1.3626933065190306e-06, 'samples': 24783360, 'steps': 48404, 'loss/train': 2.185882091522217} 02/26/2022 16:16:43 - INFO - codeparrot_training - Step 48405: {'lr': 1.360987756355736e-06, 'samples': 24783872, 'steps': 48405, 'loss/train': 0.8031267523765564} 02/26/2022 16:16:47 - INFO - codeparrot_training - Step 48406: {'lr': 1.3592832712830405e-06, 'samples': 24784384, 'steps': 48406, 'loss/train': 1.9465123414993286} 02/26/2022 16:16:53 - INFO - codeparrot_training - Step 48407: {'lr': 1.3575798513082438e-06, 'samples': 24784896, 'steps': 48407, 'loss/train': 1.2222914695739746} 02/26/2022 16:16:56 - INFO - codeparrot_training - Step 48408: {'lr': 1.3558774964386177e-06, 'samples': 24785408, 'steps': 48408, 'loss/train': 0.8967334032058716} 02/26/2022 16:17:02 - INFO - codeparrot_training - Step 48409: {'lr': 1.3541762066814346e-06, 'samples': 24785920, 'steps': 48409, 'loss/train': 0.7138988971710205} 02/26/2022 16:17:05 - INFO - codeparrot_training - Step 48410: {'lr': 1.3524759820440212e-06, 'samples': 24786432, 'steps': 48410, 'loss/train': 1.5887726545333862} 02/26/2022 16:17:11 - INFO - codeparrot_training - Step 48411: {'lr': 1.3507768225336503e-06, 'samples': 24786944, 'steps': 48411, 'loss/train': 2.3518736362457275} 02/26/2022 16:17:14 - INFO - codeparrot_training - Step 48412: {'lr': 1.3490787281575933e-06, 'samples': 24787456, 'steps': 48412, 'loss/train': 0.9374799728393555} 02/26/2022 16:17:20 - INFO - codeparrot_training - Step 48413: {'lr': 1.3473816989230947e-06, 'samples': 24787968, 'steps': 48413, 'loss/train': 1.8472918272018433} 02/26/2022 16:17:24 - INFO - codeparrot_training - Step 48414: {'lr': 1.3456857348374818e-06, 'samples': 24788480, 'steps': 48414, 'loss/train': 2.2600040435791016} 02/26/2022 16:17:29 - INFO - codeparrot_training - Step 48415: {'lr': 1.3439908359080266e-06, 'samples': 24788992, 'steps': 48415, 'loss/train': 2.49967360496521} 02/26/2022 16:17:33 - INFO - codeparrot_training - Step 48416: {'lr': 1.3422970021419178e-06, 'samples': 24789504, 'steps': 48416, 'loss/train': 0.8635879158973694} 02/26/2022 16:17:39 - INFO - codeparrot_training - Step 48417: {'lr': 1.3406042335464552e-06, 'samples': 24790016, 'steps': 48417, 'loss/train': 2.140446901321411} 02/26/2022 16:17:42 - INFO - codeparrot_training - Step 48418: {'lr': 1.3389125301289107e-06, 'samples': 24790528, 'steps': 48418, 'loss/train': 0.9869257807731628} 02/26/2022 16:17:48 - INFO - codeparrot_training - Step 48419: {'lr': 1.337221891896473e-06, 'samples': 24791040, 'steps': 48419, 'loss/train': 1.6491167545318604} 02/26/2022 16:17:51 - INFO - codeparrot_training - Step 48420: {'lr': 1.3355323188564417e-06, 'samples': 24791552, 'steps': 48420, 'loss/train': 3.9106357097625732} 02/26/2022 16:17:57 - INFO - codeparrot_training - Step 48421: {'lr': 1.3338438110160056e-06, 'samples': 24792064, 'steps': 48421, 'loss/train': 1.005301594734192} 02/26/2022 16:18:00 - INFO - codeparrot_training - Step 48422: {'lr': 1.3321563683824367e-06, 'samples': 24792576, 'steps': 48422, 'loss/train': 1.6604828834533691} 02/26/2022 16:18:06 - INFO - codeparrot_training - Step 48423: {'lr': 1.3304699909629513e-06, 'samples': 24793088, 'steps': 48423, 'loss/train': 1.4465149641036987} 02/26/2022 16:18:09 - INFO - codeparrot_training - Step 48424: {'lr': 1.3287846787647939e-06, 'samples': 24793600, 'steps': 48424, 'loss/train': 1.853302240371704} 02/26/2022 16:18:15 - INFO - codeparrot_training - Step 48425: {'lr': 1.3271004317951252e-06, 'samples': 24794112, 'steps': 48425, 'loss/train': 1.9257960319519043} 02/26/2022 16:18:18 - INFO - codeparrot_training - Step 48426: {'lr': 1.3254172500612171e-06, 'samples': 24794624, 'steps': 48426, 'loss/train': 0.797614336013794} 02/26/2022 16:18:25 - INFO - codeparrot_training - Step 48427: {'lr': 1.3237351335702308e-06, 'samples': 24795136, 'steps': 48427, 'loss/train': 1.9741054773330688} 02/26/2022 16:18:28 - INFO - codeparrot_training - Step 48428: {'lr': 1.3220540823294104e-06, 'samples': 24795648, 'steps': 48428, 'loss/train': 1.6475262641906738} 02/26/2022 16:18:34 - INFO - codeparrot_training - Step 48429: {'lr': 1.320374096345972e-06, 'samples': 24796160, 'steps': 48429, 'loss/train': 1.7777506113052368} 02/26/2022 16:18:37 - INFO - codeparrot_training - Step 48430: {'lr': 1.3186951756270493e-06, 'samples': 24796672, 'steps': 48430, 'loss/train': 1.9330613613128662} 02/26/2022 16:18:43 - INFO - codeparrot_training - Step 48431: {'lr': 1.3170173201798863e-06, 'samples': 24797184, 'steps': 48431, 'loss/train': 1.565766453742981} 02/26/2022 16:18:46 - INFO - codeparrot_training - Step 48432: {'lr': 1.3153405300116717e-06, 'samples': 24797696, 'steps': 48432, 'loss/train': 1.7955403327941895} 02/26/2022 16:18:52 - INFO - codeparrot_training - Step 48433: {'lr': 1.3136648051295663e-06, 'samples': 24798208, 'steps': 48433, 'loss/train': 1.0911246538162231} 02/26/2022 16:18:55 - INFO - codeparrot_training - Step 48434: {'lr': 1.3119901455407313e-06, 'samples': 24798720, 'steps': 48434, 'loss/train': 1.308014988899231} 02/26/2022 16:19:01 - INFO - codeparrot_training - Step 48435: {'lr': 1.310316551252383e-06, 'samples': 24799232, 'steps': 48435, 'loss/train': 1.3888943195343018} 02/26/2022 16:19:04 - INFO - codeparrot_training - Step 48436: {'lr': 1.3086440222716546e-06, 'samples': 24799744, 'steps': 48436, 'loss/train': 1.4703361988067627} 02/26/2022 16:19:10 - INFO - codeparrot_training - Step 48437: {'lr': 1.3069725586057623e-06, 'samples': 24800256, 'steps': 48437, 'loss/train': 1.1862928867340088} 02/26/2022 16:19:14 - INFO - codeparrot_training - Step 48438: {'lr': 1.305302160261812e-06, 'samples': 24800768, 'steps': 48438, 'loss/train': 0.6966497898101807} 02/26/2022 16:19:19 - INFO - codeparrot_training - Step 48439: {'lr': 1.3036328272469644e-06, 'samples': 24801280, 'steps': 48439, 'loss/train': 1.1118990182876587} 02/26/2022 16:19:23 - INFO - codeparrot_training - Step 48440: {'lr': 1.3019645595683804e-06, 'samples': 24801792, 'steps': 48440, 'loss/train': 1.9786264896392822} 02/26/2022 16:19:28 - INFO - codeparrot_training - Step 48441: {'lr': 1.3002973572332211e-06, 'samples': 24802304, 'steps': 48441, 'loss/train': 1.5097253322601318} 02/26/2022 16:19:32 - INFO - codeparrot_training - Step 48442: {'lr': 1.2986312202486195e-06, 'samples': 24802816, 'steps': 48442, 'loss/train': 0.760098397731781} 02/26/2022 16:19:37 - INFO - codeparrot_training - Step 48443: {'lr': 1.296966148621709e-06, 'samples': 24803328, 'steps': 48443, 'loss/train': 0.6391351819038391} 02/26/2022 16:19:41 - INFO - codeparrot_training - Step 48444: {'lr': 1.2953021423596223e-06, 'samples': 24803840, 'steps': 48444, 'loss/train': 2.5851023197174072} 02/26/2022 16:19:46 - INFO - codeparrot_training - Step 48445: {'lr': 1.293639201469493e-06, 'samples': 24804352, 'steps': 48445, 'loss/train': 2.5467782020568848} 02/26/2022 16:19:50 - INFO - codeparrot_training - Step 48446: {'lr': 1.2919773259584266e-06, 'samples': 24804864, 'steps': 48446, 'loss/train': 1.2552155256271362} 02/26/2022 16:19:55 - INFO - codeparrot_training - Step 48447: {'lr': 1.2903165158335838e-06, 'samples': 24805376, 'steps': 48447, 'loss/train': 1.3458584547042847} 02/26/2022 16:19:59 - INFO - codeparrot_training - Step 48448: {'lr': 1.2886567711020148e-06, 'samples': 24805888, 'steps': 48448, 'loss/train': 0.8545482754707336} 02/26/2022 16:20:04 - INFO - codeparrot_training - Step 48449: {'lr': 1.2869980917708801e-06, 'samples': 24806400, 'steps': 48449, 'loss/train': 1.8371524810791016} 02/26/2022 16:20:08 - INFO - codeparrot_training - Step 48450: {'lr': 1.2853404778472856e-06, 'samples': 24806912, 'steps': 48450, 'loss/train': 0.0968320369720459} 02/26/2022 16:20:13 - INFO - codeparrot_training - Step 48451: {'lr': 1.2836839293383085e-06, 'samples': 24807424, 'steps': 48451, 'loss/train': 2.960578203201294} 02/26/2022 16:20:17 - INFO - codeparrot_training - Step 48452: {'lr': 1.2820284462510267e-06, 'samples': 24807936, 'steps': 48452, 'loss/train': 1.594126582145691} 02/26/2022 16:20:23 - INFO - codeparrot_training - Step 48453: {'lr': 1.2803740285926013e-06, 'samples': 24808448, 'steps': 48453, 'loss/train': 1.706684947013855} 02/26/2022 16:20:26 - INFO - codeparrot_training - Step 48454: {'lr': 1.2787206763700265e-06, 'samples': 24808960, 'steps': 48454, 'loss/train': 2.9540576934814453} 02/26/2022 16:20:32 - INFO - codeparrot_training - Step 48455: {'lr': 1.277068389590491e-06, 'samples': 24809472, 'steps': 48455, 'loss/train': 1.9281927347183228} 02/26/2022 16:20:35 - INFO - codeparrot_training - Step 48456: {'lr': 1.2754171682609617e-06, 'samples': 24809984, 'steps': 48456, 'loss/train': 2.2091524600982666} 02/26/2022 16:20:41 - INFO - codeparrot_training - Step 48457: {'lr': 1.2737670123885992e-06, 'samples': 24810496, 'steps': 48457, 'loss/train': 1.5281059741973877} 02/26/2022 16:20:44 - INFO - codeparrot_training - Step 48458: {'lr': 1.2721179219804258e-06, 'samples': 24811008, 'steps': 48458, 'loss/train': 1.6805046796798706} 02/26/2022 16:20:50 - INFO - codeparrot_training - Step 48459: {'lr': 1.270469897043547e-06, 'samples': 24811520, 'steps': 48459, 'loss/train': 2.421518564224243} 02/26/2022 16:20:53 - INFO - codeparrot_training - Step 48460: {'lr': 1.2688229375849568e-06, 'samples': 24812032, 'steps': 48460, 'loss/train': 1.8304585218429565} 02/26/2022 16:20:59 - INFO - codeparrot_training - Step 48461: {'lr': 1.2671770436117613e-06, 'samples': 24812544, 'steps': 48461, 'loss/train': 0.7550714015960693} 02/26/2022 16:21:02 - INFO - codeparrot_training - Step 48462: {'lr': 1.26553221513101e-06, 'samples': 24813056, 'steps': 48462, 'loss/train': 1.0804290771484375} 02/26/2022 16:21:09 - INFO - codeparrot_training - Step 48463: {'lr': 1.263888452149725e-06, 'samples': 24813568, 'steps': 48463, 'loss/train': 1.0991851091384888} 02/26/2022 16:21:12 - INFO - codeparrot_training - Step 48464: {'lr': 1.2622457546749566e-06, 'samples': 24814080, 'steps': 48464, 'loss/train': 1.117771029472351} 02/26/2022 16:21:18 - INFO - codeparrot_training - Step 48465: {'lr': 1.2606041227137544e-06, 'samples': 24814592, 'steps': 48465, 'loss/train': 1.8303498029708862} 02/26/2022 16:21:21 - INFO - codeparrot_training - Step 48466: {'lr': 1.2589635562731405e-06, 'samples': 24815104, 'steps': 48466, 'loss/train': 1.2512751817703247} 02/26/2022 16:21:27 - INFO - codeparrot_training - Step 48467: {'lr': 1.2573240553601374e-06, 'samples': 24815616, 'steps': 48467, 'loss/train': 1.6498616933822632} 02/26/2022 16:21:30 - INFO - codeparrot_training - Step 48468: {'lr': 1.2556856199817668e-06, 'samples': 24816128, 'steps': 48468, 'loss/train': 2.1129143238067627} 02/26/2022 16:21:36 - INFO - codeparrot_training - Step 48469: {'lr': 1.254048250145079e-06, 'samples': 24816640, 'steps': 48469, 'loss/train': 1.7093323469161987} 02/26/2022 16:21:39 - INFO - codeparrot_training - Step 48470: {'lr': 1.2524119458570126e-06, 'samples': 24817152, 'steps': 48470, 'loss/train': 2.3321712017059326} 02/26/2022 16:21:45 - INFO - codeparrot_training - Step 48471: {'lr': 1.2507767071246734e-06, 'samples': 24817664, 'steps': 48471, 'loss/train': 1.8468883037567139} 02/26/2022 16:21:48 - INFO - codeparrot_training - Step 48472: {'lr': 1.249142533955e-06, 'samples': 24818176, 'steps': 48472, 'loss/train': 1.765306830406189} 02/26/2022 16:21:54 - INFO - codeparrot_training - Step 48473: {'lr': 1.2475094263550423e-06, 'samples': 24818688, 'steps': 48473, 'loss/train': 1.2802129983901978} 02/26/2022 16:21:58 - INFO - codeparrot_training - Step 48474: {'lr': 1.2458773843317118e-06, 'samples': 24819200, 'steps': 48474, 'loss/train': 1.5911240577697754} 02/26/2022 16:22:03 - INFO - codeparrot_training - Step 48475: {'lr': 1.2442464078920855e-06, 'samples': 24819712, 'steps': 48475, 'loss/train': 0.43828141689300537} 02/26/2022 16:22:07 - INFO - codeparrot_training - Step 48476: {'lr': 1.242616497043131e-06, 'samples': 24820224, 'steps': 48476, 'loss/train': 2.0459420680999756} 02/26/2022 16:22:12 - INFO - codeparrot_training - Step 48477: {'lr': 1.2409876517918139e-06, 'samples': 24820736, 'steps': 48477, 'loss/train': 1.2960195541381836} 02/26/2022 16:22:16 - INFO - codeparrot_training - Step 48478: {'lr': 1.2393598721451017e-06, 'samples': 24821248, 'steps': 48478, 'loss/train': 2.3047893047332764} 02/26/2022 16:22:22 - INFO - codeparrot_training - Step 48479: {'lr': 1.2377331581099883e-06, 'samples': 24821760, 'steps': 48479, 'loss/train': 0.30409908294677734} 02/26/2022 16:22:25 - INFO - codeparrot_training - Step 48480: {'lr': 1.2361075096934404e-06, 'samples': 24822272, 'steps': 48480, 'loss/train': 1.5186514854431152} 02/26/2022 16:22:31 - INFO - codeparrot_training - Step 48481: {'lr': 1.2344829269023972e-06, 'samples': 24822784, 'steps': 48481, 'loss/train': 0.8618290424346924} 02/26/2022 16:22:34 - INFO - codeparrot_training - Step 48482: {'lr': 1.2328594097438528e-06, 'samples': 24823296, 'steps': 48482, 'loss/train': 2.224228858947754} 02/26/2022 16:22:40 - INFO - codeparrot_training - Step 48483: {'lr': 1.2312369582247462e-06, 'samples': 24823808, 'steps': 48483, 'loss/train': 1.2652878761291504} 02/26/2022 16:22:43 - INFO - codeparrot_training - Step 48484: {'lr': 1.2296155723520163e-06, 'samples': 24824320, 'steps': 48484, 'loss/train': 2.494157314300537} 02/26/2022 16:22:49 - INFO - codeparrot_training - Step 48485: {'lr': 1.2279952521326298e-06, 'samples': 24824832, 'steps': 48485, 'loss/train': 1.5959936380386353} 02/26/2022 16:22:53 - INFO - codeparrot_training - Step 48486: {'lr': 1.2263759975735257e-06, 'samples': 24825344, 'steps': 48486, 'loss/train': 1.6957417726516724} 02/26/2022 16:22:58 - INFO - codeparrot_training - Step 48487: {'lr': 1.2247578086816148e-06, 'samples': 24825856, 'steps': 48487, 'loss/train': 1.5053510665893555} 02/26/2022 16:23:01 - INFO - codeparrot_training - Step 48488: {'lr': 1.223140685463864e-06, 'samples': 24826368, 'steps': 48488, 'loss/train': 1.4984582662582397} 02/26/2022 16:23:07 - INFO - codeparrot_training - Step 48489: {'lr': 1.2215246279271563e-06, 'samples': 24826880, 'steps': 48489, 'loss/train': 2.164133071899414} 02/26/2022 16:23:11 - INFO - codeparrot_training - Step 48490: {'lr': 1.219909636078459e-06, 'samples': 24827392, 'steps': 48490, 'loss/train': 1.4851778745651245} 02/26/2022 16:23:16 - INFO - codeparrot_training - Step 48491: {'lr': 1.2182957099246828e-06, 'samples': 24827904, 'steps': 48491, 'loss/train': 2.2463409900665283} 02/26/2022 16:23:20 - INFO - codeparrot_training - Step 48492: {'lr': 1.2166828494727112e-06, 'samples': 24828416, 'steps': 48492, 'loss/train': 1.7992091178894043} 02/26/2022 16:23:25 - INFO - codeparrot_training - Step 48493: {'lr': 1.215071054729483e-06, 'samples': 24828928, 'steps': 48493, 'loss/train': 2.279757499694824} 02/26/2022 16:23:29 - INFO - codeparrot_training - Step 48494: {'lr': 1.2134603257018817e-06, 'samples': 24829440, 'steps': 48494, 'loss/train': 2.6845078468322754} 02/26/2022 16:23:34 - INFO - codeparrot_training - Step 48495: {'lr': 1.2118506623968184e-06, 'samples': 24829952, 'steps': 48495, 'loss/train': 2.1672542095184326} 02/26/2022 16:23:38 - INFO - codeparrot_training - Step 48496: {'lr': 1.2102420648212041e-06, 'samples': 24830464, 'steps': 48496, 'loss/train': 1.6132885217666626} 02/26/2022 16:23:43 - INFO - codeparrot_training - Step 48497: {'lr': 1.2086345329818949e-06, 'samples': 24830976, 'steps': 48497, 'loss/train': 2.910344362258911} 02/26/2022 16:23:47 - INFO - codeparrot_training - Step 48498: {'lr': 1.2070280668858292e-06, 'samples': 24831488, 'steps': 48498, 'loss/train': 1.2739242315292358} 02/26/2022 16:23:53 - INFO - codeparrot_training - Step 48499: {'lr': 1.205422666539807e-06, 'samples': 24832000, 'steps': 48499, 'loss/train': 1.6384413242340088} 02/26/2022 16:23:57 - INFO - codeparrot_training - Step 48500: {'lr': 1.2038183319507957e-06, 'samples': 24832512, 'steps': 48500, 'loss/train': 2.0628714561462402} 02/26/2022 16:24:02 - INFO - codeparrot_training - Step 48501: {'lr': 1.2022150631255947e-06, 'samples': 24833024, 'steps': 48501, 'loss/train': 2.5776493549346924} 02/26/2022 16:24:06 - INFO - codeparrot_training - Step 48502: {'lr': 1.2006128600711152e-06, 'samples': 24833536, 'steps': 48502, 'loss/train': 1.632768154144287} 02/26/2022 16:24:11 - INFO - codeparrot_training - Step 48503: {'lr': 1.1990117227941855e-06, 'samples': 24834048, 'steps': 48503, 'loss/train': 1.8911515474319458} 02/26/2022 16:24:14 - INFO - codeparrot_training - Step 48504: {'lr': 1.1974116513017163e-06, 'samples': 24834560, 'steps': 48504, 'loss/train': 1.5433056354522705} 02/26/2022 16:24:20 - INFO - codeparrot_training - Step 48505: {'lr': 1.1958126456005081e-06, 'samples': 24835072, 'steps': 48505, 'loss/train': 1.962333083152771} 02/26/2022 16:24:24 - INFO - codeparrot_training - Step 48506: {'lr': 1.1942147056974717e-06, 'samples': 24835584, 'steps': 48506, 'loss/train': 1.8714172840118408} 02/26/2022 16:24:29 - INFO - codeparrot_training - Step 48507: {'lr': 1.1926178315993796e-06, 'samples': 24836096, 'steps': 48507, 'loss/train': 1.2034095525741577} 02/26/2022 16:24:33 - INFO - codeparrot_training - Step 48508: {'lr': 1.1910220233131153e-06, 'samples': 24836608, 'steps': 48508, 'loss/train': 0.48684197664260864} 02/26/2022 16:24:39 - INFO - codeparrot_training - Step 48509: {'lr': 1.1894272808455064e-06, 'samples': 24837120, 'steps': 48509, 'loss/train': 1.3360004425048828} 02/26/2022 16:24:42 - INFO - codeparrot_training - Step 48510: {'lr': 1.187833604203381e-06, 'samples': 24837632, 'steps': 48510, 'loss/train': 1.5148578882217407} 02/26/2022 16:24:48 - INFO - codeparrot_training - Step 48511: {'lr': 1.1862409933935947e-06, 'samples': 24838144, 'steps': 48511, 'loss/train': 1.8555047512054443} 02/26/2022 16:24:51 - INFO - codeparrot_training - Step 48512: {'lr': 1.1846494484229198e-06, 'samples': 24838656, 'steps': 48512, 'loss/train': 1.1072171926498413} 02/26/2022 16:24:57 - INFO - codeparrot_training - Step 48513: {'lr': 1.1830589692982118e-06, 'samples': 24839168, 'steps': 48513, 'loss/train': 1.6367236375808716} 02/26/2022 16:25:00 - INFO - codeparrot_training - Step 48514: {'lr': 1.181469556026271e-06, 'samples': 24839680, 'steps': 48514, 'loss/train': 1.1613576412200928} 02/26/2022 16:25:06 - INFO - codeparrot_training - Step 48515: {'lr': 1.1798812086138977e-06, 'samples': 24840192, 'steps': 48515, 'loss/train': 1.9894667863845825} 02/26/2022 16:25:09 - INFO - codeparrot_training - Step 48516: {'lr': 1.1782939270678917e-06, 'samples': 24840704, 'steps': 48516, 'loss/train': 1.737265944480896} 02/26/2022 16:25:15 - INFO - codeparrot_training - Step 48517: {'lr': 1.1767077113950807e-06, 'samples': 24841216, 'steps': 48517, 'loss/train': 0.4749722182750702} 02/26/2022 16:25:18 - INFO - codeparrot_training - Step 48518: {'lr': 1.1751225616022376e-06, 'samples': 24841728, 'steps': 48518, 'loss/train': 1.4081003665924072} 02/26/2022 16:25:24 - INFO - codeparrot_training - Step 48519: {'lr': 1.1735384776961622e-06, 'samples': 24842240, 'steps': 48519, 'loss/train': 2.072645425796509} 02/26/2022 16:25:28 - INFO - codeparrot_training - Step 48520: {'lr': 1.1719554596836545e-06, 'samples': 24842752, 'steps': 48520, 'loss/train': 1.3582074642181396} 02/26/2022 16:25:33 - INFO - codeparrot_training - Step 48521: {'lr': 1.1703735075714594e-06, 'samples': 24843264, 'steps': 48521, 'loss/train': 1.6174136400222778} 02/26/2022 16:25:37 - INFO - codeparrot_training - Step 48522: {'lr': 1.1687926213663769e-06, 'samples': 24843776, 'steps': 48522, 'loss/train': 1.0460364818572998} 02/26/2022 16:25:42 - INFO - codeparrot_training - Step 48523: {'lr': 1.1672128010751515e-06, 'samples': 24844288, 'steps': 48523, 'loss/train': 2.1129560470581055} 02/26/2022 16:25:46 - INFO - codeparrot_training - Step 48524: {'lr': 1.165634046704611e-06, 'samples': 24844800, 'steps': 48524, 'loss/train': 1.6464053392410278} 02/26/2022 16:25:51 - INFO - codeparrot_training - Step 48525: {'lr': 1.164056358261445e-06, 'samples': 24845312, 'steps': 48525, 'loss/train': 2.392225980758667} 02/26/2022 16:25:55 - INFO - codeparrot_training - Step 48526: {'lr': 1.162479735752453e-06, 'samples': 24845824, 'steps': 48526, 'loss/train': 1.4088304042816162} 02/26/2022 16:26:00 - INFO - codeparrot_training - Step 48527: {'lr': 1.1609041791844077e-06, 'samples': 24846336, 'steps': 48527, 'loss/train': 2.3873565196990967} 02/26/2022 16:26:03 - INFO - codeparrot_training - Step 48528: {'lr': 1.1593296885640259e-06, 'samples': 24846848, 'steps': 48528, 'loss/train': 1.412457823753357} 02/26/2022 16:26:09 - INFO - codeparrot_training - Step 48529: {'lr': 1.157756263898052e-06, 'samples': 24847360, 'steps': 48529, 'loss/train': 1.4757179021835327} 02/26/2022 16:26:12 - INFO - codeparrot_training - Step 48530: {'lr': 1.156183905193231e-06, 'samples': 24847872, 'steps': 48530, 'loss/train': 2.0022785663604736} 02/26/2022 16:26:19 - INFO - codeparrot_training - Step 48531: {'lr': 1.1546126124563072e-06, 'samples': 24848384, 'steps': 48531, 'loss/train': 1.9121297597885132} 02/26/2022 16:26:23 - INFO - codeparrot_training - Step 48532: {'lr': 1.1530423856939975e-06, 'samples': 24848896, 'steps': 48532, 'loss/train': 0.6375647187232971} 02/26/2022 16:26:28 - INFO - codeparrot_training - Step 48533: {'lr': 1.1514732249130466e-06, 'samples': 24849408, 'steps': 48533, 'loss/train': 1.8892377614974976} 02/26/2022 16:26:32 - INFO - codeparrot_training - Step 48534: {'lr': 1.1499051301201713e-06, 'samples': 24849920, 'steps': 48534, 'loss/train': 1.2111845016479492} 02/26/2022 16:26:37 - INFO - codeparrot_training - Step 48535: {'lr': 1.1483381013220607e-06, 'samples': 24850432, 'steps': 48535, 'loss/train': 1.4634150266647339} 02/26/2022 16:26:41 - INFO - codeparrot_training - Step 48536: {'lr': 1.1467721385254593e-06, 'samples': 24850944, 'steps': 48536, 'loss/train': 0.2927151620388031} 02/26/2022 16:26:46 - INFO - codeparrot_training - Step 48537: {'lr': 1.145207241737084e-06, 'samples': 24851456, 'steps': 48537, 'loss/train': 2.428645133972168} 02/26/2022 16:26:50 - INFO - codeparrot_training - Step 48538: {'lr': 1.1436434109635962e-06, 'samples': 24851968, 'steps': 48538, 'loss/train': 1.378011703491211} 02/26/2022 16:26:55 - INFO - codeparrot_training - Step 48539: {'lr': 1.1420806462117406e-06, 'samples': 24852480, 'steps': 48539, 'loss/train': 1.6328016519546509} 02/26/2022 16:26:59 - INFO - codeparrot_training - Step 48540: {'lr': 1.1405189474881783e-06, 'samples': 24852992, 'steps': 48540, 'loss/train': 1.3970850706100464} 02/26/2022 16:27:05 - INFO - codeparrot_training - Step 48541: {'lr': 1.1389583147996262e-06, 'samples': 24853504, 'steps': 48541, 'loss/train': 0.06693976372480392} 02/26/2022 16:27:09 - INFO - codeparrot_training - Step 48542: {'lr': 1.1373987481527459e-06, 'samples': 24854016, 'steps': 48542, 'loss/train': 0.04737776517868042} 02/26/2022 16:27:14 - INFO - codeparrot_training - Step 48543: {'lr': 1.1358402475542261e-06, 'samples': 24854528, 'steps': 48543, 'loss/train': 0.8508061170578003} 02/26/2022 16:27:18 - INFO - codeparrot_training - Step 48544: {'lr': 1.1342828130107286e-06, 'samples': 24855040, 'steps': 48544, 'loss/train': 2.413389205932617} 02/26/2022 16:27:23 - INFO - codeparrot_training - Step 48545: {'lr': 1.132726444528942e-06, 'samples': 24855552, 'steps': 48545, 'loss/train': 1.0434762239456177} 02/26/2022 16:27:27 - INFO - codeparrot_training - Step 48546: {'lr': 1.1311711421155556e-06, 'samples': 24856064, 'steps': 48546, 'loss/train': 1.6834280490875244} 02/26/2022 16:27:32 - INFO - codeparrot_training - Step 48547: {'lr': 1.1296169057771756e-06, 'samples': 24856576, 'steps': 48547, 'loss/train': 2.1511762142181396} 02/26/2022 16:27:36 - INFO - codeparrot_training - Step 48548: {'lr': 1.1280637355205182e-06, 'samples': 24857088, 'steps': 48548, 'loss/train': 1.6215265989303589} 02/26/2022 16:27:41 - INFO - codeparrot_training - Step 48549: {'lr': 1.1265116313521894e-06, 'samples': 24857600, 'steps': 48549, 'loss/train': 1.106916069984436} 02/26/2022 16:27:45 - INFO - codeparrot_training - Step 48550: {'lr': 1.124960593278851e-06, 'samples': 24858112, 'steps': 48550, 'loss/train': 1.8584131002426147} 02/26/2022 16:27:51 - INFO - codeparrot_training - Step 48551: {'lr': 1.1234106213071915e-06, 'samples': 24858624, 'steps': 48551, 'loss/train': 1.3429499864578247} 02/26/2022 16:27:54 - INFO - codeparrot_training - Step 48552: {'lr': 1.1218617154437893e-06, 'samples': 24859136, 'steps': 48552, 'loss/train': 0.07465500384569168} 02/26/2022 16:28:00 - INFO - codeparrot_training - Step 48553: {'lr': 1.1203138756953058e-06, 'samples': 24859648, 'steps': 48553, 'loss/train': 1.2795872688293457} 02/26/2022 16:28:03 - INFO - codeparrot_training - Step 48554: {'lr': 1.1187671020683743e-06, 'samples': 24860160, 'steps': 48554, 'loss/train': 1.0946496725082397} 02/26/2022 16:28:09 - INFO - codeparrot_training - Step 48555: {'lr': 1.1172213945696286e-06, 'samples': 24860672, 'steps': 48555, 'loss/train': 1.3925013542175293} 02/26/2022 16:28:13 - INFO - codeparrot_training - Step 48556: {'lr': 1.1156767532056467e-06, 'samples': 24861184, 'steps': 48556, 'loss/train': 2.6532912254333496} 02/26/2022 16:28:18 - INFO - codeparrot_training - Step 48557: {'lr': 1.11413317798309e-06, 'samples': 24861696, 'steps': 48557, 'loss/train': 1.9645737409591675} 02/26/2022 16:28:22 - INFO - codeparrot_training - Step 48558: {'lr': 1.1125906689085641e-06, 'samples': 24862208, 'steps': 48558, 'loss/train': 1.743770718574524} 02/26/2022 16:28:27 - INFO - codeparrot_training - Step 48559: {'lr': 1.111049225988675e-06, 'samples': 24862720, 'steps': 48559, 'loss/train': 2.0174946784973145} 02/26/2022 16:28:31 - INFO - codeparrot_training - Step 48560: {'lr': 1.109508849230001e-06, 'samples': 24863232, 'steps': 48560, 'loss/train': 1.5074763298034668} 02/26/2022 16:28:36 - INFO - codeparrot_training - Step 48561: {'lr': 1.1079695386391754e-06, 'samples': 24863744, 'steps': 48561, 'loss/train': 1.159696102142334} 02/26/2022 16:28:40 - INFO - codeparrot_training - Step 48562: {'lr': 1.1064312942227484e-06, 'samples': 24864256, 'steps': 48562, 'loss/train': 1.5916857719421387} 02/26/2022 16:28:45 - INFO - codeparrot_training - Step 48563: {'lr': 1.1048941159873538e-06, 'samples': 24864768, 'steps': 48563, 'loss/train': 1.5484447479248047} 02/26/2022 16:28:48 - INFO - codeparrot_training - Step 48564: {'lr': 1.1033580039395697e-06, 'samples': 24865280, 'steps': 48564, 'loss/train': 2.149129867553711} 02/26/2022 16:28:55 - INFO - codeparrot_training - Step 48565: {'lr': 1.101822958085974e-06, 'samples': 24865792, 'steps': 48565, 'loss/train': 1.097700595855713} 02/26/2022 16:28:58 - INFO - codeparrot_training - Step 48566: {'lr': 1.100288978433117e-06, 'samples': 24866304, 'steps': 48566, 'loss/train': 1.9912561178207397} 02/26/2022 16:29:04 - INFO - codeparrot_training - Step 48567: {'lr': 1.0987560649875771e-06, 'samples': 24866816, 'steps': 48567, 'loss/train': 1.572648286819458} 02/26/2022 16:29:07 - INFO - codeparrot_training - Step 48568: {'lr': 1.0972242177559599e-06, 'samples': 24867328, 'steps': 48568, 'loss/train': 1.4035009145736694} 02/26/2022 16:29:13 - INFO - codeparrot_training - Step 48569: {'lr': 1.0956934367447601e-06, 'samples': 24867840, 'steps': 48569, 'loss/train': 3.298931121826172} 02/26/2022 16:29:16 - INFO - codeparrot_training - Step 48570: {'lr': 1.0941637219606114e-06, 'samples': 24868352, 'steps': 48570, 'loss/train': 0.9593859314918518} 02/26/2022 16:29:20 - INFO - codeparrot_training - Step 48571: {'lr': 1.0926350734100088e-06, 'samples': 24868864, 'steps': 48571, 'loss/train': 2.4536402225494385} 02/26/2022 16:29:25 - INFO - codeparrot_training - Step 48572: {'lr': 1.09110749109953e-06, 'samples': 24869376, 'steps': 48572, 'loss/train': 2.2696402072906494} 02/26/2022 16:29:29 - INFO - codeparrot_training - Step 48573: {'lr': 1.0895809750356978e-06, 'samples': 24869888, 'steps': 48573, 'loss/train': 1.0938876867294312} 02/26/2022 16:29:34 - INFO - codeparrot_training - Step 48574: {'lr': 1.0880555252250623e-06, 'samples': 24870400, 'steps': 48574, 'loss/train': 0.8961057066917419} 02/26/2022 16:29:38 - INFO - codeparrot_training - Step 48575: {'lr': 1.0865311416741742e-06, 'samples': 24870912, 'steps': 48575, 'loss/train': 2.8931827545166016} 02/26/2022 16:29:43 - INFO - codeparrot_training - Step 48576: {'lr': 1.0850078243895278e-06, 'samples': 24871424, 'steps': 48576, 'loss/train': 1.0553991794586182} 02/26/2022 16:29:47 - INFO - codeparrot_training - Step 48577: {'lr': 1.0834855733777017e-06, 'samples': 24871936, 'steps': 48577, 'loss/train': 1.8819741010665894} 02/26/2022 16:29:53 - INFO - codeparrot_training - Step 48578: {'lr': 1.0819643886451625e-06, 'samples': 24872448, 'steps': 48578, 'loss/train': 1.4161831140518188} 02/26/2022 16:29:57 - INFO - codeparrot_training - Step 48579: {'lr': 1.0804442701984329e-06, 'samples': 24872960, 'steps': 48579, 'loss/train': 1.3909871578216553} 02/26/2022 16:30:02 - INFO - codeparrot_training - Step 48580: {'lr': 1.0789252180440635e-06, 'samples': 24873472, 'steps': 48580, 'loss/train': 1.8278183937072754} 02/26/2022 16:30:06 - INFO - codeparrot_training - Step 48581: {'lr': 1.077407232188521e-06, 'samples': 24873984, 'steps': 48581, 'loss/train': 1.3987740278244019} 02/26/2022 16:30:11 - INFO - codeparrot_training - Step 48582: {'lr': 1.075890312638328e-06, 'samples': 24874496, 'steps': 48582, 'loss/train': 0.39701828360557556} 02/26/2022 16:30:15 - INFO - codeparrot_training - Step 48583: {'lr': 1.0743744593999794e-06, 'samples': 24875008, 'steps': 48583, 'loss/train': 2.290710210800171} 02/26/2022 16:30:20 - INFO - codeparrot_training - Step 48584: {'lr': 1.07285967247997e-06, 'samples': 24875520, 'steps': 48584, 'loss/train': 1.753527283668518} 02/26/2022 16:30:24 - INFO - codeparrot_training - Step 48585: {'lr': 1.0713459518847667e-06, 'samples': 24876032, 'steps': 48585, 'loss/train': 1.1332815885543823} 02/26/2022 16:30:30 - INFO - codeparrot_training - Step 48586: {'lr': 1.0698332976209201e-06, 'samples': 24876544, 'steps': 48586, 'loss/train': 2.501163959503174} 02/26/2022 16:30:33 - INFO - codeparrot_training - Step 48587: {'lr': 1.0683217096948138e-06, 'samples': 24877056, 'steps': 48587, 'loss/train': 2.2284014225006104} 02/26/2022 16:30:39 - INFO - codeparrot_training - Step 48588: {'lr': 1.0668111881129982e-06, 'samples': 24877568, 'steps': 48588, 'loss/train': 2.13496470451355} 02/26/2022 16:30:42 - INFO - codeparrot_training - Step 48589: {'lr': 1.0653017328819126e-06, 'samples': 24878080, 'steps': 48589, 'loss/train': 1.9204634428024292} 02/26/2022 16:30:48 - INFO - codeparrot_training - Step 48590: {'lr': 1.0637933440080516e-06, 'samples': 24878592, 'steps': 48590, 'loss/train': 1.3442003726959229} 02/26/2022 16:30:51 - INFO - codeparrot_training - Step 48591: {'lr': 1.0622860214978269e-06, 'samples': 24879104, 'steps': 48591, 'loss/train': 1.264943242073059} 02/26/2022 16:30:57 - INFO - codeparrot_training - Step 48592: {'lr': 1.0607797653577334e-06, 'samples': 24879616, 'steps': 48592, 'loss/train': 1.5720057487487793} 02/26/2022 16:31:00 - INFO - codeparrot_training - Step 48593: {'lr': 1.0592745755941823e-06, 'samples': 24880128, 'steps': 48593, 'loss/train': 1.7504733800888062} 02/26/2022 16:31:05 - INFO - codeparrot_training - Step 48594: {'lr': 1.0577704522136967e-06, 'samples': 24880640, 'steps': 48594, 'loss/train': 1.8449431657791138} 02/26/2022 16:31:09 - INFO - codeparrot_training - Step 48595: {'lr': 1.056267395222632e-06, 'samples': 24881152, 'steps': 48595, 'loss/train': 1.8730652332305908} 02/26/2022 16:31:16 - INFO - codeparrot_training - Step 48596: {'lr': 1.0547654046275114e-06, 'samples': 24881664, 'steps': 48596, 'loss/train': 1.0368913412094116} 02/26/2022 16:31:19 - INFO - codeparrot_training - Step 48597: {'lr': 1.0532644804346902e-06, 'samples': 24882176, 'steps': 48597, 'loss/train': 1.398057222366333} 02/26/2022 16:31:25 - INFO - codeparrot_training - Step 48598: {'lr': 1.0517646226506638e-06, 'samples': 24882688, 'steps': 48598, 'loss/train': 1.4415875673294067} 02/26/2022 16:31:30 - INFO - codeparrot_training - Step 48599: {'lr': 1.0502658312818159e-06, 'samples': 24883200, 'steps': 48599, 'loss/train': 1.8884658813476562} 02/26/2022 16:31:34 - INFO - codeparrot_training - Step 48600: {'lr': 1.0487681063345856e-06, 'samples': 24883712, 'steps': 48600, 'loss/train': 2.6032891273498535} 02/26/2022 16:31:40 - INFO - codeparrot_training - Step 48601: {'lr': 1.0472714478153567e-06, 'samples': 24884224, 'steps': 48601, 'loss/train': 2.009718656539917} 02/26/2022 16:31:43 - INFO - codeparrot_training - Step 48602: {'lr': 1.0457758557305687e-06, 'samples': 24884736, 'steps': 48602, 'loss/train': 2.674697160720825} 02/26/2022 16:31:49 - INFO - codeparrot_training - Step 48603: {'lr': 1.0442813300866605e-06, 'samples': 24885248, 'steps': 48603, 'loss/train': 1.2860231399536133} 02/26/2022 16:31:52 - INFO - codeparrot_training - Step 48604: {'lr': 1.0427878708899608e-06, 'samples': 24885760, 'steps': 48604, 'loss/train': 0.46338459849357605} 02/26/2022 16:31:56 - INFO - codeparrot_training - Step 48605: {'lr': 1.0412954781469363e-06, 'samples': 24886272, 'steps': 48605, 'loss/train': 0.6531505584716797} 02/26/2022 16:32:02 - INFO - codeparrot_training - Step 48606: {'lr': 1.0398041518639157e-06, 'samples': 24886784, 'steps': 48606, 'loss/train': 1.8503721952438354} 02/26/2022 16:32:06 - INFO - codeparrot_training - Step 48607: {'lr': 1.0383138920473378e-06, 'samples': 24887296, 'steps': 48607, 'loss/train': 1.4208728075027466} 02/26/2022 16:32:11 - INFO - codeparrot_training - Step 48608: {'lr': 1.0368246987035868e-06, 'samples': 24887808, 'steps': 48608, 'loss/train': 0.08011391758918762} 02/26/2022 16:32:15 - INFO - codeparrot_training - Step 48609: {'lr': 1.0353365718390184e-06, 'samples': 24888320, 'steps': 48609, 'loss/train': 1.9577525854110718} 02/26/2022 16:32:20 - INFO - codeparrot_training - Step 48610: {'lr': 1.0338495114599887e-06, 'samples': 24888832, 'steps': 48610, 'loss/train': 0.777114987373352} 02/26/2022 16:32:24 - INFO - codeparrot_training - Step 48611: {'lr': 1.0323635175729095e-06, 'samples': 24889344, 'steps': 48611, 'loss/train': 1.9848923683166504} 02/26/2022 16:32:29 - INFO - codeparrot_training - Step 48612: {'lr': 1.030878590184109e-06, 'samples': 24889856, 'steps': 48612, 'loss/train': 1.8646644353866577} 02/26/2022 16:32:33 - INFO - codeparrot_training - Step 48613: {'lr': 1.0293947292999984e-06, 'samples': 24890368, 'steps': 48613, 'loss/train': 2.247981309890747} 02/26/2022 16:32:38 - INFO - codeparrot_training - Step 48614: {'lr': 1.0279119349268783e-06, 'samples': 24890880, 'steps': 48614, 'loss/train': 1.999135971069336} 02/26/2022 16:32:42 - INFO - codeparrot_training - Step 48615: {'lr': 1.0264302070711328e-06, 'samples': 24891392, 'steps': 48615, 'loss/train': 1.5612785816192627} 02/26/2022 16:32:47 - INFO - codeparrot_training - Step 48616: {'lr': 1.0249495457391177e-06, 'samples': 24891904, 'steps': 48616, 'loss/train': 2.289363384246826} 02/26/2022 16:32:51 - INFO - codeparrot_training - Step 48617: {'lr': 1.0234699509371336e-06, 'samples': 24892416, 'steps': 48617, 'loss/train': 2.0629048347473145} 02/26/2022 16:32:58 - INFO - codeparrot_training - Step 48618: {'lr': 1.0219914226715643e-06, 'samples': 24892928, 'steps': 48618, 'loss/train': 2.570551872253418} 02/26/2022 16:33:03 - INFO - codeparrot_training - Step 48619: {'lr': 1.0205139609487102e-06, 'samples': 24893440, 'steps': 48619, 'loss/train': 1.9808770418167114} 02/26/2022 16:33:07 - INFO - codeparrot_training - Step 48620: {'lr': 1.0190375657749273e-06, 'samples': 24893952, 'steps': 48620, 'loss/train': 1.8617417812347412} 02/26/2022 16:33:13 - INFO - codeparrot_training - Step 48621: {'lr': 1.017562237156544e-06, 'samples': 24894464, 'steps': 48621, 'loss/train': 1.5349764823913574} 02/26/2022 16:33:16 - INFO - codeparrot_training - Step 48622: {'lr': 1.016087975099833e-06, 'samples': 24894976, 'steps': 48622, 'loss/train': 5.161903381347656} 02/26/2022 16:33:22 - INFO - codeparrot_training - Step 48623: {'lr': 1.0146147796111505e-06, 'samples': 24895488, 'steps': 48623, 'loss/train': 1.5258817672729492} 02/26/2022 16:33:25 - INFO - codeparrot_training - Step 48624: {'lr': 1.0131426506967689e-06, 'samples': 24896000, 'steps': 48624, 'loss/train': 1.4922313690185547} 02/26/2022 16:33:31 - INFO - codeparrot_training - Step 48625: {'lr': 1.0116715883630723e-06, 'samples': 24896512, 'steps': 48625, 'loss/train': 0.6071093678474426} 02/26/2022 16:33:34 - INFO - codeparrot_training - Step 48626: {'lr': 1.0102015926162778e-06, 'samples': 24897024, 'steps': 48626, 'loss/train': 0.1705748289823532} 02/26/2022 16:33:40 - INFO - codeparrot_training - Step 48627: {'lr': 1.0087326634627415e-06, 'samples': 24897536, 'steps': 48627, 'loss/train': 0.19413359463214874} 02/26/2022 16:33:44 - INFO - codeparrot_training - Step 48628: {'lr': 1.0072648009087083e-06, 'samples': 24898048, 'steps': 48628, 'loss/train': 2.0149431228637695} 02/26/2022 16:33:50 - INFO - codeparrot_training - Step 48629: {'lr': 1.0057980049605065e-06, 'samples': 24898560, 'steps': 48629, 'loss/train': 1.406548261642456} 02/26/2022 16:33:53 - INFO - codeparrot_training - Step 48630: {'lr': 1.0043322756243812e-06, 'samples': 24899072, 'steps': 48630, 'loss/train': 1.3164587020874023} 02/26/2022 16:33:57 - INFO - codeparrot_training - Step 48631: {'lr': 1.0028676129066327e-06, 'samples': 24899584, 'steps': 48631, 'loss/train': 1.6917215585708618} 02/26/2022 16:34:02 - INFO - codeparrot_training - Step 48632: {'lr': 1.001404016813534e-06, 'samples': 24900096, 'steps': 48632, 'loss/train': 0.969963014125824} 02/26/2022 16:34:06 - INFO - codeparrot_training - Step 48633: {'lr': 9.999414873513857e-07, 'samples': 24900608, 'steps': 48633, 'loss/train': 1.9355295896530151} 02/26/2022 16:34:11 - INFO - codeparrot_training - Step 48634: {'lr': 9.984800245264047e-07, 'samples': 24901120, 'steps': 48634, 'loss/train': 3.2074432373046875} 02/26/2022 16:34:17 - INFO - codeparrot_training - Step 48635: {'lr': 9.97019628344864e-07, 'samples': 24901632, 'steps': 48635, 'loss/train': 2.0678586959838867} 02/26/2022 16:34:20 - INFO - codeparrot_training - Step 48636: {'lr': 9.955602988130085e-07, 'samples': 24902144, 'steps': 48636, 'loss/train': 2.0958006381988525} 02/26/2022 16:34:26 - INFO - codeparrot_training - Step 48637: {'lr': 9.94102035937111e-07, 'samples': 24902656, 'steps': 48637, 'loss/train': 1.2168176174163818} 02/26/2022 16:34:29 - INFO - codeparrot_training - Step 48638: {'lr': 9.926448397234167e-07, 'samples': 24903168, 'steps': 48638, 'loss/train': 1.7434332370758057} 02/26/2022 16:34:35 - INFO - codeparrot_training - Step 48639: {'lr': 9.911887101781703e-07, 'samples': 24903680, 'steps': 48639, 'loss/train': 2.728332996368408} 02/26/2022 16:34:38 - INFO - codeparrot_training - Step 48640: {'lr': 9.897336473076167e-07, 'samples': 24904192, 'steps': 48640, 'loss/train': 1.2537264823913574} 02/26/2022 16:34:42 - INFO - codeparrot_training - Step 48641: {'lr': 9.882796511179459e-07, 'samples': 24904704, 'steps': 48641, 'loss/train': 0.8997842669487} 02/26/2022 16:34:47 - INFO - codeparrot_training - Step 48642: {'lr': 9.868267216154303e-07, 'samples': 24905216, 'steps': 48642, 'loss/train': 1.4228382110595703} 02/26/2022 16:34:51 - INFO - codeparrot_training - Step 48643: {'lr': 9.853748588062873e-07, 'samples': 24905728, 'steps': 48643, 'loss/train': 1.6537553071975708} 02/26/2022 16:34:57 - INFO - codeparrot_training - Step 48644: {'lr': 9.83924062696706e-07, 'samples': 24906240, 'steps': 48644, 'loss/train': 3.461794376373291} 02/26/2022 16:35:02 - INFO - codeparrot_training - Step 48645: {'lr': 9.82474333292932e-07, 'samples': 24906752, 'steps': 48645, 'loss/train': 2.364292621612549} 02/26/2022 16:35:06 - INFO - codeparrot_training - Step 48646: {'lr': 9.810256706011545e-07, 'samples': 24907264, 'steps': 48646, 'loss/train': 1.7181931734085083} 02/26/2022 16:35:11 - INFO - codeparrot_training - Step 48647: {'lr': 9.795780746275906e-07, 'samples': 24907776, 'steps': 48647, 'loss/train': 1.9545812606811523} 02/26/2022 16:35:15 - INFO - codeparrot_training - Step 48648: {'lr': 9.78131545378458e-07, 'samples': 24908288, 'steps': 48648, 'loss/train': 0.8487032055854797} 02/26/2022 16:35:21 - INFO - codeparrot_training - Step 48649: {'lr': 9.766860828599178e-07, 'samples': 24908800, 'steps': 48649, 'loss/train': 2.215221405029297} 02/26/2022 16:35:24 - INFO - codeparrot_training - Step 48650: {'lr': 9.752416870782156e-07, 'samples': 24909312, 'steps': 48650, 'loss/train': 0.03904295712709427} 02/26/2022 16:35:29 - INFO - codeparrot_training - Step 48651: {'lr': 9.73798358039485e-07, 'samples': 24909824, 'steps': 48651, 'loss/train': 1.9457398653030396} 02/26/2022 16:35:33 - INFO - codeparrot_training - Step 48652: {'lr': 9.723560957499156e-07, 'samples': 24910336, 'steps': 48652, 'loss/train': 1.3193747997283936} 02/26/2022 16:35:39 - INFO - codeparrot_training - Step 48653: {'lr': 9.709149002156969e-07, 'samples': 24910848, 'steps': 48653, 'loss/train': 2.0491995811462402} 02/26/2022 16:35:43 - INFO - codeparrot_training - Step 48654: {'lr': 9.694747714430186e-07, 'samples': 24911360, 'steps': 48654, 'loss/train': 1.517682671546936} 02/26/2022 16:35:48 - INFO - codeparrot_training - Step 48655: {'lr': 9.680357094380144e-07, 'samples': 24911872, 'steps': 48655, 'loss/train': 2.345839500427246} 02/26/2022 16:35:51 - INFO - codeparrot_training - Step 48656: {'lr': 9.665977142068738e-07, 'samples': 24912384, 'steps': 48656, 'loss/train': 0.9726991057395935} 02/26/2022 16:35:57 - INFO - codeparrot_training - Step 48657: {'lr': 9.651607857557586e-07, 'samples': 24912896, 'steps': 48657, 'loss/train': 0.8982393145561218} 02/26/2022 16:36:01 - INFO - codeparrot_training - Step 48658: {'lr': 9.637249240908307e-07, 'samples': 24913408, 'steps': 48658, 'loss/train': 1.6993069648742676} 02/26/2022 16:36:06 - INFO - codeparrot_training - Step 48659: {'lr': 9.622901292181963e-07, 'samples': 24913920, 'steps': 48659, 'loss/train': 1.3577886819839478} 02/26/2022 16:36:09 - INFO - codeparrot_training - Step 48660: {'lr': 9.608564011440724e-07, 'samples': 24914432, 'steps': 48660, 'loss/train': 1.2054692506790161} 02/26/2022 16:36:15 - INFO - codeparrot_training - Step 48661: {'lr': 9.594237398745375e-07, 'samples': 24914944, 'steps': 48661, 'loss/train': 2.3172712326049805} 02/26/2022 16:36:18 - INFO - codeparrot_training - Step 48662: {'lr': 9.579921454157536e-07, 'samples': 24915456, 'steps': 48662, 'loss/train': 0.8853294253349304} 02/26/2022 16:36:25 - INFO - codeparrot_training - Step 48663: {'lr': 9.565616177738545e-07, 'samples': 24915968, 'steps': 48663, 'loss/train': 2.126934289932251} 02/26/2022 16:36:28 - INFO - codeparrot_training - Step 48664: {'lr': 9.551321569549743e-07, 'samples': 24916480, 'steps': 48664, 'loss/train': 1.4586352109909058} 02/26/2022 16:36:34 - INFO - codeparrot_training - Step 48665: {'lr': 9.537037629651912e-07, 'samples': 24916992, 'steps': 48665, 'loss/train': 0.7465572953224182} 02/26/2022 16:36:37 - INFO - codeparrot_training - Step 48666: {'lr': 9.52276435810695e-07, 'samples': 24917504, 'steps': 48666, 'loss/train': 2.002891778945923} 02/26/2022 16:36:43 - INFO - codeparrot_training - Step 48667: {'lr': 9.508501754975363e-07, 'samples': 24918016, 'steps': 48667, 'loss/train': 1.3721352815628052} 02/26/2022 16:36:47 - INFO - codeparrot_training - Step 48668: {'lr': 9.494249820318768e-07, 'samples': 24918528, 'steps': 48668, 'loss/train': 2.2423346042633057} 02/26/2022 16:36:50 - INFO - codeparrot_training - Step 48669: {'lr': 9.480008554197949e-07, 'samples': 24919040, 'steps': 48669, 'loss/train': 2.2046985626220703} 02/26/2022 16:36:56 - INFO - codeparrot_training - Step 48670: {'lr': 9.46577795667397e-07, 'samples': 24919552, 'steps': 48670, 'loss/train': 1.5289809703826904} 02/26/2022 16:36:59 - INFO - codeparrot_training - Step 48671: {'lr': 9.451558027807617e-07, 'samples': 24920064, 'steps': 48671, 'loss/train': 1.7409809827804565} 02/26/2022 16:37:05 - INFO - codeparrot_training - Step 48672: {'lr': 9.437348767659948e-07, 'samples': 24920576, 'steps': 48672, 'loss/train': 1.8067082166671753} 02/26/2022 16:37:08 - INFO - codeparrot_training - Step 48673: {'lr': 9.423150176292028e-07, 'samples': 24921088, 'steps': 48673, 'loss/train': 1.1441411972045898} 02/26/2022 16:37:15 - INFO - codeparrot_training - Step 48674: {'lr': 9.408962253764364e-07, 'samples': 24921600, 'steps': 48674, 'loss/train': 0.058101993054151535} 02/26/2022 16:37:19 - INFO - codeparrot_training - Step 48675: {'lr': 9.394785000138018e-07, 'samples': 24922112, 'steps': 48675, 'loss/train': 1.3000520467758179} 02/26/2022 16:37:24 - INFO - codeparrot_training - Step 48676: {'lr': 9.380618415473219e-07, 'samples': 24922624, 'steps': 48676, 'loss/train': 2.370760440826416} 02/26/2022 16:37:28 - INFO - codeparrot_training - Step 48677: {'lr': 9.366462499831307e-07, 'samples': 24923136, 'steps': 48677, 'loss/train': 1.69815194606781} 02/26/2022 16:37:33 - INFO - codeparrot_training - Step 48678: {'lr': 9.352317253272513e-07, 'samples': 24923648, 'steps': 48678, 'loss/train': 1.0883640050888062} 02/26/2022 16:37:37 - INFO - codeparrot_training - Step 48679: {'lr': 9.338182675857343e-07, 'samples': 24924160, 'steps': 48679, 'loss/train': 1.762582540512085} 02/26/2022 16:37:42 - INFO - codeparrot_training - Step 48680: {'lr': 9.324058767646859e-07, 'samples': 24924672, 'steps': 48680, 'loss/train': 1.7224711179733276} 02/26/2022 16:37:46 - INFO - codeparrot_training - Step 48681: {'lr': 9.309945528701014e-07, 'samples': 24925184, 'steps': 48681, 'loss/train': 0.5291641354560852} 02/26/2022 16:37:51 - INFO - codeparrot_training - Step 48682: {'lr': 9.295842959080592e-07, 'samples': 24925696, 'steps': 48682, 'loss/train': 1.3491417169570923} 02/26/2022 16:37:55 - INFO - codeparrot_training - Step 48683: {'lr': 9.281751058845822e-07, 'samples': 24926208, 'steps': 48683, 'loss/train': 2.165400981903076} 02/26/2022 16:38:00 - INFO - codeparrot_training - Step 48684: {'lr': 9.267669828056934e-07, 'samples': 24926720, 'steps': 48684, 'loss/train': 1.8449147939682007} 02/26/2022 16:38:04 - INFO - codeparrot_training - Step 48685: {'lr': 9.253599266774715e-07, 'samples': 24927232, 'steps': 48685, 'loss/train': 1.1529393196105957} 02/26/2022 16:38:09 - INFO - codeparrot_training - Step 48686: {'lr': 9.239539375059113e-07, 'samples': 24927744, 'steps': 48686, 'loss/train': 3.8546957969665527} 02/26/2022 16:38:12 - INFO - codeparrot_training - Step 48687: {'lr': 9.225490152970361e-07, 'samples': 24928256, 'steps': 48687, 'loss/train': 1.87710702419281} 02/26/2022 16:38:18 - INFO - codeparrot_training - Step 48688: {'lr': 9.211451600568966e-07, 'samples': 24928768, 'steps': 48688, 'loss/train': 1.5914483070373535} 02/26/2022 16:38:21 - INFO - codeparrot_training - Step 48689: {'lr': 9.197423717914321e-07, 'samples': 24929280, 'steps': 48689, 'loss/train': 2.135009765625} 02/26/2022 16:38:28 - INFO - codeparrot_training - Step 48690: {'lr': 9.183406505067494e-07, 'samples': 24929792, 'steps': 48690, 'loss/train': 1.0253850221633911} 02/26/2022 16:38:31 - INFO - codeparrot_training - Step 48691: {'lr': 9.169399962087599e-07, 'samples': 24930304, 'steps': 48691, 'loss/train': 0.8639960885047913} 02/26/2022 16:38:37 - INFO - codeparrot_training - Step 48692: {'lr': 9.155404089035424e-07, 'samples': 24930816, 'steps': 48692, 'loss/train': 1.39597749710083} 02/26/2022 16:38:41 - INFO - codeparrot_training - Step 48693: {'lr': 9.141418885970365e-07, 'samples': 24931328, 'steps': 48693, 'loss/train': 1.156754732131958} 02/26/2022 16:38:46 - INFO - codeparrot_training - Step 48694: {'lr': 9.127444352952375e-07, 'samples': 24931840, 'steps': 48694, 'loss/train': 0.9060541987419128} 02/26/2022 16:38:49 - INFO - codeparrot_training - Step 48695: {'lr': 9.113480490041681e-07, 'samples': 24932352, 'steps': 48695, 'loss/train': 1.4301373958587646} 02/26/2022 16:38:55 - INFO - codeparrot_training - Step 48696: {'lr': 9.099527297297961e-07, 'samples': 24932864, 'steps': 48696, 'loss/train': 3.0084667205810547} 02/26/2022 16:38:58 - INFO - codeparrot_training - Step 48697: {'lr': 9.085584774780886e-07, 'samples': 24933376, 'steps': 48697, 'loss/train': 0.3221356272697449} 02/26/2022 16:39:04 - INFO - codeparrot_training - Step 48698: {'lr': 9.071652922550133e-07, 'samples': 24933888, 'steps': 48698, 'loss/train': 0.17614074051380157} 02/26/2022 16:39:07 - INFO - codeparrot_training - Step 48699: {'lr': 9.057731740665376e-07, 'samples': 24934400, 'steps': 48699, 'loss/train': 1.8399994373321533} 02/26/2022 16:39:14 - INFO - codeparrot_training - Step 48700: {'lr': 9.043821229186566e-07, 'samples': 24934912, 'steps': 48700, 'loss/train': 0.6145563125610352} 02/26/2022 16:39:17 - INFO - codeparrot_training - Step 48701: {'lr': 9.029921388172823e-07, 'samples': 24935424, 'steps': 48701, 'loss/train': 1.7535086870193481} 02/26/2022 16:39:23 - INFO - codeparrot_training - Step 48702: {'lr': 9.0160322176841e-07, 'samples': 24935936, 'steps': 48702, 'loss/train': 1.829583764076233} 02/26/2022 16:39:26 - INFO - codeparrot_training - Step 48703: {'lr': 9.002153717779515e-07, 'samples': 24936448, 'steps': 48703, 'loss/train': 1.8609371185302734} 02/26/2022 16:39:32 - INFO - codeparrot_training - Step 48704: {'lr': 8.988285888519021e-07, 'samples': 24936960, 'steps': 48704, 'loss/train': 1.3831982612609863} 02/26/2022 16:39:35 - INFO - codeparrot_training - Step 48705: {'lr': 8.97442872996146e-07, 'samples': 24937472, 'steps': 48705, 'loss/train': 2.1711933612823486} 02/26/2022 16:39:41 - INFO - codeparrot_training - Step 48706: {'lr': 8.960582242166504e-07, 'samples': 24937984, 'steps': 48706, 'loss/train': 2.2444212436676025} 02/26/2022 16:39:44 - INFO - codeparrot_training - Step 48707: {'lr': 8.946746425193553e-07, 'samples': 24938496, 'steps': 48707, 'loss/train': 1.9515472650527954} 02/26/2022 16:39:50 - INFO - codeparrot_training - Step 48708: {'lr': 8.932921279101725e-07, 'samples': 24939008, 'steps': 48708, 'loss/train': 0.18177679181098938} 02/26/2022 16:39:53 - INFO - codeparrot_training - Step 48709: {'lr': 8.919106803950139e-07, 'samples': 24939520, 'steps': 48709, 'loss/train': 1.665985345840454} 02/26/2022 16:39:59 - INFO - codeparrot_training - Step 48710: {'lr': 8.905302999798192e-07, 'samples': 24940032, 'steps': 48710, 'loss/train': 0.06281256675720215} 02/26/2022 16:40:03 - INFO - codeparrot_training - Step 48711: {'lr': 8.891509866705006e-07, 'samples': 24940544, 'steps': 48711, 'loss/train': 1.3541584014892578} 02/26/2022 16:40:08 - INFO - codeparrot_training - Step 48712: {'lr': 8.877727404729419e-07, 'samples': 24941056, 'steps': 48712, 'loss/train': 2.0485923290252686} 02/26/2022 16:40:12 - INFO - codeparrot_training - Step 48713: {'lr': 8.863955613930553e-07, 'samples': 24941568, 'steps': 48713, 'loss/train': 2.7983922958374023} 02/26/2022 16:40:17 - INFO - codeparrot_training - Step 48714: {'lr': 8.850194494367525e-07, 'samples': 24942080, 'steps': 48714, 'loss/train': 0.6717244386672974} 02/26/2022 16:40:21 - INFO - codeparrot_training - Step 48715: {'lr': 8.836444046099457e-07, 'samples': 24942592, 'steps': 48715, 'loss/train': 0.8583987355232239} 02/26/2022 16:40:26 - INFO - codeparrot_training - Step 48716: {'lr': 8.822704269184633e-07, 'samples': 24943104, 'steps': 48716, 'loss/train': 2.296792507171631} 02/26/2022 16:40:30 - INFO - codeparrot_training - Step 48717: {'lr': 8.808975163682731e-07, 'samples': 24943616, 'steps': 48717, 'loss/train': 1.149545431137085} 02/26/2022 16:40:35 - INFO - codeparrot_training - Step 48718: {'lr': 8.795256729651757e-07, 'samples': 24944128, 'steps': 48718, 'loss/train': 1.2793878316879272} 02/26/2022 16:40:39 - INFO - codeparrot_training - Step 48719: {'lr': 8.781548967151109e-07, 'samples': 24944640, 'steps': 48719, 'loss/train': 1.8845187425613403} 02/26/2022 16:40:45 - INFO - codeparrot_training - Step 48720: {'lr': 8.767851876239075e-07, 'samples': 24945152, 'steps': 48720, 'loss/train': 1.430665373802185} 02/26/2022 16:40:48 - INFO - codeparrot_training - Step 48721: {'lr': 8.754165456974772e-07, 'samples': 24945664, 'steps': 48721, 'loss/train': 1.5877289772033691} 02/26/2022 16:40:54 - INFO - codeparrot_training - Step 48722: {'lr': 8.740489709416211e-07, 'samples': 24946176, 'steps': 48722, 'loss/train': 1.6608823537826538} 02/26/2022 16:40:57 - INFO - codeparrot_training - Step 48723: {'lr': 8.726824633622788e-07, 'samples': 24946688, 'steps': 48723, 'loss/train': 1.69318425655365} 02/26/2022 16:41:03 - INFO - codeparrot_training - Step 48724: {'lr': 8.713170229652234e-07, 'samples': 24947200, 'steps': 48724, 'loss/train': 1.438439130783081} 02/26/2022 16:41:07 - INFO - codeparrot_training - Step 48725: {'lr': 8.699526497563392e-07, 'samples': 24947712, 'steps': 48725, 'loss/train': 1.9111677408218384} 02/26/2022 16:41:13 - INFO - codeparrot_training - Step 48726: {'lr': 8.685893437414827e-07, 'samples': 24948224, 'steps': 48726, 'loss/train': 2.1296932697296143} 02/26/2022 16:41:16 - INFO - codeparrot_training - Step 48727: {'lr': 8.672271049264824e-07, 'samples': 24948736, 'steps': 48727, 'loss/train': 1.0653467178344727} 02/26/2022 16:41:22 - INFO - codeparrot_training - Step 48728: {'lr': 8.658659333171392e-07, 'samples': 24949248, 'steps': 48728, 'loss/train': 1.6033753156661987} 02/26/2022 16:41:25 - INFO - codeparrot_training - Step 48729: {'lr': 8.645058289193652e-07, 'samples': 24949760, 'steps': 48729, 'loss/train': 0.7593342661857605} 02/26/2022 16:41:31 - INFO - codeparrot_training - Step 48730: {'lr': 8.631467917389058e-07, 'samples': 24950272, 'steps': 48730, 'loss/train': 1.277632713317871} 02/26/2022 16:41:34 - INFO - codeparrot_training - Step 48731: {'lr': 8.61788821781645e-07, 'samples': 24950784, 'steps': 48731, 'loss/train': 1.8686374425888062} 02/26/2022 16:41:40 - INFO - codeparrot_training - Step 48732: {'lr': 8.604319190533283e-07, 'samples': 24951296, 'steps': 48732, 'loss/train': 1.7932742834091187} 02/26/2022 16:41:43 - INFO - codeparrot_training - Step 48733: {'lr': 8.590760835598399e-07, 'samples': 24951808, 'steps': 48733, 'loss/train': 1.4419950246810913} 02/26/2022 16:41:49 - INFO - codeparrot_training - Step 48734: {'lr': 8.577213153069807e-07, 'samples': 24952320, 'steps': 48734, 'loss/train': 1.365775465965271} 02/26/2022 16:41:52 - INFO - codeparrot_training - Step 48735: {'lr': 8.563676143004962e-07, 'samples': 24952832, 'steps': 48735, 'loss/train': 0.5562288165092468} 02/26/2022 16:41:59 - INFO - codeparrot_training - Step 48736: {'lr': 8.550149805462148e-07, 'samples': 24953344, 'steps': 48736, 'loss/train': 2.045912504196167} 02/26/2022 16:42:02 - INFO - codeparrot_training - Step 48737: {'lr': 8.536634140499656e-07, 'samples': 24953856, 'steps': 48737, 'loss/train': 1.2927608489990234} 02/26/2022 16:42:08 - INFO - codeparrot_training - Step 48738: {'lr': 8.523129148174935e-07, 'samples': 24954368, 'steps': 48738, 'loss/train': 0.04306618869304657} 02/26/2022 16:42:11 - INFO - codeparrot_training - Step 48739: {'lr': 8.509634828545998e-07, 'samples': 24954880, 'steps': 48739, 'loss/train': 2.117922306060791} 02/26/2022 16:42:17 - INFO - codeparrot_training - Step 48740: {'lr': 8.496151181670853e-07, 'samples': 24955392, 'steps': 48740, 'loss/train': 1.6562367677688599} 02/26/2022 16:42:20 - INFO - codeparrot_training - Step 48741: {'lr': 8.482678207606675e-07, 'samples': 24955904, 'steps': 48741, 'loss/train': 0.9660260081291199} 02/26/2022 16:42:26 - INFO - codeparrot_training - Step 48742: {'lr': 8.469215906411753e-07, 'samples': 24956416, 'steps': 48742, 'loss/train': 1.2809741497039795} 02/26/2022 16:42:29 - INFO - codeparrot_training - Step 48743: {'lr': 8.45576427814354e-07, 'samples': 24956928, 'steps': 48743, 'loss/train': 1.1946102380752563} 02/26/2022 16:42:35 - INFO - codeparrot_training - Step 48744: {'lr': 8.442323322859768e-07, 'samples': 24957440, 'steps': 48744, 'loss/train': 2.6889636516571045} 02/26/2022 16:42:38 - INFO - codeparrot_training - Step 48745: {'lr': 8.428893040617891e-07, 'samples': 24957952, 'steps': 48745, 'loss/train': 0.8061151504516602} 02/26/2022 16:42:44 - INFO - codeparrot_training - Step 48746: {'lr': 8.415473431475362e-07, 'samples': 24958464, 'steps': 48746, 'loss/train': 0.03204713016748428} 02/26/2022 16:42:48 - INFO - codeparrot_training - Step 48747: {'lr': 8.402064495489914e-07, 'samples': 24958976, 'steps': 48747, 'loss/train': 2.025420904159546} 02/26/2022 16:42:53 - INFO - codeparrot_training - Step 48748: {'lr': 8.388666232718722e-07, 'samples': 24959488, 'steps': 48748, 'loss/train': 2.071467638015747} 02/26/2022 16:42:57 - INFO - codeparrot_training - Step 48749: {'lr': 8.375278643219519e-07, 'samples': 24960000, 'steps': 48749, 'loss/train': 1.2462078332901} 02/26/2022 16:43:02 - INFO - codeparrot_training - Step 48750: {'lr': 8.361901727049204e-07, 'samples': 24960512, 'steps': 48750, 'loss/train': 1.4476274251937866} 02/26/2022 16:43:06 - INFO - codeparrot_training - Step 48751: {'lr': 8.348535484265508e-07, 'samples': 24961024, 'steps': 48751, 'loss/train': 2.1303017139434814} 02/26/2022 16:43:11 - INFO - codeparrot_training - Step 48752: {'lr': 8.335179914925328e-07, 'samples': 24961536, 'steps': 48752, 'loss/train': 2.0322623252868652} 02/26/2022 16:43:15 - INFO - codeparrot_training - Step 48753: {'lr': 8.321835019086122e-07, 'samples': 24962048, 'steps': 48753, 'loss/train': 2.335151195526123} 02/26/2022 16:43:20 - INFO - codeparrot_training - Step 48754: {'lr': 8.308500796805063e-07, 'samples': 24962560, 'steps': 48754, 'loss/train': 1.8535524606704712} 02/26/2022 16:43:24 - INFO - codeparrot_training - Step 48755: {'lr': 8.295177248139052e-07, 'samples': 24963072, 'steps': 48755, 'loss/train': 2.127371072769165} 02/26/2022 16:43:30 - INFO - codeparrot_training - Step 48756: {'lr': 8.281864373145265e-07, 'samples': 24963584, 'steps': 48756, 'loss/train': 2.833411455154419} 02/26/2022 16:43:33 - INFO - codeparrot_training - Step 48757: {'lr': 8.268562171880878e-07, 'samples': 24964096, 'steps': 48757, 'loss/train': 1.9080238342285156} 02/26/2022 16:43:39 - INFO - codeparrot_training - Step 48758: {'lr': 8.25527064440279e-07, 'samples': 24964608, 'steps': 48758, 'loss/train': 1.6851392984390259} 02/26/2022 16:43:42 - INFO - codeparrot_training - Step 48759: {'lr': 8.241989790767623e-07, 'samples': 24965120, 'steps': 48759, 'loss/train': 2.0645151138305664} 02/26/2022 16:43:48 - INFO - codeparrot_training - Step 48760: {'lr': 8.228719611032831e-07, 'samples': 24965632, 'steps': 48760, 'loss/train': 1.9589732885360718} 02/26/2022 16:43:51 - INFO - codeparrot_training - Step 48761: {'lr': 8.215460105255035e-07, 'samples': 24966144, 'steps': 48761, 'loss/train': 2.447303056716919} 02/26/2022 16:43:57 - INFO - codeparrot_training - Step 48762: {'lr': 8.202211273490857e-07, 'samples': 24966656, 'steps': 48762, 'loss/train': 2.395120143890381} 02/26/2022 16:44:01 - INFO - codeparrot_training - Step 48763: {'lr': 8.188973115796916e-07, 'samples': 24967168, 'steps': 48763, 'loss/train': 1.0769151449203491} 02/26/2022 16:44:07 - INFO - codeparrot_training - Step 48764: {'lr': 8.175745632230669e-07, 'samples': 24967680, 'steps': 48764, 'loss/train': 1.6376525163650513} 02/26/2022 16:44:10 - INFO - codeparrot_training - Step 48765: {'lr': 8.162528822847904e-07, 'samples': 24968192, 'steps': 48765, 'loss/train': 1.5445570945739746} 02/26/2022 16:44:14 - INFO - codeparrot_training - Step 48766: {'lr': 8.149322687705796e-07, 'samples': 24968704, 'steps': 48766, 'loss/train': 1.4840246438980103} 02/26/2022 16:44:19 - INFO - codeparrot_training - Step 48767: {'lr': 8.13612722686069e-07, 'samples': 24969216, 'steps': 48767, 'loss/train': 0.3326415717601776} 02/26/2022 16:44:23 - INFO - codeparrot_training - Step 48768: {'lr': 8.122942440369208e-07, 'samples': 24969728, 'steps': 48768, 'loss/train': 1.6522136926651} 02/26/2022 16:44:28 - INFO - codeparrot_training - Step 48769: {'lr': 8.109768328287692e-07, 'samples': 24970240, 'steps': 48769, 'loss/train': 1.6850029230117798} 02/26/2022 16:44:32 - INFO - codeparrot_training - Step 48770: {'lr': 8.096604890672765e-07, 'samples': 24970752, 'steps': 48770, 'loss/train': 2.3310251235961914} 02/26/2022 16:44:37 - INFO - codeparrot_training - Step 48771: {'lr': 8.083452127580493e-07, 'samples': 24971264, 'steps': 48771, 'loss/train': 0.7881247997283936} 02/26/2022 16:44:41 - INFO - codeparrot_training - Step 48772: {'lr': 8.070310039067774e-07, 'samples': 24971776, 'steps': 48772, 'loss/train': 1.2660707235336304} 02/26/2022 16:44:47 - INFO - codeparrot_training - Step 48773: {'lr': 8.057178625190397e-07, 'samples': 24972288, 'steps': 48773, 'loss/train': 2.4484686851501465} 02/26/2022 16:44:50 - INFO - codeparrot_training - Step 48774: {'lr': 8.044057886004708e-07, 'samples': 24972800, 'steps': 48774, 'loss/train': 1.5282375812530518} 02/26/2022 16:44:56 - INFO - codeparrot_training - Step 48775: {'lr': 8.030947821567047e-07, 'samples': 24973312, 'steps': 48775, 'loss/train': 2.041247844696045} 02/26/2022 16:45:00 - INFO - codeparrot_training - Step 48776: {'lr': 8.017848431933484e-07, 'samples': 24973824, 'steps': 48776, 'loss/train': 1.362694263458252} 02/26/2022 16:45:05 - INFO - codeparrot_training - Step 48777: {'lr': 8.004759717160359e-07, 'samples': 24974336, 'steps': 48777, 'loss/train': 1.7827330827713013} 02/26/2022 16:45:08 - INFO - codeparrot_training - Step 48778: {'lr': 7.991681677303186e-07, 'samples': 24974848, 'steps': 48778, 'loss/train': 1.5937379598617554} 02/26/2022 16:45:14 - INFO - codeparrot_training - Step 48779: {'lr': 7.978614312418586e-07, 'samples': 24975360, 'steps': 48779, 'loss/train': 1.102156639099121} 02/26/2022 16:45:17 - INFO - codeparrot_training - Step 48780: {'lr': 7.965557622562347e-07, 'samples': 24975872, 'steps': 48780, 'loss/train': 1.416162133216858} 02/26/2022 16:45:23 - INFO - codeparrot_training - Step 48781: {'lr': 7.952511607790258e-07, 'samples': 24976384, 'steps': 48781, 'loss/train': 1.525479793548584} 02/26/2022 16:45:26 - INFO - codeparrot_training - Step 48782: {'lr': 7.939476268158385e-07, 'samples': 24976896, 'steps': 48782, 'loss/train': 1.8329294919967651} 02/26/2022 16:45:33 - INFO - codeparrot_training - Step 48783: {'lr': 7.92645160372224e-07, 'samples': 24977408, 'steps': 48783, 'loss/train': 1.46579110622406} 02/26/2022 16:45:36 - INFO - codeparrot_training - Step 48784: {'lr': 7.913437614538166e-07, 'samples': 24977920, 'steps': 48784, 'loss/train': 1.6328544616699219} 02/26/2022 16:45:42 - INFO - codeparrot_training - Step 48785: {'lr': 7.900434300661397e-07, 'samples': 24978432, 'steps': 48785, 'loss/train': 1.7565546035766602} 02/26/2022 16:45:45 - INFO - codeparrot_training - Step 48786: {'lr': 7.887441662147998e-07, 'samples': 24978944, 'steps': 48786, 'loss/train': 1.3826297521591187} 02/26/2022 16:45:51 - INFO - codeparrot_training - Step 48787: {'lr': 7.874459699053204e-07, 'samples': 24979456, 'steps': 48787, 'loss/train': 0.5662800669670105} 02/26/2022 16:45:54 - INFO - codeparrot_training - Step 48788: {'lr': 7.861488411433082e-07, 'samples': 24979968, 'steps': 48788, 'loss/train': 1.8822400569915771} 02/26/2022 16:46:00 - INFO - codeparrot_training - Step 48789: {'lr': 7.848527799343141e-07, 'samples': 24980480, 'steps': 48789, 'loss/train': 1.7091679573059082} 02/26/2022 16:46:03 - INFO - codeparrot_training - Step 48790: {'lr': 7.835577862838339e-07, 'samples': 24980992, 'steps': 48790, 'loss/train': 1.2195299863815308} 02/26/2022 16:46:09 - INFO - codeparrot_training - Step 48791: {'lr': 7.822638601975018e-07, 'samples': 24981504, 'steps': 48791, 'loss/train': 1.2716575860977173} 02/26/2022 16:46:12 - INFO - codeparrot_training - Step 48792: {'lr': 7.809710016807859e-07, 'samples': 24982016, 'steps': 48792, 'loss/train': 2.415796995162964} 02/26/2022 16:46:18 - INFO - codeparrot_training - Step 48793: {'lr': 7.796792107392647e-07, 'samples': 24982528, 'steps': 48793, 'loss/train': 2.1237080097198486} 02/26/2022 16:46:22 - INFO - codeparrot_training - Step 48794: {'lr': 7.783884873784619e-07, 'samples': 24983040, 'steps': 48794, 'loss/train': 2.117501974105835} 02/26/2022 16:46:27 - INFO - codeparrot_training - Step 48795: {'lr': 7.770988316039008e-07, 'samples': 24983552, 'steps': 48795, 'loss/train': 2.6740622520446777} 02/26/2022 16:46:31 - INFO - codeparrot_training - Step 48796: {'lr': 7.758102434211045e-07, 'samples': 24984064, 'steps': 48796, 'loss/train': 3.1281399726867676} 02/26/2022 16:46:36 - INFO - codeparrot_training - Step 48797: {'lr': 7.745227228356244e-07, 'samples': 24984576, 'steps': 48797, 'loss/train': 1.4416662454605103} 02/26/2022 16:46:40 - INFO - codeparrot_training - Step 48798: {'lr': 7.732362698529283e-07, 'samples': 24985088, 'steps': 48798, 'loss/train': 1.584259033203125} 02/26/2022 16:46:45 - INFO - codeparrot_training - Step 48799: {'lr': 7.719508844785672e-07, 'samples': 24985600, 'steps': 48799, 'loss/train': 1.4350366592407227} 02/26/2022 16:46:49 - INFO - codeparrot_training - Step 48800: {'lr': 7.70666566718009e-07, 'samples': 24986112, 'steps': 48800, 'loss/train': 1.67714262008667} 02/26/2022 16:46:54 - INFO - codeparrot_training - Step 48801: {'lr': 7.693833165767772e-07, 'samples': 24986624, 'steps': 48801, 'loss/train': 1.9618850946426392} 02/26/2022 16:46:58 - INFO - codeparrot_training - Step 48802: {'lr': 7.681011340603949e-07, 'samples': 24987136, 'steps': 48802, 'loss/train': 0.9408183693885803} 02/26/2022 16:47:03 - INFO - codeparrot_training - Step 48803: {'lr': 7.668200191743024e-07, 'samples': 24987648, 'steps': 48803, 'loss/train': 1.1661893129348755} 02/26/2022 16:47:07 - INFO - codeparrot_training - Step 48804: {'lr': 7.65539971924023e-07, 'samples': 24988160, 'steps': 48804, 'loss/train': 1.5404794216156006} 02/26/2022 16:47:13 - INFO - codeparrot_training - Step 48805: {'lr': 7.642609923150246e-07, 'samples': 24988672, 'steps': 48805, 'loss/train': 1.5036308765411377} 02/26/2022 16:47:16 - INFO - codeparrot_training - Step 48806: {'lr': 7.629830803527749e-07, 'samples': 24989184, 'steps': 48806, 'loss/train': 1.780240535736084} 02/26/2022 16:47:22 - INFO - codeparrot_training - Step 48807: {'lr': 7.617062360427973e-07, 'samples': 24989696, 'steps': 48807, 'loss/train': 1.278171420097351} 02/26/2022 16:47:25 - INFO - codeparrot_training - Step 48808: {'lr': 7.604304593905042e-07, 'samples': 24990208, 'steps': 48808, 'loss/train': 1.5116682052612305} 02/26/2022 16:47:31 - INFO - codeparrot_training - Step 48809: {'lr': 7.591557504013913e-07, 'samples': 24990720, 'steps': 48809, 'loss/train': 1.6538307666778564} 02/26/2022 16:47:35 - INFO - codeparrot_training - Step 48810: {'lr': 7.578821090809263e-07, 'samples': 24991232, 'steps': 48810, 'loss/train': 1.1046043634414673} 02/26/2022 16:47:40 - INFO - codeparrot_training - Step 48811: {'lr': 7.566095354345215e-07, 'samples': 24991744, 'steps': 48811, 'loss/train': 0.9560911059379578} 02/26/2022 16:47:44 - INFO - codeparrot_training - Step 48812: {'lr': 7.553380294676726e-07, 'samples': 24992256, 'steps': 48812, 'loss/train': 1.5707350969314575} 02/26/2022 16:47:49 - INFO - codeparrot_training - Step 48813: {'lr': 7.540675911858197e-07, 'samples': 24992768, 'steps': 48813, 'loss/train': 1.8504838943481445} 02/26/2022 16:47:53 - INFO - codeparrot_training - Step 48814: {'lr': 7.527982205944028e-07, 'samples': 24993280, 'steps': 48814, 'loss/train': 1.5718964338302612} 02/26/2022 16:47:59 - INFO - codeparrot_training - Step 48815: {'lr': 7.515299176988344e-07, 'samples': 24993792, 'steps': 48815, 'loss/train': 1.8446481227874756} 02/26/2022 16:48:02 - INFO - codeparrot_training - Step 48816: {'lr': 7.502626825045544e-07, 'samples': 24994304, 'steps': 48816, 'loss/train': 2.2348520755767822} 02/26/2022 16:48:08 - INFO - codeparrot_training - Step 48817: {'lr': 7.489965150170586e-07, 'samples': 24994816, 'steps': 48817, 'loss/train': 1.857997179031372} 02/26/2022 16:48:11 - INFO - codeparrot_training - Step 48818: {'lr': 7.477314152416481e-07, 'samples': 24995328, 'steps': 48818, 'loss/train': 2.0076744556427} 02/26/2022 16:48:17 - INFO - codeparrot_training - Step 48819: {'lr': 7.464673831838464e-07, 'samples': 24995840, 'steps': 48819, 'loss/train': 1.8949207067489624} 02/26/2022 16:48:20 - INFO - codeparrot_training - Step 48820: {'lr': 7.452044188490381e-07, 'samples': 24996352, 'steps': 48820, 'loss/train': 2.1093645095825195} 02/26/2022 16:48:26 - INFO - codeparrot_training - Step 48821: {'lr': 7.439425222426077e-07, 'samples': 24996864, 'steps': 48821, 'loss/train': 1.644471526145935} 02/26/2022 16:48:29 - INFO - codeparrot_training - Step 48822: {'lr': 7.426816933699954e-07, 'samples': 24997376, 'steps': 48822, 'loss/train': 1.4608527421951294} 02/26/2022 16:48:35 - INFO - codeparrot_training - Step 48823: {'lr': 7.414219322365578e-07, 'samples': 24997888, 'steps': 48823, 'loss/train': 2.208585262298584} 02/26/2022 16:48:38 - INFO - codeparrot_training - Step 48824: {'lr': 7.401632388477631e-07, 'samples': 24998400, 'steps': 48824, 'loss/train': 2.027339220046997} 02/26/2022 16:48:44 - INFO - codeparrot_training - Step 48825: {'lr': 7.3890561320894e-07, 'samples': 24998912, 'steps': 48825, 'loss/train': 1.1359057426452637} 02/26/2022 16:48:47 - INFO - codeparrot_training - Step 48826: {'lr': 7.376490553254734e-07, 'samples': 24999424, 'steps': 48826, 'loss/train': 0.6023389101028442} 02/26/2022 16:48:53 - INFO - codeparrot_training - Step 48827: {'lr': 7.363935652028031e-07, 'samples': 24999936, 'steps': 48827, 'loss/train': 0.6397705078125} 02/26/2022 16:48:56 - INFO - codeparrot_training - Step 48828: {'lr': 7.351391428462584e-07, 'samples': 25000448, 'steps': 48828, 'loss/train': 0.9860379099845886} 02/26/2022 16:49:03 - INFO - codeparrot_training - Step 48829: {'lr': 7.338857882612238e-07, 'samples': 25000960, 'steps': 48829, 'loss/train': 0.6927613019943237} 02/26/2022 16:49:06 - INFO - codeparrot_training - Step 48830: {'lr': 7.326335014530561e-07, 'samples': 25001472, 'steps': 48830, 'loss/train': 1.9804426431655884} 02/26/2022 16:49:12 - INFO - codeparrot_training - Step 48831: {'lr': 7.313822824271676e-07, 'samples': 25001984, 'steps': 48831, 'loss/train': 1.6332658529281616} 02/26/2022 16:49:15 - INFO - codeparrot_training - Step 48832: {'lr': 7.301321311888876e-07, 'samples': 25002496, 'steps': 48832, 'loss/train': 1.8073056936264038} 02/26/2022 16:49:21 - INFO - codeparrot_training - Step 48833: {'lr': 7.288830477435448e-07, 'samples': 25003008, 'steps': 48833, 'loss/train': 1.422527551651001} 02/26/2022 16:49:24 - INFO - codeparrot_training - Step 48834: {'lr': 7.276350320965241e-07, 'samples': 25003520, 'steps': 48834, 'loss/train': 2.2832090854644775} 02/26/2022 16:49:30 - INFO - codeparrot_training - Step 48835: {'lr': 7.263880842531545e-07, 'samples': 25004032, 'steps': 48835, 'loss/train': 0.46577489376068115} 02/26/2022 16:49:33 - INFO - codeparrot_training - Step 48836: {'lr': 7.251422042187927e-07, 'samples': 25004544, 'steps': 48836, 'loss/train': 1.4138360023498535} 02/26/2022 16:49:39 - INFO - codeparrot_training - Step 48837: {'lr': 7.238973919987679e-07, 'samples': 25005056, 'steps': 48837, 'loss/train': 0.9904511570930481} 02/26/2022 16:49:42 - INFO - codeparrot_training - Step 48838: {'lr': 7.226536475984092e-07, 'samples': 25005568, 'steps': 48838, 'loss/train': 2.4837419986724854} 02/26/2022 16:49:49 - INFO - codeparrot_training - Step 48839: {'lr': 7.214109710230454e-07, 'samples': 25006080, 'steps': 48839, 'loss/train': 1.3925635814666748} 02/26/2022 16:49:52 - INFO - codeparrot_training - Step 48840: {'lr': 7.201693622780059e-07, 'samples': 25006592, 'steps': 48840, 'loss/train': 1.4889143705368042} 02/26/2022 16:49:58 - INFO - codeparrot_training - Step 48841: {'lr': 7.189288213686196e-07, 'samples': 25007104, 'steps': 48841, 'loss/train': 2.4012787342071533} 02/26/2022 16:50:01 - INFO - codeparrot_training - Step 48842: {'lr': 7.176893483001601e-07, 'samples': 25007616, 'steps': 48842, 'loss/train': 1.550831913948059} 02/26/2022 16:50:07 - INFO - codeparrot_training - Step 48843: {'lr': 7.164509430779842e-07, 'samples': 25008128, 'steps': 48843, 'loss/train': 0.19090385735034943} 02/26/2022 16:50:10 - INFO - codeparrot_training - Step 48844: {'lr': 7.152136057073655e-07, 'samples': 25008640, 'steps': 48844, 'loss/train': 2.9373362064361572} 02/26/2022 16:50:16 - INFO - codeparrot_training - Step 48845: {'lr': 7.139773361936053e-07, 'samples': 25009152, 'steps': 48845, 'loss/train': 2.0581512451171875} 02/26/2022 16:50:19 - INFO - codeparrot_training - Step 48846: {'lr': 7.127421345420326e-07, 'samples': 25009664, 'steps': 48846, 'loss/train': 1.744335412979126} 02/26/2022 16:50:25 - INFO - codeparrot_training - Step 48847: {'lr': 7.11508000757921e-07, 'samples': 25010176, 'steps': 48847, 'loss/train': 2.596433639526367} 02/26/2022 16:50:28 - INFO - codeparrot_training - Step 48848: {'lr': 7.102749348465165e-07, 'samples': 25010688, 'steps': 48848, 'loss/train': 1.710936427116394} 02/26/2022 16:50:34 - INFO - codeparrot_training - Step 48849: {'lr': 7.090429368131479e-07, 'samples': 25011200, 'steps': 48849, 'loss/train': 2.4011363983154297} 02/26/2022 16:50:37 - INFO - codeparrot_training - Step 48850: {'lr': 7.078120066630889e-07, 'samples': 25011712, 'steps': 48850, 'loss/train': 1.4815024137496948} 02/26/2022 16:50:44 - INFO - codeparrot_training - Step 48851: {'lr': 7.065821444016129e-07, 'samples': 25012224, 'steps': 48851, 'loss/train': 2.5250051021575928} 02/26/2022 16:50:47 - INFO - codeparrot_training - Step 48852: {'lr': 7.053533500339659e-07, 'samples': 25012736, 'steps': 48852, 'loss/train': 2.3422412872314453} 02/26/2022 16:50:52 - INFO - codeparrot_training - Step 48853: {'lr': 7.041256235654214e-07, 'samples': 25013248, 'steps': 48853, 'loss/train': 2.815161943435669} 02/26/2022 16:50:56 - INFO - codeparrot_training - Step 48854: {'lr': 7.028989650012529e-07, 'samples': 25013760, 'steps': 48854, 'loss/train': 1.9013689756393433} 02/26/2022 16:51:01 - INFO - codeparrot_training - Step 48855: {'lr': 7.016733743466786e-07, 'samples': 25014272, 'steps': 48855, 'loss/train': 1.5986576080322266} 02/26/2022 16:51:05 - INFO - codeparrot_training - Step 48856: {'lr': 7.004488516070273e-07, 'samples': 25014784, 'steps': 48856, 'loss/train': 2.037203788757324} 02/26/2022 16:51:10 - INFO - codeparrot_training - Step 48857: {'lr': 6.99225396787434e-07, 'samples': 25015296, 'steps': 48857, 'loss/train': 1.5300078392028809} 02/26/2022 16:51:14 - INFO - codeparrot_training - Step 48858: {'lr': 6.980030098932277e-07, 'samples': 25015808, 'steps': 48858, 'loss/train': 1.9404606819152832} 02/26/2022 16:51:20 - INFO - codeparrot_training - Step 48859: {'lr': 6.967816909295988e-07, 'samples': 25016320, 'steps': 48859, 'loss/train': 0.05252516269683838} 02/26/2022 16:51:23 - INFO - codeparrot_training - Step 48860: {'lr': 6.955614399018207e-07, 'samples': 25016832, 'steps': 48860, 'loss/train': 2.270653486251831} 02/26/2022 16:51:29 - INFO - codeparrot_training - Step 48861: {'lr': 6.943422568150836e-07, 'samples': 25017344, 'steps': 48861, 'loss/train': 1.713470697402954} 02/26/2022 16:51:33 - INFO - codeparrot_training - Step 48862: {'lr': 6.931241416746059e-07, 'samples': 25017856, 'steps': 48862, 'loss/train': 1.769823431968689} 02/26/2022 16:51:38 - INFO - codeparrot_training - Step 48863: {'lr': 6.919070944856332e-07, 'samples': 25018368, 'steps': 48863, 'loss/train': 0.2431567758321762} 02/26/2022 16:51:42 - INFO - codeparrot_training - Step 48864: {'lr': 6.906911152533557e-07, 'samples': 25018880, 'steps': 48864, 'loss/train': 1.2163974046707153} 02/26/2022 16:51:47 - INFO - codeparrot_training - Step 48865: {'lr': 6.894762039830193e-07, 'samples': 25019392, 'steps': 48865, 'loss/train': 1.2048203945159912} 02/26/2022 16:51:51 - INFO - codeparrot_training - Step 48866: {'lr': 6.882623606797867e-07, 'samples': 25019904, 'steps': 48866, 'loss/train': 2.3501455783843994} 02/26/2022 16:51:56 - INFO - codeparrot_training - Step 48867: {'lr': 6.870495853488756e-07, 'samples': 25020416, 'steps': 48867, 'loss/train': 1.6358106136322021} 02/26/2022 16:52:00 - INFO - codeparrot_training - Step 48868: {'lr': 6.858378779954488e-07, 'samples': 25020928, 'steps': 48868, 'loss/train': 1.8266350030899048} 02/26/2022 16:52:05 - INFO - codeparrot_training - Step 48869: {'lr': 6.84627238624752e-07, 'samples': 25021440, 'steps': 48869, 'loss/train': 1.988757610321045} 02/26/2022 16:52:09 - INFO - codeparrot_training - Step 48870: {'lr': 6.834176672419479e-07, 'samples': 25021952, 'steps': 48870, 'loss/train': 2.291236639022827} 02/26/2022 16:52:15 - INFO - codeparrot_training - Step 48871: {'lr': 6.822091638522265e-07, 'samples': 25022464, 'steps': 48871, 'loss/train': 1.7195982933044434} 02/26/2022 16:52:19 - INFO - codeparrot_training - Step 48872: {'lr': 6.810017284607229e-07, 'samples': 25022976, 'steps': 48872, 'loss/train': 2.53993558883667} 02/26/2022 16:52:24 - INFO - codeparrot_training - Step 48873: {'lr': 6.797953610726548e-07, 'samples': 25023488, 'steps': 48873, 'loss/train': 0.134123295545578} 02/26/2022 16:52:27 - INFO - codeparrot_training - Step 48874: {'lr': 6.785900616931851e-07, 'samples': 25024000, 'steps': 48874, 'loss/train': 1.010966420173645} 02/26/2022 16:52:33 - INFO - codeparrot_training - Step 48875: {'lr': 6.773858303274482e-07, 'samples': 25024512, 'steps': 48875, 'loss/train': 1.0960313081741333} 02/26/2022 16:52:36 - INFO - codeparrot_training - Step 48876: {'lr': 6.761826669806348e-07, 'samples': 25025024, 'steps': 48876, 'loss/train': 1.8134791851043701} 02/26/2022 16:52:42 - INFO - codeparrot_training - Step 48877: {'lr': 6.749805716578794e-07, 'samples': 25025536, 'steps': 48877, 'loss/train': 2.035311222076416} 02/26/2022 16:52:45 - INFO - codeparrot_training - Step 48878: {'lr': 6.737795443643446e-07, 'samples': 25026048, 'steps': 48878, 'loss/train': 2.643656015396118} 02/26/2022 16:52:51 - INFO - codeparrot_training - Step 48879: {'lr': 6.725795851051653e-07, 'samples': 25026560, 'steps': 48879, 'loss/train': 1.8357720375061035} 02/26/2022 16:52:54 - INFO - codeparrot_training - Step 48880: {'lr': 6.713806938854761e-07, 'samples': 25027072, 'steps': 48880, 'loss/train': 1.4059455394744873} 02/26/2022 16:53:00 - INFO - codeparrot_training - Step 48881: {'lr': 6.701828707104396e-07, 'samples': 25027584, 'steps': 48881, 'loss/train': 2.1597094535827637} 02/26/2022 16:53:03 - INFO - codeparrot_training - Step 48882: {'lr': 6.689861155851351e-07, 'samples': 25028096, 'steps': 48882, 'loss/train': 0.33912619948387146} 02/26/2022 16:53:09 - INFO - codeparrot_training - Step 48883: {'lr': 6.677904285147529e-07, 'samples': 25028608, 'steps': 48883, 'loss/train': 1.812813401222229} 02/26/2022 16:53:12 - INFO - codeparrot_training - Step 48884: {'lr': 6.665958095043723e-07, 'samples': 25029120, 'steps': 48884, 'loss/train': 1.9970989227294922} 02/26/2022 16:53:18 - INFO - codeparrot_training - Step 48885: {'lr': 6.65402258559128e-07, 'samples': 25029632, 'steps': 48885, 'loss/train': 1.2113577127456665} 02/26/2022 16:53:21 - INFO - codeparrot_training - Step 48886: {'lr': 6.64209775684127e-07, 'samples': 25030144, 'steps': 48886, 'loss/train': 1.9954582452774048} 02/26/2022 16:53:28 - INFO - codeparrot_training - Step 48887: {'lr': 6.630183608844764e-07, 'samples': 25030656, 'steps': 48887, 'loss/train': 2.0644354820251465} 02/26/2022 16:53:31 - INFO - codeparrot_training - Step 48888: {'lr': 6.618280141652833e-07, 'samples': 25031168, 'steps': 48888, 'loss/train': 1.4872416257858276} 02/26/2022 16:53:36 - INFO - codeparrot_training - Step 48889: {'lr': 6.606387355316545e-07, 'samples': 25031680, 'steps': 48889, 'loss/train': 1.973291277885437} 02/26/2022 16:53:40 - INFO - codeparrot_training - Step 48890: {'lr': 6.594505249886695e-07, 'samples': 25032192, 'steps': 48890, 'loss/train': 2.922168254852295} 02/26/2022 16:53:46 - INFO - codeparrot_training - Step 48891: {'lr': 6.582633825414075e-07, 'samples': 25032704, 'steps': 48891, 'loss/train': 0.972525417804718} 02/26/2022 16:53:49 - INFO - codeparrot_training - Step 48892: {'lr': 6.570773081950033e-07, 'samples': 25033216, 'steps': 48892, 'loss/train': 1.3254616260528564} 02/26/2022 16:53:55 - INFO - codeparrot_training - Step 48893: {'lr': 6.558923019544804e-07, 'samples': 25033728, 'steps': 48893, 'loss/train': 1.7696346044540405} 02/26/2022 16:53:58 - INFO - codeparrot_training - Step 48894: {'lr': 6.547083638249185e-07, 'samples': 25034240, 'steps': 48894, 'loss/train': 1.839529037475586} 02/26/2022 16:54:04 - INFO - codeparrot_training - Step 48895: {'lr': 6.535254938114521e-07, 'samples': 25034752, 'steps': 48895, 'loss/train': 2.6115753650665283} 02/26/2022 16:54:07 - INFO - codeparrot_training - Step 48896: {'lr': 6.523436919190773e-07, 'samples': 25035264, 'steps': 48896, 'loss/train': 1.5751363039016724} 02/26/2022 16:54:13 - INFO - codeparrot_training - Step 48897: {'lr': 6.511629581529011e-07, 'samples': 25035776, 'steps': 48897, 'loss/train': 1.649093747138977} 02/26/2022 16:54:17 - INFO - codeparrot_training - Step 48898: {'lr': 6.499832925179749e-07, 'samples': 25036288, 'steps': 48898, 'loss/train': 1.4541383981704712} 02/26/2022 16:54:22 - INFO - codeparrot_training - Step 48899: {'lr': 6.488046950193228e-07, 'samples': 25036800, 'steps': 48899, 'loss/train': 1.5391324758529663} 02/26/2022 16:54:26 - INFO - codeparrot_training - Step 48900: {'lr': 6.476271656620236e-07, 'samples': 25037312, 'steps': 48900, 'loss/train': 2.4783318042755127} 02/26/2022 16:54:31 - INFO - codeparrot_training - Step 48901: {'lr': 6.464507044511015e-07, 'samples': 25037824, 'steps': 48901, 'loss/train': 1.8874729871749878} 02/26/2022 16:54:35 - INFO - codeparrot_training - Step 48902: {'lr': 6.452753113916077e-07, 'samples': 25038336, 'steps': 48902, 'loss/train': 2.030632972717285} 02/26/2022 16:54:40 - INFO - codeparrot_training - Step 48903: {'lr': 6.441009864885938e-07, 'samples': 25038848, 'steps': 48903, 'loss/train': 1.3733640909194946} 02/26/2022 16:54:44 - INFO - codeparrot_training - Step 48904: {'lr': 6.42927729747056e-07, 'samples': 25039360, 'steps': 48904, 'loss/train': 0.734083354473114} 02/26/2022 16:54:49 - INFO - codeparrot_training - Step 48905: {'lr': 6.417555411720454e-07, 'samples': 25039872, 'steps': 48905, 'loss/train': 1.6773123741149902} 02/26/2022 16:54:53 - INFO - codeparrot_training - Step 48906: {'lr': 6.405844207685585e-07, 'samples': 25040384, 'steps': 48906, 'loss/train': 1.7317454814910889} 02/26/2022 16:54:59 - INFO - codeparrot_training - Step 48907: {'lr': 6.394143685416187e-07, 'samples': 25040896, 'steps': 48907, 'loss/train': 1.48530113697052} 02/26/2022 16:55:02 - INFO - codeparrot_training - Step 48908: {'lr': 6.382453844962776e-07, 'samples': 25041408, 'steps': 48908, 'loss/train': 1.9938817024230957} 02/26/2022 16:55:08 - INFO - codeparrot_training - Step 48909: {'lr': 6.370774686375036e-07, 'samples': 25041920, 'steps': 48909, 'loss/train': 1.591796636581421} 02/26/2022 16:55:11 - INFO - codeparrot_training - Step 48910: {'lr': 6.359106209702926e-07, 'samples': 25042432, 'steps': 48910, 'loss/train': 2.0834133625030518} 02/26/2022 16:55:17 - INFO - codeparrot_training - Step 48911: {'lr': 6.347448414996682e-07, 'samples': 25042944, 'steps': 48911, 'loss/train': 1.879281997680664} 02/26/2022 16:55:20 - INFO - codeparrot_training - Step 48912: {'lr': 6.335801302306265e-07, 'samples': 25043456, 'steps': 48912, 'loss/train': 1.8462363481521606} 02/26/2022 16:55:26 - INFO - codeparrot_training - Step 48913: {'lr': 6.324164871681359e-07, 'samples': 25043968, 'steps': 48913, 'loss/train': 1.70614492893219} 02/26/2022 16:55:29 - INFO - codeparrot_training - Step 48914: {'lr': 6.312539123171646e-07, 'samples': 25044480, 'steps': 48914, 'loss/train': 1.4532690048217773} 02/26/2022 16:55:35 - INFO - codeparrot_training - Step 48915: {'lr': 6.300924056827362e-07, 'samples': 25044992, 'steps': 48915, 'loss/train': 2.6289873123168945} 02/26/2022 16:55:38 - INFO - codeparrot_training - Step 48916: {'lr': 6.28931967269819e-07, 'samples': 25045504, 'steps': 48916, 'loss/train': 2.073599100112915} 02/26/2022 16:55:44 - INFO - codeparrot_training - Step 48917: {'lr': 6.277725970833815e-07, 'samples': 25046016, 'steps': 48917, 'loss/train': 1.506932258605957} 02/26/2022 16:55:48 - INFO - codeparrot_training - Step 48918: {'lr': 6.266142951283638e-07, 'samples': 25046528, 'steps': 48918, 'loss/train': 1.5356069803237915} 02/26/2022 16:55:53 - INFO - codeparrot_training - Step 48919: {'lr': 6.254570614097344e-07, 'samples': 25047040, 'steps': 48919, 'loss/train': 0.13488370180130005} 02/26/2022 16:55:57 - INFO - codeparrot_training - Step 48920: {'lr': 6.243008959324892e-07, 'samples': 25047552, 'steps': 48920, 'loss/train': 0.3730163276195526} 02/26/2022 16:56:02 - INFO - codeparrot_training - Step 48921: {'lr': 6.23145798701541e-07, 'samples': 25048064, 'steps': 48921, 'loss/train': 1.9878838062286377} 02/26/2022 16:56:06 - INFO - codeparrot_training - Step 48922: {'lr': 6.219917697218303e-07, 'samples': 25048576, 'steps': 48922, 'loss/train': 1.1903020143508911} 02/26/2022 16:56:11 - INFO - codeparrot_training - Step 48923: {'lr': 6.20838808998353e-07, 'samples': 25049088, 'steps': 48923, 'loss/train': 1.5491689443588257} 02/26/2022 16:56:15 - INFO - codeparrot_training - Step 48924: {'lr': 6.196869165359941e-07, 'samples': 25049600, 'steps': 48924, 'loss/train': 0.41483357548713684} 02/26/2022 16:56:20 - INFO - codeparrot_training - Step 48925: {'lr': 6.185360923396943e-07, 'samples': 25050112, 'steps': 48925, 'loss/train': 1.0842278003692627} 02/26/2022 16:56:24 - INFO - codeparrot_training - Step 48926: {'lr': 6.173863364144494e-07, 'samples': 25050624, 'steps': 48926, 'loss/train': 1.283677101135254} 02/26/2022 16:56:29 - INFO - codeparrot_training - Step 48927: {'lr': 6.162376487650889e-07, 'samples': 25051136, 'steps': 48927, 'loss/train': 2.1205806732177734} 02/26/2022 16:56:33 - INFO - codeparrot_training - Step 48928: {'lr': 6.150900293966089e-07, 'samples': 25051648, 'steps': 48928, 'loss/train': 2.7395167350769043} 02/26/2022 16:56:38 - INFO - codeparrot_training - Step 48929: {'lr': 6.139434783138664e-07, 'samples': 25052160, 'steps': 48929, 'loss/train': 2.0299055576324463} 02/26/2022 16:56:42 - INFO - codeparrot_training - Step 48930: {'lr': 6.127979955218299e-07, 'samples': 25052672, 'steps': 48930, 'loss/train': 0.6684316396713257} 02/26/2022 16:56:47 - INFO - codeparrot_training - Step 48931: {'lr': 6.116535810253565e-07, 'samples': 25053184, 'steps': 48931, 'loss/train': 0.9637073278427124} 02/26/2022 16:56:53 - INFO - codeparrot_training - Step 48932: {'lr': 6.10510234829359e-07, 'samples': 25053696, 'steps': 48932, 'loss/train': 1.3618954420089722} 02/26/2022 16:56:57 - INFO - codeparrot_training - Step 48933: {'lr': 6.0936795693875e-07, 'samples': 25054208, 'steps': 48933, 'loss/train': 1.6464715003967285} 02/26/2022 16:57:02 - INFO - codeparrot_training - Step 48934: {'lr': 6.082267473584424e-07, 'samples': 25054720, 'steps': 48934, 'loss/train': 2.2851829528808594} 02/26/2022 16:57:06 - INFO - codeparrot_training - Step 48935: {'lr': 6.070866060932934e-07, 'samples': 25055232, 'steps': 48935, 'loss/train': 1.779329538345337} 02/26/2022 16:57:11 - INFO - codeparrot_training - Step 48936: {'lr': 6.059475331481878e-07, 'samples': 25055744, 'steps': 48936, 'loss/train': 2.0829570293426514} 02/26/2022 16:57:15 - INFO - codeparrot_training - Step 48937: {'lr': 6.048095285280109e-07, 'samples': 25056256, 'steps': 48937, 'loss/train': 0.2626766860485077} 02/26/2022 16:57:20 - INFO - codeparrot_training - Step 48938: {'lr': 6.036725922376196e-07, 'samples': 25056768, 'steps': 48938, 'loss/train': 1.6045479774475098} 02/26/2022 16:57:24 - INFO - codeparrot_training - Step 48939: {'lr': 6.025367242819268e-07, 'samples': 25057280, 'steps': 48939, 'loss/train': 1.910038709640503} 02/26/2022 16:57:29 - INFO - codeparrot_training - Step 48940: {'lr': 6.014019246657898e-07, 'samples': 25057792, 'steps': 48940, 'loss/train': 1.8424007892608643} 02/26/2022 16:57:33 - INFO - codeparrot_training - Step 48941: {'lr': 6.002681933940102e-07, 'samples': 25058304, 'steps': 48941, 'loss/train': 0.9923708438873291} 02/26/2022 16:57:39 - INFO - codeparrot_training - Step 48942: {'lr': 5.991355304715285e-07, 'samples': 25058816, 'steps': 48942, 'loss/train': 3.176571846008301} 02/26/2022 16:57:42 - INFO - codeparrot_training - Step 48943: {'lr': 5.980039359031186e-07, 'samples': 25059328, 'steps': 48943, 'loss/train': 0.3473702073097229} 02/26/2022 16:57:48 - INFO - codeparrot_training - Step 48944: {'lr': 5.968734096936935e-07, 'samples': 25059840, 'steps': 48944, 'loss/train': 0.9416552186012268} 02/26/2022 16:57:51 - INFO - codeparrot_training - Step 48945: {'lr': 5.957439518480545e-07, 'samples': 25060352, 'steps': 48945, 'loss/train': 1.5790685415267944} 02/26/2022 16:57:57 - INFO - codeparrot_training - Step 48946: {'lr': 5.946155623710592e-07, 'samples': 25060864, 'steps': 48946, 'loss/train': 1.6193063259124756} 02/26/2022 16:58:00 - INFO - codeparrot_training - Step 48947: {'lr': 5.934882412675646e-07, 'samples': 25061376, 'steps': 48947, 'loss/train': 2.1959645748138428} 02/26/2022 16:58:06 - INFO - codeparrot_training - Step 48948: {'lr': 5.923619885423448e-07, 'samples': 25061888, 'steps': 48948, 'loss/train': 1.487581729888916} 02/26/2022 16:58:09 - INFO - codeparrot_training - Step 48949: {'lr': 5.912368042002292e-07, 'samples': 25062400, 'steps': 48949, 'loss/train': 1.0387383699417114} 02/26/2022 16:58:15 - INFO - codeparrot_training - Step 48950: {'lr': 5.90112688246075e-07, 'samples': 25062912, 'steps': 48950, 'loss/train': 1.5394786596298218} 02/26/2022 16:58:18 - INFO - codeparrot_training - Step 48951: {'lr': 5.889896406846839e-07, 'samples': 25063424, 'steps': 48951, 'loss/train': 2.4579970836639404} 02/26/2022 16:58:25 - INFO - codeparrot_training - Step 48952: {'lr': 5.878676615208856e-07, 'samples': 25063936, 'steps': 48952, 'loss/train': 1.8241870403289795} 02/26/2022 16:58:28 - INFO - codeparrot_training - Step 48953: {'lr': 5.867467507594259e-07, 'samples': 25064448, 'steps': 48953, 'loss/train': 0.7845878601074219} 02/26/2022 16:58:34 - INFO - codeparrot_training - Step 48954: {'lr': 5.856269084051624e-07, 'samples': 25064960, 'steps': 48954, 'loss/train': 0.5075451731681824} 02/26/2022 16:58:37 - INFO - codeparrot_training - Step 48955: {'lr': 5.845081344628967e-07, 'samples': 25065472, 'steps': 48955, 'loss/train': 3.1926262378692627} 02/26/2022 16:58:42 - INFO - codeparrot_training - Step 48956: {'lr': 5.833904289373748e-07, 'samples': 25065984, 'steps': 48956, 'loss/train': 2.351822853088379} 02/26/2022 16:58:46 - INFO - codeparrot_training - Step 48957: {'lr': 5.822737918333987e-07, 'samples': 25066496, 'steps': 48957, 'loss/train': 1.7050217390060425} 02/26/2022 16:58:52 - INFO - codeparrot_training - Step 48958: {'lr': 5.811582231557977e-07, 'samples': 25067008, 'steps': 48958, 'loss/train': 1.4442594051361084} 02/26/2022 16:58:55 - INFO - codeparrot_training - Step 48959: {'lr': 5.800437229092903e-07, 'samples': 25067520, 'steps': 48959, 'loss/train': 1.4676461219787598} 02/26/2022 16:59:01 - INFO - codeparrot_training - Step 48960: {'lr': 5.789302910986782e-07, 'samples': 25068032, 'steps': 48960, 'loss/train': 1.693637728691101} 02/26/2022 16:59:04 - INFO - codeparrot_training - Step 48961: {'lr': 5.778179277287354e-07, 'samples': 25068544, 'steps': 48961, 'loss/train': 3.154338836669922} 02/26/2022 16:59:10 - INFO - codeparrot_training - Step 48962: {'lr': 5.767066328042081e-07, 'samples': 25069056, 'steps': 48962, 'loss/train': 1.5748413801193237} 02/26/2022 16:59:13 - INFO - codeparrot_training - Step 48963: {'lr': 5.75596406329898e-07, 'samples': 25069568, 'steps': 48963, 'loss/train': 1.761905550956726} 02/26/2022 16:59:19 - INFO - codeparrot_training - Step 48964: {'lr': 5.744872483104957e-07, 'samples': 25070080, 'steps': 48964, 'loss/train': 1.4534127712249756} 02/26/2022 16:59:23 - INFO - codeparrot_training - Step 48965: {'lr': 5.73379158750803e-07, 'samples': 25070592, 'steps': 48965, 'loss/train': 1.1781994104385376} 02/26/2022 16:59:28 - INFO - codeparrot_training - Step 48966: {'lr': 5.722721376555662e-07, 'samples': 25071104, 'steps': 48966, 'loss/train': 1.6667585372924805} 02/26/2022 16:59:32 - INFO - codeparrot_training - Step 48967: {'lr': 5.711661850295036e-07, 'samples': 25071616, 'steps': 48967, 'loss/train': 1.5644289255142212} 02/26/2022 16:59:37 - INFO - codeparrot_training - Step 48968: {'lr': 5.700613008773336e-07, 'samples': 25072128, 'steps': 48968, 'loss/train': 2.4919207096099854} 02/26/2022 16:59:41 - INFO - codeparrot_training - Step 48969: {'lr': 5.689574852038582e-07, 'samples': 25072640, 'steps': 48969, 'loss/train': 1.1718039512634277} 02/26/2022 16:59:46 - INFO - codeparrot_training - Step 48970: {'lr': 5.6785473801374e-07, 'samples': 25073152, 'steps': 48970, 'loss/train': 1.782148838043213} 02/26/2022 16:59:50 - INFO - codeparrot_training - Step 48971: {'lr': 5.667530593117531e-07, 'samples': 25073664, 'steps': 48971, 'loss/train': 1.9602621793746948} 02/26/2022 16:59:55 - INFO - codeparrot_training - Step 48972: {'lr': 5.656524491025883e-07, 'samples': 25074176, 'steps': 48972, 'loss/train': 1.281722903251648} 02/26/2022 16:59:59 - INFO - codeparrot_training - Step 48973: {'lr': 5.645529073909361e-07, 'samples': 25074688, 'steps': 48973, 'loss/train': 2.163677453994751} 02/26/2022 17:00:04 - INFO - codeparrot_training - Step 48974: {'lr': 5.634544341815706e-07, 'samples': 25075200, 'steps': 48974, 'loss/train': 0.9127210974693298} 02/26/2022 17:00:08 - INFO - codeparrot_training - Step 48975: {'lr': 5.623570294791269e-07, 'samples': 25075712, 'steps': 48975, 'loss/train': 2.605700731277466} 02/26/2022 17:00:13 - INFO - codeparrot_training - Step 48976: {'lr': 5.612606932883513e-07, 'samples': 25076224, 'steps': 48976, 'loss/train': 1.877730131149292} 02/26/2022 17:00:17 - INFO - codeparrot_training - Step 48977: {'lr': 5.601654256139621e-07, 'samples': 25076736, 'steps': 48977, 'loss/train': 1.9526714086532593} 02/26/2022 17:00:23 - INFO - codeparrot_training - Step 48978: {'lr': 5.590712264605668e-07, 'samples': 25077248, 'steps': 48978, 'loss/train': 2.089413642883301} 02/26/2022 17:00:26 - INFO - codeparrot_training - Step 48979: {'lr': 5.579780958329394e-07, 'samples': 25077760, 'steps': 48979, 'loss/train': 1.099550724029541} 02/26/2022 17:00:32 - INFO - codeparrot_training - Step 48980: {'lr': 5.568860337357151e-07, 'samples': 25078272, 'steps': 48980, 'loss/train': 1.7878739833831787} 02/26/2022 17:00:36 - INFO - codeparrot_training - Step 48981: {'lr': 5.557950401735845e-07, 'samples': 25078784, 'steps': 48981, 'loss/train': 1.6897863149642944} 02/26/2022 17:00:41 - INFO - codeparrot_training - Step 48982: {'lr': 5.547051151512384e-07, 'samples': 25079296, 'steps': 48982, 'loss/train': 1.6523233652114868} 02/26/2022 17:00:45 - INFO - codeparrot_training - Step 48983: {'lr': 5.536162586733118e-07, 'samples': 25079808, 'steps': 48983, 'loss/train': 2.362736701965332} 02/26/2022 17:00:50 - INFO - codeparrot_training - Step 48984: {'lr': 5.525284707444678e-07, 'samples': 25080320, 'steps': 48984, 'loss/train': 0.5700541734695435} 02/26/2022 17:00:54 - INFO - codeparrot_training - Step 48985: {'lr': 5.51441751369397e-07, 'samples': 25080832, 'steps': 48985, 'loss/train': 1.792810320854187} 02/26/2022 17:00:59 - INFO - codeparrot_training - Step 48986: {'lr': 5.503561005527347e-07, 'samples': 25081344, 'steps': 48986, 'loss/train': 1.5757895708084106} 02/26/2022 17:01:03 - INFO - codeparrot_training - Step 48987: {'lr': 5.492715182991714e-07, 'samples': 25081856, 'steps': 48987, 'loss/train': 1.5620430707931519} 02/26/2022 17:01:08 - INFO - codeparrot_training - Step 48988: {'lr': 5.48188004613287e-07, 'samples': 25082368, 'steps': 48988, 'loss/train': 1.836152195930481} 02/26/2022 17:01:12 - INFO - codeparrot_training - Step 48989: {'lr': 5.471055594997443e-07, 'samples': 25082880, 'steps': 48989, 'loss/train': 0.47212541103363037} 02/26/2022 17:01:18 - INFO - codeparrot_training - Step 48990: {'lr': 5.460241829632063e-07, 'samples': 25083392, 'steps': 48990, 'loss/train': 1.2685925960540771} 02/26/2022 17:01:21 - INFO - codeparrot_training - Step 48991: {'lr': 5.44943875008308e-07, 'samples': 25083904, 'steps': 48991, 'loss/train': 1.3224447965621948} 02/26/2022 17:01:27 - INFO - codeparrot_training - Step 48992: {'lr': 5.438646356396293e-07, 'samples': 25084416, 'steps': 48992, 'loss/train': 1.1389464139938354} 02/26/2022 17:01:30 - INFO - codeparrot_training - Step 48993: {'lr': 5.427864648618608e-07, 'samples': 25084928, 'steps': 48993, 'loss/train': 1.6031832695007324} 02/26/2022 17:01:36 - INFO - codeparrot_training - Step 48994: {'lr': 5.417093626795544e-07, 'samples': 25085440, 'steps': 48994, 'loss/train': 2.1603238582611084} 02/26/2022 17:01:39 - INFO - codeparrot_training - Step 48995: {'lr': 5.406333290973453e-07, 'samples': 25085952, 'steps': 48995, 'loss/train': 1.7539637088775635} 02/26/2022 17:01:45 - INFO - codeparrot_training - Step 48996: {'lr': 5.395583641198687e-07, 'samples': 25086464, 'steps': 48996, 'loss/train': 2.132857322692871} 02/26/2022 17:01:48 - INFO - codeparrot_training - Step 48997: {'lr': 5.384844677517043e-07, 'samples': 25086976, 'steps': 48997, 'loss/train': 1.8419115543365479} 02/26/2022 17:01:54 - INFO - codeparrot_training - Step 48998: {'lr': 5.374116399974871e-07, 'samples': 25087488, 'steps': 48998, 'loss/train': 2.223346471786499} 02/26/2022 17:01:57 - INFO - codeparrot_training - Step 48999: {'lr': 5.363398808617414e-07, 'samples': 25088000, 'steps': 48999, 'loss/train': 1.8959709405899048} 02/26/2022 17:01:57 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 17:02:15 - WARNING - huggingface_hub.repository - Several commits (49) will be pushed upstream. 02/26/2022 17:02:15 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 17:02:50 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 8e26ae5..508ef45 floral-grass-11 -> floral-grass-11 02/26/2022 17:02:58 - INFO - codeparrot_training - Step 49000: {'lr': 5.352691903491303e-07, 'samples': 25088512, 'steps': 49000, 'loss/train': 1.3508716821670532} 02/26/2022 17:03:01 - INFO - codeparrot_training - Step 49001: {'lr': 5.341995684642053e-07, 'samples': 25089024, 'steps': 49001, 'loss/train': 0.7777153253555298} 02/26/2022 17:03:07 - INFO - codeparrot_training - Step 49002: {'lr': 5.331310152115465e-07, 'samples': 25089536, 'steps': 49002, 'loss/train': 1.7808958292007446} 02/26/2022 17:03:10 - INFO - codeparrot_training - Step 49003: {'lr': 5.320635305957333e-07, 'samples': 25090048, 'steps': 49003, 'loss/train': 2.8032636642456055} 02/26/2022 17:03:16 - INFO - codeparrot_training - Step 49004: {'lr': 5.309971146213732e-07, 'samples': 25090560, 'steps': 49004, 'loss/train': 1.5926603078842163} 02/26/2022 17:03:19 - INFO - codeparrot_training - Step 49005: {'lr': 5.299317672929904e-07, 'samples': 25091072, 'steps': 49005, 'loss/train': 1.5938676595687866} 02/26/2022 17:03:25 - INFO - codeparrot_training - Step 49006: {'lr': 5.288674886151645e-07, 'samples': 25091584, 'steps': 49006, 'loss/train': 1.1151988506317139} 02/26/2022 17:03:28 - INFO - codeparrot_training - Step 49007: {'lr': 5.278042785924475e-07, 'samples': 25092096, 'steps': 49007, 'loss/train': 1.8216667175292969} 02/26/2022 17:03:34 - INFO - codeparrot_training - Step 49008: {'lr': 5.26742137229419e-07, 'samples': 25092608, 'steps': 49008, 'loss/train': 2.0989344120025635} 02/26/2022 17:03:37 - INFO - codeparrot_training - Step 49009: {'lr': 5.256810645306032e-07, 'samples': 25093120, 'steps': 49009, 'loss/train': 0.9914582967758179} 02/26/2022 17:03:43 - INFO - codeparrot_training - Step 49010: {'lr': 5.246210605005519e-07, 'samples': 25093632, 'steps': 49010, 'loss/train': 1.4833296537399292} 02/26/2022 17:03:47 - INFO - codeparrot_training - Step 49011: {'lr': 5.235621251437894e-07, 'samples': 25094144, 'steps': 49011, 'loss/train': 0.8157903552055359} 02/26/2022 17:03:52 - INFO - codeparrot_training - Step 49012: {'lr': 5.225042584648954e-07, 'samples': 25094656, 'steps': 49012, 'loss/train': 2.2832565307617188} 02/26/2022 17:03:56 - INFO - codeparrot_training - Step 49013: {'lr': 5.214474604683384e-07, 'samples': 25095168, 'steps': 49013, 'loss/train': 2.02289080619812} 02/26/2022 17:04:02 - INFO - codeparrot_training - Step 49014: {'lr': 5.203917311587258e-07, 'samples': 25095680, 'steps': 49014, 'loss/train': 1.2984391450881958} 02/26/2022 17:04:05 - INFO - codeparrot_training - Step 49015: {'lr': 5.193370705404988e-07, 'samples': 25096192, 'steps': 49015, 'loss/train': 1.4447349309921265} 02/26/2022 17:04:11 - INFO - codeparrot_training - Step 49016: {'lr': 5.182834786182366e-07, 'samples': 25096704, 'steps': 49016, 'loss/train': 2.0010018348693848} 02/26/2022 17:04:14 - INFO - codeparrot_training - Step 49017: {'lr': 5.17230955396436e-07, 'samples': 25097216, 'steps': 49017, 'loss/train': 0.18538343906402588} 02/26/2022 17:04:20 - INFO - codeparrot_training - Step 49018: {'lr': 5.161795008795656e-07, 'samples': 25097728, 'steps': 49018, 'loss/train': 1.284847378730774} 02/26/2022 17:04:23 - INFO - codeparrot_training - Step 49019: {'lr': 5.151291150721771e-07, 'samples': 25098240, 'steps': 49019, 'loss/train': 1.2908810377120972} 02/26/2022 17:04:30 - INFO - codeparrot_training - Step 49020: {'lr': 5.140797979787671e-07, 'samples': 25098752, 'steps': 49020, 'loss/train': 8.608312606811523} 02/26/2022 17:04:33 - INFO - codeparrot_training - Step 49021: {'lr': 5.13031549603804e-07, 'samples': 25099264, 'steps': 49021, 'loss/train': 1.6041030883789062} 02/26/2022 17:04:39 - INFO - codeparrot_training - Step 49022: {'lr': 5.119843699518123e-07, 'samples': 25099776, 'steps': 49022, 'loss/train': 1.7628525495529175} 02/26/2022 17:04:42 - INFO - codeparrot_training - Step 49023: {'lr': 5.109382590272327e-07, 'samples': 25100288, 'steps': 49023, 'loss/train': 1.6956602334976196} 02/26/2022 17:04:48 - INFO - codeparrot_training - Step 49024: {'lr': 5.098932168345893e-07, 'samples': 25100800, 'steps': 49024, 'loss/train': 3.1024675369262695} 02/26/2022 17:04:51 - INFO - codeparrot_training - Step 49025: {'lr': 5.088492433783509e-07, 'samples': 25101312, 'steps': 49025, 'loss/train': 1.6666789054870605} 02/26/2022 17:04:57 - INFO - codeparrot_training - Step 49026: {'lr': 5.078063386629583e-07, 'samples': 25101824, 'steps': 49026, 'loss/train': 1.6158421039581299} 02/26/2022 17:05:00 - INFO - codeparrot_training - Step 49027: {'lr': 5.067645026929079e-07, 'samples': 25102336, 'steps': 49027, 'loss/train': 1.9067306518554688} 02/26/2022 17:05:06 - INFO - codeparrot_training - Step 49028: {'lr': 5.057237354726685e-07, 'samples': 25102848, 'steps': 49028, 'loss/train': 1.9076614379882812} 02/26/2022 17:05:09 - INFO - codeparrot_training - Step 49029: {'lr': 5.04684037006653e-07, 'samples': 25103360, 'steps': 49029, 'loss/train': 1.5937572717666626} 02/26/2022 17:05:15 - INFO - codeparrot_training - Step 49030: {'lr': 5.036454072993857e-07, 'samples': 25103872, 'steps': 49030, 'loss/train': 1.298364281654358} 02/26/2022 17:05:18 - INFO - codeparrot_training - Step 49031: {'lr': 5.026078463552519e-07, 'samples': 25104384, 'steps': 49031, 'loss/train': 1.7518583536148071} 02/26/2022 17:05:24 - INFO - codeparrot_training - Step 49032: {'lr': 5.015713541787204e-07, 'samples': 25104896, 'steps': 49032, 'loss/train': 2.3967034816741943} 02/26/2022 17:05:27 - INFO - codeparrot_training - Step 49033: {'lr': 5.005359307742319e-07, 'samples': 25105408, 'steps': 49033, 'loss/train': 1.2391053438186646} 02/26/2022 17:05:33 - INFO - codeparrot_training - Step 49034: {'lr': 4.995015761462551e-07, 'samples': 25105920, 'steps': 49034, 'loss/train': 1.8676375150680542} 02/26/2022 17:05:37 - INFO - codeparrot_training - Step 49035: {'lr': 4.984682902991477e-07, 'samples': 25106432, 'steps': 49035, 'loss/train': 0.9794620871543884} 02/26/2022 17:05:42 - INFO - codeparrot_training - Step 49036: {'lr': 4.974360732374062e-07, 'samples': 25106944, 'steps': 49036, 'loss/train': 0.07919516414403915} 02/26/2022 17:05:46 - INFO - codeparrot_training - Step 49037: {'lr': 4.964049249654156e-07, 'samples': 25107456, 'steps': 49037, 'loss/train': 3.1983420848846436} 02/26/2022 17:05:51 - INFO - codeparrot_training - Step 49038: {'lr': 4.953748454875895e-07, 'samples': 25107968, 'steps': 49038, 'loss/train': 1.0089688301086426} 02/26/2022 17:05:55 - INFO - codeparrot_training - Step 49039: {'lr': 4.943458348083407e-07, 'samples': 25108480, 'steps': 49039, 'loss/train': 1.2593165636062622} 02/26/2022 17:06:00 - INFO - codeparrot_training - Step 49040: {'lr': 4.933178929321103e-07, 'samples': 25108992, 'steps': 49040, 'loss/train': 0.506611704826355} 02/26/2022 17:06:04 - INFO - codeparrot_training - Step 49041: {'lr': 4.922910198632557e-07, 'samples': 25109504, 'steps': 49041, 'loss/train': 2.087669610977173} 02/26/2022 17:06:09 - INFO - codeparrot_training - Step 49042: {'lr': 4.912652156062181e-07, 'samples': 25110016, 'steps': 49042, 'loss/train': 1.889102578163147} 02/26/2022 17:06:13 - INFO - codeparrot_training - Step 49043: {'lr': 4.902404801653826e-07, 'samples': 25110528, 'steps': 49043, 'loss/train': 2.149672746658325} 02/26/2022 17:06:18 - INFO - codeparrot_training - Step 49044: {'lr': 4.892168135451069e-07, 'samples': 25111040, 'steps': 49044, 'loss/train': 0.6079110503196716} 02/26/2022 17:06:22 - INFO - codeparrot_training - Step 49045: {'lr': 4.881942157498043e-07, 'samples': 25111552, 'steps': 49045, 'loss/train': 0.18009762465953827} 02/26/2022 17:06:28 - INFO - codeparrot_training - Step 49046: {'lr': 4.8717268678386e-07, 'samples': 25112064, 'steps': 49046, 'loss/train': 1.4845880270004272} 02/26/2022 17:06:31 - INFO - codeparrot_training - Step 49047: {'lr': 4.861522266516316e-07, 'samples': 25112576, 'steps': 49047, 'loss/train': 0.9136321544647217} 02/26/2022 17:06:37 - INFO - codeparrot_training - Step 49048: {'lr': 4.851328353574768e-07, 'samples': 25113088, 'steps': 49048, 'loss/train': 2.304622173309326} 02/26/2022 17:06:40 - INFO - codeparrot_training - Step 49049: {'lr': 4.841145129058088e-07, 'samples': 25113600, 'steps': 49049, 'loss/train': 2.046469211578369} 02/26/2022 17:06:46 - INFO - codeparrot_training - Step 49050: {'lr': 4.830972593009575e-07, 'samples': 25114112, 'steps': 49050, 'loss/train': 2.2296531200408936} 02/26/2022 17:06:49 - INFO - codeparrot_training - Step 49051: {'lr': 4.820810745472803e-07, 'samples': 25114624, 'steps': 49051, 'loss/train': 1.6871041059494019} 02/26/2022 17:06:55 - INFO - codeparrot_training - Step 49052: {'lr': 4.81065958649135e-07, 'samples': 25115136, 'steps': 49052, 'loss/train': 1.5846202373504639} 02/26/2022 17:06:58 - INFO - codeparrot_training - Step 49053: {'lr': 4.800519116108792e-07, 'samples': 25115648, 'steps': 49053, 'loss/train': 1.346282958984375} 02/26/2022 17:07:04 - INFO - codeparrot_training - Step 49054: {'lr': 4.790389334368428e-07, 'samples': 25116160, 'steps': 49054, 'loss/train': 1.1217204332351685} 02/26/2022 17:07:07 - INFO - codeparrot_training - Step 49055: {'lr': 4.780270241313833e-07, 'samples': 25116672, 'steps': 49055, 'loss/train': 1.8381842374801636} 02/26/2022 17:07:13 - INFO - codeparrot_training - Step 49056: {'lr': 4.770161836988308e-07, 'samples': 25117184, 'steps': 49056, 'loss/train': 1.8852794170379639} 02/26/2022 17:07:17 - INFO - codeparrot_training - Step 49057: {'lr': 4.7600641214348706e-07, 'samples': 25117696, 'steps': 49057, 'loss/train': 1.2660834789276123} 02/26/2022 17:07:22 - INFO - codeparrot_training - Step 49058: {'lr': 4.7499770946971e-07, 'samples': 25118208, 'steps': 49058, 'loss/train': 1.4219340085983276} 02/26/2022 17:07:26 - INFO - codeparrot_training - Step 49059: {'lr': 4.7399007568180163e-07, 'samples': 25118720, 'steps': 49059, 'loss/train': 1.187503695487976} 02/26/2022 17:07:31 - INFO - codeparrot_training - Step 49060: {'lr': 4.7298351078409185e-07, 'samples': 25119232, 'steps': 49060, 'loss/train': 1.719525694847107} 02/26/2022 17:07:35 - INFO - codeparrot_training - Step 49061: {'lr': 4.7197801478088276e-07, 'samples': 25119744, 'steps': 49061, 'loss/train': 1.1388719081878662} 02/26/2022 17:07:41 - INFO - codeparrot_training - Step 49062: {'lr': 4.7097358767650423e-07, 'samples': 25120256, 'steps': 49062, 'loss/train': 1.2376247644424438} 02/26/2022 17:07:44 - INFO - codeparrot_training - Step 49063: {'lr': 4.699702294752306e-07, 'samples': 25120768, 'steps': 49063, 'loss/train': 2.1665024757385254} 02/26/2022 17:07:50 - INFO - codeparrot_training - Step 49064: {'lr': 4.6896794018136404e-07, 'samples': 25121280, 'steps': 49064, 'loss/train': 2.526989221572876} 02/26/2022 17:07:53 - INFO - codeparrot_training - Step 49065: {'lr': 4.679667197992066e-07, 'samples': 25121792, 'steps': 49065, 'loss/train': 2.167402982711792} 02/26/2022 17:07:59 - INFO - codeparrot_training - Step 49066: {'lr': 4.6696656833303265e-07, 'samples': 25122304, 'steps': 49066, 'loss/train': 0.0786314532160759} 02/26/2022 17:08:03 - INFO - codeparrot_training - Step 49067: {'lr': 4.659674857871443e-07, 'samples': 25122816, 'steps': 49067, 'loss/train': 1.325038194656372} 02/26/2022 17:08:08 - INFO - codeparrot_training - Step 49068: {'lr': 4.6496947216581597e-07, 'samples': 25123328, 'steps': 49068, 'loss/train': 1.980177402496338} 02/26/2022 17:08:12 - INFO - codeparrot_training - Step 49069: {'lr': 4.6397252747332196e-07, 'samples': 25123840, 'steps': 49069, 'loss/train': 1.6689904928207397} 02/26/2022 17:08:17 - INFO - codeparrot_training - Step 49070: {'lr': 4.6297665171393666e-07, 'samples': 25124352, 'steps': 49070, 'loss/train': 1.5461705923080444} 02/26/2022 17:08:21 - INFO - codeparrot_training - Step 49071: {'lr': 4.619818448919344e-07, 'samples': 25124864, 'steps': 49071, 'loss/train': 3.0336387157440186} 02/26/2022 17:08:26 - INFO - codeparrot_training - Step 49072: {'lr': 4.609881070115618e-07, 'samples': 25125376, 'steps': 49072, 'loss/train': 1.640491008758545} 02/26/2022 17:08:30 - INFO - codeparrot_training - Step 49073: {'lr': 4.5999543807706545e-07, 'samples': 25125888, 'steps': 49073, 'loss/train': 1.6270902156829834} 02/26/2022 17:08:35 - INFO - codeparrot_training - Step 49074: {'lr': 4.5900383809271974e-07, 'samples': 25126400, 'steps': 49074, 'loss/train': 1.3789494037628174} 02/26/2022 17:08:39 - INFO - codeparrot_training - Step 49075: {'lr': 4.580133070627435e-07, 'samples': 25126912, 'steps': 49075, 'loss/train': 1.1285934448242188} 02/26/2022 17:08:44 - INFO - codeparrot_training - Step 49076: {'lr': 4.5702384499141106e-07, 'samples': 25127424, 'steps': 49076, 'loss/train': 0.973630964756012} 02/26/2022 17:08:48 - INFO - codeparrot_training - Step 49077: {'lr': 4.56035451882969e-07, 'samples': 25127936, 'steps': 49077, 'loss/train': 2.1530659198760986} 02/26/2022 17:08:53 - INFO - codeparrot_training - Step 49078: {'lr': 4.5504812774160855e-07, 'samples': 25128448, 'steps': 49078, 'loss/train': 1.0737115144729614} 02/26/2022 17:08:57 - INFO - codeparrot_training - Step 49079: {'lr': 4.540618725716039e-07, 'samples': 25128960, 'steps': 49079, 'loss/train': 2.057705879211426} 02/26/2022 17:09:02 - INFO - codeparrot_training - Step 49080: {'lr': 4.530766863771185e-07, 'samples': 25129472, 'steps': 49080, 'loss/train': 1.524463176727295} 02/26/2022 17:09:06 - INFO - codeparrot_training - Step 49081: {'lr': 4.5209256916245443e-07, 'samples': 25129984, 'steps': 49081, 'loss/train': 1.3854734897613525} 02/26/2022 17:09:12 - INFO - codeparrot_training - Step 49082: {'lr': 4.51109520931775e-07, 'samples': 25130496, 'steps': 49082, 'loss/train': 1.759461522102356} 02/26/2022 17:09:16 - INFO - codeparrot_training - Step 49083: {'lr': 4.5012754168929916e-07, 'samples': 25131008, 'steps': 49083, 'loss/train': 1.9355318546295166} 02/26/2022 17:09:21 - INFO - codeparrot_training - Step 49084: {'lr': 4.4914663143921785e-07, 'samples': 25131520, 'steps': 49084, 'loss/train': 1.1921107769012451} 02/26/2022 17:09:25 - INFO - codeparrot_training - Step 49085: {'lr': 4.4816679018577775e-07, 'samples': 25132032, 'steps': 49085, 'loss/train': 2.0256187915802} 02/26/2022 17:09:30 - INFO - codeparrot_training - Step 49086: {'lr': 4.4718801793314224e-07, 'samples': 25132544, 'steps': 49086, 'loss/train': 0.46422213315963745} 02/26/2022 17:09:34 - INFO - codeparrot_training - Step 49087: {'lr': 4.4621031468550233e-07, 'samples': 25133056, 'steps': 49087, 'loss/train': 1.410831093788147} 02/26/2022 17:09:39 - INFO - codeparrot_training - Step 49088: {'lr': 4.4523368044704915e-07, 'samples': 25133568, 'steps': 49088, 'loss/train': 1.8755733966827393} 02/26/2022 17:09:43 - INFO - codeparrot_training - Step 49089: {'lr': 4.4425811522200154e-07, 'samples': 25134080, 'steps': 49089, 'loss/train': 0.9011662602424622} 02/26/2022 17:09:48 - INFO - codeparrot_training - Step 49090: {'lr': 4.432836190144951e-07, 'samples': 25134592, 'steps': 49090, 'loss/train': 0.1797219067811966} 02/26/2022 17:09:52 - INFO - codeparrot_training - Step 49091: {'lr': 4.4231019182869315e-07, 'samples': 25135104, 'steps': 49091, 'loss/train': 1.9175300598144531} 02/26/2022 17:09:58 - INFO - codeparrot_training - Step 49092: {'lr': 4.4133783366881455e-07, 'samples': 25135616, 'steps': 49092, 'loss/train': 0.7467092275619507} 02/26/2022 17:10:01 - INFO - codeparrot_training - Step 49093: {'lr': 4.4036654453899483e-07, 'samples': 25136128, 'steps': 49093, 'loss/train': 1.8505994081497192} 02/26/2022 17:10:07 - INFO - codeparrot_training - Step 49094: {'lr': 4.393963244434251e-07, 'samples': 25136640, 'steps': 49094, 'loss/train': 1.1297352313995361} 02/26/2022 17:10:10 - INFO - codeparrot_training - Step 49095: {'lr': 4.384271733862133e-07, 'samples': 25137152, 'steps': 49095, 'loss/train': 0.15467067062854767} 02/26/2022 17:10:16 - INFO - codeparrot_training - Step 49096: {'lr': 4.374590913715226e-07, 'samples': 25137664, 'steps': 49096, 'loss/train': 0.8506312370300293} 02/26/2022 17:10:19 - INFO - codeparrot_training - Step 49097: {'lr': 4.364920784035442e-07, 'samples': 25138176, 'steps': 49097, 'loss/train': 1.478582501411438} 02/26/2022 17:10:25 - INFO - codeparrot_training - Step 49098: {'lr': 4.3552613448635815e-07, 'samples': 25138688, 'steps': 49098, 'loss/train': 1.641661286354065} 02/26/2022 17:10:28 - INFO - codeparrot_training - Step 49099: {'lr': 4.345612596241555e-07, 'samples': 25139200, 'steps': 49099, 'loss/train': 2.45971941947937} 02/26/2022 17:10:34 - INFO - codeparrot_training - Step 49100: {'lr': 4.335974538210441e-07, 'samples': 25139712, 'steps': 49100, 'loss/train': 1.9928321838378906} 02/26/2022 17:10:37 - INFO - codeparrot_training - Step 49101: {'lr': 4.3263471708113177e-07, 'samples': 25140224, 'steps': 49101, 'loss/train': 1.9004757404327393} 02/26/2022 17:10:43 - INFO - codeparrot_training - Step 49102: {'lr': 4.3167304940858185e-07, 'samples': 25140736, 'steps': 49102, 'loss/train': 1.4035385847091675} 02/26/2022 17:10:49 - INFO - codeparrot_training - Step 49103: {'lr': 4.307124508075022e-07, 'samples': 25141248, 'steps': 49103, 'loss/train': 1.7648118734359741} 02/26/2022 17:10:52 - INFO - codeparrot_training - Step 49104: {'lr': 4.297529212820006e-07, 'samples': 25141760, 'steps': 49104, 'loss/train': 1.4929388761520386} 02/26/2022 17:10:57 - INFO - codeparrot_training - Step 49105: {'lr': 4.287944608361849e-07, 'samples': 25142272, 'steps': 49105, 'loss/train': 2.0259857177734375} 02/26/2022 17:11:01 - INFO - codeparrot_training - Step 49106: {'lr': 4.278370694741629e-07, 'samples': 25142784, 'steps': 49106, 'loss/train': 2.1846115589141846} 02/26/2022 17:11:07 - INFO - codeparrot_training - Step 49107: {'lr': 4.268807472000424e-07, 'samples': 25143296, 'steps': 49107, 'loss/train': 1.0400158166885376} 02/26/2022 17:11:10 - INFO - codeparrot_training - Step 49108: {'lr': 4.2592549401793133e-07, 'samples': 25143808, 'steps': 49108, 'loss/train': 1.0118441581726074} 02/26/2022 17:11:16 - INFO - codeparrot_training - Step 49109: {'lr': 4.2497130993188194e-07, 'samples': 25144320, 'steps': 49109, 'loss/train': 1.5299532413482666} 02/26/2022 17:11:19 - INFO - codeparrot_training - Step 49110: {'lr': 4.240181949460298e-07, 'samples': 25144832, 'steps': 49110, 'loss/train': 2.021423816680908} 02/26/2022 17:11:25 - INFO - codeparrot_training - Step 49111: {'lr': 4.230661490644272e-07, 'samples': 25145344, 'steps': 49111, 'loss/train': 1.249069094657898} 02/26/2022 17:11:28 - INFO - codeparrot_training - Step 49112: {'lr': 4.2211517229115427e-07, 'samples': 25145856, 'steps': 49112, 'loss/train': 3.2623772621154785} 02/26/2022 17:11:36 - INFO - codeparrot_training - Step 49113: {'lr': 4.21165264630291e-07, 'samples': 25146368, 'steps': 49113, 'loss/train': 1.301159143447876} 02/26/2022 17:11:39 - INFO - codeparrot_training - Step 49114: {'lr': 4.202164260858898e-07, 'samples': 25146880, 'steps': 49114, 'loss/train': 1.7586778402328491} 02/26/2022 17:11:45 - INFO - codeparrot_training - Step 49115: {'lr': 4.192686566620585e-07, 'samples': 25147392, 'steps': 49115, 'loss/train': 2.4094693660736084} 02/26/2022 17:11:48 - INFO - codeparrot_training - Step 49116: {'lr': 4.1832195636282154e-07, 'samples': 25147904, 'steps': 49116, 'loss/train': 0.9420658946037292} 02/26/2022 17:11:54 - INFO - codeparrot_training - Step 49117: {'lr': 4.1737632519223133e-07, 'samples': 25148416, 'steps': 49117, 'loss/train': 1.0134227275848389} 02/26/2022 17:11:57 - INFO - codeparrot_training - Step 49118: {'lr': 4.164317631543679e-07, 'samples': 25148928, 'steps': 49118, 'loss/train': 2.1292717456817627} 02/26/2022 17:12:03 - INFO - codeparrot_training - Step 49119: {'lr': 4.154882702532281e-07, 'samples': 25149440, 'steps': 49119, 'loss/train': 1.5120207071304321} 02/26/2022 17:12:06 - INFO - codeparrot_training - Step 49120: {'lr': 4.145458464929197e-07, 'samples': 25149952, 'steps': 49120, 'loss/train': 1.184522271156311} 02/26/2022 17:12:12 - INFO - codeparrot_training - Step 49121: {'lr': 4.1360449187741177e-07, 'samples': 25150464, 'steps': 49121, 'loss/train': 1.8079391717910767} 02/26/2022 17:12:15 - INFO - codeparrot_training - Step 49122: {'lr': 4.1266420641081215e-07, 'samples': 25150976, 'steps': 49122, 'loss/train': 1.9366484880447388} 02/26/2022 17:12:21 - INFO - codeparrot_training - Step 49123: {'lr': 4.117249900970621e-07, 'samples': 25151488, 'steps': 49123, 'loss/train': 0.13338087499141693} 02/26/2022 17:12:24 - INFO - codeparrot_training - Step 49124: {'lr': 4.107868429402417e-07, 'samples': 25152000, 'steps': 49124, 'loss/train': 2.923515796661377} 02/26/2022 17:12:32 - INFO - codeparrot_training - Step 49125: {'lr': 4.0984976494437553e-07, 'samples': 25152512, 'steps': 49125, 'loss/train': 1.2058892250061035} 02/26/2022 17:12:35 - INFO - codeparrot_training - Step 49126: {'lr': 4.089137561134604e-07, 'samples': 25153024, 'steps': 49126, 'loss/train': 0.6896197199821472} 02/26/2022 17:12:41 - INFO - codeparrot_training - Step 49127: {'lr': 4.079788164514653e-07, 'samples': 25153536, 'steps': 49127, 'loss/train': 0.6483401656150818} 02/26/2022 17:12:44 - INFO - codeparrot_training - Step 49128: {'lr': 4.070449459624703e-07, 'samples': 25154048, 'steps': 49128, 'loss/train': 1.4917714595794678} 02/26/2022 17:12:50 - INFO - codeparrot_training - Step 49129: {'lr': 4.0611214465044456e-07, 'samples': 25154560, 'steps': 49129, 'loss/train': 0.13677482306957245} 02/26/2022 17:12:53 - INFO - codeparrot_training - Step 49130: {'lr': 4.05180412519357e-07, 'samples': 25155072, 'steps': 49130, 'loss/train': 2.2253036499023438} 02/26/2022 17:12:59 - INFO - codeparrot_training - Step 49131: {'lr': 4.042497495732322e-07, 'samples': 25155584, 'steps': 49131, 'loss/train': 2.3909530639648438} 02/26/2022 17:13:02 - INFO - codeparrot_training - Step 49132: {'lr': 4.033201558160393e-07, 'samples': 25156096, 'steps': 49132, 'loss/train': 2.2927629947662354} 02/26/2022 17:13:08 - INFO - codeparrot_training - Step 49133: {'lr': 4.02391631251775e-07, 'samples': 25156608, 'steps': 49133, 'loss/train': 1.9030135869979858} 02/26/2022 17:13:11 - INFO - codeparrot_training - Step 49134: {'lr': 4.0146417588440844e-07, 'samples': 25157120, 'steps': 49134, 'loss/train': 0.7534550428390503} 02/26/2022 17:13:19 - INFO - codeparrot_training - Step 49135: {'lr': 4.0053778971790853e-07, 'samples': 25157632, 'steps': 49135, 'loss/train': 2.359708070755005} 02/26/2022 17:13:22 - INFO - codeparrot_training - Step 49136: {'lr': 3.9961247275624445e-07, 'samples': 25158144, 'steps': 49136, 'loss/train': 2.3942036628723145} 02/26/2022 17:13:28 - INFO - codeparrot_training - Step 49137: {'lr': 3.986882250033852e-07, 'samples': 25158656, 'steps': 49137, 'loss/train': 1.8561989068984985} 02/26/2022 17:13:31 - INFO - codeparrot_training - Step 49138: {'lr': 3.977650464632998e-07, 'samples': 25159168, 'steps': 49138, 'loss/train': 0.273765504360199} 02/26/2022 17:13:36 - INFO - codeparrot_training - Step 49139: {'lr': 3.968429371399296e-07, 'samples': 25159680, 'steps': 49139, 'loss/train': 2.343970775604248} 02/26/2022 17:13:40 - INFO - codeparrot_training - Step 49140: {'lr': 3.9592189703718804e-07, 'samples': 25160192, 'steps': 49140, 'loss/train': 1.297652244567871} 02/26/2022 17:13:46 - INFO - codeparrot_training - Step 49141: {'lr': 3.950019261590998e-07, 'samples': 25160704, 'steps': 49141, 'loss/train': 1.8379788398742676} 02/26/2022 17:13:49 - INFO - codeparrot_training - Step 49142: {'lr': 3.9408302450955057e-07, 'samples': 25161216, 'steps': 49142, 'loss/train': 2.3740673065185547} 02/26/2022 17:13:55 - INFO - codeparrot_training - Step 49143: {'lr': 3.9316519209250944e-07, 'samples': 25161728, 'steps': 49143, 'loss/train': 1.849450945854187} 02/26/2022 17:13:58 - INFO - codeparrot_training - Step 49144: {'lr': 3.922484289118622e-07, 'samples': 25162240, 'steps': 49144, 'loss/train': 1.4418665170669556} 02/26/2022 17:14:05 - INFO - codeparrot_training - Step 49145: {'lr': 3.9133273497157783e-07, 'samples': 25162752, 'steps': 49145, 'loss/train': 1.8954391479492188} 02/26/2022 17:14:09 - INFO - codeparrot_training - Step 49146: {'lr': 3.9041811027556994e-07, 'samples': 25163264, 'steps': 49146, 'loss/train': 2.1025638580322266} 02/26/2022 17:14:14 - INFO - codeparrot_training - Step 49147: {'lr': 3.8950455482775203e-07, 'samples': 25163776, 'steps': 49147, 'loss/train': 1.0569565296173096} 02/26/2022 17:14:18 - INFO - codeparrot_training - Step 49148: {'lr': 3.885920686320099e-07, 'samples': 25164288, 'steps': 49148, 'loss/train': 1.853453278541565} 02/26/2022 17:14:23 - INFO - codeparrot_training - Step 49149: {'lr': 3.876806516923126e-07, 'samples': 25164800, 'steps': 49149, 'loss/train': 0.3836728632450104} 02/26/2022 17:14:27 - INFO - codeparrot_training - Step 49150: {'lr': 3.8677030401249035e-07, 'samples': 25165312, 'steps': 49150, 'loss/train': 2.1091792583465576} 02/26/2022 17:14:32 - INFO - codeparrot_training - Step 49151: {'lr': 3.8586102559651226e-07, 'samples': 25165824, 'steps': 49151, 'loss/train': 0.9523171186447144} 02/26/2022 17:14:36 - INFO - codeparrot_training - Step 49152: {'lr': 3.849528164482363e-07, 'samples': 25166336, 'steps': 49152, 'loss/train': 2.497467517852783} 02/26/2022 17:14:41 - INFO - codeparrot_training - Step 49153: {'lr': 3.8404567657157607e-07, 'samples': 25166848, 'steps': 49153, 'loss/train': 1.6794003248214722} 02/26/2022 17:14:45 - INFO - codeparrot_training - Step 49154: {'lr': 3.8313960597038956e-07, 'samples': 25167360, 'steps': 49154, 'loss/train': 1.7243382930755615} 02/26/2022 17:14:52 - INFO - codeparrot_training - Step 49155: {'lr': 3.822346046485625e-07, 'samples': 25167872, 'steps': 49155, 'loss/train': 1.6368248462677002} 02/26/2022 17:14:55 - INFO - codeparrot_training - Step 49156: {'lr': 3.8133067260998076e-07, 'samples': 25168384, 'steps': 49156, 'loss/train': 1.7094802856445312} 02/26/2022 17:15:01 - INFO - codeparrot_training - Step 49157: {'lr': 3.804278098585301e-07, 'samples': 25168896, 'steps': 49157, 'loss/train': 6.278919696807861} 02/26/2022 17:15:04 - INFO - codeparrot_training - Step 49158: {'lr': 3.795260163980685e-07, 'samples': 25169408, 'steps': 49158, 'loss/train': 2.52380633354187} 02/26/2022 17:15:10 - INFO - codeparrot_training - Step 49159: {'lr': 3.7862529223242623e-07, 'samples': 25169920, 'steps': 49159, 'loss/train': 2.5490052700042725} 02/26/2022 17:15:13 - INFO - codeparrot_training - Step 49160: {'lr': 3.777256373655169e-07, 'samples': 25170432, 'steps': 49160, 'loss/train': 2.0267715454101562} 02/26/2022 17:15:19 - INFO - codeparrot_training - Step 49161: {'lr': 3.7682705180117074e-07, 'samples': 25170944, 'steps': 49161, 'loss/train': 1.6364741325378418} 02/26/2022 17:15:22 - INFO - codeparrot_training - Step 49162: {'lr': 3.759295355431902e-07, 'samples': 25171456, 'steps': 49162, 'loss/train': 1.6684178113937378} 02/26/2022 17:15:28 - INFO - codeparrot_training - Step 49163: {'lr': 3.7503308859551665e-07, 'samples': 25171968, 'steps': 49163, 'loss/train': 1.5207043886184692} 02/26/2022 17:15:31 - INFO - codeparrot_training - Step 49164: {'lr': 3.7413771096189707e-07, 'samples': 25172480, 'steps': 49164, 'loss/train': 2.12703537940979} 02/26/2022 17:15:37 - INFO - codeparrot_training - Step 49165: {'lr': 3.7324340264621724e-07, 'samples': 25172992, 'steps': 49165, 'loss/train': 1.8947240114212036} 02/26/2022 17:15:40 - INFO - codeparrot_training - Step 49166: {'lr': 3.7235016365230746e-07, 'samples': 25173504, 'steps': 49166, 'loss/train': 1.1667351722717285} 02/26/2022 17:15:46 - INFO - codeparrot_training - Step 49167: {'lr': 3.7145799398397016e-07, 'samples': 25174016, 'steps': 49167, 'loss/train': 2.478134870529175} 02/26/2022 17:15:50 - INFO - codeparrot_training - Step 49168: {'lr': 3.7056689364503573e-07, 'samples': 25174528, 'steps': 49168, 'loss/train': 1.4744770526885986} 02/26/2022 17:15:57 - INFO - codeparrot_training - Step 49169: {'lr': 3.696768626393343e-07, 'samples': 25175040, 'steps': 49169, 'loss/train': 0.8081668615341187} 02/26/2022 17:16:01 - INFO - codeparrot_training - Step 49170: {'lr': 3.687879009706963e-07, 'samples': 25175552, 'steps': 49170, 'loss/train': 2.199540615081787} 02/26/2022 17:16:06 - INFO - codeparrot_training - Step 49171: {'lr': 3.6790000864286855e-07, 'samples': 25176064, 'steps': 49171, 'loss/train': 1.9282736778259277} 02/26/2022 17:16:10 - INFO - codeparrot_training - Step 49172: {'lr': 3.670131856597092e-07, 'samples': 25176576, 'steps': 49172, 'loss/train': 1.5354024171829224} 02/26/2022 17:16:15 - INFO - codeparrot_training - Step 49173: {'lr': 3.66127432024993e-07, 'samples': 25177088, 'steps': 49173, 'loss/train': 2.1589131355285645} 02/26/2022 17:16:19 - INFO - codeparrot_training - Step 49174: {'lr': 3.6524274774252244e-07, 'samples': 25177600, 'steps': 49174, 'loss/train': 2.42380952835083} 02/26/2022 17:16:24 - INFO - codeparrot_training - Step 49175: {'lr': 3.6435913281607224e-07, 'samples': 25178112, 'steps': 49175, 'loss/train': 2.2026307582855225} 02/26/2022 17:16:28 - INFO - codeparrot_training - Step 49176: {'lr': 3.63476587249445e-07, 'samples': 25178624, 'steps': 49176, 'loss/train': 1.486001968383789} 02/26/2022 17:16:34 - INFO - codeparrot_training - Step 49177: {'lr': 3.625951110464154e-07, 'samples': 25179136, 'steps': 49177, 'loss/train': 1.705509901046753} 02/26/2022 17:16:37 - INFO - codeparrot_training - Step 49178: {'lr': 3.6171470421075825e-07, 'samples': 25179648, 'steps': 49178, 'loss/train': 1.1683207750320435} 02/26/2022 17:16:43 - INFO - codeparrot_training - Step 49179: {'lr': 3.608353667462483e-07, 'samples': 25180160, 'steps': 49179, 'loss/train': 0.4662502706050873} 02/26/2022 17:16:46 - INFO - codeparrot_training - Step 49180: {'lr': 3.599570986566603e-07, 'samples': 25180672, 'steps': 49180, 'loss/train': 0.8370378613471985} 02/26/2022 17:16:54 - INFO - codeparrot_training - Step 49181: {'lr': 3.590798999457412e-07, 'samples': 25181184, 'steps': 49181, 'loss/train': 1.2941521406173706} 02/26/2022 17:16:57 - INFO - codeparrot_training - Step 49182: {'lr': 3.582037706172381e-07, 'samples': 25181696, 'steps': 49182, 'loss/train': 1.1602590084075928} 02/26/2022 17:17:03 - INFO - codeparrot_training - Step 49183: {'lr': 3.573287106749534e-07, 'samples': 25182208, 'steps': 49183, 'loss/train': 1.65255868434906} 02/26/2022 17:17:06 - INFO - codeparrot_training - Step 49184: {'lr': 3.5645472012257876e-07, 'samples': 25182720, 'steps': 49184, 'loss/train': 2.339512586593628} 02/26/2022 17:17:10 - INFO - codeparrot_training - Step 49185: {'lr': 3.55581798963861e-07, 'samples': 25183232, 'steps': 49185, 'loss/train': 1.23650062084198} 02/26/2022 17:17:15 - INFO - codeparrot_training - Step 49186: {'lr': 3.5470994720260276e-07, 'samples': 25183744, 'steps': 49186, 'loss/train': 1.4002877473831177} 02/26/2022 17:17:19 - INFO - codeparrot_training - Step 49187: {'lr': 3.538391648424677e-07, 'samples': 25184256, 'steps': 49187, 'loss/train': 0.95508873462677} 02/26/2022 17:17:24 - INFO - codeparrot_training - Step 49188: {'lr': 3.5296945188723063e-07, 'samples': 25184768, 'steps': 49188, 'loss/train': 2.977891683578491} 02/26/2022 17:17:30 - INFO - codeparrot_training - Step 49189: {'lr': 3.52100808340583e-07, 'samples': 25185280, 'steps': 49189, 'loss/train': 2.0337424278259277} 02/26/2022 17:17:33 - INFO - codeparrot_training - Step 49190: {'lr': 3.5123323420627183e-07, 'samples': 25185792, 'steps': 49190, 'loss/train': 1.6148189306259155} 02/26/2022 17:17:41 - INFO - codeparrot_training - Step 49191: {'lr': 3.503667294879886e-07, 'samples': 25186304, 'steps': 49191, 'loss/train': 2.1883771419525146} 02/26/2022 17:17:44 - INFO - codeparrot_training - Step 49192: {'lr': 3.495012941894804e-07, 'samples': 25186816, 'steps': 49192, 'loss/train': 1.573521375656128} 02/26/2022 17:17:50 - INFO - codeparrot_training - Step 49193: {'lr': 3.486369283144386e-07, 'samples': 25187328, 'steps': 49193, 'loss/train': 2.792530059814453} 02/26/2022 17:17:53 - INFO - codeparrot_training - Step 49194: {'lr': 3.4777363186655477e-07, 'samples': 25187840, 'steps': 49194, 'loss/train': 0.9519041180610657} 02/26/2022 17:17:59 - INFO - codeparrot_training - Step 49195: {'lr': 3.469114048495481e-07, 'samples': 25188352, 'steps': 49195, 'loss/train': 1.179364800453186} 02/26/2022 17:18:02 - INFO - codeparrot_training - Step 49196: {'lr': 3.4605024726708235e-07, 'samples': 25188864, 'steps': 49196, 'loss/train': 2.2198383808135986} 02/26/2022 17:18:07 - INFO - codeparrot_training - Step 49197: {'lr': 3.4519015912287677e-07, 'samples': 25189376, 'steps': 49197, 'loss/train': 1.5995867252349854} 02/26/2022 17:18:11 - INFO - codeparrot_training - Step 49198: {'lr': 3.443311404205951e-07, 'samples': 25189888, 'steps': 49198, 'loss/train': 1.0336480140686035} 02/26/2022 17:18:16 - INFO - codeparrot_training - Step 49199: {'lr': 3.4347319116390106e-07, 'samples': 25190400, 'steps': 49199, 'loss/train': 0.5650877952575684} 02/26/2022 17:18:20 - INFO - codeparrot_training - Step 49200: {'lr': 3.426163113565417e-07, 'samples': 25190912, 'steps': 49200, 'loss/train': 1.613423228263855} 02/26/2022 17:18:26 - INFO - codeparrot_training - Step 49201: {'lr': 3.4176050100209744e-07, 'samples': 25191424, 'steps': 49201, 'loss/train': 1.7087675333023071} 02/26/2022 17:18:29 - INFO - codeparrot_training - Step 49202: {'lr': 3.409057601042875e-07, 'samples': 25191936, 'steps': 49202, 'loss/train': 2.2779128551483154} 02/26/2022 17:18:35 - INFO - codeparrot_training - Step 49203: {'lr': 3.4005208866677574e-07, 'samples': 25192448, 'steps': 49203, 'loss/train': 2.2067148685455322} 02/26/2022 17:18:38 - INFO - codeparrot_training - Step 49204: {'lr': 3.39199486693198e-07, 'samples': 25192960, 'steps': 49204, 'loss/train': 1.5900170803070068} 02/26/2022 17:18:45 - INFO - codeparrot_training - Step 49205: {'lr': 3.383479541872181e-07, 'samples': 25193472, 'steps': 49205, 'loss/train': 0.923997163772583} 02/26/2022 17:18:49 - INFO - codeparrot_training - Step 49206: {'lr': 3.3749749115247197e-07, 'samples': 25193984, 'steps': 49206, 'loss/train': 1.6662477254867554} 02/26/2022 17:18:54 - INFO - codeparrot_training - Step 49207: {'lr': 3.3664809759259565e-07, 'samples': 25194496, 'steps': 49207, 'loss/train': 0.8580371141433716} 02/26/2022 17:18:58 - INFO - codeparrot_training - Step 49208: {'lr': 3.357997735112528e-07, 'samples': 25195008, 'steps': 49208, 'loss/train': 1.279490351676941} 02/26/2022 17:19:03 - INFO - codeparrot_training - Step 49209: {'lr': 3.3495251891207946e-07, 'samples': 25195520, 'steps': 49209, 'loss/train': 1.272603988647461} 02/26/2022 17:19:07 - INFO - codeparrot_training - Step 49210: {'lr': 3.3410633379868383e-07, 'samples': 25196032, 'steps': 49210, 'loss/train': 2.558316946029663} 02/26/2022 17:19:12 - INFO - codeparrot_training - Step 49211: {'lr': 3.332612181747019e-07, 'samples': 25196544, 'steps': 49211, 'loss/train': 1.1882816553115845} 02/26/2022 17:19:16 - INFO - codeparrot_training - Step 49212: {'lr': 3.324171720437419e-07, 'samples': 25197056, 'steps': 49212, 'loss/train': 1.201464056968689} 02/26/2022 17:19:21 - INFO - codeparrot_training - Step 49213: {'lr': 3.31574195409412e-07, 'samples': 25197568, 'steps': 49213, 'loss/train': 1.187307596206665} 02/26/2022 17:19:25 - INFO - codeparrot_training - Step 49214: {'lr': 3.3073228827537606e-07, 'samples': 25198080, 'steps': 49214, 'loss/train': 2.3933403491973877} 02/26/2022 17:19:30 - INFO - codeparrot_training - Step 49215: {'lr': 3.2989145064518665e-07, 'samples': 25198592, 'steps': 49215, 'loss/train': 1.3481743335723877} 02/26/2022 17:19:33 - INFO - codeparrot_training - Step 49216: {'lr': 3.290516825224521e-07, 'samples': 25199104, 'steps': 49216, 'loss/train': 0.9960803389549255} 02/26/2022 17:19:41 - INFO - codeparrot_training - Step 49217: {'lr': 3.2821298391080835e-07, 'samples': 25199616, 'steps': 49217, 'loss/train': 1.0293229818344116} 02/26/2022 17:19:44 - INFO - codeparrot_training - Step 49218: {'lr': 3.2737535481380807e-07, 'samples': 25200128, 'steps': 49218, 'loss/train': 1.533575177192688} 02/26/2022 17:19:50 - INFO - codeparrot_training - Step 49219: {'lr': 3.265387952350318e-07, 'samples': 25200640, 'steps': 49219, 'loss/train': 0.6177148222923279} 02/26/2022 17:19:53 - INFO - codeparrot_training - Step 49220: {'lr': 3.2570330517811555e-07, 'samples': 25201152, 'steps': 49220, 'loss/train': 4.056741237640381} 02/26/2022 17:19:59 - INFO - codeparrot_training - Step 49221: {'lr': 3.2486888464658413e-07, 'samples': 25201664, 'steps': 49221, 'loss/train': 1.9567835330963135} 02/26/2022 17:20:02 - INFO - codeparrot_training - Step 49222: {'lr': 3.240355336440459e-07, 'samples': 25202176, 'steps': 49222, 'loss/train': 1.888270616531372} 02/26/2022 17:20:08 - INFO - codeparrot_training - Step 49223: {'lr': 3.2320325217405354e-07, 'samples': 25202688, 'steps': 49223, 'loss/train': 1.583760380744934} 02/26/2022 17:20:11 - INFO - codeparrot_training - Step 49224: {'lr': 3.2237204024015976e-07, 'samples': 25203200, 'steps': 49224, 'loss/train': 2.663724184036255} 02/26/2022 17:20:17 - INFO - codeparrot_training - Step 49225: {'lr': 3.215418978459728e-07, 'samples': 25203712, 'steps': 49225, 'loss/train': 2.2093865871429443} 02/26/2022 17:20:20 - INFO - codeparrot_training - Step 49226: {'lr': 3.207128249949898e-07, 'samples': 25204224, 'steps': 49226, 'loss/train': 0.3780570924282074} 02/26/2022 17:20:27 - INFO - codeparrot_training - Step 49227: {'lr': 3.19884821690819e-07, 'samples': 25204736, 'steps': 49227, 'loss/train': 1.0628679990768433} 02/26/2022 17:20:31 - INFO - codeparrot_training - Step 49228: {'lr': 3.190578879369577e-07, 'samples': 25205248, 'steps': 49228, 'loss/train': 1.1448765993118286} 02/26/2022 17:20:36 - INFO - codeparrot_training - Step 49229: {'lr': 3.182320237369862e-07, 'samples': 25205760, 'steps': 49229, 'loss/train': 1.6691434383392334} 02/26/2022 17:20:40 - INFO - codeparrot_training - Step 49230: {'lr': 3.174072290944019e-07, 'samples': 25206272, 'steps': 49230, 'loss/train': 1.346313714981079} 02/26/2022 17:20:46 - INFO - codeparrot_training - Step 49231: {'lr': 3.165835040127851e-07, 'samples': 25206784, 'steps': 49231, 'loss/train': 1.961909532546997} 02/26/2022 17:20:49 - INFO - codeparrot_training - Step 49232: {'lr': 3.1576084849563315e-07, 'samples': 25207296, 'steps': 49232, 'loss/train': 1.9196834564208984} 02/26/2022 17:20:55 - INFO - codeparrot_training - Step 49233: {'lr': 3.149392625464709e-07, 'samples': 25207808, 'steps': 49233, 'loss/train': 0.5348854660987854} 02/26/2022 17:20:58 - INFO - codeparrot_training - Step 49234: {'lr': 3.141187461688233e-07, 'samples': 25208320, 'steps': 49234, 'loss/train': 1.8401069641113281} 02/26/2022 17:21:04 - INFO - codeparrot_training - Step 49235: {'lr': 3.1329929936621536e-07, 'samples': 25208832, 'steps': 49235, 'loss/train': 2.233855962753296} 02/26/2022 17:21:08 - INFO - codeparrot_training - Step 49236: {'lr': 3.1248092214217206e-07, 'samples': 25209344, 'steps': 49236, 'loss/train': 2.24180006980896} 02/26/2022 17:21:15 - INFO - codeparrot_training - Step 49237: {'lr': 3.11663614500135e-07, 'samples': 25209856, 'steps': 49237, 'loss/train': 1.7753978967666626} 02/26/2022 17:21:18 - INFO - codeparrot_training - Step 49238: {'lr': 3.1084737644365703e-07, 'samples': 25210368, 'steps': 49238, 'loss/train': 1.365486741065979} 02/26/2022 17:21:22 - INFO - codeparrot_training - Step 49239: {'lr': 3.100322079762352e-07, 'samples': 25210880, 'steps': 49239, 'loss/train': 2.2111904621124268} 02/26/2022 17:21:27 - INFO - codeparrot_training - Step 49240: {'lr': 3.0921810910133906e-07, 'samples': 25211392, 'steps': 49240, 'loss/train': 1.0475375652313232} 02/26/2022 17:21:31 - INFO - codeparrot_training - Step 49241: {'lr': 3.0840507982249354e-07, 'samples': 25211904, 'steps': 49241, 'loss/train': 1.9657444953918457} 02/26/2022 17:21:36 - INFO - codeparrot_training - Step 49242: {'lr': 3.075931201431126e-07, 'samples': 25212416, 'steps': 49242, 'loss/train': 1.1375079154968262} 02/26/2022 17:21:40 - INFO - codeparrot_training - Step 49243: {'lr': 3.067822300667489e-07, 'samples': 25212928, 'steps': 49243, 'loss/train': 1.6300870180130005} 02/26/2022 17:21:45 - INFO - codeparrot_training - Step 49244: {'lr': 3.059724095968441e-07, 'samples': 25213440, 'steps': 49244, 'loss/train': 3.051119089126587} 02/26/2022 17:21:49 - INFO - codeparrot_training - Step 49245: {'lr': 3.0516365873684004e-07, 'samples': 25213952, 'steps': 49245, 'loss/train': 1.5433343648910522} 02/26/2022 17:21:54 - INFO - codeparrot_training - Step 49246: {'lr': 3.043559774902338e-07, 'samples': 25214464, 'steps': 49246, 'loss/train': 0.4882158041000366} 02/26/2022 17:21:58 - INFO - codeparrot_training - Step 49247: {'lr': 3.0354936586049486e-07, 'samples': 25214976, 'steps': 49247, 'loss/train': 1.7533336877822876} 02/26/2022 17:22:03 - INFO - codeparrot_training - Step 49248: {'lr': 3.027438238510649e-07, 'samples': 25215488, 'steps': 49248, 'loss/train': 1.7116378545761108} 02/26/2022 17:22:07 - INFO - codeparrot_training - Step 49249: {'lr': 3.019393514653857e-07, 'samples': 25216000, 'steps': 49249, 'loss/train': 2.323669195175171} 02/26/2022 17:22:12 - INFO - codeparrot_training - Step 49250: {'lr': 3.0113594870689873e-07, 'samples': 25216512, 'steps': 49250, 'loss/train': 1.2697666883468628} 02/26/2022 17:22:20 - INFO - codeparrot_training - Step 49251: {'lr': 3.003336155790737e-07, 'samples': 25217024, 'steps': 49251, 'loss/train': 2.723059892654419} 02/26/2022 17:22:23 - INFO - codeparrot_training - Step 49252: {'lr': 2.9953235208532437e-07, 'samples': 25217536, 'steps': 49252, 'loss/train': 1.6497389078140259} 02/26/2022 17:22:28 - INFO - codeparrot_training - Step 49253: {'lr': 2.987321582290925e-07, 'samples': 25218048, 'steps': 49253, 'loss/train': 2.370401382446289} 02/26/2022 17:22:32 - INFO - codeparrot_training - Step 49254: {'lr': 2.97933034013792e-07, 'samples': 25218560, 'steps': 49254, 'loss/train': 1.3297098875045776} 02/26/2022 17:22:37 - INFO - codeparrot_training - Step 49255: {'lr': 2.971349794428646e-07, 'samples': 25219072, 'steps': 49255, 'loss/train': 1.72870934009552} 02/26/2022 17:22:41 - INFO - codeparrot_training - Step 49256: {'lr': 2.963379945197242e-07, 'samples': 25219584, 'steps': 49256, 'loss/train': 1.5735093355178833} 02/26/2022 17:22:47 - INFO - codeparrot_training - Step 49257: {'lr': 2.9554207924778475e-07, 'samples': 25220096, 'steps': 49257, 'loss/train': 1.419816017150879} 02/26/2022 17:22:50 - INFO - codeparrot_training - Step 49258: {'lr': 2.947472336304324e-07, 'samples': 25220608, 'steps': 49258, 'loss/train': 0.5545970797538757} 02/26/2022 17:22:56 - INFO - codeparrot_training - Step 49259: {'lr': 2.939534576711089e-07, 'samples': 25221120, 'steps': 49259, 'loss/train': 2.7279176712036133} 02/26/2022 17:22:59 - INFO - codeparrot_training - Step 49260: {'lr': 2.931607513732004e-07, 'samples': 25221632, 'steps': 49260, 'loss/train': 1.7415560483932495} 02/26/2022 17:23:03 - INFO - codeparrot_training - Step 49261: {'lr': 2.923691147400931e-07, 'samples': 25222144, 'steps': 49261, 'loss/train': 0.4381670355796814} 02/26/2022 17:23:08 - INFO - codeparrot_training - Step 49262: {'lr': 2.915785477752009e-07, 'samples': 25222656, 'steps': 49262, 'loss/train': 2.497663736343384} 02/26/2022 17:23:12 - INFO - codeparrot_training - Step 49263: {'lr': 2.907890504818822e-07, 'samples': 25223168, 'steps': 49263, 'loss/train': 1.6097497940063477} 02/26/2022 17:23:19 - INFO - codeparrot_training - Step 49264: {'lr': 2.9000062286352326e-07, 'samples': 25223680, 'steps': 49264, 'loss/train': 1.1878303289413452} 02/26/2022 17:23:22 - INFO - codeparrot_training - Step 49265: {'lr': 2.892132649235102e-07, 'samples': 25224192, 'steps': 49265, 'loss/train': 1.3831366300582886} 02/26/2022 17:23:28 - INFO - codeparrot_training - Step 49266: {'lr': 2.884269766652292e-07, 'samples': 25224704, 'steps': 49266, 'loss/train': 2.0943593978881836} 02/26/2022 17:23:31 - INFO - codeparrot_training - Step 49267: {'lr': 2.8764175809203873e-07, 'samples': 25225216, 'steps': 49267, 'loss/train': 2.1908011436462402} 02/26/2022 17:23:37 - INFO - codeparrot_training - Step 49268: {'lr': 2.868576092072972e-07, 'samples': 25225728, 'steps': 49268, 'loss/train': 1.7559328079223633} 02/26/2022 17:23:41 - INFO - codeparrot_training - Step 49269: {'lr': 2.860745300143353e-07, 'samples': 25226240, 'steps': 49269, 'loss/train': 1.9213505983352661} 02/26/2022 17:23:46 - INFO - codeparrot_training - Step 49270: {'lr': 2.852925205165668e-07, 'samples': 25226752, 'steps': 49270, 'loss/train': 3.552980661392212} 02/26/2022 17:23:50 - INFO - codeparrot_training - Step 49271: {'lr': 2.845115807172949e-07, 'samples': 25227264, 'steps': 49271, 'loss/train': 1.8593324422836304} 02/26/2022 17:23:55 - INFO - codeparrot_training - Step 49272: {'lr': 2.8373171061990553e-07, 'samples': 25227776, 'steps': 49272, 'loss/train': 1.9358209371566772} 02/26/2022 17:23:59 - INFO - codeparrot_training - Step 49273: {'lr': 2.829529102277018e-07, 'samples': 25228288, 'steps': 49273, 'loss/train': 0.28513431549072266} 02/26/2022 17:24:06 - INFO - codeparrot_training - Step 49274: {'lr': 2.8217517954404194e-07, 'samples': 25228800, 'steps': 49274, 'loss/train': 2.081965684890747} 02/26/2022 17:24:09 - INFO - codeparrot_training - Step 49275: {'lr': 2.81398518572229e-07, 'samples': 25229312, 'steps': 49275, 'loss/train': 1.3904247283935547} 02/26/2022 17:24:15 - INFO - codeparrot_training - Step 49276: {'lr': 2.806229273156213e-07, 'samples': 25229824, 'steps': 49276, 'loss/train': 1.8676384687423706} 02/26/2022 17:24:18 - INFO - codeparrot_training - Step 49277: {'lr': 2.798484057775219e-07, 'samples': 25230336, 'steps': 49277, 'loss/train': 0.11385231465101242} 02/26/2022 17:24:24 - INFO - codeparrot_training - Step 49278: {'lr': 2.7907495396126137e-07, 'samples': 25230848, 'steps': 49278, 'loss/train': 1.4199883937835693} 02/26/2022 17:24:28 - INFO - codeparrot_training - Step 49279: {'lr': 2.783025718701704e-07, 'samples': 25231360, 'steps': 49279, 'loss/train': 0.138031467795372} 02/26/2022 17:24:33 - INFO - codeparrot_training - Step 49280: {'lr': 2.7753125950752413e-07, 'samples': 25231872, 'steps': 49280, 'loss/train': 1.5217351913452148} 02/26/2022 17:24:37 - INFO - codeparrot_training - Step 49281: {'lr': 2.7676101687662545e-07, 'samples': 25232384, 'steps': 49281, 'loss/train': 2.1831204891204834} 02/26/2022 17:24:42 - INFO - codeparrot_training - Step 49282: {'lr': 2.7599184398077737e-07, 'samples': 25232896, 'steps': 49282, 'loss/train': 0.7931455969810486} 02/26/2022 17:24:46 - INFO - codeparrot_training - Step 49283: {'lr': 2.752237408233105e-07, 'samples': 25233408, 'steps': 49283, 'loss/train': 2.4793100357055664} 02/26/2022 17:24:53 - INFO - codeparrot_training - Step 49284: {'lr': 2.744567074074722e-07, 'samples': 25233920, 'steps': 49284, 'loss/train': 0.9266887903213501} 02/26/2022 17:24:56 - INFO - codeparrot_training - Step 49285: {'lr': 2.7369074373656546e-07, 'samples': 25234432, 'steps': 49285, 'loss/train': 1.0195916891098022} 02/26/2022 17:25:02 - INFO - codeparrot_training - Step 49286: {'lr': 2.729258498138654e-07, 'samples': 25234944, 'steps': 49286, 'loss/train': 1.062586784362793} 02/26/2022 17:25:06 - INFO - codeparrot_training - Step 49287: {'lr': 2.721620256426749e-07, 'samples': 25235456, 'steps': 49287, 'loss/train': 2.073833465576172} 02/26/2022 17:25:11 - INFO - codeparrot_training - Step 49288: {'lr': 2.713992712262414e-07, 'samples': 25235968, 'steps': 49288, 'loss/train': 1.8696194887161255} 02/26/2022 17:25:15 - INFO - codeparrot_training - Step 49289: {'lr': 2.706375865678123e-07, 'samples': 25236480, 'steps': 49289, 'loss/train': 1.149975061416626} 02/26/2022 17:25:20 - INFO - codeparrot_training - Step 49290: {'lr': 2.698769716706906e-07, 'samples': 25236992, 'steps': 49290, 'loss/train': 1.6504911184310913} 02/26/2022 17:25:24 - INFO - codeparrot_training - Step 49291: {'lr': 2.691174265381235e-07, 'samples': 25237504, 'steps': 49291, 'loss/train': 1.89723801612854} 02/26/2022 17:25:29 - INFO - codeparrot_training - Step 49292: {'lr': 2.6835895117335864e-07, 'samples': 25238016, 'steps': 49292, 'loss/train': 1.4044740200042725} 02/26/2022 17:25:33 - INFO - codeparrot_training - Step 49293: {'lr': 2.6760154557961547e-07, 'samples': 25238528, 'steps': 49293, 'loss/train': 1.5652474164962769} 02/26/2022 17:25:38 - INFO - codeparrot_training - Step 49294: {'lr': 2.66845209760197e-07, 'samples': 25239040, 'steps': 49294, 'loss/train': 0.8309801816940308} 02/26/2022 17:25:42 - INFO - codeparrot_training - Step 49295: {'lr': 2.660899437182951e-07, 'samples': 25239552, 'steps': 49295, 'loss/train': 2.2028732299804688} 02/26/2022 17:25:47 - INFO - codeparrot_training - Step 49296: {'lr': 2.653357474571849e-07, 'samples': 25240064, 'steps': 49296, 'loss/train': 1.136959195137024} 02/26/2022 17:25:51 - INFO - codeparrot_training - Step 49297: {'lr': 2.6458262098005834e-07, 'samples': 25240576, 'steps': 49297, 'loss/train': 1.6235889196395874} 02/26/2022 17:25:56 - INFO - codeparrot_training - Step 49298: {'lr': 2.638305642901906e-07, 'samples': 25241088, 'steps': 49298, 'loss/train': 0.2817689776420593} 02/26/2022 17:26:00 - INFO - codeparrot_training - Step 49299: {'lr': 2.630795773907457e-07, 'samples': 25241600, 'steps': 49299, 'loss/train': 0.5102606415748596} 02/26/2022 17:26:07 - INFO - codeparrot_training - Step 49300: {'lr': 2.623296602849712e-07, 'samples': 25242112, 'steps': 49300, 'loss/train': 1.696257472038269} 02/26/2022 17:26:10 - INFO - codeparrot_training - Step 49301: {'lr': 2.615808129760866e-07, 'samples': 25242624, 'steps': 49301, 'loss/train': 1.798613429069519} 02/26/2022 17:26:16 - INFO - codeparrot_training - Step 49302: {'lr': 2.6083303546728387e-07, 'samples': 25243136, 'steps': 49302, 'loss/train': 0.9105786681175232} 02/26/2022 17:26:19 - INFO - codeparrot_training - Step 49303: {'lr': 2.600863277617549e-07, 'samples': 25243648, 'steps': 49303, 'loss/train': 1.6797343492507935} 02/26/2022 17:26:25 - INFO - codeparrot_training - Step 49304: {'lr': 2.593406898627193e-07, 'samples': 25244160, 'steps': 49304, 'loss/train': 1.3931041955947876} 02/26/2022 17:26:29 - INFO - codeparrot_training - Step 49305: {'lr': 2.5859612177339673e-07, 'samples': 25244672, 'steps': 49305, 'loss/train': 1.4529887437820435} 02/26/2022 17:26:34 - INFO - codeparrot_training - Step 49306: {'lr': 2.578526234969236e-07, 'samples': 25245184, 'steps': 49306, 'loss/train': 1.996252179145813} 02/26/2022 17:26:38 - INFO - codeparrot_training - Step 49307: {'lr': 2.571101950364918e-07, 'samples': 25245696, 'steps': 49307, 'loss/train': 1.5781975984573364} 02/26/2022 17:26:43 - INFO - codeparrot_training - Step 49308: {'lr': 2.5636883639532094e-07, 'samples': 25246208, 'steps': 49308, 'loss/train': 1.5069892406463623} 02/26/2022 17:26:47 - INFO - codeparrot_training - Step 49309: {'lr': 2.5562854757657516e-07, 'samples': 25246720, 'steps': 49309, 'loss/train': 1.6497517824172974} 02/26/2022 17:26:54 - INFO - codeparrot_training - Step 49310: {'lr': 2.5488932858339084e-07, 'samples': 25247232, 'steps': 49310, 'loss/train': 1.9937832355499268} 02/26/2022 17:26:58 - INFO - codeparrot_training - Step 49311: {'lr': 2.5415117941895994e-07, 'samples': 25247744, 'steps': 49311, 'loss/train': 1.3809044361114502} 02/26/2022 17:27:04 - INFO - codeparrot_training - Step 49312: {'lr': 2.534141000864465e-07, 'samples': 25248256, 'steps': 49312, 'loss/train': 1.4051517248153687} 02/26/2022 17:27:07 - INFO - codeparrot_training - Step 49313: {'lr': 2.526780905889869e-07, 'samples': 25248768, 'steps': 49313, 'loss/train': 2.258091449737549} 02/26/2022 17:27:11 - INFO - codeparrot_training - Step 49314: {'lr': 2.5194315092977315e-07, 'samples': 25249280, 'steps': 49314, 'loss/train': 3.379939079284668} 02/26/2022 17:27:16 - INFO - codeparrot_training - Step 49315: {'lr': 2.512092811119415e-07, 'samples': 25249792, 'steps': 49315, 'loss/train': 0.4511565864086151} 02/26/2022 17:27:20 - INFO - codeparrot_training - Step 49316: {'lr': 2.504764811386007e-07, 'samples': 25250304, 'steps': 49316, 'loss/train': 1.4937771558761597} 02/26/2022 17:27:25 - INFO - codeparrot_training - Step 49317: {'lr': 2.497447510129147e-07, 'samples': 25250816, 'steps': 49317, 'loss/train': 1.8493415117263794} 02/26/2022 17:27:29 - INFO - codeparrot_training - Step 49318: {'lr': 2.4901409073802006e-07, 'samples': 25251328, 'steps': 49318, 'loss/train': 1.8940860033035278} 02/26/2022 17:27:34 - INFO - codeparrot_training - Step 49319: {'lr': 2.482845003170531e-07, 'samples': 25251840, 'steps': 49319, 'loss/train': 1.6199268102645874} 02/26/2022 17:27:42 - INFO - codeparrot_training - Step 49320: {'lr': 2.475559797531224e-07, 'samples': 25252352, 'steps': 49320, 'loss/train': 0.8242297172546387} 02/26/2022 17:27:45 - INFO - codeparrot_training - Step 49321: {'lr': 2.4682852904936434e-07, 'samples': 25252864, 'steps': 49321, 'loss/train': 2.161670684814453} 02/26/2022 17:27:51 - INFO - codeparrot_training - Step 49322: {'lr': 2.4610214820888767e-07, 'samples': 25253376, 'steps': 49322, 'loss/train': 2.2019267082214355} 02/26/2022 17:27:54 - INFO - codeparrot_training - Step 49323: {'lr': 2.4537683723480087e-07, 'samples': 25253888, 'steps': 49323, 'loss/train': 0.9363074898719788} 02/26/2022 17:27:59 - INFO - codeparrot_training - Step 49324: {'lr': 2.4465259613021263e-07, 'samples': 25254400, 'steps': 49324, 'loss/train': 1.4704328775405884} 02/26/2022 17:28:03 - INFO - codeparrot_training - Step 49325: {'lr': 2.439294248982038e-07, 'samples': 25254912, 'steps': 49325, 'loss/train': 2.071948766708374} 02/26/2022 17:28:08 - INFO - codeparrot_training - Step 49326: {'lr': 2.4320732354191076e-07, 'samples': 25255424, 'steps': 49326, 'loss/train': 2.5521204471588135} 02/26/2022 17:28:12 - INFO - codeparrot_training - Step 49327: {'lr': 2.4248629206444215e-07, 'samples': 25255936, 'steps': 49327, 'loss/train': 1.24164617061615} 02/26/2022 17:28:17 - INFO - codeparrot_training - Step 49328: {'lr': 2.4176633046882337e-07, 'samples': 25256448, 'steps': 49328, 'loss/train': 1.9267997741699219} 02/26/2022 17:28:21 - INFO - codeparrot_training - Step 49329: {'lr': 2.410474387581629e-07, 'samples': 25256960, 'steps': 49329, 'loss/train': 2.0199813842773438} 02/26/2022 17:28:28 - INFO - codeparrot_training - Step 49330: {'lr': 2.4032961693556956e-07, 'samples': 25257472, 'steps': 49330, 'loss/train': 2.0545132160186768} 02/26/2022 17:28:32 - INFO - codeparrot_training - Step 49331: {'lr': 2.396128650040963e-07, 'samples': 25257984, 'steps': 49331, 'loss/train': 1.6450552940368652} 02/26/2022 17:28:37 - INFO - codeparrot_training - Step 49332: {'lr': 2.388971829667963e-07, 'samples': 25258496, 'steps': 49332, 'loss/train': 1.4598097801208496} 02/26/2022 17:28:40 - INFO - codeparrot_training - Step 49333: {'lr': 2.3818257082677818e-07, 'samples': 25259008, 'steps': 49333, 'loss/train': 1.7282825708389282} 02/26/2022 17:28:46 - INFO - codeparrot_training - Step 49334: {'lr': 2.3746902858703956e-07, 'samples': 25259520, 'steps': 49334, 'loss/train': 0.3359510004520416} 02/26/2022 17:28:49 - INFO - codeparrot_training - Step 49335: {'lr': 2.3675655625068905e-07, 'samples': 25260032, 'steps': 49335, 'loss/train': 2.07143235206604} 02/26/2022 17:28:55 - INFO - codeparrot_training - Step 49336: {'lr': 2.3604515382077974e-07, 'samples': 25260544, 'steps': 49336, 'loss/train': 0.24383077025413513} 02/26/2022 17:28:58 - INFO - codeparrot_training - Step 49337: {'lr': 2.3533482130033702e-07, 'samples': 25261056, 'steps': 49337, 'loss/train': 2.0739169120788574} 02/26/2022 17:29:04 - INFO - codeparrot_training - Step 49338: {'lr': 2.3462555869241398e-07, 'samples': 25261568, 'steps': 49338, 'loss/train': 0.6827960014343262} 02/26/2022 17:29:08 - INFO - codeparrot_training - Step 49339: {'lr': 2.33917366000036e-07, 'samples': 25262080, 'steps': 49339, 'loss/train': 1.594533920288086} 02/26/2022 17:29:13 - INFO - codeparrot_training - Step 49340: {'lr': 2.3321024322625618e-07, 'samples': 25262592, 'steps': 49340, 'loss/train': 1.655191421508789} 02/26/2022 17:29:17 - INFO - codeparrot_training - Step 49341: {'lr': 2.3250419037409986e-07, 'samples': 25263104, 'steps': 49341, 'loss/train': 1.4917610883712769} 02/26/2022 17:29:22 - INFO - codeparrot_training - Step 49342: {'lr': 2.317992074465647e-07, 'samples': 25263616, 'steps': 49342, 'loss/train': 1.2126142978668213} 02/26/2022 17:29:26 - INFO - codeparrot_training - Step 49343: {'lr': 2.3109529444670374e-07, 'samples': 25264128, 'steps': 49343, 'loss/train': 1.5893272161483765} 02/26/2022 17:29:31 - INFO - codeparrot_training - Step 49344: {'lr': 2.3039245137751463e-07, 'samples': 25264640, 'steps': 49344, 'loss/train': 1.7388670444488525} 02/26/2022 17:29:35 - INFO - codeparrot_training - Step 49345: {'lr': 2.296906782420227e-07, 'samples': 25265152, 'steps': 49345, 'loss/train': 1.2293654680252075} 02/26/2022 17:29:42 - INFO - codeparrot_training - Step 49346: {'lr': 2.289899750432256e-07, 'samples': 25265664, 'steps': 49346, 'loss/train': 2.2173004150390625} 02/26/2022 17:29:45 - INFO - codeparrot_training - Step 49347: {'lr': 2.2829034178412088e-07, 'samples': 25266176, 'steps': 49347, 'loss/train': 1.1115319728851318} 02/26/2022 17:29:51 - INFO - codeparrot_training - Step 49348: {'lr': 2.275917784677062e-07, 'samples': 25266688, 'steps': 49348, 'loss/train': 1.7598215341567993} 02/26/2022 17:29:54 - INFO - codeparrot_training - Step 49349: {'lr': 2.2689428509697907e-07, 'samples': 25267200, 'steps': 49349, 'loss/train': 2.1307475566864014} 02/26/2022 17:30:00 - INFO - codeparrot_training - Step 49350: {'lr': 2.2619786167490942e-07, 'samples': 25267712, 'steps': 49350, 'loss/train': 2.045980215072632} 02/26/2022 17:30:03 - INFO - codeparrot_training - Step 49351: {'lr': 2.2550250820452256e-07, 'samples': 25268224, 'steps': 49351, 'loss/train': 1.6557023525238037} 02/26/2022 17:30:09 - INFO - codeparrot_training - Step 49352: {'lr': 2.248082246887606e-07, 'samples': 25268736, 'steps': 49352, 'loss/train': 2.380180835723877} 02/26/2022 17:30:12 - INFO - codeparrot_training - Step 49353: {'lr': 2.241150111305934e-07, 'samples': 25269248, 'steps': 49353, 'loss/train': 2.416721820831299} 02/26/2022 17:30:18 - INFO - codeparrot_training - Step 49354: {'lr': 2.2342286753301856e-07, 'samples': 25269760, 'steps': 49354, 'loss/train': 2.3425889015197754} 02/26/2022 17:30:21 - INFO - codeparrot_training - Step 49355: {'lr': 2.2273179389897812e-07, 'samples': 25270272, 'steps': 49355, 'loss/train': 0.30048707127571106} 02/26/2022 17:30:29 - INFO - codeparrot_training - Step 49356: {'lr': 2.22041790231442e-07, 'samples': 25270784, 'steps': 49356, 'loss/train': 0.8866140246391296} 02/26/2022 17:30:32 - INFO - codeparrot_training - Step 49357: {'lr': 2.2135285653338e-07, 'samples': 25271296, 'steps': 49357, 'loss/train': 2.2440567016601562} 02/26/2022 17:30:38 - INFO - codeparrot_training - Step 49358: {'lr': 2.2066499280773421e-07, 'samples': 25271808, 'steps': 49358, 'loss/train': 1.9294134378433228} 02/26/2022 17:30:41 - INFO - codeparrot_training - Step 49359: {'lr': 2.19978199057419e-07, 'samples': 25272320, 'steps': 49359, 'loss/train': 1.5684888362884521} 02/26/2022 17:30:47 - INFO - codeparrot_training - Step 49360: {'lr': 2.1929247528540418e-07, 'samples': 25272832, 'steps': 49360, 'loss/train': 0.646882951259613} 02/26/2022 17:30:50 - INFO - codeparrot_training - Step 49361: {'lr': 2.1860782149463188e-07, 'samples': 25273344, 'steps': 49361, 'loss/train': 0.8152244687080383} 02/26/2022 17:30:56 - INFO - codeparrot_training - Step 49362: {'lr': 2.1792423768804415e-07, 'samples': 25273856, 'steps': 49362, 'loss/train': 1.682790756225586} 02/26/2022 17:30:59 - INFO - codeparrot_training - Step 49363: {'lr': 2.1724172386855535e-07, 'samples': 25274368, 'steps': 49363, 'loss/train': 1.901334524154663} 02/26/2022 17:31:05 - INFO - codeparrot_training - Step 49364: {'lr': 2.1656028003907978e-07, 'samples': 25274880, 'steps': 49364, 'loss/train': 1.6426023244857788} 02/26/2022 17:31:08 - INFO - codeparrot_training - Step 49365: {'lr': 2.1587990620253185e-07, 'samples': 25275392, 'steps': 49365, 'loss/train': 1.0824851989746094} 02/26/2022 17:31:15 - INFO - codeparrot_training - Step 49366: {'lr': 2.1520060236185357e-07, 'samples': 25275904, 'steps': 49366, 'loss/train': 1.577638864517212} 02/26/2022 17:31:19 - INFO - codeparrot_training - Step 49367: {'lr': 2.1452236851993157e-07, 'samples': 25276416, 'steps': 49367, 'loss/train': 0.9240710735321045} 02/26/2022 17:31:24 - INFO - codeparrot_training - Step 49368: {'lr': 2.1384520467968017e-07, 'samples': 25276928, 'steps': 49368, 'loss/train': 2.516341209411621} 02/26/2022 17:31:28 - INFO - codeparrot_training - Step 49369: {'lr': 2.131691108440137e-07, 'samples': 25277440, 'steps': 49369, 'loss/train': 1.1944133043289185} 02/26/2022 17:31:33 - INFO - codeparrot_training - Step 49370: {'lr': 2.1249408701581873e-07, 'samples': 25277952, 'steps': 49370, 'loss/train': 2.0850064754486084} 02/26/2022 17:31:37 - INFO - codeparrot_training - Step 49371: {'lr': 2.1182013319795412e-07, 'samples': 25278464, 'steps': 49371, 'loss/train': 0.7400929927825928} 02/26/2022 17:31:43 - INFO - codeparrot_training - Step 49372: {'lr': 2.1114724939336195e-07, 'samples': 25278976, 'steps': 49372, 'loss/train': 2.0787765979766846} 02/26/2022 17:31:46 - INFO - codeparrot_training - Step 49373: {'lr': 2.1047543560487327e-07, 'samples': 25279488, 'steps': 49373, 'loss/train': 2.363260507583618} 02/26/2022 17:31:52 - INFO - codeparrot_training - Step 49374: {'lr': 2.098046918354024e-07, 'samples': 25280000, 'steps': 49374, 'loss/train': 1.6996179819107056} 02/26/2022 17:31:55 - INFO - codeparrot_training - Step 49375: {'lr': 2.091350180878082e-07, 'samples': 25280512, 'steps': 49375, 'loss/train': 2.3067574501037598} 02/26/2022 17:32:01 - INFO - codeparrot_training - Step 49376: {'lr': 2.0846641436497726e-07, 'samples': 25281024, 'steps': 49376, 'loss/train': 0.6636629700660706} 02/26/2022 17:32:04 - INFO - codeparrot_training - Step 49377: {'lr': 2.0779888066974062e-07, 'samples': 25281536, 'steps': 49377, 'loss/train': 2.244915723800659} 02/26/2022 17:32:11 - INFO - codeparrot_training - Step 49378: {'lr': 2.0713241700498485e-07, 'samples': 25282048, 'steps': 49378, 'loss/train': 2.468876838684082} 02/26/2022 17:32:15 - INFO - codeparrot_training - Step 49379: {'lr': 2.0646702337354108e-07, 'samples': 25282560, 'steps': 49379, 'loss/train': 1.5205780267715454} 02/26/2022 17:32:20 - INFO - codeparrot_training - Step 49380: {'lr': 2.058026997782958e-07, 'samples': 25283072, 'steps': 49380, 'loss/train': 2.1277313232421875} 02/26/2022 17:32:24 - INFO - codeparrot_training - Step 49381: {'lr': 2.0513944622205239e-07, 'samples': 25283584, 'steps': 49381, 'loss/train': 0.9164019227027893} 02/26/2022 17:32:29 - INFO - codeparrot_training - Step 49382: {'lr': 2.0447726270769739e-07, 'samples': 25284096, 'steps': 49382, 'loss/train': 1.305641770362854} 02/26/2022 17:32:33 - INFO - codeparrot_training - Step 49383: {'lr': 2.0381614923803417e-07, 'samples': 25284608, 'steps': 49383, 'loss/train': 1.5540956258773804} 02/26/2022 17:32:38 - INFO - codeparrot_training - Step 49384: {'lr': 2.0315610581589372e-07, 'samples': 25285120, 'steps': 49384, 'loss/train': 0.8872753977775574} 02/26/2022 17:32:42 - INFO - codeparrot_training - Step 49385: {'lr': 2.0249713244410718e-07, 'samples': 25285632, 'steps': 49385, 'loss/train': 1.963981032371521} 02/26/2022 17:32:47 - INFO - codeparrot_training - Step 49386: {'lr': 2.0183922912553332e-07, 'samples': 25286144, 'steps': 49386, 'loss/train': 2.091745376586914} 02/26/2022 17:32:51 - INFO - codeparrot_training - Step 49387: {'lr': 2.0118239586292e-07, 'samples': 25286656, 'steps': 49387, 'loss/train': 2.2686331272125244} 02/26/2022 17:32:56 - INFO - codeparrot_training - Step 49388: {'lr': 2.0052663265915373e-07, 'samples': 25287168, 'steps': 49388, 'loss/train': 0.8810884952545166} 02/26/2022 17:33:00 - INFO - codeparrot_training - Step 49389: {'lr': 1.9987193951698235e-07, 'samples': 25287680, 'steps': 49389, 'loss/train': 1.7230876684188843} 02/26/2022 17:33:05 - INFO - codeparrot_training - Step 49390: {'lr': 1.9921831643926468e-07, 'samples': 25288192, 'steps': 49390, 'loss/train': 2.2715725898742676} 02/26/2022 17:33:09 - INFO - codeparrot_training - Step 49391: {'lr': 1.985657634287763e-07, 'samples': 25288704, 'steps': 49391, 'loss/train': 1.448055386543274} 02/26/2022 17:33:16 - INFO - codeparrot_training - Step 49392: {'lr': 1.9791428048829275e-07, 'samples': 25289216, 'steps': 49392, 'loss/train': 1.7186641693115234} 02/26/2022 17:33:20 - INFO - codeparrot_training - Step 49393: {'lr': 1.9726386762064508e-07, 'samples': 25289728, 'steps': 49393, 'loss/train': 1.3137915134429932} 02/26/2022 17:33:25 - INFO - codeparrot_training - Step 49394: {'lr': 1.966145248285811e-07, 'samples': 25290240, 'steps': 49394, 'loss/train': 0.8090012669563293} 02/26/2022 17:33:28 - INFO - codeparrot_training - Step 49395: {'lr': 1.9596625211490416e-07, 'samples': 25290752, 'steps': 49395, 'loss/train': 1.6962679624557495} 02/26/2022 17:33:34 - INFO - codeparrot_training - Step 49396: {'lr': 1.9531904948236202e-07, 'samples': 25291264, 'steps': 49396, 'loss/train': 1.3675588369369507} 02/26/2022 17:33:37 - INFO - codeparrot_training - Step 49397: {'lr': 1.9467291693378574e-07, 'samples': 25291776, 'steps': 49397, 'loss/train': 0.5769745707511902} 02/26/2022 17:33:43 - INFO - codeparrot_training - Step 49398: {'lr': 1.9402785447189542e-07, 'samples': 25292288, 'steps': 49398, 'loss/train': 2.389784812927246} 02/26/2022 17:33:47 - INFO - codeparrot_training - Step 49399: {'lr': 1.9338386209946656e-07, 'samples': 25292800, 'steps': 49399, 'loss/train': 2.068979263305664} 02/26/2022 17:33:52 - INFO - codeparrot_training - Step 49400: {'lr': 1.9274093981927476e-07, 'samples': 25293312, 'steps': 49400, 'loss/train': 2.364450693130493} 02/26/2022 17:33:55 - INFO - codeparrot_training - Step 49401: {'lr': 1.9209908763404006e-07, 'samples': 25293824, 'steps': 49401, 'loss/train': 1.3425612449645996} 02/26/2022 17:34:03 - INFO - codeparrot_training - Step 49402: {'lr': 1.91458305546538e-07, 'samples': 25294336, 'steps': 49402, 'loss/train': 1.4773885011672974} 02/26/2022 17:34:06 - INFO - codeparrot_training - Step 49403: {'lr': 1.908185935594886e-07, 'samples': 25294848, 'steps': 49403, 'loss/train': 1.7944095134735107} 02/26/2022 17:34:12 - INFO - codeparrot_training - Step 49404: {'lr': 1.901799516756675e-07, 'samples': 25295360, 'steps': 49404, 'loss/train': 2.3042733669281006} 02/26/2022 17:34:15 - INFO - codeparrot_training - Step 49405: {'lr': 1.8954237989779467e-07, 'samples': 25295872, 'steps': 49405, 'loss/train': 1.5094069242477417} 02/26/2022 17:34:21 - INFO - codeparrot_training - Step 49406: {'lr': 1.8890587822859018e-07, 'samples': 25296384, 'steps': 49406, 'loss/train': 1.779155969619751} 02/26/2022 17:34:24 - INFO - codeparrot_training - Step 49407: {'lr': 1.8827044667077408e-07, 'samples': 25296896, 'steps': 49407, 'loss/train': 2.8769690990448} 02/26/2022 17:34:30 - INFO - codeparrot_training - Step 49408: {'lr': 1.8763608522709418e-07, 'samples': 25297408, 'steps': 49408, 'loss/train': 1.1164543628692627} 02/26/2022 17:34:33 - INFO - codeparrot_training - Step 49409: {'lr': 1.8700279390027052e-07, 'samples': 25297920, 'steps': 49409, 'loss/train': 1.7498164176940918} 02/26/2022 17:34:39 - INFO - codeparrot_training - Step 49410: {'lr': 1.863705726929954e-07, 'samples': 25298432, 'steps': 49410, 'loss/train': 0.7097488641738892} 02/26/2022 17:34:42 - INFO - codeparrot_training - Step 49411: {'lr': 1.8573942160798885e-07, 'samples': 25298944, 'steps': 49411, 'loss/train': 1.2397820949554443} 02/26/2022 17:34:50 - INFO - codeparrot_training - Step 49412: {'lr': 1.8510934064791542e-07, 'samples': 25299456, 'steps': 49412, 'loss/train': 2.2480454444885254} 02/26/2022 17:34:54 - INFO - codeparrot_training - Step 49413: {'lr': 1.8448032981555064e-07, 'samples': 25299968, 'steps': 49413, 'loss/train': 2.2325031757354736} 02/26/2022 17:34:57 - INFO - codeparrot_training - Step 49414: {'lr': 1.838523891135313e-07, 'samples': 25300480, 'steps': 49414, 'loss/train': 1.7891182899475098} 02/26/2022 17:35:03 - INFO - codeparrot_training - Step 49415: {'lr': 1.8322551854454973e-07, 'samples': 25300992, 'steps': 49415, 'loss/train': 1.816938877105713} 02/26/2022 17:35:06 - INFO - codeparrot_training - Step 49416: {'lr': 1.8259971811132593e-07, 'samples': 25301504, 'steps': 49416, 'loss/train': 1.3228381872177124} 02/26/2022 17:35:11 - INFO - codeparrot_training - Step 49417: {'lr': 1.8197498781652443e-07, 'samples': 25302016, 'steps': 49417, 'loss/train': 1.1276450157165527} 02/26/2022 17:35:15 - INFO - codeparrot_training - Step 49418: {'lr': 1.8135132766278207e-07, 'samples': 25302528, 'steps': 49418, 'loss/train': 1.9976853132247925} 02/26/2022 17:35:20 - INFO - codeparrot_training - Step 49419: {'lr': 1.8072873765281883e-07, 'samples': 25303040, 'steps': 49419, 'loss/train': 1.4852638244628906} 02/26/2022 17:35:24 - INFO - codeparrot_training - Step 49420: {'lr': 1.801072177892993e-07, 'samples': 25303552, 'steps': 49420, 'loss/train': 1.1721638441085815} 02/26/2022 17:35:29 - INFO - codeparrot_training - Step 49421: {'lr': 1.794867680748602e-07, 'samples': 25304064, 'steps': 49421, 'loss/train': 1.2996807098388672} 02/26/2022 17:35:33 - INFO - codeparrot_training - Step 49422: {'lr': 1.7886738851219388e-07, 'samples': 25304576, 'steps': 49422, 'loss/train': 1.7551120519638062} 02/26/2022 17:35:38 - INFO - codeparrot_training - Step 49423: {'lr': 1.782490791039093e-07, 'samples': 25305088, 'steps': 49423, 'loss/train': 1.8207831382751465} 02/26/2022 17:35:42 - INFO - codeparrot_training - Step 49424: {'lr': 1.7763183985269881e-07, 'samples': 25305600, 'steps': 49424, 'loss/train': 1.570170521736145} 02/26/2022 17:35:49 - INFO - codeparrot_training - Step 49425: {'lr': 1.770156707611714e-07, 'samples': 25306112, 'steps': 49425, 'loss/train': 1.4419302940368652} 02/26/2022 17:35:53 - INFO - codeparrot_training - Step 49426: {'lr': 1.7640057183196389e-07, 'samples': 25306624, 'steps': 49426, 'loss/train': 1.6194907426834106} 02/26/2022 17:35:58 - INFO - codeparrot_training - Step 49427: {'lr': 1.7578654306776853e-07, 'samples': 25307136, 'steps': 49427, 'loss/train': 0.6367307901382446} 02/26/2022 17:36:02 - INFO - codeparrot_training - Step 49428: {'lr': 1.7517358447113884e-07, 'samples': 25307648, 'steps': 49428, 'loss/train': 1.4055144786834717} 02/26/2022 17:36:07 - INFO - codeparrot_training - Step 49429: {'lr': 1.7456169604476713e-07, 'samples': 25308160, 'steps': 49429, 'loss/train': 0.07487159967422485} 02/26/2022 17:36:11 - INFO - codeparrot_training - Step 49430: {'lr': 1.7395087779123465e-07, 'samples': 25308672, 'steps': 49430, 'loss/train': 1.3172026872634888} 02/26/2022 17:36:16 - INFO - codeparrot_training - Step 49431: {'lr': 1.7334112971315042e-07, 'samples': 25309184, 'steps': 49431, 'loss/train': 1.3774311542510986} 02/26/2022 17:36:20 - INFO - codeparrot_training - Step 49432: {'lr': 1.7273245181315122e-07, 'samples': 25309696, 'steps': 49432, 'loss/train': 1.0256197452545166} 02/26/2022 17:36:25 - INFO - codeparrot_training - Step 49433: {'lr': 1.721248440938461e-07, 'samples': 25310208, 'steps': 49433, 'loss/train': 1.517964482307434} 02/26/2022 17:36:28 - INFO - codeparrot_training - Step 49434: {'lr': 1.7151830655784405e-07, 'samples': 25310720, 'steps': 49434, 'loss/train': 0.5323280096054077} 02/26/2022 17:36:36 - INFO - codeparrot_training - Step 49435: {'lr': 1.7091283920769863e-07, 'samples': 25311232, 'steps': 49435, 'loss/train': 2.2097764015197754} 02/26/2022 17:36:39 - INFO - codeparrot_training - Step 49436: {'lr': 1.7030844204604657e-07, 'samples': 25311744, 'steps': 49436, 'loss/train': 2.1547555923461914} 02/26/2022 17:36:45 - INFO - codeparrot_training - Step 49437: {'lr': 1.6970511507546914e-07, 'samples': 25312256, 'steps': 49437, 'loss/train': 0.35932597517967224} 02/26/2022 17:36:50 - INFO - codeparrot_training - Step 49438: {'lr': 1.691028582985199e-07, 'samples': 25312768, 'steps': 49438, 'loss/train': 1.6998181343078613} 02/26/2022 17:36:54 - INFO - codeparrot_training - Step 49439: {'lr': 1.685016717178356e-07, 'samples': 25313280, 'steps': 49439, 'loss/train': 2.5360960960388184} 02/26/2022 17:36:59 - INFO - codeparrot_training - Step 49440: {'lr': 1.67901555335942e-07, 'samples': 25313792, 'steps': 49440, 'loss/train': 2.00775408744812} 02/26/2022 17:37:03 - INFO - codeparrot_training - Step 49441: {'lr': 1.6730250915542035e-07, 'samples': 25314304, 'steps': 49441, 'loss/train': 0.01932586543262005} 02/26/2022 17:37:08 - INFO - codeparrot_training - Step 49442: {'lr': 1.6670453317885192e-07, 'samples': 25314816, 'steps': 49442, 'loss/train': 2.3165156841278076} 02/26/2022 17:37:12 - INFO - codeparrot_training - Step 49443: {'lr': 1.6610762740879027e-07, 'samples': 25315328, 'steps': 49443, 'loss/train': 2.51115083694458} 02/26/2022 17:37:17 - INFO - codeparrot_training - Step 49444: {'lr': 1.6551179184781663e-07, 'samples': 25315840, 'steps': 49444, 'loss/train': 2.125347137451172} 02/26/2022 17:37:21 - INFO - codeparrot_training - Step 49445: {'lr': 1.64917026498429e-07, 'samples': 25316352, 'steps': 49445, 'loss/train': 1.4960134029388428} 02/26/2022 17:37:26 - INFO - codeparrot_training - Step 49446: {'lr': 1.6432333136320865e-07, 'samples': 25316864, 'steps': 49446, 'loss/train': 0.8772303462028503} 02/26/2022 17:37:30 - INFO - codeparrot_training - Step 49447: {'lr': 1.637307064447091e-07, 'samples': 25317376, 'steps': 49447, 'loss/train': 1.2380704879760742} 02/26/2022 17:37:37 - INFO - codeparrot_training - Step 49448: {'lr': 1.6313915174542836e-07, 'samples': 25317888, 'steps': 49448, 'loss/train': 1.084328532218933} 02/26/2022 17:37:40 - INFO - codeparrot_training - Step 49449: {'lr': 1.625486672679477e-07, 'samples': 25318400, 'steps': 49449, 'loss/train': 2.2731661796569824} 02/26/2022 17:37:46 - INFO - codeparrot_training - Step 49450: {'lr': 1.6195925301479287e-07, 'samples': 25318912, 'steps': 49450, 'loss/train': 2.20371150970459} 02/26/2022 17:37:49 - INFO - codeparrot_training - Step 49451: {'lr': 1.613709089884341e-07, 'samples': 25319424, 'steps': 49451, 'loss/train': 1.5246758460998535} 02/26/2022 17:37:55 - INFO - codeparrot_training - Step 49452: {'lr': 1.6078363519145268e-07, 'samples': 25319936, 'steps': 49452, 'loss/train': 2.318956136703491} 02/26/2022 17:37:58 - INFO - codeparrot_training - Step 49453: {'lr': 1.601974316263466e-07, 'samples': 25320448, 'steps': 49453, 'loss/train': 1.7194854021072388} 02/26/2022 17:38:04 - INFO - codeparrot_training - Step 49454: {'lr': 1.5961229829561385e-07, 'samples': 25320960, 'steps': 49454, 'loss/train': 1.0852497816085815} 02/26/2022 17:38:07 - INFO - codeparrot_training - Step 49455: {'lr': 1.5902823520175247e-07, 'samples': 25321472, 'steps': 49455, 'loss/train': 2.0783467292785645} 02/26/2022 17:38:13 - INFO - codeparrot_training - Step 49456: {'lr': 1.5844524234728818e-07, 'samples': 25321984, 'steps': 49456, 'loss/train': 0.9691670536994934} 02/26/2022 17:38:16 - INFO - codeparrot_training - Step 49457: {'lr': 1.57863319734719e-07, 'samples': 25322496, 'steps': 49457, 'loss/train': 2.3410816192626953} 02/26/2022 17:38:20 - INFO - codeparrot_training - Step 49458: {'lr': 1.5728246736654294e-07, 'samples': 25323008, 'steps': 49458, 'loss/train': 2.033514976501465} 02/26/2022 17:38:26 - INFO - codeparrot_training - Step 49459: {'lr': 1.5670268524520248e-07, 'samples': 25323520, 'steps': 49459, 'loss/train': 2.0586495399475098} 02/26/2022 17:38:29 - INFO - codeparrot_training - Step 49460: {'lr': 1.5612397337325114e-07, 'samples': 25324032, 'steps': 49460, 'loss/train': 0.4165288507938385} 02/26/2022 17:38:36 - INFO - codeparrot_training - Step 49461: {'lr': 1.5554633175310363e-07, 'samples': 25324544, 'steps': 49461, 'loss/train': 1.7726829051971436} 02/26/2022 17:38:40 - INFO - codeparrot_training - Step 49462: {'lr': 1.5496976038725795e-07, 'samples': 25325056, 'steps': 49462, 'loss/train': 1.6395699977874756} 02/26/2022 17:38:45 - INFO - codeparrot_training - Step 49463: {'lr': 1.5439425927821217e-07, 'samples': 25325568, 'steps': 49463, 'loss/train': 1.2925454378128052} 02/26/2022 17:38:49 - INFO - codeparrot_training - Step 49464: {'lr': 1.5381982842838093e-07, 'samples': 25326080, 'steps': 49464, 'loss/train': 1.6523593664169312} 02/26/2022 17:38:54 - INFO - codeparrot_training - Step 49465: {'lr': 1.5324646784026232e-07, 'samples': 25326592, 'steps': 49465, 'loss/train': 0.07220093905925751} 02/26/2022 17:38:58 - INFO - codeparrot_training - Step 49466: {'lr': 1.5267417751629876e-07, 'samples': 25327104, 'steps': 49466, 'loss/train': 1.0161285400390625} 02/26/2022 17:39:03 - INFO - codeparrot_training - Step 49467: {'lr': 1.5210295745893277e-07, 'samples': 25327616, 'steps': 49467, 'loss/train': 1.5583438873291016} 02/26/2022 17:39:07 - INFO - codeparrot_training - Step 49468: {'lr': 1.515328076706346e-07, 'samples': 25328128, 'steps': 49468, 'loss/train': 1.0967252254486084} 02/26/2022 17:39:13 - INFO - codeparrot_training - Step 49469: {'lr': 1.5096372815384675e-07, 'samples': 25328640, 'steps': 49469, 'loss/train': 1.5874531269073486} 02/26/2022 17:39:16 - INFO - codeparrot_training - Step 49470: {'lr': 1.503957189109839e-07, 'samples': 25329152, 'steps': 49470, 'loss/train': 1.2299761772155762} 02/26/2022 17:39:23 - INFO - codeparrot_training - Step 49471: {'lr': 1.4982877994448862e-07, 'samples': 25329664, 'steps': 49471, 'loss/train': 0.9899556636810303} 02/26/2022 17:39:27 - INFO - codeparrot_training - Step 49472: {'lr': 1.492629112567756e-07, 'samples': 25330176, 'steps': 49472, 'loss/train': 2.01727032661438} 02/26/2022 17:39:32 - INFO - codeparrot_training - Step 49473: {'lr': 1.4869811285028734e-07, 'samples': 25330688, 'steps': 49473, 'loss/train': 2.082319974899292} 02/26/2022 17:39:36 - INFO - codeparrot_training - Step 49474: {'lr': 1.4813438472746632e-07, 'samples': 25331200, 'steps': 49474, 'loss/train': 2.282454013824463} 02/26/2022 17:39:41 - INFO - codeparrot_training - Step 49475: {'lr': 1.4757172689067177e-07, 'samples': 25331712, 'steps': 49475, 'loss/train': 2.0538759231567383} 02/26/2022 17:39:45 - INFO - codeparrot_training - Step 49476: {'lr': 1.470101393423462e-07, 'samples': 25332224, 'steps': 49476, 'loss/train': 1.5414130687713623} 02/26/2022 17:39:50 - INFO - codeparrot_training - Step 49477: {'lr': 1.464496220849043e-07, 'samples': 25332736, 'steps': 49477, 'loss/train': 1.9200726747512817} 02/26/2022 17:39:54 - INFO - codeparrot_training - Step 49478: {'lr': 1.458901751207331e-07, 'samples': 25333248, 'steps': 49478, 'loss/train': 1.1467262506484985} 02/26/2022 17:39:59 - INFO - codeparrot_training - Step 49479: {'lr': 1.4533179845221956e-07, 'samples': 25333760, 'steps': 49479, 'loss/train': 1.0757455825805664} 02/26/2022 17:40:03 - INFO - codeparrot_training - Step 49480: {'lr': 1.447744920817784e-07, 'samples': 25334272, 'steps': 49480, 'loss/train': 1.1312575340270996} 02/26/2022 17:40:10 - INFO - codeparrot_training - Step 49481: {'lr': 1.442182560117966e-07, 'samples': 25334784, 'steps': 49481, 'loss/train': 1.7154914140701294} 02/26/2022 17:40:14 - INFO - codeparrot_training - Step 49482: {'lr': 1.4366309024463343e-07, 'samples': 25335296, 'steps': 49482, 'loss/train': 0.4336872696876526} 02/26/2022 17:40:19 - INFO - codeparrot_training - Step 49483: {'lr': 1.431089947826758e-07, 'samples': 25335808, 'steps': 49483, 'loss/train': 1.353283405303955} 02/26/2022 17:40:23 - INFO - codeparrot_training - Step 49484: {'lr': 1.42555969628283e-07, 'samples': 25336320, 'steps': 49484, 'loss/train': 1.3421893119812012} 02/26/2022 17:40:29 - INFO - codeparrot_training - Step 49485: {'lr': 1.420040147838697e-07, 'samples': 25336832, 'steps': 49485, 'loss/train': 2.022498846054077} 02/26/2022 17:40:33 - INFO - codeparrot_training - Step 49486: {'lr': 1.4145313025176743e-07, 'samples': 25337344, 'steps': 49486, 'loss/train': 2.103182315826416} 02/26/2022 17:40:38 - INFO - codeparrot_training - Step 49487: {'lr': 1.4090331603433536e-07, 'samples': 25337856, 'steps': 49487, 'loss/train': 1.5370337963104248} 02/26/2022 17:40:42 - INFO - codeparrot_training - Step 49488: {'lr': 1.4035457213393277e-07, 'samples': 25338368, 'steps': 49488, 'loss/train': 2.414919853210449} 02/26/2022 17:40:47 - INFO - codeparrot_training - Step 49489: {'lr': 1.3980689855291884e-07, 'samples': 25338880, 'steps': 49489, 'loss/train': 2.0758466720581055} 02/26/2022 17:40:51 - INFO - codeparrot_training - Step 49490: {'lr': 1.3926029529362506e-07, 'samples': 25339392, 'steps': 49490, 'loss/train': 2.3458642959594727} 02/26/2022 17:40:56 - INFO - codeparrot_training - Step 49491: {'lr': 1.3871476235841063e-07, 'samples': 25339904, 'steps': 49491, 'loss/train': 2.408580780029297} 02/26/2022 17:41:00 - INFO - codeparrot_training - Step 49492: {'lr': 1.381702997495793e-07, 'samples': 25340416, 'steps': 49492, 'loss/train': 1.6963386535644531} 02/26/2022 17:41:05 - INFO - codeparrot_training - Step 49493: {'lr': 1.37626907469518e-07, 'samples': 25340928, 'steps': 49493, 'loss/train': 0.04213282838463783} 02/26/2022 17:41:09 - INFO - codeparrot_training - Step 49494: {'lr': 1.3708458552050274e-07, 'samples': 25341440, 'steps': 49494, 'loss/train': 1.2074458599090576} 02/26/2022 17:41:16 - INFO - codeparrot_training - Step 49495: {'lr': 1.3654333390486494e-07, 'samples': 25341952, 'steps': 49495, 'loss/train': 0.6862615942955017} 02/26/2022 17:41:19 - INFO - codeparrot_training - Step 49496: {'lr': 1.3600315262496388e-07, 'samples': 25342464, 'steps': 49496, 'loss/train': 1.3673888444900513} 02/26/2022 17:41:25 - INFO - codeparrot_training - Step 49497: {'lr': 1.3546404168307547e-07, 'samples': 25342976, 'steps': 49497, 'loss/train': 1.2132560014724731} 02/26/2022 17:41:29 - INFO - codeparrot_training - Step 49498: {'lr': 1.3492600108150345e-07, 'samples': 25343488, 'steps': 49498, 'loss/train': 2.350691556930542} 02/26/2022 17:41:34 - INFO - codeparrot_training - Step 49499: {'lr': 1.3438903082257924e-07, 'samples': 25344000, 'steps': 49499, 'loss/train': 1.8677647113800049} 02/26/2022 17:41:37 - INFO - codeparrot_training - Step 49500: {'lr': 1.3385313090857886e-07, 'samples': 25344512, 'steps': 49500, 'loss/train': 1.8952529430389404} 02/26/2022 17:41:43 - INFO - codeparrot_training - Step 49501: {'lr': 1.3331830134183376e-07, 'samples': 25345024, 'steps': 49501, 'loss/train': 1.6151928901672363} 02/26/2022 17:41:47 - INFO - codeparrot_training - Step 49502: {'lr': 1.3278454212459213e-07, 'samples': 25345536, 'steps': 49502, 'loss/train': 1.4087148904800415} 02/26/2022 17:41:52 - INFO - codeparrot_training - Step 49503: {'lr': 1.3225185325918542e-07, 'samples': 25346048, 'steps': 49503, 'loss/train': 1.5115303993225098} 02/26/2022 17:41:56 - INFO - codeparrot_training - Step 49504: {'lr': 1.3172023474783413e-07, 'samples': 25346560, 'steps': 49504, 'loss/train': 1.823445439338684} 02/26/2022 17:42:03 - INFO - codeparrot_training - Step 49505: {'lr': 1.3118968659286967e-07, 'samples': 25347072, 'steps': 49505, 'loss/train': 2.0007190704345703} 02/26/2022 17:42:07 - INFO - codeparrot_training - Step 49506: {'lr': 1.3066020879654027e-07, 'samples': 25347584, 'steps': 49506, 'loss/train': 1.6916680335998535} 02/26/2022 17:42:12 - INFO - codeparrot_training - Step 49507: {'lr': 1.301318013611219e-07, 'samples': 25348096, 'steps': 49507, 'loss/train': 2.3506176471710205} 02/26/2022 17:42:16 - INFO - codeparrot_training - Step 49508: {'lr': 1.296044642888905e-07, 'samples': 25348608, 'steps': 49508, 'loss/train': 2.174691677093506} 02/26/2022 17:42:21 - INFO - codeparrot_training - Step 49509: {'lr': 1.2907819758206652e-07, 'samples': 25349120, 'steps': 49509, 'loss/train': 2.3295016288757324} 02/26/2022 17:42:25 - INFO - codeparrot_training - Step 49510: {'lr': 1.2855300124295367e-07, 'samples': 25349632, 'steps': 49510, 'loss/train': 1.8093880414962769} 02/26/2022 17:42:31 - INFO - codeparrot_training - Step 49511: {'lr': 1.280288752737724e-07, 'samples': 25350144, 'steps': 49511, 'loss/train': 2.0852577686309814} 02/26/2022 17:42:34 - INFO - codeparrot_training - Step 49512: {'lr': 1.2750581967674312e-07, 'samples': 25350656, 'steps': 49512, 'loss/train': 2.1863291263580322} 02/26/2022 17:42:37 - INFO - codeparrot_training - Step 49513: {'lr': 1.2698383445416962e-07, 'samples': 25351168, 'steps': 49513, 'loss/train': 0.9368155598640442} 02/26/2022 17:42:43 - INFO - codeparrot_training - Step 49514: {'lr': 1.2646291960824452e-07, 'samples': 25351680, 'steps': 49514, 'loss/train': 1.6686816215515137} 02/26/2022 17:42:46 - INFO - codeparrot_training - Step 49515: {'lr': 1.259430751411883e-07, 'samples': 25352192, 'steps': 49515, 'loss/train': 0.6560873985290527} 02/26/2022 17:42:52 - INFO - codeparrot_training - Step 49516: {'lr': 1.2542430105527693e-07, 'samples': 25352704, 'steps': 49516, 'loss/train': 0.5743892192840576} 02/26/2022 17:42:55 - INFO - codeparrot_training - Step 49517: {'lr': 1.2490659735267529e-07, 'samples': 25353216, 'steps': 49517, 'loss/train': 0.031562063843011856} 02/26/2022 17:43:03 - INFO - codeparrot_training - Step 49518: {'lr': 1.243899640356594e-07, 'samples': 25353728, 'steps': 49518, 'loss/train': 1.756022572517395} 02/26/2022 17:43:06 - INFO - codeparrot_training - Step 49519: {'lr': 1.2387440110639415e-07, 'samples': 25354240, 'steps': 49519, 'loss/train': 1.9736648797988892} 02/26/2022 17:43:12 - INFO - codeparrot_training - Step 49520: {'lr': 1.233599085671e-07, 'samples': 25354752, 'steps': 49520, 'loss/train': 1.8910934925079346} 02/26/2022 17:43:15 - INFO - codeparrot_training - Step 49521: {'lr': 1.228464864199974e-07, 'samples': 25355264, 'steps': 49521, 'loss/train': 1.1839945316314697} 02/26/2022 17:43:21 - INFO - codeparrot_training - Step 49522: {'lr': 1.2233413466727905e-07, 'samples': 25355776, 'steps': 49522, 'loss/train': 1.8331290483474731} 02/26/2022 17:43:24 - INFO - codeparrot_training - Step 49523: {'lr': 1.2182285331113764e-07, 'samples': 25356288, 'steps': 49523, 'loss/train': 1.3405741453170776} 02/26/2022 17:43:30 - INFO - codeparrot_training - Step 49524: {'lr': 1.2131264235376582e-07, 'samples': 25356800, 'steps': 49524, 'loss/train': 0.5912362337112427} 02/26/2022 17:43:33 - INFO - codeparrot_training - Step 49525: {'lr': 1.2080350179732857e-07, 'samples': 25357312, 'steps': 49525, 'loss/train': 2.036501169204712} 02/26/2022 17:43:39 - INFO - codeparrot_training - Step 49526: {'lr': 1.2029543164404633e-07, 'samples': 25357824, 'steps': 49526, 'loss/train': 0.03968025743961334} 02/26/2022 17:43:42 - INFO - codeparrot_training - Step 49527: {'lr': 1.1978843189605626e-07, 'samples': 25358336, 'steps': 49527, 'loss/train': 1.017698049545288} 02/26/2022 17:43:50 - INFO - codeparrot_training - Step 49528: {'lr': 1.1928250255555106e-07, 'samples': 25358848, 'steps': 49528, 'loss/train': 2.15105938911438} 02/26/2022 17:43:53 - INFO - codeparrot_training - Step 49529: {'lr': 1.1877764362466792e-07, 'samples': 25359360, 'steps': 49529, 'loss/train': 2.18550181388855} 02/26/2022 17:43:59 - INFO - codeparrot_training - Step 49530: {'lr': 1.1827385510562727e-07, 'samples': 25359872, 'steps': 49530, 'loss/train': 1.3240455389022827} 02/26/2022 17:44:02 - INFO - codeparrot_training - Step 49531: {'lr': 1.1777113700053854e-07, 'samples': 25360384, 'steps': 49531, 'loss/train': 1.775473713874817} 02/26/2022 17:44:08 - INFO - codeparrot_training - Step 49532: {'lr': 1.1726948931159443e-07, 'samples': 25360896, 'steps': 49532, 'loss/train': 1.051789402961731} 02/26/2022 17:44:11 - INFO - codeparrot_training - Step 49533: {'lr': 1.1676891204090434e-07, 'samples': 25361408, 'steps': 49533, 'loss/train': 1.8594964742660522} 02/26/2022 17:44:17 - INFO - codeparrot_training - Step 49534: {'lr': 1.1626940519063323e-07, 'samples': 25361920, 'steps': 49534, 'loss/train': 2.3752570152282715} 02/26/2022 17:44:20 - INFO - codeparrot_training - Step 49535: {'lr': 1.1577096876289051e-07, 'samples': 25362432, 'steps': 49535, 'loss/train': 1.1005746126174927} 02/26/2022 17:44:26 - INFO - codeparrot_training - Step 49536: {'lr': 1.1527360275986887e-07, 'samples': 25362944, 'steps': 49536, 'loss/train': 2.6564674377441406} 02/26/2022 17:44:29 - INFO - codeparrot_training - Step 49537: {'lr': 1.1477730718367774e-07, 'samples': 25363456, 'steps': 49537, 'loss/train': 2.443544387817383} 02/26/2022 17:44:35 - INFO - codeparrot_training - Step 49538: {'lr': 1.1428208203639878e-07, 'samples': 25363968, 'steps': 49538, 'loss/train': 2.4494733810424805} 02/26/2022 17:44:38 - INFO - codeparrot_training - Step 49539: {'lr': 1.1378792732019694e-07, 'samples': 25364480, 'steps': 49539, 'loss/train': 2.5757603645324707} 02/26/2022 17:44:44 - INFO - codeparrot_training - Step 49540: {'lr': 1.1329484303720939e-07, 'samples': 25364992, 'steps': 49540, 'loss/train': 1.4004231691360474} 02/26/2022 17:44:47 - INFO - codeparrot_training - Step 49541: {'lr': 1.1280282918949003e-07, 'samples': 25365504, 'steps': 49541, 'loss/train': 1.7510607242584229} 02/26/2022 17:44:54 - INFO - codeparrot_training - Step 49542: {'lr': 1.1231188577917607e-07, 'samples': 25366016, 'steps': 49542, 'loss/train': 1.7925915718078613} 02/26/2022 17:44:58 - INFO - codeparrot_training - Step 49543: {'lr': 1.1182201280837689e-07, 'samples': 25366528, 'steps': 49543, 'loss/train': 1.5280436277389526} 02/26/2022 17:45:03 - INFO - codeparrot_training - Step 49544: {'lr': 1.113332102791742e-07, 'samples': 25367040, 'steps': 49544, 'loss/train': 0.8311648368835449} 02/26/2022 17:45:07 - INFO - codeparrot_training - Step 49545: {'lr': 1.1084547819370516e-07, 'samples': 25367552, 'steps': 49545, 'loss/train': 2.5234060287475586} 02/26/2022 17:45:12 - INFO - codeparrot_training - Step 49546: {'lr': 1.1035881655399593e-07, 'samples': 25368064, 'steps': 49546, 'loss/train': 1.327113151550293} 02/26/2022 17:45:16 - INFO - codeparrot_training - Step 49547: {'lr': 1.0987322536215594e-07, 'samples': 25368576, 'steps': 49547, 'loss/train': 2.671031951904297} 02/26/2022 17:45:22 - INFO - codeparrot_training - Step 49548: {'lr': 1.0938870462026685e-07, 'samples': 25369088, 'steps': 49548, 'loss/train': 2.4321107864379883} 02/26/2022 17:45:25 - INFO - codeparrot_training - Step 49549: {'lr': 1.0890525433043808e-07, 'samples': 25369600, 'steps': 49549, 'loss/train': 0.5166988372802734} 02/26/2022 17:45:31 - INFO - codeparrot_training - Step 49550: {'lr': 1.0842287449469579e-07, 'samples': 25370112, 'steps': 49550, 'loss/train': 1.7249177694320679} 02/26/2022 17:45:34 - INFO - codeparrot_training - Step 49551: {'lr': 1.079415651150939e-07, 'samples': 25370624, 'steps': 49551, 'loss/train': 2.3032150268554688} 02/26/2022 17:45:41 - INFO - codeparrot_training - Step 49552: {'lr': 1.0746132619374182e-07, 'samples': 25371136, 'steps': 49552, 'loss/train': 1.8407658338546753} 02/26/2022 17:45:45 - INFO - codeparrot_training - Step 49553: {'lr': 1.0698215773269349e-07, 'samples': 25371648, 'steps': 49553, 'loss/train': 2.513997793197632} 02/26/2022 17:45:50 - INFO - codeparrot_training - Step 49554: {'lr': 1.0650405973397504e-07, 'samples': 25372160, 'steps': 49554, 'loss/train': 1.606418490409851} 02/26/2022 17:45:54 - INFO - codeparrot_training - Step 49555: {'lr': 1.0602703219961263e-07, 'samples': 25372672, 'steps': 49555, 'loss/train': 1.2359482049942017} 02/26/2022 17:45:59 - INFO - codeparrot_training - Step 49556: {'lr': 1.0555107513171569e-07, 'samples': 25373184, 'steps': 49556, 'loss/train': 1.5361590385437012} 02/26/2022 17:46:03 - INFO - codeparrot_training - Step 49557: {'lr': 1.0507618853225487e-07, 'samples': 25373696, 'steps': 49557, 'loss/train': 0.9107977747917175} 02/26/2022 17:46:08 - INFO - codeparrot_training - Step 49558: {'lr': 1.0460237240333959e-07, 'samples': 25374208, 'steps': 49558, 'loss/train': 3.649322271347046} 02/26/2022 17:46:12 - INFO - codeparrot_training - Step 49559: {'lr': 1.0412962674691273e-07, 'samples': 25374720, 'steps': 49559, 'loss/train': 1.5797438621520996} 02/26/2022 17:46:17 - INFO - codeparrot_training - Step 49560: {'lr': 1.0365795156508373e-07, 'samples': 25375232, 'steps': 49560, 'loss/train': 0.07243501394987106} 02/26/2022 17:46:21 - INFO - codeparrot_training - Step 49561: {'lr': 1.0318734685979548e-07, 'samples': 25375744, 'steps': 49561, 'loss/train': 1.9132602214813232} 02/26/2022 17:46:26 - INFO - codeparrot_training - Step 49562: {'lr': 1.0271781263312963e-07, 'samples': 25376256, 'steps': 49562, 'loss/train': 1.1305632591247559} 02/26/2022 17:46:30 - INFO - codeparrot_training - Step 49563: {'lr': 1.0224934888705684e-07, 'samples': 25376768, 'steps': 49563, 'loss/train': 1.9545669555664062} 02/26/2022 17:46:37 - INFO - codeparrot_training - Step 49564: {'lr': 1.0178195562360326e-07, 'samples': 25377280, 'steps': 49564, 'loss/train': 0.019778305664658546} 02/26/2022 17:46:40 - INFO - codeparrot_training - Step 49565: {'lr': 1.013156328447673e-07, 'samples': 25377792, 'steps': 49565, 'loss/train': 1.5807524919509888} 02/26/2022 17:46:46 - INFO - codeparrot_training - Step 49566: {'lr': 1.008503805525196e-07, 'samples': 25378304, 'steps': 49566, 'loss/train': 1.3367489576339722} 02/26/2022 17:46:49 - INFO - codeparrot_training - Step 49567: {'lr': 1.0038619874888633e-07, 'samples': 25378816, 'steps': 49567, 'loss/train': 2.774674654006958} 02/26/2022 17:46:55 - INFO - codeparrot_training - Step 49568: {'lr': 9.992308743586587e-08, 'samples': 25379328, 'steps': 49568, 'loss/train': 1.610655426979065} 02/26/2022 17:46:58 - INFO - codeparrot_training - Step 49569: {'lr': 9.946104661540112e-08, 'samples': 25379840, 'steps': 49569, 'loss/train': 1.7148969173431396} 02/26/2022 17:47:04 - INFO - codeparrot_training - Step 49570: {'lr': 9.900007628951824e-08, 'samples': 25380352, 'steps': 49570, 'loss/train': 1.8496601581573486} 02/26/2022 17:47:07 - INFO - codeparrot_training - Step 49571: {'lr': 9.854017646013236e-08, 'samples': 25380864, 'steps': 49571, 'loss/train': 1.7861485481262207} 02/26/2022 17:47:13 - INFO - codeparrot_training - Step 49572: {'lr': 9.808134712926963e-08, 'samples': 25381376, 'steps': 49572, 'loss/train': 2.089369535446167} 02/26/2022 17:47:16 - INFO - codeparrot_training - Step 49573: {'lr': 9.762358829887297e-08, 'samples': 25381888, 'steps': 49573, 'loss/train': 1.2879939079284668} 02/26/2022 17:47:24 - INFO - codeparrot_training - Step 49574: {'lr': 9.716689997091299e-08, 'samples': 25382400, 'steps': 49574, 'loss/train': 2.5727152824401855} 02/26/2022 17:47:27 - INFO - codeparrot_training - Step 49575: {'lr': 9.671128214730484e-08, 'samples': 25382912, 'steps': 49575, 'loss/train': 0.9690009355545044} 02/26/2022 17:47:33 - INFO - codeparrot_training - Step 49576: {'lr': 9.625673483007469e-08, 'samples': 25383424, 'steps': 49576, 'loss/train': 1.8034100532531738} 02/26/2022 17:47:36 - INFO - codeparrot_training - Step 49577: {'lr': 9.580325802108214e-08, 'samples': 25383936, 'steps': 49577, 'loss/train': 0.6103371977806091} 02/26/2022 17:47:42 - INFO - codeparrot_training - Step 49578: {'lr': 9.535085172235336e-08, 'samples': 25384448, 'steps': 49578, 'loss/train': 1.089733600616455} 02/26/2022 17:47:45 - INFO - codeparrot_training - Step 49579: {'lr': 9.489951593577573e-08, 'samples': 25384960, 'steps': 49579, 'loss/train': 1.1412036418914795} 02/26/2022 17:47:51 - INFO - codeparrot_training - Step 49580: {'lr': 9.444925066329213e-08, 'samples': 25385472, 'steps': 49580, 'loss/train': 1.5451005697250366} 02/26/2022 17:47:54 - INFO - codeparrot_training - Step 49581: {'lr': 9.40000559068177e-08, 'samples': 25385984, 'steps': 49581, 'loss/train': 2.1165547370910645} 02/26/2022 17:48:00 - INFO - codeparrot_training - Step 49582: {'lr': 9.35519316683231e-08, 'samples': 25386496, 'steps': 49582, 'loss/train': 1.5899916887283325} 02/26/2022 17:48:03 - INFO - codeparrot_training - Step 49583: {'lr': 9.310487794966793e-08, 'samples': 25387008, 'steps': 49583, 'loss/train': 1.745990514755249} 02/26/2022 17:48:09 - INFO - codeparrot_training - Step 49584: {'lr': 9.265889475282285e-08, 'samples': 25387520, 'steps': 49584, 'loss/train': 1.5534600019454956} 02/26/2022 17:48:12 - INFO - codeparrot_training - Step 49585: {'lr': 9.221398207964748e-08, 'samples': 25388032, 'steps': 49585, 'loss/train': 1.450218677520752} 02/26/2022 17:48:18 - INFO - codeparrot_training - Step 49586: {'lr': 9.177013993208471e-08, 'samples': 25388544, 'steps': 49586, 'loss/train': 1.0296509265899658} 02/26/2022 17:48:21 - INFO - codeparrot_training - Step 49587: {'lr': 9.132736831202193e-08, 'samples': 25389056, 'steps': 49587, 'loss/train': 1.8154584169387817} 02/26/2022 17:48:27 - INFO - codeparrot_training - Step 49588: {'lr': 9.088566722137425e-08, 'samples': 25389568, 'steps': 49588, 'loss/train': 1.367182970046997} 02/26/2022 17:48:30 - INFO - codeparrot_training - Step 49589: {'lr': 9.044503666200132e-08, 'samples': 25390080, 'steps': 49589, 'loss/train': 1.768072247505188} 02/26/2022 17:48:38 - INFO - codeparrot_training - Step 49590: {'lr': 9.000547663581826e-08, 'samples': 25390592, 'steps': 49590, 'loss/train': 1.7978105545043945} 02/26/2022 17:48:41 - INFO - codeparrot_training - Step 49591: {'lr': 8.956698714468469e-08, 'samples': 25391104, 'steps': 49591, 'loss/train': 0.29134321212768555} 02/26/2022 17:48:47 - INFO - codeparrot_training - Step 49592: {'lr': 8.912956819048801e-08, 'samples': 25391616, 'steps': 49592, 'loss/train': 2.074617862701416} 02/26/2022 17:48:50 - INFO - codeparrot_training - Step 49593: {'lr': 8.869321977511557e-08, 'samples': 25392128, 'steps': 49593, 'loss/train': 1.7583087682724} 02/26/2022 17:48:57 - INFO - codeparrot_training - Step 49594: {'lr': 8.825794190042703e-08, 'samples': 25392640, 'steps': 49594, 'loss/train': 1.9845067262649536} 02/26/2022 17:49:00 - INFO - codeparrot_training - Step 49595: {'lr': 8.782373456828197e-08, 'samples': 25393152, 'steps': 49595, 'loss/train': 0.9412076473236084} 02/26/2022 17:49:04 - INFO - codeparrot_training - Step 49596: {'lr': 8.739059778054004e-08, 'samples': 25393664, 'steps': 49596, 'loss/train': 2.2576260566711426} 02/26/2022 17:49:09 - INFO - codeparrot_training - Step 49597: {'lr': 8.695853153906086e-08, 'samples': 25394176, 'steps': 49597, 'loss/train': 1.793504238128662} 02/26/2022 17:49:13 - INFO - codeparrot_training - Step 49598: {'lr': 8.652753584567629e-08, 'samples': 25394688, 'steps': 49598, 'loss/train': 1.7318885326385498} 02/26/2022 17:49:18 - INFO - codeparrot_training - Step 49599: {'lr': 8.609761070227373e-08, 'samples': 25395200, 'steps': 49599, 'loss/train': 2.2207977771759033} 02/26/2022 17:49:22 - INFO - codeparrot_training - Step 49600: {'lr': 8.566875611068504e-08, 'samples': 25395712, 'steps': 49600, 'loss/train': 1.7379605770111084} 02/26/2022 17:49:29 - INFO - codeparrot_training - Step 49601: {'lr': 8.52409720727143e-08, 'samples': 25396224, 'steps': 49601, 'loss/train': 1.3713390827178955} 02/26/2022 17:49:33 - INFO - codeparrot_training - Step 49602: {'lr': 8.481425859022118e-08, 'samples': 25396736, 'steps': 49602, 'loss/train': 2.2123289108276367} 02/26/2022 17:49:38 - INFO - codeparrot_training - Step 49603: {'lr': 8.438861566503752e-08, 'samples': 25397248, 'steps': 49603, 'loss/train': 2.246407985687256} 02/26/2022 17:49:42 - INFO - codeparrot_training - Step 49604: {'lr': 8.396404329893969e-08, 'samples': 25397760, 'steps': 49604, 'loss/train': 2.2773749828338623} 02/26/2022 17:49:47 - INFO - codeparrot_training - Step 49605: {'lr': 8.354054149381507e-08, 'samples': 25398272, 'steps': 49605, 'loss/train': 2.126040458679199} 02/26/2022 17:49:51 - INFO - codeparrot_training - Step 49606: {'lr': 8.311811025144e-08, 'samples': 25398784, 'steps': 49606, 'loss/train': 2.6375536918640137} 02/26/2022 17:49:56 - INFO - codeparrot_training - Step 49607: {'lr': 8.269674957361862e-08, 'samples': 25399296, 'steps': 49607, 'loss/train': 1.9615708589553833} 02/26/2022 17:50:00 - INFO - codeparrot_training - Step 49608: {'lr': 8.227645946218276e-08, 'samples': 25399808, 'steps': 49608, 'loss/train': 1.1938278675079346} 02/26/2022 17:50:05 - INFO - codeparrot_training - Step 49609: {'lr': 8.185723991890881e-08, 'samples': 25400320, 'steps': 49609, 'loss/train': 2.205256462097168} 02/26/2022 17:50:09 - INFO - codeparrot_training - Step 49610: {'lr': 8.143909094560088e-08, 'samples': 25400832, 'steps': 49610, 'loss/train': 1.3597317934036255} 02/26/2022 17:50:16 - INFO - codeparrot_training - Step 49611: {'lr': 8.10220125440353e-08, 'samples': 25401344, 'steps': 49611, 'loss/train': 2.3371031284332275} 02/26/2022 17:50:20 - INFO - codeparrot_training - Step 49612: {'lr': 8.060600471604396e-08, 'samples': 25401856, 'steps': 49612, 'loss/train': 1.0389764308929443} 02/26/2022 17:50:25 - INFO - codeparrot_training - Step 49613: {'lr': 8.01910674633477e-08, 'samples': 25402368, 'steps': 49613, 'loss/train': 3.1536383628845215} 02/26/2022 17:50:28 - INFO - codeparrot_training - Step 49614: {'lr': 7.97772007877784e-08, 'samples': 25402880, 'steps': 49614, 'loss/train': 0.0577298179268837} 02/26/2022 17:50:34 - INFO - codeparrot_training - Step 49615: {'lr': 7.936440469105688e-08, 'samples': 25403392, 'steps': 49615, 'loss/train': 0.6977536678314209} 02/26/2022 17:50:37 - INFO - codeparrot_training - Step 49616: {'lr': 7.895267917501503e-08, 'samples': 25403904, 'steps': 49616, 'loss/train': 1.7199970483779907} 02/26/2022 17:50:43 - INFO - codeparrot_training - Step 49617: {'lr': 7.854202424134593e-08, 'samples': 25404416, 'steps': 49617, 'loss/train': 1.774612307548523} 02/26/2022 17:50:46 - INFO - codeparrot_training - Step 49618: {'lr': 7.81324398918537e-08, 'samples': 25404928, 'steps': 49618, 'loss/train': 2.8690364360809326} 02/26/2022 17:50:52 - INFO - codeparrot_training - Step 49619: {'lr': 7.772392612831469e-08, 'samples': 25405440, 'steps': 49619, 'loss/train': 0.31925228238105774} 02/26/2022 17:50:55 - INFO - codeparrot_training - Step 49620: {'lr': 7.7316482952422e-08, 'samples': 25405952, 'steps': 49620, 'loss/train': 1.7292675971984863} 02/26/2022 17:51:02 - INFO - codeparrot_training - Step 49621: {'lr': 7.691011036592421e-08, 'samples': 25406464, 'steps': 49621, 'loss/train': 2.1880085468292236} 02/26/2022 17:51:06 - INFO - codeparrot_training - Step 49622: {'lr': 7.650480837062546e-08, 'samples': 25406976, 'steps': 49622, 'loss/train': 1.6229488849639893} 02/26/2022 17:51:11 - INFO - codeparrot_training - Step 49623: {'lr': 7.610057696819106e-08, 'samples': 25407488, 'steps': 49623, 'loss/train': 1.0848151445388794} 02/26/2022 17:51:15 - INFO - codeparrot_training - Step 49624: {'lr': 7.569741616039738e-08, 'samples': 25408000, 'steps': 49624, 'loss/train': 1.605403184890747} 02/26/2022 17:51:20 - INFO - codeparrot_training - Step 49625: {'lr': 7.529532594893751e-08, 'samples': 25408512, 'steps': 49625, 'loss/train': 3.2828586101531982} 02/26/2022 17:51:24 - INFO - codeparrot_training - Step 49626: {'lr': 7.489430633556005e-08, 'samples': 25409024, 'steps': 49626, 'loss/train': 1.3865106105804443} 02/26/2022 17:51:29 - INFO - codeparrot_training - Step 49627: {'lr': 7.449435732198584e-08, 'samples': 25409536, 'steps': 49627, 'loss/train': 1.3934231996536255} 02/26/2022 17:51:33 - INFO - codeparrot_training - Step 49628: {'lr': 7.409547890993573e-08, 'samples': 25410048, 'steps': 49628, 'loss/train': 1.8808860778808594} 02/26/2022 17:51:39 - INFO - codeparrot_training - Step 49629: {'lr': 7.369767110107507e-08, 'samples': 25410560, 'steps': 49629, 'loss/train': 2.0115225315093994} 02/26/2022 17:51:42 - INFO - codeparrot_training - Step 49630: {'lr': 7.330093389712466e-08, 'samples': 25411072, 'steps': 49630, 'loss/train': 2.3045902252197266} 02/26/2022 17:51:46 - INFO - codeparrot_training - Step 49631: {'lr': 7.29052672998054e-08, 'samples': 25411584, 'steps': 49631, 'loss/train': 1.2370697259902954} 02/26/2022 17:51:51 - INFO - codeparrot_training - Step 49632: {'lr': 7.25106713107826e-08, 'samples': 25412096, 'steps': 49632, 'loss/train': 1.484943151473999} 02/26/2022 17:51:55 - INFO - codeparrot_training - Step 49633: {'lr': 7.211714593177709e-08, 'samples': 25412608, 'steps': 49633, 'loss/train': 2.5384371280670166} 02/26/2022 17:52:00 - INFO - codeparrot_training - Step 49634: {'lr': 7.172469116448199e-08, 'samples': 25413120, 'steps': 49634, 'loss/train': 1.9349697828292847} 02/26/2022 17:52:04 - INFO - codeparrot_training - Step 49635: {'lr': 7.133330701050711e-08, 'samples': 25413632, 'steps': 49635, 'loss/train': 2.10516619682312} 02/26/2022 17:52:11 - INFO - codeparrot_training - Step 49636: {'lr': 7.094299347162881e-08, 'samples': 25414144, 'steps': 49636, 'loss/train': 1.5242670774459839} 02/26/2022 17:52:14 - INFO - codeparrot_training - Step 49637: {'lr': 7.055375054942914e-08, 'samples': 25414656, 'steps': 49637, 'loss/train': 0.8418490886688232} 02/26/2022 17:52:20 - INFO - codeparrot_training - Step 49638: {'lr': 7.016557824565672e-08, 'samples': 25415168, 'steps': 49638, 'loss/train': 1.4000853300094604} 02/26/2022 17:52:23 - INFO - codeparrot_training - Step 49639: {'lr': 6.977847656189362e-08, 'samples': 25415680, 'steps': 49639, 'loss/train': 1.5796009302139282} 02/26/2022 17:52:29 - INFO - codeparrot_training - Step 49640: {'lr': 6.939244549986068e-08, 'samples': 25416192, 'steps': 49640, 'loss/train': 1.6645773649215698} 02/26/2022 17:52:33 - INFO - codeparrot_training - Step 49641: {'lr': 6.900748506119547e-08, 'samples': 25416704, 'steps': 49641, 'loss/train': 0.4310552179813385} 02/26/2022 17:52:38 - INFO - codeparrot_training - Step 49642: {'lr': 6.862359524750782e-08, 'samples': 25417216, 'steps': 49642, 'loss/train': 1.2219706773757935} 02/26/2022 17:52:41 - INFO - codeparrot_training - Step 49643: {'lr': 6.824077606049083e-08, 'samples': 25417728, 'steps': 49643, 'loss/train': 1.665999174118042} 02/26/2022 17:52:47 - INFO - codeparrot_training - Step 49644: {'lr': 6.785902750178208e-08, 'samples': 25418240, 'steps': 49644, 'loss/train': 1.4190175533294678} 02/26/2022 17:52:50 - INFO - codeparrot_training - Step 49645: {'lr': 6.747834957299137e-08, 'samples': 25418752, 'steps': 49645, 'loss/train': 1.5200834274291992} 02/26/2022 17:52:57 - INFO - codeparrot_training - Step 49646: {'lr': 6.70987422757563e-08, 'samples': 25419264, 'steps': 49646, 'loss/train': 1.9355356693267822} 02/26/2022 17:53:00 - INFO - codeparrot_training - Step 49647: {'lr': 6.672020561171443e-08, 'samples': 25419776, 'steps': 49647, 'loss/train': 2.010352849960327} 02/26/2022 17:53:06 - INFO - codeparrot_training - Step 49648: {'lr': 6.63427395824756e-08, 'samples': 25420288, 'steps': 49648, 'loss/train': 1.577196717262268} 02/26/2022 17:53:09 - INFO - codeparrot_training - Step 49649: {'lr': 6.59663441896774e-08, 'samples': 25420800, 'steps': 49649, 'loss/train': 2.2887322902679443} 02/26/2022 17:53:15 - INFO - codeparrot_training - Step 49650: {'lr': 6.559101943490186e-08, 'samples': 25421312, 'steps': 49650, 'loss/train': 1.492795467376709} 02/26/2022 17:53:18 - INFO - codeparrot_training - Step 49651: {'lr': 6.52167653197866e-08, 'samples': 25421824, 'steps': 49651, 'loss/train': 1.6093878746032715} 02/26/2022 17:53:24 - INFO - codeparrot_training - Step 49652: {'lr': 6.48435818458859e-08, 'samples': 25422336, 'steps': 49652, 'loss/train': 2.2008275985717773} 02/26/2022 17:53:27 - INFO - codeparrot_training - Step 49653: {'lr': 6.447146901486511e-08, 'samples': 25422848, 'steps': 49653, 'loss/train': 1.3510240316390991} 02/26/2022 17:53:33 - INFO - codeparrot_training - Step 49654: {'lr': 6.410042682827855e-08, 'samples': 25423360, 'steps': 49654, 'loss/train': 1.2156535387039185} 02/26/2022 17:53:36 - INFO - codeparrot_training - Step 49655: {'lr': 6.373045528770826e-08, 'samples': 25423872, 'steps': 49655, 'loss/train': 2.296717882156372} 02/26/2022 17:53:42 - INFO - codeparrot_training - Step 49656: {'lr': 6.33615543947641e-08, 'samples': 25424384, 'steps': 49656, 'loss/train': 1.5778707265853882} 02/26/2022 17:53:46 - INFO - codeparrot_training - Step 49657: {'lr': 6.299372415100036e-08, 'samples': 25424896, 'steps': 49657, 'loss/train': 1.9722676277160645} 02/26/2022 17:53:51 - INFO - codeparrot_training - Step 49658: {'lr': 6.262696455802685e-08, 'samples': 25425408, 'steps': 49658, 'loss/train': 2.2931156158447266} 02/26/2022 17:53:55 - INFO - codeparrot_training - Step 49659: {'lr': 6.226127561739792e-08, 'samples': 25425920, 'steps': 49659, 'loss/train': 2.960163116455078} 02/26/2022 17:54:00 - INFO - codeparrot_training - Step 49660: {'lr': 6.189665733066785e-08, 'samples': 25426432, 'steps': 49660, 'loss/train': 1.4128910303115845} 02/26/2022 17:54:04 - INFO - codeparrot_training - Step 49661: {'lr': 6.153310969939096e-08, 'samples': 25426944, 'steps': 49661, 'loss/train': 0.6558352708816528} 02/26/2022 17:54:09 - INFO - codeparrot_training - Step 49662: {'lr': 6.11706327251771e-08, 'samples': 25427456, 'steps': 49662, 'loss/train': 0.646000862121582} 02/26/2022 17:54:13 - INFO - codeparrot_training - Step 49663: {'lr': 6.080922640952502e-08, 'samples': 25427968, 'steps': 49663, 'loss/train': 2.069631814956665} 02/26/2022 17:54:19 - INFO - codeparrot_training - Step 49664: {'lr': 6.044889075398908e-08, 'samples': 25428480, 'steps': 49664, 'loss/train': 1.1431056261062622} 02/26/2022 17:54:23 - INFO - codeparrot_training - Step 49665: {'lr': 6.008962576015132e-08, 'samples': 25428992, 'steps': 49665, 'loss/train': 1.4981911182403564} 02/26/2022 17:54:28 - INFO - codeparrot_training - Step 49666: {'lr': 5.973143142951054e-08, 'samples': 25429504, 'steps': 49666, 'loss/train': 1.0248603820800781} 02/26/2022 17:54:32 - INFO - codeparrot_training - Step 49667: {'lr': 5.9374307763621075e-08, 'samples': 25430016, 'steps': 49667, 'loss/train': 0.1982443928718567} 02/26/2022 17:54:37 - INFO - codeparrot_training - Step 49668: {'lr': 5.9018254763981705e-08, 'samples': 25430528, 'steps': 49668, 'loss/train': 1.9074077606201172} 02/26/2022 17:54:41 - INFO - codeparrot_training - Step 49669: {'lr': 5.866327243217451e-08, 'samples': 25431040, 'steps': 49669, 'loss/train': 2.692355155944824} 02/26/2022 17:54:46 - INFO - codeparrot_training - Step 49670: {'lr': 5.830936076967053e-08, 'samples': 25431552, 'steps': 49670, 'loss/train': 1.570583701133728} 02/26/2022 17:54:50 - INFO - codeparrot_training - Step 49671: {'lr': 5.795651977802408e-08, 'samples': 25432064, 'steps': 49671, 'loss/train': 1.241619348526001} 02/26/2022 17:54:55 - INFO - codeparrot_training - Step 49672: {'lr': 5.7604749458706194e-08, 'samples': 25432576, 'steps': 49672, 'loss/train': 1.0222573280334473} 02/26/2022 17:54:59 - INFO - codeparrot_training - Step 49673: {'lr': 5.725404981324345e-08, 'samples': 25433088, 'steps': 49673, 'loss/train': 0.6399295926094055} 02/26/2022 17:55:04 - INFO - codeparrot_training - Step 49674: {'lr': 5.690442084313463e-08, 'samples': 25433600, 'steps': 49674, 'loss/train': 2.2612533569335938} 02/26/2022 17:55:08 - INFO - codeparrot_training - Step 49675: {'lr': 5.65558625499063e-08, 'samples': 25434112, 'steps': 49675, 'loss/train': 1.8178939819335938} 02/26/2022 17:55:14 - INFO - codeparrot_training - Step 49676: {'lr': 5.620837493500175e-08, 'samples': 25434624, 'steps': 49676, 'loss/train': 2.5089809894561768} 02/26/2022 17:55:17 - INFO - codeparrot_training - Step 49677: {'lr': 5.586195799994753e-08, 'samples': 25435136, 'steps': 49677, 'loss/train': 1.1936006546020508} 02/26/2022 17:55:23 - INFO - codeparrot_training - Step 49678: {'lr': 5.55166117462147e-08, 'samples': 25435648, 'steps': 49678, 'loss/train': 9.683402061462402} 02/26/2022 17:55:27 - INFO - codeparrot_training - Step 49679: {'lr': 5.5172336175274285e-08, 'samples': 25436160, 'steps': 49679, 'loss/train': 1.7998840808868408} 02/26/2022 17:55:32 - INFO - codeparrot_training - Step 49680: {'lr': 5.48291312886251e-08, 'samples': 25436672, 'steps': 49680, 'loss/train': 1.4408538341522217} 02/26/2022 17:55:36 - INFO - codeparrot_training - Step 49681: {'lr': 5.448699708771043e-08, 'samples': 25437184, 'steps': 49681, 'loss/train': 1.694828987121582} 02/26/2022 17:55:42 - INFO - codeparrot_training - Step 49682: {'lr': 5.414593357400133e-08, 'samples': 25437696, 'steps': 49682, 'loss/train': 1.576622724533081} 02/26/2022 17:55:45 - INFO - codeparrot_training - Step 49683: {'lr': 5.380594074896883e-08, 'samples': 25438208, 'steps': 49683, 'loss/train': 2.116130828857422} 02/26/2022 17:55:51 - INFO - codeparrot_training - Step 49684: {'lr': 5.346701861408398e-08, 'samples': 25438720, 'steps': 49684, 'loss/train': 1.9564499855041504} 02/26/2022 17:55:54 - INFO - codeparrot_training - Step 49685: {'lr': 5.312916717076233e-08, 'samples': 25439232, 'steps': 49685, 'loss/train': 1.2669018507003784} 02/26/2022 17:56:00 - INFO - codeparrot_training - Step 49686: {'lr': 5.27923864204749e-08, 'samples': 25439744, 'steps': 49686, 'loss/train': 1.9539506435394287} 02/26/2022 17:56:03 - INFO - codeparrot_training - Step 49687: {'lr': 5.245667636463725e-08, 'samples': 25440256, 'steps': 49687, 'loss/train': 1.2277711629867554} 02/26/2022 17:56:09 - INFO - codeparrot_training - Step 49688: {'lr': 5.21220370047204e-08, 'samples': 25440768, 'steps': 49688, 'loss/train': 2.543898105621338} 02/26/2022 17:56:12 - INFO - codeparrot_training - Step 49689: {'lr': 5.1788468342139905e-08, 'samples': 25441280, 'steps': 49689, 'loss/train': 1.5499975681304932} 02/26/2022 17:56:18 - INFO - codeparrot_training - Step 49690: {'lr': 5.145597037833904e-08, 'samples': 25441792, 'steps': 49690, 'loss/train': 1.2460930347442627} 02/26/2022 17:56:21 - INFO - codeparrot_training - Step 49691: {'lr': 5.112454311473336e-08, 'samples': 25442304, 'steps': 49691, 'loss/train': 1.6725564002990723} 02/26/2022 17:56:27 - INFO - codeparrot_training - Step 49692: {'lr': 5.0794186552738375e-08, 'samples': 25442816, 'steps': 49692, 'loss/train': 1.2821602821350098} 02/26/2022 17:56:31 - INFO - codeparrot_training - Step 49693: {'lr': 5.046490069379739e-08, 'samples': 25443328, 'steps': 49693, 'loss/train': 2.3357479572296143} 02/26/2022 17:56:36 - INFO - codeparrot_training - Step 49694: {'lr': 5.0136685539270424e-08, 'samples': 25443840, 'steps': 49694, 'loss/train': 0.6887544989585876} 02/26/2022 17:56:40 - INFO - codeparrot_training - Step 49695: {'lr': 4.9809541090573005e-08, 'samples': 25444352, 'steps': 49695, 'loss/train': 3.190800189971924} 02/26/2022 17:56:45 - INFO - codeparrot_training - Step 49696: {'lr': 4.948346734914844e-08, 'samples': 25444864, 'steps': 49696, 'loss/train': 1.0308183431625366} 02/26/2022 17:56:49 - INFO - codeparrot_training - Step 49697: {'lr': 4.9158464316384486e-08, 'samples': 25445376, 'steps': 49697, 'loss/train': 2.798590898513794} 02/26/2022 17:56:54 - INFO - codeparrot_training - Step 49698: {'lr': 4.883453199361343e-08, 'samples': 25445888, 'steps': 49698, 'loss/train': 1.6407665014266968} 02/26/2022 17:56:58 - INFO - codeparrot_training - Step 49699: {'lr': 4.851167038230631e-08, 'samples': 25446400, 'steps': 49699, 'loss/train': 2.0579278469085693} 02/26/2022 17:57:03 - INFO - codeparrot_training - Step 49700: {'lr': 4.818987948379538e-08, 'samples': 25446912, 'steps': 49700, 'loss/train': 2.0435235500335693} 02/26/2022 17:57:07 - INFO - codeparrot_training - Step 49701: {'lr': 4.7869159299468443e-08, 'samples': 25447424, 'steps': 49701, 'loss/train': 3.1509881019592285} 02/26/2022 17:57:13 - INFO - codeparrot_training - Step 49702: {'lr': 4.7549509830685514e-08, 'samples': 25447936, 'steps': 49702, 'loss/train': 1.106390118598938} 02/26/2022 17:57:17 - INFO - codeparrot_training - Step 49703: {'lr': 4.723093107883436e-08, 'samples': 25448448, 'steps': 49703, 'loss/train': 1.3935894966125488} 02/26/2022 17:57:22 - INFO - codeparrot_training - Step 49704: {'lr': 4.6913423045302773e-08, 'samples': 25448960, 'steps': 49704, 'loss/train': 1.944941520690918} 02/26/2022 17:57:26 - INFO - codeparrot_training - Step 49705: {'lr': 4.659698573139526e-08, 'samples': 25449472, 'steps': 49705, 'loss/train': 1.6275675296783447} 02/26/2022 17:57:31 - INFO - codeparrot_training - Step 49706: {'lr': 4.628161913852735e-08, 'samples': 25449984, 'steps': 49706, 'loss/train': 1.9497686624526978} 02/26/2022 17:57:35 - INFO - codeparrot_training - Step 49707: {'lr': 4.596732326800357e-08, 'samples': 25450496, 'steps': 49707, 'loss/train': 1.7830393314361572} 02/26/2022 17:57:40 - INFO - codeparrot_training - Step 49708: {'lr': 4.565409812118393e-08, 'samples': 25451008, 'steps': 49708, 'loss/train': 2.2010281085968018} 02/26/2022 17:57:44 - INFO - codeparrot_training - Step 49709: {'lr': 4.53419436994007e-08, 'samples': 25451520, 'steps': 49709, 'loss/train': 0.8849744200706482} 02/26/2022 17:57:50 - INFO - codeparrot_training - Step 49710: {'lr': 4.503086000404166e-08, 'samples': 25452032, 'steps': 49710, 'loss/train': 2.3053812980651855} 02/26/2022 17:57:53 - INFO - codeparrot_training - Step 49711: {'lr': 4.4720847036355814e-08, 'samples': 25452544, 'steps': 49711, 'loss/train': 0.9687795042991638} 02/26/2022 17:57:59 - INFO - codeparrot_training - Step 49712: {'lr': 4.4411904797758695e-08, 'samples': 25453056, 'steps': 49712, 'loss/train': 2.7189462184906006} 02/26/2022 17:58:03 - INFO - codeparrot_training - Step 49713: {'lr': 4.41040332894993e-08, 'samples': 25453568, 'steps': 49713, 'loss/train': 0.98175448179245} 02/26/2022 17:58:08 - INFO - codeparrot_training - Step 49714: {'lr': 4.3797232512937655e-08, 'samples': 25454080, 'steps': 49714, 'loss/train': 1.117993950843811} 02/26/2022 17:58:12 - INFO - codeparrot_training - Step 49715: {'lr': 4.3491502469378275e-08, 'samples': 25454592, 'steps': 49715, 'loss/train': 1.2794795036315918} 02/26/2022 17:58:17 - INFO - codeparrot_training - Step 49716: {'lr': 4.3186843160125664e-08, 'samples': 25455104, 'steps': 49716, 'loss/train': 1.9968023300170898} 02/26/2022 17:58:21 - INFO - codeparrot_training - Step 49717: {'lr': 4.28832545865121e-08, 'samples': 25455616, 'steps': 49717, 'loss/train': 2.224591016769409} 02/26/2022 17:58:27 - INFO - codeparrot_training - Step 49718: {'lr': 4.2580736749786575e-08, 'samples': 25456128, 'steps': 49718, 'loss/train': 2.2531254291534424} 02/26/2022 17:58:31 - INFO - codeparrot_training - Step 49719: {'lr': 4.227928965130912e-08, 'samples': 25456640, 'steps': 49719, 'loss/train': 1.970960021018982} 02/26/2022 17:58:34 - INFO - codeparrot_training - Step 49720: {'lr': 4.197891329230097e-08, 'samples': 25457152, 'steps': 49720, 'loss/train': 2.8190581798553467} 02/26/2022 17:58:40 - INFO - codeparrot_training - Step 49721: {'lr': 4.16796076740944e-08, 'samples': 25457664, 'steps': 49721, 'loss/train': 1.9687682390213013} 02/26/2022 17:58:43 - INFO - codeparrot_training - Step 49722: {'lr': 4.138137279799392e-08, 'samples': 25458176, 'steps': 49722, 'loss/train': 2.0611155033111572} 02/26/2022 17:58:49 - INFO - codeparrot_training - Step 49723: {'lr': 4.1084208665193025e-08, 'samples': 25458688, 'steps': 49723, 'loss/train': 1.9255492687225342} 02/26/2022 17:58:53 - INFO - codeparrot_training - Step 49724: {'lr': 4.0788115277051727e-08, 'samples': 25459200, 'steps': 49724, 'loss/train': 1.390259027481079} 02/26/2022 17:58:59 - INFO - codeparrot_training - Step 49725: {'lr': 4.0493092634791286e-08, 'samples': 25459712, 'steps': 49725, 'loss/train': 1.708520770072937} 02/26/2022 17:59:02 - INFO - codeparrot_training - Step 49726: {'lr': 4.019914073968844e-08, 'samples': 25460224, 'steps': 49726, 'loss/train': 2.6933388710021973} 02/26/2022 17:59:07 - INFO - codeparrot_training - Step 49727: {'lr': 3.990625959301997e-08, 'samples': 25460736, 'steps': 49727, 'loss/train': 1.3498625755310059} 02/26/2022 17:59:11 - INFO - codeparrot_training - Step 49728: {'lr': 3.96144491960071e-08, 'samples': 25461248, 'steps': 49728, 'loss/train': 1.728865623474121} 02/26/2022 17:59:16 - INFO - codeparrot_training - Step 49729: {'lr': 3.932370954989884e-08, 'samples': 25461760, 'steps': 49729, 'loss/train': 1.798843502998352} 02/26/2022 17:59:20 - INFO - codeparrot_training - Step 49730: {'lr': 3.903404065597194e-08, 'samples': 25462272, 'steps': 49730, 'loss/train': 1.8355815410614014} 02/26/2022 17:59:25 - INFO - codeparrot_training - Step 49731: {'lr': 3.874544251544765e-08, 'samples': 25462784, 'steps': 49731, 'loss/train': 1.0345803499221802} 02/26/2022 17:59:29 - INFO - codeparrot_training - Step 49732: {'lr': 3.845791512957497e-08, 'samples': 25463296, 'steps': 49732, 'loss/train': 1.8129682540893555} 02/26/2022 17:59:34 - INFO - codeparrot_training - Step 49733: {'lr': 3.8171458499575154e-08, 'samples': 25463808, 'steps': 49733, 'loss/train': 1.964986801147461} 02/26/2022 17:59:38 - INFO - codeparrot_training - Step 49734: {'lr': 3.7886072626669434e-08, 'samples': 25464320, 'steps': 49734, 'loss/train': 1.252440094947815} 02/26/2022 17:59:43 - INFO - codeparrot_training - Step 49735: {'lr': 3.760175751207906e-08, 'samples': 25464832, 'steps': 49735, 'loss/train': 1.0094990730285645} 02/26/2022 17:59:47 - INFO - codeparrot_training - Step 49736: {'lr': 3.7318513157053036e-08, 'samples': 25465344, 'steps': 49736, 'loss/train': 1.6370432376861572} 02/26/2022 17:59:52 - INFO - codeparrot_training - Step 49737: {'lr': 3.703633956275709e-08, 'samples': 25465856, 'steps': 49737, 'loss/train': 0.5853754878044128} 02/26/2022 17:59:56 - INFO - codeparrot_training - Step 49738: {'lr': 3.6755236730440235e-08, 'samples': 25466368, 'steps': 49738, 'loss/train': 1.8676743507385254} 02/26/2022 18:00:02 - INFO - codeparrot_training - Step 49739: {'lr': 3.6475204661295946e-08, 'samples': 25466880, 'steps': 49739, 'loss/train': 0.8303935527801514} 02/26/2022 18:00:05 - INFO - codeparrot_training - Step 49740: {'lr': 3.619624335651772e-08, 'samples': 25467392, 'steps': 49740, 'loss/train': 0.33742159605026245} 02/26/2022 18:00:11 - INFO - codeparrot_training - Step 49741: {'lr': 3.591835281729905e-08, 'samples': 25467904, 'steps': 49741, 'loss/train': 2.738327980041504} 02/26/2022 18:00:14 - INFO - codeparrot_training - Step 49742: {'lr': 3.5641533044833417e-08, 'samples': 25468416, 'steps': 49742, 'loss/train': 1.4374414682388306} 02/26/2022 18:00:20 - INFO - codeparrot_training - Step 49743: {'lr': 3.536578404031432e-08, 'samples': 25468928, 'steps': 49743, 'loss/train': 1.5922343730926514} 02/26/2022 18:00:23 - INFO - codeparrot_training - Step 49744: {'lr': 3.509110580490749e-08, 'samples': 25469440, 'steps': 49744, 'loss/train': 0.9981396794319153} 02/26/2022 18:00:29 - INFO - codeparrot_training - Step 49745: {'lr': 3.481749833977865e-08, 'samples': 25469952, 'steps': 49745, 'loss/train': 1.3962358236312866} 02/26/2022 18:00:32 - INFO - codeparrot_training - Step 49746: {'lr': 3.4544961646149065e-08, 'samples': 25470464, 'steps': 49746, 'loss/train': 2.002027988433838} 02/26/2022 18:00:38 - INFO - codeparrot_training - Step 49747: {'lr': 3.42734957251567e-08, 'samples': 25470976, 'steps': 49747, 'loss/train': 0.1796267330646515} 02/26/2022 18:00:41 - INFO - codeparrot_training - Step 49748: {'lr': 3.4003100577939536e-08, 'samples': 25471488, 'steps': 49748, 'loss/train': 1.6250418424606323} 02/26/2022 18:00:48 - INFO - codeparrot_training - Step 49749: {'lr': 3.373377620569107e-08, 'samples': 25472000, 'steps': 49749, 'loss/train': 1.3040080070495605} 02/26/2022 18:00:51 - INFO - codeparrot_training - Step 49750: {'lr': 3.346552260954927e-08, 'samples': 25472512, 'steps': 49750, 'loss/train': 1.5120213031768799} 02/26/2022 18:00:57 - INFO - codeparrot_training - Step 49751: {'lr': 3.3198339790679874e-08, 'samples': 25473024, 'steps': 49751, 'loss/train': 1.7201801538467407} 02/26/2022 18:01:00 - INFO - codeparrot_training - Step 49752: {'lr': 3.293222775022087e-08, 'samples': 25473536, 'steps': 49752, 'loss/train': 2.538609027862549} 02/26/2022 18:01:06 - INFO - codeparrot_training - Step 49753: {'lr': 3.2667186489282465e-08, 'samples': 25474048, 'steps': 49753, 'loss/train': 1.6552835702896118} 02/26/2022 18:01:09 - INFO - codeparrot_training - Step 49754: {'lr': 3.24032160090304e-08, 'samples': 25474560, 'steps': 49754, 'loss/train': 1.060267686843872} 02/26/2022 18:01:15 - INFO - codeparrot_training - Step 49755: {'lr': 3.214031631060266e-08, 'samples': 25475072, 'steps': 49755, 'loss/train': 1.078352689743042} 02/26/2022 18:01:18 - INFO - codeparrot_training - Step 49756: {'lr': 3.187848739510946e-08, 'samples': 25475584, 'steps': 49756, 'loss/train': 1.0409449338912964} 02/26/2022 18:01:24 - INFO - codeparrot_training - Step 49757: {'lr': 3.161772926366102e-08, 'samples': 25476096, 'steps': 49757, 'loss/train': 1.6763296127319336} 02/26/2022 18:01:27 - INFO - codeparrot_training - Step 49758: {'lr': 3.135804191739533e-08, 'samples': 25476608, 'steps': 49758, 'loss/train': 2.109692096710205} 02/26/2022 18:01:34 - INFO - codeparrot_training - Step 49759: {'lr': 3.1099425357422604e-08, 'samples': 25477120, 'steps': 49759, 'loss/train': 0.2643895447254181} 02/26/2022 18:01:37 - INFO - codeparrot_training - Step 49760: {'lr': 3.084187958485307e-08, 'samples': 25477632, 'steps': 49760, 'loss/train': 2.084852695465088} 02/26/2022 18:01:43 - INFO - codeparrot_training - Step 49761: {'lr': 3.05854046007692e-08, 'samples': 25478144, 'steps': 49761, 'loss/train': 2.0043864250183105} 02/26/2022 18:01:47 - INFO - codeparrot_training - Step 49762: {'lr': 3.03300004062812e-08, 'samples': 25478656, 'steps': 49762, 'loss/train': 1.6485490798950195} 02/26/2022 18:01:52 - INFO - codeparrot_training - Step 49763: {'lr': 3.007566700249931e-08, 'samples': 25479168, 'steps': 49763, 'loss/train': 1.4275230169296265} 02/26/2022 18:01:56 - INFO - codeparrot_training - Step 49764: {'lr': 2.982240439047823e-08, 'samples': 25479680, 'steps': 49764, 'loss/train': 1.6674201488494873} 02/26/2022 18:02:01 - INFO - codeparrot_training - Step 49765: {'lr': 2.9570212571355947e-08, 'samples': 25480192, 'steps': 49765, 'loss/train': 1.0222281217575073} 02/26/2022 18:02:05 - INFO - codeparrot_training - Step 49766: {'lr': 2.9319091546159415e-08, 'samples': 25480704, 'steps': 49766, 'loss/train': 1.992708444595337} 02/26/2022 18:02:11 - INFO - codeparrot_training - Step 49767: {'lr': 2.90690413159711e-08, 'samples': 25481216, 'steps': 49767, 'loss/train': 0.5209033489227295} 02/26/2022 18:02:14 - INFO - codeparrot_training - Step 49768: {'lr': 2.8820061881901226e-08, 'samples': 25481728, 'steps': 49768, 'loss/train': 1.1297340393066406} 02/26/2022 18:02:19 - INFO - codeparrot_training - Step 49769: {'lr': 2.857215324497675e-08, 'samples': 25482240, 'steps': 49769, 'loss/train': 0.6178969144821167} 02/26/2022 18:02:23 - INFO - codeparrot_training - Step 49770: {'lr': 2.832531540628014e-08, 'samples': 25482752, 'steps': 49770, 'loss/train': 1.3847692012786865} 02/26/2022 18:02:30 - INFO - codeparrot_training - Step 49771: {'lr': 2.8079548366866105e-08, 'samples': 25483264, 'steps': 49771, 'loss/train': 2.3196327686309814} 02/26/2022 18:02:33 - INFO - codeparrot_training - Step 49772: {'lr': 2.7834852127789356e-08, 'samples': 25483776, 'steps': 49772, 'loss/train': 2.1752512454986572} 02/26/2022 18:02:39 - INFO - codeparrot_training - Step 49773: {'lr': 2.7591226690076853e-08, 'samples': 25484288, 'steps': 49773, 'loss/train': 2.07833194732666} 02/26/2022 18:02:42 - INFO - codeparrot_training - Step 49774: {'lr': 2.734867205478331e-08, 'samples': 25484800, 'steps': 49774, 'loss/train': 1.5006252527236938} 02/26/2022 18:02:48 - INFO - codeparrot_training - Step 49775: {'lr': 2.7107188222991187e-08, 'samples': 25485312, 'steps': 49775, 'loss/train': 1.5359556674957275} 02/26/2022 18:02:51 - INFO - codeparrot_training - Step 49776: {'lr': 2.686677519564418e-08, 'samples': 25485824, 'steps': 49776, 'loss/train': 1.9262195825576782} 02/26/2022 18:02:57 - INFO - codeparrot_training - Step 49777: {'lr': 2.6627432973852506e-08, 'samples': 25486336, 'steps': 49777, 'loss/train': 0.4731055796146393} 02/26/2022 18:03:00 - INFO - codeparrot_training - Step 49778: {'lr': 2.638916155861537e-08, 'samples': 25486848, 'steps': 49778, 'loss/train': 2.6146676540374756} 02/26/2022 18:03:06 - INFO - codeparrot_training - Step 49779: {'lr': 2.615196095095973e-08, 'samples': 25487360, 'steps': 49779, 'loss/train': 2.111182928085327} 02/26/2022 18:03:09 - INFO - codeparrot_training - Step 49780: {'lr': 2.5915831151884784e-08, 'samples': 25487872, 'steps': 49780, 'loss/train': 1.5207401514053345} 02/26/2022 18:03:15 - INFO - codeparrot_training - Step 49781: {'lr': 2.5680772162389732e-08, 'samples': 25488384, 'steps': 49781, 'loss/train': 2.062026262283325} 02/26/2022 18:03:18 - INFO - codeparrot_training - Step 49782: {'lr': 2.5446783983529286e-08, 'samples': 25488896, 'steps': 49782, 'loss/train': 1.3513418436050415} 02/26/2022 18:03:24 - INFO - codeparrot_training - Step 49783: {'lr': 2.5213866616274895e-08, 'samples': 25489408, 'steps': 49783, 'loss/train': 1.4412165880203247} 02/26/2022 18:03:27 - INFO - codeparrot_training - Step 49784: {'lr': 2.4982020061625754e-08, 'samples': 25489920, 'steps': 49784, 'loss/train': 0.5724613070487976} 02/26/2022 18:03:33 - INFO - codeparrot_training - Step 49785: {'lr': 2.475124432058107e-08, 'samples': 25490432, 'steps': 49785, 'loss/train': 1.6561940908432007} 02/26/2022 18:03:37 - INFO - codeparrot_training - Step 49786: {'lr': 2.452153939411228e-08, 'samples': 25490944, 'steps': 49786, 'loss/train': 2.0832836627960205} 02/26/2022 18:03:42 - INFO - codeparrot_training - Step 49787: {'lr': 2.4292905283218592e-08, 'samples': 25491456, 'steps': 49787, 'loss/train': 1.4292668104171753} 02/26/2022 18:03:46 - INFO - codeparrot_training - Step 49788: {'lr': 2.4065341988899204e-08, 'samples': 25491968, 'steps': 49788, 'loss/train': 1.3877793550491333} 02/26/2022 18:03:51 - INFO - codeparrot_training - Step 49789: {'lr': 2.383884951207005e-08, 'samples': 25492480, 'steps': 49789, 'loss/train': 0.8354439735412598} 02/26/2022 18:03:55 - INFO - codeparrot_training - Step 49790: {'lr': 2.3613427853758086e-08, 'samples': 25492992, 'steps': 49790, 'loss/train': 1.7699052095413208} 02/26/2022 18:04:00 - INFO - codeparrot_training - Step 49791: {'lr': 2.3389077014934756e-08, 'samples': 25493504, 'steps': 49791, 'loss/train': 0.6568465232849121} 02/26/2022 18:04:04 - INFO - codeparrot_training - Step 49792: {'lr': 2.3165796996515997e-08, 'samples': 25494016, 'steps': 49792, 'loss/train': 1.5680688619613647} 02/26/2022 18:04:09 - INFO - codeparrot_training - Step 49793: {'lr': 2.2943587799473253e-08, 'samples': 25494528, 'steps': 49793, 'loss/train': 1.324320912361145} 02/26/2022 18:04:13 - INFO - codeparrot_training - Step 49794: {'lr': 2.2722449424750214e-08, 'samples': 25495040, 'steps': 49794, 'loss/train': 2.320014238357544} 02/26/2022 18:04:18 - INFO - codeparrot_training - Step 49795: {'lr': 2.250238187334608e-08, 'samples': 25495552, 'steps': 49795, 'loss/train': 2.499485969543457} 02/26/2022 18:04:22 - INFO - codeparrot_training - Step 49796: {'lr': 2.2283385146121272e-08, 'samples': 25496064, 'steps': 49796, 'loss/train': 2.381624698638916} 02/26/2022 18:04:28 - INFO - codeparrot_training - Step 49797: {'lr': 2.2065459244074992e-08, 'samples': 25496576, 'steps': 49797, 'loss/train': 2.4942405223846436} 02/26/2022 18:04:31 - INFO - codeparrot_training - Step 49798: {'lr': 2.184860416815093e-08, 'samples': 25497088, 'steps': 49798, 'loss/train': 0.09525682032108307} 02/26/2022 18:04:37 - INFO - codeparrot_training - Step 49799: {'lr': 2.1632819919209513e-08, 'samples': 25497600, 'steps': 49799, 'loss/train': 2.514915943145752} 02/26/2022 18:04:40 - INFO - codeparrot_training - Step 49800: {'lr': 2.1418106498249933e-08, 'samples': 25498112, 'steps': 49800, 'loss/train': 1.5702953338623047} 02/26/2022 18:04:46 - INFO - codeparrot_training - Step 49801: {'lr': 2.120446390613262e-08, 'samples': 25498624, 'steps': 49801, 'loss/train': 1.5474028587341309} 02/26/2022 18:04:49 - INFO - codeparrot_training - Step 49802: {'lr': 2.0991892143801262e-08, 'samples': 25499136, 'steps': 49802, 'loss/train': 2.387803792953491} 02/26/2022 18:04:55 - INFO - codeparrot_training - Step 49803: {'lr': 2.078039121217179e-08, 'samples': 25499648, 'steps': 49803, 'loss/train': 0.5508601665496826} 02/26/2022 18:04:58 - INFO - codeparrot_training - Step 49804: {'lr': 2.0569961112132385e-08, 'samples': 25500160, 'steps': 49804, 'loss/train': 2.0543103218078613} 02/26/2022 18:05:04 - INFO - codeparrot_training - Step 49805: {'lr': 2.0360601844626736e-08, 'samples': 25500672, 'steps': 49805, 'loss/train': 2.405283212661743} 02/26/2022 18:05:07 - INFO - codeparrot_training - Step 49806: {'lr': 2.015231341048751e-08, 'samples': 25501184, 'steps': 49806, 'loss/train': 2.0754432678222656} 02/26/2022 18:05:13 - INFO - codeparrot_training - Step 49807: {'lr': 1.9945095810630642e-08, 'samples': 25501696, 'steps': 49807, 'loss/train': 0.9231384992599487} 02/26/2022 18:05:17 - INFO - codeparrot_training - Step 49808: {'lr': 1.9738949045972064e-08, 'samples': 25502208, 'steps': 49808, 'loss/train': 1.648268699645996} 02/26/2022 18:05:22 - INFO - codeparrot_training - Step 49809: {'lr': 1.95338731173722e-08, 'samples': 25502720, 'steps': 49809, 'loss/train': 1.5111702680587769} 02/26/2022 18:05:26 - INFO - codeparrot_training - Step 49810: {'lr': 1.9329868025691478e-08, 'samples': 25503232, 'steps': 49810, 'loss/train': 1.0084084272384644} 02/26/2022 18:05:31 - INFO - codeparrot_training - Step 49811: {'lr': 1.912693377184582e-08, 'samples': 25503744, 'steps': 49811, 'loss/train': 1.7887688875198364} 02/26/2022 18:05:35 - INFO - codeparrot_training - Step 49812: {'lr': 1.8925070356667907e-08, 'samples': 25504256, 'steps': 49812, 'loss/train': 1.0794496536254883} 02/26/2022 18:05:40 - INFO - codeparrot_training - Step 49813: {'lr': 1.8724277781045906e-08, 'samples': 25504768, 'steps': 49813, 'loss/train': 0.8234782814979553} 02/26/2022 18:05:44 - INFO - codeparrot_training - Step 49814: {'lr': 1.8524556045840245e-08, 'samples': 25505280, 'steps': 49814, 'loss/train': 1.2747784852981567} 02/26/2022 18:05:49 - INFO - codeparrot_training - Step 49815: {'lr': 1.8325905151883593e-08, 'samples': 25505792, 'steps': 49815, 'loss/train': 2.13350510597229} 02/26/2022 18:05:53 - INFO - codeparrot_training - Step 49816: {'lr': 1.8128325100036367e-08, 'samples': 25506304, 'steps': 49816, 'loss/train': 2.030944585800171} 02/26/2022 18:05:59 - INFO - codeparrot_training - Step 49817: {'lr': 1.7931815891158998e-08, 'samples': 25506816, 'steps': 49817, 'loss/train': 4.86380672454834} 02/26/2022 18:06:02 - INFO - codeparrot_training - Step 49818: {'lr': 1.7736377526084145e-08, 'samples': 25507328, 'steps': 49818, 'loss/train': 2.0606706142425537} 02/26/2022 18:06:08 - INFO - codeparrot_training - Step 49819: {'lr': 1.754201000564448e-08, 'samples': 25507840, 'steps': 49819, 'loss/train': 2.158308744430542} 02/26/2022 18:06:12 - INFO - codeparrot_training - Step 49820: {'lr': 1.7348713330672673e-08, 'samples': 25508352, 'steps': 49820, 'loss/train': 1.022890329360962} 02/26/2022 18:06:17 - INFO - codeparrot_training - Step 49821: {'lr': 1.7156487502001383e-08, 'samples': 25508864, 'steps': 49821, 'loss/train': 1.3499585390090942} 02/26/2022 18:06:21 - INFO - codeparrot_training - Step 49822: {'lr': 1.6965332520463282e-08, 'samples': 25509376, 'steps': 49822, 'loss/train': 1.7400176525115967} 02/26/2022 18:06:26 - INFO - codeparrot_training - Step 49823: {'lr': 1.6775248386863283e-08, 'samples': 25509888, 'steps': 49823, 'loss/train': 1.4092994928359985} 02/26/2022 18:06:30 - INFO - codeparrot_training - Step 49824: {'lr': 1.6586235102006298e-08, 'samples': 25510400, 'steps': 49824, 'loss/train': 2.208158254623413} 02/26/2022 18:06:35 - INFO - codeparrot_training - Step 49825: {'lr': 1.639829266672499e-08, 'samples': 25510912, 'steps': 49825, 'loss/train': 1.0031169652938843} 02/26/2022 18:06:39 - INFO - codeparrot_training - Step 49826: {'lr': 1.6211421081824273e-08, 'samples': 25511424, 'steps': 49826, 'loss/train': 1.2313945293426514} 02/26/2022 18:06:44 - INFO - codeparrot_training - Step 49827: {'lr': 1.6025620348081303e-08, 'samples': 25511936, 'steps': 49827, 'loss/train': 2.046518087387085} 02/26/2022 18:06:48 - INFO - codeparrot_training - Step 49828: {'lr': 1.5840890466300995e-08, 'samples': 25512448, 'steps': 49828, 'loss/train': 1.7530125379562378} 02/26/2022 18:06:53 - INFO - codeparrot_training - Step 49829: {'lr': 1.5657231437288257e-08, 'samples': 25512960, 'steps': 49829, 'loss/train': 1.495530128479004} 02/26/2022 18:06:57 - INFO - codeparrot_training - Step 49830: {'lr': 1.5474643261820243e-08, 'samples': 25513472, 'steps': 49830, 'loss/train': 1.5152678489685059} 02/26/2022 18:07:02 - INFO - codeparrot_training - Step 49831: {'lr': 1.5293125940701868e-08, 'samples': 25513984, 'steps': 49831, 'loss/train': 0.8266059160232544} 02/26/2022 18:07:06 - INFO - codeparrot_training - Step 49832: {'lr': 1.511267947465478e-08, 'samples': 25514496, 'steps': 49832, 'loss/train': 1.1114251613616943} 02/26/2022 18:07:12 - INFO - codeparrot_training - Step 49833: {'lr': 1.4933303864511637e-08, 'samples': 25515008, 'steps': 49833, 'loss/train': 1.2117236852645874} 02/26/2022 18:07:15 - INFO - codeparrot_training - Step 49834: {'lr': 1.4754999110994094e-08, 'samples': 25515520, 'steps': 49834, 'loss/train': 0.901160478591919} 02/26/2022 18:07:21 - INFO - codeparrot_training - Step 49835: {'lr': 1.4577765214907056e-08, 'samples': 25516032, 'steps': 49835, 'loss/train': 1.5964235067367554} 02/26/2022 18:07:24 - INFO - codeparrot_training - Step 49836: {'lr': 1.4401602176972173e-08, 'samples': 25516544, 'steps': 49836, 'loss/train': 1.9681991338729858} 02/26/2022 18:07:30 - INFO - codeparrot_training - Step 49837: {'lr': 1.4226509997966596e-08, 'samples': 25517056, 'steps': 49837, 'loss/train': 1.6212447881698608} 02/26/2022 18:07:33 - INFO - codeparrot_training - Step 49838: {'lr': 1.405248867863973e-08, 'samples': 25517568, 'steps': 49838, 'loss/train': 0.8212572336196899} 02/26/2022 18:07:39 - INFO - codeparrot_training - Step 49839: {'lr': 1.3879538219713217e-08, 'samples': 25518080, 'steps': 49839, 'loss/train': 1.2232691049575806} 02/26/2022 18:07:42 - INFO - codeparrot_training - Step 49840: {'lr': 1.3707658621964214e-08, 'samples': 25518592, 'steps': 49840, 'loss/train': 1.354844570159912} 02/26/2022 18:07:48 - INFO - codeparrot_training - Step 49841: {'lr': 1.3536849886114366e-08, 'samples': 25519104, 'steps': 49841, 'loss/train': 1.6901588439941406} 02/26/2022 18:07:52 - INFO - codeparrot_training - Step 49842: {'lr': 1.3367112012885319e-08, 'samples': 25519616, 'steps': 49842, 'loss/train': 8.419468879699707} 02/26/2022 18:07:58 - INFO - codeparrot_training - Step 49843: {'lr': 1.3198445003026472e-08, 'samples': 25520128, 'steps': 49843, 'loss/train': 2.186472177505493} 02/26/2022 18:08:01 - INFO - codeparrot_training - Step 49844: {'lr': 1.3030848857231714e-08, 'samples': 25520640, 'steps': 49844, 'loss/train': 1.437972068786621} 02/26/2022 18:08:07 - INFO - codeparrot_training - Step 49845: {'lr': 1.2864323576222692e-08, 'samples': 25521152, 'steps': 49845, 'loss/train': 1.9912728071212769} 02/26/2022 18:08:10 - INFO - codeparrot_training - Step 49846: {'lr': 1.2698869160748806e-08, 'samples': 25521664, 'steps': 49846, 'loss/train': 0.8067148327827454} 02/26/2022 18:08:16 - INFO - codeparrot_training - Step 49847: {'lr': 1.2534485611476187e-08, 'samples': 25522176, 'steps': 49847, 'loss/train': 1.9356513023376465} 02/26/2022 18:08:19 - INFO - codeparrot_training - Step 49848: {'lr': 1.2371172929126484e-08, 'samples': 25522688, 'steps': 49848, 'loss/train': 2.290548324584961} 02/26/2022 18:08:25 - INFO - codeparrot_training - Step 49849: {'lr': 1.2208931114393584e-08, 'samples': 25523200, 'steps': 49849, 'loss/train': 1.7514867782592773} 02/26/2022 18:08:28 - INFO - codeparrot_training - Step 49850: {'lr': 1.2047760167999133e-08, 'samples': 25523712, 'steps': 49850, 'loss/train': 1.7774622440338135} 02/26/2022 18:08:34 - INFO - codeparrot_training - Step 49851: {'lr': 1.1887660090609264e-08, 'samples': 25524224, 'steps': 49851, 'loss/train': 1.676344394683838} 02/26/2022 18:08:37 - INFO - codeparrot_training - Step 49852: {'lr': 1.1728630882890112e-08, 'samples': 25524736, 'steps': 49852, 'loss/train': 2.046475648880005} 02/26/2022 18:08:44 - INFO - codeparrot_training - Step 49853: {'lr': 1.1570672545563321e-08, 'samples': 25525248, 'steps': 49853, 'loss/train': 1.997458815574646} 02/26/2022 18:08:47 - INFO - codeparrot_training - Step 49854: {'lr': 1.1413785079267269e-08, 'samples': 25525760, 'steps': 49854, 'loss/train': 1.489466667175293} 02/26/2022 18:08:53 - INFO - codeparrot_training - Step 49855: {'lr': 1.1257968484723602e-08, 'samples': 25526272, 'steps': 49855, 'loss/train': 2.143293857574463} 02/26/2022 18:08:56 - INFO - codeparrot_training - Step 49856: {'lr': 1.1103222762542941e-08, 'samples': 25526784, 'steps': 49856, 'loss/train': 1.6462370157241821} 02/26/2022 18:09:02 - INFO - codeparrot_training - Step 49857: {'lr': 1.0949547913446933e-08, 'samples': 25527296, 'steps': 49857, 'loss/train': 2.2388648986816406} 02/26/2022 18:09:05 - INFO - codeparrot_training - Step 49858: {'lr': 1.0796943938018445e-08, 'samples': 25527808, 'steps': 49858, 'loss/train': 2.399623155593872} 02/26/2022 18:09:11 - INFO - codeparrot_training - Step 49859: {'lr': 1.0645410837006875e-08, 'samples': 25528320, 'steps': 49859, 'loss/train': 1.6432126760482788} 02/26/2022 18:09:14 - INFO - codeparrot_training - Step 49860: {'lr': 1.0494948610967337e-08, 'samples': 25528832, 'steps': 49860, 'loss/train': 1.7454429864883423} 02/26/2022 18:09:20 - INFO - codeparrot_training - Step 49861: {'lr': 1.034555726059372e-08, 'samples': 25529344, 'steps': 49861, 'loss/train': 2.9228708744049072} 02/26/2022 18:09:23 - INFO - codeparrot_training - Step 49862: {'lr': 1.0197236786552155e-08, 'samples': 25529856, 'steps': 49862, 'loss/train': 1.7823240756988525} 02/26/2022 18:09:30 - INFO - codeparrot_training - Step 49863: {'lr': 1.0049987189397758e-08, 'samples': 25530368, 'steps': 49863, 'loss/train': 1.155539870262146} 02/26/2022 18:09:33 - INFO - codeparrot_training - Step 49864: {'lr': 9.90380846985217e-09, 'samples': 25530880, 'steps': 49864, 'loss/train': 1.9393876791000366} 02/26/2022 18:09:37 - INFO - codeparrot_training - Step 49865: {'lr': 9.758700628470507e-09, 'samples': 25531392, 'steps': 49865, 'loss/train': 0.8131659030914307} 02/26/2022 18:09:42 - INFO - codeparrot_training - Step 49866: {'lr': 9.614663665918899e-09, 'samples': 25531904, 'steps': 49866, 'loss/train': 1.6611566543579102} 02/26/2022 18:09:46 - INFO - codeparrot_training - Step 49867: {'lr': 9.471697582780214e-09, 'samples': 25532416, 'steps': 49867, 'loss/train': 2.2808587551116943} 02/26/2022 18:09:51 - INFO - codeparrot_training - Step 49868: {'lr': 9.329802379692831e-09, 'samples': 25532928, 'steps': 49868, 'loss/train': 2.0008318424224854} 02/26/2022 18:09:55 - INFO - codeparrot_training - Step 49869: {'lr': 9.188978057239616e-09, 'samples': 25533440, 'steps': 49869, 'loss/train': 1.3950893878936768} 02/26/2022 18:10:00 - INFO - codeparrot_training - Step 49870: {'lr': 9.049224616058948e-09, 'samples': 25533952, 'steps': 49870, 'loss/train': 2.2258694171905518} 02/26/2022 18:10:04 - INFO - codeparrot_training - Step 49871: {'lr': 8.910542056733695e-09, 'samples': 25534464, 'steps': 49871, 'loss/train': 1.170242190361023} 02/26/2022 18:10:09 - INFO - codeparrot_training - Step 49872: {'lr': 8.772930379846722e-09, 'samples': 25534976, 'steps': 49872, 'loss/train': 2.095367193222046} 02/26/2022 18:10:13 - INFO - codeparrot_training - Step 49873: {'lr': 8.636389585980897e-09, 'samples': 25535488, 'steps': 49873, 'loss/train': 0.40433958172798157} 02/26/2022 18:10:19 - INFO - codeparrot_training - Step 49874: {'lr': 8.500919675746843e-09, 'samples': 25536000, 'steps': 49874, 'loss/train': 0.24236752092838287} 02/26/2022 18:10:22 - INFO - codeparrot_training - Step 49875: {'lr': 8.366520649727427e-09, 'samples': 25536512, 'steps': 49875, 'loss/train': 2.23716402053833} 02/26/2022 18:10:28 - INFO - codeparrot_training - Step 49876: {'lr': 8.233192508477761e-09, 'samples': 25537024, 'steps': 49876, 'loss/train': 1.3792365789413452} 02/26/2022 18:10:31 - INFO - codeparrot_training - Step 49877: {'lr': 8.100935252552955e-09, 'samples': 25537536, 'steps': 49877, 'loss/train': 1.2408055067062378} 02/26/2022 18:10:37 - INFO - codeparrot_training - Step 49878: {'lr': 7.969748882563632e-09, 'samples': 25538048, 'steps': 49878, 'loss/train': 1.9493943452835083} 02/26/2022 18:10:40 - INFO - codeparrot_training - Step 49879: {'lr': 7.839633399064905e-09, 'samples': 25538560, 'steps': 49879, 'loss/train': 1.695171594619751} 02/26/2022 18:10:46 - INFO - codeparrot_training - Step 49880: {'lr': 7.710588802584129e-09, 'samples': 25539072, 'steps': 49880, 'loss/train': 2.728811740875244} 02/26/2022 18:10:49 - INFO - codeparrot_training - Step 49881: {'lr': 7.582615093676415e-09, 'samples': 25539584, 'steps': 49881, 'loss/train': 0.7754898071289062} 02/26/2022 18:10:55 - INFO - codeparrot_training - Step 49882: {'lr': 7.455712272924632e-09, 'samples': 25540096, 'steps': 49882, 'loss/train': 1.793603539466858} 02/26/2022 18:10:58 - INFO - codeparrot_training - Step 49883: {'lr': 7.329880340828377e-09, 'samples': 25540608, 'steps': 49883, 'loss/train': 1.2226747274398804} 02/26/2022 18:11:04 - INFO - codeparrot_training - Step 49884: {'lr': 7.20511929797052e-09, 'samples': 25541120, 'steps': 49884, 'loss/train': 1.2711772918701172} 02/26/2022 18:11:07 - INFO - codeparrot_training - Step 49885: {'lr': 7.081429144850659e-09, 'samples': 25541632, 'steps': 49885, 'loss/train': 1.431278109550476} 02/26/2022 18:11:13 - INFO - codeparrot_training - Step 49886: {'lr': 6.958809882023909e-09, 'samples': 25542144, 'steps': 49886, 'loss/train': 2.0990211963653564} 02/26/2022 18:11:16 - INFO - codeparrot_training - Step 49887: {'lr': 6.837261509989867e-09, 'samples': 25542656, 'steps': 49887, 'loss/train': 1.3828198909759521} 02/26/2022 18:11:22 - INFO - codeparrot_training - Step 49888: {'lr': 6.716784029303646e-09, 'samples': 25543168, 'steps': 49888, 'loss/train': 2.3065266609191895} 02/26/2022 18:11:25 - INFO - codeparrot_training - Step 49889: {'lr': 6.597377440437091e-09, 'samples': 25543680, 'steps': 49889, 'loss/train': 2.7996435165405273} 02/26/2022 18:11:32 - INFO - codeparrot_training - Step 49890: {'lr': 6.479041743945313e-09, 'samples': 25544192, 'steps': 49890, 'loss/train': 0.8790472149848938} 02/26/2022 18:11:36 - INFO - codeparrot_training - Step 49891: {'lr': 6.361776940300157e-09, 'samples': 25544704, 'steps': 49891, 'loss/train': 1.6156212091445923} 02/26/2022 18:11:41 - INFO - codeparrot_training - Step 49892: {'lr': 6.2455830300289785e-09, 'samples': 25545216, 'steps': 49892, 'loss/train': 0.49404293298721313} 02/26/2022 18:11:44 - INFO - codeparrot_training - Step 49893: {'lr': 6.1304600136036225e-09, 'samples': 25545728, 'steps': 49893, 'loss/train': 1.9654839038848877} 02/26/2022 18:11:50 - INFO - codeparrot_training - Step 49894: {'lr': 6.016407891551445e-09, 'samples': 25546240, 'steps': 49894, 'loss/train': 1.9663312435150146} 02/26/2022 18:11:54 - INFO - codeparrot_training - Step 49895: {'lr': 5.9034266643442915e-09, 'samples': 25546752, 'steps': 49895, 'loss/train': 1.3407528400421143} 02/26/2022 18:11:59 - INFO - codeparrot_training - Step 49896: {'lr': 5.791516332454006e-09, 'samples': 25547264, 'steps': 49896, 'loss/train': 1.2587236166000366} 02/26/2022 18:12:02 - INFO - codeparrot_training - Step 49897: {'lr': 5.680676896380188e-09, 'samples': 25547776, 'steps': 49897, 'loss/train': 2.067148208618164} 02/26/2022 18:12:08 - INFO - codeparrot_training - Step 49898: {'lr': 5.570908356566929e-09, 'samples': 25548288, 'steps': 49898, 'loss/train': 1.4106074571609497} 02/26/2022 18:12:11 - INFO - codeparrot_training - Step 49899: {'lr': 5.462210713513826e-09, 'samples': 25548800, 'steps': 49899, 'loss/train': 1.870240569114685} 02/26/2022 18:12:18 - INFO - codeparrot_training - Step 49900: {'lr': 5.354583967692728e-09, 'samples': 25549312, 'steps': 49900, 'loss/train': 1.6069060564041138} 02/26/2022 18:12:22 - INFO - codeparrot_training - Step 49901: {'lr': 5.248028119547721e-09, 'samples': 25549824, 'steps': 49901, 'loss/train': 1.6345579624176025} 02/26/2022 18:12:27 - INFO - codeparrot_training - Step 49902: {'lr': 5.1425431695506504e-09, 'samples': 25550336, 'steps': 49902, 'loss/train': 1.5215963125228882} 02/26/2022 18:12:30 - INFO - codeparrot_training - Step 49903: {'lr': 5.038129118117851e-09, 'samples': 25550848, 'steps': 49903, 'loss/train': 1.8097896575927734} 02/26/2022 18:12:36 - INFO - codeparrot_training - Step 49904: {'lr': 4.934785965721167e-09, 'samples': 25551360, 'steps': 49904, 'loss/train': 3.924865961074829} 02/26/2022 18:12:40 - INFO - codeparrot_training - Step 49905: {'lr': 4.832513712804687e-09, 'samples': 25551872, 'steps': 49905, 'loss/train': 2.7155182361602783} 02/26/2022 18:12:45 - INFO - codeparrot_training - Step 49906: {'lr': 4.731312359812501e-09, 'samples': 25552384, 'steps': 49906, 'loss/train': 1.0408412218093872} 02/26/2022 18:12:49 - INFO - codeparrot_training - Step 49907: {'lr': 4.631181907160942e-09, 'samples': 25552896, 'steps': 49907, 'loss/train': 1.5556855201721191} 02/26/2022 18:12:54 - INFO - codeparrot_training - Step 49908: {'lr': 4.5321223552663436e-09, 'samples': 25553408, 'steps': 49908, 'loss/train': 2.84206223487854} 02/26/2022 18:12:58 - INFO - codeparrot_training - Step 49909: {'lr': 4.434133704600552e-09, 'samples': 25553920, 'steps': 49909, 'loss/train': 2.1844868659973145} 02/26/2022 18:13:04 - INFO - codeparrot_training - Step 49910: {'lr': 4.3372159555521426e-09, 'samples': 25554432, 'steps': 49910, 'loss/train': 2.244971513748169} 02/26/2022 18:13:08 - INFO - codeparrot_training - Step 49911: {'lr': 4.241369108537452e-09, 'samples': 25554944, 'steps': 49911, 'loss/train': 1.7292194366455078} 02/26/2022 18:13:13 - INFO - codeparrot_training - Step 49912: {'lr': 4.146593163972812e-09, 'samples': 25555456, 'steps': 49912, 'loss/train': 2.6721560955047607} 02/26/2022 18:13:17 - INFO - codeparrot_training - Step 49913: {'lr': 4.052888122246801e-09, 'samples': 25555968, 'steps': 49913, 'loss/train': 0.9000661373138428} 02/26/2022 18:13:22 - INFO - codeparrot_training - Step 49914: {'lr': 3.960253983803508e-09, 'samples': 25556480, 'steps': 49914, 'loss/train': 1.0378044843673706} 02/26/2022 18:13:26 - INFO - codeparrot_training - Step 49915: {'lr': 3.868690749003756e-09, 'samples': 25556992, 'steps': 49915, 'loss/train': 1.2975900173187256} 02/26/2022 18:13:32 - INFO - codeparrot_training - Step 49916: {'lr': 3.7781984182361226e-09, 'samples': 25557504, 'steps': 49916, 'loss/train': 1.4350013732910156} 02/26/2022 18:13:35 - INFO - codeparrot_training - Step 49917: {'lr': 3.688776991889187e-09, 'samples': 25558016, 'steps': 49917, 'loss/train': 2.298407793045044} 02/26/2022 18:13:41 - INFO - codeparrot_training - Step 49918: {'lr': 3.600426470379281e-09, 'samples': 25558528, 'steps': 49918, 'loss/train': 1.7967218160629272} 02/26/2022 18:13:44 - INFO - codeparrot_training - Step 49919: {'lr': 3.5131468540672283e-09, 'samples': 25559040, 'steps': 49919, 'loss/train': 1.8604594469070435} 02/26/2022 18:13:49 - INFO - codeparrot_training - Step 49920: {'lr': 3.4269381433138515e-09, 'samples': 25559552, 'steps': 49920, 'loss/train': 2.1783883571624756} 02/26/2022 18:13:53 - INFO - codeparrot_training - Step 49921: {'lr': 3.341800338479972e-09, 'samples': 25560064, 'steps': 49921, 'loss/train': 0.9122430682182312} 02/26/2022 18:14:00 - INFO - codeparrot_training - Step 49922: {'lr': 3.2577334399541693e-09, 'samples': 25560576, 'steps': 49922, 'loss/train': 0.28049030900001526} 02/26/2022 18:14:03 - INFO - codeparrot_training - Step 49923: {'lr': 3.1747374480972647e-09, 'samples': 25561088, 'steps': 49923, 'loss/train': 0.6618022322654724} 02/26/2022 18:14:09 - INFO - codeparrot_training - Step 49924: {'lr': 3.092812363270081e-09, 'samples': 25561600, 'steps': 49924, 'loss/train': 1.9529082775115967} 02/26/2022 18:14:12 - INFO - codeparrot_training - Step 49925: {'lr': 3.01195818577793e-09, 'samples': 25562112, 'steps': 49925, 'loss/train': 1.6398659944534302} 02/26/2022 18:14:18 - INFO - codeparrot_training - Step 49926: {'lr': 2.932174916009389e-09, 'samples': 25562624, 'steps': 49926, 'loss/train': 1.4235341548919678} 02/26/2022 18:14:21 - INFO - codeparrot_training - Step 49927: {'lr': 2.8534625543252813e-09, 'samples': 25563136, 'steps': 49927, 'loss/train': 1.7917429208755493} 02/26/2022 18:14:27 - INFO - codeparrot_training - Step 49928: {'lr': 2.775821101003162e-09, 'samples': 25563648, 'steps': 49928, 'loss/train': 2.244459629058838} 02/26/2022 18:14:30 - INFO - codeparrot_training - Step 49929: {'lr': 2.699250556403854e-09, 'samples': 25564160, 'steps': 49929, 'loss/train': 1.7520989179611206} 02/26/2022 18:14:36 - INFO - codeparrot_training - Step 49930: {'lr': 2.6237509208881795e-09, 'samples': 25564672, 'steps': 49930, 'loss/train': 1.5223432779312134} 02/26/2022 18:14:39 - INFO - codeparrot_training - Step 49931: {'lr': 2.549322194733694e-09, 'samples': 25565184, 'steps': 49931, 'loss/train': 1.947978138923645} 02/26/2022 18:14:45 - INFO - codeparrot_training - Step 49932: {'lr': 2.4759643782734655e-09, 'samples': 25565696, 'steps': 49932, 'loss/train': 0.6739815473556519} 02/26/2022 18:14:48 - INFO - codeparrot_training - Step 49933: {'lr': 2.403677471812804e-09, 'samples': 25566208, 'steps': 49933, 'loss/train': 2.4989287853240967} 02/26/2022 18:14:54 - INFO - codeparrot_training - Step 49934: {'lr': 2.3324614756847774e-09, 'samples': 25566720, 'steps': 49934, 'loss/train': 1.0190621614456177} 02/26/2022 18:14:57 - INFO - codeparrot_training - Step 49935: {'lr': 2.2623163901669407e-09, 'samples': 25567232, 'steps': 49935, 'loss/train': 0.25735238194465637} 02/26/2022 18:15:04 - INFO - codeparrot_training - Step 49936: {'lr': 2.1932422155923616e-09, 'samples': 25567744, 'steps': 49936, 'loss/train': 0.910692036151886} 02/26/2022 18:15:07 - INFO - codeparrot_training - Step 49937: {'lr': 2.1252389522108396e-09, 'samples': 25568256, 'steps': 49937, 'loss/train': 1.597078800201416} 02/26/2022 18:15:13 - INFO - codeparrot_training - Step 49938: {'lr': 2.0583066003831975e-09, 'samples': 25568768, 'steps': 49938, 'loss/train': 2.289205312728882} 02/26/2022 18:15:16 - INFO - codeparrot_training - Step 49939: {'lr': 1.99244516033148e-09, 'samples': 25569280, 'steps': 49939, 'loss/train': 0.7344565987586975} 02/26/2022 18:15:20 - INFO - codeparrot_training - Step 49940: {'lr': 1.927654632360998e-09, 'samples': 25569792, 'steps': 49940, 'loss/train': 0.9511494040489197} 02/26/2022 18:15:25 - INFO - codeparrot_training - Step 49941: {'lr': 1.8639350167493074e-09, 'samples': 25570304, 'steps': 49941, 'loss/train': 1.2779861688613892} 02/26/2022 18:15:29 - INFO - codeparrot_training - Step 49942: {'lr': 1.8012863137462088e-09, 'samples': 25570816, 'steps': 49942, 'loss/train': 1.6683646440505981} 02/26/2022 18:15:34 - INFO - codeparrot_training - Step 49943: {'lr': 1.7397085236847687e-09, 'samples': 25571328, 'steps': 49943, 'loss/train': 1.4545079469680786} 02/26/2022 18:15:38 - INFO - codeparrot_training - Step 49944: {'lr': 1.6792016467592764e-09, 'samples': 25571840, 'steps': 49944, 'loss/train': 1.3636372089385986} 02/26/2022 18:15:43 - INFO - codeparrot_training - Step 49945: {'lr': 1.6197656832750428e-09, 'samples': 25572352, 'steps': 49945, 'loss/train': 1.4487106800079346} 02/26/2022 18:15:47 - INFO - codeparrot_training - Step 49946: {'lr': 1.561400633454113e-09, 'samples': 25572864, 'steps': 49946, 'loss/train': 2.730422019958496} 02/26/2022 18:15:53 - INFO - codeparrot_training - Step 49947: {'lr': 1.5041064975740426e-09, 'samples': 25573376, 'steps': 49947, 'loss/train': 1.2567530870437622} 02/26/2022 18:15:56 - INFO - codeparrot_training - Step 49948: {'lr': 1.4478832758568762e-09, 'samples': 25573888, 'steps': 49948, 'loss/train': 0.8132756948471069} 02/26/2022 18:16:02 - INFO - codeparrot_training - Step 49949: {'lr': 1.3927309685524136e-09, 'samples': 25574400, 'steps': 49949, 'loss/train': 1.624342679977417} 02/26/2022 18:16:06 - INFO - codeparrot_training - Step 49950: {'lr': 1.3386495758827e-09, 'samples': 25574912, 'steps': 49950, 'loss/train': 1.123450756072998} 02/26/2022 18:16:11 - INFO - codeparrot_training - Step 49951: {'lr': 1.2856390981252907e-09, 'samples': 25575424, 'steps': 49951, 'loss/train': 5.038472652435303} 02/26/2022 18:16:15 - INFO - codeparrot_training - Step 49952: {'lr': 1.2336995354467196e-09, 'samples': 25575936, 'steps': 49952, 'loss/train': 1.3611881732940674} 02/26/2022 18:16:20 - INFO - codeparrot_training - Step 49953: {'lr': 1.182830888124542e-09, 'samples': 25576448, 'steps': 49953, 'loss/train': 1.7058833837509155} 02/26/2022 18:16:24 - INFO - codeparrot_training - Step 49954: {'lr': 1.133033156353047e-09, 'samples': 25576960, 'steps': 49954, 'loss/train': 1.687996745109558} 02/26/2022 18:16:29 - INFO - codeparrot_training - Step 49955: {'lr': 1.0843063403265242e-09, 'samples': 25577472, 'steps': 49955, 'loss/train': 2.9062206745147705} 02/26/2022 18:16:33 - INFO - codeparrot_training - Step 49956: {'lr': 1.0366504402947729e-09, 'samples': 25577984, 'steps': 49956, 'loss/train': 0.9279352426528931} 02/26/2022 18:16:39 - INFO - codeparrot_training - Step 49957: {'lr': 9.900654563965716e-10, 'samples': 25578496, 'steps': 49957, 'loss/train': 0.7256073355674744} 02/26/2022 18:16:43 - INFO - codeparrot_training - Step 49958: {'lr': 9.445513889094759e-10, 'samples': 25579008, 'steps': 49958, 'loss/train': 0.22750012576580048} 02/26/2022 18:16:49 - INFO - codeparrot_training - Step 49959: {'lr': 9.001082379722636e-10, 'samples': 25579520, 'steps': 49959, 'loss/train': 1.4443467855453491} 02/26/2022 18:16:52 - INFO - codeparrot_training - Step 49960: {'lr': 8.567360038069794e-10, 'samples': 25580032, 'steps': 49960, 'loss/train': 1.0324279069900513} 02/26/2022 18:16:56 - INFO - codeparrot_training - Step 49961: {'lr': 8.144346866079122e-10, 'samples': 25580544, 'steps': 49961, 'loss/train': 1.933714509010315} 02/26/2022 18:17:01 - INFO - codeparrot_training - Step 49962: {'lr': 7.7320428651384e-10, 'samples': 25581056, 'steps': 49962, 'loss/train': 2.138439178466797} 02/26/2022 18:17:05 - INFO - codeparrot_training - Step 49963: {'lr': 7.330448037190518e-10, 'samples': 25581568, 'steps': 49963, 'loss/train': 2.0313007831573486} 02/26/2022 18:17:11 - INFO - codeparrot_training - Step 49964: {'lr': 6.939562383900811e-10, 'samples': 25582080, 'steps': 49964, 'loss/train': 1.2096962928771973} 02/26/2022 18:17:14 - INFO - codeparrot_training - Step 49965: {'lr': 6.559385907212167e-10, 'samples': 25582592, 'steps': 49965, 'loss/train': 0.6134408712387085} 02/26/2022 18:17:18 - INFO - codeparrot_training - Step 49966: {'lr': 6.189918608789924e-10, 'samples': 25583104, 'steps': 49966, 'loss/train': 1.0768870115280151} 02/26/2022 18:17:23 - INFO - codeparrot_training - Step 49967: {'lr': 5.831160489744303e-10, 'samples': 25583616, 'steps': 49967, 'loss/train': 2.3901000022888184} 02/26/2022 18:17:27 - INFO - codeparrot_training - Step 49968: {'lr': 5.483111551740638e-10, 'samples': 25584128, 'steps': 49968, 'loss/train': 1.9718189239501953} 02/26/2022 18:17:32 - INFO - codeparrot_training - Step 49969: {'lr': 5.145771796721821e-10, 'samples': 25584640, 'steps': 49969, 'loss/train': 1.3913748264312744} 02/26/2022 18:17:36 - INFO - codeparrot_training - Step 49970: {'lr': 4.819141225798074e-10, 'samples': 25585152, 'steps': 49970, 'loss/train': 1.8307862281799316} 02/26/2022 18:17:42 - INFO - codeparrot_training - Step 49971: {'lr': 4.50321984007962e-10, 'samples': 25585664, 'steps': 49971, 'loss/train': 1.5337656736373901} 02/26/2022 18:17:45 - INFO - codeparrot_training - Step 49972: {'lr': 4.198007641509349e-10, 'samples': 25586176, 'steps': 49972, 'loss/train': 2.0893256664276123} 02/26/2022 18:17:51 - INFO - codeparrot_training - Step 49973: {'lr': 3.903504631197485e-10, 'samples': 25586688, 'steps': 49973, 'loss/train': 1.5156900882720947} 02/26/2022 18:17:55 - INFO - codeparrot_training - Step 49974: {'lr': 3.619710809976695e-10, 'samples': 25587200, 'steps': 49974, 'loss/train': 1.4385193586349487} 02/26/2022 18:18:00 - INFO - codeparrot_training - Step 49975: {'lr': 3.346626179789869e-10, 'samples': 25587712, 'steps': 49975, 'loss/train': 1.695233941078186} 02/26/2022 18:18:04 - INFO - codeparrot_training - Step 49976: {'lr': 3.0842507411921185e-10, 'samples': 25588224, 'steps': 49976, 'loss/train': 1.9455724954605103} 02/26/2022 18:18:09 - INFO - codeparrot_training - Step 49977: {'lr': 2.832584495571222e-10, 'samples': 25588736, 'steps': 49977, 'loss/train': 1.5164356231689453} 02/26/2022 18:18:12 - INFO - codeparrot_training - Step 49978: {'lr': 2.5916274440374034e-10, 'samples': 25589248, 'steps': 49978, 'loss/train': 2.4412038326263428} 02/26/2022 18:18:18 - INFO - codeparrot_training - Step 49979: {'lr': 2.361379587423329e-10, 'samples': 25589760, 'steps': 49979, 'loss/train': 0.5632081031799316} 02/26/2022 18:18:21 - INFO - codeparrot_training - Step 49980: {'lr': 2.1418409271167783e-10, 'samples': 25590272, 'steps': 49980, 'loss/train': 2.2746522426605225} 02/26/2022 18:18:27 - INFO - codeparrot_training - Step 49981: {'lr': 1.9330114636728623e-10, 'samples': 25590784, 'steps': 49981, 'loss/train': 2.490490198135376} 02/26/2022 18:18:30 - INFO - codeparrot_training - Step 49982: {'lr': 1.734891197924249e-10, 'samples': 25591296, 'steps': 49982, 'loss/train': 1.5169414281845093} 02/26/2022 18:18:36 - INFO - codeparrot_training - Step 49983: {'lr': 1.5474801309811605e-10, 'samples': 25591808, 'steps': 49983, 'loss/train': 1.9936089515686035} 02/26/2022 18:18:39 - INFO - codeparrot_training - Step 49984: {'lr': 1.3707782636762644e-10, 'samples': 25592320, 'steps': 49984, 'loss/train': 1.350411057472229} 02/26/2022 18:18:45 - INFO - codeparrot_training - Step 49985: {'lr': 1.2047855968422282e-10, 'samples': 25592832, 'steps': 49985, 'loss/train': 1.2329177856445312} 02/26/2022 18:18:48 - INFO - codeparrot_training - Step 49986: {'lr': 1.0495021307566077e-10, 'samples': 25593344, 'steps': 49986, 'loss/train': 0.12649674713611603} 02/26/2022 18:18:54 - INFO - codeparrot_training - Step 49987: {'lr': 9.049278662520699e-11, 'samples': 25593856, 'steps': 49987, 'loss/train': 1.775930643081665} 02/26/2022 18:18:58 - INFO - codeparrot_training - Step 49988: {'lr': 7.710628041612822e-11, 'samples': 25594368, 'steps': 49988, 'loss/train': 1.977392554283142} 02/26/2022 18:19:03 - INFO - codeparrot_training - Step 49989: {'lr': 6.47906945039356e-11, 'samples': 25594880, 'steps': 49989, 'loss/train': 2.5542259216308594} 02/26/2022 18:19:07 - INFO - codeparrot_training - Step 49990: {'lr': 5.354602891638471e-11, 'samples': 25595392, 'steps': 49990, 'loss/train': 0.2777775526046753} 02/26/2022 18:19:12 - INFO - codeparrot_training - Step 49991: {'lr': 4.3372283708986714e-11, 'samples': 25595904, 'steps': 49991, 'loss/train': 1.014872431755066} 02/26/2022 18:19:16 - INFO - codeparrot_training - Step 49992: {'lr': 3.4269458937252754e-11, 'samples': 25596416, 'steps': 49992, 'loss/train': 1.938048005104065} 02/26/2022 18:19:21 - INFO - codeparrot_training - Step 49993: {'lr': 2.6237554656693975e-11, 'samples': 25596928, 'steps': 49993, 'loss/train': 0.7541128993034363} 02/26/2022 18:19:25 - INFO - codeparrot_training - Step 49994: {'lr': 1.927657083955481e-11, 'samples': 25597440, 'steps': 49994, 'loss/train': 2.6811628341674805} 02/26/2022 18:19:31 - INFO - codeparrot_training - Step 49995: {'lr': 1.3386507596857556e-11, 'samples': 25597952, 'steps': 49995, 'loss/train': 1.5248055458068848} 02/26/2022 18:19:34 - INFO - codeparrot_training - Step 49996: {'lr': 8.567364873091067e-12, 'samples': 25598464, 'steps': 49996, 'loss/train': 1.6005619764328003} 02/26/2022 18:19:40 - INFO - codeparrot_training - Step 49997: {'lr': 4.819142751522065e-12, 'samples': 25598976, 'steps': 49997, 'loss/train': 1.537976622581482} 02/26/2022 18:19:44 - INFO - codeparrot_training - Step 49998: {'lr': 2.1418412321505543e-12, 'samples': 25599488, 'steps': 49998, 'loss/train': 2.184359073638916} 02/26/2022 18:19:49 - INFO - codeparrot_training - Step 49999: {'lr': 5.354603149765325e-13, 'samples': 25600000, 'steps': 49999, 'loss/train': 2.408355236053467} 02/26/2022 18:19:49 - INFO - codeparrot_training - Evaluating and saving model checkpoint 02/26/2022 18:20:07 - WARNING - huggingface_hub.repository - Several commits (50) will be pushed upstream. 02/26/2022 18:20:07 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. 02/26/2022 18:20:52 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/multi-code-clippy 508ef45..5aaf99c floral-grass-11 -> floral-grass-11 02/26/2022 18:20:53 - INFO - codeparrot_training - Evaluating and saving model after training 02/26/2022 18:41:48 - INFO - codeparrot_training - Step 1600000: {'loss/eval': 1.6735893487930298, 'perplexity': 5.331269264221191}