diff --git "a/log/debug_0.log" "b/log/debug_0.log" --- "a/log/debug_0.log" +++ "b/log/debug_0.log" @@ -16050,3 +16050,2009 @@ Use FP16 precision: False 01/27/2022 10:53:39 - INFO - codeparrot_training - Step 15998: {'lr': 0.0004022163183755853, 'samples': 3071808, 'steps': 15998, 'loss/train': 1.232497125864029} 01/27/2022 10:53:43 - INFO - codeparrot_training - Step 15999: {'lr': 0.0004022033381398781, 'samples': 3072000, 'steps': 15999, 'loss/train': 0.9505569040775299} 01/27/2022 10:53:43 - INFO - codeparrot_training - Evaluating and saving model checkpoint +01/27/2022 10:54:00 - WARNING - huggingface_hub.repository - Several commits (8) will be pushed upstream. +01/27/2022 10:54:00 - WARNING - huggingface_hub.repository - The progress bars may be unreliable. +01/27/2022 10:54:35 - WARNING - huggingface_hub.repository - To https://huggingface.co/ncoop57/codeparrot-neo-125M-py + 283559f..6d4e11b royal-monkey-12 -> royal-monkey-12 + +01/27/2022 10:54:40 - INFO - codeparrot_training - Step 16000: {'lr': 0.0004021903572521802, 'samples': 3072192, 'steps': 16000, 'loss/train': 1.1661683320999146} +01/27/2022 10:54:43 - INFO - codeparrot_training - Step 16001: {'lr': 0.0004021773757125471, 'samples': 3072384, 'steps': 16001, 'loss/train': 0.5482946634292603} +01/27/2022 10:54:46 - INFO - codeparrot_training - Step 16002: {'lr': 0.0004021643935210344, 'samples': 3072576, 'steps': 16002, 'loss/train': 1.156781941652298} +01/27/2022 10:54:49 - INFO - codeparrot_training - Step 16003: {'lr': 0.0004021514106776978, 'samples': 3072768, 'steps': 16003, 'loss/train': 0.8858944773674011} +01/27/2022 10:54:52 - INFO - codeparrot_training - Step 16004: {'lr': 0.00040213842718259287, 'samples': 3072960, 'steps': 16004, 'loss/train': 0.805213987827301} +01/27/2022 10:54:56 - INFO - codeparrot_training - Step 16005: {'lr': 0.00040212544303577525, 'samples': 3073152, 'steps': 16005, 'loss/train': 1.2046675086021423} +01/27/2022 10:54:59 - INFO - codeparrot_training - Step 16006: {'lr': 0.00040211245823730047, 'samples': 3073344, 'steps': 16006, 'loss/train': 1.084903746843338} +01/27/2022 10:55:03 - INFO - codeparrot_training - Step 16007: {'lr': 0.00040209947278722425, 'samples': 3073536, 'steps': 16007, 'loss/train': 0.8159877955913544} +01/27/2022 10:55:06 - INFO - codeparrot_training - Step 16008: {'lr': 0.0004020864866856022, 'samples': 3073728, 'steps': 16008, 'loss/train': 0.36306392401456833} +01/27/2022 10:55:09 - INFO - codeparrot_training - Step 16009: {'lr': 0.0004020734999324899, 'samples': 3073920, 'steps': 16009, 'loss/train': 1.1039950847625732} +01/27/2022 10:55:13 - INFO - codeparrot_training - Step 16010: {'lr': 0.0004020605125279431, 'samples': 3074112, 'steps': 16010, 'loss/train': 0.9269937872886658} +01/27/2022 10:55:16 - INFO - codeparrot_training - Step 16011: {'lr': 0.0004020475244720173, 'samples': 3074304, 'steps': 16011, 'loss/train': 0.8019360601902008} +01/27/2022 10:55:19 - INFO - codeparrot_training - Step 16012: {'lr': 0.0004020345357647681, 'samples': 3074496, 'steps': 16012, 'loss/train': 1.1659138798713684} +01/27/2022 10:55:22 - INFO - codeparrot_training - Step 16013: {'lr': 0.0004020215464062513, 'samples': 3074688, 'steps': 16013, 'loss/train': 0.3456604778766632} +01/27/2022 10:55:25 - INFO - codeparrot_training - Step 16014: {'lr': 0.0004020085563965226, 'samples': 3074880, 'steps': 16014, 'loss/train': 0.8018893003463745} +01/27/2022 10:55:28 - INFO - codeparrot_training - Step 16015: {'lr': 0.00040199556573563736, 'samples': 3075072, 'steps': 16015, 'loss/train': 0.6280533671379089} +01/27/2022 10:55:33 - INFO - codeparrot_training - Step 16016: {'lr': 0.0004019825744236514, 'samples': 3075264, 'steps': 16016, 'loss/train': 0.978644996881485} +01/27/2022 10:55:36 - INFO - codeparrot_training - Step 16017: {'lr': 0.00040196958246062033, 'samples': 3075456, 'steps': 16017, 'loss/train': 0.6406474560499191} +01/27/2022 10:55:39 - INFO - codeparrot_training - Step 16018: {'lr': 0.00040195658984659987, 'samples': 3075648, 'steps': 16018, 'loss/train': 0.7530016601085663} +01/27/2022 10:55:42 - INFO - codeparrot_training - Step 16019: {'lr': 0.0004019435965816456, 'samples': 3075840, 'steps': 16019, 'loss/train': 0.8047084808349609} +01/27/2022 10:55:45 - INFO - codeparrot_training - Step 16020: {'lr': 0.0004019306026658132, 'samples': 3076032, 'steps': 16020, 'loss/train': 0.9200859367847443} +01/27/2022 10:55:48 - INFO - codeparrot_training - Step 16021: {'lr': 0.00040191760809915833, 'samples': 3076224, 'steps': 16021, 'loss/train': 1.126099169254303} +01/27/2022 10:55:52 - INFO - codeparrot_training - Step 16022: {'lr': 0.00040190461288173675, 'samples': 3076416, 'steps': 16022, 'loss/train': 0.8763271272182465} +01/27/2022 10:55:55 - INFO - codeparrot_training - Step 16023: {'lr': 0.000401891617013604, 'samples': 3076608, 'steps': 16023, 'loss/train': 0.7621990442276001} +01/27/2022 10:56:00 - INFO - codeparrot_training - Step 16024: {'lr': 0.00040187862049481573, 'samples': 3076800, 'steps': 16024, 'loss/train': 0.8959260284900665} +01/27/2022 10:56:03 - INFO - codeparrot_training - Step 16025: {'lr': 0.00040186562332542773, 'samples': 3076992, 'steps': 16025, 'loss/train': 0.979356586933136} +01/27/2022 10:56:06 - INFO - codeparrot_training - Step 16026: {'lr': 0.0004018526255054956, 'samples': 3077184, 'steps': 16026, 'loss/train': 1.3099334836006165} +01/27/2022 10:56:10 - INFO - codeparrot_training - Step 16027: {'lr': 0.00040183962703507515, 'samples': 3077376, 'steps': 16027, 'loss/train': 0.2690383046865463} +01/27/2022 10:56:13 - INFO - codeparrot_training - Step 16028: {'lr': 0.00040182662791422185, 'samples': 3077568, 'steps': 16028, 'loss/train': 0.8848666548728943} +01/27/2022 10:56:16 - INFO - codeparrot_training - Step 16029: {'lr': 0.0004018136281429915, 'samples': 3077760, 'steps': 16029, 'loss/train': 0.7618233561515808} +01/27/2022 10:56:19 - INFO - codeparrot_training - Step 16030: {'lr': 0.0004018006277214398, 'samples': 3077952, 'steps': 16030, 'loss/train': 0.35345611721277237} +01/27/2022 10:56:22 - INFO - codeparrot_training - Step 16031: {'lr': 0.00040178762664962235, 'samples': 3078144, 'steps': 16031, 'loss/train': 0.8115330934524536} +01/27/2022 10:56:26 - INFO - codeparrot_training - Step 16032: {'lr': 0.000401774624927595, 'samples': 3078336, 'steps': 16032, 'loss/train': 0.725628063082695} +01/27/2022 10:56:30 - INFO - codeparrot_training - Step 16033: {'lr': 0.00040176162255541325, 'samples': 3078528, 'steps': 16033, 'loss/train': 1.0257490575313568} +01/27/2022 10:56:33 - INFO - codeparrot_training - Step 16034: {'lr': 0.00040174861953313297, 'samples': 3078720, 'steps': 16034, 'loss/train': 0.7061260342597961} +01/27/2022 10:56:36 - INFO - codeparrot_training - Step 16035: {'lr': 0.00040173561586080974, 'samples': 3078912, 'steps': 16035, 'loss/train': 0.744817391037941} +01/27/2022 10:56:39 - INFO - codeparrot_training - Step 16036: {'lr': 0.0004017226115384994, 'samples': 3079104, 'steps': 16036, 'loss/train': 0.9325784146785736} +01/27/2022 10:56:42 - INFO - codeparrot_training - Step 16037: {'lr': 0.00040170960656625744, 'samples': 3079296, 'steps': 16037, 'loss/train': 0.9887999296188354} +01/27/2022 10:56:46 - INFO - codeparrot_training - Step 16038: {'lr': 0.00040169660094413977, 'samples': 3079488, 'steps': 16038, 'loss/train': 0.9525887668132782} +01/27/2022 10:56:49 - INFO - codeparrot_training - Step 16039: {'lr': 0.00040168359467220206, 'samples': 3079680, 'steps': 16039, 'loss/train': 0.4065826088190079} +01/27/2022 10:56:52 - INFO - codeparrot_training - Step 16040: {'lr': 0.00040167058775049993, 'samples': 3079872, 'steps': 16040, 'loss/train': 0.9671700596809387} +01/27/2022 10:56:55 - INFO - codeparrot_training - Step 16041: {'lr': 0.0004016575801790892, 'samples': 3080064, 'steps': 16041, 'loss/train': 1.1109221577644348} +01/27/2022 10:57:00 - INFO - codeparrot_training - Step 16042: {'lr': 0.0004016445719580256, 'samples': 3080256, 'steps': 16042, 'loss/train': 0.507965013384819} +01/27/2022 10:57:04 - INFO - codeparrot_training - Step 16043: {'lr': 0.0004016315630873647, 'samples': 3080448, 'steps': 16043, 'loss/train': 1.3136743605136871} +01/27/2022 10:57:07 - INFO - codeparrot_training - Step 16044: {'lr': 0.00040161855356716245, 'samples': 3080640, 'steps': 16044, 'loss/train': 0.75885209441185} +01/27/2022 10:57:10 - INFO - codeparrot_training - Step 16045: {'lr': 0.00040160554339747434, 'samples': 3080832, 'steps': 16045, 'loss/train': 0.6729195713996887} +01/27/2022 10:57:13 - INFO - codeparrot_training - Step 16046: {'lr': 0.00040159253257835624, 'samples': 3081024, 'steps': 16046, 'loss/train': 0.9527392387390137} +01/27/2022 10:57:16 - INFO - codeparrot_training - Step 16047: {'lr': 0.00040157952110986397, 'samples': 3081216, 'steps': 16047, 'loss/train': 0.7149668633937836} +01/27/2022 10:57:19 - INFO - codeparrot_training - Step 16048: {'lr': 0.00040156650899205305, 'samples': 3081408, 'steps': 16048, 'loss/train': 1.0416427552700043} +01/27/2022 10:57:22 - INFO - codeparrot_training - Step 16049: {'lr': 0.00040155349622497937, 'samples': 3081600, 'steps': 16049, 'loss/train': 1.0698343813419342} +01/27/2022 10:57:26 - INFO - codeparrot_training - Step 16050: {'lr': 0.0004015404828086987, 'samples': 3081792, 'steps': 16050, 'loss/train': 0.5175999104976654} +01/27/2022 10:57:30 - INFO - codeparrot_training - Step 16051: {'lr': 0.0004015274687432667, 'samples': 3081984, 'steps': 16051, 'loss/train': 0.11417432874441147} +01/27/2022 10:57:33 - INFO - codeparrot_training - Step 16052: {'lr': 0.0004015144540287391, 'samples': 3082176, 'steps': 16052, 'loss/train': 0.7645329236984253} +01/27/2022 10:57:36 - INFO - codeparrot_training - Step 16053: {'lr': 0.00040150143866517164, 'samples': 3082368, 'steps': 16053, 'loss/train': 0.9410390853881836} +01/27/2022 10:57:39 - INFO - codeparrot_training - Step 16054: {'lr': 0.0004014884226526202, 'samples': 3082560, 'steps': 16054, 'loss/train': 0.9200019836425781} +01/27/2022 10:57:42 - INFO - codeparrot_training - Step 16055: {'lr': 0.0004014754059911405, 'samples': 3082752, 'steps': 16055, 'loss/train': 0.40017472207546234} +01/27/2022 10:57:46 - INFO - codeparrot_training - Step 16056: {'lr': 0.0004014623886807882, 'samples': 3082944, 'steps': 16056, 'loss/train': 0.702796071767807} +01/27/2022 10:57:49 - INFO - codeparrot_training - Step 16057: {'lr': 0.0004014493707216191, 'samples': 3083136, 'steps': 16057, 'loss/train': 0.19270335137844086} +01/27/2022 10:57:52 - INFO - codeparrot_training - Step 16058: {'lr': 0.00040143635211368903, 'samples': 3083328, 'steps': 16058, 'loss/train': 0.830212265253067} +01/27/2022 10:57:56 - INFO - codeparrot_training - Step 16059: {'lr': 0.0004014233328570537, 'samples': 3083520, 'steps': 16059, 'loss/train': 0.5202421098947525} +01/27/2022 10:58:00 - INFO - codeparrot_training - Step 16060: {'lr': 0.0004014103129517689, 'samples': 3083712, 'steps': 16060, 'loss/train': 0.7666836082935333} +01/27/2022 10:58:03 - INFO - codeparrot_training - Step 16061: {'lr': 0.00040139729239789036, 'samples': 3083904, 'steps': 16061, 'loss/train': 0.41621115803718567} +01/27/2022 10:58:06 - INFO - codeparrot_training - Step 16062: {'lr': 0.0004013842711954739, 'samples': 3084096, 'steps': 16062, 'loss/train': 0.9253009557723999} +01/27/2022 10:58:09 - INFO - codeparrot_training - Step 16063: {'lr': 0.0004013712493445753, 'samples': 3084288, 'steps': 16063, 'loss/train': 0.9212969541549683} +01/27/2022 10:58:12 - INFO - codeparrot_training - Step 16064: {'lr': 0.00040135822684525036, 'samples': 3084480, 'steps': 16064, 'loss/train': 1.1886706352233887} +01/27/2022 10:58:15 - INFO - codeparrot_training - Step 16065: {'lr': 0.0004013452036975548, 'samples': 3084672, 'steps': 16065, 'loss/train': 0.8177409768104553} +01/27/2022 10:58:19 - INFO - codeparrot_training - Step 16066: {'lr': 0.0004013321799015445, 'samples': 3084864, 'steps': 16066, 'loss/train': 0.8625762462615967} +01/27/2022 10:58:22 - INFO - codeparrot_training - Step 16067: {'lr': 0.00040131915545727517, 'samples': 3085056, 'steps': 16067, 'loss/train': 0.7645277380943298} +01/27/2022 10:58:27 - INFO - codeparrot_training - Step 16068: {'lr': 0.00040130613036480265, 'samples': 3085248, 'steps': 16068, 'loss/train': 0.8315145671367645} +01/27/2022 10:58:30 - INFO - codeparrot_training - Step 16069: {'lr': 0.0004012931046241827, 'samples': 3085440, 'steps': 16069, 'loss/train': 0.3940197676420212} +01/27/2022 10:58:33 - INFO - codeparrot_training - Step 16070: {'lr': 0.00040128007823547106, 'samples': 3085632, 'steps': 16070, 'loss/train': 0.8903836011886597} +01/27/2022 10:58:36 - INFO - codeparrot_training - Step 16071: {'lr': 0.00040126705119872367, 'samples': 3085824, 'steps': 16071, 'loss/train': 0.939335972070694} +01/27/2022 10:58:39 - INFO - codeparrot_training - Step 16072: {'lr': 0.00040125402351399623, 'samples': 3086016, 'steps': 16072, 'loss/train': 0.7546872496604919} +01/27/2022 10:58:43 - INFO - codeparrot_training - Step 16073: {'lr': 0.0004012409951813446, 'samples': 3086208, 'steps': 16073, 'loss/train': 0.5747102648019791} +01/27/2022 10:58:46 - INFO - codeparrot_training - Step 16074: {'lr': 0.0004012279662008246, 'samples': 3086400, 'steps': 16074, 'loss/train': 0.7705071866512299} +01/27/2022 10:58:49 - INFO - codeparrot_training - Step 16075: {'lr': 0.000401214936572492, 'samples': 3086592, 'steps': 16075, 'loss/train': 0.8908843696117401} +01/27/2022 10:58:52 - INFO - codeparrot_training - Step 16076: {'lr': 0.0004012019062964026, 'samples': 3086784, 'steps': 16076, 'loss/train': 0.33995582163333893} +01/27/2022 10:58:57 - INFO - codeparrot_training - Step 16077: {'lr': 0.0004011888753726123, 'samples': 3086976, 'steps': 16077, 'loss/train': 0.9412097632884979} +01/27/2022 10:59:00 - INFO - codeparrot_training - Step 16078: {'lr': 0.00040117584380117675, 'samples': 3087168, 'steps': 16078, 'loss/train': 0.7451140880584717} +01/27/2022 10:59:03 - INFO - codeparrot_training - Step 16079: {'lr': 0.000401162811582152, 'samples': 3087360, 'steps': 16079, 'loss/train': 0.9822689294815063} +01/27/2022 10:59:06 - INFO - codeparrot_training - Step 16080: {'lr': 0.00040114977871559375, 'samples': 3087552, 'steps': 16080, 'loss/train': 0.6839516311883926} +01/27/2022 10:59:09 - INFO - codeparrot_training - Step 16081: {'lr': 0.0004011367452015578, 'samples': 3087744, 'steps': 16081, 'loss/train': 1.4525760412216187} +01/27/2022 10:59:12 - INFO - codeparrot_training - Step 16082: {'lr': 0.00040112371104010004, 'samples': 3087936, 'steps': 16082, 'loss/train': 0.7872274518013} +01/27/2022 10:59:16 - INFO - codeparrot_training - Step 16083: {'lr': 0.00040111067623127626, 'samples': 3088128, 'steps': 16083, 'loss/train': 0.6307044178247452} +01/27/2022 10:59:19 - INFO - codeparrot_training - Step 16084: {'lr': 0.0004010976407751424, 'samples': 3088320, 'steps': 16084, 'loss/train': 0.6168482601642609} +01/27/2022 10:59:22 - INFO - codeparrot_training - Step 16085: {'lr': 0.00040108460467175425, 'samples': 3088512, 'steps': 16085, 'loss/train': 0.7885091006755829} +01/27/2022 10:59:26 - INFO - codeparrot_training - Step 16086: {'lr': 0.00040107156792116753, 'samples': 3088704, 'steps': 16086, 'loss/train': 0.7441748678684235} +01/27/2022 10:59:29 - INFO - codeparrot_training - Step 16087: {'lr': 0.0004010585305234382, 'samples': 3088896, 'steps': 16087, 'loss/train': 0.7913475036621094} +01/27/2022 10:59:33 - INFO - codeparrot_training - Step 16088: {'lr': 0.00040104549247862217, 'samples': 3089088, 'steps': 16088, 'loss/train': 0.8577834069728851} +01/27/2022 10:59:36 - INFO - codeparrot_training - Step 16089: {'lr': 0.0004010324537867751, 'samples': 3089280, 'steps': 16089, 'loss/train': 0.05432242155075073} +01/27/2022 10:59:39 - INFO - codeparrot_training - Step 16090: {'lr': 0.000401019414447953, 'samples': 3089472, 'steps': 16090, 'loss/train': 0.9519981443881989} +01/27/2022 10:59:42 - INFO - codeparrot_training - Step 16091: {'lr': 0.0004010063744622117, 'samples': 3089664, 'steps': 16091, 'loss/train': 1.5884395837783813} +01/27/2022 10:59:45 - INFO - codeparrot_training - Step 16092: {'lr': 0.00040099333382960707, 'samples': 3089856, 'steps': 16092, 'loss/train': 0.6136399805545807} +01/27/2022 10:59:48 - INFO - codeparrot_training - Step 16093: {'lr': 0.00040098029255019484, 'samples': 3090048, 'steps': 16093, 'loss/train': 0.8105615973472595} +01/27/2022 10:59:51 - INFO - codeparrot_training - Step 16094: {'lr': 0.0004009672506240311, 'samples': 3090240, 'steps': 16094, 'loss/train': 0.864004522562027} +01/27/2022 10:59:56 - INFO - codeparrot_training - Step 16095: {'lr': 0.00040095420805117153, 'samples': 3090432, 'steps': 16095, 'loss/train': 1.0065853893756866} +01/27/2022 10:59:59 - INFO - codeparrot_training - Step 16096: {'lr': 0.0004009411648316721, 'samples': 3090624, 'steps': 16096, 'loss/train': 0.8987221419811249} +01/27/2022 11:00:02 - INFO - codeparrot_training - Step 16097: {'lr': 0.0004009281209655886, 'samples': 3090816, 'steps': 16097, 'loss/train': 0.4887055903673172} +01/27/2022 11:00:05 - INFO - codeparrot_training - Step 16098: {'lr': 0.000400915076452977, 'samples': 3091008, 'steps': 16098, 'loss/train': 0.9994000196456909} +01/27/2022 11:00:09 - INFO - codeparrot_training - Step 16099: {'lr': 0.0004009020312938931, 'samples': 3091200, 'steps': 16099, 'loss/train': 1.2118475139141083} +01/27/2022 11:00:12 - INFO - codeparrot_training - Step 16100: {'lr': 0.0004008889854883929, 'samples': 3091392, 'steps': 16100, 'loss/train': 0.5654051452875137} +01/27/2022 11:00:15 - INFO - codeparrot_training - Step 16101: {'lr': 0.0004008759390365321, 'samples': 3091584, 'steps': 16101, 'loss/train': 0.5668313205242157} +01/27/2022 11:00:18 - INFO - codeparrot_training - Step 16102: {'lr': 0.00040086289193836674, 'samples': 3091776, 'steps': 16102, 'loss/train': 0.8640321493148804} +01/27/2022 11:00:23 - INFO - codeparrot_training - Step 16103: {'lr': 0.00040084984419395264, 'samples': 3091968, 'steps': 16103, 'loss/train': 0.5375676602125168} +01/27/2022 11:00:27 - INFO - codeparrot_training - Step 16104: {'lr': 0.00040083679580334565, 'samples': 3092160, 'steps': 16104, 'loss/train': 1.8602385520935059} +01/27/2022 11:00:30 - INFO - codeparrot_training - Step 16105: {'lr': 0.00040082374676660176, 'samples': 3092352, 'steps': 16105, 'loss/train': 1.322939783334732} +01/27/2022 11:00:33 - INFO - codeparrot_training - Step 16106: {'lr': 0.00040081069708377686, 'samples': 3092544, 'steps': 16106, 'loss/train': 0.7416754513978958} +01/27/2022 11:00:36 - INFO - codeparrot_training - Step 16107: {'lr': 0.0004007976467549268, 'samples': 3092736, 'steps': 16107, 'loss/train': 1.258715271949768} +01/27/2022 11:00:39 - INFO - codeparrot_training - Step 16108: {'lr': 0.0004007845957801075, 'samples': 3092928, 'steps': 16108, 'loss/train': 0.9229179918766022} +01/27/2022 11:00:42 - INFO - codeparrot_training - Step 16109: {'lr': 0.0004007715441593749, 'samples': 3093120, 'steps': 16109, 'loss/train': 1.1582191586494446} +01/27/2022 11:00:45 - INFO - codeparrot_training - Step 16110: {'lr': 0.0004007584918927849, 'samples': 3093312, 'steps': 16110, 'loss/train': 0.6482914388179779} +01/27/2022 11:00:49 - INFO - codeparrot_training - Step 16111: {'lr': 0.0004007454389803933, 'samples': 3093504, 'steps': 16111, 'loss/train': 0.8451608419418335} +01/27/2022 11:00:53 - INFO - codeparrot_training - Step 16112: {'lr': 0.00040073238542225623, 'samples': 3093696, 'steps': 16112, 'loss/train': 0.18500537797808647} +01/27/2022 11:00:57 - INFO - codeparrot_training - Step 16113: {'lr': 0.00040071933121842943, 'samples': 3093888, 'steps': 16113, 'loss/train': 1.3139043152332306} +01/27/2022 11:01:00 - INFO - codeparrot_training - Step 16114: {'lr': 0.00040070627636896886, 'samples': 3094080, 'steps': 16114, 'loss/train': 0.798805832862854} +01/27/2022 11:01:03 - INFO - codeparrot_training - Step 16115: {'lr': 0.0004006932208739304, 'samples': 3094272, 'steps': 16115, 'loss/train': 0.9381473064422607} +01/27/2022 11:01:06 - INFO - codeparrot_training - Step 16116: {'lr': 0.0004006801647333701, 'samples': 3094464, 'steps': 16116, 'loss/train': 0.601113572716713} +01/27/2022 11:01:09 - INFO - codeparrot_training - Step 16117: {'lr': 0.0004006671079473438, 'samples': 3094656, 'steps': 16117, 'loss/train': 0.7330269366502762} +01/27/2022 11:01:12 - INFO - codeparrot_training - Step 16118: {'lr': 0.00040065405051590745, 'samples': 3094848, 'steps': 16118, 'loss/train': 0.8243494927883148} +01/27/2022 11:01:15 - INFO - codeparrot_training - Step 16119: {'lr': 0.000400640992439117, 'samples': 3095040, 'steps': 16119, 'loss/train': 0.1335626095533371} +01/27/2022 11:01:19 - INFO - codeparrot_training - Step 16120: {'lr': 0.0004006279337170283, 'samples': 3095232, 'steps': 16120, 'loss/train': 0.9009931683540344} +01/27/2022 11:01:23 - INFO - codeparrot_training - Step 16121: {'lr': 0.00040061487434969744, 'samples': 3095424, 'steps': 16121, 'loss/train': 0.9683931469917297} +01/27/2022 11:01:26 - INFO - codeparrot_training - Step 16122: {'lr': 0.00040060181433718037, 'samples': 3095616, 'steps': 16122, 'loss/train': 0.9955757260322571} +01/27/2022 11:01:29 - INFO - codeparrot_training - Step 16123: {'lr': 0.00040058875367953285, 'samples': 3095808, 'steps': 16123, 'loss/train': 1.0257788300514221} +01/27/2022 11:01:33 - INFO - codeparrot_training - Step 16124: {'lr': 0.0004005756923768109, 'samples': 3096000, 'steps': 16124, 'loss/train': 0.36369357258081436} +01/27/2022 11:01:36 - INFO - codeparrot_training - Step 16125: {'lr': 0.0004005626304290705, 'samples': 3096192, 'steps': 16125, 'loss/train': 0.7323752492666245} +01/27/2022 11:01:39 - INFO - codeparrot_training - Step 16126: {'lr': 0.00040054956783636765, 'samples': 3096384, 'steps': 16126, 'loss/train': 0.6439166218042374} +01/27/2022 11:01:42 - INFO - codeparrot_training - Step 16127: {'lr': 0.00040053650459875823, 'samples': 3096576, 'steps': 16127, 'loss/train': 0.6893886029720306} +01/27/2022 11:01:45 - INFO - codeparrot_training - Step 16128: {'lr': 0.0004005234407162982, 'samples': 3096768, 'steps': 16128, 'loss/train': 0.7801358699798584} +01/27/2022 11:01:50 - INFO - codeparrot_training - Step 16129: {'lr': 0.00040051037618904365, 'samples': 3096960, 'steps': 16129, 'loss/train': 0.35465774685144424} +01/27/2022 11:01:53 - INFO - codeparrot_training - Step 16130: {'lr': 0.0004004973110170503, 'samples': 3097152, 'steps': 16130, 'loss/train': 0.6837345510721207} +01/27/2022 11:01:57 - INFO - codeparrot_training - Step 16131: {'lr': 0.0004004842452003743, 'samples': 3097344, 'steps': 16131, 'loss/train': 0.6895062178373337} +01/27/2022 11:02:00 - INFO - codeparrot_training - Step 16132: {'lr': 0.0004004711787390716, 'samples': 3097536, 'steps': 16132, 'loss/train': 1.583378255367279} +01/27/2022 11:02:03 - INFO - codeparrot_training - Step 16133: {'lr': 0.0004004581116331981, 'samples': 3097728, 'steps': 16133, 'loss/train': 0.34579894691705704} +01/27/2022 11:02:06 - INFO - codeparrot_training - Step 16134: {'lr': 0.00040044504388280996, 'samples': 3097920, 'steps': 16134, 'loss/train': 0.8061041235923767} +01/27/2022 11:02:09 - INFO - codeparrot_training - Step 16135: {'lr': 0.00040043197548796295, 'samples': 3098112, 'steps': 16135, 'loss/train': 0.42744599282741547} +01/27/2022 11:02:12 - INFO - codeparrot_training - Step 16136: {'lr': 0.0004004189064487131, 'samples': 3098304, 'steps': 16136, 'loss/train': 0.8541747629642487} +01/27/2022 11:02:15 - INFO - codeparrot_training - Step 16137: {'lr': 0.00040040583676511645, 'samples': 3098496, 'steps': 16137, 'loss/train': 0.5681349188089371} +01/27/2022 11:02:20 - INFO - codeparrot_training - Step 16138: {'lr': 0.0004003927664372289, 'samples': 3098688, 'steps': 16138, 'loss/train': 0.9292392432689667} +01/27/2022 11:02:23 - INFO - codeparrot_training - Step 16139: {'lr': 0.00040037969546510653, 'samples': 3098880, 'steps': 16139, 'loss/train': 0.5920339822769165} +01/27/2022 11:02:26 - INFO - codeparrot_training - Step 16140: {'lr': 0.0004003666238488053, 'samples': 3099072, 'steps': 16140, 'loss/train': 1.0657279193401337} +01/27/2022 11:02:29 - INFO - codeparrot_training - Step 16141: {'lr': 0.00040035355158838114, 'samples': 3099264, 'steps': 16141, 'loss/train': 0.6732043325901031} +01/27/2022 11:02:32 - INFO - codeparrot_training - Step 16142: {'lr': 0.0004003404786838902, 'samples': 3099456, 'steps': 16142, 'loss/train': 0.7918139398097992} +01/27/2022 11:02:36 - INFO - codeparrot_training - Step 16143: {'lr': 0.0004003274051353884, 'samples': 3099648, 'steps': 16143, 'loss/train': 0.68306864798069} +01/27/2022 11:02:39 - INFO - codeparrot_training - Step 16144: {'lr': 0.00040031433094293167, 'samples': 3099840, 'steps': 16144, 'loss/train': 0.8096366822719574} +01/27/2022 11:02:42 - INFO - codeparrot_training - Step 16145: {'lr': 0.0004003012561065761, 'samples': 3100032, 'steps': 16145, 'loss/train': 0.7284202873706818} +01/27/2022 11:02:45 - INFO - codeparrot_training - Step 16146: {'lr': 0.0004002881806263776, 'samples': 3100224, 'steps': 16146, 'loss/train': 0.8681244850158691} +01/27/2022 11:02:51 - INFO - codeparrot_training - Step 16147: {'lr': 0.0004002751045023924, 'samples': 3100416, 'steps': 16147, 'loss/train': 1.2953082025051117} +01/27/2022 11:02:54 - INFO - codeparrot_training - Step 16148: {'lr': 0.00040026202773467623, 'samples': 3100608, 'steps': 16148, 'loss/train': 1.031703382730484} +01/27/2022 11:02:57 - INFO - codeparrot_training - Step 16149: {'lr': 0.00040024895032328536, 'samples': 3100800, 'steps': 16149, 'loss/train': 0.5150815844535828} +01/27/2022 11:03:00 - INFO - codeparrot_training - Step 16150: {'lr': 0.0004002358722682756, 'samples': 3100992, 'steps': 16150, 'loss/train': 0.39607080817222595} +01/27/2022 11:03:03 - INFO - codeparrot_training - Step 16151: {'lr': 0.00040022279356970316, 'samples': 3101184, 'steps': 16151, 'loss/train': 1.1239608228206635} +01/27/2022 11:03:06 - INFO - codeparrot_training - Step 16152: {'lr': 0.0004002097142276239, 'samples': 3101376, 'steps': 16152, 'loss/train': 0.636260524392128} +01/27/2022 11:03:10 - INFO - codeparrot_training - Step 16153: {'lr': 0.00040019663424209397, 'samples': 3101568, 'steps': 16153, 'loss/train': 0.6892773360013962} +01/27/2022 11:03:13 - INFO - codeparrot_training - Step 16154: {'lr': 0.0004001835536131693, 'samples': 3101760, 'steps': 16154, 'loss/train': 0.66307532787323} +01/27/2022 11:03:17 - INFO - codeparrot_training - Step 16155: {'lr': 0.00040017047234090596, 'samples': 3101952, 'steps': 16155, 'loss/train': 0.3124677836894989} +01/27/2022 11:03:20 - INFO - codeparrot_training - Step 16156: {'lr': 0.00040015739042536, 'samples': 3102144, 'steps': 16156, 'loss/train': 0.7959659099578857} +01/27/2022 11:03:23 - INFO - codeparrot_training - Step 16157: {'lr': 0.00040014430786658754, 'samples': 3102336, 'steps': 16157, 'loss/train': 0.6381566673517227} +01/27/2022 11:03:27 - INFO - codeparrot_training - Step 16158: {'lr': 0.0004001312246646446, 'samples': 3102528, 'steps': 16158, 'loss/train': 1.0710554122924805} +01/27/2022 11:03:30 - INFO - codeparrot_training - Step 16159: {'lr': 0.000400118140819587, 'samples': 3102720, 'steps': 16159, 'loss/train': 0.7727639973163605} +01/27/2022 11:03:33 - INFO - codeparrot_training - Step 16160: {'lr': 0.00040010505633147106, 'samples': 3102912, 'steps': 16160, 'loss/train': 0.4233178049325943} +01/27/2022 11:03:36 - INFO - codeparrot_training - Step 16161: {'lr': 0.0004000919712003526, 'samples': 3103104, 'steps': 16161, 'loss/train': 0.6450802534818649} +01/27/2022 11:03:39 - INFO - codeparrot_training - Step 16162: {'lr': 0.0004000788854262879, 'samples': 3103296, 'steps': 16162, 'loss/train': 0.7920041978359222} +01/27/2022 11:03:42 - INFO - codeparrot_training - Step 16163: {'lr': 0.00040006579900933294, 'samples': 3103488, 'steps': 16163, 'loss/train': 1.2418106496334076} +01/27/2022 11:03:47 - INFO - codeparrot_training - Step 16164: {'lr': 0.00040005271194954367, 'samples': 3103680, 'steps': 16164, 'loss/train': 0.7443996369838715} +01/27/2022 11:03:50 - INFO - codeparrot_training - Step 16165: {'lr': 0.00040003962424697625, 'samples': 3103872, 'steps': 16165, 'loss/train': 1.0874756276607513} +01/27/2022 11:03:53 - INFO - codeparrot_training - Step 16166: {'lr': 0.0004000265359016867, 'samples': 3104064, 'steps': 16166, 'loss/train': 0.9540434181690216} +01/27/2022 11:03:56 - INFO - codeparrot_training - Step 16167: {'lr': 0.0004000134469137312, 'samples': 3104256, 'steps': 16167, 'loss/train': 0.80058154463768} +01/27/2022 11:03:59 - INFO - codeparrot_training - Step 16168: {'lr': 0.00040000035728316564, 'samples': 3104448, 'steps': 16168, 'loss/train': 1.0064590573310852} +01/27/2022 11:04:02 - INFO - codeparrot_training - Step 16169: {'lr': 0.0003999872670100462, 'samples': 3104640, 'steps': 16169, 'loss/train': 0.9559896290302277} +01/27/2022 11:04:06 - INFO - codeparrot_training - Step 16170: {'lr': 0.000399974176094429, 'samples': 3104832, 'steps': 16170, 'loss/train': 0.6497704535722733} +01/27/2022 11:04:09 - INFO - codeparrot_training - Step 16171: {'lr': 0.00039996108453637, 'samples': 3105024, 'steps': 16171, 'loss/train': 0.6208857446908951} +01/27/2022 11:04:12 - INFO - codeparrot_training - Step 16172: {'lr': 0.0003999479923359253, 'samples': 3105216, 'steps': 16172, 'loss/train': 1.4152099192142487} +01/27/2022 11:04:17 - INFO - codeparrot_training - Step 16173: {'lr': 0.00039993489949315103, 'samples': 3105408, 'steps': 16173, 'loss/train': 0.744867280125618} +01/27/2022 11:04:20 - INFO - codeparrot_training - Step 16174: {'lr': 0.0003999218060081032, 'samples': 3105600, 'steps': 16174, 'loss/train': 0.48310884833335876} +01/27/2022 11:04:23 - INFO - codeparrot_training - Step 16175: {'lr': 0.0003999087118808381, 'samples': 3105792, 'steps': 16175, 'loss/train': 0.670142412185669} +01/27/2022 11:04:27 - INFO - codeparrot_training - Step 16176: {'lr': 0.0003998956171114116, 'samples': 3105984, 'steps': 16176, 'loss/train': 0.7885953783988953} +01/27/2022 11:04:30 - INFO - codeparrot_training - Step 16177: {'lr': 0.0003998825216998799, 'samples': 3106176, 'steps': 16177, 'loss/train': 0.6725649833679199} +01/27/2022 11:04:33 - INFO - codeparrot_training - Step 16178: {'lr': 0.00039986942564629904, 'samples': 3106368, 'steps': 16178, 'loss/train': 0.9765986502170563} +01/27/2022 11:04:36 - INFO - codeparrot_training - Step 16179: {'lr': 0.0003998563289507251, 'samples': 3106560, 'steps': 16179, 'loss/train': 1.6755763292312622} +01/27/2022 11:04:39 - INFO - codeparrot_training - Step 16180: {'lr': 0.0003998432316132143, 'samples': 3106752, 'steps': 16180, 'loss/train': 0.46593882143497467} +01/27/2022 11:04:42 - INFO - codeparrot_training - Step 16181: {'lr': 0.0003998301336338227, 'samples': 3106944, 'steps': 16181, 'loss/train': 0.25301726907491684} +01/27/2022 11:04:47 - INFO - codeparrot_training - Step 16182: {'lr': 0.0003998170350126064, 'samples': 3107136, 'steps': 16182, 'loss/train': 0.05575856566429138} +01/27/2022 11:04:50 - INFO - codeparrot_training - Step 16183: {'lr': 0.0003998039357496214, 'samples': 3107328, 'steps': 16183, 'loss/train': 0.7141703814268112} +01/27/2022 11:04:53 - INFO - codeparrot_training - Step 16184: {'lr': 0.000399790835844924, 'samples': 3107520, 'steps': 16184, 'loss/train': 0.8561670184135437} +01/27/2022 11:04:56 - INFO - codeparrot_training - Step 16185: {'lr': 0.00039977773529857016, 'samples': 3107712, 'steps': 16185, 'loss/train': 1.07328262925148} +01/27/2022 11:04:59 - INFO - codeparrot_training - Step 16186: {'lr': 0.00039976463411061606, 'samples': 3107904, 'steps': 16186, 'loss/train': 0.7389503717422485} +01/27/2022 11:05:03 - INFO - codeparrot_training - Step 16187: {'lr': 0.00039975153228111784, 'samples': 3108096, 'steps': 16187, 'loss/train': 1.6909891963005066} +01/27/2022 11:05:06 - INFO - codeparrot_training - Step 16188: {'lr': 0.0003997384298101316, 'samples': 3108288, 'steps': 16188, 'loss/train': 0.6677248477935791} +01/27/2022 11:05:09 - INFO - codeparrot_training - Step 16189: {'lr': 0.0003997253266977135, 'samples': 3108480, 'steps': 16189, 'loss/train': 0.8549251556396484} +01/27/2022 11:05:12 - INFO - codeparrot_training - Step 16190: {'lr': 0.0003997122229439196, 'samples': 3108672, 'steps': 16190, 'loss/train': 0.5774542540311813} +01/27/2022 11:05:17 - INFO - codeparrot_training - Step 16191: {'lr': 0.00039969911854880613, 'samples': 3108864, 'steps': 16191, 'loss/train': 0.9107017815113068} +01/27/2022 11:05:20 - INFO - codeparrot_training - Step 16192: {'lr': 0.0003996860135124292, 'samples': 3109056, 'steps': 16192, 'loss/train': 0.6766562908887863} +01/27/2022 11:05:23 - INFO - codeparrot_training - Step 16193: {'lr': 0.00039967290783484485, 'samples': 3109248, 'steps': 16193, 'loss/train': 0.5964110344648361} +01/27/2022 11:05:26 - INFO - codeparrot_training - Step 16194: {'lr': 0.00039965980151610925, 'samples': 3109440, 'steps': 16194, 'loss/train': 0.937375545501709} +01/27/2022 11:05:29 - INFO - codeparrot_training - Step 16195: {'lr': 0.0003996466945562787, 'samples': 3109632, 'steps': 16195, 'loss/train': 0.9767438471317291} +01/27/2022 11:05:32 - INFO - codeparrot_training - Step 16196: {'lr': 0.00039963358695540907, 'samples': 3109824, 'steps': 16196, 'loss/train': 0.5100443512201309} +01/27/2022 11:05:36 - INFO - codeparrot_training - Step 16197: {'lr': 0.00039962047871355686, 'samples': 3110016, 'steps': 16197, 'loss/train': 1.1521844565868378} +01/27/2022 11:05:39 - INFO - codeparrot_training - Step 16198: {'lr': 0.00039960736983077783, 'samples': 3110208, 'steps': 16198, 'loss/train': 1.3670508563518524} +01/27/2022 11:05:43 - INFO - codeparrot_training - Step 16199: {'lr': 0.0003995942603071285, 'samples': 3110400, 'steps': 16199, 'loss/train': 1.3524795770645142} +01/27/2022 11:05:46 - INFO - codeparrot_training - Step 16200: {'lr': 0.0003995811501426648, 'samples': 3110592, 'steps': 16200, 'loss/train': 0.49441806972026825} +01/27/2022 11:05:49 - INFO - codeparrot_training - Step 16201: {'lr': 0.0003995680393374429, 'samples': 3110784, 'steps': 16201, 'loss/train': 0.9774319231510162} +01/27/2022 11:05:53 - INFO - codeparrot_training - Step 16202: {'lr': 0.00039955492789151904, 'samples': 3110976, 'steps': 16202, 'loss/train': 1.0279123485088348} +01/27/2022 11:05:56 - INFO - codeparrot_training - Step 16203: {'lr': 0.0003995418158049494, 'samples': 3111168, 'steps': 16203, 'loss/train': 0.7564727962017059} +01/27/2022 11:05:59 - INFO - codeparrot_training - Step 16204: {'lr': 0.0003995287030777901, 'samples': 3111360, 'steps': 16204, 'loss/train': 0.6424032300710678} +01/27/2022 11:06:02 - INFO - codeparrot_training - Step 16205: {'lr': 0.0003995155897100973, 'samples': 3111552, 'steps': 16205, 'loss/train': 0.7752204537391663} +01/27/2022 11:06:05 - INFO - codeparrot_training - Step 16206: {'lr': 0.0003995024757019272, 'samples': 3111744, 'steps': 16206, 'loss/train': 0.3311643451452255} +01/27/2022 11:06:08 - INFO - codeparrot_training - Step 16207: {'lr': 0.00039948936105333593, 'samples': 3111936, 'steps': 16207, 'loss/train': 0.8803102076053619} +01/27/2022 11:06:14 - INFO - codeparrot_training - Step 16208: {'lr': 0.0003994762457643797, 'samples': 3112128, 'steps': 16208, 'loss/train': 0.6792906671762466} +01/27/2022 11:06:17 - INFO - codeparrot_training - Step 16209: {'lr': 0.0003994631298351148, 'samples': 3112320, 'steps': 16209, 'loss/train': 0.5388921350240707} +01/27/2022 11:06:20 - INFO - codeparrot_training - Step 16210: {'lr': 0.0003994500132655972, 'samples': 3112512, 'steps': 16210, 'loss/train': 0.5494813621044159} +01/27/2022 11:06:23 - INFO - codeparrot_training - Step 16211: {'lr': 0.0003994368960558832, 'samples': 3112704, 'steps': 16211, 'loss/train': 1.0097468197345734} +01/27/2022 11:06:26 - INFO - codeparrot_training - Step 16212: {'lr': 0.0003994237782060291, 'samples': 3112896, 'steps': 16212, 'loss/train': 1.1294676661491394} +01/27/2022 11:06:30 - INFO - codeparrot_training - Step 16213: {'lr': 0.00039941065971609084, 'samples': 3113088, 'steps': 16213, 'loss/train': 0.8697527647018433} +01/27/2022 11:06:33 - INFO - codeparrot_training - Step 16214: {'lr': 0.00039939754058612487, 'samples': 3113280, 'steps': 16214, 'loss/train': 0.687748521566391} +01/27/2022 11:06:36 - INFO - codeparrot_training - Step 16215: {'lr': 0.0003993844208161872, 'samples': 3113472, 'steps': 16215, 'loss/train': 1.127063512802124} +01/27/2022 11:06:39 - INFO - codeparrot_training - Step 16216: {'lr': 0.0003993713004063341, 'samples': 3113664, 'steps': 16216, 'loss/train': 0.7654272615909576} +01/27/2022 11:06:44 - INFO - codeparrot_training - Step 16217: {'lr': 0.0003993581793566219, 'samples': 3113856, 'steps': 16217, 'loss/train': 0.8087926805019379} +01/27/2022 11:06:47 - INFO - codeparrot_training - Step 16218: {'lr': 0.00039934505766710656, 'samples': 3114048, 'steps': 16218, 'loss/train': 0.04706752672791481} +01/27/2022 11:06:51 - INFO - codeparrot_training - Step 16219: {'lr': 0.0003993319353378445, 'samples': 3114240, 'steps': 16219, 'loss/train': 0.8821775615215302} +01/27/2022 11:06:54 - INFO - codeparrot_training - Step 16220: {'lr': 0.0003993188123688918, 'samples': 3114432, 'steps': 16220, 'loss/train': 0.7198865711688995} +01/27/2022 11:06:57 - INFO - codeparrot_training - Step 16221: {'lr': 0.00039930568876030473, 'samples': 3114624, 'steps': 16221, 'loss/train': 0.8737896680831909} +01/27/2022 11:07:00 - INFO - codeparrot_training - Step 16222: {'lr': 0.0003992925645121395, 'samples': 3114816, 'steps': 16222, 'loss/train': 0.8249053359031677} +01/27/2022 11:07:03 - INFO - codeparrot_training - Step 16223: {'lr': 0.00039927943962445234, 'samples': 3115008, 'steps': 16223, 'loss/train': 0.2131868228316307} +01/27/2022 11:07:06 - INFO - codeparrot_training - Step 16224: {'lr': 0.0003992663140972994, 'samples': 3115200, 'steps': 16224, 'loss/train': 1.0417856276035309} +01/27/2022 11:07:09 - INFO - codeparrot_training - Step 16225: {'lr': 0.0003992531879307371, 'samples': 3115392, 'steps': 16225, 'loss/train': 0.316602922976017} +01/27/2022 11:07:14 - INFO - codeparrot_training - Step 16226: {'lr': 0.0003992400611248214, 'samples': 3115584, 'steps': 16226, 'loss/train': 0.03718013595789671} +01/27/2022 11:07:18 - INFO - codeparrot_training - Step 16227: {'lr': 0.0003992269336796087, 'samples': 3115776, 'steps': 16227, 'loss/train': 0.8122426271438599} +01/27/2022 11:07:21 - INFO - codeparrot_training - Step 16228: {'lr': 0.0003992138055951552, 'samples': 3115968, 'steps': 16228, 'loss/train': 0.5651389807462692} +01/27/2022 11:07:24 - INFO - codeparrot_training - Step 16229: {'lr': 0.00039920067687151717, 'samples': 3116160, 'steps': 16229, 'loss/train': 1.0383670628070831} +01/27/2022 11:07:27 - INFO - codeparrot_training - Step 16230: {'lr': 0.0003991875475087508, 'samples': 3116352, 'steps': 16230, 'loss/train': 1.1189160346984863} +01/27/2022 11:07:30 - INFO - codeparrot_training - Step 16231: {'lr': 0.00039917441750691237, 'samples': 3116544, 'steps': 16231, 'loss/train': 0.8673271536827087} +01/27/2022 11:07:33 - INFO - codeparrot_training - Step 16232: {'lr': 0.0003991612868660581, 'samples': 3116736, 'steps': 16232, 'loss/train': 1.0767593085765839} +01/27/2022 11:07:36 - INFO - codeparrot_training - Step 16233: {'lr': 0.0003991481555862442, 'samples': 3116928, 'steps': 16233, 'loss/train': 0.9410102963447571} +01/27/2022 11:07:40 - INFO - codeparrot_training - Step 16234: {'lr': 0.00039913502366752704, 'samples': 3117120, 'steps': 16234, 'loss/train': 1.0308907628059387} +01/27/2022 11:07:44 - INFO - codeparrot_training - Step 16235: {'lr': 0.0003991218911099627, 'samples': 3117312, 'steps': 16235, 'loss/train': 0.6291737258434296} +01/27/2022 11:07:47 - INFO - codeparrot_training - Step 16236: {'lr': 0.0003991087579136076, 'samples': 3117504, 'steps': 16236, 'loss/train': 1.1416684985160828} +01/27/2022 11:07:51 - INFO - codeparrot_training - Step 16237: {'lr': 0.00039909562407851784, 'samples': 3117696, 'steps': 16237, 'loss/train': 1.071918547153473} +01/27/2022 11:07:54 - INFO - codeparrot_training - Step 16238: {'lr': 0.0003990824896047498, 'samples': 3117888, 'steps': 16238, 'loss/train': 0.4931730329990387} +01/27/2022 11:07:57 - INFO - codeparrot_training - Step 16239: {'lr': 0.00039906935449235983, 'samples': 3118080, 'steps': 16239, 'loss/train': 0.5481891632080078} +01/27/2022 11:08:00 - INFO - codeparrot_training - Step 16240: {'lr': 0.00039905621874140396, 'samples': 3118272, 'steps': 16240, 'loss/train': 0.8554481863975525} +01/27/2022 11:08:03 - INFO - codeparrot_training - Step 16241: {'lr': 0.00039904308235193866, 'samples': 3118464, 'steps': 16241, 'loss/train': 1.6575230956077576} +01/27/2022 11:08:06 - INFO - codeparrot_training - Step 16242: {'lr': 0.00039902994532402004, 'samples': 3118656, 'steps': 16242, 'loss/train': 0.9631115198135376} +01/27/2022 11:08:10 - INFO - codeparrot_training - Step 16243: {'lr': 0.0003990168076577045, 'samples': 3118848, 'steps': 16243, 'loss/train': 1.032608985900879} +01/27/2022 11:08:14 - INFO - codeparrot_training - Step 16244: {'lr': 0.00039900366935304824, 'samples': 3119040, 'steps': 16244, 'loss/train': 1.0263049900531769} +01/27/2022 11:08:17 - INFO - codeparrot_training - Step 16245: {'lr': 0.00039899053041010765, 'samples': 3119232, 'steps': 16245, 'loss/train': 0.5330948531627655} +01/27/2022 11:08:20 - INFO - codeparrot_training - Step 16246: {'lr': 0.00039897739082893883, 'samples': 3119424, 'steps': 16246, 'loss/train': 0.9059200286865234} +01/27/2022 11:08:23 - INFO - codeparrot_training - Step 16247: {'lr': 0.0003989642506095983, 'samples': 3119616, 'steps': 16247, 'loss/train': 1.2391507029533386} +01/27/2022 11:08:27 - INFO - codeparrot_training - Step 16248: {'lr': 0.0003989511097521421, 'samples': 3119808, 'steps': 16248, 'loss/train': 0.6522131860256195} +01/27/2022 11:08:30 - INFO - codeparrot_training - Step 16249: {'lr': 0.00039893796825662676, 'samples': 3120000, 'steps': 16249, 'loss/train': 0.4519360810518265} +01/27/2022 11:08:33 - INFO - codeparrot_training - Step 16250: {'lr': 0.0003989248261231084, 'samples': 3120192, 'steps': 16250, 'loss/train': 0.3935438543558121} +01/27/2022 11:08:36 - INFO - codeparrot_training - Step 16251: {'lr': 0.0003989116833516433, 'samples': 3120384, 'steps': 16251, 'loss/train': 0.3833215981721878} +01/27/2022 11:08:39 - INFO - codeparrot_training - Step 16252: {'lr': 0.000398898539942288, 'samples': 3120576, 'steps': 16252, 'loss/train': 1.5999439358711243} +01/27/2022 11:08:44 - INFO - codeparrot_training - Step 16253: {'lr': 0.0003988853958950984, 'samples': 3120768, 'steps': 16253, 'loss/train': 0.3693583533167839} +01/27/2022 11:08:47 - INFO - codeparrot_training - Step 16254: {'lr': 0.00039887225121013124, 'samples': 3120960, 'steps': 16254, 'loss/train': 0.772210031747818} +01/27/2022 11:08:51 - INFO - codeparrot_training - Step 16255: {'lr': 0.0003988591058874426, 'samples': 3121152, 'steps': 16255, 'loss/train': 0.4513072371482849} +01/27/2022 11:08:54 - INFO - codeparrot_training - Step 16256: {'lr': 0.00039884595992708877, 'samples': 3121344, 'steps': 16256, 'loss/train': 0.8744835555553436} +01/27/2022 11:08:57 - INFO - codeparrot_training - Step 16257: {'lr': 0.0003988328133291261, 'samples': 3121536, 'steps': 16257, 'loss/train': 0.38357050716876984} +01/27/2022 11:09:00 - INFO - codeparrot_training - Step 16258: {'lr': 0.000398819666093611, 'samples': 3121728, 'steps': 16258, 'loss/train': 0.6371843218803406} +01/27/2022 11:09:03 - INFO - codeparrot_training - Step 16259: {'lr': 0.0003988065182205996, 'samples': 3121920, 'steps': 16259, 'loss/train': 1.0179716348648071} +01/27/2022 11:09:06 - INFO - codeparrot_training - Step 16260: {'lr': 0.0003987933697101484, 'samples': 3122112, 'steps': 16260, 'loss/train': 1.2850188910961151} +01/27/2022 11:09:09 - INFO - codeparrot_training - Step 16261: {'lr': 0.0003987802205623136, 'samples': 3122304, 'steps': 16261, 'loss/train': 0.7786278426647186} +01/27/2022 11:09:14 - INFO - codeparrot_training - Step 16262: {'lr': 0.0003987670707771516, 'samples': 3122496, 'steps': 16262, 'loss/train': 0.9967038631439209} +01/27/2022 11:09:17 - INFO - codeparrot_training - Step 16263: {'lr': 0.0003987539203547187, 'samples': 3122688, 'steps': 16263, 'loss/train': 0.6211754232645035} +01/27/2022 11:09:20 - INFO - codeparrot_training - Step 16264: {'lr': 0.00039874076929507124, 'samples': 3122880, 'steps': 16264, 'loss/train': 1.1608092784881592} +01/27/2022 11:09:23 - INFO - codeparrot_training - Step 16265: {'lr': 0.0003987276175982656, 'samples': 3123072, 'steps': 16265, 'loss/train': 0.8607359826564789} +01/27/2022 11:09:26 - INFO - codeparrot_training - Step 16266: {'lr': 0.00039871446526435806, 'samples': 3123264, 'steps': 16266, 'loss/train': 0.853412389755249} +01/27/2022 11:09:30 - INFO - codeparrot_training - Step 16267: {'lr': 0.00039870131229340495, 'samples': 3123456, 'steps': 16267, 'loss/train': 1.112040102481842} +01/27/2022 11:09:33 - INFO - codeparrot_training - Step 16268: {'lr': 0.00039868815868546257, 'samples': 3123648, 'steps': 16268, 'loss/train': 1.001338541507721} +01/27/2022 11:09:36 - INFO - codeparrot_training - Step 16269: {'lr': 0.00039867500444058747, 'samples': 3123840, 'steps': 16269, 'loss/train': 0.6718313097953796} +01/27/2022 11:09:39 - INFO - codeparrot_training - Step 16270: {'lr': 0.0003986618495588358, 'samples': 3124032, 'steps': 16270, 'loss/train': 1.0366871058940887} +01/27/2022 11:09:43 - INFO - codeparrot_training - Step 16271: {'lr': 0.00039864869404026394, 'samples': 3124224, 'steps': 16271, 'loss/train': 0.8711869418621063} +01/27/2022 11:09:47 - INFO - codeparrot_training - Step 16272: {'lr': 0.0003986355378849283, 'samples': 3124416, 'steps': 16272, 'loss/train': 1.0948247909545898} +01/27/2022 11:09:50 - INFO - codeparrot_training - Step 16273: {'lr': 0.00039862238109288523, 'samples': 3124608, 'steps': 16273, 'loss/train': 1.1524816453456879} +01/27/2022 11:09:53 - INFO - codeparrot_training - Step 16274: {'lr': 0.0003986092236641911, 'samples': 3124800, 'steps': 16274, 'loss/train': 0.6929799020290375} +01/27/2022 11:09:56 - INFO - codeparrot_training - Step 16275: {'lr': 0.00039859606559890215, 'samples': 3124992, 'steps': 16275, 'loss/train': 0.8650272488594055} +01/27/2022 11:09:59 - INFO - codeparrot_training - Step 16276: {'lr': 0.0003985829068970749, 'samples': 3125184, 'steps': 16276, 'loss/train': 0.8608489036560059} +01/27/2022 11:10:02 - INFO - codeparrot_training - Step 16277: {'lr': 0.00039856974755876563, 'samples': 3125376, 'steps': 16277, 'loss/train': 0.7365667819976807} +01/27/2022 11:10:06 - INFO - codeparrot_training - Step 16278: {'lr': 0.0003985565875840308, 'samples': 3125568, 'steps': 16278, 'loss/train': 0.5427868813276291} +01/27/2022 11:10:10 - INFO - codeparrot_training - Step 16279: {'lr': 0.0003985434269729267, 'samples': 3125760, 'steps': 16279, 'loss/train': 0.5338517725467682} +01/27/2022 11:10:13 - INFO - codeparrot_training - Step 16280: {'lr': 0.00039853026572550965, 'samples': 3125952, 'steps': 16280, 'loss/train': 0.9932606220245361} +01/27/2022 11:10:16 - INFO - codeparrot_training - Step 16281: {'lr': 0.00039851710384183615, 'samples': 3126144, 'steps': 16281, 'loss/train': 0.502830907702446} +01/27/2022 11:10:19 - INFO - codeparrot_training - Step 16282: {'lr': 0.0003985039413219626, 'samples': 3126336, 'steps': 16282, 'loss/train': 0.8430125713348389} +01/27/2022 11:10:22 - INFO - codeparrot_training - Step 16283: {'lr': 0.0003984907781659452, 'samples': 3126528, 'steps': 16283, 'loss/train': 0.8155592679977417} +01/27/2022 11:10:26 - INFO - codeparrot_training - Step 16284: {'lr': 0.00039847761437384054, 'samples': 3126720, 'steps': 16284, 'loss/train': 0.7957356870174408} +01/27/2022 11:10:29 - INFO - codeparrot_training - Step 16285: {'lr': 0.0003984644499457049, 'samples': 3126912, 'steps': 16285, 'loss/train': 0.9324469864368439} +01/27/2022 11:10:32 - INFO - codeparrot_training - Step 16286: {'lr': 0.0003984512848815948, 'samples': 3127104, 'steps': 16286, 'loss/train': 0.6606500744819641} +01/27/2022 11:10:35 - INFO - codeparrot_training - Step 16287: {'lr': 0.00039843811918156635, 'samples': 3127296, 'steps': 16287, 'loss/train': 0.8816904723644257} +01/27/2022 11:10:40 - INFO - codeparrot_training - Step 16288: {'lr': 0.0003984249528456762, 'samples': 3127488, 'steps': 16288, 'loss/train': 0.2783263474702835} +01/27/2022 11:10:43 - INFO - codeparrot_training - Step 16289: {'lr': 0.00039841178587398074, 'samples': 3127680, 'steps': 16289, 'loss/train': 0.7745550870895386} +01/27/2022 11:10:47 - INFO - codeparrot_training - Step 16290: {'lr': 0.0003983986182665362, 'samples': 3127872, 'steps': 16290, 'loss/train': 0.5719417333602905} +01/27/2022 11:10:50 - INFO - codeparrot_training - Step 16291: {'lr': 0.00039838545002339926, 'samples': 3128064, 'steps': 16291, 'loss/train': 0.6814579367637634} +01/27/2022 11:10:53 - INFO - codeparrot_training - Step 16292: {'lr': 0.0003983722811446261, 'samples': 3128256, 'steps': 16292, 'loss/train': 0.8080460429191589} +01/27/2022 11:10:56 - INFO - codeparrot_training - Step 16293: {'lr': 0.00039835911163027315, 'samples': 3128448, 'steps': 16293, 'loss/train': 0.7708510458469391} +01/27/2022 11:10:59 - INFO - codeparrot_training - Step 16294: {'lr': 0.00039834594148039693, 'samples': 3128640, 'steps': 16294, 'loss/train': 0.754031628370285} +01/27/2022 11:11:02 - INFO - codeparrot_training - Step 16295: {'lr': 0.0003983327706950538, 'samples': 3128832, 'steps': 16295, 'loss/train': 0.89482381939888} +01/27/2022 11:11:05 - INFO - codeparrot_training - Step 16296: {'lr': 0.00039831959927430017, 'samples': 3129024, 'steps': 16296, 'loss/train': 0.6457739621400833} +01/27/2022 11:11:10 - INFO - codeparrot_training - Step 16297: {'lr': 0.00039830642721819254, 'samples': 3129216, 'steps': 16297, 'loss/train': 1.07470241189003} +01/27/2022 11:11:13 - INFO - codeparrot_training - Step 16298: {'lr': 0.0003982932545267872, 'samples': 3129408, 'steps': 16298, 'loss/train': 0.8556700050830841} +01/27/2022 11:11:16 - INFO - codeparrot_training - Step 16299: {'lr': 0.00039828008120014057, 'samples': 3129600, 'steps': 16299, 'loss/train': 0.7670690417289734} +01/27/2022 11:11:20 - INFO - codeparrot_training - Step 16300: {'lr': 0.00039826690723830926, 'samples': 3129792, 'steps': 16300, 'loss/train': 0.8765671849250793} +01/27/2022 11:11:23 - INFO - codeparrot_training - Step 16301: {'lr': 0.00039825373264134955, 'samples': 3129984, 'steps': 16301, 'loss/train': 0.7851921021938324} +01/27/2022 11:11:26 - INFO - codeparrot_training - Step 16302: {'lr': 0.00039824055740931804, 'samples': 3130176, 'steps': 16302, 'loss/train': 0.9035835564136505} +01/27/2022 11:11:29 - INFO - codeparrot_training - Step 16303: {'lr': 0.0003982273815422709, 'samples': 3130368, 'steps': 16303, 'loss/train': 0.7076370120048523} +01/27/2022 11:11:32 - INFO - codeparrot_training - Step 16304: {'lr': 0.00039821420504026486, 'samples': 3130560, 'steps': 16304, 'loss/train': 1.009582668542862} +01/27/2022 11:11:35 - INFO - codeparrot_training - Step 16305: {'lr': 0.0003982010279033561, 'samples': 3130752, 'steps': 16305, 'loss/train': 0.8883113265037537} +01/27/2022 11:11:40 - INFO - codeparrot_training - Step 16306: {'lr': 0.0003981878501316013, 'samples': 3130944, 'steps': 16306, 'loss/train': 0.7432819604873657} +01/27/2022 11:11:44 - INFO - codeparrot_training - Step 16307: {'lr': 0.0003981746717250567, 'samples': 3131136, 'steps': 16307, 'loss/train': 0.8210342824459076} +01/27/2022 11:11:47 - INFO - codeparrot_training - Step 16308: {'lr': 0.000398161492683779, 'samples': 3131328, 'steps': 16308, 'loss/train': 0.4928121864795685} +01/27/2022 11:11:50 - INFO - codeparrot_training - Step 16309: {'lr': 0.0003981483130078244, 'samples': 3131520, 'steps': 16309, 'loss/train': 1.050301730632782} +01/27/2022 11:11:53 - INFO - codeparrot_training - Step 16310: {'lr': 0.0003981351326972495, 'samples': 3131712, 'steps': 16310, 'loss/train': 1.2259369790554047} +01/27/2022 11:11:56 - INFO - codeparrot_training - Step 16311: {'lr': 0.00039812195175211075, 'samples': 3131904, 'steps': 16311, 'loss/train': 1.1644804179668427} +01/27/2022 11:11:59 - INFO - codeparrot_training - Step 16312: {'lr': 0.0003981087701724645, 'samples': 3132096, 'steps': 16312, 'loss/train': 0.8268392980098724} +01/27/2022 11:12:02 - INFO - codeparrot_training - Step 16313: {'lr': 0.00039809558795836743, 'samples': 3132288, 'steps': 16313, 'loss/train': 0.8993506729602814} +01/27/2022 11:12:07 - INFO - codeparrot_training - Step 16314: {'lr': 0.00039808240510987584, 'samples': 3132480, 'steps': 16314, 'loss/train': 0.3010287433862686} +01/27/2022 11:12:10 - INFO - codeparrot_training - Step 16315: {'lr': 0.0003980692216270462, 'samples': 3132672, 'steps': 16315, 'loss/train': 0.9246939718723297} +01/27/2022 11:12:13 - INFO - codeparrot_training - Step 16316: {'lr': 0.00039805603750993514, 'samples': 3132864, 'steps': 16316, 'loss/train': 0.7390566766262054} +01/27/2022 11:12:16 - INFO - codeparrot_training - Step 16317: {'lr': 0.0003980428527585989, 'samples': 3133056, 'steps': 16317, 'loss/train': 0.3761061429977417} +01/27/2022 11:12:20 - INFO - codeparrot_training - Step 16318: {'lr': 0.0003980296673730942, 'samples': 3133248, 'steps': 16318, 'loss/train': 0.6053046137094498} +01/27/2022 11:12:23 - INFO - codeparrot_training - Step 16319: {'lr': 0.0003980164813534773, 'samples': 3133440, 'steps': 16319, 'loss/train': 0.8428775668144226} +01/27/2022 11:12:26 - INFO - codeparrot_training - Step 16320: {'lr': 0.0003980032946998049, 'samples': 3133632, 'steps': 16320, 'loss/train': 0.6845053732395172} +01/27/2022 11:12:29 - INFO - codeparrot_training - Step 16321: {'lr': 0.00039799010741213336, 'samples': 3133824, 'steps': 16321, 'loss/train': 0.34418465942144394} +01/27/2022 11:12:32 - INFO - codeparrot_training - Step 16322: {'lr': 0.0003979769194905192, 'samples': 3134016, 'steps': 16322, 'loss/train': 0.9932771623134613} +01/27/2022 11:12:37 - INFO - codeparrot_training - Step 16323: {'lr': 0.0003979637309350188, 'samples': 3134208, 'steps': 16323, 'loss/train': 1.0993483364582062} +01/27/2022 11:12:40 - INFO - codeparrot_training - Step 16324: {'lr': 0.0003979505417456889, 'samples': 3134400, 'steps': 16324, 'loss/train': 0.7312969118356705} +01/27/2022 11:12:43 - INFO - codeparrot_training - Step 16325: {'lr': 0.00039793735192258575, 'samples': 3134592, 'steps': 16325, 'loss/train': 0.8382711410522461} +01/27/2022 11:12:46 - INFO - codeparrot_training - Step 16326: {'lr': 0.000397924161465766, 'samples': 3134784, 'steps': 16326, 'loss/train': 0.6754876524209976} +01/27/2022 11:12:49 - INFO - codeparrot_training - Step 16327: {'lr': 0.0003979109703752861, 'samples': 3134976, 'steps': 16327, 'loss/train': 0.7692593336105347} +01/27/2022 11:12:52 - INFO - codeparrot_training - Step 16328: {'lr': 0.00039789777865120257, 'samples': 3135168, 'steps': 16328, 'loss/train': 0.8904886543750763} +01/27/2022 11:12:56 - INFO - codeparrot_training - Step 16329: {'lr': 0.00039788458629357195, 'samples': 3135360, 'steps': 16329, 'loss/train': 0.2693357840180397} +01/27/2022 11:12:59 - INFO - codeparrot_training - Step 16330: {'lr': 0.0003978713933024507, 'samples': 3135552, 'steps': 16330, 'loss/train': 0.7962928712368011} +01/27/2022 11:13:02 - INFO - codeparrot_training - Step 16331: {'lr': 0.0003978581996778954, 'samples': 3135744, 'steps': 16331, 'loss/train': 0.46128053963184357} +01/27/2022 11:13:07 - INFO - codeparrot_training - Step 16332: {'lr': 0.0003978450054199625, 'samples': 3135936, 'steps': 16332, 'loss/train': 1.4301784336566925} +01/27/2022 11:13:11 - INFO - codeparrot_training - Step 16333: {'lr': 0.0003978318105287085, 'samples': 3136128, 'steps': 16333, 'loss/train': 0.9687565863132477} +01/27/2022 11:13:14 - INFO - codeparrot_training - Step 16334: {'lr': 0.00039781861500419, 'samples': 3136320, 'steps': 16334, 'loss/train': 0.1528499834239483} +01/27/2022 11:13:17 - INFO - codeparrot_training - Step 16335: {'lr': 0.00039780541884646347, 'samples': 3136512, 'steps': 16335, 'loss/train': 0.970037430524826} +01/27/2022 11:13:20 - INFO - codeparrot_training - Step 16336: {'lr': 0.0003977922220555855, 'samples': 3136704, 'steps': 16336, 'loss/train': 0.7505045235157013} +01/27/2022 11:13:23 - INFO - codeparrot_training - Step 16337: {'lr': 0.0003977790246316125, 'samples': 3136896, 'steps': 16337, 'loss/train': 0.7529717087745667} +01/27/2022 11:13:26 - INFO - codeparrot_training - Step 16338: {'lr': 0.00039776582657460115, 'samples': 3137088, 'steps': 16338, 'loss/train': 0.5797615349292755} +01/27/2022 11:13:29 - INFO - codeparrot_training - Step 16339: {'lr': 0.000397752627884608, 'samples': 3137280, 'steps': 16339, 'loss/train': 0.9103802740573883} +01/27/2022 11:13:33 - INFO - codeparrot_training - Step 16340: {'lr': 0.0003977394285616893, 'samples': 3137472, 'steps': 16340, 'loss/train': 0.9499474167823792} +01/27/2022 11:13:37 - INFO - codeparrot_training - Step 16341: {'lr': 0.000397726228605902, 'samples': 3137664, 'steps': 16341, 'loss/train': 1.1836229860782623} +01/27/2022 11:13:40 - INFO - codeparrot_training - Step 16342: {'lr': 0.00039771302801730235, 'samples': 3137856, 'steps': 16342, 'loss/train': 1.1383735835552216} +01/27/2022 11:13:43 - INFO - codeparrot_training - Step 16343: {'lr': 0.00039769982679594703, 'samples': 3138048, 'steps': 16343, 'loss/train': 0.8605233728885651} +01/27/2022 11:13:46 - INFO - codeparrot_training - Step 16344: {'lr': 0.0003976866249418925, 'samples': 3138240, 'steps': 16344, 'loss/train': 0.7122722268104553} +01/27/2022 11:13:50 - INFO - codeparrot_training - Step 16345: {'lr': 0.0003976734224551954, 'samples': 3138432, 'steps': 16345, 'loss/train': 1.660026490688324} +01/27/2022 11:13:53 - INFO - codeparrot_training - Step 16346: {'lr': 0.0003976602193359122, 'samples': 3138624, 'steps': 16346, 'loss/train': 1.1642853319644928} +01/27/2022 11:13:56 - INFO - codeparrot_training - Step 16347: {'lr': 0.00039764701558409955, 'samples': 3138816, 'steps': 16347, 'loss/train': 0.8584884703159332} +01/27/2022 11:13:59 - INFO - codeparrot_training - Step 16348: {'lr': 0.000397633811199814, 'samples': 3139008, 'steps': 16348, 'loss/train': 0.7647995352745056} +01/27/2022 11:14:02 - INFO - codeparrot_training - Step 16349: {'lr': 0.000397620606183112, 'samples': 3139200, 'steps': 16349, 'loss/train': 0.8294972777366638} +01/27/2022 11:14:07 - INFO - codeparrot_training - Step 16350: {'lr': 0.00039760740053405033, 'samples': 3139392, 'steps': 16350, 'loss/train': 0.8636766672134399} +01/27/2022 11:14:11 - INFO - codeparrot_training - Step 16351: {'lr': 0.00039759419425268526, 'samples': 3139584, 'steps': 16351, 'loss/train': 0.9668554365634918} +01/27/2022 11:14:14 - INFO - codeparrot_training - Step 16352: {'lr': 0.00039758098733907364, 'samples': 3139776, 'steps': 16352, 'loss/train': 0.9940418601036072} +01/27/2022 11:14:17 - INFO - codeparrot_training - Step 16353: {'lr': 0.00039756777979327193, 'samples': 3139968, 'steps': 16353, 'loss/train': 0.7608877122402191} +01/27/2022 11:14:20 - INFO - codeparrot_training - Step 16354: {'lr': 0.0003975545716153367, 'samples': 3140160, 'steps': 16354, 'loss/train': 0.759500652551651} +01/27/2022 11:14:23 - INFO - codeparrot_training - Step 16355: {'lr': 0.0003975413628053245, 'samples': 3140352, 'steps': 16355, 'loss/train': 0.6195583194494247} +01/27/2022 11:14:26 - INFO - codeparrot_training - Step 16356: {'lr': 0.000397528153363292, 'samples': 3140544, 'steps': 16356, 'loss/train': 0.7404384166002274} +01/27/2022 11:14:29 - INFO - codeparrot_training - Step 16357: {'lr': 0.00039751494328929565, 'samples': 3140736, 'steps': 16357, 'loss/train': 1.1545848548412323} +01/27/2022 11:14:33 - INFO - codeparrot_training - Step 16358: {'lr': 0.00039750173258339225, 'samples': 3140928, 'steps': 16358, 'loss/train': 1.3863947689533234} +01/27/2022 11:14:37 - INFO - codeparrot_training - Step 16359: {'lr': 0.00039748852124563816, 'samples': 3141120, 'steps': 16359, 'loss/train': 1.242002785205841} +01/27/2022 11:14:40 - INFO - codeparrot_training - Step 16360: {'lr': 0.0003974753092760901, 'samples': 3141312, 'steps': 16360, 'loss/train': 0.4398142397403717} +01/27/2022 11:14:44 - INFO - codeparrot_training - Step 16361: {'lr': 0.00039746209667480473, 'samples': 3141504, 'steps': 16361, 'loss/train': 0.8595219254493713} +01/27/2022 11:14:47 - INFO - codeparrot_training - Step 16362: {'lr': 0.00039744888344183846, 'samples': 3141696, 'steps': 16362, 'loss/train': 0.8012799024581909} +01/27/2022 11:14:50 - INFO - codeparrot_training - Step 16363: {'lr': 0.00039743566957724805, 'samples': 3141888, 'steps': 16363, 'loss/train': 0.24459897726774216} +01/27/2022 11:14:53 - INFO - codeparrot_training - Step 16364: {'lr': 0.00039742245508109, 'samples': 3142080, 'steps': 16364, 'loss/train': 0.8025190830230713} +01/27/2022 11:14:56 - INFO - codeparrot_training - Step 16365: {'lr': 0.000397409239953421, 'samples': 3142272, 'steps': 16365, 'loss/train': 0.6357373595237732} +01/27/2022 11:14:59 - INFO - codeparrot_training - Step 16366: {'lr': 0.00039739602419429755, 'samples': 3142464, 'steps': 16366, 'loss/train': 0.8606222569942474} +01/27/2022 11:15:02 - INFO - codeparrot_training - Step 16367: {'lr': 0.00039738280780377645, 'samples': 3142656, 'steps': 16367, 'loss/train': 0.6234107315540314} +01/27/2022 11:15:07 - INFO - codeparrot_training - Step 16368: {'lr': 0.0003973695907819141, 'samples': 3142848, 'steps': 16368, 'loss/train': 0.5991390198469162} +01/27/2022 11:15:10 - INFO - codeparrot_training - Step 16369: {'lr': 0.0003973563731287673, 'samples': 3143040, 'steps': 16369, 'loss/train': 0.7862134873867035} +01/27/2022 11:15:14 - INFO - codeparrot_training - Step 16370: {'lr': 0.00039734315484439255, 'samples': 3143232, 'steps': 16370, 'loss/train': 0.5914807319641113} +01/27/2022 11:15:17 - INFO - codeparrot_training - Step 16371: {'lr': 0.0003973299359288465, 'samples': 3143424, 'steps': 16371, 'loss/train': 0.4683467298746109} +01/27/2022 11:15:20 - INFO - codeparrot_training - Step 16372: {'lr': 0.0003973167163821858, 'samples': 3143616, 'steps': 16372, 'loss/train': 0.8322366178035736} +01/27/2022 11:15:23 - INFO - codeparrot_training - Step 16373: {'lr': 0.0003973034962044671, 'samples': 3143808, 'steps': 16373, 'loss/train': 0.7112137377262115} +01/27/2022 11:15:26 - INFO - codeparrot_training - Step 16374: {'lr': 0.00039729027539574696, 'samples': 3144000, 'steps': 16374, 'loss/train': 0.5096184611320496} +01/27/2022 11:15:29 - INFO - codeparrot_training - Step 16375: {'lr': 0.00039727705395608203, 'samples': 3144192, 'steps': 16375, 'loss/train': 1.239459067583084} +01/27/2022 11:15:34 - INFO - codeparrot_training - Step 16376: {'lr': 0.00039726383188552907, 'samples': 3144384, 'steps': 16376, 'loss/train': 0.7693246901035309} +01/27/2022 11:15:37 - INFO - codeparrot_training - Step 16377: {'lr': 0.00039725060918414446, 'samples': 3144576, 'steps': 16377, 'loss/train': 1.109573632478714} +01/27/2022 11:15:40 - INFO - codeparrot_training - Step 16378: {'lr': 0.0003972373858519851, 'samples': 3144768, 'steps': 16378, 'loss/train': 1.333444207906723} +01/27/2022 11:15:43 - INFO - codeparrot_training - Step 16379: {'lr': 0.00039722416188910754, 'samples': 3144960, 'steps': 16379, 'loss/train': 1.1786867380142212} +01/27/2022 11:15:46 - INFO - codeparrot_training - Step 16380: {'lr': 0.00039721093729556836, 'samples': 3145152, 'steps': 16380, 'loss/train': 0.9484879374504089} +01/27/2022 11:15:49 - INFO - codeparrot_training - Step 16381: {'lr': 0.0003971977120714243, 'samples': 3145344, 'steps': 16381, 'loss/train': 0.8165819048881531} +01/27/2022 11:15:53 - INFO - codeparrot_training - Step 16382: {'lr': 0.000397184486216732, 'samples': 3145536, 'steps': 16382, 'loss/train': 0.8462325632572174} +01/27/2022 11:15:56 - INFO - codeparrot_training - Step 16383: {'lr': 0.0003971712597315481, 'samples': 3145728, 'steps': 16383, 'loss/train': 1.2614806294441223} +01/27/2022 11:15:59 - INFO - codeparrot_training - Step 16384: {'lr': 0.0003971580326159292, 'samples': 3145920, 'steps': 16384, 'loss/train': 0.8884590268135071} +01/27/2022 11:16:04 - INFO - codeparrot_training - Step 16385: {'lr': 0.0003971448048699321, 'samples': 3146112, 'steps': 16385, 'loss/train': 0.6288697868585587} +01/27/2022 11:16:07 - INFO - codeparrot_training - Step 16386: {'lr': 0.00039713157649361327, 'samples': 3146304, 'steps': 16386, 'loss/train': 1.0685488879680634} +01/27/2022 11:16:10 - INFO - codeparrot_training - Step 16387: {'lr': 0.00039711834748702956, 'samples': 3146496, 'steps': 16387, 'loss/train': 0.9877242743968964} +01/27/2022 11:16:13 - INFO - codeparrot_training - Step 16388: {'lr': 0.0003971051178502375, 'samples': 3146688, 'steps': 16388, 'loss/train': 0.6256601214408875} +01/27/2022 11:16:16 - INFO - codeparrot_training - Step 16389: {'lr': 0.00039709188758329394, 'samples': 3146880, 'steps': 16389, 'loss/train': 0.6501017957925797} +01/27/2022 11:16:20 - INFO - codeparrot_training - Step 16390: {'lr': 0.0003970786566862553, 'samples': 3147072, 'steps': 16390, 'loss/train': 1.1275242269039154} +01/27/2022 11:16:23 - INFO - codeparrot_training - Step 16391: {'lr': 0.00039706542515917853, 'samples': 3147264, 'steps': 16391, 'loss/train': 0.6267930418252945} +01/27/2022 11:16:26 - INFO - codeparrot_training - Step 16392: {'lr': 0.00039705219300212015, 'samples': 3147456, 'steps': 16392, 'loss/train': 0.8122687339782715} +01/27/2022 11:16:29 - INFO - codeparrot_training - Step 16393: {'lr': 0.00039703896021513684, 'samples': 3147648, 'steps': 16393, 'loss/train': 0.8229024410247803} +01/27/2022 11:16:36 - INFO - codeparrot_training - Step 16394: {'lr': 0.0003970257267982853, 'samples': 3147840, 'steps': 16394, 'loss/train': 1.2244775891304016} +01/27/2022 11:16:39 - INFO - codeparrot_training - Step 16395: {'lr': 0.0003970124927516222, 'samples': 3148032, 'steps': 16395, 'loss/train': 0.433227002620697} +01/27/2022 11:16:42 - INFO - codeparrot_training - Step 16396: {'lr': 0.0003969992580752043, 'samples': 3148224, 'steps': 16396, 'loss/train': 0.5091579258441925} +01/27/2022 11:16:45 - INFO - codeparrot_training - Step 16397: {'lr': 0.00039698602276908826, 'samples': 3148416, 'steps': 16397, 'loss/train': 0.7512785196304321} +01/27/2022 11:16:48 - INFO - codeparrot_training - Step 16398: {'lr': 0.0003969727868333308, 'samples': 3148608, 'steps': 16398, 'loss/train': 0.41959767043590546} +01/27/2022 11:16:51 - INFO - codeparrot_training - Step 16399: {'lr': 0.00039695955026798857, 'samples': 3148800, 'steps': 16399, 'loss/train': 0.6502413153648376} +01/27/2022 11:16:55 - INFO - codeparrot_training - Step 16400: {'lr': 0.0003969463130731183, 'samples': 3148992, 'steps': 16400, 'loss/train': 0.9085874855518341} +01/27/2022 11:16:58 - INFO - codeparrot_training - Step 16401: {'lr': 0.00039693307524877664, 'samples': 3149184, 'steps': 16401, 'loss/train': 0.9677371680736542} +01/27/2022 11:17:01 - INFO - codeparrot_training - Step 16402: {'lr': 0.0003969198367950204, 'samples': 3149376, 'steps': 16402, 'loss/train': 0.7147872000932693} +01/27/2022 11:17:05 - INFO - codeparrot_training - Step 16403: {'lr': 0.00039690659771190616, 'samples': 3149568, 'steps': 16403, 'loss/train': 0.10759841278195381} +01/27/2022 11:17:08 - INFO - codeparrot_training - Step 16404: {'lr': 0.0003968933579994908, 'samples': 3149760, 'steps': 16404, 'loss/train': 0.987971305847168} +01/27/2022 11:17:12 - INFO - codeparrot_training - Step 16405: {'lr': 0.0003968801176578309, 'samples': 3149952, 'steps': 16405, 'loss/train': 0.45827150344848633} +01/27/2022 11:17:15 - INFO - codeparrot_training - Step 16406: {'lr': 0.00039686687668698316, 'samples': 3150144, 'steps': 16406, 'loss/train': 1.1288044452667236} +01/27/2022 11:17:18 - INFO - codeparrot_training - Step 16407: {'lr': 0.00039685363508700443, 'samples': 3150336, 'steps': 16407, 'loss/train': 0.8957827091217041} +01/27/2022 11:17:21 - INFO - codeparrot_training - Step 16408: {'lr': 0.00039684039285795133, 'samples': 3150528, 'steps': 16408, 'loss/train': 0.41786250472068787} +01/27/2022 11:17:24 - INFO - codeparrot_training - Step 16409: {'lr': 0.0003968271499998806, 'samples': 3150720, 'steps': 16409, 'loss/train': 0.616459384560585} +01/27/2022 11:17:27 - INFO - codeparrot_training - Step 16410: {'lr': 0.000396813906512849, 'samples': 3150912, 'steps': 16410, 'loss/train': 0.1511295922100544} +01/27/2022 11:17:30 - INFO - codeparrot_training - Step 16411: {'lr': 0.00039680066239691325, 'samples': 3151104, 'steps': 16411, 'loss/train': 0.7987346649169922} +01/27/2022 11:17:35 - INFO - codeparrot_training - Step 16412: {'lr': 0.00039678741765213006, 'samples': 3151296, 'steps': 16412, 'loss/train': 0.8277180790901184} +01/27/2022 11:17:38 - INFO - codeparrot_training - Step 16413: {'lr': 0.00039677417227855624, 'samples': 3151488, 'steps': 16413, 'loss/train': 0.8293441236019135} +01/27/2022 11:17:41 - INFO - codeparrot_training - Step 16414: {'lr': 0.0003967609262762484, 'samples': 3151680, 'steps': 16414, 'loss/train': 0.6646831780672073} +01/27/2022 11:17:44 - INFO - codeparrot_training - Step 16415: {'lr': 0.0003967476796452634, 'samples': 3151872, 'steps': 16415, 'loss/train': 1.3371411859989166} +01/27/2022 11:17:47 - INFO - codeparrot_training - Step 16416: {'lr': 0.00039673443238565786, 'samples': 3152064, 'steps': 16416, 'loss/train': 0.9867611825466156} +01/27/2022 11:17:51 - INFO - codeparrot_training - Step 16417: {'lr': 0.0003967211844974887, 'samples': 3152256, 'steps': 16417, 'loss/train': 0.7275371253490448} +01/27/2022 11:17:54 - INFO - codeparrot_training - Step 16418: {'lr': 0.0003967079359808125, 'samples': 3152448, 'steps': 16418, 'loss/train': 0.048546502366662025} +01/27/2022 11:17:57 - INFO - codeparrot_training - Step 16419: {'lr': 0.0003966946868356861, 'samples': 3152640, 'steps': 16419, 'loss/train': 0.9831126630306244} +01/27/2022 11:18:03 - INFO - codeparrot_training - Step 16420: {'lr': 0.0003966814370621663, 'samples': 3152832, 'steps': 16420, 'loss/train': 0.7818227112293243} +01/27/2022 11:18:06 - INFO - codeparrot_training - Step 16421: {'lr': 0.00039666818666030974, 'samples': 3153024, 'steps': 16421, 'loss/train': 1.1001561284065247} +01/27/2022 11:18:09 - INFO - codeparrot_training - Step 16422: {'lr': 0.0003966549356301733, 'samples': 3153216, 'steps': 16422, 'loss/train': 0.5559077560901642} +01/27/2022 11:18:13 - INFO - codeparrot_training - Step 16423: {'lr': 0.0003966416839718136, 'samples': 3153408, 'steps': 16423, 'loss/train': 0.8029322326183319} +01/27/2022 11:18:16 - INFO - codeparrot_training - Step 16424: {'lr': 0.00039662843168528756, 'samples': 3153600, 'steps': 16424, 'loss/train': 0.7819757759571075} +01/27/2022 11:18:19 - INFO - codeparrot_training - Step 16425: {'lr': 0.00039661517877065183, 'samples': 3153792, 'steps': 16425, 'loss/train': 0.7896585166454315} +01/27/2022 11:18:22 - INFO - codeparrot_training - Step 16426: {'lr': 0.0003966019252279633, 'samples': 3153984, 'steps': 16426, 'loss/train': 1.0619692504405975} +01/27/2022 11:18:25 - INFO - codeparrot_training - Step 16427: {'lr': 0.00039658867105727856, 'samples': 3154176, 'steps': 16427, 'loss/train': 0.4249438941478729} +01/27/2022 11:18:28 - INFO - codeparrot_training - Step 16428: {'lr': 0.0003965754162586547, 'samples': 3154368, 'steps': 16428, 'loss/train': 1.3989788889884949} +01/27/2022 11:18:33 - INFO - codeparrot_training - Step 16429: {'lr': 0.0003965621608321481, 'samples': 3154560, 'steps': 16429, 'loss/train': 0.5543275773525238} +01/27/2022 11:18:36 - INFO - codeparrot_training - Step 16430: {'lr': 0.0003965489047778158, 'samples': 3154752, 'steps': 16430, 'loss/train': 0.22664880752563477} +01/27/2022 11:18:39 - INFO - codeparrot_training - Step 16431: {'lr': 0.0003965356480957145, 'samples': 3154944, 'steps': 16431, 'loss/train': 0.4697231501340866} +01/27/2022 11:18:42 - INFO - codeparrot_training - Step 16432: {'lr': 0.0003965223907859011, 'samples': 3155136, 'steps': 16432, 'loss/train': 0.6264688074588776} +01/27/2022 11:18:45 - INFO - codeparrot_training - Step 16433: {'lr': 0.00039650913284843225, 'samples': 3155328, 'steps': 16433, 'loss/train': 0.8311588168144226} +01/27/2022 11:18:49 - INFO - codeparrot_training - Step 16434: {'lr': 0.00039649587428336474, 'samples': 3155520, 'steps': 16434, 'loss/train': 0.1469871997833252} +01/27/2022 11:18:52 - INFO - codeparrot_training - Step 16435: {'lr': 0.00039648261509075554, 'samples': 3155712, 'steps': 16435, 'loss/train': 0.458076149225235} +01/27/2022 11:18:55 - INFO - codeparrot_training - Step 16436: {'lr': 0.00039646935527066124, 'samples': 3155904, 'steps': 16436, 'loss/train': 1.2633817791938782} +01/27/2022 11:18:58 - INFO - codeparrot_training - Step 16437: {'lr': 0.0003964560948231388, 'samples': 3156096, 'steps': 16437, 'loss/train': 0.6900997012853622} +01/27/2022 11:19:05 - INFO - codeparrot_training - Step 16438: {'lr': 0.0003964428337482449, 'samples': 3156288, 'steps': 16438, 'loss/train': 0.8631270825862885} +01/27/2022 11:19:08 - INFO - codeparrot_training - Step 16439: {'lr': 0.00039642957204603647, 'samples': 3156480, 'steps': 16439, 'loss/train': 0.4677002280950546} +01/27/2022 11:19:11 - INFO - codeparrot_training - Step 16440: {'lr': 0.0003964163097165702, 'samples': 3156672, 'steps': 16440, 'loss/train': 0.7438879609107971} +01/27/2022 11:19:14 - INFO - codeparrot_training - Step 16441: {'lr': 0.0003964030467599029, 'samples': 3156864, 'steps': 16441, 'loss/train': 0.5567301660776138} +01/27/2022 11:19:17 - INFO - codeparrot_training - Step 16442: {'lr': 0.00039638978317609155, 'samples': 3157056, 'steps': 16442, 'loss/train': 0.8895531892776489} +01/27/2022 11:19:20 - INFO - codeparrot_training - Step 16443: {'lr': 0.0003963765189651928, 'samples': 3157248, 'steps': 16443, 'loss/train': 0.70783831179142} +01/27/2022 11:19:23 - INFO - codeparrot_training - Step 16444: {'lr': 0.0003963632541272635, 'samples': 3157440, 'steps': 16444, 'loss/train': 0.9935088157653809} +01/27/2022 11:19:27 - INFO - codeparrot_training - Step 16445: {'lr': 0.00039634998866236047, 'samples': 3157632, 'steps': 16445, 'loss/train': 0.30726467818021774} +01/27/2022 11:19:31 - INFO - codeparrot_training - Step 16446: {'lr': 0.0003963367225705406, 'samples': 3157824, 'steps': 16446, 'loss/train': 1.2245839834213257} +01/27/2022 11:19:34 - INFO - codeparrot_training - Step 16447: {'lr': 0.0003963234558518607, 'samples': 3158016, 'steps': 16447, 'loss/train': 0.6575295925140381} +01/27/2022 11:19:37 - INFO - codeparrot_training - Step 16448: {'lr': 0.0003963101885063776, 'samples': 3158208, 'steps': 16448, 'loss/train': 1.1062726378440857} +01/27/2022 11:19:41 - INFO - codeparrot_training - Step 16449: {'lr': 0.000396296920534148, 'samples': 3158400, 'steps': 16449, 'loss/train': 1.0263479053974152} +01/27/2022 11:19:44 - INFO - codeparrot_training - Step 16450: {'lr': 0.000396283651935229, 'samples': 3158592, 'steps': 16450, 'loss/train': 1.0904256999492645} +01/27/2022 11:19:47 - INFO - codeparrot_training - Step 16451: {'lr': 0.0003962703827096771, 'samples': 3158784, 'steps': 16451, 'loss/train': 1.1059216260910034} +01/27/2022 11:19:50 - INFO - codeparrot_training - Step 16452: {'lr': 0.00039625711285754943, 'samples': 3158976, 'steps': 16452, 'loss/train': 1.5165544152259827} +01/27/2022 11:19:53 - INFO - codeparrot_training - Step 16453: {'lr': 0.00039624384237890275, 'samples': 3159168, 'steps': 16453, 'loss/train': 0.34100617468357086} +01/27/2022 11:19:56 - INFO - codeparrot_training - Step 16454: {'lr': 0.00039623057127379386, 'samples': 3159360, 'steps': 16454, 'loss/train': 0.6216109246015549} +01/27/2022 11:20:01 - INFO - codeparrot_training - Step 16455: {'lr': 0.0003962172995422796, 'samples': 3159552, 'steps': 16455, 'loss/train': 1.0587633848190308} +01/27/2022 11:20:04 - INFO - codeparrot_training - Step 16456: {'lr': 0.00039620402718441687, 'samples': 3159744, 'steps': 16456, 'loss/train': 0.8901599049568176} +01/27/2022 11:20:07 - INFO - codeparrot_training - Step 16457: {'lr': 0.0003961907542002626, 'samples': 3159936, 'steps': 16457, 'loss/train': 0.12307662516832352} +01/27/2022 11:20:10 - INFO - codeparrot_training - Step 16458: {'lr': 0.00039617748058987345, 'samples': 3160128, 'steps': 16458, 'loss/train': 0.546050637960434} +01/27/2022 11:20:13 - INFO - codeparrot_training - Step 16459: {'lr': 0.0003961642063533065, 'samples': 3160320, 'steps': 16459, 'loss/train': 0.47912295162677765} +01/27/2022 11:20:17 - INFO - codeparrot_training - Step 16460: {'lr': 0.0003961509314906184, 'samples': 3160512, 'steps': 16460, 'loss/train': 0.437875896692276} +01/27/2022 11:20:20 - INFO - codeparrot_training - Step 16461: {'lr': 0.0003961376560018662, 'samples': 3160704, 'steps': 16461, 'loss/train': 0.9406188726425171} +01/27/2022 11:20:23 - INFO - codeparrot_training - Step 16462: {'lr': 0.0003961243798871066, 'samples': 3160896, 'steps': 16462, 'loss/train': 0.7530031800270081} +01/27/2022 11:20:26 - INFO - codeparrot_training - Step 16463: {'lr': 0.00039611110314639663, 'samples': 3161088, 'steps': 16463, 'loss/train': 0.8586059510707855} +01/27/2022 11:20:32 - INFO - codeparrot_training - Step 16464: {'lr': 0.00039609782577979306, 'samples': 3161280, 'steps': 16464, 'loss/train': 0.9285542070865631} +01/27/2022 11:20:36 - INFO - codeparrot_training - Step 16465: {'lr': 0.0003960845477873528, 'samples': 3161472, 'steps': 16465, 'loss/train': 0.984802633523941} +01/27/2022 11:20:39 - INFO - codeparrot_training - Step 16466: {'lr': 0.00039607126916913274, 'samples': 3161664, 'steps': 16466, 'loss/train': 0.9545544683933258} +01/27/2022 11:20:42 - INFO - codeparrot_training - Step 16467: {'lr': 0.00039605798992518973, 'samples': 3161856, 'steps': 16467, 'loss/train': 1.2074414491653442} +01/27/2022 11:20:45 - INFO - codeparrot_training - Step 16468: {'lr': 0.00039604471005558065, 'samples': 3162048, 'steps': 16468, 'loss/train': 0.8222541511058807} +01/27/2022 11:20:48 - INFO - codeparrot_training - Step 16469: {'lr': 0.0003960314295603624, 'samples': 3162240, 'steps': 16469, 'loss/train': 1.9956028461456299} +01/27/2022 11:20:51 - INFO - codeparrot_training - Step 16470: {'lr': 0.00039601814843959193, 'samples': 3162432, 'steps': 16470, 'loss/train': 1.0499126315116882} +01/27/2022 11:20:54 - INFO - codeparrot_training - Step 16471: {'lr': 0.00039600486669332603, 'samples': 3162624, 'steps': 16471, 'loss/train': 0.0696211326867342} +01/27/2022 11:20:58 - INFO - codeparrot_training - Step 16472: {'lr': 0.00039599158432162163, 'samples': 3162816, 'steps': 16472, 'loss/train': 0.3520548865199089} +01/27/2022 11:21:02 - INFO - codeparrot_training - Step 16473: {'lr': 0.0003959783013245357, 'samples': 3163008, 'steps': 16473, 'loss/train': 0.927079439163208} +01/27/2022 11:21:05 - INFO - codeparrot_training - Step 16474: {'lr': 0.000395965017702125, 'samples': 3163200, 'steps': 16474, 'loss/train': 0.4509286880493164} +01/27/2022 11:21:08 - INFO - codeparrot_training - Step 16475: {'lr': 0.00039595173345444656, 'samples': 3163392, 'steps': 16475, 'loss/train': 1.194792777299881} +01/27/2022 11:21:11 - INFO - codeparrot_training - Step 16476: {'lr': 0.0003959384485815573, 'samples': 3163584, 'steps': 16476, 'loss/train': 0.5340692549943924} +01/27/2022 11:21:15 - INFO - codeparrot_training - Step 16477: {'lr': 0.000395925163083514, 'samples': 3163776, 'steps': 16477, 'loss/train': 1.0951342284679413} +01/27/2022 11:21:18 - INFO - codeparrot_training - Step 16478: {'lr': 0.00039591187696037366, 'samples': 3163968, 'steps': 16478, 'loss/train': 0.8578071892261505} +01/27/2022 11:21:21 - INFO - codeparrot_training - Step 16479: {'lr': 0.0003958985902121931, 'samples': 3164160, 'steps': 16479, 'loss/train': 0.8020130395889282} +01/27/2022 11:21:24 - INFO - codeparrot_training - Step 16480: {'lr': 0.00039588530283902936, 'samples': 3164352, 'steps': 16480, 'loss/train': 0.36279820650815964} +01/27/2022 11:21:27 - INFO - codeparrot_training - Step 16481: {'lr': 0.00039587201484093937, 'samples': 3164544, 'steps': 16481, 'loss/train': 0.9909215569496155} +01/27/2022 11:21:33 - INFO - codeparrot_training - Step 16482: {'lr': 0.0003958587262179799, 'samples': 3164736, 'steps': 16482, 'loss/train': 0.7821823954582214} +01/27/2022 11:21:37 - INFO - codeparrot_training - Step 16483: {'lr': 0.00039584543697020804, 'samples': 3164928, 'steps': 16483, 'loss/train': 0.8446026742458344} +01/27/2022 11:21:40 - INFO - codeparrot_training - Step 16484: {'lr': 0.00039583214709768054, 'samples': 3165120, 'steps': 16484, 'loss/train': 1.0304425656795502} +01/27/2022 11:21:43 - INFO - codeparrot_training - Step 16485: {'lr': 0.00039581885660045445, 'samples': 3165312, 'steps': 16485, 'loss/train': 1.0588009357452393} +01/27/2022 11:21:46 - INFO - codeparrot_training - Step 16486: {'lr': 0.0003958055654785867, 'samples': 3165504, 'steps': 16486, 'loss/train': 0.6049697399139404} +01/27/2022 11:21:49 - INFO - codeparrot_training - Step 16487: {'lr': 0.0003957922737321343, 'samples': 3165696, 'steps': 16487, 'loss/train': 0.9308875501155853} +01/27/2022 11:21:52 - INFO - codeparrot_training - Step 16488: {'lr': 0.00039577898136115397, 'samples': 3165888, 'steps': 16488, 'loss/train': 0.728198915719986} +01/27/2022 11:21:56 - INFO - codeparrot_training - Step 16489: {'lr': 0.00039576568836570283, 'samples': 3166080, 'steps': 16489, 'loss/train': 0.5297000259160995} +01/27/2022 11:22:00 - INFO - codeparrot_training - Step 16490: {'lr': 0.0003957523947458377, 'samples': 3166272, 'steps': 16490, 'loss/train': 0.6964269429445267} +01/27/2022 11:22:03 - INFO - codeparrot_training - Step 16491: {'lr': 0.00039573910050161564, 'samples': 3166464, 'steps': 16491, 'loss/train': 0.696523904800415} +01/27/2022 11:22:06 - INFO - codeparrot_training - Step 16492: {'lr': 0.0003957258056330936, 'samples': 3166656, 'steps': 16492, 'loss/train': 1.0968183875083923} +01/27/2022 11:22:09 - INFO - codeparrot_training - Step 16493: {'lr': 0.00039571251014032847, 'samples': 3166848, 'steps': 16493, 'loss/train': 1.0563878417015076} +01/27/2022 11:22:12 - INFO - codeparrot_training - Step 16494: {'lr': 0.00039569921402337715, 'samples': 3167040, 'steps': 16494, 'loss/train': 0.7124888151884079} +01/27/2022 11:22:16 - INFO - codeparrot_training - Step 16495: {'lr': 0.00039568591728229667, 'samples': 3167232, 'steps': 16495, 'loss/train': 0.4610670804977417} +01/27/2022 11:22:19 - INFO - codeparrot_training - Step 16496: {'lr': 0.00039567261991714406, 'samples': 3167424, 'steps': 16496, 'loss/train': 1.0298214554786682} +01/27/2022 11:22:22 - INFO - codeparrot_training - Step 16497: {'lr': 0.0003956593219279761, 'samples': 3167616, 'steps': 16497, 'loss/train': 0.6108525842428207} +01/27/2022 11:22:25 - INFO - codeparrot_training - Step 16498: {'lr': 0.00039564602331484993, 'samples': 3167808, 'steps': 16498, 'loss/train': 0.7337829619646072} +01/27/2022 11:22:29 - INFO - codeparrot_training - Step 16499: {'lr': 0.0003956327240778224, 'samples': 3168000, 'steps': 16499, 'loss/train': 0.33012619614601135} +01/27/2022 11:22:33 - INFO - codeparrot_training - Step 16500: {'lr': 0.00039561942421695057, 'samples': 3168192, 'steps': 16500, 'loss/train': 1.0325500667095184} +01/27/2022 11:22:36 - INFO - codeparrot_training - Step 16501: {'lr': 0.00039560612373229135, 'samples': 3168384, 'steps': 16501, 'loss/train': 1.0396359264850616} +01/27/2022 11:22:39 - INFO - codeparrot_training - Step 16502: {'lr': 0.0003955928226239017, 'samples': 3168576, 'steps': 16502, 'loss/train': 0.7965084314346313} +01/27/2022 11:22:42 - INFO - codeparrot_training - Step 16503: {'lr': 0.00039557952089183863, 'samples': 3168768, 'steps': 16503, 'loss/train': 0.3303249031305313} +01/27/2022 11:22:45 - INFO - codeparrot_training - Step 16504: {'lr': 0.00039556621853615914, 'samples': 3168960, 'steps': 16504, 'loss/train': 0.9424332976341248} +01/27/2022 11:22:48 - INFO - codeparrot_training - Step 16505: {'lr': 0.0003955529155569202, 'samples': 3169152, 'steps': 16505, 'loss/train': 1.0110992789268494} +01/27/2022 11:22:51 - INFO - codeparrot_training - Step 16506: {'lr': 0.0003955396119541788, 'samples': 3169344, 'steps': 16506, 'loss/train': 0.050900084897875786} +01/27/2022 11:22:55 - INFO - codeparrot_training - Step 16507: {'lr': 0.00039552630772799185, 'samples': 3169536, 'steps': 16507, 'loss/train': 0.9185109436511993} +01/27/2022 11:22:59 - INFO - codeparrot_training - Step 16508: {'lr': 0.0003955130028784165, 'samples': 3169728, 'steps': 16508, 'loss/train': 0.6492849290370941} +01/27/2022 11:23:02 - INFO - codeparrot_training - Step 16509: {'lr': 0.00039549969740550954, 'samples': 3169920, 'steps': 16509, 'loss/train': 0.6369474828243256} +01/27/2022 11:23:05 - INFO - codeparrot_training - Step 16510: {'lr': 0.00039548639130932816, 'samples': 3170112, 'steps': 16510, 'loss/train': 0.8093358278274536} +01/27/2022 11:23:09 - INFO - codeparrot_training - Step 16511: {'lr': 0.00039547308458992927, 'samples': 3170304, 'steps': 16511, 'loss/train': 1.0987606644630432} +01/27/2022 11:23:12 - INFO - codeparrot_training - Step 16512: {'lr': 0.00039545977724736984, 'samples': 3170496, 'steps': 16512, 'loss/train': 0.16172479838132858} +01/27/2022 11:23:15 - INFO - codeparrot_training - Step 16513: {'lr': 0.00039544646928170695, 'samples': 3170688, 'steps': 16513, 'loss/train': 0.6994959264993668} +01/27/2022 11:23:18 - INFO - codeparrot_training - Step 16514: {'lr': 0.0003954331606929976, 'samples': 3170880, 'steps': 16514, 'loss/train': 0.30243629962205887} +01/27/2022 11:23:21 - INFO - codeparrot_training - Step 16515: {'lr': 0.00039541985148129865, 'samples': 3171072, 'steps': 16515, 'loss/train': 0.844240665435791} +01/27/2022 11:23:24 - INFO - codeparrot_training - Step 16516: {'lr': 0.00039540654164666735, 'samples': 3171264, 'steps': 16516, 'loss/train': 0.8916550576686859} +01/27/2022 11:23:31 - INFO - codeparrot_training - Step 16517: {'lr': 0.00039539323118916055, 'samples': 3171456, 'steps': 16517, 'loss/train': 1.0751637518405914} +01/27/2022 11:23:34 - INFO - codeparrot_training - Step 16518: {'lr': 0.0003953799201088353, 'samples': 3171648, 'steps': 16518, 'loss/train': 1.0046809315681458} +01/27/2022 11:23:37 - INFO - codeparrot_training - Step 16519: {'lr': 0.00039536660840574866, 'samples': 3171840, 'steps': 16519, 'loss/train': 0.5885586440563202} +01/27/2022 11:23:40 - INFO - codeparrot_training - Step 16520: {'lr': 0.0003953532960799577, 'samples': 3172032, 'steps': 16520, 'loss/train': 1.9914727807044983} +01/27/2022 11:23:43 - INFO - codeparrot_training - Step 16521: {'lr': 0.00039533998313151926, 'samples': 3172224, 'steps': 16521, 'loss/train': 0.8456672430038452} +01/27/2022 11:23:46 - INFO - codeparrot_training - Step 16522: {'lr': 0.0003953266695604906, 'samples': 3172416, 'steps': 16522, 'loss/train': 1.4317070245742798} +01/27/2022 11:23:50 - INFO - codeparrot_training - Step 16523: {'lr': 0.0003953133553669285, 'samples': 3172608, 'steps': 16523, 'loss/train': 0.8191746175289154} +01/27/2022 11:23:53 - INFO - codeparrot_training - Step 16524: {'lr': 0.0003953000405508902, 'samples': 3172800, 'steps': 16524, 'loss/train': 1.1051898300647736} +01/27/2022 11:23:56 - INFO - codeparrot_training - Step 16525: {'lr': 0.00039528672511243256, 'samples': 3172992, 'steps': 16525, 'loss/train': 0.8391530513763428} +01/27/2022 11:24:00 - INFO - codeparrot_training - Step 16526: {'lr': 0.0003952734090516129, 'samples': 3173184, 'steps': 16526, 'loss/train': 0.9399526119232178} +01/27/2022 11:24:03 - INFO - codeparrot_training - Step 16527: {'lr': 0.000395260092368488, 'samples': 3173376, 'steps': 16527, 'loss/train': 0.9037066698074341} +01/27/2022 11:24:06 - INFO - codeparrot_training - Step 16528: {'lr': 0.000395246775063115, 'samples': 3173568, 'steps': 16528, 'loss/train': 0.5166259557008743} +01/27/2022 11:24:10 - INFO - codeparrot_training - Step 16529: {'lr': 0.0003952334571355509, 'samples': 3173760, 'steps': 16529, 'loss/train': 0.9063565135002136} +01/27/2022 11:24:13 - INFO - codeparrot_training - Step 16530: {'lr': 0.0003952201385858528, 'samples': 3173952, 'steps': 16530, 'loss/train': 1.3277464807033539} +01/27/2022 11:24:16 - INFO - codeparrot_training - Step 16531: {'lr': 0.00039520681941407777, 'samples': 3174144, 'steps': 16531, 'loss/train': 0.8441122770309448} +01/27/2022 11:24:19 - INFO - codeparrot_training - Step 16532: {'lr': 0.00039519349962028276, 'samples': 3174336, 'steps': 16532, 'loss/train': 0.8823607563972473} +01/27/2022 11:24:22 - INFO - codeparrot_training - Step 16533: {'lr': 0.000395180179204525, 'samples': 3174528, 'steps': 16533, 'loss/train': 0.9904004037380219} +01/27/2022 11:24:27 - INFO - codeparrot_training - Step 16534: {'lr': 0.0003951668581668614, 'samples': 3174720, 'steps': 16534, 'loss/train': 0.7316027283668518} +01/27/2022 11:24:30 - INFO - codeparrot_training - Step 16535: {'lr': 0.0003951535365073491, 'samples': 3174912, 'steps': 16535, 'loss/train': 0.8587291538715363} +01/27/2022 11:24:33 - INFO - codeparrot_training - Step 16536: {'lr': 0.00039514021422604515, 'samples': 3175104, 'steps': 16536, 'loss/train': 0.7947175204753876} +01/27/2022 11:24:36 - INFO - codeparrot_training - Step 16537: {'lr': 0.0003951268913230066, 'samples': 3175296, 'steps': 16537, 'loss/train': 0.48509278893470764} +01/27/2022 11:24:39 - INFO - codeparrot_training - Step 16538: {'lr': 0.0003951135677982904, 'samples': 3175488, 'steps': 16538, 'loss/train': 0.36743807047605515} +01/27/2022 11:24:43 - INFO - codeparrot_training - Step 16539: {'lr': 0.000395100243651954, 'samples': 3175680, 'steps': 16539, 'loss/train': 0.961525171995163} +01/27/2022 11:24:46 - INFO - codeparrot_training - Step 16540: {'lr': 0.00039508691888405403, 'samples': 3175872, 'steps': 16540, 'loss/train': 0.731463298201561} +01/27/2022 11:24:49 - INFO - codeparrot_training - Step 16541: {'lr': 0.0003950735934946478, 'samples': 3176064, 'steps': 16541, 'loss/train': 0.6246278285980225} +01/27/2022 11:24:52 - INFO - codeparrot_training - Step 16542: {'lr': 0.0003950602674837924, 'samples': 3176256, 'steps': 16542, 'loss/train': 1.0293222069740295} +01/27/2022 11:24:58 - INFO - codeparrot_training - Step 16543: {'lr': 0.0003950469408515449, 'samples': 3176448, 'steps': 16543, 'loss/train': 1.0070224106311798} +01/27/2022 11:25:01 - INFO - codeparrot_training - Step 16544: {'lr': 0.00039503361359796235, 'samples': 3176640, 'steps': 16544, 'loss/train': 0.6476696133613586} +01/27/2022 11:25:04 - INFO - codeparrot_training - Step 16545: {'lr': 0.00039502028572310186, 'samples': 3176832, 'steps': 16545, 'loss/train': 0.918239951133728} +01/27/2022 11:25:08 - INFO - codeparrot_training - Step 16546: {'lr': 0.0003950069572270205, 'samples': 3177024, 'steps': 16546, 'loss/train': 0.9084859192371368} +01/27/2022 11:25:11 - INFO - codeparrot_training - Step 16547: {'lr': 0.00039499362810977535, 'samples': 3177216, 'steps': 16547, 'loss/train': 0.5326113402843475} +01/27/2022 11:25:14 - INFO - codeparrot_training - Step 16548: {'lr': 0.00039498029837142356, 'samples': 3177408, 'steps': 16548, 'loss/train': 0.6391774266958237} +01/27/2022 11:25:17 - INFO - codeparrot_training - Step 16549: {'lr': 0.0003949669680120223, 'samples': 3177600, 'steps': 16549, 'loss/train': 0.9387180805206299} +01/27/2022 11:25:20 - INFO - codeparrot_training - Step 16550: {'lr': 0.00039495363703162843, 'samples': 3177792, 'steps': 16550, 'loss/train': 0.5353071838617325} +01/27/2022 11:25:23 - INFO - codeparrot_training - Step 16551: {'lr': 0.00039494030543029925, 'samples': 3177984, 'steps': 16551, 'loss/train': 0.7367417514324188} +01/27/2022 11:25:28 - INFO - codeparrot_training - Step 16552: {'lr': 0.0003949269732080919, 'samples': 3178176, 'steps': 16552, 'loss/train': 0.39576494693756104} +01/27/2022 11:25:31 - INFO - codeparrot_training - Step 16553: {'lr': 0.0003949136403650633, 'samples': 3178368, 'steps': 16553, 'loss/train': 0.7498573064804077} +01/27/2022 11:25:34 - INFO - codeparrot_training - Step 16554: {'lr': 0.0003949003069012708, 'samples': 3178560, 'steps': 16554, 'loss/train': 0.422603040933609} +01/27/2022 11:25:37 - INFO - codeparrot_training - Step 16555: {'lr': 0.0003948869728167713, 'samples': 3178752, 'steps': 16555, 'loss/train': 1.057502567768097} +01/27/2022 11:25:40 - INFO - codeparrot_training - Step 16556: {'lr': 0.0003948736381116221, 'samples': 3178944, 'steps': 16556, 'loss/train': 0.19354160875082016} +01/27/2022 11:25:44 - INFO - codeparrot_training - Step 16557: {'lr': 0.0003948603027858802, 'samples': 3179136, 'steps': 16557, 'loss/train': 0.2749708369374275} +01/27/2022 11:25:47 - INFO - codeparrot_training - Step 16558: {'lr': 0.00039484696683960276, 'samples': 3179328, 'steps': 16558, 'loss/train': 0.7165920585393906} +01/27/2022 11:25:50 - INFO - codeparrot_training - Step 16559: {'lr': 0.0003948336302728469, 'samples': 3179520, 'steps': 16559, 'loss/train': 0.99613156914711} +01/27/2022 11:25:56 - INFO - codeparrot_training - Step 16560: {'lr': 0.0003948202930856697, 'samples': 3179712, 'steps': 16560, 'loss/train': 0.8270643353462219} +01/27/2022 11:25:59 - INFO - codeparrot_training - Step 16561: {'lr': 0.0003948069552781285, 'samples': 3179904, 'steps': 16561, 'loss/train': 0.676588699221611} +01/27/2022 11:26:03 - INFO - codeparrot_training - Step 16562: {'lr': 0.00039479361685028016, 'samples': 3180096, 'steps': 16562, 'loss/train': 0.783758819103241} +01/27/2022 11:26:06 - INFO - codeparrot_training - Step 16563: {'lr': 0.00039478027780218193, 'samples': 3180288, 'steps': 16563, 'loss/train': 0.6258453726768494} +01/27/2022 11:26:09 - INFO - codeparrot_training - Step 16564: {'lr': 0.00039476693813389105, 'samples': 3180480, 'steps': 16564, 'loss/train': 0.8174875974655151} +01/27/2022 11:26:12 - INFO - codeparrot_training - Step 16565: {'lr': 0.0003947535978454645, 'samples': 3180672, 'steps': 16565, 'loss/train': 0.8849004507064819} +01/27/2022 11:26:15 - INFO - codeparrot_training - Step 16566: {'lr': 0.0003947402569369596, 'samples': 3180864, 'steps': 16566, 'loss/train': 1.084971696138382} +01/27/2022 11:26:18 - INFO - codeparrot_training - Step 16567: {'lr': 0.0003947269154084333, 'samples': 3181056, 'steps': 16567, 'loss/train': 1.10421484708786} +01/27/2022 11:26:21 - INFO - codeparrot_training - Step 16568: {'lr': 0.0003947135732599428, 'samples': 3181248, 'steps': 16568, 'loss/train': 1.1927977502346039} +01/27/2022 11:26:26 - INFO - codeparrot_training - Step 16569: {'lr': 0.00039470023049154544, 'samples': 3181440, 'steps': 16569, 'loss/train': 1.3085245192050934} +01/27/2022 11:26:29 - INFO - codeparrot_training - Step 16570: {'lr': 0.00039468688710329826, 'samples': 3181632, 'steps': 16570, 'loss/train': 0.7902727425098419} +01/27/2022 11:26:32 - INFO - codeparrot_training - Step 16571: {'lr': 0.0003946735430952583, 'samples': 3181824, 'steps': 16571, 'loss/train': 0.7922590970993042} +01/27/2022 11:26:36 - INFO - codeparrot_training - Step 16572: {'lr': 0.0003946601984674828, 'samples': 3182016, 'steps': 16572, 'loss/train': 0.5704865902662277} +01/27/2022 11:26:39 - INFO - codeparrot_training - Step 16573: {'lr': 0.00039464685322002904, 'samples': 3182208, 'steps': 16573, 'loss/train': 0.6855838000774384} +01/27/2022 11:26:42 - INFO - codeparrot_training - Step 16574: {'lr': 0.000394633507352954, 'samples': 3182400, 'steps': 16574, 'loss/train': 0.7145133018493652} +01/27/2022 11:26:45 - INFO - codeparrot_training - Step 16575: {'lr': 0.00039462016086631505, 'samples': 3182592, 'steps': 16575, 'loss/train': 0.972789466381073} +01/27/2022 11:26:48 - INFO - codeparrot_training - Step 16576: {'lr': 0.00039460681376016915, 'samples': 3182784, 'steps': 16576, 'loss/train': 0.35083360970020294} +01/27/2022 11:26:51 - INFO - codeparrot_training - Step 16577: {'lr': 0.0003945934660345736, 'samples': 3182976, 'steps': 16577, 'loss/train': 0.3271612375974655} +01/27/2022 11:26:56 - INFO - codeparrot_training - Step 16578: {'lr': 0.00039458011768958557, 'samples': 3183168, 'steps': 16578, 'loss/train': 0.8080206513404846} +01/27/2022 11:26:59 - INFO - codeparrot_training - Step 16579: {'lr': 0.00039456676872526227, 'samples': 3183360, 'steps': 16579, 'loss/train': 1.1806439459323883} +01/27/2022 11:27:02 - INFO - codeparrot_training - Step 16580: {'lr': 0.00039455341914166074, 'samples': 3183552, 'steps': 16580, 'loss/train': 0.8876824378967285} +01/27/2022 11:27:05 - INFO - codeparrot_training - Step 16581: {'lr': 0.0003945400689388384, 'samples': 3183744, 'steps': 16581, 'loss/train': 0.9121866524219513} +01/27/2022 11:27:08 - INFO - codeparrot_training - Step 16582: {'lr': 0.00039452671811685214, 'samples': 3183936, 'steps': 16582, 'loss/train': 1.10377836227417} +01/27/2022 11:27:11 - INFO - codeparrot_training - Step 16583: {'lr': 0.00039451336667575945, 'samples': 3184128, 'steps': 16583, 'loss/train': 0.6815927177667618} +01/27/2022 11:27:15 - INFO - codeparrot_training - Step 16584: {'lr': 0.0003945000146156173, 'samples': 3184320, 'steps': 16584, 'loss/train': 0.2676375210285187} +01/27/2022 11:27:18 - INFO - codeparrot_training - Step 16585: {'lr': 0.00039448666193648305, 'samples': 3184512, 'steps': 16585, 'loss/train': 0.5688890665769577} +01/27/2022 11:27:21 - INFO - codeparrot_training - Step 16586: {'lr': 0.0003944733086384137, 'samples': 3184704, 'steps': 16586, 'loss/train': 0.8929148018360138} +01/27/2022 11:27:25 - INFO - codeparrot_training - Step 16587: {'lr': 0.00039445995472146665, 'samples': 3184896, 'steps': 16587, 'loss/train': 0.7771497666835785} +01/27/2022 11:27:28 - INFO - codeparrot_training - Step 16588: {'lr': 0.000394446600185699, 'samples': 3185088, 'steps': 16588, 'loss/train': 1.4753319025039673} +01/27/2022 11:27:32 - INFO - codeparrot_training - Step 16589: {'lr': 0.000394433245031168, 'samples': 3185280, 'steps': 16589, 'loss/train': 0.16334032639861107} +01/27/2022 11:27:35 - INFO - codeparrot_training - Step 16590: {'lr': 0.0003944198892579309, 'samples': 3185472, 'steps': 16590, 'loss/train': 1.1409006714820862} +01/27/2022 11:27:38 - INFO - codeparrot_training - Step 16591: {'lr': 0.0003944065328660447, 'samples': 3185664, 'steps': 16591, 'loss/train': 0.2998780086636543} +01/27/2022 11:27:41 - INFO - codeparrot_training - Step 16592: {'lr': 0.0003943931758555669, 'samples': 3185856, 'steps': 16592, 'loss/train': 1.0056205093860626} +01/27/2022 11:27:44 - INFO - codeparrot_training - Step 16593: {'lr': 0.00039437981822655453, 'samples': 3186048, 'steps': 16593, 'loss/train': 1.02582648396492} +01/27/2022 11:27:47 - INFO - codeparrot_training - Step 16594: {'lr': 0.00039436645997906487, 'samples': 3186240, 'steps': 16594, 'loss/train': 0.6022307574748993} +01/27/2022 11:27:50 - INFO - codeparrot_training - Step 16595: {'lr': 0.00039435310111315513, 'samples': 3186432, 'steps': 16595, 'loss/train': 0.6373828500509262} +01/27/2022 11:27:55 - INFO - codeparrot_training - Step 16596: {'lr': 0.00039433974162888266, 'samples': 3186624, 'steps': 16596, 'loss/train': 1.0721024572849274} +01/27/2022 11:27:58 - INFO - codeparrot_training - Step 16597: {'lr': 0.0003943263815263044, 'samples': 3186816, 'steps': 16597, 'loss/train': 0.6327148228883743} +01/27/2022 11:28:01 - INFO - codeparrot_training - Step 16598: {'lr': 0.0003943130208054778, 'samples': 3187008, 'steps': 16598, 'loss/train': 0.9300316572189331} +01/27/2022 11:28:04 - INFO - codeparrot_training - Step 16599: {'lr': 0.0003942996594664601, 'samples': 3187200, 'steps': 16599, 'loss/train': 0.8318009376525879} +01/27/2022 11:28:08 - INFO - codeparrot_training - Step 16600: {'lr': 0.00039428629750930846, 'samples': 3187392, 'steps': 16600, 'loss/train': 0.79554882645607} +01/27/2022 11:28:11 - INFO - codeparrot_training - Step 16601: {'lr': 0.0003942729349340801, 'samples': 3187584, 'steps': 16601, 'loss/train': 0.7911353409290314} +01/27/2022 11:28:14 - INFO - codeparrot_training - Step 16602: {'lr': 0.00039425957174083224, 'samples': 3187776, 'steps': 16602, 'loss/train': 0.3604074865579605} +01/27/2022 11:28:17 - INFO - codeparrot_training - Step 16603: {'lr': 0.0003942462079296223, 'samples': 3187968, 'steps': 16603, 'loss/train': 0.7541225552558899} +01/27/2022 11:28:23 - INFO - codeparrot_training - Step 16604: {'lr': 0.00039423284350050735, 'samples': 3188160, 'steps': 16604, 'loss/train': 0.6731172949075699} +01/27/2022 11:28:26 - INFO - codeparrot_training - Step 16605: {'lr': 0.00039421947845354476, 'samples': 3188352, 'steps': 16605, 'loss/train': 0.5511913150548935} +01/27/2022 11:28:29 - INFO - codeparrot_training - Step 16606: {'lr': 0.0003942061127887916, 'samples': 3188544, 'steps': 16606, 'loss/train': 0.9432304501533508} +01/27/2022 11:28:33 - INFO - codeparrot_training - Step 16607: {'lr': 0.00039419274650630536, 'samples': 3188736, 'steps': 16607, 'loss/train': 1.0078726708889008} +01/27/2022 11:28:36 - INFO - codeparrot_training - Step 16608: {'lr': 0.00039417937960614316, 'samples': 3188928, 'steps': 16608, 'loss/train': 1.0350541770458221} +01/27/2022 11:28:39 - INFO - codeparrot_training - Step 16609: {'lr': 0.0003941660120883622, 'samples': 3189120, 'steps': 16609, 'loss/train': 0.5650066137313843} +01/27/2022 11:28:42 - INFO - codeparrot_training - Step 16610: {'lr': 0.0003941526439530199, 'samples': 3189312, 'steps': 16610, 'loss/train': 1.2736766338348389} +01/27/2022 11:28:45 - INFO - codeparrot_training - Step 16611: {'lr': 0.00039413927520017347, 'samples': 3189504, 'steps': 16611, 'loss/train': 0.9660032093524933} +01/27/2022 11:28:48 - INFO - codeparrot_training - Step 16612: {'lr': 0.00039412590582988007, 'samples': 3189696, 'steps': 16612, 'loss/train': 0.38302333652973175} +01/27/2022 11:28:53 - INFO - codeparrot_training - Step 16613: {'lr': 0.00039411253584219707, 'samples': 3189888, 'steps': 16613, 'loss/train': 0.6929687261581421} +01/27/2022 11:28:56 - INFO - codeparrot_training - Step 16614: {'lr': 0.0003940991652371818, 'samples': 3190080, 'steps': 16614, 'loss/train': 0.8751102089881897} +01/27/2022 11:28:59 - INFO - codeparrot_training - Step 16615: {'lr': 0.0003940857940148914, 'samples': 3190272, 'steps': 16615, 'loss/train': 1.2635843753814697} +01/27/2022 11:29:02 - INFO - codeparrot_training - Step 16616: {'lr': 0.00039407242217538317, 'samples': 3190464, 'steps': 16616, 'loss/train': 0.8175921142101288} +01/27/2022 11:29:05 - INFO - codeparrot_training - Step 16617: {'lr': 0.00039405904971871454, 'samples': 3190656, 'steps': 16617, 'loss/train': 0.6919123381376266} +01/27/2022 11:29:08 - INFO - codeparrot_training - Step 16618: {'lr': 0.00039404567664494264, 'samples': 3190848, 'steps': 16618, 'loss/train': 0.508542001247406} +01/27/2022 11:29:12 - INFO - codeparrot_training - Step 16619: {'lr': 0.0003940323029541248, 'samples': 3191040, 'steps': 16619, 'loss/train': 0.5910428613424301} +01/27/2022 11:29:15 - INFO - codeparrot_training - Step 16620: {'lr': 0.00039401892864631826, 'samples': 3191232, 'steps': 16620, 'loss/train': 0.9494670331478119} +01/27/2022 11:29:18 - INFO - codeparrot_training - Step 16621: {'lr': 0.0003940055537215804, 'samples': 3191424, 'steps': 16621, 'loss/train': 0.7614655494689941} +01/27/2022 11:29:24 - INFO - codeparrot_training - Step 16622: {'lr': 0.0003939921781799685, 'samples': 3191616, 'steps': 16622, 'loss/train': 0.7803365886211395} +01/27/2022 11:29:27 - INFO - codeparrot_training - Step 16623: {'lr': 0.0003939788020215398, 'samples': 3191808, 'steps': 16623, 'loss/train': 0.5933124125003815} +01/27/2022 11:29:31 - INFO - codeparrot_training - Step 16624: {'lr': 0.0003939654252463517, 'samples': 3192000, 'steps': 16624, 'loss/train': 0.6415375471115112} +01/27/2022 11:29:34 - INFO - codeparrot_training - Step 16625: {'lr': 0.00039395204785446137, 'samples': 3192192, 'steps': 16625, 'loss/train': 0.9080023169517517} +01/27/2022 11:29:37 - INFO - codeparrot_training - Step 16626: {'lr': 0.00039393866984592616, 'samples': 3192384, 'steps': 16626, 'loss/train': 0.2179461568593979} +01/27/2022 11:29:40 - INFO - codeparrot_training - Step 16627: {'lr': 0.00039392529122080343, 'samples': 3192576, 'steps': 16627, 'loss/train': 0.790966272354126} +01/27/2022 11:29:43 - INFO - codeparrot_training - Step 16628: {'lr': 0.0003939119119791504, 'samples': 3192768, 'steps': 16628, 'loss/train': 0.6232640594244003} +01/27/2022 11:29:46 - INFO - codeparrot_training - Step 16629: {'lr': 0.0003938985321210245, 'samples': 3192960, 'steps': 16629, 'loss/train': 0.29216165095567703} +01/27/2022 11:29:49 - INFO - codeparrot_training - Step 16630: {'lr': 0.00039388515164648293, 'samples': 3193152, 'steps': 16630, 'loss/train': 0.6077028661966324} +01/27/2022 11:29:54 - INFO - codeparrot_training - Step 16631: {'lr': 0.0003938717705555831, 'samples': 3193344, 'steps': 16631, 'loss/train': 0.9190243184566498} +01/27/2022 11:29:57 - INFO - codeparrot_training - Step 16632: {'lr': 0.0003938583888483823, 'samples': 3193536, 'steps': 16632, 'loss/train': 0.7955794036388397} +01/27/2022 11:30:00 - INFO - codeparrot_training - Step 16633: {'lr': 0.0003938450065249378, 'samples': 3193728, 'steps': 16633, 'loss/train': 0.5880499631166458} +01/27/2022 11:30:03 - INFO - codeparrot_training - Step 16634: {'lr': 0.00039383162358530696, 'samples': 3193920, 'steps': 16634, 'loss/train': 0.5856458097696304} +01/27/2022 11:30:07 - INFO - codeparrot_training - Step 16635: {'lr': 0.0003938182400295471, 'samples': 3194112, 'steps': 16635, 'loss/train': 0.9202377498149872} +01/27/2022 11:30:10 - INFO - codeparrot_training - Step 16636: {'lr': 0.00039380485585771563, 'samples': 3194304, 'steps': 16636, 'loss/train': 0.4655890166759491} +01/27/2022 11:30:13 - INFO - codeparrot_training - Step 16637: {'lr': 0.00039379147106986985, 'samples': 3194496, 'steps': 16637, 'loss/train': 0.5169586837291718} +01/27/2022 11:30:16 - INFO - codeparrot_training - Step 16638: {'lr': 0.00039377808566606697, 'samples': 3194688, 'steps': 16638, 'loss/train': 1.0658023059368134} +01/27/2022 11:30:19 - INFO - codeparrot_training - Step 16639: {'lr': 0.00039376469964636445, 'samples': 3194880, 'steps': 16639, 'loss/train': 1.6631666421890259} +01/27/2022 11:30:24 - INFO - codeparrot_training - Step 16640: {'lr': 0.0003937513130108197, 'samples': 3195072, 'steps': 16640, 'loss/train': 0.7431111931800842} +01/27/2022 11:30:27 - INFO - codeparrot_training - Step 16641: {'lr': 0.00039373792575948986, 'samples': 3195264, 'steps': 16641, 'loss/train': 0.8739589154720306} +01/27/2022 11:30:30 - INFO - codeparrot_training - Step 16642: {'lr': 0.00039372453789243245, 'samples': 3195456, 'steps': 16642, 'loss/train': 1.3651200234889984} +01/27/2022 11:30:33 - INFO - codeparrot_training - Step 16643: {'lr': 0.0003937111494097047, 'samples': 3195648, 'steps': 16643, 'loss/train': 0.9363327026367188} +01/27/2022 11:30:36 - INFO - codeparrot_training - Step 16644: {'lr': 0.0003936977603113641, 'samples': 3195840, 'steps': 16644, 'loss/train': 0.73640276491642} +01/27/2022 11:30:39 - INFO - codeparrot_training - Step 16645: {'lr': 0.00039368437059746785, 'samples': 3196032, 'steps': 16645, 'loss/train': 0.8628147840499878} +01/27/2022 11:30:42 - INFO - codeparrot_training - Step 16646: {'lr': 0.0003936709802680734, 'samples': 3196224, 'steps': 16646, 'loss/train': 0.8485980033874512} +01/27/2022 11:30:46 - INFO - codeparrot_training - Step 16647: {'lr': 0.0003936575893232381, 'samples': 3196416, 'steps': 16647, 'loss/train': 0.6648914963006973} +01/27/2022 11:30:52 - INFO - codeparrot_training - Step 16648: {'lr': 0.0003936441977630193, 'samples': 3196608, 'steps': 16648, 'loss/train': 0.5970194488763809} +01/27/2022 11:30:56 - INFO - codeparrot_training - Step 16649: {'lr': 0.0003936308055874744, 'samples': 3196800, 'steps': 16649, 'loss/train': 0.8995825946331024} +01/27/2022 11:30:59 - INFO - codeparrot_training - Step 16650: {'lr': 0.00039361741279666065, 'samples': 3196992, 'steps': 16650, 'loss/train': 1.0453486740589142} +01/27/2022 11:31:02 - INFO - codeparrot_training - Step 16651: {'lr': 0.0003936040193906356, 'samples': 3197184, 'steps': 16651, 'loss/train': 0.5045066624879837} +01/27/2022 11:31:05 - INFO - codeparrot_training - Step 16652: {'lr': 0.00039359062536945645, 'samples': 3197376, 'steps': 16652, 'loss/train': 0.6516699492931366} +01/27/2022 11:31:08 - INFO - codeparrot_training - Step 16653: {'lr': 0.00039357723073318076, 'samples': 3197568, 'steps': 16653, 'loss/train': 0.8535715341567993} +01/27/2022 11:31:11 - INFO - codeparrot_training - Step 16654: {'lr': 0.0003935638354818657, 'samples': 3197760, 'steps': 16654, 'loss/train': 1.2044422030448914} +01/27/2022 11:31:14 - INFO - codeparrot_training - Step 16655: {'lr': 0.0003935504396155688, 'samples': 3197952, 'steps': 16655, 'loss/train': 0.6183807402849197} +01/27/2022 11:31:18 - INFO - codeparrot_training - Step 16656: {'lr': 0.00039353704313434745, 'samples': 3198144, 'steps': 16656, 'loss/train': 0.3731156140565872} +01/27/2022 11:31:22 - INFO - codeparrot_training - Step 16657: {'lr': 0.000393523646038259, 'samples': 3198336, 'steps': 16657, 'loss/train': 0.6135188341140747} +01/27/2022 11:31:25 - INFO - codeparrot_training - Step 16658: {'lr': 0.0003935102483273607, 'samples': 3198528, 'steps': 16658, 'loss/train': 0.8142899572849274} +01/27/2022 11:31:29 - INFO - codeparrot_training - Step 16659: {'lr': 0.0003934968500017101, 'samples': 3198720, 'steps': 16659, 'loss/train': 0.7979467213153839} +01/27/2022 11:31:32 - INFO - codeparrot_training - Step 16660: {'lr': 0.0003934834510613646, 'samples': 3198912, 'steps': 16660, 'loss/train': 0.4622219502925873} +01/27/2022 11:31:35 - INFO - codeparrot_training - Step 16661: {'lr': 0.00039347005150638156, 'samples': 3199104, 'steps': 16661, 'loss/train': 1.2901637256145477} +01/27/2022 11:31:38 - INFO - codeparrot_training - Step 16662: {'lr': 0.0003934566513368183, 'samples': 3199296, 'steps': 16662, 'loss/train': 0.14947481453418732} +01/27/2022 11:31:41 - INFO - codeparrot_training - Step 16663: {'lr': 0.00039344325055273236, 'samples': 3199488, 'steps': 16663, 'loss/train': 0.6703546196222305} +01/27/2022 11:31:44 - INFO - codeparrot_training - Step 16664: {'lr': 0.0003934298491541811, 'samples': 3199680, 'steps': 16664, 'loss/train': 0.8610442578792572} +01/27/2022 11:31:47 - INFO - codeparrot_training - Step 16665: {'lr': 0.00039341644714122195, 'samples': 3199872, 'steps': 16665, 'loss/train': 0.700242206454277} +01/27/2022 11:31:54 - INFO - codeparrot_training - Step 16666: {'lr': 0.00039340304451391216, 'samples': 3200064, 'steps': 16666, 'loss/train': 1.6794821619987488} +01/27/2022 11:31:57 - INFO - codeparrot_training - Step 16667: {'lr': 0.00039338964127230935, 'samples': 3200256, 'steps': 16667, 'loss/train': 0.14145388081669807} +01/27/2022 11:32:00 - INFO - codeparrot_training - Step 16668: {'lr': 0.00039337623741647084, 'samples': 3200448, 'steps': 16668, 'loss/train': 0.5690004676580429} +01/27/2022 11:32:03 - INFO - codeparrot_training - Step 16669: {'lr': 0.000393362832946454, 'samples': 3200640, 'steps': 16669, 'loss/train': 5.116445302963257} +01/27/2022 11:32:07 - INFO - codeparrot_training - Step 16670: {'lr': 0.0003933494278623164, 'samples': 3200832, 'steps': 16670, 'loss/train': 0.8027774691581726} +01/27/2022 11:32:10 - INFO - codeparrot_training - Step 16671: {'lr': 0.0003933360221641153, 'samples': 3201024, 'steps': 16671, 'loss/train': 0.8326976001262665} +01/27/2022 11:32:13 - INFO - codeparrot_training - Step 16672: {'lr': 0.0003933226158519082, 'samples': 3201216, 'steps': 16672, 'loss/train': 0.7170132994651794} +01/27/2022 11:32:16 - INFO - codeparrot_training - Step 16673: {'lr': 0.0003933092089257525, 'samples': 3201408, 'steps': 16673, 'loss/train': 1.3849074840545654} +01/27/2022 11:32:19 - INFO - codeparrot_training - Step 16674: {'lr': 0.0003932958013857057, 'samples': 3201600, 'steps': 16674, 'loss/train': 1.15932235121727} +01/27/2022 11:32:23 - INFO - codeparrot_training - Step 16675: {'lr': 0.0003932823932318252, 'samples': 3201792, 'steps': 16675, 'loss/train': 0.7461849600076675} +01/27/2022 11:32:27 - INFO - codeparrot_training - Step 16676: {'lr': 0.0003932689844641684, 'samples': 3201984, 'steps': 16676, 'loss/train': 1.2937967777252197} +01/27/2022 11:32:30 - INFO - codeparrot_training - Step 16677: {'lr': 0.00039325557508279276, 'samples': 3202176, 'steps': 16677, 'loss/train': 1.228144884109497} +01/27/2022 11:32:33 - INFO - codeparrot_training - Step 16678: {'lr': 0.00039324216508775567, 'samples': 3202368, 'steps': 16678, 'loss/train': 1.1171397864818573} +01/27/2022 11:32:36 - INFO - codeparrot_training - Step 16679: {'lr': 0.0003932287544791148, 'samples': 3202560, 'steps': 16679, 'loss/train': 0.6430477201938629} +01/27/2022 11:32:39 - INFO - codeparrot_training - Step 16680: {'lr': 0.00039321534325692726, 'samples': 3202752, 'steps': 16680, 'loss/train': 0.35831549763679504} +01/27/2022 11:32:42 - INFO - codeparrot_training - Step 16681: {'lr': 0.0003932019314212507, 'samples': 3202944, 'steps': 16681, 'loss/train': 0.5996710360050201} +01/27/2022 11:32:45 - INFO - codeparrot_training - Step 16682: {'lr': 0.0003931885189721426, 'samples': 3203136, 'steps': 16682, 'loss/train': 0.999709278345108} +01/27/2022 11:32:49 - INFO - codeparrot_training - Step 16683: {'lr': 0.00039317510590966033, 'samples': 3203328, 'steps': 16683, 'loss/train': 1.5889348983764648} +01/27/2022 11:32:53 - INFO - codeparrot_training - Step 16684: {'lr': 0.0003931616922338613, 'samples': 3203520, 'steps': 16684, 'loss/train': 0.8786716461181641} +01/27/2022 11:32:56 - INFO - codeparrot_training - Step 16685: {'lr': 0.00039314827794480314, 'samples': 3203712, 'steps': 16685, 'loss/train': 1.4999263286590576} +01/27/2022 11:32:59 - INFO - codeparrot_training - Step 16686: {'lr': 0.00039313486304254315, 'samples': 3203904, 'steps': 16686, 'loss/train': 0.8605344593524933} +01/27/2022 11:33:03 - INFO - codeparrot_training - Step 16687: {'lr': 0.00039312144752713885, 'samples': 3204096, 'steps': 16687, 'loss/train': 0.9058350026607513} +01/27/2022 11:33:06 - INFO - codeparrot_training - Step 16688: {'lr': 0.00039310803139864777, 'samples': 3204288, 'steps': 16688, 'loss/train': 1.2467730939388275} +01/27/2022 11:33:09 - INFO - codeparrot_training - Step 16689: {'lr': 0.00039309461465712725, 'samples': 3204480, 'steps': 16689, 'loss/train': 0.6246209442615509} +01/27/2022 11:33:12 - INFO - codeparrot_training - Step 16690: {'lr': 0.00039308119730263494, 'samples': 3204672, 'steps': 16690, 'loss/train': 0.9108148813247681} +01/27/2022 11:33:15 - INFO - codeparrot_training - Step 16691: {'lr': 0.00039306777933522806, 'samples': 3204864, 'steps': 16691, 'loss/train': 0.9141234755516052} +01/27/2022 11:33:18 - INFO - codeparrot_training - Step 16692: {'lr': 0.00039305436075496436, 'samples': 3205056, 'steps': 16692, 'loss/train': 0.856312483549118} +01/27/2022 11:33:23 - INFO - codeparrot_training - Step 16693: {'lr': 0.0003930409415619012, 'samples': 3205248, 'steps': 16693, 'loss/train': 0.6361477375030518} +01/27/2022 11:33:26 - INFO - codeparrot_training - Step 16694: {'lr': 0.000393027521756096, 'samples': 3205440, 'steps': 16694, 'loss/train': 0.8870329856872559} +01/27/2022 11:33:29 - INFO - codeparrot_training - Step 16695: {'lr': 0.0003930141013376064, 'samples': 3205632, 'steps': 16695, 'loss/train': 0.9518773555755615} +01/27/2022 11:33:32 - INFO - codeparrot_training - Step 16696: {'lr': 0.00039300068030648976, 'samples': 3205824, 'steps': 16696, 'loss/train': 0.4598803371191025} +01/27/2022 11:33:35 - INFO - codeparrot_training - Step 16697: {'lr': 0.0003929872586628036, 'samples': 3206016, 'steps': 16697, 'loss/train': 0.6698092371225357} +01/27/2022 11:33:38 - INFO - codeparrot_training - Step 16698: {'lr': 0.00039297383640660545, 'samples': 3206208, 'steps': 16698, 'loss/train': 0.8733636438846588} +01/27/2022 11:33:42 - INFO - codeparrot_training - Step 16699: {'lr': 0.0003929604135379528, 'samples': 3206400, 'steps': 16699, 'loss/train': 0.6255200654268265} +01/27/2022 11:33:45 - INFO - codeparrot_training - Step 16700: {'lr': 0.000392946990056903, 'samples': 3206592, 'steps': 16700, 'loss/train': 0.5678131878376007} +01/27/2022 11:33:51 - INFO - codeparrot_training - Step 16701: {'lr': 0.0003929335659635139, 'samples': 3206784, 'steps': 16701, 'loss/train': 0.9285567104816437} +01/27/2022 11:33:54 - INFO - codeparrot_training - Step 16702: {'lr': 0.00039292014125784266, 'samples': 3206976, 'steps': 16702, 'loss/train': 0.9886830747127533} +01/27/2022 11:33:57 - INFO - codeparrot_training - Step 16703: {'lr': 0.00039290671593994697, 'samples': 3207168, 'steps': 16703, 'loss/train': 0.555996760725975} +01/27/2022 11:34:01 - INFO - codeparrot_training - Step 16704: {'lr': 0.0003928932900098842, 'samples': 3207360, 'steps': 16704, 'loss/train': 0.04711763933300972} +01/27/2022 11:34:04 - INFO - codeparrot_training - Step 16705: {'lr': 0.00039287986346771205, 'samples': 3207552, 'steps': 16705, 'loss/train': 1.9233128428459167} +01/27/2022 11:34:07 - INFO - codeparrot_training - Step 16706: {'lr': 0.0003928664363134879, 'samples': 3207744, 'steps': 16706, 'loss/train': 0.8812556862831116} +01/27/2022 11:34:10 - INFO - codeparrot_training - Step 16707: {'lr': 0.00039285300854726926, 'samples': 3207936, 'steps': 16707, 'loss/train': 0.876979798078537} +01/27/2022 11:34:13 - INFO - codeparrot_training - Step 16708: {'lr': 0.00039283958016911373, 'samples': 3208128, 'steps': 16708, 'loss/train': 0.8022417426109314} +01/27/2022 11:34:16 - INFO - codeparrot_training - Step 16709: {'lr': 0.00039282615117907884, 'samples': 3208320, 'steps': 16709, 'loss/train': 0.7989880442619324} +01/27/2022 11:34:21 - INFO - codeparrot_training - Step 16710: {'lr': 0.00039281272157722205, 'samples': 3208512, 'steps': 16710, 'loss/train': 0.514949306845665} +01/27/2022 11:34:24 - INFO - codeparrot_training - Step 16711: {'lr': 0.0003927992913636008, 'samples': 3208704, 'steps': 16711, 'loss/train': 0.10661918297410011} +01/27/2022 11:34:27 - INFO - codeparrot_training - Step 16712: {'lr': 0.0003927858605382728, 'samples': 3208896, 'steps': 16712, 'loss/train': 1.2503102123737335} +01/27/2022 11:34:30 - INFO - codeparrot_training - Step 16713: {'lr': 0.0003927724291012955, 'samples': 3209088, 'steps': 16713, 'loss/train': 0.765852302312851} +01/27/2022 11:34:33 - INFO - codeparrot_training - Step 16714: {'lr': 0.00039275899705272656, 'samples': 3209280, 'steps': 16714, 'loss/train': 0.6418425589799881} +01/27/2022 11:34:37 - INFO - codeparrot_training - Step 16715: {'lr': 0.00039274556439262325, 'samples': 3209472, 'steps': 16715, 'loss/train': 0.7408009618520737} +01/27/2022 11:34:40 - INFO - codeparrot_training - Step 16716: {'lr': 0.0003927321311210434, 'samples': 3209664, 'steps': 16716, 'loss/train': 0.6624245792627335} +01/27/2022 11:34:43 - INFO - codeparrot_training - Step 16717: {'lr': 0.00039271869723804434, 'samples': 3209856, 'steps': 16717, 'loss/train': 0.47735631465911865} +01/27/2022 11:34:46 - INFO - codeparrot_training - Step 16718: {'lr': 0.0003927052627436837, 'samples': 3210048, 'steps': 16718, 'loss/train': 0.68913334608078} +01/27/2022 11:34:50 - INFO - codeparrot_training - Step 16719: {'lr': 0.000392691827638019, 'samples': 3210240, 'steps': 16719, 'loss/train': 0.9620359539985657} +01/27/2022 11:34:54 - INFO - codeparrot_training - Step 16720: {'lr': 0.000392678391921108, 'samples': 3210432, 'steps': 16720, 'loss/train': 0.8534120321273804} +01/27/2022 11:34:57 - INFO - codeparrot_training - Step 16721: {'lr': 0.00039266495559300786, 'samples': 3210624, 'steps': 16721, 'loss/train': 0.7651798725128174} +01/27/2022 11:35:00 - INFO - codeparrot_training - Step 16722: {'lr': 0.00039265151865377644, 'samples': 3210816, 'steps': 16722, 'loss/train': 0.8505843579769135} +01/27/2022 11:35:03 - INFO - codeparrot_training - Step 16723: {'lr': 0.0003926380811034712, 'samples': 3211008, 'steps': 16723, 'loss/train': 0.688309907913208} +01/27/2022 11:35:06 - INFO - codeparrot_training - Step 16724: {'lr': 0.0003926246429421497, 'samples': 3211200, 'steps': 16724, 'loss/train': 0.560474082827568} +01/27/2022 11:35:09 - INFO - codeparrot_training - Step 16725: {'lr': 0.0003926112041698696, 'samples': 3211392, 'steps': 16725, 'loss/train': 0.8942853212356567} +01/27/2022 11:35:12 - INFO - codeparrot_training - Step 16726: {'lr': 0.0003925977647866883, 'samples': 3211584, 'steps': 16726, 'loss/train': 0.6951418071985245} +01/27/2022 11:35:16 - INFO - codeparrot_training - Step 16727: {'lr': 0.0003925843247926635, 'samples': 3211776, 'steps': 16727, 'loss/train': 0.31534988433122635} +01/27/2022 11:35:22 - INFO - codeparrot_training - Step 16728: {'lr': 0.00039257088418785267, 'samples': 3211968, 'steps': 16728, 'loss/train': 1.0744222104549408} +01/27/2022 11:35:25 - INFO - codeparrot_training - Step 16729: {'lr': 0.00039255744297231354, 'samples': 3212160, 'steps': 16729, 'loss/train': 0.7197470963001251} +01/27/2022 11:35:28 - INFO - codeparrot_training - Step 16730: {'lr': 0.0003925440011461035, 'samples': 3212352, 'steps': 16730, 'loss/train': 0.9369017779827118} +01/27/2022 11:35:31 - INFO - codeparrot_training - Step 16731: {'lr': 0.0003925305587092802, 'samples': 3212544, 'steps': 16731, 'loss/train': 1.1171679496765137} +01/27/2022 11:35:34 - INFO - codeparrot_training - Step 16732: {'lr': 0.00039251711566190133, 'samples': 3212736, 'steps': 16732, 'loss/train': 0.5926946997642517} +01/27/2022 11:35:38 - INFO - codeparrot_training - Step 16733: {'lr': 0.0003925036720040244, 'samples': 3212928, 'steps': 16733, 'loss/train': 0.6009356081485748} +01/27/2022 11:35:41 - INFO - codeparrot_training - Step 16734: {'lr': 0.000392490227735707, 'samples': 3213120, 'steps': 16734, 'loss/train': 1.1935933828353882} +01/27/2022 11:35:44 - INFO - codeparrot_training - Step 16735: {'lr': 0.0003924767828570066, 'samples': 3213312, 'steps': 16735, 'loss/train': 0.8786973059177399} +01/27/2022 11:35:47 - INFO - codeparrot_training - Step 16736: {'lr': 0.00039246333736798095, 'samples': 3213504, 'steps': 16736, 'loss/train': 0.8101111650466919} +01/27/2022 11:35:51 - INFO - codeparrot_training - Step 16737: {'lr': 0.00039244989126868755, 'samples': 3213696, 'steps': 16737, 'loss/train': 0.9912816882133484} +01/27/2022 11:35:55 - INFO - codeparrot_training - Step 16738: {'lr': 0.0003924364445591842, 'samples': 3213888, 'steps': 16738, 'loss/train': 1.0460240542888641} +01/27/2022 11:35:58 - INFO - codeparrot_training - Step 16739: {'lr': 0.0003924229972395282, 'samples': 3214080, 'steps': 16739, 'loss/train': 1.021691232919693} +01/27/2022 11:36:01 - INFO - codeparrot_training - Step 16740: {'lr': 0.00039240954930977744, 'samples': 3214272, 'steps': 16740, 'loss/train': 0.6755223423242569} +01/27/2022 11:36:04 - INFO - codeparrot_training - Step 16741: {'lr': 0.0003923961007699893, 'samples': 3214464, 'steps': 16741, 'loss/train': 0.8301410973072052} +01/27/2022 11:36:07 - INFO - codeparrot_training - Step 16742: {'lr': 0.00039238265162022147, 'samples': 3214656, 'steps': 16742, 'loss/train': 1.0666493475437164} +01/27/2022 11:36:10 - INFO - codeparrot_training - Step 16743: {'lr': 0.0003923692018605316, 'samples': 3214848, 'steps': 16743, 'loss/train': 1.1022498607635498} +01/27/2022 11:36:13 - INFO - codeparrot_training - Step 16744: {'lr': 0.0003923557514909773, 'samples': 3215040, 'steps': 16744, 'loss/train': 1.1632480323314667} +01/27/2022 11:36:20 - INFO - codeparrot_training - Step 16745: {'lr': 0.00039234230051161614, 'samples': 3215232, 'steps': 16745, 'loss/train': 1.0130780339241028} +01/27/2022 11:36:23 - INFO - codeparrot_training - Step 16746: {'lr': 0.00039232884892250575, 'samples': 3215424, 'steps': 16746, 'loss/train': 0.9194342494010925} +01/27/2022 11:36:26 - INFO - codeparrot_training - Step 16747: {'lr': 0.00039231539672370376, 'samples': 3215616, 'steps': 16747, 'loss/train': 0.7772029638290405} +01/27/2022 11:36:29 - INFO - codeparrot_training - Step 16748: {'lr': 0.00039230194391526784, 'samples': 3215808, 'steps': 16748, 'loss/train': 1.2626834213733673} +01/27/2022 11:36:32 - INFO - codeparrot_training - Step 16749: {'lr': 0.0003922884904972556, 'samples': 3216000, 'steps': 16749, 'loss/train': 0.8118246495723724} +01/27/2022 11:36:36 - INFO - codeparrot_training - Step 16750: {'lr': 0.0003922750364697246, 'samples': 3216192, 'steps': 16750, 'loss/train': 0.8513975143432617} +01/27/2022 11:36:39 - INFO - codeparrot_training - Step 16751: {'lr': 0.0003922615818327325, 'samples': 3216384, 'steps': 16751, 'loss/train': 0.9015172719955444} +01/27/2022 11:36:42 - INFO - codeparrot_training - Step 16752: {'lr': 0.000392248126586337, 'samples': 3216576, 'steps': 16752, 'loss/train': 0.9106683433055878} +01/27/2022 11:36:45 - INFO - codeparrot_training - Step 16753: {'lr': 0.0003922346707305957, 'samples': 3216768, 'steps': 16753, 'loss/train': 0.7612093985080719} +01/27/2022 11:36:49 - INFO - codeparrot_training - Step 16754: {'lr': 0.00039222121426556617, 'samples': 3216960, 'steps': 16754, 'loss/train': 0.3005252256989479} +01/27/2022 11:36:53 - INFO - codeparrot_training - Step 16755: {'lr': 0.0003922077571913062, 'samples': 3217152, 'steps': 16755, 'loss/train': 0.6153963804244995} +01/27/2022 11:36:56 - INFO - codeparrot_training - Step 16756: {'lr': 0.00039219429950787326, 'samples': 3217344, 'steps': 16756, 'loss/train': 0.860878586769104} +01/27/2022 11:36:59 - INFO - codeparrot_training - Step 16757: {'lr': 0.0003921808412153252, 'samples': 3217536, 'steps': 16757, 'loss/train': 0.6527019292116165} +01/27/2022 11:37:02 - INFO - codeparrot_training - Step 16758: {'lr': 0.0003921673823137195, 'samples': 3217728, 'steps': 16758, 'loss/train': 0.58095283806324} +01/27/2022 11:37:05 - INFO - codeparrot_training - Step 16759: {'lr': 0.00039215392280311383, 'samples': 3217920, 'steps': 16759, 'loss/train': 0.9643063545227051} +01/27/2022 11:37:08 - INFO - codeparrot_training - Step 16760: {'lr': 0.000392140462683566, 'samples': 3218112, 'steps': 16760, 'loss/train': 0.8284721374511719} +01/27/2022 11:37:11 - INFO - codeparrot_training - Step 16761: {'lr': 0.0003921270019551335, 'samples': 3218304, 'steps': 16761, 'loss/train': 5.530507922172546} +01/27/2022 11:37:15 - INFO - codeparrot_training - Step 16762: {'lr': 0.00039211354061787407, 'samples': 3218496, 'steps': 16762, 'loss/train': 0.6056576818227768} +01/27/2022 11:37:20 - INFO - codeparrot_training - Step 16763: {'lr': 0.0003921000786718454, 'samples': 3218688, 'steps': 16763, 'loss/train': 0.5534693598747253} +01/27/2022 11:37:23 - INFO - codeparrot_training - Step 16764: {'lr': 0.0003920866161171051, 'samples': 3218880, 'steps': 16764, 'loss/train': 0.796954482793808} +01/27/2022 11:37:26 - INFO - codeparrot_training - Step 16765: {'lr': 0.0003920731529537108, 'samples': 3219072, 'steps': 16765, 'loss/train': 0.9207988679409027} +01/27/2022 11:37:29 - INFO - codeparrot_training - Step 16766: {'lr': 0.00039205968918172026, 'samples': 3219264, 'steps': 16766, 'loss/train': 1.3204565048217773} +01/27/2022 11:37:32 - INFO - codeparrot_training - Step 16767: {'lr': 0.00039204622480119107, 'samples': 3219456, 'steps': 16767, 'loss/train': 1.726703703403473} +01/27/2022 11:37:35 - INFO - codeparrot_training - Step 16768: {'lr': 0.000392032759812181, 'samples': 3219648, 'steps': 16768, 'loss/train': 0.7946044206619263} +01/27/2022 11:37:38 - INFO - codeparrot_training - Step 16769: {'lr': 0.0003920192942147477, 'samples': 3219840, 'steps': 16769, 'loss/train': 0.9263950288295746} +01/27/2022 11:37:42 - INFO - codeparrot_training - Step 16770: {'lr': 0.00039200582800894885, 'samples': 3220032, 'steps': 16770, 'loss/train': 1.0182404816150665} +01/27/2022 11:37:45 - INFO - codeparrot_training - Step 16771: {'lr': 0.00039199236119484207, 'samples': 3220224, 'steps': 16771, 'loss/train': 0.31293871253728867} +01/27/2022 11:37:50 - INFO - codeparrot_training - Step 16772: {'lr': 0.0003919788937724852, 'samples': 3220416, 'steps': 16772, 'loss/train': 0.9355471730232239} +01/27/2022 11:37:54 - INFO - codeparrot_training - Step 16773: {'lr': 0.0003919654257419357, 'samples': 3220608, 'steps': 16773, 'loss/train': 0.6336187720298767} +01/27/2022 11:37:57 - INFO - codeparrot_training - Step 16774: {'lr': 0.0003919519571032515, 'samples': 3220800, 'steps': 16774, 'loss/train': 1.2122949957847595} +01/27/2022 11:38:00 - INFO - codeparrot_training - Step 16775: {'lr': 0.00039193848785649016, 'samples': 3220992, 'steps': 16775, 'loss/train': 0.5471028238534927} +01/27/2022 11:38:03 - INFO - codeparrot_training - Step 16776: {'lr': 0.0003919250180017094, 'samples': 3221184, 'steps': 16776, 'loss/train': 0.05610623583197594} +01/27/2022 11:38:06 - INFO - codeparrot_training - Step 16777: {'lr': 0.00039191154753896696, 'samples': 3221376, 'steps': 16777, 'loss/train': 0.3731239289045334} +01/27/2022 11:38:09 - INFO - codeparrot_training - Step 16778: {'lr': 0.00039189807646832045, 'samples': 3221568, 'steps': 16778, 'loss/train': 0.3794650733470917} +01/27/2022 11:38:12 - INFO - codeparrot_training - Step 16779: {'lr': 0.0003918846047898277, 'samples': 3221760, 'steps': 16779, 'loss/train': 1.1759809255599976} +01/27/2022 11:38:16 - INFO - codeparrot_training - Step 16780: {'lr': 0.00039187113250354635, 'samples': 3221952, 'steps': 16780, 'loss/train': 0.6449707746505737} +01/27/2022 11:38:20 - INFO - codeparrot_training - Step 16781: {'lr': 0.00039185765960953405, 'samples': 3222144, 'steps': 16781, 'loss/train': 0.2099103033542633} +01/27/2022 11:38:23 - INFO - codeparrot_training - Step 16782: {'lr': 0.0003918441861078486, 'samples': 3222336, 'steps': 16782, 'loss/train': 0.8961345255374908} +01/27/2022 11:38:26 - INFO - codeparrot_training - Step 16783: {'lr': 0.0003918307119985477, 'samples': 3222528, 'steps': 16783, 'loss/train': 0.732565239071846} +01/27/2022 11:38:30 - INFO - codeparrot_training - Step 16784: {'lr': 0.0003918172372816892, 'samples': 3222720, 'steps': 16784, 'loss/train': 0.7673167884349823} +01/27/2022 11:38:33 - INFO - codeparrot_training - Step 16785: {'lr': 0.0003918037619573305, 'samples': 3222912, 'steps': 16785, 'loss/train': 0.6072016060352325} +01/27/2022 11:38:36 - INFO - codeparrot_training - Step 16786: {'lr': 0.0003917902860255296, 'samples': 3223104, 'steps': 16786, 'loss/train': 0.8642038106918335} +01/27/2022 11:38:39 - INFO - codeparrot_training - Step 16787: {'lr': 0.0003917768094863441, 'samples': 3223296, 'steps': 16787, 'loss/train': 0.08242721110582352} +01/27/2022 11:38:42 - INFO - codeparrot_training - Step 16788: {'lr': 0.00039176333233983187, 'samples': 3223488, 'steps': 16788, 'loss/train': 0.5627690702676773} +01/27/2022 11:38:45 - INFO - codeparrot_training - Step 16789: {'lr': 0.0003917498545860504, 'samples': 3223680, 'steps': 16789, 'loss/train': 2.5399033427238464} +01/27/2022 11:38:50 - INFO - codeparrot_training - Step 16790: {'lr': 0.0003917363762250576, 'samples': 3223872, 'steps': 16790, 'loss/train': 0.9434401094913483} +01/27/2022 11:38:53 - INFO - codeparrot_training - Step 16791: {'lr': 0.00039172289725691124, 'samples': 3224064, 'steps': 16791, 'loss/train': 0.8442219793796539} +01/27/2022 11:38:56 - INFO - codeparrot_training - Step 16792: {'lr': 0.000391709417681669, 'samples': 3224256, 'steps': 16792, 'loss/train': 0.9742890894412994} +01/27/2022 11:38:59 - INFO - codeparrot_training - Step 16793: {'lr': 0.0003916959374993885, 'samples': 3224448, 'steps': 16793, 'loss/train': 0.5274993628263474} +01/27/2022 11:39:02 - INFO - codeparrot_training - Step 16794: {'lr': 0.0003916824567101277, 'samples': 3224640, 'steps': 16794, 'loss/train': 1.027316004037857} +01/27/2022 11:39:05 - INFO - codeparrot_training - Step 16795: {'lr': 0.0003916689753139442, 'samples': 3224832, 'steps': 16795, 'loss/train': 1.153609961271286} +01/27/2022 11:39:08 - INFO - codeparrot_training - Step 16796: {'lr': 0.0003916554933108958, 'samples': 3225024, 'steps': 16796, 'loss/train': 0.6559967547655106} +01/27/2022 11:39:12 - INFO - codeparrot_training - Step 16797: {'lr': 0.0003916420107010402, 'samples': 3225216, 'steps': 16797, 'loss/train': 0.15351517125964165} +01/27/2022 11:39:18 - INFO - codeparrot_training - Step 16798: {'lr': 0.0003916285274844353, 'samples': 3225408, 'steps': 16798, 'loss/train': 0.991411417722702} +01/27/2022 11:39:21 - INFO - codeparrot_training - Step 16799: {'lr': 0.0003916150436611387, 'samples': 3225600, 'steps': 16799, 'loss/train': 0.7129653096199036} +01/27/2022 11:39:24 - INFO - codeparrot_training - Step 16800: {'lr': 0.0003916015592312082, 'samples': 3225792, 'steps': 16800, 'loss/train': 0.6359747350215912} +01/27/2022 11:39:27 - INFO - codeparrot_training - Step 16801: {'lr': 0.00039158807419470166, 'samples': 3225984, 'steps': 16801, 'loss/train': 1.3880435228347778} +01/27/2022 11:39:30 - INFO - codeparrot_training - Step 16802: {'lr': 0.0003915745885516767, 'samples': 3226176, 'steps': 16802, 'loss/train': 0.6845333129167557} +01/27/2022 11:39:34 - INFO - codeparrot_training - Step 16803: {'lr': 0.0003915611023021912, 'samples': 3226368, 'steps': 16803, 'loss/train': 0.7006503939628601} +01/27/2022 11:39:37 - INFO - codeparrot_training - Step 16804: {'lr': 0.00039154761544630287, 'samples': 3226560, 'steps': 16804, 'loss/train': 1.3191989064216614} +01/27/2022 11:39:40 - INFO - codeparrot_training - Step 16805: {'lr': 0.0003915341279840695, 'samples': 3226752, 'steps': 16805, 'loss/train': 1.1211770474910736} +01/27/2022 11:39:43 - INFO - codeparrot_training - Step 16806: {'lr': 0.00039152063991554885, 'samples': 3226944, 'steps': 16806, 'loss/train': 0.77216836810112} +01/27/2022 11:39:47 - INFO - codeparrot_training - Step 16807: {'lr': 0.0003915071512407987, 'samples': 3227136, 'steps': 16807, 'loss/train': 1.017994076013565} +01/27/2022 11:39:51 - INFO - codeparrot_training - Step 16808: {'lr': 0.0003914936619598769, 'samples': 3227328, 'steps': 16808, 'loss/train': 0.9824958443641663} +01/27/2022 11:39:54 - INFO - codeparrot_training - Step 16809: {'lr': 0.00039148017207284115, 'samples': 3227520, 'steps': 16809, 'loss/train': 0.7405836582183838} +01/27/2022 11:39:57 - INFO - codeparrot_training - Step 16810: {'lr': 0.0003914666815797493, 'samples': 3227712, 'steps': 16810, 'loss/train': 1.2057855427265167} +01/27/2022 11:40:00 - INFO - codeparrot_training - Step 16811: {'lr': 0.00039145319048065907, 'samples': 3227904, 'steps': 16811, 'loss/train': 0.8157295882701874} +01/27/2022 11:40:03 - INFO - codeparrot_training - Step 16812: {'lr': 0.00039143969877562833, 'samples': 3228096, 'steps': 16812, 'loss/train': 0.5855377167463303} +01/27/2022 11:40:06 - INFO - codeparrot_training - Step 16813: {'lr': 0.00039142620646471485, 'samples': 3228288, 'steps': 16813, 'loss/train': 0.38132715225219727} +01/27/2022 11:40:09 - INFO - codeparrot_training - Step 16814: {'lr': 0.00039141271354797635, 'samples': 3228480, 'steps': 16814, 'loss/train': 1.009420484304428} +01/27/2022 11:40:13 - INFO - codeparrot_training - Step 16815: {'lr': 0.0003913992200254707, 'samples': 3228672, 'steps': 16815, 'loss/train': 0.924778014421463} +01/27/2022 11:40:17 - INFO - codeparrot_training - Step 16816: {'lr': 0.0003913857258972557, 'samples': 3228864, 'steps': 16816, 'loss/train': 1.2576481997966766} +01/27/2022 11:40:20 - INFO - codeparrot_training - Step 16817: {'lr': 0.0003913722311633892, 'samples': 3229056, 'steps': 16817, 'loss/train': 1.390502393245697} +01/27/2022 11:40:23 - INFO - codeparrot_training - Step 16818: {'lr': 0.0003913587358239288, 'samples': 3229248, 'steps': 16818, 'loss/train': 1.2251645922660828} +01/27/2022 11:40:26 - INFO - codeparrot_training - Step 16819: {'lr': 0.0003913452398789326, 'samples': 3229440, 'steps': 16819, 'loss/train': 0.6522023230791092} +01/27/2022 11:40:30 - INFO - codeparrot_training - Step 16820: {'lr': 0.0003913317433284582, 'samples': 3229632, 'steps': 16820, 'loss/train': 0.8496522903442383} +01/27/2022 11:40:33 - INFO - codeparrot_training - Step 16821: {'lr': 0.00039131824617256354, 'samples': 3229824, 'steps': 16821, 'loss/train': 0.6969248503446579} +01/27/2022 11:40:36 - INFO - codeparrot_training - Step 16822: {'lr': 0.0003913047484113064, 'samples': 3230016, 'steps': 16822, 'loss/train': 0.4682147651910782} +01/27/2022 11:40:39 - INFO - codeparrot_training - Step 16823: {'lr': 0.0003912912500447445, 'samples': 3230208, 'steps': 16823, 'loss/train': 0.7875845432281494} +01/27/2022 11:40:42 - INFO - codeparrot_training - Step 16824: {'lr': 0.0003912777510729358, 'samples': 3230400, 'steps': 16824, 'loss/train': 0.5978379249572754} +01/27/2022 11:40:49 - INFO - codeparrot_training - Step 16825: {'lr': 0.0003912642514959381, 'samples': 3230592, 'steps': 16825, 'loss/train': 0.11844279617071152} +01/27/2022 11:40:52 - INFO - codeparrot_training - Step 16826: {'lr': 0.00039125075131380923, 'samples': 3230784, 'steps': 16826, 'loss/train': 1.0011006295681} +01/27/2022 11:40:55 - INFO - codeparrot_training - Step 16827: {'lr': 0.00039123725052660696, 'samples': 3230976, 'steps': 16827, 'loss/train': 0.04758444428443909} +01/27/2022 11:40:58 - INFO - codeparrot_training - Step 16828: {'lr': 0.00039122374913438913, 'samples': 3231168, 'steps': 16828, 'loss/train': 1.0879873931407928} +01/27/2022 11:41:01 - INFO - codeparrot_training - Step 16829: {'lr': 0.00039121024713721365, 'samples': 3231360, 'steps': 16829, 'loss/train': 0.9653832614421844} +01/27/2022 11:41:04 - INFO - codeparrot_training - Step 16830: {'lr': 0.0003911967445351382, 'samples': 3231552, 'steps': 16830, 'loss/train': 0.9355493187904358} +01/27/2022 11:41:08 - INFO - codeparrot_training - Step 16831: {'lr': 0.00039118324132822083, 'samples': 3231744, 'steps': 16831, 'loss/train': 0.6660778373479843} +01/27/2022 11:41:11 - INFO - codeparrot_training - Step 16832: {'lr': 0.0003911697375165193, 'samples': 3231936, 'steps': 16832, 'loss/train': 0.9991978704929352} +01/27/2022 11:41:15 - INFO - codeparrot_training - Step 16833: {'lr': 0.00039115623310009135, 'samples': 3232128, 'steps': 16833, 'loss/train': 0.8684123754501343} +01/27/2022 11:41:18 - INFO - codeparrot_training - Step 16834: {'lr': 0.00039114272807899496, 'samples': 3232320, 'steps': 16834, 'loss/train': 0.8950560986995697} +01/27/2022 11:41:22 - INFO - codeparrot_training - Step 16835: {'lr': 0.000391129222453288, 'samples': 3232512, 'steps': 16835, 'loss/train': 1.0092829763889313} +01/27/2022 11:41:25 - INFO - codeparrot_training - Step 16836: {'lr': 0.00039111571622302824, 'samples': 3232704, 'steps': 16836, 'loss/train': 0.9063351452350616} +01/27/2022 11:41:28 - INFO - codeparrot_training - Step 16837: {'lr': 0.0003911022093882736, 'samples': 3232896, 'steps': 16837, 'loss/train': 0.598248302936554} +01/27/2022 11:41:31 - INFO - codeparrot_training - Step 16838: {'lr': 0.00039108870194908175, 'samples': 3233088, 'steps': 16838, 'loss/train': 0.8673560321331024} +01/27/2022 11:41:34 - INFO - codeparrot_training - Step 16839: {'lr': 0.00039107519390551085, 'samples': 3233280, 'steps': 16839, 'loss/train': 0.9127008318901062} +01/27/2022 11:41:37 - INFO - codeparrot_training - Step 16840: {'lr': 0.00039106168525761855, 'samples': 3233472, 'steps': 16840, 'loss/train': 0.7007948756217957} +01/27/2022 11:41:40 - INFO - codeparrot_training - Step 16841: {'lr': 0.00039104817600546277, 'samples': 3233664, 'steps': 16841, 'loss/train': 0.7678492069244385} +01/27/2022 11:41:45 - INFO - codeparrot_training - Step 16842: {'lr': 0.00039103466614910144, 'samples': 3233856, 'steps': 16842, 'loss/train': 0.9272978603839874} +01/27/2022 11:41:48 - INFO - codeparrot_training - Step 16843: {'lr': 0.0003910211556885923, 'samples': 3234048, 'steps': 16843, 'loss/train': 0.6041042357683182} +01/27/2022 11:41:51 - INFO - codeparrot_training - Step 16844: {'lr': 0.0003910076446239934, 'samples': 3234240, 'steps': 16844, 'loss/train': 0.9683482646942139} +01/27/2022 11:41:54 - INFO - codeparrot_training - Step 16845: {'lr': 0.00039099413295536246, 'samples': 3234432, 'steps': 16845, 'loss/train': 0.6289411336183548} +01/27/2022 11:41:58 - INFO - codeparrot_training - Step 16846: {'lr': 0.0003909806206827575, 'samples': 3234624, 'steps': 16846, 'loss/train': 0.3367649093270302} +01/27/2022 11:42:01 - INFO - codeparrot_training - Step 16847: {'lr': 0.00039096710780623625, 'samples': 3234816, 'steps': 16847, 'loss/train': 1.2910743355751038} +01/27/2022 11:42:04 - INFO - codeparrot_training - Step 16848: {'lr': 0.0003909535943258567, 'samples': 3235008, 'steps': 16848, 'loss/train': 0.6661932170391083} +01/27/2022 11:42:07 - INFO - codeparrot_training - Step 16849: {'lr': 0.0003909400802416767, 'samples': 3235200, 'steps': 16849, 'loss/train': 0.4175412654876709} +01/27/2022 11:42:10 - INFO - codeparrot_training - Step 16850: {'lr': 0.00039092656555375416, 'samples': 3235392, 'steps': 16850, 'loss/train': 1.1915354132652283} +01/27/2022 11:42:16 - INFO - codeparrot_training - Step 16851: {'lr': 0.00039091305026214704, 'samples': 3235584, 'steps': 16851, 'loss/train': 0.780450314283371} +01/27/2022 11:42:19 - INFO - codeparrot_training - Step 16852: {'lr': 0.0003908995343669131, 'samples': 3235776, 'steps': 16852, 'loss/train': 0.6312952190637589} +01/27/2022 11:42:22 - INFO - codeparrot_training - Step 16853: {'lr': 0.0003908860178681102, 'samples': 3235968, 'steps': 16853, 'loss/train': 1.2336612045764923} +01/27/2022 11:42:25 - INFO - codeparrot_training - Step 16854: {'lr': 0.0003908725007657964, 'samples': 3236160, 'steps': 16854, 'loss/train': 0.5714336335659027} +01/27/2022 11:42:28 - INFO - codeparrot_training - Step 16855: {'lr': 0.0003908589830600296, 'samples': 3236352, 'steps': 16855, 'loss/train': 0.7534431517124176} +01/27/2022 11:42:31 - INFO - codeparrot_training - Step 16856: {'lr': 0.0003908454647508676, 'samples': 3236544, 'steps': 16856, 'loss/train': 1.0875763893127441} +01/27/2022 11:42:35 - INFO - codeparrot_training - Step 16857: {'lr': 0.00039083194583836836, 'samples': 3236736, 'steps': 16857, 'loss/train': 0.6401609480381012} +01/27/2022 11:42:38 - INFO - codeparrot_training - Step 16858: {'lr': 0.0003908184263225898, 'samples': 3236928, 'steps': 16858, 'loss/train': 2.4269081354141235} +01/27/2022 11:42:41 - INFO - codeparrot_training - Step 16859: {'lr': 0.0003908049062035898, 'samples': 3237120, 'steps': 16859, 'loss/train': 0.7410305142402649} +01/27/2022 11:42:45 - INFO - codeparrot_training - Step 16860: {'lr': 0.0003907913854814262, 'samples': 3237312, 'steps': 16860, 'loss/train': 0.9646310806274414} +01/27/2022 11:42:48 - INFO - codeparrot_training - Step 16861: {'lr': 0.00039077786415615714, 'samples': 3237504, 'steps': 16861, 'loss/train': 0.7373970597982407} +01/27/2022 11:42:52 - INFO - codeparrot_training - Step 16862: {'lr': 0.0003907643422278404, 'samples': 3237696, 'steps': 16862, 'loss/train': 0.5253083556890488} +01/27/2022 11:42:55 - INFO - codeparrot_training - Step 16863: {'lr': 0.00039075081969653383, 'samples': 3237888, 'steps': 16863, 'loss/train': 0.9523229598999023} +01/27/2022 11:42:58 - INFO - codeparrot_training - Step 16864: {'lr': 0.0003907372965622955, 'samples': 3238080, 'steps': 16864, 'loss/train': 0.7623406648635864} +01/27/2022 11:43:01 - INFO - codeparrot_training - Step 16865: {'lr': 0.0003907237728251833, 'samples': 3238272, 'steps': 16865, 'loss/train': 0.6992100030183792} +01/27/2022 11:43:04 - INFO - codeparrot_training - Step 16866: {'lr': 0.0003907102484852551, 'samples': 3238464, 'steps': 16866, 'loss/train': 0.8990136086940765} +01/27/2022 11:43:07 - INFO - codeparrot_training - Step 16867: {'lr': 0.0003906967235425689, 'samples': 3238656, 'steps': 16867, 'loss/train': 0.8337095081806183} +01/27/2022 11:43:11 - INFO - codeparrot_training - Step 16868: {'lr': 0.0003906831979971826, 'samples': 3238848, 'steps': 16868, 'loss/train': 0.8573309183120728} +01/27/2022 11:43:15 - INFO - codeparrot_training - Step 16869: {'lr': 0.0003906696718491541, 'samples': 3239040, 'steps': 16869, 'loss/train': 0.4679349213838577} +01/27/2022 11:43:18 - INFO - codeparrot_training - Step 16870: {'lr': 0.0003906561450985415, 'samples': 3239232, 'steps': 16870, 'loss/train': 0.4599645584821701} +01/27/2022 11:43:21 - INFO - codeparrot_training - Step 16871: {'lr': 0.00039064261774540254, 'samples': 3239424, 'steps': 16871, 'loss/train': 0.8348590135574341} +01/27/2022 11:43:24 - INFO - codeparrot_training - Step 16872: {'lr': 0.0003906290897897953, 'samples': 3239616, 'steps': 16872, 'loss/train': 0.7666544616222382} +01/27/2022 11:43:28 - INFO - codeparrot_training - Step 16873: {'lr': 0.00039061556123177777, 'samples': 3239808, 'steps': 16873, 'loss/train': 0.4653368890285492} +01/27/2022 11:43:31 - INFO - codeparrot_training - Step 16874: {'lr': 0.00039060203207140774, 'samples': 3240000, 'steps': 16874, 'loss/train': 0.7820010781288147} +01/27/2022 11:43:34 - INFO - codeparrot_training - Step 16875: {'lr': 0.0003905885023087433, 'samples': 3240192, 'steps': 16875, 'loss/train': 1.2156483829021454} +01/27/2022 11:43:37 - INFO - codeparrot_training - Step 16876: {'lr': 0.0003905749719438423, 'samples': 3240384, 'steps': 16876, 'loss/train': 0.2888473570346832} +01/27/2022 11:43:43 - INFO - codeparrot_training - Step 16877: {'lr': 0.00039056144097676285, 'samples': 3240576, 'steps': 16877, 'loss/train': 0.2498508095741272} +01/27/2022 11:43:46 - INFO - codeparrot_training - Step 16878: {'lr': 0.0003905479094075627, 'samples': 3240768, 'steps': 16878, 'loss/train': 0.9661105871200562} +01/27/2022 11:43:49 - INFO - codeparrot_training - Step 16879: {'lr': 0.00039053437723630003, 'samples': 3240960, 'steps': 16879, 'loss/train': 0.9111748337745667} +01/27/2022 11:43:53 - INFO - codeparrot_training - Step 16880: {'lr': 0.00039052084446303264, 'samples': 3241152, 'steps': 16880, 'loss/train': 0.5124429166316986} +01/27/2022 11:43:56 - INFO - codeparrot_training - Step 16881: {'lr': 0.0003905073110878186, 'samples': 3241344, 'steps': 16881, 'loss/train': 0.6561328321695328} +01/27/2022 11:43:59 - INFO - codeparrot_training - Step 16882: {'lr': 0.00039049377711071595, 'samples': 3241536, 'steps': 16882, 'loss/train': 0.9551517069339752} +01/27/2022 11:44:02 - INFO - codeparrot_training - Step 16883: {'lr': 0.00039048024253178243, 'samples': 3241728, 'steps': 16883, 'loss/train': 0.8015717267990112} +01/27/2022 11:44:05 - INFO - codeparrot_training - Step 16884: {'lr': 0.00039046670735107627, 'samples': 3241920, 'steps': 16884, 'loss/train': 0.7641280889511108} +01/27/2022 11:44:08 - INFO - codeparrot_training - Step 16885: {'lr': 0.00039045317156865525, 'samples': 3242112, 'steps': 16885, 'loss/train': 0.8805055618286133} +01/27/2022 11:44:13 - INFO - codeparrot_training - Step 16886: {'lr': 0.0003904396351845775, 'samples': 3242304, 'steps': 16886, 'loss/train': 0.6275840252637863} +01/27/2022 11:44:16 - INFO - codeparrot_training - Step 16887: {'lr': 0.00039042609819890087, 'samples': 3242496, 'steps': 16887, 'loss/train': 2.1334437131881714} +01/27/2022 11:44:19 - INFO - codeparrot_training - Step 16888: {'lr': 0.0003904125606116835, 'samples': 3242688, 'steps': 16888, 'loss/train': 0.7131605297327042} +01/27/2022 11:44:22 - INFO - codeparrot_training - Step 16889: {'lr': 0.0003903990224229833, 'samples': 3242880, 'steps': 16889, 'loss/train': 1.0128754377365112} +01/27/2022 11:44:25 - INFO - codeparrot_training - Step 16890: {'lr': 0.00039038548363285825, 'samples': 3243072, 'steps': 16890, 'loss/train': 0.5173426866531372} +01/27/2022 11:44:28 - INFO - codeparrot_training - Step 16891: {'lr': 0.00039037194424136634, 'samples': 3243264, 'steps': 16891, 'loss/train': 1.3186720311641693} +01/27/2022 11:44:32 - INFO - codeparrot_training - Step 16892: {'lr': 0.0003903584042485656, 'samples': 3243456, 'steps': 16892, 'loss/train': 0.9015749394893646} +01/27/2022 11:44:35 - INFO - codeparrot_training - Step 16893: {'lr': 0.00039034486365451405, 'samples': 3243648, 'steps': 16893, 'loss/train': 0.924083411693573} +01/27/2022 11:44:38 - INFO - codeparrot_training - Step 16894: {'lr': 0.00039033132245926974, 'samples': 3243840, 'steps': 16894, 'loss/train': 0.9485701024532318} +01/27/2022 11:44:44 - INFO - codeparrot_training - Step 16895: {'lr': 0.0003903177806628905, 'samples': 3244032, 'steps': 16895, 'loss/train': 0.5335685759782791} +01/27/2022 11:44:47 - INFO - codeparrot_training - Step 16896: {'lr': 0.00039030423826543446, 'samples': 3244224, 'steps': 16896, 'loss/train': 0.8597194254398346} +01/27/2022 11:44:50 - INFO - codeparrot_training - Step 16897: {'lr': 0.0003902906952669596, 'samples': 3244416, 'steps': 16897, 'loss/train': 1.0436486899852753} +01/27/2022 11:44:54 - INFO - codeparrot_training - Step 16898: {'lr': 0.000390277151667524, 'samples': 3244608, 'steps': 16898, 'loss/train': 1.1486335694789886} +01/27/2022 11:44:57 - INFO - codeparrot_training - Step 16899: {'lr': 0.0003902636074671856, 'samples': 3244800, 'steps': 16899, 'loss/train': 1.0118916928768158} +01/27/2022 11:45:00 - INFO - codeparrot_training - Step 16900: {'lr': 0.0003902500626660025, 'samples': 3244992, 'steps': 16900, 'loss/train': 1.076905757188797} +01/27/2022 11:45:03 - INFO - codeparrot_training - Step 16901: {'lr': 0.00039023651726403263, 'samples': 3245184, 'steps': 16901, 'loss/train': 0.7469543516635895} +01/27/2022 11:45:06 - INFO - codeparrot_training - Step 16902: {'lr': 0.00039022297126133397, 'samples': 3245376, 'steps': 16902, 'loss/train': 0.1911558285355568} +01/27/2022 11:45:09 - INFO - codeparrot_training - Step 16903: {'lr': 0.0003902094246579647, 'samples': 3245568, 'steps': 16903, 'loss/train': 0.6816944181919098} +01/27/2022 11:45:14 - INFO - codeparrot_training - Step 16904: {'lr': 0.00039019587745398276, 'samples': 3245760, 'steps': 16904, 'loss/train': 0.6018802374601364} +01/27/2022 11:45:17 - INFO - codeparrot_training - Step 16905: {'lr': 0.00039018232964944623, 'samples': 3245952, 'steps': 16905, 'loss/train': 0.9022212624549866} +01/27/2022 11:45:20 - INFO - codeparrot_training - Step 16906: {'lr': 0.0003901687812444131, 'samples': 3246144, 'steps': 16906, 'loss/train': 1.492704838514328} +01/27/2022 11:45:24 - INFO - codeparrot_training - Step 16907: {'lr': 0.0003901552322389414, 'samples': 3246336, 'steps': 16907, 'loss/train': 0.6090226918458939} +01/27/2022 11:45:27 - INFO - codeparrot_training - Step 16908: {'lr': 0.00039014168263308926, 'samples': 3246528, 'steps': 16908, 'loss/train': 0.7020551115274429} +01/27/2022 11:45:30 - INFO - codeparrot_training - Step 16909: {'lr': 0.00039012813242691454, 'samples': 3246720, 'steps': 16909, 'loss/train': 0.5655768066644669} +01/27/2022 11:45:33 - INFO - codeparrot_training - Step 16910: {'lr': 0.00039011458162047547, 'samples': 3246912, 'steps': 16910, 'loss/train': 0.5809158682823181} +01/27/2022 11:45:36 - INFO - codeparrot_training - Step 16911: {'lr': 0.00039010103021383, 'samples': 3247104, 'steps': 16911, 'loss/train': 0.8071163892745972} +01/27/2022 11:45:39 - INFO - codeparrot_training - Step 16912: {'lr': 0.00039008747820703615, 'samples': 3247296, 'steps': 16912, 'loss/train': 0.6788030862808228} +01/27/2022 11:45:44 - INFO - codeparrot_training - Step 16913: {'lr': 0.0003900739256001521, 'samples': 3247488, 'steps': 16913, 'loss/train': 0.7524007558822632} +01/27/2022 11:45:47 - INFO - codeparrot_training - Step 16914: {'lr': 0.00039006037239323584, 'samples': 3247680, 'steps': 16914, 'loss/train': 0.777653306722641} +01/27/2022 11:45:50 - INFO - codeparrot_training - Step 16915: {'lr': 0.00039004681858634537, 'samples': 3247872, 'steps': 16915, 'loss/train': 0.6411704868078232} +01/27/2022 11:45:53 - INFO - codeparrot_training - Step 16916: {'lr': 0.0003900332641795388, 'samples': 3248064, 'steps': 16916, 'loss/train': 1.066002756357193} +01/27/2022 11:45:56 - INFO - codeparrot_training - Step 16917: {'lr': 0.0003900197091728742, 'samples': 3248256, 'steps': 16917, 'loss/train': 1.163102924823761} +01/27/2022 11:45:59 - INFO - codeparrot_training - Step 16918: {'lr': 0.0003900061535664097, 'samples': 3248448, 'steps': 16918, 'loss/train': 0.4907820224761963} +01/27/2022 11:46:03 - INFO - codeparrot_training - Step 16919: {'lr': 0.0003899925973602032, 'samples': 3248640, 'steps': 16919, 'loss/train': 1.062075287103653} +01/27/2022 11:46:06 - INFO - codeparrot_training - Step 16920: {'lr': 0.0003899790405543129, 'samples': 3248832, 'steps': 16920, 'loss/train': 0.711375430226326} +01/27/2022 11:46:09 - INFO - codeparrot_training - Step 16921: {'lr': 0.0003899654831487969, 'samples': 3249024, 'steps': 16921, 'loss/train': 0.5722496062517166} +01/27/2022 11:46:14 - INFO - codeparrot_training - Step 16922: {'lr': 0.0003899519251437131, 'samples': 3249216, 'steps': 16922, 'loss/train': 0.1864769607782364} +01/27/2022 11:46:17 - INFO - codeparrot_training - Step 16923: {'lr': 0.00038993836653911974, 'samples': 3249408, 'steps': 16923, 'loss/train': 0.9627458453178406} +01/27/2022 11:46:20 - INFO - codeparrot_training - Step 16924: {'lr': 0.00038992480733507487, 'samples': 3249600, 'steps': 16924, 'loss/train': 0.4025430679321289} +01/27/2022 11:46:23 - INFO - codeparrot_training - Step 16925: {'lr': 0.0003899112475316365, 'samples': 3249792, 'steps': 16925, 'loss/train': 0.7798760533332825} +01/27/2022 11:46:26 - INFO - codeparrot_training - Step 16926: {'lr': 0.00038989768712886287, 'samples': 3249984, 'steps': 16926, 'loss/train': 1.0067017078399658} +01/27/2022 11:46:30 - INFO - codeparrot_training - Step 16927: {'lr': 0.0003898841261268119, 'samples': 3250176, 'steps': 16927, 'loss/train': 1.3577590584754944} +01/27/2022 11:46:33 - INFO - codeparrot_training - Step 16928: {'lr': 0.00038987056452554177, 'samples': 3250368, 'steps': 16928, 'loss/train': 0.967994213104248} +01/27/2022 11:46:36 - INFO - codeparrot_training - Step 16929: {'lr': 0.00038985700232511055, 'samples': 3250560, 'steps': 16929, 'loss/train': 0.9476949870586395} +01/27/2022 11:46:42 - INFO - codeparrot_training - Step 16930: {'lr': 0.0003898434395255763, 'samples': 3250752, 'steps': 16930, 'loss/train': 1.0211888551712036} +01/27/2022 11:46:45 - INFO - codeparrot_training - Step 16931: {'lr': 0.0003898298761269973, 'samples': 3250944, 'steps': 16931, 'loss/train': 0.5557549595832825} +01/27/2022 11:46:49 - INFO - codeparrot_training - Step 16932: {'lr': 0.0003898163121294314, 'samples': 3251136, 'steps': 16932, 'loss/train': 0.7581929862499237} +01/27/2022 11:46:52 - INFO - codeparrot_training - Step 16933: {'lr': 0.0003898027475329368, 'samples': 3251328, 'steps': 16933, 'loss/train': 0.6421479284763336} +01/27/2022 11:46:55 - INFO - codeparrot_training - Step 16934: {'lr': 0.00038978918233757167, 'samples': 3251520, 'steps': 16934, 'loss/train': 1.3237614333629608} +01/27/2022 11:46:58 - INFO - codeparrot_training - Step 16935: {'lr': 0.000389775616543394, 'samples': 3251712, 'steps': 16935, 'loss/train': 0.7780162990093231} +01/27/2022 11:47:01 - INFO - codeparrot_training - Step 16936: {'lr': 0.00038976205015046206, 'samples': 3251904, 'steps': 16936, 'loss/train': 0.6622701287269592} +01/27/2022 11:47:04 - INFO - codeparrot_training - Step 16937: {'lr': 0.00038974848315883383, 'samples': 3252096, 'steps': 16937, 'loss/train': 0.712713897228241} +01/27/2022 11:47:08 - INFO - codeparrot_training - Step 16938: {'lr': 0.00038973491556856755, 'samples': 3252288, 'steps': 16938, 'loss/train': 0.920296847820282} +01/27/2022 11:47:12 - INFO - codeparrot_training - Step 16939: {'lr': 0.0003897213473797212, 'samples': 3252480, 'steps': 16939, 'loss/train': 1.2752052247524261} +01/27/2022 11:47:15 - INFO - codeparrot_training - Step 16940: {'lr': 0.0003897077785923529, 'samples': 3252672, 'steps': 16940, 'loss/train': 0.941489964723587} +01/27/2022 11:47:18 - INFO - codeparrot_training - Step 16941: {'lr': 0.0003896942092065209, 'samples': 3252864, 'steps': 16941, 'loss/train': 0.8200270235538483} +01/27/2022 11:47:21 - INFO - codeparrot_training - Step 16942: {'lr': 0.0003896806392222833, 'samples': 3253056, 'steps': 16942, 'loss/train': 1.1924895644187927} +01/27/2022 11:47:25 - INFO - codeparrot_training - Step 16943: {'lr': 0.00038966706863969815, 'samples': 3253248, 'steps': 16943, 'loss/train': 0.82960644364357} +01/27/2022 11:47:28 - INFO - codeparrot_training - Step 16944: {'lr': 0.00038965349745882365, 'samples': 3253440, 'steps': 16944, 'loss/train': 0.6690224558115005} +01/27/2022 11:47:31 - INFO - codeparrot_training - Step 16945: {'lr': 0.00038963992567971794, 'samples': 3253632, 'steps': 16945, 'loss/train': 0.07631062343716621} +01/27/2022 11:47:34 - INFO - codeparrot_training - Step 16946: {'lr': 0.0003896263533024391, 'samples': 3253824, 'steps': 16946, 'loss/train': 0.6476037204265594} +01/27/2022 11:47:37 - INFO - codeparrot_training - Step 16947: {'lr': 0.0003896127803270453, 'samples': 3254016, 'steps': 16947, 'loss/train': 0.7610760927200317} +01/27/2022 11:47:41 - INFO - codeparrot_training - Step 16948: {'lr': 0.0003895992067535946, 'samples': 3254208, 'steps': 16948, 'loss/train': 0.7775023877620697} +01/27/2022 11:47:45 - INFO - codeparrot_training - Step 16949: {'lr': 0.0003895856325821454, 'samples': 3254400, 'steps': 16949, 'loss/train': 0.8749565184116364} +01/27/2022 11:47:48 - INFO - codeparrot_training - Step 16950: {'lr': 0.00038957205781275554, 'samples': 3254592, 'steps': 16950, 'loss/train': 0.7663524448871613} +01/27/2022 11:47:51 - INFO - codeparrot_training - Step 16951: {'lr': 0.00038955848244548333, 'samples': 3254784, 'steps': 16951, 'loss/train': 0.6056289821863174} +01/27/2022 11:47:54 - INFO - codeparrot_training - Step 16952: {'lr': 0.00038954490648038687, 'samples': 3254976, 'steps': 16952, 'loss/train': 1.1136817932128906} +01/27/2022 11:47:57 - INFO - codeparrot_training - Step 16953: {'lr': 0.0003895313299175244, 'samples': 3255168, 'steps': 16953, 'loss/train': 0.7329642623662949} +01/27/2022 11:48:00 - INFO - codeparrot_training - Step 16954: {'lr': 0.000389517752756954, 'samples': 3255360, 'steps': 16954, 'loss/train': 0.5365540087223053} +01/27/2022 11:48:03 - INFO - codeparrot_training - Step 16955: {'lr': 0.0003895041749987338, 'samples': 3255552, 'steps': 16955, 'loss/train': 0.8544944822788239} +01/27/2022 11:48:07 - INFO - codeparrot_training - Step 16956: {'lr': 0.00038949059664292207, 'samples': 3255744, 'steps': 16956, 'loss/train': 0.9806711375713348} +01/27/2022 11:48:13 - INFO - codeparrot_training - Step 16957: {'lr': 0.0003894770176895769, 'samples': 3255936, 'steps': 16957, 'loss/train': 0.6531978249549866} +01/27/2022 11:48:16 - INFO - codeparrot_training - Step 16958: {'lr': 0.0003894634381387565, 'samples': 3256128, 'steps': 16958, 'loss/train': 0.9627905488014221} +01/27/2022 11:48:19 - INFO - codeparrot_training - Step 16959: {'lr': 0.00038944985799051896, 'samples': 3256320, 'steps': 16959, 'loss/train': 1.0260493755340576} +01/27/2022 11:48:22 - INFO - codeparrot_training - Step 16960: {'lr': 0.0003894362772449226, 'samples': 3256512, 'steps': 16960, 'loss/train': 0.8106839060783386} +01/27/2022 11:48:26 - INFO - codeparrot_training - Step 16961: {'lr': 0.0003894226959020254, 'samples': 3256704, 'steps': 16961, 'loss/train': 1.4848626852035522} +01/27/2022 11:48:29 - INFO - codeparrot_training - Step 16962: {'lr': 0.00038940911396188573, 'samples': 3256896, 'steps': 16962, 'loss/train': 0.9038337171077728} +01/27/2022 11:48:32 - INFO - codeparrot_training - Step 16963: {'lr': 0.0003893955314245616, 'samples': 3257088, 'steps': 16963, 'loss/train': 0.5744192451238632} +01/27/2022 11:48:35 - INFO - codeparrot_training - Step 16964: {'lr': 0.0003893819482901113, 'samples': 3257280, 'steps': 16964, 'loss/train': 0.911381721496582} +01/27/2022 11:48:40 - INFO - codeparrot_training - Step 16965: {'lr': 0.000389368364558593, 'samples': 3257472, 'steps': 16965, 'loss/train': 0.8829802572727203} +01/27/2022 11:48:43 - INFO - codeparrot_training - Step 16966: {'lr': 0.00038935478023006487, 'samples': 3257664, 'steps': 16966, 'loss/train': 0.8382733762264252} +01/27/2022 11:48:46 - INFO - codeparrot_training - Step 16967: {'lr': 0.0003893411953045852, 'samples': 3257856, 'steps': 16967, 'loss/train': 0.9348568618297577} +01/27/2022 11:48:49 - INFO - codeparrot_training - Step 16968: {'lr': 0.000389327609782212, 'samples': 3258048, 'steps': 16968, 'loss/train': 1.2356248497962952} +01/27/2022 11:48:52 - INFO - codeparrot_training - Step 16969: {'lr': 0.0003893140236630036, 'samples': 3258240, 'steps': 16969, 'loss/train': 0.057005250826478004} +01/27/2022 11:48:55 - INFO - codeparrot_training - Step 16970: {'lr': 0.0003893004369470181, 'samples': 3258432, 'steps': 16970, 'loss/train': 0.5362645536661148} +01/27/2022 11:48:59 - INFO - codeparrot_training - Step 16971: {'lr': 0.00038928684963431383, 'samples': 3258624, 'steps': 16971, 'loss/train': 0.9633254706859589} +01/27/2022 11:49:02 - INFO - codeparrot_training - Step 16972: {'lr': 0.00038927326172494894, 'samples': 3258816, 'steps': 16972, 'loss/train': 0.6587191969156265} +01/27/2022 11:49:05 - INFO - codeparrot_training - Step 16973: {'lr': 0.0003892596732189816, 'samples': 3259008, 'steps': 16973, 'loss/train': 0.7833366394042969} +01/27/2022 11:49:11 - INFO - codeparrot_training - Step 16974: {'lr': 0.00038924608411647, 'samples': 3259200, 'steps': 16974, 'loss/train': 1.1013591885566711} +01/27/2022 11:49:14 - INFO - codeparrot_training - Step 16975: {'lr': 0.00038923249441747245, 'samples': 3259392, 'steps': 16975, 'loss/train': 0.8554796576499939} +01/27/2022 11:49:17 - INFO - codeparrot_training - Step 16976: {'lr': 0.000389218904122047, 'samples': 3259584, 'steps': 16976, 'loss/train': 0.35823018103837967} +01/27/2022 11:49:20 - INFO - codeparrot_training - Step 16977: {'lr': 0.00038920531323025206, 'samples': 3259776, 'steps': 16977, 'loss/train': 0.28835946321487427} +01/27/2022 11:49:24 - INFO - codeparrot_training - Step 16978: {'lr': 0.0003891917217421458, 'samples': 3259968, 'steps': 16978, 'loss/train': 0.9356774389743805} +01/27/2022 11:49:27 - INFO - codeparrot_training - Step 16979: {'lr': 0.00038917812965778625, 'samples': 3260160, 'steps': 16979, 'loss/train': 0.8370275795459747} +01/27/2022 11:49:30 - INFO - codeparrot_training - Step 16980: {'lr': 0.00038916453697723194, 'samples': 3260352, 'steps': 16980, 'loss/train': 1.3011333346366882} +01/27/2022 11:49:33 - INFO - codeparrot_training - Step 16981: {'lr': 0.00038915094370054083, 'samples': 3260544, 'steps': 16981, 'loss/train': 0.8093619346618652} +01/27/2022 11:49:36 - INFO - codeparrot_training - Step 16982: {'lr': 0.00038913734982777136, 'samples': 3260736, 'steps': 16982, 'loss/train': 0.8828321099281311} +01/27/2022 11:49:40 - INFO - codeparrot_training - Step 16983: {'lr': 0.0003891237553589816, 'samples': 3260928, 'steps': 16983, 'loss/train': 1.0371741950511932} +01/27/2022 11:49:44 - INFO - codeparrot_training - Step 16984: {'lr': 0.00038911016029422984, 'samples': 3261120, 'steps': 16984, 'loss/train': 0.4332704097032547} +01/27/2022 11:49:47 - INFO - codeparrot_training - Step 16985: {'lr': 0.0003890965646335744, 'samples': 3261312, 'steps': 16985, 'loss/train': 1.1151655912399292} +01/27/2022 11:49:50 - INFO - codeparrot_training - Step 16986: {'lr': 0.0003890829683770734, 'samples': 3261504, 'steps': 16986, 'loss/train': 0.3305295333266258} +01/27/2022 11:49:53 - INFO - codeparrot_training - Step 16987: {'lr': 0.0003890693715247851, 'samples': 3261696, 'steps': 16987, 'loss/train': 0.666034609079361} +01/27/2022 11:49:56 - INFO - codeparrot_training - Step 16988: {'lr': 0.0003890557740767678, 'samples': 3261888, 'steps': 16988, 'loss/train': 0.8982296884059906} +01/27/2022 11:49:59 - INFO - codeparrot_training - Step 16989: {'lr': 0.0003890421760330798, 'samples': 3262080, 'steps': 16989, 'loss/train': 1.086978167295456} +01/27/2022 11:50:02 - INFO - codeparrot_training - Step 16990: {'lr': 0.0003890285773937792, 'samples': 3262272, 'steps': 16990, 'loss/train': 0.8090126216411591} +01/27/2022 11:50:06 - INFO - codeparrot_training - Step 16991: {'lr': 0.0003890149781589243, 'samples': 3262464, 'steps': 16991, 'loss/train': 0.6847304552793503} +01/27/2022 11:50:10 - INFO - codeparrot_training - Step 16992: {'lr': 0.0003890013783285733, 'samples': 3262656, 'steps': 16992, 'loss/train': 0.5485451817512512} +01/27/2022 11:50:13 - INFO - codeparrot_training - Step 16993: {'lr': 0.00038898777790278465, 'samples': 3262848, 'steps': 16993, 'loss/train': 0.7054782807826996} +01/27/2022 11:50:16 - INFO - codeparrot_training - Step 16994: {'lr': 0.00038897417688161644, 'samples': 3263040, 'steps': 16994, 'loss/train': 0.948241800069809} +01/27/2022 11:50:19 - INFO - codeparrot_training - Step 16995: {'lr': 0.0003889605752651271, 'samples': 3263232, 'steps': 16995, 'loss/train': 1.1427498757839203} +01/27/2022 11:50:23 - INFO - codeparrot_training - Step 16996: {'lr': 0.0003889469730533746, 'samples': 3263424, 'steps': 16996, 'loss/train': 0.4091653972864151} +01/27/2022 11:50:26 - INFO - codeparrot_training - Step 16997: {'lr': 0.0003889333702464175, 'samples': 3263616, 'steps': 16997, 'loss/train': 1.4571962356567383} +01/27/2022 11:50:29 - INFO - codeparrot_training - Step 16998: {'lr': 0.00038891976684431395, 'samples': 3263808, 'steps': 16998, 'loss/train': 0.8151625692844391} +01/27/2022 11:50:32 - INFO - codeparrot_training - Step 16999: {'lr': 0.0003889061628471222, 'samples': 3264000, 'steps': 16999, 'loss/train': 0.5776092857122421} +01/27/2022 11:50:35 - INFO - codeparrot_training - Step 17000: {'lr': 0.00038889255825490053, 'samples': 3264192, 'steps': 17000, 'loss/train': 0.7305838465690613} +01/27/2022 11:50:42 - INFO - codeparrot_training - Step 17001: {'lr': 0.0003888789530677073, 'samples': 3264384, 'steps': 17001, 'loss/train': 0.32645507901906967} +01/27/2022 11:50:45 - INFO - codeparrot_training - Step 17002: {'lr': 0.00038886534728560073, 'samples': 3264576, 'steps': 17002, 'loss/train': 0.4996393471956253} +01/27/2022 11:50:48 - INFO - codeparrot_training - Step 17003: {'lr': 0.0003888517409086391, 'samples': 3264768, 'steps': 17003, 'loss/train': 0.9575718641281128} +01/27/2022 11:50:51 - INFO - codeparrot_training - Step 17004: {'lr': 0.0003888381339368807, 'samples': 3264960, 'steps': 17004, 'loss/train': 0.9887277781963348} +01/27/2022 11:50:54 - INFO - codeparrot_training - Step 17005: {'lr': 0.00038882452637038377, 'samples': 3265152, 'steps': 17005, 'loss/train': 0.2427799478173256} +01/27/2022 11:50:58 - INFO - codeparrot_training - Step 17006: {'lr': 0.00038881091820920676, 'samples': 3265344, 'steps': 17006, 'loss/train': 1.0604884922504425} +01/27/2022 11:51:01 - INFO - codeparrot_training - Step 17007: {'lr': 0.00038879730945340775, 'samples': 3265536, 'steps': 17007, 'loss/train': 0.8758994936943054} +01/27/2022 11:51:04 - INFO - codeparrot_training - Step 17008: {'lr': 0.0003887837001030452, 'samples': 3265728, 'steps': 17008, 'loss/train': 0.1941739171743393} +01/27/2022 11:51:08 - INFO - codeparrot_training - Step 17009: {'lr': 0.00038877009015817734, 'samples': 3265920, 'steps': 17009, 'loss/train': 1.8740796446800232} +01/27/2022 11:51:11 - INFO - codeparrot_training - Step 17010: {'lr': 0.0003887564796188625, 'samples': 3266112, 'steps': 17010, 'loss/train': 0.6667362749576569} +01/27/2022 11:51:15 - INFO - codeparrot_training - Step 17011: {'lr': 0.0003887428684851589, 'samples': 3266304, 'steps': 17011, 'loss/train': 1.0712314546108246} +01/27/2022 11:51:18 - INFO - codeparrot_training - Step 17012: {'lr': 0.00038872925675712493, 'samples': 3266496, 'steps': 17012, 'loss/train': 0.8743101954460144} +01/27/2022 11:51:21 - INFO - codeparrot_training - Step 17013: {'lr': 0.00038871564443481886, 'samples': 3266688, 'steps': 17013, 'loss/train': 0.5236360430717468} +01/27/2022 11:51:24 - INFO - codeparrot_training - Step 17014: {'lr': 0.0003887020315182991, 'samples': 3266880, 'steps': 17014, 'loss/train': 1.0370870232582092} +01/27/2022 11:51:27 - INFO - codeparrot_training - Step 17015: {'lr': 0.0003886884180076238, 'samples': 3267072, 'steps': 17015, 'loss/train': 1.2306594550609589} +01/27/2022 11:51:30 - INFO - codeparrot_training - Step 17016: {'lr': 0.0003886748039028514, 'samples': 3267264, 'steps': 17016, 'loss/train': 0.7918047308921814} +01/27/2022 11:51:33 - INFO - codeparrot_training - Step 17017: {'lr': 0.00038866118920404013, 'samples': 3267456, 'steps': 17017, 'loss/train': 1.0017724335193634} +01/27/2022 11:51:38 - INFO - codeparrot_training - Step 17018: {'lr': 0.0003886475739112484, 'samples': 3267648, 'steps': 17018, 'loss/train': 0.662424311041832} +01/27/2022 11:51:41 - INFO - codeparrot_training - Step 17019: {'lr': 0.0003886339580245344, 'samples': 3267840, 'steps': 17019, 'loss/train': 0.8934758305549622} +01/27/2022 11:51:44 - INFO - codeparrot_training - Step 17020: {'lr': 0.00038862034154395664, 'samples': 3268032, 'steps': 17020, 'loss/train': 0.9115326404571533} +01/27/2022 11:51:47 - INFO - codeparrot_training - Step 17021: {'lr': 0.00038860672446957336, 'samples': 3268224, 'steps': 17021, 'loss/train': 0.839954674243927} +01/27/2022 11:51:51 - INFO - codeparrot_training - Step 17022: {'lr': 0.00038859310680144276, 'samples': 3268416, 'steps': 17022, 'loss/train': 0.7890474200248718} +01/27/2022 11:51:54 - INFO - codeparrot_training - Step 17023: {'lr': 0.0003885794885396234, 'samples': 3268608, 'steps': 17023, 'loss/train': 0.7510886192321777} +01/27/2022 11:51:57 - INFO - codeparrot_training - Step 17024: {'lr': 0.00038856586968417353, 'samples': 3268800, 'steps': 17024, 'loss/train': 1.0419365465641022} +01/27/2022 11:52:00 - INFO - codeparrot_training - Step 17025: {'lr': 0.0003885522502351514, 'samples': 3268992, 'steps': 17025, 'loss/train': 0.8937330543994904} +01/27/2022 11:52:03 - INFO - codeparrot_training - Step 17026: {'lr': 0.0003885386301926155, 'samples': 3269184, 'steps': 17026, 'loss/train': 1.1400487124919891} +01/27/2022 11:52:08 - INFO - codeparrot_training - Step 17027: {'lr': 0.00038852500955662407, 'samples': 3269376, 'steps': 17027, 'loss/train': 0.9712927043437958} +01/27/2022 11:52:11 - INFO - codeparrot_training - Step 17028: {'lr': 0.0003885113883272355, 'samples': 3269568, 'steps': 17028, 'loss/train': 0.6891266405582428} +01/27/2022 11:52:14 - INFO - codeparrot_training - Step 17029: {'lr': 0.0003884977665045081, 'samples': 3269760, 'steps': 17029, 'loss/train': 1.0174627304077148} +01/27/2022 11:52:17 - INFO - codeparrot_training - Step 17030: {'lr': 0.0003884841440885003, 'samples': 3269952, 'steps': 17030, 'loss/train': 1.2079386413097382} +01/27/2022 11:52:20 - INFO - codeparrot_training - Step 17031: {'lr': 0.0003884705210792703, 'samples': 3270144, 'steps': 17031, 'loss/train': 0.0451526390388608} +01/27/2022 11:52:23 - INFO - codeparrot_training - Step 17032: {'lr': 0.00038845689747687664, 'samples': 3270336, 'steps': 17032, 'loss/train': 0.22059445828199387} +01/27/2022 11:52:26 - INFO - codeparrot_training - Step 17033: {'lr': 0.0003884432732813776, 'samples': 3270528, 'steps': 17033, 'loss/train': 0.7348201721906662} +01/27/2022 11:52:30 - INFO - codeparrot_training - Step 17034: {'lr': 0.00038842964849283146, 'samples': 3270720, 'steps': 17034, 'loss/train': 0.9016899168491364} +01/27/2022 11:52:33 - INFO - codeparrot_training - Step 17035: {'lr': 0.0003884160231112968, 'samples': 3270912, 'steps': 17035, 'loss/train': 0.6028322875499725} +01/27/2022 11:52:39 - INFO - codeparrot_training - Step 17036: {'lr': 0.00038840239713683165, 'samples': 3271104, 'steps': 17036, 'loss/train': 0.7409137040376663} +01/27/2022 11:52:42 - INFO - codeparrot_training - Step 17037: {'lr': 0.00038838877056949475, 'samples': 3271296, 'steps': 17037, 'loss/train': 0.6765693426132202} +01/27/2022 11:52:45 - INFO - codeparrot_training - Step 17038: {'lr': 0.00038837514340934424, 'samples': 3271488, 'steps': 17038, 'loss/train': 0.10178080201148987} +01/27/2022 11:52:48 - INFO - codeparrot_training - Step 17039: {'lr': 0.0003883615156564385, 'samples': 3271680, 'steps': 17039, 'loss/train': 0.5524561554193497} +01/27/2022 11:52:52 - INFO - codeparrot_training - Step 17040: {'lr': 0.000388347887310836, 'samples': 3271872, 'steps': 17040, 'loss/train': 0.6607821732759476} +01/27/2022 11:52:55 - INFO - codeparrot_training - Step 17041: {'lr': 0.0003883342583725952, 'samples': 3272064, 'steps': 17041, 'loss/train': 1.474579095840454} +01/27/2022 11:52:58 - INFO - codeparrot_training - Step 17042: {'lr': 0.0003883206288417742, 'samples': 3272256, 'steps': 17042, 'loss/train': 0.7803474068641663} +01/27/2022 11:53:01 - INFO - codeparrot_training - Step 17043: {'lr': 0.0003883069987184316, 'samples': 3272448, 'steps': 17043, 'loss/train': 0.8445212244987488} +01/27/2022 11:53:06 - INFO - codeparrot_training - Step 17044: {'lr': 0.0003882933680026257, 'samples': 3272640, 'steps': 17044, 'loss/train': 0.728143036365509} +01/27/2022 11:53:09 - INFO - codeparrot_training - Step 17045: {'lr': 0.000388279736694415, 'samples': 3272832, 'steps': 17045, 'loss/train': 0.9062857031822205} +01/27/2022 11:53:12 - INFO - codeparrot_training - Step 17046: {'lr': 0.00038826610479385774, 'samples': 3273024, 'steps': 17046, 'loss/train': 1.0894111096858978} +01/27/2022 11:53:15 - INFO - codeparrot_training - Step 17047: {'lr': 0.00038825247230101244, 'samples': 3273216, 'steps': 17047, 'loss/train': 0.7716476619243622} +01/27/2022 11:53:18 - INFO - codeparrot_training - Step 17048: {'lr': 0.0003882388392159375, 'samples': 3273408, 'steps': 17048, 'loss/train': 0.20042574405670166} +01/27/2022 11:53:21 - INFO - codeparrot_training - Step 17049: {'lr': 0.0003882252055386912, 'samples': 3273600, 'steps': 17049, 'loss/train': 0.42289701104164124} +01/27/2022 11:53:25 - INFO - codeparrot_training - Step 17050: {'lr': 0.00038821157126933204, 'samples': 3273792, 'steps': 17050, 'loss/train': 0.8383811116218567} +01/27/2022 11:53:28 - INFO - codeparrot_training - Step 17051: {'lr': 0.00038819793640791834, 'samples': 3273984, 'steps': 17051, 'loss/train': 0.6655778288841248} +01/27/2022 11:53:31 - INFO - codeparrot_training - Step 17052: {'lr': 0.0003881843009545086, 'samples': 3274176, 'steps': 17052, 'loss/train': 0.8225883543491364} +01/27/2022 11:53:37 - INFO - codeparrot_training - Step 17053: {'lr': 0.0003881706649091612, 'samples': 3274368, 'steps': 17053, 'loss/train': 0.4879691004753113} +01/27/2022 11:53:41 - INFO - codeparrot_training - Step 17054: {'lr': 0.0003881570282719346, 'samples': 3274560, 'steps': 17054, 'loss/train': 0.7611680030822754} +01/27/2022 11:53:44 - INFO - codeparrot_training - Step 17055: {'lr': 0.00038814339104288706, 'samples': 3274752, 'steps': 17055, 'loss/train': 0.8038306832313538} +01/27/2022 11:53:47 - INFO - codeparrot_training - Step 17056: {'lr': 0.00038812975322207713, 'samples': 3274944, 'steps': 17056, 'loss/train': 0.5895051509141922} +01/27/2022 11:53:50 - INFO - codeparrot_training - Step 17057: {'lr': 0.0003881161148095632, 'samples': 3275136, 'steps': 17057, 'loss/train': 0.7737646400928497} +01/27/2022 11:53:53 - INFO - codeparrot_training - Step 17058: {'lr': 0.0003881024758054037, 'samples': 3275328, 'steps': 17058, 'loss/train': 0.858266294002533} +01/27/2022 11:53:56 - INFO - codeparrot_training - Step 17059: {'lr': 0.00038808883620965705, 'samples': 3275520, 'steps': 17059, 'loss/train': 0.3931390643119812} +01/27/2022 11:53:59 - INFO - codeparrot_training - Step 17060: {'lr': 0.00038807519602238174, 'samples': 3275712, 'steps': 17060, 'loss/train': 0.4802197515964508} +01/27/2022 11:54:03 - INFO - codeparrot_training - Step 17061: {'lr': 0.00038806155524363594, 'samples': 3275904, 'steps': 17061, 'loss/train': 0.6924887448549271} +01/27/2022 11:54:07 - INFO - codeparrot_training - Step 17062: {'lr': 0.00038804791387347844, 'samples': 3276096, 'steps': 17062, 'loss/train': 0.8739460408687592} +01/27/2022 11:54:10 - INFO - codeparrot_training - Step 17063: {'lr': 0.0003880342719119675, 'samples': 3276288, 'steps': 17063, 'loss/train': 0.8691548109054565} +01/27/2022 11:54:13 - INFO - codeparrot_training - Step 17064: {'lr': 0.0003880206293591615, 'samples': 3276480, 'steps': 17064, 'loss/train': 0.5927909910678864} +01/27/2022 11:54:16 - INFO - codeparrot_training - Step 17065: {'lr': 0.000388006986215119, 'samples': 3276672, 'steps': 17065, 'loss/train': 0.859227329492569} +01/27/2022 11:54:20 - INFO - codeparrot_training - Step 17066: {'lr': 0.0003879933424798984, 'samples': 3276864, 'steps': 17066, 'loss/train': 0.732380211353302} +01/27/2022 11:54:23 - INFO - codeparrot_training - Step 17067: {'lr': 0.0003879796981535582, 'samples': 3277056, 'steps': 17067, 'loss/train': 0.5548969656229019} +01/27/2022 11:54:26 - INFO - codeparrot_training - Step 17068: {'lr': 0.00038796605323615664, 'samples': 3277248, 'steps': 17068, 'loss/train': 0.3936801105737686} +01/27/2022 11:54:29 - INFO - codeparrot_training - Step 17069: {'lr': 0.00038795240772775244, 'samples': 3277440, 'steps': 17069, 'loss/train': 0.7332470566034317} +01/27/2022 11:54:32 - INFO - codeparrot_training - Step 17070: {'lr': 0.0003879387616284038, 'samples': 3277632, 'steps': 17070, 'loss/train': 1.0897755324840546} +01/27/2022 11:54:37 - INFO - codeparrot_training - Step 17071: {'lr': 0.0003879251149381694, 'samples': 3277824, 'steps': 17071, 'loss/train': 0.9200559854507446} +01/27/2022 11:54:40 - INFO - codeparrot_training - Step 17072: {'lr': 0.0003879114676571076, 'samples': 3278016, 'steps': 17072, 'loss/train': 0.5846855789422989} +01/27/2022 11:54:43 - INFO - codeparrot_training - Step 17073: {'lr': 0.00038789781978527683, 'samples': 3278208, 'steps': 17073, 'loss/train': 1.0526083409786224} +01/27/2022 11:54:46 - INFO - codeparrot_training - Step 17074: {'lr': 0.0003878841713227356, 'samples': 3278400, 'steps': 17074, 'loss/train': 0.8529087603092194} +01/27/2022 11:54:49 - INFO - codeparrot_training - Step 17075: {'lr': 0.00038787052226954235, 'samples': 3278592, 'steps': 17075, 'loss/train': 0.7341953963041306} +01/27/2022 11:54:52 - INFO - codeparrot_training - Step 17076: {'lr': 0.0003878568726257556, 'samples': 3278784, 'steps': 17076, 'loss/train': 0.8961077928543091} +01/27/2022 11:54:55 - INFO - codeparrot_training - Step 17077: {'lr': 0.0003878432223914338, 'samples': 3278976, 'steps': 17077, 'loss/train': 0.7210025489330292} +01/27/2022 11:54:59 - INFO - codeparrot_training - Step 17078: {'lr': 0.00038782957156663535, 'samples': 3279168, 'steps': 17078, 'loss/train': 0.9196067154407501} +01/27/2022 11:55:02 - INFO - codeparrot_training - Step 17079: {'lr': 0.0003878159201514188, 'samples': 3279360, 'steps': 17079, 'loss/train': 0.5812731385231018} +01/27/2022 11:55:08 - INFO - codeparrot_training - Step 17080: {'lr': 0.00038780226814584263, 'samples': 3279552, 'steps': 17080, 'loss/train': 0.7888712882995605} +01/27/2022 11:55:11 - INFO - codeparrot_training - Step 17081: {'lr': 0.00038778861554996524, 'samples': 3279744, 'steps': 17081, 'loss/train': 0.8665232956409454} +01/27/2022 11:55:14 - INFO - codeparrot_training - Step 17082: {'lr': 0.00038777496236384526, 'samples': 3279936, 'steps': 17082, 'loss/train': 0.5821090936660767} +01/27/2022 11:55:17 - INFO - codeparrot_training - Step 17083: {'lr': 0.000387761308587541, 'samples': 3280128, 'steps': 17083, 'loss/train': 0.45525088906288147} +01/27/2022 11:55:21 - INFO - codeparrot_training - Step 17084: {'lr': 0.0003877476542211111, 'samples': 3280320, 'steps': 17084, 'loss/train': 0.28613872826099396} +01/27/2022 11:55:24 - INFO - codeparrot_training - Step 17085: {'lr': 0.00038773399926461395, 'samples': 3280512, 'steps': 17085, 'loss/train': 1.1982513964176178} +01/27/2022 11:55:27 - INFO - codeparrot_training - Step 17086: {'lr': 0.0003877203437181081, 'samples': 3280704, 'steps': 17086, 'loss/train': 0.7807973921298981} +01/27/2022 11:55:30 - INFO - codeparrot_training - Step 17087: {'lr': 0.0003877066875816521, 'samples': 3280896, 'steps': 17087, 'loss/train': 0.6829292625188828} +01/27/2022 11:55:34 - INFO - codeparrot_training - Step 17088: {'lr': 0.00038769303085530425, 'samples': 3281088, 'steps': 17088, 'loss/train': 1.1501461565494537} +01/27/2022 11:55:38 - INFO - codeparrot_training - Step 17089: {'lr': 0.0003876793735391233, 'samples': 3281280, 'steps': 17089, 'loss/train': 0.8927645981311798} +01/27/2022 11:55:41 - INFO - codeparrot_training - Step 17090: {'lr': 0.00038766571563316756, 'samples': 3281472, 'steps': 17090, 'loss/train': 0.6974459141492844} +01/27/2022 11:55:44 - INFO - codeparrot_training - Step 17091: {'lr': 0.00038765205713749563, 'samples': 3281664, 'steps': 17091, 'loss/train': 0.6338396072387695} +01/27/2022 11:55:47 - INFO - codeparrot_training - Step 17092: {'lr': 0.0003876383980521659, 'samples': 3281856, 'steps': 17092, 'loss/train': 1.0689464807510376} +01/27/2022 11:55:50 - INFO - codeparrot_training - Step 17093: {'lr': 0.0003876247383772371, 'samples': 3282048, 'steps': 17093, 'loss/train': 1.0504281520843506} +01/27/2022 11:55:53 - INFO - codeparrot_training - Step 17094: {'lr': 0.00038761107811276756, 'samples': 3282240, 'steps': 17094, 'loss/train': 0.5748596638441086} +01/27/2022 11:55:56 - INFO - codeparrot_training - Step 17095: {'lr': 0.00038759741725881593, 'samples': 3282432, 'steps': 17095, 'loss/train': 2.8273560404777527} +01/27/2022 11:56:00 - INFO - codeparrot_training - Step 17096: {'lr': 0.0003875837558154406, 'samples': 3282624, 'steps': 17096, 'loss/train': 0.8496375381946564} +01/27/2022 11:56:11 - INFO - codeparrot_training - Step 17097: {'lr': 0.00038757009378270014, 'samples': 3282816, 'steps': 17097, 'loss/train': 0.9595415890216827} +01/27/2022 11:56:15 - INFO - codeparrot_training - Step 17098: {'lr': 0.0003875564311606531, 'samples': 3283008, 'steps': 17098, 'loss/train': 0.9611683487892151} +01/27/2022 11:56:18 - INFO - codeparrot_training - Step 17099: {'lr': 0.000387542767949358, 'samples': 3283200, 'steps': 17099, 'loss/train': 0.5808737128973007} +01/27/2022 11:56:21 - INFO - codeparrot_training - Step 17100: {'lr': 0.0003875291041488734, 'samples': 3283392, 'steps': 17100, 'loss/train': 0.9940761923789978} +01/27/2022 11:56:24 - INFO - codeparrot_training - Step 17101: {'lr': 0.00038751543975925766, 'samples': 3283584, 'steps': 17101, 'loss/train': 0.659049466252327} +01/27/2022 11:56:27 - INFO - codeparrot_training - Step 17102: {'lr': 0.00038750177478056956, 'samples': 3283776, 'steps': 17102, 'loss/train': 0.3995637148618698} +01/27/2022 11:56:30 - INFO - codeparrot_training - Step 17103: {'lr': 0.0003874881092128675, 'samples': 3283968, 'steps': 17103, 'loss/train': 1.7050702571868896} +01/27/2022 11:56:34 - INFO - codeparrot_training - Step 17104: {'lr': 0.00038747444305621, 'samples': 3284160, 'steps': 17104, 'loss/train': 1.6842021346092224} +01/27/2022 11:56:37 - INFO - codeparrot_training - Step 17105: {'lr': 0.0003874607763106556, 'samples': 3284352, 'steps': 17105, 'loss/train': 0.5757706314325333} +01/27/2022 11:56:40 - INFO - codeparrot_training - Step 17106: {'lr': 0.00038744710897626293, 'samples': 3284544, 'steps': 17106, 'loss/train': 1.452052116394043} +01/27/2022 11:56:44 - INFO - codeparrot_training - Step 17107: {'lr': 0.00038743344105309055, 'samples': 3284736, 'steps': 17107, 'loss/train': 1.1913694739341736} +01/27/2022 11:56:47 - INFO - codeparrot_training - Step 17108: {'lr': 0.0003874197725411969, 'samples': 3284928, 'steps': 17108, 'loss/train': 0.6733482331037521} +01/27/2022 11:56:50 - INFO - codeparrot_training - Step 17109: {'lr': 0.0003874061034406405, 'samples': 3285120, 'steps': 17109, 'loss/train': 0.9188793003559113} +01/27/2022 11:56:54 - INFO - codeparrot_training - Step 17110: {'lr': 0.00038739243375148, 'samples': 3285312, 'steps': 17110, 'loss/train': 1.4206112623214722} +01/27/2022 11:56:57 - INFO - codeparrot_training - Step 17111: {'lr': 0.0003873787634737741, 'samples': 3285504, 'steps': 17111, 'loss/train': 0.8982879817485809} +01/27/2022 11:57:00 - INFO - codeparrot_training - Step 17112: {'lr': 0.00038736509260758103, 'samples': 3285696, 'steps': 17112, 'loss/train': 0.5065480917692184} +01/27/2022 11:57:03 - INFO - codeparrot_training - Step 17113: {'lr': 0.00038735142115295965, 'samples': 3285888, 'steps': 17113, 'loss/train': 1.072641670703888} +01/27/2022 11:57:06 - INFO - codeparrot_training - Step 17114: {'lr': 0.00038733774910996825, 'samples': 3286080, 'steps': 17114, 'loss/train': 0.877034604549408} +01/27/2022 11:57:09 - INFO - codeparrot_training - Step 17115: {'lr': 0.00038732407647866567, 'samples': 3286272, 'steps': 17115, 'loss/train': 0.8730863928794861} +01/27/2022 11:57:14 - INFO - codeparrot_training - Step 17116: {'lr': 0.00038731040325911027, 'samples': 3286464, 'steps': 17116, 'loss/train': 0.7605755925178528} +01/27/2022 11:57:17 - INFO - codeparrot_training - Step 17117: {'lr': 0.0003872967294513608, 'samples': 3286656, 'steps': 17117, 'loss/train': 0.720842108130455} +01/27/2022 11:57:20 - INFO - codeparrot_training - Step 17118: {'lr': 0.0003872830550554757, 'samples': 3286848, 'steps': 17118, 'loss/train': 0.7333067357540131} +01/27/2022 11:57:23 - INFO - codeparrot_training - Step 17119: {'lr': 0.0003872693800715135, 'samples': 3287040, 'steps': 17119, 'loss/train': 0.7090850472450256} +01/27/2022 11:57:26 - INFO - codeparrot_training - Step 17120: {'lr': 0.00038725570449953296, 'samples': 3287232, 'steps': 17120, 'loss/train': 0.8487252295017242} +01/27/2022 11:57:30 - INFO - codeparrot_training - Step 17121: {'lr': 0.00038724202833959254, 'samples': 3287424, 'steps': 17121, 'loss/train': 1.102437973022461} +01/27/2022 11:57:33 - INFO - codeparrot_training - Step 17122: {'lr': 0.00038722835159175087, 'samples': 3287616, 'steps': 17122, 'loss/train': 0.7238404154777527} +01/27/2022 11:57:36 - INFO - codeparrot_training - Step 17123: {'lr': 0.00038721467425606644, 'samples': 3287808, 'steps': 17123, 'loss/train': 0.7338927090167999} +01/27/2022 11:57:40 - INFO - codeparrot_training - Step 17124: {'lr': 0.000387200996332598, 'samples': 3288000, 'steps': 17124, 'loss/train': 0.5847375243902206} +01/27/2022 11:57:44 - INFO - codeparrot_training - Step 17125: {'lr': 0.000387187317821404, 'samples': 3288192, 'steps': 17125, 'loss/train': 0.43509162962436676} +01/27/2022 11:57:47 - INFO - codeparrot_training - Step 17126: {'lr': 0.0003871736387225431, 'samples': 3288384, 'steps': 17126, 'loss/train': 0.8111625909805298} +01/27/2022 11:57:50 - INFO - codeparrot_training - Step 17127: {'lr': 0.0003871599590360739, 'samples': 3288576, 'steps': 17127, 'loss/train': 0.8504003584384918} +01/27/2022 11:57:53 - INFO - codeparrot_training - Step 17128: {'lr': 0.000387146278762055, 'samples': 3288768, 'steps': 17128, 'loss/train': 0.8934609889984131} +01/27/2022 11:57:56 - INFO - codeparrot_training - Step 17129: {'lr': 0.000387132597900545, 'samples': 3288960, 'steps': 17129, 'loss/train': 1.182650774717331} +01/27/2022 11:57:59 - INFO - codeparrot_training - Step 17130: {'lr': 0.0003871189164516025, 'samples': 3289152, 'steps': 17130, 'loss/train': 0.8245615661144257} +01/27/2022 11:58:03 - INFO - codeparrot_training - Step 17131: {'lr': 0.000387105234415286, 'samples': 3289344, 'steps': 17131, 'loss/train': 0.43334394693374634} +01/27/2022 11:58:06 - INFO - codeparrot_training - Step 17132: {'lr': 0.00038709155179165436, 'samples': 3289536, 'steps': 17132, 'loss/train': 0.9035642445087433} +01/27/2022 11:58:12 - INFO - codeparrot_training - Step 17133: {'lr': 0.000387077868580766, 'samples': 3289728, 'steps': 17133, 'loss/train': 4.857071042060852} +01/27/2022 11:58:15 - INFO - codeparrot_training - Step 17134: {'lr': 0.00038706418478267945, 'samples': 3289920, 'steps': 17134, 'loss/train': 1.1010225713253021} +01/27/2022 11:58:18 - INFO - codeparrot_training - Step 17135: {'lr': 0.0003870505003974536, 'samples': 3290112, 'steps': 17135, 'loss/train': 0.35707008093595505} +01/27/2022 11:58:21 - INFO - codeparrot_training - Step 17136: {'lr': 0.0003870368154251469, 'samples': 3290304, 'steps': 17136, 'loss/train': 0.6170548796653748} +01/27/2022 11:58:25 - INFO - codeparrot_training - Step 17137: {'lr': 0.000387023129865818, 'samples': 3290496, 'steps': 17137, 'loss/train': 1.501773476600647} +01/27/2022 11:58:28 - INFO - codeparrot_training - Step 17138: {'lr': 0.00038700944371952543, 'samples': 3290688, 'steps': 17138, 'loss/train': 0.845130980014801} +01/27/2022 11:58:31 - INFO - codeparrot_training - Step 17139: {'lr': 0.00038699575698632806, 'samples': 3290880, 'steps': 17139, 'loss/train': 0.9448512196540833} +01/27/2022 11:58:34 - INFO - codeparrot_training - Step 17140: {'lr': 0.00038698206966628426, 'samples': 3291072, 'steps': 17140, 'loss/train': 0.7295672446489334} +01/27/2022 11:58:37 - INFO - codeparrot_training - Step 17141: {'lr': 0.00038696838175945284, 'samples': 3291264, 'steps': 17141, 'loss/train': 0.8344963788986206} +01/27/2022 11:58:41 - INFO - codeparrot_training - Step 17142: {'lr': 0.0003869546932658923, 'samples': 3291456, 'steps': 17142, 'loss/train': 0.9255821406841278} +01/27/2022 11:58:45 - INFO - codeparrot_training - Step 17143: {'lr': 0.0003869410041856614, 'samples': 3291648, 'steps': 17143, 'loss/train': 0.6557753831148148} +01/27/2022 11:58:48 - INFO - codeparrot_training - Step 17144: {'lr': 0.0003869273145188186, 'samples': 3291840, 'steps': 17144, 'loss/train': 0.8829418122768402} +01/27/2022 11:58:51 - INFO - codeparrot_training - Step 17145: {'lr': 0.00038691362426542273, 'samples': 3292032, 'steps': 17145, 'loss/train': 0.7441485822200775} +01/27/2022 11:58:54 - INFO - codeparrot_training - Step 17146: {'lr': 0.0003868999334255324, 'samples': 3292224, 'steps': 17146, 'loss/train': 0.7426473051309586} +01/27/2022 11:58:57 - INFO - codeparrot_training - Step 17147: {'lr': 0.00038688624199920623, 'samples': 3292416, 'steps': 17147, 'loss/train': 1.040901392698288} +01/27/2022 11:59:00 - INFO - codeparrot_training - Step 17148: {'lr': 0.0003868725499865029, 'samples': 3292608, 'steps': 17148, 'loss/train': 0.32856787741184235} +01/27/2022 11:59:03 - INFO - codeparrot_training - Step 17149: {'lr': 0.00038685885738748096, 'samples': 3292800, 'steps': 17149, 'loss/train': 0.867857962846756} +01/27/2022 11:59:07 - INFO - codeparrot_training - Step 17150: {'lr': 0.0003868451642021992, 'samples': 3292992, 'steps': 17150, 'loss/train': 1.0771827399730682} +01/27/2022 11:59:12 - INFO - codeparrot_training - Step 17151: {'lr': 0.0003868314704307161, 'samples': 3293184, 'steps': 17151, 'loss/train': 0.4406020939350128} +01/27/2022 11:59:15 - INFO - codeparrot_training - Step 17152: {'lr': 0.0003868177760730905, 'samples': 3293376, 'steps': 17152, 'loss/train': 1.188736081123352} +01/27/2022 11:59:18 - INFO - codeparrot_training - Step 17153: {'lr': 0.00038680408112938097, 'samples': 3293568, 'steps': 17153, 'loss/train': 0.5251313745975494} +01/27/2022 11:59:21 - INFO - codeparrot_training - Step 17154: {'lr': 0.00038679038559964626, 'samples': 3293760, 'steps': 17154, 'loss/train': 1.7475634217262268} +01/27/2022 11:59:24 - INFO - codeparrot_training - Step 17155: {'lr': 0.0003867766894839449, 'samples': 3293952, 'steps': 17155, 'loss/train': 1.6415782570838928} +01/27/2022 11:59:28 - INFO - codeparrot_training - Step 17156: {'lr': 0.0003867629927823357, 'samples': 3294144, 'steps': 17156, 'loss/train': 0.8078372776508331} +01/27/2022 11:59:31 - INFO - codeparrot_training - Step 17157: {'lr': 0.00038674929549487714, 'samples': 3294336, 'steps': 17157, 'loss/train': 1.0802510976791382} +01/27/2022 11:59:34 - INFO - codeparrot_training - Step 17158: {'lr': 0.00038673559762162816, 'samples': 3294528, 'steps': 17158, 'loss/train': 0.475567951798439} +01/27/2022 11:59:37 - INFO - codeparrot_training - Step 17159: {'lr': 0.0003867218991626472, 'samples': 3294720, 'steps': 17159, 'loss/train': 0.90149787068367} +01/27/2022 11:59:43 - INFO - codeparrot_training - Step 17160: {'lr': 0.0003867082001179931, 'samples': 3294912, 'steps': 17160, 'loss/train': 0.987870454788208} +01/27/2022 11:59:47 - INFO - codeparrot_training - Step 17161: {'lr': 0.0003866945004877245, 'samples': 3295104, 'steps': 17161, 'loss/train': 0.8352283537387848} +01/27/2022 11:59:50 - INFO - codeparrot_training - Step 17162: {'lr': 0.0003866808002719, 'samples': 3295296, 'steps': 17162, 'loss/train': 0.9312686026096344} +01/27/2022 11:59:53 - INFO - codeparrot_training - Step 17163: {'lr': 0.00038666709947057836, 'samples': 3295488, 'steps': 17163, 'loss/train': 0.39049476385116577} +01/27/2022 11:59:56 - INFO - codeparrot_training - Step 17164: {'lr': 0.0003866533980838183, 'samples': 3295680, 'steps': 17164, 'loss/train': 1.055180311203003} +01/27/2022 11:59:59 - INFO - codeparrot_training - Step 17165: {'lr': 0.0003866396961116785, 'samples': 3295872, 'steps': 17165, 'loss/train': 0.7023611068725586} +01/27/2022 12:00:02 - INFO - codeparrot_training - Step 17166: {'lr': 0.00038662599355421756, 'samples': 3296064, 'steps': 17166, 'loss/train': 1.2058171033859253} +01/27/2022 12:00:05 - INFO - codeparrot_training - Step 17167: {'lr': 0.00038661229041149427, 'samples': 3296256, 'steps': 17167, 'loss/train': 1.3383604288101196} +01/27/2022 12:00:09 - INFO - codeparrot_training - Step 17168: {'lr': 0.0003865985866835673, 'samples': 3296448, 'steps': 17168, 'loss/train': 0.9030220806598663} +01/27/2022 12:00:13 - INFO - codeparrot_training - Step 17169: {'lr': 0.0003865848823704954, 'samples': 3296640, 'steps': 17169, 'loss/train': 1.2047575414180756} +01/27/2022 12:00:16 - INFO - codeparrot_training - Step 17170: {'lr': 0.00038657117747233717, 'samples': 3296832, 'steps': 17170, 'loss/train': 0.788334310054779} +01/27/2022 12:00:19 - INFO - codeparrot_training - Step 17171: {'lr': 0.00038655747198915137, 'samples': 3297024, 'steps': 17171, 'loss/train': 0.8884812891483307} +01/27/2022 12:00:23 - INFO - codeparrot_training - Step 17172: {'lr': 0.0003865437659209968, 'samples': 3297216, 'steps': 17172, 'loss/train': 0.7638389468193054} +01/27/2022 12:00:26 - INFO - codeparrot_training - Step 17173: {'lr': 0.00038653005926793203, 'samples': 3297408, 'steps': 17173, 'loss/train': 0.7381017208099365} +01/27/2022 12:00:29 - INFO - codeparrot_training - Step 17174: {'lr': 0.0003865163520300159, 'samples': 3297600, 'steps': 17174, 'loss/train': 5.271773099899292} +01/27/2022 12:00:32 - INFO - codeparrot_training - Step 17175: {'lr': 0.00038650264420730707, 'samples': 3297792, 'steps': 17175, 'loss/train': 1.0470133423805237} +01/27/2022 12:00:35 - INFO - codeparrot_training - Step 17176: {'lr': 0.00038648893579986424, 'samples': 3297984, 'steps': 17176, 'loss/train': 0.8803399801254272} +01/27/2022 12:00:38 - INFO - codeparrot_training - Step 17177: {'lr': 0.00038647522680774603, 'samples': 3298176, 'steps': 17177, 'loss/train': 0.9328750669956207} +01/27/2022 12:00:44 - INFO - codeparrot_training - Step 17178: {'lr': 0.0003864615172310115, 'samples': 3298368, 'steps': 17178, 'loss/train': 0.7722419500350952} +01/27/2022 12:00:47 - INFO - codeparrot_training - Step 17179: {'lr': 0.000386447807069719, 'samples': 3298560, 'steps': 17179, 'loss/train': 0.9769330322742462} +01/27/2022 12:00:51 - INFO - codeparrot_training - Step 17180: {'lr': 0.0003864340963239275, 'samples': 3298752, 'steps': 17180, 'loss/train': 0.7703602015972137} +01/27/2022 12:00:54 - INFO - codeparrot_training - Step 17181: {'lr': 0.00038642038499369556, 'samples': 3298944, 'steps': 17181, 'loss/train': 0.6687292903661728} +01/27/2022 12:00:57 - INFO - codeparrot_training - Step 17182: {'lr': 0.0003864066730790821, 'samples': 3299136, 'steps': 17182, 'loss/train': 0.8582193553447723} +01/27/2022 12:01:00 - INFO - codeparrot_training - Step 17183: {'lr': 0.00038639296058014575, 'samples': 3299328, 'steps': 17183, 'loss/train': 2.0792430639266968} +01/27/2022 12:01:03 - INFO - codeparrot_training - Step 17184: {'lr': 0.0003863792474969453, 'samples': 3299520, 'steps': 17184, 'loss/train': 1.1491015255451202} +01/27/2022 12:01:06 - INFO - codeparrot_training - Step 17185: {'lr': 0.00038636553382953944, 'samples': 3299712, 'steps': 17185, 'loss/train': 0.8016312718391418} +01/27/2022 12:01:09 - INFO - codeparrot_training - Step 17186: {'lr': 0.00038635181957798686, 'samples': 3299904, 'steps': 17186, 'loss/train': 0.9545815587043762} +01/27/2022 12:01:14 - INFO - codeparrot_training - Step 17187: {'lr': 0.00038633810474234643, 'samples': 3300096, 'steps': 17187, 'loss/train': 0.7852778434753418} +01/27/2022 12:01:17 - INFO - codeparrot_training - Step 17188: {'lr': 0.00038632438932267686, 'samples': 3300288, 'steps': 17188, 'loss/train': 0.8019550144672394} +01/27/2022 12:01:20 - INFO - codeparrot_training - Step 17189: {'lr': 0.0003863106733190369, 'samples': 3300480, 'steps': 17189, 'loss/train': 0.20018691569566727} +01/27/2022 12:01:23 - INFO - codeparrot_training - Step 17190: {'lr': 0.0003862969567314852, 'samples': 3300672, 'steps': 17190, 'loss/train': 4.099797606468201} +01/27/2022 12:01:26 - INFO - codeparrot_training - Step 17191: {'lr': 0.0003862832395600808, 'samples': 3300864, 'steps': 17191, 'loss/train': 0.73240926861763} +01/27/2022 12:01:30 - INFO - codeparrot_training - Step 17192: {'lr': 0.0003862695218048822, 'samples': 3301056, 'steps': 17192, 'loss/train': 0.8167985379695892} +01/27/2022 12:01:33 - INFO - codeparrot_training - Step 17193: {'lr': 0.00038625580346594824, 'samples': 3301248, 'steps': 17193, 'loss/train': 0.9045749008655548} +01/27/2022 12:01:36 - INFO - codeparrot_training - Step 17194: {'lr': 0.00038624208454333763, 'samples': 3301440, 'steps': 17194, 'loss/train': 1.2144889533519745} +01/27/2022 12:01:39 - INFO - codeparrot_training - Step 17195: {'lr': 0.00038622836503710917, 'samples': 3301632, 'steps': 17195, 'loss/train': 0.7750020325183868} +01/27/2022 12:01:44 - INFO - codeparrot_training - Step 17196: {'lr': 0.00038621464494732174, 'samples': 3301824, 'steps': 17196, 'loss/train': 0.7740617394447327} +01/27/2022 12:01:47 - INFO - codeparrot_training - Step 17197: {'lr': 0.00038620092427403395, 'samples': 3302016, 'steps': 17197, 'loss/train': 0.9700503051280975} +01/27/2022 12:01:50 - INFO - codeparrot_training - Step 17198: {'lr': 0.0003861872030173047, 'samples': 3302208, 'steps': 17198, 'loss/train': 1.0479870736598969} +01/27/2022 12:01:53 - INFO - codeparrot_training - Step 17199: {'lr': 0.0003861734811771928, 'samples': 3302400, 'steps': 17199, 'loss/train': 1.3490813970565796} +01/27/2022 12:01:56 - INFO - codeparrot_training - Step 17200: {'lr': 0.00038615975875375683, 'samples': 3302592, 'steps': 17200, 'loss/train': 0.7893092036247253} +01/27/2022 12:01:59 - INFO - codeparrot_training - Step 17201: {'lr': 0.0003861460357470556, 'samples': 3302784, 'steps': 17201, 'loss/train': 0.7958753407001495} +01/27/2022 12:02:03 - INFO - codeparrot_training - Step 17202: {'lr': 0.0003861323121571482, 'samples': 3302976, 'steps': 17202, 'loss/train': 0.8382782936096191} +01/27/2022 12:02:06 - INFO - codeparrot_training - Step 17203: {'lr': 0.0003861185879840931, 'samples': 3303168, 'steps': 17203, 'loss/train': 0.6341733634471893} +01/27/2022 12:02:12 - INFO - codeparrot_training - Step 17204: {'lr': 0.00038610486322794915, 'samples': 3303360, 'steps': 17204, 'loss/train': 1.0078166127204895} +01/27/2022 12:02:15 - INFO - codeparrot_training - Step 17205: {'lr': 0.0003860911378887752, 'samples': 3303552, 'steps': 17205, 'loss/train': 0.9842389225959778} +01/27/2022 12:02:18 - INFO - codeparrot_training - Step 17206: {'lr': 0.00038607741196663005, 'samples': 3303744, 'steps': 17206, 'loss/train': 0.3809104710817337} +01/27/2022 12:02:22 - INFO - codeparrot_training - Step 17207: {'lr': 0.0003860636854615725, 'samples': 3303936, 'steps': 17207, 'loss/train': 0.9755906760692596} +01/27/2022 12:02:25 - INFO - codeparrot_training - Step 17208: {'lr': 0.0003860499583736613, 'samples': 3304128, 'steps': 17208, 'loss/train': 1.0893445014953613} +01/27/2022 12:02:28 - INFO - codeparrot_training - Step 17209: {'lr': 0.00038603623070295536, 'samples': 3304320, 'steps': 17209, 'loss/train': 1.2445896863937378} +01/27/2022 12:02:31 - INFO - codeparrot_training - Step 17210: {'lr': 0.0003860225024495133, 'samples': 3304512, 'steps': 17210, 'loss/train': 0.7934837937355042} +01/27/2022 12:02:34 - INFO - codeparrot_training - Step 17211: {'lr': 0.000386008773613394, 'samples': 3304704, 'steps': 17211, 'loss/train': 0.9850592315196991} +01/27/2022 12:02:37 - INFO - codeparrot_training - Step 17212: {'lr': 0.0003859950441946564, 'samples': 3304896, 'steps': 17212, 'loss/train': 0.8522422313690186} +01/27/2022 12:02:42 - INFO - codeparrot_training - Step 17213: {'lr': 0.0003859813141933592, 'samples': 3305088, 'steps': 17213, 'loss/train': 1.0689623951911926} +01/27/2022 12:02:45 - INFO - codeparrot_training - Step 17214: {'lr': 0.0003859675836095612, 'samples': 3305280, 'steps': 17214, 'loss/train': 0.9115203022956848} +01/27/2022 12:02:48 - INFO - codeparrot_training - Step 17215: {'lr': 0.00038595385244332125, 'samples': 3305472, 'steps': 17215, 'loss/train': 0.9020907282829285} +01/27/2022 12:02:51 - INFO - codeparrot_training - Step 17216: {'lr': 0.00038594012069469814, 'samples': 3305664, 'steps': 17216, 'loss/train': 1.027454674243927} +01/27/2022 12:02:54 - INFO - codeparrot_training - Step 17217: {'lr': 0.00038592638836375075, 'samples': 3305856, 'steps': 17217, 'loss/train': 0.9640725553035736} +01/27/2022 12:02:58 - INFO - codeparrot_training - Step 17218: {'lr': 0.0003859126554505379, 'samples': 3306048, 'steps': 17218, 'loss/train': 1.1395648419857025} +01/27/2022 12:03:01 - INFO - codeparrot_training - Step 17219: {'lr': 0.00038589892195511834, 'samples': 3306240, 'steps': 17219, 'loss/train': 1.263746201992035} +01/27/2022 12:03:04 - INFO - codeparrot_training - Step 17220: {'lr': 0.00038588518787755096, 'samples': 3306432, 'steps': 17220, 'loss/train': 0.7372184693813324} +01/27/2022 12:03:07 - INFO - codeparrot_training - Step 17221: {'lr': 0.00038587145321789456, 'samples': 3306624, 'steps': 17221, 'loss/train': 0.6328811198472977} +01/27/2022 12:03:11 - INFO - codeparrot_training - Step 17222: {'lr': 0.00038585771797620803, 'samples': 3306816, 'steps': 17222, 'loss/train': 1.1549644768238068} +01/27/2022 12:03:14 - INFO - codeparrot_training - Step 17223: {'lr': 0.00038584398215255023, 'samples': 3307008, 'steps': 17223, 'loss/train': 1.0643807351589203} +01/27/2022 12:03:18 - INFO - codeparrot_training - Step 17224: {'lr': 0.0003858302457469799, 'samples': 3307200, 'steps': 17224, 'loss/train': 0.5199750512838364} +01/27/2022 12:03:21 - INFO - codeparrot_training - Step 17225: {'lr': 0.0003858165087595559, 'samples': 3307392, 'steps': 17225, 'loss/train': 0.8222046196460724} +01/27/2022 12:03:24 - INFO - codeparrot_training - Step 17226: {'lr': 0.00038580277119033715, 'samples': 3307584, 'steps': 17226, 'loss/train': 1.2518862783908844} +01/27/2022 12:03:27 - INFO - codeparrot_training - Step 17227: {'lr': 0.0003857890330393824, 'samples': 3307776, 'steps': 17227, 'loss/train': 0.21527986228466034} +01/27/2022 12:03:30 - INFO - codeparrot_training - Step 17228: {'lr': 0.0003857752943067506, 'samples': 3307968, 'steps': 17228, 'loss/train': 0.5521108657121658} +01/27/2022 12:03:33 - INFO - codeparrot_training - Step 17229: {'lr': 0.00038576155499250056, 'samples': 3308160, 'steps': 17229, 'loss/train': 0.8426592350006104} +01/27/2022 12:03:37 - INFO - codeparrot_training - Step 17230: {'lr': 0.000385747815096691, 'samples': 3308352, 'steps': 17230, 'loss/train': 0.13608907163143158} +01/27/2022 12:03:41 - INFO - codeparrot_training - Step 17231: {'lr': 0.00038573407461938103, 'samples': 3308544, 'steps': 17231, 'loss/train': 1.2773436605930328} +01/27/2022 12:03:44 - INFO - codeparrot_training - Step 17232: {'lr': 0.0003857203335606294, 'samples': 3308736, 'steps': 17232, 'loss/train': 0.9917554557323456} +01/27/2022 12:03:47 - INFO - codeparrot_training - Step 17233: {'lr': 0.00038570659192049497, 'samples': 3308928, 'steps': 17233, 'loss/train': 1.2190384268760681} +01/27/2022 12:03:50 - INFO - codeparrot_training - Step 17234: {'lr': 0.0003856928496990364, 'samples': 3309120, 'steps': 17234, 'loss/train': 0.95966836810112} +01/27/2022 12:03:54 - INFO - codeparrot_training - Step 17235: {'lr': 0.000385679106896313, 'samples': 3309312, 'steps': 17235, 'loss/train': 1.0819309651851654} +01/27/2022 12:03:57 - INFO - codeparrot_training - Step 17236: {'lr': 0.0003856653635123832, 'samples': 3309504, 'steps': 17236, 'loss/train': 1.0137571692466736} +01/27/2022 12:04:00 - INFO - codeparrot_training - Step 17237: {'lr': 0.0003856516195473062, 'samples': 3309696, 'steps': 17237, 'loss/train': 0.754884660243988} +01/27/2022 12:04:03 - INFO - codeparrot_training - Step 17238: {'lr': 0.0003856378750011407, 'samples': 3309888, 'steps': 17238, 'loss/train': 0.8089282214641571} +01/27/2022 12:04:06 - INFO - codeparrot_training - Step 17239: {'lr': 0.0003856241298739456, 'samples': 3310080, 'steps': 17239, 'loss/train': 0.25753019750118256} +01/27/2022 12:04:12 - INFO - codeparrot_training - Step 17240: {'lr': 0.0003856103841657797, 'samples': 3310272, 'steps': 17240, 'loss/train': 0.8107045590877533} +01/27/2022 12:04:15 - INFO - codeparrot_training - Step 17241: {'lr': 0.0003855966378767021, 'samples': 3310464, 'steps': 17241, 'loss/train': 1.0111234188079834} +01/27/2022 12:04:19 - INFO - codeparrot_training - Step 17242: {'lr': 0.00038558289100677144, 'samples': 3310656, 'steps': 17242, 'loss/train': 0.8398600816726685} +01/27/2022 12:04:22 - INFO - codeparrot_training - Step 17243: {'lr': 0.00038556914355604676, 'samples': 3310848, 'steps': 17243, 'loss/train': 0.49779273569583893} +01/27/2022 12:04:25 - INFO - codeparrot_training - Step 17244: {'lr': 0.0003855553955245871, 'samples': 3311040, 'steps': 17244, 'loss/train': 0.7260745167732239} +01/27/2022 12:04:28 - INFO - codeparrot_training - Step 17245: {'lr': 0.00038554164691245095, 'samples': 3311232, 'steps': 17245, 'loss/train': 0.6497770249843597} +01/27/2022 12:04:31 - INFO - codeparrot_training - Step 17246: {'lr': 0.00038552789771969755, 'samples': 3311424, 'steps': 17246, 'loss/train': 0.8452306687831879} +01/27/2022 12:04:34 - INFO - codeparrot_training - Step 17247: {'lr': 0.00038551414794638555, 'samples': 3311616, 'steps': 17247, 'loss/train': 0.8565156161785126} +01/27/2022 12:04:39 - INFO - codeparrot_training - Step 17248: {'lr': 0.00038550039759257404, 'samples': 3311808, 'steps': 17248, 'loss/train': 0.6468378603458405} +01/27/2022 12:04:42 - INFO - codeparrot_training - Step 17249: {'lr': 0.0003854866466583219, 'samples': 3312000, 'steps': 17249, 'loss/train': 0.6167835742235184} +01/27/2022 12:04:45 - INFO - codeparrot_training - Step 17250: {'lr': 0.00038547289514368795, 'samples': 3312192, 'steps': 17250, 'loss/train': 1.1010775566101074} +01/27/2022 12:04:48 - INFO - codeparrot_training - Step 17251: {'lr': 0.00038545914304873117, 'samples': 3312384, 'steps': 17251, 'loss/train': 0.5249023139476776} +01/27/2022 12:04:51 - INFO - codeparrot_training - Step 17252: {'lr': 0.00038544539037351037, 'samples': 3312576, 'steps': 17252, 'loss/train': 0.491867333650589} +01/27/2022 12:04:54 - INFO - codeparrot_training - Step 17253: {'lr': 0.00038543163711808457, 'samples': 3312768, 'steps': 17253, 'loss/train': 0.4940957576036453} +01/27/2022 12:04:58 - INFO - codeparrot_training - Step 17254: {'lr': 0.0003854178832825126, 'samples': 3312960, 'steps': 17254, 'loss/train': 1.127654492855072} +01/27/2022 12:05:01 - INFO - codeparrot_training - Step 17255: {'lr': 0.0003854041288668534, 'samples': 3313152, 'steps': 17255, 'loss/train': 0.6838559210300446} +01/27/2022 12:05:04 - INFO - codeparrot_training - Step 17256: {'lr': 0.00038539037387116595, 'samples': 3313344, 'steps': 17256, 'loss/train': 0.8344090282917023} +01/27/2022 12:05:08 - INFO - codeparrot_training - Step 17257: {'lr': 0.0003853766182955092, 'samples': 3313536, 'steps': 17257, 'loss/train': 0.8003952205181122} +01/27/2022 12:05:11 - INFO - codeparrot_training - Step 17258: {'lr': 0.0003853628621399419, 'samples': 3313728, 'steps': 17258, 'loss/train': 0.4011252522468567} +01/27/2022 12:05:14 - INFO - codeparrot_training - Step 17259: {'lr': 0.00038534910540452305, 'samples': 3313920, 'steps': 17259, 'loss/train': 0.9199228584766388} +01/27/2022 12:05:18 - INFO - codeparrot_training - Step 17260: {'lr': 0.0003853353480893117, 'samples': 3314112, 'steps': 17260, 'loss/train': 1.2258745729923248} +01/27/2022 12:05:21 - INFO - codeparrot_training - Step 17261: {'lr': 0.0003853215901943667, 'samples': 3314304, 'steps': 17261, 'loss/train': 0.4479082077741623} +01/27/2022 12:05:24 - INFO - codeparrot_training - Step 17262: {'lr': 0.00038530783171974694, 'samples': 3314496, 'steps': 17262, 'loss/train': 1.0066815912723541} +01/27/2022 12:05:27 - INFO - codeparrot_training - Step 17263: {'lr': 0.0003852940726655114, 'samples': 3314688, 'steps': 17263, 'loss/train': 0.7901995182037354} +01/27/2022 12:05:30 - INFO - codeparrot_training - Step 17264: {'lr': 0.000385280313031719, 'samples': 3314880, 'steps': 17264, 'loss/train': 0.5094643235206604} +01/27/2022 12:05:33 - INFO - codeparrot_training - Step 17265: {'lr': 0.0003852665528184287, 'samples': 3315072, 'steps': 17265, 'loss/train': 0.9375911056995392} +01/27/2022 12:05:40 - INFO - codeparrot_training - Step 17266: {'lr': 0.0003852527920256994, 'samples': 3315264, 'steps': 17266, 'loss/train': 0.8437194228172302} +01/27/2022 12:05:43 - INFO - codeparrot_training - Step 17267: {'lr': 0.00038523903065359013, 'samples': 3315456, 'steps': 17267, 'loss/train': 0.5410618185997009} +01/27/2022 12:05:46 - INFO - codeparrot_training - Step 17268: {'lr': 0.0003852252687021598, 'samples': 3315648, 'steps': 17268, 'loss/train': 0.4964694678783417} +01/27/2022 12:05:49 - INFO - codeparrot_training - Step 17269: {'lr': 0.00038521150617146737, 'samples': 3315840, 'steps': 17269, 'loss/train': 0.6763711273670197} +01/27/2022 12:05:52 - INFO - codeparrot_training - Step 17270: {'lr': 0.00038519774306157174, 'samples': 3316032, 'steps': 17270, 'loss/train': 0.8893589079380035} +01/27/2022 12:05:55 - INFO - codeparrot_training - Step 17271: {'lr': 0.00038518397937253195, 'samples': 3316224, 'steps': 17271, 'loss/train': 0.7312084883451462} +01/27/2022 12:05:58 - INFO - codeparrot_training - Step 17272: {'lr': 0.00038517021510440694, 'samples': 3316416, 'steps': 17272, 'loss/train': 0.7250602394342422} +01/27/2022 12:06:02 - INFO - codeparrot_training - Step 17273: {'lr': 0.0003851564502572556, 'samples': 3316608, 'steps': 17273, 'loss/train': 1.248950868844986} +01/27/2022 12:06:06 - INFO - codeparrot_training - Step 17274: {'lr': 0.00038514268483113694, 'samples': 3316800, 'steps': 17274, 'loss/train': 0.7562805712223053} +01/27/2022 12:06:09 - INFO - codeparrot_training - Step 17275: {'lr': 0.00038512891882610997, 'samples': 3316992, 'steps': 17275, 'loss/train': 5.618906021118164} +01/27/2022 12:06:12 - INFO - codeparrot_training - Step 17276: {'lr': 0.0003851151522422336, 'samples': 3317184, 'steps': 17276, 'loss/train': 0.604311615228653} +01/27/2022 12:06:16 - INFO - codeparrot_training - Step 17277: {'lr': 0.0003851013850795668, 'samples': 3317376, 'steps': 17277, 'loss/train': 0.821129322052002} +01/27/2022 12:06:19 - INFO - codeparrot_training - Step 17278: {'lr': 0.00038508761733816864, 'samples': 3317568, 'steps': 17278, 'loss/train': 1.0323987901210785} +01/27/2022 12:06:22 - INFO - codeparrot_training - Step 17279: {'lr': 0.00038507384901809795, 'samples': 3317760, 'steps': 17279, 'loss/train': 0.8463170528411865} +01/27/2022 12:06:25 - INFO - codeparrot_training - Step 17280: {'lr': 0.00038506008011941376, 'samples': 3317952, 'steps': 17280, 'loss/train': 1.1559272110462189} +01/27/2022 12:06:28 - INFO - codeparrot_training - Step 17281: {'lr': 0.0003850463106421751, 'samples': 3318144, 'steps': 17281, 'loss/train': 0.8020239472389221} +01/27/2022 12:06:31 - INFO - codeparrot_training - Step 17282: {'lr': 0.000385032540586441, 'samples': 3318336, 'steps': 17282, 'loss/train': 1.0159482657909393} +01/27/2022 12:06:38 - INFO - codeparrot_training - Step 17283: {'lr': 0.00038501876995227023, 'samples': 3318528, 'steps': 17283, 'loss/train': 0.28964464366436005} +01/27/2022 12:06:41 - INFO - codeparrot_training - Step 17284: {'lr': 0.00038500499873972204, 'samples': 3318720, 'steps': 17284, 'loss/train': 0.4893749803304672} +01/27/2022 12:06:44 - INFO - codeparrot_training - Step 17285: {'lr': 0.0003849912269488552, 'samples': 3318912, 'steps': 17285, 'loss/train': 0.8351521790027618} +01/27/2022 12:06:47 - INFO - codeparrot_training - Step 17286: {'lr': 0.000384977454579729, 'samples': 3319104, 'steps': 17286, 'loss/train': 0.9028458595275879} +01/27/2022 12:06:50 - INFO - codeparrot_training - Step 17287: {'lr': 0.00038496368163240215, 'samples': 3319296, 'steps': 17287, 'loss/train': 0.7461619824171066} +01/27/2022 12:06:53 - INFO - codeparrot_training - Step 17288: {'lr': 0.00038494990810693366, 'samples': 3319488, 'steps': 17288, 'loss/train': 0.8105019629001617} +01/27/2022 12:06:57 - INFO - codeparrot_training - Step 17289: {'lr': 0.00038493613400338267, 'samples': 3319680, 'steps': 17289, 'loss/train': 0.9958089888095856} +01/27/2022 12:07:00 - INFO - codeparrot_training - Step 17290: {'lr': 0.0003849223593218082, 'samples': 3319872, 'steps': 17290, 'loss/train': 0.7786571681499481} +01/27/2022 12:07:03 - INFO - codeparrot_training - Step 17291: {'lr': 0.00038490858406226903, 'samples': 3320064, 'steps': 17291, 'loss/train': 1.0172542333602905} +01/27/2022 12:07:07 - INFO - codeparrot_training - Step 17292: {'lr': 0.00038489480822482446, 'samples': 3320256, 'steps': 17292, 'loss/train': 0.5405485779047012} +01/27/2022 12:07:10 - INFO - codeparrot_training - Step 17293: {'lr': 0.00038488103180953326, 'samples': 3320448, 'steps': 17293, 'loss/train': 0.4422297030687332} +01/27/2022 12:07:14 - INFO - codeparrot_training - Step 17294: {'lr': 0.00038486725481645467, 'samples': 3320640, 'steps': 17294, 'loss/train': 1.043163388967514} +01/27/2022 12:07:17 - INFO - codeparrot_training - Step 17295: {'lr': 0.00038485347724564746, 'samples': 3320832, 'steps': 17295, 'loss/train': 0.8734130859375} +01/27/2022 12:07:20 - INFO - codeparrot_training - Step 17296: {'lr': 0.0003848396990971709, 'samples': 3321024, 'steps': 17296, 'loss/train': 0.9932441711425781} +01/27/2022 12:07:23 - INFO - codeparrot_training - Step 17297: {'lr': 0.00038482592037108375, 'samples': 3321216, 'steps': 17297, 'loss/train': 0.7481312602758408} +01/27/2022 12:07:26 - INFO - codeparrot_training - Step 17298: {'lr': 0.0003848121410674453, 'samples': 3321408, 'steps': 17298, 'loss/train': 1.0444898307323456} +01/27/2022 12:07:29 - INFO - codeparrot_training - Step 17299: {'lr': 0.0003847983611863144, 'samples': 3321600, 'steps': 17299, 'loss/train': 0.36014796048402786} +01/27/2022 12:07:32 - INFO - codeparrot_training - Step 17300: {'lr': 0.0003847845807277501, 'samples': 3321792, 'steps': 17300, 'loss/train': 0.9920140206813812} +01/27/2022 12:07:37 - INFO - codeparrot_training - Step 17301: {'lr': 0.00038477079969181146, 'samples': 3321984, 'steps': 17301, 'loss/train': 0.7235189527273178} +01/27/2022 12:07:40 - INFO - codeparrot_training - Step 17302: {'lr': 0.00038475701807855753, 'samples': 3322176, 'steps': 17302, 'loss/train': 1.3511719107627869} +01/27/2022 12:07:43 - INFO - codeparrot_training - Step 17303: {'lr': 0.00038474323588804727, 'samples': 3322368, 'steps': 17303, 'loss/train': 0.7229753583669662} +01/27/2022 12:07:46 - INFO - codeparrot_training - Step 17304: {'lr': 0.0003847294531203398, 'samples': 3322560, 'steps': 17304, 'loss/train': 0.42376287281513214} +01/27/2022 12:07:49 - INFO - codeparrot_training - Step 17305: {'lr': 0.0003847156697754942, 'samples': 3322752, 'steps': 17305, 'loss/train': 0.7098683416843414} +01/27/2022 12:07:53 - INFO - codeparrot_training - Step 17306: {'lr': 0.00038470188585356936, 'samples': 3322944, 'steps': 17306, 'loss/train': 0.8706230521202087} +01/27/2022 12:07:56 - INFO - codeparrot_training - Step 17307: {'lr': 0.00038468810135462445, 'samples': 3323136, 'steps': 17307, 'loss/train': 0.9052166640758514} +01/27/2022 12:07:59 - INFO - codeparrot_training - Step 17308: {'lr': 0.00038467431627871844, 'samples': 3323328, 'steps': 17308, 'loss/train': 0.8047813475131989} +01/27/2022 12:08:02 - INFO - codeparrot_training - Step 17309: {'lr': 0.0003846605306259105, 'samples': 3323520, 'steps': 17309, 'loss/train': 0.6914016902446747} +01/27/2022 12:08:07 - INFO - codeparrot_training - Step 17310: {'lr': 0.0003846467443962596, 'samples': 3323712, 'steps': 17310, 'loss/train': 0.8374458253383636} +01/27/2022 12:08:11 - INFO - codeparrot_training - Step 17311: {'lr': 0.00038463295758982475, 'samples': 3323904, 'steps': 17311, 'loss/train': 1.0438011288642883} +01/27/2022 12:08:14 - INFO - codeparrot_training - Step 17312: {'lr': 0.00038461917020666506, 'samples': 3324096, 'steps': 17312, 'loss/train': 0.42586031556129456} +01/27/2022 12:08:17 - INFO - codeparrot_training - Step 17313: {'lr': 0.0003846053822468396, 'samples': 3324288, 'steps': 17313, 'loss/train': 1.1752041578292847} +01/27/2022 12:08:20 - INFO - codeparrot_training - Step 17314: {'lr': 0.00038459159371040743, 'samples': 3324480, 'steps': 17314, 'loss/train': 0.5639998465776443} +01/27/2022 12:08:23 - INFO - codeparrot_training - Step 17315: {'lr': 0.0003845778045974276, 'samples': 3324672, 'steps': 17315, 'loss/train': 0.9631868004798889} +01/27/2022 12:08:26 - INFO - codeparrot_training - Step 17316: {'lr': 0.0003845640149079592, 'samples': 3324864, 'steps': 17316, 'loss/train': 0.9095828533172607} +01/27/2022 12:08:29 - INFO - codeparrot_training - Step 17317: {'lr': 0.0003845502246420613, 'samples': 3325056, 'steps': 17317, 'loss/train': 0.9990110993385315} +01/27/2022 12:08:34 - INFO - codeparrot_training - Step 17318: {'lr': 0.00038453643379979295, 'samples': 3325248, 'steps': 17318, 'loss/train': 0.08117925748229027} +01/27/2022 12:08:37 - INFO - codeparrot_training - Step 17319: {'lr': 0.00038452264238121326, 'samples': 3325440, 'steps': 17319, 'loss/train': 1.002740889787674} +01/27/2022 12:08:40 - INFO - codeparrot_training - Step 17320: {'lr': 0.0003845088503863813, 'samples': 3325632, 'steps': 17320, 'loss/train': 0.734030082821846} +01/27/2022 12:08:43 - INFO - codeparrot_training - Step 17321: {'lr': 0.0003844950578153561, 'samples': 3325824, 'steps': 17321, 'loss/train': 1.7182039618492126} +01/27/2022 12:08:46 - INFO - codeparrot_training - Step 17322: {'lr': 0.00038448126466819675, 'samples': 3326016, 'steps': 17322, 'loss/train': 0.9405626356601715} +01/27/2022 12:08:50 - INFO - codeparrot_training - Step 17323: {'lr': 0.00038446747094496243, 'samples': 3326208, 'steps': 17323, 'loss/train': 0.6842790842056274} +01/27/2022 12:08:53 - INFO - codeparrot_training - Step 17324: {'lr': 0.00038445367664571216, 'samples': 3326400, 'steps': 17324, 'loss/train': 0.4445958137512207} +01/27/2022 12:08:56 - INFO - codeparrot_training - Step 17325: {'lr': 0.000384439881770505, 'samples': 3326592, 'steps': 17325, 'loss/train': 0.7818026840686798} +01/27/2022 12:08:59 - INFO - codeparrot_training - Step 17326: {'lr': 0.0003844260863194001, 'samples': 3326784, 'steps': 17326, 'loss/train': 0.3377355560660362} +01/27/2022 12:09:03 - INFO - codeparrot_training - Step 17327: {'lr': 0.0003844122902924565, 'samples': 3326976, 'steps': 17327, 'loss/train': 1.2799673974514008} +01/27/2022 12:09:07 - INFO - codeparrot_training - Step 17328: {'lr': 0.00038439849368973334, 'samples': 3327168, 'steps': 17328, 'loss/train': 0.967228889465332} +01/27/2022 12:09:10 - INFO - codeparrot_training - Step 17329: {'lr': 0.0003843846965112897, 'samples': 3327360, 'steps': 17329, 'loss/train': 0.611377939581871} +01/27/2022 12:09:13 - INFO - codeparrot_training - Step 17330: {'lr': 0.0003843708987571847, 'samples': 3327552, 'steps': 17330, 'loss/train': 1.3449221849441528} +01/27/2022 12:09:16 - INFO - codeparrot_training - Step 17331: {'lr': 0.0003843571004274775, 'samples': 3327744, 'steps': 17331, 'loss/train': 0.6128996014595032} +01/27/2022 12:09:19 - INFO - codeparrot_training - Step 17332: {'lr': 0.0003843433015222271, 'samples': 3327936, 'steps': 17332, 'loss/train': 0.9001774191856384} +01/27/2022 12:09:22 - INFO - codeparrot_training - Step 17333: {'lr': 0.0003843295020414926, 'samples': 3328128, 'steps': 17333, 'loss/train': 1.0729883015155792} +01/27/2022 12:09:26 - INFO - codeparrot_training - Step 17334: {'lr': 0.0003843157019853332, 'samples': 3328320, 'steps': 17334, 'loss/train': 0.8696693480014801} +01/27/2022 12:09:29 - INFO - codeparrot_training - Step 17335: {'lr': 0.00038430190135380803, 'samples': 3328512, 'steps': 17335, 'loss/train': 1.391870230436325} +01/27/2022 12:09:33 - INFO - codeparrot_training - Step 17336: {'lr': 0.00038428810014697615, 'samples': 3328704, 'steps': 17336, 'loss/train': 1.1846664547920227} +01/27/2022 12:09:36 - INFO - codeparrot_training - Step 17337: {'lr': 0.00038427429836489663, 'samples': 3328896, 'steps': 17337, 'loss/train': 1.3876873254776} +01/27/2022 12:09:39 - INFO - codeparrot_training - Step 17338: {'lr': 0.00038426049600762867, 'samples': 3329088, 'steps': 17338, 'loss/train': 1.1478002965450287} +01/27/2022 12:09:43 - INFO - codeparrot_training - Step 17339: {'lr': 0.00038424669307523135, 'samples': 3329280, 'steps': 17339, 'loss/train': 0.34568506479263306} +01/27/2022 12:09:46 - INFO - codeparrot_training - Step 17340: {'lr': 0.00038423288956776394, 'samples': 3329472, 'steps': 17340, 'loss/train': 0.7177906483411789} +01/27/2022 12:09:49 - INFO - codeparrot_training - Step 17341: {'lr': 0.00038421908548528534, 'samples': 3329664, 'steps': 17341, 'loss/train': 0.7663499414920807} +01/27/2022 12:09:52 - INFO - codeparrot_training - Step 17342: {'lr': 0.0003842052808278549, 'samples': 3329856, 'steps': 17342, 'loss/train': 0.02663511224091053} +01/27/2022 12:09:55 - INFO - codeparrot_training - Step 17343: {'lr': 0.0003841914755955315, 'samples': 3330048, 'steps': 17343, 'loss/train': 0.9332700669765472} +01/27/2022 12:09:58 - INFO - codeparrot_training - Step 17344: {'lr': 0.00038417766978837453, 'samples': 3330240, 'steps': 17344, 'loss/train': 1.1043168604373932} +01/27/2022 12:10:04 - INFO - codeparrot_training - Step 17345: {'lr': 0.00038416386340644305, 'samples': 3330432, 'steps': 17345, 'loss/train': 0.8539256751537323} +01/27/2022 12:10:08 - INFO - codeparrot_training - Step 17346: {'lr': 0.00038415005644979616, 'samples': 3330624, 'steps': 17346, 'loss/train': 0.8399129211902618} +01/27/2022 12:10:11 - INFO - codeparrot_training - Step 17347: {'lr': 0.00038413624891849295, 'samples': 3330816, 'steps': 17347, 'loss/train': 1.2483892142772675} +01/27/2022 12:10:14 - INFO - codeparrot_training - Step 17348: {'lr': 0.00038412244081259273, 'samples': 3331008, 'steps': 17348, 'loss/train': 0.7914360165596008} +01/27/2022 12:10:17 - INFO - codeparrot_training - Step 17349: {'lr': 0.00038410863213215454, 'samples': 3331200, 'steps': 17349, 'loss/train': 0.9063577651977539} +01/27/2022 12:10:20 - INFO - codeparrot_training - Step 17350: {'lr': 0.0003840948228772376, 'samples': 3331392, 'steps': 17350, 'loss/train': 0.9410102069377899} +01/27/2022 12:10:23 - INFO - codeparrot_training - Step 17351: {'lr': 0.00038408101304790096, 'samples': 3331584, 'steps': 17351, 'loss/train': 1.0353861451148987} +01/27/2022 12:10:27 - INFO - codeparrot_training - Step 17352: {'lr': 0.0003840672026442038, 'samples': 3331776, 'steps': 17352, 'loss/train': 1.4229017794132233} +01/27/2022 12:10:30 - INFO - codeparrot_training - Step 17353: {'lr': 0.0003840533916662054, 'samples': 3331968, 'steps': 17353, 'loss/train': 0.45604442059993744} +01/27/2022 12:10:34 - INFO - codeparrot_training - Step 17354: {'lr': 0.00038403958011396476, 'samples': 3332160, 'steps': 17354, 'loss/train': 0.7084895968437195} +01/27/2022 12:10:37 - INFO - codeparrot_training - Step 17355: {'lr': 0.0003840257679875412, 'samples': 3332352, 'steps': 17355, 'loss/train': 0.7983041703701019} +01/27/2022 12:10:40 - INFO - codeparrot_training - Step 17356: {'lr': 0.00038401195528699374, 'samples': 3332544, 'steps': 17356, 'loss/train': 0.577332079410553} +01/27/2022 12:10:44 - INFO - codeparrot_training - Step 17357: {'lr': 0.0003839981420123817, 'samples': 3332736, 'steps': 17357, 'loss/train': 1.1558298468589783} +01/27/2022 12:10:47 - INFO - codeparrot_training - Step 17358: {'lr': 0.00038398432816376404, 'samples': 3332928, 'steps': 17358, 'loss/train': 0.12298128381371498} +01/27/2022 12:10:50 - INFO - codeparrot_training - Step 17359: {'lr': 0.00038397051374120016, 'samples': 3333120, 'steps': 17359, 'loss/train': 1.0058072805404663} +01/27/2022 12:10:53 - INFO - codeparrot_training - Step 17360: {'lr': 0.00038395669874474915, 'samples': 3333312, 'steps': 17360, 'loss/train': 1.0320155024528503} +01/27/2022 12:10:56 - INFO - codeparrot_training - Step 17361: {'lr': 0.0003839428831744702, 'samples': 3333504, 'steps': 17361, 'loss/train': 0.7938851416110992} +01/27/2022 12:11:02 - INFO - codeparrot_training - Step 17362: {'lr': 0.0003839290670304224, 'samples': 3333696, 'steps': 17362, 'loss/train': 0.9845447838306427} +01/27/2022 12:11:05 - INFO - codeparrot_training - Step 17363: {'lr': 0.00038391525031266494, 'samples': 3333888, 'steps': 17363, 'loss/train': 0.7631959319114685} +01/27/2022 12:11:08 - INFO - codeparrot_training - Step 17364: {'lr': 0.0003839014330212572, 'samples': 3334080, 'steps': 17364, 'loss/train': 0.35819772630929947} +01/27/2022 12:11:12 - INFO - codeparrot_training - Step 17365: {'lr': 0.00038388761515625815, 'samples': 3334272, 'steps': 17365, 'loss/train': 0.6109247803688049} +01/27/2022 12:11:15 - INFO - codeparrot_training - Step 17366: {'lr': 0.0003838737967177271, 'samples': 3334464, 'steps': 17366, 'loss/train': 0.3478442654013634} +01/27/2022 12:11:18 - INFO - codeparrot_training - Step 17367: {'lr': 0.00038385997770572336, 'samples': 3334656, 'steps': 17367, 'loss/train': 0.734886422753334} +01/27/2022 12:11:21 - INFO - codeparrot_training - Step 17368: {'lr': 0.0003838461581203058, 'samples': 3334848, 'steps': 17368, 'loss/train': 0.6378163844347} +01/27/2022 12:11:24 - INFO - codeparrot_training - Step 17369: {'lr': 0.00038383233796153383, 'samples': 3335040, 'steps': 17369, 'loss/train': 0.9703617095947266} +01/27/2022 12:11:27 - INFO - codeparrot_training - Step 17370: {'lr': 0.00038381851722946663, 'samples': 3335232, 'steps': 17370, 'loss/train': 0.8987545967102051} +01/27/2022 12:11:32 - INFO - codeparrot_training - Step 17371: {'lr': 0.00038380469592416347, 'samples': 3335424, 'steps': 17371, 'loss/train': 0.7706951200962067} +01/27/2022 12:11:35 - INFO - codeparrot_training - Step 17372: {'lr': 0.00038379087404568333, 'samples': 3335616, 'steps': 17372, 'loss/train': 1.4598216712474823} +01/27/2022 12:11:38 - INFO - codeparrot_training - Step 17373: {'lr': 0.0003837770515940857, 'samples': 3335808, 'steps': 17373, 'loss/train': 0.5704761296510696} +01/27/2022 12:11:41 - INFO - codeparrot_training - Step 17374: {'lr': 0.0003837632285694296, 'samples': 3336000, 'steps': 17374, 'loss/train': 0.8199278712272644} +01/27/2022 12:11:44 - INFO - codeparrot_training - Step 17375: {'lr': 0.00038374940497177434, 'samples': 3336192, 'steps': 17375, 'loss/train': 0.8777465522289276} +01/27/2022 12:11:47 - INFO - codeparrot_training - Step 17376: {'lr': 0.000383735580801179, 'samples': 3336384, 'steps': 17376, 'loss/train': 1.1777991950511932} +01/27/2022 12:11:50 - INFO - codeparrot_training - Step 17377: {'lr': 0.00038372175605770305, 'samples': 3336576, 'steps': 17377, 'loss/train': 0.5687758773565292} +01/27/2022 12:11:54 - INFO - codeparrot_training - Step 17378: {'lr': 0.00038370793074140545, 'samples': 3336768, 'steps': 17378, 'loss/train': 0.6160747110843658} +01/27/2022 12:11:57 - INFO - codeparrot_training - Step 17379: {'lr': 0.00038369410485234557, 'samples': 3336960, 'steps': 17379, 'loss/train': 0.7546310126781464} +01/27/2022 12:12:01 - INFO - codeparrot_training - Step 17380: {'lr': 0.0003836802783905826, 'samples': 3337152, 'steps': 17380, 'loss/train': 0.4948379248380661} +01/27/2022 12:12:04 - INFO - codeparrot_training - Step 17381: {'lr': 0.0003836664513561758, 'samples': 3337344, 'steps': 17381, 'loss/train': 1.0011994242668152} +01/27/2022 12:12:07 - INFO - codeparrot_training - Step 17382: {'lr': 0.0003836526237491843, 'samples': 3337536, 'steps': 17382, 'loss/train': 0.876231461763382} +01/27/2022 12:12:10 - INFO - codeparrot_training - Step 17383: {'lr': 0.0003836387955696674, 'samples': 3337728, 'steps': 17383, 'loss/train': 0.9364541172981262} +01/27/2022 12:12:14 - INFO - codeparrot_training - Step 17384: {'lr': 0.00038362496681768434, 'samples': 3337920, 'steps': 17384, 'loss/train': 1.2400301098823547} +01/27/2022 12:12:17 - INFO - codeparrot_training - Step 17385: {'lr': 0.00038361113749329443, 'samples': 3338112, 'steps': 17385, 'loss/train': 1.4865602552890778} +01/27/2022 12:12:20 - INFO - codeparrot_training - Step 17386: {'lr': 0.00038359730759655674, 'samples': 3338304, 'steps': 17386, 'loss/train': 0.679748922586441} +01/27/2022 12:12:23 - INFO - codeparrot_training - Step 17387: {'lr': 0.00038358347712753063, 'samples': 3338496, 'steps': 17387, 'loss/train': 0.8711357116699219} +01/27/2022 12:12:26 - INFO - codeparrot_training - Step 17388: {'lr': 0.0003835696460862753, 'samples': 3338688, 'steps': 17388, 'loss/train': 0.712509423494339} +01/27/2022 12:12:33 - INFO - codeparrot_training - Step 17389: {'lr': 0.00038355581447285005, 'samples': 3338880, 'steps': 17389, 'loss/train': 0.57874895632267} +01/27/2022 12:12:36 - INFO - codeparrot_training - Step 17390: {'lr': 0.00038354198228731414, 'samples': 3339072, 'steps': 17390, 'loss/train': 0.8388234078884125} +01/27/2022 12:12:39 - INFO - codeparrot_training - Step 17391: {'lr': 0.0003835281495297267, 'samples': 3339264, 'steps': 17391, 'loss/train': 0.6834256052970886} +01/27/2022 12:12:43 - INFO - codeparrot_training - Step 17392: {'lr': 0.0003835143162001472, 'samples': 3339456, 'steps': 17392, 'loss/train': 0.38365624845027924} +01/27/2022 12:12:46 - INFO - codeparrot_training - Step 17393: {'lr': 0.0003835004822986346, 'samples': 3339648, 'steps': 17393, 'loss/train': 1.2774766087532043} +01/27/2022 12:12:49 - INFO - codeparrot_training - Step 17394: {'lr': 0.00038348664782524846, 'samples': 3339840, 'steps': 17394, 'loss/train': 0.8322208821773529} +01/27/2022 12:12:52 - INFO - codeparrot_training - Step 17395: {'lr': 0.00038347281278004774, 'samples': 3340032, 'steps': 17395, 'loss/train': 0.9303825795650482} +01/27/2022 12:12:55 - INFO - codeparrot_training - Step 17396: {'lr': 0.0003834589771630921, 'samples': 3340224, 'steps': 17396, 'loss/train': 0.9360744953155518} +01/27/2022 12:13:00 - INFO - codeparrot_training - Step 17397: {'lr': 0.0003834451409744404, 'samples': 3340416, 'steps': 17397, 'loss/train': 0.9696314334869385} +01/27/2022 12:13:03 - INFO - codeparrot_training - Step 17398: {'lr': 0.0003834313042141522, 'samples': 3340608, 'steps': 17398, 'loss/train': 0.8427042067050934} +01/27/2022 12:13:06 - INFO - codeparrot_training - Step 17399: {'lr': 0.0003834174668822865, 'samples': 3340800, 'steps': 17399, 'loss/train': 0.32703063637018204} +01/27/2022 12:13:09 - INFO - codeparrot_training - Step 17400: {'lr': 0.0003834036289789029, 'samples': 3340992, 'steps': 17400, 'loss/train': 0.49507784843444824} +01/27/2022 12:13:12 - INFO - codeparrot_training - Step 17401: {'lr': 0.0003833897905040604, 'samples': 3341184, 'steps': 17401, 'loss/train': 0.8647011816501617} +01/27/2022 12:13:15 - INFO - codeparrot_training - Step 17402: {'lr': 0.00038337595145781844, 'samples': 3341376, 'steps': 17402, 'loss/train': 0.5454166531562805} +01/27/2022 12:13:19 - INFO - codeparrot_training - Step 17403: {'lr': 0.00038336211184023634, 'samples': 3341568, 'steps': 17403, 'loss/train': 0.7052793502807617} +01/27/2022 12:13:22 - INFO - codeparrot_training - Step 17404: {'lr': 0.0003833482716513732, 'samples': 3341760, 'steps': 17404, 'loss/train': 1.0739258229732513} +01/27/2022 12:13:25 - INFO - codeparrot_training - Step 17405: {'lr': 0.0003833344308912885, 'samples': 3341952, 'steps': 17405, 'loss/train': 1.2584266662597656} +01/27/2022 12:13:31 - INFO - codeparrot_training - Step 17406: {'lr': 0.00038332058956004134, 'samples': 3342144, 'steps': 17406, 'loss/train': 0.7941047251224518} +01/27/2022 12:13:34 - INFO - codeparrot_training - Step 17407: {'lr': 0.0003833067476576911, 'samples': 3342336, 'steps': 17407, 'loss/train': 0.9457806944847107} +01/27/2022 12:13:37 - INFO - codeparrot_training - Step 17408: {'lr': 0.0003832929051842972, 'samples': 3342528, 'steps': 17408, 'loss/train': 0.9787053465843201} +01/27/2022 12:13:40 - INFO - codeparrot_training - Step 17409: {'lr': 0.0003832790621399187, 'samples': 3342720, 'steps': 17409, 'loss/train': 0.7154784053564072} +01/27/2022 12:13:44 - INFO - codeparrot_training - Step 17410: {'lr': 0.00038326521852461505, 'samples': 3342912, 'steps': 17410, 'loss/train': 0.6060643941164017} +01/27/2022 12:13:47 - INFO - codeparrot_training - Step 17411: {'lr': 0.0003832513743384456, 'samples': 3343104, 'steps': 17411, 'loss/train': 1.202091783285141} +01/27/2022 12:13:50 - INFO - codeparrot_training - Step 17412: {'lr': 0.0003832375295814695, 'samples': 3343296, 'steps': 17412, 'loss/train': 0.889755517244339} +01/27/2022 12:13:53 - INFO - codeparrot_training - Step 17413: {'lr': 0.0003832236842537461, 'samples': 3343488, 'steps': 17413, 'loss/train': 0.6376149952411652} +01/27/2022 12:13:56 - INFO - codeparrot_training - Step 17414: {'lr': 0.0003832098383553347, 'samples': 3343680, 'steps': 17414, 'loss/train': 0.6875883489847183} +01/27/2022 12:14:01 - INFO - codeparrot_training - Step 17415: {'lr': 0.00038319599188629485, 'samples': 3343872, 'steps': 17415, 'loss/train': 0.9991341233253479} +01/27/2022 12:14:04 - INFO - codeparrot_training - Step 17416: {'lr': 0.00038318214484668557, 'samples': 3344064, 'steps': 17416, 'loss/train': 0.9859030544757843} +01/27/2022 12:14:07 - INFO - codeparrot_training - Step 17417: {'lr': 0.0003831682972365662, 'samples': 3344256, 'steps': 17417, 'loss/train': 1.0002467036247253} +01/27/2022 12:14:10 - INFO - codeparrot_training - Step 17418: {'lr': 0.0003831544490559962, 'samples': 3344448, 'steps': 17418, 'loss/train': 0.6348497718572617} +01/27/2022 12:14:13 - INFO - codeparrot_training - Step 17419: {'lr': 0.00038314060030503476, 'samples': 3344640, 'steps': 17419, 'loss/train': 0.6488919407129288} +01/27/2022 12:14:16 - INFO - codeparrot_training - Step 17420: {'lr': 0.00038312675098374136, 'samples': 3344832, 'steps': 17420, 'loss/train': 0.7642919719219208} +01/27/2022 12:14:19 - INFO - codeparrot_training - Step 17421: {'lr': 0.0003831129010921751, 'samples': 3345024, 'steps': 17421, 'loss/train': 0.3899569809436798} +01/27/2022 12:14:23 - INFO - codeparrot_training - Step 17422: {'lr': 0.0003830990506303956, 'samples': 3345216, 'steps': 17422, 'loss/train': 0.7257312387228012} +01/27/2022 12:14:27 - INFO - codeparrot_training - Step 17423: {'lr': 0.0003830851995984619, 'samples': 3345408, 'steps': 17423, 'loss/train': 0.932055652141571} +01/27/2022 12:14:30 - INFO - codeparrot_training - Step 17424: {'lr': 0.0003830713479964335, 'samples': 3345600, 'steps': 17424, 'loss/train': 0.968845009803772} +01/27/2022 12:14:33 - INFO - codeparrot_training - Step 17425: {'lr': 0.0003830574958243697, 'samples': 3345792, 'steps': 17425, 'loss/train': 0.5606117695569992} +01/27/2022 12:14:37 - INFO - codeparrot_training - Step 17426: {'lr': 0.00038304364308232986, 'samples': 3345984, 'steps': 17426, 'loss/train': 0.6127442568540573} +01/27/2022 12:14:40 - INFO - codeparrot_training - Step 17427: {'lr': 0.0003830297897703733, 'samples': 3346176, 'steps': 17427, 'loss/train': 0.9435603618621826} +01/27/2022 12:14:43 - INFO - codeparrot_training - Step 17428: {'lr': 0.0003830159358885593, 'samples': 3346368, 'steps': 17428, 'loss/train': 0.7075652182102203} +01/27/2022 12:14:46 - INFO - codeparrot_training - Step 17429: {'lr': 0.00038300208143694737, 'samples': 3346560, 'steps': 17429, 'loss/train': 0.33063722401857376} +01/27/2022 12:14:49 - INFO - codeparrot_training - Step 17430: {'lr': 0.00038298822641559673, 'samples': 3346752, 'steps': 17430, 'loss/train': 0.7688887417316437} +01/27/2022 12:14:52 - INFO - codeparrot_training - Step 17431: {'lr': 0.0003829743708245667, 'samples': 3346944, 'steps': 17431, 'loss/train': 0.7742585241794586} +01/27/2022 12:14:57 - INFO - codeparrot_training - Step 17432: {'lr': 0.0003829605146639167, 'samples': 3347136, 'steps': 17432, 'loss/train': 0.7253496944904327} +01/27/2022 12:15:00 - INFO - codeparrot_training - Step 17433: {'lr': 0.0003829466579337061, 'samples': 3347328, 'steps': 17433, 'loss/train': 0.8708010613918304} +01/27/2022 12:15:03 - INFO - codeparrot_training - Step 17434: {'lr': 0.00038293280063399427, 'samples': 3347520, 'steps': 17434, 'loss/train': 0.8003986179828644} +01/27/2022 12:15:06 - INFO - codeparrot_training - Step 17435: {'lr': 0.00038291894276484053, 'samples': 3347712, 'steps': 17435, 'loss/train': 0.7516883611679077} +01/27/2022 12:15:09 - INFO - codeparrot_training - Step 17436: {'lr': 0.0003829050843263041, 'samples': 3347904, 'steps': 17436, 'loss/train': 0.8640706837177277} +01/27/2022 12:15:13 - INFO - codeparrot_training - Step 17437: {'lr': 0.0003828912253184446, 'samples': 3348096, 'steps': 17437, 'loss/train': 0.4093736708164215} +01/27/2022 12:15:16 - INFO - codeparrot_training - Step 17438: {'lr': 0.0003828773657413213, 'samples': 3348288, 'steps': 17438, 'loss/train': 0.8513549566268921} +01/27/2022 12:15:19 - INFO - codeparrot_training - Step 17439: {'lr': 0.0003828635055949935, 'samples': 3348480, 'steps': 17439, 'loss/train': 1.0270476043224335} +01/27/2022 12:15:22 - INFO - codeparrot_training - Step 17440: {'lr': 0.0003828496448795207, 'samples': 3348672, 'steps': 17440, 'loss/train': 0.7145764231681824} +01/27/2022 12:15:28 - INFO - codeparrot_training - Step 17441: {'lr': 0.0003828357835949622, 'samples': 3348864, 'steps': 17441, 'loss/train': 0.48617054522037506} +01/27/2022 12:15:31 - INFO - codeparrot_training - Step 17442: {'lr': 0.00038282192174137744, 'samples': 3349056, 'steps': 17442, 'loss/train': 1.292456030845642} +01/27/2022 12:15:35 - INFO - codeparrot_training - Step 17443: {'lr': 0.00038280805931882557, 'samples': 3349248, 'steps': 17443, 'loss/train': 0.8344650864601135} +01/27/2022 12:15:38 - INFO - codeparrot_training - Step 17444: {'lr': 0.0003827941963273663, 'samples': 3349440, 'steps': 17444, 'loss/train': 0.596059575676918} +01/27/2022 12:15:41 - INFO - codeparrot_training - Step 17445: {'lr': 0.00038278033276705875, 'samples': 3349632, 'steps': 17445, 'loss/train': 0.6150305271148682} +01/27/2022 12:15:44 - INFO - codeparrot_training - Step 17446: {'lr': 0.0003827664686379625, 'samples': 3349824, 'steps': 17446, 'loss/train': 0.24077432602643967} +01/27/2022 12:15:47 - INFO - codeparrot_training - Step 17447: {'lr': 0.00038275260394013676, 'samples': 3350016, 'steps': 17447, 'loss/train': 0.5864391177892685} +01/27/2022 12:15:50 - INFO - codeparrot_training - Step 17448: {'lr': 0.0003827387386736411, 'samples': 3350208, 'steps': 17448, 'loss/train': 0.07031954638659954} +01/27/2022 12:15:53 - INFO - codeparrot_training - Step 17449: {'lr': 0.0003827248728385349, 'samples': 3350400, 'steps': 17449, 'loss/train': 0.7389795184135437} +01/27/2022 12:15:58 - INFO - codeparrot_training - Step 17450: {'lr': 0.0003827110064348773, 'samples': 3350592, 'steps': 17450, 'loss/train': 1.1510248482227325} +01/27/2022 12:16:01 - INFO - codeparrot_training - Step 17451: {'lr': 0.000382697139462728, 'samples': 3350784, 'steps': 17451, 'loss/train': 0.8760152757167816} +01/27/2022 12:16:04 - INFO - codeparrot_training - Step 17452: {'lr': 0.00038268327192214635, 'samples': 3350976, 'steps': 17452, 'loss/train': 0.7420395165681839} +01/27/2022 12:16:07 - INFO - codeparrot_training - Step 17453: {'lr': 0.0003826694038131916, 'samples': 3351168, 'steps': 17453, 'loss/train': 0.7792365252971649} +01/27/2022 12:16:10 - INFO - codeparrot_training - Step 17454: {'lr': 0.00038265553513592334, 'samples': 3351360, 'steps': 17454, 'loss/train': 0.2625616192817688} +01/27/2022 12:16:13 - INFO - codeparrot_training - Step 17455: {'lr': 0.00038264166589040084, 'samples': 3351552, 'steps': 17455, 'loss/train': 0.5943495333194733} +01/27/2022 12:16:17 - INFO - codeparrot_training - Step 17456: {'lr': 0.00038262779607668354, 'samples': 3351744, 'steps': 17456, 'loss/train': 0.800645649433136} +01/27/2022 12:16:20 - INFO - codeparrot_training - Step 17457: {'lr': 0.00038261392569483087, 'samples': 3351936, 'steps': 17457, 'loss/train': 0.7355403453111649} +01/27/2022 12:16:23 - INFO - codeparrot_training - Step 17458: {'lr': 0.0003826000547449023, 'samples': 3352128, 'steps': 17458, 'loss/train': 1.0187968611717224} +01/27/2022 12:16:27 - INFO - codeparrot_training - Step 17459: {'lr': 0.0003825861832269571, 'samples': 3352320, 'steps': 17459, 'loss/train': 1.09365713596344} +01/27/2022 12:16:31 - INFO - codeparrot_training - Step 17460: {'lr': 0.00038257231114105495, 'samples': 3352512, 'steps': 17460, 'loss/train': 0.7036960870027542} +01/27/2022 12:16:34 - INFO - codeparrot_training - Step 17461: {'lr': 0.00038255843848725504, 'samples': 3352704, 'steps': 17461, 'loss/train': 0.8107507824897766} +01/27/2022 12:16:37 - INFO - codeparrot_training - Step 17462: {'lr': 0.0003825445652656169, 'samples': 3352896, 'steps': 17462, 'loss/train': 0.35699623078107834} +01/27/2022 12:16:40 - INFO - codeparrot_training - Step 17463: {'lr': 0.00038253069147619977, 'samples': 3353088, 'steps': 17463, 'loss/train': 0.6396523118019104} +01/27/2022 12:16:43 - INFO - codeparrot_training - Step 17464: {'lr': 0.00038251681711906345, 'samples': 3353280, 'steps': 17464, 'loss/train': 0.85936439037323} +01/27/2022 12:16:46 - INFO - codeparrot_training - Step 17465: {'lr': 0.00038250294219426706, 'samples': 3353472, 'steps': 17465, 'loss/train': 0.860742598772049} +01/27/2022 12:16:49 - INFO - codeparrot_training - Step 17466: {'lr': 0.00038248906670187017, 'samples': 3353664, 'steps': 17466, 'loss/train': 0.18838533014059067} +01/27/2022 12:16:53 - INFO - codeparrot_training - Step 17467: {'lr': 0.00038247519064193216, 'samples': 3353856, 'steps': 17467, 'loss/train': 0.7119137048721313} +01/27/2022 12:16:59 - INFO - codeparrot_training - Step 17468: {'lr': 0.0003824613140145125, 'samples': 3354048, 'steps': 17468, 'loss/train': 0.8265596330165863} +01/27/2022 12:17:02 - INFO - codeparrot_training - Step 17469: {'lr': 0.00038244743681967066, 'samples': 3354240, 'steps': 17469, 'loss/train': 0.8623052537441254} +01/27/2022 12:17:05 - INFO - codeparrot_training - Step 17470: {'lr': 0.000382433559057466, 'samples': 3354432, 'steps': 17470, 'loss/train': 1.085395485162735} +01/27/2022 12:17:08 - INFO - codeparrot_training - Step 17471: {'lr': 0.00038241968072795805, 'samples': 3354624, 'steps': 17471, 'loss/train': 0.5383385717868805} +01/27/2022 12:17:12 - INFO - codeparrot_training - Step 17472: {'lr': 0.00038240580183120624, 'samples': 3354816, 'steps': 17472, 'loss/train': 0.6566796898841858} +01/27/2022 12:17:15 - INFO - codeparrot_training - Step 17473: {'lr': 0.0003823919223672701, 'samples': 3355008, 'steps': 17473, 'loss/train': 0.8781140148639679} +01/27/2022 12:17:18 - INFO - codeparrot_training - Step 17474: {'lr': 0.00038237804233620887, 'samples': 3355200, 'steps': 17474, 'loss/train': 0.47033657133579254} +01/27/2022 12:17:21 - INFO - codeparrot_training - Step 17475: {'lr': 0.0003823641617380823, 'samples': 3355392, 'steps': 17475, 'loss/train': 0.7462128549814224} +01/27/2022 12:17:26 - INFO - codeparrot_training - Step 17476: {'lr': 0.00038235028057294953, 'samples': 3355584, 'steps': 17476, 'loss/train': 1.4378129839897156} +01/27/2022 12:17:29 - INFO - codeparrot_training - Step 17477: {'lr': 0.0003823363988408703, 'samples': 3355776, 'steps': 17477, 'loss/train': 1.2519384026527405} +01/27/2022 12:17:32 - INFO - codeparrot_training - Step 17478: {'lr': 0.00038232251654190386, 'samples': 3355968, 'steps': 17478, 'loss/train': 0.7136898636817932} +01/27/2022 12:17:35 - INFO - codeparrot_training - Step 17479: {'lr': 0.0003823086336761099, 'samples': 3356160, 'steps': 17479, 'loss/train': 1.0080040991306305} +01/27/2022 12:17:38 - INFO - codeparrot_training - Step 17480: {'lr': 0.00038229475024354766, 'samples': 3356352, 'steps': 17480, 'loss/train': 0.7654073238372803} +01/27/2022 12:17:42 - INFO - codeparrot_training - Step 17481: {'lr': 0.00038228086624427675, 'samples': 3356544, 'steps': 17481, 'loss/train': 1.0931824743747711} +01/27/2022 12:17:45 - INFO - codeparrot_training - Step 17482: {'lr': 0.0003822669816783566, 'samples': 3356736, 'steps': 17482, 'loss/train': 0.8471819758415222} +01/27/2022 12:17:48 - INFO - codeparrot_training - Step 17483: {'lr': 0.0003822530965458467, 'samples': 3356928, 'steps': 17483, 'loss/train': 0.9397275745868683} +01/27/2022 12:17:51 - INFO - codeparrot_training - Step 17484: {'lr': 0.0003822392108468066, 'samples': 3357120, 'steps': 17484, 'loss/train': 0.6372285783290863} +01/27/2022 12:17:56 - INFO - codeparrot_training - Step 17485: {'lr': 0.00038222532458129563, 'samples': 3357312, 'steps': 17485, 'loss/train': 1.007326751947403} +01/27/2022 12:17:59 - INFO - codeparrot_training - Step 17486: {'lr': 0.0003822114377493734, 'samples': 3357504, 'steps': 17486, 'loss/train': 0.0823157038539648} +01/27/2022 12:18:02 - INFO - codeparrot_training - Step 17487: {'lr': 0.0003821975503510993, 'samples': 3357696, 'steps': 17487, 'loss/train': 0.8822834193706512} +01/27/2022 12:18:05 - INFO - codeparrot_training - Step 17488: {'lr': 0.0003821836623865329, 'samples': 3357888, 'steps': 17488, 'loss/train': 0.6713958531618118} +01/27/2022 12:18:08 - INFO - codeparrot_training - Step 17489: {'lr': 0.0003821697738557337, 'samples': 3358080, 'steps': 17489, 'loss/train': 3.5097720623016357} +01/27/2022 12:18:11 - INFO - codeparrot_training - Step 17490: {'lr': 0.00038215588475876117, 'samples': 3358272, 'steps': 17490, 'loss/train': 0.7954971492290497} +01/27/2022 12:18:14 - INFO - codeparrot_training - Step 17491: {'lr': 0.0003821419950956747, 'samples': 3358464, 'steps': 17491, 'loss/train': 1.2565458118915558} +01/27/2022 12:18:17 - INFO - codeparrot_training - Step 17492: {'lr': 0.00038212810486653394, 'samples': 3358656, 'steps': 17492, 'loss/train': 1.0412573218345642} +01/27/2022 12:18:21 - INFO - codeparrot_training - Step 17493: {'lr': 0.0003821142140713983, 'samples': 3358848, 'steps': 17493, 'loss/train': 0.6795124858617783} +01/27/2022 12:18:27 - INFO - codeparrot_training - Step 17494: {'lr': 0.0003821003227103274, 'samples': 3359040, 'steps': 17494, 'loss/train': 1.0772932469844818} +01/27/2022 12:18:30 - INFO - codeparrot_training - Step 17495: {'lr': 0.00038208643078338055, 'samples': 3359232, 'steps': 17495, 'loss/train': 0.9373748302459717} +01/27/2022 12:18:33 - INFO - codeparrot_training - Step 17496: {'lr': 0.0003820725382906175, 'samples': 3359424, 'steps': 17496, 'loss/train': 0.7637994289398193} +01/27/2022 12:18:37 - INFO - codeparrot_training - Step 17497: {'lr': 0.0003820586452320975, 'samples': 3359616, 'steps': 17497, 'loss/train': 0.0925069022923708} +01/27/2022 12:18:40 - INFO - codeparrot_training - Step 17498: {'lr': 0.0003820447516078803, 'samples': 3359808, 'steps': 17498, 'loss/train': 0.43254344165325165} +01/27/2022 12:18:43 - INFO - codeparrot_training - Step 17499: {'lr': 0.0003820308574180253, 'samples': 3360000, 'steps': 17499, 'loss/train': 0.8230116963386536} +01/27/2022 12:18:46 - INFO - codeparrot_training - Step 17500: {'lr': 0.000382016962662592, 'samples': 3360192, 'steps': 17500, 'loss/train': 0.4130830317735672} +01/27/2022 12:18:49 - INFO - codeparrot_training - Step 17501: {'lr': 0.0003820030673416399, 'samples': 3360384, 'steps': 17501, 'loss/train': 0.45884111523628235} +01/27/2022 12:18:52 - INFO - codeparrot_training - Step 17502: {'lr': 0.0003819891714552287, 'samples': 3360576, 'steps': 17502, 'loss/train': 0.5004651993513107} +01/27/2022 12:18:57 - INFO - codeparrot_training - Step 17503: {'lr': 0.00038197527500341777, 'samples': 3360768, 'steps': 17503, 'loss/train': 0.9714834988117218} +01/27/2022 12:19:00 - INFO - codeparrot_training - Step 17504: {'lr': 0.00038196137798626663, 'samples': 3360960, 'steps': 17504, 'loss/train': 0.8076061606407166} +01/27/2022 12:19:03 - INFO - codeparrot_training - Step 17505: {'lr': 0.00038194748040383487, 'samples': 3361152, 'steps': 17505, 'loss/train': 1.1978643536567688} +01/27/2022 12:19:06 - INFO - codeparrot_training - Step 17506: {'lr': 0.00038193358225618195, 'samples': 3361344, 'steps': 17506, 'loss/train': 0.8706175088882446} +01/27/2022 12:19:10 - INFO - codeparrot_training - Step 17507: {'lr': 0.0003819196835433675, 'samples': 3361536, 'steps': 17507, 'loss/train': 0.5156856179237366} +01/27/2022 12:19:13 - INFO - codeparrot_training - Step 17508: {'lr': 0.000381905784265451, 'samples': 3361728, 'steps': 17508, 'loss/train': 0.6851937621831894} +01/27/2022 12:19:16 - INFO - codeparrot_training - Step 17509: {'lr': 0.000381891884422492, 'samples': 3361920, 'steps': 17509, 'loss/train': 0.5026418119668961} +01/27/2022 12:19:19 - INFO - codeparrot_training - Step 17510: {'lr': 0.0003818779840145501, 'samples': 3362112, 'steps': 17510, 'loss/train': 1.190850019454956} +01/27/2022 12:19:25 - INFO - codeparrot_training - Step 17511: {'lr': 0.00038186408304168474, 'samples': 3362304, 'steps': 17511, 'loss/train': 0.4246915429830551} +01/27/2022 12:19:28 - INFO - codeparrot_training - Step 17512: {'lr': 0.00038185018150395557, 'samples': 3362496, 'steps': 17512, 'loss/train': 0.46924401819705963} +01/27/2022 12:19:32 - INFO - codeparrot_training - Step 17513: {'lr': 0.000381836279401422, 'samples': 3362688, 'steps': 17513, 'loss/train': 0.9070177674293518} +01/27/2022 12:19:35 - INFO - codeparrot_training - Step 17514: {'lr': 0.00038182237673414375, 'samples': 3362880, 'steps': 17514, 'loss/train': 0.9267464876174927} +01/27/2022 12:19:38 - INFO - codeparrot_training - Step 17515: {'lr': 0.0003818084735021803, 'samples': 3363072, 'steps': 17515, 'loss/train': 0.6759776920080185} +01/27/2022 12:19:41 - INFO - codeparrot_training - Step 17516: {'lr': 0.00038179456970559116, 'samples': 3363264, 'steps': 17516, 'loss/train': 1.1087413430213928} +01/27/2022 12:19:44 - INFO - codeparrot_training - Step 17517: {'lr': 0.00038178066534443587, 'samples': 3363456, 'steps': 17517, 'loss/train': 0.7099573016166687} +01/27/2022 12:19:47 - INFO - codeparrot_training - Step 17518: {'lr': 0.00038176676041877424, 'samples': 3363648, 'steps': 17518, 'loss/train': 0.48759596049785614} +01/27/2022 12:19:50 - INFO - codeparrot_training - Step 17519: {'lr': 0.0003817528549286655, 'samples': 3363840, 'steps': 17519, 'loss/train': 0.6598745137453079} +01/27/2022 12:19:55 - INFO - codeparrot_training - Step 17520: {'lr': 0.00038173894887416946, 'samples': 3364032, 'steps': 17520, 'loss/train': 0.7835880517959595} +01/27/2022 12:19:58 - INFO - codeparrot_training - Step 17521: {'lr': 0.0003817250422553455, 'samples': 3364224, 'steps': 17521, 'loss/train': 0.9134827852249146} +01/27/2022 12:20:01 - INFO - codeparrot_training - Step 17522: {'lr': 0.0003817111350722533, 'samples': 3364416, 'steps': 17522, 'loss/train': 0.5483958721160889} +01/27/2022 12:20:05 - INFO - codeparrot_training - Step 17523: {'lr': 0.0003816972273249525, 'samples': 3364608, 'steps': 17523, 'loss/train': 0.6918669193983078} +01/27/2022 12:20:08 - INFO - codeparrot_training - Step 17524: {'lr': 0.00038168331901350253, 'samples': 3364800, 'steps': 17524, 'loss/train': 0.8949649035930634} +01/27/2022 12:20:11 - INFO - codeparrot_training - Step 17525: {'lr': 0.0003816694101379631, 'samples': 3364992, 'steps': 17525, 'loss/train': 0.6667816489934921} +01/27/2022 12:20:14 - INFO - codeparrot_training - Step 17526: {'lr': 0.0003816555006983936, 'samples': 3365184, 'steps': 17526, 'loss/train': 0.8910436928272247} +01/27/2022 12:20:17 - INFO - codeparrot_training - Step 17527: {'lr': 0.0003816415906948538, 'samples': 3365376, 'steps': 17527, 'loss/train': 0.7835187613964081} +01/27/2022 12:20:20 - INFO - codeparrot_training - Step 17528: {'lr': 0.00038162768012740323, 'samples': 3365568, 'steps': 17528, 'loss/train': 0.6811739355325699} +01/27/2022 12:20:25 - INFO - codeparrot_training - Step 17529: {'lr': 0.00038161376899610154, 'samples': 3365760, 'steps': 17529, 'loss/train': 1.2958041429519653} +01/27/2022 12:20:28 - INFO - codeparrot_training - Step 17530: {'lr': 0.0003815998573010082, 'samples': 3365952, 'steps': 17530, 'loss/train': 0.6818699240684509} +01/27/2022 12:20:31 - INFO - codeparrot_training - Step 17531: {'lr': 0.0003815859450421829, 'samples': 3366144, 'steps': 17531, 'loss/train': 0.8118098080158234} +01/27/2022 12:20:34 - INFO - codeparrot_training - Step 17532: {'lr': 0.00038157203221968514, 'samples': 3366336, 'steps': 17532, 'loss/train': 0.5982089638710022} +01/27/2022 12:20:37 - INFO - codeparrot_training - Step 17533: {'lr': 0.00038155811883357454, 'samples': 3366528, 'steps': 17533, 'loss/train': 0.7265773862600327} +01/27/2022 12:20:40 - INFO - codeparrot_training - Step 17534: {'lr': 0.0003815442048839108, 'samples': 3366720, 'steps': 17534, 'loss/train': 1.0042346119880676} +01/27/2022 12:20:44 - INFO - codeparrot_training - Step 17535: {'lr': 0.0003815302903707534, 'samples': 3366912, 'steps': 17535, 'loss/train': 0.8042386472225189} +01/27/2022 12:20:47 - INFO - codeparrot_training - Step 17536: {'lr': 0.0003815163752941621, 'samples': 3367104, 'steps': 17536, 'loss/train': 0.7953372895717621} +01/27/2022 12:20:50 - INFO - codeparrot_training - Step 17537: {'lr': 0.00038150245965419636, 'samples': 3367296, 'steps': 17537, 'loss/train': 0.6973494440317154} +01/27/2022 12:20:54 - INFO - codeparrot_training - Step 17538: {'lr': 0.0003814885434509158, 'samples': 3367488, 'steps': 17538, 'loss/train': 0.801744818687439} +01/27/2022 12:20:58 - INFO - codeparrot_training - Step 17539: {'lr': 0.0003814746266843801, 'samples': 3367680, 'steps': 17539, 'loss/train': 1.0285227298736572} +01/27/2022 12:21:01 - INFO - codeparrot_training - Step 17540: {'lr': 0.0003814607093546489, 'samples': 3367872, 'steps': 17540, 'loss/train': 0.6992193460464478} +01/27/2022 12:21:04 - INFO - codeparrot_training - Step 17541: {'lr': 0.00038144679146178166, 'samples': 3368064, 'steps': 17541, 'loss/train': 0.9193615615367889} +01/27/2022 12:21:07 - INFO - codeparrot_training - Step 17542: {'lr': 0.00038143287300583816, 'samples': 3368256, 'steps': 17542, 'loss/train': 0.8259209990501404} +01/27/2022 12:21:10 - INFO - codeparrot_training - Step 17543: {'lr': 0.00038141895398687806, 'samples': 3368448, 'steps': 17543, 'loss/train': 0.5555993020534515} +01/27/2022 12:21:13 - INFO - codeparrot_training - Step 17544: {'lr': 0.0003814050344049608, 'samples': 3368640, 'steps': 17544, 'loss/train': 1.0293809473514557} +01/27/2022 12:21:16 - INFO - codeparrot_training - Step 17545: {'lr': 0.00038139111426014607, 'samples': 3368832, 'steps': 17545, 'loss/train': 1.2555532157421112} +01/27/2022 12:21:20 - INFO - codeparrot_training - Step 17546: {'lr': 0.00038137719355249355, 'samples': 3369024, 'steps': 17546, 'loss/train': 0.9436326026916504} +01/27/2022 12:21:26 - INFO - codeparrot_training - Step 17547: {'lr': 0.00038136327228206285, 'samples': 3369216, 'steps': 17547, 'loss/train': 0.6809262335300446} +01/27/2022 12:21:29 - INFO - codeparrot_training - Step 17548: {'lr': 0.0003813493504489136, 'samples': 3369408, 'steps': 17548, 'loss/train': 0.9901327192783356} +01/27/2022 12:21:32 - INFO - codeparrot_training - Step 17549: {'lr': 0.0003813354280531055, 'samples': 3369600, 'steps': 17549, 'loss/train': 0.6241798549890518} +01/27/2022 12:21:35 - INFO - codeparrot_training - Step 17550: {'lr': 0.00038132150509469806, 'samples': 3369792, 'steps': 17550, 'loss/train': 0.7401200383901596} +01/27/2022 12:21:39 - INFO - codeparrot_training - Step 17551: {'lr': 0.000381307581573751, 'samples': 3369984, 'steps': 17551, 'loss/train': 0.8443306088447571} +01/27/2022 12:21:42 - INFO - codeparrot_training - Step 17552: {'lr': 0.00038129365749032395, 'samples': 3370176, 'steps': 17552, 'loss/train': 0.697687491774559} +01/27/2022 12:21:45 - INFO - codeparrot_training - Step 17553: {'lr': 0.0003812797328444766, 'samples': 3370368, 'steps': 17553, 'loss/train': 0.9391675293445587} +01/27/2022 12:21:48 - INFO - codeparrot_training - Step 17554: {'lr': 0.0003812658076362685, 'samples': 3370560, 'steps': 17554, 'loss/train': 0.3838339000940323} +01/27/2022 12:21:52 - INFO - codeparrot_training - Step 17555: {'lr': 0.00038125188186575944, 'samples': 3370752, 'steps': 17555, 'loss/train': 0.4677797108888626} +01/27/2022 12:21:55 - INFO - codeparrot_training - Step 17556: {'lr': 0.00038123795553300893, 'samples': 3370944, 'steps': 17556, 'loss/train': 0.8074381649494171} +01/27/2022 12:21:59 - INFO - codeparrot_training - Step 17557: {'lr': 0.0003812240286380767, 'samples': 3371136, 'steps': 17557, 'loss/train': 0.8397026360034943} +01/27/2022 12:22:02 - INFO - codeparrot_training - Step 17558: {'lr': 0.0003812101011810224, 'samples': 3371328, 'steps': 17558, 'loss/train': 0.6931201368570328} +01/27/2022 12:22:05 - INFO - codeparrot_training - Step 17559: {'lr': 0.0003811961731619057, 'samples': 3371520, 'steps': 17559, 'loss/train': 0.759276419878006} +01/27/2022 12:22:08 - INFO - codeparrot_training - Step 17560: {'lr': 0.0003811822445807863, 'samples': 3371712, 'steps': 17560, 'loss/train': 0.9370661973953247} +01/27/2022 12:22:11 - INFO - codeparrot_training - Step 17561: {'lr': 0.00038116831543772377, 'samples': 3371904, 'steps': 17561, 'loss/train': 0.4099772125482559} +01/27/2022 12:22:14 - INFO - codeparrot_training - Step 17562: {'lr': 0.00038115438573277784, 'samples': 3372096, 'steps': 17562, 'loss/train': 0.9384652376174927} +01/27/2022 12:22:18 - INFO - codeparrot_training - Step 17563: {'lr': 0.0003811404554660082, 'samples': 3372288, 'steps': 17563, 'loss/train': 0.5200569480657578} +01/27/2022 12:22:22 - INFO - codeparrot_training - Step 17564: {'lr': 0.00038112652463747444, 'samples': 3372480, 'steps': 17564, 'loss/train': 0.8636096119880676} +01/27/2022 12:22:25 - INFO - codeparrot_training - Step 17565: {'lr': 0.00038111259324723624, 'samples': 3372672, 'steps': 17565, 'loss/train': 0.5550636649131775} +01/27/2022 12:22:28 - INFO - codeparrot_training - Step 17566: {'lr': 0.0003810986612953534, 'samples': 3372864, 'steps': 17566, 'loss/train': 0.6710767149925232} +01/27/2022 12:22:31 - INFO - codeparrot_training - Step 17567: {'lr': 0.0003810847287818855, 'samples': 3373056, 'steps': 17567, 'loss/train': 0.7653064727783203} +01/27/2022 12:22:35 - INFO - codeparrot_training - Step 17568: {'lr': 0.0003810707957068923, 'samples': 3373248, 'steps': 17568, 'loss/train': 0.6420416235923767} +01/27/2022 12:22:38 - INFO - codeparrot_training - Step 17569: {'lr': 0.0003810568620704334, 'samples': 3373440, 'steps': 17569, 'loss/train': 0.7790663838386536} +01/27/2022 12:22:41 - INFO - codeparrot_training - Step 17570: {'lr': 0.00038104292787256844, 'samples': 3373632, 'steps': 17570, 'loss/train': 1.198676437139511} +01/27/2022 12:22:44 - INFO - codeparrot_training - Step 17571: {'lr': 0.0003810289931133573, 'samples': 3373824, 'steps': 17571, 'loss/train': 0.3762316256761551} +01/27/2022 12:22:47 - INFO - codeparrot_training - Step 17572: {'lr': 0.0003810150577928595, 'samples': 3374016, 'steps': 17572, 'loss/train': 1.0040989816188812} +01/27/2022 12:22:53 - INFO - codeparrot_training - Step 17573: {'lr': 0.0003810011219111348, 'samples': 3374208, 'steps': 17573, 'loss/train': 1.1294004321098328} +01/27/2022 12:22:56 - INFO - codeparrot_training - Step 17574: {'lr': 0.00038098718546824287, 'samples': 3374400, 'steps': 17574, 'loss/train': 0.9147258102893829} +01/27/2022 12:22:59 - INFO - codeparrot_training - Step 17575: {'lr': 0.00038097324846424354, 'samples': 3374592, 'steps': 17575, 'loss/train': 0.398577556014061} +01/27/2022 12:23:03 - INFO - codeparrot_training - Step 17576: {'lr': 0.0003809593108991962, 'samples': 3374784, 'steps': 17576, 'loss/train': 0.9043717682361603} +01/27/2022 12:23:06 - INFO - codeparrot_training - Step 17577: {'lr': 0.0003809453727731609, 'samples': 3374976, 'steps': 17577, 'loss/train': 0.7413520663976669} +01/27/2022 12:23:09 - INFO - codeparrot_training - Step 17578: {'lr': 0.00038093143408619726, 'samples': 3375168, 'steps': 17578, 'loss/train': 0.9380393922328949} +01/27/2022 12:23:12 - INFO - codeparrot_training - Step 17579: {'lr': 0.0003809174948383648, 'samples': 3375360, 'steps': 17579, 'loss/train': 0.48193618655204773} +01/27/2022 12:23:15 - INFO - codeparrot_training - Step 17580: {'lr': 0.0003809035550297234, 'samples': 3375552, 'steps': 17580, 'loss/train': 0.7141970247030258} +01/27/2022 12:23:18 - INFO - codeparrot_training - Step 17581: {'lr': 0.00038088961466033276, 'samples': 3375744, 'steps': 17581, 'loss/train': 1.3027331829071045} +01/27/2022 12:23:23 - INFO - codeparrot_training - Step 17582: {'lr': 0.00038087567373025255, 'samples': 3375936, 'steps': 17582, 'loss/train': 0.8489871919155121} +01/27/2022 12:23:26 - INFO - codeparrot_training - Step 17583: {'lr': 0.0003808617322395425, 'samples': 3376128, 'steps': 17583, 'loss/train': 0.599931076169014} +01/27/2022 12:23:29 - INFO - codeparrot_training - Step 17584: {'lr': 0.00038084779018826245, 'samples': 3376320, 'steps': 17584, 'loss/train': 0.9065417647361755} +01/27/2022 12:23:32 - INFO - codeparrot_training - Step 17585: {'lr': 0.00038083384757647186, 'samples': 3376512, 'steps': 17585, 'loss/train': 0.9769366085529327} +01/27/2022 12:23:36 - INFO - codeparrot_training - Step 17586: {'lr': 0.0003808199044042308, 'samples': 3376704, 'steps': 17586, 'loss/train': 0.7999990582466125} +01/27/2022 12:23:39 - INFO - codeparrot_training - Step 17587: {'lr': 0.00038080596067159865, 'samples': 3376896, 'steps': 17587, 'loss/train': 0.5833088010549545} +01/27/2022 12:23:42 - INFO - codeparrot_training - Step 17588: {'lr': 0.0003807920163786353, 'samples': 3377088, 'steps': 17588, 'loss/train': 1.1053553223609924} +01/27/2022 12:23:45 - INFO - codeparrot_training - Step 17589: {'lr': 0.0003807780715254006, 'samples': 3377280, 'steps': 17589, 'loss/train': 1.0071618854999542} +01/27/2022 12:23:52 - INFO - codeparrot_training - Step 17590: {'lr': 0.000380764126111954, 'samples': 3377472, 'steps': 17590, 'loss/train': 0.7911767363548279} +01/27/2022 12:23:55 - INFO - codeparrot_training - Step 17591: {'lr': 0.0003807501801383555, 'samples': 3377664, 'steps': 17591, 'loss/train': 0.935079038143158} +01/27/2022 12:23:58 - INFO - codeparrot_training - Step 17592: {'lr': 0.0003807362336046648, 'samples': 3377856, 'steps': 17592, 'loss/train': 1.7963011264801025} +01/27/2022 12:24:01 - INFO - codeparrot_training - Step 17593: {'lr': 0.00038072228651094155, 'samples': 3378048, 'steps': 17593, 'loss/train': 0.6747234910726547} +01/27/2022 12:24:04 - INFO - codeparrot_training - Step 17594: {'lr': 0.0003807083388572455, 'samples': 3378240, 'steps': 17594, 'loss/train': 0.7520534098148346} +01/27/2022 12:24:07 - INFO - codeparrot_training - Step 17595: {'lr': 0.0003806943906436364, 'samples': 3378432, 'steps': 17595, 'loss/train': 0.8757303357124329} +01/27/2022 12:24:10 - INFO - codeparrot_training - Step 17596: {'lr': 0.0003806804418701741, 'samples': 3378624, 'steps': 17596, 'loss/train': 0.901152491569519} +01/27/2022 12:24:14 - INFO - codeparrot_training - Step 17597: {'lr': 0.0003806664925369183, 'samples': 3378816, 'steps': 17597, 'loss/train': 0.6177244484424591} +01/27/2022 12:24:17 - INFO - codeparrot_training - Step 17598: {'lr': 0.0003806525426439287, 'samples': 3379008, 'steps': 17598, 'loss/train': 0.42204125225543976} +01/27/2022 12:24:21 - INFO - codeparrot_training - Step 17599: {'lr': 0.00038063859219126514, 'samples': 3379200, 'steps': 17599, 'loss/train': 0.8274960815906525} +01/27/2022 12:24:24 - INFO - codeparrot_training - Step 17600: {'lr': 0.0003806246411789872, 'samples': 3379392, 'steps': 17600, 'loss/train': 0.6486605554819107} +01/27/2022 12:24:27 - INFO - codeparrot_training - Step 17601: {'lr': 0.00038061068960715494, 'samples': 3379584, 'steps': 17601, 'loss/train': 0.8894058465957642} +01/27/2022 12:24:31 - INFO - codeparrot_training - Step 17602: {'lr': 0.00038059673747582783, 'samples': 3379776, 'steps': 17602, 'loss/train': 0.5349358767271042} +01/27/2022 12:24:34 - INFO - codeparrot_training - Step 17603: {'lr': 0.00038058278478506584, 'samples': 3379968, 'steps': 17603, 'loss/train': 0.5744879096746445} +01/27/2022 12:24:37 - INFO - codeparrot_training - Step 17604: {'lr': 0.0003805688315349286, 'samples': 3380160, 'steps': 17604, 'loss/train': 0.4258686304092407} +01/27/2022 12:24:40 - INFO - codeparrot_training - Step 17605: {'lr': 0.00038055487772547603, 'samples': 3380352, 'steps': 17605, 'loss/train': 0.8670195043087006} +01/27/2022 12:24:43 - INFO - codeparrot_training - Step 17606: {'lr': 0.00038054092335676774, 'samples': 3380544, 'steps': 17606, 'loss/train': 0.5635651499032974} +01/27/2022 12:24:46 - INFO - codeparrot_training - Step 17607: {'lr': 0.00038052696842886364, 'samples': 3380736, 'steps': 17607, 'loss/train': 0.7464081197977066} +01/27/2022 12:24:51 - INFO - codeparrot_training - Step 17608: {'lr': 0.0003805130129418235, 'samples': 3380928, 'steps': 17608, 'loss/train': 0.7785082161426544} +01/27/2022 12:24:54 - INFO - codeparrot_training - Step 17609: {'lr': 0.00038049905689570697, 'samples': 3381120, 'steps': 17609, 'loss/train': 0.9789430797100067} +01/27/2022 12:24:57 - INFO - codeparrot_training - Step 17610: {'lr': 0.00038048510029057393, 'samples': 3381312, 'steps': 17610, 'loss/train': 1.1393182575702667} +01/27/2022 12:25:00 - INFO - codeparrot_training - Step 17611: {'lr': 0.00038047114312648414, 'samples': 3381504, 'steps': 17611, 'loss/train': 0.9507156014442444} +01/27/2022 12:25:03 - INFO - codeparrot_training - Step 17612: {'lr': 0.0003804571854034975, 'samples': 3381696, 'steps': 17612, 'loss/train': 0.7753790616989136} +01/27/2022 12:25:06 - INFO - codeparrot_training - Step 17613: {'lr': 0.0003804432271216736, 'samples': 3381888, 'steps': 17613, 'loss/train': 0.6945864111185074} +01/27/2022 12:25:10 - INFO - codeparrot_training - Step 17614: {'lr': 0.0003804292682810724, 'samples': 3382080, 'steps': 17614, 'loss/train': 0.8680836260318756} +01/27/2022 12:25:13 - INFO - codeparrot_training - Step 17615: {'lr': 0.00038041530888175356, 'samples': 3382272, 'steps': 17615, 'loss/train': 0.8822807371616364} +01/27/2022 12:25:16 - INFO - codeparrot_training - Step 17616: {'lr': 0.00038040134892377696, 'samples': 3382464, 'steps': 17616, 'loss/train': 0.7911579608917236} +01/27/2022 12:25:23 - INFO - codeparrot_training - Step 17617: {'lr': 0.00038038738840720244, 'samples': 3382656, 'steps': 17617, 'loss/train': 1.108813852071762} +01/27/2022 12:25:26 - INFO - codeparrot_training - Step 17618: {'lr': 0.0003803734273320897, 'samples': 3382848, 'steps': 17618, 'loss/train': 1.0023580491542816} +01/27/2022 12:25:29 - INFO - codeparrot_training - Step 17619: {'lr': 0.0003803594656984986, 'samples': 3383040, 'steps': 17619, 'loss/train': 0.7388640940189362} +01/27/2022 12:25:32 - INFO - codeparrot_training - Step 17620: {'lr': 0.000380345503506489, 'samples': 3383232, 'steps': 17620, 'loss/train': 0.8199170529842377} +01/27/2022 12:25:35 - INFO - codeparrot_training - Step 17621: {'lr': 0.00038033154075612063, 'samples': 3383424, 'steps': 17621, 'loss/train': 0.9106069207191467} +01/27/2022 12:25:38 - INFO - codeparrot_training - Step 17622: {'lr': 0.00038031757744745327, 'samples': 3383616, 'steps': 17622, 'loss/train': 1.0614877045154572} +01/27/2022 12:25:42 - INFO - codeparrot_training - Step 17623: {'lr': 0.0003803036135805469, 'samples': 3383808, 'steps': 17623, 'loss/train': 0.706055760383606} +01/27/2022 12:25:45 - INFO - codeparrot_training - Step 17624: {'lr': 0.00038028964915546107, 'samples': 3384000, 'steps': 17624, 'loss/train': 0.8001103699207306} +01/27/2022 12:25:49 - INFO - codeparrot_training - Step 17625: {'lr': 0.00038027568417225586, 'samples': 3384192, 'steps': 17625, 'loss/train': 0.7546382546424866} +01/27/2022 12:25:52 - INFO - codeparrot_training - Step 17626: {'lr': 0.00038026171863099093, 'samples': 3384384, 'steps': 17626, 'loss/train': 0.7697919309139252} +01/27/2022 12:25:56 - INFO - codeparrot_training - Step 17627: {'lr': 0.0003802477525317263, 'samples': 3384576, 'steps': 17627, 'loss/train': 0.9213748276233673} +01/27/2022 12:25:59 - INFO - codeparrot_training - Step 17628: {'lr': 0.00038023378587452144, 'samples': 3384768, 'steps': 17628, 'loss/train': 0.7434128075838089} +01/27/2022 12:26:02 - INFO - codeparrot_training - Step 17629: {'lr': 0.0003802198186594366, 'samples': 3384960, 'steps': 17629, 'loss/train': 0.8375399708747864} +01/27/2022 12:26:05 - INFO - codeparrot_training - Step 17630: {'lr': 0.00038020585088653126, 'samples': 3385152, 'steps': 17630, 'loss/train': 0.7644470930099487} +01/27/2022 12:26:08 - INFO - codeparrot_training - Step 17631: {'lr': 0.00038019188255586546, 'samples': 3385344, 'steps': 17631, 'loss/train': 0.4997047930955887} +01/27/2022 12:26:11 - INFO - codeparrot_training - Step 17632: {'lr': 0.00038017791366749896, 'samples': 3385536, 'steps': 17632, 'loss/train': 0.8453731834888458} +01/27/2022 12:26:14 - INFO - codeparrot_training - Step 17633: {'lr': 0.0003801639442214916, 'samples': 3385728, 'steps': 17633, 'loss/train': 1.596295416355133} +01/27/2022 12:26:21 - INFO - codeparrot_training - Step 17634: {'lr': 0.0003801499742179033, 'samples': 3385920, 'steps': 17634, 'loss/train': 0.7018159478902817} +01/27/2022 12:26:24 - INFO - codeparrot_training - Step 17635: {'lr': 0.0003801360036567938, 'samples': 3386112, 'steps': 17635, 'loss/train': 1.0462090373039246} +01/27/2022 12:26:27 - INFO - codeparrot_training - Step 17636: {'lr': 0.000380122032538223, 'samples': 3386304, 'steps': 17636, 'loss/train': 0.40242840349674225} +01/27/2022 12:26:30 - INFO - codeparrot_training - Step 17637: {'lr': 0.0003801080608622507, 'samples': 3386496, 'steps': 17637, 'loss/train': 0.5320856720209122} +01/27/2022 12:26:33 - INFO - codeparrot_training - Step 17638: {'lr': 0.0003800940886289368, 'samples': 3386688, 'steps': 17638, 'loss/train': 0.7203127294778824} +01/27/2022 12:26:36 - INFO - codeparrot_training - Step 17639: {'lr': 0.0003800801158383411, 'samples': 3386880, 'steps': 17639, 'loss/train': 0.4303123354911804} +01/27/2022 12:26:40 - INFO - codeparrot_training - Step 17640: {'lr': 0.00038006614249052353, 'samples': 3387072, 'steps': 17640, 'loss/train': 0.8024492561817169} +01/27/2022 12:26:43 - INFO - codeparrot_training - Step 17641: {'lr': 0.0003800521685855439, 'samples': 3387264, 'steps': 17641, 'loss/train': 1.227532535791397} +01/27/2022 12:26:46 - INFO - codeparrot_training - Step 17642: {'lr': 0.000380038194123462, 'samples': 3387456, 'steps': 17642, 'loss/train': 0.6158351451158524} +01/27/2022 12:26:50 - INFO - codeparrot_training - Step 17643: {'lr': 0.0003800242191043379, 'samples': 3387648, 'steps': 17643, 'loss/train': 1.0035163164138794} +01/27/2022 12:26:53 - INFO - codeparrot_training - Step 17644: {'lr': 0.00038001024352823123, 'samples': 3387840, 'steps': 17644, 'loss/train': 2.6012887358665466} +01/27/2022 12:26:56 - INFO - codeparrot_training - Step 17645: {'lr': 0.00037999626739520197, 'samples': 3388032, 'steps': 17645, 'loss/train': 1.234763503074646} +01/27/2022 12:27:00 - INFO - codeparrot_training - Step 17646: {'lr': 0.00037998229070531, 'samples': 3388224, 'steps': 17646, 'loss/train': 1.2936771512031555} +01/27/2022 12:27:03 - INFO - codeparrot_training - Step 17647: {'lr': 0.0003799683134586152, 'samples': 3388416, 'steps': 17647, 'loss/train': 0.7855320274829865} +01/27/2022 12:27:06 - INFO - codeparrot_training - Step 17648: {'lr': 0.0003799543356551773, 'samples': 3388608, 'steps': 17648, 'loss/train': 0.6672200113534927} +01/27/2022 12:27:09 - INFO - codeparrot_training - Step 17649: {'lr': 0.0003799403572950565, 'samples': 3388800, 'steps': 17649, 'loss/train': 0.6547168493270874} +01/27/2022 12:27:12 - INFO - codeparrot_training - Step 17650: {'lr': 0.00037992637837831235, 'samples': 3388992, 'steps': 17650, 'loss/train': 1.073117583990097} +01/27/2022 12:27:17 - INFO - codeparrot_training - Step 17651: {'lr': 0.00037991239890500483, 'samples': 3389184, 'steps': 17651, 'loss/train': 0.43304094672203064} +01/27/2022 12:27:20 - INFO - codeparrot_training - Step 17652: {'lr': 0.00037989841887519385, 'samples': 3389376, 'steps': 17652, 'loss/train': 0.5141791999340057} +01/27/2022 12:27:23 - INFO - codeparrot_training - Step 17653: {'lr': 0.00037988443828893936, 'samples': 3389568, 'steps': 17653, 'loss/train': 0.8803211152553558} +01/27/2022 12:27:26 - INFO - codeparrot_training - Step 17654: {'lr': 0.0003798704571463011, 'samples': 3389760, 'steps': 17654, 'loss/train': 0.9142555296421051} +01/27/2022 12:27:29 - INFO - codeparrot_training - Step 17655: {'lr': 0.00037985647544733903, 'samples': 3389952, 'steps': 17655, 'loss/train': 1.0644122064113617} +01/27/2022 12:27:33 - INFO - codeparrot_training - Step 17656: {'lr': 0.0003798424931921131, 'samples': 3390144, 'steps': 17656, 'loss/train': 0.7376801669597626} +01/27/2022 12:27:36 - INFO - codeparrot_training - Step 17657: {'lr': 0.0003798285103806831, 'samples': 3390336, 'steps': 17657, 'loss/train': 0.8549390137195587} +01/27/2022 12:27:39 - INFO - codeparrot_training - Step 17658: {'lr': 0.0003798145270131091, 'samples': 3390528, 'steps': 17658, 'loss/train': 0.39151203632354736} +01/27/2022 12:27:42 - INFO - codeparrot_training - Step 17659: {'lr': 0.00037980054308945076, 'samples': 3390720, 'steps': 17659, 'loss/train': 1.1263719499111176} +01/27/2022 12:27:48 - INFO - codeparrot_training - Step 17660: {'lr': 0.00037978655860976826, 'samples': 3390912, 'steps': 17660, 'loss/train': 1.106958121061325} +01/27/2022 12:27:51 - INFO - codeparrot_training - Step 17661: {'lr': 0.0003797725735741212, 'samples': 3391104, 'steps': 17661, 'loss/train': 0.5523396581411362} +01/27/2022 12:27:54 - INFO - codeparrot_training - Step 17662: {'lr': 0.0003797585879825698, 'samples': 3391296, 'steps': 17662, 'loss/train': 0.8923992812633514} +01/27/2022 12:27:57 - INFO - codeparrot_training - Step 17663: {'lr': 0.00037974460183517366, 'samples': 3391488, 'steps': 17663, 'loss/train': 0.6943888664245605} +01/27/2022 12:28:00 - INFO - codeparrot_training - Step 17664: {'lr': 0.0003797306151319929, 'samples': 3391680, 'steps': 17664, 'loss/train': 0.6808324903249741} +01/27/2022 12:28:03 - INFO - codeparrot_training - Step 17665: {'lr': 0.00037971662787308734, 'samples': 3391872, 'steps': 17665, 'loss/train': 1.6969846487045288} +01/27/2022 12:28:07 - INFO - codeparrot_training - Step 17666: {'lr': 0.00037970264005851703, 'samples': 3392064, 'steps': 17666, 'loss/train': 1.025417000055313} +01/27/2022 12:28:10 - INFO - codeparrot_training - Step 17667: {'lr': 0.0003796886516883418, 'samples': 3392256, 'steps': 17667, 'loss/train': 2.6182313561439514} +01/27/2022 12:28:13 - INFO - codeparrot_training - Step 17668: {'lr': 0.0003796746627626214, 'samples': 3392448, 'steps': 17668, 'loss/train': 0.5651771575212479} +01/27/2022 12:28:18 - INFO - codeparrot_training - Step 17669: {'lr': 0.00037966067328141606, 'samples': 3392640, 'steps': 17669, 'loss/train': 0.7804397642612457} +01/27/2022 12:28:21 - INFO - codeparrot_training - Step 17670: {'lr': 0.0003796466832447856, 'samples': 3392832, 'steps': 17670, 'loss/train': 0.45675739645957947} +01/27/2022 12:28:24 - INFO - codeparrot_training - Step 17671: {'lr': 0.00037963269265278986, 'samples': 3393024, 'steps': 17671, 'loss/train': 0.6983577758073807} +01/27/2022 12:28:27 - INFO - codeparrot_training - Step 17672: {'lr': 0.0003796187015054888, 'samples': 3393216, 'steps': 17672, 'loss/train': 1.200396180152893} +01/27/2022 12:28:30 - INFO - codeparrot_training - Step 17673: {'lr': 0.0003796047098029424, 'samples': 3393408, 'steps': 17673, 'loss/train': 0.9162155091762543} +01/27/2022 12:28:33 - INFO - codeparrot_training - Step 17674: {'lr': 0.0003795907175452106, 'samples': 3393600, 'steps': 17674, 'loss/train': 1.4042057991027832} +01/27/2022 12:28:36 - INFO - codeparrot_training - Step 17675: {'lr': 0.0003795767247323533, 'samples': 3393792, 'steps': 17675, 'loss/train': 1.6248448491096497} +01/27/2022 12:28:40 - INFO - codeparrot_training - Step 17676: {'lr': 0.00037956273136443056, 'samples': 3393984, 'steps': 17676, 'loss/train': 0.7945581078529358} +01/27/2022 12:28:43 - INFO - codeparrot_training - Step 17677: {'lr': 0.000379548737441502, 'samples': 3394176, 'steps': 17677, 'loss/train': 0.3918060064315796} +01/27/2022 12:28:48 - INFO - codeparrot_training - Step 17678: {'lr': 0.00037953474296362796, 'samples': 3394368, 'steps': 17678, 'loss/train': 1.3041239082813263} +01/27/2022 12:28:51 - INFO - codeparrot_training - Step 17679: {'lr': 0.0003795207479308681, 'samples': 3394560, 'steps': 17679, 'loss/train': 1.0668982565402985} +01/27/2022 12:28:54 - INFO - codeparrot_training - Step 17680: {'lr': 0.00037950675234328256, 'samples': 3394752, 'steps': 17680, 'loss/train': 0.4474448561668396} +01/27/2022 12:28:57 - INFO - codeparrot_training - Step 17681: {'lr': 0.00037949275620093124, 'samples': 3394944, 'steps': 17681, 'loss/train': 0.5069480091333389} +01/27/2022 12:29:00 - INFO - codeparrot_training - Step 17682: {'lr': 0.000379478759503874, 'samples': 3395136, 'steps': 17682, 'loss/train': 1.1141667366027832} +01/27/2022 12:29:03 - INFO - codeparrot_training - Step 17683: {'lr': 0.00037946476225217087, 'samples': 3395328, 'steps': 17683, 'loss/train': 0.39164958894252777} +01/27/2022 12:29:06 - INFO - codeparrot_training - Step 17684: {'lr': 0.0003794507644458819, 'samples': 3395520, 'steps': 17684, 'loss/train': 1.0509098768234253} +01/27/2022 12:29:10 - INFO - codeparrot_training - Step 17685: {'lr': 0.00037943676608506683, 'samples': 3395712, 'steps': 17685, 'loss/train': 1.186167687177658} +01/27/2022 12:29:14 - INFO - codeparrot_training - Step 17686: {'lr': 0.00037942276716978584, 'samples': 3395904, 'steps': 17686, 'loss/train': 1.0408623218536377} +01/27/2022 12:29:17 - INFO - codeparrot_training - Step 17687: {'lr': 0.0003794087677000988, 'samples': 3396096, 'steps': 17687, 'loss/train': 0.8815032541751862} +01/27/2022 12:29:20 - INFO - codeparrot_training - Step 17688: {'lr': 0.0003793947676760657, 'samples': 3396288, 'steps': 17688, 'loss/train': 1.6696629524230957} +01/27/2022 12:29:24 - INFO - codeparrot_training - Step 17689: {'lr': 0.00037938076709774645, 'samples': 3396480, 'steps': 17689, 'loss/train': 0.6941552013158798} +01/27/2022 12:29:27 - INFO - codeparrot_training - Step 17690: {'lr': 0.0003793667659652011, 'samples': 3396672, 'steps': 17690, 'loss/train': 0.9092665314674377} +01/27/2022 12:29:30 - INFO - codeparrot_training - Step 17691: {'lr': 0.0003793527642784896, 'samples': 3396864, 'steps': 17691, 'loss/train': 0.7028269171714783} +01/27/2022 12:29:33 - INFO - codeparrot_training - Step 17692: {'lr': 0.0003793387620376719, 'samples': 3397056, 'steps': 17692, 'loss/train': 1.2979472279548645} +01/27/2022 12:29:36 - INFO - codeparrot_training - Step 17693: {'lr': 0.0003793247592428081, 'samples': 3397248, 'steps': 17693, 'loss/train': 1.18478125333786} +01/27/2022 12:29:39 - INFO - codeparrot_training - Step 17694: {'lr': 0.00037931075589395805, 'samples': 3397440, 'steps': 17694, 'loss/train': 0.37494785338640213} +01/27/2022 12:29:45 - INFO - codeparrot_training - Step 17695: {'lr': 0.00037929675199118183, 'samples': 3397632, 'steps': 17695, 'loss/train': 0.6255338788032532} +01/27/2022 12:29:48 - INFO - codeparrot_training - Step 17696: {'lr': 0.0003792827475345393, 'samples': 3397824, 'steps': 17696, 'loss/train': 0.731553465127945} +01/27/2022 12:29:51 - INFO - codeparrot_training - Step 17697: {'lr': 0.0003792687425240906, 'samples': 3398016, 'steps': 17697, 'loss/train': 1.0597515106201172} +01/27/2022 12:29:54 - INFO - codeparrot_training - Step 17698: {'lr': 0.0003792547369598956, 'samples': 3398208, 'steps': 17698, 'loss/train': 0.6732043772935867} +01/27/2022 12:29:57 - INFO - codeparrot_training - Step 17699: {'lr': 0.0003792407308420144, 'samples': 3398400, 'steps': 17699, 'loss/train': 1.070304036140442} +01/27/2022 12:30:01 - INFO - codeparrot_training - Step 17700: {'lr': 0.00037922672417050685, 'samples': 3398592, 'steps': 17700, 'loss/train': 0.8304772675037384} +01/27/2022 12:30:04 - INFO - codeparrot_training - Step 17701: {'lr': 0.00037921271694543317, 'samples': 3398784, 'steps': 17701, 'loss/train': 1.18225559592247} +01/27/2022 12:30:07 - INFO - codeparrot_training - Step 17702: {'lr': 0.0003791987091668532, 'samples': 3398976, 'steps': 17702, 'loss/train': 0.9208369553089142} +01/27/2022 12:30:10 - INFO - codeparrot_training - Step 17703: {'lr': 0.00037918470083482693, 'samples': 3399168, 'steps': 17703, 'loss/train': 1.2328450977802277} +01/27/2022 12:30:14 - INFO - codeparrot_training - Step 17704: {'lr': 0.0003791706919494145, 'samples': 3399360, 'steps': 17704, 'loss/train': 0.9051245748996735} +01/27/2022 12:30:18 - INFO - codeparrot_training - Step 17705: {'lr': 0.0003791566825106758, 'samples': 3399552, 'steps': 17705, 'loss/train': 0.769343376159668} +01/27/2022 12:30:21 - INFO - codeparrot_training - Step 17706: {'lr': 0.0003791426725186709, 'samples': 3399744, 'steps': 17706, 'loss/train': 0.621571496129036} +01/27/2022 12:30:24 - INFO - codeparrot_training - Step 17707: {'lr': 0.0003791286619734597, 'samples': 3399936, 'steps': 17707, 'loss/train': 0.4595901221036911} +01/27/2022 12:30:27 - INFO - codeparrot_training - Step 17708: {'lr': 0.0003791146508751025, 'samples': 3400128, 'steps': 17708, 'loss/train': 0.8479340672492981} +01/27/2022 12:30:30 - INFO - codeparrot_training - Step 17709: {'lr': 0.00037910063922365903, 'samples': 3400320, 'steps': 17709, 'loss/train': 1.076926052570343} +01/27/2022 12:30:33 - INFO - codeparrot_training - Step 17710: {'lr': 0.00037908662701918944, 'samples': 3400512, 'steps': 17710, 'loss/train': 0.6719412803649902} +01/27/2022 12:30:36 - INFO - codeparrot_training - Step 17711: {'lr': 0.00037907261426175365, 'samples': 3400704, 'steps': 17711, 'loss/train': 0.6751941740512848} +01/27/2022 12:30:40 - INFO - codeparrot_training - Step 17712: {'lr': 0.0003790586009514119, 'samples': 3400896, 'steps': 17712, 'loss/train': 0.9057995975017548} +01/27/2022 12:30:45 - INFO - codeparrot_training - Step 17713: {'lr': 0.000379044587088224, 'samples': 3401088, 'steps': 17713, 'loss/train': 0.6229561418294907} +01/27/2022 12:30:48 - INFO - codeparrot_training - Step 17714: {'lr': 0.0003790305726722501, 'samples': 3401280, 'steps': 17714, 'loss/train': 0.8434931337833405} +01/27/2022 12:30:52 - INFO - codeparrot_training - Step 17715: {'lr': 0.00037901655770355015, 'samples': 3401472, 'steps': 17715, 'loss/train': 1.2113399505615234} +01/27/2022 12:30:55 - INFO - codeparrot_training - Step 17716: {'lr': 0.0003790025421821843, 'samples': 3401664, 'steps': 17716, 'loss/train': 0.7394381761550903} +01/27/2022 12:30:58 - INFO - codeparrot_training - Step 17717: {'lr': 0.0003789885261082124, 'samples': 3401856, 'steps': 17717, 'loss/train': 0.28448040783405304} +01/27/2022 12:31:01 - INFO - codeparrot_training - Step 17718: {'lr': 0.00037897450948169476, 'samples': 3402048, 'steps': 17718, 'loss/train': 1.1871600151062012} +01/27/2022 12:31:04 - INFO - codeparrot_training - Step 17719: {'lr': 0.0003789604923026912, 'samples': 3402240, 'steps': 17719, 'loss/train': 0.1921374723315239} +01/27/2022 12:31:07 - INFO - codeparrot_training - Step 17720: {'lr': 0.00037894647457126186, 'samples': 3402432, 'steps': 17720, 'loss/train': 0.21144280582666397} +01/27/2022 12:31:10 - INFO - codeparrot_training - Step 17721: {'lr': 0.0003789324562874668, 'samples': 3402624, 'steps': 17721, 'loss/train': 0.2824966236948967} +01/27/2022 12:31:15 - INFO - codeparrot_training - Step 17722: {'lr': 0.000378918437451366, 'samples': 3402816, 'steps': 17722, 'loss/train': 1.2950773537158966} +01/27/2022 12:31:18 - INFO - codeparrot_training - Step 17723: {'lr': 0.00037890441806301954, 'samples': 3403008, 'steps': 17723, 'loss/train': 0.850862592458725} +01/27/2022 12:31:21 - INFO - codeparrot_training - Step 17724: {'lr': 0.0003788903981224875, 'samples': 3403200, 'steps': 17724, 'loss/train': 1.2488656640052795} +01/27/2022 12:31:24 - INFO - codeparrot_training - Step 17725: {'lr': 0.00037887637762982996, 'samples': 3403392, 'steps': 17725, 'loss/train': 0.6582675576210022} +01/27/2022 12:31:27 - INFO - codeparrot_training - Step 17726: {'lr': 0.0003788623565851068, 'samples': 3403584, 'steps': 17726, 'loss/train': 0.7107021063566208} +01/27/2022 12:31:31 - INFO - codeparrot_training - Step 17727: {'lr': 0.00037884833498837833, 'samples': 3403776, 'steps': 17727, 'loss/train': 0.5869897753000259} +01/27/2022 12:31:34 - INFO - codeparrot_training - Step 17728: {'lr': 0.00037883431283970454, 'samples': 3403968, 'steps': 17728, 'loss/train': 0.29002057760953903} +01/27/2022 12:31:37 - INFO - codeparrot_training - Step 17729: {'lr': 0.00037882029013914544, 'samples': 3404160, 'steps': 17729, 'loss/train': 0.5466892719268799} +01/27/2022 12:31:41 - INFO - codeparrot_training - Step 17730: {'lr': 0.0003788062668867611, 'samples': 3404352, 'steps': 17730, 'loss/train': 0.41187572479248047} +01/27/2022 12:31:44 - INFO - codeparrot_training - Step 17731: {'lr': 0.00037879224308261163, 'samples': 3404544, 'steps': 17731, 'loss/train': 0.934820830821991} +01/27/2022 12:31:48 - INFO - codeparrot_training - Step 17732: {'lr': 0.00037877821872675705, 'samples': 3404736, 'steps': 17732, 'loss/train': 1.040990173816681} +01/27/2022 12:31:51 - INFO - codeparrot_training - Step 17733: {'lr': 0.0003787641938192575, 'samples': 3404928, 'steps': 17733, 'loss/train': 0.78511181473732} +01/27/2022 12:31:54 - INFO - codeparrot_training - Step 17734: {'lr': 0.00037875016836017304, 'samples': 3405120, 'steps': 17734, 'loss/train': 0.7643394470214844} +01/27/2022 12:31:57 - INFO - codeparrot_training - Step 17735: {'lr': 0.0003787361423495637, 'samples': 3405312, 'steps': 17735, 'loss/train': 0.06435721926391125} +01/27/2022 12:32:00 - INFO - codeparrot_training - Step 17736: {'lr': 0.0003787221157874897, 'samples': 3405504, 'steps': 17736, 'loss/train': 1.6632746458053589} +01/27/2022 12:32:03 - INFO - codeparrot_training - Step 17737: {'lr': 0.00037870808867401085, 'samples': 3405696, 'steps': 17737, 'loss/train': 0.923375129699707} +01/27/2022 12:32:06 - INFO - codeparrot_training - Step 17738: {'lr': 0.00037869406100918756, 'samples': 3405888, 'steps': 17738, 'loss/train': 0.9484435021877289} +01/27/2022 12:32:12 - INFO - codeparrot_training - Step 17739: {'lr': 0.0003786800327930797, 'samples': 3406080, 'steps': 17739, 'loss/train': 0.32493162900209427} +01/27/2022 12:32:15 - INFO - codeparrot_training - Step 17740: {'lr': 0.0003786660040257475, 'samples': 3406272, 'steps': 17740, 'loss/train': 0.8537809252738953} +01/27/2022 12:32:18 - INFO - codeparrot_training - Step 17741: {'lr': 0.00037865197470725103, 'samples': 3406464, 'steps': 17741, 'loss/train': 0.8624403476715088} +01/27/2022 12:32:22 - INFO - codeparrot_training - Step 17742: {'lr': 0.0003786379448376503, 'samples': 3406656, 'steps': 17742, 'loss/train': 1.2402792870998383} +01/27/2022 12:32:25 - INFO - codeparrot_training - Step 17743: {'lr': 0.0003786239144170055, 'samples': 3406848, 'steps': 17743, 'loss/train': 0.645303949713707} +01/27/2022 12:32:28 - INFO - codeparrot_training - Step 17744: {'lr': 0.0003786098834453766, 'samples': 3407040, 'steps': 17744, 'loss/train': 0.6766538769006729} +01/27/2022 12:32:31 - INFO - codeparrot_training - Step 17745: {'lr': 0.00037859585192282386, 'samples': 3407232, 'steps': 17745, 'loss/train': 1.2060854136943817} +01/27/2022 12:32:34 - INFO - codeparrot_training - Step 17746: {'lr': 0.00037858181984940734, 'samples': 3407424, 'steps': 17746, 'loss/train': 0.17646407335996628} +01/27/2022 12:32:37 - INFO - codeparrot_training - Step 17747: {'lr': 0.0003785677872251871, 'samples': 3407616, 'steps': 17747, 'loss/train': 0.7733698189258575} +01/27/2022 12:32:42 - INFO - codeparrot_training - Step 17748: {'lr': 0.0003785537540502233, 'samples': 3407808, 'steps': 17748, 'loss/train': 1.0578591227531433} +01/27/2022 12:32:45 - INFO - codeparrot_training - Step 17749: {'lr': 0.0003785397203245761, 'samples': 3408000, 'steps': 17749, 'loss/train': 0.7659654915332794} +01/27/2022 12:32:48 - INFO - codeparrot_training - Step 17750: {'lr': 0.0003785256860483054, 'samples': 3408192, 'steps': 17750, 'loss/train': 1.178657054901123} +01/27/2022 12:32:51 - INFO - codeparrot_training - Step 17751: {'lr': 0.0003785116512214716, 'samples': 3408384, 'steps': 17751, 'loss/train': 1.1738046705722809} +01/27/2022 12:32:54 - INFO - codeparrot_training - Step 17752: {'lr': 0.0003784976158441347, 'samples': 3408576, 'steps': 17752, 'loss/train': 0.04111298080533743} +01/27/2022 12:32:57 - INFO - codeparrot_training - Step 17753: {'lr': 0.0003784835799163547, 'samples': 3408768, 'steps': 17753, 'loss/train': 0.7648500502109528} +01/27/2022 12:33:01 - INFO - codeparrot_training - Step 17754: {'lr': 0.00037846954343819195, 'samples': 3408960, 'steps': 17754, 'loss/train': 0.7031485587358475} +01/27/2022 12:33:04 - INFO - codeparrot_training - Step 17755: {'lr': 0.00037845550640970636, 'samples': 3409152, 'steps': 17755, 'loss/train': 0.6835691928863525} +01/27/2022 12:33:07 - INFO - codeparrot_training - Step 17756: {'lr': 0.0003784414688309583, 'samples': 3409344, 'steps': 17756, 'loss/train': 0.999997079372406} +01/27/2022 12:33:11 - INFO - codeparrot_training - Step 17757: {'lr': 0.00037842743070200767, 'samples': 3409536, 'steps': 17757, 'loss/train': 0.6320207566022873} +01/27/2022 12:33:15 - INFO - codeparrot_training - Step 17758: {'lr': 0.0003784133920229148, 'samples': 3409728, 'steps': 17758, 'loss/train': 1.1022075712680817} +01/27/2022 12:33:18 - INFO - codeparrot_training - Step 17759: {'lr': 0.0003783993527937397, 'samples': 3409920, 'steps': 17759, 'loss/train': 1.1408203840255737} +01/27/2022 12:33:21 - INFO - codeparrot_training - Step 17760: {'lr': 0.0003783853130145425, 'samples': 3410112, 'steps': 17760, 'loss/train': 0.6201674938201904} +01/27/2022 12:33:24 - INFO - codeparrot_training - Step 17761: {'lr': 0.0003783712726853835, 'samples': 3410304, 'steps': 17761, 'loss/train': 0.5229577124118805} +01/27/2022 12:33:27 - INFO - codeparrot_training - Step 17762: {'lr': 0.00037835723180632263, 'samples': 3410496, 'steps': 17762, 'loss/train': 0.9081635177135468} +01/27/2022 12:33:30 - INFO - codeparrot_training - Step 17763: {'lr': 0.00037834319037742016, 'samples': 3410688, 'steps': 17763, 'loss/train': 0.42422300577163696} +01/27/2022 12:33:33 - INFO - codeparrot_training - Step 17764: {'lr': 0.00037832914839873623, 'samples': 3410880, 'steps': 17764, 'loss/train': 0.7212165892124176} +01/27/2022 12:33:40 - INFO - codeparrot_training - Step 17765: {'lr': 0.0003783151058703309, 'samples': 3411072, 'steps': 17765, 'loss/train': 1.0867212116718292} +01/27/2022 12:33:43 - INFO - codeparrot_training - Step 17766: {'lr': 0.0003783010627922645, 'samples': 3411264, 'steps': 17766, 'loss/train': 1.6776471734046936} +01/27/2022 12:33:46 - INFO - codeparrot_training - Step 17767: {'lr': 0.0003782870191645971, 'samples': 3411456, 'steps': 17767, 'loss/train': 0.5070259273052216} +01/27/2022 12:33:49 - INFO - codeparrot_training - Step 17768: {'lr': 0.0003782729749873887, 'samples': 3411648, 'steps': 17768, 'loss/train': 0.8707565367221832} +01/27/2022 12:33:53 - INFO - codeparrot_training - Step 17769: {'lr': 0.00037825893026069977, 'samples': 3411840, 'steps': 17769, 'loss/train': 1.1111523807048798} +01/27/2022 12:33:56 - INFO - codeparrot_training - Step 17770: {'lr': 0.0003782448849845902, 'samples': 3412032, 'steps': 17770, 'loss/train': 0.8579094707965851} +01/27/2022 12:33:59 - INFO - codeparrot_training - Step 17771: {'lr': 0.0003782308391591203, 'samples': 3412224, 'steps': 17771, 'loss/train': 0.7559394836425781} +01/27/2022 12:34:02 - INFO - codeparrot_training - Step 17772: {'lr': 0.00037821679278435017, 'samples': 3412416, 'steps': 17772, 'loss/train': 1.293379157781601} +01/27/2022 12:34:05 - INFO - codeparrot_training - Step 17773: {'lr': 0.0003782027458603401, 'samples': 3412608, 'steps': 17773, 'loss/train': 1.2864887416362762} +01/27/2022 12:34:10 - INFO - codeparrot_training - Step 17774: {'lr': 0.0003781886983871501, 'samples': 3412800, 'steps': 17774, 'loss/train': 0.8114047050476074} +01/27/2022 12:34:13 - INFO - codeparrot_training - Step 17775: {'lr': 0.00037817465036484043, 'samples': 3412992, 'steps': 17775, 'loss/train': 0.9736630618572235} +01/27/2022 12:34:16 - INFO - codeparrot_training - Step 17776: {'lr': 0.0003781606017934713, 'samples': 3413184, 'steps': 17776, 'loss/train': 0.747618243098259} +01/27/2022 12:34:19 - INFO - codeparrot_training - Step 17777: {'lr': 0.0003781465526731028, 'samples': 3413376, 'steps': 17777, 'loss/train': 0.7631416618824005} +01/27/2022 12:34:22 - INFO - codeparrot_training - Step 17778: {'lr': 0.0003781325030037952, 'samples': 3413568, 'steps': 17778, 'loss/train': 0.38618794083595276} +01/27/2022 12:34:25 - INFO - codeparrot_training - Step 17779: {'lr': 0.00037811845278560864, 'samples': 3413760, 'steps': 17779, 'loss/train': 0.606358990073204} +01/27/2022 12:34:28 - INFO - codeparrot_training - Step 17780: {'lr': 0.0003781044020186033, 'samples': 3413952, 'steps': 17780, 'loss/train': 0.17154952883720398} +01/27/2022 12:34:31 - INFO - codeparrot_training - Step 17781: {'lr': 0.0003780903507028393, 'samples': 3414144, 'steps': 17781, 'loss/train': 0.9237264096736908} +01/27/2022 12:34:35 - INFO - codeparrot_training - Step 17782: {'lr': 0.00037807629883837703, 'samples': 3414336, 'steps': 17782, 'loss/train': 0.40533529222011566} +01/27/2022 12:34:40 - INFO - codeparrot_training - Step 17783: {'lr': 0.00037806224642527653, 'samples': 3414528, 'steps': 17783, 'loss/train': 0.7241041213274002} +01/27/2022 12:34:43 - INFO - codeparrot_training - Step 17784: {'lr': 0.000378048193463598, 'samples': 3414720, 'steps': 17784, 'loss/train': 0.8158173859119415} +01/27/2022 12:34:46 - INFO - codeparrot_training - Step 17785: {'lr': 0.0003780341399534017, 'samples': 3414912, 'steps': 17785, 'loss/train': 0.7599489390850067} +01/27/2022 12:34:50 - INFO - codeparrot_training - Step 17786: {'lr': 0.00037802008589474777, 'samples': 3415104, 'steps': 17786, 'loss/train': 1.3232737183570862} +01/27/2022 12:34:53 - INFO - codeparrot_training - Step 17787: {'lr': 0.0003780060312876965, 'samples': 3415296, 'steps': 17787, 'loss/train': 0.8640496730804443} +01/27/2022 12:34:56 - INFO - codeparrot_training - Step 17788: {'lr': 0.00037799197613230795, 'samples': 3415488, 'steps': 17788, 'loss/train': 0.9790913164615631} +01/27/2022 12:34:59 - INFO - codeparrot_training - Step 17789: {'lr': 0.00037797792042864247, 'samples': 3415680, 'steps': 17789, 'loss/train': 0.9319522082805634} +01/27/2022 12:35:02 - INFO - codeparrot_training - Step 17790: {'lr': 0.0003779638641767602, 'samples': 3415872, 'steps': 17790, 'loss/train': 0.44592931866645813} +01/27/2022 12:35:05 - INFO - codeparrot_training - Step 17791: {'lr': 0.0003779498073767214, 'samples': 3416064, 'steps': 17791, 'loss/train': 0.9241170287132263} +01/27/2022 12:35:10 - INFO - codeparrot_training - Step 17792: {'lr': 0.00037793575002858625, 'samples': 3416256, 'steps': 17792, 'loss/train': 1.2508339583873749} +01/27/2022 12:35:13 - INFO - codeparrot_training - Step 17793: {'lr': 0.00037792169213241494, 'samples': 3416448, 'steps': 17793, 'loss/train': 0.818234771490097} +01/27/2022 12:35:16 - INFO - codeparrot_training - Step 17794: {'lr': 0.00037790763368826774, 'samples': 3416640, 'steps': 17794, 'loss/train': 1.0224229395389557} +01/27/2022 12:35:19 - INFO - codeparrot_training - Step 17795: {'lr': 0.00037789357469620487, 'samples': 3416832, 'steps': 17795, 'loss/train': 0.9671560227870941} +01/27/2022 12:35:22 - INFO - codeparrot_training - Step 17796: {'lr': 0.0003778795151562865, 'samples': 3417024, 'steps': 17796, 'loss/train': 0.18831110000610352} +01/27/2022 12:35:25 - INFO - codeparrot_training - Step 17797: {'lr': 0.00037786545506857295, 'samples': 3417216, 'steps': 17797, 'loss/train': 0.7587812840938568} +01/27/2022 12:35:29 - INFO - codeparrot_training - Step 17798: {'lr': 0.0003778513944331243, 'samples': 3417408, 'steps': 17798, 'loss/train': 1.1899133026599884} +01/27/2022 12:35:32 - INFO - codeparrot_training - Step 17799: {'lr': 0.0003778373332500009, 'samples': 3417600, 'steps': 17799, 'loss/train': 0.8800412714481354} +01/27/2022 12:35:35 - INFO - codeparrot_training - Step 17800: {'lr': 0.00037782327151926297, 'samples': 3417792, 'steps': 17800, 'loss/train': 0.3728077858686447} +01/27/2022 12:35:39 - INFO - codeparrot_training - Step 17801: {'lr': 0.00037780920924097085, 'samples': 3417984, 'steps': 17801, 'loss/train': 0.7881158888339996} +01/27/2022 12:35:42 - INFO - codeparrot_training - Step 17802: {'lr': 0.00037779514641518455, 'samples': 3418176, 'steps': 17802, 'loss/train': 0.8491335511207581} +01/27/2022 12:35:46 - INFO - codeparrot_training - Step 17803: {'lr': 0.0003777810830419644, 'samples': 3418368, 'steps': 17803, 'loss/train': 0.9170521795749664} +01/27/2022 12:35:49 - INFO - codeparrot_training - Step 17804: {'lr': 0.00037776701912137066, 'samples': 3418560, 'steps': 17804, 'loss/train': 0.8212615549564362} +01/27/2022 12:35:52 - INFO - codeparrot_training - Step 17805: {'lr': 0.00037775295465346373, 'samples': 3418752, 'steps': 17805, 'loss/train': 0.5756540447473526} +01/27/2022 12:35:55 - INFO - codeparrot_training - Step 17806: {'lr': 0.0003777388896383035, 'samples': 3418944, 'steps': 17806, 'loss/train': 1.1730384528636932} +01/27/2022 12:35:58 - INFO - codeparrot_training - Step 17807: {'lr': 0.00037772482407595056, 'samples': 3419136, 'steps': 17807, 'loss/train': 0.7193736881017685} +01/27/2022 12:36:01 - INFO - codeparrot_training - Step 17808: {'lr': 0.000377710757966465, 'samples': 3419328, 'steps': 17808, 'loss/train': 0.8357550501823425} +01/27/2022 12:36:04 - INFO - codeparrot_training - Step 17809: {'lr': 0.0003776966913099071, 'samples': 3419520, 'steps': 17809, 'loss/train': 0.6030236184597015} +01/27/2022 12:36:09 - INFO - codeparrot_training - Step 17810: {'lr': 0.00037768262410633715, 'samples': 3419712, 'steps': 17810, 'loss/train': 1.5301258563995361} +01/27/2022 12:36:12 - INFO - codeparrot_training - Step 17811: {'lr': 0.0003776685563558153, 'samples': 3419904, 'steps': 17811, 'loss/train': 0.6363756358623505} +01/27/2022 12:36:15 - INFO - codeparrot_training - Step 17812: {'lr': 0.00037765448805840196, 'samples': 3420096, 'steps': 17812, 'loss/train': 0.8167665302753448} +01/27/2022 12:36:18 - INFO - codeparrot_training - Step 17813: {'lr': 0.00037764041921415736, 'samples': 3420288, 'steps': 17813, 'loss/train': 5.503037810325623} +01/27/2022 12:36:22 - INFO - codeparrot_training - Step 17814: {'lr': 0.00037762634982314164, 'samples': 3420480, 'steps': 17814, 'loss/train': 0.6128091663122177} +01/27/2022 12:36:25 - INFO - codeparrot_training - Step 17815: {'lr': 0.00037761227988541523, 'samples': 3420672, 'steps': 17815, 'loss/train': 1.2876260876655579} +01/27/2022 12:36:28 - INFO - codeparrot_training - Step 17816: {'lr': 0.00037759820940103827, 'samples': 3420864, 'steps': 17816, 'loss/train': 0.6718382835388184} +01/27/2022 12:36:31 - INFO - codeparrot_training - Step 17817: {'lr': 0.00037758413837007124, 'samples': 3421056, 'steps': 17817, 'loss/train': 0.7165743559598923} +01/27/2022 12:36:37 - INFO - codeparrot_training - Step 17818: {'lr': 0.0003775700667925741, 'samples': 3421248, 'steps': 17818, 'loss/train': 0.8752159774303436} +01/27/2022 12:36:40 - INFO - codeparrot_training - Step 17819: {'lr': 0.0003775559946686075, 'samples': 3421440, 'steps': 17819, 'loss/train': 0.968622475862503} +01/27/2022 12:36:43 - INFO - codeparrot_training - Step 17820: {'lr': 0.00037754192199823135, 'samples': 3421632, 'steps': 17820, 'loss/train': 1.2922334969043732} +01/27/2022 12:36:46 - INFO - codeparrot_training - Step 17821: {'lr': 0.00037752784878150613, 'samples': 3421824, 'steps': 17821, 'loss/train': 0.7727307379245758} +01/27/2022 12:36:49 - INFO - codeparrot_training - Step 17822: {'lr': 0.00037751377501849215, 'samples': 3422016, 'steps': 17822, 'loss/train': 0.8676148653030396} +01/27/2022 12:36:53 - INFO - codeparrot_training - Step 17823: {'lr': 0.0003774997007092496, 'samples': 3422208, 'steps': 17823, 'loss/train': 0.9425946772098541} +01/27/2022 12:36:56 - INFO - codeparrot_training - Step 17824: {'lr': 0.00037748562585383886, 'samples': 3422400, 'steps': 17824, 'loss/train': 0.9956670105457306} +01/27/2022 12:36:59 - INFO - codeparrot_training - Step 17825: {'lr': 0.00037747155045232016, 'samples': 3422592, 'steps': 17825, 'loss/train': 0.7316556125879288} +01/27/2022 12:37:02 - INFO - codeparrot_training - Step 17826: {'lr': 0.0003774574745047539, 'samples': 3422784, 'steps': 17826, 'loss/train': 0.9868363738059998} +01/27/2022 12:37:05 - INFO - codeparrot_training - Step 17827: {'lr': 0.0003774433980112001, 'samples': 3422976, 'steps': 17827, 'loss/train': 0.16220640018582344} +01/27/2022 12:37:10 - INFO - codeparrot_training - Step 17828: {'lr': 0.00037742932097171945, 'samples': 3423168, 'steps': 17828, 'loss/train': 0.806742399930954} +01/27/2022 12:37:13 - INFO - codeparrot_training - Step 17829: {'lr': 0.0003774152433863719, 'samples': 3423360, 'steps': 17829, 'loss/train': 0.6970316469669342} +01/27/2022 12:37:16 - INFO - codeparrot_training - Step 17830: {'lr': 0.000377401165255218, 'samples': 3423552, 'steps': 17830, 'loss/train': 0.7486716359853745} +01/27/2022 12:37:19 - INFO - codeparrot_training - Step 17831: {'lr': 0.0003773870865783179, 'samples': 3423744, 'steps': 17831, 'loss/train': 0.48460453748703003} +01/27/2022 12:37:23 - INFO - codeparrot_training - Step 17832: {'lr': 0.00037737300735573204, 'samples': 3423936, 'steps': 17832, 'loss/train': 1.0549482107162476} +01/27/2022 12:37:26 - INFO - codeparrot_training - Step 17833: {'lr': 0.00037735892758752063, 'samples': 3424128, 'steps': 17833, 'loss/train': 1.028437614440918} +01/27/2022 12:37:29 - INFO - codeparrot_training - Step 17834: {'lr': 0.000377344847273744, 'samples': 3424320, 'steps': 17834, 'loss/train': 1.5317743420600891} +01/27/2022 12:37:32 - INFO - codeparrot_training - Step 17835: {'lr': 0.0003773307664144625, 'samples': 3424512, 'steps': 17835, 'loss/train': 1.03023362159729} +01/27/2022 12:37:35 - INFO - codeparrot_training - Step 17836: {'lr': 0.00037731668500973637, 'samples': 3424704, 'steps': 17836, 'loss/train': 0.6799329668283463} +01/27/2022 12:37:40 - INFO - codeparrot_training - Step 17837: {'lr': 0.00037730260305962604, 'samples': 3424896, 'steps': 17837, 'loss/train': 0.8980058133602142} +01/27/2022 12:37:43 - INFO - codeparrot_training - Step 17838: {'lr': 0.00037728852056419183, 'samples': 3425088, 'steps': 17838, 'loss/train': 0.7650082111358643} +01/27/2022 12:37:46 - INFO - codeparrot_training - Step 17839: {'lr': 0.000377274437523494, 'samples': 3425280, 'steps': 17839, 'loss/train': 1.0262062847614288} +01/27/2022 12:37:49 - INFO - codeparrot_training - Step 17840: {'lr': 0.00037726035393759286, 'samples': 3425472, 'steps': 17840, 'loss/train': 1.0618519484996796} +01/27/2022 12:37:52 - INFO - codeparrot_training - Step 17841: {'lr': 0.00037724626980654877, 'samples': 3425664, 'steps': 17841, 'loss/train': 1.1169151067733765} +01/27/2022 12:37:55 - INFO - codeparrot_training - Step 17842: {'lr': 0.00037723218513042203, 'samples': 3425856, 'steps': 17842, 'loss/train': 0.891390323638916} +01/27/2022 12:37:59 - INFO - codeparrot_training - Step 17843: {'lr': 0.0003772180999092731, 'samples': 3426048, 'steps': 17843, 'loss/train': 1.1056045889854431} +01/27/2022 12:38:02 - INFO - codeparrot_training - Step 17844: {'lr': 0.00037720401414316213, 'samples': 3426240, 'steps': 17844, 'loss/train': 1.03940749168396} +01/27/2022 12:38:07 - INFO - codeparrot_training - Step 17845: {'lr': 0.00037718992783214965, 'samples': 3426432, 'steps': 17845, 'loss/train': 0.9667668342590332} +01/27/2022 12:38:11 - INFO - codeparrot_training - Step 17846: {'lr': 0.0003771758409762958, 'samples': 3426624, 'steps': 17846, 'loss/train': 0.3835897296667099} +01/27/2022 12:38:14 - INFO - codeparrot_training - Step 17847: {'lr': 0.0003771617535756611, 'samples': 3426816, 'steps': 17847, 'loss/train': 0.45386382937431335} +01/27/2022 12:38:17 - INFO - codeparrot_training - Step 17848: {'lr': 0.00037714766563030585, 'samples': 3427008, 'steps': 17848, 'loss/train': 1.0644161403179169} +01/27/2022 12:38:20 - INFO - codeparrot_training - Step 17849: {'lr': 0.00037713357714029035, 'samples': 3427200, 'steps': 17849, 'loss/train': 0.8950679898262024} +01/27/2022 12:38:23 - INFO - codeparrot_training - Step 17850: {'lr': 0.000377119488105675, 'samples': 3427392, 'steps': 17850, 'loss/train': 0.5049117654561996} +01/27/2022 12:38:26 - INFO - codeparrot_training - Step 17851: {'lr': 0.00037710539852652003, 'samples': 3427584, 'steps': 17851, 'loss/train': 1.0289808511734009} +01/27/2022 12:38:30 - INFO - codeparrot_training - Step 17852: {'lr': 0.00037709130840288605, 'samples': 3427776, 'steps': 17852, 'loss/train': 0.047231873497366905} +01/27/2022 12:38:33 - INFO - codeparrot_training - Step 17853: {'lr': 0.0003770772177348331, 'samples': 3427968, 'steps': 17853, 'loss/train': 0.7396425157785416} +01/27/2022 12:38:37 - INFO - codeparrot_training - Step 17854: {'lr': 0.0003770631265224218, 'samples': 3428160, 'steps': 17854, 'loss/train': 0.47764505445957184} +01/27/2022 12:38:40 - INFO - codeparrot_training - Step 17855: {'lr': 0.0003770490347657124, 'samples': 3428352, 'steps': 17855, 'loss/train': 0.8726488351821899} +01/27/2022 12:38:43 - INFO - codeparrot_training - Step 17856: {'lr': 0.00037703494246476524, 'samples': 3428544, 'steps': 17856, 'loss/train': 0.7103675007820129} +01/27/2022 12:38:47 - INFO - codeparrot_training - Step 17857: {'lr': 0.00037702084961964075, 'samples': 3428736, 'steps': 17857, 'loss/train': 0.842792272567749} +01/27/2022 12:38:50 - INFO - codeparrot_training - Step 17858: {'lr': 0.00037700675623039925, 'samples': 3428928, 'steps': 17858, 'loss/train': 0.7852815985679626} +01/27/2022 12:38:53 - INFO - codeparrot_training - Step 17859: {'lr': 0.00037699266229710115, 'samples': 3429120, 'steps': 17859, 'loss/train': 0.5865557491779327} +01/27/2022 12:38:56 - INFO - codeparrot_training - Step 17860: {'lr': 0.0003769785678198068, 'samples': 3429312, 'steps': 17860, 'loss/train': 0.5531609952449799} +01/27/2022 12:38:59 - INFO - codeparrot_training - Step 17861: {'lr': 0.0003769644727985766, 'samples': 3429504, 'steps': 17861, 'loss/train': 0.8094476759433746} +01/27/2022 12:39:02 - INFO - codeparrot_training - Step 17862: {'lr': 0.00037695037723347094, 'samples': 3429696, 'steps': 17862, 'loss/train': 0.43270061910152435} +01/27/2022 12:39:08 - INFO - codeparrot_training - Step 17863: {'lr': 0.00037693628112455015, 'samples': 3429888, 'steps': 17863, 'loss/train': 0.7407862544059753} +01/27/2022 12:39:11 - INFO - codeparrot_training - Step 17864: {'lr': 0.0003769221844718746, 'samples': 3430080, 'steps': 17864, 'loss/train': 0.7080651372671127} +01/27/2022 12:39:14 - INFO - codeparrot_training - Step 17865: {'lr': 0.00037690808727550477, 'samples': 3430272, 'steps': 17865, 'loss/train': 0.7394076436758041} +01/27/2022 12:39:17 - INFO - codeparrot_training - Step 17866: {'lr': 0.0003768939895355009, 'samples': 3430464, 'steps': 17866, 'loss/train': 0.3710573762655258} +01/27/2022 12:39:21 - INFO - codeparrot_training - Step 17867: {'lr': 0.0003768798912519236, 'samples': 3430656, 'steps': 17867, 'loss/train': 0.8106890916824341} +01/27/2022 12:39:24 - INFO - codeparrot_training - Step 17868: {'lr': 0.0003768657924248331, 'samples': 3430848, 'steps': 17868, 'loss/train': 0.7833350300788879} +01/27/2022 12:39:27 - INFO - codeparrot_training - Step 17869: {'lr': 0.0003768516930542898, 'samples': 3431040, 'steps': 17869, 'loss/train': 0.9321344196796417} +01/27/2022 12:39:30 - INFO - codeparrot_training - Step 17870: {'lr': 0.00037683759314035414, 'samples': 3431232, 'steps': 17870, 'loss/train': 0.7998980283737183} +01/27/2022 12:39:33 - INFO - codeparrot_training - Step 17871: {'lr': 0.0003768234926830865, 'samples': 3431424, 'steps': 17871, 'loss/train': 0.7183521240949631} +01/27/2022 12:39:38 - INFO - codeparrot_training - Step 17872: {'lr': 0.0003768093916825473, 'samples': 3431616, 'steps': 17872, 'loss/train': 0.7743760049343109} +01/27/2022 12:39:41 - INFO - codeparrot_training - Step 17873: {'lr': 0.00037679529013879686, 'samples': 3431808, 'steps': 17873, 'loss/train': 1.0870539844036102} +01/27/2022 12:39:44 - INFO - codeparrot_training - Step 17874: {'lr': 0.00037678118805189575, 'samples': 3432000, 'steps': 17874, 'loss/train': 0.9091189205646515} +01/27/2022 12:39:47 - INFO - codeparrot_training - Step 17875: {'lr': 0.0003767670854219043, 'samples': 3432192, 'steps': 17875, 'loss/train': 1.3348017632961273} +01/27/2022 12:39:50 - INFO - codeparrot_training - Step 17876: {'lr': 0.00037675298224888287, 'samples': 3432384, 'steps': 17876, 'loss/train': 1.0138724148273468} +01/27/2022 12:39:53 - INFO - codeparrot_training - Step 17877: {'lr': 0.0003767388785328919, 'samples': 3432576, 'steps': 17877, 'loss/train': 0.8898642361164093} +01/27/2022 12:39:56 - INFO - codeparrot_training - Step 17878: {'lr': 0.0003767247742739918, 'samples': 3432768, 'steps': 17878, 'loss/train': 0.4390469491481781} +01/27/2022 12:40:00 - INFO - codeparrot_training - Step 17879: {'lr': 0.0003767106694722431, 'samples': 3432960, 'steps': 17879, 'loss/train': 0.5408623069524765} +01/27/2022 12:40:04 - INFO - codeparrot_training - Step 17880: {'lr': 0.000376696564127706, 'samples': 3433152, 'steps': 17880, 'loss/train': 0.7494834065437317} +01/27/2022 12:40:07 - INFO - codeparrot_training - Step 17881: {'lr': 0.0003766824582404411, 'samples': 3433344, 'steps': 17881, 'loss/train': 0.8861905932426453} +01/27/2022 12:40:10 - INFO - codeparrot_training - Step 17882: {'lr': 0.00037666835181050887, 'samples': 3433536, 'steps': 17882, 'loss/train': 1.2920831143856049} +01/27/2022 12:40:14 - INFO - codeparrot_training - Step 17883: {'lr': 0.0003766542448379695, 'samples': 3433728, 'steps': 17883, 'loss/train': 0.8060343861579895} +01/27/2022 12:40:17 - INFO - codeparrot_training - Step 17884: {'lr': 0.0003766401373228836, 'samples': 3433920, 'steps': 17884, 'loss/train': 0.8285459876060486} +01/27/2022 12:40:20 - INFO - codeparrot_training - Step 17885: {'lr': 0.00037662602926531166, 'samples': 3434112, 'steps': 17885, 'loss/train': 0.9514822661876678} +01/27/2022 12:40:23 - INFO - codeparrot_training - Step 17886: {'lr': 0.0003766119206653139, 'samples': 3434304, 'steps': 17886, 'loss/train': 0.1768232323229313} +01/27/2022 12:40:26 - INFO - codeparrot_training - Step 17887: {'lr': 0.00037659781152295094, 'samples': 3434496, 'steps': 17887, 'loss/train': 0.5970326364040375} +01/27/2022 12:40:29 - INFO - codeparrot_training - Step 17888: {'lr': 0.0003765837018382831, 'samples': 3434688, 'steps': 17888, 'loss/train': 0.8003620505332947} +01/27/2022 12:40:35 - INFO - codeparrot_training - Step 17889: {'lr': 0.00037656959161137094, 'samples': 3434880, 'steps': 17889, 'loss/train': 0.8234326243400574} +01/27/2022 12:40:38 - INFO - codeparrot_training - Step 17890: {'lr': 0.00037655548084227484, 'samples': 3435072, 'steps': 17890, 'loss/train': 0.6930548250675201} +01/27/2022 12:40:41 - INFO - codeparrot_training - Step 17891: {'lr': 0.0003765413695310552, 'samples': 3435264, 'steps': 17891, 'loss/train': 0.8977358043193817} +01/27/2022 12:40:45 - INFO - codeparrot_training - Step 17892: {'lr': 0.00037652725767777255, 'samples': 3435456, 'steps': 17892, 'loss/train': 0.42084574699401855} +01/27/2022 12:40:48 - INFO - codeparrot_training - Step 17893: {'lr': 0.00037651314528248724, 'samples': 3435648, 'steps': 17893, 'loss/train': 0.40352632105350494} +01/27/2022 12:40:51 - INFO - codeparrot_training - Step 17894: {'lr': 0.00037649903234525996, 'samples': 3435840, 'steps': 17894, 'loss/train': 1.012834757566452} +01/27/2022 12:40:54 - INFO - codeparrot_training - Step 17895: {'lr': 0.00037648491886615077, 'samples': 3436032, 'steps': 17895, 'loss/train': 1.1155715882778168} +01/27/2022 12:40:57 - INFO - codeparrot_training - Step 17896: {'lr': 0.0003764708048452205, 'samples': 3436224, 'steps': 17896, 'loss/train': 0.9174368977546692} +01/27/2022 12:41:00 - INFO - codeparrot_training - Step 17897: {'lr': 0.0003764566902825294, 'samples': 3436416, 'steps': 17897, 'loss/train': 0.7664187848567963} +01/27/2022 12:41:03 - INFO - codeparrot_training - Step 17898: {'lr': 0.0003764425751781381, 'samples': 3436608, 'steps': 17898, 'loss/train': 0.6826503574848175} +01/27/2022 12:41:08 - INFO - codeparrot_training - Step 17899: {'lr': 0.0003764284595321068, 'samples': 3436800, 'steps': 17899, 'loss/train': 0.7936385571956635} +01/27/2022 12:41:11 - INFO - codeparrot_training - Step 17900: {'lr': 0.0003764143433444962, 'samples': 3436992, 'steps': 17900, 'loss/train': 0.6782476007938385} +01/27/2022 12:41:15 - INFO - codeparrot_training - Step 17901: {'lr': 0.00037640022661536665, 'samples': 3437184, 'steps': 17901, 'loss/train': 0.9428253471851349} +01/27/2022 12:41:18 - INFO - codeparrot_training - Step 17902: {'lr': 0.0003763861093447787, 'samples': 3437376, 'steps': 17902, 'loss/train': 1.848872423171997} +01/27/2022 12:41:21 - INFO - codeparrot_training - Step 17903: {'lr': 0.0003763719915327928, 'samples': 3437568, 'steps': 17903, 'loss/train': 0.9533161818981171} +01/27/2022 12:41:24 - INFO - codeparrot_training - Step 17904: {'lr': 0.00037635787317946945, 'samples': 3437760, 'steps': 17904, 'loss/train': 0.5548265129327774} +01/27/2022 12:41:27 - INFO - codeparrot_training - Step 17905: {'lr': 0.000376343754284869, 'samples': 3437952, 'steps': 17905, 'loss/train': 0.9358738660812378} +01/27/2022 12:41:30 - INFO - codeparrot_training - Step 17906: {'lr': 0.00037632963484905213, 'samples': 3438144, 'steps': 17906, 'loss/train': 1.0115763545036316} +01/27/2022 12:41:35 - INFO - codeparrot_training - Step 17907: {'lr': 0.0003763155148720791, 'samples': 3438336, 'steps': 17907, 'loss/train': 0.6282755881547928} +01/27/2022 12:41:38 - INFO - codeparrot_training - Step 17908: {'lr': 0.00037630139435401055, 'samples': 3438528, 'steps': 17908, 'loss/train': 0.8558928072452545} +01/27/2022 12:41:41 - INFO - codeparrot_training - Step 17909: {'lr': 0.000376287273294907, 'samples': 3438720, 'steps': 17909, 'loss/train': 1.0716103613376617} +01/27/2022 12:41:44 - INFO - codeparrot_training - Step 17910: {'lr': 0.0003762731516948288, 'samples': 3438912, 'steps': 17910, 'loss/train': 0.6623565852642059} +01/27/2022 12:41:47 - INFO - codeparrot_training - Step 17911: {'lr': 0.00037625902955383664, 'samples': 3439104, 'steps': 17911, 'loss/train': 1.1786466836929321} +01/27/2022 12:41:51 - INFO - codeparrot_training - Step 17912: {'lr': 0.0003762449068719907, 'samples': 3439296, 'steps': 17912, 'loss/train': 0.8526237308979034} +01/27/2022 12:41:54 - INFO - codeparrot_training - Step 17913: {'lr': 0.0003762307836493518, 'samples': 3439488, 'steps': 17913, 'loss/train': 0.711031123995781} +01/27/2022 12:41:57 - INFO - codeparrot_training - Step 17914: {'lr': 0.00037621665988598024, 'samples': 3439680, 'steps': 17914, 'loss/train': 1.1132463812828064} +01/27/2022 12:42:00 - INFO - codeparrot_training - Step 17915: {'lr': 0.0003762025355819366, 'samples': 3439872, 'steps': 17915, 'loss/train': 0.9389703869819641} +01/27/2022 12:42:04 - INFO - codeparrot_training - Step 17916: {'lr': 0.0003761884107372814, 'samples': 3440064, 'steps': 17916, 'loss/train': 1.1480253338813782} +01/27/2022 12:42:07 - INFO - codeparrot_training - Step 17917: {'lr': 0.0003761742853520751, 'samples': 3440256, 'steps': 17917, 'loss/train': 0.64650097489357} +01/27/2022 12:42:11 - INFO - codeparrot_training - Step 17918: {'lr': 0.00037616015942637824, 'samples': 3440448, 'steps': 17918, 'loss/train': 0.7604730427265167} +01/27/2022 12:42:14 - INFO - codeparrot_training - Step 17919: {'lr': 0.0003761460329602513, 'samples': 3440640, 'steps': 17919, 'loss/train': 0.5834275335073471} +01/27/2022 12:42:17 - INFO - codeparrot_training - Step 17920: {'lr': 0.0003761319059537548, 'samples': 3440832, 'steps': 17920, 'loss/train': 0.9251265227794647} +01/27/2022 12:42:20 - INFO - codeparrot_training - Step 17921: {'lr': 0.0003761177784069493, 'samples': 3441024, 'steps': 17921, 'loss/train': 0.4388592392206192} +01/27/2022 12:42:23 - INFO - codeparrot_training - Step 17922: {'lr': 0.00037610365031989524, 'samples': 3441216, 'steps': 17922, 'loss/train': 1.2855013310909271} +01/27/2022 12:42:26 - INFO - codeparrot_training - Step 17923: {'lr': 0.0003760895216926532, 'samples': 3441408, 'steps': 17923, 'loss/train': 1.125196248292923} +01/27/2022 12:42:30 - INFO - codeparrot_training - Step 17924: {'lr': 0.0003760753925252838, 'samples': 3441600, 'steps': 17924, 'loss/train': 0.3592373952269554} +01/27/2022 12:42:36 - INFO - codeparrot_training - Step 17925: {'lr': 0.00037606126281784725, 'samples': 3441792, 'steps': 17925, 'loss/train': 0.8088610768318176} +01/27/2022 12:42:39 - INFO - codeparrot_training - Step 17926: {'lr': 0.0003760471325704045, 'samples': 3441984, 'steps': 17926, 'loss/train': 0.7711814045906067} +01/27/2022 12:42:42 - INFO - codeparrot_training - Step 17927: {'lr': 0.0003760330017830157, 'samples': 3442176, 'steps': 17927, 'loss/train': 1.3911140263080597} +01/27/2022 12:42:45 - INFO - codeparrot_training - Step 17928: {'lr': 0.00037601887045574155, 'samples': 3442368, 'steps': 17928, 'loss/train': 0.8131018280982971} +01/27/2022 12:42:48 - INFO - codeparrot_training - Step 17929: {'lr': 0.0003760047385886426, 'samples': 3442560, 'steps': 17929, 'loss/train': 0.23695486038923264} +01/27/2022 12:42:52 - INFO - codeparrot_training - Step 17930: {'lr': 0.0003759906061817794, 'samples': 3442752, 'steps': 17930, 'loss/train': 0.27896542847156525} +01/27/2022 12:42:55 - INFO - codeparrot_training - Step 17931: {'lr': 0.00037597647323521234, 'samples': 3442944, 'steps': 17931, 'loss/train': 1.0130764245986938} +01/27/2022 12:42:58 - INFO - codeparrot_training - Step 17932: {'lr': 0.0003759623397490022, 'samples': 3443136, 'steps': 17932, 'loss/train': 0.7405153959989548} +01/27/2022 12:43:01 - INFO - codeparrot_training - Step 17933: {'lr': 0.00037594820572320933, 'samples': 3443328, 'steps': 17933, 'loss/train': 0.7939205467700958} +01/27/2022 12:43:05 - INFO - codeparrot_training - Step 17934: {'lr': 0.0003759340711578944, 'samples': 3443520, 'steps': 17934, 'loss/train': 0.5022717118263245} +01/27/2022 12:43:08 - INFO - codeparrot_training - Step 17935: {'lr': 0.0003759199360531178, 'samples': 3443712, 'steps': 17935, 'loss/train': 1.0045287609100342} +01/27/2022 12:43:12 - INFO - codeparrot_training - Step 17936: {'lr': 0.00037590580040894024, 'samples': 3443904, 'steps': 17936, 'loss/train': 0.8051446080207825} +01/27/2022 12:43:15 - INFO - codeparrot_training - Step 17937: {'lr': 0.0003758916642254222, 'samples': 3444096, 'steps': 17937, 'loss/train': 0.9203100800514221} +01/27/2022 12:43:18 - INFO - codeparrot_training - Step 17938: {'lr': 0.00037587752750262426, 'samples': 3444288, 'steps': 17938, 'loss/train': 0.30617518723011017} +01/27/2022 12:43:21 - INFO - codeparrot_training - Step 17939: {'lr': 0.00037586339024060696, 'samples': 3444480, 'steps': 17939, 'loss/train': 0.6931801289319992} +01/27/2022 12:43:24 - INFO - codeparrot_training - Step 17940: {'lr': 0.0003758492524394308, 'samples': 3444672, 'steps': 17940, 'loss/train': 0.46619197726249695} +01/27/2022 12:43:27 - INFO - codeparrot_training - Step 17941: {'lr': 0.0003758351140991565, 'samples': 3444864, 'steps': 17941, 'loss/train': 0.8829436898231506} +01/27/2022 12:43:33 - INFO - codeparrot_training - Step 17942: {'lr': 0.0003758209752198444, 'samples': 3445056, 'steps': 17942, 'loss/train': 0.6434184908866882} +01/27/2022 12:43:37 - INFO - codeparrot_training - Step 17943: {'lr': 0.0003758068358015553, 'samples': 3445248, 'steps': 17943, 'loss/train': 0.48153796792030334} +01/27/2022 12:43:40 - INFO - codeparrot_training - Step 17944: {'lr': 0.0003757926958443496, 'samples': 3445440, 'steps': 17944, 'loss/train': 0.7031002342700958} +01/27/2022 12:43:43 - INFO - codeparrot_training - Step 17945: {'lr': 0.000375778555348288, 'samples': 3445632, 'steps': 17945, 'loss/train': 1.113799273967743} +01/27/2022 12:43:46 - INFO - codeparrot_training - Step 17946: {'lr': 0.000375764414313431, 'samples': 3445824, 'steps': 17946, 'loss/train': 0.6074004918336868} +01/27/2022 12:43:49 - INFO - codeparrot_training - Step 17947: {'lr': 0.0003757502727398391, 'samples': 3446016, 'steps': 17947, 'loss/train': 0.9972122311592102} +01/27/2022 12:43:52 - INFO - codeparrot_training - Step 17948: {'lr': 0.00037573613062757304, 'samples': 3446208, 'steps': 17948, 'loss/train': 0.8115869164466858} +01/27/2022 12:43:55 - INFO - codeparrot_training - Step 17949: {'lr': 0.0003757219879766933, 'samples': 3446400, 'steps': 17949, 'loss/train': 0.9752765893936157} +01/27/2022 12:43:59 - INFO - codeparrot_training - Step 17950: {'lr': 0.00037570784478726057, 'samples': 3446592, 'steps': 17950, 'loss/train': 0.7239139080047607} +01/27/2022 12:44:03 - INFO - codeparrot_training - Step 17951: {'lr': 0.00037569370105933523, 'samples': 3446784, 'steps': 17951, 'loss/train': 0.3716737926006317} +01/27/2022 12:44:06 - INFO - codeparrot_training - Step 17952: {'lr': 0.00037567955679297806, 'samples': 3446976, 'steps': 17952, 'loss/train': 0.6268345713615417} +01/27/2022 12:44:10 - INFO - codeparrot_training - Step 17953: {'lr': 0.0003756654119882496, 'samples': 3447168, 'steps': 17953, 'loss/train': 0.5763550400733948} +01/27/2022 12:44:13 - INFO - codeparrot_training - Step 17954: {'lr': 0.0003756512666452103, 'samples': 3447360, 'steps': 17954, 'loss/train': 0.262355737388134} +01/27/2022 12:44:16 - INFO - codeparrot_training - Step 17955: {'lr': 0.0003756371207639209, 'samples': 3447552, 'steps': 17955, 'loss/train': 0.9930324554443359} +01/27/2022 12:44:19 - INFO - codeparrot_training - Step 17956: {'lr': 0.00037562297434444203, 'samples': 3447744, 'steps': 17956, 'loss/train': 1.2762583494186401} +01/27/2022 12:44:22 - INFO - codeparrot_training - Step 17957: {'lr': 0.0003756088273868342, 'samples': 3447936, 'steps': 17957, 'loss/train': 0.8031097054481506} +01/27/2022 12:44:25 - INFO - codeparrot_training - Step 17958: {'lr': 0.00037559467989115806, 'samples': 3448128, 'steps': 17958, 'loss/train': 0.5294175893068314} +01/27/2022 12:44:29 - INFO - codeparrot_training - Step 17959: {'lr': 0.00037558053185747416, 'samples': 3448320, 'steps': 17959, 'loss/train': 0.7924456894397736} +01/27/2022 12:44:33 - INFO - codeparrot_training - Step 17960: {'lr': 0.00037556638328584314, 'samples': 3448512, 'steps': 17960, 'loss/train': 0.827648788690567} +01/27/2022 12:44:37 - INFO - codeparrot_training - Step 17961: {'lr': 0.00037555223417632565, 'samples': 3448704, 'steps': 17961, 'loss/train': 0.6637343019247055} +01/27/2022 12:44:40 - INFO - codeparrot_training - Step 17962: {'lr': 0.0003755380845289822, 'samples': 3448896, 'steps': 17962, 'loss/train': 1.1260019838809967} +01/27/2022 12:44:43 - INFO - codeparrot_training - Step 17963: {'lr': 0.0003755239343438735, 'samples': 3449088, 'steps': 17963, 'loss/train': 0.785476416349411} +01/27/2022 12:44:46 - INFO - codeparrot_training - Step 17964: {'lr': 0.00037550978362106, 'samples': 3449280, 'steps': 17964, 'loss/train': 1.199933409690857} +01/27/2022 12:44:49 - INFO - codeparrot_training - Step 17965: {'lr': 0.0003754956323606026, 'samples': 3449472, 'steps': 17965, 'loss/train': 0.7535121738910675} +01/27/2022 12:44:52 - INFO - codeparrot_training - Step 17966: {'lr': 0.0003754814805625617, 'samples': 3449664, 'steps': 17966, 'loss/train': 0.4501775801181793} +01/27/2022 12:44:56 - INFO - codeparrot_training - Step 17967: {'lr': 0.00037546732822699803, 'samples': 3449856, 'steps': 17967, 'loss/train': 0.41887445747852325} +01/27/2022 12:44:59 - INFO - codeparrot_training - Step 17968: {'lr': 0.0003754531753539721, 'samples': 3450048, 'steps': 17968, 'loss/train': 0.5372181683778763} +01/27/2022 12:45:07 - INFO - codeparrot_training - Step 17969: {'lr': 0.0003754390219435446, 'samples': 3450240, 'steps': 17969, 'loss/train': 0.8379674255847931} +01/27/2022 12:45:10 - INFO - codeparrot_training - Step 17970: {'lr': 0.00037542486799577624, 'samples': 3450432, 'steps': 17970, 'loss/train': 0.9128230512142181} +01/27/2022 12:45:13 - INFO - codeparrot_training - Step 17971: {'lr': 0.00037541071351072746, 'samples': 3450624, 'steps': 17971, 'loss/train': 0.5380378067493439} +01/27/2022 12:45:17 - INFO - codeparrot_training - Step 17972: {'lr': 0.0003753965584884591, 'samples': 3450816, 'steps': 17972, 'loss/train': 0.6869356334209442} +01/27/2022 12:45:20 - INFO - codeparrot_training - Step 17973: {'lr': 0.00037538240292903167, 'samples': 3451008, 'steps': 17973, 'loss/train': 0.3067547678947449} +01/27/2022 12:45:23 - INFO - codeparrot_training - Step 17974: {'lr': 0.0003753682468325059, 'samples': 3451200, 'steps': 17974, 'loss/train': 0.6076724678277969} +01/27/2022 12:45:26 - INFO - codeparrot_training - Step 17975: {'lr': 0.0003753540901989422, 'samples': 3451392, 'steps': 17975, 'loss/train': 0.49873723089694977} +01/27/2022 12:45:29 - INFO - codeparrot_training - Step 17976: {'lr': 0.00037533993302840153, 'samples': 3451584, 'steps': 17976, 'loss/train': 0.3396426737308502} +01/27/2022 12:45:32 - INFO - codeparrot_training - Step 17977: {'lr': 0.00037532577532094436, 'samples': 3451776, 'steps': 17977, 'loss/train': 1.7281374335289001} +01/27/2022 12:45:37 - INFO - codeparrot_training - Step 17978: {'lr': 0.00037531161707663136, 'samples': 3451968, 'steps': 17978, 'loss/train': 0.6896446198225021} +01/27/2022 12:45:40 - INFO - codeparrot_training - Step 17979: {'lr': 0.0003752974582955232, 'samples': 3452160, 'steps': 17979, 'loss/train': 0.6871363073587418} +01/27/2022 12:45:43 - INFO - codeparrot_training - Step 17980: {'lr': 0.0003752832989776804, 'samples': 3452352, 'steps': 17980, 'loss/train': 0.7854113280773163} +01/27/2022 12:45:46 - INFO - codeparrot_training - Step 17981: {'lr': 0.0003752691391231639, 'samples': 3452544, 'steps': 17981, 'loss/train': 0.8390191197395325} +01/27/2022 12:45:49 - INFO - codeparrot_training - Step 17982: {'lr': 0.00037525497873203405, 'samples': 3452736, 'steps': 17982, 'loss/train': 2.324696123600006} +01/27/2022 12:45:53 - INFO - codeparrot_training - Step 17983: {'lr': 0.0003752408178043518, 'samples': 3452928, 'steps': 17983, 'loss/train': 0.8827296495437622} +01/27/2022 12:45:56 - INFO - codeparrot_training - Step 17984: {'lr': 0.0003752266563401775, 'samples': 3453120, 'steps': 17984, 'loss/train': 0.8177593946456909} +01/27/2022 12:45:59 - INFO - codeparrot_training - Step 17985: {'lr': 0.00037521249433957203, 'samples': 3453312, 'steps': 17985, 'loss/train': 0.7612828016281128} +01/27/2022 12:46:02 - INFO - codeparrot_training - Step 17986: {'lr': 0.000375198331802596, 'samples': 3453504, 'steps': 17986, 'loss/train': 0.8115473985671997} +01/27/2022 12:46:06 - INFO - codeparrot_training - Step 17987: {'lr': 0.00037518416872931007, 'samples': 3453696, 'steps': 17987, 'loss/train': 1.4162762761116028} +01/27/2022 12:46:10 - INFO - codeparrot_training - Step 17988: {'lr': 0.00037517000511977486, 'samples': 3453888, 'steps': 17988, 'loss/train': 1.028537929058075} +01/27/2022 12:46:13 - INFO - codeparrot_training - Step 17989: {'lr': 0.00037515584097405115, 'samples': 3454080, 'steps': 17989, 'loss/train': 0.6136468201875687} +01/27/2022 12:46:16 - INFO - codeparrot_training - Step 17990: {'lr': 0.00037514167629219955, 'samples': 3454272, 'steps': 17990, 'loss/train': 0.60965596139431} +01/27/2022 12:46:19 - INFO - codeparrot_training - Step 17991: {'lr': 0.0003751275110742807, 'samples': 3454464, 'steps': 17991, 'loss/train': 0.6522223949432373} +01/27/2022 12:46:22 - INFO - codeparrot_training - Step 17992: {'lr': 0.00037511334532035537, 'samples': 3454656, 'steps': 17992, 'loss/train': 0.7448626756668091} +01/27/2022 12:46:25 - INFO - codeparrot_training - Step 17993: {'lr': 0.00037509917903048417, 'samples': 3454848, 'steps': 17993, 'loss/train': 0.7819700539112091} +01/27/2022 12:46:28 - INFO - codeparrot_training - Step 17994: {'lr': 0.00037508501220472783, 'samples': 3455040, 'steps': 17994, 'loss/train': 0.9454674124717712} +01/27/2022 12:46:35 - INFO - codeparrot_training - Step 17995: {'lr': 0.000375070844843147, 'samples': 3455232, 'steps': 17995, 'loss/train': 0.8421874344348907} +01/27/2022 12:46:38 - INFO - codeparrot_training - Step 17996: {'lr': 0.00037505667694580244, 'samples': 3455424, 'steps': 17996, 'loss/train': 1.1058039665222168} +01/27/2022 12:46:41 - INFO - codeparrot_training - Step 17997: {'lr': 0.00037504250851275466, 'samples': 3455616, 'steps': 17997, 'loss/train': 0.9111911952495575} +01/27/2022 12:46:44 - INFO - codeparrot_training - Step 17998: {'lr': 0.0003750283395440647, 'samples': 3455808, 'steps': 17998, 'loss/train': 0.6563953310251236} +01/27/2022 12:46:47 - INFO - codeparrot_training - Step 17999: {'lr': 0.0003750141700397928, 'samples': 3456000, 'steps': 17999, 'loss/train': 0.7130634784698486} +01/27/2022 12:46:47 - INFO - codeparrot_training - Evaluating and saving model checkpoint